1epoch_extra_verified / trainer_state.json
sedrickkeh's picture
End of training
76d851c verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.999815668202765,
"eval_steps": 500,
"global_step": 1808,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0005529953917050691,
"grad_norm": 5.9790191650390625,
"learning_rate": 5.524861878453039e-08,
"loss": 0.8575,
"step": 1
},
{
"epoch": 0.0011059907834101382,
"grad_norm": 5.738432884216309,
"learning_rate": 1.1049723756906078e-07,
"loss": 0.8535,
"step": 2
},
{
"epoch": 0.0016589861751152074,
"grad_norm": 5.905203342437744,
"learning_rate": 1.6574585635359117e-07,
"loss": 0.8907,
"step": 3
},
{
"epoch": 0.0022119815668202765,
"grad_norm": 5.97217321395874,
"learning_rate": 2.2099447513812156e-07,
"loss": 0.8695,
"step": 4
},
{
"epoch": 0.0027649769585253456,
"grad_norm": 6.003969669342041,
"learning_rate": 2.7624309392265196e-07,
"loss": 0.8746,
"step": 5
},
{
"epoch": 0.0033179723502304147,
"grad_norm": 5.878486156463623,
"learning_rate": 3.3149171270718233e-07,
"loss": 0.866,
"step": 6
},
{
"epoch": 0.003870967741935484,
"grad_norm": 5.74072265625,
"learning_rate": 3.867403314917127e-07,
"loss": 0.8351,
"step": 7
},
{
"epoch": 0.004423963133640553,
"grad_norm": 5.6800150871276855,
"learning_rate": 4.419889502762431e-07,
"loss": 0.8534,
"step": 8
},
{
"epoch": 0.004976958525345622,
"grad_norm": 5.634627819061279,
"learning_rate": 4.972375690607735e-07,
"loss": 0.8592,
"step": 9
},
{
"epoch": 0.005529953917050691,
"grad_norm": 5.71673059463501,
"learning_rate": 5.524861878453039e-07,
"loss": 0.8825,
"step": 10
},
{
"epoch": 0.00608294930875576,
"grad_norm": 5.533100605010986,
"learning_rate": 6.077348066298343e-07,
"loss": 0.8662,
"step": 11
},
{
"epoch": 0.0066359447004608295,
"grad_norm": 5.378372669219971,
"learning_rate": 6.629834254143647e-07,
"loss": 0.8688,
"step": 12
},
{
"epoch": 0.007188940092165899,
"grad_norm": 5.282576560974121,
"learning_rate": 7.18232044198895e-07,
"loss": 0.8179,
"step": 13
},
{
"epoch": 0.007741935483870968,
"grad_norm": 4.443999290466309,
"learning_rate": 7.734806629834254e-07,
"loss": 0.829,
"step": 14
},
{
"epoch": 0.008294930875576038,
"grad_norm": 4.481590270996094,
"learning_rate": 8.287292817679559e-07,
"loss": 0.8441,
"step": 15
},
{
"epoch": 0.008847926267281106,
"grad_norm": 4.04002571105957,
"learning_rate": 8.839779005524863e-07,
"loss": 0.7884,
"step": 16
},
{
"epoch": 0.009400921658986176,
"grad_norm": 3.9912734031677246,
"learning_rate": 9.392265193370166e-07,
"loss": 0.7851,
"step": 17
},
{
"epoch": 0.009953917050691244,
"grad_norm": 3.7548842430114746,
"learning_rate": 9.94475138121547e-07,
"loss": 0.7836,
"step": 18
},
{
"epoch": 0.010506912442396314,
"grad_norm": 2.2732059955596924,
"learning_rate": 1.0497237569060774e-06,
"loss": 0.7572,
"step": 19
},
{
"epoch": 0.011059907834101382,
"grad_norm": 2.244140625,
"learning_rate": 1.1049723756906078e-06,
"loss": 0.7637,
"step": 20
},
{
"epoch": 0.011612903225806452,
"grad_norm": 2.1460845470428467,
"learning_rate": 1.160220994475138e-06,
"loss": 0.7849,
"step": 21
},
{
"epoch": 0.01216589861751152,
"grad_norm": 2.0333800315856934,
"learning_rate": 1.2154696132596686e-06,
"loss": 0.8008,
"step": 22
},
{
"epoch": 0.01271889400921659,
"grad_norm": 1.962172508239746,
"learning_rate": 1.270718232044199e-06,
"loss": 0.7914,
"step": 23
},
{
"epoch": 0.013271889400921659,
"grad_norm": 1.9014993906021118,
"learning_rate": 1.3259668508287293e-06,
"loss": 0.754,
"step": 24
},
{
"epoch": 0.013824884792626729,
"grad_norm": 1.5539250373840332,
"learning_rate": 1.3812154696132598e-06,
"loss": 0.7218,
"step": 25
},
{
"epoch": 0.014377880184331797,
"grad_norm": 2.1412103176116943,
"learning_rate": 1.43646408839779e-06,
"loss": 0.7083,
"step": 26
},
{
"epoch": 0.014930875576036867,
"grad_norm": 2.793086290359497,
"learning_rate": 1.4917127071823205e-06,
"loss": 0.7505,
"step": 27
},
{
"epoch": 0.015483870967741935,
"grad_norm": 2.779872179031372,
"learning_rate": 1.5469613259668508e-06,
"loss": 0.7609,
"step": 28
},
{
"epoch": 0.016036866359447004,
"grad_norm": 2.9033734798431396,
"learning_rate": 1.6022099447513815e-06,
"loss": 0.7388,
"step": 29
},
{
"epoch": 0.016589861751152075,
"grad_norm": 2.6718666553497314,
"learning_rate": 1.6574585635359118e-06,
"loss": 0.7395,
"step": 30
},
{
"epoch": 0.017142857142857144,
"grad_norm": 2.3672306537628174,
"learning_rate": 1.7127071823204422e-06,
"loss": 0.736,
"step": 31
},
{
"epoch": 0.017695852534562212,
"grad_norm": 2.0016157627105713,
"learning_rate": 1.7679558011049725e-06,
"loss": 0.7169,
"step": 32
},
{
"epoch": 0.01824884792626728,
"grad_norm": 1.755001425743103,
"learning_rate": 1.823204419889503e-06,
"loss": 0.7358,
"step": 33
},
{
"epoch": 0.018801843317972352,
"grad_norm": 1.084032416343689,
"learning_rate": 1.8784530386740332e-06,
"loss": 0.654,
"step": 34
},
{
"epoch": 0.01935483870967742,
"grad_norm": 1.0225807428359985,
"learning_rate": 1.933701657458564e-06,
"loss": 0.6793,
"step": 35
},
{
"epoch": 0.01990783410138249,
"grad_norm": 1.0972223281860352,
"learning_rate": 1.988950276243094e-06,
"loss": 0.6625,
"step": 36
},
{
"epoch": 0.020460829493087557,
"grad_norm": 1.1796579360961914,
"learning_rate": 2.0441988950276245e-06,
"loss": 0.6905,
"step": 37
},
{
"epoch": 0.02101382488479263,
"grad_norm": 1.1285828351974487,
"learning_rate": 2.0994475138121547e-06,
"loss": 0.6433,
"step": 38
},
{
"epoch": 0.021566820276497697,
"grad_norm": 1.0303235054016113,
"learning_rate": 2.1546961325966854e-06,
"loss": 0.637,
"step": 39
},
{
"epoch": 0.022119815668202765,
"grad_norm": 0.9121658205986023,
"learning_rate": 2.2099447513812157e-06,
"loss": 0.6506,
"step": 40
},
{
"epoch": 0.022672811059907833,
"grad_norm": 0.8387944102287292,
"learning_rate": 2.265193370165746e-06,
"loss": 0.6217,
"step": 41
},
{
"epoch": 0.023225806451612905,
"grad_norm": 0.762744665145874,
"learning_rate": 2.320441988950276e-06,
"loss": 0.6131,
"step": 42
},
{
"epoch": 0.023778801843317973,
"grad_norm": 0.8466107845306396,
"learning_rate": 2.375690607734807e-06,
"loss": 0.6313,
"step": 43
},
{
"epoch": 0.02433179723502304,
"grad_norm": 0.7411535382270813,
"learning_rate": 2.430939226519337e-06,
"loss": 0.6178,
"step": 44
},
{
"epoch": 0.02488479262672811,
"grad_norm": 0.7712140083312988,
"learning_rate": 2.486187845303868e-06,
"loss": 0.6218,
"step": 45
},
{
"epoch": 0.02543778801843318,
"grad_norm": 1.0630971193313599,
"learning_rate": 2.541436464088398e-06,
"loss": 0.6497,
"step": 46
},
{
"epoch": 0.02599078341013825,
"grad_norm": 0.8186689615249634,
"learning_rate": 2.5966850828729284e-06,
"loss": 0.6493,
"step": 47
},
{
"epoch": 0.026543778801843318,
"grad_norm": 0.7290191650390625,
"learning_rate": 2.6519337016574586e-06,
"loss": 0.6278,
"step": 48
},
{
"epoch": 0.027096774193548386,
"grad_norm": 0.6419755816459656,
"learning_rate": 2.707182320441989e-06,
"loss": 0.6495,
"step": 49
},
{
"epoch": 0.027649769585253458,
"grad_norm": 0.6034302711486816,
"learning_rate": 2.7624309392265196e-06,
"loss": 0.6262,
"step": 50
},
{
"epoch": 0.028202764976958526,
"grad_norm": 0.7188774943351746,
"learning_rate": 2.81767955801105e-06,
"loss": 0.6416,
"step": 51
},
{
"epoch": 0.028755760368663594,
"grad_norm": 0.7371018528938293,
"learning_rate": 2.87292817679558e-06,
"loss": 0.6181,
"step": 52
},
{
"epoch": 0.029308755760368663,
"grad_norm": 0.7526310086250305,
"learning_rate": 2.9281767955801104e-06,
"loss": 0.6644,
"step": 53
},
{
"epoch": 0.029861751152073734,
"grad_norm": 0.6002934575080872,
"learning_rate": 2.983425414364641e-06,
"loss": 0.57,
"step": 54
},
{
"epoch": 0.030414746543778803,
"grad_norm": 0.535647451877594,
"learning_rate": 3.0386740331491713e-06,
"loss": 0.6158,
"step": 55
},
{
"epoch": 0.03096774193548387,
"grad_norm": 0.5333572030067444,
"learning_rate": 3.0939226519337016e-06,
"loss": 0.6237,
"step": 56
},
{
"epoch": 0.03152073732718894,
"grad_norm": 0.6033058166503906,
"learning_rate": 3.149171270718232e-06,
"loss": 0.5905,
"step": 57
},
{
"epoch": 0.03207373271889401,
"grad_norm": 0.5824403166770935,
"learning_rate": 3.204419889502763e-06,
"loss": 0.5934,
"step": 58
},
{
"epoch": 0.03262672811059908,
"grad_norm": 0.5389731526374817,
"learning_rate": 3.2596685082872933e-06,
"loss": 0.6149,
"step": 59
},
{
"epoch": 0.03317972350230415,
"grad_norm": 0.5317662954330444,
"learning_rate": 3.3149171270718235e-06,
"loss": 0.5953,
"step": 60
},
{
"epoch": 0.033732718894009216,
"grad_norm": 0.5464670658111572,
"learning_rate": 3.370165745856354e-06,
"loss": 0.5848,
"step": 61
},
{
"epoch": 0.03428571428571429,
"grad_norm": 0.5087230205535889,
"learning_rate": 3.4254143646408845e-06,
"loss": 0.5912,
"step": 62
},
{
"epoch": 0.03483870967741935,
"grad_norm": 0.4756256937980652,
"learning_rate": 3.4806629834254147e-06,
"loss": 0.5842,
"step": 63
},
{
"epoch": 0.035391705069124424,
"grad_norm": 0.49870121479034424,
"learning_rate": 3.535911602209945e-06,
"loss": 0.6009,
"step": 64
},
{
"epoch": 0.035944700460829496,
"grad_norm": 0.44439196586608887,
"learning_rate": 3.5911602209944757e-06,
"loss": 0.5832,
"step": 65
},
{
"epoch": 0.03649769585253456,
"grad_norm": 0.4260464608669281,
"learning_rate": 3.646408839779006e-06,
"loss": 0.5896,
"step": 66
},
{
"epoch": 0.03705069124423963,
"grad_norm": 0.4569395184516907,
"learning_rate": 3.7016574585635362e-06,
"loss": 0.6184,
"step": 67
},
{
"epoch": 0.037603686635944704,
"grad_norm": 0.4692537486553192,
"learning_rate": 3.7569060773480665e-06,
"loss": 0.5623,
"step": 68
},
{
"epoch": 0.03815668202764977,
"grad_norm": 0.40080568194389343,
"learning_rate": 3.812154696132597e-06,
"loss": 0.5832,
"step": 69
},
{
"epoch": 0.03870967741935484,
"grad_norm": 0.47561773657798767,
"learning_rate": 3.867403314917128e-06,
"loss": 0.5801,
"step": 70
},
{
"epoch": 0.039262672811059905,
"grad_norm": 0.47063136100769043,
"learning_rate": 3.922651933701658e-06,
"loss": 0.5913,
"step": 71
},
{
"epoch": 0.03981566820276498,
"grad_norm": 0.435703843832016,
"learning_rate": 3.977900552486188e-06,
"loss": 0.583,
"step": 72
},
{
"epoch": 0.04036866359447005,
"grad_norm": 0.44961071014404297,
"learning_rate": 4.033149171270719e-06,
"loss": 0.6015,
"step": 73
},
{
"epoch": 0.04092165898617511,
"grad_norm": 0.43612730503082275,
"learning_rate": 4.088397790055249e-06,
"loss": 0.5916,
"step": 74
},
{
"epoch": 0.041474654377880185,
"grad_norm": 0.4045799970626831,
"learning_rate": 4.143646408839779e-06,
"loss": 0.5365,
"step": 75
},
{
"epoch": 0.04202764976958526,
"grad_norm": 0.47303467988967896,
"learning_rate": 4.1988950276243095e-06,
"loss": 0.587,
"step": 76
},
{
"epoch": 0.04258064516129032,
"grad_norm": 0.3942665159702301,
"learning_rate": 4.2541436464088406e-06,
"loss": 0.5682,
"step": 77
},
{
"epoch": 0.04313364055299539,
"grad_norm": 0.40599432587623596,
"learning_rate": 4.309392265193371e-06,
"loss": 0.5686,
"step": 78
},
{
"epoch": 0.04368663594470046,
"grad_norm": 0.41502925753593445,
"learning_rate": 4.364640883977901e-06,
"loss": 0.5644,
"step": 79
},
{
"epoch": 0.04423963133640553,
"grad_norm": 0.4402730166912079,
"learning_rate": 4.419889502762431e-06,
"loss": 0.5458,
"step": 80
},
{
"epoch": 0.0447926267281106,
"grad_norm": 0.4110409915447235,
"learning_rate": 4.475138121546962e-06,
"loss": 0.5447,
"step": 81
},
{
"epoch": 0.045345622119815666,
"grad_norm": 0.46120530366897583,
"learning_rate": 4.530386740331492e-06,
"loss": 0.584,
"step": 82
},
{
"epoch": 0.04589861751152074,
"grad_norm": 0.4089435040950775,
"learning_rate": 4.585635359116022e-06,
"loss": 0.555,
"step": 83
},
{
"epoch": 0.04645161290322581,
"grad_norm": 0.4719107449054718,
"learning_rate": 4.640883977900552e-06,
"loss": 0.5714,
"step": 84
},
{
"epoch": 0.047004608294930875,
"grad_norm": 0.45129814743995667,
"learning_rate": 4.6961325966850835e-06,
"loss": 0.5443,
"step": 85
},
{
"epoch": 0.047557603686635946,
"grad_norm": 0.4321492612361908,
"learning_rate": 4.751381215469614e-06,
"loss": 0.5624,
"step": 86
},
{
"epoch": 0.04811059907834101,
"grad_norm": 0.4300953149795532,
"learning_rate": 4.806629834254144e-06,
"loss": 0.5497,
"step": 87
},
{
"epoch": 0.04866359447004608,
"grad_norm": 0.4590497314929962,
"learning_rate": 4.861878453038674e-06,
"loss": 0.5652,
"step": 88
},
{
"epoch": 0.049216589861751155,
"grad_norm": 0.47058796882629395,
"learning_rate": 4.9171270718232054e-06,
"loss": 0.5768,
"step": 89
},
{
"epoch": 0.04976958525345622,
"grad_norm": 0.43738317489624023,
"learning_rate": 4.972375690607736e-06,
"loss": 0.58,
"step": 90
},
{
"epoch": 0.05032258064516129,
"grad_norm": 0.4055286943912506,
"learning_rate": 5.027624309392266e-06,
"loss": 0.5503,
"step": 91
},
{
"epoch": 0.05087557603686636,
"grad_norm": 0.44679558277130127,
"learning_rate": 5.082872928176796e-06,
"loss": 0.5347,
"step": 92
},
{
"epoch": 0.05142857142857143,
"grad_norm": 0.46003517508506775,
"learning_rate": 5.1381215469613265e-06,
"loss": 0.5854,
"step": 93
},
{
"epoch": 0.0519815668202765,
"grad_norm": 0.44815686345100403,
"learning_rate": 5.193370165745857e-06,
"loss": 0.5316,
"step": 94
},
{
"epoch": 0.052534562211981564,
"grad_norm": 0.4559434950351715,
"learning_rate": 5.248618784530387e-06,
"loss": 0.5724,
"step": 95
},
{
"epoch": 0.053087557603686636,
"grad_norm": 0.45906394720077515,
"learning_rate": 5.303867403314917e-06,
"loss": 0.565,
"step": 96
},
{
"epoch": 0.05364055299539171,
"grad_norm": 0.3892553448677063,
"learning_rate": 5.3591160220994476e-06,
"loss": 0.5397,
"step": 97
},
{
"epoch": 0.05419354838709677,
"grad_norm": 0.4995177090167999,
"learning_rate": 5.414364640883978e-06,
"loss": 0.5509,
"step": 98
},
{
"epoch": 0.054746543778801844,
"grad_norm": 0.4675018787384033,
"learning_rate": 5.469613259668509e-06,
"loss": 0.5432,
"step": 99
},
{
"epoch": 0.055299539170506916,
"grad_norm": 0.4121154844760895,
"learning_rate": 5.524861878453039e-06,
"loss": 0.5463,
"step": 100
},
{
"epoch": 0.05585253456221198,
"grad_norm": 0.4763623774051666,
"learning_rate": 5.5801104972375695e-06,
"loss": 0.5673,
"step": 101
},
{
"epoch": 0.05640552995391705,
"grad_norm": 0.44080275297164917,
"learning_rate": 5.6353591160221e-06,
"loss": 0.5684,
"step": 102
},
{
"epoch": 0.05695852534562212,
"grad_norm": 0.4623638391494751,
"learning_rate": 5.69060773480663e-06,
"loss": 0.5738,
"step": 103
},
{
"epoch": 0.05751152073732719,
"grad_norm": 0.379423588514328,
"learning_rate": 5.74585635359116e-06,
"loss": 0.5503,
"step": 104
},
{
"epoch": 0.05806451612903226,
"grad_norm": 0.45095521211624146,
"learning_rate": 5.8011049723756905e-06,
"loss": 0.5683,
"step": 105
},
{
"epoch": 0.058617511520737325,
"grad_norm": 0.4157678484916687,
"learning_rate": 5.856353591160221e-06,
"loss": 0.5459,
"step": 106
},
{
"epoch": 0.0591705069124424,
"grad_norm": 0.4484108090400696,
"learning_rate": 5.911602209944752e-06,
"loss": 0.5465,
"step": 107
},
{
"epoch": 0.05972350230414747,
"grad_norm": 0.41227927803993225,
"learning_rate": 5.966850828729282e-06,
"loss": 0.5737,
"step": 108
},
{
"epoch": 0.060276497695852534,
"grad_norm": 0.4771617352962494,
"learning_rate": 6.0220994475138124e-06,
"loss": 0.5295,
"step": 109
},
{
"epoch": 0.060829493087557605,
"grad_norm": 0.42870959639549255,
"learning_rate": 6.077348066298343e-06,
"loss": 0.5432,
"step": 110
},
{
"epoch": 0.06138248847926267,
"grad_norm": 0.47716712951660156,
"learning_rate": 6.132596685082873e-06,
"loss": 0.5695,
"step": 111
},
{
"epoch": 0.06193548387096774,
"grad_norm": 0.43211421370506287,
"learning_rate": 6.187845303867403e-06,
"loss": 0.5408,
"step": 112
},
{
"epoch": 0.062488479262672814,
"grad_norm": 0.49489325284957886,
"learning_rate": 6.2430939226519335e-06,
"loss": 0.5488,
"step": 113
},
{
"epoch": 0.06304147465437789,
"grad_norm": 0.4600902199745178,
"learning_rate": 6.298342541436464e-06,
"loss": 0.5445,
"step": 114
},
{
"epoch": 0.06359447004608294,
"grad_norm": 0.41128382086753845,
"learning_rate": 6.353591160220996e-06,
"loss": 0.5507,
"step": 115
},
{
"epoch": 0.06414746543778801,
"grad_norm": 0.38756921887397766,
"learning_rate": 6.408839779005526e-06,
"loss": 0.5072,
"step": 116
},
{
"epoch": 0.06470046082949309,
"grad_norm": 0.4521966278553009,
"learning_rate": 6.464088397790056e-06,
"loss": 0.5597,
"step": 117
},
{
"epoch": 0.06525345622119816,
"grad_norm": 0.39012429118156433,
"learning_rate": 6.5193370165745865e-06,
"loss": 0.5017,
"step": 118
},
{
"epoch": 0.06580645161290323,
"grad_norm": 0.4543527364730835,
"learning_rate": 6.574585635359117e-06,
"loss": 0.5342,
"step": 119
},
{
"epoch": 0.0663594470046083,
"grad_norm": 0.4338400065898895,
"learning_rate": 6.629834254143647e-06,
"loss": 0.5489,
"step": 120
},
{
"epoch": 0.06691244239631336,
"grad_norm": 0.461480975151062,
"learning_rate": 6.685082872928177e-06,
"loss": 0.5327,
"step": 121
},
{
"epoch": 0.06746543778801843,
"grad_norm": 0.4241288900375366,
"learning_rate": 6.740331491712708e-06,
"loss": 0.5297,
"step": 122
},
{
"epoch": 0.0680184331797235,
"grad_norm": 0.5540463924407959,
"learning_rate": 6.795580110497239e-06,
"loss": 0.5262,
"step": 123
},
{
"epoch": 0.06857142857142857,
"grad_norm": 0.4794096350669861,
"learning_rate": 6.850828729281769e-06,
"loss": 0.5638,
"step": 124
},
{
"epoch": 0.06912442396313365,
"grad_norm": 0.43071404099464417,
"learning_rate": 6.906077348066299e-06,
"loss": 0.5254,
"step": 125
},
{
"epoch": 0.0696774193548387,
"grad_norm": 0.4853145480155945,
"learning_rate": 6.9613259668508295e-06,
"loss": 0.5524,
"step": 126
},
{
"epoch": 0.07023041474654378,
"grad_norm": 0.47362276911735535,
"learning_rate": 7.01657458563536e-06,
"loss": 0.5469,
"step": 127
},
{
"epoch": 0.07078341013824885,
"grad_norm": 0.45325443148612976,
"learning_rate": 7.07182320441989e-06,
"loss": 0.5476,
"step": 128
},
{
"epoch": 0.07133640552995392,
"grad_norm": 0.43244412541389465,
"learning_rate": 7.12707182320442e-06,
"loss": 0.5367,
"step": 129
},
{
"epoch": 0.07188940092165899,
"grad_norm": 0.5289068222045898,
"learning_rate": 7.182320441988951e-06,
"loss": 0.5796,
"step": 130
},
{
"epoch": 0.07244239631336405,
"grad_norm": 0.4485274851322174,
"learning_rate": 7.237569060773482e-06,
"loss": 0.5396,
"step": 131
},
{
"epoch": 0.07299539170506912,
"grad_norm": 0.42793214321136475,
"learning_rate": 7.292817679558012e-06,
"loss": 0.5532,
"step": 132
},
{
"epoch": 0.07354838709677419,
"grad_norm": 0.4250226318836212,
"learning_rate": 7.348066298342542e-06,
"loss": 0.5847,
"step": 133
},
{
"epoch": 0.07410138248847926,
"grad_norm": 0.44791412353515625,
"learning_rate": 7.4033149171270724e-06,
"loss": 0.5216,
"step": 134
},
{
"epoch": 0.07465437788018434,
"grad_norm": 0.48473551869392395,
"learning_rate": 7.458563535911603e-06,
"loss": 0.5434,
"step": 135
},
{
"epoch": 0.07520737327188941,
"grad_norm": 0.5042223334312439,
"learning_rate": 7.513812154696133e-06,
"loss": 0.5292,
"step": 136
},
{
"epoch": 0.07576036866359447,
"grad_norm": 0.5112206935882568,
"learning_rate": 7.569060773480663e-06,
"loss": 0.573,
"step": 137
},
{
"epoch": 0.07631336405529954,
"grad_norm": 0.4625641107559204,
"learning_rate": 7.624309392265194e-06,
"loss": 0.5656,
"step": 138
},
{
"epoch": 0.07686635944700461,
"grad_norm": 0.507746160030365,
"learning_rate": 7.679558011049725e-06,
"loss": 0.5352,
"step": 139
},
{
"epoch": 0.07741935483870968,
"grad_norm": 0.4666261672973633,
"learning_rate": 7.734806629834256e-06,
"loss": 0.5465,
"step": 140
},
{
"epoch": 0.07797235023041475,
"grad_norm": 0.44455286860466003,
"learning_rate": 7.790055248618785e-06,
"loss": 0.5395,
"step": 141
},
{
"epoch": 0.07852534562211981,
"grad_norm": 0.4286342263221741,
"learning_rate": 7.845303867403316e-06,
"loss": 0.5045,
"step": 142
},
{
"epoch": 0.07907834101382488,
"grad_norm": 0.4748035669326782,
"learning_rate": 7.900552486187846e-06,
"loss": 0.552,
"step": 143
},
{
"epoch": 0.07963133640552995,
"grad_norm": 0.44316938519477844,
"learning_rate": 7.955801104972377e-06,
"loss": 0.5394,
"step": 144
},
{
"epoch": 0.08018433179723503,
"grad_norm": 0.4846290647983551,
"learning_rate": 8.011049723756906e-06,
"loss": 0.5267,
"step": 145
},
{
"epoch": 0.0807373271889401,
"grad_norm": 0.4428083300590515,
"learning_rate": 8.066298342541437e-06,
"loss": 0.4941,
"step": 146
},
{
"epoch": 0.08129032258064516,
"grad_norm": 0.44677111506462097,
"learning_rate": 8.121546961325968e-06,
"loss": 0.4885,
"step": 147
},
{
"epoch": 0.08184331797235023,
"grad_norm": 0.4421268105506897,
"learning_rate": 8.176795580110498e-06,
"loss": 0.5765,
"step": 148
},
{
"epoch": 0.0823963133640553,
"grad_norm": 0.4853808581829071,
"learning_rate": 8.232044198895029e-06,
"loss": 0.5297,
"step": 149
},
{
"epoch": 0.08294930875576037,
"grad_norm": 0.47924932837486267,
"learning_rate": 8.287292817679558e-06,
"loss": 0.5192,
"step": 150
},
{
"epoch": 0.08350230414746544,
"grad_norm": 0.5630611181259155,
"learning_rate": 8.34254143646409e-06,
"loss": 0.5439,
"step": 151
},
{
"epoch": 0.08405529953917051,
"grad_norm": 0.6081127524375916,
"learning_rate": 8.397790055248619e-06,
"loss": 0.5012,
"step": 152
},
{
"epoch": 0.08460829493087557,
"grad_norm": 0.44527262449264526,
"learning_rate": 8.45303867403315e-06,
"loss": 0.5305,
"step": 153
},
{
"epoch": 0.08516129032258064,
"grad_norm": 0.4948616623878479,
"learning_rate": 8.508287292817681e-06,
"loss": 0.5109,
"step": 154
},
{
"epoch": 0.08571428571428572,
"grad_norm": 0.5024743676185608,
"learning_rate": 8.56353591160221e-06,
"loss": 0.5632,
"step": 155
},
{
"epoch": 0.08626728110599079,
"grad_norm": 0.45424172282218933,
"learning_rate": 8.618784530386742e-06,
"loss": 0.5029,
"step": 156
},
{
"epoch": 0.08682027649769586,
"grad_norm": 0.5445595979690552,
"learning_rate": 8.674033149171271e-06,
"loss": 0.5451,
"step": 157
},
{
"epoch": 0.08737327188940092,
"grad_norm": 0.495613157749176,
"learning_rate": 8.729281767955802e-06,
"loss": 0.527,
"step": 158
},
{
"epoch": 0.08792626728110599,
"grad_norm": 0.6915062665939331,
"learning_rate": 8.784530386740332e-06,
"loss": 0.5378,
"step": 159
},
{
"epoch": 0.08847926267281106,
"grad_norm": 0.4711109697818756,
"learning_rate": 8.839779005524863e-06,
"loss": 0.5105,
"step": 160
},
{
"epoch": 0.08903225806451613,
"grad_norm": 0.5774613618850708,
"learning_rate": 8.895027624309392e-06,
"loss": 0.5268,
"step": 161
},
{
"epoch": 0.0895852534562212,
"grad_norm": 0.534370481967926,
"learning_rate": 8.950276243093923e-06,
"loss": 0.5275,
"step": 162
},
{
"epoch": 0.09013824884792626,
"grad_norm": 0.5167669057846069,
"learning_rate": 9.005524861878454e-06,
"loss": 0.5173,
"step": 163
},
{
"epoch": 0.09069124423963133,
"grad_norm": 0.5220374464988708,
"learning_rate": 9.060773480662984e-06,
"loss": 0.5117,
"step": 164
},
{
"epoch": 0.0912442396313364,
"grad_norm": 0.6151928305625916,
"learning_rate": 9.116022099447515e-06,
"loss": 0.5236,
"step": 165
},
{
"epoch": 0.09179723502304148,
"grad_norm": 0.54198157787323,
"learning_rate": 9.171270718232044e-06,
"loss": 0.5113,
"step": 166
},
{
"epoch": 0.09235023041474655,
"grad_norm": 0.540729820728302,
"learning_rate": 9.226519337016575e-06,
"loss": 0.5293,
"step": 167
},
{
"epoch": 0.09290322580645162,
"grad_norm": 0.4339202642440796,
"learning_rate": 9.281767955801105e-06,
"loss": 0.5105,
"step": 168
},
{
"epoch": 0.09345622119815668,
"grad_norm": 0.5427307486534119,
"learning_rate": 9.337016574585636e-06,
"loss": 0.5442,
"step": 169
},
{
"epoch": 0.09400921658986175,
"grad_norm": 0.5696178674697876,
"learning_rate": 9.392265193370167e-06,
"loss": 0.5198,
"step": 170
},
{
"epoch": 0.09456221198156682,
"grad_norm": 0.475676029920578,
"learning_rate": 9.447513812154696e-06,
"loss": 0.5275,
"step": 171
},
{
"epoch": 0.09511520737327189,
"grad_norm": 0.6084780693054199,
"learning_rate": 9.502762430939228e-06,
"loss": 0.5344,
"step": 172
},
{
"epoch": 0.09566820276497696,
"grad_norm": 0.45912471413612366,
"learning_rate": 9.558011049723757e-06,
"loss": 0.5128,
"step": 173
},
{
"epoch": 0.09622119815668202,
"grad_norm": 0.5522315502166748,
"learning_rate": 9.613259668508288e-06,
"loss": 0.4985,
"step": 174
},
{
"epoch": 0.0967741935483871,
"grad_norm": 0.5598754286766052,
"learning_rate": 9.668508287292818e-06,
"loss": 0.5438,
"step": 175
},
{
"epoch": 0.09732718894009217,
"grad_norm": 0.4842788577079773,
"learning_rate": 9.723756906077349e-06,
"loss": 0.5098,
"step": 176
},
{
"epoch": 0.09788018433179724,
"grad_norm": 0.5201513171195984,
"learning_rate": 9.779005524861878e-06,
"loss": 0.5155,
"step": 177
},
{
"epoch": 0.09843317972350231,
"grad_norm": 0.4839669466018677,
"learning_rate": 9.834254143646411e-06,
"loss": 0.5201,
"step": 178
},
{
"epoch": 0.09898617511520737,
"grad_norm": 0.5859594941139221,
"learning_rate": 9.88950276243094e-06,
"loss": 0.5603,
"step": 179
},
{
"epoch": 0.09953917050691244,
"grad_norm": 0.48163536190986633,
"learning_rate": 9.944751381215471e-06,
"loss": 0.5279,
"step": 180
},
{
"epoch": 0.10009216589861751,
"grad_norm": 0.508065938949585,
"learning_rate": 1e-05,
"loss": 0.5186,
"step": 181
},
{
"epoch": 0.10064516129032258,
"grad_norm": 0.47829869389533997,
"learning_rate": 9.999990678956964e-06,
"loss": 0.5224,
"step": 182
},
{
"epoch": 0.10119815668202765,
"grad_norm": 0.5300391912460327,
"learning_rate": 9.999962715862601e-06,
"loss": 0.5128,
"step": 183
},
{
"epoch": 0.10175115207373273,
"grad_norm": 0.47900837659835815,
"learning_rate": 9.999916110821174e-06,
"loss": 0.5232,
"step": 184
},
{
"epoch": 0.10230414746543778,
"grad_norm": 0.4351784288883209,
"learning_rate": 9.999850864006444e-06,
"loss": 0.5274,
"step": 185
},
{
"epoch": 0.10285714285714286,
"grad_norm": 0.48379045724868774,
"learning_rate": 9.99976697566168e-06,
"loss": 0.4964,
"step": 186
},
{
"epoch": 0.10341013824884793,
"grad_norm": 0.4367005228996277,
"learning_rate": 9.999664446099651e-06,
"loss": 0.5311,
"step": 187
},
{
"epoch": 0.103963133640553,
"grad_norm": 0.4944308400154114,
"learning_rate": 9.999543275702632e-06,
"loss": 0.5093,
"step": 188
},
{
"epoch": 0.10451612903225807,
"grad_norm": 0.44800275564193726,
"learning_rate": 9.999403464922393e-06,
"loss": 0.5124,
"step": 189
},
{
"epoch": 0.10506912442396313,
"grad_norm": 0.5029193758964539,
"learning_rate": 9.99924501428021e-06,
"loss": 0.5538,
"step": 190
},
{
"epoch": 0.1056221198156682,
"grad_norm": 0.4435194730758667,
"learning_rate": 9.999067924366854e-06,
"loss": 0.5291,
"step": 191
},
{
"epoch": 0.10617511520737327,
"grad_norm": 0.45577532052993774,
"learning_rate": 9.998872195842588e-06,
"loss": 0.5107,
"step": 192
},
{
"epoch": 0.10672811059907834,
"grad_norm": 0.4373863637447357,
"learning_rate": 9.998657829437171e-06,
"loss": 0.5137,
"step": 193
},
{
"epoch": 0.10728110599078342,
"grad_norm": 0.4408453404903412,
"learning_rate": 9.998424825949848e-06,
"loss": 0.5061,
"step": 194
},
{
"epoch": 0.10783410138248847,
"grad_norm": 0.5477350950241089,
"learning_rate": 9.998173186249357e-06,
"loss": 0.5363,
"step": 195
},
{
"epoch": 0.10838709677419354,
"grad_norm": 0.43939393758773804,
"learning_rate": 9.99790291127391e-06,
"loss": 0.5095,
"step": 196
},
{
"epoch": 0.10894009216589862,
"grad_norm": 0.47420209646224976,
"learning_rate": 9.997614002031211e-06,
"loss": 0.5297,
"step": 197
},
{
"epoch": 0.10949308755760369,
"grad_norm": 0.45546746253967285,
"learning_rate": 9.99730645959843e-06,
"loss": 0.5139,
"step": 198
},
{
"epoch": 0.11004608294930876,
"grad_norm": 0.4920537769794464,
"learning_rate": 9.996980285122218e-06,
"loss": 0.5344,
"step": 199
},
{
"epoch": 0.11059907834101383,
"grad_norm": 0.562940776348114,
"learning_rate": 9.996635479818683e-06,
"loss": 0.535,
"step": 200
},
{
"epoch": 0.11115207373271889,
"grad_norm": 0.5914273262023926,
"learning_rate": 9.99627204497341e-06,
"loss": 0.5134,
"step": 201
},
{
"epoch": 0.11170506912442396,
"grad_norm": 0.4358191192150116,
"learning_rate": 9.995889981941432e-06,
"loss": 0.4886,
"step": 202
},
{
"epoch": 0.11225806451612903,
"grad_norm": 0.569436252117157,
"learning_rate": 9.995489292147238e-06,
"loss": 0.5302,
"step": 203
},
{
"epoch": 0.1128110599078341,
"grad_norm": 0.5561317801475525,
"learning_rate": 9.995069977084769e-06,
"loss": 0.5214,
"step": 204
},
{
"epoch": 0.11336405529953918,
"grad_norm": 0.4526347219944,
"learning_rate": 9.994632038317407e-06,
"loss": 0.5106,
"step": 205
},
{
"epoch": 0.11391705069124423,
"grad_norm": 0.6103693246841431,
"learning_rate": 9.99417547747797e-06,
"loss": 0.5187,
"step": 206
},
{
"epoch": 0.1144700460829493,
"grad_norm": 0.5255654454231262,
"learning_rate": 9.993700296268705e-06,
"loss": 0.5253,
"step": 207
},
{
"epoch": 0.11502304147465438,
"grad_norm": 0.48559436202049255,
"learning_rate": 9.993206496461287e-06,
"loss": 0.5127,
"step": 208
},
{
"epoch": 0.11557603686635945,
"grad_norm": 0.5230923891067505,
"learning_rate": 9.992694079896812e-06,
"loss": 0.5078,
"step": 209
},
{
"epoch": 0.11612903225806452,
"grad_norm": 0.45681560039520264,
"learning_rate": 9.992163048485776e-06,
"loss": 0.522,
"step": 210
},
{
"epoch": 0.11668202764976958,
"grad_norm": 0.510287344455719,
"learning_rate": 9.99161340420809e-06,
"loss": 0.509,
"step": 211
},
{
"epoch": 0.11723502304147465,
"grad_norm": 0.4897903501987457,
"learning_rate": 9.991045149113055e-06,
"loss": 0.5118,
"step": 212
},
{
"epoch": 0.11778801843317972,
"grad_norm": 0.516163170337677,
"learning_rate": 9.990458285319362e-06,
"loss": 0.4915,
"step": 213
},
{
"epoch": 0.1183410138248848,
"grad_norm": 0.5389583110809326,
"learning_rate": 9.98985281501509e-06,
"loss": 0.5192,
"step": 214
},
{
"epoch": 0.11889400921658987,
"grad_norm": 0.6268092393875122,
"learning_rate": 9.989228740457679e-06,
"loss": 0.524,
"step": 215
},
{
"epoch": 0.11944700460829494,
"grad_norm": 0.447966605424881,
"learning_rate": 9.988586063973942e-06,
"loss": 0.4864,
"step": 216
},
{
"epoch": 0.12,
"grad_norm": 0.6052084565162659,
"learning_rate": 9.987924787960043e-06,
"loss": 0.5005,
"step": 217
},
{
"epoch": 0.12055299539170507,
"grad_norm": 0.5076256990432739,
"learning_rate": 9.987244914881498e-06,
"loss": 0.5219,
"step": 218
},
{
"epoch": 0.12110599078341014,
"grad_norm": 0.5539697408676147,
"learning_rate": 9.986546447273153e-06,
"loss": 0.5024,
"step": 219
},
{
"epoch": 0.12165898617511521,
"grad_norm": 0.4377414584159851,
"learning_rate": 9.985829387739192e-06,
"loss": 0.5291,
"step": 220
},
{
"epoch": 0.12221198156682028,
"grad_norm": 0.5455598831176758,
"learning_rate": 9.985093738953108e-06,
"loss": 0.5126,
"step": 221
},
{
"epoch": 0.12276497695852534,
"grad_norm": 0.462424099445343,
"learning_rate": 9.98433950365771e-06,
"loss": 0.5225,
"step": 222
},
{
"epoch": 0.12331797235023041,
"grad_norm": 0.49225398898124695,
"learning_rate": 9.983566684665097e-06,
"loss": 0.5328,
"step": 223
},
{
"epoch": 0.12387096774193548,
"grad_norm": 0.5304319262504578,
"learning_rate": 9.982775284856665e-06,
"loss": 0.5212,
"step": 224
},
{
"epoch": 0.12442396313364056,
"grad_norm": 0.4655149579048157,
"learning_rate": 9.981965307183081e-06,
"loss": 0.5161,
"step": 225
},
{
"epoch": 0.12497695852534563,
"grad_norm": 0.6474348902702332,
"learning_rate": 9.98113675466428e-06,
"loss": 0.4983,
"step": 226
},
{
"epoch": 0.12552995391705069,
"grad_norm": 0.43538525700569153,
"learning_rate": 9.980289630389453e-06,
"loss": 0.5286,
"step": 227
},
{
"epoch": 0.12608294930875577,
"grad_norm": 0.5731073617935181,
"learning_rate": 9.97942393751703e-06,
"loss": 0.5414,
"step": 228
},
{
"epoch": 0.12663594470046083,
"grad_norm": 0.4664866626262665,
"learning_rate": 9.978539679274675e-06,
"loss": 0.5052,
"step": 229
},
{
"epoch": 0.1271889400921659,
"grad_norm": 0.5314398407936096,
"learning_rate": 9.977636858959274e-06,
"loss": 0.527,
"step": 230
},
{
"epoch": 0.12774193548387097,
"grad_norm": 0.538694441318512,
"learning_rate": 9.976715479936916e-06,
"loss": 0.5094,
"step": 231
},
{
"epoch": 0.12829493087557603,
"grad_norm": 0.4854016900062561,
"learning_rate": 9.975775545642889e-06,
"loss": 0.5078,
"step": 232
},
{
"epoch": 0.12884792626728112,
"grad_norm": 0.5078600645065308,
"learning_rate": 9.974817059581656e-06,
"loss": 0.5236,
"step": 233
},
{
"epoch": 0.12940092165898617,
"grad_norm": 0.455322802066803,
"learning_rate": 9.973840025326858e-06,
"loss": 0.5137,
"step": 234
},
{
"epoch": 0.12995391705069123,
"grad_norm": 0.5216564536094666,
"learning_rate": 9.972844446521281e-06,
"loss": 0.5289,
"step": 235
},
{
"epoch": 0.13050691244239632,
"grad_norm": 0.4569193422794342,
"learning_rate": 9.971830326876864e-06,
"loss": 0.5137,
"step": 236
},
{
"epoch": 0.13105990783410137,
"grad_norm": 0.5937883853912354,
"learning_rate": 9.970797670174663e-06,
"loss": 0.5142,
"step": 237
},
{
"epoch": 0.13161290322580646,
"grad_norm": 0.48625534772872925,
"learning_rate": 9.969746480264855e-06,
"loss": 0.5279,
"step": 238
},
{
"epoch": 0.13216589861751152,
"grad_norm": 0.47141098976135254,
"learning_rate": 9.968676761066714e-06,
"loss": 0.5042,
"step": 239
},
{
"epoch": 0.1327188940092166,
"grad_norm": 0.4889548122882843,
"learning_rate": 9.967588516568601e-06,
"loss": 0.5154,
"step": 240
},
{
"epoch": 0.13327188940092166,
"grad_norm": 0.4120868444442749,
"learning_rate": 9.966481750827943e-06,
"loss": 0.504,
"step": 241
},
{
"epoch": 0.13382488479262672,
"grad_norm": 0.5358414053916931,
"learning_rate": 9.965356467971228e-06,
"loss": 0.5264,
"step": 242
},
{
"epoch": 0.1343778801843318,
"grad_norm": 0.472572922706604,
"learning_rate": 9.964212672193978e-06,
"loss": 0.5176,
"step": 243
},
{
"epoch": 0.13493087557603686,
"grad_norm": 0.4870736598968506,
"learning_rate": 9.96305036776074e-06,
"loss": 0.5071,
"step": 244
},
{
"epoch": 0.13548387096774195,
"grad_norm": 0.5704755783081055,
"learning_rate": 9.96186955900507e-06,
"loss": 0.5198,
"step": 245
},
{
"epoch": 0.136036866359447,
"grad_norm": 0.5287049412727356,
"learning_rate": 9.960670250329517e-06,
"loss": 0.5216,
"step": 246
},
{
"epoch": 0.13658986175115206,
"grad_norm": 0.5611394643783569,
"learning_rate": 9.959452446205603e-06,
"loss": 0.5188,
"step": 247
},
{
"epoch": 0.13714285714285715,
"grad_norm": 0.4512503445148468,
"learning_rate": 9.958216151173812e-06,
"loss": 0.4979,
"step": 248
},
{
"epoch": 0.1376958525345622,
"grad_norm": 0.47202053666114807,
"learning_rate": 9.956961369843565e-06,
"loss": 0.4834,
"step": 249
},
{
"epoch": 0.1382488479262673,
"grad_norm": 0.5006586909294128,
"learning_rate": 9.95568810689321e-06,
"loss": 0.5285,
"step": 250
},
{
"epoch": 0.13880184331797235,
"grad_norm": 0.4660753011703491,
"learning_rate": 9.954396367070006e-06,
"loss": 0.5192,
"step": 251
},
{
"epoch": 0.1393548387096774,
"grad_norm": 0.4885999262332916,
"learning_rate": 9.953086155190095e-06,
"loss": 0.5028,
"step": 252
},
{
"epoch": 0.1399078341013825,
"grad_norm": 0.49867549538612366,
"learning_rate": 9.951757476138495e-06,
"loss": 0.5039,
"step": 253
},
{
"epoch": 0.14046082949308755,
"grad_norm": 0.5823776125907898,
"learning_rate": 9.950410334869075e-06,
"loss": 0.535,
"step": 254
},
{
"epoch": 0.14101382488479264,
"grad_norm": 0.47015827894210815,
"learning_rate": 9.949044736404538e-06,
"loss": 0.528,
"step": 255
},
{
"epoch": 0.1415668202764977,
"grad_norm": 0.5392245054244995,
"learning_rate": 9.94766068583641e-06,
"loss": 0.4855,
"step": 256
},
{
"epoch": 0.14211981566820275,
"grad_norm": 0.5323054194450378,
"learning_rate": 9.946258188325003e-06,
"loss": 0.5269,
"step": 257
},
{
"epoch": 0.14267281105990784,
"grad_norm": 0.5459584593772888,
"learning_rate": 9.944837249099418e-06,
"loss": 0.4874,
"step": 258
},
{
"epoch": 0.1432258064516129,
"grad_norm": 0.4532824158668518,
"learning_rate": 9.943397873457503e-06,
"loss": 0.5111,
"step": 259
},
{
"epoch": 0.14377880184331798,
"grad_norm": 0.46974876523017883,
"learning_rate": 9.94194006676586e-06,
"loss": 0.5261,
"step": 260
},
{
"epoch": 0.14433179723502304,
"grad_norm": 0.5061238408088684,
"learning_rate": 9.94046383445979e-06,
"loss": 0.5217,
"step": 261
},
{
"epoch": 0.1448847926267281,
"grad_norm": 0.4556806981563568,
"learning_rate": 9.938969182043312e-06,
"loss": 0.4969,
"step": 262
},
{
"epoch": 0.14543778801843318,
"grad_norm": 0.48571473360061646,
"learning_rate": 9.937456115089108e-06,
"loss": 0.5057,
"step": 263
},
{
"epoch": 0.14599078341013824,
"grad_norm": 0.497122198343277,
"learning_rate": 9.935924639238526e-06,
"loss": 0.5205,
"step": 264
},
{
"epoch": 0.14654377880184333,
"grad_norm": 0.5786699652671814,
"learning_rate": 9.934374760201546e-06,
"loss": 0.5122,
"step": 265
},
{
"epoch": 0.14709677419354839,
"grad_norm": 0.5201807022094727,
"learning_rate": 9.932806483756763e-06,
"loss": 0.5328,
"step": 266
},
{
"epoch": 0.14764976958525344,
"grad_norm": 0.5433037877082825,
"learning_rate": 9.931219815751368e-06,
"loss": 0.5153,
"step": 267
},
{
"epoch": 0.14820276497695853,
"grad_norm": 0.47332048416137695,
"learning_rate": 9.929614762101117e-06,
"loss": 0.5004,
"step": 268
},
{
"epoch": 0.1487557603686636,
"grad_norm": 0.5838091373443604,
"learning_rate": 9.927991328790324e-06,
"loss": 0.4889,
"step": 269
},
{
"epoch": 0.14930875576036867,
"grad_norm": 0.4618660807609558,
"learning_rate": 9.926349521871824e-06,
"loss": 0.5265,
"step": 270
},
{
"epoch": 0.14986175115207373,
"grad_norm": 0.5086191892623901,
"learning_rate": 9.924689347466959e-06,
"loss": 0.5106,
"step": 271
},
{
"epoch": 0.15041474654377882,
"grad_norm": 0.457042932510376,
"learning_rate": 9.92301081176555e-06,
"loss": 0.518,
"step": 272
},
{
"epoch": 0.15096774193548387,
"grad_norm": 0.5256728529930115,
"learning_rate": 9.92131392102588e-06,
"loss": 0.531,
"step": 273
},
{
"epoch": 0.15152073732718893,
"grad_norm": 0.4683758318424225,
"learning_rate": 9.919598681574665e-06,
"loss": 0.4896,
"step": 274
},
{
"epoch": 0.15207373271889402,
"grad_norm": 0.4681558609008789,
"learning_rate": 9.917865099807034e-06,
"loss": 0.5089,
"step": 275
},
{
"epoch": 0.15262672811059907,
"grad_norm": 0.4770297706127167,
"learning_rate": 9.916113182186503e-06,
"loss": 0.4973,
"step": 276
},
{
"epoch": 0.15317972350230416,
"grad_norm": 0.4939088225364685,
"learning_rate": 9.91434293524495e-06,
"loss": 0.5388,
"step": 277
},
{
"epoch": 0.15373271889400922,
"grad_norm": 0.5165495276451111,
"learning_rate": 9.912554365582596e-06,
"loss": 0.5046,
"step": 278
},
{
"epoch": 0.15428571428571428,
"grad_norm": 0.4685964584350586,
"learning_rate": 9.910747479867975e-06,
"loss": 0.5373,
"step": 279
},
{
"epoch": 0.15483870967741936,
"grad_norm": 0.5078836679458618,
"learning_rate": 9.908922284837911e-06,
"loss": 0.5105,
"step": 280
},
{
"epoch": 0.15539170506912442,
"grad_norm": 0.4230501651763916,
"learning_rate": 9.90707878729749e-06,
"loss": 0.4749,
"step": 281
},
{
"epoch": 0.1559447004608295,
"grad_norm": 0.483981728553772,
"learning_rate": 9.905216994120044e-06,
"loss": 0.4954,
"step": 282
},
{
"epoch": 0.15649769585253456,
"grad_norm": 0.4270246922969818,
"learning_rate": 9.90333691224711e-06,
"loss": 0.5007,
"step": 283
},
{
"epoch": 0.15705069124423962,
"grad_norm": 0.4967236816883087,
"learning_rate": 9.901438548688423e-06,
"loss": 0.5308,
"step": 284
},
{
"epoch": 0.1576036866359447,
"grad_norm": 0.4965408444404602,
"learning_rate": 9.89952191052187e-06,
"loss": 0.5093,
"step": 285
},
{
"epoch": 0.15815668202764976,
"grad_norm": 0.45975926518440247,
"learning_rate": 9.89758700489348e-06,
"loss": 0.5122,
"step": 286
},
{
"epoch": 0.15870967741935485,
"grad_norm": 0.5180716514587402,
"learning_rate": 9.895633839017387e-06,
"loss": 0.5515,
"step": 287
},
{
"epoch": 0.1592626728110599,
"grad_norm": 0.43234091997146606,
"learning_rate": 9.893662420175809e-06,
"loss": 0.4901,
"step": 288
},
{
"epoch": 0.15981566820276497,
"grad_norm": 0.4894983768463135,
"learning_rate": 9.89167275571902e-06,
"loss": 0.5052,
"step": 289
},
{
"epoch": 0.16036866359447005,
"grad_norm": 0.4710214138031006,
"learning_rate": 9.889664853065315e-06,
"loss": 0.5,
"step": 290
},
{
"epoch": 0.1609216589861751,
"grad_norm": 0.5677446126937866,
"learning_rate": 9.887638719700996e-06,
"loss": 0.5169,
"step": 291
},
{
"epoch": 0.1614746543778802,
"grad_norm": 0.45073893666267395,
"learning_rate": 9.88559436318033e-06,
"loss": 0.5258,
"step": 292
},
{
"epoch": 0.16202764976958525,
"grad_norm": 0.5394282341003418,
"learning_rate": 9.883531791125538e-06,
"loss": 0.4799,
"step": 293
},
{
"epoch": 0.1625806451612903,
"grad_norm": 0.4784766435623169,
"learning_rate": 9.881451011226742e-06,
"loss": 0.4693,
"step": 294
},
{
"epoch": 0.1631336405529954,
"grad_norm": 0.435619980096817,
"learning_rate": 9.87935203124196e-06,
"loss": 0.5196,
"step": 295
},
{
"epoch": 0.16368663594470045,
"grad_norm": 0.5438254475593567,
"learning_rate": 9.877234858997066e-06,
"loss": 0.5078,
"step": 296
},
{
"epoch": 0.16423963133640554,
"grad_norm": 0.4646199643611908,
"learning_rate": 9.875099502385761e-06,
"loss": 0.5239,
"step": 297
},
{
"epoch": 0.1647926267281106,
"grad_norm": 0.549284040927887,
"learning_rate": 9.872945969369546e-06,
"loss": 0.5029,
"step": 298
},
{
"epoch": 0.16534562211981566,
"grad_norm": 0.5235300064086914,
"learning_rate": 9.87077426797769e-06,
"loss": 0.5225,
"step": 299
},
{
"epoch": 0.16589861751152074,
"grad_norm": 0.525598406791687,
"learning_rate": 9.8685844063072e-06,
"loss": 0.5366,
"step": 300
},
{
"epoch": 0.1664516129032258,
"grad_norm": 0.5001644492149353,
"learning_rate": 9.866376392522798e-06,
"loss": 0.5025,
"step": 301
},
{
"epoch": 0.16700460829493088,
"grad_norm": 0.4980228841304779,
"learning_rate": 9.864150234856876e-06,
"loss": 0.5332,
"step": 302
},
{
"epoch": 0.16755760368663594,
"grad_norm": 0.44216716289520264,
"learning_rate": 9.861905941609482e-06,
"loss": 0.4831,
"step": 303
},
{
"epoch": 0.16811059907834103,
"grad_norm": 0.45178863406181335,
"learning_rate": 9.859643521148275e-06,
"loss": 0.5164,
"step": 304
},
{
"epoch": 0.16866359447004609,
"grad_norm": 0.4833186864852905,
"learning_rate": 9.857362981908505e-06,
"loss": 0.5128,
"step": 305
},
{
"epoch": 0.16921658986175114,
"grad_norm": 0.5167268514633179,
"learning_rate": 9.855064332392972e-06,
"loss": 0.5149,
"step": 306
},
{
"epoch": 0.16976958525345623,
"grad_norm": 0.4465697109699249,
"learning_rate": 9.852747581172002e-06,
"loss": 0.5016,
"step": 307
},
{
"epoch": 0.1703225806451613,
"grad_norm": 0.5105251669883728,
"learning_rate": 9.850412736883408e-06,
"loss": 0.5021,
"step": 308
},
{
"epoch": 0.17087557603686637,
"grad_norm": 0.5230706334114075,
"learning_rate": 9.848059808232464e-06,
"loss": 0.4857,
"step": 309
},
{
"epoch": 0.17142857142857143,
"grad_norm": 0.5510496497154236,
"learning_rate": 9.845688803991873e-06,
"loss": 0.5305,
"step": 310
},
{
"epoch": 0.1719815668202765,
"grad_norm": 0.4802294671535492,
"learning_rate": 9.843299733001723e-06,
"loss": 0.5084,
"step": 311
},
{
"epoch": 0.17253456221198157,
"grad_norm": 0.5729020833969116,
"learning_rate": 9.840892604169473e-06,
"loss": 0.4797,
"step": 312
},
{
"epoch": 0.17308755760368663,
"grad_norm": 0.6035499572753906,
"learning_rate": 9.838467426469897e-06,
"loss": 0.5069,
"step": 313
},
{
"epoch": 0.17364055299539172,
"grad_norm": 0.4519449770450592,
"learning_rate": 9.836024208945074e-06,
"loss": 0.4959,
"step": 314
},
{
"epoch": 0.17419354838709677,
"grad_norm": 0.5990521907806396,
"learning_rate": 9.833562960704336e-06,
"loss": 0.522,
"step": 315
},
{
"epoch": 0.17474654377880183,
"grad_norm": 0.4257313013076782,
"learning_rate": 9.831083690924246e-06,
"loss": 0.4769,
"step": 316
},
{
"epoch": 0.17529953917050692,
"grad_norm": 0.5456443428993225,
"learning_rate": 9.828586408848553e-06,
"loss": 0.5129,
"step": 317
},
{
"epoch": 0.17585253456221198,
"grad_norm": 0.4802759885787964,
"learning_rate": 9.82607112378817e-06,
"loss": 0.5043,
"step": 318
},
{
"epoch": 0.17640552995391706,
"grad_norm": 0.6043391823768616,
"learning_rate": 9.823537845121126e-06,
"loss": 0.5063,
"step": 319
},
{
"epoch": 0.17695852534562212,
"grad_norm": 0.4696696698665619,
"learning_rate": 9.82098658229254e-06,
"loss": 0.5164,
"step": 320
},
{
"epoch": 0.17751152073732718,
"grad_norm": 0.5530909299850464,
"learning_rate": 9.818417344814587e-06,
"loss": 0.5275,
"step": 321
},
{
"epoch": 0.17806451612903226,
"grad_norm": 0.609788179397583,
"learning_rate": 9.815830142266457e-06,
"loss": 0.4716,
"step": 322
},
{
"epoch": 0.17861751152073732,
"grad_norm": 0.46327316761016846,
"learning_rate": 9.813224984294318e-06,
"loss": 0.4887,
"step": 323
},
{
"epoch": 0.1791705069124424,
"grad_norm": 0.5943720936775208,
"learning_rate": 9.810601880611286e-06,
"loss": 0.5009,
"step": 324
},
{
"epoch": 0.17972350230414746,
"grad_norm": 0.4262332320213318,
"learning_rate": 9.807960840997387e-06,
"loss": 0.4747,
"step": 325
},
{
"epoch": 0.18027649769585252,
"grad_norm": 0.47295740246772766,
"learning_rate": 9.805301875299518e-06,
"loss": 0.5092,
"step": 326
},
{
"epoch": 0.1808294930875576,
"grad_norm": 0.4910508394241333,
"learning_rate": 9.802624993431414e-06,
"loss": 0.5195,
"step": 327
},
{
"epoch": 0.18138248847926267,
"grad_norm": 0.5312582850456238,
"learning_rate": 9.799930205373605e-06,
"loss": 0.5077,
"step": 328
},
{
"epoch": 0.18193548387096775,
"grad_norm": 0.49236226081848145,
"learning_rate": 9.797217521173385e-06,
"loss": 0.5228,
"step": 329
},
{
"epoch": 0.1824884792626728,
"grad_norm": 0.5442107319831848,
"learning_rate": 9.794486950944775e-06,
"loss": 0.4909,
"step": 330
},
{
"epoch": 0.18304147465437787,
"grad_norm": 0.4408930242061615,
"learning_rate": 9.79173850486848e-06,
"loss": 0.4919,
"step": 331
},
{
"epoch": 0.18359447004608295,
"grad_norm": 0.4888366460800171,
"learning_rate": 9.78897219319185e-06,
"loss": 0.5247,
"step": 332
},
{
"epoch": 0.184147465437788,
"grad_norm": 0.48040100932121277,
"learning_rate": 9.786188026228853e-06,
"loss": 0.4746,
"step": 333
},
{
"epoch": 0.1847004608294931,
"grad_norm": 0.46840646862983704,
"learning_rate": 9.783386014360024e-06,
"loss": 0.5174,
"step": 334
},
{
"epoch": 0.18525345622119815,
"grad_norm": 0.4824281930923462,
"learning_rate": 9.780566168032432e-06,
"loss": 0.4944,
"step": 335
},
{
"epoch": 0.18580645161290324,
"grad_norm": 0.4401717483997345,
"learning_rate": 9.77772849775964e-06,
"loss": 0.5074,
"step": 336
},
{
"epoch": 0.1863594470046083,
"grad_norm": 0.5038358569145203,
"learning_rate": 9.774873014121667e-06,
"loss": 0.5052,
"step": 337
},
{
"epoch": 0.18691244239631336,
"grad_norm": 0.47577494382858276,
"learning_rate": 9.771999727764946e-06,
"loss": 0.4856,
"step": 338
},
{
"epoch": 0.18746543778801844,
"grad_norm": 0.4989863634109497,
"learning_rate": 9.76910864940229e-06,
"loss": 0.4839,
"step": 339
},
{
"epoch": 0.1880184331797235,
"grad_norm": 0.49821236729621887,
"learning_rate": 9.766199789812845e-06,
"loss": 0.498,
"step": 340
},
{
"epoch": 0.18857142857142858,
"grad_norm": 0.44591957330703735,
"learning_rate": 9.763273159842052e-06,
"loss": 0.4832,
"step": 341
},
{
"epoch": 0.18912442396313364,
"grad_norm": 0.4908744990825653,
"learning_rate": 9.76032877040161e-06,
"loss": 0.4927,
"step": 342
},
{
"epoch": 0.1896774193548387,
"grad_norm": 0.55243319272995,
"learning_rate": 9.757366632469427e-06,
"loss": 0.5005,
"step": 343
},
{
"epoch": 0.19023041474654379,
"grad_norm": 0.4234575629234314,
"learning_rate": 9.754386757089596e-06,
"loss": 0.4799,
"step": 344
},
{
"epoch": 0.19078341013824884,
"grad_norm": 0.49401965737342834,
"learning_rate": 9.751389155372329e-06,
"loss": 0.4942,
"step": 345
},
{
"epoch": 0.19133640552995393,
"grad_norm": 0.4622778594493866,
"learning_rate": 9.74837383849394e-06,
"loss": 0.4644,
"step": 346
},
{
"epoch": 0.191889400921659,
"grad_norm": 0.4600273072719574,
"learning_rate": 9.745340817696787e-06,
"loss": 0.5019,
"step": 347
},
{
"epoch": 0.19244239631336404,
"grad_norm": 0.4513567388057709,
"learning_rate": 9.742290104289237e-06,
"loss": 0.4704,
"step": 348
},
{
"epoch": 0.19299539170506913,
"grad_norm": 0.5180230736732483,
"learning_rate": 9.739221709645622e-06,
"loss": 0.4917,
"step": 349
},
{
"epoch": 0.1935483870967742,
"grad_norm": 0.44200998544692993,
"learning_rate": 9.736135645206198e-06,
"loss": 0.5047,
"step": 350
},
{
"epoch": 0.19410138248847927,
"grad_norm": 0.5660358667373657,
"learning_rate": 9.7330319224771e-06,
"loss": 0.5023,
"step": 351
},
{
"epoch": 0.19465437788018433,
"grad_norm": 0.44904786348342896,
"learning_rate": 9.729910553030304e-06,
"loss": 0.4998,
"step": 352
},
{
"epoch": 0.1952073732718894,
"grad_norm": 0.523826003074646,
"learning_rate": 9.726771548503575e-06,
"loss": 0.4966,
"step": 353
},
{
"epoch": 0.19576036866359448,
"grad_norm": 0.4515267312526703,
"learning_rate": 9.72361492060043e-06,
"loss": 0.4996,
"step": 354
},
{
"epoch": 0.19631336405529953,
"grad_norm": 0.4984499216079712,
"learning_rate": 9.7204406810901e-06,
"loss": 0.4985,
"step": 355
},
{
"epoch": 0.19686635944700462,
"grad_norm": 0.4689009189605713,
"learning_rate": 9.71724884180747e-06,
"loss": 0.4936,
"step": 356
},
{
"epoch": 0.19741935483870968,
"grad_norm": 0.46954217553138733,
"learning_rate": 9.714039414653047e-06,
"loss": 0.491,
"step": 357
},
{
"epoch": 0.19797235023041473,
"grad_norm": 0.5114353895187378,
"learning_rate": 9.71081241159292e-06,
"loss": 0.488,
"step": 358
},
{
"epoch": 0.19852534562211982,
"grad_norm": 0.4611765742301941,
"learning_rate": 9.707567844658698e-06,
"loss": 0.4749,
"step": 359
},
{
"epoch": 0.19907834101382488,
"grad_norm": 0.5052332878112793,
"learning_rate": 9.704305725947483e-06,
"loss": 0.5234,
"step": 360
},
{
"epoch": 0.19963133640552996,
"grad_norm": 0.5130250453948975,
"learning_rate": 9.701026067621813e-06,
"loss": 0.5006,
"step": 361
},
{
"epoch": 0.20018433179723502,
"grad_norm": 0.47562843561172485,
"learning_rate": 9.697728881909622e-06,
"loss": 0.529,
"step": 362
},
{
"epoch": 0.20073732718894008,
"grad_norm": 0.48228660225868225,
"learning_rate": 9.694414181104197e-06,
"loss": 0.5057,
"step": 363
},
{
"epoch": 0.20129032258064516,
"grad_norm": 0.501700222492218,
"learning_rate": 9.691081977564124e-06,
"loss": 0.4949,
"step": 364
},
{
"epoch": 0.20184331797235022,
"grad_norm": 0.4578889012336731,
"learning_rate": 9.687732283713247e-06,
"loss": 0.4989,
"step": 365
},
{
"epoch": 0.2023963133640553,
"grad_norm": 0.5223492980003357,
"learning_rate": 9.684365112040622e-06,
"loss": 0.4772,
"step": 366
},
{
"epoch": 0.20294930875576037,
"grad_norm": 0.4984261393547058,
"learning_rate": 9.680980475100471e-06,
"loss": 0.5132,
"step": 367
},
{
"epoch": 0.20350230414746545,
"grad_norm": 0.4805150032043457,
"learning_rate": 9.677578385512135e-06,
"loss": 0.5069,
"step": 368
},
{
"epoch": 0.2040552995391705,
"grad_norm": 0.46725666522979736,
"learning_rate": 9.674158855960017e-06,
"loss": 0.4806,
"step": 369
},
{
"epoch": 0.20460829493087557,
"grad_norm": 0.47330862283706665,
"learning_rate": 9.670721899193556e-06,
"loss": 0.5127,
"step": 370
},
{
"epoch": 0.20516129032258065,
"grad_norm": 0.5010029077529907,
"learning_rate": 9.667267528027157e-06,
"loss": 0.5189,
"step": 371
},
{
"epoch": 0.2057142857142857,
"grad_norm": 0.5415194034576416,
"learning_rate": 9.663795755340159e-06,
"loss": 0.5053,
"step": 372
},
{
"epoch": 0.2062672811059908,
"grad_norm": 0.4000757932662964,
"learning_rate": 9.660306594076779e-06,
"loss": 0.4939,
"step": 373
},
{
"epoch": 0.20682027649769585,
"grad_norm": 0.5064148306846619,
"learning_rate": 9.656800057246065e-06,
"loss": 0.4937,
"step": 374
},
{
"epoch": 0.2073732718894009,
"grad_norm": 0.467369943857193,
"learning_rate": 9.65327615792185e-06,
"loss": 0.5215,
"step": 375
},
{
"epoch": 0.207926267281106,
"grad_norm": 0.4753814935684204,
"learning_rate": 9.649734909242699e-06,
"loss": 0.4903,
"step": 376
},
{
"epoch": 0.20847926267281106,
"grad_norm": 0.4203657805919647,
"learning_rate": 9.646176324411867e-06,
"loss": 0.4874,
"step": 377
},
{
"epoch": 0.20903225806451614,
"grad_norm": 0.4930543601512909,
"learning_rate": 9.642600416697242e-06,
"loss": 0.4816,
"step": 378
},
{
"epoch": 0.2095852534562212,
"grad_norm": 0.4189959764480591,
"learning_rate": 9.639007199431298e-06,
"loss": 0.491,
"step": 379
},
{
"epoch": 0.21013824884792626,
"grad_norm": 0.48571497201919556,
"learning_rate": 9.635396686011052e-06,
"loss": 0.5005,
"step": 380
},
{
"epoch": 0.21069124423963134,
"grad_norm": 0.4344756007194519,
"learning_rate": 9.631768889898004e-06,
"loss": 0.5113,
"step": 381
},
{
"epoch": 0.2112442396313364,
"grad_norm": 0.5239787101745605,
"learning_rate": 9.628123824618087e-06,
"loss": 0.5013,
"step": 382
},
{
"epoch": 0.21179723502304149,
"grad_norm": 0.40745410323143005,
"learning_rate": 9.624461503761628e-06,
"loss": 0.4824,
"step": 383
},
{
"epoch": 0.21235023041474654,
"grad_norm": 0.5618027448654175,
"learning_rate": 9.620781940983288e-06,
"loss": 0.4852,
"step": 384
},
{
"epoch": 0.2129032258064516,
"grad_norm": 0.5030412673950195,
"learning_rate": 9.617085150002012e-06,
"loss": 0.4868,
"step": 385
},
{
"epoch": 0.2134562211981567,
"grad_norm": 0.5133850574493408,
"learning_rate": 9.613371144600976e-06,
"loss": 0.5076,
"step": 386
},
{
"epoch": 0.21400921658986174,
"grad_norm": 0.47164881229400635,
"learning_rate": 9.609639938627546e-06,
"loss": 0.5076,
"step": 387
},
{
"epoch": 0.21456221198156683,
"grad_norm": 0.4972430467605591,
"learning_rate": 9.605891545993214e-06,
"loss": 0.4843,
"step": 388
},
{
"epoch": 0.2151152073732719,
"grad_norm": 0.4716038405895233,
"learning_rate": 9.60212598067355e-06,
"loss": 0.4782,
"step": 389
},
{
"epoch": 0.21566820276497695,
"grad_norm": 0.4822791814804077,
"learning_rate": 9.59834325670815e-06,
"loss": 0.4874,
"step": 390
},
{
"epoch": 0.21622119815668203,
"grad_norm": 0.4538308382034302,
"learning_rate": 9.594543388200592e-06,
"loss": 0.501,
"step": 391
},
{
"epoch": 0.2167741935483871,
"grad_norm": 0.4863637387752533,
"learning_rate": 9.59072638931837e-06,
"loss": 0.5115,
"step": 392
},
{
"epoch": 0.21732718894009218,
"grad_norm": 0.4747362434864044,
"learning_rate": 9.586892274292846e-06,
"loss": 0.5262,
"step": 393
},
{
"epoch": 0.21788018433179723,
"grad_norm": 0.4688088595867157,
"learning_rate": 9.583041057419203e-06,
"loss": 0.517,
"step": 394
},
{
"epoch": 0.2184331797235023,
"grad_norm": 0.49005988240242004,
"learning_rate": 9.579172753056383e-06,
"loss": 0.4737,
"step": 395
},
{
"epoch": 0.21898617511520738,
"grad_norm": 0.4679119288921356,
"learning_rate": 9.575287375627037e-06,
"loss": 0.4978,
"step": 396
},
{
"epoch": 0.21953917050691243,
"grad_norm": 0.47708967328071594,
"learning_rate": 9.571384939617478e-06,
"loss": 0.5005,
"step": 397
},
{
"epoch": 0.22009216589861752,
"grad_norm": 0.47183945775032043,
"learning_rate": 9.567465459577613e-06,
"loss": 0.4708,
"step": 398
},
{
"epoch": 0.22064516129032258,
"grad_norm": 0.47582703828811646,
"learning_rate": 9.563528950120895e-06,
"loss": 0.4983,
"step": 399
},
{
"epoch": 0.22119815668202766,
"grad_norm": 0.48260581493377686,
"learning_rate": 9.559575425924279e-06,
"loss": 0.4894,
"step": 400
},
{
"epoch": 0.22175115207373272,
"grad_norm": 0.5306166410446167,
"learning_rate": 9.55560490172815e-06,
"loss": 0.5068,
"step": 401
},
{
"epoch": 0.22230414746543778,
"grad_norm": 0.49763691425323486,
"learning_rate": 9.551617392336281e-06,
"loss": 0.5028,
"step": 402
},
{
"epoch": 0.22285714285714286,
"grad_norm": 0.5285375118255615,
"learning_rate": 9.547612912615769e-06,
"loss": 0.484,
"step": 403
},
{
"epoch": 0.22341013824884792,
"grad_norm": 0.5528775453567505,
"learning_rate": 9.543591477496985e-06,
"loss": 0.4774,
"step": 404
},
{
"epoch": 0.223963133640553,
"grad_norm": 0.48812106251716614,
"learning_rate": 9.539553101973516e-06,
"loss": 0.5042,
"step": 405
},
{
"epoch": 0.22451612903225807,
"grad_norm": 0.5272664427757263,
"learning_rate": 9.535497801102114e-06,
"loss": 0.4987,
"step": 406
},
{
"epoch": 0.22506912442396312,
"grad_norm": 0.5162638425827026,
"learning_rate": 9.531425590002629e-06,
"loss": 0.4942,
"step": 407
},
{
"epoch": 0.2256221198156682,
"grad_norm": 0.45863449573516846,
"learning_rate": 9.527336483857965e-06,
"loss": 0.5165,
"step": 408
},
{
"epoch": 0.22617511520737327,
"grad_norm": 0.5142653584480286,
"learning_rate": 9.523230497914016e-06,
"loss": 0.4883,
"step": 409
},
{
"epoch": 0.22672811059907835,
"grad_norm": 0.4717908203601837,
"learning_rate": 9.519107647479609e-06,
"loss": 0.4859,
"step": 410
},
{
"epoch": 0.2272811059907834,
"grad_norm": 0.4761084020137787,
"learning_rate": 9.514967947926453e-06,
"loss": 0.5022,
"step": 411
},
{
"epoch": 0.22783410138248847,
"grad_norm": 0.4796617925167084,
"learning_rate": 9.510811414689073e-06,
"loss": 0.5058,
"step": 412
},
{
"epoch": 0.22838709677419355,
"grad_norm": 0.5060638189315796,
"learning_rate": 9.506638063264759e-06,
"loss": 0.4859,
"step": 413
},
{
"epoch": 0.2289400921658986,
"grad_norm": 0.42853257060050964,
"learning_rate": 9.502447909213508e-06,
"loss": 0.5004,
"step": 414
},
{
"epoch": 0.2294930875576037,
"grad_norm": 0.45876848697662354,
"learning_rate": 9.498240968157962e-06,
"loss": 0.4898,
"step": 415
},
{
"epoch": 0.23004608294930876,
"grad_norm": 0.4455052614212036,
"learning_rate": 9.49401725578335e-06,
"loss": 0.4972,
"step": 416
},
{
"epoch": 0.2305990783410138,
"grad_norm": 0.41394197940826416,
"learning_rate": 9.489776787837438e-06,
"loss": 0.5006,
"step": 417
},
{
"epoch": 0.2311520737327189,
"grad_norm": 0.44113320112228394,
"learning_rate": 9.485519580130456e-06,
"loss": 0.4644,
"step": 418
},
{
"epoch": 0.23170506912442396,
"grad_norm": 0.4882364869117737,
"learning_rate": 9.481245648535053e-06,
"loss": 0.4726,
"step": 419
},
{
"epoch": 0.23225806451612904,
"grad_norm": 0.4658440053462982,
"learning_rate": 9.476955008986228e-06,
"loss": 0.497,
"step": 420
},
{
"epoch": 0.2328110599078341,
"grad_norm": 0.4944906532764435,
"learning_rate": 9.472647677481275e-06,
"loss": 0.4743,
"step": 421
},
{
"epoch": 0.23336405529953916,
"grad_norm": 0.4508844316005707,
"learning_rate": 9.468323670079725e-06,
"loss": 0.4884,
"step": 422
},
{
"epoch": 0.23391705069124424,
"grad_norm": 0.470225065946579,
"learning_rate": 9.463983002903278e-06,
"loss": 0.4929,
"step": 423
},
{
"epoch": 0.2344700460829493,
"grad_norm": 0.5286823511123657,
"learning_rate": 9.459625692135756e-06,
"loss": 0.4938,
"step": 424
},
{
"epoch": 0.2350230414746544,
"grad_norm": 0.45246848464012146,
"learning_rate": 9.45525175402303e-06,
"loss": 0.4834,
"step": 425
},
{
"epoch": 0.23557603686635945,
"grad_norm": 0.48045429587364197,
"learning_rate": 9.450861204872965e-06,
"loss": 0.4877,
"step": 426
},
{
"epoch": 0.2361290322580645,
"grad_norm": 0.5548929572105408,
"learning_rate": 9.44645406105536e-06,
"loss": 0.5185,
"step": 427
},
{
"epoch": 0.2366820276497696,
"grad_norm": 0.4929436147212982,
"learning_rate": 9.44203033900189e-06,
"loss": 0.5048,
"step": 428
},
{
"epoch": 0.23723502304147465,
"grad_norm": 0.46492716670036316,
"learning_rate": 9.437590055206032e-06,
"loss": 0.4939,
"step": 429
},
{
"epoch": 0.23778801843317973,
"grad_norm": 0.4560331404209137,
"learning_rate": 9.433133226223018e-06,
"loss": 0.5016,
"step": 430
},
{
"epoch": 0.2383410138248848,
"grad_norm": 0.4622935652732849,
"learning_rate": 9.428659868669765e-06,
"loss": 0.4748,
"step": 431
},
{
"epoch": 0.23889400921658988,
"grad_norm": 0.4686061441898346,
"learning_rate": 9.424169999224819e-06,
"loss": 0.5012,
"step": 432
},
{
"epoch": 0.23944700460829493,
"grad_norm": 0.47005850076675415,
"learning_rate": 9.419663634628283e-06,
"loss": 0.4902,
"step": 433
},
{
"epoch": 0.24,
"grad_norm": 0.5422276854515076,
"learning_rate": 9.415140791681767e-06,
"loss": 0.4923,
"step": 434
},
{
"epoch": 0.24055299539170508,
"grad_norm": 0.42137962579727173,
"learning_rate": 9.410601487248315e-06,
"loss": 0.4762,
"step": 435
},
{
"epoch": 0.24110599078341013,
"grad_norm": 0.518340528011322,
"learning_rate": 9.406045738252349e-06,
"loss": 0.5257,
"step": 436
},
{
"epoch": 0.24165898617511522,
"grad_norm": 0.5296422243118286,
"learning_rate": 9.4014735616796e-06,
"loss": 0.4959,
"step": 437
},
{
"epoch": 0.24221198156682028,
"grad_norm": 0.4660092890262604,
"learning_rate": 9.396884974577052e-06,
"loss": 0.4902,
"step": 438
},
{
"epoch": 0.24276497695852534,
"grad_norm": 0.5939272046089172,
"learning_rate": 9.392279994052872e-06,
"loss": 0.4859,
"step": 439
},
{
"epoch": 0.24331797235023042,
"grad_norm": 0.4680648446083069,
"learning_rate": 9.387658637276348e-06,
"loss": 0.4856,
"step": 440
},
{
"epoch": 0.24387096774193548,
"grad_norm": 0.4640378952026367,
"learning_rate": 9.383020921477824e-06,
"loss": 0.4976,
"step": 441
},
{
"epoch": 0.24442396313364056,
"grad_norm": 0.5372658371925354,
"learning_rate": 9.378366863948644e-06,
"loss": 0.4731,
"step": 442
},
{
"epoch": 0.24497695852534562,
"grad_norm": 0.48611563444137573,
"learning_rate": 9.373696482041072e-06,
"loss": 0.474,
"step": 443
},
{
"epoch": 0.24552995391705068,
"grad_norm": 0.4766850173473358,
"learning_rate": 9.369009793168243e-06,
"loss": 0.4878,
"step": 444
},
{
"epoch": 0.24608294930875577,
"grad_norm": 0.503322184085846,
"learning_rate": 9.364306814804086e-06,
"loss": 0.49,
"step": 445
},
{
"epoch": 0.24663594470046082,
"grad_norm": 0.5735645294189453,
"learning_rate": 9.35958756448327e-06,
"loss": 0.5357,
"step": 446
},
{
"epoch": 0.2471889400921659,
"grad_norm": 0.5782787799835205,
"learning_rate": 9.354852059801127e-06,
"loss": 0.5052,
"step": 447
},
{
"epoch": 0.24774193548387097,
"grad_norm": 0.48679542541503906,
"learning_rate": 9.350100318413594e-06,
"loss": 0.4824,
"step": 448
},
{
"epoch": 0.24829493087557603,
"grad_norm": 0.535241961479187,
"learning_rate": 9.345332358037147e-06,
"loss": 0.4814,
"step": 449
},
{
"epoch": 0.2488479262672811,
"grad_norm": 0.5126506090164185,
"learning_rate": 9.340548196448729e-06,
"loss": 0.4903,
"step": 450
},
{
"epoch": 0.24940092165898617,
"grad_norm": 0.459888219833374,
"learning_rate": 9.33574785148569e-06,
"loss": 0.4764,
"step": 451
},
{
"epoch": 0.24995391705069125,
"grad_norm": 0.47858306765556335,
"learning_rate": 9.330931341045723e-06,
"loss": 0.4759,
"step": 452
},
{
"epoch": 0.2505069124423963,
"grad_norm": 0.4925948977470398,
"learning_rate": 9.326098683086786e-06,
"loss": 0.5092,
"step": 453
},
{
"epoch": 0.25105990783410137,
"grad_norm": 0.4197613000869751,
"learning_rate": 9.321249895627043e-06,
"loss": 0.4855,
"step": 454
},
{
"epoch": 0.25161290322580643,
"grad_norm": 0.4368072748184204,
"learning_rate": 9.316384996744798e-06,
"loss": 0.4731,
"step": 455
},
{
"epoch": 0.25216589861751154,
"grad_norm": 0.4712340533733368,
"learning_rate": 9.311504004578425e-06,
"loss": 0.4864,
"step": 456
},
{
"epoch": 0.2527188940092166,
"grad_norm": 0.4518265426158905,
"learning_rate": 9.306606937326295e-06,
"loss": 0.5119,
"step": 457
},
{
"epoch": 0.25327188940092166,
"grad_norm": 0.48842400312423706,
"learning_rate": 9.301693813246721e-06,
"loss": 0.4732,
"step": 458
},
{
"epoch": 0.2538248847926267,
"grad_norm": 0.5565005540847778,
"learning_rate": 9.296764650657881e-06,
"loss": 0.5018,
"step": 459
},
{
"epoch": 0.2543778801843318,
"grad_norm": 0.4874143600463867,
"learning_rate": 9.291819467937746e-06,
"loss": 0.4712,
"step": 460
},
{
"epoch": 0.2549308755760369,
"grad_norm": 0.5408375859260559,
"learning_rate": 9.286858283524022e-06,
"loss": 0.4785,
"step": 461
},
{
"epoch": 0.25548387096774194,
"grad_norm": 0.4784450829029083,
"learning_rate": 9.281881115914075e-06,
"loss": 0.4893,
"step": 462
},
{
"epoch": 0.256036866359447,
"grad_norm": 0.4540750980377197,
"learning_rate": 9.276887983664861e-06,
"loss": 0.4873,
"step": 463
},
{
"epoch": 0.25658986175115206,
"grad_norm": 0.5902926921844482,
"learning_rate": 9.271878905392863e-06,
"loss": 0.4903,
"step": 464
},
{
"epoch": 0.2571428571428571,
"grad_norm": 0.44395557045936584,
"learning_rate": 9.26685389977401e-06,
"loss": 0.4862,
"step": 465
},
{
"epoch": 0.25769585253456223,
"grad_norm": 0.44266945123672485,
"learning_rate": 9.261812985543625e-06,
"loss": 0.4875,
"step": 466
},
{
"epoch": 0.2582488479262673,
"grad_norm": 0.5877528190612793,
"learning_rate": 9.256756181496334e-06,
"loss": 0.498,
"step": 467
},
{
"epoch": 0.25880184331797235,
"grad_norm": 0.43955346941947937,
"learning_rate": 9.251683506486016e-06,
"loss": 0.4782,
"step": 468
},
{
"epoch": 0.2593548387096774,
"grad_norm": 0.5338640809059143,
"learning_rate": 9.246594979425719e-06,
"loss": 0.4834,
"step": 469
},
{
"epoch": 0.25990783410138246,
"grad_norm": 0.4647727608680725,
"learning_rate": 9.241490619287593e-06,
"loss": 0.4957,
"step": 470
},
{
"epoch": 0.2604608294930876,
"grad_norm": 0.3950575590133667,
"learning_rate": 9.236370445102825e-06,
"loss": 0.4864,
"step": 471
},
{
"epoch": 0.26101382488479263,
"grad_norm": 0.5759503245353699,
"learning_rate": 9.231234475961559e-06,
"loss": 0.5049,
"step": 472
},
{
"epoch": 0.2615668202764977,
"grad_norm": 0.49982547760009766,
"learning_rate": 9.22608273101283e-06,
"loss": 0.5099,
"step": 473
},
{
"epoch": 0.26211981566820275,
"grad_norm": 0.45275378227233887,
"learning_rate": 9.220915229464496e-06,
"loss": 0.4768,
"step": 474
},
{
"epoch": 0.2626728110599078,
"grad_norm": 0.4712145924568176,
"learning_rate": 9.215731990583155e-06,
"loss": 0.4418,
"step": 475
},
{
"epoch": 0.2632258064516129,
"grad_norm": 0.4405660927295685,
"learning_rate": 9.210533033694084e-06,
"loss": 0.4938,
"step": 476
},
{
"epoch": 0.263778801843318,
"grad_norm": 0.4714353382587433,
"learning_rate": 9.205318378181167e-06,
"loss": 0.4806,
"step": 477
},
{
"epoch": 0.26433179723502304,
"grad_norm": 0.5084551572799683,
"learning_rate": 9.200088043486813e-06,
"loss": 0.487,
"step": 478
},
{
"epoch": 0.2648847926267281,
"grad_norm": 0.4339804947376251,
"learning_rate": 9.194842049111889e-06,
"loss": 0.4703,
"step": 479
},
{
"epoch": 0.2654377880184332,
"grad_norm": 0.5167462825775146,
"learning_rate": 9.189580414615658e-06,
"loss": 0.495,
"step": 480
},
{
"epoch": 0.26599078341013827,
"grad_norm": 0.5027214288711548,
"learning_rate": 9.184303159615682e-06,
"loss": 0.487,
"step": 481
},
{
"epoch": 0.2665437788018433,
"grad_norm": 0.5376026034355164,
"learning_rate": 9.179010303787772e-06,
"loss": 0.4974,
"step": 482
},
{
"epoch": 0.2670967741935484,
"grad_norm": 0.4745848774909973,
"learning_rate": 9.173701866865905e-06,
"loss": 0.4761,
"step": 483
},
{
"epoch": 0.26764976958525344,
"grad_norm": 0.615476667881012,
"learning_rate": 9.168377868642142e-06,
"loss": 0.4903,
"step": 484
},
{
"epoch": 0.26820276497695855,
"grad_norm": 0.43914616107940674,
"learning_rate": 9.163038328966578e-06,
"loss": 0.5018,
"step": 485
},
{
"epoch": 0.2687557603686636,
"grad_norm": 0.5516188144683838,
"learning_rate": 9.157683267747239e-06,
"loss": 0.5042,
"step": 486
},
{
"epoch": 0.26930875576036867,
"grad_norm": 0.5231643319129944,
"learning_rate": 9.152312704950028e-06,
"loss": 0.4745,
"step": 487
},
{
"epoch": 0.2698617511520737,
"grad_norm": 0.4931274354457855,
"learning_rate": 9.146926660598646e-06,
"loss": 0.4891,
"step": 488
},
{
"epoch": 0.2704147465437788,
"grad_norm": 0.5022376179695129,
"learning_rate": 9.141525154774513e-06,
"loss": 0.4707,
"step": 489
},
{
"epoch": 0.2709677419354839,
"grad_norm": 0.4848520755767822,
"learning_rate": 9.136108207616694e-06,
"loss": 0.5176,
"step": 490
},
{
"epoch": 0.27152073732718895,
"grad_norm": 0.5185691118240356,
"learning_rate": 9.130675839321829e-06,
"loss": 0.4951,
"step": 491
},
{
"epoch": 0.272073732718894,
"grad_norm": 0.4987480938434601,
"learning_rate": 9.125228070144056e-06,
"loss": 0.4865,
"step": 492
},
{
"epoch": 0.27262672811059907,
"grad_norm": 0.47905007004737854,
"learning_rate": 9.119764920394928e-06,
"loss": 0.4931,
"step": 493
},
{
"epoch": 0.27317972350230413,
"grad_norm": 0.5060631632804871,
"learning_rate": 9.114286410443349e-06,
"loss": 0.4912,
"step": 494
},
{
"epoch": 0.27373271889400924,
"grad_norm": 0.5127936005592346,
"learning_rate": 9.108792560715487e-06,
"loss": 0.5004,
"step": 495
},
{
"epoch": 0.2742857142857143,
"grad_norm": 0.5518050193786621,
"learning_rate": 9.103283391694711e-06,
"loss": 0.5051,
"step": 496
},
{
"epoch": 0.27483870967741936,
"grad_norm": 0.5080561637878418,
"learning_rate": 9.097758923921495e-06,
"loss": 0.4679,
"step": 497
},
{
"epoch": 0.2753917050691244,
"grad_norm": 0.5666584372520447,
"learning_rate": 9.092219177993365e-06,
"loss": 0.4861,
"step": 498
},
{
"epoch": 0.2759447004608295,
"grad_norm": 0.48917266726493835,
"learning_rate": 9.086664174564804e-06,
"loss": 0.4616,
"step": 499
},
{
"epoch": 0.2764976958525346,
"grad_norm": 0.4538862407207489,
"learning_rate": 9.081093934347178e-06,
"loss": 0.4917,
"step": 500
},
{
"epoch": 0.27705069124423964,
"grad_norm": 0.6453654170036316,
"learning_rate": 9.075508478108674e-06,
"loss": 0.477,
"step": 501
},
{
"epoch": 0.2776036866359447,
"grad_norm": 0.5007541179656982,
"learning_rate": 9.069907826674199e-06,
"loss": 0.4748,
"step": 502
},
{
"epoch": 0.27815668202764976,
"grad_norm": 0.5134044289588928,
"learning_rate": 9.064292000925316e-06,
"loss": 0.5122,
"step": 503
},
{
"epoch": 0.2787096774193548,
"grad_norm": 0.5208799242973328,
"learning_rate": 9.05866102180017e-06,
"loss": 0.4908,
"step": 504
},
{
"epoch": 0.27926267281105993,
"grad_norm": 0.4694979786872864,
"learning_rate": 9.053014910293399e-06,
"loss": 0.5006,
"step": 505
},
{
"epoch": 0.279815668202765,
"grad_norm": 0.4957534670829773,
"learning_rate": 9.047353687456064e-06,
"loss": 0.4901,
"step": 506
},
{
"epoch": 0.28036866359447005,
"grad_norm": 0.5231026411056519,
"learning_rate": 9.041677374395563e-06,
"loss": 0.5062,
"step": 507
},
{
"epoch": 0.2809216589861751,
"grad_norm": 0.5049005150794983,
"learning_rate": 9.035985992275562e-06,
"loss": 0.501,
"step": 508
},
{
"epoch": 0.28147465437788016,
"grad_norm": 0.48721855878829956,
"learning_rate": 9.030279562315907e-06,
"loss": 0.4881,
"step": 509
},
{
"epoch": 0.2820276497695853,
"grad_norm": 0.47729742527008057,
"learning_rate": 9.02455810579255e-06,
"loss": 0.4734,
"step": 510
},
{
"epoch": 0.28258064516129033,
"grad_norm": 0.5377198457717896,
"learning_rate": 9.018821644037466e-06,
"loss": 0.4655,
"step": 511
},
{
"epoch": 0.2831336405529954,
"grad_norm": 0.5237789154052734,
"learning_rate": 9.01307019843858e-06,
"loss": 0.5081,
"step": 512
},
{
"epoch": 0.28368663594470045,
"grad_norm": 0.6095771193504333,
"learning_rate": 9.00730379043968e-06,
"loss": 0.4938,
"step": 513
},
{
"epoch": 0.2842396313364055,
"grad_norm": 0.5196386575698853,
"learning_rate": 9.001522441540342e-06,
"loss": 0.4901,
"step": 514
},
{
"epoch": 0.2847926267281106,
"grad_norm": 0.5163764357566833,
"learning_rate": 8.995726173295845e-06,
"loss": 0.4901,
"step": 515
},
{
"epoch": 0.2853456221198157,
"grad_norm": 0.5501850247383118,
"learning_rate": 8.989915007317095e-06,
"loss": 0.4884,
"step": 516
},
{
"epoch": 0.28589861751152074,
"grad_norm": 0.4907573461532593,
"learning_rate": 8.984088965270547e-06,
"loss": 0.4803,
"step": 517
},
{
"epoch": 0.2864516129032258,
"grad_norm": 0.4490540325641632,
"learning_rate": 8.978248068878113e-06,
"loss": 0.5025,
"step": 518
},
{
"epoch": 0.28700460829493085,
"grad_norm": 0.541654109954834,
"learning_rate": 8.97239233991709e-06,
"loss": 0.4844,
"step": 519
},
{
"epoch": 0.28755760368663597,
"grad_norm": 0.4671088457107544,
"learning_rate": 8.966521800220084e-06,
"loss": 0.4775,
"step": 520
},
{
"epoch": 0.288110599078341,
"grad_norm": 0.5631542801856995,
"learning_rate": 8.960636471674913e-06,
"loss": 0.4723,
"step": 521
},
{
"epoch": 0.2886635944700461,
"grad_norm": 0.5776397585868835,
"learning_rate": 8.954736376224536e-06,
"loss": 0.4949,
"step": 522
},
{
"epoch": 0.28921658986175114,
"grad_norm": 0.5042744874954224,
"learning_rate": 8.948821535866973e-06,
"loss": 0.4934,
"step": 523
},
{
"epoch": 0.2897695852534562,
"grad_norm": 0.5939375162124634,
"learning_rate": 8.942891972655217e-06,
"loss": 0.4884,
"step": 524
},
{
"epoch": 0.2903225806451613,
"grad_norm": 0.4696662127971649,
"learning_rate": 8.93694770869715e-06,
"loss": 0.5052,
"step": 525
},
{
"epoch": 0.29087557603686637,
"grad_norm": 0.45712941884994507,
"learning_rate": 8.930988766155472e-06,
"loss": 0.5488,
"step": 526
},
{
"epoch": 0.2914285714285714,
"grad_norm": 0.5518166422843933,
"learning_rate": 8.925015167247604e-06,
"loss": 0.5129,
"step": 527
},
{
"epoch": 0.2919815668202765,
"grad_norm": 0.446274071931839,
"learning_rate": 8.919026934245618e-06,
"loss": 0.504,
"step": 528
},
{
"epoch": 0.29253456221198154,
"grad_norm": 0.5146806836128235,
"learning_rate": 8.91302408947614e-06,
"loss": 0.4779,
"step": 529
},
{
"epoch": 0.29308755760368665,
"grad_norm": 0.4867715537548065,
"learning_rate": 8.907006655320287e-06,
"loss": 0.4743,
"step": 530
},
{
"epoch": 0.2936405529953917,
"grad_norm": 0.47047367691993713,
"learning_rate": 8.90097465421356e-06,
"loss": 0.4881,
"step": 531
},
{
"epoch": 0.29419354838709677,
"grad_norm": 0.46602970361709595,
"learning_rate": 8.894928108645772e-06,
"loss": 0.486,
"step": 532
},
{
"epoch": 0.29474654377880183,
"grad_norm": 0.45530185103416443,
"learning_rate": 8.888867041160975e-06,
"loss": 0.4726,
"step": 533
},
{
"epoch": 0.2952995391705069,
"grad_norm": 0.4860335886478424,
"learning_rate": 8.882791474357354e-06,
"loss": 0.4818,
"step": 534
},
{
"epoch": 0.295852534562212,
"grad_norm": 0.4780627191066742,
"learning_rate": 8.876701430887156e-06,
"loss": 0.4937,
"step": 535
},
{
"epoch": 0.29640552995391706,
"grad_norm": 0.488271027803421,
"learning_rate": 8.870596933456603e-06,
"loss": 0.4988,
"step": 536
},
{
"epoch": 0.2969585253456221,
"grad_norm": 0.48193061351776123,
"learning_rate": 8.864478004825812e-06,
"loss": 0.4925,
"step": 537
},
{
"epoch": 0.2975115207373272,
"grad_norm": 0.5294972658157349,
"learning_rate": 8.8583446678087e-06,
"loss": 0.4787,
"step": 538
},
{
"epoch": 0.29806451612903223,
"grad_norm": 0.48833131790161133,
"learning_rate": 8.852196945272906e-06,
"loss": 0.5055,
"step": 539
},
{
"epoch": 0.29861751152073734,
"grad_norm": 0.4671526551246643,
"learning_rate": 8.846034860139706e-06,
"loss": 0.4775,
"step": 540
},
{
"epoch": 0.2991705069124424,
"grad_norm": 0.6074989438056946,
"learning_rate": 8.839858435383918e-06,
"loss": 0.4894,
"step": 541
},
{
"epoch": 0.29972350230414746,
"grad_norm": 0.46650996804237366,
"learning_rate": 8.833667694033841e-06,
"loss": 0.4913,
"step": 542
},
{
"epoch": 0.3002764976958525,
"grad_norm": 0.7234087586402893,
"learning_rate": 8.827462659171134e-06,
"loss": 0.4794,
"step": 543
},
{
"epoch": 0.30082949308755763,
"grad_norm": 0.5737038850784302,
"learning_rate": 8.821243353930756e-06,
"loss": 0.5169,
"step": 544
},
{
"epoch": 0.3013824884792627,
"grad_norm": 0.6461325883865356,
"learning_rate": 8.815009801500875e-06,
"loss": 0.4846,
"step": 545
},
{
"epoch": 0.30193548387096775,
"grad_norm": 0.4608705937862396,
"learning_rate": 8.808762025122774e-06,
"loss": 0.4763,
"step": 546
},
{
"epoch": 0.3024884792626728,
"grad_norm": 0.6168321371078491,
"learning_rate": 8.802500048090769e-06,
"loss": 0.4906,
"step": 547
},
{
"epoch": 0.30304147465437786,
"grad_norm": 0.5032297968864441,
"learning_rate": 8.796223893752125e-06,
"loss": 0.4809,
"step": 548
},
{
"epoch": 0.303594470046083,
"grad_norm": 0.48676756024360657,
"learning_rate": 8.789933585506962e-06,
"loss": 0.4844,
"step": 549
},
{
"epoch": 0.30414746543778803,
"grad_norm": 0.5273764133453369,
"learning_rate": 8.783629146808175e-06,
"loss": 0.4979,
"step": 550
},
{
"epoch": 0.3047004608294931,
"grad_norm": 0.47446364164352417,
"learning_rate": 8.77731060116134e-06,
"loss": 0.4633,
"step": 551
},
{
"epoch": 0.30525345622119815,
"grad_norm": 0.5045877695083618,
"learning_rate": 8.77097797212463e-06,
"loss": 0.4907,
"step": 552
},
{
"epoch": 0.3058064516129032,
"grad_norm": 0.4806043803691864,
"learning_rate": 8.764631283308733e-06,
"loss": 0.4794,
"step": 553
},
{
"epoch": 0.3063594470046083,
"grad_norm": 0.5195322036743164,
"learning_rate": 8.758270558376752e-06,
"loss": 0.4921,
"step": 554
},
{
"epoch": 0.3069124423963134,
"grad_norm": 0.5429467558860779,
"learning_rate": 8.751895821044118e-06,
"loss": 0.4825,
"step": 555
},
{
"epoch": 0.30746543778801844,
"grad_norm": 0.48403802514076233,
"learning_rate": 8.745507095078515e-06,
"loss": 0.4997,
"step": 556
},
{
"epoch": 0.3080184331797235,
"grad_norm": 0.5179433822631836,
"learning_rate": 8.739104404299781e-06,
"loss": 0.4444,
"step": 557
},
{
"epoch": 0.30857142857142855,
"grad_norm": 0.5344632863998413,
"learning_rate": 8.732687772579816e-06,
"loss": 0.4695,
"step": 558
},
{
"epoch": 0.30912442396313367,
"grad_norm": 0.5367107391357422,
"learning_rate": 8.7262572238425e-06,
"loss": 0.453,
"step": 559
},
{
"epoch": 0.3096774193548387,
"grad_norm": 0.5187183618545532,
"learning_rate": 8.719812782063603e-06,
"loss": 0.4975,
"step": 560
},
{
"epoch": 0.3102304147465438,
"grad_norm": 0.4541904330253601,
"learning_rate": 8.713354471270691e-06,
"loss": 0.5075,
"step": 561
},
{
"epoch": 0.31078341013824884,
"grad_norm": 0.4493202865123749,
"learning_rate": 8.706882315543043e-06,
"loss": 0.458,
"step": 562
},
{
"epoch": 0.3113364055299539,
"grad_norm": 0.571712076663971,
"learning_rate": 8.700396339011554e-06,
"loss": 0.5057,
"step": 563
},
{
"epoch": 0.311889400921659,
"grad_norm": 0.4367184042930603,
"learning_rate": 8.69389656585865e-06,
"loss": 0.4725,
"step": 564
},
{
"epoch": 0.31244239631336407,
"grad_norm": 0.5127226710319519,
"learning_rate": 8.6873830203182e-06,
"loss": 0.4615,
"step": 565
},
{
"epoch": 0.3129953917050691,
"grad_norm": 0.5464975833892822,
"learning_rate": 8.680855726675419e-06,
"loss": 0.4661,
"step": 566
},
{
"epoch": 0.3135483870967742,
"grad_norm": 0.5116366147994995,
"learning_rate": 8.674314709266778e-06,
"loss": 0.4789,
"step": 567
},
{
"epoch": 0.31410138248847924,
"grad_norm": 0.5597204566001892,
"learning_rate": 8.667759992479922e-06,
"loss": 0.4903,
"step": 568
},
{
"epoch": 0.31465437788018435,
"grad_norm": 0.4036201536655426,
"learning_rate": 8.661191600753567e-06,
"loss": 0.5093,
"step": 569
},
{
"epoch": 0.3152073732718894,
"grad_norm": 0.6131529808044434,
"learning_rate": 8.65460955857742e-06,
"loss": 0.4572,
"step": 570
},
{
"epoch": 0.31576036866359447,
"grad_norm": 0.532192587852478,
"learning_rate": 8.648013890492081e-06,
"loss": 0.4679,
"step": 571
},
{
"epoch": 0.31631336405529953,
"grad_norm": 0.4397170841693878,
"learning_rate": 8.641404621088951e-06,
"loss": 0.4841,
"step": 572
},
{
"epoch": 0.3168663594470046,
"grad_norm": 0.44879838824272156,
"learning_rate": 8.634781775010144e-06,
"loss": 0.4853,
"step": 573
},
{
"epoch": 0.3174193548387097,
"grad_norm": 0.5257536768913269,
"learning_rate": 8.628145376948392e-06,
"loss": 0.4915,
"step": 574
},
{
"epoch": 0.31797235023041476,
"grad_norm": 0.4205341637134552,
"learning_rate": 8.621495451646958e-06,
"loss": 0.4895,
"step": 575
},
{
"epoch": 0.3185253456221198,
"grad_norm": 0.4227660298347473,
"learning_rate": 8.614832023899536e-06,
"loss": 0.488,
"step": 576
},
{
"epoch": 0.3190783410138249,
"grad_norm": 0.441709965467453,
"learning_rate": 8.608155118550167e-06,
"loss": 0.4756,
"step": 577
},
{
"epoch": 0.31963133640552993,
"grad_norm": 0.4320451021194458,
"learning_rate": 8.601464760493138e-06,
"loss": 0.4726,
"step": 578
},
{
"epoch": 0.32018433179723504,
"grad_norm": 0.3868134617805481,
"learning_rate": 8.594760974672898e-06,
"loss": 0.475,
"step": 579
},
{
"epoch": 0.3207373271889401,
"grad_norm": 0.43438073992729187,
"learning_rate": 8.588043786083952e-06,
"loss": 0.4874,
"step": 580
},
{
"epoch": 0.32129032258064516,
"grad_norm": 0.40946924686431885,
"learning_rate": 8.581313219770788e-06,
"loss": 0.4843,
"step": 581
},
{
"epoch": 0.3218433179723502,
"grad_norm": 0.4077948331832886,
"learning_rate": 8.57456930082776e-06,
"loss": 0.4656,
"step": 582
},
{
"epoch": 0.3223963133640553,
"grad_norm": 0.4392508566379547,
"learning_rate": 8.567812054399016e-06,
"loss": 0.4911,
"step": 583
},
{
"epoch": 0.3229493087557604,
"grad_norm": 0.4225884974002838,
"learning_rate": 8.561041505678385e-06,
"loss": 0.4907,
"step": 584
},
{
"epoch": 0.32350230414746545,
"grad_norm": 0.4444757103919983,
"learning_rate": 8.5542576799093e-06,
"loss": 0.484,
"step": 585
},
{
"epoch": 0.3240552995391705,
"grad_norm": 0.4226742088794708,
"learning_rate": 8.547460602384697e-06,
"loss": 0.4947,
"step": 586
},
{
"epoch": 0.32460829493087556,
"grad_norm": 0.5053589344024658,
"learning_rate": 8.540650298446912e-06,
"loss": 0.4984,
"step": 587
},
{
"epoch": 0.3251612903225806,
"grad_norm": 0.4462142586708069,
"learning_rate": 8.533826793487601e-06,
"loss": 0.4624,
"step": 588
},
{
"epoch": 0.32571428571428573,
"grad_norm": 0.5019125938415527,
"learning_rate": 8.526990112947636e-06,
"loss": 0.4597,
"step": 589
},
{
"epoch": 0.3262672811059908,
"grad_norm": 0.4495023787021637,
"learning_rate": 8.520140282317018e-06,
"loss": 0.4619,
"step": 590
},
{
"epoch": 0.32682027649769585,
"grad_norm": 0.45633986592292786,
"learning_rate": 8.51327732713477e-06,
"loss": 0.4997,
"step": 591
},
{
"epoch": 0.3273732718894009,
"grad_norm": 0.45402443408966064,
"learning_rate": 8.506401272988854e-06,
"loss": 0.4726,
"step": 592
},
{
"epoch": 0.32792626728110597,
"grad_norm": 0.4144057631492615,
"learning_rate": 8.499512145516068e-06,
"loss": 0.483,
"step": 593
},
{
"epoch": 0.3284792626728111,
"grad_norm": 0.506801962852478,
"learning_rate": 8.492609970401955e-06,
"loss": 0.4936,
"step": 594
},
{
"epoch": 0.32903225806451614,
"grad_norm": 0.45201730728149414,
"learning_rate": 8.4856947733807e-06,
"loss": 0.5055,
"step": 595
},
{
"epoch": 0.3295852534562212,
"grad_norm": 0.4555196762084961,
"learning_rate": 8.478766580235044e-06,
"loss": 0.4847,
"step": 596
},
{
"epoch": 0.33013824884792625,
"grad_norm": 0.49041327834129333,
"learning_rate": 8.471825416796185e-06,
"loss": 0.4662,
"step": 597
},
{
"epoch": 0.3306912442396313,
"grad_norm": 0.466457724571228,
"learning_rate": 8.464871308943673e-06,
"loss": 0.4794,
"step": 598
},
{
"epoch": 0.3312442396313364,
"grad_norm": 0.5273913145065308,
"learning_rate": 8.457904282605324e-06,
"loss": 0.482,
"step": 599
},
{
"epoch": 0.3317972350230415,
"grad_norm": 0.42694202065467834,
"learning_rate": 8.45092436375712e-06,
"loss": 0.4663,
"step": 600
},
{
"epoch": 0.33235023041474654,
"grad_norm": 0.467692494392395,
"learning_rate": 8.443931578423108e-06,
"loss": 0.4606,
"step": 601
},
{
"epoch": 0.3329032258064516,
"grad_norm": 0.5536178946495056,
"learning_rate": 8.436925952675312e-06,
"loss": 0.49,
"step": 602
},
{
"epoch": 0.33345622119815665,
"grad_norm": 0.5051153898239136,
"learning_rate": 8.429907512633626e-06,
"loss": 0.5041,
"step": 603
},
{
"epoch": 0.33400921658986177,
"grad_norm": 0.4975510835647583,
"learning_rate": 8.422876284465722e-06,
"loss": 0.491,
"step": 604
},
{
"epoch": 0.3345622119815668,
"grad_norm": 0.5645290613174438,
"learning_rate": 8.415832294386956e-06,
"loss": 0.4942,
"step": 605
},
{
"epoch": 0.3351152073732719,
"grad_norm": 0.5082492828369141,
"learning_rate": 8.408775568660256e-06,
"loss": 0.4829,
"step": 606
},
{
"epoch": 0.33566820276497694,
"grad_norm": 0.4222676753997803,
"learning_rate": 8.401706133596047e-06,
"loss": 0.4837,
"step": 607
},
{
"epoch": 0.33622119815668206,
"grad_norm": 0.5251336693763733,
"learning_rate": 8.394624015552124e-06,
"loss": 0.492,
"step": 608
},
{
"epoch": 0.3367741935483871,
"grad_norm": 0.4806731641292572,
"learning_rate": 8.387529240933585e-06,
"loss": 0.4791,
"step": 609
},
{
"epoch": 0.33732718894009217,
"grad_norm": 0.5003568530082703,
"learning_rate": 8.380421836192705e-06,
"loss": 0.4768,
"step": 610
},
{
"epoch": 0.33788018433179723,
"grad_norm": 0.48693132400512695,
"learning_rate": 8.373301827828858e-06,
"loss": 0.5219,
"step": 611
},
{
"epoch": 0.3384331797235023,
"grad_norm": 0.43921712040901184,
"learning_rate": 8.366169242388402e-06,
"loss": 0.4668,
"step": 612
},
{
"epoch": 0.3389861751152074,
"grad_norm": 0.5484280586242676,
"learning_rate": 8.359024106464594e-06,
"loss": 0.4669,
"step": 613
},
{
"epoch": 0.33953917050691246,
"grad_norm": 0.4520418643951416,
"learning_rate": 8.351866446697483e-06,
"loss": 0.4973,
"step": 614
},
{
"epoch": 0.3400921658986175,
"grad_norm": 0.588077962398529,
"learning_rate": 8.344696289773806e-06,
"loss": 0.4749,
"step": 615
},
{
"epoch": 0.3406451612903226,
"grad_norm": 0.4226244390010834,
"learning_rate": 8.337513662426904e-06,
"loss": 0.482,
"step": 616
},
{
"epoch": 0.34119815668202763,
"grad_norm": 0.5818718671798706,
"learning_rate": 8.330318591436606e-06,
"loss": 0.4963,
"step": 617
},
{
"epoch": 0.34175115207373274,
"grad_norm": 0.4959838092327118,
"learning_rate": 8.32311110362914e-06,
"loss": 0.4928,
"step": 618
},
{
"epoch": 0.3423041474654378,
"grad_norm": 0.43233931064605713,
"learning_rate": 8.315891225877026e-06,
"loss": 0.4619,
"step": 619
},
{
"epoch": 0.34285714285714286,
"grad_norm": 0.506004273891449,
"learning_rate": 8.308658985098983e-06,
"loss": 0.4638,
"step": 620
},
{
"epoch": 0.3434101382488479,
"grad_norm": 0.5699809193611145,
"learning_rate": 8.301414408259821e-06,
"loss": 0.4666,
"step": 621
},
{
"epoch": 0.343963133640553,
"grad_norm": 0.4087035059928894,
"learning_rate": 8.294157522370343e-06,
"loss": 0.4792,
"step": 622
},
{
"epoch": 0.3445161290322581,
"grad_norm": 0.482768714427948,
"learning_rate": 8.28688835448725e-06,
"loss": 0.4943,
"step": 623
},
{
"epoch": 0.34506912442396315,
"grad_norm": 0.4292857348918915,
"learning_rate": 8.27960693171303e-06,
"loss": 0.4809,
"step": 624
},
{
"epoch": 0.3456221198156682,
"grad_norm": 0.43942004442214966,
"learning_rate": 8.272313281195868e-06,
"loss": 0.488,
"step": 625
},
{
"epoch": 0.34617511520737326,
"grad_norm": 0.5329846739768982,
"learning_rate": 8.265007430129535e-06,
"loss": 0.494,
"step": 626
},
{
"epoch": 0.3467281105990783,
"grad_norm": 0.41792014241218567,
"learning_rate": 8.25768940575329e-06,
"loss": 0.4838,
"step": 627
},
{
"epoch": 0.34728110599078343,
"grad_norm": 0.522320568561554,
"learning_rate": 8.250359235351783e-06,
"loss": 0.4778,
"step": 628
},
{
"epoch": 0.3478341013824885,
"grad_norm": 0.4983687996864319,
"learning_rate": 8.243016946254947e-06,
"loss": 0.4761,
"step": 629
},
{
"epoch": 0.34838709677419355,
"grad_norm": 0.5216838717460632,
"learning_rate": 8.235662565837901e-06,
"loss": 0.4949,
"step": 630
},
{
"epoch": 0.3489400921658986,
"grad_norm": 0.5560440421104431,
"learning_rate": 8.22829612152084e-06,
"loss": 0.4653,
"step": 631
},
{
"epoch": 0.34949308755760367,
"grad_norm": 0.505742073059082,
"learning_rate": 8.220917640768943e-06,
"loss": 0.4981,
"step": 632
},
{
"epoch": 0.3500460829493088,
"grad_norm": 0.4594072103500366,
"learning_rate": 8.213527151092267e-06,
"loss": 0.4522,
"step": 633
},
{
"epoch": 0.35059907834101384,
"grad_norm": 0.40718746185302734,
"learning_rate": 8.20612468004564e-06,
"loss": 0.4728,
"step": 634
},
{
"epoch": 0.3511520737327189,
"grad_norm": 0.500005304813385,
"learning_rate": 8.19871025522856e-06,
"loss": 0.4667,
"step": 635
},
{
"epoch": 0.35170506912442395,
"grad_norm": 0.4374920427799225,
"learning_rate": 8.191283904285098e-06,
"loss": 0.4801,
"step": 636
},
{
"epoch": 0.352258064516129,
"grad_norm": 0.4499951899051666,
"learning_rate": 8.183845654903789e-06,
"loss": 0.4588,
"step": 637
},
{
"epoch": 0.3528110599078341,
"grad_norm": 0.41557565331459045,
"learning_rate": 8.17639553481753e-06,
"loss": 0.4953,
"step": 638
},
{
"epoch": 0.3533640552995392,
"grad_norm": 0.44742459058761597,
"learning_rate": 8.168933571803475e-06,
"loss": 0.4711,
"step": 639
},
{
"epoch": 0.35391705069124424,
"grad_norm": 0.4027085304260254,
"learning_rate": 8.161459793682937e-06,
"loss": 0.4872,
"step": 640
},
{
"epoch": 0.3544700460829493,
"grad_norm": 0.4093347191810608,
"learning_rate": 8.153974228321279e-06,
"loss": 0.477,
"step": 641
},
{
"epoch": 0.35502304147465436,
"grad_norm": 0.45894891023635864,
"learning_rate": 8.146476903627813e-06,
"loss": 0.4844,
"step": 642
},
{
"epoch": 0.35557603686635947,
"grad_norm": 0.44424736499786377,
"learning_rate": 8.138967847555693e-06,
"loss": 0.4754,
"step": 643
},
{
"epoch": 0.3561290322580645,
"grad_norm": 0.4633028209209442,
"learning_rate": 8.13144708810181e-06,
"loss": 0.4873,
"step": 644
},
{
"epoch": 0.3566820276497696,
"grad_norm": 0.46212074160575867,
"learning_rate": 8.123914653306695e-06,
"loss": 0.4576,
"step": 645
},
{
"epoch": 0.35723502304147464,
"grad_norm": 0.5635020732879639,
"learning_rate": 8.116370571254407e-06,
"loss": 0.4651,
"step": 646
},
{
"epoch": 0.3577880184331797,
"grad_norm": 0.43892374634742737,
"learning_rate": 8.108814870072434e-06,
"loss": 0.5018,
"step": 647
},
{
"epoch": 0.3583410138248848,
"grad_norm": 0.4993932843208313,
"learning_rate": 8.10124757793158e-06,
"loss": 0.4848,
"step": 648
},
{
"epoch": 0.35889400921658987,
"grad_norm": 0.5206483006477356,
"learning_rate": 8.093668723045866e-06,
"loss": 0.4845,
"step": 649
},
{
"epoch": 0.35944700460829493,
"grad_norm": 0.4889640808105469,
"learning_rate": 8.08607833367243e-06,
"loss": 0.4731,
"step": 650
},
{
"epoch": 0.36,
"grad_norm": 0.5545540452003479,
"learning_rate": 8.078476438111404e-06,
"loss": 0.4919,
"step": 651
},
{
"epoch": 0.36055299539170504,
"grad_norm": 0.47018927335739136,
"learning_rate": 8.07086306470583e-06,
"loss": 0.4902,
"step": 652
},
{
"epoch": 0.36110599078341016,
"grad_norm": 0.5213498473167419,
"learning_rate": 8.06323824184154e-06,
"loss": 0.4918,
"step": 653
},
{
"epoch": 0.3616589861751152,
"grad_norm": 0.4931851923465729,
"learning_rate": 8.055601997947056e-06,
"loss": 0.4868,
"step": 654
},
{
"epoch": 0.3622119815668203,
"grad_norm": 0.4996345341205597,
"learning_rate": 8.04795436149348e-06,
"loss": 0.4908,
"step": 655
},
{
"epoch": 0.36276497695852533,
"grad_norm": 0.48268017172813416,
"learning_rate": 8.04029536099439e-06,
"loss": 0.4975,
"step": 656
},
{
"epoch": 0.3633179723502304,
"grad_norm": 0.5365729928016663,
"learning_rate": 8.032625025005736e-06,
"loss": 0.4891,
"step": 657
},
{
"epoch": 0.3638709677419355,
"grad_norm": 0.5524755716323853,
"learning_rate": 8.024943382125732e-06,
"loss": 0.4833,
"step": 658
},
{
"epoch": 0.36442396313364056,
"grad_norm": 0.4499988257884979,
"learning_rate": 8.017250460994748e-06,
"loss": 0.4724,
"step": 659
},
{
"epoch": 0.3649769585253456,
"grad_norm": 0.44743797183036804,
"learning_rate": 8.0095462902952e-06,
"loss": 0.503,
"step": 660
},
{
"epoch": 0.3655299539170507,
"grad_norm": 0.502610445022583,
"learning_rate": 8.001830898751455e-06,
"loss": 0.4813,
"step": 661
},
{
"epoch": 0.36608294930875573,
"grad_norm": 0.41622281074523926,
"learning_rate": 7.99410431512971e-06,
"loss": 0.5157,
"step": 662
},
{
"epoch": 0.36663594470046085,
"grad_norm": 0.4399860203266144,
"learning_rate": 7.986366568237893e-06,
"loss": 0.4741,
"step": 663
},
{
"epoch": 0.3671889400921659,
"grad_norm": 0.43299001455307007,
"learning_rate": 7.97861768692555e-06,
"loss": 0.4727,
"step": 664
},
{
"epoch": 0.36774193548387096,
"grad_norm": 0.44035786390304565,
"learning_rate": 7.970857700083747e-06,
"loss": 0.4662,
"step": 665
},
{
"epoch": 0.368294930875576,
"grad_norm": 0.422536164522171,
"learning_rate": 7.96308663664495e-06,
"loss": 0.4969,
"step": 666
},
{
"epoch": 0.3688479262672811,
"grad_norm": 0.44590070843696594,
"learning_rate": 7.955304525582928e-06,
"loss": 0.4591,
"step": 667
},
{
"epoch": 0.3694009216589862,
"grad_norm": 0.43369054794311523,
"learning_rate": 7.947511395912635e-06,
"loss": 0.4941,
"step": 668
},
{
"epoch": 0.36995391705069125,
"grad_norm": 0.40852341055870056,
"learning_rate": 7.939707276690111e-06,
"loss": 0.4859,
"step": 669
},
{
"epoch": 0.3705069124423963,
"grad_norm": 0.3855592906475067,
"learning_rate": 7.93189219701237e-06,
"loss": 0.4834,
"step": 670
},
{
"epoch": 0.37105990783410137,
"grad_norm": 0.40031880140304565,
"learning_rate": 7.924066186017288e-06,
"loss": 0.5146,
"step": 671
},
{
"epoch": 0.3716129032258065,
"grad_norm": 0.42813777923583984,
"learning_rate": 7.916229272883499e-06,
"loss": 0.4875,
"step": 672
},
{
"epoch": 0.37216589861751154,
"grad_norm": 0.4341396689414978,
"learning_rate": 7.908381486830286e-06,
"loss": 0.4787,
"step": 673
},
{
"epoch": 0.3727188940092166,
"grad_norm": 0.40330180525779724,
"learning_rate": 7.90052285711747e-06,
"loss": 0.4866,
"step": 674
},
{
"epoch": 0.37327188940092165,
"grad_norm": 0.4432542026042938,
"learning_rate": 7.892653413045298e-06,
"loss": 0.4574,
"step": 675
},
{
"epoch": 0.3738248847926267,
"grad_norm": 0.4223710298538208,
"learning_rate": 7.884773183954344e-06,
"loss": 0.4619,
"step": 676
},
{
"epoch": 0.3743778801843318,
"grad_norm": 0.39498183131217957,
"learning_rate": 7.876882199225388e-06,
"loss": 0.4717,
"step": 677
},
{
"epoch": 0.3749308755760369,
"grad_norm": 0.4838169813156128,
"learning_rate": 7.868980488279316e-06,
"loss": 0.4756,
"step": 678
},
{
"epoch": 0.37548387096774194,
"grad_norm": 0.3836471140384674,
"learning_rate": 7.861068080576998e-06,
"loss": 0.462,
"step": 679
},
{
"epoch": 0.376036866359447,
"grad_norm": 0.503296971321106,
"learning_rate": 7.853145005619199e-06,
"loss": 0.4963,
"step": 680
},
{
"epoch": 0.37658986175115206,
"grad_norm": 0.3729606568813324,
"learning_rate": 7.84521129294644e-06,
"loss": 0.4684,
"step": 681
},
{
"epoch": 0.37714285714285717,
"grad_norm": 0.5140215158462524,
"learning_rate": 7.837266972138918e-06,
"loss": 0.5041,
"step": 682
},
{
"epoch": 0.3776958525345622,
"grad_norm": 0.4366711676120758,
"learning_rate": 7.82931207281637e-06,
"loss": 0.5007,
"step": 683
},
{
"epoch": 0.3782488479262673,
"grad_norm": 0.43705347180366516,
"learning_rate": 7.821346624637984e-06,
"loss": 0.4988,
"step": 684
},
{
"epoch": 0.37880184331797234,
"grad_norm": 0.4190782308578491,
"learning_rate": 7.813370657302273e-06,
"loss": 0.4815,
"step": 685
},
{
"epoch": 0.3793548387096774,
"grad_norm": 0.43545421957969666,
"learning_rate": 7.80538420054697e-06,
"loss": 0.486,
"step": 686
},
{
"epoch": 0.3799078341013825,
"grad_norm": 0.4012760818004608,
"learning_rate": 7.797387284148919e-06,
"loss": 0.456,
"step": 687
},
{
"epoch": 0.38046082949308757,
"grad_norm": 0.45634332299232483,
"learning_rate": 7.789379937923958e-06,
"loss": 0.4827,
"step": 688
},
{
"epoch": 0.38101382488479263,
"grad_norm": 0.45957717299461365,
"learning_rate": 7.781362191726818e-06,
"loss": 0.4976,
"step": 689
},
{
"epoch": 0.3815668202764977,
"grad_norm": 0.4902954399585724,
"learning_rate": 7.773334075451e-06,
"loss": 0.4952,
"step": 690
},
{
"epoch": 0.38211981566820274,
"grad_norm": 0.4607703387737274,
"learning_rate": 7.765295619028672e-06,
"loss": 0.4995,
"step": 691
},
{
"epoch": 0.38267281105990786,
"grad_norm": 0.49557220935821533,
"learning_rate": 7.757246852430553e-06,
"loss": 0.4921,
"step": 692
},
{
"epoch": 0.3832258064516129,
"grad_norm": 0.4447820782661438,
"learning_rate": 7.749187805665801e-06,
"loss": 0.4589,
"step": 693
},
{
"epoch": 0.383778801843318,
"grad_norm": 0.4028923809528351,
"learning_rate": 7.741118508781906e-06,
"loss": 0.4732,
"step": 694
},
{
"epoch": 0.38433179723502303,
"grad_norm": 0.4806511700153351,
"learning_rate": 7.733038991864576e-06,
"loss": 0.4968,
"step": 695
},
{
"epoch": 0.3848847926267281,
"grad_norm": 0.4322826862335205,
"learning_rate": 7.724949285037619e-06,
"loss": 0.4904,
"step": 696
},
{
"epoch": 0.3854377880184332,
"grad_norm": 0.44291332364082336,
"learning_rate": 7.716849418462836e-06,
"loss": 0.4829,
"step": 697
},
{
"epoch": 0.38599078341013826,
"grad_norm": 0.4092819094657898,
"learning_rate": 7.70873942233991e-06,
"loss": 0.4809,
"step": 698
},
{
"epoch": 0.3865437788018433,
"grad_norm": 0.4542578160762787,
"learning_rate": 7.700619326906288e-06,
"loss": 0.4713,
"step": 699
},
{
"epoch": 0.3870967741935484,
"grad_norm": 0.4618943929672241,
"learning_rate": 7.69248916243708e-06,
"loss": 0.4443,
"step": 700
},
{
"epoch": 0.38764976958525343,
"grad_norm": 0.43175920844078064,
"learning_rate": 7.684348959244924e-06,
"loss": 0.4761,
"step": 701
},
{
"epoch": 0.38820276497695855,
"grad_norm": 0.4224050045013428,
"learning_rate": 7.676198747679898e-06,
"loss": 0.4943,
"step": 702
},
{
"epoch": 0.3887557603686636,
"grad_norm": 0.4592989683151245,
"learning_rate": 7.66803855812939e-06,
"loss": 0.4529,
"step": 703
},
{
"epoch": 0.38930875576036866,
"grad_norm": 0.4457801878452301,
"learning_rate": 7.659868421017989e-06,
"loss": 0.4785,
"step": 704
},
{
"epoch": 0.3898617511520737,
"grad_norm": 0.4617078900337219,
"learning_rate": 7.651688366807378e-06,
"loss": 0.4755,
"step": 705
},
{
"epoch": 0.3904147465437788,
"grad_norm": 0.43977999687194824,
"learning_rate": 7.64349842599621e-06,
"loss": 0.5098,
"step": 706
},
{
"epoch": 0.3909677419354839,
"grad_norm": 0.4340912699699402,
"learning_rate": 7.635298629120002e-06,
"loss": 0.4452,
"step": 707
},
{
"epoch": 0.39152073732718895,
"grad_norm": 0.5204182863235474,
"learning_rate": 7.6270890067510196e-06,
"loss": 0.4677,
"step": 708
},
{
"epoch": 0.392073732718894,
"grad_norm": 0.47488752007484436,
"learning_rate": 7.618869589498157e-06,
"loss": 0.465,
"step": 709
},
{
"epoch": 0.39262672811059907,
"grad_norm": 0.4327450692653656,
"learning_rate": 7.610640408006832e-06,
"loss": 0.4945,
"step": 710
},
{
"epoch": 0.3931797235023041,
"grad_norm": 0.5072281360626221,
"learning_rate": 7.602401492958868e-06,
"loss": 0.4987,
"step": 711
},
{
"epoch": 0.39373271889400924,
"grad_norm": 0.4239625632762909,
"learning_rate": 7.594152875072376e-06,
"loss": 0.4715,
"step": 712
},
{
"epoch": 0.3942857142857143,
"grad_norm": 0.4698258936405182,
"learning_rate": 7.5858945851016455e-06,
"loss": 0.4779,
"step": 713
},
{
"epoch": 0.39483870967741935,
"grad_norm": 0.42160850763320923,
"learning_rate": 7.577626653837028e-06,
"loss": 0.45,
"step": 714
},
{
"epoch": 0.3953917050691244,
"grad_norm": 0.4377395808696747,
"learning_rate": 7.5693491121048194e-06,
"loss": 0.5144,
"step": 715
},
{
"epoch": 0.39594470046082947,
"grad_norm": 0.4012657403945923,
"learning_rate": 7.561061990767149e-06,
"loss": 0.4812,
"step": 716
},
{
"epoch": 0.3964976958525346,
"grad_norm": 0.41947153210639954,
"learning_rate": 7.552765320721865e-06,
"loss": 0.4407,
"step": 717
},
{
"epoch": 0.39705069124423964,
"grad_norm": 0.48914283514022827,
"learning_rate": 7.544459132902409e-06,
"loss": 0.4689,
"step": 718
},
{
"epoch": 0.3976036866359447,
"grad_norm": 0.3954436480998993,
"learning_rate": 7.5361434582777205e-06,
"loss": 0.4866,
"step": 719
},
{
"epoch": 0.39815668202764976,
"grad_norm": 0.45554065704345703,
"learning_rate": 7.527818327852101e-06,
"loss": 0.4499,
"step": 720
},
{
"epoch": 0.3987096774193548,
"grad_norm": 0.3696426451206207,
"learning_rate": 7.519483772665113e-06,
"loss": 0.4963,
"step": 721
},
{
"epoch": 0.3992626728110599,
"grad_norm": 0.42578935623168945,
"learning_rate": 7.511139823791452e-06,
"loss": 0.4642,
"step": 722
},
{
"epoch": 0.399815668202765,
"grad_norm": 0.3758449852466583,
"learning_rate": 7.502786512340841e-06,
"loss": 0.516,
"step": 723
},
{
"epoch": 0.40036866359447004,
"grad_norm": 0.4170108139514923,
"learning_rate": 7.494423869457912e-06,
"loss": 0.4824,
"step": 724
},
{
"epoch": 0.4009216589861751,
"grad_norm": 0.3795914053916931,
"learning_rate": 7.4860519263220865e-06,
"loss": 0.4744,
"step": 725
},
{
"epoch": 0.40147465437788016,
"grad_norm": 0.389635294675827,
"learning_rate": 7.477670714147461e-06,
"loss": 0.4957,
"step": 726
},
{
"epoch": 0.40202764976958527,
"grad_norm": 0.44072335958480835,
"learning_rate": 7.469280264182689e-06,
"loss": 0.5172,
"step": 727
},
{
"epoch": 0.40258064516129033,
"grad_norm": 0.372860312461853,
"learning_rate": 7.460880607710872e-06,
"loss": 0.4465,
"step": 728
},
{
"epoch": 0.4031336405529954,
"grad_norm": 0.47300985455513,
"learning_rate": 7.452471776049432e-06,
"loss": 0.4636,
"step": 729
},
{
"epoch": 0.40368663594470044,
"grad_norm": 0.44356971979141235,
"learning_rate": 7.444053800550004e-06,
"loss": 0.4911,
"step": 730
},
{
"epoch": 0.4042396313364055,
"grad_norm": 0.4640626311302185,
"learning_rate": 7.435626712598312e-06,
"loss": 0.4873,
"step": 731
},
{
"epoch": 0.4047926267281106,
"grad_norm": 0.4652840197086334,
"learning_rate": 7.427190543614053e-06,
"loss": 0.4641,
"step": 732
},
{
"epoch": 0.4053456221198157,
"grad_norm": 0.46554264426231384,
"learning_rate": 7.418745325050787e-06,
"loss": 0.4797,
"step": 733
},
{
"epoch": 0.40589861751152073,
"grad_norm": 0.42912858724594116,
"learning_rate": 7.410291088395812e-06,
"loss": 0.5064,
"step": 734
},
{
"epoch": 0.4064516129032258,
"grad_norm": 0.41365447640419006,
"learning_rate": 7.401827865170047e-06,
"loss": 0.4873,
"step": 735
},
{
"epoch": 0.4070046082949309,
"grad_norm": 0.5490487813949585,
"learning_rate": 7.3933556869279235e-06,
"loss": 0.4583,
"step": 736
},
{
"epoch": 0.40755760368663596,
"grad_norm": 0.39917993545532227,
"learning_rate": 7.384874585257255e-06,
"loss": 0.4625,
"step": 737
},
{
"epoch": 0.408110599078341,
"grad_norm": 0.4784587621688843,
"learning_rate": 7.3763845917791245e-06,
"loss": 0.5073,
"step": 738
},
{
"epoch": 0.4086635944700461,
"grad_norm": 0.4989032447338104,
"learning_rate": 7.367885738147773e-06,
"loss": 0.4652,
"step": 739
},
{
"epoch": 0.40921658986175113,
"grad_norm": 0.4629876911640167,
"learning_rate": 7.359378056050472e-06,
"loss": 0.4601,
"step": 740
},
{
"epoch": 0.40976958525345625,
"grad_norm": 0.4878072440624237,
"learning_rate": 7.35086157720741e-06,
"loss": 0.4713,
"step": 741
},
{
"epoch": 0.4103225806451613,
"grad_norm": 0.49611517786979675,
"learning_rate": 7.3423363333715726e-06,
"loss": 0.4727,
"step": 742
},
{
"epoch": 0.41087557603686636,
"grad_norm": 0.5412954688072205,
"learning_rate": 7.333802356328624e-06,
"loss": 0.4795,
"step": 743
},
{
"epoch": 0.4114285714285714,
"grad_norm": 0.387992262840271,
"learning_rate": 7.3252596778967954e-06,
"loss": 0.492,
"step": 744
},
{
"epoch": 0.4119815668202765,
"grad_norm": 0.47084909677505493,
"learning_rate": 7.316708329926754e-06,
"loss": 0.4546,
"step": 745
},
{
"epoch": 0.4125345622119816,
"grad_norm": 0.5451803803443909,
"learning_rate": 7.308148344301491e-06,
"loss": 0.5053,
"step": 746
},
{
"epoch": 0.41308755760368665,
"grad_norm": 0.4724251329898834,
"learning_rate": 7.2995797529362075e-06,
"loss": 0.4937,
"step": 747
},
{
"epoch": 0.4136405529953917,
"grad_norm": 0.48225685954093933,
"learning_rate": 7.291002587778184e-06,
"loss": 0.4939,
"step": 748
},
{
"epoch": 0.41419354838709677,
"grad_norm": 0.4945637881755829,
"learning_rate": 7.282416880806673e-06,
"loss": 0.4934,
"step": 749
},
{
"epoch": 0.4147465437788018,
"grad_norm": 0.48181742429733276,
"learning_rate": 7.273822664032771e-06,
"loss": 0.4763,
"step": 750
},
{
"epoch": 0.41529953917050694,
"grad_norm": 0.41117456555366516,
"learning_rate": 7.265219969499302e-06,
"loss": 0.4928,
"step": 751
},
{
"epoch": 0.415852534562212,
"grad_norm": 0.5093626976013184,
"learning_rate": 7.256608829280705e-06,
"loss": 0.4563,
"step": 752
},
{
"epoch": 0.41640552995391705,
"grad_norm": 0.45837345719337463,
"learning_rate": 7.2479892754828995e-06,
"loss": 0.4736,
"step": 753
},
{
"epoch": 0.4169585253456221,
"grad_norm": 0.4117204546928406,
"learning_rate": 7.239361340243179e-06,
"loss": 0.4788,
"step": 754
},
{
"epoch": 0.41751152073732717,
"grad_norm": 0.44084909558296204,
"learning_rate": 7.230725055730088e-06,
"loss": 0.4781,
"step": 755
},
{
"epoch": 0.4180645161290323,
"grad_norm": 0.5877032279968262,
"learning_rate": 7.222080454143296e-06,
"loss": 0.4782,
"step": 756
},
{
"epoch": 0.41861751152073734,
"grad_norm": 0.3853990137577057,
"learning_rate": 7.213427567713485e-06,
"loss": 0.4498,
"step": 757
},
{
"epoch": 0.4191705069124424,
"grad_norm": 0.4723830819129944,
"learning_rate": 7.204766428702226e-06,
"loss": 0.4601,
"step": 758
},
{
"epoch": 0.41972350230414746,
"grad_norm": 0.43980076909065247,
"learning_rate": 7.196097069401857e-06,
"loss": 0.4783,
"step": 759
},
{
"epoch": 0.4202764976958525,
"grad_norm": 0.4152413606643677,
"learning_rate": 7.1874195221353706e-06,
"loss": 0.4873,
"step": 760
},
{
"epoch": 0.4208294930875576,
"grad_norm": 0.3723086714744568,
"learning_rate": 7.17873381925628e-06,
"loss": 0.4664,
"step": 761
},
{
"epoch": 0.4213824884792627,
"grad_norm": 0.44318974018096924,
"learning_rate": 7.17003999314851e-06,
"loss": 0.4865,
"step": 762
},
{
"epoch": 0.42193548387096774,
"grad_norm": 0.40795907378196716,
"learning_rate": 7.161338076226272e-06,
"loss": 0.4795,
"step": 763
},
{
"epoch": 0.4224884792626728,
"grad_norm": 0.4073677659034729,
"learning_rate": 7.1526281009339426e-06,
"loss": 0.4815,
"step": 764
},
{
"epoch": 0.42304147465437786,
"grad_norm": 0.4583486020565033,
"learning_rate": 7.143910099745944e-06,
"loss": 0.4872,
"step": 765
},
{
"epoch": 0.42359447004608297,
"grad_norm": 0.4036937654018402,
"learning_rate": 7.13518410516662e-06,
"loss": 0.4616,
"step": 766
},
{
"epoch": 0.42414746543778803,
"grad_norm": 0.42568910121917725,
"learning_rate": 7.126450149730122e-06,
"loss": 0.489,
"step": 767
},
{
"epoch": 0.4247004608294931,
"grad_norm": 0.42301973700523376,
"learning_rate": 7.1177082660002784e-06,
"loss": 0.4685,
"step": 768
},
{
"epoch": 0.42525345622119815,
"grad_norm": 0.45005002617836,
"learning_rate": 7.108958486570479e-06,
"loss": 0.4555,
"step": 769
},
{
"epoch": 0.4258064516129032,
"grad_norm": 0.45795372128486633,
"learning_rate": 7.1002008440635515e-06,
"loss": 0.4678,
"step": 770
},
{
"epoch": 0.4263594470046083,
"grad_norm": 0.41580426692962646,
"learning_rate": 7.091435371131642e-06,
"loss": 0.5058,
"step": 771
},
{
"epoch": 0.4269124423963134,
"grad_norm": 0.4396505355834961,
"learning_rate": 7.082662100456089e-06,
"loss": 0.492,
"step": 772
},
{
"epoch": 0.42746543778801843,
"grad_norm": 0.4275103211402893,
"learning_rate": 7.073881064747308e-06,
"loss": 0.4624,
"step": 773
},
{
"epoch": 0.4280184331797235,
"grad_norm": 0.37376993894577026,
"learning_rate": 7.0650922967446634e-06,
"loss": 0.4626,
"step": 774
},
{
"epoch": 0.42857142857142855,
"grad_norm": 0.40532317757606506,
"learning_rate": 7.056295829216347e-06,
"loss": 0.5042,
"step": 775
},
{
"epoch": 0.42912442396313366,
"grad_norm": 0.39862802624702454,
"learning_rate": 7.047491694959263e-06,
"loss": 0.4398,
"step": 776
},
{
"epoch": 0.4296774193548387,
"grad_norm": 0.4326246380805969,
"learning_rate": 7.038679926798895e-06,
"loss": 0.4902,
"step": 777
},
{
"epoch": 0.4302304147465438,
"grad_norm": 0.43664461374282837,
"learning_rate": 7.029860557589191e-06,
"loss": 0.4817,
"step": 778
},
{
"epoch": 0.43078341013824883,
"grad_norm": 0.4556635320186615,
"learning_rate": 7.02103362021244e-06,
"loss": 0.4825,
"step": 779
},
{
"epoch": 0.4313364055299539,
"grad_norm": 0.40778669714927673,
"learning_rate": 7.012199147579146e-06,
"loss": 0.496,
"step": 780
},
{
"epoch": 0.431889400921659,
"grad_norm": 0.41946834325790405,
"learning_rate": 7.00335717262791e-06,
"loss": 0.482,
"step": 781
},
{
"epoch": 0.43244239631336406,
"grad_norm": 0.4511401653289795,
"learning_rate": 6.994507728325304e-06,
"loss": 0.4763,
"step": 782
},
{
"epoch": 0.4329953917050691,
"grad_norm": 0.43290090560913086,
"learning_rate": 6.985650847665747e-06,
"loss": 0.4804,
"step": 783
},
{
"epoch": 0.4335483870967742,
"grad_norm": 0.4241175949573517,
"learning_rate": 6.976786563671386e-06,
"loss": 0.4776,
"step": 784
},
{
"epoch": 0.43410138248847924,
"grad_norm": 0.4226067364215851,
"learning_rate": 6.967914909391971e-06,
"loss": 0.4839,
"step": 785
},
{
"epoch": 0.43465437788018435,
"grad_norm": 0.39866408705711365,
"learning_rate": 6.959035917904728e-06,
"loss": 0.4596,
"step": 786
},
{
"epoch": 0.4352073732718894,
"grad_norm": 0.37730705738067627,
"learning_rate": 6.950149622314245e-06,
"loss": 0.4784,
"step": 787
},
{
"epoch": 0.43576036866359447,
"grad_norm": 0.41297316551208496,
"learning_rate": 6.941256055752337e-06,
"loss": 0.4887,
"step": 788
},
{
"epoch": 0.4363133640552995,
"grad_norm": 0.4532250761985779,
"learning_rate": 6.932355251377932e-06,
"loss": 0.4897,
"step": 789
},
{
"epoch": 0.4368663594470046,
"grad_norm": 0.4239371418952942,
"learning_rate": 6.923447242376942e-06,
"loss": 0.4873,
"step": 790
},
{
"epoch": 0.4374193548387097,
"grad_norm": 0.33959221839904785,
"learning_rate": 6.91453206196214e-06,
"loss": 0.4697,
"step": 791
},
{
"epoch": 0.43797235023041475,
"grad_norm": 0.4459823668003082,
"learning_rate": 6.90560974337304e-06,
"loss": 0.4799,
"step": 792
},
{
"epoch": 0.4385253456221198,
"grad_norm": 0.4143071174621582,
"learning_rate": 6.896680319875767e-06,
"loss": 0.4556,
"step": 793
},
{
"epoch": 0.43907834101382487,
"grad_norm": 0.4591521620750427,
"learning_rate": 6.887743824762937e-06,
"loss": 0.4841,
"step": 794
},
{
"epoch": 0.4396313364055299,
"grad_norm": 0.41245412826538086,
"learning_rate": 6.878800291353533e-06,
"loss": 0.5083,
"step": 795
},
{
"epoch": 0.44018433179723504,
"grad_norm": 0.4207186996936798,
"learning_rate": 6.8698497529927784e-06,
"loss": 0.4653,
"step": 796
},
{
"epoch": 0.4407373271889401,
"grad_norm": 0.42010927200317383,
"learning_rate": 6.860892243052015e-06,
"loss": 0.4595,
"step": 797
},
{
"epoch": 0.44129032258064516,
"grad_norm": 0.41488680243492126,
"learning_rate": 6.851927794928578e-06,
"loss": 0.4685,
"step": 798
},
{
"epoch": 0.4418433179723502,
"grad_norm": 0.43417924642562866,
"learning_rate": 6.842956442045668e-06,
"loss": 0.4753,
"step": 799
},
{
"epoch": 0.4423963133640553,
"grad_norm": 0.35748791694641113,
"learning_rate": 6.833978217852233e-06,
"loss": 0.4677,
"step": 800
},
{
"epoch": 0.4429493087557604,
"grad_norm": 0.4993172883987427,
"learning_rate": 6.824993155822838e-06,
"loss": 0.4729,
"step": 801
},
{
"epoch": 0.44350230414746544,
"grad_norm": 0.36592617630958557,
"learning_rate": 6.816001289457542e-06,
"loss": 0.4565,
"step": 802
},
{
"epoch": 0.4440552995391705,
"grad_norm": 0.3910996615886688,
"learning_rate": 6.807002652281777e-06,
"loss": 0.4619,
"step": 803
},
{
"epoch": 0.44460829493087556,
"grad_norm": 0.41243669390678406,
"learning_rate": 6.797997277846215e-06,
"loss": 0.46,
"step": 804
},
{
"epoch": 0.44516129032258067,
"grad_norm": 0.3570329546928406,
"learning_rate": 6.788985199726648e-06,
"loss": 0.4624,
"step": 805
},
{
"epoch": 0.44571428571428573,
"grad_norm": 0.4660005271434784,
"learning_rate": 6.779966451523866e-06,
"loss": 0.494,
"step": 806
},
{
"epoch": 0.4462672811059908,
"grad_norm": 0.38170328736305237,
"learning_rate": 6.770941066863523e-06,
"loss": 0.4848,
"step": 807
},
{
"epoch": 0.44682027649769585,
"grad_norm": 0.410656601190567,
"learning_rate": 6.761909079396018e-06,
"loss": 0.4783,
"step": 808
},
{
"epoch": 0.4473732718894009,
"grad_norm": 0.39199960231781006,
"learning_rate": 6.752870522796372e-06,
"loss": 0.4706,
"step": 809
},
{
"epoch": 0.447926267281106,
"grad_norm": 0.4048680067062378,
"learning_rate": 6.743825430764091e-06,
"loss": 0.4841,
"step": 810
},
{
"epoch": 0.4484792626728111,
"grad_norm": 0.3699415922164917,
"learning_rate": 6.734773837023054e-06,
"loss": 0.4479,
"step": 811
},
{
"epoch": 0.44903225806451613,
"grad_norm": 0.3935936391353607,
"learning_rate": 6.725715775321379e-06,
"loss": 0.4802,
"step": 812
},
{
"epoch": 0.4495852534562212,
"grad_norm": 0.39852485060691833,
"learning_rate": 6.7166512794312986e-06,
"loss": 0.4652,
"step": 813
},
{
"epoch": 0.45013824884792625,
"grad_norm": 0.38947340846061707,
"learning_rate": 6.707580383149035e-06,
"loss": 0.4925,
"step": 814
},
{
"epoch": 0.45069124423963136,
"grad_norm": 0.4168023467063904,
"learning_rate": 6.698503120294674e-06,
"loss": 0.4829,
"step": 815
},
{
"epoch": 0.4512442396313364,
"grad_norm": 0.39369869232177734,
"learning_rate": 6.6894195247120396e-06,
"loss": 0.4562,
"step": 816
},
{
"epoch": 0.4517972350230415,
"grad_norm": 0.3529163599014282,
"learning_rate": 6.680329630268565e-06,
"loss": 0.4628,
"step": 817
},
{
"epoch": 0.45235023041474653,
"grad_norm": 0.3897452652454376,
"learning_rate": 6.671233470855171e-06,
"loss": 0.4635,
"step": 818
},
{
"epoch": 0.4529032258064516,
"grad_norm": 0.4104883074760437,
"learning_rate": 6.662131080386132e-06,
"loss": 0.4716,
"step": 819
},
{
"epoch": 0.4534562211981567,
"grad_norm": 0.39856767654418945,
"learning_rate": 6.653022492798959e-06,
"loss": 0.4666,
"step": 820
},
{
"epoch": 0.45400921658986176,
"grad_norm": 0.3940970301628113,
"learning_rate": 6.643907742054267e-06,
"loss": 0.4899,
"step": 821
},
{
"epoch": 0.4545622119815668,
"grad_norm": 0.4091474413871765,
"learning_rate": 6.63478686213565e-06,
"loss": 0.4829,
"step": 822
},
{
"epoch": 0.4551152073732719,
"grad_norm": 0.4372076690196991,
"learning_rate": 6.625659887049554e-06,
"loss": 0.4872,
"step": 823
},
{
"epoch": 0.45566820276497694,
"grad_norm": 0.4298078417778015,
"learning_rate": 6.6165268508251455e-06,
"loss": 0.4748,
"step": 824
},
{
"epoch": 0.45622119815668205,
"grad_norm": 0.4123404622077942,
"learning_rate": 6.607387787514199e-06,
"loss": 0.4784,
"step": 825
},
{
"epoch": 0.4567741935483871,
"grad_norm": 0.48749467730522156,
"learning_rate": 6.598242731190955e-06,
"loss": 0.4617,
"step": 826
},
{
"epoch": 0.45732718894009217,
"grad_norm": 0.44510942697525024,
"learning_rate": 6.589091715951996e-06,
"loss": 0.5012,
"step": 827
},
{
"epoch": 0.4578801843317972,
"grad_norm": 0.4342193603515625,
"learning_rate": 6.579934775916128e-06,
"loss": 0.4686,
"step": 828
},
{
"epoch": 0.4584331797235023,
"grad_norm": 0.46518048644065857,
"learning_rate": 6.570771945224243e-06,
"loss": 0.4695,
"step": 829
},
{
"epoch": 0.4589861751152074,
"grad_norm": 0.4383966326713562,
"learning_rate": 6.561603258039195e-06,
"loss": 0.4719,
"step": 830
},
{
"epoch": 0.45953917050691245,
"grad_norm": 0.41184401512145996,
"learning_rate": 6.552428748545677e-06,
"loss": 0.4789,
"step": 831
},
{
"epoch": 0.4600921658986175,
"grad_norm": 0.4380422830581665,
"learning_rate": 6.543248450950087e-06,
"loss": 0.4907,
"step": 832
},
{
"epoch": 0.46064516129032257,
"grad_norm": 0.453457236289978,
"learning_rate": 6.534062399480405e-06,
"loss": 0.4747,
"step": 833
},
{
"epoch": 0.4611981566820276,
"grad_norm": 0.36671724915504456,
"learning_rate": 6.524870628386064e-06,
"loss": 0.4905,
"step": 834
},
{
"epoch": 0.46175115207373274,
"grad_norm": 0.42992615699768066,
"learning_rate": 6.515673171937822e-06,
"loss": 0.4634,
"step": 835
},
{
"epoch": 0.4623041474654378,
"grad_norm": 0.40917208790779114,
"learning_rate": 6.5064700644276315e-06,
"loss": 0.4661,
"step": 836
},
{
"epoch": 0.46285714285714286,
"grad_norm": 0.4510115385055542,
"learning_rate": 6.497261340168519e-06,
"loss": 0.4504,
"step": 837
},
{
"epoch": 0.4634101382488479,
"grad_norm": 0.45596301555633545,
"learning_rate": 6.4880470334944515e-06,
"loss": 0.4551,
"step": 838
},
{
"epoch": 0.46396313364055297,
"grad_norm": 0.4473353624343872,
"learning_rate": 6.478827178760205e-06,
"loss": 0.4834,
"step": 839
},
{
"epoch": 0.4645161290322581,
"grad_norm": 0.4285813271999359,
"learning_rate": 6.469601810341247e-06,
"loss": 0.498,
"step": 840
},
{
"epoch": 0.46506912442396314,
"grad_norm": 0.38439249992370605,
"learning_rate": 6.4603709626336e-06,
"loss": 0.4754,
"step": 841
},
{
"epoch": 0.4656221198156682,
"grad_norm": 0.3741195499897003,
"learning_rate": 6.451134670053716e-06,
"loss": 0.4887,
"step": 842
},
{
"epoch": 0.46617511520737326,
"grad_norm": 0.4147772789001465,
"learning_rate": 6.441892967038346e-06,
"loss": 0.5023,
"step": 843
},
{
"epoch": 0.4667281105990783,
"grad_norm": 0.3958702087402344,
"learning_rate": 6.4326458880444155e-06,
"loss": 0.507,
"step": 844
},
{
"epoch": 0.46728110599078343,
"grad_norm": 0.4042072892189026,
"learning_rate": 6.423393467548893e-06,
"loss": 0.444,
"step": 845
},
{
"epoch": 0.4678341013824885,
"grad_norm": 0.387441486120224,
"learning_rate": 6.414135740048662e-06,
"loss": 0.4446,
"step": 846
},
{
"epoch": 0.46838709677419355,
"grad_norm": 0.4227030277252197,
"learning_rate": 6.404872740060392e-06,
"loss": 0.5104,
"step": 847
},
{
"epoch": 0.4689400921658986,
"grad_norm": 0.3976932764053345,
"learning_rate": 6.395604502120412e-06,
"loss": 0.4686,
"step": 848
},
{
"epoch": 0.46949308755760366,
"grad_norm": 0.41935819387435913,
"learning_rate": 6.386331060784583e-06,
"loss": 0.4802,
"step": 849
},
{
"epoch": 0.4700460829493088,
"grad_norm": 0.3910142183303833,
"learning_rate": 6.377052450628159e-06,
"loss": 0.465,
"step": 850
},
{
"epoch": 0.47059907834101383,
"grad_norm": 0.3708650469779968,
"learning_rate": 6.367768706245674e-06,
"loss": 0.4844,
"step": 851
},
{
"epoch": 0.4711520737327189,
"grad_norm": 0.39117151498794556,
"learning_rate": 6.358479862250796e-06,
"loss": 0.4826,
"step": 852
},
{
"epoch": 0.47170506912442395,
"grad_norm": 0.36844393610954285,
"learning_rate": 6.349185953276216e-06,
"loss": 0.4531,
"step": 853
},
{
"epoch": 0.472258064516129,
"grad_norm": 0.3579118549823761,
"learning_rate": 6.339887013973498e-06,
"loss": 0.4579,
"step": 854
},
{
"epoch": 0.4728110599078341,
"grad_norm": 0.4207668900489807,
"learning_rate": 6.330583079012972e-06,
"loss": 0.4769,
"step": 855
},
{
"epoch": 0.4733640552995392,
"grad_norm": 0.37506356835365295,
"learning_rate": 6.321274183083588e-06,
"loss": 0.4592,
"step": 856
},
{
"epoch": 0.47391705069124423,
"grad_norm": 0.4109135568141937,
"learning_rate": 6.3119603608927914e-06,
"loss": 0.4721,
"step": 857
},
{
"epoch": 0.4744700460829493,
"grad_norm": 0.4288492202758789,
"learning_rate": 6.302641647166402e-06,
"loss": 0.4766,
"step": 858
},
{
"epoch": 0.47502304147465435,
"grad_norm": 0.37833884358406067,
"learning_rate": 6.293318076648469e-06,
"loss": 0.4779,
"step": 859
},
{
"epoch": 0.47557603686635946,
"grad_norm": 0.4019952714443207,
"learning_rate": 6.283989684101155e-06,
"loss": 0.487,
"step": 860
},
{
"epoch": 0.4761290322580645,
"grad_norm": 0.45245492458343506,
"learning_rate": 6.274656504304598e-06,
"loss": 0.4582,
"step": 861
},
{
"epoch": 0.4766820276497696,
"grad_norm": 0.4037781357765198,
"learning_rate": 6.265318572056787e-06,
"loss": 0.4536,
"step": 862
},
{
"epoch": 0.47723502304147464,
"grad_norm": 0.4167582392692566,
"learning_rate": 6.25597592217343e-06,
"loss": 0.4737,
"step": 863
},
{
"epoch": 0.47778801843317975,
"grad_norm": 0.38921552896499634,
"learning_rate": 6.2466285894878206e-06,
"loss": 0.4889,
"step": 864
},
{
"epoch": 0.4783410138248848,
"grad_norm": 0.3599966764450073,
"learning_rate": 6.2372766088507195e-06,
"loss": 0.4421,
"step": 865
},
{
"epoch": 0.47889400921658987,
"grad_norm": 0.440920889377594,
"learning_rate": 6.227920015130209e-06,
"loss": 0.4959,
"step": 866
},
{
"epoch": 0.4794470046082949,
"grad_norm": 0.39445987343788147,
"learning_rate": 6.218558843211576e-06,
"loss": 0.4784,
"step": 867
},
{
"epoch": 0.48,
"grad_norm": 0.4886539578437805,
"learning_rate": 6.209193127997173e-06,
"loss": 0.4702,
"step": 868
},
{
"epoch": 0.4805529953917051,
"grad_norm": 0.43818390369415283,
"learning_rate": 6.199822904406295e-06,
"loss": 0.4685,
"step": 869
},
{
"epoch": 0.48110599078341015,
"grad_norm": 0.4038926064968109,
"learning_rate": 6.190448207375046e-06,
"loss": 0.4572,
"step": 870
},
{
"epoch": 0.4816589861751152,
"grad_norm": 0.5002937912940979,
"learning_rate": 6.1810690718562065e-06,
"loss": 0.4899,
"step": 871
},
{
"epoch": 0.48221198156682027,
"grad_norm": 0.39266350865364075,
"learning_rate": 6.1716855328191064e-06,
"loss": 0.4682,
"step": 872
},
{
"epoch": 0.4827649769585253,
"grad_norm": 0.4712425470352173,
"learning_rate": 6.162297625249494e-06,
"loss": 0.473,
"step": 873
},
{
"epoch": 0.48331797235023044,
"grad_norm": 0.33758747577667236,
"learning_rate": 6.1529053841494065e-06,
"loss": 0.4771,
"step": 874
},
{
"epoch": 0.4838709677419355,
"grad_norm": 0.442381888628006,
"learning_rate": 6.143508844537038e-06,
"loss": 0.4834,
"step": 875
},
{
"epoch": 0.48442396313364056,
"grad_norm": 0.43109333515167236,
"learning_rate": 6.134108041446609e-06,
"loss": 0.4717,
"step": 876
},
{
"epoch": 0.4849769585253456,
"grad_norm": 0.4240846037864685,
"learning_rate": 6.124703009928233e-06,
"loss": 0.491,
"step": 877
},
{
"epoch": 0.48552995391705067,
"grad_norm": 0.41452300548553467,
"learning_rate": 6.115293785047793e-06,
"loss": 0.4789,
"step": 878
},
{
"epoch": 0.4860829493087558,
"grad_norm": 0.397856742143631,
"learning_rate": 6.105880401886804e-06,
"loss": 0.4655,
"step": 879
},
{
"epoch": 0.48663594470046084,
"grad_norm": 0.3984890878200531,
"learning_rate": 6.096462895542288e-06,
"loss": 0.4864,
"step": 880
},
{
"epoch": 0.4871889400921659,
"grad_norm": 0.4527179002761841,
"learning_rate": 6.087041301126636e-06,
"loss": 0.4884,
"step": 881
},
{
"epoch": 0.48774193548387096,
"grad_norm": 0.4393288791179657,
"learning_rate": 6.077615653767484e-06,
"loss": 0.45,
"step": 882
},
{
"epoch": 0.488294930875576,
"grad_norm": 0.4285862445831299,
"learning_rate": 6.068185988607575e-06,
"loss": 0.4858,
"step": 883
},
{
"epoch": 0.48884792626728113,
"grad_norm": 0.4243047833442688,
"learning_rate": 6.058752340804639e-06,
"loss": 0.4801,
"step": 884
},
{
"epoch": 0.4894009216589862,
"grad_norm": 0.37787267565727234,
"learning_rate": 6.0493147455312475e-06,
"loss": 0.4584,
"step": 885
},
{
"epoch": 0.48995391705069125,
"grad_norm": 0.4291735589504242,
"learning_rate": 6.039873237974695e-06,
"loss": 0.5038,
"step": 886
},
{
"epoch": 0.4905069124423963,
"grad_norm": 0.3942468464374542,
"learning_rate": 6.03042785333686e-06,
"loss": 0.4805,
"step": 887
},
{
"epoch": 0.49105990783410136,
"grad_norm": 0.42691728472709656,
"learning_rate": 6.020978626834077e-06,
"loss": 0.4587,
"step": 888
},
{
"epoch": 0.4916129032258065,
"grad_norm": 0.3891282379627228,
"learning_rate": 6.0115255936970056e-06,
"loss": 0.4756,
"step": 889
},
{
"epoch": 0.49216589861751153,
"grad_norm": 0.39589089155197144,
"learning_rate": 6.002068789170497e-06,
"loss": 0.4796,
"step": 890
},
{
"epoch": 0.4927188940092166,
"grad_norm": 0.4499064087867737,
"learning_rate": 5.992608248513465e-06,
"loss": 0.4667,
"step": 891
},
{
"epoch": 0.49327188940092165,
"grad_norm": 0.4281262457370758,
"learning_rate": 5.983144006998751e-06,
"loss": 0.4762,
"step": 892
},
{
"epoch": 0.4938248847926267,
"grad_norm": 0.4095862805843353,
"learning_rate": 5.973676099912996e-06,
"loss": 0.4691,
"step": 893
},
{
"epoch": 0.4943778801843318,
"grad_norm": 0.4401033818721771,
"learning_rate": 5.964204562556508e-06,
"loss": 0.4688,
"step": 894
},
{
"epoch": 0.4949308755760369,
"grad_norm": 0.41105031967163086,
"learning_rate": 5.954729430243129e-06,
"loss": 0.4877,
"step": 895
},
{
"epoch": 0.49548387096774194,
"grad_norm": 0.39041486382484436,
"learning_rate": 5.945250738300108e-06,
"loss": 0.4504,
"step": 896
},
{
"epoch": 0.496036866359447,
"grad_norm": 0.4423321485519409,
"learning_rate": 5.935768522067962e-06,
"loss": 0.4858,
"step": 897
},
{
"epoch": 0.49658986175115205,
"grad_norm": 0.4127327799797058,
"learning_rate": 5.9262828169003476e-06,
"loss": 0.4767,
"step": 898
},
{
"epoch": 0.49714285714285716,
"grad_norm": 0.41609784960746765,
"learning_rate": 5.9167936581639325e-06,
"loss": 0.4901,
"step": 899
},
{
"epoch": 0.4976958525345622,
"grad_norm": 0.3699811100959778,
"learning_rate": 5.9073010812382595e-06,
"loss": 0.4595,
"step": 900
},
{
"epoch": 0.4982488479262673,
"grad_norm": 0.45880934596061707,
"learning_rate": 5.897805121515616e-06,
"loss": 0.4514,
"step": 901
},
{
"epoch": 0.49880184331797234,
"grad_norm": 0.38559094071388245,
"learning_rate": 5.888305814400901e-06,
"loss": 0.4871,
"step": 902
},
{
"epoch": 0.4993548387096774,
"grad_norm": 0.40912917256355286,
"learning_rate": 5.878803195311496e-06,
"loss": 0.4726,
"step": 903
},
{
"epoch": 0.4999078341013825,
"grad_norm": 0.4119787812232971,
"learning_rate": 5.869297299677128e-06,
"loss": 0.459,
"step": 904
},
{
"epoch": 0.5004608294930876,
"grad_norm": 0.4300439953804016,
"learning_rate": 5.8597881629397435e-06,
"loss": 0.4764,
"step": 905
},
{
"epoch": 0.5010138248847926,
"grad_norm": 0.4216427803039551,
"learning_rate": 5.850275820553368e-06,
"loss": 0.4795,
"step": 906
},
{
"epoch": 0.5015668202764977,
"grad_norm": 0.578460693359375,
"learning_rate": 5.840760307983988e-06,
"loss": 0.4463,
"step": 907
},
{
"epoch": 0.5021198156682027,
"grad_norm": 0.38813382387161255,
"learning_rate": 5.831241660709402e-06,
"loss": 0.479,
"step": 908
},
{
"epoch": 0.5026728110599078,
"grad_norm": 0.4912916123867035,
"learning_rate": 5.8217199142190975e-06,
"loss": 0.4846,
"step": 909
},
{
"epoch": 0.5032258064516129,
"grad_norm": 0.4877387285232544,
"learning_rate": 5.812195104014119e-06,
"loss": 0.5051,
"step": 910
},
{
"epoch": 0.503778801843318,
"grad_norm": 0.4228856861591339,
"learning_rate": 5.802667265606933e-06,
"loss": 0.4853,
"step": 911
},
{
"epoch": 0.5043317972350231,
"grad_norm": 0.4995596706867218,
"learning_rate": 5.793136434521296e-06,
"loss": 0.4603,
"step": 912
},
{
"epoch": 0.5048847926267281,
"grad_norm": 0.5033881664276123,
"learning_rate": 5.783602646292123e-06,
"loss": 0.4793,
"step": 913
},
{
"epoch": 0.5054377880184332,
"grad_norm": 0.41643261909484863,
"learning_rate": 5.774065936465352e-06,
"loss": 0.4744,
"step": 914
},
{
"epoch": 0.5059907834101383,
"grad_norm": 0.4517223834991455,
"learning_rate": 5.764526340597818e-06,
"loss": 0.4659,
"step": 915
},
{
"epoch": 0.5065437788018433,
"grad_norm": 0.3723732531070709,
"learning_rate": 5.754983894257114e-06,
"loss": 0.4727,
"step": 916
},
{
"epoch": 0.5070967741935484,
"grad_norm": 0.46375709772109985,
"learning_rate": 5.745438633021462e-06,
"loss": 0.4977,
"step": 917
},
{
"epoch": 0.5076497695852534,
"grad_norm": 0.42543715238571167,
"learning_rate": 5.7358905924795775e-06,
"loss": 0.4931,
"step": 918
},
{
"epoch": 0.5082027649769585,
"grad_norm": 0.3639947772026062,
"learning_rate": 5.726339808230539e-06,
"loss": 0.4608,
"step": 919
},
{
"epoch": 0.5087557603686635,
"grad_norm": 0.47503146529197693,
"learning_rate": 5.716786315883657e-06,
"loss": 0.4746,
"step": 920
},
{
"epoch": 0.5093087557603687,
"grad_norm": 0.4439897835254669,
"learning_rate": 5.707230151058334e-06,
"loss": 0.47,
"step": 921
},
{
"epoch": 0.5098617511520738,
"grad_norm": 0.43655335903167725,
"learning_rate": 5.697671349383939e-06,
"loss": 0.4668,
"step": 922
},
{
"epoch": 0.5104147465437788,
"grad_norm": 0.4014508128166199,
"learning_rate": 5.688109946499676e-06,
"loss": 0.4537,
"step": 923
},
{
"epoch": 0.5109677419354839,
"grad_norm": 0.4049840271472931,
"learning_rate": 5.678545978054443e-06,
"loss": 0.4713,
"step": 924
},
{
"epoch": 0.511520737327189,
"grad_norm": 0.4141314923763275,
"learning_rate": 5.668979479706703e-06,
"loss": 0.4882,
"step": 925
},
{
"epoch": 0.512073732718894,
"grad_norm": 0.4375897943973541,
"learning_rate": 5.659410487124355e-06,
"loss": 0.4625,
"step": 926
},
{
"epoch": 0.5126267281105991,
"grad_norm": 0.3951698839664459,
"learning_rate": 5.649839035984597e-06,
"loss": 0.4841,
"step": 927
},
{
"epoch": 0.5131797235023041,
"grad_norm": 0.367379367351532,
"learning_rate": 5.6402651619737865e-06,
"loss": 0.4818,
"step": 928
},
{
"epoch": 0.5137327188940092,
"grad_norm": 0.3982902765274048,
"learning_rate": 5.630688900787326e-06,
"loss": 0.4835,
"step": 929
},
{
"epoch": 0.5142857142857142,
"grad_norm": 0.41130173206329346,
"learning_rate": 5.621110288129509e-06,
"loss": 0.4592,
"step": 930
},
{
"epoch": 0.5148387096774194,
"grad_norm": 0.4003121554851532,
"learning_rate": 5.6115293597134015e-06,
"loss": 0.4511,
"step": 931
},
{
"epoch": 0.5153917050691245,
"grad_norm": 0.3777535855770111,
"learning_rate": 5.601946151260702e-06,
"loss": 0.4943,
"step": 932
},
{
"epoch": 0.5159447004608295,
"grad_norm": 0.46429774165153503,
"learning_rate": 5.592360698501609e-06,
"loss": 0.4709,
"step": 933
},
{
"epoch": 0.5164976958525346,
"grad_norm": 0.38206857442855835,
"learning_rate": 5.58277303717469e-06,
"loss": 0.4817,
"step": 934
},
{
"epoch": 0.5170506912442396,
"grad_norm": 0.39985257387161255,
"learning_rate": 5.573183203026747e-06,
"loss": 0.4646,
"step": 935
},
{
"epoch": 0.5176036866359447,
"grad_norm": 0.4025905728340149,
"learning_rate": 5.563591231812683e-06,
"loss": 0.471,
"step": 936
},
{
"epoch": 0.5181566820276498,
"grad_norm": 0.41740676760673523,
"learning_rate": 5.553997159295366e-06,
"loss": 0.4798,
"step": 937
},
{
"epoch": 0.5187096774193548,
"grad_norm": 0.395632266998291,
"learning_rate": 5.544401021245505e-06,
"loss": 0.4905,
"step": 938
},
{
"epoch": 0.5192626728110599,
"grad_norm": 0.3581371605396271,
"learning_rate": 5.534802853441503e-06,
"loss": 0.4563,
"step": 939
},
{
"epoch": 0.5198156682027649,
"grad_norm": 0.3925972282886505,
"learning_rate": 5.525202691669335e-06,
"loss": 0.4553,
"step": 940
},
{
"epoch": 0.5203686635944701,
"grad_norm": 0.3491620421409607,
"learning_rate": 5.515600571722412e-06,
"loss": 0.4602,
"step": 941
},
{
"epoch": 0.5209216589861752,
"grad_norm": 0.37922897934913635,
"learning_rate": 5.505996529401442e-06,
"loss": 0.4588,
"step": 942
},
{
"epoch": 0.5214746543778802,
"grad_norm": 0.37720414996147156,
"learning_rate": 5.496390600514298e-06,
"loss": 0.4688,
"step": 943
},
{
"epoch": 0.5220276497695853,
"grad_norm": 0.38247910141944885,
"learning_rate": 5.486782820875895e-06,
"loss": 0.4619,
"step": 944
},
{
"epoch": 0.5225806451612903,
"grad_norm": 0.3531082570552826,
"learning_rate": 5.477173226308042e-06,
"loss": 0.4611,
"step": 945
},
{
"epoch": 0.5231336405529954,
"grad_norm": 0.35197713971138,
"learning_rate": 5.4675618526393185e-06,
"loss": 0.483,
"step": 946
},
{
"epoch": 0.5236866359447004,
"grad_norm": 0.42967379093170166,
"learning_rate": 5.457948735704933e-06,
"loss": 0.469,
"step": 947
},
{
"epoch": 0.5242396313364055,
"grad_norm": 0.41465815901756287,
"learning_rate": 5.448333911346598e-06,
"loss": 0.4792,
"step": 948
},
{
"epoch": 0.5247926267281106,
"grad_norm": 0.385215163230896,
"learning_rate": 5.43871741541239e-06,
"loss": 0.4789,
"step": 949
},
{
"epoch": 0.5253456221198156,
"grad_norm": 0.39359477162361145,
"learning_rate": 5.429099283756618e-06,
"loss": 0.4356,
"step": 950
},
{
"epoch": 0.5258986175115208,
"grad_norm": 0.37385135889053345,
"learning_rate": 5.419479552239689e-06,
"loss": 0.4713,
"step": 951
},
{
"epoch": 0.5264516129032258,
"grad_norm": 0.40417802333831787,
"learning_rate": 5.4098582567279755e-06,
"loss": 0.4714,
"step": 952
},
{
"epoch": 0.5270046082949309,
"grad_norm": 0.415487140417099,
"learning_rate": 5.400235433093682e-06,
"loss": 0.4696,
"step": 953
},
{
"epoch": 0.527557603686636,
"grad_norm": 0.40797826647758484,
"learning_rate": 5.39061111721471e-06,
"loss": 0.4861,
"step": 954
},
{
"epoch": 0.528110599078341,
"grad_norm": 0.4314478039741516,
"learning_rate": 5.380985344974524e-06,
"loss": 0.4885,
"step": 955
},
{
"epoch": 0.5286635944700461,
"grad_norm": 0.3722938895225525,
"learning_rate": 5.371358152262019e-06,
"loss": 0.4576,
"step": 956
},
{
"epoch": 0.5292165898617511,
"grad_norm": 0.4123469293117523,
"learning_rate": 5.361729574971387e-06,
"loss": 0.4657,
"step": 957
},
{
"epoch": 0.5297695852534562,
"grad_norm": 0.4114021956920624,
"learning_rate": 5.352099649001979e-06,
"loss": 0.4705,
"step": 958
},
{
"epoch": 0.5303225806451612,
"grad_norm": 0.38689500093460083,
"learning_rate": 5.3424684102581795e-06,
"loss": 0.4661,
"step": 959
},
{
"epoch": 0.5308755760368664,
"grad_norm": 0.4139274060726166,
"learning_rate": 5.3328358946492634e-06,
"loss": 0.4721,
"step": 960
},
{
"epoch": 0.5314285714285715,
"grad_norm": 0.3889627754688263,
"learning_rate": 5.323202138089266e-06,
"loss": 0.4632,
"step": 961
},
{
"epoch": 0.5319815668202765,
"grad_norm": 0.3822690546512604,
"learning_rate": 5.313567176496854e-06,
"loss": 0.4522,
"step": 962
},
{
"epoch": 0.5325345622119816,
"grad_norm": 0.4444742798805237,
"learning_rate": 5.303931045795184e-06,
"loss": 0.4627,
"step": 963
},
{
"epoch": 0.5330875576036866,
"grad_norm": 0.4218832552433014,
"learning_rate": 5.294293781911769e-06,
"loss": 0.4762,
"step": 964
},
{
"epoch": 0.5336405529953917,
"grad_norm": 0.40190473198890686,
"learning_rate": 5.284655420778353e-06,
"loss": 0.4399,
"step": 965
},
{
"epoch": 0.5341935483870968,
"grad_norm": 0.44916805624961853,
"learning_rate": 5.275015998330765e-06,
"loss": 0.4457,
"step": 966
},
{
"epoch": 0.5347465437788018,
"grad_norm": 0.4550216794013977,
"learning_rate": 5.265375550508793e-06,
"loss": 0.4719,
"step": 967
},
{
"epoch": 0.5352995391705069,
"grad_norm": 0.42324307560920715,
"learning_rate": 5.255734113256051e-06,
"loss": 0.46,
"step": 968
},
{
"epoch": 0.5358525345622119,
"grad_norm": 0.3758416473865509,
"learning_rate": 5.246091722519839e-06,
"loss": 0.4441,
"step": 969
},
{
"epoch": 0.5364055299539171,
"grad_norm": 0.37716051936149597,
"learning_rate": 5.236448414251012e-06,
"loss": 0.4643,
"step": 970
},
{
"epoch": 0.5369585253456222,
"grad_norm": 0.3788035809993744,
"learning_rate": 5.226804224403846e-06,
"loss": 0.4592,
"step": 971
},
{
"epoch": 0.5375115207373272,
"grad_norm": 0.38398197293281555,
"learning_rate": 5.217159188935907e-06,
"loss": 0.4744,
"step": 972
},
{
"epoch": 0.5380645161290323,
"grad_norm": 0.3767092823982239,
"learning_rate": 5.207513343807907e-06,
"loss": 0.4698,
"step": 973
},
{
"epoch": 0.5386175115207373,
"grad_norm": 0.39843782782554626,
"learning_rate": 5.1978667249835855e-06,
"loss": 0.4803,
"step": 974
},
{
"epoch": 0.5391705069124424,
"grad_norm": 0.38011598587036133,
"learning_rate": 5.18821936842956e-06,
"loss": 0.4398,
"step": 975
},
{
"epoch": 0.5397235023041475,
"grad_norm": 0.36936551332473755,
"learning_rate": 5.1785713101152e-06,
"loss": 0.4662,
"step": 976
},
{
"epoch": 0.5402764976958525,
"grad_norm": 0.3832260072231293,
"learning_rate": 5.168922586012495e-06,
"loss": 0.488,
"step": 977
},
{
"epoch": 0.5408294930875576,
"grad_norm": 0.38859879970550537,
"learning_rate": 5.159273232095911e-06,
"loss": 0.4982,
"step": 978
},
{
"epoch": 0.5413824884792626,
"grad_norm": 0.4032990038394928,
"learning_rate": 5.1496232843422665e-06,
"loss": 0.4919,
"step": 979
},
{
"epoch": 0.5419354838709678,
"grad_norm": 0.3422912061214447,
"learning_rate": 5.139972778730593e-06,
"loss": 0.4658,
"step": 980
},
{
"epoch": 0.5424884792626729,
"grad_norm": 0.38321739435195923,
"learning_rate": 5.130321751242002e-06,
"loss": 0.4738,
"step": 981
},
{
"epoch": 0.5430414746543779,
"grad_norm": 0.3506300151348114,
"learning_rate": 5.12067023785955e-06,
"loss": 0.4612,
"step": 982
},
{
"epoch": 0.543594470046083,
"grad_norm": 0.3691524565219879,
"learning_rate": 5.111018274568106e-06,
"loss": 0.4461,
"step": 983
},
{
"epoch": 0.544147465437788,
"grad_norm": 0.37267613410949707,
"learning_rate": 5.101365897354215e-06,
"loss": 0.4891,
"step": 984
},
{
"epoch": 0.5447004608294931,
"grad_norm": 0.3920471668243408,
"learning_rate": 5.091713142205967e-06,
"loss": 0.4645,
"step": 985
},
{
"epoch": 0.5452534562211981,
"grad_norm": 0.37700968980789185,
"learning_rate": 5.082060045112862e-06,
"loss": 0.4721,
"step": 986
},
{
"epoch": 0.5458064516129032,
"grad_norm": 0.4021558165550232,
"learning_rate": 5.072406642065671e-06,
"loss": 0.4825,
"step": 987
},
{
"epoch": 0.5463594470046083,
"grad_norm": 0.40953153371810913,
"learning_rate": 5.062752969056311e-06,
"loss": 0.4478,
"step": 988
},
{
"epoch": 0.5469124423963133,
"grad_norm": 0.4201180934906006,
"learning_rate": 5.0530990620777e-06,
"loss": 0.4405,
"step": 989
},
{
"epoch": 0.5474654377880185,
"grad_norm": 0.42839857935905457,
"learning_rate": 5.0434449571236314e-06,
"loss": 0.4774,
"step": 990
},
{
"epoch": 0.5480184331797235,
"grad_norm": 0.40306076407432556,
"learning_rate": 5.033790690188637e-06,
"loss": 0.475,
"step": 991
},
{
"epoch": 0.5485714285714286,
"grad_norm": 0.3931422829627991,
"learning_rate": 5.024136297267851e-06,
"loss": 0.4813,
"step": 992
},
{
"epoch": 0.5491244239631337,
"grad_norm": 0.41347619891166687,
"learning_rate": 5.01448181435688e-06,
"loss": 0.4726,
"step": 993
},
{
"epoch": 0.5496774193548387,
"grad_norm": 0.41298919916152954,
"learning_rate": 5.004827277451661e-06,
"loss": 0.4581,
"step": 994
},
{
"epoch": 0.5502304147465438,
"grad_norm": 0.3852965235710144,
"learning_rate": 4.9951727225483395e-06,
"loss": 0.4638,
"step": 995
},
{
"epoch": 0.5507834101382488,
"grad_norm": 0.37419193983078003,
"learning_rate": 4.985518185643122e-06,
"loss": 0.4675,
"step": 996
},
{
"epoch": 0.5513364055299539,
"grad_norm": 0.36649253964424133,
"learning_rate": 4.975863702732151e-06,
"loss": 0.4584,
"step": 997
},
{
"epoch": 0.551889400921659,
"grad_norm": 0.4342942237854004,
"learning_rate": 4.966209309811364e-06,
"loss": 0.4731,
"step": 998
},
{
"epoch": 0.552442396313364,
"grad_norm": 0.37662389874458313,
"learning_rate": 4.95655504287637e-06,
"loss": 0.4696,
"step": 999
},
{
"epoch": 0.5529953917050692,
"grad_norm": 0.40196794271469116,
"learning_rate": 4.946900937922302e-06,
"loss": 0.4496,
"step": 1000
},
{
"epoch": 0.5535483870967742,
"grad_norm": 0.3595891296863556,
"learning_rate": 4.93724703094369e-06,
"loss": 0.4621,
"step": 1001
},
{
"epoch": 0.5541013824884793,
"grad_norm": 0.4029429256916046,
"learning_rate": 4.927593357934329e-06,
"loss": 0.4885,
"step": 1002
},
{
"epoch": 0.5546543778801843,
"grad_norm": 0.37034833431243896,
"learning_rate": 4.917939954887139e-06,
"loss": 0.462,
"step": 1003
},
{
"epoch": 0.5552073732718894,
"grad_norm": 0.4130437672138214,
"learning_rate": 4.908286857794034e-06,
"loss": 0.4646,
"step": 1004
},
{
"epoch": 0.5557603686635945,
"grad_norm": 0.39467036724090576,
"learning_rate": 4.8986341026457875e-06,
"loss": 0.4557,
"step": 1005
},
{
"epoch": 0.5563133640552995,
"grad_norm": 0.37382972240448,
"learning_rate": 4.888981725431897e-06,
"loss": 0.4589,
"step": 1006
},
{
"epoch": 0.5568663594470046,
"grad_norm": 0.3812313377857208,
"learning_rate": 4.879329762140453e-06,
"loss": 0.476,
"step": 1007
},
{
"epoch": 0.5574193548387096,
"grad_norm": 0.3706028461456299,
"learning_rate": 4.8696782487580005e-06,
"loss": 0.4534,
"step": 1008
},
{
"epoch": 0.5579723502304147,
"grad_norm": 0.40674856305122375,
"learning_rate": 4.8600272212694094e-06,
"loss": 0.4555,
"step": 1009
},
{
"epoch": 0.5585253456221199,
"grad_norm": 0.3949076235294342,
"learning_rate": 4.850376715657736e-06,
"loss": 0.4569,
"step": 1010
},
{
"epoch": 0.5590783410138249,
"grad_norm": 0.3829311430454254,
"learning_rate": 4.840726767904089e-06,
"loss": 0.4589,
"step": 1011
},
{
"epoch": 0.55963133640553,
"grad_norm": 0.36710280179977417,
"learning_rate": 4.831077413987505e-06,
"loss": 0.4763,
"step": 1012
},
{
"epoch": 0.560184331797235,
"grad_norm": 0.37851566076278687,
"learning_rate": 4.8214286898848e-06,
"loss": 0.4723,
"step": 1013
},
{
"epoch": 0.5607373271889401,
"grad_norm": 0.3961973488330841,
"learning_rate": 4.81178063157044e-06,
"loss": 0.456,
"step": 1014
},
{
"epoch": 0.5612903225806452,
"grad_norm": 0.38343876600265503,
"learning_rate": 4.802133275016415e-06,
"loss": 0.4626,
"step": 1015
},
{
"epoch": 0.5618433179723502,
"grad_norm": 0.3635197877883911,
"learning_rate": 4.792486656192094e-06,
"loss": 0.4692,
"step": 1016
},
{
"epoch": 0.5623963133640553,
"grad_norm": 0.3797365128993988,
"learning_rate": 4.782840811064095e-06,
"loss": 0.4587,
"step": 1017
},
{
"epoch": 0.5629493087557603,
"grad_norm": 0.35782304406166077,
"learning_rate": 4.773195775596155e-06,
"loss": 0.4503,
"step": 1018
},
{
"epoch": 0.5635023041474654,
"grad_norm": 0.370126336812973,
"learning_rate": 4.763551585748989e-06,
"loss": 0.4588,
"step": 1019
},
{
"epoch": 0.5640552995391706,
"grad_norm": 0.3596176207065582,
"learning_rate": 4.753908277480162e-06,
"loss": 0.4919,
"step": 1020
},
{
"epoch": 0.5646082949308756,
"grad_norm": 0.343191921710968,
"learning_rate": 4.74426588674395e-06,
"loss": 0.4667,
"step": 1021
},
{
"epoch": 0.5651612903225807,
"grad_norm": 0.3751341998577118,
"learning_rate": 4.734624449491208e-06,
"loss": 0.4529,
"step": 1022
},
{
"epoch": 0.5657142857142857,
"grad_norm": 0.3533148169517517,
"learning_rate": 4.724984001669237e-06,
"loss": 0.4491,
"step": 1023
},
{
"epoch": 0.5662672811059908,
"grad_norm": 0.38230666518211365,
"learning_rate": 4.715344579221649e-06,
"loss": 0.4961,
"step": 1024
},
{
"epoch": 0.5668202764976958,
"grad_norm": 0.40656420588493347,
"learning_rate": 4.705706218088232e-06,
"loss": 0.4773,
"step": 1025
},
{
"epoch": 0.5673732718894009,
"grad_norm": 0.3879065215587616,
"learning_rate": 4.696068954204817e-06,
"loss": 0.5097,
"step": 1026
},
{
"epoch": 0.567926267281106,
"grad_norm": 0.40097224712371826,
"learning_rate": 4.686432823503147e-06,
"loss": 0.4661,
"step": 1027
},
{
"epoch": 0.568479262672811,
"grad_norm": 0.37987810373306274,
"learning_rate": 4.676797861910735e-06,
"loss": 0.4429,
"step": 1028
},
{
"epoch": 0.5690322580645162,
"grad_norm": 0.38563069701194763,
"learning_rate": 4.667164105350739e-06,
"loss": 0.4719,
"step": 1029
},
{
"epoch": 0.5695852534562212,
"grad_norm": 0.3681628406047821,
"learning_rate": 4.657531589741822e-06,
"loss": 0.4757,
"step": 1030
},
{
"epoch": 0.5701382488479263,
"grad_norm": 0.4479837119579315,
"learning_rate": 4.647900350998022e-06,
"loss": 0.471,
"step": 1031
},
{
"epoch": 0.5706912442396314,
"grad_norm": 0.39742401242256165,
"learning_rate": 4.638270425028614e-06,
"loss": 0.4762,
"step": 1032
},
{
"epoch": 0.5712442396313364,
"grad_norm": 0.38390353322029114,
"learning_rate": 4.628641847737982e-06,
"loss": 0.4475,
"step": 1033
},
{
"epoch": 0.5717972350230415,
"grad_norm": 0.4764840602874756,
"learning_rate": 4.6190146550254775e-06,
"loss": 0.4827,
"step": 1034
},
{
"epoch": 0.5723502304147465,
"grad_norm": 0.41221946477890015,
"learning_rate": 4.609388882785291e-06,
"loss": 0.495,
"step": 1035
},
{
"epoch": 0.5729032258064516,
"grad_norm": 0.35299742221832275,
"learning_rate": 4.599764566906319e-06,
"loss": 0.4614,
"step": 1036
},
{
"epoch": 0.5734562211981566,
"grad_norm": 0.45965003967285156,
"learning_rate": 4.590141743272026e-06,
"loss": 0.462,
"step": 1037
},
{
"epoch": 0.5740092165898617,
"grad_norm": 0.37869590520858765,
"learning_rate": 4.5805204477603135e-06,
"loss": 0.4442,
"step": 1038
},
{
"epoch": 0.5745622119815669,
"grad_norm": 0.37475693225860596,
"learning_rate": 4.570900716243385e-06,
"loss": 0.4468,
"step": 1039
},
{
"epoch": 0.5751152073732719,
"grad_norm": 0.38735684752464294,
"learning_rate": 4.561282584587612e-06,
"loss": 0.4444,
"step": 1040
},
{
"epoch": 0.575668202764977,
"grad_norm": 0.3577326834201813,
"learning_rate": 4.551666088653404e-06,
"loss": 0.4864,
"step": 1041
},
{
"epoch": 0.576221198156682,
"grad_norm": 0.3724598288536072,
"learning_rate": 4.542051264295069e-06,
"loss": 0.4556,
"step": 1042
},
{
"epoch": 0.5767741935483871,
"grad_norm": 0.4160480797290802,
"learning_rate": 4.532438147360684e-06,
"loss": 0.4599,
"step": 1043
},
{
"epoch": 0.5773271889400922,
"grad_norm": 0.38303178548812866,
"learning_rate": 4.522826773691958e-06,
"loss": 0.4574,
"step": 1044
},
{
"epoch": 0.5778801843317972,
"grad_norm": 0.4100678861141205,
"learning_rate": 4.513217179124106e-06,
"loss": 0.4637,
"step": 1045
},
{
"epoch": 0.5784331797235023,
"grad_norm": 0.4295979142189026,
"learning_rate": 4.5036093994857025e-06,
"loss": 0.4747,
"step": 1046
},
{
"epoch": 0.5789861751152073,
"grad_norm": 0.36032918095588684,
"learning_rate": 4.49400347059856e-06,
"loss": 0.4682,
"step": 1047
},
{
"epoch": 0.5795391705069124,
"grad_norm": 0.4092552959918976,
"learning_rate": 4.484399428277589e-06,
"loss": 0.463,
"step": 1048
},
{
"epoch": 0.5800921658986176,
"grad_norm": 0.3771011233329773,
"learning_rate": 4.474797308330665e-06,
"loss": 0.4865,
"step": 1049
},
{
"epoch": 0.5806451612903226,
"grad_norm": 0.37506765127182007,
"learning_rate": 4.465197146558498e-06,
"loss": 0.4558,
"step": 1050
},
{
"epoch": 0.5811981566820277,
"grad_norm": 0.34741005301475525,
"learning_rate": 4.455598978754496e-06,
"loss": 0.445,
"step": 1051
},
{
"epoch": 0.5817511520737327,
"grad_norm": 0.41416463255882263,
"learning_rate": 4.4460028407046344e-06,
"loss": 0.4755,
"step": 1052
},
{
"epoch": 0.5823041474654378,
"grad_norm": 0.414931982755661,
"learning_rate": 4.436408768187319e-06,
"loss": 0.4879,
"step": 1053
},
{
"epoch": 0.5828571428571429,
"grad_norm": 0.37698793411254883,
"learning_rate": 4.426816796973254e-06,
"loss": 0.4896,
"step": 1054
},
{
"epoch": 0.5834101382488479,
"grad_norm": 0.4369506537914276,
"learning_rate": 4.417226962825311e-06,
"loss": 0.4723,
"step": 1055
},
{
"epoch": 0.583963133640553,
"grad_norm": 0.3790305256843567,
"learning_rate": 4.407639301498392e-06,
"loss": 0.4466,
"step": 1056
},
{
"epoch": 0.584516129032258,
"grad_norm": 0.3516313135623932,
"learning_rate": 4.398053848739299e-06,
"loss": 0.4483,
"step": 1057
},
{
"epoch": 0.5850691244239631,
"grad_norm": 0.42550036311149597,
"learning_rate": 4.388470640286599e-06,
"loss": 0.4633,
"step": 1058
},
{
"epoch": 0.5856221198156683,
"grad_norm": 0.3732847571372986,
"learning_rate": 4.378889711870492e-06,
"loss": 0.4909,
"step": 1059
},
{
"epoch": 0.5861751152073733,
"grad_norm": 0.3723104000091553,
"learning_rate": 4.369311099212676e-06,
"loss": 0.4612,
"step": 1060
},
{
"epoch": 0.5867281105990784,
"grad_norm": 0.4225739538669586,
"learning_rate": 4.359734838026214e-06,
"loss": 0.4576,
"step": 1061
},
{
"epoch": 0.5872811059907834,
"grad_norm": 0.4037158489227295,
"learning_rate": 4.3501609640154056e-06,
"loss": 0.4803,
"step": 1062
},
{
"epoch": 0.5878341013824885,
"grad_norm": 0.3930150866508484,
"learning_rate": 4.340589512875646e-06,
"loss": 0.4484,
"step": 1063
},
{
"epoch": 0.5883870967741935,
"grad_norm": 0.4097388684749603,
"learning_rate": 4.331020520293298e-06,
"loss": 0.4699,
"step": 1064
},
{
"epoch": 0.5889400921658986,
"grad_norm": 0.3482785224914551,
"learning_rate": 4.321454021945559e-06,
"loss": 0.4632,
"step": 1065
},
{
"epoch": 0.5894930875576037,
"grad_norm": 0.45652082562446594,
"learning_rate": 4.311890053500326e-06,
"loss": 0.4699,
"step": 1066
},
{
"epoch": 0.5900460829493087,
"grad_norm": 0.3783150613307953,
"learning_rate": 4.302328650616062e-06,
"loss": 0.4662,
"step": 1067
},
{
"epoch": 0.5905990783410138,
"grad_norm": 0.36994531750679016,
"learning_rate": 4.292769848941669e-06,
"loss": 0.4581,
"step": 1068
},
{
"epoch": 0.5911520737327189,
"grad_norm": 0.3842531144618988,
"learning_rate": 4.283213684116346e-06,
"loss": 0.4631,
"step": 1069
},
{
"epoch": 0.591705069124424,
"grad_norm": 0.4432550370693207,
"learning_rate": 4.273660191769463e-06,
"loss": 0.4759,
"step": 1070
},
{
"epoch": 0.5922580645161291,
"grad_norm": 0.37499210238456726,
"learning_rate": 4.264109407520425e-06,
"loss": 0.4703,
"step": 1071
},
{
"epoch": 0.5928110599078341,
"grad_norm": 0.383779376745224,
"learning_rate": 4.2545613669785405e-06,
"loss": 0.4793,
"step": 1072
},
{
"epoch": 0.5933640552995392,
"grad_norm": 0.43077680468559265,
"learning_rate": 4.245016105742889e-06,
"loss": 0.447,
"step": 1073
},
{
"epoch": 0.5939170506912442,
"grad_norm": 0.3452954888343811,
"learning_rate": 4.2354736594021854e-06,
"loss": 0.457,
"step": 1074
},
{
"epoch": 0.5944700460829493,
"grad_norm": 0.3940063714981079,
"learning_rate": 4.2259340635346515e-06,
"loss": 0.4579,
"step": 1075
},
{
"epoch": 0.5950230414746543,
"grad_norm": 0.37174075841903687,
"learning_rate": 4.21639735370788e-06,
"loss": 0.4718,
"step": 1076
},
{
"epoch": 0.5955760368663594,
"grad_norm": 0.38617780804634094,
"learning_rate": 4.206863565478705e-06,
"loss": 0.4698,
"step": 1077
},
{
"epoch": 0.5961290322580645,
"grad_norm": 0.36345598101615906,
"learning_rate": 4.197332734393067e-06,
"loss": 0.4828,
"step": 1078
},
{
"epoch": 0.5966820276497696,
"grad_norm": 0.37622806429862976,
"learning_rate": 4.187804895985881e-06,
"loss": 0.4584,
"step": 1079
},
{
"epoch": 0.5972350230414747,
"grad_norm": 0.41539904475212097,
"learning_rate": 4.1782800857809025e-06,
"loss": 0.4606,
"step": 1080
},
{
"epoch": 0.5977880184331797,
"grad_norm": 0.37827736139297485,
"learning_rate": 4.168758339290599e-06,
"loss": 0.4576,
"step": 1081
},
{
"epoch": 0.5983410138248848,
"grad_norm": 0.36408886313438416,
"learning_rate": 4.159239692016012e-06,
"loss": 0.4792,
"step": 1082
},
{
"epoch": 0.5988940092165899,
"grad_norm": 0.347615510225296,
"learning_rate": 4.149724179446631e-06,
"loss": 0.4627,
"step": 1083
},
{
"epoch": 0.5994470046082949,
"grad_norm": 0.36233338713645935,
"learning_rate": 4.140211837060258e-06,
"loss": 0.4595,
"step": 1084
},
{
"epoch": 0.6,
"grad_norm": 0.385028213262558,
"learning_rate": 4.130702700322873e-06,
"loss": 0.4469,
"step": 1085
},
{
"epoch": 0.600552995391705,
"grad_norm": 0.3965871036052704,
"learning_rate": 4.121196804688506e-06,
"loss": 0.4761,
"step": 1086
},
{
"epoch": 0.6011059907834101,
"grad_norm": 0.36717289686203003,
"learning_rate": 4.111694185599101e-06,
"loss": 0.4401,
"step": 1087
},
{
"epoch": 0.6016589861751153,
"grad_norm": 0.3657437264919281,
"learning_rate": 4.102194878484386e-06,
"loss": 0.4613,
"step": 1088
},
{
"epoch": 0.6022119815668203,
"grad_norm": 0.4054916799068451,
"learning_rate": 4.092698918761742e-06,
"loss": 0.489,
"step": 1089
},
{
"epoch": 0.6027649769585254,
"grad_norm": 0.40190011262893677,
"learning_rate": 4.083206341836069e-06,
"loss": 0.4703,
"step": 1090
},
{
"epoch": 0.6033179723502304,
"grad_norm": 0.4301840662956238,
"learning_rate": 4.073717183099654e-06,
"loss": 0.4676,
"step": 1091
},
{
"epoch": 0.6038709677419355,
"grad_norm": 0.3311327397823334,
"learning_rate": 4.06423147793204e-06,
"loss": 0.4695,
"step": 1092
},
{
"epoch": 0.6044239631336406,
"grad_norm": 0.3920708894729614,
"learning_rate": 4.054749261699893e-06,
"loss": 0.4834,
"step": 1093
},
{
"epoch": 0.6049769585253456,
"grad_norm": 0.33984097838401794,
"learning_rate": 4.045270569756872e-06,
"loss": 0.4633,
"step": 1094
},
{
"epoch": 0.6055299539170507,
"grad_norm": 0.36689960956573486,
"learning_rate": 4.0357954374434936e-06,
"loss": 0.4778,
"step": 1095
},
{
"epoch": 0.6060829493087557,
"grad_norm": 0.3129914402961731,
"learning_rate": 4.026323900087006e-06,
"loss": 0.4389,
"step": 1096
},
{
"epoch": 0.6066359447004608,
"grad_norm": 0.3611215651035309,
"learning_rate": 4.016855993001251e-06,
"loss": 0.4687,
"step": 1097
},
{
"epoch": 0.607188940092166,
"grad_norm": 0.36361318826675415,
"learning_rate": 4.007391751486536e-06,
"loss": 0.4592,
"step": 1098
},
{
"epoch": 0.607741935483871,
"grad_norm": 0.3965631425380707,
"learning_rate": 3.997931210829503e-06,
"loss": 0.4403,
"step": 1099
},
{
"epoch": 0.6082949308755761,
"grad_norm": 0.371737539768219,
"learning_rate": 3.988474406302995e-06,
"loss": 0.4433,
"step": 1100
},
{
"epoch": 0.6088479262672811,
"grad_norm": 0.4065137207508087,
"learning_rate": 3.979021373165924e-06,
"loss": 0.4548,
"step": 1101
},
{
"epoch": 0.6094009216589862,
"grad_norm": 0.36007609963417053,
"learning_rate": 3.9695721466631426e-06,
"loss": 0.4664,
"step": 1102
},
{
"epoch": 0.6099539170506912,
"grad_norm": 0.39393535256385803,
"learning_rate": 3.9601267620253075e-06,
"loss": 0.4899,
"step": 1103
},
{
"epoch": 0.6105069124423963,
"grad_norm": 0.40391650795936584,
"learning_rate": 3.950685254468755e-06,
"loss": 0.4657,
"step": 1104
},
{
"epoch": 0.6110599078341014,
"grad_norm": 0.3698117434978485,
"learning_rate": 3.941247659195364e-06,
"loss": 0.4524,
"step": 1105
},
{
"epoch": 0.6116129032258064,
"grad_norm": 0.35965868830680847,
"learning_rate": 3.931814011392427e-06,
"loss": 0.4629,
"step": 1106
},
{
"epoch": 0.6121658986175115,
"grad_norm": 0.4399069547653198,
"learning_rate": 3.922384346232519e-06,
"loss": 0.491,
"step": 1107
},
{
"epoch": 0.6127188940092166,
"grad_norm": 0.3660935163497925,
"learning_rate": 3.912958698873366e-06,
"loss": 0.474,
"step": 1108
},
{
"epoch": 0.6132718894009217,
"grad_norm": 0.3902994990348816,
"learning_rate": 3.903537104457713e-06,
"loss": 0.4979,
"step": 1109
},
{
"epoch": 0.6138248847926268,
"grad_norm": 0.4067245125770569,
"learning_rate": 3.894119598113196e-06,
"loss": 0.4724,
"step": 1110
},
{
"epoch": 0.6143778801843318,
"grad_norm": 0.3835393190383911,
"learning_rate": 3.8847062149522075e-06,
"loss": 0.4558,
"step": 1111
},
{
"epoch": 0.6149308755760369,
"grad_norm": 0.3831106126308441,
"learning_rate": 3.875296990071768e-06,
"loss": 0.4619,
"step": 1112
},
{
"epoch": 0.6154838709677419,
"grad_norm": 0.3794800043106079,
"learning_rate": 3.8658919585533916e-06,
"loss": 0.4535,
"step": 1113
},
{
"epoch": 0.616036866359447,
"grad_norm": 0.3396313786506653,
"learning_rate": 3.8564911554629616e-06,
"loss": 0.4892,
"step": 1114
},
{
"epoch": 0.616589861751152,
"grad_norm": 0.37773552536964417,
"learning_rate": 3.847094615850593e-06,
"loss": 0.46,
"step": 1115
},
{
"epoch": 0.6171428571428571,
"grad_norm": 0.38637781143188477,
"learning_rate": 3.837702374750507e-06,
"loss": 0.4701,
"step": 1116
},
{
"epoch": 0.6176958525345622,
"grad_norm": 0.3425704538822174,
"learning_rate": 3.828314467180895e-06,
"loss": 0.4389,
"step": 1117
},
{
"epoch": 0.6182488479262673,
"grad_norm": 0.35482358932495117,
"learning_rate": 3.818930928143796e-06,
"loss": 0.477,
"step": 1118
},
{
"epoch": 0.6188018433179724,
"grad_norm": 0.3860636353492737,
"learning_rate": 3.8095517926249557e-06,
"loss": 0.4748,
"step": 1119
},
{
"epoch": 0.6193548387096774,
"grad_norm": 0.412280797958374,
"learning_rate": 3.800177095593706e-06,
"loss": 0.4624,
"step": 1120
},
{
"epoch": 0.6199078341013825,
"grad_norm": 0.3511303961277008,
"learning_rate": 3.790806872002828e-06,
"loss": 0.4867,
"step": 1121
},
{
"epoch": 0.6204608294930876,
"grad_norm": 0.3653261065483093,
"learning_rate": 3.7814411567884256e-06,
"loss": 0.4737,
"step": 1122
},
{
"epoch": 0.6210138248847926,
"grad_norm": 0.3349493145942688,
"learning_rate": 3.7720799848697924e-06,
"loss": 0.4705,
"step": 1123
},
{
"epoch": 0.6215668202764977,
"grad_norm": 0.3643210828304291,
"learning_rate": 3.762723391149282e-06,
"loss": 0.4934,
"step": 1124
},
{
"epoch": 0.6221198156682027,
"grad_norm": 0.3928675055503845,
"learning_rate": 3.75337141051218e-06,
"loss": 0.4726,
"step": 1125
},
{
"epoch": 0.6226728110599078,
"grad_norm": 0.36943259835243225,
"learning_rate": 3.7440240778265725e-06,
"loss": 0.4557,
"step": 1126
},
{
"epoch": 0.6232258064516129,
"grad_norm": 0.37895265221595764,
"learning_rate": 3.734681427943214e-06,
"loss": 0.4989,
"step": 1127
},
{
"epoch": 0.623778801843318,
"grad_norm": 0.34087231755256653,
"learning_rate": 3.725343495695404e-06,
"loss": 0.4418,
"step": 1128
},
{
"epoch": 0.6243317972350231,
"grad_norm": 0.3866007626056671,
"learning_rate": 3.716010315898847e-06,
"loss": 0.4515,
"step": 1129
},
{
"epoch": 0.6248847926267281,
"grad_norm": 0.4217049777507782,
"learning_rate": 3.706681923351533e-06,
"loss": 0.4962,
"step": 1130
},
{
"epoch": 0.6254377880184332,
"grad_norm": 0.4117668867111206,
"learning_rate": 3.6973583528336e-06,
"loss": 0.4704,
"step": 1131
},
{
"epoch": 0.6259907834101383,
"grad_norm": 0.3604739308357239,
"learning_rate": 3.68803963910721e-06,
"loss": 0.4839,
"step": 1132
},
{
"epoch": 0.6265437788018433,
"grad_norm": 0.4002184569835663,
"learning_rate": 3.678725816916415e-06,
"loss": 0.4853,
"step": 1133
},
{
"epoch": 0.6270967741935484,
"grad_norm": 0.3674345910549164,
"learning_rate": 3.6694169209870305e-06,
"loss": 0.4471,
"step": 1134
},
{
"epoch": 0.6276497695852534,
"grad_norm": 0.37487921118736267,
"learning_rate": 3.660112986026504e-06,
"loss": 0.4646,
"step": 1135
},
{
"epoch": 0.6282027649769585,
"grad_norm": 0.36266854405403137,
"learning_rate": 3.6508140467237873e-06,
"loss": 0.4802,
"step": 1136
},
{
"epoch": 0.6287557603686635,
"grad_norm": 0.3480757176876068,
"learning_rate": 3.6415201377492048e-06,
"loss": 0.458,
"step": 1137
},
{
"epoch": 0.6293087557603687,
"grad_norm": 0.3710945248603821,
"learning_rate": 3.632231293754328e-06,
"loss": 0.4648,
"step": 1138
},
{
"epoch": 0.6298617511520738,
"grad_norm": 0.42196762561798096,
"learning_rate": 3.622947549371841e-06,
"loss": 0.4802,
"step": 1139
},
{
"epoch": 0.6304147465437788,
"grad_norm": 0.3192515969276428,
"learning_rate": 3.6136689392154186e-06,
"loss": 0.4559,
"step": 1140
},
{
"epoch": 0.6309677419354839,
"grad_norm": 0.37404048442840576,
"learning_rate": 3.6043954978795868e-06,
"loss": 0.4647,
"step": 1141
},
{
"epoch": 0.6315207373271889,
"grad_norm": 0.37971508502960205,
"learning_rate": 3.5951272599396086e-06,
"loss": 0.4347,
"step": 1142
},
{
"epoch": 0.632073732718894,
"grad_norm": 0.410057932138443,
"learning_rate": 3.5858642599513394e-06,
"loss": 0.4759,
"step": 1143
},
{
"epoch": 0.6326267281105991,
"grad_norm": 0.367079496383667,
"learning_rate": 3.576606532451108e-06,
"loss": 0.4864,
"step": 1144
},
{
"epoch": 0.6331797235023041,
"grad_norm": 0.3866696357727051,
"learning_rate": 3.567354111955585e-06,
"loss": 0.4746,
"step": 1145
},
{
"epoch": 0.6337327188940092,
"grad_norm": 0.3390635848045349,
"learning_rate": 3.5581070329616543e-06,
"loss": 0.4593,
"step": 1146
},
{
"epoch": 0.6342857142857142,
"grad_norm": 0.41744962334632874,
"learning_rate": 3.5488653299462844e-06,
"loss": 0.4514,
"step": 1147
},
{
"epoch": 0.6348387096774194,
"grad_norm": 0.3437979817390442,
"learning_rate": 3.5396290373664e-06,
"loss": 0.4689,
"step": 1148
},
{
"epoch": 0.6353917050691245,
"grad_norm": 0.32637640833854675,
"learning_rate": 3.5303981896587534e-06,
"loss": 0.4657,
"step": 1149
},
{
"epoch": 0.6359447004608295,
"grad_norm": 0.3600415885448456,
"learning_rate": 3.521172821239796e-06,
"loss": 0.4637,
"step": 1150
},
{
"epoch": 0.6364976958525346,
"grad_norm": 0.4324146807193756,
"learning_rate": 3.5119529665055506e-06,
"loss": 0.4715,
"step": 1151
},
{
"epoch": 0.6370506912442396,
"grad_norm": 0.3583220839500427,
"learning_rate": 3.502738659831483e-06,
"loss": 0.4666,
"step": 1152
},
{
"epoch": 0.6376036866359447,
"grad_norm": 0.35306817293167114,
"learning_rate": 3.49352993557237e-06,
"loss": 0.4554,
"step": 1153
},
{
"epoch": 0.6381566820276497,
"grad_norm": 0.39276352524757385,
"learning_rate": 3.48432682806218e-06,
"loss": 0.4492,
"step": 1154
},
{
"epoch": 0.6387096774193548,
"grad_norm": 0.3977315425872803,
"learning_rate": 3.4751293716139366e-06,
"loss": 0.4523,
"step": 1155
},
{
"epoch": 0.6392626728110599,
"grad_norm": 0.36172613501548767,
"learning_rate": 3.4659376005195956e-06,
"loss": 0.5053,
"step": 1156
},
{
"epoch": 0.639815668202765,
"grad_norm": 0.34491607546806335,
"learning_rate": 3.4567515490499135e-06,
"loss": 0.4633,
"step": 1157
},
{
"epoch": 0.6403686635944701,
"grad_norm": 0.3682193160057068,
"learning_rate": 3.4475712514543246e-06,
"loss": 0.4376,
"step": 1158
},
{
"epoch": 0.6409216589861751,
"grad_norm": 0.41406625509262085,
"learning_rate": 3.438396741960806e-06,
"loss": 0.4669,
"step": 1159
},
{
"epoch": 0.6414746543778802,
"grad_norm": 0.38740435242652893,
"learning_rate": 3.4292280547757586e-06,
"loss": 0.4516,
"step": 1160
},
{
"epoch": 0.6420276497695853,
"grad_norm": 0.35590702295303345,
"learning_rate": 3.420065224083873e-06,
"loss": 0.4406,
"step": 1161
},
{
"epoch": 0.6425806451612903,
"grad_norm": 0.3844543397426605,
"learning_rate": 3.4109082840480045e-06,
"loss": 0.4675,
"step": 1162
},
{
"epoch": 0.6431336405529954,
"grad_norm": 0.40717223286628723,
"learning_rate": 3.4017572688090467e-06,
"loss": 0.4862,
"step": 1163
},
{
"epoch": 0.6436866359447004,
"grad_norm": 0.37645992636680603,
"learning_rate": 3.392612212485802e-06,
"loss": 0.4813,
"step": 1164
},
{
"epoch": 0.6442396313364055,
"grad_norm": 0.33362942934036255,
"learning_rate": 3.3834731491748558e-06,
"loss": 0.4576,
"step": 1165
},
{
"epoch": 0.6447926267281106,
"grad_norm": 0.3458590805530548,
"learning_rate": 3.3743401129504496e-06,
"loss": 0.4664,
"step": 1166
},
{
"epoch": 0.6453456221198157,
"grad_norm": 0.39395076036453247,
"learning_rate": 3.3652131378643515e-06,
"loss": 0.4801,
"step": 1167
},
{
"epoch": 0.6458986175115208,
"grad_norm": 0.36739885807037354,
"learning_rate": 3.356092257945733e-06,
"loss": 0.4662,
"step": 1168
},
{
"epoch": 0.6464516129032258,
"grad_norm": 0.35936224460601807,
"learning_rate": 3.3469775072010417e-06,
"loss": 0.4678,
"step": 1169
},
{
"epoch": 0.6470046082949309,
"grad_norm": 0.35847219824790955,
"learning_rate": 3.337868919613869e-06,
"loss": 0.4595,
"step": 1170
},
{
"epoch": 0.647557603686636,
"grad_norm": 0.33475178480148315,
"learning_rate": 3.3287665291448323e-06,
"loss": 0.4631,
"step": 1171
},
{
"epoch": 0.648110599078341,
"grad_norm": 0.3567461669445038,
"learning_rate": 3.319670369731437e-06,
"loss": 0.475,
"step": 1172
},
{
"epoch": 0.6486635944700461,
"grad_norm": 0.4065379798412323,
"learning_rate": 3.310580475287963e-06,
"loss": 0.4596,
"step": 1173
},
{
"epoch": 0.6492165898617511,
"grad_norm": 0.34394586086273193,
"learning_rate": 3.3014968797053263e-06,
"loss": 0.4717,
"step": 1174
},
{
"epoch": 0.6497695852534562,
"grad_norm": 0.3613266944885254,
"learning_rate": 3.2924196168509657e-06,
"loss": 0.4933,
"step": 1175
},
{
"epoch": 0.6503225806451612,
"grad_norm": 0.38999873399734497,
"learning_rate": 3.283348720568702e-06,
"loss": 0.4756,
"step": 1176
},
{
"epoch": 0.6508755760368664,
"grad_norm": 0.3532698154449463,
"learning_rate": 3.274284224678621e-06,
"loss": 0.4575,
"step": 1177
},
{
"epoch": 0.6514285714285715,
"grad_norm": 0.33012717962265015,
"learning_rate": 3.2652261629769457e-06,
"loss": 0.4473,
"step": 1178
},
{
"epoch": 0.6519815668202765,
"grad_norm": 0.3971370458602905,
"learning_rate": 3.256174569235909e-06,
"loss": 0.5044,
"step": 1179
},
{
"epoch": 0.6525345622119816,
"grad_norm": 0.3590445816516876,
"learning_rate": 3.2471294772036287e-06,
"loss": 0.4675,
"step": 1180
},
{
"epoch": 0.6530875576036866,
"grad_norm": 0.3668496012687683,
"learning_rate": 3.238090920603981e-06,
"loss": 0.4419,
"step": 1181
},
{
"epoch": 0.6536405529953917,
"grad_norm": 0.3250981867313385,
"learning_rate": 3.2290589331364787e-06,
"loss": 0.4597,
"step": 1182
},
{
"epoch": 0.6541935483870968,
"grad_norm": 0.38982975482940674,
"learning_rate": 3.2200335484761352e-06,
"loss": 0.4638,
"step": 1183
},
{
"epoch": 0.6547465437788018,
"grad_norm": 0.38048577308654785,
"learning_rate": 3.2110148002733534e-06,
"loss": 0.473,
"step": 1184
},
{
"epoch": 0.6552995391705069,
"grad_norm": 0.35395413637161255,
"learning_rate": 3.2020027221537864e-06,
"loss": 0.4832,
"step": 1185
},
{
"epoch": 0.6558525345622119,
"grad_norm": 0.3707823157310486,
"learning_rate": 3.192997347718224e-06,
"loss": 0.4958,
"step": 1186
},
{
"epoch": 0.6564055299539171,
"grad_norm": 0.36011847853660583,
"learning_rate": 3.1839987105424586e-06,
"loss": 0.472,
"step": 1187
},
{
"epoch": 0.6569585253456222,
"grad_norm": 0.3693712055683136,
"learning_rate": 3.1750068441771637e-06,
"loss": 0.4649,
"step": 1188
},
{
"epoch": 0.6575115207373272,
"grad_norm": 0.34001612663269043,
"learning_rate": 3.1660217821477686e-06,
"loss": 0.4764,
"step": 1189
},
{
"epoch": 0.6580645161290323,
"grad_norm": 0.332956999540329,
"learning_rate": 3.1570435579543333e-06,
"loss": 0.4652,
"step": 1190
},
{
"epoch": 0.6586175115207373,
"grad_norm": 0.35701560974121094,
"learning_rate": 3.148072205071423e-06,
"loss": 0.4695,
"step": 1191
},
{
"epoch": 0.6591705069124424,
"grad_norm": 0.3385053277015686,
"learning_rate": 3.1391077569479856e-06,
"loss": 0.4524,
"step": 1192
},
{
"epoch": 0.6597235023041474,
"grad_norm": 0.35991233587265015,
"learning_rate": 3.130150247007222e-06,
"loss": 0.4322,
"step": 1193
},
{
"epoch": 0.6602764976958525,
"grad_norm": 0.36623886227607727,
"learning_rate": 3.1211997086464683e-06,
"loss": 0.4754,
"step": 1194
},
{
"epoch": 0.6608294930875576,
"grad_norm": 0.33661210536956787,
"learning_rate": 3.112256175237064e-06,
"loss": 0.464,
"step": 1195
},
{
"epoch": 0.6613824884792626,
"grad_norm": 0.3260611295700073,
"learning_rate": 3.103319680124235e-06,
"loss": 0.4608,
"step": 1196
},
{
"epoch": 0.6619354838709678,
"grad_norm": 0.38113391399383545,
"learning_rate": 3.0943902566269613e-06,
"loss": 0.4963,
"step": 1197
},
{
"epoch": 0.6624884792626728,
"grad_norm": 0.3196834623813629,
"learning_rate": 3.085467938037861e-06,
"loss": 0.4631,
"step": 1198
},
{
"epoch": 0.6630414746543779,
"grad_norm": 0.3713325262069702,
"learning_rate": 3.076552757623059e-06,
"loss": 0.4559,
"step": 1199
},
{
"epoch": 0.663594470046083,
"grad_norm": 0.35076281428337097,
"learning_rate": 3.0676447486220705e-06,
"loss": 0.4601,
"step": 1200
},
{
"epoch": 0.664147465437788,
"grad_norm": 0.34032174944877625,
"learning_rate": 3.058743944247665e-06,
"loss": 0.4702,
"step": 1201
},
{
"epoch": 0.6647004608294931,
"grad_norm": 0.37965843081474304,
"learning_rate": 3.0498503776857576e-06,
"loss": 0.4806,
"step": 1202
},
{
"epoch": 0.6652534562211981,
"grad_norm": 0.3258879780769348,
"learning_rate": 3.0409640820952735e-06,
"loss": 0.4331,
"step": 1203
},
{
"epoch": 0.6658064516129032,
"grad_norm": 0.3517114222049713,
"learning_rate": 3.0320850906080325e-06,
"loss": 0.4521,
"step": 1204
},
{
"epoch": 0.6663594470046083,
"grad_norm": 0.3746355473995209,
"learning_rate": 3.0232134363286163e-06,
"loss": 0.4702,
"step": 1205
},
{
"epoch": 0.6669124423963133,
"grad_norm": 0.3362243175506592,
"learning_rate": 3.0143491523342562e-06,
"loss": 0.4382,
"step": 1206
},
{
"epoch": 0.6674654377880185,
"grad_norm": 0.3588075041770935,
"learning_rate": 3.005492271674697e-06,
"loss": 0.452,
"step": 1207
},
{
"epoch": 0.6680184331797235,
"grad_norm": 0.4125935137271881,
"learning_rate": 2.9966428273720904e-06,
"loss": 0.4715,
"step": 1208
},
{
"epoch": 0.6685714285714286,
"grad_norm": 0.34157875180244446,
"learning_rate": 2.987800852420855e-06,
"loss": 0.4543,
"step": 1209
},
{
"epoch": 0.6691244239631337,
"grad_norm": 0.34906935691833496,
"learning_rate": 2.9789663797875614e-06,
"loss": 0.4535,
"step": 1210
},
{
"epoch": 0.6696774193548387,
"grad_norm": 0.3961377739906311,
"learning_rate": 2.970139442410811e-06,
"loss": 0.4992,
"step": 1211
},
{
"epoch": 0.6702304147465438,
"grad_norm": 0.3486752510070801,
"learning_rate": 2.961320073201107e-06,
"loss": 0.4603,
"step": 1212
},
{
"epoch": 0.6707834101382488,
"grad_norm": 0.3723919093608856,
"learning_rate": 2.952508305040739e-06,
"loss": 0.4735,
"step": 1213
},
{
"epoch": 0.6713364055299539,
"grad_norm": 0.33436325192451477,
"learning_rate": 2.9437041707836532e-06,
"loss": 0.4558,
"step": 1214
},
{
"epoch": 0.6718894009216589,
"grad_norm": 0.3649885058403015,
"learning_rate": 2.9349077032553387e-06,
"loss": 0.4734,
"step": 1215
},
{
"epoch": 0.6724423963133641,
"grad_norm": 0.4229625463485718,
"learning_rate": 2.9261189352526932e-06,
"loss": 0.4734,
"step": 1216
},
{
"epoch": 0.6729953917050692,
"grad_norm": 0.31533482670783997,
"learning_rate": 2.9173378995439107e-06,
"loss": 0.4596,
"step": 1217
},
{
"epoch": 0.6735483870967742,
"grad_norm": 0.39130347967147827,
"learning_rate": 2.9085646288683587e-06,
"loss": 0.4913,
"step": 1218
},
{
"epoch": 0.6741013824884793,
"grad_norm": 0.3557525873184204,
"learning_rate": 2.8997991559364493e-06,
"loss": 0.4708,
"step": 1219
},
{
"epoch": 0.6746543778801843,
"grad_norm": 0.36776086688041687,
"learning_rate": 2.8910415134295216e-06,
"loss": 0.4685,
"step": 1220
},
{
"epoch": 0.6752073732718894,
"grad_norm": 0.3936042785644531,
"learning_rate": 2.8822917339997237e-06,
"loss": 0.4667,
"step": 1221
},
{
"epoch": 0.6757603686635945,
"grad_norm": 0.3256804645061493,
"learning_rate": 2.87354985026988e-06,
"loss": 0.428,
"step": 1222
},
{
"epoch": 0.6763133640552995,
"grad_norm": 0.3750945031642914,
"learning_rate": 2.8648158948333817e-06,
"loss": 0.458,
"step": 1223
},
{
"epoch": 0.6768663594470046,
"grad_norm": 0.3525260090827942,
"learning_rate": 2.856089900254059e-06,
"loss": 0.4469,
"step": 1224
},
{
"epoch": 0.6774193548387096,
"grad_norm": 0.32836800813674927,
"learning_rate": 2.847371899066059e-06,
"loss": 0.4581,
"step": 1225
},
{
"epoch": 0.6779723502304148,
"grad_norm": 0.34020939469337463,
"learning_rate": 2.83866192377373e-06,
"loss": 0.444,
"step": 1226
},
{
"epoch": 0.6785253456221199,
"grad_norm": 0.31056728959083557,
"learning_rate": 2.829960006851492e-06,
"loss": 0.4748,
"step": 1227
},
{
"epoch": 0.6790783410138249,
"grad_norm": 0.3614257574081421,
"learning_rate": 2.8212661807437226e-06,
"loss": 0.487,
"step": 1228
},
{
"epoch": 0.67963133640553,
"grad_norm": 0.33884182572364807,
"learning_rate": 2.8125804778646315e-06,
"loss": 0.4871,
"step": 1229
},
{
"epoch": 0.680184331797235,
"grad_norm": 0.3698228597640991,
"learning_rate": 2.803902930598144e-06,
"loss": 0.4783,
"step": 1230
},
{
"epoch": 0.6807373271889401,
"grad_norm": 0.3055351674556732,
"learning_rate": 2.7952335712977764e-06,
"loss": 0.4488,
"step": 1231
},
{
"epoch": 0.6812903225806451,
"grad_norm": 0.3399738073348999,
"learning_rate": 2.7865724322865174e-06,
"loss": 0.4489,
"step": 1232
},
{
"epoch": 0.6818433179723502,
"grad_norm": 0.3493901193141937,
"learning_rate": 2.7779195458567067e-06,
"loss": 0.4508,
"step": 1233
},
{
"epoch": 0.6823963133640553,
"grad_norm": 0.38333308696746826,
"learning_rate": 2.7692749442699147e-06,
"loss": 0.4733,
"step": 1234
},
{
"epoch": 0.6829493087557603,
"grad_norm": 0.33638396859169006,
"learning_rate": 2.7606386597568223e-06,
"loss": 0.4521,
"step": 1235
},
{
"epoch": 0.6835023041474655,
"grad_norm": 0.33842137455940247,
"learning_rate": 2.7520107245171035e-06,
"loss": 0.4359,
"step": 1236
},
{
"epoch": 0.6840552995391705,
"grad_norm": 0.34261924028396606,
"learning_rate": 2.7433911707192976e-06,
"loss": 0.4713,
"step": 1237
},
{
"epoch": 0.6846082949308756,
"grad_norm": 0.33021679520606995,
"learning_rate": 2.7347800305007e-06,
"loss": 0.4563,
"step": 1238
},
{
"epoch": 0.6851612903225807,
"grad_norm": 0.33682435750961304,
"learning_rate": 2.7261773359672306e-06,
"loss": 0.4429,
"step": 1239
},
{
"epoch": 0.6857142857142857,
"grad_norm": 0.3547312319278717,
"learning_rate": 2.7175831191933275e-06,
"loss": 0.4846,
"step": 1240
},
{
"epoch": 0.6862672811059908,
"grad_norm": 0.3378848433494568,
"learning_rate": 2.7089974122218165e-06,
"loss": 0.4681,
"step": 1241
},
{
"epoch": 0.6868202764976958,
"grad_norm": 0.3516218066215515,
"learning_rate": 2.700420247063793e-06,
"loss": 0.4796,
"step": 1242
},
{
"epoch": 0.6873732718894009,
"grad_norm": 0.3513849973678589,
"learning_rate": 2.6918516556985088e-06,
"loss": 0.4732,
"step": 1243
},
{
"epoch": 0.687926267281106,
"grad_norm": 0.37055304646492004,
"learning_rate": 2.6832916700732466e-06,
"loss": 0.492,
"step": 1244
},
{
"epoch": 0.688479262672811,
"grad_norm": 0.3224850594997406,
"learning_rate": 2.674740322103204e-06,
"loss": 0.4618,
"step": 1245
},
{
"epoch": 0.6890322580645162,
"grad_norm": 0.3502301573753357,
"learning_rate": 2.666197643671375e-06,
"loss": 0.4562,
"step": 1246
},
{
"epoch": 0.6895852534562212,
"grad_norm": 0.37586522102355957,
"learning_rate": 2.657663666628428e-06,
"loss": 0.448,
"step": 1247
},
{
"epoch": 0.6901382488479263,
"grad_norm": 0.38795405626296997,
"learning_rate": 2.6491384227925903e-06,
"loss": 0.472,
"step": 1248
},
{
"epoch": 0.6906912442396314,
"grad_norm": 0.3674571216106415,
"learning_rate": 2.640621943949527e-06,
"loss": 0.4707,
"step": 1249
},
{
"epoch": 0.6912442396313364,
"grad_norm": 0.36492788791656494,
"learning_rate": 2.6321142618522288e-06,
"loss": 0.4644,
"step": 1250
},
{
"epoch": 0.6917972350230415,
"grad_norm": 0.345634400844574,
"learning_rate": 2.6236154082208776e-06,
"loss": 0.4556,
"step": 1251
},
{
"epoch": 0.6923502304147465,
"grad_norm": 0.3678940534591675,
"learning_rate": 2.6151254147427486e-06,
"loss": 0.4983,
"step": 1252
},
{
"epoch": 0.6929032258064516,
"grad_norm": 0.34437957406044006,
"learning_rate": 2.6066443130720786e-06,
"loss": 0.4613,
"step": 1253
},
{
"epoch": 0.6934562211981566,
"grad_norm": 0.3727521300315857,
"learning_rate": 2.598172134829955e-06,
"loss": 0.4771,
"step": 1254
},
{
"epoch": 0.6940092165898617,
"grad_norm": 0.3770442008972168,
"learning_rate": 2.5897089116041918e-06,
"loss": 0.4627,
"step": 1255
},
{
"epoch": 0.6945622119815669,
"grad_norm": 0.3384653329849243,
"learning_rate": 2.5812546749492163e-06,
"loss": 0.4494,
"step": 1256
},
{
"epoch": 0.6951152073732719,
"grad_norm": 0.3460019528865814,
"learning_rate": 2.5728094563859495e-06,
"loss": 0.4726,
"step": 1257
},
{
"epoch": 0.695668202764977,
"grad_norm": 0.328002005815506,
"learning_rate": 2.5643732874016903e-06,
"loss": 0.4437,
"step": 1258
},
{
"epoch": 0.696221198156682,
"grad_norm": 0.40374118089675903,
"learning_rate": 2.5559461994499968e-06,
"loss": 0.4766,
"step": 1259
},
{
"epoch": 0.6967741935483871,
"grad_norm": 0.3571721017360687,
"learning_rate": 2.5475282239505685e-06,
"loss": 0.459,
"step": 1260
},
{
"epoch": 0.6973271889400922,
"grad_norm": 0.42431801557540894,
"learning_rate": 2.5391193922891288e-06,
"loss": 0.4641,
"step": 1261
},
{
"epoch": 0.6978801843317972,
"grad_norm": 0.35599109530448914,
"learning_rate": 2.5307197358173126e-06,
"loss": 0.4601,
"step": 1262
},
{
"epoch": 0.6984331797235023,
"grad_norm": 0.3282775282859802,
"learning_rate": 2.5223292858525423e-06,
"loss": 0.4352,
"step": 1263
},
{
"epoch": 0.6989861751152073,
"grad_norm": 0.3075070381164551,
"learning_rate": 2.5139480736779164e-06,
"loss": 0.4373,
"step": 1264
},
{
"epoch": 0.6995391705069124,
"grad_norm": 0.3571772277355194,
"learning_rate": 2.5055761305420907e-06,
"loss": 0.4938,
"step": 1265
},
{
"epoch": 0.7000921658986176,
"grad_norm": 0.3431309461593628,
"learning_rate": 2.4972134876591618e-06,
"loss": 0.4413,
"step": 1266
},
{
"epoch": 0.7006451612903226,
"grad_norm": 0.3382079601287842,
"learning_rate": 2.4888601762085518e-06,
"loss": 0.502,
"step": 1267
},
{
"epoch": 0.7011981566820277,
"grad_norm": 0.31066352128982544,
"learning_rate": 2.4805162273348905e-06,
"loss": 0.4596,
"step": 1268
},
{
"epoch": 0.7017511520737327,
"grad_norm": 0.31966033577919006,
"learning_rate": 2.4721816721479007e-06,
"loss": 0.4769,
"step": 1269
},
{
"epoch": 0.7023041474654378,
"grad_norm": 0.3357180058956146,
"learning_rate": 2.4638565417222816e-06,
"loss": 0.437,
"step": 1270
},
{
"epoch": 0.7028571428571428,
"grad_norm": 0.335115522146225,
"learning_rate": 2.4555408670975928e-06,
"loss": 0.4498,
"step": 1271
},
{
"epoch": 0.7034101382488479,
"grad_norm": 0.33428287506103516,
"learning_rate": 2.4472346792781366e-06,
"loss": 0.4661,
"step": 1272
},
{
"epoch": 0.703963133640553,
"grad_norm": 0.3672966957092285,
"learning_rate": 2.438938009232851e-06,
"loss": 0.4775,
"step": 1273
},
{
"epoch": 0.704516129032258,
"grad_norm": 0.3439885079860687,
"learning_rate": 2.4306508878951805e-06,
"loss": 0.4611,
"step": 1274
},
{
"epoch": 0.7050691244239631,
"grad_norm": 0.3395749628543854,
"learning_rate": 2.4223733461629716e-06,
"loss": 0.4498,
"step": 1275
},
{
"epoch": 0.7056221198156682,
"grad_norm": 0.35915130376815796,
"learning_rate": 2.4141054148983532e-06,
"loss": 0.4767,
"step": 1276
},
{
"epoch": 0.7061751152073733,
"grad_norm": 0.393310546875,
"learning_rate": 2.4058471249276232e-06,
"loss": 0.4594,
"step": 1277
},
{
"epoch": 0.7067281105990784,
"grad_norm": 0.3339405357837677,
"learning_rate": 2.397598507041132e-06,
"loss": 0.4732,
"step": 1278
},
{
"epoch": 0.7072811059907834,
"grad_norm": 0.35175976157188416,
"learning_rate": 2.3893595919931673e-06,
"loss": 0.4608,
"step": 1279
},
{
"epoch": 0.7078341013824885,
"grad_norm": 0.33776116371154785,
"learning_rate": 2.381130410501845e-06,
"loss": 0.4568,
"step": 1280
},
{
"epoch": 0.7083870967741935,
"grad_norm": 0.34300360083580017,
"learning_rate": 2.3729109932489825e-06,
"loss": 0.4802,
"step": 1281
},
{
"epoch": 0.7089400921658986,
"grad_norm": 0.36740466952323914,
"learning_rate": 2.3647013708799987e-06,
"loss": 0.4755,
"step": 1282
},
{
"epoch": 0.7094930875576037,
"grad_norm": 0.3558812439441681,
"learning_rate": 2.3565015740037915e-06,
"loss": 0.4623,
"step": 1283
},
{
"epoch": 0.7100460829493087,
"grad_norm": 0.4075135886669159,
"learning_rate": 2.3483116331926244e-06,
"loss": 0.4612,
"step": 1284
},
{
"epoch": 0.7105990783410139,
"grad_norm": 0.3450377285480499,
"learning_rate": 2.340131578982013e-06,
"loss": 0.4695,
"step": 1285
},
{
"epoch": 0.7111520737327189,
"grad_norm": 0.3357861042022705,
"learning_rate": 2.3319614418706127e-06,
"loss": 0.4826,
"step": 1286
},
{
"epoch": 0.711705069124424,
"grad_norm": 0.3738630712032318,
"learning_rate": 2.3238012523201035e-06,
"loss": 0.4905,
"step": 1287
},
{
"epoch": 0.712258064516129,
"grad_norm": 0.3660827577114105,
"learning_rate": 2.315651040755077e-06,
"loss": 0.4754,
"step": 1288
},
{
"epoch": 0.7128110599078341,
"grad_norm": 0.3476543426513672,
"learning_rate": 2.3075108375629218e-06,
"loss": 0.4728,
"step": 1289
},
{
"epoch": 0.7133640552995392,
"grad_norm": 0.33124610781669617,
"learning_rate": 2.299380673093712e-06,
"loss": 0.474,
"step": 1290
},
{
"epoch": 0.7139170506912442,
"grad_norm": 0.33316168189048767,
"learning_rate": 2.291260577660092e-06,
"loss": 0.4853,
"step": 1291
},
{
"epoch": 0.7144700460829493,
"grad_norm": 0.3274078369140625,
"learning_rate": 2.2831505815371656e-06,
"loss": 0.4571,
"step": 1292
},
{
"epoch": 0.7150230414746543,
"grad_norm": 0.3307998776435852,
"learning_rate": 2.275050714962383e-06,
"loss": 0.4498,
"step": 1293
},
{
"epoch": 0.7155760368663594,
"grad_norm": 0.3073872923851013,
"learning_rate": 2.2669610081354254e-06,
"loss": 0.4661,
"step": 1294
},
{
"epoch": 0.7161290322580646,
"grad_norm": 0.37436577677726746,
"learning_rate": 2.2588814912180944e-06,
"loss": 0.4746,
"step": 1295
},
{
"epoch": 0.7166820276497696,
"grad_norm": 0.3833792507648468,
"learning_rate": 2.2508121943342008e-06,
"loss": 0.4502,
"step": 1296
},
{
"epoch": 0.7172350230414747,
"grad_norm": 0.3415040075778961,
"learning_rate": 2.2427531475694496e-06,
"loss": 0.4652,
"step": 1297
},
{
"epoch": 0.7177880184331797,
"grad_norm": 0.3346996605396271,
"learning_rate": 2.2347043809713306e-06,
"loss": 0.4474,
"step": 1298
},
{
"epoch": 0.7183410138248848,
"grad_norm": 0.34259018301963806,
"learning_rate": 2.2266659245490017e-06,
"loss": 0.4832,
"step": 1299
},
{
"epoch": 0.7188940092165899,
"grad_norm": 0.3417535424232483,
"learning_rate": 2.218637808273184e-06,
"loss": 0.4674,
"step": 1300
},
{
"epoch": 0.7194470046082949,
"grad_norm": 0.37401384115219116,
"learning_rate": 2.210620062076044e-06,
"loss": 0.4859,
"step": 1301
},
{
"epoch": 0.72,
"grad_norm": 0.33238619565963745,
"learning_rate": 2.2026127158510843e-06,
"loss": 0.4441,
"step": 1302
},
{
"epoch": 0.720552995391705,
"grad_norm": 0.32708191871643066,
"learning_rate": 2.194615799453032e-06,
"loss": 0.4793,
"step": 1303
},
{
"epoch": 0.7211059907834101,
"grad_norm": 0.30514681339263916,
"learning_rate": 2.186629342697727e-06,
"loss": 0.4409,
"step": 1304
},
{
"epoch": 0.7216589861751153,
"grad_norm": 0.35654860734939575,
"learning_rate": 2.1786533753620155e-06,
"loss": 0.5136,
"step": 1305
},
{
"epoch": 0.7222119815668203,
"grad_norm": 0.3565172255039215,
"learning_rate": 2.170687927183629e-06,
"loss": 0.4683,
"step": 1306
},
{
"epoch": 0.7227649769585254,
"grad_norm": 0.3977636396884918,
"learning_rate": 2.1627330278610824e-06,
"loss": 0.4781,
"step": 1307
},
{
"epoch": 0.7233179723502304,
"grad_norm": 0.37379372119903564,
"learning_rate": 2.154788707053559e-06,
"loss": 0.4713,
"step": 1308
},
{
"epoch": 0.7238709677419355,
"grad_norm": 0.3560510575771332,
"learning_rate": 2.1468549943808033e-06,
"loss": 0.4725,
"step": 1309
},
{
"epoch": 0.7244239631336405,
"grad_norm": 0.3701375424861908,
"learning_rate": 2.1389319194230017e-06,
"loss": 0.4887,
"step": 1310
},
{
"epoch": 0.7249769585253456,
"grad_norm": 0.35575488209724426,
"learning_rate": 2.1310195117206863e-06,
"loss": 0.4628,
"step": 1311
},
{
"epoch": 0.7255299539170507,
"grad_norm": 0.34897711873054504,
"learning_rate": 2.1231178007746136e-06,
"loss": 0.443,
"step": 1312
},
{
"epoch": 0.7260829493087557,
"grad_norm": 0.36126595735549927,
"learning_rate": 2.1152268160456584e-06,
"loss": 0.4622,
"step": 1313
},
{
"epoch": 0.7266359447004608,
"grad_norm": 0.3314565420150757,
"learning_rate": 2.1073465869547043e-06,
"loss": 0.4418,
"step": 1314
},
{
"epoch": 0.727188940092166,
"grad_norm": 0.32749220728874207,
"learning_rate": 2.0994771428825332e-06,
"loss": 0.4325,
"step": 1315
},
{
"epoch": 0.727741935483871,
"grad_norm": 0.39038175344467163,
"learning_rate": 2.0916185131697163e-06,
"loss": 0.4928,
"step": 1316
},
{
"epoch": 0.7282949308755761,
"grad_norm": 0.3242311477661133,
"learning_rate": 2.083770727116503e-06,
"loss": 0.4708,
"step": 1317
},
{
"epoch": 0.7288479262672811,
"grad_norm": 0.3813181221485138,
"learning_rate": 2.0759338139827145e-06,
"loss": 0.4959,
"step": 1318
},
{
"epoch": 0.7294009216589862,
"grad_norm": 0.31590506434440613,
"learning_rate": 2.0681078029876322e-06,
"loss": 0.451,
"step": 1319
},
{
"epoch": 0.7299539170506912,
"grad_norm": 0.3604491651058197,
"learning_rate": 2.0602927233098908e-06,
"loss": 0.4817,
"step": 1320
},
{
"epoch": 0.7305069124423963,
"grad_norm": 0.3593979775905609,
"learning_rate": 2.0524886040873676e-06,
"loss": 0.4707,
"step": 1321
},
{
"epoch": 0.7310599078341014,
"grad_norm": 0.38674402236938477,
"learning_rate": 2.0446954744170748e-06,
"loss": 0.4522,
"step": 1322
},
{
"epoch": 0.7316129032258064,
"grad_norm": 0.3926672637462616,
"learning_rate": 2.036913363355052e-06,
"loss": 0.4582,
"step": 1323
},
{
"epoch": 0.7321658986175115,
"grad_norm": 0.3277740776538849,
"learning_rate": 2.029142299916255e-06,
"loss": 0.4913,
"step": 1324
},
{
"epoch": 0.7327188940092166,
"grad_norm": 0.3636924922466278,
"learning_rate": 2.0213823130744516e-06,
"loss": 0.4624,
"step": 1325
},
{
"epoch": 0.7332718894009217,
"grad_norm": 0.36329200863838196,
"learning_rate": 2.0136334317621093e-06,
"loss": 0.4551,
"step": 1326
},
{
"epoch": 0.7338248847926268,
"grad_norm": 0.3718715012073517,
"learning_rate": 2.005895684870291e-06,
"loss": 0.4717,
"step": 1327
},
{
"epoch": 0.7343778801843318,
"grad_norm": 0.3874320685863495,
"learning_rate": 1.9981691012485455e-06,
"loss": 0.4757,
"step": 1328
},
{
"epoch": 0.7349308755760369,
"grad_norm": 0.29239824414253235,
"learning_rate": 1.9904537097048004e-06,
"loss": 0.4554,
"step": 1329
},
{
"epoch": 0.7354838709677419,
"grad_norm": 0.3096185326576233,
"learning_rate": 1.982749539005254e-06,
"loss": 0.4559,
"step": 1330
},
{
"epoch": 0.736036866359447,
"grad_norm": 0.34961435198783875,
"learning_rate": 1.9750566178742697e-06,
"loss": 0.4376,
"step": 1331
},
{
"epoch": 0.736589861751152,
"grad_norm": 0.36215296387672424,
"learning_rate": 1.9673749749942655e-06,
"loss": 0.4544,
"step": 1332
},
{
"epoch": 0.7371428571428571,
"grad_norm": 0.3274637758731842,
"learning_rate": 1.959704639005613e-06,
"loss": 0.4468,
"step": 1333
},
{
"epoch": 0.7376958525345622,
"grad_norm": 0.324142187833786,
"learning_rate": 1.952045638506523e-06,
"loss": 0.4775,
"step": 1334
},
{
"epoch": 0.7382488479262673,
"grad_norm": 0.30846402049064636,
"learning_rate": 1.9443980020529456e-06,
"loss": 0.4507,
"step": 1335
},
{
"epoch": 0.7388018433179724,
"grad_norm": 0.36270537972450256,
"learning_rate": 1.9367617581584606e-06,
"loss": 0.4526,
"step": 1336
},
{
"epoch": 0.7393548387096774,
"grad_norm": 0.35891664028167725,
"learning_rate": 1.9291369352941696e-06,
"loss": 0.4606,
"step": 1337
},
{
"epoch": 0.7399078341013825,
"grad_norm": 0.3416483700275421,
"learning_rate": 1.9215235618885964e-06,
"loss": 0.4976,
"step": 1338
},
{
"epoch": 0.7404608294930876,
"grad_norm": 0.30863887071609497,
"learning_rate": 1.9139216663275727e-06,
"loss": 0.44,
"step": 1339
},
{
"epoch": 0.7410138248847926,
"grad_norm": 0.313449501991272,
"learning_rate": 1.9063312769541348e-06,
"loss": 0.4643,
"step": 1340
},
{
"epoch": 0.7415668202764977,
"grad_norm": 0.35145795345306396,
"learning_rate": 1.8987524220684222e-06,
"loss": 0.4689,
"step": 1341
},
{
"epoch": 0.7421198156682027,
"grad_norm": 0.34510427713394165,
"learning_rate": 1.8911851299275675e-06,
"loss": 0.498,
"step": 1342
},
{
"epoch": 0.7426728110599078,
"grad_norm": 0.3454337418079376,
"learning_rate": 1.8836294287455936e-06,
"loss": 0.4488,
"step": 1343
},
{
"epoch": 0.743225806451613,
"grad_norm": 0.3266771733760834,
"learning_rate": 1.8760853466933072e-06,
"loss": 0.4855,
"step": 1344
},
{
"epoch": 0.743778801843318,
"grad_norm": 0.3165615200996399,
"learning_rate": 1.8685529118981926e-06,
"loss": 0.4556,
"step": 1345
},
{
"epoch": 0.7443317972350231,
"grad_norm": 0.33930885791778564,
"learning_rate": 1.8610321524443099e-06,
"loss": 0.4401,
"step": 1346
},
{
"epoch": 0.7448847926267281,
"grad_norm": 0.35839566588401794,
"learning_rate": 1.8535230963721884e-06,
"loss": 0.4965,
"step": 1347
},
{
"epoch": 0.7454377880184332,
"grad_norm": 0.33815956115722656,
"learning_rate": 1.8460257716787216e-06,
"loss": 0.4868,
"step": 1348
},
{
"epoch": 0.7459907834101382,
"grad_norm": 0.34670326113700867,
"learning_rate": 1.8385402063170643e-06,
"loss": 0.4696,
"step": 1349
},
{
"epoch": 0.7465437788018433,
"grad_norm": 0.32649490237236023,
"learning_rate": 1.8310664281965268e-06,
"loss": 0.4951,
"step": 1350
},
{
"epoch": 0.7470967741935484,
"grad_norm": 0.32630065083503723,
"learning_rate": 1.8236044651824725e-06,
"loss": 0.4733,
"step": 1351
},
{
"epoch": 0.7476497695852534,
"grad_norm": 0.2883811593055725,
"learning_rate": 1.8161543450962127e-06,
"loss": 0.4637,
"step": 1352
},
{
"epoch": 0.7482027649769585,
"grad_norm": 0.34161558747291565,
"learning_rate": 1.8087160957149036e-06,
"loss": 0.4629,
"step": 1353
},
{
"epoch": 0.7487557603686636,
"grad_norm": 0.3212891221046448,
"learning_rate": 1.8012897447714417e-06,
"loss": 0.4703,
"step": 1354
},
{
"epoch": 0.7493087557603687,
"grad_norm": 0.3223973214626312,
"learning_rate": 1.7938753199543618e-06,
"loss": 0.4536,
"step": 1355
},
{
"epoch": 0.7498617511520738,
"grad_norm": 0.3456159830093384,
"learning_rate": 1.7864728489077332e-06,
"loss": 0.4671,
"step": 1356
},
{
"epoch": 0.7504147465437788,
"grad_norm": 0.35950833559036255,
"learning_rate": 1.779082359231057e-06,
"loss": 0.4674,
"step": 1357
},
{
"epoch": 0.7509677419354839,
"grad_norm": 0.33876293897628784,
"learning_rate": 1.7717038784791612e-06,
"loss": 0.4614,
"step": 1358
},
{
"epoch": 0.7515207373271889,
"grad_norm": 0.34674006700515747,
"learning_rate": 1.7643374341621006e-06,
"loss": 0.4434,
"step": 1359
},
{
"epoch": 0.752073732718894,
"grad_norm": 0.3747800290584564,
"learning_rate": 1.7569830537450533e-06,
"loss": 0.4607,
"step": 1360
},
{
"epoch": 0.752626728110599,
"grad_norm": 0.3502337336540222,
"learning_rate": 1.7496407646482182e-06,
"loss": 0.471,
"step": 1361
},
{
"epoch": 0.7531797235023041,
"grad_norm": 0.3303203582763672,
"learning_rate": 1.7423105942467116e-06,
"loss": 0.4801,
"step": 1362
},
{
"epoch": 0.7537327188940092,
"grad_norm": 0.3284555673599243,
"learning_rate": 1.7349925698704673e-06,
"loss": 0.4742,
"step": 1363
},
{
"epoch": 0.7542857142857143,
"grad_norm": 0.3339201509952545,
"learning_rate": 1.7276867188041336e-06,
"loss": 0.4588,
"step": 1364
},
{
"epoch": 0.7548387096774194,
"grad_norm": 0.35038480162620544,
"learning_rate": 1.7203930682869707e-06,
"loss": 0.4921,
"step": 1365
},
{
"epoch": 0.7553917050691245,
"grad_norm": 0.3145638406276703,
"learning_rate": 1.7131116455127518e-06,
"loss": 0.428,
"step": 1366
},
{
"epoch": 0.7559447004608295,
"grad_norm": 0.34270545840263367,
"learning_rate": 1.7058424776296583e-06,
"loss": 0.4599,
"step": 1367
},
{
"epoch": 0.7564976958525346,
"grad_norm": 0.33915436267852783,
"learning_rate": 1.698585591740181e-06,
"loss": 0.4713,
"step": 1368
},
{
"epoch": 0.7570506912442396,
"grad_norm": 0.3200076222419739,
"learning_rate": 1.6913410149010179e-06,
"loss": 0.4469,
"step": 1369
},
{
"epoch": 0.7576036866359447,
"grad_norm": 0.31938573718070984,
"learning_rate": 1.6841087741229745e-06,
"loss": 0.4563,
"step": 1370
},
{
"epoch": 0.7581566820276497,
"grad_norm": 0.3671538233757019,
"learning_rate": 1.6768888963708612e-06,
"loss": 0.471,
"step": 1371
},
{
"epoch": 0.7587096774193548,
"grad_norm": 0.33220696449279785,
"learning_rate": 1.669681408563395e-06,
"loss": 0.444,
"step": 1372
},
{
"epoch": 0.7592626728110599,
"grad_norm": 0.3261774182319641,
"learning_rate": 1.6624863375730977e-06,
"loss": 0.4775,
"step": 1373
},
{
"epoch": 0.759815668202765,
"grad_norm": 0.31591543555259705,
"learning_rate": 1.6553037102261955e-06,
"loss": 0.4763,
"step": 1374
},
{
"epoch": 0.7603686635944701,
"grad_norm": 0.345234215259552,
"learning_rate": 1.6481335533025195e-06,
"loss": 0.443,
"step": 1375
},
{
"epoch": 0.7609216589861751,
"grad_norm": 0.3390832245349884,
"learning_rate": 1.6409758935354065e-06,
"loss": 0.4622,
"step": 1376
},
{
"epoch": 0.7614746543778802,
"grad_norm": 0.34640833735466003,
"learning_rate": 1.6338307576115987e-06,
"loss": 0.4606,
"step": 1377
},
{
"epoch": 0.7620276497695853,
"grad_norm": 0.3271695375442505,
"learning_rate": 1.6266981721711438e-06,
"loss": 0.4354,
"step": 1378
},
{
"epoch": 0.7625806451612903,
"grad_norm": 0.34254854917526245,
"learning_rate": 1.619578163807296e-06,
"loss": 0.4601,
"step": 1379
},
{
"epoch": 0.7631336405529954,
"grad_norm": 0.3350234925746918,
"learning_rate": 1.6124707590664168e-06,
"loss": 0.4652,
"step": 1380
},
{
"epoch": 0.7636866359447004,
"grad_norm": 0.3148898780345917,
"learning_rate": 1.6053759844478768e-06,
"loss": 0.4702,
"step": 1381
},
{
"epoch": 0.7642396313364055,
"grad_norm": 0.3196204602718353,
"learning_rate": 1.5982938664039555e-06,
"loss": 0.464,
"step": 1382
},
{
"epoch": 0.7647926267281105,
"grad_norm": 0.3282592296600342,
"learning_rate": 1.591224431339744e-06,
"loss": 0.4631,
"step": 1383
},
{
"epoch": 0.7653456221198157,
"grad_norm": 0.34297293424606323,
"learning_rate": 1.584167705613046e-06,
"loss": 0.4494,
"step": 1384
},
{
"epoch": 0.7658986175115208,
"grad_norm": 0.34436681866645813,
"learning_rate": 1.5771237155342784e-06,
"loss": 0.4501,
"step": 1385
},
{
"epoch": 0.7664516129032258,
"grad_norm": 0.2950928807258606,
"learning_rate": 1.5700924873663758e-06,
"loss": 0.4493,
"step": 1386
},
{
"epoch": 0.7670046082949309,
"grad_norm": 0.3339753746986389,
"learning_rate": 1.5630740473246896e-06,
"loss": 0.4793,
"step": 1387
},
{
"epoch": 0.767557603686636,
"grad_norm": 0.3539128005504608,
"learning_rate": 1.5560684215768935e-06,
"loss": 0.4459,
"step": 1388
},
{
"epoch": 0.768110599078341,
"grad_norm": 0.35491037368774414,
"learning_rate": 1.549075636242882e-06,
"loss": 0.4814,
"step": 1389
},
{
"epoch": 0.7686635944700461,
"grad_norm": 0.3396543562412262,
"learning_rate": 1.5420957173946772e-06,
"loss": 0.4526,
"step": 1390
},
{
"epoch": 0.7692165898617511,
"grad_norm": 0.35023194551467896,
"learning_rate": 1.5351286910563278e-06,
"loss": 0.432,
"step": 1391
},
{
"epoch": 0.7697695852534562,
"grad_norm": 0.3362998366355896,
"learning_rate": 1.5281745832038159e-06,
"loss": 0.4518,
"step": 1392
},
{
"epoch": 0.7703225806451612,
"grad_norm": 0.3237576186656952,
"learning_rate": 1.5212334197649564e-06,
"loss": 0.4732,
"step": 1393
},
{
"epoch": 0.7708755760368664,
"grad_norm": 0.3349010646343231,
"learning_rate": 1.514305226619302e-06,
"loss": 0.4793,
"step": 1394
},
{
"epoch": 0.7714285714285715,
"grad_norm": 0.3398427367210388,
"learning_rate": 1.5073900295980481e-06,
"loss": 0.471,
"step": 1395
},
{
"epoch": 0.7719815668202765,
"grad_norm": 0.33397191762924194,
"learning_rate": 1.5004878544839335e-06,
"loss": 0.4398,
"step": 1396
},
{
"epoch": 0.7725345622119816,
"grad_norm": 0.3291890025138855,
"learning_rate": 1.493598727011148e-06,
"loss": 0.4632,
"step": 1397
},
{
"epoch": 0.7730875576036866,
"grad_norm": 0.33560484647750854,
"learning_rate": 1.4867226728652318e-06,
"loss": 0.4661,
"step": 1398
},
{
"epoch": 0.7736405529953917,
"grad_norm": 0.32114970684051514,
"learning_rate": 1.4798597176829844e-06,
"loss": 0.4558,
"step": 1399
},
{
"epoch": 0.7741935483870968,
"grad_norm": 0.29689663648605347,
"learning_rate": 1.4730098870523652e-06,
"loss": 0.4567,
"step": 1400
},
{
"epoch": 0.7747465437788018,
"grad_norm": 0.33287614583969116,
"learning_rate": 1.4661732065124012e-06,
"loss": 0.4458,
"step": 1401
},
{
"epoch": 0.7752995391705069,
"grad_norm": 0.32061877846717834,
"learning_rate": 1.45934970155309e-06,
"loss": 0.4482,
"step": 1402
},
{
"epoch": 0.7758525345622119,
"grad_norm": 0.32134944200515747,
"learning_rate": 1.4525393976153046e-06,
"loss": 0.4669,
"step": 1403
},
{
"epoch": 0.7764055299539171,
"grad_norm": 0.31328532099723816,
"learning_rate": 1.4457423200906994e-06,
"loss": 0.474,
"step": 1404
},
{
"epoch": 0.7769585253456222,
"grad_norm": 0.31957975029945374,
"learning_rate": 1.4389584943216156e-06,
"loss": 0.4478,
"step": 1405
},
{
"epoch": 0.7775115207373272,
"grad_norm": 0.318330317735672,
"learning_rate": 1.4321879456009858e-06,
"loss": 0.4738,
"step": 1406
},
{
"epoch": 0.7780645161290323,
"grad_norm": 0.3517070710659027,
"learning_rate": 1.4254306991722406e-06,
"loss": 0.49,
"step": 1407
},
{
"epoch": 0.7786175115207373,
"grad_norm": 0.3362846374511719,
"learning_rate": 1.4186867802292132e-06,
"loss": 0.4756,
"step": 1408
},
{
"epoch": 0.7791705069124424,
"grad_norm": 0.32801997661590576,
"learning_rate": 1.411956213916048e-06,
"loss": 0.4745,
"step": 1409
},
{
"epoch": 0.7797235023041474,
"grad_norm": 0.32685017585754395,
"learning_rate": 1.4052390253271037e-06,
"loss": 0.4618,
"step": 1410
},
{
"epoch": 0.7802764976958525,
"grad_norm": 0.32221168279647827,
"learning_rate": 1.3985352395068618e-06,
"loss": 0.4373,
"step": 1411
},
{
"epoch": 0.7808294930875576,
"grad_norm": 0.3476710915565491,
"learning_rate": 1.3918448814498336e-06,
"loss": 0.4463,
"step": 1412
},
{
"epoch": 0.7813824884792627,
"grad_norm": 0.3333290219306946,
"learning_rate": 1.3851679761004644e-06,
"loss": 0.4592,
"step": 1413
},
{
"epoch": 0.7819354838709678,
"grad_norm": 0.32516762614250183,
"learning_rate": 1.3785045483530435e-06,
"loss": 0.476,
"step": 1414
},
{
"epoch": 0.7824884792626728,
"grad_norm": 0.3284921944141388,
"learning_rate": 1.3718546230516095e-06,
"loss": 0.4669,
"step": 1415
},
{
"epoch": 0.7830414746543779,
"grad_norm": 0.3187192976474762,
"learning_rate": 1.3652182249898583e-06,
"loss": 0.478,
"step": 1416
},
{
"epoch": 0.783594470046083,
"grad_norm": 0.3000369369983673,
"learning_rate": 1.3585953789110506e-06,
"loss": 0.4763,
"step": 1417
},
{
"epoch": 0.784147465437788,
"grad_norm": 0.3406629264354706,
"learning_rate": 1.3519861095079202e-06,
"loss": 0.451,
"step": 1418
},
{
"epoch": 0.7847004608294931,
"grad_norm": 0.3349382281303406,
"learning_rate": 1.3453904414225799e-06,
"loss": 0.4573,
"step": 1419
},
{
"epoch": 0.7852534562211981,
"grad_norm": 0.3069670796394348,
"learning_rate": 1.3388083992464335e-06,
"loss": 0.4739,
"step": 1420
},
{
"epoch": 0.7858064516129032,
"grad_norm": 0.31198862195014954,
"learning_rate": 1.3322400075200792e-06,
"loss": 0.454,
"step": 1421
},
{
"epoch": 0.7863594470046082,
"grad_norm": 0.30845198035240173,
"learning_rate": 1.325685290733223e-06,
"loss": 0.4498,
"step": 1422
},
{
"epoch": 0.7869124423963134,
"grad_norm": 0.31618401408195496,
"learning_rate": 1.3191442733245824e-06,
"loss": 0.491,
"step": 1423
},
{
"epoch": 0.7874654377880185,
"grad_norm": 0.356711208820343,
"learning_rate": 1.3126169796818001e-06,
"loss": 0.4422,
"step": 1424
},
{
"epoch": 0.7880184331797235,
"grad_norm": 0.35151243209838867,
"learning_rate": 1.3061034341413497e-06,
"loss": 0.469,
"step": 1425
},
{
"epoch": 0.7885714285714286,
"grad_norm": 0.30518248677253723,
"learning_rate": 1.2996036609884478e-06,
"loss": 0.4778,
"step": 1426
},
{
"epoch": 0.7891244239631336,
"grad_norm": 0.30732637643814087,
"learning_rate": 1.2931176844569588e-06,
"loss": 0.4627,
"step": 1427
},
{
"epoch": 0.7896774193548387,
"grad_norm": 0.3301125466823578,
"learning_rate": 1.2866455287293094e-06,
"loss": 0.4711,
"step": 1428
},
{
"epoch": 0.7902304147465438,
"grad_norm": 0.3300682008266449,
"learning_rate": 1.2801872179363978e-06,
"loss": 0.497,
"step": 1429
},
{
"epoch": 0.7907834101382488,
"grad_norm": 0.3206273019313812,
"learning_rate": 1.2737427761575006e-06,
"loss": 0.4593,
"step": 1430
},
{
"epoch": 0.7913364055299539,
"grad_norm": 0.35528701543807983,
"learning_rate": 1.2673122274201844e-06,
"loss": 0.452,
"step": 1431
},
{
"epoch": 0.7918894009216589,
"grad_norm": 0.3094896972179413,
"learning_rate": 1.2608955957002196e-06,
"loss": 0.4426,
"step": 1432
},
{
"epoch": 0.7924423963133641,
"grad_norm": 0.3296172618865967,
"learning_rate": 1.2544929049214843e-06,
"loss": 0.4751,
"step": 1433
},
{
"epoch": 0.7929953917050692,
"grad_norm": 0.3449612855911255,
"learning_rate": 1.248104178955883e-06,
"loss": 0.473,
"step": 1434
},
{
"epoch": 0.7935483870967742,
"grad_norm": 0.30491313338279724,
"learning_rate": 1.2417294416232505e-06,
"loss": 0.447,
"step": 1435
},
{
"epoch": 0.7941013824884793,
"grad_norm": 0.32028716802597046,
"learning_rate": 1.235368716691267e-06,
"loss": 0.4358,
"step": 1436
},
{
"epoch": 0.7946543778801843,
"grad_norm": 0.3335442543029785,
"learning_rate": 1.22902202787537e-06,
"loss": 0.476,
"step": 1437
},
{
"epoch": 0.7952073732718894,
"grad_norm": 0.3434390723705292,
"learning_rate": 1.2226893988386618e-06,
"loss": 0.4447,
"step": 1438
},
{
"epoch": 0.7957603686635945,
"grad_norm": 0.3305012881755829,
"learning_rate": 1.2163708531918267e-06,
"loss": 0.4644,
"step": 1439
},
{
"epoch": 0.7963133640552995,
"grad_norm": 0.32044875621795654,
"learning_rate": 1.210066414493039e-06,
"loss": 0.4525,
"step": 1440
},
{
"epoch": 0.7968663594470046,
"grad_norm": 0.3219109773635864,
"learning_rate": 1.2037761062478759e-06,
"loss": 0.4614,
"step": 1441
},
{
"epoch": 0.7974193548387096,
"grad_norm": 0.3471842408180237,
"learning_rate": 1.1974999519092311e-06,
"loss": 0.4541,
"step": 1442
},
{
"epoch": 0.7979723502304148,
"grad_norm": 0.2974897623062134,
"learning_rate": 1.1912379748772267e-06,
"loss": 0.4344,
"step": 1443
},
{
"epoch": 0.7985253456221199,
"grad_norm": 0.31679412722587585,
"learning_rate": 1.1849901984991253e-06,
"loss": 0.456,
"step": 1444
},
{
"epoch": 0.7990783410138249,
"grad_norm": 0.30493679642677307,
"learning_rate": 1.1787566460692446e-06,
"loss": 0.4674,
"step": 1445
},
{
"epoch": 0.79963133640553,
"grad_norm": 0.3283561170101166,
"learning_rate": 1.1725373408288682e-06,
"loss": 0.4621,
"step": 1446
},
{
"epoch": 0.800184331797235,
"grad_norm": 0.32548418641090393,
"learning_rate": 1.1663323059661609e-06,
"loss": 0.4422,
"step": 1447
},
{
"epoch": 0.8007373271889401,
"grad_norm": 0.29554906487464905,
"learning_rate": 1.1601415646160813e-06,
"loss": 0.4535,
"step": 1448
},
{
"epoch": 0.8012903225806451,
"grad_norm": 0.31748977303504944,
"learning_rate": 1.153965139860297e-06,
"loss": 0.4579,
"step": 1449
},
{
"epoch": 0.8018433179723502,
"grad_norm": 0.30266064405441284,
"learning_rate": 1.147803054727095e-06,
"loss": 0.446,
"step": 1450
},
{
"epoch": 0.8023963133640553,
"grad_norm": 0.32854562997817993,
"learning_rate": 1.1416553321913009e-06,
"loss": 0.4632,
"step": 1451
},
{
"epoch": 0.8029493087557603,
"grad_norm": 0.29594528675079346,
"learning_rate": 1.1355219951741881e-06,
"loss": 0.4704,
"step": 1452
},
{
"epoch": 0.8035023041474655,
"grad_norm": 0.2916601300239563,
"learning_rate": 1.1294030665433969e-06,
"loss": 0.4287,
"step": 1453
},
{
"epoch": 0.8040552995391705,
"grad_norm": 0.31938451528549194,
"learning_rate": 1.1232985691128457e-06,
"loss": 0.4242,
"step": 1454
},
{
"epoch": 0.8046082949308756,
"grad_norm": 0.3538084030151367,
"learning_rate": 1.1172085256426473e-06,
"loss": 0.4705,
"step": 1455
},
{
"epoch": 0.8051612903225807,
"grad_norm": 0.3275561034679413,
"learning_rate": 1.1111329588390253e-06,
"loss": 0.4745,
"step": 1456
},
{
"epoch": 0.8057142857142857,
"grad_norm": 0.30866336822509766,
"learning_rate": 1.1050718913542275e-06,
"loss": 0.453,
"step": 1457
},
{
"epoch": 0.8062672811059908,
"grad_norm": 0.3329750597476959,
"learning_rate": 1.0990253457864418e-06,
"loss": 0.4598,
"step": 1458
},
{
"epoch": 0.8068202764976958,
"grad_norm": 0.32834047079086304,
"learning_rate": 1.0929933446797136e-06,
"loss": 0.4649,
"step": 1459
},
{
"epoch": 0.8073732718894009,
"grad_norm": 0.3380810022354126,
"learning_rate": 1.0869759105238592e-06,
"loss": 0.4605,
"step": 1460
},
{
"epoch": 0.807926267281106,
"grad_norm": 0.3055575489997864,
"learning_rate": 1.0809730657543838e-06,
"loss": 0.4852,
"step": 1461
},
{
"epoch": 0.808479262672811,
"grad_norm": 0.30661699175834656,
"learning_rate": 1.0749848327523966e-06,
"loss": 0.4824,
"step": 1462
},
{
"epoch": 0.8090322580645162,
"grad_norm": 0.32859787344932556,
"learning_rate": 1.0690112338445292e-06,
"loss": 0.464,
"step": 1463
},
{
"epoch": 0.8095852534562212,
"grad_norm": 0.3309606611728668,
"learning_rate": 1.0630522913028508e-06,
"loss": 0.4585,
"step": 1464
},
{
"epoch": 0.8101382488479263,
"grad_norm": 0.33870929479599,
"learning_rate": 1.0571080273447858e-06,
"loss": 0.476,
"step": 1465
},
{
"epoch": 0.8106912442396313,
"grad_norm": 0.33712446689605713,
"learning_rate": 1.0511784641330286e-06,
"loss": 0.4476,
"step": 1466
},
{
"epoch": 0.8112442396313364,
"grad_norm": 0.2965930700302124,
"learning_rate": 1.0452636237754642e-06,
"loss": 0.4766,
"step": 1467
},
{
"epoch": 0.8117972350230415,
"grad_norm": 0.3398984372615814,
"learning_rate": 1.0393635283250885e-06,
"loss": 0.4716,
"step": 1468
},
{
"epoch": 0.8123502304147465,
"grad_norm": 0.3270639181137085,
"learning_rate": 1.0334781997799165e-06,
"loss": 0.4556,
"step": 1469
},
{
"epoch": 0.8129032258064516,
"grad_norm": 0.3468005955219269,
"learning_rate": 1.0276076600829094e-06,
"loss": 0.4725,
"step": 1470
},
{
"epoch": 0.8134562211981566,
"grad_norm": 0.31104719638824463,
"learning_rate": 1.0217519311218882e-06,
"loss": 0.4735,
"step": 1471
},
{
"epoch": 0.8140092165898618,
"grad_norm": 0.30012935400009155,
"learning_rate": 1.0159110347294536e-06,
"loss": 0.4624,
"step": 1472
},
{
"epoch": 0.8145622119815669,
"grad_norm": 0.3302370309829712,
"learning_rate": 1.0100849926829043e-06,
"loss": 0.4755,
"step": 1473
},
{
"epoch": 0.8151152073732719,
"grad_norm": 0.3273729681968689,
"learning_rate": 1.0042738267041552e-06,
"loss": 0.4558,
"step": 1474
},
{
"epoch": 0.815668202764977,
"grad_norm": 0.313931405544281,
"learning_rate": 9.984775584596585e-07,
"loss": 0.4778,
"step": 1475
},
{
"epoch": 0.816221198156682,
"grad_norm": 0.31616419553756714,
"learning_rate": 9.926962095603199e-07,
"loss": 0.4358,
"step": 1476
},
{
"epoch": 0.8167741935483871,
"grad_norm": 0.33137091994285583,
"learning_rate": 9.869298015614198e-07,
"loss": 0.4558,
"step": 1477
},
{
"epoch": 0.8173271889400922,
"grad_norm": 0.3267715573310852,
"learning_rate": 9.811783559625343e-07,
"loss": 0.475,
"step": 1478
},
{
"epoch": 0.8178801843317972,
"grad_norm": 0.37252572178840637,
"learning_rate": 9.754418942074512e-07,
"loss": 0.4829,
"step": 1479
},
{
"epoch": 0.8184331797235023,
"grad_norm": 0.3060474395751953,
"learning_rate": 9.697204376840936e-07,
"loss": 0.4627,
"step": 1480
},
{
"epoch": 0.8189861751152073,
"grad_norm": 0.32007962465286255,
"learning_rate": 9.640140077244382e-07,
"loss": 0.4698,
"step": 1481
},
{
"epoch": 0.8195391705069125,
"grad_norm": 0.2968813478946686,
"learning_rate": 9.58322625604437e-07,
"loss": 0.4316,
"step": 1482
},
{
"epoch": 0.8200921658986176,
"grad_norm": 0.31622204184532166,
"learning_rate": 9.526463125439372e-07,
"loss": 0.4408,
"step": 1483
},
{
"epoch": 0.8206451612903226,
"grad_norm": 0.2871773838996887,
"learning_rate": 9.469850897066014e-07,
"loss": 0.4563,
"step": 1484
},
{
"epoch": 0.8211981566820277,
"grad_norm": 0.30365270376205444,
"learning_rate": 9.413389781998311e-07,
"loss": 0.4519,
"step": 1485
},
{
"epoch": 0.8217511520737327,
"grad_norm": 0.3301312029361725,
"learning_rate": 9.357079990746853e-07,
"loss": 0.4543,
"step": 1486
},
{
"epoch": 0.8223041474654378,
"grad_norm": 0.3503333628177643,
"learning_rate": 9.300921733258039e-07,
"loss": 0.4753,
"step": 1487
},
{
"epoch": 0.8228571428571428,
"grad_norm": 0.30572423338890076,
"learning_rate": 9.244915218913275e-07,
"loss": 0.4694,
"step": 1488
},
{
"epoch": 0.8234101382488479,
"grad_norm": 0.30958428978919983,
"learning_rate": 9.18906065652822e-07,
"loss": 0.4544,
"step": 1489
},
{
"epoch": 0.823963133640553,
"grad_norm": 0.3019687831401825,
"learning_rate": 9.133358254351982e-07,
"loss": 0.4751,
"step": 1490
},
{
"epoch": 0.824516129032258,
"grad_norm": 0.3120213747024536,
"learning_rate": 9.077808220066359e-07,
"loss": 0.4586,
"step": 1491
},
{
"epoch": 0.8250691244239632,
"grad_norm": 0.3264663815498352,
"learning_rate": 9.022410760785055e-07,
"loss": 0.4569,
"step": 1492
},
{
"epoch": 0.8256221198156682,
"grad_norm": 0.3079594075679779,
"learning_rate": 8.967166083052908e-07,
"loss": 0.5003,
"step": 1493
},
{
"epoch": 0.8261751152073733,
"grad_norm": 0.30771124362945557,
"learning_rate": 8.912074392845144e-07,
"loss": 0.4654,
"step": 1494
},
{
"epoch": 0.8267281105990784,
"grad_norm": 0.31957051157951355,
"learning_rate": 8.857135895566538e-07,
"loss": 0.4722,
"step": 1495
},
{
"epoch": 0.8272811059907834,
"grad_norm": 0.30836620926856995,
"learning_rate": 8.80235079605074e-07,
"loss": 0.4557,
"step": 1496
},
{
"epoch": 0.8278341013824885,
"grad_norm": 0.30665266513824463,
"learning_rate": 8.747719298559465e-07,
"loss": 0.4516,
"step": 1497
},
{
"epoch": 0.8283870967741935,
"grad_norm": 0.33325862884521484,
"learning_rate": 8.693241606781728e-07,
"loss": 0.4667,
"step": 1498
},
{
"epoch": 0.8289400921658986,
"grad_norm": 0.3295386731624603,
"learning_rate": 8.638917923833074e-07,
"loss": 0.4602,
"step": 1499
},
{
"epoch": 0.8294930875576036,
"grad_norm": 0.3496408462524414,
"learning_rate": 8.584748452254888e-07,
"loss": 0.4601,
"step": 1500
},
{
"epoch": 0.8300460829493087,
"grad_norm": 0.2922409474849701,
"learning_rate": 8.530733394013546e-07,
"loss": 0.4787,
"step": 1501
},
{
"epoch": 0.8305990783410139,
"grad_norm": 0.3221864104270935,
"learning_rate": 8.476872950499726e-07,
"loss": 0.4401,
"step": 1502
},
{
"epoch": 0.8311520737327189,
"grad_norm": 0.33632180094718933,
"learning_rate": 8.423167322527626e-07,
"loss": 0.4636,
"step": 1503
},
{
"epoch": 0.831705069124424,
"grad_norm": 0.3501831293106079,
"learning_rate": 8.369616710334233e-07,
"loss": 0.4573,
"step": 1504
},
{
"epoch": 0.832258064516129,
"grad_norm": 0.31366902589797974,
"learning_rate": 8.316221313578576e-07,
"loss": 0.4583,
"step": 1505
},
{
"epoch": 0.8328110599078341,
"grad_norm": 0.36373165249824524,
"learning_rate": 8.262981331340969e-07,
"loss": 0.4633,
"step": 1506
},
{
"epoch": 0.8333640552995392,
"grad_norm": 0.31518691778182983,
"learning_rate": 8.209896962122282e-07,
"loss": 0.4599,
"step": 1507
},
{
"epoch": 0.8339170506912442,
"grad_norm": 0.3142349123954773,
"learning_rate": 8.15696840384319e-07,
"loss": 0.465,
"step": 1508
},
{
"epoch": 0.8344700460829493,
"grad_norm": 0.3050268292427063,
"learning_rate": 8.104195853843433e-07,
"loss": 0.4617,
"step": 1509
},
{
"epoch": 0.8350230414746543,
"grad_norm": 0.3328079879283905,
"learning_rate": 8.051579508881107e-07,
"loss": 0.4558,
"step": 1510
},
{
"epoch": 0.8355760368663594,
"grad_norm": 0.3367290496826172,
"learning_rate": 7.999119565131891e-07,
"loss": 0.4662,
"step": 1511
},
{
"epoch": 0.8361290322580646,
"grad_norm": 0.2858041226863861,
"learning_rate": 7.946816218188347e-07,
"loss": 0.4645,
"step": 1512
},
{
"epoch": 0.8366820276497696,
"grad_norm": 0.3110063374042511,
"learning_rate": 7.894669663059168e-07,
"loss": 0.4616,
"step": 1513
},
{
"epoch": 0.8372350230414747,
"grad_norm": 0.3313262462615967,
"learning_rate": 7.84268009416847e-07,
"loss": 0.4423,
"step": 1514
},
{
"epoch": 0.8377880184331797,
"grad_norm": 0.33478546142578125,
"learning_rate": 7.790847705355059e-07,
"loss": 0.4551,
"step": 1515
},
{
"epoch": 0.8383410138248848,
"grad_norm": 0.3241022527217865,
"learning_rate": 7.739172689871705e-07,
"loss": 0.4749,
"step": 1516
},
{
"epoch": 0.8388940092165899,
"grad_norm": 0.30135831236839294,
"learning_rate": 7.687655240384423e-07,
"loss": 0.4432,
"step": 1517
},
{
"epoch": 0.8394470046082949,
"grad_norm": 0.3093847930431366,
"learning_rate": 7.636295548971762e-07,
"loss": 0.4398,
"step": 1518
},
{
"epoch": 0.84,
"grad_norm": 0.31386277079582214,
"learning_rate": 7.585093807124077e-07,
"loss": 0.4308,
"step": 1519
},
{
"epoch": 0.840552995391705,
"grad_norm": 0.32683154940605164,
"learning_rate": 7.534050205742827e-07,
"loss": 0.4671,
"step": 1520
},
{
"epoch": 0.8411059907834101,
"grad_norm": 0.32057997584342957,
"learning_rate": 7.483164935139847e-07,
"loss": 0.4777,
"step": 1521
},
{
"epoch": 0.8416589861751153,
"grad_norm": 0.3294847011566162,
"learning_rate": 7.432438185036667e-07,
"loss": 0.4872,
"step": 1522
},
{
"epoch": 0.8422119815668203,
"grad_norm": 0.3206014931201935,
"learning_rate": 7.381870144563763e-07,
"loss": 0.4719,
"step": 1523
},
{
"epoch": 0.8427649769585254,
"grad_norm": 0.291071355342865,
"learning_rate": 7.33146100225991e-07,
"loss": 0.4496,
"step": 1524
},
{
"epoch": 0.8433179723502304,
"grad_norm": 0.31741753220558167,
"learning_rate": 7.281210946071393e-07,
"loss": 0.4305,
"step": 1525
},
{
"epoch": 0.8438709677419355,
"grad_norm": 0.3081185817718506,
"learning_rate": 7.231120163351396e-07,
"loss": 0.4796,
"step": 1526
},
{
"epoch": 0.8444239631336405,
"grad_norm": 0.31715425848960876,
"learning_rate": 7.181188840859266e-07,
"loss": 0.4827,
"step": 1527
},
{
"epoch": 0.8449769585253456,
"grad_norm": 0.3110186755657196,
"learning_rate": 7.131417164759791e-07,
"loss": 0.4652,
"step": 1528
},
{
"epoch": 0.8455299539170507,
"grad_norm": 0.3567166030406952,
"learning_rate": 7.081805320622559e-07,
"loss": 0.485,
"step": 1529
},
{
"epoch": 0.8460829493087557,
"grad_norm": 0.3137590289115906,
"learning_rate": 7.032353493421213e-07,
"loss": 0.445,
"step": 1530
},
{
"epoch": 0.8466359447004608,
"grad_norm": 0.3127385675907135,
"learning_rate": 6.983061867532798e-07,
"loss": 0.4691,
"step": 1531
},
{
"epoch": 0.8471889400921659,
"grad_norm": 0.3255513608455658,
"learning_rate": 6.933930626737057e-07,
"loss": 0.4418,
"step": 1532
},
{
"epoch": 0.847741935483871,
"grad_norm": 0.2911495566368103,
"learning_rate": 6.88495995421577e-07,
"loss": 0.4578,
"step": 1533
},
{
"epoch": 0.8482949308755761,
"grad_norm": 0.29997825622558594,
"learning_rate": 6.83615003255203e-07,
"loss": 0.4543,
"step": 1534
},
{
"epoch": 0.8488479262672811,
"grad_norm": 0.2882632315158844,
"learning_rate": 6.787501043729577e-07,
"loss": 0.4531,
"step": 1535
},
{
"epoch": 0.8494009216589862,
"grad_norm": 0.32424575090408325,
"learning_rate": 6.739013169132153e-07,
"loss": 0.4516,
"step": 1536
},
{
"epoch": 0.8499539170506912,
"grad_norm": 0.3189219534397125,
"learning_rate": 6.690686589542778e-07,
"loss": 0.456,
"step": 1537
},
{
"epoch": 0.8505069124423963,
"grad_norm": 0.2835007309913635,
"learning_rate": 6.642521485143099e-07,
"loss": 0.4874,
"step": 1538
},
{
"epoch": 0.8510599078341013,
"grad_norm": 0.30130714178085327,
"learning_rate": 6.594518035512726e-07,
"loss": 0.4698,
"step": 1539
},
{
"epoch": 0.8516129032258064,
"grad_norm": 0.3199770748615265,
"learning_rate": 6.546676419628545e-07,
"loss": 0.4822,
"step": 1540
},
{
"epoch": 0.8521658986175116,
"grad_norm": 0.31458160281181335,
"learning_rate": 6.498996815864068e-07,
"loss": 0.4737,
"step": 1541
},
{
"epoch": 0.8527188940092166,
"grad_norm": 0.33827102184295654,
"learning_rate": 6.451479401988736e-07,
"loss": 0.4607,
"step": 1542
},
{
"epoch": 0.8532718894009217,
"grad_norm": 0.3004834055900574,
"learning_rate": 6.404124355167302e-07,
"loss": 0.4621,
"step": 1543
},
{
"epoch": 0.8538248847926267,
"grad_norm": 0.2976799011230469,
"learning_rate": 6.356931851959136e-07,
"loss": 0.4571,
"step": 1544
},
{
"epoch": 0.8543778801843318,
"grad_norm": 0.31761622428894043,
"learning_rate": 6.30990206831758e-07,
"loss": 0.4647,
"step": 1545
},
{
"epoch": 0.8549308755760369,
"grad_norm": 0.30549031496047974,
"learning_rate": 6.263035179589288e-07,
"loss": 0.4619,
"step": 1546
},
{
"epoch": 0.8554838709677419,
"grad_norm": 0.3004506528377533,
"learning_rate": 6.216331360513572e-07,
"loss": 0.4786,
"step": 1547
},
{
"epoch": 0.856036866359447,
"grad_norm": 0.3054359257221222,
"learning_rate": 6.169790785221763e-07,
"loss": 0.4652,
"step": 1548
},
{
"epoch": 0.856589861751152,
"grad_norm": 0.3147340416908264,
"learning_rate": 6.123413627236536e-07,
"loss": 0.4573,
"step": 1549
},
{
"epoch": 0.8571428571428571,
"grad_norm": 0.3203125,
"learning_rate": 6.077200059471289e-07,
"loss": 0.4498,
"step": 1550
},
{
"epoch": 0.8576958525345623,
"grad_norm": 0.3389976918697357,
"learning_rate": 6.031150254229484e-07,
"loss": 0.4756,
"step": 1551
},
{
"epoch": 0.8582488479262673,
"grad_norm": 0.3188997507095337,
"learning_rate": 5.985264383204003e-07,
"loss": 0.4579,
"step": 1552
},
{
"epoch": 0.8588018433179724,
"grad_norm": 0.2967085540294647,
"learning_rate": 5.939542617476529e-07,
"loss": 0.4648,
"step": 1553
},
{
"epoch": 0.8593548387096774,
"grad_norm": 0.2970874011516571,
"learning_rate": 5.893985127516866e-07,
"loss": 0.4654,
"step": 1554
},
{
"epoch": 0.8599078341013825,
"grad_norm": 0.3088582158088684,
"learning_rate": 5.848592083182348e-07,
"loss": 0.4444,
"step": 1555
},
{
"epoch": 0.8604608294930876,
"grad_norm": 0.3454407751560211,
"learning_rate": 5.803363653717187e-07,
"loss": 0.4796,
"step": 1556
},
{
"epoch": 0.8610138248847926,
"grad_norm": 0.34124165773391724,
"learning_rate": 5.758300007751832e-07,
"loss": 0.475,
"step": 1557
},
{
"epoch": 0.8615668202764977,
"grad_norm": 0.31380608677864075,
"learning_rate": 5.713401313302358e-07,
"loss": 0.4714,
"step": 1558
},
{
"epoch": 0.8621198156682027,
"grad_norm": 0.3094062805175781,
"learning_rate": 5.668667737769834e-07,
"loss": 0.4348,
"step": 1559
},
{
"epoch": 0.8626728110599078,
"grad_norm": 0.3069341778755188,
"learning_rate": 5.624099447939696e-07,
"loss": 0.4677,
"step": 1560
},
{
"epoch": 0.863225806451613,
"grad_norm": 0.34499165415763855,
"learning_rate": 5.579696609981117e-07,
"loss": 0.4673,
"step": 1561
},
{
"epoch": 0.863778801843318,
"grad_norm": 0.32358747720718384,
"learning_rate": 5.535459389446401e-07,
"loss": 0.4627,
"step": 1562
},
{
"epoch": 0.8643317972350231,
"grad_norm": 0.33145248889923096,
"learning_rate": 5.491387951270366e-07,
"loss": 0.4535,
"step": 1563
},
{
"epoch": 0.8648847926267281,
"grad_norm": 0.3262435495853424,
"learning_rate": 5.447482459769709e-07,
"loss": 0.4613,
"step": 1564
},
{
"epoch": 0.8654377880184332,
"grad_norm": 0.31697559356689453,
"learning_rate": 5.403743078642443e-07,
"loss": 0.4503,
"step": 1565
},
{
"epoch": 0.8659907834101382,
"grad_norm": 0.3151688873767853,
"learning_rate": 5.360169970967221e-07,
"loss": 0.4816,
"step": 1566
},
{
"epoch": 0.8665437788018433,
"grad_norm": 0.3067736029624939,
"learning_rate": 5.316763299202766e-07,
"loss": 0.4631,
"step": 1567
},
{
"epoch": 0.8670967741935484,
"grad_norm": 0.2979617714881897,
"learning_rate": 5.273523225187255e-07,
"loss": 0.4403,
"step": 1568
},
{
"epoch": 0.8676497695852534,
"grad_norm": 0.3092879056930542,
"learning_rate": 5.23044991013773e-07,
"loss": 0.4669,
"step": 1569
},
{
"epoch": 0.8682027649769585,
"grad_norm": 0.29821327328681946,
"learning_rate": 5.187543514649479e-07,
"loss": 0.4678,
"step": 1570
},
{
"epoch": 0.8687557603686636,
"grad_norm": 0.3069091737270355,
"learning_rate": 5.144804198695447e-07,
"loss": 0.4834,
"step": 1571
},
{
"epoch": 0.8693087557603687,
"grad_norm": 0.3062274754047394,
"learning_rate": 5.102232121625633e-07,
"loss": 0.4583,
"step": 1572
},
{
"epoch": 0.8698617511520738,
"grad_norm": 0.3618842661380768,
"learning_rate": 5.059827442166504e-07,
"loss": 0.4348,
"step": 1573
},
{
"epoch": 0.8704147465437788,
"grad_norm": 0.3038586676120758,
"learning_rate": 5.017590318420395e-07,
"loss": 0.4601,
"step": 1574
},
{
"epoch": 0.8709677419354839,
"grad_norm": 0.30661341547966003,
"learning_rate": 4.975520907864928e-07,
"loss": 0.4432,
"step": 1575
},
{
"epoch": 0.8715207373271889,
"grad_norm": 0.3259396255016327,
"learning_rate": 4.93361936735241e-07,
"loss": 0.4602,
"step": 1576
},
{
"epoch": 0.872073732718894,
"grad_norm": 0.32900482416152954,
"learning_rate": 4.891885853109279e-07,
"loss": 0.473,
"step": 1577
},
{
"epoch": 0.872626728110599,
"grad_norm": 0.280129075050354,
"learning_rate": 4.85032052073548e-07,
"loss": 0.4581,
"step": 1578
},
{
"epoch": 0.8731797235023041,
"grad_norm": 0.31960612535476685,
"learning_rate": 4.808923525203912e-07,
"loss": 0.4472,
"step": 1579
},
{
"epoch": 0.8737327188940092,
"grad_norm": 0.28361162543296814,
"learning_rate": 4.767695020859847e-07,
"loss": 0.4538,
"step": 1580
},
{
"epoch": 0.8742857142857143,
"grad_norm": 0.3138439953327179,
"learning_rate": 4.726635161420351e-07,
"loss": 0.4746,
"step": 1581
},
{
"epoch": 0.8748387096774194,
"grad_norm": 0.30272024869918823,
"learning_rate": 4.685744099973716e-07,
"loss": 0.4713,
"step": 1582
},
{
"epoch": 0.8753917050691244,
"grad_norm": 0.31137874722480774,
"learning_rate": 4.6450219889788816e-07,
"loss": 0.4569,
"step": 1583
},
{
"epoch": 0.8759447004608295,
"grad_norm": 0.34966525435447693,
"learning_rate": 4.6044689802648534e-07,
"loss": 0.4646,
"step": 1584
},
{
"epoch": 0.8764976958525346,
"grad_norm": 0.29577386379241943,
"learning_rate": 4.564085225030174e-07,
"loss": 0.4642,
"step": 1585
},
{
"epoch": 0.8770506912442396,
"grad_norm": 0.2946975529193878,
"learning_rate": 4.523870873842329e-07,
"loss": 0.4312,
"step": 1586
},
{
"epoch": 0.8776036866359447,
"grad_norm": 0.30097469687461853,
"learning_rate": 4.4838260766372044e-07,
"loss": 0.4666,
"step": 1587
},
{
"epoch": 0.8781566820276497,
"grad_norm": 0.2847994565963745,
"learning_rate": 4.4439509827185034e-07,
"loss": 0.4165,
"step": 1588
},
{
"epoch": 0.8787096774193548,
"grad_norm": 0.2944558560848236,
"learning_rate": 4.4042457407572235e-07,
"loss": 0.4721,
"step": 1589
},
{
"epoch": 0.8792626728110599,
"grad_norm": 0.3086279034614563,
"learning_rate": 4.3647104987910636e-07,
"loss": 0.45,
"step": 1590
},
{
"epoch": 0.879815668202765,
"grad_norm": 0.3032161593437195,
"learning_rate": 4.3253454042239016e-07,
"loss": 0.458,
"step": 1591
},
{
"epoch": 0.8803686635944701,
"grad_norm": 0.3120705485343933,
"learning_rate": 4.286150603825234e-07,
"loss": 0.474,
"step": 1592
},
{
"epoch": 0.8809216589861751,
"grad_norm": 0.332922101020813,
"learning_rate": 4.2471262437296326e-07,
"loss": 0.4702,
"step": 1593
},
{
"epoch": 0.8814746543778802,
"grad_norm": 0.2810650169849396,
"learning_rate": 4.208272469436192e-07,
"loss": 0.4509,
"step": 1594
},
{
"epoch": 0.8820276497695853,
"grad_norm": 0.31393900513648987,
"learning_rate": 4.1695894258079903e-07,
"loss": 0.4734,
"step": 1595
},
{
"epoch": 0.8825806451612903,
"grad_norm": 0.2897551357746124,
"learning_rate": 4.1310772570715586e-07,
"loss": 0.4442,
"step": 1596
},
{
"epoch": 0.8831336405529954,
"grad_norm": 0.3010689616203308,
"learning_rate": 4.092736106816314e-07,
"loss": 0.4477,
"step": 1597
},
{
"epoch": 0.8836866359447004,
"grad_norm": 0.29473552107810974,
"learning_rate": 4.0545661179940854e-07,
"loss": 0.4595,
"step": 1598
},
{
"epoch": 0.8842396313364055,
"grad_norm": 0.3020123839378357,
"learning_rate": 4.0165674329185055e-07,
"loss": 0.461,
"step": 1599
},
{
"epoch": 0.8847926267281107,
"grad_norm": 0.3016417920589447,
"learning_rate": 3.978740193264524e-07,
"loss": 0.4802,
"step": 1600
},
{
"epoch": 0.8853456221198157,
"grad_norm": 0.3054351806640625,
"learning_rate": 3.941084540067874e-07,
"loss": 0.461,
"step": 1601
},
{
"epoch": 0.8858986175115208,
"grad_norm": 0.31593799591064453,
"learning_rate": 3.9036006137245396e-07,
"loss": 0.485,
"step": 1602
},
{
"epoch": 0.8864516129032258,
"grad_norm": 0.2995027005672455,
"learning_rate": 3.866288553990233e-07,
"loss": 0.4544,
"step": 1603
},
{
"epoch": 0.8870046082949309,
"grad_norm": 0.3163018524646759,
"learning_rate": 3.82914849997989e-07,
"loss": 0.4675,
"step": 1604
},
{
"epoch": 0.8875576036866359,
"grad_norm": 0.3466692268848419,
"learning_rate": 3.7921805901671273e-07,
"loss": 0.455,
"step": 1605
},
{
"epoch": 0.888110599078341,
"grad_norm": 0.30836209654808044,
"learning_rate": 3.7553849623837237e-07,
"loss": 0.4983,
"step": 1606
},
{
"epoch": 0.8886635944700461,
"grad_norm": 0.3018787205219269,
"learning_rate": 3.7187617538191446e-07,
"loss": 0.4714,
"step": 1607
},
{
"epoch": 0.8892165898617511,
"grad_norm": 0.3216530382633209,
"learning_rate": 3.682311101019981e-07,
"loss": 0.4857,
"step": 1608
},
{
"epoch": 0.8897695852534562,
"grad_norm": 0.32293516397476196,
"learning_rate": 3.646033139889482e-07,
"loss": 0.4508,
"step": 1609
},
{
"epoch": 0.8903225806451613,
"grad_norm": 0.29628053307533264,
"learning_rate": 3.6099280056870136e-07,
"loss": 0.4271,
"step": 1610
},
{
"epoch": 0.8908755760368664,
"grad_norm": 0.30662253499031067,
"learning_rate": 3.5739958330275936e-07,
"loss": 0.4833,
"step": 1611
},
{
"epoch": 0.8914285714285715,
"grad_norm": 0.30339956283569336,
"learning_rate": 3.538236755881341e-07,
"loss": 0.4675,
"step": 1612
},
{
"epoch": 0.8919815668202765,
"grad_norm": 0.29642388224601746,
"learning_rate": 3.502650907573024e-07,
"loss": 0.4529,
"step": 1613
},
{
"epoch": 0.8925345622119816,
"grad_norm": 0.3087359666824341,
"learning_rate": 3.467238420781527e-07,
"loss": 0.4794,
"step": 1614
},
{
"epoch": 0.8930875576036866,
"grad_norm": 0.34186214208602905,
"learning_rate": 3.431999427539368e-07,
"loss": 0.4951,
"step": 1615
},
{
"epoch": 0.8936405529953917,
"grad_norm": 0.31149131059646606,
"learning_rate": 3.396934059232226e-07,
"loss": 0.4476,
"step": 1616
},
{
"epoch": 0.8941935483870967,
"grad_norm": 0.3173343241214752,
"learning_rate": 3.362042446598418e-07,
"loss": 0.4735,
"step": 1617
},
{
"epoch": 0.8947465437788018,
"grad_norm": 0.29687947034835815,
"learning_rate": 3.3273247197284366e-07,
"loss": 0.4588,
"step": 1618
},
{
"epoch": 0.8952995391705069,
"grad_norm": 0.29158180952072144,
"learning_rate": 3.292781008064455e-07,
"loss": 0.4477,
"step": 1619
},
{
"epoch": 0.895852534562212,
"grad_norm": 0.31376034021377563,
"learning_rate": 3.258411440399839e-07,
"loss": 0.4613,
"step": 1620
},
{
"epoch": 0.8964055299539171,
"grad_norm": 0.3553878366947174,
"learning_rate": 3.2242161448786724e-07,
"loss": 0.4638,
"step": 1621
},
{
"epoch": 0.8969585253456221,
"grad_norm": 0.31276994943618774,
"learning_rate": 3.1901952489952927e-07,
"loss": 0.4297,
"step": 1622
},
{
"epoch": 0.8975115207373272,
"grad_norm": 0.2925131320953369,
"learning_rate": 3.15634887959379e-07,
"loss": 0.4642,
"step": 1623
},
{
"epoch": 0.8980645161290323,
"grad_norm": 0.3111323416233063,
"learning_rate": 3.122677162867549e-07,
"loss": 0.4628,
"step": 1624
},
{
"epoch": 0.8986175115207373,
"grad_norm": 0.28449130058288574,
"learning_rate": 3.089180224358773e-07,
"loss": 0.4619,
"step": 1625
},
{
"epoch": 0.8991705069124424,
"grad_norm": 0.2895408868789673,
"learning_rate": 3.0558581889580383e-07,
"loss": 0.4486,
"step": 1626
},
{
"epoch": 0.8997235023041474,
"grad_norm": 0.28121551871299744,
"learning_rate": 3.0227111809037815e-07,
"loss": 0.4523,
"step": 1627
},
{
"epoch": 0.9002764976958525,
"grad_norm": 0.3156479001045227,
"learning_rate": 2.9897393237818873e-07,
"loss": 0.4504,
"step": 1628
},
{
"epoch": 0.9008294930875576,
"grad_norm": 0.30593904852867126,
"learning_rate": 2.956942740525187e-07,
"loss": 0.4832,
"step": 1629
},
{
"epoch": 0.9013824884792627,
"grad_norm": 0.2923428416252136,
"learning_rate": 2.924321553413029e-07,
"loss": 0.4473,
"step": 1630
},
{
"epoch": 0.9019354838709678,
"grad_norm": 0.29185551404953003,
"learning_rate": 2.891875884070816e-07,
"loss": 0.4514,
"step": 1631
},
{
"epoch": 0.9024884792626728,
"grad_norm": 0.3090588450431824,
"learning_rate": 2.859605853469533e-07,
"loss": 0.492,
"step": 1632
},
{
"epoch": 0.9030414746543779,
"grad_norm": 0.2918623387813568,
"learning_rate": 2.8275115819253184e-07,
"loss": 0.4378,
"step": 1633
},
{
"epoch": 0.903594470046083,
"grad_norm": 0.2900620698928833,
"learning_rate": 2.795593189099016e-07,
"loss": 0.4702,
"step": 1634
},
{
"epoch": 0.904147465437788,
"grad_norm": 0.30443379282951355,
"learning_rate": 2.7638507939956926e-07,
"loss": 0.4562,
"step": 1635
},
{
"epoch": 0.9047004608294931,
"grad_norm": 0.2894691526889801,
"learning_rate": 2.7322845149642564e-07,
"loss": 0.4405,
"step": 1636
},
{
"epoch": 0.9052534562211981,
"grad_norm": 0.30414119362831116,
"learning_rate": 2.700894469696963e-07,
"loss": 0.4618,
"step": 1637
},
{
"epoch": 0.9058064516129032,
"grad_norm": 0.29795682430267334,
"learning_rate": 2.669680775228989e-07,
"loss": 0.4717,
"step": 1638
},
{
"epoch": 0.9063594470046082,
"grad_norm": 0.2925158739089966,
"learning_rate": 2.638643547938019e-07,
"loss": 0.4788,
"step": 1639
},
{
"epoch": 0.9069124423963134,
"grad_norm": 0.29635146260261536,
"learning_rate": 2.607782903543782e-07,
"loss": 0.4521,
"step": 1640
},
{
"epoch": 0.9074654377880185,
"grad_norm": 0.29401543736457825,
"learning_rate": 2.5770989571076375e-07,
"loss": 0.4401,
"step": 1641
},
{
"epoch": 0.9080184331797235,
"grad_norm": 0.2778622508049011,
"learning_rate": 2.5465918230321464e-07,
"loss": 0.4522,
"step": 1642
},
{
"epoch": 0.9085714285714286,
"grad_norm": 0.2870117127895355,
"learning_rate": 2.5162616150606167e-07,
"loss": 0.4686,
"step": 1643
},
{
"epoch": 0.9091244239631336,
"grad_norm": 0.3375545144081116,
"learning_rate": 2.486108446276725e-07,
"loss": 0.4741,
"step": 1644
},
{
"epoch": 0.9096774193548387,
"grad_norm": 0.285396546125412,
"learning_rate": 2.4561324291040636e-07,
"loss": 0.4749,
"step": 1645
},
{
"epoch": 0.9102304147465438,
"grad_norm": 0.31609663367271423,
"learning_rate": 2.4263336753057364e-07,
"loss": 0.4552,
"step": 1646
},
{
"epoch": 0.9107834101382488,
"grad_norm": 0.32548245787620544,
"learning_rate": 2.396712295983922e-07,
"loss": 0.4707,
"step": 1647
},
{
"epoch": 0.9113364055299539,
"grad_norm": 0.28127720952033997,
"learning_rate": 2.3672684015794923e-07,
"loss": 0.4808,
"step": 1648
},
{
"epoch": 0.9118894009216589,
"grad_norm": 0.3091859817504883,
"learning_rate": 2.338002101871556e-07,
"loss": 0.4469,
"step": 1649
},
{
"epoch": 0.9124423963133641,
"grad_norm": 0.3138236403465271,
"learning_rate": 2.3089135059771007e-07,
"loss": 0.4848,
"step": 1650
},
{
"epoch": 0.9129953917050692,
"grad_norm": 0.30295220017433167,
"learning_rate": 2.2800027223505438e-07,
"loss": 0.4567,
"step": 1651
},
{
"epoch": 0.9135483870967742,
"grad_norm": 0.3260992467403412,
"learning_rate": 2.2512698587833502e-07,
"loss": 0.4776,
"step": 1652
},
{
"epoch": 0.9141013824884793,
"grad_norm": 0.30410924553871155,
"learning_rate": 2.222715022403621e-07,
"loss": 0.4565,
"step": 1653
},
{
"epoch": 0.9146543778801843,
"grad_norm": 0.296068400144577,
"learning_rate": 2.1943383196756984e-07,
"loss": 0.4868,
"step": 1654
},
{
"epoch": 0.9152073732718894,
"grad_norm": 0.3140951991081238,
"learning_rate": 2.1661398563997737e-07,
"loss": 0.4645,
"step": 1655
},
{
"epoch": 0.9157603686635944,
"grad_norm": 0.2961066663265228,
"learning_rate": 2.138119737711475e-07,
"loss": 0.4376,
"step": 1656
},
{
"epoch": 0.9163133640552995,
"grad_norm": 0.3166109621524811,
"learning_rate": 2.110278068081506e-07,
"loss": 0.4713,
"step": 1657
},
{
"epoch": 0.9168663594470046,
"grad_norm": 0.32516512274742126,
"learning_rate": 2.082614951315215e-07,
"loss": 0.4516,
"step": 1658
},
{
"epoch": 0.9174193548387096,
"grad_norm": 0.293089896440506,
"learning_rate": 2.055130490552254e-07,
"loss": 0.469,
"step": 1659
},
{
"epoch": 0.9179723502304148,
"grad_norm": 0.31795766949653625,
"learning_rate": 2.0278247882661584e-07,
"loss": 0.4497,
"step": 1660
},
{
"epoch": 0.9185253456221198,
"grad_norm": 0.2847367525100708,
"learning_rate": 2.0006979462639686e-07,
"loss": 0.4447,
"step": 1661
},
{
"epoch": 0.9190783410138249,
"grad_norm": 0.2693861126899719,
"learning_rate": 1.9737500656858754e-07,
"loss": 0.4406,
"step": 1662
},
{
"epoch": 0.91963133640553,
"grad_norm": 0.3001803755760193,
"learning_rate": 1.94698124700482e-07,
"loss": 0.4964,
"step": 1663
},
{
"epoch": 0.920184331797235,
"grad_norm": 0.3073177933692932,
"learning_rate": 1.9203915900261327e-07,
"loss": 0.4737,
"step": 1664
},
{
"epoch": 0.9207373271889401,
"grad_norm": 0.31089356541633606,
"learning_rate": 1.8939811938871456e-07,
"loss": 0.4539,
"step": 1665
},
{
"epoch": 0.9212903225806451,
"grad_norm": 0.27542582154273987,
"learning_rate": 1.86775015705683e-07,
"loss": 0.434,
"step": 1666
},
{
"epoch": 0.9218433179723502,
"grad_norm": 0.32208341360092163,
"learning_rate": 1.841698577335438e-07,
"loss": 0.4476,
"step": 1667
},
{
"epoch": 0.9223963133640553,
"grad_norm": 0.3119848668575287,
"learning_rate": 1.8158265518541274e-07,
"loss": 0.4484,
"step": 1668
},
{
"epoch": 0.9229493087557604,
"grad_norm": 0.2974831759929657,
"learning_rate": 1.790134177074604e-07,
"loss": 0.4574,
"step": 1669
},
{
"epoch": 0.9235023041474655,
"grad_norm": 0.29929405450820923,
"learning_rate": 1.7646215487887587e-07,
"loss": 0.4685,
"step": 1670
},
{
"epoch": 0.9240552995391705,
"grad_norm": 0.2898726165294647,
"learning_rate": 1.739288762118313e-07,
"loss": 0.455,
"step": 1671
},
{
"epoch": 0.9246082949308756,
"grad_norm": 0.3034640848636627,
"learning_rate": 1.714135911514475e-07,
"loss": 0.4541,
"step": 1672
},
{
"epoch": 0.9251612903225807,
"grad_norm": 0.3259904682636261,
"learning_rate": 1.6891630907575562e-07,
"loss": 0.4678,
"step": 1673
},
{
"epoch": 0.9257142857142857,
"grad_norm": 0.3421642780303955,
"learning_rate": 1.664370392956649e-07,
"loss": 0.4518,
"step": 1674
},
{
"epoch": 0.9262672811059908,
"grad_norm": 0.29688096046447754,
"learning_rate": 1.6397579105492778e-07,
"loss": 0.4222,
"step": 1675
},
{
"epoch": 0.9268202764976958,
"grad_norm": 0.3092136085033417,
"learning_rate": 1.6153257353010433e-07,
"loss": 0.4679,
"step": 1676
},
{
"epoch": 0.9273732718894009,
"grad_norm": 0.3253113627433777,
"learning_rate": 1.5910739583053002e-07,
"loss": 0.4495,
"step": 1677
},
{
"epoch": 0.9279262672811059,
"grad_norm": 0.286625474691391,
"learning_rate": 1.5670026699827757e-07,
"loss": 0.4652,
"step": 1678
},
{
"epoch": 0.9284792626728111,
"grad_norm": 0.2905667722225189,
"learning_rate": 1.5431119600812837e-07,
"loss": 0.4529,
"step": 1679
},
{
"epoch": 0.9290322580645162,
"grad_norm": 0.32414594292640686,
"learning_rate": 1.5194019176753615e-07,
"loss": 0.4878,
"step": 1680
},
{
"epoch": 0.9295852534562212,
"grad_norm": 0.29310277104377747,
"learning_rate": 1.495872631165929e-07,
"loss": 0.476,
"step": 1681
},
{
"epoch": 0.9301382488479263,
"grad_norm": 0.3012104332447052,
"learning_rate": 1.4725241882799946e-07,
"loss": 0.4823,
"step": 1682
},
{
"epoch": 0.9306912442396313,
"grad_norm": 0.2896125018596649,
"learning_rate": 1.449356676070285e-07,
"loss": 0.4429,
"step": 1683
},
{
"epoch": 0.9312442396313364,
"grad_norm": 0.2954270839691162,
"learning_rate": 1.42637018091496e-07,
"loss": 0.473,
"step": 1684
},
{
"epoch": 0.9317972350230415,
"grad_norm": 0.29838648438453674,
"learning_rate": 1.4035647885172533e-07,
"loss": 0.4615,
"step": 1685
},
{
"epoch": 0.9323502304147465,
"grad_norm": 0.3438728451728821,
"learning_rate": 1.3809405839051937e-07,
"loss": 0.4924,
"step": 1686
},
{
"epoch": 0.9329032258064516,
"grad_norm": 0.2971719205379486,
"learning_rate": 1.358497651431251e-07,
"loss": 0.4438,
"step": 1687
},
{
"epoch": 0.9334562211981566,
"grad_norm": 0.3192031979560852,
"learning_rate": 1.3362360747720405e-07,
"loss": 0.4559,
"step": 1688
},
{
"epoch": 0.9340092165898618,
"grad_norm": 0.3306371867656708,
"learning_rate": 1.314155936928002e-07,
"loss": 0.4633,
"step": 1689
},
{
"epoch": 0.9345622119815669,
"grad_norm": 0.28587308526039124,
"learning_rate": 1.2922573202231114e-07,
"loss": 0.4577,
"step": 1690
},
{
"epoch": 0.9351152073732719,
"grad_norm": 0.31524068117141724,
"learning_rate": 1.2705403063045464e-07,
"loss": 0.4457,
"step": 1691
},
{
"epoch": 0.935668202764977,
"grad_norm": 0.2901270389556885,
"learning_rate": 1.2490049761423927e-07,
"loss": 0.4396,
"step": 1692
},
{
"epoch": 0.936221198156682,
"grad_norm": 0.2995949685573578,
"learning_rate": 1.2276514100293403e-07,
"loss": 0.4633,
"step": 1693
},
{
"epoch": 0.9367741935483871,
"grad_norm": 0.2743770182132721,
"learning_rate": 1.206479687580403e-07,
"loss": 0.4655,
"step": 1694
},
{
"epoch": 0.9373271889400921,
"grad_norm": 0.3132480978965759,
"learning_rate": 1.1854898877325882e-07,
"loss": 0.4641,
"step": 1695
},
{
"epoch": 0.9378801843317972,
"grad_norm": 0.2994924783706665,
"learning_rate": 1.1646820887446342e-07,
"loss": 0.4491,
"step": 1696
},
{
"epoch": 0.9384331797235023,
"grad_norm": 0.2822410762310028,
"learning_rate": 1.1440563681966943e-07,
"loss": 0.4663,
"step": 1697
},
{
"epoch": 0.9389861751152073,
"grad_norm": 0.3047800660133362,
"learning_rate": 1.1236128029900539e-07,
"loss": 0.4638,
"step": 1698
},
{
"epoch": 0.9395391705069125,
"grad_norm": 0.31529849767684937,
"learning_rate": 1.1033514693468584e-07,
"loss": 0.4269,
"step": 1699
},
{
"epoch": 0.9400921658986175,
"grad_norm": 0.28443610668182373,
"learning_rate": 1.0832724428098185e-07,
"loss": 0.4568,
"step": 1700
},
{
"epoch": 0.9406451612903226,
"grad_norm": 0.2651902735233307,
"learning_rate": 1.0633757982419169e-07,
"loss": 0.4508,
"step": 1701
},
{
"epoch": 0.9411981566820277,
"grad_norm": 0.29510411620140076,
"learning_rate": 1.0436616098261409e-07,
"loss": 0.4643,
"step": 1702
},
{
"epoch": 0.9417511520737327,
"grad_norm": 0.2974132299423218,
"learning_rate": 1.0241299510652114e-07,
"loss": 0.429,
"step": 1703
},
{
"epoch": 0.9423041474654378,
"grad_norm": 0.2929219901561737,
"learning_rate": 1.0047808947813097e-07,
"loss": 0.4851,
"step": 1704
},
{
"epoch": 0.9428571428571428,
"grad_norm": 0.29706960916519165,
"learning_rate": 9.85614513115779e-08,
"loss": 0.4559,
"step": 1705
},
{
"epoch": 0.9434101382488479,
"grad_norm": 0.298990935087204,
"learning_rate": 9.666308775289013e-08,
"loss": 0.479,
"step": 1706
},
{
"epoch": 0.943963133640553,
"grad_norm": 0.29148900508880615,
"learning_rate": 9.47830058799576e-08,
"loss": 0.4577,
"step": 1707
},
{
"epoch": 0.944516129032258,
"grad_norm": 0.3139622211456299,
"learning_rate": 9.292121270251031e-08,
"loss": 0.4569,
"step": 1708
},
{
"epoch": 0.9450691244239632,
"grad_norm": 0.29921236634254456,
"learning_rate": 9.107771516209008e-08,
"loss": 0.4558,
"step": 1709
},
{
"epoch": 0.9456221198156682,
"grad_norm": 0.28523844480514526,
"learning_rate": 8.925252013202545e-08,
"loss": 0.4709,
"step": 1710
},
{
"epoch": 0.9461751152073733,
"grad_norm": 0.3282380998134613,
"learning_rate": 8.744563441740461e-08,
"loss": 0.4485,
"step": 1711
},
{
"epoch": 0.9467281105990784,
"grad_norm": 0.30116501450538635,
"learning_rate": 8.565706475505086e-08,
"loss": 0.4633,
"step": 1712
},
{
"epoch": 0.9472811059907834,
"grad_norm": 0.2834503948688507,
"learning_rate": 8.388681781349828e-08,
"loss": 0.4569,
"step": 1713
},
{
"epoch": 0.9478341013824885,
"grad_norm": 0.31157195568084717,
"learning_rate": 8.213490019296666e-08,
"loss": 0.4666,
"step": 1714
},
{
"epoch": 0.9483870967741935,
"grad_norm": 0.30053144693374634,
"learning_rate": 8.04013184253355e-08,
"loss": 0.4645,
"step": 1715
},
{
"epoch": 0.9489400921658986,
"grad_norm": 0.3080219328403473,
"learning_rate": 7.868607897412062e-08,
"loss": 0.4409,
"step": 1716
},
{
"epoch": 0.9494930875576036,
"grad_norm": 0.2975987493991852,
"learning_rate": 7.69891882344509e-08,
"loss": 0.4691,
"step": 1717
},
{
"epoch": 0.9500460829493087,
"grad_norm": 0.3106543719768524,
"learning_rate": 7.53106525330427e-08,
"loss": 0.4816,
"step": 1718
},
{
"epoch": 0.9505990783410139,
"grad_norm": 0.29965740442276,
"learning_rate": 7.36504781281766e-08,
"loss": 0.469,
"step": 1719
},
{
"epoch": 0.9511520737327189,
"grad_norm": 0.31763914227485657,
"learning_rate": 7.20086712096768e-08,
"loss": 0.4615,
"step": 1720
},
{
"epoch": 0.951705069124424,
"grad_norm": 0.2858491837978363,
"learning_rate": 7.038523789888397e-08,
"loss": 0.457,
"step": 1721
},
{
"epoch": 0.952258064516129,
"grad_norm": 0.2787969410419464,
"learning_rate": 6.878018424863408e-08,
"loss": 0.4485,
"step": 1722
},
{
"epoch": 0.9528110599078341,
"grad_norm": 0.32252055406570435,
"learning_rate": 6.7193516243238e-08,
"loss": 0.4708,
"step": 1723
},
{
"epoch": 0.9533640552995392,
"grad_norm": 0.3171110451221466,
"learning_rate": 6.562523979845525e-08,
"loss": 0.466,
"step": 1724
},
{
"epoch": 0.9539170506912442,
"grad_norm": 0.31782266497612,
"learning_rate": 6.407536076147525e-08,
"loss": 0.4624,
"step": 1725
},
{
"epoch": 0.9544700460829493,
"grad_norm": 0.2952912449836731,
"learning_rate": 6.254388491089281e-08,
"loss": 0.4537,
"step": 1726
},
{
"epoch": 0.9550230414746543,
"grad_norm": 0.28047212958335876,
"learning_rate": 6.10308179566893e-08,
"loss": 0.473,
"step": 1727
},
{
"epoch": 0.9555760368663595,
"grad_norm": 0.30235886573791504,
"learning_rate": 5.9536165540209914e-08,
"loss": 0.489,
"step": 1728
},
{
"epoch": 0.9561290322580646,
"grad_norm": 0.29142141342163086,
"learning_rate": 5.80599332341425e-08,
"loss": 0.4718,
"step": 1729
},
{
"epoch": 0.9566820276497696,
"grad_norm": 0.2832047939300537,
"learning_rate": 5.6602126542496525e-08,
"loss": 0.4562,
"step": 1730
},
{
"epoch": 0.9572350230414747,
"grad_norm": 0.2992711365222931,
"learning_rate": 5.516275090058476e-08,
"loss": 0.4816,
"step": 1731
},
{
"epoch": 0.9577880184331797,
"grad_norm": 0.29784855246543884,
"learning_rate": 5.3741811674998254e-08,
"loss": 0.4818,
"step": 1732
},
{
"epoch": 0.9583410138248848,
"grad_norm": 0.3014211058616638,
"learning_rate": 5.233931416359195e-08,
"loss": 0.4914,
"step": 1733
},
{
"epoch": 0.9588940092165898,
"grad_norm": 0.30964842438697815,
"learning_rate": 5.095526359546243e-08,
"loss": 0.4643,
"step": 1734
},
{
"epoch": 0.9594470046082949,
"grad_norm": 0.3023122251033783,
"learning_rate": 4.958966513092689e-08,
"loss": 0.4499,
"step": 1735
},
{
"epoch": 0.96,
"grad_norm": 0.295585572719574,
"learning_rate": 4.8242523861506405e-08,
"loss": 0.4248,
"step": 1736
},
{
"epoch": 0.960552995391705,
"grad_norm": 0.2833447754383087,
"learning_rate": 4.691384480990602e-08,
"loss": 0.4491,
"step": 1737
},
{
"epoch": 0.9611059907834102,
"grad_norm": 0.309023916721344,
"learning_rate": 4.56036329299947e-08,
"loss": 0.4404,
"step": 1738
},
{
"epoch": 0.9616589861751152,
"grad_norm": 0.3236503303050995,
"learning_rate": 4.4311893106789847e-08,
"loss": 0.4591,
"step": 1739
},
{
"epoch": 0.9622119815668203,
"grad_norm": 0.29482167959213257,
"learning_rate": 4.3038630156436166e-08,
"loss": 0.4558,
"step": 1740
},
{
"epoch": 0.9627649769585254,
"grad_norm": 0.2854726016521454,
"learning_rate": 4.1783848826189025e-08,
"loss": 0.4561,
"step": 1741
},
{
"epoch": 0.9633179723502304,
"grad_norm": 0.28575026988983154,
"learning_rate": 4.0547553794397235e-08,
"loss": 0.4541,
"step": 1742
},
{
"epoch": 0.9638709677419355,
"grad_norm": 0.30080512166023254,
"learning_rate": 3.932974967048364e-08,
"loss": 0.4475,
"step": 1743
},
{
"epoch": 0.9644239631336405,
"grad_norm": 0.2857537269592285,
"learning_rate": 3.813044099493068e-08,
"loss": 0.4521,
"step": 1744
},
{
"epoch": 0.9649769585253456,
"grad_norm": 0.28055858612060547,
"learning_rate": 3.6949632239261514e-08,
"loss": 0.4388,
"step": 1745
},
{
"epoch": 0.9655299539170507,
"grad_norm": 0.27776503562927246,
"learning_rate": 3.578732780602334e-08,
"loss": 0.4483,
"step": 1746
},
{
"epoch": 0.9660829493087557,
"grad_norm": 0.322313517332077,
"learning_rate": 3.464353202877302e-08,
"loss": 0.4589,
"step": 1747
},
{
"epoch": 0.9666359447004609,
"grad_norm": 0.315902978181839,
"learning_rate": 3.351824917205704e-08,
"loss": 0.4696,
"step": 1748
},
{
"epoch": 0.9671889400921659,
"grad_norm": 0.3116399645805359,
"learning_rate": 3.2411483431400435e-08,
"loss": 0.4532,
"step": 1749
},
{
"epoch": 0.967741935483871,
"grad_norm": 0.27709704637527466,
"learning_rate": 3.1323238933286814e-08,
"loss": 0.4808,
"step": 1750
},
{
"epoch": 0.968294930875576,
"grad_norm": 0.28587040305137634,
"learning_rate": 3.0253519735146674e-08,
"loss": 0.4498,
"step": 1751
},
{
"epoch": 0.9688479262672811,
"grad_norm": 0.3236626088619232,
"learning_rate": 2.920232982533855e-08,
"loss": 0.4618,
"step": 1752
},
{
"epoch": 0.9694009216589862,
"grad_norm": 0.29707276821136475,
"learning_rate": 2.8169673123137896e-08,
"loss": 0.4516,
"step": 1753
},
{
"epoch": 0.9699539170506912,
"grad_norm": 0.3110201954841614,
"learning_rate": 2.7155553478719343e-08,
"loss": 0.4534,
"step": 1754
},
{
"epoch": 0.9705069124423963,
"grad_norm": 0.308906227350235,
"learning_rate": 2.6159974673143352e-08,
"loss": 0.447,
"step": 1755
},
{
"epoch": 0.9710599078341013,
"grad_norm": 0.30986642837524414,
"learning_rate": 2.5182940418344016e-08,
"loss": 0.4587,
"step": 1756
},
{
"epoch": 0.9716129032258064,
"grad_norm": 0.31161195039749146,
"learning_rate": 2.4224454357112404e-08,
"loss": 0.4616,
"step": 1757
},
{
"epoch": 0.9721658986175116,
"grad_norm": 0.2983948290348053,
"learning_rate": 2.328452006308435e-08,
"loss": 0.4696,
"step": 1758
},
{
"epoch": 0.9727188940092166,
"grad_norm": 0.28888341784477234,
"learning_rate": 2.2363141040727123e-08,
"loss": 0.4859,
"step": 1759
},
{
"epoch": 0.9732718894009217,
"grad_norm": 0.2775828242301941,
"learning_rate": 2.1460320725326113e-08,
"loss": 0.4354,
"step": 1760
},
{
"epoch": 0.9738248847926267,
"grad_norm": 0.34634268283843994,
"learning_rate": 2.057606248297206e-08,
"loss": 0.4431,
"step": 1761
},
{
"epoch": 0.9743778801843318,
"grad_norm": 0.31715697050094604,
"learning_rate": 1.971036961054884e-08,
"loss": 0.4566,
"step": 1762
},
{
"epoch": 0.9749308755760369,
"grad_norm": 0.2788535952568054,
"learning_rate": 1.886324533572015e-08,
"loss": 0.4492,
"step": 1763
},
{
"epoch": 0.9754838709677419,
"grad_norm": 0.3046756684780121,
"learning_rate": 1.8034692816919497e-08,
"loss": 0.4461,
"step": 1764
},
{
"epoch": 0.976036866359447,
"grad_norm": 0.2936743497848511,
"learning_rate": 1.7224715143335235e-08,
"loss": 0.4635,
"step": 1765
},
{
"epoch": 0.976589861751152,
"grad_norm": 0.31344956159591675,
"learning_rate": 1.643331533490333e-08,
"loss": 0.434,
"step": 1766
},
{
"epoch": 0.9771428571428571,
"grad_norm": 0.2932872176170349,
"learning_rate": 1.5660496342291833e-08,
"loss": 0.4707,
"step": 1767
},
{
"epoch": 0.9776958525345623,
"grad_norm": 0.29183509945869446,
"learning_rate": 1.4906261046892523e-08,
"loss": 0.4711,
"step": 1768
},
{
"epoch": 0.9782488479262673,
"grad_norm": 0.29039350152015686,
"learning_rate": 1.4170612260808736e-08,
"loss": 0.4802,
"step": 1769
},
{
"epoch": 0.9788018433179724,
"grad_norm": 0.30380529165267944,
"learning_rate": 1.3453552726847008e-08,
"loss": 0.4868,
"step": 1770
},
{
"epoch": 0.9793548387096774,
"grad_norm": 0.29868564009666443,
"learning_rate": 1.2755085118503762e-08,
"loss": 0.4625,
"step": 1771
},
{
"epoch": 0.9799078341013825,
"grad_norm": 0.30664679408073425,
"learning_rate": 1.207521203995754e-08,
"loss": 0.4369,
"step": 1772
},
{
"epoch": 0.9804608294930875,
"grad_norm": 0.28604456782341003,
"learning_rate": 1.1413936026059558e-08,
"loss": 0.4559,
"step": 1773
},
{
"epoch": 0.9810138248847926,
"grad_norm": 0.33186855912208557,
"learning_rate": 1.0771259542322055e-08,
"loss": 0.4735,
"step": 1774
},
{
"epoch": 0.9815668202764977,
"grad_norm": 0.3165326714515686,
"learning_rate": 1.0147184984911073e-08,
"loss": 0.4411,
"step": 1775
},
{
"epoch": 0.9821198156682027,
"grad_norm": 0.30226629972457886,
"learning_rate": 9.541714680637582e-09,
"loss": 0.4274,
"step": 1776
},
{
"epoch": 0.9826728110599078,
"grad_norm": 0.3025604486465454,
"learning_rate": 8.95485088694692e-09,
"loss": 0.4461,
"step": 1777
},
{
"epoch": 0.983225806451613,
"grad_norm": 0.3139297068119049,
"learning_rate": 8.386595791912145e-09,
"loss": 0.4501,
"step": 1778
},
{
"epoch": 0.983778801843318,
"grad_norm": 0.28405094146728516,
"learning_rate": 7.836951514225144e-09,
"loss": 0.4686,
"step": 1779
},
{
"epoch": 0.9843317972350231,
"grad_norm": 0.2899510860443115,
"learning_rate": 7.3059201031899786e-09,
"loss": 0.4738,
"step": 1780
},
{
"epoch": 0.9848847926267281,
"grad_norm": 0.3046450614929199,
"learning_rate": 6.7935035387128865e-09,
"loss": 0.4531,
"step": 1781
},
{
"epoch": 0.9854377880184332,
"grad_norm": 0.2701072692871094,
"learning_rate": 6.299703731296181e-09,
"loss": 0.4532,
"step": 1782
},
{
"epoch": 0.9859907834101382,
"grad_norm": 0.271697461605072,
"learning_rate": 5.8245225220321385e-09,
"loss": 0.4349,
"step": 1783
},
{
"epoch": 0.9865437788018433,
"grad_norm": 0.29448968172073364,
"learning_rate": 5.367961682594125e-09,
"loss": 0.4664,
"step": 1784
},
{
"epoch": 0.9870967741935484,
"grad_norm": 0.3071184754371643,
"learning_rate": 4.930022915231591e-09,
"loss": 0.4397,
"step": 1785
},
{
"epoch": 0.9876497695852534,
"grad_norm": 0.2995312213897705,
"learning_rate": 4.510707852762864e-09,
"loss": 0.4592,
"step": 1786
},
{
"epoch": 0.9882027649769585,
"grad_norm": 0.30661311745643616,
"learning_rate": 4.110018058570142e-09,
"loss": 0.4835,
"step": 1787
},
{
"epoch": 0.9887557603686636,
"grad_norm": 0.2768559157848358,
"learning_rate": 3.727955026591179e-09,
"loss": 0.4506,
"step": 1788
},
{
"epoch": 0.9893087557603687,
"grad_norm": 0.3139650821685791,
"learning_rate": 3.3645201813170546e-09,
"loss": 0.4483,
"step": 1789
},
{
"epoch": 0.9898617511520738,
"grad_norm": 0.318562775850296,
"learning_rate": 3.0197148777838524e-09,
"loss": 0.4669,
"step": 1790
},
{
"epoch": 0.9904147465437788,
"grad_norm": 0.31068408489227295,
"learning_rate": 2.693540401569883e-09,
"loss": 0.4519,
"step": 1791
},
{
"epoch": 0.9909677419354839,
"grad_norm": 0.30719509720802307,
"learning_rate": 2.3859979687901326e-09,
"loss": 0.4284,
"step": 1792
},
{
"epoch": 0.9915207373271889,
"grad_norm": 0.28648641705513,
"learning_rate": 2.0970887260907123e-09,
"loss": 0.4556,
"step": 1793
},
{
"epoch": 0.992073732718894,
"grad_norm": 0.2835429608821869,
"learning_rate": 1.8268137506455286e-09,
"loss": 0.4764,
"step": 1794
},
{
"epoch": 0.992626728110599,
"grad_norm": 0.3018244206905365,
"learning_rate": 1.57517405015295e-09,
"loss": 0.4505,
"step": 1795
},
{
"epoch": 0.9931797235023041,
"grad_norm": 0.30132514238357544,
"learning_rate": 1.3421705628302584e-09,
"loss": 0.4686,
"step": 1796
},
{
"epoch": 0.9937327188940093,
"grad_norm": 0.292498379945755,
"learning_rate": 1.1278041574125376e-09,
"loss": 0.4295,
"step": 1797
},
{
"epoch": 0.9942857142857143,
"grad_norm": 0.2868688702583313,
"learning_rate": 9.320756331465675e-10,
"loss": 0.4246,
"step": 1798
},
{
"epoch": 0.9948387096774194,
"grad_norm": 0.2882429361343384,
"learning_rate": 7.549857197897137e-10,
"loss": 0.4539,
"step": 1799
},
{
"epoch": 0.9953917050691244,
"grad_norm": 0.2899753749370575,
"learning_rate": 5.965350776071521e-10,
"loss": 0.4748,
"step": 1800
},
{
"epoch": 0.9959447004608295,
"grad_norm": 0.293375700712204,
"learning_rate": 4.567242973696484e-10,
"loss": 0.4613,
"step": 1801
},
{
"epoch": 0.9964976958525346,
"grad_norm": 0.2795793116092682,
"learning_rate": 3.3555390034967214e-10,
"loss": 0.4732,
"step": 1802
},
{
"epoch": 0.9970506912442396,
"grad_norm": 0.29951760172843933,
"learning_rate": 2.330243383208419e-10,
"loss": 0.4612,
"step": 1803
},
{
"epoch": 0.9976036866359447,
"grad_norm": 0.2993388772010803,
"learning_rate": 1.4913599355625975e-10,
"loss": 0.4653,
"step": 1804
},
{
"epoch": 0.9981566820276497,
"grad_norm": 0.31436586380004883,
"learning_rate": 8.388917882684589e-11,
"loss": 0.4712,
"step": 1805
},
{
"epoch": 0.9987096774193548,
"grad_norm": 0.2935016453266144,
"learning_rate": 3.728413739967351e-11,
"loss": 0.4381,
"step": 1806
},
{
"epoch": 0.99926267281106,
"grad_norm": 0.29614418745040894,
"learning_rate": 9.321043037968657e-12,
"loss": 0.4752,
"step": 1807
},
{
"epoch": 0.999815668202765,
"grad_norm": 0.294443815946579,
"learning_rate": 0.0,
"loss": 0.4617,
"step": 1808
},
{
"epoch": 0.999815668202765,
"step": 1808,
"total_flos": 2730506599792640.0,
"train_loss": 0.488344039471276,
"train_runtime": 63700.3598,
"train_samples_per_second": 2.725,
"train_steps_per_second": 0.028
}
],
"logging_steps": 1.0,
"max_steps": 1808,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 2730506599792640.0,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}