diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,12698 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.999815668202765, + "eval_steps": 500, + "global_step": 1808, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0005529953917050691, + "grad_norm": 5.9790191650390625, + "learning_rate": 5.524861878453039e-08, + "loss": 0.8575, + "step": 1 + }, + { + "epoch": 0.0011059907834101382, + "grad_norm": 5.738432884216309, + "learning_rate": 1.1049723756906078e-07, + "loss": 0.8535, + "step": 2 + }, + { + "epoch": 0.0016589861751152074, + "grad_norm": 5.905203342437744, + "learning_rate": 1.6574585635359117e-07, + "loss": 0.8907, + "step": 3 + }, + { + "epoch": 0.0022119815668202765, + "grad_norm": 5.97217321395874, + "learning_rate": 2.2099447513812156e-07, + "loss": 0.8695, + "step": 4 + }, + { + "epoch": 0.0027649769585253456, + "grad_norm": 6.003969669342041, + "learning_rate": 2.7624309392265196e-07, + "loss": 0.8746, + "step": 5 + }, + { + "epoch": 0.0033179723502304147, + "grad_norm": 5.878486156463623, + "learning_rate": 3.3149171270718233e-07, + "loss": 0.866, + "step": 6 + }, + { + "epoch": 0.003870967741935484, + "grad_norm": 5.74072265625, + "learning_rate": 3.867403314917127e-07, + "loss": 0.8351, + "step": 7 + }, + { + "epoch": 0.004423963133640553, + "grad_norm": 5.6800150871276855, + "learning_rate": 4.419889502762431e-07, + "loss": 0.8534, + "step": 8 + }, + { + "epoch": 0.004976958525345622, + "grad_norm": 5.634627819061279, + "learning_rate": 4.972375690607735e-07, + "loss": 0.8592, + "step": 9 + }, + { + "epoch": 0.005529953917050691, + "grad_norm": 5.71673059463501, + "learning_rate": 5.524861878453039e-07, + "loss": 0.8825, + "step": 10 + }, + { + "epoch": 0.00608294930875576, + "grad_norm": 5.533100605010986, + "learning_rate": 6.077348066298343e-07, + "loss": 0.8662, + "step": 11 + }, + { + "epoch": 0.0066359447004608295, + "grad_norm": 5.378372669219971, + "learning_rate": 6.629834254143647e-07, + "loss": 0.8688, + "step": 12 + }, + { + "epoch": 0.007188940092165899, + "grad_norm": 5.282576560974121, + "learning_rate": 7.18232044198895e-07, + "loss": 0.8179, + "step": 13 + }, + { + "epoch": 0.007741935483870968, + "grad_norm": 4.443999290466309, + "learning_rate": 7.734806629834254e-07, + "loss": 0.829, + "step": 14 + }, + { + "epoch": 0.008294930875576038, + "grad_norm": 4.481590270996094, + "learning_rate": 8.287292817679559e-07, + "loss": 0.8441, + "step": 15 + }, + { + "epoch": 0.008847926267281106, + "grad_norm": 4.04002571105957, + "learning_rate": 8.839779005524863e-07, + "loss": 0.7884, + "step": 16 + }, + { + "epoch": 0.009400921658986176, + "grad_norm": 3.9912734031677246, + "learning_rate": 9.392265193370166e-07, + "loss": 0.7851, + "step": 17 + }, + { + "epoch": 0.009953917050691244, + "grad_norm": 3.7548842430114746, + "learning_rate": 9.94475138121547e-07, + "loss": 0.7836, + "step": 18 + }, + { + "epoch": 0.010506912442396314, + "grad_norm": 2.2732059955596924, + "learning_rate": 1.0497237569060774e-06, + "loss": 0.7572, + "step": 19 + }, + { + "epoch": 0.011059907834101382, + "grad_norm": 2.244140625, + "learning_rate": 1.1049723756906078e-06, + "loss": 0.7637, + "step": 20 + }, + { + "epoch": 0.011612903225806452, + "grad_norm": 2.1460845470428467, + "learning_rate": 1.160220994475138e-06, + "loss": 0.7849, + "step": 21 + }, + { + "epoch": 0.01216589861751152, + "grad_norm": 2.0333800315856934, + "learning_rate": 1.2154696132596686e-06, + "loss": 0.8008, + "step": 22 + }, + { + "epoch": 0.01271889400921659, + "grad_norm": 1.962172508239746, + "learning_rate": 1.270718232044199e-06, + "loss": 0.7914, + "step": 23 + }, + { + "epoch": 0.013271889400921659, + "grad_norm": 1.9014993906021118, + "learning_rate": 1.3259668508287293e-06, + "loss": 0.754, + "step": 24 + }, + { + "epoch": 0.013824884792626729, + "grad_norm": 1.5539250373840332, + "learning_rate": 1.3812154696132598e-06, + "loss": 0.7218, + "step": 25 + }, + { + "epoch": 0.014377880184331797, + "grad_norm": 2.1412103176116943, + "learning_rate": 1.43646408839779e-06, + "loss": 0.7083, + "step": 26 + }, + { + "epoch": 0.014930875576036867, + "grad_norm": 2.793086290359497, + "learning_rate": 1.4917127071823205e-06, + "loss": 0.7505, + "step": 27 + }, + { + "epoch": 0.015483870967741935, + "grad_norm": 2.779872179031372, + "learning_rate": 1.5469613259668508e-06, + "loss": 0.7609, + "step": 28 + }, + { + "epoch": 0.016036866359447004, + "grad_norm": 2.9033734798431396, + "learning_rate": 1.6022099447513815e-06, + "loss": 0.7388, + "step": 29 + }, + { + "epoch": 0.016589861751152075, + "grad_norm": 2.6718666553497314, + "learning_rate": 1.6574585635359118e-06, + "loss": 0.7395, + "step": 30 + }, + { + "epoch": 0.017142857142857144, + "grad_norm": 2.3672306537628174, + "learning_rate": 1.7127071823204422e-06, + "loss": 0.736, + "step": 31 + }, + { + "epoch": 0.017695852534562212, + "grad_norm": 2.0016157627105713, + "learning_rate": 1.7679558011049725e-06, + "loss": 0.7169, + "step": 32 + }, + { + "epoch": 0.01824884792626728, + "grad_norm": 1.755001425743103, + "learning_rate": 1.823204419889503e-06, + "loss": 0.7358, + "step": 33 + }, + { + "epoch": 0.018801843317972352, + "grad_norm": 1.084032416343689, + "learning_rate": 1.8784530386740332e-06, + "loss": 0.654, + "step": 34 + }, + { + "epoch": 0.01935483870967742, + "grad_norm": 1.0225807428359985, + "learning_rate": 1.933701657458564e-06, + "loss": 0.6793, + "step": 35 + }, + { + "epoch": 0.01990783410138249, + "grad_norm": 1.0972223281860352, + "learning_rate": 1.988950276243094e-06, + "loss": 0.6625, + "step": 36 + }, + { + "epoch": 0.020460829493087557, + "grad_norm": 1.1796579360961914, + "learning_rate": 2.0441988950276245e-06, + "loss": 0.6905, + "step": 37 + }, + { + "epoch": 0.02101382488479263, + "grad_norm": 1.1285828351974487, + "learning_rate": 2.0994475138121547e-06, + "loss": 0.6433, + "step": 38 + }, + { + "epoch": 0.021566820276497697, + "grad_norm": 1.0303235054016113, + "learning_rate": 2.1546961325966854e-06, + "loss": 0.637, + "step": 39 + }, + { + "epoch": 0.022119815668202765, + "grad_norm": 0.9121658205986023, + "learning_rate": 2.2099447513812157e-06, + "loss": 0.6506, + "step": 40 + }, + { + "epoch": 0.022672811059907833, + "grad_norm": 0.8387944102287292, + "learning_rate": 2.265193370165746e-06, + "loss": 0.6217, + "step": 41 + }, + { + "epoch": 0.023225806451612905, + "grad_norm": 0.762744665145874, + "learning_rate": 2.320441988950276e-06, + "loss": 0.6131, + "step": 42 + }, + { + "epoch": 0.023778801843317973, + "grad_norm": 0.8466107845306396, + "learning_rate": 2.375690607734807e-06, + "loss": 0.6313, + "step": 43 + }, + { + "epoch": 0.02433179723502304, + "grad_norm": 0.7411535382270813, + "learning_rate": 2.430939226519337e-06, + "loss": 0.6178, + "step": 44 + }, + { + "epoch": 0.02488479262672811, + "grad_norm": 0.7712140083312988, + "learning_rate": 2.486187845303868e-06, + "loss": 0.6218, + "step": 45 + }, + { + "epoch": 0.02543778801843318, + "grad_norm": 1.0630971193313599, + "learning_rate": 2.541436464088398e-06, + "loss": 0.6497, + "step": 46 + }, + { + "epoch": 0.02599078341013825, + "grad_norm": 0.8186689615249634, + "learning_rate": 2.5966850828729284e-06, + "loss": 0.6493, + "step": 47 + }, + { + "epoch": 0.026543778801843318, + "grad_norm": 0.7290191650390625, + "learning_rate": 2.6519337016574586e-06, + "loss": 0.6278, + "step": 48 + }, + { + "epoch": 0.027096774193548386, + "grad_norm": 0.6419755816459656, + "learning_rate": 2.707182320441989e-06, + "loss": 0.6495, + "step": 49 + }, + { + "epoch": 0.027649769585253458, + "grad_norm": 0.6034302711486816, + "learning_rate": 2.7624309392265196e-06, + "loss": 0.6262, + "step": 50 + }, + { + "epoch": 0.028202764976958526, + "grad_norm": 0.7188774943351746, + "learning_rate": 2.81767955801105e-06, + "loss": 0.6416, + "step": 51 + }, + { + "epoch": 0.028755760368663594, + "grad_norm": 0.7371018528938293, + "learning_rate": 2.87292817679558e-06, + "loss": 0.6181, + "step": 52 + }, + { + "epoch": 0.029308755760368663, + "grad_norm": 0.7526310086250305, + "learning_rate": 2.9281767955801104e-06, + "loss": 0.6644, + "step": 53 + }, + { + "epoch": 0.029861751152073734, + "grad_norm": 0.6002934575080872, + "learning_rate": 2.983425414364641e-06, + "loss": 0.57, + "step": 54 + }, + { + "epoch": 0.030414746543778803, + "grad_norm": 0.535647451877594, + "learning_rate": 3.0386740331491713e-06, + "loss": 0.6158, + "step": 55 + }, + { + "epoch": 0.03096774193548387, + "grad_norm": 0.5333572030067444, + "learning_rate": 3.0939226519337016e-06, + "loss": 0.6237, + "step": 56 + }, + { + "epoch": 0.03152073732718894, + "grad_norm": 0.6033058166503906, + "learning_rate": 3.149171270718232e-06, + "loss": 0.5905, + "step": 57 + }, + { + "epoch": 0.03207373271889401, + "grad_norm": 0.5824403166770935, + "learning_rate": 3.204419889502763e-06, + "loss": 0.5934, + "step": 58 + }, + { + "epoch": 0.03262672811059908, + "grad_norm": 0.5389731526374817, + "learning_rate": 3.2596685082872933e-06, + "loss": 0.6149, + "step": 59 + }, + { + "epoch": 0.03317972350230415, + "grad_norm": 0.5317662954330444, + "learning_rate": 3.3149171270718235e-06, + "loss": 0.5953, + "step": 60 + }, + { + "epoch": 0.033732718894009216, + "grad_norm": 0.5464670658111572, + "learning_rate": 3.370165745856354e-06, + "loss": 0.5848, + "step": 61 + }, + { + "epoch": 0.03428571428571429, + "grad_norm": 0.5087230205535889, + "learning_rate": 3.4254143646408845e-06, + "loss": 0.5912, + "step": 62 + }, + { + "epoch": 0.03483870967741935, + "grad_norm": 0.4756256937980652, + "learning_rate": 3.4806629834254147e-06, + "loss": 0.5842, + "step": 63 + }, + { + "epoch": 0.035391705069124424, + "grad_norm": 0.49870121479034424, + "learning_rate": 3.535911602209945e-06, + "loss": 0.6009, + "step": 64 + }, + { + "epoch": 0.035944700460829496, + "grad_norm": 0.44439196586608887, + "learning_rate": 3.5911602209944757e-06, + "loss": 0.5832, + "step": 65 + }, + { + "epoch": 0.03649769585253456, + "grad_norm": 0.4260464608669281, + "learning_rate": 3.646408839779006e-06, + "loss": 0.5896, + "step": 66 + }, + { + "epoch": 0.03705069124423963, + "grad_norm": 0.4569395184516907, + "learning_rate": 3.7016574585635362e-06, + "loss": 0.6184, + "step": 67 + }, + { + "epoch": 0.037603686635944704, + "grad_norm": 0.4692537486553192, + "learning_rate": 3.7569060773480665e-06, + "loss": 0.5623, + "step": 68 + }, + { + "epoch": 0.03815668202764977, + "grad_norm": 0.40080568194389343, + "learning_rate": 3.812154696132597e-06, + "loss": 0.5832, + "step": 69 + }, + { + "epoch": 0.03870967741935484, + "grad_norm": 0.47561773657798767, + "learning_rate": 3.867403314917128e-06, + "loss": 0.5801, + "step": 70 + }, + { + "epoch": 0.039262672811059905, + "grad_norm": 0.47063136100769043, + "learning_rate": 3.922651933701658e-06, + "loss": 0.5913, + "step": 71 + }, + { + "epoch": 0.03981566820276498, + "grad_norm": 0.435703843832016, + "learning_rate": 3.977900552486188e-06, + "loss": 0.583, + "step": 72 + }, + { + "epoch": 0.04036866359447005, + "grad_norm": 0.44961071014404297, + "learning_rate": 4.033149171270719e-06, + "loss": 0.6015, + "step": 73 + }, + { + "epoch": 0.04092165898617511, + "grad_norm": 0.43612730503082275, + "learning_rate": 4.088397790055249e-06, + "loss": 0.5916, + "step": 74 + }, + { + "epoch": 0.041474654377880185, + "grad_norm": 0.4045799970626831, + "learning_rate": 4.143646408839779e-06, + "loss": 0.5365, + "step": 75 + }, + { + "epoch": 0.04202764976958526, + "grad_norm": 0.47303467988967896, + "learning_rate": 4.1988950276243095e-06, + "loss": 0.587, + "step": 76 + }, + { + "epoch": 0.04258064516129032, + "grad_norm": 0.3942665159702301, + "learning_rate": 4.2541436464088406e-06, + "loss": 0.5682, + "step": 77 + }, + { + "epoch": 0.04313364055299539, + "grad_norm": 0.40599432587623596, + "learning_rate": 4.309392265193371e-06, + "loss": 0.5686, + "step": 78 + }, + { + "epoch": 0.04368663594470046, + "grad_norm": 0.41502925753593445, + "learning_rate": 4.364640883977901e-06, + "loss": 0.5644, + "step": 79 + }, + { + "epoch": 0.04423963133640553, + "grad_norm": 0.4402730166912079, + "learning_rate": 4.419889502762431e-06, + "loss": 0.5458, + "step": 80 + }, + { + "epoch": 0.0447926267281106, + "grad_norm": 0.4110409915447235, + "learning_rate": 4.475138121546962e-06, + "loss": 0.5447, + "step": 81 + }, + { + "epoch": 0.045345622119815666, + "grad_norm": 0.46120530366897583, + "learning_rate": 4.530386740331492e-06, + "loss": 0.584, + "step": 82 + }, + { + "epoch": 0.04589861751152074, + "grad_norm": 0.4089435040950775, + "learning_rate": 4.585635359116022e-06, + "loss": 0.555, + "step": 83 + }, + { + "epoch": 0.04645161290322581, + "grad_norm": 0.4719107449054718, + "learning_rate": 4.640883977900552e-06, + "loss": 0.5714, + "step": 84 + }, + { + "epoch": 0.047004608294930875, + "grad_norm": 0.45129814743995667, + "learning_rate": 4.6961325966850835e-06, + "loss": 0.5443, + "step": 85 + }, + { + "epoch": 0.047557603686635946, + "grad_norm": 0.4321492612361908, + "learning_rate": 4.751381215469614e-06, + "loss": 0.5624, + "step": 86 + }, + { + "epoch": 0.04811059907834101, + "grad_norm": 0.4300953149795532, + "learning_rate": 4.806629834254144e-06, + "loss": 0.5497, + "step": 87 + }, + { + "epoch": 0.04866359447004608, + "grad_norm": 0.4590497314929962, + "learning_rate": 4.861878453038674e-06, + "loss": 0.5652, + "step": 88 + }, + { + "epoch": 0.049216589861751155, + "grad_norm": 0.47058796882629395, + "learning_rate": 4.9171270718232054e-06, + "loss": 0.5768, + "step": 89 + }, + { + "epoch": 0.04976958525345622, + "grad_norm": 0.43738317489624023, + "learning_rate": 4.972375690607736e-06, + "loss": 0.58, + "step": 90 + }, + { + "epoch": 0.05032258064516129, + "grad_norm": 0.4055286943912506, + "learning_rate": 5.027624309392266e-06, + "loss": 0.5503, + "step": 91 + }, + { + "epoch": 0.05087557603686636, + "grad_norm": 0.44679558277130127, + "learning_rate": 5.082872928176796e-06, + "loss": 0.5347, + "step": 92 + }, + { + "epoch": 0.05142857142857143, + "grad_norm": 0.46003517508506775, + "learning_rate": 5.1381215469613265e-06, + "loss": 0.5854, + "step": 93 + }, + { + "epoch": 0.0519815668202765, + "grad_norm": 0.44815686345100403, + "learning_rate": 5.193370165745857e-06, + "loss": 0.5316, + "step": 94 + }, + { + "epoch": 0.052534562211981564, + "grad_norm": 0.4559434950351715, + "learning_rate": 5.248618784530387e-06, + "loss": 0.5724, + "step": 95 + }, + { + "epoch": 0.053087557603686636, + "grad_norm": 0.45906394720077515, + "learning_rate": 5.303867403314917e-06, + "loss": 0.565, + "step": 96 + }, + { + "epoch": 0.05364055299539171, + "grad_norm": 0.3892553448677063, + "learning_rate": 5.3591160220994476e-06, + "loss": 0.5397, + "step": 97 + }, + { + "epoch": 0.05419354838709677, + "grad_norm": 0.4995177090167999, + "learning_rate": 5.414364640883978e-06, + "loss": 0.5509, + "step": 98 + }, + { + "epoch": 0.054746543778801844, + "grad_norm": 0.4675018787384033, + "learning_rate": 5.469613259668509e-06, + "loss": 0.5432, + "step": 99 + }, + { + "epoch": 0.055299539170506916, + "grad_norm": 0.4121154844760895, + "learning_rate": 5.524861878453039e-06, + "loss": 0.5463, + "step": 100 + }, + { + "epoch": 0.05585253456221198, + "grad_norm": 0.4763623774051666, + "learning_rate": 5.5801104972375695e-06, + "loss": 0.5673, + "step": 101 + }, + { + "epoch": 0.05640552995391705, + "grad_norm": 0.44080275297164917, + "learning_rate": 5.6353591160221e-06, + "loss": 0.5684, + "step": 102 + }, + { + "epoch": 0.05695852534562212, + "grad_norm": 0.4623638391494751, + "learning_rate": 5.69060773480663e-06, + "loss": 0.5738, + "step": 103 + }, + { + "epoch": 0.05751152073732719, + "grad_norm": 0.379423588514328, + "learning_rate": 5.74585635359116e-06, + "loss": 0.5503, + "step": 104 + }, + { + "epoch": 0.05806451612903226, + "grad_norm": 0.45095521211624146, + "learning_rate": 5.8011049723756905e-06, + "loss": 0.5683, + "step": 105 + }, + { + "epoch": 0.058617511520737325, + "grad_norm": 0.4157678484916687, + "learning_rate": 5.856353591160221e-06, + "loss": 0.5459, + "step": 106 + }, + { + "epoch": 0.0591705069124424, + "grad_norm": 0.4484108090400696, + "learning_rate": 5.911602209944752e-06, + "loss": 0.5465, + "step": 107 + }, + { + "epoch": 0.05972350230414747, + "grad_norm": 0.41227927803993225, + "learning_rate": 5.966850828729282e-06, + "loss": 0.5737, + "step": 108 + }, + { + "epoch": 0.060276497695852534, + "grad_norm": 0.4771617352962494, + "learning_rate": 6.0220994475138124e-06, + "loss": 0.5295, + "step": 109 + }, + { + "epoch": 0.060829493087557605, + "grad_norm": 0.42870959639549255, + "learning_rate": 6.077348066298343e-06, + "loss": 0.5432, + "step": 110 + }, + { + "epoch": 0.06138248847926267, + "grad_norm": 0.47716712951660156, + "learning_rate": 6.132596685082873e-06, + "loss": 0.5695, + "step": 111 + }, + { + "epoch": 0.06193548387096774, + "grad_norm": 0.43211421370506287, + "learning_rate": 6.187845303867403e-06, + "loss": 0.5408, + "step": 112 + }, + { + "epoch": 0.062488479262672814, + "grad_norm": 0.49489325284957886, + "learning_rate": 6.2430939226519335e-06, + "loss": 0.5488, + "step": 113 + }, + { + "epoch": 0.06304147465437789, + "grad_norm": 0.4600902199745178, + "learning_rate": 6.298342541436464e-06, + "loss": 0.5445, + "step": 114 + }, + { + "epoch": 0.06359447004608294, + "grad_norm": 0.41128382086753845, + "learning_rate": 6.353591160220996e-06, + "loss": 0.5507, + "step": 115 + }, + { + "epoch": 0.06414746543778801, + "grad_norm": 0.38756921887397766, + "learning_rate": 6.408839779005526e-06, + "loss": 0.5072, + "step": 116 + }, + { + "epoch": 0.06470046082949309, + "grad_norm": 0.4521966278553009, + "learning_rate": 6.464088397790056e-06, + "loss": 0.5597, + "step": 117 + }, + { + "epoch": 0.06525345622119816, + "grad_norm": 0.39012429118156433, + "learning_rate": 6.5193370165745865e-06, + "loss": 0.5017, + "step": 118 + }, + { + "epoch": 0.06580645161290323, + "grad_norm": 0.4543527364730835, + "learning_rate": 6.574585635359117e-06, + "loss": 0.5342, + "step": 119 + }, + { + "epoch": 0.0663594470046083, + "grad_norm": 0.4338400065898895, + "learning_rate": 6.629834254143647e-06, + "loss": 0.5489, + "step": 120 + }, + { + "epoch": 0.06691244239631336, + "grad_norm": 0.461480975151062, + "learning_rate": 6.685082872928177e-06, + "loss": 0.5327, + "step": 121 + }, + { + "epoch": 0.06746543778801843, + "grad_norm": 0.4241288900375366, + "learning_rate": 6.740331491712708e-06, + "loss": 0.5297, + "step": 122 + }, + { + "epoch": 0.0680184331797235, + "grad_norm": 0.5540463924407959, + "learning_rate": 6.795580110497239e-06, + "loss": 0.5262, + "step": 123 + }, + { + "epoch": 0.06857142857142857, + "grad_norm": 0.4794096350669861, + "learning_rate": 6.850828729281769e-06, + "loss": 0.5638, + "step": 124 + }, + { + "epoch": 0.06912442396313365, + "grad_norm": 0.43071404099464417, + "learning_rate": 6.906077348066299e-06, + "loss": 0.5254, + "step": 125 + }, + { + "epoch": 0.0696774193548387, + "grad_norm": 0.4853145480155945, + "learning_rate": 6.9613259668508295e-06, + "loss": 0.5524, + "step": 126 + }, + { + "epoch": 0.07023041474654378, + "grad_norm": 0.47362276911735535, + "learning_rate": 7.01657458563536e-06, + "loss": 0.5469, + "step": 127 + }, + { + "epoch": 0.07078341013824885, + "grad_norm": 0.45325443148612976, + "learning_rate": 7.07182320441989e-06, + "loss": 0.5476, + "step": 128 + }, + { + "epoch": 0.07133640552995392, + "grad_norm": 0.43244412541389465, + "learning_rate": 7.12707182320442e-06, + "loss": 0.5367, + "step": 129 + }, + { + "epoch": 0.07188940092165899, + "grad_norm": 0.5289068222045898, + "learning_rate": 7.182320441988951e-06, + "loss": 0.5796, + "step": 130 + }, + { + "epoch": 0.07244239631336405, + "grad_norm": 0.4485274851322174, + "learning_rate": 7.237569060773482e-06, + "loss": 0.5396, + "step": 131 + }, + { + "epoch": 0.07299539170506912, + "grad_norm": 0.42793214321136475, + "learning_rate": 7.292817679558012e-06, + "loss": 0.5532, + "step": 132 + }, + { + "epoch": 0.07354838709677419, + "grad_norm": 0.4250226318836212, + "learning_rate": 7.348066298342542e-06, + "loss": 0.5847, + "step": 133 + }, + { + "epoch": 0.07410138248847926, + "grad_norm": 0.44791412353515625, + "learning_rate": 7.4033149171270724e-06, + "loss": 0.5216, + "step": 134 + }, + { + "epoch": 0.07465437788018434, + "grad_norm": 0.48473551869392395, + "learning_rate": 7.458563535911603e-06, + "loss": 0.5434, + "step": 135 + }, + { + "epoch": 0.07520737327188941, + "grad_norm": 0.5042223334312439, + "learning_rate": 7.513812154696133e-06, + "loss": 0.5292, + "step": 136 + }, + { + "epoch": 0.07576036866359447, + "grad_norm": 0.5112206935882568, + "learning_rate": 7.569060773480663e-06, + "loss": 0.573, + "step": 137 + }, + { + "epoch": 0.07631336405529954, + "grad_norm": 0.4625641107559204, + "learning_rate": 7.624309392265194e-06, + "loss": 0.5656, + "step": 138 + }, + { + "epoch": 0.07686635944700461, + "grad_norm": 0.507746160030365, + "learning_rate": 7.679558011049725e-06, + "loss": 0.5352, + "step": 139 + }, + { + "epoch": 0.07741935483870968, + "grad_norm": 0.4666261672973633, + "learning_rate": 7.734806629834256e-06, + "loss": 0.5465, + "step": 140 + }, + { + "epoch": 0.07797235023041475, + "grad_norm": 0.44455286860466003, + "learning_rate": 7.790055248618785e-06, + "loss": 0.5395, + "step": 141 + }, + { + "epoch": 0.07852534562211981, + "grad_norm": 0.4286342263221741, + "learning_rate": 7.845303867403316e-06, + "loss": 0.5045, + "step": 142 + }, + { + "epoch": 0.07907834101382488, + "grad_norm": 0.4748035669326782, + "learning_rate": 7.900552486187846e-06, + "loss": 0.552, + "step": 143 + }, + { + "epoch": 0.07963133640552995, + "grad_norm": 0.44316938519477844, + "learning_rate": 7.955801104972377e-06, + "loss": 0.5394, + "step": 144 + }, + { + "epoch": 0.08018433179723503, + "grad_norm": 0.4846290647983551, + "learning_rate": 8.011049723756906e-06, + "loss": 0.5267, + "step": 145 + }, + { + "epoch": 0.0807373271889401, + "grad_norm": 0.4428083300590515, + "learning_rate": 8.066298342541437e-06, + "loss": 0.4941, + "step": 146 + }, + { + "epoch": 0.08129032258064516, + "grad_norm": 0.44677111506462097, + "learning_rate": 8.121546961325968e-06, + "loss": 0.4885, + "step": 147 + }, + { + "epoch": 0.08184331797235023, + "grad_norm": 0.4421268105506897, + "learning_rate": 8.176795580110498e-06, + "loss": 0.5765, + "step": 148 + }, + { + "epoch": 0.0823963133640553, + "grad_norm": 0.4853808581829071, + "learning_rate": 8.232044198895029e-06, + "loss": 0.5297, + "step": 149 + }, + { + "epoch": 0.08294930875576037, + "grad_norm": 0.47924932837486267, + "learning_rate": 8.287292817679558e-06, + "loss": 0.5192, + "step": 150 + }, + { + "epoch": 0.08350230414746544, + "grad_norm": 0.5630611181259155, + "learning_rate": 8.34254143646409e-06, + "loss": 0.5439, + "step": 151 + }, + { + "epoch": 0.08405529953917051, + "grad_norm": 0.6081127524375916, + "learning_rate": 8.397790055248619e-06, + "loss": 0.5012, + "step": 152 + }, + { + "epoch": 0.08460829493087557, + "grad_norm": 0.44527262449264526, + "learning_rate": 8.45303867403315e-06, + "loss": 0.5305, + "step": 153 + }, + { + "epoch": 0.08516129032258064, + "grad_norm": 0.4948616623878479, + "learning_rate": 8.508287292817681e-06, + "loss": 0.5109, + "step": 154 + }, + { + "epoch": 0.08571428571428572, + "grad_norm": 0.5024743676185608, + "learning_rate": 8.56353591160221e-06, + "loss": 0.5632, + "step": 155 + }, + { + "epoch": 0.08626728110599079, + "grad_norm": 0.45424172282218933, + "learning_rate": 8.618784530386742e-06, + "loss": 0.5029, + "step": 156 + }, + { + "epoch": 0.08682027649769586, + "grad_norm": 0.5445595979690552, + "learning_rate": 8.674033149171271e-06, + "loss": 0.5451, + "step": 157 + }, + { + "epoch": 0.08737327188940092, + "grad_norm": 0.495613157749176, + "learning_rate": 8.729281767955802e-06, + "loss": 0.527, + "step": 158 + }, + { + "epoch": 0.08792626728110599, + "grad_norm": 0.6915062665939331, + "learning_rate": 8.784530386740332e-06, + "loss": 0.5378, + "step": 159 + }, + { + "epoch": 0.08847926267281106, + "grad_norm": 0.4711109697818756, + "learning_rate": 8.839779005524863e-06, + "loss": 0.5105, + "step": 160 + }, + { + "epoch": 0.08903225806451613, + "grad_norm": 0.5774613618850708, + "learning_rate": 8.895027624309392e-06, + "loss": 0.5268, + "step": 161 + }, + { + "epoch": 0.0895852534562212, + "grad_norm": 0.534370481967926, + "learning_rate": 8.950276243093923e-06, + "loss": 0.5275, + "step": 162 + }, + { + "epoch": 0.09013824884792626, + "grad_norm": 0.5167669057846069, + "learning_rate": 9.005524861878454e-06, + "loss": 0.5173, + "step": 163 + }, + { + "epoch": 0.09069124423963133, + "grad_norm": 0.5220374464988708, + "learning_rate": 9.060773480662984e-06, + "loss": 0.5117, + "step": 164 + }, + { + "epoch": 0.0912442396313364, + "grad_norm": 0.6151928305625916, + "learning_rate": 9.116022099447515e-06, + "loss": 0.5236, + "step": 165 + }, + { + "epoch": 0.09179723502304148, + "grad_norm": 0.54198157787323, + "learning_rate": 9.171270718232044e-06, + "loss": 0.5113, + "step": 166 + }, + { + "epoch": 0.09235023041474655, + "grad_norm": 0.540729820728302, + "learning_rate": 9.226519337016575e-06, + "loss": 0.5293, + "step": 167 + }, + { + "epoch": 0.09290322580645162, + "grad_norm": 0.4339202642440796, + "learning_rate": 9.281767955801105e-06, + "loss": 0.5105, + "step": 168 + }, + { + "epoch": 0.09345622119815668, + "grad_norm": 0.5427307486534119, + "learning_rate": 9.337016574585636e-06, + "loss": 0.5442, + "step": 169 + }, + { + "epoch": 0.09400921658986175, + "grad_norm": 0.5696178674697876, + "learning_rate": 9.392265193370167e-06, + "loss": 0.5198, + "step": 170 + }, + { + "epoch": 0.09456221198156682, + "grad_norm": 0.475676029920578, + "learning_rate": 9.447513812154696e-06, + "loss": 0.5275, + "step": 171 + }, + { + "epoch": 0.09511520737327189, + "grad_norm": 0.6084780693054199, + "learning_rate": 9.502762430939228e-06, + "loss": 0.5344, + "step": 172 + }, + { + "epoch": 0.09566820276497696, + "grad_norm": 0.45912471413612366, + "learning_rate": 9.558011049723757e-06, + "loss": 0.5128, + "step": 173 + }, + { + "epoch": 0.09622119815668202, + "grad_norm": 0.5522315502166748, + "learning_rate": 9.613259668508288e-06, + "loss": 0.4985, + "step": 174 + }, + { + "epoch": 0.0967741935483871, + "grad_norm": 0.5598754286766052, + "learning_rate": 9.668508287292818e-06, + "loss": 0.5438, + "step": 175 + }, + { + "epoch": 0.09732718894009217, + "grad_norm": 0.4842788577079773, + "learning_rate": 9.723756906077349e-06, + "loss": 0.5098, + "step": 176 + }, + { + "epoch": 0.09788018433179724, + "grad_norm": 0.5201513171195984, + "learning_rate": 9.779005524861878e-06, + "loss": 0.5155, + "step": 177 + }, + { + "epoch": 0.09843317972350231, + "grad_norm": 0.4839669466018677, + "learning_rate": 9.834254143646411e-06, + "loss": 0.5201, + "step": 178 + }, + { + "epoch": 0.09898617511520737, + "grad_norm": 0.5859594941139221, + "learning_rate": 9.88950276243094e-06, + "loss": 0.5603, + "step": 179 + }, + { + "epoch": 0.09953917050691244, + "grad_norm": 0.48163536190986633, + "learning_rate": 9.944751381215471e-06, + "loss": 0.5279, + "step": 180 + }, + { + "epoch": 0.10009216589861751, + "grad_norm": 0.508065938949585, + "learning_rate": 1e-05, + "loss": 0.5186, + "step": 181 + }, + { + "epoch": 0.10064516129032258, + "grad_norm": 0.47829869389533997, + "learning_rate": 9.999990678956964e-06, + "loss": 0.5224, + "step": 182 + }, + { + "epoch": 0.10119815668202765, + "grad_norm": 0.5300391912460327, + "learning_rate": 9.999962715862601e-06, + "loss": 0.5128, + "step": 183 + }, + { + "epoch": 0.10175115207373273, + "grad_norm": 0.47900837659835815, + "learning_rate": 9.999916110821174e-06, + "loss": 0.5232, + "step": 184 + }, + { + "epoch": 0.10230414746543778, + "grad_norm": 0.4351784288883209, + "learning_rate": 9.999850864006444e-06, + "loss": 0.5274, + "step": 185 + }, + { + "epoch": 0.10285714285714286, + "grad_norm": 0.48379045724868774, + "learning_rate": 9.99976697566168e-06, + "loss": 0.4964, + "step": 186 + }, + { + "epoch": 0.10341013824884793, + "grad_norm": 0.4367005228996277, + "learning_rate": 9.999664446099651e-06, + "loss": 0.5311, + "step": 187 + }, + { + "epoch": 0.103963133640553, + "grad_norm": 0.4944308400154114, + "learning_rate": 9.999543275702632e-06, + "loss": 0.5093, + "step": 188 + }, + { + "epoch": 0.10451612903225807, + "grad_norm": 0.44800275564193726, + "learning_rate": 9.999403464922393e-06, + "loss": 0.5124, + "step": 189 + }, + { + "epoch": 0.10506912442396313, + "grad_norm": 0.5029193758964539, + "learning_rate": 9.99924501428021e-06, + "loss": 0.5538, + "step": 190 + }, + { + "epoch": 0.1056221198156682, + "grad_norm": 0.4435194730758667, + "learning_rate": 9.999067924366854e-06, + "loss": 0.5291, + "step": 191 + }, + { + "epoch": 0.10617511520737327, + "grad_norm": 0.45577532052993774, + "learning_rate": 9.998872195842588e-06, + "loss": 0.5107, + "step": 192 + }, + { + "epoch": 0.10672811059907834, + "grad_norm": 0.4373863637447357, + "learning_rate": 9.998657829437171e-06, + "loss": 0.5137, + "step": 193 + }, + { + "epoch": 0.10728110599078342, + "grad_norm": 0.4408453404903412, + "learning_rate": 9.998424825949848e-06, + "loss": 0.5061, + "step": 194 + }, + { + "epoch": 0.10783410138248847, + "grad_norm": 0.5477350950241089, + "learning_rate": 9.998173186249357e-06, + "loss": 0.5363, + "step": 195 + }, + { + "epoch": 0.10838709677419354, + "grad_norm": 0.43939393758773804, + "learning_rate": 9.99790291127391e-06, + "loss": 0.5095, + "step": 196 + }, + { + "epoch": 0.10894009216589862, + "grad_norm": 0.47420209646224976, + "learning_rate": 9.997614002031211e-06, + "loss": 0.5297, + "step": 197 + }, + { + "epoch": 0.10949308755760369, + "grad_norm": 0.45546746253967285, + "learning_rate": 9.99730645959843e-06, + "loss": 0.5139, + "step": 198 + }, + { + "epoch": 0.11004608294930876, + "grad_norm": 0.4920537769794464, + "learning_rate": 9.996980285122218e-06, + "loss": 0.5344, + "step": 199 + }, + { + "epoch": 0.11059907834101383, + "grad_norm": 0.562940776348114, + "learning_rate": 9.996635479818683e-06, + "loss": 0.535, + "step": 200 + }, + { + "epoch": 0.11115207373271889, + "grad_norm": 0.5914273262023926, + "learning_rate": 9.99627204497341e-06, + "loss": 0.5134, + "step": 201 + }, + { + "epoch": 0.11170506912442396, + "grad_norm": 0.4358191192150116, + "learning_rate": 9.995889981941432e-06, + "loss": 0.4886, + "step": 202 + }, + { + "epoch": 0.11225806451612903, + "grad_norm": 0.569436252117157, + "learning_rate": 9.995489292147238e-06, + "loss": 0.5302, + "step": 203 + }, + { + "epoch": 0.1128110599078341, + "grad_norm": 0.5561317801475525, + "learning_rate": 9.995069977084769e-06, + "loss": 0.5214, + "step": 204 + }, + { + "epoch": 0.11336405529953918, + "grad_norm": 0.4526347219944, + "learning_rate": 9.994632038317407e-06, + "loss": 0.5106, + "step": 205 + }, + { + "epoch": 0.11391705069124423, + "grad_norm": 0.6103693246841431, + "learning_rate": 9.99417547747797e-06, + "loss": 0.5187, + "step": 206 + }, + { + "epoch": 0.1144700460829493, + "grad_norm": 0.5255654454231262, + "learning_rate": 9.993700296268705e-06, + "loss": 0.5253, + "step": 207 + }, + { + "epoch": 0.11502304147465438, + "grad_norm": 0.48559436202049255, + "learning_rate": 9.993206496461287e-06, + "loss": 0.5127, + "step": 208 + }, + { + "epoch": 0.11557603686635945, + "grad_norm": 0.5230923891067505, + "learning_rate": 9.992694079896812e-06, + "loss": 0.5078, + "step": 209 + }, + { + "epoch": 0.11612903225806452, + "grad_norm": 0.45681560039520264, + "learning_rate": 9.992163048485776e-06, + "loss": 0.522, + "step": 210 + }, + { + "epoch": 0.11668202764976958, + "grad_norm": 0.510287344455719, + "learning_rate": 9.99161340420809e-06, + "loss": 0.509, + "step": 211 + }, + { + "epoch": 0.11723502304147465, + "grad_norm": 0.4897903501987457, + "learning_rate": 9.991045149113055e-06, + "loss": 0.5118, + "step": 212 + }, + { + "epoch": 0.11778801843317972, + "grad_norm": 0.516163170337677, + "learning_rate": 9.990458285319362e-06, + "loss": 0.4915, + "step": 213 + }, + { + "epoch": 0.1183410138248848, + "grad_norm": 0.5389583110809326, + "learning_rate": 9.98985281501509e-06, + "loss": 0.5192, + "step": 214 + }, + { + "epoch": 0.11889400921658987, + "grad_norm": 0.6268092393875122, + "learning_rate": 9.989228740457679e-06, + "loss": 0.524, + "step": 215 + }, + { + "epoch": 0.11944700460829494, + "grad_norm": 0.447966605424881, + "learning_rate": 9.988586063973942e-06, + "loss": 0.4864, + "step": 216 + }, + { + "epoch": 0.12, + "grad_norm": 0.6052084565162659, + "learning_rate": 9.987924787960043e-06, + "loss": 0.5005, + "step": 217 + }, + { + "epoch": 0.12055299539170507, + "grad_norm": 0.5076256990432739, + "learning_rate": 9.987244914881498e-06, + "loss": 0.5219, + "step": 218 + }, + { + "epoch": 0.12110599078341014, + "grad_norm": 0.5539697408676147, + "learning_rate": 9.986546447273153e-06, + "loss": 0.5024, + "step": 219 + }, + { + "epoch": 0.12165898617511521, + "grad_norm": 0.4377414584159851, + "learning_rate": 9.985829387739192e-06, + "loss": 0.5291, + "step": 220 + }, + { + "epoch": 0.12221198156682028, + "grad_norm": 0.5455598831176758, + "learning_rate": 9.985093738953108e-06, + "loss": 0.5126, + "step": 221 + }, + { + "epoch": 0.12276497695852534, + "grad_norm": 0.462424099445343, + "learning_rate": 9.98433950365771e-06, + "loss": 0.5225, + "step": 222 + }, + { + "epoch": 0.12331797235023041, + "grad_norm": 0.49225398898124695, + "learning_rate": 9.983566684665097e-06, + "loss": 0.5328, + "step": 223 + }, + { + "epoch": 0.12387096774193548, + "grad_norm": 0.5304319262504578, + "learning_rate": 9.982775284856665e-06, + "loss": 0.5212, + "step": 224 + }, + { + "epoch": 0.12442396313364056, + "grad_norm": 0.4655149579048157, + "learning_rate": 9.981965307183081e-06, + "loss": 0.5161, + "step": 225 + }, + { + "epoch": 0.12497695852534563, + "grad_norm": 0.6474348902702332, + "learning_rate": 9.98113675466428e-06, + "loss": 0.4983, + "step": 226 + }, + { + "epoch": 0.12552995391705069, + "grad_norm": 0.43538525700569153, + "learning_rate": 9.980289630389453e-06, + "loss": 0.5286, + "step": 227 + }, + { + "epoch": 0.12608294930875577, + "grad_norm": 0.5731073617935181, + "learning_rate": 9.97942393751703e-06, + "loss": 0.5414, + "step": 228 + }, + { + "epoch": 0.12663594470046083, + "grad_norm": 0.4664866626262665, + "learning_rate": 9.978539679274675e-06, + "loss": 0.5052, + "step": 229 + }, + { + "epoch": 0.1271889400921659, + "grad_norm": 0.5314398407936096, + "learning_rate": 9.977636858959274e-06, + "loss": 0.527, + "step": 230 + }, + { + "epoch": 0.12774193548387097, + "grad_norm": 0.538694441318512, + "learning_rate": 9.976715479936916e-06, + "loss": 0.5094, + "step": 231 + }, + { + "epoch": 0.12829493087557603, + "grad_norm": 0.4854016900062561, + "learning_rate": 9.975775545642889e-06, + "loss": 0.5078, + "step": 232 + }, + { + "epoch": 0.12884792626728112, + "grad_norm": 0.5078600645065308, + "learning_rate": 9.974817059581656e-06, + "loss": 0.5236, + "step": 233 + }, + { + "epoch": 0.12940092165898617, + "grad_norm": 0.455322802066803, + "learning_rate": 9.973840025326858e-06, + "loss": 0.5137, + "step": 234 + }, + { + "epoch": 0.12995391705069123, + "grad_norm": 0.5216564536094666, + "learning_rate": 9.972844446521281e-06, + "loss": 0.5289, + "step": 235 + }, + { + "epoch": 0.13050691244239632, + "grad_norm": 0.4569193422794342, + "learning_rate": 9.971830326876864e-06, + "loss": 0.5137, + "step": 236 + }, + { + "epoch": 0.13105990783410137, + "grad_norm": 0.5937883853912354, + "learning_rate": 9.970797670174663e-06, + "loss": 0.5142, + "step": 237 + }, + { + "epoch": 0.13161290322580646, + "grad_norm": 0.48625534772872925, + "learning_rate": 9.969746480264855e-06, + "loss": 0.5279, + "step": 238 + }, + { + "epoch": 0.13216589861751152, + "grad_norm": 0.47141098976135254, + "learning_rate": 9.968676761066714e-06, + "loss": 0.5042, + "step": 239 + }, + { + "epoch": 0.1327188940092166, + "grad_norm": 0.4889548122882843, + "learning_rate": 9.967588516568601e-06, + "loss": 0.5154, + "step": 240 + }, + { + "epoch": 0.13327188940092166, + "grad_norm": 0.4120868444442749, + "learning_rate": 9.966481750827943e-06, + "loss": 0.504, + "step": 241 + }, + { + "epoch": 0.13382488479262672, + "grad_norm": 0.5358414053916931, + "learning_rate": 9.965356467971228e-06, + "loss": 0.5264, + "step": 242 + }, + { + "epoch": 0.1343778801843318, + "grad_norm": 0.472572922706604, + "learning_rate": 9.964212672193978e-06, + "loss": 0.5176, + "step": 243 + }, + { + "epoch": 0.13493087557603686, + "grad_norm": 0.4870736598968506, + "learning_rate": 9.96305036776074e-06, + "loss": 0.5071, + "step": 244 + }, + { + "epoch": 0.13548387096774195, + "grad_norm": 0.5704755783081055, + "learning_rate": 9.96186955900507e-06, + "loss": 0.5198, + "step": 245 + }, + { + "epoch": 0.136036866359447, + "grad_norm": 0.5287049412727356, + "learning_rate": 9.960670250329517e-06, + "loss": 0.5216, + "step": 246 + }, + { + "epoch": 0.13658986175115206, + "grad_norm": 0.5611394643783569, + "learning_rate": 9.959452446205603e-06, + "loss": 0.5188, + "step": 247 + }, + { + "epoch": 0.13714285714285715, + "grad_norm": 0.4512503445148468, + "learning_rate": 9.958216151173812e-06, + "loss": 0.4979, + "step": 248 + }, + { + "epoch": 0.1376958525345622, + "grad_norm": 0.47202053666114807, + "learning_rate": 9.956961369843565e-06, + "loss": 0.4834, + "step": 249 + }, + { + "epoch": 0.1382488479262673, + "grad_norm": 0.5006586909294128, + "learning_rate": 9.95568810689321e-06, + "loss": 0.5285, + "step": 250 + }, + { + "epoch": 0.13880184331797235, + "grad_norm": 0.4660753011703491, + "learning_rate": 9.954396367070006e-06, + "loss": 0.5192, + "step": 251 + }, + { + "epoch": 0.1393548387096774, + "grad_norm": 0.4885999262332916, + "learning_rate": 9.953086155190095e-06, + "loss": 0.5028, + "step": 252 + }, + { + "epoch": 0.1399078341013825, + "grad_norm": 0.49867549538612366, + "learning_rate": 9.951757476138495e-06, + "loss": 0.5039, + "step": 253 + }, + { + "epoch": 0.14046082949308755, + "grad_norm": 0.5823776125907898, + "learning_rate": 9.950410334869075e-06, + "loss": 0.535, + "step": 254 + }, + { + "epoch": 0.14101382488479264, + "grad_norm": 0.47015827894210815, + "learning_rate": 9.949044736404538e-06, + "loss": 0.528, + "step": 255 + }, + { + "epoch": 0.1415668202764977, + "grad_norm": 0.5392245054244995, + "learning_rate": 9.94766068583641e-06, + "loss": 0.4855, + "step": 256 + }, + { + "epoch": 0.14211981566820275, + "grad_norm": 0.5323054194450378, + "learning_rate": 9.946258188325003e-06, + "loss": 0.5269, + "step": 257 + }, + { + "epoch": 0.14267281105990784, + "grad_norm": 0.5459584593772888, + "learning_rate": 9.944837249099418e-06, + "loss": 0.4874, + "step": 258 + }, + { + "epoch": 0.1432258064516129, + "grad_norm": 0.4532824158668518, + "learning_rate": 9.943397873457503e-06, + "loss": 0.5111, + "step": 259 + }, + { + "epoch": 0.14377880184331798, + "grad_norm": 0.46974876523017883, + "learning_rate": 9.94194006676586e-06, + "loss": 0.5261, + "step": 260 + }, + { + "epoch": 0.14433179723502304, + "grad_norm": 0.5061238408088684, + "learning_rate": 9.94046383445979e-06, + "loss": 0.5217, + "step": 261 + }, + { + "epoch": 0.1448847926267281, + "grad_norm": 0.4556806981563568, + "learning_rate": 9.938969182043312e-06, + "loss": 0.4969, + "step": 262 + }, + { + "epoch": 0.14543778801843318, + "grad_norm": 0.48571473360061646, + "learning_rate": 9.937456115089108e-06, + "loss": 0.5057, + "step": 263 + }, + { + "epoch": 0.14599078341013824, + "grad_norm": 0.497122198343277, + "learning_rate": 9.935924639238526e-06, + "loss": 0.5205, + "step": 264 + }, + { + "epoch": 0.14654377880184333, + "grad_norm": 0.5786699652671814, + "learning_rate": 9.934374760201546e-06, + "loss": 0.5122, + "step": 265 + }, + { + "epoch": 0.14709677419354839, + "grad_norm": 0.5201807022094727, + "learning_rate": 9.932806483756763e-06, + "loss": 0.5328, + "step": 266 + }, + { + "epoch": 0.14764976958525344, + "grad_norm": 0.5433037877082825, + "learning_rate": 9.931219815751368e-06, + "loss": 0.5153, + "step": 267 + }, + { + "epoch": 0.14820276497695853, + "grad_norm": 0.47332048416137695, + "learning_rate": 9.929614762101117e-06, + "loss": 0.5004, + "step": 268 + }, + { + "epoch": 0.1487557603686636, + "grad_norm": 0.5838091373443604, + "learning_rate": 9.927991328790324e-06, + "loss": 0.4889, + "step": 269 + }, + { + "epoch": 0.14930875576036867, + "grad_norm": 0.4618660807609558, + "learning_rate": 9.926349521871824e-06, + "loss": 0.5265, + "step": 270 + }, + { + "epoch": 0.14986175115207373, + "grad_norm": 0.5086191892623901, + "learning_rate": 9.924689347466959e-06, + "loss": 0.5106, + "step": 271 + }, + { + "epoch": 0.15041474654377882, + "grad_norm": 0.457042932510376, + "learning_rate": 9.92301081176555e-06, + "loss": 0.518, + "step": 272 + }, + { + "epoch": 0.15096774193548387, + "grad_norm": 0.5256728529930115, + "learning_rate": 9.92131392102588e-06, + "loss": 0.531, + "step": 273 + }, + { + "epoch": 0.15152073732718893, + "grad_norm": 0.4683758318424225, + "learning_rate": 9.919598681574665e-06, + "loss": 0.4896, + "step": 274 + }, + { + "epoch": 0.15207373271889402, + "grad_norm": 0.4681558609008789, + "learning_rate": 9.917865099807034e-06, + "loss": 0.5089, + "step": 275 + }, + { + "epoch": 0.15262672811059907, + "grad_norm": 0.4770297706127167, + "learning_rate": 9.916113182186503e-06, + "loss": 0.4973, + "step": 276 + }, + { + "epoch": 0.15317972350230416, + "grad_norm": 0.4939088225364685, + "learning_rate": 9.91434293524495e-06, + "loss": 0.5388, + "step": 277 + }, + { + "epoch": 0.15373271889400922, + "grad_norm": 0.5165495276451111, + "learning_rate": 9.912554365582596e-06, + "loss": 0.5046, + "step": 278 + }, + { + "epoch": 0.15428571428571428, + "grad_norm": 0.4685964584350586, + "learning_rate": 9.910747479867975e-06, + "loss": 0.5373, + "step": 279 + }, + { + "epoch": 0.15483870967741936, + "grad_norm": 0.5078836679458618, + "learning_rate": 9.908922284837911e-06, + "loss": 0.5105, + "step": 280 + }, + { + "epoch": 0.15539170506912442, + "grad_norm": 0.4230501651763916, + "learning_rate": 9.90707878729749e-06, + "loss": 0.4749, + "step": 281 + }, + { + "epoch": 0.1559447004608295, + "grad_norm": 0.483981728553772, + "learning_rate": 9.905216994120044e-06, + "loss": 0.4954, + "step": 282 + }, + { + "epoch": 0.15649769585253456, + "grad_norm": 0.4270246922969818, + "learning_rate": 9.90333691224711e-06, + "loss": 0.5007, + "step": 283 + }, + { + "epoch": 0.15705069124423962, + "grad_norm": 0.4967236816883087, + "learning_rate": 9.901438548688423e-06, + "loss": 0.5308, + "step": 284 + }, + { + "epoch": 0.1576036866359447, + "grad_norm": 0.4965408444404602, + "learning_rate": 9.89952191052187e-06, + "loss": 0.5093, + "step": 285 + }, + { + "epoch": 0.15815668202764976, + "grad_norm": 0.45975926518440247, + "learning_rate": 9.89758700489348e-06, + "loss": 0.5122, + "step": 286 + }, + { + "epoch": 0.15870967741935485, + "grad_norm": 0.5180716514587402, + "learning_rate": 9.895633839017387e-06, + "loss": 0.5515, + "step": 287 + }, + { + "epoch": 0.1592626728110599, + "grad_norm": 0.43234091997146606, + "learning_rate": 9.893662420175809e-06, + "loss": 0.4901, + "step": 288 + }, + { + "epoch": 0.15981566820276497, + "grad_norm": 0.4894983768463135, + "learning_rate": 9.89167275571902e-06, + "loss": 0.5052, + "step": 289 + }, + { + "epoch": 0.16036866359447005, + "grad_norm": 0.4710214138031006, + "learning_rate": 9.889664853065315e-06, + "loss": 0.5, + "step": 290 + }, + { + "epoch": 0.1609216589861751, + "grad_norm": 0.5677446126937866, + "learning_rate": 9.887638719700996e-06, + "loss": 0.5169, + "step": 291 + }, + { + "epoch": 0.1614746543778802, + "grad_norm": 0.45073893666267395, + "learning_rate": 9.88559436318033e-06, + "loss": 0.5258, + "step": 292 + }, + { + "epoch": 0.16202764976958525, + "grad_norm": 0.5394282341003418, + "learning_rate": 9.883531791125538e-06, + "loss": 0.4799, + "step": 293 + }, + { + "epoch": 0.1625806451612903, + "grad_norm": 0.4784766435623169, + "learning_rate": 9.881451011226742e-06, + "loss": 0.4693, + "step": 294 + }, + { + "epoch": 0.1631336405529954, + "grad_norm": 0.435619980096817, + "learning_rate": 9.87935203124196e-06, + "loss": 0.5196, + "step": 295 + }, + { + "epoch": 0.16368663594470045, + "grad_norm": 0.5438254475593567, + "learning_rate": 9.877234858997066e-06, + "loss": 0.5078, + "step": 296 + }, + { + "epoch": 0.16423963133640554, + "grad_norm": 0.4646199643611908, + "learning_rate": 9.875099502385761e-06, + "loss": 0.5239, + "step": 297 + }, + { + "epoch": 0.1647926267281106, + "grad_norm": 0.549284040927887, + "learning_rate": 9.872945969369546e-06, + "loss": 0.5029, + "step": 298 + }, + { + "epoch": 0.16534562211981566, + "grad_norm": 0.5235300064086914, + "learning_rate": 9.87077426797769e-06, + "loss": 0.5225, + "step": 299 + }, + { + "epoch": 0.16589861751152074, + "grad_norm": 0.525598406791687, + "learning_rate": 9.8685844063072e-06, + "loss": 0.5366, + "step": 300 + }, + { + "epoch": 0.1664516129032258, + "grad_norm": 0.5001644492149353, + "learning_rate": 9.866376392522798e-06, + "loss": 0.5025, + "step": 301 + }, + { + "epoch": 0.16700460829493088, + "grad_norm": 0.4980228841304779, + "learning_rate": 9.864150234856876e-06, + "loss": 0.5332, + "step": 302 + }, + { + "epoch": 0.16755760368663594, + "grad_norm": 0.44216716289520264, + "learning_rate": 9.861905941609482e-06, + "loss": 0.4831, + "step": 303 + }, + { + "epoch": 0.16811059907834103, + "grad_norm": 0.45178863406181335, + "learning_rate": 9.859643521148275e-06, + "loss": 0.5164, + "step": 304 + }, + { + "epoch": 0.16866359447004609, + "grad_norm": 0.4833186864852905, + "learning_rate": 9.857362981908505e-06, + "loss": 0.5128, + "step": 305 + }, + { + "epoch": 0.16921658986175114, + "grad_norm": 0.5167268514633179, + "learning_rate": 9.855064332392972e-06, + "loss": 0.5149, + "step": 306 + }, + { + "epoch": 0.16976958525345623, + "grad_norm": 0.4465697109699249, + "learning_rate": 9.852747581172002e-06, + "loss": 0.5016, + "step": 307 + }, + { + "epoch": 0.1703225806451613, + "grad_norm": 0.5105251669883728, + "learning_rate": 9.850412736883408e-06, + "loss": 0.5021, + "step": 308 + }, + { + "epoch": 0.17087557603686637, + "grad_norm": 0.5230706334114075, + "learning_rate": 9.848059808232464e-06, + "loss": 0.4857, + "step": 309 + }, + { + "epoch": 0.17142857142857143, + "grad_norm": 0.5510496497154236, + "learning_rate": 9.845688803991873e-06, + "loss": 0.5305, + "step": 310 + }, + { + "epoch": 0.1719815668202765, + "grad_norm": 0.4802294671535492, + "learning_rate": 9.843299733001723e-06, + "loss": 0.5084, + "step": 311 + }, + { + "epoch": 0.17253456221198157, + "grad_norm": 0.5729020833969116, + "learning_rate": 9.840892604169473e-06, + "loss": 0.4797, + "step": 312 + }, + { + "epoch": 0.17308755760368663, + "grad_norm": 0.6035499572753906, + "learning_rate": 9.838467426469897e-06, + "loss": 0.5069, + "step": 313 + }, + { + "epoch": 0.17364055299539172, + "grad_norm": 0.4519449770450592, + "learning_rate": 9.836024208945074e-06, + "loss": 0.4959, + "step": 314 + }, + { + "epoch": 0.17419354838709677, + "grad_norm": 0.5990521907806396, + "learning_rate": 9.833562960704336e-06, + "loss": 0.522, + "step": 315 + }, + { + "epoch": 0.17474654377880183, + "grad_norm": 0.4257313013076782, + "learning_rate": 9.831083690924246e-06, + "loss": 0.4769, + "step": 316 + }, + { + "epoch": 0.17529953917050692, + "grad_norm": 0.5456443428993225, + "learning_rate": 9.828586408848553e-06, + "loss": 0.5129, + "step": 317 + }, + { + "epoch": 0.17585253456221198, + "grad_norm": 0.4802759885787964, + "learning_rate": 9.82607112378817e-06, + "loss": 0.5043, + "step": 318 + }, + { + "epoch": 0.17640552995391706, + "grad_norm": 0.6043391823768616, + "learning_rate": 9.823537845121126e-06, + "loss": 0.5063, + "step": 319 + }, + { + "epoch": 0.17695852534562212, + "grad_norm": 0.4696696698665619, + "learning_rate": 9.82098658229254e-06, + "loss": 0.5164, + "step": 320 + }, + { + "epoch": 0.17751152073732718, + "grad_norm": 0.5530909299850464, + "learning_rate": 9.818417344814587e-06, + "loss": 0.5275, + "step": 321 + }, + { + "epoch": 0.17806451612903226, + "grad_norm": 0.609788179397583, + "learning_rate": 9.815830142266457e-06, + "loss": 0.4716, + "step": 322 + }, + { + "epoch": 0.17861751152073732, + "grad_norm": 0.46327316761016846, + "learning_rate": 9.813224984294318e-06, + "loss": 0.4887, + "step": 323 + }, + { + "epoch": 0.1791705069124424, + "grad_norm": 0.5943720936775208, + "learning_rate": 9.810601880611286e-06, + "loss": 0.5009, + "step": 324 + }, + { + "epoch": 0.17972350230414746, + "grad_norm": 0.4262332320213318, + "learning_rate": 9.807960840997387e-06, + "loss": 0.4747, + "step": 325 + }, + { + "epoch": 0.18027649769585252, + "grad_norm": 0.47295740246772766, + "learning_rate": 9.805301875299518e-06, + "loss": 0.5092, + "step": 326 + }, + { + "epoch": 0.1808294930875576, + "grad_norm": 0.4910508394241333, + "learning_rate": 9.802624993431414e-06, + "loss": 0.5195, + "step": 327 + }, + { + "epoch": 0.18138248847926267, + "grad_norm": 0.5312582850456238, + "learning_rate": 9.799930205373605e-06, + "loss": 0.5077, + "step": 328 + }, + { + "epoch": 0.18193548387096775, + "grad_norm": 0.49236226081848145, + "learning_rate": 9.797217521173385e-06, + "loss": 0.5228, + "step": 329 + }, + { + "epoch": 0.1824884792626728, + "grad_norm": 0.5442107319831848, + "learning_rate": 9.794486950944775e-06, + "loss": 0.4909, + "step": 330 + }, + { + "epoch": 0.18304147465437787, + "grad_norm": 0.4408930242061615, + "learning_rate": 9.79173850486848e-06, + "loss": 0.4919, + "step": 331 + }, + { + "epoch": 0.18359447004608295, + "grad_norm": 0.4888366460800171, + "learning_rate": 9.78897219319185e-06, + "loss": 0.5247, + "step": 332 + }, + { + "epoch": 0.184147465437788, + "grad_norm": 0.48040100932121277, + "learning_rate": 9.786188026228853e-06, + "loss": 0.4746, + "step": 333 + }, + { + "epoch": 0.1847004608294931, + "grad_norm": 0.46840646862983704, + "learning_rate": 9.783386014360024e-06, + "loss": 0.5174, + "step": 334 + }, + { + "epoch": 0.18525345622119815, + "grad_norm": 0.4824281930923462, + "learning_rate": 9.780566168032432e-06, + "loss": 0.4944, + "step": 335 + }, + { + "epoch": 0.18580645161290324, + "grad_norm": 0.4401717483997345, + "learning_rate": 9.77772849775964e-06, + "loss": 0.5074, + "step": 336 + }, + { + "epoch": 0.1863594470046083, + "grad_norm": 0.5038358569145203, + "learning_rate": 9.774873014121667e-06, + "loss": 0.5052, + "step": 337 + }, + { + "epoch": 0.18691244239631336, + "grad_norm": 0.47577494382858276, + "learning_rate": 9.771999727764946e-06, + "loss": 0.4856, + "step": 338 + }, + { + "epoch": 0.18746543778801844, + "grad_norm": 0.4989863634109497, + "learning_rate": 9.76910864940229e-06, + "loss": 0.4839, + "step": 339 + }, + { + "epoch": 0.1880184331797235, + "grad_norm": 0.49821236729621887, + "learning_rate": 9.766199789812845e-06, + "loss": 0.498, + "step": 340 + }, + { + "epoch": 0.18857142857142858, + "grad_norm": 0.44591957330703735, + "learning_rate": 9.763273159842052e-06, + "loss": 0.4832, + "step": 341 + }, + { + "epoch": 0.18912442396313364, + "grad_norm": 0.4908744990825653, + "learning_rate": 9.76032877040161e-06, + "loss": 0.4927, + "step": 342 + }, + { + "epoch": 0.1896774193548387, + "grad_norm": 0.55243319272995, + "learning_rate": 9.757366632469427e-06, + "loss": 0.5005, + "step": 343 + }, + { + "epoch": 0.19023041474654379, + "grad_norm": 0.4234575629234314, + "learning_rate": 9.754386757089596e-06, + "loss": 0.4799, + "step": 344 + }, + { + "epoch": 0.19078341013824884, + "grad_norm": 0.49401965737342834, + "learning_rate": 9.751389155372329e-06, + "loss": 0.4942, + "step": 345 + }, + { + "epoch": 0.19133640552995393, + "grad_norm": 0.4622778594493866, + "learning_rate": 9.74837383849394e-06, + "loss": 0.4644, + "step": 346 + }, + { + "epoch": 0.191889400921659, + "grad_norm": 0.4600273072719574, + "learning_rate": 9.745340817696787e-06, + "loss": 0.5019, + "step": 347 + }, + { + "epoch": 0.19244239631336404, + "grad_norm": 0.4513567388057709, + "learning_rate": 9.742290104289237e-06, + "loss": 0.4704, + "step": 348 + }, + { + "epoch": 0.19299539170506913, + "grad_norm": 0.5180230736732483, + "learning_rate": 9.739221709645622e-06, + "loss": 0.4917, + "step": 349 + }, + { + "epoch": 0.1935483870967742, + "grad_norm": 0.44200998544692993, + "learning_rate": 9.736135645206198e-06, + "loss": 0.5047, + "step": 350 + }, + { + "epoch": 0.19410138248847927, + "grad_norm": 0.5660358667373657, + "learning_rate": 9.7330319224771e-06, + "loss": 0.5023, + "step": 351 + }, + { + "epoch": 0.19465437788018433, + "grad_norm": 0.44904786348342896, + "learning_rate": 9.729910553030304e-06, + "loss": 0.4998, + "step": 352 + }, + { + "epoch": 0.1952073732718894, + "grad_norm": 0.523826003074646, + "learning_rate": 9.726771548503575e-06, + "loss": 0.4966, + "step": 353 + }, + { + "epoch": 0.19576036866359448, + "grad_norm": 0.4515267312526703, + "learning_rate": 9.72361492060043e-06, + "loss": 0.4996, + "step": 354 + }, + { + "epoch": 0.19631336405529953, + "grad_norm": 0.4984499216079712, + "learning_rate": 9.7204406810901e-06, + "loss": 0.4985, + "step": 355 + }, + { + "epoch": 0.19686635944700462, + "grad_norm": 0.4689009189605713, + "learning_rate": 9.71724884180747e-06, + "loss": 0.4936, + "step": 356 + }, + { + "epoch": 0.19741935483870968, + "grad_norm": 0.46954217553138733, + "learning_rate": 9.714039414653047e-06, + "loss": 0.491, + "step": 357 + }, + { + "epoch": 0.19797235023041473, + "grad_norm": 0.5114353895187378, + "learning_rate": 9.71081241159292e-06, + "loss": 0.488, + "step": 358 + }, + { + "epoch": 0.19852534562211982, + "grad_norm": 0.4611765742301941, + "learning_rate": 9.707567844658698e-06, + "loss": 0.4749, + "step": 359 + }, + { + "epoch": 0.19907834101382488, + "grad_norm": 0.5052332878112793, + "learning_rate": 9.704305725947483e-06, + "loss": 0.5234, + "step": 360 + }, + { + "epoch": 0.19963133640552996, + "grad_norm": 0.5130250453948975, + "learning_rate": 9.701026067621813e-06, + "loss": 0.5006, + "step": 361 + }, + { + "epoch": 0.20018433179723502, + "grad_norm": 0.47562843561172485, + "learning_rate": 9.697728881909622e-06, + "loss": 0.529, + "step": 362 + }, + { + "epoch": 0.20073732718894008, + "grad_norm": 0.48228660225868225, + "learning_rate": 9.694414181104197e-06, + "loss": 0.5057, + "step": 363 + }, + { + "epoch": 0.20129032258064516, + "grad_norm": 0.501700222492218, + "learning_rate": 9.691081977564124e-06, + "loss": 0.4949, + "step": 364 + }, + { + "epoch": 0.20184331797235022, + "grad_norm": 0.4578889012336731, + "learning_rate": 9.687732283713247e-06, + "loss": 0.4989, + "step": 365 + }, + { + "epoch": 0.2023963133640553, + "grad_norm": 0.5223492980003357, + "learning_rate": 9.684365112040622e-06, + "loss": 0.4772, + "step": 366 + }, + { + "epoch": 0.20294930875576037, + "grad_norm": 0.4984261393547058, + "learning_rate": 9.680980475100471e-06, + "loss": 0.5132, + "step": 367 + }, + { + "epoch": 0.20350230414746545, + "grad_norm": 0.4805150032043457, + "learning_rate": 9.677578385512135e-06, + "loss": 0.5069, + "step": 368 + }, + { + "epoch": 0.2040552995391705, + "grad_norm": 0.46725666522979736, + "learning_rate": 9.674158855960017e-06, + "loss": 0.4806, + "step": 369 + }, + { + "epoch": 0.20460829493087557, + "grad_norm": 0.47330862283706665, + "learning_rate": 9.670721899193556e-06, + "loss": 0.5127, + "step": 370 + }, + { + "epoch": 0.20516129032258065, + "grad_norm": 0.5010029077529907, + "learning_rate": 9.667267528027157e-06, + "loss": 0.5189, + "step": 371 + }, + { + "epoch": 0.2057142857142857, + "grad_norm": 0.5415194034576416, + "learning_rate": 9.663795755340159e-06, + "loss": 0.5053, + "step": 372 + }, + { + "epoch": 0.2062672811059908, + "grad_norm": 0.4000757932662964, + "learning_rate": 9.660306594076779e-06, + "loss": 0.4939, + "step": 373 + }, + { + "epoch": 0.20682027649769585, + "grad_norm": 0.5064148306846619, + "learning_rate": 9.656800057246065e-06, + "loss": 0.4937, + "step": 374 + }, + { + "epoch": 0.2073732718894009, + "grad_norm": 0.467369943857193, + "learning_rate": 9.65327615792185e-06, + "loss": 0.5215, + "step": 375 + }, + { + "epoch": 0.207926267281106, + "grad_norm": 0.4753814935684204, + "learning_rate": 9.649734909242699e-06, + "loss": 0.4903, + "step": 376 + }, + { + "epoch": 0.20847926267281106, + "grad_norm": 0.4203657805919647, + "learning_rate": 9.646176324411867e-06, + "loss": 0.4874, + "step": 377 + }, + { + "epoch": 0.20903225806451614, + "grad_norm": 0.4930543601512909, + "learning_rate": 9.642600416697242e-06, + "loss": 0.4816, + "step": 378 + }, + { + "epoch": 0.2095852534562212, + "grad_norm": 0.4189959764480591, + "learning_rate": 9.639007199431298e-06, + "loss": 0.491, + "step": 379 + }, + { + "epoch": 0.21013824884792626, + "grad_norm": 0.48571497201919556, + "learning_rate": 9.635396686011052e-06, + "loss": 0.5005, + "step": 380 + }, + { + "epoch": 0.21069124423963134, + "grad_norm": 0.4344756007194519, + "learning_rate": 9.631768889898004e-06, + "loss": 0.5113, + "step": 381 + }, + { + "epoch": 0.2112442396313364, + "grad_norm": 0.5239787101745605, + "learning_rate": 9.628123824618087e-06, + "loss": 0.5013, + "step": 382 + }, + { + "epoch": 0.21179723502304149, + "grad_norm": 0.40745410323143005, + "learning_rate": 9.624461503761628e-06, + "loss": 0.4824, + "step": 383 + }, + { + "epoch": 0.21235023041474654, + "grad_norm": 0.5618027448654175, + "learning_rate": 9.620781940983288e-06, + "loss": 0.4852, + "step": 384 + }, + { + "epoch": 0.2129032258064516, + "grad_norm": 0.5030412673950195, + "learning_rate": 9.617085150002012e-06, + "loss": 0.4868, + "step": 385 + }, + { + "epoch": 0.2134562211981567, + "grad_norm": 0.5133850574493408, + "learning_rate": 9.613371144600976e-06, + "loss": 0.5076, + "step": 386 + }, + { + "epoch": 0.21400921658986174, + "grad_norm": 0.47164881229400635, + "learning_rate": 9.609639938627546e-06, + "loss": 0.5076, + "step": 387 + }, + { + "epoch": 0.21456221198156683, + "grad_norm": 0.4972430467605591, + "learning_rate": 9.605891545993214e-06, + "loss": 0.4843, + "step": 388 + }, + { + "epoch": 0.2151152073732719, + "grad_norm": 0.4716038405895233, + "learning_rate": 9.60212598067355e-06, + "loss": 0.4782, + "step": 389 + }, + { + "epoch": 0.21566820276497695, + "grad_norm": 0.4822791814804077, + "learning_rate": 9.59834325670815e-06, + "loss": 0.4874, + "step": 390 + }, + { + "epoch": 0.21622119815668203, + "grad_norm": 0.4538308382034302, + "learning_rate": 9.594543388200592e-06, + "loss": 0.501, + "step": 391 + }, + { + "epoch": 0.2167741935483871, + "grad_norm": 0.4863637387752533, + "learning_rate": 9.59072638931837e-06, + "loss": 0.5115, + "step": 392 + }, + { + "epoch": 0.21732718894009218, + "grad_norm": 0.4747362434864044, + "learning_rate": 9.586892274292846e-06, + "loss": 0.5262, + "step": 393 + }, + { + "epoch": 0.21788018433179723, + "grad_norm": 0.4688088595867157, + "learning_rate": 9.583041057419203e-06, + "loss": 0.517, + "step": 394 + }, + { + "epoch": 0.2184331797235023, + "grad_norm": 0.49005988240242004, + "learning_rate": 9.579172753056383e-06, + "loss": 0.4737, + "step": 395 + }, + { + "epoch": 0.21898617511520738, + "grad_norm": 0.4679119288921356, + "learning_rate": 9.575287375627037e-06, + "loss": 0.4978, + "step": 396 + }, + { + "epoch": 0.21953917050691243, + "grad_norm": 0.47708967328071594, + "learning_rate": 9.571384939617478e-06, + "loss": 0.5005, + "step": 397 + }, + { + "epoch": 0.22009216589861752, + "grad_norm": 0.47183945775032043, + "learning_rate": 9.567465459577613e-06, + "loss": 0.4708, + "step": 398 + }, + { + "epoch": 0.22064516129032258, + "grad_norm": 0.47582703828811646, + "learning_rate": 9.563528950120895e-06, + "loss": 0.4983, + "step": 399 + }, + { + "epoch": 0.22119815668202766, + "grad_norm": 0.48260581493377686, + "learning_rate": 9.559575425924279e-06, + "loss": 0.4894, + "step": 400 + }, + { + "epoch": 0.22175115207373272, + "grad_norm": 0.5306166410446167, + "learning_rate": 9.55560490172815e-06, + "loss": 0.5068, + "step": 401 + }, + { + "epoch": 0.22230414746543778, + "grad_norm": 0.49763691425323486, + "learning_rate": 9.551617392336281e-06, + "loss": 0.5028, + "step": 402 + }, + { + "epoch": 0.22285714285714286, + "grad_norm": 0.5285375118255615, + "learning_rate": 9.547612912615769e-06, + "loss": 0.484, + "step": 403 + }, + { + "epoch": 0.22341013824884792, + "grad_norm": 0.5528775453567505, + "learning_rate": 9.543591477496985e-06, + "loss": 0.4774, + "step": 404 + }, + { + "epoch": 0.223963133640553, + "grad_norm": 0.48812106251716614, + "learning_rate": 9.539553101973516e-06, + "loss": 0.5042, + "step": 405 + }, + { + "epoch": 0.22451612903225807, + "grad_norm": 0.5272664427757263, + "learning_rate": 9.535497801102114e-06, + "loss": 0.4987, + "step": 406 + }, + { + "epoch": 0.22506912442396312, + "grad_norm": 0.5162638425827026, + "learning_rate": 9.531425590002629e-06, + "loss": 0.4942, + "step": 407 + }, + { + "epoch": 0.2256221198156682, + "grad_norm": 0.45863449573516846, + "learning_rate": 9.527336483857965e-06, + "loss": 0.5165, + "step": 408 + }, + { + "epoch": 0.22617511520737327, + "grad_norm": 0.5142653584480286, + "learning_rate": 9.523230497914016e-06, + "loss": 0.4883, + "step": 409 + }, + { + "epoch": 0.22672811059907835, + "grad_norm": 0.4717908203601837, + "learning_rate": 9.519107647479609e-06, + "loss": 0.4859, + "step": 410 + }, + { + "epoch": 0.2272811059907834, + "grad_norm": 0.4761084020137787, + "learning_rate": 9.514967947926453e-06, + "loss": 0.5022, + "step": 411 + }, + { + "epoch": 0.22783410138248847, + "grad_norm": 0.4796617925167084, + "learning_rate": 9.510811414689073e-06, + "loss": 0.5058, + "step": 412 + }, + { + "epoch": 0.22838709677419355, + "grad_norm": 0.5060638189315796, + "learning_rate": 9.506638063264759e-06, + "loss": 0.4859, + "step": 413 + }, + { + "epoch": 0.2289400921658986, + "grad_norm": 0.42853257060050964, + "learning_rate": 9.502447909213508e-06, + "loss": 0.5004, + "step": 414 + }, + { + "epoch": 0.2294930875576037, + "grad_norm": 0.45876848697662354, + "learning_rate": 9.498240968157962e-06, + "loss": 0.4898, + "step": 415 + }, + { + "epoch": 0.23004608294930876, + "grad_norm": 0.4455052614212036, + "learning_rate": 9.49401725578335e-06, + "loss": 0.4972, + "step": 416 + }, + { + "epoch": 0.2305990783410138, + "grad_norm": 0.41394197940826416, + "learning_rate": 9.489776787837438e-06, + "loss": 0.5006, + "step": 417 + }, + { + "epoch": 0.2311520737327189, + "grad_norm": 0.44113320112228394, + "learning_rate": 9.485519580130456e-06, + "loss": 0.4644, + "step": 418 + }, + { + "epoch": 0.23170506912442396, + "grad_norm": 0.4882364869117737, + "learning_rate": 9.481245648535053e-06, + "loss": 0.4726, + "step": 419 + }, + { + "epoch": 0.23225806451612904, + "grad_norm": 0.4658440053462982, + "learning_rate": 9.476955008986228e-06, + "loss": 0.497, + "step": 420 + }, + { + "epoch": 0.2328110599078341, + "grad_norm": 0.4944906532764435, + "learning_rate": 9.472647677481275e-06, + "loss": 0.4743, + "step": 421 + }, + { + "epoch": 0.23336405529953916, + "grad_norm": 0.4508844316005707, + "learning_rate": 9.468323670079725e-06, + "loss": 0.4884, + "step": 422 + }, + { + "epoch": 0.23391705069124424, + "grad_norm": 0.470225065946579, + "learning_rate": 9.463983002903278e-06, + "loss": 0.4929, + "step": 423 + }, + { + "epoch": 0.2344700460829493, + "grad_norm": 0.5286823511123657, + "learning_rate": 9.459625692135756e-06, + "loss": 0.4938, + "step": 424 + }, + { + "epoch": 0.2350230414746544, + "grad_norm": 0.45246848464012146, + "learning_rate": 9.45525175402303e-06, + "loss": 0.4834, + "step": 425 + }, + { + "epoch": 0.23557603686635945, + "grad_norm": 0.48045429587364197, + "learning_rate": 9.450861204872965e-06, + "loss": 0.4877, + "step": 426 + }, + { + "epoch": 0.2361290322580645, + "grad_norm": 0.5548929572105408, + "learning_rate": 9.44645406105536e-06, + "loss": 0.5185, + "step": 427 + }, + { + "epoch": 0.2366820276497696, + "grad_norm": 0.4929436147212982, + "learning_rate": 9.44203033900189e-06, + "loss": 0.5048, + "step": 428 + }, + { + "epoch": 0.23723502304147465, + "grad_norm": 0.46492716670036316, + "learning_rate": 9.437590055206032e-06, + "loss": 0.4939, + "step": 429 + }, + { + "epoch": 0.23778801843317973, + "grad_norm": 0.4560331404209137, + "learning_rate": 9.433133226223018e-06, + "loss": 0.5016, + "step": 430 + }, + { + "epoch": 0.2383410138248848, + "grad_norm": 0.4622935652732849, + "learning_rate": 9.428659868669765e-06, + "loss": 0.4748, + "step": 431 + }, + { + "epoch": 0.23889400921658988, + "grad_norm": 0.4686061441898346, + "learning_rate": 9.424169999224819e-06, + "loss": 0.5012, + "step": 432 + }, + { + "epoch": 0.23944700460829493, + "grad_norm": 0.47005850076675415, + "learning_rate": 9.419663634628283e-06, + "loss": 0.4902, + "step": 433 + }, + { + "epoch": 0.24, + "grad_norm": 0.5422276854515076, + "learning_rate": 9.415140791681767e-06, + "loss": 0.4923, + "step": 434 + }, + { + "epoch": 0.24055299539170508, + "grad_norm": 0.42137962579727173, + "learning_rate": 9.410601487248315e-06, + "loss": 0.4762, + "step": 435 + }, + { + "epoch": 0.24110599078341013, + "grad_norm": 0.518340528011322, + "learning_rate": 9.406045738252349e-06, + "loss": 0.5257, + "step": 436 + }, + { + "epoch": 0.24165898617511522, + "grad_norm": 0.5296422243118286, + "learning_rate": 9.4014735616796e-06, + "loss": 0.4959, + "step": 437 + }, + { + "epoch": 0.24221198156682028, + "grad_norm": 0.4660092890262604, + "learning_rate": 9.396884974577052e-06, + "loss": 0.4902, + "step": 438 + }, + { + "epoch": 0.24276497695852534, + "grad_norm": 0.5939272046089172, + "learning_rate": 9.392279994052872e-06, + "loss": 0.4859, + "step": 439 + }, + { + "epoch": 0.24331797235023042, + "grad_norm": 0.4680648446083069, + "learning_rate": 9.387658637276348e-06, + "loss": 0.4856, + "step": 440 + }, + { + "epoch": 0.24387096774193548, + "grad_norm": 0.4640378952026367, + "learning_rate": 9.383020921477824e-06, + "loss": 0.4976, + "step": 441 + }, + { + "epoch": 0.24442396313364056, + "grad_norm": 0.5372658371925354, + "learning_rate": 9.378366863948644e-06, + "loss": 0.4731, + "step": 442 + }, + { + "epoch": 0.24497695852534562, + "grad_norm": 0.48611563444137573, + "learning_rate": 9.373696482041072e-06, + "loss": 0.474, + "step": 443 + }, + { + "epoch": 0.24552995391705068, + "grad_norm": 0.4766850173473358, + "learning_rate": 9.369009793168243e-06, + "loss": 0.4878, + "step": 444 + }, + { + "epoch": 0.24608294930875577, + "grad_norm": 0.503322184085846, + "learning_rate": 9.364306814804086e-06, + "loss": 0.49, + "step": 445 + }, + { + "epoch": 0.24663594470046082, + "grad_norm": 0.5735645294189453, + "learning_rate": 9.35958756448327e-06, + "loss": 0.5357, + "step": 446 + }, + { + "epoch": 0.2471889400921659, + "grad_norm": 0.5782787799835205, + "learning_rate": 9.354852059801127e-06, + "loss": 0.5052, + "step": 447 + }, + { + "epoch": 0.24774193548387097, + "grad_norm": 0.48679542541503906, + "learning_rate": 9.350100318413594e-06, + "loss": 0.4824, + "step": 448 + }, + { + "epoch": 0.24829493087557603, + "grad_norm": 0.535241961479187, + "learning_rate": 9.345332358037147e-06, + "loss": 0.4814, + "step": 449 + }, + { + "epoch": 0.2488479262672811, + "grad_norm": 0.5126506090164185, + "learning_rate": 9.340548196448729e-06, + "loss": 0.4903, + "step": 450 + }, + { + "epoch": 0.24940092165898617, + "grad_norm": 0.459888219833374, + "learning_rate": 9.33574785148569e-06, + "loss": 0.4764, + "step": 451 + }, + { + "epoch": 0.24995391705069125, + "grad_norm": 0.47858306765556335, + "learning_rate": 9.330931341045723e-06, + "loss": 0.4759, + "step": 452 + }, + { + "epoch": 0.2505069124423963, + "grad_norm": 0.4925948977470398, + "learning_rate": 9.326098683086786e-06, + "loss": 0.5092, + "step": 453 + }, + { + "epoch": 0.25105990783410137, + "grad_norm": 0.4197613000869751, + "learning_rate": 9.321249895627043e-06, + "loss": 0.4855, + "step": 454 + }, + { + "epoch": 0.25161290322580643, + "grad_norm": 0.4368072748184204, + "learning_rate": 9.316384996744798e-06, + "loss": 0.4731, + "step": 455 + }, + { + "epoch": 0.25216589861751154, + "grad_norm": 0.4712340533733368, + "learning_rate": 9.311504004578425e-06, + "loss": 0.4864, + "step": 456 + }, + { + "epoch": 0.2527188940092166, + "grad_norm": 0.4518265426158905, + "learning_rate": 9.306606937326295e-06, + "loss": 0.5119, + "step": 457 + }, + { + "epoch": 0.25327188940092166, + "grad_norm": 0.48842400312423706, + "learning_rate": 9.301693813246721e-06, + "loss": 0.4732, + "step": 458 + }, + { + "epoch": 0.2538248847926267, + "grad_norm": 0.5565005540847778, + "learning_rate": 9.296764650657881e-06, + "loss": 0.5018, + "step": 459 + }, + { + "epoch": 0.2543778801843318, + "grad_norm": 0.4874143600463867, + "learning_rate": 9.291819467937746e-06, + "loss": 0.4712, + "step": 460 + }, + { + "epoch": 0.2549308755760369, + "grad_norm": 0.5408375859260559, + "learning_rate": 9.286858283524022e-06, + "loss": 0.4785, + "step": 461 + }, + { + "epoch": 0.25548387096774194, + "grad_norm": 0.4784450829029083, + "learning_rate": 9.281881115914075e-06, + "loss": 0.4893, + "step": 462 + }, + { + "epoch": 0.256036866359447, + "grad_norm": 0.4540750980377197, + "learning_rate": 9.276887983664861e-06, + "loss": 0.4873, + "step": 463 + }, + { + "epoch": 0.25658986175115206, + "grad_norm": 0.5902926921844482, + "learning_rate": 9.271878905392863e-06, + "loss": 0.4903, + "step": 464 + }, + { + "epoch": 0.2571428571428571, + "grad_norm": 0.44395557045936584, + "learning_rate": 9.26685389977401e-06, + "loss": 0.4862, + "step": 465 + }, + { + "epoch": 0.25769585253456223, + "grad_norm": 0.44266945123672485, + "learning_rate": 9.261812985543625e-06, + "loss": 0.4875, + "step": 466 + }, + { + "epoch": 0.2582488479262673, + "grad_norm": 0.5877528190612793, + "learning_rate": 9.256756181496334e-06, + "loss": 0.498, + "step": 467 + }, + { + "epoch": 0.25880184331797235, + "grad_norm": 0.43955346941947937, + "learning_rate": 9.251683506486016e-06, + "loss": 0.4782, + "step": 468 + }, + { + "epoch": 0.2593548387096774, + "grad_norm": 0.5338640809059143, + "learning_rate": 9.246594979425719e-06, + "loss": 0.4834, + "step": 469 + }, + { + "epoch": 0.25990783410138246, + "grad_norm": 0.4647727608680725, + "learning_rate": 9.241490619287593e-06, + "loss": 0.4957, + "step": 470 + }, + { + "epoch": 0.2604608294930876, + "grad_norm": 0.3950575590133667, + "learning_rate": 9.236370445102825e-06, + "loss": 0.4864, + "step": 471 + }, + { + "epoch": 0.26101382488479263, + "grad_norm": 0.5759503245353699, + "learning_rate": 9.231234475961559e-06, + "loss": 0.5049, + "step": 472 + }, + { + "epoch": 0.2615668202764977, + "grad_norm": 0.49982547760009766, + "learning_rate": 9.22608273101283e-06, + "loss": 0.5099, + "step": 473 + }, + { + "epoch": 0.26211981566820275, + "grad_norm": 0.45275378227233887, + "learning_rate": 9.220915229464496e-06, + "loss": 0.4768, + "step": 474 + }, + { + "epoch": 0.2626728110599078, + "grad_norm": 0.4712145924568176, + "learning_rate": 9.215731990583155e-06, + "loss": 0.4418, + "step": 475 + }, + { + "epoch": 0.2632258064516129, + "grad_norm": 0.4405660927295685, + "learning_rate": 9.210533033694084e-06, + "loss": 0.4938, + "step": 476 + }, + { + "epoch": 0.263778801843318, + "grad_norm": 0.4714353382587433, + "learning_rate": 9.205318378181167e-06, + "loss": 0.4806, + "step": 477 + }, + { + "epoch": 0.26433179723502304, + "grad_norm": 0.5084551572799683, + "learning_rate": 9.200088043486813e-06, + "loss": 0.487, + "step": 478 + }, + { + "epoch": 0.2648847926267281, + "grad_norm": 0.4339804947376251, + "learning_rate": 9.194842049111889e-06, + "loss": 0.4703, + "step": 479 + }, + { + "epoch": 0.2654377880184332, + "grad_norm": 0.5167462825775146, + "learning_rate": 9.189580414615658e-06, + "loss": 0.495, + "step": 480 + }, + { + "epoch": 0.26599078341013827, + "grad_norm": 0.5027214288711548, + "learning_rate": 9.184303159615682e-06, + "loss": 0.487, + "step": 481 + }, + { + "epoch": 0.2665437788018433, + "grad_norm": 0.5376026034355164, + "learning_rate": 9.179010303787772e-06, + "loss": 0.4974, + "step": 482 + }, + { + "epoch": 0.2670967741935484, + "grad_norm": 0.4745848774909973, + "learning_rate": 9.173701866865905e-06, + "loss": 0.4761, + "step": 483 + }, + { + "epoch": 0.26764976958525344, + "grad_norm": 0.615476667881012, + "learning_rate": 9.168377868642142e-06, + "loss": 0.4903, + "step": 484 + }, + { + "epoch": 0.26820276497695855, + "grad_norm": 0.43914616107940674, + "learning_rate": 9.163038328966578e-06, + "loss": 0.5018, + "step": 485 + }, + { + "epoch": 0.2687557603686636, + "grad_norm": 0.5516188144683838, + "learning_rate": 9.157683267747239e-06, + "loss": 0.5042, + "step": 486 + }, + { + "epoch": 0.26930875576036867, + "grad_norm": 0.5231643319129944, + "learning_rate": 9.152312704950028e-06, + "loss": 0.4745, + "step": 487 + }, + { + "epoch": 0.2698617511520737, + "grad_norm": 0.4931274354457855, + "learning_rate": 9.146926660598646e-06, + "loss": 0.4891, + "step": 488 + }, + { + "epoch": 0.2704147465437788, + "grad_norm": 0.5022376179695129, + "learning_rate": 9.141525154774513e-06, + "loss": 0.4707, + "step": 489 + }, + { + "epoch": 0.2709677419354839, + "grad_norm": 0.4848520755767822, + "learning_rate": 9.136108207616694e-06, + "loss": 0.5176, + "step": 490 + }, + { + "epoch": 0.27152073732718895, + "grad_norm": 0.5185691118240356, + "learning_rate": 9.130675839321829e-06, + "loss": 0.4951, + "step": 491 + }, + { + "epoch": 0.272073732718894, + "grad_norm": 0.4987480938434601, + "learning_rate": 9.125228070144056e-06, + "loss": 0.4865, + "step": 492 + }, + { + "epoch": 0.27262672811059907, + "grad_norm": 0.47905007004737854, + "learning_rate": 9.119764920394928e-06, + "loss": 0.4931, + "step": 493 + }, + { + "epoch": 0.27317972350230413, + "grad_norm": 0.5060631632804871, + "learning_rate": 9.114286410443349e-06, + "loss": 0.4912, + "step": 494 + }, + { + "epoch": 0.27373271889400924, + "grad_norm": 0.5127936005592346, + "learning_rate": 9.108792560715487e-06, + "loss": 0.5004, + "step": 495 + }, + { + "epoch": 0.2742857142857143, + "grad_norm": 0.5518050193786621, + "learning_rate": 9.103283391694711e-06, + "loss": 0.5051, + "step": 496 + }, + { + "epoch": 0.27483870967741936, + "grad_norm": 0.5080561637878418, + "learning_rate": 9.097758923921495e-06, + "loss": 0.4679, + "step": 497 + }, + { + "epoch": 0.2753917050691244, + "grad_norm": 0.5666584372520447, + "learning_rate": 9.092219177993365e-06, + "loss": 0.4861, + "step": 498 + }, + { + "epoch": 0.2759447004608295, + "grad_norm": 0.48917266726493835, + "learning_rate": 9.086664174564804e-06, + "loss": 0.4616, + "step": 499 + }, + { + "epoch": 0.2764976958525346, + "grad_norm": 0.4538862407207489, + "learning_rate": 9.081093934347178e-06, + "loss": 0.4917, + "step": 500 + }, + { + "epoch": 0.27705069124423964, + "grad_norm": 0.6453654170036316, + "learning_rate": 9.075508478108674e-06, + "loss": 0.477, + "step": 501 + }, + { + "epoch": 0.2776036866359447, + "grad_norm": 0.5007541179656982, + "learning_rate": 9.069907826674199e-06, + "loss": 0.4748, + "step": 502 + }, + { + "epoch": 0.27815668202764976, + "grad_norm": 0.5134044289588928, + "learning_rate": 9.064292000925316e-06, + "loss": 0.5122, + "step": 503 + }, + { + "epoch": 0.2787096774193548, + "grad_norm": 0.5208799242973328, + "learning_rate": 9.05866102180017e-06, + "loss": 0.4908, + "step": 504 + }, + { + "epoch": 0.27926267281105993, + "grad_norm": 0.4694979786872864, + "learning_rate": 9.053014910293399e-06, + "loss": 0.5006, + "step": 505 + }, + { + "epoch": 0.279815668202765, + "grad_norm": 0.4957534670829773, + "learning_rate": 9.047353687456064e-06, + "loss": 0.4901, + "step": 506 + }, + { + "epoch": 0.28036866359447005, + "grad_norm": 0.5231026411056519, + "learning_rate": 9.041677374395563e-06, + "loss": 0.5062, + "step": 507 + }, + { + "epoch": 0.2809216589861751, + "grad_norm": 0.5049005150794983, + "learning_rate": 9.035985992275562e-06, + "loss": 0.501, + "step": 508 + }, + { + "epoch": 0.28147465437788016, + "grad_norm": 0.48721855878829956, + "learning_rate": 9.030279562315907e-06, + "loss": 0.4881, + "step": 509 + }, + { + "epoch": 0.2820276497695853, + "grad_norm": 0.47729742527008057, + "learning_rate": 9.02455810579255e-06, + "loss": 0.4734, + "step": 510 + }, + { + "epoch": 0.28258064516129033, + "grad_norm": 0.5377198457717896, + "learning_rate": 9.018821644037466e-06, + "loss": 0.4655, + "step": 511 + }, + { + "epoch": 0.2831336405529954, + "grad_norm": 0.5237789154052734, + "learning_rate": 9.01307019843858e-06, + "loss": 0.5081, + "step": 512 + }, + { + "epoch": 0.28368663594470045, + "grad_norm": 0.6095771193504333, + "learning_rate": 9.00730379043968e-06, + "loss": 0.4938, + "step": 513 + }, + { + "epoch": 0.2842396313364055, + "grad_norm": 0.5196386575698853, + "learning_rate": 9.001522441540342e-06, + "loss": 0.4901, + "step": 514 + }, + { + "epoch": 0.2847926267281106, + "grad_norm": 0.5163764357566833, + "learning_rate": 8.995726173295845e-06, + "loss": 0.4901, + "step": 515 + }, + { + "epoch": 0.2853456221198157, + "grad_norm": 0.5501850247383118, + "learning_rate": 8.989915007317095e-06, + "loss": 0.4884, + "step": 516 + }, + { + "epoch": 0.28589861751152074, + "grad_norm": 0.4907573461532593, + "learning_rate": 8.984088965270547e-06, + "loss": 0.4803, + "step": 517 + }, + { + "epoch": 0.2864516129032258, + "grad_norm": 0.4490540325641632, + "learning_rate": 8.978248068878113e-06, + "loss": 0.5025, + "step": 518 + }, + { + "epoch": 0.28700460829493085, + "grad_norm": 0.541654109954834, + "learning_rate": 8.97239233991709e-06, + "loss": 0.4844, + "step": 519 + }, + { + "epoch": 0.28755760368663597, + "grad_norm": 0.4671088457107544, + "learning_rate": 8.966521800220084e-06, + "loss": 0.4775, + "step": 520 + }, + { + "epoch": 0.288110599078341, + "grad_norm": 0.5631542801856995, + "learning_rate": 8.960636471674913e-06, + "loss": 0.4723, + "step": 521 + }, + { + "epoch": 0.2886635944700461, + "grad_norm": 0.5776397585868835, + "learning_rate": 8.954736376224536e-06, + "loss": 0.4949, + "step": 522 + }, + { + "epoch": 0.28921658986175114, + "grad_norm": 0.5042744874954224, + "learning_rate": 8.948821535866973e-06, + "loss": 0.4934, + "step": 523 + }, + { + "epoch": 0.2897695852534562, + "grad_norm": 0.5939375162124634, + "learning_rate": 8.942891972655217e-06, + "loss": 0.4884, + "step": 524 + }, + { + "epoch": 0.2903225806451613, + "grad_norm": 0.4696662127971649, + "learning_rate": 8.93694770869715e-06, + "loss": 0.5052, + "step": 525 + }, + { + "epoch": 0.29087557603686637, + "grad_norm": 0.45712941884994507, + "learning_rate": 8.930988766155472e-06, + "loss": 0.5488, + "step": 526 + }, + { + "epoch": 0.2914285714285714, + "grad_norm": 0.5518166422843933, + "learning_rate": 8.925015167247604e-06, + "loss": 0.5129, + "step": 527 + }, + { + "epoch": 0.2919815668202765, + "grad_norm": 0.446274071931839, + "learning_rate": 8.919026934245618e-06, + "loss": 0.504, + "step": 528 + }, + { + "epoch": 0.29253456221198154, + "grad_norm": 0.5146806836128235, + "learning_rate": 8.91302408947614e-06, + "loss": 0.4779, + "step": 529 + }, + { + "epoch": 0.29308755760368665, + "grad_norm": 0.4867715537548065, + "learning_rate": 8.907006655320287e-06, + "loss": 0.4743, + "step": 530 + }, + { + "epoch": 0.2936405529953917, + "grad_norm": 0.47047367691993713, + "learning_rate": 8.90097465421356e-06, + "loss": 0.4881, + "step": 531 + }, + { + "epoch": 0.29419354838709677, + "grad_norm": 0.46602970361709595, + "learning_rate": 8.894928108645772e-06, + "loss": 0.486, + "step": 532 + }, + { + "epoch": 0.29474654377880183, + "grad_norm": 0.45530185103416443, + "learning_rate": 8.888867041160975e-06, + "loss": 0.4726, + "step": 533 + }, + { + "epoch": 0.2952995391705069, + "grad_norm": 0.4860335886478424, + "learning_rate": 8.882791474357354e-06, + "loss": 0.4818, + "step": 534 + }, + { + "epoch": 0.295852534562212, + "grad_norm": 0.4780627191066742, + "learning_rate": 8.876701430887156e-06, + "loss": 0.4937, + "step": 535 + }, + { + "epoch": 0.29640552995391706, + "grad_norm": 0.488271027803421, + "learning_rate": 8.870596933456603e-06, + "loss": 0.4988, + "step": 536 + }, + { + "epoch": 0.2969585253456221, + "grad_norm": 0.48193061351776123, + "learning_rate": 8.864478004825812e-06, + "loss": 0.4925, + "step": 537 + }, + { + "epoch": 0.2975115207373272, + "grad_norm": 0.5294972658157349, + "learning_rate": 8.8583446678087e-06, + "loss": 0.4787, + "step": 538 + }, + { + "epoch": 0.29806451612903223, + "grad_norm": 0.48833131790161133, + "learning_rate": 8.852196945272906e-06, + "loss": 0.5055, + "step": 539 + }, + { + "epoch": 0.29861751152073734, + "grad_norm": 0.4671526551246643, + "learning_rate": 8.846034860139706e-06, + "loss": 0.4775, + "step": 540 + }, + { + "epoch": 0.2991705069124424, + "grad_norm": 0.6074989438056946, + "learning_rate": 8.839858435383918e-06, + "loss": 0.4894, + "step": 541 + }, + { + "epoch": 0.29972350230414746, + "grad_norm": 0.46650996804237366, + "learning_rate": 8.833667694033841e-06, + "loss": 0.4913, + "step": 542 + }, + { + "epoch": 0.3002764976958525, + "grad_norm": 0.7234087586402893, + "learning_rate": 8.827462659171134e-06, + "loss": 0.4794, + "step": 543 + }, + { + "epoch": 0.30082949308755763, + "grad_norm": 0.5737038850784302, + "learning_rate": 8.821243353930756e-06, + "loss": 0.5169, + "step": 544 + }, + { + "epoch": 0.3013824884792627, + "grad_norm": 0.6461325883865356, + "learning_rate": 8.815009801500875e-06, + "loss": 0.4846, + "step": 545 + }, + { + "epoch": 0.30193548387096775, + "grad_norm": 0.4608705937862396, + "learning_rate": 8.808762025122774e-06, + "loss": 0.4763, + "step": 546 + }, + { + "epoch": 0.3024884792626728, + "grad_norm": 0.6168321371078491, + "learning_rate": 8.802500048090769e-06, + "loss": 0.4906, + "step": 547 + }, + { + "epoch": 0.30304147465437786, + "grad_norm": 0.5032297968864441, + "learning_rate": 8.796223893752125e-06, + "loss": 0.4809, + "step": 548 + }, + { + "epoch": 0.303594470046083, + "grad_norm": 0.48676756024360657, + "learning_rate": 8.789933585506962e-06, + "loss": 0.4844, + "step": 549 + }, + { + "epoch": 0.30414746543778803, + "grad_norm": 0.5273764133453369, + "learning_rate": 8.783629146808175e-06, + "loss": 0.4979, + "step": 550 + }, + { + "epoch": 0.3047004608294931, + "grad_norm": 0.47446364164352417, + "learning_rate": 8.77731060116134e-06, + "loss": 0.4633, + "step": 551 + }, + { + "epoch": 0.30525345622119815, + "grad_norm": 0.5045877695083618, + "learning_rate": 8.77097797212463e-06, + "loss": 0.4907, + "step": 552 + }, + { + "epoch": 0.3058064516129032, + "grad_norm": 0.4806043803691864, + "learning_rate": 8.764631283308733e-06, + "loss": 0.4794, + "step": 553 + }, + { + "epoch": 0.3063594470046083, + "grad_norm": 0.5195322036743164, + "learning_rate": 8.758270558376752e-06, + "loss": 0.4921, + "step": 554 + }, + { + "epoch": 0.3069124423963134, + "grad_norm": 0.5429467558860779, + "learning_rate": 8.751895821044118e-06, + "loss": 0.4825, + "step": 555 + }, + { + "epoch": 0.30746543778801844, + "grad_norm": 0.48403802514076233, + "learning_rate": 8.745507095078515e-06, + "loss": 0.4997, + "step": 556 + }, + { + "epoch": 0.3080184331797235, + "grad_norm": 0.5179433822631836, + "learning_rate": 8.739104404299781e-06, + "loss": 0.4444, + "step": 557 + }, + { + "epoch": 0.30857142857142855, + "grad_norm": 0.5344632863998413, + "learning_rate": 8.732687772579816e-06, + "loss": 0.4695, + "step": 558 + }, + { + "epoch": 0.30912442396313367, + "grad_norm": 0.5367107391357422, + "learning_rate": 8.7262572238425e-06, + "loss": 0.453, + "step": 559 + }, + { + "epoch": 0.3096774193548387, + "grad_norm": 0.5187183618545532, + "learning_rate": 8.719812782063603e-06, + "loss": 0.4975, + "step": 560 + }, + { + "epoch": 0.3102304147465438, + "grad_norm": 0.4541904330253601, + "learning_rate": 8.713354471270691e-06, + "loss": 0.5075, + "step": 561 + }, + { + "epoch": 0.31078341013824884, + "grad_norm": 0.4493202865123749, + "learning_rate": 8.706882315543043e-06, + "loss": 0.458, + "step": 562 + }, + { + "epoch": 0.3113364055299539, + "grad_norm": 0.571712076663971, + "learning_rate": 8.700396339011554e-06, + "loss": 0.5057, + "step": 563 + }, + { + "epoch": 0.311889400921659, + "grad_norm": 0.4367184042930603, + "learning_rate": 8.69389656585865e-06, + "loss": 0.4725, + "step": 564 + }, + { + "epoch": 0.31244239631336407, + "grad_norm": 0.5127226710319519, + "learning_rate": 8.6873830203182e-06, + "loss": 0.4615, + "step": 565 + }, + { + "epoch": 0.3129953917050691, + "grad_norm": 0.5464975833892822, + "learning_rate": 8.680855726675419e-06, + "loss": 0.4661, + "step": 566 + }, + { + "epoch": 0.3135483870967742, + "grad_norm": 0.5116366147994995, + "learning_rate": 8.674314709266778e-06, + "loss": 0.4789, + "step": 567 + }, + { + "epoch": 0.31410138248847924, + "grad_norm": 0.5597204566001892, + "learning_rate": 8.667759992479922e-06, + "loss": 0.4903, + "step": 568 + }, + { + "epoch": 0.31465437788018435, + "grad_norm": 0.4036201536655426, + "learning_rate": 8.661191600753567e-06, + "loss": 0.5093, + "step": 569 + }, + { + "epoch": 0.3152073732718894, + "grad_norm": 0.6131529808044434, + "learning_rate": 8.65460955857742e-06, + "loss": 0.4572, + "step": 570 + }, + { + "epoch": 0.31576036866359447, + "grad_norm": 0.532192587852478, + "learning_rate": 8.648013890492081e-06, + "loss": 0.4679, + "step": 571 + }, + { + "epoch": 0.31631336405529953, + "grad_norm": 0.4397170841693878, + "learning_rate": 8.641404621088951e-06, + "loss": 0.4841, + "step": 572 + }, + { + "epoch": 0.3168663594470046, + "grad_norm": 0.44879838824272156, + "learning_rate": 8.634781775010144e-06, + "loss": 0.4853, + "step": 573 + }, + { + "epoch": 0.3174193548387097, + "grad_norm": 0.5257536768913269, + "learning_rate": 8.628145376948392e-06, + "loss": 0.4915, + "step": 574 + }, + { + "epoch": 0.31797235023041476, + "grad_norm": 0.4205341637134552, + "learning_rate": 8.621495451646958e-06, + "loss": 0.4895, + "step": 575 + }, + { + "epoch": 0.3185253456221198, + "grad_norm": 0.4227660298347473, + "learning_rate": 8.614832023899536e-06, + "loss": 0.488, + "step": 576 + }, + { + "epoch": 0.3190783410138249, + "grad_norm": 0.441709965467453, + "learning_rate": 8.608155118550167e-06, + "loss": 0.4756, + "step": 577 + }, + { + "epoch": 0.31963133640552993, + "grad_norm": 0.4320451021194458, + "learning_rate": 8.601464760493138e-06, + "loss": 0.4726, + "step": 578 + }, + { + "epoch": 0.32018433179723504, + "grad_norm": 0.3868134617805481, + "learning_rate": 8.594760974672898e-06, + "loss": 0.475, + "step": 579 + }, + { + "epoch": 0.3207373271889401, + "grad_norm": 0.43438073992729187, + "learning_rate": 8.588043786083952e-06, + "loss": 0.4874, + "step": 580 + }, + { + "epoch": 0.32129032258064516, + "grad_norm": 0.40946924686431885, + "learning_rate": 8.581313219770788e-06, + "loss": 0.4843, + "step": 581 + }, + { + "epoch": 0.3218433179723502, + "grad_norm": 0.4077948331832886, + "learning_rate": 8.57456930082776e-06, + "loss": 0.4656, + "step": 582 + }, + { + "epoch": 0.3223963133640553, + "grad_norm": 0.4392508566379547, + "learning_rate": 8.567812054399016e-06, + "loss": 0.4911, + "step": 583 + }, + { + "epoch": 0.3229493087557604, + "grad_norm": 0.4225884974002838, + "learning_rate": 8.561041505678385e-06, + "loss": 0.4907, + "step": 584 + }, + { + "epoch": 0.32350230414746545, + "grad_norm": 0.4444757103919983, + "learning_rate": 8.5542576799093e-06, + "loss": 0.484, + "step": 585 + }, + { + "epoch": 0.3240552995391705, + "grad_norm": 0.4226742088794708, + "learning_rate": 8.547460602384697e-06, + "loss": 0.4947, + "step": 586 + }, + { + "epoch": 0.32460829493087556, + "grad_norm": 0.5053589344024658, + "learning_rate": 8.540650298446912e-06, + "loss": 0.4984, + "step": 587 + }, + { + "epoch": 0.3251612903225806, + "grad_norm": 0.4462142586708069, + "learning_rate": 8.533826793487601e-06, + "loss": 0.4624, + "step": 588 + }, + { + "epoch": 0.32571428571428573, + "grad_norm": 0.5019125938415527, + "learning_rate": 8.526990112947636e-06, + "loss": 0.4597, + "step": 589 + }, + { + "epoch": 0.3262672811059908, + "grad_norm": 0.4495023787021637, + "learning_rate": 8.520140282317018e-06, + "loss": 0.4619, + "step": 590 + }, + { + "epoch": 0.32682027649769585, + "grad_norm": 0.45633986592292786, + "learning_rate": 8.51327732713477e-06, + "loss": 0.4997, + "step": 591 + }, + { + "epoch": 0.3273732718894009, + "grad_norm": 0.45402443408966064, + "learning_rate": 8.506401272988854e-06, + "loss": 0.4726, + "step": 592 + }, + { + "epoch": 0.32792626728110597, + "grad_norm": 0.4144057631492615, + "learning_rate": 8.499512145516068e-06, + "loss": 0.483, + "step": 593 + }, + { + "epoch": 0.3284792626728111, + "grad_norm": 0.506801962852478, + "learning_rate": 8.492609970401955e-06, + "loss": 0.4936, + "step": 594 + }, + { + "epoch": 0.32903225806451614, + "grad_norm": 0.45201730728149414, + "learning_rate": 8.4856947733807e-06, + "loss": 0.5055, + "step": 595 + }, + { + "epoch": 0.3295852534562212, + "grad_norm": 0.4555196762084961, + "learning_rate": 8.478766580235044e-06, + "loss": 0.4847, + "step": 596 + }, + { + "epoch": 0.33013824884792625, + "grad_norm": 0.49041327834129333, + "learning_rate": 8.471825416796185e-06, + "loss": 0.4662, + "step": 597 + }, + { + "epoch": 0.3306912442396313, + "grad_norm": 0.466457724571228, + "learning_rate": 8.464871308943673e-06, + "loss": 0.4794, + "step": 598 + }, + { + "epoch": 0.3312442396313364, + "grad_norm": 0.5273913145065308, + "learning_rate": 8.457904282605324e-06, + "loss": 0.482, + "step": 599 + }, + { + "epoch": 0.3317972350230415, + "grad_norm": 0.42694202065467834, + "learning_rate": 8.45092436375712e-06, + "loss": 0.4663, + "step": 600 + }, + { + "epoch": 0.33235023041474654, + "grad_norm": 0.467692494392395, + "learning_rate": 8.443931578423108e-06, + "loss": 0.4606, + "step": 601 + }, + { + "epoch": 0.3329032258064516, + "grad_norm": 0.5536178946495056, + "learning_rate": 8.436925952675312e-06, + "loss": 0.49, + "step": 602 + }, + { + "epoch": 0.33345622119815665, + "grad_norm": 0.5051153898239136, + "learning_rate": 8.429907512633626e-06, + "loss": 0.5041, + "step": 603 + }, + { + "epoch": 0.33400921658986177, + "grad_norm": 0.4975510835647583, + "learning_rate": 8.422876284465722e-06, + "loss": 0.491, + "step": 604 + }, + { + "epoch": 0.3345622119815668, + "grad_norm": 0.5645290613174438, + "learning_rate": 8.415832294386956e-06, + "loss": 0.4942, + "step": 605 + }, + { + "epoch": 0.3351152073732719, + "grad_norm": 0.5082492828369141, + "learning_rate": 8.408775568660256e-06, + "loss": 0.4829, + "step": 606 + }, + { + "epoch": 0.33566820276497694, + "grad_norm": 0.4222676753997803, + "learning_rate": 8.401706133596047e-06, + "loss": 0.4837, + "step": 607 + }, + { + "epoch": 0.33622119815668206, + "grad_norm": 0.5251336693763733, + "learning_rate": 8.394624015552124e-06, + "loss": 0.492, + "step": 608 + }, + { + "epoch": 0.3367741935483871, + "grad_norm": 0.4806731641292572, + "learning_rate": 8.387529240933585e-06, + "loss": 0.4791, + "step": 609 + }, + { + "epoch": 0.33732718894009217, + "grad_norm": 0.5003568530082703, + "learning_rate": 8.380421836192705e-06, + "loss": 0.4768, + "step": 610 + }, + { + "epoch": 0.33788018433179723, + "grad_norm": 0.48693132400512695, + "learning_rate": 8.373301827828858e-06, + "loss": 0.5219, + "step": 611 + }, + { + "epoch": 0.3384331797235023, + "grad_norm": 0.43921712040901184, + "learning_rate": 8.366169242388402e-06, + "loss": 0.4668, + "step": 612 + }, + { + "epoch": 0.3389861751152074, + "grad_norm": 0.5484280586242676, + "learning_rate": 8.359024106464594e-06, + "loss": 0.4669, + "step": 613 + }, + { + "epoch": 0.33953917050691246, + "grad_norm": 0.4520418643951416, + "learning_rate": 8.351866446697483e-06, + "loss": 0.4973, + "step": 614 + }, + { + "epoch": 0.3400921658986175, + "grad_norm": 0.588077962398529, + "learning_rate": 8.344696289773806e-06, + "loss": 0.4749, + "step": 615 + }, + { + "epoch": 0.3406451612903226, + "grad_norm": 0.4226244390010834, + "learning_rate": 8.337513662426904e-06, + "loss": 0.482, + "step": 616 + }, + { + "epoch": 0.34119815668202763, + "grad_norm": 0.5818718671798706, + "learning_rate": 8.330318591436606e-06, + "loss": 0.4963, + "step": 617 + }, + { + "epoch": 0.34175115207373274, + "grad_norm": 0.4959838092327118, + "learning_rate": 8.32311110362914e-06, + "loss": 0.4928, + "step": 618 + }, + { + "epoch": 0.3423041474654378, + "grad_norm": 0.43233931064605713, + "learning_rate": 8.315891225877026e-06, + "loss": 0.4619, + "step": 619 + }, + { + "epoch": 0.34285714285714286, + "grad_norm": 0.506004273891449, + "learning_rate": 8.308658985098983e-06, + "loss": 0.4638, + "step": 620 + }, + { + "epoch": 0.3434101382488479, + "grad_norm": 0.5699809193611145, + "learning_rate": 8.301414408259821e-06, + "loss": 0.4666, + "step": 621 + }, + { + "epoch": 0.343963133640553, + "grad_norm": 0.4087035059928894, + "learning_rate": 8.294157522370343e-06, + "loss": 0.4792, + "step": 622 + }, + { + "epoch": 0.3445161290322581, + "grad_norm": 0.482768714427948, + "learning_rate": 8.28688835448725e-06, + "loss": 0.4943, + "step": 623 + }, + { + "epoch": 0.34506912442396315, + "grad_norm": 0.4292857348918915, + "learning_rate": 8.27960693171303e-06, + "loss": 0.4809, + "step": 624 + }, + { + "epoch": 0.3456221198156682, + "grad_norm": 0.43942004442214966, + "learning_rate": 8.272313281195868e-06, + "loss": 0.488, + "step": 625 + }, + { + "epoch": 0.34617511520737326, + "grad_norm": 0.5329846739768982, + "learning_rate": 8.265007430129535e-06, + "loss": 0.494, + "step": 626 + }, + { + "epoch": 0.3467281105990783, + "grad_norm": 0.41792014241218567, + "learning_rate": 8.25768940575329e-06, + "loss": 0.4838, + "step": 627 + }, + { + "epoch": 0.34728110599078343, + "grad_norm": 0.522320568561554, + "learning_rate": 8.250359235351783e-06, + "loss": 0.4778, + "step": 628 + }, + { + "epoch": 0.3478341013824885, + "grad_norm": 0.4983687996864319, + "learning_rate": 8.243016946254947e-06, + "loss": 0.4761, + "step": 629 + }, + { + "epoch": 0.34838709677419355, + "grad_norm": 0.5216838717460632, + "learning_rate": 8.235662565837901e-06, + "loss": 0.4949, + "step": 630 + }, + { + "epoch": 0.3489400921658986, + "grad_norm": 0.5560440421104431, + "learning_rate": 8.22829612152084e-06, + "loss": 0.4653, + "step": 631 + }, + { + "epoch": 0.34949308755760367, + "grad_norm": 0.505742073059082, + "learning_rate": 8.220917640768943e-06, + "loss": 0.4981, + "step": 632 + }, + { + "epoch": 0.3500460829493088, + "grad_norm": 0.4594072103500366, + "learning_rate": 8.213527151092267e-06, + "loss": 0.4522, + "step": 633 + }, + { + "epoch": 0.35059907834101384, + "grad_norm": 0.40718746185302734, + "learning_rate": 8.20612468004564e-06, + "loss": 0.4728, + "step": 634 + }, + { + "epoch": 0.3511520737327189, + "grad_norm": 0.500005304813385, + "learning_rate": 8.19871025522856e-06, + "loss": 0.4667, + "step": 635 + }, + { + "epoch": 0.35170506912442395, + "grad_norm": 0.4374920427799225, + "learning_rate": 8.191283904285098e-06, + "loss": 0.4801, + "step": 636 + }, + { + "epoch": 0.352258064516129, + "grad_norm": 0.4499951899051666, + "learning_rate": 8.183845654903789e-06, + "loss": 0.4588, + "step": 637 + }, + { + "epoch": 0.3528110599078341, + "grad_norm": 0.41557565331459045, + "learning_rate": 8.17639553481753e-06, + "loss": 0.4953, + "step": 638 + }, + { + "epoch": 0.3533640552995392, + "grad_norm": 0.44742459058761597, + "learning_rate": 8.168933571803475e-06, + "loss": 0.4711, + "step": 639 + }, + { + "epoch": 0.35391705069124424, + "grad_norm": 0.4027085304260254, + "learning_rate": 8.161459793682937e-06, + "loss": 0.4872, + "step": 640 + }, + { + "epoch": 0.3544700460829493, + "grad_norm": 0.4093347191810608, + "learning_rate": 8.153974228321279e-06, + "loss": 0.477, + "step": 641 + }, + { + "epoch": 0.35502304147465436, + "grad_norm": 0.45894891023635864, + "learning_rate": 8.146476903627813e-06, + "loss": 0.4844, + "step": 642 + }, + { + "epoch": 0.35557603686635947, + "grad_norm": 0.44424736499786377, + "learning_rate": 8.138967847555693e-06, + "loss": 0.4754, + "step": 643 + }, + { + "epoch": 0.3561290322580645, + "grad_norm": 0.4633028209209442, + "learning_rate": 8.13144708810181e-06, + "loss": 0.4873, + "step": 644 + }, + { + "epoch": 0.3566820276497696, + "grad_norm": 0.46212074160575867, + "learning_rate": 8.123914653306695e-06, + "loss": 0.4576, + "step": 645 + }, + { + "epoch": 0.35723502304147464, + "grad_norm": 0.5635020732879639, + "learning_rate": 8.116370571254407e-06, + "loss": 0.4651, + "step": 646 + }, + { + "epoch": 0.3577880184331797, + "grad_norm": 0.43892374634742737, + "learning_rate": 8.108814870072434e-06, + "loss": 0.5018, + "step": 647 + }, + { + "epoch": 0.3583410138248848, + "grad_norm": 0.4993932843208313, + "learning_rate": 8.10124757793158e-06, + "loss": 0.4848, + "step": 648 + }, + { + "epoch": 0.35889400921658987, + "grad_norm": 0.5206483006477356, + "learning_rate": 8.093668723045866e-06, + "loss": 0.4845, + "step": 649 + }, + { + "epoch": 0.35944700460829493, + "grad_norm": 0.4889640808105469, + "learning_rate": 8.08607833367243e-06, + "loss": 0.4731, + "step": 650 + }, + { + "epoch": 0.36, + "grad_norm": 0.5545540452003479, + "learning_rate": 8.078476438111404e-06, + "loss": 0.4919, + "step": 651 + }, + { + "epoch": 0.36055299539170504, + "grad_norm": 0.47018927335739136, + "learning_rate": 8.07086306470583e-06, + "loss": 0.4902, + "step": 652 + }, + { + "epoch": 0.36110599078341016, + "grad_norm": 0.5213498473167419, + "learning_rate": 8.06323824184154e-06, + "loss": 0.4918, + "step": 653 + }, + { + "epoch": 0.3616589861751152, + "grad_norm": 0.4931851923465729, + "learning_rate": 8.055601997947056e-06, + "loss": 0.4868, + "step": 654 + }, + { + "epoch": 0.3622119815668203, + "grad_norm": 0.4996345341205597, + "learning_rate": 8.04795436149348e-06, + "loss": 0.4908, + "step": 655 + }, + { + "epoch": 0.36276497695852533, + "grad_norm": 0.48268017172813416, + "learning_rate": 8.04029536099439e-06, + "loss": 0.4975, + "step": 656 + }, + { + "epoch": 0.3633179723502304, + "grad_norm": 0.5365729928016663, + "learning_rate": 8.032625025005736e-06, + "loss": 0.4891, + "step": 657 + }, + { + "epoch": 0.3638709677419355, + "grad_norm": 0.5524755716323853, + "learning_rate": 8.024943382125732e-06, + "loss": 0.4833, + "step": 658 + }, + { + "epoch": 0.36442396313364056, + "grad_norm": 0.4499988257884979, + "learning_rate": 8.017250460994748e-06, + "loss": 0.4724, + "step": 659 + }, + { + "epoch": 0.3649769585253456, + "grad_norm": 0.44743797183036804, + "learning_rate": 8.0095462902952e-06, + "loss": 0.503, + "step": 660 + }, + { + "epoch": 0.3655299539170507, + "grad_norm": 0.502610445022583, + "learning_rate": 8.001830898751455e-06, + "loss": 0.4813, + "step": 661 + }, + { + "epoch": 0.36608294930875573, + "grad_norm": 0.41622281074523926, + "learning_rate": 7.99410431512971e-06, + "loss": 0.5157, + "step": 662 + }, + { + "epoch": 0.36663594470046085, + "grad_norm": 0.4399860203266144, + "learning_rate": 7.986366568237893e-06, + "loss": 0.4741, + "step": 663 + }, + { + "epoch": 0.3671889400921659, + "grad_norm": 0.43299001455307007, + "learning_rate": 7.97861768692555e-06, + "loss": 0.4727, + "step": 664 + }, + { + "epoch": 0.36774193548387096, + "grad_norm": 0.44035786390304565, + "learning_rate": 7.970857700083747e-06, + "loss": 0.4662, + "step": 665 + }, + { + "epoch": 0.368294930875576, + "grad_norm": 0.422536164522171, + "learning_rate": 7.96308663664495e-06, + "loss": 0.4969, + "step": 666 + }, + { + "epoch": 0.3688479262672811, + "grad_norm": 0.44590070843696594, + "learning_rate": 7.955304525582928e-06, + "loss": 0.4591, + "step": 667 + }, + { + "epoch": 0.3694009216589862, + "grad_norm": 0.43369054794311523, + "learning_rate": 7.947511395912635e-06, + "loss": 0.4941, + "step": 668 + }, + { + "epoch": 0.36995391705069125, + "grad_norm": 0.40852341055870056, + "learning_rate": 7.939707276690111e-06, + "loss": 0.4859, + "step": 669 + }, + { + "epoch": 0.3705069124423963, + "grad_norm": 0.3855592906475067, + "learning_rate": 7.93189219701237e-06, + "loss": 0.4834, + "step": 670 + }, + { + "epoch": 0.37105990783410137, + "grad_norm": 0.40031880140304565, + "learning_rate": 7.924066186017288e-06, + "loss": 0.5146, + "step": 671 + }, + { + "epoch": 0.3716129032258065, + "grad_norm": 0.42813777923583984, + "learning_rate": 7.916229272883499e-06, + "loss": 0.4875, + "step": 672 + }, + { + "epoch": 0.37216589861751154, + "grad_norm": 0.4341396689414978, + "learning_rate": 7.908381486830286e-06, + "loss": 0.4787, + "step": 673 + }, + { + "epoch": 0.3727188940092166, + "grad_norm": 0.40330180525779724, + "learning_rate": 7.90052285711747e-06, + "loss": 0.4866, + "step": 674 + }, + { + "epoch": 0.37327188940092165, + "grad_norm": 0.4432542026042938, + "learning_rate": 7.892653413045298e-06, + "loss": 0.4574, + "step": 675 + }, + { + "epoch": 0.3738248847926267, + "grad_norm": 0.4223710298538208, + "learning_rate": 7.884773183954344e-06, + "loss": 0.4619, + "step": 676 + }, + { + "epoch": 0.3743778801843318, + "grad_norm": 0.39498183131217957, + "learning_rate": 7.876882199225388e-06, + "loss": 0.4717, + "step": 677 + }, + { + "epoch": 0.3749308755760369, + "grad_norm": 0.4838169813156128, + "learning_rate": 7.868980488279316e-06, + "loss": 0.4756, + "step": 678 + }, + { + "epoch": 0.37548387096774194, + "grad_norm": 0.3836471140384674, + "learning_rate": 7.861068080576998e-06, + "loss": 0.462, + "step": 679 + }, + { + "epoch": 0.376036866359447, + "grad_norm": 0.503296971321106, + "learning_rate": 7.853145005619199e-06, + "loss": 0.4963, + "step": 680 + }, + { + "epoch": 0.37658986175115206, + "grad_norm": 0.3729606568813324, + "learning_rate": 7.84521129294644e-06, + "loss": 0.4684, + "step": 681 + }, + { + "epoch": 0.37714285714285717, + "grad_norm": 0.5140215158462524, + "learning_rate": 7.837266972138918e-06, + "loss": 0.5041, + "step": 682 + }, + { + "epoch": 0.3776958525345622, + "grad_norm": 0.4366711676120758, + "learning_rate": 7.82931207281637e-06, + "loss": 0.5007, + "step": 683 + }, + { + "epoch": 0.3782488479262673, + "grad_norm": 0.43705347180366516, + "learning_rate": 7.821346624637984e-06, + "loss": 0.4988, + "step": 684 + }, + { + "epoch": 0.37880184331797234, + "grad_norm": 0.4190782308578491, + "learning_rate": 7.813370657302273e-06, + "loss": 0.4815, + "step": 685 + }, + { + "epoch": 0.3793548387096774, + "grad_norm": 0.43545421957969666, + "learning_rate": 7.80538420054697e-06, + "loss": 0.486, + "step": 686 + }, + { + "epoch": 0.3799078341013825, + "grad_norm": 0.4012760818004608, + "learning_rate": 7.797387284148919e-06, + "loss": 0.456, + "step": 687 + }, + { + "epoch": 0.38046082949308757, + "grad_norm": 0.45634332299232483, + "learning_rate": 7.789379937923958e-06, + "loss": 0.4827, + "step": 688 + }, + { + "epoch": 0.38101382488479263, + "grad_norm": 0.45957717299461365, + "learning_rate": 7.781362191726818e-06, + "loss": 0.4976, + "step": 689 + }, + { + "epoch": 0.3815668202764977, + "grad_norm": 0.4902954399585724, + "learning_rate": 7.773334075451e-06, + "loss": 0.4952, + "step": 690 + }, + { + "epoch": 0.38211981566820274, + "grad_norm": 0.4607703387737274, + "learning_rate": 7.765295619028672e-06, + "loss": 0.4995, + "step": 691 + }, + { + "epoch": 0.38267281105990786, + "grad_norm": 0.49557220935821533, + "learning_rate": 7.757246852430553e-06, + "loss": 0.4921, + "step": 692 + }, + { + "epoch": 0.3832258064516129, + "grad_norm": 0.4447820782661438, + "learning_rate": 7.749187805665801e-06, + "loss": 0.4589, + "step": 693 + }, + { + "epoch": 0.383778801843318, + "grad_norm": 0.4028923809528351, + "learning_rate": 7.741118508781906e-06, + "loss": 0.4732, + "step": 694 + }, + { + "epoch": 0.38433179723502303, + "grad_norm": 0.4806511700153351, + "learning_rate": 7.733038991864576e-06, + "loss": 0.4968, + "step": 695 + }, + { + "epoch": 0.3848847926267281, + "grad_norm": 0.4322826862335205, + "learning_rate": 7.724949285037619e-06, + "loss": 0.4904, + "step": 696 + }, + { + "epoch": 0.3854377880184332, + "grad_norm": 0.44291332364082336, + "learning_rate": 7.716849418462836e-06, + "loss": 0.4829, + "step": 697 + }, + { + "epoch": 0.38599078341013826, + "grad_norm": 0.4092819094657898, + "learning_rate": 7.70873942233991e-06, + "loss": 0.4809, + "step": 698 + }, + { + "epoch": 0.3865437788018433, + "grad_norm": 0.4542578160762787, + "learning_rate": 7.700619326906288e-06, + "loss": 0.4713, + "step": 699 + }, + { + "epoch": 0.3870967741935484, + "grad_norm": 0.4618943929672241, + "learning_rate": 7.69248916243708e-06, + "loss": 0.4443, + "step": 700 + }, + { + "epoch": 0.38764976958525343, + "grad_norm": 0.43175920844078064, + "learning_rate": 7.684348959244924e-06, + "loss": 0.4761, + "step": 701 + }, + { + "epoch": 0.38820276497695855, + "grad_norm": 0.4224050045013428, + "learning_rate": 7.676198747679898e-06, + "loss": 0.4943, + "step": 702 + }, + { + "epoch": 0.3887557603686636, + "grad_norm": 0.4592989683151245, + "learning_rate": 7.66803855812939e-06, + "loss": 0.4529, + "step": 703 + }, + { + "epoch": 0.38930875576036866, + "grad_norm": 0.4457801878452301, + "learning_rate": 7.659868421017989e-06, + "loss": 0.4785, + "step": 704 + }, + { + "epoch": 0.3898617511520737, + "grad_norm": 0.4617078900337219, + "learning_rate": 7.651688366807378e-06, + "loss": 0.4755, + "step": 705 + }, + { + "epoch": 0.3904147465437788, + "grad_norm": 0.43977999687194824, + "learning_rate": 7.64349842599621e-06, + "loss": 0.5098, + "step": 706 + }, + { + "epoch": 0.3909677419354839, + "grad_norm": 0.4340912699699402, + "learning_rate": 7.635298629120002e-06, + "loss": 0.4452, + "step": 707 + }, + { + "epoch": 0.39152073732718895, + "grad_norm": 0.5204182863235474, + "learning_rate": 7.6270890067510196e-06, + "loss": 0.4677, + "step": 708 + }, + { + "epoch": 0.392073732718894, + "grad_norm": 0.47488752007484436, + "learning_rate": 7.618869589498157e-06, + "loss": 0.465, + "step": 709 + }, + { + "epoch": 0.39262672811059907, + "grad_norm": 0.4327450692653656, + "learning_rate": 7.610640408006832e-06, + "loss": 0.4945, + "step": 710 + }, + { + "epoch": 0.3931797235023041, + "grad_norm": 0.5072281360626221, + "learning_rate": 7.602401492958868e-06, + "loss": 0.4987, + "step": 711 + }, + { + "epoch": 0.39373271889400924, + "grad_norm": 0.4239625632762909, + "learning_rate": 7.594152875072376e-06, + "loss": 0.4715, + "step": 712 + }, + { + "epoch": 0.3942857142857143, + "grad_norm": 0.4698258936405182, + "learning_rate": 7.5858945851016455e-06, + "loss": 0.4779, + "step": 713 + }, + { + "epoch": 0.39483870967741935, + "grad_norm": 0.42160850763320923, + "learning_rate": 7.577626653837028e-06, + "loss": 0.45, + "step": 714 + }, + { + "epoch": 0.3953917050691244, + "grad_norm": 0.4377395808696747, + "learning_rate": 7.5693491121048194e-06, + "loss": 0.5144, + "step": 715 + }, + { + "epoch": 0.39594470046082947, + "grad_norm": 0.4012657403945923, + "learning_rate": 7.561061990767149e-06, + "loss": 0.4812, + "step": 716 + }, + { + "epoch": 0.3964976958525346, + "grad_norm": 0.41947153210639954, + "learning_rate": 7.552765320721865e-06, + "loss": 0.4407, + "step": 717 + }, + { + "epoch": 0.39705069124423964, + "grad_norm": 0.48914283514022827, + "learning_rate": 7.544459132902409e-06, + "loss": 0.4689, + "step": 718 + }, + { + "epoch": 0.3976036866359447, + "grad_norm": 0.3954436480998993, + "learning_rate": 7.5361434582777205e-06, + "loss": 0.4866, + "step": 719 + }, + { + "epoch": 0.39815668202764976, + "grad_norm": 0.45554065704345703, + "learning_rate": 7.527818327852101e-06, + "loss": 0.4499, + "step": 720 + }, + { + "epoch": 0.3987096774193548, + "grad_norm": 0.3696426451206207, + "learning_rate": 7.519483772665113e-06, + "loss": 0.4963, + "step": 721 + }, + { + "epoch": 0.3992626728110599, + "grad_norm": 0.42578935623168945, + "learning_rate": 7.511139823791452e-06, + "loss": 0.4642, + "step": 722 + }, + { + "epoch": 0.399815668202765, + "grad_norm": 0.3758449852466583, + "learning_rate": 7.502786512340841e-06, + "loss": 0.516, + "step": 723 + }, + { + "epoch": 0.40036866359447004, + "grad_norm": 0.4170108139514923, + "learning_rate": 7.494423869457912e-06, + "loss": 0.4824, + "step": 724 + }, + { + "epoch": 0.4009216589861751, + "grad_norm": 0.3795914053916931, + "learning_rate": 7.4860519263220865e-06, + "loss": 0.4744, + "step": 725 + }, + { + "epoch": 0.40147465437788016, + "grad_norm": 0.389635294675827, + "learning_rate": 7.477670714147461e-06, + "loss": 0.4957, + "step": 726 + }, + { + "epoch": 0.40202764976958527, + "grad_norm": 0.44072335958480835, + "learning_rate": 7.469280264182689e-06, + "loss": 0.5172, + "step": 727 + }, + { + "epoch": 0.40258064516129033, + "grad_norm": 0.372860312461853, + "learning_rate": 7.460880607710872e-06, + "loss": 0.4465, + "step": 728 + }, + { + "epoch": 0.4031336405529954, + "grad_norm": 0.47300985455513, + "learning_rate": 7.452471776049432e-06, + "loss": 0.4636, + "step": 729 + }, + { + "epoch": 0.40368663594470044, + "grad_norm": 0.44356971979141235, + "learning_rate": 7.444053800550004e-06, + "loss": 0.4911, + "step": 730 + }, + { + "epoch": 0.4042396313364055, + "grad_norm": 0.4640626311302185, + "learning_rate": 7.435626712598312e-06, + "loss": 0.4873, + "step": 731 + }, + { + "epoch": 0.4047926267281106, + "grad_norm": 0.4652840197086334, + "learning_rate": 7.427190543614053e-06, + "loss": 0.4641, + "step": 732 + }, + { + "epoch": 0.4053456221198157, + "grad_norm": 0.46554264426231384, + "learning_rate": 7.418745325050787e-06, + "loss": 0.4797, + "step": 733 + }, + { + "epoch": 0.40589861751152073, + "grad_norm": 0.42912858724594116, + "learning_rate": 7.410291088395812e-06, + "loss": 0.5064, + "step": 734 + }, + { + "epoch": 0.4064516129032258, + "grad_norm": 0.41365447640419006, + "learning_rate": 7.401827865170047e-06, + "loss": 0.4873, + "step": 735 + }, + { + "epoch": 0.4070046082949309, + "grad_norm": 0.5490487813949585, + "learning_rate": 7.3933556869279235e-06, + "loss": 0.4583, + "step": 736 + }, + { + "epoch": 0.40755760368663596, + "grad_norm": 0.39917993545532227, + "learning_rate": 7.384874585257255e-06, + "loss": 0.4625, + "step": 737 + }, + { + "epoch": 0.408110599078341, + "grad_norm": 0.4784587621688843, + "learning_rate": 7.3763845917791245e-06, + "loss": 0.5073, + "step": 738 + }, + { + "epoch": 0.4086635944700461, + "grad_norm": 0.4989032447338104, + "learning_rate": 7.367885738147773e-06, + "loss": 0.4652, + "step": 739 + }, + { + "epoch": 0.40921658986175113, + "grad_norm": 0.4629876911640167, + "learning_rate": 7.359378056050472e-06, + "loss": 0.4601, + "step": 740 + }, + { + "epoch": 0.40976958525345625, + "grad_norm": 0.4878072440624237, + "learning_rate": 7.35086157720741e-06, + "loss": 0.4713, + "step": 741 + }, + { + "epoch": 0.4103225806451613, + "grad_norm": 0.49611517786979675, + "learning_rate": 7.3423363333715726e-06, + "loss": 0.4727, + "step": 742 + }, + { + "epoch": 0.41087557603686636, + "grad_norm": 0.5412954688072205, + "learning_rate": 7.333802356328624e-06, + "loss": 0.4795, + "step": 743 + }, + { + "epoch": 0.4114285714285714, + "grad_norm": 0.387992262840271, + "learning_rate": 7.3252596778967954e-06, + "loss": 0.492, + "step": 744 + }, + { + "epoch": 0.4119815668202765, + "grad_norm": 0.47084909677505493, + "learning_rate": 7.316708329926754e-06, + "loss": 0.4546, + "step": 745 + }, + { + "epoch": 0.4125345622119816, + "grad_norm": 0.5451803803443909, + "learning_rate": 7.308148344301491e-06, + "loss": 0.5053, + "step": 746 + }, + { + "epoch": 0.41308755760368665, + "grad_norm": 0.4724251329898834, + "learning_rate": 7.2995797529362075e-06, + "loss": 0.4937, + "step": 747 + }, + { + "epoch": 0.4136405529953917, + "grad_norm": 0.48225685954093933, + "learning_rate": 7.291002587778184e-06, + "loss": 0.4939, + "step": 748 + }, + { + "epoch": 0.41419354838709677, + "grad_norm": 0.4945637881755829, + "learning_rate": 7.282416880806673e-06, + "loss": 0.4934, + "step": 749 + }, + { + "epoch": 0.4147465437788018, + "grad_norm": 0.48181742429733276, + "learning_rate": 7.273822664032771e-06, + "loss": 0.4763, + "step": 750 + }, + { + "epoch": 0.41529953917050694, + "grad_norm": 0.41117456555366516, + "learning_rate": 7.265219969499302e-06, + "loss": 0.4928, + "step": 751 + }, + { + "epoch": 0.415852534562212, + "grad_norm": 0.5093626976013184, + "learning_rate": 7.256608829280705e-06, + "loss": 0.4563, + "step": 752 + }, + { + "epoch": 0.41640552995391705, + "grad_norm": 0.45837345719337463, + "learning_rate": 7.2479892754828995e-06, + "loss": 0.4736, + "step": 753 + }, + { + "epoch": 0.4169585253456221, + "grad_norm": 0.4117204546928406, + "learning_rate": 7.239361340243179e-06, + "loss": 0.4788, + "step": 754 + }, + { + "epoch": 0.41751152073732717, + "grad_norm": 0.44084909558296204, + "learning_rate": 7.230725055730088e-06, + "loss": 0.4781, + "step": 755 + }, + { + "epoch": 0.4180645161290323, + "grad_norm": 0.5877032279968262, + "learning_rate": 7.222080454143296e-06, + "loss": 0.4782, + "step": 756 + }, + { + "epoch": 0.41861751152073734, + "grad_norm": 0.3853990137577057, + "learning_rate": 7.213427567713485e-06, + "loss": 0.4498, + "step": 757 + }, + { + "epoch": 0.4191705069124424, + "grad_norm": 0.4723830819129944, + "learning_rate": 7.204766428702226e-06, + "loss": 0.4601, + "step": 758 + }, + { + "epoch": 0.41972350230414746, + "grad_norm": 0.43980076909065247, + "learning_rate": 7.196097069401857e-06, + "loss": 0.4783, + "step": 759 + }, + { + "epoch": 0.4202764976958525, + "grad_norm": 0.4152413606643677, + "learning_rate": 7.1874195221353706e-06, + "loss": 0.4873, + "step": 760 + }, + { + "epoch": 0.4208294930875576, + "grad_norm": 0.3723086714744568, + "learning_rate": 7.17873381925628e-06, + "loss": 0.4664, + "step": 761 + }, + { + "epoch": 0.4213824884792627, + "grad_norm": 0.44318974018096924, + "learning_rate": 7.17003999314851e-06, + "loss": 0.4865, + "step": 762 + }, + { + "epoch": 0.42193548387096774, + "grad_norm": 0.40795907378196716, + "learning_rate": 7.161338076226272e-06, + "loss": 0.4795, + "step": 763 + }, + { + "epoch": 0.4224884792626728, + "grad_norm": 0.4073677659034729, + "learning_rate": 7.1526281009339426e-06, + "loss": 0.4815, + "step": 764 + }, + { + "epoch": 0.42304147465437786, + "grad_norm": 0.4583486020565033, + "learning_rate": 7.143910099745944e-06, + "loss": 0.4872, + "step": 765 + }, + { + "epoch": 0.42359447004608297, + "grad_norm": 0.4036937654018402, + "learning_rate": 7.13518410516662e-06, + "loss": 0.4616, + "step": 766 + }, + { + "epoch": 0.42414746543778803, + "grad_norm": 0.42568910121917725, + "learning_rate": 7.126450149730122e-06, + "loss": 0.489, + "step": 767 + }, + { + "epoch": 0.4247004608294931, + "grad_norm": 0.42301973700523376, + "learning_rate": 7.1177082660002784e-06, + "loss": 0.4685, + "step": 768 + }, + { + "epoch": 0.42525345622119815, + "grad_norm": 0.45005002617836, + "learning_rate": 7.108958486570479e-06, + "loss": 0.4555, + "step": 769 + }, + { + "epoch": 0.4258064516129032, + "grad_norm": 0.45795372128486633, + "learning_rate": 7.1002008440635515e-06, + "loss": 0.4678, + "step": 770 + }, + { + "epoch": 0.4263594470046083, + "grad_norm": 0.41580426692962646, + "learning_rate": 7.091435371131642e-06, + "loss": 0.5058, + "step": 771 + }, + { + "epoch": 0.4269124423963134, + "grad_norm": 0.4396505355834961, + "learning_rate": 7.082662100456089e-06, + "loss": 0.492, + "step": 772 + }, + { + "epoch": 0.42746543778801843, + "grad_norm": 0.4275103211402893, + "learning_rate": 7.073881064747308e-06, + "loss": 0.4624, + "step": 773 + }, + { + "epoch": 0.4280184331797235, + "grad_norm": 0.37376993894577026, + "learning_rate": 7.0650922967446634e-06, + "loss": 0.4626, + "step": 774 + }, + { + "epoch": 0.42857142857142855, + "grad_norm": 0.40532317757606506, + "learning_rate": 7.056295829216347e-06, + "loss": 0.5042, + "step": 775 + }, + { + "epoch": 0.42912442396313366, + "grad_norm": 0.39862802624702454, + "learning_rate": 7.047491694959263e-06, + "loss": 0.4398, + "step": 776 + }, + { + "epoch": 0.4296774193548387, + "grad_norm": 0.4326246380805969, + "learning_rate": 7.038679926798895e-06, + "loss": 0.4902, + "step": 777 + }, + { + "epoch": 0.4302304147465438, + "grad_norm": 0.43664461374282837, + "learning_rate": 7.029860557589191e-06, + "loss": 0.4817, + "step": 778 + }, + { + "epoch": 0.43078341013824883, + "grad_norm": 0.4556635320186615, + "learning_rate": 7.02103362021244e-06, + "loss": 0.4825, + "step": 779 + }, + { + "epoch": 0.4313364055299539, + "grad_norm": 0.40778669714927673, + "learning_rate": 7.012199147579146e-06, + "loss": 0.496, + "step": 780 + }, + { + "epoch": 0.431889400921659, + "grad_norm": 0.41946834325790405, + "learning_rate": 7.00335717262791e-06, + "loss": 0.482, + "step": 781 + }, + { + "epoch": 0.43244239631336406, + "grad_norm": 0.4511401653289795, + "learning_rate": 6.994507728325304e-06, + "loss": 0.4763, + "step": 782 + }, + { + "epoch": 0.4329953917050691, + "grad_norm": 0.43290090560913086, + "learning_rate": 6.985650847665747e-06, + "loss": 0.4804, + "step": 783 + }, + { + "epoch": 0.4335483870967742, + "grad_norm": 0.4241175949573517, + "learning_rate": 6.976786563671386e-06, + "loss": 0.4776, + "step": 784 + }, + { + "epoch": 0.43410138248847924, + "grad_norm": 0.4226067364215851, + "learning_rate": 6.967914909391971e-06, + "loss": 0.4839, + "step": 785 + }, + { + "epoch": 0.43465437788018435, + "grad_norm": 0.39866408705711365, + "learning_rate": 6.959035917904728e-06, + "loss": 0.4596, + "step": 786 + }, + { + "epoch": 0.4352073732718894, + "grad_norm": 0.37730705738067627, + "learning_rate": 6.950149622314245e-06, + "loss": 0.4784, + "step": 787 + }, + { + "epoch": 0.43576036866359447, + "grad_norm": 0.41297316551208496, + "learning_rate": 6.941256055752337e-06, + "loss": 0.4887, + "step": 788 + }, + { + "epoch": 0.4363133640552995, + "grad_norm": 0.4532250761985779, + "learning_rate": 6.932355251377932e-06, + "loss": 0.4897, + "step": 789 + }, + { + "epoch": 0.4368663594470046, + "grad_norm": 0.4239371418952942, + "learning_rate": 6.923447242376942e-06, + "loss": 0.4873, + "step": 790 + }, + { + "epoch": 0.4374193548387097, + "grad_norm": 0.33959221839904785, + "learning_rate": 6.91453206196214e-06, + "loss": 0.4697, + "step": 791 + }, + { + "epoch": 0.43797235023041475, + "grad_norm": 0.4459823668003082, + "learning_rate": 6.90560974337304e-06, + "loss": 0.4799, + "step": 792 + }, + { + "epoch": 0.4385253456221198, + "grad_norm": 0.4143071174621582, + "learning_rate": 6.896680319875767e-06, + "loss": 0.4556, + "step": 793 + }, + { + "epoch": 0.43907834101382487, + "grad_norm": 0.4591521620750427, + "learning_rate": 6.887743824762937e-06, + "loss": 0.4841, + "step": 794 + }, + { + "epoch": 0.4396313364055299, + "grad_norm": 0.41245412826538086, + "learning_rate": 6.878800291353533e-06, + "loss": 0.5083, + "step": 795 + }, + { + "epoch": 0.44018433179723504, + "grad_norm": 0.4207186996936798, + "learning_rate": 6.8698497529927784e-06, + "loss": 0.4653, + "step": 796 + }, + { + "epoch": 0.4407373271889401, + "grad_norm": 0.42010927200317383, + "learning_rate": 6.860892243052015e-06, + "loss": 0.4595, + "step": 797 + }, + { + "epoch": 0.44129032258064516, + "grad_norm": 0.41488680243492126, + "learning_rate": 6.851927794928578e-06, + "loss": 0.4685, + "step": 798 + }, + { + "epoch": 0.4418433179723502, + "grad_norm": 0.43417924642562866, + "learning_rate": 6.842956442045668e-06, + "loss": 0.4753, + "step": 799 + }, + { + "epoch": 0.4423963133640553, + "grad_norm": 0.35748791694641113, + "learning_rate": 6.833978217852233e-06, + "loss": 0.4677, + "step": 800 + }, + { + "epoch": 0.4429493087557604, + "grad_norm": 0.4993172883987427, + "learning_rate": 6.824993155822838e-06, + "loss": 0.4729, + "step": 801 + }, + { + "epoch": 0.44350230414746544, + "grad_norm": 0.36592617630958557, + "learning_rate": 6.816001289457542e-06, + "loss": 0.4565, + "step": 802 + }, + { + "epoch": 0.4440552995391705, + "grad_norm": 0.3910996615886688, + "learning_rate": 6.807002652281777e-06, + "loss": 0.4619, + "step": 803 + }, + { + "epoch": 0.44460829493087556, + "grad_norm": 0.41243669390678406, + "learning_rate": 6.797997277846215e-06, + "loss": 0.46, + "step": 804 + }, + { + "epoch": 0.44516129032258067, + "grad_norm": 0.3570329546928406, + "learning_rate": 6.788985199726648e-06, + "loss": 0.4624, + "step": 805 + }, + { + "epoch": 0.44571428571428573, + "grad_norm": 0.4660005271434784, + "learning_rate": 6.779966451523866e-06, + "loss": 0.494, + "step": 806 + }, + { + "epoch": 0.4462672811059908, + "grad_norm": 0.38170328736305237, + "learning_rate": 6.770941066863523e-06, + "loss": 0.4848, + "step": 807 + }, + { + "epoch": 0.44682027649769585, + "grad_norm": 0.410656601190567, + "learning_rate": 6.761909079396018e-06, + "loss": 0.4783, + "step": 808 + }, + { + "epoch": 0.4473732718894009, + "grad_norm": 0.39199960231781006, + "learning_rate": 6.752870522796372e-06, + "loss": 0.4706, + "step": 809 + }, + { + "epoch": 0.447926267281106, + "grad_norm": 0.4048680067062378, + "learning_rate": 6.743825430764091e-06, + "loss": 0.4841, + "step": 810 + }, + { + "epoch": 0.4484792626728111, + "grad_norm": 0.3699415922164917, + "learning_rate": 6.734773837023054e-06, + "loss": 0.4479, + "step": 811 + }, + { + "epoch": 0.44903225806451613, + "grad_norm": 0.3935936391353607, + "learning_rate": 6.725715775321379e-06, + "loss": 0.4802, + "step": 812 + }, + { + "epoch": 0.4495852534562212, + "grad_norm": 0.39852485060691833, + "learning_rate": 6.7166512794312986e-06, + "loss": 0.4652, + "step": 813 + }, + { + "epoch": 0.45013824884792625, + "grad_norm": 0.38947340846061707, + "learning_rate": 6.707580383149035e-06, + "loss": 0.4925, + "step": 814 + }, + { + "epoch": 0.45069124423963136, + "grad_norm": 0.4168023467063904, + "learning_rate": 6.698503120294674e-06, + "loss": 0.4829, + "step": 815 + }, + { + "epoch": 0.4512442396313364, + "grad_norm": 0.39369869232177734, + "learning_rate": 6.6894195247120396e-06, + "loss": 0.4562, + "step": 816 + }, + { + "epoch": 0.4517972350230415, + "grad_norm": 0.3529163599014282, + "learning_rate": 6.680329630268565e-06, + "loss": 0.4628, + "step": 817 + }, + { + "epoch": 0.45235023041474653, + "grad_norm": 0.3897452652454376, + "learning_rate": 6.671233470855171e-06, + "loss": 0.4635, + "step": 818 + }, + { + "epoch": 0.4529032258064516, + "grad_norm": 0.4104883074760437, + "learning_rate": 6.662131080386132e-06, + "loss": 0.4716, + "step": 819 + }, + { + "epoch": 0.4534562211981567, + "grad_norm": 0.39856767654418945, + "learning_rate": 6.653022492798959e-06, + "loss": 0.4666, + "step": 820 + }, + { + "epoch": 0.45400921658986176, + "grad_norm": 0.3940970301628113, + "learning_rate": 6.643907742054267e-06, + "loss": 0.4899, + "step": 821 + }, + { + "epoch": 0.4545622119815668, + "grad_norm": 0.4091474413871765, + "learning_rate": 6.63478686213565e-06, + "loss": 0.4829, + "step": 822 + }, + { + "epoch": 0.4551152073732719, + "grad_norm": 0.4372076690196991, + "learning_rate": 6.625659887049554e-06, + "loss": 0.4872, + "step": 823 + }, + { + "epoch": 0.45566820276497694, + "grad_norm": 0.4298078417778015, + "learning_rate": 6.6165268508251455e-06, + "loss": 0.4748, + "step": 824 + }, + { + "epoch": 0.45622119815668205, + "grad_norm": 0.4123404622077942, + "learning_rate": 6.607387787514199e-06, + "loss": 0.4784, + "step": 825 + }, + { + "epoch": 0.4567741935483871, + "grad_norm": 0.48749467730522156, + "learning_rate": 6.598242731190955e-06, + "loss": 0.4617, + "step": 826 + }, + { + "epoch": 0.45732718894009217, + "grad_norm": 0.44510942697525024, + "learning_rate": 6.589091715951996e-06, + "loss": 0.5012, + "step": 827 + }, + { + "epoch": 0.4578801843317972, + "grad_norm": 0.4342193603515625, + "learning_rate": 6.579934775916128e-06, + "loss": 0.4686, + "step": 828 + }, + { + "epoch": 0.4584331797235023, + "grad_norm": 0.46518048644065857, + "learning_rate": 6.570771945224243e-06, + "loss": 0.4695, + "step": 829 + }, + { + "epoch": 0.4589861751152074, + "grad_norm": 0.4383966326713562, + "learning_rate": 6.561603258039195e-06, + "loss": 0.4719, + "step": 830 + }, + { + "epoch": 0.45953917050691245, + "grad_norm": 0.41184401512145996, + "learning_rate": 6.552428748545677e-06, + "loss": 0.4789, + "step": 831 + }, + { + "epoch": 0.4600921658986175, + "grad_norm": 0.4380422830581665, + "learning_rate": 6.543248450950087e-06, + "loss": 0.4907, + "step": 832 + }, + { + "epoch": 0.46064516129032257, + "grad_norm": 0.453457236289978, + "learning_rate": 6.534062399480405e-06, + "loss": 0.4747, + "step": 833 + }, + { + "epoch": 0.4611981566820276, + "grad_norm": 0.36671724915504456, + "learning_rate": 6.524870628386064e-06, + "loss": 0.4905, + "step": 834 + }, + { + "epoch": 0.46175115207373274, + "grad_norm": 0.42992615699768066, + "learning_rate": 6.515673171937822e-06, + "loss": 0.4634, + "step": 835 + }, + { + "epoch": 0.4623041474654378, + "grad_norm": 0.40917208790779114, + "learning_rate": 6.5064700644276315e-06, + "loss": 0.4661, + "step": 836 + }, + { + "epoch": 0.46285714285714286, + "grad_norm": 0.4510115385055542, + "learning_rate": 6.497261340168519e-06, + "loss": 0.4504, + "step": 837 + }, + { + "epoch": 0.4634101382488479, + "grad_norm": 0.45596301555633545, + "learning_rate": 6.4880470334944515e-06, + "loss": 0.4551, + "step": 838 + }, + { + "epoch": 0.46396313364055297, + "grad_norm": 0.4473353624343872, + "learning_rate": 6.478827178760205e-06, + "loss": 0.4834, + "step": 839 + }, + { + "epoch": 0.4645161290322581, + "grad_norm": 0.4285813271999359, + "learning_rate": 6.469601810341247e-06, + "loss": 0.498, + "step": 840 + }, + { + "epoch": 0.46506912442396314, + "grad_norm": 0.38439249992370605, + "learning_rate": 6.4603709626336e-06, + "loss": 0.4754, + "step": 841 + }, + { + "epoch": 0.4656221198156682, + "grad_norm": 0.3741195499897003, + "learning_rate": 6.451134670053716e-06, + "loss": 0.4887, + "step": 842 + }, + { + "epoch": 0.46617511520737326, + "grad_norm": 0.4147772789001465, + "learning_rate": 6.441892967038346e-06, + "loss": 0.5023, + "step": 843 + }, + { + "epoch": 0.4667281105990783, + "grad_norm": 0.3958702087402344, + "learning_rate": 6.4326458880444155e-06, + "loss": 0.507, + "step": 844 + }, + { + "epoch": 0.46728110599078343, + "grad_norm": 0.4042072892189026, + "learning_rate": 6.423393467548893e-06, + "loss": 0.444, + "step": 845 + }, + { + "epoch": 0.4678341013824885, + "grad_norm": 0.387441486120224, + "learning_rate": 6.414135740048662e-06, + "loss": 0.4446, + "step": 846 + }, + { + "epoch": 0.46838709677419355, + "grad_norm": 0.4227030277252197, + "learning_rate": 6.404872740060392e-06, + "loss": 0.5104, + "step": 847 + }, + { + "epoch": 0.4689400921658986, + "grad_norm": 0.3976932764053345, + "learning_rate": 6.395604502120412e-06, + "loss": 0.4686, + "step": 848 + }, + { + "epoch": 0.46949308755760366, + "grad_norm": 0.41935819387435913, + "learning_rate": 6.386331060784583e-06, + "loss": 0.4802, + "step": 849 + }, + { + "epoch": 0.4700460829493088, + "grad_norm": 0.3910142183303833, + "learning_rate": 6.377052450628159e-06, + "loss": 0.465, + "step": 850 + }, + { + "epoch": 0.47059907834101383, + "grad_norm": 0.3708650469779968, + "learning_rate": 6.367768706245674e-06, + "loss": 0.4844, + "step": 851 + }, + { + "epoch": 0.4711520737327189, + "grad_norm": 0.39117151498794556, + "learning_rate": 6.358479862250796e-06, + "loss": 0.4826, + "step": 852 + }, + { + "epoch": 0.47170506912442395, + "grad_norm": 0.36844393610954285, + "learning_rate": 6.349185953276216e-06, + "loss": 0.4531, + "step": 853 + }, + { + "epoch": 0.472258064516129, + "grad_norm": 0.3579118549823761, + "learning_rate": 6.339887013973498e-06, + "loss": 0.4579, + "step": 854 + }, + { + "epoch": 0.4728110599078341, + "grad_norm": 0.4207668900489807, + "learning_rate": 6.330583079012972e-06, + "loss": 0.4769, + "step": 855 + }, + { + "epoch": 0.4733640552995392, + "grad_norm": 0.37506356835365295, + "learning_rate": 6.321274183083588e-06, + "loss": 0.4592, + "step": 856 + }, + { + "epoch": 0.47391705069124423, + "grad_norm": 0.4109135568141937, + "learning_rate": 6.3119603608927914e-06, + "loss": 0.4721, + "step": 857 + }, + { + "epoch": 0.4744700460829493, + "grad_norm": 0.4288492202758789, + "learning_rate": 6.302641647166402e-06, + "loss": 0.4766, + "step": 858 + }, + { + "epoch": 0.47502304147465435, + "grad_norm": 0.37833884358406067, + "learning_rate": 6.293318076648469e-06, + "loss": 0.4779, + "step": 859 + }, + { + "epoch": 0.47557603686635946, + "grad_norm": 0.4019952714443207, + "learning_rate": 6.283989684101155e-06, + "loss": 0.487, + "step": 860 + }, + { + "epoch": 0.4761290322580645, + "grad_norm": 0.45245492458343506, + "learning_rate": 6.274656504304598e-06, + "loss": 0.4582, + "step": 861 + }, + { + "epoch": 0.4766820276497696, + "grad_norm": 0.4037781357765198, + "learning_rate": 6.265318572056787e-06, + "loss": 0.4536, + "step": 862 + }, + { + "epoch": 0.47723502304147464, + "grad_norm": 0.4167582392692566, + "learning_rate": 6.25597592217343e-06, + "loss": 0.4737, + "step": 863 + }, + { + "epoch": 0.47778801843317975, + "grad_norm": 0.38921552896499634, + "learning_rate": 6.2466285894878206e-06, + "loss": 0.4889, + "step": 864 + }, + { + "epoch": 0.4783410138248848, + "grad_norm": 0.3599966764450073, + "learning_rate": 6.2372766088507195e-06, + "loss": 0.4421, + "step": 865 + }, + { + "epoch": 0.47889400921658987, + "grad_norm": 0.440920889377594, + "learning_rate": 6.227920015130209e-06, + "loss": 0.4959, + "step": 866 + }, + { + "epoch": 0.4794470046082949, + "grad_norm": 0.39445987343788147, + "learning_rate": 6.218558843211576e-06, + "loss": 0.4784, + "step": 867 + }, + { + "epoch": 0.48, + "grad_norm": 0.4886539578437805, + "learning_rate": 6.209193127997173e-06, + "loss": 0.4702, + "step": 868 + }, + { + "epoch": 0.4805529953917051, + "grad_norm": 0.43818390369415283, + "learning_rate": 6.199822904406295e-06, + "loss": 0.4685, + "step": 869 + }, + { + "epoch": 0.48110599078341015, + "grad_norm": 0.4038926064968109, + "learning_rate": 6.190448207375046e-06, + "loss": 0.4572, + "step": 870 + }, + { + "epoch": 0.4816589861751152, + "grad_norm": 0.5002937912940979, + "learning_rate": 6.1810690718562065e-06, + "loss": 0.4899, + "step": 871 + }, + { + "epoch": 0.48221198156682027, + "grad_norm": 0.39266350865364075, + "learning_rate": 6.1716855328191064e-06, + "loss": 0.4682, + "step": 872 + }, + { + "epoch": 0.4827649769585253, + "grad_norm": 0.4712425470352173, + "learning_rate": 6.162297625249494e-06, + "loss": 0.473, + "step": 873 + }, + { + "epoch": 0.48331797235023044, + "grad_norm": 0.33758747577667236, + "learning_rate": 6.1529053841494065e-06, + "loss": 0.4771, + "step": 874 + }, + { + "epoch": 0.4838709677419355, + "grad_norm": 0.442381888628006, + "learning_rate": 6.143508844537038e-06, + "loss": 0.4834, + "step": 875 + }, + { + "epoch": 0.48442396313364056, + "grad_norm": 0.43109333515167236, + "learning_rate": 6.134108041446609e-06, + "loss": 0.4717, + "step": 876 + }, + { + "epoch": 0.4849769585253456, + "grad_norm": 0.4240846037864685, + "learning_rate": 6.124703009928233e-06, + "loss": 0.491, + "step": 877 + }, + { + "epoch": 0.48552995391705067, + "grad_norm": 0.41452300548553467, + "learning_rate": 6.115293785047793e-06, + "loss": 0.4789, + "step": 878 + }, + { + "epoch": 0.4860829493087558, + "grad_norm": 0.397856742143631, + "learning_rate": 6.105880401886804e-06, + "loss": 0.4655, + "step": 879 + }, + { + "epoch": 0.48663594470046084, + "grad_norm": 0.3984890878200531, + "learning_rate": 6.096462895542288e-06, + "loss": 0.4864, + "step": 880 + }, + { + "epoch": 0.4871889400921659, + "grad_norm": 0.4527179002761841, + "learning_rate": 6.087041301126636e-06, + "loss": 0.4884, + "step": 881 + }, + { + "epoch": 0.48774193548387096, + "grad_norm": 0.4393288791179657, + "learning_rate": 6.077615653767484e-06, + "loss": 0.45, + "step": 882 + }, + { + "epoch": 0.488294930875576, + "grad_norm": 0.4285862445831299, + "learning_rate": 6.068185988607575e-06, + "loss": 0.4858, + "step": 883 + }, + { + "epoch": 0.48884792626728113, + "grad_norm": 0.4243047833442688, + "learning_rate": 6.058752340804639e-06, + "loss": 0.4801, + "step": 884 + }, + { + "epoch": 0.4894009216589862, + "grad_norm": 0.37787267565727234, + "learning_rate": 6.0493147455312475e-06, + "loss": 0.4584, + "step": 885 + }, + { + "epoch": 0.48995391705069125, + "grad_norm": 0.4291735589504242, + "learning_rate": 6.039873237974695e-06, + "loss": 0.5038, + "step": 886 + }, + { + "epoch": 0.4905069124423963, + "grad_norm": 0.3942468464374542, + "learning_rate": 6.03042785333686e-06, + "loss": 0.4805, + "step": 887 + }, + { + "epoch": 0.49105990783410136, + "grad_norm": 0.42691728472709656, + "learning_rate": 6.020978626834077e-06, + "loss": 0.4587, + "step": 888 + }, + { + "epoch": 0.4916129032258065, + "grad_norm": 0.3891282379627228, + "learning_rate": 6.0115255936970056e-06, + "loss": 0.4756, + "step": 889 + }, + { + "epoch": 0.49216589861751153, + "grad_norm": 0.39589089155197144, + "learning_rate": 6.002068789170497e-06, + "loss": 0.4796, + "step": 890 + }, + { + "epoch": 0.4927188940092166, + "grad_norm": 0.4499064087867737, + "learning_rate": 5.992608248513465e-06, + "loss": 0.4667, + "step": 891 + }, + { + "epoch": 0.49327188940092165, + "grad_norm": 0.4281262457370758, + "learning_rate": 5.983144006998751e-06, + "loss": 0.4762, + "step": 892 + }, + { + "epoch": 0.4938248847926267, + "grad_norm": 0.4095862805843353, + "learning_rate": 5.973676099912996e-06, + "loss": 0.4691, + "step": 893 + }, + { + "epoch": 0.4943778801843318, + "grad_norm": 0.4401033818721771, + "learning_rate": 5.964204562556508e-06, + "loss": 0.4688, + "step": 894 + }, + { + "epoch": 0.4949308755760369, + "grad_norm": 0.41105031967163086, + "learning_rate": 5.954729430243129e-06, + "loss": 0.4877, + "step": 895 + }, + { + "epoch": 0.49548387096774194, + "grad_norm": 0.39041486382484436, + "learning_rate": 5.945250738300108e-06, + "loss": 0.4504, + "step": 896 + }, + { + "epoch": 0.496036866359447, + "grad_norm": 0.4423321485519409, + "learning_rate": 5.935768522067962e-06, + "loss": 0.4858, + "step": 897 + }, + { + "epoch": 0.49658986175115205, + "grad_norm": 0.4127327799797058, + "learning_rate": 5.9262828169003476e-06, + "loss": 0.4767, + "step": 898 + }, + { + "epoch": 0.49714285714285716, + "grad_norm": 0.41609784960746765, + "learning_rate": 5.9167936581639325e-06, + "loss": 0.4901, + "step": 899 + }, + { + "epoch": 0.4976958525345622, + "grad_norm": 0.3699811100959778, + "learning_rate": 5.9073010812382595e-06, + "loss": 0.4595, + "step": 900 + }, + { + "epoch": 0.4982488479262673, + "grad_norm": 0.45880934596061707, + "learning_rate": 5.897805121515616e-06, + "loss": 0.4514, + "step": 901 + }, + { + "epoch": 0.49880184331797234, + "grad_norm": 0.38559094071388245, + "learning_rate": 5.888305814400901e-06, + "loss": 0.4871, + "step": 902 + }, + { + "epoch": 0.4993548387096774, + "grad_norm": 0.40912917256355286, + "learning_rate": 5.878803195311496e-06, + "loss": 0.4726, + "step": 903 + }, + { + "epoch": 0.4999078341013825, + "grad_norm": 0.4119787812232971, + "learning_rate": 5.869297299677128e-06, + "loss": 0.459, + "step": 904 + }, + { + "epoch": 0.5004608294930876, + "grad_norm": 0.4300439953804016, + "learning_rate": 5.8597881629397435e-06, + "loss": 0.4764, + "step": 905 + }, + { + "epoch": 0.5010138248847926, + "grad_norm": 0.4216427803039551, + "learning_rate": 5.850275820553368e-06, + "loss": 0.4795, + "step": 906 + }, + { + "epoch": 0.5015668202764977, + "grad_norm": 0.578460693359375, + "learning_rate": 5.840760307983988e-06, + "loss": 0.4463, + "step": 907 + }, + { + "epoch": 0.5021198156682027, + "grad_norm": 0.38813382387161255, + "learning_rate": 5.831241660709402e-06, + "loss": 0.479, + "step": 908 + }, + { + "epoch": 0.5026728110599078, + "grad_norm": 0.4912916123867035, + "learning_rate": 5.8217199142190975e-06, + "loss": 0.4846, + "step": 909 + }, + { + "epoch": 0.5032258064516129, + "grad_norm": 0.4877387285232544, + "learning_rate": 5.812195104014119e-06, + "loss": 0.5051, + "step": 910 + }, + { + "epoch": 0.503778801843318, + "grad_norm": 0.4228856861591339, + "learning_rate": 5.802667265606933e-06, + "loss": 0.4853, + "step": 911 + }, + { + "epoch": 0.5043317972350231, + "grad_norm": 0.4995596706867218, + "learning_rate": 5.793136434521296e-06, + "loss": 0.4603, + "step": 912 + }, + { + "epoch": 0.5048847926267281, + "grad_norm": 0.5033881664276123, + "learning_rate": 5.783602646292123e-06, + "loss": 0.4793, + "step": 913 + }, + { + "epoch": 0.5054377880184332, + "grad_norm": 0.41643261909484863, + "learning_rate": 5.774065936465352e-06, + "loss": 0.4744, + "step": 914 + }, + { + "epoch": 0.5059907834101383, + "grad_norm": 0.4517223834991455, + "learning_rate": 5.764526340597818e-06, + "loss": 0.4659, + "step": 915 + }, + { + "epoch": 0.5065437788018433, + "grad_norm": 0.3723732531070709, + "learning_rate": 5.754983894257114e-06, + "loss": 0.4727, + "step": 916 + }, + { + "epoch": 0.5070967741935484, + "grad_norm": 0.46375709772109985, + "learning_rate": 5.745438633021462e-06, + "loss": 0.4977, + "step": 917 + }, + { + "epoch": 0.5076497695852534, + "grad_norm": 0.42543715238571167, + "learning_rate": 5.7358905924795775e-06, + "loss": 0.4931, + "step": 918 + }, + { + "epoch": 0.5082027649769585, + "grad_norm": 0.3639947772026062, + "learning_rate": 5.726339808230539e-06, + "loss": 0.4608, + "step": 919 + }, + { + "epoch": 0.5087557603686635, + "grad_norm": 0.47503146529197693, + "learning_rate": 5.716786315883657e-06, + "loss": 0.4746, + "step": 920 + }, + { + "epoch": 0.5093087557603687, + "grad_norm": 0.4439897835254669, + "learning_rate": 5.707230151058334e-06, + "loss": 0.47, + "step": 921 + }, + { + "epoch": 0.5098617511520738, + "grad_norm": 0.43655335903167725, + "learning_rate": 5.697671349383939e-06, + "loss": 0.4668, + "step": 922 + }, + { + "epoch": 0.5104147465437788, + "grad_norm": 0.4014508128166199, + "learning_rate": 5.688109946499676e-06, + "loss": 0.4537, + "step": 923 + }, + { + "epoch": 0.5109677419354839, + "grad_norm": 0.4049840271472931, + "learning_rate": 5.678545978054443e-06, + "loss": 0.4713, + "step": 924 + }, + { + "epoch": 0.511520737327189, + "grad_norm": 0.4141314923763275, + "learning_rate": 5.668979479706703e-06, + "loss": 0.4882, + "step": 925 + }, + { + "epoch": 0.512073732718894, + "grad_norm": 0.4375897943973541, + "learning_rate": 5.659410487124355e-06, + "loss": 0.4625, + "step": 926 + }, + { + "epoch": 0.5126267281105991, + "grad_norm": 0.3951698839664459, + "learning_rate": 5.649839035984597e-06, + "loss": 0.4841, + "step": 927 + }, + { + "epoch": 0.5131797235023041, + "grad_norm": 0.367379367351532, + "learning_rate": 5.6402651619737865e-06, + "loss": 0.4818, + "step": 928 + }, + { + "epoch": 0.5137327188940092, + "grad_norm": 0.3982902765274048, + "learning_rate": 5.630688900787326e-06, + "loss": 0.4835, + "step": 929 + }, + { + "epoch": 0.5142857142857142, + "grad_norm": 0.41130173206329346, + "learning_rate": 5.621110288129509e-06, + "loss": 0.4592, + "step": 930 + }, + { + "epoch": 0.5148387096774194, + "grad_norm": 0.4003121554851532, + "learning_rate": 5.6115293597134015e-06, + "loss": 0.4511, + "step": 931 + }, + { + "epoch": 0.5153917050691245, + "grad_norm": 0.3777535855770111, + "learning_rate": 5.601946151260702e-06, + "loss": 0.4943, + "step": 932 + }, + { + "epoch": 0.5159447004608295, + "grad_norm": 0.46429774165153503, + "learning_rate": 5.592360698501609e-06, + "loss": 0.4709, + "step": 933 + }, + { + "epoch": 0.5164976958525346, + "grad_norm": 0.38206857442855835, + "learning_rate": 5.58277303717469e-06, + "loss": 0.4817, + "step": 934 + }, + { + "epoch": 0.5170506912442396, + "grad_norm": 0.39985257387161255, + "learning_rate": 5.573183203026747e-06, + "loss": 0.4646, + "step": 935 + }, + { + "epoch": 0.5176036866359447, + "grad_norm": 0.4025905728340149, + "learning_rate": 5.563591231812683e-06, + "loss": 0.471, + "step": 936 + }, + { + "epoch": 0.5181566820276498, + "grad_norm": 0.41740676760673523, + "learning_rate": 5.553997159295366e-06, + "loss": 0.4798, + "step": 937 + }, + { + "epoch": 0.5187096774193548, + "grad_norm": 0.395632266998291, + "learning_rate": 5.544401021245505e-06, + "loss": 0.4905, + "step": 938 + }, + { + "epoch": 0.5192626728110599, + "grad_norm": 0.3581371605396271, + "learning_rate": 5.534802853441503e-06, + "loss": 0.4563, + "step": 939 + }, + { + "epoch": 0.5198156682027649, + "grad_norm": 0.3925972282886505, + "learning_rate": 5.525202691669335e-06, + "loss": 0.4553, + "step": 940 + }, + { + "epoch": 0.5203686635944701, + "grad_norm": 0.3491620421409607, + "learning_rate": 5.515600571722412e-06, + "loss": 0.4602, + "step": 941 + }, + { + "epoch": 0.5209216589861752, + "grad_norm": 0.37922897934913635, + "learning_rate": 5.505996529401442e-06, + "loss": 0.4588, + "step": 942 + }, + { + "epoch": 0.5214746543778802, + "grad_norm": 0.37720414996147156, + "learning_rate": 5.496390600514298e-06, + "loss": 0.4688, + "step": 943 + }, + { + "epoch": 0.5220276497695853, + "grad_norm": 0.38247910141944885, + "learning_rate": 5.486782820875895e-06, + "loss": 0.4619, + "step": 944 + }, + { + "epoch": 0.5225806451612903, + "grad_norm": 0.3531082570552826, + "learning_rate": 5.477173226308042e-06, + "loss": 0.4611, + "step": 945 + }, + { + "epoch": 0.5231336405529954, + "grad_norm": 0.35197713971138, + "learning_rate": 5.4675618526393185e-06, + "loss": 0.483, + "step": 946 + }, + { + "epoch": 0.5236866359447004, + "grad_norm": 0.42967379093170166, + "learning_rate": 5.457948735704933e-06, + "loss": 0.469, + "step": 947 + }, + { + "epoch": 0.5242396313364055, + "grad_norm": 0.41465815901756287, + "learning_rate": 5.448333911346598e-06, + "loss": 0.4792, + "step": 948 + }, + { + "epoch": 0.5247926267281106, + "grad_norm": 0.385215163230896, + "learning_rate": 5.43871741541239e-06, + "loss": 0.4789, + "step": 949 + }, + { + "epoch": 0.5253456221198156, + "grad_norm": 0.39359477162361145, + "learning_rate": 5.429099283756618e-06, + "loss": 0.4356, + "step": 950 + }, + { + "epoch": 0.5258986175115208, + "grad_norm": 0.37385135889053345, + "learning_rate": 5.419479552239689e-06, + "loss": 0.4713, + "step": 951 + }, + { + "epoch": 0.5264516129032258, + "grad_norm": 0.40417802333831787, + "learning_rate": 5.4098582567279755e-06, + "loss": 0.4714, + "step": 952 + }, + { + "epoch": 0.5270046082949309, + "grad_norm": 0.415487140417099, + "learning_rate": 5.400235433093682e-06, + "loss": 0.4696, + "step": 953 + }, + { + "epoch": 0.527557603686636, + "grad_norm": 0.40797826647758484, + "learning_rate": 5.39061111721471e-06, + "loss": 0.4861, + "step": 954 + }, + { + "epoch": 0.528110599078341, + "grad_norm": 0.4314478039741516, + "learning_rate": 5.380985344974524e-06, + "loss": 0.4885, + "step": 955 + }, + { + "epoch": 0.5286635944700461, + "grad_norm": 0.3722938895225525, + "learning_rate": 5.371358152262019e-06, + "loss": 0.4576, + "step": 956 + }, + { + "epoch": 0.5292165898617511, + "grad_norm": 0.4123469293117523, + "learning_rate": 5.361729574971387e-06, + "loss": 0.4657, + "step": 957 + }, + { + "epoch": 0.5297695852534562, + "grad_norm": 0.4114021956920624, + "learning_rate": 5.352099649001979e-06, + "loss": 0.4705, + "step": 958 + }, + { + "epoch": 0.5303225806451612, + "grad_norm": 0.38689500093460083, + "learning_rate": 5.3424684102581795e-06, + "loss": 0.4661, + "step": 959 + }, + { + "epoch": 0.5308755760368664, + "grad_norm": 0.4139274060726166, + "learning_rate": 5.3328358946492634e-06, + "loss": 0.4721, + "step": 960 + }, + { + "epoch": 0.5314285714285715, + "grad_norm": 0.3889627754688263, + "learning_rate": 5.323202138089266e-06, + "loss": 0.4632, + "step": 961 + }, + { + "epoch": 0.5319815668202765, + "grad_norm": 0.3822690546512604, + "learning_rate": 5.313567176496854e-06, + "loss": 0.4522, + "step": 962 + }, + { + "epoch": 0.5325345622119816, + "grad_norm": 0.4444742798805237, + "learning_rate": 5.303931045795184e-06, + "loss": 0.4627, + "step": 963 + }, + { + "epoch": 0.5330875576036866, + "grad_norm": 0.4218832552433014, + "learning_rate": 5.294293781911769e-06, + "loss": 0.4762, + "step": 964 + }, + { + "epoch": 0.5336405529953917, + "grad_norm": 0.40190473198890686, + "learning_rate": 5.284655420778353e-06, + "loss": 0.4399, + "step": 965 + }, + { + "epoch": 0.5341935483870968, + "grad_norm": 0.44916805624961853, + "learning_rate": 5.275015998330765e-06, + "loss": 0.4457, + "step": 966 + }, + { + "epoch": 0.5347465437788018, + "grad_norm": 0.4550216794013977, + "learning_rate": 5.265375550508793e-06, + "loss": 0.4719, + "step": 967 + }, + { + "epoch": 0.5352995391705069, + "grad_norm": 0.42324307560920715, + "learning_rate": 5.255734113256051e-06, + "loss": 0.46, + "step": 968 + }, + { + "epoch": 0.5358525345622119, + "grad_norm": 0.3758416473865509, + "learning_rate": 5.246091722519839e-06, + "loss": 0.4441, + "step": 969 + }, + { + "epoch": 0.5364055299539171, + "grad_norm": 0.37716051936149597, + "learning_rate": 5.236448414251012e-06, + "loss": 0.4643, + "step": 970 + }, + { + "epoch": 0.5369585253456222, + "grad_norm": 0.3788035809993744, + "learning_rate": 5.226804224403846e-06, + "loss": 0.4592, + "step": 971 + }, + { + "epoch": 0.5375115207373272, + "grad_norm": 0.38398197293281555, + "learning_rate": 5.217159188935907e-06, + "loss": 0.4744, + "step": 972 + }, + { + "epoch": 0.5380645161290323, + "grad_norm": 0.3767092823982239, + "learning_rate": 5.207513343807907e-06, + "loss": 0.4698, + "step": 973 + }, + { + "epoch": 0.5386175115207373, + "grad_norm": 0.39843782782554626, + "learning_rate": 5.1978667249835855e-06, + "loss": 0.4803, + "step": 974 + }, + { + "epoch": 0.5391705069124424, + "grad_norm": 0.38011598587036133, + "learning_rate": 5.18821936842956e-06, + "loss": 0.4398, + "step": 975 + }, + { + "epoch": 0.5397235023041475, + "grad_norm": 0.36936551332473755, + "learning_rate": 5.1785713101152e-06, + "loss": 0.4662, + "step": 976 + }, + { + "epoch": 0.5402764976958525, + "grad_norm": 0.3832260072231293, + "learning_rate": 5.168922586012495e-06, + "loss": 0.488, + "step": 977 + }, + { + "epoch": 0.5408294930875576, + "grad_norm": 0.38859879970550537, + "learning_rate": 5.159273232095911e-06, + "loss": 0.4982, + "step": 978 + }, + { + "epoch": 0.5413824884792626, + "grad_norm": 0.4032990038394928, + "learning_rate": 5.1496232843422665e-06, + "loss": 0.4919, + "step": 979 + }, + { + "epoch": 0.5419354838709678, + "grad_norm": 0.3422912061214447, + "learning_rate": 5.139972778730593e-06, + "loss": 0.4658, + "step": 980 + }, + { + "epoch": 0.5424884792626729, + "grad_norm": 0.38321739435195923, + "learning_rate": 5.130321751242002e-06, + "loss": 0.4738, + "step": 981 + }, + { + "epoch": 0.5430414746543779, + "grad_norm": 0.3506300151348114, + "learning_rate": 5.12067023785955e-06, + "loss": 0.4612, + "step": 982 + }, + { + "epoch": 0.543594470046083, + "grad_norm": 0.3691524565219879, + "learning_rate": 5.111018274568106e-06, + "loss": 0.4461, + "step": 983 + }, + { + "epoch": 0.544147465437788, + "grad_norm": 0.37267613410949707, + "learning_rate": 5.101365897354215e-06, + "loss": 0.4891, + "step": 984 + }, + { + "epoch": 0.5447004608294931, + "grad_norm": 0.3920471668243408, + "learning_rate": 5.091713142205967e-06, + "loss": 0.4645, + "step": 985 + }, + { + "epoch": 0.5452534562211981, + "grad_norm": 0.37700968980789185, + "learning_rate": 5.082060045112862e-06, + "loss": 0.4721, + "step": 986 + }, + { + "epoch": 0.5458064516129032, + "grad_norm": 0.4021558165550232, + "learning_rate": 5.072406642065671e-06, + "loss": 0.4825, + "step": 987 + }, + { + "epoch": 0.5463594470046083, + "grad_norm": 0.40953153371810913, + "learning_rate": 5.062752969056311e-06, + "loss": 0.4478, + "step": 988 + }, + { + "epoch": 0.5469124423963133, + "grad_norm": 0.4201180934906006, + "learning_rate": 5.0530990620777e-06, + "loss": 0.4405, + "step": 989 + }, + { + "epoch": 0.5474654377880185, + "grad_norm": 0.42839857935905457, + "learning_rate": 5.0434449571236314e-06, + "loss": 0.4774, + "step": 990 + }, + { + "epoch": 0.5480184331797235, + "grad_norm": 0.40306076407432556, + "learning_rate": 5.033790690188637e-06, + "loss": 0.475, + "step": 991 + }, + { + "epoch": 0.5485714285714286, + "grad_norm": 0.3931422829627991, + "learning_rate": 5.024136297267851e-06, + "loss": 0.4813, + "step": 992 + }, + { + "epoch": 0.5491244239631337, + "grad_norm": 0.41347619891166687, + "learning_rate": 5.01448181435688e-06, + "loss": 0.4726, + "step": 993 + }, + { + "epoch": 0.5496774193548387, + "grad_norm": 0.41298919916152954, + "learning_rate": 5.004827277451661e-06, + "loss": 0.4581, + "step": 994 + }, + { + "epoch": 0.5502304147465438, + "grad_norm": 0.3852965235710144, + "learning_rate": 4.9951727225483395e-06, + "loss": 0.4638, + "step": 995 + }, + { + "epoch": 0.5507834101382488, + "grad_norm": 0.37419193983078003, + "learning_rate": 4.985518185643122e-06, + "loss": 0.4675, + "step": 996 + }, + { + "epoch": 0.5513364055299539, + "grad_norm": 0.36649253964424133, + "learning_rate": 4.975863702732151e-06, + "loss": 0.4584, + "step": 997 + }, + { + "epoch": 0.551889400921659, + "grad_norm": 0.4342942237854004, + "learning_rate": 4.966209309811364e-06, + "loss": 0.4731, + "step": 998 + }, + { + "epoch": 0.552442396313364, + "grad_norm": 0.37662389874458313, + "learning_rate": 4.95655504287637e-06, + "loss": 0.4696, + "step": 999 + }, + { + "epoch": 0.5529953917050692, + "grad_norm": 0.40196794271469116, + "learning_rate": 4.946900937922302e-06, + "loss": 0.4496, + "step": 1000 + }, + { + "epoch": 0.5535483870967742, + "grad_norm": 0.3595891296863556, + "learning_rate": 4.93724703094369e-06, + "loss": 0.4621, + "step": 1001 + }, + { + "epoch": 0.5541013824884793, + "grad_norm": 0.4029429256916046, + "learning_rate": 4.927593357934329e-06, + "loss": 0.4885, + "step": 1002 + }, + { + "epoch": 0.5546543778801843, + "grad_norm": 0.37034833431243896, + "learning_rate": 4.917939954887139e-06, + "loss": 0.462, + "step": 1003 + }, + { + "epoch": 0.5552073732718894, + "grad_norm": 0.4130437672138214, + "learning_rate": 4.908286857794034e-06, + "loss": 0.4646, + "step": 1004 + }, + { + "epoch": 0.5557603686635945, + "grad_norm": 0.39467036724090576, + "learning_rate": 4.8986341026457875e-06, + "loss": 0.4557, + "step": 1005 + }, + { + "epoch": 0.5563133640552995, + "grad_norm": 0.37382972240448, + "learning_rate": 4.888981725431897e-06, + "loss": 0.4589, + "step": 1006 + }, + { + "epoch": 0.5568663594470046, + "grad_norm": 0.3812313377857208, + "learning_rate": 4.879329762140453e-06, + "loss": 0.476, + "step": 1007 + }, + { + "epoch": 0.5574193548387096, + "grad_norm": 0.3706028461456299, + "learning_rate": 4.8696782487580005e-06, + "loss": 0.4534, + "step": 1008 + }, + { + "epoch": 0.5579723502304147, + "grad_norm": 0.40674856305122375, + "learning_rate": 4.8600272212694094e-06, + "loss": 0.4555, + "step": 1009 + }, + { + "epoch": 0.5585253456221199, + "grad_norm": 0.3949076235294342, + "learning_rate": 4.850376715657736e-06, + "loss": 0.4569, + "step": 1010 + }, + { + "epoch": 0.5590783410138249, + "grad_norm": 0.3829311430454254, + "learning_rate": 4.840726767904089e-06, + "loss": 0.4589, + "step": 1011 + }, + { + "epoch": 0.55963133640553, + "grad_norm": 0.36710280179977417, + "learning_rate": 4.831077413987505e-06, + "loss": 0.4763, + "step": 1012 + }, + { + "epoch": 0.560184331797235, + "grad_norm": 0.37851566076278687, + "learning_rate": 4.8214286898848e-06, + "loss": 0.4723, + "step": 1013 + }, + { + "epoch": 0.5607373271889401, + "grad_norm": 0.3961973488330841, + "learning_rate": 4.81178063157044e-06, + "loss": 0.456, + "step": 1014 + }, + { + "epoch": 0.5612903225806452, + "grad_norm": 0.38343876600265503, + "learning_rate": 4.802133275016415e-06, + "loss": 0.4626, + "step": 1015 + }, + { + "epoch": 0.5618433179723502, + "grad_norm": 0.3635197877883911, + "learning_rate": 4.792486656192094e-06, + "loss": 0.4692, + "step": 1016 + }, + { + "epoch": 0.5623963133640553, + "grad_norm": 0.3797365128993988, + "learning_rate": 4.782840811064095e-06, + "loss": 0.4587, + "step": 1017 + }, + { + "epoch": 0.5629493087557603, + "grad_norm": 0.35782304406166077, + "learning_rate": 4.773195775596155e-06, + "loss": 0.4503, + "step": 1018 + }, + { + "epoch": 0.5635023041474654, + "grad_norm": 0.370126336812973, + "learning_rate": 4.763551585748989e-06, + "loss": 0.4588, + "step": 1019 + }, + { + "epoch": 0.5640552995391706, + "grad_norm": 0.3596176207065582, + "learning_rate": 4.753908277480162e-06, + "loss": 0.4919, + "step": 1020 + }, + { + "epoch": 0.5646082949308756, + "grad_norm": 0.343191921710968, + "learning_rate": 4.74426588674395e-06, + "loss": 0.4667, + "step": 1021 + }, + { + "epoch": 0.5651612903225807, + "grad_norm": 0.3751341998577118, + "learning_rate": 4.734624449491208e-06, + "loss": 0.4529, + "step": 1022 + }, + { + "epoch": 0.5657142857142857, + "grad_norm": 0.3533148169517517, + "learning_rate": 4.724984001669237e-06, + "loss": 0.4491, + "step": 1023 + }, + { + "epoch": 0.5662672811059908, + "grad_norm": 0.38230666518211365, + "learning_rate": 4.715344579221649e-06, + "loss": 0.4961, + "step": 1024 + }, + { + "epoch": 0.5668202764976958, + "grad_norm": 0.40656420588493347, + "learning_rate": 4.705706218088232e-06, + "loss": 0.4773, + "step": 1025 + }, + { + "epoch": 0.5673732718894009, + "grad_norm": 0.3879065215587616, + "learning_rate": 4.696068954204817e-06, + "loss": 0.5097, + "step": 1026 + }, + { + "epoch": 0.567926267281106, + "grad_norm": 0.40097224712371826, + "learning_rate": 4.686432823503147e-06, + "loss": 0.4661, + "step": 1027 + }, + { + "epoch": 0.568479262672811, + "grad_norm": 0.37987810373306274, + "learning_rate": 4.676797861910735e-06, + "loss": 0.4429, + "step": 1028 + }, + { + "epoch": 0.5690322580645162, + "grad_norm": 0.38563069701194763, + "learning_rate": 4.667164105350739e-06, + "loss": 0.4719, + "step": 1029 + }, + { + "epoch": 0.5695852534562212, + "grad_norm": 0.3681628406047821, + "learning_rate": 4.657531589741822e-06, + "loss": 0.4757, + "step": 1030 + }, + { + "epoch": 0.5701382488479263, + "grad_norm": 0.4479837119579315, + "learning_rate": 4.647900350998022e-06, + "loss": 0.471, + "step": 1031 + }, + { + "epoch": 0.5706912442396314, + "grad_norm": 0.39742401242256165, + "learning_rate": 4.638270425028614e-06, + "loss": 0.4762, + "step": 1032 + }, + { + "epoch": 0.5712442396313364, + "grad_norm": 0.38390353322029114, + "learning_rate": 4.628641847737982e-06, + "loss": 0.4475, + "step": 1033 + }, + { + "epoch": 0.5717972350230415, + "grad_norm": 0.4764840602874756, + "learning_rate": 4.6190146550254775e-06, + "loss": 0.4827, + "step": 1034 + }, + { + "epoch": 0.5723502304147465, + "grad_norm": 0.41221946477890015, + "learning_rate": 4.609388882785291e-06, + "loss": 0.495, + "step": 1035 + }, + { + "epoch": 0.5729032258064516, + "grad_norm": 0.35299742221832275, + "learning_rate": 4.599764566906319e-06, + "loss": 0.4614, + "step": 1036 + }, + { + "epoch": 0.5734562211981566, + "grad_norm": 0.45965003967285156, + "learning_rate": 4.590141743272026e-06, + "loss": 0.462, + "step": 1037 + }, + { + "epoch": 0.5740092165898617, + "grad_norm": 0.37869590520858765, + "learning_rate": 4.5805204477603135e-06, + "loss": 0.4442, + "step": 1038 + }, + { + "epoch": 0.5745622119815669, + "grad_norm": 0.37475693225860596, + "learning_rate": 4.570900716243385e-06, + "loss": 0.4468, + "step": 1039 + }, + { + "epoch": 0.5751152073732719, + "grad_norm": 0.38735684752464294, + "learning_rate": 4.561282584587612e-06, + "loss": 0.4444, + "step": 1040 + }, + { + "epoch": 0.575668202764977, + "grad_norm": 0.3577326834201813, + "learning_rate": 4.551666088653404e-06, + "loss": 0.4864, + "step": 1041 + }, + { + "epoch": 0.576221198156682, + "grad_norm": 0.3724598288536072, + "learning_rate": 4.542051264295069e-06, + "loss": 0.4556, + "step": 1042 + }, + { + "epoch": 0.5767741935483871, + "grad_norm": 0.4160480797290802, + "learning_rate": 4.532438147360684e-06, + "loss": 0.4599, + "step": 1043 + }, + { + "epoch": 0.5773271889400922, + "grad_norm": 0.38303178548812866, + "learning_rate": 4.522826773691958e-06, + "loss": 0.4574, + "step": 1044 + }, + { + "epoch": 0.5778801843317972, + "grad_norm": 0.4100678861141205, + "learning_rate": 4.513217179124106e-06, + "loss": 0.4637, + "step": 1045 + }, + { + "epoch": 0.5784331797235023, + "grad_norm": 0.4295979142189026, + "learning_rate": 4.5036093994857025e-06, + "loss": 0.4747, + "step": 1046 + }, + { + "epoch": 0.5789861751152073, + "grad_norm": 0.36032918095588684, + "learning_rate": 4.49400347059856e-06, + "loss": 0.4682, + "step": 1047 + }, + { + "epoch": 0.5795391705069124, + "grad_norm": 0.4092552959918976, + "learning_rate": 4.484399428277589e-06, + "loss": 0.463, + "step": 1048 + }, + { + "epoch": 0.5800921658986176, + "grad_norm": 0.3771011233329773, + "learning_rate": 4.474797308330665e-06, + "loss": 0.4865, + "step": 1049 + }, + { + "epoch": 0.5806451612903226, + "grad_norm": 0.37506765127182007, + "learning_rate": 4.465197146558498e-06, + "loss": 0.4558, + "step": 1050 + }, + { + "epoch": 0.5811981566820277, + "grad_norm": 0.34741005301475525, + "learning_rate": 4.455598978754496e-06, + "loss": 0.445, + "step": 1051 + }, + { + "epoch": 0.5817511520737327, + "grad_norm": 0.41416463255882263, + "learning_rate": 4.4460028407046344e-06, + "loss": 0.4755, + "step": 1052 + }, + { + "epoch": 0.5823041474654378, + "grad_norm": 0.414931982755661, + "learning_rate": 4.436408768187319e-06, + "loss": 0.4879, + "step": 1053 + }, + { + "epoch": 0.5828571428571429, + "grad_norm": 0.37698793411254883, + "learning_rate": 4.426816796973254e-06, + "loss": 0.4896, + "step": 1054 + }, + { + "epoch": 0.5834101382488479, + "grad_norm": 0.4369506537914276, + "learning_rate": 4.417226962825311e-06, + "loss": 0.4723, + "step": 1055 + }, + { + "epoch": 0.583963133640553, + "grad_norm": 0.3790305256843567, + "learning_rate": 4.407639301498392e-06, + "loss": 0.4466, + "step": 1056 + }, + { + "epoch": 0.584516129032258, + "grad_norm": 0.3516313135623932, + "learning_rate": 4.398053848739299e-06, + "loss": 0.4483, + "step": 1057 + }, + { + "epoch": 0.5850691244239631, + "grad_norm": 0.42550036311149597, + "learning_rate": 4.388470640286599e-06, + "loss": 0.4633, + "step": 1058 + }, + { + "epoch": 0.5856221198156683, + "grad_norm": 0.3732847571372986, + "learning_rate": 4.378889711870492e-06, + "loss": 0.4909, + "step": 1059 + }, + { + "epoch": 0.5861751152073733, + "grad_norm": 0.3723104000091553, + "learning_rate": 4.369311099212676e-06, + "loss": 0.4612, + "step": 1060 + }, + { + "epoch": 0.5867281105990784, + "grad_norm": 0.4225739538669586, + "learning_rate": 4.359734838026214e-06, + "loss": 0.4576, + "step": 1061 + }, + { + "epoch": 0.5872811059907834, + "grad_norm": 0.4037158489227295, + "learning_rate": 4.3501609640154056e-06, + "loss": 0.4803, + "step": 1062 + }, + { + "epoch": 0.5878341013824885, + "grad_norm": 0.3930150866508484, + "learning_rate": 4.340589512875646e-06, + "loss": 0.4484, + "step": 1063 + }, + { + "epoch": 0.5883870967741935, + "grad_norm": 0.4097388684749603, + "learning_rate": 4.331020520293298e-06, + "loss": 0.4699, + "step": 1064 + }, + { + "epoch": 0.5889400921658986, + "grad_norm": 0.3482785224914551, + "learning_rate": 4.321454021945559e-06, + "loss": 0.4632, + "step": 1065 + }, + { + "epoch": 0.5894930875576037, + "grad_norm": 0.45652082562446594, + "learning_rate": 4.311890053500326e-06, + "loss": 0.4699, + "step": 1066 + }, + { + "epoch": 0.5900460829493087, + "grad_norm": 0.3783150613307953, + "learning_rate": 4.302328650616062e-06, + "loss": 0.4662, + "step": 1067 + }, + { + "epoch": 0.5905990783410138, + "grad_norm": 0.36994531750679016, + "learning_rate": 4.292769848941669e-06, + "loss": 0.4581, + "step": 1068 + }, + { + "epoch": 0.5911520737327189, + "grad_norm": 0.3842531144618988, + "learning_rate": 4.283213684116346e-06, + "loss": 0.4631, + "step": 1069 + }, + { + "epoch": 0.591705069124424, + "grad_norm": 0.4432550370693207, + "learning_rate": 4.273660191769463e-06, + "loss": 0.4759, + "step": 1070 + }, + { + "epoch": 0.5922580645161291, + "grad_norm": 0.37499210238456726, + "learning_rate": 4.264109407520425e-06, + "loss": 0.4703, + "step": 1071 + }, + { + "epoch": 0.5928110599078341, + "grad_norm": 0.383779376745224, + "learning_rate": 4.2545613669785405e-06, + "loss": 0.4793, + "step": 1072 + }, + { + "epoch": 0.5933640552995392, + "grad_norm": 0.43077680468559265, + "learning_rate": 4.245016105742889e-06, + "loss": 0.447, + "step": 1073 + }, + { + "epoch": 0.5939170506912442, + "grad_norm": 0.3452954888343811, + "learning_rate": 4.2354736594021854e-06, + "loss": 0.457, + "step": 1074 + }, + { + "epoch": 0.5944700460829493, + "grad_norm": 0.3940063714981079, + "learning_rate": 4.2259340635346515e-06, + "loss": 0.4579, + "step": 1075 + }, + { + "epoch": 0.5950230414746543, + "grad_norm": 0.37174075841903687, + "learning_rate": 4.21639735370788e-06, + "loss": 0.4718, + "step": 1076 + }, + { + "epoch": 0.5955760368663594, + "grad_norm": 0.38617780804634094, + "learning_rate": 4.206863565478705e-06, + "loss": 0.4698, + "step": 1077 + }, + { + "epoch": 0.5961290322580645, + "grad_norm": 0.36345598101615906, + "learning_rate": 4.197332734393067e-06, + "loss": 0.4828, + "step": 1078 + }, + { + "epoch": 0.5966820276497696, + "grad_norm": 0.37622806429862976, + "learning_rate": 4.187804895985881e-06, + "loss": 0.4584, + "step": 1079 + }, + { + "epoch": 0.5972350230414747, + "grad_norm": 0.41539904475212097, + "learning_rate": 4.1782800857809025e-06, + "loss": 0.4606, + "step": 1080 + }, + { + "epoch": 0.5977880184331797, + "grad_norm": 0.37827736139297485, + "learning_rate": 4.168758339290599e-06, + "loss": 0.4576, + "step": 1081 + }, + { + "epoch": 0.5983410138248848, + "grad_norm": 0.36408886313438416, + "learning_rate": 4.159239692016012e-06, + "loss": 0.4792, + "step": 1082 + }, + { + "epoch": 0.5988940092165899, + "grad_norm": 0.347615510225296, + "learning_rate": 4.149724179446631e-06, + "loss": 0.4627, + "step": 1083 + }, + { + "epoch": 0.5994470046082949, + "grad_norm": 0.36233338713645935, + "learning_rate": 4.140211837060258e-06, + "loss": 0.4595, + "step": 1084 + }, + { + "epoch": 0.6, + "grad_norm": 0.385028213262558, + "learning_rate": 4.130702700322873e-06, + "loss": 0.4469, + "step": 1085 + }, + { + "epoch": 0.600552995391705, + "grad_norm": 0.3965871036052704, + "learning_rate": 4.121196804688506e-06, + "loss": 0.4761, + "step": 1086 + }, + { + "epoch": 0.6011059907834101, + "grad_norm": 0.36717289686203003, + "learning_rate": 4.111694185599101e-06, + "loss": 0.4401, + "step": 1087 + }, + { + "epoch": 0.6016589861751153, + "grad_norm": 0.3657437264919281, + "learning_rate": 4.102194878484386e-06, + "loss": 0.4613, + "step": 1088 + }, + { + "epoch": 0.6022119815668203, + "grad_norm": 0.4054916799068451, + "learning_rate": 4.092698918761742e-06, + "loss": 0.489, + "step": 1089 + }, + { + "epoch": 0.6027649769585254, + "grad_norm": 0.40190011262893677, + "learning_rate": 4.083206341836069e-06, + "loss": 0.4703, + "step": 1090 + }, + { + "epoch": 0.6033179723502304, + "grad_norm": 0.4301840662956238, + "learning_rate": 4.073717183099654e-06, + "loss": 0.4676, + "step": 1091 + }, + { + "epoch": 0.6038709677419355, + "grad_norm": 0.3311327397823334, + "learning_rate": 4.06423147793204e-06, + "loss": 0.4695, + "step": 1092 + }, + { + "epoch": 0.6044239631336406, + "grad_norm": 0.3920708894729614, + "learning_rate": 4.054749261699893e-06, + "loss": 0.4834, + "step": 1093 + }, + { + "epoch": 0.6049769585253456, + "grad_norm": 0.33984097838401794, + "learning_rate": 4.045270569756872e-06, + "loss": 0.4633, + "step": 1094 + }, + { + "epoch": 0.6055299539170507, + "grad_norm": 0.36689960956573486, + "learning_rate": 4.0357954374434936e-06, + "loss": 0.4778, + "step": 1095 + }, + { + "epoch": 0.6060829493087557, + "grad_norm": 0.3129914402961731, + "learning_rate": 4.026323900087006e-06, + "loss": 0.4389, + "step": 1096 + }, + { + "epoch": 0.6066359447004608, + "grad_norm": 0.3611215651035309, + "learning_rate": 4.016855993001251e-06, + "loss": 0.4687, + "step": 1097 + }, + { + "epoch": 0.607188940092166, + "grad_norm": 0.36361318826675415, + "learning_rate": 4.007391751486536e-06, + "loss": 0.4592, + "step": 1098 + }, + { + "epoch": 0.607741935483871, + "grad_norm": 0.3965631425380707, + "learning_rate": 3.997931210829503e-06, + "loss": 0.4403, + "step": 1099 + }, + { + "epoch": 0.6082949308755761, + "grad_norm": 0.371737539768219, + "learning_rate": 3.988474406302995e-06, + "loss": 0.4433, + "step": 1100 + }, + { + "epoch": 0.6088479262672811, + "grad_norm": 0.4065137207508087, + "learning_rate": 3.979021373165924e-06, + "loss": 0.4548, + "step": 1101 + }, + { + "epoch": 0.6094009216589862, + "grad_norm": 0.36007609963417053, + "learning_rate": 3.9695721466631426e-06, + "loss": 0.4664, + "step": 1102 + }, + { + "epoch": 0.6099539170506912, + "grad_norm": 0.39393535256385803, + "learning_rate": 3.9601267620253075e-06, + "loss": 0.4899, + "step": 1103 + }, + { + "epoch": 0.6105069124423963, + "grad_norm": 0.40391650795936584, + "learning_rate": 3.950685254468755e-06, + "loss": 0.4657, + "step": 1104 + }, + { + "epoch": 0.6110599078341014, + "grad_norm": 0.3698117434978485, + "learning_rate": 3.941247659195364e-06, + "loss": 0.4524, + "step": 1105 + }, + { + "epoch": 0.6116129032258064, + "grad_norm": 0.35965868830680847, + "learning_rate": 3.931814011392427e-06, + "loss": 0.4629, + "step": 1106 + }, + { + "epoch": 0.6121658986175115, + "grad_norm": 0.4399069547653198, + "learning_rate": 3.922384346232519e-06, + "loss": 0.491, + "step": 1107 + }, + { + "epoch": 0.6127188940092166, + "grad_norm": 0.3660935163497925, + "learning_rate": 3.912958698873366e-06, + "loss": 0.474, + "step": 1108 + }, + { + "epoch": 0.6132718894009217, + "grad_norm": 0.3902994990348816, + "learning_rate": 3.903537104457713e-06, + "loss": 0.4979, + "step": 1109 + }, + { + "epoch": 0.6138248847926268, + "grad_norm": 0.4067245125770569, + "learning_rate": 3.894119598113196e-06, + "loss": 0.4724, + "step": 1110 + }, + { + "epoch": 0.6143778801843318, + "grad_norm": 0.3835393190383911, + "learning_rate": 3.8847062149522075e-06, + "loss": 0.4558, + "step": 1111 + }, + { + "epoch": 0.6149308755760369, + "grad_norm": 0.3831106126308441, + "learning_rate": 3.875296990071768e-06, + "loss": 0.4619, + "step": 1112 + }, + { + "epoch": 0.6154838709677419, + "grad_norm": 0.3794800043106079, + "learning_rate": 3.8658919585533916e-06, + "loss": 0.4535, + "step": 1113 + }, + { + "epoch": 0.616036866359447, + "grad_norm": 0.3396313786506653, + "learning_rate": 3.8564911554629616e-06, + "loss": 0.4892, + "step": 1114 + }, + { + "epoch": 0.616589861751152, + "grad_norm": 0.37773552536964417, + "learning_rate": 3.847094615850593e-06, + "loss": 0.46, + "step": 1115 + }, + { + "epoch": 0.6171428571428571, + "grad_norm": 0.38637781143188477, + "learning_rate": 3.837702374750507e-06, + "loss": 0.4701, + "step": 1116 + }, + { + "epoch": 0.6176958525345622, + "grad_norm": 0.3425704538822174, + "learning_rate": 3.828314467180895e-06, + "loss": 0.4389, + "step": 1117 + }, + { + "epoch": 0.6182488479262673, + "grad_norm": 0.35482358932495117, + "learning_rate": 3.818930928143796e-06, + "loss": 0.477, + "step": 1118 + }, + { + "epoch": 0.6188018433179724, + "grad_norm": 0.3860636353492737, + "learning_rate": 3.8095517926249557e-06, + "loss": 0.4748, + "step": 1119 + }, + { + "epoch": 0.6193548387096774, + "grad_norm": 0.412280797958374, + "learning_rate": 3.800177095593706e-06, + "loss": 0.4624, + "step": 1120 + }, + { + "epoch": 0.6199078341013825, + "grad_norm": 0.3511303961277008, + "learning_rate": 3.790806872002828e-06, + "loss": 0.4867, + "step": 1121 + }, + { + "epoch": 0.6204608294930876, + "grad_norm": 0.3653261065483093, + "learning_rate": 3.7814411567884256e-06, + "loss": 0.4737, + "step": 1122 + }, + { + "epoch": 0.6210138248847926, + "grad_norm": 0.3349493145942688, + "learning_rate": 3.7720799848697924e-06, + "loss": 0.4705, + "step": 1123 + }, + { + "epoch": 0.6215668202764977, + "grad_norm": 0.3643210828304291, + "learning_rate": 3.762723391149282e-06, + "loss": 0.4934, + "step": 1124 + }, + { + "epoch": 0.6221198156682027, + "grad_norm": 0.3928675055503845, + "learning_rate": 3.75337141051218e-06, + "loss": 0.4726, + "step": 1125 + }, + { + "epoch": 0.6226728110599078, + "grad_norm": 0.36943259835243225, + "learning_rate": 3.7440240778265725e-06, + "loss": 0.4557, + "step": 1126 + }, + { + "epoch": 0.6232258064516129, + "grad_norm": 0.37895265221595764, + "learning_rate": 3.734681427943214e-06, + "loss": 0.4989, + "step": 1127 + }, + { + "epoch": 0.623778801843318, + "grad_norm": 0.34087231755256653, + "learning_rate": 3.725343495695404e-06, + "loss": 0.4418, + "step": 1128 + }, + { + "epoch": 0.6243317972350231, + "grad_norm": 0.3866007626056671, + "learning_rate": 3.716010315898847e-06, + "loss": 0.4515, + "step": 1129 + }, + { + "epoch": 0.6248847926267281, + "grad_norm": 0.4217049777507782, + "learning_rate": 3.706681923351533e-06, + "loss": 0.4962, + "step": 1130 + }, + { + "epoch": 0.6254377880184332, + "grad_norm": 0.4117668867111206, + "learning_rate": 3.6973583528336e-06, + "loss": 0.4704, + "step": 1131 + }, + { + "epoch": 0.6259907834101383, + "grad_norm": 0.3604739308357239, + "learning_rate": 3.68803963910721e-06, + "loss": 0.4839, + "step": 1132 + }, + { + "epoch": 0.6265437788018433, + "grad_norm": 0.4002184569835663, + "learning_rate": 3.678725816916415e-06, + "loss": 0.4853, + "step": 1133 + }, + { + "epoch": 0.6270967741935484, + "grad_norm": 0.3674345910549164, + "learning_rate": 3.6694169209870305e-06, + "loss": 0.4471, + "step": 1134 + }, + { + "epoch": 0.6276497695852534, + "grad_norm": 0.37487921118736267, + "learning_rate": 3.660112986026504e-06, + "loss": 0.4646, + "step": 1135 + }, + { + "epoch": 0.6282027649769585, + "grad_norm": 0.36266854405403137, + "learning_rate": 3.6508140467237873e-06, + "loss": 0.4802, + "step": 1136 + }, + { + "epoch": 0.6287557603686635, + "grad_norm": 0.3480757176876068, + "learning_rate": 3.6415201377492048e-06, + "loss": 0.458, + "step": 1137 + }, + { + "epoch": 0.6293087557603687, + "grad_norm": 0.3710945248603821, + "learning_rate": 3.632231293754328e-06, + "loss": 0.4648, + "step": 1138 + }, + { + "epoch": 0.6298617511520738, + "grad_norm": 0.42196762561798096, + "learning_rate": 3.622947549371841e-06, + "loss": 0.4802, + "step": 1139 + }, + { + "epoch": 0.6304147465437788, + "grad_norm": 0.3192515969276428, + "learning_rate": 3.6136689392154186e-06, + "loss": 0.4559, + "step": 1140 + }, + { + "epoch": 0.6309677419354839, + "grad_norm": 0.37404048442840576, + "learning_rate": 3.6043954978795868e-06, + "loss": 0.4647, + "step": 1141 + }, + { + "epoch": 0.6315207373271889, + "grad_norm": 0.37971508502960205, + "learning_rate": 3.5951272599396086e-06, + "loss": 0.4347, + "step": 1142 + }, + { + "epoch": 0.632073732718894, + "grad_norm": 0.410057932138443, + "learning_rate": 3.5858642599513394e-06, + "loss": 0.4759, + "step": 1143 + }, + { + "epoch": 0.6326267281105991, + "grad_norm": 0.367079496383667, + "learning_rate": 3.576606532451108e-06, + "loss": 0.4864, + "step": 1144 + }, + { + "epoch": 0.6331797235023041, + "grad_norm": 0.3866696357727051, + "learning_rate": 3.567354111955585e-06, + "loss": 0.4746, + "step": 1145 + }, + { + "epoch": 0.6337327188940092, + "grad_norm": 0.3390635848045349, + "learning_rate": 3.5581070329616543e-06, + "loss": 0.4593, + "step": 1146 + }, + { + "epoch": 0.6342857142857142, + "grad_norm": 0.41744962334632874, + "learning_rate": 3.5488653299462844e-06, + "loss": 0.4514, + "step": 1147 + }, + { + "epoch": 0.6348387096774194, + "grad_norm": 0.3437979817390442, + "learning_rate": 3.5396290373664e-06, + "loss": 0.4689, + "step": 1148 + }, + { + "epoch": 0.6353917050691245, + "grad_norm": 0.32637640833854675, + "learning_rate": 3.5303981896587534e-06, + "loss": 0.4657, + "step": 1149 + }, + { + "epoch": 0.6359447004608295, + "grad_norm": 0.3600415885448456, + "learning_rate": 3.521172821239796e-06, + "loss": 0.4637, + "step": 1150 + }, + { + "epoch": 0.6364976958525346, + "grad_norm": 0.4324146807193756, + "learning_rate": 3.5119529665055506e-06, + "loss": 0.4715, + "step": 1151 + }, + { + "epoch": 0.6370506912442396, + "grad_norm": 0.3583220839500427, + "learning_rate": 3.502738659831483e-06, + "loss": 0.4666, + "step": 1152 + }, + { + "epoch": 0.6376036866359447, + "grad_norm": 0.35306817293167114, + "learning_rate": 3.49352993557237e-06, + "loss": 0.4554, + "step": 1153 + }, + { + "epoch": 0.6381566820276497, + "grad_norm": 0.39276352524757385, + "learning_rate": 3.48432682806218e-06, + "loss": 0.4492, + "step": 1154 + }, + { + "epoch": 0.6387096774193548, + "grad_norm": 0.3977315425872803, + "learning_rate": 3.4751293716139366e-06, + "loss": 0.4523, + "step": 1155 + }, + { + "epoch": 0.6392626728110599, + "grad_norm": 0.36172613501548767, + "learning_rate": 3.4659376005195956e-06, + "loss": 0.5053, + "step": 1156 + }, + { + "epoch": 0.639815668202765, + "grad_norm": 0.34491607546806335, + "learning_rate": 3.4567515490499135e-06, + "loss": 0.4633, + "step": 1157 + }, + { + "epoch": 0.6403686635944701, + "grad_norm": 0.3682193160057068, + "learning_rate": 3.4475712514543246e-06, + "loss": 0.4376, + "step": 1158 + }, + { + "epoch": 0.6409216589861751, + "grad_norm": 0.41406625509262085, + "learning_rate": 3.438396741960806e-06, + "loss": 0.4669, + "step": 1159 + }, + { + "epoch": 0.6414746543778802, + "grad_norm": 0.38740435242652893, + "learning_rate": 3.4292280547757586e-06, + "loss": 0.4516, + "step": 1160 + }, + { + "epoch": 0.6420276497695853, + "grad_norm": 0.35590702295303345, + "learning_rate": 3.420065224083873e-06, + "loss": 0.4406, + "step": 1161 + }, + { + "epoch": 0.6425806451612903, + "grad_norm": 0.3844543397426605, + "learning_rate": 3.4109082840480045e-06, + "loss": 0.4675, + "step": 1162 + }, + { + "epoch": 0.6431336405529954, + "grad_norm": 0.40717223286628723, + "learning_rate": 3.4017572688090467e-06, + "loss": 0.4862, + "step": 1163 + }, + { + "epoch": 0.6436866359447004, + "grad_norm": 0.37645992636680603, + "learning_rate": 3.392612212485802e-06, + "loss": 0.4813, + "step": 1164 + }, + { + "epoch": 0.6442396313364055, + "grad_norm": 0.33362942934036255, + "learning_rate": 3.3834731491748558e-06, + "loss": 0.4576, + "step": 1165 + }, + { + "epoch": 0.6447926267281106, + "grad_norm": 0.3458590805530548, + "learning_rate": 3.3743401129504496e-06, + "loss": 0.4664, + "step": 1166 + }, + { + "epoch": 0.6453456221198157, + "grad_norm": 0.39395076036453247, + "learning_rate": 3.3652131378643515e-06, + "loss": 0.4801, + "step": 1167 + }, + { + "epoch": 0.6458986175115208, + "grad_norm": 0.36739885807037354, + "learning_rate": 3.356092257945733e-06, + "loss": 0.4662, + "step": 1168 + }, + { + "epoch": 0.6464516129032258, + "grad_norm": 0.35936224460601807, + "learning_rate": 3.3469775072010417e-06, + "loss": 0.4678, + "step": 1169 + }, + { + "epoch": 0.6470046082949309, + "grad_norm": 0.35847219824790955, + "learning_rate": 3.337868919613869e-06, + "loss": 0.4595, + "step": 1170 + }, + { + "epoch": 0.647557603686636, + "grad_norm": 0.33475178480148315, + "learning_rate": 3.3287665291448323e-06, + "loss": 0.4631, + "step": 1171 + }, + { + "epoch": 0.648110599078341, + "grad_norm": 0.3567461669445038, + "learning_rate": 3.319670369731437e-06, + "loss": 0.475, + "step": 1172 + }, + { + "epoch": 0.6486635944700461, + "grad_norm": 0.4065379798412323, + "learning_rate": 3.310580475287963e-06, + "loss": 0.4596, + "step": 1173 + }, + { + "epoch": 0.6492165898617511, + "grad_norm": 0.34394586086273193, + "learning_rate": 3.3014968797053263e-06, + "loss": 0.4717, + "step": 1174 + }, + { + "epoch": 0.6497695852534562, + "grad_norm": 0.3613266944885254, + "learning_rate": 3.2924196168509657e-06, + "loss": 0.4933, + "step": 1175 + }, + { + "epoch": 0.6503225806451612, + "grad_norm": 0.38999873399734497, + "learning_rate": 3.283348720568702e-06, + "loss": 0.4756, + "step": 1176 + }, + { + "epoch": 0.6508755760368664, + "grad_norm": 0.3532698154449463, + "learning_rate": 3.274284224678621e-06, + "loss": 0.4575, + "step": 1177 + }, + { + "epoch": 0.6514285714285715, + "grad_norm": 0.33012717962265015, + "learning_rate": 3.2652261629769457e-06, + "loss": 0.4473, + "step": 1178 + }, + { + "epoch": 0.6519815668202765, + "grad_norm": 0.3971370458602905, + "learning_rate": 3.256174569235909e-06, + "loss": 0.5044, + "step": 1179 + }, + { + "epoch": 0.6525345622119816, + "grad_norm": 0.3590445816516876, + "learning_rate": 3.2471294772036287e-06, + "loss": 0.4675, + "step": 1180 + }, + { + "epoch": 0.6530875576036866, + "grad_norm": 0.3668496012687683, + "learning_rate": 3.238090920603981e-06, + "loss": 0.4419, + "step": 1181 + }, + { + "epoch": 0.6536405529953917, + "grad_norm": 0.3250981867313385, + "learning_rate": 3.2290589331364787e-06, + "loss": 0.4597, + "step": 1182 + }, + { + "epoch": 0.6541935483870968, + "grad_norm": 0.38982975482940674, + "learning_rate": 3.2200335484761352e-06, + "loss": 0.4638, + "step": 1183 + }, + { + "epoch": 0.6547465437788018, + "grad_norm": 0.38048577308654785, + "learning_rate": 3.2110148002733534e-06, + "loss": 0.473, + "step": 1184 + }, + { + "epoch": 0.6552995391705069, + "grad_norm": 0.35395413637161255, + "learning_rate": 3.2020027221537864e-06, + "loss": 0.4832, + "step": 1185 + }, + { + "epoch": 0.6558525345622119, + "grad_norm": 0.3707823157310486, + "learning_rate": 3.192997347718224e-06, + "loss": 0.4958, + "step": 1186 + }, + { + "epoch": 0.6564055299539171, + "grad_norm": 0.36011847853660583, + "learning_rate": 3.1839987105424586e-06, + "loss": 0.472, + "step": 1187 + }, + { + "epoch": 0.6569585253456222, + "grad_norm": 0.3693712055683136, + "learning_rate": 3.1750068441771637e-06, + "loss": 0.4649, + "step": 1188 + }, + { + "epoch": 0.6575115207373272, + "grad_norm": 0.34001612663269043, + "learning_rate": 3.1660217821477686e-06, + "loss": 0.4764, + "step": 1189 + }, + { + "epoch": 0.6580645161290323, + "grad_norm": 0.332956999540329, + "learning_rate": 3.1570435579543333e-06, + "loss": 0.4652, + "step": 1190 + }, + { + "epoch": 0.6586175115207373, + "grad_norm": 0.35701560974121094, + "learning_rate": 3.148072205071423e-06, + "loss": 0.4695, + "step": 1191 + }, + { + "epoch": 0.6591705069124424, + "grad_norm": 0.3385053277015686, + "learning_rate": 3.1391077569479856e-06, + "loss": 0.4524, + "step": 1192 + }, + { + "epoch": 0.6597235023041474, + "grad_norm": 0.35991233587265015, + "learning_rate": 3.130150247007222e-06, + "loss": 0.4322, + "step": 1193 + }, + { + "epoch": 0.6602764976958525, + "grad_norm": 0.36623886227607727, + "learning_rate": 3.1211997086464683e-06, + "loss": 0.4754, + "step": 1194 + }, + { + "epoch": 0.6608294930875576, + "grad_norm": 0.33661210536956787, + "learning_rate": 3.112256175237064e-06, + "loss": 0.464, + "step": 1195 + }, + { + "epoch": 0.6613824884792626, + "grad_norm": 0.3260611295700073, + "learning_rate": 3.103319680124235e-06, + "loss": 0.4608, + "step": 1196 + }, + { + "epoch": 0.6619354838709678, + "grad_norm": 0.38113391399383545, + "learning_rate": 3.0943902566269613e-06, + "loss": 0.4963, + "step": 1197 + }, + { + "epoch": 0.6624884792626728, + "grad_norm": 0.3196834623813629, + "learning_rate": 3.085467938037861e-06, + "loss": 0.4631, + "step": 1198 + }, + { + "epoch": 0.6630414746543779, + "grad_norm": 0.3713325262069702, + "learning_rate": 3.076552757623059e-06, + "loss": 0.4559, + "step": 1199 + }, + { + "epoch": 0.663594470046083, + "grad_norm": 0.35076281428337097, + "learning_rate": 3.0676447486220705e-06, + "loss": 0.4601, + "step": 1200 + }, + { + "epoch": 0.664147465437788, + "grad_norm": 0.34032174944877625, + "learning_rate": 3.058743944247665e-06, + "loss": 0.4702, + "step": 1201 + }, + { + "epoch": 0.6647004608294931, + "grad_norm": 0.37965843081474304, + "learning_rate": 3.0498503776857576e-06, + "loss": 0.4806, + "step": 1202 + }, + { + "epoch": 0.6652534562211981, + "grad_norm": 0.3258879780769348, + "learning_rate": 3.0409640820952735e-06, + "loss": 0.4331, + "step": 1203 + }, + { + "epoch": 0.6658064516129032, + "grad_norm": 0.3517114222049713, + "learning_rate": 3.0320850906080325e-06, + "loss": 0.4521, + "step": 1204 + }, + { + "epoch": 0.6663594470046083, + "grad_norm": 0.3746355473995209, + "learning_rate": 3.0232134363286163e-06, + "loss": 0.4702, + "step": 1205 + }, + { + "epoch": 0.6669124423963133, + "grad_norm": 0.3362243175506592, + "learning_rate": 3.0143491523342562e-06, + "loss": 0.4382, + "step": 1206 + }, + { + "epoch": 0.6674654377880185, + "grad_norm": 0.3588075041770935, + "learning_rate": 3.005492271674697e-06, + "loss": 0.452, + "step": 1207 + }, + { + "epoch": 0.6680184331797235, + "grad_norm": 0.4125935137271881, + "learning_rate": 2.9966428273720904e-06, + "loss": 0.4715, + "step": 1208 + }, + { + "epoch": 0.6685714285714286, + "grad_norm": 0.34157875180244446, + "learning_rate": 2.987800852420855e-06, + "loss": 0.4543, + "step": 1209 + }, + { + "epoch": 0.6691244239631337, + "grad_norm": 0.34906935691833496, + "learning_rate": 2.9789663797875614e-06, + "loss": 0.4535, + "step": 1210 + }, + { + "epoch": 0.6696774193548387, + "grad_norm": 0.3961377739906311, + "learning_rate": 2.970139442410811e-06, + "loss": 0.4992, + "step": 1211 + }, + { + "epoch": 0.6702304147465438, + "grad_norm": 0.3486752510070801, + "learning_rate": 2.961320073201107e-06, + "loss": 0.4603, + "step": 1212 + }, + { + "epoch": 0.6707834101382488, + "grad_norm": 0.3723919093608856, + "learning_rate": 2.952508305040739e-06, + "loss": 0.4735, + "step": 1213 + }, + { + "epoch": 0.6713364055299539, + "grad_norm": 0.33436325192451477, + "learning_rate": 2.9437041707836532e-06, + "loss": 0.4558, + "step": 1214 + }, + { + "epoch": 0.6718894009216589, + "grad_norm": 0.3649885058403015, + "learning_rate": 2.9349077032553387e-06, + "loss": 0.4734, + "step": 1215 + }, + { + "epoch": 0.6724423963133641, + "grad_norm": 0.4229625463485718, + "learning_rate": 2.9261189352526932e-06, + "loss": 0.4734, + "step": 1216 + }, + { + "epoch": 0.6729953917050692, + "grad_norm": 0.31533482670783997, + "learning_rate": 2.9173378995439107e-06, + "loss": 0.4596, + "step": 1217 + }, + { + "epoch": 0.6735483870967742, + "grad_norm": 0.39130347967147827, + "learning_rate": 2.9085646288683587e-06, + "loss": 0.4913, + "step": 1218 + }, + { + "epoch": 0.6741013824884793, + "grad_norm": 0.3557525873184204, + "learning_rate": 2.8997991559364493e-06, + "loss": 0.4708, + "step": 1219 + }, + { + "epoch": 0.6746543778801843, + "grad_norm": 0.36776086688041687, + "learning_rate": 2.8910415134295216e-06, + "loss": 0.4685, + "step": 1220 + }, + { + "epoch": 0.6752073732718894, + "grad_norm": 0.3936042785644531, + "learning_rate": 2.8822917339997237e-06, + "loss": 0.4667, + "step": 1221 + }, + { + "epoch": 0.6757603686635945, + "grad_norm": 0.3256804645061493, + "learning_rate": 2.87354985026988e-06, + "loss": 0.428, + "step": 1222 + }, + { + "epoch": 0.6763133640552995, + "grad_norm": 0.3750945031642914, + "learning_rate": 2.8648158948333817e-06, + "loss": 0.458, + "step": 1223 + }, + { + "epoch": 0.6768663594470046, + "grad_norm": 0.3525260090827942, + "learning_rate": 2.856089900254059e-06, + "loss": 0.4469, + "step": 1224 + }, + { + "epoch": 0.6774193548387096, + "grad_norm": 0.32836800813674927, + "learning_rate": 2.847371899066059e-06, + "loss": 0.4581, + "step": 1225 + }, + { + "epoch": 0.6779723502304148, + "grad_norm": 0.34020939469337463, + "learning_rate": 2.83866192377373e-06, + "loss": 0.444, + "step": 1226 + }, + { + "epoch": 0.6785253456221199, + "grad_norm": 0.31056728959083557, + "learning_rate": 2.829960006851492e-06, + "loss": 0.4748, + "step": 1227 + }, + { + "epoch": 0.6790783410138249, + "grad_norm": 0.3614257574081421, + "learning_rate": 2.8212661807437226e-06, + "loss": 0.487, + "step": 1228 + }, + { + "epoch": 0.67963133640553, + "grad_norm": 0.33884182572364807, + "learning_rate": 2.8125804778646315e-06, + "loss": 0.4871, + "step": 1229 + }, + { + "epoch": 0.680184331797235, + "grad_norm": 0.3698228597640991, + "learning_rate": 2.803902930598144e-06, + "loss": 0.4783, + "step": 1230 + }, + { + "epoch": 0.6807373271889401, + "grad_norm": 0.3055351674556732, + "learning_rate": 2.7952335712977764e-06, + "loss": 0.4488, + "step": 1231 + }, + { + "epoch": 0.6812903225806451, + "grad_norm": 0.3399738073348999, + "learning_rate": 2.7865724322865174e-06, + "loss": 0.4489, + "step": 1232 + }, + { + "epoch": 0.6818433179723502, + "grad_norm": 0.3493901193141937, + "learning_rate": 2.7779195458567067e-06, + "loss": 0.4508, + "step": 1233 + }, + { + "epoch": 0.6823963133640553, + "grad_norm": 0.38333308696746826, + "learning_rate": 2.7692749442699147e-06, + "loss": 0.4733, + "step": 1234 + }, + { + "epoch": 0.6829493087557603, + "grad_norm": 0.33638396859169006, + "learning_rate": 2.7606386597568223e-06, + "loss": 0.4521, + "step": 1235 + }, + { + "epoch": 0.6835023041474655, + "grad_norm": 0.33842137455940247, + "learning_rate": 2.7520107245171035e-06, + "loss": 0.4359, + "step": 1236 + }, + { + "epoch": 0.6840552995391705, + "grad_norm": 0.34261924028396606, + "learning_rate": 2.7433911707192976e-06, + "loss": 0.4713, + "step": 1237 + }, + { + "epoch": 0.6846082949308756, + "grad_norm": 0.33021679520606995, + "learning_rate": 2.7347800305007e-06, + "loss": 0.4563, + "step": 1238 + }, + { + "epoch": 0.6851612903225807, + "grad_norm": 0.33682435750961304, + "learning_rate": 2.7261773359672306e-06, + "loss": 0.4429, + "step": 1239 + }, + { + "epoch": 0.6857142857142857, + "grad_norm": 0.3547312319278717, + "learning_rate": 2.7175831191933275e-06, + "loss": 0.4846, + "step": 1240 + }, + { + "epoch": 0.6862672811059908, + "grad_norm": 0.3378848433494568, + "learning_rate": 2.7089974122218165e-06, + "loss": 0.4681, + "step": 1241 + }, + { + "epoch": 0.6868202764976958, + "grad_norm": 0.3516218066215515, + "learning_rate": 2.700420247063793e-06, + "loss": 0.4796, + "step": 1242 + }, + { + "epoch": 0.6873732718894009, + "grad_norm": 0.3513849973678589, + "learning_rate": 2.6918516556985088e-06, + "loss": 0.4732, + "step": 1243 + }, + { + "epoch": 0.687926267281106, + "grad_norm": 0.37055304646492004, + "learning_rate": 2.6832916700732466e-06, + "loss": 0.492, + "step": 1244 + }, + { + "epoch": 0.688479262672811, + "grad_norm": 0.3224850594997406, + "learning_rate": 2.674740322103204e-06, + "loss": 0.4618, + "step": 1245 + }, + { + "epoch": 0.6890322580645162, + "grad_norm": 0.3502301573753357, + "learning_rate": 2.666197643671375e-06, + "loss": 0.4562, + "step": 1246 + }, + { + "epoch": 0.6895852534562212, + "grad_norm": 0.37586522102355957, + "learning_rate": 2.657663666628428e-06, + "loss": 0.448, + "step": 1247 + }, + { + "epoch": 0.6901382488479263, + "grad_norm": 0.38795405626296997, + "learning_rate": 2.6491384227925903e-06, + "loss": 0.472, + "step": 1248 + }, + { + "epoch": 0.6906912442396314, + "grad_norm": 0.3674571216106415, + "learning_rate": 2.640621943949527e-06, + "loss": 0.4707, + "step": 1249 + }, + { + "epoch": 0.6912442396313364, + "grad_norm": 0.36492788791656494, + "learning_rate": 2.6321142618522288e-06, + "loss": 0.4644, + "step": 1250 + }, + { + "epoch": 0.6917972350230415, + "grad_norm": 0.345634400844574, + "learning_rate": 2.6236154082208776e-06, + "loss": 0.4556, + "step": 1251 + }, + { + "epoch": 0.6923502304147465, + "grad_norm": 0.3678940534591675, + "learning_rate": 2.6151254147427486e-06, + "loss": 0.4983, + "step": 1252 + }, + { + "epoch": 0.6929032258064516, + "grad_norm": 0.34437957406044006, + "learning_rate": 2.6066443130720786e-06, + "loss": 0.4613, + "step": 1253 + }, + { + "epoch": 0.6934562211981566, + "grad_norm": 0.3727521300315857, + "learning_rate": 2.598172134829955e-06, + "loss": 0.4771, + "step": 1254 + }, + { + "epoch": 0.6940092165898617, + "grad_norm": 0.3770442008972168, + "learning_rate": 2.5897089116041918e-06, + "loss": 0.4627, + "step": 1255 + }, + { + "epoch": 0.6945622119815669, + "grad_norm": 0.3384653329849243, + "learning_rate": 2.5812546749492163e-06, + "loss": 0.4494, + "step": 1256 + }, + { + "epoch": 0.6951152073732719, + "grad_norm": 0.3460019528865814, + "learning_rate": 2.5728094563859495e-06, + "loss": 0.4726, + "step": 1257 + }, + { + "epoch": 0.695668202764977, + "grad_norm": 0.328002005815506, + "learning_rate": 2.5643732874016903e-06, + "loss": 0.4437, + "step": 1258 + }, + { + "epoch": 0.696221198156682, + "grad_norm": 0.40374118089675903, + "learning_rate": 2.5559461994499968e-06, + "loss": 0.4766, + "step": 1259 + }, + { + "epoch": 0.6967741935483871, + "grad_norm": 0.3571721017360687, + "learning_rate": 2.5475282239505685e-06, + "loss": 0.459, + "step": 1260 + }, + { + "epoch": 0.6973271889400922, + "grad_norm": 0.42431801557540894, + "learning_rate": 2.5391193922891288e-06, + "loss": 0.4641, + "step": 1261 + }, + { + "epoch": 0.6978801843317972, + "grad_norm": 0.35599109530448914, + "learning_rate": 2.5307197358173126e-06, + "loss": 0.4601, + "step": 1262 + }, + { + "epoch": 0.6984331797235023, + "grad_norm": 0.3282775282859802, + "learning_rate": 2.5223292858525423e-06, + "loss": 0.4352, + "step": 1263 + }, + { + "epoch": 0.6989861751152073, + "grad_norm": 0.3075070381164551, + "learning_rate": 2.5139480736779164e-06, + "loss": 0.4373, + "step": 1264 + }, + { + "epoch": 0.6995391705069124, + "grad_norm": 0.3571772277355194, + "learning_rate": 2.5055761305420907e-06, + "loss": 0.4938, + "step": 1265 + }, + { + "epoch": 0.7000921658986176, + "grad_norm": 0.3431309461593628, + "learning_rate": 2.4972134876591618e-06, + "loss": 0.4413, + "step": 1266 + }, + { + "epoch": 0.7006451612903226, + "grad_norm": 0.3382079601287842, + "learning_rate": 2.4888601762085518e-06, + "loss": 0.502, + "step": 1267 + }, + { + "epoch": 0.7011981566820277, + "grad_norm": 0.31066352128982544, + "learning_rate": 2.4805162273348905e-06, + "loss": 0.4596, + "step": 1268 + }, + { + "epoch": 0.7017511520737327, + "grad_norm": 0.31966033577919006, + "learning_rate": 2.4721816721479007e-06, + "loss": 0.4769, + "step": 1269 + }, + { + "epoch": 0.7023041474654378, + "grad_norm": 0.3357180058956146, + "learning_rate": 2.4638565417222816e-06, + "loss": 0.437, + "step": 1270 + }, + { + "epoch": 0.7028571428571428, + "grad_norm": 0.335115522146225, + "learning_rate": 2.4555408670975928e-06, + "loss": 0.4498, + "step": 1271 + }, + { + "epoch": 0.7034101382488479, + "grad_norm": 0.33428287506103516, + "learning_rate": 2.4472346792781366e-06, + "loss": 0.4661, + "step": 1272 + }, + { + "epoch": 0.703963133640553, + "grad_norm": 0.3672966957092285, + "learning_rate": 2.438938009232851e-06, + "loss": 0.4775, + "step": 1273 + }, + { + "epoch": 0.704516129032258, + "grad_norm": 0.3439885079860687, + "learning_rate": 2.4306508878951805e-06, + "loss": 0.4611, + "step": 1274 + }, + { + "epoch": 0.7050691244239631, + "grad_norm": 0.3395749628543854, + "learning_rate": 2.4223733461629716e-06, + "loss": 0.4498, + "step": 1275 + }, + { + "epoch": 0.7056221198156682, + "grad_norm": 0.35915130376815796, + "learning_rate": 2.4141054148983532e-06, + "loss": 0.4767, + "step": 1276 + }, + { + "epoch": 0.7061751152073733, + "grad_norm": 0.393310546875, + "learning_rate": 2.4058471249276232e-06, + "loss": 0.4594, + "step": 1277 + }, + { + "epoch": 0.7067281105990784, + "grad_norm": 0.3339405357837677, + "learning_rate": 2.397598507041132e-06, + "loss": 0.4732, + "step": 1278 + }, + { + "epoch": 0.7072811059907834, + "grad_norm": 0.35175976157188416, + "learning_rate": 2.3893595919931673e-06, + "loss": 0.4608, + "step": 1279 + }, + { + "epoch": 0.7078341013824885, + "grad_norm": 0.33776116371154785, + "learning_rate": 2.381130410501845e-06, + "loss": 0.4568, + "step": 1280 + }, + { + "epoch": 0.7083870967741935, + "grad_norm": 0.34300360083580017, + "learning_rate": 2.3729109932489825e-06, + "loss": 0.4802, + "step": 1281 + }, + { + "epoch": 0.7089400921658986, + "grad_norm": 0.36740466952323914, + "learning_rate": 2.3647013708799987e-06, + "loss": 0.4755, + "step": 1282 + }, + { + "epoch": 0.7094930875576037, + "grad_norm": 0.3558812439441681, + "learning_rate": 2.3565015740037915e-06, + "loss": 0.4623, + "step": 1283 + }, + { + "epoch": 0.7100460829493087, + "grad_norm": 0.4075135886669159, + "learning_rate": 2.3483116331926244e-06, + "loss": 0.4612, + "step": 1284 + }, + { + "epoch": 0.7105990783410139, + "grad_norm": 0.3450377285480499, + "learning_rate": 2.340131578982013e-06, + "loss": 0.4695, + "step": 1285 + }, + { + "epoch": 0.7111520737327189, + "grad_norm": 0.3357861042022705, + "learning_rate": 2.3319614418706127e-06, + "loss": 0.4826, + "step": 1286 + }, + { + "epoch": 0.711705069124424, + "grad_norm": 0.3738630712032318, + "learning_rate": 2.3238012523201035e-06, + "loss": 0.4905, + "step": 1287 + }, + { + "epoch": 0.712258064516129, + "grad_norm": 0.3660827577114105, + "learning_rate": 2.315651040755077e-06, + "loss": 0.4754, + "step": 1288 + }, + { + "epoch": 0.7128110599078341, + "grad_norm": 0.3476543426513672, + "learning_rate": 2.3075108375629218e-06, + "loss": 0.4728, + "step": 1289 + }, + { + "epoch": 0.7133640552995392, + "grad_norm": 0.33124610781669617, + "learning_rate": 2.299380673093712e-06, + "loss": 0.474, + "step": 1290 + }, + { + "epoch": 0.7139170506912442, + "grad_norm": 0.33316168189048767, + "learning_rate": 2.291260577660092e-06, + "loss": 0.4853, + "step": 1291 + }, + { + "epoch": 0.7144700460829493, + "grad_norm": 0.3274078369140625, + "learning_rate": 2.2831505815371656e-06, + "loss": 0.4571, + "step": 1292 + }, + { + "epoch": 0.7150230414746543, + "grad_norm": 0.3307998776435852, + "learning_rate": 2.275050714962383e-06, + "loss": 0.4498, + "step": 1293 + }, + { + "epoch": 0.7155760368663594, + "grad_norm": 0.3073872923851013, + "learning_rate": 2.2669610081354254e-06, + "loss": 0.4661, + "step": 1294 + }, + { + "epoch": 0.7161290322580646, + "grad_norm": 0.37436577677726746, + "learning_rate": 2.2588814912180944e-06, + "loss": 0.4746, + "step": 1295 + }, + { + "epoch": 0.7166820276497696, + "grad_norm": 0.3833792507648468, + "learning_rate": 2.2508121943342008e-06, + "loss": 0.4502, + "step": 1296 + }, + { + "epoch": 0.7172350230414747, + "grad_norm": 0.3415040075778961, + "learning_rate": 2.2427531475694496e-06, + "loss": 0.4652, + "step": 1297 + }, + { + "epoch": 0.7177880184331797, + "grad_norm": 0.3346996605396271, + "learning_rate": 2.2347043809713306e-06, + "loss": 0.4474, + "step": 1298 + }, + { + "epoch": 0.7183410138248848, + "grad_norm": 0.34259018301963806, + "learning_rate": 2.2266659245490017e-06, + "loss": 0.4832, + "step": 1299 + }, + { + "epoch": 0.7188940092165899, + "grad_norm": 0.3417535424232483, + "learning_rate": 2.218637808273184e-06, + "loss": 0.4674, + "step": 1300 + }, + { + "epoch": 0.7194470046082949, + "grad_norm": 0.37401384115219116, + "learning_rate": 2.210620062076044e-06, + "loss": 0.4859, + "step": 1301 + }, + { + "epoch": 0.72, + "grad_norm": 0.33238619565963745, + "learning_rate": 2.2026127158510843e-06, + "loss": 0.4441, + "step": 1302 + }, + { + "epoch": 0.720552995391705, + "grad_norm": 0.32708191871643066, + "learning_rate": 2.194615799453032e-06, + "loss": 0.4793, + "step": 1303 + }, + { + "epoch": 0.7211059907834101, + "grad_norm": 0.30514681339263916, + "learning_rate": 2.186629342697727e-06, + "loss": 0.4409, + "step": 1304 + }, + { + "epoch": 0.7216589861751153, + "grad_norm": 0.35654860734939575, + "learning_rate": 2.1786533753620155e-06, + "loss": 0.5136, + "step": 1305 + }, + { + "epoch": 0.7222119815668203, + "grad_norm": 0.3565172255039215, + "learning_rate": 2.170687927183629e-06, + "loss": 0.4683, + "step": 1306 + }, + { + "epoch": 0.7227649769585254, + "grad_norm": 0.3977636396884918, + "learning_rate": 2.1627330278610824e-06, + "loss": 0.4781, + "step": 1307 + }, + { + "epoch": 0.7233179723502304, + "grad_norm": 0.37379372119903564, + "learning_rate": 2.154788707053559e-06, + "loss": 0.4713, + "step": 1308 + }, + { + "epoch": 0.7238709677419355, + "grad_norm": 0.3560510575771332, + "learning_rate": 2.1468549943808033e-06, + "loss": 0.4725, + "step": 1309 + }, + { + "epoch": 0.7244239631336405, + "grad_norm": 0.3701375424861908, + "learning_rate": 2.1389319194230017e-06, + "loss": 0.4887, + "step": 1310 + }, + { + "epoch": 0.7249769585253456, + "grad_norm": 0.35575488209724426, + "learning_rate": 2.1310195117206863e-06, + "loss": 0.4628, + "step": 1311 + }, + { + "epoch": 0.7255299539170507, + "grad_norm": 0.34897711873054504, + "learning_rate": 2.1231178007746136e-06, + "loss": 0.443, + "step": 1312 + }, + { + "epoch": 0.7260829493087557, + "grad_norm": 0.36126595735549927, + "learning_rate": 2.1152268160456584e-06, + "loss": 0.4622, + "step": 1313 + }, + { + "epoch": 0.7266359447004608, + "grad_norm": 0.3314565420150757, + "learning_rate": 2.1073465869547043e-06, + "loss": 0.4418, + "step": 1314 + }, + { + "epoch": 0.727188940092166, + "grad_norm": 0.32749220728874207, + "learning_rate": 2.0994771428825332e-06, + "loss": 0.4325, + "step": 1315 + }, + { + "epoch": 0.727741935483871, + "grad_norm": 0.39038175344467163, + "learning_rate": 2.0916185131697163e-06, + "loss": 0.4928, + "step": 1316 + }, + { + "epoch": 0.7282949308755761, + "grad_norm": 0.3242311477661133, + "learning_rate": 2.083770727116503e-06, + "loss": 0.4708, + "step": 1317 + }, + { + "epoch": 0.7288479262672811, + "grad_norm": 0.3813181221485138, + "learning_rate": 2.0759338139827145e-06, + "loss": 0.4959, + "step": 1318 + }, + { + "epoch": 0.7294009216589862, + "grad_norm": 0.31590506434440613, + "learning_rate": 2.0681078029876322e-06, + "loss": 0.451, + "step": 1319 + }, + { + "epoch": 0.7299539170506912, + "grad_norm": 0.3604491651058197, + "learning_rate": 2.0602927233098908e-06, + "loss": 0.4817, + "step": 1320 + }, + { + "epoch": 0.7305069124423963, + "grad_norm": 0.3593979775905609, + "learning_rate": 2.0524886040873676e-06, + "loss": 0.4707, + "step": 1321 + }, + { + "epoch": 0.7310599078341014, + "grad_norm": 0.38674402236938477, + "learning_rate": 2.0446954744170748e-06, + "loss": 0.4522, + "step": 1322 + }, + { + "epoch": 0.7316129032258064, + "grad_norm": 0.3926672637462616, + "learning_rate": 2.036913363355052e-06, + "loss": 0.4582, + "step": 1323 + }, + { + "epoch": 0.7321658986175115, + "grad_norm": 0.3277740776538849, + "learning_rate": 2.029142299916255e-06, + "loss": 0.4913, + "step": 1324 + }, + { + "epoch": 0.7327188940092166, + "grad_norm": 0.3636924922466278, + "learning_rate": 2.0213823130744516e-06, + "loss": 0.4624, + "step": 1325 + }, + { + "epoch": 0.7332718894009217, + "grad_norm": 0.36329200863838196, + "learning_rate": 2.0136334317621093e-06, + "loss": 0.4551, + "step": 1326 + }, + { + "epoch": 0.7338248847926268, + "grad_norm": 0.3718715012073517, + "learning_rate": 2.005895684870291e-06, + "loss": 0.4717, + "step": 1327 + }, + { + "epoch": 0.7343778801843318, + "grad_norm": 0.3874320685863495, + "learning_rate": 1.9981691012485455e-06, + "loss": 0.4757, + "step": 1328 + }, + { + "epoch": 0.7349308755760369, + "grad_norm": 0.29239824414253235, + "learning_rate": 1.9904537097048004e-06, + "loss": 0.4554, + "step": 1329 + }, + { + "epoch": 0.7354838709677419, + "grad_norm": 0.3096185326576233, + "learning_rate": 1.982749539005254e-06, + "loss": 0.4559, + "step": 1330 + }, + { + "epoch": 0.736036866359447, + "grad_norm": 0.34961435198783875, + "learning_rate": 1.9750566178742697e-06, + "loss": 0.4376, + "step": 1331 + }, + { + "epoch": 0.736589861751152, + "grad_norm": 0.36215296387672424, + "learning_rate": 1.9673749749942655e-06, + "loss": 0.4544, + "step": 1332 + }, + { + "epoch": 0.7371428571428571, + "grad_norm": 0.3274637758731842, + "learning_rate": 1.959704639005613e-06, + "loss": 0.4468, + "step": 1333 + }, + { + "epoch": 0.7376958525345622, + "grad_norm": 0.324142187833786, + "learning_rate": 1.952045638506523e-06, + "loss": 0.4775, + "step": 1334 + }, + { + "epoch": 0.7382488479262673, + "grad_norm": 0.30846402049064636, + "learning_rate": 1.9443980020529456e-06, + "loss": 0.4507, + "step": 1335 + }, + { + "epoch": 0.7388018433179724, + "grad_norm": 0.36270537972450256, + "learning_rate": 1.9367617581584606e-06, + "loss": 0.4526, + "step": 1336 + }, + { + "epoch": 0.7393548387096774, + "grad_norm": 0.35891664028167725, + "learning_rate": 1.9291369352941696e-06, + "loss": 0.4606, + "step": 1337 + }, + { + "epoch": 0.7399078341013825, + "grad_norm": 0.3416483700275421, + "learning_rate": 1.9215235618885964e-06, + "loss": 0.4976, + "step": 1338 + }, + { + "epoch": 0.7404608294930876, + "grad_norm": 0.30863887071609497, + "learning_rate": 1.9139216663275727e-06, + "loss": 0.44, + "step": 1339 + }, + { + "epoch": 0.7410138248847926, + "grad_norm": 0.313449501991272, + "learning_rate": 1.9063312769541348e-06, + "loss": 0.4643, + "step": 1340 + }, + { + "epoch": 0.7415668202764977, + "grad_norm": 0.35145795345306396, + "learning_rate": 1.8987524220684222e-06, + "loss": 0.4689, + "step": 1341 + }, + { + "epoch": 0.7421198156682027, + "grad_norm": 0.34510427713394165, + "learning_rate": 1.8911851299275675e-06, + "loss": 0.498, + "step": 1342 + }, + { + "epoch": 0.7426728110599078, + "grad_norm": 0.3454337418079376, + "learning_rate": 1.8836294287455936e-06, + "loss": 0.4488, + "step": 1343 + }, + { + "epoch": 0.743225806451613, + "grad_norm": 0.3266771733760834, + "learning_rate": 1.8760853466933072e-06, + "loss": 0.4855, + "step": 1344 + }, + { + "epoch": 0.743778801843318, + "grad_norm": 0.3165615200996399, + "learning_rate": 1.8685529118981926e-06, + "loss": 0.4556, + "step": 1345 + }, + { + "epoch": 0.7443317972350231, + "grad_norm": 0.33930885791778564, + "learning_rate": 1.8610321524443099e-06, + "loss": 0.4401, + "step": 1346 + }, + { + "epoch": 0.7448847926267281, + "grad_norm": 0.35839566588401794, + "learning_rate": 1.8535230963721884e-06, + "loss": 0.4965, + "step": 1347 + }, + { + "epoch": 0.7454377880184332, + "grad_norm": 0.33815956115722656, + "learning_rate": 1.8460257716787216e-06, + "loss": 0.4868, + "step": 1348 + }, + { + "epoch": 0.7459907834101382, + "grad_norm": 0.34670326113700867, + "learning_rate": 1.8385402063170643e-06, + "loss": 0.4696, + "step": 1349 + }, + { + "epoch": 0.7465437788018433, + "grad_norm": 0.32649490237236023, + "learning_rate": 1.8310664281965268e-06, + "loss": 0.4951, + "step": 1350 + }, + { + "epoch": 0.7470967741935484, + "grad_norm": 0.32630065083503723, + "learning_rate": 1.8236044651824725e-06, + "loss": 0.4733, + "step": 1351 + }, + { + "epoch": 0.7476497695852534, + "grad_norm": 0.2883811593055725, + "learning_rate": 1.8161543450962127e-06, + "loss": 0.4637, + "step": 1352 + }, + { + "epoch": 0.7482027649769585, + "grad_norm": 0.34161558747291565, + "learning_rate": 1.8087160957149036e-06, + "loss": 0.4629, + "step": 1353 + }, + { + "epoch": 0.7487557603686636, + "grad_norm": 0.3212891221046448, + "learning_rate": 1.8012897447714417e-06, + "loss": 0.4703, + "step": 1354 + }, + { + "epoch": 0.7493087557603687, + "grad_norm": 0.3223973214626312, + "learning_rate": 1.7938753199543618e-06, + "loss": 0.4536, + "step": 1355 + }, + { + "epoch": 0.7498617511520738, + "grad_norm": 0.3456159830093384, + "learning_rate": 1.7864728489077332e-06, + "loss": 0.4671, + "step": 1356 + }, + { + "epoch": 0.7504147465437788, + "grad_norm": 0.35950833559036255, + "learning_rate": 1.779082359231057e-06, + "loss": 0.4674, + "step": 1357 + }, + { + "epoch": 0.7509677419354839, + "grad_norm": 0.33876293897628784, + "learning_rate": 1.7717038784791612e-06, + "loss": 0.4614, + "step": 1358 + }, + { + "epoch": 0.7515207373271889, + "grad_norm": 0.34674006700515747, + "learning_rate": 1.7643374341621006e-06, + "loss": 0.4434, + "step": 1359 + }, + { + "epoch": 0.752073732718894, + "grad_norm": 0.3747800290584564, + "learning_rate": 1.7569830537450533e-06, + "loss": 0.4607, + "step": 1360 + }, + { + "epoch": 0.752626728110599, + "grad_norm": 0.3502337336540222, + "learning_rate": 1.7496407646482182e-06, + "loss": 0.471, + "step": 1361 + }, + { + "epoch": 0.7531797235023041, + "grad_norm": 0.3303203582763672, + "learning_rate": 1.7423105942467116e-06, + "loss": 0.4801, + "step": 1362 + }, + { + "epoch": 0.7537327188940092, + "grad_norm": 0.3284555673599243, + "learning_rate": 1.7349925698704673e-06, + "loss": 0.4742, + "step": 1363 + }, + { + "epoch": 0.7542857142857143, + "grad_norm": 0.3339201509952545, + "learning_rate": 1.7276867188041336e-06, + "loss": 0.4588, + "step": 1364 + }, + { + "epoch": 0.7548387096774194, + "grad_norm": 0.35038480162620544, + "learning_rate": 1.7203930682869707e-06, + "loss": 0.4921, + "step": 1365 + }, + { + "epoch": 0.7553917050691245, + "grad_norm": 0.3145638406276703, + "learning_rate": 1.7131116455127518e-06, + "loss": 0.428, + "step": 1366 + }, + { + "epoch": 0.7559447004608295, + "grad_norm": 0.34270545840263367, + "learning_rate": 1.7058424776296583e-06, + "loss": 0.4599, + "step": 1367 + }, + { + "epoch": 0.7564976958525346, + "grad_norm": 0.33915436267852783, + "learning_rate": 1.698585591740181e-06, + "loss": 0.4713, + "step": 1368 + }, + { + "epoch": 0.7570506912442396, + "grad_norm": 0.3200076222419739, + "learning_rate": 1.6913410149010179e-06, + "loss": 0.4469, + "step": 1369 + }, + { + "epoch": 0.7576036866359447, + "grad_norm": 0.31938573718070984, + "learning_rate": 1.6841087741229745e-06, + "loss": 0.4563, + "step": 1370 + }, + { + "epoch": 0.7581566820276497, + "grad_norm": 0.3671538233757019, + "learning_rate": 1.6768888963708612e-06, + "loss": 0.471, + "step": 1371 + }, + { + "epoch": 0.7587096774193548, + "grad_norm": 0.33220696449279785, + "learning_rate": 1.669681408563395e-06, + "loss": 0.444, + "step": 1372 + }, + { + "epoch": 0.7592626728110599, + "grad_norm": 0.3261774182319641, + "learning_rate": 1.6624863375730977e-06, + "loss": 0.4775, + "step": 1373 + }, + { + "epoch": 0.759815668202765, + "grad_norm": 0.31591543555259705, + "learning_rate": 1.6553037102261955e-06, + "loss": 0.4763, + "step": 1374 + }, + { + "epoch": 0.7603686635944701, + "grad_norm": 0.345234215259552, + "learning_rate": 1.6481335533025195e-06, + "loss": 0.443, + "step": 1375 + }, + { + "epoch": 0.7609216589861751, + "grad_norm": 0.3390832245349884, + "learning_rate": 1.6409758935354065e-06, + "loss": 0.4622, + "step": 1376 + }, + { + "epoch": 0.7614746543778802, + "grad_norm": 0.34640833735466003, + "learning_rate": 1.6338307576115987e-06, + "loss": 0.4606, + "step": 1377 + }, + { + "epoch": 0.7620276497695853, + "grad_norm": 0.3271695375442505, + "learning_rate": 1.6266981721711438e-06, + "loss": 0.4354, + "step": 1378 + }, + { + "epoch": 0.7625806451612903, + "grad_norm": 0.34254854917526245, + "learning_rate": 1.619578163807296e-06, + "loss": 0.4601, + "step": 1379 + }, + { + "epoch": 0.7631336405529954, + "grad_norm": 0.3350234925746918, + "learning_rate": 1.6124707590664168e-06, + "loss": 0.4652, + "step": 1380 + }, + { + "epoch": 0.7636866359447004, + "grad_norm": 0.3148898780345917, + "learning_rate": 1.6053759844478768e-06, + "loss": 0.4702, + "step": 1381 + }, + { + "epoch": 0.7642396313364055, + "grad_norm": 0.3196204602718353, + "learning_rate": 1.5982938664039555e-06, + "loss": 0.464, + "step": 1382 + }, + { + "epoch": 0.7647926267281105, + "grad_norm": 0.3282592296600342, + "learning_rate": 1.591224431339744e-06, + "loss": 0.4631, + "step": 1383 + }, + { + "epoch": 0.7653456221198157, + "grad_norm": 0.34297293424606323, + "learning_rate": 1.584167705613046e-06, + "loss": 0.4494, + "step": 1384 + }, + { + "epoch": 0.7658986175115208, + "grad_norm": 0.34436681866645813, + "learning_rate": 1.5771237155342784e-06, + "loss": 0.4501, + "step": 1385 + }, + { + "epoch": 0.7664516129032258, + "grad_norm": 0.2950928807258606, + "learning_rate": 1.5700924873663758e-06, + "loss": 0.4493, + "step": 1386 + }, + { + "epoch": 0.7670046082949309, + "grad_norm": 0.3339753746986389, + "learning_rate": 1.5630740473246896e-06, + "loss": 0.4793, + "step": 1387 + }, + { + "epoch": 0.767557603686636, + "grad_norm": 0.3539128005504608, + "learning_rate": 1.5560684215768935e-06, + "loss": 0.4459, + "step": 1388 + }, + { + "epoch": 0.768110599078341, + "grad_norm": 0.35491037368774414, + "learning_rate": 1.549075636242882e-06, + "loss": 0.4814, + "step": 1389 + }, + { + "epoch": 0.7686635944700461, + "grad_norm": 0.3396543562412262, + "learning_rate": 1.5420957173946772e-06, + "loss": 0.4526, + "step": 1390 + }, + { + "epoch": 0.7692165898617511, + "grad_norm": 0.35023194551467896, + "learning_rate": 1.5351286910563278e-06, + "loss": 0.432, + "step": 1391 + }, + { + "epoch": 0.7697695852534562, + "grad_norm": 0.3362998366355896, + "learning_rate": 1.5281745832038159e-06, + "loss": 0.4518, + "step": 1392 + }, + { + "epoch": 0.7703225806451612, + "grad_norm": 0.3237576186656952, + "learning_rate": 1.5212334197649564e-06, + "loss": 0.4732, + "step": 1393 + }, + { + "epoch": 0.7708755760368664, + "grad_norm": 0.3349010646343231, + "learning_rate": 1.514305226619302e-06, + "loss": 0.4793, + "step": 1394 + }, + { + "epoch": 0.7714285714285715, + "grad_norm": 0.3398427367210388, + "learning_rate": 1.5073900295980481e-06, + "loss": 0.471, + "step": 1395 + }, + { + "epoch": 0.7719815668202765, + "grad_norm": 0.33397191762924194, + "learning_rate": 1.5004878544839335e-06, + "loss": 0.4398, + "step": 1396 + }, + { + "epoch": 0.7725345622119816, + "grad_norm": 0.3291890025138855, + "learning_rate": 1.493598727011148e-06, + "loss": 0.4632, + "step": 1397 + }, + { + "epoch": 0.7730875576036866, + "grad_norm": 0.33560484647750854, + "learning_rate": 1.4867226728652318e-06, + "loss": 0.4661, + "step": 1398 + }, + { + "epoch": 0.7736405529953917, + "grad_norm": 0.32114970684051514, + "learning_rate": 1.4798597176829844e-06, + "loss": 0.4558, + "step": 1399 + }, + { + "epoch": 0.7741935483870968, + "grad_norm": 0.29689663648605347, + "learning_rate": 1.4730098870523652e-06, + "loss": 0.4567, + "step": 1400 + }, + { + "epoch": 0.7747465437788018, + "grad_norm": 0.33287614583969116, + "learning_rate": 1.4661732065124012e-06, + "loss": 0.4458, + "step": 1401 + }, + { + "epoch": 0.7752995391705069, + "grad_norm": 0.32061877846717834, + "learning_rate": 1.45934970155309e-06, + "loss": 0.4482, + "step": 1402 + }, + { + "epoch": 0.7758525345622119, + "grad_norm": 0.32134944200515747, + "learning_rate": 1.4525393976153046e-06, + "loss": 0.4669, + "step": 1403 + }, + { + "epoch": 0.7764055299539171, + "grad_norm": 0.31328532099723816, + "learning_rate": 1.4457423200906994e-06, + "loss": 0.474, + "step": 1404 + }, + { + "epoch": 0.7769585253456222, + "grad_norm": 0.31957975029945374, + "learning_rate": 1.4389584943216156e-06, + "loss": 0.4478, + "step": 1405 + }, + { + "epoch": 0.7775115207373272, + "grad_norm": 0.318330317735672, + "learning_rate": 1.4321879456009858e-06, + "loss": 0.4738, + "step": 1406 + }, + { + "epoch": 0.7780645161290323, + "grad_norm": 0.3517070710659027, + "learning_rate": 1.4254306991722406e-06, + "loss": 0.49, + "step": 1407 + }, + { + "epoch": 0.7786175115207373, + "grad_norm": 0.3362846374511719, + "learning_rate": 1.4186867802292132e-06, + "loss": 0.4756, + "step": 1408 + }, + { + "epoch": 0.7791705069124424, + "grad_norm": 0.32801997661590576, + "learning_rate": 1.411956213916048e-06, + "loss": 0.4745, + "step": 1409 + }, + { + "epoch": 0.7797235023041474, + "grad_norm": 0.32685017585754395, + "learning_rate": 1.4052390253271037e-06, + "loss": 0.4618, + "step": 1410 + }, + { + "epoch": 0.7802764976958525, + "grad_norm": 0.32221168279647827, + "learning_rate": 1.3985352395068618e-06, + "loss": 0.4373, + "step": 1411 + }, + { + "epoch": 0.7808294930875576, + "grad_norm": 0.3476710915565491, + "learning_rate": 1.3918448814498336e-06, + "loss": 0.4463, + "step": 1412 + }, + { + "epoch": 0.7813824884792627, + "grad_norm": 0.3333290219306946, + "learning_rate": 1.3851679761004644e-06, + "loss": 0.4592, + "step": 1413 + }, + { + "epoch": 0.7819354838709678, + "grad_norm": 0.32516762614250183, + "learning_rate": 1.3785045483530435e-06, + "loss": 0.476, + "step": 1414 + }, + { + "epoch": 0.7824884792626728, + "grad_norm": 0.3284921944141388, + "learning_rate": 1.3718546230516095e-06, + "loss": 0.4669, + "step": 1415 + }, + { + "epoch": 0.7830414746543779, + "grad_norm": 0.3187192976474762, + "learning_rate": 1.3652182249898583e-06, + "loss": 0.478, + "step": 1416 + }, + { + "epoch": 0.783594470046083, + "grad_norm": 0.3000369369983673, + "learning_rate": 1.3585953789110506e-06, + "loss": 0.4763, + "step": 1417 + }, + { + "epoch": 0.784147465437788, + "grad_norm": 0.3406629264354706, + "learning_rate": 1.3519861095079202e-06, + "loss": 0.451, + "step": 1418 + }, + { + "epoch": 0.7847004608294931, + "grad_norm": 0.3349382281303406, + "learning_rate": 1.3453904414225799e-06, + "loss": 0.4573, + "step": 1419 + }, + { + "epoch": 0.7852534562211981, + "grad_norm": 0.3069670796394348, + "learning_rate": 1.3388083992464335e-06, + "loss": 0.4739, + "step": 1420 + }, + { + "epoch": 0.7858064516129032, + "grad_norm": 0.31198862195014954, + "learning_rate": 1.3322400075200792e-06, + "loss": 0.454, + "step": 1421 + }, + { + "epoch": 0.7863594470046082, + "grad_norm": 0.30845198035240173, + "learning_rate": 1.325685290733223e-06, + "loss": 0.4498, + "step": 1422 + }, + { + "epoch": 0.7869124423963134, + "grad_norm": 0.31618401408195496, + "learning_rate": 1.3191442733245824e-06, + "loss": 0.491, + "step": 1423 + }, + { + "epoch": 0.7874654377880185, + "grad_norm": 0.356711208820343, + "learning_rate": 1.3126169796818001e-06, + "loss": 0.4422, + "step": 1424 + }, + { + "epoch": 0.7880184331797235, + "grad_norm": 0.35151243209838867, + "learning_rate": 1.3061034341413497e-06, + "loss": 0.469, + "step": 1425 + }, + { + "epoch": 0.7885714285714286, + "grad_norm": 0.30518248677253723, + "learning_rate": 1.2996036609884478e-06, + "loss": 0.4778, + "step": 1426 + }, + { + "epoch": 0.7891244239631336, + "grad_norm": 0.30732637643814087, + "learning_rate": 1.2931176844569588e-06, + "loss": 0.4627, + "step": 1427 + }, + { + "epoch": 0.7896774193548387, + "grad_norm": 0.3301125466823578, + "learning_rate": 1.2866455287293094e-06, + "loss": 0.4711, + "step": 1428 + }, + { + "epoch": 0.7902304147465438, + "grad_norm": 0.3300682008266449, + "learning_rate": 1.2801872179363978e-06, + "loss": 0.497, + "step": 1429 + }, + { + "epoch": 0.7907834101382488, + "grad_norm": 0.3206273019313812, + "learning_rate": 1.2737427761575006e-06, + "loss": 0.4593, + "step": 1430 + }, + { + "epoch": 0.7913364055299539, + "grad_norm": 0.35528701543807983, + "learning_rate": 1.2673122274201844e-06, + "loss": 0.452, + "step": 1431 + }, + { + "epoch": 0.7918894009216589, + "grad_norm": 0.3094896972179413, + "learning_rate": 1.2608955957002196e-06, + "loss": 0.4426, + "step": 1432 + }, + { + "epoch": 0.7924423963133641, + "grad_norm": 0.3296172618865967, + "learning_rate": 1.2544929049214843e-06, + "loss": 0.4751, + "step": 1433 + }, + { + "epoch": 0.7929953917050692, + "grad_norm": 0.3449612855911255, + "learning_rate": 1.248104178955883e-06, + "loss": 0.473, + "step": 1434 + }, + { + "epoch": 0.7935483870967742, + "grad_norm": 0.30491313338279724, + "learning_rate": 1.2417294416232505e-06, + "loss": 0.447, + "step": 1435 + }, + { + "epoch": 0.7941013824884793, + "grad_norm": 0.32028716802597046, + "learning_rate": 1.235368716691267e-06, + "loss": 0.4358, + "step": 1436 + }, + { + "epoch": 0.7946543778801843, + "grad_norm": 0.3335442543029785, + "learning_rate": 1.22902202787537e-06, + "loss": 0.476, + "step": 1437 + }, + { + "epoch": 0.7952073732718894, + "grad_norm": 0.3434390723705292, + "learning_rate": 1.2226893988386618e-06, + "loss": 0.4447, + "step": 1438 + }, + { + "epoch": 0.7957603686635945, + "grad_norm": 0.3305012881755829, + "learning_rate": 1.2163708531918267e-06, + "loss": 0.4644, + "step": 1439 + }, + { + "epoch": 0.7963133640552995, + "grad_norm": 0.32044875621795654, + "learning_rate": 1.210066414493039e-06, + "loss": 0.4525, + "step": 1440 + }, + { + "epoch": 0.7968663594470046, + "grad_norm": 0.3219109773635864, + "learning_rate": 1.2037761062478759e-06, + "loss": 0.4614, + "step": 1441 + }, + { + "epoch": 0.7974193548387096, + "grad_norm": 0.3471842408180237, + "learning_rate": 1.1974999519092311e-06, + "loss": 0.4541, + "step": 1442 + }, + { + "epoch": 0.7979723502304148, + "grad_norm": 0.2974897623062134, + "learning_rate": 1.1912379748772267e-06, + "loss": 0.4344, + "step": 1443 + }, + { + "epoch": 0.7985253456221199, + "grad_norm": 0.31679412722587585, + "learning_rate": 1.1849901984991253e-06, + "loss": 0.456, + "step": 1444 + }, + { + "epoch": 0.7990783410138249, + "grad_norm": 0.30493679642677307, + "learning_rate": 1.1787566460692446e-06, + "loss": 0.4674, + "step": 1445 + }, + { + "epoch": 0.79963133640553, + "grad_norm": 0.3283561170101166, + "learning_rate": 1.1725373408288682e-06, + "loss": 0.4621, + "step": 1446 + }, + { + "epoch": 0.800184331797235, + "grad_norm": 0.32548418641090393, + "learning_rate": 1.1663323059661609e-06, + "loss": 0.4422, + "step": 1447 + }, + { + "epoch": 0.8007373271889401, + "grad_norm": 0.29554906487464905, + "learning_rate": 1.1601415646160813e-06, + "loss": 0.4535, + "step": 1448 + }, + { + "epoch": 0.8012903225806451, + "grad_norm": 0.31748977303504944, + "learning_rate": 1.153965139860297e-06, + "loss": 0.4579, + "step": 1449 + }, + { + "epoch": 0.8018433179723502, + "grad_norm": 0.30266064405441284, + "learning_rate": 1.147803054727095e-06, + "loss": 0.446, + "step": 1450 + }, + { + "epoch": 0.8023963133640553, + "grad_norm": 0.32854562997817993, + "learning_rate": 1.1416553321913009e-06, + "loss": 0.4632, + "step": 1451 + }, + { + "epoch": 0.8029493087557603, + "grad_norm": 0.29594528675079346, + "learning_rate": 1.1355219951741881e-06, + "loss": 0.4704, + "step": 1452 + }, + { + "epoch": 0.8035023041474655, + "grad_norm": 0.2916601300239563, + "learning_rate": 1.1294030665433969e-06, + "loss": 0.4287, + "step": 1453 + }, + { + "epoch": 0.8040552995391705, + "grad_norm": 0.31938451528549194, + "learning_rate": 1.1232985691128457e-06, + "loss": 0.4242, + "step": 1454 + }, + { + "epoch": 0.8046082949308756, + "grad_norm": 0.3538084030151367, + "learning_rate": 1.1172085256426473e-06, + "loss": 0.4705, + "step": 1455 + }, + { + "epoch": 0.8051612903225807, + "grad_norm": 0.3275561034679413, + "learning_rate": 1.1111329588390253e-06, + "loss": 0.4745, + "step": 1456 + }, + { + "epoch": 0.8057142857142857, + "grad_norm": 0.30866336822509766, + "learning_rate": 1.1050718913542275e-06, + "loss": 0.453, + "step": 1457 + }, + { + "epoch": 0.8062672811059908, + "grad_norm": 0.3329750597476959, + "learning_rate": 1.0990253457864418e-06, + "loss": 0.4598, + "step": 1458 + }, + { + "epoch": 0.8068202764976958, + "grad_norm": 0.32834047079086304, + "learning_rate": 1.0929933446797136e-06, + "loss": 0.4649, + "step": 1459 + }, + { + "epoch": 0.8073732718894009, + "grad_norm": 0.3380810022354126, + "learning_rate": 1.0869759105238592e-06, + "loss": 0.4605, + "step": 1460 + }, + { + "epoch": 0.807926267281106, + "grad_norm": 0.3055575489997864, + "learning_rate": 1.0809730657543838e-06, + "loss": 0.4852, + "step": 1461 + }, + { + "epoch": 0.808479262672811, + "grad_norm": 0.30661699175834656, + "learning_rate": 1.0749848327523966e-06, + "loss": 0.4824, + "step": 1462 + }, + { + "epoch": 0.8090322580645162, + "grad_norm": 0.32859787344932556, + "learning_rate": 1.0690112338445292e-06, + "loss": 0.464, + "step": 1463 + }, + { + "epoch": 0.8095852534562212, + "grad_norm": 0.3309606611728668, + "learning_rate": 1.0630522913028508e-06, + "loss": 0.4585, + "step": 1464 + }, + { + "epoch": 0.8101382488479263, + "grad_norm": 0.33870929479599, + "learning_rate": 1.0571080273447858e-06, + "loss": 0.476, + "step": 1465 + }, + { + "epoch": 0.8106912442396313, + "grad_norm": 0.33712446689605713, + "learning_rate": 1.0511784641330286e-06, + "loss": 0.4476, + "step": 1466 + }, + { + "epoch": 0.8112442396313364, + "grad_norm": 0.2965930700302124, + "learning_rate": 1.0452636237754642e-06, + "loss": 0.4766, + "step": 1467 + }, + { + "epoch": 0.8117972350230415, + "grad_norm": 0.3398984372615814, + "learning_rate": 1.0393635283250885e-06, + "loss": 0.4716, + "step": 1468 + }, + { + "epoch": 0.8123502304147465, + "grad_norm": 0.3270639181137085, + "learning_rate": 1.0334781997799165e-06, + "loss": 0.4556, + "step": 1469 + }, + { + "epoch": 0.8129032258064516, + "grad_norm": 0.3468005955219269, + "learning_rate": 1.0276076600829094e-06, + "loss": 0.4725, + "step": 1470 + }, + { + "epoch": 0.8134562211981566, + "grad_norm": 0.31104719638824463, + "learning_rate": 1.0217519311218882e-06, + "loss": 0.4735, + "step": 1471 + }, + { + "epoch": 0.8140092165898618, + "grad_norm": 0.30012935400009155, + "learning_rate": 1.0159110347294536e-06, + "loss": 0.4624, + "step": 1472 + }, + { + "epoch": 0.8145622119815669, + "grad_norm": 0.3302370309829712, + "learning_rate": 1.0100849926829043e-06, + "loss": 0.4755, + "step": 1473 + }, + { + "epoch": 0.8151152073732719, + "grad_norm": 0.3273729681968689, + "learning_rate": 1.0042738267041552e-06, + "loss": 0.4558, + "step": 1474 + }, + { + "epoch": 0.815668202764977, + "grad_norm": 0.313931405544281, + "learning_rate": 9.984775584596585e-07, + "loss": 0.4778, + "step": 1475 + }, + { + "epoch": 0.816221198156682, + "grad_norm": 0.31616419553756714, + "learning_rate": 9.926962095603199e-07, + "loss": 0.4358, + "step": 1476 + }, + { + "epoch": 0.8167741935483871, + "grad_norm": 0.33137091994285583, + "learning_rate": 9.869298015614198e-07, + "loss": 0.4558, + "step": 1477 + }, + { + "epoch": 0.8173271889400922, + "grad_norm": 0.3267715573310852, + "learning_rate": 9.811783559625343e-07, + "loss": 0.475, + "step": 1478 + }, + { + "epoch": 0.8178801843317972, + "grad_norm": 0.37252572178840637, + "learning_rate": 9.754418942074512e-07, + "loss": 0.4829, + "step": 1479 + }, + { + "epoch": 0.8184331797235023, + "grad_norm": 0.3060474395751953, + "learning_rate": 9.697204376840936e-07, + "loss": 0.4627, + "step": 1480 + }, + { + "epoch": 0.8189861751152073, + "grad_norm": 0.32007962465286255, + "learning_rate": 9.640140077244382e-07, + "loss": 0.4698, + "step": 1481 + }, + { + "epoch": 0.8195391705069125, + "grad_norm": 0.2968813478946686, + "learning_rate": 9.58322625604437e-07, + "loss": 0.4316, + "step": 1482 + }, + { + "epoch": 0.8200921658986176, + "grad_norm": 0.31622204184532166, + "learning_rate": 9.526463125439372e-07, + "loss": 0.4408, + "step": 1483 + }, + { + "epoch": 0.8206451612903226, + "grad_norm": 0.2871773838996887, + "learning_rate": 9.469850897066014e-07, + "loss": 0.4563, + "step": 1484 + }, + { + "epoch": 0.8211981566820277, + "grad_norm": 0.30365270376205444, + "learning_rate": 9.413389781998311e-07, + "loss": 0.4519, + "step": 1485 + }, + { + "epoch": 0.8217511520737327, + "grad_norm": 0.3301312029361725, + "learning_rate": 9.357079990746853e-07, + "loss": 0.4543, + "step": 1486 + }, + { + "epoch": 0.8223041474654378, + "grad_norm": 0.3503333628177643, + "learning_rate": 9.300921733258039e-07, + "loss": 0.4753, + "step": 1487 + }, + { + "epoch": 0.8228571428571428, + "grad_norm": 0.30572423338890076, + "learning_rate": 9.244915218913275e-07, + "loss": 0.4694, + "step": 1488 + }, + { + "epoch": 0.8234101382488479, + "grad_norm": 0.30958428978919983, + "learning_rate": 9.18906065652822e-07, + "loss": 0.4544, + "step": 1489 + }, + { + "epoch": 0.823963133640553, + "grad_norm": 0.3019687831401825, + "learning_rate": 9.133358254351982e-07, + "loss": 0.4751, + "step": 1490 + }, + { + "epoch": 0.824516129032258, + "grad_norm": 0.3120213747024536, + "learning_rate": 9.077808220066359e-07, + "loss": 0.4586, + "step": 1491 + }, + { + "epoch": 0.8250691244239632, + "grad_norm": 0.3264663815498352, + "learning_rate": 9.022410760785055e-07, + "loss": 0.4569, + "step": 1492 + }, + { + "epoch": 0.8256221198156682, + "grad_norm": 0.3079594075679779, + "learning_rate": 8.967166083052908e-07, + "loss": 0.5003, + "step": 1493 + }, + { + "epoch": 0.8261751152073733, + "grad_norm": 0.30771124362945557, + "learning_rate": 8.912074392845144e-07, + "loss": 0.4654, + "step": 1494 + }, + { + "epoch": 0.8267281105990784, + "grad_norm": 0.31957051157951355, + "learning_rate": 8.857135895566538e-07, + "loss": 0.4722, + "step": 1495 + }, + { + "epoch": 0.8272811059907834, + "grad_norm": 0.30836620926856995, + "learning_rate": 8.80235079605074e-07, + "loss": 0.4557, + "step": 1496 + }, + { + "epoch": 0.8278341013824885, + "grad_norm": 0.30665266513824463, + "learning_rate": 8.747719298559465e-07, + "loss": 0.4516, + "step": 1497 + }, + { + "epoch": 0.8283870967741935, + "grad_norm": 0.33325862884521484, + "learning_rate": 8.693241606781728e-07, + "loss": 0.4667, + "step": 1498 + }, + { + "epoch": 0.8289400921658986, + "grad_norm": 0.3295386731624603, + "learning_rate": 8.638917923833074e-07, + "loss": 0.4602, + "step": 1499 + }, + { + "epoch": 0.8294930875576036, + "grad_norm": 0.3496408462524414, + "learning_rate": 8.584748452254888e-07, + "loss": 0.4601, + "step": 1500 + }, + { + "epoch": 0.8300460829493087, + "grad_norm": 0.2922409474849701, + "learning_rate": 8.530733394013546e-07, + "loss": 0.4787, + "step": 1501 + }, + { + "epoch": 0.8305990783410139, + "grad_norm": 0.3221864104270935, + "learning_rate": 8.476872950499726e-07, + "loss": 0.4401, + "step": 1502 + }, + { + "epoch": 0.8311520737327189, + "grad_norm": 0.33632180094718933, + "learning_rate": 8.423167322527626e-07, + "loss": 0.4636, + "step": 1503 + }, + { + "epoch": 0.831705069124424, + "grad_norm": 0.3501831293106079, + "learning_rate": 8.369616710334233e-07, + "loss": 0.4573, + "step": 1504 + }, + { + "epoch": 0.832258064516129, + "grad_norm": 0.31366902589797974, + "learning_rate": 8.316221313578576e-07, + "loss": 0.4583, + "step": 1505 + }, + { + "epoch": 0.8328110599078341, + "grad_norm": 0.36373165249824524, + "learning_rate": 8.262981331340969e-07, + "loss": 0.4633, + "step": 1506 + }, + { + "epoch": 0.8333640552995392, + "grad_norm": 0.31518691778182983, + "learning_rate": 8.209896962122282e-07, + "loss": 0.4599, + "step": 1507 + }, + { + "epoch": 0.8339170506912442, + "grad_norm": 0.3142349123954773, + "learning_rate": 8.15696840384319e-07, + "loss": 0.465, + "step": 1508 + }, + { + "epoch": 0.8344700460829493, + "grad_norm": 0.3050268292427063, + "learning_rate": 8.104195853843433e-07, + "loss": 0.4617, + "step": 1509 + }, + { + "epoch": 0.8350230414746543, + "grad_norm": 0.3328079879283905, + "learning_rate": 8.051579508881107e-07, + "loss": 0.4558, + "step": 1510 + }, + { + "epoch": 0.8355760368663594, + "grad_norm": 0.3367290496826172, + "learning_rate": 7.999119565131891e-07, + "loss": 0.4662, + "step": 1511 + }, + { + "epoch": 0.8361290322580646, + "grad_norm": 0.2858041226863861, + "learning_rate": 7.946816218188347e-07, + "loss": 0.4645, + "step": 1512 + }, + { + "epoch": 0.8366820276497696, + "grad_norm": 0.3110063374042511, + "learning_rate": 7.894669663059168e-07, + "loss": 0.4616, + "step": 1513 + }, + { + "epoch": 0.8372350230414747, + "grad_norm": 0.3313262462615967, + "learning_rate": 7.84268009416847e-07, + "loss": 0.4423, + "step": 1514 + }, + { + "epoch": 0.8377880184331797, + "grad_norm": 0.33478546142578125, + "learning_rate": 7.790847705355059e-07, + "loss": 0.4551, + "step": 1515 + }, + { + "epoch": 0.8383410138248848, + "grad_norm": 0.3241022527217865, + "learning_rate": 7.739172689871705e-07, + "loss": 0.4749, + "step": 1516 + }, + { + "epoch": 0.8388940092165899, + "grad_norm": 0.30135831236839294, + "learning_rate": 7.687655240384423e-07, + "loss": 0.4432, + "step": 1517 + }, + { + "epoch": 0.8394470046082949, + "grad_norm": 0.3093847930431366, + "learning_rate": 7.636295548971762e-07, + "loss": 0.4398, + "step": 1518 + }, + { + "epoch": 0.84, + "grad_norm": 0.31386277079582214, + "learning_rate": 7.585093807124077e-07, + "loss": 0.4308, + "step": 1519 + }, + { + "epoch": 0.840552995391705, + "grad_norm": 0.32683154940605164, + "learning_rate": 7.534050205742827e-07, + "loss": 0.4671, + "step": 1520 + }, + { + "epoch": 0.8411059907834101, + "grad_norm": 0.32057997584342957, + "learning_rate": 7.483164935139847e-07, + "loss": 0.4777, + "step": 1521 + }, + { + "epoch": 0.8416589861751153, + "grad_norm": 0.3294847011566162, + "learning_rate": 7.432438185036667e-07, + "loss": 0.4872, + "step": 1522 + }, + { + "epoch": 0.8422119815668203, + "grad_norm": 0.3206014931201935, + "learning_rate": 7.381870144563763e-07, + "loss": 0.4719, + "step": 1523 + }, + { + "epoch": 0.8427649769585254, + "grad_norm": 0.291071355342865, + "learning_rate": 7.33146100225991e-07, + "loss": 0.4496, + "step": 1524 + }, + { + "epoch": 0.8433179723502304, + "grad_norm": 0.31741753220558167, + "learning_rate": 7.281210946071393e-07, + "loss": 0.4305, + "step": 1525 + }, + { + "epoch": 0.8438709677419355, + "grad_norm": 0.3081185817718506, + "learning_rate": 7.231120163351396e-07, + "loss": 0.4796, + "step": 1526 + }, + { + "epoch": 0.8444239631336405, + "grad_norm": 0.31715425848960876, + "learning_rate": 7.181188840859266e-07, + "loss": 0.4827, + "step": 1527 + }, + { + "epoch": 0.8449769585253456, + "grad_norm": 0.3110186755657196, + "learning_rate": 7.131417164759791e-07, + "loss": 0.4652, + "step": 1528 + }, + { + "epoch": 0.8455299539170507, + "grad_norm": 0.3567166030406952, + "learning_rate": 7.081805320622559e-07, + "loss": 0.485, + "step": 1529 + }, + { + "epoch": 0.8460829493087557, + "grad_norm": 0.3137590289115906, + "learning_rate": 7.032353493421213e-07, + "loss": 0.445, + "step": 1530 + }, + { + "epoch": 0.8466359447004608, + "grad_norm": 0.3127385675907135, + "learning_rate": 6.983061867532798e-07, + "loss": 0.4691, + "step": 1531 + }, + { + "epoch": 0.8471889400921659, + "grad_norm": 0.3255513608455658, + "learning_rate": 6.933930626737057e-07, + "loss": 0.4418, + "step": 1532 + }, + { + "epoch": 0.847741935483871, + "grad_norm": 0.2911495566368103, + "learning_rate": 6.88495995421577e-07, + "loss": 0.4578, + "step": 1533 + }, + { + "epoch": 0.8482949308755761, + "grad_norm": 0.29997825622558594, + "learning_rate": 6.83615003255203e-07, + "loss": 0.4543, + "step": 1534 + }, + { + "epoch": 0.8488479262672811, + "grad_norm": 0.2882632315158844, + "learning_rate": 6.787501043729577e-07, + "loss": 0.4531, + "step": 1535 + }, + { + "epoch": 0.8494009216589862, + "grad_norm": 0.32424575090408325, + "learning_rate": 6.739013169132153e-07, + "loss": 0.4516, + "step": 1536 + }, + { + "epoch": 0.8499539170506912, + "grad_norm": 0.3189219534397125, + "learning_rate": 6.690686589542778e-07, + "loss": 0.456, + "step": 1537 + }, + { + "epoch": 0.8505069124423963, + "grad_norm": 0.2835007309913635, + "learning_rate": 6.642521485143099e-07, + "loss": 0.4874, + "step": 1538 + }, + { + "epoch": 0.8510599078341013, + "grad_norm": 0.30130714178085327, + "learning_rate": 6.594518035512726e-07, + "loss": 0.4698, + "step": 1539 + }, + { + "epoch": 0.8516129032258064, + "grad_norm": 0.3199770748615265, + "learning_rate": 6.546676419628545e-07, + "loss": 0.4822, + "step": 1540 + }, + { + "epoch": 0.8521658986175116, + "grad_norm": 0.31458160281181335, + "learning_rate": 6.498996815864068e-07, + "loss": 0.4737, + "step": 1541 + }, + { + "epoch": 0.8527188940092166, + "grad_norm": 0.33827102184295654, + "learning_rate": 6.451479401988736e-07, + "loss": 0.4607, + "step": 1542 + }, + { + "epoch": 0.8532718894009217, + "grad_norm": 0.3004834055900574, + "learning_rate": 6.404124355167302e-07, + "loss": 0.4621, + "step": 1543 + }, + { + "epoch": 0.8538248847926267, + "grad_norm": 0.2976799011230469, + "learning_rate": 6.356931851959136e-07, + "loss": 0.4571, + "step": 1544 + }, + { + "epoch": 0.8543778801843318, + "grad_norm": 0.31761622428894043, + "learning_rate": 6.30990206831758e-07, + "loss": 0.4647, + "step": 1545 + }, + { + "epoch": 0.8549308755760369, + "grad_norm": 0.30549031496047974, + "learning_rate": 6.263035179589288e-07, + "loss": 0.4619, + "step": 1546 + }, + { + "epoch": 0.8554838709677419, + "grad_norm": 0.3004506528377533, + "learning_rate": 6.216331360513572e-07, + "loss": 0.4786, + "step": 1547 + }, + { + "epoch": 0.856036866359447, + "grad_norm": 0.3054359257221222, + "learning_rate": 6.169790785221763e-07, + "loss": 0.4652, + "step": 1548 + }, + { + "epoch": 0.856589861751152, + "grad_norm": 0.3147340416908264, + "learning_rate": 6.123413627236536e-07, + "loss": 0.4573, + "step": 1549 + }, + { + "epoch": 0.8571428571428571, + "grad_norm": 0.3203125, + "learning_rate": 6.077200059471289e-07, + "loss": 0.4498, + "step": 1550 + }, + { + "epoch": 0.8576958525345623, + "grad_norm": 0.3389976918697357, + "learning_rate": 6.031150254229484e-07, + "loss": 0.4756, + "step": 1551 + }, + { + "epoch": 0.8582488479262673, + "grad_norm": 0.3188997507095337, + "learning_rate": 5.985264383204003e-07, + "loss": 0.4579, + "step": 1552 + }, + { + "epoch": 0.8588018433179724, + "grad_norm": 0.2967085540294647, + "learning_rate": 5.939542617476529e-07, + "loss": 0.4648, + "step": 1553 + }, + { + "epoch": 0.8593548387096774, + "grad_norm": 0.2970874011516571, + "learning_rate": 5.893985127516866e-07, + "loss": 0.4654, + "step": 1554 + }, + { + "epoch": 0.8599078341013825, + "grad_norm": 0.3088582158088684, + "learning_rate": 5.848592083182348e-07, + "loss": 0.4444, + "step": 1555 + }, + { + "epoch": 0.8604608294930876, + "grad_norm": 0.3454407751560211, + "learning_rate": 5.803363653717187e-07, + "loss": 0.4796, + "step": 1556 + }, + { + "epoch": 0.8610138248847926, + "grad_norm": 0.34124165773391724, + "learning_rate": 5.758300007751832e-07, + "loss": 0.475, + "step": 1557 + }, + { + "epoch": 0.8615668202764977, + "grad_norm": 0.31380608677864075, + "learning_rate": 5.713401313302358e-07, + "loss": 0.4714, + "step": 1558 + }, + { + "epoch": 0.8621198156682027, + "grad_norm": 0.3094062805175781, + "learning_rate": 5.668667737769834e-07, + "loss": 0.4348, + "step": 1559 + }, + { + "epoch": 0.8626728110599078, + "grad_norm": 0.3069341778755188, + "learning_rate": 5.624099447939696e-07, + "loss": 0.4677, + "step": 1560 + }, + { + "epoch": 0.863225806451613, + "grad_norm": 0.34499165415763855, + "learning_rate": 5.579696609981117e-07, + "loss": 0.4673, + "step": 1561 + }, + { + "epoch": 0.863778801843318, + "grad_norm": 0.32358747720718384, + "learning_rate": 5.535459389446401e-07, + "loss": 0.4627, + "step": 1562 + }, + { + "epoch": 0.8643317972350231, + "grad_norm": 0.33145248889923096, + "learning_rate": 5.491387951270366e-07, + "loss": 0.4535, + "step": 1563 + }, + { + "epoch": 0.8648847926267281, + "grad_norm": 0.3262435495853424, + "learning_rate": 5.447482459769709e-07, + "loss": 0.4613, + "step": 1564 + }, + { + "epoch": 0.8654377880184332, + "grad_norm": 0.31697559356689453, + "learning_rate": 5.403743078642443e-07, + "loss": 0.4503, + "step": 1565 + }, + { + "epoch": 0.8659907834101382, + "grad_norm": 0.3151688873767853, + "learning_rate": 5.360169970967221e-07, + "loss": 0.4816, + "step": 1566 + }, + { + "epoch": 0.8665437788018433, + "grad_norm": 0.3067736029624939, + "learning_rate": 5.316763299202766e-07, + "loss": 0.4631, + "step": 1567 + }, + { + "epoch": 0.8670967741935484, + "grad_norm": 0.2979617714881897, + "learning_rate": 5.273523225187255e-07, + "loss": 0.4403, + "step": 1568 + }, + { + "epoch": 0.8676497695852534, + "grad_norm": 0.3092879056930542, + "learning_rate": 5.23044991013773e-07, + "loss": 0.4669, + "step": 1569 + }, + { + "epoch": 0.8682027649769585, + "grad_norm": 0.29821327328681946, + "learning_rate": 5.187543514649479e-07, + "loss": 0.4678, + "step": 1570 + }, + { + "epoch": 0.8687557603686636, + "grad_norm": 0.3069091737270355, + "learning_rate": 5.144804198695447e-07, + "loss": 0.4834, + "step": 1571 + }, + { + "epoch": 0.8693087557603687, + "grad_norm": 0.3062274754047394, + "learning_rate": 5.102232121625633e-07, + "loss": 0.4583, + "step": 1572 + }, + { + "epoch": 0.8698617511520738, + "grad_norm": 0.3618842661380768, + "learning_rate": 5.059827442166504e-07, + "loss": 0.4348, + "step": 1573 + }, + { + "epoch": 0.8704147465437788, + "grad_norm": 0.3038586676120758, + "learning_rate": 5.017590318420395e-07, + "loss": 0.4601, + "step": 1574 + }, + { + "epoch": 0.8709677419354839, + "grad_norm": 0.30661341547966003, + "learning_rate": 4.975520907864928e-07, + "loss": 0.4432, + "step": 1575 + }, + { + "epoch": 0.8715207373271889, + "grad_norm": 0.3259396255016327, + "learning_rate": 4.93361936735241e-07, + "loss": 0.4602, + "step": 1576 + }, + { + "epoch": 0.872073732718894, + "grad_norm": 0.32900482416152954, + "learning_rate": 4.891885853109279e-07, + "loss": 0.473, + "step": 1577 + }, + { + "epoch": 0.872626728110599, + "grad_norm": 0.280129075050354, + "learning_rate": 4.85032052073548e-07, + "loss": 0.4581, + "step": 1578 + }, + { + "epoch": 0.8731797235023041, + "grad_norm": 0.31960612535476685, + "learning_rate": 4.808923525203912e-07, + "loss": 0.4472, + "step": 1579 + }, + { + "epoch": 0.8737327188940092, + "grad_norm": 0.28361162543296814, + "learning_rate": 4.767695020859847e-07, + "loss": 0.4538, + "step": 1580 + }, + { + "epoch": 0.8742857142857143, + "grad_norm": 0.3138439953327179, + "learning_rate": 4.726635161420351e-07, + "loss": 0.4746, + "step": 1581 + }, + { + "epoch": 0.8748387096774194, + "grad_norm": 0.30272024869918823, + "learning_rate": 4.685744099973716e-07, + "loss": 0.4713, + "step": 1582 + }, + { + "epoch": 0.8753917050691244, + "grad_norm": 0.31137874722480774, + "learning_rate": 4.6450219889788816e-07, + "loss": 0.4569, + "step": 1583 + }, + { + "epoch": 0.8759447004608295, + "grad_norm": 0.34966525435447693, + "learning_rate": 4.6044689802648534e-07, + "loss": 0.4646, + "step": 1584 + }, + { + "epoch": 0.8764976958525346, + "grad_norm": 0.29577386379241943, + "learning_rate": 4.564085225030174e-07, + "loss": 0.4642, + "step": 1585 + }, + { + "epoch": 0.8770506912442396, + "grad_norm": 0.2946975529193878, + "learning_rate": 4.523870873842329e-07, + "loss": 0.4312, + "step": 1586 + }, + { + "epoch": 0.8776036866359447, + "grad_norm": 0.30097469687461853, + "learning_rate": 4.4838260766372044e-07, + "loss": 0.4666, + "step": 1587 + }, + { + "epoch": 0.8781566820276497, + "grad_norm": 0.2847994565963745, + "learning_rate": 4.4439509827185034e-07, + "loss": 0.4165, + "step": 1588 + }, + { + "epoch": 0.8787096774193548, + "grad_norm": 0.2944558560848236, + "learning_rate": 4.4042457407572235e-07, + "loss": 0.4721, + "step": 1589 + }, + { + "epoch": 0.8792626728110599, + "grad_norm": 0.3086279034614563, + "learning_rate": 4.3647104987910636e-07, + "loss": 0.45, + "step": 1590 + }, + { + "epoch": 0.879815668202765, + "grad_norm": 0.3032161593437195, + "learning_rate": 4.3253454042239016e-07, + "loss": 0.458, + "step": 1591 + }, + { + "epoch": 0.8803686635944701, + "grad_norm": 0.3120705485343933, + "learning_rate": 4.286150603825234e-07, + "loss": 0.474, + "step": 1592 + }, + { + "epoch": 0.8809216589861751, + "grad_norm": 0.332922101020813, + "learning_rate": 4.2471262437296326e-07, + "loss": 0.4702, + "step": 1593 + }, + { + "epoch": 0.8814746543778802, + "grad_norm": 0.2810650169849396, + "learning_rate": 4.208272469436192e-07, + "loss": 0.4509, + "step": 1594 + }, + { + "epoch": 0.8820276497695853, + "grad_norm": 0.31393900513648987, + "learning_rate": 4.1695894258079903e-07, + "loss": 0.4734, + "step": 1595 + }, + { + "epoch": 0.8825806451612903, + "grad_norm": 0.2897551357746124, + "learning_rate": 4.1310772570715586e-07, + "loss": 0.4442, + "step": 1596 + }, + { + "epoch": 0.8831336405529954, + "grad_norm": 0.3010689616203308, + "learning_rate": 4.092736106816314e-07, + "loss": 0.4477, + "step": 1597 + }, + { + "epoch": 0.8836866359447004, + "grad_norm": 0.29473552107810974, + "learning_rate": 4.0545661179940854e-07, + "loss": 0.4595, + "step": 1598 + }, + { + "epoch": 0.8842396313364055, + "grad_norm": 0.3020123839378357, + "learning_rate": 4.0165674329185055e-07, + "loss": 0.461, + "step": 1599 + }, + { + "epoch": 0.8847926267281107, + "grad_norm": 0.3016417920589447, + "learning_rate": 3.978740193264524e-07, + "loss": 0.4802, + "step": 1600 + }, + { + "epoch": 0.8853456221198157, + "grad_norm": 0.3054351806640625, + "learning_rate": 3.941084540067874e-07, + "loss": 0.461, + "step": 1601 + }, + { + "epoch": 0.8858986175115208, + "grad_norm": 0.31593799591064453, + "learning_rate": 3.9036006137245396e-07, + "loss": 0.485, + "step": 1602 + }, + { + "epoch": 0.8864516129032258, + "grad_norm": 0.2995027005672455, + "learning_rate": 3.866288553990233e-07, + "loss": 0.4544, + "step": 1603 + }, + { + "epoch": 0.8870046082949309, + "grad_norm": 0.3163018524646759, + "learning_rate": 3.82914849997989e-07, + "loss": 0.4675, + "step": 1604 + }, + { + "epoch": 0.8875576036866359, + "grad_norm": 0.3466692268848419, + "learning_rate": 3.7921805901671273e-07, + "loss": 0.455, + "step": 1605 + }, + { + "epoch": 0.888110599078341, + "grad_norm": 0.30836209654808044, + "learning_rate": 3.7553849623837237e-07, + "loss": 0.4983, + "step": 1606 + }, + { + "epoch": 0.8886635944700461, + "grad_norm": 0.3018787205219269, + "learning_rate": 3.7187617538191446e-07, + "loss": 0.4714, + "step": 1607 + }, + { + "epoch": 0.8892165898617511, + "grad_norm": 0.3216530382633209, + "learning_rate": 3.682311101019981e-07, + "loss": 0.4857, + "step": 1608 + }, + { + "epoch": 0.8897695852534562, + "grad_norm": 0.32293516397476196, + "learning_rate": 3.646033139889482e-07, + "loss": 0.4508, + "step": 1609 + }, + { + "epoch": 0.8903225806451613, + "grad_norm": 0.29628053307533264, + "learning_rate": 3.6099280056870136e-07, + "loss": 0.4271, + "step": 1610 + }, + { + "epoch": 0.8908755760368664, + "grad_norm": 0.30662253499031067, + "learning_rate": 3.5739958330275936e-07, + "loss": 0.4833, + "step": 1611 + }, + { + "epoch": 0.8914285714285715, + "grad_norm": 0.30339956283569336, + "learning_rate": 3.538236755881341e-07, + "loss": 0.4675, + "step": 1612 + }, + { + "epoch": 0.8919815668202765, + "grad_norm": 0.29642388224601746, + "learning_rate": 3.502650907573024e-07, + "loss": 0.4529, + "step": 1613 + }, + { + "epoch": 0.8925345622119816, + "grad_norm": 0.3087359666824341, + "learning_rate": 3.467238420781527e-07, + "loss": 0.4794, + "step": 1614 + }, + { + "epoch": 0.8930875576036866, + "grad_norm": 0.34186214208602905, + "learning_rate": 3.431999427539368e-07, + "loss": 0.4951, + "step": 1615 + }, + { + "epoch": 0.8936405529953917, + "grad_norm": 0.31149131059646606, + "learning_rate": 3.396934059232226e-07, + "loss": 0.4476, + "step": 1616 + }, + { + "epoch": 0.8941935483870967, + "grad_norm": 0.3173343241214752, + "learning_rate": 3.362042446598418e-07, + "loss": 0.4735, + "step": 1617 + }, + { + "epoch": 0.8947465437788018, + "grad_norm": 0.29687947034835815, + "learning_rate": 3.3273247197284366e-07, + "loss": 0.4588, + "step": 1618 + }, + { + "epoch": 0.8952995391705069, + "grad_norm": 0.29158180952072144, + "learning_rate": 3.292781008064455e-07, + "loss": 0.4477, + "step": 1619 + }, + { + "epoch": 0.895852534562212, + "grad_norm": 0.31376034021377563, + "learning_rate": 3.258411440399839e-07, + "loss": 0.4613, + "step": 1620 + }, + { + "epoch": 0.8964055299539171, + "grad_norm": 0.3553878366947174, + "learning_rate": 3.2242161448786724e-07, + "loss": 0.4638, + "step": 1621 + }, + { + "epoch": 0.8969585253456221, + "grad_norm": 0.31276994943618774, + "learning_rate": 3.1901952489952927e-07, + "loss": 0.4297, + "step": 1622 + }, + { + "epoch": 0.8975115207373272, + "grad_norm": 0.2925131320953369, + "learning_rate": 3.15634887959379e-07, + "loss": 0.4642, + "step": 1623 + }, + { + "epoch": 0.8980645161290323, + "grad_norm": 0.3111323416233063, + "learning_rate": 3.122677162867549e-07, + "loss": 0.4628, + "step": 1624 + }, + { + "epoch": 0.8986175115207373, + "grad_norm": 0.28449130058288574, + "learning_rate": 3.089180224358773e-07, + "loss": 0.4619, + "step": 1625 + }, + { + "epoch": 0.8991705069124424, + "grad_norm": 0.2895408868789673, + "learning_rate": 3.0558581889580383e-07, + "loss": 0.4486, + "step": 1626 + }, + { + "epoch": 0.8997235023041474, + "grad_norm": 0.28121551871299744, + "learning_rate": 3.0227111809037815e-07, + "loss": 0.4523, + "step": 1627 + }, + { + "epoch": 0.9002764976958525, + "grad_norm": 0.3156479001045227, + "learning_rate": 2.9897393237818873e-07, + "loss": 0.4504, + "step": 1628 + }, + { + "epoch": 0.9008294930875576, + "grad_norm": 0.30593904852867126, + "learning_rate": 2.956942740525187e-07, + "loss": 0.4832, + "step": 1629 + }, + { + "epoch": 0.9013824884792627, + "grad_norm": 0.2923428416252136, + "learning_rate": 2.924321553413029e-07, + "loss": 0.4473, + "step": 1630 + }, + { + "epoch": 0.9019354838709678, + "grad_norm": 0.29185551404953003, + "learning_rate": 2.891875884070816e-07, + "loss": 0.4514, + "step": 1631 + }, + { + "epoch": 0.9024884792626728, + "grad_norm": 0.3090588450431824, + "learning_rate": 2.859605853469533e-07, + "loss": 0.492, + "step": 1632 + }, + { + "epoch": 0.9030414746543779, + "grad_norm": 0.2918623387813568, + "learning_rate": 2.8275115819253184e-07, + "loss": 0.4378, + "step": 1633 + }, + { + "epoch": 0.903594470046083, + "grad_norm": 0.2900620698928833, + "learning_rate": 2.795593189099016e-07, + "loss": 0.4702, + "step": 1634 + }, + { + "epoch": 0.904147465437788, + "grad_norm": 0.30443379282951355, + "learning_rate": 2.7638507939956926e-07, + "loss": 0.4562, + "step": 1635 + }, + { + "epoch": 0.9047004608294931, + "grad_norm": 0.2894691526889801, + "learning_rate": 2.7322845149642564e-07, + "loss": 0.4405, + "step": 1636 + }, + { + "epoch": 0.9052534562211981, + "grad_norm": 0.30414119362831116, + "learning_rate": 2.700894469696963e-07, + "loss": 0.4618, + "step": 1637 + }, + { + "epoch": 0.9058064516129032, + "grad_norm": 0.29795682430267334, + "learning_rate": 2.669680775228989e-07, + "loss": 0.4717, + "step": 1638 + }, + { + "epoch": 0.9063594470046082, + "grad_norm": 0.2925158739089966, + "learning_rate": 2.638643547938019e-07, + "loss": 0.4788, + "step": 1639 + }, + { + "epoch": 0.9069124423963134, + "grad_norm": 0.29635146260261536, + "learning_rate": 2.607782903543782e-07, + "loss": 0.4521, + "step": 1640 + }, + { + "epoch": 0.9074654377880185, + "grad_norm": 0.29401543736457825, + "learning_rate": 2.5770989571076375e-07, + "loss": 0.4401, + "step": 1641 + }, + { + "epoch": 0.9080184331797235, + "grad_norm": 0.2778622508049011, + "learning_rate": 2.5465918230321464e-07, + "loss": 0.4522, + "step": 1642 + }, + { + "epoch": 0.9085714285714286, + "grad_norm": 0.2870117127895355, + "learning_rate": 2.5162616150606167e-07, + "loss": 0.4686, + "step": 1643 + }, + { + "epoch": 0.9091244239631336, + "grad_norm": 0.3375545144081116, + "learning_rate": 2.486108446276725e-07, + "loss": 0.4741, + "step": 1644 + }, + { + "epoch": 0.9096774193548387, + "grad_norm": 0.285396546125412, + "learning_rate": 2.4561324291040636e-07, + "loss": 0.4749, + "step": 1645 + }, + { + "epoch": 0.9102304147465438, + "grad_norm": 0.31609663367271423, + "learning_rate": 2.4263336753057364e-07, + "loss": 0.4552, + "step": 1646 + }, + { + "epoch": 0.9107834101382488, + "grad_norm": 0.32548245787620544, + "learning_rate": 2.396712295983922e-07, + "loss": 0.4707, + "step": 1647 + }, + { + "epoch": 0.9113364055299539, + "grad_norm": 0.28127720952033997, + "learning_rate": 2.3672684015794923e-07, + "loss": 0.4808, + "step": 1648 + }, + { + "epoch": 0.9118894009216589, + "grad_norm": 0.3091859817504883, + "learning_rate": 2.338002101871556e-07, + "loss": 0.4469, + "step": 1649 + }, + { + "epoch": 0.9124423963133641, + "grad_norm": 0.3138236403465271, + "learning_rate": 2.3089135059771007e-07, + "loss": 0.4848, + "step": 1650 + }, + { + "epoch": 0.9129953917050692, + "grad_norm": 0.30295220017433167, + "learning_rate": 2.2800027223505438e-07, + "loss": 0.4567, + "step": 1651 + }, + { + "epoch": 0.9135483870967742, + "grad_norm": 0.3260992467403412, + "learning_rate": 2.2512698587833502e-07, + "loss": 0.4776, + "step": 1652 + }, + { + "epoch": 0.9141013824884793, + "grad_norm": 0.30410924553871155, + "learning_rate": 2.222715022403621e-07, + "loss": 0.4565, + "step": 1653 + }, + { + "epoch": 0.9146543778801843, + "grad_norm": 0.296068400144577, + "learning_rate": 2.1943383196756984e-07, + "loss": 0.4868, + "step": 1654 + }, + { + "epoch": 0.9152073732718894, + "grad_norm": 0.3140951991081238, + "learning_rate": 2.1661398563997737e-07, + "loss": 0.4645, + "step": 1655 + }, + { + "epoch": 0.9157603686635944, + "grad_norm": 0.2961066663265228, + "learning_rate": 2.138119737711475e-07, + "loss": 0.4376, + "step": 1656 + }, + { + "epoch": 0.9163133640552995, + "grad_norm": 0.3166109621524811, + "learning_rate": 2.110278068081506e-07, + "loss": 0.4713, + "step": 1657 + }, + { + "epoch": 0.9168663594470046, + "grad_norm": 0.32516512274742126, + "learning_rate": 2.082614951315215e-07, + "loss": 0.4516, + "step": 1658 + }, + { + "epoch": 0.9174193548387096, + "grad_norm": 0.293089896440506, + "learning_rate": 2.055130490552254e-07, + "loss": 0.469, + "step": 1659 + }, + { + "epoch": 0.9179723502304148, + "grad_norm": 0.31795766949653625, + "learning_rate": 2.0278247882661584e-07, + "loss": 0.4497, + "step": 1660 + }, + { + "epoch": 0.9185253456221198, + "grad_norm": 0.2847367525100708, + "learning_rate": 2.0006979462639686e-07, + "loss": 0.4447, + "step": 1661 + }, + { + "epoch": 0.9190783410138249, + "grad_norm": 0.2693861126899719, + "learning_rate": 1.9737500656858754e-07, + "loss": 0.4406, + "step": 1662 + }, + { + "epoch": 0.91963133640553, + "grad_norm": 0.3001803755760193, + "learning_rate": 1.94698124700482e-07, + "loss": 0.4964, + "step": 1663 + }, + { + "epoch": 0.920184331797235, + "grad_norm": 0.3073177933692932, + "learning_rate": 1.9203915900261327e-07, + "loss": 0.4737, + "step": 1664 + }, + { + "epoch": 0.9207373271889401, + "grad_norm": 0.31089356541633606, + "learning_rate": 1.8939811938871456e-07, + "loss": 0.4539, + "step": 1665 + }, + { + "epoch": 0.9212903225806451, + "grad_norm": 0.27542582154273987, + "learning_rate": 1.86775015705683e-07, + "loss": 0.434, + "step": 1666 + }, + { + "epoch": 0.9218433179723502, + "grad_norm": 0.32208341360092163, + "learning_rate": 1.841698577335438e-07, + "loss": 0.4476, + "step": 1667 + }, + { + "epoch": 0.9223963133640553, + "grad_norm": 0.3119848668575287, + "learning_rate": 1.8158265518541274e-07, + "loss": 0.4484, + "step": 1668 + }, + { + "epoch": 0.9229493087557604, + "grad_norm": 0.2974831759929657, + "learning_rate": 1.790134177074604e-07, + "loss": 0.4574, + "step": 1669 + }, + { + "epoch": 0.9235023041474655, + "grad_norm": 0.29929405450820923, + "learning_rate": 1.7646215487887587e-07, + "loss": 0.4685, + "step": 1670 + }, + { + "epoch": 0.9240552995391705, + "grad_norm": 0.2898726165294647, + "learning_rate": 1.739288762118313e-07, + "loss": 0.455, + "step": 1671 + }, + { + "epoch": 0.9246082949308756, + "grad_norm": 0.3034640848636627, + "learning_rate": 1.714135911514475e-07, + "loss": 0.4541, + "step": 1672 + }, + { + "epoch": 0.9251612903225807, + "grad_norm": 0.3259904682636261, + "learning_rate": 1.6891630907575562e-07, + "loss": 0.4678, + "step": 1673 + }, + { + "epoch": 0.9257142857142857, + "grad_norm": 0.3421642780303955, + "learning_rate": 1.664370392956649e-07, + "loss": 0.4518, + "step": 1674 + }, + { + "epoch": 0.9262672811059908, + "grad_norm": 0.29688096046447754, + "learning_rate": 1.6397579105492778e-07, + "loss": 0.4222, + "step": 1675 + }, + { + "epoch": 0.9268202764976958, + "grad_norm": 0.3092136085033417, + "learning_rate": 1.6153257353010433e-07, + "loss": 0.4679, + "step": 1676 + }, + { + "epoch": 0.9273732718894009, + "grad_norm": 0.3253113627433777, + "learning_rate": 1.5910739583053002e-07, + "loss": 0.4495, + "step": 1677 + }, + { + "epoch": 0.9279262672811059, + "grad_norm": 0.286625474691391, + "learning_rate": 1.5670026699827757e-07, + "loss": 0.4652, + "step": 1678 + }, + { + "epoch": 0.9284792626728111, + "grad_norm": 0.2905667722225189, + "learning_rate": 1.5431119600812837e-07, + "loss": 0.4529, + "step": 1679 + }, + { + "epoch": 0.9290322580645162, + "grad_norm": 0.32414594292640686, + "learning_rate": 1.5194019176753615e-07, + "loss": 0.4878, + "step": 1680 + }, + { + "epoch": 0.9295852534562212, + "grad_norm": 0.29310277104377747, + "learning_rate": 1.495872631165929e-07, + "loss": 0.476, + "step": 1681 + }, + { + "epoch": 0.9301382488479263, + "grad_norm": 0.3012104332447052, + "learning_rate": 1.4725241882799946e-07, + "loss": 0.4823, + "step": 1682 + }, + { + "epoch": 0.9306912442396313, + "grad_norm": 0.2896125018596649, + "learning_rate": 1.449356676070285e-07, + "loss": 0.4429, + "step": 1683 + }, + { + "epoch": 0.9312442396313364, + "grad_norm": 0.2954270839691162, + "learning_rate": 1.42637018091496e-07, + "loss": 0.473, + "step": 1684 + }, + { + "epoch": 0.9317972350230415, + "grad_norm": 0.29838648438453674, + "learning_rate": 1.4035647885172533e-07, + "loss": 0.4615, + "step": 1685 + }, + { + "epoch": 0.9323502304147465, + "grad_norm": 0.3438728451728821, + "learning_rate": 1.3809405839051937e-07, + "loss": 0.4924, + "step": 1686 + }, + { + "epoch": 0.9329032258064516, + "grad_norm": 0.2971719205379486, + "learning_rate": 1.358497651431251e-07, + "loss": 0.4438, + "step": 1687 + }, + { + "epoch": 0.9334562211981566, + "grad_norm": 0.3192031979560852, + "learning_rate": 1.3362360747720405e-07, + "loss": 0.4559, + "step": 1688 + }, + { + "epoch": 0.9340092165898618, + "grad_norm": 0.3306371867656708, + "learning_rate": 1.314155936928002e-07, + "loss": 0.4633, + "step": 1689 + }, + { + "epoch": 0.9345622119815669, + "grad_norm": 0.28587308526039124, + "learning_rate": 1.2922573202231114e-07, + "loss": 0.4577, + "step": 1690 + }, + { + "epoch": 0.9351152073732719, + "grad_norm": 0.31524068117141724, + "learning_rate": 1.2705403063045464e-07, + "loss": 0.4457, + "step": 1691 + }, + { + "epoch": 0.935668202764977, + "grad_norm": 0.2901270389556885, + "learning_rate": 1.2490049761423927e-07, + "loss": 0.4396, + "step": 1692 + }, + { + "epoch": 0.936221198156682, + "grad_norm": 0.2995949685573578, + "learning_rate": 1.2276514100293403e-07, + "loss": 0.4633, + "step": 1693 + }, + { + "epoch": 0.9367741935483871, + "grad_norm": 0.2743770182132721, + "learning_rate": 1.206479687580403e-07, + "loss": 0.4655, + "step": 1694 + }, + { + "epoch": 0.9373271889400921, + "grad_norm": 0.3132480978965759, + "learning_rate": 1.1854898877325882e-07, + "loss": 0.4641, + "step": 1695 + }, + { + "epoch": 0.9378801843317972, + "grad_norm": 0.2994924783706665, + "learning_rate": 1.1646820887446342e-07, + "loss": 0.4491, + "step": 1696 + }, + { + "epoch": 0.9384331797235023, + "grad_norm": 0.2822410762310028, + "learning_rate": 1.1440563681966943e-07, + "loss": 0.4663, + "step": 1697 + }, + { + "epoch": 0.9389861751152073, + "grad_norm": 0.3047800660133362, + "learning_rate": 1.1236128029900539e-07, + "loss": 0.4638, + "step": 1698 + }, + { + "epoch": 0.9395391705069125, + "grad_norm": 0.31529849767684937, + "learning_rate": 1.1033514693468584e-07, + "loss": 0.4269, + "step": 1699 + }, + { + "epoch": 0.9400921658986175, + "grad_norm": 0.28443610668182373, + "learning_rate": 1.0832724428098185e-07, + "loss": 0.4568, + "step": 1700 + }, + { + "epoch": 0.9406451612903226, + "grad_norm": 0.2651902735233307, + "learning_rate": 1.0633757982419169e-07, + "loss": 0.4508, + "step": 1701 + }, + { + "epoch": 0.9411981566820277, + "grad_norm": 0.29510411620140076, + "learning_rate": 1.0436616098261409e-07, + "loss": 0.4643, + "step": 1702 + }, + { + "epoch": 0.9417511520737327, + "grad_norm": 0.2974132299423218, + "learning_rate": 1.0241299510652114e-07, + "loss": 0.429, + "step": 1703 + }, + { + "epoch": 0.9423041474654378, + "grad_norm": 0.2929219901561737, + "learning_rate": 1.0047808947813097e-07, + "loss": 0.4851, + "step": 1704 + }, + { + "epoch": 0.9428571428571428, + "grad_norm": 0.29706960916519165, + "learning_rate": 9.85614513115779e-08, + "loss": 0.4559, + "step": 1705 + }, + { + "epoch": 0.9434101382488479, + "grad_norm": 0.298990935087204, + "learning_rate": 9.666308775289013e-08, + "loss": 0.479, + "step": 1706 + }, + { + "epoch": 0.943963133640553, + "grad_norm": 0.29148900508880615, + "learning_rate": 9.47830058799576e-08, + "loss": 0.4577, + "step": 1707 + }, + { + "epoch": 0.944516129032258, + "grad_norm": 0.3139622211456299, + "learning_rate": 9.292121270251031e-08, + "loss": 0.4569, + "step": 1708 + }, + { + "epoch": 0.9450691244239632, + "grad_norm": 0.29921236634254456, + "learning_rate": 9.107771516209008e-08, + "loss": 0.4558, + "step": 1709 + }, + { + "epoch": 0.9456221198156682, + "grad_norm": 0.28523844480514526, + "learning_rate": 8.925252013202545e-08, + "loss": 0.4709, + "step": 1710 + }, + { + "epoch": 0.9461751152073733, + "grad_norm": 0.3282380998134613, + "learning_rate": 8.744563441740461e-08, + "loss": 0.4485, + "step": 1711 + }, + { + "epoch": 0.9467281105990784, + "grad_norm": 0.30116501450538635, + "learning_rate": 8.565706475505086e-08, + "loss": 0.4633, + "step": 1712 + }, + { + "epoch": 0.9472811059907834, + "grad_norm": 0.2834503948688507, + "learning_rate": 8.388681781349828e-08, + "loss": 0.4569, + "step": 1713 + }, + { + "epoch": 0.9478341013824885, + "grad_norm": 0.31157195568084717, + "learning_rate": 8.213490019296666e-08, + "loss": 0.4666, + "step": 1714 + }, + { + "epoch": 0.9483870967741935, + "grad_norm": 0.30053144693374634, + "learning_rate": 8.04013184253355e-08, + "loss": 0.4645, + "step": 1715 + }, + { + "epoch": 0.9489400921658986, + "grad_norm": 0.3080219328403473, + "learning_rate": 7.868607897412062e-08, + "loss": 0.4409, + "step": 1716 + }, + { + "epoch": 0.9494930875576036, + "grad_norm": 0.2975987493991852, + "learning_rate": 7.69891882344509e-08, + "loss": 0.4691, + "step": 1717 + }, + { + "epoch": 0.9500460829493087, + "grad_norm": 0.3106543719768524, + "learning_rate": 7.53106525330427e-08, + "loss": 0.4816, + "step": 1718 + }, + { + "epoch": 0.9505990783410139, + "grad_norm": 0.29965740442276, + "learning_rate": 7.36504781281766e-08, + "loss": 0.469, + "step": 1719 + }, + { + "epoch": 0.9511520737327189, + "grad_norm": 0.31763914227485657, + "learning_rate": 7.20086712096768e-08, + "loss": 0.4615, + "step": 1720 + }, + { + "epoch": 0.951705069124424, + "grad_norm": 0.2858491837978363, + "learning_rate": 7.038523789888397e-08, + "loss": 0.457, + "step": 1721 + }, + { + "epoch": 0.952258064516129, + "grad_norm": 0.2787969410419464, + "learning_rate": 6.878018424863408e-08, + "loss": 0.4485, + "step": 1722 + }, + { + "epoch": 0.9528110599078341, + "grad_norm": 0.32252055406570435, + "learning_rate": 6.7193516243238e-08, + "loss": 0.4708, + "step": 1723 + }, + { + "epoch": 0.9533640552995392, + "grad_norm": 0.3171110451221466, + "learning_rate": 6.562523979845525e-08, + "loss": 0.466, + "step": 1724 + }, + { + "epoch": 0.9539170506912442, + "grad_norm": 0.31782266497612, + "learning_rate": 6.407536076147525e-08, + "loss": 0.4624, + "step": 1725 + }, + { + "epoch": 0.9544700460829493, + "grad_norm": 0.2952912449836731, + "learning_rate": 6.254388491089281e-08, + "loss": 0.4537, + "step": 1726 + }, + { + "epoch": 0.9550230414746543, + "grad_norm": 0.28047212958335876, + "learning_rate": 6.10308179566893e-08, + "loss": 0.473, + "step": 1727 + }, + { + "epoch": 0.9555760368663595, + "grad_norm": 0.30235886573791504, + "learning_rate": 5.9536165540209914e-08, + "loss": 0.489, + "step": 1728 + }, + { + "epoch": 0.9561290322580646, + "grad_norm": 0.29142141342163086, + "learning_rate": 5.80599332341425e-08, + "loss": 0.4718, + "step": 1729 + }, + { + "epoch": 0.9566820276497696, + "grad_norm": 0.2832047939300537, + "learning_rate": 5.6602126542496525e-08, + "loss": 0.4562, + "step": 1730 + }, + { + "epoch": 0.9572350230414747, + "grad_norm": 0.2992711365222931, + "learning_rate": 5.516275090058476e-08, + "loss": 0.4816, + "step": 1731 + }, + { + "epoch": 0.9577880184331797, + "grad_norm": 0.29784855246543884, + "learning_rate": 5.3741811674998254e-08, + "loss": 0.4818, + "step": 1732 + }, + { + "epoch": 0.9583410138248848, + "grad_norm": 0.3014211058616638, + "learning_rate": 5.233931416359195e-08, + "loss": 0.4914, + "step": 1733 + }, + { + "epoch": 0.9588940092165898, + "grad_norm": 0.30964842438697815, + "learning_rate": 5.095526359546243e-08, + "loss": 0.4643, + "step": 1734 + }, + { + "epoch": 0.9594470046082949, + "grad_norm": 0.3023122251033783, + "learning_rate": 4.958966513092689e-08, + "loss": 0.4499, + "step": 1735 + }, + { + "epoch": 0.96, + "grad_norm": 0.295585572719574, + "learning_rate": 4.8242523861506405e-08, + "loss": 0.4248, + "step": 1736 + }, + { + "epoch": 0.960552995391705, + "grad_norm": 0.2833447754383087, + "learning_rate": 4.691384480990602e-08, + "loss": 0.4491, + "step": 1737 + }, + { + "epoch": 0.9611059907834102, + "grad_norm": 0.309023916721344, + "learning_rate": 4.56036329299947e-08, + "loss": 0.4404, + "step": 1738 + }, + { + "epoch": 0.9616589861751152, + "grad_norm": 0.3236503303050995, + "learning_rate": 4.4311893106789847e-08, + "loss": 0.4591, + "step": 1739 + }, + { + "epoch": 0.9622119815668203, + "grad_norm": 0.29482167959213257, + "learning_rate": 4.3038630156436166e-08, + "loss": 0.4558, + "step": 1740 + }, + { + "epoch": 0.9627649769585254, + "grad_norm": 0.2854726016521454, + "learning_rate": 4.1783848826189025e-08, + "loss": 0.4561, + "step": 1741 + }, + { + "epoch": 0.9633179723502304, + "grad_norm": 0.28575026988983154, + "learning_rate": 4.0547553794397235e-08, + "loss": 0.4541, + "step": 1742 + }, + { + "epoch": 0.9638709677419355, + "grad_norm": 0.30080512166023254, + "learning_rate": 3.932974967048364e-08, + "loss": 0.4475, + "step": 1743 + }, + { + "epoch": 0.9644239631336405, + "grad_norm": 0.2857537269592285, + "learning_rate": 3.813044099493068e-08, + "loss": 0.4521, + "step": 1744 + }, + { + "epoch": 0.9649769585253456, + "grad_norm": 0.28055858612060547, + "learning_rate": 3.6949632239261514e-08, + "loss": 0.4388, + "step": 1745 + }, + { + "epoch": 0.9655299539170507, + "grad_norm": 0.27776503562927246, + "learning_rate": 3.578732780602334e-08, + "loss": 0.4483, + "step": 1746 + }, + { + "epoch": 0.9660829493087557, + "grad_norm": 0.322313517332077, + "learning_rate": 3.464353202877302e-08, + "loss": 0.4589, + "step": 1747 + }, + { + "epoch": 0.9666359447004609, + "grad_norm": 0.315902978181839, + "learning_rate": 3.351824917205704e-08, + "loss": 0.4696, + "step": 1748 + }, + { + "epoch": 0.9671889400921659, + "grad_norm": 0.3116399645805359, + "learning_rate": 3.2411483431400435e-08, + "loss": 0.4532, + "step": 1749 + }, + { + "epoch": 0.967741935483871, + "grad_norm": 0.27709704637527466, + "learning_rate": 3.1323238933286814e-08, + "loss": 0.4808, + "step": 1750 + }, + { + "epoch": 0.968294930875576, + "grad_norm": 0.28587040305137634, + "learning_rate": 3.0253519735146674e-08, + "loss": 0.4498, + "step": 1751 + }, + { + "epoch": 0.9688479262672811, + "grad_norm": 0.3236626088619232, + "learning_rate": 2.920232982533855e-08, + "loss": 0.4618, + "step": 1752 + }, + { + "epoch": 0.9694009216589862, + "grad_norm": 0.29707276821136475, + "learning_rate": 2.8169673123137896e-08, + "loss": 0.4516, + "step": 1753 + }, + { + "epoch": 0.9699539170506912, + "grad_norm": 0.3110201954841614, + "learning_rate": 2.7155553478719343e-08, + "loss": 0.4534, + "step": 1754 + }, + { + "epoch": 0.9705069124423963, + "grad_norm": 0.308906227350235, + "learning_rate": 2.6159974673143352e-08, + "loss": 0.447, + "step": 1755 + }, + { + "epoch": 0.9710599078341013, + "grad_norm": 0.30986642837524414, + "learning_rate": 2.5182940418344016e-08, + "loss": 0.4587, + "step": 1756 + }, + { + "epoch": 0.9716129032258064, + "grad_norm": 0.31161195039749146, + "learning_rate": 2.4224454357112404e-08, + "loss": 0.4616, + "step": 1757 + }, + { + "epoch": 0.9721658986175116, + "grad_norm": 0.2983948290348053, + "learning_rate": 2.328452006308435e-08, + "loss": 0.4696, + "step": 1758 + }, + { + "epoch": 0.9727188940092166, + "grad_norm": 0.28888341784477234, + "learning_rate": 2.2363141040727123e-08, + "loss": 0.4859, + "step": 1759 + }, + { + "epoch": 0.9732718894009217, + "grad_norm": 0.2775828242301941, + "learning_rate": 2.1460320725326113e-08, + "loss": 0.4354, + "step": 1760 + }, + { + "epoch": 0.9738248847926267, + "grad_norm": 0.34634268283843994, + "learning_rate": 2.057606248297206e-08, + "loss": 0.4431, + "step": 1761 + }, + { + "epoch": 0.9743778801843318, + "grad_norm": 0.31715697050094604, + "learning_rate": 1.971036961054884e-08, + "loss": 0.4566, + "step": 1762 + }, + { + "epoch": 0.9749308755760369, + "grad_norm": 0.2788535952568054, + "learning_rate": 1.886324533572015e-08, + "loss": 0.4492, + "step": 1763 + }, + { + "epoch": 0.9754838709677419, + "grad_norm": 0.3046756684780121, + "learning_rate": 1.8034692816919497e-08, + "loss": 0.4461, + "step": 1764 + }, + { + "epoch": 0.976036866359447, + "grad_norm": 0.2936743497848511, + "learning_rate": 1.7224715143335235e-08, + "loss": 0.4635, + "step": 1765 + }, + { + "epoch": 0.976589861751152, + "grad_norm": 0.31344956159591675, + "learning_rate": 1.643331533490333e-08, + "loss": 0.434, + "step": 1766 + }, + { + "epoch": 0.9771428571428571, + "grad_norm": 0.2932872176170349, + "learning_rate": 1.5660496342291833e-08, + "loss": 0.4707, + "step": 1767 + }, + { + "epoch": 0.9776958525345623, + "grad_norm": 0.29183509945869446, + "learning_rate": 1.4906261046892523e-08, + "loss": 0.4711, + "step": 1768 + }, + { + "epoch": 0.9782488479262673, + "grad_norm": 0.29039350152015686, + "learning_rate": 1.4170612260808736e-08, + "loss": 0.4802, + "step": 1769 + }, + { + "epoch": 0.9788018433179724, + "grad_norm": 0.30380529165267944, + "learning_rate": 1.3453552726847008e-08, + "loss": 0.4868, + "step": 1770 + }, + { + "epoch": 0.9793548387096774, + "grad_norm": 0.29868564009666443, + "learning_rate": 1.2755085118503762e-08, + "loss": 0.4625, + "step": 1771 + }, + { + "epoch": 0.9799078341013825, + "grad_norm": 0.30664679408073425, + "learning_rate": 1.207521203995754e-08, + "loss": 0.4369, + "step": 1772 + }, + { + "epoch": 0.9804608294930875, + "grad_norm": 0.28604456782341003, + "learning_rate": 1.1413936026059558e-08, + "loss": 0.4559, + "step": 1773 + }, + { + "epoch": 0.9810138248847926, + "grad_norm": 0.33186855912208557, + "learning_rate": 1.0771259542322055e-08, + "loss": 0.4735, + "step": 1774 + }, + { + "epoch": 0.9815668202764977, + "grad_norm": 0.3165326714515686, + "learning_rate": 1.0147184984911073e-08, + "loss": 0.4411, + "step": 1775 + }, + { + "epoch": 0.9821198156682027, + "grad_norm": 0.30226629972457886, + "learning_rate": 9.541714680637582e-09, + "loss": 0.4274, + "step": 1776 + }, + { + "epoch": 0.9826728110599078, + "grad_norm": 0.3025604486465454, + "learning_rate": 8.95485088694692e-09, + "loss": 0.4461, + "step": 1777 + }, + { + "epoch": 0.983225806451613, + "grad_norm": 0.3139297068119049, + "learning_rate": 8.386595791912145e-09, + "loss": 0.4501, + "step": 1778 + }, + { + "epoch": 0.983778801843318, + "grad_norm": 0.28405094146728516, + "learning_rate": 7.836951514225144e-09, + "loss": 0.4686, + "step": 1779 + }, + { + "epoch": 0.9843317972350231, + "grad_norm": 0.2899510860443115, + "learning_rate": 7.3059201031899786e-09, + "loss": 0.4738, + "step": 1780 + }, + { + "epoch": 0.9848847926267281, + "grad_norm": 0.3046450614929199, + "learning_rate": 6.7935035387128865e-09, + "loss": 0.4531, + "step": 1781 + }, + { + "epoch": 0.9854377880184332, + "grad_norm": 0.2701072692871094, + "learning_rate": 6.299703731296181e-09, + "loss": 0.4532, + "step": 1782 + }, + { + "epoch": 0.9859907834101382, + "grad_norm": 0.271697461605072, + "learning_rate": 5.8245225220321385e-09, + "loss": 0.4349, + "step": 1783 + }, + { + "epoch": 0.9865437788018433, + "grad_norm": 0.29448968172073364, + "learning_rate": 5.367961682594125e-09, + "loss": 0.4664, + "step": 1784 + }, + { + "epoch": 0.9870967741935484, + "grad_norm": 0.3071184754371643, + "learning_rate": 4.930022915231591e-09, + "loss": 0.4397, + "step": 1785 + }, + { + "epoch": 0.9876497695852534, + "grad_norm": 0.2995312213897705, + "learning_rate": 4.510707852762864e-09, + "loss": 0.4592, + "step": 1786 + }, + { + "epoch": 0.9882027649769585, + "grad_norm": 0.30661311745643616, + "learning_rate": 4.110018058570142e-09, + "loss": 0.4835, + "step": 1787 + }, + { + "epoch": 0.9887557603686636, + "grad_norm": 0.2768559157848358, + "learning_rate": 3.727955026591179e-09, + "loss": 0.4506, + "step": 1788 + }, + { + "epoch": 0.9893087557603687, + "grad_norm": 0.3139650821685791, + "learning_rate": 3.3645201813170546e-09, + "loss": 0.4483, + "step": 1789 + }, + { + "epoch": 0.9898617511520738, + "grad_norm": 0.318562775850296, + "learning_rate": 3.0197148777838524e-09, + "loss": 0.4669, + "step": 1790 + }, + { + "epoch": 0.9904147465437788, + "grad_norm": 0.31068408489227295, + "learning_rate": 2.693540401569883e-09, + "loss": 0.4519, + "step": 1791 + }, + { + "epoch": 0.9909677419354839, + "grad_norm": 0.30719509720802307, + "learning_rate": 2.3859979687901326e-09, + "loss": 0.4284, + "step": 1792 + }, + { + "epoch": 0.9915207373271889, + "grad_norm": 0.28648641705513, + "learning_rate": 2.0970887260907123e-09, + "loss": 0.4556, + "step": 1793 + }, + { + "epoch": 0.992073732718894, + "grad_norm": 0.2835429608821869, + "learning_rate": 1.8268137506455286e-09, + "loss": 0.4764, + "step": 1794 + }, + { + "epoch": 0.992626728110599, + "grad_norm": 0.3018244206905365, + "learning_rate": 1.57517405015295e-09, + "loss": 0.4505, + "step": 1795 + }, + { + "epoch": 0.9931797235023041, + "grad_norm": 0.30132514238357544, + "learning_rate": 1.3421705628302584e-09, + "loss": 0.4686, + "step": 1796 + }, + { + "epoch": 0.9937327188940093, + "grad_norm": 0.292498379945755, + "learning_rate": 1.1278041574125376e-09, + "loss": 0.4295, + "step": 1797 + }, + { + "epoch": 0.9942857142857143, + "grad_norm": 0.2868688702583313, + "learning_rate": 9.320756331465675e-10, + "loss": 0.4246, + "step": 1798 + }, + { + "epoch": 0.9948387096774194, + "grad_norm": 0.2882429361343384, + "learning_rate": 7.549857197897137e-10, + "loss": 0.4539, + "step": 1799 + }, + { + "epoch": 0.9953917050691244, + "grad_norm": 0.2899753749370575, + "learning_rate": 5.965350776071521e-10, + "loss": 0.4748, + "step": 1800 + }, + { + "epoch": 0.9959447004608295, + "grad_norm": 0.293375700712204, + "learning_rate": 4.567242973696484e-10, + "loss": 0.4613, + "step": 1801 + }, + { + "epoch": 0.9964976958525346, + "grad_norm": 0.2795793116092682, + "learning_rate": 3.3555390034967214e-10, + "loss": 0.4732, + "step": 1802 + }, + { + "epoch": 0.9970506912442396, + "grad_norm": 0.29951760172843933, + "learning_rate": 2.330243383208419e-10, + "loss": 0.4612, + "step": 1803 + }, + { + "epoch": 0.9976036866359447, + "grad_norm": 0.2993388772010803, + "learning_rate": 1.4913599355625975e-10, + "loss": 0.4653, + "step": 1804 + }, + { + "epoch": 0.9981566820276497, + "grad_norm": 0.31436586380004883, + "learning_rate": 8.388917882684589e-11, + "loss": 0.4712, + "step": 1805 + }, + { + "epoch": 0.9987096774193548, + "grad_norm": 0.2935016453266144, + "learning_rate": 3.728413739967351e-11, + "loss": 0.4381, + "step": 1806 + }, + { + "epoch": 0.99926267281106, + "grad_norm": 0.29614418745040894, + "learning_rate": 9.321043037968657e-12, + "loss": 0.4752, + "step": 1807 + }, + { + "epoch": 0.999815668202765, + "grad_norm": 0.294443815946579, + "learning_rate": 0.0, + "loss": 0.4617, + "step": 1808 + }, + { + "epoch": 0.999815668202765, + "step": 1808, + "total_flos": 2730506599792640.0, + "train_loss": 0.488344039471276, + "train_runtime": 63700.3598, + "train_samples_per_second": 2.725, + "train_steps_per_second": 0.028 + } + ], + "logging_steps": 1.0, + "max_steps": 1808, + "num_input_tokens_seen": 0, + "num_train_epochs": 1, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 2730506599792640.0, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +}