| { |
| "best_global_step": null, |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 0.9992229992229992, |
| "eval_steps": 500, |
| "global_step": 643, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.001554001554001554, |
| "grad_norm": 1.0753202631674548, |
| "learning_rate": 1e-05, |
| "loss": 0.5508, |
| "step": 1 |
| }, |
| { |
| "epoch": 0.003108003108003108, |
| "grad_norm": 1.394325863785031, |
| "learning_rate": 9.999940321631158e-06, |
| "loss": 0.6286, |
| "step": 2 |
| }, |
| { |
| "epoch": 0.004662004662004662, |
| "grad_norm": 0.8138734995659078, |
| "learning_rate": 9.999761287949237e-06, |
| "loss": 0.4343, |
| "step": 3 |
| }, |
| { |
| "epoch": 0.006216006216006216, |
| "grad_norm": 1.4834629168746152, |
| "learning_rate": 9.99946290322801e-06, |
| "loss": 0.5391, |
| "step": 4 |
| }, |
| { |
| "epoch": 0.00777000777000777, |
| "grad_norm": 0.6636994391139434, |
| "learning_rate": 9.999045174590324e-06, |
| "loss": 0.5696, |
| "step": 5 |
| }, |
| { |
| "epoch": 0.009324009324009324, |
| "grad_norm": 0.8144572025546154, |
| "learning_rate": 9.998508112007925e-06, |
| "loss": 0.8279, |
| "step": 6 |
| }, |
| { |
| "epoch": 0.010878010878010878, |
| "grad_norm": 0.6969071538976613, |
| "learning_rate": 9.997851728301219e-06, |
| "loss": 0.537, |
| "step": 7 |
| }, |
| { |
| "epoch": 0.012432012432012432, |
| "grad_norm": 0.8735249365273472, |
| "learning_rate": 9.99707603913897e-06, |
| "loss": 0.6761, |
| "step": 8 |
| }, |
| { |
| "epoch": 0.013986013986013986, |
| "grad_norm": 0.7027334655844945, |
| "learning_rate": 9.996181063037924e-06, |
| "loss": 0.5301, |
| "step": 9 |
| }, |
| { |
| "epoch": 0.01554001554001554, |
| "grad_norm": 0.8373746380293126, |
| "learning_rate": 9.995166821362368e-06, |
| "loss": 0.6674, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.017094017094017096, |
| "grad_norm": 0.984420779214608, |
| "learning_rate": 9.994033338323612e-06, |
| "loss": 0.5113, |
| "step": 11 |
| }, |
| { |
| "epoch": 0.018648018648018648, |
| "grad_norm": 0.7622743125468773, |
| "learning_rate": 9.99278064097943e-06, |
| "loss": 0.6431, |
| "step": 12 |
| }, |
| { |
| "epoch": 0.020202020202020204, |
| "grad_norm": 0.7377189725798596, |
| "learning_rate": 9.991408759233394e-06, |
| "loss": 0.6131, |
| "step": 13 |
| }, |
| { |
| "epoch": 0.021756021756021756, |
| "grad_norm": 0.9016901262329209, |
| "learning_rate": 9.989917725834166e-06, |
| "loss": 0.5208, |
| "step": 14 |
| }, |
| { |
| "epoch": 0.023310023310023312, |
| "grad_norm": 0.7060988613807518, |
| "learning_rate": 9.988307576374727e-06, |
| "loss": 0.5591, |
| "step": 15 |
| }, |
| { |
| "epoch": 0.024864024864024864, |
| "grad_norm": 1.965431280490754, |
| "learning_rate": 9.986578349291514e-06, |
| "loss": 0.6333, |
| "step": 16 |
| }, |
| { |
| "epoch": 0.02641802641802642, |
| "grad_norm": 0.6204824016696763, |
| "learning_rate": 9.984730085863504e-06, |
| "loss": 0.4729, |
| "step": 17 |
| }, |
| { |
| "epoch": 0.027972027972027972, |
| "grad_norm": 0.6629661100589302, |
| "learning_rate": 9.982762830211239e-06, |
| "loss": 0.41, |
| "step": 18 |
| }, |
| { |
| "epoch": 0.029526029526029528, |
| "grad_norm": 0.679676793160211, |
| "learning_rate": 9.980676629295763e-06, |
| "loss": 0.5619, |
| "step": 19 |
| }, |
| { |
| "epoch": 0.03108003108003108, |
| "grad_norm": 0.7228425066613519, |
| "learning_rate": 9.9784715329175e-06, |
| "loss": 0.5637, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.03263403263403263, |
| "grad_norm": 1.4494500979247202, |
| "learning_rate": 9.976147593715074e-06, |
| "loss": 0.4784, |
| "step": 21 |
| }, |
| { |
| "epoch": 0.03418803418803419, |
| "grad_norm": 0.9490720634807803, |
| "learning_rate": 9.973704867164044e-06, |
| "loss": 0.775, |
| "step": 22 |
| }, |
| { |
| "epoch": 0.035742035742035744, |
| "grad_norm": 0.7173074724185691, |
| "learning_rate": 9.971143411575585e-06, |
| "loss": 0.5456, |
| "step": 23 |
| }, |
| { |
| "epoch": 0.037296037296037296, |
| "grad_norm": 1.813950013627374, |
| "learning_rate": 9.968463288095096e-06, |
| "loss": 0.5494, |
| "step": 24 |
| }, |
| { |
| "epoch": 0.03885003885003885, |
| "grad_norm": 0.764359484896435, |
| "learning_rate": 9.965664560700734e-06, |
| "loss": 0.6135, |
| "step": 25 |
| }, |
| { |
| "epoch": 0.04040404040404041, |
| "grad_norm": 0.8961015996035883, |
| "learning_rate": 9.962747296201891e-06, |
| "loss": 0.4535, |
| "step": 26 |
| }, |
| { |
| "epoch": 0.04195804195804196, |
| "grad_norm": 0.7061951973387899, |
| "learning_rate": 9.959711564237603e-06, |
| "loss": 0.5766, |
| "step": 27 |
| }, |
| { |
| "epoch": 0.04351204351204351, |
| "grad_norm": 0.9186274326988478, |
| "learning_rate": 9.956557437274887e-06, |
| "loss": 0.4126, |
| "step": 28 |
| }, |
| { |
| "epoch": 0.045066045066045064, |
| "grad_norm": 1.0005571619677003, |
| "learning_rate": 9.953284990607e-06, |
| "loss": 0.5344, |
| "step": 29 |
| }, |
| { |
| "epoch": 0.046620046620046623, |
| "grad_norm": 1.0384488624813175, |
| "learning_rate": 9.949894302351653e-06, |
| "loss": 0.686, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.048174048174048176, |
| "grad_norm": 1.1091587053056344, |
| "learning_rate": 9.946385453449145e-06, |
| "loss": 0.7023, |
| "step": 31 |
| }, |
| { |
| "epoch": 0.04972804972804973, |
| "grad_norm": 0.8221842041481426, |
| "learning_rate": 9.942758527660429e-06, |
| "loss": 0.4769, |
| "step": 32 |
| }, |
| { |
| "epoch": 0.05128205128205128, |
| "grad_norm": 0.8497884172706046, |
| "learning_rate": 9.93901361156511e-06, |
| "loss": 0.5504, |
| "step": 33 |
| }, |
| { |
| "epoch": 0.05283605283605284, |
| "grad_norm": 0.6872331365190385, |
| "learning_rate": 9.935150794559379e-06, |
| "loss": 0.5369, |
| "step": 34 |
| }, |
| { |
| "epoch": 0.05439005439005439, |
| "grad_norm": 0.8777872780147628, |
| "learning_rate": 9.931170168853886e-06, |
| "loss": 0.657, |
| "step": 35 |
| }, |
| { |
| "epoch": 0.055944055944055944, |
| "grad_norm": 0.6487410029703777, |
| "learning_rate": 9.927071829471531e-06, |
| "loss": 0.4006, |
| "step": 36 |
| }, |
| { |
| "epoch": 0.057498057498057496, |
| "grad_norm": 0.6468395282548087, |
| "learning_rate": 9.922855874245197e-06, |
| "loss": 0.4699, |
| "step": 37 |
| }, |
| { |
| "epoch": 0.059052059052059055, |
| "grad_norm": 0.821140351961526, |
| "learning_rate": 9.918522403815414e-06, |
| "loss": 0.5927, |
| "step": 38 |
| }, |
| { |
| "epoch": 0.06060606060606061, |
| "grad_norm": 0.7910492113722175, |
| "learning_rate": 9.914071521627964e-06, |
| "loss": 0.4792, |
| "step": 39 |
| }, |
| { |
| "epoch": 0.06216006216006216, |
| "grad_norm": 1.036910491968146, |
| "learning_rate": 9.909503333931402e-06, |
| "loss": 0.699, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.06371406371406371, |
| "grad_norm": 0.8196334764356895, |
| "learning_rate": 9.904817949774524e-06, |
| "loss": 0.7333, |
| "step": 41 |
| }, |
| { |
| "epoch": 0.06526806526806526, |
| "grad_norm": 0.6798779713989535, |
| "learning_rate": 9.900015481003762e-06, |
| "loss": 0.6433, |
| "step": 42 |
| }, |
| { |
| "epoch": 0.06682206682206682, |
| "grad_norm": 0.9642768659293228, |
| "learning_rate": 9.895096042260517e-06, |
| "loss": 0.5257, |
| "step": 43 |
| }, |
| { |
| "epoch": 0.06837606837606838, |
| "grad_norm": 1.7689631037060651, |
| "learning_rate": 9.890059750978425e-06, |
| "loss": 0.6034, |
| "step": 44 |
| }, |
| { |
| "epoch": 0.06993006993006994, |
| "grad_norm": 0.7443934495191462, |
| "learning_rate": 9.88490672738054e-06, |
| "loss": 0.6089, |
| "step": 45 |
| }, |
| { |
| "epoch": 0.07148407148407149, |
| "grad_norm": 1.3800385551010166, |
| "learning_rate": 9.879637094476482e-06, |
| "loss": 0.555, |
| "step": 46 |
| }, |
| { |
| "epoch": 0.07303807303807304, |
| "grad_norm": 0.6934112455074183, |
| "learning_rate": 9.874250978059489e-06, |
| "loss": 0.5253, |
| "step": 47 |
| }, |
| { |
| "epoch": 0.07459207459207459, |
| "grad_norm": 0.728895851371575, |
| "learning_rate": 9.86874850670342e-06, |
| "loss": 0.5111, |
| "step": 48 |
| }, |
| { |
| "epoch": 0.07614607614607614, |
| "grad_norm": 0.705671764219133, |
| "learning_rate": 9.863129811759678e-06, |
| "loss": 0.5166, |
| "step": 49 |
| }, |
| { |
| "epoch": 0.0777000777000777, |
| "grad_norm": 0.9004429060865304, |
| "learning_rate": 9.857395027354085e-06, |
| "loss": 0.4981, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.07925407925407925, |
| "grad_norm": 0.6186440018603123, |
| "learning_rate": 9.85154429038367e-06, |
| "loss": 0.3806, |
| "step": 51 |
| }, |
| { |
| "epoch": 0.08080808080808081, |
| "grad_norm": 0.7725834636904632, |
| "learning_rate": 9.845577740513409e-06, |
| "loss": 0.4948, |
| "step": 52 |
| }, |
| { |
| "epoch": 0.08236208236208237, |
| "grad_norm": 0.9253638464116425, |
| "learning_rate": 9.83949552017289e-06, |
| "loss": 0.5285, |
| "step": 53 |
| }, |
| { |
| "epoch": 0.08391608391608392, |
| "grad_norm": 0.7656584472395311, |
| "learning_rate": 9.833297774552905e-06, |
| "loss": 0.6342, |
| "step": 54 |
| }, |
| { |
| "epoch": 0.08547008547008547, |
| "grad_norm": 0.8998211569830732, |
| "learning_rate": 9.826984651601998e-06, |
| "loss": 0.6178, |
| "step": 55 |
| }, |
| { |
| "epoch": 0.08702408702408702, |
| "grad_norm": 0.8685832029444266, |
| "learning_rate": 9.820556302022916e-06, |
| "loss": 0.6903, |
| "step": 56 |
| }, |
| { |
| "epoch": 0.08857808857808858, |
| "grad_norm": 0.7304781373527222, |
| "learning_rate": 9.814012879269031e-06, |
| "loss": 0.5054, |
| "step": 57 |
| }, |
| { |
| "epoch": 0.09013209013209013, |
| "grad_norm": 0.9941249676710022, |
| "learning_rate": 9.80735453954066e-06, |
| "loss": 0.5752, |
| "step": 58 |
| }, |
| { |
| "epoch": 0.09168609168609168, |
| "grad_norm": 0.8636774782845266, |
| "learning_rate": 9.800581441781342e-06, |
| "loss": 0.472, |
| "step": 59 |
| }, |
| { |
| "epoch": 0.09324009324009325, |
| "grad_norm": 0.8037019830533219, |
| "learning_rate": 9.79369374767405e-06, |
| "loss": 0.451, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.0947940947940948, |
| "grad_norm": 0.7569996171253018, |
| "learning_rate": 9.786691621637322e-06, |
| "loss": 0.5862, |
| "step": 61 |
| }, |
| { |
| "epoch": 0.09634809634809635, |
| "grad_norm": 0.718132158990271, |
| "learning_rate": 9.779575230821344e-06, |
| "loss": 0.57, |
| "step": 62 |
| }, |
| { |
| "epoch": 0.0979020979020979, |
| "grad_norm": 0.8861885775634778, |
| "learning_rate": 9.772344745103955e-06, |
| "loss": 0.661, |
| "step": 63 |
| }, |
| { |
| "epoch": 0.09945609945609946, |
| "grad_norm": 0.7647017941329705, |
| "learning_rate": 9.76500033708659e-06, |
| "loss": 0.5032, |
| "step": 64 |
| }, |
| { |
| "epoch": 0.10101010101010101, |
| "grad_norm": 0.6503958573653673, |
| "learning_rate": 9.757542182090165e-06, |
| "loss": 0.5113, |
| "step": 65 |
| }, |
| { |
| "epoch": 0.10256410256410256, |
| "grad_norm": 0.8514535986961426, |
| "learning_rate": 9.749970458150893e-06, |
| "loss": 0.4582, |
| "step": 66 |
| }, |
| { |
| "epoch": 0.10411810411810411, |
| "grad_norm": 0.8115540624516915, |
| "learning_rate": 9.742285346016024e-06, |
| "loss": 0.5324, |
| "step": 67 |
| }, |
| { |
| "epoch": 0.10567210567210568, |
| "grad_norm": 0.8716261250005453, |
| "learning_rate": 9.734487029139544e-06, |
| "loss": 0.5622, |
| "step": 68 |
| }, |
| { |
| "epoch": 0.10722610722610723, |
| "grad_norm": 0.6674304497161411, |
| "learning_rate": 9.726575693677782e-06, |
| "loss": 0.4194, |
| "step": 69 |
| }, |
| { |
| "epoch": 0.10878010878010878, |
| "grad_norm": 1.0158040848497383, |
| "learning_rate": 9.718551528484979e-06, |
| "loss": 0.4802, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.11033411033411034, |
| "grad_norm": 0.595576459482197, |
| "learning_rate": 9.710414725108771e-06, |
| "loss": 0.4594, |
| "step": 71 |
| }, |
| { |
| "epoch": 0.11188811188811189, |
| "grad_norm": 0.9680714108769358, |
| "learning_rate": 9.702165477785618e-06, |
| "loss": 0.5109, |
| "step": 72 |
| }, |
| { |
| "epoch": 0.11344211344211344, |
| "grad_norm": 0.7314195273632106, |
| "learning_rate": 9.69380398343617e-06, |
| "loss": 0.4759, |
| "step": 73 |
| }, |
| { |
| "epoch": 0.11499611499611499, |
| "grad_norm": 0.8578361036597735, |
| "learning_rate": 9.685330441660564e-06, |
| "loss": 0.5533, |
| "step": 74 |
| }, |
| { |
| "epoch": 0.11655011655011654, |
| "grad_norm": 1.890522114467283, |
| "learning_rate": 9.676745054733661e-06, |
| "loss": 0.6364, |
| "step": 75 |
| }, |
| { |
| "epoch": 0.11810411810411811, |
| "grad_norm": 0.8238570929416833, |
| "learning_rate": 9.668048027600217e-06, |
| "loss": 0.5262, |
| "step": 76 |
| }, |
| { |
| "epoch": 0.11965811965811966, |
| "grad_norm": 0.7000587405463369, |
| "learning_rate": 9.659239567869989e-06, |
| "loss": 0.5541, |
| "step": 77 |
| }, |
| { |
| "epoch": 0.12121212121212122, |
| "grad_norm": 1.0275499560212522, |
| "learning_rate": 9.650319885812777e-06, |
| "loss": 0.6356, |
| "step": 78 |
| }, |
| { |
| "epoch": 0.12276612276612277, |
| "grad_norm": 0.9584707284083808, |
| "learning_rate": 9.641289194353418e-06, |
| "loss": 0.5393, |
| "step": 79 |
| }, |
| { |
| "epoch": 0.12432012432012432, |
| "grad_norm": 0.7661012737973905, |
| "learning_rate": 9.632147709066682e-06, |
| "loss": 0.5864, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.1258741258741259, |
| "grad_norm": 0.8981339358976846, |
| "learning_rate": 9.622895648172141e-06, |
| "loss": 0.5001, |
| "step": 81 |
| }, |
| { |
| "epoch": 0.12742812742812742, |
| "grad_norm": 0.9136632663354124, |
| "learning_rate": 9.613533232528956e-06, |
| "loss": 0.5989, |
| "step": 82 |
| }, |
| { |
| "epoch": 0.128982128982129, |
| "grad_norm": 1.0752006190192398, |
| "learning_rate": 9.604060685630608e-06, |
| "loss": 0.6551, |
| "step": 83 |
| }, |
| { |
| "epoch": 0.13053613053613053, |
| "grad_norm": 0.6746196580403023, |
| "learning_rate": 9.594478233599551e-06, |
| "loss": 0.5763, |
| "step": 84 |
| }, |
| { |
| "epoch": 0.1320901320901321, |
| "grad_norm": 1.0560596133360578, |
| "learning_rate": 9.584786105181831e-06, |
| "loss": 0.626, |
| "step": 85 |
| }, |
| { |
| "epoch": 0.13364413364413363, |
| "grad_norm": 1.9322491414214575, |
| "learning_rate": 9.574984531741613e-06, |
| "loss": 0.6028, |
| "step": 86 |
| }, |
| { |
| "epoch": 0.1351981351981352, |
| "grad_norm": 1.1199274751748922, |
| "learning_rate": 9.565073747255665e-06, |
| "loss": 0.5536, |
| "step": 87 |
| }, |
| { |
| "epoch": 0.13675213675213677, |
| "grad_norm": 1.2142928266989612, |
| "learning_rate": 9.555053988307764e-06, |
| "loss": 0.5881, |
| "step": 88 |
| }, |
| { |
| "epoch": 0.1383061383061383, |
| "grad_norm": 0.886265217836807, |
| "learning_rate": 9.544925494083062e-06, |
| "loss": 0.4521, |
| "step": 89 |
| }, |
| { |
| "epoch": 0.13986013986013987, |
| "grad_norm": 0.7758293999026016, |
| "learning_rate": 9.53468850636236e-06, |
| "loss": 0.6255, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.1414141414141414, |
| "grad_norm": 0.8222144113614714, |
| "learning_rate": 9.524343269516354e-06, |
| "loss": 0.5643, |
| "step": 91 |
| }, |
| { |
| "epoch": 0.14296814296814297, |
| "grad_norm": 0.7971504019254859, |
| "learning_rate": 9.513890030499786e-06, |
| "loss": 0.5469, |
| "step": 92 |
| }, |
| { |
| "epoch": 0.1445221445221445, |
| "grad_norm": 1.0988108202778002, |
| "learning_rate": 9.503329038845556e-06, |
| "loss": 0.6065, |
| "step": 93 |
| }, |
| { |
| "epoch": 0.14607614607614608, |
| "grad_norm": 0.9770953296063096, |
| "learning_rate": 9.492660546658771e-06, |
| "loss": 0.4678, |
| "step": 94 |
| }, |
| { |
| "epoch": 0.14763014763014762, |
| "grad_norm": 0.7001359835404304, |
| "learning_rate": 9.481884808610712e-06, |
| "loss": 0.5186, |
| "step": 95 |
| }, |
| { |
| "epoch": 0.14918414918414918, |
| "grad_norm": 1.028392902002761, |
| "learning_rate": 9.471002081932767e-06, |
| "loss": 0.5347, |
| "step": 96 |
| }, |
| { |
| "epoch": 0.15073815073815075, |
| "grad_norm": 0.6722329878713879, |
| "learning_rate": 9.460012626410286e-06, |
| "loss": 0.5138, |
| "step": 97 |
| }, |
| { |
| "epoch": 0.1522921522921523, |
| "grad_norm": 1.0396069509533319, |
| "learning_rate": 9.448916704376384e-06, |
| "loss": 0.5388, |
| "step": 98 |
| }, |
| { |
| "epoch": 0.15384615384615385, |
| "grad_norm": 0.788486690631997, |
| "learning_rate": 9.437714580705671e-06, |
| "loss": 0.4044, |
| "step": 99 |
| }, |
| { |
| "epoch": 0.1554001554001554, |
| "grad_norm": 0.7359447373312309, |
| "learning_rate": 9.426406522807932e-06, |
| "loss": 0.6637, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.15695415695415696, |
| "grad_norm": 1.0558555591940144, |
| "learning_rate": 9.414992800621749e-06, |
| "loss": 0.4842, |
| "step": 101 |
| }, |
| { |
| "epoch": 0.1585081585081585, |
| "grad_norm": 0.8538402867666735, |
| "learning_rate": 9.40347368660805e-06, |
| "loss": 0.51, |
| "step": 102 |
| }, |
| { |
| "epoch": 0.16006216006216006, |
| "grad_norm": 0.5016420650271773, |
| "learning_rate": 9.39184945574361e-06, |
| "loss": 0.3733, |
| "step": 103 |
| }, |
| { |
| "epoch": 0.16161616161616163, |
| "grad_norm": 0.8043995407764333, |
| "learning_rate": 9.380120385514484e-06, |
| "loss": 0.6424, |
| "step": 104 |
| }, |
| { |
| "epoch": 0.16317016317016317, |
| "grad_norm": 0.7432969785115889, |
| "learning_rate": 9.368286755909383e-06, |
| "loss": 0.5918, |
| "step": 105 |
| }, |
| { |
| "epoch": 0.16472416472416473, |
| "grad_norm": 0.7800858860573905, |
| "learning_rate": 9.356348849412991e-06, |
| "loss": 0.5557, |
| "step": 106 |
| }, |
| { |
| "epoch": 0.16627816627816627, |
| "grad_norm": 0.8792710300993529, |
| "learning_rate": 9.344306950999226e-06, |
| "loss": 0.5359, |
| "step": 107 |
| }, |
| { |
| "epoch": 0.16783216783216784, |
| "grad_norm": 1.0983876112460325, |
| "learning_rate": 9.332161348124426e-06, |
| "loss": 0.6833, |
| "step": 108 |
| }, |
| { |
| "epoch": 0.16938616938616938, |
| "grad_norm": 0.8026939716617563, |
| "learning_rate": 9.319912330720502e-06, |
| "loss": 0.5503, |
| "step": 109 |
| }, |
| { |
| "epoch": 0.17094017094017094, |
| "grad_norm": 0.7137107427499185, |
| "learning_rate": 9.307560191188e-06, |
| "loss": 0.4826, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.17249417249417248, |
| "grad_norm": 0.8145843383585528, |
| "learning_rate": 9.295105224389144e-06, |
| "loss": 0.5508, |
| "step": 111 |
| }, |
| { |
| "epoch": 0.17404817404817405, |
| "grad_norm": 0.6415967276646262, |
| "learning_rate": 9.282547727640767e-06, |
| "loss": 0.6075, |
| "step": 112 |
| }, |
| { |
| "epoch": 0.17560217560217561, |
| "grad_norm": 0.6695998673267839, |
| "learning_rate": 9.269888000707243e-06, |
| "loss": 0.4768, |
| "step": 113 |
| }, |
| { |
| "epoch": 0.17715617715617715, |
| "grad_norm": 0.692586482200001, |
| "learning_rate": 9.25712634579331e-06, |
| "loss": 0.5651, |
| "step": 114 |
| }, |
| { |
| "epoch": 0.17871017871017872, |
| "grad_norm": 0.9490969710584504, |
| "learning_rate": 9.244263067536872e-06, |
| "loss": 0.4607, |
| "step": 115 |
| }, |
| { |
| "epoch": 0.18026418026418026, |
| "grad_norm": 0.7448331565841488, |
| "learning_rate": 9.23129847300171e-06, |
| "loss": 0.5287, |
| "step": 116 |
| }, |
| { |
| "epoch": 0.18181818181818182, |
| "grad_norm": 0.7568020783154017, |
| "learning_rate": 9.218232871670168e-06, |
| "loss": 0.5876, |
| "step": 117 |
| }, |
| { |
| "epoch": 0.18337218337218336, |
| "grad_norm": 0.703790461691835, |
| "learning_rate": 9.205066575435754e-06, |
| "loss": 0.5477, |
| "step": 118 |
| }, |
| { |
| "epoch": 0.18492618492618493, |
| "grad_norm": 0.7355757023568825, |
| "learning_rate": 9.191799898595706e-06, |
| "loss": 0.4819, |
| "step": 119 |
| }, |
| { |
| "epoch": 0.1864801864801865, |
| "grad_norm": 0.7718259998249563, |
| "learning_rate": 9.178433157843474e-06, |
| "loss": 0.5529, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.18803418803418803, |
| "grad_norm": 0.7325746328877619, |
| "learning_rate": 9.164966672261171e-06, |
| "loss": 0.471, |
| "step": 121 |
| }, |
| { |
| "epoch": 0.1895881895881896, |
| "grad_norm": 0.7807440267754963, |
| "learning_rate": 9.151400763311958e-06, |
| "loss": 0.5616, |
| "step": 122 |
| }, |
| { |
| "epoch": 0.19114219114219114, |
| "grad_norm": 0.6694238548360774, |
| "learning_rate": 9.13773575483236e-06, |
| "loss": 0.5637, |
| "step": 123 |
| }, |
| { |
| "epoch": 0.1926961926961927, |
| "grad_norm": 0.7619298527686099, |
| "learning_rate": 9.123971973024543e-06, |
| "loss": 0.5222, |
| "step": 124 |
| }, |
| { |
| "epoch": 0.19425019425019424, |
| "grad_norm": 0.8077572349157733, |
| "learning_rate": 9.110109746448527e-06, |
| "loss": 0.4792, |
| "step": 125 |
| }, |
| { |
| "epoch": 0.1958041958041958, |
| "grad_norm": 0.7156882422608634, |
| "learning_rate": 9.09614940601434e-06, |
| "loss": 0.5079, |
| "step": 126 |
| }, |
| { |
| "epoch": 0.19735819735819735, |
| "grad_norm": 0.8839790656748958, |
| "learning_rate": 9.08209128497412e-06, |
| "loss": 0.6167, |
| "step": 127 |
| }, |
| { |
| "epoch": 0.1989121989121989, |
| "grad_norm": 0.7361442782777586, |
| "learning_rate": 9.06793571891416e-06, |
| "loss": 0.573, |
| "step": 128 |
| }, |
| { |
| "epoch": 0.20046620046620048, |
| "grad_norm": 0.6874615113965296, |
| "learning_rate": 9.053683045746897e-06, |
| "loss": 0.5625, |
| "step": 129 |
| }, |
| { |
| "epoch": 0.20202020202020202, |
| "grad_norm": 1.1187739918068127, |
| "learning_rate": 9.039333605702844e-06, |
| "loss": 0.5386, |
| "step": 130 |
| }, |
| { |
| "epoch": 0.20357420357420358, |
| "grad_norm": 0.7935185219891059, |
| "learning_rate": 9.024887741322475e-06, |
| "loss": 0.4354, |
| "step": 131 |
| }, |
| { |
| "epoch": 0.20512820512820512, |
| "grad_norm": 0.6755878221667591, |
| "learning_rate": 9.010345797448037e-06, |
| "loss": 0.4871, |
| "step": 132 |
| }, |
| { |
| "epoch": 0.2066822066822067, |
| "grad_norm": 0.8821400973796808, |
| "learning_rate": 8.995708121215325e-06, |
| "loss": 0.5988, |
| "step": 133 |
| }, |
| { |
| "epoch": 0.20823620823620823, |
| "grad_norm": 0.8963468623381899, |
| "learning_rate": 8.980975062045398e-06, |
| "loss": 0.4808, |
| "step": 134 |
| }, |
| { |
| "epoch": 0.2097902097902098, |
| "grad_norm": 0.7605330355025913, |
| "learning_rate": 8.96614697163623e-06, |
| "loss": 0.5307, |
| "step": 135 |
| }, |
| { |
| "epoch": 0.21134421134421136, |
| "grad_norm": 0.8377449496608493, |
| "learning_rate": 8.95122420395432e-06, |
| "loss": 0.5855, |
| "step": 136 |
| }, |
| { |
| "epoch": 0.2128982128982129, |
| "grad_norm": 0.6874533009837056, |
| "learning_rate": 8.936207115226242e-06, |
| "loss": 0.5025, |
| "step": 137 |
| }, |
| { |
| "epoch": 0.21445221445221446, |
| "grad_norm": 1.0615886301023565, |
| "learning_rate": 8.921096063930141e-06, |
| "loss": 0.5858, |
| "step": 138 |
| }, |
| { |
| "epoch": 0.216006216006216, |
| "grad_norm": 0.7440356443459726, |
| "learning_rate": 8.905891410787174e-06, |
| "loss": 0.4721, |
| "step": 139 |
| }, |
| { |
| "epoch": 0.21756021756021757, |
| "grad_norm": 0.9780044181786849, |
| "learning_rate": 8.8905935187529e-06, |
| "loss": 0.5506, |
| "step": 140 |
| }, |
| { |
| "epoch": 0.2191142191142191, |
| "grad_norm": 0.7269242314361574, |
| "learning_rate": 8.875202753008614e-06, |
| "loss": 0.5308, |
| "step": 141 |
| }, |
| { |
| "epoch": 0.22066822066822067, |
| "grad_norm": 0.8667053692132669, |
| "learning_rate": 8.859719480952637e-06, |
| "loss": 0.5679, |
| "step": 142 |
| }, |
| { |
| "epoch": 0.2222222222222222, |
| "grad_norm": 0.7257081688438616, |
| "learning_rate": 8.844144072191537e-06, |
| "loss": 0.4858, |
| "step": 143 |
| }, |
| { |
| "epoch": 0.22377622377622378, |
| "grad_norm": 0.6542430837557276, |
| "learning_rate": 8.828476898531308e-06, |
| "loss": 0.5328, |
| "step": 144 |
| }, |
| { |
| "epoch": 0.22533022533022534, |
| "grad_norm": 0.7971170503328316, |
| "learning_rate": 8.812718333968498e-06, |
| "loss": 0.5695, |
| "step": 145 |
| }, |
| { |
| "epoch": 0.22688422688422688, |
| "grad_norm": 0.8486913376544397, |
| "learning_rate": 8.79686875468128e-06, |
| "loss": 0.5336, |
| "step": 146 |
| }, |
| { |
| "epoch": 0.22843822843822845, |
| "grad_norm": 1.0063509221057145, |
| "learning_rate": 8.780928539020467e-06, |
| "loss": 0.4731, |
| "step": 147 |
| }, |
| { |
| "epoch": 0.22999222999222999, |
| "grad_norm": 1.084598842228243, |
| "learning_rate": 8.764898067500488e-06, |
| "loss": 0.5336, |
| "step": 148 |
| }, |
| { |
| "epoch": 0.23154623154623155, |
| "grad_norm": 0.8477855123394982, |
| "learning_rate": 8.7487777227903e-06, |
| "loss": 0.7133, |
| "step": 149 |
| }, |
| { |
| "epoch": 0.2331002331002331, |
| "grad_norm": 0.7433328633942683, |
| "learning_rate": 8.732567889704253e-06, |
| "loss": 0.4906, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.23465423465423466, |
| "grad_norm": 1.3157111994706745, |
| "learning_rate": 8.716268955192908e-06, |
| "loss": 0.4115, |
| "step": 151 |
| }, |
| { |
| "epoch": 0.23620823620823622, |
| "grad_norm": 0.7915085431469301, |
| "learning_rate": 8.699881308333794e-06, |
| "loss": 0.4529, |
| "step": 152 |
| }, |
| { |
| "epoch": 0.23776223776223776, |
| "grad_norm": 0.7093953259845076, |
| "learning_rate": 8.683405340322123e-06, |
| "loss": 0.5795, |
| "step": 153 |
| }, |
| { |
| "epoch": 0.23931623931623933, |
| "grad_norm": 0.9148926328978699, |
| "learning_rate": 8.666841444461456e-06, |
| "loss": 0.5371, |
| "step": 154 |
| }, |
| { |
| "epoch": 0.24087024087024086, |
| "grad_norm": 0.8451708726467831, |
| "learning_rate": 8.650190016154307e-06, |
| "loss": 0.6872, |
| "step": 155 |
| }, |
| { |
| "epoch": 0.24242424242424243, |
| "grad_norm": 0.7575098139209872, |
| "learning_rate": 8.633451452892707e-06, |
| "loss": 0.5262, |
| "step": 156 |
| }, |
| { |
| "epoch": 0.24397824397824397, |
| "grad_norm": 0.8039436989228282, |
| "learning_rate": 8.616626154248717e-06, |
| "loss": 0.5936, |
| "step": 157 |
| }, |
| { |
| "epoch": 0.24553224553224554, |
| "grad_norm": 0.9111018315972883, |
| "learning_rate": 8.59971452186489e-06, |
| "loss": 0.433, |
| "step": 158 |
| }, |
| { |
| "epoch": 0.24708624708624707, |
| "grad_norm": 0.7730492291768862, |
| "learning_rate": 8.582716959444679e-06, |
| "loss": 0.4906, |
| "step": 159 |
| }, |
| { |
| "epoch": 0.24864024864024864, |
| "grad_norm": 0.7508532423751979, |
| "learning_rate": 8.565633872742803e-06, |
| "loss": 0.5951, |
| "step": 160 |
| }, |
| { |
| "epoch": 0.2501942501942502, |
| "grad_norm": 0.7979075113160532, |
| "learning_rate": 8.548465669555564e-06, |
| "loss": 0.5487, |
| "step": 161 |
| }, |
| { |
| "epoch": 0.2517482517482518, |
| "grad_norm": 2.822907214340435, |
| "learning_rate": 8.531212759711103e-06, |
| "loss": 0.5006, |
| "step": 162 |
| }, |
| { |
| "epoch": 0.2533022533022533, |
| "grad_norm": 0.7951356785770967, |
| "learning_rate": 8.51387555505963e-06, |
| "loss": 0.4786, |
| "step": 163 |
| }, |
| { |
| "epoch": 0.25485625485625485, |
| "grad_norm": 0.7944890316200554, |
| "learning_rate": 8.496454469463583e-06, |
| "loss": 0.5428, |
| "step": 164 |
| }, |
| { |
| "epoch": 0.2564102564102564, |
| "grad_norm": 0.885097829322448, |
| "learning_rate": 8.478949918787746e-06, |
| "loss": 0.5559, |
| "step": 165 |
| }, |
| { |
| "epoch": 0.257964257964258, |
| "grad_norm": 0.8765537416075171, |
| "learning_rate": 8.461362320889338e-06, |
| "loss": 0.6328, |
| "step": 166 |
| }, |
| { |
| "epoch": 0.2595182595182595, |
| "grad_norm": 0.7600338630468001, |
| "learning_rate": 8.443692095608019e-06, |
| "loss": 0.4644, |
| "step": 167 |
| }, |
| { |
| "epoch": 0.26107226107226106, |
| "grad_norm": 0.7771144701749865, |
| "learning_rate": 8.425939664755874e-06, |
| "loss": 0.4637, |
| "step": 168 |
| }, |
| { |
| "epoch": 0.26262626262626265, |
| "grad_norm": 1.066507788727023, |
| "learning_rate": 8.408105452107353e-06, |
| "loss": 0.4977, |
| "step": 169 |
| }, |
| { |
| "epoch": 0.2641802641802642, |
| "grad_norm": 1.0331827740490165, |
| "learning_rate": 8.390189883389143e-06, |
| "loss": 0.4936, |
| "step": 170 |
| }, |
| { |
| "epoch": 0.26573426573426573, |
| "grad_norm": 0.6598501232172324, |
| "learning_rate": 8.37219338627001e-06, |
| "loss": 0.4534, |
| "step": 171 |
| }, |
| { |
| "epoch": 0.26728826728826727, |
| "grad_norm": 0.7813638132356793, |
| "learning_rate": 8.354116390350594e-06, |
| "loss": 0.5658, |
| "step": 172 |
| }, |
| { |
| "epoch": 0.26884226884226886, |
| "grad_norm": 0.9169947439806584, |
| "learning_rate": 8.335959327153148e-06, |
| "loss": 0.6649, |
| "step": 173 |
| }, |
| { |
| "epoch": 0.2703962703962704, |
| "grad_norm": 0.9173506379895737, |
| "learning_rate": 8.317722630111233e-06, |
| "loss": 0.6389, |
| "step": 174 |
| }, |
| { |
| "epoch": 0.27195027195027194, |
| "grad_norm": 0.9071927842627939, |
| "learning_rate": 8.299406734559385e-06, |
| "loss": 0.428, |
| "step": 175 |
| }, |
| { |
| "epoch": 0.27350427350427353, |
| "grad_norm": 0.7688021245301622, |
| "learning_rate": 8.281012077722712e-06, |
| "loss": 0.5133, |
| "step": 176 |
| }, |
| { |
| "epoch": 0.27505827505827507, |
| "grad_norm": 0.6639447866723064, |
| "learning_rate": 8.26253909870646e-06, |
| "loss": 0.5077, |
| "step": 177 |
| }, |
| { |
| "epoch": 0.2766122766122766, |
| "grad_norm": 0.6947258135733688, |
| "learning_rate": 8.24398823848553e-06, |
| "loss": 0.5527, |
| "step": 178 |
| }, |
| { |
| "epoch": 0.27816627816627815, |
| "grad_norm": 0.8705601245175114, |
| "learning_rate": 8.225359939893954e-06, |
| "loss": 0.4838, |
| "step": 179 |
| }, |
| { |
| "epoch": 0.27972027972027974, |
| "grad_norm": 0.6618526787787719, |
| "learning_rate": 8.206654647614323e-06, |
| "loss": 0.3804, |
| "step": 180 |
| }, |
| { |
| "epoch": 0.2812742812742813, |
| "grad_norm": 1.0338275501153558, |
| "learning_rate": 8.18787280816717e-06, |
| "loss": 0.5362, |
| "step": 181 |
| }, |
| { |
| "epoch": 0.2828282828282828, |
| "grad_norm": 0.7923177578469405, |
| "learning_rate": 8.169014869900308e-06, |
| "loss": 0.5842, |
| "step": 182 |
| }, |
| { |
| "epoch": 0.28438228438228436, |
| "grad_norm": 0.6396399402988676, |
| "learning_rate": 8.150081282978139e-06, |
| "loss": 0.4271, |
| "step": 183 |
| }, |
| { |
| "epoch": 0.28593628593628595, |
| "grad_norm": 0.8865011660122173, |
| "learning_rate": 8.131072499370897e-06, |
| "loss": 0.5489, |
| "step": 184 |
| }, |
| { |
| "epoch": 0.2874902874902875, |
| "grad_norm": 0.9138845420481327, |
| "learning_rate": 8.111988972843859e-06, |
| "loss": 0.5943, |
| "step": 185 |
| }, |
| { |
| "epoch": 0.289044289044289, |
| "grad_norm": 1.1268837902712043, |
| "learning_rate": 8.09283115894652e-06, |
| "loss": 0.5491, |
| "step": 186 |
| }, |
| { |
| "epoch": 0.2905982905982906, |
| "grad_norm": 0.7444814008859322, |
| "learning_rate": 8.073599515001713e-06, |
| "loss": 0.4997, |
| "step": 187 |
| }, |
| { |
| "epoch": 0.29215229215229216, |
| "grad_norm": 0.8044881882512651, |
| "learning_rate": 8.054294500094697e-06, |
| "loss": 0.5325, |
| "step": 188 |
| }, |
| { |
| "epoch": 0.2937062937062937, |
| "grad_norm": 0.8146942609420201, |
| "learning_rate": 8.034916575062188e-06, |
| "loss": 0.5437, |
| "step": 189 |
| }, |
| { |
| "epoch": 0.29526029526029524, |
| "grad_norm": 0.7872597021358523, |
| "learning_rate": 8.015466202481371e-06, |
| "loss": 0.4712, |
| "step": 190 |
| }, |
| { |
| "epoch": 0.29681429681429683, |
| "grad_norm": 0.7193071970265177, |
| "learning_rate": 7.995943846658852e-06, |
| "loss": 0.5397, |
| "step": 191 |
| }, |
| { |
| "epoch": 0.29836829836829837, |
| "grad_norm": 0.8391823684267, |
| "learning_rate": 7.976349973619567e-06, |
| "loss": 0.5765, |
| "step": 192 |
| }, |
| { |
| "epoch": 0.2999222999222999, |
| "grad_norm": 0.6845170782635249, |
| "learning_rate": 7.956685051095672e-06, |
| "loss": 0.56, |
| "step": 193 |
| }, |
| { |
| "epoch": 0.3014763014763015, |
| "grad_norm": 0.7252420419163962, |
| "learning_rate": 7.936949548515364e-06, |
| "loss": 0.605, |
| "step": 194 |
| }, |
| { |
| "epoch": 0.30303030303030304, |
| "grad_norm": 0.7875283859706272, |
| "learning_rate": 7.917143936991688e-06, |
| "loss": 0.5004, |
| "step": 195 |
| }, |
| { |
| "epoch": 0.3045843045843046, |
| "grad_norm": 0.6047235742618552, |
| "learning_rate": 7.897268689311278e-06, |
| "loss": 0.4562, |
| "step": 196 |
| }, |
| { |
| "epoch": 0.3061383061383061, |
| "grad_norm": 0.5825984478127255, |
| "learning_rate": 7.877324279923078e-06, |
| "loss": 0.451, |
| "step": 197 |
| }, |
| { |
| "epoch": 0.3076923076923077, |
| "grad_norm": 0.7369125343894832, |
| "learning_rate": 7.857311184927015e-06, |
| "loss": 0.5195, |
| "step": 198 |
| }, |
| { |
| "epoch": 0.30924630924630925, |
| "grad_norm": 0.6104305536066545, |
| "learning_rate": 7.837229882062638e-06, |
| "loss": 0.531, |
| "step": 199 |
| }, |
| { |
| "epoch": 0.3108003108003108, |
| "grad_norm": 0.9139389379071885, |
| "learning_rate": 7.817080850697705e-06, |
| "loss": 0.4996, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.3123543123543124, |
| "grad_norm": 0.8513425536888274, |
| "learning_rate": 7.796864571816745e-06, |
| "loss": 0.3977, |
| "step": 201 |
| }, |
| { |
| "epoch": 0.3139083139083139, |
| "grad_norm": 0.7739897197051825, |
| "learning_rate": 7.77658152800958e-06, |
| "loss": 0.5569, |
| "step": 202 |
| }, |
| { |
| "epoch": 0.31546231546231546, |
| "grad_norm": 0.7242668166031772, |
| "learning_rate": 7.756232203459794e-06, |
| "loss": 0.4652, |
| "step": 203 |
| }, |
| { |
| "epoch": 0.317016317016317, |
| "grad_norm": 0.7649888193815342, |
| "learning_rate": 7.735817083933189e-06, |
| "loss": 0.4699, |
| "step": 204 |
| }, |
| { |
| "epoch": 0.3185703185703186, |
| "grad_norm": 1.0274699034423136, |
| "learning_rate": 7.715336656766176e-06, |
| "loss": 0.5284, |
| "step": 205 |
| }, |
| { |
| "epoch": 0.3201243201243201, |
| "grad_norm": 0.6976554827980787, |
| "learning_rate": 7.69479141085415e-06, |
| "loss": 0.4875, |
| "step": 206 |
| }, |
| { |
| "epoch": 0.32167832167832167, |
| "grad_norm": 1.130786196712028, |
| "learning_rate": 7.674181836639819e-06, |
| "loss": 0.4214, |
| "step": 207 |
| }, |
| { |
| "epoch": 0.32323232323232326, |
| "grad_norm": 0.7607350344632271, |
| "learning_rate": 7.653508426101488e-06, |
| "loss": 0.4803, |
| "step": 208 |
| }, |
| { |
| "epoch": 0.3247863247863248, |
| "grad_norm": 0.7396245146929729, |
| "learning_rate": 7.632771672741326e-06, |
| "loss": 0.5967, |
| "step": 209 |
| }, |
| { |
| "epoch": 0.32634032634032634, |
| "grad_norm": 0.9687489401453343, |
| "learning_rate": 7.611972071573579e-06, |
| "loss": 0.5753, |
| "step": 210 |
| }, |
| { |
| "epoch": 0.3278943278943279, |
| "grad_norm": 0.7739827653387099, |
| "learning_rate": 7.591110119112757e-06, |
| "loss": 0.495, |
| "step": 211 |
| }, |
| { |
| "epoch": 0.32944832944832947, |
| "grad_norm": 0.8201270448540147, |
| "learning_rate": 7.5701863133617735e-06, |
| "loss": 0.562, |
| "step": 212 |
| }, |
| { |
| "epoch": 0.331002331002331, |
| "grad_norm": 0.8576029389701612, |
| "learning_rate": 7.549201153800073e-06, |
| "loss": 0.5381, |
| "step": 213 |
| }, |
| { |
| "epoch": 0.33255633255633255, |
| "grad_norm": 0.6841266898837828, |
| "learning_rate": 7.528155141371688e-06, |
| "loss": 0.467, |
| "step": 214 |
| }, |
| { |
| "epoch": 0.3341103341103341, |
| "grad_norm": 0.7802991357810948, |
| "learning_rate": 7.507048778473296e-06, |
| "loss": 0.4966, |
| "step": 215 |
| }, |
| { |
| "epoch": 0.3356643356643357, |
| "grad_norm": 0.7218564126997288, |
| "learning_rate": 7.485882568942222e-06, |
| "loss": 0.5112, |
| "step": 216 |
| }, |
| { |
| "epoch": 0.3372183372183372, |
| "grad_norm": 0.7400107616507341, |
| "learning_rate": 7.464657018044411e-06, |
| "loss": 0.6446, |
| "step": 217 |
| }, |
| { |
| "epoch": 0.33877233877233875, |
| "grad_norm": 0.9014902134428892, |
| "learning_rate": 7.443372632462363e-06, |
| "loss": 0.5053, |
| "step": 218 |
| }, |
| { |
| "epoch": 0.34032634032634035, |
| "grad_norm": 0.7329909522650422, |
| "learning_rate": 7.422029920283044e-06, |
| "loss": 0.5443, |
| "step": 219 |
| }, |
| { |
| "epoch": 0.3418803418803419, |
| "grad_norm": 0.6892541253744358, |
| "learning_rate": 7.400629390985753e-06, |
| "loss": 0.4837, |
| "step": 220 |
| }, |
| { |
| "epoch": 0.3434343434343434, |
| "grad_norm": 0.7723348206252342, |
| "learning_rate": 7.379171555429965e-06, |
| "loss": 0.5755, |
| "step": 221 |
| }, |
| { |
| "epoch": 0.34498834498834496, |
| "grad_norm": 7.664546431285865, |
| "learning_rate": 7.357656925843125e-06, |
| "loss": 0.4786, |
| "step": 222 |
| }, |
| { |
| "epoch": 0.34654234654234656, |
| "grad_norm": 0.7853221971786, |
| "learning_rate": 7.336086015808439e-06, |
| "loss": 0.4729, |
| "step": 223 |
| }, |
| { |
| "epoch": 0.3480963480963481, |
| "grad_norm": 0.7864246657192466, |
| "learning_rate": 7.314459340252593e-06, |
| "loss": 0.5848, |
| "step": 224 |
| }, |
| { |
| "epoch": 0.34965034965034963, |
| "grad_norm": 0.7588019542197073, |
| "learning_rate": 7.2927774154334765e-06, |
| "loss": 0.4884, |
| "step": 225 |
| }, |
| { |
| "epoch": 0.35120435120435123, |
| "grad_norm": 0.6496638579237419, |
| "learning_rate": 7.271040758927852e-06, |
| "loss": 0.4323, |
| "step": 226 |
| }, |
| { |
| "epoch": 0.35275835275835277, |
| "grad_norm": 0.6618044442655401, |
| "learning_rate": 7.2492498896190015e-06, |
| "loss": 0.5797, |
| "step": 227 |
| }, |
| { |
| "epoch": 0.3543123543123543, |
| "grad_norm": 0.8288426526834968, |
| "learning_rate": 7.227405327684339e-06, |
| "loss": 0.5113, |
| "step": 228 |
| }, |
| { |
| "epoch": 0.35586635586635584, |
| "grad_norm": 0.6693277294293136, |
| "learning_rate": 7.205507594582994e-06, |
| "loss": 0.4158, |
| "step": 229 |
| }, |
| { |
| "epoch": 0.35742035742035744, |
| "grad_norm": 0.834098167457632, |
| "learning_rate": 7.183557213043365e-06, |
| "loss": 0.4826, |
| "step": 230 |
| }, |
| { |
| "epoch": 0.358974358974359, |
| "grad_norm": 0.7563577310691308, |
| "learning_rate": 7.161554707050637e-06, |
| "loss": 0.4767, |
| "step": 231 |
| }, |
| { |
| "epoch": 0.3605283605283605, |
| "grad_norm": 0.718266653620817, |
| "learning_rate": 7.1395006018342774e-06, |
| "loss": 0.5423, |
| "step": 232 |
| }, |
| { |
| "epoch": 0.3620823620823621, |
| "grad_norm": 0.7794408706255476, |
| "learning_rate": 7.117395423855496e-06, |
| "loss": 0.5394, |
| "step": 233 |
| }, |
| { |
| "epoch": 0.36363636363636365, |
| "grad_norm": 0.7269215366898196, |
| "learning_rate": 7.09523970079468e-06, |
| "loss": 0.5705, |
| "step": 234 |
| }, |
| { |
| "epoch": 0.3651903651903652, |
| "grad_norm": 0.9408390684558995, |
| "learning_rate": 7.073033961538793e-06, |
| "loss": 0.6389, |
| "step": 235 |
| }, |
| { |
| "epoch": 0.3667443667443667, |
| "grad_norm": 0.6638828664952492, |
| "learning_rate": 7.050778736168757e-06, |
| "loss": 0.5542, |
| "step": 236 |
| }, |
| { |
| "epoch": 0.3682983682983683, |
| "grad_norm": 0.8460797947106127, |
| "learning_rate": 7.028474555946787e-06, |
| "loss": 0.5444, |
| "step": 237 |
| }, |
| { |
| "epoch": 0.36985236985236986, |
| "grad_norm": 0.7665734815326051, |
| "learning_rate": 7.006121953303724e-06, |
| "loss": 0.5229, |
| "step": 238 |
| }, |
| { |
| "epoch": 0.3714063714063714, |
| "grad_norm": 0.8309718789678989, |
| "learning_rate": 6.983721461826312e-06, |
| "loss": 0.5784, |
| "step": 239 |
| }, |
| { |
| "epoch": 0.372960372960373, |
| "grad_norm": 0.8222921080795965, |
| "learning_rate": 6.9612736162444695e-06, |
| "loss": 0.4292, |
| "step": 240 |
| }, |
| { |
| "epoch": 0.3745143745143745, |
| "grad_norm": 0.7527448252012365, |
| "learning_rate": 6.938778952418519e-06, |
| "loss": 0.4729, |
| "step": 241 |
| }, |
| { |
| "epoch": 0.37606837606837606, |
| "grad_norm": 0.8985906116004181, |
| "learning_rate": 6.916238007326399e-06, |
| "loss": 0.5655, |
| "step": 242 |
| }, |
| { |
| "epoch": 0.3776223776223776, |
| "grad_norm": 0.6113924311221383, |
| "learning_rate": 6.893651319050842e-06, |
| "loss": 0.5263, |
| "step": 243 |
| }, |
| { |
| "epoch": 0.3791763791763792, |
| "grad_norm": 0.7817006095897586, |
| "learning_rate": 6.871019426766537e-06, |
| "loss": 0.4632, |
| "step": 244 |
| }, |
| { |
| "epoch": 0.38073038073038074, |
| "grad_norm": 1.7226586227932226, |
| "learning_rate": 6.8483428707272456e-06, |
| "loss": 0.5609, |
| "step": 245 |
| }, |
| { |
| "epoch": 0.3822843822843823, |
| "grad_norm": 0.8080544757443882, |
| "learning_rate": 6.825622192252922e-06, |
| "loss": 0.5424, |
| "step": 246 |
| }, |
| { |
| "epoch": 0.3838383838383838, |
| "grad_norm": 0.8602129359285826, |
| "learning_rate": 6.802857933716774e-06, |
| "loss": 0.4986, |
| "step": 247 |
| }, |
| { |
| "epoch": 0.3853923853923854, |
| "grad_norm": 0.7014559071633292, |
| "learning_rate": 6.7800506385323335e-06, |
| "loss": 0.5079, |
| "step": 248 |
| }, |
| { |
| "epoch": 0.38694638694638694, |
| "grad_norm": 0.8917239030684251, |
| "learning_rate": 6.757200851140468e-06, |
| "loss": 0.5281, |
| "step": 249 |
| }, |
| { |
| "epoch": 0.3885003885003885, |
| "grad_norm": 0.7447990452878435, |
| "learning_rate": 6.734309116996392e-06, |
| "loss": 0.6071, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.3900543900543901, |
| "grad_norm": 0.7472396252981742, |
| "learning_rate": 6.711375982556648e-06, |
| "loss": 0.5418, |
| "step": 251 |
| }, |
| { |
| "epoch": 0.3916083916083916, |
| "grad_norm": 0.6987079676338328, |
| "learning_rate": 6.688401995266061e-06, |
| "loss": 0.4444, |
| "step": 252 |
| }, |
| { |
| "epoch": 0.39316239316239315, |
| "grad_norm": 0.8025151202499369, |
| "learning_rate": 6.665387703544661e-06, |
| "loss": 0.5082, |
| "step": 253 |
| }, |
| { |
| "epoch": 0.3947163947163947, |
| "grad_norm": 0.8425567278002095, |
| "learning_rate": 6.642333656774607e-06, |
| "loss": 0.4212, |
| "step": 254 |
| }, |
| { |
| "epoch": 0.3962703962703963, |
| "grad_norm": 0.7875874739440923, |
| "learning_rate": 6.61924040528706e-06, |
| "loss": 0.4969, |
| "step": 255 |
| }, |
| { |
| "epoch": 0.3978243978243978, |
| "grad_norm": 0.645057270258725, |
| "learning_rate": 6.596108500349054e-06, |
| "loss": 0.4857, |
| "step": 256 |
| }, |
| { |
| "epoch": 0.39937839937839936, |
| "grad_norm": 0.670409599043594, |
| "learning_rate": 6.572938494150332e-06, |
| "loss": 0.5544, |
| "step": 257 |
| }, |
| { |
| "epoch": 0.40093240093240096, |
| "grad_norm": 0.6198181436713772, |
| "learning_rate": 6.54973093979016e-06, |
| "loss": 0.4725, |
| "step": 258 |
| }, |
| { |
| "epoch": 0.4024864024864025, |
| "grad_norm": 0.7662890791531904, |
| "learning_rate": 6.526486391264137e-06, |
| "loss": 0.5041, |
| "step": 259 |
| }, |
| { |
| "epoch": 0.40404040404040403, |
| "grad_norm": 0.71665149094791, |
| "learning_rate": 6.503205403450957e-06, |
| "loss": 0.5193, |
| "step": 260 |
| }, |
| { |
| "epoch": 0.40559440559440557, |
| "grad_norm": 0.8140997620947699, |
| "learning_rate": 6.479888532099175e-06, |
| "loss": 0.5243, |
| "step": 261 |
| }, |
| { |
| "epoch": 0.40714840714840717, |
| "grad_norm": 0.624858848374034, |
| "learning_rate": 6.4565363338139245e-06, |
| "loss": 0.3996, |
| "step": 262 |
| }, |
| { |
| "epoch": 0.4087024087024087, |
| "grad_norm": 0.7647591367675671, |
| "learning_rate": 6.433149366043652e-06, |
| "loss": 0.4658, |
| "step": 263 |
| }, |
| { |
| "epoch": 0.41025641025641024, |
| "grad_norm": 0.6633892543296002, |
| "learning_rate": 6.409728187066789e-06, |
| "loss": 0.4452, |
| "step": 264 |
| }, |
| { |
| "epoch": 0.41181041181041184, |
| "grad_norm": 0.9241797213745755, |
| "learning_rate": 6.386273355978442e-06, |
| "loss": 0.5173, |
| "step": 265 |
| }, |
| { |
| "epoch": 0.4133644133644134, |
| "grad_norm": 0.9790400835539028, |
| "learning_rate": 6.3627854326770326e-06, |
| "loss": 0.5611, |
| "step": 266 |
| }, |
| { |
| "epoch": 0.4149184149184149, |
| "grad_norm": 0.7010295847973484, |
| "learning_rate": 6.339264977850943e-06, |
| "loss": 0.6185, |
| "step": 267 |
| }, |
| { |
| "epoch": 0.41647241647241645, |
| "grad_norm": 0.9024237646983181, |
| "learning_rate": 6.3157125529651205e-06, |
| "loss": 0.4375, |
| "step": 268 |
| }, |
| { |
| "epoch": 0.41802641802641805, |
| "grad_norm": 0.728576137938516, |
| "learning_rate": 6.292128720247692e-06, |
| "loss": 0.416, |
| "step": 269 |
| }, |
| { |
| "epoch": 0.4195804195804196, |
| "grad_norm": 0.7209803232506808, |
| "learning_rate": 6.268514042676519e-06, |
| "loss": 0.4931, |
| "step": 270 |
| }, |
| { |
| "epoch": 0.4211344211344211, |
| "grad_norm": 0.9852718978800553, |
| "learning_rate": 6.244869083965777e-06, |
| "loss": 0.6138, |
| "step": 271 |
| }, |
| { |
| "epoch": 0.4226884226884227, |
| "grad_norm": 0.8544813586101946, |
| "learning_rate": 6.221194408552494e-06, |
| "loss": 0.4828, |
| "step": 272 |
| }, |
| { |
| "epoch": 0.42424242424242425, |
| "grad_norm": 0.5339139693029649, |
| "learning_rate": 6.197490581583078e-06, |
| "loss": 0.3763, |
| "step": 273 |
| }, |
| { |
| "epoch": 0.4257964257964258, |
| "grad_norm": 0.7029889613516929, |
| "learning_rate": 6.173758168899814e-06, |
| "loss": 0.5003, |
| "step": 274 |
| }, |
| { |
| "epoch": 0.42735042735042733, |
| "grad_norm": 0.6697709635144842, |
| "learning_rate": 6.149997737027377e-06, |
| "loss": 0.5023, |
| "step": 275 |
| }, |
| { |
| "epoch": 0.4289044289044289, |
| "grad_norm": 0.7711140122998004, |
| "learning_rate": 6.126209853159293e-06, |
| "loss": 0.5458, |
| "step": 276 |
| }, |
| { |
| "epoch": 0.43045843045843046, |
| "grad_norm": 0.7959237764492787, |
| "learning_rate": 6.102395085144406e-06, |
| "loss": 0.5436, |
| "step": 277 |
| }, |
| { |
| "epoch": 0.432012432012432, |
| "grad_norm": 0.7725608290998155, |
| "learning_rate": 6.078554001473317e-06, |
| "loss": 0.5174, |
| "step": 278 |
| }, |
| { |
| "epoch": 0.43356643356643354, |
| "grad_norm": 1.0803231102952509, |
| "learning_rate": 6.054687171264822e-06, |
| "loss": 0.6228, |
| "step": 279 |
| }, |
| { |
| "epoch": 0.43512043512043513, |
| "grad_norm": 0.8198164427592709, |
| "learning_rate": 6.030795164252321e-06, |
| "loss": 0.4489, |
| "step": 280 |
| }, |
| { |
| "epoch": 0.4366744366744367, |
| "grad_norm": 0.9826854880546106, |
| "learning_rate": 6.006878550770213e-06, |
| "loss": 0.4776, |
| "step": 281 |
| }, |
| { |
| "epoch": 0.4382284382284382, |
| "grad_norm": 0.6853341082656256, |
| "learning_rate": 5.982937901740296e-06, |
| "loss": 0.5168, |
| "step": 282 |
| }, |
| { |
| "epoch": 0.4397824397824398, |
| "grad_norm": 0.8527642374724453, |
| "learning_rate": 5.958973788658115e-06, |
| "loss": 0.4427, |
| "step": 283 |
| }, |
| { |
| "epoch": 0.44133644133644134, |
| "grad_norm": 0.820049645998347, |
| "learning_rate": 5.934986783579349e-06, |
| "loss": 0.5325, |
| "step": 284 |
| }, |
| { |
| "epoch": 0.4428904428904429, |
| "grad_norm": 0.8635505053006802, |
| "learning_rate": 5.91097745910613e-06, |
| "loss": 0.5158, |
| "step": 285 |
| }, |
| { |
| "epoch": 0.4444444444444444, |
| "grad_norm": 0.6163515215798748, |
| "learning_rate": 5.886946388373387e-06, |
| "loss": 0.5613, |
| "step": 286 |
| }, |
| { |
| "epoch": 0.445998445998446, |
| "grad_norm": 0.7114308768259194, |
| "learning_rate": 5.862894145035158e-06, |
| "loss": 0.5812, |
| "step": 287 |
| }, |
| { |
| "epoch": 0.44755244755244755, |
| "grad_norm": 0.7695278464891274, |
| "learning_rate": 5.83882130325091e-06, |
| "loss": 0.5216, |
| "step": 288 |
| }, |
| { |
| "epoch": 0.4491064491064491, |
| "grad_norm": 0.7953550029299292, |
| "learning_rate": 5.814728437671808e-06, |
| "loss": 0.3821, |
| "step": 289 |
| }, |
| { |
| "epoch": 0.4506604506604507, |
| "grad_norm": 0.8080489538709861, |
| "learning_rate": 5.7906161234270234e-06, |
| "loss": 0.4982, |
| "step": 290 |
| }, |
| { |
| "epoch": 0.4522144522144522, |
| "grad_norm": 0.7349261540661857, |
| "learning_rate": 5.766484936109987e-06, |
| "loss": 0.4715, |
| "step": 291 |
| }, |
| { |
| "epoch": 0.45376845376845376, |
| "grad_norm": 1.0100168672849925, |
| "learning_rate": 5.7423354517646616e-06, |
| "loss": 0.4585, |
| "step": 292 |
| }, |
| { |
| "epoch": 0.4553224553224553, |
| "grad_norm": 0.7711000819084326, |
| "learning_rate": 5.718168246871775e-06, |
| "loss": 0.5841, |
| "step": 293 |
| }, |
| { |
| "epoch": 0.4568764568764569, |
| "grad_norm": 0.9388643582133742, |
| "learning_rate": 5.69398389833508e-06, |
| "loss": 0.5857, |
| "step": 294 |
| }, |
| { |
| "epoch": 0.45843045843045843, |
| "grad_norm": 0.6391229423057062, |
| "learning_rate": 5.669782983467562e-06, |
| "loss": 0.5786, |
| "step": 295 |
| }, |
| { |
| "epoch": 0.45998445998445997, |
| "grad_norm": 0.7381704493689959, |
| "learning_rate": 5.645566079977673e-06, |
| "loss": 0.4531, |
| "step": 296 |
| }, |
| { |
| "epoch": 0.46153846153846156, |
| "grad_norm": 0.6427062310992607, |
| "learning_rate": 5.621333765955529e-06, |
| "loss": 0.4516, |
| "step": 297 |
| }, |
| { |
| "epoch": 0.4630924630924631, |
| "grad_norm": 0.6045207793918917, |
| "learning_rate": 5.5970866198591235e-06, |
| "loss": 0.5904, |
| "step": 298 |
| }, |
| { |
| "epoch": 0.46464646464646464, |
| "grad_norm": 0.7486268299657493, |
| "learning_rate": 5.572825220500505e-06, |
| "loss": 0.5559, |
| "step": 299 |
| }, |
| { |
| "epoch": 0.4662004662004662, |
| "grad_norm": 0.8393377063443249, |
| "learning_rate": 5.548550147031971e-06, |
| "loss": 0.2951, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.4677544677544678, |
| "grad_norm": 0.5563021839864102, |
| "learning_rate": 5.524261978932234e-06, |
| "loss": 0.4705, |
| "step": 301 |
| }, |
| { |
| "epoch": 0.4693084693084693, |
| "grad_norm": 0.5244277581252728, |
| "learning_rate": 5.4999612959925995e-06, |
| "loss": 0.5826, |
| "step": 302 |
| }, |
| { |
| "epoch": 0.47086247086247085, |
| "grad_norm": 0.8967222238475693, |
| "learning_rate": 5.475648678303112e-06, |
| "loss": 0.4326, |
| "step": 303 |
| }, |
| { |
| "epoch": 0.47241647241647244, |
| "grad_norm": 0.8597788121495409, |
| "learning_rate": 5.451324706238721e-06, |
| "loss": 0.5024, |
| "step": 304 |
| }, |
| { |
| "epoch": 0.473970473970474, |
| "grad_norm": 0.6219585515878757, |
| "learning_rate": 5.426989960445415e-06, |
| "loss": 0.4228, |
| "step": 305 |
| }, |
| { |
| "epoch": 0.4755244755244755, |
| "grad_norm": 0.87442477445858, |
| "learning_rate": 5.402645021826367e-06, |
| "loss": 0.5451, |
| "step": 306 |
| }, |
| { |
| "epoch": 0.47707847707847706, |
| "grad_norm": 0.7018801509337828, |
| "learning_rate": 5.3782904715280705e-06, |
| "loss": 0.454, |
| "step": 307 |
| }, |
| { |
| "epoch": 0.47863247863247865, |
| "grad_norm": 0.7870901274907686, |
| "learning_rate": 5.35392689092646e-06, |
| "loss": 0.6209, |
| "step": 308 |
| }, |
| { |
| "epoch": 0.4801864801864802, |
| "grad_norm": 0.8404067379482166, |
| "learning_rate": 5.329554861613031e-06, |
| "loss": 0.4577, |
| "step": 309 |
| }, |
| { |
| "epoch": 0.48174048174048173, |
| "grad_norm": 0.8609659868189161, |
| "learning_rate": 5.3051749653809685e-06, |
| "loss": 0.516, |
| "step": 310 |
| }, |
| { |
| "epoch": 0.48329448329448327, |
| "grad_norm": 0.6451369228669567, |
| "learning_rate": 5.2807877842112475e-06, |
| "loss": 0.529, |
| "step": 311 |
| }, |
| { |
| "epoch": 0.48484848484848486, |
| "grad_norm": 0.7089279571799519, |
| "learning_rate": 5.256393900258747e-06, |
| "loss": 0.55, |
| "step": 312 |
| }, |
| { |
| "epoch": 0.4864024864024864, |
| "grad_norm": 0.6009435199914196, |
| "learning_rate": 5.231993895838348e-06, |
| "loss": 0.4524, |
| "step": 313 |
| }, |
| { |
| "epoch": 0.48795648795648794, |
| "grad_norm": 0.7399390807603545, |
| "learning_rate": 5.207588353411032e-06, |
| "loss": 0.5826, |
| "step": 314 |
| }, |
| { |
| "epoch": 0.48951048951048953, |
| "grad_norm": 0.6358169721821743, |
| "learning_rate": 5.183177855569989e-06, |
| "loss": 0.4686, |
| "step": 315 |
| }, |
| { |
| "epoch": 0.49106449106449107, |
| "grad_norm": 0.834159373189627, |
| "learning_rate": 5.158762985026694e-06, |
| "loss": 0.6194, |
| "step": 316 |
| }, |
| { |
| "epoch": 0.4926184926184926, |
| "grad_norm": 0.672932649828031, |
| "learning_rate": 5.1343443245970095e-06, |
| "loss": 0.5593, |
| "step": 317 |
| }, |
| { |
| "epoch": 0.49417249417249415, |
| "grad_norm": 0.7305995766455743, |
| "learning_rate": 5.10992245718726e-06, |
| "loss": 0.4943, |
| "step": 318 |
| }, |
| { |
| "epoch": 0.49572649572649574, |
| "grad_norm": 1.1658079579901208, |
| "learning_rate": 5.085497965780335e-06, |
| "loss": 0.5031, |
| "step": 319 |
| }, |
| { |
| "epoch": 0.4972804972804973, |
| "grad_norm": 0.9841093713953577, |
| "learning_rate": 5.061071433421754e-06, |
| "loss": 0.5508, |
| "step": 320 |
| }, |
| { |
| "epoch": 0.4988344988344988, |
| "grad_norm": 0.7841115815090715, |
| "learning_rate": 5.0366434432057624e-06, |
| "loss": 0.563, |
| "step": 321 |
| }, |
| { |
| "epoch": 0.5003885003885004, |
| "grad_norm": 0.5645683948415432, |
| "learning_rate": 5.012214578261402e-06, |
| "loss": 0.4011, |
| "step": 322 |
| }, |
| { |
| "epoch": 0.5019425019425019, |
| "grad_norm": 0.8438933215498013, |
| "learning_rate": 4.987785421738599e-06, |
| "loss": 0.5832, |
| "step": 323 |
| }, |
| { |
| "epoch": 0.5034965034965035, |
| "grad_norm": 0.5640735635986946, |
| "learning_rate": 4.963356556794238e-06, |
| "loss": 0.401, |
| "step": 324 |
| }, |
| { |
| "epoch": 0.5050505050505051, |
| "grad_norm": 1.2747685452336512, |
| "learning_rate": 4.938928566578247e-06, |
| "loss": 0.5392, |
| "step": 325 |
| }, |
| { |
| "epoch": 0.5066045066045066, |
| "grad_norm": 0.6877059270681696, |
| "learning_rate": 4.914502034219667e-06, |
| "loss": 0.439, |
| "step": 326 |
| }, |
| { |
| "epoch": 0.5081585081585082, |
| "grad_norm": 0.705842825427767, |
| "learning_rate": 4.890077542812742e-06, |
| "loss": 0.5187, |
| "step": 327 |
| }, |
| { |
| "epoch": 0.5097125097125097, |
| "grad_norm": 0.5947821340173784, |
| "learning_rate": 4.865655675402993e-06, |
| "loss": 0.4178, |
| "step": 328 |
| }, |
| { |
| "epoch": 0.5112665112665112, |
| "grad_norm": 0.6757902851130476, |
| "learning_rate": 4.841237014973305e-06, |
| "loss": 0.493, |
| "step": 329 |
| }, |
| { |
| "epoch": 0.5128205128205128, |
| "grad_norm": 0.6256933230220865, |
| "learning_rate": 4.8168221444300124e-06, |
| "loss": 0.4499, |
| "step": 330 |
| }, |
| { |
| "epoch": 0.5143745143745144, |
| "grad_norm": 0.6390396367290697, |
| "learning_rate": 4.7924116465889684e-06, |
| "loss": 0.4783, |
| "step": 331 |
| }, |
| { |
| "epoch": 0.515928515928516, |
| "grad_norm": 0.7039920639754556, |
| "learning_rate": 4.768006104161655e-06, |
| "loss": 0.4862, |
| "step": 332 |
| }, |
| { |
| "epoch": 0.5174825174825175, |
| "grad_norm": 0.6996136377759004, |
| "learning_rate": 4.743606099741255e-06, |
| "loss": 0.43, |
| "step": 333 |
| }, |
| { |
| "epoch": 0.519036519036519, |
| "grad_norm": 0.5635901754632139, |
| "learning_rate": 4.719212215788753e-06, |
| "loss": 0.3489, |
| "step": 334 |
| }, |
| { |
| "epoch": 0.5205905205905206, |
| "grad_norm": 1.0174320744839043, |
| "learning_rate": 4.6948250346190315e-06, |
| "loss": 0.4664, |
| "step": 335 |
| }, |
| { |
| "epoch": 0.5221445221445221, |
| "grad_norm": 0.7281418344534993, |
| "learning_rate": 4.670445138386971e-06, |
| "loss": 0.4357, |
| "step": 336 |
| }, |
| { |
| "epoch": 0.5236985236985237, |
| "grad_norm": 0.7444913580894218, |
| "learning_rate": 4.646073109073542e-06, |
| "loss": 0.5164, |
| "step": 337 |
| }, |
| { |
| "epoch": 0.5252525252525253, |
| "grad_norm": 0.6453883418849363, |
| "learning_rate": 4.621709528471931e-06, |
| "loss": 0.4876, |
| "step": 338 |
| }, |
| { |
| "epoch": 0.5268065268065268, |
| "grad_norm": 0.6332116311781928, |
| "learning_rate": 4.5973549781736335e-06, |
| "loss": 0.4512, |
| "step": 339 |
| }, |
| { |
| "epoch": 0.5283605283605284, |
| "grad_norm": 0.7839883177892899, |
| "learning_rate": 4.573010039554587e-06, |
| "loss": 0.4864, |
| "step": 340 |
| }, |
| { |
| "epoch": 0.5299145299145299, |
| "grad_norm": 0.574123724971613, |
| "learning_rate": 4.548675293761281e-06, |
| "loss": 0.5942, |
| "step": 341 |
| }, |
| { |
| "epoch": 0.5314685314685315, |
| "grad_norm": 0.7341359415325991, |
| "learning_rate": 4.524351321696889e-06, |
| "loss": 0.6479, |
| "step": 342 |
| }, |
| { |
| "epoch": 0.533022533022533, |
| "grad_norm": 0.7042514483644169, |
| "learning_rate": 4.500038704007402e-06, |
| "loss": 0.5163, |
| "step": 343 |
| }, |
| { |
| "epoch": 0.5345765345765345, |
| "grad_norm": 0.7905795068611852, |
| "learning_rate": 4.475738021067768e-06, |
| "loss": 0.4238, |
| "step": 344 |
| }, |
| { |
| "epoch": 0.5361305361305362, |
| "grad_norm": 0.6659802197095699, |
| "learning_rate": 4.451449852968031e-06, |
| "loss": 0.5456, |
| "step": 345 |
| }, |
| { |
| "epoch": 0.5376845376845377, |
| "grad_norm": 0.790921006825035, |
| "learning_rate": 4.427174779499498e-06, |
| "loss": 0.3841, |
| "step": 346 |
| }, |
| { |
| "epoch": 0.5392385392385393, |
| "grad_norm": 0.7752025731889072, |
| "learning_rate": 4.402913380140878e-06, |
| "loss": 0.5737, |
| "step": 347 |
| }, |
| { |
| "epoch": 0.5407925407925408, |
| "grad_norm": 0.69589495472164, |
| "learning_rate": 4.378666234044471e-06, |
| "loss": 0.5731, |
| "step": 348 |
| }, |
| { |
| "epoch": 0.5423465423465423, |
| "grad_norm": 0.6620238929419696, |
| "learning_rate": 4.354433920022328e-06, |
| "loss": 0.6002, |
| "step": 349 |
| }, |
| { |
| "epoch": 0.5439005439005439, |
| "grad_norm": 0.7640658489542932, |
| "learning_rate": 4.3302170165324385e-06, |
| "loss": 0.5681, |
| "step": 350 |
| }, |
| { |
| "epoch": 0.5454545454545454, |
| "grad_norm": 0.8733971833765328, |
| "learning_rate": 4.306016101664921e-06, |
| "loss": 0.5128, |
| "step": 351 |
| }, |
| { |
| "epoch": 0.5470085470085471, |
| "grad_norm": 1.0078351927972078, |
| "learning_rate": 4.281831753128226e-06, |
| "loss": 0.5179, |
| "step": 352 |
| }, |
| { |
| "epoch": 0.5485625485625486, |
| "grad_norm": 0.5776801607508973, |
| "learning_rate": 4.257664548235341e-06, |
| "loss": 0.5191, |
| "step": 353 |
| }, |
| { |
| "epoch": 0.5501165501165501, |
| "grad_norm": 0.7595818677545988, |
| "learning_rate": 4.233515063890013e-06, |
| "loss": 0.4912, |
| "step": 354 |
| }, |
| { |
| "epoch": 0.5516705516705517, |
| "grad_norm": 0.6606204189468279, |
| "learning_rate": 4.209383876572977e-06, |
| "loss": 0.5694, |
| "step": 355 |
| }, |
| { |
| "epoch": 0.5532245532245532, |
| "grad_norm": 0.7228853812417981, |
| "learning_rate": 4.1852715623281934e-06, |
| "loss": 0.5009, |
| "step": 356 |
| }, |
| { |
| "epoch": 0.5547785547785548, |
| "grad_norm": 0.679061167051694, |
| "learning_rate": 4.161178696749092e-06, |
| "loss": 0.4821, |
| "step": 357 |
| }, |
| { |
| "epoch": 0.5563325563325563, |
| "grad_norm": 1.034495766490782, |
| "learning_rate": 4.1371058549648425e-06, |
| "loss": 0.5308, |
| "step": 358 |
| }, |
| { |
| "epoch": 0.5578865578865578, |
| "grad_norm": 0.7686980644117477, |
| "learning_rate": 4.1130536116266155e-06, |
| "loss": 0.489, |
| "step": 359 |
| }, |
| { |
| "epoch": 0.5594405594405595, |
| "grad_norm": 0.9681286219055543, |
| "learning_rate": 4.089022540893871e-06, |
| "loss": 0.5282, |
| "step": 360 |
| }, |
| { |
| "epoch": 0.560994560994561, |
| "grad_norm": 0.8743676508733846, |
| "learning_rate": 4.0650132164206515e-06, |
| "loss": 0.5241, |
| "step": 361 |
| }, |
| { |
| "epoch": 0.5625485625485626, |
| "grad_norm": 0.6062796295682528, |
| "learning_rate": 4.041026211341886e-06, |
| "loss": 0.5007, |
| "step": 362 |
| }, |
| { |
| "epoch": 0.5641025641025641, |
| "grad_norm": 0.8866048922703949, |
| "learning_rate": 4.017062098259707e-06, |
| "loss": 0.5478, |
| "step": 363 |
| }, |
| { |
| "epoch": 0.5656565656565656, |
| "grad_norm": 0.6320850311198043, |
| "learning_rate": 3.9931214492297875e-06, |
| "loss": 0.5351, |
| "step": 364 |
| }, |
| { |
| "epoch": 0.5672105672105672, |
| "grad_norm": 0.6435052023704503, |
| "learning_rate": 3.969204835747681e-06, |
| "loss": 0.4535, |
| "step": 365 |
| }, |
| { |
| "epoch": 0.5687645687645687, |
| "grad_norm": 0.6824957419803489, |
| "learning_rate": 3.945312828735179e-06, |
| "loss": 0.4962, |
| "step": 366 |
| }, |
| { |
| "epoch": 0.5703185703185704, |
| "grad_norm": 0.6630522920673861, |
| "learning_rate": 3.921445998526684e-06, |
| "loss": 0.5484, |
| "step": 367 |
| }, |
| { |
| "epoch": 0.5718725718725719, |
| "grad_norm": 0.6203973778780169, |
| "learning_rate": 3.897604914855596e-06, |
| "loss": 0.4307, |
| "step": 368 |
| }, |
| { |
| "epoch": 0.5734265734265734, |
| "grad_norm": 0.8691644246672308, |
| "learning_rate": 3.873790146840709e-06, |
| "loss": 0.4494, |
| "step": 369 |
| }, |
| { |
| "epoch": 0.574980574980575, |
| "grad_norm": 0.6583038284757291, |
| "learning_rate": 3.8500022629726246e-06, |
| "loss": 0.5461, |
| "step": 370 |
| }, |
| { |
| "epoch": 0.5765345765345765, |
| "grad_norm": 0.6205452786275579, |
| "learning_rate": 3.8262418311001884e-06, |
| "loss": 0.441, |
| "step": 371 |
| }, |
| { |
| "epoch": 0.578088578088578, |
| "grad_norm": 0.8090078519631057, |
| "learning_rate": 3.8025094184169254e-06, |
| "loss": 0.4396, |
| "step": 372 |
| }, |
| { |
| "epoch": 0.5796425796425796, |
| "grad_norm": 3.7824694886306442, |
| "learning_rate": 3.778805591447505e-06, |
| "loss": 0.5243, |
| "step": 373 |
| }, |
| { |
| "epoch": 0.5811965811965812, |
| "grad_norm": 0.7728659065744149, |
| "learning_rate": 3.7551309160342233e-06, |
| "loss": 0.492, |
| "step": 374 |
| }, |
| { |
| "epoch": 0.5827505827505828, |
| "grad_norm": 0.6144384200403961, |
| "learning_rate": 3.731485957323483e-06, |
| "loss": 0.5093, |
| "step": 375 |
| }, |
| { |
| "epoch": 0.5843045843045843, |
| "grad_norm": 0.7022871161510597, |
| "learning_rate": 3.707871279752309e-06, |
| "loss": 0.5471, |
| "step": 376 |
| }, |
| { |
| "epoch": 0.5858585858585859, |
| "grad_norm": 0.7563905294320689, |
| "learning_rate": 3.68428744703488e-06, |
| "loss": 0.5169, |
| "step": 377 |
| }, |
| { |
| "epoch": 0.5874125874125874, |
| "grad_norm": 0.7546808388382039, |
| "learning_rate": 3.6607350221490593e-06, |
| "loss": 0.5465, |
| "step": 378 |
| }, |
| { |
| "epoch": 0.5889665889665889, |
| "grad_norm": 2.017091506116899, |
| "learning_rate": 3.6372145673229683e-06, |
| "loss": 0.5379, |
| "step": 379 |
| }, |
| { |
| "epoch": 0.5905205905205905, |
| "grad_norm": 1.0384918566794876, |
| "learning_rate": 3.613726644021559e-06, |
| "loss": 0.5533, |
| "step": 380 |
| }, |
| { |
| "epoch": 0.5920745920745921, |
| "grad_norm": 0.7843829856867381, |
| "learning_rate": 3.590271812933212e-06, |
| "loss": 0.5192, |
| "step": 381 |
| }, |
| { |
| "epoch": 0.5936285936285937, |
| "grad_norm": 0.8917403925552857, |
| "learning_rate": 3.5668506339563502e-06, |
| "loss": 0.5273, |
| "step": 382 |
| }, |
| { |
| "epoch": 0.5951825951825952, |
| "grad_norm": 0.8080693974342459, |
| "learning_rate": 3.5434636661860776e-06, |
| "loss": 0.4502, |
| "step": 383 |
| }, |
| { |
| "epoch": 0.5967365967365967, |
| "grad_norm": 0.8589025653140446, |
| "learning_rate": 3.5201114679008286e-06, |
| "loss": 0.4597, |
| "step": 384 |
| }, |
| { |
| "epoch": 0.5982905982905983, |
| "grad_norm": 0.8451337336955351, |
| "learning_rate": 3.4967945965490434e-06, |
| "loss": 0.5479, |
| "step": 385 |
| }, |
| { |
| "epoch": 0.5998445998445998, |
| "grad_norm": 0.6022399299860023, |
| "learning_rate": 3.4735136087358646e-06, |
| "loss": 0.3617, |
| "step": 386 |
| }, |
| { |
| "epoch": 0.6013986013986014, |
| "grad_norm": 1.091419134601723, |
| "learning_rate": 3.450269060209841e-06, |
| "loss": 0.4422, |
| "step": 387 |
| }, |
| { |
| "epoch": 0.602952602952603, |
| "grad_norm": 0.7663320014407212, |
| "learning_rate": 3.42706150584967e-06, |
| "loss": 0.5793, |
| "step": 388 |
| }, |
| { |
| "epoch": 0.6045066045066045, |
| "grad_norm": 0.8338372079331721, |
| "learning_rate": 3.4038914996509464e-06, |
| "loss": 0.6015, |
| "step": 389 |
| }, |
| { |
| "epoch": 0.6060606060606061, |
| "grad_norm": 0.8601272441851022, |
| "learning_rate": 3.3807595947129405e-06, |
| "loss": 0.6391, |
| "step": 390 |
| }, |
| { |
| "epoch": 0.6076146076146076, |
| "grad_norm": 0.8548829786482054, |
| "learning_rate": 3.357666343225396e-06, |
| "loss": 0.5055, |
| "step": 391 |
| }, |
| { |
| "epoch": 0.6091686091686092, |
| "grad_norm": 0.7149586625595873, |
| "learning_rate": 3.3346122964553407e-06, |
| "loss": 0.4305, |
| "step": 392 |
| }, |
| { |
| "epoch": 0.6107226107226107, |
| "grad_norm": 0.7734311233562257, |
| "learning_rate": 3.3115980047339415e-06, |
| "loss": 0.4676, |
| "step": 393 |
| }, |
| { |
| "epoch": 0.6122766122766122, |
| "grad_norm": 0.7524053415700651, |
| "learning_rate": 3.288624017443353e-06, |
| "loss": 0.6177, |
| "step": 394 |
| }, |
| { |
| "epoch": 0.6138306138306139, |
| "grad_norm": 0.6577165811691974, |
| "learning_rate": 3.265690883003609e-06, |
| "loss": 0.4062, |
| "step": 395 |
| }, |
| { |
| "epoch": 0.6153846153846154, |
| "grad_norm": 0.7818702560735775, |
| "learning_rate": 3.2427991488595334e-06, |
| "loss": 0.7139, |
| "step": 396 |
| }, |
| { |
| "epoch": 0.616938616938617, |
| "grad_norm": 0.8337417209233989, |
| "learning_rate": 3.219949361467668e-06, |
| "loss": 0.3658, |
| "step": 397 |
| }, |
| { |
| "epoch": 0.6184926184926185, |
| "grad_norm": 0.6408051420484582, |
| "learning_rate": 3.197142066283225e-06, |
| "loss": 0.4037, |
| "step": 398 |
| }, |
| { |
| "epoch": 0.62004662004662, |
| "grad_norm": 0.6924590018612481, |
| "learning_rate": 3.174377807747079e-06, |
| "loss": 0.3915, |
| "step": 399 |
| }, |
| { |
| "epoch": 0.6216006216006216, |
| "grad_norm": 0.7521284979964744, |
| "learning_rate": 3.1516571292727553e-06, |
| "loss": 0.5316, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.6231546231546231, |
| "grad_norm": 0.6902359570029046, |
| "learning_rate": 3.128980573233465e-06, |
| "loss": 0.3485, |
| "step": 401 |
| }, |
| { |
| "epoch": 0.6247086247086248, |
| "grad_norm": 0.8616508180715082, |
| "learning_rate": 3.1063486809491595e-06, |
| "loss": 0.5077, |
| "step": 402 |
| }, |
| { |
| "epoch": 0.6262626262626263, |
| "grad_norm": 0.7958497698850551, |
| "learning_rate": 3.0837619926736027e-06, |
| "loss": 0.4225, |
| "step": 403 |
| }, |
| { |
| "epoch": 0.6278166278166278, |
| "grad_norm": 0.7109428935226295, |
| "learning_rate": 3.061221047581482e-06, |
| "loss": 0.6677, |
| "step": 404 |
| }, |
| { |
| "epoch": 0.6293706293706294, |
| "grad_norm": 0.7470001438392854, |
| "learning_rate": 3.038726383755531e-06, |
| "loss": 0.4242, |
| "step": 405 |
| }, |
| { |
| "epoch": 0.6309246309246309, |
| "grad_norm": 0.6742001176378809, |
| "learning_rate": 3.0162785381736893e-06, |
| "loss": 0.4731, |
| "step": 406 |
| }, |
| { |
| "epoch": 0.6324786324786325, |
| "grad_norm": 0.8148615875421817, |
| "learning_rate": 2.9938780466962768e-06, |
| "loss": 0.5124, |
| "step": 407 |
| }, |
| { |
| "epoch": 0.634032634032634, |
| "grad_norm": 0.629279814363528, |
| "learning_rate": 2.9715254440532147e-06, |
| "loss": 0.5542, |
| "step": 408 |
| }, |
| { |
| "epoch": 0.6355866355866356, |
| "grad_norm": 0.7893368478824807, |
| "learning_rate": 2.9492212638312458e-06, |
| "loss": 0.4712, |
| "step": 409 |
| }, |
| { |
| "epoch": 0.6371406371406372, |
| "grad_norm": 0.700072426705422, |
| "learning_rate": 2.9269660384612064e-06, |
| "loss": 0.5445, |
| "step": 410 |
| }, |
| { |
| "epoch": 0.6386946386946387, |
| "grad_norm": 0.6730347447205193, |
| "learning_rate": 2.90476029920532e-06, |
| "loss": 0.569, |
| "step": 411 |
| }, |
| { |
| "epoch": 0.6402486402486403, |
| "grad_norm": 1.8040910895752915, |
| "learning_rate": 2.882604576144505e-06, |
| "loss": 0.5713, |
| "step": 412 |
| }, |
| { |
| "epoch": 0.6418026418026418, |
| "grad_norm": 0.6061113607030476, |
| "learning_rate": 2.8604993981657247e-06, |
| "loss": 0.4493, |
| "step": 413 |
| }, |
| { |
| "epoch": 0.6433566433566433, |
| "grad_norm": 1.350624074574079, |
| "learning_rate": 2.8384452929493645e-06, |
| "loss": 0.4243, |
| "step": 414 |
| }, |
| { |
| "epoch": 0.6449106449106449, |
| "grad_norm": 1.020418478286547, |
| "learning_rate": 2.8164427869566367e-06, |
| "loss": 0.4695, |
| "step": 415 |
| }, |
| { |
| "epoch": 0.6464646464646465, |
| "grad_norm": 1.073163390573267, |
| "learning_rate": 2.7944924054170087e-06, |
| "loss": 0.4878, |
| "step": 416 |
| }, |
| { |
| "epoch": 0.6480186480186481, |
| "grad_norm": 0.8939568549949852, |
| "learning_rate": 2.7725946723156626e-06, |
| "loss": 0.564, |
| "step": 417 |
| }, |
| { |
| "epoch": 0.6495726495726496, |
| "grad_norm": 0.7664017459877794, |
| "learning_rate": 2.750750110381001e-06, |
| "loss": 0.499, |
| "step": 418 |
| }, |
| { |
| "epoch": 0.6511266511266511, |
| "grad_norm": 0.8801185511706636, |
| "learning_rate": 2.728959241072149e-06, |
| "loss": 0.4769, |
| "step": 419 |
| }, |
| { |
| "epoch": 0.6526806526806527, |
| "grad_norm": 0.645638892422625, |
| "learning_rate": 2.7072225845665256e-06, |
| "loss": 0.5266, |
| "step": 420 |
| }, |
| { |
| "epoch": 0.6542346542346542, |
| "grad_norm": 0.8936830405965801, |
| "learning_rate": 2.6855406597474098e-06, |
| "loss": 0.5236, |
| "step": 421 |
| }, |
| { |
| "epoch": 0.6557886557886557, |
| "grad_norm": 0.8974442709328094, |
| "learning_rate": 2.6639139841915628e-06, |
| "loss": 0.5072, |
| "step": 422 |
| }, |
| { |
| "epoch": 0.6573426573426573, |
| "grad_norm": 0.8176617831786887, |
| "learning_rate": 2.6423430741568746e-06, |
| "loss": 0.6122, |
| "step": 423 |
| }, |
| { |
| "epoch": 0.6588966588966589, |
| "grad_norm": 1.027696072983406, |
| "learning_rate": 2.6208284445700373e-06, |
| "loss": 0.5373, |
| "step": 424 |
| }, |
| { |
| "epoch": 0.6604506604506605, |
| "grad_norm": 1.254318562097477, |
| "learning_rate": 2.5993706090142484e-06, |
| "loss": 0.5659, |
| "step": 425 |
| }, |
| { |
| "epoch": 0.662004662004662, |
| "grad_norm": 0.9412050785188167, |
| "learning_rate": 2.577970079716959e-06, |
| "loss": 0.6532, |
| "step": 426 |
| }, |
| { |
| "epoch": 0.6635586635586636, |
| "grad_norm": 0.669172885578478, |
| "learning_rate": 2.5566273675376386e-06, |
| "loss": 0.4321, |
| "step": 427 |
| }, |
| { |
| "epoch": 0.6651126651126651, |
| "grad_norm": 0.7839761385934717, |
| "learning_rate": 2.535342981955591e-06, |
| "loss": 0.5029, |
| "step": 428 |
| }, |
| { |
| "epoch": 0.6666666666666666, |
| "grad_norm": 0.6291916677121909, |
| "learning_rate": 2.5141174310577774e-06, |
| "loss": 0.5619, |
| "step": 429 |
| }, |
| { |
| "epoch": 0.6682206682206682, |
| "grad_norm": 0.6009527872848712, |
| "learning_rate": 2.492951221526705e-06, |
| "loss": 0.4838, |
| "step": 430 |
| }, |
| { |
| "epoch": 0.6697746697746698, |
| "grad_norm": 0.7557317812271531, |
| "learning_rate": 2.4718448586283126e-06, |
| "loss": 0.5686, |
| "step": 431 |
| }, |
| { |
| "epoch": 0.6713286713286714, |
| "grad_norm": 0.6594827809287767, |
| "learning_rate": 2.4507988461999283e-06, |
| "loss": 0.4434, |
| "step": 432 |
| }, |
| { |
| "epoch": 0.6728826728826729, |
| "grad_norm": 0.6721258490052082, |
| "learning_rate": 2.429813686638227e-06, |
| "loss": 0.5448, |
| "step": 433 |
| }, |
| { |
| "epoch": 0.6744366744366744, |
| "grad_norm": 0.6253425210970757, |
| "learning_rate": 2.408889880887246e-06, |
| "loss": 0.5399, |
| "step": 434 |
| }, |
| { |
| "epoch": 0.675990675990676, |
| "grad_norm": 0.7132068704320651, |
| "learning_rate": 2.38802792842642e-06, |
| "loss": 0.4131, |
| "step": 435 |
| }, |
| { |
| "epoch": 0.6775446775446775, |
| "grad_norm": 0.9992884448946338, |
| "learning_rate": 2.3672283272586745e-06, |
| "loss": 0.6681, |
| "step": 436 |
| }, |
| { |
| "epoch": 0.679098679098679, |
| "grad_norm": 0.6736697588900878, |
| "learning_rate": 2.346491573898513e-06, |
| "loss": 0.5672, |
| "step": 437 |
| }, |
| { |
| "epoch": 0.6806526806526807, |
| "grad_norm": 0.8547599476784933, |
| "learning_rate": 2.3258181633601836e-06, |
| "loss": 0.5186, |
| "step": 438 |
| }, |
| { |
| "epoch": 0.6822066822066822, |
| "grad_norm": 0.7608589363389933, |
| "learning_rate": 2.30520858914585e-06, |
| "loss": 0.4367, |
| "step": 439 |
| }, |
| { |
| "epoch": 0.6837606837606838, |
| "grad_norm": 0.6900628130745625, |
| "learning_rate": 2.2846633432338256e-06, |
| "loss": 0.5083, |
| "step": 440 |
| }, |
| { |
| "epoch": 0.6853146853146853, |
| "grad_norm": 0.7581972429614234, |
| "learning_rate": 2.2641829160668137e-06, |
| "loss": 0.4485, |
| "step": 441 |
| }, |
| { |
| "epoch": 0.6868686868686869, |
| "grad_norm": 0.7973753083321776, |
| "learning_rate": 2.243767796540207e-06, |
| "loss": 0.4911, |
| "step": 442 |
| }, |
| { |
| "epoch": 0.6884226884226884, |
| "grad_norm": 0.8430463031780523, |
| "learning_rate": 2.223418471990421e-06, |
| "loss": 0.6334, |
| "step": 443 |
| }, |
| { |
| "epoch": 0.6899766899766899, |
| "grad_norm": 0.6780160144046566, |
| "learning_rate": 2.2031354281832555e-06, |
| "loss": 0.4788, |
| "step": 444 |
| }, |
| { |
| "epoch": 0.6915306915306916, |
| "grad_norm": 0.8261099182920753, |
| "learning_rate": 2.1829191493022974e-06, |
| "loss": 0.5132, |
| "step": 445 |
| }, |
| { |
| "epoch": 0.6930846930846931, |
| "grad_norm": 0.9687007131393034, |
| "learning_rate": 2.1627701179373645e-06, |
| "loss": 0.5751, |
| "step": 446 |
| }, |
| { |
| "epoch": 0.6946386946386947, |
| "grad_norm": 0.6461088689056241, |
| "learning_rate": 2.142688815072986e-06, |
| "loss": 0.5866, |
| "step": 447 |
| }, |
| { |
| "epoch": 0.6961926961926962, |
| "grad_norm": 0.8088646308620774, |
| "learning_rate": 2.1226757200769225e-06, |
| "loss": 0.6883, |
| "step": 448 |
| }, |
| { |
| "epoch": 0.6977466977466977, |
| "grad_norm": 1.6821696576244451, |
| "learning_rate": 2.102731310688723e-06, |
| "loss": 0.5542, |
| "step": 449 |
| }, |
| { |
| "epoch": 0.6993006993006993, |
| "grad_norm": 0.6856586399011556, |
| "learning_rate": 2.0828560630083127e-06, |
| "loss": 0.3613, |
| "step": 450 |
| }, |
| { |
| "epoch": 0.7008547008547008, |
| "grad_norm": 0.8799640778401495, |
| "learning_rate": 2.0630504514846372e-06, |
| "loss": 0.6021, |
| "step": 451 |
| }, |
| { |
| "epoch": 0.7024087024087025, |
| "grad_norm": 1.0346792136769154, |
| "learning_rate": 2.0433149489043296e-06, |
| "loss": 0.6368, |
| "step": 452 |
| }, |
| { |
| "epoch": 0.703962703962704, |
| "grad_norm": 0.7233064199559888, |
| "learning_rate": 2.0236500263804355e-06, |
| "loss": 0.5101, |
| "step": 453 |
| }, |
| { |
| "epoch": 0.7055167055167055, |
| "grad_norm": 0.6101595125764867, |
| "learning_rate": 2.0040561533411494e-06, |
| "loss": 0.4283, |
| "step": 454 |
| }, |
| { |
| "epoch": 0.7070707070707071, |
| "grad_norm": 0.6838739790158097, |
| "learning_rate": 1.9845337975186297e-06, |
| "loss": 0.4044, |
| "step": 455 |
| }, |
| { |
| "epoch": 0.7086247086247086, |
| "grad_norm": 0.6512670498735073, |
| "learning_rate": 1.9650834249378125e-06, |
| "loss": 0.5122, |
| "step": 456 |
| }, |
| { |
| "epoch": 0.7101787101787101, |
| "grad_norm": 0.7448486863111607, |
| "learning_rate": 1.945705499905305e-06, |
| "loss": 0.4664, |
| "step": 457 |
| }, |
| { |
| "epoch": 0.7117327117327117, |
| "grad_norm": 0.6714430637930826, |
| "learning_rate": 1.926400484998289e-06, |
| "loss": 0.4605, |
| "step": 458 |
| }, |
| { |
| "epoch": 0.7132867132867133, |
| "grad_norm": 0.7034702671220221, |
| "learning_rate": 1.9071688410534828e-06, |
| "loss": 0.4332, |
| "step": 459 |
| }, |
| { |
| "epoch": 0.7148407148407149, |
| "grad_norm": 0.920215915504864, |
| "learning_rate": 1.8880110271561415e-06, |
| "loss": 0.582, |
| "step": 460 |
| }, |
| { |
| "epoch": 0.7163947163947164, |
| "grad_norm": 0.7704916182387226, |
| "learning_rate": 1.8689275006291035e-06, |
| "loss": 0.3934, |
| "step": 461 |
| }, |
| { |
| "epoch": 0.717948717948718, |
| "grad_norm": 0.7275954100685235, |
| "learning_rate": 1.8499187170218614e-06, |
| "loss": 0.6786, |
| "step": 462 |
| }, |
| { |
| "epoch": 0.7195027195027195, |
| "grad_norm": 0.9130677264258454, |
| "learning_rate": 1.8309851300996934e-06, |
| "loss": 0.4669, |
| "step": 463 |
| }, |
| { |
| "epoch": 0.721056721056721, |
| "grad_norm": 0.627613808696502, |
| "learning_rate": 1.8121271918328314e-06, |
| "loss": 0.3791, |
| "step": 464 |
| }, |
| { |
| "epoch": 0.7226107226107226, |
| "grad_norm": 0.7966353111195642, |
| "learning_rate": 1.793345352385678e-06, |
| "loss": 0.5131, |
| "step": 465 |
| }, |
| { |
| "epoch": 0.7241647241647242, |
| "grad_norm": 0.7257869097486319, |
| "learning_rate": 1.7746400601060476e-06, |
| "loss": 0.4698, |
| "step": 466 |
| }, |
| { |
| "epoch": 0.7257187257187258, |
| "grad_norm": 0.8274666640783022, |
| "learning_rate": 1.7560117615144717e-06, |
| "loss": 0.5127, |
| "step": 467 |
| }, |
| { |
| "epoch": 0.7272727272727273, |
| "grad_norm": 0.919410912868769, |
| "learning_rate": 1.7374609012935412e-06, |
| "loss": 0.5077, |
| "step": 468 |
| }, |
| { |
| "epoch": 0.7288267288267288, |
| "grad_norm": 0.8104780902360221, |
| "learning_rate": 1.7189879222772894e-06, |
| "loss": 0.4689, |
| "step": 469 |
| }, |
| { |
| "epoch": 0.7303807303807304, |
| "grad_norm": 0.7461709338199918, |
| "learning_rate": 1.7005932654406165e-06, |
| "loss": 0.3507, |
| "step": 470 |
| }, |
| { |
| "epoch": 0.7319347319347319, |
| "grad_norm": 0.6643952482627672, |
| "learning_rate": 1.682277369888769e-06, |
| "loss": 0.4538, |
| "step": 471 |
| }, |
| { |
| "epoch": 0.7334887334887334, |
| "grad_norm": 0.8395310070807439, |
| "learning_rate": 1.6640406728468534e-06, |
| "loss": 0.4401, |
| "step": 472 |
| }, |
| { |
| "epoch": 0.7350427350427351, |
| "grad_norm": 0.9088380178834886, |
| "learning_rate": 1.6458836096494046e-06, |
| "loss": 0.6829, |
| "step": 473 |
| }, |
| { |
| "epoch": 0.7365967365967366, |
| "grad_norm": 0.8158423552015206, |
| "learning_rate": 1.6278066137299898e-06, |
| "loss": 0.5124, |
| "step": 474 |
| }, |
| { |
| "epoch": 0.7381507381507382, |
| "grad_norm": 0.6628148397288011, |
| "learning_rate": 1.6098101166108593e-06, |
| "loss": 0.4556, |
| "step": 475 |
| }, |
| { |
| "epoch": 0.7397047397047397, |
| "grad_norm": 0.7157818223192728, |
| "learning_rate": 1.5918945478926484e-06, |
| "loss": 0.5349, |
| "step": 476 |
| }, |
| { |
| "epoch": 0.7412587412587412, |
| "grad_norm": 0.9140000377609874, |
| "learning_rate": 1.5740603352441281e-06, |
| "loss": 0.6558, |
| "step": 477 |
| }, |
| { |
| "epoch": 0.7428127428127428, |
| "grad_norm": 1.8219839216709126, |
| "learning_rate": 1.5563079043919843e-06, |
| "loss": 0.6314, |
| "step": 478 |
| }, |
| { |
| "epoch": 0.7443667443667443, |
| "grad_norm": 0.6702786236421273, |
| "learning_rate": 1.5386376791106627e-06, |
| "loss": 0.4835, |
| "step": 479 |
| }, |
| { |
| "epoch": 0.745920745920746, |
| "grad_norm": 0.8589486289983319, |
| "learning_rate": 1.5210500812122548e-06, |
| "loss": 0.28, |
| "step": 480 |
| }, |
| { |
| "epoch": 0.7474747474747475, |
| "grad_norm": 0.6696692189402571, |
| "learning_rate": 1.5035455305364188e-06, |
| "loss": 0.4483, |
| "step": 481 |
| }, |
| { |
| "epoch": 0.749028749028749, |
| "grad_norm": 0.832181293716284, |
| "learning_rate": 1.4861244449403717e-06, |
| "loss": 0.5884, |
| "step": 482 |
| }, |
| { |
| "epoch": 0.7505827505827506, |
| "grad_norm": 0.9644123682005706, |
| "learning_rate": 1.4687872402888991e-06, |
| "loss": 0.4757, |
| "step": 483 |
| }, |
| { |
| "epoch": 0.7521367521367521, |
| "grad_norm": 0.8646333036074157, |
| "learning_rate": 1.451534330444438e-06, |
| "loss": 0.4525, |
| "step": 484 |
| }, |
| { |
| "epoch": 0.7536907536907537, |
| "grad_norm": 0.7543556415051755, |
| "learning_rate": 1.4343661272571967e-06, |
| "loss": 0.453, |
| "step": 485 |
| }, |
| { |
| "epoch": 0.7552447552447552, |
| "grad_norm": 0.7262516356092494, |
| "learning_rate": 1.4172830405553216e-06, |
| "loss": 0.3922, |
| "step": 486 |
| }, |
| { |
| "epoch": 0.7567987567987567, |
| "grad_norm": 0.7925066399718954, |
| "learning_rate": 1.4002854781351104e-06, |
| "loss": 0.517, |
| "step": 487 |
| }, |
| { |
| "epoch": 0.7583527583527584, |
| "grad_norm": 0.7148060275837002, |
| "learning_rate": 1.3833738457512842e-06, |
| "loss": 0.3847, |
| "step": 488 |
| }, |
| { |
| "epoch": 0.7599067599067599, |
| "grad_norm": 0.6309982281142955, |
| "learning_rate": 1.3665485471072937e-06, |
| "loss": 0.4778, |
| "step": 489 |
| }, |
| { |
| "epoch": 0.7614607614607615, |
| "grad_norm": 0.7642090785940528, |
| "learning_rate": 1.3498099838456947e-06, |
| "loss": 0.4309, |
| "step": 490 |
| }, |
| { |
| "epoch": 0.763014763014763, |
| "grad_norm": 0.7640039314162583, |
| "learning_rate": 1.3331585555385458e-06, |
| "loss": 0.4696, |
| "step": 491 |
| }, |
| { |
| "epoch": 0.7645687645687645, |
| "grad_norm": 0.8643751125446875, |
| "learning_rate": 1.3165946596778773e-06, |
| "loss": 0.4918, |
| "step": 492 |
| }, |
| { |
| "epoch": 0.7661227661227661, |
| "grad_norm": 0.6804715835256383, |
| "learning_rate": 1.3001186916662066e-06, |
| "loss": 0.4724, |
| "step": 493 |
| }, |
| { |
| "epoch": 0.7676767676767676, |
| "grad_norm": 1.0793870282462463, |
| "learning_rate": 1.2837310448070929e-06, |
| "loss": 0.5909, |
| "step": 494 |
| }, |
| { |
| "epoch": 0.7692307692307693, |
| "grad_norm": 0.7827368042523198, |
| "learning_rate": 1.2674321102957476e-06, |
| "loss": 0.5138, |
| "step": 495 |
| }, |
| { |
| "epoch": 0.7707847707847708, |
| "grad_norm": 0.847544394369445, |
| "learning_rate": 1.251222277209702e-06, |
| "loss": 0.603, |
| "step": 496 |
| }, |
| { |
| "epoch": 0.7723387723387724, |
| "grad_norm": 0.9723527581718034, |
| "learning_rate": 1.2351019324995128e-06, |
| "loss": 0.6304, |
| "step": 497 |
| }, |
| { |
| "epoch": 0.7738927738927739, |
| "grad_norm": 0.6468908302032275, |
| "learning_rate": 1.2190714609795334e-06, |
| "loss": 0.521, |
| "step": 498 |
| }, |
| { |
| "epoch": 0.7754467754467754, |
| "grad_norm": 0.7114004263355455, |
| "learning_rate": 1.203131245318721e-06, |
| "loss": 0.4607, |
| "step": 499 |
| }, |
| { |
| "epoch": 0.777000777000777, |
| "grad_norm": 0.6224249787588289, |
| "learning_rate": 1.1872816660315029e-06, |
| "loss": 0.3998, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.7785547785547785, |
| "grad_norm": 0.8157807813857092, |
| "learning_rate": 1.171523101468693e-06, |
| "loss": 0.5267, |
| "step": 501 |
| }, |
| { |
| "epoch": 0.7801087801087802, |
| "grad_norm": 0.799774654796459, |
| "learning_rate": 1.1558559278084647e-06, |
| "loss": 0.4909, |
| "step": 502 |
| }, |
| { |
| "epoch": 0.7816627816627817, |
| "grad_norm": 0.712975958103021, |
| "learning_rate": 1.1402805190473649e-06, |
| "loss": 0.4833, |
| "step": 503 |
| }, |
| { |
| "epoch": 0.7832167832167832, |
| "grad_norm": 0.7414921854529274, |
| "learning_rate": 1.124797246991387e-06, |
| "loss": 0.4083, |
| "step": 504 |
| }, |
| { |
| "epoch": 0.7847707847707848, |
| "grad_norm": 1.0834283945062337, |
| "learning_rate": 1.1094064812471028e-06, |
| "loss": 0.5327, |
| "step": 505 |
| }, |
| { |
| "epoch": 0.7863247863247863, |
| "grad_norm": 0.7386660114893862, |
| "learning_rate": 1.0941085892128272e-06, |
| "loss": 0.4922, |
| "step": 506 |
| }, |
| { |
| "epoch": 0.7878787878787878, |
| "grad_norm": 0.5879874025861683, |
| "learning_rate": 1.07890393606986e-06, |
| "loss": 0.3765, |
| "step": 507 |
| }, |
| { |
| "epoch": 0.7894327894327894, |
| "grad_norm": 0.8290075934295431, |
| "learning_rate": 1.0637928847737594e-06, |
| "loss": 0.5165, |
| "step": 508 |
| }, |
| { |
| "epoch": 0.790986790986791, |
| "grad_norm": 0.9222992948806642, |
| "learning_rate": 1.0487757960456812e-06, |
| "loss": 0.5239, |
| "step": 509 |
| }, |
| { |
| "epoch": 0.7925407925407926, |
| "grad_norm": 0.894510374178912, |
| "learning_rate": 1.0338530283637704e-06, |
| "loss": 0.5294, |
| "step": 510 |
| }, |
| { |
| "epoch": 0.7940947940947941, |
| "grad_norm": 0.9067683678931795, |
| "learning_rate": 1.0190249379546024e-06, |
| "loss": 0.6271, |
| "step": 511 |
| }, |
| { |
| "epoch": 0.7956487956487956, |
| "grad_norm": 0.8420348253099458, |
| "learning_rate": 1.0042918787846757e-06, |
| "loss": 0.4476, |
| "step": 512 |
| }, |
| { |
| "epoch": 0.7972027972027972, |
| "grad_norm": 2.2314012317895555, |
| "learning_rate": 9.896542025519645e-07, |
| "loss": 0.4875, |
| "step": 513 |
| }, |
| { |
| "epoch": 0.7987567987567987, |
| "grad_norm": 1.112553561220701, |
| "learning_rate": 9.751122586775253e-07, |
| "loss": 0.4668, |
| "step": 514 |
| }, |
| { |
| "epoch": 0.8003108003108003, |
| "grad_norm": 0.8631165830420672, |
| "learning_rate": 9.606663942971568e-07, |
| "loss": 0.4823, |
| "step": 515 |
| }, |
| { |
| "epoch": 0.8018648018648019, |
| "grad_norm": 0.8259570757220709, |
| "learning_rate": 9.463169542531059e-07, |
| "loss": 0.5339, |
| "step": 516 |
| }, |
| { |
| "epoch": 0.8034188034188035, |
| "grad_norm": 1.166843823225754, |
| "learning_rate": 9.320642810858421e-07, |
| "loss": 0.5133, |
| "step": 517 |
| }, |
| { |
| "epoch": 0.804972804972805, |
| "grad_norm": 0.9069458506573143, |
| "learning_rate": 9.179087150258814e-07, |
| "loss": 0.4574, |
| "step": 518 |
| }, |
| { |
| "epoch": 0.8065268065268065, |
| "grad_norm": 0.6083374762695839, |
| "learning_rate": 9.038505939856612e-07, |
| "loss": 0.5352, |
| "step": 519 |
| }, |
| { |
| "epoch": 0.8080808080808081, |
| "grad_norm": 0.7051533869734491, |
| "learning_rate": 8.898902535514747e-07, |
| "loss": 0.486, |
| "step": 520 |
| }, |
| { |
| "epoch": 0.8096348096348096, |
| "grad_norm": 1.1361715723258188, |
| "learning_rate": 8.76028026975459e-07, |
| "loss": 0.5395, |
| "step": 521 |
| }, |
| { |
| "epoch": 0.8111888111888111, |
| "grad_norm": 0.7251363387868152, |
| "learning_rate": 8.62264245167641e-07, |
| "loss": 0.4734, |
| "step": 522 |
| }, |
| { |
| "epoch": 0.8127428127428128, |
| "grad_norm": 0.7015732791695378, |
| "learning_rate": 8.485992366880419e-07, |
| "loss": 0.5281, |
| "step": 523 |
| }, |
| { |
| "epoch": 0.8142968142968143, |
| "grad_norm": 1.5424433933643296, |
| "learning_rate": 8.35033327738829e-07, |
| "loss": 0.4043, |
| "step": 524 |
| }, |
| { |
| "epoch": 0.8158508158508159, |
| "grad_norm": 0.9218463524060198, |
| "learning_rate": 8.215668421565276e-07, |
| "loss": 0.5614, |
| "step": 525 |
| }, |
| { |
| "epoch": 0.8174048174048174, |
| "grad_norm": 0.764325971788183, |
| "learning_rate": 8.082001014042945e-07, |
| "loss": 0.5032, |
| "step": 526 |
| }, |
| { |
| "epoch": 0.818958818958819, |
| "grad_norm": 0.652486590264396, |
| "learning_rate": 7.949334245642459e-07, |
| "loss": 0.4544, |
| "step": 527 |
| }, |
| { |
| "epoch": 0.8205128205128205, |
| "grad_norm": 0.7176940824971066, |
| "learning_rate": 7.817671283298345e-07, |
| "loss": 0.5098, |
| "step": 528 |
| }, |
| { |
| "epoch": 0.822066822066822, |
| "grad_norm": 1.823594644970129, |
| "learning_rate": 7.687015269982917e-07, |
| "loss": 0.4841, |
| "step": 529 |
| }, |
| { |
| "epoch": 0.8236208236208237, |
| "grad_norm": 0.7669592956201216, |
| "learning_rate": 7.557369324631303e-07, |
| "loss": 0.5751, |
| "step": 530 |
| }, |
| { |
| "epoch": 0.8251748251748252, |
| "grad_norm": 0.7791697796343787, |
| "learning_rate": 7.4287365420669e-07, |
| "loss": 0.5334, |
| "step": 531 |
| }, |
| { |
| "epoch": 0.8267288267288267, |
| "grad_norm": 0.7675193579164078, |
| "learning_rate": 7.301119992927585e-07, |
| "loss": 0.4749, |
| "step": 532 |
| }, |
| { |
| "epoch": 0.8282828282828283, |
| "grad_norm": 0.8274131519826066, |
| "learning_rate": 7.174522723592342e-07, |
| "loss": 0.5556, |
| "step": 533 |
| }, |
| { |
| "epoch": 0.8298368298368298, |
| "grad_norm": 0.6248155156332447, |
| "learning_rate": 7.048947756108576e-07, |
| "loss": 0.4449, |
| "step": 534 |
| }, |
| { |
| "epoch": 0.8313908313908314, |
| "grad_norm": 0.764818832164719, |
| "learning_rate": 6.924398088119988e-07, |
| "loss": 0.5908, |
| "step": 535 |
| }, |
| { |
| "epoch": 0.8329448329448329, |
| "grad_norm": 0.7925584720042309, |
| "learning_rate": 6.800876692794994e-07, |
| "loss": 0.5557, |
| "step": 536 |
| }, |
| { |
| "epoch": 0.8344988344988346, |
| "grad_norm": 0.8639156535546165, |
| "learning_rate": 6.678386518755747e-07, |
| "loss": 0.5577, |
| "step": 537 |
| }, |
| { |
| "epoch": 0.8360528360528361, |
| "grad_norm": 0.8390195478931005, |
| "learning_rate": 6.556930490007762e-07, |
| "loss": 0.5484, |
| "step": 538 |
| }, |
| { |
| "epoch": 0.8376068376068376, |
| "grad_norm": 0.63339201522811, |
| "learning_rate": 6.436511505870091e-07, |
| "loss": 0.5014, |
| "step": 539 |
| }, |
| { |
| "epoch": 0.8391608391608392, |
| "grad_norm": 0.889532359805588, |
| "learning_rate": 6.317132440906188e-07, |
| "loss": 0.5266, |
| "step": 540 |
| }, |
| { |
| "epoch": 0.8407148407148407, |
| "grad_norm": 0.5624920627332638, |
| "learning_rate": 6.198796144855168e-07, |
| "loss": 0.4393, |
| "step": 541 |
| }, |
| { |
| "epoch": 0.8422688422688422, |
| "grad_norm": 0.6994504769690357, |
| "learning_rate": 6.081505442563912e-07, |
| "loss": 0.4758, |
| "step": 542 |
| }, |
| { |
| "epoch": 0.8438228438228438, |
| "grad_norm": 0.677345489523822, |
| "learning_rate": 5.965263133919508e-07, |
| "loss": 0.4376, |
| "step": 543 |
| }, |
| { |
| "epoch": 0.8453768453768454, |
| "grad_norm": 1.1792412849625011, |
| "learning_rate": 5.850071993782525e-07, |
| "loss": 0.557, |
| "step": 544 |
| }, |
| { |
| "epoch": 0.846930846930847, |
| "grad_norm": 0.8383140877354333, |
| "learning_rate": 5.735934771920704e-07, |
| "loss": 0.4129, |
| "step": 545 |
| }, |
| { |
| "epoch": 0.8484848484848485, |
| "grad_norm": 0.7877126864302514, |
| "learning_rate": 5.622854192943317e-07, |
| "loss": 0.5008, |
| "step": 546 |
| }, |
| { |
| "epoch": 0.85003885003885, |
| "grad_norm": 0.9108037190809487, |
| "learning_rate": 5.510832956236173e-07, |
| "loss": 0.4783, |
| "step": 547 |
| }, |
| { |
| "epoch": 0.8515928515928516, |
| "grad_norm": 1.0094511865353648, |
| "learning_rate": 5.399873735897137e-07, |
| "loss": 0.4469, |
| "step": 548 |
| }, |
| { |
| "epoch": 0.8531468531468531, |
| "grad_norm": 0.663812075111233, |
| "learning_rate": 5.289979180672344e-07, |
| "loss": 0.6368, |
| "step": 549 |
| }, |
| { |
| "epoch": 0.8547008547008547, |
| "grad_norm": 0.7540552593487783, |
| "learning_rate": 5.181151913892896e-07, |
| "loss": 0.4343, |
| "step": 550 |
| }, |
| { |
| "epoch": 0.8562548562548562, |
| "grad_norm": 1.0037008402911451, |
| "learning_rate": 5.073394533412296e-07, |
| "loss": 0.5469, |
| "step": 551 |
| }, |
| { |
| "epoch": 0.8578088578088578, |
| "grad_norm": 0.5605077409329356, |
| "learning_rate": 4.966709611544435e-07, |
| "loss": 0.4462, |
| "step": 552 |
| }, |
| { |
| "epoch": 0.8593628593628594, |
| "grad_norm": 0.8732352825291733, |
| "learning_rate": 4.861099695002158e-07, |
| "loss": 0.5251, |
| "step": 553 |
| }, |
| { |
| "epoch": 0.8609168609168609, |
| "grad_norm": 0.7053119794249835, |
| "learning_rate": 4.7565673048364735e-07, |
| "loss": 0.4842, |
| "step": 554 |
| }, |
| { |
| "epoch": 0.8624708624708625, |
| "grad_norm": 0.821971296932637, |
| "learning_rate": 4.6531149363764126e-07, |
| "loss": 0.4487, |
| "step": 555 |
| }, |
| { |
| "epoch": 0.864024864024864, |
| "grad_norm": 1.2794248867833895, |
| "learning_rate": 4.550745059169398e-07, |
| "loss": 0.5328, |
| "step": 556 |
| }, |
| { |
| "epoch": 0.8655788655788655, |
| "grad_norm": 0.7489672108536733, |
| "learning_rate": 4.4494601169223715e-07, |
| "loss": 0.5069, |
| "step": 557 |
| }, |
| { |
| "epoch": 0.8671328671328671, |
| "grad_norm": 0.6659188193571587, |
| "learning_rate": 4.349262527443371e-07, |
| "loss": 0.5734, |
| "step": 558 |
| }, |
| { |
| "epoch": 0.8686868686868687, |
| "grad_norm": 0.7454236276103242, |
| "learning_rate": 4.2501546825838735e-07, |
| "loss": 0.358, |
| "step": 559 |
| }, |
| { |
| "epoch": 0.8702408702408703, |
| "grad_norm": 0.9570678482900348, |
| "learning_rate": 4.152138948181689e-07, |
| "loss": 0.538, |
| "step": 560 |
| }, |
| { |
| "epoch": 0.8717948717948718, |
| "grad_norm": 0.9593262810731289, |
| "learning_rate": 4.0552176640045017e-07, |
| "loss": 0.6003, |
| "step": 561 |
| }, |
| { |
| "epoch": 0.8733488733488733, |
| "grad_norm": 0.6814712625667729, |
| "learning_rate": 3.959393143693946e-07, |
| "loss": 0.4296, |
| "step": 562 |
| }, |
| { |
| "epoch": 0.8749028749028749, |
| "grad_norm": 0.8468228674134194, |
| "learning_rate": 3.864667674710454e-07, |
| "loss": 0.478, |
| "step": 563 |
| }, |
| { |
| "epoch": 0.8764568764568764, |
| "grad_norm": 0.8214210844483506, |
| "learning_rate": 3.7710435182786053e-07, |
| "loss": 0.4921, |
| "step": 564 |
| }, |
| { |
| "epoch": 0.878010878010878, |
| "grad_norm": 0.7775961282959882, |
| "learning_rate": 3.6785229093331987e-07, |
| "loss": 0.5158, |
| "step": 565 |
| }, |
| { |
| "epoch": 0.8795648795648796, |
| "grad_norm": 0.7706689173433338, |
| "learning_rate": 3.587108056465827e-07, |
| "loss": 0.4987, |
| "step": 566 |
| }, |
| { |
| "epoch": 0.8811188811188811, |
| "grad_norm": 0.7115920841531194, |
| "learning_rate": 3.496801141872225e-07, |
| "loss": 0.5508, |
| "step": 567 |
| }, |
| { |
| "epoch": 0.8826728826728827, |
| "grad_norm": 0.7360248498356703, |
| "learning_rate": 3.407604321300123e-07, |
| "loss": 0.4622, |
| "step": 568 |
| }, |
| { |
| "epoch": 0.8842268842268842, |
| "grad_norm": 0.6656952275189396, |
| "learning_rate": 3.3195197239978384e-07, |
| "loss": 0.4616, |
| "step": 569 |
| }, |
| { |
| "epoch": 0.8857808857808858, |
| "grad_norm": 1.0438892609670738, |
| "learning_rate": 3.232549452663403e-07, |
| "loss": 0.6725, |
| "step": 570 |
| }, |
| { |
| "epoch": 0.8873348873348873, |
| "grad_norm": 0.9087915959905298, |
| "learning_rate": 3.146695583394377e-07, |
| "loss": 0.5397, |
| "step": 571 |
| }, |
| { |
| "epoch": 0.8888888888888888, |
| "grad_norm": 0.7459812770941501, |
| "learning_rate": 3.061960165638317e-07, |
| "loss": 0.4763, |
| "step": 572 |
| }, |
| { |
| "epoch": 0.8904428904428905, |
| "grad_norm": 0.9368206700984494, |
| "learning_rate": 2.9783452221438304e-07, |
| "loss": 0.4894, |
| "step": 573 |
| }, |
| { |
| "epoch": 0.891996891996892, |
| "grad_norm": 0.7581701388162561, |
| "learning_rate": 2.895852748912298e-07, |
| "loss": 0.4285, |
| "step": 574 |
| }, |
| { |
| "epoch": 0.8935508935508936, |
| "grad_norm": 0.7295548456707804, |
| "learning_rate": 2.814484715150212e-07, |
| "loss": 0.4201, |
| "step": 575 |
| }, |
| { |
| "epoch": 0.8951048951048951, |
| "grad_norm": 0.6762252893935657, |
| "learning_rate": 2.734243063222181e-07, |
| "loss": 0.4842, |
| "step": 576 |
| }, |
| { |
| "epoch": 0.8966588966588966, |
| "grad_norm": 0.7273733934803849, |
| "learning_rate": 2.655129708604576e-07, |
| "loss": 0.3651, |
| "step": 577 |
| }, |
| { |
| "epoch": 0.8982128982128982, |
| "grad_norm": 0.9666761581157529, |
| "learning_rate": 2.5771465398397757e-07, |
| "loss": 0.7778, |
| "step": 578 |
| }, |
| { |
| "epoch": 0.8997668997668997, |
| "grad_norm": 0.6776269970027221, |
| "learning_rate": 2.5002954184910887e-07, |
| "loss": 0.5259, |
| "step": 579 |
| }, |
| { |
| "epoch": 0.9013209013209014, |
| "grad_norm": 0.5934838544460401, |
| "learning_rate": 2.424578179098358e-07, |
| "loss": 0.5417, |
| "step": 580 |
| }, |
| { |
| "epoch": 0.9028749028749029, |
| "grad_norm": 0.8173118127489211, |
| "learning_rate": 2.3499966291341213e-07, |
| "loss": 0.4431, |
| "step": 581 |
| }, |
| { |
| "epoch": 0.9044289044289044, |
| "grad_norm": 0.6926431286616653, |
| "learning_rate": 2.2765525489604702e-07, |
| "loss": 0.3946, |
| "step": 582 |
| }, |
| { |
| "epoch": 0.905982905982906, |
| "grad_norm": 0.8058462854601025, |
| "learning_rate": 2.2042476917865706e-07, |
| "loss": 0.4716, |
| "step": 583 |
| }, |
| { |
| "epoch": 0.9075369075369075, |
| "grad_norm": 0.735345958366895, |
| "learning_rate": 2.1330837836267882e-07, |
| "loss": 0.4664, |
| "step": 584 |
| }, |
| { |
| "epoch": 0.9090909090909091, |
| "grad_norm": 0.6326682057182371, |
| "learning_rate": 2.0630625232595126e-07, |
| "loss": 0.395, |
| "step": 585 |
| }, |
| { |
| "epoch": 0.9106449106449106, |
| "grad_norm": 0.7934491290140667, |
| "learning_rate": 1.9941855821865918e-07, |
| "loss": 0.4923, |
| "step": 586 |
| }, |
| { |
| "epoch": 0.9121989121989122, |
| "grad_norm": 0.7620750821145263, |
| "learning_rate": 1.9264546045934196e-07, |
| "loss": 0.4928, |
| "step": 587 |
| }, |
| { |
| "epoch": 0.9137529137529138, |
| "grad_norm": 0.6744132607863633, |
| "learning_rate": 1.859871207309688e-07, |
| "loss": 0.544, |
| "step": 588 |
| }, |
| { |
| "epoch": 0.9153069153069153, |
| "grad_norm": 0.850939097017279, |
| "learning_rate": 1.7944369797708362e-07, |
| "loss": 0.4553, |
| "step": 589 |
| }, |
| { |
| "epoch": 0.9168609168609169, |
| "grad_norm": 0.6424420190959574, |
| "learning_rate": 1.7301534839800348e-07, |
| "loss": 0.3915, |
| "step": 590 |
| }, |
| { |
| "epoch": 0.9184149184149184, |
| "grad_norm": 0.7046012463166428, |
| "learning_rate": 1.6670222544709515e-07, |
| "loss": 0.4708, |
| "step": 591 |
| }, |
| { |
| "epoch": 0.9199689199689199, |
| "grad_norm": 1.0364283887763577, |
| "learning_rate": 1.6050447982711214e-07, |
| "loss": 0.4111, |
| "step": 592 |
| }, |
| { |
| "epoch": 0.9215229215229215, |
| "grad_norm": 0.7049190344022591, |
| "learning_rate": 1.5442225948659183e-07, |
| "loss": 0.5696, |
| "step": 593 |
| }, |
| { |
| "epoch": 0.9230769230769231, |
| "grad_norm": 0.8155103152236225, |
| "learning_rate": 1.4845570961633192e-07, |
| "loss": 0.5059, |
| "step": 594 |
| }, |
| { |
| "epoch": 0.9246309246309247, |
| "grad_norm": 0.5949094353172898, |
| "learning_rate": 1.426049726459172e-07, |
| "loss": 0.4638, |
| "step": 595 |
| }, |
| { |
| "epoch": 0.9261849261849262, |
| "grad_norm": 0.6815208521528887, |
| "learning_rate": 1.368701882403234e-07, |
| "loss": 0.4551, |
| "step": 596 |
| }, |
| { |
| "epoch": 0.9277389277389277, |
| "grad_norm": 0.9296974247938309, |
| "learning_rate": 1.3125149329658083e-07, |
| "loss": 0.4026, |
| "step": 597 |
| }, |
| { |
| "epoch": 0.9292929292929293, |
| "grad_norm": 1.2038986087224606, |
| "learning_rate": 1.2574902194050996e-07, |
| "loss": 0.4656, |
| "step": 598 |
| }, |
| { |
| "epoch": 0.9308469308469308, |
| "grad_norm": 1.0581872023301784, |
| "learning_rate": 1.2036290552351838e-07, |
| "loss": 0.6128, |
| "step": 599 |
| }, |
| { |
| "epoch": 0.9324009324009324, |
| "grad_norm": 0.9158733284924282, |
| "learning_rate": 1.150932726194609e-07, |
| "loss": 0.4848, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.933954933954934, |
| "grad_norm": 0.7543888048930436, |
| "learning_rate": 1.0994024902157674e-07, |
| "loss": 0.4701, |
| "step": 601 |
| }, |
| { |
| "epoch": 0.9355089355089355, |
| "grad_norm": 0.7081866182048879, |
| "learning_rate": 1.0490395773948336e-07, |
| "loss": 0.5253, |
| "step": 602 |
| }, |
| { |
| "epoch": 0.9370629370629371, |
| "grad_norm": 0.6949901814621636, |
| "learning_rate": 9.998451899624007e-08, |
| "loss": 0.4287, |
| "step": 603 |
| }, |
| { |
| "epoch": 0.9386169386169386, |
| "grad_norm": 0.8441421847918561, |
| "learning_rate": 9.5182050225478e-08, |
| "loss": 0.5441, |
| "step": 604 |
| }, |
| { |
| "epoch": 0.9401709401709402, |
| "grad_norm": 0.7034318164083617, |
| "learning_rate": 9.049666606859852e-08, |
| "loss": 0.4834, |
| "step": 605 |
| }, |
| { |
| "epoch": 0.9417249417249417, |
| "grad_norm": 0.8895517754609739, |
| "learning_rate": 8.592847837203655e-08, |
| "loss": 0.4564, |
| "step": 606 |
| }, |
| { |
| "epoch": 0.9432789432789432, |
| "grad_norm": 0.7017208181125264, |
| "learning_rate": 8.147759618458706e-08, |
| "loss": 0.4238, |
| "step": 607 |
| }, |
| { |
| "epoch": 0.9448329448329449, |
| "grad_norm": 0.7595723199577458, |
| "learning_rate": 7.714412575480556e-08, |
| "loss": 0.541, |
| "step": 608 |
| }, |
| { |
| "epoch": 0.9463869463869464, |
| "grad_norm": 0.701349856315337, |
| "learning_rate": 7.292817052847068e-08, |
| "loss": 0.5784, |
| "step": 609 |
| }, |
| { |
| "epoch": 0.947940947940948, |
| "grad_norm": 0.6902057779636126, |
| "learning_rate": 6.882983114611497e-08, |
| "loss": 0.5877, |
| "step": 610 |
| }, |
| { |
| "epoch": 0.9494949494949495, |
| "grad_norm": 0.7053000221431716, |
| "learning_rate": 6.484920544062245e-08, |
| "loss": 0.3299, |
| "step": 611 |
| }, |
| { |
| "epoch": 0.951048951048951, |
| "grad_norm": 0.7668065733213358, |
| "learning_rate": 6.098638843489213e-08, |
| "loss": 0.6096, |
| "step": 612 |
| }, |
| { |
| "epoch": 0.9526029526029526, |
| "grad_norm": 0.9203339601219532, |
| "learning_rate": 5.7241472339572e-08, |
| "loss": 0.6287, |
| "step": 613 |
| }, |
| { |
| "epoch": 0.9541569541569541, |
| "grad_norm": 0.7745409282534764, |
| "learning_rate": 5.361454655085529e-08, |
| "loss": 0.4677, |
| "step": 614 |
| }, |
| { |
| "epoch": 0.9557109557109557, |
| "grad_norm": 0.7385280228665362, |
| "learning_rate": 5.0105697648347716e-08, |
| "loss": 0.5869, |
| "step": 615 |
| }, |
| { |
| "epoch": 0.9572649572649573, |
| "grad_norm": 0.7825960999337372, |
| "learning_rate": 4.671500939300133e-08, |
| "loss": 0.498, |
| "step": 616 |
| }, |
| { |
| "epoch": 0.9588189588189588, |
| "grad_norm": 0.753511461886924, |
| "learning_rate": 4.344256272511338e-08, |
| "loss": 0.4964, |
| "step": 617 |
| }, |
| { |
| "epoch": 0.9603729603729604, |
| "grad_norm": 0.6840524141170308, |
| "learning_rate": 4.0288435762396164e-08, |
| "loss": 0.4294, |
| "step": 618 |
| }, |
| { |
| "epoch": 0.9619269619269619, |
| "grad_norm": 1.035162046106296, |
| "learning_rate": 3.725270379811019e-08, |
| "loss": 0.5584, |
| "step": 619 |
| }, |
| { |
| "epoch": 0.9634809634809635, |
| "grad_norm": 0.9156660131303417, |
| "learning_rate": 3.4335439299268414e-08, |
| "loss": 0.4166, |
| "step": 620 |
| }, |
| { |
| "epoch": 0.965034965034965, |
| "grad_norm": 0.7719462038166263, |
| "learning_rate": 3.1536711904904816e-08, |
| "loss": 0.4776, |
| "step": 621 |
| }, |
| { |
| "epoch": 0.9665889665889665, |
| "grad_norm": 0.6723067728580475, |
| "learning_rate": 2.8856588424414632e-08, |
| "loss": 0.4692, |
| "step": 622 |
| }, |
| { |
| "epoch": 0.9681429681429682, |
| "grad_norm": 0.8273069565821325, |
| "learning_rate": 2.6295132835956748e-08, |
| "loss": 0.4465, |
| "step": 623 |
| }, |
| { |
| "epoch": 0.9696969696969697, |
| "grad_norm": 0.7689357247441089, |
| "learning_rate": 2.3852406284927687e-08, |
| "loss": 0.4544, |
| "step": 624 |
| }, |
| { |
| "epoch": 0.9712509712509713, |
| "grad_norm": 0.7619653211518902, |
| "learning_rate": 2.152846708250167e-08, |
| "loss": 0.5586, |
| "step": 625 |
| }, |
| { |
| "epoch": 0.9728049728049728, |
| "grad_norm": 1.5148882415320697, |
| "learning_rate": 1.93233707042384e-08, |
| "loss": 0.4416, |
| "step": 626 |
| }, |
| { |
| "epoch": 0.9743589743589743, |
| "grad_norm": 0.723885981000404, |
| "learning_rate": 1.723716978876133e-08, |
| "loss": 0.366, |
| "step": 627 |
| }, |
| { |
| "epoch": 0.9759129759129759, |
| "grad_norm": 0.9896178053638355, |
| "learning_rate": 1.5269914136497033e-08, |
| "loss": 0.6383, |
| "step": 628 |
| }, |
| { |
| "epoch": 0.9774669774669774, |
| "grad_norm": 0.9450685315761042, |
| "learning_rate": 1.3421650708487777e-08, |
| "loss": 0.4261, |
| "step": 629 |
| }, |
| { |
| "epoch": 0.9790209790209791, |
| "grad_norm": 0.7776928320411156, |
| "learning_rate": 1.1692423625273563e-08, |
| "loss": 0.3718, |
| "step": 630 |
| }, |
| { |
| "epoch": 0.9805749805749806, |
| "grad_norm": 2.709950931518374, |
| "learning_rate": 1.0082274165834627e-08, |
| "loss": 0.4794, |
| "step": 631 |
| }, |
| { |
| "epoch": 0.9821289821289821, |
| "grad_norm": 0.812744833469983, |
| "learning_rate": 8.591240766607779e-09, |
| "loss": 0.5243, |
| "step": 632 |
| }, |
| { |
| "epoch": 0.9836829836829837, |
| "grad_norm": 0.7580063402130408, |
| "learning_rate": 7.219359020570471e-09, |
| "loss": 0.4821, |
| "step": 633 |
| }, |
| { |
| "epoch": 0.9852369852369852, |
| "grad_norm": 1.012231114852923, |
| "learning_rate": 5.966661676388152e-09, |
| "loss": 0.4589, |
| "step": 634 |
| }, |
| { |
| "epoch": 0.9867909867909868, |
| "grad_norm": 0.7714041439030666, |
| "learning_rate": 4.833178637633773e-09, |
| "loss": 0.5265, |
| "step": 635 |
| }, |
| { |
| "epoch": 0.9883449883449883, |
| "grad_norm": 0.7785451239864247, |
| "learning_rate": 3.818936962076136e-09, |
| "loss": 0.5888, |
| "step": 636 |
| }, |
| { |
| "epoch": 0.98989898989899, |
| "grad_norm": 0.8911993119168611, |
| "learning_rate": 2.9239608610298618e-09, |
| "loss": 0.415, |
| "step": 637 |
| }, |
| { |
| "epoch": 0.9914529914529915, |
| "grad_norm": 0.9270156128048929, |
| "learning_rate": 2.148271698781401e-09, |
| "loss": 0.4057, |
| "step": 638 |
| }, |
| { |
| "epoch": 0.993006993006993, |
| "grad_norm": 0.7334067869030746, |
| "learning_rate": 1.4918879920750029e-09, |
| "loss": 0.4826, |
| "step": 639 |
| }, |
| { |
| "epoch": 0.9945609945609946, |
| "grad_norm": 0.7633076997899867, |
| "learning_rate": 9.548254096752862e-10, |
| "loss": 0.6209, |
| "step": 640 |
| }, |
| { |
| "epoch": 0.9961149961149961, |
| "grad_norm": 1.0315005661760475, |
| "learning_rate": 5.370967719897646e-10, |
| "loss": 0.4908, |
| "step": 641 |
| }, |
| { |
| "epoch": 0.9976689976689976, |
| "grad_norm": 0.8907517067968038, |
| "learning_rate": 2.387120507629792e-10, |
| "loss": 0.467, |
| "step": 642 |
| }, |
| { |
| "epoch": 0.9992229992229992, |
| "grad_norm": 0.9387605136360067, |
| "learning_rate": 5.967836884168687e-11, |
| "loss": 0.5826, |
| "step": 643 |
| } |
| ], |
| "logging_steps": 1, |
| "max_steps": 643, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 1, |
| "save_steps": 500, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 70581768388608.0, |
| "train_batch_size": 1, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|