diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,5923 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 3.0, + "eval_steps": 500, + "global_step": 1680, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0035714285714285713, + "grad_norm": 1.015763759613037, + "learning_rate": 2.3809523809523811e-07, + "loss": 1.6102378368377686, + "step": 2 + }, + { + "epoch": 0.007142857142857143, + "grad_norm": 2.673253297805786, + "learning_rate": 7.142857142857143e-07, + "loss": 2.1662302017211914, + "step": 4 + }, + { + "epoch": 0.010714285714285714, + "grad_norm": 1.0624771118164062, + "learning_rate": 1.1904761904761906e-06, + "loss": 1.9342565536499023, + "step": 6 + }, + { + "epoch": 0.014285714285714285, + "grad_norm": 0.48025915026664734, + "learning_rate": 1.6666666666666667e-06, + "loss": 1.6917049884796143, + "step": 8 + }, + { + "epoch": 0.017857142857142856, + "grad_norm": 0.8735950589179993, + "learning_rate": 2.1428571428571427e-06, + "loss": 2.5385754108428955, + "step": 10 + }, + { + "epoch": 0.02142857142857143, + "grad_norm": 0.611692488193512, + "learning_rate": 2.6190476190476192e-06, + "loss": 2.1326732635498047, + "step": 12 + }, + { + "epoch": 0.025, + "grad_norm": 1.3469291925430298, + "learning_rate": 3.0952380952380957e-06, + "loss": 2.2317585945129395, + "step": 14 + }, + { + "epoch": 0.02857142857142857, + "grad_norm": 0.6198003888130188, + "learning_rate": 3.5714285714285718e-06, + "loss": 1.9082640409469604, + "step": 16 + }, + { + "epoch": 0.03214285714285714, + "grad_norm": 0.7136287689208984, + "learning_rate": 4.047619047619048e-06, + "loss": 1.882780909538269, + "step": 18 + }, + { + "epoch": 0.03571428571428571, + "grad_norm": 1.2058864831924438, + "learning_rate": 4.523809523809524e-06, + "loss": 2.1093122959136963, + "step": 20 + }, + { + "epoch": 0.039285714285714285, + "grad_norm": 0.44349122047424316, + "learning_rate": 5e-06, + "loss": 1.7804787158966064, + "step": 22 + }, + { + "epoch": 0.04285714285714286, + "grad_norm": 1.3120089769363403, + "learning_rate": 5.476190476190477e-06, + "loss": 2.0871825218200684, + "step": 24 + }, + { + "epoch": 0.04642857142857143, + "grad_norm": 0.6059308052062988, + "learning_rate": 5.9523809523809525e-06, + "loss": 2.06781005859375, + "step": 26 + }, + { + "epoch": 0.05, + "grad_norm": 0.42657509446144104, + "learning_rate": 6.4285714285714295e-06, + "loss": 1.5501041412353516, + "step": 28 + }, + { + "epoch": 0.05357142857142857, + "grad_norm": 1.5947790145874023, + "learning_rate": 6.9047619047619055e-06, + "loss": 2.1851985454559326, + "step": 30 + }, + { + "epoch": 0.05714285714285714, + "grad_norm": 1.2271709442138672, + "learning_rate": 7.380952380952382e-06, + "loss": 2.1415982246398926, + "step": 32 + }, + { + "epoch": 0.060714285714285714, + "grad_norm": 0.4892466068267822, + "learning_rate": 7.857142857142858e-06, + "loss": 1.8040955066680908, + "step": 34 + }, + { + "epoch": 0.06428571428571428, + "grad_norm": 1.1410105228424072, + "learning_rate": 8.333333333333334e-06, + "loss": 1.7697616815567017, + "step": 36 + }, + { + "epoch": 0.06785714285714285, + "grad_norm": 0.6576052308082581, + "learning_rate": 8.80952380952381e-06, + "loss": 1.7375702857971191, + "step": 38 + }, + { + "epoch": 0.07142857142857142, + "grad_norm": 0.5573370456695557, + "learning_rate": 9.285714285714288e-06, + "loss": 1.9562886953353882, + "step": 40 + }, + { + "epoch": 0.075, + "grad_norm": 0.4374576508998871, + "learning_rate": 9.761904761904762e-06, + "loss": 2.0028505325317383, + "step": 42 + }, + { + "epoch": 0.07857142857142857, + "grad_norm": 1.975783348083496, + "learning_rate": 1.0238095238095238e-05, + "loss": 2.158412456512451, + "step": 44 + }, + { + "epoch": 0.08214285714285714, + "grad_norm": 0.34883615374565125, + "learning_rate": 1.0714285714285714e-05, + "loss": 1.8357443809509277, + "step": 46 + }, + { + "epoch": 0.08571428571428572, + "grad_norm": 0.507429838180542, + "learning_rate": 1.1190476190476192e-05, + "loss": 1.6925463676452637, + "step": 48 + }, + { + "epoch": 0.08928571428571429, + "grad_norm": 0.5762679576873779, + "learning_rate": 1.1666666666666668e-05, + "loss": 1.9667983055114746, + "step": 50 + }, + { + "epoch": 0.09285714285714286, + "grad_norm": 0.5237306952476501, + "learning_rate": 1.2142857142857142e-05, + "loss": 1.739395022392273, + "step": 52 + }, + { + "epoch": 0.09642857142857143, + "grad_norm": 1.05526602268219, + "learning_rate": 1.261904761904762e-05, + "loss": 1.9076595306396484, + "step": 54 + }, + { + "epoch": 0.1, + "grad_norm": 0.5703083872795105, + "learning_rate": 1.3095238095238096e-05, + "loss": 1.6693158149719238, + "step": 56 + }, + { + "epoch": 0.10357142857142858, + "grad_norm": 0.36678144335746765, + "learning_rate": 1.3571428571428574e-05, + "loss": 1.9416877031326294, + "step": 58 + }, + { + "epoch": 0.10714285714285714, + "grad_norm": 0.5519128441810608, + "learning_rate": 1.4047619047619048e-05, + "loss": 1.5721981525421143, + "step": 60 + }, + { + "epoch": 0.11071428571428571, + "grad_norm": 1.7103147506713867, + "learning_rate": 1.4523809523809524e-05, + "loss": 1.9566872119903564, + "step": 62 + }, + { + "epoch": 0.11428571428571428, + "grad_norm": 0.46710196137428284, + "learning_rate": 1.5000000000000002e-05, + "loss": 1.7038791179656982, + "step": 64 + }, + { + "epoch": 0.11785714285714285, + "grad_norm": 0.5537970066070557, + "learning_rate": 1.5476190476190476e-05, + "loss": 1.7286560535430908, + "step": 66 + }, + { + "epoch": 0.12142857142857143, + "grad_norm": 0.4240347445011139, + "learning_rate": 1.5952380952380954e-05, + "loss": 1.4021289348602295, + "step": 68 + }, + { + "epoch": 0.125, + "grad_norm": 0.5624739527702332, + "learning_rate": 1.642857142857143e-05, + "loss": 1.7000582218170166, + "step": 70 + }, + { + "epoch": 0.12857142857142856, + "grad_norm": 0.929057776927948, + "learning_rate": 1.6904761904761906e-05, + "loss": 1.57151198387146, + "step": 72 + }, + { + "epoch": 0.13214285714285715, + "grad_norm": 0.45938676595687866, + "learning_rate": 1.7380952380952384e-05, + "loss": 1.7509145736694336, + "step": 74 + }, + { + "epoch": 0.1357142857142857, + "grad_norm": 0.7960886359214783, + "learning_rate": 1.785714285714286e-05, + "loss": 1.627264142036438, + "step": 76 + }, + { + "epoch": 0.1392857142857143, + "grad_norm": 0.6913440227508545, + "learning_rate": 1.8333333333333333e-05, + "loss": 1.5597522258758545, + "step": 78 + }, + { + "epoch": 0.14285714285714285, + "grad_norm": 0.3351443111896515, + "learning_rate": 1.880952380952381e-05, + "loss": 1.5047800540924072, + "step": 80 + }, + { + "epoch": 0.14642857142857144, + "grad_norm": 4.245250701904297, + "learning_rate": 1.928571428571429e-05, + "loss": 1.6616361141204834, + "step": 82 + }, + { + "epoch": 0.15, + "grad_norm": 0.8213376402854919, + "learning_rate": 1.9761904761904763e-05, + "loss": 1.633697509765625, + "step": 84 + }, + { + "epoch": 0.15357142857142858, + "grad_norm": 0.5915255546569824, + "learning_rate": 1.9999982564020695e-05, + "loss": 1.3151729106903076, + "step": 86 + }, + { + "epoch": 0.15714285714285714, + "grad_norm": 0.5589431524276733, + "learning_rate": 1.9999843076591598e-05, + "loss": 1.5256870985031128, + "step": 88 + }, + { + "epoch": 0.16071428571428573, + "grad_norm": 0.5508841872215271, + "learning_rate": 1.9999564103895265e-05, + "loss": 1.4112975597381592, + "step": 90 + }, + { + "epoch": 0.16428571428571428, + "grad_norm": 0.6648878455162048, + "learning_rate": 1.9999145650255392e-05, + "loss": 1.4721081256866455, + "step": 92 + }, + { + "epoch": 0.16785714285714284, + "grad_norm": 0.7325057983398438, + "learning_rate": 1.999858772215744e-05, + "loss": 1.45548415184021, + "step": 94 + }, + { + "epoch": 0.17142857142857143, + "grad_norm": 4.314115524291992, + "learning_rate": 1.9997890328248536e-05, + "loss": 1.6468373537063599, + "step": 96 + }, + { + "epoch": 0.175, + "grad_norm": 0.4537746012210846, + "learning_rate": 1.9997053479337327e-05, + "loss": 1.4209429025650024, + "step": 98 + }, + { + "epoch": 0.17857142857142858, + "grad_norm": 0.35222765803337097, + "learning_rate": 1.9996077188393826e-05, + "loss": 1.390677809715271, + "step": 100 + }, + { + "epoch": 0.18214285714285713, + "grad_norm": 1.7469472885131836, + "learning_rate": 1.9994961470549216e-05, + "loss": 1.7277623414993286, + "step": 102 + }, + { + "epoch": 0.18571428571428572, + "grad_norm": 0.1474483162164688, + "learning_rate": 1.999370634309559e-05, + "loss": 1.1081668138504028, + "step": 104 + }, + { + "epoch": 0.18928571428571428, + "grad_norm": 0.46639060974121094, + "learning_rate": 1.999231182548571e-05, + "loss": 1.410254955291748, + "step": 106 + }, + { + "epoch": 0.19285714285714287, + "grad_norm": 0.43072256445884705, + "learning_rate": 1.9990777939332697e-05, + "loss": 1.404090404510498, + "step": 108 + }, + { + "epoch": 0.19642857142857142, + "grad_norm": 0.5892950892448425, + "learning_rate": 1.998910470840969e-05, + "loss": 1.591571569442749, + "step": 110 + }, + { + "epoch": 0.2, + "grad_norm": 0.7321836352348328, + "learning_rate": 1.9987292158649477e-05, + "loss": 1.7062772512435913, + "step": 112 + }, + { + "epoch": 0.20357142857142857, + "grad_norm": 0.5677797794342041, + "learning_rate": 1.9985340318144104e-05, + "loss": 1.6029880046844482, + "step": 114 + }, + { + "epoch": 0.20714285714285716, + "grad_norm": 0.5012320280075073, + "learning_rate": 1.998324921714443e-05, + "loss": 1.6100475788116455, + "step": 116 + }, + { + "epoch": 0.21071428571428572, + "grad_norm": 0.3957848846912384, + "learning_rate": 1.9981018888059666e-05, + "loss": 1.4384129047393799, + "step": 118 + }, + { + "epoch": 0.21428571428571427, + "grad_norm": 0.5229449272155762, + "learning_rate": 1.997864936545686e-05, + "loss": 1.3900649547576904, + "step": 120 + }, + { + "epoch": 0.21785714285714286, + "grad_norm": 1.3295955657958984, + "learning_rate": 1.997614068606038e-05, + "loss": 1.7176506519317627, + "step": 122 + }, + { + "epoch": 0.22142857142857142, + "grad_norm": 0.44528570771217346, + "learning_rate": 1.997349288875132e-05, + "loss": 1.4366806745529175, + "step": 124 + }, + { + "epoch": 0.225, + "grad_norm": 0.39040693640708923, + "learning_rate": 1.997070601456693e-05, + "loss": 1.3515652418136597, + "step": 126 + }, + { + "epoch": 0.22857142857142856, + "grad_norm": 0.5782378315925598, + "learning_rate": 1.9967780106699938e-05, + "loss": 1.352698564529419, + "step": 128 + }, + { + "epoch": 0.23214285714285715, + "grad_norm": 0.4463367760181427, + "learning_rate": 1.9964715210497926e-05, + "loss": 1.2373923063278198, + "step": 130 + }, + { + "epoch": 0.2357142857142857, + "grad_norm": 0.4424625635147095, + "learning_rate": 1.996151137346259e-05, + "loss": 1.4218852519989014, + "step": 132 + }, + { + "epoch": 0.2392857142857143, + "grad_norm": 0.39771443605422974, + "learning_rate": 1.9958168645249036e-05, + "loss": 1.3959071636199951, + "step": 134 + }, + { + "epoch": 0.24285714285714285, + "grad_norm": 0.995601236820221, + "learning_rate": 1.995468707766497e-05, + "loss": 1.5231602191925049, + "step": 136 + }, + { + "epoch": 0.24642857142857144, + "grad_norm": 0.8687110543251038, + "learning_rate": 1.995106672466994e-05, + "loss": 1.5624220371246338, + "step": 138 + }, + { + "epoch": 0.25, + "grad_norm": 2.4341928958892822, + "learning_rate": 1.9947307642374466e-05, + "loss": 1.1912024021148682, + "step": 140 + }, + { + "epoch": 0.25357142857142856, + "grad_norm": 0.3811090886592865, + "learning_rate": 1.9943409889039186e-05, + "loss": 1.145674467086792, + "step": 142 + }, + { + "epoch": 0.2571428571428571, + "grad_norm": 0.3321725130081177, + "learning_rate": 1.9939373525073946e-05, + "loss": 1.5421075820922852, + "step": 144 + }, + { + "epoch": 0.26071428571428573, + "grad_norm": 0.9177084565162659, + "learning_rate": 1.9935198613036877e-05, + "loss": 1.5951645374298096, + "step": 146 + }, + { + "epoch": 0.2642857142857143, + "grad_norm": 0.4631420969963074, + "learning_rate": 1.9930885217633408e-05, + "loss": 1.4165486097335815, + "step": 148 + }, + { + "epoch": 0.26785714285714285, + "grad_norm": 0.31056374311447144, + "learning_rate": 1.992643340571527e-05, + "loss": 1.337796688079834, + "step": 150 + }, + { + "epoch": 0.2714285714285714, + "grad_norm": 0.7542070746421814, + "learning_rate": 1.992184324627946e-05, + "loss": 1.378330111503601, + "step": 152 + }, + { + "epoch": 0.275, + "grad_norm": 0.32567477226257324, + "learning_rate": 1.9917114810467187e-05, + "loss": 1.3354138135910034, + "step": 154 + }, + { + "epoch": 0.2785714285714286, + "grad_norm": 0.39338260889053345, + "learning_rate": 1.9912248171562732e-05, + "loss": 1.454702615737915, + "step": 156 + }, + { + "epoch": 0.28214285714285714, + "grad_norm": 1.2659757137298584, + "learning_rate": 1.9907243404992357e-05, + "loss": 1.430185317993164, + "step": 158 + }, + { + "epoch": 0.2857142857142857, + "grad_norm": 0.7528992891311646, + "learning_rate": 1.9902100588323095e-05, + "loss": 1.4888527393341064, + "step": 160 + }, + { + "epoch": 0.2892857142857143, + "grad_norm": 0.3230721652507782, + "learning_rate": 1.9896819801261575e-05, + "loss": 1.372198462486267, + "step": 162 + }, + { + "epoch": 0.29285714285714287, + "grad_norm": 0.5240757465362549, + "learning_rate": 1.9891401125652788e-05, + "loss": 1.3045603036880493, + "step": 164 + }, + { + "epoch": 0.29642857142857143, + "grad_norm": 0.41186925768852234, + "learning_rate": 1.988584464547879e-05, + "loss": 1.3587464094161987, + "step": 166 + }, + { + "epoch": 0.3, + "grad_norm": 0.5823759436607361, + "learning_rate": 1.9880150446857435e-05, + "loss": 1.2455697059631348, + "step": 168 + }, + { + "epoch": 0.30357142857142855, + "grad_norm": 0.6033863425254822, + "learning_rate": 1.987431861804102e-05, + "loss": 1.5122787952423096, + "step": 170 + }, + { + "epoch": 0.30714285714285716, + "grad_norm": 0.5195621848106384, + "learning_rate": 1.9868349249414918e-05, + "loss": 1.5619277954101562, + "step": 172 + }, + { + "epoch": 0.3107142857142857, + "grad_norm": 0.3413119614124298, + "learning_rate": 1.9862242433496185e-05, + "loss": 1.2759490013122559, + "step": 174 + }, + { + "epoch": 0.3142857142857143, + "grad_norm": 0.43644359707832336, + "learning_rate": 1.9855998264932118e-05, + "loss": 1.525347352027893, + "step": 176 + }, + { + "epoch": 0.31785714285714284, + "grad_norm": 2.2331385612487793, + "learning_rate": 1.9849616840498807e-05, + "loss": 1.4077792167663574, + "step": 178 + }, + { + "epoch": 0.32142857142857145, + "grad_norm": 1.2636653184890747, + "learning_rate": 1.9843098259099597e-05, + "loss": 1.494966745376587, + "step": 180 + }, + { + "epoch": 0.325, + "grad_norm": 1.3332237005233765, + "learning_rate": 1.9836442621763593e-05, + "loss": 1.2962632179260254, + "step": 182 + }, + { + "epoch": 0.32857142857142857, + "grad_norm": 0.9052379131317139, + "learning_rate": 1.982965003164408e-05, + "loss": 1.4094679355621338, + "step": 184 + }, + { + "epoch": 0.33214285714285713, + "grad_norm": 0.3790226876735687, + "learning_rate": 1.982272059401692e-05, + "loss": 1.308838129043579, + "step": 186 + }, + { + "epoch": 0.3357142857142857, + "grad_norm": 0.5489211678504944, + "learning_rate": 1.9815654416278924e-05, + "loss": 1.2358474731445312, + "step": 188 + }, + { + "epoch": 0.3392857142857143, + "grad_norm": 0.32167771458625793, + "learning_rate": 1.980845160794619e-05, + "loss": 1.24757719039917, + "step": 190 + }, + { + "epoch": 0.34285714285714286, + "grad_norm": 0.5797743201255798, + "learning_rate": 1.9801112280652406e-05, + "loss": 0.9474111795425415, + "step": 192 + }, + { + "epoch": 0.3464285714285714, + "grad_norm": 0.5825701951980591, + "learning_rate": 1.979363654814711e-05, + "loss": 1.6204115152359009, + "step": 194 + }, + { + "epoch": 0.35, + "grad_norm": 0.2390112429857254, + "learning_rate": 1.9786024526293943e-05, + "loss": 1.2390871047973633, + "step": 196 + }, + { + "epoch": 0.3535714285714286, + "grad_norm": 0.49600404500961304, + "learning_rate": 1.9778276333068833e-05, + "loss": 1.4131810665130615, + "step": 198 + }, + { + "epoch": 0.35714285714285715, + "grad_norm": 0.678787350654602, + "learning_rate": 1.977039208855819e-05, + "loss": 1.23861563205719, + "step": 200 + }, + { + "epoch": 0.3607142857142857, + "grad_norm": 0.8061485290527344, + "learning_rate": 1.9762371914957027e-05, + "loss": 1.4984278678894043, + "step": 202 + }, + { + "epoch": 0.36428571428571427, + "grad_norm": 1.1672908067703247, + "learning_rate": 1.9754215936567077e-05, + "loss": 1.346609354019165, + "step": 204 + }, + { + "epoch": 0.3678571428571429, + "grad_norm": 0.42324623465538025, + "learning_rate": 1.9745924279794853e-05, + "loss": 1.4815120697021484, + "step": 206 + }, + { + "epoch": 0.37142857142857144, + "grad_norm": 1.4115076065063477, + "learning_rate": 1.97374970731497e-05, + "loss": 1.2184674739837646, + "step": 208 + }, + { + "epoch": 0.375, + "grad_norm": 0.3797534704208374, + "learning_rate": 1.9728934447241815e-05, + "loss": 1.419222116470337, + "step": 210 + }, + { + "epoch": 0.37857142857142856, + "grad_norm": 0.49382659792900085, + "learning_rate": 1.9720236534780184e-05, + "loss": 1.3580292463302612, + "step": 212 + }, + { + "epoch": 0.3821428571428571, + "grad_norm": 0.838750958442688, + "learning_rate": 1.971140347057057e-05, + "loss": 1.3154736757278442, + "step": 214 + }, + { + "epoch": 0.38571428571428573, + "grad_norm": 0.43846485018730164, + "learning_rate": 1.970243539151339e-05, + "loss": 1.4407868385314941, + "step": 216 + }, + { + "epoch": 0.3892857142857143, + "grad_norm": 0.4752633571624756, + "learning_rate": 1.9693332436601616e-05, + "loss": 1.6369309425354004, + "step": 218 + }, + { + "epoch": 0.39285714285714285, + "grad_norm": 0.7840119004249573, + "learning_rate": 1.968409474691861e-05, + "loss": 1.2841938734054565, + "step": 220 + }, + { + "epoch": 0.3964285714285714, + "grad_norm": 0.3010779917240143, + "learning_rate": 1.967472246563593e-05, + "loss": 1.5273520946502686, + "step": 222 + }, + { + "epoch": 0.4, + "grad_norm": 0.37263405323028564, + "learning_rate": 1.966521573801113e-05, + "loss": 1.2460472583770752, + "step": 224 + }, + { + "epoch": 0.4035714285714286, + "grad_norm": 0.2715803384780884, + "learning_rate": 1.9655574711385497e-05, + "loss": 1.2670462131500244, + "step": 226 + }, + { + "epoch": 0.40714285714285714, + "grad_norm": 0.7070139646530151, + "learning_rate": 1.9645799535181767e-05, + "loss": 1.363389253616333, + "step": 228 + }, + { + "epoch": 0.4107142857142857, + "grad_norm": 0.6753621101379395, + "learning_rate": 1.9635890360901805e-05, + "loss": 1.5315601825714111, + "step": 230 + }, + { + "epoch": 0.4142857142857143, + "grad_norm": 0.5578319430351257, + "learning_rate": 1.9625847342124278e-05, + "loss": 1.2386059761047363, + "step": 232 + }, + { + "epoch": 0.41785714285714287, + "grad_norm": 0.5665930509567261, + "learning_rate": 1.961567063450224e-05, + "loss": 1.4771205186843872, + "step": 234 + }, + { + "epoch": 0.42142857142857143, + "grad_norm": 0.3672884404659271, + "learning_rate": 1.960536039576076e-05, + "loss": 1.2611351013183594, + "step": 236 + }, + { + "epoch": 0.425, + "grad_norm": 0.37258124351501465, + "learning_rate": 1.959491678569444e-05, + "loss": 1.3256354331970215, + "step": 238 + }, + { + "epoch": 0.42857142857142855, + "grad_norm": 0.5384905934333801, + "learning_rate": 1.958433996616497e-05, + "loss": 1.2469019889831543, + "step": 240 + }, + { + "epoch": 0.43214285714285716, + "grad_norm": 0.7136536240577698, + "learning_rate": 1.957363010109859e-05, + "loss": 1.1012486219406128, + "step": 242 + }, + { + "epoch": 0.4357142857142857, + "grad_norm": 0.6285988092422485, + "learning_rate": 1.9562787356483573e-05, + "loss": 1.3433374166488647, + "step": 244 + }, + { + "epoch": 0.4392857142857143, + "grad_norm": 1.8126535415649414, + "learning_rate": 1.9551811900367642e-05, + "loss": 1.2772047519683838, + "step": 246 + }, + { + "epoch": 0.44285714285714284, + "grad_norm": 0.7285305857658386, + "learning_rate": 1.954070390285537e-05, + "loss": 1.4760041236877441, + "step": 248 + }, + { + "epoch": 0.44642857142857145, + "grad_norm": 0.685204803943634, + "learning_rate": 1.9529463536105525e-05, + "loss": 1.436897873878479, + "step": 250 + }, + { + "epoch": 0.45, + "grad_norm": 0.7495101094245911, + "learning_rate": 1.951809097432844e-05, + "loss": 1.3357012271881104, + "step": 252 + }, + { + "epoch": 0.45357142857142857, + "grad_norm": 0.39518865942955017, + "learning_rate": 1.9506586393783278e-05, + "loss": 1.3316457271575928, + "step": 254 + }, + { + "epoch": 0.45714285714285713, + "grad_norm": 0.3139326870441437, + "learning_rate": 1.949494997277531e-05, + "loss": 1.2338916063308716, + "step": 256 + }, + { + "epoch": 0.4607142857142857, + "grad_norm": 0.7414034008979797, + "learning_rate": 1.948318189165316e-05, + "loss": 0.9024492502212524, + "step": 258 + }, + { + "epoch": 0.4642857142857143, + "grad_norm": 0.31645119190216064, + "learning_rate": 1.9471282332805996e-05, + "loss": 1.4956551790237427, + "step": 260 + }, + { + "epoch": 0.46785714285714286, + "grad_norm": 0.40859320759773254, + "learning_rate": 1.9459251480660726e-05, + "loss": 1.5328614711761475, + "step": 262 + }, + { + "epoch": 0.4714285714285714, + "grad_norm": 0.5480664372444153, + "learning_rate": 1.944708952167911e-05, + "loss": 1.263709306716919, + "step": 264 + }, + { + "epoch": 0.475, + "grad_norm": 0.4685518741607666, + "learning_rate": 1.9434796644354885e-05, + "loss": 1.2509859800338745, + "step": 266 + }, + { + "epoch": 0.4785714285714286, + "grad_norm": 0.4723505973815918, + "learning_rate": 1.942237303921086e-05, + "loss": 1.1997942924499512, + "step": 268 + }, + { + "epoch": 0.48214285714285715, + "grad_norm": 0.49579185247421265, + "learning_rate": 1.9409818898795924e-05, + "loss": 1.4426811933517456, + "step": 270 + }, + { + "epoch": 0.4857142857142857, + "grad_norm": 0.7510737776756287, + "learning_rate": 1.93971344176821e-05, + "loss": 1.2704529762268066, + "step": 272 + }, + { + "epoch": 0.48928571428571427, + "grad_norm": 0.4258019030094147, + "learning_rate": 1.9384319792461513e-05, + "loss": 1.2604095935821533, + "step": 274 + }, + { + "epoch": 0.4928571428571429, + "grad_norm": 0.7846337556838989, + "learning_rate": 1.9371375221743333e-05, + "loss": 1.2763457298278809, + "step": 276 + }, + { + "epoch": 0.49642857142857144, + "grad_norm": 0.6551533341407776, + "learning_rate": 1.9358300906150715e-05, + "loss": 1.4955462217330933, + "step": 278 + }, + { + "epoch": 0.5, + "grad_norm": 0.39320385456085205, + "learning_rate": 1.934509704831768e-05, + "loss": 1.305250644683838, + "step": 280 + }, + { + "epoch": 0.5035714285714286, + "grad_norm": 0.4256862998008728, + "learning_rate": 1.9331763852885988e-05, + "loss": 1.2777750492095947, + "step": 282 + }, + { + "epoch": 0.5071428571428571, + "grad_norm": 1.6539260149002075, + "learning_rate": 1.931830152650193e-05, + "loss": 1.3221278190612793, + "step": 284 + }, + { + "epoch": 0.5107142857142857, + "grad_norm": 0.7155967950820923, + "learning_rate": 1.9304710277813182e-05, + "loss": 1.5129222869873047, + "step": 286 + }, + { + "epoch": 0.5142857142857142, + "grad_norm": 0.4225437343120575, + "learning_rate": 1.929099031746551e-05, + "loss": 1.4677280187606812, + "step": 288 + }, + { + "epoch": 0.5178571428571429, + "grad_norm": 1.5834101438522339, + "learning_rate": 1.9277141858099552e-05, + "loss": 1.324265480041504, + "step": 290 + }, + { + "epoch": 0.5214285714285715, + "grad_norm": 0.42641639709472656, + "learning_rate": 1.9263165114347503e-05, + "loss": 1.5003204345703125, + "step": 292 + }, + { + "epoch": 0.525, + "grad_norm": 0.32519879937171936, + "learning_rate": 1.924906030282979e-05, + "loss": 1.2406567335128784, + "step": 294 + }, + { + "epoch": 0.5285714285714286, + "grad_norm": 0.8007866144180298, + "learning_rate": 1.9234827642151705e-05, + "loss": 1.5307186841964722, + "step": 296 + }, + { + "epoch": 0.5321428571428571, + "grad_norm": 0.2881964147090912, + "learning_rate": 1.922046735290004e-05, + "loss": 1.3302806615829468, + "step": 298 + }, + { + "epoch": 0.5357142857142857, + "grad_norm": 0.39077696204185486, + "learning_rate": 1.9205979657639658e-05, + "loss": 1.3356809616088867, + "step": 300 + }, + { + "epoch": 0.5392857142857143, + "grad_norm": 2.1859304904937744, + "learning_rate": 1.919136478091003e-05, + "loss": 1.58237886428833, + "step": 302 + }, + { + "epoch": 0.5428571428571428, + "grad_norm": 0.5423356890678406, + "learning_rate": 1.9176622949221776e-05, + "loss": 1.2764296531677246, + "step": 304 + }, + { + "epoch": 0.5464285714285714, + "grad_norm": 0.6123110055923462, + "learning_rate": 1.9161754391053127e-05, + "loss": 1.3649842739105225, + "step": 306 + }, + { + "epoch": 0.55, + "grad_norm": 0.40666741132736206, + "learning_rate": 1.9146759336846418e-05, + "loss": 1.3004515171051025, + "step": 308 + }, + { + "epoch": 0.5535714285714286, + "grad_norm": 0.4282005727291107, + "learning_rate": 1.91316380190045e-05, + "loss": 1.4622749090194702, + "step": 310 + }, + { + "epoch": 0.5571428571428572, + "grad_norm": 0.7790765166282654, + "learning_rate": 1.911639067188713e-05, + "loss": 1.2346100807189941, + "step": 312 + }, + { + "epoch": 0.5607142857142857, + "grad_norm": 0.8703919649124146, + "learning_rate": 1.9101017531807344e-05, + "loss": 1.2121374607086182, + "step": 314 + }, + { + "epoch": 0.5642857142857143, + "grad_norm": 0.6184125542640686, + "learning_rate": 1.9085518837027812e-05, + "loss": 1.2770280838012695, + "step": 316 + }, + { + "epoch": 0.5678571428571428, + "grad_norm": 1.1969032287597656, + "learning_rate": 1.9069894827757112e-05, + "loss": 1.2650474309921265, + "step": 318 + }, + { + "epoch": 0.5714285714285714, + "grad_norm": 0.6322106719017029, + "learning_rate": 1.905414574614604e-05, + "loss": 1.2580130100250244, + "step": 320 + }, + { + "epoch": 0.575, + "grad_norm": 0.24475856125354767, + "learning_rate": 1.9038271836283826e-05, + "loss": 1.1550788879394531, + "step": 322 + }, + { + "epoch": 0.5785714285714286, + "grad_norm": 0.3760751783847809, + "learning_rate": 1.9022273344194388e-05, + "loss": 1.4843223094940186, + "step": 324 + }, + { + "epoch": 0.5821428571428572, + "grad_norm": 0.46275073289871216, + "learning_rate": 1.9006150517832482e-05, + "loss": 1.448726773262024, + "step": 326 + }, + { + "epoch": 0.5857142857142857, + "grad_norm": 0.4800632894039154, + "learning_rate": 1.8989903607079885e-05, + "loss": 1.2668516635894775, + "step": 328 + }, + { + "epoch": 0.5892857142857143, + "grad_norm": 0.3428266644477844, + "learning_rate": 1.8973532863741504e-05, + "loss": 1.222381353378296, + "step": 330 + }, + { + "epoch": 0.5928571428571429, + "grad_norm": 0.4723515510559082, + "learning_rate": 1.895703854154149e-05, + "loss": 1.3355088233947754, + "step": 332 + }, + { + "epoch": 0.5964285714285714, + "grad_norm": 0.3576123118400574, + "learning_rate": 1.894042089611929e-05, + "loss": 1.3003771305084229, + "step": 334 + }, + { + "epoch": 0.6, + "grad_norm": 0.3541733920574188, + "learning_rate": 1.89236801850257e-05, + "loss": 1.207617998123169, + "step": 336 + }, + { + "epoch": 0.6035714285714285, + "grad_norm": 0.6847231984138489, + "learning_rate": 1.8906816667718854e-05, + "loss": 1.2353670597076416, + "step": 338 + }, + { + "epoch": 0.6071428571428571, + "grad_norm": 0.28425782918930054, + "learning_rate": 1.8889830605560234e-05, + "loss": 1.3947781324386597, + "step": 340 + }, + { + "epoch": 0.6107142857142858, + "grad_norm": 0.6358458995819092, + "learning_rate": 1.8872722261810576e-05, + "loss": 1.209977626800537, + "step": 342 + }, + { + "epoch": 0.6142857142857143, + "grad_norm": 0.5836315155029297, + "learning_rate": 1.8855491901625835e-05, + "loss": 1.2867995500564575, + "step": 344 + }, + { + "epoch": 0.6178571428571429, + "grad_norm": 0.5231335163116455, + "learning_rate": 1.883813979205304e-05, + "loss": 1.2911901473999023, + "step": 346 + }, + { + "epoch": 0.6214285714285714, + "grad_norm": 0.43874913454055786, + "learning_rate": 1.8820666202026172e-05, + "loss": 1.2187429666519165, + "step": 348 + }, + { + "epoch": 0.625, + "grad_norm": 0.7828032970428467, + "learning_rate": 1.8803071402361995e-05, + "loss": 1.492908000946045, + "step": 350 + }, + { + "epoch": 0.6285714285714286, + "grad_norm": 0.5225529074668884, + "learning_rate": 1.8785355665755864e-05, + "loss": 1.5098028182983398, + "step": 352 + }, + { + "epoch": 0.6321428571428571, + "grad_norm": 0.47736281156539917, + "learning_rate": 1.8767519266777473e-05, + "loss": 1.176377296447754, + "step": 354 + }, + { + "epoch": 0.6357142857142857, + "grad_norm": 0.27933818101882935, + "learning_rate": 1.8749562481866632e-05, + "loss": 1.3661816120147705, + "step": 356 + }, + { + "epoch": 0.6392857142857142, + "grad_norm": 0.9762703776359558, + "learning_rate": 1.8731485589328968e-05, + "loss": 1.6452014446258545, + "step": 358 + }, + { + "epoch": 0.6428571428571429, + "grad_norm": 0.41242146492004395, + "learning_rate": 1.8713288869331608e-05, + "loss": 1.4132412672042847, + "step": 360 + }, + { + "epoch": 0.6464285714285715, + "grad_norm": 0.4835861623287201, + "learning_rate": 1.8694972603898834e-05, + "loss": 1.2498993873596191, + "step": 362 + }, + { + "epoch": 0.65, + "grad_norm": 0.26400306820869446, + "learning_rate": 1.867653707690774e-05, + "loss": 1.52461576461792, + "step": 364 + }, + { + "epoch": 0.6535714285714286, + "grad_norm": 0.38874751329421997, + "learning_rate": 1.8657982574083784e-05, + "loss": 1.5055546760559082, + "step": 366 + }, + { + "epoch": 0.6571428571428571, + "grad_norm": 0.48226460814476013, + "learning_rate": 1.863930938299641e-05, + "loss": 1.2804465293884277, + "step": 368 + }, + { + "epoch": 0.6607142857142857, + "grad_norm": 0.5018228888511658, + "learning_rate": 1.862051779305456e-05, + "loss": 1.2952604293823242, + "step": 370 + }, + { + "epoch": 0.6642857142857143, + "grad_norm": 0.7374165058135986, + "learning_rate": 1.8601608095502186e-05, + "loss": 1.203336238861084, + "step": 372 + }, + { + "epoch": 0.6678571428571428, + "grad_norm": 0.48408135771751404, + "learning_rate": 1.8582580583413762e-05, + "loss": 1.3007687330245972, + "step": 374 + }, + { + "epoch": 0.6714285714285714, + "grad_norm": 0.4113704562187195, + "learning_rate": 1.8563435551689714e-05, + "loss": 1.354430913925171, + "step": 376 + }, + { + "epoch": 0.675, + "grad_norm": 1.2474658489227295, + "learning_rate": 1.8544173297051873e-05, + "loss": 0.8642697334289551, + "step": 378 + }, + { + "epoch": 0.6785714285714286, + "grad_norm": 0.4104273021221161, + "learning_rate": 1.852479411803886e-05, + "loss": 1.2896859645843506, + "step": 380 + }, + { + "epoch": 0.6821428571428572, + "grad_norm": 0.5409207940101624, + "learning_rate": 1.850529831500146e-05, + "loss": 1.4960516691207886, + "step": 382 + }, + { + "epoch": 0.6857142857142857, + "grad_norm": 0.48803240060806274, + "learning_rate": 1.8485686190097975e-05, + "loss": 1.7281725406646729, + "step": 384 + }, + { + "epoch": 0.6892857142857143, + "grad_norm": 0.9719665050506592, + "learning_rate": 1.8465958047289535e-05, + "loss": 1.2362210750579834, + "step": 386 + }, + { + "epoch": 0.6928571428571428, + "grad_norm": 0.3114999234676361, + "learning_rate": 1.8446114192335393e-05, + "loss": 1.2179937362670898, + "step": 388 + }, + { + "epoch": 0.6964285714285714, + "grad_norm": 0.3131682872772217, + "learning_rate": 1.8426154932788176e-05, + "loss": 1.2936997413635254, + "step": 390 + }, + { + "epoch": 0.7, + "grad_norm": 0.32079121470451355, + "learning_rate": 1.8406080577989132e-05, + "loss": 1.2510591745376587, + "step": 392 + }, + { + "epoch": 0.7035714285714286, + "grad_norm": 0.9477353692054749, + "learning_rate": 1.8385891439063325e-05, + "loss": 1.2309963703155518, + "step": 394 + }, + { + "epoch": 0.7071428571428572, + "grad_norm": 0.4169057309627533, + "learning_rate": 1.8365587828914804e-05, + "loss": 1.4043102264404297, + "step": 396 + }, + { + "epoch": 0.7107142857142857, + "grad_norm": 1.3407492637634277, + "learning_rate": 1.834517006222179e-05, + "loss": 1.486006259918213, + "step": 398 + }, + { + "epoch": 0.7142857142857143, + "grad_norm": 0.4785768389701843, + "learning_rate": 1.8324638455431755e-05, + "loss": 1.158205270767212, + "step": 400 + }, + { + "epoch": 0.7178571428571429, + "grad_norm": 3.806015729904175, + "learning_rate": 1.8303993326756543e-05, + "loss": 1.4163193702697754, + "step": 402 + }, + { + "epoch": 0.7214285714285714, + "grad_norm": 0.3417515158653259, + "learning_rate": 1.8283234996167434e-05, + "loss": 1.2975351810455322, + "step": 404 + }, + { + "epoch": 0.725, + "grad_norm": 1.0901787281036377, + "learning_rate": 1.8262363785390177e-05, + "loss": 1.4190306663513184, + "step": 406 + }, + { + "epoch": 0.7285714285714285, + "grad_norm": 0.4423401355743408, + "learning_rate": 1.8241380017900015e-05, + "loss": 1.5579520463943481, + "step": 408 + }, + { + "epoch": 0.7321428571428571, + "grad_norm": 0.5694108605384827, + "learning_rate": 1.8220284018916667e-05, + "loss": 1.2755184173583984, + "step": 410 + }, + { + "epoch": 0.7357142857142858, + "grad_norm": 0.292386919260025, + "learning_rate": 1.8199076115399285e-05, + "loss": 1.4448673725128174, + "step": 412 + }, + { + "epoch": 0.7392857142857143, + "grad_norm": 0.3983137309551239, + "learning_rate": 1.817775663604138e-05, + "loss": 0.8253436088562012, + "step": 414 + }, + { + "epoch": 0.7428571428571429, + "grad_norm": 0.39176154136657715, + "learning_rate": 1.8156325911265756e-05, + "loss": 1.2878429889678955, + "step": 416 + }, + { + "epoch": 0.7464285714285714, + "grad_norm": 0.38600635528564453, + "learning_rate": 1.8134784273219345e-05, + "loss": 1.1995046138763428, + "step": 418 + }, + { + "epoch": 0.75, + "grad_norm": 0.3774026036262512, + "learning_rate": 1.8113132055768102e-05, + "loss": 1.284184217453003, + "step": 420 + }, + { + "epoch": 0.7535714285714286, + "grad_norm": 0.3247928321361542, + "learning_rate": 1.8091369594491805e-05, + "loss": 1.392996072769165, + "step": 422 + }, + { + "epoch": 0.7571428571428571, + "grad_norm": 0.41736844182014465, + "learning_rate": 1.8069497226678853e-05, + "loss": 1.5332679748535156, + "step": 424 + }, + { + "epoch": 0.7607142857142857, + "grad_norm": 0.6577937006950378, + "learning_rate": 1.8047515291321062e-05, + "loss": 1.2635902166366577, + "step": 426 + }, + { + "epoch": 0.7642857142857142, + "grad_norm": 0.5530592799186707, + "learning_rate": 1.802542412910838e-05, + "loss": 1.4461865425109863, + "step": 428 + }, + { + "epoch": 0.7678571428571429, + "grad_norm": 0.2581753432750702, + "learning_rate": 1.8003224082423634e-05, + "loss": 1.0822758674621582, + "step": 430 + }, + { + "epoch": 0.7714285714285715, + "grad_norm": 1.1362017393112183, + "learning_rate": 1.79809154953372e-05, + "loss": 1.272403359413147, + "step": 432 + }, + { + "epoch": 0.775, + "grad_norm": 0.2005966454744339, + "learning_rate": 1.7958498713601692e-05, + "loss": 1.1220377683639526, + "step": 434 + }, + { + "epoch": 0.7785714285714286, + "grad_norm": 0.5320938229560852, + "learning_rate": 1.7935974084646585e-05, + "loss": 0.9426363706588745, + "step": 436 + }, + { + "epoch": 0.7821428571428571, + "grad_norm": 0.5724827647209167, + "learning_rate": 1.7913341957572846e-05, + "loss": 1.35481858253479, + "step": 438 + }, + { + "epoch": 0.7857142857142857, + "grad_norm": 0.7783690690994263, + "learning_rate": 1.7890602683147515e-05, + "loss": 1.2968411445617676, + "step": 440 + }, + { + "epoch": 0.7892857142857143, + "grad_norm": 1.0548959970474243, + "learning_rate": 1.786775661379826e-05, + "loss": 1.4600225687026978, + "step": 442 + }, + { + "epoch": 0.7928571428571428, + "grad_norm": 0.41852042078971863, + "learning_rate": 1.7844804103607935e-05, + "loss": 1.274878978729248, + "step": 444 + }, + { + "epoch": 0.7964285714285714, + "grad_norm": 1.4088575839996338, + "learning_rate": 1.782174550830908e-05, + "loss": 1.4733103513717651, + "step": 446 + }, + { + "epoch": 0.8, + "grad_norm": 0.2099597156047821, + "learning_rate": 1.77985811852784e-05, + "loss": 1.112687349319458, + "step": 448 + }, + { + "epoch": 0.8035714285714286, + "grad_norm": 0.44812679290771484, + "learning_rate": 1.777531149353125e-05, + "loss": 1.2501091957092285, + "step": 450 + }, + { + "epoch": 0.8071428571428572, + "grad_norm": 1.0644121170043945, + "learning_rate": 1.7751936793716045e-05, + "loss": 1.2617871761322021, + "step": 452 + }, + { + "epoch": 0.8107142857142857, + "grad_norm": 0.6780351996421814, + "learning_rate": 1.7728457448108683e-05, + "loss": 1.270803689956665, + "step": 454 + }, + { + "epoch": 0.8142857142857143, + "grad_norm": 0.5940962433815002, + "learning_rate": 1.7704873820606932e-05, + "loss": 1.253537893295288, + "step": 456 + }, + { + "epoch": 0.8178571428571428, + "grad_norm": 0.3232758045196533, + "learning_rate": 1.768118627672479e-05, + "loss": 1.240506887435913, + "step": 458 + }, + { + "epoch": 0.8214285714285714, + "grad_norm": 0.7501121163368225, + "learning_rate": 1.765739518358681e-05, + "loss": 1.2568142414093018, + "step": 460 + }, + { + "epoch": 0.825, + "grad_norm": 1.307832956314087, + "learning_rate": 1.7633500909922413e-05, + "loss": 1.516510009765625, + "step": 462 + }, + { + "epoch": 0.8285714285714286, + "grad_norm": 0.7196856141090393, + "learning_rate": 1.760950382606019e-05, + "loss": 1.267121434211731, + "step": 464 + }, + { + "epoch": 0.8321428571428572, + "grad_norm": 0.27105146646499634, + "learning_rate": 1.7585404303922147e-05, + "loss": 1.211654782295227, + "step": 466 + }, + { + "epoch": 0.8357142857142857, + "grad_norm": 3.3737258911132812, + "learning_rate": 1.7561202717017933e-05, + "loss": 1.4474252462387085, + "step": 468 + }, + { + "epoch": 0.8392857142857143, + "grad_norm": 0.6952375769615173, + "learning_rate": 1.7536899440439066e-05, + "loss": 1.3239004611968994, + "step": 470 + }, + { + "epoch": 0.8428571428571429, + "grad_norm": 0.34531912207603455, + "learning_rate": 1.751249485085312e-05, + "loss": 1.492138147354126, + "step": 472 + }, + { + "epoch": 0.8464285714285714, + "grad_norm": 0.23176339268684387, + "learning_rate": 1.7487989326497878e-05, + "loss": 1.1729906797409058, + "step": 474 + }, + { + "epoch": 0.85, + "grad_norm": 0.6648816466331482, + "learning_rate": 1.746338324717548e-05, + "loss": 1.2582666873931885, + "step": 476 + }, + { + "epoch": 0.8535714285714285, + "grad_norm": 0.6908137202262878, + "learning_rate": 1.7438676994246515e-05, + "loss": 1.216418743133545, + "step": 478 + }, + { + "epoch": 0.8571428571428571, + "grad_norm": 0.8889988660812378, + "learning_rate": 1.7413870950624146e-05, + "loss": 1.2808120250701904, + "step": 480 + }, + { + "epoch": 0.8607142857142858, + "grad_norm": 0.5396649241447449, + "learning_rate": 1.7388965500768138e-05, + "loss": 1.3011693954467773, + "step": 482 + }, + { + "epoch": 0.8642857142857143, + "grad_norm": 0.4081900119781494, + "learning_rate": 1.736396103067893e-05, + "loss": 1.322187066078186, + "step": 484 + }, + { + "epoch": 0.8678571428571429, + "grad_norm": 3.7584619522094727, + "learning_rate": 1.733885792789163e-05, + "loss": 1.2870206832885742, + "step": 486 + }, + { + "epoch": 0.8714285714285714, + "grad_norm": 1.3534671068191528, + "learning_rate": 1.7313656581470025e-05, + "loss": 1.52611243724823, + "step": 488 + }, + { + "epoch": 0.875, + "grad_norm": 0.27716994285583496, + "learning_rate": 1.7288357382000544e-05, + "loss": 1.1492018699645996, + "step": 490 + }, + { + "epoch": 0.8785714285714286, + "grad_norm": 0.3399287760257721, + "learning_rate": 1.726296072158619e-05, + "loss": 1.170601725578308, + "step": 492 + }, + { + "epoch": 0.8821428571428571, + "grad_norm": 0.49777647852897644, + "learning_rate": 1.72374669938405e-05, + "loss": 1.689265251159668, + "step": 494 + }, + { + "epoch": 0.8857142857142857, + "grad_norm": 0.4312261641025543, + "learning_rate": 1.7211876593881404e-05, + "loss": 1.3613009452819824, + "step": 496 + }, + { + "epoch": 0.8892857142857142, + "grad_norm": 0.4204741418361664, + "learning_rate": 1.718618991832513e-05, + "loss": 1.2536442279815674, + "step": 498 + }, + { + "epoch": 0.8928571428571429, + "grad_norm": 1.1467082500457764, + "learning_rate": 1.716040736528004e-05, + "loss": 1.47482168674469, + "step": 500 + }, + { + "epoch": 0.8964285714285715, + "grad_norm": 0.3066917359828949, + "learning_rate": 1.7134529334340465e-05, + "loss": 1.1949682235717773, + "step": 502 + }, + { + "epoch": 0.9, + "grad_norm": 0.28838351368904114, + "learning_rate": 1.7108556226580524e-05, + "loss": 0.8847708702087402, + "step": 504 + }, + { + "epoch": 0.9035714285714286, + "grad_norm": 0.6763066053390503, + "learning_rate": 1.7082488444547883e-05, + "loss": 1.4644018411636353, + "step": 506 + }, + { + "epoch": 0.9071428571428571, + "grad_norm": 0.4099554121494293, + "learning_rate": 1.7056326392257535e-05, + "loss": 1.2191145420074463, + "step": 508 + }, + { + "epoch": 0.9107142857142857, + "grad_norm": 0.42622798681259155, + "learning_rate": 1.703007047518554e-05, + "loss": 1.4777641296386719, + "step": 510 + }, + { + "epoch": 0.9142857142857143, + "grad_norm": 0.29898732900619507, + "learning_rate": 1.7003721100262723e-05, + "loss": 1.1827527284622192, + "step": 512 + }, + { + "epoch": 0.9178571428571428, + "grad_norm": 0.1540851891040802, + "learning_rate": 1.6977278675868376e-05, + "loss": 1.115896463394165, + "step": 514 + }, + { + "epoch": 0.9214285714285714, + "grad_norm": 0.31116917729377747, + "learning_rate": 1.695074361182395e-05, + "loss": 1.376265287399292, + "step": 516 + }, + { + "epoch": 0.925, + "grad_norm": 0.3563463091850281, + "learning_rate": 1.6924116319386665e-05, + "loss": 1.10237455368042, + "step": 518 + }, + { + "epoch": 0.9285714285714286, + "grad_norm": 0.8681960701942444, + "learning_rate": 1.689739721124316e-05, + "loss": 1.2694408893585205, + "step": 520 + }, + { + "epoch": 0.9321428571428572, + "grad_norm": 0.3502778708934784, + "learning_rate": 1.687058670150309e-05, + "loss": 1.1879040002822876, + "step": 522 + }, + { + "epoch": 0.9357142857142857, + "grad_norm": 0.6757563352584839, + "learning_rate": 1.6843685205692724e-05, + "loss": 1.612572431564331, + "step": 524 + }, + { + "epoch": 0.9392857142857143, + "grad_norm": 0.5540065765380859, + "learning_rate": 1.681669314074847e-05, + "loss": 1.2043344974517822, + "step": 526 + }, + { + "epoch": 0.9428571428571428, + "grad_norm": 0.5378439426422119, + "learning_rate": 1.6789610925010448e-05, + "loss": 1.2300772666931152, + "step": 528 + }, + { + "epoch": 0.9464285714285714, + "grad_norm": 0.3303406536579132, + "learning_rate": 1.6762438978215984e-05, + "loss": 0.998758852481842, + "step": 530 + }, + { + "epoch": 0.95, + "grad_norm": 0.5480644106864929, + "learning_rate": 1.673517772149312e-05, + "loss": 1.2909802198410034, + "step": 532 + }, + { + "epoch": 0.9535714285714286, + "grad_norm": 0.2534736096858978, + "learning_rate": 1.6707827577354072e-05, + "loss": 1.2177340984344482, + "step": 534 + }, + { + "epoch": 0.9571428571428572, + "grad_norm": 0.28689199686050415, + "learning_rate": 1.66803889696887e-05, + "loss": 1.5511302947998047, + "step": 536 + }, + { + "epoch": 0.9607142857142857, + "grad_norm": 0.3565874993801117, + "learning_rate": 1.6652862323757914e-05, + "loss": 1.2498445510864258, + "step": 538 + }, + { + "epoch": 0.9642857142857143, + "grad_norm": 0.6810623407363892, + "learning_rate": 1.662524806618711e-05, + "loss": 1.2677786350250244, + "step": 540 + }, + { + "epoch": 0.9678571428571429, + "grad_norm": 0.9204868078231812, + "learning_rate": 1.6597546624959534e-05, + "loss": 1.2525708675384521, + "step": 542 + }, + { + "epoch": 0.9714285714285714, + "grad_norm": 0.39553532004356384, + "learning_rate": 1.656975842940967e-05, + "loss": 1.5089502334594727, + "step": 544 + }, + { + "epoch": 0.975, + "grad_norm": 2.675163984298706, + "learning_rate": 1.6541883910216562e-05, + "loss": 1.4569265842437744, + "step": 546 + }, + { + "epoch": 0.9785714285714285, + "grad_norm": 0.6487977504730225, + "learning_rate": 1.6513923499397165e-05, + "loss": 1.218340516090393, + "step": 548 + }, + { + "epoch": 0.9821428571428571, + "grad_norm": 0.7652715444564819, + "learning_rate": 1.6485877630299633e-05, + "loss": 0.6461201310157776, + "step": 550 + }, + { + "epoch": 0.9857142857142858, + "grad_norm": 0.3120361566543579, + "learning_rate": 1.6457746737596608e-05, + "loss": 0.6979402303695679, + "step": 552 + }, + { + "epoch": 0.9892857142857143, + "grad_norm": 0.3177029490470886, + "learning_rate": 1.642953125727847e-05, + "loss": 1.2062575817108154, + "step": 554 + }, + { + "epoch": 0.9928571428571429, + "grad_norm": 0.3227110803127289, + "learning_rate": 1.6401231626646612e-05, + "loss": 1.3390659093856812, + "step": 556 + }, + { + "epoch": 0.9964285714285714, + "grad_norm": 7.338903903961182, + "learning_rate": 1.637284828430662e-05, + "loss": 0.8427339792251587, + "step": 558 + }, + { + "epoch": 1.0, + "grad_norm": 0.5971918106079102, + "learning_rate": 1.6344381670161514e-05, + "loss": 1.6298896074295044, + "step": 560 + }, + { + "epoch": 1.0035714285714286, + "grad_norm": 0.4188767075538635, + "learning_rate": 1.6315832225404905e-05, + "loss": 1.122739315032959, + "step": 562 + }, + { + "epoch": 1.0071428571428571, + "grad_norm": 0.6393467783927917, + "learning_rate": 1.6287200392514172e-05, + "loss": 1.4107418060302734, + "step": 564 + }, + { + "epoch": 1.0107142857142857, + "grad_norm": 0.7803854942321777, + "learning_rate": 1.6258486615243583e-05, + "loss": 1.3382797241210938, + "step": 566 + }, + { + "epoch": 1.0142857142857142, + "grad_norm": 0.8164628744125366, + "learning_rate": 1.6229691338617447e-05, + "loss": 1.170792579650879, + "step": 568 + }, + { + "epoch": 1.0178571428571428, + "grad_norm": 0.5098171234130859, + "learning_rate": 1.620081500892319e-05, + "loss": 1.1756045818328857, + "step": 570 + }, + { + "epoch": 1.0214285714285714, + "grad_norm": 0.35069453716278076, + "learning_rate": 1.6171858073704472e-05, + "loss": 1.2409474849700928, + "step": 572 + }, + { + "epoch": 1.025, + "grad_norm": 0.5415301322937012, + "learning_rate": 1.6142820981754194e-05, + "loss": 1.2160595655441284, + "step": 574 + }, + { + "epoch": 1.0285714285714285, + "grad_norm": 0.3839040994644165, + "learning_rate": 1.61137041831076e-05, + "loss": 1.2669594287872314, + "step": 576 + }, + { + "epoch": 1.032142857142857, + "grad_norm": 0.3261508047580719, + "learning_rate": 1.6084508129035285e-05, + "loss": 1.2169692516326904, + "step": 578 + }, + { + "epoch": 1.0357142857142858, + "grad_norm": 2.2598116397857666, + "learning_rate": 1.605523327203617e-05, + "loss": 1.412977933883667, + "step": 580 + }, + { + "epoch": 1.0392857142857144, + "grad_norm": 0.43175196647644043, + "learning_rate": 1.6025880065830527e-05, + "loss": 1.138507604598999, + "step": 582 + }, + { + "epoch": 1.042857142857143, + "grad_norm": 1.0159385204315186, + "learning_rate": 1.5996448965352946e-05, + "loss": 1.2821019887924194, + "step": 584 + }, + { + "epoch": 1.0464285714285715, + "grad_norm": 0.5554090738296509, + "learning_rate": 1.596694042674525e-05, + "loss": 1.3544402122497559, + "step": 586 + }, + { + "epoch": 1.05, + "grad_norm": 0.4888118505477905, + "learning_rate": 1.593735490734946e-05, + "loss": 1.1748173236846924, + "step": 588 + }, + { + "epoch": 1.0535714285714286, + "grad_norm": 0.8884871602058411, + "learning_rate": 1.590769286570069e-05, + "loss": 1.4871482849121094, + "step": 590 + }, + { + "epoch": 1.0571428571428572, + "grad_norm": 1.121586799621582, + "learning_rate": 1.587795476152005e-05, + "loss": 1.3661131858825684, + "step": 592 + }, + { + "epoch": 1.0607142857142857, + "grad_norm": 0.33232274651527405, + "learning_rate": 1.58481410557075e-05, + "loss": 1.1384379863739014, + "step": 594 + }, + { + "epoch": 1.0642857142857143, + "grad_norm": 0.5710160136222839, + "learning_rate": 1.5818252210334746e-05, + "loss": 1.2044832706451416, + "step": 596 + }, + { + "epoch": 1.0678571428571428, + "grad_norm": 0.4304794669151306, + "learning_rate": 1.578828868863803e-05, + "loss": 1.1721159219741821, + "step": 598 + }, + { + "epoch": 1.0714285714285714, + "grad_norm": 0.29414913058280945, + "learning_rate": 1.575825095501099e-05, + "loss": 1.3607425689697266, + "step": 600 + }, + { + "epoch": 1.075, + "grad_norm": 0.3909018933773041, + "learning_rate": 1.5728139474997445e-05, + "loss": 1.3812944889068604, + "step": 602 + }, + { + "epoch": 1.0785714285714285, + "grad_norm": 0.7157077789306641, + "learning_rate": 1.5697954715284177e-05, + "loss": 1.3088464736938477, + "step": 604 + }, + { + "epoch": 1.082142857142857, + "grad_norm": 0.4474976062774658, + "learning_rate": 1.566769714369371e-05, + "loss": 1.2116174697875977, + "step": 606 + }, + { + "epoch": 1.0857142857142856, + "grad_norm": 0.3516829013824463, + "learning_rate": 1.5637367229177046e-05, + "loss": 1.199174404144287, + "step": 608 + }, + { + "epoch": 1.0892857142857142, + "grad_norm": 0.4679516851902008, + "learning_rate": 1.560696544180641e-05, + "loss": 1.4164745807647705, + "step": 610 + }, + { + "epoch": 1.092857142857143, + "grad_norm": 0.47901520133018494, + "learning_rate": 1.5576492252767954e-05, + "loss": 1.189131259918213, + "step": 612 + }, + { + "epoch": 1.0964285714285715, + "grad_norm": 2.753019094467163, + "learning_rate": 1.554594813435446e-05, + "loss": 1.3877692222595215, + "step": 614 + }, + { + "epoch": 1.1, + "grad_norm": 0.3647393584251404, + "learning_rate": 1.5515333559958015e-05, + "loss": 1.1784098148345947, + "step": 616 + }, + { + "epoch": 1.1035714285714286, + "grad_norm": 0.517415463924408, + "learning_rate": 1.548464900406268e-05, + "loss": 1.1289467811584473, + "step": 618 + }, + { + "epoch": 1.1071428571428572, + "grad_norm": 1.6620982885360718, + "learning_rate": 1.545389494223714e-05, + "loss": 1.1143990755081177, + "step": 620 + }, + { + "epoch": 1.1107142857142858, + "grad_norm": 0.9535795450210571, + "learning_rate": 1.542307185112731e-05, + "loss": 1.2819159030914307, + "step": 622 + }, + { + "epoch": 1.1142857142857143, + "grad_norm": 0.34654900431632996, + "learning_rate": 1.5392180208448984e-05, + "loss": 1.1823941469192505, + "step": 624 + }, + { + "epoch": 1.1178571428571429, + "grad_norm": 0.46996039152145386, + "learning_rate": 1.5361220492980398e-05, + "loss": 1.1919305324554443, + "step": 626 + }, + { + "epoch": 1.1214285714285714, + "grad_norm": 0.2986307144165039, + "learning_rate": 1.533019318455483e-05, + "loss": 1.0924913883209229, + "step": 628 + }, + { + "epoch": 1.125, + "grad_norm": 0.25168025493621826, + "learning_rate": 1.529909876405315e-05, + "loss": 0.9394223690032959, + "step": 630 + }, + { + "epoch": 1.1285714285714286, + "grad_norm": 0.6047491431236267, + "learning_rate": 1.5267937713396384e-05, + "loss": 1.1416627168655396, + "step": 632 + }, + { + "epoch": 1.1321428571428571, + "grad_norm": 0.2771312892436981, + "learning_rate": 1.5236710515538223e-05, + "loss": 1.3538787364959717, + "step": 634 + }, + { + "epoch": 1.1357142857142857, + "grad_norm": 0.9521869421005249, + "learning_rate": 1.5205417654457559e-05, + "loss": 1.3267796039581299, + "step": 636 + }, + { + "epoch": 1.1392857142857142, + "grad_norm": 0.33584362268447876, + "learning_rate": 1.5174059615150965e-05, + "loss": 1.0853066444396973, + "step": 638 + }, + { + "epoch": 1.1428571428571428, + "grad_norm": 0.3648921251296997, + "learning_rate": 1.5142636883625197e-05, + "loss": 1.1428154706954956, + "step": 640 + }, + { + "epoch": 1.1464285714285714, + "grad_norm": 1.2066301107406616, + "learning_rate": 1.511114994688964e-05, + "loss": 0.8075799345970154, + "step": 642 + }, + { + "epoch": 1.15, + "grad_norm": 3.704089879989624, + "learning_rate": 1.5079599292948785e-05, + "loss": 1.2435736656188965, + "step": 644 + }, + { + "epoch": 1.1535714285714285, + "grad_norm": 0.711872935295105, + "learning_rate": 1.5047985410794641e-05, + "loss": 1.0868031978607178, + "step": 646 + }, + { + "epoch": 1.157142857142857, + "grad_norm": 0.5934475064277649, + "learning_rate": 1.5016308790399183e-05, + "loss": 1.1679191589355469, + "step": 648 + }, + { + "epoch": 1.1607142857142858, + "grad_norm": 0.2927475571632385, + "learning_rate": 1.4984569922706722e-05, + "loss": 0.7266023755073547, + "step": 650 + }, + { + "epoch": 1.1642857142857144, + "grad_norm": 0.4356221556663513, + "learning_rate": 1.4952769299626335e-05, + "loss": 1.1431584358215332, + "step": 652 + }, + { + "epoch": 1.167857142857143, + "grad_norm": 0.5513753294944763, + "learning_rate": 1.4920907414024215e-05, + "loss": 1.1460659503936768, + "step": 654 + }, + { + "epoch": 1.1714285714285715, + "grad_norm": 0.5610360503196716, + "learning_rate": 1.4888984759716041e-05, + "loss": 1.1650118827819824, + "step": 656 + }, + { + "epoch": 1.175, + "grad_norm": 0.405282586812973, + "learning_rate": 1.4857001831459326e-05, + "loss": 1.1098073720932007, + "step": 658 + }, + { + "epoch": 1.1785714285714286, + "grad_norm": 0.3328110873699188, + "learning_rate": 1.4824959124945746e-05, + "loss": 1.1764960289001465, + "step": 660 + }, + { + "epoch": 1.1821428571428572, + "grad_norm": 1.3929952383041382, + "learning_rate": 1.4792857136793457e-05, + "loss": 1.2145479917526245, + "step": 662 + }, + { + "epoch": 1.1857142857142857, + "grad_norm": 0.8867573738098145, + "learning_rate": 1.4760696364539402e-05, + "loss": 0.9081100225448608, + "step": 664 + }, + { + "epoch": 1.1892857142857143, + "grad_norm": 0.3415173590183258, + "learning_rate": 1.472847730663159e-05, + "loss": 1.1348457336425781, + "step": 666 + }, + { + "epoch": 1.1928571428571428, + "grad_norm": 0.4172976016998291, + "learning_rate": 1.4696200462421393e-05, + "loss": 1.1158576011657715, + "step": 668 + }, + { + "epoch": 1.1964285714285714, + "grad_norm": 0.447843998670578, + "learning_rate": 1.4663866332155772e-05, + "loss": 1.2849993705749512, + "step": 670 + }, + { + "epoch": 1.2, + "grad_norm": 0.9494065046310425, + "learning_rate": 1.463147541696956e-05, + "loss": 1.3352279663085938, + "step": 672 + }, + { + "epoch": 1.2035714285714285, + "grad_norm": 0.621944785118103, + "learning_rate": 1.459902821887767e-05, + "loss": 1.2335541248321533, + "step": 674 + }, + { + "epoch": 1.207142857142857, + "grad_norm": 0.48863765597343445, + "learning_rate": 1.4566525240767328e-05, + "loss": 1.3280656337738037, + "step": 676 + }, + { + "epoch": 1.2107142857142856, + "grad_norm": 0.5822184681892395, + "learning_rate": 1.4533966986390266e-05, + "loss": 1.265816569328308, + "step": 678 + }, + { + "epoch": 1.2142857142857142, + "grad_norm": 0.5110518932342529, + "learning_rate": 1.4501353960354935e-05, + "loss": 1.225487232208252, + "step": 680 + }, + { + "epoch": 1.217857142857143, + "grad_norm": 0.8750176429748535, + "learning_rate": 1.4468686668118663e-05, + "loss": 1.2301876544952393, + "step": 682 + }, + { + "epoch": 1.2214285714285715, + "grad_norm": 0.8533264398574829, + "learning_rate": 1.443596561597983e-05, + "loss": 1.161440134048462, + "step": 684 + }, + { + "epoch": 1.225, + "grad_norm": 0.44423893094062805, + "learning_rate": 1.4403191311070022e-05, + "loss": 1.0979464054107666, + "step": 686 + }, + { + "epoch": 1.2285714285714286, + "grad_norm": 0.43870681524276733, + "learning_rate": 1.4370364261346175e-05, + "loss": 1.1086313724517822, + "step": 688 + }, + { + "epoch": 1.2321428571428572, + "grad_norm": 0.49559494853019714, + "learning_rate": 1.433748497558269e-05, + "loss": 1.0904359817504883, + "step": 690 + }, + { + "epoch": 1.2357142857142858, + "grad_norm": 0.37396514415740967, + "learning_rate": 1.4304553963363563e-05, + "loss": 1.1759669780731201, + "step": 692 + }, + { + "epoch": 1.2392857142857143, + "grad_norm": 0.4118036925792694, + "learning_rate": 1.427157173507447e-05, + "loss": 1.157477855682373, + "step": 694 + }, + { + "epoch": 1.2428571428571429, + "grad_norm": 0.7096890211105347, + "learning_rate": 1.4238538801894875e-05, + "loss": 1.1737724542617798, + "step": 696 + }, + { + "epoch": 1.2464285714285714, + "grad_norm": 0.6166890263557434, + "learning_rate": 1.4205455675790097e-05, + "loss": 1.2456121444702148, + "step": 698 + }, + { + "epoch": 1.25, + "grad_norm": 1.1848427057266235, + "learning_rate": 1.4172322869503368e-05, + "loss": 0.8353923559188843, + "step": 700 + }, + { + "epoch": 1.2535714285714286, + "grad_norm": 0.38646364212036133, + "learning_rate": 1.4139140896547902e-05, + "loss": 0.7373632192611694, + "step": 702 + }, + { + "epoch": 1.2571428571428571, + "grad_norm": 0.6813467144966125, + "learning_rate": 1.4105910271198937e-05, + "loss": 1.3092610836029053, + "step": 704 + }, + { + "epoch": 1.2607142857142857, + "grad_norm": 0.3635100722312927, + "learning_rate": 1.407263150848574e-05, + "loss": 1.2204773426055908, + "step": 706 + }, + { + "epoch": 1.2642857142857142, + "grad_norm": 0.507089376449585, + "learning_rate": 1.4039305124183653e-05, + "loss": 1.182018756866455, + "step": 708 + }, + { + "epoch": 1.2678571428571428, + "grad_norm": 0.27979063987731934, + "learning_rate": 1.4005931634806085e-05, + "loss": 1.095738410949707, + "step": 710 + }, + { + "epoch": 1.2714285714285714, + "grad_norm": 0.4857190251350403, + "learning_rate": 1.3972511557596506e-05, + "loss": 1.1410452127456665, + "step": 712 + }, + { + "epoch": 1.275, + "grad_norm": 0.3612891435623169, + "learning_rate": 1.3939045410520435e-05, + "loss": 1.1210098266601562, + "step": 714 + }, + { + "epoch": 1.2785714285714285, + "grad_norm": 0.3975180685520172, + "learning_rate": 1.3905533712257418e-05, + "loss": 1.2363688945770264, + "step": 716 + }, + { + "epoch": 1.282142857142857, + "grad_norm": 0.9986791610717773, + "learning_rate": 1.3871976982192971e-05, + "loss": 1.1617156267166138, + "step": 718 + }, + { + "epoch": 1.2857142857142856, + "grad_norm": 1.0810168981552124, + "learning_rate": 1.383837574041055e-05, + "loss": 1.2290289402008057, + "step": 720 + }, + { + "epoch": 1.2892857142857144, + "grad_norm": 0.4217285215854645, + "learning_rate": 1.3804730507683473e-05, + "loss": 1.1525651216506958, + "step": 722 + }, + { + "epoch": 1.292857142857143, + "grad_norm": 0.40972352027893066, + "learning_rate": 1.377104180546687e-05, + "loss": 1.1072614192962646, + "step": 724 + }, + { + "epoch": 1.2964285714285715, + "grad_norm": 0.3110302686691284, + "learning_rate": 1.3737310155889575e-05, + "loss": 1.1427106857299805, + "step": 726 + }, + { + "epoch": 1.3, + "grad_norm": 0.4348861873149872, + "learning_rate": 1.370353608174606e-05, + "loss": 0.9627160429954529, + "step": 728 + }, + { + "epoch": 1.3035714285714286, + "grad_norm": 0.7683548331260681, + "learning_rate": 1.3669720106488308e-05, + "loss": 1.2495923042297363, + "step": 730 + }, + { + "epoch": 1.3071428571428572, + "grad_norm": 1.5341452360153198, + "learning_rate": 1.3635862754217725e-05, + "loss": 1.1818249225616455, + "step": 732 + }, + { + "epoch": 1.3107142857142857, + "grad_norm": 0.359321653842926, + "learning_rate": 1.360196454967699e-05, + "loss": 1.0857856273651123, + "step": 734 + }, + { + "epoch": 1.3142857142857143, + "grad_norm": 0.5810532569885254, + "learning_rate": 1.356802601824195e-05, + "loss": 1.1771756410598755, + "step": 736 + }, + { + "epoch": 1.3178571428571428, + "grad_norm": 0.7885570526123047, + "learning_rate": 1.353404768591345e-05, + "loss": 1.1475200653076172, + "step": 738 + }, + { + "epoch": 1.3214285714285714, + "grad_norm": 1.2528671026229858, + "learning_rate": 1.3500030079309206e-05, + "loss": 1.1383891105651855, + "step": 740 + }, + { + "epoch": 1.325, + "grad_norm": 4.210450649261475, + "learning_rate": 1.3465973725655625e-05, + "loss": 1.0303577184677124, + "step": 742 + }, + { + "epoch": 1.3285714285714285, + "grad_norm": 1.3920871019363403, + "learning_rate": 1.3431879152779643e-05, + "loss": 1.1899462938308716, + "step": 744 + }, + { + "epoch": 1.332142857142857, + "grad_norm": 0.3303053081035614, + "learning_rate": 1.3397746889100542e-05, + "loss": 1.118945598602295, + "step": 746 + }, + { + "epoch": 1.3357142857142856, + "grad_norm": 0.4941954016685486, + "learning_rate": 1.336357746362176e-05, + "loss": 1.1191518306732178, + "step": 748 + }, + { + "epoch": 1.3392857142857144, + "grad_norm": 0.4031634032726288, + "learning_rate": 1.3329371405922688e-05, + "loss": 0.8386397361755371, + "step": 750 + }, + { + "epoch": 1.342857142857143, + "grad_norm": 0.44179248809814453, + "learning_rate": 1.3295129246150472e-05, + "loss": 0.6230043172836304, + "step": 752 + }, + { + "epoch": 1.3464285714285715, + "grad_norm": 0.4847067594528198, + "learning_rate": 1.3260851515011788e-05, + "loss": 1.3626277446746826, + "step": 754 + }, + { + "epoch": 1.35, + "grad_norm": 0.2796545624732971, + "learning_rate": 1.3226538743764617e-05, + "loss": 1.1133283376693726, + "step": 756 + }, + { + "epoch": 1.3535714285714286, + "grad_norm": 1.6174930334091187, + "learning_rate": 1.3192191464210023e-05, + "loss": 1.1741091012954712, + "step": 758 + }, + { + "epoch": 1.3571428571428572, + "grad_norm": 0.7611403465270996, + "learning_rate": 1.3157810208683887e-05, + "loss": 0.9258888959884644, + "step": 760 + }, + { + "epoch": 1.3607142857142858, + "grad_norm": 0.38805851340293884, + "learning_rate": 1.3123395510048687e-05, + "loss": 1.196305274963379, + "step": 762 + }, + { + "epoch": 1.3642857142857143, + "grad_norm": 1.080810785293579, + "learning_rate": 1.3088947901685212e-05, + "loss": 0.9452080726623535, + "step": 764 + }, + { + "epoch": 1.3678571428571429, + "grad_norm": 0.8032243847846985, + "learning_rate": 1.3054467917484308e-05, + "loss": 1.3384945392608643, + "step": 766 + }, + { + "epoch": 1.3714285714285714, + "grad_norm": 0.7725457549095154, + "learning_rate": 1.3019956091838614e-05, + "loss": 1.093704104423523, + "step": 768 + }, + { + "epoch": 1.375, + "grad_norm": 0.5379224419593811, + "learning_rate": 1.298541295963425e-05, + "loss": 0.9955064058303833, + "step": 770 + }, + { + "epoch": 1.3785714285714286, + "grad_norm": 0.6937380433082581, + "learning_rate": 1.2950839056242557e-05, + "loss": 1.1630027294158936, + "step": 772 + }, + { + "epoch": 1.3821428571428571, + "grad_norm": 0.6541162729263306, + "learning_rate": 1.291623491751178e-05, + "loss": 1.1202328205108643, + "step": 774 + }, + { + "epoch": 1.3857142857142857, + "grad_norm": 0.4741293787956238, + "learning_rate": 1.2881601079758784e-05, + "loss": 1.1508142948150635, + "step": 776 + }, + { + "epoch": 1.3892857142857142, + "grad_norm": 0.631596565246582, + "learning_rate": 1.284693807976071e-05, + "loss": 1.3536005020141602, + "step": 778 + }, + { + "epoch": 1.3928571428571428, + "grad_norm": 0.30716219544410706, + "learning_rate": 1.2812246454746687e-05, + "loss": 1.1645984649658203, + "step": 780 + }, + { + "epoch": 1.3964285714285714, + "grad_norm": 0.4866527318954468, + "learning_rate": 1.2777526742389483e-05, + "loss": 1.1908864974975586, + "step": 782 + }, + { + "epoch": 1.4, + "grad_norm": 0.5402504801750183, + "learning_rate": 1.2742779480797194e-05, + "loss": 1.0720609426498413, + "step": 784 + }, + { + "epoch": 1.4035714285714285, + "grad_norm": 0.5585693120956421, + "learning_rate": 1.270800520850488e-05, + "loss": 1.1451340913772583, + "step": 786 + }, + { + "epoch": 1.407142857142857, + "grad_norm": 0.6004486680030823, + "learning_rate": 1.2673204464466233e-05, + "loss": 1.1766554117202759, + "step": 788 + }, + { + "epoch": 1.4107142857142856, + "grad_norm": 0.8663070797920227, + "learning_rate": 1.2638377788045223e-05, + "loss": 1.087449312210083, + "step": 790 + }, + { + "epoch": 1.4142857142857144, + "grad_norm": 0.8936390280723572, + "learning_rate": 1.2603525719007738e-05, + "loss": 1.0832384824752808, + "step": 792 + }, + { + "epoch": 1.417857142857143, + "grad_norm": 0.62984699010849, + "learning_rate": 1.2568648797513212e-05, + "loss": 1.1712042093276978, + "step": 794 + }, + { + "epoch": 1.4214285714285715, + "grad_norm": 0.6213272213935852, + "learning_rate": 1.2533747564106262e-05, + "loss": 1.0903995037078857, + "step": 796 + }, + { + "epoch": 1.425, + "grad_norm": 0.5741475224494934, + "learning_rate": 1.2498822559708304e-05, + "loss": 1.1080608367919922, + "step": 798 + }, + { + "epoch": 1.4285714285714286, + "grad_norm": 0.6108075380325317, + "learning_rate": 1.2463874325609168e-05, + "loss": 1.1283464431762695, + "step": 800 + }, + { + "epoch": 1.4321428571428572, + "grad_norm": 1.3531321287155151, + "learning_rate": 1.2428903403458725e-05, + "loss": 0.9932132959365845, + "step": 802 + }, + { + "epoch": 1.4357142857142857, + "grad_norm": 0.31752732396125793, + "learning_rate": 1.2393910335258472e-05, + "loss": 1.1548457145690918, + "step": 804 + }, + { + "epoch": 1.4392857142857143, + "grad_norm": 0.3582463562488556, + "learning_rate": 1.2358895663353132e-05, + "loss": 1.1124224662780762, + "step": 806 + }, + { + "epoch": 1.4428571428571428, + "grad_norm": 0.7238558530807495, + "learning_rate": 1.232385993042227e-05, + "loss": 1.2339575290679932, + "step": 808 + }, + { + "epoch": 1.4464285714285714, + "grad_norm": 0.6423646211624146, + "learning_rate": 1.2288803679471861e-05, + "loss": 0.7599235773086548, + "step": 810 + }, + { + "epoch": 1.45, + "grad_norm": 0.41928139328956604, + "learning_rate": 1.225372745382588e-05, + "loss": 1.160172939300537, + "step": 812 + }, + { + "epoch": 1.4535714285714285, + "grad_norm": 0.5137944221496582, + "learning_rate": 1.2218631797117885e-05, + "loss": 1.1503641605377197, + "step": 814 + }, + { + "epoch": 1.457142857142857, + "grad_norm": 0.3738017678260803, + "learning_rate": 1.2183517253282591e-05, + "loss": 1.0606850385665894, + "step": 816 + }, + { + "epoch": 1.4607142857142856, + "grad_norm": 0.9808754920959473, + "learning_rate": 1.2148384366547428e-05, + "loss": 0.6364516019821167, + "step": 818 + }, + { + "epoch": 1.4642857142857144, + "grad_norm": 0.4231683313846588, + "learning_rate": 1.211323368142413e-05, + "loss": 1.2867590188980103, + "step": 820 + }, + { + "epoch": 1.467857142857143, + "grad_norm": 0.40807411074638367, + "learning_rate": 1.2078065742700272e-05, + "loss": 1.2712998390197754, + "step": 822 + }, + { + "epoch": 1.4714285714285715, + "grad_norm": 0.4421752393245697, + "learning_rate": 1.2042881095430836e-05, + "loss": 1.084946632385254, + "step": 824 + }, + { + "epoch": 1.475, + "grad_norm": 0.43936818838119507, + "learning_rate": 1.2007680284929773e-05, + "loss": 1.1451640129089355, + "step": 826 + }, + { + "epoch": 1.4785714285714286, + "grad_norm": 0.4863755702972412, + "learning_rate": 1.1972463856761529e-05, + "loss": 0.8599765300750732, + "step": 828 + }, + { + "epoch": 1.4821428571428572, + "grad_norm": 0.5459559559822083, + "learning_rate": 1.1937232356732609e-05, + "loss": 1.2189276218414307, + "step": 830 + }, + { + "epoch": 1.4857142857142858, + "grad_norm": 0.7105582356452942, + "learning_rate": 1.190198633088312e-05, + "loss": 1.143498420715332, + "step": 832 + }, + { + "epoch": 1.4892857142857143, + "grad_norm": 0.5701984763145447, + "learning_rate": 1.1866726325478277e-05, + "loss": 1.1005926132202148, + "step": 834 + }, + { + "epoch": 1.4928571428571429, + "grad_norm": 0.609856903553009, + "learning_rate": 1.1831452886999984e-05, + "loss": 1.1103310585021973, + "step": 836 + }, + { + "epoch": 1.4964285714285714, + "grad_norm": 0.7349211573600769, + "learning_rate": 1.179616656213832e-05, + "loss": 0.8592168092727661, + "step": 838 + }, + { + "epoch": 1.5, + "grad_norm": 0.4150165617465973, + "learning_rate": 1.1760867897783097e-05, + "loss": 1.1124815940856934, + "step": 840 + }, + { + "epoch": 1.5035714285714286, + "grad_norm": 0.36004072427749634, + "learning_rate": 1.1725557441015369e-05, + "loss": 1.1074330806732178, + "step": 842 + }, + { + "epoch": 1.5071428571428571, + "grad_norm": 0.5040198564529419, + "learning_rate": 1.1690235739098953e-05, + "loss": 1.1491334438323975, + "step": 844 + }, + { + "epoch": 1.5107142857142857, + "grad_norm": 1.0856701135635376, + "learning_rate": 1.1654903339471954e-05, + "loss": 1.2160296440124512, + "step": 846 + }, + { + "epoch": 1.5142857142857142, + "grad_norm": 0.4830666780471802, + "learning_rate": 1.161956078973828e-05, + "loss": 1.051498293876648, + "step": 848 + }, + { + "epoch": 1.5178571428571428, + "grad_norm": 1.711748480796814, + "learning_rate": 1.158420863765914e-05, + "loss": 0.8320801258087158, + "step": 850 + }, + { + "epoch": 1.5214285714285714, + "grad_norm": 0.5154921412467957, + "learning_rate": 1.1548847431144578e-05, + "loss": 1.3413938283920288, + "step": 852 + }, + { + "epoch": 1.525, + "grad_norm": 0.41332578659057617, + "learning_rate": 1.1513477718244967e-05, + "loss": 1.0733758211135864, + "step": 854 + }, + { + "epoch": 1.5285714285714285, + "grad_norm": 0.9633269309997559, + "learning_rate": 1.1478100047142516e-05, + "loss": 1.2646903991699219, + "step": 856 + }, + { + "epoch": 1.532142857142857, + "grad_norm": 0.38085678219795227, + "learning_rate": 1.1442714966142773e-05, + "loss": 1.1131600141525269, + "step": 858 + }, + { + "epoch": 1.5357142857142856, + "grad_norm": 0.6264724731445312, + "learning_rate": 1.1407323023666127e-05, + "loss": 1.1462935209274292, + "step": 860 + }, + { + "epoch": 1.5392857142857141, + "grad_norm": 3.2465555667877197, + "learning_rate": 1.137192476823932e-05, + "loss": 1.2240748405456543, + "step": 862 + }, + { + "epoch": 1.5428571428571427, + "grad_norm": 0.5931512117385864, + "learning_rate": 1.1336520748486934e-05, + "loss": 1.1074802875518799, + "step": 864 + }, + { + "epoch": 1.5464285714285713, + "grad_norm": 0.6654783487319946, + "learning_rate": 1.1301111513122877e-05, + "loss": 1.1554946899414062, + "step": 866 + }, + { + "epoch": 1.55, + "grad_norm": 0.8768860101699829, + "learning_rate": 1.1265697610941915e-05, + "loss": 1.1281225681304932, + "step": 868 + }, + { + "epoch": 1.5535714285714286, + "grad_norm": 0.4611034691333771, + "learning_rate": 1.1230279590811118e-05, + "loss": 1.2249135971069336, + "step": 870 + }, + { + "epoch": 1.5571428571428572, + "grad_norm": 0.5411264896392822, + "learning_rate": 1.11948580016614e-05, + "loss": 1.0645391941070557, + "step": 872 + }, + { + "epoch": 1.5607142857142857, + "grad_norm": 0.31846562027931213, + "learning_rate": 1.1159433392478973e-05, + "loss": 1.0551997423171997, + "step": 874 + }, + { + "epoch": 1.5642857142857143, + "grad_norm": 0.3621332347393036, + "learning_rate": 1.1124006312296869e-05, + "loss": 1.0810638666152954, + "step": 876 + }, + { + "epoch": 1.5678571428571428, + "grad_norm": 0.5022686123847961, + "learning_rate": 1.1088577310186406e-05, + "loss": 1.1075555086135864, + "step": 878 + }, + { + "epoch": 1.5714285714285714, + "grad_norm": 1.8955113887786865, + "learning_rate": 1.1053146935248701e-05, + "loss": 1.0974161624908447, + "step": 880 + }, + { + "epoch": 1.575, + "grad_norm": 0.26390373706817627, + "learning_rate": 1.1017715736606137e-05, + "loss": 0.9959229230880737, + "step": 882 + }, + { + "epoch": 1.5785714285714287, + "grad_norm": 0.32717210054397583, + "learning_rate": 1.0982284263393868e-05, + "loss": 1.2565701007843018, + "step": 884 + }, + { + "epoch": 1.5821428571428573, + "grad_norm": 0.36438724398612976, + "learning_rate": 1.0946853064751301e-05, + "loss": 1.256880760192871, + "step": 886 + }, + { + "epoch": 1.5857142857142859, + "grad_norm": 0.41332828998565674, + "learning_rate": 1.0911422689813594e-05, + "loss": 1.090340495109558, + "step": 888 + }, + { + "epoch": 1.5892857142857144, + "grad_norm": 0.6816041469573975, + "learning_rate": 1.0875993687703134e-05, + "loss": 1.1211128234863281, + "step": 890 + }, + { + "epoch": 1.592857142857143, + "grad_norm": 0.4995070993900299, + "learning_rate": 1.084056660752103e-05, + "loss": 1.1537326574325562, + "step": 892 + }, + { + "epoch": 1.5964285714285715, + "grad_norm": 0.5382480621337891, + "learning_rate": 1.0805141998338607e-05, + "loss": 1.1503942012786865, + "step": 894 + }, + { + "epoch": 1.6, + "grad_norm": 0.22450421750545502, + "learning_rate": 1.0769720409188883e-05, + "loss": 1.068176031112671, + "step": 896 + }, + { + "epoch": 1.6035714285714286, + "grad_norm": 0.7616103291511536, + "learning_rate": 1.073430238905809e-05, + "loss": 1.0579860210418701, + "step": 898 + }, + { + "epoch": 1.6071428571428572, + "grad_norm": 0.4256003499031067, + "learning_rate": 1.0698888486877126e-05, + "loss": 1.016850471496582, + "step": 900 + }, + { + "epoch": 1.6107142857142858, + "grad_norm": 0.43691617250442505, + "learning_rate": 1.066347925151307e-05, + "loss": 1.052515983581543, + "step": 902 + }, + { + "epoch": 1.6142857142857143, + "grad_norm": 1.1696960926055908, + "learning_rate": 1.0628075231760682e-05, + "loss": 1.1436378955841064, + "step": 904 + }, + { + "epoch": 1.6178571428571429, + "grad_norm": 0.3926475942134857, + "learning_rate": 1.0592676976333877e-05, + "loss": 1.0450983047485352, + "step": 906 + }, + { + "epoch": 1.6214285714285714, + "grad_norm": 0.3293057978153229, + "learning_rate": 1.0557285033857234e-05, + "loss": 1.0548239946365356, + "step": 908 + }, + { + "epoch": 1.625, + "grad_norm": 2.7279651165008545, + "learning_rate": 1.052189995285749e-05, + "loss": 0.8727450966835022, + "step": 910 + }, + { + "epoch": 1.6285714285714286, + "grad_norm": 0.513361930847168, + "learning_rate": 1.0486522281755034e-05, + "loss": 1.1863266229629517, + "step": 912 + }, + { + "epoch": 1.6321428571428571, + "grad_norm": 0.7765159606933594, + "learning_rate": 1.0451152568855424e-05, + "loss": 1.0367705821990967, + "step": 914 + }, + { + "epoch": 1.6357142857142857, + "grad_norm": 0.6320871114730835, + "learning_rate": 1.0415791362340864e-05, + "loss": 1.191881775856018, + "step": 916 + }, + { + "epoch": 1.6392857142857142, + "grad_norm": 2.4710655212402344, + "learning_rate": 1.0380439210261726e-05, + "loss": 1.0561764240264893, + "step": 918 + }, + { + "epoch": 1.6428571428571428, + "grad_norm": 0.6536027193069458, + "learning_rate": 1.0345096660528047e-05, + "loss": 0.9318988919258118, + "step": 920 + }, + { + "epoch": 1.6464285714285714, + "grad_norm": 1.375860333442688, + "learning_rate": 1.030976426090105e-05, + "loss": 1.101020336151123, + "step": 922 + }, + { + "epoch": 1.65, + "grad_norm": 0.2552533745765686, + "learning_rate": 1.0274442558984634e-05, + "loss": 1.2564234733581543, + "step": 924 + }, + { + "epoch": 1.6535714285714285, + "grad_norm": 0.5600205659866333, + "learning_rate": 1.0239132102216906e-05, + "loss": 1.2616956233978271, + "step": 926 + }, + { + "epoch": 1.657142857142857, + "grad_norm": 0.5227077007293701, + "learning_rate": 1.020383343786168e-05, + "loss": 1.1424009799957275, + "step": 928 + }, + { + "epoch": 1.6607142857142856, + "grad_norm": 1.377939224243164, + "learning_rate": 1.016854711300002e-05, + "loss": 1.0064213275909424, + "step": 930 + }, + { + "epoch": 1.6642857142857141, + "grad_norm": 0.4621258080005646, + "learning_rate": 1.0133273674521726e-05, + "loss": 1.0637047290802002, + "step": 932 + }, + { + "epoch": 1.6678571428571427, + "grad_norm": 0.4151983857154846, + "learning_rate": 1.0098013669116886e-05, + "loss": 1.1586838960647583, + "step": 934 + }, + { + "epoch": 1.6714285714285713, + "grad_norm": 0.45863673090934753, + "learning_rate": 1.006276764326739e-05, + "loss": 1.1697707176208496, + "step": 936 + }, + { + "epoch": 1.675, + "grad_norm": 0.6595650315284729, + "learning_rate": 1.0027536143238474e-05, + "loss": 0.6215699911117554, + "step": 938 + }, + { + "epoch": 1.6785714285714286, + "grad_norm": 0.5221896171569824, + "learning_rate": 9.992319715070231e-06, + "loss": 1.132964849472046, + "step": 940 + }, + { + "epoch": 1.6821428571428572, + "grad_norm": 0.499324768781662, + "learning_rate": 9.957118904569167e-06, + "loss": 1.2663060426712036, + "step": 942 + }, + { + "epoch": 1.6857142857142857, + "grad_norm": 0.766228437423706, + "learning_rate": 9.921934257299731e-06, + "loss": 1.3311316967010498, + "step": 944 + }, + { + "epoch": 1.6892857142857143, + "grad_norm": 0.7923848628997803, + "learning_rate": 9.886766318575871e-06, + "loss": 0.8178808689117432, + "step": 946 + }, + { + "epoch": 1.6928571428571428, + "grad_norm": 1.8682267665863037, + "learning_rate": 9.851615633452577e-06, + "loss": 1.1221369504928589, + "step": 948 + }, + { + "epoch": 1.6964285714285714, + "grad_norm": 0.32414621114730835, + "learning_rate": 9.816482746717415e-06, + "loss": 1.1450505256652832, + "step": 950 + }, + { + "epoch": 1.7, + "grad_norm": 0.48412182927131653, + "learning_rate": 9.781368202882118e-06, + "loss": 1.1050865650177002, + "step": 952 + }, + { + "epoch": 1.7035714285714287, + "grad_norm": 0.7146490812301636, + "learning_rate": 9.746272546174122e-06, + "loss": 0.9220115542411804, + "step": 954 + }, + { + "epoch": 1.7071428571428573, + "grad_norm": 0.4684576392173767, + "learning_rate": 9.711196320528142e-06, + "loss": 1.2169872522354126, + "step": 956 + }, + { + "epoch": 1.7107142857142859, + "grad_norm": 0.9166097640991211, + "learning_rate": 9.67614006957773e-06, + "loss": 1.2747939825057983, + "step": 958 + }, + { + "epoch": 1.7142857142857144, + "grad_norm": 0.3798523247241974, + "learning_rate": 9.641104336646868e-06, + "loss": 1.029056429862976, + "step": 960 + }, + { + "epoch": 1.717857142857143, + "grad_norm": 1.7487893104553223, + "learning_rate": 9.60608966474153e-06, + "loss": 1.1041202545166016, + "step": 962 + }, + { + "epoch": 1.7214285714285715, + "grad_norm": 0.5260602235794067, + "learning_rate": 9.571096596541279e-06, + "loss": 1.1489906311035156, + "step": 964 + }, + { + "epoch": 1.725, + "grad_norm": 1.3120121955871582, + "learning_rate": 9.536125674390834e-06, + "loss": 1.1129286289215088, + "step": 966 + }, + { + "epoch": 1.7285714285714286, + "grad_norm": 0.6453574299812317, + "learning_rate": 9.501177440291697e-06, + "loss": 1.0809402465820312, + "step": 968 + }, + { + "epoch": 1.7321428571428572, + "grad_norm": 0.4150318205356598, + "learning_rate": 9.46625243589374e-06, + "loss": 1.142215609550476, + "step": 970 + }, + { + "epoch": 1.7357142857142858, + "grad_norm": 0.30773821473121643, + "learning_rate": 9.431351202486792e-06, + "loss": 1.0678926706314087, + "step": 972 + }, + { + "epoch": 1.7392857142857143, + "grad_norm": 0.5969177484512329, + "learning_rate": 9.396474280992265e-06, + "loss": 0.61910080909729, + "step": 974 + }, + { + "epoch": 1.7428571428571429, + "grad_norm": 0.3318014442920685, + "learning_rate": 9.36162221195478e-06, + "loss": 1.1399273872375488, + "step": 976 + }, + { + "epoch": 1.7464285714285714, + "grad_norm": 0.5802143216133118, + "learning_rate": 9.32679553553377e-06, + "loss": 1.0623743534088135, + "step": 978 + }, + { + "epoch": 1.75, + "grad_norm": 0.6982766389846802, + "learning_rate": 9.291994791495125e-06, + "loss": 1.130384922027588, + "step": 980 + }, + { + "epoch": 1.7535714285714286, + "grad_norm": 0.9474343061447144, + "learning_rate": 9.257220519202812e-06, + "loss": 1.1633626222610474, + "step": 982 + }, + { + "epoch": 1.7571428571428571, + "grad_norm": 0.823501706123352, + "learning_rate": 9.222473257610519e-06, + "loss": 1.3282148838043213, + "step": 984 + }, + { + "epoch": 1.7607142857142857, + "grad_norm": 0.8721451163291931, + "learning_rate": 9.187753545253318e-06, + "loss": 1.1099696159362793, + "step": 986 + }, + { + "epoch": 1.7642857142857142, + "grad_norm": 1.4551762342453003, + "learning_rate": 9.153061920239291e-06, + "loss": 1.2010962963104248, + "step": 988 + }, + { + "epoch": 1.7678571428571428, + "grad_norm": 0.22830167412757874, + "learning_rate": 9.11839892024122e-06, + "loss": 0.9588916301727295, + "step": 990 + }, + { + "epoch": 1.7714285714285714, + "grad_norm": 0.7184726595878601, + "learning_rate": 9.08376508248822e-06, + "loss": 0.9959466457366943, + "step": 992 + }, + { + "epoch": 1.775, + "grad_norm": 0.8741506934165955, + "learning_rate": 9.049160943757447e-06, + "loss": 1.015051007270813, + "step": 994 + }, + { + "epoch": 1.7785714285714285, + "grad_norm": 0.341207891702652, + "learning_rate": 9.014587040365754e-06, + "loss": 0.6152805685997009, + "step": 996 + }, + { + "epoch": 1.782142857142857, + "grad_norm": 0.5346217155456543, + "learning_rate": 8.98004390816139e-06, + "loss": 1.0692589282989502, + "step": 998 + }, + { + "epoch": 1.7857142857142856, + "grad_norm": 1.2236759662628174, + "learning_rate": 8.945532082515692e-06, + "loss": 1.1441328525543213, + "step": 1000 + }, + { + "epoch": 1.7892857142857141, + "grad_norm": 0.7906798124313354, + "learning_rate": 8.911052098314791e-06, + "loss": 0.9982239007949829, + "step": 1002 + }, + { + "epoch": 1.7928571428571427, + "grad_norm": 0.48270416259765625, + "learning_rate": 8.876604489951317e-06, + "loss": 1.1276839971542358, + "step": 1004 + }, + { + "epoch": 1.7964285714285713, + "grad_norm": 1.3317867517471313, + "learning_rate": 8.842189791316116e-06, + "loss": 1.1620763540267944, + "step": 1006 + }, + { + "epoch": 1.8, + "grad_norm": 0.13755838572978973, + "learning_rate": 8.807808535789982e-06, + "loss": 0.9927979707717896, + "step": 1008 + }, + { + "epoch": 1.8035714285714286, + "grad_norm": 0.6385722160339355, + "learning_rate": 8.773461256235385e-06, + "loss": 1.0914297103881836, + "step": 1010 + }, + { + "epoch": 1.8071428571428572, + "grad_norm": 0.46439942717552185, + "learning_rate": 8.739148484988216e-06, + "loss": 1.1059390306472778, + "step": 1012 + }, + { + "epoch": 1.8107142857142857, + "grad_norm": 0.5348365306854248, + "learning_rate": 8.704870753849533e-06, + "loss": 1.0481466054916382, + "step": 1014 + }, + { + "epoch": 1.8142857142857143, + "grad_norm": 0.503119707107544, + "learning_rate": 8.670628594077313e-06, + "loss": 1.0685755014419556, + "step": 1016 + }, + { + "epoch": 1.8178571428571428, + "grad_norm": 0.45715904235839844, + "learning_rate": 8.636422536378241e-06, + "loss": 1.1131443977355957, + "step": 1018 + }, + { + "epoch": 1.8214285714285714, + "grad_norm": 0.4219473600387573, + "learning_rate": 8.602253110899461e-06, + "loss": 1.1202515363693237, + "step": 1020 + }, + { + "epoch": 1.825, + "grad_norm": 1.506514549255371, + "learning_rate": 8.56812084722036e-06, + "loss": 1.0852363109588623, + "step": 1022 + }, + { + "epoch": 1.8285714285714287, + "grad_norm": 1.1264326572418213, + "learning_rate": 8.534026274344378e-06, + "loss": 1.1459194421768188, + "step": 1024 + }, + { + "epoch": 1.8321428571428573, + "grad_norm": 0.19118832051753998, + "learning_rate": 8.499969920690799e-06, + "loss": 1.0966734886169434, + "step": 1026 + }, + { + "epoch": 1.8357142857142859, + "grad_norm": 5.950002670288086, + "learning_rate": 8.465952314086554e-06, + "loss": 0.9605913162231445, + "step": 1028 + }, + { + "epoch": 1.8392857142857144, + "grad_norm": 0.9554546475410461, + "learning_rate": 8.431973981758055e-06, + "loss": 0.8658078908920288, + "step": 1030 + }, + { + "epoch": 1.842857142857143, + "grad_norm": 0.35784661769866943, + "learning_rate": 8.39803545032301e-06, + "loss": 1.21623957157135, + "step": 1032 + }, + { + "epoch": 1.8464285714285715, + "grad_norm": 0.3995438516139984, + "learning_rate": 8.364137245782278e-06, + "loss": 1.0300588607788086, + "step": 1034 + }, + { + "epoch": 1.85, + "grad_norm": 0.4243587255477905, + "learning_rate": 8.330279893511695e-06, + "loss": 1.1283214092254639, + "step": 1036 + }, + { + "epoch": 1.8535714285714286, + "grad_norm": 1.56551992893219, + "learning_rate": 8.296463918253945e-06, + "loss": 1.0735278129577637, + "step": 1038 + }, + { + "epoch": 1.8571428571428572, + "grad_norm": 0.5571054220199585, + "learning_rate": 8.262689844110426e-06, + "loss": 1.1387474536895752, + "step": 1040 + }, + { + "epoch": 1.8607142857142858, + "grad_norm": 0.8355203866958618, + "learning_rate": 8.228958194533134e-06, + "loss": 1.1476329565048218, + "step": 1042 + }, + { + "epoch": 1.8642857142857143, + "grad_norm": 1.1273201704025269, + "learning_rate": 8.195269492316531e-06, + "loss": 1.1415425539016724, + "step": 1044 + }, + { + "epoch": 1.8678571428571429, + "grad_norm": 0.48498058319091797, + "learning_rate": 8.161624259589456e-06, + "loss": 1.2414381504058838, + "step": 1046 + }, + { + "epoch": 1.8714285714285714, + "grad_norm": 2.3717753887176514, + "learning_rate": 8.128023017807032e-06, + "loss": 1.0288283824920654, + "step": 1048 + }, + { + "epoch": 1.875, + "grad_norm": 0.4442752003669739, + "learning_rate": 8.094466287742583e-06, + "loss": 1.0729954242706299, + "step": 1050 + }, + { + "epoch": 1.8785714285714286, + "grad_norm": 1.309718370437622, + "learning_rate": 8.060954589479566e-06, + "loss": 1.0899989604949951, + "step": 1052 + }, + { + "epoch": 1.8821428571428571, + "grad_norm": 0.6105849146842957, + "learning_rate": 8.027488442403499e-06, + "loss": 1.247680425643921, + "step": 1054 + }, + { + "epoch": 1.8857142857142857, + "grad_norm": 0.5027915835380554, + "learning_rate": 7.994068365193916e-06, + "loss": 1.1438732147216797, + "step": 1056 + }, + { + "epoch": 1.8892857142857142, + "grad_norm": 0.8934045433998108, + "learning_rate": 7.96069487581635e-06, + "loss": 1.1180412769317627, + "step": 1058 + }, + { + "epoch": 1.8928571428571428, + "grad_norm": 1.2043625116348267, + "learning_rate": 7.927368491514263e-06, + "loss": 1.289278507232666, + "step": 1060 + }, + { + "epoch": 1.8964285714285714, + "grad_norm": 0.3672593832015991, + "learning_rate": 7.894089728801069e-06, + "loss": 1.0988218784332275, + "step": 1062 + }, + { + "epoch": 1.9, + "grad_norm": 0.4244534373283386, + "learning_rate": 7.860859103452099e-06, + "loss": 0.6889010667800903, + "step": 1064 + }, + { + "epoch": 1.9035714285714285, + "grad_norm": 4.469996452331543, + "learning_rate": 7.827677130496635e-06, + "loss": 0.995285153388977, + "step": 1066 + }, + { + "epoch": 1.907142857142857, + "grad_norm": 0.39180439710617065, + "learning_rate": 7.794544324209909e-06, + "loss": 1.084833025932312, + "step": 1068 + }, + { + "epoch": 1.9107142857142856, + "grad_norm": 2.244345188140869, + "learning_rate": 7.76146119810513e-06, + "loss": 1.1369450092315674, + "step": 1070 + }, + { + "epoch": 1.9142857142857141, + "grad_norm": 0.685583233833313, + "learning_rate": 7.728428264925531e-06, + "loss": 0.9934486150741577, + "step": 1072 + }, + { + "epoch": 1.9178571428571427, + "grad_norm": 0.3774887025356293, + "learning_rate": 7.69544603663644e-06, + "loss": 0.9882639646530151, + "step": 1074 + }, + { + "epoch": 1.9214285714285713, + "grad_norm": 0.53889000415802, + "learning_rate": 7.662515024417315e-06, + "loss": 1.1985116004943848, + "step": 1076 + }, + { + "epoch": 1.925, + "grad_norm": 1.1523969173431396, + "learning_rate": 7.62963573865383e-06, + "loss": 0.9715709686279297, + "step": 1078 + }, + { + "epoch": 1.9285714285714286, + "grad_norm": 0.540876030921936, + "learning_rate": 7.5968086889299795e-06, + "loss": 1.1391596794128418, + "step": 1080 + }, + { + "epoch": 1.9321428571428572, + "grad_norm": 0.4845031797885895, + "learning_rate": 7.564034384020174e-06, + "loss": 1.0560252666473389, + "step": 1082 + }, + { + "epoch": 1.9357142857142857, + "grad_norm": 1.0255088806152344, + "learning_rate": 7.53131333188134e-06, + "loss": 1.1791609525680542, + "step": 1084 + }, + { + "epoch": 1.9392857142857143, + "grad_norm": 0.47408270835876465, + "learning_rate": 7.498646039645068e-06, + "loss": 1.0764998197555542, + "step": 1086 + }, + { + "epoch": 1.9428571428571428, + "grad_norm": 0.5158260464668274, + "learning_rate": 7.4660330136097345e-06, + "loss": 1.1066300868988037, + "step": 1088 + }, + { + "epoch": 1.9464285714285714, + "grad_norm": 0.636821985244751, + "learning_rate": 7.433474759232675e-06, + "loss": 0.7171211242675781, + "step": 1090 + }, + { + "epoch": 1.95, + "grad_norm": 0.4269849956035614, + "learning_rate": 7.400971781122334e-06, + "loss": 1.1467523574829102, + "step": 1092 + }, + { + "epoch": 1.9535714285714287, + "grad_norm": 0.4313041865825653, + "learning_rate": 7.3685245830304455e-06, + "loss": 1.088484287261963, + "step": 1094 + }, + { + "epoch": 1.9571428571428573, + "grad_norm": 0.607107937335968, + "learning_rate": 7.336133667844232e-06, + "loss": 1.1424229145050049, + "step": 1096 + }, + { + "epoch": 1.9607142857142859, + "grad_norm": 0.5271368622779846, + "learning_rate": 7.3037995375786105e-06, + "loss": 1.1263647079467773, + "step": 1098 + }, + { + "epoch": 1.9642857142857144, + "grad_norm": 0.5421050190925598, + "learning_rate": 7.271522693368412e-06, + "loss": 1.1368845701217651, + "step": 1100 + }, + { + "epoch": 1.967857142857143, + "grad_norm": 0.45558372139930725, + "learning_rate": 7.239303635460604e-06, + "loss": 1.0905189514160156, + "step": 1102 + }, + { + "epoch": 1.9714285714285715, + "grad_norm": 0.6813379526138306, + "learning_rate": 7.207142863206544e-06, + "loss": 1.300194501876831, + "step": 1104 + }, + { + "epoch": 1.975, + "grad_norm": 3.9907617568969727, + "learning_rate": 7.175040875054256e-06, + "loss": 1.2036380767822266, + "step": 1106 + }, + { + "epoch": 1.9785714285714286, + "grad_norm": 0.3055593967437744, + "learning_rate": 7.142998168540676e-06, + "loss": 1.092109203338623, + "step": 1108 + }, + { + "epoch": 1.9821428571428572, + "grad_norm": 0.5111721158027649, + "learning_rate": 7.111015240283963e-06, + "loss": 0.5738495588302612, + "step": 1110 + }, + { + "epoch": 1.9857142857142858, + "grad_norm": 0.3026096522808075, + "learning_rate": 7.079092585975789e-06, + "loss": 0.5167784094810486, + "step": 1112 + }, + { + "epoch": 1.9892857142857143, + "grad_norm": 0.37242579460144043, + "learning_rate": 7.047230700373669e-06, + "loss": 1.086807131767273, + "step": 1114 + }, + { + "epoch": 1.9928571428571429, + "grad_norm": 0.5165676474571228, + "learning_rate": 7.015430077293281e-06, + "loss": 1.1821894645690918, + "step": 1116 + }, + { + "epoch": 1.9964285714285714, + "grad_norm": 2.6484367847442627, + "learning_rate": 6.983691209600821e-06, + "loss": 0.7227582931518555, + "step": 1118 + }, + { + "epoch": 2.0, + "grad_norm": 0.8464747071266174, + "learning_rate": 6.952014589205357e-06, + "loss": 1.2078351974487305, + "step": 1120 + }, + { + "epoch": 2.0035714285714286, + "grad_norm": 0.5015007853507996, + "learning_rate": 6.92040070705122e-06, + "loss": 1.0274121761322021, + "step": 1122 + }, + { + "epoch": 2.007142857142857, + "grad_norm": 0.7175948023796082, + "learning_rate": 6.888850053110364e-06, + "loss": 1.2103437185287476, + "step": 1124 + }, + { + "epoch": 2.0107142857142857, + "grad_norm": 0.7134751081466675, + "learning_rate": 6.857363116374809e-06, + "loss": 1.1368117332458496, + "step": 1126 + }, + { + "epoch": 2.0142857142857142, + "grad_norm": 0.363921582698822, + "learning_rate": 6.825940384849035e-06, + "loss": 1.0515990257263184, + "step": 1128 + }, + { + "epoch": 2.017857142857143, + "grad_norm": 1.2383047342300415, + "learning_rate": 6.794582345542442e-06, + "loss": 0.8165490627288818, + "step": 1130 + }, + { + "epoch": 2.0214285714285714, + "grad_norm": 0.3281400799751282, + "learning_rate": 6.763289484461777e-06, + "loss": 0.9551868438720703, + "step": 1132 + }, + { + "epoch": 2.025, + "grad_norm": 2.324735641479492, + "learning_rate": 6.732062286603622e-06, + "loss": 0.8535771369934082, + "step": 1134 + }, + { + "epoch": 2.0285714285714285, + "grad_norm": 0.450276643037796, + "learning_rate": 6.700901235946851e-06, + "loss": 1.1220966577529907, + "step": 1136 + }, + { + "epoch": 2.032142857142857, + "grad_norm": 0.4659401476383209, + "learning_rate": 6.669806815445174e-06, + "loss": 1.1011137962341309, + "step": 1138 + }, + { + "epoch": 2.0357142857142856, + "grad_norm": 0.7880271077156067, + "learning_rate": 6.638779507019606e-06, + "loss": 1.2896149158477783, + "step": 1140 + }, + { + "epoch": 2.039285714285714, + "grad_norm": 0.27034834027290344, + "learning_rate": 6.60781979155102e-06, + "loss": 1.01216459274292, + "step": 1142 + }, + { + "epoch": 2.0428571428571427, + "grad_norm": 1.3774323463439941, + "learning_rate": 6.576928148872692e-06, + "loss": 1.0622344017028809, + "step": 1144 + }, + { + "epoch": 2.0464285714285713, + "grad_norm": 0.41495487093925476, + "learning_rate": 6.546105057762866e-06, + "loss": 1.0661730766296387, + "step": 1146 + }, + { + "epoch": 2.05, + "grad_norm": 0.49956122040748596, + "learning_rate": 6.515350995937322e-06, + "loss": 1.1045374870300293, + "step": 1148 + }, + { + "epoch": 2.0535714285714284, + "grad_norm": 1.3499337434768677, + "learning_rate": 6.484666440041989e-06, + "loss": 1.1958565711975098, + "step": 1150 + }, + { + "epoch": 2.057142857142857, + "grad_norm": 0.8459187746047974, + "learning_rate": 6.454051865645541e-06, + "loss": 1.0263601541519165, + "step": 1152 + }, + { + "epoch": 2.0607142857142855, + "grad_norm": 0.8016642928123474, + "learning_rate": 6.423507747232047e-06, + "loss": 1.0019943714141846, + "step": 1154 + }, + { + "epoch": 2.064285714285714, + "grad_norm": 0.3757397532463074, + "learning_rate": 6.3930345581935934e-06, + "loss": 1.091320514678955, + "step": 1156 + }, + { + "epoch": 2.067857142857143, + "grad_norm": 0.6635928153991699, + "learning_rate": 6.3626327708229585e-06, + "loss": 1.0494961738586426, + "step": 1158 + }, + { + "epoch": 2.0714285714285716, + "grad_norm": 0.3476162254810333, + "learning_rate": 6.332302856306293e-06, + "loss": 1.1452181339263916, + "step": 1160 + }, + { + "epoch": 2.075, + "grad_norm": 0.5809051990509033, + "learning_rate": 6.302045284715825e-06, + "loss": 1.1838127374649048, + "step": 1162 + }, + { + "epoch": 2.0785714285714287, + "grad_norm": 0.7286864519119263, + "learning_rate": 6.271860525002558e-06, + "loss": 0.9906060695648193, + "step": 1164 + }, + { + "epoch": 2.0821428571428573, + "grad_norm": 0.7779916524887085, + "learning_rate": 6.241749044989012e-06, + "loss": 1.0825626850128174, + "step": 1166 + }, + { + "epoch": 2.085714285714286, + "grad_norm": 0.624515175819397, + "learning_rate": 6.211711311361972e-06, + "loss": 1.0792040824890137, + "step": 1168 + }, + { + "epoch": 2.0892857142857144, + "grad_norm": 0.5379583239555359, + "learning_rate": 6.181747789665256e-06, + "loss": 1.233987808227539, + "step": 1170 + }, + { + "epoch": 2.092857142857143, + "grad_norm": 0.5427027940750122, + "learning_rate": 6.1518589442925e-06, + "loss": 1.0380905866622925, + "step": 1172 + }, + { + "epoch": 2.0964285714285715, + "grad_norm": 0.8613371253013611, + "learning_rate": 6.122045238479953e-06, + "loss": 1.1964986324310303, + "step": 1174 + }, + { + "epoch": 2.1, + "grad_norm": 0.749973714351654, + "learning_rate": 6.0923071342993075e-06, + "loss": 1.0324103832244873, + "step": 1176 + }, + { + "epoch": 2.1035714285714286, + "grad_norm": 0.41390734910964966, + "learning_rate": 6.062645092650543e-06, + "loss": 0.7683989405632019, + "step": 1178 + }, + { + "epoch": 2.107142857142857, + "grad_norm": 0.4426082968711853, + "learning_rate": 6.033059573254753e-06, + "loss": 0.9940497875213623, + "step": 1180 + }, + { + "epoch": 2.1107142857142858, + "grad_norm": 1.0037376880645752, + "learning_rate": 6.003551034647059e-06, + "loss": 0.975098192691803, + "step": 1182 + }, + { + "epoch": 2.1142857142857143, + "grad_norm": 1.797518014907837, + "learning_rate": 5.974119934169473e-06, + "loss": 1.0679240226745605, + "step": 1184 + }, + { + "epoch": 2.117857142857143, + "grad_norm": 0.9084307551383972, + "learning_rate": 5.944766727963834e-06, + "loss": 0.9722939729690552, + "step": 1186 + }, + { + "epoch": 2.1214285714285714, + "grad_norm": 0.4284418523311615, + "learning_rate": 5.9154918709647204e-06, + "loss": 0.9965044260025024, + "step": 1188 + }, + { + "epoch": 2.125, + "grad_norm": 0.9147275686264038, + "learning_rate": 5.8862958168924025e-06, + "loss": 0.6787570714950562, + "step": 1190 + }, + { + "epoch": 2.1285714285714286, + "grad_norm": 0.6834171414375305, + "learning_rate": 5.8571790182458085e-06, + "loss": 1.0144072771072388, + "step": 1192 + }, + { + "epoch": 2.132142857142857, + "grad_norm": 0.5649306178092957, + "learning_rate": 5.828141926295533e-06, + "loss": 1.1785552501678467, + "step": 1194 + }, + { + "epoch": 2.1357142857142857, + "grad_norm": 0.9754287004470825, + "learning_rate": 5.7991849910768096e-06, + "loss": 1.1542279720306396, + "step": 1196 + }, + { + "epoch": 2.1392857142857142, + "grad_norm": 0.540823757648468, + "learning_rate": 5.770308661382556e-06, + "loss": 0.8885504007339478, + "step": 1198 + }, + { + "epoch": 2.142857142857143, + "grad_norm": 0.378530889749527, + "learning_rate": 5.741513384756421e-06, + "loss": 1.0100677013397217, + "step": 1200 + }, + { + "epoch": 2.1464285714285714, + "grad_norm": 1.1241729259490967, + "learning_rate": 5.712799607485832e-06, + "loss": 0.654653787612915, + "step": 1202 + }, + { + "epoch": 2.15, + "grad_norm": 1.2759966850280762, + "learning_rate": 5.6841677745950965e-06, + "loss": 1.1008085012435913, + "step": 1204 + }, + { + "epoch": 2.1535714285714285, + "grad_norm": 0.34266769886016846, + "learning_rate": 5.6556183298384885e-06, + "loss": 1.011092185974121, + "step": 1206 + }, + { + "epoch": 2.157142857142857, + "grad_norm": 0.4193215072154999, + "learning_rate": 5.62715171569338e-06, + "loss": 1.0412278175354004, + "step": 1208 + }, + { + "epoch": 2.1607142857142856, + "grad_norm": 0.32657843828201294, + "learning_rate": 5.598768373353392e-06, + "loss": 0.5627393126487732, + "step": 1210 + }, + { + "epoch": 2.164285714285714, + "grad_norm": 0.5068192481994629, + "learning_rate": 5.570468742721532e-06, + "loss": 1.0111093521118164, + "step": 1212 + }, + { + "epoch": 2.1678571428571427, + "grad_norm": 0.8486136794090271, + "learning_rate": 5.542253262403397e-06, + "loss": 1.0036768913269043, + "step": 1214 + }, + { + "epoch": 2.1714285714285713, + "grad_norm": 0.4339812695980072, + "learning_rate": 5.514122369700366e-06, + "loss": 0.858208179473877, + "step": 1216 + }, + { + "epoch": 2.175, + "grad_norm": 0.3655393421649933, + "learning_rate": 5.486076500602836e-06, + "loss": 0.9979751110076904, + "step": 1218 + }, + { + "epoch": 2.1785714285714284, + "grad_norm": 0.7398764491081238, + "learning_rate": 5.458116089783441e-06, + "loss": 1.089611530303955, + "step": 1220 + }, + { + "epoch": 2.182142857142857, + "grad_norm": 2.556060314178467, + "learning_rate": 5.430241570590335e-06, + "loss": 0.9004594087600708, + "step": 1222 + }, + { + "epoch": 2.185714285714286, + "grad_norm": 0.24670109152793884, + "learning_rate": 5.40245337504047e-06, + "loss": 0.8174819946289062, + "step": 1224 + }, + { + "epoch": 2.189285714285714, + "grad_norm": 0.414289265871048, + "learning_rate": 5.374751933812895e-06, + "loss": 1.019258737564087, + "step": 1226 + }, + { + "epoch": 2.192857142857143, + "grad_norm": 0.4929758310317993, + "learning_rate": 5.34713767624209e-06, + "loss": 0.9936923384666443, + "step": 1228 + }, + { + "epoch": 2.1964285714285716, + "grad_norm": 0.42652517557144165, + "learning_rate": 5.319611030311306e-06, + "loss": 1.1034317016601562, + "step": 1230 + }, + { + "epoch": 2.2, + "grad_norm": 1.5116170644760132, + "learning_rate": 5.292172422645931e-06, + "loss": 1.107032299041748, + "step": 1232 + }, + { + "epoch": 2.2035714285714287, + "grad_norm": 1.4664469957351685, + "learning_rate": 5.2648222785068845e-06, + "loss": 1.0061404705047607, + "step": 1234 + }, + { + "epoch": 2.2071428571428573, + "grad_norm": 0.6140238642692566, + "learning_rate": 5.237561021784021e-06, + "loss": 1.1637604236602783, + "step": 1236 + }, + { + "epoch": 2.210714285714286, + "grad_norm": 0.9620886445045471, + "learning_rate": 5.210389074989558e-06, + "loss": 1.1427053213119507, + "step": 1238 + }, + { + "epoch": 2.2142857142857144, + "grad_norm": 0.43619751930236816, + "learning_rate": 5.183306859251531e-06, + "loss": 1.1551034450531006, + "step": 1240 + }, + { + "epoch": 2.217857142857143, + "grad_norm": 1.3034915924072266, + "learning_rate": 5.1563147943072775e-06, + "loss": 0.9217376708984375, + "step": 1242 + }, + { + "epoch": 2.2214285714285715, + "grad_norm": 0.7653930187225342, + "learning_rate": 5.129413298496913e-06, + "loss": 1.0179123878479004, + "step": 1244 + }, + { + "epoch": 2.225, + "grad_norm": 0.39587950706481934, + "learning_rate": 5.102602788756847e-06, + "loss": 0.9925167560577393, + "step": 1246 + }, + { + "epoch": 2.2285714285714286, + "grad_norm": 0.6801050305366516, + "learning_rate": 5.075883680613338e-06, + "loss": 0.9864404797554016, + "step": 1248 + }, + { + "epoch": 2.232142857142857, + "grad_norm": 0.5718074440956116, + "learning_rate": 5.049256388176054e-06, + "loss": 1.0245519876480103, + "step": 1250 + }, + { + "epoch": 2.2357142857142858, + "grad_norm": 0.4339280426502228, + "learning_rate": 5.022721324131626e-06, + "loss": 1.0415318012237549, + "step": 1252 + }, + { + "epoch": 2.2392857142857143, + "grad_norm": 1.2172452211380005, + "learning_rate": 4.996278899737283e-06, + "loss": 1.045757532119751, + "step": 1254 + }, + { + "epoch": 2.242857142857143, + "grad_norm": 1.0536972284317017, + "learning_rate": 4.969929524814464e-06, + "loss": 0.9379353523254395, + "step": 1256 + }, + { + "epoch": 2.2464285714285714, + "grad_norm": 0.9581257104873657, + "learning_rate": 4.943673607742466e-06, + "loss": 1.0260355472564697, + "step": 1258 + }, + { + "epoch": 2.25, + "grad_norm": 1.8986560106277466, + "learning_rate": 4.91751155545212e-06, + "loss": 0.6572850942611694, + "step": 1260 + }, + { + "epoch": 2.2535714285714286, + "grad_norm": 0.6189560890197754, + "learning_rate": 4.891443773419479e-06, + "loss": 0.5532552599906921, + "step": 1262 + }, + { + "epoch": 2.257142857142857, + "grad_norm": 0.4603727459907532, + "learning_rate": 4.865470665659535e-06, + "loss": 1.104417324066162, + "step": 1264 + }, + { + "epoch": 2.2607142857142857, + "grad_norm": 0.7528354525566101, + "learning_rate": 4.839592634719966e-06, + "loss": 0.906780481338501, + "step": 1266 + }, + { + "epoch": 2.2642857142857142, + "grad_norm": 0.8098517060279846, + "learning_rate": 4.813810081674875e-06, + "loss": 1.0541425943374634, + "step": 1268 + }, + { + "epoch": 2.267857142857143, + "grad_norm": 1.1421626806259155, + "learning_rate": 4.7881234061186e-06, + "loss": 0.977279782295227, + "step": 1270 + }, + { + "epoch": 2.2714285714285714, + "grad_norm": 0.5860005617141724, + "learning_rate": 4.7625330061595025e-06, + "loss": 1.0465322732925415, + "step": 1272 + }, + { + "epoch": 2.275, + "grad_norm": 0.753535270690918, + "learning_rate": 4.73703927841381e-06, + "loss": 1.0138700008392334, + "step": 1274 + }, + { + "epoch": 2.2785714285714285, + "grad_norm": 0.5131919384002686, + "learning_rate": 4.711642617999461e-06, + "loss": 1.0608528852462769, + "step": 1276 + }, + { + "epoch": 2.282142857142857, + "grad_norm": 2.5389108657836914, + "learning_rate": 4.6863434185299784e-06, + "loss": 0.9582719802856445, + "step": 1278 + }, + { + "epoch": 2.2857142857142856, + "grad_norm": 0.9607290029525757, + "learning_rate": 4.661142072108371e-06, + "loss": 1.059136152267456, + "step": 1280 + }, + { + "epoch": 2.289285714285714, + "grad_norm": 0.3775724470615387, + "learning_rate": 4.636038969321073e-06, + "loss": 1.0284271240234375, + "step": 1282 + }, + { + "epoch": 2.2928571428571427, + "grad_norm": 0.4234330356121063, + "learning_rate": 4.611034499231865e-06, + "loss": 1.0011167526245117, + "step": 1284 + }, + { + "epoch": 2.2964285714285713, + "grad_norm": 0.36131852865219116, + "learning_rate": 4.586129049375857e-06, + "loss": 1.0314903259277344, + "step": 1286 + }, + { + "epoch": 2.3, + "grad_norm": 0.7630230188369751, + "learning_rate": 4.561323005753489e-06, + "loss": 0.6767295598983765, + "step": 1288 + }, + { + "epoch": 2.3035714285714284, + "grad_norm": 1.119634985923767, + "learning_rate": 4.536616752824525e-06, + "loss": 1.0448781251907349, + "step": 1290 + }, + { + "epoch": 2.307142857142857, + "grad_norm": 1.1402679681777954, + "learning_rate": 4.512010673502125e-06, + "loss": 1.046932578086853, + "step": 1292 + }, + { + "epoch": 2.310714285714286, + "grad_norm": 0.5093026161193848, + "learning_rate": 4.48750514914688e-06, + "loss": 0.9864715337753296, + "step": 1294 + }, + { + "epoch": 2.314285714285714, + "grad_norm": 0.5392947196960449, + "learning_rate": 4.463100559560935e-06, + "loss": 0.959062933921814, + "step": 1296 + }, + { + "epoch": 2.317857142857143, + "grad_norm": 1.170287847518921, + "learning_rate": 4.438797282982069e-06, + "loss": 0.8960469961166382, + "step": 1298 + }, + { + "epoch": 2.3214285714285716, + "grad_norm": 1.2749667167663574, + "learning_rate": 4.414595696077857e-06, + "loss": 0.829791247844696, + "step": 1300 + }, + { + "epoch": 2.325, + "grad_norm": 1.079351782798767, + "learning_rate": 4.390496173939808e-06, + "loss": 0.9560054540634155, + "step": 1302 + }, + { + "epoch": 2.3285714285714287, + "grad_norm": 1.3681961297988892, + "learning_rate": 4.366499090077587e-06, + "loss": 1.0540683269500732, + "step": 1304 + }, + { + "epoch": 2.3321428571428573, + "grad_norm": 0.41432955861091614, + "learning_rate": 4.342604816413193e-06, + "loss": 1.0061728954315186, + "step": 1306 + }, + { + "epoch": 2.335714285714286, + "grad_norm": 1.46927809715271, + "learning_rate": 4.318813723275211e-06, + "loss": 1.0353810787200928, + "step": 1308 + }, + { + "epoch": 2.3392857142857144, + "grad_norm": 0.609827995300293, + "learning_rate": 4.295126179393067e-06, + "loss": 0.6233819127082825, + "step": 1310 + }, + { + "epoch": 2.342857142857143, + "grad_norm": 0.36403512954711914, + "learning_rate": 4.271542551891319e-06, + "loss": 0.5161199569702148, + "step": 1312 + }, + { + "epoch": 2.3464285714285715, + "grad_norm": 0.453372985124588, + "learning_rate": 4.248063206283959e-06, + "loss": 1.1536060571670532, + "step": 1314 + }, + { + "epoch": 2.35, + "grad_norm": 0.2660079896450043, + "learning_rate": 4.224688506468754e-06, + "loss": 1.0320401191711426, + "step": 1316 + }, + { + "epoch": 2.3535714285714286, + "grad_norm": 0.5758374333381653, + "learning_rate": 4.201418814721599e-06, + "loss": 1.025177240371704, + "step": 1318 + }, + { + "epoch": 2.357142857142857, + "grad_norm": 0.6305975317955017, + "learning_rate": 4.178254491690923e-06, + "loss": 0.7249777317047119, + "step": 1320 + }, + { + "epoch": 2.3607142857142858, + "grad_norm": 0.4613308310508728, + "learning_rate": 4.155195896392069e-06, + "loss": 0.9757188558578491, + "step": 1322 + }, + { + "epoch": 2.3642857142857143, + "grad_norm": 0.8076472878456116, + "learning_rate": 4.132243386201743e-06, + "loss": 0.7056589126586914, + "step": 1324 + }, + { + "epoch": 2.367857142857143, + "grad_norm": 0.3006990849971771, + "learning_rate": 4.109397316852488e-06, + "loss": 1.2015197277069092, + "step": 1326 + }, + { + "epoch": 2.3714285714285714, + "grad_norm": 0.4321019649505615, + "learning_rate": 4.086658042427154e-06, + "loss": 1.0216331481933594, + "step": 1328 + }, + { + "epoch": 2.375, + "grad_norm": 0.21614263951778412, + "learning_rate": 4.0640259153534165e-06, + "loss": 0.7432563900947571, + "step": 1330 + }, + { + "epoch": 2.3785714285714286, + "grad_norm": 0.5420116782188416, + "learning_rate": 4.041501286398311e-06, + "loss": 1.0380558967590332, + "step": 1332 + }, + { + "epoch": 2.382142857142857, + "grad_norm": 0.6743971705436707, + "learning_rate": 4.019084504662803e-06, + "loss": 0.9956204295158386, + "step": 1334 + }, + { + "epoch": 2.3857142857142857, + "grad_norm": 0.8979402780532837, + "learning_rate": 3.99677591757637e-06, + "loss": 0.9765022993087769, + "step": 1336 + }, + { + "epoch": 2.3892857142857142, + "grad_norm": 0.6747384667396545, + "learning_rate": 3.974575870891622e-06, + "loss": 1.1052476167678833, + "step": 1338 + }, + { + "epoch": 2.392857142857143, + "grad_norm": 0.27344566583633423, + "learning_rate": 3.952484708678942e-06, + "loss": 1.0951298475265503, + "step": 1340 + }, + { + "epoch": 2.3964285714285714, + "grad_norm": 0.4220183193683624, + "learning_rate": 3.93050277332115e-06, + "loss": 1.0463672876358032, + "step": 1342 + }, + { + "epoch": 2.4, + "grad_norm": 0.434129923582077, + "learning_rate": 3.9086304055082005e-06, + "loss": 0.9638611078262329, + "step": 1344 + }, + { + "epoch": 2.4035714285714285, + "grad_norm": 0.30925750732421875, + "learning_rate": 3.886867944231901e-06, + "loss": 1.068469762802124, + "step": 1346 + }, + { + "epoch": 2.407142857142857, + "grad_norm": 0.7762970328330994, + "learning_rate": 3.865215726780658e-06, + "loss": 1.0488204956054688, + "step": 1348 + }, + { + "epoch": 2.4107142857142856, + "grad_norm": 0.9941971898078918, + "learning_rate": 3.8436740887342464e-06, + "loss": 0.9059903621673584, + "step": 1350 + }, + { + "epoch": 2.414285714285714, + "grad_norm": 0.7018589377403259, + "learning_rate": 3.82224336395862e-06, + "loss": 0.9785033464431763, + "step": 1352 + }, + { + "epoch": 2.4178571428571427, + "grad_norm": 0.4142831861972809, + "learning_rate": 3.800923884600718e-06, + "loss": 0.9408230781555176, + "step": 1354 + }, + { + "epoch": 2.4214285714285713, + "grad_norm": 0.3824380040168762, + "learning_rate": 3.7797159810833356e-06, + "loss": 0.979665994644165, + "step": 1356 + }, + { + "epoch": 2.425, + "grad_norm": 0.4297451972961426, + "learning_rate": 3.758619982099985e-06, + "loss": 0.9747478365898132, + "step": 1358 + }, + { + "epoch": 2.4285714285714284, + "grad_norm": 0.6103563904762268, + "learning_rate": 3.737636214609825e-06, + "loss": 1.060492992401123, + "step": 1360 + }, + { + "epoch": 2.432142857142857, + "grad_norm": 0.26877671480178833, + "learning_rate": 3.7167650038325685e-06, + "loss": 0.9182797074317932, + "step": 1362 + }, + { + "epoch": 2.435714285714286, + "grad_norm": 0.7906038761138916, + "learning_rate": 3.696006673243458e-06, + "loss": 1.034062147140503, + "step": 1364 + }, + { + "epoch": 2.439285714285714, + "grad_norm": 0.4611876308917999, + "learning_rate": 3.6753615445682463e-06, + "loss": 1.0039832592010498, + "step": 1366 + }, + { + "epoch": 2.442857142857143, + "grad_norm": 0.9368420243263245, + "learning_rate": 3.6548299377782113e-06, + "loss": 1.0611416101455688, + "step": 1368 + }, + { + "epoch": 2.4464285714285716, + "grad_norm": 0.7252188324928284, + "learning_rate": 3.634412171085197e-06, + "loss": 0.5861155986785889, + "step": 1370 + }, + { + "epoch": 2.45, + "grad_norm": 1.0302289724349976, + "learning_rate": 3.614108560936681e-06, + "loss": 1.044318675994873, + "step": 1372 + }, + { + "epoch": 2.4535714285714287, + "grad_norm": 0.5836187601089478, + "learning_rate": 3.5939194220108687e-06, + "loss": 1.0356674194335938, + "step": 1374 + }, + { + "epoch": 2.4571428571428573, + "grad_norm": 0.49724525213241577, + "learning_rate": 3.5738450672118265e-06, + "loss": 0.9499672651290894, + "step": 1376 + }, + { + "epoch": 2.460714285714286, + "grad_norm": 0.6018553376197815, + "learning_rate": 3.5538858076646115e-06, + "loss": 0.501020073890686, + "step": 1378 + }, + { + "epoch": 2.4642857142857144, + "grad_norm": 0.47356775403022766, + "learning_rate": 3.5340419527104685e-06, + "loss": 1.16387140750885, + "step": 1380 + }, + { + "epoch": 2.467857142857143, + "grad_norm": 0.3433539867401123, + "learning_rate": 3.514313809902028e-06, + "loss": 1.0770752429962158, + "step": 1382 + }, + { + "epoch": 2.4714285714285715, + "grad_norm": 0.5331788063049316, + "learning_rate": 3.494701684998542e-06, + "loss": 0.9788726568222046, + "step": 1384 + }, + { + "epoch": 2.475, + "grad_norm": 0.6456915140151978, + "learning_rate": 3.4752058819611417e-06, + "loss": 1.0846669673919678, + "step": 1386 + }, + { + "epoch": 2.4785714285714286, + "grad_norm": 0.5334643721580505, + "learning_rate": 3.455826702948129e-06, + "loss": 0.6731972694396973, + "step": 1388 + }, + { + "epoch": 2.482142857142857, + "grad_norm": 0.543170154094696, + "learning_rate": 3.436564448310287e-06, + "loss": 1.0429730415344238, + "step": 1390 + }, + { + "epoch": 2.4857142857142858, + "grad_norm": 0.9272890686988831, + "learning_rate": 3.417419416586242e-06, + "loss": 1.0558149814605713, + "step": 1392 + }, + { + "epoch": 2.4892857142857143, + "grad_norm": 0.6048328876495361, + "learning_rate": 3.3983919044978163e-06, + "loss": 1.000851035118103, + "step": 1394 + }, + { + "epoch": 2.492857142857143, + "grad_norm": 0.4908715486526489, + "learning_rate": 3.3794822069454434e-06, + "loss": 1.0060486793518066, + "step": 1396 + }, + { + "epoch": 2.4964285714285714, + "grad_norm": 1.0622693300247192, + "learning_rate": 3.36069061700359e-06, + "loss": 0.6252748966217041, + "step": 1398 + }, + { + "epoch": 2.5, + "grad_norm": 0.41150182485580444, + "learning_rate": 3.3420174259162173e-06, + "loss": 0.9904586672782898, + "step": 1400 + }, + { + "epoch": 2.5035714285714286, + "grad_norm": 0.5609564185142517, + "learning_rate": 3.323462923092265e-06, + "loss": 1.0023245811462402, + "step": 1402 + }, + { + "epoch": 2.507142857142857, + "grad_norm": 0.529987633228302, + "learning_rate": 3.305027396101167e-06, + "loss": 0.9853721857070923, + "step": 1404 + }, + { + "epoch": 2.5107142857142857, + "grad_norm": 1.71652090549469, + "learning_rate": 3.286711130668395e-06, + "loss": 0.9943090677261353, + "step": 1406 + }, + { + "epoch": 2.5142857142857142, + "grad_norm": 0.6390452980995178, + "learning_rate": 3.268514410671033e-06, + "loss": 0.9306870102882385, + "step": 1408 + }, + { + "epoch": 2.517857142857143, + "grad_norm": 1.2343145608901978, + "learning_rate": 3.25043751813337e-06, + "loss": 0.6592239141464233, + "step": 1410 + }, + { + "epoch": 2.5214285714285714, + "grad_norm": 0.4998447000980377, + "learning_rate": 3.2324807332225307e-06, + "loss": 1.1522630453109741, + "step": 1412 + }, + { + "epoch": 2.525, + "grad_norm": 0.384732186794281, + "learning_rate": 3.2146443342441384e-06, + "loss": 0.9651750922203064, + "step": 1414 + }, + { + "epoch": 2.5285714285714285, + "grad_norm": 1.3452273607254028, + "learning_rate": 3.1969285976380044e-06, + "loss": 1.0436429977416992, + "step": 1416 + }, + { + "epoch": 2.532142857142857, + "grad_norm": 0.4325752556324005, + "learning_rate": 3.1793337979738303e-06, + "loss": 1.0330166816711426, + "step": 1418 + }, + { + "epoch": 2.5357142857142856, + "grad_norm": 0.6546037793159485, + "learning_rate": 3.161860207946963e-06, + "loss": 1.0312299728393555, + "step": 1420 + }, + { + "epoch": 2.539285714285714, + "grad_norm": 2.958420515060425, + "learning_rate": 3.1445080983741676e-06, + "loss": 1.0116913318634033, + "step": 1422 + }, + { + "epoch": 2.5428571428571427, + "grad_norm": 0.3540860712528229, + "learning_rate": 3.127277738189425e-06, + "loss": 0.994167149066925, + "step": 1424 + }, + { + "epoch": 2.5464285714285713, + "grad_norm": 1.0055928230285645, + "learning_rate": 3.1101693944397702e-06, + "loss": 1.0106935501098633, + "step": 1426 + }, + { + "epoch": 2.55, + "grad_norm": 0.5099295377731323, + "learning_rate": 3.0931833322811467e-06, + "loss": 1.0180671215057373, + "step": 1428 + }, + { + "epoch": 2.553571428571429, + "grad_norm": 0.4464087188243866, + "learning_rate": 3.0763198149743024e-06, + "loss": 1.0863144397735596, + "step": 1430 + }, + { + "epoch": 2.557142857142857, + "grad_norm": 0.4935389459133148, + "learning_rate": 3.0595791038807127e-06, + "loss": 0.9617230892181396, + "step": 1432 + }, + { + "epoch": 2.560714285714286, + "grad_norm": 0.3790127635002136, + "learning_rate": 3.0429614584585134e-06, + "loss": 0.9597178101539612, + "step": 1434 + }, + { + "epoch": 2.564285714285714, + "grad_norm": 0.44204089045524597, + "learning_rate": 3.0264671362584983e-06, + "loss": 0.9430738687515259, + "step": 1436 + }, + { + "epoch": 2.567857142857143, + "grad_norm": 0.6700682640075684, + "learning_rate": 3.0100963929201165e-06, + "loss": 1.0032017230987549, + "step": 1438 + }, + { + "epoch": 2.571428571428571, + "grad_norm": 0.6156457662582397, + "learning_rate": 2.993849482167518e-06, + "loss": 1.018100380897522, + "step": 1440 + }, + { + "epoch": 2.575, + "grad_norm": 0.21810433268547058, + "learning_rate": 2.9777266558056136e-06, + "loss": 0.8953680992126465, + "step": 1442 + }, + { + "epoch": 2.5785714285714287, + "grad_norm": 0.3726331293582916, + "learning_rate": 2.961728163716177e-06, + "loss": 1.0808253288269043, + "step": 1444 + }, + { + "epoch": 2.5821428571428573, + "grad_norm": 0.46072059869766235, + "learning_rate": 2.9458542538539646e-06, + "loss": 1.1591756343841553, + "step": 1446 + }, + { + "epoch": 2.585714285714286, + "grad_norm": 0.5146509408950806, + "learning_rate": 2.9301051722428908e-06, + "loss": 0.9750396013259888, + "step": 1448 + }, + { + "epoch": 2.5892857142857144, + "grad_norm": 0.27188050746917725, + "learning_rate": 2.914481162972191e-06, + "loss": 1.0626856088638306, + "step": 1450 + }, + { + "epoch": 2.592857142857143, + "grad_norm": 2.597723960876465, + "learning_rate": 2.8989824681926573e-06, + "loss": 1.0312139987945557, + "step": 1452 + }, + { + "epoch": 2.5964285714285715, + "grad_norm": 0.5178307890892029, + "learning_rate": 2.883609328112874e-06, + "loss": 1.0592162609100342, + "step": 1454 + }, + { + "epoch": 2.6, + "grad_norm": 0.3589671850204468, + "learning_rate": 2.8683619809955022e-06, + "loss": 0.9793047904968262, + "step": 1456 + }, + { + "epoch": 2.6035714285714286, + "grad_norm": 1.327312707901001, + "learning_rate": 2.8532406631535835e-06, + "loss": 0.9267488718032837, + "step": 1458 + }, + { + "epoch": 2.607142857142857, + "grad_norm": 0.6031576991081238, + "learning_rate": 2.8382456089468774e-06, + "loss": 0.7647952437400818, + "step": 1460 + }, + { + "epoch": 2.6107142857142858, + "grad_norm": 0.740556001663208, + "learning_rate": 2.8233770507782284e-06, + "loss": 0.9490369558334351, + "step": 1462 + }, + { + "epoch": 2.6142857142857143, + "grad_norm": 0.4929530918598175, + "learning_rate": 2.8086352190899685e-06, + "loss": 1.063023567199707, + "step": 1464 + }, + { + "epoch": 2.617857142857143, + "grad_norm": 0.36942094564437866, + "learning_rate": 2.7940203423603424e-06, + "loss": 0.8732975721359253, + "step": 1466 + }, + { + "epoch": 2.6214285714285714, + "grad_norm": 1.9331592321395874, + "learning_rate": 2.7795326470999594e-06, + "loss": 0.9476880431175232, + "step": 1468 + }, + { + "epoch": 2.625, + "grad_norm": 1.912407636642456, + "learning_rate": 2.765172357848298e-06, + "loss": 0.5515610575675964, + "step": 1470 + }, + { + "epoch": 2.6285714285714286, + "grad_norm": 0.522169291973114, + "learning_rate": 2.7509396971702148e-06, + "loss": 0.9627550840377808, + "step": 1472 + }, + { + "epoch": 2.632142857142857, + "grad_norm": 0.4970167279243469, + "learning_rate": 2.736834885652498e-06, + "loss": 0.9455366134643555, + "step": 1474 + }, + { + "epoch": 2.6357142857142857, + "grad_norm": 0.6564792394638062, + "learning_rate": 2.7228581419004484e-06, + "loss": 1.0734994411468506, + "step": 1476 + }, + { + "epoch": 2.6392857142857142, + "grad_norm": 1.9194797277450562, + "learning_rate": 2.7090096825344918e-06, + "loss": 0.7485166192054749, + "step": 1478 + }, + { + "epoch": 2.642857142857143, + "grad_norm": 0.6008222699165344, + "learning_rate": 2.6952897221868208e-06, + "loss": 0.7062904834747314, + "step": 1480 + }, + { + "epoch": 2.6464285714285714, + "grad_norm": 0.586191713809967, + "learning_rate": 2.681698473498069e-06, + "loss": 1.01906156539917, + "step": 1482 + }, + { + "epoch": 2.65, + "grad_norm": 0.5794109106063843, + "learning_rate": 2.6682361471140162e-06, + "loss": 1.0513708591461182, + "step": 1484 + }, + { + "epoch": 2.6535714285714285, + "grad_norm": 0.5425417423248291, + "learning_rate": 2.654902951682319e-06, + "loss": 1.0853989124298096, + "step": 1486 + }, + { + "epoch": 2.657142857142857, + "grad_norm": 0.755917489528656, + "learning_rate": 2.641699093849289e-06, + "loss": 1.0674422979354858, + "step": 1488 + }, + { + "epoch": 2.6607142857142856, + "grad_norm": 0.7524024844169617, + "learning_rate": 2.628624778256671e-06, + "loss": 0.8497341275215149, + "step": 1490 + }, + { + "epoch": 2.664285714285714, + "grad_norm": 0.8485654592514038, + "learning_rate": 2.6156802075384903e-06, + "loss": 0.9786584377288818, + "step": 1492 + }, + { + "epoch": 2.6678571428571427, + "grad_norm": 0.47498440742492676, + "learning_rate": 2.602865582317899e-06, + "loss": 1.0709497928619385, + "step": 1494 + }, + { + "epoch": 2.6714285714285713, + "grad_norm": 0.4130082428455353, + "learning_rate": 2.5901811012040766e-06, + "loss": 1.049452781677246, + "step": 1496 + }, + { + "epoch": 2.675, + "grad_norm": 1.847033977508545, + "learning_rate": 2.5776269607891426e-06, + "loss": 0.5213537216186523, + "step": 1498 + }, + { + "epoch": 2.678571428571429, + "grad_norm": 0.40808621048927307, + "learning_rate": 2.5652033556451156e-06, + "loss": 1.0305453538894653, + "step": 1500 + }, + { + "epoch": 2.682142857142857, + "grad_norm": 0.6981632709503174, + "learning_rate": 2.5529104783208936e-06, + "loss": 1.158233404159546, + "step": 1502 + }, + { + "epoch": 2.685714285714286, + "grad_norm": 1.0872377157211304, + "learning_rate": 2.5407485193392765e-06, + "loss": 1.0444018840789795, + "step": 1504 + }, + { + "epoch": 2.689285714285714, + "grad_norm": 1.1252286434173584, + "learning_rate": 2.5287176671940043e-06, + "loss": 0.6784479022026062, + "step": 1506 + }, + { + "epoch": 2.692857142857143, + "grad_norm": 0.35642820596694946, + "learning_rate": 2.5168181083468433e-06, + "loss": 1.0736439228057861, + "step": 1508 + }, + { + "epoch": 2.696428571428571, + "grad_norm": 0.5642552375793457, + "learning_rate": 2.505050027224692e-06, + "loss": 1.065712809562683, + "step": 1510 + }, + { + "epoch": 2.7, + "grad_norm": 0.44292396306991577, + "learning_rate": 2.4934136062167245e-06, + "loss": 1.0252857208251953, + "step": 1512 + }, + { + "epoch": 2.7035714285714287, + "grad_norm": 0.8530491590499878, + "learning_rate": 2.481909025671561e-06, + "loss": 0.7706226706504822, + "step": 1514 + }, + { + "epoch": 2.7071428571428573, + "grad_norm": 1.9386380910873413, + "learning_rate": 2.470536463894476e-06, + "loss": 1.1056604385375977, + "step": 1516 + }, + { + "epoch": 2.710714285714286, + "grad_norm": 1.086133360862732, + "learning_rate": 2.459296097144634e-06, + "loss": 1.1189930438995361, + "step": 1518 + }, + { + "epoch": 2.7142857142857144, + "grad_norm": 0.3476833999156952, + "learning_rate": 2.4481880996323595e-06, + "loss": 0.9641977548599243, + "step": 1520 + }, + { + "epoch": 2.717857142857143, + "grad_norm": 1.6515637636184692, + "learning_rate": 2.4372126435164287e-06, + "loss": 0.7884814143180847, + "step": 1522 + }, + { + "epoch": 2.7214285714285715, + "grad_norm": 0.8657196760177612, + "learning_rate": 2.426369898901412e-06, + "loss": 1.0643302202224731, + "step": 1524 + }, + { + "epoch": 2.725, + "grad_norm": 1.4171736240386963, + "learning_rate": 2.4156600338350315e-06, + "loss": 0.894715428352356, + "step": 1526 + }, + { + "epoch": 2.7285714285714286, + "grad_norm": 0.8196776509284973, + "learning_rate": 2.40508321430556e-06, + "loss": 0.7902459502220154, + "step": 1528 + }, + { + "epoch": 2.732142857142857, + "grad_norm": 0.982892632484436, + "learning_rate": 2.3946396042392415e-06, + "loss": 1.0700161457061768, + "step": 1530 + }, + { + "epoch": 2.7357142857142858, + "grad_norm": 0.47785484790802, + "learning_rate": 2.3843293654977613e-06, + "loss": 0.8019880056381226, + "step": 1532 + }, + { + "epoch": 2.7392857142857143, + "grad_norm": 0.591884970664978, + "learning_rate": 2.374152657875726e-06, + "loss": 0.5509794354438782, + "step": 1534 + }, + { + "epoch": 2.742857142857143, + "grad_norm": 0.584327220916748, + "learning_rate": 2.364109639098196e-06, + "loss": 1.0576118230819702, + "step": 1536 + }, + { + "epoch": 2.7464285714285714, + "grad_norm": 0.3715934455394745, + "learning_rate": 2.3542004648182366e-06, + "loss": 0.9892575740814209, + "step": 1538 + }, + { + "epoch": 2.75, + "grad_norm": 0.7555781602859497, + "learning_rate": 2.344425288614504e-06, + "loss": 1.0496368408203125, + "step": 1540 + }, + { + "epoch": 2.7535714285714286, + "grad_norm": 0.47145044803619385, + "learning_rate": 2.334784261988871e-06, + "loss": 1.0199828147888184, + "step": 1542 + }, + { + "epoch": 2.757142857142857, + "grad_norm": 0.4803292155265808, + "learning_rate": 2.3252775343640726e-06, + "loss": 1.2034311294555664, + "step": 1544 + }, + { + "epoch": 2.7607142857142857, + "grad_norm": 0.5480961203575134, + "learning_rate": 2.3159052530813944e-06, + "loss": 1.0263570547103882, + "step": 1546 + }, + { + "epoch": 2.7642857142857142, + "grad_norm": 0.7437337040901184, + "learning_rate": 2.3066675633983863e-06, + "loss": 1.065530776977539, + "step": 1548 + }, + { + "epoch": 2.767857142857143, + "grad_norm": 0.33723706007003784, + "learning_rate": 2.2975646084866126e-06, + "loss": 0.8913886547088623, + "step": 1550 + }, + { + "epoch": 2.7714285714285714, + "grad_norm": 0.7276679277420044, + "learning_rate": 2.2885965294294334e-06, + "loss": 0.8413328528404236, + "step": 1552 + }, + { + "epoch": 2.775, + "grad_norm": 0.4004630744457245, + "learning_rate": 2.2797634652198187e-06, + "loss": 0.957332968711853, + "step": 1554 + }, + { + "epoch": 2.7785714285714285, + "grad_norm": 0.6412128806114197, + "learning_rate": 2.2710655527581886e-06, + "loss": 0.3783353269100189, + "step": 1556 + }, + { + "epoch": 2.782142857142857, + "grad_norm": 1.0582815408706665, + "learning_rate": 2.2625029268502984e-06, + "loss": 0.8345463275909424, + "step": 1558 + }, + { + "epoch": 2.7857142857142856, + "grad_norm": 0.5835173726081848, + "learning_rate": 2.25407572020515e-06, + "loss": 1.0623098611831665, + "step": 1560 + }, + { + "epoch": 2.789285714285714, + "grad_norm": 2.572283983230591, + "learning_rate": 2.245784063432925e-06, + "loss": 0.7072303295135498, + "step": 1562 + }, + { + "epoch": 2.7928571428571427, + "grad_norm": 0.5214061737060547, + "learning_rate": 2.237628085042972e-06, + "loss": 1.0455104112625122, + "step": 1564 + }, + { + "epoch": 2.7964285714285713, + "grad_norm": 1.3393778800964355, + "learning_rate": 2.2296079114418113e-06, + "loss": 0.9467559456825256, + "step": 1566 + }, + { + "epoch": 2.8, + "grad_norm": 0.25553128123283386, + "learning_rate": 2.2217236669311687e-06, + "loss": 0.928917407989502, + "step": 1568 + }, + { + "epoch": 2.803571428571429, + "grad_norm": 0.4923466145992279, + "learning_rate": 2.2139754737060606e-06, + "loss": 1.0020575523376465, + "step": 1570 + }, + { + "epoch": 2.807142857142857, + "grad_norm": 0.5088347792625427, + "learning_rate": 2.206363451852891e-06, + "loss": 1.0174009799957275, + "step": 1572 + }, + { + "epoch": 2.810714285714286, + "grad_norm": 1.0624254941940308, + "learning_rate": 2.1988877193475942e-06, + "loss": 0.943864107131958, + "step": 1574 + }, + { + "epoch": 2.814285714285714, + "grad_norm": 1.2148317098617554, + "learning_rate": 2.1915483920538098e-06, + "loss": 0.9655317068099976, + "step": 1576 + }, + { + "epoch": 2.817857142857143, + "grad_norm": 0.40001875162124634, + "learning_rate": 2.184345583721078e-06, + "loss": 1.03749680519104, + "step": 1578 + }, + { + "epoch": 2.821428571428571, + "grad_norm": 0.5738964080810547, + "learning_rate": 2.177279405983083e-06, + "loss": 1.0436820983886719, + "step": 1580 + }, + { + "epoch": 2.825, + "grad_norm": 1.9478952884674072, + "learning_rate": 2.1703499683559218e-06, + "loss": 0.823712944984436, + "step": 1582 + }, + { + "epoch": 2.8285714285714287, + "grad_norm": 0.5372344255447388, + "learning_rate": 2.1635573782364083e-06, + "loss": 1.0900218486785889, + "step": 1584 + }, + { + "epoch": 2.8321428571428573, + "grad_norm": 0.4769842028617859, + "learning_rate": 2.1569017409004058e-06, + "loss": 1.032717227935791, + "step": 1586 + }, + { + "epoch": 2.835714285714286, + "grad_norm": 3.2297580242156982, + "learning_rate": 2.1503831595011972e-06, + "loss": 0.6889293193817139, + "step": 1588 + }, + { + "epoch": 2.8392857142857144, + "grad_norm": 1.6764355897903442, + "learning_rate": 2.1440017350678806e-06, + "loss": 0.5947354435920715, + "step": 1590 + }, + { + "epoch": 2.842857142857143, + "grad_norm": 0.44973084330558777, + "learning_rate": 2.137757566503817e-06, + "loss": 1.1642112731933594, + "step": 1592 + }, + { + "epoch": 2.8464285714285715, + "grad_norm": 0.27710893750190735, + "learning_rate": 2.1316507505850835e-06, + "loss": 0.9468015432357788, + "step": 1594 + }, + { + "epoch": 2.85, + "grad_norm": 0.4122758209705353, + "learning_rate": 2.125681381958982e-06, + "loss": 1.0558395385742188, + "step": 1596 + }, + { + "epoch": 2.8535714285714286, + "grad_norm": 0.8191478848457336, + "learning_rate": 2.1198495531425663e-06, + "loss": 1.0057258605957031, + "step": 1598 + }, + { + "epoch": 2.857142857142857, + "grad_norm": 0.8358668088912964, + "learning_rate": 2.114155354521211e-06, + "loss": 1.0604225397109985, + "step": 1600 + }, + { + "epoch": 2.8607142857142858, + "grad_norm": 0.897541880607605, + "learning_rate": 2.1085988743472153e-06, + "loss": 1.0604546070098877, + "step": 1602 + }, + { + "epoch": 2.8642857142857143, + "grad_norm": 0.9542982578277588, + "learning_rate": 2.1031801987384255e-06, + "loss": 1.0365536212921143, + "step": 1604 + }, + { + "epoch": 2.867857142857143, + "grad_norm": 0.7912425398826599, + "learning_rate": 2.097899411676908e-06, + "loss": 0.9559098482131958, + "step": 1606 + }, + { + "epoch": 2.8714285714285714, + "grad_norm": 1.4533398151397705, + "learning_rate": 2.092756595007645e-06, + "loss": 0.7953978776931763, + "step": 1608 + }, + { + "epoch": 2.875, + "grad_norm": 0.358332097530365, + "learning_rate": 2.087751828437267e-06, + "loss": 1.0427325963974, + "step": 1610 + }, + { + "epoch": 2.8785714285714286, + "grad_norm": 0.2145242542028427, + "learning_rate": 2.082885189532815e-06, + "loss": 1.0498052835464478, + "step": 1612 + }, + { + "epoch": 2.882142857142857, + "grad_norm": 1.0103882551193237, + "learning_rate": 2.0781567537205387e-06, + "loss": 1.0514143705368042, + "step": 1614 + }, + { + "epoch": 2.8857142857142857, + "grad_norm": 0.669739305973053, + "learning_rate": 2.0735665942847328e-06, + "loss": 1.0135245323181152, + "step": 1616 + }, + { + "epoch": 2.8892857142857142, + "grad_norm": 0.576935887336731, + "learning_rate": 2.0691147823665953e-06, + "loss": 1.0457009077072144, + "step": 1618 + }, + { + "epoch": 2.892857142857143, + "grad_norm": 1.905529499053955, + "learning_rate": 2.0648013869631246e-06, + "loss": 1.1598899364471436, + "step": 1620 + }, + { + "epoch": 2.8964285714285714, + "grad_norm": 0.44723814725875854, + "learning_rate": 2.060626474926055e-06, + "loss": 1.0284682512283325, + "step": 1622 + }, + { + "epoch": 2.9, + "grad_norm": 0.4723893702030182, + "learning_rate": 2.056590110960817e-06, + "loss": 0.5460047721862793, + "step": 1624 + }, + { + "epoch": 2.9035714285714285, + "grad_norm": 0.8015878796577454, + "learning_rate": 2.0526923576255363e-06, + "loss": 0.7874211668968201, + "step": 1626 + }, + { + "epoch": 2.907142857142857, + "grad_norm": 0.6421289443969727, + "learning_rate": 2.0489332753300613e-06, + "loss": 1.013157844543457, + "step": 1628 + }, + { + "epoch": 2.9107142857142856, + "grad_norm": 0.5567857027053833, + "learning_rate": 2.0453129223350293e-06, + "loss": 0.9200034737586975, + "step": 1630 + }, + { + "epoch": 2.914285714285714, + "grad_norm": 0.8033483624458313, + "learning_rate": 2.041831354750966e-06, + "loss": 0.9019091129302979, + "step": 1632 + }, + { + "epoch": 2.9178571428571427, + "grad_norm": 0.24038927257061005, + "learning_rate": 2.0384886265374078e-06, + "loss": 0.9200948476791382, + "step": 1634 + }, + { + "epoch": 2.9214285714285713, + "grad_norm": 0.48659539222717285, + "learning_rate": 2.0352847895020754e-06, + "loss": 1.117785096168518, + "step": 1636 + }, + { + "epoch": 2.925, + "grad_norm": 0.6561591029167175, + "learning_rate": 2.032219893300064e-06, + "loss": 0.909339189529419, + "step": 1638 + }, + { + "epoch": 2.928571428571429, + "grad_norm": 0.4943861663341522, + "learning_rate": 2.029293985433074e-06, + "loss": 1.068084478378296, + "step": 1640 + }, + { + "epoch": 2.932142857142857, + "grad_norm": 0.4287692606449127, + "learning_rate": 2.02650711124868e-06, + "loss": 0.9846458435058594, + "step": 1642 + }, + { + "epoch": 2.935714285714286, + "grad_norm": 0.9221391677856445, + "learning_rate": 2.0238593139396225e-06, + "loss": 0.9131702184677124, + "step": 1644 + }, + { + "epoch": 2.939285714285714, + "grad_norm": 0.5019553303718567, + "learning_rate": 2.02135063454314e-06, + "loss": 0.990814208984375, + "step": 1646 + }, + { + "epoch": 2.942857142857143, + "grad_norm": 0.3528009057044983, + "learning_rate": 2.0189811119403356e-06, + "loss": 1.043567419052124, + "step": 1648 + }, + { + "epoch": 2.946428571428571, + "grad_norm": 1.9814422130584717, + "learning_rate": 2.0167507828555718e-06, + "loss": 0.600018322467804, + "step": 1650 + }, + { + "epoch": 2.95, + "grad_norm": 0.4034280776977539, + "learning_rate": 2.014659681855898e-06, + "loss": 1.066345453262329, + "step": 1652 + }, + { + "epoch": 2.9535714285714287, + "grad_norm": 0.35513538122177124, + "learning_rate": 2.012707841350526e-06, + "loss": 1.023656964302063, + "step": 1654 + }, + { + "epoch": 2.9571428571428573, + "grad_norm": 0.3866799771785736, + "learning_rate": 2.0108952915903134e-06, + "loss": 0.89092618227005, + "step": 1656 + }, + { + "epoch": 2.960714285714286, + "grad_norm": 0.4807373583316803, + "learning_rate": 2.0092220606673037e-06, + "loss": 1.0590665340423584, + "step": 1658 + }, + { + "epoch": 2.9642857142857144, + "grad_norm": 1.271697998046875, + "learning_rate": 2.0076881745142912e-06, + "loss": 1.0663166046142578, + "step": 1660 + }, + { + "epoch": 2.967857142857143, + "grad_norm": 0.5819157361984253, + "learning_rate": 2.0062936569044127e-06, + "loss": 1.0056861639022827, + "step": 1662 + }, + { + "epoch": 2.9714285714285715, + "grad_norm": 1.1669540405273438, + "learning_rate": 2.0050385294507877e-06, + "loss": 1.175034761428833, + "step": 1664 + }, + { + "epoch": 2.975, + "grad_norm": 1.3715438842773438, + "learning_rate": 2.0039228116061747e-06, + "loss": 1.0603545904159546, + "step": 1666 + }, + { + "epoch": 2.9785714285714286, + "grad_norm": 0.3443331718444824, + "learning_rate": 2.002946520662675e-06, + "loss": 1.029945731163025, + "step": 1668 + }, + { + "epoch": 2.982142857142857, + "grad_norm": 0.9261486530303955, + "learning_rate": 2.0021096717514655e-06, + "loss": 0.5534359812736511, + "step": 1670 + }, + { + "epoch": 2.9857142857142858, + "grad_norm": 0.34926119446754456, + "learning_rate": 2.001412277842559e-06, + "loss": 0.4894912838935852, + "step": 1672 + }, + { + "epoch": 2.9892857142857143, + "grad_norm": 0.3799968361854553, + "learning_rate": 2.0008543497446085e-06, + "loss": 1.021775722503662, + "step": 1674 + }, + { + "epoch": 2.992857142857143, + "grad_norm": 0.33220550417900085, + "learning_rate": 2.0004358961047375e-06, + "loss": 1.0939745903015137, + "step": 1676 + }, + { + "epoch": 2.9964285714285714, + "grad_norm": 1.253113031387329, + "learning_rate": 2.0001569234084046e-06, + "loss": 0.5844765901565552, + "step": 1678 + }, + { + "epoch": 3.0, + "grad_norm": 1.1644887924194336, + "learning_rate": 2.0000174359793066e-06, + "loss": 0.9837155342102051, + "step": 1680 + }, + { + "epoch": 3.0, + "step": 1680, + "total_flos": 2.50959750918203e+18, + "train_loss": 1.167248018547183, + "train_runtime": 8368.8806, + "train_samples_per_second": 3.212, + "train_steps_per_second": 0.201 + } + ], + "logging_steps": 2, + "max_steps": 1680, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 9999999, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 2.50959750918203e+18, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +}