9b-35 / trainer_state.json
furproxy's picture
Upload folder using huggingface_hub
9eeb250 verified
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 3.0,
"eval_steps": 500,
"global_step": 1410,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.004259850905218318,
"grad_norm": 1.125,
"learning_rate": 2.8169014084507043e-07,
"loss": 1.820163607597351,
"step": 2
},
{
"epoch": 0.008519701810436636,
"grad_norm": 1.2734375,
"learning_rate": 8.450704225352114e-07,
"loss": 1.8241215944290161,
"step": 4
},
{
"epoch": 0.012779552715654952,
"grad_norm": 0.96875,
"learning_rate": 1.4084507042253523e-06,
"loss": 1.7736045122146606,
"step": 6
},
{
"epoch": 0.01703940362087327,
"grad_norm": 1.6953125,
"learning_rate": 1.971830985915493e-06,
"loss": 1.9067131280899048,
"step": 8
},
{
"epoch": 0.021299254526091587,
"grad_norm": 1.453125,
"learning_rate": 2.535211267605634e-06,
"loss": 1.8296759128570557,
"step": 10
},
{
"epoch": 0.025559105431309903,
"grad_norm": 1.3984375,
"learning_rate": 3.0985915492957746e-06,
"loss": 1.698632001876831,
"step": 12
},
{
"epoch": 0.029818956336528223,
"grad_norm": 1.734375,
"learning_rate": 3.6619718309859158e-06,
"loss": 1.7496201992034912,
"step": 14
},
{
"epoch": 0.03407880724174654,
"grad_norm": 1.0703125,
"learning_rate": 4.225352112676057e-06,
"loss": 1.7501145601272583,
"step": 16
},
{
"epoch": 0.038338658146964855,
"grad_norm": 1.3984375,
"learning_rate": 4.788732394366197e-06,
"loss": 1.8459789752960205,
"step": 18
},
{
"epoch": 0.042598509052183174,
"grad_norm": 0.83203125,
"learning_rate": 5.352112676056338e-06,
"loss": 1.8062622547149658,
"step": 20
},
{
"epoch": 0.046858359957401494,
"grad_norm": 1.9140625,
"learning_rate": 5.915492957746479e-06,
"loss": 1.6213020086288452,
"step": 22
},
{
"epoch": 0.051118210862619806,
"grad_norm": 0.921875,
"learning_rate": 6.478873239436621e-06,
"loss": 1.6766525506973267,
"step": 24
},
{
"epoch": 0.055378061767838126,
"grad_norm": 0.97265625,
"learning_rate": 7.042253521126761e-06,
"loss": 1.6821609735488892,
"step": 26
},
{
"epoch": 0.059637912673056445,
"grad_norm": 0.84765625,
"learning_rate": 7.6056338028169015e-06,
"loss": 1.6819344758987427,
"step": 28
},
{
"epoch": 0.06389776357827476,
"grad_norm": 1.0546875,
"learning_rate": 8.169014084507043e-06,
"loss": 1.668290376663208,
"step": 30
},
{
"epoch": 0.06815761448349308,
"grad_norm": 0.796875,
"learning_rate": 8.732394366197183e-06,
"loss": 1.7105621099472046,
"step": 32
},
{
"epoch": 0.0724174653887114,
"grad_norm": 0.70703125,
"learning_rate": 9.295774647887325e-06,
"loss": 1.640573501586914,
"step": 34
},
{
"epoch": 0.07667731629392971,
"grad_norm": 0.90625,
"learning_rate": 9.859154929577466e-06,
"loss": 1.7408779859542847,
"step": 36
},
{
"epoch": 0.08093716719914804,
"grad_norm": 0.87109375,
"learning_rate": 1.0422535211267606e-05,
"loss": 1.7062400579452515,
"step": 38
},
{
"epoch": 0.08519701810436635,
"grad_norm": 0.85546875,
"learning_rate": 1.0985915492957748e-05,
"loss": 1.6299972534179688,
"step": 40
},
{
"epoch": 0.08945686900958466,
"grad_norm": 0.71484375,
"learning_rate": 1.1549295774647888e-05,
"loss": 1.7348750829696655,
"step": 42
},
{
"epoch": 0.09371671991480299,
"grad_norm": 0.63671875,
"learning_rate": 1.211267605633803e-05,
"loss": 1.5983866453170776,
"step": 44
},
{
"epoch": 0.0979765708200213,
"grad_norm": 0.9296875,
"learning_rate": 1.2676056338028171e-05,
"loss": 1.6268177032470703,
"step": 46
},
{
"epoch": 0.10223642172523961,
"grad_norm": 0.609375,
"learning_rate": 1.323943661971831e-05,
"loss": 1.6371753215789795,
"step": 48
},
{
"epoch": 0.10649627263045794,
"grad_norm": 0.52734375,
"learning_rate": 1.380281690140845e-05,
"loss": 1.4547175168991089,
"step": 50
},
{
"epoch": 0.11075612353567625,
"grad_norm": 0.734375,
"learning_rate": 1.4366197183098594e-05,
"loss": 1.5619021654129028,
"step": 52
},
{
"epoch": 0.11501597444089456,
"grad_norm": 1.359375,
"learning_rate": 1.4929577464788734e-05,
"loss": 1.5433554649353027,
"step": 54
},
{
"epoch": 0.11927582534611289,
"grad_norm": 1.0625,
"learning_rate": 1.5492957746478872e-05,
"loss": 1.4989991188049316,
"step": 56
},
{
"epoch": 0.1235356762513312,
"grad_norm": 0.63671875,
"learning_rate": 1.6056338028169017e-05,
"loss": 1.549402117729187,
"step": 58
},
{
"epoch": 0.12779552715654952,
"grad_norm": 0.6953125,
"learning_rate": 1.6619718309859155e-05,
"loss": 1.434061050415039,
"step": 60
},
{
"epoch": 0.13205537806176784,
"grad_norm": 0.73828125,
"learning_rate": 1.7183098591549297e-05,
"loss": 1.5171629190444946,
"step": 62
},
{
"epoch": 0.13631522896698617,
"grad_norm": 0.7734375,
"learning_rate": 1.774647887323944e-05,
"loss": 1.3912996053695679,
"step": 64
},
{
"epoch": 0.14057507987220447,
"grad_norm": 0.62890625,
"learning_rate": 1.830985915492958e-05,
"loss": 1.4431164264678955,
"step": 66
},
{
"epoch": 0.1448349307774228,
"grad_norm": 0.625,
"learning_rate": 1.887323943661972e-05,
"loss": 1.385206699371338,
"step": 68
},
{
"epoch": 0.14909478168264112,
"grad_norm": 0.74609375,
"learning_rate": 1.943661971830986e-05,
"loss": 1.4210329055786133,
"step": 70
},
{
"epoch": 0.15335463258785942,
"grad_norm": 0.57421875,
"learning_rate": 2e-05,
"loss": 1.4347320795059204,
"step": 72
},
{
"epoch": 0.15761448349307774,
"grad_norm": 0.6484375,
"learning_rate": 1.9999911923960593e-05,
"loss": 1.4454214572906494,
"step": 74
},
{
"epoch": 0.16187433439829607,
"grad_norm": 0.71484375,
"learning_rate": 1.9999647697781703e-05,
"loss": 1.34307861328125,
"step": 76
},
{
"epoch": 0.16613418530351437,
"grad_norm": 0.71484375,
"learning_rate": 1.9999207327281333e-05,
"loss": 1.3641051054000854,
"step": 78
},
{
"epoch": 0.1703940362087327,
"grad_norm": 1.2890625,
"learning_rate": 1.9998590822156014e-05,
"loss": 1.3956571817398071,
"step": 80
},
{
"epoch": 0.17465388711395102,
"grad_norm": 0.69140625,
"learning_rate": 1.999779819598057e-05,
"loss": 1.3472541570663452,
"step": 82
},
{
"epoch": 0.17891373801916932,
"grad_norm": 0.703125,
"learning_rate": 1.999682946620784e-05,
"loss": 1.3230623006820679,
"step": 84
},
{
"epoch": 0.18317358892438765,
"grad_norm": 0.64453125,
"learning_rate": 1.999568465416831e-05,
"loss": 1.3147826194763184,
"step": 86
},
{
"epoch": 0.18743343982960597,
"grad_norm": 1.21875,
"learning_rate": 1.9994363785069595e-05,
"loss": 1.3436126708984375,
"step": 88
},
{
"epoch": 0.19169329073482427,
"grad_norm": 1.0390625,
"learning_rate": 1.9992866887995928e-05,
"loss": 1.2736291885375977,
"step": 90
},
{
"epoch": 0.1959531416400426,
"grad_norm": 0.8828125,
"learning_rate": 1.99911939959075e-05,
"loss": 1.4134427309036255,
"step": 92
},
{
"epoch": 0.20021299254526093,
"grad_norm": 1.6171875,
"learning_rate": 1.9989345145639734e-05,
"loss": 1.3293910026550293,
"step": 94
},
{
"epoch": 0.20447284345047922,
"grad_norm": 0.609375,
"learning_rate": 1.998732037790249e-05,
"loss": 1.3805630207061768,
"step": 96
},
{
"epoch": 0.20873269435569755,
"grad_norm": 0.578125,
"learning_rate": 1.9985119737279156e-05,
"loss": 1.3184444904327393,
"step": 98
},
{
"epoch": 0.21299254526091588,
"grad_norm": 0.6640625,
"learning_rate": 1.9982743272225637e-05,
"loss": 1.2683892250061035,
"step": 100
},
{
"epoch": 0.21725239616613418,
"grad_norm": 0.609375,
"learning_rate": 1.9980191035069357e-05,
"loss": 1.2953828573226929,
"step": 102
},
{
"epoch": 0.2215122470713525,
"grad_norm": 0.609375,
"learning_rate": 1.9977463082008048e-05,
"loss": 1.3181703090667725,
"step": 104
},
{
"epoch": 0.22577209797657083,
"grad_norm": 1.4375,
"learning_rate": 1.997455947310854e-05,
"loss": 1.3661359548568726,
"step": 106
},
{
"epoch": 0.23003194888178913,
"grad_norm": 1.2578125,
"learning_rate": 1.997148027230541e-05,
"loss": 1.277463436126709,
"step": 108
},
{
"epoch": 0.23429179978700745,
"grad_norm": 0.7265625,
"learning_rate": 1.9968225547399624e-05,
"loss": 1.244004249572754,
"step": 110
},
{
"epoch": 0.23855165069222578,
"grad_norm": 0.77734375,
"learning_rate": 1.9964795370057004e-05,
"loss": 1.247603416442871,
"step": 112
},
{
"epoch": 0.24281150159744408,
"grad_norm": 1.0234375,
"learning_rate": 1.996118981580665e-05,
"loss": 1.2560734748840332,
"step": 114
},
{
"epoch": 0.2470713525026624,
"grad_norm": 1.0859375,
"learning_rate": 1.99574089640393e-05,
"loss": 1.2559478282928467,
"step": 116
},
{
"epoch": 0.25133120340788073,
"grad_norm": 0.65625,
"learning_rate": 1.9953452898005564e-05,
"loss": 1.2884942293167114,
"step": 118
},
{
"epoch": 0.25559105431309903,
"grad_norm": 1.4921875,
"learning_rate": 1.9949321704814103e-05,
"loss": 1.2233449220657349,
"step": 120
},
{
"epoch": 0.2598509052183174,
"grad_norm": 0.69921875,
"learning_rate": 1.9945015475429704e-05,
"loss": 1.282986044883728,
"step": 122
},
{
"epoch": 0.2641107561235357,
"grad_norm": 0.5546875,
"learning_rate": 1.9940534304671266e-05,
"loss": 1.2828234434127808,
"step": 124
},
{
"epoch": 0.268370607028754,
"grad_norm": 0.609375,
"learning_rate": 1.9935878291209737e-05,
"loss": 1.2087836265563965,
"step": 126
},
{
"epoch": 0.27263045793397234,
"grad_norm": 0.86328125,
"learning_rate": 1.9931047537565917e-05,
"loss": 1.2843331098556519,
"step": 128
},
{
"epoch": 0.27689030883919064,
"grad_norm": 0.85546875,
"learning_rate": 1.9926042150108228e-05,
"loss": 1.337433099746704,
"step": 130
},
{
"epoch": 0.28115015974440893,
"grad_norm": 0.515625,
"learning_rate": 1.9920862239050333e-05,
"loss": 1.244279146194458,
"step": 132
},
{
"epoch": 0.2854100106496273,
"grad_norm": 0.6796875,
"learning_rate": 1.9915507918448763e-05,
"loss": 1.191616415977478,
"step": 134
},
{
"epoch": 0.2896698615548456,
"grad_norm": 1.7421875,
"learning_rate": 1.9909979306200337e-05,
"loss": 1.3028016090393066,
"step": 136
},
{
"epoch": 0.2939297124600639,
"grad_norm": 0.84765625,
"learning_rate": 1.9904276524039633e-05,
"loss": 1.2914124727249146,
"step": 138
},
{
"epoch": 0.29818956336528224,
"grad_norm": 0.66015625,
"learning_rate": 1.9898399697536263e-05,
"loss": 1.318739652633667,
"step": 140
},
{
"epoch": 0.30244941427050054,
"grad_norm": 0.6328125,
"learning_rate": 1.9892348956092136e-05,
"loss": 1.2955219745635986,
"step": 142
},
{
"epoch": 0.30670926517571884,
"grad_norm": 0.5859375,
"learning_rate": 1.9886124432938582e-05,
"loss": 1.2421311140060425,
"step": 144
},
{
"epoch": 0.3109691160809372,
"grad_norm": 1.34375,
"learning_rate": 1.9879726265133432e-05,
"loss": 1.2619209289550781,
"step": 146
},
{
"epoch": 0.3152289669861555,
"grad_norm": 1.875,
"learning_rate": 1.987315459355801e-05,
"loss": 1.2909760475158691,
"step": 148
},
{
"epoch": 0.3194888178913738,
"grad_norm": 0.7734375,
"learning_rate": 1.9866409562914022e-05,
"loss": 1.2878739833831787,
"step": 150
},
{
"epoch": 0.32374866879659214,
"grad_norm": 2.515625,
"learning_rate": 1.985949132172036e-05,
"loss": 1.2502833604812622,
"step": 152
},
{
"epoch": 0.32800851970181044,
"grad_norm": 0.58984375,
"learning_rate": 1.9852400022309845e-05,
"loss": 1.2763196229934692,
"step": 154
},
{
"epoch": 0.33226837060702874,
"grad_norm": 0.609375,
"learning_rate": 1.984513582082587e-05,
"loss": 1.3070451021194458,
"step": 156
},
{
"epoch": 0.3365282215122471,
"grad_norm": 0.67578125,
"learning_rate": 1.9837698877218955e-05,
"loss": 1.2980157136917114,
"step": 158
},
{
"epoch": 0.3407880724174654,
"grad_norm": 0.482421875,
"learning_rate": 1.983008935524324e-05,
"loss": 1.2509483098983765,
"step": 160
},
{
"epoch": 0.3450479233226837,
"grad_norm": 0.640625,
"learning_rate": 1.9822307422452862e-05,
"loss": 1.1746965646743774,
"step": 162
},
{
"epoch": 0.34930777422790205,
"grad_norm": 0.546875,
"learning_rate": 1.9814353250198275e-05,
"loss": 1.1982603073120117,
"step": 164
},
{
"epoch": 0.35356762513312034,
"grad_norm": 0.62890625,
"learning_rate": 1.9806227013622483e-05,
"loss": 1.2470253705978394,
"step": 166
},
{
"epoch": 0.35782747603833864,
"grad_norm": 1.734375,
"learning_rate": 1.9797928891657156e-05,
"loss": 1.2994393110275269,
"step": 168
},
{
"epoch": 0.362087326943557,
"grad_norm": 0.76171875,
"learning_rate": 1.9789459067018733e-05,
"loss": 1.2793241739273071,
"step": 170
},
{
"epoch": 0.3663471778487753,
"grad_norm": 0.67578125,
"learning_rate": 1.9780817726204363e-05,
"loss": 1.2065937519073486,
"step": 172
},
{
"epoch": 0.3706070287539936,
"grad_norm": 0.5390625,
"learning_rate": 1.9772005059487815e-05,
"loss": 1.290358543395996,
"step": 174
},
{
"epoch": 0.37486687965921195,
"grad_norm": 0.6171875,
"learning_rate": 1.9763021260915283e-05,
"loss": 1.2577617168426514,
"step": 176
},
{
"epoch": 0.37912673056443025,
"grad_norm": 0.4765625,
"learning_rate": 1.9753866528301128e-05,
"loss": 1.1449509859085083,
"step": 178
},
{
"epoch": 0.38338658146964855,
"grad_norm": 0.6171875,
"learning_rate": 1.9744541063223477e-05,
"loss": 1.2480742931365967,
"step": 180
},
{
"epoch": 0.3876464323748669,
"grad_norm": 0.54296875,
"learning_rate": 1.9735045071019853e-05,
"loss": 1.245701789855957,
"step": 182
},
{
"epoch": 0.3919062832800852,
"grad_norm": 0.7578125,
"learning_rate": 1.9725378760782593e-05,
"loss": 1.2594965696334839,
"step": 184
},
{
"epoch": 0.3961661341853035,
"grad_norm": 0.71875,
"learning_rate": 1.9715542345354285e-05,
"loss": 1.112318992614746,
"step": 186
},
{
"epoch": 0.40042598509052185,
"grad_norm": 0.71484375,
"learning_rate": 1.9705536041323056e-05,
"loss": 1.240614891052246,
"step": 188
},
{
"epoch": 0.40468583599574015,
"grad_norm": 0.435546875,
"learning_rate": 1.9695360069017798e-05,
"loss": 1.219789981842041,
"step": 190
},
{
"epoch": 0.40894568690095845,
"grad_norm": 0.69921875,
"learning_rate": 1.9685014652503368e-05,
"loss": 1.2046880722045898,
"step": 192
},
{
"epoch": 0.4132055378061768,
"grad_norm": 0.65625,
"learning_rate": 1.9674500019575585e-05,
"loss": 1.3095979690551758,
"step": 194
},
{
"epoch": 0.4174653887113951,
"grad_norm": 2.515625,
"learning_rate": 1.966381640175625e-05,
"loss": 1.236427903175354,
"step": 196
},
{
"epoch": 0.4217252396166134,
"grad_norm": 0.62109375,
"learning_rate": 1.9652964034288054e-05,
"loss": 1.2324111461639404,
"step": 198
},
{
"epoch": 0.42598509052183176,
"grad_norm": 0.72265625,
"learning_rate": 1.9641943156129385e-05,
"loss": 1.296306848526001,
"step": 200
},
{
"epoch": 0.43024494142705005,
"grad_norm": 0.91015625,
"learning_rate": 1.9630754009949062e-05,
"loss": 1.2453858852386475,
"step": 202
},
{
"epoch": 0.43450479233226835,
"grad_norm": 0.984375,
"learning_rate": 1.961939684212101e-05,
"loss": 1.2566311359405518,
"step": 204
},
{
"epoch": 0.4387646432374867,
"grad_norm": 0.546875,
"learning_rate": 1.9607871902718817e-05,
"loss": 1.1997429132461548,
"step": 206
},
{
"epoch": 0.443024494142705,
"grad_norm": 0.55078125,
"learning_rate": 1.959617944551024e-05,
"loss": 1.2592103481292725,
"step": 208
},
{
"epoch": 0.4472843450479233,
"grad_norm": 1.5546875,
"learning_rate": 1.9584319727951608e-05,
"loss": 1.2379335165023804,
"step": 210
},
{
"epoch": 0.45154419595314166,
"grad_norm": 0.3984375,
"learning_rate": 1.9572293011182153e-05,
"loss": 1.2336797714233398,
"step": 212
},
{
"epoch": 0.45580404685835996,
"grad_norm": 1.3671875,
"learning_rate": 1.9560099560018276e-05,
"loss": 1.2236690521240234,
"step": 214
},
{
"epoch": 0.46006389776357826,
"grad_norm": 0.53515625,
"learning_rate": 1.95477396429477e-05,
"loss": 1.2254308462142944,
"step": 216
},
{
"epoch": 0.4643237486687966,
"grad_norm": 0.578125,
"learning_rate": 1.9535213532123554e-05,
"loss": 1.1878576278686523,
"step": 218
},
{
"epoch": 0.4685835995740149,
"grad_norm": 0.5078125,
"learning_rate": 1.95225215033584e-05,
"loss": 1.1797441244125366,
"step": 220
},
{
"epoch": 0.4728434504792332,
"grad_norm": 0.8203125,
"learning_rate": 1.950966383611815e-05,
"loss": 1.2125729322433472,
"step": 222
},
{
"epoch": 0.47710330138445156,
"grad_norm": 0.486328125,
"learning_rate": 1.9496640813515896e-05,
"loss": 1.3302891254425049,
"step": 224
},
{
"epoch": 0.48136315228966986,
"grad_norm": 0.486328125,
"learning_rate": 1.9483452722305705e-05,
"loss": 1.206823468208313,
"step": 226
},
{
"epoch": 0.48562300319488816,
"grad_norm": 1.1484375,
"learning_rate": 1.9470099852876284e-05,
"loss": 1.1781184673309326,
"step": 228
},
{
"epoch": 0.4898828541001065,
"grad_norm": 1.2265625,
"learning_rate": 1.9456582499244607e-05,
"loss": 1.1779451370239258,
"step": 230
},
{
"epoch": 0.4941427050053248,
"grad_norm": 0.51953125,
"learning_rate": 1.9442900959049402e-05,
"loss": 1.2148332595825195,
"step": 232
},
{
"epoch": 0.4984025559105431,
"grad_norm": 0.9453125,
"learning_rate": 1.942905553354465e-05,
"loss": 1.2204636335372925,
"step": 234
},
{
"epoch": 0.5026624068157615,
"grad_norm": 0.5859375,
"learning_rate": 1.9415046527592905e-05,
"loss": 1.19704008102417,
"step": 236
},
{
"epoch": 0.5069222577209798,
"grad_norm": 0.6796875,
"learning_rate": 1.9400874249658606e-05,
"loss": 1.230122685432434,
"step": 238
},
{
"epoch": 0.5111821086261981,
"grad_norm": 1.015625,
"learning_rate": 1.9386539011801288e-05,
"loss": 1.2034316062927246,
"step": 240
},
{
"epoch": 0.5154419595314164,
"grad_norm": 0.6484375,
"learning_rate": 1.9372041129668688e-05,
"loss": 1.310076355934143,
"step": 242
},
{
"epoch": 0.5197018104366348,
"grad_norm": 0.64453125,
"learning_rate": 1.935738092248981e-05,
"loss": 1.2447279691696167,
"step": 244
},
{
"epoch": 0.5239616613418531,
"grad_norm": 0.72265625,
"learning_rate": 1.9342558713067916e-05,
"loss": 1.249355673789978,
"step": 246
},
{
"epoch": 0.5282215122470714,
"grad_norm": 0.54296875,
"learning_rate": 1.932757482777337e-05,
"loss": 1.2074915170669556,
"step": 248
},
{
"epoch": 0.5324813631522897,
"grad_norm": 0.87109375,
"learning_rate": 1.9312429596536493e-05,
"loss": 1.2255703210830688,
"step": 250
},
{
"epoch": 0.536741214057508,
"grad_norm": 0.9609375,
"learning_rate": 1.929712335284028e-05,
"loss": 1.174513339996338,
"step": 252
},
{
"epoch": 0.5410010649627263,
"grad_norm": 1.2265625,
"learning_rate": 1.9281656433713065e-05,
"loss": 1.186693787574768,
"step": 254
},
{
"epoch": 0.5452609158679447,
"grad_norm": 0.84765625,
"learning_rate": 1.9266029179721087e-05,
"loss": 1.1817073822021484,
"step": 256
},
{
"epoch": 0.549520766773163,
"grad_norm": 0.4453125,
"learning_rate": 1.925024193496101e-05,
"loss": 1.2056523561477661,
"step": 258
},
{
"epoch": 0.5537806176783813,
"grad_norm": 0.87109375,
"learning_rate": 1.923429504705233e-05,
"loss": 1.2128726243972778,
"step": 260
},
{
"epoch": 0.5580404685835996,
"grad_norm": 0.55078125,
"learning_rate": 1.921818886712973e-05,
"loss": 1.2569330930709839,
"step": 262
},
{
"epoch": 0.5623003194888179,
"grad_norm": 1.1484375,
"learning_rate": 1.9201923749835343e-05,
"loss": 1.2348226308822632,
"step": 264
},
{
"epoch": 0.5665601703940362,
"grad_norm": 0.73828125,
"learning_rate": 1.9185500053310947e-05,
"loss": 1.2059990167617798,
"step": 266
},
{
"epoch": 0.5708200212992546,
"grad_norm": 0.5546875,
"learning_rate": 1.916891813919008e-05,
"loss": 1.2185040712356567,
"step": 268
},
{
"epoch": 0.5750798722044729,
"grad_norm": 0.90234375,
"learning_rate": 1.915217837259007e-05,
"loss": 1.1790536642074585,
"step": 270
},
{
"epoch": 0.5793397231096912,
"grad_norm": 0.72265625,
"learning_rate": 1.9135281122104e-05,
"loss": 1.2463805675506592,
"step": 272
},
{
"epoch": 0.5835995740149095,
"grad_norm": 0.60546875,
"learning_rate": 1.9118226759792606e-05,
"loss": 1.2532294988632202,
"step": 274
},
{
"epoch": 0.5878594249201278,
"grad_norm": 1.1171875,
"learning_rate": 1.910101566117605e-05,
"loss": 1.257046103477478,
"step": 276
},
{
"epoch": 0.5921192758253461,
"grad_norm": 0.57421875,
"learning_rate": 1.908364820522568e-05,
"loss": 1.2225136756896973,
"step": 278
},
{
"epoch": 0.5963791267305645,
"grad_norm": 0.6328125,
"learning_rate": 1.906612477435569e-05,
"loss": 1.1029458045959473,
"step": 280
},
{
"epoch": 0.6006389776357828,
"grad_norm": 0.65625,
"learning_rate": 1.9048445754414667e-05,
"loss": 1.2707245349884033,
"step": 282
},
{
"epoch": 0.6048988285410011,
"grad_norm": 0.72265625,
"learning_rate": 1.903061153467713e-05,
"loss": 1.2144739627838135,
"step": 284
},
{
"epoch": 0.6091586794462194,
"grad_norm": 0.52734375,
"learning_rate": 1.9012622507834943e-05,
"loss": 1.267938494682312,
"step": 286
},
{
"epoch": 0.6134185303514377,
"grad_norm": 0.828125,
"learning_rate": 1.8994479069988654e-05,
"loss": 1.158606767654419,
"step": 288
},
{
"epoch": 0.617678381256656,
"grad_norm": 0.96484375,
"learning_rate": 1.897618162063881e-05,
"loss": 1.3172389268875122,
"step": 290
},
{
"epoch": 0.6219382321618744,
"grad_norm": 0.55859375,
"learning_rate": 1.895773056267712e-05,
"loss": 1.1937861442565918,
"step": 292
},
{
"epoch": 0.6261980830670927,
"grad_norm": 0.484375,
"learning_rate": 1.893912630237762e-05,
"loss": 1.2631405591964722,
"step": 294
},
{
"epoch": 0.630457933972311,
"grad_norm": 0.67578125,
"learning_rate": 1.8920369249387697e-05,
"loss": 1.2060716152191162,
"step": 296
},
{
"epoch": 0.6347177848775293,
"grad_norm": 0.78125,
"learning_rate": 1.8901459816719085e-05,
"loss": 1.1968940496444702,
"step": 298
},
{
"epoch": 0.6389776357827476,
"grad_norm": 0.6953125,
"learning_rate": 1.8882398420738766e-05,
"loss": 1.2191784381866455,
"step": 300
},
{
"epoch": 0.6432374866879659,
"grad_norm": 0.4765625,
"learning_rate": 1.8863185481159815e-05,
"loss": 1.1612508296966553,
"step": 302
},
{
"epoch": 0.6474973375931843,
"grad_norm": 0.84765625,
"learning_rate": 1.8843821421032124e-05,
"loss": 1.2527700662612915,
"step": 304
},
{
"epoch": 0.6517571884984026,
"grad_norm": 2.484375,
"learning_rate": 1.8824306666733133e-05,
"loss": 1.2492940425872803,
"step": 306
},
{
"epoch": 0.6560170394036209,
"grad_norm": 0.91796875,
"learning_rate": 1.880464164795841e-05,
"loss": 1.2013378143310547,
"step": 308
},
{
"epoch": 0.6602768903088392,
"grad_norm": 0.6015625,
"learning_rate": 1.8784826797712195e-05,
"loss": 1.1568100452423096,
"step": 310
},
{
"epoch": 0.6645367412140575,
"grad_norm": 3.8125,
"learning_rate": 1.8764862552297865e-05,
"loss": 1.2038320302963257,
"step": 312
},
{
"epoch": 0.6687965921192758,
"grad_norm": 0.5234375,
"learning_rate": 1.874474935130835e-05,
"loss": 1.1380341053009033,
"step": 314
},
{
"epoch": 0.6730564430244942,
"grad_norm": 0.578125,
"learning_rate": 1.8724487637616405e-05,
"loss": 1.1762800216674805,
"step": 316
},
{
"epoch": 0.6773162939297125,
"grad_norm": 0.69921875,
"learning_rate": 1.8704077857364913e-05,
"loss": 1.2073653936386108,
"step": 318
},
{
"epoch": 0.6815761448349308,
"grad_norm": 0.76953125,
"learning_rate": 1.8683520459957027e-05,
"loss": 1.2313804626464844,
"step": 320
},
{
"epoch": 0.6858359957401491,
"grad_norm": 0.53125,
"learning_rate": 1.8662815898046277e-05,
"loss": 1.1902257204055786,
"step": 322
},
{
"epoch": 0.6900958466453674,
"grad_norm": 0.7578125,
"learning_rate": 1.8641964627526623e-05,
"loss": 1.177359938621521,
"step": 324
},
{
"epoch": 0.6943556975505857,
"grad_norm": 0.640625,
"learning_rate": 1.8620967107522393e-05,
"loss": 1.1724225282669067,
"step": 326
},
{
"epoch": 0.6986155484558041,
"grad_norm": 0.8203125,
"learning_rate": 1.8599823800378183e-05,
"loss": 1.2089399099349976,
"step": 328
},
{
"epoch": 0.7028753993610224,
"grad_norm": 0.90234375,
"learning_rate": 1.8578535171648683e-05,
"loss": 1.1843369007110596,
"step": 330
},
{
"epoch": 0.7071352502662407,
"grad_norm": 0.625,
"learning_rate": 1.8557101690088423e-05,
"loss": 1.2020741701126099,
"step": 332
},
{
"epoch": 0.711395101171459,
"grad_norm": 0.578125,
"learning_rate": 1.853552382764144e-05,
"loss": 1.2178312540054321,
"step": 334
},
{
"epoch": 0.7156549520766773,
"grad_norm": 0.5625,
"learning_rate": 1.8513802059430906e-05,
"loss": 1.3234312534332275,
"step": 336
},
{
"epoch": 0.7199148029818956,
"grad_norm": 0.65625,
"learning_rate": 1.8491936863748638e-05,
"loss": 1.2087228298187256,
"step": 338
},
{
"epoch": 0.724174653887114,
"grad_norm": 0.6171875,
"learning_rate": 1.8469928722044607e-05,
"loss": 1.213451862335205,
"step": 340
},
{
"epoch": 0.7284345047923323,
"grad_norm": 0.58984375,
"learning_rate": 1.8447778118916297e-05,
"loss": 1.2303829193115234,
"step": 342
},
{
"epoch": 0.7326943556975506,
"grad_norm": 0.65625,
"learning_rate": 1.8425485542098055e-05,
"loss": 1.226832389831543,
"step": 344
},
{
"epoch": 0.7369542066027689,
"grad_norm": 0.94921875,
"learning_rate": 1.8403051482450352e-05,
"loss": 1.1740474700927734,
"step": 346
},
{
"epoch": 0.7412140575079872,
"grad_norm": 0.44921875,
"learning_rate": 1.8380476433948967e-05,
"loss": 1.234078288078308,
"step": 348
},
{
"epoch": 0.7454739084132055,
"grad_norm": 0.62109375,
"learning_rate": 1.8357760893674118e-05,
"loss": 1.1722126007080078,
"step": 350
},
{
"epoch": 0.7497337593184239,
"grad_norm": 0.455078125,
"learning_rate": 1.8334905361799505e-05,
"loss": 1.1766189336776733,
"step": 352
},
{
"epoch": 0.7539936102236422,
"grad_norm": 0.734375,
"learning_rate": 1.8311910341581317e-05,
"loss": 1.2379800081253052,
"step": 354
},
{
"epoch": 0.7582534611288605,
"grad_norm": 1.9375,
"learning_rate": 1.8288776339347123e-05,
"loss": 1.3067048788070679,
"step": 356
},
{
"epoch": 0.7625133120340788,
"grad_norm": 0.984375,
"learning_rate": 1.826550386448475e-05,
"loss": 1.249035120010376,
"step": 358
},
{
"epoch": 0.7667731629392971,
"grad_norm": 0.609375,
"learning_rate": 1.8242093429431054e-05,
"loss": 1.153590440750122,
"step": 360
},
{
"epoch": 0.7710330138445154,
"grad_norm": 1.515625,
"learning_rate": 1.8218545549660627e-05,
"loss": 1.2533551454544067,
"step": 362
},
{
"epoch": 0.7752928647497338,
"grad_norm": 0.9375,
"learning_rate": 1.8194860743674475e-05,
"loss": 1.2324496507644653,
"step": 364
},
{
"epoch": 0.7795527156549521,
"grad_norm": 0.5078125,
"learning_rate": 1.817103953298857e-05,
"loss": 1.1727190017700195,
"step": 366
},
{
"epoch": 0.7838125665601704,
"grad_norm": 0.435546875,
"learning_rate": 1.814708244212238e-05,
"loss": 1.202933430671692,
"step": 368
},
{
"epoch": 0.7880724174653887,
"grad_norm": 0.70703125,
"learning_rate": 1.8122989998587334e-05,
"loss": 1.1554442644119263,
"step": 370
},
{
"epoch": 0.792332268370607,
"grad_norm": 0.38671875,
"learning_rate": 1.8098762732875176e-05,
"loss": 1.2414895296096802,
"step": 372
},
{
"epoch": 0.7965921192758253,
"grad_norm": 0.70703125,
"learning_rate": 1.807440117844632e-05,
"loss": 1.196638584136963,
"step": 374
},
{
"epoch": 0.8008519701810437,
"grad_norm": 0.625,
"learning_rate": 1.8049905871718057e-05,
"loss": 1.1261889934539795,
"step": 376
},
{
"epoch": 0.805111821086262,
"grad_norm": 0.5703125,
"learning_rate": 1.80252773520528e-05,
"loss": 1.1604783535003662,
"step": 378
},
{
"epoch": 0.8093716719914803,
"grad_norm": 1.28125,
"learning_rate": 1.8000516161746158e-05,
"loss": 1.2182838916778564,
"step": 380
},
{
"epoch": 0.8136315228966986,
"grad_norm": 0.60546875,
"learning_rate": 1.7975622846015025e-05,
"loss": 1.118109941482544,
"step": 382
},
{
"epoch": 0.8178913738019169,
"grad_norm": 0.734375,
"learning_rate": 1.795059795298557e-05,
"loss": 1.171169638633728,
"step": 384
},
{
"epoch": 0.8221512247071352,
"grad_norm": 0.6640625,
"learning_rate": 1.792544203368116e-05,
"loss": 1.2411582469940186,
"step": 386
},
{
"epoch": 0.8264110756123536,
"grad_norm": 0.9140625,
"learning_rate": 1.7900155642010224e-05,
"loss": 1.2089407444000244,
"step": 388
},
{
"epoch": 0.8306709265175719,
"grad_norm": 0.5078125,
"learning_rate": 1.7874739334754072e-05,
"loss": 1.1152267456054688,
"step": 390
},
{
"epoch": 0.8349307774227902,
"grad_norm": 0.625,
"learning_rate": 1.7849193671554625e-05,
"loss": 1.1154651641845703,
"step": 392
},
{
"epoch": 0.8391906283280085,
"grad_norm": 0.458984375,
"learning_rate": 1.7823519214902093e-05,
"loss": 1.1766343116760254,
"step": 394
},
{
"epoch": 0.8434504792332268,
"grad_norm": 0.56640625,
"learning_rate": 1.7797716530122588e-05,
"loss": 1.061235785484314,
"step": 396
},
{
"epoch": 0.8477103301384451,
"grad_norm": 0.54296875,
"learning_rate": 1.777178618536568e-05,
"loss": 1.1713075637817383,
"step": 398
},
{
"epoch": 0.8519701810436635,
"grad_norm": 0.5,
"learning_rate": 1.774572875159189e-05,
"loss": 1.1845115423202515,
"step": 400
},
{
"epoch": 0.8562300319488818,
"grad_norm": 0.51953125,
"learning_rate": 1.77195448025601e-05,
"loss": 1.1638309955596924,
"step": 402
},
{
"epoch": 0.8604898828541001,
"grad_norm": 0.6875,
"learning_rate": 1.7693234914814955e-05,
"loss": 1.1716026067733765,
"step": 404
},
{
"epoch": 0.8647497337593184,
"grad_norm": 0.80078125,
"learning_rate": 1.7666799667674123e-05,
"loss": 1.2654908895492554,
"step": 406
},
{
"epoch": 0.8690095846645367,
"grad_norm": 1.140625,
"learning_rate": 1.7640239643215574e-05,
"loss": 1.219713568687439,
"step": 408
},
{
"epoch": 0.873269435569755,
"grad_norm": 1.15625,
"learning_rate": 1.761355542626475e-05,
"loss": 1.3021862506866455,
"step": 410
},
{
"epoch": 0.8775292864749734,
"grad_norm": 0.52734375,
"learning_rate": 1.758674760438168e-05,
"loss": 1.2263654470443726,
"step": 412
},
{
"epoch": 0.8817891373801917,
"grad_norm": 0.75390625,
"learning_rate": 1.7559816767848063e-05,
"loss": 1.212475061416626,
"step": 414
},
{
"epoch": 0.88604898828541,
"grad_norm": 1.0859375,
"learning_rate": 1.7532763509654257e-05,
"loss": 1.1879194974899292,
"step": 416
},
{
"epoch": 0.8903088391906283,
"grad_norm": 0.53125,
"learning_rate": 1.7505588425486216e-05,
"loss": 1.1831648349761963,
"step": 418
},
{
"epoch": 0.8945686900958466,
"grad_norm": 0.75,
"learning_rate": 1.7478292113712387e-05,
"loss": 1.2148785591125488,
"step": 420
},
{
"epoch": 0.898828541001065,
"grad_norm": 0.48046875,
"learning_rate": 1.7450875175370532e-05,
"loss": 1.1164610385894775,
"step": 422
},
{
"epoch": 0.9030883919062833,
"grad_norm": 0.478515625,
"learning_rate": 1.7423338214154476e-05,
"loss": 1.1142120361328125,
"step": 424
},
{
"epoch": 0.9073482428115016,
"grad_norm": 0.69921875,
"learning_rate": 1.7395681836400856e-05,
"loss": 1.1910451650619507,
"step": 426
},
{
"epoch": 0.9116080937167199,
"grad_norm": 0.828125,
"learning_rate": 1.736790665107571e-05,
"loss": 1.1739782094955444,
"step": 428
},
{
"epoch": 0.9158679446219382,
"grad_norm": 0.431640625,
"learning_rate": 1.7340013269761127e-05,
"loss": 1.156859278678894,
"step": 430
},
{
"epoch": 0.9201277955271565,
"grad_norm": 0.46875,
"learning_rate": 1.731200230664174e-05,
"loss": 1.1874017715454102,
"step": 432
},
{
"epoch": 0.9243876464323749,
"grad_norm": 0.376953125,
"learning_rate": 1.7283874378491226e-05,
"loss": 1.1091042757034302,
"step": 434
},
{
"epoch": 0.9286474973375932,
"grad_norm": 0.546875,
"learning_rate": 1.7255630104658707e-05,
"loss": 1.2352137565612793,
"step": 436
},
{
"epoch": 0.9329073482428115,
"grad_norm": 0.55859375,
"learning_rate": 1.7227270107055134e-05,
"loss": 1.1861075162887573,
"step": 438
},
{
"epoch": 0.9371671991480298,
"grad_norm": 0.84375,
"learning_rate": 1.7198795010139565e-05,
"loss": 1.1825189590454102,
"step": 440
},
{
"epoch": 0.9414270500532481,
"grad_norm": 0.51953125,
"learning_rate": 1.7170205440905444e-05,
"loss": 1.1355581283569336,
"step": 442
},
{
"epoch": 0.9456869009584664,
"grad_norm": 0.578125,
"learning_rate": 1.7141502028866782e-05,
"loss": 1.1689903736114502,
"step": 444
},
{
"epoch": 0.9499467518636848,
"grad_norm": 0.66015625,
"learning_rate": 1.7112685406044283e-05,
"loss": 1.1596486568450928,
"step": 446
},
{
"epoch": 0.9542066027689031,
"grad_norm": 0.62109375,
"learning_rate": 1.7083756206951457e-05,
"loss": 1.2047994136810303,
"step": 448
},
{
"epoch": 0.9584664536741214,
"grad_norm": 0.9140625,
"learning_rate": 1.705471506858062e-05,
"loss": 1.1618238687515259,
"step": 450
},
{
"epoch": 0.9627263045793397,
"grad_norm": 0.70703125,
"learning_rate": 1.702556263038888e-05,
"loss": 1.158185601234436,
"step": 452
},
{
"epoch": 0.966986155484558,
"grad_norm": 0.41015625,
"learning_rate": 1.699629953428407e-05,
"loss": 1.182377576828003,
"step": 454
},
{
"epoch": 0.9712460063897763,
"grad_norm": 0.609375,
"learning_rate": 1.696692642461058e-05,
"loss": 1.1322381496429443,
"step": 456
},
{
"epoch": 0.9755058572949947,
"grad_norm": 0.58203125,
"learning_rate": 1.693744394813521e-05,
"loss": 1.1632483005523682,
"step": 458
},
{
"epoch": 0.979765708200213,
"grad_norm": 0.5859375,
"learning_rate": 1.6907852754032897e-05,
"loss": 1.1526472568511963,
"step": 460
},
{
"epoch": 0.9840255591054313,
"grad_norm": 0.68359375,
"learning_rate": 1.6878153493872438e-05,
"loss": 1.212873101234436,
"step": 462
},
{
"epoch": 0.9882854100106496,
"grad_norm": 0.46484375,
"learning_rate": 1.684834682160213e-05,
"loss": 1.1329699754714966,
"step": 464
},
{
"epoch": 0.9925452609158679,
"grad_norm": 0.56640625,
"learning_rate": 1.681843339353539e-05,
"loss": 1.2156919240951538,
"step": 466
},
{
"epoch": 0.9968051118210862,
"grad_norm": 0.515625,
"learning_rate": 1.678841386833628e-05,
"loss": 1.1542456150054932,
"step": 468
},
{
"epoch": 1.0,
"grad_norm": 0.6328125,
"learning_rate": 1.675828890700502e-05,
"loss": 1.2353312969207764,
"step": 470
},
{
"epoch": 1.0042598509052183,
"grad_norm": 0.4296875,
"learning_rate": 1.6728059172863434e-05,
"loss": 1.0001182556152344,
"step": 472
},
{
"epoch": 1.0085197018104366,
"grad_norm": 0.447265625,
"learning_rate": 1.6697725331540337e-05,
"loss": 1.0624783039093018,
"step": 474
},
{
"epoch": 1.012779552715655,
"grad_norm": 0.458984375,
"learning_rate": 1.666728805095688e-05,
"loss": 1.0327106714248657,
"step": 476
},
{
"epoch": 1.0170394036208732,
"grad_norm": 0.609375,
"learning_rate": 1.663674800131184e-05,
"loss": 0.9451117515563965,
"step": 478
},
{
"epoch": 1.0212992545260915,
"grad_norm": 0.515625,
"learning_rate": 1.660610585506687e-05,
"loss": 1.0440083742141724,
"step": 480
},
{
"epoch": 1.0255591054313098,
"grad_norm": 0.4921875,
"learning_rate": 1.65753622869317e-05,
"loss": 1.0152018070220947,
"step": 482
},
{
"epoch": 1.0298189563365283,
"grad_norm": 3.34375,
"learning_rate": 1.6544517973849247e-05,
"loss": 1.005466341972351,
"step": 484
},
{
"epoch": 1.0340788072417466,
"grad_norm": 0.50390625,
"learning_rate": 1.6513573594980753e-05,
"loss": 1.0366133451461792,
"step": 486
},
{
"epoch": 1.038338658146965,
"grad_norm": 1.171875,
"learning_rate": 1.6482529831690805e-05,
"loss": 1.0141640901565552,
"step": 488
},
{
"epoch": 1.0425985090521832,
"grad_norm": 0.54296875,
"learning_rate": 1.6451387367532316e-05,
"loss": 0.9699265956878662,
"step": 490
},
{
"epoch": 1.0468583599574015,
"grad_norm": 0.546875,
"learning_rate": 1.642014688823153e-05,
"loss": 1.0510493516921997,
"step": 492
},
{
"epoch": 1.0511182108626198,
"grad_norm": 0.490234375,
"learning_rate": 1.6388809081672854e-05,
"loss": 0.9604178071022034,
"step": 494
},
{
"epoch": 1.055378061767838,
"grad_norm": 0.59375,
"learning_rate": 1.635737463788377e-05,
"loss": 1.0432132482528687,
"step": 496
},
{
"epoch": 1.0596379126730564,
"grad_norm": 0.921875,
"learning_rate": 1.63258442490196e-05,
"loss": 0.9879993796348572,
"step": 498
},
{
"epoch": 1.0638977635782747,
"grad_norm": 1.8671875,
"learning_rate": 1.6294218609348292e-05,
"loss": 0.989188015460968,
"step": 500
},
{
"epoch": 1.068157614483493,
"grad_norm": 0.490234375,
"learning_rate": 1.626249841523511e-05,
"loss": 1.0020525455474854,
"step": 502
},
{
"epoch": 1.0724174653887113,
"grad_norm": 0.6328125,
"learning_rate": 1.623068436512732e-05,
"loss": 1.0052525997161865,
"step": 504
},
{
"epoch": 1.0766773162939298,
"grad_norm": 0.50390625,
"learning_rate": 1.619877715953882e-05,
"loss": 1.030534029006958,
"step": 506
},
{
"epoch": 1.0809371671991481,
"grad_norm": 0.53515625,
"learning_rate": 1.616677750103467e-05,
"loss": 1.1060301065444946,
"step": 508
},
{
"epoch": 1.0851970181043664,
"grad_norm": 1.5703125,
"learning_rate": 1.6134686094215663e-05,
"loss": 0.976584255695343,
"step": 510
},
{
"epoch": 1.0894568690095847,
"grad_norm": 0.46484375,
"learning_rate": 1.6102503645702813e-05,
"loss": 0.9827842712402344,
"step": 512
},
{
"epoch": 1.093716719914803,
"grad_norm": 0.5703125,
"learning_rate": 1.6070230864121765e-05,
"loss": 0.9941512942314148,
"step": 514
},
{
"epoch": 1.0979765708200213,
"grad_norm": 0.640625,
"learning_rate": 1.603786846008722e-05,
"loss": 0.9916704893112183,
"step": 516
},
{
"epoch": 1.1022364217252396,
"grad_norm": 0.7734375,
"learning_rate": 1.6005417146187262e-05,
"loss": 0.9758402109146118,
"step": 518
},
{
"epoch": 1.106496272630458,
"grad_norm": 0.546875,
"learning_rate": 1.59728776369677e-05,
"loss": 1.0404748916625977,
"step": 520
},
{
"epoch": 1.1107561235356762,
"grad_norm": 0.63671875,
"learning_rate": 1.5940250648916305e-05,
"loss": 1.0453121662139893,
"step": 522
},
{
"epoch": 1.1150159744408945,
"grad_norm": 0.796875,
"learning_rate": 1.5907536900447046e-05,
"loss": 0.9855365753173828,
"step": 524
},
{
"epoch": 1.1192758253461128,
"grad_norm": 0.76953125,
"learning_rate": 1.587473711188428e-05,
"loss": 0.9536042213439941,
"step": 526
},
{
"epoch": 1.123535676251331,
"grad_norm": 0.50390625,
"learning_rate": 1.5841852005446866e-05,
"loss": 1.0321474075317383,
"step": 528
},
{
"epoch": 1.1277955271565494,
"grad_norm": 0.5625,
"learning_rate": 1.5808882305232284e-05,
"loss": 0.9870891571044922,
"step": 530
},
{
"epoch": 1.132055378061768,
"grad_norm": 0.6484375,
"learning_rate": 1.5775828737200708e-05,
"loss": 1.0295721292495728,
"step": 532
},
{
"epoch": 1.1363152289669862,
"grad_norm": 0.439453125,
"learning_rate": 1.5742692029158956e-05,
"loss": 0.9844380021095276,
"step": 534
},
{
"epoch": 1.1405750798722045,
"grad_norm": 0.53125,
"learning_rate": 1.5709472910744535e-05,
"loss": 1.0327517986297607,
"step": 536
},
{
"epoch": 1.1448349307774228,
"grad_norm": 0.54296875,
"learning_rate": 1.567617211340955e-05,
"loss": 1.015358328819275,
"step": 538
},
{
"epoch": 1.1490947816826411,
"grad_norm": 0.5390625,
"learning_rate": 1.5642790370404572e-05,
"loss": 0.9756560325622559,
"step": 540
},
{
"epoch": 1.1533546325878594,
"grad_norm": 0.62890625,
"learning_rate": 1.5609328416762538e-05,
"loss": 0.9365598559379578,
"step": 542
},
{
"epoch": 1.1576144834930777,
"grad_norm": 8.3125,
"learning_rate": 1.557578698928253e-05,
"loss": 1.0103881359100342,
"step": 544
},
{
"epoch": 1.161874334398296,
"grad_norm": 1.25,
"learning_rate": 1.5542166826513576e-05,
"loss": 0.972273051738739,
"step": 546
},
{
"epoch": 1.1661341853035143,
"grad_norm": 0.90625,
"learning_rate": 1.550846866873837e-05,
"loss": 1.016709804534912,
"step": 548
},
{
"epoch": 1.1703940362087326,
"grad_norm": 0.55859375,
"learning_rate": 1.5474693257956976e-05,
"loss": 1.0085346698760986,
"step": 550
},
{
"epoch": 1.174653887113951,
"grad_norm": 1.328125,
"learning_rate": 1.5440841337870497e-05,
"loss": 0.9733573794364929,
"step": 552
},
{
"epoch": 1.1789137380191694,
"grad_norm": 0.64453125,
"learning_rate": 1.54069136538647e-05,
"loss": 1.0755430459976196,
"step": 554
},
{
"epoch": 1.1831735889243877,
"grad_norm": 0.640625,
"learning_rate": 1.5372910952993588e-05,
"loss": 1.1009080410003662,
"step": 556
},
{
"epoch": 1.187433439829606,
"grad_norm": 0.953125,
"learning_rate": 1.5338833983962972e-05,
"loss": 0.9900831580162048,
"step": 558
},
{
"epoch": 1.1916932907348243,
"grad_norm": 0.466796875,
"learning_rate": 1.5304683497113963e-05,
"loss": 1.0196163654327393,
"step": 560
},
{
"epoch": 1.1959531416400426,
"grad_norm": 0.73046875,
"learning_rate": 1.5270460244406467e-05,
"loss": 1.012573480606079,
"step": 562
},
{
"epoch": 1.200212992545261,
"grad_norm": 0.5703125,
"learning_rate": 1.523616497940263e-05,
"loss": 0.9771086573600769,
"step": 564
},
{
"epoch": 1.2044728434504792,
"grad_norm": 0.45703125,
"learning_rate": 1.5201798457250227e-05,
"loss": 1.048552393913269,
"step": 566
},
{
"epoch": 1.2087326943556975,
"grad_norm": 0.4609375,
"learning_rate": 1.5167361434666044e-05,
"loss": 0.9992176294326782,
"step": 568
},
{
"epoch": 1.2129925452609158,
"grad_norm": 0.765625,
"learning_rate": 1.5132854669919223e-05,
"loss": 0.9978233575820923,
"step": 570
},
{
"epoch": 1.2172523961661341,
"grad_norm": 0.42578125,
"learning_rate": 1.5098278922814555e-05,
"loss": 0.9533007740974426,
"step": 572
},
{
"epoch": 1.2215122470713524,
"grad_norm": 0.59375,
"learning_rate": 1.5063634954675758e-05,
"loss": 1.05226469039917,
"step": 574
},
{
"epoch": 1.225772097976571,
"grad_norm": 0.70703125,
"learning_rate": 1.5028923528328712e-05,
"loss": 1.0399305820465088,
"step": 576
},
{
"epoch": 1.230031948881789,
"grad_norm": 0.439453125,
"learning_rate": 1.499414540808465e-05,
"loss": 1.0016851425170898,
"step": 578
},
{
"epoch": 1.2342917997870075,
"grad_norm": 0.42578125,
"learning_rate": 1.4959301359723348e-05,
"loss": 0.9840012192726135,
"step": 580
},
{
"epoch": 1.2385516506922258,
"grad_norm": 0.4609375,
"learning_rate": 1.4924392150476247e-05,
"loss": 0.9944754838943481,
"step": 582
},
{
"epoch": 1.2428115015974441,
"grad_norm": 0.58984375,
"learning_rate": 1.488941854900957e-05,
"loss": 1.0165916681289673,
"step": 584
},
{
"epoch": 1.2470713525026624,
"grad_norm": 0.6484375,
"learning_rate": 1.4854381325407393e-05,
"loss": 1.0132522583007812,
"step": 586
},
{
"epoch": 1.2513312034078807,
"grad_norm": 0.51171875,
"learning_rate": 1.4819281251154686e-05,
"loss": 1.0050327777862549,
"step": 588
},
{
"epoch": 1.255591054313099,
"grad_norm": 0.703125,
"learning_rate": 1.4784119099120338e-05,
"loss": 1.0251795053482056,
"step": 590
},
{
"epoch": 1.2598509052183173,
"grad_norm": 0.671875,
"learning_rate": 1.4748895643540109e-05,
"loss": 1.0751694440841675,
"step": 592
},
{
"epoch": 1.2641107561235356,
"grad_norm": 0.66015625,
"learning_rate": 1.4713611659999618e-05,
"loss": 0.9825355410575867,
"step": 594
},
{
"epoch": 1.268370607028754,
"grad_norm": 0.578125,
"learning_rate": 1.4678267925417254e-05,
"loss": 1.0707963705062866,
"step": 596
},
{
"epoch": 1.2726304579339724,
"grad_norm": 0.466796875,
"learning_rate": 1.4642865218027047e-05,
"loss": 0.9957266449928284,
"step": 598
},
{
"epoch": 1.2768903088391905,
"grad_norm": 1.7265625,
"learning_rate": 1.4607404317361566e-05,
"loss": 0.9615009427070618,
"step": 600
},
{
"epoch": 1.281150159744409,
"grad_norm": 0.515625,
"learning_rate": 1.4571886004234728e-05,
"loss": 0.9674661159515381,
"step": 602
},
{
"epoch": 1.2854100106496273,
"grad_norm": 0.484375,
"learning_rate": 1.4536311060724622e-05,
"loss": 1.0484395027160645,
"step": 604
},
{
"epoch": 1.2896698615548456,
"grad_norm": 0.5625,
"learning_rate": 1.4500680270156285e-05,
"loss": 1.0365606546401978,
"step": 606
},
{
"epoch": 1.293929712460064,
"grad_norm": 3.34375,
"learning_rate": 1.4464994417084434e-05,
"loss": 0.9990244507789612,
"step": 608
},
{
"epoch": 1.2981895633652822,
"grad_norm": 0.5234375,
"learning_rate": 1.442925428727623e-05,
"loss": 0.9851716756820679,
"step": 610
},
{
"epoch": 1.3024494142705005,
"grad_norm": 0.61328125,
"learning_rate": 1.4393460667693937e-05,
"loss": 1.04202139377594,
"step": 612
},
{
"epoch": 1.3067092651757188,
"grad_norm": 0.58203125,
"learning_rate": 1.4357614346477622e-05,
"loss": 1.027279019355774,
"step": 614
},
{
"epoch": 1.3109691160809371,
"grad_norm": 0.40234375,
"learning_rate": 1.4321716112927777e-05,
"loss": 0.987048864364624,
"step": 616
},
{
"epoch": 1.3152289669861554,
"grad_norm": 0.66015625,
"learning_rate": 1.4285766757487966e-05,
"loss": 0.9855493307113647,
"step": 618
},
{
"epoch": 1.3194888178913737,
"grad_norm": 0.68359375,
"learning_rate": 1.4249767071727406e-05,
"loss": 0.9904506802558899,
"step": 620
},
{
"epoch": 1.323748668796592,
"grad_norm": 0.50390625,
"learning_rate": 1.4213717848323523e-05,
"loss": 1.0218966007232666,
"step": 622
},
{
"epoch": 1.3280085197018106,
"grad_norm": 0.64453125,
"learning_rate": 1.4177619881044527e-05,
"loss": 1.0790345668792725,
"step": 624
},
{
"epoch": 1.3322683706070286,
"grad_norm": 0.478515625,
"learning_rate": 1.4141473964731915e-05,
"loss": 0.9817459583282471,
"step": 626
},
{
"epoch": 1.3365282215122471,
"grad_norm": 1.0859375,
"learning_rate": 1.4105280895282972e-05,
"loss": 0.9989323019981384,
"step": 628
},
{
"epoch": 1.3407880724174654,
"grad_norm": 0.61328125,
"learning_rate": 1.4069041469633251e-05,
"loss": 0.9893896579742432,
"step": 630
},
{
"epoch": 1.3450479233226837,
"grad_norm": 0.58203125,
"learning_rate": 1.4032756485739025e-05,
"loss": 1.0551786422729492,
"step": 632
},
{
"epoch": 1.349307774227902,
"grad_norm": 0.48828125,
"learning_rate": 1.3996426742559712e-05,
"loss": 0.9710611701011658,
"step": 634
},
{
"epoch": 1.3535676251331203,
"grad_norm": 0.546875,
"learning_rate": 1.3960053040040284e-05,
"loss": 0.9912348985671997,
"step": 636
},
{
"epoch": 1.3578274760383386,
"grad_norm": 0.6171875,
"learning_rate": 1.3923636179093653e-05,
"loss": 0.9898839592933655,
"step": 638
},
{
"epoch": 1.362087326943557,
"grad_norm": 0.6171875,
"learning_rate": 1.3887176961583043e-05,
"loss": 0.9827688932418823,
"step": 640
},
{
"epoch": 1.3663471778487752,
"grad_norm": 0.478515625,
"learning_rate": 1.3850676190304322e-05,
"loss": 0.9909817576408386,
"step": 642
},
{
"epoch": 1.3706070287539935,
"grad_norm": 1.1796875,
"learning_rate": 1.3814134668968331e-05,
"loss": 1.054543375968933,
"step": 644
},
{
"epoch": 1.374866879659212,
"grad_norm": 0.6015625,
"learning_rate": 1.3777553202183186e-05,
"loss": 1.0329318046569824,
"step": 646
},
{
"epoch": 1.3791267305644301,
"grad_norm": 0.56640625,
"learning_rate": 1.3740932595436574e-05,
"loss": 0.9633697867393494,
"step": 648
},
{
"epoch": 1.3833865814696487,
"grad_norm": 0.48046875,
"learning_rate": 1.3704273655077986e-05,
"loss": 0.9963237643241882,
"step": 650
},
{
"epoch": 1.387646432374867,
"grad_norm": 0.486328125,
"learning_rate": 1.3667577188300994e-05,
"loss": 0.9815477728843689,
"step": 652
},
{
"epoch": 1.3919062832800853,
"grad_norm": 0.69921875,
"learning_rate": 1.3630844003125466e-05,
"loss": 0.9907623529434204,
"step": 654
},
{
"epoch": 1.3961661341853036,
"grad_norm": 0.76171875,
"learning_rate": 1.3594074908379759e-05,
"loss": 1.0208251476287842,
"step": 656
},
{
"epoch": 1.4004259850905219,
"grad_norm": 0.5859375,
"learning_rate": 1.3557270713682944e-05,
"loss": 1.0356820821762085,
"step": 658
},
{
"epoch": 1.4046858359957402,
"grad_norm": 0.4453125,
"learning_rate": 1.3520432229426933e-05,
"loss": 0.9749876260757446,
"step": 660
},
{
"epoch": 1.4089456869009584,
"grad_norm": 0.486328125,
"learning_rate": 1.3483560266758688e-05,
"loss": 0.9936063885688782,
"step": 662
},
{
"epoch": 1.4132055378061767,
"grad_norm": 0.5859375,
"learning_rate": 1.344665563756231e-05,
"loss": 0.9732779264450073,
"step": 664
},
{
"epoch": 1.417465388711395,
"grad_norm": 0.482421875,
"learning_rate": 1.3409719154441194e-05,
"loss": 1.0013114213943481,
"step": 666
},
{
"epoch": 1.4217252396166133,
"grad_norm": 0.78125,
"learning_rate": 1.3372751630700127e-05,
"loss": 0.9554456472396851,
"step": 668
},
{
"epoch": 1.4259850905218316,
"grad_norm": 0.5546875,
"learning_rate": 1.3335753880327378e-05,
"loss": 1.013936161994934,
"step": 670
},
{
"epoch": 1.4302449414270502,
"grad_norm": 0.5078125,
"learning_rate": 1.3298726717976776e-05,
"loss": 0.9992549419403076,
"step": 672
},
{
"epoch": 1.4345047923322682,
"grad_norm": 0.546875,
"learning_rate": 1.3261670958949778e-05,
"loss": 0.9462178945541382,
"step": 674
},
{
"epoch": 1.4387646432374868,
"grad_norm": 0.486328125,
"learning_rate": 1.32245874191775e-05,
"loss": 1.0281662940979004,
"step": 676
},
{
"epoch": 1.443024494142705,
"grad_norm": 0.79296875,
"learning_rate": 1.3187476915202772e-05,
"loss": 1.080543875694275,
"step": 678
},
{
"epoch": 1.4472843450479234,
"grad_norm": 0.80078125,
"learning_rate": 1.315034026416215e-05,
"loss": 0.9739678502082825,
"step": 680
},
{
"epoch": 1.4515441959531417,
"grad_norm": 0.58984375,
"learning_rate": 1.3113178283767916e-05,
"loss": 1.058993935585022,
"step": 682
},
{
"epoch": 1.45580404685836,
"grad_norm": 0.515625,
"learning_rate": 1.3075991792290087e-05,
"loss": 0.9686048626899719,
"step": 684
},
{
"epoch": 1.4600638977635783,
"grad_norm": 0.494140625,
"learning_rate": 1.303878160853838e-05,
"loss": 1.0147266387939453,
"step": 686
},
{
"epoch": 1.4643237486687966,
"grad_norm": 0.5234375,
"learning_rate": 1.3001548551844198e-05,
"loss": 0.9957408905029297,
"step": 688
},
{
"epoch": 1.4685835995740149,
"grad_norm": 0.64453125,
"learning_rate": 1.2964293442042583e-05,
"loss": 1.0241044759750366,
"step": 690
},
{
"epoch": 1.4728434504792332,
"grad_norm": 0.6484375,
"learning_rate": 1.2927017099454174e-05,
"loss": 0.9795851707458496,
"step": 692
},
{
"epoch": 1.4771033013844517,
"grad_norm": 0.5390625,
"learning_rate": 1.288972034486711e-05,
"loss": 1.0294287204742432,
"step": 694
},
{
"epoch": 1.4813631522896697,
"grad_norm": 0.45703125,
"learning_rate": 1.2852403999519009e-05,
"loss": 1.04385507106781,
"step": 696
},
{
"epoch": 1.4856230031948883,
"grad_norm": 0.462890625,
"learning_rate": 1.2815068885078848e-05,
"loss": 1.0272358655929565,
"step": 698
},
{
"epoch": 1.4898828541001066,
"grad_norm": 0.5625,
"learning_rate": 1.2777715823628869e-05,
"loss": 1.0158170461654663,
"step": 700
},
{
"epoch": 1.4941427050053249,
"grad_norm": 3.25,
"learning_rate": 1.274034563764651e-05,
"loss": 1.0180643796920776,
"step": 702
},
{
"epoch": 1.4984025559105432,
"grad_norm": 0.55859375,
"learning_rate": 1.2702959149986263e-05,
"loss": 0.9400717616081238,
"step": 704
},
{
"epoch": 1.5026624068157615,
"grad_norm": 1.46875,
"learning_rate": 1.2665557183861563e-05,
"loss": 1.0463502407073975,
"step": 706
},
{
"epoch": 1.5069222577209798,
"grad_norm": 0.4453125,
"learning_rate": 1.2628140562826674e-05,
"loss": 0.9682408571243286,
"step": 708
},
{
"epoch": 1.511182108626198,
"grad_norm": 2.0,
"learning_rate": 1.2590710110758542e-05,
"loss": 0.9887672066688538,
"step": 710
},
{
"epoch": 1.5154419595314164,
"grad_norm": 0.65234375,
"learning_rate": 1.2553266651838664e-05,
"loss": 0.9790204167366028,
"step": 712
},
{
"epoch": 1.5197018104366347,
"grad_norm": 0.55078125,
"learning_rate": 1.2515811010534926e-05,
"loss": 1.048098087310791,
"step": 714
},
{
"epoch": 1.5239616613418532,
"grad_norm": 0.91796875,
"learning_rate": 1.2478344011583463e-05,
"loss": 0.9795944094657898,
"step": 716
},
{
"epoch": 1.5282215122470713,
"grad_norm": 0.60546875,
"learning_rate": 1.2440866479970501e-05,
"loss": 1.1048449277877808,
"step": 718
},
{
"epoch": 1.5324813631522898,
"grad_norm": 0.5,
"learning_rate": 1.2403379240914171e-05,
"loss": 1.1177802085876465,
"step": 720
},
{
"epoch": 1.5367412140575079,
"grad_norm": 0.5859375,
"learning_rate": 1.2365883119846358e-05,
"loss": 0.9775673747062683,
"step": 722
},
{
"epoch": 1.5410010649627264,
"grad_norm": 0.56640625,
"learning_rate": 1.2328378942394526e-05,
"loss": 1.0114903450012207,
"step": 724
},
{
"epoch": 1.5452609158679447,
"grad_norm": 0.6640625,
"learning_rate": 1.2290867534363523e-05,
"loss": 0.8942697048187256,
"step": 726
},
{
"epoch": 1.549520766773163,
"grad_norm": 0.490234375,
"learning_rate": 1.225334972171742e-05,
"loss": 1.0175893306732178,
"step": 728
},
{
"epoch": 1.5537806176783813,
"grad_norm": 0.578125,
"learning_rate": 1.2215826330561296e-05,
"loss": 1.027167797088623,
"step": 730
},
{
"epoch": 1.5580404685835996,
"grad_norm": 0.5625,
"learning_rate": 1.2178298187123072e-05,
"loss": 0.9816038608551025,
"step": 732
},
{
"epoch": 1.5623003194888179,
"grad_norm": 0.78515625,
"learning_rate": 1.2140766117735309e-05,
"loss": 1.032719612121582,
"step": 734
},
{
"epoch": 1.5665601703940362,
"grad_norm": 0.85546875,
"learning_rate": 1.210323094881701e-05,
"loss": 1.034834384918213,
"step": 736
},
{
"epoch": 1.5708200212992547,
"grad_norm": 0.6171875,
"learning_rate": 1.2065693506855438e-05,
"loss": 0.9969629645347595,
"step": 738
},
{
"epoch": 1.5750798722044728,
"grad_norm": 0.6484375,
"learning_rate": 1.2028154618387888e-05,
"loss": 0.9849120378494263,
"step": 740
},
{
"epoch": 1.5793397231096913,
"grad_norm": 0.41796875,
"learning_rate": 1.1990615109983516e-05,
"loss": 1.042475700378418,
"step": 742
},
{
"epoch": 1.5835995740149094,
"grad_norm": 0.416015625,
"learning_rate": 1.1953075808225134e-05,
"loss": 0.9866620898246765,
"step": 744
},
{
"epoch": 1.5878594249201279,
"grad_norm": 0.484375,
"learning_rate": 1.191553753969099e-05,
"loss": 1.012044906616211,
"step": 746
},
{
"epoch": 1.592119275825346,
"grad_norm": 0.50390625,
"learning_rate": 1.1878001130936596e-05,
"loss": 0.9986369609832764,
"step": 748
},
{
"epoch": 1.5963791267305645,
"grad_norm": 0.67578125,
"learning_rate": 1.1840467408476495e-05,
"loss": 0.9213187098503113,
"step": 750
},
{
"epoch": 1.6006389776357828,
"grad_norm": 0.5234375,
"learning_rate": 1.1802937198766105e-05,
"loss": 1.0106372833251953,
"step": 752
},
{
"epoch": 1.604898828541001,
"grad_norm": 0.66796875,
"learning_rate": 1.1765411328183474e-05,
"loss": 1.030936360359192,
"step": 754
},
{
"epoch": 1.6091586794462194,
"grad_norm": 0.421875,
"learning_rate": 1.172789062301112e-05,
"loss": 0.9806941747665405,
"step": 756
},
{
"epoch": 1.6134185303514377,
"grad_norm": 0.466796875,
"learning_rate": 1.1690375909417817e-05,
"loss": 1.0275804996490479,
"step": 758
},
{
"epoch": 1.617678381256656,
"grad_norm": 0.48046875,
"learning_rate": 1.165286801344042e-05,
"loss": 0.9850936532020569,
"step": 760
},
{
"epoch": 1.6219382321618743,
"grad_norm": 0.5,
"learning_rate": 1.1615367760965652e-05,
"loss": 0.9898473620414734,
"step": 762
},
{
"epoch": 1.6261980830670928,
"grad_norm": 0.71875,
"learning_rate": 1.1577875977711943e-05,
"loss": 0.9166401624679565,
"step": 764
},
{
"epoch": 1.6304579339723109,
"grad_norm": 0.90234375,
"learning_rate": 1.1540393489211246e-05,
"loss": 1.0430673360824585,
"step": 766
},
{
"epoch": 1.6347177848775294,
"grad_norm": 0.86328125,
"learning_rate": 1.1502921120790832e-05,
"loss": 0.9577189683914185,
"step": 768
},
{
"epoch": 1.6389776357827475,
"grad_norm": 0.6015625,
"learning_rate": 1.146545969755516e-05,
"loss": 1.046250581741333,
"step": 770
},
{
"epoch": 1.643237486687966,
"grad_norm": 0.4453125,
"learning_rate": 1.1428010044367676e-05,
"loss": 0.9433165788650513,
"step": 772
},
{
"epoch": 1.6474973375931843,
"grad_norm": 0.53125,
"learning_rate": 1.139057298583266e-05,
"loss": 1.0670093297958374,
"step": 774
},
{
"epoch": 1.6517571884984026,
"grad_norm": 0.8671875,
"learning_rate": 1.1353149346277071e-05,
"loss": 1.0747054815292358,
"step": 776
},
{
"epoch": 1.6560170394036209,
"grad_norm": 0.462890625,
"learning_rate": 1.1315739949732403e-05,
"loss": 0.9709829092025757,
"step": 778
},
{
"epoch": 1.6602768903088392,
"grad_norm": 0.55859375,
"learning_rate": 1.1278345619916519e-05,
"loss": 0.9456352591514587,
"step": 780
},
{
"epoch": 1.6645367412140575,
"grad_norm": 0.61328125,
"learning_rate": 1.1240967180215546e-05,
"loss": 0.9906789064407349,
"step": 782
},
{
"epoch": 1.6687965921192758,
"grad_norm": 0.9453125,
"learning_rate": 1.1203605453665706e-05,
"loss": 0.9821105599403381,
"step": 784
},
{
"epoch": 1.6730564430244943,
"grad_norm": 0.76953125,
"learning_rate": 1.1166261262935234e-05,
"loss": 0.9833469390869141,
"step": 786
},
{
"epoch": 1.6773162939297124,
"grad_norm": 0.384765625,
"learning_rate": 1.1128935430306224e-05,
"loss": 0.9621955752372742,
"step": 788
},
{
"epoch": 1.681576144834931,
"grad_norm": 0.76953125,
"learning_rate": 1.1091628777656558e-05,
"loss": 1.0520445108413696,
"step": 790
},
{
"epoch": 1.685835995740149,
"grad_norm": 1.734375,
"learning_rate": 1.105434212644179e-05,
"loss": 1.00932776927948,
"step": 792
},
{
"epoch": 1.6900958466453675,
"grad_norm": 2.109375,
"learning_rate": 1.1017076297677054e-05,
"loss": 1.0297919511795044,
"step": 794
},
{
"epoch": 1.6943556975505856,
"grad_norm": 0.6171875,
"learning_rate": 1.0979832111919002e-05,
"loss": 0.9979485273361206,
"step": 796
},
{
"epoch": 1.698615548455804,
"grad_norm": 0.71875,
"learning_rate": 1.0942610389247735e-05,
"loss": 1.0736881494522095,
"step": 798
},
{
"epoch": 1.7028753993610224,
"grad_norm": 0.48046875,
"learning_rate": 1.0905411949248722e-05,
"loss": 0.9282923936843872,
"step": 800
},
{
"epoch": 1.7071352502662407,
"grad_norm": 1.2734375,
"learning_rate": 1.086823761099478e-05,
"loss": 0.9751482605934143,
"step": 802
},
{
"epoch": 1.711395101171459,
"grad_norm": 0.61328125,
"learning_rate": 1.0831088193028037e-05,
"loss": 1.050538182258606,
"step": 804
},
{
"epoch": 1.7156549520766773,
"grad_norm": 0.4375,
"learning_rate": 1.0793964513341892e-05,
"loss": 0.8432624936103821,
"step": 806
},
{
"epoch": 1.7199148029818956,
"grad_norm": 0.8671875,
"learning_rate": 1.0756867389363007e-05,
"loss": 1.0375334024429321,
"step": 808
},
{
"epoch": 1.7241746538871139,
"grad_norm": 0.55078125,
"learning_rate": 1.0719797637933325e-05,
"loss": 1.0315303802490234,
"step": 810
},
{
"epoch": 1.7284345047923324,
"grad_norm": 0.50390625,
"learning_rate": 1.068275607529207e-05,
"loss": 1.0522267818450928,
"step": 812
},
{
"epoch": 1.7326943556975505,
"grad_norm": 0.7265625,
"learning_rate": 1.0645743517057766e-05,
"loss": 0.9734303951263428,
"step": 814
},
{
"epoch": 1.736954206602769,
"grad_norm": 0.5,
"learning_rate": 1.0608760778210305e-05,
"loss": 0.956866979598999,
"step": 816
},
{
"epoch": 1.741214057507987,
"grad_norm": 0.58984375,
"learning_rate": 1.057180867307298e-05,
"loss": 0.9936807751655579,
"step": 818
},
{
"epoch": 1.7454739084132056,
"grad_norm": 0.78515625,
"learning_rate": 1.053488801529455e-05,
"loss": 1.0136879682540894,
"step": 820
},
{
"epoch": 1.749733759318424,
"grad_norm": 0.5078125,
"learning_rate": 1.0497999617831352e-05,
"loss": 0.9871541261672974,
"step": 822
},
{
"epoch": 1.7539936102236422,
"grad_norm": 0.76171875,
"learning_rate": 1.0461144292929361e-05,
"loss": 0.990527868270874,
"step": 824
},
{
"epoch": 1.7582534611288605,
"grad_norm": 1.25,
"learning_rate": 1.0424322852106354e-05,
"loss": 1.039267897605896,
"step": 826
},
{
"epoch": 1.7625133120340788,
"grad_norm": 0.470703125,
"learning_rate": 1.0387536106133989e-05,
"loss": 1.069290280342102,
"step": 828
},
{
"epoch": 1.766773162939297,
"grad_norm": 0.6328125,
"learning_rate": 1.0350784865019987e-05,
"loss": 0.9809550642967224,
"step": 830
},
{
"epoch": 1.7710330138445154,
"grad_norm": 0.48046875,
"learning_rate": 1.0314069937990298e-05,
"loss": 0.9778714776039124,
"step": 832
},
{
"epoch": 1.775292864749734,
"grad_norm": 1.015625,
"learning_rate": 1.027739213347125e-05,
"loss": 0.9834805130958557,
"step": 834
},
{
"epoch": 1.779552715654952,
"grad_norm": 0.6640625,
"learning_rate": 1.0240752259071792e-05,
"loss": 1.0088666677474976,
"step": 836
},
{
"epoch": 1.7838125665601705,
"grad_norm": 0.47265625,
"learning_rate": 1.0204151121565675e-05,
"loss": 0.958771288394928,
"step": 838
},
{
"epoch": 1.7880724174653886,
"grad_norm": 0.59375,
"learning_rate": 1.0167589526873706e-05,
"loss": 1.0218135118484497,
"step": 840
},
{
"epoch": 1.792332268370607,
"grad_norm": 0.71484375,
"learning_rate": 1.0131068280045997e-05,
"loss": 1.027093768119812,
"step": 842
},
{
"epoch": 1.7965921192758252,
"grad_norm": 0.91015625,
"learning_rate": 1.0094588185244248e-05,
"loss": 1.0224063396453857,
"step": 844
},
{
"epoch": 1.8008519701810437,
"grad_norm": 0.60546875,
"learning_rate": 1.0058150045724022e-05,
"loss": 1.0029429197311401,
"step": 846
},
{
"epoch": 1.805111821086262,
"grad_norm": 0.6484375,
"learning_rate": 1.002175466381707e-05,
"loss": 0.9247448444366455,
"step": 848
},
{
"epoch": 1.8093716719914803,
"grad_norm": 0.59375,
"learning_rate": 9.985402840913666e-06,
"loss": 1.0142505168914795,
"step": 850
},
{
"epoch": 1.8136315228966986,
"grad_norm": 0.5078125,
"learning_rate": 9.949095377444961e-06,
"loss": 0.9922072291374207,
"step": 852
},
{
"epoch": 1.817891373801917,
"grad_norm": 0.59375,
"learning_rate": 9.912833072865345e-06,
"loss": 1.0142475366592407,
"step": 854
},
{
"epoch": 1.8221512247071352,
"grad_norm": 0.91015625,
"learning_rate": 9.876616725634864e-06,
"loss": 0.9615439176559448,
"step": 856
},
{
"epoch": 1.8264110756123535,
"grad_norm": 0.408203125,
"learning_rate": 9.840447133201623e-06,
"loss": 0.9525391459465027,
"step": 858
},
{
"epoch": 1.830670926517572,
"grad_norm": 0.57421875,
"learning_rate": 9.804325091984234e-06,
"loss": 1.0148870944976807,
"step": 860
},
{
"epoch": 1.83493077742279,
"grad_norm": 0.53125,
"learning_rate": 9.768251397354276e-06,
"loss": 0.9480238556861877,
"step": 862
},
{
"epoch": 1.8391906283280086,
"grad_norm": 0.8046875,
"learning_rate": 9.73222684361879e-06,
"loss": 1.0006335973739624,
"step": 864
},
{
"epoch": 1.8434504792332267,
"grad_norm": 0.71484375,
"learning_rate": 9.696252224002783e-06,
"loss": 0.9949763417243958,
"step": 866
},
{
"epoch": 1.8477103301384452,
"grad_norm": 0.51953125,
"learning_rate": 9.660328330631752e-06,
"loss": 0.9215972423553467,
"step": 868
},
{
"epoch": 1.8519701810436635,
"grad_norm": 0.53515625,
"learning_rate": 9.624455954514258e-06,
"loss": 0.9809213280677795,
"step": 870
},
{
"epoch": 1.8562300319488818,
"grad_norm": 0.55078125,
"learning_rate": 9.588635885524514e-06,
"loss": 0.9926417469978333,
"step": 872
},
{
"epoch": 1.8604898828541,
"grad_norm": 3.5625,
"learning_rate": 9.552868912384962e-06,
"loss": 1.0632542371749878,
"step": 874
},
{
"epoch": 1.8647497337593184,
"grad_norm": 0.54296875,
"learning_rate": 9.517155822648938e-06,
"loss": 1.0109238624572754,
"step": 876
},
{
"epoch": 1.8690095846645367,
"grad_norm": 0.66015625,
"learning_rate": 9.481497402683325e-06,
"loss": 0.9798102974891663,
"step": 878
},
{
"epoch": 1.873269435569755,
"grad_norm": 0.50390625,
"learning_rate": 9.445894437651211e-06,
"loss": 1.0319267511367798,
"step": 880
},
{
"epoch": 1.8775292864749735,
"grad_norm": 0.451171875,
"learning_rate": 9.41034771149464e-06,
"loss": 1.0264664888381958,
"step": 882
},
{
"epoch": 1.8817891373801916,
"grad_norm": 0.51171875,
"learning_rate": 9.374858006917328e-06,
"loss": 1.014363169670105,
"step": 884
},
{
"epoch": 1.8860489882854101,
"grad_norm": 0.53125,
"learning_rate": 9.339426105367428e-06,
"loss": 0.9093478322029114,
"step": 886
},
{
"epoch": 1.8903088391906282,
"grad_norm": 0.4921875,
"learning_rate": 9.304052787020326e-06,
"loss": 0.9274075627326965,
"step": 888
},
{
"epoch": 1.8945686900958467,
"grad_norm": 0.484375,
"learning_rate": 9.26873883076147e-06,
"loss": 1.0133625268936157,
"step": 890
},
{
"epoch": 1.898828541001065,
"grad_norm": 0.5234375,
"learning_rate": 9.233485014169214e-06,
"loss": 1.0017213821411133,
"step": 892
},
{
"epoch": 1.9030883919062833,
"grad_norm": 0.515625,
"learning_rate": 9.198292113497686e-06,
"loss": 1.0270392894744873,
"step": 894
},
{
"epoch": 1.9073482428115016,
"grad_norm": 0.546875,
"learning_rate": 9.163160903659715e-06,
"loss": 0.9596851468086243,
"step": 896
},
{
"epoch": 1.91160809371672,
"grad_norm": 0.54296875,
"learning_rate": 9.128092158209762e-06,
"loss": 1.0112367868423462,
"step": 898
},
{
"epoch": 1.9158679446219382,
"grad_norm": 0.6796875,
"learning_rate": 9.093086649326873e-06,
"loss": 1.0255273580551147,
"step": 900
},
{
"epoch": 1.9201277955271565,
"grad_norm": 0.62890625,
"learning_rate": 9.05814514779769e-06,
"loss": 0.9857978820800781,
"step": 902
},
{
"epoch": 1.924387646432375,
"grad_norm": 0.546875,
"learning_rate": 9.02326842299949e-06,
"loss": 0.9689497351646423,
"step": 904
},
{
"epoch": 1.928647497337593,
"grad_norm": 0.640625,
"learning_rate": 8.98845724288321e-06,
"loss": 1.039789080619812,
"step": 906
},
{
"epoch": 1.9329073482428116,
"grad_norm": 0.38671875,
"learning_rate": 8.95371237395657e-06,
"loss": 0.9854997396469116,
"step": 908
},
{
"epoch": 1.9371671991480297,
"grad_norm": 0.5859375,
"learning_rate": 8.919034581267185e-06,
"loss": 0.863085150718689,
"step": 910
},
{
"epoch": 1.9414270500532482,
"grad_norm": 0.6328125,
"learning_rate": 8.884424628385713e-06,
"loss": 0.9845672845840454,
"step": 912
},
{
"epoch": 1.9456869009584663,
"grad_norm": 0.609375,
"learning_rate": 8.849883277389046e-06,
"loss": 0.9393159747123718,
"step": 914
},
{
"epoch": 1.9499467518636848,
"grad_norm": 1.0546875,
"learning_rate": 8.815411288843531e-06,
"loss": 1.019649624824524,
"step": 916
},
{
"epoch": 1.9542066027689031,
"grad_norm": 0.53125,
"learning_rate": 8.781009421788226e-06,
"loss": 1.001703143119812,
"step": 918
},
{
"epoch": 1.9584664536741214,
"grad_norm": 0.478515625,
"learning_rate": 8.74667843371818e-06,
"loss": 1.0148494243621826,
"step": 920
},
{
"epoch": 1.9627263045793397,
"grad_norm": 0.60546875,
"learning_rate": 8.712419080567753e-06,
"loss": 0.9227896332740784,
"step": 922
},
{
"epoch": 1.966986155484558,
"grad_norm": 0.578125,
"learning_rate": 8.678232116693988e-06,
"loss": 0.9633923768997192,
"step": 924
},
{
"epoch": 1.9712460063897763,
"grad_norm": 0.6640625,
"learning_rate": 8.644118294859972e-06,
"loss": 0.9478899836540222,
"step": 926
},
{
"epoch": 1.9755058572949946,
"grad_norm": 0.55078125,
"learning_rate": 8.610078366218287e-06,
"loss": 1.0053707361221313,
"step": 928
},
{
"epoch": 1.9797657082002131,
"grad_norm": 0.640625,
"learning_rate": 8.576113080294459e-06,
"loss": 0.9101724624633789,
"step": 930
},
{
"epoch": 1.9840255591054312,
"grad_norm": 0.63671875,
"learning_rate": 8.542223184970451e-06,
"loss": 1.0562280416488647,
"step": 932
},
{
"epoch": 1.9882854100106497,
"grad_norm": 0.494140625,
"learning_rate": 8.5084094264682e-06,
"loss": 0.9434181451797485,
"step": 934
},
{
"epoch": 1.9925452609158678,
"grad_norm": 0.75390625,
"learning_rate": 8.474672549333191e-06,
"loss": 1.0050933361053467,
"step": 936
},
{
"epoch": 1.9968051118210863,
"grad_norm": 0.61328125,
"learning_rate": 8.441013296418057e-06,
"loss": 1.014664649963379,
"step": 938
},
{
"epoch": 2.0,
"grad_norm": 1.6171875,
"learning_rate": 8.407432408866211e-06,
"loss": 0.9959129095077515,
"step": 940
},
{
"epoch": 2.0042598509052185,
"grad_norm": 0.91015625,
"learning_rate": 8.373930626095557e-06,
"loss": 0.8519871830940247,
"step": 942
},
{
"epoch": 2.0085197018104366,
"grad_norm": 0.50390625,
"learning_rate": 8.340508685782173e-06,
"loss": 0.8345139622688293,
"step": 944
},
{
"epoch": 2.012779552715655,
"grad_norm": 0.64453125,
"learning_rate": 8.30716732384409e-06,
"loss": 0.8144312500953674,
"step": 946
},
{
"epoch": 2.017039403620873,
"grad_norm": 0.6796875,
"learning_rate": 8.27390727442509e-06,
"loss": 0.8890555500984192,
"step": 948
},
{
"epoch": 2.0212992545260917,
"grad_norm": 0.54296875,
"learning_rate": 8.24072926987853e-06,
"loss": 0.9254493117332458,
"step": 950
},
{
"epoch": 2.02555910543131,
"grad_norm": 0.431640625,
"learning_rate": 8.207634040751218e-06,
"loss": 0.8629240989685059,
"step": 952
},
{
"epoch": 2.0298189563365283,
"grad_norm": 0.419921875,
"learning_rate": 8.174622315767325e-06,
"loss": 0.9234293699264526,
"step": 954
},
{
"epoch": 2.0340788072417464,
"grad_norm": 0.67578125,
"learning_rate": 8.141694821812352e-06,
"loss": 0.8904613852500916,
"step": 956
},
{
"epoch": 2.038338658146965,
"grad_norm": 0.56640625,
"learning_rate": 8.108852283917116e-06,
"loss": 0.8621918559074402,
"step": 958
},
{
"epoch": 2.042598509052183,
"grad_norm": 0.6796875,
"learning_rate": 8.07609542524178e-06,
"loss": 0.8154129981994629,
"step": 960
},
{
"epoch": 2.0468583599574015,
"grad_norm": 0.388671875,
"learning_rate": 8.043424967059936e-06,
"loss": 0.7776715755462646,
"step": 962
},
{
"epoch": 2.0511182108626196,
"grad_norm": 0.490234375,
"learning_rate": 8.010841628742728e-06,
"loss": 0.9061995148658752,
"step": 964
},
{
"epoch": 2.055378061767838,
"grad_norm": 0.45703125,
"learning_rate": 7.978346127742999e-06,
"loss": 0.8465338349342346,
"step": 966
},
{
"epoch": 2.0596379126730566,
"grad_norm": 0.71484375,
"learning_rate": 7.945939179579509e-06,
"loss": 0.8736094832420349,
"step": 968
},
{
"epoch": 2.0638977635782747,
"grad_norm": 0.625,
"learning_rate": 7.913621497821174e-06,
"loss": 0.8707051277160645,
"step": 970
},
{
"epoch": 2.0681576144834932,
"grad_norm": 0.6796875,
"learning_rate": 7.881393794071338e-06,
"loss": 0.8639911413192749,
"step": 972
},
{
"epoch": 2.0724174653887113,
"grad_norm": 0.65625,
"learning_rate": 7.84925677795213e-06,
"loss": 0.8878441452980042,
"step": 974
},
{
"epoch": 2.07667731629393,
"grad_norm": 0.73828125,
"learning_rate": 7.817211157088823e-06,
"loss": 0.8697801232337952,
"step": 976
},
{
"epoch": 2.080937167199148,
"grad_norm": 0.6015625,
"learning_rate": 7.785257637094265e-06,
"loss": 0.9132862687110901,
"step": 978
},
{
"epoch": 2.0851970181043664,
"grad_norm": 0.81640625,
"learning_rate": 7.753396921553325e-06,
"loss": 0.9692128896713257,
"step": 980
},
{
"epoch": 2.0894568690095845,
"grad_norm": 0.6953125,
"learning_rate": 7.721629712007408e-06,
"loss": 0.8117839694023132,
"step": 982
},
{
"epoch": 2.093716719914803,
"grad_norm": 0.65234375,
"learning_rate": 7.68995670793902e-06,
"loss": 0.8652855157852173,
"step": 984
},
{
"epoch": 2.097976570820021,
"grad_norm": 0.671875,
"learning_rate": 7.658378606756342e-06,
"loss": 0.8414390683174133,
"step": 986
},
{
"epoch": 2.1022364217252396,
"grad_norm": 0.78125,
"learning_rate": 7.626896103777901e-06,
"loss": 0.975841224193573,
"step": 988
},
{
"epoch": 2.106496272630458,
"grad_norm": 0.5078125,
"learning_rate": 7.595509892217243e-06,
"loss": 0.8733735680580139,
"step": 990
},
{
"epoch": 2.110756123535676,
"grad_norm": 0.7578125,
"learning_rate": 7.5642206631676655e-06,
"loss": 0.8554286360740662,
"step": 992
},
{
"epoch": 2.1150159744408947,
"grad_norm": 0.7109375,
"learning_rate": 7.533029105587006e-06,
"loss": 0.8418490290641785,
"step": 994
},
{
"epoch": 2.119275825346113,
"grad_norm": 0.578125,
"learning_rate": 7.501935906282486e-06,
"loss": 0.8224227428436279,
"step": 996
},
{
"epoch": 2.1235356762513313,
"grad_norm": 1.1015625,
"learning_rate": 7.4709417498955676e-06,
"loss": 0.8177536129951477,
"step": 998
},
{
"epoch": 2.1277955271565494,
"grad_norm": 0.60546875,
"learning_rate": 7.440047318886882e-06,
"loss": 0.8571978211402893,
"step": 1000
},
{
"epoch": 2.132055378061768,
"grad_norm": 0.64453125,
"learning_rate": 7.409253293521209e-06,
"loss": 0.8153810501098633,
"step": 1002
},
{
"epoch": 2.136315228966986,
"grad_norm": 0.75,
"learning_rate": 7.378560351852502e-06,
"loss": 0.8463664650917053,
"step": 1004
},
{
"epoch": 2.1405750798722045,
"grad_norm": 0.578125,
"learning_rate": 7.347969169708937e-06,
"loss": 0.8821415305137634,
"step": 1006
},
{
"epoch": 2.1448349307774226,
"grad_norm": 0.7265625,
"learning_rate": 7.317480420678064e-06,
"loss": 0.9291739463806152,
"step": 1008
},
{
"epoch": 2.149094781682641,
"grad_norm": 0.70703125,
"learning_rate": 7.2870947760919486e-06,
"loss": 0.772282063961029,
"step": 1010
},
{
"epoch": 2.1533546325878596,
"grad_norm": 0.48046875,
"learning_rate": 7.256812905012389e-06,
"loss": 0.8808871507644653,
"step": 1012
},
{
"epoch": 2.1576144834930777,
"grad_norm": 0.486328125,
"learning_rate": 7.2266354742162095e-06,
"loss": 0.8753591179847717,
"step": 1014
},
{
"epoch": 2.1618743343982962,
"grad_norm": 0.7890625,
"learning_rate": 7.196563148180554e-06,
"loss": 0.8668736815452576,
"step": 1016
},
{
"epoch": 2.1661341853035143,
"grad_norm": 0.50390625,
"learning_rate": 7.166596589068271e-06,
"loss": 0.8054329752922058,
"step": 1018
},
{
"epoch": 2.170394036208733,
"grad_norm": 1.953125,
"learning_rate": 7.136736456713316e-06,
"loss": 0.8477678298950195,
"step": 1020
},
{
"epoch": 2.174653887113951,
"grad_norm": 0.53125,
"learning_rate": 7.1069834086062366e-06,
"loss": 0.7990297079086304,
"step": 1022
},
{
"epoch": 2.1789137380191694,
"grad_norm": 0.84765625,
"learning_rate": 7.077338099879696e-06,
"loss": 0.87007075548172,
"step": 1024
},
{
"epoch": 2.1831735889243875,
"grad_norm": 0.8671875,
"learning_rate": 7.047801183294034e-06,
"loss": 0.8642436265945435,
"step": 1026
},
{
"epoch": 2.187433439829606,
"grad_norm": 0.546875,
"learning_rate": 7.018373309222914e-06,
"loss": 0.9285038113594055,
"step": 1028
},
{
"epoch": 2.191693290734824,
"grad_norm": 0.6640625,
"learning_rate": 6.989055125638982e-06,
"loss": 0.861228346824646,
"step": 1030
},
{
"epoch": 2.1959531416400426,
"grad_norm": 0.6328125,
"learning_rate": 6.95984727809961e-06,
"loss": 0.8538438677787781,
"step": 1032
},
{
"epoch": 2.2002129925452607,
"grad_norm": 0.62890625,
"learning_rate": 6.930750409732679e-06,
"loss": 0.9002454876899719,
"step": 1034
},
{
"epoch": 2.2044728434504792,
"grad_norm": 0.828125,
"learning_rate": 6.901765161222426e-06,
"loss": 0.8670125603675842,
"step": 1036
},
{
"epoch": 2.2087326943556977,
"grad_norm": 0.53515625,
"learning_rate": 6.8728921707953225e-06,
"loss": 0.8965079188346863,
"step": 1038
},
{
"epoch": 2.212992545260916,
"grad_norm": 0.63671875,
"learning_rate": 6.844132074206028e-06,
"loss": 0.9224348664283752,
"step": 1040
},
{
"epoch": 2.2172523961661343,
"grad_norm": 0.859375,
"learning_rate": 6.815485504723389e-06,
"loss": 0.8701453804969788,
"step": 1042
},
{
"epoch": 2.2215122470713524,
"grad_norm": 1.0234375,
"learning_rate": 6.786953093116508e-06,
"loss": 0.8607268333435059,
"step": 1044
},
{
"epoch": 2.225772097976571,
"grad_norm": 0.6875,
"learning_rate": 6.758535467640832e-06,
"loss": 0.8190476298332214,
"step": 1046
},
{
"epoch": 2.230031948881789,
"grad_norm": 0.51171875,
"learning_rate": 6.730233254024339e-06,
"loss": 0.7581257820129395,
"step": 1048
},
{
"epoch": 2.2342917997870075,
"grad_norm": 0.53125,
"learning_rate": 6.702047075453751e-06,
"loss": 0.867296576499939,
"step": 1050
},
{
"epoch": 2.2385516506922256,
"grad_norm": 0.8203125,
"learning_rate": 6.673977552560804e-06,
"loss": 0.8582524061203003,
"step": 1052
},
{
"epoch": 2.242811501597444,
"grad_norm": 0.5078125,
"learning_rate": 6.646025303408606e-06,
"loss": 0.8571093082427979,
"step": 1054
},
{
"epoch": 2.247071352502662,
"grad_norm": 0.6484375,
"learning_rate": 6.618190943478003e-06,
"loss": 0.867594301700592,
"step": 1056
},
{
"epoch": 2.2513312034078807,
"grad_norm": 0.66015625,
"learning_rate": 6.590475085654047e-06,
"loss": 0.7921283841133118,
"step": 1058
},
{
"epoch": 2.255591054313099,
"grad_norm": 0.8125,
"learning_rate": 6.56287834021248e-06,
"loss": 0.8648768067359924,
"step": 1060
},
{
"epoch": 2.2598509052183173,
"grad_norm": 0.82421875,
"learning_rate": 6.535401314806309e-06,
"loss": 0.8589727282524109,
"step": 1062
},
{
"epoch": 2.264110756123536,
"grad_norm": 0.48828125,
"learning_rate": 6.508044614452431e-06,
"loss": 0.8992140889167786,
"step": 1064
},
{
"epoch": 2.268370607028754,
"grad_norm": 0.69921875,
"learning_rate": 6.480808841518298e-06,
"loss": 0.8318920135498047,
"step": 1066
},
{
"epoch": 2.2726304579339724,
"grad_norm": 0.80859375,
"learning_rate": 6.453694595708667e-06,
"loss": 0.8525157570838928,
"step": 1068
},
{
"epoch": 2.2768903088391905,
"grad_norm": 0.6953125,
"learning_rate": 6.426702474052381e-06,
"loss": 0.8547057509422302,
"step": 1070
},
{
"epoch": 2.281150159744409,
"grad_norm": 0.890625,
"learning_rate": 6.39983307088923e-06,
"loss": 0.8861790299415588,
"step": 1072
},
{
"epoch": 2.285410010649627,
"grad_norm": 0.62890625,
"learning_rate": 6.373086977856868e-06,
"loss": 0.8417205810546875,
"step": 1074
},
{
"epoch": 2.2896698615548456,
"grad_norm": 0.46484375,
"learning_rate": 6.346464783877784e-06,
"loss": 0.9120402932167053,
"step": 1076
},
{
"epoch": 2.2939297124600637,
"grad_norm": 0.53515625,
"learning_rate": 6.319967075146332e-06,
"loss": 0.8795621991157532,
"step": 1078
},
{
"epoch": 2.2981895633652822,
"grad_norm": 1.5078125,
"learning_rate": 6.293594435115814e-06,
"loss": 0.884793758392334,
"step": 1080
},
{
"epoch": 2.3024494142705008,
"grad_norm": 0.482421875,
"learning_rate": 6.267347444485654e-06,
"loss": 0.9300041198730469,
"step": 1082
},
{
"epoch": 2.306709265175719,
"grad_norm": 0.6796875,
"learning_rate": 6.2412266811886005e-06,
"loss": 0.8604894876480103,
"step": 1084
},
{
"epoch": 2.3109691160809374,
"grad_norm": 0.65625,
"learning_rate": 6.215232720377991e-06,
"loss": 0.8723834753036499,
"step": 1086
},
{
"epoch": 2.3152289669861554,
"grad_norm": 0.482421875,
"learning_rate": 6.189366134415112e-06,
"loss": 0.8270635008811951,
"step": 1088
},
{
"epoch": 2.319488817891374,
"grad_norm": 0.84375,
"learning_rate": 6.163627492856572e-06,
"loss": 0.873162567615509,
"step": 1090
},
{
"epoch": 2.323748668796592,
"grad_norm": 0.6796875,
"learning_rate": 6.13801736244177e-06,
"loss": 0.8589759469032288,
"step": 1092
},
{
"epoch": 2.3280085197018106,
"grad_norm": 1.078125,
"learning_rate": 6.112536307080422e-06,
"loss": 0.8987887501716614,
"step": 1094
},
{
"epoch": 2.3322683706070286,
"grad_norm": 0.6640625,
"learning_rate": 6.087184887840139e-06,
"loss": 0.8068996071815491,
"step": 1096
},
{
"epoch": 2.336528221512247,
"grad_norm": 0.4921875,
"learning_rate": 6.061963662934072e-06,
"loss": 0.802603006362915,
"step": 1098
},
{
"epoch": 2.3407880724174652,
"grad_norm": 0.5234375,
"learning_rate": 6.036873187708615e-06,
"loss": 0.8735599517822266,
"step": 1100
},
{
"epoch": 2.3450479233226837,
"grad_norm": 0.5546875,
"learning_rate": 6.011914014631191e-06,
"loss": 0.7997400760650635,
"step": 1102
},
{
"epoch": 2.349307774227902,
"grad_norm": 0.59765625,
"learning_rate": 5.987086693278083e-06,
"loss": 0.8167827129364014,
"step": 1104
},
{
"epoch": 2.3535676251331203,
"grad_norm": 1.078125,
"learning_rate": 5.962391770322319e-06,
"loss": 0.853871762752533,
"step": 1106
},
{
"epoch": 2.357827476038339,
"grad_norm": 0.59765625,
"learning_rate": 5.937829789521655e-06,
"loss": 0.8599565029144287,
"step": 1108
},
{
"epoch": 2.362087326943557,
"grad_norm": 0.5625,
"learning_rate": 5.913401291706585e-06,
"loss": 0.8716296553611755,
"step": 1110
},
{
"epoch": 2.3663471778487755,
"grad_norm": 0.671875,
"learning_rate": 5.889106814768442e-06,
"loss": 0.8599187731742859,
"step": 1112
},
{
"epoch": 2.3706070287539935,
"grad_norm": 0.96484375,
"learning_rate": 5.864946893647556e-06,
"loss": 0.8140082955360413,
"step": 1114
},
{
"epoch": 2.374866879659212,
"grad_norm": 0.8125,
"learning_rate": 5.8409220603214704e-06,
"loss": 0.8339557647705078,
"step": 1116
},
{
"epoch": 2.37912673056443,
"grad_norm": 0.447265625,
"learning_rate": 5.817032843793222e-06,
"loss": 0.828040599822998,
"step": 1118
},
{
"epoch": 2.3833865814696487,
"grad_norm": 0.80078125,
"learning_rate": 5.793279770079706e-06,
"loss": 0.869745135307312,
"step": 1120
},
{
"epoch": 2.3876464323748667,
"grad_norm": 1.140625,
"learning_rate": 5.769663362200086e-06,
"loss": 0.8580743074417114,
"step": 1122
},
{
"epoch": 2.3919062832800853,
"grad_norm": 0.70703125,
"learning_rate": 5.746184140164285e-06,
"loss": 0.8233816027641296,
"step": 1124
},
{
"epoch": 2.3961661341853033,
"grad_norm": 0.4375,
"learning_rate": 5.722842620961518e-06,
"loss": 0.8709803819656372,
"step": 1126
},
{
"epoch": 2.400425985090522,
"grad_norm": 0.58203125,
"learning_rate": 5.69963931854893e-06,
"loss": 0.9047388434410095,
"step": 1128
},
{
"epoch": 2.40468583599574,
"grad_norm": 1.890625,
"learning_rate": 5.676574743840264e-06,
"loss": 0.8460210561752319,
"step": 1130
},
{
"epoch": 2.4089456869009584,
"grad_norm": 0.73046875,
"learning_rate": 5.6536494046946174e-06,
"loss": 0.9233959913253784,
"step": 1132
},
{
"epoch": 2.413205537806177,
"grad_norm": 0.40234375,
"learning_rate": 5.630863805905259e-06,
"loss": 0.8673460483551025,
"step": 1134
},
{
"epoch": 2.417465388711395,
"grad_norm": 1.0234375,
"learning_rate": 5.608218449188516e-06,
"loss": 0.9345380663871765,
"step": 1136
},
{
"epoch": 2.4217252396166136,
"grad_norm": 1.1328125,
"learning_rate": 5.585713833172718e-06,
"loss": 0.9205328226089478,
"step": 1138
},
{
"epoch": 2.4259850905218316,
"grad_norm": 0.61328125,
"learning_rate": 5.563350453387224e-06,
"loss": 0.7964959740638733,
"step": 1140
},
{
"epoch": 2.43024494142705,
"grad_norm": 0.49609375,
"learning_rate": 5.541128802251518e-06,
"loss": 0.811107873916626,
"step": 1142
},
{
"epoch": 2.4345047923322682,
"grad_norm": 0.51953125,
"learning_rate": 5.5190493690643566e-06,
"loss": 0.787108838558197,
"step": 1144
},
{
"epoch": 2.4387646432374868,
"grad_norm": 17.375,
"learning_rate": 5.497112639992993e-06,
"loss": 0.9524678587913513,
"step": 1146
},
{
"epoch": 2.443024494142705,
"grad_norm": 1.109375,
"learning_rate": 5.4753190980624834e-06,
"loss": 0.7983661890029907,
"step": 1148
},
{
"epoch": 2.4472843450479234,
"grad_norm": 1.109375,
"learning_rate": 5.453669223145042e-06,
"loss": 0.8548356294631958,
"step": 1150
},
{
"epoch": 2.451544195953142,
"grad_norm": 0.6328125,
"learning_rate": 5.432163491949475e-06,
"loss": 0.9406867027282715,
"step": 1152
},
{
"epoch": 2.45580404685836,
"grad_norm": 0.6015625,
"learning_rate": 5.410802378010691e-06,
"loss": 0.8843863010406494,
"step": 1154
},
{
"epoch": 2.460063897763578,
"grad_norm": 1.125,
"learning_rate": 5.389586351679272e-06,
"loss": 0.8945695757865906,
"step": 1156
},
{
"epoch": 2.4643237486687966,
"grad_norm": 0.58203125,
"learning_rate": 5.368515880111107e-06,
"loss": 0.8422825336456299,
"step": 1158
},
{
"epoch": 2.468583599574015,
"grad_norm": 0.84375,
"learning_rate": 5.347591427257115e-06,
"loss": 0.940974771976471,
"step": 1160
},
{
"epoch": 2.472843450479233,
"grad_norm": 0.484375,
"learning_rate": 5.326813453853033e-06,
"loss": 0.849635124206543,
"step": 1162
},
{
"epoch": 2.4771033013844517,
"grad_norm": 0.62109375,
"learning_rate": 5.3061824174092645e-06,
"loss": 0.8890464901924133,
"step": 1164
},
{
"epoch": 2.4813631522896697,
"grad_norm": 1.1015625,
"learning_rate": 5.2856987722008004e-06,
"loss": 0.8929099440574646,
"step": 1166
},
{
"epoch": 2.4856230031948883,
"grad_norm": 0.52734375,
"learning_rate": 5.265362969257225e-06,
"loss": 0.811379075050354,
"step": 1168
},
{
"epoch": 2.4898828541001063,
"grad_norm": 0.56640625,
"learning_rate": 5.245175456352788e-06,
"loss": 0.8641577959060669,
"step": 1170
},
{
"epoch": 2.494142705005325,
"grad_norm": 0.79296875,
"learning_rate": 5.225136677996531e-06,
"loss": 0.8928030133247375,
"step": 1172
},
{
"epoch": 2.498402555910543,
"grad_norm": 1.0546875,
"learning_rate": 5.205247075422514e-06,
"loss": 0.8985695838928223,
"step": 1174
},
{
"epoch": 2.5026624068157615,
"grad_norm": 0.53515625,
"learning_rate": 5.185507086580094e-06,
"loss": 0.825484037399292,
"step": 1176
},
{
"epoch": 2.50692225772098,
"grad_norm": 0.71484375,
"learning_rate": 5.165917146124276e-06,
"loss": 0.8295057415962219,
"step": 1178
},
{
"epoch": 2.511182108626198,
"grad_norm": 0.66796875,
"learning_rate": 5.146477685406154e-06,
"loss": 0.8838967084884644,
"step": 1180
},
{
"epoch": 2.515441959531416,
"grad_norm": 0.71875,
"learning_rate": 5.127189132463404e-06,
"loss": 0.8865538239479065,
"step": 1182
},
{
"epoch": 2.5197018104366347,
"grad_norm": 0.984375,
"learning_rate": 5.108051912010865e-06,
"loss": 0.839586079120636,
"step": 1184
},
{
"epoch": 2.523961661341853,
"grad_norm": 0.70703125,
"learning_rate": 5.089066445431183e-06,
"loss": 0.8859398365020752,
"step": 1186
},
{
"epoch": 2.5282215122470713,
"grad_norm": 1.1328125,
"learning_rate": 5.070233150765531e-06,
"loss": 0.8947311639785767,
"step": 1188
},
{
"epoch": 2.5324813631522898,
"grad_norm": 0.7265625,
"learning_rate": 5.051552442704411e-06,
"loss": 0.8858785033226013,
"step": 1190
},
{
"epoch": 2.536741214057508,
"grad_norm": 0.5546875,
"learning_rate": 5.033024732578518e-06,
"loss": 0.8125391006469727,
"step": 1192
},
{
"epoch": 2.5410010649627264,
"grad_norm": 1.0703125,
"learning_rate": 5.014650428349683e-06,
"loss": 0.8854655027389526,
"step": 1194
},
{
"epoch": 2.545260915867945,
"grad_norm": 0.65234375,
"learning_rate": 4.996429934601895e-06,
"loss": 0.8128026127815247,
"step": 1196
},
{
"epoch": 2.549520766773163,
"grad_norm": 0.9140625,
"learning_rate": 4.978363652532383e-06,
"loss": 0.8405453562736511,
"step": 1198
},
{
"epoch": 2.553780617678381,
"grad_norm": 0.6953125,
"learning_rate": 4.960451979942789e-06,
"loss": 0.959130585193634,
"step": 1200
},
{
"epoch": 2.5580404685835996,
"grad_norm": 5.28125,
"learning_rate": 4.942695311230408e-06,
"loss": 0.7953628897666931,
"step": 1202
},
{
"epoch": 2.562300319488818,
"grad_norm": 0.84765625,
"learning_rate": 4.925094037379507e-06,
"loss": 0.7662147283554077,
"step": 1204
},
{
"epoch": 2.566560170394036,
"grad_norm": 0.9453125,
"learning_rate": 4.907648545952706e-06,
"loss": 0.8059567213058472,
"step": 1206
},
{
"epoch": 2.5708200212992547,
"grad_norm": 0.7578125,
"learning_rate": 4.890359221082451e-06,
"loss": 0.9323968291282654,
"step": 1208
},
{
"epoch": 2.5750798722044728,
"grad_norm": 1.1171875,
"learning_rate": 4.87322644346256e-06,
"loss": 0.8555244207382202,
"step": 1210
},
{
"epoch": 2.5793397231096913,
"grad_norm": 0.5390625,
"learning_rate": 4.856250590339828e-06,
"loss": 0.8834168314933777,
"step": 1212
},
{
"epoch": 2.5835995740149094,
"grad_norm": 0.58984375,
"learning_rate": 4.8394320355057355e-06,
"loss": 0.9264652132987976,
"step": 1214
},
{
"epoch": 2.587859424920128,
"grad_norm": 0.69140625,
"learning_rate": 4.822771149288208e-06,
"loss": 0.8777004480361938,
"step": 1216
},
{
"epoch": 2.592119275825346,
"grad_norm": 0.63671875,
"learning_rate": 4.80626829854346e-06,
"loss": 0.8719021081924438,
"step": 1218
},
{
"epoch": 2.5963791267305645,
"grad_norm": 0.62109375,
"learning_rate": 4.789923846647926e-06,
"loss": 0.8178717494010925,
"step": 1220
},
{
"epoch": 2.600638977635783,
"grad_norm": 0.70703125,
"learning_rate": 4.773738153490252e-06,
"loss": 0.9265211820602417,
"step": 1222
},
{
"epoch": 2.604898828541001,
"grad_norm": 0.765625,
"learning_rate": 4.75771157546338e-06,
"loss": 0.9368568658828735,
"step": 1224
},
{
"epoch": 2.609158679446219,
"grad_norm": 0.4453125,
"learning_rate": 4.7418444654566845e-06,
"loss": 0.8576453328132629,
"step": 1226
},
{
"epoch": 2.6134185303514377,
"grad_norm": 0.64453125,
"learning_rate": 4.726137172848219e-06,
"loss": 0.8761544823646545,
"step": 1228
},
{
"epoch": 2.617678381256656,
"grad_norm": 0.58203125,
"learning_rate": 4.710590043497013e-06,
"loss": 0.8335532546043396,
"step": 1230
},
{
"epoch": 2.6219382321618743,
"grad_norm": 0.7578125,
"learning_rate": 4.695203419735462e-06,
"loss": 0.8370216488838196,
"step": 1232
},
{
"epoch": 2.626198083067093,
"grad_norm": 0.578125,
"learning_rate": 4.6799776403617875e-06,
"loss": 0.844344973564148,
"step": 1234
},
{
"epoch": 2.630457933972311,
"grad_norm": 0.89453125,
"learning_rate": 4.6649130406325726e-06,
"loss": 0.9227336645126343,
"step": 1236
},
{
"epoch": 2.6347177848775294,
"grad_norm": 0.61328125,
"learning_rate": 4.650009952255389e-06,
"loss": 0.8581987023353577,
"step": 1238
},
{
"epoch": 2.6389776357827475,
"grad_norm": 1.4140625,
"learning_rate": 4.635268703381484e-06,
"loss": 0.8400872349739075,
"step": 1240
},
{
"epoch": 2.643237486687966,
"grad_norm": 0.6796875,
"learning_rate": 4.620689618598568e-06,
"loss": 0.8582939505577087,
"step": 1242
},
{
"epoch": 2.647497337593184,
"grad_norm": 0.6953125,
"learning_rate": 4.606273018923644e-06,
"loss": 0.9465791583061218,
"step": 1244
},
{
"epoch": 2.6517571884984026,
"grad_norm": 0.70703125,
"learning_rate": 4.592019221795969e-06,
"loss": 0.8765722513198853,
"step": 1246
},
{
"epoch": 2.656017039403621,
"grad_norm": 0.8125,
"learning_rate": 4.57792854107004e-06,
"loss": 0.8363880515098572,
"step": 1248
},
{
"epoch": 2.660276890308839,
"grad_norm": 0.625,
"learning_rate": 4.564001287008695e-06,
"loss": 0.9550789594650269,
"step": 1250
},
{
"epoch": 2.6645367412140573,
"grad_norm": 0.52734375,
"learning_rate": 4.550237766276277e-06,
"loss": 0.8685204982757568,
"step": 1252
},
{
"epoch": 2.668796592119276,
"grad_norm": 0.58984375,
"learning_rate": 4.536638281931887e-06,
"loss": 0.8562531471252441,
"step": 1254
},
{
"epoch": 2.6730564430244943,
"grad_norm": 0.69921875,
"learning_rate": 4.523203133422705e-06,
"loss": 0.8494756817817688,
"step": 1256
},
{
"epoch": 2.6773162939297124,
"grad_norm": 1.5390625,
"learning_rate": 4.509932616577394e-06,
"loss": 0.899334728717804,
"step": 1258
},
{
"epoch": 2.681576144834931,
"grad_norm": 0.796875,
"learning_rate": 4.4968270235995995e-06,
"loss": 0.8993234634399414,
"step": 1260
},
{
"epoch": 2.685835995740149,
"grad_norm": 0.51953125,
"learning_rate": 4.483886643061501e-06,
"loss": 0.8710511326789856,
"step": 1262
},
{
"epoch": 2.6900958466453675,
"grad_norm": 0.66796875,
"learning_rate": 4.471111759897464e-06,
"loss": 0.8785170912742615,
"step": 1264
},
{
"epoch": 2.6943556975505856,
"grad_norm": 1.6796875,
"learning_rate": 4.45850265539777e-06,
"loss": 0.8322556018829346,
"step": 1266
},
{
"epoch": 2.698615548455804,
"grad_norm": 0.5390625,
"learning_rate": 4.4460596072024114e-06,
"loss": 0.8931154012680054,
"step": 1268
},
{
"epoch": 2.702875399361022,
"grad_norm": 0.60546875,
"learning_rate": 4.433782889294992e-06,
"loss": 0.8494217395782471,
"step": 1270
},
{
"epoch": 2.7071352502662407,
"grad_norm": 1.0,
"learning_rate": 4.4216727719966825e-06,
"loss": 0.8380395174026489,
"step": 1272
},
{
"epoch": 2.711395101171459,
"grad_norm": 0.79296875,
"learning_rate": 4.409729521960276e-06,
"loss": 0.89030522108078,
"step": 1274
},
{
"epoch": 2.7156549520766773,
"grad_norm": 0.74609375,
"learning_rate": 4.397953402164312e-06,
"loss": 0.8374807238578796,
"step": 1276
},
{
"epoch": 2.7199148029818954,
"grad_norm": 0.59765625,
"learning_rate": 4.386344671907289e-06,
"loss": 0.8489108085632324,
"step": 1278
},
{
"epoch": 2.724174653887114,
"grad_norm": 0.53515625,
"learning_rate": 4.374903586801952e-06,
"loss": 0.9123415946960449,
"step": 1280
},
{
"epoch": 2.7284345047923324,
"grad_norm": 0.5625,
"learning_rate": 4.363630398769667e-06,
"loss": 0.8218215107917786,
"step": 1282
},
{
"epoch": 2.7326943556975505,
"grad_norm": 0.6015625,
"learning_rate": 4.352525356034874e-06,
"loss": 0.8863515853881836,
"step": 1284
},
{
"epoch": 2.736954206602769,
"grad_norm": 0.494140625,
"learning_rate": 4.341588703119615e-06,
"loss": 0.8300585150718689,
"step": 1286
},
{
"epoch": 2.741214057507987,
"grad_norm": 1.96875,
"learning_rate": 4.330820680838162e-06,
"loss": 0.8769638538360596,
"step": 1288
},
{
"epoch": 2.7454739084132056,
"grad_norm": 0.5234375,
"learning_rate": 4.320221526291701e-06,
"loss": 0.8285037279129028,
"step": 1290
},
{
"epoch": 2.749733759318424,
"grad_norm": 0.81640625,
"learning_rate": 4.309791472863121e-06,
"loss": 0.8763640522956848,
"step": 1292
},
{
"epoch": 2.753993610223642,
"grad_norm": 0.6640625,
"learning_rate": 4.2995307502118735e-06,
"loss": 0.8772199749946594,
"step": 1294
},
{
"epoch": 2.7582534611288603,
"grad_norm": 1.0234375,
"learning_rate": 4.289439584268909e-06,
"loss": 0.9215362668037415,
"step": 1296
},
{
"epoch": 2.762513312034079,
"grad_norm": 0.6015625,
"learning_rate": 4.279518197231709e-06,
"loss": 0.781262218952179,
"step": 1298
},
{
"epoch": 2.7667731629392973,
"grad_norm": 2.09375,
"learning_rate": 4.2697668075593955e-06,
"loss": 0.8765016198158264,
"step": 1300
},
{
"epoch": 2.7710330138445154,
"grad_norm": 0.455078125,
"learning_rate": 4.260185629967912e-06,
"loss": 0.9362728595733643,
"step": 1302
},
{
"epoch": 2.775292864749734,
"grad_norm": 0.671875,
"learning_rate": 4.250774875425303e-06,
"loss": 0.9154821634292603,
"step": 1304
},
{
"epoch": 2.779552715654952,
"grad_norm": 0.5859375,
"learning_rate": 4.241534751147065e-06,
"loss": 0.8477409482002258,
"step": 1306
},
{
"epoch": 2.7838125665601705,
"grad_norm": 0.81640625,
"learning_rate": 4.232465460591584e-06,
"loss": 0.8393524289131165,
"step": 1308
},
{
"epoch": 2.7880724174653886,
"grad_norm": 0.7109375,
"learning_rate": 4.223567203455662e-06,
"loss": 1.008418083190918,
"step": 1310
},
{
"epoch": 2.792332268370607,
"grad_norm": 0.5859375,
"learning_rate": 4.214840175670106e-06,
"loss": 0.8442473411560059,
"step": 1312
},
{
"epoch": 2.796592119275825,
"grad_norm": 0.5703125,
"learning_rate": 4.2062845693954315e-06,
"loss": 0.8037950992584229,
"step": 1314
},
{
"epoch": 2.8008519701810437,
"grad_norm": 0.57421875,
"learning_rate": 4.197900573017613e-06,
"loss": 0.9449222683906555,
"step": 1316
},
{
"epoch": 2.8051118210862622,
"grad_norm": 1.71875,
"learning_rate": 4.189688371143952e-06,
"loss": 0.9002764225006104,
"step": 1318
},
{
"epoch": 2.8093716719914803,
"grad_norm": 0.74609375,
"learning_rate": 4.181648144599001e-06,
"loss": 0.9393887519836426,
"step": 1320
},
{
"epoch": 2.8136315228966984,
"grad_norm": 0.69921875,
"learning_rate": 4.17378007042059e-06,
"loss": 0.9079067707061768,
"step": 1322
},
{
"epoch": 2.817891373801917,
"grad_norm": 0.7109375,
"learning_rate": 4.166084321855919e-06,
"loss": 0.9096862077713013,
"step": 1324
},
{
"epoch": 2.8221512247071354,
"grad_norm": 1.1640625,
"learning_rate": 4.158561068357751e-06,
"loss": 0.9297478199005127,
"step": 1326
},
{
"epoch": 2.8264110756123535,
"grad_norm": 0.55859375,
"learning_rate": 4.151210475580683e-06,
"loss": 0.7916253805160522,
"step": 1328
},
{
"epoch": 2.830670926517572,
"grad_norm": 0.71875,
"learning_rate": 4.144032705377486e-06,
"loss": 0.7997951507568359,
"step": 1330
},
{
"epoch": 2.83493077742279,
"grad_norm": 0.5625,
"learning_rate": 4.137027915795555e-06,
"loss": 0.8519441485404968,
"step": 1332
},
{
"epoch": 2.8391906283280086,
"grad_norm": 0.5,
"learning_rate": 4.130196261073419e-06,
"loss": 0.8826906681060791,
"step": 1334
},
{
"epoch": 2.8434504792332267,
"grad_norm": 0.62109375,
"learning_rate": 4.123537891637352e-06,
"loss": 0.9272059798240662,
"step": 1336
},
{
"epoch": 2.847710330138445,
"grad_norm": 0.66015625,
"learning_rate": 4.117052954098055e-06,
"loss": 0.8484643697738647,
"step": 1338
},
{
"epoch": 2.8519701810436633,
"grad_norm": 0.92578125,
"learning_rate": 4.110741591247433e-06,
"loss": 0.8416329026222229,
"step": 1340
},
{
"epoch": 2.856230031948882,
"grad_norm": 1.21875,
"learning_rate": 4.1046039420554465e-06,
"loss": 0.7760175466537476,
"step": 1342
},
{
"epoch": 2.8604898828541003,
"grad_norm": 0.67578125,
"learning_rate": 4.0986401416670535e-06,
"loss": 0.8102884888648987,
"step": 1344
},
{
"epoch": 2.8647497337593184,
"grad_norm": 0.65625,
"learning_rate": 4.092850321399232e-06,
"loss": 0.8387068510055542,
"step": 1346
},
{
"epoch": 2.8690095846645365,
"grad_norm": 0.72265625,
"learning_rate": 4.0872346087380924e-06,
"loss": 0.8194290399551392,
"step": 1348
},
{
"epoch": 2.873269435569755,
"grad_norm": 0.6796875,
"learning_rate": 4.0817931273360686e-06,
"loss": 0.8201800584793091,
"step": 1350
},
{
"epoch": 2.8775292864749735,
"grad_norm": 0.51953125,
"learning_rate": 4.076525997009191e-06,
"loss": 0.8607783913612366,
"step": 1352
},
{
"epoch": 2.8817891373801916,
"grad_norm": 0.83984375,
"learning_rate": 4.0714333337344565e-06,
"loss": 0.8402998447418213,
"step": 1354
},
{
"epoch": 2.88604898828541,
"grad_norm": 0.52734375,
"learning_rate": 4.066515249647266e-06,
"loss": 0.8459066152572632,
"step": 1356
},
{
"epoch": 2.890308839190628,
"grad_norm": 0.63671875,
"learning_rate": 4.061771853038961e-06,
"loss": 0.8821372985839844,
"step": 1358
},
{
"epoch": 2.8945686900958467,
"grad_norm": 0.53515625,
"learning_rate": 4.05720324835444e-06,
"loss": 0.8806694746017456,
"step": 1360
},
{
"epoch": 2.8988285410010652,
"grad_norm": 0.7421875,
"learning_rate": 4.052809536189854e-06,
"loss": 0.9144023656845093,
"step": 1362
},
{
"epoch": 2.9030883919062833,
"grad_norm": 1.25,
"learning_rate": 4.048590813290395e-06,
"loss": 0.8206741213798523,
"step": 1364
},
{
"epoch": 2.9073482428115014,
"grad_norm": 0.67578125,
"learning_rate": 4.044547172548163e-06,
"loss": 0.8178682923316956,
"step": 1366
},
{
"epoch": 2.91160809371672,
"grad_norm": 0.66015625,
"learning_rate": 4.040678703000123e-06,
"loss": 0.8602153658866882,
"step": 1368
},
{
"epoch": 2.9158679446219384,
"grad_norm": 0.53515625,
"learning_rate": 4.036985489826145e-06,
"loss": 0.8640646934509277,
"step": 1370
},
{
"epoch": 2.9201277955271565,
"grad_norm": 0.494140625,
"learning_rate": 4.033467614347126e-06,
"loss": 0.8199517130851746,
"step": 1372
},
{
"epoch": 2.924387646432375,
"grad_norm": 0.462890625,
"learning_rate": 4.030125154023201e-06,
"loss": 0.8197891712188721,
"step": 1374
},
{
"epoch": 2.928647497337593,
"grad_norm": 0.6015625,
"learning_rate": 4.026958182452037e-06,
"loss": 0.9104003310203552,
"step": 1376
},
{
"epoch": 2.9329073482428116,
"grad_norm": 0.78125,
"learning_rate": 4.023966769367212e-06,
"loss": 0.9194601774215698,
"step": 1378
},
{
"epoch": 2.9371671991480297,
"grad_norm": 0.6875,
"learning_rate": 4.021150980636679e-06,
"loss": 0.8961591720581055,
"step": 1380
},
{
"epoch": 2.9414270500532482,
"grad_norm": 1.1640625,
"learning_rate": 4.01851087826132e-06,
"loss": 0.9252373576164246,
"step": 1382
},
{
"epoch": 2.9456869009584663,
"grad_norm": 0.703125,
"learning_rate": 4.016046520373573e-06,
"loss": 0.8721043467521667,
"step": 1384
},
{
"epoch": 2.949946751863685,
"grad_norm": 0.55859375,
"learning_rate": 4.013757961236162e-06,
"loss": 0.8307064771652222,
"step": 1386
},
{
"epoch": 2.9542066027689033,
"grad_norm": 0.87109375,
"learning_rate": 4.011645251240889e-06,
"loss": 0.8707568049430847,
"step": 1388
},
{
"epoch": 2.9584664536741214,
"grad_norm": 0.69921875,
"learning_rate": 4.009708436907538e-06,
"loss": 0.826111376285553,
"step": 1390
},
{
"epoch": 2.9627263045793395,
"grad_norm": 0.7734375,
"learning_rate": 4.007947560882844e-06,
"loss": 0.8147805333137512,
"step": 1392
},
{
"epoch": 2.966986155484558,
"grad_norm": 5.5625,
"learning_rate": 4.006362661939552e-06,
"loss": 0.9187301993370056,
"step": 1394
},
{
"epoch": 2.9712460063897765,
"grad_norm": 0.94140625,
"learning_rate": 4.0049537749755685e-06,
"loss": 0.7911329865455627,
"step": 1396
},
{
"epoch": 2.9755058572949946,
"grad_norm": 0.60546875,
"learning_rate": 4.003720931013188e-06,
"loss": 0.8627235293388367,
"step": 1398
},
{
"epoch": 2.979765708200213,
"grad_norm": 1.75,
"learning_rate": 4.002664157198416e-06,
"loss": 0.7859567999839783,
"step": 1400
},
{
"epoch": 2.984025559105431,
"grad_norm": 1.5703125,
"learning_rate": 4.001783476800364e-06,
"loss": 0.8008852005004883,
"step": 1402
},
{
"epoch": 2.9882854100106497,
"grad_norm": 0.78515625,
"learning_rate": 4.001078909210742e-06,
"loss": 0.8270421624183655,
"step": 1404
},
{
"epoch": 2.992545260915868,
"grad_norm": 0.6171875,
"learning_rate": 4.000550469943431e-06,
"loss": 0.8458138704299927,
"step": 1406
},
{
"epoch": 2.9968051118210863,
"grad_norm": 0.5546875,
"learning_rate": 4.0001981706341416e-06,
"loss": 0.8161488771438599,
"step": 1408
},
{
"epoch": 3.0,
"grad_norm": 1.1015625,
"learning_rate": 4.000022019040156e-06,
"loss": 0.8455473780632019,
"step": 1410
},
{
"epoch": 3.0,
"step": 1410,
"total_flos": 4.4194007019526554e+18,
"train_loss": 1.0514326345836018,
"train_runtime": 11229.2105,
"train_samples_per_second": 4.014,
"train_steps_per_second": 0.126
}
],
"logging_steps": 2,
"max_steps": 1410,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 99999,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 4.4194007019526554e+18,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}