9b-37 / trainer_state.json
furproxy's picture
Upload folder using huggingface_hub
92fd788 verified
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 3.0,
"eval_steps": 500,
"global_step": 1410,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.00425531914893617,
"grad_norm": 1.125,
"learning_rate": 8.450704225352114e-07,
"loss": 1.4721390008926392,
"step": 2
},
{
"epoch": 0.00851063829787234,
"grad_norm": 1.046875,
"learning_rate": 2.535211267605634e-06,
"loss": 1.9258784055709839,
"step": 4
},
{
"epoch": 0.01276595744680851,
"grad_norm": 5.78125,
"learning_rate": 4.225352112676057e-06,
"loss": 1.9619797468185425,
"step": 6
},
{
"epoch": 0.01702127659574468,
"grad_norm": 2.5625,
"learning_rate": 5.915492957746478e-06,
"loss": 1.6346324682235718,
"step": 8
},
{
"epoch": 0.02127659574468085,
"grad_norm": 1.21875,
"learning_rate": 7.6056338028169015e-06,
"loss": 1.9556920528411865,
"step": 10
},
{
"epoch": 0.02553191489361702,
"grad_norm": 0.87109375,
"learning_rate": 9.295774647887323e-06,
"loss": 1.470306634902954,
"step": 12
},
{
"epoch": 0.029787234042553193,
"grad_norm": 1.3984375,
"learning_rate": 1.0985915492957748e-05,
"loss": 2.068174362182617,
"step": 14
},
{
"epoch": 0.03404255319148936,
"grad_norm": 1.03125,
"learning_rate": 1.267605633802817e-05,
"loss": 1.903701663017273,
"step": 16
},
{
"epoch": 0.03829787234042553,
"grad_norm": 1.0703125,
"learning_rate": 1.436619718309859e-05,
"loss": 1.765465259552002,
"step": 18
},
{
"epoch": 0.0425531914893617,
"grad_norm": 0.87890625,
"learning_rate": 1.6056338028169014e-05,
"loss": 1.9205501079559326,
"step": 20
},
{
"epoch": 0.04680851063829787,
"grad_norm": 0.8359375,
"learning_rate": 1.7746478873239435e-05,
"loss": 1.6455962657928467,
"step": 22
},
{
"epoch": 0.05106382978723404,
"grad_norm": 0.83984375,
"learning_rate": 1.943661971830986e-05,
"loss": 1.6872483491897583,
"step": 24
},
{
"epoch": 0.05531914893617021,
"grad_norm": 2.234375,
"learning_rate": 2.112676056338028e-05,
"loss": 1.4278151988983154,
"step": 26
},
{
"epoch": 0.059574468085106386,
"grad_norm": 0.76171875,
"learning_rate": 2.2816901408450703e-05,
"loss": 1.7643554210662842,
"step": 28
},
{
"epoch": 0.06382978723404255,
"grad_norm": 0.71484375,
"learning_rate": 2.4507042253521128e-05,
"loss": 1.686973214149475,
"step": 30
},
{
"epoch": 0.06808510638297872,
"grad_norm": 0.63671875,
"learning_rate": 2.619718309859155e-05,
"loss": 1.4633193016052246,
"step": 32
},
{
"epoch": 0.07234042553191489,
"grad_norm": 0.60546875,
"learning_rate": 2.788732394366197e-05,
"loss": 1.6693249940872192,
"step": 34
},
{
"epoch": 0.07659574468085106,
"grad_norm": 0.55859375,
"learning_rate": 2.9577464788732395e-05,
"loss": 1.657246708869934,
"step": 36
},
{
"epoch": 0.08085106382978724,
"grad_norm": 1.0703125,
"learning_rate": 3.126760563380282e-05,
"loss": 1.882036805152893,
"step": 38
},
{
"epoch": 0.0851063829787234,
"grad_norm": 0.765625,
"learning_rate": 3.2957746478873245e-05,
"loss": 1.618768572807312,
"step": 40
},
{
"epoch": 0.08936170212765958,
"grad_norm": 1.0078125,
"learning_rate": 3.4647887323943666e-05,
"loss": 1.29584801197052,
"step": 42
},
{
"epoch": 0.09361702127659574,
"grad_norm": 0.5546875,
"learning_rate": 3.633802816901408e-05,
"loss": 1.5222253799438477,
"step": 44
},
{
"epoch": 0.09787234042553192,
"grad_norm": 1.2890625,
"learning_rate": 3.802816901408451e-05,
"loss": 1.3532958030700684,
"step": 46
},
{
"epoch": 0.10212765957446808,
"grad_norm": 1.1640625,
"learning_rate": 3.971830985915493e-05,
"loss": 1.5933119058609009,
"step": 48
},
{
"epoch": 0.10638297872340426,
"grad_norm": 1.671875,
"learning_rate": 4.140845070422535e-05,
"loss": 1.6478880643844604,
"step": 50
},
{
"epoch": 0.11063829787234042,
"grad_norm": 0.578125,
"learning_rate": 4.309859154929578e-05,
"loss": 1.3426867723464966,
"step": 52
},
{
"epoch": 0.1148936170212766,
"grad_norm": 0.5078125,
"learning_rate": 4.47887323943662e-05,
"loss": 1.4301996231079102,
"step": 54
},
{
"epoch": 0.11914893617021277,
"grad_norm": 0.44140625,
"learning_rate": 4.6478873239436617e-05,
"loss": 1.2384978532791138,
"step": 56
},
{
"epoch": 0.12340425531914893,
"grad_norm": 2.328125,
"learning_rate": 4.8169014084507045e-05,
"loss": 1.4298828840255737,
"step": 58
},
{
"epoch": 0.1276595744680851,
"grad_norm": 0.71875,
"learning_rate": 4.9859154929577466e-05,
"loss": 1.3327827453613281,
"step": 60
},
{
"epoch": 0.13191489361702127,
"grad_norm": 0.59375,
"learning_rate": 5.154929577464789e-05,
"loss": 1.281138300895691,
"step": 62
},
{
"epoch": 0.13617021276595745,
"grad_norm": 1.0703125,
"learning_rate": 5.3239436619718316e-05,
"loss": 1.064106822013855,
"step": 64
},
{
"epoch": 0.14042553191489363,
"grad_norm": 0.314453125,
"learning_rate": 5.492957746478874e-05,
"loss": 1.4334735870361328,
"step": 66
},
{
"epoch": 0.14468085106382977,
"grad_norm": 0.78125,
"learning_rate": 5.661971830985915e-05,
"loss": 1.3736847639083862,
"step": 68
},
{
"epoch": 0.14893617021276595,
"grad_norm": 1.2578125,
"learning_rate": 5.830985915492958e-05,
"loss": 1.3596861362457275,
"step": 70
},
{
"epoch": 0.15319148936170213,
"grad_norm": 0.671875,
"learning_rate": 6e-05,
"loss": 1.4395864009857178,
"step": 72
},
{
"epoch": 0.1574468085106383,
"grad_norm": 0.84765625,
"learning_rate": 5.9999735771881775e-05,
"loss": 1.5454157590866089,
"step": 74
},
{
"epoch": 0.16170212765957448,
"grad_norm": 0.56640625,
"learning_rate": 5.99989430933451e-05,
"loss": 1.3853117227554321,
"step": 76
},
{
"epoch": 0.16595744680851063,
"grad_norm": 0.408203125,
"learning_rate": 5.9997621981843994e-05,
"loss": 1.2754864692687988,
"step": 78
},
{
"epoch": 0.1702127659574468,
"grad_norm": 0.92578125,
"learning_rate": 5.999577246646804e-05,
"loss": 1.5860013961791992,
"step": 80
},
{
"epoch": 0.17446808510638298,
"grad_norm": 0.5859375,
"learning_rate": 5.99933945879417e-05,
"loss": 1.237565517425537,
"step": 82
},
{
"epoch": 0.17872340425531916,
"grad_norm": 2.171875,
"learning_rate": 5.999048839862352e-05,
"loss": 1.2814902067184448,
"step": 84
},
{
"epoch": 0.1829787234042553,
"grad_norm": 1.4296875,
"learning_rate": 5.998705396250493e-05,
"loss": 1.3300844430923462,
"step": 86
},
{
"epoch": 0.18723404255319148,
"grad_norm": 0.462890625,
"learning_rate": 5.998309135520878e-05,
"loss": 1.3447200059890747,
"step": 88
},
{
"epoch": 0.19148936170212766,
"grad_norm": 0.55078125,
"learning_rate": 5.997860066398778e-05,
"loss": 1.3371561765670776,
"step": 90
},
{
"epoch": 0.19574468085106383,
"grad_norm": 0.6328125,
"learning_rate": 5.997358198772249e-05,
"loss": 1.3265268802642822,
"step": 92
},
{
"epoch": 0.2,
"grad_norm": 0.51953125,
"learning_rate": 5.9968035436919206e-05,
"loss": 1.1992994546890259,
"step": 94
},
{
"epoch": 0.20425531914893616,
"grad_norm": 0.5390625,
"learning_rate": 5.996196113370748e-05,
"loss": 1.2414522171020508,
"step": 96
},
{
"epoch": 0.20851063829787234,
"grad_norm": 0.4921875,
"learning_rate": 5.9955359211837465e-05,
"loss": 1.314779281616211,
"step": 98
},
{
"epoch": 0.2127659574468085,
"grad_norm": 0.6484375,
"learning_rate": 5.994822981667691e-05,
"loss": 1.2792837619781494,
"step": 100
},
{
"epoch": 0.2170212765957447,
"grad_norm": 0.494140625,
"learning_rate": 5.994057310520807e-05,
"loss": 1.234731912612915,
"step": 102
},
{
"epoch": 0.22127659574468084,
"grad_norm": 0.859375,
"learning_rate": 5.993238924602414e-05,
"loss": 1.1855965852737427,
"step": 104
},
{
"epoch": 0.225531914893617,
"grad_norm": 0.427734375,
"learning_rate": 5.9923678419325616e-05,
"loss": 1.3185398578643799,
"step": 106
},
{
"epoch": 0.2297872340425532,
"grad_norm": 0.8359375,
"learning_rate": 5.9914440816916236e-05,
"loss": 1.3415206670761108,
"step": 108
},
{
"epoch": 0.23404255319148937,
"grad_norm": 1.2421875,
"learning_rate": 5.990467664219887e-05,
"loss": 1.7355936765670776,
"step": 110
},
{
"epoch": 0.23829787234042554,
"grad_norm": 0.474609375,
"learning_rate": 5.989438611017101e-05,
"loss": 1.299019455909729,
"step": 112
},
{
"epoch": 0.2425531914893617,
"grad_norm": 0.73046875,
"learning_rate": 5.9883569447419946e-05,
"loss": 1.2734484672546387,
"step": 114
},
{
"epoch": 0.24680851063829787,
"grad_norm": 0.83203125,
"learning_rate": 5.98722268921179e-05,
"loss": 1.3722704648971558,
"step": 116
},
{
"epoch": 0.251063829787234,
"grad_norm": 1.4453125,
"learning_rate": 5.9860358694016695e-05,
"loss": 1.241133451461792,
"step": 118
},
{
"epoch": 0.2553191489361702,
"grad_norm": 0.451171875,
"learning_rate": 5.984796511444231e-05,
"loss": 1.2068897485733032,
"step": 120
},
{
"epoch": 0.25957446808510637,
"grad_norm": 0.66796875,
"learning_rate": 5.983504642628911e-05,
"loss": 1.0534322261810303,
"step": 122
},
{
"epoch": 0.26382978723404255,
"grad_norm": 0.5078125,
"learning_rate": 5.9821602914013794e-05,
"loss": 1.306460976600647,
"step": 124
},
{
"epoch": 0.2680851063829787,
"grad_norm": 0.5625,
"learning_rate": 5.9807634873629205e-05,
"loss": 1.3372706174850464,
"step": 126
},
{
"epoch": 0.2723404255319149,
"grad_norm": 1.546875,
"learning_rate": 5.979314261269775e-05,
"loss": 1.3372623920440674,
"step": 128
},
{
"epoch": 0.2765957446808511,
"grad_norm": 3.953125,
"learning_rate": 5.9778126450324674e-05,
"loss": 1.303137183189392,
"step": 130
},
{
"epoch": 0.28085106382978725,
"grad_norm": 0.546875,
"learning_rate": 5.9762586717151e-05,
"loss": 1.2485641241073608,
"step": 132
},
{
"epoch": 0.2851063829787234,
"grad_norm": 0.45703125,
"learning_rate": 5.9746523755346284e-05,
"loss": 1.2784473896026611,
"step": 134
},
{
"epoch": 0.28936170212765955,
"grad_norm": 0.9375,
"learning_rate": 5.972993791860101e-05,
"loss": 1.3001835346221924,
"step": 136
},
{
"epoch": 0.2936170212765957,
"grad_norm": 1.2734375,
"learning_rate": 5.97128295721189e-05,
"loss": 1.2696309089660645,
"step": 138
},
{
"epoch": 0.2978723404255319,
"grad_norm": 0.796875,
"learning_rate": 5.969519909260879e-05,
"loss": 1.4176753759384155,
"step": 140
},
{
"epoch": 0.3021276595744681,
"grad_norm": 0.5546875,
"learning_rate": 5.967704686827641e-05,
"loss": 1.3172787427902222,
"step": 142
},
{
"epoch": 0.30638297872340425,
"grad_norm": 14.0,
"learning_rate": 5.965837329881574e-05,
"loss": 1.4879260063171387,
"step": 144
},
{
"epoch": 0.31063829787234043,
"grad_norm": 0.4453125,
"learning_rate": 5.9639178795400295e-05,
"loss": 1.1774471998214722,
"step": 146
},
{
"epoch": 0.3148936170212766,
"grad_norm": 0.71875,
"learning_rate": 5.9619463780674034e-05,
"loss": 1.2439987659454346,
"step": 148
},
{
"epoch": 0.3191489361702128,
"grad_norm": 0.48828125,
"learning_rate": 5.959922868874206e-05,
"loss": 1.205959439277649,
"step": 150
},
{
"epoch": 0.32340425531914896,
"grad_norm": 1.1796875,
"learning_rate": 5.9578473965161075e-05,
"loss": 1.330673098564148,
"step": 152
},
{
"epoch": 0.3276595744680851,
"grad_norm": 0.455078125,
"learning_rate": 5.9557200066929534e-05,
"loss": 1.2126266956329346,
"step": 154
},
{
"epoch": 0.33191489361702126,
"grad_norm": 0.4140625,
"learning_rate": 5.95354074624776e-05,
"loss": 1.2708547115325928,
"step": 156
},
{
"epoch": 0.33617021276595743,
"grad_norm": 0.6953125,
"learning_rate": 5.951309663165686e-05,
"loss": 1.2793720960617065,
"step": 158
},
{
"epoch": 0.3404255319148936,
"grad_norm": 0.310546875,
"learning_rate": 5.9490268065729724e-05,
"loss": 1.2078324556350708,
"step": 160
},
{
"epoch": 0.3446808510638298,
"grad_norm": 0.53515625,
"learning_rate": 5.946692226735858e-05,
"loss": 1.2282007932662964,
"step": 162
},
{
"epoch": 0.34893617021276596,
"grad_norm": 0.74609375,
"learning_rate": 5.944305975059482e-05,
"loss": 1.2256958484649658,
"step": 164
},
{
"epoch": 0.35319148936170214,
"grad_norm": 0.56640625,
"learning_rate": 5.9418681040867445e-05,
"loss": 1.2895530462265015,
"step": 166
},
{
"epoch": 0.3574468085106383,
"grad_norm": 0.50390625,
"learning_rate": 5.939378667497147e-05,
"loss": 1.1319845914840698,
"step": 168
},
{
"epoch": 0.3617021276595745,
"grad_norm": 0.828125,
"learning_rate": 5.9368377201056195e-05,
"loss": 1.3089065551757812,
"step": 170
},
{
"epoch": 0.3659574468085106,
"grad_norm": 0.474609375,
"learning_rate": 5.934245317861308e-05,
"loss": 1.2677334547042847,
"step": 172
},
{
"epoch": 0.3702127659574468,
"grad_norm": 0.45703125,
"learning_rate": 5.9316015178463446e-05,
"loss": 1.2161953449249268,
"step": 174
},
{
"epoch": 0.37446808510638296,
"grad_norm": 0.890625,
"learning_rate": 5.928906378274585e-05,
"loss": 1.4091310501098633,
"step": 176
},
{
"epoch": 0.37872340425531914,
"grad_norm": 0.8984375,
"learning_rate": 5.9261599584903375e-05,
"loss": 1.3989202976226807,
"step": 178
},
{
"epoch": 0.3829787234042553,
"grad_norm": 0.55859375,
"learning_rate": 5.923362318967043e-05,
"loss": 1.1342294216156006,
"step": 180
},
{
"epoch": 0.3872340425531915,
"grad_norm": 0.390625,
"learning_rate": 5.920513521305955e-05,
"loss": 1.2139039039611816,
"step": 182
},
{
"epoch": 0.39148936170212767,
"grad_norm": 0.76953125,
"learning_rate": 5.9176136282347777e-05,
"loss": 1.12371027469635,
"step": 184
},
{
"epoch": 0.39574468085106385,
"grad_norm": 0.84765625,
"learning_rate": 5.914662703606285e-05,
"loss": 1.3441473245620728,
"step": 186
},
{
"epoch": 0.4,
"grad_norm": 0.9140625,
"learning_rate": 5.911660812396916e-05,
"loss": 1.2006664276123047,
"step": 188
},
{
"epoch": 0.40425531914893614,
"grad_norm": 0.5,
"learning_rate": 5.90860802070534e-05,
"loss": 1.2136142253875732,
"step": 190
},
{
"epoch": 0.4085106382978723,
"grad_norm": 0.59375,
"learning_rate": 5.905504395751011e-05,
"loss": 1.2749568223953247,
"step": 192
},
{
"epoch": 0.4127659574468085,
"grad_norm": 0.40234375,
"learning_rate": 5.9023500058726754e-05,
"loss": 1.369842529296875,
"step": 194
},
{
"epoch": 0.41702127659574467,
"grad_norm": 0.6171875,
"learning_rate": 5.899144920526876e-05,
"loss": 1.2179538011550903,
"step": 196
},
{
"epoch": 0.42127659574468085,
"grad_norm": 0.765625,
"learning_rate": 5.895889210286416e-05,
"loss": 1.3043186664581299,
"step": 198
},
{
"epoch": 0.425531914893617,
"grad_norm": 0.80859375,
"learning_rate": 5.8925829468388156e-05,
"loss": 1.309710144996643,
"step": 200
},
{
"epoch": 0.4297872340425532,
"grad_norm": 0.36328125,
"learning_rate": 5.8892262029847186e-05,
"loss": 1.233137607574463,
"step": 202
},
{
"epoch": 0.4340425531914894,
"grad_norm": 0.55859375,
"learning_rate": 5.885819052636303e-05,
"loss": 1.1934456825256348,
"step": 204
},
{
"epoch": 0.43829787234042555,
"grad_norm": 0.625,
"learning_rate": 5.882361570815645e-05,
"loss": 1.2160348892211914,
"step": 206
},
{
"epoch": 0.4425531914893617,
"grad_norm": 0.87890625,
"learning_rate": 5.8788538336530715e-05,
"loss": 1.2657216787338257,
"step": 208
},
{
"epoch": 0.44680851063829785,
"grad_norm": 0.58203125,
"learning_rate": 5.8752959183854816e-05,
"loss": 1.2517626285552979,
"step": 210
},
{
"epoch": 0.451063829787234,
"grad_norm": 0.66796875,
"learning_rate": 5.871687903354646e-05,
"loss": 1.1521042585372925,
"step": 212
},
{
"epoch": 0.4553191489361702,
"grad_norm": 0.61328125,
"learning_rate": 5.8680298680054824e-05,
"loss": 1.2568317651748657,
"step": 214
},
{
"epoch": 0.4595744680851064,
"grad_norm": 1.46875,
"learning_rate": 5.864321892884309e-05,
"loss": 1.2429876327514648,
"step": 216
},
{
"epoch": 0.46382978723404256,
"grad_norm": 0.64453125,
"learning_rate": 5.860564059637066e-05,
"loss": 1.327577829360962,
"step": 218
},
{
"epoch": 0.46808510638297873,
"grad_norm": 3.3125,
"learning_rate": 5.85675645100752e-05,
"loss": 1.416573166847229,
"step": 220
},
{
"epoch": 0.4723404255319149,
"grad_norm": 0.828125,
"learning_rate": 5.852899150835445e-05,
"loss": 1.241747498512268,
"step": 222
},
{
"epoch": 0.4765957446808511,
"grad_norm": 0.48046875,
"learning_rate": 5.848992244054768e-05,
"loss": 1.2418628931045532,
"step": 224
},
{
"epoch": 0.4808510638297872,
"grad_norm": 0.58203125,
"learning_rate": 5.845035816691711e-05,
"loss": 1.1999211311340332,
"step": 226
},
{
"epoch": 0.4851063829787234,
"grad_norm": 0.71875,
"learning_rate": 5.841029955862885e-05,
"loss": 1.4046475887298584,
"step": 228
},
{
"epoch": 0.48936170212765956,
"grad_norm": 0.5234375,
"learning_rate": 5.836974749773382e-05,
"loss": 1.2456966638565063,
"step": 230
},
{
"epoch": 0.49361702127659574,
"grad_norm": 0.578125,
"learning_rate": 5.83287028771482e-05,
"loss": 1.2392256259918213,
"step": 232
},
{
"epoch": 0.4978723404255319,
"grad_norm": 0.6171875,
"learning_rate": 5.828716660063395e-05,
"loss": 1.3959895372390747,
"step": 234
},
{
"epoch": 0.502127659574468,
"grad_norm": 0.482421875,
"learning_rate": 5.824513958277871e-05,
"loss": 1.2444478273391724,
"step": 236
},
{
"epoch": 0.5063829787234042,
"grad_norm": 0.6328125,
"learning_rate": 5.8202622748975817e-05,
"loss": 1.294707179069519,
"step": 238
},
{
"epoch": 0.5106382978723404,
"grad_norm": 1.5703125,
"learning_rate": 5.815961703540386e-05,
"loss": 1.3415395021438599,
"step": 240
},
{
"epoch": 0.5148936170212766,
"grad_norm": 1.328125,
"learning_rate": 5.811612338900606e-05,
"loss": 1.1040756702423096,
"step": 242
},
{
"epoch": 0.5191489361702127,
"grad_norm": 0.26953125,
"learning_rate": 5.807214276746943e-05,
"loss": 1.1671684980392456,
"step": 244
},
{
"epoch": 0.5234042553191489,
"grad_norm": 0.474609375,
"learning_rate": 5.802767613920375e-05,
"loss": 1.1621744632720947,
"step": 246
},
{
"epoch": 0.5276595744680851,
"grad_norm": 0.6171875,
"learning_rate": 5.7982724483320105e-05,
"loss": 1.2460886240005493,
"step": 248
},
{
"epoch": 0.5319148936170213,
"grad_norm": 0.734375,
"learning_rate": 5.793728878960947e-05,
"loss": 1.3446158170700073,
"step": 250
},
{
"epoch": 0.5361702127659574,
"grad_norm": 0.35546875,
"learning_rate": 5.789137005852084e-05,
"loss": 1.4579079151153564,
"step": 252
},
{
"epoch": 0.5404255319148936,
"grad_norm": 0.74609375,
"learning_rate": 5.7844969301139194e-05,
"loss": 1.4334136247634888,
"step": 254
},
{
"epoch": 0.5446808510638298,
"grad_norm": 1.421875,
"learning_rate": 5.779808753916325e-05,
"loss": 1.6397161483764648,
"step": 256
},
{
"epoch": 0.548936170212766,
"grad_norm": 0.6953125,
"learning_rate": 5.7750725804883034e-05,
"loss": 1.301457166671753,
"step": 258
},
{
"epoch": 0.5531914893617021,
"grad_norm": 0.7578125,
"learning_rate": 5.770288514115698e-05,
"loss": 1.2889872789382935,
"step": 260
},
{
"epoch": 0.5574468085106383,
"grad_norm": 0.765625,
"learning_rate": 5.765456660138919e-05,
"loss": 1.2275917530059814,
"step": 262
},
{
"epoch": 0.5617021276595745,
"grad_norm": 0.546875,
"learning_rate": 5.760577124950603e-05,
"loss": 1.2264875173568726,
"step": 264
},
{
"epoch": 0.5659574468085107,
"grad_norm": 0.75,
"learning_rate": 5.7556500159932835e-05,
"loss": 1.3288713693618774,
"step": 266
},
{
"epoch": 0.5702127659574469,
"grad_norm": 0.4609375,
"learning_rate": 5.750675441757024e-05,
"loss": 1.1868747472763062,
"step": 268
},
{
"epoch": 0.574468085106383,
"grad_norm": 1.53125,
"learning_rate": 5.7456535117770204e-05,
"loss": 1.3627235889434814,
"step": 270
},
{
"epoch": 0.5787234042553191,
"grad_norm": 0.73828125,
"learning_rate": 5.7405843366312e-05,
"loss": 1.2165353298187256,
"step": 272
},
{
"epoch": 0.5829787234042553,
"grad_norm": 0.6484375,
"learning_rate": 5.735468027937782e-05,
"loss": 1.3962359428405762,
"step": 274
},
{
"epoch": 0.5872340425531914,
"grad_norm": 0.49609375,
"learning_rate": 5.730304698352815e-05,
"loss": 1.2368425130844116,
"step": 276
},
{
"epoch": 0.5914893617021276,
"grad_norm": 0.65625,
"learning_rate": 5.7250944615677036e-05,
"loss": 1.2761774063110352,
"step": 278
},
{
"epoch": 0.5957446808510638,
"grad_norm": 0.6328125,
"learning_rate": 5.719837432306707e-05,
"loss": 1.146309494972229,
"step": 280
},
{
"epoch": 0.6,
"grad_norm": 1.0859375,
"learning_rate": 5.7145337263244e-05,
"loss": 1.3055988550186157,
"step": 282
},
{
"epoch": 0.6042553191489362,
"grad_norm": 0.6875,
"learning_rate": 5.7091834604031386e-05,
"loss": 1.2853789329528809,
"step": 284
},
{
"epoch": 0.6085106382978723,
"grad_norm": 0.5078125,
"learning_rate": 5.703786752350482e-05,
"loss": 1.2626879215240479,
"step": 286
},
{
"epoch": 0.6127659574468085,
"grad_norm": 0.8515625,
"learning_rate": 5.698343720996596e-05,
"loss": 1.0880780220031738,
"step": 288
},
{
"epoch": 0.6170212765957447,
"grad_norm": 0.5859375,
"learning_rate": 5.6928544861916416e-05,
"loss": 1.4804942607879639,
"step": 290
},
{
"epoch": 0.6212765957446809,
"grad_norm": 0.482421875,
"learning_rate": 5.687319168803136e-05,
"loss": 1.2268519401550293,
"step": 292
},
{
"epoch": 0.625531914893617,
"grad_norm": 0.6015625,
"learning_rate": 5.681737890713286e-05,
"loss": 1.2773425579071045,
"step": 294
},
{
"epoch": 0.6297872340425532,
"grad_norm": 0.40625,
"learning_rate": 5.676110774816309e-05,
"loss": 1.3065530061721802,
"step": 296
},
{
"epoch": 0.6340425531914894,
"grad_norm": 0.50390625,
"learning_rate": 5.670437945015725e-05,
"loss": 1.2505621910095215,
"step": 298
},
{
"epoch": 0.6382978723404256,
"grad_norm": 1.515625,
"learning_rate": 5.6647195262216296e-05,
"loss": 1.2247607707977295,
"step": 300
},
{
"epoch": 0.6425531914893617,
"grad_norm": 0.296875,
"learning_rate": 5.658955644347944e-05,
"loss": 1.3000612258911133,
"step": 302
},
{
"epoch": 0.6468085106382979,
"grad_norm": 0.50390625,
"learning_rate": 5.653146426309637e-05,
"loss": 1.2301831245422363,
"step": 304
},
{
"epoch": 0.6510638297872341,
"grad_norm": 2.875,
"learning_rate": 5.64729200001994e-05,
"loss": 1.23581862449646,
"step": 306
},
{
"epoch": 0.6553191489361702,
"grad_norm": 0.68359375,
"learning_rate": 5.6413924943875225e-05,
"loss": 1.2415351867675781,
"step": 308
},
{
"epoch": 0.6595744680851063,
"grad_norm": 0.44921875,
"learning_rate": 5.635448039313658e-05,
"loss": 1.054430365562439,
"step": 310
},
{
"epoch": 0.6638297872340425,
"grad_norm": 0.640625,
"learning_rate": 5.62945876568936e-05,
"loss": 1.1587709188461304,
"step": 312
},
{
"epoch": 0.6680851063829787,
"grad_norm": 0.61328125,
"learning_rate": 5.623424805392504e-05,
"loss": 1.2375378608703613,
"step": 314
},
{
"epoch": 0.6723404255319149,
"grad_norm": 0.44140625,
"learning_rate": 5.617346291284922e-05,
"loss": 1.2096360921859741,
"step": 316
},
{
"epoch": 0.676595744680851,
"grad_norm": 3.484375,
"learning_rate": 5.611223357209474e-05,
"loss": 1.4315375089645386,
"step": 318
},
{
"epoch": 0.6808510638297872,
"grad_norm": 0.515625,
"learning_rate": 5.6050561379871077e-05,
"loss": 1.2253354787826538,
"step": 320
},
{
"epoch": 0.6851063829787234,
"grad_norm": 0.431640625,
"learning_rate": 5.5988447694138834e-05,
"loss": 1.1945621967315674,
"step": 322
},
{
"epoch": 0.6893617021276596,
"grad_norm": 0.72265625,
"learning_rate": 5.5925893882579864e-05,
"loss": 1.2644530534744263,
"step": 324
},
{
"epoch": 0.6936170212765957,
"grad_norm": 0.9921875,
"learning_rate": 5.5862901322567175e-05,
"loss": 1.2183250188827515,
"step": 326
},
{
"epoch": 0.6978723404255319,
"grad_norm": 1.2421875,
"learning_rate": 5.5799471401134543e-05,
"loss": 1.4243056774139404,
"step": 328
},
{
"epoch": 0.7021276595744681,
"grad_norm": 0.671875,
"learning_rate": 5.5735605514946046e-05,
"loss": 1.2356775999069214,
"step": 330
},
{
"epoch": 0.7063829787234043,
"grad_norm": 2.078125,
"learning_rate": 5.567130507026527e-05,
"loss": 1.245442509651184,
"step": 332
},
{
"epoch": 0.7106382978723405,
"grad_norm": 4.28125,
"learning_rate": 5.560657148292432e-05,
"loss": 1.0984729528427124,
"step": 334
},
{
"epoch": 0.7148936170212766,
"grad_norm": 0.361328125,
"learning_rate": 5.554140617829271e-05,
"loss": 1.2282965183258057,
"step": 336
},
{
"epoch": 0.7191489361702128,
"grad_norm": 0.875,
"learning_rate": 5.547581059124591e-05,
"loss": 1.3061587810516357,
"step": 338
},
{
"epoch": 0.723404255319149,
"grad_norm": 0.85546875,
"learning_rate": 5.540978616613381e-05,
"loss": 1.2117385864257812,
"step": 340
},
{
"epoch": 0.7276595744680852,
"grad_norm": 0.78515625,
"learning_rate": 5.534333435674889e-05,
"loss": 1.105977177619934,
"step": 342
},
{
"epoch": 0.7319148936170212,
"grad_norm": 0.462890625,
"learning_rate": 5.527645662629417e-05,
"loss": 1.1953120231628418,
"step": 344
},
{
"epoch": 0.7361702127659574,
"grad_norm": 0.59375,
"learning_rate": 5.520915444735106e-05,
"loss": 1.2646780014038086,
"step": 346
},
{
"epoch": 0.7404255319148936,
"grad_norm": 0.390625,
"learning_rate": 5.514142930184689e-05,
"loss": 1.2484480142593384,
"step": 348
},
{
"epoch": 0.7446808510638298,
"grad_norm": 1.8203125,
"learning_rate": 5.5073282681022354e-05,
"loss": 1.274402379989624,
"step": 350
},
{
"epoch": 0.7489361702127659,
"grad_norm": 0.37890625,
"learning_rate": 5.5004716085398515e-05,
"loss": 1.182061791419983,
"step": 352
},
{
"epoch": 0.7531914893617021,
"grad_norm": 1.015625,
"learning_rate": 5.493573102474395e-05,
"loss": 1.1438305377960205,
"step": 354
},
{
"epoch": 0.7574468085106383,
"grad_norm": 0.80078125,
"learning_rate": 5.486632901804137e-05,
"loss": 1.2968688011169434,
"step": 356
},
{
"epoch": 0.7617021276595745,
"grad_norm": 0.42578125,
"learning_rate": 5.4796511593454254e-05,
"loss": 1.1422514915466309,
"step": 358
},
{
"epoch": 0.7659574468085106,
"grad_norm": 0.53515625,
"learning_rate": 5.4726280288293156e-05,
"loss": 1.20034658908844,
"step": 360
},
{
"epoch": 0.7702127659574468,
"grad_norm": 0.453125,
"learning_rate": 5.4655636648981876e-05,
"loss": 1.224547266960144,
"step": 362
},
{
"epoch": 0.774468085106383,
"grad_norm": 0.53515625,
"learning_rate": 5.458458223102342e-05,
"loss": 1.2625551223754883,
"step": 364
},
{
"epoch": 0.7787234042553192,
"grad_norm": 0.60546875,
"learning_rate": 5.45131185989657e-05,
"loss": 1.170459508895874,
"step": 366
},
{
"epoch": 0.7829787234042553,
"grad_norm": 0.439453125,
"learning_rate": 5.444124732636714e-05,
"loss": 1.1499803066253662,
"step": 368
},
{
"epoch": 0.7872340425531915,
"grad_norm": 0.6640625,
"learning_rate": 5.4368969995762e-05,
"loss": 1.2377893924713135,
"step": 370
},
{
"epoch": 0.7914893617021277,
"grad_norm": 0.83203125,
"learning_rate": 5.429628819862553e-05,
"loss": 1.1148861646652222,
"step": 372
},
{
"epoch": 0.7957446808510639,
"grad_norm": 0.546875,
"learning_rate": 5.4223203535338945e-05,
"loss": 1.0856741666793823,
"step": 374
},
{
"epoch": 0.8,
"grad_norm": 1.0390625,
"learning_rate": 5.414971761515417e-05,
"loss": 1.150919795036316,
"step": 376
},
{
"epoch": 0.8042553191489362,
"grad_norm": 0.60546875,
"learning_rate": 5.4075832056158395e-05,
"loss": 1.21478271484375,
"step": 378
},
{
"epoch": 0.8085106382978723,
"grad_norm": 0.66796875,
"learning_rate": 5.400154848523847e-05,
"loss": 1.2340768575668335,
"step": 380
},
{
"epoch": 0.8127659574468085,
"grad_norm": 0.58984375,
"learning_rate": 5.392686853804508e-05,
"loss": 1.1131625175476074,
"step": 382
},
{
"epoch": 0.8170212765957446,
"grad_norm": 1.7578125,
"learning_rate": 5.385179385895671e-05,
"loss": 1.203751564025879,
"step": 384
},
{
"epoch": 0.8212765957446808,
"grad_norm": 1.5625,
"learning_rate": 5.377632610104347e-05,
"loss": 1.090701937675476,
"step": 386
},
{
"epoch": 0.825531914893617,
"grad_norm": 0.61328125,
"learning_rate": 5.370046692603067e-05,
"loss": 1.1695449352264404,
"step": 388
},
{
"epoch": 0.8297872340425532,
"grad_norm": 0.416015625,
"learning_rate": 5.362421800426221e-05,
"loss": 1.2987892627716064,
"step": 390
},
{
"epoch": 0.8340425531914893,
"grad_norm": 0.5703125,
"learning_rate": 5.354758101466387e-05,
"loss": 1.145818829536438,
"step": 392
},
{
"epoch": 0.8382978723404255,
"grad_norm": 0.76171875,
"learning_rate": 5.347055764470628e-05,
"loss": 1.0881308317184448,
"step": 394
},
{
"epoch": 0.8425531914893617,
"grad_norm": 0.45703125,
"learning_rate": 5.3393149590367764e-05,
"loss": 1.312870740890503,
"step": 396
},
{
"epoch": 0.8468085106382979,
"grad_norm": 1.3203125,
"learning_rate": 5.331535855609704e-05,
"loss": 1.239678144454956,
"step": 398
},
{
"epoch": 0.851063829787234,
"grad_norm": 0.42578125,
"learning_rate": 5.323718625477566e-05,
"loss": 1.1566566228866577,
"step": 400
},
{
"epoch": 0.8553191489361702,
"grad_norm": 3.984375,
"learning_rate": 5.315863440768031e-05,
"loss": 1.2142958641052246,
"step": 402
},
{
"epoch": 0.8595744680851064,
"grad_norm": 0.7265625,
"learning_rate": 5.307970474444487e-05,
"loss": 1.1810637712478638,
"step": 404
},
{
"epoch": 0.8638297872340426,
"grad_norm": 0.484375,
"learning_rate": 5.300039900302237e-05,
"loss": 1.216611623764038,
"step": 406
},
{
"epoch": 0.8680851063829788,
"grad_norm": 0.43359375,
"learning_rate": 5.292071892964672e-05,
"loss": 1.2652095556259155,
"step": 408
},
{
"epoch": 0.8723404255319149,
"grad_norm": 0.9921875,
"learning_rate": 5.2840666278794244e-05,
"loss": 1.1579747200012207,
"step": 410
},
{
"epoch": 0.8765957446808511,
"grad_norm": 0.62109375,
"learning_rate": 5.276024281314504e-05,
"loss": 1.2436968088150024,
"step": 412
},
{
"epoch": 0.8808510638297873,
"grad_norm": 1.03125,
"learning_rate": 5.267945030354419e-05,
"loss": 1.1159509420394897,
"step": 414
},
{
"epoch": 0.8851063829787233,
"grad_norm": 0.421875,
"learning_rate": 5.259829052896277e-05,
"loss": 1.1389038562774658,
"step": 416
},
{
"epoch": 0.8893617021276595,
"grad_norm": 0.59765625,
"learning_rate": 5.251676527645864e-05,
"loss": 1.179782509803772,
"step": 418
},
{
"epoch": 0.8936170212765957,
"grad_norm": 0.83203125,
"learning_rate": 5.243487634113716e-05,
"loss": 1.2923263311386108,
"step": 420
},
{
"epoch": 0.8978723404255319,
"grad_norm": 2.09375,
"learning_rate": 5.235262552611159e-05,
"loss": 1.0829358100891113,
"step": 422
},
{
"epoch": 0.902127659574468,
"grad_norm": 0.859375,
"learning_rate": 5.227001464246343e-05,
"loss": 1.155044674873352,
"step": 424
},
{
"epoch": 0.9063829787234042,
"grad_norm": 0.5234375,
"learning_rate": 5.218704550920257e-05,
"loss": 1.3217086791992188,
"step": 426
},
{
"epoch": 0.9106382978723404,
"grad_norm": 0.63671875,
"learning_rate": 5.210371995322713e-05,
"loss": 1.2187201976776123,
"step": 428
},
{
"epoch": 0.9148936170212766,
"grad_norm": 0.55078125,
"learning_rate": 5.202003980928338e-05,
"loss": 1.1969107389450073,
"step": 430
},
{
"epoch": 0.9191489361702128,
"grad_norm": 0.76953125,
"learning_rate": 5.1936006919925216e-05,
"loss": 1.2421386241912842,
"step": 432
},
{
"epoch": 0.9234042553191489,
"grad_norm": 0.609375,
"learning_rate": 5.185162313547368e-05,
"loss": 1.5777398347854614,
"step": 434
},
{
"epoch": 0.9276595744680851,
"grad_norm": 0.59765625,
"learning_rate": 5.176689031397612e-05,
"loss": 1.2189395427703857,
"step": 436
},
{
"epoch": 0.9319148936170213,
"grad_norm": 0.546875,
"learning_rate": 5.16818103211654e-05,
"loss": 1.191472053527832,
"step": 438
},
{
"epoch": 0.9361702127659575,
"grad_norm": 0.515625,
"learning_rate": 5.1596385030418686e-05,
"loss": 1.0312530994415283,
"step": 440
},
{
"epoch": 0.9404255319148936,
"grad_norm": 0.72265625,
"learning_rate": 5.151061632271633e-05,
"loss": 1.2013260126113892,
"step": 442
},
{
"epoch": 0.9446808510638298,
"grad_norm": 0.6796875,
"learning_rate": 5.142450608660035e-05,
"loss": 1.1814100742340088,
"step": 444
},
{
"epoch": 0.948936170212766,
"grad_norm": 0.796875,
"learning_rate": 5.133805621813285e-05,
"loss": 1.3198750019073486,
"step": 446
},
{
"epoch": 0.9531914893617022,
"grad_norm": 0.421875,
"learning_rate": 5.1251268620854375e-05,
"loss": 1.1754573583602905,
"step": 448
},
{
"epoch": 0.9574468085106383,
"grad_norm": 0.396484375,
"learning_rate": 5.116414520574186e-05,
"loss": 1.1766588687896729,
"step": 450
},
{
"epoch": 0.9617021276595744,
"grad_norm": 0.41796875,
"learning_rate": 5.107668789116664e-05,
"loss": 1.1669294834136963,
"step": 452
},
{
"epoch": 0.9659574468085106,
"grad_norm": 0.6171875,
"learning_rate": 5.0988898602852204e-05,
"loss": 1.0226119756698608,
"step": 454
},
{
"epoch": 0.9702127659574468,
"grad_norm": 0.392578125,
"learning_rate": 5.090077927383173e-05,
"loss": 1.184398889541626,
"step": 456
},
{
"epoch": 0.9744680851063829,
"grad_norm": 0.71875,
"learning_rate": 5.081233184440562e-05,
"loss": 1.2115238904953003,
"step": 458
},
{
"epoch": 0.9787234042553191,
"grad_norm": 0.412109375,
"learning_rate": 5.072355826209869e-05,
"loss": 1.169049620628357,
"step": 460
},
{
"epoch": 0.9829787234042553,
"grad_norm": 0.5703125,
"learning_rate": 5.063446048161731e-05,
"loss": 1.2979567050933838,
"step": 462
},
{
"epoch": 0.9872340425531915,
"grad_norm": 0.71484375,
"learning_rate": 5.054504046480638e-05,
"loss": 1.046877145767212,
"step": 464
},
{
"epoch": 0.9914893617021276,
"grad_norm": 2.078125,
"learning_rate": 5.0455300180606165e-05,
"loss": 1.2601969242095947,
"step": 466
},
{
"epoch": 0.9957446808510638,
"grad_norm": 0.59375,
"learning_rate": 5.036524160500883e-05,
"loss": 1.1848105192184448,
"step": 468
},
{
"epoch": 1.0,
"grad_norm": 0.486328125,
"learning_rate": 5.027486672101506e-05,
"loss": 1.2240030765533447,
"step": 470
},
{
"epoch": 1.004255319148936,
"grad_norm": 0.482421875,
"learning_rate": 5.0184177518590294e-05,
"loss": 0.9678438901901245,
"step": 472
},
{
"epoch": 1.0085106382978724,
"grad_norm": 1.21875,
"learning_rate": 5.0093175994621006e-05,
"loss": 0.8098158836364746,
"step": 474
},
{
"epoch": 1.0127659574468084,
"grad_norm": 0.5390625,
"learning_rate": 5.000186415287064e-05,
"loss": 1.0377544164657593,
"step": 476
},
{
"epoch": 1.0170212765957447,
"grad_norm": 0.578125,
"learning_rate": 4.9910244003935514e-05,
"loss": 0.9136829376220703,
"step": 478
},
{
"epoch": 1.0212765957446808,
"grad_norm": 0.96875,
"learning_rate": 4.981831756520061e-05,
"loss": 0.8520618081092834,
"step": 480
},
{
"epoch": 1.025531914893617,
"grad_norm": 0.875,
"learning_rate": 4.972608686079509e-05,
"loss": 0.7956000566482544,
"step": 482
},
{
"epoch": 1.0297872340425531,
"grad_norm": 0.3984375,
"learning_rate": 4.963355392154774e-05,
"loss": 0.9468518495559692,
"step": 484
},
{
"epoch": 1.0340425531914894,
"grad_norm": 0.671875,
"learning_rate": 4.954072078494226e-05,
"loss": 1.131507158279419,
"step": 486
},
{
"epoch": 1.0382978723404255,
"grad_norm": 0.37890625,
"learning_rate": 4.944758949507241e-05,
"loss": 1.0291956663131714,
"step": 488
},
{
"epoch": 1.0425531914893618,
"grad_norm": 0.65625,
"learning_rate": 4.9354162102596946e-05,
"loss": 1.1023054122924805,
"step": 490
},
{
"epoch": 1.0468085106382978,
"grad_norm": 0.51171875,
"learning_rate": 4.926044066469459e-05,
"loss": 0.9480533599853516,
"step": 492
},
{
"epoch": 1.0510638297872341,
"grad_norm": 0.58984375,
"learning_rate": 4.916642724501856e-05,
"loss": 0.898105263710022,
"step": 494
},
{
"epoch": 1.0553191489361702,
"grad_norm": 0.38671875,
"learning_rate": 4.9072123913651306e-05,
"loss": 0.9935942888259888,
"step": 496
},
{
"epoch": 1.0595744680851065,
"grad_norm": 0.703125,
"learning_rate": 4.89775327470588e-05,
"loss": 0.7786449193954468,
"step": 498
},
{
"epoch": 1.0638297872340425,
"grad_norm": 0.498046875,
"learning_rate": 4.888265582804487e-05,
"loss": 0.8598610758781433,
"step": 500
},
{
"epoch": 1.0680851063829788,
"grad_norm": 0.42578125,
"learning_rate": 4.878749524570533e-05,
"loss": 0.9269306063652039,
"step": 502
},
{
"epoch": 1.0723404255319149,
"grad_norm": 0.5390625,
"learning_rate": 4.869205309538197e-05,
"loss": 0.6500522494316101,
"step": 504
},
{
"epoch": 1.076595744680851,
"grad_norm": 0.765625,
"learning_rate": 4.8596331478616454e-05,
"loss": 0.9429140090942383,
"step": 506
},
{
"epoch": 1.0808510638297872,
"grad_norm": 0.48828125,
"learning_rate": 4.8500332503103996e-05,
"loss": 0.9708827137947083,
"step": 508
},
{
"epoch": 1.0851063829787233,
"grad_norm": 0.91796875,
"learning_rate": 4.8404058282646985e-05,
"loss": 0.9373714327812195,
"step": 510
},
{
"epoch": 1.0893617021276596,
"grad_norm": 1.671875,
"learning_rate": 4.830751093710844e-05,
"loss": 1.1464526653289795,
"step": 512
},
{
"epoch": 1.0936170212765957,
"grad_norm": 1.1875,
"learning_rate": 4.8210692592365296e-05,
"loss": 0.8451336026191711,
"step": 514
},
{
"epoch": 1.097872340425532,
"grad_norm": 0.8671875,
"learning_rate": 4.811360538026165e-05,
"loss": 1.058915376663208,
"step": 516
},
{
"epoch": 1.102127659574468,
"grad_norm": 1.1796875,
"learning_rate": 4.801625143856179e-05,
"loss": 1.225139856338501,
"step": 518
},
{
"epoch": 1.1063829787234043,
"grad_norm": 0.4609375,
"learning_rate": 4.79186329109031e-05,
"loss": 0.6482216119766235,
"step": 520
},
{
"epoch": 1.1106382978723404,
"grad_norm": 0.515625,
"learning_rate": 4.782075194674892e-05,
"loss": 0.802341103553772,
"step": 522
},
{
"epoch": 1.1148936170212767,
"grad_norm": 0.48828125,
"learning_rate": 4.772261070134113e-05,
"loss": 0.8717508316040039,
"step": 524
},
{
"epoch": 1.1191489361702127,
"grad_norm": 0.455078125,
"learning_rate": 4.762421133565284e-05,
"loss": 0.8982354402542114,
"step": 526
},
{
"epoch": 1.123404255319149,
"grad_norm": 0.83203125,
"learning_rate": 4.752555601634059e-05,
"loss": 1.0001122951507568,
"step": 528
},
{
"epoch": 1.127659574468085,
"grad_norm": 0.451171875,
"learning_rate": 4.742664691569685e-05,
"loss": 0.8902131915092468,
"step": 530
},
{
"epoch": 1.1319148936170214,
"grad_norm": 0.79296875,
"learning_rate": 4.732748621160212e-05,
"loss": 1.2693915367126465,
"step": 532
},
{
"epoch": 1.1361702127659574,
"grad_norm": 0.79296875,
"learning_rate": 4.7228076087476864e-05,
"loss": 0.8774423599243164,
"step": 534
},
{
"epoch": 1.1404255319148937,
"grad_norm": 0.671875,
"learning_rate": 4.71284187322336e-05,
"loss": 0.9756125211715698,
"step": 536
},
{
"epoch": 1.1446808510638298,
"grad_norm": 0.5546875,
"learning_rate": 4.702851634022864e-05,
"loss": 0.8841724991798401,
"step": 538
},
{
"epoch": 1.148936170212766,
"grad_norm": 0.275390625,
"learning_rate": 4.692837111121371e-05,
"loss": 0.794608473777771,
"step": 540
},
{
"epoch": 1.1531914893617021,
"grad_norm": 0.55859375,
"learning_rate": 4.6827985250287616e-05,
"loss": 0.9677450060844421,
"step": 542
},
{
"epoch": 1.1574468085106382,
"grad_norm": 0.51953125,
"learning_rate": 4.672736096784759e-05,
"loss": 0.9043182134628296,
"step": 544
},
{
"epoch": 1.1617021276595745,
"grad_norm": 0.34765625,
"learning_rate": 4.662650047954073e-05,
"loss": 0.9864247441291809,
"step": 546
},
{
"epoch": 1.1659574468085105,
"grad_norm": 0.58984375,
"learning_rate": 4.652540600621512e-05,
"loss": 0.7724499106407166,
"step": 548
},
{
"epoch": 1.1702127659574468,
"grad_norm": 0.357421875,
"learning_rate": 4.642407977387093e-05,
"loss": 0.8427982330322266,
"step": 550
},
{
"epoch": 1.174468085106383,
"grad_norm": 0.71484375,
"learning_rate": 4.632252401361149e-05,
"loss": 0.9444398283958435,
"step": 552
},
{
"epoch": 1.1787234042553192,
"grad_norm": 0.439453125,
"learning_rate": 4.622074096159409e-05,
"loss": 0.9146491289138794,
"step": 554
},
{
"epoch": 1.1829787234042553,
"grad_norm": 0.5390625,
"learning_rate": 4.6118732858980764e-05,
"loss": 0.9469373226165771,
"step": 556
},
{
"epoch": 1.1872340425531915,
"grad_norm": 0.3046875,
"learning_rate": 4.6016501951888916e-05,
"loss": 0.7791444659233093,
"step": 558
},
{
"epoch": 1.1914893617021276,
"grad_norm": 0.59765625,
"learning_rate": 4.591405049134189e-05,
"loss": 0.8431161046028137,
"step": 560
},
{
"epoch": 1.195744680851064,
"grad_norm": 0.609375,
"learning_rate": 4.5811380733219405e-05,
"loss": 0.9778469204902649,
"step": 562
},
{
"epoch": 1.2,
"grad_norm": 1.34375,
"learning_rate": 4.570849493820789e-05,
"loss": 1.0207704305648804,
"step": 564
},
{
"epoch": 1.2042553191489362,
"grad_norm": 0.4140625,
"learning_rate": 4.560539537175068e-05,
"loss": 0.9259340167045593,
"step": 566
},
{
"epoch": 1.2085106382978723,
"grad_norm": 0.71875,
"learning_rate": 4.550208430399813e-05,
"loss": 0.908126711845398,
"step": 568
},
{
"epoch": 1.2127659574468086,
"grad_norm": 0.4375,
"learning_rate": 4.539856400975767e-05,
"loss": 0.9130632877349854,
"step": 570
},
{
"epoch": 1.2170212765957447,
"grad_norm": 0.609375,
"learning_rate": 4.529483676844366e-05,
"loss": 0.9781907796859741,
"step": 572
},
{
"epoch": 1.2212765957446807,
"grad_norm": 0.51953125,
"learning_rate": 4.519090486402727e-05,
"loss": 1.0180959701538086,
"step": 574
},
{
"epoch": 1.225531914893617,
"grad_norm": 0.6328125,
"learning_rate": 4.5086770584986135e-05,
"loss": 0.8333351016044617,
"step": 576
},
{
"epoch": 1.2297872340425533,
"grad_norm": 0.396484375,
"learning_rate": 4.498243622425395e-05,
"loss": 0.7459168434143066,
"step": 578
},
{
"epoch": 1.2340425531914894,
"grad_norm": 0.39453125,
"learning_rate": 4.4877904079170046e-05,
"loss": 0.7476868033409119,
"step": 580
},
{
"epoch": 1.2382978723404254,
"grad_norm": 0.640625,
"learning_rate": 4.477317645142874e-05,
"loss": 0.7878999710083008,
"step": 582
},
{
"epoch": 1.2425531914893617,
"grad_norm": 0.7265625,
"learning_rate": 4.4668255647028706e-05,
"loss": 0.8205560445785522,
"step": 584
},
{
"epoch": 1.2468085106382978,
"grad_norm": 0.71875,
"learning_rate": 4.456314397622217e-05,
"loss": 0.9549564719200134,
"step": 586
},
{
"epoch": 1.251063829787234,
"grad_norm": 0.44921875,
"learning_rate": 4.445784375346406e-05,
"loss": 0.7480695247650146,
"step": 588
},
{
"epoch": 1.2553191489361701,
"grad_norm": 0.4765625,
"learning_rate": 4.435235729736101e-05,
"loss": 1.0009480714797974,
"step": 590
},
{
"epoch": 1.2595744680851064,
"grad_norm": 0.75390625,
"learning_rate": 4.4246686930620326e-05,
"loss": 0.7730486989021301,
"step": 592
},
{
"epoch": 1.2638297872340425,
"grad_norm": 0.62890625,
"learning_rate": 4.4140834979998853e-05,
"loss": 1.0018788576126099,
"step": 594
},
{
"epoch": 1.2680851063829788,
"grad_norm": 0.55078125,
"learning_rate": 4.403480377625176e-05,
"loss": 0.9256978631019592,
"step": 596
},
{
"epoch": 1.2723404255319148,
"grad_norm": 0.376953125,
"learning_rate": 4.392859565408114e-05,
"loss": 0.9084795713424683,
"step": 598
},
{
"epoch": 1.2765957446808511,
"grad_norm": 0.74609375,
"learning_rate": 4.38222129520847e-05,
"loss": 1.0896780490875244,
"step": 600
},
{
"epoch": 1.2808510638297872,
"grad_norm": 0.7890625,
"learning_rate": 4.3715658012704184e-05,
"loss": 0.992649495601654,
"step": 602
},
{
"epoch": 1.2851063829787235,
"grad_norm": 0.76171875,
"learning_rate": 4.360893318217386e-05,
"loss": 0.8589400053024292,
"step": 604
},
{
"epoch": 1.2893617021276595,
"grad_norm": 0.609375,
"learning_rate": 4.350204081046885e-05,
"loss": 1.02162766456604,
"step": 606
},
{
"epoch": 1.2936170212765958,
"grad_norm": 1.0234375,
"learning_rate": 4.3394983251253303e-05,
"loss": 0.83441162109375,
"step": 608
},
{
"epoch": 1.297872340425532,
"grad_norm": 0.69921875,
"learning_rate": 4.328776286182869e-05,
"loss": 1.0881268978118896,
"step": 610
},
{
"epoch": 1.302127659574468,
"grad_norm": 0.375,
"learning_rate": 4.3180382003081806e-05,
"loss": 0.9433637857437134,
"step": 612
},
{
"epoch": 1.3063829787234043,
"grad_norm": 0.55859375,
"learning_rate": 4.307284303943286e-05,
"loss": 1.0009987354278564,
"step": 614
},
{
"epoch": 1.3106382978723405,
"grad_norm": 0.70703125,
"learning_rate": 4.296514833878333e-05,
"loss": 0.9870180487632751,
"step": 616
},
{
"epoch": 1.3148936170212766,
"grad_norm": 1.0390625,
"learning_rate": 4.2857300272463896e-05,
"loss": 0.7985630631446838,
"step": 618
},
{
"epoch": 1.3191489361702127,
"grad_norm": 0.52734375,
"learning_rate": 4.274930121518221e-05,
"loss": 0.9055668115615845,
"step": 620
},
{
"epoch": 1.323404255319149,
"grad_norm": 0.49609375,
"learning_rate": 4.264115354497057e-05,
"loss": 0.9184504747390747,
"step": 622
},
{
"epoch": 1.327659574468085,
"grad_norm": 0.88671875,
"learning_rate": 4.253285964313358e-05,
"loss": 0.8388556241989136,
"step": 624
},
{
"epoch": 1.3319148936170213,
"grad_norm": 0.328125,
"learning_rate": 4.2424421894195746e-05,
"loss": 0.9362179040908813,
"step": 626
},
{
"epoch": 1.3361702127659574,
"grad_norm": 0.390625,
"learning_rate": 4.2315842685848914e-05,
"loss": 0.9783342480659485,
"step": 628
},
{
"epoch": 1.3404255319148937,
"grad_norm": 0.46875,
"learning_rate": 4.220712440889975e-05,
"loss": 1.012178659439087,
"step": 630
},
{
"epoch": 1.3446808510638297,
"grad_norm": 0.47265625,
"learning_rate": 4.2098269457217074e-05,
"loss": 0.9540913701057434,
"step": 632
},
{
"epoch": 1.348936170212766,
"grad_norm": 0.39453125,
"learning_rate": 4.1989280227679136e-05,
"loss": 0.9662917852401733,
"step": 634
},
{
"epoch": 1.353191489361702,
"grad_norm": 0.38671875,
"learning_rate": 4.188015912012085e-05,
"loss": 0.9110101461410522,
"step": 636
},
{
"epoch": 1.3574468085106384,
"grad_norm": 0.451171875,
"learning_rate": 4.177090853728096e-05,
"loss": 0.9384497404098511,
"step": 638
},
{
"epoch": 1.3617021276595744,
"grad_norm": 0.5234375,
"learning_rate": 4.1661530884749125e-05,
"loss": 0.7670997977256775,
"step": 640
},
{
"epoch": 1.3659574468085105,
"grad_norm": 0.50390625,
"learning_rate": 4.155202857091296e-05,
"loss": 0.8632846474647522,
"step": 642
},
{
"epoch": 1.3702127659574468,
"grad_norm": 0.474609375,
"learning_rate": 4.144240400690499e-05,
"loss": 0.6109669804573059,
"step": 644
},
{
"epoch": 1.374468085106383,
"grad_norm": 0.37890625,
"learning_rate": 4.133265960654956e-05,
"loss": 0.9433888792991638,
"step": 646
},
{
"epoch": 1.3787234042553191,
"grad_norm": 0.390625,
"learning_rate": 4.122279778630972e-05,
"loss": 0.9552747011184692,
"step": 648
},
{
"epoch": 1.3829787234042552,
"grad_norm": 0.546875,
"learning_rate": 4.1112820965233954e-05,
"loss": 1.0235886573791504,
"step": 650
},
{
"epoch": 1.3872340425531915,
"grad_norm": 0.54296875,
"learning_rate": 4.100273156490298e-05,
"loss": 0.9218517541885376,
"step": 652
},
{
"epoch": 1.3914893617021278,
"grad_norm": 0.4453125,
"learning_rate": 4.089253200937639e-05,
"loss": 0.9519026875495911,
"step": 654
},
{
"epoch": 1.3957446808510638,
"grad_norm": 0.5625,
"learning_rate": 4.078222472513928e-05,
"loss": 1.0031726360321045,
"step": 656
},
{
"epoch": 1.4,
"grad_norm": 0.3515625,
"learning_rate": 4.067181214104883e-05,
"loss": 0.9014978408813477,
"step": 658
},
{
"epoch": 1.4042553191489362,
"grad_norm": 0.78125,
"learning_rate": 4.05612966882808e-05,
"loss": 0.8434354066848755,
"step": 660
},
{
"epoch": 1.4085106382978723,
"grad_norm": 0.44140625,
"learning_rate": 4.0450680800276065e-05,
"loss": 0.9269939064979553,
"step": 662
},
{
"epoch": 1.4127659574468086,
"grad_norm": 0.62109375,
"learning_rate": 4.033996691268693e-05,
"loss": 1.1910121440887451,
"step": 664
},
{
"epoch": 1.4170212765957446,
"grad_norm": 0.326171875,
"learning_rate": 4.022915746332358e-05,
"loss": 0.9383742213249207,
"step": 666
},
{
"epoch": 1.421276595744681,
"grad_norm": 0.333984375,
"learning_rate": 4.011825489210038e-05,
"loss": 0.9260106086730957,
"step": 668
},
{
"epoch": 1.425531914893617,
"grad_norm": 0.5,
"learning_rate": 4.000726164098213e-05,
"loss": 0.8756862282752991,
"step": 670
},
{
"epoch": 1.4297872340425533,
"grad_norm": 0.39453125,
"learning_rate": 3.9896180153930326e-05,
"loss": 0.9409113526344299,
"step": 672
},
{
"epoch": 1.4340425531914893,
"grad_norm": 0.376953125,
"learning_rate": 3.978501287684933e-05,
"loss": 0.9107382893562317,
"step": 674
},
{
"epoch": 1.4382978723404256,
"grad_norm": 0.48046875,
"learning_rate": 3.9673762257532496e-05,
"loss": 0.9422937035560608,
"step": 676
},
{
"epoch": 1.4425531914893617,
"grad_norm": 0.400390625,
"learning_rate": 3.9562430745608315e-05,
"loss": 0.9622249007225037,
"step": 678
},
{
"epoch": 1.4468085106382977,
"grad_norm": 0.416015625,
"learning_rate": 3.945102079248645e-05,
"loss": 0.9257479906082153,
"step": 680
},
{
"epoch": 1.451063829787234,
"grad_norm": 0.361328125,
"learning_rate": 3.9339534851303746e-05,
"loss": 0.8421862125396729,
"step": 682
},
{
"epoch": 1.4553191489361703,
"grad_norm": 0.53125,
"learning_rate": 3.9227975376870264e-05,
"loss": 0.8206483125686646,
"step": 684
},
{
"epoch": 1.4595744680851064,
"grad_norm": 0.96484375,
"learning_rate": 3.911634482561514e-05,
"loss": 0.8526564836502075,
"step": 686
},
{
"epoch": 1.4638297872340424,
"grad_norm": 0.5625,
"learning_rate": 3.900464565553259e-05,
"loss": 1.036257266998291,
"step": 688
},
{
"epoch": 1.4680851063829787,
"grad_norm": 0.60546875,
"learning_rate": 3.889288032612775e-05,
"loss": 0.6551811695098877,
"step": 690
},
{
"epoch": 1.472340425531915,
"grad_norm": 0.376953125,
"learning_rate": 3.878105129836252e-05,
"loss": 0.9458101391792297,
"step": 692
},
{
"epoch": 1.476595744680851,
"grad_norm": 0.345703125,
"learning_rate": 3.8669161034601336e-05,
"loss": 0.8864946961402893,
"step": 694
},
{
"epoch": 1.4808510638297872,
"grad_norm": 0.435546875,
"learning_rate": 3.8557211998557025e-05,
"loss": 0.9007453322410583,
"step": 696
},
{
"epoch": 1.4851063829787234,
"grad_norm": 0.373046875,
"learning_rate": 3.844520665523654e-05,
"loss": 0.7618290185928345,
"step": 698
},
{
"epoch": 1.4893617021276595,
"grad_norm": 0.625,
"learning_rate": 3.8333147470886604e-05,
"loss": 0.7166705131530762,
"step": 700
},
{
"epoch": 1.4936170212765958,
"grad_norm": 0.63671875,
"learning_rate": 3.822103691293953e-05,
"loss": 0.9165350794792175,
"step": 702
},
{
"epoch": 1.4978723404255319,
"grad_norm": 0.3515625,
"learning_rate": 3.810887744995878e-05,
"loss": 0.6734737157821655,
"step": 704
},
{
"epoch": 1.5021276595744681,
"grad_norm": 0.76171875,
"learning_rate": 3.7996671551584686e-05,
"loss": 1.038259506225586,
"step": 706
},
{
"epoch": 1.5063829787234042,
"grad_norm": 0.58203125,
"learning_rate": 3.788442168848002e-05,
"loss": 1.081926941871643,
"step": 708
},
{
"epoch": 1.5106382978723403,
"grad_norm": 0.375,
"learning_rate": 3.777213033227562e-05,
"loss": 0.9264968633651733,
"step": 710
},
{
"epoch": 1.5148936170212766,
"grad_norm": 0.48828125,
"learning_rate": 3.765979995551599e-05,
"loss": 0.9294235110282898,
"step": 712
},
{
"epoch": 1.5191489361702128,
"grad_norm": 0.490234375,
"learning_rate": 3.7547433031604774e-05,
"loss": 0.9344536662101746,
"step": 714
},
{
"epoch": 1.523404255319149,
"grad_norm": 1.0390625,
"learning_rate": 3.7435032034750385e-05,
"loss": 0.9146227240562439,
"step": 716
},
{
"epoch": 1.527659574468085,
"grad_norm": 0.412109375,
"learning_rate": 3.73225994399115e-05,
"loss": 0.9178793430328369,
"step": 718
},
{
"epoch": 1.5319148936170213,
"grad_norm": 0.337890625,
"learning_rate": 3.721013772274251e-05,
"loss": 0.8365082144737244,
"step": 720
},
{
"epoch": 1.5361702127659576,
"grad_norm": 0.66796875,
"learning_rate": 3.7097649359539075e-05,
"loss": 0.8762301802635193,
"step": 722
},
{
"epoch": 1.5404255319148936,
"grad_norm": 0.37890625,
"learning_rate": 3.6985136827183575e-05,
"loss": 0.8902574181556702,
"step": 724
},
{
"epoch": 1.5446808510638297,
"grad_norm": 0.5,
"learning_rate": 3.6872602603090564e-05,
"loss": 0.9932896494865417,
"step": 726
},
{
"epoch": 1.548936170212766,
"grad_norm": 0.443359375,
"learning_rate": 3.6760049165152256e-05,
"loss": 0.896535336971283,
"step": 728
},
{
"epoch": 1.5531914893617023,
"grad_norm": 0.388671875,
"learning_rate": 3.6647478991683885e-05,
"loss": 1.0637593269348145,
"step": 730
},
{
"epoch": 1.5574468085106383,
"grad_norm": 0.361328125,
"learning_rate": 3.6534894561369214e-05,
"loss": 0.7936927676200867,
"step": 732
},
{
"epoch": 1.5617021276595744,
"grad_norm": 0.578125,
"learning_rate": 3.642229835320593e-05,
"loss": 0.8339595794677734,
"step": 734
},
{
"epoch": 1.5659574468085107,
"grad_norm": 0.38671875,
"learning_rate": 3.630969284645103e-05,
"loss": 0.9330006837844849,
"step": 736
},
{
"epoch": 1.570212765957447,
"grad_norm": 0.42578125,
"learning_rate": 3.6197080520566315e-05,
"loss": 0.8958064913749695,
"step": 738
},
{
"epoch": 1.574468085106383,
"grad_norm": 0.333984375,
"learning_rate": 3.6084463855163666e-05,
"loss": 0.8008121848106384,
"step": 740
},
{
"epoch": 1.578723404255319,
"grad_norm": 0.373046875,
"learning_rate": 3.597184532995055e-05,
"loss": 0.950303316116333,
"step": 742
},
{
"epoch": 1.5829787234042554,
"grad_norm": 0.52734375,
"learning_rate": 3.58592274246754e-05,
"loss": 0.9128738641738892,
"step": 744
},
{
"epoch": 1.5872340425531914,
"grad_norm": 0.40234375,
"learning_rate": 3.574661261907297e-05,
"loss": 0.8636385798454285,
"step": 746
},
{
"epoch": 1.5914893617021275,
"grad_norm": 0.5625,
"learning_rate": 3.563400339280979e-05,
"loss": 0.8232119083404541,
"step": 748
},
{
"epoch": 1.5957446808510638,
"grad_norm": 0.3984375,
"learning_rate": 3.5521402225429485e-05,
"loss": 0.8971787095069885,
"step": 750
},
{
"epoch": 1.6,
"grad_norm": 1.28125,
"learning_rate": 3.540881159629831e-05,
"loss": 0.9066179394721985,
"step": 752
},
{
"epoch": 1.6042553191489362,
"grad_norm": 0.32421875,
"learning_rate": 3.529623398455042e-05,
"loss": 0.8429763913154602,
"step": 754
},
{
"epoch": 1.6085106382978722,
"grad_norm": 1.25,
"learning_rate": 3.5183671869033355e-05,
"loss": 0.8769048452377319,
"step": 756
},
{
"epoch": 1.6127659574468085,
"grad_norm": 0.400390625,
"learning_rate": 3.507112772825345e-05,
"loss": 1.064902424812317,
"step": 758
},
{
"epoch": 1.6170212765957448,
"grad_norm": 0.345703125,
"learning_rate": 3.4958604040321254e-05,
"loss": 1.0025050640106201,
"step": 760
},
{
"epoch": 1.6212765957446809,
"grad_norm": 0.3671875,
"learning_rate": 3.4846103282896956e-05,
"loss": 0.9082741737365723,
"step": 762
},
{
"epoch": 1.625531914893617,
"grad_norm": 0.482421875,
"learning_rate": 3.473362793313583e-05,
"loss": 1.00468909740448,
"step": 764
},
{
"epoch": 1.6297872340425532,
"grad_norm": 0.62890625,
"learning_rate": 3.4621180467633736e-05,
"loss": 1.1201874017715454,
"step": 766
},
{
"epoch": 1.6340425531914895,
"grad_norm": 0.73828125,
"learning_rate": 3.4508763362372495e-05,
"loss": 0.8420827388763428,
"step": 768
},
{
"epoch": 1.6382978723404256,
"grad_norm": 0.4453125,
"learning_rate": 3.4396379092665476e-05,
"loss": 0.9371001720428467,
"step": 770
},
{
"epoch": 1.6425531914893616,
"grad_norm": 0.375,
"learning_rate": 3.428403013310303e-05,
"loss": 1.0956753492355347,
"step": 772
},
{
"epoch": 1.646808510638298,
"grad_norm": 0.94921875,
"learning_rate": 3.4171718957497976e-05,
"loss": 1.0000417232513428,
"step": 774
},
{
"epoch": 1.6510638297872342,
"grad_norm": 0.51953125,
"learning_rate": 3.405944803883121e-05,
"loss": 0.7758415937423706,
"step": 776
},
{
"epoch": 1.65531914893617,
"grad_norm": 0.353515625,
"learning_rate": 3.394721984919721e-05,
"loss": 0.9292746186256409,
"step": 778
},
{
"epoch": 1.6595744680851063,
"grad_norm": 0.42578125,
"learning_rate": 3.383503685974956e-05,
"loss": 0.8925681114196777,
"step": 780
},
{
"epoch": 1.6638297872340426,
"grad_norm": 0.3203125,
"learning_rate": 3.3722901540646634e-05,
"loss": 0.9213247299194336,
"step": 782
},
{
"epoch": 1.6680851063829787,
"grad_norm": 0.57421875,
"learning_rate": 3.361081636099712e-05,
"loss": 1.0045268535614014,
"step": 784
},
{
"epoch": 1.6723404255319148,
"grad_norm": 1.4453125,
"learning_rate": 3.34987837888057e-05,
"loss": 1.0464414358139038,
"step": 786
},
{
"epoch": 1.676595744680851,
"grad_norm": 0.9375,
"learning_rate": 3.338680629091867e-05,
"loss": 0.7823745608329773,
"step": 788
},
{
"epoch": 1.6808510638297873,
"grad_norm": 0.5703125,
"learning_rate": 3.327488633296967e-05,
"loss": 1.0422827005386353,
"step": 790
},
{
"epoch": 1.6851063829787234,
"grad_norm": 0.62890625,
"learning_rate": 3.316302637932537e-05,
"loss": 0.8115458488464355,
"step": 792
},
{
"epoch": 1.6893617021276595,
"grad_norm": 0.38671875,
"learning_rate": 3.305122889303116e-05,
"loss": 1.0222609043121338,
"step": 794
},
{
"epoch": 1.6936170212765957,
"grad_norm": 0.54296875,
"learning_rate": 3.2939496335757004e-05,
"loss": 0.9007856249809265,
"step": 796
},
{
"epoch": 1.697872340425532,
"grad_norm": 0.498046875,
"learning_rate": 3.28278311677432e-05,
"loss": 0.9559044241905212,
"step": 798
},
{
"epoch": 1.702127659574468,
"grad_norm": 0.41015625,
"learning_rate": 3.271623584774616e-05,
"loss": 0.9334125518798828,
"step": 800
},
{
"epoch": 1.7063829787234042,
"grad_norm": 0.49609375,
"learning_rate": 3.260471283298434e-05,
"loss": 0.9849218130111694,
"step": 802
},
{
"epoch": 1.7106382978723405,
"grad_norm": 1.0078125,
"learning_rate": 3.249326457908411e-05,
"loss": 0.8784253001213074,
"step": 804
},
{
"epoch": 1.7148936170212767,
"grad_norm": 0.482421875,
"learning_rate": 3.2381893540025676e-05,
"loss": 0.9033365845680237,
"step": 806
},
{
"epoch": 1.7191489361702128,
"grad_norm": 1.0234375,
"learning_rate": 3.227060216808902e-05,
"loss": 1.0640370845794678,
"step": 808
},
{
"epoch": 1.7234042553191489,
"grad_norm": 1.2734375,
"learning_rate": 3.2159392913799974e-05,
"loss": 0.8058955669403076,
"step": 810
},
{
"epoch": 1.7276595744680852,
"grad_norm": 0.46875,
"learning_rate": 3.2048268225876204e-05,
"loss": 0.9892784357070923,
"step": 812
},
{
"epoch": 1.7319148936170212,
"grad_norm": 0.63671875,
"learning_rate": 3.1937230551173295e-05,
"loss": 0.7961447834968567,
"step": 814
},
{
"epoch": 1.7361702127659573,
"grad_norm": 0.318359375,
"learning_rate": 3.182628233463091e-05,
"loss": 0.9170913100242615,
"step": 816
},
{
"epoch": 1.7404255319148936,
"grad_norm": 0.419921875,
"learning_rate": 3.1715426019218936e-05,
"loss": 0.9619283080101013,
"step": 818
},
{
"epoch": 1.7446808510638299,
"grad_norm": 0.390625,
"learning_rate": 3.1604664045883653e-05,
"loss": 0.9417982697486877,
"step": 820
},
{
"epoch": 1.748936170212766,
"grad_norm": 0.453125,
"learning_rate": 3.1493998853494055e-05,
"loss": 0.8605011105537415,
"step": 822
},
{
"epoch": 1.753191489361702,
"grad_norm": 0.41015625,
"learning_rate": 3.1383432878788086e-05,
"loss": 0.9855290651321411,
"step": 824
},
{
"epoch": 1.7574468085106383,
"grad_norm": 0.482421875,
"learning_rate": 3.127296855631906e-05,
"loss": 1.0079026222229004,
"step": 826
},
{
"epoch": 1.7617021276595746,
"grad_norm": 0.49609375,
"learning_rate": 3.116260831840196e-05,
"loss": 0.9938555955886841,
"step": 828
},
{
"epoch": 1.7659574468085106,
"grad_norm": 3.65625,
"learning_rate": 3.105235459505996e-05,
"loss": 0.930496871471405,
"step": 830
},
{
"epoch": 1.7702127659574467,
"grad_norm": 0.3203125,
"learning_rate": 3.0942209813970894e-05,
"loss": 0.9760335683822632,
"step": 832
},
{
"epoch": 1.774468085106383,
"grad_norm": 0.431640625,
"learning_rate": 3.0832176400413745e-05,
"loss": 0.9947340488433838,
"step": 834
},
{
"epoch": 1.7787234042553193,
"grad_norm": 0.84765625,
"learning_rate": 3.072225677721537e-05,
"loss": 0.994687020778656,
"step": 836
},
{
"epoch": 1.7829787234042553,
"grad_norm": 0.33984375,
"learning_rate": 3.0612453364697025e-05,
"loss": 0.6977970004081726,
"step": 838
},
{
"epoch": 1.7872340425531914,
"grad_norm": 0.94140625,
"learning_rate": 3.050276858062112e-05,
"loss": 0.9086024761199951,
"step": 840
},
{
"epoch": 1.7914893617021277,
"grad_norm": 0.390625,
"learning_rate": 3.039320484013799e-05,
"loss": 0.7375553250312805,
"step": 842
},
{
"epoch": 1.795744680851064,
"grad_norm": 1.234375,
"learning_rate": 3.028376455573274e-05,
"loss": 0.9830509424209595,
"step": 844
},
{
"epoch": 1.8,
"grad_norm": 0.365234375,
"learning_rate": 3.0174450137172063e-05,
"loss": 0.8254852890968323,
"step": 846
},
{
"epoch": 1.804255319148936,
"grad_norm": 0.55859375,
"learning_rate": 3.0065263991451205e-05,
"loss": 0.865581750869751,
"step": 848
},
{
"epoch": 1.8085106382978724,
"grad_norm": 0.56640625,
"learning_rate": 2.9956208522740998e-05,
"loss": 0.9538697600364685,
"step": 850
},
{
"epoch": 1.8127659574468085,
"grad_norm": 0.73828125,
"learning_rate": 2.9847286132334884e-05,
"loss": 0.9136937856674194,
"step": 852
},
{
"epoch": 1.8170212765957445,
"grad_norm": 0.59765625,
"learning_rate": 2.9738499218596033e-05,
"loss": 0.9238496422767639,
"step": 854
},
{
"epoch": 1.8212765957446808,
"grad_norm": 0.7890625,
"learning_rate": 2.962985017690459e-05,
"loss": 0.870966374874115,
"step": 856
},
{
"epoch": 1.825531914893617,
"grad_norm": 0.328125,
"learning_rate": 2.9521341399604866e-05,
"loss": 0.9596646428108215,
"step": 858
},
{
"epoch": 1.8297872340425532,
"grad_norm": 0.80859375,
"learning_rate": 2.9412975275952698e-05,
"loss": 0.8268457651138306,
"step": 860
},
{
"epoch": 1.8340425531914892,
"grad_norm": 0.58984375,
"learning_rate": 2.9304754192062825e-05,
"loss": 0.9566388130187988,
"step": 862
},
{
"epoch": 1.8382978723404255,
"grad_norm": 0.3515625,
"learning_rate": 2.919668053085637e-05,
"loss": 0.9026366472244263,
"step": 864
},
{
"epoch": 1.8425531914893618,
"grad_norm": 1.5859375,
"learning_rate": 2.908875667200835e-05,
"loss": 1.1470530033111572,
"step": 866
},
{
"epoch": 1.8468085106382979,
"grad_norm": 0.53515625,
"learning_rate": 2.898098499189525e-05,
"loss": 0.9321303367614746,
"step": 868
},
{
"epoch": 1.851063829787234,
"grad_norm": 0.462890625,
"learning_rate": 2.887336786354277e-05,
"loss": 0.8843462467193604,
"step": 870
},
{
"epoch": 1.8553191489361702,
"grad_norm": 0.412109375,
"learning_rate": 2.8765907656573538e-05,
"loss": 0.9543187022209167,
"step": 872
},
{
"epoch": 1.8595744680851065,
"grad_norm": 0.40234375,
"learning_rate": 2.8658606737154885e-05,
"loss": 0.9508803486824036,
"step": 874
},
{
"epoch": 1.8638297872340426,
"grad_norm": 0.953125,
"learning_rate": 2.8551467467946817e-05,
"loss": 0.7834142446517944,
"step": 876
},
{
"epoch": 1.8680851063829786,
"grad_norm": 0.5703125,
"learning_rate": 2.844449220804997e-05,
"loss": 0.9817957282066345,
"step": 878
},
{
"epoch": 1.872340425531915,
"grad_norm": 0.5859375,
"learning_rate": 2.8337683312953634e-05,
"loss": 0.9959681630134583,
"step": 880
},
{
"epoch": 1.8765957446808512,
"grad_norm": 0.40234375,
"learning_rate": 2.823104313448392e-05,
"loss": 0.8245176672935486,
"step": 882
},
{
"epoch": 1.8808510638297873,
"grad_norm": 0.3203125,
"learning_rate": 2.8124574020751983e-05,
"loss": 0.9558523297309875,
"step": 884
},
{
"epoch": 1.8851063829787233,
"grad_norm": 0.37890625,
"learning_rate": 2.8018278316102283e-05,
"loss": 0.8767422437667847,
"step": 886
},
{
"epoch": 1.8893617021276596,
"grad_norm": 0.40234375,
"learning_rate": 2.7912158361060976e-05,
"loss": 0.966520369052887,
"step": 888
},
{
"epoch": 1.8936170212765957,
"grad_norm": 0.455078125,
"learning_rate": 2.7806216492284407e-05,
"loss": 0.6318687200546265,
"step": 890
},
{
"epoch": 1.8978723404255318,
"grad_norm": 0.78125,
"learning_rate": 2.7700455042507637e-05,
"loss": 0.9377596378326416,
"step": 892
},
{
"epoch": 1.902127659574468,
"grad_norm": 0.60546875,
"learning_rate": 2.759487634049306e-05,
"loss": 1.0316444635391235,
"step": 894
},
{
"epoch": 1.9063829787234043,
"grad_norm": 0.455078125,
"learning_rate": 2.7489482710979147e-05,
"loss": 0.7911099195480347,
"step": 896
},
{
"epoch": 1.9106382978723404,
"grad_norm": 2.09375,
"learning_rate": 2.7384276474629283e-05,
"loss": 0.8916456699371338,
"step": 898
},
{
"epoch": 1.9148936170212765,
"grad_norm": 0.53125,
"learning_rate": 2.7279259947980615e-05,
"loss": 1.0238900184631348,
"step": 900
},
{
"epoch": 1.9191489361702128,
"grad_norm": 0.373046875,
"learning_rate": 2.717443544339307e-05,
"loss": 0.9721688628196716,
"step": 902
},
{
"epoch": 1.923404255319149,
"grad_norm": 0.423828125,
"learning_rate": 2.7069805268998467e-05,
"loss": 0.8731375336647034,
"step": 904
},
{
"epoch": 1.9276595744680851,
"grad_norm": 0.482421875,
"learning_rate": 2.6965371728649632e-05,
"loss": 0.9867290258407593,
"step": 906
},
{
"epoch": 1.9319148936170212,
"grad_norm": 0.484375,
"learning_rate": 2.686113712186971e-05,
"loss": 0.8383113741874695,
"step": 908
},
{
"epoch": 1.9361702127659575,
"grad_norm": 0.79296875,
"learning_rate": 2.6757103743801555e-05,
"loss": 0.9159626960754395,
"step": 910
},
{
"epoch": 1.9404255319148938,
"grad_norm": 1.4765625,
"learning_rate": 2.665327388515714e-05,
"loss": 0.9763681292533875,
"step": 912
},
{
"epoch": 1.9446808510638298,
"grad_norm": 1.0546875,
"learning_rate": 2.6549649832167138e-05,
"loss": 1.0418541431427002,
"step": 914
},
{
"epoch": 1.9489361702127659,
"grad_norm": 0.73828125,
"learning_rate": 2.644623386653059e-05,
"loss": 1.002961277961731,
"step": 916
},
{
"epoch": 1.9531914893617022,
"grad_norm": 0.5078125,
"learning_rate": 2.6343028265364677e-05,
"loss": 0.8668839335441589,
"step": 918
},
{
"epoch": 1.9574468085106385,
"grad_norm": 0.470703125,
"learning_rate": 2.6240035301154533e-05,
"loss": 1.0359348058700562,
"step": 920
},
{
"epoch": 1.9617021276595743,
"grad_norm": 0.44140625,
"learning_rate": 2.6137257241703254e-05,
"loss": 0.9270017147064209,
"step": 922
},
{
"epoch": 1.9659574468085106,
"grad_norm": 0.29296875,
"learning_rate": 2.6034696350081965e-05,
"loss": 0.9452486634254456,
"step": 924
},
{
"epoch": 1.9702127659574469,
"grad_norm": 0.53125,
"learning_rate": 2.5932354884579916e-05,
"loss": 0.9649438261985779,
"step": 926
},
{
"epoch": 1.974468085106383,
"grad_norm": 0.6171875,
"learning_rate": 2.5830235098654857e-05,
"loss": 0.7984356880187988,
"step": 928
},
{
"epoch": 1.978723404255319,
"grad_norm": 0.380859375,
"learning_rate": 2.5728339240883376e-05,
"loss": 1.008255958557129,
"step": 930
},
{
"epoch": 1.9829787234042553,
"grad_norm": 0.359375,
"learning_rate": 2.5626669554911353e-05,
"loss": 1.021424651145935,
"step": 932
},
{
"epoch": 1.9872340425531916,
"grad_norm": 0.53125,
"learning_rate": 2.55252282794046e-05,
"loss": 0.8926823735237122,
"step": 934
},
{
"epoch": 1.9914893617021276,
"grad_norm": 0.4140625,
"learning_rate": 2.5424017647999574e-05,
"loss": 0.7046434283256531,
"step": 936
},
{
"epoch": 1.9957446808510637,
"grad_norm": 0.5078125,
"learning_rate": 2.532303988925417e-05,
"loss": 1.097957730293274,
"step": 938
},
{
"epoch": 2.0,
"grad_norm": 0.90234375,
"learning_rate": 2.5222297226598633e-05,
"loss": 0.8573867082595825,
"step": 940
},
{
"epoch": 2.0042553191489363,
"grad_norm": 0.455078125,
"learning_rate": 2.512179187828667e-05,
"loss": 0.6483351588249207,
"step": 942
},
{
"epoch": 2.008510638297872,
"grad_norm": 0.380859375,
"learning_rate": 2.5021526057346518e-05,
"loss": 0.6100251078605652,
"step": 944
},
{
"epoch": 2.0127659574468084,
"grad_norm": 0.365234375,
"learning_rate": 2.4921501971532267e-05,
"loss": 0.6785602569580078,
"step": 946
},
{
"epoch": 2.0170212765957447,
"grad_norm": 0.515625,
"learning_rate": 2.4821721823275266e-05,
"loss": 0.6127380728721619,
"step": 948
},
{
"epoch": 2.021276595744681,
"grad_norm": 2.140625,
"learning_rate": 2.472218780963559e-05,
"loss": 0.845111072063446,
"step": 950
},
{
"epoch": 2.025531914893617,
"grad_norm": 0.375,
"learning_rate": 2.462290212225365e-05,
"loss": 0.6631003022193909,
"step": 952
},
{
"epoch": 2.029787234042553,
"grad_norm": 0.435546875,
"learning_rate": 2.452386694730197e-05,
"loss": 0.5201372504234314,
"step": 954
},
{
"epoch": 2.0340425531914894,
"grad_norm": 0.48046875,
"learning_rate": 2.4425084465437053e-05,
"loss": 0.7710279226303101,
"step": 956
},
{
"epoch": 2.0382978723404257,
"grad_norm": 0.64453125,
"learning_rate": 2.4326556851751346e-05,
"loss": 0.6157738566398621,
"step": 958
},
{
"epoch": 2.0425531914893615,
"grad_norm": 0.49609375,
"learning_rate": 2.422828627572534e-05,
"loss": 0.7093011736869812,
"step": 960
},
{
"epoch": 2.046808510638298,
"grad_norm": 0.79296875,
"learning_rate": 2.4130274901179803e-05,
"loss": 0.6860568523406982,
"step": 962
},
{
"epoch": 2.051063829787234,
"grad_norm": 1.359375,
"learning_rate": 2.4032524886228184e-05,
"loss": 0.719150960445404,
"step": 964
},
{
"epoch": 2.0553191489361704,
"grad_norm": 0.6171875,
"learning_rate": 2.3935038383229e-05,
"loss": 0.6379442811012268,
"step": 966
},
{
"epoch": 2.0595744680851062,
"grad_norm": 0.369140625,
"learning_rate": 2.3837817538738525e-05,
"loss": 0.7576701641082764,
"step": 968
},
{
"epoch": 2.0638297872340425,
"grad_norm": 0.3671875,
"learning_rate": 2.374086449346352e-05,
"loss": 0.4662551283836365,
"step": 970
},
{
"epoch": 2.068085106382979,
"grad_norm": 0.57421875,
"learning_rate": 2.3644181382214013e-05,
"loss": 0.5384807586669922,
"step": 972
},
{
"epoch": 2.072340425531915,
"grad_norm": 0.515625,
"learning_rate": 2.3547770333856386e-05,
"loss": 0.6011054515838623,
"step": 974
},
{
"epoch": 2.076595744680851,
"grad_norm": 0.83203125,
"learning_rate": 2.345163347126647e-05,
"loss": 0.5131646990776062,
"step": 976
},
{
"epoch": 2.0808510638297872,
"grad_norm": 0.52734375,
"learning_rate": 2.3355772911282796e-05,
"loss": 0.8288999199867249,
"step": 978
},
{
"epoch": 2.0851063829787235,
"grad_norm": 0.2080078125,
"learning_rate": 2.3260190764659976e-05,
"loss": 0.47640979290008545,
"step": 980
},
{
"epoch": 2.0893617021276594,
"grad_norm": 0.55859375,
"learning_rate": 2.316488913602222e-05,
"loss": 0.5666584372520447,
"step": 982
},
{
"epoch": 2.0936170212765957,
"grad_norm": 1.0390625,
"learning_rate": 2.3069870123817056e-05,
"loss": 0.5758649110794067,
"step": 984
},
{
"epoch": 2.097872340425532,
"grad_norm": 0.314453125,
"learning_rate": 2.2975135820269026e-05,
"loss": 0.5656971335411072,
"step": 986
},
{
"epoch": 2.1021276595744682,
"grad_norm": 0.6171875,
"learning_rate": 2.2880688311333702e-05,
"loss": 0.3263399302959442,
"step": 988
},
{
"epoch": 2.106382978723404,
"grad_norm": 0.85546875,
"learning_rate": 2.278652967665173e-05,
"loss": 0.690028965473175,
"step": 990
},
{
"epoch": 2.1106382978723404,
"grad_norm": 0.7890625,
"learning_rate": 2.2692661989502995e-05,
"loss": 0.6520633101463318,
"step": 992
},
{
"epoch": 2.1148936170212767,
"grad_norm": 0.36328125,
"learning_rate": 2.2599087316761018e-05,
"loss": 0.7235679626464844,
"step": 994
},
{
"epoch": 2.119148936170213,
"grad_norm": 0.84375,
"learning_rate": 2.2505807718847456e-05,
"loss": 0.6303759217262268,
"step": 996
},
{
"epoch": 2.123404255319149,
"grad_norm": 0.65625,
"learning_rate": 2.24128252496867e-05,
"loss": 0.6719092130661011,
"step": 998
},
{
"epoch": 2.127659574468085,
"grad_norm": 0.5,
"learning_rate": 2.2320141956660646e-05,
"loss": 0.28417664766311646,
"step": 1000
},
{
"epoch": 2.1319148936170214,
"grad_norm": 0.6171875,
"learning_rate": 2.2227759880563626e-05,
"loss": 0.7599141597747803,
"step": 1002
},
{
"epoch": 2.1361702127659576,
"grad_norm": 0.58984375,
"learning_rate": 2.2135681055557504e-05,
"loss": 0.6200002431869507,
"step": 1004
},
{
"epoch": 2.1404255319148935,
"grad_norm": 0.7578125,
"learning_rate": 2.2043907509126812e-05,
"loss": 0.7208251953125,
"step": 1006
},
{
"epoch": 2.1446808510638298,
"grad_norm": 0.90234375,
"learning_rate": 2.1952441262034188e-05,
"loss": 0.5411021709442139,
"step": 1008
},
{
"epoch": 2.148936170212766,
"grad_norm": 0.41796875,
"learning_rate": 2.1861284328275845e-05,
"loss": 0.5026164650917053,
"step": 1010
},
{
"epoch": 2.153191489361702,
"grad_norm": 0.44140625,
"learning_rate": 2.1770438715037165e-05,
"loss": 0.6443688273429871,
"step": 1012
},
{
"epoch": 2.157446808510638,
"grad_norm": 0.578125,
"learning_rate": 2.1679906422648626e-05,
"loss": 0.6200368404388428,
"step": 1014
},
{
"epoch": 2.1617021276595745,
"grad_norm": 0.75390625,
"learning_rate": 2.1589689444541662e-05,
"loss": 0.6573978066444397,
"step": 1016
},
{
"epoch": 2.1659574468085108,
"grad_norm": 1.390625,
"learning_rate": 2.1499789767204812e-05,
"loss": 0.6872032880783081,
"step": 1018
},
{
"epoch": 2.1702127659574466,
"grad_norm": 2.890625,
"learning_rate": 2.1410209370139945e-05,
"loss": 0.5005927085876465,
"step": 1020
},
{
"epoch": 2.174468085106383,
"grad_norm": 0.58203125,
"learning_rate": 2.1320950225818707e-05,
"loss": 0.7184480428695679,
"step": 1022
},
{
"epoch": 2.178723404255319,
"grad_norm": 0.54296875,
"learning_rate": 2.1232014299639085e-05,
"loss": 0.6513587832450867,
"step": 1024
},
{
"epoch": 2.1829787234042555,
"grad_norm": 0.76171875,
"learning_rate": 2.11434035498821e-05,
"loss": 0.5126093626022339,
"step": 1026
},
{
"epoch": 2.1872340425531913,
"grad_norm": 0.447265625,
"learning_rate": 2.105511992766874e-05,
"loss": 0.6408154964447021,
"step": 1028
},
{
"epoch": 2.1914893617021276,
"grad_norm": 0.59765625,
"learning_rate": 2.0967165376916945e-05,
"loss": 0.7048395872116089,
"step": 1030
},
{
"epoch": 2.195744680851064,
"grad_norm": 0.84375,
"learning_rate": 2.0879541834298828e-05,
"loss": 0.662340521812439,
"step": 1032
},
{
"epoch": 2.2,
"grad_norm": 0.70703125,
"learning_rate": 2.0792251229198035e-05,
"loss": 0.7315186858177185,
"step": 1034
},
{
"epoch": 2.204255319148936,
"grad_norm": 0.435546875,
"learning_rate": 2.0705295483667274e-05,
"loss": 0.47347530722618103,
"step": 1036
},
{
"epoch": 2.2085106382978723,
"grad_norm": 0.79296875,
"learning_rate": 2.0618676512385965e-05,
"loss": 0.5597242712974548,
"step": 1038
},
{
"epoch": 2.2127659574468086,
"grad_norm": 0.8515625,
"learning_rate": 2.0532396222618083e-05,
"loss": 0.6009190082550049,
"step": 1040
},
{
"epoch": 2.217021276595745,
"grad_norm": 0.4140625,
"learning_rate": 2.0446456514170166e-05,
"loss": 0.7078320384025574,
"step": 1042
},
{
"epoch": 2.2212765957446807,
"grad_norm": 0.6171875,
"learning_rate": 2.0360859279349523e-05,
"loss": 0.6109620928764343,
"step": 1044
},
{
"epoch": 2.225531914893617,
"grad_norm": 1.8984375,
"learning_rate": 2.0275606402922496e-05,
"loss": 0.6881995797157288,
"step": 1046
},
{
"epoch": 2.2297872340425533,
"grad_norm": 0.65625,
"learning_rate": 2.0190699762073015e-05,
"loss": 0.8076979517936707,
"step": 1048
},
{
"epoch": 2.2340425531914896,
"grad_norm": 0.70703125,
"learning_rate": 2.010614122636125e-05,
"loss": 0.5962254405021667,
"step": 1050
},
{
"epoch": 2.2382978723404254,
"grad_norm": 0.494140625,
"learning_rate": 2.002193265768241e-05,
"loss": 0.5154112577438354,
"step": 1052
},
{
"epoch": 2.2425531914893617,
"grad_norm": 0.478515625,
"learning_rate": 1.9938075910225816e-05,
"loss": 0.5857576131820679,
"step": 1054
},
{
"epoch": 2.246808510638298,
"grad_norm": 1.671875,
"learning_rate": 1.985457283043401e-05,
"loss": 0.7445710301399231,
"step": 1056
},
{
"epoch": 2.251063829787234,
"grad_norm": 0.5234375,
"learning_rate": 1.977142525696214e-05,
"loss": 0.4012032449245453,
"step": 1058
},
{
"epoch": 2.25531914893617,
"grad_norm": 0.419921875,
"learning_rate": 1.9688635020637438e-05,
"loss": 0.7297635078430176,
"step": 1060
},
{
"epoch": 2.2595744680851064,
"grad_norm": 0.458984375,
"learning_rate": 1.9606203944418924e-05,
"loss": 0.5968733429908752,
"step": 1062
},
{
"epoch": 2.2638297872340427,
"grad_norm": 0.498046875,
"learning_rate": 1.9524133843357294e-05,
"loss": 0.7591136693954468,
"step": 1064
},
{
"epoch": 2.2680851063829786,
"grad_norm": 0.421875,
"learning_rate": 1.9442426524554893e-05,
"loss": 0.6319488286972046,
"step": 1066
},
{
"epoch": 2.272340425531915,
"grad_norm": 0.58984375,
"learning_rate": 1.9361083787126e-05,
"loss": 0.6187411546707153,
"step": 1068
},
{
"epoch": 2.276595744680851,
"grad_norm": 0.49609375,
"learning_rate": 1.9280107422157143e-05,
"loss": 0.6891050934791565,
"step": 1070
},
{
"epoch": 2.2808510638297874,
"grad_norm": 0.3984375,
"learning_rate": 1.9199499212667688e-05,
"loss": 0.507449209690094,
"step": 1072
},
{
"epoch": 2.2851063829787233,
"grad_norm": 0.515625,
"learning_rate": 1.9119260933570603e-05,
"loss": 0.563790500164032,
"step": 1074
},
{
"epoch": 2.2893617021276595,
"grad_norm": 0.48828125,
"learning_rate": 1.903939435163335e-05,
"loss": 0.5918843150138855,
"step": 1076
},
{
"epoch": 2.293617021276596,
"grad_norm": 0.6640625,
"learning_rate": 1.8959901225439e-05,
"loss": 0.6510270833969116,
"step": 1078
},
{
"epoch": 2.297872340425532,
"grad_norm": 0.6796875,
"learning_rate": 1.888078330534744e-05,
"loss": 0.4389875531196594,
"step": 1080
},
{
"epoch": 2.302127659574468,
"grad_norm": 1.78125,
"learning_rate": 1.880204233345696e-05,
"loss": 0.3430854082107544,
"step": 1082
},
{
"epoch": 2.3063829787234043,
"grad_norm": 0.51953125,
"learning_rate": 1.8723680043565798e-05,
"loss": 0.7696226835250854,
"step": 1084
},
{
"epoch": 2.3106382978723405,
"grad_norm": 0.453125,
"learning_rate": 1.8645698161133972e-05,
"loss": 0.5286341905593872,
"step": 1086
},
{
"epoch": 2.3148936170212764,
"grad_norm": 0.76953125,
"learning_rate": 1.8568098403245336e-05,
"loss": 0.567283570766449,
"step": 1088
},
{
"epoch": 2.3191489361702127,
"grad_norm": 0.439453125,
"learning_rate": 1.8490882478569716e-05,
"loss": 0.5737777352333069,
"step": 1090
},
{
"epoch": 2.323404255319149,
"grad_norm": 0.54296875,
"learning_rate": 1.8414052087325308e-05,
"loss": 0.728583037853241,
"step": 1092
},
{
"epoch": 2.3276595744680852,
"grad_norm": 0.408203125,
"learning_rate": 1.8337608921241267e-05,
"loss": 0.600398600101471,
"step": 1094
},
{
"epoch": 2.331914893617021,
"grad_norm": 0.609375,
"learning_rate": 1.8261554663520416e-05,
"loss": 0.6284059286117554,
"step": 1096
},
{
"epoch": 2.3361702127659574,
"grad_norm": 0.71875,
"learning_rate": 1.8185890988802214e-05,
"loss": 0.6262116432189941,
"step": 1098
},
{
"epoch": 2.3404255319148937,
"grad_norm": 1.0703125,
"learning_rate": 1.8110619563125844e-05,
"loss": 0.5292909741401672,
"step": 1100
},
{
"epoch": 2.34468085106383,
"grad_norm": 0.44140625,
"learning_rate": 1.8035742043893575e-05,
"loss": 0.625741720199585,
"step": 1102
},
{
"epoch": 2.348936170212766,
"grad_norm": 0.416015625,
"learning_rate": 1.796126007983425e-05,
"loss": 0.7988000512123108,
"step": 1104
},
{
"epoch": 2.353191489361702,
"grad_norm": 0.6640625,
"learning_rate": 1.7887175310966956e-05,
"loss": 0.6517726182937622,
"step": 1106
},
{
"epoch": 2.3574468085106384,
"grad_norm": 0.96875,
"learning_rate": 1.7813489368564965e-05,
"loss": 0.6426142454147339,
"step": 1108
},
{
"epoch": 2.3617021276595747,
"grad_norm": 0.474609375,
"learning_rate": 1.7740203875119755e-05,
"loss": 0.5290454626083374,
"step": 1110
},
{
"epoch": 2.3659574468085105,
"grad_norm": 0.578125,
"learning_rate": 1.7667320444305326e-05,
"loss": 0.5435438752174377,
"step": 1112
},
{
"epoch": 2.370212765957447,
"grad_norm": 0.59375,
"learning_rate": 1.7594840680942667e-05,
"loss": 0.4323336184024811,
"step": 1114
},
{
"epoch": 2.374468085106383,
"grad_norm": 0.431640625,
"learning_rate": 1.752276618096441e-05,
"loss": 0.5053521394729614,
"step": 1116
},
{
"epoch": 2.378723404255319,
"grad_norm": 1.1640625,
"learning_rate": 1.7451098531379666e-05,
"loss": 0.63356614112854,
"step": 1118
},
{
"epoch": 2.382978723404255,
"grad_norm": 0.58984375,
"learning_rate": 1.7379839310239118e-05,
"loss": 0.7377380728721619,
"step": 1120
},
{
"epoch": 2.3872340425531915,
"grad_norm": 0.9296875,
"learning_rate": 1.7308990086600258e-05,
"loss": 0.7407448887825012,
"step": 1122
},
{
"epoch": 2.391489361702128,
"grad_norm": 0.396484375,
"learning_rate": 1.7238552420492854e-05,
"loss": 0.5100683569908142,
"step": 1124
},
{
"epoch": 2.395744680851064,
"grad_norm": 0.52734375,
"learning_rate": 1.716852786288455e-05,
"loss": 0.6132201552391052,
"step": 1126
},
{
"epoch": 2.4,
"grad_norm": 0.47265625,
"learning_rate": 1.709891795564679e-05,
"loss": 0.5412701368331909,
"step": 1128
},
{
"epoch": 2.404255319148936,
"grad_norm": 0.453125,
"learning_rate": 1.7029724231520792e-05,
"loss": 0.8436723351478577,
"step": 1130
},
{
"epoch": 2.4085106382978725,
"grad_norm": 0.4375,
"learning_rate": 1.696094821408385e-05,
"loss": 0.6744006872177124,
"step": 1132
},
{
"epoch": 2.4127659574468083,
"grad_norm": 1.1171875,
"learning_rate": 1.6892591417715775e-05,
"loss": 0.7410330176353455,
"step": 1134
},
{
"epoch": 2.4170212765957446,
"grad_norm": 1.3984375,
"learning_rate": 1.682465534756555e-05,
"loss": 0.6016606092453003,
"step": 1136
},
{
"epoch": 2.421276595744681,
"grad_norm": 0.4921875,
"learning_rate": 1.6757141499518153e-05,
"loss": 0.4425470530986786,
"step": 1138
},
{
"epoch": 2.425531914893617,
"grad_norm": 0.546875,
"learning_rate": 1.6690051360161673e-05,
"loss": 0.7017032504081726,
"step": 1140
},
{
"epoch": 2.429787234042553,
"grad_norm": 0.7265625,
"learning_rate": 1.6623386406754555e-05,
"loss": 0.5560771822929382,
"step": 1142
},
{
"epoch": 2.4340425531914893,
"grad_norm": 1.0390625,
"learning_rate": 1.655714810719307e-05,
"loss": 0.6498077511787415,
"step": 1144
},
{
"epoch": 2.4382978723404256,
"grad_norm": 0.466796875,
"learning_rate": 1.6491337919978978e-05,
"loss": 0.6854323148727417,
"step": 1146
},
{
"epoch": 2.4425531914893615,
"grad_norm": 0.86328125,
"learning_rate": 1.642595729418745e-05,
"loss": 0.4551085829734802,
"step": 1148
},
{
"epoch": 2.4468085106382977,
"grad_norm": 0.76171875,
"learning_rate": 1.6361007669435126e-05,
"loss": 0.5184612274169922,
"step": 1150
},
{
"epoch": 2.451063829787234,
"grad_norm": 0.5234375,
"learning_rate": 1.6296490475848424e-05,
"loss": 0.5938658118247986,
"step": 1152
},
{
"epoch": 2.4553191489361703,
"grad_norm": 0.69921875,
"learning_rate": 1.623240713403207e-05,
"loss": 0.5654492974281311,
"step": 1154
},
{
"epoch": 2.4595744680851066,
"grad_norm": 0.47265625,
"learning_rate": 1.6168759055037817e-05,
"loss": 0.6496747732162476,
"step": 1156
},
{
"epoch": 2.4638297872340424,
"grad_norm": 1.921875,
"learning_rate": 1.610554764033332e-05,
"loss": 0.7038140296936035,
"step": 1158
},
{
"epoch": 2.4680851063829787,
"grad_norm": 0.58203125,
"learning_rate": 1.6042774281771345e-05,
"loss": 0.641715407371521,
"step": 1160
},
{
"epoch": 2.472340425531915,
"grad_norm": 0.80859375,
"learning_rate": 1.59804403615591e-05,
"loss": 0.5276774168014526,
"step": 1162
},
{
"epoch": 2.476595744680851,
"grad_norm": 0.45703125,
"learning_rate": 1.5918547252227793e-05,
"loss": 0.6699836850166321,
"step": 1164
},
{
"epoch": 2.480851063829787,
"grad_norm": 0.69140625,
"learning_rate": 1.58570963166024e-05,
"loss": 0.6293801069259644,
"step": 1166
},
{
"epoch": 2.4851063829787234,
"grad_norm": 0.435546875,
"learning_rate": 1.5796088907771674e-05,
"loss": 0.6323214769363403,
"step": 1168
},
{
"epoch": 2.4893617021276597,
"grad_norm": 0.55859375,
"learning_rate": 1.5735526369058364e-05,
"loss": 0.6430104970932007,
"step": 1170
},
{
"epoch": 2.4936170212765956,
"grad_norm": 0.380859375,
"learning_rate": 1.5675410033989592e-05,
"loss": 0.5185415744781494,
"step": 1172
},
{
"epoch": 2.497872340425532,
"grad_norm": 0.78125,
"learning_rate": 1.561574122626754e-05,
"loss": 0.47881028056144714,
"step": 1174
},
{
"epoch": 2.502127659574468,
"grad_norm": 0.4921875,
"learning_rate": 1.555652125974028e-05,
"loss": 0.7108798027038574,
"step": 1176
},
{
"epoch": 2.506382978723404,
"grad_norm": 1.09375,
"learning_rate": 1.5497751438372827e-05,
"loss": 0.5829865336418152,
"step": 1178
},
{
"epoch": 2.5106382978723403,
"grad_norm": 0.486328125,
"learning_rate": 1.543943305621846e-05,
"loss": 0.5929511785507202,
"step": 1180
},
{
"epoch": 2.5148936170212766,
"grad_norm": 0.5703125,
"learning_rate": 1.538156739739021e-05,
"loss": 0.5175199508666992,
"step": 1182
},
{
"epoch": 2.519148936170213,
"grad_norm": 0.298828125,
"learning_rate": 1.5324155736032595e-05,
"loss": 0.7886143326759338,
"step": 1184
},
{
"epoch": 2.523404255319149,
"grad_norm": 0.416015625,
"learning_rate": 1.526719933629355e-05,
"loss": 0.6642839908599854,
"step": 1186
},
{
"epoch": 2.527659574468085,
"grad_norm": 0.703125,
"learning_rate": 1.5210699452296592e-05,
"loss": 0.8588666915893555,
"step": 1188
},
{
"epoch": 2.5319148936170213,
"grad_norm": 0.6171875,
"learning_rate": 1.5154657328113233e-05,
"loss": 0.7579631209373474,
"step": 1190
},
{
"epoch": 2.5361702127659576,
"grad_norm": 0.46484375,
"learning_rate": 1.5099074197735552e-05,
"loss": 0.6104500889778137,
"step": 1192
},
{
"epoch": 2.5404255319148934,
"grad_norm": 0.58203125,
"learning_rate": 1.504395128504905e-05,
"loss": 0.34646666049957275,
"step": 1194
},
{
"epoch": 2.5446808510638297,
"grad_norm": 0.7890625,
"learning_rate": 1.4989289803805685e-05,
"loss": 0.5463817119598389,
"step": 1196
},
{
"epoch": 2.548936170212766,
"grad_norm": 0.83984375,
"learning_rate": 1.4935090957597149e-05,
"loss": 0.61968594789505,
"step": 1198
},
{
"epoch": 2.5531914893617023,
"grad_norm": 0.61328125,
"learning_rate": 1.4881355939828364e-05,
"loss": 0.6242573857307434,
"step": 1200
},
{
"epoch": 2.5574468085106385,
"grad_norm": 0.51953125,
"learning_rate": 1.4828085933691223e-05,
"loss": 0.6220880150794983,
"step": 1202
},
{
"epoch": 2.5617021276595744,
"grad_norm": 0.58984375,
"learning_rate": 1.477528211213852e-05,
"loss": 0.6225618124008179,
"step": 1204
},
{
"epoch": 2.5659574468085107,
"grad_norm": 0.3515625,
"learning_rate": 1.4722945637858116e-05,
"loss": 0.4512316584587097,
"step": 1206
},
{
"epoch": 2.570212765957447,
"grad_norm": 0.9921875,
"learning_rate": 1.4671077663247351e-05,
"loss": 0.5912795662879944,
"step": 1208
},
{
"epoch": 2.574468085106383,
"grad_norm": 0.4921875,
"learning_rate": 1.4619679330387679e-05,
"loss": 0.6551855802536011,
"step": 1210
},
{
"epoch": 2.578723404255319,
"grad_norm": 0.62109375,
"learning_rate": 1.4568751771019482e-05,
"loss": 0.5984119772911072,
"step": 1212
},
{
"epoch": 2.5829787234042554,
"grad_norm": 0.65625,
"learning_rate": 1.4518296106517206e-05,
"loss": 0.6674041152000427,
"step": 1214
},
{
"epoch": 2.5872340425531917,
"grad_norm": 0.58203125,
"learning_rate": 1.4468313447864624e-05,
"loss": 0.48521387577056885,
"step": 1216
},
{
"epoch": 2.5914893617021275,
"grad_norm": 3.734375,
"learning_rate": 1.441880489563038e-05,
"loss": 0.6522884964942932,
"step": 1218
},
{
"epoch": 2.595744680851064,
"grad_norm": 1.109375,
"learning_rate": 1.4369771539943776e-05,
"loss": 0.7783772945404053,
"step": 1220
},
{
"epoch": 2.6,
"grad_norm": 0.6640625,
"learning_rate": 1.4321214460470755e-05,
"loss": 0.5732651948928833,
"step": 1222
},
{
"epoch": 2.604255319148936,
"grad_norm": 0.53125,
"learning_rate": 1.4273134726390138e-05,
"loss": 0.6476449966430664,
"step": 1224
},
{
"epoch": 2.608510638297872,
"grad_norm": 1.3515625,
"learning_rate": 1.4225533396370053e-05,
"loss": 0.7929537296295166,
"step": 1226
},
{
"epoch": 2.6127659574468085,
"grad_norm": 0.43359375,
"learning_rate": 1.4178411518544654e-05,
"loss": 0.3998229205608368,
"step": 1228
},
{
"epoch": 2.617021276595745,
"grad_norm": 1.6640625,
"learning_rate": 1.413177013049104e-05,
"loss": 0.4514097273349762,
"step": 1230
},
{
"epoch": 2.621276595744681,
"grad_norm": 0.9765625,
"learning_rate": 1.4085610259206387e-05,
"loss": 0.6864388585090637,
"step": 1232
},
{
"epoch": 2.625531914893617,
"grad_norm": 0.421875,
"learning_rate": 1.4039932921085362e-05,
"loss": 0.5575997233390808,
"step": 1234
},
{
"epoch": 2.629787234042553,
"grad_norm": 0.49609375,
"learning_rate": 1.3994739121897718e-05,
"loss": 0.653709888458252,
"step": 1236
},
{
"epoch": 2.6340425531914895,
"grad_norm": 0.53125,
"learning_rate": 1.3950029856766165e-05,
"loss": 0.6980820298194885,
"step": 1238
},
{
"epoch": 2.6382978723404253,
"grad_norm": 0.515625,
"learning_rate": 1.3905806110144452e-05,
"loss": 0.6501242518424988,
"step": 1240
},
{
"epoch": 2.6425531914893616,
"grad_norm": 0.46875,
"learning_rate": 1.3862068855795701e-05,
"loss": 0.7544458508491516,
"step": 1242
},
{
"epoch": 2.646808510638298,
"grad_norm": 1.0,
"learning_rate": 1.3818819056770932e-05,
"loss": 0.5266544222831726,
"step": 1244
},
{
"epoch": 2.651063829787234,
"grad_norm": 0.9140625,
"learning_rate": 1.3776057665387907e-05,
"loss": 0.6442818641662598,
"step": 1246
},
{
"epoch": 2.65531914893617,
"grad_norm": 0.375,
"learning_rate": 1.373378562321012e-05,
"loss": 0.6095808148384094,
"step": 1248
},
{
"epoch": 2.6595744680851063,
"grad_norm": 0.453125,
"learning_rate": 1.3692003861026083e-05,
"loss": 0.7218018770217896,
"step": 1250
},
{
"epoch": 2.6638297872340426,
"grad_norm": 0.52734375,
"learning_rate": 1.365071329882883e-05,
"loss": 0.6180848479270935,
"step": 1252
},
{
"epoch": 2.6680851063829785,
"grad_norm": 0.73828125,
"learning_rate": 1.360991484579566e-05,
"loss": 0.4784125089645386,
"step": 1254
},
{
"epoch": 2.6723404255319148,
"grad_norm": 0.51953125,
"learning_rate": 1.3569609400268112e-05,
"loss": 0.7386208772659302,
"step": 1256
},
{
"epoch": 2.676595744680851,
"grad_norm": 0.33984375,
"learning_rate": 1.3529797849732183e-05,
"loss": 0.5939998626708984,
"step": 1258
},
{
"epoch": 2.6808510638297873,
"grad_norm": 0.494140625,
"learning_rate": 1.3490481070798797e-05,
"loss": 0.6285054683685303,
"step": 1260
},
{
"epoch": 2.6851063829787236,
"grad_norm": 0.6171875,
"learning_rate": 1.34516599291845e-05,
"loss": 0.7830681800842285,
"step": 1262
},
{
"epoch": 2.6893617021276595,
"grad_norm": 2.609375,
"learning_rate": 1.3413335279692392e-05,
"loss": 0.7019430994987488,
"step": 1264
},
{
"epoch": 2.6936170212765957,
"grad_norm": 1.1328125,
"learning_rate": 1.3375507966193309e-05,
"loss": 0.5392411947250366,
"step": 1266
},
{
"epoch": 2.697872340425532,
"grad_norm": 0.4609375,
"learning_rate": 1.3338178821607234e-05,
"loss": 0.8191847205162048,
"step": 1268
},
{
"epoch": 2.702127659574468,
"grad_norm": 0.53515625,
"learning_rate": 1.3301348667884975e-05,
"loss": 0.30988848209381104,
"step": 1270
},
{
"epoch": 2.706382978723404,
"grad_norm": 0.474609375,
"learning_rate": 1.3265018315990046e-05,
"loss": 0.5801700353622437,
"step": 1272
},
{
"epoch": 2.7106382978723405,
"grad_norm": 0.390625,
"learning_rate": 1.3229188565880827e-05,
"loss": 0.6832275390625,
"step": 1274
},
{
"epoch": 2.7148936170212767,
"grad_norm": 0.435546875,
"learning_rate": 1.3193860206492936e-05,
"loss": 0.6654208898544312,
"step": 1276
},
{
"epoch": 2.719148936170213,
"grad_norm": 1.078125,
"learning_rate": 1.3159034015721865e-05,
"loss": 0.7127547860145569,
"step": 1278
},
{
"epoch": 2.723404255319149,
"grad_norm": 0.5859375,
"learning_rate": 1.3124710760405853e-05,
"loss": 0.537979781627655,
"step": 1280
},
{
"epoch": 2.727659574468085,
"grad_norm": 0.546875,
"learning_rate": 1.3090891196309e-05,
"loss": 0.6351765990257263,
"step": 1282
},
{
"epoch": 2.731914893617021,
"grad_norm": 0.390625,
"learning_rate": 1.3057576068104621e-05,
"loss": 0.601453959941864,
"step": 1284
},
{
"epoch": 2.7361702127659573,
"grad_norm": 0.53515625,
"learning_rate": 1.3024766109358845e-05,
"loss": 0.7181084752082825,
"step": 1286
},
{
"epoch": 2.7404255319148936,
"grad_norm": 0.59765625,
"learning_rate": 1.2992462042514483e-05,
"loss": 0.5218726396560669,
"step": 1288
},
{
"epoch": 2.74468085106383,
"grad_norm": 0.51953125,
"learning_rate": 1.2960664578875104e-05,
"loss": 0.7979943156242371,
"step": 1290
},
{
"epoch": 2.748936170212766,
"grad_norm": 0.64453125,
"learning_rate": 1.2929374418589363e-05,
"loss": 0.600561797618866,
"step": 1292
},
{
"epoch": 2.753191489361702,
"grad_norm": 0.578125,
"learning_rate": 1.289859225063562e-05,
"loss": 0.5805540084838867,
"step": 1294
},
{
"epoch": 2.7574468085106383,
"grad_norm": 0.5390625,
"learning_rate": 1.2868318752806726e-05,
"loss": 0.5937775373458862,
"step": 1296
},
{
"epoch": 2.7617021276595746,
"grad_norm": 0.6328125,
"learning_rate": 1.2838554591695126e-05,
"loss": 0.4572460353374481,
"step": 1298
},
{
"epoch": 2.7659574468085104,
"grad_norm": 0.447265625,
"learning_rate": 1.2809300422678187e-05,
"loss": 0.7637752294540405,
"step": 1300
},
{
"epoch": 2.7702127659574467,
"grad_norm": 0.53515625,
"learning_rate": 1.2780556889903737e-05,
"loss": 0.48286503553390503,
"step": 1302
},
{
"epoch": 2.774468085106383,
"grad_norm": 0.828125,
"learning_rate": 1.275232462627591e-05,
"loss": 0.5692922472953796,
"step": 1304
},
{
"epoch": 2.7787234042553193,
"grad_norm": 0.7734375,
"learning_rate": 1.2724604253441195e-05,
"loss": 0.49659186601638794,
"step": 1306
},
{
"epoch": 2.7829787234042556,
"grad_norm": 0.671875,
"learning_rate": 1.2697396381774753e-05,
"loss": 0.7237148880958557,
"step": 1308
},
{
"epoch": 2.7872340425531914,
"grad_norm": 0.44921875,
"learning_rate": 1.2670701610366985e-05,
"loss": 0.5669375658035278,
"step": 1310
},
{
"epoch": 2.7914893617021277,
"grad_norm": 0.38671875,
"learning_rate": 1.2644520527010319e-05,
"loss": 0.49942082166671753,
"step": 1312
},
{
"epoch": 2.795744680851064,
"grad_norm": 0.47265625,
"learning_rate": 1.2618853708186294e-05,
"loss": 0.5854433178901672,
"step": 1314
},
{
"epoch": 2.8,
"grad_norm": 0.890625,
"learning_rate": 1.2593701719052839e-05,
"loss": 0.4207378625869751,
"step": 1316
},
{
"epoch": 2.804255319148936,
"grad_norm": 0.65625,
"learning_rate": 1.2569065113431854e-05,
"loss": 0.6822559237480164,
"step": 1318
},
{
"epoch": 2.8085106382978724,
"grad_norm": 0.52734375,
"learning_rate": 1.2544944433797002e-05,
"loss": 0.5026736855506897,
"step": 1320
},
{
"epoch": 2.8127659574468087,
"grad_norm": 0.7109375,
"learning_rate": 1.252134021126177e-05,
"loss": 0.7221361398696899,
"step": 1322
},
{
"epoch": 2.8170212765957445,
"grad_norm": 0.6171875,
"learning_rate": 1.2498252965567755e-05,
"loss": 0.6420372724533081,
"step": 1324
},
{
"epoch": 2.821276595744681,
"grad_norm": 0.60546875,
"learning_rate": 1.2475683205073255e-05,
"loss": 0.5234490633010864,
"step": 1326
},
{
"epoch": 2.825531914893617,
"grad_norm": 0.482421875,
"learning_rate": 1.2453631426742047e-05,
"loss": 0.51298987865448,
"step": 1328
},
{
"epoch": 2.829787234042553,
"grad_norm": 0.5625,
"learning_rate": 1.2432098116132458e-05,
"loss": 0.5665377378463745,
"step": 1330
},
{
"epoch": 2.8340425531914892,
"grad_norm": 0.5703125,
"learning_rate": 1.2411083747386662e-05,
"loss": 0.7207722067832947,
"step": 1332
},
{
"epoch": 2.8382978723404255,
"grad_norm": 0.50390625,
"learning_rate": 1.2390588783220257e-05,
"loss": 0.595878005027771,
"step": 1334
},
{
"epoch": 2.842553191489362,
"grad_norm": 0.94921875,
"learning_rate": 1.2370613674912055e-05,
"loss": 0.7001873850822449,
"step": 1336
},
{
"epoch": 2.846808510638298,
"grad_norm": 0.46484375,
"learning_rate": 1.2351158862294165e-05,
"loss": 0.5404136776924133,
"step": 1338
},
{
"epoch": 2.851063829787234,
"grad_norm": 0.52734375,
"learning_rate": 1.2332224773742298e-05,
"loss": 0.5846556425094604,
"step": 1340
},
{
"epoch": 2.8553191489361702,
"grad_norm": 0.4609375,
"learning_rate": 1.231381182616634e-05,
"loss": 0.4478878080844879,
"step": 1342
},
{
"epoch": 2.8595744680851065,
"grad_norm": 0.625,
"learning_rate": 1.229592042500116e-05,
"loss": 0.49717509746551514,
"step": 1344
},
{
"epoch": 2.8638297872340424,
"grad_norm": 0.57421875,
"learning_rate": 1.2278550964197693e-05,
"loss": 0.650644063949585,
"step": 1346
},
{
"epoch": 2.8680851063829786,
"grad_norm": 0.375,
"learning_rate": 1.2261703826214275e-05,
"loss": 0.6666759252548218,
"step": 1348
},
{
"epoch": 2.872340425531915,
"grad_norm": 0.5546875,
"learning_rate": 1.2245379382008206e-05,
"loss": 0.7478127479553223,
"step": 1350
},
{
"epoch": 2.876595744680851,
"grad_norm": 0.482421875,
"learning_rate": 1.2229577991027574e-05,
"loss": 0.7170307636260986,
"step": 1352
},
{
"epoch": 2.8808510638297875,
"grad_norm": 0.67578125,
"learning_rate": 1.2214300001203369e-05,
"loss": 0.8161506056785583,
"step": 1354
},
{
"epoch": 2.8851063829787233,
"grad_norm": 0.9375,
"learning_rate": 1.2199545748941797e-05,
"loss": 0.49168941378593445,
"step": 1356
},
{
"epoch": 2.8893617021276596,
"grad_norm": 0.50390625,
"learning_rate": 1.2185315559116883e-05,
"loss": 0.46535876393318176,
"step": 1358
},
{
"epoch": 2.8936170212765955,
"grad_norm": 0.57421875,
"learning_rate": 1.217160974506332e-05,
"loss": 0.770456850528717,
"step": 1360
},
{
"epoch": 2.8978723404255318,
"grad_norm": 0.578125,
"learning_rate": 1.2158428608569563e-05,
"loss": 0.6479332447052002,
"step": 1362
},
{
"epoch": 2.902127659574468,
"grad_norm": 0.71484375,
"learning_rate": 1.2145772439871186e-05,
"loss": 0.5973125100135803,
"step": 1364
},
{
"epoch": 2.9063829787234043,
"grad_norm": 0.6484375,
"learning_rate": 1.2133641517644488e-05,
"loss": 0.5936951637268066,
"step": 1366
},
{
"epoch": 2.9106382978723406,
"grad_norm": 0.64453125,
"learning_rate": 1.2122036109000369e-05,
"loss": 0.5945168733596802,
"step": 1368
},
{
"epoch": 2.9148936170212765,
"grad_norm": 0.42578125,
"learning_rate": 1.2110956469478434e-05,
"loss": 0.799973726272583,
"step": 1370
},
{
"epoch": 2.9191489361702128,
"grad_norm": 0.4453125,
"learning_rate": 1.2100402843041378e-05,
"loss": 0.5132302641868591,
"step": 1372
},
{
"epoch": 2.923404255319149,
"grad_norm": 0.578125,
"learning_rate": 1.2090375462069602e-05,
"loss": 0.6212731599807739,
"step": 1374
},
{
"epoch": 2.927659574468085,
"grad_norm": 0.416015625,
"learning_rate": 1.208087454735611e-05,
"loss": 0.6145610809326172,
"step": 1376
},
{
"epoch": 2.931914893617021,
"grad_norm": 0.65234375,
"learning_rate": 1.2071900308101635e-05,
"loss": 0.7530128359794617,
"step": 1378
},
{
"epoch": 2.9361702127659575,
"grad_norm": 0.54296875,
"learning_rate": 1.2063452941910037e-05,
"loss": 0.5762054324150085,
"step": 1380
},
{
"epoch": 2.9404255319148938,
"grad_norm": 0.490234375,
"learning_rate": 1.205553263478396e-05,
"loss": 0.7723484039306641,
"step": 1382
},
{
"epoch": 2.94468085106383,
"grad_norm": 0.5234375,
"learning_rate": 1.2048139561120719e-05,
"loss": 0.6207926869392395,
"step": 1384
},
{
"epoch": 2.948936170212766,
"grad_norm": 0.58984375,
"learning_rate": 1.2041273883708483e-05,
"loss": 0.6677027940750122,
"step": 1386
},
{
"epoch": 2.953191489361702,
"grad_norm": 0.47265625,
"learning_rate": 1.2034935753722666e-05,
"loss": 0.7704895734786987,
"step": 1388
},
{
"epoch": 2.9574468085106385,
"grad_norm": 0.4453125,
"learning_rate": 1.2029125310722613e-05,
"loss": 0.7197574377059937,
"step": 1390
},
{
"epoch": 2.9617021276595743,
"grad_norm": 0.51953125,
"learning_rate": 1.202384268264853e-05,
"loss": 0.7283585667610168,
"step": 1392
},
{
"epoch": 2.9659574468085106,
"grad_norm": 0.5078125,
"learning_rate": 1.2019087985818655e-05,
"loss": 0.5656808018684387,
"step": 1394
},
{
"epoch": 2.970212765957447,
"grad_norm": 0.87109375,
"learning_rate": 1.2014861324926705e-05,
"loss": 0.49777036905288696,
"step": 1396
},
{
"epoch": 2.974468085106383,
"grad_norm": 0.61328125,
"learning_rate": 1.2011162793039564e-05,
"loss": 0.6693958640098572,
"step": 1398
},
{
"epoch": 2.978723404255319,
"grad_norm": 0.439453125,
"learning_rate": 1.2007992471595247e-05,
"loss": 0.5529562830924988,
"step": 1400
},
{
"epoch": 2.9829787234042553,
"grad_norm": 0.4375,
"learning_rate": 1.2005350430401092e-05,
"loss": 0.5693247318267822,
"step": 1402
},
{
"epoch": 2.9872340425531916,
"grad_norm": 0.3359375,
"learning_rate": 1.2003236727632224e-05,
"loss": 0.5829908847808838,
"step": 1404
},
{
"epoch": 2.9914893617021274,
"grad_norm": 0.484375,
"learning_rate": 1.2001651409830295e-05,
"loss": 0.6101734638214111,
"step": 1406
},
{
"epoch": 2.9957446808510637,
"grad_norm": 0.43359375,
"learning_rate": 1.2000594511902426e-05,
"loss": 0.6352876424789429,
"step": 1408
},
{
"epoch": 3.0,
"grad_norm": 0.83203125,
"learning_rate": 1.2000066057120467e-05,
"loss": 0.3924168646335602,
"step": 1410
},
{
"epoch": 3.0,
"step": 1410,
"total_flos": 4.1743170019314893e+18,
"train_loss": 0.9459603985573383,
"train_runtime": 10552.9203,
"train_samples_per_second": 4.276,
"train_steps_per_second": 0.134
}
],
"logging_steps": 2,
"max_steps": 1410,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 99999,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 4.1743170019314893e+18,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}