diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,6748 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 2.0, + "eval_steps": 500, + "global_step": 958, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0020876826722338203, + "grad_norm": 0.8805100452450126, + "learning_rate": 1.3793103448275863e-05, + "loss": 3.4375, + "step": 1 + }, + { + "epoch": 0.0041753653444676405, + "grad_norm": 0.8962056177448029, + "learning_rate": 2.7586206896551727e-05, + "loss": 3.4531, + "step": 2 + }, + { + "epoch": 0.006263048016701462, + "grad_norm": 0.8975533488683494, + "learning_rate": 4.1379310344827587e-05, + "loss": 3.375, + "step": 3 + }, + { + "epoch": 0.008350730688935281, + "grad_norm": 0.8562154080719621, + "learning_rate": 5.517241379310345e-05, + "loss": 3.0312, + "step": 4 + }, + { + "epoch": 0.010438413361169102, + "grad_norm": 0.7722844092130382, + "learning_rate": 6.896551724137931e-05, + "loss": 2.0938, + "step": 5 + }, + { + "epoch": 0.012526096033402923, + "grad_norm": 0.5469423252625637, + "learning_rate": 8.275862068965517e-05, + "loss": 1.5391, + "step": 6 + }, + { + "epoch": 0.014613778705636743, + "grad_norm": 0.269527877565377, + "learning_rate": 9.655172413793105e-05, + "loss": 1.3281, + "step": 7 + }, + { + "epoch": 0.016701461377870562, + "grad_norm": 0.2129561647095634, + "learning_rate": 0.0001103448275862069, + "loss": 1.0547, + "step": 8 + }, + { + "epoch": 0.018789144050104383, + "grad_norm": 0.209174835903919, + "learning_rate": 0.00012413793103448277, + "loss": 0.8672, + "step": 9 + }, + { + "epoch": 0.020876826722338204, + "grad_norm": 0.16723166688980518, + "learning_rate": 0.00013793103448275863, + "loss": 0.7461, + "step": 10 + }, + { + "epoch": 0.022964509394572025, + "grad_norm": 0.13523071310549678, + "learning_rate": 0.00015172413793103449, + "loss": 0.75, + "step": 11 + }, + { + "epoch": 0.025052192066805846, + "grad_norm": 0.13783786374844356, + "learning_rate": 0.00016551724137931035, + "loss": 0.7578, + "step": 12 + }, + { + "epoch": 0.027139874739039668, + "grad_norm": 0.11159117473743674, + "learning_rate": 0.0001793103448275862, + "loss": 0.6914, + "step": 13 + }, + { + "epoch": 0.029227557411273485, + "grad_norm": 0.10486826556036052, + "learning_rate": 0.0001931034482758621, + "loss": 0.6484, + "step": 14 + }, + { + "epoch": 0.031315240083507306, + "grad_norm": 0.1012538060823961, + "learning_rate": 0.00020689655172413795, + "loss": 0.6562, + "step": 15 + }, + { + "epoch": 0.033402922755741124, + "grad_norm": 0.10310591282634196, + "learning_rate": 0.0002206896551724138, + "loss": 0.6602, + "step": 16 + }, + { + "epoch": 0.03549060542797495, + "grad_norm": 0.07838991965173675, + "learning_rate": 0.00023448275862068965, + "loss": 0.5469, + "step": 17 + }, + { + "epoch": 0.037578288100208766, + "grad_norm": 0.07804122247550843, + "learning_rate": 0.00024827586206896553, + "loss": 0.543, + "step": 18 + }, + { + "epoch": 0.03966597077244259, + "grad_norm": 0.08565464666867292, + "learning_rate": 0.00026206896551724137, + "loss": 0.5625, + "step": 19 + }, + { + "epoch": 0.04175365344467641, + "grad_norm": 0.07879295016594862, + "learning_rate": 0.00027586206896551725, + "loss": 0.4551, + "step": 20 + }, + { + "epoch": 0.04384133611691023, + "grad_norm": 0.07549627256843491, + "learning_rate": 0.00028965517241379314, + "loss": 0.4863, + "step": 21 + }, + { + "epoch": 0.04592901878914405, + "grad_norm": 0.07816061161931821, + "learning_rate": 0.00030344827586206897, + "loss": 0.459, + "step": 22 + }, + { + "epoch": 0.04801670146137787, + "grad_norm": 0.07796400588496231, + "learning_rate": 0.00031724137931034486, + "loss": 0.5156, + "step": 23 + }, + { + "epoch": 0.05010438413361169, + "grad_norm": 0.06699113066481945, + "learning_rate": 0.0003310344827586207, + "loss": 0.5508, + "step": 24 + }, + { + "epoch": 0.05219206680584551, + "grad_norm": 0.07261576114010894, + "learning_rate": 0.0003448275862068965, + "loss": 0.5469, + "step": 25 + }, + { + "epoch": 0.054279749478079335, + "grad_norm": 0.07464032217874282, + "learning_rate": 0.0003586206896551724, + "loss": 0.4805, + "step": 26 + }, + { + "epoch": 0.05636743215031315, + "grad_norm": 0.06913377612077472, + "learning_rate": 0.0003724137931034483, + "loss": 0.5117, + "step": 27 + }, + { + "epoch": 0.05845511482254697, + "grad_norm": 0.061903733751750956, + "learning_rate": 0.0003862068965517242, + "loss": 0.4844, + "step": 28 + }, + { + "epoch": 0.060542797494780795, + "grad_norm": 0.06908123825325603, + "learning_rate": 0.0004, + "loss": 0.4434, + "step": 29 + }, + { + "epoch": 0.06263048016701461, + "grad_norm": 0.05444037852658766, + "learning_rate": 0.00039999885641643965, + "loss": 0.4336, + "step": 30 + }, + { + "epoch": 0.06471816283924843, + "grad_norm": 0.06329218054783699, + "learning_rate": 0.00039999542567883625, + "loss": 0.5078, + "step": 31 + }, + { + "epoch": 0.06680584551148225, + "grad_norm": 0.05617311135086726, + "learning_rate": 0.0003999897078264233, + "loss": 0.3418, + "step": 32 + }, + { + "epoch": 0.06889352818371608, + "grad_norm": 0.0631466065976358, + "learning_rate": 0.00039998170292458913, + "loss": 0.4512, + "step": 33 + }, + { + "epoch": 0.0709812108559499, + "grad_norm": 0.05939776609982831, + "learning_rate": 0.0003999714110648765, + "loss": 0.4727, + "step": 34 + }, + { + "epoch": 0.07306889352818371, + "grad_norm": 0.07015310870913878, + "learning_rate": 0.0003999588323649815, + "loss": 0.4316, + "step": 35 + }, + { + "epoch": 0.07515657620041753, + "grad_norm": 0.05541574151493628, + "learning_rate": 0.00039994396696875197, + "loss": 0.3691, + "step": 36 + }, + { + "epoch": 0.07724425887265135, + "grad_norm": 0.04781737127842468, + "learning_rate": 0.0003999268150461862, + "loss": 0.3711, + "step": 37 + }, + { + "epoch": 0.07933194154488518, + "grad_norm": 0.052081481582784105, + "learning_rate": 0.00039990737679343073, + "loss": 0.3633, + "step": 38 + }, + { + "epoch": 0.081419624217119, + "grad_norm": 0.04824424451620144, + "learning_rate": 0.0003998856524327782, + "loss": 0.3867, + "step": 39 + }, + { + "epoch": 0.08350730688935282, + "grad_norm": 0.048331445388276904, + "learning_rate": 0.0003998616422126649, + "loss": 0.4199, + "step": 40 + }, + { + "epoch": 0.08559498956158663, + "grad_norm": 0.06796823383551855, + "learning_rate": 0.00039983534640766766, + "loss": 0.4629, + "step": 41 + }, + { + "epoch": 0.08768267223382047, + "grad_norm": 0.04720979824092969, + "learning_rate": 0.0003998067653185011, + "loss": 0.3594, + "step": 42 + }, + { + "epoch": 0.08977035490605428, + "grad_norm": 0.05374730136315078, + "learning_rate": 0.00039977589927201373, + "loss": 0.4414, + "step": 43 + }, + { + "epoch": 0.0918580375782881, + "grad_norm": 0.050894908072710146, + "learning_rate": 0.0003997427486211847, + "loss": 0.4043, + "step": 44 + }, + { + "epoch": 0.09394572025052192, + "grad_norm": 0.04383506588833311, + "learning_rate": 0.0003997073137451194, + "loss": 0.3281, + "step": 45 + }, + { + "epoch": 0.09603340292275574, + "grad_norm": 0.055686594769476794, + "learning_rate": 0.00039966959504904515, + "loss": 0.4062, + "step": 46 + }, + { + "epoch": 0.09812108559498957, + "grad_norm": 0.05244693300757331, + "learning_rate": 0.00039962959296430676, + "loss": 0.3652, + "step": 47 + }, + { + "epoch": 0.10020876826722339, + "grad_norm": 0.05926741920735249, + "learning_rate": 0.0003995873079483616, + "loss": 0.4844, + "step": 48 + }, + { + "epoch": 0.1022964509394572, + "grad_norm": 0.04551719284385615, + "learning_rate": 0.00039954274048477405, + "loss": 0.3223, + "step": 49 + }, + { + "epoch": 0.10438413361169102, + "grad_norm": 0.04385484814869444, + "learning_rate": 0.0003994958910832104, + "loss": 0.4004, + "step": 50 + }, + { + "epoch": 0.10647181628392484, + "grad_norm": 0.04780597607586747, + "learning_rate": 0.0003994467602794327, + "loss": 0.3125, + "step": 51 + }, + { + "epoch": 0.10855949895615867, + "grad_norm": 0.08970784709982177, + "learning_rate": 0.00039939534863529265, + "loss": 0.3359, + "step": 52 + }, + { + "epoch": 0.11064718162839249, + "grad_norm": 0.055961833035488864, + "learning_rate": 0.00039934165673872536, + "loss": 0.4023, + "step": 53 + }, + { + "epoch": 0.1127348643006263, + "grad_norm": 0.051294633986921585, + "learning_rate": 0.00039928568520374263, + "loss": 0.3848, + "step": 54 + }, + { + "epoch": 0.11482254697286012, + "grad_norm": 0.05439699876212045, + "learning_rate": 0.0003992274346704257, + "loss": 0.4785, + "step": 55 + }, + { + "epoch": 0.11691022964509394, + "grad_norm": 0.04990096696869215, + "learning_rate": 0.00039916690580491806, + "loss": 0.377, + "step": 56 + }, + { + "epoch": 0.11899791231732777, + "grad_norm": 0.04880913842131259, + "learning_rate": 0.00039910409929941787, + "loss": 0.3809, + "step": 57 + }, + { + "epoch": 0.12108559498956159, + "grad_norm": 0.048110591694635114, + "learning_rate": 0.0003990390158721701, + "loss": 0.375, + "step": 58 + }, + { + "epoch": 0.12317327766179541, + "grad_norm": 0.05425382131888079, + "learning_rate": 0.0003989716562674579, + "loss": 0.4297, + "step": 59 + }, + { + "epoch": 0.12526096033402923, + "grad_norm": 0.04140427495861459, + "learning_rate": 0.0003989020212555948, + "loss": 0.3164, + "step": 60 + }, + { + "epoch": 0.12734864300626306, + "grad_norm": 0.055270560900777276, + "learning_rate": 0.00039883011163291545, + "loss": 0.418, + "step": 61 + }, + { + "epoch": 0.12943632567849686, + "grad_norm": 0.04546442173482473, + "learning_rate": 0.0003987559282217662, + "loss": 0.3125, + "step": 62 + }, + { + "epoch": 0.1315240083507307, + "grad_norm": 0.05835477454974449, + "learning_rate": 0.00039867947187049654, + "loss": 0.4199, + "step": 63 + }, + { + "epoch": 0.1336116910229645, + "grad_norm": 0.06512707863408194, + "learning_rate": 0.00039860074345344875, + "loss": 0.4238, + "step": 64 + }, + { + "epoch": 0.13569937369519833, + "grad_norm": 0.04896390157005381, + "learning_rate": 0.0003985197438709479, + "loss": 0.2598, + "step": 65 + }, + { + "epoch": 0.13778705636743216, + "grad_norm": 0.048613271081049014, + "learning_rate": 0.00039843647404929207, + "loss": 0.3633, + "step": 66 + }, + { + "epoch": 0.13987473903966596, + "grad_norm": 0.050233778334495785, + "learning_rate": 0.0003983509349407412, + "loss": 0.3926, + "step": 67 + }, + { + "epoch": 0.1419624217118998, + "grad_norm": 0.051442044526522436, + "learning_rate": 0.0003982631275235064, + "loss": 0.4531, + "step": 68 + }, + { + "epoch": 0.1440501043841336, + "grad_norm": 0.04624056126157752, + "learning_rate": 0.00039817305280173893, + "loss": 0.373, + "step": 69 + }, + { + "epoch": 0.14613778705636743, + "grad_norm": 0.04350984771291989, + "learning_rate": 0.0003980807118055185, + "loss": 0.377, + "step": 70 + }, + { + "epoch": 0.14822546972860126, + "grad_norm": 0.04395725697208839, + "learning_rate": 0.0003979861055908415, + "loss": 0.3633, + "step": 71 + }, + { + "epoch": 0.15031315240083507, + "grad_norm": 0.04706297130636851, + "learning_rate": 0.0003978892352396091, + "loss": 0.4199, + "step": 72 + }, + { + "epoch": 0.1524008350730689, + "grad_norm": 0.044497773696428565, + "learning_rate": 0.00039779010185961473, + "loss": 0.4707, + "step": 73 + }, + { + "epoch": 0.1544885177453027, + "grad_norm": 0.0411155822363719, + "learning_rate": 0.0003976887065845314, + "loss": 0.3262, + "step": 74 + }, + { + "epoch": 0.15657620041753653, + "grad_norm": 0.043812847498547546, + "learning_rate": 0.0003975850505738988, + "loss": 0.3359, + "step": 75 + }, + { + "epoch": 0.15866388308977036, + "grad_norm": 0.0554940446413714, + "learning_rate": 0.0003974791350131101, + "loss": 0.3867, + "step": 76 + }, + { + "epoch": 0.16075156576200417, + "grad_norm": 0.04898321080003146, + "learning_rate": 0.00039737096111339804, + "loss": 0.4648, + "step": 77 + }, + { + "epoch": 0.162839248434238, + "grad_norm": 0.04703455829390756, + "learning_rate": 0.00039726053011182176, + "loss": 0.3945, + "step": 78 + }, + { + "epoch": 0.1649269311064718, + "grad_norm": 0.048291736323549404, + "learning_rate": 0.00039714784327125196, + "loss": 0.3555, + "step": 79 + }, + { + "epoch": 0.16701461377870563, + "grad_norm": 0.048158996103635526, + "learning_rate": 0.00039703290188035687, + "loss": 0.3848, + "step": 80 + }, + { + "epoch": 0.16910229645093947, + "grad_norm": 0.041070665599436396, + "learning_rate": 0.00039691570725358726, + "loss": 0.3672, + "step": 81 + }, + { + "epoch": 0.17118997912317327, + "grad_norm": 0.04514739709583794, + "learning_rate": 0.0003967962607311617, + "loss": 0.416, + "step": 82 + }, + { + "epoch": 0.1732776617954071, + "grad_norm": 0.05018154319088866, + "learning_rate": 0.0003966745636790509, + "loss": 0.4219, + "step": 83 + }, + { + "epoch": 0.17536534446764093, + "grad_norm": 0.04615841699661321, + "learning_rate": 0.00039655061748896243, + "loss": 0.3594, + "step": 84 + }, + { + "epoch": 0.17745302713987474, + "grad_norm": 0.04770173243618234, + "learning_rate": 0.00039642442357832447, + "loss": 0.3789, + "step": 85 + }, + { + "epoch": 0.17954070981210857, + "grad_norm": 0.0529492248920703, + "learning_rate": 0.00039629598339026984, + "loss": 0.4102, + "step": 86 + }, + { + "epoch": 0.18162839248434237, + "grad_norm": 0.04108643805111346, + "learning_rate": 0.00039616529839361946, + "loss": 0.3691, + "step": 87 + }, + { + "epoch": 0.1837160751565762, + "grad_norm": 0.04230844261748932, + "learning_rate": 0.00039603237008286544, + "loss": 0.3672, + "step": 88 + }, + { + "epoch": 0.18580375782881003, + "grad_norm": 0.057536412046600406, + "learning_rate": 0.00039589719997815413, + "loss": 0.3574, + "step": 89 + }, + { + "epoch": 0.18789144050104384, + "grad_norm": 0.03943187889859641, + "learning_rate": 0.0003957597896252685, + "loss": 0.332, + "step": 90 + }, + { + "epoch": 0.18997912317327767, + "grad_norm": 0.04024451449784526, + "learning_rate": 0.000395620140595611, + "loss": 0.3438, + "step": 91 + }, + { + "epoch": 0.19206680584551147, + "grad_norm": 0.04428176826646225, + "learning_rate": 0.0003954782544861847, + "loss": 0.3574, + "step": 92 + }, + { + "epoch": 0.1941544885177453, + "grad_norm": 0.045692937481720604, + "learning_rate": 0.00039533413291957596, + "loss": 0.3574, + "step": 93 + }, + { + "epoch": 0.19624217118997914, + "grad_norm": 0.04688683747789243, + "learning_rate": 0.00039518777754393535, + "loss": 0.3867, + "step": 94 + }, + { + "epoch": 0.19832985386221294, + "grad_norm": 0.05536707488449957, + "learning_rate": 0.0003950391900329588, + "loss": 0.3223, + "step": 95 + }, + { + "epoch": 0.20041753653444677, + "grad_norm": 0.04606231136571311, + "learning_rate": 0.0003948883720858687, + "loss": 0.3535, + "step": 96 + }, + { + "epoch": 0.20250521920668058, + "grad_norm": 0.04099133317121465, + "learning_rate": 0.00039473532542739435, + "loss": 0.3066, + "step": 97 + }, + { + "epoch": 0.2045929018789144, + "grad_norm": 0.055252229005040635, + "learning_rate": 0.0003945800518077521, + "loss": 0.3809, + "step": 98 + }, + { + "epoch": 0.20668058455114824, + "grad_norm": 0.04490759353014913, + "learning_rate": 0.00039442255300262553, + "loss": 0.373, + "step": 99 + }, + { + "epoch": 0.20876826722338204, + "grad_norm": 0.04889359614788617, + "learning_rate": 0.0003942628308131452, + "loss": 0.377, + "step": 100 + }, + { + "epoch": 0.21085594989561587, + "grad_norm": 0.04331101172153666, + "learning_rate": 0.0003941008870658677, + "loss": 0.3926, + "step": 101 + }, + { + "epoch": 0.21294363256784968, + "grad_norm": 0.039340556954521134, + "learning_rate": 0.00039393672361275506, + "loss": 0.3379, + "step": 102 + }, + { + "epoch": 0.2150313152400835, + "grad_norm": 0.047692728640419506, + "learning_rate": 0.0003937703423311537, + "loss": 0.375, + "step": 103 + }, + { + "epoch": 0.21711899791231734, + "grad_norm": 0.04193899617132502, + "learning_rate": 0.00039360174512377246, + "loss": 0.3203, + "step": 104 + }, + { + "epoch": 0.21920668058455114, + "grad_norm": 0.04450971768791051, + "learning_rate": 0.0003934309339186614, + "loss": 0.3789, + "step": 105 + }, + { + "epoch": 0.22129436325678498, + "grad_norm": 0.04610401641341007, + "learning_rate": 0.00039325791066918925, + "loss": 0.4004, + "step": 106 + }, + { + "epoch": 0.22338204592901878, + "grad_norm": 0.0394316515696973, + "learning_rate": 0.00039308267735402156, + "loss": 0.2734, + "step": 107 + }, + { + "epoch": 0.2254697286012526, + "grad_norm": 0.04552211598084851, + "learning_rate": 0.00039290523597709765, + "loss": 0.3555, + "step": 108 + }, + { + "epoch": 0.22755741127348644, + "grad_norm": 0.044355074180719295, + "learning_rate": 0.00039272558856760797, + "loss": 0.3418, + "step": 109 + }, + { + "epoch": 0.22964509394572025, + "grad_norm": 0.04664869616431499, + "learning_rate": 0.0003925437371799708, + "loss": 0.3301, + "step": 110 + }, + { + "epoch": 0.23173277661795408, + "grad_norm": 0.04421582409863553, + "learning_rate": 0.0003923596838938086, + "loss": 0.3242, + "step": 111 + }, + { + "epoch": 0.23382045929018788, + "grad_norm": 0.04914905461390972, + "learning_rate": 0.00039217343081392463, + "loss": 0.3027, + "step": 112 + }, + { + "epoch": 0.2359081419624217, + "grad_norm": 0.04801403896872833, + "learning_rate": 0.00039198498007027836, + "loss": 0.377, + "step": 113 + }, + { + "epoch": 0.23799582463465555, + "grad_norm": 0.039909379565673524, + "learning_rate": 0.0003917943338179616, + "loss": 0.2949, + "step": 114 + }, + { + "epoch": 0.24008350730688935, + "grad_norm": 0.04562132710784955, + "learning_rate": 0.00039160149423717353, + "loss": 0.3828, + "step": 115 + }, + { + "epoch": 0.24217118997912318, + "grad_norm": 0.04218187984586478, + "learning_rate": 0.00039140646353319584, + "loss": 0.3379, + "step": 116 + }, + { + "epoch": 0.24425887265135698, + "grad_norm": 0.13995497079449562, + "learning_rate": 0.0003912092439363677, + "loss": 0.3887, + "step": 117 + }, + { + "epoch": 0.24634655532359082, + "grad_norm": 0.03956892044992931, + "learning_rate": 0.00039100983770205995, + "loss": 0.3027, + "step": 118 + }, + { + "epoch": 0.24843423799582465, + "grad_norm": 0.043965744196170775, + "learning_rate": 0.00039080824711064946, + "loss": 0.3223, + "step": 119 + }, + { + "epoch": 0.25052192066805845, + "grad_norm": 0.044644460638659476, + "learning_rate": 0.00039060447446749314, + "loss": 0.3496, + "step": 120 + }, + { + "epoch": 0.25260960334029225, + "grad_norm": 0.05275646603548191, + "learning_rate": 0.00039039852210290147, + "loss": 0.3438, + "step": 121 + }, + { + "epoch": 0.2546972860125261, + "grad_norm": 0.04064451365674439, + "learning_rate": 0.00039019039237211175, + "loss": 0.3359, + "step": 122 + }, + { + "epoch": 0.2567849686847599, + "grad_norm": 0.048323141931080164, + "learning_rate": 0.00038998008765526136, + "loss": 0.377, + "step": 123 + }, + { + "epoch": 0.2588726513569937, + "grad_norm": 0.03725290930480903, + "learning_rate": 0.00038976761035736064, + "loss": 0.2988, + "step": 124 + }, + { + "epoch": 0.2609603340292276, + "grad_norm": 0.04052181522164268, + "learning_rate": 0.00038955296290826483, + "loss": 0.3203, + "step": 125 + }, + { + "epoch": 0.2630480167014614, + "grad_norm": 0.04516920645500193, + "learning_rate": 0.00038933614776264695, + "loss": 0.3496, + "step": 126 + }, + { + "epoch": 0.2651356993736952, + "grad_norm": 0.04887108820007405, + "learning_rate": 0.00038911716739996945, + "loss": 0.2734, + "step": 127 + }, + { + "epoch": 0.267223382045929, + "grad_norm": 0.04662438292740599, + "learning_rate": 0.0003888960243244556, + "loss": 0.2949, + "step": 128 + }, + { + "epoch": 0.26931106471816285, + "grad_norm": 0.0462101738931956, + "learning_rate": 0.0003886727210650613, + "loss": 0.3711, + "step": 129 + }, + { + "epoch": 0.27139874739039666, + "grad_norm": 0.045514397288015096, + "learning_rate": 0.00038844726017544606, + "loss": 0.3945, + "step": 130 + }, + { + "epoch": 0.27348643006263046, + "grad_norm": 0.04564176989369504, + "learning_rate": 0.00038821964423394335, + "loss": 0.3262, + "step": 131 + }, + { + "epoch": 0.2755741127348643, + "grad_norm": 0.0480634271209225, + "learning_rate": 0.00038798987584353175, + "loss": 0.4297, + "step": 132 + }, + { + "epoch": 0.2776617954070981, + "grad_norm": 0.04064367156377542, + "learning_rate": 0.0003877579576318048, + "loss": 0.3066, + "step": 133 + }, + { + "epoch": 0.2797494780793319, + "grad_norm": 0.06187073759126634, + "learning_rate": 0.000387523892250941, + "loss": 0.3828, + "step": 134 + }, + { + "epoch": 0.2818371607515658, + "grad_norm": 0.047235774434653544, + "learning_rate": 0.0003872876823776737, + "loss": 0.4551, + "step": 135 + }, + { + "epoch": 0.2839248434237996, + "grad_norm": 0.04613794323687757, + "learning_rate": 0.00038704933071326004, + "loss": 0.3555, + "step": 136 + }, + { + "epoch": 0.2860125260960334, + "grad_norm": 0.04123661418763717, + "learning_rate": 0.00038680883998345046, + "loss": 0.3848, + "step": 137 + }, + { + "epoch": 0.2881002087682672, + "grad_norm": 0.03990534224519547, + "learning_rate": 0.0003865662129384575, + "loss": 0.3223, + "step": 138 + }, + { + "epoch": 0.29018789144050106, + "grad_norm": 0.03964108882883799, + "learning_rate": 0.00038632145235292407, + "loss": 0.3145, + "step": 139 + }, + { + "epoch": 0.29227557411273486, + "grad_norm": 0.04771985690919652, + "learning_rate": 0.00038607456102589217, + "loss": 0.3555, + "step": 140 + }, + { + "epoch": 0.29436325678496866, + "grad_norm": 0.041459959268485676, + "learning_rate": 0.0003858255417807701, + "loss": 0.3203, + "step": 141 + }, + { + "epoch": 0.2964509394572025, + "grad_norm": 0.04488781487797441, + "learning_rate": 0.0003855743974653014, + "loss": 0.3516, + "step": 142 + }, + { + "epoch": 0.2985386221294363, + "grad_norm": 0.045595447329589846, + "learning_rate": 0.00038532113095153086, + "loss": 0.4121, + "step": 143 + }, + { + "epoch": 0.30062630480167013, + "grad_norm": 0.051789411309846184, + "learning_rate": 0.0003850657451357729, + "loss": 0.3945, + "step": 144 + }, + { + "epoch": 0.302713987473904, + "grad_norm": 0.04655813617947128, + "learning_rate": 0.00038480824293857756, + "loss": 0.377, + "step": 145 + }, + { + "epoch": 0.3048016701461378, + "grad_norm": 0.07648712725066147, + "learning_rate": 0.00038454862730469773, + "loss": 0.3438, + "step": 146 + }, + { + "epoch": 0.3068893528183716, + "grad_norm": 0.05310195833383878, + "learning_rate": 0.0003842869012030551, + "loss": 0.3594, + "step": 147 + }, + { + "epoch": 0.3089770354906054, + "grad_norm": 0.04391887584861711, + "learning_rate": 0.00038402306762670636, + "loss": 0.3555, + "step": 148 + }, + { + "epoch": 0.31106471816283926, + "grad_norm": 0.038985625260836355, + "learning_rate": 0.000383757129592809, + "loss": 0.2969, + "step": 149 + }, + { + "epoch": 0.31315240083507306, + "grad_norm": 0.050635143690619895, + "learning_rate": 0.00038348909014258647, + "loss": 0.3398, + "step": 150 + }, + { + "epoch": 0.31524008350730687, + "grad_norm": 0.04475955558328409, + "learning_rate": 0.00038321895234129394, + "loss": 0.3594, + "step": 151 + }, + { + "epoch": 0.3173277661795407, + "grad_norm": 0.046705792421880006, + "learning_rate": 0.0003829467192781829, + "loss": 0.3906, + "step": 152 + }, + { + "epoch": 0.31941544885177453, + "grad_norm": 0.04135295822439203, + "learning_rate": 0.00038267239406646595, + "loss": 0.3203, + "step": 153 + }, + { + "epoch": 0.32150313152400833, + "grad_norm": 0.03936648551856702, + "learning_rate": 0.000382395979843281, + "loss": 0.3086, + "step": 154 + }, + { + "epoch": 0.3235908141962422, + "grad_norm": 0.04411342571947294, + "learning_rate": 0.0003821174797696558, + "loss": 0.3145, + "step": 155 + }, + { + "epoch": 0.325678496868476, + "grad_norm": 0.0440989009123292, + "learning_rate": 0.0003818368970304713, + "loss": 0.3398, + "step": 156 + }, + { + "epoch": 0.3277661795407098, + "grad_norm": 0.04623084733338215, + "learning_rate": 0.00038155423483442565, + "loss": 0.373, + "step": 157 + }, + { + "epoch": 0.3298538622129436, + "grad_norm": 0.03851250195995776, + "learning_rate": 0.00038126949641399725, + "loss": 0.2773, + "step": 158 + }, + { + "epoch": 0.33194154488517746, + "grad_norm": 0.04376944212115029, + "learning_rate": 0.0003809826850254078, + "loss": 0.3613, + "step": 159 + }, + { + "epoch": 0.33402922755741127, + "grad_norm": 0.044952970087986426, + "learning_rate": 0.00038069380394858526, + "loss": 0.2734, + "step": 160 + }, + { + "epoch": 0.33611691022964507, + "grad_norm": 0.04338817418637471, + "learning_rate": 0.0003804028564871262, + "loss": 0.334, + "step": 161 + }, + { + "epoch": 0.33820459290187893, + "grad_norm": 0.04540609013390802, + "learning_rate": 0.00038010984596825774, + "loss": 0.3086, + "step": 162 + }, + { + "epoch": 0.34029227557411273, + "grad_norm": 0.040675218084428574, + "learning_rate": 0.00037981477574280015, + "loss": 0.3105, + "step": 163 + }, + { + "epoch": 0.34237995824634654, + "grad_norm": 0.04628740772476471, + "learning_rate": 0.000379517649185128, + "loss": 0.375, + "step": 164 + }, + { + "epoch": 0.3444676409185804, + "grad_norm": 0.045320126800110626, + "learning_rate": 0.0003792184696931318, + "loss": 0.3926, + "step": 165 + }, + { + "epoch": 0.3465553235908142, + "grad_norm": 0.047079880323708634, + "learning_rate": 0.00037891724068817896, + "loss": 0.3613, + "step": 166 + }, + { + "epoch": 0.348643006263048, + "grad_norm": 0.042006341910680284, + "learning_rate": 0.0003786139656150749, + "loss": 0.3848, + "step": 167 + }, + { + "epoch": 0.35073068893528186, + "grad_norm": 0.042457959635259945, + "learning_rate": 0.0003783086479420235, + "loss": 0.3594, + "step": 168 + }, + { + "epoch": 0.35281837160751567, + "grad_norm": 0.04729445116359038, + "learning_rate": 0.0003780012911605875, + "loss": 0.4023, + "step": 169 + }, + { + "epoch": 0.35490605427974947, + "grad_norm": 0.04490557924422057, + "learning_rate": 0.00037769189878564843, + "loss": 0.3184, + "step": 170 + }, + { + "epoch": 0.3569937369519833, + "grad_norm": 0.04659805592024877, + "learning_rate": 0.0003773804743553667, + "loss": 0.3613, + "step": 171 + }, + { + "epoch": 0.35908141962421714, + "grad_norm": 0.0457558395135587, + "learning_rate": 0.00037706702143114085, + "loss": 0.3848, + "step": 172 + }, + { + "epoch": 0.36116910229645094, + "grad_norm": 0.042225345661087835, + "learning_rate": 0.0003767515435975671, + "loss": 0.3652, + "step": 173 + }, + { + "epoch": 0.36325678496868474, + "grad_norm": 0.03862535581909019, + "learning_rate": 0.00037643404446239805, + "loss": 0.2578, + "step": 174 + }, + { + "epoch": 0.3653444676409186, + "grad_norm": 0.04511245438490728, + "learning_rate": 0.0003761145276565015, + "loss": 0.3789, + "step": 175 + }, + { + "epoch": 0.3674321503131524, + "grad_norm": 0.04418007608895449, + "learning_rate": 0.0003757929968338193, + "loss": 0.375, + "step": 176 + }, + { + "epoch": 0.3695198329853862, + "grad_norm": 0.04131570246472954, + "learning_rate": 0.00037546945567132486, + "loss": 0.3262, + "step": 177 + }, + { + "epoch": 0.37160751565762007, + "grad_norm": 0.04432885957928793, + "learning_rate": 0.0003751439078689819, + "loss": 0.3105, + "step": 178 + }, + { + "epoch": 0.3736951983298539, + "grad_norm": 0.04130281228669527, + "learning_rate": 0.0003748163571497015, + "loss": 0.3145, + "step": 179 + }, + { + "epoch": 0.3757828810020877, + "grad_norm": 0.041235332477716484, + "learning_rate": 0.00037448680725929987, + "loss": 0.3086, + "step": 180 + }, + { + "epoch": 0.3778705636743215, + "grad_norm": 0.046591187622845885, + "learning_rate": 0.00037415526196645535, + "loss": 0.377, + "step": 181 + }, + { + "epoch": 0.37995824634655534, + "grad_norm": 0.043131270347982746, + "learning_rate": 0.00037382172506266544, + "loss": 0.3223, + "step": 182 + }, + { + "epoch": 0.38204592901878914, + "grad_norm": 0.04113286864493325, + "learning_rate": 0.00037348620036220325, + "loss": 0.3379, + "step": 183 + }, + { + "epoch": 0.38413361169102295, + "grad_norm": 0.045126082835651934, + "learning_rate": 0.00037314869170207427, + "loss": 0.3906, + "step": 184 + }, + { + "epoch": 0.3862212943632568, + "grad_norm": 0.0468308767634132, + "learning_rate": 0.00037280920294197186, + "loss": 0.418, + "step": 185 + }, + { + "epoch": 0.3883089770354906, + "grad_norm": 0.041909756539268275, + "learning_rate": 0.0003724677379642338, + "loss": 0.3008, + "step": 186 + }, + { + "epoch": 0.3903966597077244, + "grad_norm": 0.039249830774930286, + "learning_rate": 0.0003721243006737973, + "loss": 0.293, + "step": 187 + }, + { + "epoch": 0.3924843423799583, + "grad_norm": 0.043933299859950536, + "learning_rate": 0.0003717788949981549, + "loss": 0.3281, + "step": 188 + }, + { + "epoch": 0.3945720250521921, + "grad_norm": 0.045457197762689416, + "learning_rate": 0.0003714315248873091, + "loss": 0.416, + "step": 189 + }, + { + "epoch": 0.3966597077244259, + "grad_norm": 0.038856886092809004, + "learning_rate": 0.00037108219431372734, + "loss": 0.3242, + "step": 190 + }, + { + "epoch": 0.3987473903966597, + "grad_norm": 0.03887481602609131, + "learning_rate": 0.0003707309072722966, + "loss": 0.3047, + "step": 191 + }, + { + "epoch": 0.40083507306889354, + "grad_norm": 0.04338847469385484, + "learning_rate": 0.0003703776677802778, + "loss": 0.3945, + "step": 192 + }, + { + "epoch": 0.40292275574112735, + "grad_norm": 0.04874209932196092, + "learning_rate": 0.0003700224798772597, + "loss": 0.4316, + "step": 193 + }, + { + "epoch": 0.40501043841336115, + "grad_norm": 0.0409033493154823, + "learning_rate": 0.00036966534762511276, + "loss": 0.3301, + "step": 194 + }, + { + "epoch": 0.407098121085595, + "grad_norm": 0.040056909386716535, + "learning_rate": 0.0003693062751079427, + "loss": 0.3887, + "step": 195 + }, + { + "epoch": 0.4091858037578288, + "grad_norm": 0.03764639519212107, + "learning_rate": 0.0003689452664320437, + "loss": 0.3125, + "step": 196 + }, + { + "epoch": 0.4112734864300626, + "grad_norm": 0.04012224378405559, + "learning_rate": 0.00036858232572585173, + "loss": 0.3574, + "step": 197 + }, + { + "epoch": 0.4133611691022965, + "grad_norm": 0.04149192047957985, + "learning_rate": 0.00036821745713989706, + "loss": 0.3242, + "step": 198 + }, + { + "epoch": 0.4154488517745303, + "grad_norm": 0.05017075921477889, + "learning_rate": 0.00036785066484675683, + "loss": 0.3652, + "step": 199 + }, + { + "epoch": 0.4175365344467641, + "grad_norm": 0.04227941718590804, + "learning_rate": 0.00036748195304100735, + "loss": 0.3242, + "step": 200 + }, + { + "epoch": 0.4196242171189979, + "grad_norm": 0.03768968002704521, + "learning_rate": 0.00036711132593917635, + "loss": 0.3105, + "step": 201 + }, + { + "epoch": 0.42171189979123175, + "grad_norm": 0.04194019808640771, + "learning_rate": 0.00036673878777969426, + "loss": 0.2891, + "step": 202 + }, + { + "epoch": 0.42379958246346555, + "grad_norm": 0.04542209099244845, + "learning_rate": 0.00036636434282284633, + "loss": 0.3359, + "step": 203 + }, + { + "epoch": 0.42588726513569936, + "grad_norm": 0.03795072966054866, + "learning_rate": 0.0003659879953507235, + "loss": 0.2178, + "step": 204 + }, + { + "epoch": 0.4279749478079332, + "grad_norm": 0.0676312462445889, + "learning_rate": 0.0003656097496671737, + "loss": 0.4004, + "step": 205 + }, + { + "epoch": 0.430062630480167, + "grad_norm": 0.06595187551394353, + "learning_rate": 0.00036522961009775217, + "loss": 0.3359, + "step": 206 + }, + { + "epoch": 0.4321503131524008, + "grad_norm": 0.048238875202028735, + "learning_rate": 0.0003648475809896727, + "loss": 0.375, + "step": 207 + }, + { + "epoch": 0.4342379958246347, + "grad_norm": 0.04774287102162698, + "learning_rate": 0.00036446366671175735, + "loss": 0.3613, + "step": 208 + }, + { + "epoch": 0.4363256784968685, + "grad_norm": 0.04447429752378728, + "learning_rate": 0.0003640778716543866, + "loss": 0.3594, + "step": 209 + }, + { + "epoch": 0.4384133611691023, + "grad_norm": 0.04069940254145397, + "learning_rate": 0.00036369020022944946, + "loss": 0.3105, + "step": 210 + }, + { + "epoch": 0.4405010438413361, + "grad_norm": 0.03646181755503713, + "learning_rate": 0.0003633006568702924, + "loss": 0.3047, + "step": 211 + }, + { + "epoch": 0.44258872651356995, + "grad_norm": 0.041622368664595316, + "learning_rate": 0.00036290924603166945, + "loss": 0.3418, + "step": 212 + }, + { + "epoch": 0.44467640918580376, + "grad_norm": 0.04687265549200718, + "learning_rate": 0.00036251597218969045, + "loss": 0.3926, + "step": 213 + }, + { + "epoch": 0.44676409185803756, + "grad_norm": 0.05629699568400707, + "learning_rate": 0.0003621208398417705, + "loss": 0.3848, + "step": 214 + }, + { + "epoch": 0.4488517745302714, + "grad_norm": 0.04583925885988264, + "learning_rate": 0.0003617238535065782, + "loss": 0.334, + "step": 215 + }, + { + "epoch": 0.4509394572025052, + "grad_norm": 0.04693126530215214, + "learning_rate": 0.0003613250177239838, + "loss": 0.3691, + "step": 216 + }, + { + "epoch": 0.453027139874739, + "grad_norm": 0.05342936402981673, + "learning_rate": 0.00036092433705500797, + "loss": 0.3301, + "step": 217 + }, + { + "epoch": 0.4551148225469729, + "grad_norm": 0.04089550326814837, + "learning_rate": 0.000360521816081769, + "loss": 0.2793, + "step": 218 + }, + { + "epoch": 0.4572025052192067, + "grad_norm": 0.04725556069406023, + "learning_rate": 0.00036011745940743036, + "loss": 0.3887, + "step": 219 + }, + { + "epoch": 0.4592901878914405, + "grad_norm": 0.04629637553798221, + "learning_rate": 0.0003597112716561486, + "loss": 0.3477, + "step": 220 + }, + { + "epoch": 0.4613778705636743, + "grad_norm": 0.042829084723361624, + "learning_rate": 0.0003593032574730201, + "loss": 0.3379, + "step": 221 + }, + { + "epoch": 0.46346555323590816, + "grad_norm": 0.04290003185482568, + "learning_rate": 0.000358893421524028, + "loss": 0.3828, + "step": 222 + }, + { + "epoch": 0.46555323590814196, + "grad_norm": 0.03695634767631775, + "learning_rate": 0.0003584817684959888, + "loss": 0.2539, + "step": 223 + }, + { + "epoch": 0.46764091858037576, + "grad_norm": 0.043249882204696076, + "learning_rate": 0.00035806830309649887, + "loss": 0.291, + "step": 224 + }, + { + "epoch": 0.4697286012526096, + "grad_norm": 0.0426499927716365, + "learning_rate": 0.0003576530300538805, + "loss": 0.3262, + "step": 225 + }, + { + "epoch": 0.4718162839248434, + "grad_norm": 0.038269992068262225, + "learning_rate": 0.00035723595411712793, + "loss": 0.2598, + "step": 226 + }, + { + "epoch": 0.47390396659707723, + "grad_norm": 0.039227521535724244, + "learning_rate": 0.0003568170800558531, + "loss": 0.2812, + "step": 227 + }, + { + "epoch": 0.4759916492693111, + "grad_norm": 0.0414687007362602, + "learning_rate": 0.0003563964126602308, + "loss": 0.3184, + "step": 228 + }, + { + "epoch": 0.4780793319415449, + "grad_norm": 0.06563073201350628, + "learning_rate": 0.0003559739567409443, + "loss": 0.2969, + "step": 229 + }, + { + "epoch": 0.4801670146137787, + "grad_norm": 0.04223079865704053, + "learning_rate": 0.00035554971712913004, + "loss": 0.334, + "step": 230 + }, + { + "epoch": 0.4822546972860125, + "grad_norm": 0.042708614625140126, + "learning_rate": 0.00035512369867632236, + "loss": 0.3418, + "step": 231 + }, + { + "epoch": 0.48434237995824636, + "grad_norm": 0.04063103807491919, + "learning_rate": 0.00035469590625439843, + "loss": 0.3477, + "step": 232 + }, + { + "epoch": 0.48643006263048016, + "grad_norm": 0.045147549230686396, + "learning_rate": 0.00035426634475552187, + "loss": 0.3516, + "step": 233 + }, + { + "epoch": 0.48851774530271397, + "grad_norm": 0.04106108808167586, + "learning_rate": 0.00035383501909208755, + "loss": 0.2812, + "step": 234 + }, + { + "epoch": 0.4906054279749478, + "grad_norm": 0.0428728366141268, + "learning_rate": 0.0003534019341966647, + "loss": 0.3398, + "step": 235 + }, + { + "epoch": 0.49269311064718163, + "grad_norm": 0.047772722106343216, + "learning_rate": 0.00035296709502194115, + "loss": 0.3848, + "step": 236 + }, + { + "epoch": 0.49478079331941544, + "grad_norm": 0.04629142912114304, + "learning_rate": 0.000352530506540666, + "loss": 0.2715, + "step": 237 + }, + { + "epoch": 0.4968684759916493, + "grad_norm": 0.05018937450568595, + "learning_rate": 0.0003520921737455935, + "loss": 0.3027, + "step": 238 + }, + { + "epoch": 0.4989561586638831, + "grad_norm": 0.04431763914166162, + "learning_rate": 0.00035165210164942535, + "loss": 0.3477, + "step": 239 + }, + { + "epoch": 0.5010438413361169, + "grad_norm": 0.04080914710858152, + "learning_rate": 0.00035121029528475386, + "loss": 0.3086, + "step": 240 + }, + { + "epoch": 0.5031315240083507, + "grad_norm": 0.04222028074641329, + "learning_rate": 0.0003507667597040038, + "loss": 0.3672, + "step": 241 + }, + { + "epoch": 0.5052192066805845, + "grad_norm": 0.04367534710869874, + "learning_rate": 0.0003503214999793752, + "loss": 0.3223, + "step": 242 + }, + { + "epoch": 0.5073068893528184, + "grad_norm": 0.04813953280858226, + "learning_rate": 0.00034987452120278505, + "loss": 0.3984, + "step": 243 + }, + { + "epoch": 0.5093945720250522, + "grad_norm": 0.046413128010797784, + "learning_rate": 0.00034942582848580925, + "loss": 0.3848, + "step": 244 + }, + { + "epoch": 0.511482254697286, + "grad_norm": 0.036986531982911595, + "learning_rate": 0.0003489754269596239, + "loss": 0.3066, + "step": 245 + }, + { + "epoch": 0.5135699373695198, + "grad_norm": 0.040001114155947204, + "learning_rate": 0.00034852332177494685, + "loss": 0.2852, + "step": 246 + }, + { + "epoch": 0.5156576200417536, + "grad_norm": 0.04605884877382518, + "learning_rate": 0.0003480695181019786, + "loss": 0.3535, + "step": 247 + }, + { + "epoch": 0.5177453027139874, + "grad_norm": 0.0406658437248984, + "learning_rate": 0.0003476140211303433, + "loss": 0.334, + "step": 248 + }, + { + "epoch": 0.5198329853862212, + "grad_norm": 0.03909067687413047, + "learning_rate": 0.0003471568360690296, + "loss": 0.2871, + "step": 249 + }, + { + "epoch": 0.5219206680584552, + "grad_norm": 0.04831389973643467, + "learning_rate": 0.00034669796814633055, + "loss": 0.3398, + "step": 250 + }, + { + "epoch": 0.524008350730689, + "grad_norm": 0.04408464727595837, + "learning_rate": 0.0003462374226097844, + "loss": 0.3242, + "step": 251 + }, + { + "epoch": 0.5260960334029228, + "grad_norm": 0.03798160874377182, + "learning_rate": 0.00034577520472611414, + "loss": 0.252, + "step": 252 + }, + { + "epoch": 0.5281837160751566, + "grad_norm": 0.04625762493599334, + "learning_rate": 0.0003453113197811675, + "loss": 0.334, + "step": 253 + }, + { + "epoch": 0.5302713987473904, + "grad_norm": 0.03757333278335248, + "learning_rate": 0.0003448457730798564, + "loss": 0.2734, + "step": 254 + }, + { + "epoch": 0.5323590814196242, + "grad_norm": 0.038808091023111425, + "learning_rate": 0.0003443785699460965, + "loss": 0.3652, + "step": 255 + }, + { + "epoch": 0.534446764091858, + "grad_norm": 0.03797009162534055, + "learning_rate": 0.0003439097157227459, + "loss": 0.3223, + "step": 256 + }, + { + "epoch": 0.5365344467640919, + "grad_norm": 0.038601658815291935, + "learning_rate": 0.0003434392157715445, + "loss": 0.373, + "step": 257 + }, + { + "epoch": 0.5386221294363257, + "grad_norm": 0.04680506137895324, + "learning_rate": 0.00034296707547305235, + "loss": 0.334, + "step": 258 + }, + { + "epoch": 0.5407098121085595, + "grad_norm": 0.041499005757660336, + "learning_rate": 0.0003424933002265883, + "loss": 0.3457, + "step": 259 + }, + { + "epoch": 0.5427974947807933, + "grad_norm": 0.04115196540841461, + "learning_rate": 0.00034201789545016824, + "loss": 0.3242, + "step": 260 + }, + { + "epoch": 0.5448851774530271, + "grad_norm": 0.04340048389863411, + "learning_rate": 0.00034154086658044297, + "loss": 0.2949, + "step": 261 + }, + { + "epoch": 0.5469728601252609, + "grad_norm": 0.04378711119505926, + "learning_rate": 0.0003410622190726362, + "loss": 0.3066, + "step": 262 + }, + { + "epoch": 0.5490605427974948, + "grad_norm": 0.04643412727247773, + "learning_rate": 0.00034058195840048225, + "loss": 0.3652, + "step": 263 + }, + { + "epoch": 0.5511482254697286, + "grad_norm": 0.042043153196747375, + "learning_rate": 0.0003401000900561631, + "loss": 0.3789, + "step": 264 + }, + { + "epoch": 0.5532359081419624, + "grad_norm": 0.040040695052475275, + "learning_rate": 0.00033961661955024595, + "loss": 0.3145, + "step": 265 + }, + { + "epoch": 0.5553235908141962, + "grad_norm": 0.038423365462145156, + "learning_rate": 0.0003391315524116202, + "loss": 0.3223, + "step": 266 + }, + { + "epoch": 0.55741127348643, + "grad_norm": 0.038539884775496434, + "learning_rate": 0.00033864489418743355, + "loss": 0.3086, + "step": 267 + }, + { + "epoch": 0.5594989561586639, + "grad_norm": 0.04422443806977669, + "learning_rate": 0.0003381566504430298, + "loss": 0.3945, + "step": 268 + }, + { + "epoch": 0.5615866388308977, + "grad_norm": 0.0377661934221125, + "learning_rate": 0.00033766682676188386, + "loss": 0.3398, + "step": 269 + }, + { + "epoch": 0.5636743215031316, + "grad_norm": 0.036534819703584714, + "learning_rate": 0.00033717542874553896, + "loss": 0.3418, + "step": 270 + }, + { + "epoch": 0.5657620041753654, + "grad_norm": 0.04440346943789282, + "learning_rate": 0.0003366824620135421, + "loss": 0.3984, + "step": 271 + }, + { + "epoch": 0.5678496868475992, + "grad_norm": 0.037034787962705526, + "learning_rate": 0.00033618793220337965, + "loss": 0.2812, + "step": 272 + }, + { + "epoch": 0.569937369519833, + "grad_norm": 0.03927620752143223, + "learning_rate": 0.0003356918449704133, + "loss": 0.2695, + "step": 273 + }, + { + "epoch": 0.5720250521920668, + "grad_norm": 0.04369797108825467, + "learning_rate": 0.00033519420598781505, + "loss": 0.3477, + "step": 274 + }, + { + "epoch": 0.5741127348643006, + "grad_norm": 0.0379874901088521, + "learning_rate": 0.0003346950209465025, + "loss": 0.248, + "step": 275 + }, + { + "epoch": 0.5762004175365344, + "grad_norm": 0.042755557159503056, + "learning_rate": 0.0003341942955550738, + "loss": 0.3242, + "step": 276 + }, + { + "epoch": 0.5782881002087683, + "grad_norm": 0.039892604797522456, + "learning_rate": 0.0003336920355397421, + "loss": 0.3125, + "step": 277 + }, + { + "epoch": 0.5803757828810021, + "grad_norm": 0.037020438684393274, + "learning_rate": 0.0003331882466442704, + "loss": 0.2832, + "step": 278 + }, + { + "epoch": 0.5824634655532359, + "grad_norm": 0.0448544789689836, + "learning_rate": 0.0003326829346299057, + "loss": 0.3613, + "step": 279 + }, + { + "epoch": 0.5845511482254697, + "grad_norm": 0.037923744133296695, + "learning_rate": 0.00033217610527531315, + "loss": 0.3262, + "step": 280 + }, + { + "epoch": 0.5866388308977035, + "grad_norm": 0.039536585438320955, + "learning_rate": 0.00033166776437650985, + "loss": 0.3301, + "step": 281 + }, + { + "epoch": 0.5887265135699373, + "grad_norm": 0.042309874793804436, + "learning_rate": 0.0003311579177467988, + "loss": 0.3281, + "step": 282 + }, + { + "epoch": 0.5908141962421712, + "grad_norm": 0.04247524788442754, + "learning_rate": 0.0003306465712167023, + "loss": 0.3281, + "step": 283 + }, + { + "epoch": 0.592901878914405, + "grad_norm": 0.038298773282535274, + "learning_rate": 0.0003301337306338951, + "loss": 0.3066, + "step": 284 + }, + { + "epoch": 0.5949895615866388, + "grad_norm": 0.04845680190635201, + "learning_rate": 0.00032961940186313784, + "loss": 0.3867, + "step": 285 + }, + { + "epoch": 0.5970772442588727, + "grad_norm": 0.04425999769437242, + "learning_rate": 0.00032910359078620974, + "loss": 0.3848, + "step": 286 + }, + { + "epoch": 0.5991649269311065, + "grad_norm": 0.04350677333004508, + "learning_rate": 0.0003285863033018416, + "loss": 0.3359, + "step": 287 + }, + { + "epoch": 0.6012526096033403, + "grad_norm": 0.04313248762187342, + "learning_rate": 0.00032806754532564795, + "loss": 0.3281, + "step": 288 + }, + { + "epoch": 0.6033402922755741, + "grad_norm": 0.04172934704529679, + "learning_rate": 0.00032754732279005974, + "loss": 0.3887, + "step": 289 + }, + { + "epoch": 0.605427974947808, + "grad_norm": 0.053530546182614214, + "learning_rate": 0.0003270256416442564, + "loss": 0.4512, + "step": 290 + }, + { + "epoch": 0.6075156576200418, + "grad_norm": 0.0585588589735363, + "learning_rate": 0.0003265025078540977, + "loss": 0.4023, + "step": 291 + }, + { + "epoch": 0.6096033402922756, + "grad_norm": 0.046794215765212366, + "learning_rate": 0.00032597792740205573, + "loss": 0.2871, + "step": 292 + }, + { + "epoch": 0.6116910229645094, + "grad_norm": 0.04076085980117636, + "learning_rate": 0.0003254519062871462, + "loss": 0.3027, + "step": 293 + }, + { + "epoch": 0.6137787056367432, + "grad_norm": 0.039930806408977594, + "learning_rate": 0.0003249244505248603, + "loss": 0.3086, + "step": 294 + }, + { + "epoch": 0.615866388308977, + "grad_norm": 0.047339086645206876, + "learning_rate": 0.0003243955661470952, + "loss": 0.373, + "step": 295 + }, + { + "epoch": 0.6179540709812108, + "grad_norm": 0.041035006122051215, + "learning_rate": 0.00032386525920208574, + "loss": 0.3281, + "step": 296 + }, + { + "epoch": 0.6200417536534447, + "grad_norm": 0.05042533676976384, + "learning_rate": 0.000323333535754335, + "loss": 0.293, + "step": 297 + }, + { + "epoch": 0.6221294363256785, + "grad_norm": 0.043959165085212505, + "learning_rate": 0.00032280040188454494, + "loss": 0.3125, + "step": 298 + }, + { + "epoch": 0.6242171189979123, + "grad_norm": 0.04458635918949398, + "learning_rate": 0.0003222658636895469, + "loss": 0.3105, + "step": 299 + }, + { + "epoch": 0.6263048016701461, + "grad_norm": 0.04146656432155535, + "learning_rate": 0.0003217299272822316, + "loss": 0.3672, + "step": 300 + }, + { + "epoch": 0.6283924843423799, + "grad_norm": 0.04285146681080291, + "learning_rate": 0.00032119259879147987, + "loss": 0.2773, + "step": 301 + }, + { + "epoch": 0.6304801670146137, + "grad_norm": 0.04012756626164825, + "learning_rate": 0.0003206538843620919, + "loss": 0.3145, + "step": 302 + }, + { + "epoch": 0.6325678496868476, + "grad_norm": 0.03395547556102837, + "learning_rate": 0.0003201137901547175, + "loss": 0.2344, + "step": 303 + }, + { + "epoch": 0.6346555323590815, + "grad_norm": 0.038474167468473125, + "learning_rate": 0.00031957232234578506, + "loss": 0.3105, + "step": 304 + }, + { + "epoch": 0.6367432150313153, + "grad_norm": 0.04345850995580467, + "learning_rate": 0.00031902948712743146, + "loss": 0.3711, + "step": 305 + }, + { + "epoch": 0.6388308977035491, + "grad_norm": 0.04561991595493648, + "learning_rate": 0.00031848529070743113, + "loss": 0.3281, + "step": 306 + }, + { + "epoch": 0.6409185803757829, + "grad_norm": 0.043639261730465534, + "learning_rate": 0.00031793973930912465, + "loss": 0.3535, + "step": 307 + }, + { + "epoch": 0.6430062630480167, + "grad_norm": 0.041614849807266444, + "learning_rate": 0.00031739283917134835, + "loss": 0.293, + "step": 308 + }, + { + "epoch": 0.6450939457202505, + "grad_norm": 0.037829924843835225, + "learning_rate": 0.00031684459654836224, + "loss": 0.2793, + "step": 309 + }, + { + "epoch": 0.6471816283924844, + "grad_norm": 0.043593747485495736, + "learning_rate": 0.00031629501770977865, + "loss": 0.2871, + "step": 310 + }, + { + "epoch": 0.6492693110647182, + "grad_norm": 0.043504619259611736, + "learning_rate": 0.0003157441089404911, + "loss": 0.377, + "step": 311 + }, + { + "epoch": 0.651356993736952, + "grad_norm": 0.042969570910675, + "learning_rate": 0.00031519187654060153, + "loss": 0.3535, + "step": 312 + }, + { + "epoch": 0.6534446764091858, + "grad_norm": 0.04444146348618118, + "learning_rate": 0.0003146383268253489, + "loss": 0.3633, + "step": 313 + }, + { + "epoch": 0.6555323590814196, + "grad_norm": 0.0579233541293356, + "learning_rate": 0.00031408346612503684, + "loss": 0.2852, + "step": 314 + }, + { + "epoch": 0.6576200417536534, + "grad_norm": 0.04116612592402649, + "learning_rate": 0.00031352730078496103, + "loss": 0.3184, + "step": 315 + }, + { + "epoch": 0.6597077244258872, + "grad_norm": 0.042619485954835536, + "learning_rate": 0.0003129698371653368, + "loss": 0.3555, + "step": 316 + }, + { + "epoch": 0.6617954070981211, + "grad_norm": 0.03978568537420636, + "learning_rate": 0.00031241108164122667, + "loss": 0.3184, + "step": 317 + }, + { + "epoch": 0.6638830897703549, + "grad_norm": 0.04398595731711137, + "learning_rate": 0.0003118510406024667, + "loss": 0.3262, + "step": 318 + }, + { + "epoch": 0.6659707724425887, + "grad_norm": 0.03557319884714253, + "learning_rate": 0.0003112897204535943, + "loss": 0.2539, + "step": 319 + }, + { + "epoch": 0.6680584551148225, + "grad_norm": 0.04098653755334443, + "learning_rate": 0.0003107271276137745, + "loss": 0.3711, + "step": 320 + }, + { + "epoch": 0.6701461377870563, + "grad_norm": 0.0377819264468114, + "learning_rate": 0.0003101632685167263, + "loss": 0.2891, + "step": 321 + }, + { + "epoch": 0.6722338204592901, + "grad_norm": 0.038417025071128603, + "learning_rate": 0.00030959814961064973, + "loss": 0.2891, + "step": 322 + }, + { + "epoch": 0.6743215031315241, + "grad_norm": 0.038602243878265344, + "learning_rate": 0.0003090317773581518, + "loss": 0.2734, + "step": 323 + }, + { + "epoch": 0.6764091858037579, + "grad_norm": 0.0487821791964808, + "learning_rate": 0.0003084641582361723, + "loss": 0.3574, + "step": 324 + }, + { + "epoch": 0.6784968684759917, + "grad_norm": 0.04544872266451946, + "learning_rate": 0.00030789529873591034, + "loss": 0.3008, + "step": 325 + }, + { + "epoch": 0.6805845511482255, + "grad_norm": 0.0373160417363858, + "learning_rate": 0.0003073252053627496, + "loss": 0.2871, + "step": 326 + }, + { + "epoch": 0.6826722338204593, + "grad_norm": 0.04358868533626725, + "learning_rate": 0.0003067538846361841, + "loss": 0.3066, + "step": 327 + }, + { + "epoch": 0.6847599164926931, + "grad_norm": 0.040357136071415334, + "learning_rate": 0.0003061813430897439, + "loss": 0.3516, + "step": 328 + }, + { + "epoch": 0.6868475991649269, + "grad_norm": 0.04862770193509928, + "learning_rate": 0.0003056075872709199, + "loss": 0.4062, + "step": 329 + }, + { + "epoch": 0.6889352818371608, + "grad_norm": 0.04688881879011172, + "learning_rate": 0.00030503262374108927, + "loss": 0.4434, + "step": 330 + }, + { + "epoch": 0.6910229645093946, + "grad_norm": 0.04276287817278344, + "learning_rate": 0.00030445645907544057, + "loss": 0.3105, + "step": 331 + }, + { + "epoch": 0.6931106471816284, + "grad_norm": 0.03644902974160357, + "learning_rate": 0.0003038790998628981, + "loss": 0.249, + "step": 332 + }, + { + "epoch": 0.6951983298538622, + "grad_norm": 0.03602308126572102, + "learning_rate": 0.0003033005527060469, + "loss": 0.2891, + "step": 333 + }, + { + "epoch": 0.697286012526096, + "grad_norm": 0.0397502232733115, + "learning_rate": 0.0003027208242210571, + "loss": 0.334, + "step": 334 + }, + { + "epoch": 0.6993736951983298, + "grad_norm": 0.042614553937913115, + "learning_rate": 0.0003021399210376084, + "loss": 0.3281, + "step": 335 + }, + { + "epoch": 0.7014613778705637, + "grad_norm": 0.040360864280806166, + "learning_rate": 0.0003015578497988142, + "loss": 0.3145, + "step": 336 + }, + { + "epoch": 0.7035490605427975, + "grad_norm": 0.03697519450333998, + "learning_rate": 0.0003009746171611454, + "loss": 0.2852, + "step": 337 + }, + { + "epoch": 0.7056367432150313, + "grad_norm": 0.03920168369252357, + "learning_rate": 0.0003003902297943545, + "loss": 0.2578, + "step": 338 + }, + { + "epoch": 0.7077244258872651, + "grad_norm": 0.04596701530986532, + "learning_rate": 0.00029980469438139947, + "loss": 0.3066, + "step": 339 + }, + { + "epoch": 0.7098121085594989, + "grad_norm": 0.03590053569859233, + "learning_rate": 0.00029921801761836696, + "loss": 0.2598, + "step": 340 + }, + { + "epoch": 0.7118997912317327, + "grad_norm": 0.039137959610818826, + "learning_rate": 0.0002986302062143961, + "loss": 0.3398, + "step": 341 + }, + { + "epoch": 0.7139874739039666, + "grad_norm": 0.04445491792602957, + "learning_rate": 0.0002980412668916013, + "loss": 0.3398, + "step": 342 + }, + { + "epoch": 0.7160751565762005, + "grad_norm": 0.040340445047287156, + "learning_rate": 0.00029745120638499596, + "loss": 0.3516, + "step": 343 + }, + { + "epoch": 0.7181628392484343, + "grad_norm": 0.04520397395484933, + "learning_rate": 0.00029686003144241494, + "loss": 0.3555, + "step": 344 + }, + { + "epoch": 0.7202505219206681, + "grad_norm": 0.045695108134425295, + "learning_rate": 0.0002962677488244378, + "loss": 0.2871, + "step": 345 + }, + { + "epoch": 0.7223382045929019, + "grad_norm": 0.0414553610811136, + "learning_rate": 0.0002956743653043111, + "loss": 0.3828, + "step": 346 + }, + { + "epoch": 0.7244258872651357, + "grad_norm": 0.043267589906942225, + "learning_rate": 0.00029507988766787135, + "loss": 0.3262, + "step": 347 + }, + { + "epoch": 0.7265135699373695, + "grad_norm": 0.03665181460868117, + "learning_rate": 0.0002944843227134669, + "loss": 0.3301, + "step": 348 + }, + { + "epoch": 0.7286012526096033, + "grad_norm": 0.04220684212991339, + "learning_rate": 0.00029388767725188077, + "loss": 0.3223, + "step": 349 + }, + { + "epoch": 0.7306889352818372, + "grad_norm": 0.04468007941423499, + "learning_rate": 0.0002932899581062524, + "loss": 0.2949, + "step": 350 + }, + { + "epoch": 0.732776617954071, + "grad_norm": 0.044433764048658723, + "learning_rate": 0.00029269117211199955, + "loss": 0.377, + "step": 351 + }, + { + "epoch": 0.7348643006263048, + "grad_norm": 0.046209035174827495, + "learning_rate": 0.0002920913261167405, + "loss": 0.3691, + "step": 352 + }, + { + "epoch": 0.7369519832985386, + "grad_norm": 0.040034133910176115, + "learning_rate": 0.00029149042698021544, + "loss": 0.3105, + "step": 353 + }, + { + "epoch": 0.7390396659707724, + "grad_norm": 0.04609016944022695, + "learning_rate": 0.00029088848157420813, + "loss": 0.3984, + "step": 354 + }, + { + "epoch": 0.7411273486430062, + "grad_norm": 0.04480775819697267, + "learning_rate": 0.00029028549678246715, + "loss": 0.332, + "step": 355 + }, + { + "epoch": 0.7432150313152401, + "grad_norm": 0.04387465996335384, + "learning_rate": 0.0002896814795006277, + "loss": 0.3984, + "step": 356 + }, + { + "epoch": 0.7453027139874739, + "grad_norm": 0.041864619992452806, + "learning_rate": 0.0002890764366361318, + "loss": 0.3633, + "step": 357 + }, + { + "epoch": 0.7473903966597077, + "grad_norm": 0.03717012201708029, + "learning_rate": 0.0002884703751081504, + "loss": 0.2324, + "step": 358 + }, + { + "epoch": 0.7494780793319415, + "grad_norm": 0.043251857860819434, + "learning_rate": 0.0002878633018475035, + "loss": 0.3066, + "step": 359 + }, + { + "epoch": 0.7515657620041754, + "grad_norm": 0.03591327407860244, + "learning_rate": 0.00028725522379658097, + "loss": 0.2852, + "step": 360 + }, + { + "epoch": 0.7536534446764092, + "grad_norm": 0.0381908896371197, + "learning_rate": 0.00028664614790926355, + "loss": 0.3477, + "step": 361 + }, + { + "epoch": 0.755741127348643, + "grad_norm": 0.04392871145759733, + "learning_rate": 0.000286036081150843, + "loss": 0.4414, + "step": 362 + }, + { + "epoch": 0.7578288100208769, + "grad_norm": 0.04392250979416697, + "learning_rate": 0.0002854250304979423, + "loss": 0.2832, + "step": 363 + }, + { + "epoch": 0.7599164926931107, + "grad_norm": 0.03445036859255453, + "learning_rate": 0.0002848130029384363, + "loss": 0.3027, + "step": 364 + }, + { + "epoch": 0.7620041753653445, + "grad_norm": 0.03588315051937214, + "learning_rate": 0.00028420000547137175, + "loss": 0.2793, + "step": 365 + }, + { + "epoch": 0.7640918580375783, + "grad_norm": 0.043623631697390225, + "learning_rate": 0.00028358604510688666, + "loss": 0.3301, + "step": 366 + }, + { + "epoch": 0.7661795407098121, + "grad_norm": 0.044899266192686, + "learning_rate": 0.000282971128866131, + "loss": 0.2988, + "step": 367 + }, + { + "epoch": 0.7682672233820459, + "grad_norm": 0.044441044360934975, + "learning_rate": 0.0002823552637811857, + "loss": 0.3262, + "step": 368 + }, + { + "epoch": 0.7703549060542797, + "grad_norm": 0.04617336800060674, + "learning_rate": 0.00028173845689498286, + "loss": 0.2812, + "step": 369 + }, + { + "epoch": 0.7724425887265136, + "grad_norm": 0.03481084134760086, + "learning_rate": 0.0002811207152612243, + "loss": 0.2637, + "step": 370 + }, + { + "epoch": 0.7745302713987474, + "grad_norm": 0.04359599728927467, + "learning_rate": 0.000280502045944302, + "loss": 0.3516, + "step": 371 + }, + { + "epoch": 0.7766179540709812, + "grad_norm": 0.03910171209605865, + "learning_rate": 0.0002798824560192164, + "loss": 0.2793, + "step": 372 + }, + { + "epoch": 0.778705636743215, + "grad_norm": 0.04498875604580853, + "learning_rate": 0.0002792619525714962, + "loss": 0.334, + "step": 373 + }, + { + "epoch": 0.7807933194154488, + "grad_norm": 0.046945206399684386, + "learning_rate": 0.00027864054269711666, + "loss": 0.3262, + "step": 374 + }, + { + "epoch": 0.7828810020876826, + "grad_norm": 0.03720210905131818, + "learning_rate": 0.0002780182335024191, + "loss": 0.252, + "step": 375 + }, + { + "epoch": 0.7849686847599165, + "grad_norm": 0.03971464889707817, + "learning_rate": 0.00027739503210402907, + "loss": 0.334, + "step": 376 + }, + { + "epoch": 0.7870563674321504, + "grad_norm": 0.047941221203412314, + "learning_rate": 0.00027677094562877534, + "loss": 0.3066, + "step": 377 + }, + { + "epoch": 0.7891440501043842, + "grad_norm": 0.04116312254250454, + "learning_rate": 0.0002761459812136083, + "loss": 0.3652, + "step": 378 + }, + { + "epoch": 0.791231732776618, + "grad_norm": 0.0408608818482939, + "learning_rate": 0.00027552014600551813, + "loss": 0.3359, + "step": 379 + }, + { + "epoch": 0.7933194154488518, + "grad_norm": 0.039519399095894796, + "learning_rate": 0.00027489344716145344, + "loss": 0.3105, + "step": 380 + }, + { + "epoch": 0.7954070981210856, + "grad_norm": 0.04468756187207934, + "learning_rate": 0.0002742658918482392, + "loss": 0.334, + "step": 381 + }, + { + "epoch": 0.7974947807933194, + "grad_norm": 0.03799026650779495, + "learning_rate": 0.00027363748724249485, + "loss": 0.3125, + "step": 382 + }, + { + "epoch": 0.7995824634655533, + "grad_norm": 0.05260052294591301, + "learning_rate": 0.000273008240530552, + "loss": 0.4316, + "step": 383 + }, + { + "epoch": 0.8016701461377871, + "grad_norm": 0.04069771345832873, + "learning_rate": 0.00027237815890837283, + "loss": 0.2969, + "step": 384 + }, + { + "epoch": 0.8037578288100209, + "grad_norm": 0.039739122060491815, + "learning_rate": 0.000271747249581467, + "loss": 0.3555, + "step": 385 + }, + { + "epoch": 0.8058455114822547, + "grad_norm": 0.04610276901781518, + "learning_rate": 0.00027111551976480996, + "loss": 0.3574, + "step": 386 + }, + { + "epoch": 0.8079331941544885, + "grad_norm": 0.04374672870028719, + "learning_rate": 0.00027048297668276, + "loss": 0.3047, + "step": 387 + }, + { + "epoch": 0.8100208768267223, + "grad_norm": 0.04631564065384605, + "learning_rate": 0.0002698496275689758, + "loss": 0.2871, + "step": 388 + }, + { + "epoch": 0.8121085594989561, + "grad_norm": 0.049176525271089545, + "learning_rate": 0.00026921547966633373, + "loss": 0.2871, + "step": 389 + }, + { + "epoch": 0.81419624217119, + "grad_norm": 0.040234387326007176, + "learning_rate": 0.0002685805402268449, + "loss": 0.3301, + "step": 390 + }, + { + "epoch": 0.8162839248434238, + "grad_norm": 0.044678234660364975, + "learning_rate": 0.0002679448165115725, + "loss": 0.2852, + "step": 391 + }, + { + "epoch": 0.8183716075156576, + "grad_norm": 0.043749807508504976, + "learning_rate": 0.00026730831579054823, + "loss": 0.3418, + "step": 392 + }, + { + "epoch": 0.8204592901878914, + "grad_norm": 0.03702597280717622, + "learning_rate": 0.00026667104534268984, + "loss": 0.334, + "step": 393 + }, + { + "epoch": 0.8225469728601252, + "grad_norm": 0.04051649281576064, + "learning_rate": 0.00026603301245571733, + "loss": 0.3652, + "step": 394 + }, + { + "epoch": 0.824634655532359, + "grad_norm": 0.04070349409408308, + "learning_rate": 0.00026539422442607, + "loss": 0.3164, + "step": 395 + }, + { + "epoch": 0.826722338204593, + "grad_norm": 0.03685594277413145, + "learning_rate": 0.0002647546885588227, + "loss": 0.2695, + "step": 396 + }, + { + "epoch": 0.8288100208768268, + "grad_norm": 0.033959522628423805, + "learning_rate": 0.00026411441216760243, + "loss": 0.2422, + "step": 397 + }, + { + "epoch": 0.8308977035490606, + "grad_norm": 0.03649515909769935, + "learning_rate": 0.00026347340257450473, + "loss": 0.252, + "step": 398 + }, + { + "epoch": 0.8329853862212944, + "grad_norm": 0.05036202800279952, + "learning_rate": 0.00026283166711001, + "loss": 0.3203, + "step": 399 + }, + { + "epoch": 0.8350730688935282, + "grad_norm": 0.04724871652205951, + "learning_rate": 0.0002621892131128994, + "loss": 0.3555, + "step": 400 + }, + { + "epoch": 0.837160751565762, + "grad_norm": 0.04523467438521327, + "learning_rate": 0.0002615460479301714, + "loss": 0.3809, + "step": 401 + }, + { + "epoch": 0.8392484342379958, + "grad_norm": 0.04240307801356352, + "learning_rate": 0.0002609021789169571, + "loss": 0.3574, + "step": 402 + }, + { + "epoch": 0.8413361169102297, + "grad_norm": 0.03998645664485665, + "learning_rate": 0.0002602576134364369, + "loss": 0.3008, + "step": 403 + }, + { + "epoch": 0.8434237995824635, + "grad_norm": 0.045774083398434376, + "learning_rate": 0.0002596123588597555, + "loss": 0.4043, + "step": 404 + }, + { + "epoch": 0.8455114822546973, + "grad_norm": 0.043894774464113946, + "learning_rate": 0.0002589664225659382, + "loss": 0.3301, + "step": 405 + }, + { + "epoch": 0.8475991649269311, + "grad_norm": 0.04133334314623751, + "learning_rate": 0.0002583198119418063, + "loss": 0.2637, + "step": 406 + }, + { + "epoch": 0.8496868475991649, + "grad_norm": 0.037286338050560974, + "learning_rate": 0.00025767253438189255, + "loss": 0.3164, + "step": 407 + }, + { + "epoch": 0.8517745302713987, + "grad_norm": 0.04288387043023777, + "learning_rate": 0.0002570245972883568, + "loss": 0.3164, + "step": 408 + }, + { + "epoch": 0.8538622129436325, + "grad_norm": 0.04117605927774004, + "learning_rate": 0.000256376008070901, + "loss": 0.3457, + "step": 409 + }, + { + "epoch": 0.8559498956158664, + "grad_norm": 0.040660730568059064, + "learning_rate": 0.00025572677414668497, + "loss": 0.3691, + "step": 410 + }, + { + "epoch": 0.8580375782881002, + "grad_norm": 0.03473424285365033, + "learning_rate": 0.00025507690294024105, + "loss": 0.252, + "step": 411 + }, + { + "epoch": 0.860125260960334, + "grad_norm": 0.04106073793621654, + "learning_rate": 0.00025442640188338945, + "loss": 0.2871, + "step": 412 + }, + { + "epoch": 0.8622129436325678, + "grad_norm": 0.03780921487044816, + "learning_rate": 0.00025377527841515343, + "loss": 0.2969, + "step": 413 + }, + { + "epoch": 0.8643006263048016, + "grad_norm": 0.03779696762506096, + "learning_rate": 0.00025312353998167397, + "loss": 0.291, + "step": 414 + }, + { + "epoch": 0.8663883089770354, + "grad_norm": 0.04781732258675552, + "learning_rate": 0.00025247119403612456, + "loss": 0.3301, + "step": 415 + }, + { + "epoch": 0.8684759916492694, + "grad_norm": 0.03912752520944136, + "learning_rate": 0.0002518182480386261, + "loss": 0.3477, + "step": 416 + }, + { + "epoch": 0.8705636743215032, + "grad_norm": 0.043540503401128916, + "learning_rate": 0.0002511647094561619, + "loss": 0.3398, + "step": 417 + }, + { + "epoch": 0.872651356993737, + "grad_norm": 0.04207586943763798, + "learning_rate": 0.00025051058576249153, + "loss": 0.3398, + "step": 418 + }, + { + "epoch": 0.8747390396659708, + "grad_norm": 0.03596881591749821, + "learning_rate": 0.00024985588443806624, + "loss": 0.2715, + "step": 419 + }, + { + "epoch": 0.8768267223382046, + "grad_norm": 0.042287007835322306, + "learning_rate": 0.00024920061296994253, + "loss": 0.3086, + "step": 420 + }, + { + "epoch": 0.8789144050104384, + "grad_norm": 0.03856865323635185, + "learning_rate": 0.00024854477885169734, + "loss": 0.2871, + "step": 421 + }, + { + "epoch": 0.8810020876826722, + "grad_norm": 0.03879769693027103, + "learning_rate": 0.0002478883895833417, + "loss": 0.2617, + "step": 422 + }, + { + "epoch": 0.8830897703549061, + "grad_norm": 0.04500551833152804, + "learning_rate": 0.00024723145267123556, + "loss": 0.2793, + "step": 423 + }, + { + "epoch": 0.8851774530271399, + "grad_norm": 0.03925914590146329, + "learning_rate": 0.0002465739756280013, + "loss": 0.3086, + "step": 424 + }, + { + "epoch": 0.8872651356993737, + "grad_norm": 0.04390031842385643, + "learning_rate": 0.0002459159659724383, + "loss": 0.3594, + "step": 425 + }, + { + "epoch": 0.8893528183716075, + "grad_norm": 0.04015458555102483, + "learning_rate": 0.00024525743122943684, + "loss": 0.2773, + "step": 426 + }, + { + "epoch": 0.8914405010438413, + "grad_norm": 0.043251502572368115, + "learning_rate": 0.000244598378929892, + "loss": 0.3633, + "step": 427 + }, + { + "epoch": 0.8935281837160751, + "grad_norm": 0.041718720849523466, + "learning_rate": 0.00024393881661061747, + "loss": 0.3418, + "step": 428 + }, + { + "epoch": 0.8956158663883089, + "grad_norm": 0.03667112112306563, + "learning_rate": 0.0002432787518142596, + "loss": 0.2754, + "step": 429 + }, + { + "epoch": 0.8977035490605428, + "grad_norm": 0.04034875824232465, + "learning_rate": 0.0002426181920892108, + "loss": 0.3242, + "step": 430 + }, + { + "epoch": 0.8997912317327766, + "grad_norm": 0.0416779068349274, + "learning_rate": 0.0002419571449895236, + "loss": 0.2793, + "step": 431 + }, + { + "epoch": 0.9018789144050104, + "grad_norm": 0.04492298155955886, + "learning_rate": 0.00024129561807482378, + "loss": 0.373, + "step": 432 + }, + { + "epoch": 0.9039665970772442, + "grad_norm": 0.0358568800030184, + "learning_rate": 0.00024063361891022455, + "loss": 0.293, + "step": 433 + }, + { + "epoch": 0.906054279749478, + "grad_norm": 0.041240679259523606, + "learning_rate": 0.00023997115506623948, + "loss": 0.3477, + "step": 434 + }, + { + "epoch": 0.9081419624217119, + "grad_norm": 0.04776747282663658, + "learning_rate": 0.00023930823411869617, + "loss": 0.3633, + "step": 435 + }, + { + "epoch": 0.9102296450939458, + "grad_norm": 0.03739632540656231, + "learning_rate": 0.00023864486364864963, + "loss": 0.3027, + "step": 436 + }, + { + "epoch": 0.9123173277661796, + "grad_norm": 0.04013506353404083, + "learning_rate": 0.00023798105124229536, + "loss": 0.2793, + "step": 437 + }, + { + "epoch": 0.9144050104384134, + "grad_norm": 0.044295244931196354, + "learning_rate": 0.00023731680449088312, + "loss": 0.377, + "step": 438 + }, + { + "epoch": 0.9164926931106472, + "grad_norm": 0.03805031241226096, + "learning_rate": 0.00023665213099062938, + "loss": 0.2559, + "step": 439 + }, + { + "epoch": 0.918580375782881, + "grad_norm": 0.03767845062036613, + "learning_rate": 0.00023598703834263113, + "loss": 0.3125, + "step": 440 + }, + { + "epoch": 0.9206680584551148, + "grad_norm": 0.04968215089246422, + "learning_rate": 0.0002353215341527785, + "loss": 0.3223, + "step": 441 + }, + { + "epoch": 0.9227557411273486, + "grad_norm": 0.036246340480605606, + "learning_rate": 0.00023465562603166808, + "loss": 0.2734, + "step": 442 + }, + { + "epoch": 0.9248434237995825, + "grad_norm": 0.034358684978586476, + "learning_rate": 0.00023398932159451557, + "loss": 0.3457, + "step": 443 + }, + { + "epoch": 0.9269311064718163, + "grad_norm": 0.03631746144581004, + "learning_rate": 0.000233322628461069, + "loss": 0.3672, + "step": 444 + }, + { + "epoch": 0.9290187891440501, + "grad_norm": 0.037623854588027666, + "learning_rate": 0.0002326555542555215, + "loss": 0.3047, + "step": 445 + }, + { + "epoch": 0.9311064718162839, + "grad_norm": 0.037886888034629966, + "learning_rate": 0.0002319881066064239, + "loss": 0.2812, + "step": 446 + }, + { + "epoch": 0.9331941544885177, + "grad_norm": 0.039683782098775454, + "learning_rate": 0.00023132029314659795, + "loss": 0.2969, + "step": 447 + }, + { + "epoch": 0.9352818371607515, + "grad_norm": 0.032859528564560474, + "learning_rate": 0.00023065212151304843, + "loss": 0.2617, + "step": 448 + }, + { + "epoch": 0.9373695198329853, + "grad_norm": 0.032092006779359715, + "learning_rate": 0.00022998359934687645, + "loss": 0.2656, + "step": 449 + }, + { + "epoch": 0.9394572025052192, + "grad_norm": 0.03973715339510534, + "learning_rate": 0.00022931473429319143, + "loss": 0.3242, + "step": 450 + }, + { + "epoch": 0.941544885177453, + "grad_norm": 0.040296933695116964, + "learning_rate": 0.00022864553400102437, + "loss": 0.3281, + "step": 451 + }, + { + "epoch": 0.9436325678496869, + "grad_norm": 0.03972360152628702, + "learning_rate": 0.00022797600612323965, + "loss": 0.3301, + "step": 452 + }, + { + "epoch": 0.9457202505219207, + "grad_norm": 0.03196535128099967, + "learning_rate": 0.00022730615831644807, + "loss": 0.2617, + "step": 453 + }, + { + "epoch": 0.9478079331941545, + "grad_norm": 0.041518745854850714, + "learning_rate": 0.00022663599824091892, + "loss": 0.3535, + "step": 454 + }, + { + "epoch": 0.9498956158663883, + "grad_norm": 0.03987544052335102, + "learning_rate": 0.00022596553356049283, + "loss": 0.3184, + "step": 455 + }, + { + "epoch": 0.9519832985386222, + "grad_norm": 0.03790469940548919, + "learning_rate": 0.0002252947719424935, + "loss": 0.2988, + "step": 456 + }, + { + "epoch": 0.954070981210856, + "grad_norm": 0.03494334490842234, + "learning_rate": 0.00022462372105764065, + "loss": 0.2793, + "step": 457 + }, + { + "epoch": 0.9561586638830898, + "grad_norm": 0.03463196563976093, + "learning_rate": 0.0002239523885799618, + "loss": 0.2637, + "step": 458 + }, + { + "epoch": 0.9582463465553236, + "grad_norm": 0.03593899157528181, + "learning_rate": 0.0002232807821867048, + "loss": 0.3184, + "step": 459 + }, + { + "epoch": 0.9603340292275574, + "grad_norm": 0.03890353431042402, + "learning_rate": 0.00022260890955825001, + "loss": 0.3008, + "step": 460 + }, + { + "epoch": 0.9624217118997912, + "grad_norm": 0.033343558090590815, + "learning_rate": 0.00022193677837802235, + "loss": 0.2891, + "step": 461 + }, + { + "epoch": 0.964509394572025, + "grad_norm": 0.037945187747727475, + "learning_rate": 0.0002212643963324035, + "loss": 0.3574, + "step": 462 + }, + { + "epoch": 0.9665970772442589, + "grad_norm": 0.03770022520852902, + "learning_rate": 0.00022059177111064391, + "loss": 0.3262, + "step": 463 + }, + { + "epoch": 0.9686847599164927, + "grad_norm": 0.03875032196468757, + "learning_rate": 0.00021991891040477516, + "loss": 0.3145, + "step": 464 + }, + { + "epoch": 0.9707724425887265, + "grad_norm": 0.03788954429707523, + "learning_rate": 0.00021924582190952153, + "loss": 0.3008, + "step": 465 + }, + { + "epoch": 0.9728601252609603, + "grad_norm": 0.035023715952115834, + "learning_rate": 0.00021857251332221256, + "loss": 0.2412, + "step": 466 + }, + { + "epoch": 0.9749478079331941, + "grad_norm": 0.037210987286251065, + "learning_rate": 0.0002178989923426944, + "loss": 0.3398, + "step": 467 + }, + { + "epoch": 0.9770354906054279, + "grad_norm": 0.03207771649660305, + "learning_rate": 0.0002172252666732424, + "loss": 0.2578, + "step": 468 + }, + { + "epoch": 0.9791231732776617, + "grad_norm": 0.031951869684034837, + "learning_rate": 0.00021655134401847245, + "loss": 0.2695, + "step": 469 + }, + { + "epoch": 0.9812108559498957, + "grad_norm": 0.03674391020210522, + "learning_rate": 0.00021587723208525334, + "loss": 0.3066, + "step": 470 + }, + { + "epoch": 0.9832985386221295, + "grad_norm": 0.035932078354231324, + "learning_rate": 0.00021520293858261823, + "loss": 0.3086, + "step": 471 + }, + { + "epoch": 0.9853862212943633, + "grad_norm": 0.03876657300516048, + "learning_rate": 0.00021452847122167677, + "loss": 0.3008, + "step": 472 + }, + { + "epoch": 0.9874739039665971, + "grad_norm": 0.032766584147543416, + "learning_rate": 0.00021385383771552684, + "loss": 0.3047, + "step": 473 + }, + { + "epoch": 0.9895615866388309, + "grad_norm": 0.04086665364700512, + "learning_rate": 0.00021317904577916635, + "loss": 0.3223, + "step": 474 + }, + { + "epoch": 0.9916492693110647, + "grad_norm": 0.043130532139654114, + "learning_rate": 0.0002125041031294049, + "loss": 0.3242, + "step": 475 + }, + { + "epoch": 0.9937369519832986, + "grad_norm": 0.04273564060035245, + "learning_rate": 0.00021182901748477564, + "loss": 0.4023, + "step": 476 + }, + { + "epoch": 0.9958246346555324, + "grad_norm": 0.03628393896960175, + "learning_rate": 0.00021115379656544715, + "loss": 0.2969, + "step": 477 + }, + { + "epoch": 0.9979123173277662, + "grad_norm": 0.03800046175114251, + "learning_rate": 0.00021047844809313467, + "loss": 0.332, + "step": 478 + }, + { + "epoch": 1.0, + "grad_norm": 0.038197115177252855, + "learning_rate": 0.00020980297979101252, + "loss": 0.2695, + "step": 479 + }, + { + "epoch": 1.0020876826722338, + "grad_norm": 0.032423776045183476, + "learning_rate": 0.000209127399383625, + "loss": 0.2021, + "step": 480 + }, + { + "epoch": 1.0041753653444676, + "grad_norm": 0.03224992988614844, + "learning_rate": 0.00020845171459679866, + "loss": 0.2139, + "step": 481 + }, + { + "epoch": 1.0062630480167014, + "grad_norm": 0.03676757603527534, + "learning_rate": 0.00020777593315755358, + "loss": 0.249, + "step": 482 + }, + { + "epoch": 1.0083507306889352, + "grad_norm": 0.03889498255916948, + "learning_rate": 0.00020710006279401531, + "loss": 0.3047, + "step": 483 + }, + { + "epoch": 1.010438413361169, + "grad_norm": 0.03394341377626214, + "learning_rate": 0.00020642411123532605, + "loss": 0.2598, + "step": 484 + }, + { + "epoch": 1.0125260960334028, + "grad_norm": 0.03195140878316827, + "learning_rate": 0.0002057480862115569, + "loss": 0.2266, + "step": 485 + }, + { + "epoch": 1.0146137787056368, + "grad_norm": 0.03236779419185144, + "learning_rate": 0.0002050719954536187, + "loss": 0.2373, + "step": 486 + }, + { + "epoch": 1.0167014613778707, + "grad_norm": 0.033762294027972366, + "learning_rate": 0.0002043958466931744, + "loss": 0.2324, + "step": 487 + }, + { + "epoch": 1.0187891440501045, + "grad_norm": 0.03225364955535692, + "learning_rate": 0.00020371964766254988, + "loss": 0.2246, + "step": 488 + }, + { + "epoch": 1.0208768267223383, + "grad_norm": 0.03549374570066917, + "learning_rate": 0.00020304340609464627, + "loss": 0.2559, + "step": 489 + }, + { + "epoch": 1.022964509394572, + "grad_norm": 0.039376031742883684, + "learning_rate": 0.00020236712972285088, + "loss": 0.291, + "step": 490 + }, + { + "epoch": 1.0250521920668059, + "grad_norm": 0.033112169954733074, + "learning_rate": 0.0002016908262809491, + "loss": 0.2188, + "step": 491 + }, + { + "epoch": 1.0271398747390397, + "grad_norm": 0.03556499196371877, + "learning_rate": 0.00020101450350303596, + "loss": 0.2363, + "step": 492 + }, + { + "epoch": 1.0292275574112735, + "grad_norm": 0.03747952002586393, + "learning_rate": 0.0002003381691234275, + "loss": 0.2734, + "step": 493 + }, + { + "epoch": 1.0313152400835073, + "grad_norm": 0.039212245318626486, + "learning_rate": 0.00019966183087657255, + "loss": 0.2812, + "step": 494 + }, + { + "epoch": 1.033402922755741, + "grad_norm": 0.038556823852449205, + "learning_rate": 0.00019898549649696409, + "loss": 0.252, + "step": 495 + }, + { + "epoch": 1.0354906054279749, + "grad_norm": 0.03678903694409147, + "learning_rate": 0.00019830917371905095, + "loss": 0.2227, + "step": 496 + }, + { + "epoch": 1.0375782881002087, + "grad_norm": 0.035536317646152164, + "learning_rate": 0.00019763287027714917, + "loss": 0.2451, + "step": 497 + }, + { + "epoch": 1.0396659707724425, + "grad_norm": 0.03669268286478292, + "learning_rate": 0.00019695659390535377, + "loss": 0.2158, + "step": 498 + }, + { + "epoch": 1.0417536534446765, + "grad_norm": 0.0365044069944666, + "learning_rate": 0.0001962803523374501, + "loss": 0.2334, + "step": 499 + }, + { + "epoch": 1.0438413361169103, + "grad_norm": 0.04311308542212124, + "learning_rate": 0.00019560415330682567, + "loss": 0.2754, + "step": 500 + }, + { + "epoch": 1.0459290187891441, + "grad_norm": 0.03807877662781654, + "learning_rate": 0.00019492800454638128, + "loss": 0.2207, + "step": 501 + }, + { + "epoch": 1.048016701461378, + "grad_norm": 0.037347719889777004, + "learning_rate": 0.00019425191378844315, + "loss": 0.2598, + "step": 502 + }, + { + "epoch": 1.0501043841336117, + "grad_norm": 0.041720879231187474, + "learning_rate": 0.00019357588876467403, + "loss": 0.2832, + "step": 503 + }, + { + "epoch": 1.0521920668058455, + "grad_norm": 0.03611878979996833, + "learning_rate": 0.0001928999372059848, + "loss": 0.2305, + "step": 504 + }, + { + "epoch": 1.0542797494780793, + "grad_norm": 0.03914121207636423, + "learning_rate": 0.00019222406684244652, + "loss": 0.2715, + "step": 505 + }, + { + "epoch": 1.0563674321503131, + "grad_norm": 0.045008260139354196, + "learning_rate": 0.00019154828540320141, + "loss": 0.3496, + "step": 506 + }, + { + "epoch": 1.058455114822547, + "grad_norm": 0.06925075294932237, + "learning_rate": 0.00019087260061637507, + "loss": 0.21, + "step": 507 + }, + { + "epoch": 1.0605427974947808, + "grad_norm": 0.042949891828889544, + "learning_rate": 0.00019019702020898753, + "loss": 0.3164, + "step": 508 + }, + { + "epoch": 1.0626304801670146, + "grad_norm": 0.03336924307296428, + "learning_rate": 0.00018952155190686535, + "loss": 0.2334, + "step": 509 + }, + { + "epoch": 1.0647181628392484, + "grad_norm": 0.037171672304770545, + "learning_rate": 0.00018884620343455292, + "loss": 0.2432, + "step": 510 + }, + { + "epoch": 1.0668058455114822, + "grad_norm": 0.03845099190077751, + "learning_rate": 0.0001881709825152244, + "loss": 0.2793, + "step": 511 + }, + { + "epoch": 1.068893528183716, + "grad_norm": 0.03411076783878538, + "learning_rate": 0.00018749589687059513, + "loss": 0.2041, + "step": 512 + }, + { + "epoch": 1.07098121085595, + "grad_norm": 0.03701566955647055, + "learning_rate": 0.00018682095422083373, + "loss": 0.2305, + "step": 513 + }, + { + "epoch": 1.0730688935281838, + "grad_norm": 0.03954320852756579, + "learning_rate": 0.00018614616228447318, + "loss": 0.2354, + "step": 514 + }, + { + "epoch": 1.0751565762004176, + "grad_norm": 0.04313297344221612, + "learning_rate": 0.00018547152877832328, + "loss": 0.2656, + "step": 515 + }, + { + "epoch": 1.0772442588726514, + "grad_norm": 0.03799814412465588, + "learning_rate": 0.0001847970614173818, + "loss": 0.2285, + "step": 516 + }, + { + "epoch": 1.0793319415448852, + "grad_norm": 0.04096742193043013, + "learning_rate": 0.0001841227679147467, + "loss": 0.25, + "step": 517 + }, + { + "epoch": 1.081419624217119, + "grad_norm": 0.03795069745138144, + "learning_rate": 0.00018344865598152754, + "loss": 0.2471, + "step": 518 + }, + { + "epoch": 1.0835073068893528, + "grad_norm": 0.04325821353197005, + "learning_rate": 0.00018277473332675758, + "loss": 0.2754, + "step": 519 + }, + { + "epoch": 1.0855949895615866, + "grad_norm": 0.04070425486872679, + "learning_rate": 0.0001821010076573056, + "loss": 0.2617, + "step": 520 + }, + { + "epoch": 1.0876826722338204, + "grad_norm": 0.031587246842812465, + "learning_rate": 0.00018142748667778746, + "loss": 0.1777, + "step": 521 + }, + { + "epoch": 1.0897703549060542, + "grad_norm": 0.036539737613616244, + "learning_rate": 0.00018075417809047857, + "loss": 0.2773, + "step": 522 + }, + { + "epoch": 1.091858037578288, + "grad_norm": 0.041178092025696736, + "learning_rate": 0.00018008108959522497, + "loss": 0.2676, + "step": 523 + }, + { + "epoch": 1.0939457202505218, + "grad_norm": 0.04046942553591543, + "learning_rate": 0.00017940822888935616, + "loss": 0.2217, + "step": 524 + }, + { + "epoch": 1.0960334029227556, + "grad_norm": 0.03449629028853095, + "learning_rate": 0.0001787356036675966, + "loss": 0.2061, + "step": 525 + }, + { + "epoch": 1.0981210855949897, + "grad_norm": 0.03986588393895996, + "learning_rate": 0.0001780632216219777, + "loss": 0.3125, + "step": 526 + }, + { + "epoch": 1.1002087682672235, + "grad_norm": 0.0333875432706545, + "learning_rate": 0.00017739109044175003, + "loss": 0.1787, + "step": 527 + }, + { + "epoch": 1.1022964509394573, + "grad_norm": 0.04046361146111627, + "learning_rate": 0.00017671921781329522, + "loss": 0.2715, + "step": 528 + }, + { + "epoch": 1.104384133611691, + "grad_norm": 0.04262435868048576, + "learning_rate": 0.00017604761142003827, + "loss": 0.2559, + "step": 529 + }, + { + "epoch": 1.1064718162839249, + "grad_norm": 0.03428115599372914, + "learning_rate": 0.0001753762789423594, + "loss": 0.2061, + "step": 530 + }, + { + "epoch": 1.1085594989561587, + "grad_norm": 0.03868138089293832, + "learning_rate": 0.00017470522805750654, + "loss": 0.2236, + "step": 531 + }, + { + "epoch": 1.1106471816283925, + "grad_norm": 0.03475622072699226, + "learning_rate": 0.00017403446643950724, + "loss": 0.248, + "step": 532 + }, + { + "epoch": 1.1127348643006263, + "grad_norm": 0.03979226115172334, + "learning_rate": 0.00017336400175908112, + "loss": 0.2988, + "step": 533 + }, + { + "epoch": 1.11482254697286, + "grad_norm": 0.036336277616070035, + "learning_rate": 0.000172693841683552, + "loss": 0.2236, + "step": 534 + }, + { + "epoch": 1.116910229645094, + "grad_norm": 0.039263581746387936, + "learning_rate": 0.0001720239938767604, + "loss": 0.332, + "step": 535 + }, + { + "epoch": 1.1189979123173277, + "grad_norm": 0.04395872414762282, + "learning_rate": 0.00017135446599897565, + "loss": 0.2734, + "step": 536 + }, + { + "epoch": 1.1210855949895615, + "grad_norm": 0.03867598581621191, + "learning_rate": 0.00017068526570680853, + "loss": 0.249, + "step": 537 + }, + { + "epoch": 1.1231732776617953, + "grad_norm": 0.035662617079617494, + "learning_rate": 0.00017001640065312357, + "loss": 0.2041, + "step": 538 + }, + { + "epoch": 1.1252609603340291, + "grad_norm": 0.04078186211455469, + "learning_rate": 0.00016934787848695156, + "loss": 0.2119, + "step": 539 + }, + { + "epoch": 1.1273486430062631, + "grad_norm": 0.03947326929596921, + "learning_rate": 0.00016867970685340215, + "loss": 0.2393, + "step": 540 + }, + { + "epoch": 1.129436325678497, + "grad_norm": 0.04449645249301931, + "learning_rate": 0.00016801189339357616, + "loss": 0.3203, + "step": 541 + }, + { + "epoch": 1.1315240083507307, + "grad_norm": 0.046340346999829446, + "learning_rate": 0.00016734444574447862, + "loss": 0.2598, + "step": 542 + }, + { + "epoch": 1.1336116910229646, + "grad_norm": 0.03957931201804222, + "learning_rate": 0.00016667737153893105, + "loss": 0.2402, + "step": 543 + }, + { + "epoch": 1.1356993736951984, + "grad_norm": 0.03703757920078097, + "learning_rate": 0.00016601067840548453, + "loss": 0.1924, + "step": 544 + }, + { + "epoch": 1.1377870563674322, + "grad_norm": 0.03682994100461504, + "learning_rate": 0.000165344373968332, + "loss": 0.2383, + "step": 545 + }, + { + "epoch": 1.139874739039666, + "grad_norm": 0.03920731523127231, + "learning_rate": 0.00016467846584722153, + "loss": 0.2734, + "step": 546 + }, + { + "epoch": 1.1419624217118998, + "grad_norm": 0.04276682539134295, + "learning_rate": 0.0001640129616573689, + "loss": 0.2852, + "step": 547 + }, + { + "epoch": 1.1440501043841336, + "grad_norm": 0.03614032190490994, + "learning_rate": 0.00016334786900937067, + "loss": 0.2354, + "step": 548 + }, + { + "epoch": 1.1461377870563674, + "grad_norm": 0.041006439179447914, + "learning_rate": 0.00016268319550911695, + "loss": 0.2432, + "step": 549 + }, + { + "epoch": 1.1482254697286012, + "grad_norm": 0.044517563506872825, + "learning_rate": 0.00016201894875770466, + "loss": 0.2217, + "step": 550 + }, + { + "epoch": 1.150313152400835, + "grad_norm": 0.03655321665381233, + "learning_rate": 0.00016135513635135045, + "loss": 0.2031, + "step": 551 + }, + { + "epoch": 1.152400835073069, + "grad_norm": 0.04079189189547784, + "learning_rate": 0.00016069176588130388, + "loss": 0.3027, + "step": 552 + }, + { + "epoch": 1.1544885177453028, + "grad_norm": 0.03485321031694839, + "learning_rate": 0.00016002884493376054, + "loss": 0.2275, + "step": 553 + }, + { + "epoch": 1.1565762004175366, + "grad_norm": 0.03052141642274577, + "learning_rate": 0.00015936638108977546, + "loss": 0.1992, + "step": 554 + }, + { + "epoch": 1.1586638830897704, + "grad_norm": 0.03724204209659217, + "learning_rate": 0.0001587043819251762, + "loss": 0.2207, + "step": 555 + }, + { + "epoch": 1.1607515657620042, + "grad_norm": 0.03701626710699115, + "learning_rate": 0.00015804285501047646, + "loss": 0.2041, + "step": 556 + }, + { + "epoch": 1.162839248434238, + "grad_norm": 0.04171363907307022, + "learning_rate": 0.00015738180791078922, + "loss": 0.2715, + "step": 557 + }, + { + "epoch": 1.1649269311064718, + "grad_norm": 0.0392243815886844, + "learning_rate": 0.00015672124818574042, + "loss": 0.2734, + "step": 558 + }, + { + "epoch": 1.1670146137787056, + "grad_norm": 0.034576244303318175, + "learning_rate": 0.0001560611833893826, + "loss": 0.207, + "step": 559 + }, + { + "epoch": 1.1691022964509394, + "grad_norm": 0.03659195217316462, + "learning_rate": 0.0001554016210701081, + "loss": 0.2178, + "step": 560 + }, + { + "epoch": 1.1711899791231732, + "grad_norm": 0.04227483515507797, + "learning_rate": 0.00015474256877056326, + "loss": 0.3086, + "step": 561 + }, + { + "epoch": 1.173277661795407, + "grad_norm": 0.0359857035728927, + "learning_rate": 0.00015408403402756177, + "loss": 0.2617, + "step": 562 + }, + { + "epoch": 1.1753653444676408, + "grad_norm": 0.03680442854270533, + "learning_rate": 0.0001534260243719988, + "loss": 0.209, + "step": 563 + }, + { + "epoch": 1.1774530271398747, + "grad_norm": 0.03670404417452906, + "learning_rate": 0.0001527685473287645, + "loss": 0.2197, + "step": 564 + }, + { + "epoch": 1.1795407098121085, + "grad_norm": 0.0393433887764882, + "learning_rate": 0.00015211161041665832, + "loss": 0.2324, + "step": 565 + }, + { + "epoch": 1.1816283924843423, + "grad_norm": 0.03642939621217633, + "learning_rate": 0.00015145522114830268, + "loss": 0.2393, + "step": 566 + }, + { + "epoch": 1.1837160751565763, + "grad_norm": 0.03878579783362816, + "learning_rate": 0.00015079938703005752, + "loss": 0.2559, + "step": 567 + }, + { + "epoch": 1.18580375782881, + "grad_norm": 0.040970554736426454, + "learning_rate": 0.0001501441155619338, + "loss": 0.2139, + "step": 568 + }, + { + "epoch": 1.187891440501044, + "grad_norm": 0.036820951922352504, + "learning_rate": 0.0001494894142375085, + "loss": 0.1719, + "step": 569 + }, + { + "epoch": 1.1899791231732777, + "grad_norm": 0.04473872009108077, + "learning_rate": 0.00014883529054383817, + "loss": 0.2734, + "step": 570 + }, + { + "epoch": 1.1920668058455115, + "grad_norm": 0.037460855453091094, + "learning_rate": 0.00014818175196137392, + "loss": 0.2334, + "step": 571 + }, + { + "epoch": 1.1941544885177453, + "grad_norm": 0.03980728407908227, + "learning_rate": 0.00014752880596387551, + "loss": 0.2295, + "step": 572 + }, + { + "epoch": 1.196242171189979, + "grad_norm": 0.03562870229391469, + "learning_rate": 0.00014687646001832608, + "loss": 0.1953, + "step": 573 + }, + { + "epoch": 1.198329853862213, + "grad_norm": 0.0412827193725494, + "learning_rate": 0.00014622472158484654, + "loss": 0.25, + "step": 574 + }, + { + "epoch": 1.2004175365344467, + "grad_norm": 0.03634424947703343, + "learning_rate": 0.00014557359811661054, + "loss": 0.207, + "step": 575 + }, + { + "epoch": 1.2025052192066805, + "grad_norm": 0.036441371511458535, + "learning_rate": 0.000144923097059759, + "loss": 0.249, + "step": 576 + }, + { + "epoch": 1.2045929018789143, + "grad_norm": 0.04200794927737242, + "learning_rate": 0.00014427322585331505, + "loss": 0.3027, + "step": 577 + }, + { + "epoch": 1.2066805845511483, + "grad_norm": 0.03620959513479551, + "learning_rate": 0.00014362399192909908, + "loss": 0.2324, + "step": 578 + }, + { + "epoch": 1.2087682672233822, + "grad_norm": 0.03930165573165628, + "learning_rate": 0.0001429754027116433, + "loss": 0.2812, + "step": 579 + }, + { + "epoch": 1.210855949895616, + "grad_norm": 0.040322537436277076, + "learning_rate": 0.00014232746561810747, + "loss": 0.2773, + "step": 580 + }, + { + "epoch": 1.2129436325678498, + "grad_norm": 0.03781024017014511, + "learning_rate": 0.00014168018805819376, + "loss": 0.2354, + "step": 581 + }, + { + "epoch": 1.2150313152400836, + "grad_norm": 0.038595664424750396, + "learning_rate": 0.00014103357743406182, + "loss": 0.2715, + "step": 582 + }, + { + "epoch": 1.2171189979123174, + "grad_norm": 0.034975311451833, + "learning_rate": 0.00014038764114024456, + "loss": 0.2041, + "step": 583 + }, + { + "epoch": 1.2192066805845512, + "grad_norm": 0.03825635258528123, + "learning_rate": 0.00013974238656356313, + "loss": 0.2344, + "step": 584 + }, + { + "epoch": 1.221294363256785, + "grad_norm": 0.03735073552974533, + "learning_rate": 0.0001390978210830429, + "loss": 0.2559, + "step": 585 + }, + { + "epoch": 1.2233820459290188, + "grad_norm": 0.040138739862162154, + "learning_rate": 0.00013845395206982863, + "loss": 0.2656, + "step": 586 + }, + { + "epoch": 1.2254697286012526, + "grad_norm": 0.04418658015359039, + "learning_rate": 0.00013781078688710063, + "loss": 0.291, + "step": 587 + }, + { + "epoch": 1.2275574112734864, + "grad_norm": 0.04239030702531795, + "learning_rate": 0.00013716833288999005, + "loss": 0.2637, + "step": 588 + }, + { + "epoch": 1.2296450939457202, + "grad_norm": 0.03818928317637622, + "learning_rate": 0.00013652659742549532, + "loss": 0.252, + "step": 589 + }, + { + "epoch": 1.231732776617954, + "grad_norm": 0.033789428188020935, + "learning_rate": 0.00013588558783239762, + "loss": 0.1826, + "step": 590 + }, + { + "epoch": 1.2338204592901878, + "grad_norm": 0.03479799715006895, + "learning_rate": 0.00013524531144117736, + "loss": 0.208, + "step": 591 + }, + { + "epoch": 1.2359081419624216, + "grad_norm": 0.0407139048904091, + "learning_rate": 0.00013460577557393, + "loss": 0.2832, + "step": 592 + }, + { + "epoch": 1.2379958246346556, + "grad_norm": 0.04211809721321409, + "learning_rate": 0.00013396698754428266, + "loss": 0.2402, + "step": 593 + }, + { + "epoch": 1.2400835073068894, + "grad_norm": 0.038313063106902565, + "learning_rate": 0.00013332895465731018, + "loss": 0.2539, + "step": 594 + }, + { + "epoch": 1.2421711899791232, + "grad_norm": 0.034451480337492464, + "learning_rate": 0.00013269168420945178, + "loss": 0.209, + "step": 595 + }, + { + "epoch": 1.244258872651357, + "grad_norm": 0.042570134603052, + "learning_rate": 0.0001320551834884276, + "loss": 0.2734, + "step": 596 + }, + { + "epoch": 1.2463465553235908, + "grad_norm": 0.041331289867428476, + "learning_rate": 0.00013141945977315515, + "loss": 0.3066, + "step": 597 + }, + { + "epoch": 1.2484342379958246, + "grad_norm": 0.03567452834168155, + "learning_rate": 0.00013078452033366635, + "loss": 0.2129, + "step": 598 + }, + { + "epoch": 1.2505219206680585, + "grad_norm": 0.03702357216857016, + "learning_rate": 0.00013015037243102425, + "loss": 0.2471, + "step": 599 + }, + { + "epoch": 1.2526096033402923, + "grad_norm": 0.03423169810999704, + "learning_rate": 0.00012951702331724007, + "loss": 0.2236, + "step": 600 + }, + { + "epoch": 1.254697286012526, + "grad_norm": 0.04865070295457581, + "learning_rate": 0.00012888448023519006, + "loss": 0.3223, + "step": 601 + }, + { + "epoch": 1.2567849686847599, + "grad_norm": 0.031921521655174964, + "learning_rate": 0.000128252750418533, + "loss": 0.1934, + "step": 602 + }, + { + "epoch": 1.2588726513569937, + "grad_norm": 0.03207714489804328, + "learning_rate": 0.0001276218410916272, + "loss": 0.1768, + "step": 603 + }, + { + "epoch": 1.2609603340292277, + "grad_norm": 0.03764702756962339, + "learning_rate": 0.000126991759469448, + "loss": 0.207, + "step": 604 + }, + { + "epoch": 1.2630480167014615, + "grad_norm": 0.03494076286574369, + "learning_rate": 0.0001263625127575052, + "loss": 0.2129, + "step": 605 + }, + { + "epoch": 1.2651356993736953, + "grad_norm": 0.03530986655804941, + "learning_rate": 0.00012573410815176083, + "loss": 0.2617, + "step": 606 + }, + { + "epoch": 1.267223382045929, + "grad_norm": 0.03741233532866471, + "learning_rate": 0.00012510655283854658, + "loss": 0.1914, + "step": 607 + }, + { + "epoch": 1.269311064718163, + "grad_norm": 0.03133581344814849, + "learning_rate": 0.00012447985399448194, + "loss": 0.1631, + "step": 608 + }, + { + "epoch": 1.2713987473903967, + "grad_norm": 0.03819672658693745, + "learning_rate": 0.00012385401878639175, + "loss": 0.2891, + "step": 609 + }, + { + "epoch": 1.2734864300626305, + "grad_norm": 0.04476513417542424, + "learning_rate": 0.00012322905437122468, + "loss": 0.3379, + "step": 610 + }, + { + "epoch": 1.2755741127348643, + "grad_norm": 0.037096901415977844, + "learning_rate": 0.00012260496789597093, + "loss": 0.2637, + "step": 611 + }, + { + "epoch": 1.2776617954070981, + "grad_norm": 0.033560968570005754, + "learning_rate": 0.00012198176649758091, + "loss": 0.1768, + "step": 612 + }, + { + "epoch": 1.279749478079332, + "grad_norm": 0.03915286043749694, + "learning_rate": 0.00012135945730288333, + "loss": 0.2402, + "step": 613 + }, + { + "epoch": 1.2818371607515657, + "grad_norm": 0.04243780465562693, + "learning_rate": 0.00012073804742850385, + "loss": 0.2266, + "step": 614 + }, + { + "epoch": 1.2839248434237995, + "grad_norm": 0.035340613835254, + "learning_rate": 0.00012011754398078369, + "loss": 0.2021, + "step": 615 + }, + { + "epoch": 1.2860125260960333, + "grad_norm": 0.038064762794890517, + "learning_rate": 0.00011949795405569813, + "loss": 0.2373, + "step": 616 + }, + { + "epoch": 1.2881002087682671, + "grad_norm": 0.03189685893047627, + "learning_rate": 0.00011887928473877581, + "loss": 0.2178, + "step": 617 + }, + { + "epoch": 1.290187891440501, + "grad_norm": 0.04119691632833074, + "learning_rate": 0.00011826154310501723, + "loss": 0.2832, + "step": 618 + }, + { + "epoch": 1.2922755741127347, + "grad_norm": 0.035970451962649, + "learning_rate": 0.00011764473621881427, + "loss": 0.2295, + "step": 619 + }, + { + "epoch": 1.2943632567849686, + "grad_norm": 0.03644862411137844, + "learning_rate": 0.000117028871133869, + "loss": 0.293, + "step": 620 + }, + { + "epoch": 1.2964509394572026, + "grad_norm": 0.04232916111512329, + "learning_rate": 0.00011641395489311334, + "loss": 0.2432, + "step": 621 + }, + { + "epoch": 1.2985386221294364, + "grad_norm": 0.03815713518197317, + "learning_rate": 0.00011579999452862834, + "loss": 0.2363, + "step": 622 + }, + { + "epoch": 1.3006263048016702, + "grad_norm": 0.03522231596610688, + "learning_rate": 0.00011518699706156373, + "loss": 0.2275, + "step": 623 + }, + { + "epoch": 1.302713987473904, + "grad_norm": 0.03763846715271956, + "learning_rate": 0.00011457496950205784, + "loss": 0.2139, + "step": 624 + }, + { + "epoch": 1.3048016701461378, + "grad_norm": 0.03937303225722648, + "learning_rate": 0.00011396391884915707, + "loss": 0.2246, + "step": 625 + }, + { + "epoch": 1.3068893528183716, + "grad_norm": 0.03259155185424707, + "learning_rate": 0.00011335385209073645, + "loss": 0.2148, + "step": 626 + }, + { + "epoch": 1.3089770354906054, + "grad_norm": 0.036437191147767146, + "learning_rate": 0.00011274477620341906, + "loss": 0.2168, + "step": 627 + }, + { + "epoch": 1.3110647181628392, + "grad_norm": 0.03676944565811173, + "learning_rate": 0.00011213669815249659, + "loss": 0.21, + "step": 628 + }, + { + "epoch": 1.313152400835073, + "grad_norm": 0.03659844810941453, + "learning_rate": 0.00011152962489184955, + "loss": 0.2139, + "step": 629 + }, + { + "epoch": 1.3152400835073068, + "grad_norm": 0.03855990494361526, + "learning_rate": 0.0001109235633638682, + "loss": 0.2373, + "step": 630 + }, + { + "epoch": 1.3173277661795408, + "grad_norm": 0.034719164079528235, + "learning_rate": 0.00011031852049937237, + "loss": 0.2236, + "step": 631 + }, + { + "epoch": 1.3194154488517746, + "grad_norm": 0.03650236752350518, + "learning_rate": 0.00010971450321753276, + "loss": 0.2197, + "step": 632 + }, + { + "epoch": 1.3215031315240084, + "grad_norm": 0.035674494077417296, + "learning_rate": 0.00010911151842579195, + "loss": 0.25, + "step": 633 + }, + { + "epoch": 1.3235908141962422, + "grad_norm": 0.040631215715819284, + "learning_rate": 0.00010850957301978462, + "loss": 0.2891, + "step": 634 + }, + { + "epoch": 1.325678496868476, + "grad_norm": 0.03124048788618308, + "learning_rate": 0.00010790867388325951, + "loss": 0.208, + "step": 635 + }, + { + "epoch": 1.3277661795407099, + "grad_norm": 0.04271594640915922, + "learning_rate": 0.0001073088278880005, + "loss": 0.3184, + "step": 636 + }, + { + "epoch": 1.3298538622129437, + "grad_norm": 0.03507030432270385, + "learning_rate": 0.0001067100418937477, + "loss": 0.208, + "step": 637 + }, + { + "epoch": 1.3319415448851775, + "grad_norm": 0.04020676296142397, + "learning_rate": 0.00010611232274811929, + "loss": 0.2988, + "step": 638 + }, + { + "epoch": 1.3340292275574113, + "grad_norm": 0.034794700042634165, + "learning_rate": 0.00010551567728653311, + "loss": 0.2061, + "step": 639 + }, + { + "epoch": 1.336116910229645, + "grad_norm": 0.04251059198573156, + "learning_rate": 0.00010492011233212871, + "loss": 0.2578, + "step": 640 + }, + { + "epoch": 1.3382045929018789, + "grad_norm": 0.031789932037755535, + "learning_rate": 0.0001043256346956889, + "loss": 0.2109, + "step": 641 + }, + { + "epoch": 1.3402922755741127, + "grad_norm": 0.039155457576247846, + "learning_rate": 0.00010373225117556224, + "loss": 0.2422, + "step": 642 + }, + { + "epoch": 1.3423799582463465, + "grad_norm": 0.035388665066062634, + "learning_rate": 0.00010313996855758504, + "loss": 0.2002, + "step": 643 + }, + { + "epoch": 1.3444676409185803, + "grad_norm": 0.0385841616358797, + "learning_rate": 0.00010254879361500407, + "loss": 0.2041, + "step": 644 + }, + { + "epoch": 1.346555323590814, + "grad_norm": 0.03971686342813934, + "learning_rate": 0.00010195873310839874, + "loss": 0.2256, + "step": 645 + }, + { + "epoch": 1.348643006263048, + "grad_norm": 0.04152102958645101, + "learning_rate": 0.00010136979378560398, + "loss": 0.2734, + "step": 646 + }, + { + "epoch": 1.350730688935282, + "grad_norm": 0.036651673265288884, + "learning_rate": 0.00010078198238163299, + "loss": 0.2148, + "step": 647 + }, + { + "epoch": 1.3528183716075157, + "grad_norm": 0.032824107867213306, + "learning_rate": 0.00010019530561860051, + "loss": 0.208, + "step": 648 + }, + { + "epoch": 1.3549060542797495, + "grad_norm": 0.03503308809553616, + "learning_rate": 9.96097702056455e-05, + "loss": 0.2041, + "step": 649 + }, + { + "epoch": 1.3569937369519833, + "grad_norm": 0.0404355664516907, + "learning_rate": 9.902538283885465e-05, + "loss": 0.2021, + "step": 650 + }, + { + "epoch": 1.3590814196242171, + "grad_norm": 0.047056225295053385, + "learning_rate": 9.844215020118576e-05, + "loss": 0.3125, + "step": 651 + }, + { + "epoch": 1.361169102296451, + "grad_norm": 0.0328129234031199, + "learning_rate": 9.78600789623916e-05, + "loss": 0.1973, + "step": 652 + }, + { + "epoch": 1.3632567849686847, + "grad_norm": 0.03175285628250321, + "learning_rate": 9.727917577894297e-05, + "loss": 0.1914, + "step": 653 + }, + { + "epoch": 1.3653444676409185, + "grad_norm": 0.04520466671433204, + "learning_rate": 9.669944729395316e-05, + "loss": 0.3125, + "step": 654 + }, + { + "epoch": 1.3674321503131524, + "grad_norm": 0.03901825127204986, + "learning_rate": 9.612090013710195e-05, + "loss": 0.2412, + "step": 655 + }, + { + "epoch": 1.3695198329853862, + "grad_norm": 0.037647523878022705, + "learning_rate": 9.554354092455949e-05, + "loss": 0.2227, + "step": 656 + }, + { + "epoch": 1.3716075156576202, + "grad_norm": 0.03796904304768066, + "learning_rate": 9.496737625891076e-05, + "loss": 0.2402, + "step": 657 + }, + { + "epoch": 1.373695198329854, + "grad_norm": 0.036620990435136655, + "learning_rate": 9.439241272908012e-05, + "loss": 0.21, + "step": 658 + }, + { + "epoch": 1.3757828810020878, + "grad_norm": 0.03753406391892656, + "learning_rate": 9.381865691025613e-05, + "loss": 0.2471, + "step": 659 + }, + { + "epoch": 1.3778705636743216, + "grad_norm": 0.03675743788954959, + "learning_rate": 9.324611536381591e-05, + "loss": 0.2256, + "step": 660 + }, + { + "epoch": 1.3799582463465554, + "grad_norm": 0.04379732725831618, + "learning_rate": 9.267479463725048e-05, + "loss": 0.3184, + "step": 661 + }, + { + "epoch": 1.3820459290187892, + "grad_norm": 0.04108282244231379, + "learning_rate": 9.210470126408966e-05, + "loss": 0.3145, + "step": 662 + }, + { + "epoch": 1.384133611691023, + "grad_norm": 0.033523837380523475, + "learning_rate": 9.15358417638277e-05, + "loss": 0.2021, + "step": 663 + }, + { + "epoch": 1.3862212943632568, + "grad_norm": 0.03998207042720944, + "learning_rate": 9.096822264184825e-05, + "loss": 0.2441, + "step": 664 + }, + { + "epoch": 1.3883089770354906, + "grad_norm": 0.037635714292856164, + "learning_rate": 9.040185038935029e-05, + "loss": 0.1934, + "step": 665 + }, + { + "epoch": 1.3903966597077244, + "grad_norm": 0.04063187756459359, + "learning_rate": 8.983673148327369e-05, + "loss": 0.2812, + "step": 666 + }, + { + "epoch": 1.3924843423799582, + "grad_norm": 0.03568609472936473, + "learning_rate": 8.927287238622555e-05, + "loss": 0.2578, + "step": 667 + }, + { + "epoch": 1.394572025052192, + "grad_norm": 0.044075873922673904, + "learning_rate": 8.871027954640567e-05, + "loss": 0.2949, + "step": 668 + }, + { + "epoch": 1.3966597077244258, + "grad_norm": 0.04018365813830774, + "learning_rate": 8.814895939753331e-05, + "loss": 0.3203, + "step": 669 + }, + { + "epoch": 1.3987473903966596, + "grad_norm": 0.038371138538880856, + "learning_rate": 8.758891835877335e-05, + "loss": 0.2451, + "step": 670 + }, + { + "epoch": 1.4008350730688934, + "grad_norm": 0.03644115907045684, + "learning_rate": 8.703016283466323e-05, + "loss": 0.2109, + "step": 671 + }, + { + "epoch": 1.4029227557411272, + "grad_norm": 0.0355184313530291, + "learning_rate": 8.64726992150391e-05, + "loss": 0.2256, + "step": 672 + }, + { + "epoch": 1.405010438413361, + "grad_norm": 0.032473676946887235, + "learning_rate": 8.59165338749632e-05, + "loss": 0.2266, + "step": 673 + }, + { + "epoch": 1.407098121085595, + "grad_norm": 0.03821259918901156, + "learning_rate": 8.536167317465114e-05, + "loss": 0.2305, + "step": 674 + }, + { + "epoch": 1.4091858037578289, + "grad_norm": 0.035842354722894494, + "learning_rate": 8.480812345939855e-05, + "loss": 0.2363, + "step": 675 + }, + { + "epoch": 1.4112734864300627, + "grad_norm": 0.034936854456220405, + "learning_rate": 8.4255891059509e-05, + "loss": 0.2344, + "step": 676 + }, + { + "epoch": 1.4133611691022965, + "grad_norm": 0.0327536953633252, + "learning_rate": 8.370498229022133e-05, + "loss": 0.2354, + "step": 677 + }, + { + "epoch": 1.4154488517745303, + "grad_norm": 0.03332967518468726, + "learning_rate": 8.315540345163783e-05, + "loss": 0.21, + "step": 678 + }, + { + "epoch": 1.417536534446764, + "grad_norm": 0.039757185578627195, + "learning_rate": 8.26071608286517e-05, + "loss": 0.2451, + "step": 679 + }, + { + "epoch": 1.4196242171189979, + "grad_norm": 0.033351639676043855, + "learning_rate": 8.206026069087538e-05, + "loss": 0.1963, + "step": 680 + }, + { + "epoch": 1.4217118997912317, + "grad_norm": 0.036626791510792185, + "learning_rate": 8.151470929256893e-05, + "loss": 0.2471, + "step": 681 + }, + { + "epoch": 1.4237995824634655, + "grad_norm": 0.03663750757496257, + "learning_rate": 8.097051287256854e-05, + "loss": 0.2559, + "step": 682 + }, + { + "epoch": 1.4258872651356993, + "grad_norm": 0.03533804105379715, + "learning_rate": 8.042767765421499e-05, + "loss": 0.2334, + "step": 683 + }, + { + "epoch": 1.4279749478079333, + "grad_norm": 0.03548267765082115, + "learning_rate": 7.988620984528248e-05, + "loss": 0.1992, + "step": 684 + }, + { + "epoch": 1.4300626304801671, + "grad_norm": 0.03365970412354974, + "learning_rate": 7.934611563790803e-05, + "loss": 0.2129, + "step": 685 + }, + { + "epoch": 1.432150313152401, + "grad_norm": 0.036222173404344804, + "learning_rate": 7.880740120852012e-05, + "loss": 0.2441, + "step": 686 + }, + { + "epoch": 1.4342379958246347, + "grad_norm": 0.03545069244831471, + "learning_rate": 7.827007271776843e-05, + "loss": 0.2246, + "step": 687 + }, + { + "epoch": 1.4363256784968685, + "grad_norm": 0.036141104648819485, + "learning_rate": 7.773413631045314e-05, + "loss": 0.2334, + "step": 688 + }, + { + "epoch": 1.4384133611691023, + "grad_norm": 0.03552223808507066, + "learning_rate": 7.719959811545512e-05, + "loss": 0.207, + "step": 689 + }, + { + "epoch": 1.4405010438413361, + "grad_norm": 0.040531388046511914, + "learning_rate": 7.666646424566508e-05, + "loss": 0.25, + "step": 690 + }, + { + "epoch": 1.44258872651357, + "grad_norm": 0.03642114412895087, + "learning_rate": 7.613474079791432e-05, + "loss": 0.2275, + "step": 691 + }, + { + "epoch": 1.4446764091858038, + "grad_norm": 0.03334766371093395, + "learning_rate": 7.56044338529049e-05, + "loss": 0.2246, + "step": 692 + }, + { + "epoch": 1.4467640918580376, + "grad_norm": 0.038435220776801955, + "learning_rate": 7.50755494751398e-05, + "loss": 0.2871, + "step": 693 + }, + { + "epoch": 1.4488517745302714, + "grad_norm": 0.03639976375126217, + "learning_rate": 7.454809371285381e-05, + "loss": 0.2236, + "step": 694 + }, + { + "epoch": 1.4509394572025052, + "grad_norm": 0.04139491524391803, + "learning_rate": 7.402207259794428e-05, + "loss": 0.2617, + "step": 695 + }, + { + "epoch": 1.453027139874739, + "grad_norm": 0.031147759796875986, + "learning_rate": 7.34974921459023e-05, + "loss": 0.1904, + "step": 696 + }, + { + "epoch": 1.4551148225469728, + "grad_norm": 0.04026433329347746, + "learning_rate": 7.297435835574362e-05, + "loss": 0.2393, + "step": 697 + }, + { + "epoch": 1.4572025052192066, + "grad_norm": 0.03965005806725716, + "learning_rate": 7.24526772099403e-05, + "loss": 0.2715, + "step": 698 + }, + { + "epoch": 1.4592901878914404, + "grad_norm": 0.040893406654203844, + "learning_rate": 7.193245467435206e-05, + "loss": 0.2471, + "step": 699 + }, + { + "epoch": 1.4613778705636742, + "grad_norm": 0.03636456124885245, + "learning_rate": 7.141369669815841e-05, + "loss": 0.2168, + "step": 700 + }, + { + "epoch": 1.4634655532359082, + "grad_norm": 0.03647208734905981, + "learning_rate": 7.089640921379026e-05, + "loss": 0.2178, + "step": 701 + }, + { + "epoch": 1.465553235908142, + "grad_norm": 0.0398294556445215, + "learning_rate": 7.038059813686224e-05, + "loss": 0.2695, + "step": 702 + }, + { + "epoch": 1.4676409185803758, + "grad_norm": 0.03444283718095077, + "learning_rate": 6.986626936610491e-05, + "loss": 0.2109, + "step": 703 + }, + { + "epoch": 1.4697286012526096, + "grad_norm": 0.036117772890174496, + "learning_rate": 6.935342878329774e-05, + "loss": 0.1855, + "step": 704 + }, + { + "epoch": 1.4718162839248434, + "grad_norm": 0.03485285291792577, + "learning_rate": 6.884208225320121e-05, + "loss": 0.1953, + "step": 705 + }, + { + "epoch": 1.4739039665970772, + "grad_norm": 0.03913119058204658, + "learning_rate": 6.833223562349018e-05, + "loss": 0.291, + "step": 706 + }, + { + "epoch": 1.475991649269311, + "grad_norm": 0.03619007457603297, + "learning_rate": 6.782389472468687e-05, + "loss": 0.2256, + "step": 707 + }, + { + "epoch": 1.4780793319415448, + "grad_norm": 0.03954136112257214, + "learning_rate": 6.731706537009437e-05, + "loss": 0.2148, + "step": 708 + }, + { + "epoch": 1.4801670146137786, + "grad_norm": 0.039048253382048426, + "learning_rate": 6.68117533557297e-05, + "loss": 0.2119, + "step": 709 + }, + { + "epoch": 1.4822546972860124, + "grad_norm": 0.03633979845664376, + "learning_rate": 6.630796446025793e-05, + "loss": 0.207, + "step": 710 + }, + { + "epoch": 1.4843423799582465, + "grad_norm": 0.04042749139792089, + "learning_rate": 6.580570444492626e-05, + "loss": 0.2617, + "step": 711 + }, + { + "epoch": 1.4864300626304803, + "grad_norm": 0.04112986991098296, + "learning_rate": 6.530497905349753e-05, + "loss": 0.2578, + "step": 712 + }, + { + "epoch": 1.488517745302714, + "grad_norm": 0.04014939278609065, + "learning_rate": 6.480579401218502e-05, + "loss": 0.2324, + "step": 713 + }, + { + "epoch": 1.4906054279749479, + "grad_norm": 0.03662270065596823, + "learning_rate": 6.430815502958674e-05, + "loss": 0.1689, + "step": 714 + }, + { + "epoch": 1.4926931106471817, + "grad_norm": 0.033001164537146806, + "learning_rate": 6.381206779662039e-05, + "loss": 0.1826, + "step": 715 + }, + { + "epoch": 1.4947807933194155, + "grad_norm": 0.04093484928652971, + "learning_rate": 6.331753798645796e-05, + "loss": 0.249, + "step": 716 + }, + { + "epoch": 1.4968684759916493, + "grad_norm": 0.03400770412940897, + "learning_rate": 6.282457125446109e-05, + "loss": 0.1709, + "step": 717 + }, + { + "epoch": 1.498956158663883, + "grad_norm": 0.03449610130369155, + "learning_rate": 6.233317323811615e-05, + "loss": 0.2168, + "step": 718 + }, + { + "epoch": 1.501043841336117, + "grad_norm": 0.03943007646918304, + "learning_rate": 6.184334955697028e-05, + "loss": 0.2617, + "step": 719 + }, + { + "epoch": 1.5031315240083507, + "grad_norm": 0.04046270086854637, + "learning_rate": 6.135510581256647e-05, + "loss": 0.2715, + "step": 720 + }, + { + "epoch": 1.5052192066805845, + "grad_norm": 0.036042076960770504, + "learning_rate": 6.086844758837991e-05, + "loss": 0.2275, + "step": 721 + }, + { + "epoch": 1.5073068893528183, + "grad_norm": 0.038478916176100005, + "learning_rate": 6.0383380449754004e-05, + "loss": 0.1924, + "step": 722 + }, + { + "epoch": 1.5093945720250521, + "grad_norm": 0.0578411545976244, + "learning_rate": 5.98999099438369e-05, + "loss": 0.2852, + "step": 723 + }, + { + "epoch": 1.511482254697286, + "grad_norm": 0.03739145944754586, + "learning_rate": 5.941804159951778e-05, + "loss": 0.2197, + "step": 724 + }, + { + "epoch": 1.5135699373695197, + "grad_norm": 0.036832588153257496, + "learning_rate": 5.893778092736382e-05, + "loss": 0.2422, + "step": 725 + }, + { + "epoch": 1.5156576200417535, + "grad_norm": 0.03671021108564163, + "learning_rate": 5.845913341955711e-05, + "loss": 0.1572, + "step": 726 + }, + { + "epoch": 1.5177453027139873, + "grad_norm": 0.03573073453794959, + "learning_rate": 5.798210454983184e-05, + "loss": 0.2168, + "step": 727 + }, + { + "epoch": 1.5198329853862211, + "grad_norm": 0.035959632617295005, + "learning_rate": 5.7506699773411764e-05, + "loss": 0.2451, + "step": 728 + }, + { + "epoch": 1.5219206680584552, + "grad_norm": 0.038728578675763334, + "learning_rate": 5.703292452694771e-05, + "loss": 0.2139, + "step": 729 + }, + { + "epoch": 1.524008350730689, + "grad_norm": 0.035308915371426546, + "learning_rate": 5.6560784228455586e-05, + "loss": 0.2021, + "step": 730 + }, + { + "epoch": 1.5260960334029228, + "grad_norm": 0.039840996287124834, + "learning_rate": 5.609028427725418e-05, + "loss": 0.1953, + "step": 731 + }, + { + "epoch": 1.5281837160751566, + "grad_norm": 0.04087634507351788, + "learning_rate": 5.562143005390361e-05, + "loss": 0.2617, + "step": 732 + }, + { + "epoch": 1.5302713987473904, + "grad_norm": 0.03746689779610756, + "learning_rate": 5.5154226920143626e-05, + "loss": 0.2197, + "step": 733 + }, + { + "epoch": 1.5323590814196242, + "grad_norm": 0.04226091666932121, + "learning_rate": 5.468868021883256e-05, + "loss": 0.2773, + "step": 734 + }, + { + "epoch": 1.534446764091858, + "grad_norm": 0.03255698564490662, + "learning_rate": 5.422479527388591e-05, + "loss": 0.1816, + "step": 735 + }, + { + "epoch": 1.536534446764092, + "grad_norm": 0.03534998758506692, + "learning_rate": 5.376257739021564e-05, + "loss": 0.2158, + "step": 736 + }, + { + "epoch": 1.5386221294363258, + "grad_norm": 0.04286917750609741, + "learning_rate": 5.330203185366942e-05, + "loss": 0.2559, + "step": 737 + }, + { + "epoch": 1.5407098121085596, + "grad_norm": 0.03987124501813775, + "learning_rate": 5.284316393097042e-05, + "loss": 0.2197, + "step": 738 + }, + { + "epoch": 1.5427974947807934, + "grad_norm": 0.0357279862009252, + "learning_rate": 5.2385978869656705e-05, + "loss": 0.209, + "step": 739 + }, + { + "epoch": 1.5448851774530272, + "grad_norm": 0.03917856556644758, + "learning_rate": 5.193048189802143e-05, + "loss": 0.2402, + "step": 740 + }, + { + "epoch": 1.546972860125261, + "grad_norm": 0.05340411374410496, + "learning_rate": 5.147667822505318e-05, + "loss": 0.2695, + "step": 741 + }, + { + "epoch": 1.5490605427974948, + "grad_norm": 0.041333285407966595, + "learning_rate": 5.1024573040376115e-05, + "loss": 0.2969, + "step": 742 + }, + { + "epoch": 1.5511482254697286, + "grad_norm": 0.039402130485741114, + "learning_rate": 5.057417151419077e-05, + "loss": 0.2402, + "step": 743 + }, + { + "epoch": 1.5532359081419624, + "grad_norm": 0.035671385682246794, + "learning_rate": 5.012547879721494e-05, + "loss": 0.2158, + "step": 744 + }, + { + "epoch": 1.5553235908141962, + "grad_norm": 0.038417929606667435, + "learning_rate": 4.967850002062491e-05, + "loss": 0.2236, + "step": 745 + }, + { + "epoch": 1.55741127348643, + "grad_norm": 0.03949905474465816, + "learning_rate": 4.923324029599632e-05, + "loss": 0.2715, + "step": 746 + }, + { + "epoch": 1.5594989561586639, + "grad_norm": 0.0352045032702957, + "learning_rate": 4.878970471524622e-05, + "loss": 0.2275, + "step": 747 + }, + { + "epoch": 1.5615866388308977, + "grad_norm": 0.036303248719977955, + "learning_rate": 4.834789835057465e-05, + "loss": 0.2021, + "step": 748 + }, + { + "epoch": 1.5636743215031315, + "grad_norm": 0.03443800350271572, + "learning_rate": 4.790782625440655e-05, + "loss": 0.1992, + "step": 749 + }, + { + "epoch": 1.5657620041753653, + "grad_norm": 0.04231633067945213, + "learning_rate": 4.7469493459334066e-05, + "loss": 0.2617, + "step": 750 + }, + { + "epoch": 1.567849686847599, + "grad_norm": 0.037830454141616246, + "learning_rate": 4.70329049780589e-05, + "loss": 0.2217, + "step": 751 + }, + { + "epoch": 1.5699373695198329, + "grad_norm": 0.037662635732679015, + "learning_rate": 4.65980658033353e-05, + "loss": 0.2188, + "step": 752 + }, + { + "epoch": 1.5720250521920667, + "grad_norm": 0.04067293838178101, + "learning_rate": 4.616498090791248e-05, + "loss": 0.293, + "step": 753 + }, + { + "epoch": 1.5741127348643005, + "grad_norm": 0.03901581955251283, + "learning_rate": 4.5733655244478146e-05, + "loss": 0.2266, + "step": 754 + }, + { + "epoch": 1.5762004175365343, + "grad_norm": 0.03554948290229258, + "learning_rate": 4.5304093745601605e-05, + "loss": 0.1943, + "step": 755 + }, + { + "epoch": 1.5782881002087683, + "grad_norm": 0.0405280460733433, + "learning_rate": 4.487630132367764e-05, + "loss": 0.2617, + "step": 756 + }, + { + "epoch": 1.5803757828810021, + "grad_norm": 0.03823926366289705, + "learning_rate": 4.445028287087003e-05, + "loss": 0.252, + "step": 757 + }, + { + "epoch": 1.582463465553236, + "grad_norm": 0.03721532941881517, + "learning_rate": 4.402604325905575e-05, + "loss": 0.2129, + "step": 758 + }, + { + "epoch": 1.5845511482254697, + "grad_norm": 0.029782665453401677, + "learning_rate": 4.360358733976919e-05, + "loss": 0.1826, + "step": 759 + }, + { + "epoch": 1.5866388308977035, + "grad_norm": 0.03989589641175334, + "learning_rate": 4.3182919944146936e-05, + "loss": 0.2734, + "step": 760 + }, + { + "epoch": 1.5887265135699373, + "grad_norm": 0.04383942244265336, + "learning_rate": 4.2764045882872086e-05, + "loss": 0.3008, + "step": 761 + }, + { + "epoch": 1.5908141962421714, + "grad_norm": 0.034818289814213685, + "learning_rate": 4.2346969946119574e-05, + "loss": 0.2188, + "step": 762 + }, + { + "epoch": 1.5929018789144052, + "grad_norm": 0.03770186022731002, + "learning_rate": 4.1931696903501204e-05, + "loss": 0.2314, + "step": 763 + }, + { + "epoch": 1.594989561586639, + "grad_norm": 0.03973578816213652, + "learning_rate": 4.151823150401124e-05, + "loss": 0.2402, + "step": 764 + }, + { + "epoch": 1.5970772442588728, + "grad_norm": 0.03236007667945323, + "learning_rate": 4.1106578475972035e-05, + "loss": 0.1924, + "step": 765 + }, + { + "epoch": 1.5991649269311066, + "grad_norm": 0.03764508247239107, + "learning_rate": 4.069674252697988e-05, + "loss": 0.1748, + "step": 766 + }, + { + "epoch": 1.6012526096033404, + "grad_norm": 0.03731281581422325, + "learning_rate": 4.02887283438514e-05, + "loss": 0.2559, + "step": 767 + }, + { + "epoch": 1.6033402922755742, + "grad_norm": 0.0372624393342072, + "learning_rate": 3.988254059256971e-05, + "loss": 0.2188, + "step": 768 + }, + { + "epoch": 1.605427974947808, + "grad_norm": 0.039361210964991795, + "learning_rate": 3.9478183918231106e-05, + "loss": 0.2578, + "step": 769 + }, + { + "epoch": 1.6075156576200418, + "grad_norm": 0.03750381742110697, + "learning_rate": 3.907566294499201e-05, + "loss": 0.1982, + "step": 770 + }, + { + "epoch": 1.6096033402922756, + "grad_norm": 0.03771991482783373, + "learning_rate": 3.8674982276016205e-05, + "loss": 0.2412, + "step": 771 + }, + { + "epoch": 1.6116910229645094, + "grad_norm": 0.04091075279365844, + "learning_rate": 3.82761464934219e-05, + "loss": 0.2793, + "step": 772 + }, + { + "epoch": 1.6137787056367432, + "grad_norm": 0.03859340968976196, + "learning_rate": 3.787916015822954e-05, + "loss": 0.2471, + "step": 773 + }, + { + "epoch": 1.615866388308977, + "grad_norm": 0.03200544736921389, + "learning_rate": 3.748402781030955e-05, + "loss": 0.1826, + "step": 774 + }, + { + "epoch": 1.6179540709812108, + "grad_norm": 0.0340269199931104, + "learning_rate": 3.709075396833057e-05, + "loss": 0.1855, + "step": 775 + }, + { + "epoch": 1.6200417536534446, + "grad_norm": 0.038072978162411425, + "learning_rate": 3.66993431297076e-05, + "loss": 0.2432, + "step": 776 + }, + { + "epoch": 1.6221294363256784, + "grad_norm": 0.04102889623967079, + "learning_rate": 3.63097997705506e-05, + "loss": 0.3145, + "step": 777 + }, + { + "epoch": 1.6242171189979122, + "grad_norm": 0.04007477882751308, + "learning_rate": 3.592212834561337e-05, + "loss": 0.2432, + "step": 778 + }, + { + "epoch": 1.626304801670146, + "grad_norm": 0.03819036990721501, + "learning_rate": 3.553633328824266e-05, + "loss": 0.1973, + "step": 779 + }, + { + "epoch": 1.6283924843423798, + "grad_norm": 0.038350798236554316, + "learning_rate": 3.515241901032731e-05, + "loss": 0.2676, + "step": 780 + }, + { + "epoch": 1.6304801670146136, + "grad_norm": 0.03548951603993466, + "learning_rate": 3.477038990224786e-05, + "loss": 0.2002, + "step": 781 + }, + { + "epoch": 1.6325678496868476, + "grad_norm": 0.034615335672358714, + "learning_rate": 3.439025033282639e-05, + "loss": 0.1963, + "step": 782 + }, + { + "epoch": 1.6346555323590815, + "grad_norm": 0.04345337713073096, + "learning_rate": 3.401200464927654e-05, + "loss": 0.2578, + "step": 783 + }, + { + "epoch": 1.6367432150313153, + "grad_norm": 0.03791537054424221, + "learning_rate": 3.363565717715373e-05, + "loss": 0.2227, + "step": 784 + }, + { + "epoch": 1.638830897703549, + "grad_norm": 0.04113195871715081, + "learning_rate": 3.326121222030578e-05, + "loss": 0.2012, + "step": 785 + }, + { + "epoch": 1.6409185803757829, + "grad_norm": 0.041884239195247946, + "learning_rate": 3.288867406082372e-05, + "loss": 0.2236, + "step": 786 + }, + { + "epoch": 1.6430062630480167, + "grad_norm": 0.03652594444269533, + "learning_rate": 3.251804695899267e-05, + "loss": 0.2139, + "step": 787 + }, + { + "epoch": 1.6450939457202505, + "grad_norm": 0.05643896674418451, + "learning_rate": 3.214933515324323e-05, + "loss": 0.2422, + "step": 788 + }, + { + "epoch": 1.6471816283924845, + "grad_norm": 0.04098991740754866, + "learning_rate": 3.178254286010296e-05, + "loss": 0.2422, + "step": 789 + }, + { + "epoch": 1.6492693110647183, + "grad_norm": 0.04098022159134268, + "learning_rate": 3.1417674274148276e-05, + "loss": 0.2402, + "step": 790 + }, + { + "epoch": 1.651356993736952, + "grad_norm": 0.037087942509587524, + "learning_rate": 3.105473356795634e-05, + "loss": 0.2295, + "step": 791 + }, + { + "epoch": 1.653444676409186, + "grad_norm": 0.03775161487801411, + "learning_rate": 3.0693724892057396e-05, + "loss": 0.207, + "step": 792 + }, + { + "epoch": 1.6555323590814197, + "grad_norm": 0.04001811834655189, + "learning_rate": 3.0334652374887263e-05, + "loss": 0.2773, + "step": 793 + }, + { + "epoch": 1.6576200417536535, + "grad_norm": 0.03971800362165568, + "learning_rate": 2.997752012274031e-05, + "loss": 0.2139, + "step": 794 + }, + { + "epoch": 1.6597077244258873, + "grad_norm": 0.03971183328914648, + "learning_rate": 2.9622332219722193e-05, + "loss": 0.2422, + "step": 795 + }, + { + "epoch": 1.6617954070981211, + "grad_norm": 0.03937647373747184, + "learning_rate": 2.9269092727703375e-05, + "loss": 0.252, + "step": 796 + }, + { + "epoch": 1.663883089770355, + "grad_norm": 0.03693074285981622, + "learning_rate": 2.891780568627267e-05, + "loss": 0.1992, + "step": 797 + }, + { + "epoch": 1.6659707724425887, + "grad_norm": 0.03819446049329271, + "learning_rate": 2.8568475112690918e-05, + "loss": 0.2383, + "step": 798 + }, + { + "epoch": 1.6680584551148225, + "grad_norm": 0.03385681858741079, + "learning_rate": 2.8221105001845095e-05, + "loss": 0.1914, + "step": 799 + }, + { + "epoch": 1.6701461377870563, + "grad_norm": 0.03481284448136041, + "learning_rate": 2.7875699326202663e-05, + "loss": 0.1943, + "step": 800 + }, + { + "epoch": 1.6722338204592901, + "grad_norm": 0.038887635380198286, + "learning_rate": 2.7532262035766288e-05, + "loss": 0.2236, + "step": 801 + }, + { + "epoch": 1.674321503131524, + "grad_norm": 0.039268882753983984, + "learning_rate": 2.7190797058028207e-05, + "loss": 0.2402, + "step": 802 + }, + { + "epoch": 1.6764091858037578, + "grad_norm": 0.03561285649555011, + "learning_rate": 2.685130829792577e-05, + "loss": 0.2041, + "step": 803 + }, + { + "epoch": 1.6784968684759916, + "grad_norm": 0.03548955540319579, + "learning_rate": 2.6513799637796745e-05, + "loss": 0.2188, + "step": 804 + }, + { + "epoch": 1.6805845511482254, + "grad_norm": 0.03987654406934741, + "learning_rate": 2.61782749373346e-05, + "loss": 0.2334, + "step": 805 + }, + { + "epoch": 1.6826722338204592, + "grad_norm": 0.03591235670312746, + "learning_rate": 2.584473803354468e-05, + "loss": 0.2217, + "step": 806 + }, + { + "epoch": 1.684759916492693, + "grad_norm": 0.03563754130765487, + "learning_rate": 2.5513192740700144e-05, + "loss": 0.2061, + "step": 807 + }, + { + "epoch": 1.6868475991649268, + "grad_norm": 0.03453991302292082, + "learning_rate": 2.5183642850298505e-05, + "loss": 0.207, + "step": 808 + }, + { + "epoch": 1.6889352818371608, + "grad_norm": 0.040646656045912735, + "learning_rate": 2.4856092131018137e-05, + "loss": 0.293, + "step": 809 + }, + { + "epoch": 1.6910229645093946, + "grad_norm": 0.03639422717002704, + "learning_rate": 2.4530544328675186e-05, + "loss": 0.2236, + "step": 810 + }, + { + "epoch": 1.6931106471816284, + "grad_norm": 0.04124184507078617, + "learning_rate": 2.4207003166180765e-05, + "loss": 0.2344, + "step": 811 + }, + { + "epoch": 1.6951983298538622, + "grad_norm": 0.04821861121500109, + "learning_rate": 2.3885472343498515e-05, + "loss": 0.2227, + "step": 812 + }, + { + "epoch": 1.697286012526096, + "grad_norm": 0.036261622126831795, + "learning_rate": 2.3565955537602014e-05, + "loss": 0.2266, + "step": 813 + }, + { + "epoch": 1.6993736951983298, + "grad_norm": 0.03833360109114772, + "learning_rate": 2.3248456402432918e-05, + "loss": 0.2598, + "step": 814 + }, + { + "epoch": 1.7014613778705638, + "grad_norm": 0.03811210564749943, + "learning_rate": 2.293297856885912e-05, + "loss": 0.2197, + "step": 815 + }, + { + "epoch": 1.7035490605427976, + "grad_norm": 0.03440881800931742, + "learning_rate": 2.261952564463332e-05, + "loss": 0.1865, + "step": 816 + }, + { + "epoch": 1.7056367432150314, + "grad_norm": 0.03784666030148996, + "learning_rate": 2.2308101214351628e-05, + "loss": 0.2324, + "step": 817 + }, + { + "epoch": 1.7077244258872653, + "grad_norm": 0.03706179096561306, + "learning_rate": 2.1998708839412597e-05, + "loss": 0.21, + "step": 818 + }, + { + "epoch": 1.709812108559499, + "grad_norm": 0.036230331032483534, + "learning_rate": 2.1691352057976565e-05, + "loss": 0.2471, + "step": 819 + }, + { + "epoch": 1.7118997912317329, + "grad_norm": 0.04090973690460557, + "learning_rate": 2.138603438492517e-05, + "loss": 0.2637, + "step": 820 + }, + { + "epoch": 1.7139874739039667, + "grad_norm": 0.051662701812649245, + "learning_rate": 2.108275931182111e-05, + "loss": 0.1953, + "step": 821 + }, + { + "epoch": 1.7160751565762005, + "grad_norm": 0.038240986750146325, + "learning_rate": 2.0781530306868246e-05, + "loss": 0.2393, + "step": 822 + }, + { + "epoch": 1.7181628392484343, + "grad_norm": 0.04106815130974131, + "learning_rate": 2.048235081487202e-05, + "loss": 0.2637, + "step": 823 + }, + { + "epoch": 1.720250521920668, + "grad_norm": 0.03673610429332063, + "learning_rate": 2.0185224257199908e-05, + "loss": 0.2207, + "step": 824 + }, + { + "epoch": 1.7223382045929019, + "grad_norm": 0.04334868227291063, + "learning_rate": 1.9890154031742325e-05, + "loss": 0.3223, + "step": 825 + }, + { + "epoch": 1.7244258872651357, + "grad_norm": 0.04017653068637208, + "learning_rate": 1.9597143512873872e-05, + "loss": 0.2656, + "step": 826 + }, + { + "epoch": 1.7265135699373695, + "grad_norm": 0.04143141277717955, + "learning_rate": 1.9306196051414728e-05, + "loss": 0.252, + "step": 827 + }, + { + "epoch": 1.7286012526096033, + "grad_norm": 0.030689616701832047, + "learning_rate": 1.9017314974592203e-05, + "loss": 0.1982, + "step": 828 + }, + { + "epoch": 1.730688935281837, + "grad_norm": 0.04149364456774439, + "learning_rate": 1.873050358600279e-05, + "loss": 0.2451, + "step": 829 + }, + { + "epoch": 1.732776617954071, + "grad_norm": 0.037822515436121175, + "learning_rate": 1.8445765165574324e-05, + "loss": 0.2246, + "step": 830 + }, + { + "epoch": 1.7348643006263047, + "grad_norm": 0.036300250288018185, + "learning_rate": 1.8163102969528702e-05, + "loss": 0.1709, + "step": 831 + }, + { + "epoch": 1.7369519832985385, + "grad_norm": 0.03813579853432033, + "learning_rate": 1.7882520230344225e-05, + "loss": 0.1982, + "step": 832 + }, + { + "epoch": 1.7390396659707723, + "grad_norm": 0.03490543301641405, + "learning_rate": 1.7604020156719026e-05, + "loss": 0.2256, + "step": 833 + }, + { + "epoch": 1.7411273486430061, + "grad_norm": 0.03767801651422133, + "learning_rate": 1.7327605933534087e-05, + "loss": 0.248, + "step": 834 + }, + { + "epoch": 1.7432150313152401, + "grad_norm": 0.03559965744869205, + "learning_rate": 1.705328072181711e-05, + "loss": 0.1963, + "step": 835 + }, + { + "epoch": 1.745302713987474, + "grad_norm": 0.03620103246900227, + "learning_rate": 1.6781047658706094e-05, + "loss": 0.2363, + "step": 836 + }, + { + "epoch": 1.7473903966597077, + "grad_norm": 0.03363208108698153, + "learning_rate": 1.6510909857413593e-05, + "loss": 0.208, + "step": 837 + }, + { + "epoch": 1.7494780793319415, + "grad_norm": 0.038864402335742426, + "learning_rate": 1.624287040719108e-05, + "loss": 0.2217, + "step": 838 + }, + { + "epoch": 1.7515657620041754, + "grad_norm": 0.03604996210348334, + "learning_rate": 1.597693237329363e-05, + "loss": 0.2363, + "step": 839 + }, + { + "epoch": 1.7536534446764092, + "grad_norm": 0.03870805026122142, + "learning_rate": 1.5713098796944913e-05, + "loss": 0.2363, + "step": 840 + }, + { + "epoch": 1.755741127348643, + "grad_norm": 0.037967280022889786, + "learning_rate": 1.545137269530228e-05, + "loss": 0.2031, + "step": 841 + }, + { + "epoch": 1.757828810020877, + "grad_norm": 0.032100601105533956, + "learning_rate": 1.5191757061422484e-05, + "loss": 0.1963, + "step": 842 + }, + { + "epoch": 1.7599164926931108, + "grad_norm": 0.03677180115256404, + "learning_rate": 1.493425486422717e-05, + "loss": 0.2324, + "step": 843 + }, + { + "epoch": 1.7620041753653446, + "grad_norm": 0.04179333438984321, + "learning_rate": 1.467886904846918e-05, + "loss": 0.2197, + "step": 844 + }, + { + "epoch": 1.7640918580375784, + "grad_norm": 0.03330069993588093, + "learning_rate": 1.4425602534698645e-05, + "loss": 0.1992, + "step": 845 + }, + { + "epoch": 1.7661795407098122, + "grad_norm": 0.03921593097418753, + "learning_rate": 1.4174458219229868e-05, + "loss": 0.2734, + "step": 846 + }, + { + "epoch": 1.768267223382046, + "grad_norm": 0.0384869660070795, + "learning_rate": 1.3925438974107918e-05, + "loss": 0.2285, + "step": 847 + }, + { + "epoch": 1.7703549060542798, + "grad_norm": 0.03235612278723638, + "learning_rate": 1.3678547647075946e-05, + "loss": 0.2021, + "step": 848 + }, + { + "epoch": 1.7724425887265136, + "grad_norm": 0.034557637083813396, + "learning_rate": 1.3433787061542525e-05, + "loss": 0.1982, + "step": 849 + }, + { + "epoch": 1.7745302713987474, + "grad_norm": 0.036691512115311166, + "learning_rate": 1.3191160016549564e-05, + "loss": 0.2695, + "step": 850 + }, + { + "epoch": 1.7766179540709812, + "grad_norm": 0.036890816812044876, + "learning_rate": 1.2950669286740024e-05, + "loss": 0.2119, + "step": 851 + }, + { + "epoch": 1.778705636743215, + "grad_norm": 0.03801656481599902, + "learning_rate": 1.271231762232632e-05, + "loss": 0.2354, + "step": 852 + }, + { + "epoch": 1.7807933194154488, + "grad_norm": 0.0431132757880389, + "learning_rate": 1.2476107749058986e-05, + "loss": 0.2305, + "step": 853 + }, + { + "epoch": 1.7828810020876826, + "grad_norm": 0.034619924337342727, + "learning_rate": 1.2242042368195218e-05, + "loss": 0.1914, + "step": 854 + }, + { + "epoch": 1.7849686847599164, + "grad_norm": 0.034941322602160714, + "learning_rate": 1.2010124156468294e-05, + "loss": 0.2051, + "step": 855 + }, + { + "epoch": 1.7870563674321502, + "grad_norm": 0.03577282594283845, + "learning_rate": 1.1780355766056694e-05, + "loss": 0.2334, + "step": 856 + }, + { + "epoch": 1.789144050104384, + "grad_norm": 0.03923352027261708, + "learning_rate": 1.1552739824554026e-05, + "loss": 0.2217, + "step": 857 + }, + { + "epoch": 1.7912317327766178, + "grad_norm": 0.03464671275803716, + "learning_rate": 1.1327278934938723e-05, + "loss": 0.1992, + "step": 858 + }, + { + "epoch": 1.7933194154488517, + "grad_norm": 0.04157357259303964, + "learning_rate": 1.1103975675544443e-05, + "loss": 0.2461, + "step": 859 + }, + { + "epoch": 1.7954070981210855, + "grad_norm": 0.040227594429237123, + "learning_rate": 1.0882832600030624e-05, + "loss": 0.2441, + "step": 860 + }, + { + "epoch": 1.7974947807933193, + "grad_norm": 0.03488250123821703, + "learning_rate": 1.0663852237353067e-05, + "loss": 0.1963, + "step": 861 + }, + { + "epoch": 1.7995824634655533, + "grad_norm": 0.03979475512791435, + "learning_rate": 1.0447037091735223e-05, + "loss": 0.2471, + "step": 862 + }, + { + "epoch": 1.801670146137787, + "grad_norm": 0.03997223499064563, + "learning_rate": 1.0232389642639395e-05, + "loss": 0.2471, + "step": 863 + }, + { + "epoch": 1.803757828810021, + "grad_norm": 0.04038899286733726, + "learning_rate": 1.0019912344738625e-05, + "loss": 0.2598, + "step": 864 + }, + { + "epoch": 1.8058455114822547, + "grad_norm": 0.03748451896837008, + "learning_rate": 9.809607627888296e-06, + "loss": 0.2002, + "step": 865 + }, + { + "epoch": 1.8079331941544885, + "grad_norm": 0.03427034514456546, + "learning_rate": 9.601477897098576e-06, + "loss": 0.1729, + "step": 866 + }, + { + "epoch": 1.8100208768267223, + "grad_norm": 0.04471055633543089, + "learning_rate": 9.395525532506866e-06, + "loss": 0.2256, + "step": 867 + }, + { + "epoch": 1.812108559498956, + "grad_norm": 0.03475782341698621, + "learning_rate": 9.191752889350547e-06, + "loss": 0.2207, + "step": 868 + }, + { + "epoch": 1.8141962421711901, + "grad_norm": 0.03828798248246293, + "learning_rate": 8.990162297940097e-06, + "loss": 0.2109, + "step": 869 + }, + { + "epoch": 1.816283924843424, + "grad_norm": 0.04180601763110725, + "learning_rate": 8.79075606363231e-06, + "loss": 0.2676, + "step": 870 + }, + { + "epoch": 1.8183716075156577, + "grad_norm": 0.03646735664376503, + "learning_rate": 8.593536466804142e-06, + "loss": 0.207, + "step": 871 + }, + { + "epoch": 1.8204592901878915, + "grad_norm": 0.03218572550634406, + "learning_rate": 8.398505762826503e-06, + "loss": 0.1797, + "step": 872 + }, + { + "epoch": 1.8225469728601253, + "grad_norm": 0.03982709976184709, + "learning_rate": 8.205666182038418e-06, + "loss": 0.2432, + "step": 873 + }, + { + "epoch": 1.8246346555323592, + "grad_norm": 0.037982327489911684, + "learning_rate": 8.015019929721668e-06, + "loss": 0.2344, + "step": 874 + }, + { + "epoch": 1.826722338204593, + "grad_norm": 0.03580893689768604, + "learning_rate": 7.826569186075428e-06, + "loss": 0.252, + "step": 875 + }, + { + "epoch": 1.8288100208768268, + "grad_norm": 0.035854237119299116, + "learning_rate": 7.640316106191403e-06, + "loss": 0.1963, + "step": 876 + }, + { + "epoch": 1.8308977035490606, + "grad_norm": 0.03548915356785303, + "learning_rate": 7.4562628200292475e-06, + "loss": 0.2148, + "step": 877 + }, + { + "epoch": 1.8329853862212944, + "grad_norm": 0.03888900644009877, + "learning_rate": 7.274411432392026e-06, + "loss": 0.2812, + "step": 878 + }, + { + "epoch": 1.8350730688935282, + "grad_norm": 0.036848786139394796, + "learning_rate": 7.0947640229023675e-06, + "loss": 0.1904, + "step": 879 + }, + { + "epoch": 1.837160751565762, + "grad_norm": 0.032594039620939604, + "learning_rate": 6.917322645978463e-06, + "loss": 0.1904, + "step": 880 + }, + { + "epoch": 1.8392484342379958, + "grad_norm": 0.038807798544498055, + "learning_rate": 6.742089330810774e-06, + "loss": 0.2383, + "step": 881 + }, + { + "epoch": 1.8413361169102296, + "grad_norm": 0.042279119809553053, + "learning_rate": 6.5690660813386484e-06, + "loss": 0.3184, + "step": 882 + }, + { + "epoch": 1.8434237995824634, + "grad_norm": 0.03575523378084652, + "learning_rate": 6.398254876227561e-06, + "loss": 0.1875, + "step": 883 + }, + { + "epoch": 1.8455114822546972, + "grad_norm": 0.03438492023581685, + "learning_rate": 6.229657668846333e-06, + "loss": 0.21, + "step": 884 + }, + { + "epoch": 1.847599164926931, + "grad_norm": 0.04355612584591555, + "learning_rate": 6.0632763872449535e-06, + "loss": 0.2754, + "step": 885 + }, + { + "epoch": 1.8496868475991648, + "grad_norm": 0.03533181091838774, + "learning_rate": 5.899112934132345e-06, + "loss": 0.1689, + "step": 886 + }, + { + "epoch": 1.8517745302713986, + "grad_norm": 0.040660080641486476, + "learning_rate": 5.73716918685483e-06, + "loss": 0.2471, + "step": 887 + }, + { + "epoch": 1.8538622129436324, + "grad_norm": 0.03235795768355878, + "learning_rate": 5.577446997374458e-06, + "loss": 0.1826, + "step": 888 + }, + { + "epoch": 1.8559498956158664, + "grad_norm": 0.03609407610302125, + "learning_rate": 5.4199481922479146e-06, + "loss": 0.1816, + "step": 889 + }, + { + "epoch": 1.8580375782881002, + "grad_norm": 0.03647085365814291, + "learning_rate": 5.264674572605665e-06, + "loss": 0.2393, + "step": 890 + }, + { + "epoch": 1.860125260960334, + "grad_norm": 0.03892481200719314, + "learning_rate": 5.11162791413129e-06, + "loss": 0.2021, + "step": 891 + }, + { + "epoch": 1.8622129436325678, + "grad_norm": 0.03859819801659933, + "learning_rate": 4.9608099670412025e-06, + "loss": 0.2402, + "step": 892 + }, + { + "epoch": 1.8643006263048016, + "grad_norm": 0.03638768037937149, + "learning_rate": 4.812222456064697e-06, + "loss": 0.1729, + "step": 893 + }, + { + "epoch": 1.8663883089770354, + "grad_norm": 0.03578385304837865, + "learning_rate": 4.665867080424046e-06, + "loss": 0.2217, + "step": 894 + }, + { + "epoch": 1.8684759916492695, + "grad_norm": 0.03899470496061886, + "learning_rate": 4.521745513815345e-06, + "loss": 0.2236, + "step": 895 + }, + { + "epoch": 1.8705636743215033, + "grad_norm": 0.03565199656237529, + "learning_rate": 4.379859404389075e-06, + "loss": 0.2158, + "step": 896 + }, + { + "epoch": 1.872651356993737, + "grad_norm": 0.03700782502510176, + "learning_rate": 4.240210374731479e-06, + "loss": 0.2441, + "step": 897 + }, + { + "epoch": 1.8747390396659709, + "grad_norm": 0.034012775228610695, + "learning_rate": 4.10280002184591e-06, + "loss": 0.1934, + "step": 898 + }, + { + "epoch": 1.8768267223382047, + "grad_norm": 0.032231872294556166, + "learning_rate": 3.967629917134574e-06, + "loss": 0.1689, + "step": 899 + }, + { + "epoch": 1.8789144050104385, + "grad_norm": 0.039125390388194826, + "learning_rate": 3.834701606380575e-06, + "loss": 0.2285, + "step": 900 + }, + { + "epoch": 1.8810020876826723, + "grad_norm": 0.04416962700457997, + "learning_rate": 3.7040166097301877e-06, + "loss": 0.2656, + "step": 901 + }, + { + "epoch": 1.883089770354906, + "grad_norm": 0.04111412261545973, + "learning_rate": 3.575576421675586e-06, + "loss": 0.248, + "step": 902 + }, + { + "epoch": 1.88517745302714, + "grad_norm": 0.03397805823890809, + "learning_rate": 3.449382511037613e-06, + "loss": 0.1904, + "step": 903 + }, + { + "epoch": 1.8872651356993737, + "grad_norm": 0.041722415283545866, + "learning_rate": 3.3254363209491046e-06, + "loss": 0.3301, + "step": 904 + }, + { + "epoch": 1.8893528183716075, + "grad_norm": 0.047301660349510964, + "learning_rate": 3.203739268838324e-06, + "loss": 0.2988, + "step": 905 + }, + { + "epoch": 1.8914405010438413, + "grad_norm": 0.03091322302919706, + "learning_rate": 3.0842927464127537e-06, + "loss": 0.1592, + "step": 906 + }, + { + "epoch": 1.8935281837160751, + "grad_norm": 0.038250787995771175, + "learning_rate": 2.9670981196431525e-06, + "loss": 0.248, + "step": 907 + }, + { + "epoch": 1.895615866388309, + "grad_norm": 0.04184624767282992, + "learning_rate": 2.8521567287480344e-06, + "loss": 0.2461, + "step": 908 + }, + { + "epoch": 1.8977035490605427, + "grad_norm": 0.03669678971086905, + "learning_rate": 2.7394698881782367e-06, + "loss": 0.1895, + "step": 909 + }, + { + "epoch": 1.8997912317327765, + "grad_norm": 0.03332356389562326, + "learning_rate": 2.6290388866019533e-06, + "loss": 0.1934, + "step": 910 + }, + { + "epoch": 1.9018789144050103, + "grad_norm": 0.03567454628961636, + "learning_rate": 2.5208649868899704e-06, + "loss": 0.2207, + "step": 911 + }, + { + "epoch": 1.9039665970772441, + "grad_norm": 0.03723478614951914, + "learning_rate": 2.4149494261012097e-06, + "loss": 0.2002, + "step": 912 + }, + { + "epoch": 1.906054279749478, + "grad_norm": 0.03840886510699952, + "learning_rate": 2.3112934154686296e-06, + "loss": 0.1982, + "step": 913 + }, + { + "epoch": 1.9081419624217117, + "grad_norm": 0.03658052263392586, + "learning_rate": 2.2098981403852804e-06, + "loss": 0.2295, + "step": 914 + }, + { + "epoch": 1.9102296450939458, + "grad_norm": 0.042943855139820596, + "learning_rate": 2.1107647603908933e-06, + "loss": 0.2422, + "step": 915 + }, + { + "epoch": 1.9123173277661796, + "grad_norm": 0.038181568479852754, + "learning_rate": 2.0138944091585123e-06, + "loss": 0.1768, + "step": 916 + }, + { + "epoch": 1.9144050104384134, + "grad_norm": 0.04044402128707661, + "learning_rate": 1.91928819448155e-06, + "loss": 0.249, + "step": 917 + }, + { + "epoch": 1.9164926931106472, + "grad_norm": 0.04021345081714543, + "learning_rate": 1.8269471982611086e-06, + "loss": 0.2578, + "step": 918 + }, + { + "epoch": 1.918580375782881, + "grad_norm": 0.03765027742041835, + "learning_rate": 1.7368724764936118e-06, + "loss": 0.2275, + "step": 919 + }, + { + "epoch": 1.9206680584551148, + "grad_norm": 0.03276750255084186, + "learning_rate": 1.6490650592588363e-06, + "loss": 0.1797, + "step": 920 + }, + { + "epoch": 1.9227557411273486, + "grad_norm": 0.03836166999423338, + "learning_rate": 1.5635259507079224e-06, + "loss": 0.2656, + "step": 921 + }, + { + "epoch": 1.9248434237995826, + "grad_norm": 0.040716455351115934, + "learning_rate": 1.4802561290520934e-06, + "loss": 0.2373, + "step": 922 + }, + { + "epoch": 1.9269311064718164, + "grad_norm": 0.044722992742083724, + "learning_rate": 1.3992565465512863e-06, + "loss": 0.2832, + "step": 923 + }, + { + "epoch": 1.9290187891440502, + "grad_norm": 0.035718501429723326, + "learning_rate": 1.3205281295034512e-06, + "loss": 0.2109, + "step": 924 + }, + { + "epoch": 1.931106471816284, + "grad_norm": 0.040897036162254374, + "learning_rate": 1.2440717782338018e-06, + "loss": 0.248, + "step": 925 + }, + { + "epoch": 1.9331941544885178, + "grad_norm": 0.037701179364082175, + "learning_rate": 1.169888367084626e-06, + "loss": 0.2021, + "step": 926 + }, + { + "epoch": 1.9352818371607516, + "grad_norm": 0.03800126744938023, + "learning_rate": 1.0979787444051815e-06, + "loss": 0.2051, + "step": 927 + }, + { + "epoch": 1.9373695198329854, + "grad_norm": 0.0376053096363303, + "learning_rate": 1.0283437325421252e-06, + "loss": 0.25, + "step": 928 + }, + { + "epoch": 1.9394572025052192, + "grad_norm": 0.03852373604309701, + "learning_rate": 9.60984127829989e-07, + "loss": 0.1895, + "step": 929 + }, + { + "epoch": 1.941544885177453, + "grad_norm": 0.05141719488909021, + "learning_rate": 8.95900700582164e-07, + "loss": 0.2969, + "step": 930 + }, + { + "epoch": 1.9436325678496869, + "grad_norm": 0.0422557444402279, + "learning_rate": 8.330941950819738e-07, + "loss": 0.2793, + "step": 931 + }, + { + "epoch": 1.9457202505219207, + "grad_norm": 0.03878673278759856, + "learning_rate": 7.725653295743484e-07, + "loss": 0.2412, + "step": 932 + }, + { + "epoch": 1.9478079331941545, + "grad_norm": 0.04515324231545769, + "learning_rate": 7.14314796257387e-07, + "loss": 0.2715, + "step": 933 + }, + { + "epoch": 1.9498956158663883, + "grad_norm": 0.04060496658163304, + "learning_rate": 6.58343261274652e-07, + "loss": 0.2559, + "step": 934 + }, + { + "epoch": 1.951983298538622, + "grad_norm": 0.0410883973174675, + "learning_rate": 6.046513647074203e-07, + "loss": 0.2676, + "step": 935 + }, + { + "epoch": 1.9540709812108559, + "grad_norm": 0.037425826708713576, + "learning_rate": 5.532397205673556e-07, + "loss": 0.208, + "step": 936 + }, + { + "epoch": 1.9561586638830897, + "grad_norm": 0.03711360512013879, + "learning_rate": 5.04108916789603e-07, + "loss": 0.2246, + "step": 937 + }, + { + "epoch": 1.9582463465553235, + "grad_norm": 0.04129502522262085, + "learning_rate": 4.5725951522592734e-07, + "loss": 0.2812, + "step": 938 + }, + { + "epoch": 1.9603340292275573, + "grad_norm": 0.042768062565203456, + "learning_rate": 4.126920516384303e-07, + "loss": 0.2734, + "step": 939 + }, + { + "epoch": 1.962421711899791, + "grad_norm": 0.03636718305173902, + "learning_rate": 3.704070356932432e-07, + "loss": 0.2275, + "step": 940 + }, + { + "epoch": 1.964509394572025, + "grad_norm": 0.03914519289291409, + "learning_rate": 3.3040495095491006e-07, + "loss": 0.2051, + "step": 941 + }, + { + "epoch": 1.966597077244259, + "grad_norm": 0.036435492987632524, + "learning_rate": 2.926862548806364e-07, + "loss": 0.21, + "step": 942 + }, + { + "epoch": 1.9686847599164927, + "grad_norm": 0.04520438858168889, + "learning_rate": 2.5725137881529306e-07, + "loss": 0.2852, + "step": 943 + }, + { + "epoch": 1.9707724425887265, + "grad_norm": 0.036719387383711895, + "learning_rate": 2.2410072798624283e-07, + "loss": 0.2422, + "step": 944 + }, + { + "epoch": 1.9728601252609603, + "grad_norm": 0.03890839218462144, + "learning_rate": 1.9323468149892165e-07, + "loss": 0.2617, + "step": 945 + }, + { + "epoch": 1.9749478079331941, + "grad_norm": 0.04401800316091134, + "learning_rate": 1.6465359233233114e-07, + "loss": 0.2314, + "step": 946 + }, + { + "epoch": 1.977035490605428, + "grad_norm": 0.042526472328545395, + "learning_rate": 1.383577873351305e-07, + "loss": 0.2734, + "step": 947 + }, + { + "epoch": 1.9791231732776617, + "grad_norm": 0.04123640104260365, + "learning_rate": 1.143475672218175e-07, + "loss": 0.2715, + "step": 948 + }, + { + "epoch": 1.9812108559498958, + "grad_norm": 0.041143122691388076, + "learning_rate": 9.26232065693089e-08, + "loss": 0.2578, + "step": 949 + }, + { + "epoch": 1.9832985386221296, + "grad_norm": 0.03646932624947557, + "learning_rate": 7.318495381383184e-08, + "loss": 0.1953, + "step": 950 + }, + { + "epoch": 1.9853862212943634, + "grad_norm": 0.042625916797236235, + "learning_rate": 5.603303124803727e-08, + "loss": 0.2344, + "step": 951 + }, + { + "epoch": 1.9874739039665972, + "grad_norm": 0.039620902455050594, + "learning_rate": 4.1167635018513064e-08, + "loss": 0.2598, + "step": 952 + }, + { + "epoch": 1.989561586638831, + "grad_norm": 0.0329100227494744, + "learning_rate": 2.8588935123474714e-08, + "loss": 0.1758, + "step": 953 + }, + { + "epoch": 1.9916492693110648, + "grad_norm": 0.03829557846462901, + "learning_rate": 1.8297075410877997e-08, + "loss": 0.2676, + "step": 954 + }, + { + "epoch": 1.9937369519832986, + "grad_norm": 0.037593692719248624, + "learning_rate": 1.02921735767314e-08, + "loss": 0.1943, + "step": 955 + }, + { + "epoch": 1.9958246346555324, + "grad_norm": 0.03806855037144764, + "learning_rate": 4.574321163763884e-09, + "loss": 0.2383, + "step": 956 + }, + { + "epoch": 1.9979123173277662, + "grad_norm": 0.03559582629136701, + "learning_rate": 1.1435835604034495e-09, + "loss": 0.1787, + "step": 957 + }, + { + "epoch": 2.0, + "grad_norm": 0.03239360059033573, + "learning_rate": 0.0, + "loss": 0.167, + "step": 958 + }, + { + "epoch": 2.0, + "step": 958, + "total_flos": 5.650597015940235e+18, + "train_loss": 0.3088689816022965, + "train_runtime": 6060.4121, + "train_samples_per_second": 90.951, + "train_steps_per_second": 0.158 + } + ], + "logging_steps": 1.0, + "max_steps": 958, + "num_input_tokens_seen": 0, + "num_train_epochs": 2, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 5.650597015940235e+18, + "train_batch_size": 72, + "trial_name": null, + "trial_params": null +}