| { |
| "best_global_step": null, |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 1.0, |
| "eval_steps": 500, |
| "global_step": 1656, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.006041839740200891, |
| "grad_norm": 8.275440544361343, |
| "learning_rate": 5.421686746987952e-07, |
| "loss": 0.818, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.012083679480401783, |
| "grad_norm": 2.7914609484184627, |
| "learning_rate": 1.1445783132530121e-06, |
| "loss": 0.6987, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.018125519220602675, |
| "grad_norm": 1.3664105415610248, |
| "learning_rate": 1.7469879518072292e-06, |
| "loss": 0.5802, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.024167358960803565, |
| "grad_norm": 0.8594872429754616, |
| "learning_rate": 2.349397590361446e-06, |
| "loss": 0.5258, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.030209198701004455, |
| "grad_norm": 0.7434393729226938, |
| "learning_rate": 2.9518072289156627e-06, |
| "loss": 0.5145, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.03625103844120535, |
| "grad_norm": 0.8872227027342147, |
| "learning_rate": 3.5542168674698798e-06, |
| "loss": 0.5015, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.042292878181406236, |
| "grad_norm": 0.8429713488596823, |
| "learning_rate": 4.156626506024097e-06, |
| "loss": 0.4901, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.04833471792160713, |
| "grad_norm": 0.9065488209295363, |
| "learning_rate": 4.759036144578314e-06, |
| "loss": 0.4796, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.05437655766180802, |
| "grad_norm": 0.9651498125714989, |
| "learning_rate": 5.361445783132531e-06, |
| "loss": 0.4758, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.06041839740200891, |
| "grad_norm": 0.8301785431193227, |
| "learning_rate": 5.963855421686747e-06, |
| "loss": 0.4829, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.0664602371422098, |
| "grad_norm": 0.8886113651753562, |
| "learning_rate": 6.566265060240964e-06, |
| "loss": 0.4812, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.0725020768824107, |
| "grad_norm": 0.9862797318907437, |
| "learning_rate": 7.168674698795182e-06, |
| "loss": 0.4789, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.07854391662261158, |
| "grad_norm": 1.0637568246181046, |
| "learning_rate": 7.771084337349398e-06, |
| "loss": 0.4767, |
| "step": 130 |
| }, |
| { |
| "epoch": 0.08458575636281247, |
| "grad_norm": 1.019587639434787, |
| "learning_rate": 8.373493975903614e-06, |
| "loss": 0.4748, |
| "step": 140 |
| }, |
| { |
| "epoch": 0.09062759610301337, |
| "grad_norm": 1.0494143456183387, |
| "learning_rate": 8.975903614457832e-06, |
| "loss": 0.471, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.09666943584321426, |
| "grad_norm": 1.0167856442156524, |
| "learning_rate": 9.57831325301205e-06, |
| "loss": 0.482, |
| "step": 160 |
| }, |
| { |
| "epoch": 0.10271127558341515, |
| "grad_norm": 0.9927765526681102, |
| "learning_rate": 9.99989997506481e-06, |
| "loss": 0.4867, |
| "step": 170 |
| }, |
| { |
| "epoch": 0.10875311532361603, |
| "grad_norm": 0.8252285160527216, |
| "learning_rate": 9.998121865323769e-06, |
| "loss": 0.4832, |
| "step": 180 |
| }, |
| { |
| "epoch": 0.11479495506381693, |
| "grad_norm": 1.0494152999510933, |
| "learning_rate": 9.994121889085426e-06, |
| "loss": 0.4836, |
| "step": 190 |
| }, |
| { |
| "epoch": 0.12083679480401782, |
| "grad_norm": 0.8683126435696938, |
| "learning_rate": 9.987901824500342e-06, |
| "loss": 0.4796, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.12687863454421872, |
| "grad_norm": 1.0765238837274924, |
| "learning_rate": 9.979464436637771e-06, |
| "loss": 0.479, |
| "step": 210 |
| }, |
| { |
| "epoch": 0.1329204742844196, |
| "grad_norm": 0.9747782968326728, |
| "learning_rate": 9.968813476256483e-06, |
| "loss": 0.4774, |
| "step": 220 |
| }, |
| { |
| "epoch": 0.1389623140246205, |
| "grad_norm": 1.0422947760995596, |
| "learning_rate": 9.955953678137397e-06, |
| "loss": 0.4775, |
| "step": 230 |
| }, |
| { |
| "epoch": 0.1450041537648214, |
| "grad_norm": 0.8638869125481701, |
| "learning_rate": 9.940890758978781e-06, |
| "loss": 0.4809, |
| "step": 240 |
| }, |
| { |
| "epoch": 0.1510459935050223, |
| "grad_norm": 0.9908759029078118, |
| "learning_rate": 9.923631414854946e-06, |
| "loss": 0.4823, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.15708783324522316, |
| "grad_norm": 0.8594590054265251, |
| "learning_rate": 9.904183318239573e-06, |
| "loss": 0.4732, |
| "step": 260 |
| }, |
| { |
| "epoch": 0.16312967298542405, |
| "grad_norm": 0.8537166815636835, |
| "learning_rate": 9.882555114594994e-06, |
| "loss": 0.4746, |
| "step": 270 |
| }, |
| { |
| "epoch": 0.16917151272562495, |
| "grad_norm": 0.7988607086870596, |
| "learning_rate": 9.858756418528928e-06, |
| "loss": 0.4771, |
| "step": 280 |
| }, |
| { |
| "epoch": 0.17521335246582584, |
| "grad_norm": 0.8128733549061342, |
| "learning_rate": 9.832797809520404e-06, |
| "loss": 0.4703, |
| "step": 290 |
| }, |
| { |
| "epoch": 0.18125519220602673, |
| "grad_norm": 0.8670976233152772, |
| "learning_rate": 9.804690827216764e-06, |
| "loss": 0.4763, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.18729703194622763, |
| "grad_norm": 0.9163703807052254, |
| "learning_rate": 9.77444796630381e-06, |
| "loss": 0.4665, |
| "step": 310 |
| }, |
| { |
| "epoch": 0.19333887168642852, |
| "grad_norm": 1.0505077270930072, |
| "learning_rate": 9.742082670951423e-06, |
| "loss": 0.4781, |
| "step": 320 |
| }, |
| { |
| "epoch": 0.19938071142662941, |
| "grad_norm": 0.9771619804123264, |
| "learning_rate": 9.707609328837085e-06, |
| "loss": 0.4695, |
| "step": 330 |
| }, |
| { |
| "epoch": 0.2054225511668303, |
| "grad_norm": 0.8443160970596196, |
| "learning_rate": 9.67104326474998e-06, |
| "loss": 0.4709, |
| "step": 340 |
| }, |
| { |
| "epoch": 0.2114643909070312, |
| "grad_norm": 0.7405442085507775, |
| "learning_rate": 9.632400733778504e-06, |
| "loss": 0.4687, |
| "step": 350 |
| }, |
| { |
| "epoch": 0.21750623064723207, |
| "grad_norm": 1.06910781919214, |
| "learning_rate": 9.591698914084224e-06, |
| "loss": 0.4727, |
| "step": 360 |
| }, |
| { |
| "epoch": 0.22354807038743296, |
| "grad_norm": 0.9783321165505232, |
| "learning_rate": 9.548955899265495e-06, |
| "loss": 0.4757, |
| "step": 370 |
| }, |
| { |
| "epoch": 0.22958991012763386, |
| "grad_norm": 0.856614778477577, |
| "learning_rate": 9.504190690314124e-06, |
| "loss": 0.4628, |
| "step": 380 |
| }, |
| { |
| "epoch": 0.23563174986783475, |
| "grad_norm": 0.8407994854798428, |
| "learning_rate": 9.457423187168667e-06, |
| "loss": 0.4751, |
| "step": 390 |
| }, |
| { |
| "epoch": 0.24167358960803564, |
| "grad_norm": 0.835330085907184, |
| "learning_rate": 9.408674179868101e-06, |
| "loss": 0.4755, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.24771542934823654, |
| "grad_norm": 0.8184694750958316, |
| "learning_rate": 9.357965339309823e-06, |
| "loss": 0.4688, |
| "step": 410 |
| }, |
| { |
| "epoch": 0.25375726908843743, |
| "grad_norm": 0.8556998098855371, |
| "learning_rate": 9.305319207616048e-06, |
| "loss": 0.468, |
| "step": 420 |
| }, |
| { |
| "epoch": 0.2597991088286383, |
| "grad_norm": 0.7125631433834922, |
| "learning_rate": 9.250759188112944e-06, |
| "loss": 0.4721, |
| "step": 430 |
| }, |
| { |
| "epoch": 0.2658409485688392, |
| "grad_norm": 0.8684064304089488, |
| "learning_rate": 9.194309534926896e-06, |
| "loss": 0.4719, |
| "step": 440 |
| }, |
| { |
| "epoch": 0.2718827883090401, |
| "grad_norm": 0.8581281745281804, |
| "learning_rate": 9.13599534220258e-06, |
| "loss": 0.4676, |
| "step": 450 |
| }, |
| { |
| "epoch": 0.277924628049241, |
| "grad_norm": 0.8073909041822805, |
| "learning_rate": 9.075842532947605e-06, |
| "loss": 0.4729, |
| "step": 460 |
| }, |
| { |
| "epoch": 0.2839664677894419, |
| "grad_norm": 0.7760503268897376, |
| "learning_rate": 9.013877847508684e-06, |
| "loss": 0.4735, |
| "step": 470 |
| }, |
| { |
| "epoch": 0.2900083075296428, |
| "grad_norm": 1.007511777834999, |
| "learning_rate": 8.95012883168448e-06, |
| "loss": 0.4655, |
| "step": 480 |
| }, |
| { |
| "epoch": 0.2960501472698437, |
| "grad_norm": 0.8792795612951879, |
| "learning_rate": 8.884623824480388e-06, |
| "loss": 0.4604, |
| "step": 490 |
| }, |
| { |
| "epoch": 0.3020919870100446, |
| "grad_norm": 0.7649844789740483, |
| "learning_rate": 8.817391945510697e-06, |
| "loss": 0.4727, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.3081338267502455, |
| "grad_norm": 0.7112188721793731, |
| "learning_rate": 8.748463082053765e-06, |
| "loss": 0.4696, |
| "step": 510 |
| }, |
| { |
| "epoch": 0.3141756664904463, |
| "grad_norm": 0.982736127204613, |
| "learning_rate": 8.677867875765904e-06, |
| "loss": 0.4667, |
| "step": 520 |
| }, |
| { |
| "epoch": 0.3202175062306472, |
| "grad_norm": 0.7380835460961964, |
| "learning_rate": 8.605637709059937e-06, |
| "loss": 0.4665, |
| "step": 530 |
| }, |
| { |
| "epoch": 0.3262593459708481, |
| "grad_norm": 0.889341527636709, |
| "learning_rate": 8.531804691154454e-06, |
| "loss": 0.4596, |
| "step": 540 |
| }, |
| { |
| "epoch": 0.332301185711049, |
| "grad_norm": 0.8717590020119719, |
| "learning_rate": 8.45640164379996e-06, |
| "loss": 0.4635, |
| "step": 550 |
| }, |
| { |
| "epoch": 0.3383430254512499, |
| "grad_norm": 0.8593981074116124, |
| "learning_rate": 8.379462086688302e-06, |
| "loss": 0.4614, |
| "step": 560 |
| }, |
| { |
| "epoch": 0.3443848651914508, |
| "grad_norm": 0.7452870255566968, |
| "learning_rate": 8.301020222551795e-06, |
| "loss": 0.4613, |
| "step": 570 |
| }, |
| { |
| "epoch": 0.3504267049316517, |
| "grad_norm": 0.6738237626130682, |
| "learning_rate": 8.221110921958748e-06, |
| "loss": 0.4625, |
| "step": 580 |
| }, |
| { |
| "epoch": 0.35646854467185257, |
| "grad_norm": 0.7475917228131528, |
| "learning_rate": 8.139769707812083e-06, |
| "loss": 0.4643, |
| "step": 590 |
| }, |
| { |
| "epoch": 0.36251038441205347, |
| "grad_norm": 0.8625312431506086, |
| "learning_rate": 8.057032739557973e-06, |
| "loss": 0.4647, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.36855222415225436, |
| "grad_norm": 0.7686541350728158, |
| "learning_rate": 7.972936797111515e-06, |
| "loss": 0.4664, |
| "step": 610 |
| }, |
| { |
| "epoch": 0.37459406389245525, |
| "grad_norm": 0.793893791173724, |
| "learning_rate": 7.887519264506577e-06, |
| "loss": 0.4687, |
| "step": 620 |
| }, |
| { |
| "epoch": 0.38063590363265615, |
| "grad_norm": 0.8889160126310801, |
| "learning_rate": 7.800818113277085e-06, |
| "loss": 0.4675, |
| "step": 630 |
| }, |
| { |
| "epoch": 0.38667774337285704, |
| "grad_norm": 0.73855111981923, |
| "learning_rate": 7.712871885577147e-06, |
| "loss": 0.4696, |
| "step": 640 |
| }, |
| { |
| "epoch": 0.39271958311305794, |
| "grad_norm": 0.9038531350448926, |
| "learning_rate": 7.623719677047521e-06, |
| "loss": 0.4679, |
| "step": 650 |
| }, |
| { |
| "epoch": 0.39876142285325883, |
| "grad_norm": 0.8258158060448149, |
| "learning_rate": 7.533401119436012e-06, |
| "loss": 0.4636, |
| "step": 660 |
| }, |
| { |
| "epoch": 0.4048032625934597, |
| "grad_norm": 0.7984996623335708, |
| "learning_rate": 7.441956362979584e-06, |
| "loss": 0.4573, |
| "step": 670 |
| }, |
| { |
| "epoch": 0.4108451023336606, |
| "grad_norm": 0.9068733833902644, |
| "learning_rate": 7.349426058555943e-06, |
| "loss": 0.467, |
| "step": 680 |
| }, |
| { |
| "epoch": 0.4168869420738615, |
| "grad_norm": 0.7305483298541591, |
| "learning_rate": 7.255851339612597e-06, |
| "loss": 0.4619, |
| "step": 690 |
| }, |
| { |
| "epoch": 0.4229287818140624, |
| "grad_norm": 0.7441065423891078, |
| "learning_rate": 7.161273803881381e-06, |
| "loss": 0.4491, |
| "step": 700 |
| }, |
| { |
| "epoch": 0.4289706215542633, |
| "grad_norm": 0.9276808876681103, |
| "learning_rate": 7.0657354948865786e-06, |
| "loss": 0.4532, |
| "step": 710 |
| }, |
| { |
| "epoch": 0.43501246129446414, |
| "grad_norm": 0.6605412717277142, |
| "learning_rate": 6.969278883254896e-06, |
| "loss": 0.4646, |
| "step": 720 |
| }, |
| { |
| "epoch": 0.44105430103466503, |
| "grad_norm": 0.6593545188739028, |
| "learning_rate": 6.871946847835548e-06, |
| "loss": 0.4542, |
| "step": 730 |
| }, |
| { |
| "epoch": 0.4470961407748659, |
| "grad_norm": 0.7295116693541216, |
| "learning_rate": 6.773782656638889e-06, |
| "loss": 0.4532, |
| "step": 740 |
| }, |
| { |
| "epoch": 0.4531379805150668, |
| "grad_norm": 0.7877222556349484, |
| "learning_rate": 6.674829947602034e-06, |
| "loss": 0.464, |
| "step": 750 |
| }, |
| { |
| "epoch": 0.4591798202552677, |
| "grad_norm": 0.7208466663864777, |
| "learning_rate": 6.575132709190041e-06, |
| "loss": 0.4604, |
| "step": 760 |
| }, |
| { |
| "epoch": 0.4652216599954686, |
| "grad_norm": 0.8296261142467121, |
| "learning_rate": 6.474735260841264e-06, |
| "loss": 0.4604, |
| "step": 770 |
| }, |
| { |
| "epoch": 0.4712634997356695, |
| "grad_norm": 0.7971141568003346, |
| "learning_rate": 6.373682233265581e-06, |
| "loss": 0.464, |
| "step": 780 |
| }, |
| { |
| "epoch": 0.4773053394758704, |
| "grad_norm": 0.8736349174540989, |
| "learning_rate": 6.2720185486042405e-06, |
| "loss": 0.4679, |
| "step": 790 |
| }, |
| { |
| "epoch": 0.4833471792160713, |
| "grad_norm": 0.9309796515999401, |
| "learning_rate": 6.169789400460167e-06, |
| "loss": 0.4613, |
| "step": 800 |
| }, |
| { |
| "epoch": 0.4893890189562722, |
| "grad_norm": 0.7752657579479202, |
| "learning_rate": 6.067040233807579e-06, |
| "loss": 0.4561, |
| "step": 810 |
| }, |
| { |
| "epoch": 0.4954308586964731, |
| "grad_norm": 0.8534120955438902, |
| "learning_rate": 5.963816724789868e-06, |
| "loss": 0.4488, |
| "step": 820 |
| }, |
| { |
| "epoch": 0.501472698436674, |
| "grad_norm": 0.8496125433352629, |
| "learning_rate": 5.860164760414715e-06, |
| "loss": 0.4618, |
| "step": 830 |
| }, |
| { |
| "epoch": 0.5075145381768749, |
| "grad_norm": 0.8251415980778435, |
| "learning_rate": 5.7561304181554626e-06, |
| "loss": 0.4593, |
| "step": 840 |
| }, |
| { |
| "epoch": 0.5135563779170758, |
| "grad_norm": 0.7139950510608393, |
| "learning_rate": 5.651759945467829e-06, |
| "loss": 0.4502, |
| "step": 850 |
| }, |
| { |
| "epoch": 0.5195982176572767, |
| "grad_norm": 0.601943553224525, |
| "learning_rate": 5.5470997392310475e-06, |
| "loss": 0.4619, |
| "step": 860 |
| }, |
| { |
| "epoch": 0.5256400573974775, |
| "grad_norm": 0.8153866464407951, |
| "learning_rate": 5.442196325122583e-06, |
| "loss": 0.4588, |
| "step": 870 |
| }, |
| { |
| "epoch": 0.5316818971376784, |
| "grad_norm": 0.8344346217932002, |
| "learning_rate": 5.337096336935596e-06, |
| "loss": 0.4606, |
| "step": 880 |
| }, |
| { |
| "epoch": 0.5377237368778793, |
| "grad_norm": 0.7492780666259381, |
| "learning_rate": 5.231846495848343e-06, |
| "loss": 0.4408, |
| "step": 890 |
| }, |
| { |
| "epoch": 0.5437655766180802, |
| "grad_norm": 0.7261206428098568, |
| "learning_rate": 5.12649358965473e-06, |
| "loss": 0.4538, |
| "step": 900 |
| }, |
| { |
| "epoch": 0.5498074163582811, |
| "grad_norm": 0.7279393955754347, |
| "learning_rate": 5.021084451965257e-06, |
| "loss": 0.4532, |
| "step": 910 |
| }, |
| { |
| "epoch": 0.555849256098482, |
| "grad_norm": 0.825743775909579, |
| "learning_rate": 4.915665941387589e-06, |
| "loss": 0.4496, |
| "step": 920 |
| }, |
| { |
| "epoch": 0.5618910958386829, |
| "grad_norm": 0.7524696024763261, |
| "learning_rate": 4.8102849206960144e-06, |
| "loss": 0.4521, |
| "step": 930 |
| }, |
| { |
| "epoch": 0.5679329355788838, |
| "grad_norm": 0.6103762891076722, |
| "learning_rate": 4.704988235999059e-06, |
| "loss": 0.4537, |
| "step": 940 |
| }, |
| { |
| "epoch": 0.5739747753190847, |
| "grad_norm": 0.7615515295808474, |
| "learning_rate": 4.5998226959145e-06, |
| "loss": 0.4437, |
| "step": 950 |
| }, |
| { |
| "epoch": 0.5800166150592856, |
| "grad_norm": 0.8148357524689337, |
| "learning_rate": 4.494835050761055e-06, |
| "loss": 0.4506, |
| "step": 960 |
| }, |
| { |
| "epoch": 0.5860584547994865, |
| "grad_norm": 0.746845262451762, |
| "learning_rate": 4.390071971775977e-06, |
| "loss": 0.4496, |
| "step": 970 |
| }, |
| { |
| "epoch": 0.5921002945396874, |
| "grad_norm": 0.7650124837023387, |
| "learning_rate": 4.285580030367812e-06, |
| "loss": 0.4442, |
| "step": 980 |
| }, |
| { |
| "epoch": 0.5981421342798883, |
| "grad_norm": 0.7044872850784959, |
| "learning_rate": 4.18140567741353e-06, |
| "loss": 0.4515, |
| "step": 990 |
| }, |
| { |
| "epoch": 0.6041839740200892, |
| "grad_norm": 0.6835203586178121, |
| "learning_rate": 4.07759522260924e-06, |
| "loss": 0.4484, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.6102258137602901, |
| "grad_norm": 0.7875522158264618, |
| "learning_rate": 3.974194813883672e-06, |
| "loss": 0.4487, |
| "step": 1010 |
| }, |
| { |
| "epoch": 0.616267653500491, |
| "grad_norm": 1.0731871419616035, |
| "learning_rate": 3.871250416883547e-06, |
| "loss": 0.4487, |
| "step": 1020 |
| }, |
| { |
| "epoch": 0.6223094932406917, |
| "grad_norm": 0.7645842030603645, |
| "learning_rate": 3.7688077945400135e-06, |
| "loss": 0.4476, |
| "step": 1030 |
| }, |
| { |
| "epoch": 0.6283513329808926, |
| "grad_norm": 0.8262016423595615, |
| "learning_rate": 3.6669124867251705e-06, |
| "loss": 0.4419, |
| "step": 1040 |
| }, |
| { |
| "epoch": 0.6343931727210935, |
| "grad_norm": 0.8630082766433129, |
| "learning_rate": 3.56560979000776e-06, |
| "loss": 0.4433, |
| "step": 1050 |
| }, |
| { |
| "epoch": 0.6404350124612944, |
| "grad_norm": 0.6739630234814128, |
| "learning_rate": 3.4649447375170243e-06, |
| "loss": 0.434, |
| "step": 1060 |
| }, |
| { |
| "epoch": 0.6464768522014953, |
| "grad_norm": 0.8005468876181695, |
| "learning_rate": 3.3649620789236613e-06, |
| "loss": 0.4449, |
| "step": 1070 |
| }, |
| { |
| "epoch": 0.6525186919416962, |
| "grad_norm": 0.6927017507913021, |
| "learning_rate": 3.265706260546793e-06, |
| "loss": 0.4487, |
| "step": 1080 |
| }, |
| { |
| "epoch": 0.6585605316818971, |
| "grad_norm": 0.7470116822150206, |
| "learning_rate": 3.1672214055957885e-06, |
| "loss": 0.4492, |
| "step": 1090 |
| }, |
| { |
| "epoch": 0.664602371422098, |
| "grad_norm": 0.7165981129271416, |
| "learning_rate": 3.0695512945557175e-06, |
| "loss": 0.4435, |
| "step": 1100 |
| }, |
| { |
| "epoch": 0.6706442111622989, |
| "grad_norm": 0.6713758708690007, |
| "learning_rate": 2.972739345725163e-06, |
| "loss": 0.445, |
| "step": 1110 |
| }, |
| { |
| "epoch": 0.6766860509024998, |
| "grad_norm": 0.7528881198034107, |
| "learning_rate": 2.876828595915043e-06, |
| "loss": 0.4458, |
| "step": 1120 |
| }, |
| { |
| "epoch": 0.6827278906427007, |
| "grad_norm": 1.3727779517766678, |
| "learning_rate": 2.781861681317004e-06, |
| "loss": 0.4362, |
| "step": 1130 |
| }, |
| { |
| "epoch": 0.6887697303829016, |
| "grad_norm": 0.8146352120444986, |
| "learning_rate": 2.687880818549927e-06, |
| "loss": 0.4396, |
| "step": 1140 |
| }, |
| { |
| "epoch": 0.6948115701231025, |
| "grad_norm": 0.6965825401455522, |
| "learning_rate": 2.5949277858929297e-06, |
| "loss": 0.4362, |
| "step": 1150 |
| }, |
| { |
| "epoch": 0.7008534098633034, |
| "grad_norm": 0.6495955670121708, |
| "learning_rate": 2.5030439047132484e-06, |
| "loss": 0.4453, |
| "step": 1160 |
| }, |
| { |
| "epoch": 0.7068952496035042, |
| "grad_norm": 0.6114269361750394, |
| "learning_rate": 2.4122700210972218e-06, |
| "loss": 0.4436, |
| "step": 1170 |
| }, |
| { |
| "epoch": 0.7129370893437051, |
| "grad_norm": 0.851795747172313, |
| "learning_rate": 2.322646487692556e-06, |
| "loss": 0.4437, |
| "step": 1180 |
| }, |
| { |
| "epoch": 0.718978929083906, |
| "grad_norm": 0.775135762171182, |
| "learning_rate": 2.2342131457699575e-06, |
| "loss": 0.4495, |
| "step": 1190 |
| }, |
| { |
| "epoch": 0.7250207688241069, |
| "grad_norm": 0.6479257056864712, |
| "learning_rate": 2.1470093075120686e-06, |
| "loss": 0.4383, |
| "step": 1200 |
| }, |
| { |
| "epoch": 0.7310626085643078, |
| "grad_norm": 0.8263286588325557, |
| "learning_rate": 2.061073738537635e-06, |
| "loss": 0.4463, |
| "step": 1210 |
| }, |
| { |
| "epoch": 0.7371044483045087, |
| "grad_norm": 0.6982438518266665, |
| "learning_rate": 1.9764446406686177e-06, |
| "loss": 0.4436, |
| "step": 1220 |
| }, |
| { |
| "epoch": 0.7431462880447096, |
| "grad_norm": 0.6700592747869402, |
| "learning_rate": 1.8931596349479385e-06, |
| "loss": 0.4432, |
| "step": 1230 |
| }, |
| { |
| "epoch": 0.7491881277849105, |
| "grad_norm": 0.5967953856950272, |
| "learning_rate": 1.8112557449154316e-06, |
| "loss": 0.4338, |
| "step": 1240 |
| }, |
| { |
| "epoch": 0.7552299675251114, |
| "grad_norm": 0.6985853030295077, |
| "learning_rate": 1.7307693801493619e-06, |
| "loss": 0.4407, |
| "step": 1250 |
| }, |
| { |
| "epoch": 0.7612718072653123, |
| "grad_norm": 0.7016362891096635, |
| "learning_rate": 1.6517363200809222e-06, |
| "loss": 0.4488, |
| "step": 1260 |
| }, |
| { |
| "epoch": 0.7673136470055132, |
| "grad_norm": 0.8402985241208492, |
| "learning_rate": 1.5741916980888267e-06, |
| "loss": 0.435, |
| "step": 1270 |
| }, |
| { |
| "epoch": 0.7733554867457141, |
| "grad_norm": 0.8240304279049092, |
| "learning_rate": 1.4981699858811038e-06, |
| "loss": 0.4436, |
| "step": 1280 |
| }, |
| { |
| "epoch": 0.779397326485915, |
| "grad_norm": 0.6660945380738031, |
| "learning_rate": 1.423704978171046e-06, |
| "loss": 0.4407, |
| "step": 1290 |
| }, |
| { |
| "epoch": 0.7854391662261159, |
| "grad_norm": 0.5874782271939601, |
| "learning_rate": 1.3508297776540845e-06, |
| "loss": 0.4498, |
| "step": 1300 |
| }, |
| { |
| "epoch": 0.7914810059663168, |
| "grad_norm": 0.5796653208549384, |
| "learning_rate": 1.2795767802923192e-06, |
| "loss": 0.4352, |
| "step": 1310 |
| }, |
| { |
| "epoch": 0.7975228457065177, |
| "grad_norm": 0.9463374855160357, |
| "learning_rate": 1.2099776609132048e-06, |
| "loss": 0.4384, |
| "step": 1320 |
| }, |
| { |
| "epoch": 0.8035646854467186, |
| "grad_norm": 0.6012369149099516, |
| "learning_rate": 1.1420633591288072e-06, |
| "loss": 0.441, |
| "step": 1330 |
| }, |
| { |
| "epoch": 0.8096065251869194, |
| "grad_norm": 0.7550415373921882, |
| "learning_rate": 1.0758640655819107e-06, |
| "loss": 0.4375, |
| "step": 1340 |
| }, |
| { |
| "epoch": 0.8156483649271203, |
| "grad_norm": 0.7006495875182843, |
| "learning_rate": 1.0114092085250566e-06, |
| "loss": 0.4439, |
| "step": 1350 |
| }, |
| { |
| "epoch": 0.8216902046673212, |
| "grad_norm": 0.6206161322638641, |
| "learning_rate": 9.487274407384972e-07, |
| "loss": 0.4359, |
| "step": 1360 |
| }, |
| { |
| "epoch": 0.8277320444075221, |
| "grad_norm": 0.6290155968597686, |
| "learning_rate": 8.878466267928814e-07, |
| "loss": 0.4366, |
| "step": 1370 |
| }, |
| { |
| "epoch": 0.833773884147723, |
| "grad_norm": 0.6195315234512612, |
| "learning_rate": 8.287938306623349e-07, |
| "loss": 0.4398, |
| "step": 1380 |
| }, |
| { |
| "epoch": 0.8398157238879239, |
| "grad_norm": 0.8102446928428507, |
| "learning_rate": 7.715953036934304e-07, |
| "loss": 0.4409, |
| "step": 1390 |
| }, |
| { |
| "epoch": 0.8458575636281248, |
| "grad_norm": 0.789309132604837, |
| "learning_rate": 7.162764729354033e-07, |
| "loss": 0.4407, |
| "step": 1400 |
| }, |
| { |
| "epoch": 0.8518994033683257, |
| "grad_norm": 0.6558463034047459, |
| "learning_rate": 6.628619298368133e-07, |
| "loss": 0.4297, |
| "step": 1410 |
| }, |
| { |
| "epoch": 0.8579412431085266, |
| "grad_norm": 0.6589202868449741, |
| "learning_rate": 6.11375419313644e-07, |
| "loss": 0.4397, |
| "step": 1420 |
| }, |
| { |
| "epoch": 0.8639830828487275, |
| "grad_norm": 0.6171089586498126, |
| "learning_rate": 5.618398291937393e-07, |
| "loss": 0.4473, |
| "step": 1430 |
| }, |
| { |
| "epoch": 0.8700249225889283, |
| "grad_norm": 0.8171417567736048, |
| "learning_rate": 5.142771800422403e-07, |
| "loss": 0.4424, |
| "step": 1440 |
| }, |
| { |
| "epoch": 0.8760667623291292, |
| "grad_norm": 0.813136783771392, |
| "learning_rate": 4.687086153725534e-07, |
| "loss": 0.4412, |
| "step": 1450 |
| }, |
| { |
| "epoch": 0.8821086020693301, |
| "grad_norm": 0.6976453824184702, |
| "learning_rate": 4.2515439224721066e-07, |
| "loss": 0.4332, |
| "step": 1460 |
| }, |
| { |
| "epoch": 0.888150441809531, |
| "grad_norm": 0.5748251232369562, |
| "learning_rate": 3.8363387227278947e-07, |
| "loss": 0.4348, |
| "step": 1470 |
| }, |
| { |
| "epoch": 0.8941922815497318, |
| "grad_norm": 0.720005515762108, |
| "learning_rate": 3.441655129928972e-07, |
| "loss": 0.4442, |
| "step": 1480 |
| }, |
| { |
| "epoch": 0.9002341212899327, |
| "grad_norm": 0.7252408165512142, |
| "learning_rate": 3.067668596830481e-07, |
| "loss": 0.4395, |
| "step": 1490 |
| }, |
| { |
| "epoch": 0.9062759610301336, |
| "grad_norm": 0.7144631857492548, |
| "learning_rate": 2.714545375510852e-07, |
| "loss": 0.4366, |
| "step": 1500 |
| }, |
| { |
| "epoch": 0.9123178007703345, |
| "grad_norm": 0.6635911944336934, |
| "learning_rate": 2.382442443466043e-07, |
| "loss": 0.4321, |
| "step": 1510 |
| }, |
| { |
| "epoch": 0.9183596405105354, |
| "grad_norm": 0.7758826608638922, |
| "learning_rate": 2.0715074338266915e-07, |
| "loss": 0.4312, |
| "step": 1520 |
| }, |
| { |
| "epoch": 0.9244014802507363, |
| "grad_norm": 0.7230232505488107, |
| "learning_rate": 1.7818785697292895e-07, |
| "loss": 0.4388, |
| "step": 1530 |
| }, |
| { |
| "epoch": 0.9304433199909372, |
| "grad_norm": 0.6580184350979571, |
| "learning_rate": 1.5136846028704132e-07, |
| "loss": 0.4491, |
| "step": 1540 |
| }, |
| { |
| "epoch": 0.9364851597311381, |
| "grad_norm": 0.5586825913101796, |
| "learning_rate": 1.2670447562713684e-07, |
| "loss": 0.4322, |
| "step": 1550 |
| }, |
| { |
| "epoch": 0.942526999471339, |
| "grad_norm": 0.647195128183336, |
| "learning_rate": 1.042068671278823e-07, |
| "loss": 0.4354, |
| "step": 1560 |
| }, |
| { |
| "epoch": 0.9485688392115399, |
| "grad_norm": 0.6035271594962103, |
| "learning_rate": 8.388563588247523e-08, |
| "loss": 0.4316, |
| "step": 1570 |
| }, |
| { |
| "epoch": 0.9546106789517408, |
| "grad_norm": 0.6113112433338288, |
| "learning_rate": 6.574981549676007e-08, |
| "loss": 0.4354, |
| "step": 1580 |
| }, |
| { |
| "epoch": 0.9606525186919417, |
| "grad_norm": 0.7146951417286433, |
| "learning_rate": 4.980746807342285e-08, |
| "loss": 0.4331, |
| "step": 1590 |
| }, |
| { |
| "epoch": 0.9666943584321426, |
| "grad_norm": 0.6521596489801758, |
| "learning_rate": 3.6065680628062924e-08, |
| "loss": 0.4397, |
| "step": 1600 |
| }, |
| { |
| "epoch": 0.9727361981723435, |
| "grad_norm": 0.8208153994914311, |
| "learning_rate": 2.4530561938729825e-08, |
| "loss": 0.4266, |
| "step": 1610 |
| }, |
| { |
| "epoch": 0.9787780379125444, |
| "grad_norm": 0.6500288352676025, |
| "learning_rate": 1.520723983032324e-08, |
| "loss": 0.4335, |
| "step": 1620 |
| }, |
| { |
| "epoch": 0.9848198776527453, |
| "grad_norm": 0.6538530895079608, |
| "learning_rate": 8.099858895072587e-09, |
| "loss": 0.4373, |
| "step": 1630 |
| }, |
| { |
| "epoch": 0.9908617173929462, |
| "grad_norm": 0.8433082091701005, |
| "learning_rate": 3.2115786500924728e-09, |
| "loss": 0.4371, |
| "step": 1640 |
| }, |
| { |
| "epoch": 0.996903557133147, |
| "grad_norm": 0.6909298344072548, |
| "learning_rate": 5.445721328567466e-10, |
| "loss": 0.4329, |
| "step": 1650 |
| }, |
| { |
| "epoch": 1.0, |
| "step": 1656, |
| "total_flos": 2042833294327808.0, |
| "train_loss": 0.4611289887036678, |
| "train_runtime": 68178.0817, |
| "train_samples_per_second": 1.554, |
| "train_steps_per_second": 0.024 |
| } |
| ], |
| "logging_steps": 10, |
| "max_steps": 1656, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 1, |
| "save_steps": 500, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 2042833294327808.0, |
| "train_batch_size": 1, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|