| { |
| "best_global_step": 9530, |
| "best_metric": 0.6577488780021667, |
| "best_model_checkpoint": "./finbert_weighted_final/checkpoint-9530", |
| "epoch": 2.0, |
| "eval_steps": 500, |
| "global_step": 19060, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.01049317943336831, |
| "grad_norm": 4.134560585021973, |
| "learning_rate": 1.9896117523609656e-05, |
| "loss": 0.7054, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.02098635886673662, |
| "grad_norm": 3.1322972774505615, |
| "learning_rate": 1.979118572927597e-05, |
| "loss": 0.7018, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.03147953830010493, |
| "grad_norm": 1.025286316871643, |
| "learning_rate": 1.968625393494229e-05, |
| "loss": 0.6951, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.04197271773347324, |
| "grad_norm": 2.2497143745422363, |
| "learning_rate": 1.9581322140608606e-05, |
| "loss": 0.6978, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.05246589716684155, |
| "grad_norm": 1.647702932357788, |
| "learning_rate": 1.9476390346274924e-05, |
| "loss": 0.6892, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.06295907660020986, |
| "grad_norm": 3.1609485149383545, |
| "learning_rate": 1.937145855194124e-05, |
| "loss": 0.6925, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.07345225603357818, |
| "grad_norm": 1.7412222623825073, |
| "learning_rate": 1.9266526757607556e-05, |
| "loss": 0.6903, |
| "step": 700 |
| }, |
| { |
| "epoch": 0.08394543546694648, |
| "grad_norm": 2.2502248287200928, |
| "learning_rate": 1.9161594963273874e-05, |
| "loss": 0.6942, |
| "step": 800 |
| }, |
| { |
| "epoch": 0.0944386149003148, |
| "grad_norm": 2.354647636413574, |
| "learning_rate": 1.905666316894019e-05, |
| "loss": 0.6878, |
| "step": 900 |
| }, |
| { |
| "epoch": 0.1049317943336831, |
| "grad_norm": 4.833475589752197, |
| "learning_rate": 1.895173137460651e-05, |
| "loss": 0.6893, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.11542497376705142, |
| "grad_norm": 2.3594470024108887, |
| "learning_rate": 1.8846799580272824e-05, |
| "loss": 0.6848, |
| "step": 1100 |
| }, |
| { |
| "epoch": 0.1259181532004197, |
| "grad_norm": 2.3646111488342285, |
| "learning_rate": 1.874186778593914e-05, |
| "loss": 0.6885, |
| "step": 1200 |
| }, |
| { |
| "epoch": 0.13641133263378805, |
| "grad_norm": 1.8924506902694702, |
| "learning_rate": 1.863693599160546e-05, |
| "loss": 0.6828, |
| "step": 1300 |
| }, |
| { |
| "epoch": 0.14690451206715635, |
| "grad_norm": 1.7312902212142944, |
| "learning_rate": 1.8532004197271777e-05, |
| "loss": 0.6885, |
| "step": 1400 |
| }, |
| { |
| "epoch": 0.15739769150052466, |
| "grad_norm": 2.434694528579712, |
| "learning_rate": 1.842707240293809e-05, |
| "loss": 0.6929, |
| "step": 1500 |
| }, |
| { |
| "epoch": 0.16789087093389296, |
| "grad_norm": 2.263343572616577, |
| "learning_rate": 1.832214060860441e-05, |
| "loss": 0.6828, |
| "step": 1600 |
| }, |
| { |
| "epoch": 0.1783840503672613, |
| "grad_norm": 1.6868016719818115, |
| "learning_rate": 1.8217208814270727e-05, |
| "loss": 0.6706, |
| "step": 1700 |
| }, |
| { |
| "epoch": 0.1888772298006296, |
| "grad_norm": 3.530144691467285, |
| "learning_rate": 1.8112277019937045e-05, |
| "loss": 0.6833, |
| "step": 1800 |
| }, |
| { |
| "epoch": 0.1993704092339979, |
| "grad_norm": 5.920458793640137, |
| "learning_rate": 1.800734522560336e-05, |
| "loss": 0.682, |
| "step": 1900 |
| }, |
| { |
| "epoch": 0.2098635886673662, |
| "grad_norm": 1.9053469896316528, |
| "learning_rate": 1.7902413431269677e-05, |
| "loss": 0.6747, |
| "step": 2000 |
| }, |
| { |
| "epoch": 0.2203567681007345, |
| "grad_norm": 1.9532723426818848, |
| "learning_rate": 1.7797481636935994e-05, |
| "loss": 0.6695, |
| "step": 2100 |
| }, |
| { |
| "epoch": 0.23084994753410285, |
| "grad_norm": 1.5969197750091553, |
| "learning_rate": 1.769254984260231e-05, |
| "loss": 0.6836, |
| "step": 2200 |
| }, |
| { |
| "epoch": 0.24134312696747115, |
| "grad_norm": 2.2583141326904297, |
| "learning_rate": 1.7587618048268627e-05, |
| "loss": 0.6748, |
| "step": 2300 |
| }, |
| { |
| "epoch": 0.2518363064008394, |
| "grad_norm": 6.070931911468506, |
| "learning_rate": 1.7482686253934944e-05, |
| "loss": 0.682, |
| "step": 2400 |
| }, |
| { |
| "epoch": 0.2623294858342078, |
| "grad_norm": 4.128929138183594, |
| "learning_rate": 1.7377754459601262e-05, |
| "loss": 0.6849, |
| "step": 2500 |
| }, |
| { |
| "epoch": 0.2728226652675761, |
| "grad_norm": 5.5497565269470215, |
| "learning_rate": 1.7272822665267576e-05, |
| "loss": 0.6805, |
| "step": 2600 |
| }, |
| { |
| "epoch": 0.2833158447009444, |
| "grad_norm": 4.123847961425781, |
| "learning_rate": 1.7167890870933894e-05, |
| "loss": 0.6876, |
| "step": 2700 |
| }, |
| { |
| "epoch": 0.2938090241343127, |
| "grad_norm": 2.535865068435669, |
| "learning_rate": 1.7062959076600212e-05, |
| "loss": 0.6794, |
| "step": 2800 |
| }, |
| { |
| "epoch": 0.304302203567681, |
| "grad_norm": 2.873096466064453, |
| "learning_rate": 1.6958027282266526e-05, |
| "loss": 0.6625, |
| "step": 2900 |
| }, |
| { |
| "epoch": 0.3147953830010493, |
| "grad_norm": 2.8748557567596436, |
| "learning_rate": 1.6853095487932844e-05, |
| "loss": 0.6791, |
| "step": 3000 |
| }, |
| { |
| "epoch": 0.3252885624344176, |
| "grad_norm": 5.24050235748291, |
| "learning_rate": 1.6748163693599162e-05, |
| "loss": 0.6848, |
| "step": 3100 |
| }, |
| { |
| "epoch": 0.3357817418677859, |
| "grad_norm": 2.885096311569214, |
| "learning_rate": 1.664323189926548e-05, |
| "loss": 0.6849, |
| "step": 3200 |
| }, |
| { |
| "epoch": 0.3462749213011542, |
| "grad_norm": 3.2616255283355713, |
| "learning_rate": 1.6538300104931794e-05, |
| "loss": 0.6751, |
| "step": 3300 |
| }, |
| { |
| "epoch": 0.3567681007345226, |
| "grad_norm": 4.990833282470703, |
| "learning_rate": 1.643336831059811e-05, |
| "loss": 0.6715, |
| "step": 3400 |
| }, |
| { |
| "epoch": 0.3672612801678909, |
| "grad_norm": 4.264138221740723, |
| "learning_rate": 1.632843651626443e-05, |
| "loss": 0.6831, |
| "step": 3500 |
| }, |
| { |
| "epoch": 0.3777544596012592, |
| "grad_norm": 3.7659502029418945, |
| "learning_rate": 1.6223504721930747e-05, |
| "loss": 0.6794, |
| "step": 3600 |
| }, |
| { |
| "epoch": 0.3882476390346275, |
| "grad_norm": 3.0774195194244385, |
| "learning_rate": 1.611857292759706e-05, |
| "loss": 0.6787, |
| "step": 3700 |
| }, |
| { |
| "epoch": 0.3987408184679958, |
| "grad_norm": 4.424105167388916, |
| "learning_rate": 1.601364113326338e-05, |
| "loss": 0.6606, |
| "step": 3800 |
| }, |
| { |
| "epoch": 0.4092339979013641, |
| "grad_norm": 5.348263263702393, |
| "learning_rate": 1.5908709338929697e-05, |
| "loss": 0.6757, |
| "step": 3900 |
| }, |
| { |
| "epoch": 0.4197271773347324, |
| "grad_norm": 13.849942207336426, |
| "learning_rate": 1.580377754459601e-05, |
| "loss": 0.683, |
| "step": 4000 |
| }, |
| { |
| "epoch": 0.4302203567681007, |
| "grad_norm": 3.790433168411255, |
| "learning_rate": 1.569884575026233e-05, |
| "loss": 0.6713, |
| "step": 4100 |
| }, |
| { |
| "epoch": 0.440713536201469, |
| "grad_norm": 3.099285840988159, |
| "learning_rate": 1.5593913955928647e-05, |
| "loss": 0.6792, |
| "step": 4200 |
| }, |
| { |
| "epoch": 0.45120671563483733, |
| "grad_norm": 6.984988689422607, |
| "learning_rate": 1.5488982161594965e-05, |
| "loss": 0.6628, |
| "step": 4300 |
| }, |
| { |
| "epoch": 0.4616998950682057, |
| "grad_norm": 2.477212905883789, |
| "learning_rate": 1.538405036726128e-05, |
| "loss": 0.6701, |
| "step": 4400 |
| }, |
| { |
| "epoch": 0.472193074501574, |
| "grad_norm": 3.1472208499908447, |
| "learning_rate": 1.5279118572927597e-05, |
| "loss": 0.6749, |
| "step": 4500 |
| }, |
| { |
| "epoch": 0.4826862539349423, |
| "grad_norm": 3.234646797180176, |
| "learning_rate": 1.5174186778593916e-05, |
| "loss": 0.6721, |
| "step": 4600 |
| }, |
| { |
| "epoch": 0.4931794333683106, |
| "grad_norm": 4.641822338104248, |
| "learning_rate": 1.5069254984260234e-05, |
| "loss": 0.6657, |
| "step": 4700 |
| }, |
| { |
| "epoch": 0.5036726128016789, |
| "grad_norm": 2.161076545715332, |
| "learning_rate": 1.4964323189926548e-05, |
| "loss": 0.6814, |
| "step": 4800 |
| }, |
| { |
| "epoch": 0.5141657922350472, |
| "grad_norm": 3.0307767391204834, |
| "learning_rate": 1.4859391395592866e-05, |
| "loss": 0.6657, |
| "step": 4900 |
| }, |
| { |
| "epoch": 0.5246589716684156, |
| "grad_norm": 4.245023727416992, |
| "learning_rate": 1.4754459601259184e-05, |
| "loss": 0.6689, |
| "step": 5000 |
| }, |
| { |
| "epoch": 0.5351521511017838, |
| "grad_norm": 1.969188928604126, |
| "learning_rate": 1.4649527806925498e-05, |
| "loss": 0.6808, |
| "step": 5100 |
| }, |
| { |
| "epoch": 0.5456453305351522, |
| "grad_norm": 2.4927730560302734, |
| "learning_rate": 1.4544596012591816e-05, |
| "loss": 0.6771, |
| "step": 5200 |
| }, |
| { |
| "epoch": 0.5561385099685204, |
| "grad_norm": 2.748563289642334, |
| "learning_rate": 1.4439664218258134e-05, |
| "loss": 0.6631, |
| "step": 5300 |
| }, |
| { |
| "epoch": 0.5666316894018888, |
| "grad_norm": 2.897603988647461, |
| "learning_rate": 1.4334732423924451e-05, |
| "loss": 0.6673, |
| "step": 5400 |
| }, |
| { |
| "epoch": 0.577124868835257, |
| "grad_norm": 4.7978434562683105, |
| "learning_rate": 1.4229800629590766e-05, |
| "loss": 0.6658, |
| "step": 5500 |
| }, |
| { |
| "epoch": 0.5876180482686254, |
| "grad_norm": 2.886988878250122, |
| "learning_rate": 1.4124868835257084e-05, |
| "loss": 0.6654, |
| "step": 5600 |
| }, |
| { |
| "epoch": 0.5981112277019937, |
| "grad_norm": 1.771572470664978, |
| "learning_rate": 1.4019937040923401e-05, |
| "loss": 0.6699, |
| "step": 5700 |
| }, |
| { |
| "epoch": 0.608604407135362, |
| "grad_norm": 5.75907564163208, |
| "learning_rate": 1.3915005246589717e-05, |
| "loss": 0.6611, |
| "step": 5800 |
| }, |
| { |
| "epoch": 0.6190975865687304, |
| "grad_norm": 2.745943307876587, |
| "learning_rate": 1.3810073452256033e-05, |
| "loss": 0.6725, |
| "step": 5900 |
| }, |
| { |
| "epoch": 0.6295907660020986, |
| "grad_norm": 5.2041096687316895, |
| "learning_rate": 1.3705141657922351e-05, |
| "loss": 0.6726, |
| "step": 6000 |
| }, |
| { |
| "epoch": 0.640083945435467, |
| "grad_norm": 2.5480427742004395, |
| "learning_rate": 1.3600209863588669e-05, |
| "loss": 0.6696, |
| "step": 6100 |
| }, |
| { |
| "epoch": 0.6505771248688352, |
| "grad_norm": 3.4199516773223877, |
| "learning_rate": 1.3495278069254985e-05, |
| "loss": 0.6661, |
| "step": 6200 |
| }, |
| { |
| "epoch": 0.6610703043022036, |
| "grad_norm": 2.6629273891448975, |
| "learning_rate": 1.3390346274921303e-05, |
| "loss": 0.662, |
| "step": 6300 |
| }, |
| { |
| "epoch": 0.6715634837355718, |
| "grad_norm": 3.0862207412719727, |
| "learning_rate": 1.3285414480587619e-05, |
| "loss": 0.6717, |
| "step": 6400 |
| }, |
| { |
| "epoch": 0.6820566631689402, |
| "grad_norm": 3.0523600578308105, |
| "learning_rate": 1.3180482686253937e-05, |
| "loss": 0.6652, |
| "step": 6500 |
| }, |
| { |
| "epoch": 0.6925498426023085, |
| "grad_norm": 2.254683494567871, |
| "learning_rate": 1.3075550891920253e-05, |
| "loss": 0.6719, |
| "step": 6600 |
| }, |
| { |
| "epoch": 0.7030430220356768, |
| "grad_norm": 2.2026445865631104, |
| "learning_rate": 1.297061909758657e-05, |
| "loss": 0.6637, |
| "step": 6700 |
| }, |
| { |
| "epoch": 0.7135362014690452, |
| "grad_norm": 1.478987455368042, |
| "learning_rate": 1.2865687303252886e-05, |
| "loss": 0.6684, |
| "step": 6800 |
| }, |
| { |
| "epoch": 0.7240293809024134, |
| "grad_norm": 5.10315465927124, |
| "learning_rate": 1.2760755508919203e-05, |
| "loss": 0.6715, |
| "step": 6900 |
| }, |
| { |
| "epoch": 0.7345225603357818, |
| "grad_norm": 3.607482433319092, |
| "learning_rate": 1.265582371458552e-05, |
| "loss": 0.6831, |
| "step": 7000 |
| }, |
| { |
| "epoch": 0.74501573976915, |
| "grad_norm": 3.3320655822753906, |
| "learning_rate": 1.2550891920251838e-05, |
| "loss": 0.6608, |
| "step": 7100 |
| }, |
| { |
| "epoch": 0.7555089192025184, |
| "grad_norm": 2.87164306640625, |
| "learning_rate": 1.2445960125918156e-05, |
| "loss": 0.6716, |
| "step": 7200 |
| }, |
| { |
| "epoch": 0.7660020986358866, |
| "grad_norm": 3.3101062774658203, |
| "learning_rate": 1.234102833158447e-05, |
| "loss": 0.6709, |
| "step": 7300 |
| }, |
| { |
| "epoch": 0.776495278069255, |
| "grad_norm": 3.6823277473449707, |
| "learning_rate": 1.2236096537250788e-05, |
| "loss": 0.6616, |
| "step": 7400 |
| }, |
| { |
| "epoch": 0.7869884575026233, |
| "grad_norm": 5.88408088684082, |
| "learning_rate": 1.2131164742917106e-05, |
| "loss": 0.6623, |
| "step": 7500 |
| }, |
| { |
| "epoch": 0.7974816369359916, |
| "grad_norm": 1.64940345287323, |
| "learning_rate": 1.2026232948583423e-05, |
| "loss": 0.6784, |
| "step": 7600 |
| }, |
| { |
| "epoch": 0.8079748163693599, |
| "grad_norm": 2.7837324142456055, |
| "learning_rate": 1.1921301154249738e-05, |
| "loss": 0.6661, |
| "step": 7700 |
| }, |
| { |
| "epoch": 0.8184679958027282, |
| "grad_norm": 2.218055248260498, |
| "learning_rate": 1.1816369359916056e-05, |
| "loss": 0.6651, |
| "step": 7800 |
| }, |
| { |
| "epoch": 0.8289611752360966, |
| "grad_norm": 2.035811185836792, |
| "learning_rate": 1.1711437565582373e-05, |
| "loss": 0.6671, |
| "step": 7900 |
| }, |
| { |
| "epoch": 0.8394543546694648, |
| "grad_norm": 2.858041524887085, |
| "learning_rate": 1.1606505771248688e-05, |
| "loss": 0.6729, |
| "step": 8000 |
| }, |
| { |
| "epoch": 0.8499475341028332, |
| "grad_norm": 3.871108293533325, |
| "learning_rate": 1.1501573976915005e-05, |
| "loss": 0.6607, |
| "step": 8100 |
| }, |
| { |
| "epoch": 0.8604407135362014, |
| "grad_norm": 2.4229085445404053, |
| "learning_rate": 1.1396642182581323e-05, |
| "loss": 0.6654, |
| "step": 8200 |
| }, |
| { |
| "epoch": 0.8709338929695698, |
| "grad_norm": 5.086295127868652, |
| "learning_rate": 1.1291710388247641e-05, |
| "loss": 0.6782, |
| "step": 8300 |
| }, |
| { |
| "epoch": 0.881427072402938, |
| "grad_norm": 2.7656631469726562, |
| "learning_rate": 1.1186778593913957e-05, |
| "loss": 0.6678, |
| "step": 8400 |
| }, |
| { |
| "epoch": 0.8919202518363064, |
| "grad_norm": 4.321835041046143, |
| "learning_rate": 1.1081846799580273e-05, |
| "loss": 0.6627, |
| "step": 8500 |
| }, |
| { |
| "epoch": 0.9024134312696747, |
| "grad_norm": 3.4984049797058105, |
| "learning_rate": 1.097691500524659e-05, |
| "loss": 0.6616, |
| "step": 8600 |
| }, |
| { |
| "epoch": 0.912906610703043, |
| "grad_norm": 2.042647123336792, |
| "learning_rate": 1.0871983210912907e-05, |
| "loss": 0.6488, |
| "step": 8700 |
| }, |
| { |
| "epoch": 0.9233997901364114, |
| "grad_norm": 2.3309738636016846, |
| "learning_rate": 1.0767051416579225e-05, |
| "loss": 0.6608, |
| "step": 8800 |
| }, |
| { |
| "epoch": 0.9338929695697796, |
| "grad_norm": 2.525007486343384, |
| "learning_rate": 1.066211962224554e-05, |
| "loss": 0.658, |
| "step": 8900 |
| }, |
| { |
| "epoch": 0.944386149003148, |
| "grad_norm": 2.533332347869873, |
| "learning_rate": 1.0557187827911858e-05, |
| "loss": 0.6626, |
| "step": 9000 |
| }, |
| { |
| "epoch": 0.9548793284365162, |
| "grad_norm": 2.3524672985076904, |
| "learning_rate": 1.0452256033578174e-05, |
| "loss": 0.6698, |
| "step": 9100 |
| }, |
| { |
| "epoch": 0.9653725078698846, |
| "grad_norm": 2.1326098442077637, |
| "learning_rate": 1.0347324239244492e-05, |
| "loss": 0.6622, |
| "step": 9200 |
| }, |
| { |
| "epoch": 0.9758656873032528, |
| "grad_norm": 2.8159704208374023, |
| "learning_rate": 1.024239244491081e-05, |
| "loss": 0.6686, |
| "step": 9300 |
| }, |
| { |
| "epoch": 0.9863588667366212, |
| "grad_norm": 3.572868824005127, |
| "learning_rate": 1.0137460650577126e-05, |
| "loss": 0.6637, |
| "step": 9400 |
| }, |
| { |
| "epoch": 0.9968520461699895, |
| "grad_norm": 2.862517833709717, |
| "learning_rate": 1.0032528856243442e-05, |
| "loss": 0.66, |
| "step": 9500 |
| }, |
| { |
| "epoch": 1.0, |
| "eval_accuracy": 0.5864637985309549, |
| "eval_f1": 0.5137569401603949, |
| "eval_loss": 0.6577488780021667, |
| "eval_precision": 0.4279767716737756, |
| "eval_recall": 0.6425430136563537, |
| "eval_runtime": 41.917, |
| "eval_samples_per_second": 909.416, |
| "eval_steps_per_second": 113.677, |
| "step": 9530 |
| }, |
| { |
| "epoch": 1.0073452256033577, |
| "grad_norm": 4.176674842834473, |
| "learning_rate": 9.92759706190976e-06, |
| "loss": 0.6298, |
| "step": 9600 |
| }, |
| { |
| "epoch": 1.0178384050367262, |
| "grad_norm": 5.948734760284424, |
| "learning_rate": 9.822665267576076e-06, |
| "loss": 0.6389, |
| "step": 9700 |
| }, |
| { |
| "epoch": 1.0283315844700944, |
| "grad_norm": 2.9435501098632812, |
| "learning_rate": 9.717733473242394e-06, |
| "loss": 0.6418, |
| "step": 9800 |
| }, |
| { |
| "epoch": 1.0388247639034627, |
| "grad_norm": 3.332951545715332, |
| "learning_rate": 9.61280167890871e-06, |
| "loss": 0.6393, |
| "step": 9900 |
| }, |
| { |
| "epoch": 1.0493179433368311, |
| "grad_norm": 3.7592711448669434, |
| "learning_rate": 9.507869884575027e-06, |
| "loss": 0.6334, |
| "step": 10000 |
| }, |
| { |
| "epoch": 1.0598111227701994, |
| "grad_norm": 4.113369464874268, |
| "learning_rate": 9.402938090241344e-06, |
| "loss": 0.6324, |
| "step": 10100 |
| }, |
| { |
| "epoch": 1.0703043022035676, |
| "grad_norm": 4.169111251831055, |
| "learning_rate": 9.298006295907661e-06, |
| "loss": 0.6195, |
| "step": 10200 |
| }, |
| { |
| "epoch": 1.080797481636936, |
| "grad_norm": 6.074437141418457, |
| "learning_rate": 9.193074501573977e-06, |
| "loss": 0.6336, |
| "step": 10300 |
| }, |
| { |
| "epoch": 1.0912906610703044, |
| "grad_norm": 5.635167121887207, |
| "learning_rate": 9.088142707240295e-06, |
| "loss": 0.6235, |
| "step": 10400 |
| }, |
| { |
| "epoch": 1.1017838405036726, |
| "grad_norm": 3.1099581718444824, |
| "learning_rate": 8.983210912906611e-06, |
| "loss": 0.6332, |
| "step": 10500 |
| }, |
| { |
| "epoch": 1.1122770199370409, |
| "grad_norm": 3.938689708709717, |
| "learning_rate": 8.878279118572927e-06, |
| "loss": 0.6337, |
| "step": 10600 |
| }, |
| { |
| "epoch": 1.1227701993704091, |
| "grad_norm": 3.1300745010375977, |
| "learning_rate": 8.773347324239245e-06, |
| "loss": 0.626, |
| "step": 10700 |
| }, |
| { |
| "epoch": 1.1332633788037776, |
| "grad_norm": 4.025566101074219, |
| "learning_rate": 8.668415529905561e-06, |
| "loss": 0.6316, |
| "step": 10800 |
| }, |
| { |
| "epoch": 1.1437565582371458, |
| "grad_norm": 2.0796873569488525, |
| "learning_rate": 8.563483735571879e-06, |
| "loss": 0.6445, |
| "step": 10900 |
| }, |
| { |
| "epoch": 1.154249737670514, |
| "grad_norm": 6.824429512023926, |
| "learning_rate": 8.458551941238195e-06, |
| "loss": 0.6291, |
| "step": 11000 |
| }, |
| { |
| "epoch": 1.1647429171038826, |
| "grad_norm": 3.644653797149658, |
| "learning_rate": 8.353620146904513e-06, |
| "loss": 0.6391, |
| "step": 11100 |
| }, |
| { |
| "epoch": 1.1752360965372508, |
| "grad_norm": 3.249530792236328, |
| "learning_rate": 8.248688352570829e-06, |
| "loss": 0.645, |
| "step": 11200 |
| }, |
| { |
| "epoch": 1.185729275970619, |
| "grad_norm": 2.6394639015197754, |
| "learning_rate": 8.143756558237146e-06, |
| "loss": 0.6523, |
| "step": 11300 |
| }, |
| { |
| "epoch": 1.1962224554039875, |
| "grad_norm": 4.585180282592773, |
| "learning_rate": 8.038824763903464e-06, |
| "loss": 0.6355, |
| "step": 11400 |
| }, |
| { |
| "epoch": 1.2067156348373558, |
| "grad_norm": 3.901735782623291, |
| "learning_rate": 7.93389296956978e-06, |
| "loss": 0.646, |
| "step": 11500 |
| }, |
| { |
| "epoch": 1.217208814270724, |
| "grad_norm": 3.5547521114349365, |
| "learning_rate": 7.828961175236098e-06, |
| "loss": 0.6289, |
| "step": 11600 |
| }, |
| { |
| "epoch": 1.2277019937040923, |
| "grad_norm": 3.6690990924835205, |
| "learning_rate": 7.724029380902414e-06, |
| "loss": 0.6365, |
| "step": 11700 |
| }, |
| { |
| "epoch": 1.2381951731374607, |
| "grad_norm": 3.4228599071502686, |
| "learning_rate": 7.619097586568731e-06, |
| "loss": 0.6313, |
| "step": 11800 |
| }, |
| { |
| "epoch": 1.248688352570829, |
| "grad_norm": 5.294908046722412, |
| "learning_rate": 7.514165792235048e-06, |
| "loss": 0.6206, |
| "step": 11900 |
| }, |
| { |
| "epoch": 1.2591815320041972, |
| "grad_norm": 4.170722007751465, |
| "learning_rate": 7.409233997901365e-06, |
| "loss": 0.6131, |
| "step": 12000 |
| }, |
| { |
| "epoch": 1.2696747114375655, |
| "grad_norm": 4.095099925994873, |
| "learning_rate": 7.304302203567682e-06, |
| "loss": 0.6353, |
| "step": 12100 |
| }, |
| { |
| "epoch": 1.280167890870934, |
| "grad_norm": 3.773933172225952, |
| "learning_rate": 7.1993704092339985e-06, |
| "loss": 0.6431, |
| "step": 12200 |
| }, |
| { |
| "epoch": 1.2906610703043022, |
| "grad_norm": 4.165128707885742, |
| "learning_rate": 7.0944386149003154e-06, |
| "loss": 0.6279, |
| "step": 12300 |
| }, |
| { |
| "epoch": 1.3011542497376705, |
| "grad_norm": 5.304838180541992, |
| "learning_rate": 6.989506820566632e-06, |
| "loss": 0.6168, |
| "step": 12400 |
| }, |
| { |
| "epoch": 1.311647429171039, |
| "grad_norm": 4.008569717407227, |
| "learning_rate": 6.884575026232949e-06, |
| "loss": 0.6162, |
| "step": 12500 |
| }, |
| { |
| "epoch": 1.3221406086044072, |
| "grad_norm": 7.424976348876953, |
| "learning_rate": 6.779643231899265e-06, |
| "loss": 0.6382, |
| "step": 12600 |
| }, |
| { |
| "epoch": 1.3326337880377754, |
| "grad_norm": 3.9119601249694824, |
| "learning_rate": 6.674711437565583e-06, |
| "loss": 0.6359, |
| "step": 12700 |
| }, |
| { |
| "epoch": 1.3431269674711437, |
| "grad_norm": 2.6989879608154297, |
| "learning_rate": 6.569779643231899e-06, |
| "loss": 0.6365, |
| "step": 12800 |
| }, |
| { |
| "epoch": 1.353620146904512, |
| "grad_norm": 3.548513650894165, |
| "learning_rate": 6.464847848898217e-06, |
| "loss": 0.6291, |
| "step": 12900 |
| }, |
| { |
| "epoch": 1.3641133263378804, |
| "grad_norm": 4.189005374908447, |
| "learning_rate": 6.359916054564533e-06, |
| "loss": 0.6171, |
| "step": 13000 |
| }, |
| { |
| "epoch": 1.3746065057712487, |
| "grad_norm": 4.647156715393066, |
| "learning_rate": 6.254984260230851e-06, |
| "loss": 0.6121, |
| "step": 13100 |
| }, |
| { |
| "epoch": 1.385099685204617, |
| "grad_norm": 6.423859119415283, |
| "learning_rate": 6.150052465897168e-06, |
| "loss": 0.6229, |
| "step": 13200 |
| }, |
| { |
| "epoch": 1.3955928646379854, |
| "grad_norm": 3.7313272953033447, |
| "learning_rate": 6.0451206715634845e-06, |
| "loss": 0.6193, |
| "step": 13300 |
| }, |
| { |
| "epoch": 1.4060860440713536, |
| "grad_norm": 4.396394729614258, |
| "learning_rate": 5.940188877229801e-06, |
| "loss": 0.6303, |
| "step": 13400 |
| }, |
| { |
| "epoch": 1.4165792235047219, |
| "grad_norm": 4.366927146911621, |
| "learning_rate": 5.8352570828961175e-06, |
| "loss": 0.6349, |
| "step": 13500 |
| }, |
| { |
| "epoch": 1.4270724029380903, |
| "grad_norm": 3.679366111755371, |
| "learning_rate": 5.730325288562435e-06, |
| "loss": 0.6188, |
| "step": 13600 |
| }, |
| { |
| "epoch": 1.4375655823714586, |
| "grad_norm": 6.429540157318115, |
| "learning_rate": 5.625393494228751e-06, |
| "loss": 0.6161, |
| "step": 13700 |
| }, |
| { |
| "epoch": 1.4480587618048268, |
| "grad_norm": 4.448658466339111, |
| "learning_rate": 5.520461699895069e-06, |
| "loss": 0.6179, |
| "step": 13800 |
| }, |
| { |
| "epoch": 1.458551941238195, |
| "grad_norm": 5.556158542633057, |
| "learning_rate": 5.415529905561385e-06, |
| "loss": 0.6176, |
| "step": 13900 |
| }, |
| { |
| "epoch": 1.4690451206715636, |
| "grad_norm": 6.288188457489014, |
| "learning_rate": 5.310598111227703e-06, |
| "loss": 0.6243, |
| "step": 14000 |
| }, |
| { |
| "epoch": 1.4795383001049318, |
| "grad_norm": 4.509946823120117, |
| "learning_rate": 5.205666316894019e-06, |
| "loss": 0.6395, |
| "step": 14100 |
| }, |
| { |
| "epoch": 1.4900314795383, |
| "grad_norm": 2.5123424530029297, |
| "learning_rate": 5.100734522560337e-06, |
| "loss": 0.642, |
| "step": 14200 |
| }, |
| { |
| "epoch": 1.5005246589716683, |
| "grad_norm": 5.663776397705078, |
| "learning_rate": 4.995802728226653e-06, |
| "loss": 0.6174, |
| "step": 14300 |
| }, |
| { |
| "epoch": 1.5110178384050368, |
| "grad_norm": 3.452349901199341, |
| "learning_rate": 4.89087093389297e-06, |
| "loss": 0.6026, |
| "step": 14400 |
| }, |
| { |
| "epoch": 1.521511017838405, |
| "grad_norm": 3.6340229511260986, |
| "learning_rate": 4.7859391395592865e-06, |
| "loss": 0.6163, |
| "step": 14500 |
| }, |
| { |
| "epoch": 1.5320041972717733, |
| "grad_norm": 3.506894588470459, |
| "learning_rate": 4.6810073452256034e-06, |
| "loss": 0.6287, |
| "step": 14600 |
| }, |
| { |
| "epoch": 1.5424973767051418, |
| "grad_norm": 4.924068450927734, |
| "learning_rate": 4.576075550891921e-06, |
| "loss": 0.6254, |
| "step": 14700 |
| }, |
| { |
| "epoch": 1.55299055613851, |
| "grad_norm": 6.727036476135254, |
| "learning_rate": 4.471143756558238e-06, |
| "loss": 0.6187, |
| "step": 14800 |
| }, |
| { |
| "epoch": 1.5634837355718783, |
| "grad_norm": 4.4192399978637695, |
| "learning_rate": 4.366211962224554e-06, |
| "loss": 0.6308, |
| "step": 14900 |
| }, |
| { |
| "epoch": 1.5739769150052467, |
| "grad_norm": 3.18361496925354, |
| "learning_rate": 4.261280167890871e-06, |
| "loss": 0.6423, |
| "step": 15000 |
| }, |
| { |
| "epoch": 1.5844700944386148, |
| "grad_norm": 3.8575243949890137, |
| "learning_rate": 4.156348373557188e-06, |
| "loss": 0.6202, |
| "step": 15100 |
| }, |
| { |
| "epoch": 1.5949632738719832, |
| "grad_norm": 5.774935722351074, |
| "learning_rate": 4.051416579223505e-06, |
| "loss": 0.6226, |
| "step": 15200 |
| }, |
| { |
| "epoch": 1.6054564533053515, |
| "grad_norm": 5.742249488830566, |
| "learning_rate": 3.946484784889822e-06, |
| "loss": 0.6291, |
| "step": 15300 |
| }, |
| { |
| "epoch": 1.6159496327387197, |
| "grad_norm": 4.477142810821533, |
| "learning_rate": 3.841552990556139e-06, |
| "loss": 0.627, |
| "step": 15400 |
| }, |
| { |
| "epoch": 1.6264428121720882, |
| "grad_norm": 4.507772922515869, |
| "learning_rate": 3.7366211962224556e-06, |
| "loss": 0.6309, |
| "step": 15500 |
| }, |
| { |
| "epoch": 1.6369359916054564, |
| "grad_norm": 4.460930347442627, |
| "learning_rate": 3.631689401888773e-06, |
| "loss": 0.6242, |
| "step": 15600 |
| }, |
| { |
| "epoch": 1.6474291710388247, |
| "grad_norm": 4.099055290222168, |
| "learning_rate": 3.52675760755509e-06, |
| "loss": 0.6242, |
| "step": 15700 |
| }, |
| { |
| "epoch": 1.6579223504721932, |
| "grad_norm": 5.501063346862793, |
| "learning_rate": 3.4218258132214067e-06, |
| "loss": 0.6207, |
| "step": 15800 |
| }, |
| { |
| "epoch": 1.6684155299055614, |
| "grad_norm": 5.495902061462402, |
| "learning_rate": 3.316894018887723e-06, |
| "loss": 0.6197, |
| "step": 15900 |
| }, |
| { |
| "epoch": 1.6789087093389297, |
| "grad_norm": 5.21685266494751, |
| "learning_rate": 3.21196222455404e-06, |
| "loss": 0.6275, |
| "step": 16000 |
| }, |
| { |
| "epoch": 1.6894018887722981, |
| "grad_norm": 3.404468536376953, |
| "learning_rate": 3.107030430220357e-06, |
| "loss": 0.6333, |
| "step": 16100 |
| }, |
| { |
| "epoch": 1.6998950682056662, |
| "grad_norm": 5.716117858886719, |
| "learning_rate": 3.002098635886674e-06, |
| "loss": 0.6244, |
| "step": 16200 |
| }, |
| { |
| "epoch": 1.7103882476390346, |
| "grad_norm": 4.773796081542969, |
| "learning_rate": 2.897166841552991e-06, |
| "loss": 0.6203, |
| "step": 16300 |
| }, |
| { |
| "epoch": 1.720881427072403, |
| "grad_norm": 4.482935905456543, |
| "learning_rate": 2.7922350472193077e-06, |
| "loss": 0.6143, |
| "step": 16400 |
| }, |
| { |
| "epoch": 1.7313746065057711, |
| "grad_norm": 4.8394341468811035, |
| "learning_rate": 2.6873032528856246e-06, |
| "loss": 0.6325, |
| "step": 16500 |
| }, |
| { |
| "epoch": 1.7418677859391396, |
| "grad_norm": 5.177644729614258, |
| "learning_rate": 2.5823714585519415e-06, |
| "loss": 0.6093, |
| "step": 16600 |
| }, |
| { |
| "epoch": 1.7523609653725079, |
| "grad_norm": 5.235119342803955, |
| "learning_rate": 2.4774396642182585e-06, |
| "loss": 0.6218, |
| "step": 16700 |
| }, |
| { |
| "epoch": 1.762854144805876, |
| "grad_norm": 4.899617671966553, |
| "learning_rate": 2.3725078698845754e-06, |
| "loss": 0.6102, |
| "step": 16800 |
| }, |
| { |
| "epoch": 1.7733473242392446, |
| "grad_norm": 6.910520076751709, |
| "learning_rate": 2.2675760755508923e-06, |
| "loss": 0.6138, |
| "step": 16900 |
| }, |
| { |
| "epoch": 1.7838405036726128, |
| "grad_norm": 4.011341094970703, |
| "learning_rate": 2.162644281217209e-06, |
| "loss": 0.6367, |
| "step": 17000 |
| }, |
| { |
| "epoch": 1.794333683105981, |
| "grad_norm": 4.2056427001953125, |
| "learning_rate": 2.0577124868835257e-06, |
| "loss": 0.6297, |
| "step": 17100 |
| }, |
| { |
| "epoch": 1.8048268625393495, |
| "grad_norm": 5.336106777191162, |
| "learning_rate": 1.9527806925498426e-06, |
| "loss": 0.6315, |
| "step": 17200 |
| }, |
| { |
| "epoch": 1.8153200419727176, |
| "grad_norm": 5.930999279022217, |
| "learning_rate": 1.8478488982161597e-06, |
| "loss": 0.6202, |
| "step": 17300 |
| }, |
| { |
| "epoch": 1.825813221406086, |
| "grad_norm": 3.3669204711914062, |
| "learning_rate": 1.7429171038824766e-06, |
| "loss": 0.6275, |
| "step": 17400 |
| }, |
| { |
| "epoch": 1.8363064008394545, |
| "grad_norm": 3.4519994258880615, |
| "learning_rate": 1.6379853095487935e-06, |
| "loss": 0.6257, |
| "step": 17500 |
| }, |
| { |
| "epoch": 1.8467995802728225, |
| "grad_norm": 4.662219524383545, |
| "learning_rate": 1.5330535152151102e-06, |
| "loss": 0.599, |
| "step": 17600 |
| }, |
| { |
| "epoch": 1.857292759706191, |
| "grad_norm": 5.023181438446045, |
| "learning_rate": 1.428121720881427e-06, |
| "loss": 0.6366, |
| "step": 17700 |
| }, |
| { |
| "epoch": 1.8677859391395593, |
| "grad_norm": 7.75839900970459, |
| "learning_rate": 1.323189926547744e-06, |
| "loss": 0.6181, |
| "step": 17800 |
| }, |
| { |
| "epoch": 1.8782791185729275, |
| "grad_norm": 3.3768789768218994, |
| "learning_rate": 1.2182581322140611e-06, |
| "loss": 0.6293, |
| "step": 17900 |
| }, |
| { |
| "epoch": 1.888772298006296, |
| "grad_norm": 4.408294200897217, |
| "learning_rate": 1.1133263378803778e-06, |
| "loss": 0.6053, |
| "step": 18000 |
| }, |
| { |
| "epoch": 1.8992654774396642, |
| "grad_norm": 5.284770488739014, |
| "learning_rate": 1.0083945435466947e-06, |
| "loss": 0.6257, |
| "step": 18100 |
| }, |
| { |
| "epoch": 1.9097586568730325, |
| "grad_norm": 3.856238842010498, |
| "learning_rate": 9.034627492130116e-07, |
| "loss": 0.6239, |
| "step": 18200 |
| }, |
| { |
| "epoch": 1.920251836306401, |
| "grad_norm": 3.986468553543091, |
| "learning_rate": 7.985309548793285e-07, |
| "loss": 0.6058, |
| "step": 18300 |
| }, |
| { |
| "epoch": 1.9307450157397692, |
| "grad_norm": 4.508366584777832, |
| "learning_rate": 6.935991605456453e-07, |
| "loss": 0.6168, |
| "step": 18400 |
| }, |
| { |
| "epoch": 1.9412381951731374, |
| "grad_norm": 3.7954189777374268, |
| "learning_rate": 5.886673662119622e-07, |
| "loss": 0.6143, |
| "step": 18500 |
| }, |
| { |
| "epoch": 1.951731374606506, |
| "grad_norm": 4.71843957901001, |
| "learning_rate": 4.837355718782791e-07, |
| "loss": 0.6153, |
| "step": 18600 |
| }, |
| { |
| "epoch": 1.962224554039874, |
| "grad_norm": 5.512548446655273, |
| "learning_rate": 3.7880377754459604e-07, |
| "loss": 0.6101, |
| "step": 18700 |
| }, |
| { |
| "epoch": 1.9727177334732424, |
| "grad_norm": 4.672413349151611, |
| "learning_rate": 2.7387198321091295e-07, |
| "loss": 0.6149, |
| "step": 18800 |
| }, |
| { |
| "epoch": 1.9832109129066107, |
| "grad_norm": 3.992551326751709, |
| "learning_rate": 1.689401888772298e-07, |
| "loss": 0.6431, |
| "step": 18900 |
| }, |
| { |
| "epoch": 1.993704092339979, |
| "grad_norm": 3.957615375518799, |
| "learning_rate": 6.40083945435467e-08, |
| "loss": 0.6107, |
| "step": 19000 |
| }, |
| { |
| "epoch": 2.0, |
| "eval_accuracy": 0.6031217208814271, |
| "eval_f1": 0.515732530968919, |
| "eval_loss": 0.6643054485321045, |
| "eval_precision": 0.44070021881838073, |
| "eval_recall": 0.6215569786281923, |
| "eval_runtime": 42.0167, |
| "eval_samples_per_second": 907.258, |
| "eval_steps_per_second": 113.407, |
| "step": 19060 |
| } |
| ], |
| "logging_steps": 100, |
| "max_steps": 19060, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 2, |
| "save_steps": 500, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 1.00297934303232e+16, |
| "train_batch_size": 16, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|