{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 3.0, "eval_steps": 500, "global_step": 1773, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.08460236886632826, "grad_norm": 4.897884368896484, "learning_rate": 1.9447264523406657e-05, "loss": 2.026262054443359, "step": 50 }, { "epoch": 0.1692047377326565, "grad_norm": 9.024230003356934, "learning_rate": 1.888324873096447e-05, "loss": 1.0631247711181642, "step": 100 }, { "epoch": 0.25380710659898476, "grad_norm": 8.573076248168945, "learning_rate": 1.831923293852228e-05, "loss": 0.8047801971435546, "step": 150 }, { "epoch": 0.338409475465313, "grad_norm": 9.223526954650879, "learning_rate": 1.7755217146080094e-05, "loss": 0.6834676361083984, "step": 200 }, { "epoch": 0.4230118443316413, "grad_norm": 6.011892795562744, "learning_rate": 1.7191201353637904e-05, "loss": 0.5958165359497071, "step": 250 }, { "epoch": 0.5076142131979695, "grad_norm": 7.655128002166748, "learning_rate": 1.6627185561195714e-05, "loss": 0.6095886993408203, "step": 300 }, { "epoch": 0.5922165820642978, "grad_norm": 6.9969353675842285, "learning_rate": 1.6063169768753528e-05, "loss": 0.5533372116088867, "step": 350 }, { "epoch": 0.676818950930626, "grad_norm": 9.313193321228027, "learning_rate": 1.5499153976311338e-05, "loss": 0.5110644149780273, "step": 400 }, { "epoch": 0.7614213197969543, "grad_norm": 12.71023941040039, "learning_rate": 1.493513818386915e-05, "loss": 0.5459028244018554, "step": 450 }, { "epoch": 0.8460236886632826, "grad_norm": 8.374627113342285, "learning_rate": 1.4371122391426961e-05, "loss": 0.46964054107666015, "step": 500 }, { "epoch": 0.9306260575296108, "grad_norm": 9.238364219665527, "learning_rate": 1.3807106598984771e-05, "loss": 0.5043766021728515, "step": 550 }, { "epoch": 1.0, "eval_f1": 0.7749147704749069, "eval_loss": 0.43065306544303894, "eval_precision": 0.791417004048583, "eval_recall": 0.7590866728797764, "eval_runtime": 107.5887, "eval_samples_per_second": 2.444, "eval_steps_per_second": 0.613, "step": 591 }, { "epoch": 1.015228426395939, "grad_norm": 5.481167793273926, "learning_rate": 1.3243090806542585e-05, "loss": 0.4575632858276367, "step": 600 }, { "epoch": 1.0998307952622675, "grad_norm": 6.10057258605957, "learning_rate": 1.2679075014100395e-05, "loss": 0.43560176849365234, "step": 650 }, { "epoch": 1.1844331641285957, "grad_norm": 9.03114128112793, "learning_rate": 1.2115059221658207e-05, "loss": 0.46458480834960936, "step": 700 }, { "epoch": 1.2690355329949239, "grad_norm": 9.81558895111084, "learning_rate": 1.155104342921602e-05, "loss": 0.40803119659423825, "step": 750 }, { "epoch": 1.353637901861252, "grad_norm": 15.4000825881958, "learning_rate": 1.098702763677383e-05, "loss": 0.42147384643554686, "step": 800 }, { "epoch": 1.4382402707275803, "grad_norm": 10.52159595489502, "learning_rate": 1.0423011844331642e-05, "loss": 0.4100672149658203, "step": 850 }, { "epoch": 1.5228426395939088, "grad_norm": 15.26666259765625, "learning_rate": 9.858996051889454e-06, "loss": 0.3965869903564453, "step": 900 }, { "epoch": 1.6074450084602367, "grad_norm": 9.637870788574219, "learning_rate": 9.294980259447264e-06, "loss": 0.3966701126098633, "step": 950 }, { "epoch": 1.6920473773265652, "grad_norm": 4.112008571624756, "learning_rate": 8.730964467005076e-06, "loss": 0.40687416076660154, "step": 1000 }, { "epoch": 1.7766497461928934, "grad_norm": 11.278714179992676, "learning_rate": 8.16694867456289e-06, "loss": 0.41602436065673826, "step": 1050 }, { "epoch": 1.8612521150592216, "grad_norm": 8.418536186218262, "learning_rate": 7.602932882120701e-06, "loss": 0.40810283660888674, "step": 1100 }, { "epoch": 1.94585448392555, "grad_norm": 4.924833297729492, "learning_rate": 7.038917089678512e-06, "loss": 0.3806829833984375, "step": 1150 }, { "epoch": 2.0, "eval_f1": 0.8019422037747671, "eval_loss": 0.3661089539527893, "eval_precision": 0.8087190017374822, "eval_recall": 0.795278036657347, "eval_runtime": 105.196, "eval_samples_per_second": 2.5, "eval_steps_per_second": 0.627, "step": 1182 }, { "epoch": 2.030456852791878, "grad_norm": 10.482388496398926, "learning_rate": 6.4749012972363234e-06, "loss": 0.36735050201416014, "step": 1200 }, { "epoch": 2.1150592216582065, "grad_norm": 6.7676496505737305, "learning_rate": 5.910885504794135e-06, "loss": 0.3464337158203125, "step": 1250 }, { "epoch": 2.199661590524535, "grad_norm": 8.209813117980957, "learning_rate": 5.346869712351946e-06, "loss": 0.34701999664306643, "step": 1300 }, { "epoch": 2.284263959390863, "grad_norm": 9.558784484863281, "learning_rate": 4.782853919909758e-06, "loss": 0.3395441818237305, "step": 1350 }, { "epoch": 2.3688663282571913, "grad_norm": 6.974896430969238, "learning_rate": 4.218838127467569e-06, "loss": 0.3871475601196289, "step": 1400 }, { "epoch": 2.4534686971235193, "grad_norm": 9.049581527709961, "learning_rate": 3.654822335025381e-06, "loss": 0.3264299774169922, "step": 1450 }, { "epoch": 2.5380710659898478, "grad_norm": 16.217670440673828, "learning_rate": 3.0908065425831923e-06, "loss": 0.3252403259277344, "step": 1500 }, { "epoch": 2.6226734348561758, "grad_norm": 7.654144763946533, "learning_rate": 2.5267907501410045e-06, "loss": 0.3354499816894531, "step": 1550 }, { "epoch": 2.707275803722504, "grad_norm": 8.84153938293457, "learning_rate": 1.9627749576988155e-06, "loss": 0.31287595748901365, "step": 1600 }, { "epoch": 2.7918781725888326, "grad_norm": 8.133597373962402, "learning_rate": 1.3987591652566273e-06, "loss": 0.33677963256835936, "step": 1650 }, { "epoch": 2.8764805414551606, "grad_norm": 5.950267791748047, "learning_rate": 8.347433728144388e-07, "loss": 0.32882442474365237, "step": 1700 }, { "epoch": 2.961082910321489, "grad_norm": 9.175284385681152, "learning_rate": 2.7072758037225044e-07, "loss": 0.3340682220458984, "step": 1750 }, { "epoch": 3.0, "eval_f1": 0.8073293488518403, "eval_loss": 0.3556199371814728, "eval_precision": 0.8176170755017521, "eval_recall": 0.7972972972972973, "eval_runtime": 105.9353, "eval_samples_per_second": 2.483, "eval_steps_per_second": 0.623, "step": 1773 } ], "logging_steps": 50, "max_steps": 1773, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 463350353826816.0, "train_batch_size": 4, "trial_name": null, "trial_params": null }