| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 1.905829596412556, |
| "eval_steps": 500, |
| "global_step": 1700, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.11210762331838565, |
| "grad_norm": 2.0963618755340576, |
| "learning_rate": 2.793296089385475e-05, |
| "loss": 4.9671, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.2242152466367713, |
| "grad_norm": 0.3255109190940857, |
| "learning_rate": 4.9408099688473526e-05, |
| "loss": 4.7197, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.336322869955157, |
| "grad_norm": 1.1150901317596436, |
| "learning_rate": 4.632398753894081e-05, |
| "loss": 4.5727, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.4484304932735426, |
| "grad_norm": 4.738381862640381, |
| "learning_rate": 4.320872274143302e-05, |
| "loss": 4.5585, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.5605381165919282, |
| "grad_norm": 0.3568013906478882, |
| "learning_rate": 4.0093457943925236e-05, |
| "loss": 4.5399, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.672645739910314, |
| "grad_norm": 0.558647871017456, |
| "learning_rate": 3.700934579439253e-05, |
| "loss": 4.4905, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.7847533632286996, |
| "grad_norm": 19.64227867126465, |
| "learning_rate": 3.3894080996884734e-05, |
| "loss": 4.4371, |
| "step": 700 |
| }, |
| { |
| "epoch": 0.8968609865470852, |
| "grad_norm": 2.4077465534210205, |
| "learning_rate": 3.077881619937695e-05, |
| "loss": 4.4867, |
| "step": 800 |
| }, |
| { |
| "epoch": 1.0089686098654709, |
| "grad_norm": 1.3140878677368164, |
| "learning_rate": 2.7663551401869157e-05, |
| "loss": 4.4675, |
| "step": 900 |
| }, |
| { |
| "epoch": 1.1210762331838564, |
| "grad_norm": 1.7104419469833374, |
| "learning_rate": 2.4579439252336452e-05, |
| "loss": 4.432, |
| "step": 1000 |
| }, |
| { |
| "epoch": 1.2331838565022422, |
| "grad_norm": 7.763595104217529, |
| "learning_rate": 2.1464174454828662e-05, |
| "loss": 4.4185, |
| "step": 1100 |
| }, |
| { |
| "epoch": 1.3452914798206277, |
| "grad_norm": 0.4148617088794708, |
| "learning_rate": 1.8348909657320872e-05, |
| "loss": 4.428, |
| "step": 1200 |
| }, |
| { |
| "epoch": 1.4573991031390134, |
| "grad_norm": 2.4640185832977295, |
| "learning_rate": 1.5233644859813085e-05, |
| "loss": 4.4133, |
| "step": 1300 |
| }, |
| { |
| "epoch": 1.5695067264573992, |
| "grad_norm": 0.4486638903617859, |
| "learning_rate": 1.2118380062305297e-05, |
| "loss": 4.3019, |
| "step": 1400 |
| }, |
| { |
| "epoch": 1.6816143497757847, |
| "grad_norm": 191.92808532714844, |
| "learning_rate": 9.003115264797508e-06, |
| "loss": 4.4209, |
| "step": 1500 |
| }, |
| { |
| "epoch": 1.7937219730941703, |
| "grad_norm": 2.8736190795898438, |
| "learning_rate": 5.88785046728972e-06, |
| "loss": 4.3696, |
| "step": 1600 |
| }, |
| { |
| "epoch": 1.905829596412556, |
| "grad_norm": 0.13583791255950928, |
| "learning_rate": 2.7725856697819316e-06, |
| "loss": 4.3962, |
| "step": 1700 |
| } |
| ], |
| "logging_steps": 100, |
| "max_steps": 1784, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 2, |
| "save_steps": 100, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": false |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 0.0, |
| "train_batch_size": 16, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|