| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 4.864, |
| "eval_steps": 500, |
| "global_step": 155, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.16, |
| "grad_norm": 0.04806969687342644, |
| "learning_rate": 6.249999999999999e-07, |
| "loss": 1.1372, |
| "mean_token_accuracy": 0.7175553351640701, |
| "step": 5 |
| }, |
| { |
| "epoch": 0.32, |
| "grad_norm": 0.04902997612953186, |
| "learning_rate": 9.99543333708549e-07, |
| "loss": 1.1327, |
| "mean_token_accuracy": 0.7127479765564203, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.48, |
| "grad_norm": 0.052965614944696426, |
| "learning_rate": 9.944154131125642e-07, |
| "loss": 1.1403, |
| "mean_token_accuracy": 0.7101998340338469, |
| "step": 15 |
| }, |
| { |
| "epoch": 0.64, |
| "grad_norm": 0.049492210149765015, |
| "learning_rate": 9.836474315195147e-07, |
| "loss": 1.1492, |
| "mean_token_accuracy": 0.7093485150486231, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.8, |
| "grad_norm": 0.04770048335194588, |
| "learning_rate": 9.673622250534155e-07, |
| "loss": 1.0896, |
| "mean_token_accuracy": 0.722656001150608, |
| "step": 25 |
| }, |
| { |
| "epoch": 0.96, |
| "grad_norm": 0.04960792139172554, |
| "learning_rate": 9.457455677726447e-07, |
| "loss": 1.1031, |
| "mean_token_accuracy": 0.7249287318438291, |
| "step": 30 |
| }, |
| { |
| "epoch": 1.096, |
| "grad_norm": 0.05211903899908066, |
| "learning_rate": 9.190440524459202e-07, |
| "loss": 1.0969, |
| "mean_token_accuracy": 0.7173009343883571, |
| "step": 35 |
| }, |
| { |
| "epoch": 1.256, |
| "grad_norm": 0.05225418508052826, |
| "learning_rate": 8.875622775367259e-07, |
| "loss": 1.1024, |
| "mean_token_accuracy": 0.7184804223477841, |
| "step": 40 |
| }, |
| { |
| "epoch": 1.416, |
| "grad_norm": 0.053441036492586136, |
| "learning_rate": 8.516593724857597e-07, |
| "loss": 1.1062, |
| "mean_token_accuracy": 0.7180147383362054, |
| "step": 45 |
| }, |
| { |
| "epoch": 1.576, |
| "grad_norm": 0.05642622709274292, |
| "learning_rate": 8.117449009293668e-07, |
| "loss": 1.1143, |
| "mean_token_accuracy": 0.7181682541966439, |
| "step": 50 |
| }, |
| { |
| "epoch": 1.736, |
| "grad_norm": 0.06203881651163101, |
| "learning_rate": 7.682741885881314e-07, |
| "loss": 1.1619, |
| "mean_token_accuracy": 0.7082567475736141, |
| "step": 55 |
| }, |
| { |
| "epoch": 1.896, |
| "grad_norm": 0.05656283721327782, |
| "learning_rate": 7.217431291229067e-07, |
| "loss": 1.1153, |
| "mean_token_accuracy": 0.722714689001441, |
| "step": 60 |
| }, |
| { |
| "epoch": 2.032, |
| "grad_norm": 0.061042506247758865, |
| "learning_rate": 6.726825272106538e-07, |
| "loss": 1.1493, |
| "mean_token_accuracy": 0.7165276228505022, |
| "step": 65 |
| }, |
| { |
| "epoch": 2.192, |
| "grad_norm": 0.058021657168865204, |
| "learning_rate": 6.216520433716544e-07, |
| "loss": 1.1358, |
| "mean_token_accuracy": 0.7132049109786749, |
| "step": 70 |
| }, |
| { |
| "epoch": 2.352, |
| "grad_norm": 0.059838637709617615, |
| "learning_rate": 5.69233809622687e-07, |
| "loss": 1.1148, |
| "mean_token_accuracy": 0.7178624272346497, |
| "step": 75 |
| }, |
| { |
| "epoch": 2.512, |
| "grad_norm": 0.06404729187488556, |
| "learning_rate": 5.160257887858277e-07, |
| "loss": 1.1569, |
| "mean_token_accuracy": 0.7101449474692345, |
| "step": 80 |
| }, |
| { |
| "epoch": 2.672, |
| "grad_norm": 0.06028318777680397, |
| "learning_rate": 4.626349532067879e-07, |
| "loss": 1.0867, |
| "mean_token_accuracy": 0.7239396564662457, |
| "step": 85 |
| }, |
| { |
| "epoch": 2.832, |
| "grad_norm": 0.06384854018688202, |
| "learning_rate": 4.096703606968006e-07, |
| "loss": 1.1188, |
| "mean_token_accuracy": 0.7177162211388349, |
| "step": 90 |
| }, |
| { |
| "epoch": 2.992, |
| "grad_norm": 0.06171684339642525, |
| "learning_rate": 3.577362066844838e-07, |
| "loss": 1.1188, |
| "mean_token_accuracy": 0.7161729197949172, |
| "step": 95 |
| }, |
| { |
| "epoch": 3.128, |
| "grad_norm": 0.0635930597782135, |
| "learning_rate": 3.0742493183550454e-07, |
| "loss": 1.0999, |
| "mean_token_accuracy": 0.714051765992361, |
| "step": 100 |
| }, |
| { |
| "epoch": 3.288, |
| "grad_norm": 0.06290465593338013, |
| "learning_rate": 2.593104637651087e-07, |
| "loss": 1.138, |
| "mean_token_accuracy": 0.7161103874444962, |
| "step": 105 |
| }, |
| { |
| "epoch": 3.448, |
| "grad_norm": 0.06601794064044952, |
| "learning_rate": 2.1394166993891526e-07, |
| "loss": 1.1073, |
| "mean_token_accuracy": 0.7206942658871412, |
| "step": 110 |
| }, |
| { |
| "epoch": 3.608, |
| "grad_norm": 0.06451869755983353, |
| "learning_rate": 1.7183609644824092e-07, |
| "loss": 1.105, |
| "mean_token_accuracy": 0.7173311490565538, |
| "step": 115 |
| }, |
| { |
| "epoch": 3.768, |
| "grad_norm": 0.06503637880086899, |
| "learning_rate": 1.3347406408508694e-07, |
| "loss": 1.1026, |
| "mean_token_accuracy": 0.7226938724517822, |
| "step": 120 |
| }, |
| { |
| "epoch": 3.928, |
| "grad_norm": 0.06369519233703613, |
| "learning_rate": 9.929318906602174e-08, |
| "loss": 1.136, |
| "mean_token_accuracy": 0.7119275834411383, |
| "step": 125 |
| }, |
| { |
| "epoch": 4.064, |
| "grad_norm": 0.0652015432715416, |
| "learning_rate": 6.968339090999186e-08, |
| "loss": 1.1041, |
| "mean_token_accuracy": 0.7163285718244665, |
| "step": 130 |
| }, |
| { |
| "epoch": 4.224, |
| "grad_norm": 0.06359368562698364, |
| "learning_rate": 4.498244441786675e-08, |
| "loss": 1.1359, |
| "mean_token_accuracy": 0.7156530544161797, |
| "step": 135 |
| }, |
| { |
| "epoch": 4.384, |
| "grad_norm": 0.06807275116443634, |
| "learning_rate": 2.547212649466568e-08, |
| "loss": 1.1229, |
| "mean_token_accuracy": 0.7130052808672189, |
| "step": 140 |
| }, |
| { |
| "epoch": 4.5440000000000005, |
| "grad_norm": 0.06558381021022797, |
| "learning_rate": 1.1375001769727999e-08, |
| "loss": 1.0983, |
| "mean_token_accuracy": 0.7218218572437763, |
| "step": 145 |
| }, |
| { |
| "epoch": 4.704, |
| "grad_norm": 0.06619936972856522, |
| "learning_rate": 2.851883682973233e-09, |
| "loss": 1.1149, |
| "mean_token_accuracy": 0.7202201712876558, |
| "step": 150 |
| }, |
| { |
| "epoch": 4.864, |
| "grad_norm": 0.0672445297241211, |
| "learning_rate": 0.0, |
| "loss": 1.0885, |
| "mean_token_accuracy": 0.7229270905256271, |
| "step": 155 |
| } |
| ], |
| "logging_steps": 5, |
| "max_steps": 155, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 5, |
| "save_steps": 500, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 9.40617510529582e+18, |
| "train_batch_size": 1, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|