| { |
| "best_metric": 0.636893310208111, |
| "best_model_checkpoint": "./save_models/cola/roberta-base_lr1e-05_run0/checkpoint-3848", |
| "epoch": 10.0, |
| "eval_steps": 500, |
| "global_step": 4810, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 1.0, |
| "eval_loss": 0.45413094758987427, |
| "eval_matthews_correlation": 0.47785514722622213, |
| "eval_runtime": 0.5981, |
| "eval_samples_per_second": 1431.104, |
| "eval_steps_per_second": 90.28, |
| "step": 481 |
| }, |
| { |
| "epoch": 1.04, |
| "learning_rate": 9.533289095332891e-06, |
| "loss": 0.5451, |
| "step": 500 |
| }, |
| { |
| "epoch": 2.0, |
| "eval_loss": 0.4294413626194, |
| "eval_matthews_correlation": 0.5559024469662472, |
| "eval_runtime": 0.6646, |
| "eval_samples_per_second": 1288.013, |
| "eval_steps_per_second": 81.253, |
| "step": 962 |
| }, |
| { |
| "epoch": 2.08, |
| "learning_rate": 8.427339084273391e-06, |
| "loss": 0.3643, |
| "step": 1000 |
| }, |
| { |
| "epoch": 3.0, |
| "eval_loss": 0.5282544493675232, |
| "eval_matthews_correlation": 0.6061655951679263, |
| "eval_runtime": 0.5226, |
| "eval_samples_per_second": 1638.098, |
| "eval_steps_per_second": 103.338, |
| "step": 1443 |
| }, |
| { |
| "epoch": 3.12, |
| "learning_rate": 7.3213890732138915e-06, |
| "loss": 0.263, |
| "step": 1500 |
| }, |
| { |
| "epoch": 4.0, |
| "eval_loss": 0.5712747573852539, |
| "eval_matthews_correlation": 0.6315329091741146, |
| "eval_runtime": 0.5375, |
| "eval_samples_per_second": 1592.64, |
| "eval_steps_per_second": 100.47, |
| "step": 1924 |
| }, |
| { |
| "epoch": 4.16, |
| "learning_rate": 6.2154390621543915e-06, |
| "loss": 0.2115, |
| "step": 2000 |
| }, |
| { |
| "epoch": 5.0, |
| "eval_loss": 0.6360122561454773, |
| "eval_matthews_correlation": 0.6312355301626161, |
| "eval_runtime": 0.5144, |
| "eval_samples_per_second": 1664.168, |
| "eval_steps_per_second": 104.983, |
| "step": 2405 |
| }, |
| { |
| "epoch": 5.2, |
| "learning_rate": 5.1094890510948916e-06, |
| "loss": 0.1785, |
| "step": 2500 |
| }, |
| { |
| "epoch": 6.0, |
| "eval_loss": 0.8011564612388611, |
| "eval_matthews_correlation": 0.6138643516919586, |
| "eval_runtime": 0.5202, |
| "eval_samples_per_second": 1645.404, |
| "eval_steps_per_second": 103.799, |
| "step": 2886 |
| }, |
| { |
| "epoch": 6.24, |
| "learning_rate": 4.003539040035391e-06, |
| "loss": 0.1659, |
| "step": 3000 |
| }, |
| { |
| "epoch": 7.0, |
| "eval_loss": 0.812069833278656, |
| "eval_matthews_correlation": 0.6262286365318072, |
| "eval_runtime": 0.5378, |
| "eval_samples_per_second": 1591.669, |
| "eval_steps_per_second": 100.409, |
| "step": 3367 |
| }, |
| { |
| "epoch": 7.28, |
| "learning_rate": 2.897589028975891e-06, |
| "loss": 0.1218, |
| "step": 3500 |
| }, |
| { |
| "epoch": 8.0, |
| "eval_loss": 0.7686376571655273, |
| "eval_matthews_correlation": 0.636893310208111, |
| "eval_runtime": 0.5131, |
| "eval_samples_per_second": 1668.28, |
| "eval_steps_per_second": 105.242, |
| "step": 3848 |
| }, |
| { |
| "epoch": 8.32, |
| "learning_rate": 1.7916390179163902e-06, |
| "loss": 0.1063, |
| "step": 4000 |
| }, |
| { |
| "epoch": 9.0, |
| "eval_loss": 0.8879063129425049, |
| "eval_matthews_correlation": 0.6314014927014334, |
| "eval_runtime": 0.5155, |
| "eval_samples_per_second": 1660.431, |
| "eval_steps_per_second": 104.747, |
| "step": 4329 |
| }, |
| { |
| "epoch": 9.36, |
| "learning_rate": 6.856890068568902e-07, |
| "loss": 0.0838, |
| "step": 4500 |
| }, |
| { |
| "epoch": 10.0, |
| "eval_loss": 0.8707919716835022, |
| "eval_matthews_correlation": 0.6253690295963492, |
| "eval_runtime": 0.5401, |
| "eval_samples_per_second": 1584.804, |
| "eval_steps_per_second": 99.976, |
| "step": 4810 |
| }, |
| { |
| "epoch": 10.0, |
| "step": 4810, |
| "total_flos": 808993009136460.0, |
| "train_loss": 0.21800182366321588, |
| "train_runtime": 226.2198, |
| "train_samples_per_second": 340.156, |
| "train_steps_per_second": 21.263 |
| } |
| ], |
| "logging_steps": 500, |
| "max_steps": 4810, |
| "num_train_epochs": 10, |
| "save_steps": 500, |
| "total_flos": 808993009136460.0, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|