| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 11.0, |
| "global_step": 221540, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.25, |
| "learning_rate": 9.909609120521173e-05, |
| "loss": 2.4473, |
| "step": 5000 |
| }, |
| { |
| "epoch": 0.5, |
| "learning_rate": 9.796530039811799e-05, |
| "loss": 1.8064, |
| "step": 10000 |
| }, |
| { |
| "epoch": 0.74, |
| "learning_rate": 9.683473579442635e-05, |
| "loss": 1.6957, |
| "step": 15000 |
| }, |
| { |
| "epoch": 0.99, |
| "learning_rate": 9.570417119073471e-05, |
| "loss": 1.6302, |
| "step": 20000 |
| }, |
| { |
| "epoch": 1.24, |
| "learning_rate": 9.457383279044517e-05, |
| "loss": 1.5838, |
| "step": 25000 |
| }, |
| { |
| "epoch": 1.49, |
| "learning_rate": 9.344304198335143e-05, |
| "loss": 1.552, |
| "step": 30000 |
| }, |
| { |
| "epoch": 1.74, |
| "learning_rate": 9.23127035830619e-05, |
| "loss": 1.5235, |
| "step": 35000 |
| }, |
| { |
| "epoch": 1.99, |
| "learning_rate": 9.118236518277235e-05, |
| "loss": 1.5019, |
| "step": 40000 |
| }, |
| { |
| "epoch": 2.23, |
| "learning_rate": 9.005157437567862e-05, |
| "loss": 1.4807, |
| "step": 45000 |
| }, |
| { |
| "epoch": 2.48, |
| "learning_rate": 8.892100977198697e-05, |
| "loss": 1.4627, |
| "step": 50000 |
| }, |
| { |
| "epoch": 2.73, |
| "learning_rate": 8.779044516829533e-05, |
| "loss": 1.449, |
| "step": 55000 |
| }, |
| { |
| "epoch": 2.98, |
| "learning_rate": 8.665965436120159e-05, |
| "loss": 1.4371, |
| "step": 60000 |
| }, |
| { |
| "epoch": 3.23, |
| "learning_rate": 8.552931596091206e-05, |
| "loss": 1.4229, |
| "step": 65000 |
| }, |
| { |
| "epoch": 3.48, |
| "learning_rate": 8.439852515381831e-05, |
| "loss": 1.4127, |
| "step": 70000 |
| }, |
| { |
| "epoch": 3.72, |
| "learning_rate": 8.326773434672458e-05, |
| "loss": 1.4034, |
| "step": 75000 |
| }, |
| { |
| "epoch": 3.97, |
| "learning_rate": 8.213716974303293e-05, |
| "loss": 1.394, |
| "step": 80000 |
| }, |
| { |
| "epoch": 4.22, |
| "learning_rate": 8.100683134274339e-05, |
| "loss": 1.3822, |
| "step": 85000 |
| }, |
| { |
| "epoch": 4.47, |
| "learning_rate": 7.987581433224755e-05, |
| "loss": 1.3746, |
| "step": 90000 |
| }, |
| { |
| "epoch": 4.72, |
| "learning_rate": 7.874524972855593e-05, |
| "loss": 1.3659, |
| "step": 95000 |
| }, |
| { |
| "epoch": 4.97, |
| "learning_rate": 7.761468512486428e-05, |
| "loss": 1.3603, |
| "step": 100000 |
| }, |
| { |
| "epoch": 5.21, |
| "learning_rate": 7.648412052117264e-05, |
| "loss": 1.3509, |
| "step": 105000 |
| }, |
| { |
| "epoch": 5.46, |
| "learning_rate": 7.53533297140789e-05, |
| "loss": 1.3439, |
| "step": 110000 |
| }, |
| { |
| "epoch": 5.71, |
| "learning_rate": 7.422276511038726e-05, |
| "loss": 1.3389, |
| "step": 115000 |
| }, |
| { |
| "epoch": 5.96, |
| "learning_rate": 7.309220050669564e-05, |
| "loss": 1.3318, |
| "step": 120000 |
| }, |
| { |
| "epoch": 6.21, |
| "learning_rate": 7.196163590300398e-05, |
| "loss": 1.3233, |
| "step": 125000 |
| }, |
| { |
| "epoch": 6.45, |
| "learning_rate": 7.083107129931235e-05, |
| "loss": 1.3169, |
| "step": 130000 |
| }, |
| { |
| "epoch": 6.7, |
| "learning_rate": 6.97002804922186e-05, |
| "loss": 1.3104, |
| "step": 135000 |
| }, |
| { |
| "epoch": 6.95, |
| "learning_rate": 6.856994209192906e-05, |
| "loss": 1.3064, |
| "step": 140000 |
| }, |
| { |
| "epoch": 7.2, |
| "learning_rate": 6.743937748823743e-05, |
| "loss": 1.3007, |
| "step": 145000 |
| }, |
| { |
| "epoch": 7.45, |
| "learning_rate": 6.630858668114369e-05, |
| "loss": 1.2962, |
| "step": 150000 |
| }, |
| { |
| "epoch": 7.7, |
| "learning_rate": 6.517779587404995e-05, |
| "loss": 1.2915, |
| "step": 155000 |
| }, |
| { |
| "epoch": 7.94, |
| "learning_rate": 6.40470050669562e-05, |
| "loss": 1.2877, |
| "step": 160000 |
| }, |
| { |
| "epoch": 8.19, |
| "learning_rate": 6.291644046326457e-05, |
| "loss": 1.2783, |
| "step": 165000 |
| }, |
| { |
| "epoch": 8.44, |
| "learning_rate": 6.178564965617082e-05, |
| "loss": 1.2741, |
| "step": 170000 |
| }, |
| { |
| "epoch": 8.69, |
| "learning_rate": 6.065485884907709e-05, |
| "loss": 1.2705, |
| "step": 175000 |
| }, |
| { |
| "epoch": 8.94, |
| "learning_rate": 5.9524746652189655e-05, |
| "loss": 1.267, |
| "step": 180000 |
| }, |
| { |
| "epoch": 9.19, |
| "learning_rate": 5.839395584509592e-05, |
| "loss": 1.26, |
| "step": 185000 |
| }, |
| { |
| "epoch": 9.43, |
| "learning_rate": 5.7263391241404275e-05, |
| "loss": 1.2558, |
| "step": 190000 |
| }, |
| { |
| "epoch": 9.68, |
| "learning_rate": 5.613260043431053e-05, |
| "loss": 1.2531, |
| "step": 195000 |
| }, |
| { |
| "epoch": 9.93, |
| "learning_rate": 5.5002035830618894e-05, |
| "loss": 1.2492, |
| "step": 200000 |
| }, |
| { |
| "epoch": 10.18, |
| "learning_rate": 5.387124502352515e-05, |
| "loss": 1.2432, |
| "step": 205000 |
| }, |
| { |
| "epoch": 10.43, |
| "learning_rate": 5.274113282663772e-05, |
| "loss": 1.2391, |
| "step": 210000 |
| }, |
| { |
| "epoch": 10.68, |
| "learning_rate": 5.1610342019543975e-05, |
| "loss": 1.2347, |
| "step": 215000 |
| }, |
| { |
| "epoch": 10.92, |
| "learning_rate": 5.047977741585234e-05, |
| "loss": 1.231, |
| "step": 220000 |
| } |
| ], |
| "max_steps": 443080, |
| "num_train_epochs": 22, |
| "total_flos": 7.46536345693738e+18, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|