| { |
| "best_metric": 0.5330753326416016, |
| "best_model_checkpoint": "/scratch/project_2006600/dif_models/fi_bert_reg/diff_bert_base_sampled_unfrozen_early_real_test/checkpoint-3500", |
| "epoch": 10.666666666666666, |
| "eval_steps": 100, |
| "global_step": 4000, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.26666666666666666, |
| "grad_norm": 37.87092208862305, |
| "learning_rate": 5.444444444444444e-07, |
| "loss": 13.2633, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.26666666666666666, |
| "eval_loss": 10.746745109558105, |
| "eval_mse": 10.746745109558105, |
| "eval_r2": -3.070736885070801, |
| "eval_rmse": 3.2782227993011475, |
| "eval_runtime": 57.8905, |
| "eval_samples_per_second": 103.644, |
| "eval_steps_per_second": 25.911, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.5333333333333333, |
| "grad_norm": 11.965848922729492, |
| "learning_rate": 1.0944444444444445e-06, |
| "loss": 6.0044, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.5333333333333333, |
| "eval_loss": 2.2986323833465576, |
| "eval_mse": 2.2986323833465576, |
| "eval_r2": 0.129305899143219, |
| "eval_rmse": 1.5161241292953491, |
| "eval_runtime": 57.8766, |
| "eval_samples_per_second": 103.669, |
| "eval_steps_per_second": 25.917, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.8, |
| "grad_norm": 16.900310516357422, |
| "learning_rate": 1.6500000000000003e-06, |
| "loss": 1.9777, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.8, |
| "eval_loss": 1.0734479427337646, |
| "eval_mse": 1.0734479427337646, |
| "eval_r2": 0.593390941619873, |
| "eval_rmse": 1.0360733270645142, |
| "eval_runtime": 57.8679, |
| "eval_samples_per_second": 103.684, |
| "eval_steps_per_second": 25.921, |
| "step": 300 |
| }, |
| { |
| "epoch": 1.0666666666666667, |
| "grad_norm": 36.441951751708984, |
| "learning_rate": 2.2e-06, |
| "loss": 0.9918, |
| "step": 400 |
| }, |
| { |
| "epoch": 1.0666666666666667, |
| "eval_loss": 0.7707585692405701, |
| "eval_mse": 0.7707585692405701, |
| "eval_r2": 0.708046019077301, |
| "eval_rmse": 0.8779285550117493, |
| "eval_runtime": 57.8738, |
| "eval_samples_per_second": 103.674, |
| "eval_steps_per_second": 25.918, |
| "step": 400 |
| }, |
| { |
| "epoch": 1.3333333333333333, |
| "grad_norm": 35.2764892578125, |
| "learning_rate": 2.755555555555556e-06, |
| "loss": 0.7404, |
| "step": 500 |
| }, |
| { |
| "epoch": 1.3333333333333333, |
| "eval_loss": 0.8621485829353333, |
| "eval_mse": 0.8621485829353333, |
| "eval_r2": 0.6734285354614258, |
| "eval_rmse": 0.9285195469856262, |
| "eval_runtime": 57.8776, |
| "eval_samples_per_second": 103.667, |
| "eval_steps_per_second": 25.917, |
| "step": 500 |
| }, |
| { |
| "epoch": 1.6, |
| "grad_norm": 25.652307510375977, |
| "learning_rate": 3.3055555555555558e-06, |
| "loss": 0.5897, |
| "step": 600 |
| }, |
| { |
| "epoch": 1.6, |
| "eval_loss": 0.7726595997810364, |
| "eval_mse": 0.7726595997810364, |
| "eval_r2": 0.7073259353637695, |
| "eval_rmse": 0.8790105581283569, |
| "eval_runtime": 57.8605, |
| "eval_samples_per_second": 103.698, |
| "eval_steps_per_second": 25.924, |
| "step": 600 |
| }, |
| { |
| "epoch": 1.8666666666666667, |
| "grad_norm": 20.405384063720703, |
| "learning_rate": 3.861111111111112e-06, |
| "loss": 0.5903, |
| "step": 700 |
| }, |
| { |
| "epoch": 1.8666666666666667, |
| "eval_loss": 0.7032095789909363, |
| "eval_mse": 0.7032095789909363, |
| "eval_r2": 0.7336327433586121, |
| "eval_rmse": 0.8385758996009827, |
| "eval_runtime": 57.8612, |
| "eval_samples_per_second": 103.696, |
| "eval_steps_per_second": 25.924, |
| "step": 700 |
| }, |
| { |
| "epoch": 2.1333333333333333, |
| "grad_norm": 17.7819766998291, |
| "learning_rate": 4.416666666666667e-06, |
| "loss": 0.4654, |
| "step": 800 |
| }, |
| { |
| "epoch": 2.1333333333333333, |
| "eval_loss": 0.6557766199111938, |
| "eval_mse": 0.6557766199111938, |
| "eval_r2": 0.7515997886657715, |
| "eval_rmse": 0.809800386428833, |
| "eval_runtime": 57.8596, |
| "eval_samples_per_second": 103.699, |
| "eval_steps_per_second": 25.925, |
| "step": 800 |
| }, |
| { |
| "epoch": 2.4, |
| "grad_norm": 16.64554786682129, |
| "learning_rate": 4.9722222222222224e-06, |
| "loss": 0.4457, |
| "step": 900 |
| }, |
| { |
| "epoch": 2.4, |
| "eval_loss": 0.7730452418327332, |
| "eval_mse": 0.7730452418327332, |
| "eval_r2": 0.707179844379425, |
| "eval_rmse": 0.8792299032211304, |
| "eval_runtime": 57.8442, |
| "eval_samples_per_second": 103.727, |
| "eval_steps_per_second": 25.932, |
| "step": 900 |
| }, |
| { |
| "epoch": 2.6666666666666665, |
| "grad_norm": 23.086116790771484, |
| "learning_rate": 5.527777777777779e-06, |
| "loss": 0.3555, |
| "step": 1000 |
| }, |
| { |
| "epoch": 2.6666666666666665, |
| "eval_loss": 0.6905973553657532, |
| "eval_mse": 0.6905973553657532, |
| "eval_r2": 0.7384101152420044, |
| "eval_rmse": 0.8310218453407288, |
| "eval_runtime": 57.8673, |
| "eval_samples_per_second": 103.685, |
| "eval_steps_per_second": 25.921, |
| "step": 1000 |
| }, |
| { |
| "epoch": 2.9333333333333336, |
| "grad_norm": 12.183361053466797, |
| "learning_rate": 6.083333333333333e-06, |
| "loss": 0.3482, |
| "step": 1100 |
| }, |
| { |
| "epoch": 2.9333333333333336, |
| "eval_loss": 0.7757130265235901, |
| "eval_mse": 0.7757130265235901, |
| "eval_r2": 0.7061693072319031, |
| "eval_rmse": 0.8807457089424133, |
| "eval_runtime": 57.8693, |
| "eval_samples_per_second": 103.682, |
| "eval_steps_per_second": 25.921, |
| "step": 1100 |
| }, |
| { |
| "epoch": 3.2, |
| "grad_norm": 7.989579677581787, |
| "learning_rate": 6.6388888888888895e-06, |
| "loss": 0.2977, |
| "step": 1200 |
| }, |
| { |
| "epoch": 3.2, |
| "eval_loss": 0.756076991558075, |
| "eval_mse": 0.7560770511627197, |
| "eval_r2": 0.7136071920394897, |
| "eval_rmse": 0.8695269227027893, |
| "eval_runtime": 57.8756, |
| "eval_samples_per_second": 103.671, |
| "eval_steps_per_second": 25.918, |
| "step": 1200 |
| }, |
| { |
| "epoch": 3.466666666666667, |
| "grad_norm": 23.822872161865234, |
| "learning_rate": 7.194444444444445e-06, |
| "loss": 0.312, |
| "step": 1300 |
| }, |
| { |
| "epoch": 3.466666666666667, |
| "eval_loss": 0.9145962595939636, |
| "eval_mse": 0.9145963788032532, |
| "eval_r2": 0.6535619497299194, |
| "eval_rmse": 0.9563453197479248, |
| "eval_runtime": 57.8768, |
| "eval_samples_per_second": 103.669, |
| "eval_steps_per_second": 25.917, |
| "step": 1300 |
| }, |
| { |
| "epoch": 3.7333333333333334, |
| "grad_norm": 15.19206428527832, |
| "learning_rate": 7.75e-06, |
| "loss": 0.2681, |
| "step": 1400 |
| }, |
| { |
| "epoch": 3.7333333333333334, |
| "eval_loss": 0.6494566202163696, |
| "eval_mse": 0.6494566202163696, |
| "eval_r2": 0.7539936900138855, |
| "eval_rmse": 0.8058887124061584, |
| "eval_runtime": 57.883, |
| "eval_samples_per_second": 103.657, |
| "eval_steps_per_second": 25.914, |
| "step": 1400 |
| }, |
| { |
| "epoch": 4.0, |
| "grad_norm": 10.404943466186523, |
| "learning_rate": 8.305555555555557e-06, |
| "loss": 0.2744, |
| "step": 1500 |
| }, |
| { |
| "epoch": 4.0, |
| "eval_loss": 0.6918351054191589, |
| "eval_mse": 0.6918351054191589, |
| "eval_r2": 0.7379412651062012, |
| "eval_rmse": 0.8317662477493286, |
| "eval_runtime": 57.8842, |
| "eval_samples_per_second": 103.655, |
| "eval_steps_per_second": 25.914, |
| "step": 1500 |
| }, |
| { |
| "epoch": 4.266666666666667, |
| "grad_norm": 11.18205738067627, |
| "learning_rate": 8.861111111111111e-06, |
| "loss": 0.2142, |
| "step": 1600 |
| }, |
| { |
| "epoch": 4.266666666666667, |
| "eval_loss": 0.7790309190750122, |
| "eval_mse": 0.779030978679657, |
| "eval_r2": 0.7049124836921692, |
| "eval_rmse": 0.8826273083686829, |
| "eval_runtime": 58.0438, |
| "eval_samples_per_second": 103.37, |
| "eval_steps_per_second": 25.843, |
| "step": 1600 |
| }, |
| { |
| "epoch": 4.533333333333333, |
| "grad_norm": 15.371255874633789, |
| "learning_rate": 9.416666666666667e-06, |
| "loss": 0.2103, |
| "step": 1700 |
| }, |
| { |
| "epoch": 4.533333333333333, |
| "eval_loss": 0.6177367568016052, |
| "eval_mse": 0.6177367568016052, |
| "eval_r2": 0.7660087943077087, |
| "eval_rmse": 0.7859622836112976, |
| "eval_runtime": 57.8859, |
| "eval_samples_per_second": 103.652, |
| "eval_steps_per_second": 25.913, |
| "step": 1700 |
| }, |
| { |
| "epoch": 4.8, |
| "grad_norm": 26.033283233642578, |
| "learning_rate": 9.972222222222224e-06, |
| "loss": 0.2009, |
| "step": 1800 |
| }, |
| { |
| "epoch": 4.8, |
| "eval_loss": 0.4925681948661804, |
| "eval_mse": 0.4925681948661804, |
| "eval_r2": 0.8134211301803589, |
| "eval_rmse": 0.7018320560455322, |
| "eval_runtime": 57.8886, |
| "eval_samples_per_second": 103.647, |
| "eval_steps_per_second": 25.912, |
| "step": 1800 |
| }, |
| { |
| "epoch": 5.066666666666666, |
| "grad_norm": 14.781371116638184, |
| "learning_rate": 1.0527777777777778e-05, |
| "loss": 0.2077, |
| "step": 1900 |
| }, |
| { |
| "epoch": 5.066666666666666, |
| "eval_loss": 0.731107234954834, |
| "eval_mse": 0.7311073541641235, |
| "eval_r2": 0.7230653762817383, |
| "eval_rmse": 0.8550481796264648, |
| "eval_runtime": 57.8849, |
| "eval_samples_per_second": 103.654, |
| "eval_steps_per_second": 25.913, |
| "step": 1900 |
| }, |
| { |
| "epoch": 5.333333333333333, |
| "grad_norm": 10.876953125, |
| "learning_rate": 1.1083333333333335e-05, |
| "loss": 0.1621, |
| "step": 2000 |
| }, |
| { |
| "epoch": 5.333333333333333, |
| "eval_loss": 0.9120460748672485, |
| "eval_mse": 0.9120460748672485, |
| "eval_r2": 0.654528021812439, |
| "eval_rmse": 0.9550110101699829, |
| "eval_runtime": 57.8944, |
| "eval_samples_per_second": 103.637, |
| "eval_steps_per_second": 25.909, |
| "step": 2000 |
| }, |
| { |
| "epoch": 5.6, |
| "grad_norm": 9.433772087097168, |
| "learning_rate": 1.163888888888889e-05, |
| "loss": 0.172, |
| "step": 2100 |
| }, |
| { |
| "epoch": 5.6, |
| "eval_loss": 0.7684391140937805, |
| "eval_mse": 0.7684392333030701, |
| "eval_r2": 0.7089245319366455, |
| "eval_rmse": 0.8766066431999207, |
| "eval_runtime": 57.8919, |
| "eval_samples_per_second": 103.641, |
| "eval_steps_per_second": 25.91, |
| "step": 2100 |
| }, |
| { |
| "epoch": 5.866666666666667, |
| "grad_norm": 8.749499320983887, |
| "learning_rate": 1.2194444444444447e-05, |
| "loss": 0.205, |
| "step": 2200 |
| }, |
| { |
| "epoch": 5.866666666666667, |
| "eval_loss": 0.9115136861801147, |
| "eval_mse": 0.9115136861801147, |
| "eval_r2": 0.6547296643257141, |
| "eval_rmse": 0.9547322392463684, |
| "eval_runtime": 57.8967, |
| "eval_samples_per_second": 103.633, |
| "eval_steps_per_second": 25.908, |
| "step": 2200 |
| }, |
| { |
| "epoch": 6.133333333333334, |
| "grad_norm": 13.432199478149414, |
| "learning_rate": 1.275e-05, |
| "loss": 0.1548, |
| "step": 2300 |
| }, |
| { |
| "epoch": 6.133333333333334, |
| "eval_loss": 0.7451738119125366, |
| "eval_mse": 0.7451738119125366, |
| "eval_r2": 0.7177371978759766, |
| "eval_rmse": 0.8632345199584961, |
| "eval_runtime": 57.8927, |
| "eval_samples_per_second": 103.64, |
| "eval_steps_per_second": 25.91, |
| "step": 2300 |
| }, |
| { |
| "epoch": 6.4, |
| "grad_norm": 13.872139930725098, |
| "learning_rate": 1.3305555555555556e-05, |
| "loss": 0.1593, |
| "step": 2400 |
| }, |
| { |
| "epoch": 6.4, |
| "eval_loss": 0.6607070565223694, |
| "eval_mse": 0.6607070565223694, |
| "eval_r2": 0.7497321963310242, |
| "eval_rmse": 0.8128389120101929, |
| "eval_runtime": 57.8803, |
| "eval_samples_per_second": 103.662, |
| "eval_steps_per_second": 25.916, |
| "step": 2400 |
| }, |
| { |
| "epoch": 6.666666666666667, |
| "grad_norm": 10.801892280578613, |
| "learning_rate": 1.3855555555555558e-05, |
| "loss": 0.1684, |
| "step": 2500 |
| }, |
| { |
| "epoch": 6.666666666666667, |
| "eval_loss": 0.8161285519599915, |
| "eval_mse": 0.8161285519599915, |
| "eval_r2": 0.690860390663147, |
| "eval_rmse": 0.903398334980011, |
| "eval_runtime": 57.8953, |
| "eval_samples_per_second": 103.635, |
| "eval_steps_per_second": 25.909, |
| "step": 2500 |
| }, |
| { |
| "epoch": 6.933333333333334, |
| "grad_norm": 13.156838417053223, |
| "learning_rate": 1.441111111111111e-05, |
| "loss": 0.1823, |
| "step": 2600 |
| }, |
| { |
| "epoch": 6.933333333333334, |
| "eval_loss": 1.0073966979980469, |
| "eval_mse": 1.0073966979980469, |
| "eval_r2": 0.6184103488922119, |
| "eval_rmse": 1.003691554069519, |
| "eval_runtime": 57.9026, |
| "eval_samples_per_second": 103.622, |
| "eval_steps_per_second": 25.906, |
| "step": 2600 |
| }, |
| { |
| "epoch": 7.2, |
| "grad_norm": 16.002506256103516, |
| "learning_rate": 1.4966666666666667e-05, |
| "loss": 0.1936, |
| "step": 2700 |
| }, |
| { |
| "epoch": 7.2, |
| "eval_loss": 0.8159008622169495, |
| "eval_mse": 0.8159008622169495, |
| "eval_r2": 0.690946638584137, |
| "eval_rmse": 0.9032723307609558, |
| "eval_runtime": 57.9032, |
| "eval_samples_per_second": 103.621, |
| "eval_steps_per_second": 25.905, |
| "step": 2700 |
| }, |
| { |
| "epoch": 7.466666666666667, |
| "grad_norm": 12.294295310974121, |
| "learning_rate": 1.5522222222222223e-05, |
| "loss": 0.1506, |
| "step": 2800 |
| }, |
| { |
| "epoch": 7.466666666666667, |
| "eval_loss": 0.8324363827705383, |
| "eval_mse": 0.8324363827705383, |
| "eval_r2": 0.6846832036972046, |
| "eval_rmse": 0.9123795032501221, |
| "eval_runtime": 57.9057, |
| "eval_samples_per_second": 103.617, |
| "eval_steps_per_second": 25.904, |
| "step": 2800 |
| }, |
| { |
| "epoch": 7.733333333333333, |
| "grad_norm": 5.195356369018555, |
| "learning_rate": 1.607777777777778e-05, |
| "loss": 0.1499, |
| "step": 2900 |
| }, |
| { |
| "epoch": 7.733333333333333, |
| "eval_loss": 0.8613407611846924, |
| "eval_mse": 0.8613407611846924, |
| "eval_r2": 0.6737345457077026, |
| "eval_rmse": 0.9280844330787659, |
| "eval_runtime": 57.9123, |
| "eval_samples_per_second": 103.605, |
| "eval_steps_per_second": 25.901, |
| "step": 2900 |
| }, |
| { |
| "epoch": 8.0, |
| "grad_norm": 11.357749938964844, |
| "learning_rate": 1.6633333333333336e-05, |
| "loss": 0.1631, |
| "step": 3000 |
| }, |
| { |
| "epoch": 8.0, |
| "eval_loss": 0.6052080988883972, |
| "eval_mse": 0.6052080988883972, |
| "eval_r2": 0.7707545161247253, |
| "eval_rmse": 0.7779512405395508, |
| "eval_runtime": 57.8963, |
| "eval_samples_per_second": 103.634, |
| "eval_steps_per_second": 25.908, |
| "step": 3000 |
| }, |
| { |
| "epoch": 8.266666666666667, |
| "grad_norm": 20.62454605102539, |
| "learning_rate": 1.7188888888888892e-05, |
| "loss": 0.1559, |
| "step": 3100 |
| }, |
| { |
| "epoch": 8.266666666666667, |
| "eval_loss": 0.584992527961731, |
| "eval_mse": 0.584992527961731, |
| "eval_r2": 0.7784119248390198, |
| "eval_rmse": 0.7648480534553528, |
| "eval_runtime": 57.8939, |
| "eval_samples_per_second": 103.638, |
| "eval_steps_per_second": 25.909, |
| "step": 3100 |
| }, |
| { |
| "epoch": 8.533333333333333, |
| "grad_norm": 5.493505954742432, |
| "learning_rate": 1.7744444444444445e-05, |
| "loss": 0.1434, |
| "step": 3200 |
| }, |
| { |
| "epoch": 8.533333333333333, |
| "eval_loss": 0.6216039061546326, |
| "eval_mse": 0.6216039061546326, |
| "eval_r2": 0.7645439505577087, |
| "eval_rmse": 0.7884185910224915, |
| "eval_runtime": 57.9149, |
| "eval_samples_per_second": 103.6, |
| "eval_steps_per_second": 25.9, |
| "step": 3200 |
| }, |
| { |
| "epoch": 8.8, |
| "grad_norm": 8.887247085571289, |
| "learning_rate": 1.83e-05, |
| "loss": 0.1485, |
| "step": 3300 |
| }, |
| { |
| "epoch": 8.8, |
| "eval_loss": 0.6688755750656128, |
| "eval_mse": 0.6688756346702576, |
| "eval_r2": 0.7466380000114441, |
| "eval_rmse": 0.8178481459617615, |
| "eval_runtime": 57.8949, |
| "eval_samples_per_second": 103.636, |
| "eval_steps_per_second": 25.909, |
| "step": 3300 |
| }, |
| { |
| "epoch": 9.066666666666666, |
| "grad_norm": 54.376461029052734, |
| "learning_rate": 1.8855555555555557e-05, |
| "loss": 0.1487, |
| "step": 3400 |
| }, |
| { |
| "epoch": 9.066666666666666, |
| "eval_loss": 0.6742914915084839, |
| "eval_mse": 0.6742914915084839, |
| "eval_r2": 0.7445865869522095, |
| "eval_rmse": 0.8211525678634644, |
| "eval_runtime": 57.9237, |
| "eval_samples_per_second": 103.584, |
| "eval_steps_per_second": 25.896, |
| "step": 3400 |
| }, |
| { |
| "epoch": 9.333333333333334, |
| "grad_norm": 2.6493732929229736, |
| "learning_rate": 1.941111111111111e-05, |
| "loss": 0.1248, |
| "step": 3500 |
| }, |
| { |
| "epoch": 9.333333333333334, |
| "eval_loss": 0.5330753326416016, |
| "eval_mse": 0.5330753326416016, |
| "eval_r2": 0.7980775237083435, |
| "eval_rmse": 0.730120062828064, |
| "eval_runtime": 57.9187, |
| "eval_samples_per_second": 103.593, |
| "eval_steps_per_second": 25.898, |
| "step": 3500 |
| }, |
| { |
| "epoch": 9.6, |
| "grad_norm": 22.65760040283203, |
| "learning_rate": 1.9966666666666666e-05, |
| "loss": 0.127, |
| "step": 3600 |
| }, |
| { |
| "epoch": 9.6, |
| "eval_loss": 0.760678231716156, |
| "eval_mse": 0.760678231716156, |
| "eval_r2": 0.7118643522262573, |
| "eval_rmse": 0.8721687197685242, |
| "eval_runtime": 57.9121, |
| "eval_samples_per_second": 103.605, |
| "eval_steps_per_second": 25.901, |
| "step": 3600 |
| }, |
| { |
| "epoch": 9.866666666666667, |
| "grad_norm": 11.861601829528809, |
| "learning_rate": 1.999958463257905e-05, |
| "loss": 0.1409, |
| "step": 3700 |
| }, |
| { |
| "epoch": 9.866666666666667, |
| "eval_loss": 0.6627658009529114, |
| "eval_mse": 0.6627658009529114, |
| "eval_r2": 0.748952329158783, |
| "eval_rmse": 0.8141043186187744, |
| "eval_runtime": 57.9148, |
| "eval_samples_per_second": 103.601, |
| "eval_steps_per_second": 25.9, |
| "step": 3700 |
| }, |
| { |
| "epoch": 10.133333333333333, |
| "grad_norm": 10.758604049682617, |
| "learning_rate": 1.999823082667316e-05, |
| "loss": 0.1352, |
| "step": 3800 |
| }, |
| { |
| "epoch": 10.133333333333333, |
| "eval_loss": 0.7061165571212769, |
| "eval_mse": 0.7061165571212769, |
| "eval_r2": 0.7325316071510315, |
| "eval_rmse": 0.8403074145317078, |
| "eval_runtime": 58.8761, |
| "eval_samples_per_second": 101.909, |
| "eval_steps_per_second": 25.477, |
| "step": 3800 |
| }, |
| { |
| "epoch": 10.4, |
| "grad_norm": 4.267393112182617, |
| "learning_rate": 1.999593701724414e-05, |
| "loss": 0.1418, |
| "step": 3900 |
| }, |
| { |
| "epoch": 10.4, |
| "eval_loss": 0.6572955846786499, |
| "eval_mse": 0.6572955846786499, |
| "eval_r2": 0.7510244250297546, |
| "eval_rmse": 0.810737669467926, |
| "eval_runtime": 57.921, |
| "eval_samples_per_second": 103.589, |
| "eval_steps_per_second": 25.897, |
| "step": 3900 |
| }, |
| { |
| "epoch": 10.666666666666666, |
| "grad_norm": 4.308215618133545, |
| "learning_rate": 1.9992703419949032e-05, |
| "loss": 0.138, |
| "step": 4000 |
| }, |
| { |
| "epoch": 10.666666666666666, |
| "eval_loss": 0.6316142678260803, |
| "eval_mse": 0.6316142678260803, |
| "eval_r2": 0.7607522010803223, |
| "eval_rmse": 0.7947416305541992, |
| "eval_runtime": 57.92, |
| "eval_samples_per_second": 103.591, |
| "eval_steps_per_second": 25.898, |
| "step": 4000 |
| } |
| ], |
| "logging_steps": 100, |
| "max_steps": 9000, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 24, |
| "save_steps": 500, |
| "stateful_callbacks": { |
| "EarlyStoppingCallback": { |
| "args": { |
| "early_stopping_patience": 6, |
| "early_stopping_threshold": 0.0 |
| }, |
| "attributes": { |
| "early_stopping_patience_counter": 0 |
| } |
| }, |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": false |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 1.6858183041024e+16, |
| "train_batch_size": 4, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|