| { |
| "best_global_step": null, |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 8.0, |
| "eval_steps": 500, |
| "global_step": 480, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.08333333333333333, |
| "grad_norm": 41.758705139160156, |
| "learning_rate": 4.9986614686909146e-05, |
| "loss": 39.5171, |
| "num_input_tokens_seen": 47056, |
| "step": 5 |
| }, |
| { |
| "epoch": 0.16666666666666666, |
| "grad_norm": 35.72260665893555, |
| "learning_rate": 4.994647308096509e-05, |
| "loss": 3.2466, |
| "num_input_tokens_seen": 93544, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.25, |
| "grad_norm": 40.3800163269043, |
| "learning_rate": 4.987961816680492e-05, |
| "loss": 3.3122, |
| "num_input_tokens_seen": 141120, |
| "step": 15 |
| }, |
| { |
| "epoch": 0.3333333333333333, |
| "grad_norm": 22.71551513671875, |
| "learning_rate": 4.9786121534345265e-05, |
| "loss": 3.0265, |
| "num_input_tokens_seen": 188544, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.4166666666666667, |
| "grad_norm": 16.711044311523438, |
| "learning_rate": 4.966608330212198e-05, |
| "loss": 2.6989, |
| "num_input_tokens_seen": 235128, |
| "step": 25 |
| }, |
| { |
| "epoch": 0.5, |
| "grad_norm": 10.093056678771973, |
| "learning_rate": 4.951963201008076e-05, |
| "loss": 2.2284, |
| "num_input_tokens_seen": 281904, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.5833333333333334, |
| "grad_norm": 12.554229736328125, |
| "learning_rate": 4.934692448193334e-05, |
| "loss": 2.2164, |
| "num_input_tokens_seen": 330096, |
| "step": 35 |
| }, |
| { |
| "epoch": 0.6666666666666666, |
| "grad_norm": 20.774412155151367, |
| "learning_rate": 4.914814565722671e-05, |
| "loss": 2.6854, |
| "num_input_tokens_seen": 376824, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.75, |
| "grad_norm": 12.924601554870605, |
| "learning_rate": 4.892350839330522e-05, |
| "loss": 2.4108, |
| "num_input_tokens_seen": 424048, |
| "step": 45 |
| }, |
| { |
| "epoch": 0.8333333333333334, |
| "grad_norm": 23.568218231201172, |
| "learning_rate": 4.867325323737765e-05, |
| "loss": 2.419, |
| "num_input_tokens_seen": 470432, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.9166666666666666, |
| "grad_norm": 29.64082908630371, |
| "learning_rate": 4.839764816893315e-05, |
| "loss": 2.3183, |
| "num_input_tokens_seen": 517624, |
| "step": 55 |
| }, |
| { |
| "epoch": 1.0, |
| "grad_norm": 17.5496883392334, |
| "learning_rate": 4.8096988312782174e-05, |
| "loss": 1.9734, |
| "num_input_tokens_seen": 564920, |
| "step": 60 |
| }, |
| { |
| "epoch": 1.0833333333333333, |
| "grad_norm": 14.56460189819336, |
| "learning_rate": 4.7771595623029394e-05, |
| "loss": 1.8086, |
| "num_input_tokens_seen": 611240, |
| "step": 65 |
| }, |
| { |
| "epoch": 1.1666666666666667, |
| "grad_norm": 11.394974708557129, |
| "learning_rate": 4.742181853831721e-05, |
| "loss": 2.4591, |
| "num_input_tokens_seen": 658904, |
| "step": 70 |
| }, |
| { |
| "epoch": 1.25, |
| "grad_norm": 5.737662315368652, |
| "learning_rate": 4.7048031608708876e-05, |
| "loss": 1.7169, |
| "num_input_tokens_seen": 706480, |
| "step": 75 |
| }, |
| { |
| "epoch": 1.3333333333333333, |
| "grad_norm": 18.580102920532227, |
| "learning_rate": 4.665063509461097e-05, |
| "loss": 1.809, |
| "num_input_tokens_seen": 752904, |
| "step": 80 |
| }, |
| { |
| "epoch": 1.4166666666666667, |
| "grad_norm": 18.77831268310547, |
| "learning_rate": 4.6230054538164474e-05, |
| "loss": 2.3907, |
| "num_input_tokens_seen": 800304, |
| "step": 85 |
| }, |
| { |
| "epoch": 1.5, |
| "grad_norm": 7.019700527191162, |
| "learning_rate": 4.5786740307563636e-05, |
| "loss": 2.0662, |
| "num_input_tokens_seen": 846792, |
| "step": 90 |
| }, |
| { |
| "epoch": 1.5833333333333335, |
| "grad_norm": 12.335309028625488, |
| "learning_rate": 4.5321167114790385e-05, |
| "loss": 2.0185, |
| "num_input_tokens_seen": 894552, |
| "step": 95 |
| }, |
| { |
| "epoch": 1.6666666666666665, |
| "grad_norm": 8.229971885681152, |
| "learning_rate": 4.4833833507280884e-05, |
| "loss": 1.9831, |
| "num_input_tokens_seen": 941040, |
| "step": 100 |
| }, |
| { |
| "epoch": 1.75, |
| "grad_norm": 13.02308177947998, |
| "learning_rate": 4.4325261334068426e-05, |
| "loss": 2.0873, |
| "num_input_tokens_seen": 987776, |
| "step": 105 |
| }, |
| { |
| "epoch": 1.8333333333333335, |
| "grad_norm": 6.7481513023376465, |
| "learning_rate": 4.379599518697444e-05, |
| "loss": 1.9163, |
| "num_input_tokens_seen": 1036072, |
| "step": 110 |
| }, |
| { |
| "epoch": 1.9166666666666665, |
| "grad_norm": 8.03829288482666, |
| "learning_rate": 4.324660181744589e-05, |
| "loss": 1.4848, |
| "num_input_tokens_seen": 1083328, |
| "step": 115 |
| }, |
| { |
| "epoch": 2.0, |
| "grad_norm": 9.837620735168457, |
| "learning_rate": 4.267766952966369e-05, |
| "loss": 1.7719, |
| "num_input_tokens_seen": 1129840, |
| "step": 120 |
| }, |
| { |
| "epoch": 2.0833333333333335, |
| "grad_norm": 14.701643943786621, |
| "learning_rate": 4.208980755057178e-05, |
| "loss": 1.4241, |
| "num_input_tokens_seen": 1177616, |
| "step": 125 |
| }, |
| { |
| "epoch": 2.1666666666666665, |
| "grad_norm": 30.390798568725586, |
| "learning_rate": 4.148364537750172e-05, |
| "loss": 1.6277, |
| "num_input_tokens_seen": 1225296, |
| "step": 130 |
| }, |
| { |
| "epoch": 2.25, |
| "grad_norm": 9.536770820617676, |
| "learning_rate": 4.085983210409114e-05, |
| "loss": 0.9761, |
| "num_input_tokens_seen": 1272160, |
| "step": 135 |
| }, |
| { |
| "epoch": 2.3333333333333335, |
| "grad_norm": 11.45682430267334, |
| "learning_rate": 4.021903572521802e-05, |
| "loss": 1.0144, |
| "num_input_tokens_seen": 1319432, |
| "step": 140 |
| }, |
| { |
| "epoch": 2.4166666666666665, |
| "grad_norm": 11.267831802368164, |
| "learning_rate": 3.956194242169506e-05, |
| "loss": 1.3858, |
| "num_input_tokens_seen": 1365992, |
| "step": 145 |
| }, |
| { |
| "epoch": 2.5, |
| "grad_norm": 11.92896556854248, |
| "learning_rate": 3.888925582549006e-05, |
| "loss": 0.9969, |
| "num_input_tokens_seen": 1412712, |
| "step": 150 |
| }, |
| { |
| "epoch": 2.5833333333333335, |
| "grad_norm": 15.724352836608887, |
| "learning_rate": 3.82016962662592e-05, |
| "loss": 1.0721, |
| "num_input_tokens_seen": 1459448, |
| "step": 155 |
| }, |
| { |
| "epoch": 2.6666666666666665, |
| "grad_norm": 12.68454360961914, |
| "learning_rate": 3.7500000000000003e-05, |
| "loss": 1.6622, |
| "num_input_tokens_seen": 1506944, |
| "step": 160 |
| }, |
| { |
| "epoch": 2.75, |
| "grad_norm": 6.482985496520996, |
| "learning_rate": 3.678491842064995e-05, |
| "loss": 1.059, |
| "num_input_tokens_seen": 1554216, |
| "step": 165 |
| }, |
| { |
| "epoch": 2.8333333333333335, |
| "grad_norm": 15.581390380859375, |
| "learning_rate": 3.6057217255475034e-05, |
| "loss": 0.774, |
| "num_input_tokens_seen": 1600608, |
| "step": 170 |
| }, |
| { |
| "epoch": 2.9166666666666665, |
| "grad_norm": 3.8697454929351807, |
| "learning_rate": 3.5317675745109866e-05, |
| "loss": 1.0403, |
| "num_input_tokens_seen": 1647272, |
| "step": 175 |
| }, |
| { |
| "epoch": 3.0, |
| "grad_norm": 54.316123962402344, |
| "learning_rate": 3.456708580912725e-05, |
| "loss": 2.173, |
| "num_input_tokens_seen": 1694760, |
| "step": 180 |
| }, |
| { |
| "epoch": 3.0833333333333335, |
| "grad_norm": 57.840152740478516, |
| "learning_rate": 3.380625119803084e-05, |
| "loss": 1.9583, |
| "num_input_tokens_seen": 1741936, |
| "step": 185 |
| }, |
| { |
| "epoch": 3.1666666666666665, |
| "grad_norm": 26.283540725708008, |
| "learning_rate": 3.303598663257904e-05, |
| "loss": 1.8407, |
| "num_input_tokens_seen": 1788624, |
| "step": 190 |
| }, |
| { |
| "epoch": 3.25, |
| "grad_norm": 15.37193489074707, |
| "learning_rate": 3.225711693136156e-05, |
| "loss": 1.1398, |
| "num_input_tokens_seen": 1835336, |
| "step": 195 |
| }, |
| { |
| "epoch": 3.3333333333333335, |
| "grad_norm": 53.772212982177734, |
| "learning_rate": 3.147047612756302e-05, |
| "loss": 0.8378, |
| "num_input_tokens_seen": 1882584, |
| "step": 200 |
| }, |
| { |
| "epoch": 3.4166666666666665, |
| "grad_norm": 11.857513427734375, |
| "learning_rate": 3.0676906575859334e-05, |
| "loss": 1.4633, |
| "num_input_tokens_seen": 1929344, |
| "step": 205 |
| }, |
| { |
| "epoch": 3.5, |
| "grad_norm": 5.734513282775879, |
| "learning_rate": 2.9877258050403212e-05, |
| "loss": 0.7289, |
| "num_input_tokens_seen": 1976064, |
| "step": 210 |
| }, |
| { |
| "epoch": 3.5833333333333335, |
| "grad_norm": 7.579070091247559, |
| "learning_rate": 2.9072386834864724e-05, |
| "loss": 0.9374, |
| "num_input_tokens_seen": 2023392, |
| "step": 215 |
| }, |
| { |
| "epoch": 3.6666666666666665, |
| "grad_norm": 15.9894437789917, |
| "learning_rate": 2.8263154805501297e-05, |
| "loss": 0.8814, |
| "num_input_tokens_seen": 2070752, |
| "step": 220 |
| }, |
| { |
| "epoch": 3.75, |
| "grad_norm": 17.481538772583008, |
| "learning_rate": 2.7450428508239024e-05, |
| "loss": 0.8873, |
| "num_input_tokens_seen": 2118464, |
| "step": 225 |
| }, |
| { |
| "epoch": 3.8333333333333335, |
| "grad_norm": 9.79324722290039, |
| "learning_rate": 2.663507823075358e-05, |
| "loss": 0.5864, |
| "num_input_tokens_seen": 2165488, |
| "step": 230 |
| }, |
| { |
| "epoch": 3.9166666666666665, |
| "grad_norm": 29.084766387939453, |
| "learning_rate": 2.5817977070544407e-05, |
| "loss": 0.7186, |
| "num_input_tokens_seen": 2212104, |
| "step": 235 |
| }, |
| { |
| "epoch": 4.0, |
| "grad_norm": 31.556673049926758, |
| "learning_rate": 2.5e-05, |
| "loss": 0.8153, |
| "num_input_tokens_seen": 2259680, |
| "step": 240 |
| }, |
| { |
| "epoch": 4.083333333333333, |
| "grad_norm": 13.266711235046387, |
| "learning_rate": 2.41820229294556e-05, |
| "loss": 0.5078, |
| "num_input_tokens_seen": 2307200, |
| "step": 245 |
| }, |
| { |
| "epoch": 4.166666666666667, |
| "grad_norm": 9.629940032958984, |
| "learning_rate": 2.3364921769246423e-05, |
| "loss": 0.5766, |
| "num_input_tokens_seen": 2354416, |
| "step": 250 |
| }, |
| { |
| "epoch": 4.25, |
| "grad_norm": 3.589583396911621, |
| "learning_rate": 2.2549571491760986e-05, |
| "loss": 0.2176, |
| "num_input_tokens_seen": 2400880, |
| "step": 255 |
| }, |
| { |
| "epoch": 4.333333333333333, |
| "grad_norm": 7.237700939178467, |
| "learning_rate": 2.173684519449872e-05, |
| "loss": 0.6257, |
| "num_input_tokens_seen": 2448016, |
| "step": 260 |
| }, |
| { |
| "epoch": 4.416666666666667, |
| "grad_norm": 34.4775505065918, |
| "learning_rate": 2.0927613165135285e-05, |
| "loss": 0.4594, |
| "num_input_tokens_seen": 2495032, |
| "step": 265 |
| }, |
| { |
| "epoch": 4.5, |
| "grad_norm": 1.8570725917816162, |
| "learning_rate": 2.0122741949596797e-05, |
| "loss": 0.1073, |
| "num_input_tokens_seen": 2542264, |
| "step": 270 |
| }, |
| { |
| "epoch": 4.583333333333333, |
| "grad_norm": 26.749910354614258, |
| "learning_rate": 1.932309342414067e-05, |
| "loss": 0.7155, |
| "num_input_tokens_seen": 2589368, |
| "step": 275 |
| }, |
| { |
| "epoch": 4.666666666666667, |
| "grad_norm": 11.495745658874512, |
| "learning_rate": 1.852952387243698e-05, |
| "loss": 1.0752, |
| "num_input_tokens_seen": 2636712, |
| "step": 280 |
| }, |
| { |
| "epoch": 4.75, |
| "grad_norm": 79.2107925415039, |
| "learning_rate": 1.7742883068638447e-05, |
| "loss": 0.7887, |
| "num_input_tokens_seen": 2682328, |
| "step": 285 |
| }, |
| { |
| "epoch": 4.833333333333333, |
| "grad_norm": 6.504589080810547, |
| "learning_rate": 1.6964013367420966e-05, |
| "loss": 0.1958, |
| "num_input_tokens_seen": 2729264, |
| "step": 290 |
| }, |
| { |
| "epoch": 4.916666666666667, |
| "grad_norm": 17.09076499938965, |
| "learning_rate": 1.6193748801969163e-05, |
| "loss": 0.1749, |
| "num_input_tokens_seen": 2777032, |
| "step": 295 |
| }, |
| { |
| "epoch": 5.0, |
| "grad_norm": 33.851985931396484, |
| "learning_rate": 1.5432914190872757e-05, |
| "loss": 0.6815, |
| "num_input_tokens_seen": 2824600, |
| "step": 300 |
| }, |
| { |
| "epoch": 5.083333333333333, |
| "grad_norm": 26.060302734375, |
| "learning_rate": 1.4682324254890134e-05, |
| "loss": 0.4818, |
| "num_input_tokens_seen": 2871552, |
| "step": 305 |
| }, |
| { |
| "epoch": 5.166666666666667, |
| "grad_norm": 23.55987548828125, |
| "learning_rate": 1.3942782744524973e-05, |
| "loss": 0.0944, |
| "num_input_tokens_seen": 2918968, |
| "step": 310 |
| }, |
| { |
| "epoch": 5.25, |
| "grad_norm": 0.07318225502967834, |
| "learning_rate": 1.3215081579350058e-05, |
| "loss": 0.0347, |
| "num_input_tokens_seen": 2966784, |
| "step": 315 |
| }, |
| { |
| "epoch": 5.333333333333333, |
| "grad_norm": 0.25874024629592896, |
| "learning_rate": 1.2500000000000006e-05, |
| "loss": 0.1513, |
| "num_input_tokens_seen": 3013000, |
| "step": 320 |
| }, |
| { |
| "epoch": 5.416666666666667, |
| "grad_norm": 9.690403938293457, |
| "learning_rate": 1.1798303733740802e-05, |
| "loss": 0.6741, |
| "num_input_tokens_seen": 3059640, |
| "step": 325 |
| }, |
| { |
| "epoch": 5.5, |
| "grad_norm": 0.252065509557724, |
| "learning_rate": 1.1110744174509952e-05, |
| "loss": 0.1384, |
| "num_input_tokens_seen": 3106528, |
| "step": 330 |
| }, |
| { |
| "epoch": 5.583333333333333, |
| "grad_norm": 68.76900482177734, |
| "learning_rate": 1.043805757830495e-05, |
| "loss": 0.1767, |
| "num_input_tokens_seen": 3154400, |
| "step": 335 |
| }, |
| { |
| "epoch": 5.666666666666667, |
| "grad_norm": 0.04263976961374283, |
| "learning_rate": 9.780964274781984e-06, |
| "loss": 0.0103, |
| "num_input_tokens_seen": 3201256, |
| "step": 340 |
| }, |
| { |
| "epoch": 5.75, |
| "grad_norm": 51.393367767333984, |
| "learning_rate": 9.140167895908867e-06, |
| "loss": 0.2753, |
| "num_input_tokens_seen": 3248464, |
| "step": 345 |
| }, |
| { |
| "epoch": 5.833333333333333, |
| "grad_norm": 0.051287949085235596, |
| "learning_rate": 8.51635462249828e-06, |
| "loss": 0.1367, |
| "num_input_tokens_seen": 3294448, |
| "step": 350 |
| }, |
| { |
| "epoch": 5.916666666666667, |
| "grad_norm": 133.54603576660156, |
| "learning_rate": 7.910192449428217e-06, |
| "loss": 1.0681, |
| "num_input_tokens_seen": 3341088, |
| "step": 355 |
| }, |
| { |
| "epoch": 6.0, |
| "grad_norm": 0.007671155035495758, |
| "learning_rate": 7.3223304703363135e-06, |
| "loss": 0.0855, |
| "num_input_tokens_seen": 3389520, |
| "step": 360 |
| }, |
| { |
| "epoch": 6.083333333333333, |
| "grad_norm": 0.07004520297050476, |
| "learning_rate": 6.753398182554116e-06, |
| "loss": 0.007, |
| "num_input_tokens_seen": 3435816, |
| "step": 365 |
| }, |
| { |
| "epoch": 6.166666666666667, |
| "grad_norm": 0.034883052110672, |
| "learning_rate": 6.204004813025568e-06, |
| "loss": 0.0161, |
| "num_input_tokens_seen": 3482312, |
| "step": 370 |
| }, |
| { |
| "epoch": 6.25, |
| "grad_norm": 0.08256684988737106, |
| "learning_rate": 5.674738665931575e-06, |
| "loss": 0.002, |
| "num_input_tokens_seen": 3529280, |
| "step": 375 |
| }, |
| { |
| "epoch": 6.333333333333333, |
| "grad_norm": 9.55763053894043, |
| "learning_rate": 5.166166492719124e-06, |
| "loss": 0.0501, |
| "num_input_tokens_seen": 3575952, |
| "step": 380 |
| }, |
| { |
| "epoch": 6.416666666666667, |
| "grad_norm": 130.89366149902344, |
| "learning_rate": 4.678832885209622e-06, |
| "loss": 0.2303, |
| "num_input_tokens_seen": 3623744, |
| "step": 385 |
| }, |
| { |
| "epoch": 6.5, |
| "grad_norm": 0.029895585030317307, |
| "learning_rate": 4.213259692436367e-06, |
| "loss": 0.0015, |
| "num_input_tokens_seen": 3671256, |
| "step": 390 |
| }, |
| { |
| "epoch": 6.583333333333333, |
| "grad_norm": 0.13803993165493011, |
| "learning_rate": 3.769945461835531e-06, |
| "loss": 0.0139, |
| "num_input_tokens_seen": 3718144, |
| "step": 395 |
| }, |
| { |
| "epoch": 6.666666666666667, |
| "grad_norm": 0.004999049007892609, |
| "learning_rate": 3.3493649053890326e-06, |
| "loss": 0.0001, |
| "num_input_tokens_seen": 3766064, |
| "step": 400 |
| }, |
| { |
| "epoch": 6.75, |
| "grad_norm": 0.4988614618778229, |
| "learning_rate": 2.9519683912911266e-06, |
| "loss": 0.0013, |
| "num_input_tokens_seen": 3812744, |
| "step": 405 |
| }, |
| { |
| "epoch": 6.833333333333333, |
| "grad_norm": 0.0008273015846498311, |
| "learning_rate": 2.578181461682794e-06, |
| "loss": 0.0275, |
| "num_input_tokens_seen": 3859792, |
| "step": 410 |
| }, |
| { |
| "epoch": 6.916666666666667, |
| "grad_norm": 0.04480042681097984, |
| "learning_rate": 2.2284043769706027e-06, |
| "loss": 0.0002, |
| "num_input_tokens_seen": 3906864, |
| "step": 415 |
| }, |
| { |
| "epoch": 7.0, |
| "grad_norm": 0.07072815299034119, |
| "learning_rate": 1.9030116872178316e-06, |
| "loss": 0.0004, |
| "num_input_tokens_seen": 3954440, |
| "step": 420 |
| }, |
| { |
| "epoch": 7.083333333333333, |
| "grad_norm": 0.002131677232682705, |
| "learning_rate": 1.6023518310668618e-06, |
| "loss": 0.0001, |
| "num_input_tokens_seen": 4001944, |
| "step": 425 |
| }, |
| { |
| "epoch": 7.166666666666667, |
| "grad_norm": 0.01150690671056509, |
| "learning_rate": 1.3267467626223606e-06, |
| "loss": 0.0003, |
| "num_input_tokens_seen": 4047736, |
| "step": 430 |
| }, |
| { |
| "epoch": 7.25, |
| "grad_norm": 0.0005230719689279795, |
| "learning_rate": 1.0764916066947794e-06, |
| "loss": 0.0001, |
| "num_input_tokens_seen": 4095480, |
| "step": 435 |
| }, |
| { |
| "epoch": 7.333333333333333, |
| "grad_norm": 0.0005898113595321774, |
| "learning_rate": 8.51854342773295e-07, |
| "loss": 0.0001, |
| "num_input_tokens_seen": 4142520, |
| "step": 440 |
| }, |
| { |
| "epoch": 7.416666666666667, |
| "grad_norm": 0.004632467869669199, |
| "learning_rate": 6.530755180666592e-07, |
| "loss": 0.001, |
| "num_input_tokens_seen": 4188920, |
| "step": 445 |
| }, |
| { |
| "epoch": 7.5, |
| "grad_norm": 0.007595964707434177, |
| "learning_rate": 4.803679899192392e-07, |
| "loss": 0.0001, |
| "num_input_tokens_seen": 4236368, |
| "step": 450 |
| }, |
| { |
| "epoch": 7.583333333333333, |
| "grad_norm": 0.00039777965866960585, |
| "learning_rate": 3.339166978780256e-07, |
| "loss": 0.0, |
| "num_input_tokens_seen": 4283360, |
| "step": 455 |
| }, |
| { |
| "epoch": 7.666666666666667, |
| "grad_norm": 0.12269078195095062, |
| "learning_rate": 2.1387846565474045e-07, |
| "loss": 0.0001, |
| "num_input_tokens_seen": 4330912, |
| "step": 460 |
| }, |
| { |
| "epoch": 7.75, |
| "grad_norm": 0.011688662692904472, |
| "learning_rate": 1.2038183319507955e-07, |
| "loss": 0.0191, |
| "num_input_tokens_seen": 4378920, |
| "step": 465 |
| }, |
| { |
| "epoch": 7.833333333333333, |
| "grad_norm": 0.32097485661506653, |
| "learning_rate": 5.352691903491303e-08, |
| "loss": 0.0016, |
| "num_input_tokens_seen": 4425608, |
| "step": 470 |
| }, |
| { |
| "epoch": 7.916666666666667, |
| "grad_norm": 0.001609671046026051, |
| "learning_rate": 1.3385313090857887e-08, |
| "loss": 0.0016, |
| "num_input_tokens_seen": 4472872, |
| "step": 475 |
| }, |
| { |
| "epoch": 8.0, |
| "grad_norm": 0.0008315809536725283, |
| "learning_rate": 0.0, |
| "loss": 0.0, |
| "num_input_tokens_seen": 4519360, |
| "step": 480 |
| }, |
| { |
| "epoch": 8.0, |
| "num_input_tokens_seen": 4519360, |
| "step": 480, |
| "total_flos": 7.178223503576986e+17, |
| "train_loss": 1.3478541066272858, |
| "train_runtime": 3084.094, |
| "train_samples_per_second": 1.245, |
| "train_steps_per_second": 0.156 |
| } |
| ], |
| "logging_steps": 5, |
| "max_steps": 480, |
| "num_input_tokens_seen": 4519360, |
| "num_train_epochs": 8, |
| "save_steps": 100, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 7.178223503576986e+17, |
| "train_batch_size": 1, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|