| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 2.997333333333333, |
| "eval_steps": 10, |
| "global_step": 2811, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "completion_length": 121.971875, |
| "epoch": 0.010666666666666666, |
| "grad_norm": 0.156667098402977, |
| "kl": 2.0313262939453126e-05, |
| "learning_rate": 1.0638297872340426e-05, |
| "loss": 0.001, |
| "reward": 0.0125, |
| "reward_std": 0.025, |
| "rewards/accuracy_reward": 0.009375, |
| "rewards/format_reward": 0.003125, |
| "step": 10 |
| }, |
| { |
| "completion_length": 122.521875, |
| "epoch": 0.021333333333333333, |
| "grad_norm": 0.0012713409960269928, |
| "kl": 0.00021836161613464355, |
| "learning_rate": 2.1276595744680852e-05, |
| "loss": 0.0051, |
| "reward": 0.015625, |
| "reward_std": 0.025966878235340118, |
| "rewards/accuracy_reward": 0.015625, |
| "rewards/format_reward": 0.0, |
| "step": 20 |
| }, |
| { |
| "completion_length": 117.5125, |
| "epoch": 0.032, |
| "grad_norm": 0.002654253738000989, |
| "kl": 0.0003068089485168457, |
| "learning_rate": 3.1914893617021275e-05, |
| "loss": -0.0002, |
| "reward": 0.00625, |
| "reward_std": 0.007216878235340118, |
| "rewards/accuracy_reward": 0.00625, |
| "rewards/format_reward": 0.0, |
| "step": 30 |
| }, |
| { |
| "completion_length": 118.871875, |
| "epoch": 0.042666666666666665, |
| "grad_norm": 0.00353299081325531, |
| "kl": 0.000412750244140625, |
| "learning_rate": 4.2553191489361704e-05, |
| "loss": 0.0055, |
| "reward": 0.009375, |
| "reward_std": 0.01875, |
| "rewards/accuracy_reward": 0.00625, |
| "rewards/format_reward": 0.003125, |
| "step": 40 |
| }, |
| { |
| "completion_length": 121.046875, |
| "epoch": 0.05333333333333334, |
| "grad_norm": 0.003619612194597721, |
| "kl": 0.0004070043563842773, |
| "learning_rate": 5.319148936170213e-05, |
| "loss": 0.0, |
| "reward": 0.0, |
| "reward_std": 0.0, |
| "rewards/accuracy_reward": 0.0, |
| "rewards/format_reward": 0.0, |
| "step": 50 |
| }, |
| { |
| "completion_length": 119.4375, |
| "epoch": 0.064, |
| "grad_norm": 0.11174867302179337, |
| "kl": 0.00045168399810791016, |
| "learning_rate": 6.382978723404255e-05, |
| "loss": 0.0064, |
| "reward": 0.01875, |
| "reward_std": 0.0375, |
| "rewards/accuracy_reward": 0.015625, |
| "rewards/format_reward": 0.003125, |
| "step": 60 |
| }, |
| { |
| "completion_length": 119.05625, |
| "epoch": 0.07466666666666667, |
| "grad_norm": 0.006828859448432922, |
| "kl": 0.0011888980865478516, |
| "learning_rate": 7.446808510638297e-05, |
| "loss": 0.0012, |
| "reward": 0.0125, |
| "reward_std": 0.025, |
| "rewards/accuracy_reward": 0.0125, |
| "rewards/format_reward": 0.0, |
| "step": 70 |
| }, |
| { |
| "completion_length": 120.28125, |
| "epoch": 0.08533333333333333, |
| "grad_norm": 0.0064537739381194115, |
| "kl": 0.0019659996032714844, |
| "learning_rate": 8.510638297872341e-05, |
| "loss": 0.0028, |
| "reward": 0.0125, |
| "reward_std": 0.025, |
| "rewards/accuracy_reward": 0.0125, |
| "rewards/format_reward": 0.0, |
| "step": 80 |
| }, |
| { |
| "completion_length": 117.559375, |
| "epoch": 0.096, |
| "grad_norm": 0.09068689495325089, |
| "kl": 0.0025023460388183595, |
| "learning_rate": 9.574468085106382e-05, |
| "loss": 0.003, |
| "reward": 0.021875, |
| "reward_std": 0.04375, |
| "rewards/accuracy_reward": 0.01875, |
| "rewards/format_reward": 0.003125, |
| "step": 90 |
| }, |
| { |
| "completion_length": 117.68125, |
| "epoch": 0.10666666666666667, |
| "grad_norm": 0.16541939973831177, |
| "kl": 0.00291900634765625, |
| "learning_rate": 0.00010638297872340425, |
| "loss": 0.0008, |
| "reward": 0.021875, |
| "reward_std": 0.03125, |
| "rewards/accuracy_reward": 0.021875, |
| "rewards/format_reward": 0.0, |
| "step": 100 |
| }, |
| { |
| "completion_length": 116.971875, |
| "epoch": 0.11733333333333333, |
| "grad_norm": 0.07206544280052185, |
| "kl": 0.0038990020751953126, |
| "learning_rate": 0.00011702127659574467, |
| "loss": 0.0026, |
| "reward": 0.015625, |
| "reward_std": 0.03125, |
| "rewards/accuracy_reward": 0.0125, |
| "rewards/format_reward": 0.003125, |
| "step": 110 |
| }, |
| { |
| "completion_length": 114.996875, |
| "epoch": 0.128, |
| "grad_norm": 0.02286006510257721, |
| "kl": 0.007346725463867188, |
| "learning_rate": 0.0001276595744680851, |
| "loss": 0.0076, |
| "reward": 0.025, |
| "reward_std": 0.05, |
| "rewards/accuracy_reward": 0.015625, |
| "rewards/format_reward": 0.009375, |
| "step": 120 |
| }, |
| { |
| "completion_length": 119.315625, |
| "epoch": 0.13866666666666666, |
| "grad_norm": 0.015629781410098076, |
| "kl": 0.008090972900390625, |
| "learning_rate": 0.00013829787234042552, |
| "loss": 0.0011, |
| "reward": 0.009375, |
| "reward_std": 0.01875, |
| "rewards/accuracy_reward": 0.009375, |
| "rewards/format_reward": 0.0, |
| "step": 130 |
| }, |
| { |
| "completion_length": 121.821875, |
| "epoch": 0.14933333333333335, |
| "grad_norm": 0.15498439967632294, |
| "kl": 0.006272506713867187, |
| "learning_rate": 0.00014893617021276593, |
| "loss": -0.0012, |
| "reward": 0.021875, |
| "reward_std": 0.03846687823534012, |
| "rewards/accuracy_reward": 0.01875, |
| "rewards/format_reward": 0.003125, |
| "step": 140 |
| }, |
| { |
| "completion_length": 121.409375, |
| "epoch": 0.16, |
| "grad_norm": 0.18756870925426483, |
| "kl": 0.00465240478515625, |
| "learning_rate": 0.00015957446808510637, |
| "loss": 0.0012, |
| "reward": 0.021875, |
| "reward_std": 0.03846687823534012, |
| "rewards/accuracy_reward": 0.01875, |
| "rewards/format_reward": 0.003125, |
| "step": 150 |
| }, |
| { |
| "completion_length": 118.26875, |
| "epoch": 0.17066666666666666, |
| "grad_norm": 0.011626984924077988, |
| "kl": 0.01092681884765625, |
| "learning_rate": 0.00017021276595744682, |
| "loss": -0.0011, |
| "reward": 0.021875, |
| "reward_std": 0.03318375647068024, |
| "rewards/accuracy_reward": 0.01875, |
| "rewards/format_reward": 0.003125, |
| "step": 160 |
| }, |
| { |
| "completion_length": 119.25, |
| "epoch": 0.18133333333333335, |
| "grad_norm": 0.00764912273734808, |
| "kl": 0.00976104736328125, |
| "learning_rate": 0.0001808510638297872, |
| "loss": 0.0045, |
| "reward": 0.021875, |
| "reward_std": 0.04375, |
| "rewards/accuracy_reward": 0.021875, |
| "rewards/format_reward": 0.0, |
| "step": 170 |
| }, |
| { |
| "completion_length": 115.35, |
| "epoch": 0.192, |
| "grad_norm": 0.0785018652677536, |
| "kl": 0.014077377319335938, |
| "learning_rate": 0.00019148936170212765, |
| "loss": 0.0037, |
| "reward": 0.025, |
| "reward_std": 0.04471687823534012, |
| "rewards/accuracy_reward": 0.009375, |
| "rewards/format_reward": 0.015625, |
| "step": 180 |
| }, |
| { |
| "completion_length": 108.6125, |
| "epoch": 0.20266666666666666, |
| "grad_norm": 0.13107918202877045, |
| "kl": 0.039361572265625, |
| "learning_rate": 0.00020212765957446807, |
| "loss": 0.0412, |
| "reward": 0.11875, |
| "reward_std": 0.18907372057437896, |
| "rewards/accuracy_reward": 0.021875, |
| "rewards/format_reward": 0.096875, |
| "step": 190 |
| }, |
| { |
| "completion_length": 89.915625, |
| "epoch": 0.21333333333333335, |
| "grad_norm": 0.19012346863746643, |
| "kl": 0.08895263671875, |
| "learning_rate": 0.0002127659574468085, |
| "loss": 0.1321, |
| "reward": 0.46875, |
| "reward_std": 0.41404569447040557, |
| "rewards/accuracy_reward": 0.015625, |
| "rewards/format_reward": 0.453125, |
| "step": 200 |
| }, |
| { |
| "completion_length": 47.74375, |
| "epoch": 0.224, |
| "grad_norm": 0.4668453335762024, |
| "kl": 0.26416015625, |
| "learning_rate": 0.0002234042553191489, |
| "loss": 0.0712, |
| "reward": 0.871875, |
| "reward_std": 0.19805223047733306, |
| "rewards/accuracy_reward": 0.01875, |
| "rewards/format_reward": 0.853125, |
| "step": 210 |
| }, |
| { |
| "completion_length": 45.15625, |
| "epoch": 0.23466666666666666, |
| "grad_norm": 0.21052278578281403, |
| "kl": 0.3112213134765625, |
| "learning_rate": 0.00023404255319148934, |
| "loss": 0.0464, |
| "reward": 0.890625, |
| "reward_std": 0.11346687823534012, |
| "rewards/accuracy_reward": 0.0125, |
| "rewards/format_reward": 0.878125, |
| "step": 220 |
| }, |
| { |
| "completion_length": 57.2875, |
| "epoch": 0.24533333333333332, |
| "grad_norm": 0.16618619859218597, |
| "kl": 0.254522705078125, |
| "learning_rate": 0.00024468085106382976, |
| "loss": 0.0589, |
| "reward": 0.834375, |
| "reward_std": 0.12261751294136047, |
| "rewards/accuracy_reward": 0.0125, |
| "rewards/format_reward": 0.821875, |
| "step": 230 |
| }, |
| { |
| "completion_length": 68.153125, |
| "epoch": 0.256, |
| "grad_norm": 0.17739807069301605, |
| "kl": 0.214471435546875, |
| "learning_rate": 0.0002553191489361702, |
| "loss": 0.1375, |
| "reward": 0.659375, |
| "reward_std": 0.28527562469244006, |
| "rewards/accuracy_reward": 0.009375, |
| "rewards/format_reward": 0.65, |
| "step": 240 |
| }, |
| { |
| "completion_length": 52.709375, |
| "epoch": 0.26666666666666666, |
| "grad_norm": 0.09843996912240982, |
| "kl": 0.2847900390625, |
| "learning_rate": 0.0002659574468085106, |
| "loss": 0.1085, |
| "reward": 0.834375, |
| "reward_std": 0.290549997985363, |
| "rewards/accuracy_reward": 0.034375, |
| "rewards/format_reward": 0.8, |
| "step": 250 |
| }, |
| { |
| "completion_length": 51.55625, |
| "epoch": 0.2773333333333333, |
| "grad_norm": 0.1133696436882019, |
| "kl": 0.276953125, |
| "learning_rate": 0.00027659574468085103, |
| "loss": 0.0437, |
| "reward": 0.903125, |
| "reward_std": 0.1361730858683586, |
| "rewards/accuracy_reward": 0.025, |
| "rewards/format_reward": 0.878125, |
| "step": 260 |
| }, |
| { |
| "completion_length": 55.046875, |
| "epoch": 0.288, |
| "grad_norm": 0.14536090195178986, |
| "kl": 0.2501953125, |
| "learning_rate": 0.0002872340425531915, |
| "loss": 0.0588, |
| "reward": 0.878125, |
| "reward_std": 0.13846687823534012, |
| "rewards/accuracy_reward": 0.00625, |
| "rewards/format_reward": 0.871875, |
| "step": 270 |
| }, |
| { |
| "completion_length": 55.853125, |
| "epoch": 0.2986666666666667, |
| "grad_norm": 0.1799221634864807, |
| "kl": 0.3143310546875, |
| "learning_rate": 0.00029787234042553186, |
| "loss": 0.0609, |
| "reward": 0.90625, |
| "reward_std": 0.18080126941204072, |
| "rewards/accuracy_reward": 0.021875, |
| "rewards/format_reward": 0.884375, |
| "step": 280 |
| }, |
| { |
| "completion_length": 59.85, |
| "epoch": 0.30933333333333335, |
| "grad_norm": 0.10688479989767075, |
| "kl": 0.20706787109375, |
| "learning_rate": 0.0002999925930442553, |
| "loss": 0.0522, |
| "reward": 0.815625, |
| "reward_std": 0.2959165498614311, |
| "rewards/accuracy_reward": 0.065625, |
| "rewards/format_reward": 0.75, |
| "step": 290 |
| }, |
| { |
| "completion_length": 64.625, |
| "epoch": 0.32, |
| "grad_norm": 0.03851361572742462, |
| "kl": 0.201220703125, |
| "learning_rate": 0.00029996250354024344, |
| "loss": 0.0815, |
| "reward": 0.8625, |
| "reward_std": 0.21301814764738083, |
| "rewards/accuracy_reward": 0.0125, |
| "rewards/format_reward": 0.85, |
| "step": 300 |
| }, |
| { |
| "completion_length": 57.95, |
| "epoch": 0.33066666666666666, |
| "grad_norm": 0.23480646312236786, |
| "kl": 0.221240234375, |
| "learning_rate": 0.0002999092731927958, |
| "loss": 0.0292, |
| "reward": 0.921875, |
| "reward_std": 0.15895397514104842, |
| "rewards/accuracy_reward": 0.0375, |
| "rewards/format_reward": 0.884375, |
| "step": 310 |
| }, |
| { |
| "completion_length": 64.196875, |
| "epoch": 0.3413333333333333, |
| "grad_norm": 0.1151675432920456, |
| "kl": 0.20123291015625, |
| "learning_rate": 0.0002998329102159332, |
| "loss": 0.0491, |
| "reward": 0.83125, |
| "reward_std": 0.19258119761943818, |
| "rewards/accuracy_reward": 0.01875, |
| "rewards/format_reward": 0.8125, |
| "step": 320 |
| }, |
| { |
| "completion_length": 70.628125, |
| "epoch": 0.352, |
| "grad_norm": 0.1377689391374588, |
| "kl": 0.1906005859375, |
| "learning_rate": 0.0002997334263932927, |
| "loss": 0.0841, |
| "reward": 0.846875, |
| "reward_std": 0.21890811175107955, |
| "rewards/accuracy_reward": 0.015625, |
| "rewards/format_reward": 0.83125, |
| "step": 330 |
| }, |
| { |
| "completion_length": 61.54375, |
| "epoch": 0.3626666666666667, |
| "grad_norm": 0.0947548896074295, |
| "kl": 0.21240234375, |
| "learning_rate": 0.0002996108370763087, |
| "loss": 0.062, |
| "reward": 0.88125, |
| "reward_std": 0.13713996410369872, |
| "rewards/accuracy_reward": 0.015625, |
| "rewards/format_reward": 0.865625, |
| "step": 340 |
| }, |
| { |
| "completion_length": 60.109375, |
| "epoch": 0.37333333333333335, |
| "grad_norm": 0.14599719643592834, |
| "kl": 0.2236083984375, |
| "learning_rate": 0.0002994651611818448, |
| "loss": 0.0408, |
| "reward": 0.928125, |
| "reward_std": 0.18282372057437896, |
| "rewards/accuracy_reward": 0.028125, |
| "rewards/format_reward": 0.9, |
| "step": 350 |
| }, |
| { |
| "completion_length": 62.90625, |
| "epoch": 0.384, |
| "grad_norm": 0.3738599121570587, |
| "kl": 0.2464111328125, |
| "learning_rate": 0.00029929642118927394, |
| "loss": 0.0753, |
| "reward": 0.834375, |
| "reward_std": 0.20676814764738083, |
| "rewards/accuracy_reward": 0.025, |
| "rewards/format_reward": 0.809375, |
| "step": 360 |
| }, |
| { |
| "completion_length": 70.1625, |
| "epoch": 0.39466666666666667, |
| "grad_norm": 2.8762810230255127, |
| "kl": 0.88681640625, |
| "learning_rate": 0.00029910464313701013, |
| "loss": 0.2053, |
| "reward": 0.640625, |
| "reward_std": 0.38192625939846037, |
| "rewards/accuracy_reward": 0.009375, |
| "rewards/format_reward": 0.63125, |
| "step": 370 |
| }, |
| { |
| "completion_length": 43.578125, |
| "epoch": 0.4053333333333333, |
| "grad_norm": 0.9355350136756897, |
| "kl": 2.06865234375, |
| "learning_rate": 0.0002988898566184902, |
| "loss": 0.2631, |
| "reward": 0.725, |
| "reward_std": 0.32462068647146225, |
| "rewards/accuracy_reward": 0.0, |
| "rewards/format_reward": 0.725, |
| "step": 380 |
| }, |
| { |
| "completion_length": 44.19375, |
| "epoch": 0.416, |
| "grad_norm": 0.709173858165741, |
| "kl": 3.47939453125, |
| "learning_rate": 0.0002986520947776074, |
| "loss": 0.3225, |
| "reward": 0.6125, |
| "reward_std": 0.3950331017374992, |
| "rewards/accuracy_reward": 0.009375, |
| "rewards/format_reward": 0.603125, |
| "step": 390 |
| }, |
| { |
| "completion_length": 54.775, |
| "epoch": 0.4266666666666667, |
| "grad_norm": 0.6549698114395142, |
| "kl": 4.3202392578125, |
| "learning_rate": 0.0002983913943035968, |
| "loss": 0.3808, |
| "reward": 0.66875, |
| "reward_std": 0.3901100158691406, |
| "rewards/accuracy_reward": 0.01875, |
| "rewards/format_reward": 0.65, |
| "step": 400 |
| }, |
| { |
| "completion_length": 53.871875, |
| "epoch": 0.43733333333333335, |
| "grad_norm": 0.01826515607535839, |
| "kl": 2.477734375, |
| "learning_rate": 0.00029810779542537355, |
| "loss": 0.2661, |
| "reward": 0.79375, |
| "reward_std": 0.22999776750802994, |
| "rewards/accuracy_reward": 0.025, |
| "rewards/format_reward": 0.76875, |
| "step": 410 |
| }, |
| { |
| "completion_length": 49.434375, |
| "epoch": 0.448, |
| "grad_norm": 0.5134692192077637, |
| "kl": 2.07587890625, |
| "learning_rate": 0.0002978013419053255, |
| "loss": 0.2091, |
| "reward": 0.771875, |
| "reward_std": 0.26785253882408144, |
| "rewards/accuracy_reward": 0.0, |
| "rewards/format_reward": 0.771875, |
| "step": 420 |
| }, |
| { |
| "completion_length": 59.475, |
| "epoch": 0.45866666666666667, |
| "grad_norm": 0.7835673689842224, |
| "kl": 2.516943359375, |
| "learning_rate": 0.00029747208103256, |
| "loss": 0.2312, |
| "reward": 0.740625, |
| "reward_std": 0.31220938116312025, |
| "rewards/accuracy_reward": 0.0125, |
| "rewards/format_reward": 0.728125, |
| "step": 430 |
| }, |
| { |
| "completion_length": 58.815625, |
| "epoch": 0.4693333333333333, |
| "grad_norm": 0.021143430843949318, |
| "kl": 2.1959716796875, |
| "learning_rate": 0.0002971200636156068, |
| "loss": 0.2386, |
| "reward": 0.796875, |
| "reward_std": 0.2231356605887413, |
| "rewards/accuracy_reward": 0.0, |
| "rewards/format_reward": 0.796875, |
| "step": 440 |
| }, |
| { |
| "completion_length": 55.696875, |
| "epoch": 0.48, |
| "grad_norm": 1.8231980800628662, |
| "kl": 2.9314697265625, |
| "learning_rate": 0.00029674534397457745, |
| "loss": 0.3506, |
| "reward": 0.796875, |
| "reward_std": 0.2616912335157394, |
| "rewards/accuracy_reward": 0.021875, |
| "rewards/format_reward": 0.775, |
| "step": 450 |
| }, |
| { |
| "completion_length": 49.778125, |
| "epoch": 0.49066666666666664, |
| "grad_norm": 0.5252532362937927, |
| "kl": 1.47607421875, |
| "learning_rate": 0.00029634797993278333, |
| "loss": 0.2026, |
| "reward": 0.89375, |
| "reward_std": 0.11293471753597259, |
| "rewards/accuracy_reward": 0.0125, |
| "rewards/format_reward": 0.88125, |
| "step": 460 |
| }, |
| { |
| "completion_length": 57.903125, |
| "epoch": 0.5013333333333333, |
| "grad_norm": 0.10718824714422226, |
| "kl": 1.683837890625, |
| "learning_rate": 0.000295928032807813, |
| "loss": 0.1887, |
| "reward": 0.859375, |
| "reward_std": 0.1423343911767006, |
| "rewards/accuracy_reward": 0.00625, |
| "rewards/format_reward": 0.853125, |
| "step": 470 |
| }, |
| { |
| "completion_length": 63.734375, |
| "epoch": 0.512, |
| "grad_norm": 0.32101932168006897, |
| "kl": 2.9671875, |
| "learning_rate": 0.00029548556740206994, |
| "loss": 0.3254, |
| "reward": 0.79375, |
| "reward_std": 0.29874250292778015, |
| "rewards/accuracy_reward": 0.009375, |
| "rewards/format_reward": 0.784375, |
| "step": 480 |
| }, |
| { |
| "completion_length": 66.253125, |
| "epoch": 0.5226666666666666, |
| "grad_norm": 0.7132259011268616, |
| "kl": 2.6101806640625, |
| "learning_rate": 0.0002950206519927731, |
| "loss": 0.2574, |
| "reward": 0.728125, |
| "reward_std": 0.3086773693561554, |
| "rewards/accuracy_reward": 0.01875, |
| "rewards/format_reward": 0.709375, |
| "step": 490 |
| }, |
| { |
| "completion_length": 64.765625, |
| "epoch": 0.5333333333333333, |
| "grad_norm": 0.7293491959571838, |
| "kl": 3.3251220703125, |
| "learning_rate": 0.00029453335832142075, |
| "loss": 0.3315, |
| "reward": 0.75625, |
| "reward_std": 0.27030970752239225, |
| "rewards/accuracy_reward": 0.009375, |
| "rewards/format_reward": 0.746875, |
| "step": 500 |
| }, |
| { |
| "completion_length": 61.140625, |
| "epoch": 0.544, |
| "grad_norm": 0.4900813400745392, |
| "kl": 1.6069091796875, |
| "learning_rate": 0.0002940237615827202, |
| "loss": 0.162, |
| "reward": 0.86875, |
| "reward_std": 0.21899680644273758, |
| "rewards/accuracy_reward": 0.053125, |
| "rewards/format_reward": 0.815625, |
| "step": 510 |
| }, |
| { |
| "completion_length": 59.621875, |
| "epoch": 0.5546666666666666, |
| "grad_norm": 0.4984245002269745, |
| "kl": 1.695751953125, |
| "learning_rate": 0.00029349194041298435, |
| "loss": 0.2075, |
| "reward": 0.903125, |
| "reward_std": 0.16081304997205734, |
| "rewards/accuracy_reward": 0.034375, |
| "rewards/format_reward": 0.86875, |
| "step": 520 |
| }, |
| { |
| "completion_length": 62.209375, |
| "epoch": 0.5653333333333334, |
| "grad_norm": 0.25215986371040344, |
| "kl": 1.5575439453125, |
| "learning_rate": 0.0002929379768779971, |
| "loss": 0.1648, |
| "reward": 0.890625, |
| "reward_std": 0.17983439117670058, |
| "rewards/accuracy_reward": 0.040625, |
| "rewards/format_reward": 0.85, |
| "step": 530 |
| }, |
| { |
| "completion_length": 65.865625, |
| "epoch": 0.576, |
| "grad_norm": 0.1489488184452057, |
| "kl": 2.0063720703125, |
| "learning_rate": 0.0002923619564603501, |
| "loss": 0.187, |
| "reward": 0.78125, |
| "reward_std": 0.2043856605887413, |
| "rewards/accuracy_reward": 0.03125, |
| "rewards/format_reward": 0.75, |
| "step": 540 |
| }, |
| { |
| "completion_length": 68.68125, |
| "epoch": 0.5866666666666667, |
| "grad_norm": 0.24991311132907867, |
| "kl": 1.0421142578125, |
| "learning_rate": 0.00029176396804625135, |
| "loss": 0.0977, |
| "reward": 0.909375, |
| "reward_std": 0.1441847175359726, |
| "rewards/accuracy_reward": 0.021875, |
| "rewards/format_reward": 0.8875, |
| "step": 550 |
| }, |
| { |
| "completion_length": 64.2, |
| "epoch": 0.5973333333333334, |
| "grad_norm": 0.7193971872329712, |
| "kl": 2.2302490234375, |
| "learning_rate": 0.00029114410391180946, |
| "loss": 0.2166, |
| "reward": 0.834375, |
| "reward_std": 0.2048343911767006, |
| "rewards/accuracy_reward": 0.015625, |
| "rewards/format_reward": 0.81875, |
| "step": 560 |
| }, |
| { |
| "completion_length": 66.01875, |
| "epoch": 0.608, |
| "grad_norm": 0.59996098279953, |
| "kl": 2.652294921875, |
| "learning_rate": 0.0002905024597087945, |
| "loss": 0.2907, |
| "reward": 0.815625, |
| "reward_std": 0.20596464574337006, |
| "rewards/accuracy_reward": 0.028125, |
| "rewards/format_reward": 0.7875, |
| "step": 570 |
| }, |
| { |
| "completion_length": 69.009375, |
| "epoch": 0.6186666666666667, |
| "grad_norm": 0.32363754510879517, |
| "kl": 0.75888671875, |
| "learning_rate": 0.0002898391344498775, |
| "loss": 0.112, |
| "reward": 0.896875, |
| "reward_std": 0.14761751294136047, |
| "rewards/accuracy_reward": 0.034375, |
| "rewards/format_reward": 0.8625, |
| "step": 580 |
| }, |
| { |
| "completion_length": 58.490625, |
| "epoch": 0.6293333333333333, |
| "grad_norm": 0.5817243456840515, |
| "kl": 3.2273193359375, |
| "learning_rate": 0.0002891542304933521, |
| "loss": 0.3775, |
| "reward": 0.796875, |
| "reward_std": 0.24620190411806106, |
| "rewards/accuracy_reward": 0.00625, |
| "rewards/format_reward": 0.790625, |
| "step": 590 |
| }, |
| { |
| "completion_length": 56.890625, |
| "epoch": 0.64, |
| "grad_norm": 0.561817467212677, |
| "kl": 1.345703125, |
| "learning_rate": 0.00028844785352733924, |
| "loss": 0.1409, |
| "reward": 0.884375, |
| "reward_std": 0.1315855011343956, |
| "rewards/accuracy_reward": 0.021875, |
| "rewards/format_reward": 0.8625, |
| "step": 600 |
| }, |
| { |
| "completion_length": 54.61875, |
| "epoch": 0.6506666666666666, |
| "grad_norm": 0.41451311111450195, |
| "kl": 1.3966552734375, |
| "learning_rate": 0.00028772011255347873, |
| "loss": 0.1476, |
| "reward": 0.890625, |
| "reward_std": 0.16838996410369872, |
| "rewards/accuracy_reward": 0.01875, |
| "rewards/format_reward": 0.871875, |
| "step": 610 |
| }, |
| { |
| "completion_length": 58.15, |
| "epoch": 0.6613333333333333, |
| "grad_norm": 0.38927924633026123, |
| "kl": 2.2388427734375, |
| "learning_rate": 0.00028697111987010865, |
| "loss": 0.2576, |
| "reward": 0.871875, |
| "reward_std": 0.1995512694120407, |
| "rewards/accuracy_reward": 0.028125, |
| "rewards/format_reward": 0.84375, |
| "step": 620 |
| }, |
| { |
| "completion_length": 57.20625, |
| "epoch": 0.672, |
| "grad_norm": 0.09751415997743607, |
| "kl": 0.793994140625, |
| "learning_rate": 0.0002862009910549369, |
| "loss": 0.0629, |
| "reward": 0.9375, |
| "reward_std": 0.10386751294136047, |
| "rewards/accuracy_reward": 0.01875, |
| "rewards/format_reward": 0.91875, |
| "step": 630 |
| }, |
| { |
| "completion_length": 65.846875, |
| "epoch": 0.6826666666666666, |
| "grad_norm": 0.2675510048866272, |
| "kl": 2.466015625, |
| "learning_rate": 0.0002854098449472061, |
| "loss": 0.2627, |
| "reward": 0.79375, |
| "reward_std": 0.2520918682217598, |
| "rewards/accuracy_reward": 0.015625, |
| "rewards/format_reward": 0.778125, |
| "step": 640 |
| }, |
| { |
| "completion_length": 62.98125, |
| "epoch": 0.6933333333333334, |
| "grad_norm": 0.15855202078819275, |
| "kl": 1.8398193359375, |
| "learning_rate": 0.00028459780362935527, |
| "loss": 0.177, |
| "reward": 0.91875, |
| "reward_std": 0.15879059880971907, |
| "rewards/accuracy_reward": 0.0125, |
| "rewards/format_reward": 0.90625, |
| "step": 650 |
| }, |
| { |
| "completion_length": 62.6, |
| "epoch": 0.704, |
| "grad_norm": 0.12087615579366684, |
| "kl": 2.0813720703125, |
| "learning_rate": 0.0002837649924081816, |
| "loss": 0.1866, |
| "reward": 0.90625, |
| "reward_std": 0.19479155987501146, |
| "rewards/accuracy_reward": 0.03125, |
| "rewards/format_reward": 0.875, |
| "step": 660 |
| }, |
| { |
| "completion_length": 65.925, |
| "epoch": 0.7146666666666667, |
| "grad_norm": 0.39411771297454834, |
| "kl": 1.392919921875, |
| "learning_rate": 0.00028291153979550387, |
| "loss": 0.2015, |
| "reward": 0.915625, |
| "reward_std": 0.1775405988097191, |
| "rewards/accuracy_reward": 0.028125, |
| "rewards/format_reward": 0.8875, |
| "step": 670 |
| }, |
| { |
| "completion_length": 64.378125, |
| "epoch": 0.7253333333333334, |
| "grad_norm": 1.1659783124923706, |
| "kl": 2.8165283203125, |
| "learning_rate": 0.00028203757748833174, |
| "loss": 0.3109, |
| "reward": 0.778125, |
| "reward_std": 0.21169123351573943, |
| "rewards/accuracy_reward": 0.003125, |
| "rewards/format_reward": 0.775, |
| "step": 680 |
| }, |
| { |
| "completion_length": 57.334375, |
| "epoch": 0.736, |
| "grad_norm": 0.27627384662628174, |
| "kl": 1.0085693359375, |
| "learning_rate": 0.0002811432403485437, |
| "loss": 0.1226, |
| "reward": 0.859375, |
| "reward_std": 0.11540063470602036, |
| "rewards/accuracy_reward": 0.009375, |
| "rewards/format_reward": 0.85, |
| "step": 690 |
| }, |
| { |
| "completion_length": 54.6, |
| "epoch": 0.7466666666666667, |
| "grad_norm": 0.4506663382053375, |
| "kl": 2.3274169921875, |
| "learning_rate": 0.00028022866638207624, |
| "loss": 0.2726, |
| "reward": 0.853125, |
| "reward_std": 0.2143363133072853, |
| "rewards/accuracy_reward": 0.025, |
| "rewards/format_reward": 0.828125, |
| "step": 700 |
| }, |
| { |
| "completion_length": 64.646875, |
| "epoch": 0.7573333333333333, |
| "grad_norm": 0.24161870777606964, |
| "kl": 1.10751953125, |
| "learning_rate": 0.00027929399671762793, |
| "loss": 0.1497, |
| "reward": 0.878125, |
| "reward_std": 0.18096464574337007, |
| "rewards/accuracy_reward": 0.04375, |
| "rewards/format_reward": 0.834375, |
| "step": 710 |
| }, |
| { |
| "completion_length": 65.128125, |
| "epoch": 0.768, |
| "grad_norm": 0.22652657330036163, |
| "kl": 1.7567138671875, |
| "learning_rate": 0.00027833937558488183, |
| "loss": 0.1692, |
| "reward": 0.865625, |
| "reward_std": 0.19575843811035157, |
| "rewards/accuracy_reward": 0.059375, |
| "rewards/format_reward": 0.80625, |
| "step": 720 |
| }, |
| { |
| "completion_length": 76.259375, |
| "epoch": 0.7786666666666666, |
| "grad_norm": 0.46417316794395447, |
| "kl": 3.4798828125, |
| "learning_rate": 0.0002773649502922495, |
| "loss": 0.3618, |
| "reward": 0.7125, |
| "reward_std": 0.31879488229751585, |
| "rewards/accuracy_reward": 0.0125, |
| "rewards/format_reward": 0.7, |
| "step": 730 |
| }, |
| { |
| "completion_length": 64.928125, |
| "epoch": 0.7893333333333333, |
| "grad_norm": 0.8743041753768921, |
| "kl": 2.166650390625, |
| "learning_rate": 0.00027637087120413933, |
| "loss": 0.2562, |
| "reward": 0.840625, |
| "reward_std": 0.2851921945810318, |
| "rewards/accuracy_reward": 0.0375, |
| "rewards/format_reward": 0.803125, |
| "step": 740 |
| }, |
| { |
| "completion_length": 58.028125, |
| "epoch": 0.8, |
| "grad_norm": 0.18655003607273102, |
| "kl": 1.73994140625, |
| "learning_rate": 0.000275357291717754, |
| "loss": 0.191, |
| "reward": 0.909375, |
| "reward_std": 0.19460364878177644, |
| "rewards/accuracy_reward": 0.053125, |
| "rewards/format_reward": 0.85625, |
| "step": 750 |
| }, |
| { |
| "completion_length": 60.803125, |
| "epoch": 0.8106666666666666, |
| "grad_norm": 0.04459076747298241, |
| "kl": 1.7782470703125, |
| "learning_rate": 0.0002743243682394195, |
| "loss": 0.2117, |
| "reward": 0.83125, |
| "reward_std": 0.16213996410369874, |
| "rewards/accuracy_reward": 0.025, |
| "rewards/format_reward": 0.80625, |
| "step": 760 |
| }, |
| { |
| "completion_length": 56.203125, |
| "epoch": 0.8213333333333334, |
| "grad_norm": 0.04220689460635185, |
| "kl": 1.7406494140625, |
| "learning_rate": 0.00027327226016044963, |
| "loss": 0.1999, |
| "reward": 0.878125, |
| "reward_std": 0.1423343911767006, |
| "rewards/accuracy_reward": 0.034375, |
| "rewards/format_reward": 0.84375, |
| "step": 770 |
| }, |
| { |
| "completion_length": 58.375, |
| "epoch": 0.832, |
| "grad_norm": 0.3807085156440735, |
| "kl": 1.8222412109375, |
| "learning_rate": 0.00027220112983255087, |
| "loss": 0.2296, |
| "reward": 0.903125, |
| "reward_std": 0.20482564270496367, |
| "rewards/accuracy_reward": 0.034375, |
| "rewards/format_reward": 0.86875, |
| "step": 780 |
| }, |
| { |
| "completion_length": 63.378125, |
| "epoch": 0.8426666666666667, |
| "grad_norm": 0.01206011138856411, |
| "kl": 2.458740234375, |
| "learning_rate": 0.00027111114254276913, |
| "loss": 0.3096, |
| "reward": 0.84375, |
| "reward_std": 0.2114198923110962, |
| "rewards/accuracy_reward": 0.021875, |
| "rewards/format_reward": 0.821875, |
| "step": 790 |
| }, |
| { |
| "completion_length": 59.7625, |
| "epoch": 0.8533333333333334, |
| "grad_norm": 0.40591439604759216, |
| "kl": 1.378076171875, |
| "learning_rate": 0.00027000246648798456, |
| "loss": 0.1403, |
| "reward": 0.934375, |
| "reward_std": 0.14083535224199295, |
| "rewards/accuracy_reward": 0.03125, |
| "rewards/format_reward": 0.903125, |
| "step": 800 |
| }, |
| { |
| "completion_length": 62.284375, |
| "epoch": 0.864, |
| "grad_norm": 0.27511999011039734, |
| "kl": 2.2107177734375, |
| "learning_rate": 0.0002688752727489565, |
| "loss": 0.2636, |
| "reward": 0.8875, |
| "reward_std": 0.21739855110645295, |
| "rewards/accuracy_reward": 0.028125, |
| "rewards/format_reward": 0.859375, |
| "step": 810 |
| }, |
| { |
| "completion_length": 65.265625, |
| "epoch": 0.8746666666666667, |
| "grad_norm": 0.2582601010799408, |
| "kl": 2.3897705078125, |
| "learning_rate": 0.00026772973526392453, |
| "loss": 0.2965, |
| "reward": 0.83125, |
| "reward_std": 0.2494538262486458, |
| "rewards/accuracy_reward": 0.028125, |
| "rewards/format_reward": 0.803125, |
| "step": 820 |
| }, |
| { |
| "completion_length": 54.865625, |
| "epoch": 0.8853333333333333, |
| "grad_norm": 0.23494267463684082, |
| "kl": 2.6015625, |
| "learning_rate": 0.0002665660308017671, |
| "loss": 0.252, |
| "reward": 0.9, |
| "reward_std": 0.23950843811035155, |
| "rewards/accuracy_reward": 0.04375, |
| "rewards/format_reward": 0.85625, |
| "step": 830 |
| }, |
| { |
| "completion_length": 55.503125, |
| "epoch": 0.896, |
| "grad_norm": 0.20798054337501526, |
| "kl": 1.5889892578125, |
| "learning_rate": 0.000265384338934725, |
| "loss": 0.1996, |
| "reward": 0.9375, |
| "reward_std": 0.20120493620634078, |
| "rewards/accuracy_reward": 0.0625, |
| "rewards/format_reward": 0.875, |
| "step": 840 |
| }, |
| { |
| "completion_length": 59.909375, |
| "epoch": 0.9066666666666666, |
| "grad_norm": 0.23807695508003235, |
| "kl": 1.656982421875, |
| "learning_rate": 0.00026418484201069055, |
| "loss": 0.194, |
| "reward": 0.840625, |
| "reward_std": 0.17524680644273757, |
| "rewards/accuracy_reward": 0.034375, |
| "rewards/format_reward": 0.80625, |
| "step": 850 |
| }, |
| { |
| "completion_length": 56.74375, |
| "epoch": 0.9173333333333333, |
| "grad_norm": 0.21559438109397888, |
| "kl": 0.813427734375, |
| "learning_rate": 0.00026296772512507025, |
| "loss": 0.1054, |
| "reward": 0.884375, |
| "reward_std": 0.13916241526603698, |
| "rewards/accuracy_reward": 0.021875, |
| "rewards/format_reward": 0.8625, |
| "step": 860 |
| }, |
| { |
| "completion_length": 62.390625, |
| "epoch": 0.928, |
| "grad_norm": 0.1291944831609726, |
| "kl": 1.9663330078125, |
| "learning_rate": 0.0002617331760922218, |
| "loss": 0.2316, |
| "reward": 0.85625, |
| "reward_std": 0.15685684233903885, |
| "rewards/accuracy_reward": 0.0125, |
| "rewards/format_reward": 0.84375, |
| "step": 870 |
| }, |
| { |
| "completion_length": 56.009375, |
| "epoch": 0.9386666666666666, |
| "grad_norm": 1.045857548713684, |
| "kl": 1.652001953125, |
| "learning_rate": 0.0002604813854164726, |
| "loss": 0.1616, |
| "reward": 0.9375, |
| "reward_std": 0.16336943507194518, |
| "rewards/accuracy_reward": 0.028125, |
| "rewards/format_reward": 0.909375, |
| "step": 880 |
| }, |
| { |
| "completion_length": 64.446875, |
| "epoch": 0.9493333333333334, |
| "grad_norm": 0.33091413974761963, |
| "kl": 3.235400390625, |
| "learning_rate": 0.0002592125462627231, |
| "loss": 0.3973, |
| "reward": 0.796875, |
| "reward_std": 0.2716366216540337, |
| "rewards/accuracy_reward": 0.015625, |
| "rewards/format_reward": 0.78125, |
| "step": 890 |
| }, |
| { |
| "completion_length": 59.0, |
| "epoch": 0.96, |
| "grad_norm": 0.25974419713020325, |
| "kl": 1.752197265625, |
| "learning_rate": 0.00025792685442663877, |
| "loss": 0.1938, |
| "reward": 0.89375, |
| "reward_std": 0.1826515957713127, |
| "rewards/accuracy_reward": 0.028125, |
| "rewards/format_reward": 0.865625, |
| "step": 900 |
| }, |
| { |
| "completion_length": 57.4125, |
| "epoch": 0.9706666666666667, |
| "grad_norm": 0.2569887936115265, |
| "kl": 2.5720703125, |
| "learning_rate": 0.00025662450830443733, |
| "loss": 0.3213, |
| "reward": 0.846875, |
| "reward_std": 0.22065922170877456, |
| "rewards/accuracy_reward": 0.03125, |
| "rewards/format_reward": 0.815625, |
| "step": 910 |
| }, |
| { |
| "completion_length": 51.140625, |
| "epoch": 0.9813333333333333, |
| "grad_norm": 0.18798935413360596, |
| "kl": 1.6124267578125, |
| "learning_rate": 0.0002553057088622736, |
| "loss": 0.2214, |
| "reward": 0.925, |
| "reward_std": 0.1477062076330185, |
| "rewards/accuracy_reward": 0.015625, |
| "rewards/format_reward": 0.909375, |
| "step": 920 |
| }, |
| { |
| "completion_length": 53.921875, |
| "epoch": 0.992, |
| "grad_norm": 0.8309330940246582, |
| "kl": 1.651806640625, |
| "learning_rate": 0.0002539706596052286, |
| "loss": 0.1893, |
| "reward": 0.909375, |
| "reward_std": 0.16504059880971908, |
| "rewards/accuracy_reward": 0.03125, |
| "rewards/format_reward": 0.878125, |
| "step": 930 |
| }, |
| { |
| "completion_length": 59.96052631578947, |
| "epoch": 1.0021333333333333, |
| "grad_norm": 0.3670661151409149, |
| "kl": 3.8713250411184212, |
| "learning_rate": 0.000252619566545906, |
| "loss": 0.4192, |
| "reward": 0.7796052631578947, |
| "reward_std": 0.2917690135930714, |
| "rewards/accuracy_reward": 0.01644736842105263, |
| "rewards/format_reward": 0.7631578947368421, |
| "step": 940 |
| }, |
| { |
| "completion_length": 59.1125, |
| "epoch": 1.0128, |
| "grad_norm": 0.2787770926952362, |
| "kl": 2.920068359375, |
| "learning_rate": 0.0002512526381726427, |
| "loss": 0.4194, |
| "reward": 0.734375, |
| "reward_std": 0.3439827933907509, |
| "rewards/accuracy_reward": 0.01875, |
| "rewards/format_reward": 0.715625, |
| "step": 950 |
| }, |
| { |
| "completion_length": 57.63125, |
| "epoch": 1.0234666666666667, |
| "grad_norm": 0.15397749841213226, |
| "kl": 2.7442626953125, |
| "learning_rate": 0.00024987008541733663, |
| "loss": 0.3308, |
| "reward": 0.81875, |
| "reward_std": 0.24064744114875794, |
| "rewards/accuracy_reward": 0.0125, |
| "rewards/format_reward": 0.80625, |
| "step": 960 |
| }, |
| { |
| "completion_length": 51.134375, |
| "epoch": 1.0341333333333333, |
| "grad_norm": 0.41957736015319824, |
| "kl": 1.035986328125, |
| "learning_rate": 0.0002484721216228974, |
| "loss": 0.1489, |
| "reward": 0.9625, |
| "reward_std": 0.10561862289905548, |
| "rewards/accuracy_reward": 0.021875, |
| "rewards/format_reward": 0.940625, |
| "step": 970 |
| }, |
| { |
| "completion_length": 56.68125, |
| "epoch": 1.0448, |
| "grad_norm": 0.11578945815563202, |
| "kl": 1.0271240234375, |
| "learning_rate": 0.0002470589625103255, |
| "loss": 0.1162, |
| "reward": 0.9, |
| "reward_std": 0.13415063470602034, |
| "rewards/accuracy_reward": 0.034375, |
| "rewards/format_reward": 0.865625, |
| "step": 980 |
| }, |
| { |
| "completion_length": 55.071875, |
| "epoch": 1.0554666666666668, |
| "grad_norm": 0.006299301981925964, |
| "kl": 1.3626953125, |
| "learning_rate": 0.0002456308261454241, |
| "loss": 0.1452, |
| "reward": 0.909375, |
| "reward_std": 0.1264015957713127, |
| "rewards/accuracy_reward": 0.021875, |
| "rewards/format_reward": 0.8875, |
| "step": 990 |
| }, |
| { |
| "completion_length": 58.0875, |
| "epoch": 1.0661333333333334, |
| "grad_norm": 0.007178621832281351, |
| "kl": 1.4033203125, |
| "learning_rate": 0.00024418793290514906, |
| "loss": 0.1534, |
| "reward": 0.86875, |
| "reward_std": 0.13291241526603698, |
| "rewards/accuracy_reward": 0.01875, |
| "rewards/format_reward": 0.85, |
| "step": 1000 |
| }, |
| { |
| "completion_length": 58.828125, |
| "epoch": 1.0768, |
| "grad_norm": 0.1767469048500061, |
| "kl": 1.26591796875, |
| "learning_rate": 0.0002427305054436024, |
| "loss": 0.1309, |
| "reward": 0.915625, |
| "reward_std": 0.15447435528039932, |
| "rewards/accuracy_reward": 0.034375, |
| "rewards/format_reward": 0.88125, |
| "step": 1010 |
| }, |
| { |
| "completion_length": 60.309375, |
| "epoch": 1.0874666666666666, |
| "grad_norm": 0.20854564011096954, |
| "kl": 2.313330078125, |
| "learning_rate": 0.00024125876865767438, |
| "loss": 0.2191, |
| "reward": 0.90625, |
| "reward_std": 0.17595286518335343, |
| "rewards/accuracy_reward": 0.053125, |
| "rewards/format_reward": 0.853125, |
| "step": 1020 |
| }, |
| { |
| "completion_length": 60.540625, |
| "epoch": 1.0981333333333334, |
| "grad_norm": 0.13962095975875854, |
| "kl": 2.0788818359375, |
| "learning_rate": 0.0002397729496523396, |
| "loss": 0.226, |
| "reward": 0.890625, |
| "reward_std": 0.18282372057437896, |
| "rewards/accuracy_reward": 0.025, |
| "rewards/format_reward": 0.865625, |
| "step": 1030 |
| }, |
| { |
| "completion_length": 58.89375, |
| "epoch": 1.1088, |
| "grad_norm": 0.25990164279937744, |
| "kl": 1.064794921875, |
| "learning_rate": 0.0002382732777056119, |
| "loss": 0.1602, |
| "reward": 0.915625, |
| "reward_std": 0.16433631330728532, |
| "rewards/accuracy_reward": 0.025, |
| "rewards/format_reward": 0.890625, |
| "step": 1040 |
| }, |
| { |
| "completion_length": 58.478125, |
| "epoch": 1.1194666666666666, |
| "grad_norm": 0.21004174649715424, |
| "kl": 1.47158203125, |
| "learning_rate": 0.00023675998423316457, |
| "loss": 0.1682, |
| "reward": 0.953125, |
| "reward_std": 0.18810684233903885, |
| "rewards/accuracy_reward": 0.05, |
| "rewards/format_reward": 0.903125, |
| "step": 1050 |
| }, |
| { |
| "completion_length": 59.39375, |
| "epoch": 1.1301333333333332, |
| "grad_norm": 0.08877279609441757, |
| "kl": 1.388037109375, |
| "learning_rate": 0.00023523330275262037, |
| "loss": 0.1636, |
| "reward": 0.9125, |
| "reward_std": 0.1637136846780777, |
| "rewards/accuracy_reward": 0.0375, |
| "rewards/format_reward": 0.875, |
| "step": 1060 |
| }, |
| { |
| "completion_length": 63.709375, |
| "epoch": 1.1408, |
| "grad_norm": 0.218344584107399, |
| "kl": 1.704052734375, |
| "learning_rate": 0.00023369346884751706, |
| "loss": 0.2163, |
| "reward": 0.884375, |
| "reward_std": 0.23439744114875793, |
| "rewards/accuracy_reward": 0.046875, |
| "rewards/format_reward": 0.8375, |
| "step": 1070 |
| }, |
| { |
| "completion_length": 60.93125, |
| "epoch": 1.1514666666666666, |
| "grad_norm": 0.24491117894649506, |
| "kl": 1.15849609375, |
| "learning_rate": 0.00023214072013095434, |
| "loss": 0.1445, |
| "reward": 0.925, |
| "reward_std": 0.14206304997205735, |
| "rewards/accuracy_reward": 0.034375, |
| "rewards/format_reward": 0.890625, |
| "step": 1080 |
| }, |
| { |
| "completion_length": 55.40625, |
| "epoch": 1.1621333333333332, |
| "grad_norm": 1.4203561544418335, |
| "kl": 1.4958740234375, |
| "learning_rate": 0.00023057529620892773, |
| "loss": 0.2111, |
| "reward": 0.946875, |
| "reward_std": 0.18555223047733307, |
| "rewards/accuracy_reward": 0.040625, |
| "rewards/format_reward": 0.90625, |
| "step": 1090 |
| }, |
| { |
| "completion_length": 61.6375, |
| "epoch": 1.1728, |
| "grad_norm": 0.07859649509191513, |
| "kl": 2.87578125, |
| "learning_rate": 0.00022899743864335462, |
| "loss": 0.3232, |
| "reward": 0.825, |
| "reward_std": 0.22410253882408143, |
| "rewards/accuracy_reward": 0.0125, |
| "rewards/format_reward": 0.8125, |
| "step": 1100 |
| }, |
| { |
| "completion_length": 56.74375, |
| "epoch": 1.1834666666666667, |
| "grad_norm": 0.6798639893531799, |
| "kl": 2.165380859375, |
| "learning_rate": 0.0002274073909147986, |
| "loss": 0.29, |
| "reward": 0.86875, |
| "reward_std": 0.17693375647068024, |
| "rewards/accuracy_reward": 0.0125, |
| "rewards/format_reward": 0.85625, |
| "step": 1110 |
| }, |
| { |
| "completion_length": 56.08125, |
| "epoch": 1.1941333333333333, |
| "grad_norm": 0.46122825145721436, |
| "kl": 2.6029052734375, |
| "learning_rate": 0.000225805398384898, |
| "loss": 0.2877, |
| "reward": 0.85625, |
| "reward_std": 0.15879059880971907, |
| "rewards/accuracy_reward": 0.01875, |
| "rewards/format_reward": 0.8375, |
| "step": 1120 |
| }, |
| { |
| "completion_length": 53.4875, |
| "epoch": 1.2048, |
| "grad_norm": 0.0872046947479248, |
| "kl": 1.120751953125, |
| "learning_rate": 0.0002241917082585036, |
| "loss": 0.1583, |
| "reward": 0.959375, |
| "reward_std": 0.16838996410369872, |
| "rewards/accuracy_reward": 0.046875, |
| "rewards/format_reward": 0.9125, |
| "step": 1130 |
| }, |
| { |
| "completion_length": 52.09375, |
| "epoch": 1.2154666666666667, |
| "grad_norm": 0.2135591208934784, |
| "kl": 1.0230712890625, |
| "learning_rate": 0.00022256656954553245, |
| "loss": 0.1191, |
| "reward": 0.9625, |
| "reward_std": 0.14858439117670058, |
| "rewards/accuracy_reward": 0.05625, |
| "rewards/format_reward": 0.90625, |
| "step": 1140 |
| }, |
| { |
| "completion_length": 50.771875, |
| "epoch": 1.2261333333333333, |
| "grad_norm": 0.2628862261772156, |
| "kl": 2.2327880859375, |
| "learning_rate": 0.00022093023302254295, |
| "loss": 0.2802, |
| "reward": 0.953125, |
| "reward_std": 0.19831304997205734, |
| "rewards/accuracy_reward": 0.053125, |
| "rewards/format_reward": 0.9, |
| "step": 1150 |
| }, |
| { |
| "completion_length": 55.653125, |
| "epoch": 1.2368000000000001, |
| "grad_norm": 0.01043323241174221, |
| "kl": 1.43515625, |
| "learning_rate": 0.0002192829511940371, |
| "loss": 0.216, |
| "reward": 0.871875, |
| "reward_std": 0.17604155987501144, |
| "rewards/accuracy_reward": 0.03125, |
| "rewards/format_reward": 0.840625, |
| "step": 1160 |
| }, |
| { |
| "completion_length": 56.01875, |
| "epoch": 1.2474666666666667, |
| "grad_norm": 0.30781543254852295, |
| "kl": 1.352490234375, |
| "learning_rate": 0.00021762497825349663, |
| "loss": 0.1604, |
| "reward": 0.875, |
| "reward_std": 0.18801814764738084, |
| "rewards/accuracy_reward": 0.040625, |
| "rewards/format_reward": 0.834375, |
| "step": 1170 |
| }, |
| { |
| "completion_length": 57.009375, |
| "epoch": 1.2581333333333333, |
| "grad_norm": 0.268877774477005, |
| "kl": 1.6961669921875, |
| "learning_rate": 0.00021595657004415777, |
| "loss": 0.2207, |
| "reward": 0.896875, |
| "reward_std": 0.18351925760507584, |
| "rewards/accuracy_reward": 0.03125, |
| "rewards/format_reward": 0.865625, |
| "step": 1180 |
| }, |
| { |
| "completion_length": 53.909375, |
| "epoch": 1.2688, |
| "grad_norm": 0.10004394501447678, |
| "kl": 1.055419921875, |
| "learning_rate": 0.00021427798401953233, |
| "loss": 0.1206, |
| "reward": 0.915625, |
| "reward_std": 0.10359617173671723, |
| "rewards/accuracy_reward": 0.0375, |
| "rewards/format_reward": 0.878125, |
| "step": 1190 |
| }, |
| { |
| "completion_length": 55.93125, |
| "epoch": 1.2794666666666665, |
| "grad_norm": 0.16535454988479614, |
| "kl": 2.3431884765625, |
| "learning_rate": 0.0002125894792036794, |
| "loss": 0.3288, |
| "reward": 0.903125, |
| "reward_std": 0.22568152397871016, |
| "rewards/accuracy_reward": 0.0375, |
| "rewards/format_reward": 0.865625, |
| "step": 1200 |
| }, |
| { |
| "completion_length": 55.09375, |
| "epoch": 1.2901333333333334, |
| "grad_norm": 0.19274021685123444, |
| "kl": 1.0988037109375, |
| "learning_rate": 0.0002108913161512354, |
| "loss": 0.1432, |
| "reward": 0.9375, |
| "reward_std": 0.12358439117670059, |
| "rewards/accuracy_reward": 0.034375, |
| "rewards/format_reward": 0.903125, |
| "step": 1210 |
| }, |
| { |
| "completion_length": 55.675, |
| "epoch": 1.3008, |
| "grad_norm": 0.15594810247421265, |
| "kl": 1.42451171875, |
| "learning_rate": 0.0002091837569072076, |
| "loss": 0.1693, |
| "reward": 0.94375, |
| "reward_std": 0.16636751294136048, |
| "rewards/accuracy_reward": 0.046875, |
| "rewards/format_reward": 0.896875, |
| "step": 1220 |
| }, |
| { |
| "completion_length": 56.88125, |
| "epoch": 1.3114666666666666, |
| "grad_norm": 0.3196319341659546, |
| "kl": 1.6291015625, |
| "learning_rate": 0.00020746706496653765, |
| "loss": 0.2144, |
| "reward": 0.915625, |
| "reward_std": 0.1927691087126732, |
| "rewards/accuracy_reward": 0.046875, |
| "rewards/format_reward": 0.86875, |
| "step": 1230 |
| }, |
| { |
| "completion_length": 58.7875, |
| "epoch": 1.3221333333333334, |
| "grad_norm": 0.13602705299854279, |
| "kl": 1.5197509765625, |
| "learning_rate": 0.00020574150523344152, |
| "loss": 0.1651, |
| "reward": 0.94375, |
| "reward_std": 0.16670301407575608, |
| "rewards/accuracy_reward": 0.0375, |
| "rewards/format_reward": 0.90625, |
| "step": 1240 |
| }, |
| { |
| "completion_length": 62.996875, |
| "epoch": 1.3328, |
| "grad_norm": 0.05853046849370003, |
| "kl": 1.815576171875, |
| "learning_rate": 0.00020400734398053186, |
| "loss": 0.1795, |
| "reward": 0.86875, |
| "reward_std": 0.19752006977796555, |
| "rewards/accuracy_reward": 0.0375, |
| "rewards/format_reward": 0.83125, |
| "step": 1250 |
| }, |
| { |
| "completion_length": 62.54375, |
| "epoch": 1.3434666666666666, |
| "grad_norm": 0.007495929021388292, |
| "kl": 1.3890869140625, |
| "learning_rate": 0.0002022648488077294, |
| "loss": 0.1695, |
| "reward": 0.884375, |
| "reward_std": 0.1775405988097191, |
| "rewards/accuracy_reward": 0.028125, |
| "rewards/format_reward": 0.85625, |
| "step": 1260 |
| }, |
| { |
| "completion_length": 61.278125, |
| "epoch": 1.3541333333333334, |
| "grad_norm": 0.1842016726732254, |
| "kl": 1.7712646484375, |
| "learning_rate": 0.0002005142886009691, |
| "loss": 0.2379, |
| "reward": 0.875, |
| "reward_std": 0.16706304997205734, |
| "rewards/accuracy_reward": 0.021875, |
| "rewards/format_reward": 0.853125, |
| "step": 1270 |
| }, |
| { |
| "completion_length": 62.025, |
| "epoch": 1.3648, |
| "grad_norm": 0.16039888560771942, |
| "kl": 1.9799560546875, |
| "learning_rate": 0.00019875593349070832, |
| "loss": 0.2323, |
| "reward": 0.91875, |
| "reward_std": 0.20685684233903884, |
| "rewards/accuracy_reward": 0.046875, |
| "rewards/format_reward": 0.871875, |
| "step": 1280 |
| }, |
| { |
| "completion_length": 61.0375, |
| "epoch": 1.3754666666666666, |
| "grad_norm": 0.15333615243434906, |
| "kl": 2.3615478515625, |
| "learning_rate": 0.0001969900548102427, |
| "loss": 0.2778, |
| "reward": 0.84375, |
| "reward_std": 0.20719234347343446, |
| "rewards/accuracy_reward": 0.01875, |
| "rewards/format_reward": 0.825, |
| "step": 1290 |
| }, |
| { |
| "completion_length": 58.328125, |
| "epoch": 1.3861333333333334, |
| "grad_norm": 0.07369455695152283, |
| "kl": 2.516748046875, |
| "learning_rate": 0.00019521692505383657, |
| "loss": 0.3136, |
| "reward": 0.85, |
| "reward_std": 0.19249776750802994, |
| "rewards/accuracy_reward": 0.025, |
| "rewards/format_reward": 0.825, |
| "step": 1300 |
| }, |
| { |
| "completion_length": 55.35625, |
| "epoch": 1.3968, |
| "grad_norm": 0.2832612693309784, |
| "kl": 1.73642578125, |
| "learning_rate": 0.000193436817834674, |
| "loss": 0.2319, |
| "reward": 0.925, |
| "reward_std": 0.21794123351573944, |
| "rewards/accuracy_reward": 0.034375, |
| "rewards/format_reward": 0.890625, |
| "step": 1310 |
| }, |
| { |
| "completion_length": 59.0875, |
| "epoch": 1.4074666666666666, |
| "grad_norm": 0.24120619893074036, |
| "kl": 2.8723876953125, |
| "learning_rate": 0.0001916500078426373, |
| "loss": 0.3392, |
| "reward": 0.8375, |
| "reward_std": 0.22023502588272095, |
| "rewards/accuracy_reward": 0.025, |
| "rewards/format_reward": 0.8125, |
| "step": 1320 |
| }, |
| { |
| "completion_length": 55.025, |
| "epoch": 1.4181333333333335, |
| "grad_norm": 0.18106360733509064, |
| "kl": 1.674365234375, |
| "learning_rate": 0.0001898567708019196, |
| "loss": 0.2313, |
| "reward": 0.88125, |
| "reward_std": 0.17693375647068024, |
| "rewards/accuracy_reward": 0.01875, |
| "rewards/format_reward": 0.8625, |
| "step": 1330 |
| }, |
| { |
| "completion_length": 55.15625, |
| "epoch": 1.4288, |
| "grad_norm": 0.1365566849708557, |
| "kl": 2.2543212890625, |
| "learning_rate": 0.00018805738342847727, |
| "loss": 0.3175, |
| "reward": 0.878125, |
| "reward_std": 0.2048343911767006, |
| "rewards/accuracy_reward": 0.028125, |
| "rewards/format_reward": 0.85, |
| "step": 1340 |
| }, |
| { |
| "completion_length": 55.5625, |
| "epoch": 1.4394666666666667, |
| "grad_norm": 0.10590548813343048, |
| "kl": 1.425244140625, |
| "learning_rate": 0.00018625212338733, |
| "loss": 0.1371, |
| "reward": 0.8875, |
| "reward_std": 0.10915063470602035, |
| "rewards/accuracy_reward": 0.028125, |
| "rewards/format_reward": 0.859375, |
| "step": 1350 |
| }, |
| { |
| "completion_length": 53.225, |
| "epoch": 1.4501333333333333, |
| "grad_norm": 0.15103192627429962, |
| "kl": 1.1132080078125, |
| "learning_rate": 0.00018444126924971387, |
| "loss": 0.1228, |
| "reward": 0.953125, |
| "reward_std": 0.11838996410369873, |
| "rewards/accuracy_reward": 0.0375, |
| "rewards/format_reward": 0.915625, |
| "step": 1360 |
| }, |
| { |
| "completion_length": 57.265625, |
| "epoch": 1.4607999999999999, |
| "grad_norm": 0.05684982240200043, |
| "kl": 1.13671875, |
| "learning_rate": 0.0001826251004500947, |
| "loss": 0.1437, |
| "reward": 0.9125, |
| "reward_std": 0.19223694801330565, |
| "rewards/accuracy_reward": 0.065625, |
| "rewards/format_reward": 0.846875, |
| "step": 1370 |
| }, |
| { |
| "completion_length": 59.325, |
| "epoch": 1.4714666666666667, |
| "grad_norm": 0.17307031154632568, |
| "kl": 1.8568115234375, |
| "learning_rate": 0.0001808038972430486, |
| "loss": 0.2279, |
| "reward": 0.871875, |
| "reward_std": 0.16398502588272096, |
| "rewards/accuracy_reward": 0.01875, |
| "rewards/format_reward": 0.853125, |
| "step": 1380 |
| }, |
| { |
| "completion_length": 59.646875, |
| "epoch": 1.4821333333333333, |
| "grad_norm": 0.007796150632202625, |
| "kl": 2.4506591796875, |
| "learning_rate": 0.00017897794066001524, |
| "loss": 0.2992, |
| "reward": 0.84375, |
| "reward_std": 0.20420301407575608, |
| "rewards/accuracy_reward": 0.034375, |
| "rewards/format_reward": 0.809375, |
| "step": 1390 |
| }, |
| { |
| "completion_length": 58.3, |
| "epoch": 1.4928, |
| "grad_norm": 0.1803148239850998, |
| "kl": 1.1746826171875, |
| "learning_rate": 0.00017714751246593197, |
| "loss": 0.1374, |
| "reward": 0.85625, |
| "reward_std": 0.09665063470602035, |
| "rewards/accuracy_reward": 0.0125, |
| "rewards/format_reward": 0.84375, |
| "step": 1400 |
| }, |
| { |
| "completion_length": 51.30625, |
| "epoch": 1.5034666666666667, |
| "grad_norm": 0.12448029220104218, |
| "kl": 1.7854248046875, |
| "learning_rate": 0.00017531289511575425, |
| "loss": 0.2174, |
| "reward": 0.940625, |
| "reward_std": 0.17032372057437897, |
| "rewards/accuracy_reward": 0.04375, |
| "rewards/format_reward": 0.896875, |
| "step": 1410 |
| }, |
| { |
| "completion_length": 50.446875, |
| "epoch": 1.5141333333333333, |
| "grad_norm": 0.1189781054854393, |
| "kl": 1.4447998046875, |
| "learning_rate": 0.0001734743717108699, |
| "loss": 0.1672, |
| "reward": 0.959375, |
| "reward_std": 0.13282372057437897, |
| "rewards/accuracy_reward": 0.0375, |
| "rewards/format_reward": 0.921875, |
| "step": 1420 |
| }, |
| { |
| "completion_length": 55.278125, |
| "epoch": 1.5248, |
| "grad_norm": 0.10093328356742859, |
| "kl": 1.5136474609375, |
| "learning_rate": 0.0001716322259554132, |
| "loss": 0.1768, |
| "reward": 0.96875, |
| "reward_std": 0.19523502588272096, |
| "rewards/accuracy_reward": 0.065625, |
| "rewards/format_reward": 0.903125, |
| "step": 1430 |
| }, |
| { |
| "completion_length": 59.934375, |
| "epoch": 1.5354666666666668, |
| "grad_norm": 0.2610551714897156, |
| "kl": 1.778369140625, |
| "learning_rate": 0.00016978674211248673, |
| "loss": 0.2314, |
| "reward": 0.84375, |
| "reward_std": 0.20420301407575608, |
| "rewards/accuracy_reward": 0.03125, |
| "rewards/format_reward": 0.8125, |
| "step": 1440 |
| }, |
| { |
| "completion_length": 54.96875, |
| "epoch": 1.5461333333333334, |
| "grad_norm": 0.19814546406269073, |
| "kl": 2.6193603515625, |
| "learning_rate": 0.00016793820496029623, |
| "loss": 0.3738, |
| "reward": 0.88125, |
| "reward_std": 0.2356409251689911, |
| "rewards/accuracy_reward": 0.025, |
| "rewards/format_reward": 0.85625, |
| "step": 1450 |
| }, |
| { |
| "completion_length": 51.703125, |
| "epoch": 1.5568, |
| "grad_norm": 0.1247173473238945, |
| "kl": 2.2877197265625, |
| "learning_rate": 0.000166086899748206, |
| "loss": 0.236, |
| "reward": 0.940625, |
| "reward_std": 0.2197028651833534, |
| "rewards/accuracy_reward": 0.053125, |
| "rewards/format_reward": 0.8875, |
| "step": 1460 |
| }, |
| { |
| "completion_length": 51.096875, |
| "epoch": 1.5674666666666668, |
| "grad_norm": 0.16843904554843903, |
| "kl": 1.2693115234375, |
| "learning_rate": 0.0001642331121527223, |
| "loss": 0.1801, |
| "reward": 0.959375, |
| "reward_std": 0.14867308586835862, |
| "rewards/accuracy_reward": 0.04375, |
| "rewards/format_reward": 0.915625, |
| "step": 1470 |
| }, |
| { |
| "completion_length": 55.465625, |
| "epoch": 1.5781333333333334, |
| "grad_norm": 0.12530925869941711, |
| "kl": 1.403857421875, |
| "learning_rate": 0.0001623771282334099, |
| "loss": 0.1621, |
| "reward": 0.921875, |
| "reward_std": 0.17568152397871017, |
| "rewards/accuracy_reward": 0.046875, |
| "rewards/format_reward": 0.875, |
| "step": 1480 |
| }, |
| { |
| "completion_length": 58.23125, |
| "epoch": 1.5888, |
| "grad_norm": 0.016702894121408463, |
| "kl": 0.8815673828125, |
| "learning_rate": 0.00016051923438875035, |
| "loss": 0.0918, |
| "reward": 0.8875, |
| "reward_std": 0.0879347175359726, |
| "rewards/accuracy_reward": 0.034375, |
| "rewards/format_reward": 0.853125, |
| "step": 1490 |
| }, |
| { |
| "completion_length": 53.565625, |
| "epoch": 1.5994666666666668, |
| "grad_norm": 0.09671846032142639, |
| "kl": 0.7630126953125, |
| "learning_rate": 0.00015865971731194738, |
| "loss": 0.0861, |
| "reward": 0.98125, |
| "reward_std": 0.11971687823534012, |
| "rewards/accuracy_reward": 0.040625, |
| "rewards/format_reward": 0.940625, |
| "step": 1500 |
| }, |
| { |
| "completion_length": 57.234375, |
| "epoch": 1.6101333333333332, |
| "grad_norm": 0.16410210728645325, |
| "kl": 0.922265625, |
| "learning_rate": 0.00015679886394668707, |
| "loss": 0.1231, |
| "reward": 0.925, |
| "reward_std": 0.10915063470602035, |
| "rewards/accuracy_reward": 0.028125, |
| "rewards/format_reward": 0.896875, |
| "step": 1510 |
| }, |
| { |
| "completion_length": 95.821875, |
| "epoch": 1.6208, |
| "grad_norm": 0.09543804824352264, |
| "kl": 6.0548583984375, |
| "learning_rate": 0.00015493696144285935, |
| "loss": 0.4094, |
| "reward": 0.284375, |
| "reward_std": 0.2681046098470688, |
| "rewards/accuracy_reward": 0.025, |
| "rewards/format_reward": 0.259375, |
| "step": 1520 |
| }, |
| { |
| "completion_length": 113.446875, |
| "epoch": 1.6314666666666666, |
| "grad_norm": 0.006384687032550573, |
| "kl": 0.34674072265625, |
| "learning_rate": 0.00015307429711224754, |
| "loss": 0.0323, |
| "reward": 0.05, |
| "reward_std": 0.08221687823534012, |
| "rewards/accuracy_reward": 0.03125, |
| "rewards/format_reward": 0.01875, |
| "step": 1530 |
| }, |
| { |
| "completion_length": 116.7, |
| "epoch": 1.6421333333333332, |
| "grad_norm": 0.03033365309238434, |
| "kl": 0.18465576171875, |
| "learning_rate": 0.0001512111583841933, |
| "loss": 0.0367, |
| "reward": 0.053125, |
| "reward_std": 0.10625, |
| "rewards/accuracy_reward": 0.028125, |
| "rewards/format_reward": 0.025, |
| "step": 1540 |
| }, |
| { |
| "completion_length": 117.84375, |
| "epoch": 1.6528, |
| "grad_norm": 0.007193129975348711, |
| "kl": 0.1654541015625, |
| "learning_rate": 0.00014934783276124278, |
| "loss": 0.0284, |
| "reward": 0.059375, |
| "reward_std": 0.08318375647068024, |
| "rewards/accuracy_reward": 0.025, |
| "rewards/format_reward": 0.034375, |
| "step": 1550 |
| }, |
| { |
| "completion_length": 111.459375, |
| "epoch": 1.6634666666666666, |
| "grad_norm": 0.024860132485628128, |
| "kl": 0.17579345703125, |
| "learning_rate": 0.00014748460777478208, |
| "loss": 0.0751, |
| "reward": 0.13125, |
| "reward_std": 0.20580126941204072, |
| "rewards/accuracy_reward": 0.009375, |
| "rewards/format_reward": 0.121875, |
| "step": 1560 |
| }, |
| { |
| "completion_length": 91.90625, |
| "epoch": 1.6741333333333333, |
| "grad_norm": 0.07996781170368195, |
| "kl": 0.19302978515625, |
| "learning_rate": 0.00014562177094066812, |
| "loss": 0.1666, |
| "reward": 0.478125, |
| "reward_std": 0.42759600281715393, |
| "rewards/accuracy_reward": 0.021875, |
| "rewards/format_reward": 0.45625, |
| "step": 1570 |
| }, |
| { |
| "completion_length": 73.571875, |
| "epoch": 1.6848, |
| "grad_norm": 0.07270823419094086, |
| "kl": 0.2430419921875, |
| "learning_rate": 0.0001437596097148615, |
| "loss": 0.1744, |
| "reward": 0.76875, |
| "reward_std": 0.332449671626091, |
| "rewards/accuracy_reward": 0.028125, |
| "rewards/format_reward": 0.740625, |
| "step": 1580 |
| }, |
| { |
| "completion_length": 64.4875, |
| "epoch": 1.6954666666666667, |
| "grad_norm": 0.1185784786939621, |
| "kl": 0.27333984375, |
| "learning_rate": 0.00014189841144906926, |
| "loss": 0.1684, |
| "reward": 0.80625, |
| "reward_std": 0.28343056291341784, |
| "rewards/accuracy_reward": 0.028125, |
| "rewards/format_reward": 0.778125, |
| "step": 1590 |
| }, |
| { |
| "completion_length": 59.8375, |
| "epoch": 1.7061333333333333, |
| "grad_norm": 0.30180656909942627, |
| "kl": 0.5548828125, |
| "learning_rate": 0.00014003846334640323, |
| "loss": 0.2054, |
| "reward": 0.740625, |
| "reward_std": 0.29256718456745145, |
| "rewards/accuracy_reward": 0.0125, |
| "rewards/format_reward": 0.728125, |
| "step": 1600 |
| }, |
| { |
| "completion_length": 59.284375, |
| "epoch": 1.7168, |
| "grad_norm": 0.26433998346328735, |
| "kl": 2.827392578125, |
| "learning_rate": 0.00013818005241706145, |
| "loss": 0.469, |
| "reward": 0.70625, |
| "reward_std": 0.33853629529476165, |
| "rewards/accuracy_reward": 0.00625, |
| "rewards/format_reward": 0.7, |
| "step": 1610 |
| }, |
| { |
| "completion_length": 53.78125, |
| "epoch": 1.7274666666666667, |
| "grad_norm": 0.2971569299697876, |
| "kl": 3.8951171875, |
| "learning_rate": 0.00013632346543403947, |
| "loss": 0.451, |
| "reward": 0.81875, |
| "reward_std": 0.23273502588272094, |
| "rewards/accuracy_reward": 0.0125, |
| "rewards/format_reward": 0.80625, |
| "step": 1620 |
| }, |
| { |
| "completion_length": 53.903125, |
| "epoch": 1.7381333333333333, |
| "grad_norm": 0.0691404640674591, |
| "kl": 1.3744140625, |
| "learning_rate": 0.00013446898888887804, |
| "loss": 0.1657, |
| "reward": 0.93125, |
| "reward_std": 0.13943375647068024, |
| "rewards/accuracy_reward": 0.015625, |
| "rewards/format_reward": 0.915625, |
| "step": 1630 |
| }, |
| { |
| "completion_length": 57.153125, |
| "epoch": 1.7488000000000001, |
| "grad_norm": 0.03758076950907707, |
| "kl": 0.640966796875, |
| "learning_rate": 0.00013261690894745442, |
| "loss": 0.0775, |
| "reward": 0.884375, |
| "reward_std": 0.07596687823534012, |
| "rewards/accuracy_reward": 0.0125, |
| "rewards/format_reward": 0.871875, |
| "step": 1640 |
| }, |
| { |
| "completion_length": 58.0125, |
| "epoch": 1.7594666666666665, |
| "grad_norm": 0.009222053922712803, |
| "kl": 1.1447265625, |
| "learning_rate": 0.00013076751140582394, |
| "loss": 0.1472, |
| "reward": 0.88125, |
| "reward_std": 0.14963996410369873, |
| "rewards/accuracy_reward": 0.05, |
| "rewards/format_reward": 0.83125, |
| "step": 1650 |
| }, |
| { |
| "completion_length": 55.746875, |
| "epoch": 1.7701333333333333, |
| "grad_norm": 0.029693789780139923, |
| "kl": 1.351123046875, |
| "learning_rate": 0.00012892108164611857, |
| "loss": 0.161, |
| "reward": 0.890625, |
| "reward_std": 0.14867308586835862, |
| "rewards/accuracy_reward": 0.021875, |
| "rewards/format_reward": 0.86875, |
| "step": 1660 |
| }, |
| { |
| "completion_length": 53.184375, |
| "epoch": 1.7808000000000002, |
| "grad_norm": 0.23443636298179626, |
| "kl": 1.46533203125, |
| "learning_rate": 0.00012707790459250904, |
| "loss": 0.1583, |
| "reward": 0.9625, |
| "reward_std": 0.14665063470602036, |
| "rewards/accuracy_reward": 0.034375, |
| "rewards/format_reward": 0.928125, |
| "step": 1670 |
| }, |
| { |
| "completion_length": 52.959375, |
| "epoch": 1.7914666666666665, |
| "grad_norm": 0.010363437235355377, |
| "kl": 1.73623046875, |
| "learning_rate": 0.0001252382646672384, |
| "loss": 0.1757, |
| "reward": 0.921875, |
| "reward_std": 0.16504059880971908, |
| "rewards/accuracy_reward": 0.021875, |
| "rewards/format_reward": 0.9, |
| "step": 1680 |
| }, |
| { |
| "completion_length": 55.971875, |
| "epoch": 1.8021333333333334, |
| "grad_norm": 0.0745161771774292, |
| "kl": 1.418701171875, |
| "learning_rate": 0.00012340244574673238, |
| "loss": 0.1882, |
| "reward": 0.91875, |
| "reward_std": 0.1851816728711128, |
| "rewards/accuracy_reward": 0.046875, |
| "rewards/format_reward": 0.871875, |
| "step": 1690 |
| }, |
| { |
| "completion_length": 57.903125, |
| "epoch": 1.8128, |
| "grad_norm": 0.09005508571863174, |
| "kl": 1.6597900390625, |
| "learning_rate": 0.000121570731117794, |
| "loss": 0.1998, |
| "reward": 0.8875, |
| "reward_std": 0.1477808892726898, |
| "rewards/accuracy_reward": 0.01875, |
| "rewards/format_reward": 0.86875, |
| "step": 1700 |
| }, |
| { |
| "completion_length": 58.240625, |
| "epoch": 1.8234666666666666, |
| "grad_norm": 0.09877178072929382, |
| "kl": 1.344873046875, |
| "learning_rate": 0.00011974340343388972, |
| "loss": 0.163, |
| "reward": 0.921875, |
| "reward_std": 0.18511751294136047, |
| "rewards/accuracy_reward": 0.034375, |
| "rewards/format_reward": 0.8875, |
| "step": 1710 |
| }, |
| { |
| "completion_length": 66.65, |
| "epoch": 1.8341333333333334, |
| "grad_norm": 0.09796544909477234, |
| "kl": 2.161865234375, |
| "learning_rate": 0.00011792074467153248, |
| "loss": 0.2649, |
| "reward": 0.81875, |
| "reward_std": 0.20843056291341783, |
| "rewards/accuracy_reward": 0.034375, |
| "rewards/format_reward": 0.784375, |
| "step": 1720 |
| }, |
| { |
| "completion_length": 65.9, |
| "epoch": 1.8448, |
| "grad_norm": 0.12498176097869873, |
| "kl": 2.451318359375, |
| "learning_rate": 0.00011610303608677008, |
| "loss": 0.3047, |
| "reward": 0.809375, |
| "reward_std": 0.268188039958477, |
| "rewards/accuracy_reward": 0.040625, |
| "rewards/format_reward": 0.76875, |
| "step": 1730 |
| }, |
| { |
| "completion_length": 61.346875, |
| "epoch": 1.8554666666666666, |
| "grad_norm": 0.40386486053466797, |
| "kl": 3.258935546875, |
| "learning_rate": 0.00011429055817178411, |
| "loss": 0.3857, |
| "reward": 0.846875, |
| "reward_std": 0.28369315564632414, |
| "rewards/accuracy_reward": 0.028125, |
| "rewards/format_reward": 0.81875, |
| "step": 1740 |
| }, |
| { |
| "completion_length": 58.859375, |
| "epoch": 1.8661333333333334, |
| "grad_norm": 0.12184485048055649, |
| "kl": 3.230029296875, |
| "learning_rate": 0.00011248359061160698, |
| "loss": 0.3751, |
| "reward": 0.86875, |
| "reward_std": 0.2617799282073975, |
| "rewards/accuracy_reward": 0.03125, |
| "rewards/format_reward": 0.8375, |
| "step": 1750 |
| }, |
| { |
| "completion_length": 57.69375, |
| "epoch": 1.8768, |
| "grad_norm": 0.1783817708492279, |
| "kl": 2.4540283203125, |
| "learning_rate": 0.00011068241224096347, |
| "loss": 0.2785, |
| "reward": 0.875, |
| "reward_std": 0.20430223047733306, |
| "rewards/accuracy_reward": 0.01875, |
| "rewards/format_reward": 0.85625, |
| "step": 1760 |
| }, |
| { |
| "completion_length": 64.459375, |
| "epoch": 1.8874666666666666, |
| "grad_norm": 0.038270145654678345, |
| "kl": 2.579345703125, |
| "learning_rate": 0.00010888730100124353, |
| "loss": 0.2952, |
| "reward": 0.815625, |
| "reward_std": 0.24689744114875795, |
| "rewards/accuracy_reward": 0.025, |
| "rewards/format_reward": 0.790625, |
| "step": 1770 |
| }, |
| { |
| "completion_length": 61.403125, |
| "epoch": 1.8981333333333335, |
| "grad_norm": 0.07670488953590393, |
| "kl": 2.2768310546875, |
| "learning_rate": 0.00010709853389761286, |
| "loss": 0.3084, |
| "reward": 0.884375, |
| "reward_std": 0.23545301407575608, |
| "rewards/accuracy_reward": 0.0375, |
| "rewards/format_reward": 0.846875, |
| "step": 1780 |
| }, |
| { |
| "completion_length": 62.98125, |
| "epoch": 1.9088, |
| "grad_norm": 0.21110066771507263, |
| "kl": 3.0247314453125, |
| "learning_rate": 0.00010531638695626811, |
| "loss": 0.3866, |
| "reward": 0.8, |
| "reward_std": 0.2520918682217598, |
| "rewards/accuracy_reward": 0.03125, |
| "rewards/format_reward": 0.76875, |
| "step": 1790 |
| }, |
| { |
| "completion_length": 64.06875, |
| "epoch": 1.9194666666666667, |
| "grad_norm": 0.19934044778347015, |
| "kl": 4.1655029296875, |
| "learning_rate": 0.00010354113518184303, |
| "loss": 0.4661, |
| "reward": 0.784375, |
| "reward_std": 0.2908942475914955, |
| "rewards/accuracy_reward": 0.028125, |
| "rewards/format_reward": 0.75625, |
| "step": 1800 |
| }, |
| { |
| "completion_length": 57.765625, |
| "epoch": 1.9301333333333335, |
| "grad_norm": 0.07659115642309189, |
| "kl": 1.541015625, |
| "learning_rate": 0.000101773052514972, |
| "loss": 0.1994, |
| "reward": 0.865625, |
| "reward_std": 0.17630237936973572, |
| "rewards/accuracy_reward": 0.025, |
| "rewards/format_reward": 0.840625, |
| "step": 1810 |
| }, |
| { |
| "completion_length": 58.328125, |
| "epoch": 1.9407999999999999, |
| "grad_norm": 0.053584493696689606, |
| "kl": 1.616796875, |
| "learning_rate": 0.00010001241179001836, |
| "loss": 0.2072, |
| "reward": 0.88125, |
| "reward_std": 0.15386751294136047, |
| "rewards/accuracy_reward": 0.0125, |
| "rewards/format_reward": 0.86875, |
| "step": 1820 |
| }, |
| { |
| "completion_length": 59.36875, |
| "epoch": 1.9514666666666667, |
| "grad_norm": 0.06573835760354996, |
| "kl": 1.567236328125, |
| "learning_rate": 9.825948469297301e-05, |
| "loss": 0.2048, |
| "reward": 0.921875, |
| "reward_std": 0.19004059880971907, |
| "rewards/accuracy_reward": 0.053125, |
| "rewards/format_reward": 0.86875, |
| "step": 1830 |
| }, |
| { |
| "completion_length": 55.890625, |
| "epoch": 1.9621333333333333, |
| "grad_norm": 0.18969739973545074, |
| "kl": 1.7246337890625, |
| "learning_rate": 9.651454171953012e-05, |
| "loss": 0.2224, |
| "reward": 0.946875, |
| "reward_std": 0.1664562076330185, |
| "rewards/accuracy_reward": 0.040625, |
| "rewards/format_reward": 0.90625, |
| "step": 1840 |
| }, |
| { |
| "completion_length": 60.59375, |
| "epoch": 1.9727999999999999, |
| "grad_norm": 0.04010459780693054, |
| "kl": 2.70771484375, |
| "learning_rate": 9.477785213334706e-05, |
| "loss": 0.3228, |
| "reward": 0.834375, |
| "reward_std": 0.22910557091236114, |
| "rewards/accuracy_reward": 0.025, |
| "rewards/format_reward": 0.809375, |
| "step": 1850 |
| }, |
| { |
| "completion_length": 63.846875, |
| "epoch": 1.9834666666666667, |
| "grad_norm": 0.2044885903596878, |
| "kl": 3.81484375, |
| "learning_rate": 9.30496839244936e-05, |
| "loss": 0.4808, |
| "reward": 0.753125, |
| "reward_std": 0.307637582719326, |
| "rewards/accuracy_reward": 0.028125, |
| "rewards/format_reward": 0.725, |
| "step": 1860 |
| }, |
| { |
| "completion_length": 63.05, |
| "epoch": 1.9941333333333333, |
| "grad_norm": 0.32608747482299805, |
| "kl": 2.995166015625, |
| "learning_rate": 9.133030376809867e-05, |
| "loss": 0.4066, |
| "reward": 0.734375, |
| "reward_std": 0.31027562469244, |
| "rewards/accuracy_reward": 0.009375, |
| "rewards/format_reward": 0.725, |
| "step": 1870 |
| }, |
| { |
| "completion_length": 62.14473684210526, |
| "epoch": 2.0042666666666666, |
| "grad_norm": 0.19676542282104492, |
| "kl": 2.9421258223684212, |
| "learning_rate": 8.961997698319912e-05, |
| "loss": 0.3671, |
| "reward": 0.7861842105263158, |
| "reward_std": 0.24553519801089638, |
| "rewards/accuracy_reward": 0.013157894736842105, |
| "rewards/format_reward": 0.7730263157894737, |
| "step": 1880 |
| }, |
| { |
| "completion_length": 58.6125, |
| "epoch": 2.0149333333333335, |
| "grad_norm": 0.3623450994491577, |
| "kl": 1.9839111328125, |
| "learning_rate": 8.79189674917983e-05, |
| "loss": 0.2562, |
| "reward": 0.840625, |
| "reward_std": 0.20684282928705217, |
| "rewards/accuracy_reward": 0.03125, |
| "rewards/format_reward": 0.809375, |
| "step": 1890 |
| }, |
| { |
| "completion_length": 58.20625, |
| "epoch": 2.0256, |
| "grad_norm": 0.14830775558948517, |
| "kl": 1.731103515625, |
| "learning_rate": 8.622753777813978e-05, |
| "loss": 0.2133, |
| "reward": 0.909375, |
| "reward_std": 0.16573613584041597, |
| "rewards/accuracy_reward": 0.0625, |
| "rewards/format_reward": 0.846875, |
| "step": 1900 |
| }, |
| { |
| "completion_length": 55.796875, |
| "epoch": 2.0362666666666667, |
| "grad_norm": 0.1913866549730301, |
| "kl": 1.334716796875, |
| "learning_rate": 8.454594884820358e-05, |
| "loss": 0.1778, |
| "reward": 0.921875, |
| "reward_std": 0.14040063470602035, |
| "rewards/accuracy_reward": 0.01875, |
| "rewards/format_reward": 0.903125, |
| "step": 1910 |
| }, |
| { |
| "completion_length": 55.759375, |
| "epoch": 2.0469333333333335, |
| "grad_norm": 0.21011610329151154, |
| "kl": 1.82216796875, |
| "learning_rate": 8.287446018942971e-05, |
| "loss": 0.204, |
| "reward": 0.94375, |
| "reward_std": 0.16476925760507583, |
| "rewards/accuracy_reward": 0.03125, |
| "rewards/format_reward": 0.9125, |
| "step": 1920 |
| }, |
| { |
| "completion_length": 60.0, |
| "epoch": 2.0576, |
| "grad_norm": 0.28949812054634094, |
| "kl": 2.074462890625, |
| "learning_rate": 8.121332973067665e-05, |
| "loss": 0.2961, |
| "reward": 0.871875, |
| "reward_std": 0.19040063470602037, |
| "rewards/accuracy_reward": 0.025, |
| "rewards/format_reward": 0.846875, |
| "step": 1930 |
| }, |
| { |
| "completion_length": 61.309375, |
| "epoch": 2.0682666666666667, |
| "grad_norm": 0.24213965237140656, |
| "kl": 2.480908203125, |
| "learning_rate": 7.956281380242029e-05, |
| "loss": 0.29, |
| "reward": 0.81875, |
| "reward_std": 0.21265811175107957, |
| "rewards/accuracy_reward": 0.03125, |
| "rewards/format_reward": 0.7875, |
| "step": 1940 |
| }, |
| { |
| "completion_length": 56.003125, |
| "epoch": 2.0789333333333335, |
| "grad_norm": 0.20885713398456573, |
| "kl": 2.0755615234375, |
| "learning_rate": 7.792316709719874e-05, |
| "loss": 0.287, |
| "reward": 0.925, |
| "reward_std": 0.22208535224199294, |
| "rewards/accuracy_reward": 0.046875, |
| "rewards/format_reward": 0.878125, |
| "step": 1950 |
| }, |
| { |
| "completion_length": 59.065625, |
| "epoch": 2.0896, |
| "grad_norm": 0.18546347320079803, |
| "kl": 1.8335693359375, |
| "learning_rate": 7.629464263031132e-05, |
| "loss": 0.2215, |
| "reward": 0.91875, |
| "reward_std": 0.1766484022140503, |
| "rewards/accuracy_reward": 0.028125, |
| "rewards/format_reward": 0.890625, |
| "step": 1960 |
| }, |
| { |
| "completion_length": 56.5, |
| "epoch": 2.1002666666666667, |
| "grad_norm": 0.25101250410079956, |
| "kl": 1.4716064453125, |
| "learning_rate": 7.46774917007749e-05, |
| "loss": 0.1931, |
| "reward": 0.9125, |
| "reward_std": 0.14136751294136046, |
| "rewards/accuracy_reward": 0.021875, |
| "rewards/format_reward": 0.890625, |
| "step": 1970 |
| }, |
| { |
| "completion_length": 59.453125, |
| "epoch": 2.1109333333333336, |
| "grad_norm": 0.2480248659849167, |
| "kl": 1.67666015625, |
| "learning_rate": 7.30719638525462e-05, |
| "loss": 0.1832, |
| "reward": 0.903125, |
| "reward_std": 0.1623033404350281, |
| "rewards/accuracy_reward": 0.046875, |
| "rewards/format_reward": 0.85625, |
| "step": 1980 |
| }, |
| { |
| "completion_length": 58.46875, |
| "epoch": 2.1216, |
| "grad_norm": 0.19903503358364105, |
| "kl": 1.7174560546875, |
| "learning_rate": 7.147830683601388e-05, |
| "loss": 0.1958, |
| "reward": 0.921875, |
| "reward_std": 0.20825843811035155, |
| "rewards/accuracy_reward": 0.05, |
| "rewards/format_reward": 0.871875, |
| "step": 1990 |
| }, |
| { |
| "completion_length": 57.909375, |
| "epoch": 2.1322666666666668, |
| "grad_norm": 0.16613556444644928, |
| "kl": 2.1371826171875, |
| "learning_rate": 6.989676656976868e-05, |
| "loss": 0.3016, |
| "reward": 0.896875, |
| "reward_std": 0.20526910871267318, |
| "rewards/accuracy_reward": 0.034375, |
| "rewards/format_reward": 0.8625, |
| "step": 2000 |
| }, |
| { |
| "completion_length": 54.175, |
| "epoch": 2.142933333333333, |
| "grad_norm": 0.37461480498313904, |
| "kl": 1.8574951171875, |
| "learning_rate": 6.832758710265492e-05, |
| "loss": 0.2408, |
| "reward": 0.93125, |
| "reward_std": 0.17427992820739746, |
| "rewards/accuracy_reward": 0.040625, |
| "rewards/format_reward": 0.890625, |
| "step": 2010 |
| }, |
| { |
| "completion_length": 60.190625, |
| "epoch": 2.1536, |
| "grad_norm": 0.15081331133842468, |
| "kl": 2.0186279296875, |
| "learning_rate": 6.677101057611133e-05, |
| "loss": 0.229, |
| "reward": 0.8875, |
| "reward_std": 0.2087818503379822, |
| "rewards/accuracy_reward": 0.05625, |
| "rewards/format_reward": 0.83125, |
| "step": 2020 |
| }, |
| { |
| "completion_length": 56.49375, |
| "epoch": 2.164266666666667, |
| "grad_norm": 0.28891250491142273, |
| "kl": 1.998779296875, |
| "learning_rate": 6.522727718680623e-05, |
| "loss": 0.2448, |
| "reward": 0.875, |
| "reward_std": 0.18149680644273758, |
| "rewards/accuracy_reward": 0.025, |
| "rewards/format_reward": 0.85, |
| "step": 2030 |
| }, |
| { |
| "completion_length": 55.95, |
| "epoch": 2.174933333333333, |
| "grad_norm": 0.2829490900039673, |
| "kl": 2.0568115234375, |
| "learning_rate": 6.36966251495719e-05, |
| "loss": 0.2872, |
| "reward": 0.884375, |
| "reward_std": 0.1923343911767006, |
| "rewards/accuracy_reward": 0.015625, |
| "rewards/format_reward": 0.86875, |
| "step": 2040 |
| }, |
| { |
| "completion_length": 56.1, |
| "epoch": 2.1856, |
| "grad_norm": 0.46314913034439087, |
| "kl": 1.8526611328125, |
| "learning_rate": 6.217929066064636e-05, |
| "loss": 0.2034, |
| "reward": 0.890625, |
| "reward_std": 0.15861847400665283, |
| "rewards/accuracy_reward": 0.040625, |
| "rewards/format_reward": 0.85, |
| "step": 2050 |
| }, |
| { |
| "completion_length": 52.740625, |
| "epoch": 2.196266666666667, |
| "grad_norm": 0.07802554219961166, |
| "kl": 1.4162841796875, |
| "learning_rate": 6.067550786122496e-05, |
| "loss": 0.174, |
| "reward": 0.953125, |
| "reward_std": 0.16124776750802994, |
| "rewards/accuracy_reward": 0.059375, |
| "rewards/format_reward": 0.89375, |
| "step": 2060 |
| }, |
| { |
| "completion_length": 53.54375, |
| "epoch": 2.206933333333333, |
| "grad_norm": 0.4088180959224701, |
| "kl": 1.8029052734375, |
| "learning_rate": 5.918550880133017e-05, |
| "loss": 0.2458, |
| "reward": 0.9125, |
| "reward_std": 0.20420301407575608, |
| "rewards/accuracy_reward": 0.0375, |
| "rewards/format_reward": 0.875, |
| "step": 2070 |
| }, |
| { |
| "completion_length": 54.8, |
| "epoch": 2.2176, |
| "grad_norm": 0.3543941378593445, |
| "kl": 1.8623046875, |
| "learning_rate": 5.77095234040039e-05, |
| "loss": 0.2642, |
| "reward": 0.884375, |
| "reward_std": 0.18361847400665282, |
| "rewards/accuracy_reward": 0.03125, |
| "rewards/format_reward": 0.853125, |
| "step": 2080 |
| }, |
| { |
| "completion_length": 54.496875, |
| "epoch": 2.228266666666667, |
| "grad_norm": 0.20835542678833008, |
| "kl": 1.9629638671875, |
| "learning_rate": 5.624777942982734e-05, |
| "loss": 0.2696, |
| "reward": 0.8625, |
| "reward_std": 0.17620493620634078, |
| "rewards/accuracy_reward": 0.021875, |
| "rewards/format_reward": 0.840625, |
| "step": 2090 |
| }, |
| { |
| "completion_length": 58.75625, |
| "epoch": 2.238933333333333, |
| "grad_norm": 0.39805445075035095, |
| "kl": 3.082763671875, |
| "learning_rate": 5.480050244177572e-05, |
| "loss": 0.4061, |
| "reward": 0.8375, |
| "reward_std": 0.28123598694801333, |
| "rewards/accuracy_reward": 0.0375, |
| "rewards/format_reward": 0.8, |
| "step": 2100 |
| }, |
| { |
| "completion_length": 54.8125, |
| "epoch": 2.2496, |
| "grad_norm": 0.24060776829719543, |
| "kl": 2.3712890625, |
| "learning_rate": 5.33679157704109e-05, |
| "loss": 0.3109, |
| "reward": 0.89375, |
| "reward_std": 0.1928578034043312, |
| "rewards/accuracy_reward": 0.04375, |
| "rewards/format_reward": 0.85, |
| "step": 2110 |
| }, |
| { |
| "completion_length": 58.071875, |
| "epoch": 2.2602666666666664, |
| "grad_norm": 0.22757568955421448, |
| "kl": 1.7650634765625, |
| "learning_rate": 5.195024047941955e-05, |
| "loss": 0.2339, |
| "reward": 0.859375, |
| "reward_std": 0.19197435528039933, |
| "rewards/accuracy_reward": 0.0375, |
| "rewards/format_reward": 0.821875, |
| "step": 2120 |
| }, |
| { |
| "completion_length": 56.8125, |
| "epoch": 2.2709333333333332, |
| "grad_norm": 0.1922132819890976, |
| "kl": 2.193359375, |
| "learning_rate": 5.054769533149998e-05, |
| "loss": 0.282, |
| "reward": 0.86875, |
| "reward_std": 0.20499776750802995, |
| "rewards/accuracy_reward": 0.03125, |
| "rewards/format_reward": 0.8375, |
| "step": 2130 |
| }, |
| { |
| "completion_length": 53.096875, |
| "epoch": 2.2816, |
| "grad_norm": 0.1918802708387375, |
| "kl": 1.8790283203125, |
| "learning_rate": 4.916049675460543e-05, |
| "loss": 0.2532, |
| "reward": 0.9125, |
| "reward_std": 0.19136751294136048, |
| "rewards/accuracy_reward": 0.0375, |
| "rewards/format_reward": 0.875, |
| "step": 2140 |
| }, |
| { |
| "completion_length": 56.621875, |
| "epoch": 2.2922666666666665, |
| "grad_norm": 0.1691223531961441, |
| "kl": 2.2410888671875, |
| "learning_rate": 4.7788858808546275e-05, |
| "loss": 0.3084, |
| "reward": 0.871875, |
| "reward_std": 0.2093140110373497, |
| "rewards/accuracy_reward": 0.05, |
| "rewards/format_reward": 0.821875, |
| "step": 2150 |
| }, |
| { |
| "completion_length": 53.76875, |
| "epoch": 2.3029333333333333, |
| "grad_norm": 0.1904859095811844, |
| "kl": 2.6338623046875, |
| "learning_rate": 4.6432993151958546e-05, |
| "loss": 0.3819, |
| "reward": 0.884375, |
| "reward_std": 0.20359617173671724, |
| "rewards/accuracy_reward": 0.028125, |
| "rewards/format_reward": 0.85625, |
| "step": 2160 |
| }, |
| { |
| "completion_length": 52.234375, |
| "epoch": 2.3136, |
| "grad_norm": 0.20980538427829742, |
| "kl": 1.95830078125, |
| "learning_rate": 4.5093109009642855e-05, |
| "loss": 0.2513, |
| "reward": 0.915625, |
| "reward_std": 0.1976921945810318, |
| "rewards/accuracy_reward": 0.040625, |
| "rewards/format_reward": 0.875, |
| "step": 2170 |
| }, |
| { |
| "completion_length": 56.2375, |
| "epoch": 2.3242666666666665, |
| "grad_norm": 1.081332802772522, |
| "kl": 1.9156494140625, |
| "learning_rate": 4.376941314027819e-05, |
| "loss": 0.2811, |
| "reward": 0.86875, |
| "reward_std": 0.1936367705464363, |
| "rewards/accuracy_reward": 0.03125, |
| "rewards/format_reward": 0.8375, |
| "step": 2180 |
| }, |
| { |
| "completion_length": 50.08125, |
| "epoch": 2.3349333333333333, |
| "grad_norm": 0.263333261013031, |
| "kl": 2.2359375, |
| "learning_rate": 4.246210980451749e-05, |
| "loss": 0.3355, |
| "reward": 0.903125, |
| "reward_std": 0.20254059880971909, |
| "rewards/accuracy_reward": 0.01875, |
| "rewards/format_reward": 0.884375, |
| "step": 2190 |
| }, |
| { |
| "completion_length": 59.2125, |
| "epoch": 2.3456, |
| "grad_norm": 0.4202430248260498, |
| "kl": 3.7255859375, |
| "learning_rate": 4.117140073346749e-05, |
| "loss": 0.515, |
| "reward": 0.78125, |
| "reward_std": 0.2895918682217598, |
| "rewards/accuracy_reward": 0.009375, |
| "rewards/format_reward": 0.771875, |
| "step": 2200 |
| }, |
| { |
| "completion_length": 55.38125, |
| "epoch": 2.3562666666666665, |
| "grad_norm": 0.2832404673099518, |
| "kl": 2.9238525390625, |
| "learning_rate": 3.9897485097559684e-05, |
| "loss": 0.4053, |
| "reward": 0.875, |
| "reward_std": 0.2679324850440025, |
| "rewards/accuracy_reward": 0.0375, |
| "rewards/format_reward": 0.8375, |
| "step": 2210 |
| }, |
| { |
| "completion_length": 59.365625, |
| "epoch": 2.3669333333333333, |
| "grad_norm": 0.2592329978942871, |
| "kl": 2.598388671875, |
| "learning_rate": 3.864055947581605e-05, |
| "loss": 0.3084, |
| "reward": 0.815625, |
| "reward_std": 0.22225747704505922, |
| "rewards/accuracy_reward": 0.025, |
| "rewards/format_reward": 0.790625, |
| "step": 2220 |
| }, |
| { |
| "completion_length": 53.565625, |
| "epoch": 2.3776, |
| "grad_norm": 0.12750284373760223, |
| "kl": 2.2623291015625, |
| "learning_rate": 3.740081782551477e-05, |
| "loss": 0.3018, |
| "reward": 0.9125, |
| "reward_std": 0.19629059880971908, |
| "rewards/accuracy_reward": 0.028125, |
| "rewards/format_reward": 0.884375, |
| "step": 2230 |
| }, |
| { |
| "completion_length": 55.828125, |
| "epoch": 2.3882666666666665, |
| "grad_norm": 0.12283790111541748, |
| "kl": 1.74697265625, |
| "learning_rate": 3.617845145226067e-05, |
| "loss": 0.2485, |
| "reward": 0.921875, |
| "reward_std": 0.18282372057437896, |
| "rewards/accuracy_reward": 0.0375, |
| "rewards/format_reward": 0.884375, |
| "step": 2240 |
| }, |
| { |
| "completion_length": 50.853125, |
| "epoch": 2.3989333333333334, |
| "grad_norm": 0.1259896457195282, |
| "kl": 1.4160400390625, |
| "learning_rate": 3.4973648980464446e-05, |
| "loss": 0.1784, |
| "reward": 0.928125, |
| "reward_std": 0.13282372057437897, |
| "rewards/accuracy_reward": 0.028125, |
| "rewards/format_reward": 0.9, |
| "step": 2250 |
| }, |
| { |
| "completion_length": 51.35, |
| "epoch": 2.4096, |
| "grad_norm": 0.1746770590543747, |
| "kl": 1.1815673828125, |
| "learning_rate": 3.378659632423616e-05, |
| "loss": 0.1853, |
| "reward": 0.9625, |
| "reward_std": 0.14136751294136046, |
| "rewards/accuracy_reward": 0.046875, |
| "rewards/format_reward": 0.915625, |
| "step": 2260 |
| }, |
| { |
| "completion_length": 50.85625, |
| "epoch": 2.4202666666666666, |
| "grad_norm": 0.2644580006599426, |
| "kl": 1.69404296875, |
| "learning_rate": 3.261747665869621e-05, |
| "loss": 0.2072, |
| "reward": 0.965625, |
| "reward_std": 0.16124776750802994, |
| "rewards/accuracy_reward": 0.053125, |
| "rewards/format_reward": 0.9125, |
| "step": 2270 |
| }, |
| { |
| "completion_length": 56.90625, |
| "epoch": 2.4309333333333334, |
| "grad_norm": 0.1671292930841446, |
| "kl": 1.673974609375, |
| "learning_rate": 3.146647039171002e-05, |
| "loss": 0.1965, |
| "reward": 0.85, |
| "reward_std": 0.1655640110373497, |
| "rewards/accuracy_reward": 0.021875, |
| "rewards/format_reward": 0.828125, |
| "step": 2280 |
| }, |
| { |
| "completion_length": 57.678125, |
| "epoch": 2.4416, |
| "grad_norm": 0.25073471665382385, |
| "kl": 1.8162353515625, |
| "learning_rate": 3.0333755136048664e-05, |
| "loss": 0.2599, |
| "reward": 0.859375, |
| "reward_std": 0.20060684233903886, |
| "rewards/accuracy_reward": 0.05, |
| "rewards/format_reward": 0.809375, |
| "step": 2290 |
| }, |
| { |
| "completion_length": 54.340625, |
| "epoch": 2.4522666666666666, |
| "grad_norm": 0.2447560727596283, |
| "kl": 1.910791015625, |
| "learning_rate": 2.9219505681981524e-05, |
| "loss": 0.2393, |
| "reward": 0.91875, |
| "reward_std": 0.18528088927268982, |
| "rewards/accuracy_reward": 0.05, |
| "rewards/format_reward": 0.86875, |
| "step": 2300 |
| }, |
| { |
| "completion_length": 54.85625, |
| "epoch": 2.4629333333333334, |
| "grad_norm": 0.5278003215789795, |
| "kl": 1.694287109375, |
| "learning_rate": 2.812389397030415e-05, |
| "loss": 0.2223, |
| "reward": 0.88125, |
| "reward_std": 0.1822168782353401, |
| "rewards/accuracy_reward": 0.0375, |
| "rewards/format_reward": 0.84375, |
| "step": 2310 |
| }, |
| { |
| "completion_length": 57.6375, |
| "epoch": 2.4736000000000002, |
| "grad_norm": 0.1791255623102188, |
| "kl": 2.213330078125, |
| "learning_rate": 2.7047089065805977e-05, |
| "loss": 0.2678, |
| "reward": 0.84375, |
| "reward_std": 0.1837905988097191, |
| "rewards/accuracy_reward": 0.040625, |
| "rewards/format_reward": 0.803125, |
| "step": 2320 |
| }, |
| { |
| "completion_length": 54.99375, |
| "epoch": 2.4842666666666666, |
| "grad_norm": 0.06923685222864151, |
| "kl": 2.011474609375, |
| "learning_rate": 2.5989257131181874e-05, |
| "loss": 0.2702, |
| "reward": 0.890625, |
| "reward_std": 0.17790063470602036, |
| "rewards/accuracy_reward": 0.0375, |
| "rewards/format_reward": 0.853125, |
| "step": 2330 |
| }, |
| { |
| "completion_length": 55.078125, |
| "epoch": 2.4949333333333334, |
| "grad_norm": 0.2437313050031662, |
| "kl": 1.8513427734375, |
| "learning_rate": 2.4950561401391188e-05, |
| "loss": 0.2786, |
| "reward": 0.878125, |
| "reward_std": 0.1824636846780777, |
| "rewards/accuracy_reward": 0.03125, |
| "rewards/format_reward": 0.846875, |
| "step": 2340 |
| }, |
| { |
| "completion_length": 59.078125, |
| "epoch": 2.5056000000000003, |
| "grad_norm": 0.07926956564188004, |
| "kl": 1.988330078125, |
| "learning_rate": 2.3931162158469085e-05, |
| "loss": 0.2581, |
| "reward": 0.846875, |
| "reward_std": 0.16882468163967132, |
| "rewards/accuracy_reward": 0.0375, |
| "rewards/format_reward": 0.809375, |
| "step": 2350 |
| }, |
| { |
| "completion_length": 50.0875, |
| "epoch": 2.5162666666666667, |
| "grad_norm": 0.059707965701818466, |
| "kl": 1.9446533203125, |
| "learning_rate": 2.2931216706793136e-05, |
| "loss": 0.2305, |
| "reward": 0.940625, |
| "reward_std": 0.1668909251689911, |
| "rewards/accuracy_reward": 0.05625, |
| "rewards/format_reward": 0.884375, |
| "step": 2360 |
| }, |
| { |
| "completion_length": 50.753125, |
| "epoch": 2.5269333333333335, |
| "grad_norm": 0.12831467390060425, |
| "kl": 1.78369140625, |
| "learning_rate": 2.1950879348809546e-05, |
| "loss": 0.2416, |
| "reward": 0.94375, |
| "reward_std": 0.2137136846780777, |
| "rewards/accuracy_reward": 0.059375, |
| "rewards/format_reward": 0.884375, |
| "step": 2370 |
| }, |
| { |
| "completion_length": 60.515625, |
| "epoch": 2.5376, |
| "grad_norm": 0.11030708253383636, |
| "kl": 1.88526611328125, |
| "learning_rate": 2.099030136122276e-05, |
| "loss": 0.2618, |
| "reward": 0.8375, |
| "reward_std": 0.17886751294136047, |
| "rewards/accuracy_reward": 0.034375, |
| "rewards/format_reward": 0.803125, |
| "step": 2380 |
| }, |
| { |
| "completion_length": 51.6125, |
| "epoch": 2.5482666666666667, |
| "grad_norm": 1.4983174800872803, |
| "kl": 1.6604248046875, |
| "learning_rate": 2.0049630971651576e-05, |
| "loss": 0.2423, |
| "reward": 0.96875, |
| "reward_std": 0.1933012694120407, |
| "rewards/accuracy_reward": 0.059375, |
| "rewards/format_reward": 0.909375, |
| "step": 2390 |
| }, |
| { |
| "completion_length": 54.178125, |
| "epoch": 2.558933333333333, |
| "grad_norm": 0.2519737482070923, |
| "kl": 2.282080078125, |
| "learning_rate": 1.9129013335756316e-05, |
| "loss": 0.3116, |
| "reward": 0.890625, |
| "reward_std": 0.23096464574337006, |
| "rewards/accuracy_reward": 0.034375, |
| "rewards/format_reward": 0.85625, |
| "step": 2400 |
| }, |
| { |
| "completion_length": 56.959375, |
| "epoch": 2.5696, |
| "grad_norm": 0.16114293038845062, |
| "kl": 1.4489013671875, |
| "learning_rate": 1.822859051483932e-05, |
| "loss": 0.1778, |
| "reward": 0.925, |
| "reward_std": 0.1691707044839859, |
| "rewards/accuracy_reward": 0.05625, |
| "rewards/format_reward": 0.86875, |
| "step": 2410 |
| }, |
| { |
| "completion_length": 53.259375, |
| "epoch": 2.5802666666666667, |
| "grad_norm": 0.11866087466478348, |
| "kl": 1.4989501953125, |
| "learning_rate": 1.7348501453923826e-05, |
| "loss": 0.1715, |
| "reward": 0.90625, |
| "reward_std": 0.1558012694120407, |
| "rewards/accuracy_reward": 0.03125, |
| "rewards/format_reward": 0.875, |
| "step": 2420 |
| }, |
| { |
| "completion_length": 54.046875, |
| "epoch": 2.590933333333333, |
| "grad_norm": 0.24277645349502563, |
| "kl": 1.7145751953125, |
| "learning_rate": 1.64888819603129e-05, |
| "loss": 0.2216, |
| "reward": 0.925, |
| "reward_std": 0.16935684233903886, |
| "rewards/accuracy_reward": 0.034375, |
| "rewards/format_reward": 0.890625, |
| "step": 2430 |
| }, |
| { |
| "completion_length": 57.071875, |
| "epoch": 2.6016, |
| "grad_norm": 0.04312370344996452, |
| "kl": 2.06767578125, |
| "learning_rate": 1.564986468263298e-05, |
| "loss": 0.2706, |
| "reward": 0.8875, |
| "reward_std": 0.19136751294136048, |
| "rewards/accuracy_reward": 0.03125, |
| "rewards/format_reward": 0.85625, |
| "step": 2440 |
| }, |
| { |
| "completion_length": 54.8625, |
| "epoch": 2.6122666666666667, |
| "grad_norm": 0.2207166850566864, |
| "kl": 2.34619140625, |
| "learning_rate": 1.4831579090364832e-05, |
| "loss": 0.2831, |
| "reward": 0.871875, |
| "reward_std": 0.20552992820739746, |
| "rewards/accuracy_reward": 0.025, |
| "rewards/format_reward": 0.846875, |
| "step": 2450 |
| }, |
| { |
| "completion_length": 51.11875, |
| "epoch": 2.622933333333333, |
| "grad_norm": 0.2076699137687683, |
| "kl": 1.8322509765625, |
| "learning_rate": 1.4034151453864845e-05, |
| "loss": 0.2557, |
| "reward": 0.940625, |
| "reward_std": 0.20333535224199295, |
| "rewards/accuracy_reward": 0.05625, |
| "rewards/format_reward": 0.884375, |
| "step": 2460 |
| }, |
| { |
| "completion_length": 55.934375, |
| "epoch": 2.6336, |
| "grad_norm": 0.27051660418510437, |
| "kl": 2.1045166015625, |
| "learning_rate": 1.3257704824880377e-05, |
| "loss": 0.2683, |
| "reward": 0.915625, |
| "reward_std": 0.23369315564632415, |
| "rewards/accuracy_reward": 0.0625, |
| "rewards/format_reward": 0.853125, |
| "step": 2470 |
| }, |
| { |
| "completion_length": 54.821875, |
| "epoch": 2.6442666666666668, |
| "grad_norm": 0.32035326957702637, |
| "kl": 2.1305419921875, |
| "learning_rate": 1.2502359017561297e-05, |
| "loss": 0.2833, |
| "reward": 0.928125, |
| "reward_std": 0.22261751294136048, |
| "rewards/accuracy_reward": 0.05625, |
| "rewards/format_reward": 0.871875, |
| "step": 2480 |
| }, |
| { |
| "completion_length": 52.3625, |
| "epoch": 2.654933333333333, |
| "grad_norm": 0.0916447713971138, |
| "kl": 1.389013671875, |
| "learning_rate": 1.1768230589971455e-05, |
| "loss": 0.1857, |
| "reward": 0.95625, |
| "reward_std": 0.17930223047733307, |
| "rewards/accuracy_reward": 0.06875, |
| "rewards/format_reward": 0.8875, |
| "step": 2490 |
| }, |
| { |
| "completion_length": 55.6625, |
| "epoch": 2.6656, |
| "grad_norm": 0.23957239091396332, |
| "kl": 1.5188232421875, |
| "learning_rate": 1.1055432826102589e-05, |
| "loss": 0.223, |
| "reward": 0.946875, |
| "reward_std": 0.1785961717367172, |
| "rewards/accuracy_reward": 0.075, |
| "rewards/format_reward": 0.871875, |
| "step": 2500 |
| }, |
| { |
| "completion_length": 52.58125, |
| "epoch": 2.676266666666667, |
| "grad_norm": 0.14451664686203003, |
| "kl": 1.84560546875, |
| "learning_rate": 1.0364075718393172e-05, |
| "loss": 0.224, |
| "reward": 0.9625, |
| "reward_std": 0.22999776750802994, |
| "rewards/accuracy_reward": 0.084375, |
| "rewards/format_reward": 0.878125, |
| "step": 2510 |
| }, |
| { |
| "completion_length": 58.35625, |
| "epoch": 2.686933333333333, |
| "grad_norm": 0.18883152306079865, |
| "kl": 2.30576171875, |
| "learning_rate": 9.69426595075566e-06, |
| "loss": 0.3123, |
| "reward": 0.865625, |
| "reward_std": 0.20895397514104844, |
| "rewards/accuracy_reward": 0.0375, |
| "rewards/format_reward": 0.828125, |
| "step": 2520 |
| }, |
| { |
| "completion_length": 58.309375, |
| "epoch": 2.6976, |
| "grad_norm": 0.3094783127307892, |
| "kl": 1.9437255859375, |
| "learning_rate": 9.046106882113751e-06, |
| "loss": 0.2174, |
| "reward": 0.875, |
| "reward_std": 0.15157372057437896, |
| "rewards/accuracy_reward": 0.0375, |
| "rewards/format_reward": 0.8375, |
| "step": 2530 |
| }, |
| { |
| "completion_length": 51.671875, |
| "epoch": 2.708266666666667, |
| "grad_norm": 0.3006289601325989, |
| "kl": 2.19658203125, |
| "learning_rate": 8.419698530453095e-06, |
| "loss": 0.3126, |
| "reward": 0.91875, |
| "reward_std": 0.1887136846780777, |
| "rewards/accuracy_reward": 0.021875, |
| "rewards/format_reward": 0.896875, |
| "step": 2540 |
| }, |
| { |
| "completion_length": 51.45, |
| "epoch": 2.718933333333333, |
| "grad_norm": 0.3948608338832855, |
| "kl": 2.006982421875, |
| "learning_rate": 7.815137557387419e-06, |
| "loss": 0.2757, |
| "reward": 0.95, |
| "reward_std": 0.2034987285733223, |
| "rewards/accuracy_reward": 0.040625, |
| "rewards/format_reward": 0.909375, |
| "step": 2550 |
| }, |
| { |
| "completion_length": 51.890625, |
| "epoch": 2.7296, |
| "grad_norm": 0.13767310976982117, |
| "kl": 1.519189453125, |
| "learning_rate": 7.232517253242636e-06, |
| "loss": 0.1851, |
| "reward": 0.925, |
| "reward_std": 0.16177992820739745, |
| "rewards/accuracy_reward": 0.03125, |
| "rewards/format_reward": 0.89375, |
| "step": 2560 |
| }, |
| { |
| "completion_length": 53.684375, |
| "epoch": 2.740266666666667, |
| "grad_norm": 0.15098528563976288, |
| "kl": 2.033837890625, |
| "learning_rate": 6.671927522661047e-06, |
| "loss": 0.2932, |
| "reward": 0.975, |
| "reward_std": 0.2466260999441147, |
| "rewards/accuracy_reward": 0.0875, |
| "rewards/format_reward": 0.8875, |
| "step": 2570 |
| }, |
| { |
| "completion_length": 53.775, |
| "epoch": 2.7509333333333332, |
| "grad_norm": 0.1634114533662796, |
| "kl": 1.951416015625, |
| "learning_rate": 6.133454870728111e-06, |
| "loss": 0.2477, |
| "reward": 0.91875, |
| "reward_std": 0.198650324344635, |
| "rewards/accuracy_reward": 0.04375, |
| "rewards/format_reward": 0.875, |
| "step": 2580 |
| }, |
| { |
| "completion_length": 56.625, |
| "epoch": 2.7616, |
| "grad_norm": 0.1946033537387848, |
| "kl": 1.8022705078125, |
| "learning_rate": 5.617182389623814e-06, |
| "loss": 0.2218, |
| "reward": 0.93125, |
| "reward_std": 0.20193375647068024, |
| "rewards/accuracy_reward": 0.06875, |
| "rewards/format_reward": 0.8625, |
| "step": 2590 |
| }, |
| { |
| "completion_length": 55.571875, |
| "epoch": 2.772266666666667, |
| "grad_norm": 0.0797191932797432, |
| "kl": 2.0580810546875, |
| "learning_rate": 5.1231897458005765e-06, |
| "loss": 0.223, |
| "reward": 0.88125, |
| "reward_std": 0.2035074770450592, |
| "rewards/accuracy_reward": 0.03125, |
| "rewards/format_reward": 0.85, |
| "step": 2600 |
| }, |
| { |
| "completion_length": 58.39375, |
| "epoch": 2.7829333333333333, |
| "grad_norm": 0.2864690124988556, |
| "kl": 2.5972412109375, |
| "learning_rate": 4.651553167689931e-06, |
| "loss": 0.3517, |
| "reward": 0.88125, |
| "reward_std": 0.24293248504400253, |
| "rewards/accuracy_reward": 0.05625, |
| "rewards/format_reward": 0.825, |
| "step": 2610 |
| }, |
| { |
| "completion_length": 54.334375, |
| "epoch": 2.7936, |
| "grad_norm": 0.21106931567192078, |
| "kl": 1.7190673828125, |
| "learning_rate": 4.202345433939552e-06, |
| "loss": 0.2163, |
| "reward": 0.878125, |
| "reward_std": 0.19532372057437897, |
| "rewards/accuracy_reward": 0.040625, |
| "rewards/format_reward": 0.8375, |
| "step": 2620 |
| }, |
| { |
| "completion_length": 53.89375, |
| "epoch": 2.804266666666667, |
| "grad_norm": 0.21579229831695557, |
| "kl": 1.726171875, |
| "learning_rate": 3.7756358621827907e-06, |
| "loss": 0.2158, |
| "reward": 0.90625, |
| "reward_std": 0.1833471328020096, |
| "rewards/accuracy_reward": 0.034375, |
| "rewards/format_reward": 0.871875, |
| "step": 2630 |
| }, |
| { |
| "completion_length": 54.48125, |
| "epoch": 2.8149333333333333, |
| "grad_norm": 0.11399048566818237, |
| "kl": 1.6014404296875, |
| "learning_rate": 3.371490298342194e-06, |
| "loss": 0.1927, |
| "reward": 0.93125, |
| "reward_std": 0.16407372057437897, |
| "rewards/accuracy_reward": 0.05, |
| "rewards/format_reward": 0.88125, |
| "step": 2640 |
| }, |
| { |
| "completion_length": 55.609375, |
| "epoch": 2.8256, |
| "grad_norm": 0.31620094180107117, |
| "kl": 2.4648193359375, |
| "learning_rate": 2.989971106468736e-06, |
| "loss": 0.3298, |
| "reward": 0.909375, |
| "reward_std": 0.24601925760507584, |
| "rewards/accuracy_reward": 0.0625, |
| "rewards/format_reward": 0.846875, |
| "step": 2650 |
| }, |
| { |
| "completion_length": 54.99375, |
| "epoch": 2.836266666666667, |
| "grad_norm": 0.3356408178806305, |
| "kl": 2.2318359375, |
| "learning_rate": 2.631137159118446e-06, |
| "loss": 0.2667, |
| "reward": 0.9, |
| "reward_std": 0.2262136846780777, |
| "rewards/accuracy_reward": 0.040625, |
| "rewards/format_reward": 0.859375, |
| "step": 2660 |
| }, |
| { |
| "completion_length": 55.634375, |
| "epoch": 2.8469333333333333, |
| "grad_norm": 0.6469197273254395, |
| "kl": 2.4311767578125, |
| "learning_rate": 2.2950438282676453e-06, |
| "loss": 0.3333, |
| "reward": 0.90625, |
| "reward_std": 0.25358215868473055, |
| "rewards/accuracy_reward": 0.053125, |
| "rewards/format_reward": 0.853125, |
| "step": 2670 |
| }, |
| { |
| "completion_length": 59.8375, |
| "epoch": 2.8576, |
| "grad_norm": 0.17483676970005035, |
| "kl": 2.5937255859375, |
| "learning_rate": 1.9817429767684466e-06, |
| "loss": 0.3489, |
| "reward": 0.8625, |
| "reward_std": 0.2525265857577324, |
| "rewards/accuracy_reward": 0.05, |
| "rewards/format_reward": 0.8125, |
| "step": 2680 |
| }, |
| { |
| "completion_length": 55.5, |
| "epoch": 2.8682666666666665, |
| "grad_norm": 0.2512027621269226, |
| "kl": 2.55244140625, |
| "learning_rate": 1.6912829503458569e-06, |
| "loss": 0.3347, |
| "reward": 0.896875, |
| "reward_std": 0.24197435528039932, |
| "rewards/accuracy_reward": 0.034375, |
| "rewards/format_reward": 0.8625, |
| "step": 2690 |
| }, |
| { |
| "completion_length": 55.325, |
| "epoch": 2.8789333333333333, |
| "grad_norm": 0.16474415361881256, |
| "kl": 1.9730712890625, |
| "learning_rate": 1.4237085701374107e-06, |
| "loss": 0.2434, |
| "reward": 0.8875, |
| "reward_std": 0.22551814764738082, |
| "rewards/accuracy_reward": 0.0375, |
| "rewards/format_reward": 0.85, |
| "step": 2700 |
| }, |
| { |
| "completion_length": 55.084375, |
| "epoch": 2.8895999999999997, |
| "grad_norm": 0.06464667618274689, |
| "kl": 2.05751953125, |
| "learning_rate": 1.1790611257767868e-06, |
| "loss": 0.2714, |
| "reward": 0.925, |
| "reward_std": 0.2470608174800873, |
| "rewards/accuracy_reward": 0.0625, |
| "rewards/format_reward": 0.8625, |
| "step": 2710 |
| }, |
| { |
| "completion_length": 57.228125, |
| "epoch": 2.9002666666666665, |
| "grad_norm": 0.20830662548542023, |
| "kl": 2.081640625, |
| "learning_rate": 9.573783690224213e-07, |
| "loss": 0.2514, |
| "reward": 0.903125, |
| "reward_std": 0.22295301407575607, |
| "rewards/accuracy_reward": 0.059375, |
| "rewards/format_reward": 0.84375, |
| "step": 2720 |
| }, |
| { |
| "completion_length": 56.6375, |
| "epoch": 2.9109333333333334, |
| "grad_norm": 0.20990578830242157, |
| "kl": 1.5892578125, |
| "learning_rate": 7.586945079319673e-07, |
| "loss": 0.2132, |
| "reward": 0.9, |
| "reward_std": 0.1721687823534012, |
| "rewards/accuracy_reward": 0.05, |
| "rewards/format_reward": 0.85, |
| "step": 2730 |
| }, |
| { |
| "completion_length": 58.346875, |
| "epoch": 2.9215999999999998, |
| "grad_norm": 0.010507356375455856, |
| "kl": 1.9132568359375, |
| "learning_rate": 5.830402015836178e-07, |
| "loss": 0.2316, |
| "reward": 0.853125, |
| "reward_std": 0.18854155987501145, |
| "rewards/accuracy_reward": 0.03125, |
| "rewards/format_reward": 0.821875, |
| "step": 2740 |
| }, |
| { |
| "completion_length": 57.3125, |
| "epoch": 2.9322666666666666, |
| "grad_norm": 0.4105237126350403, |
| "kl": 2.6902587890625, |
| "learning_rate": 4.304425553450896e-07, |
| "loss": 0.3641, |
| "reward": 0.865625, |
| "reward_std": 0.2506815239787102, |
| "rewards/accuracy_reward": 0.03125, |
| "rewards/format_reward": 0.834375, |
| "step": 2750 |
| }, |
| { |
| "completion_length": 53.96875, |
| "epoch": 2.9429333333333334, |
| "grad_norm": 0.11192090809345245, |
| "kl": 2.0924072265625, |
| "learning_rate": 3.009251166909699e-07, |
| "loss": 0.2969, |
| "reward": 0.928125, |
| "reward_std": 0.2324636846780777, |
| "rewards/accuracy_reward": 0.046875, |
| "rewards/format_reward": 0.88125, |
| "step": 2760 |
| }, |
| { |
| "completion_length": 57.19375, |
| "epoch": 2.9536, |
| "grad_norm": 0.16923871636390686, |
| "kl": 1.6264404296875, |
| "learning_rate": 1.9450787156907177e-07, |
| "loss": 0.209, |
| "reward": 0.865625, |
| "reward_std": 0.1672264263033867, |
| "rewards/accuracy_reward": 0.03125, |
| "rewards/format_reward": 0.834375, |
| "step": 2770 |
| }, |
| { |
| "completion_length": 55.79375, |
| "epoch": 2.9642666666666666, |
| "grad_norm": 0.19571039080619812, |
| "kl": 1.5086181640625, |
| "learning_rate": 1.1120724131638558e-07, |
| "loss": 0.186, |
| "reward": 0.90625, |
| "reward_std": 0.16512929350137712, |
| "rewards/accuracy_reward": 0.0375, |
| "rewards/format_reward": 0.86875, |
| "step": 2780 |
| }, |
| { |
| "completion_length": 54.446875, |
| "epoch": 2.9749333333333334, |
| "grad_norm": 0.2531106173992157, |
| "kl": 2.208544921875, |
| "learning_rate": 5.103608012512195e-08, |
| "loss": 0.2793, |
| "reward": 0.95625, |
| "reward_std": 0.2638788342475891, |
| "rewards/accuracy_reward": 0.0875, |
| "rewards/format_reward": 0.86875, |
| "step": 2790 |
| }, |
| { |
| "completion_length": 54.8875, |
| "epoch": 2.9856, |
| "grad_norm": 0.5443268418312073, |
| "kl": 2.21259765625, |
| "learning_rate": 1.400367305913197e-08, |
| "loss": 0.2949, |
| "reward": 0.89375, |
| "reward_std": 0.20729155987501144, |
| "rewards/accuracy_reward": 0.034375, |
| "rewards/format_reward": 0.859375, |
| "step": 2800 |
| }, |
| { |
| "completion_length": 56.165625, |
| "epoch": 2.9962666666666666, |
| "grad_norm": 0.1890791356563568, |
| "kl": 1.9867919921875, |
| "learning_rate": 1.1573462112002274e-10, |
| "loss": 0.2607, |
| "reward": 0.85625, |
| "reward_std": 0.19206304997205734, |
| "rewards/accuracy_reward": 0.034375, |
| "rewards/format_reward": 0.821875, |
| "step": 2810 |
| }, |
| { |
| "completion_length": 56.4375, |
| "epoch": 2.997333333333333, |
| "kl": 4.046875, |
| "reward": 0.875, |
| "reward_std": 0.3221687823534012, |
| "rewards/accuracy_reward": 0.03125, |
| "rewards/format_reward": 0.84375, |
| "step": 2811, |
| "total_flos": 0.0, |
| "train_loss": 0.21294908598650353, |
| "train_runtime": 95825.8938, |
| "train_samples_per_second": 0.235, |
| "train_steps_per_second": 0.029 |
| } |
| ], |
| "logging_steps": 10, |
| "max_steps": 2811, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 3, |
| "save_steps": 500, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 0.0, |
| "train_batch_size": 8, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|