diff --git "a/checkpoint-600/trainer_state.json" "b/checkpoint-600/trainer_state.json" new file mode 100644--- /dev/null +++ "b/checkpoint-600/trainer_state.json" @@ -0,0 +1,7534 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.8391608391608392, + "eval_steps": 500, + "global_step": 600, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "clip_ratio": 0.0, + "completion_length": 94.14286041259766, + "epoch": 0.0013986013986013986, + "grad_norm": 0.09098726750251449, + "kl": 0.0003147125244140625, + "learning_rate": 3.496503496503497e-07, + "loss": -0.0001, + "num_tokens": 19338.0, + "reward": 1.3853150606155396, + "reward_std": 0.4827312231063843, + "rewards/check_gptzero_func": 0.3214285671710968, + "rewards/check_perplexity_diff_func": 0.6428571343421936, + "rewards/check_winston_local_func": 0.42102929949760437, + "step": 1 + }, + { + "clip_ratio": 0.0, + "epoch": 0.002797202797202797, + "grad_norm": 0.09100260462019598, + "kl": 0.0003147125244140625, + "learning_rate": 6.993006993006994e-07, + "loss": -0.0001, + "step": 2 + }, + { + "clip_ratio": 0.0031990089919418097, + "completion_length": 97.73214721679688, + "epoch": 0.004195804195804196, + "grad_norm": 0.16978190242842756, + "kl": 0.0003662109375, + "learning_rate": 1.0489510489510491e-06, + "loss": -0.0017, + "num_tokens": 39487.0, + "reward": 0.9710169434547424, + "reward_std": 0.6376833319664001, + "rewards/check_gptzero_func": 0.125, + "rewards/check_perplexity_diff_func": 0.5, + "rewards/check_winston_local_func": 0.34601688385009766, + "step": 3 + }, + { + "clip_ratio": 0.0018373647471889853, + "epoch": 0.005594405594405594, + "grad_norm": 0.17663660241053314, + "kl": 0.000423431396484375, + "learning_rate": 1.3986013986013987e-06, + "loss": -0.0013, + "step": 4 + }, + { + "clip_ratio": 0.0016871786210685968, + "completion_length": 72.83928680419922, + "epoch": 0.006993006993006993, + "grad_norm": 0.09466520543424979, + "kl": 0.00034332275390625, + "learning_rate": 1.7482517482517485e-06, + "loss": 0.0008, + "num_tokens": 56082.0, + "reward": 1.0736479759216309, + "reward_std": 0.6862176656723022, + "rewards/check_gptzero_func": 0.2321428507566452, + "rewards/check_perplexity_diff_func": 0.3928571343421936, + "rewards/check_winston_local_func": 0.4486479163169861, + "step": 5 + }, + { + "clip_ratio": 0.0010098000057041645, + "epoch": 0.008391608391608392, + "grad_norm": 0.09539550614656178, + "kl": 0.0003376007080078125, + "learning_rate": 2.0979020979020983e-06, + "loss": 0.001, + "step": 6 + }, + { + "clip_ratio": 0.00184189947322011, + "completion_length": 95.92857360839844, + "epoch": 0.009790209790209791, + "grad_norm": 0.10573631695411624, + "kl": 0.0003757476806640625, + "learning_rate": 2.4475524475524477e-06, + "loss": 0.001, + "num_tokens": 75906.0, + "reward": 1.135968804359436, + "reward_std": 0.48974505066871643, + "rewards/check_gptzero_func": 0.2142857164144516, + "rewards/check_perplexity_diff_func": 0.5, + "rewards/check_winston_local_func": 0.42168304324150085, + "step": 7 + }, + { + "clip_ratio": 0.0018530809320509434, + "epoch": 0.011188811188811189, + "grad_norm": 0.10614532740192534, + "kl": 0.0004119873046875, + "learning_rate": 2.7972027972027974e-06, + "loss": 0.0011, + "step": 8 + }, + { + "clip_ratio": 0.0017129909247159958, + "completion_length": 106.76786041259766, + "epoch": 0.012587412587412588, + "grad_norm": 0.07916450383171113, + "kl": 0.000377655029296875, + "learning_rate": 3.1468531468531472e-06, + "loss": -0.0012, + "num_tokens": 96989.0, + "reward": 0.9590713977813721, + "reward_std": 0.5206417441368103, + "rewards/check_gptzero_func": 0.2321428507566452, + "rewards/check_perplexity_diff_func": 0.3928571343421936, + "rewards/check_winston_local_func": 0.3340713679790497, + "step": 9 + }, + { + "clip_ratio": 0.0034840807784348726, + "epoch": 0.013986013986013986, + "grad_norm": 0.08061029037170311, + "kl": 0.00037384033203125, + "learning_rate": 3.496503496503497e-06, + "loss": -0.0008, + "step": 10 + }, + { + "clip_ratio": 0.0017933619674295187, + "completion_length": 102.85714721679688, + "epoch": 0.015384615384615385, + "grad_norm": 0.07930375667041344, + "kl": 0.00030517578125, + "learning_rate": 3.846153846153847e-06, + "loss": -0.0038, + "num_tokens": 117271.0, + "reward": 1.1903148889541626, + "reward_std": 0.4722224473953247, + "rewards/check_gptzero_func": 0.2857142984867096, + "rewards/check_perplexity_diff_func": 0.5357142686843872, + "rewards/check_winston_local_func": 0.368886262178421, + "step": 11 + }, + { + "clip_ratio": 0.0015239757485687733, + "epoch": 0.016783216783216783, + "grad_norm": 0.07898151410343454, + "kl": 0.0003871917724609375, + "learning_rate": 4.195804195804197e-06, + "loss": -0.0041, + "step": 12 + }, + { + "clip_ratio": 0.0029670600779354572, + "completion_length": 104.37500762939453, + "epoch": 0.01818181818181818, + "grad_norm": 0.09247854491588844, + "kl": 0.0003986358642578125, + "learning_rate": 4.5454545454545455e-06, + "loss": -0.0049, + "num_tokens": 138174.0, + "reward": 0.9787324070930481, + "reward_std": 0.63657546043396, + "rewards/check_gptzero_func": 0.1964285671710968, + "rewards/check_perplexity_diff_func": 0.5714285969734192, + "rewards/check_winston_local_func": 0.21087531745433807, + "step": 13 + }, + { + "clip_ratio": 0.003217793768271804, + "epoch": 0.019580419580419582, + "grad_norm": 0.09293749933034297, + "kl": 0.0003204345703125, + "learning_rate": 4.895104895104895e-06, + "loss": -0.0045, + "step": 14 + }, + { + "clip_ratio": 0.002677519340068102, + "completion_length": 67.41072082519531, + "epoch": 0.02097902097902098, + "grad_norm": 0.14020673218962584, + "kl": 0.000499725341796875, + "learning_rate": 5.244755244755245e-06, + "loss": -0.008, + "num_tokens": 154385.0, + "reward": 1.0573337078094482, + "reward_std": 0.5388314127922058, + "rewards/check_gptzero_func": 0.2857142984867096, + "rewards/check_perplexity_diff_func": 0.3571428656578064, + "rewards/check_winston_local_func": 0.41447654366493225, + "step": 15 + }, + { + "clip_ratio": 0.0021408216562122107, + "epoch": 0.022377622377622378, + "grad_norm": 0.16110555979669544, + "kl": 0.000530242919921875, + "learning_rate": 5.594405594405595e-06, + "loss": -0.0081, + "step": 16 + }, + { + "clip_ratio": 0.002852825215086341, + "completion_length": 89.83928680419922, + "epoch": 0.023776223776223775, + "grad_norm": 0.1066932843002634, + "kl": 0.000415802001953125, + "learning_rate": 5.944055944055945e-06, + "loss": 0.0166, + "num_tokens": 173158.0, + "reward": 1.2226747274398804, + "reward_std": 0.45578521490097046, + "rewards/check_gptzero_func": 0.3392857015132904, + "rewards/check_perplexity_diff_func": 0.5, + "rewards/check_winston_local_func": 0.3833889663219452, + "step": 17 + }, + { + "clip_ratio": 0.002018069615587592, + "epoch": 0.025174825174825177, + "grad_norm": 0.10578254537451186, + "kl": 0.0003833770751953125, + "learning_rate": 6.2937062937062944e-06, + "loss": 0.0164, + "step": 18 + }, + { + "clip_ratio": 0.0032663617748767138, + "completion_length": 105.83928680419922, + "epoch": 0.026573426573426574, + "grad_norm": 0.11554707163597037, + "kl": 0.0003719329833984375, + "learning_rate": 6.643356643356643e-06, + "loss": -0.0005, + "num_tokens": 194649.0, + "reward": 0.8957912921905518, + "reward_std": 0.47702187299728394, + "rewards/check_gptzero_func": 0.1785714328289032, + "rewards/check_perplexity_diff_func": 0.3571428656578064, + "rewards/check_winston_local_func": 0.3600769340991974, + "step": 19 + }, + { + "clip_ratio": 0.0024422537535429, + "epoch": 0.027972027972027972, + "grad_norm": 0.11417666172406592, + "kl": 0.0003814697265625, + "learning_rate": 6.993006993006994e-06, + "loss": -0.0008, + "step": 20 + }, + { + "clip_ratio": 0.0013639701064676046, + "completion_length": 107.39286041259766, + "epoch": 0.02937062937062937, + "grad_norm": 0.09634196153318929, + "kl": 0.000396728515625, + "learning_rate": 7.342657342657343e-06, + "loss": -0.0011, + "num_tokens": 215987.0, + "reward": 0.9197577238082886, + "reward_std": 0.39100831747055054, + "rewards/check_gptzero_func": 0.2321428507566452, + "rewards/check_perplexity_diff_func": 0.3571428656578064, + "rewards/check_winston_local_func": 0.33047202229499817, + "step": 21 + }, + { + "clip_ratio": 0.003595889313146472, + "epoch": 0.03076923076923077, + "grad_norm": 0.09779857975187953, + "kl": 0.00052642822265625, + "learning_rate": 7.692307692307694e-06, + "loss": -0.0013, + "step": 22 + }, + { + "clip_ratio": 0.0024316231720149517, + "completion_length": 89.08928680419922, + "epoch": 0.032167832167832165, + "grad_norm": 0.13189477135999747, + "kl": 0.0003414154052734375, + "learning_rate": 8.041958041958042e-06, + "loss": -0.0009, + "num_tokens": 234678.0, + "reward": 0.9593304395675659, + "reward_std": 0.5972030758857727, + "rewards/check_gptzero_func": 0.1964285671710968, + "rewards/check_perplexity_diff_func": 0.4285714328289032, + "rewards/check_winston_local_func": 0.33433040976524353, + "step": 23 + }, + { + "clip_ratio": 0.000728283659555018, + "epoch": 0.033566433566433566, + "grad_norm": 0.12953291972672934, + "kl": 0.000335693359375, + "learning_rate": 8.391608391608393e-06, + "loss": -0.0011, + "step": 24 + }, + { + "clip_ratio": 0.0021546650677919388, + "completion_length": 90.98214721679688, + "epoch": 0.03496503496503497, + "grad_norm": 0.07682474826469807, + "kl": 0.0004558563232421875, + "learning_rate": 8.741258741258741e-06, + "loss": 0.0077, + "num_tokens": 253833.0, + "reward": 0.8255766034126282, + "reward_std": 0.3988816440105438, + "rewards/check_gptzero_func": 0.2678571343421936, + "rewards/check_perplexity_diff_func": 0.2857142984867096, + "rewards/check_winston_local_func": 0.272005170583725, + "step": 25 + }, + { + "clip_ratio": 0.0019250252516940236, + "epoch": 0.03636363636363636, + "grad_norm": 0.07873511079065725, + "kl": 0.0003490447998046875, + "learning_rate": 9.090909090909091e-06, + "loss": 0.0075, + "step": 26 + }, + { + "clip_ratio": 0.001886485842987895, + "completion_length": 74.4464340209961, + "epoch": 0.03776223776223776, + "grad_norm": 0.06765277577538054, + "kl": 0.000354766845703125, + "learning_rate": 9.44055944055944e-06, + "loss": -0.0003, + "num_tokens": 270908.0, + "reward": 1.3086735010147095, + "reward_std": 0.3881511986255646, + "rewards/check_gptzero_func": 0.4107142984867096, + "rewards/check_perplexity_diff_func": 0.4642857015132904, + "rewards/check_winston_local_func": 0.4336733818054199, + "step": 27 + }, + { + "clip_ratio": 0.0020898371003568172, + "epoch": 0.039160839160839164, + "grad_norm": 0.06722581226675682, + "kl": 0.000453948974609375, + "learning_rate": 9.79020979020979e-06, + "loss": -0.0006, + "step": 28 + }, + { + "clip_ratio": 0.002855924190953374, + "completion_length": 103.64286041259766, + "epoch": 0.04055944055944056, + "grad_norm": 0.16301390061045865, + "kl": 0.0003948211669921875, + "learning_rate": 1.013986013986014e-05, + "loss": 0.0063, + "num_tokens": 292060.0, + "reward": 1.22488272190094, + "reward_std": 0.8093323111534119, + "rewards/check_gptzero_func": 0.2678571343421936, + "rewards/check_perplexity_diff_func": 0.6071428656578064, + "rewards/check_winston_local_func": 0.34988275170326233, + "step": 29 + }, + { + "clip_ratio": 0.0016208746237680316, + "epoch": 0.04195804195804196, + "grad_norm": 0.17335125351231856, + "kl": 0.000507354736328125, + "learning_rate": 1.048951048951049e-05, + "loss": 0.0059, + "step": 30 + }, + { + "clip_ratio": 0.0011572305811569095, + "completion_length": 82.91072082519531, + "epoch": 0.043356643356643354, + "grad_norm": 0.11830097508590764, + "kl": 0.0006256103515625, + "learning_rate": 1.0839160839160838e-05, + "loss": 0.002, + "num_tokens": 310163.0, + "reward": 0.9069231748580933, + "reward_std": 0.4621748924255371, + "rewards/check_gptzero_func": 0.25, + "rewards/check_perplexity_diff_func": 0.3571428656578064, + "rewards/check_winston_local_func": 0.29978030920028687, + "step": 31 + }, + { + "clip_ratio": 0.00402362085878849, + "epoch": 0.044755244755244755, + "grad_norm": 0.11861334422671065, + "kl": 0.000762939453125, + "learning_rate": 1.118881118881119e-05, + "loss": 0.0019, + "step": 32 + }, + { + "clip_ratio": 0.0017474278574809432, + "completion_length": 90.9464340209961, + "epoch": 0.046153846153846156, + "grad_norm": 0.1025496804751534, + "kl": 0.000701904296875, + "learning_rate": 1.153846153846154e-05, + "loss": -0.0014, + "num_tokens": 328922.0, + "reward": 1.1198338270187378, + "reward_std": 0.5045111775398254, + "rewards/check_gptzero_func": 0.0892857164144516, + "rewards/check_perplexity_diff_func": 0.7857142686843872, + "rewards/check_winston_local_func": 0.24483375251293182, + "step": 33 + }, + { + "clip_ratio": 0.0012307984288781881, + "epoch": 0.04755244755244755, + "grad_norm": 0.10274008519956918, + "kl": 0.00087738037109375, + "learning_rate": 1.188811188811189e-05, + "loss": -0.0015, + "step": 34 + }, + { + "clip_ratio": 0.0029886537231504917, + "completion_length": 93.00000762939453, + "epoch": 0.04895104895104895, + "grad_norm": 0.1311802256012859, + "kl": 0.001190185546875, + "learning_rate": 1.2237762237762239e-05, + "loss": -0.0141, + "num_tokens": 348852.0, + "reward": 0.9456299543380737, + "reward_std": 0.6374980807304382, + "rewards/check_gptzero_func": 0.2321428507566452, + "rewards/check_perplexity_diff_func": 0.4285714328289032, + "rewards/check_winston_local_func": 0.28491565585136414, + "step": 35 + }, + { + "clip_ratio": 0.0028984802775084972, + "epoch": 0.05034965034965035, + "grad_norm": 0.13267305607921692, + "kl": 0.00145721435546875, + "learning_rate": 1.2587412587412589e-05, + "loss": -0.0147, + "step": 36 + }, + { + "clip_ratio": 0.0018746532732620835, + "completion_length": 89.85714721679688, + "epoch": 0.05174825174825175, + "grad_norm": 0.1497182815907709, + "kl": 0.002044677734375, + "learning_rate": 1.2937062937062939e-05, + "loss": -0.0253, + "num_tokens": 367482.0, + "reward": 1.2605940103530884, + "reward_std": 0.66156405210495, + "rewards/check_gptzero_func": 0.1785714328289032, + "rewards/check_perplexity_diff_func": 0.75, + "rewards/check_winston_local_func": 0.3320225179195404, + "step": 37 + }, + { + "clip_ratio": 0.0020914783235639334, + "epoch": 0.05314685314685315, + "grad_norm": 0.15027543310666272, + "kl": 0.0026092529296875, + "learning_rate": 1.3286713286713287e-05, + "loss": -0.0257, + "step": 38 + }, + { + "clip_ratio": 0.0017028081929311156, + "completion_length": 101.58928680419922, + "epoch": 0.05454545454545454, + "grad_norm": 0.1306748742984494, + "kl": 0.0027008056640625, + "learning_rate": 1.3636363636363637e-05, + "loss": -0.0091, + "num_tokens": 388365.0, + "reward": 1.1187876462936401, + "reward_std": 0.4044211804866791, + "rewards/check_gptzero_func": 0.2321428507566452, + "rewards/check_perplexity_diff_func": 0.5357142686843872, + "rewards/check_winston_local_func": 0.3509305417537689, + "step": 39 + }, + { + "clip_ratio": 0.003386714030057192, + "epoch": 0.055944055944055944, + "grad_norm": 0.12880289464580882, + "kl": 0.003265380859375, + "learning_rate": 1.3986013986013988e-05, + "loss": -0.0086, + "step": 40 + }, + { + "clip_ratio": 0.0024523374158889055, + "completion_length": 100.41072082519531, + "epoch": 0.057342657342657345, + "grad_norm": 0.12385816200639352, + "kl": 0.004547119140625, + "learning_rate": 1.4335664335664336e-05, + "loss": -0.0161, + "num_tokens": 409074.0, + "reward": 0.804709792137146, + "reward_std": 0.5129754543304443, + "rewards/check_gptzero_func": 0.1428571492433548, + "rewards/check_perplexity_diff_func": 0.3928571343421936, + "rewards/check_winston_local_func": 0.2689954936504364, + "step": 41 + }, + { + "clip_ratio": 0.004267544951289892, + "epoch": 0.05874125874125874, + "grad_norm": 0.12301160152684054, + "kl": 0.005218505859375, + "learning_rate": 1.4685314685314686e-05, + "loss": -0.0161, + "step": 42 + }, + { + "clip_ratio": 0.0019341636216267943, + "completion_length": 90.46428680419922, + "epoch": 0.06013986013986014, + "grad_norm": 0.1390908713431379, + "kl": 0.0048828125, + "learning_rate": 1.5034965034965034e-05, + "loss": 0.0068, + "num_tokens": 428262.0, + "reward": 1.087609887123108, + "reward_std": 0.6448312997817993, + "rewards/check_gptzero_func": 0.2321428507566452, + "rewards/check_perplexity_diff_func": 0.3928571343421936, + "rewards/check_winston_local_func": 0.46260982751846313, + "step": 43 + }, + { + "clip_ratio": 0.0016832423862069845, + "epoch": 0.06153846153846154, + "grad_norm": 0.13873080950458352, + "kl": 0.00567626953125, + "learning_rate": 1.5384615384615387e-05, + "loss": 0.0065, + "step": 44 + }, + { + "clip_ratio": 0.0019954824820160866, + "completion_length": 96.92857360839844, + "epoch": 0.06293706293706294, + "grad_norm": 0.11630124487143968, + "kl": 0.00897216796875, + "learning_rate": 1.5734265734265734e-05, + "loss": 0.008, + "num_tokens": 448480.0, + "reward": 1.0604994297027588, + "reward_std": 0.5378546118736267, + "rewards/check_gptzero_func": 0.2321428507566452, + "rewards/check_perplexity_diff_func": 0.4642857015132904, + "rewards/check_winston_local_func": 0.36407074332237244, + "step": 45 + }, + { + "clip_ratio": 0.002663462422788143, + "epoch": 0.06433566433566433, + "grad_norm": 0.1149796219701467, + "kl": 0.01007080078125, + "learning_rate": 1.6083916083916083e-05, + "loss": 0.0081, + "step": 46 + }, + { + "clip_ratio": 0.0018039485439658165, + "completion_length": 79.51786041259766, + "epoch": 0.06573426573426573, + "grad_norm": 0.12721969778213826, + "kl": 0.01123046875, + "learning_rate": 1.6433566433566433e-05, + "loss": 0.0112, + "num_tokens": 465671.0, + "reward": 0.9921315312385559, + "reward_std": 0.38579052686691284, + "rewards/check_gptzero_func": 0.1607142835855484, + "rewards/check_perplexity_diff_func": 0.3928571343421936, + "rewards/check_winston_local_func": 0.43856000900268555, + "step": 47 + }, + { + "clip_ratio": 0.0027954555116593838, + "epoch": 0.06713286713286713, + "grad_norm": 0.12930272448754576, + "kl": 0.01300048828125, + "learning_rate": 1.6783216783216786e-05, + "loss": 0.0107, + "step": 48 + }, + { + "clip_ratio": 0.0028269642498344183, + "completion_length": 116.5714340209961, + "epoch": 0.06853146853146853, + "grad_norm": 0.12430052891449103, + "kl": 0.01226806640625, + "learning_rate": 1.7132867132867133e-05, + "loss": 0.025, + "num_tokens": 488367.0, + "reward": 0.8830849528312683, + "reward_std": 0.5276607871055603, + "rewards/check_gptzero_func": 0.25, + "rewards/check_perplexity_diff_func": 0.3214285671710968, + "rewards/check_winston_local_func": 0.31165632605552673, + "step": 49 + }, + { + "clip_ratio": 0.0023628019262105227, + "epoch": 0.06993006993006994, + "grad_norm": 0.12293590613717971, + "kl": 0.0142822265625, + "learning_rate": 1.7482517482517483e-05, + "loss": 0.0249, + "step": 50 + }, + { + "clip_ratio": 0.0033425339497625828, + "completion_length": 83.39286041259766, + "epoch": 0.07132867132867132, + "grad_norm": 0.11106847343390151, + "kl": 0.01177978515625, + "learning_rate": 1.7832167832167836e-05, + "loss": -0.0041, + "num_tokens": 506551.0, + "reward": 1.029296875, + "reward_std": 0.4901208281517029, + "rewards/check_gptzero_func": 0.2857142984867096, + "rewards/check_perplexity_diff_func": 0.3214285671710968, + "rewards/check_winston_local_func": 0.42215394973754883, + "step": 51 + }, + { + "clip_ratio": 0.0022109400015324354, + "epoch": 0.07272727272727272, + "grad_norm": 0.11017795132415774, + "kl": 0.013671875, + "learning_rate": 1.8181818181818182e-05, + "loss": -0.0043, + "step": 52 + }, + { + "clip_ratio": 0.0019256824161857367, + "completion_length": 86.5714340209961, + "epoch": 0.07412587412587412, + "grad_norm": 0.11010712395827207, + "kl": 0.0208740234375, + "learning_rate": 1.8531468531468532e-05, + "loss": -0.0032, + "num_tokens": 524863.0, + "reward": 0.962996244430542, + "reward_std": 0.6607940196990967, + "rewards/check_gptzero_func": 0.1071428582072258, + "rewards/check_perplexity_diff_func": 0.5, + "rewards/check_winston_local_func": 0.355853408575058, + "step": 53 + }, + { + "clip_ratio": 0.0025832760147750378, + "epoch": 0.07552447552447553, + "grad_norm": 0.10695378216267973, + "kl": 0.0250244140625, + "learning_rate": 1.888111888111888e-05, + "loss": -0.0034, + "step": 54 + }, + { + "clip_ratio": 0.0017639577854424715, + "completion_length": 96.08928680419922, + "epoch": 0.07692307692307693, + "grad_norm": 0.17472941592839772, + "kl": 0.0322265625, + "learning_rate": 1.923076923076923e-05, + "loss": -0.0141, + "num_tokens": 544736.0, + "reward": 1.3478963375091553, + "reward_std": 0.8002303242683411, + "rewards/check_gptzero_func": 0.2857142984867096, + "rewards/check_perplexity_diff_func": 0.7142857313156128, + "rewards/check_winston_local_func": 0.34789615869522095, + "step": 55 + }, + { + "clip_ratio": 0.0034089265391230583, + "epoch": 0.07832167832167833, + "grad_norm": 0.14666884899599142, + "kl": 0.041748046875, + "learning_rate": 1.958041958041958e-05, + "loss": -0.0155, + "step": 56 + }, + { + "clip_ratio": 0.0026615143287926912, + "completion_length": 96.91072082519531, + "epoch": 0.07972027972027972, + "grad_norm": 0.1400337476406121, + "kl": 0.035400390625, + "learning_rate": 1.993006993006993e-05, + "loss": -0.027, + "num_tokens": 564803.0, + "reward": 1.0594887733459473, + "reward_std": 0.6912091970443726, + "rewards/check_gptzero_func": 0.2857142984867096, + "rewards/check_perplexity_diff_func": 0.4285714328289032, + "rewards/check_winston_local_func": 0.34520307183265686, + "step": 57 + }, + { + "clip_ratio": 0.003153084311634302, + "epoch": 0.08111888111888112, + "grad_norm": 0.138816359725178, + "kl": 0.042236328125, + "learning_rate": 2.027972027972028e-05, + "loss": -0.0274, + "step": 58 + }, + { + "clip_ratio": 0.0012626759707927704, + "completion_length": 76.35714721679688, + "epoch": 0.08251748251748252, + "grad_norm": 0.15940476459351663, + "kl": 0.0556640625, + "learning_rate": 2.062937062937063e-05, + "loss": -0.0057, + "num_tokens": 582265.0, + "reward": 0.992591381072998, + "reward_std": 0.5938137769699097, + "rewards/check_gptzero_func": 0.125, + "rewards/check_perplexity_diff_func": 0.4642857015132904, + "rewards/check_winston_local_func": 0.4033055901527405, + "step": 59 + }, + { + "clip_ratio": 0.004643784370273352, + "epoch": 0.08391608391608392, + "grad_norm": 0.15305834024157844, + "kl": 0.07421875, + "learning_rate": 2.097902097902098e-05, + "loss": -0.0064, + "step": 60 + }, + { + "clip_ratio": 0.0032869603019207716, + "completion_length": 103.33928680419922, + "epoch": 0.08531468531468532, + "grad_norm": 0.12260760902895611, + "kl": 0.0810546875, + "learning_rate": 2.132867132867133e-05, + "loss": 0.0066, + "num_tokens": 603454.0, + "reward": 1.191124439239502, + "reward_std": 0.5557684898376465, + "rewards/check_gptzero_func": 0.2321428507566452, + "rewards/check_perplexity_diff_func": 0.5357142686843872, + "rewards/check_winston_local_func": 0.4232672154903412, + "step": 61 + }, + { + "clip_ratio": 0.008058370091021061, + "epoch": 0.08671328671328671, + "grad_norm": 0.10224499878092017, + "kl": 0.1162109375, + "learning_rate": 2.1678321678321677e-05, + "loss": 0.0058, + "step": 62 + }, + { + "clip_ratio": 0.002202474046498537, + "completion_length": 91.4464340209961, + "epoch": 0.08811188811188811, + "grad_norm": 0.09776882919444682, + "kl": 0.2138671875, + "learning_rate": 2.202797202797203e-05, + "loss": 0.0021, + "num_tokens": 622705.0, + "reward": 1.1320958137512207, + "reward_std": 0.4060821831226349, + "rewards/check_gptzero_func": 0.3928571343421936, + "rewards/check_perplexity_diff_func": 0.3214285671710968, + "rewards/check_winston_local_func": 0.41781002283096313, + "step": 63 + }, + { + "clip_ratio": 0.0025903189089149237, + "epoch": 0.08951048951048951, + "grad_norm": 0.0963149975780195, + "kl": 0.2109375, + "learning_rate": 2.237762237762238e-05, + "loss": 0.0014, + "step": 64 + }, + { + "clip_ratio": 0.0016666523879393935, + "completion_length": 100.00000762939453, + "epoch": 0.09090909090909091, + "grad_norm": 0.12490001498293682, + "kl": 0.1220703125, + "learning_rate": 2.272727272727273e-05, + "loss": 0.0081, + "num_tokens": 643191.0, + "reward": 1.274580478668213, + "reward_std": 0.5277792811393738, + "rewards/check_gptzero_func": 0.2321428507566452, + "rewards/check_perplexity_diff_func": 0.5357142686843872, + "rewards/check_winston_local_func": 0.5067232847213745, + "step": 65 + }, + { + "clip_ratio": 0.0032784033101052046, + "epoch": 0.09230769230769231, + "grad_norm": 0.12114304348028067, + "kl": 0.134765625, + "learning_rate": 2.307692307692308e-05, + "loss": 0.0074, + "step": 66 + }, + { + "clip_ratio": 0.0017520035617053509, + "completion_length": 95.16072082519531, + "epoch": 0.0937062937062937, + "grad_norm": 0.14292853522837679, + "kl": 0.216796875, + "learning_rate": 2.342657342657343e-05, + "loss": -0.0018, + "num_tokens": 662736.0, + "reward": 1.1007850170135498, + "reward_std": 0.45985397696495056, + "rewards/check_gptzero_func": 0.25, + "rewards/check_perplexity_diff_func": 0.4642857015132904, + "rewards/check_winston_local_func": 0.38649922609329224, + "step": 67 + }, + { + "clip_ratio": 0.00579434959217906, + "epoch": 0.0951048951048951, + "grad_norm": 0.14340586759573087, + "kl": 0.2431640625, + "learning_rate": 2.377622377622378e-05, + "loss": -0.0029, + "step": 68 + }, + { + "clip_ratio": 0.0022546499967575073, + "completion_length": 86.17857360839844, + "epoch": 0.0965034965034965, + "grad_norm": 0.17287042335415018, + "kl": 0.26171875, + "learning_rate": 2.4125874125874125e-05, + "loss": 0.0059, + "num_tokens": 681030.0, + "reward": 1.4391711950302124, + "reward_std": 0.6516181230545044, + "rewards/check_gptzero_func": 0.5178571343421936, + "rewards/check_perplexity_diff_func": 0.4642857015132904, + "rewards/check_winston_local_func": 0.45702821016311646, + "step": 69 + }, + { + "clip_ratio": 0.004240955226123333, + "epoch": 0.0979020979020979, + "grad_norm": 0.16371494613893736, + "kl": 0.28125, + "learning_rate": 2.4475524475524478e-05, + "loss": 0.0048, + "step": 70 + }, + { + "clip_ratio": 0.003548440057784319, + "completion_length": 102.42857360839844, + "epoch": 0.0993006993006993, + "grad_norm": 0.1615418523026911, + "kl": 0.1708984375, + "learning_rate": 2.4825174825174828e-05, + "loss": 0.0015, + "num_tokens": 701874.0, + "reward": 1.4929367303848267, + "reward_std": 0.6083499193191528, + "rewards/check_gptzero_func": 0.3214285671710968, + "rewards/check_perplexity_diff_func": 0.7142857313156128, + "rewards/check_winston_local_func": 0.4572224020957947, + "step": 71 + }, + { + "clip_ratio": 0.004331245087087154, + "epoch": 0.1006993006993007, + "grad_norm": 0.12876017762018707, + "kl": 0.216796875, + "learning_rate": 2.5174825174825178e-05, + "loss": 0.0, + "step": 72 + }, + { + "clip_ratio": 0.003951852675527334, + "completion_length": 73.46428680419922, + "epoch": 0.1020979020979021, + "grad_norm": 0.20791748542634808, + "kl": 0.365234375, + "learning_rate": 2.5524475524475528e-05, + "loss": -0.0037, + "num_tokens": 718488.0, + "reward": 1.7953797578811646, + "reward_std": 0.5393027067184448, + "rewards/check_gptzero_func": 0.4464285671710968, + "rewards/check_perplexity_diff_func": 0.8571428656578064, + "rewards/check_winston_local_func": 0.49180838465690613, + "step": 73 + }, + { + "clip_ratio": 0.008312534540891647, + "epoch": 0.1034965034965035, + "grad_norm": 0.19232465302245624, + "kl": 0.5390625, + "learning_rate": 2.5874125874125877e-05, + "loss": -0.0053, + "step": 74 + }, + { + "clip_ratio": 0.0021736263297498226, + "completion_length": 97.73214721679688, + "epoch": 0.1048951048951049, + "grad_norm": 0.18504738647098173, + "kl": 0.2421875, + "learning_rate": 2.6223776223776224e-05, + "loss": -0.0023, + "num_tokens": 738619.0, + "reward": 1.284185767173767, + "reward_std": 0.6959513425827026, + "rewards/check_gptzero_func": 0.4464285671710968, + "rewards/check_perplexity_diff_func": 0.3928571343421936, + "rewards/check_winston_local_func": 0.4449000358581543, + "step": 75 + }, + { + "clip_ratio": 0.0033756059128791094, + "epoch": 0.1062937062937063, + "grad_norm": 0.17585639043073228, + "kl": 0.255859375, + "learning_rate": 2.6573426573426574e-05, + "loss": -0.0043, + "step": 76 + }, + { + "clip_ratio": 0.003740633837878704, + "completion_length": 95.67857360839844, + "epoch": 0.1076923076923077, + "grad_norm": 0.21573775826044048, + "kl": 0.2451171875, + "learning_rate": 2.6923076923076923e-05, + "loss": 0.0105, + "num_tokens": 758111.0, + "reward": 1.4672702550888062, + "reward_std": 0.85318922996521, + "rewards/check_gptzero_func": 0.25, + "rewards/check_perplexity_diff_func": 0.7857142686843872, + "rewards/check_winston_local_func": 0.43155592679977417, + "step": 77 + }, + { + "clip_ratio": 0.004517109598964453, + "epoch": 0.10909090909090909, + "grad_norm": 0.2071235680791843, + "kl": 0.302734375, + "learning_rate": 2.7272727272727273e-05, + "loss": 0.009, + "step": 78 + }, + { + "clip_ratio": 0.004641580395400524, + "completion_length": 117.75000762939453, + "epoch": 0.11048951048951049, + "grad_norm": 0.19278645376555428, + "kl": 0.259765625, + "learning_rate": 2.762237762237762e-05, + "loss": -0.0004, + "num_tokens": 780403.0, + "reward": 1.4974778890609741, + "reward_std": 0.59557044506073, + "rewards/check_gptzero_func": 0.4285714328289032, + "rewards/check_perplexity_diff_func": 0.6785714030265808, + "rewards/check_winston_local_func": 0.39033493399620056, + "step": 79 + }, + { + "clip_ratio": 0.006305322516709566, + "epoch": 0.11188811188811189, + "grad_norm": 0.13786850726930885, + "kl": 0.3125, + "learning_rate": 2.7972027972027976e-05, + "loss": -0.0019, + "step": 80 + }, + { + "clip_ratio": 0.0017778041074052453, + "completion_length": 101.03572082519531, + "epoch": 0.11328671328671329, + "grad_norm": 0.18099911272050398, + "kl": 1.7578125, + "learning_rate": 2.8321678321678326e-05, + "loss": 0.0053, + "num_tokens": 800231.0, + "reward": 1.420699954032898, + "reward_std": 0.5104399919509888, + "rewards/check_gptzero_func": 0.2678571343421936, + "rewards/check_perplexity_diff_func": 0.6428571343421936, + "rewards/check_winston_local_func": 0.5099858045578003, + "step": 81 + }, + { + "clip_ratio": 0.0030425102449953556, + "epoch": 0.11468531468531469, + "grad_norm": 0.10725025403922342, + "kl": 0.9921875, + "learning_rate": 2.8671328671328672e-05, + "loss": 0.0043, + "step": 82 + }, + { + "clip_ratio": 0.0030319676734507084, + "completion_length": 93.3214340209961, + "epoch": 0.11608391608391608, + "grad_norm": 0.22858957476463732, + "kl": 0.35546875, + "learning_rate": 2.9020979020979022e-05, + "loss": 0.0021, + "num_tokens": 819773.0, + "reward": 1.4201573133468628, + "reward_std": 0.761705756187439, + "rewards/check_gptzero_func": 0.3392857015132904, + "rewards/check_perplexity_diff_func": 0.5714285969734192, + "rewards/check_winston_local_func": 0.5094431042671204, + "step": 83 + }, + { + "clip_ratio": 0.0027519434224814177, + "epoch": 0.11748251748251748, + "grad_norm": 0.180287994009811, + "kl": 0.392578125, + "learning_rate": 2.9370629370629372e-05, + "loss": -0.0002, + "step": 84 + }, + { + "clip_ratio": 0.0034523813519626856, + "completion_length": 102.66072082519531, + "epoch": 0.11888111888111888, + "grad_norm": 0.25571068617285075, + "kl": 0.494140625, + "learning_rate": 2.972027972027972e-05, + "loss": -0.0103, + "num_tokens": 840280.0, + "reward": 1.7226576805114746, + "reward_std": 0.7545516490936279, + "rewards/check_gptzero_func": 0.5714285969734192, + "rewards/check_perplexity_diff_func": 0.6785714030265808, + "rewards/check_winston_local_func": 0.4726576805114746, + "step": 85 + }, + { + "clip_ratio": 0.00493992306292057, + "epoch": 0.12027972027972028, + "grad_norm": 0.2172932526081326, + "kl": 0.490234375, + "learning_rate": 3.0069930069930068e-05, + "loss": -0.013, + "step": 86 + }, + { + "clip_ratio": 0.003154533449560404, + "completion_length": 98.96428680419922, + "epoch": 0.12167832167832168, + "grad_norm": 0.3651883135439141, + "kl": 0.5859375, + "learning_rate": 3.0419580419580425e-05, + "loss": -0.0129, + "num_tokens": 860422.0, + "reward": 1.6449967622756958, + "reward_std": 0.7163441777229309, + "rewards/check_gptzero_func": 0.3928571343421936, + "rewards/check_perplexity_diff_func": 0.7142857313156128, + "rewards/check_winston_local_func": 0.5378537178039551, + "step": 87 + }, + { + "clip_ratio": 0.007247431669384241, + "epoch": 0.12307692307692308, + "grad_norm": 0.2167770735292921, + "kl": 0.61328125, + "learning_rate": 3.0769230769230774e-05, + "loss": -0.0168, + "step": 88 + }, + { + "clip_ratio": 0.0027036736719310284, + "completion_length": 108.03572082519531, + "epoch": 0.12447552447552447, + "grad_norm": 0.12604084880729077, + "kl": 0.291015625, + "learning_rate": 3.111888111888112e-05, + "loss": -0.0039, + "num_tokens": 881364.0, + "reward": 1.3799673318862915, + "reward_std": 0.6881024837493896, + "rewards/check_gptzero_func": 0.3928571343421936, + "rewards/check_perplexity_diff_func": 0.5714285969734192, + "rewards/check_winston_local_func": 0.4156815707683563, + "step": 89 + }, + { + "clip_ratio": 0.005238416139036417, + "epoch": 0.1258741258741259, + "grad_norm": 0.12103303780018655, + "kl": 0.30859375, + "learning_rate": 3.146853146853147e-05, + "loss": -0.005, + "step": 90 + }, + { + "clip_ratio": 0.003768512513488531, + "completion_length": 122.80357360839844, + "epoch": 0.12727272727272726, + "grad_norm": 0.18423815092289067, + "kl": 0.67578125, + "learning_rate": 3.181818181818182e-05, + "loss": -0.0171, + "num_tokens": 904487.0, + "reward": 1.6101170778274536, + "reward_std": 0.6241902709007263, + "rewards/check_gptzero_func": 0.3392857015132904, + "rewards/check_perplexity_diff_func": 0.7857142686843872, + "rewards/check_winston_local_func": 0.48511695861816406, + "step": 91 + }, + { + "clip_ratio": 0.005407070741057396, + "epoch": 0.12867132867132866, + "grad_norm": 0.15457101650134836, + "kl": 0.66796875, + "learning_rate": 3.216783216783217e-05, + "loss": -0.0182, + "step": 92 + }, + { + "clip_ratio": 0.002042532665655017, + "completion_length": 108.17857360839844, + "epoch": 0.13006993006993006, + "grad_norm": 0.19588538280583437, + "kl": 1.75, + "learning_rate": 3.251748251748252e-05, + "loss": -0.0023, + "num_tokens": 925633.0, + "reward": 1.3932582139968872, + "reward_std": 0.5087055563926697, + "rewards/check_gptzero_func": 0.4107142984867096, + "rewards/check_perplexity_diff_func": 0.4642857015132904, + "rewards/check_winston_local_func": 0.5182580351829529, + "step": 93 + }, + { + "clip_ratio": 0.00876330491155386, + "epoch": 0.13146853146853146, + "grad_norm": 0.1506256243597478, + "kl": 0.80078125, + "learning_rate": 3.2867132867132866e-05, + "loss": -0.0035, + "step": 94 + }, + { + "clip_ratio": 0.003118924330919981, + "completion_length": 77.4464340209961, + "epoch": 0.13286713286713286, + "grad_norm": 0.34784170699852607, + "kl": 0.72265625, + "learning_rate": 3.321678321678322e-05, + "loss": -0.0134, + "num_tokens": 942892.0, + "reward": 1.832355260848999, + "reward_std": 1.064102292060852, + "rewards/check_gptzero_func": 0.4821428656578064, + "rewards/check_perplexity_diff_func": 0.8214285969734192, + "rewards/check_winston_local_func": 0.5287837982177734, + "step": 95 + }, + { + "clip_ratio": 0.00509096821770072, + "epoch": 0.13426573426573427, + "grad_norm": 0.25968424971577175, + "kl": 0.96875, + "learning_rate": 3.356643356643357e-05, + "loss": -0.0176, + "step": 96 + }, + { + "clip_ratio": 0.0037272910121828318, + "completion_length": 95.26786041259766, + "epoch": 0.13566433566433567, + "grad_norm": 0.16914051199341906, + "kl": 0.349609375, + "learning_rate": 3.391608391608392e-05, + "loss": 0.0037, + "num_tokens": 962317.0, + "reward": 1.805726170539856, + "reward_std": 0.831261396408081, + "rewards/check_gptzero_func": 0.4107142984867096, + "rewards/check_perplexity_diff_func": 0.9285714030265808, + "rewards/check_winston_local_func": 0.46644020080566406, + "step": 97 + }, + { + "clip_ratio": 0.0034601751249283552, + "epoch": 0.13706293706293707, + "grad_norm": 0.16538870434198547, + "kl": 0.388671875, + "learning_rate": 3.4265734265734265e-05, + "loss": 0.0019, + "step": 98 + }, + { + "clip_ratio": 0.004487877711653709, + "completion_length": 111.8214340209961, + "epoch": 0.13846153846153847, + "grad_norm": 0.12415078751808917, + "kl": 0.384765625, + "learning_rate": 3.461538461538462e-05, + "loss": -0.001, + "num_tokens": 983513.0, + "reward": 1.2062333822250366, + "reward_std": 0.591974139213562, + "rewards/check_gptzero_func": 0.3214285671710968, + "rewards/check_perplexity_diff_func": 0.3571428656578064, + "rewards/check_winston_local_func": 0.5276618599891663, + "step": 99 + }, + { + "clip_ratio": 0.00284260674379766, + "epoch": 0.13986013986013987, + "grad_norm": 0.1197068572237121, + "kl": 0.376953125, + "learning_rate": 3.4965034965034965e-05, + "loss": -0.0017, + "step": 100 + }, + { + "clip_ratio": 0.0023206709884107113, + "completion_length": 101.08928680419922, + "epoch": 0.14125874125874127, + "grad_norm": 0.16400691003183243, + "kl": 0.30078125, + "learning_rate": 3.531468531468531e-05, + "loss": 0.0033, + "num_tokens": 1003878.0, + "reward": 1.631854772567749, + "reward_std": 0.5223387479782104, + "rewards/check_gptzero_func": 0.3928571343421936, + "rewards/check_perplexity_diff_func": 0.7142857313156128, + "rewards/check_winston_local_func": 0.5247119665145874, + "step": 101 + }, + { + "clip_ratio": 0.0016618981026113033, + "epoch": 0.14265734265734265, + "grad_norm": 0.1573929247787687, + "kl": 0.326171875, + "learning_rate": 3.566433566433567e-05, + "loss": 0.0012, + "step": 102 + }, + { + "clip_ratio": 0.0022201864048838615, + "completion_length": 124.46429443359375, + "epoch": 0.14405594405594405, + "grad_norm": 0.1353861256020412, + "kl": 0.3828125, + "learning_rate": 3.601398601398602e-05, + "loss": 0.005, + "num_tokens": 1027066.0, + "reward": 1.6920486688613892, + "reward_std": 0.48560601472854614, + "rewards/check_gptzero_func": 0.3928571343421936, + "rewards/check_perplexity_diff_func": 0.75, + "rewards/check_winston_local_func": 0.5491914749145508, + "step": 103 + }, + { + "clip_ratio": 0.0029075751081109047, + "epoch": 0.14545454545454545, + "grad_norm": 0.13328591024866146, + "kl": 0.39453125, + "learning_rate": 3.6363636363636364e-05, + "loss": 0.0034, + "step": 104 + }, + { + "clip_ratio": 0.0019614642951637506, + "completion_length": 85.26786041259766, + "epoch": 0.14685314685314685, + "grad_norm": 0.19674478002774554, + "kl": 0.333984375, + "learning_rate": 3.671328671328672e-05, + "loss": -0.0188, + "num_tokens": 1045567.0, + "reward": 1.6822034120559692, + "reward_std": 0.7559517621994019, + "rewards/check_gptzero_func": 0.5357142686843872, + "rewards/check_perplexity_diff_func": 0.6071428656578064, + "rewards/check_winston_local_func": 0.5393460988998413, + "step": 105 + }, + { + "clip_ratio": 0.004534203093498945, + "epoch": 0.14825174825174825, + "grad_norm": 0.16787611299305724, + "kl": 0.3671875, + "learning_rate": 3.7062937062937064e-05, + "loss": -0.0217, + "step": 106 + }, + { + "clip_ratio": 0.003009920008480549, + "completion_length": 106.55357360839844, + "epoch": 0.14965034965034965, + "grad_norm": 0.19751132540015684, + "kl": 0.455078125, + "learning_rate": 3.741258741258741e-05, + "loss": 0.0111, + "num_tokens": 1066466.0, + "reward": 2.178619861602783, + "reward_std": 0.7016831040382385, + "rewards/check_gptzero_func": 0.6071428656578064, + "rewards/check_perplexity_diff_func": 0.8928571343421936, + "rewards/check_winston_local_func": 0.6786197423934937, + "step": 107 + }, + { + "clip_ratio": 0.004473494831472635, + "epoch": 0.15104895104895105, + "grad_norm": 0.1431470008572649, + "kl": 0.4375, + "learning_rate": 3.776223776223776e-05, + "loss": 0.0095, + "step": 108 + }, + { + "clip_ratio": 0.0026541000697761774, + "completion_length": 100.6964340209961, + "epoch": 0.15244755244755245, + "grad_norm": 0.33135604940901486, + "kl": 0.310546875, + "learning_rate": 3.811188811188811e-05, + "loss": 0.0138, + "num_tokens": 1086751.0, + "reward": 1.6315226554870605, + "reward_std": 0.7806248068809509, + "rewards/check_gptzero_func": 0.4642857015132904, + "rewards/check_perplexity_diff_func": 0.6071428656578064, + "rewards/check_winston_local_func": 0.5600939393043518, + "step": 109 + }, + { + "clip_ratio": 0.0028892713598906994, + "epoch": 0.15384615384615385, + "grad_norm": 0.24217207616877234, + "kl": 0.361328125, + "learning_rate": 3.846153846153846e-05, + "loss": 0.0074, + "step": 110 + }, + { + "clip_ratio": 0.0021747422870248556, + "completion_length": 84.76786041259766, + "epoch": 0.15524475524475526, + "grad_norm": 0.15803622964815373, + "kl": 0.66796875, + "learning_rate": 3.8811188811188816e-05, + "loss": -0.0045, + "num_tokens": 1104518.0, + "reward": 1.6692500114440918, + "reward_std": 0.5965060591697693, + "rewards/check_gptzero_func": 0.5535714030265808, + "rewards/check_perplexity_diff_func": 0.6071428656578064, + "rewards/check_winston_local_func": 0.5085356831550598, + "step": 111 + }, + { + "clip_ratio": 0.004626331850886345, + "epoch": 0.15664335664335666, + "grad_norm": 0.14920406573741435, + "kl": 0.7578125, + "learning_rate": 3.916083916083916e-05, + "loss": -0.0064, + "step": 112 + }, + { + "clip_ratio": 0.002694516209885478, + "completion_length": 92.17857360839844, + "epoch": 0.15804195804195803, + "grad_norm": 0.161427063551978, + "kl": 0.361328125, + "learning_rate": 3.9510489510489516e-05, + "loss": 0.0058, + "num_tokens": 1123466.0, + "reward": 1.353637456893921, + "reward_std": 0.5352396965026855, + "rewards/check_gptzero_func": 0.4821428656578064, + "rewards/check_perplexity_diff_func": 0.3571428656578064, + "rewards/check_winston_local_func": 0.5143517255783081, + "step": 113 + }, + { + "clip_ratio": 0.00391918933019042, + "epoch": 0.15944055944055943, + "grad_norm": 0.14561571783883442, + "kl": 0.39453125, + "learning_rate": 3.986013986013986e-05, + "loss": 0.0036, + "step": 114 + }, + { + "clip_ratio": 0.001278667594306171, + "completion_length": 107.12500762939453, + "epoch": 0.16083916083916083, + "grad_norm": 0.1805624176716803, + "kl": 0.3046875, + "learning_rate": 4.020979020979021e-05, + "loss": -0.001, + "num_tokens": 1144619.0, + "reward": 1.6282455921173096, + "reward_std": 0.7325619459152222, + "rewards/check_gptzero_func": 0.4464285671710968, + "rewards/check_perplexity_diff_func": 0.5357142686843872, + "rewards/check_winston_local_func": 0.6461027264595032, + "step": 115 + }, + { + "clip_ratio": 0.0017050534952431917, + "epoch": 0.16223776223776223, + "grad_norm": 0.16942268791902212, + "kl": 0.337890625, + "learning_rate": 4.055944055944056e-05, + "loss": -0.0027, + "step": 116 + }, + { + "clip_ratio": 0.0015462420415133238, + "completion_length": 114.8214340209961, + "epoch": 0.16363636363636364, + "grad_norm": 0.1777133538888581, + "kl": 0.322265625, + "learning_rate": 4.0909090909090915e-05, + "loss": -0.0028, + "num_tokens": 1166185.0, + "reward": 1.566156268119812, + "reward_std": 0.38663557171821594, + "rewards/check_gptzero_func": 0.4642857015132904, + "rewards/check_perplexity_diff_func": 0.5, + "rewards/check_winston_local_func": 0.6018705368041992, + "step": 117 + }, + { + "clip_ratio": 0.0030793240293860435, + "epoch": 0.16503496503496504, + "grad_norm": 0.15084856022622706, + "kl": 0.2890625, + "learning_rate": 4.125874125874126e-05, + "loss": -0.0048, + "step": 118 + }, + { + "clip_ratio": 0.0018360918620601296, + "completion_length": 100.25000762939453, + "epoch": 0.16643356643356644, + "grad_norm": 0.10943003413749244, + "kl": 0.83984375, + "learning_rate": 4.1608391608391614e-05, + "loss": 0.0043, + "num_tokens": 1186317.0, + "reward": 1.7977957725524902, + "reward_std": 0.5172301530838013, + "rewards/check_gptzero_func": 0.5714285969734192, + "rewards/check_perplexity_diff_func": 0.7142857313156128, + "rewards/check_winston_local_func": 0.5120813250541687, + "step": 119 + }, + { + "clip_ratio": 0.0025589358992874622, + "epoch": 0.16783216783216784, + "grad_norm": 0.10431032316313237, + "kl": 0.68359375, + "learning_rate": 4.195804195804196e-05, + "loss": 0.0029, + "step": 120 + }, + { + "clip_ratio": 0.002800893737003207, + "completion_length": 120.91072082519531, + "epoch": 0.16923076923076924, + "grad_norm": 0.1567960287886374, + "kl": 0.37890625, + "learning_rate": 4.230769230769231e-05, + "loss": -0.0016, + "num_tokens": 1208530.0, + "reward": 1.5369055271148682, + "reward_std": 0.7073518633842468, + "rewards/check_gptzero_func": 0.3214285671710968, + "rewards/check_perplexity_diff_func": 0.6785714030265808, + "rewards/check_winston_local_func": 0.5369054079055786, + "step": 121 + }, + { + "clip_ratio": 0.0026409339625388384, + "epoch": 0.17062937062937064, + "grad_norm": 0.15013150065673506, + "kl": 0.37890625, + "learning_rate": 4.265734265734266e-05, + "loss": -0.0035, + "step": 122 + }, + { + "clip_ratio": 0.002211854327470064, + "completion_length": 110.66072082519531, + "epoch": 0.17202797202797201, + "grad_norm": 0.17274409069851862, + "kl": 0.380859375, + "learning_rate": 4.300699300699301e-05, + "loss": -0.0062, + "num_tokens": 1229171.0, + "reward": 1.6140996217727661, + "reward_std": 0.7221139669418335, + "rewards/check_gptzero_func": 0.4642857015132904, + "rewards/check_perplexity_diff_func": 0.5357142686843872, + "rewards/check_winston_local_func": 0.6140995025634766, + "step": 123 + }, + { + "clip_ratio": 0.005159804597496986, + "epoch": 0.17342657342657342, + "grad_norm": 0.13306053466541726, + "kl": 0.443359375, + "learning_rate": 4.335664335664335e-05, + "loss": -0.009, + "step": 124 + }, + { + "clip_ratio": 0.0031610180158168077, + "completion_length": 104.30357360839844, + "epoch": 0.17482517482517482, + "grad_norm": 0.19261275892706695, + "kl": 0.30078125, + "learning_rate": 4.370629370629371e-05, + "loss": -0.0129, + "num_tokens": 1249350.0, + "reward": 1.9047484397888184, + "reward_std": 0.6767317652702332, + "rewards/check_gptzero_func": 0.6071428656578064, + "rewards/check_perplexity_diff_func": 0.6071428656578064, + "rewards/check_winston_local_func": 0.6904626488685608, + "step": 125 + }, + { + "clip_ratio": 0.004231306724250317, + "epoch": 0.17622377622377622, + "grad_norm": 0.1752737652506695, + "kl": 0.33203125, + "learning_rate": 4.405594405594406e-05, + "loss": -0.0162, + "step": 126 + }, + { + "clip_ratio": 0.003981932066380978, + "completion_length": 109.53572082519531, + "epoch": 0.17762237762237762, + "grad_norm": 0.1423250496935692, + "kl": 0.48828125, + "learning_rate": 4.4405594405594406e-05, + "loss": -0.0102, + "num_tokens": 1269848.0, + "reward": 1.7591207027435303, + "reward_std": 0.5321380496025085, + "rewards/check_gptzero_func": 0.5, + "rewards/check_perplexity_diff_func": 0.6785714030265808, + "rewards/check_winston_local_func": 0.5805493593215942, + "step": 127 + }, + { + "clip_ratio": 0.005287667270749807, + "epoch": 0.17902097902097902, + "grad_norm": 0.13255848380623775, + "kl": 0.498046875, + "learning_rate": 4.475524475524476e-05, + "loss": -0.0121, + "step": 128 + }, + { + "clip_ratio": 0.0025025398936122656, + "completion_length": 105.30357360839844, + "epoch": 0.18041958041958042, + "grad_norm": 0.1177341259986552, + "kl": 0.283203125, + "learning_rate": 4.5104895104895105e-05, + "loss": -0.0016, + "num_tokens": 1290033.0, + "reward": 1.7174798250198364, + "reward_std": 0.5096268057823181, + "rewards/check_gptzero_func": 0.5357142686843872, + "rewards/check_perplexity_diff_func": 0.5357142686843872, + "rewards/check_winston_local_func": 0.6460510492324829, + "step": 129 + }, + { + "clip_ratio": 0.0033291254658252, + "epoch": 0.18181818181818182, + "grad_norm": 0.11141937591016414, + "kl": 0.294921875, + "learning_rate": 4.545454545454546e-05, + "loss": -0.0032, + "step": 130 + }, + { + "clip_ratio": 0.0024425899609923363, + "completion_length": 96.30357360839844, + "epoch": 0.18321678321678322, + "grad_norm": 0.23593816935965856, + "kl": 0.341796875, + "learning_rate": 4.5804195804195805e-05, + "loss": -0.006, + "num_tokens": 1309356.0, + "reward": 1.7568891048431396, + "reward_std": 0.7255779504776001, + "rewards/check_gptzero_func": 0.5892857313156128, + "rewards/check_perplexity_diff_func": 0.5357142686843872, + "rewards/check_winston_local_func": 0.6318890452384949, + "step": 131 + }, + { + "clip_ratio": 0.004331678152084351, + "epoch": 0.18461538461538463, + "grad_norm": 0.2135627059179365, + "kl": 0.349609375, + "learning_rate": 4.615384615384616e-05, + "loss": -0.0089, + "step": 132 + }, + { + "clip_ratio": 0.0028827630449086428, + "completion_length": 111.6964340209961, + "epoch": 0.18601398601398603, + "grad_norm": 0.1622976058137894, + "kl": 0.369140625, + "learning_rate": 4.6503496503496505e-05, + "loss": 0.014, + "num_tokens": 1330017.0, + "reward": 1.6692737340927124, + "reward_std": 0.628279983997345, + "rewards/check_gptzero_func": 0.4464285671710968, + "rewards/check_perplexity_diff_func": 0.6428571343421936, + "rewards/check_winston_local_func": 0.5799878835678101, + "step": 133 + }, + { + "clip_ratio": 0.004608546383678913, + "epoch": 0.1874125874125874, + "grad_norm": 0.16503328062657166, + "kl": 0.4140625, + "learning_rate": 4.685314685314686e-05, + "loss": 0.0121, + "step": 134 + }, + { + "clip_ratio": 0.0027426625601947308, + "completion_length": 107.58928680419922, + "epoch": 0.1888111888111888, + "grad_norm": 0.27926954484170907, + "kl": 0.5859375, + "learning_rate": 4.7202797202797204e-05, + "loss": 0.0139, + "num_tokens": 1350164.0, + "reward": 1.9428951740264893, + "reward_std": 0.6565932035446167, + "rewards/check_gptzero_func": 0.6785714030265808, + "rewards/check_perplexity_diff_func": 0.5714285969734192, + "rewards/check_winston_local_func": 0.6928950548171997, + "step": 135 + }, + { + "clip_ratio": 0.00439854059368372, + "epoch": 0.1902097902097902, + "grad_norm": 0.1925514269864402, + "kl": 0.5546875, + "learning_rate": 4.755244755244756e-05, + "loss": 0.0074, + "step": 136 + }, + { + "clip_ratio": 0.0018738384824246168, + "completion_length": 116.4464340209961, + "epoch": 0.1916083916083916, + "grad_norm": 0.14918417137816656, + "kl": 0.58984375, + "learning_rate": 4.7902097902097904e-05, + "loss": 0.0037, + "num_tokens": 1371507.0, + "reward": 2.0137577056884766, + "reward_std": 0.5378711819648743, + "rewards/check_gptzero_func": 0.6071428656578064, + "rewards/check_perplexity_diff_func": 0.6785714030265808, + "rewards/check_winston_local_func": 0.7280434370040894, + "step": 137 + }, + { + "clip_ratio": 0.003529219189658761, + "epoch": 0.193006993006993, + "grad_norm": 0.1301060608041501, + "kl": 0.6171875, + "learning_rate": 4.825174825174825e-05, + "loss": 0.0016, + "step": 138 + }, + { + "clip_ratio": 0.0011985624441877007, + "completion_length": 112.6964340209961, + "epoch": 0.1944055944055944, + "grad_norm": 0.13034934146859325, + "kl": 0.322265625, + "learning_rate": 4.86013986013986e-05, + "loss": -0.0024, + "num_tokens": 1392414.0, + "reward": 1.9355616569519043, + "reward_std": 0.4726971983909607, + "rewards/check_gptzero_func": 0.5892857313156128, + "rewards/check_perplexity_diff_func": 0.75, + "rewards/check_winston_local_func": 0.596275806427002, + "step": 139 + }, + { + "clip_ratio": 0.0032699662260711193, + "epoch": 0.1958041958041958, + "grad_norm": 0.12091032490165694, + "kl": 0.318359375, + "learning_rate": 4.8951048951048956e-05, + "loss": -0.0039, + "step": 140 + }, + { + "clip_ratio": 0.0013384540798142552, + "completion_length": 132.05357360839844, + "epoch": 0.1972027972027972, + "grad_norm": 0.12728573210952063, + "kl": 0.59765625, + "learning_rate": 4.93006993006993e-05, + "loss": 0.0045, + "num_tokens": 1416177.0, + "reward": 1.7924094200134277, + "reward_std": 0.7202263474464417, + "rewards/check_gptzero_func": 0.5714285969734192, + "rewards/check_perplexity_diff_func": 0.7142857313156128, + "rewards/check_winston_local_func": 0.5066950917243958, + "step": 141 + }, + { + "clip_ratio": 0.0017050639726221561, + "epoch": 0.1986013986013986, + "grad_norm": 0.12231361158637834, + "kl": 0.65234375, + "learning_rate": 4.9650349650349656e-05, + "loss": 0.0025, + "step": 142 + }, + { + "clip_ratio": 0.0015860958956182003, + "completion_length": 99.73214721679688, + "epoch": 0.2, + "grad_norm": 0.1853946177277763, + "kl": 0.58984375, + "learning_rate": 5e-05, + "loss": 0.0021, + "num_tokens": 1435732.0, + "reward": 2.0557591915130615, + "reward_std": 0.503829836845398, + "rewards/check_gptzero_func": 0.6964285969734192, + "rewards/check_perplexity_diff_func": 0.75, + "rewards/check_winston_local_func": 0.6093305945396423, + "step": 143 + }, + { + "clip_ratio": 0.002123563550412655, + "epoch": 0.2013986013986014, + "grad_norm": 0.13470126319919157, + "kl": 0.423828125, + "learning_rate": 5.0349650349650356e-05, + "loss": -0.0009, + "step": 144 + }, + { + "clip_ratio": 0.0018757216166704893, + "completion_length": 95.73214721679688, + "epoch": 0.20279720279720279, + "grad_norm": 0.14930190515465938, + "kl": 0.388671875, + "learning_rate": 5.06993006993007e-05, + "loss": -0.0026, + "num_tokens": 1454705.0, + "reward": 2.0560998916625977, + "reward_std": 0.5541732311248779, + "rewards/check_gptzero_func": 0.6964285969734192, + "rewards/check_perplexity_diff_func": 0.5714285969734192, + "rewards/check_winston_local_func": 0.788242518901825, + "step": 145 + }, + { + "clip_ratio": 0.003540371311828494, + "epoch": 0.2041958041958042, + "grad_norm": 0.1304251509197979, + "kl": 0.41796875, + "learning_rate": 5.1048951048951055e-05, + "loss": -0.0046, + "step": 146 + }, + { + "clip_ratio": 0.0031265008728951216, + "completion_length": 80.9464340209961, + "epoch": 0.2055944055944056, + "grad_norm": 0.2087929657840581, + "kl": 1.1328125, + "learning_rate": 5.1398601398601395e-05, + "loss": 0.0026, + "num_tokens": 1472182.0, + "reward": 1.9076076745986938, + "reward_std": 0.46633273363113403, + "rewards/check_gptzero_func": 0.6428571343421936, + "rewards/check_perplexity_diff_func": 0.5, + "rewards/check_winston_local_func": 0.7647504210472107, + "step": 147 + }, + { + "clip_ratio": 0.0029583375435322523, + "epoch": 0.206993006993007, + "grad_norm": 0.178151823438605, + "kl": 0.66015625, + "learning_rate": 5.1748251748251755e-05, + "loss": -0.0014, + "step": 148 + }, + { + "clip_ratio": 0.0020277961157262325, + "completion_length": 126.58929443359375, + "epoch": 0.2083916083916084, + "grad_norm": 0.13736153026044381, + "kl": 0.373046875, + "learning_rate": 5.2097902097902094e-05, + "loss": 0.0127, + "num_tokens": 1494581.0, + "reward": 1.870171308517456, + "reward_std": 0.4412032663822174, + "rewards/check_gptzero_func": 0.4464285671710968, + "rewards/check_perplexity_diff_func": 0.8214285969734192, + "rewards/check_winston_local_func": 0.6023141741752625, + "step": 149 + }, + { + "clip_ratio": 0.0022643795236945152, + "epoch": 0.2097902097902098, + "grad_norm": 0.12942677801199462, + "kl": 0.40625, + "learning_rate": 5.244755244755245e-05, + "loss": 0.0106, + "step": 150 + }, + { + "clip_ratio": 0.001886948710307479, + "completion_length": 95.14286041259766, + "epoch": 0.2111888111888112, + "grad_norm": 0.15555312890089615, + "kl": 0.62890625, + "learning_rate": 5.279720279720281e-05, + "loss": 0.0013, + "num_tokens": 1513403.0, + "reward": 1.9735006093978882, + "reward_std": 0.5218394994735718, + "rewards/check_gptzero_func": 0.5535714030265808, + "rewards/check_perplexity_diff_func": 0.6071428656578064, + "rewards/check_winston_local_func": 0.8127861022949219, + "step": 151 + }, + { + "clip_ratio": 0.003476199461147189, + "epoch": 0.2125874125874126, + "grad_norm": 0.13833602394582134, + "kl": 0.5390625, + "learning_rate": 5.314685314685315e-05, + "loss": -0.0015, + "step": 152 + }, + { + "clip_ratio": 0.0029247915372252464, + "completion_length": 104.1964340209961, + "epoch": 0.213986013986014, + "grad_norm": 0.18332277821774023, + "kl": 0.451171875, + "learning_rate": 5.34965034965035e-05, + "loss": -0.0047, + "num_tokens": 1533154.0, + "reward": 1.977178931236267, + "reward_std": 0.5758928656578064, + "rewards/check_gptzero_func": 0.6785714030265808, + "rewards/check_perplexity_diff_func": 0.6785714030265808, + "rewards/check_winston_local_func": 0.6200361251831055, + "step": 153 + }, + { + "clip_ratio": 0.002894646255299449, + "epoch": 0.2153846153846154, + "grad_norm": 0.12417162413759981, + "kl": 0.48828125, + "learning_rate": 5.384615384615385e-05, + "loss": -0.0082, + "step": 154 + }, + { + "clip_ratio": 0.002533347113057971, + "completion_length": 106.6964340209961, + "epoch": 0.21678321678321677, + "grad_norm": 0.2165608004438418, + "kl": 0.4140625, + "learning_rate": 5.41958041958042e-05, + "loss": 0.0367, + "num_tokens": 1554005.0, + "reward": 2.254620313644409, + "reward_std": 0.6860859394073486, + "rewards/check_gptzero_func": 0.7857142686843872, + "rewards/check_perplexity_diff_func": 0.7142857313156128, + "rewards/check_winston_local_func": 0.7546200752258301, + "step": 155 + }, + { + "clip_ratio": 0.005156368017196655, + "epoch": 0.21818181818181817, + "grad_norm": 0.16185524517286934, + "kl": 0.4609375, + "learning_rate": 5.4545454545454546e-05, + "loss": 0.0322, + "step": 156 + }, + { + "clip_ratio": 0.002192203886806965, + "completion_length": 129.55357360839844, + "epoch": 0.21958041958041957, + "grad_norm": 0.13686573639431518, + "kl": 0.421875, + "learning_rate": 5.48951048951049e-05, + "loss": 0.0095, + "num_tokens": 1576988.0, + "reward": 2.202296495437622, + "reward_std": 0.4840867817401886, + "rewards/check_gptzero_func": 0.6428571343421936, + "rewards/check_perplexity_diff_func": 0.8928571343421936, + "rewards/check_winston_local_func": 0.6665821075439453, + "step": 157 + }, + { + "clip_ratio": 0.0037054666317999363, + "epoch": 0.22097902097902097, + "grad_norm": 0.12040805378810834, + "kl": 0.45703125, + "learning_rate": 5.524475524475524e-05, + "loss": 0.0077, + "step": 158 + }, + { + "clip_ratio": 0.002324200002476573, + "completion_length": 141.10714721679688, + "epoch": 0.22237762237762237, + "grad_norm": 0.1722210877266639, + "kl": 0.9765625, + "learning_rate": 5.55944055944056e-05, + "loss": -0.0049, + "num_tokens": 1600826.0, + "reward": 2.0245378017425537, + "reward_std": 0.45877185463905334, + "rewards/check_gptzero_func": 0.5892857313156128, + "rewards/check_perplexity_diff_func": 0.75, + "rewards/check_winston_local_func": 0.6852518916130066, + "step": 159 + }, + { + "clip_ratio": 0.003470373572781682, + "epoch": 0.22377622377622378, + "grad_norm": 0.13356363193434526, + "kl": 0.65234375, + "learning_rate": 5.594405594405595e-05, + "loss": -0.0085, + "step": 160 + }, + { + "clip_ratio": 0.0026535126380622387, + "completion_length": 109.76786041259766, + "epoch": 0.22517482517482518, + "grad_norm": 0.409404915148595, + "kl": 0.515625, + "learning_rate": 5.629370629370629e-05, + "loss": -0.0051, + "num_tokens": 1621103.0, + "reward": 2.0040969848632812, + "reward_std": 0.5942196249961853, + "rewards/check_gptzero_func": 0.5178571343421936, + "rewards/check_perplexity_diff_func": 0.7857142686843872, + "rewards/check_winston_local_func": 0.7005256414413452, + "step": 161 + }, + { + "clip_ratio": 0.0058495416305959225, + "epoch": 0.22657342657342658, + "grad_norm": 0.16376328147020022, + "kl": 0.5078125, + "learning_rate": 5.664335664335665e-05, + "loss": -0.0129, + "step": 162 + }, + { + "clip_ratio": 0.002924378262832761, + "completion_length": 114.3214340209961, + "epoch": 0.22797202797202798, + "grad_norm": 0.17698045766822795, + "kl": 0.62109375, + "learning_rate": 5.699300699300699e-05, + "loss": -0.0053, + "num_tokens": 1642629.0, + "reward": 2.0913121700286865, + "reward_std": 0.533393144607544, + "rewards/check_gptzero_func": 0.6785714030265808, + "rewards/check_perplexity_diff_func": 0.6785714030265808, + "rewards/check_winston_local_func": 0.7341693639755249, + "step": 163 + }, + { + "clip_ratio": 0.006533453240990639, + "epoch": 0.22937062937062938, + "grad_norm": 0.15566114152311913, + "kl": 0.66796875, + "learning_rate": 5.7342657342657345e-05, + "loss": -0.0085, + "step": 164 + }, + { + "clip_ratio": 0.002033352619037032, + "completion_length": 110.48214721679688, + "epoch": 0.23076923076923078, + "grad_norm": 0.19340316318251602, + "kl": 0.4453125, + "learning_rate": 5.769230769230769e-05, + "loss": 0.0054, + "num_tokens": 1663000.0, + "reward": 1.809409499168396, + "reward_std": 0.5209853053092957, + "rewards/check_gptzero_func": 0.6428571343421936, + "rewards/check_perplexity_diff_func": 0.3571428656578064, + "rewards/check_winston_local_func": 0.809409499168396, + "step": 165 + }, + { + "clip_ratio": 0.005329853855073452, + "epoch": 0.23216783216783216, + "grad_norm": 0.15299329983895626, + "kl": 0.55078125, + "learning_rate": 5.8041958041958044e-05, + "loss": 0.0018, + "step": 166 + }, + { + "clip_ratio": 0.0025733087677508593, + "completion_length": 95.73214721679688, + "epoch": 0.23356643356643356, + "grad_norm": 0.16102285972903455, + "kl": 0.458984375, + "learning_rate": 5.83916083916084e-05, + "loss": -0.0008, + "num_tokens": 1682123.0, + "reward": 1.9705681800842285, + "reward_std": 0.5465096235275269, + "rewards/check_gptzero_func": 0.6428571343421936, + "rewards/check_perplexity_diff_func": 0.6428571343421936, + "rewards/check_winston_local_func": 0.6848538517951965, + "step": 167 + }, + { + "clip_ratio": 0.0034487086813896894, + "epoch": 0.23496503496503496, + "grad_norm": 0.12896214559568192, + "kl": 0.474609375, + "learning_rate": 5.8741258741258744e-05, + "loss": -0.0039, + "step": 168 + }, + { + "clip_ratio": 0.0012721805833280087, + "completion_length": 72.125, + "epoch": 0.23636363636363636, + "grad_norm": 0.19976827264055144, + "kl": 0.5703125, + "learning_rate": 5.90909090909091e-05, + "loss": -0.0046, + "num_tokens": 1698078.0, + "reward": 1.8698246479034424, + "reward_std": 0.4430284798145294, + "rewards/check_gptzero_func": 0.7142857313156128, + "rewards/check_perplexity_diff_func": 0.3571428656578064, + "rewards/check_winston_local_func": 0.7983959913253784, + "step": 169 + }, + { + "clip_ratio": 0.005269515328109264, + "epoch": 0.23776223776223776, + "grad_norm": 0.14855282442633314, + "kl": 0.61328125, + "learning_rate": 5.944055944055944e-05, + "loss": -0.0079, + "step": 170 + }, + { + "clip_ratio": 0.0027731256559491158, + "completion_length": 117.83929443359375, + "epoch": 0.23916083916083916, + "grad_norm": 0.23058249563979205, + "kl": 0.412109375, + "learning_rate": 5.9790209790209796e-05, + "loss": 0.0055, + "num_tokens": 1719673.0, + "reward": 2.2663450241088867, + "reward_std": 0.46630793809890747, + "rewards/check_gptzero_func": 0.6785714030265808, + "rewards/check_perplexity_diff_func": 0.7142857313156128, + "rewards/check_winston_local_func": 0.8734878301620483, + "step": 171 + }, + { + "clip_ratio": 0.0034511894918978214, + "epoch": 0.24055944055944056, + "grad_norm": 0.17368207545699044, + "kl": 0.373046875, + "learning_rate": 6.0139860139860136e-05, + "loss": -0.0007, + "step": 172 + }, + { + "clip_ratio": 0.0032431341242045164, + "completion_length": 126.85714721679688, + "epoch": 0.24195804195804196, + "grad_norm": 0.17995690811350934, + "kl": 0.353515625, + "learning_rate": 6.048951048951049e-05, + "loss": -0.0011, + "num_tokens": 1742569.0, + "reward": 2.258990526199341, + "reward_std": 0.6008436679840088, + "rewards/check_gptzero_func": 0.75, + "rewards/check_perplexity_diff_func": 0.75, + "rewards/check_winston_local_func": 0.7589904069900513, + "step": 173 + }, + { + "clip_ratio": 0.006950656417757273, + "epoch": 0.24335664335664337, + "grad_norm": 0.13892848528506246, + "kl": 0.359375, + "learning_rate": 6.083916083916085e-05, + "loss": -0.0033, + "step": 174 + }, + { + "clip_ratio": 0.0027662119828164577, + "completion_length": 130.8928680419922, + "epoch": 0.24475524475524477, + "grad_norm": 0.10040261340554028, + "kl": 0.345703125, + "learning_rate": 6.118881118881119e-05, + "loss": -0.0032, + "num_tokens": 1765643.0, + "reward": 2.074061632156372, + "reward_std": 0.3745954632759094, + "rewards/check_gptzero_func": 0.6607142686843872, + "rewards/check_perplexity_diff_func": 0.6785714030265808, + "rewards/check_winston_local_func": 0.7347758412361145, + "step": 175 + }, + { + "clip_ratio": 0.0030742601957172155, + "epoch": 0.24615384615384617, + "grad_norm": 0.0921134439221548, + "kl": 0.3359375, + "learning_rate": 6.153846153846155e-05, + "loss": -0.0044, + "step": 176 + }, + { + "clip_ratio": 0.002027927665039897, + "completion_length": 89.73214721679688, + "epoch": 0.24755244755244754, + "grad_norm": 0.1709976566266231, + "kl": 0.50390625, + "learning_rate": 6.188811188811188e-05, + "loss": 0.0051, + "num_tokens": 1783674.0, + "reward": 1.7810758352279663, + "reward_std": 0.4943030774593353, + "rewards/check_gptzero_func": 0.6428571343421936, + "rewards/check_perplexity_diff_func": 0.4285714328289032, + "rewards/check_winston_local_func": 0.7096471786499023, + "step": 177 + }, + { + "clip_ratio": 0.0020612890366464853, + "epoch": 0.24895104895104894, + "grad_norm": 0.1477275348028994, + "kl": 0.50390625, + "learning_rate": 6.223776223776224e-05, + "loss": 0.0011, + "step": 178 + }, + { + "clip_ratio": 0.0021104796323925257, + "completion_length": 132.10714721679688, + "epoch": 0.25034965034965034, + "grad_norm": 0.08842848405697774, + "kl": 0.39453125, + "learning_rate": 6.258741258741259e-05, + "loss": 0.0005, + "num_tokens": 1806754.0, + "reward": 2.107767105102539, + "reward_std": 0.269815057516098, + "rewards/check_gptzero_func": 0.75, + "rewards/check_perplexity_diff_func": 0.5357142686843872, + "rewards/check_winston_local_func": 0.8220529556274414, + "step": 179 + }, + { + "clip_ratio": 0.004407648928463459, + "epoch": 0.2517482517482518, + "grad_norm": 0.07542595290297384, + "kl": 0.416015625, + "learning_rate": 6.293706293706293e-05, + "loss": -0.0006, + "step": 180 + }, + { + "clip_ratio": 0.0018615310546010733, + "completion_length": 89.30357360839844, + "epoch": 0.25314685314685315, + "grad_norm": 0.2895549842864223, + "kl": 0.498046875, + "learning_rate": 6.32867132867133e-05, + "loss": -0.0082, + "num_tokens": 1824795.0, + "reward": 1.8801069259643555, + "reward_std": 0.5523244738578796, + "rewards/check_gptzero_func": 0.6071428656578064, + "rewards/check_perplexity_diff_func": 0.5, + "rewards/check_winston_local_func": 0.7729640603065491, + "step": 181 + }, + { + "clip_ratio": 0.007555535528808832, + "epoch": 0.2545454545454545, + "grad_norm": 0.20874610098022578, + "kl": 0.59375, + "learning_rate": 6.363636363636364e-05, + "loss": -0.016, + "step": 182 + }, + { + "clip_ratio": 0.0016618422232568264, + "completion_length": 84.46428680419922, + "epoch": 0.25594405594405595, + "grad_norm": 0.20653753743630154, + "kl": 0.5625, + "learning_rate": 6.398601398601399e-05, + "loss": 0.0058, + "num_tokens": 1842573.0, + "reward": 2.1319751739501953, + "reward_std": 0.5030335187911987, + "rewards/check_gptzero_func": 0.7321428656578064, + "rewards/check_perplexity_diff_func": 0.5714285969734192, + "rewards/check_winston_local_func": 0.8284037709236145, + "step": 183 + }, + { + "clip_ratio": 0.00579726742580533, + "epoch": 0.2573426573426573, + "grad_norm": 0.15782001907156346, + "kl": 0.578125, + "learning_rate": 6.433566433566433e-05, + "loss": 0.0001, + "step": 184 + }, + { + "clip_ratio": 0.0016176491044461727, + "completion_length": 125.39286041259766, + "epoch": 0.25874125874125875, + "grad_norm": 0.12385736188741701, + "kl": 0.4921875, + "learning_rate": 6.46853146853147e-05, + "loss": -0.0027, + "num_tokens": 1864297.0, + "reward": 1.7732529640197754, + "reward_std": 0.3774341642856598, + "rewards/check_gptzero_func": 0.625, + "rewards/check_perplexity_diff_func": 0.4285714328289032, + "rewards/check_winston_local_func": 0.7196813821792603, + "step": 185 + }, + { + "clip_ratio": 0.00623862212523818, + "epoch": 0.2601398601398601, + "grad_norm": 0.10877531810926387, + "kl": 0.53125, + "learning_rate": 6.503496503496504e-05, + "loss": -0.0048, + "step": 186 + }, + { + "clip_ratio": 0.0031279984395951033, + "completion_length": 98.00000762939453, + "epoch": 0.26153846153846155, + "grad_norm": 0.3661429387824046, + "kl": 0.6328125, + "learning_rate": 6.538461538461539e-05, + "loss": -0.0063, + "num_tokens": 1882931.0, + "reward": 2.0694050788879395, + "reward_std": 0.6090536117553711, + "rewards/check_gptzero_func": 0.6785714030265808, + "rewards/check_perplexity_diff_func": 0.5714285969734192, + "rewards/check_winston_local_func": 0.819405198097229, + "step": 187 + }, + { + "clip_ratio": 0.010017934255301952, + "epoch": 0.2629370629370629, + "grad_norm": 0.34977841113727764, + "kl": 0.65234375, + "learning_rate": 6.573426573426573e-05, + "loss": -0.0185, + "step": 188 + }, + { + "clip_ratio": 0.00222155568189919, + "completion_length": 133.44644165039062, + "epoch": 0.26433566433566436, + "grad_norm": 0.18715243199492715, + "kl": 0.46484375, + "learning_rate": 6.608391608391609e-05, + "loss": 0.0062, + "num_tokens": 1906084.0, + "reward": 2.364468812942505, + "reward_std": 0.5593475699424744, + "rewards/check_gptzero_func": 0.6964285969734192, + "rewards/check_perplexity_diff_func": 0.8571428656578064, + "rewards/check_winston_local_func": 0.8108974099159241, + "step": 189 + }, + { + "clip_ratio": 0.008103223517537117, + "epoch": 0.26573426573426573, + "grad_norm": 0.14058389303467, + "kl": 0.45703125, + "learning_rate": 6.643356643356644e-05, + "loss": 0.0025, + "step": 190 + }, + { + "clip_ratio": 0.0019668787717819214, + "completion_length": 142.07144165039062, + "epoch": 0.26713286713286716, + "grad_norm": 0.14399812178023874, + "kl": 0.43359375, + "learning_rate": 6.678321678321679e-05, + "loss": -0.0031, + "num_tokens": 1930224.0, + "reward": 2.1098926067352295, + "reward_std": 0.5390018820762634, + "rewards/check_gptzero_func": 0.8214285969734192, + "rewards/check_perplexity_diff_func": 0.4642857015132904, + "rewards/check_winston_local_func": 0.8241782784461975, + "step": 191 + }, + { + "clip_ratio": 0.005082913674414158, + "epoch": 0.26853146853146853, + "grad_norm": 0.12024703070308183, + "kl": 0.447265625, + "learning_rate": 6.713286713286715e-05, + "loss": -0.0055, + "step": 192 + }, + { + "clip_ratio": 0.0032251765951514244, + "completion_length": 84.98214721679688, + "epoch": 0.2699300699300699, + "grad_norm": 0.32527529059153654, + "kl": 0.6640625, + "learning_rate": 6.748251748251748e-05, + "loss": 0.0186, + "num_tokens": 1947399.0, + "reward": 1.9443087577819824, + "reward_std": 0.6417390704154968, + "rewards/check_gptzero_func": 0.7678571343421936, + "rewards/check_perplexity_diff_func": 0.2857142984867096, + "rewards/check_winston_local_func": 0.8907372355461121, + "step": 193 + }, + { + "clip_ratio": 0.008085422217845917, + "epoch": 0.27132867132867133, + "grad_norm": 0.21806984416977268, + "kl": 0.69140625, + "learning_rate": 6.783216783216784e-05, + "loss": 0.0068, + "step": 194 + }, + { + "clip_ratio": 0.0025381618179380894, + "completion_length": 91.66072082519531, + "epoch": 0.2727272727272727, + "grad_norm": 0.1831840025015104, + "kl": 0.53125, + "learning_rate": 6.818181818181818e-05, + "loss": 0.0014, + "num_tokens": 1965402.0, + "reward": 2.3979477882385254, + "reward_std": 0.45264866948127747, + "rewards/check_gptzero_func": 0.7678571343421936, + "rewards/check_perplexity_diff_func": 0.7857142686843872, + "rewards/check_winston_local_func": 0.8443759083747864, + "step": 195 + }, + { + "clip_ratio": 0.005082620773464441, + "epoch": 0.27412587412587414, + "grad_norm": 0.14399845631168656, + "kl": 0.515625, + "learning_rate": 6.853146853146853e-05, + "loss": -0.0023, + "step": 196 + }, + { + "clip_ratio": 0.0025359569117426872, + "completion_length": 118.80357360839844, + "epoch": 0.2755244755244755, + "grad_norm": 0.1617404225919446, + "kl": 0.5625, + "learning_rate": 6.888111888111889e-05, + "loss": 0.0146, + "num_tokens": 1986731.0, + "reward": 2.078895330429077, + "reward_std": 0.31289052963256836, + "rewards/check_gptzero_func": 0.8035714030265808, + "rewards/check_perplexity_diff_func": 0.3928571343421936, + "rewards/check_winston_local_func": 0.882466733455658, + "step": 197 + }, + { + "clip_ratio": 0.0062335156835615635, + "epoch": 0.27692307692307694, + "grad_norm": 0.12006362566688389, + "kl": 0.69921875, + "learning_rate": 6.923076923076924e-05, + "loss": 0.0125, + "step": 198 + }, + { + "clip_ratio": 0.0023628328926861286, + "completion_length": 113.76786041259766, + "epoch": 0.2783216783216783, + "grad_norm": 0.19235514876641358, + "kl": 0.58984375, + "learning_rate": 6.958041958041958e-05, + "loss": -0.0069, + "num_tokens": 2007822.0, + "reward": 2.0453529357910156, + "reward_std": 0.576248288154602, + "rewards/check_gptzero_func": 0.7321428656578064, + "rewards/check_perplexity_diff_func": 0.5714285969734192, + "rewards/check_winston_local_func": 0.7417814135551453, + "step": 199 + }, + { + "clip_ratio": 0.0030245708767324686, + "epoch": 0.27972027972027974, + "grad_norm": 0.15320753759426498, + "kl": 0.578125, + "learning_rate": 6.993006993006993e-05, + "loss": -0.0115, + "step": 200 + }, + { + "clip_ratio": 0.0028001146856695414, + "completion_length": 121.51786041259766, + "epoch": 0.2811188811188811, + "grad_norm": 0.2770839119590159, + "kl": 0.5859375, + "learning_rate": 7.027972027972029e-05, + "loss": -0.0321, + "num_tokens": 2030153.0, + "reward": 2.223505735397339, + "reward_std": 0.580276370048523, + "rewards/check_gptzero_func": 0.8392857313156128, + "rewards/check_perplexity_diff_func": 0.5, + "rewards/check_winston_local_func": 0.8842198252677917, + "step": 201 + }, + { + "clip_ratio": 0.004435424692928791, + "epoch": 0.28251748251748254, + "grad_norm": 0.1752123363810394, + "kl": 0.61328125, + "learning_rate": 7.062937062937062e-05, + "loss": -0.0398, + "step": 202 + }, + { + "clip_ratio": 0.003775001736357808, + "completion_length": 132.96429443359375, + "epoch": 0.2839160839160839, + "grad_norm": 0.20528144857768493, + "kl": 0.5390625, + "learning_rate": 7.097902097902098e-05, + "loss": -0.0054, + "num_tokens": 2053611.0, + "reward": 2.4398131370544434, + "reward_std": 0.6120996475219727, + "rewards/check_gptzero_func": 0.8214285969734192, + "rewards/check_perplexity_diff_func": 0.8571428656578064, + "rewards/check_winston_local_func": 0.761241614818573, + "step": 203 + }, + { + "clip_ratio": 0.0037133130244910717, + "epoch": 0.2853146853146853, + "grad_norm": 0.14916355666745199, + "kl": 0.55859375, + "learning_rate": 7.132867132867134e-05, + "loss": -0.0112, + "step": 204 + }, + { + "clip_ratio": 0.002573356730863452, + "completion_length": 109.51786041259766, + "epoch": 0.2867132867132867, + "grad_norm": 0.15672915743695726, + "kl": 1.4921875, + "learning_rate": 7.167832167832168e-05, + "loss": 0.0032, + "num_tokens": 2073846.0, + "reward": 2.221467971801758, + "reward_std": 0.492183119058609, + "rewards/check_gptzero_func": 0.6964285969734192, + "rewards/check_perplexity_diff_func": 0.6428571343421936, + "rewards/check_winston_local_func": 0.8821821212768555, + "step": 205 + }, + { + "clip_ratio": 0.0036338225472718477, + "epoch": 0.2881118881118881, + "grad_norm": 0.27785877634425765, + "kl": 0.74609375, + "learning_rate": 7.202797202797204e-05, + "loss": 0.0013, + "step": 206 + }, + { + "clip_ratio": 0.004580673761665821, + "completion_length": 95.05357360839844, + "epoch": 0.2895104895104895, + "grad_norm": 0.1746155777114636, + "kl": 0.671875, + "learning_rate": 7.237762237762238e-05, + "loss": -0.0068, + "num_tokens": 2092355.0, + "reward": 2.0429205894470215, + "reward_std": 0.36572587490081787, + "rewards/check_gptzero_func": 0.6964285969734192, + "rewards/check_perplexity_diff_func": 0.4642857015132904, + "rewards/check_winston_local_func": 0.8822061419487, + "step": 207 + }, + { + "clip_ratio": 0.007289381232112646, + "epoch": 0.2909090909090909, + "grad_norm": 0.12704093759252294, + "kl": 0.75390625, + "learning_rate": 7.272727272727273e-05, + "loss": -0.0111, + "step": 208 + }, + { + "clip_ratio": 0.0038105440326035023, + "completion_length": 113.4464340209961, + "epoch": 0.2923076923076923, + "grad_norm": 0.20958397036994925, + "kl": 0.6015625, + "learning_rate": 7.307692307692307e-05, + "loss": 0.0082, + "num_tokens": 2113542.0, + "reward": 2.162785530090332, + "reward_std": 0.49807849526405334, + "rewards/check_gptzero_func": 0.75, + "rewards/check_perplexity_diff_func": 0.5714285969734192, + "rewards/check_winston_local_func": 0.8413568139076233, + "step": 209 + }, + { + "clip_ratio": 0.007606918923556805, + "epoch": 0.2937062937062937, + "grad_norm": 0.1766466418371376, + "kl": 0.63671875, + "learning_rate": 7.342657342657343e-05, + "loss": 0.0025, + "step": 210 + }, + { + "clip_ratio": 0.0019986790139228106, + "completion_length": 92.73214721679688, + "epoch": 0.2951048951048951, + "grad_norm": 0.2716166901527731, + "kl": 0.6171875, + "learning_rate": 7.377622377622378e-05, + "loss": -0.0005, + "num_tokens": 2131873.0, + "reward": 2.099912405014038, + "reward_std": 0.5661742687225342, + "rewards/check_gptzero_func": 0.8214285969734192, + "rewards/check_perplexity_diff_func": 0.4285714328289032, + "rewards/check_winston_local_func": 0.8499122858047485, + "step": 211 + }, + { + "clip_ratio": 0.005917501635849476, + "epoch": 0.2965034965034965, + "grad_norm": 0.19270405158731238, + "kl": 0.62890625, + "learning_rate": 7.412587412587413e-05, + "loss": -0.0079, + "step": 212 + }, + { + "clip_ratio": 0.002530448604375124, + "completion_length": 96.51786041259766, + "epoch": 0.29790209790209793, + "grad_norm": 0.22723256705117592, + "kl": 0.69921875, + "learning_rate": 7.447552447552449e-05, + "loss": -0.009, + "num_tokens": 2150462.0, + "reward": 2.1355700492858887, + "reward_std": 0.3738899827003479, + "rewards/check_gptzero_func": 0.8571428656578064, + "rewards/check_perplexity_diff_func": 0.3571428656578064, + "rewards/check_winston_local_func": 0.9212842583656311, + "step": 213 + }, + { + "clip_ratio": 0.011397747322916985, + "epoch": 0.2993006993006993, + "grad_norm": 0.18189402035580596, + "kl": 0.7265625, + "learning_rate": 7.482517482517482e-05, + "loss": -0.0139, + "step": 214 + }, + { + "clip_ratio": 0.004513947293162346, + "completion_length": 94.66072082519531, + "epoch": 0.3006993006993007, + "grad_norm": 0.2675295151019955, + "kl": 0.67578125, + "learning_rate": 7.517482517482518e-05, + "loss": 0.0014, + "num_tokens": 2169243.0, + "reward": 2.2817916870117188, + "reward_std": 0.4118366539478302, + "rewards/check_gptzero_func": 0.8571428656578064, + "rewards/check_perplexity_diff_func": 0.4642857015132904, + "rewards/check_winston_local_func": 0.9603630900382996, + "step": 215 + }, + { + "clip_ratio": 0.010964194312691689, + "epoch": 0.3020979020979021, + "grad_norm": 0.17031623307626814, + "kl": 0.875, + "learning_rate": 7.552447552447553e-05, + "loss": -0.0057, + "step": 216 + }, + { + "clip_ratio": 0.003371638245880604, + "completion_length": 118.17857360839844, + "epoch": 0.3034965034965035, + "grad_norm": 0.1847364030092435, + "kl": 0.60546875, + "learning_rate": 7.587412587412587e-05, + "loss": -0.0022, + "num_tokens": 2190571.0, + "reward": 2.1225900650024414, + "reward_std": 0.5403507947921753, + "rewards/check_gptzero_func": 0.75, + "rewards/check_perplexity_diff_func": 0.4642857015132904, + "rewards/check_winston_local_func": 0.9083043336868286, + "step": 217 + }, + { + "clip_ratio": 0.01002263929694891, + "epoch": 0.3048951048951049, + "grad_norm": 0.1449002041728082, + "kl": 0.61328125, + "learning_rate": 7.622377622377622e-05, + "loss": -0.0067, + "step": 218 + }, + { + "clip_ratio": 0.0013253266224637628, + "completion_length": 129.33929443359375, + "epoch": 0.3062937062937063, + "grad_norm": 0.1504069021149342, + "kl": 0.578125, + "learning_rate": 7.657342657342658e-05, + "loss": -0.0075, + "num_tokens": 2213408.0, + "reward": 2.1974618434906006, + "reward_std": 0.4602964520454407, + "rewards/check_gptzero_func": 0.7857142686843872, + "rewards/check_perplexity_diff_func": 0.5, + "rewards/check_winston_local_func": 0.911747395992279, + "step": 219 + }, + { + "clip_ratio": 0.006876418832689524, + "epoch": 0.3076923076923077, + "grad_norm": 0.11992971275284438, + "kl": 0.57421875, + "learning_rate": 7.692307692307693e-05, + "loss": -0.0101, + "step": 220 + }, + { + "clip_ratio": 0.0024474726524204016, + "completion_length": 89.67857360839844, + "epoch": 0.3090909090909091, + "grad_norm": 0.46915411380362765, + "kl": 0.73046875, + "learning_rate": 7.727272727272727e-05, + "loss": 0.0333, + "num_tokens": 2231248.0, + "reward": 2.161350727081299, + "reward_std": 0.47136664390563965, + "rewards/check_gptzero_func": 0.8214285969734192, + "rewards/check_perplexity_diff_func": 0.4642857015132904, + "rewards/check_winston_local_func": 0.8756363987922668, + "step": 221 + }, + { + "clip_ratio": 0.02464432455599308, + "epoch": 0.3104895104895105, + "grad_norm": 0.30204889343007607, + "kl": 0.8046875, + "learning_rate": 7.762237762237763e-05, + "loss": 0.0211, + "step": 222 + }, + { + "clip_ratio": 0.0017527465242892504, + "completion_length": 134.17857360839844, + "epoch": 0.3118881118881119, + "grad_norm": 0.18948229022832336, + "kl": 0.6328125, + "learning_rate": 7.797202797202798e-05, + "loss": -0.0049, + "num_tokens": 2255162.0, + "reward": 2.4446465969085693, + "reward_std": 0.5171672105789185, + "rewards/check_gptzero_func": 0.7678571343421936, + "rewards/check_perplexity_diff_func": 0.7857142686843872, + "rewards/check_winston_local_func": 0.8910752534866333, + "step": 223 + }, + { + "clip_ratio": 0.006593752186745405, + "epoch": 0.3132867132867133, + "grad_norm": 0.15013602627394823, + "kl": 0.640625, + "learning_rate": 7.832167832167832e-05, + "loss": -0.0097, + "step": 224 + }, + { + "clip_ratio": 0.0019926591776311398, + "completion_length": 127.00000762939453, + "epoch": 0.3146853146853147, + "grad_norm": 0.22875481051435598, + "kl": 0.6875, + "learning_rate": 7.867132867132867e-05, + "loss": 0.0021, + "num_tokens": 2277320.0, + "reward": 2.2040371894836426, + "reward_std": 0.5528277158737183, + "rewards/check_gptzero_func": 0.7321428656578064, + "rewards/check_perplexity_diff_func": 0.5357142686843872, + "rewards/check_winston_local_func": 0.936180055141449, + "step": 225 + }, + { + "clip_ratio": 0.005839359946548939, + "epoch": 0.31608391608391606, + "grad_norm": 0.17829200080123941, + "kl": 0.68359375, + "learning_rate": 7.902097902097903e-05, + "loss": -0.0053, + "step": 226 + }, + { + "clip_ratio": 0.0026252593379467726, + "completion_length": 130.0357208251953, + "epoch": 0.3174825174825175, + "grad_norm": 0.16076101957107508, + "kl": 0.66015625, + "learning_rate": 7.937062937062938e-05, + "loss": -0.0033, + "num_tokens": 2299818.0, + "reward": 2.3278117179870605, + "reward_std": 0.48680880665779114, + "rewards/check_gptzero_func": 0.8035714030265808, + "rewards/check_perplexity_diff_func": 0.6428571343421936, + "rewards/check_winston_local_func": 0.8813830614089966, + "step": 227 + }, + { + "clip_ratio": 0.004954828415066004, + "epoch": 0.31888111888111886, + "grad_norm": 0.12734264227867195, + "kl": 0.6875, + "learning_rate": 7.972027972027972e-05, + "loss": -0.0071, + "step": 228 + }, + { + "clip_ratio": 0.0034341278951615095, + "completion_length": 133.7857208251953, + "epoch": 0.3202797202797203, + "grad_norm": 0.24065024754984402, + "kl": 0.6875, + "learning_rate": 8.006993006993007e-05, + "loss": -0.0052, + "num_tokens": 2322948.0, + "reward": 2.5632615089416504, + "reward_std": 0.6206622123718262, + "rewards/check_gptzero_func": 0.9642857313156128, + "rewards/check_perplexity_diff_func": 0.6785714030265808, + "rewards/check_winston_local_func": 0.9204041361808777, + "step": 229 + }, + { + "clip_ratio": 0.006521092262119055, + "epoch": 0.32167832167832167, + "grad_norm": 0.18602361781154253, + "kl": 0.6796875, + "learning_rate": 8.041958041958042e-05, + "loss": -0.013, + "step": 230 + }, + { + "clip_ratio": 0.004207184072583914, + "completion_length": 116.53572082519531, + "epoch": 0.3230769230769231, + "grad_norm": 0.36704569865090053, + "kl": 0.69921875, + "learning_rate": 8.076923076923078e-05, + "loss": 0.0055, + "num_tokens": 2344106.0, + "reward": 2.4917171001434326, + "reward_std": 0.6073458790779114, + "rewards/check_gptzero_func": 0.8392857313156128, + "rewards/check_perplexity_diff_func": 0.75, + "rewards/check_winston_local_func": 0.902431309223175, + "step": 231 + }, + { + "clip_ratio": 0.013317321427166462, + "epoch": 0.32447552447552447, + "grad_norm": 0.13844634287775134, + "kl": 0.73046875, + "learning_rate": 8.111888111888112e-05, + "loss": -0.0039, + "step": 232 + }, + { + "clip_ratio": 0.004024635534733534, + "completion_length": 134.92857360839844, + "epoch": 0.3258741258741259, + "grad_norm": 0.34168858248313216, + "kl": 0.64453125, + "learning_rate": 8.146853146853147e-05, + "loss": -0.0057, + "num_tokens": 2367724.0, + "reward": 2.20210599899292, + "reward_std": 0.5888614654541016, + "rewards/check_gptzero_func": 0.7321428656578064, + "rewards/check_perplexity_diff_func": 0.6071428656578064, + "rewards/check_winston_local_func": 0.8628200888633728, + "step": 233 + }, + { + "clip_ratio": 0.009708845987915993, + "epoch": 0.32727272727272727, + "grad_norm": 0.19135959991877635, + "kl": 0.67578125, + "learning_rate": 8.181818181818183e-05, + "loss": -0.0146, + "step": 234 + }, + { + "clip_ratio": 0.005571329966187477, + "completion_length": 98.75000762939453, + "epoch": 0.32867132867132864, + "grad_norm": 0.32328761096657005, + "kl": 0.77734375, + "learning_rate": 8.216783216783218e-05, + "loss": 0.0128, + "num_tokens": 2386636.0, + "reward": 2.2358410358428955, + "reward_std": 0.599101185798645, + "rewards/check_gptzero_func": 0.8214285969734192, + "rewards/check_perplexity_diff_func": 0.5357142686843872, + "rewards/check_winston_local_func": 0.8786982297897339, + "step": 235 + }, + { + "clip_ratio": 0.009436835534870625, + "epoch": 0.3300699300699301, + "grad_norm": 0.19557085227388898, + "kl": 0.8515625, + "learning_rate": 8.251748251748252e-05, + "loss": 0.0022, + "step": 236 + }, + { + "clip_ratio": 0.006164635997265577, + "completion_length": 79.75, + "epoch": 0.33146853146853145, + "grad_norm": 0.3250433111655094, + "kl": 0.98828125, + "learning_rate": 8.286713286713287e-05, + "loss": -0.0126, + "num_tokens": 2403400.0, + "reward": 2.1074206829071045, + "reward_std": 0.41642776131629944, + "rewards/check_gptzero_func": 0.9107142686843872, + "rewards/check_perplexity_diff_func": 0.25, + "rewards/check_winston_local_func": 0.946706235408783, + "step": 237 + }, + { + "clip_ratio": 0.0066314926370978355, + "epoch": 0.3328671328671329, + "grad_norm": 0.1612667131666411, + "kl": 0.98828125, + "learning_rate": 8.321678321678323e-05, + "loss": -0.0225, + "step": 238 + }, + { + "clip_ratio": 0.003079179208725691, + "completion_length": 117.10714721679688, + "epoch": 0.33426573426573425, + "grad_norm": 0.2505534626607256, + "kl": 0.6796875, + "learning_rate": 8.356643356643356e-05, + "loss": -0.0103, + "num_tokens": 2425354.0, + "reward": 2.3166799545288086, + "reward_std": 0.5270359516143799, + "rewards/check_gptzero_func": 0.8214285969734192, + "rewards/check_perplexity_diff_func": 0.5714285969734192, + "rewards/check_winston_local_func": 0.9238227605819702, + "step": 239 + }, + { + "clip_ratio": 0.006801524665206671, + "epoch": 0.3356643356643357, + "grad_norm": 0.16826419365505382, + "kl": 0.671875, + "learning_rate": 8.391608391608392e-05, + "loss": -0.0182, + "step": 240 + }, + { + "clip_ratio": 0.004480496048927307, + "completion_length": 125.05357360839844, + "epoch": 0.33706293706293705, + "grad_norm": 0.26807661638826913, + "kl": 0.60546875, + "learning_rate": 8.426573426573428e-05, + "loss": -0.0086, + "num_tokens": 2447333.0, + "reward": 2.2939677238464355, + "reward_std": 0.39558398723602295, + "rewards/check_gptzero_func": 0.9107142686843872, + "rewards/check_perplexity_diff_func": 0.5357142686843872, + "rewards/check_winston_local_func": 0.8475390672683716, + "step": 241 + }, + { + "clip_ratio": 0.007587091531604528, + "epoch": 0.3384615384615385, + "grad_norm": 0.13602673124742348, + "kl": 0.57421875, + "learning_rate": 8.461538461538461e-05, + "loss": -0.0152, + "step": 242 + }, + { + "clip_ratio": 0.0034923183266073465, + "completion_length": 98.71428680419922, + "epoch": 0.33986013986013985, + "grad_norm": 0.7773422467135167, + "kl": 6.5, + "learning_rate": 8.496503496503497e-05, + "loss": 0.0071, + "num_tokens": 2466435.0, + "reward": 2.4026196002960205, + "reward_std": 0.4264836013317108, + "rewards/check_gptzero_func": 0.8571428656578064, + "rewards/check_perplexity_diff_func": 0.6785714030265808, + "rewards/check_winston_local_func": 0.8669052124023438, + "step": 243 + }, + { + "clip_ratio": 0.026189187541604042, + "epoch": 0.3412587412587413, + "grad_norm": 130.9971047023429, + "kl": 0.70703125, + "learning_rate": 8.531468531468532e-05, + "loss": 1.8998, + "step": 244 + }, + { + "clip_ratio": 0.004780410788953304, + "completion_length": 135.0, + "epoch": 0.34265734265734266, + "grad_norm": 0.2329479243774761, + "kl": 0.87109375, + "learning_rate": 8.566433566433567e-05, + "loss": 0.0011, + "num_tokens": 2489651.0, + "reward": 2.2166569232940674, + "reward_std": 0.5797281265258789, + "rewards/check_gptzero_func": 0.9107142686843872, + "rewards/check_perplexity_diff_func": 0.4285714328289032, + "rewards/check_winston_local_func": 0.8773713111877441, + "step": 245 + }, + { + "clip_ratio": 0.021273447200655937, + "epoch": 0.34405594405594403, + "grad_norm": 0.22672502397660207, + "kl": 0.9609375, + "learning_rate": 8.601398601398601e-05, + "loss": -0.0045, + "step": 246 + }, + { + "clip_ratio": 0.002856873208656907, + "completion_length": 122.66072082519531, + "epoch": 0.34545454545454546, + "grad_norm": 0.19772681964626543, + "kl": 0.984375, + "learning_rate": 8.636363636363637e-05, + "loss": 0.0028, + "num_tokens": 2510506.0, + "reward": 2.323245048522949, + "reward_std": 0.46726977825164795, + "rewards/check_gptzero_func": 0.9107142686843872, + "rewards/check_perplexity_diff_func": 0.5, + "rewards/check_winston_local_func": 0.9125306010246277, + "step": 247 + }, + { + "clip_ratio": 0.02299562282860279, + "epoch": 0.34685314685314683, + "grad_norm": 0.1812005085005865, + "kl": 0.87890625, + "learning_rate": 8.67132867132867e-05, + "loss": -0.001, + "step": 248 + }, + { + "clip_ratio": 0.0026335662696510553, + "completion_length": 116.85714721679688, + "epoch": 0.34825174825174826, + "grad_norm": 0.28731547808119623, + "kl": 0.76953125, + "learning_rate": 8.706293706293707e-05, + "loss": 0.024, + "num_tokens": 2531254.0, + "reward": 2.5312695503234863, + "reward_std": 0.5851892828941345, + "rewards/check_gptzero_func": 0.8571428656578064, + "rewards/check_perplexity_diff_func": 0.8214285969734192, + "rewards/check_winston_local_func": 0.8526979088783264, + "step": 249 + }, + { + "clip_ratio": 0.012148472480475903, + "epoch": 0.34965034965034963, + "grad_norm": 0.21250398409595742, + "kl": 0.734375, + "learning_rate": 8.741258741258743e-05, + "loss": 0.0156, + "step": 250 + }, + { + "clip_ratio": 0.0036881309933960438, + "completion_length": 109.12500762939453, + "epoch": 0.35104895104895106, + "grad_norm": 0.278787459589393, + "kl": 1.3671875, + "learning_rate": 8.776223776223776e-05, + "loss": -0.0127, + "num_tokens": 2551265.0, + "reward": 2.24423885345459, + "reward_std": 0.6221095323562622, + "rewards/check_gptzero_func": 0.875, + "rewards/check_perplexity_diff_func": 0.5, + "rewards/check_winston_local_func": 0.8692389130592346, + "step": 251 + }, + { + "clip_ratio": 0.011438491754233837, + "epoch": 0.35244755244755244, + "grad_norm": 0.21324998809887244, + "kl": 1.0859375, + "learning_rate": 8.811188811188812e-05, + "loss": -0.023, + "step": 252 + }, + { + "clip_ratio": 0.0039367591962218285, + "completion_length": 112.14286041259766, + "epoch": 0.35384615384615387, + "grad_norm": 0.31069817037713615, + "kl": 0.91015625, + "learning_rate": 8.846153846153847e-05, + "loss": 0.0167, + "num_tokens": 2572119.0, + "reward": 2.492161989212036, + "reward_std": 0.6407585144042969, + "rewards/check_gptzero_func": 0.875, + "rewards/check_perplexity_diff_func": 0.75, + "rewards/check_winston_local_func": 0.8671619296073914, + "step": 253 + }, + { + "clip_ratio": 0.005880096461623907, + "epoch": 0.35524475524475524, + "grad_norm": 0.19008007229653096, + "kl": 0.90234375, + "learning_rate": 8.881118881118881e-05, + "loss": 0.0057, + "step": 254 + }, + { + "clip_ratio": 0.003151817014440894, + "completion_length": 120.78572082519531, + "epoch": 0.35664335664335667, + "grad_norm": 0.4113945542101958, + "kl": 0.890625, + "learning_rate": 8.916083916083916e-05, + "loss": 0.0079, + "num_tokens": 2593081.0, + "reward": 2.3228745460510254, + "reward_std": 0.5329480171203613, + "rewards/check_gptzero_func": 0.875, + "rewards/check_perplexity_diff_func": 0.6428571343421936, + "rewards/check_winston_local_func": 0.8050170540809631, + "step": 255 + }, + { + "clip_ratio": 0.01224872563034296, + "epoch": 0.35804195804195804, + "grad_norm": 0.24677625747594412, + "kl": 1.0078125, + "learning_rate": 8.951048951048952e-05, + "loss": 0.0006, + "step": 256 + }, + { + "clip_ratio": 0.003666934324428439, + "completion_length": 146.5178680419922, + "epoch": 0.3594405594405594, + "grad_norm": 0.21537926236506497, + "kl": 0.484375, + "learning_rate": 8.986013986013986e-05, + "loss": 0.0068, + "num_tokens": 2618758.0, + "reward": 2.198406934738159, + "reward_std": 0.3875991106033325, + "rewards/check_gptzero_func": 0.9285714030265808, + "rewards/check_perplexity_diff_func": 0.4285714328289032, + "rewards/check_winston_local_func": 0.8412641882896423, + "step": 257 + }, + { + "clip_ratio": 0.012731654569506645, + "epoch": 0.36083916083916084, + "grad_norm": 0.1543455831571161, + "kl": 0.478515625, + "learning_rate": 9.020979020979021e-05, + "loss": 0.0025, + "step": 258 + }, + { + "clip_ratio": 0.0012050624936819077, + "completion_length": 129.375, + "epoch": 0.3622377622377622, + "grad_norm": 0.16107873823673136, + "kl": 1.6640625, + "learning_rate": 9.055944055944057e-05, + "loss": 0.0054, + "num_tokens": 2641815.0, + "reward": 2.4805288314819336, + "reward_std": 0.3510296940803528, + "rewards/check_gptzero_func": 0.8392857313156128, + "rewards/check_perplexity_diff_func": 0.8214285969734192, + "rewards/check_winston_local_func": 0.8198142647743225, + "step": 259 + }, + { + "clip_ratio": 0.006208478473126888, + "epoch": 0.36363636363636365, + "grad_norm": 0.10546346547579268, + "kl": 0.8046875, + "learning_rate": 9.090909090909092e-05, + "loss": 0.0019, + "step": 260 + }, + { + "clip_ratio": 0.0025593352038413286, + "completion_length": 128.57144165039062, + "epoch": 0.365034965034965, + "grad_norm": 0.1986264685447053, + "kl": 0.6171875, + "learning_rate": 9.125874125874126e-05, + "loss": -0.0025, + "num_tokens": 2664343.0, + "reward": 2.261420249938965, + "reward_std": 0.5704914927482605, + "rewards/check_gptzero_func": 0.7857142686843872, + "rewards/check_perplexity_diff_func": 0.6428571343421936, + "rewards/check_winston_local_func": 0.8328486680984497, + "step": 261 + }, + { + "clip_ratio": 0.002767725382000208, + "epoch": 0.36643356643356645, + "grad_norm": 0.14267481856078115, + "kl": 0.66015625, + "learning_rate": 9.160839160839161e-05, + "loss": -0.0071, + "step": 262 + }, + { + "clip_ratio": 0.0036019759718328714, + "completion_length": 108.30357360839844, + "epoch": 0.3678321678321678, + "grad_norm": 0.25429899311591847, + "kl": 0.61328125, + "learning_rate": 9.195804195804196e-05, + "loss": 0.0114, + "num_tokens": 2684460.0, + "reward": 2.1853349208831787, + "reward_std": 0.5545978546142578, + "rewards/check_gptzero_func": 0.7142857313156128, + "rewards/check_perplexity_diff_func": 0.6428571343421936, + "rewards/check_winston_local_func": 0.828191876411438, + "step": 263 + }, + { + "clip_ratio": 0.002995865885168314, + "epoch": 0.36923076923076925, + "grad_norm": 0.17454926732818832, + "kl": 0.62109375, + "learning_rate": 9.230769230769232e-05, + "loss": 0.0028, + "step": 264 + }, + { + "clip_ratio": 0.002558046253398061, + "completion_length": 102.80357360839844, + "epoch": 0.3706293706293706, + "grad_norm": 0.18701691067416304, + "kl": 0.734375, + "learning_rate": 9.265734265734266e-05, + "loss": 0.0401, + "num_tokens": 2704225.0, + "reward": 2.1770143508911133, + "reward_std": 0.573693037033081, + "rewards/check_gptzero_func": 0.8035714030265808, + "rewards/check_perplexity_diff_func": 0.5357142686843872, + "rewards/check_winston_local_func": 0.8377286195755005, + "step": 265 + }, + { + "clip_ratio": 0.006838452070951462, + "epoch": 0.37202797202797205, + "grad_norm": 0.13942096443359253, + "kl": 0.71875, + "learning_rate": 9.300699300699301e-05, + "loss": 0.0354, + "step": 266 + }, + { + "clip_ratio": 0.00375689216889441, + "completion_length": 106.37500762939453, + "epoch": 0.3734265734265734, + "grad_norm": 0.25266992310982356, + "kl": 0.7421875, + "learning_rate": 9.335664335664336e-05, + "loss": -0.0156, + "num_tokens": 2724530.0, + "reward": 2.11887526512146, + "reward_std": 0.5366143584251404, + "rewards/check_gptzero_func": 0.875, + "rewards/check_perplexity_diff_func": 0.4642857015132904, + "rewards/check_winston_local_func": 0.7795897126197815, + "step": 267 + }, + { + "clip_ratio": 0.006593361962586641, + "epoch": 0.3748251748251748, + "grad_norm": 0.18497903526945206, + "kl": 0.7265625, + "learning_rate": 9.370629370629372e-05, + "loss": -0.0243, + "step": 268 + }, + { + "clip_ratio": 0.004569682292640209, + "completion_length": 124.39286041259766, + "epoch": 0.37622377622377623, + "grad_norm": 0.21331332062489702, + "kl": 0.953125, + "learning_rate": 9.405594405594406e-05, + "loss": 0.0346, + "num_tokens": 2746246.0, + "reward": 2.403602123260498, + "reward_std": 0.5621734857559204, + "rewards/check_gptzero_func": 0.7857142686843872, + "rewards/check_perplexity_diff_func": 0.75, + "rewards/check_winston_local_func": 0.8678879141807556, + "step": 269 + }, + { + "clip_ratio": 0.006006812676787376, + "epoch": 0.3776223776223776, + "grad_norm": 0.15420322253654764, + "kl": 0.8671875, + "learning_rate": 9.440559440559441e-05, + "loss": 0.0283, + "step": 270 + }, + { + "clip_ratio": 0.002397142816334963, + "completion_length": 102.41072082519531, + "epoch": 0.37902097902097903, + "grad_norm": 0.22567114655173642, + "kl": 0.609375, + "learning_rate": 9.475524475524477e-05, + "loss": 0.0231, + "num_tokens": 2765203.0, + "reward": 2.649076223373413, + "reward_std": 0.47521141171455383, + "rewards/check_gptzero_func": 0.875, + "rewards/check_perplexity_diff_func": 1.0, + "rewards/check_winston_local_func": 0.7740762829780579, + "step": 271 + }, + { + "clip_ratio": 0.007639365270733833, + "epoch": 0.3804195804195804, + "grad_norm": 0.18662930190408594, + "kl": 0.578125, + "learning_rate": 9.510489510489511e-05, + "loss": 0.0165, + "step": 272 + }, + { + "clip_ratio": 0.0010056205792352557, + "completion_length": 115.28572082519531, + "epoch": 0.38181818181818183, + "grad_norm": 0.14954423121719349, + "kl": 0.625, + "learning_rate": 9.545454545454546e-05, + "loss": 0.0237, + "num_tokens": 2785985.0, + "reward": 2.3541975021362305, + "reward_std": 0.5445494651794434, + "rewards/check_gptzero_func": 0.8392857313156128, + "rewards/check_perplexity_diff_func": 0.75, + "rewards/check_winston_local_func": 0.7649118304252625, + "step": 273 + }, + { + "clip_ratio": 0.003297280054539442, + "epoch": 0.3832167832167832, + "grad_norm": 0.11713806873969078, + "kl": 0.640625, + "learning_rate": 9.580419580419581e-05, + "loss": 0.0206, + "step": 274 + }, + { + "clip_ratio": 0.0016172031173482537, + "completion_length": 128.73214721679688, + "epoch": 0.38461538461538464, + "grad_norm": 0.155009547701029, + "kl": 0.51171875, + "learning_rate": 9.615384615384617e-05, + "loss": 0.0191, + "num_tokens": 2808414.0, + "reward": 2.376694679260254, + "reward_std": 0.33215656876564026, + "rewards/check_gptzero_func": 0.8392857313156128, + "rewards/check_perplexity_diff_func": 0.8571428656578064, + "rewards/check_winston_local_func": 0.6802659630775452, + "step": 275 + }, + { + "clip_ratio": 0.004294655751436949, + "epoch": 0.386013986013986, + "grad_norm": 0.11409067239204185, + "kl": 0.515625, + "learning_rate": 9.65034965034965e-05, + "loss": 0.0155, + "step": 276 + }, + { + "clip_ratio": 0.002925862092524767, + "completion_length": 116.41072082519531, + "epoch": 0.38741258741258744, + "grad_norm": 0.2764503095198425, + "kl": 0.578125, + "learning_rate": 9.685314685314686e-05, + "loss": 0.01, + "num_tokens": 2829285.0, + "reward": 2.3352210521698, + "reward_std": 0.7207943201065063, + "rewards/check_gptzero_func": 0.6607142686843872, + "rewards/check_perplexity_diff_func": 0.9642857313156128, + "rewards/check_winston_local_func": 0.7102211713790894, + "step": 277 + }, + { + "clip_ratio": 0.005557883996516466, + "epoch": 0.3888111888111888, + "grad_norm": 0.2140511595370052, + "kl": 0.5859375, + "learning_rate": 9.72027972027972e-05, + "loss": -0.0004, + "step": 278 + }, + { + "clip_ratio": 0.0029403052758425474, + "completion_length": 114.25000762939453, + "epoch": 0.3902097902097902, + "grad_norm": 0.21363596109825803, + "kl": 0.63671875, + "learning_rate": 9.755244755244755e-05, + "loss": 0.0259, + "num_tokens": 2849989.0, + "reward": 2.6570937633514404, + "reward_std": 0.5007120370864868, + "rewards/check_gptzero_func": 0.75, + "rewards/check_perplexity_diff_func": 1.2142857313156128, + "rewards/check_winston_local_func": 0.6928080320358276, + "step": 279 + }, + { + "clip_ratio": 0.00384224159643054, + "epoch": 0.3916083916083916, + "grad_norm": 0.15102706915256958, + "kl": 0.6875, + "learning_rate": 9.790209790209791e-05, + "loss": 0.0185, + "step": 280 + }, + { + "clip_ratio": 0.0027283646631985903, + "completion_length": 90.25000762939453, + "epoch": 0.393006993006993, + "grad_norm": 0.3585134615134873, + "kl": 0.73046875, + "learning_rate": 9.825174825174826e-05, + "loss": -0.006, + "num_tokens": 2867567.0, + "reward": 2.4528698921203613, + "reward_std": 0.6383817791938782, + "rewards/check_gptzero_func": 0.6071428656578064, + "rewards/check_perplexity_diff_func": 1.1428571939468384, + "rewards/check_winston_local_func": 0.7028695940971375, + "step": 281 + }, + { + "clip_ratio": 0.016295742243528366, + "epoch": 0.3944055944055944, + "grad_norm": 0.2358766583499307, + "kl": 0.73046875, + "learning_rate": 9.86013986013986e-05, + "loss": -0.0151, + "step": 282 + }, + { + "clip_ratio": 0.0025735762901604176, + "completion_length": 132.125, + "epoch": 0.3958041958041958, + "grad_norm": 0.12427534341566092, + "kl": 0.515625, + "learning_rate": 9.895104895104895e-05, + "loss": 0.006, + "num_tokens": 2890954.0, + "reward": 2.172079086303711, + "reward_std": 0.31105130910873413, + "rewards/check_gptzero_func": 0.6607142686843872, + "rewards/check_perplexity_diff_func": 0.7857142686843872, + "rewards/check_winston_local_func": 0.7256504893302917, + "step": 283 + }, + { + "clip_ratio": 0.008568070828914642, + "epoch": 0.3972027972027972, + "grad_norm": 0.10751149742872082, + "kl": 0.5234375, + "learning_rate": 9.930069930069931e-05, + "loss": 0.0038, + "step": 284 + }, + { + "clip_ratio": 0.0013303400482982397, + "completion_length": 115.50000762939453, + "epoch": 0.3986013986013986, + "grad_norm": 0.11155145645226237, + "kl": 0.427734375, + "learning_rate": 9.965034965034964e-05, + "loss": 0.0136, + "num_tokens": 2912320.0, + "reward": 2.499699354171753, + "reward_std": 0.22870054841041565, + "rewards/check_gptzero_func": 0.6071428656578064, + "rewards/check_perplexity_diff_func": 1.3214285373687744, + "rewards/check_winston_local_func": 0.5711276531219482, + "step": 285 + }, + { + "clip_ratio": 0.004690885543823242, + "epoch": 0.4, + "grad_norm": 0.08246302090980155, + "kl": 0.443359375, + "learning_rate": 0.0001, + "loss": 0.0117, + "step": 286 + }, + { + "clip_ratio": 0.0013457606546580791, + "completion_length": 153.60714721679688, + "epoch": 0.4013986013986014, + "grad_norm": 0.0959414180998777, + "kl": 0.458984375, + "learning_rate": 9.999996275889018e-05, + "loss": -0.0037, + "num_tokens": 2937286.0, + "reward": 2.60290789604187, + "reward_std": 0.2643657624721527, + "rewards/check_gptzero_func": 0.625, + "rewards/check_perplexity_diff_func": 1.3214285373687744, + "rewards/check_winston_local_func": 0.6564791798591614, + "step": 287 + }, + { + "clip_ratio": 0.0025680752005428076, + "epoch": 0.4027972027972028, + "grad_norm": 0.08295696905412306, + "kl": 0.453125, + "learning_rate": 9.999985103561615e-05, + "loss": -0.0053, + "step": 288 + }, + { + "clip_ratio": 0.0, + "completion_length": 106.39286041259766, + "epoch": 0.4041958041958042, + "grad_norm": 0.12524695486629228, + "kl": 0.3671875, + "learning_rate": 9.999966483034437e-05, + "loss": -0.0031, + "num_tokens": 2957572.0, + "reward": 2.607825994491577, + "reward_std": 0.2113000899553299, + "rewards/check_gptzero_func": 0.5535714030265808, + "rewards/check_perplexity_diff_func": 1.4642857313156128, + "rewards/check_winston_local_func": 0.5899689793586731, + "step": 289 + }, + { + "clip_ratio": 0.0015743181575089693, + "epoch": 0.40559440559440557, + "grad_norm": 0.08958540244405325, + "kl": 0.36328125, + "learning_rate": 9.999940414335222e-05, + "loss": -0.0046, + "step": 290 + }, + { + "clip_ratio": 0.0017556955572217703, + "completion_length": 96.0714340209961, + "epoch": 0.406993006993007, + "grad_norm": 0.19185238369967877, + "kl": 2.109375, + "learning_rate": 9.999906897502803e-05, + "loss": 0.008, + "num_tokens": 2976076.0, + "reward": 2.6148557662963867, + "reward_std": 0.24822747707366943, + "rewards/check_gptzero_func": 0.5178571343421936, + "rewards/check_perplexity_diff_func": 1.4285714626312256, + "rewards/check_winston_local_func": 0.6684269309043884, + "step": 291 + }, + { + "clip_ratio": 0.00337741756811738, + "epoch": 0.4083916083916084, + "grad_norm": 0.08769186584743054, + "kl": 0.671875, + "learning_rate": 9.999865932587107e-05, + "loss": 0.0046, + "step": 292 + }, + { + "clip_ratio": 0.0012849332997575402, + "completion_length": 91.87500762939453, + "epoch": 0.4097902097902098, + "grad_norm": 0.13036265265723118, + "kl": 0.4453125, + "learning_rate": 9.999817519649158e-05, + "loss": 0.0127, + "num_tokens": 2994773.0, + "reward": 2.552300214767456, + "reward_std": 0.14500018954277039, + "rewards/check_gptzero_func": 0.4285714328289032, + "rewards/check_perplexity_diff_func": 1.5, + "rewards/check_winston_local_func": 0.6237286329269409, + "step": 293 + }, + { + "clip_ratio": 0.0027297178748995066, + "epoch": 0.4111888111888112, + "grad_norm": 0.0976002240930887, + "kl": 0.439453125, + "learning_rate": 9.99976165876107e-05, + "loss": 0.0101, + "step": 294 + }, + { + "clip_ratio": 0.0, + "completion_length": 113.35714721679688, + "epoch": 0.4125874125874126, + "grad_norm": 0.02609480806956132, + "kl": 0.1591796875, + "learning_rate": 9.999698350006063e-05, + "loss": 0.0012, + "num_tokens": 3015993.0, + "reward": 2.874926805496216, + "reward_std": 0.0874238833785057, + "rewards/check_gptzero_func": 0.4821428656578064, + "rewards/check_perplexity_diff_func": 1.8928571939468384, + "rewards/check_winston_local_func": 0.4999266266822815, + "step": 295 + }, + { + "clip_ratio": 9.294544724980369e-05, + "epoch": 0.413986013986014, + "grad_norm": 0.02050573548257188, + "kl": 0.16015625, + "learning_rate": 9.999627593478442e-05, + "loss": 0.0011, + "step": 296 + }, + { + "clip_ratio": 0.00012664640962611884, + "completion_length": 98.5714340209961, + "epoch": 0.4153846153846154, + "grad_norm": 0.01680753751461254, + "kl": 0.201171875, + "learning_rate": 9.999549389283606e-05, + "loss": -0.0003, + "num_tokens": 3035155.0, + "reward": 3.0224545001983643, + "reward_std": 0.016296973451972008, + "rewards/check_gptzero_func": 0.4107142984867096, + "rewards/check_perplexity_diff_func": 1.9642857313156128, + "rewards/check_winston_local_func": 0.647454559803009, + "step": 297 + }, + { + "clip_ratio": 0.0, + "epoch": 0.4167832167832168, + "grad_norm": 0.012884367061070623, + "kl": 0.203125, + "learning_rate": 9.999463737538053e-05, + "loss": -0.0003, + "step": 298 + }, + { + "clip_ratio": 0.0003541912592481822, + "completion_length": 95.76786041259766, + "epoch": 0.41818181818181815, + "grad_norm": 0.04526732622569909, + "kl": 0.291015625, + "learning_rate": 9.999370638369377e-05, + "loss": 0.0002, + "num_tokens": 3054174.0, + "reward": 2.3303959369659424, + "reward_std": 0.08971218019723892, + "rewards/check_gptzero_func": 0.3035714328289032, + "rewards/check_perplexity_diff_func": 1.6071428060531616, + "rewards/check_winston_local_func": 0.4196813404560089, + "step": 299 + }, + { + "clip_ratio": 0.00011806376278400421, + "epoch": 0.4195804195804196, + "grad_norm": 0.03862986955837794, + "kl": 0.296875, + "learning_rate": 9.999270091916257e-05, + "loss": -0.0002, + "step": 300 + }, + { + "clip_ratio": 0.0012040403671562672, + "completion_length": 101.73214721679688, + "epoch": 0.42097902097902096, + "grad_norm": 0.12616717395678975, + "kl": 0.35546875, + "learning_rate": 9.999162098328474e-05, + "loss": 0.0033, + "num_tokens": 3073401.0, + "reward": 2.8153316974639893, + "reward_std": 0.18918544054031372, + "rewards/check_gptzero_func": 0.4285714328289032, + "rewards/check_perplexity_diff_func": 1.7142857313156128, + "rewards/check_winston_local_func": 0.6724745631217957, + "step": 301 + }, + { + "clip_ratio": 0.0013847867958247662, + "epoch": 0.4223776223776224, + "grad_norm": 0.06929991057149441, + "kl": 0.412109375, + "learning_rate": 9.999046657766903e-05, + "loss": 0.0012, + "step": 302 + }, + { + "clip_ratio": 0.0, + "completion_length": 81.3214340209961, + "epoch": 0.42377622377622376, + "grad_norm": 0.06730614630696258, + "kl": 0.2099609375, + "learning_rate": 9.998923770403505e-05, + "loss": -0.0019, + "num_tokens": 3090625.0, + "reward": 2.802220106124878, + "reward_std": 0.03781326860189438, + "rewards/check_gptzero_func": 0.375, + "rewards/check_perplexity_diff_func": 2.0, + "rewards/check_winston_local_func": 0.42722010612487793, + "step": 303 + }, + { + "clip_ratio": 0.0, + "epoch": 0.4251748251748252, + "grad_norm": 0.022175315902565728, + "kl": 0.2109375, + "learning_rate": 9.99879343642134e-05, + "loss": -0.0024, + "step": 304 + }, + { + "clip_ratio": 0.00013683634460903704, + "completion_length": 105.53572082519531, + "epoch": 0.42657342657342656, + "grad_norm": 0.04867704636152166, + "kl": 0.173828125, + "learning_rate": 9.998655656014561e-05, + "loss": 0.0117, + "num_tokens": 3110877.0, + "reward": 2.6083197593688965, + "reward_std": 0.13502858579158783, + "rewards/check_gptzero_func": 0.3571428656578064, + "rewards/check_perplexity_diff_func": 1.8928571939468384, + "rewards/check_winston_local_func": 0.35831958055496216, + "step": 305 + }, + { + "clip_ratio": 0.0, + "epoch": 0.427972027972028, + "grad_norm": 0.043987709695075865, + "kl": 0.1728515625, + "learning_rate": 9.99851042938841e-05, + "loss": 0.0113, + "step": 306 + }, + { + "clip_ratio": 0.0004155792994424701, + "completion_length": 124.03572082519531, + "epoch": 0.42937062937062936, + "grad_norm": 0.03684760048760383, + "kl": 0.2080078125, + "learning_rate": 9.998357756759222e-05, + "loss": -0.0258, + "num_tokens": 3134097.0, + "reward": 2.7439539432525635, + "reward_std": 0.11075940728187561, + "rewards/check_gptzero_func": 0.4642857015132904, + "rewards/check_perplexity_diff_func": 1.7857142686843872, + "rewards/check_winston_local_func": 0.4939536452293396, + "step": 307 + }, + { + "clip_ratio": 0.0, + "epoch": 0.4307692307692308, + "grad_norm": 0.013648247572273129, + "kl": 0.20703125, + "learning_rate": 9.998197638354428e-05, + "loss": -0.026, + "step": 308 + }, + { + "clip_ratio": 0.0, + "completion_length": 89.12500762939453, + "epoch": 0.43216783216783217, + "grad_norm": 0.02465210474674629, + "kl": 0.443359375, + "learning_rate": 9.998030074412545e-05, + "loss": 0.0014, + "num_tokens": 3152358.0, + "reward": 2.420307159423828, + "reward_std": 0.11704181134700775, + "rewards/check_gptzero_func": 0.2321428507566452, + "rewards/check_perplexity_diff_func": 1.75, + "rewards/check_winston_local_func": 0.4381641745567322, + "step": 309 + }, + { + "clip_ratio": 0.0, + "epoch": 0.43356643356643354, + "grad_norm": 0.021184461362474653, + "kl": 0.48828125, + "learning_rate": 9.997855065183184e-05, + "loss": 0.0014, + "step": 310 + }, + { + "clip_ratio": 0.0, + "completion_length": 115.73214721679688, + "epoch": 0.43496503496503497, + "grad_norm": 0.0049134080725686995, + "kl": 0.162109375, + "learning_rate": 9.997672610927047e-05, + "loss": -0.0047, + "num_tokens": 3173997.0, + "reward": 2.61786150932312, + "reward_std": 0.013112460263073444, + "rewards/check_gptzero_func": 0.375, + "rewards/check_perplexity_diff_func": 1.8214285373687744, + "rewards/check_winston_local_func": 0.4214327931404114, + "step": 311 + }, + { + "clip_ratio": 0.0, + "epoch": 0.43636363636363634, + "grad_norm": 0.00595537475288129, + "kl": 0.173828125, + "learning_rate": 9.997482711915927e-05, + "loss": -0.0047, + "step": 312 + }, + { + "clip_ratio": 0.0001360544265480712, + "completion_length": 90.42857360839844, + "epoch": 0.43776223776223777, + "grad_norm": 0.007409687643468917, + "kl": 0.2275390625, + "learning_rate": 9.997285368432703e-05, + "loss": 0.0006, + "num_tokens": 3192417.0, + "reward": 2.682650566101074, + "reward_std": 0.02417343109846115, + "rewards/check_gptzero_func": 0.375, + "rewards/check_perplexity_diff_func": 1.8214285373687744, + "rewards/check_winston_local_func": 0.4862218499183655, + "step": 313 + }, + { + "clip_ratio": 0.0, + "epoch": 0.43916083916083914, + "grad_norm": 0.007129921917131849, + "kl": 0.2275390625, + "learning_rate": 9.997080580771349e-05, + "loss": 0.0006, + "step": 314 + }, + { + "clip_ratio": 0.0, + "completion_length": 83.14286041259766, + "epoch": 0.4405594405594406, + "grad_norm": 1.3606081670938062e-05, + "kl": 0.2255859375, + "learning_rate": 9.996868349236927e-05, + "loss": 0.0002, + "num_tokens": 3209875.0, + "reward": 2.671295404434204, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.3571428656578064, + "rewards/check_perplexity_diff_func": 1.8571428060531616, + "rewards/check_winston_local_func": 0.45700937509536743, + "step": 315 + }, + { + "clip_ratio": 0.0, + "epoch": 0.44195804195804195, + "grad_norm": 1.3537787639105168e-05, + "kl": 0.2255859375, + "learning_rate": 9.996648674145583e-05, + "loss": 0.0002, + "step": 316 + }, + { + "clip_ratio": 0.0, + "completion_length": 126.21429443359375, + "epoch": 0.4433566433566434, + "grad_norm": 0.00991208286904361, + "kl": 0.15625, + "learning_rate": 9.996421555824558e-05, + "loss": 0.0103, + "num_tokens": 3232963.0, + "reward": 2.6654393672943115, + "reward_std": 0.0725974440574646, + "rewards/check_gptzero_func": 0.5357142686843872, + "rewards/check_perplexity_diff_func": 1.6071428060531616, + "rewards/check_winston_local_func": 0.5225821733474731, + "step": 317 + }, + { + "clip_ratio": 0.0, + "epoch": 0.44475524475524475, + "grad_norm": 0.010677817125212348, + "kl": 0.154296875, + "learning_rate": 9.996186994612176e-05, + "loss": 0.0102, + "step": 318 + }, + { + "clip_ratio": 0.0, + "completion_length": 103.4464340209961, + "epoch": 0.4461538461538462, + "grad_norm": 5.275581365835224e-05, + "kl": 0.16796875, + "learning_rate": 9.995944990857849e-05, + "loss": 0.0002, + "num_tokens": 3252898.0, + "reward": 3.0420854091644287, + "reward_std": 0.00039901022682897747, + "rewards/check_gptzero_func": 0.5714285969734192, + "rewards/check_perplexity_diff_func": 1.9285714626312256, + "rewards/check_winston_local_func": 0.5420854687690735, + "step": 319 + }, + { + "clip_ratio": 0.0, + "epoch": 0.44755244755244755, + "grad_norm": 6.10762506510971e-05, + "kl": 0.16796875, + "learning_rate": 9.995695544922077e-05, + "loss": 0.0002, + "step": 320 + }, + { + "clip_ratio": 0.0, + "completion_length": 102.28572082519531, + "epoch": 0.4489510489510489, + "grad_norm": 0.010696195320132886, + "kl": 0.1669921875, + "learning_rate": 9.995438657176448e-05, + "loss": -0.0177, + "num_tokens": 3272574.0, + "reward": 2.7592146396636963, + "reward_std": 0.03917063772678375, + "rewards/check_gptzero_func": 0.3571428656578064, + "rewards/check_perplexity_diff_func": 1.9642857313156128, + "rewards/check_winston_local_func": 0.437785804271698, + "step": 321 + }, + { + "clip_ratio": 0.0, + "epoch": 0.45034965034965035, + "grad_norm": 0.01083841197384593, + "kl": 0.1669921875, + "learning_rate": 9.995174328003631e-05, + "loss": -0.0177, + "step": 322 + }, + { + "clip_ratio": 8.473140769638121e-05, + "completion_length": 117.41072082519531, + "epoch": 0.45174825174825173, + "grad_norm": 0.004522641361347185, + "kl": 0.1640625, + "learning_rate": 9.994902557797382e-05, + "loss": -0.0002, + "num_tokens": 3294025.0, + "reward": 2.92221736907959, + "reward_std": 0.025547320023179054, + "rewards/check_gptzero_func": 0.5535714030265808, + "rewards/check_perplexity_diff_func": 2.0, + "rewards/check_winston_local_func": 0.36864587664604187, + "step": 323 + }, + { + "clip_ratio": 0.0, + "epoch": 0.45314685314685316, + "grad_norm": 0.004737684573095642, + "kl": 0.1650390625, + "learning_rate": 9.994623346962544e-05, + "loss": -0.0002, + "step": 324 + }, + { + "clip_ratio": 0.0, + "completion_length": 95.26786041259766, + "epoch": 0.45454545454545453, + "grad_norm": 3.372088632491828e-06, + "kl": 0.1796875, + "learning_rate": 9.99433669591504e-05, + "loss": 0.0002, + "num_tokens": 3312902.0, + "reward": 2.595499038696289, + "reward_std": 4.388691013446078e-05, + "rewards/check_gptzero_func": 0.25, + "rewards/check_perplexity_diff_func": 2.0, + "rewards/check_winston_local_func": 0.34549909830093384, + "step": 325 + }, + { + "clip_ratio": 0.0, + "epoch": 0.45594405594405596, + "grad_norm": 3.349405120445447e-06, + "kl": 0.1796875, + "learning_rate": 9.994042605081879e-05, + "loss": 0.0002, + "step": 326 + }, + { + "clip_ratio": 0.0, + "completion_length": 97.80357360839844, + "epoch": 0.45734265734265733, + "grad_norm": 0.010175912326793437, + "kl": 0.177734375, + "learning_rate": 9.993741074901153e-05, + "loss": 0.0003, + "num_tokens": 3331815.0, + "reward": 2.496483325958252, + "reward_std": 0.04960284009575844, + "rewards/check_gptzero_func": 0.2142857164144516, + "rewards/check_perplexity_diff_func": 1.9642857313156128, + "rewards/check_winston_local_func": 0.31791168451309204, + "step": 327 + }, + { + "clip_ratio": 0.0, + "epoch": 0.45874125874125876, + "grad_norm": 0.01017493586702021, + "kl": 0.177734375, + "learning_rate": 9.993432105822034e-05, + "loss": 0.0003, + "step": 328 + }, + { + "clip_ratio": 0.0, + "completion_length": 96.30357360839844, + "epoch": 0.46013986013986014, + "grad_norm": 0.0165654292436505, + "kl": 0.1572265625, + "learning_rate": 9.993115698304774e-05, + "loss": 0.0003, + "num_tokens": 3351210.0, + "reward": 2.8074374198913574, + "reward_std": 0.07505974173545837, + "rewards/check_gptzero_func": 0.5178571343421936, + "rewards/check_perplexity_diff_func": 1.8571428060531616, + "rewards/check_winston_local_func": 0.4324372708797455, + "step": 329 + }, + { + "clip_ratio": 0.0, + "epoch": 0.46153846153846156, + "grad_norm": 0.014211382809754432, + "kl": 0.16015625, + "learning_rate": 9.992791852820709e-05, + "loss": 0.0003, + "step": 330 + }, + { + "clip_ratio": 0.0, + "completion_length": 86.1964340209961, + "epoch": 0.46293706293706294, + "grad_norm": 0.0580909978469744, + "kl": 0.2119140625, + "learning_rate": 9.992460569852256e-05, + "loss": 0.0013, + "num_tokens": 3369025.0, + "reward": 2.986760377883911, + "reward_std": 0.10425443947315216, + "rewards/check_gptzero_func": 0.4642857015132904, + "rewards/check_perplexity_diff_func": 1.8571428060531616, + "rewards/check_winston_local_func": 0.6653318405151367, + "step": 331 + }, + { + "clip_ratio": 0.0, + "epoch": 0.4643356643356643, + "grad_norm": 0.0012843504186470114, + "kl": 0.2109375, + "learning_rate": 9.992121849892904e-05, + "loss": 0.0012, + "step": 332 + }, + { + "clip_ratio": 0.0, + "completion_length": 113.42857360839844, + "epoch": 0.46573426573426574, + "grad_norm": 0.03054414776989089, + "kl": 0.1494140625, + "learning_rate": 9.99177569344723e-05, + "loss": -0.0, + "num_tokens": 3390093.0, + "reward": 2.5890934467315674, + "reward_std": 0.1023683100938797, + "rewards/check_gptzero_func": 0.25, + "rewards/check_perplexity_diff_func": 1.9285714626312256, + "rewards/check_winston_local_func": 0.41052165627479553, + "step": 333 + }, + { + "clip_ratio": 0.00022784233442507684, + "epoch": 0.4671328671328671, + "grad_norm": 0.006141110067038759, + "kl": 0.150390625, + "learning_rate": 9.99142210103088e-05, + "loss": -0.0001, + "step": 334 + }, + { + "clip_ratio": 0.0, + "completion_length": 109.00000762939453, + "epoch": 0.46853146853146854, + "grad_norm": 0.011702000814979486, + "kl": 0.158203125, + "learning_rate": 9.991061073170585e-05, + "loss": 0.0001, + "num_tokens": 3410511.0, + "reward": 2.601181745529175, + "reward_std": 0.04859553650021553, + "rewards/check_gptzero_func": 0.2857142984867096, + "rewards/check_perplexity_diff_func": 1.8928571939468384, + "rewards/check_winston_local_func": 0.42261001467704773, + "step": 335 + }, + { + "clip_ratio": 0.0, + "epoch": 0.4699300699300699, + "grad_norm": 0.004380253375033637, + "kl": 0.158203125, + "learning_rate": 9.990692610404145e-05, + "loss": 0.0001, + "step": 336 + }, + { + "clip_ratio": 0.00015908369095996022, + "completion_length": 124.37500762939453, + "epoch": 0.47132867132867134, + "grad_norm": 0.011635533606345248, + "kl": 0.1318359375, + "learning_rate": 9.99031671328044e-05, + "loss": 0.0002, + "num_tokens": 3433250.0, + "reward": 2.719287395477295, + "reward_std": 0.05448899790644646, + "rewards/check_gptzero_func": 0.5, + "rewards/check_perplexity_diff_func": 1.75, + "rewards/check_winston_local_func": 0.4692873954772949, + "step": 337 + }, + { + "clip_ratio": 0.0, + "epoch": 0.4727272727272727, + "grad_norm": 0.009393665715625162, + "kl": 0.1318359375, + "learning_rate": 9.989933382359422e-05, + "loss": 0.0001, + "step": 338 + }, + { + "clip_ratio": 0.0, + "completion_length": 110.8214340209961, + "epoch": 0.47412587412587415, + "grad_norm": 6.954019260271828e-07, + "kl": 0.15625, + "learning_rate": 9.98954261821212e-05, + "loss": 0.0002, + "num_tokens": 3454202.0, + "reward": 2.5842461585998535, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.3928571343421936, + "rewards/check_perplexity_diff_func": 1.7857142686843872, + "rewards/check_winston_local_func": 0.40567442774772644, + "step": 339 + }, + { + "clip_ratio": 0.0, + "epoch": 0.4755244755244755, + "grad_norm": 6.417345751201852e-07, + "kl": 0.15625, + "learning_rate": 9.98914442142063e-05, + "loss": 0.0002, + "step": 340 + }, + { + "clip_ratio": 0.0, + "completion_length": 76.55357360839844, + "epoch": 0.47692307692307695, + "grad_norm": 0.008322229771779063, + "kl": 0.263671875, + "learning_rate": 9.988738792578126e-05, + "loss": 0.0005, + "num_tokens": 3471041.0, + "reward": 2.916919231414795, + "reward_std": 0.05185602605342865, + "rewards/check_gptzero_func": 0.3214285671710968, + "rewards/check_perplexity_diff_func": 1.9642857313156128, + "rewards/check_winston_local_func": 0.631205141544342, + "step": 341 + }, + { + "clip_ratio": 0.0, + "epoch": 0.4783216783216783, + "grad_norm": 0.010116392524351476, + "kl": 0.263671875, + "learning_rate": 9.988325732288851e-05, + "loss": 0.0004, + "step": 342 + }, + { + "clip_ratio": 0.0, + "completion_length": 122.21429443359375, + "epoch": 0.4797202797202797, + "grad_norm": 1.4137707959903356e-05, + "kl": 0.134765625, + "learning_rate": 9.987905241168117e-05, + "loss": 0.0001, + "num_tokens": 3493483.0, + "reward": 2.7157416343688965, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.3928571343421936, + "rewards/check_perplexity_diff_func": 1.8571428060531616, + "rewards/check_winston_local_func": 0.46574172377586365, + "step": 343 + }, + { + "clip_ratio": 0.0, + "epoch": 0.4811188811188811, + "grad_norm": 1.3729789968238485e-05, + "kl": 0.134765625, + "learning_rate": 9.987477319842307e-05, + "loss": 0.0001, + "step": 344 + }, + { + "clip_ratio": 0.0, + "completion_length": 93.53572082519531, + "epoch": 0.4825174825174825, + "grad_norm": 7.468121696306164e-07, + "kl": 0.1845703125, + "learning_rate": 9.987041968948869e-05, + "loss": 0.0002, + "num_tokens": 3512009.0, + "reward": 2.5505218505859375, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.1785714328289032, + "rewards/check_perplexity_diff_func": 1.9285714626312256, + "rewards/check_winston_local_func": 0.44337886571884155, + "step": 345 + }, + { + "clip_ratio": 0.0, + "epoch": 0.48391608391608393, + "grad_norm": 7.137709602226676e-07, + "kl": 0.1845703125, + "learning_rate": 9.986599189136325e-05, + "loss": 0.0002, + "step": 346 + }, + { + "clip_ratio": 0.00031959093757905066, + "completion_length": 102.58928680419922, + "epoch": 0.4853146853146853, + "grad_norm": 0.04707770778424934, + "kl": 0.1982421875, + "learning_rate": 9.986148981064258e-05, + "loss": 0.0004, + "num_tokens": 3531574.0, + "reward": 2.7341363430023193, + "reward_std": 0.0231277234852314, + "rewards/check_gptzero_func": 0.4107142984867096, + "rewards/check_perplexity_diff_func": 1.8928571939468384, + "rewards/check_winston_local_func": 0.430564820766449, + "step": 347 + }, + { + "clip_ratio": 0.0, + "epoch": 0.48671328671328673, + "grad_norm": 0.020919799034125892, + "kl": 0.2060546875, + "learning_rate": 9.985691345403316e-05, + "loss": 0.0002, + "step": 348 + }, + { + "clip_ratio": 0.0, + "completion_length": 127.67857360839844, + "epoch": 0.4881118881118881, + "grad_norm": 6.517786151476589e-05, + "kl": 0.1328125, + "learning_rate": 9.985226282835216e-05, + "loss": 0.0001, + "num_tokens": 3554440.0, + "reward": 2.926541566848755, + "reward_std": 0.0005768488626927137, + "rewards/check_gptzero_func": 0.5357142686843872, + "rewards/check_perplexity_diff_func": 1.8571428060531616, + "rewards/check_winston_local_func": 0.5336844325065613, + "step": 349 + }, + { + "clip_ratio": 0.0, + "epoch": 0.48951048951048953, + "grad_norm": 6.370254129427888e-05, + "kl": 0.1328125, + "learning_rate": 9.984753794052735e-05, + "loss": 0.0001, + "step": 350 + }, + { + "clip_ratio": 0.0, + "completion_length": 98.75000762939453, + "epoch": 0.4909090909090909, + "grad_norm": 2.085540464425244e-05, + "kl": 0.205078125, + "learning_rate": 9.984273879759713e-05, + "loss": 0.0002, + "num_tokens": 3573976.0, + "reward": 2.8204286098480225, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.3928571343421936, + "rewards/check_perplexity_diff_func": 1.9285714626312256, + "rewards/check_winston_local_func": 0.4989997446537018, + "step": 351 + }, + { + "clip_ratio": 0.0, + "epoch": 0.49230769230769234, + "grad_norm": 2.009578048518353e-05, + "kl": 0.205078125, + "learning_rate": 9.983786540671051e-05, + "loss": 0.0002, + "step": 352 + }, + { + "clip_ratio": 0.0, + "completion_length": 85.21428680419922, + "epoch": 0.4937062937062937, + "grad_norm": 2.1499404490472988e-06, + "kl": 0.1826171875, + "learning_rate": 9.983291777512711e-05, + "loss": 0.0002, + "num_tokens": 3591644.0, + "reward": 2.8764142990112305, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.4285714328289032, + "rewards/check_perplexity_diff_func": 2.0, + "rewards/check_winston_local_func": 0.44784265756607056, + "step": 353 + }, + { + "clip_ratio": 0.0, + "epoch": 0.4951048951048951, + "grad_norm": 2.160124410750654e-06, + "kl": 0.1826171875, + "learning_rate": 9.982789591021715e-05, + "loss": 0.0002, + "step": 354 + }, + { + "clip_ratio": 0.0, + "completion_length": 112.00000762939453, + "epoch": 0.4965034965034965, + "grad_norm": 0.00014814082463704463, + "kl": 0.1669921875, + "learning_rate": 9.982279981946143e-05, + "loss": 0.0002, + "num_tokens": 3612596.0, + "reward": 2.8398191928863525, + "reward_std": 0.0006443771999329329, + "rewards/check_gptzero_func": 0.3928571343421936, + "rewards/check_perplexity_diff_func": 1.9285714626312256, + "rewards/check_winston_local_func": 0.5183902382850647, + "step": 355 + }, + { + "clip_ratio": 0.0, + "epoch": 0.4979020979020979, + "grad_norm": 0.00014606763289714295, + "kl": 0.1669921875, + "learning_rate": 9.98176295104513e-05, + "loss": 0.0002, + "step": 356 + }, + { + "clip_ratio": 0.0, + "completion_length": 88.50000762939453, + "epoch": 0.4993006993006993, + "grad_norm": 3.33550016470454e-07, + "kl": 0.1669921875, + "learning_rate": 9.98123849908887e-05, + "loss": 0.0002, + "num_tokens": 3630862.0, + "reward": 2.655492067337036, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.3214285671710968, + "rewards/check_perplexity_diff_func": 1.8571428060531616, + "rewards/check_winston_local_func": 0.47692039608955383, + "step": 357 + }, + { + "clip_ratio": 0.0, + "epoch": 0.5006993006993007, + "grad_norm": 3.3462192763241657e-07, + "kl": 0.1669921875, + "learning_rate": 9.980706626858607e-05, + "loss": 0.0002, + "step": 358 + }, + { + "clip_ratio": 0.0, + "completion_length": 126.39286041259766, + "epoch": 0.5020979020979021, + "grad_norm": 3.4860786526471938e-06, + "kl": 0.1416015625, + "learning_rate": 9.980167335146642e-05, + "loss": 0.0001, + "num_tokens": 3653738.0, + "reward": 2.765826463699341, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.5, + "rewards/check_perplexity_diff_func": 1.8571428060531616, + "rewards/check_winston_local_func": 0.4086834490299225, + "step": 359 + }, + { + "clip_ratio": 0.0, + "epoch": 0.5034965034965035, + "grad_norm": 3.71619859471226e-06, + "kl": 0.1416015625, + "learning_rate": 9.979620624756329e-05, + "loss": 0.0001, + "step": 360 + }, + { + "clip_ratio": 0.0, + "completion_length": 129.92857360839844, + "epoch": 0.5048951048951049, + "grad_norm": 8.195642286342138e-07, + "kl": 0.1376953125, + "learning_rate": 9.979066496502074e-05, + "loss": 0.0001, + "num_tokens": 3677274.0, + "reward": 2.8696835041046143, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.5357142686843872, + "rewards/check_perplexity_diff_func": 1.9285714626312256, + "rewards/check_winston_local_func": 0.4053979218006134, + "step": 361 + }, + { + "clip_ratio": 0.0, + "epoch": 0.5062937062937063, + "grad_norm": 8.760444459307852e-07, + "kl": 0.1376953125, + "learning_rate": 9.978504951209327e-05, + "loss": 0.0001, + "step": 362 + }, + { + "clip_ratio": 0.0, + "completion_length": 118.0714340209961, + "epoch": 0.5076923076923077, + "grad_norm": 1.595860947799347e-06, + "kl": 0.162109375, + "learning_rate": 9.977935989714595e-05, + "loss": 0.0002, + "num_tokens": 3698896.0, + "reward": 2.6664552688598633, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.2857142984867096, + "rewards/check_perplexity_diff_func": 1.9285714626312256, + "rewards/check_winston_local_func": 0.45216941833496094, + "step": 363 + }, + { + "clip_ratio": 0.0, + "epoch": 0.509090909090909, + "grad_norm": 1.6260194368178069e-06, + "kl": 0.162109375, + "learning_rate": 9.977359612865423e-05, + "loss": 0.0002, + "step": 364 + }, + { + "clip_ratio": 0.0, + "completion_length": 85.96428680419922, + "epoch": 0.5104895104895105, + "grad_norm": 4.357301744161908e-06, + "kl": 0.2138671875, + "learning_rate": 9.976775821520412e-05, + "loss": 0.0002, + "num_tokens": 3716536.0, + "reward": 2.8877956867218018, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.3928571343421936, + "rewards/check_perplexity_diff_func": 1.9285714626312256, + "rewards/check_winston_local_func": 0.566366970539093, + "step": 365 + }, + { + "clip_ratio": 0.0, + "epoch": 0.5118881118881119, + "grad_norm": 4.353902612581284e-06, + "kl": 0.2138671875, + "learning_rate": 9.976184616549203e-05, + "loss": 0.0002, + "step": 366 + }, + { + "clip_ratio": 0.0, + "completion_length": 116.96429443359375, + "epoch": 0.5132867132867133, + "grad_norm": 1.5148419096170459e-06, + "kl": 0.146484375, + "learning_rate": 9.97558599883248e-05, + "loss": 0.0001, + "num_tokens": 3737950.0, + "reward": 2.687312602996826, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.3928571343421936, + "rewards/check_perplexity_diff_func": 1.9285714626312256, + "rewards/check_winston_local_func": 0.36588403582572937, + "step": 367 + }, + { + "clip_ratio": 0.0, + "epoch": 0.5146853146853146, + "grad_norm": 1.4470246412763422e-06, + "kl": 0.146484375, + "learning_rate": 9.97497996926197e-05, + "loss": 0.0001, + "step": 368 + }, + { + "clip_ratio": 0.0, + "completion_length": 94.83928680419922, + "epoch": 0.5160839160839161, + "grad_norm": 0.0004647591690002515, + "kl": 0.2138671875, + "learning_rate": 9.974366528740441e-05, + "loss": 0.0002, + "num_tokens": 3756935.0, + "reward": 2.634523391723633, + "reward_std": 0.0005499019753187895, + "rewards/check_gptzero_func": 0.2857142984867096, + "rewards/check_perplexity_diff_func": 1.9285714626312256, + "rewards/check_winston_local_func": 0.4202377200126648, + "step": 369 + }, + { + "clip_ratio": 0.0, + "epoch": 0.5174825174825175, + "grad_norm": 0.0004885464229234562, + "kl": 0.2138671875, + "learning_rate": 9.973745678181705e-05, + "loss": 0.0002, + "step": 370 + }, + { + "clip_ratio": 0.0, + "completion_length": 100.17857360839844, + "epoch": 0.5188811188811189, + "grad_norm": 3.0418176986251483e-06, + "kl": 0.1640625, + "learning_rate": 9.973117418510605e-05, + "loss": 0.0002, + "num_tokens": 3776087.0, + "reward": 2.5216457843780518, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.2857142984867096, + "rewards/check_perplexity_diff_func": 1.8571428060531616, + "rewards/check_winston_local_func": 0.3787885308265686, + "step": 371 + }, + { + "clip_ratio": 0.0, + "epoch": 0.5202797202797202, + "grad_norm": 3.0723635480561103e-06, + "kl": 0.1640625, + "learning_rate": 9.972481750663026e-05, + "loss": 0.0002, + "step": 372 + }, + { + "clip_ratio": 0.0, + "completion_length": 89.66072082519531, + "epoch": 0.5216783216783217, + "grad_norm": 0.008590656277289149, + "kl": 0.40234375, + "learning_rate": 9.971838675585888e-05, + "loss": 0.0007, + "num_tokens": 3794550.0, + "reward": 2.5267839431762695, + "reward_std": 0.05061079189181328, + "rewards/check_gptzero_func": 0.3214285671710968, + "rewards/check_perplexity_diff_func": 1.8214285373687744, + "rewards/check_winston_local_func": 0.383926659822464, + "step": 373 + }, + { + "clip_ratio": 0.0, + "epoch": 0.5230769230769231, + "grad_norm": 0.003770576536622864, + "kl": 0.4296875, + "learning_rate": 9.97118819423714e-05, + "loss": 0.0007, + "step": 374 + }, + { + "clip_ratio": 0.0, + "completion_length": 80.17857360839844, + "epoch": 0.5244755244755245, + "grad_norm": 3.560813902520289e-07, + "kl": 0.2216796875, + "learning_rate": 9.970530307585773e-05, + "loss": 0.0002, + "num_tokens": 3811756.0, + "reward": 2.883481740951538, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.4285714328289032, + "rewards/check_perplexity_diff_func": 1.9285714626312256, + "rewards/check_winston_local_func": 0.5263389348983765, + "step": 375 + }, + { + "clip_ratio": 0.0, + "epoch": 0.5258741258741259, + "grad_norm": 3.5463275242856317e-07, + "kl": 0.2216796875, + "learning_rate": 9.9698650166118e-05, + "loss": 0.0002, + "step": 376 + }, + { + "clip_ratio": 0.0, + "completion_length": 86.75000762939453, + "epoch": 0.5272727272727272, + "grad_norm": 4.776682547251814e-07, + "kl": 0.205078125, + "learning_rate": 9.969192322306271e-05, + "loss": 0.0002, + "num_tokens": 3829616.0, + "reward": 2.9529268741607666, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.4642857015132904, + "rewards/check_perplexity_diff_func": 1.9285714626312256, + "rewards/check_winston_local_func": 0.5600695610046387, + "step": 377 + }, + { + "clip_ratio": 0.0, + "epoch": 0.5286713286713287, + "grad_norm": 4.804658823710114e-07, + "kl": 0.205078125, + "learning_rate": 9.96851222567126e-05, + "loss": 0.0002, + "step": 378 + }, + { + "clip_ratio": 0.0, + "completion_length": 117.89286041259766, + "epoch": 0.5300699300699301, + "grad_norm": 9.143102926453055e-06, + "kl": 0.158203125, + "learning_rate": 9.96782472771987e-05, + "loss": 0.0002, + "num_tokens": 3851170.0, + "reward": 2.7617037296295166, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.4285714328289032, + "rewards/check_perplexity_diff_func": 1.9285714626312256, + "rewards/check_winston_local_func": 0.40456071496009827, + "step": 379 + }, + { + "clip_ratio": 0.0, + "epoch": 0.5314685314685315, + "grad_norm": 9.41817239216292e-06, + "kl": 0.158203125, + "learning_rate": 9.967129829476228e-05, + "loss": 0.0002, + "step": 380 + }, + { + "clip_ratio": 0.000286861730273813, + "completion_length": 80.8214340209961, + "epoch": 0.5328671328671328, + "grad_norm": 0.03337399461775748, + "kl": 0.224609375, + "learning_rate": 9.966427531975483e-05, + "loss": -0.0402, + "num_tokens": 3868508.0, + "reward": 2.820319652557373, + "reward_std": 0.0946396142244339, + "rewards/check_gptzero_func": 0.4285714328289032, + "rewards/check_perplexity_diff_func": 1.8571428060531616, + "rewards/check_winston_local_func": 0.5346053838729858, + "step": 381 + }, + { + "clip_ratio": 0.0, + "epoch": 0.5342657342657343, + "grad_norm": 0.00034351439093039966, + "kl": 0.224609375, + "learning_rate": 9.965717836263812e-05, + "loss": -0.0403, + "step": 382 + }, + { + "clip_ratio": 0.0, + "completion_length": 95.03572082519531, + "epoch": 0.5356643356643357, + "grad_norm": 3.3945579906577137e-07, + "kl": 0.216796875, + "learning_rate": 9.965000743398408e-05, + "loss": 0.0002, + "num_tokens": 3887044.0, + "reward": 2.8529021739959717, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.4285714328289032, + "rewards/check_perplexity_diff_func": 2.0, + "rewards/check_winston_local_func": 0.4243304431438446, + "step": 383 + }, + { + "clip_ratio": 0.0, + "epoch": 0.5370629370629371, + "grad_norm": 2.8390305863820515e-07, + "kl": 0.216796875, + "learning_rate": 9.964276254447484e-05, + "loss": 0.0002, + "step": 384 + }, + { + "clip_ratio": 0.0, + "completion_length": 140.1428680419922, + "epoch": 0.5384615384615384, + "grad_norm": 8.521451122880038e-07, + "kl": 0.134765625, + "learning_rate": 9.96354437049027e-05, + "loss": 0.0001, + "num_tokens": 3911264.0, + "reward": 2.4770619869232178, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.2857142984867096, + "rewards/check_perplexity_diff_func": 1.8571428060531616, + "rewards/check_winston_local_func": 0.33420488238334656, + "step": 385 + }, + { + "clip_ratio": 0.0, + "epoch": 0.5398601398601398, + "grad_norm": 6.753306754213034e-07, + "kl": 0.134765625, + "learning_rate": 9.962805092617016e-05, + "loss": 0.0001, + "step": 386 + }, + { + "clip_ratio": 0.0, + "completion_length": 97.64286041259766, + "epoch": 0.5412587412587413, + "grad_norm": 0.08784199050747857, + "kl": 2.28125, + "learning_rate": 9.962058421928979e-05, + "loss": 0.0023, + "num_tokens": 3930478.0, + "reward": 2.897810459136963, + "reward_std": 0.002231778111308813, + "rewards/check_gptzero_func": 0.4285714328289032, + "rewards/check_perplexity_diff_func": 2.0, + "rewards/check_winston_local_func": 0.4692386984825134, + "step": 387 + }, + { + "clip_ratio": 0.0, + "epoch": 0.5426573426573427, + "grad_norm": 0.03620083942117924, + "kl": 0.98828125, + "learning_rate": 9.961304359538437e-05, + "loss": 0.001, + "step": 388 + }, + { + "clip_ratio": 0.0, + "completion_length": 102.21428680419922, + "epoch": 0.544055944055944, + "grad_norm": 0.025447754609912356, + "kl": 0.1806640625, + "learning_rate": 9.96054290656867e-05, + "loss": 0.0002, + "num_tokens": 3950362.0, + "reward": 2.8110532760620117, + "reward_std": 0.04895726963877678, + "rewards/check_gptzero_func": 0.5357142686843872, + "rewards/check_perplexity_diff_func": 1.8214285373687744, + "rewards/check_winston_local_func": 0.45391014218330383, + "step": 389 + }, + { + "clip_ratio": 0.0, + "epoch": 0.5454545454545454, + "grad_norm": 5.1964811610508515e-05, + "kl": 0.1806640625, + "learning_rate": 9.959774064153977e-05, + "loss": 0.0001, + "step": 390 + }, + { + "clip_ratio": 0.0, + "completion_length": 103.5714340209961, + "epoch": 0.5468531468531469, + "grad_norm": 3.816406536402076e-07, + "kl": 0.173828125, + "learning_rate": 9.958997833439657e-05, + "loss": 0.0002, + "num_tokens": 3970164.0, + "reward": 2.5441431999206543, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.3214285671710968, + "rewards/check_perplexity_diff_func": 1.8571428060531616, + "rewards/check_winston_local_func": 0.3655718266963959, + "step": 391 + }, + { + "clip_ratio": 0.0, + "epoch": 0.5482517482517483, + "grad_norm": 4.199010025491445e-07, + "kl": 0.173828125, + "learning_rate": 9.958214215582018e-05, + "loss": 0.0002, + "step": 392 + }, + { + "clip_ratio": 0.0, + "completion_length": 126.8214340209961, + "epoch": 0.5496503496503496, + "grad_norm": 6.309179959293235e-06, + "kl": 0.1533203125, + "learning_rate": 9.957423211748374e-05, + "loss": 0.0002, + "num_tokens": 3992994.0, + "reward": 2.7432734966278076, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.3928571343421936, + "rewards/check_perplexity_diff_func": 1.9285714626312256, + "rewards/check_winston_local_func": 0.42184481024742126, + "step": 393 + }, + { + "clip_ratio": 0.0, + "epoch": 0.551048951048951, + "grad_norm": 5.547540566387994e-06, + "kl": 0.1533203125, + "learning_rate": 9.956624823117036e-05, + "loss": 0.0002, + "step": 394 + }, + { + "clip_ratio": 0.0, + "completion_length": 125.62500762939453, + "epoch": 0.5524475524475524, + "grad_norm": 0.007987942179452279, + "kl": 0.142578125, + "learning_rate": 9.955819050877321e-05, + "loss": 0.0001, + "num_tokens": 4015669.0, + "reward": 3.091567039489746, + "reward_std": 0.018998777493834496, + "rewards/check_gptzero_func": 0.5535714030265808, + "rewards/check_perplexity_diff_func": 1.9285714626312256, + "rewards/check_winston_local_func": 0.6094242334365845, + "step": 395 + }, + { + "clip_ratio": 0.0, + "epoch": 0.5538461538461539, + "grad_norm": 0.0034758785486018436, + "kl": 0.142578125, + "learning_rate": 9.955005896229543e-05, + "loss": 0.0, + "step": 396 + }, + { + "clip_ratio": 0.0, + "completion_length": 96.75000762939453, + "epoch": 0.5552447552447553, + "grad_norm": 0.011495641299277456, + "kl": 0.283203125, + "learning_rate": 9.954185360385013e-05, + "loss": 0.0006, + "num_tokens": 4035041.0, + "reward": 2.950004816055298, + "reward_std": 0.01496803853660822, + "rewards/check_gptzero_func": 0.4821428656578064, + "rewards/check_perplexity_diff_func": 1.8928571939468384, + "rewards/check_winston_local_func": 0.5750047564506531, + "step": 397 + }, + { + "clip_ratio": 0.0, + "epoch": 0.5566433566433566, + "grad_norm": 0.010082500068390739, + "kl": 0.283203125, + "learning_rate": 9.953357444566039e-05, + "loss": 0.0006, + "step": 398 + }, + { + "clip_ratio": 0.0, + "completion_length": 110.8214340209961, + "epoch": 0.558041958041958, + "grad_norm": 8.489376613416e-06, + "kl": 0.1787109375, + "learning_rate": 9.952522150005919e-05, + "loss": 0.0002, + "num_tokens": 4055855.0, + "reward": 2.913240909576416, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.4642857015132904, + "rewards/check_perplexity_diff_func": 1.9285714626312256, + "rewards/check_winston_local_func": 0.5203836560249329, + "step": 399 + }, + { + "clip_ratio": 0.0, + "epoch": 0.5594405594405595, + "grad_norm": 8.766485288799225e-06, + "kl": 0.1787109375, + "learning_rate": 9.951679477948947e-05, + "loss": 0.0002, + "step": 400 + }, + { + "clip_ratio": 0.0, + "completion_length": 105.46428680419922, + "epoch": 0.5608391608391609, + "grad_norm": 8.652849000461684e-06, + "kl": 0.150390625, + "learning_rate": 9.9508294296504e-05, + "loss": 0.0002, + "num_tokens": 4076051.0, + "reward": 2.871354818344116, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.4285714328289032, + "rewards/check_perplexity_diff_func": 2.0, + "rewards/check_winston_local_func": 0.44278329610824585, + "step": 401 + }, + { + "clip_ratio": 0.0, + "epoch": 0.5622377622377622, + "grad_norm": 8.18246821524729e-06, + "kl": 0.150390625, + "learning_rate": 9.949972006376556e-05, + "loss": 0.0002, + "step": 402 + }, + { + "clip_ratio": 0.0, + "completion_length": 75.25, + "epoch": 0.5636363636363636, + "grad_norm": 7.351342876079361e-07, + "kl": 0.1953125, + "learning_rate": 9.949107209404665e-05, + "loss": 0.0002, + "num_tokens": 4092527.0, + "reward": 2.9322519302368164, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.3571428656578064, + "rewards/check_perplexity_diff_func": 2.0, + "rewards/check_winston_local_func": 0.5751089453697205, + "step": 403 + }, + { + "clip_ratio": 0.0, + "epoch": 0.5650349650349651, + "grad_norm": 7.136344464618312e-07, + "kl": 0.1953125, + "learning_rate": 9.948235040022966e-05, + "loss": 0.0002, + "step": 404 + }, + { + "clip_ratio": 0.0, + "completion_length": 108.08928680419922, + "epoch": 0.5664335664335665, + "grad_norm": 0.007956878335048736, + "kl": 0.169921875, + "learning_rate": 9.947355499530683e-05, + "loss": 0.0049, + "num_tokens": 4113282.0, + "reward": 3.1798436641693115, + "reward_std": 0.06495096534490585, + "rewards/check_gptzero_func": 0.6785714030265808, + "rewards/check_perplexity_diff_func": 1.9642857313156128, + "rewards/check_winston_local_func": 0.5369864702224731, + "step": 405 + }, + { + "clip_ratio": 0.0, + "epoch": 0.5678321678321678, + "grad_norm": 0.006640329404323574, + "kl": 0.1728515625, + "learning_rate": 9.946468589238021e-05, + "loss": 0.0049, + "step": 406 + }, + { + "clip_ratio": 0.0, + "completion_length": 121.39286041259766, + "epoch": 0.5692307692307692, + "grad_norm": 1.5620797981026132e-06, + "kl": 0.1396484375, + "learning_rate": 9.945574310466159e-05, + "loss": 0.0001, + "num_tokens": 4135506.0, + "reward": 2.573127269744873, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.2857142984867096, + "rewards/check_perplexity_diff_func": 1.9285714626312256, + "rewards/check_winston_local_func": 0.3588416278362274, + "step": 407 + }, + { + "clip_ratio": 0.0, + "epoch": 0.5706293706293706, + "grad_norm": 1.51593967204227e-06, + "kl": 0.1396484375, + "learning_rate": 9.944672664547252e-05, + "loss": 0.0001, + "step": 408 + }, + { + "clip_ratio": 0.0, + "completion_length": 107.8214340209961, + "epoch": 0.5720279720279721, + "grad_norm": 0.0007775272826913423, + "kl": 0.17578125, + "learning_rate": 9.943763652824436e-05, + "loss": 0.0002, + "num_tokens": 4156060.0, + "reward": 2.55350661277771, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.3571428656578064, + "rewards/check_perplexity_diff_func": 1.7142857313156128, + "rewards/check_winston_local_func": 0.48207801580429077, + "step": 409 + }, + { + "clip_ratio": 0.0, + "epoch": 0.5734265734265734, + "grad_norm": 0.0008914625567854122, + "kl": 0.1767578125, + "learning_rate": 9.942847276651811e-05, + "loss": 0.0002, + "step": 410 + }, + { + "clip_ratio": 0.0, + "completion_length": 101.62500762939453, + "epoch": 0.5748251748251748, + "grad_norm": 4.9539161428254635e-05, + "kl": 0.1474609375, + "learning_rate": 9.941923537394456e-05, + "loss": 0.0001, + "num_tokens": 4176019.0, + "reward": 2.8906424045562744, + "reward_std": 0.00012528452498372644, + "rewards/check_gptzero_func": 0.4285714328289032, + "rewards/check_perplexity_diff_func": 1.9285714626312256, + "rewards/check_winston_local_func": 0.5334992408752441, + "step": 411 + }, + { + "clip_ratio": 0.0, + "epoch": 0.5762237762237762, + "grad_norm": 5.844422779033556e-05, + "kl": 0.1474609375, + "learning_rate": 9.940992436428409e-05, + "loss": 0.0001, + "step": 412 + }, + { + "clip_ratio": 0.0, + "completion_length": 105.28572082519531, + "epoch": 0.5776223776223777, + "grad_norm": 0.0006008878616346907, + "kl": 0.2216796875, + "learning_rate": 9.940053975140684e-05, + "loss": 0.0002, + "num_tokens": 4196667.0, + "reward": 2.943253755569458, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.5357142686843872, + "rewards/check_perplexity_diff_func": 1.8571428060531616, + "rewards/check_winston_local_func": 0.5503963828086853, + "step": 413 + }, + { + "clip_ratio": 0.0, + "epoch": 0.579020979020979, + "grad_norm": 0.0003365492175069851, + "kl": 0.220703125, + "learning_rate": 9.939108154929252e-05, + "loss": 0.0002, + "step": 414 + }, + { + "clip_ratio": 0.0, + "completion_length": 114.60714721679688, + "epoch": 0.5804195804195804, + "grad_norm": 0.008680207230787525, + "kl": 0.171875, + "learning_rate": 9.938154977203049e-05, + "loss": -0.0005, + "num_tokens": 4217727.0, + "reward": 2.7773444652557373, + "reward_std": 0.051937274634838104, + "rewards/check_gptzero_func": 0.4285714328289032, + "rewards/check_perplexity_diff_func": 1.8214285373687744, + "rewards/check_winston_local_func": 0.5273441076278687, + "step": 415 + }, + { + "clip_ratio": 0.0, + "epoch": 0.5818181818181818, + "grad_norm": 0.00033864647249595475, + "kl": 0.171875, + "learning_rate": 9.937194443381972e-05, + "loss": -0.0005, + "step": 416 + }, + { + "clip_ratio": 0.0, + "completion_length": 100.60714721679688, + "epoch": 0.5832167832167832, + "grad_norm": 4.017405087628847e-06, + "kl": 0.193359375, + "learning_rate": 9.936226554896875e-05, + "loss": 0.0002, + "num_tokens": 4237189.0, + "reward": 2.7684807777404785, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.3928571343421936, + "rewards/check_perplexity_diff_func": 1.9285714626312256, + "rewards/check_winston_local_func": 0.44705215096473694, + "step": 417 + }, + { + "clip_ratio": 0.0, + "epoch": 0.5846153846153846, + "grad_norm": 3.880139616189675e-06, + "kl": 0.193359375, + "learning_rate": 9.935251313189564e-05, + "loss": 0.0002, + "step": 418 + }, + { + "clip_ratio": 0.0, + "completion_length": 107.75000762939453, + "epoch": 0.586013986013986, + "grad_norm": 0.00032680437267053027, + "kl": 0.1826171875, + "learning_rate": 9.934268719712807e-05, + "loss": 0.0002, + "num_tokens": 4258213.0, + "reward": 2.577641487121582, + "reward_std": 0.0015505586052313447, + "rewards/check_gptzero_func": 0.2857142984867096, + "rewards/check_perplexity_diff_func": 1.8571428060531616, + "rewards/check_winston_local_func": 0.4347843825817108, + "step": 419 + }, + { + "clip_ratio": 0.0, + "epoch": 0.5874125874125874, + "grad_norm": 0.0003238465648274268, + "kl": 0.1826171875, + "learning_rate": 9.933278775930317e-05, + "loss": 0.0002, + "step": 420 + }, + { + "clip_ratio": 0.0, + "completion_length": 108.67857360839844, + "epoch": 0.5888111888111888, + "grad_norm": 0.015591092439371165, + "kl": 0.16015625, + "learning_rate": 9.932281483316758e-05, + "loss": 0.0004, + "num_tokens": 4278847.0, + "reward": 2.615447998046875, + "reward_std": 0.05428782477974892, + "rewards/check_gptzero_func": 0.3214285671710968, + "rewards/check_perplexity_diff_func": 1.8214285373687744, + "rewards/check_winston_local_func": 0.47259071469306946, + "step": 421 + }, + { + "clip_ratio": 0.0, + "epoch": 0.5902097902097903, + "grad_norm": 0.00016911180587449477, + "kl": 0.16015625, + "learning_rate": 9.931276843357742e-05, + "loss": 0.0003, + "step": 422 + }, + { + "clip_ratio": 0.0, + "completion_length": 111.16072082519531, + "epoch": 0.5916083916083916, + "grad_norm": 1.6847767080478394e-05, + "kl": 0.1630859375, + "learning_rate": 9.930264857549825e-05, + "loss": 0.0002, + "num_tokens": 4300096.0, + "reward": 2.857067823410034, + "reward_std": 0.00101040443405509, + "rewards/check_gptzero_func": 0.3928571343421936, + "rewards/check_perplexity_diff_func": 1.8571428060531616, + "rewards/check_winston_local_func": 0.6070676445960999, + "step": 423 + }, + { + "clip_ratio": 0.0, + "epoch": 0.593006993006993, + "grad_norm": 2.453025078669727e-05, + "kl": 0.1630859375, + "learning_rate": 9.929245527400503e-05, + "loss": 0.0002, + "step": 424 + }, + { + "clip_ratio": 0.0, + "completion_length": 102.50000762939453, + "epoch": 0.5944055944055944, + "grad_norm": 0.0015183266942611472, + "kl": 0.173828125, + "learning_rate": 9.928218854428221e-05, + "loss": 0.0002, + "num_tokens": 4320064.0, + "reward": 2.6430578231811523, + "reward_std": 5.020291791879572e-05, + "rewards/check_gptzero_func": 0.3571428656578064, + "rewards/check_perplexity_diff_func": 1.7857142686843872, + "rewards/check_winston_local_func": 0.5002006888389587, + "step": 425 + }, + { + "clip_ratio": 0.0, + "epoch": 0.5958041958041959, + "grad_norm": 0.0013596061771866538, + "kl": 0.1708984375, + "learning_rate": 9.927184840162354e-05, + "loss": 0.0002, + "step": 426 + }, + { + "clip_ratio": 0.0, + "completion_length": 110.66072082519531, + "epoch": 0.5972027972027972, + "grad_norm": 0.12076121626694686, + "kl": 0.189453125, + "learning_rate": 9.926143486143214e-05, + "loss": -0.001, + "num_tokens": 4341145.0, + "reward": 2.6579177379608154, + "reward_std": 0.05154726281762123, + "rewards/check_gptzero_func": 0.4285714328289032, + "rewards/check_perplexity_diff_func": 1.8214285373687744, + "rewards/check_winston_local_func": 0.4079175889492035, + "step": 427 + }, + { + "clip_ratio": 0.0, + "epoch": 0.5986013986013986, + "grad_norm": 0.0008289218843548021, + "kl": 0.19140625, + "learning_rate": 9.92509479392205e-05, + "loss": -0.0013, + "step": 428 + }, + { + "clip_ratio": 0.0, + "completion_length": 89.67857360839844, + "epoch": 0.6, + "grad_norm": 0.005379412774279031, + "kl": 0.1787109375, + "learning_rate": 9.924038765061042e-05, + "loss": 0.0001, + "num_tokens": 4359277.0, + "reward": 2.633777141571045, + "reward_std": 0.05039219558238983, + "rewards/check_gptzero_func": 0.2857142984867096, + "rewards/check_perplexity_diff_func": 1.9642857313156128, + "rewards/check_winston_local_func": 0.38377735018730164, + "step": 429 + }, + { + "clip_ratio": 0.0, + "epoch": 0.6013986013986014, + "grad_norm": 0.010439668945298157, + "kl": 0.1787109375, + "learning_rate": 9.922975401133293e-05, + "loss": 0.0002, + "step": 430 + }, + { + "clip_ratio": 0.0, + "completion_length": 102.67857360839844, + "epoch": 0.6027972027972028, + "grad_norm": 0.06317955454404899, + "kl": 0.201171875, + "learning_rate": 9.92190470372284e-05, + "loss": 0.0068, + "num_tokens": 4379109.0, + "reward": 2.6650938987731934, + "reward_std": 0.05930115655064583, + "rewards/check_gptzero_func": 0.25, + "rewards/check_perplexity_diff_func": 1.9642857313156128, + "rewards/check_winston_local_func": 0.4508081078529358, + "step": 431 + }, + { + "clip_ratio": 0.0002587991766631603, + "epoch": 0.6041958041958042, + "grad_norm": 0.05122842630547611, + "kl": 0.1865234375, + "learning_rate": 9.920826674424642e-05, + "loss": 0.0063, + "step": 432 + }, + { + "clip_ratio": 0.0, + "completion_length": 80.35714721679688, + "epoch": 0.6055944055944056, + "grad_norm": 7.162658137250354e-07, + "kl": 0.1875, + "learning_rate": 9.919741314844577e-05, + "loss": 0.0002, + "num_tokens": 4396693.0, + "reward": 3.0637807846069336, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.5357142686843872, + "rewards/check_perplexity_diff_func": 1.8571428060531616, + "rewards/check_winston_local_func": 0.6709235310554504, + "step": 433 + }, + { + "clip_ratio": 0.0, + "epoch": 0.606993006993007, + "grad_norm": 8.451796450955314e-07, + "kl": 0.1875, + "learning_rate": 9.918648626599447e-05, + "loss": 0.0002, + "step": 434 + }, + { + "clip_ratio": 0.0, + "completion_length": 124.98214721679688, + "epoch": 0.6083916083916084, + "grad_norm": 0.016099687268680437, + "kl": 0.1455078125, + "learning_rate": 9.91754861131697e-05, + "loss": -0.0001, + "num_tokens": 4418950.0, + "reward": 2.94807505607605, + "reward_std": 0.050821539014577866, + "rewards/check_gptzero_func": 0.5357142686843872, + "rewards/check_perplexity_diff_func": 1.8928571939468384, + "rewards/check_winston_local_func": 0.5195035338401794, + "step": 435 + }, + { + "clip_ratio": 0.0, + "epoch": 0.6097902097902098, + "grad_norm": 0.014390989758817311, + "kl": 0.1455078125, + "learning_rate": 9.916441270635772e-05, + "loss": -0.0001, + "step": 436 + }, + { + "clip_ratio": 0.0, + "completion_length": 106.64286041259766, + "epoch": 0.6111888111888112, + "grad_norm": 1.968332582426962e-06, + "kl": 0.14453125, + "learning_rate": 9.915326606205404e-05, + "loss": 0.0001, + "num_tokens": 4439018.0, + "reward": 2.77024507522583, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.4285714328289032, + "rewards/check_perplexity_diff_func": 1.9285714626312256, + "rewards/check_winston_local_func": 0.4131018817424774, + "step": 437 + }, + { + "clip_ratio": 0.0, + "epoch": 0.6125874125874126, + "grad_norm": 1.8961430119120163e-06, + "kl": 0.14453125, + "learning_rate": 9.914204619686314e-05, + "loss": 0.0001, + "step": 438 + }, + { + "clip_ratio": 0.0, + "completion_length": 102.21428680419922, + "epoch": 0.6139860139860139, + "grad_norm": 0.000668351376631818, + "kl": 0.1728515625, + "learning_rate": 9.913075312749866e-05, + "loss": 0.0002, + "num_tokens": 4458806.0, + "reward": 2.8182172775268555, + "reward_std": 0.0025242711417376995, + "rewards/check_gptzero_func": 0.5357142686843872, + "rewards/check_perplexity_diff_func": 1.8571428060531616, + "rewards/check_winston_local_func": 0.42535990476608276, + "step": 439 + }, + { + "clip_ratio": 0.0, + "epoch": 0.6153846153846154, + "grad_norm": 0.0006751756209535994, + "kl": 0.1728515625, + "learning_rate": 9.911938687078324e-05, + "loss": 0.0002, + "step": 440 + }, + { + "clip_ratio": 0.0, + "completion_length": 81.25, + "epoch": 0.6167832167832168, + "grad_norm": 0.007904133813646444, + "kl": 0.240234375, + "learning_rate": 9.910794744364857e-05, + "loss": -0.0001, + "num_tokens": 4475982.0, + "reward": 2.8269970417022705, + "reward_std": 0.05014092102646828, + "rewards/check_gptzero_func": 0.3928571343421936, + "rewards/check_perplexity_diff_func": 1.8928571939468384, + "rewards/check_winston_local_func": 0.5412828326225281, + "step": 441 + }, + { + "clip_ratio": 0.0, + "epoch": 0.6181818181818182, + "grad_norm": 0.00827597712943806, + "kl": 0.240234375, + "learning_rate": 9.909643486313533e-05, + "loss": -0.0001, + "step": 442 + }, + { + "clip_ratio": 0.0, + "completion_length": 113.00000762939453, + "epoch": 0.6195804195804195, + "grad_norm": 6.016032313066646e-06, + "kl": 0.16796875, + "learning_rate": 9.908484914639318e-05, + "loss": 0.0002, + "num_tokens": 4497200.0, + "reward": 2.9243876934051514, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.3928571343421936, + "rewards/check_perplexity_diff_func": 2.0, + "rewards/check_winston_local_func": 0.5315303206443787, + "step": 443 + }, + { + "clip_ratio": 0.0, + "epoch": 0.620979020979021, + "grad_norm": 7.198804696892823e-06, + "kl": 0.16796875, + "learning_rate": 9.90731903106807e-05, + "loss": 0.0002, + "step": 444 + }, + { + "clip_ratio": 0.0, + "completion_length": 129.375, + "epoch": 0.6223776223776224, + "grad_norm": 0.02075011817403209, + "kl": 0.158203125, + "learning_rate": 9.90614583733654e-05, + "loss": -0.0001, + "num_tokens": 4519807.0, + "reward": 2.866011619567871, + "reward_std": 0.10101933032274246, + "rewards/check_gptzero_func": 0.4642857015132904, + "rewards/check_perplexity_diff_func": 1.9285714626312256, + "rewards/check_winston_local_func": 0.47315436601638794, + "step": 445 + }, + { + "clip_ratio": 0.0, + "epoch": 0.6237762237762238, + "grad_norm": 0.0034619333956659956, + "kl": 0.1591796875, + "learning_rate": 9.904965335192373e-05, + "loss": -0.0002, + "step": 446 + }, + { + "clip_ratio": 0.0, + "completion_length": 98.92857360839844, + "epoch": 0.6251748251748251, + "grad_norm": 0.010057352537761423, + "kl": 0.212890625, + "learning_rate": 9.903777526394094e-05, + "loss": -0.0086, + "num_tokens": 4539039.0, + "reward": 2.980358362197876, + "reward_std": 0.05100167542695999, + "rewards/check_gptzero_func": 0.4285714328289032, + "rewards/check_perplexity_diff_func": 1.9642857313156128, + "rewards/check_winston_local_func": 0.587501049041748, + "step": 447 + }, + { + "clip_ratio": 0.0, + "epoch": 0.6265734265734266, + "grad_norm": 0.006958390301115274, + "kl": 0.21875, + "learning_rate": 9.90258241271112e-05, + "loss": -0.0087, + "step": 448 + }, + { + "clip_ratio": 0.0, + "completion_length": 111.71429443359375, + "epoch": 0.627972027972028, + "grad_norm": 1.6536770511152556e-06, + "kl": 0.142578125, + "learning_rate": 9.901379995923738e-05, + "loss": 0.0001, + "num_tokens": 4559675.0, + "reward": 2.681791067123413, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.2857142984867096, + "rewards/check_perplexity_diff_func": 1.9285714626312256, + "rewards/check_winston_local_func": 0.4675052762031555, + "step": 449 + }, + { + "clip_ratio": 0.0, + "epoch": 0.6293706293706294, + "grad_norm": 1.6773170531853214e-06, + "kl": 0.142578125, + "learning_rate": 9.900170277823129e-05, + "loss": 0.0001, + "step": 450 + }, + { + "clip_ratio": 0.0, + "completion_length": 107.35714721679688, + "epoch": 0.6307692307692307, + "grad_norm": 2.0958023033337787e-06, + "kl": 0.2109375, + "learning_rate": 9.898953260211338e-05, + "loss": 0.0002, + "num_tokens": 4579995.0, + "reward": 2.7400293350219727, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.3928571343421936, + "rewards/check_perplexity_diff_func": 1.9285714626312256, + "rewards/check_winston_local_func": 0.4186006188392639, + "step": 451 + }, + { + "clip_ratio": 0.0, + "epoch": 0.6321678321678321, + "grad_norm": 1.982971590769392e-06, + "kl": 0.2109375, + "learning_rate": 9.897728944901292e-05, + "loss": 0.0002, + "step": 452 + }, + { + "clip_ratio": 0.0, + "completion_length": 118.26786041259766, + "epoch": 0.6335664335664336, + "grad_norm": 0.009450375112303724, + "kl": 0.1416015625, + "learning_rate": 9.896497333716783e-05, + "loss": -0.0051, + "num_tokens": 4602048.0, + "reward": 2.640944004058838, + "reward_std": 0.043768420815467834, + "rewards/check_gptzero_func": 0.5357142686843872, + "rewards/check_perplexity_diff_func": 1.5357142686843872, + "rewards/check_winston_local_func": 0.5695151090621948, + "step": 453 + }, + { + "clip_ratio": 0.0, + "epoch": 0.634965034965035, + "grad_norm": 0.0017713963386990425, + "kl": 0.1416015625, + "learning_rate": 9.895258428492475e-05, + "loss": -0.0051, + "step": 454 + }, + { + "clip_ratio": 0.0, + "completion_length": 98.42857360839844, + "epoch": 0.6363636363636364, + "grad_norm": 0.014431904681184264, + "kl": 0.1943359375, + "learning_rate": 9.894012231073894e-05, + "loss": 0.0002, + "num_tokens": 4621654.0, + "reward": 2.7700040340423584, + "reward_std": 0.03728308901190758, + "rewards/check_gptzero_func": 0.4107142984867096, + "rewards/check_perplexity_diff_func": 1.9285714626312256, + "rewards/check_winston_local_func": 0.43071839213371277, + "step": 455 + }, + { + "clip_ratio": 0.0, + "epoch": 0.6377622377622377, + "grad_norm": 0.008239966094546339, + "kl": 0.1943359375, + "learning_rate": 9.892758743317434e-05, + "loss": 0.0002, + "step": 456 + }, + { + "clip_ratio": 0.0, + "completion_length": 90.50000762939453, + "epoch": 0.6391608391608392, + "grad_norm": 2.579298856684113e-06, + "kl": 0.1923828125, + "learning_rate": 9.891497967090344e-05, + "loss": 0.0002, + "num_tokens": 4640140.0, + "reward": 2.6383018493652344, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.25, + "rewards/check_perplexity_diff_func": 1.9285714626312256, + "rewards/check_winston_local_func": 0.4597306251525879, + "step": 457 + }, + { + "clip_ratio": 0.0, + "epoch": 0.6405594405594406, + "grad_norm": 2.6625863420088633e-06, + "kl": 0.1923828125, + "learning_rate": 9.890229904270731e-05, + "loss": 0.0002, + "step": 458 + }, + { + "clip_ratio": 0.0, + "completion_length": 113.03572082519531, + "epoch": 0.641958041958042, + "grad_norm": 1.7932858894376123e-05, + "kl": 0.1748046875, + "learning_rate": 9.888954556747563e-05, + "loss": 0.0002, + "num_tokens": 4661276.0, + "reward": 2.74210524559021, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.3571428656578064, + "rewards/check_perplexity_diff_func": 2.0, + "rewards/check_winston_local_func": 0.3849623203277588, + "step": 459 + }, + { + "clip_ratio": 0.0, + "epoch": 0.6433566433566433, + "grad_norm": 1.7905922488331998e-05, + "kl": 0.1748046875, + "learning_rate": 9.887671926420648e-05, + "loss": 0.0002, + "step": 460 + }, + { + "clip_ratio": 0.0, + "completion_length": 112.42857360839844, + "epoch": 0.6447552447552447, + "grad_norm": 0.014077582219597583, + "kl": 0.173828125, + "learning_rate": 9.886382015200652e-05, + "loss": 0.0002, + "num_tokens": 4682450.0, + "reward": 2.8680295944213867, + "reward_std": 0.05049533396959305, + "rewards/check_gptzero_func": 0.4642857015132904, + "rewards/check_perplexity_diff_func": 1.9642857313156128, + "rewards/check_winston_local_func": 0.43945807218551636, + "step": 461 + }, + { + "clip_ratio": 0.0, + "epoch": 0.6461538461538462, + "grad_norm": 0.010796297977350395, + "kl": 0.173828125, + "learning_rate": 9.885084825009086e-05, + "loss": 0.0001, + "step": 462 + }, + { + "clip_ratio": 0.0, + "completion_length": 105.14286041259766, + "epoch": 0.6475524475524476, + "grad_norm": 0.02303930816984521, + "kl": 0.17578125, + "learning_rate": 9.883780357778299e-05, + "loss": 0.0003, + "num_tokens": 4702710.0, + "reward": 2.885434150695801, + "reward_std": 0.06019994616508484, + "rewards/check_gptzero_func": 0.4285714328289032, + "rewards/check_perplexity_diff_func": 1.75, + "rewards/check_winston_local_func": 0.7068625092506409, + "step": 463 + }, + { + "clip_ratio": 0.0, + "epoch": 0.6489510489510489, + "grad_norm": 0.0015050223064125838, + "kl": 0.1767578125, + "learning_rate": 9.882468615451484e-05, + "loss": 0.0002, + "step": 464 + }, + { + "clip_ratio": 0.0, + "completion_length": 127.5714340209961, + "epoch": 0.6503496503496503, + "grad_norm": 3.0460982652821775e-06, + "kl": 0.1650390625, + "learning_rate": 9.881149599982671e-05, + "loss": 0.0002, + "num_tokens": 4725654.0, + "reward": 2.407341957092285, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.2857142984867096, + "rewards/check_perplexity_diff_func": 1.8571428060531616, + "rewards/check_winston_local_func": 0.2644847333431244, + "step": 465 + }, + { + "clip_ratio": 0.0, + "epoch": 0.6517482517482518, + "grad_norm": 2.8166666383288673e-06, + "kl": 0.1650390625, + "learning_rate": 9.879823313336722e-05, + "loss": 0.0002, + "step": 466 + }, + { + "clip_ratio": 0.0, + "completion_length": 122.16072082519531, + "epoch": 0.6531468531468532, + "grad_norm": 0.0001021700279542065, + "kl": 0.140625, + "learning_rate": 9.878489757489337e-05, + "loss": 0.0001, + "num_tokens": 4748375.0, + "reward": 2.6119117736816406, + "reward_std": 2.7804879209725186e-05, + "rewards/check_gptzero_func": 0.5357142686843872, + "rewards/check_perplexity_diff_func": 1.5714285373687744, + "rewards/check_winston_local_func": 0.5047687888145447, + "step": 467 + }, + { + "clip_ratio": 0.0, + "epoch": 0.6545454545454545, + "grad_norm": 9.33433928584176e-05, + "kl": 0.140625, + "learning_rate": 9.877148934427037e-05, + "loss": 0.0001, + "step": 468 + }, + { + "clip_ratio": 0.0006711409660056233, + "completion_length": 95.64286041259766, + "epoch": 0.6559440559440559, + "grad_norm": 0.0004684591425632982, + "kl": 0.171875, + "learning_rate": 9.87580084614717e-05, + "loss": 0.0002, + "num_tokens": 4767439.0, + "reward": 2.3873257637023926, + "reward_std": 0.0017833748133853078, + "rewards/check_gptzero_func": 0.25, + "rewards/check_perplexity_diff_func": 1.7142857313156128, + "rewards/check_winston_local_func": 0.42303988337516785, + "step": 469 + }, + { + "clip_ratio": 0.0004793864209204912, + "epoch": 0.6573426573426573, + "grad_norm": 0.00047799981287926325, + "kl": 0.171875, + "learning_rate": 9.874445494657911e-05, + "loss": 0.0002, + "step": 470 + }, + { + "clip_ratio": 0.0, + "completion_length": 79.39286041259766, + "epoch": 0.6587412587412588, + "grad_norm": 0.011441113949782547, + "kl": 0.216796875, + "learning_rate": 9.873082881978251e-05, + "loss": -0.0064, + "num_tokens": 4784715.0, + "reward": 3.037661075592041, + "reward_std": 0.023328183218836784, + "rewards/check_gptzero_func": 0.4821428656578064, + "rewards/check_perplexity_diff_func": 1.9642857313156128, + "rewards/check_winston_local_func": 0.5912323594093323, + "step": 471 + }, + { + "clip_ratio": 0.0, + "epoch": 0.6601398601398601, + "grad_norm": 0.008433986682608557, + "kl": 0.216796875, + "learning_rate": 9.871713010137997e-05, + "loss": -0.0064, + "step": 472 + }, + { + "clip_ratio": 0.0, + "completion_length": 119.42857360839844, + "epoch": 0.6615384615384615, + "grad_norm": 7.638914800437301e-06, + "kl": 0.1435546875, + "learning_rate": 9.870335881177774e-05, + "loss": 0.0001, + "num_tokens": 4806207.0, + "reward": 2.9706830978393555, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.4642857015132904, + "rewards/check_perplexity_diff_func": 2.0, + "rewards/check_winston_local_func": 0.5063972473144531, + "step": 473 + }, + { + "clip_ratio": 0.0, + "epoch": 0.6629370629370629, + "grad_norm": 7.505737589947172e-06, + "kl": 0.1435546875, + "learning_rate": 9.868951497149011e-05, + "loss": 0.0001, + "step": 474 + }, + { + "clip_ratio": 0.00019073051225859672, + "completion_length": 89.87500762939453, + "epoch": 0.6643356643356644, + "grad_norm": 0.03662860311910791, + "kl": 0.2060546875, + "learning_rate": 9.86755986011395e-05, + "loss": 0.0065, + "num_tokens": 4824332.0, + "reward": 2.643022060394287, + "reward_std": 0.04776443541049957, + "rewards/check_gptzero_func": 0.3214285671710968, + "rewards/check_perplexity_diff_func": 1.8928571939468384, + "rewards/check_winston_local_func": 0.42873620986938477, + "step": 475 + }, + { + "clip_ratio": 0.0, + "epoch": 0.6657342657342658, + "grad_norm": 0.03126231173733144, + "kl": 0.2060546875, + "learning_rate": 9.866160972145634e-05, + "loss": 0.0062, + "step": 476 + }, + { + "clip_ratio": 0.0, + "completion_length": 123.10714721679688, + "epoch": 0.6671328671328671, + "grad_norm": 0.0002559996362709361, + "kl": 0.1357421875, + "learning_rate": 9.864754835327909e-05, + "loss": 0.0001, + "num_tokens": 4847094.0, + "reward": 2.767104387283325, + "reward_std": 0.0007265785825438797, + "rewards/check_gptzero_func": 0.5, + "rewards/check_perplexity_diff_func": 1.7142857313156128, + "rewards/check_winston_local_func": 0.5528185963630676, + "step": 477 + }, + { + "clip_ratio": 8.97343925316818e-05, + "epoch": 0.6685314685314685, + "grad_norm": 0.00042633622080268514, + "kl": 0.1357421875, + "learning_rate": 9.86334145175542e-05, + "loss": 0.0001, + "step": 478 + }, + { + "clip_ratio": 0.0, + "completion_length": 96.67857360839844, + "epoch": 0.66993006993007, + "grad_norm": 0.0008693689003033054, + "kl": 0.16015625, + "learning_rate": 9.861920823533606e-05, + "loss": 0.0002, + "num_tokens": 4866394.0, + "reward": 2.5761330127716064, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.3571428656578064, + "rewards/check_perplexity_diff_func": 1.7142857313156128, + "rewards/check_winston_local_func": 0.504704475402832, + "step": 479 + }, + { + "clip_ratio": 0.0, + "epoch": 0.6713286713286714, + "grad_norm": 0.0008435837049153919, + "kl": 0.16015625, + "learning_rate": 9.860492952778696e-05, + "loss": 0.0002, + "step": 480 + }, + { + "clip_ratio": 0.0, + "completion_length": 94.39286041259766, + "epoch": 0.6727272727272727, + "grad_norm": 3.349433978828943e-06, + "kl": 0.177734375, + "learning_rate": 9.859057841617709e-05, + "loss": 0.0002, + "num_tokens": 4885318.0, + "reward": 2.8612334728240967, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.3571428656578064, + "rewards/check_perplexity_diff_func": 1.9285714626312256, + "rewards/check_winston_local_func": 0.5755191445350647, + "step": 481 + }, + { + "clip_ratio": 0.0, + "epoch": 0.6741258741258741, + "grad_norm": 3.5095106255593373e-06, + "kl": 0.177734375, + "learning_rate": 9.857615492188452e-05, + "loss": 0.0002, + "step": 482 + }, + { + "clip_ratio": 0.0, + "completion_length": 87.98214721679688, + "epoch": 0.6755244755244755, + "grad_norm": 0.00016582191324414645, + "kl": 0.208984375, + "learning_rate": 9.856165906639513e-05, + "loss": 0.0002, + "num_tokens": 4903401.0, + "reward": 2.797102928161621, + "reward_std": 0.0008326892857439816, + "rewards/check_gptzero_func": 0.3928571343421936, + "rewards/check_perplexity_diff_func": 1.9285714626312256, + "rewards/check_winston_local_func": 0.4756740629673004, + "step": 483 + }, + { + "clip_ratio": 0.0, + "epoch": 0.676923076923077, + "grad_norm": 0.00017592290286556547, + "kl": 0.208984375, + "learning_rate": 9.85470908713026e-05, + "loss": 0.0002, + "step": 484 + }, + { + "clip_ratio": 0.0, + "completion_length": 122.14286041259766, + "epoch": 0.6783216783216783, + "grad_norm": 3.840262302649335e-05, + "kl": 0.140625, + "learning_rate": 9.853245035830834e-05, + "loss": 0.0001, + "num_tokens": 4925777.0, + "reward": 2.80380916595459, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.4642857015132904, + "rewards/check_perplexity_diff_func": 1.8571428060531616, + "rewards/check_winston_local_func": 0.4823804199695587, + "step": 485 + }, + { + "clip_ratio": 0.0, + "epoch": 0.6797202797202797, + "grad_norm": 3.855750270963149e-05, + "kl": 0.140625, + "learning_rate": 9.851773754922152e-05, + "loss": 0.0001, + "step": 486 + }, + { + "clip_ratio": 0.0, + "completion_length": 110.5714340209961, + "epoch": 0.6811188811188811, + "grad_norm": 1.0000521495087345e-06, + "kl": 0.1572265625, + "learning_rate": 9.850295246595898e-05, + "loss": 0.0002, + "num_tokens": 4946515.0, + "reward": 2.9591448307037354, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.5, + "rewards/check_perplexity_diff_func": 1.9285714626312256, + "rewards/check_winston_local_func": 0.5305731296539307, + "step": 487 + }, + { + "clip_ratio": 0.0, + "epoch": 0.6825174825174826, + "grad_norm": 9.579988148024946e-07, + "kl": 0.1572265625, + "learning_rate": 9.848809513054523e-05, + "loss": 0.0002, + "step": 488 + }, + { + "clip_ratio": 0.0, + "completion_length": 104.67857360839844, + "epoch": 0.6839160839160839, + "grad_norm": 4.4290499511336513e-07, + "kl": 0.1728515625, + "learning_rate": 9.847316556511245e-05, + "loss": 0.0002, + "num_tokens": 4966537.0, + "reward": 3.0352413654327393, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.4642857015132904, + "rewards/check_perplexity_diff_func": 2.0, + "rewards/check_winston_local_func": 0.5709553956985474, + "step": 489 + }, + { + "clip_ratio": 0.0, + "epoch": 0.6853146853146853, + "grad_norm": 4.228282290104698e-07, + "kl": 0.1728515625, + "learning_rate": 9.845816379190036e-05, + "loss": 0.0002, + "step": 490 + }, + { + "clip_ratio": 0.0, + "completion_length": 106.92857360839844, + "epoch": 0.6867132867132867, + "grad_norm": 7.161264654013381e-07, + "kl": 0.1865234375, + "learning_rate": 9.844308983325625e-05, + "loss": 0.0002, + "num_tokens": 4986837.0, + "reward": 2.8011527061462402, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.4642857015132904, + "rewards/check_perplexity_diff_func": 1.9285714626312256, + "rewards/check_winston_local_func": 0.40829533338546753, + "step": 491 + }, + { + "clip_ratio": 0.0, + "epoch": 0.6881118881118881, + "grad_norm": 6.606409864369399e-07, + "kl": 0.1865234375, + "learning_rate": 9.842794371163501e-05, + "loss": 0.0002, + "step": 492 + }, + { + "clip_ratio": 0.0, + "completion_length": 127.3214340209961, + "epoch": 0.6895104895104895, + "grad_norm": 3.5056088482557826e-05, + "kl": 0.11767578125, + "learning_rate": 9.841272544959892e-05, + "loss": 0.0001, + "num_tokens": 5009783.0, + "reward": 2.6788315773010254, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.5, + "rewards/check_perplexity_diff_func": 1.7857142686843872, + "rewards/check_winston_local_func": 0.393117219209671, + "step": 493 + }, + { + "clip_ratio": 0.0, + "epoch": 0.6909090909090909, + "grad_norm": 3.5363245519599165e-05, + "kl": 0.11767578125, + "learning_rate": 9.839743506981782e-05, + "loss": 0.0001, + "step": 494 + }, + { + "clip_ratio": 0.0004093328316230327, + "completion_length": 99.08928680419922, + "epoch": 0.6923076923076923, + "grad_norm": 0.0028759704706919223, + "kl": 0.1650390625, + "learning_rate": 9.838207259506891e-05, + "loss": 0.0002, + "num_tokens": 5029238.0, + "reward": 2.5216493606567383, + "reward_std": 0.0011310166446492076, + "rewards/check_gptzero_func": 0.3571428656578064, + "rewards/check_perplexity_diff_func": 1.7142857313156128, + "rewards/check_winston_local_func": 0.45022064447402954, + "step": 495 + }, + { + "clip_ratio": 0.0004093328316230327, + "epoch": 0.6937062937062937, + "grad_norm": 0.0035050811312865673, + "kl": 0.1650390625, + "learning_rate": 9.836663804823683e-05, + "loss": 0.0002, + "step": 496 + }, + { + "clip_ratio": 0.0, + "completion_length": 124.85714721679688, + "epoch": 0.6951048951048951, + "grad_norm": 3.8191875557546916e-05, + "kl": 0.158203125, + "learning_rate": 9.835113145231356e-05, + "loss": 0.0002, + "num_tokens": 5051408.0, + "reward": 2.8089945316314697, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.4642857015132904, + "rewards/check_perplexity_diff_func": 1.9285714626312256, + "rewards/check_winston_local_func": 0.4161372184753418, + "step": 497 + }, + { + "clip_ratio": 0.0, + "epoch": 0.6965034965034965, + "grad_norm": 3.1854181850026e-05, + "kl": 0.158203125, + "learning_rate": 9.833555283039842e-05, + "loss": 0.0002, + "step": 498 + }, + { + "clip_ratio": 0.0, + "completion_length": 115.17857360839844, + "epoch": 0.6979020979020979, + "grad_norm": 3.735234931957505e-07, + "kl": 0.15234375, + "learning_rate": 9.831990220569801e-05, + "loss": 0.0002, + "num_tokens": 5072486.0, + "reward": 2.7535111904144287, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.3571428656578064, + "rewards/check_perplexity_diff_func": 1.9285714626312256, + "rewards/check_winston_local_func": 0.46779707074165344, + "step": 499 + }, + { + "clip_ratio": 0.0, + "epoch": 0.6993006993006993, + "grad_norm": 3.749561048042854e-07, + "kl": 0.15234375, + "learning_rate": 9.83041796015262e-05, + "loss": 0.0002, + "step": 500 + }, + { + "clip_ratio": 0.0, + "completion_length": 137.85714721679688, + "epoch": 0.7006993006993008, + "grad_norm": 0.004581651252064326, + "kl": 0.1357421875, + "learning_rate": 9.828838504130406e-05, + "loss": 0.0001, + "num_tokens": 24128.0, + "reward": 2.5751242637634277, + "reward_std": 0.050465863198041916, + "rewards/check_gptzero_func": 0.3928571343421936, + "rewards/check_perplexity_diff_func": 1.8214285373687744, + "rewards/check_winston_local_func": 0.36083847284317017, + "step": 501 + }, + { + "clip_ratio": 0.0, + "epoch": 0.7020979020979021, + "grad_norm": 0.004623783518082634, + "kl": 0.1357421875, + "learning_rate": 9.827251854855991e-05, + "loss": 0.0001, + "step": 502 + }, + { + "clip_ratio": 0.0, + "completion_length": 84.0714340209961, + "epoch": 0.7034965034965035, + "grad_norm": 5.4224474596128455e-06, + "kl": 0.1962890625, + "learning_rate": 9.825658014692914e-05, + "loss": 0.0002, + "num_tokens": 41660.0, + "reward": 2.881941556930542, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.3928571343421936, + "rewards/check_perplexity_diff_func": 2.0, + "rewards/check_winston_local_func": 0.48908427357673645, + "step": 503 + }, + { + "clip_ratio": 0.0, + "epoch": 0.7048951048951049, + "grad_norm": 5.774418529053615e-06, + "kl": 0.1962890625, + "learning_rate": 9.824056986015433e-05, + "loss": 0.0002, + "step": 504 + }, + { + "clip_ratio": 0.0, + "completion_length": 125.76786041259766, + "epoch": 0.7062937062937062, + "grad_norm": 0.012190639094338367, + "kl": 0.158203125, + "learning_rate": 9.82244877120851e-05, + "loss": -0.0, + "num_tokens": 63809.0, + "reward": 2.9153220653533936, + "reward_std": 0.05151599273085594, + "rewards/check_gptzero_func": 0.5, + "rewards/check_perplexity_diff_func": 1.9642857313156128, + "rewards/check_winston_local_func": 0.45103612542152405, + "step": 505 + }, + { + "clip_ratio": 0.0, + "epoch": 0.7076923076923077, + "grad_norm": 0.0008706312902749691, + "kl": 0.1572265625, + "learning_rate": 9.820833372667812e-05, + "loss": -0.0001, + "step": 506 + }, + { + "clip_ratio": 0.0, + "completion_length": 106.67857360839844, + "epoch": 0.7090909090909091, + "grad_norm": 7.4245877265908374e-06, + "kl": 0.171875, + "learning_rate": 9.819210792799712e-05, + "loss": 0.0002, + "num_tokens": 83877.0, + "reward": 2.742607831954956, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.3571428656578064, + "rewards/check_perplexity_diff_func": 2.0, + "rewards/check_winston_local_func": 0.3854646682739258, + "step": 507 + }, + { + "clip_ratio": 0.0, + "epoch": 0.7104895104895105, + "grad_norm": 5.9863315776137925e-06, + "kl": 0.171875, + "learning_rate": 9.817581034021272e-05, + "loss": 0.0002, + "step": 508 + }, + { + "clip_ratio": 0.0, + "completion_length": 112.21429443359375, + "epoch": 0.7118881118881119, + "grad_norm": 1.5219106062929897e-05, + "kl": 0.154296875, + "learning_rate": 9.815944098760257e-05, + "loss": 0.0002, + "num_tokens": 105225.0, + "reward": 2.886364698410034, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.3928571343421936, + "rewards/check_perplexity_diff_func": 1.9285714626312256, + "rewards/check_winston_local_func": 0.5649359226226807, + "step": 509 + }, + { + "clip_ratio": 0.0, + "epoch": 0.7132867132867133, + "grad_norm": 3.826854189322478e-05, + "kl": 0.154296875, + "learning_rate": 9.814299989455117e-05, + "loss": 0.0002, + "step": 510 + }, + { + "clip_ratio": 0.0, + "completion_length": 93.50000762939453, + "epoch": 0.7146853146853147, + "grad_norm": 3.6634275067756628e-06, + "kl": 0.18359375, + "learning_rate": 9.81264870855499e-05, + "loss": 0.0002, + "num_tokens": 124237.0, + "reward": 2.9750092029571533, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.5, + "rewards/check_perplexity_diff_func": 1.9285714626312256, + "rewards/check_winston_local_func": 0.5464377403259277, + "step": 511 + }, + { + "clip_ratio": 0.0, + "epoch": 0.7160839160839161, + "grad_norm": 3.3120330925968346e-06, + "kl": 0.18359375, + "learning_rate": 9.810990258519699e-05, + "loss": 0.0002, + "step": 512 + }, + { + "clip_ratio": 0.0, + "completion_length": 114.5714340209961, + "epoch": 0.7174825174825175, + "grad_norm": 1.5887563588226216e-07, + "kl": 0.1474609375, + "learning_rate": 9.809324641819741e-05, + "loss": 0.0001, + "num_tokens": 145557.0, + "reward": 2.767850637435913, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.3571428656578064, + "rewards/check_perplexity_diff_func": 2.0, + "rewards/check_winston_local_func": 0.4107076823711395, + "step": 513 + }, + { + "clip_ratio": 0.0, + "epoch": 0.7188811188811188, + "grad_norm": 1.5359302478120885e-07, + "kl": 0.1474609375, + "learning_rate": 9.807651860936297e-05, + "loss": 0.0001, + "step": 514 + }, + { + "clip_ratio": 0.0, + "completion_length": 110.3214340209961, + "epoch": 0.7202797202797203, + "grad_norm": 1.620447571811333e-07, + "kl": 0.1630859375, + "learning_rate": 9.805971918361214e-05, + "loss": 0.0002, + "num_tokens": 165997.0, + "reward": 2.9055941104888916, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.3928571343421936, + "rewards/check_perplexity_diff_func": 2.0, + "rewards/check_winston_local_func": 0.5127367377281189, + "step": 515 + }, + { + "clip_ratio": 0.0, + "epoch": 0.7216783216783217, + "grad_norm": 1.472040658632873e-07, + "kl": 0.1630859375, + "learning_rate": 9.804284816597008e-05, + "loss": 0.0002, + "step": 516 + }, + { + "clip_ratio": 0.0, + "completion_length": 91.67857360839844, + "epoch": 0.7230769230769231, + "grad_norm": 2.4623250489986615e-06, + "kl": 0.142578125, + "learning_rate": 9.802590558156862e-05, + "loss": 0.0001, + "num_tokens": 184617.0, + "reward": 2.853912830352783, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.3928571343421936, + "rewards/check_perplexity_diff_func": 2.0, + "rewards/check_winston_local_func": 0.46105554699897766, + "step": 517 + }, + { + "clip_ratio": 0.0, + "epoch": 0.7244755244755244, + "grad_norm": 2.0559549189734124e-06, + "kl": 0.142578125, + "learning_rate": 9.800889145564617e-05, + "loss": 0.0001, + "step": 518 + }, + { + "clip_ratio": 0.0, + "completion_length": 127.21429443359375, + "epoch": 0.7258741258741259, + "grad_norm": 2.3530034340292295e-07, + "kl": 0.1552734375, + "learning_rate": 9.799180581354774e-05, + "loss": 0.0002, + "num_tokens": 207223.0, + "reward": 2.7364511489868164, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.3571428656578064, + "rewards/check_perplexity_diff_func": 1.9285714626312256, + "rewards/check_winston_local_func": 0.45073673129081726, + "step": 519 + }, + { + "clip_ratio": 0.0, + "epoch": 0.7272727272727273, + "grad_norm": 2.2322343800833895e-07, + "kl": 0.1552734375, + "learning_rate": 9.797464868072488e-05, + "loss": 0.0002, + "step": 520 + }, + { + "clip_ratio": 0.0, + "completion_length": 104.5714340209961, + "epoch": 0.7286713286713287, + "grad_norm": 2.968346809067501e-07, + "kl": 0.177734375, + "learning_rate": 9.795742008273558e-05, + "loss": 0.0002, + "num_tokens": 227059.0, + "reward": 2.6932995319366455, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.3571428656578064, + "rewards/check_perplexity_diff_func": 2.0, + "rewards/check_winston_local_func": 0.3361565172672272, + "step": 521 + }, + { + "clip_ratio": 0.0, + "epoch": 0.73006993006993, + "grad_norm": 2.852980195095417e-07, + "kl": 0.177734375, + "learning_rate": 9.794012004524434e-05, + "loss": 0.0002, + "step": 522 + }, + { + "clip_ratio": 0.0, + "completion_length": 105.42857360839844, + "epoch": 0.7314685314685314, + "grad_norm": 3.3796985539550523e-06, + "kl": 0.1640625, + "learning_rate": 9.792274859402205e-05, + "loss": 0.0002, + "num_tokens": 246949.0, + "reward": 2.6483211517333984, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.3214285671710968, + "rewards/check_perplexity_diff_func": 1.9285714626312256, + "rewards/check_winston_local_func": 0.3983212113380432, + "step": 523 + }, + { + "clip_ratio": 0.0, + "epoch": 0.7328671328671329, + "grad_norm": 3.1476330793325015e-06, + "kl": 0.1640625, + "learning_rate": 9.790530575494603e-05, + "loss": 0.0002, + "step": 524 + }, + { + "clip_ratio": 0.0, + "completion_length": 125.03572082519531, + "epoch": 0.7342657342657343, + "grad_norm": 0.0060337949589422705, + "kl": 0.16796875, + "learning_rate": 9.788779155399987e-05, + "loss": -0.0001, + "num_tokens": 269441.0, + "reward": 2.768662214279175, + "reward_std": 0.049707408994436264, + "rewards/check_gptzero_func": 0.3928571343421936, + "rewards/check_perplexity_diff_func": 1.9642857313156128, + "rewards/check_winston_local_func": 0.4115191400051117, + "step": 525 + }, + { + "clip_ratio": 0.0, + "epoch": 0.7356643356643356, + "grad_norm": 0.0068588182961818464, + "kl": 0.16796875, + "learning_rate": 9.787020601727352e-05, + "loss": -0.0, + "step": 526 + }, + { + "clip_ratio": 0.0, + "completion_length": 78.39286041259766, + "epoch": 0.737062937062937, + "grad_norm": 1.3153821115529096e-05, + "kl": 0.2353515625, + "learning_rate": 9.785254917096318e-05, + "loss": 0.0002, + "num_tokens": 286345.0, + "reward": 3.074557304382324, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.5, + "rewards/check_perplexity_diff_func": 2.0, + "rewards/check_winston_local_func": 0.5745573043823242, + "step": 527 + }, + { + "clip_ratio": 0.0, + "epoch": 0.7384615384615385, + "grad_norm": 1.256427796379324e-05, + "kl": 0.2353515625, + "learning_rate": 9.783482104137127e-05, + "loss": 0.0002, + "step": 528 + }, + { + "clip_ratio": 0.0, + "completion_length": 124.39286041259766, + "epoch": 0.7398601398601399, + "grad_norm": 1.3517164643713064e-05, + "kl": 0.12353515625, + "learning_rate": 9.781702165490639e-05, + "loss": 0.0001, + "num_tokens": 309195.0, + "reward": 2.4804210662841797, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.3928571343421936, + "rewards/check_perplexity_diff_func": 1.6428571939468384, + "rewards/check_winston_local_func": 0.44470664858818054, + "step": 529 + }, + { + "clip_ratio": 0.0, + "epoch": 0.7412587412587412, + "grad_norm": 1.3131407312669822e-05, + "kl": 0.12353515625, + "learning_rate": 9.779915103808328e-05, + "loss": 0.0001, + "step": 530 + }, + { + "clip_ratio": 0.0, + "completion_length": 111.53572082519531, + "epoch": 0.7426573426573426, + "grad_norm": 7.637416736333039e-07, + "kl": 0.15625, + "learning_rate": 9.778120921752285e-05, + "loss": 0.0002, + "num_tokens": 330505.0, + "reward": 3.070021390914917, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.5357142686843872, + "rewards/check_perplexity_diff_func": 2.0, + "rewards/check_winston_local_func": 0.5343068838119507, + "step": 531 + }, + { + "clip_ratio": 0.0, + "epoch": 0.7440559440559441, + "grad_norm": 7.316834780823122e-07, + "kl": 0.15625, + "learning_rate": 9.776319621995201e-05, + "loss": 0.0002, + "step": 532 + }, + { + "clip_ratio": 0.0, + "completion_length": 106.35714721679688, + "epoch": 0.7454545454545455, + "grad_norm": 3.663080215251399e-07, + "kl": 0.193359375, + "learning_rate": 9.77451120722037e-05, + "loss": 0.0002, + "num_tokens": 350813.0, + "reward": 2.63775634765625, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.3214285671710968, + "rewards/check_perplexity_diff_func": 1.9285714626312256, + "rewards/check_winston_local_func": 0.3877563774585724, + "step": 533 + }, + { + "clip_ratio": 0.0, + "epoch": 0.7468531468531469, + "grad_norm": 3.8678973306591443e-07, + "kl": 0.193359375, + "learning_rate": 9.77269568012169e-05, + "loss": 0.0002, + "step": 534 + }, + { + "clip_ratio": 0.0, + "completion_length": 106.67857360839844, + "epoch": 0.7482517482517482, + "grad_norm": 6.123173544504598e-06, + "kl": 0.2216796875, + "learning_rate": 9.770873043403648e-05, + "loss": 0.0002, + "num_tokens": 371585.0, + "reward": 2.874614953994751, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.4285714328289032, + "rewards/check_perplexity_diff_func": 1.9285714626312256, + "rewards/check_winston_local_func": 0.5174719095230103, + "step": 535 + }, + { + "clip_ratio": 0.0, + "epoch": 0.7496503496503496, + "grad_norm": 6.141198966103098e-06, + "kl": 0.2216796875, + "learning_rate": 9.769043299781327e-05, + "loss": 0.0002, + "step": 536 + }, + { + "clip_ratio": 0.0, + "completion_length": 104.71428680419922, + "epoch": 0.7510489510489511, + "grad_norm": 6.024782598667822e-07, + "kl": 0.212890625, + "learning_rate": 9.767206451980394e-05, + "loss": 0.0002, + "num_tokens": 391201.0, + "reward": 2.666369676589966, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.25, + "rewards/check_perplexity_diff_func": 2.0, + "rewards/check_winston_local_func": 0.4163695275783539, + "step": 537 + }, + { + "clip_ratio": 0.0, + "epoch": 0.7524475524475525, + "grad_norm": 5.651815000751003e-07, + "kl": 0.212890625, + "learning_rate": 9.765362502737097e-05, + "loss": 0.0002, + "step": 538 + }, + { + "clip_ratio": 0.0, + "completion_length": 107.21428680419922, + "epoch": 0.7538461538461538, + "grad_norm": 4.5359076717402366e-07, + "kl": 0.19140625, + "learning_rate": 9.763511454798268e-05, + "loss": 0.0002, + "num_tokens": 412087.0, + "reward": 2.660142660140991, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.25, + "rewards/check_perplexity_diff_func": 1.9285714626312256, + "rewards/check_winston_local_func": 0.4815710484981537, + "step": 539 + }, + { + "clip_ratio": 0.0, + "epoch": 0.7552447552447552, + "grad_norm": 7.72545386297213e-07, + "kl": 0.19140625, + "learning_rate": 9.761653310921307e-05, + "loss": 0.0002, + "step": 540 + }, + { + "clip_ratio": 0.0, + "completion_length": 133.85714721679688, + "epoch": 0.7566433566433567, + "grad_norm": 8.430058973039463e-08, + "kl": 0.146484375, + "learning_rate": 9.759788073874189e-05, + "loss": 0.0001, + "num_tokens": 435535.0, + "reward": 2.720244884490967, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.4285714328289032, + "rewards/check_perplexity_diff_func": 1.9285714626312256, + "rewards/check_winston_local_func": 0.3631021976470947, + "step": 541 + }, + { + "clip_ratio": 0.0, + "epoch": 0.7580419580419581, + "grad_norm": 8.327469845321351e-08, + "kl": 0.146484375, + "learning_rate": 9.757915746435453e-05, + "loss": 0.0001, + "step": 542 + }, + { + "clip_ratio": 0.0002040816325461492, + "completion_length": 102.89286041259766, + "epoch": 0.7594405594405594, + "grad_norm": 0.000356312443076848, + "kl": 0.203125, + "learning_rate": 9.756036331394202e-05, + "loss": 0.0002, + "num_tokens": 455137.0, + "reward": 2.6731011867523193, + "reward_std": 1.646135569899343e-05, + "rewards/check_gptzero_func": 0.3928571343421936, + "rewards/check_perplexity_diff_func": 1.9285714626312256, + "rewards/check_winston_local_func": 0.351672500371933, + "step": 543 + }, + { + "clip_ratio": 0.0, + "epoch": 0.7608391608391608, + "grad_norm": 0.00031516713281171804, + "kl": 0.203125, + "learning_rate": 9.754149831550098e-05, + "loss": 0.0002, + "step": 544 + }, + { + "clip_ratio": 0.0, + "completion_length": 89.35714721679688, + "epoch": 0.7622377622377622, + "grad_norm": 2.70311309473727e-07, + "kl": 0.2080078125, + "learning_rate": 9.752256249713351e-05, + "loss": 0.0002, + "num_tokens": 473297.0, + "reward": 2.869070291519165, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.3571428656578064, + "rewards/check_perplexity_diff_func": 2.0, + "rewards/check_winston_local_func": 0.5119272470474243, + "step": 545 + }, + { + "clip_ratio": 0.0, + "epoch": 0.7636363636363637, + "grad_norm": 2.830814999627113e-07, + "kl": 0.2080078125, + "learning_rate": 9.750355588704727e-05, + "loss": 0.0002, + "step": 546 + }, + { + "clip_ratio": 0.0, + "completion_length": 132.44644165039062, + "epoch": 0.765034965034965, + "grad_norm": 0.0009688545197921594, + "kl": 0.1806640625, + "learning_rate": 9.748447851355535e-05, + "loss": 0.0002, + "num_tokens": 496926.0, + "reward": 2.7388832569122314, + "reward_std": 0.0023702967446297407, + "rewards/check_gptzero_func": 0.5357142686843872, + "rewards/check_perplexity_diff_func": 1.7142857313156128, + "rewards/check_winston_local_func": 0.4888834059238434, + "step": 547 + }, + { + "clip_ratio": 0.0, + "epoch": 0.7664335664335664, + "grad_norm": 0.0016734864168510978, + "kl": 0.1806640625, + "learning_rate": 9.746533040507624e-05, + "loss": 0.0002, + "step": 548 + }, + { + "clip_ratio": 0.0, + "completion_length": 97.03572082519531, + "epoch": 0.7678321678321678, + "grad_norm": 1.9163803223436476e-06, + "kl": 0.177734375, + "learning_rate": 9.744611159013381e-05, + "loss": 0.0002, + "num_tokens": 516080.0, + "reward": 2.9889118671417236, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.5, + "rewards/check_perplexity_diff_func": 2.0, + "rewards/check_winston_local_func": 0.48891177773475647, + "step": 549 + }, + { + "clip_ratio": 0.0, + "epoch": 0.7692307692307693, + "grad_norm": 1.2761616262245509e-06, + "kl": 0.177734375, + "learning_rate": 9.742682209735727e-05, + "loss": 0.0002, + "step": 550 + }, + { + "clip_ratio": 0.0, + "completion_length": 92.39286041259766, + "epoch": 0.7706293706293706, + "grad_norm": 1.896094459987994e-07, + "kl": 0.1923828125, + "learning_rate": 9.740746195548112e-05, + "loss": 0.0002, + "num_tokens": 534554.0, + "reward": 2.943678855895996, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.4285714328289032, + "rewards/check_perplexity_diff_func": 2.0, + "rewards/check_winston_local_func": 0.5151072144508362, + "step": 551 + }, + { + "clip_ratio": 0.0, + "epoch": 0.772027972027972, + "grad_norm": 2.1137044966091938e-07, + "kl": 0.1923828125, + "learning_rate": 9.738803119334506e-05, + "loss": 0.0002, + "step": 552 + }, + { + "clip_ratio": 0.0, + "completion_length": 103.21428680419922, + "epoch": 0.7734265734265734, + "grad_norm": 4.705318347406964e-07, + "kl": 0.1826171875, + "learning_rate": 9.736852983989404e-05, + "loss": 0.0002, + "num_tokens": 554220.0, + "reward": 2.9958415031433105, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.5, + "rewards/check_perplexity_diff_func": 2.0, + "rewards/check_winston_local_func": 0.4958415627479553, + "step": 553 + }, + { + "clip_ratio": 0.0, + "epoch": 0.7748251748251749, + "grad_norm": 5.24917119837944e-07, + "kl": 0.1826171875, + "learning_rate": 9.734895792417811e-05, + "loss": 0.0002, + "step": 554 + }, + { + "clip_ratio": 0.0, + "completion_length": 102.53572082519531, + "epoch": 0.7762237762237763, + "grad_norm": 0.0005595837692618245, + "kl": 0.1611328125, + "learning_rate": 9.73293154753525e-05, + "loss": -0.0001, + "num_tokens": 574520.0, + "reward": 2.660358190536499, + "reward_std": 0.05061452463269234, + "rewards/check_gptzero_func": 0.2857142984867096, + "rewards/check_perplexity_diff_func": 1.75, + "rewards/check_winston_local_func": 0.6246438026428223, + "step": 555 + }, + { + "clip_ratio": 0.0, + "epoch": 0.7776223776223776, + "grad_norm": 0.0005476655329061088, + "kl": 0.1611328125, + "learning_rate": 9.730960252267743e-05, + "loss": -0.0001, + "step": 556 + }, + { + "clip_ratio": 0.0, + "completion_length": 110.28572082519531, + "epoch": 0.779020979020979, + "grad_norm": 1.9994241055330475e-07, + "kl": 0.162109375, + "learning_rate": 9.728981909551824e-05, + "loss": 0.0002, + "num_tokens": 595056.0, + "reward": 2.7659287452697754, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.3571428656578064, + "rewards/check_perplexity_diff_func": 2.0, + "rewards/check_winston_local_func": 0.40878555178642273, + "step": 557 + }, + { + "clip_ratio": 0.0, + "epoch": 0.7804195804195804, + "grad_norm": 2.102720872162732e-07, + "kl": 0.162109375, + "learning_rate": 9.726996522334516e-05, + "loss": 0.0002, + "step": 558 + }, + { + "clip_ratio": 0.0, + "completion_length": 116.3214340209961, + "epoch": 0.7818181818181819, + "grad_norm": 1.4418577967265551e-06, + "kl": 0.197265625, + "learning_rate": 9.725004093573342e-05, + "loss": 0.0002, + "num_tokens": 616218.0, + "reward": 2.90523099899292, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.4285714328289032, + "rewards/check_perplexity_diff_func": 2.0, + "rewards/check_winston_local_func": 0.4766596257686615, + "step": 559 + }, + { + "clip_ratio": 0.0, + "epoch": 0.7832167832167832, + "grad_norm": 1.4761501812795077e-06, + "kl": 0.197265625, + "learning_rate": 9.723004626236314e-05, + "loss": 0.0002, + "step": 560 + }, + { + "clip_ratio": 0.0, + "completion_length": 92.28572082519531, + "epoch": 0.7846153846153846, + "grad_norm": 6.640556191873379e-07, + "kl": 0.1904296875, + "learning_rate": 9.720998123301923e-05, + "loss": 0.0002, + "num_tokens": 634450.0, + "reward": 2.910466432571411, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.4642857015132904, + "rewards/check_perplexity_diff_func": 2.0, + "rewards/check_winston_local_func": 0.4461804926395416, + "step": 561 + }, + { + "clip_ratio": 0.0, + "epoch": 0.786013986013986, + "grad_norm": 6.785626739636823e-07, + "kl": 0.1904296875, + "learning_rate": 9.718984587759148e-05, + "loss": 0.0002, + "step": 562 + }, + { + "clip_ratio": 0.0, + "completion_length": 97.67857360839844, + "epoch": 0.7874125874125875, + "grad_norm": 0.0002364011375932314, + "kl": 0.1953125, + "learning_rate": 9.71696402260744e-05, + "loss": 0.0002, + "num_tokens": 653730.0, + "reward": 3.09153151512146, + "reward_std": 4.5663102355320007e-05, + "rewards/check_gptzero_func": 0.5714285969734192, + "rewards/check_perplexity_diff_func": 2.0, + "rewards/check_winston_local_func": 0.5201030373573303, + "step": 563 + }, + { + "clip_ratio": 0.00014228800137061626, + "epoch": 0.7888111888111888, + "grad_norm": 0.00020388090420570156, + "kl": 0.1953125, + "learning_rate": 9.714936430856723e-05, + "loss": 0.0002, + "step": 564 + }, + { + "clip_ratio": 0.0, + "completion_length": 120.66072082519531, + "epoch": 0.7902097902097902, + "grad_norm": 8.089432221384178e-05, + "kl": 0.1357421875, + "learning_rate": 9.712901815527386e-05, + "loss": 0.0001, + "num_tokens": 675657.0, + "reward": 2.674098253250122, + "reward_std": 0.0001108883589040488, + "rewards/check_gptzero_func": 0.3571428656578064, + "rewards/check_perplexity_diff_func": 1.9285714626312256, + "rewards/check_winston_local_func": 0.38838380575180054, + "step": 565 + }, + { + "clip_ratio": 0.0, + "epoch": 0.7916083916083916, + "grad_norm": 8.889658368646956e-05, + "kl": 0.1357421875, + "learning_rate": 9.710860179650287e-05, + "loss": 0.0001, + "step": 566 + }, + { + "clip_ratio": 0.0, + "completion_length": 107.3214340209961, + "epoch": 0.793006993006993, + "grad_norm": 0.0001037228042730343, + "kl": 0.17578125, + "learning_rate": 9.70881152626673e-05, + "loss": 0.0002, + "num_tokens": 695807.0, + "reward": 2.615325927734375, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.25, + "rewards/check_perplexity_diff_func": 1.9285714626312256, + "rewards/check_winston_local_func": 0.4367544949054718, + "step": 567 + }, + { + "clip_ratio": 0.0, + "epoch": 0.7944055944055944, + "grad_norm": 9.62410190276379e-05, + "kl": 0.17578125, + "learning_rate": 9.706755858428486e-05, + "loss": 0.0002, + "step": 568 + }, + { + "clip_ratio": 0.0, + "completion_length": 128.3928680419922, + "epoch": 0.7958041958041958, + "grad_norm": 3.2052372287941215e-05, + "kl": 0.12890625, + "learning_rate": 9.704693179197767e-05, + "loss": 0.0001, + "num_tokens": 719201.0, + "reward": 2.616921901702881, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.4285714328289032, + "rewards/check_perplexity_diff_func": 1.7857142686843872, + "rewards/check_winston_local_func": 0.40263620018959045, + "step": 569 + }, + { + "clip_ratio": 0.0, + "epoch": 0.7972027972027972, + "grad_norm": 3.1589569913845654e-05, + "kl": 0.12890625, + "learning_rate": 9.702623491647233e-05, + "loss": 0.0001, + "step": 570 + }, + { + "clip_ratio": 0.0, + "completion_length": 97.3214340209961, + "epoch": 0.7986013986013986, + "grad_norm": 2.966365002312992e-07, + "kl": 0.1572265625, + "learning_rate": 9.70054679885998e-05, + "loss": 0.0002, + "num_tokens": 738947.0, + "reward": 2.835333824157715, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.4285714328289032, + "rewards/check_perplexity_diff_func": 1.9285714626312256, + "rewards/check_winston_local_func": 0.47819074988365173, + "step": 571 + }, + { + "clip_ratio": 0.0, + "epoch": 0.8, + "grad_norm": 4.1573650424054196e-07, + "kl": 0.1572265625, + "learning_rate": 9.698463103929542e-05, + "loss": 0.0002, + "step": 572 + }, + { + "clip_ratio": 0.0, + "completion_length": 105.78572082519531, + "epoch": 0.8013986013986014, + "grad_norm": 0.000588551803293507, + "kl": 0.1611328125, + "learning_rate": 9.696372409959886e-05, + "loss": 0.0002, + "num_tokens": 759359.0, + "reward": 2.967468738555908, + "reward_std": 0.0020603849552571774, + "rewards/check_gptzero_func": 0.5, + "rewards/check_perplexity_diff_func": 1.9285714626312256, + "rewards/check_winston_local_func": 0.5388973355293274, + "step": 573 + }, + { + "clip_ratio": 0.0, + "epoch": 0.8027972027972028, + "grad_norm": 0.0006215594211925717, + "kl": 0.1611328125, + "learning_rate": 9.694274720065399e-05, + "loss": 0.0002, + "step": 574 + }, + { + "clip_ratio": 0.0, + "completion_length": 97.14286041259766, + "epoch": 0.8041958041958042, + "grad_norm": 3.3967306751168834e-06, + "kl": 0.17578125, + "learning_rate": 9.692170037370898e-05, + "loss": 0.0002, + "num_tokens": 778637.0, + "reward": 2.9381115436553955, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.4285714328289032, + "rewards/check_perplexity_diff_func": 2.0, + "rewards/check_winston_local_func": 0.509539783000946, + "step": 575 + }, + { + "clip_ratio": 0.0, + "epoch": 0.8055944055944056, + "grad_norm": 3.2036118908588705e-06, + "kl": 0.17578125, + "learning_rate": 9.690058365011607e-05, + "loss": 0.0002, + "step": 576 + }, + { + "clip_ratio": 0.0, + "completion_length": 102.64286041259766, + "epoch": 0.806993006993007, + "grad_norm": 8.125129903430204e-06, + "kl": 0.1630859375, + "learning_rate": 9.687939706133168e-05, + "loss": 0.0002, + "num_tokens": 798473.0, + "reward": 2.781018018722534, + "reward_std": 0.025253813713788986, + "rewards/check_gptzero_func": 0.4107142984867096, + "rewards/check_perplexity_diff_func": 1.9285714626312256, + "rewards/check_winston_local_func": 0.4417320787906647, + "step": 577 + }, + { + "clip_ratio": 0.0, + "epoch": 0.8083916083916084, + "grad_norm": 7.98774787404577e-06, + "kl": 0.1630859375, + "learning_rate": 9.685814063891631e-05, + "loss": 0.0002, + "step": 578 + }, + { + "clip_ratio": 0.0, + "completion_length": 126.85714721679688, + "epoch": 0.8097902097902098, + "grad_norm": 3.679599498892351e-07, + "kl": 0.138671875, + "learning_rate": 9.683681441453445e-05, + "loss": 0.0001, + "num_tokens": 821081.0, + "reward": 2.8629186153411865, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.4285714328289032, + "rewards/check_perplexity_diff_func": 2.0, + "rewards/check_winston_local_func": 0.43434715270996094, + "step": 579 + }, + { + "clip_ratio": 0.0, + "epoch": 0.8111888111888111, + "grad_norm": 3.93663211553172e-07, + "kl": 0.138671875, + "learning_rate": 9.681541841995461e-05, + "loss": 0.0001, + "step": 580 + }, + { + "clip_ratio": 0.0, + "completion_length": 106.00000762939453, + "epoch": 0.8125874125874126, + "grad_norm": 8.022029708518222e-07, + "kl": 0.162109375, + "learning_rate": 9.67939526870492e-05, + "loss": 0.0002, + "num_tokens": 841529.0, + "reward": 2.8752760887145996, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.3928571343421936, + "rewards/check_perplexity_diff_func": 1.9285714626312256, + "rewards/check_winston_local_func": 0.55384761095047, + "step": 581 + }, + { + "clip_ratio": 0.0, + "epoch": 0.813986013986014, + "grad_norm": 8.501682817147847e-07, + "kl": 0.162109375, + "learning_rate": 9.677241724779453e-05, + "loss": 0.0002, + "step": 582 + }, + { + "clip_ratio": 0.0, + "completion_length": 109.85714721679688, + "epoch": 0.8153846153846154, + "grad_norm": 1.7930012533014129e-06, + "kl": 0.1689453125, + "learning_rate": 9.675081213427076e-05, + "loss": 0.0002, + "num_tokens": 862213.0, + "reward": 2.9139628410339355, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.4285714328289032, + "rewards/check_perplexity_diff_func": 2.0, + "rewards/check_winston_local_func": 0.48539119958877563, + "step": 583 + }, + { + "clip_ratio": 0.0, + "epoch": 0.8167832167832167, + "grad_norm": 1.6519465345892093e-06, + "kl": 0.1689453125, + "learning_rate": 9.672913737866179e-05, + "loss": 0.0002, + "step": 584 + }, + { + "clip_ratio": 0.0, + "completion_length": 98.66072082519531, + "epoch": 0.8181818181818182, + "grad_norm": 0.002787236207669909, + "kl": 0.25, + "learning_rate": 9.670739301325534e-05, + "loss": 0.0002, + "num_tokens": 881842.0, + "reward": 2.6230576038360596, + "reward_std": 0.006216124631464481, + "rewards/check_gptzero_func": 0.3571428656578064, + "rewards/check_perplexity_diff_func": 1.7857142686843872, + "rewards/check_winston_local_func": 0.48020049929618835, + "step": 585 + }, + { + "clip_ratio": 0.0, + "epoch": 0.8195804195804196, + "grad_norm": 0.002885109231591189, + "kl": 0.2490234375, + "learning_rate": 9.668557907044276e-05, + "loss": 0.0002, + "step": 586 + }, + { + "clip_ratio": 0.0, + "completion_length": 79.14286041259766, + "epoch": 0.820979020979021, + "grad_norm": 4.462250819797621e-06, + "kl": 0.1923828125, + "learning_rate": 9.666369558271909e-05, + "loss": 0.0002, + "num_tokens": 898716.0, + "reward": 2.900649309158325, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.3928571343421936, + "rewards/check_perplexity_diff_func": 2.0, + "rewards/check_winston_local_func": 0.5077921152114868, + "step": 587 + }, + { + "clip_ratio": 0.0, + "epoch": 0.8223776223776224, + "grad_norm": 4.488075066816524e-06, + "kl": 0.1923828125, + "learning_rate": 9.66417425826829e-05, + "loss": 0.0002, + "step": 588 + }, + { + "clip_ratio": 0.0, + "completion_length": 112.03572082519531, + "epoch": 0.8237762237762237, + "grad_norm": 1.3821264354883407e-07, + "kl": 0.14453125, + "learning_rate": 9.661972010303641e-05, + "loss": 0.0001, + "num_tokens": 919636.0, + "reward": 2.7668612003326416, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.4642857015132904, + "rewards/check_perplexity_diff_func": 1.8571428060531616, + "rewards/check_winston_local_func": 0.4454323649406433, + "step": 589 + }, + { + "clip_ratio": 0.0, + "epoch": 0.8251748251748252, + "grad_norm": 1.4174011116816794e-07, + "kl": 0.14453125, + "learning_rate": 9.659762817658524e-05, + "loss": 0.0001, + "step": 590 + }, + { + "clip_ratio": 0.0, + "completion_length": 107.92857360839844, + "epoch": 0.8265734265734266, + "grad_norm": 8.773734381605369e-07, + "kl": 0.1572265625, + "learning_rate": 9.65754668362385e-05, + "loss": 0.0002, + "num_tokens": 940096.0, + "reward": 2.8289589881896973, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.4642857015132904, + "rewards/check_perplexity_diff_func": 1.9285714626312256, + "rewards/check_winston_local_func": 0.4361015856266022, + "step": 591 + }, + { + "clip_ratio": 0.0, + "epoch": 0.827972027972028, + "grad_norm": 1.0023469469060766e-06, + "kl": 0.1572265625, + "learning_rate": 9.655323611500875e-05, + "loss": 0.0002, + "step": 592 + }, + { + "clip_ratio": 0.0, + "completion_length": 113.85714721679688, + "epoch": 0.8293706293706293, + "grad_norm": 6.1541718025779055e-06, + "kl": 0.166015625, + "learning_rate": 9.653093604601183e-05, + "loss": 0.0002, + "num_tokens": 961340.0, + "reward": 2.9606730937957764, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.5357142686843872, + "rewards/check_perplexity_diff_func": 2.0, + "rewards/check_winston_local_func": 0.4249587059020996, + "step": 593 + }, + { + "clip_ratio": 0.0, + "epoch": 0.8307692307692308, + "grad_norm": 6.512813773208274e-06, + "kl": 0.166015625, + "learning_rate": 9.650856666246693e-05, + "loss": 0.0002, + "step": 594 + }, + { + "clip_ratio": 0.0, + "completion_length": 91.60714721679688, + "epoch": 0.8321678321678322, + "grad_norm": 3.2190340652283214e-06, + "kl": 0.1767578125, + "learning_rate": 9.648612799769644e-05, + "loss": 0.0002, + "num_tokens": 979916.0, + "reward": 3.0096070766448975, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.4285714328289032, + "rewards/check_perplexity_diff_func": 2.0, + "rewards/check_winston_local_func": 0.5810357332229614, + "step": 595 + }, + { + "clip_ratio": 0.0, + "epoch": 0.8335664335664336, + "grad_norm": 3.1137752081162388e-06, + "kl": 0.1767578125, + "learning_rate": 9.646362008512602e-05, + "loss": 0.0002, + "step": 596 + }, + { + "clip_ratio": 0.0, + "completion_length": 106.53572082519531, + "epoch": 0.8349650349650349, + "grad_norm": 6.812051267845749e-06, + "kl": 0.185546875, + "learning_rate": 9.644104295828447e-05, + "loss": 0.0002, + "num_tokens": 1000300.0, + "reward": 2.7739078998565674, + "reward_std": 0.0, + "rewards/check_gptzero_func": 0.4285714328289032, + "rewards/check_perplexity_diff_func": 1.9285714626312256, + "rewards/check_winston_local_func": 0.4167649447917938, + "step": 597 + }, + { + "clip_ratio": 0.0, + "epoch": 0.8363636363636363, + "grad_norm": 2.398683493432653e-06, + "kl": 0.185546875, + "learning_rate": 9.641839665080363e-05, + "loss": 0.0002, + "step": 598 + }, + { + "clip_ratio": 0.0, + "completion_length": 107.41072082519531, + "epoch": 0.8377622377622378, + "grad_norm": 0.0008785473557125962, + "kl": 0.1796875, + "learning_rate": 9.63956811964185e-05, + "loss": 0.0002, + "num_tokens": 1021105.0, + "reward": 2.91171932220459, + "reward_std": 0.00400555832311511, + "rewards/check_gptzero_func": 0.5357142686843872, + "rewards/check_perplexity_diff_func": 1.8571428060531616, + "rewards/check_winston_local_func": 0.518862247467041, + "step": 599 + }, + { + "clip_ratio": 0.0, + "epoch": 0.8391608391608392, + "grad_norm": 0.000892784560673435, + "kl": 0.1796875, + "learning_rate": 9.6372896628967e-05, + "loss": 0.0002, + "step": 600 + } + ], + "logging_steps": 1, + "max_steps": 2860, + "num_input_tokens_seen": 0, + "num_train_epochs": 4, + "save_steps": 100, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 0.0, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +}