diff --git "a/checkpoint_1300/trainer_state.json" "b/checkpoint_1300/trainer_state.json" new file mode 100644--- /dev/null +++ "b/checkpoint_1300/trainer_state.json" @@ -0,0 +1,6014 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.8684034736138945, + "eval_steps": 100, + "global_step": 1300, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "learning_rate": 1.111111111111111e-08, + "loss": 0.0176, + "step": 2, + "train/R_acc": 0.003125, + "train/R_penalty": -0.671875, + "train/R_reason": 0.1771926448030963, + "train/R_vocab": 0.27042410714285714 + }, + { + "learning_rate": 3.3333333333333334e-08, + "loss": 0.0255, + "step": 4, + "train/R_acc": 0.00625, + "train/R_penalty": -0.578125, + "train/R_reason": 0.20292794300576886, + "train/R_vocab": 0.13051835317460317 + }, + { + "learning_rate": 5.555555555555555e-08, + "loss": -0.0402, + "step": 6, + "train/R_acc": 0.003125, + "train/R_penalty": -0.65625, + "train/R_reason": 0.21681085047252255, + "train/R_vocab": 0.3141617063492063 + }, + { + "learning_rate": 7.777777777777778e-08, + "loss": -0.0512, + "step": 8, + "train/R_acc": 0.034375, + "train/R_penalty": -0.53125, + "train/R_reason": 0.22407962891811065, + "train/R_vocab": 0.2578125 + }, + { + "learning_rate": 1e-07, + "loss": 0.0653, + "step": 10, + "train/R_acc": 0.003125, + "train/R_penalty": -0.640625, + "train/R_reason": 0.24027254265023745, + "train/R_vocab": 0.290625 + }, + { + "learning_rate": 1.2222222222222222e-07, + "loss": 0.0747, + "step": 12, + "train/R_acc": 0.0, + "train/R_penalty": -0.59375, + "train/R_reason": 0.2086565261070774, + "train/R_vocab": 0.2722098214285714 + }, + { + "learning_rate": 1.4444444444444442e-07, + "loss": -0.0073, + "step": 14, + "train/R_acc": 0.034375, + "train/R_penalty": -0.78125, + "train/R_reason": 0.18557344633053446, + "train/R_vocab": 0.23597470238095236 + }, + { + "learning_rate": 1.6666666666666665e-07, + "loss": 0.0011, + "step": 16, + "train/R_acc": 0.00625, + "train/R_penalty": -0.5625, + "train/R_reason": 0.18460204443700487, + "train/R_vocab": 0.28095238095238095 + }, + { + "learning_rate": 1.8888888888888888e-07, + "loss": 0.0511, + "step": 18, + "train/R_acc": 0.003125, + "train/R_penalty": -0.6875, + "train/R_reason": 0.1691650605005834, + "train/R_vocab": 0.22633928571428572 + }, + { + "learning_rate": 2.111111111111111e-07, + "loss": -0.0247, + "step": 20, + "train/R_acc": 0.03125, + "train/R_penalty": -0.484375, + "train/R_reason": 0.23429908163761645, + "train/R_vocab": 0.3410714285714286 + }, + { + "learning_rate": 2.3333333333333333e-07, + "loss": 0.0088, + "step": 22, + "train/R_acc": 0.0, + "train/R_penalty": -0.65625, + "train/R_reason": 0.15088100811244767, + "train/R_vocab": 0.2056919642857143 + }, + { + "learning_rate": 2.5555555555555553e-07, + "loss": 0.0366, + "step": 24, + "train/R_acc": 0.0, + "train/R_penalty": -0.6875, + "train/R_reason": 0.2605457318222165, + "train/R_vocab": 0.29765625 + }, + { + "learning_rate": 2.7777777777777776e-07, + "loss": 0.0558, + "step": 26, + "train/R_acc": 0.009375000000000001, + "train/R_penalty": -0.5, + "train/R_reason": 0.24175666877540153, + "train/R_vocab": 0.26484375000000004 + }, + { + "learning_rate": 3e-07, + "loss": 0.0588, + "step": 28, + "train/R_acc": 0.0, + "train/R_penalty": -0.540625, + "train/R_reason": 0.2054192775694245, + "train/R_vocab": 0.2109747023809524 + }, + { + "learning_rate": 3.222222222222222e-07, + "loss": -0.1103, + "step": 30, + "train/R_acc": 0.003125, + "train/R_penalty": -0.59375, + "train/R_reason": 0.18972739597143928, + "train/R_vocab": 0.2044642857142857 + }, + { + "learning_rate": 3.4444444444444444e-07, + "loss": 0.0162, + "step": 32, + "train/R_acc": 0.003125, + "train/R_penalty": -0.609375, + "train/R_reason": 0.16386884044279187, + "train/R_vocab": 0.26328125 + }, + { + "learning_rate": 3.666666666666666e-07, + "loss": 0.0241, + "step": 34, + "train/R_acc": 0.065625, + "train/R_penalty": -0.6875, + "train/R_reason": 0.1991015810053379, + "train/R_vocab": 0.3072916666666667 + }, + { + "learning_rate": 3.888888888888889e-07, + "loss": 0.0114, + "step": 36, + "train/R_acc": 0.003125, + "train/R_penalty": -0.65625, + "train/R_reason": 0.1542880167619275, + "train/R_vocab": 0.15633680555555557 + }, + { + "learning_rate": 4.1111111111111107e-07, + "loss": 0.0418, + "step": 38, + "train/R_acc": 0.03125, + "train/R_penalty": -0.59375, + "train/R_reason": 0.21628990497316553, + "train/R_vocab": 0.32326388888888885 + }, + { + "learning_rate": 4.3333333333333335e-07, + "loss": -0.0085, + "step": 40, + "train/R_acc": 0.065625, + "train/R_penalty": -0.540625, + "train/R_reason": 0.16669781447815488, + "train/R_vocab": 0.21458333333333335 + }, + { + "learning_rate": 4.555555555555555e-07, + "loss": -0.0614, + "step": 42, + "train/R_acc": 0.003125, + "train/R_penalty": -0.671875, + "train/R_reason": 0.23833932777613548, + "train/R_vocab": 0.24285714285714288 + }, + { + "learning_rate": 4.777777777777778e-07, + "loss": 0.004, + "step": 44, + "train/R_acc": 0.003125, + "train/R_penalty": -0.578125, + "train/R_reason": 0.2297590496740362, + "train/R_vocab": 0.1607142857142857 + }, + { + "learning_rate": 5e-07, + "loss": 0.0491, + "step": 46, + "train/R_acc": 0.03125, + "train/R_penalty": -0.6875, + "train/R_reason": 0.2037377285506986, + "train/R_vocab": 0.2598958333333333 + }, + { + "learning_rate": 4.999976528920434e-07, + "loss": -0.083, + "step": 48, + "train/R_acc": 0.0, + "train/R_penalty": -0.671875, + "train/R_reason": 0.22468232799435314, + "train/R_vocab": 0.2745907738095238 + }, + { + "learning_rate": 4.999906116122447e-07, + "loss": -0.0036, + "step": 50, + "train/R_acc": 0.0, + "train/R_penalty": -0.640625, + "train/R_reason": 0.22593521940016348, + "train/R_vocab": 0.2805803571428572 + }, + { + "learning_rate": 4.999788762928172e-07, + "loss": 0.0107, + "step": 52, + "train/R_acc": 0.003125, + "train/R_penalty": -0.821875, + "train/R_reason": 0.20522814513215548, + "train/R_vocab": 0.14702380952380953 + }, + { + "learning_rate": 4.999624471541134e-07, + "loss": 0.029, + "step": 54, + "train/R_acc": 0.0, + "train/R_penalty": -0.78125, + "train/R_reason": 0.21737758529192291, + "train/R_vocab": 0.19330357142857144 + }, + { + "learning_rate": 4.99941324504621e-07, + "loss": -0.0246, + "step": 56, + "train/R_acc": 0.00625, + "train/R_penalty": -0.546875, + "train/R_reason": 0.2251848744560886, + "train/R_vocab": 0.28158482142857144 + }, + { + "learning_rate": 4.999155087409571e-07, + "loss": -0.0517, + "step": 58, + "train/R_acc": 0.03125, + "train/R_penalty": -0.571875, + "train/R_reason": 0.18310813181259872, + "train/R_vocab": 0.24776785714285715 + }, + { + "learning_rate": 4.998850003478606e-07, + "loss": 0.0598, + "step": 60, + "train/R_acc": 0.034375, + "train/R_penalty": -0.703125, + "train/R_reason": 0.2082140947052848, + "train/R_vocab": 0.26149553571428574 + }, + { + "learning_rate": 4.998497998981838e-07, + "loss": 0.0027, + "step": 62, + "train/R_acc": 0.0, + "train/R_penalty": -0.828125, + "train/R_reason": 0.2611822631590372, + "train/R_vocab": 0.39639136904761907 + }, + { + "learning_rate": 4.998099080528803e-07, + "loss": 0.0385, + "step": 64, + "train/R_acc": 0.0, + "train/R_penalty": -0.53125, + "train/R_reason": 0.2213369943749453, + "train/R_vocab": 0.3606026785714286 + }, + { + "learning_rate": 4.997653255609941e-07, + "loss": -0.0523, + "step": 66, + "train/R_acc": 0.065625, + "train/R_penalty": -0.578125, + "train/R_reason": 0.20860609604648056, + "train/R_vocab": 0.2251488095238095 + }, + { + "learning_rate": 4.997160532596447e-07, + "loss": 0.0352, + "step": 68, + "train/R_acc": 0.003125, + "train/R_penalty": -0.790625, + "train/R_reason": 0.19482757334666323, + "train/R_vocab": 0.23776041666666667 + }, + { + "learning_rate": 4.996620920740111e-07, + "loss": 0.0103, + "step": 70, + "train/R_acc": 0.03125, + "train/R_penalty": -0.671875, + "train/R_reason": 0.25304049429757086, + "train/R_vocab": 0.299516369047619 + }, + { + "learning_rate": 4.996034430173153e-07, + "loss": 0.0247, + "step": 72, + "train/R_acc": 0.003125, + "train/R_penalty": -0.58125, + "train/R_reason": 0.27724208008758466, + "train/R_vocab": 0.3654637896825397 + }, + { + "learning_rate": 4.995401071908025e-07, + "loss": -0.0689, + "step": 74, + "train/R_acc": 0.0, + "train/R_penalty": -0.546875, + "train/R_reason": 0.18864259893071442, + "train/R_vocab": 0.265625 + }, + { + "learning_rate": 4.99472085783721e-07, + "loss": 0.0669, + "step": 76, + "train/R_acc": 0.003125, + "train/R_penalty": -0.775, + "train/R_reason": 0.24872595893355823, + "train/R_vocab": 0.3293526785714286 + }, + { + "learning_rate": 4.993993800732995e-07, + "loss": -0.0369, + "step": 78, + "train/R_acc": 0.034375, + "train/R_penalty": -0.546875, + "train/R_reason": 0.26486544822313024, + "train/R_vocab": 0.3398065476190476 + }, + { + "learning_rate": 4.993219914247231e-07, + "loss": 0.0685, + "step": 80, + "train/R_acc": 0.03125, + "train/R_penalty": -0.75, + "train/R_reason": 0.1711866679882719, + "train/R_vocab": 0.2575892857142857 + }, + { + "learning_rate": 4.992399212911082e-07, + "loss": 0.0382, + "step": 82, + "train/R_acc": 0.037500000000000006, + "train/R_penalty": -0.640625, + "train/R_reason": 0.22934020748837214, + "train/R_vocab": 0.31875 + }, + { + "learning_rate": 4.99153171213474e-07, + "loss": 0.083, + "step": 84, + "train/R_acc": 0.003125, + "train/R_penalty": -0.703125, + "train/R_reason": 0.16303826493835916, + "train/R_vocab": 0.18002232142857144 + }, + { + "learning_rate": 4.990617428207153e-07, + "loss": 0.0433, + "step": 86, + "train/R_acc": 0.03125, + "train/R_penalty": -0.84375, + "train/R_reason": 0.21139857471142573, + "train/R_vocab": 0.28266369047619044 + }, + { + "learning_rate": 4.989656378295704e-07, + "loss": 0.0554, + "step": 88, + "train/R_acc": 0.00625, + "train/R_penalty": -0.5, + "train/R_reason": 0.2142568618009861, + "train/R_vocab": 0.3390625 + }, + { + "learning_rate": 4.988648580445895e-07, + "loss": 0.0648, + "step": 90, + "train/R_acc": 0.003125, + "train/R_penalty": -0.65625, + "train/R_reason": 0.17382571925607848, + "train/R_vocab": 0.1894345238095238 + }, + { + "learning_rate": 4.987594053581012e-07, + "loss": -0.0241, + "step": 92, + "train/R_acc": 0.003125, + "train/R_penalty": -0.703125, + "train/R_reason": 0.21781325533595836, + "train/R_vocab": 0.33205605158730156 + }, + { + "learning_rate": 4.98649281750176e-07, + "loss": 0.1022, + "step": 94, + "train/R_acc": 0.003125, + "train/R_penalty": -0.703125, + "train/R_reason": 0.1999237909283059, + "train/R_vocab": 0.32064732142857144 + }, + { + "learning_rate": 4.985344892885899e-07, + "loss": 0.0821, + "step": 96, + "train/R_acc": 0.00625, + "train/R_penalty": -0.578125, + "train/R_reason": 0.23594255343132425, + "train/R_vocab": 0.30199652777777775 + }, + { + "learning_rate": 4.984150301287853e-07, + "loss": 0.0096, + "step": 98, + "train/R_acc": 0.0, + "train/R_penalty": -0.80625, + "train/R_reason": 0.1900727268525148, + "train/R_vocab": 0.20625000000000002 + }, + { + "learning_rate": 4.982909065138305e-07, + "loss": 0.0247, + "step": 100, + "train/R_acc": 0.003125, + "train/R_penalty": -0.6875, + "train/R_reason": 0.21411060850160504, + "train/R_vocab": 0.22946428571428573 + }, + { + "eval/R_acc": 0.002777777777777778, + "eval/R_penalty": -0.6847222222222222, + "eval/R_reason": 0.2572056199703138, + "eval/R_vocab": 0.3255787037037037, + "step": 100 + }, + { + "step": 100 + }, + { + "learning_rate": 4.981621207743778e-07, + "loss": -0.0004, + "step": 102, + "train/R_acc": 0.0, + "train/R_penalty": -0.765625, + "train/R_reason": 0.1735908404636029, + "train/R_vocab": 0.15747767857142858 + }, + { + "learning_rate": 4.980286753286194e-07, + "loss": 0.0155, + "step": 104, + "train/R_acc": 0.003125, + "train/R_penalty": -0.578125, + "train/R_reason": 0.2016416402158515, + "train/R_vocab": 0.3698660714285714 + }, + { + "learning_rate": 4.978905726822423e-07, + "loss": -0.0257, + "step": 106, + "train/R_acc": 0.003125, + "train/R_penalty": -0.640625, + "train/R_reason": 0.2019264344648219, + "train/R_vocab": 0.2499627976190476 + }, + { + "learning_rate": 4.97747815428381e-07, + "loss": 0.0757, + "step": 108, + "train/R_acc": 0.0, + "train/R_penalty": -0.625, + "train/R_reason": 0.2609304111036691, + "train/R_vocab": 0.22472718253968255 + }, + { + "learning_rate": 4.976004062475691e-07, + "loss": -0.0376, + "step": 110, + "train/R_acc": 0.03125, + "train/R_penalty": -0.55625, + "train/R_reason": 0.24456802402586483, + "train/R_vocab": 0.30167410714285714 + }, + { + "learning_rate": 4.974483479076885e-07, + "loss": -0.0155, + "step": 112, + "train/R_acc": 0.0, + "train/R_penalty": -0.734375, + "train/R_reason": 0.15732482982112658, + "train/R_vocab": 0.20078125 + }, + { + "learning_rate": 4.972916432639181e-07, + "loss": 0.0107, + "step": 114, + "train/R_acc": 0.0625, + "train/R_penalty": -0.6125, + "train/R_reason": 0.19719755508097864, + "train/R_vocab": 0.18292410714285715 + }, + { + "learning_rate": 4.971302952586796e-07, + "loss": -0.0536, + "step": 116, + "train/R_acc": 0.003125, + "train/R_penalty": -0.6875, + "train/R_reason": 0.1615191195936923, + "train/R_vocab": 0.20714285714285713 + }, + { + "learning_rate": 4.969643069215823e-07, + "loss": 0.03, + "step": 118, + "train/R_acc": 0.003125, + "train/R_penalty": -0.703125, + "train/R_reason": 0.2057090874466078, + "train/R_vocab": 0.26006944444444446 + }, + { + "learning_rate": 4.967936813693668e-07, + "loss": 0.0669, + "step": 120, + "train/R_acc": 0.0, + "train/R_penalty": -0.65625, + "train/R_reason": 0.17123245123013492, + "train/R_vocab": 0.22881944444444446 + }, + { + "learning_rate": 4.966184218058457e-07, + "loss": -0.01, + "step": 122, + "train/R_acc": 0.065625, + "train/R_penalty": -0.640625, + "train/R_reason": 0.25008522769707353, + "train/R_vocab": 0.2768725198412698 + }, + { + "learning_rate": 4.96438531521844e-07, + "loss": 0.0227, + "step": 124, + "train/R_acc": 0.0, + "train/R_penalty": -0.5625, + "train/R_reason": 0.228776542397296, + "train/R_vocab": 0.19776785714285716 + }, + { + "learning_rate": 4.962540138951371e-07, + "loss": 0.0342, + "step": 126, + "train/R_acc": 0.0, + "train/R_penalty": -0.578125, + "train/R_reason": 0.20827985353158485, + "train/R_vocab": 0.3024440836940837 + }, + { + "learning_rate": 4.960648723903872e-07, + "loss": 0.0418, + "step": 128, + "train/R_acc": 0.0, + "train/R_penalty": -0.671875, + "train/R_reason": 0.19681757054800564, + "train/R_vocab": 0.22455357142857144 + }, + { + "learning_rate": 4.958711105590785e-07, + "loss": -0.0196, + "step": 130, + "train/R_acc": 0.009375000000000001, + "train/R_penalty": -0.5625, + "train/R_reason": 0.22729280568853277, + "train/R_vocab": 0.19475446428571427 + }, + { + "learning_rate": 4.956727320394506e-07, + "loss": 0.0963, + "step": 132, + "train/R_acc": 0.003125, + "train/R_penalty": -0.84375, + "train/R_reason": 0.20146228335738753, + "train/R_vocab": 0.24609995039682542 + }, + { + "learning_rate": 4.954697405564299e-07, + "loss": 0.0259, + "step": 134, + "train/R_acc": 0.009375000000000001, + "train/R_penalty": -0.625, + "train/R_reason": 0.228208707691525, + "train/R_vocab": 0.23055217352092353 + }, + { + "learning_rate": 4.952621399215597e-07, + "loss": 0.0781, + "step": 136, + "train/R_acc": 0.003125, + "train/R_penalty": -0.8125, + "train/R_reason": 0.1811115376120021, + "train/R_vocab": 0.2921875 + }, + { + "learning_rate": 4.95049934032929e-07, + "loss": 0.0659, + "step": 138, + "train/R_acc": 0.0, + "train/R_penalty": -0.640625, + "train/R_reason": 0.2199008430371848, + "train/R_vocab": 0.22447916666666667 + }, + { + "learning_rate": 4.948331268750988e-07, + "loss": -0.0087, + "step": 140, + "train/R_acc": 0.03125, + "train/R_penalty": -0.796875, + "train/R_reason": 0.16414720324183496, + "train/R_vocab": 0.20345982142857144 + }, + { + "learning_rate": 4.946117225190273e-07, + "loss": 0.0109, + "step": 142, + "train/R_acc": 0.03125, + "train/R_penalty": -0.634375, + "train/R_reason": 0.20413731109417932, + "train/R_vocab": 0.27366071428571426 + }, + { + "learning_rate": 4.943857251219942e-07, + "loss": -0.0291, + "step": 144, + "train/R_acc": 0.0, + "train/R_penalty": -0.703125, + "train/R_reason": 0.2114743293926566, + "train/R_vocab": 0.24921875 + }, + { + "learning_rate": 4.941551389275217e-07, + "loss": -0.0249, + "step": 146, + "train/R_acc": 0.0, + "train/R_penalty": -0.609375, + "train/R_reason": 0.19512092302735518, + "train/R_vocab": 0.17879464285714286 + }, + { + "learning_rate": 4.939199682652952e-07, + "loss": -0.03, + "step": 148, + "train/R_acc": 0.003125, + "train/R_penalty": -0.6125, + "train/R_reason": 0.19263372463344114, + "train/R_vocab": 0.2714285714285714 + }, + { + "learning_rate": 4.936802175510823e-07, + "loss": -0.1025, + "step": 150, + "train/R_acc": 0.00625, + "train/R_penalty": -0.665625, + "train/R_reason": 0.21358490777949846, + "train/R_vocab": 0.3302455357142857 + }, + { + "learning_rate": 4.934358912866497e-07, + "loss": 0.0293, + "step": 152, + "train/R_acc": 0.00625, + "train/R_penalty": -0.484375, + "train/R_reason": 0.16662975968037924, + "train/R_vocab": 0.1699280753968254 + }, + { + "learning_rate": 4.931869940596779e-07, + "loss": 0.0662, + "step": 154, + "train/R_acc": 0.00625, + "train/R_penalty": -0.59375, + "train/R_reason": 0.20091746530920346, + "train/R_vocab": 0.27994791666666663 + }, + { + "learning_rate": 4.929335305436764e-07, + "loss": 0.0355, + "step": 156, + "train/R_acc": 0.003125, + "train/R_penalty": -0.6875, + "train/R_reason": 0.22218261967350866, + "train/R_vocab": 0.3209821428571429 + }, + { + "learning_rate": 4.926755054978951e-07, + "loss": 0.0457, + "step": 158, + "train/R_acc": 0.0, + "train/R_penalty": -0.703125, + "train/R_reason": 0.22181716803364154, + "train/R_vocab": 0.29573863636363634 + }, + { + "learning_rate": 4.92412923767235e-07, + "loss": -0.0201, + "step": 160, + "train/R_acc": 0.00625, + "train/R_penalty": -0.625, + "train/R_reason": 0.24103408318415312, + "train/R_vocab": 0.23645833333333333 + }, + { + "learning_rate": 4.921457902821578e-07, + "loss": -0.0013, + "step": 162, + "train/R_acc": 0.0, + "train/R_penalty": -0.625, + "train/R_reason": 0.21205287864979985, + "train/R_vocab": 0.24444444444444446 + }, + { + "learning_rate": 4.918741100585921e-07, + "loss": -0.0549, + "step": 164, + "train/R_acc": 0.0, + "train/R_penalty": -0.75, + "train/R_reason": 0.21139952333180925, + "train/R_vocab": 0.36577380952380956 + }, + { + "learning_rate": 4.915978881978406e-07, + "loss": 0.0116, + "step": 166, + "train/R_acc": 0.003125, + "train/R_penalty": -0.671875, + "train/R_reason": 0.21990626286225867, + "train/R_vocab": 0.3359623015873016 + }, + { + "learning_rate": 4.913171298864836e-07, + "loss": -0.0194, + "step": 168, + "train/R_acc": 0.0, + "train/R_penalty": -0.6875, + "train/R_reason": 0.19859495503742927, + "train/R_vocab": 0.3138020833333333 + }, + { + "learning_rate": 4.910318403962813e-07, + "loss": 0.0306, + "step": 170, + "train/R_acc": 0.0, + "train/R_penalty": -0.75, + "train/R_reason": 0.15012936918136355, + "train/R_vocab": 0.2549107142857143 + }, + { + "learning_rate": 4.907420250840761e-07, + "loss": 0.0817, + "step": 172, + "train/R_acc": 0.03125, + "train/R_penalty": -0.68125, + "train/R_reason": 0.20995580823960475, + "train/R_vocab": 0.20602678571428573 + }, + { + "learning_rate": 4.9044768939169e-07, + "loss": -0.104, + "step": 174, + "train/R_acc": 0.03125, + "train/R_penalty": -0.625, + "train/R_reason": 0.18253049494341367, + "train/R_vocab": 0.20379464285714285 + }, + { + "learning_rate": 4.901488388458247e-07, + "loss": 0.0418, + "step": 176, + "train/R_acc": 0.0625, + "train/R_penalty": -0.5625, + "train/R_reason": 0.2305279249641788, + "train/R_vocab": 0.38442460317460314 + }, + { + "learning_rate": 4.898454790579558e-07, + "loss": -0.0033, + "step": 178, + "train/R_acc": 0.0, + "train/R_penalty": -0.671875, + "train/R_reason": 0.18641040374410697, + "train/R_vocab": 0.25188492063492063 + }, + { + "learning_rate": 4.895376157242288e-07, + "loss": 0.0036, + "step": 180, + "train/R_acc": 0.003125, + "train/R_penalty": -0.634375, + "train/R_reason": 0.19709747248451465, + "train/R_vocab": 0.28185763888888893 + }, + { + "learning_rate": 4.892252546253514e-07, + "loss": -0.0037, + "step": 182, + "train/R_acc": 0.0, + "train/R_penalty": -0.546875, + "train/R_reason": 0.2449271246503006, + "train/R_vocab": 0.30446428571428574 + }, + { + "learning_rate": 4.889084016264858e-07, + "loss": 0.0279, + "step": 184, + "train/R_acc": 0.0, + "train/R_penalty": -0.671875, + "train/R_reason": 0.24078946289847405, + "train/R_vocab": 0.3054935515873016 + }, + { + "learning_rate": 4.88587062677137e-07, + "loss": 0.0824, + "step": 186, + "train/R_acc": 0.00625, + "train/R_penalty": -0.625, + "train/R_reason": 0.16765366722116085, + "train/R_vocab": 0.22786458333333331 + }, + { + "learning_rate": 4.882612438110429e-07, + "loss": 0.0193, + "step": 188, + "train/R_acc": 0.003125, + "train/R_penalty": -0.71875, + "train/R_reason": 0.21543693992639024, + "train/R_vocab": 0.3402901785714285 + }, + { + "learning_rate": 4.8793095114606e-07, + "loss": -0.0257, + "step": 190, + "train/R_acc": 0.003125, + "train/R_penalty": -0.625, + "train/R_reason": 0.154932479362758, + "train/R_vocab": 0.2091393849206349 + }, + { + "learning_rate": 4.875961908840485e-07, + "loss": -0.0201, + "step": 192, + "train/R_acc": 0.003125, + "train/R_penalty": -0.640625, + "train/R_reason": 0.20005776839850115, + "train/R_vocab": 0.26316964285714284 + }, + { + "learning_rate": 4.872569693107563e-07, + "loss": 0.0163, + "step": 194, + "train/R_acc": 0.0, + "train/R_penalty": -0.671875, + "train/R_reason": 0.2028767825912614, + "train/R_vocab": 0.22935267857142855 + }, + { + "learning_rate": 4.869132927957006e-07, + "loss": -0.0312, + "step": 196, + "train/R_acc": 0.0625, + "train/R_penalty": -0.6875, + "train/R_reason": 0.20598954317804136, + "train/R_vocab": 0.2894097222222222 + }, + { + "learning_rate": 4.865651677920483e-07, + "loss": -0.041, + "step": 198, + "train/R_acc": 0.003125, + "train/R_penalty": -0.5625, + "train/R_reason": 0.19447689115700523, + "train/R_vocab": 0.21484375 + }, + { + "learning_rate": 4.862126008364954e-07, + "loss": -0.0021, + "step": 200, + "train/R_acc": 0.003125, + "train/R_penalty": -0.578125, + "train/R_reason": 0.24708335127072298, + "train/R_vocab": 0.3235491071428571 + }, + { + "eval/R_acc": 0.001388888888888889, + "eval/R_penalty": -0.6875, + "eval/R_reason": 0.26545862362041284, + "eval/R_vocab": 0.3180500440917107, + "step": 200 + }, + { + "step": 200 + }, + { + "learning_rate": 4.858555985491434e-07, + "loss": 0.0539, + "step": 202, + "train/R_acc": 0.003125, + "train/R_penalty": -0.71875, + "train/R_reason": 0.12246968403633852, + "train/R_vocab": 0.12857142857142856 + }, + { + "learning_rate": 4.854941676333756e-07, + "loss": -0.0105, + "step": 204, + "train/R_acc": 0.009375000000000001, + "train/R_penalty": -0.59375, + "train/R_reason": 0.20647794674779765, + "train/R_vocab": 0.2294642857142857 + }, + { + "learning_rate": 4.85128314875731e-07, + "loss": 0.0418, + "step": 206, + "train/R_acc": 0.0, + "train/R_penalty": -0.65625, + "train/R_reason": 0.22341563734148095, + "train/R_vocab": 0.24494047619047618 + }, + { + "learning_rate": 4.84758047145777e-07, + "loss": 0.0338, + "step": 208, + "train/R_acc": 0.0, + "train/R_penalty": -0.55625, + "train/R_reason": 0.21091575917752947, + "train/R_vocab": 0.25234375 + }, + { + "learning_rate": 4.843833713959802e-07, + "loss": 0.0127, + "step": 210, + "train/R_acc": 0.003125, + "train/R_penalty": -0.78125, + "train/R_reason": 0.15006747203000467, + "train/R_vocab": 0.23049355158730156 + }, + { + "learning_rate": 4.840042946615761e-07, + "loss": 0.0429, + "step": 212, + "train/R_acc": 0.003125, + "train/R_penalty": -0.59375, + "train/R_reason": 0.23417842297662064, + "train/R_vocab": 0.28020833333333334 + }, + { + "learning_rate": 4.836208240604368e-07, + "loss": -0.0, + "step": 214, + "train/R_acc": 0.0, + "train/R_penalty": -0.65, + "train/R_reason": 0.20300438739011717, + "train/R_vocab": 0.30078125 + }, + { + "learning_rate": 4.832329667929376e-07, + "loss": 0.0223, + "step": 216, + "train/R_acc": 0.0, + "train/R_penalty": -0.696875, + "train/R_reason": 0.21329021924994432, + "train/R_vocab": 0.2752604166666667 + }, + { + "learning_rate": 4.828407301418217e-07, + "loss": -0.0077, + "step": 218, + "train/R_acc": 0.0, + "train/R_penalty": -0.75, + "train/R_reason": 0.22684321233993932, + "train/R_vocab": 0.2927331349206349 + }, + { + "learning_rate": 4.824441214720628e-07, + "loss": 0.0196, + "step": 220, + "train/R_acc": 0.003125, + "train/R_penalty": -0.68125, + "train/R_reason": 0.22135178051709037, + "train/R_vocab": 0.3354166666666667 + }, + { + "learning_rate": 4.820431482307281e-07, + "loss": 0.0335, + "step": 222, + "train/R_acc": 0.034375, + "train/R_penalty": -0.609375, + "train/R_reason": 0.23161254238918147, + "train/R_vocab": 0.3784598214285714 + }, + { + "learning_rate": 4.816378179468374e-07, + "loss": 0.0416, + "step": 224, + "train/R_acc": 0.003125, + "train/R_penalty": -0.728125, + "train/R_reason": 0.20212408310335858, + "train/R_vocab": 0.2705357142857143 + }, + { + "learning_rate": 4.812281382312223e-07, + "loss": 0.0156, + "step": 226, + "train/R_acc": 0.03125, + "train/R_penalty": -0.640625, + "train/R_reason": 0.19926301994985757, + "train/R_vocab": 0.35390625000000003 + }, + { + "learning_rate": 4.808141167763826e-07, + "loss": 0.0111, + "step": 228, + "train/R_acc": 0.003125, + "train/R_penalty": -0.484375, + "train/R_reason": 0.24773363470223553, + "train/R_vocab": 0.3498015873015873 + }, + { + "learning_rate": 4.803957613563431e-07, + "loss": 0.1022, + "step": 230, + "train/R_acc": 0.00625, + "train/R_penalty": -0.609375, + "train/R_reason": 0.19832487010930644, + "train/R_vocab": 0.2538690476190476 + }, + { + "learning_rate": 4.799730798265063e-07, + "loss": 0.1914, + "step": 232, + "train/R_acc": 0.003125, + "train/R_penalty": -0.609375, + "train/R_reason": 0.22378486607186504, + "train/R_vocab": 0.28828125000000004 + }, + { + "learning_rate": 4.795460801235058e-07, + "loss": -0.036, + "step": 234, + "train/R_acc": 0.0, + "train/R_penalty": -0.671875, + "train/R_reason": 0.17550109595621496, + "train/R_vocab": 0.24095982142857145 + }, + { + "learning_rate": 4.791147702650565e-07, + "loss": 0.1009, + "step": 236, + "train/R_acc": 0.034375, + "train/R_penalty": -0.61875, + "train/R_reason": 0.2616533979112148, + "train/R_vocab": 0.3763888888888889 + }, + { + "learning_rate": 4.786791583498051e-07, + "loss": -0.0144, + "step": 238, + "train/R_acc": 0.0, + "train/R_penalty": -0.671875, + "train/R_reason": 0.20754413577165548, + "train/R_vocab": 0.22721974206349208 + }, + { + "learning_rate": 4.78239252557177e-07, + "loss": 0.0564, + "step": 240, + "train/R_acc": 0.0, + "train/R_penalty": -0.59375, + "train/R_reason": 0.2701327994702766, + "train/R_vocab": 0.3259424603174603 + }, + { + "learning_rate": 4.777950611472233e-07, + "loss": -0.0099, + "step": 242, + "train/R_acc": 0.0, + "train/R_penalty": -0.703125, + "train/R_reason": 0.21945774918879585, + "train/R_vocab": 0.2623883928571429 + }, + { + "learning_rate": 4.773465924604656e-07, + "loss": 0.0141, + "step": 244, + "train/R_acc": 0.065625, + "train/R_penalty": -0.671875, + "train/R_reason": 0.2201195420166109, + "train/R_vocab": 0.26729910714285715 + }, + { + "learning_rate": 4.768938549177392e-07, + "loss": 0.052, + "step": 246, + "train/R_acc": 0.0, + "train/R_penalty": -0.6875, + "train/R_reason": 0.2231016124642528, + "train/R_vocab": 0.26313244047619044 + }, + { + "learning_rate": 4.764368570200353e-07, + "loss": 0.0239, + "step": 248, + "train/R_acc": 0.003125, + "train/R_penalty": -0.5625, + "train/R_reason": 0.12548248807792622, + "train/R_vocab": 0.15256696428571428 + }, + { + "learning_rate": 4.759756073483411e-07, + "loss": 0.0264, + "step": 250, + "train/R_acc": 0.0, + "train/R_penalty": -0.765625, + "train/R_reason": 0.17495784675781162, + "train/R_vocab": 0.13359375 + }, + { + "learning_rate": 4.7551011456347876e-07, + "loss": 0.0271, + "step": 252, + "train/R_acc": 0.065625, + "train/R_penalty": -0.759375, + "train/R_reason": 0.2455444828427244, + "train/R_vocab": 0.3765625 + }, + { + "learning_rate": 4.750403874059428e-07, + "loss": 0.035, + "step": 254, + "train/R_acc": 0.003125, + "train/R_penalty": -0.61875, + "train/R_reason": 0.2624393029582254, + "train/R_vocab": 0.3123883928571428 + }, + { + "learning_rate": 4.745664346957361e-07, + "loss": -0.0167, + "step": 256, + "train/R_acc": 0.0, + "train/R_penalty": -0.734375, + "train/R_reason": 0.2061035486282405, + "train/R_vocab": 0.2132688492063492 + }, + { + "learning_rate": 4.740882653322039e-07, + "loss": 0.0071, + "step": 258, + "train/R_acc": 0.0, + "train/R_penalty": -0.609375, + "train/R_reason": 0.25927699094783563, + "train/R_vocab": 0.3599330357142857 + }, + { + "learning_rate": 4.7360588829386736e-07, + "loss": -0.0231, + "step": 260, + "train/R_acc": 0.0, + "train/R_penalty": -0.7125, + "train/R_reason": 0.2040610708874982, + "train/R_vocab": 0.23035714285714287 + }, + { + "learning_rate": 4.7311931263825434e-07, + "loss": 0.0593, + "step": 262, + "train/R_acc": 0.0, + "train/R_penalty": -0.65625, + "train/R_reason": 0.14067559915832142, + "train/R_vocab": 0.171875 + }, + { + "learning_rate": 4.726285475017294e-07, + "loss": 0.0213, + "step": 264, + "train/R_acc": 0.003125, + "train/R_penalty": -0.640625, + "train/R_reason": 0.2728806346759708, + "train/R_vocab": 0.34206349206349207 + }, + { + "learning_rate": 4.721336020993228e-07, + "loss": -0.0176, + "step": 266, + "train/R_acc": 0.003125, + "train/R_penalty": -0.6875, + "train/R_reason": 0.22780919072556438, + "train/R_vocab": 0.26149553571428574 + }, + { + "learning_rate": 4.716344857245567e-07, + "loss": 0.0807, + "step": 268, + "train/R_acc": 0.0, + "train/R_penalty": -0.61875, + "train/R_reason": 0.23184996778054867, + "train/R_vocab": 0.27254464285714286 + }, + { + "learning_rate": 4.7113120774927136e-07, + "loss": 0.0114, + "step": 270, + "train/R_acc": 0.0, + "train/R_penalty": -0.640625, + "train/R_reason": 0.24027473582911105, + "train/R_vocab": 0.41076388888888893 + }, + { + "learning_rate": 4.706237776234486e-07, + "loss": 0.0187, + "step": 272, + "train/R_acc": 0.03125, + "train/R_penalty": -0.609375, + "train/R_reason": 0.1528155999223981, + "train/R_vocab": 0.1519097222222222 + }, + { + "learning_rate": 4.7011220487503476e-07, + "loss": 0.0688, + "step": 274, + "train/R_acc": 0.034375, + "train/R_penalty": -0.609375, + "train/R_reason": 0.15041989042686232, + "train/R_vocab": 0.1591889880952381 + }, + { + "learning_rate": 4.695964991097616e-07, + "loss": 0.0498, + "step": 276, + "train/R_acc": 0.0, + "train/R_penalty": -0.578125, + "train/R_reason": 0.20384797152303086, + "train/R_vocab": 0.35360863095238093 + }, + { + "learning_rate": 4.6907667001096585e-07, + "loss": -0.0992, + "step": 278, + "train/R_acc": 0.00625, + "train/R_penalty": -0.578125, + "train/R_reason": 0.23776322399396135, + "train/R_vocab": 0.4219246031746032 + }, + { + "learning_rate": 4.685527273394078e-07, + "loss": -0.0568, + "step": 280, + "train/R_acc": 0.003125, + "train/R_penalty": -0.5625, + "train/R_reason": 0.2305205389308027, + "train/R_vocab": 0.21640624999999997 + }, + { + "learning_rate": 4.680246809330874e-07, + "loss": -0.0174, + "step": 282, + "train/R_acc": 0.0, + "train/R_penalty": -0.640625, + "train/R_reason": 0.27570159502906066, + "train/R_vocab": 0.23663194444444444 + }, + { + "learning_rate": 4.6749254070706013e-07, + "loss": -0.053, + "step": 284, + "train/R_acc": 0.0, + "train/R_penalty": -0.65625, + "train/R_reason": 0.17788848371405203, + "train/R_vocab": 0.159375 + }, + { + "learning_rate": 4.669563166532503e-07, + "loss": 0.0595, + "step": 286, + "train/R_acc": 0.03125, + "train/R_penalty": -0.6875, + "train/R_reason": 0.22202849960365972, + "train/R_vocab": 0.2922619047619048 + }, + { + "learning_rate": 4.6641601884026407e-07, + "loss": -0.0247, + "step": 288, + "train/R_acc": 0.003125, + "train/R_penalty": -0.609375, + "train/R_reason": 0.20278024644809003, + "train/R_vocab": 0.2072544642857143 + }, + { + "learning_rate": 4.6587165741319967e-07, + "loss": 0.0055, + "step": 290, + "train/R_acc": 0.0, + "train/R_penalty": -0.71875, + "train/R_reason": 0.20358950838718215, + "train/R_vocab": 0.30781250000000004 + }, + { + "learning_rate": 4.6532324259345743e-07, + "loss": 0.013, + "step": 292, + "train/R_acc": 0.0, + "train/R_penalty": -0.68125, + "train/R_reason": 0.19116956517383865, + "train/R_vocab": 0.19854910714285712 + }, + { + "learning_rate": 4.647707846785477e-07, + "loss": -0.0187, + "step": 294, + "train/R_acc": 0.00625, + "train/R_penalty": -0.71875, + "train/R_reason": 0.18667014063523668, + "train/R_vocab": 0.27327178030303034 + }, + { + "learning_rate": 4.642142940418973e-07, + "loss": 0.0376, + "step": 296, + "train/R_acc": 0.034375, + "train/R_penalty": -0.546875, + "train/R_reason": 0.193981736932492, + "train/R_vocab": 0.19051339285714286 + }, + { + "learning_rate": 4.6365378113265505e-07, + "loss": -0.0587, + "step": 298, + "train/R_acc": 0.003125, + "train/R_penalty": -0.76875, + "train/R_reason": 0.21676354110088458, + "train/R_vocab": 0.2875 + }, + { + "learning_rate": 4.630892564754956e-07, + "loss": 0.0012, + "step": 300, + "train/R_acc": 0.0, + "train/R_penalty": -0.59375, + "train/R_reason": 0.24459838346088703, + "train/R_vocab": 0.33515625 + }, + { + "eval/R_acc": 0.001388888888888889, + "eval/R_penalty": -0.6499999999999999, + "eval/R_reason": 0.2340619843391708, + "eval/R_vocab": 0.2772872574955908, + "step": 300 + }, + { + "step": 300 + }, + { + "learning_rate": 4.6252073067042127e-07, + "loss": -0.0006, + "step": 302, + "train/R_acc": 0.0, + "train/R_penalty": -0.65625, + "train/R_reason": 0.20949847516693765, + "train/R_vocab": 0.24609375 + }, + { + "learning_rate": 4.6194821439256373e-07, + "loss": 0.0001, + "step": 304, + "train/R_acc": 0.0, + "train/R_penalty": -0.546875, + "train/R_reason": 0.20496328530187652, + "train/R_vocab": 0.20011160714285714 + }, + { + "learning_rate": 4.6137171839198297e-07, + "loss": 0.0792, + "step": 306, + "train/R_acc": 0.0, + "train/R_penalty": -0.609375, + "train/R_reason": 0.1994046532974053, + "train/R_vocab": 0.2868551587301587 + }, + { + "learning_rate": 4.6079125349346576e-07, + "loss": 0.0068, + "step": 308, + "train/R_acc": 0.0, + "train/R_penalty": -0.625, + "train/R_reason": 0.18547339942836427, + "train/R_vocab": 0.2426711309523809 + }, + { + "learning_rate": 4.602068305963224e-07, + "loss": 0.0044, + "step": 310, + "train/R_acc": 0.0, + "train/R_penalty": -0.71875, + "train/R_reason": 0.2112060437994439, + "train/R_vocab": 0.2083829365079365 + }, + { + "learning_rate": 4.59618460674182e-07, + "loss": 0.0279, + "step": 312, + "train/R_acc": 0.034375, + "train/R_penalty": -0.53125, + "train/R_reason": 0.272254341125526, + "train/R_vocab": 0.45736607142857144 + }, + { + "learning_rate": 4.5902615477478636e-07, + "loss": -0.0181, + "step": 314, + "train/R_acc": 0.0, + "train/R_penalty": -0.671875, + "train/R_reason": 0.1449225433066495, + "train/R_vocab": 0.16674107142857142 + }, + { + "learning_rate": 4.5842992401978256e-07, + "loss": 0.0225, + "step": 316, + "train/R_acc": 0.0, + "train/R_penalty": -0.828125, + "train/R_reason": 0.2286200296039895, + "train/R_vocab": 0.2551339285714286 + }, + { + "learning_rate": 4.5782977960451414e-07, + "loss": 0.0192, + "step": 318, + "train/R_acc": 0.003125, + "train/R_penalty": -0.68125, + "train/R_reason": 0.21916647791965457, + "train/R_vocab": 0.2505580357142857 + }, + { + "learning_rate": 4.57225732797811e-07, + "loss": 0.0485, + "step": 320, + "train/R_acc": 0.0, + "train/R_penalty": -0.625, + "train/R_reason": 0.22755603236630983, + "train/R_vocab": 0.22947668650793654 + }, + { + "learning_rate": 4.566177949417777e-07, + "loss": 0.0163, + "step": 322, + "train/R_acc": 0.0, + "train/R_penalty": -0.515625, + "train/R_reason": 0.22636069312789864, + "train/R_vocab": 0.21104910714285716 + }, + { + "learning_rate": 4.560059774515804e-07, + "loss": 0.0143, + "step": 324, + "train/R_acc": 0.003125, + "train/R_penalty": -0.578125, + "train/R_reason": 0.15769657140460813, + "train/R_vocab": 0.22265625000000003 + }, + { + "learning_rate": 4.5539029181523284e-07, + "loss": -0.0796, + "step": 326, + "train/R_acc": 0.03125, + "train/R_penalty": -0.75, + "train/R_reason": 0.1961929211100274, + "train/R_vocab": 0.27142857142857146 + }, + { + "learning_rate": 4.5477074959338015e-07, + "loss": 0.0627, + "step": 328, + "train/R_acc": 0.003125, + "train/R_penalty": -0.75, + "train/R_reason": 0.22035420658260274, + "train/R_vocab": 0.33661954365079366 + }, + { + "learning_rate": 4.5414736241908214e-07, + "loss": -0.015, + "step": 330, + "train/R_acc": 0.0, + "train/R_penalty": -0.703125, + "train/R_reason": 0.21972588070640325, + "train/R_vocab": 0.30012400793650795 + }, + { + "learning_rate": 4.535201419975948e-07, + "loss": 0.0515, + "step": 332, + "train/R_acc": 0.009375000000000001, + "train/R_penalty": -0.46875, + "train/R_reason": 0.17121075988051718, + "train/R_vocab": 0.234375 + }, + { + "learning_rate": 4.5288910010615053e-07, + "loss": 0.0479, + "step": 334, + "train/R_acc": 0.037500000000000006, + "train/R_penalty": -0.5625, + "train/R_reason": 0.21868874143295755, + "train/R_vocab": 0.2630580357142857 + }, + { + "learning_rate": 4.5225424859373684e-07, + "loss": 0.0442, + "step": 336, + "train/R_acc": 0.00625, + "train/R_penalty": -0.578125, + "train/R_reason": 0.22348722064953433, + "train/R_vocab": 0.3254464285714286 + }, + { + "learning_rate": 4.51615599380874e-07, + "loss": 0.0299, + "step": 338, + "train/R_acc": 0.0, + "train/R_penalty": -0.578125, + "train/R_reason": 0.18059697532933602, + "train/R_vocab": 0.21510416666666665 + }, + { + "learning_rate": 4.5097316445939124e-07, + "loss": 0.0359, + "step": 340, + "train/R_acc": 0.003125, + "train/R_penalty": -0.546875, + "train/R_reason": 0.1941301451022681, + "train/R_vocab": 0.32220982142857146 + }, + { + "learning_rate": 4.503269558922015e-07, + "loss": -0.0107, + "step": 342, + "train/R_acc": 0.034375, + "train/R_penalty": -0.625, + "train/R_reason": 0.21181567459611045, + "train/R_vocab": 0.2482514880952381 + }, + { + "learning_rate": 4.496769858130748e-07, + "loss": 0.0179, + "step": 344, + "train/R_acc": 0.0, + "train/R_penalty": -0.625, + "train/R_reason": 0.23230602042254933, + "train/R_vocab": 0.3271701388888889 + }, + { + "learning_rate": 4.490232664264109e-07, + "loss": 0.042, + "step": 346, + "train/R_acc": 0.03125, + "train/R_penalty": -0.696875, + "train/R_reason": 0.2775747137079635, + "train/R_vocab": 0.3890625 + }, + { + "learning_rate": 4.4836581000700944e-07, + "loss": 0.0274, + "step": 348, + "train/R_acc": 0.0, + "train/R_penalty": -0.59375, + "train/R_reason": 0.22799018822395078, + "train/R_vocab": 0.27528521825396823 + }, + { + "learning_rate": 4.477046288998401e-07, + "loss": 0.0259, + "step": 350, + "train/R_acc": 0.13125, + "train/R_penalty": -0.546875, + "train/R_reason": 0.20698908553226353, + "train/R_vocab": 0.29140625 + }, + { + "learning_rate": 4.470397355198102e-07, + "loss": 0.0362, + "step": 352, + "train/R_acc": 0.0125, + "train/R_penalty": -0.525, + "train/R_reason": 0.210767134421379, + "train/R_vocab": 0.3446180555555556 + }, + { + "learning_rate": 4.463711423515323e-07, + "loss": 0.0077, + "step": 354, + "train/R_acc": 0.0, + "train/R_penalty": -0.796875, + "train/R_reason": 0.16874434359341295, + "train/R_vocab": 0.20223214285714286 + }, + { + "learning_rate": 4.456988619490889e-07, + "loss": 0.0364, + "step": 356, + "train/R_acc": 0.00625, + "train/R_penalty": -0.665625, + "train/R_reason": 0.2266184174999622, + "train/R_vocab": 0.32075892857142857 + }, + { + "learning_rate": 4.450229069357977e-07, + "loss": 0.0624, + "step": 358, + "train/R_acc": 0.0, + "train/R_penalty": -0.640625, + "train/R_reason": 0.19035143640898555, + "train/R_vocab": 0.1884548611111111 + }, + { + "learning_rate": 4.4434329000397363e-07, + "loss": -0.0017, + "step": 360, + "train/R_acc": 0.0, + "train/R_penalty": -0.625, + "train/R_reason": 0.18822223579328265, + "train/R_vocab": 0.2318452380952381 + }, + { + "learning_rate": 4.4366002391469126e-07, + "loss": 0.0383, + "step": 362, + "train/R_acc": 0.034375, + "train/R_penalty": -0.71875, + "train/R_reason": 0.18399620476664255, + "train/R_vocab": 0.17989831349206348 + }, + { + "learning_rate": 4.4297312149754477e-07, + "loss": 0.0709, + "step": 364, + "train/R_acc": 0.03125, + "train/R_penalty": -0.65625, + "train/R_reason": 0.2374170544763758, + "train/R_vocab": 0.2957589285714286 + }, + { + "learning_rate": 4.422825956504072e-07, + "loss": 0.012, + "step": 366, + "train/R_acc": 0.034375, + "train/R_penalty": -0.775, + "train/R_reason": 0.18171410574759275, + "train/R_vocab": 0.27645089285714286 + }, + { + "learning_rate": 4.415884593391882e-07, + "loss": -0.0291, + "step": 368, + "train/R_acc": 0.0, + "train/R_penalty": -0.728125, + "train/R_reason": 0.20255323919085794, + "train/R_vocab": 0.3200396825396825 + }, + { + "learning_rate": 4.4089072559759065e-07, + "loss": 0.0228, + "step": 370, + "train/R_acc": 0.0, + "train/R_penalty": -0.546875, + "train/R_reason": 0.25184199643437555, + "train/R_vocab": 0.31328125 + }, + { + "learning_rate": 4.40189407526866e-07, + "loss": -0.022, + "step": 372, + "train/R_acc": 0.03125, + "train/R_penalty": -0.65625, + "train/R_reason": 0.21868958770018052, + "train/R_vocab": 0.21980406746031747 + }, + { + "learning_rate": 4.3948451829556775e-07, + "loss": 0.0437, + "step": 374, + "train/R_acc": 0.0, + "train/R_penalty": -0.603125, + "train/R_reason": 0.23191700315429237, + "train/R_vocab": 0.25416666666666665 + }, + { + "learning_rate": 4.3877607113930516e-07, + "loss": -0.0032, + "step": 376, + "train/R_acc": 0.0, + "train/R_penalty": -0.5, + "train/R_reason": 0.1725605699858892, + "train/R_vocab": 0.23932291666666666 + }, + { + "learning_rate": 4.380640793604938e-07, + "loss": -0.0522, + "step": 378, + "train/R_acc": 0.0, + "train/R_penalty": -0.578125, + "train/R_reason": 0.1570635238253007, + "train/R_vocab": 0.209375 + }, + { + "learning_rate": 4.373485563281062e-07, + "loss": 0.0522, + "step": 380, + "train/R_acc": 0.0, + "train/R_penalty": -0.625, + "train/R_reason": 0.25503011482100246, + "train/R_vocab": 0.2538690476190476 + }, + { + "learning_rate": 4.3662951547742075e-07, + "loss": 0.0447, + "step": 382, + "train/R_acc": 0.0, + "train/R_penalty": -0.671875, + "train/R_reason": 0.1739243830311305, + "train/R_vocab": 0.23214285714285715 + }, + { + "learning_rate": 4.3590697030976965e-07, + "loss": -0.006, + "step": 384, + "train/R_acc": 0.03125, + "train/R_penalty": -0.65625, + "train/R_reason": 0.15723816285274234, + "train/R_vocab": 0.2829996392496392 + }, + { + "learning_rate": 4.3518093439228474e-07, + "loss": 0.0162, + "step": 386, + "train/R_acc": 0.0, + "train/R_penalty": -0.625, + "train/R_reason": 0.16239281076258544, + "train/R_vocab": 0.1482142857142857 + }, + { + "learning_rate": 4.3445142135764367e-07, + "loss": -0.0038, + "step": 388, + "train/R_acc": 0.034375, + "train/R_penalty": -0.6875, + "train/R_reason": 0.19918031683009202, + "train/R_vocab": 0.27239583333333334 + }, + { + "learning_rate": 4.33718444903813e-07, + "loss": 0.0341, + "step": 390, + "train/R_acc": 0.0, + "train/R_penalty": -0.640625, + "train/R_reason": 0.20130303529176516, + "train/R_vocab": 0.29593253968253963 + }, + { + "learning_rate": 4.329820187937919e-07, + "loss": 0.0876, + "step": 392, + "train/R_acc": 0.034375, + "train/R_penalty": -0.5, + "train/R_reason": 0.23979456377125202, + "train/R_vocab": 0.24832589285714285 + }, + { + "learning_rate": 4.3224215685535287e-07, + "loss": 0.0139, + "step": 394, + "train/R_acc": 0.0, + "train/R_penalty": -0.8125, + "train/R_reason": 0.1709315188515677, + "train/R_vocab": 0.18359375 + }, + { + "learning_rate": 4.314988729807827e-07, + "loss": -0.021, + "step": 396, + "train/R_acc": 0.0, + "train/R_penalty": -0.65625, + "train/R_reason": 0.17646043188459717, + "train/R_vocab": 0.17589285714285713 + }, + { + "learning_rate": 4.3075218112662135e-07, + "loss": 0.0513, + "step": 398, + "train/R_acc": 0.0, + "train/R_penalty": -0.609375, + "train/R_reason": 0.1973938417696664, + "train/R_vocab": 0.19837549603174603 + }, + { + "learning_rate": 4.3000209531339996e-07, + "loss": 0.0043, + "step": 400, + "train/R_acc": 0.0, + "train/R_penalty": -0.5625, + "train/R_reason": 0.24184598292042236, + "train/R_vocab": 0.33288690476190474 + }, + { + "eval/R_acc": 0.0, + "eval/R_penalty": -0.7916666666666666, + "eval/R_reason": 0.27821714712822276, + "eval/R_vocab": 0.32563381834215166, + "step": 400 + }, + { + "step": 400 + }, + { + "learning_rate": 4.2943367562278896e-07, + "loss": -0.0155, + "step": 402, + "train/R_acc": 0.065625, + "train/R_penalty": -0.515625, + "train/R_reason": 0.20335033205254746, + "train/R_vocab": 0.24464285714285713 + }, + { + "learning_rate": 4.286787117443108e-07, + "loss": -0.036, + "step": 404, + "train/R_acc": 0.04062500000000001, + "train/R_penalty": -0.5625, + "train/R_reason": 0.20780436238210176, + "train/R_vocab": 0.2703993055555556 + }, + { + "learning_rate": 4.2792040207614e-07, + "loss": -0.0044, + "step": 406, + "train/R_acc": 0.03125, + "train/R_penalty": -0.734375, + "train/R_reason": 0.2366017128148703, + "train/R_vocab": 0.3632440476190476 + }, + { + "learning_rate": 4.27158760817756e-07, + "loss": 0.0245, + "step": 408, + "train/R_acc": 0.0, + "train/R_penalty": -0.671875, + "train/R_reason": 0.19598878934576208, + "train/R_vocab": 0.28031994047619047 + }, + { + "learning_rate": 4.263938022310226e-07, + "loss": -0.036, + "step": 410, + "train/R_acc": 0.034375, + "train/R_penalty": -0.671875, + "train/R_reason": 0.23507477335166782, + "train/R_vocab": 0.27189980158730165 + }, + { + "learning_rate": 4.2562554063992127e-07, + "loss": 0.024, + "step": 412, + "train/R_acc": 0.0625, + "train/R_penalty": -0.7125, + "train/R_reason": 0.19770138255909644, + "train/R_vocab": 0.24754464285714284 + }, + { + "learning_rate": 4.248539904302829e-07, + "loss": 0.0091, + "step": 414, + "train/R_acc": 0.00625, + "train/R_penalty": -0.59375, + "train/R_reason": 0.2098416268444374, + "train/R_vocab": 0.2533234126984127 + }, + { + "learning_rate": 4.240791660495182e-07, + "loss": 0.0125, + "step": 416, + "train/R_acc": 0.003125, + "train/R_penalty": -0.603125, + "train/R_reason": 0.164986564236729, + "train/R_vocab": 0.1863219246031746 + }, + { + "learning_rate": 4.2330108200634723e-07, + "loss": 0.0353, + "step": 418, + "train/R_acc": 0.03125, + "train/R_penalty": -0.703125, + "train/R_reason": 0.16427460147948922, + "train/R_vocab": 0.1800595238095238 + }, + { + "learning_rate": 4.22519752870528e-07, + "loss": -0.0002, + "step": 420, + "train/R_acc": 0.003125, + "train/R_penalty": -0.703125, + "train/R_reason": 0.21440123784001752, + "train/R_vocab": 0.26026785714285716 + }, + { + "learning_rate": 4.2173519327258325e-07, + "loss": 0.0336, + "step": 422, + "train/R_acc": 0.003125, + "train/R_penalty": -0.59375, + "train/R_reason": 0.2248050023022144, + "train/R_vocab": 0.22937184343434344 + }, + { + "learning_rate": 4.2094741790352673e-07, + "loss": 0.0256, + "step": 424, + "train/R_acc": 0.0, + "train/R_penalty": -0.578125, + "train/R_reason": 0.22803044267397005, + "train/R_vocab": 0.28350694444444446 + }, + { + "learning_rate": 4.2015644151458827e-07, + "loss": -0.0319, + "step": 426, + "train/R_acc": 0.003125, + "train/R_penalty": -0.65625, + "train/R_reason": 0.20821167099100335, + "train/R_vocab": 0.21744791666666666 + }, + { + "learning_rate": 4.19362278916937e-07, + "loss": 0.0344, + "step": 428, + "train/R_acc": 0.03125, + "train/R_penalty": -0.671875, + "train/R_reason": 0.1891472011902095, + "train/R_vocab": 0.20641233766233769 + }, + { + "learning_rate": 4.185649449814045e-07, + "loss": 0.0712, + "step": 430, + "train/R_acc": 0.00625, + "train/R_penalty": -0.6875, + "train/R_reason": 0.2279409653415437, + "train/R_vocab": 0.2996651785714286 + }, + { + "learning_rate": 4.177644546382063e-07, + "loss": 0.0645, + "step": 432, + "train/R_acc": 0.037500000000000006, + "train/R_penalty": -0.625, + "train/R_reason": 0.21120852841137872, + "train/R_vocab": 0.19676339285714284 + }, + { + "learning_rate": 4.1696082287666217e-07, + "loss": 0.0481, + "step": 434, + "train/R_acc": 0.003125, + "train/R_penalty": -0.59375, + "train/R_reason": 0.20179254286468842, + "train/R_vocab": 0.3659598214285714 + }, + { + "learning_rate": 4.1615406474491535e-07, + "loss": -0.0453, + "step": 436, + "train/R_acc": 0.034375, + "train/R_penalty": -0.78125, + "train/R_reason": 0.21685467943636602, + "train/R_vocab": 0.2571428571428571 + }, + { + "learning_rate": 4.1534419534965105e-07, + "loss": 0.0985, + "step": 438, + "train/R_acc": 0.03125, + "train/R_penalty": -0.696875, + "train/R_reason": 0.20741733073153354, + "train/R_vocab": 0.27142857142857146 + }, + { + "learning_rate": 4.145312298558133e-07, + "loss": -0.0621, + "step": 440, + "train/R_acc": 0.003125, + "train/R_penalty": -0.5625, + "train/R_reason": 0.17062418658581344, + "train/R_vocab": 0.2 + }, + { + "learning_rate": 4.137151834863213e-07, + "loss": -0.0448, + "step": 442, + "train/R_acc": 0.03125, + "train/R_penalty": -0.5, + "train/R_reason": 0.23174115282287933, + "train/R_vocab": 0.36713789682539677 + }, + { + "learning_rate": 4.128960715217839e-07, + "loss": 0.0187, + "step": 444, + "train/R_acc": 0.003125, + "train/R_penalty": -0.59375, + "train/R_reason": 0.15167707696946453, + "train/R_vocab": 0.20052083333333334 + }, + { + "learning_rate": 4.1207390930021394e-07, + "loss": -0.0156, + "step": 446, + "train/R_acc": 0.003125, + "train/R_penalty": -0.546875, + "train/R_reason": 0.2000569888049995, + "train/R_vocab": 0.24027777777777778 + }, + { + "learning_rate": 4.11248712216741e-07, + "loss": 0.0897, + "step": 448, + "train/R_acc": 0.0, + "train/R_penalty": -0.68125, + "train/R_reason": 0.21749146764720112, + "train/R_vocab": 0.2636160714285714 + }, + { + "learning_rate": 4.104204957233225e-07, + "loss": 0.0032, + "step": 450, + "train/R_acc": 0.00625, + "train/R_penalty": -0.609375, + "train/R_reason": 0.19073873404612746, + "train/R_vocab": 0.30033482142857143 + }, + { + "learning_rate": 4.095892753284553e-07, + "loss": -0.0985, + "step": 452, + "train/R_acc": 0.0, + "train/R_penalty": -0.609375, + "train/R_reason": 0.25767563611959865, + "train/R_vocab": 0.41929563492063493 + }, + { + "learning_rate": 4.087550665968846e-07, + "loss": 0.0021, + "step": 454, + "train/R_acc": 0.0, + "train/R_penalty": -0.53125, + "train/R_reason": 0.1972933583803858, + "train/R_vocab": 0.23854166666666668 + }, + { + "learning_rate": 4.079178851493127e-07, + "loss": 0.0433, + "step": 456, + "train/R_acc": 0.0, + "train/R_penalty": -0.609375, + "train/R_reason": 0.2107089160600319, + "train/R_vocab": 0.25 + }, + { + "learning_rate": 4.070777466621067e-07, + "loss": -0.0103, + "step": 458, + "train/R_acc": 0.03125, + "train/R_penalty": -0.515625, + "train/R_reason": 0.2429741695318347, + "train/R_vocab": 0.38958333333333334 + }, + { + "learning_rate": 4.062346668670046e-07, + "loss": -0.0092, + "step": 460, + "train/R_acc": 0.003125, + "train/R_penalty": -0.734375, + "train/R_reason": 0.2328364572562428, + "train/R_vocab": 0.2795758928571429 + }, + { + "learning_rate": 4.0538866155082094e-07, + "loss": -0.0198, + "step": 462, + "train/R_acc": 0.040625, + "train/R_penalty": -0.4625, + "train/R_reason": 0.2483761704003692, + "train/R_vocab": 0.2885168650793651 + }, + { + "learning_rate": 4.045397465551513e-07, + "loss": -0.0249, + "step": 464, + "train/R_acc": 0.003125, + "train/R_penalty": -0.5875, + "train/R_reason": 0.17307827497369785, + "train/R_vocab": 0.2544642857142857 + }, + { + "learning_rate": 4.036879377760752e-07, + "loss": 0.0683, + "step": 466, + "train/R_acc": 0.03125, + "train/R_penalty": -0.671875, + "train/R_reason": 0.22330642900293768, + "train/R_vocab": 0.37135416666666665 + }, + { + "learning_rate": 4.02833251163859e-07, + "loss": -0.0192, + "step": 468, + "train/R_acc": 0.003125, + "train/R_penalty": -0.65, + "train/R_reason": 0.21129365536420686, + "train/R_vocab": 0.2589285714285714 + }, + { + "learning_rate": 4.0197570272265704e-07, + "loss": 0.0497, + "step": 470, + "train/R_acc": 0.003125, + "train/R_penalty": -0.515625, + "train/R_reason": 0.2052435879372262, + "train/R_vocab": 0.2556547619047619 + }, + { + "learning_rate": 4.011153085102116e-07, + "loss": 0.0309, + "step": 472, + "train/R_acc": 0.003125, + "train/R_penalty": -0.671875, + "train/R_reason": 0.21023509045617045, + "train/R_vocab": 0.2963169642857143 + }, + { + "learning_rate": 4.0025208463755274e-07, + "loss": 0.0424, + "step": 474, + "train/R_acc": 0.0, + "train/R_penalty": -0.784375, + "train/R_reason": 0.1959827825898999, + "train/R_vocab": 0.35357142857142854 + }, + { + "learning_rate": 3.9938604726869636e-07, + "loss": -0.0506, + "step": 476, + "train/R_acc": 0.0, + "train/R_penalty": -0.59375, + "train/R_reason": 0.19807364686139228, + "train/R_vocab": 0.17491319444444442 + }, + { + "learning_rate": 3.9851721262034157e-07, + "loss": 0.0724, + "step": 478, + "train/R_acc": 0.003125, + "train/R_penalty": -0.5, + "train/R_reason": 0.20997218794964279, + "train/R_vocab": 0.27433035714285714 + }, + { + "learning_rate": 3.9764559696156697e-07, + "loss": 0.0338, + "step": 480, + "train/R_acc": 0.003125, + "train/R_penalty": -0.665625, + "train/R_reason": 0.21410868507940445, + "train/R_vocab": 0.25691964285714286 + }, + { + "learning_rate": 3.9677121661352607e-07, + "loss": -0.0087, + "step": 482, + "train/R_acc": 0.0, + "train/R_penalty": -0.515625, + "train/R_reason": 0.21506411602738199, + "train/R_vocab": 0.3130952380952381 + }, + { + "learning_rate": 3.958940879491418e-07, + "loss": -0.0375, + "step": 484, + "train/R_acc": 0.0, + "train/R_penalty": -0.71875, + "train/R_reason": 0.18895673088287218, + "train/R_vocab": 0.21395089285714286 + }, + { + "learning_rate": 3.9501422739279953e-07, + "loss": -0.0485, + "step": 486, + "train/R_acc": 0.0, + "train/R_penalty": -0.5625, + "train/R_reason": 0.20816202695918376, + "train/R_vocab": 0.22013888888888888 + }, + { + "learning_rate": 3.9413165142004e-07, + "loss": -0.0479, + "step": 488, + "train/R_acc": 0.03125, + "train/R_penalty": -0.5625, + "train/R_reason": 0.1929871719837118, + "train/R_vocab": 0.3703125 + }, + { + "learning_rate": 3.932463765572505e-07, + "loss": -0.0312, + "step": 490, + "train/R_acc": 0.0, + "train/R_penalty": -0.75, + "train/R_reason": 0.2513287746728735, + "train/R_vocab": 0.27569444444444446 + }, + { + "learning_rate": 3.923584193813555e-07, + "loss": 0.03, + "step": 492, + "train/R_acc": 0.003125, + "train/R_penalty": -0.609375, + "train/R_reason": 0.23851168147176444, + "train/R_vocab": 0.313219246031746 + }, + { + "learning_rate": 3.914677965195062e-07, + "loss": -0.0009, + "step": 494, + "train/R_acc": 0.0, + "train/R_penalty": -0.625, + "train/R_reason": 0.15729111795287057, + "train/R_vocab": 0.1830357142857143 + }, + { + "learning_rate": 3.9057452464876946e-07, + "loss": 0.049, + "step": 496, + "train/R_acc": 0.003125, + "train/R_penalty": -0.703125, + "train/R_reason": 0.15483905838498802, + "train/R_vocab": 0.18125000000000002 + }, + { + "learning_rate": 3.89678620495815e-07, + "loss": 0.0521, + "step": 498, + "train/R_acc": 0.0, + "train/R_penalty": -0.75, + "train/R_reason": 0.2600686346082261, + "train/R_vocab": 0.253125 + }, + { + "learning_rate": 3.887801008366025e-07, + "loss": -0.005, + "step": 500, + "train/R_acc": 0.003125, + "train/R_penalty": -0.790625, + "train/R_reason": 0.22427641116364, + "train/R_vocab": 0.3156125992063492 + }, + { + "eval/R_acc": 0.0025, + "eval/R_penalty": -0.7075, + "eval/R_reason": 0.2610032668269429, + "eval/R_vocab": 0.21906295093795097, + "step": 500 + }, + { + "step": 500 + }, + { + "learning_rate": 3.8787898249606767e-07, + "loss": 0.0252, + "step": 502, + "train/R_acc": 0.0625, + "train/R_penalty": -0.609375, + "train/R_reason": 0.2684211960918422, + "train/R_vocab": 0.3091517857142857 + }, + { + "learning_rate": 3.8697528234780674e-07, + "loss": 0.019, + "step": 504, + "train/R_acc": 0.0, + "train/R_penalty": -0.640625, + "train/R_reason": 0.23017420694962337, + "train/R_vocab": 0.29288194444444443 + }, + { + "learning_rate": 3.86069017313761e-07, + "loss": 0.013, + "step": 506, + "train/R_acc": 0.03125, + "train/R_penalty": -0.578125, + "train/R_reason": 0.22614899753752782, + "train/R_vocab": 0.3089905753968254 + }, + { + "learning_rate": 3.851602043638994e-07, + "loss": -0.0248, + "step": 508, + "train/R_acc": 0.034375, + "train/R_penalty": -0.515625, + "train/R_reason": 0.2017065165275432, + "train/R_vocab": 0.26004464285714285 + }, + { + "learning_rate": 3.8424886051590115e-07, + "loss": 0.0296, + "step": 510, + "train/R_acc": 0.0, + "train/R_penalty": -0.7375, + "train/R_reason": 0.17507956245908884, + "train/R_vocab": 0.17238343253968252 + }, + { + "learning_rate": 3.83335002834837e-07, + "loss": -0.0622, + "step": 512, + "train/R_acc": 0.0, + "train/R_penalty": -0.640625, + "train/R_reason": 0.22332916829962618, + "train/R_vocab": 0.34873511904761906 + }, + { + "learning_rate": 3.8241864843284964e-07, + "loss": -0.0075, + "step": 514, + "train/R_acc": 0.003125, + "train/R_penalty": -0.640625, + "train/R_reason": 0.132802326682096, + "train/R_vocab": 0.16104910714285714 + }, + { + "learning_rate": 3.814998144688333e-07, + "loss": 0.0227, + "step": 516, + "train/R_acc": 0.034375, + "train/R_penalty": -0.59375, + "train/R_reason": 0.1754689583879656, + "train/R_vocab": 0.19609374999999998 + }, + { + "learning_rate": 3.805785181481123e-07, + "loss": -0.0118, + "step": 518, + "train/R_acc": 0.003125, + "train/R_penalty": -0.540625, + "train/R_reason": 0.23541827525810197, + "train/R_vocab": 0.25967261904761907 + }, + { + "learning_rate": 3.796547767221193e-07, + "loss": -0.0924, + "step": 520, + "train/R_acc": 0.0, + "train/R_penalty": -0.59375, + "train/R_reason": 0.2376371992311745, + "train/R_vocab": 0.3128968253968254 + }, + { + "learning_rate": 3.787286074880718e-07, + "loss": -0.0122, + "step": 522, + "train/R_acc": 0.0, + "train/R_penalty": -0.55625, + "train/R_reason": 0.22978465999581174, + "train/R_vocab": 0.2617931547619048 + }, + { + "learning_rate": 3.778000277886483e-07, + "loss": -0.0054, + "step": 524, + "train/R_acc": 0.0, + "train/R_penalty": -0.53125, + "train/R_reason": 0.22061871520509158, + "train/R_vocab": 0.30894209956709956 + }, + { + "learning_rate": 3.768690550116639e-07, + "loss": 0.0092, + "step": 526, + "train/R_acc": 0.03125, + "train/R_penalty": -0.546875, + "train/R_reason": 0.22469643175166784, + "train/R_vocab": 0.34073660714285714 + }, + { + "learning_rate": 3.7593570658974436e-07, + "loss": 0.0264, + "step": 528, + "train/R_acc": 0.03125, + "train/R_penalty": -0.640625, + "train/R_reason": 0.26427900699025786, + "train/R_vocab": 0.32444196428571426 + }, + { + "learning_rate": 3.75e-07, + "loss": 0.0264, + "step": 530, + "train/R_acc": 0.003125, + "train/R_penalty": -0.625, + "train/R_reason": 0.16190914297300985, + "train/R_vocab": 0.18397817460317462 + }, + { + "learning_rate": 3.740619527636979e-07, + "loss": -0.0014, + "step": 532, + "train/R_acc": 0.003125, + "train/R_penalty": -0.540625, + "train/R_reason": 0.2013910995332763, + "train/R_vocab": 0.16294642857142855 + }, + { + "learning_rate": 3.731215824459344e-07, + "loss": 0.0653, + "step": 534, + "train/R_acc": 0.037500000000000006, + "train/R_penalty": -0.640625, + "train/R_reason": 0.1980956136240478, + "train/R_vocab": 0.26245039682539684 + }, + { + "learning_rate": 3.7217890665530587e-07, + "loss": -0.0052, + "step": 536, + "train/R_acc": 0.003125, + "train/R_penalty": -0.609375, + "train/R_reason": 0.193118823374168, + "train/R_vocab": 0.18813131313131315 + }, + { + "learning_rate": 3.712339430435792e-07, + "loss": 0.044, + "step": 538, + "train/R_acc": 0.00625, + "train/R_penalty": -0.65625, + "train/R_reason": 0.21065931054504308, + "train/R_vocab": 0.2915054563492063 + }, + { + "learning_rate": 3.7028670930536077e-07, + "loss": 0.0443, + "step": 540, + "train/R_acc": 0.0, + "train/R_penalty": -0.625, + "train/R_reason": 0.21326431514536787, + "train/R_vocab": 0.23958333333333334 + }, + { + "learning_rate": 3.693372231777658e-07, + "loss": 0.0062, + "step": 542, + "train/R_acc": 0.003125, + "train/R_penalty": -0.86875, + "train/R_reason": 0.2361538113041483, + "train/R_vocab": 0.26145833333333335 + }, + { + "learning_rate": 3.6838550244008573e-07, + "loss": 0.0011, + "step": 544, + "train/R_acc": 0.0, + "train/R_penalty": -0.671875, + "train/R_reason": 0.2560180166234486, + "train/R_vocab": 0.3419642857142857 + }, + { + "learning_rate": 3.6743156491345564e-07, + "loss": -0.1304, + "step": 546, + "train/R_acc": 0.0, + "train/R_penalty": -0.515625, + "train/R_reason": 0.22700866641314227, + "train/R_vocab": 0.3409474206349206 + }, + { + "learning_rate": 3.6647542846052003e-07, + "loss": 0.0374, + "step": 548, + "train/R_acc": 0.003125, + "train/R_penalty": -0.515625, + "train/R_reason": 0.18355650415358743, + "train/R_vocab": 0.23628472222222222 + }, + { + "learning_rate": 3.65517110985099e-07, + "loss": -0.0193, + "step": 550, + "train/R_acc": 0.0625, + "train/R_penalty": -0.546875, + "train/R_reason": 0.1497471083396625, + "train/R_vocab": 0.1660342261904762 + }, + { + "learning_rate": 3.645566304318526e-07, + "loss": -0.0168, + "step": 552, + "train/R_acc": 0.00625, + "train/R_penalty": -0.71875, + "train/R_reason": 0.17760395766642206, + "train/R_vocab": 0.20212053571428573 + }, + { + "learning_rate": 3.6359400478594473e-07, + "loss": 0.0519, + "step": 554, + "train/R_acc": 0.03125, + "train/R_penalty": -0.68125, + "train/R_reason": 0.2032957807067698, + "train/R_vocab": 0.3547991071428571 + }, + { + "learning_rate": 3.6262925207270666e-07, + "loss": -0.0077, + "step": 556, + "train/R_acc": 0.034375, + "train/R_penalty": -0.53125, + "train/R_reason": 0.24608095001365107, + "train/R_vocab": 0.2532366071428571 + }, + { + "learning_rate": 3.616623903572994e-07, + "loss": -0.0273, + "step": 558, + "train/R_acc": 0.03125, + "train/R_penalty": -0.46875, + "train/R_reason": 0.25668514206671234, + "train/R_vocab": 0.32946428571428565 + }, + { + "learning_rate": 3.6069343774437516e-07, + "loss": 0.0356, + "step": 560, + "train/R_acc": 0.00625, + "train/R_penalty": -0.853125, + "train/R_reason": 0.22578641617093248, + "train/R_vocab": 0.3117063492063492 + }, + { + "learning_rate": 3.597224123777389e-07, + "loss": 0.0074, + "step": 562, + "train/R_acc": 0.00625, + "train/R_penalty": -0.61875, + "train/R_reason": 0.1608838946497146, + "train/R_vocab": 0.22940228174603175 + }, + { + "learning_rate": 3.58749332440008e-07, + "loss": 0.0218, + "step": 564, + "train/R_acc": 0.0, + "train/R_penalty": -0.5, + "train/R_reason": 0.2042184580919646, + "train/R_vocab": 0.23277529761904764 + }, + { + "learning_rate": 3.5777421615227207e-07, + "loss": 0.0446, + "step": 566, + "train/R_acc": 0.003125, + "train/R_penalty": -0.65625, + "train/R_reason": 0.21492820594376136, + "train/R_vocab": 0.2902901785714286 + }, + { + "learning_rate": 3.567970817737518e-07, + "loss": 0.0771, + "step": 568, + "train/R_acc": 0.0, + "train/R_penalty": -0.625, + "train/R_reason": 0.2497326839973281, + "train/R_vocab": 0.30859375 + }, + { + "learning_rate": 3.5581794760145696e-07, + "loss": 0.0354, + "step": 570, + "train/R_acc": 0.0, + "train/R_penalty": -0.609375, + "train/R_reason": 0.20181240026890324, + "train/R_vocab": 0.22098214285714285 + }, + { + "learning_rate": 3.548368319698437e-07, + "loss": 0.0731, + "step": 572, + "train/R_acc": 0.00625, + "train/R_penalty": -0.696875, + "train/R_reason": 0.16856543760413237, + "train/R_vocab": 0.1800595238095238 + }, + { + "learning_rate": 3.5385375325047163e-07, + "loss": 0.0291, + "step": 574, + "train/R_acc": 0.0, + "train/R_penalty": -0.59375, + "train/R_reason": 0.2099860707148752, + "train/R_vocab": 0.2753348214285714 + }, + { + "learning_rate": 3.528687298516591e-07, + "loss": -0.0002, + "step": 576, + "train/R_acc": 0.0, + "train/R_penalty": -0.625, + "train/R_reason": 0.17694118112711849, + "train/R_vocab": 0.12705853174603174 + }, + { + "learning_rate": 3.5188178021813925e-07, + "loss": 0.028, + "step": 578, + "train/R_acc": 0.037500000000000006, + "train/R_penalty": -0.609375, + "train/R_reason": 0.24354058691683123, + "train/R_vocab": 0.39140624999999996 + }, + { + "learning_rate": 3.5089292283071417e-07, + "loss": 0.0398, + "step": 580, + "train/R_acc": 0.00625, + "train/R_penalty": -0.53125, + "train/R_reason": 0.22343115261429752, + "train/R_vocab": 0.32678571428571423 + }, + { + "learning_rate": 3.499021762059089e-07, + "loss": 0.0565, + "step": 582, + "train/R_acc": 0.03125, + "train/R_penalty": -0.7125, + "train/R_reason": 0.19496747632032374, + "train/R_vocab": 0.26284722222222223 + }, + { + "learning_rate": 3.489095588956249e-07, + "loss": 0.0092, + "step": 584, + "train/R_acc": 0.0, + "train/R_penalty": -0.5625, + "train/R_reason": 0.2074570772291119, + "train/R_vocab": 0.30972222222222223 + }, + { + "learning_rate": 3.479150894867926e-07, + "loss": -0.0015, + "step": 586, + "train/R_acc": 0.0625, + "train/R_penalty": -0.703125, + "train/R_reason": 0.20443537938998835, + "train/R_vocab": 0.2761904761904762 + }, + { + "learning_rate": 3.46918786601023e-07, + "loss": 0.1059, + "step": 588, + "train/R_acc": 0.0, + "train/R_penalty": -0.68125, + "train/R_reason": 0.20239388628972388, + "train/R_vocab": 0.22630208333333335 + }, + { + "learning_rate": 3.459206688942596e-07, + "loss": 0.0608, + "step": 590, + "train/R_acc": 0.0, + "train/R_penalty": -0.75, + "train/R_reason": 0.24682579728602413, + "train/R_vocab": 0.4008928571428571 + }, + { + "learning_rate": 3.4492075505642847e-07, + "loss": 0.055, + "step": 592, + "train/R_acc": 0.0, + "train/R_penalty": -0.5, + "train/R_reason": 0.19593385729165314, + "train/R_vocab": 0.31875 + }, + { + "learning_rate": 3.439190638110888e-07, + "loss": 0.041, + "step": 594, + "train/R_acc": 0.0, + "train/R_penalty": -0.59375, + "train/R_reason": 0.1711366333418505, + "train/R_vocab": 0.26268601190476193 + }, + { + "learning_rate": 3.4291561391508185e-07, + "loss": 0.0807, + "step": 596, + "train/R_acc": 0.003125, + "train/R_penalty": -0.59375, + "train/R_reason": 0.240933143584435, + "train/R_vocab": 0.4119791666666667 + }, + { + "learning_rate": 3.4191042415818e-07, + "loss": -0.0247, + "step": 598, + "train/R_acc": 0.003125, + "train/R_penalty": -0.74375, + "train/R_reason": 0.2539401489710603, + "train/R_vocab": 0.2673363095238095 + }, + { + "learning_rate": 3.4090351336273474e-07, + "loss": -0.0238, + "step": 600, + "train/R_acc": 0.0, + "train/R_penalty": -0.703125, + "train/R_reason": 0.19356852065501937, + "train/R_vocab": 0.24319196428571427 + }, + { + "eval/R_acc": 0.005, + "eval/R_penalty": -0.625, + "eval/R_reason": 0.24855003165355338, + "eval/R_vocab": 0.22628517316017316, + "step": 600 + }, + { + "step": 600 + }, + { + "learning_rate": 3.398949003833246e-07, + "loss": 0.0122, + "step": 602, + "train/R_acc": 0.00625, + "train/R_penalty": -0.578125, + "train/R_reason": 0.22168416427879684, + "train/R_vocab": 0.24791666666666667 + }, + { + "learning_rate": 3.388846041064012e-07, + "loss": 0.0797, + "step": 604, + "train/R_acc": 0.09375, + "train/R_penalty": -0.546875, + "train/R_reason": 0.1777229928874741, + "train/R_vocab": 0.29464285714285715 + }, + { + "learning_rate": 3.378726434499368e-07, + "loss": 0.033, + "step": 606, + "train/R_acc": 0.0, + "train/R_penalty": -0.7125, + "train/R_reason": 0.1956222769670857, + "train/R_vocab": 0.3197916666666667 + }, + { + "learning_rate": 3.368590373630692e-07, + "loss": -0.0275, + "step": 608, + "train/R_acc": 0.06875, + "train/R_penalty": -0.46875, + "train/R_reason": 0.17503223220226105, + "train/R_vocab": 0.2591517857142857 + }, + { + "learning_rate": 3.3584380482574717e-07, + "loss": -0.058, + "step": 610, + "train/R_acc": 0.003125, + "train/R_penalty": -0.546875, + "train/R_reason": 0.1902514010231657, + "train/R_vocab": 0.23125 + }, + { + "learning_rate": 3.348269648483749e-07, + "loss": -0.054, + "step": 612, + "train/R_acc": 0.0, + "train/R_penalty": -0.53125, + "train/R_reason": 0.17467640906171372, + "train/R_vocab": 0.190625 + }, + { + "learning_rate": 3.3380853647145656e-07, + "loss": -0.0371, + "step": 614, + "train/R_acc": 0.00625, + "train/R_penalty": -0.65625, + "train/R_reason": 0.17121010785732182, + "train/R_vocab": 0.2569444444444444 + }, + { + "learning_rate": 3.327885387652391e-07, + "loss": -0.0013, + "step": 616, + "train/R_acc": 0.0, + "train/R_penalty": -0.6875, + "train/R_reason": 0.19415455674882087, + "train/R_vocab": 0.23277529761904764 + }, + { + "learning_rate": 3.317669908293554e-07, + "loss": 0.0253, + "step": 618, + "train/R_acc": 0.003125, + "train/R_penalty": -0.609375, + "train/R_reason": 0.24193135939319937, + "train/R_vocab": 0.25 + }, + { + "learning_rate": 3.307439117924668e-07, + "loss": 0.0588, + "step": 620, + "train/R_acc": 0.0, + "train/R_penalty": -0.6875, + "train/R_reason": 0.20416377806011923, + "train/R_vocab": 0.2056423611111111 + }, + { + "learning_rate": 3.297193208119047e-07, + "loss": -0.0064, + "step": 622, + "train/R_acc": 0.0, + "train/R_penalty": -0.484375, + "train/R_reason": 0.211192554787049, + "train/R_vocab": 0.30390625000000004 + }, + { + "learning_rate": 3.2869323707331176e-07, + "loss": 0.0335, + "step": 624, + "train/R_acc": 0.003125, + "train/R_penalty": -0.59375, + "train/R_reason": 0.23403307094336184, + "train/R_vocab": 0.35301339285714284 + }, + { + "learning_rate": 3.2766567979028324e-07, + "loss": -0.0308, + "step": 626, + "train/R_acc": 0.00625, + "train/R_penalty": -0.578125, + "train/R_reason": 0.25211879339329, + "train/R_vocab": 0.278125 + }, + { + "learning_rate": 3.2663666820400625e-07, + "loss": 0.0239, + "step": 628, + "train/R_acc": 0.03125, + "train/R_penalty": -0.609375, + "train/R_reason": 0.23123528000410434, + "train/R_vocab": 0.328844246031746 + }, + { + "learning_rate": 3.2560622158290025e-07, + "loss": 0.0127, + "step": 630, + "train/R_acc": 0.003125, + "train/R_penalty": -0.5, + "train/R_reason": 0.22103995585722944, + "train/R_vocab": 0.26785714285714285 + }, + { + "learning_rate": 3.2457435922225603e-07, + "loss": 0.0445, + "step": 632, + "train/R_acc": 0.003125, + "train/R_penalty": -0.609375, + "train/R_reason": 0.185424618889319, + "train/R_vocab": 0.28928571428571426 + }, + { + "learning_rate": 3.235411004438741e-07, + "loss": 0.1006, + "step": 634, + "train/R_acc": 0.003125, + "train/R_penalty": -0.625, + "train/R_reason": 0.23390812732295468, + "train/R_vocab": 0.29355158730158726 + }, + { + "learning_rate": 3.2250646459570343e-07, + "loss": 0.0611, + "step": 636, + "train/R_acc": 0.0, + "train/R_penalty": -0.65625, + "train/R_reason": 0.17366692788791044, + "train/R_vocab": 0.21026785714285715 + }, + { + "learning_rate": 3.214704710514786e-07, + "loss": 0.0463, + "step": 638, + "train/R_acc": 0.0, + "train/R_penalty": -0.625, + "train/R_reason": 0.23856693280209076, + "train/R_vocab": 0.3723214285714286 + }, + { + "learning_rate": 3.204331392103574e-07, + "loss": -0.0048, + "step": 640, + "train/R_acc": 0.0, + "train/R_penalty": -0.609375, + "train/R_reason": 0.2248622406893548, + "train/R_vocab": 0.38828125 + }, + { + "learning_rate": 3.193944884965576e-07, + "loss": 0.0172, + "step": 642, + "train/R_acc": 0.065625, + "train/R_penalty": -0.671875, + "train/R_reason": 0.2000315806348117, + "train/R_vocab": 0.32477678571428575 + }, + { + "learning_rate": 3.183545383589927e-07, + "loss": -0.0164, + "step": 644, + "train/R_acc": 0.003125, + "train/R_penalty": -0.55625, + "train/R_reason": 0.1901960813142597, + "train/R_vocab": 0.2540178571428572 + }, + { + "learning_rate": 3.173133082709086e-07, + "loss": 0.0815, + "step": 646, + "train/R_acc": 0.003125, + "train/R_penalty": -0.68125, + "train/R_reason": 0.20267428182016378, + "train/R_vocab": 0.2115079365079365 + }, + { + "learning_rate": 3.1627081772951815e-07, + "loss": 0.0279, + "step": 648, + "train/R_acc": 0.03125, + "train/R_penalty": -0.484375, + "train/R_reason": 0.18044257928711105, + "train/R_vocab": 0.21806795634920634 + }, + { + "learning_rate": 3.152270862556367e-07, + "loss": 0.0707, + "step": 650, + "train/R_acc": 0.003125, + "train/R_penalty": -0.515625, + "train/R_reason": 0.26331484572063035, + "train/R_vocab": 0.24206349206349206 + }, + { + "learning_rate": 3.1418213339331576e-07, + "loss": -0.0396, + "step": 652, + "train/R_acc": 0.003125, + "train/R_penalty": -0.53125, + "train/R_reason": 0.23778099952829604, + "train/R_vocab": 0.284375 + }, + { + "learning_rate": 3.1313597870947816e-07, + "loss": 0.0883, + "step": 654, + "train/R_acc": 0.003125, + "train/R_penalty": -0.546875, + "train/R_reason": 0.22377350102735885, + "train/R_vocab": 0.3117931547619047 + }, + { + "learning_rate": 3.1208864179355074e-07, + "loss": 0.0009, + "step": 656, + "train/R_acc": 0.03125, + "train/R_penalty": -0.571875, + "train/R_reason": 0.20656730207307256, + "train/R_vocab": 0.2512276785714286 + }, + { + "learning_rate": 3.1104014225709784e-07, + "loss": 0.0828, + "step": 658, + "train/R_acc": 0.003125, + "train/R_penalty": -0.65, + "train/R_reason": 0.2410817269424846, + "train/R_vocab": 0.265625 + }, + { + "learning_rate": 3.099904997334541e-07, + "loss": -0.0317, + "step": 660, + "train/R_acc": 0.03125, + "train/R_penalty": -0.5875, + "train/R_reason": 0.25759275591542546, + "train/R_vocab": 0.24017857142857144 + }, + { + "learning_rate": 3.0893973387735683e-07, + "loss": -0.0235, + "step": 662, + "train/R_acc": 0.003125, + "train/R_penalty": -0.625, + "train/R_reason": 0.16711049206280193, + "train/R_vocab": 0.21428571428571427 + }, + { + "learning_rate": 3.078878643645778e-07, + "loss": 0.0242, + "step": 664, + "train/R_acc": 0.00625, + "train/R_penalty": -0.696875, + "train/R_reason": 0.22549504519978125, + "train/R_vocab": 0.33912450396825394 + }, + { + "learning_rate": 3.068349108915553e-07, + "loss": 0.0147, + "step": 666, + "train/R_acc": 0.003125, + "train/R_penalty": -0.546875, + "train/R_reason": 0.2824812700031851, + "train/R_vocab": 0.24285714285714285 + }, + { + "learning_rate": 3.0578089317502436e-07, + "loss": -0.0134, + "step": 668, + "train/R_acc": 0.034375, + "train/R_penalty": -0.525, + "train/R_reason": 0.1605276972684076, + "train/R_vocab": 0.22779017857142858 + }, + { + "learning_rate": 3.0472583095164873e-07, + "loss": -0.0496, + "step": 670, + "train/R_acc": 0.0, + "train/R_penalty": -0.609375, + "train/R_reason": 0.19456119250844006, + "train/R_vocab": 0.23214285714285715 + }, + { + "learning_rate": 3.036697439776504e-07, + "loss": 0.0554, + "step": 672, + "train/R_acc": 0.0, + "train/R_penalty": -0.515625, + "train/R_reason": 0.2515289920392415, + "train/R_vocab": 0.2724330357142857 + }, + { + "learning_rate": 3.0261265202844016e-07, + "loss": 0.02, + "step": 674, + "train/R_acc": 0.0, + "train/R_penalty": -0.65625, + "train/R_reason": 0.18580953264755184, + "train/R_vocab": 0.25388144841269844 + }, + { + "learning_rate": 3.01554574898247e-07, + "loss": 0.0155, + "step": 676, + "train/R_acc": 0.0, + "train/R_penalty": -0.55625, + "train/R_reason": 0.22948436053810994, + "train/R_vocab": 0.22013888888888888 + }, + { + "learning_rate": 3.004955323997478e-07, + "loss": -0.0179, + "step": 678, + "train/R_acc": 0.00625, + "train/R_penalty": -0.59375, + "train/R_reason": 0.21153443586764287, + "train/R_vocab": 0.23524305555555558 + }, + { + "learning_rate": 2.994355443636958e-07, + "loss": -0.0362, + "step": 680, + "train/R_acc": 0.003125, + "train/R_penalty": -0.625, + "train/R_reason": 0.1840845890109734, + "train/R_vocab": 0.31015625 + }, + { + "learning_rate": 2.983746306385499e-07, + "loss": 0.0297, + "step": 682, + "train/R_acc": 0.0625, + "train/R_penalty": -0.453125, + "train/R_reason": 0.2084560166468069, + "train/R_vocab": 0.31228918650793647 + }, + { + "learning_rate": 2.9731281109010253e-07, + "loss": 0.0534, + "step": 684, + "train/R_acc": 0.0, + "train/R_penalty": -0.71875, + "train/R_reason": 0.16422497838360026, + "train/R_vocab": 0.23582589285714284 + }, + { + "learning_rate": 2.9625010560110787e-07, + "loss": 0.0183, + "step": 686, + "train/R_acc": 0.003125, + "train/R_penalty": -0.634375, + "train/R_reason": 0.24390149995516267, + "train/R_vocab": 0.29910714285714285 + }, + { + "learning_rate": 2.951865340709095e-07, + "loss": 0.0683, + "step": 688, + "train/R_acc": 0.034375, + "train/R_penalty": -0.64375, + "train/R_reason": 0.23521821731229603, + "train/R_vocab": 0.3359747023809524 + }, + { + "learning_rate": 2.9412211641506773e-07, + "loss": -0.001, + "step": 690, + "train/R_acc": 0.0, + "train/R_penalty": -0.65625, + "train/R_reason": 0.18437836963522442, + "train/R_vocab": 0.209375 + }, + { + "learning_rate": 2.930568725649867e-07, + "loss": 0.0194, + "step": 692, + "train/R_acc": 0.03125, + "train/R_penalty": -0.65, + "train/R_reason": 0.21411232248047601, + "train/R_vocab": 0.2759548611111111 + }, + { + "learning_rate": 2.919908224675412e-07, + "loss": 0.0579, + "step": 694, + "train/R_acc": 0.003125, + "train/R_penalty": -0.5625, + "train/R_reason": 0.1540599514957384, + "train/R_vocab": 0.18344494047619048 + }, + { + "learning_rate": 2.9092398608470326e-07, + "loss": -0.0106, + "step": 696, + "train/R_acc": 0.03125, + "train/R_penalty": -0.59375, + "train/R_reason": 0.2501350490350204, + "train/R_vocab": 0.3291294642857143 + }, + { + "learning_rate": 2.898563833931681e-07, + "loss": 0.0156, + "step": 698, + "train/R_acc": 0.0, + "train/R_penalty": -0.609375, + "train/R_reason": 0.20865047120336916, + "train/R_vocab": 0.2130332341269841 + }, + { + "learning_rate": 2.8878803438398015e-07, + "loss": 0.0192, + "step": 700, + "train/R_acc": 0.0, + "train/R_penalty": -0.515625, + "train/R_reason": 0.19110667770899964, + "train/R_vocab": 0.24427083333333333 + }, + { + "eval/R_acc": 0.0025, + "eval/R_penalty": -0.5875, + "eval/R_reason": 0.25186260885534084, + "eval/R_vocab": 0.2177263708513708, + "step": 700 + }, + { + "step": 700 + }, + { + "learning_rate": 2.8771895906215895e-07, + "loss": 0.05, + "step": 702, + "train/R_acc": 0.003125, + "train/R_penalty": -0.53125, + "train/R_reason": 0.2448799886759996, + "train/R_vocab": 0.3176339285714286 + }, + { + "learning_rate": 2.8664917744632423e-07, + "loss": 0.0057, + "step": 704, + "train/R_acc": 0.003125, + "train/R_penalty": -0.578125, + "train/R_reason": 0.22652983918213118, + "train/R_vocab": 0.41149553571428577 + }, + { + "learning_rate": 2.8557870956832133e-07, + "loss": 0.0356, + "step": 706, + "train/R_acc": 0.0, + "train/R_penalty": -0.71875, + "train/R_reason": 0.22872856053626722, + "train/R_vocab": 0.2622767857142857 + }, + { + "learning_rate": 2.8450757547284576e-07, + "loss": 0.0236, + "step": 708, + "train/R_acc": 0.0, + "train/R_penalty": -0.609375, + "train/R_reason": 0.19584017346681087, + "train/R_vocab": 0.20625000000000002 + }, + { + "learning_rate": 2.8343579521706823e-07, + "loss": -0.0244, + "step": 710, + "train/R_acc": 0.003125, + "train/R_penalty": -0.640625, + "train/R_reason": 0.17909964580588045, + "train/R_vocab": 0.23783482142857143 + }, + { + "learning_rate": 2.8236338887025886e-07, + "loss": 0.0155, + "step": 712, + "train/R_acc": 0.034375, + "train/R_penalty": -0.484375, + "train/R_reason": 0.26302313003150035, + "train/R_vocab": 0.3813988095238095 + }, + { + "learning_rate": 2.812903765134115e-07, + "loss": -0.0178, + "step": 714, + "train/R_acc": 0.0, + "train/R_penalty": -0.546875, + "train/R_reason": 0.2631388468076138, + "train/R_vocab": 0.4032242063492063 + }, + { + "learning_rate": 2.8021677823886754e-07, + "loss": -0.0467, + "step": 716, + "train/R_acc": 0.003125, + "train/R_penalty": -0.65625, + "train/R_reason": 0.19911790925777817, + "train/R_vocab": 0.23608630952380952 + }, + { + "learning_rate": 2.7914261414993976e-07, + "loss": 0.0054, + "step": 718, + "train/R_acc": 0.003125, + "train/R_penalty": -0.603125, + "train/R_reason": 0.2387146239629917, + "train/R_vocab": 0.2571428571428571 + }, + { + "learning_rate": 2.780679043605361e-07, + "loss": -0.0082, + "step": 720, + "train/R_acc": 0.03125, + "train/R_penalty": -0.571875, + "train/R_reason": 0.2269094768328706, + "train/R_vocab": 0.31436011904761907 + }, + { + "learning_rate": 2.7699266899478274e-07, + "loss": 0.0178, + "step": 722, + "train/R_acc": 0.034375, + "train/R_penalty": -0.703125, + "train/R_reason": 0.24130363361070928, + "train/R_vocab": 0.22456597222222222 + }, + { + "learning_rate": 2.759169281866472e-07, + "loss": 0.0661, + "step": 724, + "train/R_acc": 0.03125, + "train/R_penalty": -0.64375, + "train/R_reason": 0.21308869490175653, + "train/R_vocab": 0.2772321428571428 + }, + { + "learning_rate": 2.748407020795617e-07, + "loss": 0.1056, + "step": 726, + "train/R_acc": 0.096875, + "train/R_penalty": -0.421875, + "train/R_reason": 0.23621705469287618, + "train/R_vocab": 0.2981770833333333 + }, + { + "learning_rate": 2.737640108260456e-07, + "loss": 0.0812, + "step": 728, + "train/R_acc": 0.03125, + "train/R_penalty": -0.6875, + "train/R_reason": 0.2310603125451809, + "train/R_vocab": 0.30590277777777775 + }, + { + "learning_rate": 2.726868745873286e-07, + "loss": -0.0637, + "step": 730, + "train/R_acc": 0.034375, + "train/R_penalty": -0.696875, + "train/R_reason": 0.19990588785781663, + "train/R_vocab": 0.265625 + }, + { + "learning_rate": 2.716093135329722e-07, + "loss": 0.0348, + "step": 732, + "train/R_acc": 0.0375, + "train/R_penalty": -0.421875, + "train/R_reason": 0.2095006410516756, + "train/R_vocab": 0.32751736111111107 + }, + { + "learning_rate": 2.7053134784049316e-07, + "loss": 0.0875, + "step": 734, + "train/R_acc": 0.003125, + "train/R_penalty": -0.71875, + "train/R_reason": 0.17118221219882784, + "train/R_vocab": 0.17343750000000002 + }, + { + "learning_rate": 2.694529976949849e-07, + "loss": -0.0448, + "step": 736, + "train/R_acc": 0.037500000000000006, + "train/R_penalty": -0.609375, + "train/R_reason": 0.20588894905097077, + "train/R_vocab": 0.34609375 + }, + { + "learning_rate": 2.6837428328874007e-07, + "loss": -0.0244, + "step": 738, + "train/R_acc": 0.06875, + "train/R_penalty": -0.6125, + "train/R_reason": 0.26183079845209467, + "train/R_vocab": 0.2656994047619048 + }, + { + "learning_rate": 2.6729522482087165e-07, + "loss": 0.0854, + "step": 740, + "train/R_acc": 0.003125, + "train/R_penalty": -0.59375, + "train/R_reason": 0.20793241180273003, + "train/R_vocab": 0.21996527777777775 + }, + { + "learning_rate": 2.662158424969357e-07, + "loss": 0.0149, + "step": 742, + "train/R_acc": 0.003125, + "train/R_penalty": -0.46875, + "train/R_reason": 0.18410838343666375, + "train/R_vocab": 0.1974330357142857 + }, + { + "learning_rate": 2.6513615652855246e-07, + "loss": -0.0144, + "step": 744, + "train/R_acc": 0.0, + "train/R_penalty": -0.640625, + "train/R_reason": 0.24293661434589603, + "train/R_vocab": 0.3267857142857143 + }, + { + "learning_rate": 2.6405618713302783e-07, + "loss": -0.0678, + "step": 746, + "train/R_acc": 0.00625, + "train/R_penalty": -0.61875, + "train/R_reason": 0.135718109682479, + "train/R_vocab": 0.22979910714285715 + }, + { + "learning_rate": 2.629759545329749e-07, + "loss": 0.1246, + "step": 748, + "train/R_acc": 0.0, + "train/R_penalty": -0.634375, + "train/R_reason": 0.20181279701198124, + "train/R_vocab": 0.3018849206349207 + }, + { + "learning_rate": 2.618954789559356e-07, + "loss": 0.0933, + "step": 750, + "train/R_acc": 0.003125, + "train/R_penalty": -0.578125, + "train/R_reason": 0.20252563848526106, + "train/R_vocab": 0.2291294642857143 + }, + { + "learning_rate": 2.6081478063400124e-07, + "loss": -0.0234, + "step": 752, + "train/R_acc": 0.003125, + "train/R_penalty": -0.609375, + "train/R_reason": 0.19106041589978606, + "train/R_vocab": 0.20889136904761904 + }, + { + "learning_rate": 2.597338798034344e-07, + "loss": -0.0631, + "step": 754, + "train/R_acc": 0.0, + "train/R_penalty": -0.65625, + "train/R_reason": 0.1867345308525643, + "train/R_vocab": 0.22708333333333333 + }, + { + "learning_rate": 2.5865279670428956e-07, + "loss": 0.0483, + "step": 756, + "train/R_acc": 0.003125, + "train/R_penalty": -0.625, + "train/R_reason": 0.17135153722070606, + "train/R_vocab": 0.18147321428571428 + }, + { + "learning_rate": 2.5757155158003414e-07, + "loss": 0.0295, + "step": 758, + "train/R_acc": 0.0, + "train/R_penalty": -0.703125, + "train/R_reason": 0.1663155804467654, + "train/R_vocab": 0.22377232142857145 + }, + { + "learning_rate": 2.5649016467716954e-07, + "loss": 0.0283, + "step": 760, + "train/R_acc": 0.04062500000000001, + "train/R_penalty": -0.375, + "train/R_reason": 0.19205683929152886, + "train/R_vocab": 0.2728422619047619 + }, + { + "learning_rate": 2.5540865624485216e-07, + "loss": -0.0491, + "step": 762, + "train/R_acc": 0.0, + "train/R_penalty": -0.703125, + "train/R_reason": 0.2550380457423676, + "train/R_vocab": 0.2734375 + }, + { + "learning_rate": 2.5432704653451374e-07, + "loss": 0.0441, + "step": 764, + "train/R_acc": 0.03125, + "train/R_penalty": -0.53125, + "train/R_reason": 0.20927672505440928, + "train/R_vocab": 0.30055803571428574 + }, + { + "learning_rate": 2.532453557994827e-07, + "loss": 0.048, + "step": 766, + "train/R_acc": 0.003125, + "train/R_penalty": -0.515625, + "train/R_reason": 0.20293042206204454, + "train/R_vocab": 0.2447668650793651 + }, + { + "learning_rate": 2.521636042946046e-07, + "loss": 0.0256, + "step": 768, + "train/R_acc": 0.04062500000000001, + "train/R_penalty": -0.55625, + "train/R_reason": 0.2228340738544955, + "train/R_vocab": 0.21646825396825398 + }, + { + "learning_rate": 2.5108181227586307e-07, + "loss": -0.0186, + "step": 770, + "train/R_acc": 0.03125, + "train/R_penalty": -0.703125, + "train/R_reason": 0.15339300119866717, + "train/R_vocab": 0.24015376984126985 + }, + { + "learning_rate": 2.5e-07, + "loss": 0.0082, + "step": 772, + "train/R_acc": 0.034375, + "train/R_penalty": -0.59375, + "train/R_reason": 0.25634859172881913, + "train/R_vocab": 0.2591517857142857 + }, + { + "learning_rate": 2.48918187724137e-07, + "loss": -0.0372, + "step": 774, + "train/R_acc": 0.0, + "train/R_penalty": -0.484375, + "train/R_reason": 0.15909137824419298, + "train/R_vocab": 0.2725953733766234 + }, + { + "learning_rate": 2.4783639570539535e-07, + "loss": -0.1172, + "step": 776, + "train/R_acc": 0.00625, + "train/R_penalty": -0.46875, + "train/R_reason": 0.21610677086076593, + "train/R_vocab": 0.21164434523809522 + }, + { + "learning_rate": 2.467546442005173e-07, + "loss": 0.0088, + "step": 778, + "train/R_acc": 0.06875, + "train/R_penalty": -0.390625, + "train/R_reason": 0.19689824765738337, + "train/R_vocab": 0.1890625 + }, + { + "learning_rate": 2.4567295346548634e-07, + "loss": 0.1308, + "step": 780, + "train/R_acc": 0.03125, + "train/R_penalty": -0.625, + "train/R_reason": 0.16858555636857672, + "train/R_vocab": 0.29520089285714285 + }, + { + "learning_rate": 2.4459134375514787e-07, + "loss": -0.0356, + "step": 782, + "train/R_acc": 0.003125, + "train/R_penalty": -0.5875, + "train/R_reason": 0.20190652121448122, + "train/R_vocab": 0.2455357142857143 + }, + { + "learning_rate": 2.4350983532283043e-07, + "loss": -0.0182, + "step": 784, + "train/R_acc": 0.003125, + "train/R_penalty": -0.703125, + "train/R_reason": 0.1856742513120948, + "train/R_vocab": 0.275 + }, + { + "learning_rate": 2.4242844841996583e-07, + "loss": 0.0167, + "step": 786, + "train/R_acc": 0.0, + "train/R_penalty": -0.625, + "train/R_reason": 0.20132693663438478, + "train/R_vocab": 0.18971974206349207 + }, + { + "learning_rate": 2.4134720329571047e-07, + "loss": 0.0408, + "step": 788, + "train/R_acc": 0.037500000000000006, + "train/R_penalty": -0.571875, + "train/R_reason": 0.22452081729968582, + "train/R_vocab": 0.21640625 + }, + { + "learning_rate": 2.4026612019656556e-07, + "loss": 0.0107, + "step": 790, + "train/R_acc": 0.034375, + "train/R_penalty": -0.609375, + "train/R_reason": 0.18352974837528543, + "train/R_vocab": 0.25814732142857144 + }, + { + "learning_rate": 2.3918521936599874e-07, + "loss": 0.036, + "step": 792, + "train/R_acc": 0.03125, + "train/R_penalty": -0.609375, + "train/R_reason": 0.27166955319028896, + "train/R_vocab": 0.35963541666666665 + }, + { + "learning_rate": 2.381045210440644e-07, + "loss": -0.0346, + "step": 794, + "train/R_acc": 0.0, + "train/R_penalty": -0.625, + "train/R_reason": 0.22416891439419612, + "train/R_vocab": 0.32734375000000004 + }, + { + "learning_rate": 2.3702404546702509e-07, + "loss": 0.0911, + "step": 796, + "train/R_acc": 0.034375, + "train/R_penalty": -0.546875, + "train/R_reason": 0.24303349052516854, + "train/R_vocab": 0.27745535714285713 + }, + { + "learning_rate": 2.3594381286697215e-07, + "loss": 0.0121, + "step": 798, + "train/R_acc": 0.0, + "train/R_penalty": -0.609375, + "train/R_reason": 0.23618598427535661, + "train/R_vocab": 0.28122519841269844 + }, + { + "learning_rate": 2.3486384347144752e-07, + "loss": 0.0064, + "step": 800, + "train/R_acc": 0.0, + "train/R_penalty": -0.640625, + "train/R_reason": 0.25934359162168097, + "train/R_vocab": 0.4203125 + }, + { + "eval/R_acc": 0.0025, + "eval/R_penalty": -0.5875, + "eval/R_reason": 0.29083729720667373, + "eval/R_vocab": 0.22747113997114, + "step": 800 + }, + { + "step": 800 + }, + { + "learning_rate": 2.337841575030642e-07, + "loss": -0.0305, + "step": 802, + "train/R_acc": 0.03125, + "train/R_penalty": -0.59375, + "train/R_reason": 0.20417958820529003, + "train/R_vocab": 0.2702008928571429 + }, + { + "learning_rate": 2.3270477517912835e-07, + "loss": 0.0361, + "step": 804, + "train/R_acc": 0.00625, + "train/R_penalty": -0.621875, + "train/R_reason": 0.2872431102010518, + "train/R_vocab": 0.37351190476190477 + }, + { + "learning_rate": 2.3162571671126e-07, + "loss": 0.0184, + "step": 806, + "train/R_acc": 0.0, + "train/R_penalty": -0.546875, + "train/R_reason": 0.2281768901636334, + "train/R_vocab": 0.30616319444444445 + }, + { + "learning_rate": 2.3054700230501502e-07, + "loss": 0.0128, + "step": 808, + "train/R_acc": 0.03125, + "train/R_penalty": -0.671875, + "train/R_reason": 0.20789617508651104, + "train/R_vocab": 0.3257440476190476 + }, + { + "learning_rate": 2.2946865215950685e-07, + "loss": 0.0158, + "step": 810, + "train/R_acc": 0.003125, + "train/R_penalty": -0.671875, + "train/R_reason": 0.22696589858153038, + "train/R_vocab": 0.21540178571428575 + }, + { + "learning_rate": 2.2839068646702786e-07, + "loss": 1.5239, + "step": 812, + "train/R_acc": 0.0375, + "train/R_penalty": -0.734375, + "train/R_reason": 0.1726926027420783, + "train/R_vocab": 0.25167410714285715 + }, + { + "learning_rate": 2.2731312541267143e-07, + "loss": 0.0348, + "step": 814, + "train/R_acc": 0.009375000000000001, + "train/R_penalty": -0.53125, + "train/R_reason": 0.18792369164667982, + "train/R_vocab": 0.2892485119047619 + }, + { + "learning_rate": 2.2623598917395436e-07, + "loss": 0.0032, + "step": 816, + "train/R_acc": 0.003125, + "train/R_penalty": -0.625, + "train/R_reason": 0.16929829805580582, + "train/R_vocab": 0.16987847222222222 + }, + { + "learning_rate": 2.251592979204383e-07, + "loss": 0.0127, + "step": 818, + "train/R_acc": 0.03125, + "train/R_penalty": -0.390625, + "train/R_reason": 0.28610113589945263, + "train/R_vocab": 0.35 + }, + { + "learning_rate": 2.2408307181335285e-07, + "loss": 0.0528, + "step": 820, + "train/R_acc": 0.0, + "train/R_penalty": -0.509375, + "train/R_reason": 0.18837133598647154, + "train/R_vocab": 0.21272321428571428 + }, + { + "learning_rate": 2.2300733100521732e-07, + "loss": 0.0454, + "step": 822, + "train/R_acc": 0.003125, + "train/R_penalty": -0.671875, + "train/R_reason": 0.1814823048344199, + "train/R_vocab": 0.13814484126984128 + }, + { + "learning_rate": 2.2193209563946382e-07, + "loss": 0.0355, + "step": 824, + "train/R_acc": 0.0, + "train/R_penalty": -0.6875, + "train/R_reason": 0.18663805046129311, + "train/R_vocab": 0.1985863095238095 + }, + { + "learning_rate": 2.2085738585006021e-07, + "loss": 0.0026, + "step": 826, + "train/R_acc": 0.0, + "train/R_penalty": -0.49375, + "train/R_reason": 0.18273024857160083, + "train/R_vocab": 0.21136363636363636 + }, + { + "learning_rate": 2.1978322176113251e-07, + "loss": -0.0318, + "step": 828, + "train/R_acc": 0.034375, + "train/R_penalty": -0.578125, + "train/R_reason": 0.19883211536778234, + "train/R_vocab": 0.27878224206349206 + }, + { + "learning_rate": 2.1870962348658852e-07, + "loss": -0.0335, + "step": 830, + "train/R_acc": 0.009375000000000001, + "train/R_penalty": -0.546875, + "train/R_reason": 0.1985312923562238, + "train/R_vocab": 0.2785714285714286 + }, + { + "learning_rate": 2.1763661112974115e-07, + "loss": -0.0012, + "step": 832, + "train/R_acc": 0.0, + "train/R_penalty": -0.59375, + "train/R_reason": 0.2787590605964265, + "train/R_vocab": 0.35 + }, + { + "learning_rate": 2.165642047829318e-07, + "loss": -0.0375, + "step": 834, + "train/R_acc": 0.03125, + "train/R_penalty": -0.578125, + "train/R_reason": 0.15530706954823303, + "train/R_vocab": 0.22005208333333334 + }, + { + "learning_rate": 2.1549242452715427e-07, + "loss": -0.0387, + "step": 836, + "train/R_acc": 0.003125, + "train/R_penalty": -0.5, + "train/R_reason": 0.24527110960119072, + "train/R_vocab": 0.38571428571428573 + }, + { + "learning_rate": 2.1442129043167873e-07, + "loss": 0.0321, + "step": 838, + "train/R_acc": 0.003125, + "train/R_penalty": -0.703125, + "train/R_reason": 0.21774443607948502, + "train/R_vocab": 0.26517857142857143 + }, + { + "learning_rate": 2.1335082255367572e-07, + "loss": 0.0288, + "step": 840, + "train/R_acc": 0.03125, + "train/R_penalty": -0.546875, + "train/R_reason": 0.22299875389864712, + "train/R_vocab": 0.27976190476190477 + }, + { + "learning_rate": 2.1228104093784108e-07, + "loss": 0.0203, + "step": 842, + "train/R_acc": 0.003125, + "train/R_penalty": -0.515625, + "train/R_reason": 0.25459030399590693, + "train/R_vocab": 0.36163194444444446 + }, + { + "learning_rate": 2.112119656160199e-07, + "loss": 0.0337, + "step": 844, + "train/R_acc": 0.003125, + "train/R_penalty": -0.515625, + "train/R_reason": 0.2547007960169786, + "train/R_vocab": 0.34011656746031743 + }, + { + "learning_rate": 2.1014361660683194e-07, + "loss": -0.0236, + "step": 846, + "train/R_acc": 0.03125, + "train/R_penalty": -0.609375, + "train/R_reason": 0.23028272641397346, + "train/R_vocab": 0.25 + }, + { + "learning_rate": 2.0907601391529677e-07, + "loss": -0.0204, + "step": 848, + "train/R_acc": 0.003125, + "train/R_penalty": -0.546875, + "train/R_reason": 0.18682793317030633, + "train/R_vocab": 0.21324066558441557 + }, + { + "learning_rate": 2.0800917753245875e-07, + "loss": 0.1057, + "step": 850, + "train/R_acc": 0.00625, + "train/R_penalty": -0.609375, + "train/R_reason": 0.2020052493242343, + "train/R_vocab": 0.27734375 + }, + { + "learning_rate": 2.0694312743501334e-07, + "loss": -0.0585, + "step": 852, + "train/R_acc": 0.0, + "train/R_penalty": -0.640625, + "train/R_reason": 0.23457174043904147, + "train/R_vocab": 0.3036830357142857 + }, + { + "learning_rate": 2.0587788358493235e-07, + "loss": -0.0318, + "step": 854, + "train/R_acc": 0.003125, + "train/R_penalty": -0.34375, + "train/R_reason": 0.2538105699194751, + "train/R_vocab": 0.40078125 + }, + { + "learning_rate": 2.048134659290905e-07, + "loss": -0.0723, + "step": 856, + "train/R_acc": 0.0, + "train/R_penalty": -0.6875, + "train/R_reason": 0.17298681086107648, + "train/R_vocab": 0.15267857142857144 + }, + { + "learning_rate": 2.037498943988921e-07, + "loss": -0.0058, + "step": 858, + "train/R_acc": 0.0, + "train/R_penalty": -0.5625, + "train/R_reason": 0.28539022868157543, + "train/R_vocab": 0.3465593434343434 + }, + { + "learning_rate": 2.0268718890989752e-07, + "loss": -0.0416, + "step": 860, + "train/R_acc": 0.03125, + "train/R_penalty": -0.509375, + "train/R_reason": 0.20216319087856421, + "train/R_vocab": 0.3190848214285714 + }, + { + "learning_rate": 2.0162536936145008e-07, + "loss": -0.0399, + "step": 862, + "train/R_acc": 0.003125, + "train/R_penalty": -0.696875, + "train/R_reason": 0.2420341418631836, + "train/R_vocab": 0.27738095238095234 + }, + { + "learning_rate": 2.0056445563630423e-07, + "loss": 0.0372, + "step": 864, + "train/R_acc": 0.03125, + "train/R_penalty": -0.55625, + "train/R_reason": 0.21099684301910737, + "train/R_vocab": 0.26659226190476193 + }, + { + "learning_rate": 1.9950446760025219e-07, + "loss": 0.0326, + "step": 866, + "train/R_acc": 0.0625, + "train/R_penalty": -0.61875, + "train/R_reason": 0.19938542697531797, + "train/R_vocab": 0.25535714285714284 + }, + { + "learning_rate": 1.98445425101753e-07, + "loss": 0.0211, + "step": 868, + "train/R_acc": 0.003125, + "train/R_penalty": -0.671875, + "train/R_reason": 0.22302710668155856, + "train/R_vocab": 0.2036830357142857 + }, + { + "learning_rate": 1.9738734797155987e-07, + "loss": 0.1391, + "step": 870, + "train/R_acc": 0.0125, + "train/R_penalty": -0.703125, + "train/R_reason": 0.20507300880803225, + "train/R_vocab": 0.29453125 + }, + { + "learning_rate": 1.963302560223496e-07, + "loss": 0.0314, + "step": 872, + "train/R_acc": 0.0, + "train/R_penalty": -0.515625, + "train/R_reason": 0.1946319904099672, + "train/R_vocab": 0.32533482142857145 + }, + { + "learning_rate": 1.9527416904835132e-07, + "loss": 0.0562, + "step": 874, + "train/R_acc": 0.0, + "train/R_penalty": -0.625, + "train/R_reason": 0.2123123501156333, + "train/R_vocab": 0.24363839285714287 + }, + { + "learning_rate": 1.9421910682497572e-07, + "loss": 0.0335, + "step": 876, + "train/R_acc": 0.0625, + "train/R_penalty": -0.578125, + "train/R_reason": 0.24626829087757965, + "train/R_vocab": 0.28229166666666666 + }, + { + "learning_rate": 1.9316508910844476e-07, + "loss": 0.0255, + "step": 878, + "train/R_acc": 0.034375, + "train/R_penalty": -0.59375, + "train/R_reason": 0.18789237365926498, + "train/R_vocab": 0.278125 + }, + { + "learning_rate": 1.9211213563542218e-07, + "loss": -0.0622, + "step": 880, + "train/R_acc": 0.0, + "train/R_penalty": -0.5, + "train/R_reason": 0.19592139018937305, + "train/R_vocab": 0.26670386904761906 + }, + { + "learning_rate": 1.9106026612264315e-07, + "loss": 0.0249, + "step": 882, + "train/R_acc": 0.065625, + "train/R_penalty": -0.5, + "train/R_reason": 0.17202561289857762, + "train/R_vocab": 0.24394841269841272 + }, + { + "learning_rate": 1.900095002665459e-07, + "loss": 0.0627, + "step": 884, + "train/R_acc": 0.003125, + "train/R_penalty": -0.5625, + "train/R_reason": 0.20212251182315438, + "train/R_vocab": 0.234375 + }, + { + "learning_rate": 1.889598577429022e-07, + "loss": 0.0251, + "step": 886, + "train/R_acc": 0.00625, + "train/R_penalty": -0.6875, + "train/R_reason": 0.21931575978869752, + "train/R_vocab": 0.20535714285714285 + }, + { + "learning_rate": 1.8791135820644922e-07, + "loss": 0.0228, + "step": 888, + "train/R_acc": 0.0, + "train/R_penalty": -0.5625, + "train/R_reason": 0.23473900860844063, + "train/R_vocab": 0.2440104166666667 + }, + { + "learning_rate": 1.8686402129052181e-07, + "loss": 0.0602, + "step": 890, + "train/R_acc": 0.034375, + "train/R_penalty": -0.5625, + "train/R_reason": 0.2188358452056008, + "train/R_vocab": 0.24036458333333333 + }, + { + "learning_rate": 1.858178666066843e-07, + "loss": 0.0042, + "step": 892, + "train/R_acc": 0.034375, + "train/R_penalty": -0.59375, + "train/R_reason": 0.1951297685148197, + "train/R_vocab": 0.23270089285714285 + }, + { + "learning_rate": 1.8477291374436338e-07, + "loss": -0.0364, + "step": 894, + "train/R_acc": 0.0, + "train/R_penalty": -0.640625, + "train/R_reason": 0.11877722404091096, + "train/R_vocab": 0.12317708333333333 + }, + { + "learning_rate": 1.8372918227048183e-07, + "loss": 0.3093, + "step": 896, + "train/R_acc": 0.003125, + "train/R_penalty": -0.578125, + "train/R_reason": 0.1855972158801493, + "train/R_vocab": 0.25017361111111114 + }, + { + "learning_rate": 1.8268669172909136e-07, + "loss": 0.0017, + "step": 898, + "train/R_acc": 0.003125, + "train/R_penalty": -0.625, + "train/R_reason": 0.18804079477258429, + "train/R_vocab": 0.16622023809523812 + }, + { + "learning_rate": 1.8164546164100728e-07, + "loss": -0.0186, + "step": 900, + "train/R_acc": 0.003125, + "train/R_penalty": -0.546875, + "train/R_reason": 0.24246348014867208, + "train/R_vocab": 0.29110863095238093 + }, + { + "eval/R_acc": 0.030000000000000006, + "eval/R_penalty": -0.6, + "eval/R_reason": 0.29157417644976924, + "eval/R_vocab": 0.37875, + "step": 900 + }, + { + "step": 900 + }, + { + "learning_rate": 1.806055115034425e-07, + "loss": -0.0579, + "step": 902, + "train/R_acc": 0.003125, + "train/R_penalty": -0.515625, + "train/R_reason": 0.23201039495721226, + "train/R_vocab": 0.2525669642857143 + }, + { + "learning_rate": 1.7956686078964255e-07, + "loss": 0.0383, + "step": 904, + "train/R_acc": 0.03125, + "train/R_penalty": -0.61875, + "train/R_reason": 0.19469584258694483, + "train/R_vocab": 0.21830357142857146 + }, + { + "learning_rate": 1.7852952894852148e-07, + "loss": 0.0293, + "step": 906, + "train/R_acc": 0.009375000000000001, + "train/R_penalty": -0.625, + "train/R_reason": 0.25663321405687106, + "train/R_vocab": 0.32265625000000003 + }, + { + "learning_rate": 1.7749353540429663e-07, + "loss": -0.0106, + "step": 908, + "train/R_acc": 0.003125, + "train/R_penalty": -0.546875, + "train/R_reason": 0.19230140159133588, + "train/R_vocab": 0.16744791666666667 + }, + { + "learning_rate": 1.7645889955612592e-07, + "loss": -0.0295, + "step": 910, + "train/R_acc": 0.003125, + "train/R_penalty": -0.578125, + "train/R_reason": 0.12090342349306576, + "train/R_vocab": 0.1583457341269841 + }, + { + "learning_rate": 1.7542564077774403e-07, + "loss": 0.0045, + "step": 912, + "train/R_acc": 0.034375, + "train/R_penalty": -0.675, + "train/R_reason": 0.2114914680859179, + "train/R_vocab": 0.3030505952380952 + }, + { + "learning_rate": 1.7439377841709973e-07, + "loss": -0.0017, + "step": 914, + "train/R_acc": 0.0625, + "train/R_penalty": -0.609375, + "train/R_reason": 0.217770400372361, + "train/R_vocab": 0.25592757936507937 + }, + { + "learning_rate": 1.7336333179599378e-07, + "loss": 0.0849, + "step": 916, + "train/R_acc": 0.00625, + "train/R_penalty": -0.609375, + "train/R_reason": 0.2602397752431368, + "train/R_vocab": 0.4150173611111111 + }, + { + "learning_rate": 1.7233432020971684e-07, + "loss": 0.0793, + "step": 918, + "train/R_acc": 0.0, + "train/R_penalty": -0.55625, + "train/R_reason": 0.2766625564419999, + "train/R_vocab": 0.23649553571428572 + }, + { + "learning_rate": 1.7130676292668816e-07, + "loss": 0.0446, + "step": 920, + "train/R_acc": 0.003125, + "train/R_penalty": -0.6875, + "train/R_reason": 0.22695619844003265, + "train/R_vocab": 0.20081845238095236 + }, + { + "learning_rate": 1.7028067918809535e-07, + "loss": -0.036, + "step": 922, + "train/R_acc": 0.0, + "train/R_penalty": -0.578125, + "train/R_reason": 0.17932276860695562, + "train/R_vocab": 0.2564732142857143 + }, + { + "learning_rate": 1.6925608820753325e-07, + "loss": -0.0333, + "step": 924, + "train/R_acc": 0.0, + "train/R_penalty": -0.5625, + "train/R_reason": 0.1710392126096163, + "train/R_vocab": 0.11149553571428572 + }, + { + "learning_rate": 1.6823300917064458e-07, + "loss": -0.0427, + "step": 926, + "train/R_acc": 0.0, + "train/R_penalty": -0.515625, + "train/R_reason": 0.1853385177642702, + "train/R_vocab": 0.20937499999999998 + }, + { + "learning_rate": 1.6721146123476093e-07, + "loss": 0.122, + "step": 928, + "train/R_acc": 0.03125, + "train/R_penalty": -0.634375, + "train/R_reason": 0.21111988917807223, + "train/R_vocab": 0.28467261904761904 + }, + { + "learning_rate": 1.6619146352854334e-07, + "loss": 0.0335, + "step": 930, + "train/R_acc": 0.00625, + "train/R_penalty": -0.525, + "train/R_reason": 0.2012206313966573, + "train/R_vocab": 0.2486359126984127 + }, + { + "learning_rate": 1.6517303515162505e-07, + "loss": -0.0597, + "step": 932, + "train/R_acc": 0.034375, + "train/R_penalty": -0.5625, + "train/R_reason": 0.16395669599613677, + "train/R_vocab": 0.2521329365079365 + }, + { + "learning_rate": 1.6415619517425294e-07, + "loss": 0.0165, + "step": 934, + "train/R_acc": 0.003125, + "train/R_penalty": -0.65, + "train/R_reason": 0.14773459331491504, + "train/R_vocab": 0.2604910714285714 + }, + { + "learning_rate": 1.6314096263693078e-07, + "loss": -0.0171, + "step": 936, + "train/R_acc": 0.0, + "train/R_penalty": -0.61875, + "train/R_reason": 0.24617103354461017, + "train/R_vocab": 0.25078125 + }, + { + "learning_rate": 1.6212735655006322e-07, + "loss": -0.0329, + "step": 938, + "train/R_acc": 0.0, + "train/R_penalty": -0.609375, + "train/R_reason": 0.2476563881397631, + "train/R_vocab": 0.22760416666666666 + }, + { + "learning_rate": 1.6111539589359885e-07, + "loss": -0.0157, + "step": 940, + "train/R_acc": 0.034375, + "train/R_penalty": -0.55625, + "train/R_reason": 0.18716442508640668, + "train/R_vocab": 0.21462053571428572 + }, + { + "learning_rate": 1.6010509961667546e-07, + "loss": 0.0015, + "step": 942, + "train/R_acc": 0.00625, + "train/R_penalty": -0.5625, + "train/R_reason": 0.2163363343470736, + "train/R_vocab": 0.3014508928571429 + }, + { + "learning_rate": 1.590964866372652e-07, + "loss": -0.0064, + "step": 944, + "train/R_acc": 0.0, + "train/R_penalty": -0.5875, + "train/R_reason": 0.22361495696140293, + "train/R_vocab": 0.278125 + }, + { + "learning_rate": 1.5808957584181994e-07, + "loss": -0.0227, + "step": 946, + "train/R_acc": 0.03125, + "train/R_penalty": -0.596875, + "train/R_reason": 0.2240925844127057, + "train/R_vocab": 0.30379464285714286 + }, + { + "learning_rate": 1.5708438608491815e-07, + "loss": 0.0113, + "step": 948, + "train/R_acc": 0.0, + "train/R_penalty": -0.578125, + "train/R_reason": 0.19887587350757036, + "train/R_vocab": 0.2051846590909091 + }, + { + "learning_rate": 1.560809361889112e-07, + "loss": 0.0187, + "step": 950, + "train/R_acc": 0.0, + "train/R_penalty": -0.671875, + "train/R_reason": 0.1577485102536454, + "train/R_vocab": 0.20122767857142856 + }, + { + "learning_rate": 1.5507924494357154e-07, + "loss": -0.0467, + "step": 952, + "train/R_acc": 0.034375, + "train/R_penalty": -0.5625, + "train/R_reason": 0.14574324556072643, + "train/R_vocab": 0.20066964285714284 + }, + { + "learning_rate": 1.5407933110574044e-07, + "loss": 0.0123, + "step": 954, + "train/R_acc": 0.0, + "train/R_penalty": -0.5625, + "train/R_reason": 0.21002736065355815, + "train/R_vocab": 0.27578125 + }, + { + "learning_rate": 1.5308121339897705e-07, + "loss": 0.0131, + "step": 956, + "train/R_acc": 0.0, + "train/R_penalty": -0.5625, + "train/R_reason": 0.24345409494067652, + "train/R_vocab": 0.33035714285714285 + }, + { + "learning_rate": 1.5208491051320744e-07, + "loss": -0.1197, + "step": 958, + "train/R_acc": 0.037500000000000006, + "train/R_penalty": -0.578125, + "train/R_reason": 0.21702138753711128, + "train/R_vocab": 0.28426339285714286 + }, + { + "learning_rate": 1.510904411043751e-07, + "loss": -0.0818, + "step": 960, + "train/R_acc": 0.0, + "train/R_penalty": -0.71875, + "train/R_reason": 0.23023067131726802, + "train/R_vocab": 0.3091517857142857 + }, + { + "learning_rate": 1.5009782379409108e-07, + "loss": -0.1087, + "step": 962, + "train/R_acc": 0.003125, + "train/R_penalty": -0.765625, + "train/R_reason": 0.17687649620358462, + "train/R_vocab": 0.21863839285714284 + }, + { + "learning_rate": 1.4910707716928586e-07, + "loss": 0.0956, + "step": 964, + "train/R_acc": 0.009375000000000001, + "train/R_penalty": -0.46875, + "train/R_reason": 0.22474928414381412, + "train/R_vocab": 0.26093750000000004 + }, + { + "learning_rate": 1.4811821978186078e-07, + "loss": 0.0536, + "step": 966, + "train/R_acc": 0.003125, + "train/R_penalty": -0.65625, + "train/R_reason": 0.22651502510275934, + "train/R_vocab": 0.20833333333333331 + }, + { + "learning_rate": 1.4713127014834088e-07, + "loss": 0.0205, + "step": 968, + "train/R_acc": 0.0, + "train/R_penalty": -0.5625, + "train/R_reason": 0.2021908958276862, + "train/R_vocab": 0.1978794642857143 + }, + { + "learning_rate": 1.461462467495284e-07, + "loss": 0.0427, + "step": 970, + "train/R_acc": 0.0, + "train/R_penalty": -0.625, + "train/R_reason": 0.15024320937786767, + "train/R_vocab": 0.20468750000000002 + }, + { + "learning_rate": 1.4516316803015627e-07, + "loss": 0.0126, + "step": 972, + "train/R_acc": 0.00625, + "train/R_penalty": -0.59375, + "train/R_reason": 0.18895812905862736, + "train/R_vocab": 0.14308035714285713 + }, + { + "learning_rate": 1.441820523985431e-07, + "loss": -0.0349, + "step": 974, + "train/R_acc": 0.0, + "train/R_penalty": -0.515625, + "train/R_reason": 0.2382146965819557, + "train/R_vocab": 0.2640997023809524 + }, + { + "learning_rate": 1.432029182262482e-07, + "loss": 0.0445, + "step": 976, + "train/R_acc": 0.09375, + "train/R_penalty": -0.59375, + "train/R_reason": 0.2085522733677449, + "train/R_vocab": 0.26365327380952386 + }, + { + "learning_rate": 1.42225783847728e-07, + "loss": 0.0309, + "step": 978, + "train/R_acc": 0.0, + "train/R_penalty": -0.640625, + "train/R_reason": 0.27484186892341045, + "train/R_vocab": 0.3541170634920635 + }, + { + "learning_rate": 1.41250667559992e-07, + "loss": 0.0122, + "step": 980, + "train/R_acc": 0.034375, + "train/R_penalty": -0.53125, + "train/R_reason": 0.22480109886784352, + "train/R_vocab": 0.30301339285714285 + }, + { + "learning_rate": 1.4027758762226107e-07, + "loss": -0.0861, + "step": 982, + "train/R_acc": 0.003125, + "train/R_penalty": -0.665625, + "train/R_reason": 0.21184840132463886, + "train/R_vocab": 0.34201388888888884 + }, + { + "learning_rate": 1.3930656225562474e-07, + "loss": 0.0047, + "step": 984, + "train/R_acc": 0.0, + "train/R_penalty": -0.625, + "train/R_reason": 0.2253757708880576, + "train/R_vocab": 0.30904017857142857 + }, + { + "learning_rate": 1.3833760964270058e-07, + "loss": 0.0296, + "step": 986, + "train/R_acc": 0.03125, + "train/R_penalty": -0.546875, + "train/R_reason": 0.2222406698557359, + "train/R_vocab": 0.305406746031746 + }, + { + "learning_rate": 1.3737074792729332e-07, + "loss": 0.0262, + "step": 988, + "train/R_acc": 0.0, + "train/R_penalty": -0.59375, + "train/R_reason": 0.2037288188497659, + "train/R_vocab": 0.17431795634920633 + }, + { + "learning_rate": 1.3640599521405522e-07, + "loss": 0.0671, + "step": 990, + "train/R_acc": 0.040625, + "train/R_penalty": -0.53125, + "train/R_reason": 0.19595219511870288, + "train/R_vocab": 0.2941964285714286 + }, + { + "learning_rate": 1.354433695681474e-07, + "loss": -0.0366, + "step": 992, + "train/R_acc": 0.0, + "train/R_penalty": -0.609375, + "train/R_reason": 0.1895803639379184, + "train/R_vocab": 0.2979910714285714 + }, + { + "learning_rate": 1.3448288901490092e-07, + "loss": 0.0929, + "step": 994, + "train/R_acc": 0.034375, + "train/R_penalty": -0.665625, + "train/R_reason": 0.21089602634201932, + "train/R_vocab": 0.19340277777777776 + }, + { + "learning_rate": 1.3352457153947997e-07, + "loss": 0.0515, + "step": 996, + "train/R_acc": 0.009375000000000001, + "train/R_penalty": -0.546875, + "train/R_reason": 0.21176803496084237, + "train/R_vocab": 0.3035094246031746 + }, + { + "learning_rate": 1.3256843508654444e-07, + "loss": 0.0103, + "step": 998, + "train/R_acc": 0.0, + "train/R_penalty": -0.65625, + "train/R_reason": 0.15247972660677223, + "train/R_vocab": 0.21183035714285717 + }, + { + "learning_rate": 1.3161449755991425e-07, + "loss": -0.082, + "step": 1000, + "train/R_acc": 0.13125, + "train/R_penalty": -0.515625, + "train/R_reason": 0.2110448967120055, + "train/R_vocab": 0.25863095238095235 + }, + { + "eval/R_acc": 0.0025, + "eval/R_penalty": -0.55, + "eval/R_reason": 0.2655294246752871, + "eval/R_vocab": 0.34914772727272725, + "step": 1000 + }, + { + "step": 1000 + }, + { + "learning_rate": 1.3066277682223425e-07, + "loss": -0.0096, + "step": 1002, + "train/R_acc": 0.003125, + "train/R_penalty": -0.640625, + "train/R_reason": 0.181312457358332, + "train/R_vocab": 0.21517857142857144 + }, + { + "learning_rate": 1.2971329069463932e-07, + "loss": -0.0918, + "step": 1004, + "train/R_acc": 0.003125, + "train/R_penalty": -0.75, + "train/R_reason": 0.12020329779967928, + "train/R_vocab": 0.12425595238095238 + }, + { + "learning_rate": 1.2876605695642084e-07, + "loss": -0.0157, + "step": 1006, + "train/R_acc": 0.03125, + "train/R_penalty": -0.625, + "train/R_reason": 0.2288382776444276, + "train/R_vocab": 0.3042410714285715 + }, + { + "learning_rate": 1.2782109334469409e-07, + "loss": 0.0782, + "step": 1008, + "train/R_acc": 0.003125, + "train/R_penalty": -0.578125, + "train/R_reason": 0.23151180047594122, + "train/R_vocab": 0.30937499999999996 + }, + { + "learning_rate": 1.2687841755406554e-07, + "loss": -0.0145, + "step": 1010, + "train/R_acc": 0.003125, + "train/R_penalty": -0.578125, + "train/R_reason": 0.23526236082480567, + "train/R_vocab": 0.30009920634920634 + }, + { + "learning_rate": 1.2593804723630209e-07, + "loss": 0.0254, + "step": 1012, + "train/R_acc": 0.0, + "train/R_penalty": -0.625, + "train/R_reason": 0.2194206796665749, + "train/R_vocab": 0.3340401785714286 + }, + { + "learning_rate": 1.2500000000000005e-07, + "loss": 0.0277, + "step": 1014, + "train/R_acc": 0.003125, + "train/R_penalty": -0.578125, + "train/R_reason": 0.18922638449707688, + "train/R_vocab": 0.3610491071428572 + }, + { + "learning_rate": 1.2406429341025556e-07, + "loss": 0.0668, + "step": 1016, + "train/R_acc": 0.00625, + "train/R_penalty": -0.55625, + "train/R_reason": 0.2346744057550262, + "train/R_vocab": 0.2239955357142857 + }, + { + "learning_rate": 1.231309449883361e-07, + "loss": 0.0338, + "step": 1018, + "train/R_acc": 0.003125, + "train/R_penalty": -0.53125, + "train/R_reason": 0.19074387902354337, + "train/R_vocab": 0.21744791666666666 + }, + { + "learning_rate": 1.2219997221135177e-07, + "loss": 0.0082, + "step": 1020, + "train/R_acc": 0.003125, + "train/R_penalty": -0.609375, + "train/R_reason": 0.24590138406147055, + "train/R_vocab": 0.2600446428571429 + }, + { + "learning_rate": 1.2127139251192823e-07, + "loss": 0.0268, + "step": 1022, + "train/R_acc": 0.0, + "train/R_penalty": -0.53125, + "train/R_reason": 0.24386600100476474, + "train/R_vocab": 0.3195560515873016 + }, + { + "learning_rate": 1.203452232778807e-07, + "loss": 0.0183, + "step": 1024, + "train/R_acc": 0.00625, + "train/R_penalty": -0.671875, + "train/R_reason": 0.171127249096387, + "train/R_vocab": 0.29609375000000004 + }, + { + "learning_rate": 1.1942148185188763e-07, + "loss": 0.0272, + "step": 1026, + "train/R_acc": 0.00625, + "train/R_penalty": -0.5625, + "train/R_reason": 0.17080492182756957, + "train/R_vocab": 0.21568627450980393 + }, + { + "learning_rate": 1.1850018553116676e-07, + "loss": -0.0052, + "step": 1028, + "train/R_acc": 0.003125, + "train/R_penalty": -0.5, + "train/R_reason": 0.2604828445455006, + "train/R_vocab": 0.23158482142857142 + }, + { + "learning_rate": 1.1758135156715041e-07, + "loss": -0.0154, + "step": 1030, + "train/R_acc": 0.065625, + "train/R_penalty": -0.578125, + "train/R_reason": 0.1946716410122738, + "train/R_vocab": 0.22801339285714284 + }, + { + "learning_rate": 1.1666499716516302e-07, + "loss": 0.0101, + "step": 1032, + "train/R_acc": 0.003125, + "train/R_penalty": -0.625, + "train/R_reason": 0.18422370796707505, + "train/R_vocab": 0.21763392857142855 + }, + { + "learning_rate": 1.157511394840989e-07, + "loss": -0.0049, + "step": 1034, + "train/R_acc": 0.00625, + "train/R_penalty": -0.7125, + "train/R_reason": 0.17663980590936243, + "train/R_vocab": 0.24196428571428574 + }, + { + "learning_rate": 1.1483979563610069e-07, + "loss": 0.0098, + "step": 1036, + "train/R_acc": 0.003125, + "train/R_penalty": -0.65625, + "train/R_reason": 0.26512284555646376, + "train/R_vocab": 0.375 + }, + { + "learning_rate": 1.1393098268623902e-07, + "loss": 0.1176, + "step": 1038, + "train/R_acc": 0.0625, + "train/R_penalty": -0.634375, + "train/R_reason": 0.26504285268931027, + "train/R_vocab": 0.30766369047619047 + }, + { + "learning_rate": 1.1302471765219327e-07, + "loss": 0.0428, + "step": 1040, + "train/R_acc": 0.03125, + "train/R_penalty": -0.59375, + "train/R_reason": 0.2008627221719819, + "train/R_vocab": 0.253125 + }, + { + "learning_rate": 1.1212101750393235e-07, + "loss": 0.0175, + "step": 1042, + "train/R_acc": 0.0625, + "train/R_penalty": -0.409375, + "train/R_reason": 0.19803424974179828, + "train/R_vocab": 0.17265625 + }, + { + "learning_rate": 1.1121989916339756e-07, + "loss": 0.0292, + "step": 1044, + "train/R_acc": 0.0, + "train/R_penalty": -0.61875, + "train/R_reason": 0.21600135050493885, + "train/R_vocab": 0.3185267857142857 + }, + { + "learning_rate": 1.1032137950418514e-07, + "loss": 0.0669, + "step": 1046, + "train/R_acc": 0.00625, + "train/R_penalty": -0.59375, + "train/R_reason": 0.23236107648223334, + "train/R_vocab": 0.22142857142857142 + }, + { + "learning_rate": 1.0942547535123057e-07, + "loss": -0.0038, + "step": 1048, + "train/R_acc": 0.00625, + "train/R_penalty": -0.509375, + "train/R_reason": 0.24041357438290228, + "train/R_vocab": 0.2955357142857143 + }, + { + "learning_rate": 1.085322034804938e-07, + "loss": 0.1258, + "step": 1050, + "train/R_acc": 0.003125, + "train/R_penalty": -0.665625, + "train/R_reason": 0.20610390132448592, + "train/R_vocab": 0.24229910714285713 + }, + { + "learning_rate": 1.0764158061864448e-07, + "loss": -0.0495, + "step": 1052, + "train/R_acc": 0.003125, + "train/R_penalty": -0.65, + "train/R_reason": 0.2606824569537362, + "train/R_vocab": 0.39392361111111107 + }, + { + "learning_rate": 1.0675362344274952e-07, + "loss": -0.0224, + "step": 1054, + "train/R_acc": 0.040625, + "train/R_penalty": -0.53125, + "train/R_reason": 0.2362594455371677, + "train/R_vocab": 0.19696631493506495 + }, + { + "learning_rate": 1.0586834857995996e-07, + "loss": 0.0405, + "step": 1056, + "train/R_acc": 0.0, + "train/R_penalty": -0.59375, + "train/R_reason": 0.22383388089744485, + "train/R_vocab": 0.2839285714285714 + }, + { + "learning_rate": 1.0498577260720048e-07, + "loss": 0.006, + "step": 1058, + "train/R_acc": 0.003125, + "train/R_penalty": -0.5625, + "train/R_reason": 0.2154017687323606, + "train/R_vocab": 0.18723958333333335 + }, + { + "learning_rate": 1.0410591205085817e-07, + "loss": 0.0547, + "step": 1060, + "train/R_acc": 0.03125, + "train/R_penalty": -0.64375, + "train/R_reason": 0.22723913109425195, + "train/R_vocab": 0.32499999999999996 + }, + { + "learning_rate": 1.0322878338647392e-07, + "loss": 0.1005, + "step": 1062, + "train/R_acc": 0.0, + "train/R_penalty": -0.65, + "train/R_reason": 0.21853491707209402, + "train/R_vocab": 0.19251217532467532 + }, + { + "learning_rate": 1.0235440303843302e-07, + "loss": 0.0902, + "step": 1064, + "train/R_acc": 0.03125, + "train/R_penalty": -0.546875, + "train/R_reason": 0.21816371211780805, + "train/R_vocab": 0.22801339285714284 + }, + { + "learning_rate": 1.0148278737965844e-07, + "loss": -0.0553, + "step": 1066, + "train/R_acc": 0.0, + "train/R_penalty": -0.578125, + "train/R_reason": 0.19269070794229914, + "train/R_vocab": 0.18387896825396824 + }, + { + "learning_rate": 1.0061395273130368e-07, + "loss": 0.039, + "step": 1068, + "train/R_acc": 0.0, + "train/R_penalty": -0.65625, + "train/R_reason": 0.21827443734801666, + "train/R_vocab": 0.2911830357142857 + }, + { + "learning_rate": 9.974791536244726e-08, + "loss": 0.0343, + "step": 1070, + "train/R_acc": 0.0, + "train/R_penalty": -0.61875, + "train/R_reason": 0.2733091389175897, + "train/R_vocab": 0.3764136904761905 + }, + { + "learning_rate": 9.888469148978843e-08, + "loss": -0.0473, + "step": 1072, + "train/R_acc": 0.003125, + "train/R_penalty": -0.484375, + "train/R_reason": 0.24157353834104733, + "train/R_vocab": 0.3257496843434343 + }, + { + "learning_rate": 9.802429727734294e-08, + "loss": -0.0036, + "step": 1074, + "train/R_acc": 0.040625, + "train/R_penalty": -0.421875, + "train/R_reason": 0.2823229532013337, + "train/R_vocab": 0.43273809523809526 + }, + { + "learning_rate": 9.716674883614092e-08, + "loss": 0.0124, + "step": 1076, + "train/R_acc": 0.037500000000000006, + "train/R_penalty": -0.453125, + "train/R_reason": 0.2879044257474095, + "train/R_vocab": 0.33738839285714284 + }, + { + "learning_rate": 9.631206222392479e-08, + "loss": -0.0157, + "step": 1078, + "train/R_acc": 0.003125, + "train/R_penalty": -0.546875, + "train/R_reason": 0.26644330439869424, + "train/R_vocab": 0.27585565476190477 + }, + { + "learning_rate": 9.546025344484868e-08, + "loss": 0.0013, + "step": 1080, + "train/R_acc": 0.003125, + "train/R_penalty": -0.65625, + "train/R_reason": 0.17772787504592696, + "train/R_vocab": 0.22031250000000002 + }, + { + "learning_rate": 9.461133844917901e-08, + "loss": -0.0364, + "step": 1082, + "train/R_acc": 0.0625, + "train/R_penalty": -0.696875, + "train/R_reason": 0.14647117758570943, + "train/R_vocab": 0.19616815476190477 + }, + { + "learning_rate": 9.376533313299542e-08, + "loss": -0.0051, + "step": 1084, + "train/R_acc": 0.003125, + "train/R_penalty": -0.578125, + "train/R_reason": 0.18199736371919578, + "train/R_vocab": 0.2505580357142857 + }, + { + "learning_rate": 9.292225333789327e-08, + "loss": 0.0575, + "step": 1086, + "train/R_acc": 0.03125, + "train/R_penalty": -0.603125, + "train/R_reason": 0.1892425633255841, + "train/R_vocab": 0.31540178571428573 + }, + { + "learning_rate": 9.208211485068729e-08, + "loss": 0.0395, + "step": 1088, + "train/R_acc": 0.003125, + "train/R_penalty": -0.578125, + "train/R_reason": 0.15220888566600485, + "train/R_vocab": 0.21587301587301586 + }, + { + "learning_rate": 9.124493340311537e-08, + "loss": -0.016, + "step": 1090, + "train/R_acc": 0.0, + "train/R_penalty": -0.578125, + "train/R_reason": 0.18263365867722547, + "train/R_vocab": 0.25424107142857144 + }, + { + "learning_rate": 9.041072467154471e-08, + "loss": -0.0291, + "step": 1092, + "train/R_acc": 0.0, + "train/R_penalty": -0.375, + "train/R_reason": 0.25466359857700704, + "train/R_vocab": 0.34862351190476193 + }, + { + "learning_rate": 8.957950427667751e-08, + "loss": 0.0296, + "step": 1094, + "train/R_acc": 0.03125, + "train/R_penalty": -0.53125, + "train/R_reason": 0.23575074339334445, + "train/R_vocab": 0.2983010912698413 + }, + { + "learning_rate": 8.875128778325902e-08, + "loss": -0.0389, + "step": 1096, + "train/R_acc": 0.0, + "train/R_penalty": -0.53125, + "train/R_reason": 0.1681449082491554, + "train/R_vocab": 0.23660714285714285 + }, + { + "learning_rate": 8.792609069978602e-08, + "loss": 0.0201, + "step": 1098, + "train/R_acc": 0.0125, + "train/R_penalty": -0.765625, + "train/R_reason": 0.24261987678361474, + "train/R_vocab": 0.23020833333333335 + }, + { + "learning_rate": 8.710392847821615e-08, + "loss": -0.0508, + "step": 1100, + "train/R_acc": 0.03125, + "train/R_penalty": -0.53125, + "train/R_reason": 0.19600584280495498, + "train/R_vocab": 0.24397321428571428 + }, + { + "eval/R_acc": 0.0025, + "eval/R_penalty": -0.4875, + "eval/R_reason": 0.24036428951274935, + "eval/R_vocab": 0.29469967532467534, + "step": 1100 + }, + { + "step": 1100 + }, + { + "learning_rate": 8.628481651367875e-08, + "loss": 0.0336, + "step": 1102, + "train/R_acc": 0.003125, + "train/R_penalty": -0.609375, + "train/R_reason": 0.22866834230086636, + "train/R_vocab": 0.19769345238095237 + }, + { + "learning_rate": 8.546877014418671e-08, + "loss": -0.0387, + "step": 1104, + "train/R_acc": 0.03125, + "train/R_penalty": -0.609375, + "train/R_reason": 0.21012454085545956, + "train/R_vocab": 0.23225446428571428 + }, + { + "learning_rate": 8.465580465034894e-08, + "loss": 0.0231, + "step": 1106, + "train/R_acc": 0.03125, + "train/R_penalty": -0.609375, + "train/R_reason": 0.20315946724213801, + "train/R_vocab": 0.2909598214285714 + }, + { + "learning_rate": 8.384593525508466e-08, + "loss": 0.0343, + "step": 1108, + "train/R_acc": 0.0, + "train/R_penalty": -0.509375, + "train/R_reason": 0.20224091353679885, + "train/R_vocab": 0.19077380952380954 + }, + { + "learning_rate": 8.303917712333788e-08, + "loss": 0.0208, + "step": 1110, + "train/R_acc": 0.065625, + "train/R_penalty": -0.5625, + "train/R_reason": 0.23770102329722845, + "train/R_vocab": 0.21218998015873014 + }, + { + "learning_rate": 8.223554536179367e-08, + "loss": 0.0016, + "step": 1112, + "train/R_acc": 0.0, + "train/R_penalty": -0.5, + "train/R_reason": 0.24531693423546427, + "train/R_vocab": 0.3191964285714286 + }, + { + "learning_rate": 8.143505501859551e-08, + "loss": 0.0352, + "step": 1114, + "train/R_acc": 0.0, + "train/R_penalty": -0.49375, + "train/R_reason": 0.2098236201971303, + "train/R_vocab": 0.2838169642857143 + }, + { + "learning_rate": 8.06377210830631e-08, + "loss": 0.0216, + "step": 1116, + "train/R_acc": 0.003125, + "train/R_penalty": -0.640625, + "train/R_reason": 0.20338203537474456, + "train/R_vocab": 0.3302455357142857 + }, + { + "learning_rate": 7.984355848541175e-08, + "loss": -0.0481, + "step": 1118, + "train/R_acc": 0.003125, + "train/R_penalty": -0.53125, + "train/R_reason": 0.17974317204155837, + "train/R_vocab": 0.32105654761904767 + }, + { + "learning_rate": 7.905258209647325e-08, + "loss": 0.0544, + "step": 1120, + "train/R_acc": 0.03125, + "train/R_penalty": -0.603125, + "train/R_reason": 0.24957606202713686, + "train/R_vocab": 0.30379464285714286 + }, + { + "learning_rate": 7.826480672741676e-08, + "loss": -0.0196, + "step": 1122, + "train/R_acc": 0.0625, + "train/R_penalty": -0.5625, + "train/R_reason": 0.21989961574561373, + "train/R_vocab": 0.24270833333333333 + }, + { + "learning_rate": 7.748024712947204e-08, + "loss": 0.0036, + "step": 1124, + "train/R_acc": 0.003125, + "train/R_penalty": -0.53125, + "train/R_reason": 0.20629454131030475, + "train/R_vocab": 0.2872767857142857 + }, + { + "learning_rate": 7.669891799365282e-08, + "loss": 0.0386, + "step": 1126, + "train/R_acc": 0.00625, + "train/R_penalty": -0.671875, + "train/R_reason": 0.22259933007352922, + "train/R_vocab": 0.36199156746031746 + }, + { + "learning_rate": 7.592083395048182e-08, + "loss": 0.0643, + "step": 1128, + "train/R_acc": 0.03125, + "train/R_penalty": -0.49375, + "train/R_reason": 0.18965226315756645, + "train/R_vocab": 0.2122767857142857 + }, + { + "learning_rate": 7.514600956971712e-08, + "loss": 0.0678, + "step": 1130, + "train/R_acc": 0.0, + "train/R_penalty": -0.5875, + "train/R_reason": 0.23261509390210205, + "train/R_vocab": 0.23307291666666669 + }, + { + "learning_rate": 7.437445936007867e-08, + "loss": 0.01, + "step": 1132, + "train/R_acc": 0.03125, + "train/R_penalty": -0.59375, + "train/R_reason": 0.2093342551771361, + "train/R_vocab": 0.2585565476190476 + }, + { + "learning_rate": 7.360619776897742e-08, + "loss": -0.0392, + "step": 1134, + "train/R_acc": 0.0, + "train/R_penalty": -0.525, + "train/R_reason": 0.22079774921664902, + "train/R_vocab": 0.27708333333333335 + }, + { + "learning_rate": 7.284123918224397e-08, + "loss": 0.0161, + "step": 1136, + "train/R_acc": 0.03125, + "train/R_penalty": -0.59375, + "train/R_reason": 0.22520667797905555, + "train/R_vocab": 0.3407118055555556 + }, + { + "learning_rate": 7.207959792385998e-08, + "loss": 0.0479, + "step": 1138, + "train/R_acc": 0.00625, + "train/R_penalty": -0.515625, + "train/R_reason": 0.19075869235061965, + "train/R_vocab": 0.2729166666666667 + }, + { + "learning_rate": 7.132128825568914e-08, + "loss": -0.0027, + "step": 1140, + "train/R_acc": 0.03125, + "train/R_penalty": -0.634375, + "train/R_reason": 0.23292150900288705, + "train/R_vocab": 0.21964285714285714 + }, + { + "learning_rate": 7.056632437721099e-08, + "loss": -0.0379, + "step": 1142, + "train/R_acc": 0.0625, + "train/R_penalty": -0.578125, + "train/R_reason": 0.17068707794196553, + "train/R_vocab": 0.23448660714285716 + }, + { + "learning_rate": 6.981472042525416e-08, + "loss": -0.0097, + "step": 1144, + "train/R_acc": 0.003125, + "train/R_penalty": -0.484375, + "train/R_reason": 0.1982660319129968, + "train/R_vocab": 0.18515624999999997 + }, + { + "learning_rate": 6.906649047373245e-08, + "loss": -0.0136, + "step": 1146, + "train/R_acc": 0.0, + "train/R_penalty": -0.571875, + "train/R_reason": 0.18697098939403184, + "train/R_vocab": 0.22890624999999998 + }, + { + "learning_rate": 6.832164853338066e-08, + "loss": 0.0075, + "step": 1148, + "train/R_acc": 0.065625, + "train/R_penalty": -0.515625, + "train/R_reason": 0.26961296459836603, + "train/R_vocab": 0.3282118055555555 + }, + { + "learning_rate": 6.758020855149249e-08, + "loss": -0.0081, + "step": 1150, + "train/R_acc": 0.03125, + "train/R_penalty": -0.55625, + "train/R_reason": 0.24886312239764444, + "train/R_vocab": 0.3654017857142857 + }, + { + "learning_rate": 6.684218441165962e-08, + "loss": 0.0666, + "step": 1152, + "train/R_acc": 0.0, + "train/R_penalty": -0.61875, + "train/R_reason": 0.2088855316383599, + "train/R_vocab": 0.22730654761904762 + }, + { + "learning_rate": 6.610758993351123e-08, + "loss": 0.0395, + "step": 1154, + "train/R_acc": 0.034375, + "train/R_penalty": -0.53125, + "train/R_reason": 0.22720806482016326, + "train/R_vocab": 0.24931795634920634 + }, + { + "learning_rate": 6.537643887245573e-08, + "loss": 0.0465, + "step": 1156, + "train/R_acc": 0.03125, + "train/R_penalty": -0.515625, + "train/R_reason": 0.1968795889130503, + "train/R_vocab": 0.2591517857142857 + }, + { + "learning_rate": 6.464874491942279e-08, + "loss": 0.0203, + "step": 1158, + "train/R_acc": 0.0, + "train/R_penalty": -0.665625, + "train/R_reason": 0.19558365060317057, + "train/R_vocab": 0.1140625 + }, + { + "learning_rate": 6.392452170060706e-08, + "loss": 0.0173, + "step": 1160, + "train/R_acc": 0.03125, + "train/R_penalty": -0.540625, + "train/R_reason": 0.25112434395380084, + "train/R_vocab": 0.23022073412698413 + }, + { + "learning_rate": 6.320378277721342e-08, + "loss": -0.0386, + "step": 1162, + "train/R_acc": 0.003125, + "train/R_penalty": -0.478125, + "train/R_reason": 0.24359373155022102, + "train/R_vocab": 0.3531498015873016 + }, + { + "learning_rate": 6.248654164520237e-08, + "loss": -0.0419, + "step": 1164, + "train/R_acc": 0.003125, + "train/R_penalty": -0.71875, + "train/R_reason": 0.20550711716499057, + "train/R_vocab": 0.3539434523809524 + }, + { + "learning_rate": 6.177281173503778e-08, + "loss": 0.0405, + "step": 1166, + "train/R_acc": 0.003125, + "train/R_penalty": -0.640625, + "train/R_reason": 0.20957454414924856, + "train/R_vocab": 0.2280133928571429 + }, + { + "learning_rate": 6.106260641143546e-08, + "loss": -0.0833, + "step": 1168, + "train/R_acc": 0.0, + "train/R_penalty": -0.546875, + "train/R_reason": 0.20663920365929042, + "train/R_vocab": 0.21852678571428574 + }, + { + "learning_rate": 6.035593897311239e-08, + "loss": -0.0156, + "step": 1170, + "train/R_acc": 0.003125, + "train/R_penalty": -0.61875, + "train/R_reason": 0.2371425092299592, + "train/R_vocab": 0.3323784722222222 + }, + { + "learning_rate": 5.965282265253838e-08, + "loss": 0.0651, + "step": 1172, + "train/R_acc": 0.034375, + "train/R_penalty": -0.390625, + "train/R_reason": 0.22832625758068797, + "train/R_vocab": 0.30301339285714285 + }, + { + "learning_rate": 5.895327061568775e-08, + "loss": -0.0379, + "step": 1174, + "train/R_acc": 0.03125, + "train/R_penalty": -0.578125, + "train/R_reason": 0.19140193598424352, + "train/R_vocab": 0.2755952380952381 + }, + { + "learning_rate": 5.8257295961792936e-08, + "loss": -0.1099, + "step": 1176, + "train/R_acc": 0.0, + "train/R_penalty": -0.640625, + "train/R_reason": 0.2013441433056279, + "train/R_vocab": 0.2722222222222222 + }, + { + "learning_rate": 5.756491172309952e-08, + "loss": -0.0236, + "step": 1178, + "train/R_acc": 0.003125, + "train/R_penalty": -0.53125, + "train/R_reason": 0.2409798833837463, + "train/R_vocab": 0.2924107142857143 + }, + { + "learning_rate": 5.68761308646217e-08, + "loss": -0.0326, + "step": 1180, + "train/R_acc": 0.003125, + "train/R_penalty": -0.665625, + "train/R_reason": 0.16268249986159153, + "train/R_vocab": 0.17838541666666669 + }, + { + "learning_rate": 5.6190966283899774e-08, + "loss": 0.0286, + "step": 1182, + "train/R_acc": 0.003125, + "train/R_penalty": -0.65625, + "train/R_reason": 0.2404998117876481, + "train/R_vocab": 0.34227430555555555 + }, + { + "learning_rate": 5.5509430810758817e-08, + "loss": 0.004, + "step": 1184, + "train/R_acc": 0.0, + "train/R_penalty": -0.609375, + "train/R_reason": 0.22824560961553947, + "train/R_vocab": 0.3375 + }, + { + "learning_rate": 5.483153720706798e-08, + "loss": -0.0074, + "step": 1186, + "train/R_acc": 0.1, + "train/R_penalty": -0.53125, + "train/R_reason": 0.21326554603657105, + "train/R_vocab": 0.20881696428571428 + }, + { + "learning_rate": 5.415729816650202e-08, + "loss": 0.0246, + "step": 1188, + "train/R_acc": 0.0, + "train/R_penalty": -0.609375, + "train/R_reason": 0.2444787127154557, + "train/R_vocab": 0.35230654761904756 + }, + { + "learning_rate": 5.3486726314303175e-08, + "loss": 0.0987, + "step": 1190, + "train/R_acc": 0.003125, + "train/R_penalty": -0.634375, + "train/R_reason": 0.22288328491646994, + "train/R_vocab": 0.2479910714285714 + }, + { + "learning_rate": 5.2819834207044974e-08, + "loss": -0.0208, + "step": 1192, + "train/R_acc": 0.128125, + "train/R_penalty": -0.4375, + "train/R_reason": 0.27609449087124366, + "train/R_vocab": 0.3087425595238095 + }, + { + "learning_rate": 5.215663433239728e-08, + "loss": -0.006, + "step": 1194, + "train/R_acc": 0.034375, + "train/R_penalty": -0.609375, + "train/R_reason": 0.22109386819415358, + "train/R_vocab": 0.24497767857142858 + }, + { + "learning_rate": 5.149713910889206e-08, + "loss": -0.0461, + "step": 1196, + "train/R_acc": 0.034375, + "train/R_penalty": -0.421875, + "train/R_reason": 0.2398768408665738, + "train/R_vocab": 0.3560267857142857 + }, + { + "learning_rate": 5.0841360885690996e-08, + "loss": 0.0184, + "step": 1198, + "train/R_acc": 0.037500000000000006, + "train/R_penalty": -0.5625, + "train/R_reason": 0.21147620601357403, + "train/R_vocab": 0.20399305555555555 + }, + { + "learning_rate": 5.01893119423546e-08, + "loss": 0.0505, + "step": 1200, + "train/R_acc": 0.065625, + "train/R_penalty": -0.671875, + "train/R_reason": 0.2487143037586896, + "train/R_vocab": 0.3544642857142857 + }, + { + "eval/R_acc": 0.0, + "eval/R_penalty": -0.5875, + "eval/R_reason": 0.20266877330422212, + "eval/R_vocab": 0.2709090909090909, + "step": 1200 + }, + { + "step": 1200 + }, + { + "learning_rate": 4.954100448861165e-08, + "loss": -0.009, + "step": 1202, + "train/R_acc": 0.034375, + "train/R_penalty": -0.446875, + "train/R_reason": 0.2773491666866028, + "train/R_vocab": 0.29509943181818177 + }, + { + "learning_rate": 4.889645066413112e-08, + "loss": 0.0306, + "step": 1204, + "train/R_acc": 0.003125, + "train/R_penalty": -0.734375, + "train/R_reason": 0.2607235600350966, + "train/R_vocab": 0.1830729166666667 + }, + { + "learning_rate": 4.82556625382945e-08, + "loss": -0.0188, + "step": 1206, + "train/R_acc": 0.03125, + "train/R_penalty": -0.665625, + "train/R_reason": 0.20154795093112635, + "train/R_vocab": 0.19722222222222224 + }, + { + "learning_rate": 4.76186521099699e-08, + "loss": -0.0385, + "step": 1208, + "train/R_acc": 0.0, + "train/R_penalty": -0.625, + "train/R_reason": 0.2240002350916589, + "train/R_vocab": 0.3001116071428571 + }, + { + "learning_rate": 4.698543130728755e-08, + "loss": 0.0571, + "step": 1210, + "train/R_acc": 0.0, + "train/R_penalty": -0.515625, + "train/R_reason": 0.15973930176485518, + "train/R_vocab": 0.21350446428571426 + }, + { + "learning_rate": 4.635601198741607e-08, + "loss": 0.0042, + "step": 1212, + "train/R_acc": 0.06875, + "train/R_penalty": -0.5, + "train/R_reason": 0.19491433422937576, + "train/R_vocab": 0.24547483766233766 + }, + { + "learning_rate": 4.573040593634092e-08, + "loss": 0.0204, + "step": 1214, + "train/R_acc": 0.0, + "train/R_penalty": -0.7125, + "train/R_reason": 0.2036836381635984, + "train/R_vocab": 0.2583333333333333 + }, + { + "learning_rate": 4.510862486864317e-08, + "loss": -0.0139, + "step": 1216, + "train/R_acc": 0.03125, + "train/R_penalty": -0.5, + "train/R_reason": 0.1935253778055071, + "train/R_vocab": 0.2539434523809524 + }, + { + "learning_rate": 4.449068042728063e-08, + "loss": -0.0221, + "step": 1218, + "train/R_acc": 0.03125, + "train/R_penalty": -0.671875, + "train/R_reason": 0.17478719789421077, + "train/R_vocab": 0.2385044642857143 + }, + { + "learning_rate": 4.3876584183369454e-08, + "loss": 0.0401, + "step": 1220, + "train/R_acc": 0.0, + "train/R_penalty": -0.65625, + "train/R_reason": 0.12396243023486506, + "train/R_vocab": 0.16953125 + }, + { + "learning_rate": 4.326634763596784e-08, + "loss": 0.0362, + "step": 1222, + "train/R_acc": 0.003125, + "train/R_penalty": -0.634375, + "train/R_reason": 0.21970153796136732, + "train/R_vocab": 0.23229166666666667 + }, + { + "learning_rate": 4.265998221186023e-08, + "loss": 0.0579, + "step": 1224, + "train/R_acc": 0.003125, + "train/R_penalty": -0.696875, + "train/R_reason": 0.25118929349410624, + "train/R_vocab": 0.3527901785714286 + }, + { + "learning_rate": 4.205749926534391e-08, + "loss": 0.0325, + "step": 1226, + "train/R_acc": 0.0, + "train/R_penalty": -0.53125, + "train/R_reason": 0.20960885628620934, + "train/R_vocab": 0.24062499999999998 + }, + { + "learning_rate": 4.145891007801588e-08, + "loss": -0.012, + "step": 1228, + "train/R_acc": 0.037500000000000006, + "train/R_penalty": -0.375, + "train/R_reason": 0.2440708995563252, + "train/R_vocab": 0.31674107142857144 + }, + { + "learning_rate": 4.086422585856189e-08, + "loss": 0.0635, + "step": 1230, + "train/R_acc": 0.03125, + "train/R_penalty": -0.5, + "train/R_reason": 0.16689617467212345, + "train/R_vocab": 0.24068700396825393 + }, + { + "learning_rate": 4.0273457742546566e-08, + "loss": -0.0426, + "step": 1232, + "train/R_acc": 0.003125, + "train/R_penalty": -0.546875, + "train/R_reason": 0.192806526715886, + "train/R_vocab": 0.2780505952380953 + }, + { + "learning_rate": 3.968661679220467e-08, + "loss": 0.0384, + "step": 1234, + "train/R_acc": 0.0, + "train/R_penalty": -0.546875, + "train/R_reason": 0.17224863139672328, + "train/R_vocab": 0.23426339285714287 + }, + { + "learning_rate": 3.9103713996234355e-08, + "loss": 0.0036, + "step": 1236, + "train/R_acc": 0.03125, + "train/R_penalty": -0.734375, + "train/R_reason": 0.17680126111067432, + "train/R_vocab": 0.2612723214285714 + }, + { + "learning_rate": 3.852476026959089e-08, + "loss": 0.0172, + "step": 1238, + "train/R_acc": 0.009375000000000001, + "train/R_penalty": -0.640625, + "train/R_reason": 0.23625775975212537, + "train/R_vocab": 0.2917410714285715 + }, + { + "learning_rate": 3.794976645328266e-08, + "loss": 0.0608, + "step": 1240, + "train/R_acc": 0.00625, + "train/R_penalty": -0.49375, + "train/R_reason": 0.20085612003970227, + "train/R_vocab": 0.29375 + }, + { + "learning_rate": 3.737874331416807e-08, + "loss": -0.0099, + "step": 1242, + "train/R_acc": 0.06875, + "train/R_penalty": -0.53125, + "train/R_reason": 0.2067606031031175, + "train/R_vocab": 0.1792782738095238 + }, + { + "learning_rate": 3.681170154475391e-08, + "loss": -0.0105, + "step": 1244, + "train/R_acc": 0.003125, + "train/R_penalty": -0.609375, + "train/R_reason": 0.24829657601476124, + "train/R_vocab": 0.3546875 + }, + { + "learning_rate": 3.624865176299499e-08, + "loss": -0.0273, + "step": 1246, + "train/R_acc": 0.003125, + "train/R_penalty": -0.484375, + "train/R_reason": 0.2728674749829552, + "train/R_vocab": 0.42109375000000004 + }, + { + "learning_rate": 3.568960451209574e-08, + "loss": 0.0602, + "step": 1248, + "train/R_acc": 0.0375, + "train/R_penalty": -0.546875, + "train/R_reason": 0.2555940677098582, + "train/R_vocab": 0.41767113095238095 + }, + { + "learning_rate": 3.513457026031216e-08, + "loss": -0.0627, + "step": 1250, + "train/R_acc": 0.0, + "train/R_penalty": -0.53125, + "train/R_reason": 0.20621535738161093, + "train/R_vocab": 0.16438492063492063 + }, + { + "learning_rate": 3.458355940075652e-08, + "loss": 0.0626, + "step": 1252, + "train/R_acc": 0.03125, + "train/R_penalty": -0.53125, + "train/R_reason": 0.25723943167289565, + "train/R_vocab": 0.290625 + }, + { + "learning_rate": 3.403658225120212e-08, + "loss": -0.0653, + "step": 1254, + "train/R_acc": 0.03125, + "train/R_penalty": -0.53125, + "train/R_reason": 0.2280530069453393, + "train/R_vocab": 0.26473214285714286 + }, + { + "learning_rate": 3.349364905389032e-08, + "loss": -0.013, + "step": 1256, + "train/R_acc": 0.034375, + "train/R_penalty": -0.4375, + "train/R_reason": 0.20137595355438587, + "train/R_vocab": 0.3265625 + }, + { + "learning_rate": 3.295476997533905e-08, + "loss": 0.0298, + "step": 1258, + "train/R_acc": 0.034375, + "train/R_penalty": -0.59375, + "train/R_reason": 0.20510172934689136, + "train/R_vocab": 0.3380580357142857 + }, + { + "learning_rate": 3.2419955106151893e-08, + "loss": -0.0064, + "step": 1260, + "train/R_acc": 0.00625, + "train/R_penalty": -0.53125, + "train/R_reason": 0.24581030179992447, + "train/R_vocab": 0.27968750000000003 + }, + { + "learning_rate": 3.188921446082946e-08, + "loss": 0.0136, + "step": 1262, + "train/R_acc": 0.0, + "train/R_penalty": -0.5, + "train/R_reason": 0.19498686377833271, + "train/R_vocab": 0.21171875 + }, + { + "learning_rate": 3.1362557977582e-08, + "loss": 0.0327, + "step": 1264, + "train/R_acc": 0.003125, + "train/R_penalty": -0.65625, + "train/R_reason": 0.2546377151605037, + "train/R_vocab": 0.38132440476190477 + }, + { + "learning_rate": 3.0839995518142844e-08, + "loss": 0.0813, + "step": 1266, + "train/R_acc": 0.009375000000000001, + "train/R_penalty": -0.578125, + "train/R_reason": 0.23407475867218072, + "train/R_vocab": 0.25262896825396824 + }, + { + "learning_rate": 3.032153686758432e-08, + "loss": 0.0111, + "step": 1268, + "train/R_acc": 0.0, + "train/R_penalty": -0.75, + "train/R_reason": 0.13893382364177903, + "train/R_vocab": 0.1873139880952381 + }, + { + "learning_rate": 2.980719173413396e-08, + "loss": -0.0655, + "step": 1270, + "train/R_acc": 0.003125, + "train/R_penalty": -0.546875, + "train/R_reason": 0.241975792190938, + "train/R_vocab": 0.359375 + }, + { + "learning_rate": 2.9296969748993093e-08, + "loss": 0.0107, + "step": 1272, + "train/R_acc": 0.03125, + "train/R_penalty": -0.53125, + "train/R_reason": 0.24308125423194502, + "train/R_vocab": 0.18370535714285716 + }, + { + "learning_rate": 2.8790880466156464e-08, + "loss": 0.0368, + "step": 1274, + "train/R_acc": 0.0, + "train/R_penalty": -0.571875, + "train/R_reason": 0.21833878646152022, + "train/R_vocab": 0.2832217261904762 + }, + { + "learning_rate": 2.8288933362233148e-08, + "loss": 0.0016, + "step": 1276, + "train/R_acc": 0.0, + "train/R_penalty": -0.609375, + "train/R_reason": 0.205687843726902, + "train/R_vocab": 0.23705357142857142 + }, + { + "learning_rate": 2.7791137836269158e-08, + "loss": -0.0029, + "step": 1278, + "train/R_acc": 0.003125, + "train/R_penalty": -0.640625, + "train/R_reason": 0.2181107503992247, + "train/R_vocab": 0.20520833333333333 + }, + { + "learning_rate": 2.7297503209571665e-08, + "loss": 0.1301, + "step": 1280, + "train/R_acc": 0.0, + "train/R_penalty": -0.578125, + "train/R_reason": 0.16072625927287004, + "train/R_vocab": 0.22611607142857143 + }, + { + "learning_rate": 2.680803872553408e-08, + "loss": 0.0204, + "step": 1282, + "train/R_acc": 0.00625, + "train/R_penalty": -0.525, + "train/R_reason": 0.18120856948997982, + "train/R_vocab": 0.2267981150793651 + }, + { + "learning_rate": 2.632275354946342e-08, + "loss": 0.0174, + "step": 1284, + "train/R_acc": 0.009375000000000001, + "train/R_penalty": -0.55625, + "train/R_reason": 0.19887443410752542, + "train/R_vocab": 0.20744047619047618 + }, + { + "learning_rate": 2.5841656768408216e-08, + "loss": 0.0125, + "step": 1286, + "train/R_acc": 0.003125, + "train/R_penalty": -0.53125, + "train/R_reason": 0.23872779990823945, + "train/R_vocab": 0.28046875 + }, + { + "learning_rate": 2.5364757390988616e-08, + "loss": 0.0652, + "step": 1288, + "train/R_acc": 0.003125, + "train/R_penalty": -0.546875, + "train/R_reason": 0.19387491540524032, + "train/R_vocab": 0.23203125 + }, + { + "learning_rate": 2.489206434722785e-08, + "loss": -0.012, + "step": 1290, + "train/R_acc": 0.0, + "train/R_penalty": -0.515625, + "train/R_reason": 0.18941429133549656, + "train/R_vocab": 0.23415178571428572 + }, + { + "learning_rate": 2.4423586488384524e-08, + "loss": 0.0179, + "step": 1292, + "train/R_acc": 0.0, + "train/R_penalty": -0.578125, + "train/R_reason": 0.248278845576544, + "train/R_vocab": 0.22329545454545455 + }, + { + "learning_rate": 2.395933258678745e-08, + "loss": 0.0178, + "step": 1294, + "train/R_acc": 0.0, + "train/R_penalty": -0.609375, + "train/R_reason": 0.20548558553050647, + "train/R_vocab": 0.3214905753968254 + }, + { + "learning_rate": 2.349931133567096e-08, + "loss": -0.0004, + "step": 1296, + "train/R_acc": 0.0, + "train/R_penalty": -0.578125, + "train/R_reason": 0.2386328952490884, + "train/R_vocab": 0.23169642857142858 + }, + { + "learning_rate": 2.3043531349012353e-08, + "loss": -0.0659, + "step": 1298, + "train/R_acc": 0.034375, + "train/R_penalty": -0.671875, + "train/R_reason": 0.15007159038519916, + "train/R_vocab": 0.19921875 + }, + { + "learning_rate": 2.259200116137039e-08, + "loss": 0.0095, + "step": 1300, + "train/R_acc": 0.0, + "train/R_penalty": -0.65625, + "train/R_reason": 0.20368911805265966, + "train/R_vocab": 0.2474330357142857 + }, + { + "eval/R_acc": 0.0025, + "eval/R_penalty": -0.5625, + "eval/R_reason": 0.22623212762462228, + "eval/R_vocab": 0.3164285714285714, + "step": 1300 + }, + { + "step": 1300 + } + ], + "logging_steps": 2, + "max_steps": 1497, + "num_input_tokens_seen": 13235132, + "num_train_epochs": 1, + "save_steps": 50, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 0.0, + "train_batch_size": 16, + "trial_name": null, + "trial_params": null +}