diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,4578 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 2.990024577128813, + "eval_steps": 500, + "global_step": 648, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.004626283070695388, + "grad_norm": 0.6155030131340027, + "learning_rate": 9.98456790123457e-06, + "loss": 17.5936, + "step": 1 + }, + { + "epoch": 0.009252566141390776, + "grad_norm": 0.6313583254814148, + "learning_rate": 9.969135802469136e-06, + "loss": 21.7773, + "step": 2 + }, + { + "epoch": 0.013878849212086165, + "grad_norm": 0.4570765793323517, + "learning_rate": 9.953703703703704e-06, + "loss": 17.1664, + "step": 3 + }, + { + "epoch": 0.018505132282781552, + "grad_norm": 0.3836732506752014, + "learning_rate": 9.938271604938273e-06, + "loss": 18.3922, + "step": 4 + }, + { + "epoch": 0.02313141535347694, + "grad_norm": 0.3599857985973358, + "learning_rate": 9.92283950617284e-06, + "loss": 16.9993, + "step": 5 + }, + { + "epoch": 0.02775769842417233, + "grad_norm": 0.4342029094696045, + "learning_rate": 9.907407407407408e-06, + "loss": 17.4883, + "step": 6 + }, + { + "epoch": 0.032383981494867715, + "grad_norm": 0.40308186411857605, + "learning_rate": 9.891975308641975e-06, + "loss": 18.2562, + "step": 7 + }, + { + "epoch": 0.037010264565563104, + "grad_norm": 0.39312881231307983, + "learning_rate": 9.876543209876543e-06, + "loss": 17.9318, + "step": 8 + }, + { + "epoch": 0.04163654763625849, + "grad_norm": 0.6749095916748047, + "learning_rate": 9.861111111111112e-06, + "loss": 18.5244, + "step": 9 + }, + { + "epoch": 0.04626283070695388, + "grad_norm": 0.3622874617576599, + "learning_rate": 9.84567901234568e-06, + "loss": 16.4386, + "step": 10 + }, + { + "epoch": 0.05088911377764927, + "grad_norm": 0.24345389008522034, + "learning_rate": 9.830246913580249e-06, + "loss": 17.5414, + "step": 11 + }, + { + "epoch": 0.05551539684834466, + "grad_norm": 0.34016355872154236, + "learning_rate": 9.814814814814815e-06, + "loss": 16.9405, + "step": 12 + }, + { + "epoch": 0.06014167991904005, + "grad_norm": 0.25097739696502686, + "learning_rate": 9.799382716049384e-06, + "loss": 14.4692, + "step": 13 + }, + { + "epoch": 0.06476796298973543, + "grad_norm": 0.2524467408657074, + "learning_rate": 9.783950617283952e-06, + "loss": 16.2185, + "step": 14 + }, + { + "epoch": 0.06939424606043082, + "grad_norm": 0.3705744743347168, + "learning_rate": 9.768518518518519e-06, + "loss": 16.3573, + "step": 15 + }, + { + "epoch": 0.07402052913112621, + "grad_norm": 0.2848168909549713, + "learning_rate": 9.753086419753087e-06, + "loss": 16.331, + "step": 16 + }, + { + "epoch": 0.0786468122018216, + "grad_norm": 0.300182044506073, + "learning_rate": 9.737654320987654e-06, + "loss": 15.6623, + "step": 17 + }, + { + "epoch": 0.08327309527251699, + "grad_norm": 0.36958807706832886, + "learning_rate": 9.722222222222223e-06, + "loss": 18.5857, + "step": 18 + }, + { + "epoch": 0.08789937834321238, + "grad_norm": 0.3009611666202545, + "learning_rate": 9.706790123456791e-06, + "loss": 17.0655, + "step": 19 + }, + { + "epoch": 0.09252566141390776, + "grad_norm": 0.24450023472309113, + "learning_rate": 9.691358024691358e-06, + "loss": 15.3945, + "step": 20 + }, + { + "epoch": 0.09715194448460315, + "grad_norm": 0.3174358010292053, + "learning_rate": 9.675925925925926e-06, + "loss": 15.3045, + "step": 21 + }, + { + "epoch": 0.10177822755529854, + "grad_norm": 0.34466007351875305, + "learning_rate": 9.660493827160495e-06, + "loss": 16.7186, + "step": 22 + }, + { + "epoch": 0.10640451062599393, + "grad_norm": 0.294209361076355, + "learning_rate": 9.645061728395062e-06, + "loss": 16.2893, + "step": 23 + }, + { + "epoch": 0.11103079369668932, + "grad_norm": 0.3418446183204651, + "learning_rate": 9.62962962962963e-06, + "loss": 16.1827, + "step": 24 + }, + { + "epoch": 0.11565707676738471, + "grad_norm": 0.2584611177444458, + "learning_rate": 9.614197530864198e-06, + "loss": 15.7934, + "step": 25 + }, + { + "epoch": 0.1202833598380801, + "grad_norm": 0.3344869911670685, + "learning_rate": 9.598765432098767e-06, + "loss": 14.2338, + "step": 26 + }, + { + "epoch": 0.12490964290877549, + "grad_norm": 0.28123536705970764, + "learning_rate": 9.583333333333335e-06, + "loss": 16.2229, + "step": 27 + }, + { + "epoch": 0.12953592597947086, + "grad_norm": 0.22181656956672668, + "learning_rate": 9.567901234567902e-06, + "loss": 13.7744, + "step": 28 + }, + { + "epoch": 0.13416220905016626, + "grad_norm": 0.45139142870903015, + "learning_rate": 9.55246913580247e-06, + "loss": 15.4782, + "step": 29 + }, + { + "epoch": 0.13878849212086164, + "grad_norm": 0.21577508747577667, + "learning_rate": 9.537037037037037e-06, + "loss": 15.6752, + "step": 30 + }, + { + "epoch": 0.14341477519155704, + "grad_norm": 0.29512590169906616, + "learning_rate": 9.521604938271606e-06, + "loss": 14.7965, + "step": 31 + }, + { + "epoch": 0.14804105826225242, + "grad_norm": 0.28356751799583435, + "learning_rate": 9.506172839506174e-06, + "loss": 14.9097, + "step": 32 + }, + { + "epoch": 0.15266734133294782, + "grad_norm": 0.25781649351119995, + "learning_rate": 9.490740740740741e-06, + "loss": 14.494, + "step": 33 + }, + { + "epoch": 0.1572936244036432, + "grad_norm": 0.1967274248600006, + "learning_rate": 9.47530864197531e-06, + "loss": 15.0075, + "step": 34 + }, + { + "epoch": 0.1619199074743386, + "grad_norm": 0.42894116044044495, + "learning_rate": 9.459876543209878e-06, + "loss": 14.8055, + "step": 35 + }, + { + "epoch": 0.16654619054503397, + "grad_norm": 0.2417590320110321, + "learning_rate": 9.444444444444445e-06, + "loss": 14.7145, + "step": 36 + }, + { + "epoch": 0.17117247361572938, + "grad_norm": 0.15463685989379883, + "learning_rate": 9.429012345679013e-06, + "loss": 14.4273, + "step": 37 + }, + { + "epoch": 0.17579875668642475, + "grad_norm": 0.21183425188064575, + "learning_rate": 9.413580246913581e-06, + "loss": 16.6545, + "step": 38 + }, + { + "epoch": 0.18042503975712013, + "grad_norm": 0.41926464438438416, + "learning_rate": 9.398148148148148e-06, + "loss": 13.4609, + "step": 39 + }, + { + "epoch": 0.18505132282781553, + "grad_norm": 0.21585555374622345, + "learning_rate": 9.382716049382717e-06, + "loss": 14.1515, + "step": 40 + }, + { + "epoch": 0.1896776058985109, + "grad_norm": 0.28051456809043884, + "learning_rate": 9.367283950617285e-06, + "loss": 14.0035, + "step": 41 + }, + { + "epoch": 0.1943038889692063, + "grad_norm": 0.18281345069408417, + "learning_rate": 9.351851851851854e-06, + "loss": 14.6411, + "step": 42 + }, + { + "epoch": 0.19893017203990168, + "grad_norm": 0.1747712790966034, + "learning_rate": 9.33641975308642e-06, + "loss": 14.4986, + "step": 43 + }, + { + "epoch": 0.20355645511059708, + "grad_norm": 0.2613980174064636, + "learning_rate": 9.320987654320989e-06, + "loss": 14.2572, + "step": 44 + }, + { + "epoch": 0.20818273818129246, + "grad_norm": 0.19127604365348816, + "learning_rate": 9.305555555555557e-06, + "loss": 14.3084, + "step": 45 + }, + { + "epoch": 0.21280902125198786, + "grad_norm": 0.24545663595199585, + "learning_rate": 9.290123456790124e-06, + "loss": 13.3381, + "step": 46 + }, + { + "epoch": 0.21743530432268324, + "grad_norm": 0.25891631841659546, + "learning_rate": 9.274691358024692e-06, + "loss": 13.1076, + "step": 47 + }, + { + "epoch": 0.22206158739337864, + "grad_norm": 0.23256707191467285, + "learning_rate": 9.25925925925926e-06, + "loss": 13.3061, + "step": 48 + }, + { + "epoch": 0.22668787046407401, + "grad_norm": 0.2427562177181244, + "learning_rate": 9.243827160493828e-06, + "loss": 16.0646, + "step": 49 + }, + { + "epoch": 0.23131415353476942, + "grad_norm": 0.22760462760925293, + "learning_rate": 9.228395061728396e-06, + "loss": 12.937, + "step": 50 + }, + { + "epoch": 0.2359404366054648, + "grad_norm": 0.27145224809646606, + "learning_rate": 9.212962962962963e-06, + "loss": 13.8299, + "step": 51 + }, + { + "epoch": 0.2405667196761602, + "grad_norm": 0.3120661973953247, + "learning_rate": 9.197530864197531e-06, + "loss": 14.074, + "step": 52 + }, + { + "epoch": 0.24519300274685557, + "grad_norm": 0.21604514122009277, + "learning_rate": 9.1820987654321e-06, + "loss": 13.6731, + "step": 53 + }, + { + "epoch": 0.24981928581755097, + "grad_norm": 0.22970032691955566, + "learning_rate": 9.166666666666666e-06, + "loss": 14.0123, + "step": 54 + }, + { + "epoch": 0.2544455688882464, + "grad_norm": 0.24079963564872742, + "learning_rate": 9.151234567901235e-06, + "loss": 13.5458, + "step": 55 + }, + { + "epoch": 0.2590718519589417, + "grad_norm": 0.21451319754123688, + "learning_rate": 9.135802469135803e-06, + "loss": 12.5504, + "step": 56 + }, + { + "epoch": 0.2636981350296371, + "grad_norm": 0.2508305609226227, + "learning_rate": 9.120370370370372e-06, + "loss": 14.8642, + "step": 57 + }, + { + "epoch": 0.26832441810033253, + "grad_norm": 0.24022799730300903, + "learning_rate": 9.10493827160494e-06, + "loss": 15.1011, + "step": 58 + }, + { + "epoch": 0.2729507011710279, + "grad_norm": 0.3794403076171875, + "learning_rate": 9.089506172839507e-06, + "loss": 14.3993, + "step": 59 + }, + { + "epoch": 0.2775769842417233, + "grad_norm": 0.28563258051872253, + "learning_rate": 9.074074074074075e-06, + "loss": 14.294, + "step": 60 + }, + { + "epoch": 0.2822032673124187, + "grad_norm": 0.28061386942863464, + "learning_rate": 9.058641975308642e-06, + "loss": 13.033, + "step": 61 + }, + { + "epoch": 0.2868295503831141, + "grad_norm": 0.3351198136806488, + "learning_rate": 9.04320987654321e-06, + "loss": 14.5075, + "step": 62 + }, + { + "epoch": 0.29145583345380943, + "grad_norm": 0.23749324679374695, + "learning_rate": 9.027777777777779e-06, + "loss": 13.7952, + "step": 63 + }, + { + "epoch": 0.29608211652450483, + "grad_norm": 0.16919538378715515, + "learning_rate": 9.012345679012346e-06, + "loss": 14.7146, + "step": 64 + }, + { + "epoch": 0.30070839959520024, + "grad_norm": 0.18566973507404327, + "learning_rate": 8.996913580246914e-06, + "loss": 13.9967, + "step": 65 + }, + { + "epoch": 0.30533468266589564, + "grad_norm": 0.25144919753074646, + "learning_rate": 8.981481481481483e-06, + "loss": 13.0544, + "step": 66 + }, + { + "epoch": 0.309960965736591, + "grad_norm": 0.1971070021390915, + "learning_rate": 8.96604938271605e-06, + "loss": 12.7932, + "step": 67 + }, + { + "epoch": 0.3145872488072864, + "grad_norm": 0.2236046940088272, + "learning_rate": 8.950617283950618e-06, + "loss": 14.4338, + "step": 68 + }, + { + "epoch": 0.3192135318779818, + "grad_norm": 0.20062977075576782, + "learning_rate": 8.935185185185186e-06, + "loss": 13.6965, + "step": 69 + }, + { + "epoch": 0.3238398149486772, + "grad_norm": 0.3102545738220215, + "learning_rate": 8.919753086419753e-06, + "loss": 14.174, + "step": 70 + }, + { + "epoch": 0.32846609801937254, + "grad_norm": 0.282172828912735, + "learning_rate": 8.904320987654322e-06, + "loss": 14.012, + "step": 71 + }, + { + "epoch": 0.33309238109006795, + "grad_norm": 0.31374603509902954, + "learning_rate": 8.888888888888888e-06, + "loss": 12.9434, + "step": 72 + }, + { + "epoch": 0.33771866416076335, + "grad_norm": 0.24876756966114044, + "learning_rate": 8.873456790123458e-06, + "loss": 13.0801, + "step": 73 + }, + { + "epoch": 0.34234494723145875, + "grad_norm": 0.2828700840473175, + "learning_rate": 8.858024691358025e-06, + "loss": 12.3676, + "step": 74 + }, + { + "epoch": 0.3469712303021541, + "grad_norm": 0.23395580053329468, + "learning_rate": 8.842592592592594e-06, + "loss": 12.8463, + "step": 75 + }, + { + "epoch": 0.3515975133728495, + "grad_norm": 0.26694121956825256, + "learning_rate": 8.827160493827162e-06, + "loss": 11.5822, + "step": 76 + }, + { + "epoch": 0.3562237964435449, + "grad_norm": 0.2129782736301422, + "learning_rate": 8.811728395061729e-06, + "loss": 12.6834, + "step": 77 + }, + { + "epoch": 0.36085007951424025, + "grad_norm": 0.20400603115558624, + "learning_rate": 8.796296296296297e-06, + "loss": 14.2065, + "step": 78 + }, + { + "epoch": 0.36547636258493565, + "grad_norm": 0.45560839772224426, + "learning_rate": 8.780864197530866e-06, + "loss": 13.1033, + "step": 79 + }, + { + "epoch": 0.37010264565563106, + "grad_norm": 0.26329922676086426, + "learning_rate": 8.765432098765432e-06, + "loss": 11.9119, + "step": 80 + }, + { + "epoch": 0.37472892872632646, + "grad_norm": 0.21817731857299805, + "learning_rate": 8.750000000000001e-06, + "loss": 12.815, + "step": 81 + }, + { + "epoch": 0.3793552117970218, + "grad_norm": 0.2743465304374695, + "learning_rate": 8.73456790123457e-06, + "loss": 12.2468, + "step": 82 + }, + { + "epoch": 0.3839814948677172, + "grad_norm": 0.2570422291755676, + "learning_rate": 8.719135802469136e-06, + "loss": 11.9801, + "step": 83 + }, + { + "epoch": 0.3886077779384126, + "grad_norm": 0.21237581968307495, + "learning_rate": 8.703703703703705e-06, + "loss": 11.7092, + "step": 84 + }, + { + "epoch": 0.393234061009108, + "grad_norm": 0.27444881200790405, + "learning_rate": 8.688271604938271e-06, + "loss": 13.3074, + "step": 85 + }, + { + "epoch": 0.39786034407980336, + "grad_norm": 0.22574838995933533, + "learning_rate": 8.67283950617284e-06, + "loss": 13.1573, + "step": 86 + }, + { + "epoch": 0.40248662715049877, + "grad_norm": 0.252755731344223, + "learning_rate": 8.657407407407408e-06, + "loss": 12.3472, + "step": 87 + }, + { + "epoch": 0.40711291022119417, + "grad_norm": 0.2729659378528595, + "learning_rate": 8.641975308641975e-06, + "loss": 12.8413, + "step": 88 + }, + { + "epoch": 0.41173919329188957, + "grad_norm": 0.21728309988975525, + "learning_rate": 8.626543209876543e-06, + "loss": 11.3523, + "step": 89 + }, + { + "epoch": 0.4163654763625849, + "grad_norm": 0.264457106590271, + "learning_rate": 8.611111111111112e-06, + "loss": 13.1175, + "step": 90 + }, + { + "epoch": 0.4209917594332803, + "grad_norm": 0.20264536142349243, + "learning_rate": 8.59567901234568e-06, + "loss": 12.8442, + "step": 91 + }, + { + "epoch": 0.4256180425039757, + "grad_norm": 0.3003804385662079, + "learning_rate": 8.580246913580249e-06, + "loss": 14.5755, + "step": 92 + }, + { + "epoch": 0.4302443255746711, + "grad_norm": 0.294029176235199, + "learning_rate": 8.564814814814816e-06, + "loss": 13.5777, + "step": 93 + }, + { + "epoch": 0.4348706086453665, + "grad_norm": 0.21502816677093506, + "learning_rate": 8.549382716049384e-06, + "loss": 11.9804, + "step": 94 + }, + { + "epoch": 0.4394968917160619, + "grad_norm": 0.21898359060287476, + "learning_rate": 8.53395061728395e-06, + "loss": 13.6382, + "step": 95 + }, + { + "epoch": 0.4441231747867573, + "grad_norm": 0.20010338723659515, + "learning_rate": 8.518518518518519e-06, + "loss": 12.4733, + "step": 96 + }, + { + "epoch": 0.4487494578574526, + "grad_norm": 0.2247803956270218, + "learning_rate": 8.503086419753088e-06, + "loss": 12.1589, + "step": 97 + }, + { + "epoch": 0.45337574092814803, + "grad_norm": 0.29326918721199036, + "learning_rate": 8.487654320987654e-06, + "loss": 12.8177, + "step": 98 + }, + { + "epoch": 0.45800202399884343, + "grad_norm": 0.18403670191764832, + "learning_rate": 8.472222222222223e-06, + "loss": 13.8527, + "step": 99 + }, + { + "epoch": 0.46262830706953884, + "grad_norm": 0.24531710147857666, + "learning_rate": 8.456790123456791e-06, + "loss": 11.9126, + "step": 100 + }, + { + "epoch": 0.4672545901402342, + "grad_norm": 0.24183672666549683, + "learning_rate": 8.441358024691358e-06, + "loss": 12.952, + "step": 101 + }, + { + "epoch": 0.4718808732109296, + "grad_norm": 0.29563647508621216, + "learning_rate": 8.425925925925926e-06, + "loss": 14.462, + "step": 102 + }, + { + "epoch": 0.476507156281625, + "grad_norm": 0.21156498789787292, + "learning_rate": 8.410493827160495e-06, + "loss": 13.4846, + "step": 103 + }, + { + "epoch": 0.4811334393523204, + "grad_norm": 0.24183927476406097, + "learning_rate": 8.395061728395062e-06, + "loss": 11.428, + "step": 104 + }, + { + "epoch": 0.48575972242301574, + "grad_norm": 0.21241174638271332, + "learning_rate": 8.37962962962963e-06, + "loss": 13.8068, + "step": 105 + }, + { + "epoch": 0.49038600549371114, + "grad_norm": 0.2149106115102768, + "learning_rate": 8.364197530864199e-06, + "loss": 12.7659, + "step": 106 + }, + { + "epoch": 0.49501228856440654, + "grad_norm": 0.21057994663715363, + "learning_rate": 8.348765432098767e-06, + "loss": 12.4651, + "step": 107 + }, + { + "epoch": 0.49963857163510195, + "grad_norm": 0.18793097138404846, + "learning_rate": 8.333333333333334e-06, + "loss": 11.7854, + "step": 108 + }, + { + "epoch": 0.5042648547057973, + "grad_norm": 0.1801634281873703, + "learning_rate": 8.317901234567902e-06, + "loss": 12.9306, + "step": 109 + }, + { + "epoch": 0.5088911377764928, + "grad_norm": 0.21523930132389069, + "learning_rate": 8.30246913580247e-06, + "loss": 11.6702, + "step": 110 + }, + { + "epoch": 0.5135174208471881, + "grad_norm": 0.22931678593158722, + "learning_rate": 8.287037037037037e-06, + "loss": 11.5856, + "step": 111 + }, + { + "epoch": 0.5181437039178834, + "grad_norm": 0.23802338540554047, + "learning_rate": 8.271604938271606e-06, + "loss": 12.7508, + "step": 112 + }, + { + "epoch": 0.5227699869885789, + "grad_norm": 0.23633763194084167, + "learning_rate": 8.256172839506174e-06, + "loss": 12.0906, + "step": 113 + }, + { + "epoch": 0.5273962700592743, + "grad_norm": 0.22553692758083344, + "learning_rate": 8.240740740740741e-06, + "loss": 12.4982, + "step": 114 + }, + { + "epoch": 0.5320225531299696, + "grad_norm": 0.2265913337469101, + "learning_rate": 8.22530864197531e-06, + "loss": 11.4945, + "step": 115 + }, + { + "epoch": 0.5366488362006651, + "grad_norm": 0.1848449558019638, + "learning_rate": 8.209876543209876e-06, + "loss": 11.866, + "step": 116 + }, + { + "epoch": 0.5412751192713604, + "grad_norm": 0.16533511877059937, + "learning_rate": 8.194444444444445e-06, + "loss": 12.3876, + "step": 117 + }, + { + "epoch": 0.5459014023420558, + "grad_norm": 0.19875134527683258, + "learning_rate": 8.179012345679013e-06, + "loss": 12.0186, + "step": 118 + }, + { + "epoch": 0.5505276854127512, + "grad_norm": 0.20982632040977478, + "learning_rate": 8.16358024691358e-06, + "loss": 13.1202, + "step": 119 + }, + { + "epoch": 0.5551539684834466, + "grad_norm": 0.2689734101295471, + "learning_rate": 8.148148148148148e-06, + "loss": 13.9764, + "step": 120 + }, + { + "epoch": 0.559780251554142, + "grad_norm": 0.14883318543434143, + "learning_rate": 8.132716049382717e-06, + "loss": 11.7656, + "step": 121 + }, + { + "epoch": 0.5644065346248374, + "grad_norm": 0.18115845322608948, + "learning_rate": 8.117283950617285e-06, + "loss": 11.8902, + "step": 122 + }, + { + "epoch": 0.5690328176955327, + "grad_norm": 0.23553551733493805, + "learning_rate": 8.101851851851854e-06, + "loss": 13.0425, + "step": 123 + }, + { + "epoch": 0.5736591007662282, + "grad_norm": 0.2110109180212021, + "learning_rate": 8.08641975308642e-06, + "loss": 13.6931, + "step": 124 + }, + { + "epoch": 0.5782853838369235, + "grad_norm": 0.2748431861400604, + "learning_rate": 8.070987654320989e-06, + "loss": 12.7852, + "step": 125 + }, + { + "epoch": 0.5829116669076189, + "grad_norm": 0.1847638040781021, + "learning_rate": 8.055555555555557e-06, + "loss": 12.9106, + "step": 126 + }, + { + "epoch": 0.5875379499783143, + "grad_norm": 0.48550117015838623, + "learning_rate": 8.040123456790124e-06, + "loss": 11.5484, + "step": 127 + }, + { + "epoch": 0.5921642330490097, + "grad_norm": 0.2579441964626312, + "learning_rate": 8.024691358024692e-06, + "loss": 11.9785, + "step": 128 + }, + { + "epoch": 0.5967905161197051, + "grad_norm": 0.19037479162216187, + "learning_rate": 8.00925925925926e-06, + "loss": 11.0466, + "step": 129 + }, + { + "epoch": 0.6014167991904005, + "grad_norm": 0.22901126742362976, + "learning_rate": 7.993827160493828e-06, + "loss": 11.1139, + "step": 130 + }, + { + "epoch": 0.6060430822610958, + "grad_norm": 0.16406911611557007, + "learning_rate": 7.978395061728396e-06, + "loss": 10.6613, + "step": 131 + }, + { + "epoch": 0.6106693653317913, + "grad_norm": 0.23492039740085602, + "learning_rate": 7.962962962962963e-06, + "loss": 11.3782, + "step": 132 + }, + { + "epoch": 0.6152956484024866, + "grad_norm": 0.21692270040512085, + "learning_rate": 7.947530864197531e-06, + "loss": 11.2788, + "step": 133 + }, + { + "epoch": 0.619921931473182, + "grad_norm": 0.17832407355308533, + "learning_rate": 7.9320987654321e-06, + "loss": 11.7617, + "step": 134 + }, + { + "epoch": 0.6245482145438774, + "grad_norm": 0.20935167372226715, + "learning_rate": 7.916666666666667e-06, + "loss": 10.9103, + "step": 135 + }, + { + "epoch": 0.6291744976145728, + "grad_norm": 0.199855774641037, + "learning_rate": 7.901234567901235e-06, + "loss": 11.5339, + "step": 136 + }, + { + "epoch": 0.6338007806852681, + "grad_norm": 0.1627349704504013, + "learning_rate": 7.885802469135803e-06, + "loss": 10.3593, + "step": 137 + }, + { + "epoch": 0.6384270637559636, + "grad_norm": 0.18615403771400452, + "learning_rate": 7.870370370370372e-06, + "loss": 12.3204, + "step": 138 + }, + { + "epoch": 0.6430533468266589, + "grad_norm": 0.17866994440555573, + "learning_rate": 7.854938271604939e-06, + "loss": 11.548, + "step": 139 + }, + { + "epoch": 0.6476796298973544, + "grad_norm": 0.22644223272800446, + "learning_rate": 7.839506172839507e-06, + "loss": 12.0289, + "step": 140 + }, + { + "epoch": 0.6523059129680497, + "grad_norm": 0.18510523438453674, + "learning_rate": 7.824074074074076e-06, + "loss": 12.1312, + "step": 141 + }, + { + "epoch": 0.6569321960387451, + "grad_norm": 0.1948799043893814, + "learning_rate": 7.808641975308642e-06, + "loss": 13.4831, + "step": 142 + }, + { + "epoch": 0.6615584791094405, + "grad_norm": 0.2084604948759079, + "learning_rate": 7.79320987654321e-06, + "loss": 11.1289, + "step": 143 + }, + { + "epoch": 0.6661847621801359, + "grad_norm": 0.194437637925148, + "learning_rate": 7.77777777777778e-06, + "loss": 11.5961, + "step": 144 + }, + { + "epoch": 0.6708110452508312, + "grad_norm": 0.2627541124820709, + "learning_rate": 7.762345679012346e-06, + "loss": 12.0198, + "step": 145 + }, + { + "epoch": 0.6754373283215267, + "grad_norm": 0.21856093406677246, + "learning_rate": 7.746913580246914e-06, + "loss": 12.5285, + "step": 146 + }, + { + "epoch": 0.680063611392222, + "grad_norm": 0.15228690207004547, + "learning_rate": 7.731481481481483e-06, + "loss": 10.8653, + "step": 147 + }, + { + "epoch": 0.6846898944629175, + "grad_norm": 0.22951403260231018, + "learning_rate": 7.71604938271605e-06, + "loss": 10.8751, + "step": 148 + }, + { + "epoch": 0.6893161775336129, + "grad_norm": 0.25641652941703796, + "learning_rate": 7.700617283950618e-06, + "loss": 11.0195, + "step": 149 + }, + { + "epoch": 0.6939424606043082, + "grad_norm": 0.17132331430912018, + "learning_rate": 7.685185185185185e-06, + "loss": 13.9163, + "step": 150 + }, + { + "epoch": 0.6985687436750037, + "grad_norm": 0.24479900300502777, + "learning_rate": 7.669753086419753e-06, + "loss": 11.4487, + "step": 151 + }, + { + "epoch": 0.703195026745699, + "grad_norm": 0.1990821659564972, + "learning_rate": 7.654320987654322e-06, + "loss": 10.6267, + "step": 152 + }, + { + "epoch": 0.7078213098163944, + "grad_norm": 0.17914509773254395, + "learning_rate": 7.638888888888888e-06, + "loss": 11.5696, + "step": 153 + }, + { + "epoch": 0.7124475928870898, + "grad_norm": 0.19086718559265137, + "learning_rate": 7.623456790123458e-06, + "loss": 12.3853, + "step": 154 + }, + { + "epoch": 0.7170738759577852, + "grad_norm": 0.1417158991098404, + "learning_rate": 7.608024691358026e-06, + "loss": 10.8805, + "step": 155 + }, + { + "epoch": 0.7217001590284805, + "grad_norm": 0.32373055815696716, + "learning_rate": 7.592592592592594e-06, + "loss": 12.586, + "step": 156 + }, + { + "epoch": 0.726326442099176, + "grad_norm": 0.1995164453983307, + "learning_rate": 7.577160493827161e-06, + "loss": 11.4977, + "step": 157 + }, + { + "epoch": 0.7309527251698713, + "grad_norm": 0.16449247300624847, + "learning_rate": 7.561728395061729e-06, + "loss": 11.8027, + "step": 158 + }, + { + "epoch": 0.7355790082405668, + "grad_norm": 0.21888476610183716, + "learning_rate": 7.546296296296297e-06, + "loss": 11.8753, + "step": 159 + }, + { + "epoch": 0.7402052913112621, + "grad_norm": 0.19433487951755524, + "learning_rate": 7.530864197530865e-06, + "loss": 11.5252, + "step": 160 + }, + { + "epoch": 0.7448315743819575, + "grad_norm": 0.19709180295467377, + "learning_rate": 7.515432098765433e-06, + "loss": 10.7059, + "step": 161 + }, + { + "epoch": 0.7494578574526529, + "grad_norm": 0.17522279918193817, + "learning_rate": 7.500000000000001e-06, + "loss": 11.2329, + "step": 162 + }, + { + "epoch": 0.7540841405233483, + "grad_norm": 0.2038147747516632, + "learning_rate": 7.484567901234569e-06, + "loss": 11.425, + "step": 163 + }, + { + "epoch": 0.7587104235940436, + "grad_norm": 0.29561126232147217, + "learning_rate": 7.469135802469136e-06, + "loss": 13.192, + "step": 164 + }, + { + "epoch": 0.7633367066647391, + "grad_norm": 0.2295159548521042, + "learning_rate": 7.453703703703704e-06, + "loss": 10.9714, + "step": 165 + }, + { + "epoch": 0.7679629897354344, + "grad_norm": 0.1554422378540039, + "learning_rate": 7.438271604938272e-06, + "loss": 10.5617, + "step": 166 + }, + { + "epoch": 0.7725892728061299, + "grad_norm": 0.19565965235233307, + "learning_rate": 7.42283950617284e-06, + "loss": 10.8574, + "step": 167 + }, + { + "epoch": 0.7772155558768252, + "grad_norm": 0.2645648419857025, + "learning_rate": 7.4074074074074075e-06, + "loss": 11.4472, + "step": 168 + }, + { + "epoch": 0.7818418389475206, + "grad_norm": 0.2832958996295929, + "learning_rate": 7.391975308641975e-06, + "loss": 11.628, + "step": 169 + }, + { + "epoch": 0.786468122018216, + "grad_norm": 0.2097349464893341, + "learning_rate": 7.3765432098765435e-06, + "loss": 12.2379, + "step": 170 + }, + { + "epoch": 0.7910944050889114, + "grad_norm": 0.19168758392333984, + "learning_rate": 7.361111111111112e-06, + "loss": 10.0489, + "step": 171 + }, + { + "epoch": 0.7957206881596067, + "grad_norm": 0.22470812499523163, + "learning_rate": 7.34567901234568e-06, + "loss": 11.8545, + "step": 172 + }, + { + "epoch": 0.8003469712303022, + "grad_norm": 0.2722543179988861, + "learning_rate": 7.330246913580248e-06, + "loss": 10.4316, + "step": 173 + }, + { + "epoch": 0.8049732543009975, + "grad_norm": 0.1686207354068756, + "learning_rate": 7.314814814814816e-06, + "loss": 12.5917, + "step": 174 + }, + { + "epoch": 0.8095995373716929, + "grad_norm": 0.2805918753147125, + "learning_rate": 7.299382716049383e-06, + "loss": 10.6457, + "step": 175 + }, + { + "epoch": 0.8142258204423883, + "grad_norm": 0.28346166014671326, + "learning_rate": 7.283950617283952e-06, + "loss": 11.6686, + "step": 176 + }, + { + "epoch": 0.8188521035130837, + "grad_norm": 0.23147298395633698, + "learning_rate": 7.268518518518519e-06, + "loss": 10.316, + "step": 177 + }, + { + "epoch": 0.8234783865837791, + "grad_norm": 0.17895270884037018, + "learning_rate": 7.253086419753087e-06, + "loss": 10.5762, + "step": 178 + }, + { + "epoch": 0.8281046696544745, + "grad_norm": 0.2260911464691162, + "learning_rate": 7.2376543209876545e-06, + "loss": 10.9989, + "step": 179 + }, + { + "epoch": 0.8327309527251698, + "grad_norm": 0.2163412719964981, + "learning_rate": 7.222222222222223e-06, + "loss": 10.7737, + "step": 180 + }, + { + "epoch": 0.8373572357958653, + "grad_norm": 0.21451716125011444, + "learning_rate": 7.2067901234567905e-06, + "loss": 12.4478, + "step": 181 + }, + { + "epoch": 0.8419835188665606, + "grad_norm": 0.14181958138942719, + "learning_rate": 7.191358024691358e-06, + "loss": 11.1375, + "step": 182 + }, + { + "epoch": 0.846609801937256, + "grad_norm": 0.18365442752838135, + "learning_rate": 7.1759259259259266e-06, + "loss": 11.3768, + "step": 183 + }, + { + "epoch": 0.8512360850079514, + "grad_norm": 0.20560483634471893, + "learning_rate": 7.160493827160494e-06, + "loss": 10.6342, + "step": 184 + }, + { + "epoch": 0.8558623680786468, + "grad_norm": 0.3396760821342468, + "learning_rate": 7.145061728395062e-06, + "loss": 12.9821, + "step": 185 + }, + { + "epoch": 0.8604886511493423, + "grad_norm": 0.18169118463993073, + "learning_rate": 7.129629629629629e-06, + "loss": 11.067, + "step": 186 + }, + { + "epoch": 0.8651149342200376, + "grad_norm": 0.18200387060642242, + "learning_rate": 7.114197530864199e-06, + "loss": 10.6416, + "step": 187 + }, + { + "epoch": 0.869741217290733, + "grad_norm": 0.23793809115886688, + "learning_rate": 7.098765432098766e-06, + "loss": 12.6389, + "step": 188 + }, + { + "epoch": 0.8743675003614284, + "grad_norm": 0.6948350667953491, + "learning_rate": 7.083333333333335e-06, + "loss": 11.5131, + "step": 189 + }, + { + "epoch": 0.8789937834321238, + "grad_norm": 0.19556772708892822, + "learning_rate": 7.067901234567902e-06, + "loss": 12.9335, + "step": 190 + }, + { + "epoch": 0.8836200665028191, + "grad_norm": 0.21277347207069397, + "learning_rate": 7.05246913580247e-06, + "loss": 11.533, + "step": 191 + }, + { + "epoch": 0.8882463495735146, + "grad_norm": 0.13241365551948547, + "learning_rate": 7.0370370370370375e-06, + "loss": 11.7588, + "step": 192 + }, + { + "epoch": 0.8928726326442099, + "grad_norm": 0.16065774857997894, + "learning_rate": 7.021604938271606e-06, + "loss": 10.907, + "step": 193 + }, + { + "epoch": 0.8974989157149053, + "grad_norm": 0.2425934374332428, + "learning_rate": 7.0061728395061735e-06, + "loss": 12.0332, + "step": 194 + }, + { + "epoch": 0.9021251987856007, + "grad_norm": 0.32493147253990173, + "learning_rate": 6.990740740740741e-06, + "loss": 11.0481, + "step": 195 + }, + { + "epoch": 0.9067514818562961, + "grad_norm": 0.20006415247917175, + "learning_rate": 6.975308641975309e-06, + "loss": 10.3825, + "step": 196 + }, + { + "epoch": 0.9113777649269915, + "grad_norm": 0.14633627235889435, + "learning_rate": 6.959876543209877e-06, + "loss": 11.3131, + "step": 197 + }, + { + "epoch": 0.9160040479976869, + "grad_norm": 0.1695437729358673, + "learning_rate": 6.944444444444445e-06, + "loss": 11.6805, + "step": 198 + }, + { + "epoch": 0.9206303310683822, + "grad_norm": 0.22382433712482452, + "learning_rate": 6.929012345679012e-06, + "loss": 12.4138, + "step": 199 + }, + { + "epoch": 0.9252566141390777, + "grad_norm": 0.18652208149433136, + "learning_rate": 6.913580246913581e-06, + "loss": 11.1235, + "step": 200 + }, + { + "epoch": 0.929882897209773, + "grad_norm": 0.21566657721996307, + "learning_rate": 6.898148148148148e-06, + "loss": 11.1278, + "step": 201 + }, + { + "epoch": 0.9345091802804684, + "grad_norm": 0.17630067467689514, + "learning_rate": 6.882716049382716e-06, + "loss": 12.1383, + "step": 202 + }, + { + "epoch": 0.9391354633511638, + "grad_norm": 0.22753016650676727, + "learning_rate": 6.867283950617285e-06, + "loss": 12.0065, + "step": 203 + }, + { + "epoch": 0.9437617464218592, + "grad_norm": 0.12988047301769257, + "learning_rate": 6.851851851851853e-06, + "loss": 11.9477, + "step": 204 + }, + { + "epoch": 0.9483880294925546, + "grad_norm": 0.2507847547531128, + "learning_rate": 6.8364197530864205e-06, + "loss": 11.5818, + "step": 205 + }, + { + "epoch": 0.95301431256325, + "grad_norm": 0.18030938506126404, + "learning_rate": 6.820987654320988e-06, + "loss": 13.1558, + "step": 206 + }, + { + "epoch": 0.9576405956339453, + "grad_norm": 0.21600230038166046, + "learning_rate": 6.8055555555555566e-06, + "loss": 10.7038, + "step": 207 + }, + { + "epoch": 0.9622668787046408, + "grad_norm": 0.1989142745733261, + "learning_rate": 6.790123456790124e-06, + "loss": 11.9244, + "step": 208 + }, + { + "epoch": 0.9668931617753361, + "grad_norm": 0.2517797648906708, + "learning_rate": 6.774691358024692e-06, + "loss": 11.0019, + "step": 209 + }, + { + "epoch": 0.9715194448460315, + "grad_norm": 0.24413681030273438, + "learning_rate": 6.75925925925926e-06, + "loss": 12.4322, + "step": 210 + }, + { + "epoch": 0.9761457279167269, + "grad_norm": 0.20667782425880432, + "learning_rate": 6.743827160493828e-06, + "loss": 10.6134, + "step": 211 + }, + { + "epoch": 0.9807720109874223, + "grad_norm": 0.17336109280586243, + "learning_rate": 6.728395061728395e-06, + "loss": 10.4673, + "step": 212 + }, + { + "epoch": 0.9853982940581176, + "grad_norm": 0.17548175156116486, + "learning_rate": 6.712962962962963e-06, + "loss": 10.1377, + "step": 213 + }, + { + "epoch": 0.9900245771288131, + "grad_norm": 0.1717563420534134, + "learning_rate": 6.6975308641975314e-06, + "loss": 11.5888, + "step": 214 + }, + { + "epoch": 0.9946508601995084, + "grad_norm": 0.1841057986021042, + "learning_rate": 6.682098765432099e-06, + "loss": 12.2213, + "step": 215 + }, + { + "epoch": 0.9992771432702039, + "grad_norm": 0.3000975251197815, + "learning_rate": 6.666666666666667e-06, + "loss": 11.046, + "step": 216 + }, + { + "epoch": 1.0, + "grad_norm": 0.3000975251197815, + "learning_rate": 6.666666666666667e-06, + "loss": 1.4199, + "step": 217 + }, + { + "epoch": 1.0046262830706953, + "grad_norm": 0.4262640178203583, + "learning_rate": 6.651234567901235e-06, + "loss": 10.5265, + "step": 218 + }, + { + "epoch": 1.0092525661413907, + "grad_norm": 0.21694016456604004, + "learning_rate": 6.635802469135803e-06, + "loss": 10.4657, + "step": 219 + }, + { + "epoch": 1.0138788492120863, + "grad_norm": 0.2662133276462555, + "learning_rate": 6.620370370370371e-06, + "loss": 10.7531, + "step": 220 + }, + { + "epoch": 1.0185051322827816, + "grad_norm": 0.4052404463291168, + "learning_rate": 6.60493827160494e-06, + "loss": 11.1598, + "step": 221 + }, + { + "epoch": 1.023131415353477, + "grad_norm": 0.23900097608566284, + "learning_rate": 6.589506172839507e-06, + "loss": 11.5257, + "step": 222 + }, + { + "epoch": 1.0277576984241723, + "grad_norm": 0.22569526731967926, + "learning_rate": 6.574074074074075e-06, + "loss": 11.5564, + "step": 223 + }, + { + "epoch": 1.0323839814948677, + "grad_norm": 0.29018473625183105, + "learning_rate": 6.558641975308642e-06, + "loss": 11.8655, + "step": 224 + }, + { + "epoch": 1.037010264565563, + "grad_norm": 0.2856806516647339, + "learning_rate": 6.543209876543211e-06, + "loss": 11.8046, + "step": 225 + }, + { + "epoch": 1.0416365476362586, + "grad_norm": 0.16691464185714722, + "learning_rate": 6.5277777777777784e-06, + "loss": 12.3704, + "step": 226 + }, + { + "epoch": 1.046262830706954, + "grad_norm": 0.24121911823749542, + "learning_rate": 6.512345679012346e-06, + "loss": 10.7236, + "step": 227 + }, + { + "epoch": 1.0508891137776493, + "grad_norm": 0.205901101231575, + "learning_rate": 6.4969135802469145e-06, + "loss": 10.9316, + "step": 228 + }, + { + "epoch": 1.0555153968483446, + "grad_norm": 0.18375830352306366, + "learning_rate": 6.481481481481482e-06, + "loss": 9.422, + "step": 229 + }, + { + "epoch": 1.06014167991904, + "grad_norm": 0.2132750004529953, + "learning_rate": 6.46604938271605e-06, + "loss": 12.6065, + "step": 230 + }, + { + "epoch": 1.0647679629897355, + "grad_norm": 0.23490285873413086, + "learning_rate": 6.450617283950617e-06, + "loss": 11.2282, + "step": 231 + }, + { + "epoch": 1.0693942460604309, + "grad_norm": 0.2604386806488037, + "learning_rate": 6.435185185185186e-06, + "loss": 10.8157, + "step": 232 + }, + { + "epoch": 1.0740205291311262, + "grad_norm": 0.21887636184692383, + "learning_rate": 6.419753086419753e-06, + "loss": 10.6996, + "step": 233 + }, + { + "epoch": 1.0786468122018216, + "grad_norm": 0.16946089267730713, + "learning_rate": 6.404320987654321e-06, + "loss": 10.3805, + "step": 234 + }, + { + "epoch": 1.083273095272517, + "grad_norm": 0.20204631984233856, + "learning_rate": 6.3888888888888885e-06, + "loss": 11.4941, + "step": 235 + }, + { + "epoch": 1.0878993783432125, + "grad_norm": 0.202669158577919, + "learning_rate": 6.373456790123458e-06, + "loss": 11.7944, + "step": 236 + }, + { + "epoch": 1.0925256614139078, + "grad_norm": 0.22706535458564758, + "learning_rate": 6.358024691358025e-06, + "loss": 12.6031, + "step": 237 + }, + { + "epoch": 1.0971519444846032, + "grad_norm": 0.17785371840000153, + "learning_rate": 6.342592592592594e-06, + "loss": 10.5881, + "step": 238 + }, + { + "epoch": 1.1017782275552985, + "grad_norm": 0.17540261149406433, + "learning_rate": 6.3271604938271615e-06, + "loss": 10.996, + "step": 239 + }, + { + "epoch": 1.1064045106259939, + "grad_norm": 0.22038735449314117, + "learning_rate": 6.311728395061729e-06, + "loss": 10.0293, + "step": 240 + }, + { + "epoch": 1.1110307936966892, + "grad_norm": 0.17491741478443146, + "learning_rate": 6.296296296296297e-06, + "loss": 10.0996, + "step": 241 + }, + { + "epoch": 1.1156570767673848, + "grad_norm": 0.2827455699443817, + "learning_rate": 6.280864197530865e-06, + "loss": 11.0431, + "step": 242 + }, + { + "epoch": 1.1202833598380801, + "grad_norm": 0.21907830238342285, + "learning_rate": 6.265432098765433e-06, + "loss": 10.4707, + "step": 243 + }, + { + "epoch": 1.1249096429087755, + "grad_norm": 0.19320356845855713, + "learning_rate": 6.25e-06, + "loss": 10.7269, + "step": 244 + }, + { + "epoch": 1.1295359259794708, + "grad_norm": 0.22684165835380554, + "learning_rate": 6.234567901234569e-06, + "loss": 11.3143, + "step": 245 + }, + { + "epoch": 1.1341622090501662, + "grad_norm": 0.21541574597358704, + "learning_rate": 6.219135802469136e-06, + "loss": 10.397, + "step": 246 + }, + { + "epoch": 1.1387884921208618, + "grad_norm": 0.21086207032203674, + "learning_rate": 6.203703703703704e-06, + "loss": 11.0086, + "step": 247 + }, + { + "epoch": 1.143414775191557, + "grad_norm": 0.26652148365974426, + "learning_rate": 6.1882716049382715e-06, + "loss": 11.5252, + "step": 248 + }, + { + "epoch": 1.1480410582622524, + "grad_norm": 0.1995311677455902, + "learning_rate": 6.17283950617284e-06, + "loss": 9.3802, + "step": 249 + }, + { + "epoch": 1.1526673413329478, + "grad_norm": 0.19715267419815063, + "learning_rate": 6.157407407407408e-06, + "loss": 11.6554, + "step": 250 + }, + { + "epoch": 1.1572936244036431, + "grad_norm": 0.16357477009296417, + "learning_rate": 6.141975308641975e-06, + "loss": 10.7531, + "step": 251 + }, + { + "epoch": 1.1619199074743385, + "grad_norm": 0.2201809138059616, + "learning_rate": 6.126543209876543e-06, + "loss": 11.4639, + "step": 252 + }, + { + "epoch": 1.166546190545034, + "grad_norm": 0.24760210514068604, + "learning_rate": 6.111111111111112e-06, + "loss": 10.1224, + "step": 253 + }, + { + "epoch": 1.1711724736157294, + "grad_norm": 0.18471167981624603, + "learning_rate": 6.09567901234568e-06, + "loss": 10.8261, + "step": 254 + }, + { + "epoch": 1.1757987566864248, + "grad_norm": 0.20257946848869324, + "learning_rate": 6.080246913580248e-06, + "loss": 9.8945, + "step": 255 + }, + { + "epoch": 1.18042503975712, + "grad_norm": 0.15809153020381927, + "learning_rate": 6.064814814814816e-06, + "loss": 10.4489, + "step": 256 + }, + { + "epoch": 1.1850513228278154, + "grad_norm": 0.2175232172012329, + "learning_rate": 6.049382716049383e-06, + "loss": 10.6301, + "step": 257 + }, + { + "epoch": 1.189677605898511, + "grad_norm": 0.22652743756771088, + "learning_rate": 6.033950617283951e-06, + "loss": 9.6189, + "step": 258 + }, + { + "epoch": 1.1943038889692064, + "grad_norm": 0.21281598508358002, + "learning_rate": 6.018518518518519e-06, + "loss": 10.8016, + "step": 259 + }, + { + "epoch": 1.1989301720399017, + "grad_norm": 0.20046480000019073, + "learning_rate": 6.003086419753087e-06, + "loss": 10.1304, + "step": 260 + }, + { + "epoch": 1.203556455110597, + "grad_norm": 0.19008351862430573, + "learning_rate": 5.9876543209876546e-06, + "loss": 10.6718, + "step": 261 + }, + { + "epoch": 1.2081827381812924, + "grad_norm": 0.2159290611743927, + "learning_rate": 5.972222222222222e-06, + "loss": 10.1963, + "step": 262 + }, + { + "epoch": 1.2128090212519878, + "grad_norm": 0.1703040450811386, + "learning_rate": 5.956790123456791e-06, + "loss": 10.9394, + "step": 263 + }, + { + "epoch": 1.2174353043226833, + "grad_norm": 0.20243868231773376, + "learning_rate": 5.941358024691358e-06, + "loss": 9.7358, + "step": 264 + }, + { + "epoch": 1.2220615873933787, + "grad_norm": 0.21622534096240997, + "learning_rate": 5.925925925925926e-06, + "loss": 9.2896, + "step": 265 + }, + { + "epoch": 1.226687870464074, + "grad_norm": 0.19408589601516724, + "learning_rate": 5.910493827160494e-06, + "loss": 10.0141, + "step": 266 + }, + { + "epoch": 1.2313141535347694, + "grad_norm": 0.19652226567268372, + "learning_rate": 5.895061728395062e-06, + "loss": 10.6724, + "step": 267 + }, + { + "epoch": 1.2359404366054647, + "grad_norm": 0.22394828498363495, + "learning_rate": 5.8796296296296295e-06, + "loss": 11.7472, + "step": 268 + }, + { + "epoch": 1.2405667196761603, + "grad_norm": 0.18796682357788086, + "learning_rate": 5.864197530864199e-06, + "loss": 11.3051, + "step": 269 + }, + { + "epoch": 1.2451930027468556, + "grad_norm": 0.1897086501121521, + "learning_rate": 5.848765432098766e-06, + "loss": 9.9475, + "step": 270 + }, + { + "epoch": 1.249819285817551, + "grad_norm": 0.19912078976631165, + "learning_rate": 5.833333333333334e-06, + "loss": 10.6907, + "step": 271 + }, + { + "epoch": 1.2544455688882463, + "grad_norm": 0.25889354944229126, + "learning_rate": 5.817901234567902e-06, + "loss": 9.9069, + "step": 272 + }, + { + "epoch": 1.2590718519589417, + "grad_norm": 0.19028806686401367, + "learning_rate": 5.80246913580247e-06, + "loss": 10.0748, + "step": 273 + }, + { + "epoch": 1.263698135029637, + "grad_norm": 0.1661495566368103, + "learning_rate": 5.787037037037038e-06, + "loss": 10.1342, + "step": 274 + }, + { + "epoch": 1.2683244181003326, + "grad_norm": 0.17706620693206787, + "learning_rate": 5.771604938271605e-06, + "loss": 12.4703, + "step": 275 + }, + { + "epoch": 1.272950701171028, + "grad_norm": 0.26582854986190796, + "learning_rate": 5.756172839506174e-06, + "loss": 12.6464, + "step": 276 + }, + { + "epoch": 1.2775769842417233, + "grad_norm": 0.178712397813797, + "learning_rate": 5.740740740740741e-06, + "loss": 9.6011, + "step": 277 + }, + { + "epoch": 1.2822032673124186, + "grad_norm": 0.18910570442676544, + "learning_rate": 5.725308641975309e-06, + "loss": 10.8773, + "step": 278 + }, + { + "epoch": 1.2868295503831142, + "grad_norm": 0.24754610657691956, + "learning_rate": 5.7098765432098764e-06, + "loss": 10.8981, + "step": 279 + }, + { + "epoch": 1.2914558334538095, + "grad_norm": 0.14676739275455475, + "learning_rate": 5.694444444444445e-06, + "loss": 10.1851, + "step": 280 + }, + { + "epoch": 1.296082116524505, + "grad_norm": 0.21573857963085175, + "learning_rate": 5.6790123456790125e-06, + "loss": 10.2565, + "step": 281 + }, + { + "epoch": 1.3007083995952002, + "grad_norm": 0.24850860238075256, + "learning_rate": 5.66358024691358e-06, + "loss": 9.9282, + "step": 282 + }, + { + "epoch": 1.3053346826658956, + "grad_norm": 0.180882066488266, + "learning_rate": 5.6481481481481485e-06, + "loss": 11.6456, + "step": 283 + }, + { + "epoch": 1.309960965736591, + "grad_norm": 0.19305004179477692, + "learning_rate": 5.632716049382716e-06, + "loss": 10.0129, + "step": 284 + }, + { + "epoch": 1.3145872488072863, + "grad_norm": 0.18499980866909027, + "learning_rate": 5.617283950617285e-06, + "loss": 10.6136, + "step": 285 + }, + { + "epoch": 1.3192135318779818, + "grad_norm": 0.2523798942565918, + "learning_rate": 5.601851851851853e-06, + "loss": 9.1629, + "step": 286 + }, + { + "epoch": 1.3238398149486772, + "grad_norm": 0.19359612464904785, + "learning_rate": 5.586419753086421e-06, + "loss": 9.9917, + "step": 287 + }, + { + "epoch": 1.3284660980193725, + "grad_norm": 0.15306031703948975, + "learning_rate": 5.570987654320988e-06, + "loss": 9.6058, + "step": 288 + }, + { + "epoch": 1.333092381090068, + "grad_norm": 0.2070600688457489, + "learning_rate": 5.555555555555557e-06, + "loss": 10.2776, + "step": 289 + }, + { + "epoch": 1.3377186641607635, + "grad_norm": 0.1619105190038681, + "learning_rate": 5.540123456790124e-06, + "loss": 9.1465, + "step": 290 + }, + { + "epoch": 1.3423449472314588, + "grad_norm": 0.16202016174793243, + "learning_rate": 5.524691358024692e-06, + "loss": 10.0362, + "step": 291 + }, + { + "epoch": 1.3469712303021542, + "grad_norm": 0.19075003266334534, + "learning_rate": 5.5092592592592595e-06, + "loss": 10.0189, + "step": 292 + }, + { + "epoch": 1.3515975133728495, + "grad_norm": 0.16484172642230988, + "learning_rate": 5.493827160493828e-06, + "loss": 12.3743, + "step": 293 + }, + { + "epoch": 1.3562237964435448, + "grad_norm": 0.2151799499988556, + "learning_rate": 5.4783950617283955e-06, + "loss": 9.9846, + "step": 294 + }, + { + "epoch": 1.3608500795142402, + "grad_norm": 0.20363835990428925, + "learning_rate": 5.462962962962963e-06, + "loss": 9.671, + "step": 295 + }, + { + "epoch": 1.3654763625849355, + "grad_norm": 0.15196539461612701, + "learning_rate": 5.447530864197531e-06, + "loss": 9.97, + "step": 296 + }, + { + "epoch": 1.3701026456556311, + "grad_norm": 0.17592975497245789, + "learning_rate": 5.432098765432099e-06, + "loss": 10.8506, + "step": 297 + }, + { + "epoch": 1.3747289287263265, + "grad_norm": 0.20012719929218292, + "learning_rate": 5.416666666666667e-06, + "loss": 10.4503, + "step": 298 + }, + { + "epoch": 1.3793552117970218, + "grad_norm": 0.19312864542007446, + "learning_rate": 5.401234567901234e-06, + "loss": 9.7727, + "step": 299 + }, + { + "epoch": 1.3839814948677172, + "grad_norm": 0.14697419106960297, + "learning_rate": 5.385802469135803e-06, + "loss": 11.9123, + "step": 300 + }, + { + "epoch": 1.3886077779384127, + "grad_norm": 0.15580600500106812, + "learning_rate": 5.370370370370371e-06, + "loss": 10.3442, + "step": 301 + }, + { + "epoch": 1.393234061009108, + "grad_norm": 0.19586268067359924, + "learning_rate": 5.354938271604939e-06, + "loss": 8.9558, + "step": 302 + }, + { + "epoch": 1.3978603440798034, + "grad_norm": 0.19394423067569733, + "learning_rate": 5.339506172839507e-06, + "loss": 10.8312, + "step": 303 + }, + { + "epoch": 1.4024866271504988, + "grad_norm": 0.21832065284252167, + "learning_rate": 5.324074074074075e-06, + "loss": 9.9943, + "step": 304 + }, + { + "epoch": 1.4071129102211941, + "grad_norm": 0.16751375794410706, + "learning_rate": 5.3086419753086425e-06, + "loss": 11.3979, + "step": 305 + }, + { + "epoch": 1.4117391932918895, + "grad_norm": 0.1763811856508255, + "learning_rate": 5.29320987654321e-06, + "loss": 11.175, + "step": 306 + }, + { + "epoch": 1.4163654763625848, + "grad_norm": 0.1916954666376114, + "learning_rate": 5.2777777777777785e-06, + "loss": 8.8938, + "step": 307 + }, + { + "epoch": 1.4209917594332804, + "grad_norm": 0.12443282455205917, + "learning_rate": 5.262345679012346e-06, + "loss": 9.6263, + "step": 308 + }, + { + "epoch": 1.4256180425039757, + "grad_norm": 0.18852603435516357, + "learning_rate": 5.246913580246914e-06, + "loss": 10.6811, + "step": 309 + }, + { + "epoch": 1.430244325574671, + "grad_norm": 0.1854352355003357, + "learning_rate": 5.231481481481482e-06, + "loss": 9.9107, + "step": 310 + }, + { + "epoch": 1.4348706086453664, + "grad_norm": 0.2382035106420517, + "learning_rate": 5.21604938271605e-06, + "loss": 10.8001, + "step": 311 + }, + { + "epoch": 1.439496891716062, + "grad_norm": 0.18645620346069336, + "learning_rate": 5.200617283950617e-06, + "loss": 9.686, + "step": 312 + }, + { + "epoch": 1.4441231747867573, + "grad_norm": 0.1910058856010437, + "learning_rate": 5.185185185185185e-06, + "loss": 10.3407, + "step": 313 + }, + { + "epoch": 1.4487494578574527, + "grad_norm": 0.23428039252758026, + "learning_rate": 5.1697530864197534e-06, + "loss": 9.2624, + "step": 314 + }, + { + "epoch": 1.453375740928148, + "grad_norm": 0.17926917970180511, + "learning_rate": 5.154320987654321e-06, + "loss": 9.5153, + "step": 315 + }, + { + "epoch": 1.4580020239988434, + "grad_norm": 0.22386913001537323, + "learning_rate": 5.138888888888889e-06, + "loss": 10.6948, + "step": 316 + }, + { + "epoch": 1.4626283070695387, + "grad_norm": 0.20423167943954468, + "learning_rate": 5.123456790123458e-06, + "loss": 10.3966, + "step": 317 + }, + { + "epoch": 1.467254590140234, + "grad_norm": 0.15411533415317535, + "learning_rate": 5.1080246913580255e-06, + "loss": 10.2981, + "step": 318 + }, + { + "epoch": 1.4718808732109296, + "grad_norm": 0.11654549837112427, + "learning_rate": 5.092592592592593e-06, + "loss": 10.1807, + "step": 319 + }, + { + "epoch": 1.476507156281625, + "grad_norm": 0.20476582646369934, + "learning_rate": 5.0771604938271616e-06, + "loss": 9.9201, + "step": 320 + }, + { + "epoch": 1.4811334393523203, + "grad_norm": 0.19140280783176422, + "learning_rate": 5.061728395061729e-06, + "loss": 10.3292, + "step": 321 + }, + { + "epoch": 1.4857597224230157, + "grad_norm": 0.16844482719898224, + "learning_rate": 5.046296296296297e-06, + "loss": 10.0726, + "step": 322 + }, + { + "epoch": 1.4903860054937113, + "grad_norm": 0.22298437356948853, + "learning_rate": 5.030864197530864e-06, + "loss": 10.9447, + "step": 323 + }, + { + "epoch": 1.4950122885644066, + "grad_norm": 0.1775355190038681, + "learning_rate": 5.015432098765433e-06, + "loss": 12.1205, + "step": 324 + }, + { + "epoch": 1.499638571635102, + "grad_norm": 0.198349267244339, + "learning_rate": 5e-06, + "loss": 9.0277, + "step": 325 + }, + { + "epoch": 1.5042648547057973, + "grad_norm": 0.16838030517101288, + "learning_rate": 4.984567901234568e-06, + "loss": 10.3551, + "step": 326 + }, + { + "epoch": 1.5088911377764926, + "grad_norm": 0.15806855261325836, + "learning_rate": 4.9691358024691365e-06, + "loss": 10.0048, + "step": 327 + }, + { + "epoch": 1.513517420847188, + "grad_norm": 0.21915429830551147, + "learning_rate": 4.953703703703704e-06, + "loss": 10.554, + "step": 328 + }, + { + "epoch": 1.5181437039178833, + "grad_norm": 0.23262116312980652, + "learning_rate": 4.938271604938272e-06, + "loss": 10.8198, + "step": 329 + }, + { + "epoch": 1.522769986988579, + "grad_norm": 0.18927887082099915, + "learning_rate": 4.92283950617284e-06, + "loss": 10.7398, + "step": 330 + }, + { + "epoch": 1.5273962700592743, + "grad_norm": 0.138357013463974, + "learning_rate": 4.907407407407408e-06, + "loss": 9.8479, + "step": 331 + }, + { + "epoch": 1.5320225531299696, + "grad_norm": 0.20223885774612427, + "learning_rate": 4.891975308641976e-06, + "loss": 11.1796, + "step": 332 + }, + { + "epoch": 1.5366488362006652, + "grad_norm": 0.17738379538059235, + "learning_rate": 4.876543209876544e-06, + "loss": 9.1805, + "step": 333 + }, + { + "epoch": 1.5412751192713605, + "grad_norm": 0.18442702293395996, + "learning_rate": 4.861111111111111e-06, + "loss": 8.4714, + "step": 334 + }, + { + "epoch": 1.5459014023420559, + "grad_norm": 0.19169731438159943, + "learning_rate": 4.845679012345679e-06, + "loss": 9.1503, + "step": 335 + }, + { + "epoch": 1.5505276854127512, + "grad_norm": 0.19981014728546143, + "learning_rate": 4.830246913580247e-06, + "loss": 9.6427, + "step": 336 + }, + { + "epoch": 1.5551539684834466, + "grad_norm": 0.1916283518075943, + "learning_rate": 4.814814814814815e-06, + "loss": 8.8991, + "step": 337 + }, + { + "epoch": 1.559780251554142, + "grad_norm": 0.14049910008907318, + "learning_rate": 4.7993827160493834e-06, + "loss": 9.2888, + "step": 338 + }, + { + "epoch": 1.5644065346248373, + "grad_norm": 0.14010216295719147, + "learning_rate": 4.783950617283951e-06, + "loss": 10.4533, + "step": 339 + }, + { + "epoch": 1.5690328176955326, + "grad_norm": 0.1919959932565689, + "learning_rate": 4.768518518518519e-06, + "loss": 10.149, + "step": 340 + }, + { + "epoch": 1.5736591007662282, + "grad_norm": 0.1598203033208847, + "learning_rate": 4.753086419753087e-06, + "loss": 11.1529, + "step": 341 + }, + { + "epoch": 1.5782853838369235, + "grad_norm": 0.19053767621517181, + "learning_rate": 4.737654320987655e-06, + "loss": 10.5543, + "step": 342 + }, + { + "epoch": 1.5829116669076189, + "grad_norm": 0.21635524928569794, + "learning_rate": 4.722222222222222e-06, + "loss": 8.7965, + "step": 343 + }, + { + "epoch": 1.5875379499783144, + "grad_norm": 0.2088153064250946, + "learning_rate": 4.706790123456791e-06, + "loss": 9.1113, + "step": 344 + }, + { + "epoch": 1.5921642330490098, + "grad_norm": 0.1993483006954193, + "learning_rate": 4.691358024691358e-06, + "loss": 9.7291, + "step": 345 + }, + { + "epoch": 1.5967905161197051, + "grad_norm": 0.24835029244422913, + "learning_rate": 4.675925925925927e-06, + "loss": 9.5215, + "step": 346 + }, + { + "epoch": 1.6014167991904005, + "grad_norm": 0.24942326545715332, + "learning_rate": 4.660493827160494e-06, + "loss": 9.3023, + "step": 347 + }, + { + "epoch": 1.6060430822610958, + "grad_norm": 0.18829648196697235, + "learning_rate": 4.645061728395062e-06, + "loss": 9.6869, + "step": 348 + }, + { + "epoch": 1.6106693653317912, + "grad_norm": 0.17725762724876404, + "learning_rate": 4.62962962962963e-06, + "loss": 10.0154, + "step": 349 + }, + { + "epoch": 1.6152956484024865, + "grad_norm": 0.145020991563797, + "learning_rate": 4.614197530864198e-06, + "loss": 10.6015, + "step": 350 + }, + { + "epoch": 1.6199219314731819, + "grad_norm": 0.12719136476516724, + "learning_rate": 4.598765432098766e-06, + "loss": 9.9244, + "step": 351 + }, + { + "epoch": 1.6245482145438774, + "grad_norm": 0.1771680861711502, + "learning_rate": 4.583333333333333e-06, + "loss": 9.4468, + "step": 352 + }, + { + "epoch": 1.6291744976145728, + "grad_norm": 0.1684349775314331, + "learning_rate": 4.567901234567902e-06, + "loss": 11.5601, + "step": 353 + }, + { + "epoch": 1.6338007806852681, + "grad_norm": 0.1284870207309723, + "learning_rate": 4.55246913580247e-06, + "loss": 9.5658, + "step": 354 + }, + { + "epoch": 1.6384270637559637, + "grad_norm": 0.15432706475257874, + "learning_rate": 4.537037037037038e-06, + "loss": 10.2547, + "step": 355 + }, + { + "epoch": 1.643053346826659, + "grad_norm": 0.19519680738449097, + "learning_rate": 4.521604938271605e-06, + "loss": 9.2126, + "step": 356 + }, + { + "epoch": 1.6476796298973544, + "grad_norm": 0.1388140469789505, + "learning_rate": 4.506172839506173e-06, + "loss": 9.9578, + "step": 357 + }, + { + "epoch": 1.6523059129680497, + "grad_norm": 0.19714529812335968, + "learning_rate": 4.490740740740741e-06, + "loss": 11.7677, + "step": 358 + }, + { + "epoch": 1.656932196038745, + "grad_norm": 0.22792014479637146, + "learning_rate": 4.475308641975309e-06, + "loss": 9.6164, + "step": 359 + }, + { + "epoch": 1.6615584791094404, + "grad_norm": 0.1519528478384018, + "learning_rate": 4.4598765432098765e-06, + "loss": 9.4579, + "step": 360 + }, + { + "epoch": 1.6661847621801358, + "grad_norm": 0.17547698318958282, + "learning_rate": 4.444444444444444e-06, + "loss": 8.2953, + "step": 361 + }, + { + "epoch": 1.6708110452508311, + "grad_norm": 0.2348184585571289, + "learning_rate": 4.429012345679013e-06, + "loss": 9.8739, + "step": 362 + }, + { + "epoch": 1.6754373283215267, + "grad_norm": 0.16753868758678436, + "learning_rate": 4.413580246913581e-06, + "loss": 9.8006, + "step": 363 + }, + { + "epoch": 1.680063611392222, + "grad_norm": 0.16872891783714294, + "learning_rate": 4.398148148148149e-06, + "loss": 10.2161, + "step": 364 + }, + { + "epoch": 1.6846898944629176, + "grad_norm": 0.18189671635627747, + "learning_rate": 4.382716049382716e-06, + "loss": 9.5235, + "step": 365 + }, + { + "epoch": 1.689316177533613, + "grad_norm": 0.12333816289901733, + "learning_rate": 4.367283950617285e-06, + "loss": 10.2188, + "step": 366 + }, + { + "epoch": 1.6939424606043083, + "grad_norm": 0.22701486945152283, + "learning_rate": 4.351851851851852e-06, + "loss": 10.7382, + "step": 367 + }, + { + "epoch": 1.6985687436750037, + "grad_norm": 0.20510244369506836, + "learning_rate": 4.33641975308642e-06, + "loss": 8.7558, + "step": 368 + }, + { + "epoch": 1.703195026745699, + "grad_norm": 0.1589343398809433, + "learning_rate": 4.3209876543209875e-06, + "loss": 9.924, + "step": 369 + }, + { + "epoch": 1.7078213098163944, + "grad_norm": 0.1433723270893097, + "learning_rate": 4.305555555555556e-06, + "loss": 11.0008, + "step": 370 + }, + { + "epoch": 1.7124475928870897, + "grad_norm": 0.17607755959033966, + "learning_rate": 4.290123456790124e-06, + "loss": 10.0096, + "step": 371 + }, + { + "epoch": 1.717073875957785, + "grad_norm": 0.16202040016651154, + "learning_rate": 4.274691358024692e-06, + "loss": 9.9295, + "step": 372 + }, + { + "epoch": 1.7217001590284804, + "grad_norm": 0.21179550886154175, + "learning_rate": 4.2592592592592596e-06, + "loss": 8.9856, + "step": 373 + }, + { + "epoch": 1.726326442099176, + "grad_norm": 0.35698196291923523, + "learning_rate": 4.243827160493827e-06, + "loss": 9.8793, + "step": 374 + }, + { + "epoch": 1.7309527251698713, + "grad_norm": 0.18081314861774445, + "learning_rate": 4.228395061728396e-06, + "loss": 10.8514, + "step": 375 + }, + { + "epoch": 1.7355790082405669, + "grad_norm": 0.19385862350463867, + "learning_rate": 4.212962962962963e-06, + "loss": 8.6034, + "step": 376 + }, + { + "epoch": 1.7402052913112622, + "grad_norm": 0.17623740434646606, + "learning_rate": 4.197530864197531e-06, + "loss": 10.7457, + "step": 377 + }, + { + "epoch": 1.7448315743819576, + "grad_norm": 0.1551784873008728, + "learning_rate": 4.182098765432099e-06, + "loss": 9.2378, + "step": 378 + }, + { + "epoch": 1.749457857452653, + "grad_norm": 0.15923646092414856, + "learning_rate": 4.166666666666667e-06, + "loss": 10.0731, + "step": 379 + }, + { + "epoch": 1.7540841405233483, + "grad_norm": 0.16185562312602997, + "learning_rate": 4.151234567901235e-06, + "loss": 8.9162, + "step": 380 + }, + { + "epoch": 1.7587104235940436, + "grad_norm": 0.23534594476222992, + "learning_rate": 4.135802469135803e-06, + "loss": 9.3805, + "step": 381 + }, + { + "epoch": 1.763336706664739, + "grad_norm": 0.17162498831748962, + "learning_rate": 4.1203703703703705e-06, + "loss": 10.0609, + "step": 382 + }, + { + "epoch": 1.7679629897354343, + "grad_norm": 0.15775223076343536, + "learning_rate": 4.104938271604938e-06, + "loss": 9.3534, + "step": 383 + }, + { + "epoch": 1.7725892728061299, + "grad_norm": 0.2251134216785431, + "learning_rate": 4.0895061728395066e-06, + "loss": 9.746, + "step": 384 + }, + { + "epoch": 1.7772155558768252, + "grad_norm": 0.19613635540008545, + "learning_rate": 4.074074074074074e-06, + "loss": 10.42, + "step": 385 + }, + { + "epoch": 1.7818418389475206, + "grad_norm": 0.16256873309612274, + "learning_rate": 4.058641975308643e-06, + "loss": 9.4703, + "step": 386 + }, + { + "epoch": 1.7864681220182161, + "grad_norm": 0.16490335762500763, + "learning_rate": 4.04320987654321e-06, + "loss": 9.5418, + "step": 387 + }, + { + "epoch": 1.7910944050889115, + "grad_norm": 0.159736767411232, + "learning_rate": 4.027777777777779e-06, + "loss": 10.1403, + "step": 388 + }, + { + "epoch": 1.7957206881596068, + "grad_norm": 0.1494799256324768, + "learning_rate": 4.012345679012346e-06, + "loss": 9.1583, + "step": 389 + }, + { + "epoch": 1.8003469712303022, + "grad_norm": 0.19815048575401306, + "learning_rate": 3.996913580246914e-06, + "loss": 10.2705, + "step": 390 + }, + { + "epoch": 1.8049732543009975, + "grad_norm": 0.17572833597660065, + "learning_rate": 3.9814814814814814e-06, + "loss": 9.4608, + "step": 391 + }, + { + "epoch": 1.8095995373716929, + "grad_norm": 0.23244017362594604, + "learning_rate": 3.96604938271605e-06, + "loss": 10.6273, + "step": 392 + }, + { + "epoch": 1.8142258204423882, + "grad_norm": 0.17120610177516937, + "learning_rate": 3.9506172839506175e-06, + "loss": 10.0444, + "step": 393 + }, + { + "epoch": 1.8188521035130836, + "grad_norm": 0.27220579981803894, + "learning_rate": 3.935185185185186e-06, + "loss": 10.64, + "step": 394 + }, + { + "epoch": 1.8234783865837791, + "grad_norm": 0.1648477017879486, + "learning_rate": 3.9197530864197535e-06, + "loss": 9.6264, + "step": 395 + }, + { + "epoch": 1.8281046696544745, + "grad_norm": 0.15771055221557617, + "learning_rate": 3.904320987654321e-06, + "loss": 10.5564, + "step": 396 + }, + { + "epoch": 1.8327309527251698, + "grad_norm": 0.21713097393512726, + "learning_rate": 3.88888888888889e-06, + "loss": 10.6551, + "step": 397 + }, + { + "epoch": 1.8373572357958654, + "grad_norm": 0.17076119780540466, + "learning_rate": 3.873456790123457e-06, + "loss": 11.1632, + "step": 398 + }, + { + "epoch": 1.8419835188665608, + "grad_norm": 0.160685196518898, + "learning_rate": 3.858024691358025e-06, + "loss": 9.4666, + "step": 399 + }, + { + "epoch": 1.846609801937256, + "grad_norm": 0.17023253440856934, + "learning_rate": 3.842592592592592e-06, + "loss": 9.8118, + "step": 400 + }, + { + "epoch": 1.8512360850079514, + "grad_norm": 0.2186374068260193, + "learning_rate": 3.827160493827161e-06, + "loss": 10.4903, + "step": 401 + }, + { + "epoch": 1.8558623680786468, + "grad_norm": 0.17954406142234802, + "learning_rate": 3.811728395061729e-06, + "loss": 11.3069, + "step": 402 + }, + { + "epoch": 1.8604886511493421, + "grad_norm": 0.2778474986553192, + "learning_rate": 3.796296296296297e-06, + "loss": 10.5854, + "step": 403 + }, + { + "epoch": 1.8651149342200375, + "grad_norm": 0.27708959579467773, + "learning_rate": 3.7808641975308645e-06, + "loss": 10.2884, + "step": 404 + }, + { + "epoch": 1.8697412172907328, + "grad_norm": 0.14399868249893188, + "learning_rate": 3.7654320987654325e-06, + "loss": 10.1108, + "step": 405 + }, + { + "epoch": 1.8743675003614284, + "grad_norm": 0.19901257753372192, + "learning_rate": 3.7500000000000005e-06, + "loss": 10.2892, + "step": 406 + }, + { + "epoch": 1.8789937834321238, + "grad_norm": 0.21875345706939697, + "learning_rate": 3.734567901234568e-06, + "loss": 9.0521, + "step": 407 + }, + { + "epoch": 1.883620066502819, + "grad_norm": 0.14913184940814972, + "learning_rate": 3.719135802469136e-06, + "loss": 10.5905, + "step": 408 + }, + { + "epoch": 1.8882463495735147, + "grad_norm": 0.22289720177650452, + "learning_rate": 3.7037037037037037e-06, + "loss": 9.1175, + "step": 409 + }, + { + "epoch": 1.89287263264421, + "grad_norm": 0.16322648525238037, + "learning_rate": 3.6882716049382718e-06, + "loss": 10.1569, + "step": 410 + }, + { + "epoch": 1.8974989157149054, + "grad_norm": 0.13389474153518677, + "learning_rate": 3.67283950617284e-06, + "loss": 9.7421, + "step": 411 + }, + { + "epoch": 1.9021251987856007, + "grad_norm": 0.17904877662658691, + "learning_rate": 3.657407407407408e-06, + "loss": 10.306, + "step": 412 + }, + { + "epoch": 1.906751481856296, + "grad_norm": 0.22490617632865906, + "learning_rate": 3.641975308641976e-06, + "loss": 10.0275, + "step": 413 + }, + { + "epoch": 1.9113777649269914, + "grad_norm": 0.20119769871234894, + "learning_rate": 3.6265432098765434e-06, + "loss": 9.5942, + "step": 414 + }, + { + "epoch": 1.9160040479976868, + "grad_norm": 0.156976118683815, + "learning_rate": 3.6111111111111115e-06, + "loss": 10.3435, + "step": 415 + }, + { + "epoch": 1.920630331068382, + "grad_norm": 0.13303594291210175, + "learning_rate": 3.595679012345679e-06, + "loss": 8.8616, + "step": 416 + }, + { + "epoch": 1.9252566141390777, + "grad_norm": 0.21186932921409607, + "learning_rate": 3.580246913580247e-06, + "loss": 9.8981, + "step": 417 + }, + { + "epoch": 1.929882897209773, + "grad_norm": 0.20669791102409363, + "learning_rate": 3.5648148148148147e-06, + "loss": 10.3689, + "step": 418 + }, + { + "epoch": 1.9345091802804684, + "grad_norm": 0.20095276832580566, + "learning_rate": 3.549382716049383e-06, + "loss": 9.136, + "step": 419 + }, + { + "epoch": 1.939135463351164, + "grad_norm": 0.16348451375961304, + "learning_rate": 3.533950617283951e-06, + "loss": 10.4419, + "step": 420 + }, + { + "epoch": 1.9437617464218593, + "grad_norm": 0.16439932584762573, + "learning_rate": 3.5185185185185187e-06, + "loss": 10.0692, + "step": 421 + }, + { + "epoch": 1.9483880294925546, + "grad_norm": 0.18562501668930054, + "learning_rate": 3.5030864197530868e-06, + "loss": 10.4452, + "step": 422 + }, + { + "epoch": 1.95301431256325, + "grad_norm": 0.15668661892414093, + "learning_rate": 3.4876543209876544e-06, + "loss": 9.1397, + "step": 423 + }, + { + "epoch": 1.9576405956339453, + "grad_norm": 0.129827082157135, + "learning_rate": 3.4722222222222224e-06, + "loss": 8.428, + "step": 424 + }, + { + "epoch": 1.9622668787046407, + "grad_norm": 0.14903448522090912, + "learning_rate": 3.4567901234567904e-06, + "loss": 9.0113, + "step": 425 + }, + { + "epoch": 1.966893161775336, + "grad_norm": 0.2033424973487854, + "learning_rate": 3.441358024691358e-06, + "loss": 8.6822, + "step": 426 + }, + { + "epoch": 1.9715194448460314, + "grad_norm": 0.1724868267774582, + "learning_rate": 3.4259259259259265e-06, + "loss": 9.3924, + "step": 427 + }, + { + "epoch": 1.976145727916727, + "grad_norm": 0.1788475513458252, + "learning_rate": 3.410493827160494e-06, + "loss": 7.8849, + "step": 428 + }, + { + "epoch": 1.9807720109874223, + "grad_norm": 0.13475210964679718, + "learning_rate": 3.395061728395062e-06, + "loss": 10.321, + "step": 429 + }, + { + "epoch": 1.9853982940581176, + "grad_norm": 0.15269149839878082, + "learning_rate": 3.37962962962963e-06, + "loss": 9.5632, + "step": 430 + }, + { + "epoch": 1.9900245771288132, + "grad_norm": 0.20584136247634888, + "learning_rate": 3.3641975308641977e-06, + "loss": 9.0406, + "step": 431 + }, + { + "epoch": 1.9946508601995085, + "grad_norm": 0.1267031729221344, + "learning_rate": 3.3487654320987657e-06, + "loss": 9.1269, + "step": 432 + }, + { + "epoch": 1.999277143270204, + "grad_norm": 0.20229946076869965, + "learning_rate": 3.3333333333333333e-06, + "loss": 9.1644, + "step": 433 + }, + { + "epoch": 2.0, + "grad_norm": 0.20229946076869965, + "learning_rate": 3.3333333333333333e-06, + "loss": 1.6753, + "step": 434 + }, + { + "epoch": 2.0046262830706953, + "grad_norm": 0.19674813747406006, + "learning_rate": 3.3179012345679013e-06, + "loss": 8.9472, + "step": 435 + }, + { + "epoch": 2.0092525661413907, + "grad_norm": 0.17176903784275055, + "learning_rate": 3.30246913580247e-06, + "loss": 8.1618, + "step": 436 + }, + { + "epoch": 2.013878849212086, + "grad_norm": 0.21751543879508972, + "learning_rate": 3.2870370370370374e-06, + "loss": 11.298, + "step": 437 + }, + { + "epoch": 2.0185051322827814, + "grad_norm": 0.15177738666534424, + "learning_rate": 3.2716049382716054e-06, + "loss": 8.9182, + "step": 438 + }, + { + "epoch": 2.0231314153534767, + "grad_norm": 0.13782791793346405, + "learning_rate": 3.256172839506173e-06, + "loss": 9.138, + "step": 439 + }, + { + "epoch": 2.0277576984241725, + "grad_norm": 0.1871718168258667, + "learning_rate": 3.240740740740741e-06, + "loss": 9.3935, + "step": 440 + }, + { + "epoch": 2.032383981494868, + "grad_norm": 0.13550569117069244, + "learning_rate": 3.2253086419753086e-06, + "loss": 9.4065, + "step": 441 + }, + { + "epoch": 2.0370102645655632, + "grad_norm": 0.1414715200662613, + "learning_rate": 3.2098765432098767e-06, + "loss": 10.0772, + "step": 442 + }, + { + "epoch": 2.0416365476362586, + "grad_norm": 0.16839131712913513, + "learning_rate": 3.1944444444444443e-06, + "loss": 8.4968, + "step": 443 + }, + { + "epoch": 2.046262830706954, + "grad_norm": 0.181600421667099, + "learning_rate": 3.1790123456790127e-06, + "loss": 8.0076, + "step": 444 + }, + { + "epoch": 2.0508891137776493, + "grad_norm": 0.1106899231672287, + "learning_rate": 3.1635802469135807e-06, + "loss": 8.5825, + "step": 445 + }, + { + "epoch": 2.0555153968483446, + "grad_norm": 0.17763055860996246, + "learning_rate": 3.1481481481481483e-06, + "loss": 9.2981, + "step": 446 + }, + { + "epoch": 2.06014167991904, + "grad_norm": 0.19595400989055634, + "learning_rate": 3.1327160493827164e-06, + "loss": 10.5243, + "step": 447 + }, + { + "epoch": 2.0647679629897353, + "grad_norm": 0.18254916369915009, + "learning_rate": 3.1172839506172844e-06, + "loss": 10.9755, + "step": 448 + }, + { + "epoch": 2.0693942460604307, + "grad_norm": 0.18021439015865326, + "learning_rate": 3.101851851851852e-06, + "loss": 9.3634, + "step": 449 + }, + { + "epoch": 2.074020529131126, + "grad_norm": 0.20369423925876617, + "learning_rate": 3.08641975308642e-06, + "loss": 10.1676, + "step": 450 + }, + { + "epoch": 2.078646812201822, + "grad_norm": 0.1611659973859787, + "learning_rate": 3.0709876543209876e-06, + "loss": 10.5072, + "step": 451 + }, + { + "epoch": 2.083273095272517, + "grad_norm": 0.16623584926128387, + "learning_rate": 3.055555555555556e-06, + "loss": 8.6923, + "step": 452 + }, + { + "epoch": 2.0878993783432125, + "grad_norm": 0.20521709322929382, + "learning_rate": 3.040123456790124e-06, + "loss": 10.3398, + "step": 453 + }, + { + "epoch": 2.092525661413908, + "grad_norm": 0.150822713971138, + "learning_rate": 3.0246913580246917e-06, + "loss": 8.1734, + "step": 454 + }, + { + "epoch": 2.097151944484603, + "grad_norm": 0.14610332250595093, + "learning_rate": 3.0092592592592597e-06, + "loss": 8.0033, + "step": 455 + }, + { + "epoch": 2.1017782275552985, + "grad_norm": 0.16596634685993195, + "learning_rate": 2.9938271604938273e-06, + "loss": 9.923, + "step": 456 + }, + { + "epoch": 2.106404510625994, + "grad_norm": 0.15604960918426514, + "learning_rate": 2.9783950617283953e-06, + "loss": 9.5128, + "step": 457 + }, + { + "epoch": 2.1110307936966892, + "grad_norm": 0.1488286554813385, + "learning_rate": 2.962962962962963e-06, + "loss": 8.704, + "step": 458 + }, + { + "epoch": 2.1156570767673846, + "grad_norm": 0.1483355164527893, + "learning_rate": 2.947530864197531e-06, + "loss": 8.4275, + "step": 459 + }, + { + "epoch": 2.12028335983808, + "grad_norm": 0.2545453906059265, + "learning_rate": 2.9320987654320994e-06, + "loss": 9.332, + "step": 460 + }, + { + "epoch": 2.1249096429087757, + "grad_norm": 0.18272359669208527, + "learning_rate": 2.916666666666667e-06, + "loss": 9.4908, + "step": 461 + }, + { + "epoch": 2.129535925979471, + "grad_norm": 0.19491931796073914, + "learning_rate": 2.901234567901235e-06, + "loss": 8.5668, + "step": 462 + }, + { + "epoch": 2.1341622090501664, + "grad_norm": 0.171804741024971, + "learning_rate": 2.8858024691358026e-06, + "loss": 8.9085, + "step": 463 + }, + { + "epoch": 2.1387884921208618, + "grad_norm": 0.22477325797080994, + "learning_rate": 2.8703703703703706e-06, + "loss": 9.5643, + "step": 464 + }, + { + "epoch": 2.143414775191557, + "grad_norm": 0.2512359023094177, + "learning_rate": 2.8549382716049382e-06, + "loss": 10.6236, + "step": 465 + }, + { + "epoch": 2.1480410582622524, + "grad_norm": 0.15513500571250916, + "learning_rate": 2.8395061728395062e-06, + "loss": 9.3385, + "step": 466 + }, + { + "epoch": 2.152667341332948, + "grad_norm": 0.18065772950649261, + "learning_rate": 2.8240740740740743e-06, + "loss": 9.3942, + "step": 467 + }, + { + "epoch": 2.157293624403643, + "grad_norm": 0.19600172340869904, + "learning_rate": 2.8086419753086423e-06, + "loss": 9.0732, + "step": 468 + }, + { + "epoch": 2.1619199074743385, + "grad_norm": 0.17671746015548706, + "learning_rate": 2.7932098765432103e-06, + "loss": 10.023, + "step": 469 + }, + { + "epoch": 2.166546190545034, + "grad_norm": 0.16406480967998505, + "learning_rate": 2.7777777777777783e-06, + "loss": 9.8105, + "step": 470 + }, + { + "epoch": 2.171172473615729, + "grad_norm": 0.11806418746709824, + "learning_rate": 2.762345679012346e-06, + "loss": 8.2566, + "step": 471 + }, + { + "epoch": 2.175798756686425, + "grad_norm": 0.32428812980651855, + "learning_rate": 2.746913580246914e-06, + "loss": 9.1191, + "step": 472 + }, + { + "epoch": 2.1804250397571203, + "grad_norm": 0.22312945127487183, + "learning_rate": 2.7314814814814816e-06, + "loss": 9.9265, + "step": 473 + }, + { + "epoch": 2.1850513228278157, + "grad_norm": 0.32287222146987915, + "learning_rate": 2.7160493827160496e-06, + "loss": 10.6421, + "step": 474 + }, + { + "epoch": 2.189677605898511, + "grad_norm": 0.15497097373008728, + "learning_rate": 2.700617283950617e-06, + "loss": 8.9862, + "step": 475 + }, + { + "epoch": 2.1943038889692064, + "grad_norm": 0.13556216657161713, + "learning_rate": 2.6851851851851856e-06, + "loss": 8.658, + "step": 476 + }, + { + "epoch": 2.1989301720399017, + "grad_norm": 0.17666535079479218, + "learning_rate": 2.6697530864197536e-06, + "loss": 10.2978, + "step": 477 + }, + { + "epoch": 2.203556455110597, + "grad_norm": 0.16342324018478394, + "learning_rate": 2.6543209876543212e-06, + "loss": 8.4887, + "step": 478 + }, + { + "epoch": 2.2081827381812924, + "grad_norm": 0.2034367471933365, + "learning_rate": 2.6388888888888893e-06, + "loss": 7.8828, + "step": 479 + }, + { + "epoch": 2.2128090212519878, + "grad_norm": 0.23170992732048035, + "learning_rate": 2.623456790123457e-06, + "loss": 9.6533, + "step": 480 + }, + { + "epoch": 2.217435304322683, + "grad_norm": 0.16840022802352905, + "learning_rate": 2.608024691358025e-06, + "loss": 8.1141, + "step": 481 + }, + { + "epoch": 2.2220615873933784, + "grad_norm": 0.1619078814983368, + "learning_rate": 2.5925925925925925e-06, + "loss": 8.6294, + "step": 482 + }, + { + "epoch": 2.2266878704640742, + "grad_norm": 0.22204962372779846, + "learning_rate": 2.5771604938271605e-06, + "loss": 10.4391, + "step": 483 + }, + { + "epoch": 2.2313141535347696, + "grad_norm": 0.1876753270626068, + "learning_rate": 2.561728395061729e-06, + "loss": 9.9287, + "step": 484 + }, + { + "epoch": 2.235940436605465, + "grad_norm": 0.18601296842098236, + "learning_rate": 2.5462962962962966e-06, + "loss": 9.3439, + "step": 485 + }, + { + "epoch": 2.2405667196761603, + "grad_norm": 0.17974725365638733, + "learning_rate": 2.5308641975308646e-06, + "loss": 9.6812, + "step": 486 + }, + { + "epoch": 2.2451930027468556, + "grad_norm": 0.1832571029663086, + "learning_rate": 2.515432098765432e-06, + "loss": 7.6883, + "step": 487 + }, + { + "epoch": 2.249819285817551, + "grad_norm": 0.18652378022670746, + "learning_rate": 2.5e-06, + "loss": 8.9935, + "step": 488 + }, + { + "epoch": 2.2544455688882463, + "grad_norm": 0.20331954956054688, + "learning_rate": 2.4845679012345682e-06, + "loss": 9.1709, + "step": 489 + }, + { + "epoch": 2.2590718519589417, + "grad_norm": 0.15439340472221375, + "learning_rate": 2.469135802469136e-06, + "loss": 9.0578, + "step": 490 + }, + { + "epoch": 2.263698135029637, + "grad_norm": 0.16174978017807007, + "learning_rate": 2.453703703703704e-06, + "loss": 9.2311, + "step": 491 + }, + { + "epoch": 2.2683244181003324, + "grad_norm": 0.20344924926757812, + "learning_rate": 2.438271604938272e-06, + "loss": 9.8727, + "step": 492 + }, + { + "epoch": 2.2729507011710277, + "grad_norm": 0.19355060160160065, + "learning_rate": 2.4228395061728395e-06, + "loss": 9.3433, + "step": 493 + }, + { + "epoch": 2.2775769842417235, + "grad_norm": 0.17609727382659912, + "learning_rate": 2.4074074074074075e-06, + "loss": 8.8289, + "step": 494 + }, + { + "epoch": 2.282203267312419, + "grad_norm": 0.16068409383296967, + "learning_rate": 2.3919753086419755e-06, + "loss": 9.1044, + "step": 495 + }, + { + "epoch": 2.286829550383114, + "grad_norm": 0.15089063346385956, + "learning_rate": 2.3765432098765435e-06, + "loss": 8.8363, + "step": 496 + }, + { + "epoch": 2.2914558334538095, + "grad_norm": 0.3163600265979767, + "learning_rate": 2.361111111111111e-06, + "loss": 9.3244, + "step": 497 + }, + { + "epoch": 2.296082116524505, + "grad_norm": 0.18847279250621796, + "learning_rate": 2.345679012345679e-06, + "loss": 9.1338, + "step": 498 + }, + { + "epoch": 2.3007083995952002, + "grad_norm": 0.23115108907222748, + "learning_rate": 2.330246913580247e-06, + "loss": 9.6425, + "step": 499 + }, + { + "epoch": 2.3053346826658956, + "grad_norm": 0.15251374244689941, + "learning_rate": 2.314814814814815e-06, + "loss": 9.6591, + "step": 500 + }, + { + "epoch": 2.309960965736591, + "grad_norm": 0.1706661432981491, + "learning_rate": 2.299382716049383e-06, + "loss": 10.3708, + "step": 501 + }, + { + "epoch": 2.3145872488072863, + "grad_norm": 0.16492144763469696, + "learning_rate": 2.283950617283951e-06, + "loss": 8.8977, + "step": 502 + }, + { + "epoch": 2.3192135318779816, + "grad_norm": 0.15373440086841583, + "learning_rate": 2.268518518518519e-06, + "loss": 10.9027, + "step": 503 + }, + { + "epoch": 2.323839814948677, + "grad_norm": 0.18270155787467957, + "learning_rate": 2.2530864197530865e-06, + "loss": 9.8633, + "step": 504 + }, + { + "epoch": 2.3284660980193728, + "grad_norm": 0.22537460923194885, + "learning_rate": 2.2376543209876545e-06, + "loss": 8.6262, + "step": 505 + }, + { + "epoch": 2.333092381090068, + "grad_norm": 0.1788664162158966, + "learning_rate": 2.222222222222222e-06, + "loss": 8.3806, + "step": 506 + }, + { + "epoch": 2.3377186641607635, + "grad_norm": 0.2230851948261261, + "learning_rate": 2.2067901234567905e-06, + "loss": 9.6656, + "step": 507 + }, + { + "epoch": 2.342344947231459, + "grad_norm": 0.157254159450531, + "learning_rate": 2.191358024691358e-06, + "loss": 9.0586, + "step": 508 + }, + { + "epoch": 2.346971230302154, + "grad_norm": 0.21109546720981598, + "learning_rate": 2.175925925925926e-06, + "loss": 10.4577, + "step": 509 + }, + { + "epoch": 2.3515975133728495, + "grad_norm": 0.16909867525100708, + "learning_rate": 2.1604938271604937e-06, + "loss": 9.5222, + "step": 510 + }, + { + "epoch": 2.356223796443545, + "grad_norm": 0.1652510166168213, + "learning_rate": 2.145061728395062e-06, + "loss": 9.5384, + "step": 511 + }, + { + "epoch": 2.36085007951424, + "grad_norm": 0.14154984056949615, + "learning_rate": 2.1296296296296298e-06, + "loss": 9.1477, + "step": 512 + }, + { + "epoch": 2.3654763625849355, + "grad_norm": 0.18620611727237701, + "learning_rate": 2.114197530864198e-06, + "loss": 9.8372, + "step": 513 + }, + { + "epoch": 2.370102645655631, + "grad_norm": 0.17792826890945435, + "learning_rate": 2.0987654320987654e-06, + "loss": 9.4475, + "step": 514 + }, + { + "epoch": 2.3747289287263262, + "grad_norm": 0.17278362810611725, + "learning_rate": 2.0833333333333334e-06, + "loss": 8.2241, + "step": 515 + }, + { + "epoch": 2.379355211797022, + "grad_norm": 0.17132383584976196, + "learning_rate": 2.0679012345679015e-06, + "loss": 10.9298, + "step": 516 + }, + { + "epoch": 2.3839814948677174, + "grad_norm": 0.15168847143650055, + "learning_rate": 2.052469135802469e-06, + "loss": 9.0072, + "step": 517 + }, + { + "epoch": 2.3886077779384127, + "grad_norm": 0.186196431517601, + "learning_rate": 2.037037037037037e-06, + "loss": 9.0923, + "step": 518 + }, + { + "epoch": 2.393234061009108, + "grad_norm": 0.12113353610038757, + "learning_rate": 2.021604938271605e-06, + "loss": 10.696, + "step": 519 + }, + { + "epoch": 2.3978603440798034, + "grad_norm": 0.25200334191322327, + "learning_rate": 2.006172839506173e-06, + "loss": 8.5775, + "step": 520 + }, + { + "epoch": 2.4024866271504988, + "grad_norm": 0.1717706322669983, + "learning_rate": 1.9907407407407407e-06, + "loss": 9.8223, + "step": 521 + }, + { + "epoch": 2.407112910221194, + "grad_norm": 0.21262796223163605, + "learning_rate": 1.9753086419753087e-06, + "loss": 9.1006, + "step": 522 + }, + { + "epoch": 2.4117391932918895, + "grad_norm": 0.15240874886512756, + "learning_rate": 1.9598765432098768e-06, + "loss": 8.8133, + "step": 523 + }, + { + "epoch": 2.416365476362585, + "grad_norm": 0.20154468715190887, + "learning_rate": 1.944444444444445e-06, + "loss": 8.9349, + "step": 524 + }, + { + "epoch": 2.42099175943328, + "grad_norm": 0.2465580701828003, + "learning_rate": 1.9290123456790124e-06, + "loss": 9.7463, + "step": 525 + }, + { + "epoch": 2.4256180425039755, + "grad_norm": 0.262588769197464, + "learning_rate": 1.9135802469135804e-06, + "loss": 8.9882, + "step": 526 + }, + { + "epoch": 2.4302443255746713, + "grad_norm": 0.1886514276266098, + "learning_rate": 1.8981481481481484e-06, + "loss": 10.1046, + "step": 527 + }, + { + "epoch": 2.4348706086453666, + "grad_norm": 0.15457630157470703, + "learning_rate": 1.8827160493827162e-06, + "loss": 9.1876, + "step": 528 + }, + { + "epoch": 2.439496891716062, + "grad_norm": 0.18536067008972168, + "learning_rate": 1.867283950617284e-06, + "loss": 8.9485, + "step": 529 + }, + { + "epoch": 2.4441231747867573, + "grad_norm": 0.16247576475143433, + "learning_rate": 1.8518518518518519e-06, + "loss": 9.7982, + "step": 530 + }, + { + "epoch": 2.4487494578574527, + "grad_norm": 0.13770771026611328, + "learning_rate": 1.83641975308642e-06, + "loss": 9.9136, + "step": 531 + }, + { + "epoch": 2.453375740928148, + "grad_norm": 0.1654641479253769, + "learning_rate": 1.820987654320988e-06, + "loss": 9.1883, + "step": 532 + }, + { + "epoch": 2.4580020239988434, + "grad_norm": 0.17362841963768005, + "learning_rate": 1.8055555555555557e-06, + "loss": 9.0925, + "step": 533 + }, + { + "epoch": 2.4626283070695387, + "grad_norm": 0.15451891720294952, + "learning_rate": 1.7901234567901235e-06, + "loss": 9.477, + "step": 534 + }, + { + "epoch": 2.467254590140234, + "grad_norm": 0.14848832786083221, + "learning_rate": 1.7746913580246916e-06, + "loss": 10.4986, + "step": 535 + }, + { + "epoch": 2.4718808732109294, + "grad_norm": 0.18573352694511414, + "learning_rate": 1.7592592592592594e-06, + "loss": 9.2772, + "step": 536 + }, + { + "epoch": 2.4765071562816248, + "grad_norm": 0.16295011341571808, + "learning_rate": 1.7438271604938272e-06, + "loss": 10.3985, + "step": 537 + }, + { + "epoch": 2.4811334393523206, + "grad_norm": 0.1890224814414978, + "learning_rate": 1.7283950617283952e-06, + "loss": 8.1752, + "step": 538 + }, + { + "epoch": 2.485759722423016, + "grad_norm": 0.1372375637292862, + "learning_rate": 1.7129629629629632e-06, + "loss": 8.903, + "step": 539 + }, + { + "epoch": 2.4903860054937113, + "grad_norm": 0.17448656260967255, + "learning_rate": 1.697530864197531e-06, + "loss": 8.5784, + "step": 540 + }, + { + "epoch": 2.4950122885644066, + "grad_norm": 0.1701819896697998, + "learning_rate": 1.6820987654320989e-06, + "loss": 8.7201, + "step": 541 + }, + { + "epoch": 2.499638571635102, + "grad_norm": 0.14620746672153473, + "learning_rate": 1.6666666666666667e-06, + "loss": 9.211, + "step": 542 + }, + { + "epoch": 2.5042648547057973, + "grad_norm": 0.16162265837192535, + "learning_rate": 1.651234567901235e-06, + "loss": 9.4963, + "step": 543 + }, + { + "epoch": 2.5088911377764926, + "grad_norm": 0.15886524319648743, + "learning_rate": 1.6358024691358027e-06, + "loss": 8.8424, + "step": 544 + }, + { + "epoch": 2.513517420847188, + "grad_norm": 0.14556364715099335, + "learning_rate": 1.6203703703703705e-06, + "loss": 7.9769, + "step": 545 + }, + { + "epoch": 2.5181437039178833, + "grad_norm": 0.21578781306743622, + "learning_rate": 1.6049382716049383e-06, + "loss": 8.9909, + "step": 546 + }, + { + "epoch": 2.522769986988579, + "grad_norm": 0.18850085139274597, + "learning_rate": 1.5895061728395064e-06, + "loss": 9.8237, + "step": 547 + }, + { + "epoch": 2.527396270059274, + "grad_norm": 0.1415654569864273, + "learning_rate": 1.5740740740740742e-06, + "loss": 8.9615, + "step": 548 + }, + { + "epoch": 2.53202255312997, + "grad_norm": 0.17242810130119324, + "learning_rate": 1.5586419753086422e-06, + "loss": 8.9936, + "step": 549 + }, + { + "epoch": 2.536648836200665, + "grad_norm": 0.20581774413585663, + "learning_rate": 1.54320987654321e-06, + "loss": 9.376, + "step": 550 + }, + { + "epoch": 2.5412751192713605, + "grad_norm": 0.17876863479614258, + "learning_rate": 1.527777777777778e-06, + "loss": 8.9055, + "step": 551 + }, + { + "epoch": 2.545901402342056, + "grad_norm": 0.1470121592283249, + "learning_rate": 1.5123456790123458e-06, + "loss": 9.3424, + "step": 552 + }, + { + "epoch": 2.550527685412751, + "grad_norm": 0.23757006227970123, + "learning_rate": 1.4969135802469136e-06, + "loss": 9.7866, + "step": 553 + }, + { + "epoch": 2.5551539684834466, + "grad_norm": 0.16553503274917603, + "learning_rate": 1.4814814814814815e-06, + "loss": 8.6868, + "step": 554 + }, + { + "epoch": 2.559780251554142, + "grad_norm": 0.12771116197109222, + "learning_rate": 1.4660493827160497e-06, + "loss": 8.9301, + "step": 555 + }, + { + "epoch": 2.5644065346248373, + "grad_norm": 0.35117506980895996, + "learning_rate": 1.4506172839506175e-06, + "loss": 9.3037, + "step": 556 + }, + { + "epoch": 2.5690328176955326, + "grad_norm": 0.1245264932513237, + "learning_rate": 1.4351851851851853e-06, + "loss": 9.0822, + "step": 557 + }, + { + "epoch": 2.5736591007662284, + "grad_norm": 0.18772459030151367, + "learning_rate": 1.4197530864197531e-06, + "loss": 9.146, + "step": 558 + }, + { + "epoch": 2.5782853838369233, + "grad_norm": 0.18729011714458466, + "learning_rate": 1.4043209876543211e-06, + "loss": 8.183, + "step": 559 + }, + { + "epoch": 2.582911666907619, + "grad_norm": 0.12940698862075806, + "learning_rate": 1.3888888888888892e-06, + "loss": 9.4745, + "step": 560 + }, + { + "epoch": 2.5875379499783144, + "grad_norm": 0.16963091492652893, + "learning_rate": 1.373456790123457e-06, + "loss": 8.8193, + "step": 561 + }, + { + "epoch": 2.59216423304901, + "grad_norm": 0.17672613263130188, + "learning_rate": 1.3580246913580248e-06, + "loss": 8.2093, + "step": 562 + }, + { + "epoch": 2.596790516119705, + "grad_norm": 0.24918967485427856, + "learning_rate": 1.3425925925925928e-06, + "loss": 9.2257, + "step": 563 + }, + { + "epoch": 2.6014167991904005, + "grad_norm": 0.15989799797534943, + "learning_rate": 1.3271604938271606e-06, + "loss": 9.08, + "step": 564 + }, + { + "epoch": 2.606043082261096, + "grad_norm": 0.14997775852680206, + "learning_rate": 1.3117283950617284e-06, + "loss": 9.8745, + "step": 565 + }, + { + "epoch": 2.610669365331791, + "grad_norm": 0.17529721558094025, + "learning_rate": 1.2962962962962962e-06, + "loss": 8.8959, + "step": 566 + }, + { + "epoch": 2.6152956484024865, + "grad_norm": 0.16525782644748688, + "learning_rate": 1.2808641975308645e-06, + "loss": 8.2331, + "step": 567 + }, + { + "epoch": 2.619921931473182, + "grad_norm": 0.14114739000797272, + "learning_rate": 1.2654320987654323e-06, + "loss": 8.1592, + "step": 568 + }, + { + "epoch": 2.6245482145438777, + "grad_norm": 0.21027730405330658, + "learning_rate": 1.25e-06, + "loss": 9.1182, + "step": 569 + }, + { + "epoch": 2.6291744976145726, + "grad_norm": 0.13383808732032776, + "learning_rate": 1.234567901234568e-06, + "loss": 8.8642, + "step": 570 + }, + { + "epoch": 2.6338007806852684, + "grad_norm": 0.15454170107841492, + "learning_rate": 1.219135802469136e-06, + "loss": 7.1809, + "step": 571 + }, + { + "epoch": 2.6384270637559637, + "grad_norm": 0.1575554460287094, + "learning_rate": 1.2037037037037037e-06, + "loss": 8.5097, + "step": 572 + }, + { + "epoch": 2.643053346826659, + "grad_norm": 0.16009144484996796, + "learning_rate": 1.1882716049382718e-06, + "loss": 8.8008, + "step": 573 + }, + { + "epoch": 2.6476796298973544, + "grad_norm": 0.18514862656593323, + "learning_rate": 1.1728395061728396e-06, + "loss": 8.3351, + "step": 574 + }, + { + "epoch": 2.6523059129680497, + "grad_norm": 0.16186843812465668, + "learning_rate": 1.1574074074074076e-06, + "loss": 8.702, + "step": 575 + }, + { + "epoch": 2.656932196038745, + "grad_norm": 0.19223704934120178, + "learning_rate": 1.1419753086419754e-06, + "loss": 9.1931, + "step": 576 + }, + { + "epoch": 2.6615584791094404, + "grad_norm": 0.1866094172000885, + "learning_rate": 1.1265432098765432e-06, + "loss": 9.5196, + "step": 577 + }, + { + "epoch": 2.666184762180136, + "grad_norm": 0.2299501746892929, + "learning_rate": 1.111111111111111e-06, + "loss": 9.4269, + "step": 578 + }, + { + "epoch": 2.670811045250831, + "grad_norm": 0.19736136496067047, + "learning_rate": 1.095679012345679e-06, + "loss": 9.3998, + "step": 579 + }, + { + "epoch": 2.675437328321527, + "grad_norm": 0.16987043619155884, + "learning_rate": 1.0802469135802469e-06, + "loss": 8.2513, + "step": 580 + }, + { + "epoch": 2.680063611392222, + "grad_norm": 0.18940360844135284, + "learning_rate": 1.0648148148148149e-06, + "loss": 7.8771, + "step": 581 + }, + { + "epoch": 2.6846898944629176, + "grad_norm": 0.16819283366203308, + "learning_rate": 1.0493827160493827e-06, + "loss": 8.9367, + "step": 582 + }, + { + "epoch": 2.689316177533613, + "grad_norm": 0.14170341193675995, + "learning_rate": 1.0339506172839507e-06, + "loss": 7.1555, + "step": 583 + }, + { + "epoch": 2.6939424606043083, + "grad_norm": 0.16584675014019012, + "learning_rate": 1.0185185185185185e-06, + "loss": 9.7405, + "step": 584 + }, + { + "epoch": 2.6985687436750037, + "grad_norm": 0.15231122076511383, + "learning_rate": 1.0030864197530866e-06, + "loss": 9.8814, + "step": 585 + }, + { + "epoch": 2.703195026745699, + "grad_norm": 0.14071619510650635, + "learning_rate": 9.876543209876544e-07, + "loss": 8.0466, + "step": 586 + }, + { + "epoch": 2.7078213098163944, + "grad_norm": 0.21130667626857758, + "learning_rate": 9.722222222222224e-07, + "loss": 10.1551, + "step": 587 + }, + { + "epoch": 2.7124475928870897, + "grad_norm": 0.14736154675483704, + "learning_rate": 9.567901234567902e-07, + "loss": 7.5648, + "step": 588 + }, + { + "epoch": 2.717073875957785, + "grad_norm": 0.18979448080062866, + "learning_rate": 9.413580246913581e-07, + "loss": 9.5863, + "step": 589 + }, + { + "epoch": 2.7217001590284804, + "grad_norm": 0.14803054928779602, + "learning_rate": 9.259259259259259e-07, + "loss": 9.2012, + "step": 590 + }, + { + "epoch": 2.726326442099176, + "grad_norm": 0.24644902348518372, + "learning_rate": 9.10493827160494e-07, + "loss": 7.9091, + "step": 591 + }, + { + "epoch": 2.730952725169871, + "grad_norm": 0.18411195278167725, + "learning_rate": 8.950617283950618e-07, + "loss": 9.0894, + "step": 592 + }, + { + "epoch": 2.735579008240567, + "grad_norm": 0.21505457162857056, + "learning_rate": 8.796296296296297e-07, + "loss": 7.4988, + "step": 593 + }, + { + "epoch": 2.7402052913112622, + "grad_norm": 0.18174538016319275, + "learning_rate": 8.641975308641976e-07, + "loss": 8.2626, + "step": 594 + }, + { + "epoch": 2.7448315743819576, + "grad_norm": 0.1982118785381317, + "learning_rate": 8.487654320987655e-07, + "loss": 8.415, + "step": 595 + }, + { + "epoch": 2.749457857452653, + "grad_norm": 0.18941333889961243, + "learning_rate": 8.333333333333333e-07, + "loss": 8.931, + "step": 596 + }, + { + "epoch": 2.7540841405233483, + "grad_norm": 0.2093527615070343, + "learning_rate": 8.179012345679014e-07, + "loss": 9.6871, + "step": 597 + }, + { + "epoch": 2.7587104235940436, + "grad_norm": 0.25371572375297546, + "learning_rate": 8.024691358024692e-07, + "loss": 9.719, + "step": 598 + }, + { + "epoch": 2.763336706664739, + "grad_norm": 0.13767634332180023, + "learning_rate": 7.870370370370371e-07, + "loss": 7.0931, + "step": 599 + }, + { + "epoch": 2.7679629897354343, + "grad_norm": 0.18252308666706085, + "learning_rate": 7.71604938271605e-07, + "loss": 9.0813, + "step": 600 + }, + { + "epoch": 2.7725892728061297, + "grad_norm": 0.13574370741844177, + "learning_rate": 7.561728395061729e-07, + "loss": 8.8218, + "step": 601 + }, + { + "epoch": 2.7772155558768254, + "grad_norm": 0.17202846705913544, + "learning_rate": 7.407407407407407e-07, + "loss": 8.954, + "step": 602 + }, + { + "epoch": 2.7818418389475204, + "grad_norm": 0.18582294881343842, + "learning_rate": 7.253086419753087e-07, + "loss": 7.7078, + "step": 603 + }, + { + "epoch": 2.786468122018216, + "grad_norm": 0.27104949951171875, + "learning_rate": 7.098765432098766e-07, + "loss": 8.3435, + "step": 604 + }, + { + "epoch": 2.7910944050889115, + "grad_norm": 0.32489752769470215, + "learning_rate": 6.944444444444446e-07, + "loss": 8.5035, + "step": 605 + }, + { + "epoch": 2.795720688159607, + "grad_norm": 0.1454378217458725, + "learning_rate": 6.790123456790124e-07, + "loss": 8.4067, + "step": 606 + }, + { + "epoch": 2.800346971230302, + "grad_norm": 0.20940132439136505, + "learning_rate": 6.635802469135803e-07, + "loss": 9.4978, + "step": 607 + }, + { + "epoch": 2.8049732543009975, + "grad_norm": 0.17158959805965424, + "learning_rate": 6.481481481481481e-07, + "loss": 9.3416, + "step": 608 + }, + { + "epoch": 2.809599537371693, + "grad_norm": 0.17591050267219543, + "learning_rate": 6.327160493827161e-07, + "loss": 8.7415, + "step": 609 + }, + { + "epoch": 2.8142258204423882, + "grad_norm": 0.18380632996559143, + "learning_rate": 6.17283950617284e-07, + "loss": 8.2376, + "step": 610 + }, + { + "epoch": 2.8188521035130836, + "grad_norm": 0.19443170726299286, + "learning_rate": 6.018518518518519e-07, + "loss": 7.562, + "step": 611 + }, + { + "epoch": 2.823478386583779, + "grad_norm": 0.24004080891609192, + "learning_rate": 5.864197530864198e-07, + "loss": 9.346, + "step": 612 + }, + { + "epoch": 2.8281046696544747, + "grad_norm": 0.19434408843517303, + "learning_rate": 5.709876543209877e-07, + "loss": 6.8662, + "step": 613 + }, + { + "epoch": 2.8327309527251696, + "grad_norm": 0.1413394808769226, + "learning_rate": 5.555555555555555e-07, + "loss": 7.7949, + "step": 614 + }, + { + "epoch": 2.8373572357958654, + "grad_norm": 0.21421702206134796, + "learning_rate": 5.401234567901234e-07, + "loss": 8.6497, + "step": 615 + }, + { + "epoch": 2.8419835188665608, + "grad_norm": 0.18243171274662018, + "learning_rate": 5.246913580246914e-07, + "loss": 9.0258, + "step": 616 + }, + { + "epoch": 2.846609801937256, + "grad_norm": 0.15553410351276398, + "learning_rate": 5.092592592592593e-07, + "loss": 9.0613, + "step": 617 + }, + { + "epoch": 2.8512360850079514, + "grad_norm": 0.1925041526556015, + "learning_rate": 4.938271604938272e-07, + "loss": 8.8279, + "step": 618 + }, + { + "epoch": 2.855862368078647, + "grad_norm": 0.25465813279151917, + "learning_rate": 4.783950617283951e-07, + "loss": 8.5684, + "step": 619 + }, + { + "epoch": 2.860488651149342, + "grad_norm": 0.17943061888217926, + "learning_rate": 4.6296296296296297e-07, + "loss": 8.9017, + "step": 620 + }, + { + "epoch": 2.8651149342200375, + "grad_norm": 0.22506891191005707, + "learning_rate": 4.475308641975309e-07, + "loss": 8.7791, + "step": 621 + }, + { + "epoch": 2.869741217290733, + "grad_norm": 0.1924242228269577, + "learning_rate": 4.320987654320988e-07, + "loss": 8.4592, + "step": 622 + }, + { + "epoch": 2.874367500361428, + "grad_norm": 0.1336522102355957, + "learning_rate": 4.1666666666666667e-07, + "loss": 9.2921, + "step": 623 + }, + { + "epoch": 2.878993783432124, + "grad_norm": 0.16631512343883514, + "learning_rate": 4.012345679012346e-07, + "loss": 8.7771, + "step": 624 + }, + { + "epoch": 2.883620066502819, + "grad_norm": 0.13000570237636566, + "learning_rate": 3.858024691358025e-07, + "loss": 7.6173, + "step": 625 + }, + { + "epoch": 2.8882463495735147, + "grad_norm": 0.16806651651859283, + "learning_rate": 3.7037037037037036e-07, + "loss": 8.2835, + "step": 626 + }, + { + "epoch": 2.89287263264421, + "grad_norm": 0.16624124348163605, + "learning_rate": 3.549382716049383e-07, + "loss": 9.0295, + "step": 627 + }, + { + "epoch": 2.8974989157149054, + "grad_norm": 0.18300163745880127, + "learning_rate": 3.395061728395062e-07, + "loss": 9.3673, + "step": 628 + }, + { + "epoch": 2.9021251987856007, + "grad_norm": 0.1591711938381195, + "learning_rate": 3.2407407407407406e-07, + "loss": 9.2437, + "step": 629 + }, + { + "epoch": 2.906751481856296, + "grad_norm": 0.1935225874185562, + "learning_rate": 3.08641975308642e-07, + "loss": 8.1388, + "step": 630 + }, + { + "epoch": 2.9113777649269914, + "grad_norm": 0.18095123767852783, + "learning_rate": 2.932098765432099e-07, + "loss": 8.8998, + "step": 631 + }, + { + "epoch": 2.9160040479976868, + "grad_norm": 0.15141603350639343, + "learning_rate": 2.7777777777777776e-07, + "loss": 9.5553, + "step": 632 + }, + { + "epoch": 2.920630331068382, + "grad_norm": 0.1664353758096695, + "learning_rate": 2.623456790123457e-07, + "loss": 9.2671, + "step": 633 + }, + { + "epoch": 2.9252566141390774, + "grad_norm": 0.14757901430130005, + "learning_rate": 2.469135802469136e-07, + "loss": 9.3753, + "step": 634 + }, + { + "epoch": 2.9298828972097732, + "grad_norm": 0.12832246720790863, + "learning_rate": 2.3148148148148148e-07, + "loss": 7.55, + "step": 635 + }, + { + "epoch": 2.934509180280468, + "grad_norm": 0.14796899259090424, + "learning_rate": 2.160493827160494e-07, + "loss": 8.9455, + "step": 636 + }, + { + "epoch": 2.939135463351164, + "grad_norm": 0.16375280916690826, + "learning_rate": 2.006172839506173e-07, + "loss": 9.1212, + "step": 637 + }, + { + "epoch": 2.9437617464218593, + "grad_norm": 0.15986521542072296, + "learning_rate": 1.8518518518518518e-07, + "loss": 8.0966, + "step": 638 + }, + { + "epoch": 2.9483880294925546, + "grad_norm": 0.14000752568244934, + "learning_rate": 1.697530864197531e-07, + "loss": 8.1283, + "step": 639 + }, + { + "epoch": 2.95301431256325, + "grad_norm": 0.10660137236118317, + "learning_rate": 1.54320987654321e-07, + "loss": 7.7013, + "step": 640 + }, + { + "epoch": 2.9576405956339453, + "grad_norm": 0.12879547476768494, + "learning_rate": 1.3888888888888888e-07, + "loss": 8.2472, + "step": 641 + }, + { + "epoch": 2.9622668787046407, + "grad_norm": 0.24027042090892792, + "learning_rate": 1.234567901234568e-07, + "loss": 9.3202, + "step": 642 + }, + { + "epoch": 2.966893161775336, + "grad_norm": 0.18989317119121552, + "learning_rate": 1.080246913580247e-07, + "loss": 8.6135, + "step": 643 + }, + { + "epoch": 2.9715194448460314, + "grad_norm": 0.20127813518047333, + "learning_rate": 9.259259259259259e-08, + "loss": 8.3557, + "step": 644 + }, + { + "epoch": 2.9761457279167267, + "grad_norm": 0.20816563069820404, + "learning_rate": 7.71604938271605e-08, + "loss": 9.6997, + "step": 645 + }, + { + "epoch": 2.9807720109874225, + "grad_norm": 0.2171671837568283, + "learning_rate": 6.17283950617284e-08, + "loss": 9.6809, + "step": 646 + }, + { + "epoch": 2.9853982940581174, + "grad_norm": 0.18569743633270264, + "learning_rate": 4.6296296296296295e-08, + "loss": 8.4721, + "step": 647 + }, + { + "epoch": 2.990024577128813, + "grad_norm": 0.16961927711963654, + "learning_rate": 3.08641975308642e-08, + "loss": 9.1963, + "step": 648 + }, + { + "epoch": 2.990024577128813, + "step": 648, + "total_flos": 2.9427853089130414e+18, + "train_loss": 10.740490390930647, + "train_runtime": 52216.2733, + "train_samples_per_second": 1.59, + "train_steps_per_second": 0.012 + } + ], + "logging_steps": 1.0, + "max_steps": 648, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 50, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 2.9427853089130414e+18, + "train_batch_size": 2, + "trial_name": null, + "trial_params": null +}