{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 500, "global_step": 327, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0061162079510703364, "grad_norm": 0.8125, "learning_rate": 2.3529411764705885e-06, "loss": 1.96957528591156, "step": 2 }, { "epoch": 0.012232415902140673, "grad_norm": 0.65234375, "learning_rate": 7.058823529411766e-06, "loss": 2.0347838401794434, "step": 4 }, { "epoch": 0.01834862385321101, "grad_norm": 0.578125, "learning_rate": 1.1764705882352942e-05, "loss": 1.9940425157546997, "step": 6 }, { "epoch": 0.024464831804281346, "grad_norm": 0.640625, "learning_rate": 1.647058823529412e-05, "loss": 2.06846284866333, "step": 8 }, { "epoch": 0.03058103975535168, "grad_norm": 0.80859375, "learning_rate": 2.1176470588235296e-05, "loss": 1.9072656631469727, "step": 10 }, { "epoch": 0.03669724770642202, "grad_norm": 0.62109375, "learning_rate": 2.5882352941176475e-05, "loss": 1.8469384908676147, "step": 12 }, { "epoch": 0.04281345565749235, "grad_norm": 0.67578125, "learning_rate": 3.0588235294117644e-05, "loss": 1.822761058807373, "step": 14 }, { "epoch": 0.04892966360856269, "grad_norm": 0.59765625, "learning_rate": 3.529411764705883e-05, "loss": 1.7572168111801147, "step": 16 }, { "epoch": 0.05504587155963303, "grad_norm": 0.45703125, "learning_rate": 4e-05, "loss": 1.7028628587722778, "step": 18 }, { "epoch": 0.06116207951070336, "grad_norm": 0.6796875, "learning_rate": 3.999630287622123e-05, "loss": 1.8212647438049316, "step": 20 }, { "epoch": 0.0672782874617737, "grad_norm": 0.40234375, "learning_rate": 3.9985213023632064e-05, "loss": 1.6563327312469482, "step": 22 }, { "epoch": 0.07339449541284404, "grad_norm": 0.4921875, "learning_rate": 3.996673499785002e-05, "loss": 1.6510090827941895, "step": 24 }, { "epoch": 0.07951070336391437, "grad_norm": 0.7109375, "learning_rate": 3.99408763894916e-05, "loss": 1.5338315963745117, "step": 26 }, { "epoch": 0.0856269113149847, "grad_norm": 0.404296875, "learning_rate": 3.9907647821054114e-05, "loss": 1.451381802558899, "step": 28 }, { "epoch": 0.09174311926605505, "grad_norm": 0.30859375, "learning_rate": 3.986706294255207e-05, "loss": 1.5903527736663818, "step": 30 }, { "epoch": 0.09785932721712538, "grad_norm": 0.5078125, "learning_rate": 3.981913842590985e-05, "loss": 1.5307936668395996, "step": 32 }, { "epoch": 0.10397553516819572, "grad_norm": 0.314453125, "learning_rate": 3.9763893958113005e-05, "loss": 1.413094401359558, "step": 34 }, { "epoch": 0.11009174311926606, "grad_norm": 0.625, "learning_rate": 3.970135223312103e-05, "loss": 1.4268063306808472, "step": 36 }, { "epoch": 0.1162079510703364, "grad_norm": 0.291015625, "learning_rate": 3.96315389425449e-05, "loss": 1.4923566579818726, "step": 38 }, { "epoch": 0.12232415902140673, "grad_norm": 0.298828125, "learning_rate": 3.9554482765093134e-05, "loss": 1.574055790901184, "step": 40 }, { "epoch": 0.12844036697247707, "grad_norm": 0.474609375, "learning_rate": 3.947021535479083e-05, "loss": 1.415814995765686, "step": 42 }, { "epoch": 0.1345565749235474, "grad_norm": 0.37109375, "learning_rate": 3.937877132797649e-05, "loss": 1.4422385692596436, "step": 44 }, { "epoch": 0.14067278287461774, "grad_norm": 0.375, "learning_rate": 3.928018824908187e-05, "loss": 1.4337655305862427, "step": 46 }, { "epoch": 0.14678899082568808, "grad_norm": 0.478515625, "learning_rate": 3.917450661520088e-05, "loss": 1.5849357843399048, "step": 48 }, { "epoch": 0.1529051987767584, "grad_norm": 0.578125, "learning_rate": 3.906176983945371e-05, "loss": 1.5441187620162964, "step": 50 }, { "epoch": 0.15902140672782875, "grad_norm": 0.310546875, "learning_rate": 3.894202423315306e-05, "loss": 1.4055734872817993, "step": 52 }, { "epoch": 0.1651376146788991, "grad_norm": 0.36328125, "learning_rate": 3.8815318986779875e-05, "loss": 1.543579339981079, "step": 54 }, { "epoch": 0.1712538226299694, "grad_norm": 0.328125, "learning_rate": 3.868170614977628e-05, "loss": 1.4346128702163696, "step": 56 }, { "epoch": 0.17737003058103976, "grad_norm": 0.310546875, "learning_rate": 3.854124060916415e-05, "loss": 1.5318936109542847, "step": 58 }, { "epoch": 0.1834862385321101, "grad_norm": 0.4140625, "learning_rate": 3.839398006699797e-05, "loss": 1.3749366998672485, "step": 60 }, { "epoch": 0.18960244648318042, "grad_norm": 0.47265625, "learning_rate": 3.823998501666131e-05, "loss": 1.436793565750122, "step": 62 }, { "epoch": 0.19571865443425077, "grad_norm": 0.40625, "learning_rate": 3.8079318718016666e-05, "loss": 1.3264566659927368, "step": 64 }, { "epoch": 0.2018348623853211, "grad_norm": 1.34375, "learning_rate": 3.7912047171418815e-05, "loss": 1.4578938484191895, "step": 66 }, { "epoch": 0.20795107033639143, "grad_norm": 0.490234375, "learning_rate": 3.773823909060248e-05, "loss": 1.3888816833496094, "step": 68 }, { "epoch": 0.21406727828746178, "grad_norm": 0.33984375, "learning_rate": 3.755796587445528e-05, "loss": 1.4028871059417725, "step": 70 }, { "epoch": 0.22018348623853212, "grad_norm": 0.287109375, "learning_rate": 3.7371301577687666e-05, "loss": 1.3480572700500488, "step": 72 }, { "epoch": 0.22629969418960244, "grad_norm": 1.4453125, "learning_rate": 3.717832288041188e-05, "loss": 1.4603177309036255, "step": 74 }, { "epoch": 0.2324159021406728, "grad_norm": 0.3984375, "learning_rate": 3.697910905664243e-05, "loss": 1.4285566806793213, "step": 76 }, { "epoch": 0.23853211009174313, "grad_norm": 0.353515625, "learning_rate": 3.6773741941730975e-05, "loss": 1.462870478630066, "step": 78 }, { "epoch": 0.24464831804281345, "grad_norm": 0.578125, "learning_rate": 3.656230589874905e-05, "loss": 1.3728106021881104, "step": 80 }, { "epoch": 0.25076452599388377, "grad_norm": 0.35546875, "learning_rate": 3.6344887783832474e-05, "loss": 1.3789210319519043, "step": 82 }, { "epoch": 0.25688073394495414, "grad_norm": 0.318359375, "learning_rate": 3.6121576910501517e-05, "loss": 1.438659429550171, "step": 84 }, { "epoch": 0.26299694189602446, "grad_norm": 0.6796875, "learning_rate": 3.589246501297172e-05, "loss": 1.4267497062683105, "step": 86 }, { "epoch": 0.2691131498470948, "grad_norm": 0.265625, "learning_rate": 3.565764620847024e-05, "loss": 1.297946572303772, "step": 88 }, { "epoch": 0.27522935779816515, "grad_norm": 0.50390625, "learning_rate": 3.541721695857328e-05, "loss": 1.3422534465789795, "step": 90 }, { "epoch": 0.28134556574923547, "grad_norm": 0.302734375, "learning_rate": 3.5171276029580485e-05, "loss": 1.3573137521743774, "step": 92 }, { "epoch": 0.2874617737003058, "grad_norm": 0.58203125, "learning_rate": 3.4919924451942625e-05, "loss": 1.4575246572494507, "step": 94 }, { "epoch": 0.29357798165137616, "grad_norm": 0.376953125, "learning_rate": 3.466326547875907e-05, "loss": 1.3683665990829468, "step": 96 }, { "epoch": 0.2996941896024465, "grad_norm": 0.244140625, "learning_rate": 3.440140454336236e-05, "loss": 1.2624497413635254, "step": 98 }, { "epoch": 0.3058103975535168, "grad_norm": 0.369140625, "learning_rate": 3.413444921600705e-05, "loss": 1.4250640869140625, "step": 100 }, { "epoch": 0.3119266055045872, "grad_norm": 0.69921875, "learning_rate": 3.3862509159680775e-05, "loss": 1.4060899019241333, "step": 102 }, { "epoch": 0.3180428134556575, "grad_norm": 0.416015625, "learning_rate": 3.3585696085055594e-05, "loss": 1.44656503200531, "step": 104 }, { "epoch": 0.3241590214067278, "grad_norm": 0.63671875, "learning_rate": 3.330412370459823e-05, "loss": 1.5084904432296753, "step": 106 }, { "epoch": 0.3302752293577982, "grad_norm": 0.408203125, "learning_rate": 3.301790768585793e-05, "loss": 1.4001638889312744, "step": 108 }, { "epoch": 0.3363914373088685, "grad_norm": 0.337890625, "learning_rate": 3.272716560395123e-05, "loss": 1.4370101690292358, "step": 110 }, { "epoch": 0.3425076452599388, "grad_norm": 0.61328125, "learning_rate": 3.243201689326306e-05, "loss": 1.4086581468582153, "step": 112 }, { "epoch": 0.3486238532110092, "grad_norm": 0.48828125, "learning_rate": 3.213258279838416e-05, "loss": 1.389402151107788, "step": 114 }, { "epoch": 0.3547400611620795, "grad_norm": 0.3515625, "learning_rate": 3.182898632430479e-05, "loss": 1.371553897857666, "step": 116 }, { "epoch": 0.36085626911314983, "grad_norm": 0.392578125, "learning_rate": 3.1521352185885326e-05, "loss": 1.4507074356079102, "step": 118 }, { "epoch": 0.3669724770642202, "grad_norm": 0.703125, "learning_rate": 3.120980675662449e-05, "loss": 1.3349779844284058, "step": 120 }, { "epoch": 0.3730886850152905, "grad_norm": 0.357421875, "learning_rate": 3.0894478016746106e-05, "loss": 1.3503204584121704, "step": 122 }, { "epoch": 0.37920489296636084, "grad_norm": 0.435546875, "learning_rate": 3.0575495500626015e-05, "loss": 1.4363230466842651, "step": 124 }, { "epoch": 0.3853211009174312, "grad_norm": 0.25390625, "learning_rate": 3.025299024358036e-05, "loss": 1.3254035711288452, "step": 126 }, { "epoch": 0.39143730886850153, "grad_norm": 0.357421875, "learning_rate": 2.9927094728037422e-05, "loss": 1.410750389099121, "step": 128 }, { "epoch": 0.39755351681957185, "grad_norm": 0.4609375, "learning_rate": 2.9597942829114976e-05, "loss": 1.397456169128418, "step": 130 }, { "epoch": 0.4036697247706422, "grad_norm": 0.5, "learning_rate": 2.926566975962551e-05, "loss": 1.3679834604263306, "step": 132 }, { "epoch": 0.40978593272171254, "grad_norm": 0.380859375, "learning_rate": 2.8930412014531924e-05, "loss": 1.4191218614578247, "step": 134 }, { "epoch": 0.41590214067278286, "grad_norm": 0.41796875, "learning_rate": 2.859230731487661e-05, "loss": 1.2690709829330444, "step": 136 }, { "epoch": 0.42201834862385323, "grad_norm": 0.439453125, "learning_rate": 2.8251494551206767e-05, "loss": 1.3012070655822754, "step": 138 }, { "epoch": 0.42813455657492355, "grad_norm": 0.3515625, "learning_rate": 2.7908113726519356e-05, "loss": 1.2863593101501465, "step": 140 }, { "epoch": 0.43425076452599387, "grad_norm": 0.498046875, "learning_rate": 2.7562305898749054e-05, "loss": 1.442170262336731, "step": 142 }, { "epoch": 0.44036697247706424, "grad_norm": 0.3984375, "learning_rate": 2.7214213122822864e-05, "loss": 1.3614513874053955, "step": 144 }, { "epoch": 0.44648318042813456, "grad_norm": 0.28125, "learning_rate": 2.6863978392305118e-05, "loss": 1.2583218812942505, "step": 146 }, { "epoch": 0.4525993883792049, "grad_norm": 0.349609375, "learning_rate": 2.651174558065697e-05, "loss": 1.4321554899215698, "step": 148 }, { "epoch": 0.45871559633027525, "grad_norm": 0.341796875, "learning_rate": 2.6157659382134384e-05, "loss": 1.414106845855713, "step": 150 }, { "epoch": 0.4648318042813456, "grad_norm": 0.57421875, "learning_rate": 2.5801865252348935e-05, "loss": 1.4366655349731445, "step": 152 }, { "epoch": 0.4709480122324159, "grad_norm": 0.392578125, "learning_rate": 2.5444509348515912e-05, "loss": 1.3970258235931396, "step": 154 }, { "epoch": 0.47706422018348627, "grad_norm": 3.078125, "learning_rate": 2.508573846941417e-05, "loss": 1.4000701904296875, "step": 156 }, { "epoch": 0.4831804281345566, "grad_norm": 0.431640625, "learning_rate": 2.472569999508238e-05, "loss": 1.4267001152038574, "step": 158 }, { "epoch": 0.4892966360856269, "grad_norm": 0.33203125, "learning_rate": 2.4364541826276638e-05, "loss": 1.4192265272140503, "step": 160 }, { "epoch": 0.4954128440366973, "grad_norm": 0.283203125, "learning_rate": 2.4002412323714026e-05, "loss": 1.314334511756897, "step": 162 }, { "epoch": 0.5015290519877675, "grad_norm": 0.56640625, "learning_rate": 2.363946024712732e-05, "loss": 1.355038046836853, "step": 164 }, { "epoch": 0.5076452599388379, "grad_norm": 0.6484375, "learning_rate": 2.3275834694155716e-05, "loss": 1.2806464433670044, "step": 166 }, { "epoch": 0.5137614678899083, "grad_norm": 0.6015625, "learning_rate": 2.2911685039096834e-05, "loss": 1.3307536840438843, "step": 168 }, { "epoch": 0.5198776758409785, "grad_norm": 0.333984375, "learning_rate": 2.2547160871544973e-05, "loss": 1.3525010347366333, "step": 170 }, { "epoch": 0.5259938837920489, "grad_norm": 0.404296875, "learning_rate": 2.2182411934941004e-05, "loss": 1.3771812915802002, "step": 172 }, { "epoch": 0.5321100917431193, "grad_norm": 0.337890625, "learning_rate": 2.1817588065059008e-05, "loss": 1.342699408531189, "step": 174 }, { "epoch": 0.5382262996941896, "grad_norm": 0.52734375, "learning_rate": 2.145283912845504e-05, "loss": 1.3631731271743774, "step": 176 }, { "epoch": 0.5443425076452599, "grad_norm": 0.4375, "learning_rate": 2.1088314960903172e-05, "loss": 1.3468397855758667, "step": 178 }, { "epoch": 0.5504587155963303, "grad_norm": 0.341796875, "learning_rate": 2.072416530584429e-05, "loss": 1.3079278469085693, "step": 180 }, { "epoch": 0.5565749235474006, "grad_norm": 0.478515625, "learning_rate": 2.0360539752872688e-05, "loss": 1.3927726745605469, "step": 182 }, { "epoch": 0.5626911314984709, "grad_norm": 0.71875, "learning_rate": 1.9997587676285976e-05, "loss": 1.3745498657226562, "step": 184 }, { "epoch": 0.5688073394495413, "grad_norm": 0.259765625, "learning_rate": 1.9635458173723365e-05, "loss": 1.3934822082519531, "step": 186 }, { "epoch": 0.5749235474006116, "grad_norm": 0.3203125, "learning_rate": 1.9274300004917625e-05, "loss": 1.4668513536453247, "step": 188 }, { "epoch": 0.581039755351682, "grad_norm": 0.375, "learning_rate": 1.8914261530585842e-05, "loss": 1.2894922494888306, "step": 190 }, { "epoch": 0.5871559633027523, "grad_norm": 0.322265625, "learning_rate": 1.855549065148409e-05, "loss": 1.3256540298461914, "step": 192 }, { "epoch": 0.5932721712538226, "grad_norm": 0.92578125, "learning_rate": 1.8198134747651067e-05, "loss": 1.3802309036254883, "step": 194 }, { "epoch": 0.599388379204893, "grad_norm": 0.33203125, "learning_rate": 1.7842340617865625e-05, "loss": 1.3032485246658325, "step": 196 }, { "epoch": 0.6055045871559633, "grad_norm": 0.54296875, "learning_rate": 1.748825441934303e-05, "loss": 1.236058235168457, "step": 198 }, { "epoch": 0.6116207951070336, "grad_norm": 0.376953125, "learning_rate": 1.713602160769489e-05, "loss": 1.4403380155563354, "step": 200 }, { "epoch": 0.617737003058104, "grad_norm": 0.359375, "learning_rate": 1.6785786877177145e-05, "loss": 1.3596861362457275, "step": 202 }, { "epoch": 0.6238532110091743, "grad_norm": 0.408203125, "learning_rate": 1.643769410125095e-05, "loss": 1.3777844905853271, "step": 204 }, { "epoch": 0.6299694189602446, "grad_norm": 0.4609375, "learning_rate": 1.609188627348065e-05, "loss": 1.3549814224243164, "step": 206 }, { "epoch": 0.636085626911315, "grad_norm": 0.60546875, "learning_rate": 1.574850544879324e-05, "loss": 1.2403396368026733, "step": 208 }, { "epoch": 0.6422018348623854, "grad_norm": 0.421875, "learning_rate": 1.5407692685123392e-05, "loss": 1.3685152530670166, "step": 210 }, { "epoch": 0.6483180428134556, "grad_norm": 0.298828125, "learning_rate": 1.5069587985468078e-05, "loss": 1.3501070737838745, "step": 212 }, { "epoch": 0.654434250764526, "grad_norm": 0.625, "learning_rate": 1.4734330240374504e-05, "loss": 1.292812466621399, "step": 214 }, { "epoch": 0.6605504587155964, "grad_norm": 0.3125, "learning_rate": 1.4402057170885026e-05, "loss": 1.287298560142517, "step": 216 }, { "epoch": 0.6666666666666666, "grad_norm": 0.27734375, "learning_rate": 1.4072905271962585e-05, "loss": 1.3475459814071655, "step": 218 }, { "epoch": 0.672782874617737, "grad_norm": 0.376953125, "learning_rate": 1.3747009756419657e-05, "loss": 1.2376188039779663, "step": 220 }, { "epoch": 0.6788990825688074, "grad_norm": 0.47265625, "learning_rate": 1.3424504499373994e-05, "loss": 1.315477728843689, "step": 222 }, { "epoch": 0.6850152905198776, "grad_norm": 0.365234375, "learning_rate": 1.31055219832539e-05, "loss": 1.364182949066162, "step": 224 }, { "epoch": 0.691131498470948, "grad_norm": 0.326171875, "learning_rate": 1.2790193243375521e-05, "loss": 1.3197802305221558, "step": 226 }, { "epoch": 0.6972477064220184, "grad_norm": 0.26171875, "learning_rate": 1.2478647814114683e-05, "loss": 1.2909021377563477, "step": 228 }, { "epoch": 0.7033639143730887, "grad_norm": 0.3984375, "learning_rate": 1.2171013675695222e-05, "loss": 1.2902398109436035, "step": 230 }, { "epoch": 0.709480122324159, "grad_norm": 0.8515625, "learning_rate": 1.1867417201615848e-05, "loss": 1.3986810445785522, "step": 232 }, { "epoch": 0.7155963302752294, "grad_norm": 0.29296875, "learning_rate": 1.1567983106736946e-05, "loss": 1.2758734226226807, "step": 234 }, { "epoch": 0.7217125382262997, "grad_norm": 0.37109375, "learning_rate": 1.1272834396048777e-05, "loss": 1.317265272140503, "step": 236 }, { "epoch": 0.72782874617737, "grad_norm": 3.296875, "learning_rate": 1.0982092314142068e-05, "loss": 1.4039820432662964, "step": 238 }, { "epoch": 0.7339449541284404, "grad_norm": 0.40234375, "learning_rate": 1.0695876295401771e-05, "loss": 1.330233097076416, "step": 240 }, { "epoch": 0.7400611620795107, "grad_norm": 1.7890625, "learning_rate": 1.041430391494441e-05, "loss": 1.3207037448883057, "step": 242 }, { "epoch": 0.746177370030581, "grad_norm": 0.2451171875, "learning_rate": 1.013749084031923e-05, "loss": 1.3986611366271973, "step": 244 }, { "epoch": 0.7522935779816514, "grad_norm": 0.42578125, "learning_rate": 9.86555078399295e-06, "loss": 1.3146181106567383, "step": 246 }, { "epoch": 0.7584097859327217, "grad_norm": 0.310546875, "learning_rate": 9.598595456637642e-06, "loss": 1.3460875749588013, "step": 248 }, { "epoch": 0.764525993883792, "grad_norm": 0.267578125, "learning_rate": 9.336734521240934e-06, "loss": 1.2672019004821777, "step": 250 }, { "epoch": 0.7706422018348624, "grad_norm": 0.291015625, "learning_rate": 9.080075548057383e-06, "loss": 1.3179923295974731, "step": 252 }, { "epoch": 0.7767584097859327, "grad_norm": 0.703125, "learning_rate": 8.828723970419519e-06, "loss": 1.4186333417892456, "step": 254 }, { "epoch": 0.7828746177370031, "grad_norm": 0.390625, "learning_rate": 8.582783041426728e-06, "loss": 1.2970129251480103, "step": 256 }, { "epoch": 0.7889908256880734, "grad_norm": 0.21875, "learning_rate": 8.342353791529765e-06, "loss": 1.3273588418960571, "step": 258 }, { "epoch": 0.7951070336391437, "grad_norm": 0.37109375, "learning_rate": 8.107534987028286e-06, "loss": 1.3187005519866943, "step": 260 }, { "epoch": 0.8012232415902141, "grad_norm": 0.345703125, "learning_rate": 7.878423089498492e-06, "loss": 1.288523554801941, "step": 262 }, { "epoch": 0.8073394495412844, "grad_norm": 0.6484375, "learning_rate": 7.655112216167533e-06, "loss": 1.3323516845703125, "step": 264 }, { "epoch": 0.8134556574923547, "grad_norm": 0.345703125, "learning_rate": 7.437694101250949e-06, "loss": 1.3667104244232178, "step": 266 }, { "epoch": 0.8195718654434251, "grad_norm": 0.294921875, "learning_rate": 7.226258058269031e-06, "loss": 1.354952096939087, "step": 268 }, { "epoch": 0.8256880733944955, "grad_norm": 0.42578125, "learning_rate": 7.020890943357573e-06, "loss": 1.2679226398468018, "step": 270 }, { "epoch": 0.8318042813455657, "grad_norm": 0.447265625, "learning_rate": 6.821677119588124e-06, "loss": 1.3207221031188965, "step": 272 }, { "epoch": 0.8379204892966361, "grad_norm": 0.62109375, "learning_rate": 6.628698422312341e-06, "loss": 1.4085071086883545, "step": 274 }, { "epoch": 0.8440366972477065, "grad_norm": 0.474609375, "learning_rate": 6.442034125544726e-06, "loss": 1.3547308444976807, "step": 276 }, { "epoch": 0.8501529051987767, "grad_norm": 0.31640625, "learning_rate": 6.261760909397523e-06, "loss": 1.346556305885315, "step": 278 }, { "epoch": 0.8562691131498471, "grad_norm": 0.46875, "learning_rate": 6.087952828581187e-06, "loss": 1.3009809255599976, "step": 280 }, { "epoch": 0.8623853211009175, "grad_norm": 0.8671875, "learning_rate": 5.920681281983339e-06, "loss": 1.3994947671890259, "step": 282 }, { "epoch": 0.8685015290519877, "grad_norm": 0.23046875, "learning_rate": 5.760014983338687e-06, "loss": 1.291093111038208, "step": 284 }, { "epoch": 0.8746177370030581, "grad_norm": 0.83984375, "learning_rate": 5.606019933002025e-06, "loss": 1.2565807104110718, "step": 286 }, { "epoch": 0.8807339449541285, "grad_norm": 0.326171875, "learning_rate": 5.458759390835851e-06, "loss": 1.2513775825500488, "step": 288 }, { "epoch": 0.8868501529051988, "grad_norm": 0.49609375, "learning_rate": 5.318293850223726e-06, "loss": 1.3214612007141113, "step": 290 }, { "epoch": 0.8929663608562691, "grad_norm": 0.5859375, "learning_rate": 5.184681013220134e-06, "loss": 1.4418059587478638, "step": 292 }, { "epoch": 0.8990825688073395, "grad_norm": 0.29296875, "learning_rate": 5.057975766846941e-06, "loss": 1.3208688497543335, "step": 294 }, { "epoch": 0.9051987767584098, "grad_norm": 0.66796875, "learning_rate": 4.938230160546292e-06, "loss": 1.4112577438354492, "step": 296 }, { "epoch": 0.9113149847094801, "grad_norm": 0.322265625, "learning_rate": 4.825493384799122e-06, "loss": 1.356927514076233, "step": 298 }, { "epoch": 0.9174311926605505, "grad_norm": 0.357421875, "learning_rate": 4.719811750918136e-06, "loss": 1.3079042434692383, "step": 300 }, { "epoch": 0.9235474006116208, "grad_norm": 0.40234375, "learning_rate": 4.621228672023517e-06, "loss": 1.2854888439178467, "step": 302 }, { "epoch": 0.9296636085626911, "grad_norm": 0.2890625, "learning_rate": 4.529784645209172e-06, "loss": 1.2977776527404785, "step": 304 }, { "epoch": 0.9357798165137615, "grad_norm": 0.30859375, "learning_rate": 4.4455172349068696e-06, "loss": 1.2907675504684448, "step": 306 }, { "epoch": 0.9418960244648318, "grad_norm": 0.53515625, "learning_rate": 4.3684610574551e-06, "loss": 1.2298390865325928, "step": 308 }, { "epoch": 0.9480122324159022, "grad_norm": 0.33984375, "learning_rate": 4.298647766878974e-06, "loss": 1.3057259321212769, "step": 310 }, { "epoch": 0.9541284403669725, "grad_norm": 0.3125, "learning_rate": 4.236106041887007e-06, "loss": 1.3313639163970947, "step": 312 }, { "epoch": 0.9602446483180428, "grad_norm": 0.66015625, "learning_rate": 4.180861574090156e-06, "loss": 1.3252257108688354, "step": 314 }, { "epoch": 0.9663608562691132, "grad_norm": 0.6171875, "learning_rate": 4.1329370574479296e-06, "loss": 1.258653163909912, "step": 316 }, { "epoch": 0.9724770642201835, "grad_norm": 0.6328125, "learning_rate": 4.0923521789458876e-06, "loss": 1.3827545642852783, "step": 318 }, { "epoch": 0.9785932721712538, "grad_norm": 1.2578125, "learning_rate": 4.059123610508407e-06, "loss": 1.2586948871612549, "step": 320 }, { "epoch": 0.9847094801223242, "grad_norm": 0.333984375, "learning_rate": 4.033265002149985e-06, "loss": 1.1728885173797607, "step": 322 }, { "epoch": 0.9908256880733946, "grad_norm": 0.37109375, "learning_rate": 4.014786976367939e-06, "loss": 1.230262041091919, "step": 324 }, { "epoch": 0.9969418960244648, "grad_norm": 0.78125, "learning_rate": 4.00369712377877e-06, "loss": 1.3188743591308594, "step": 326 }, { "epoch": 1.0, "step": 327, "total_flos": 8.183535699912294e+17, "train_loss": 1.4031060467437138, "train_runtime": 3980.916, "train_samples_per_second": 2.629, "train_steps_per_second": 0.082 } ], "logging_steps": 2, "max_steps": 327, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 9999999, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 8.183535699912294e+17, "train_batch_size": 1, "trial_name": null, "trial_params": null }