diff --git "a/checkpoint-5222/trainer_state.json" "b/checkpoint-5222/trainer_state.json" new file mode 100644--- /dev/null +++ "b/checkpoint-5222/trainer_state.json" @@ -0,0 +1,36587 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.999808539153743, + "eval_steps": 500, + "global_step": 5222, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003829216925138809, + "grad_norm": 1.5127280950546265, + "learning_rate": 5.0000000000000004e-08, + "loss": 0.9601, + "step": 1 + }, + { + "epoch": 0.0007658433850277618, + "grad_norm": 1.4570214748382568, + "learning_rate": 1.0000000000000001e-07, + "loss": 0.9752, + "step": 2 + }, + { + "epoch": 0.0011487650775416428, + "grad_norm": 1.4853640794754028, + "learning_rate": 1.5000000000000002e-07, + "loss": 0.9485, + "step": 3 + }, + { + "epoch": 0.0015316867700555237, + "grad_norm": 1.6019998788833618, + "learning_rate": 2.0000000000000002e-07, + "loss": 0.9711, + "step": 4 + }, + { + "epoch": 0.0019146084625694046, + "grad_norm": 1.4343098402023315, + "learning_rate": 2.5000000000000004e-07, + "loss": 0.947, + "step": 5 + }, + { + "epoch": 0.0022975301550832855, + "grad_norm": 1.5140701532363892, + "learning_rate": 3.0000000000000004e-07, + "loss": 0.9413, + "step": 6 + }, + { + "epoch": 0.0026804518475971664, + "grad_norm": 1.5071738958358765, + "learning_rate": 3.5000000000000004e-07, + "loss": 0.9399, + "step": 7 + }, + { + "epoch": 0.0030633735401110473, + "grad_norm": 1.6219124794006348, + "learning_rate": 4.0000000000000003e-07, + "loss": 0.9393, + "step": 8 + }, + { + "epoch": 0.0034462952326249283, + "grad_norm": 1.548447847366333, + "learning_rate": 4.5000000000000003e-07, + "loss": 0.978, + "step": 9 + }, + { + "epoch": 0.003829216925138809, + "grad_norm": 1.5101799964904785, + "learning_rate": 5.000000000000001e-07, + "loss": 0.968, + "step": 10 + }, + { + "epoch": 0.00421213861765269, + "grad_norm": 1.3384488821029663, + "learning_rate": 5.5e-07, + "loss": 0.9415, + "step": 11 + }, + { + "epoch": 0.004595060310166571, + "grad_norm": 1.3894020318984985, + "learning_rate": 6.000000000000001e-07, + "loss": 0.985, + "step": 12 + }, + { + "epoch": 0.004977982002680452, + "grad_norm": 1.3622794151306152, + "learning_rate": 6.5e-07, + "loss": 0.9326, + "step": 13 + }, + { + "epoch": 0.005360903695194333, + "grad_norm": 1.3284450769424438, + "learning_rate": 7.000000000000001e-07, + "loss": 0.9442, + "step": 14 + }, + { + "epoch": 0.005743825387708214, + "grad_norm": 1.2537797689437866, + "learning_rate": 7.5e-07, + "loss": 0.9426, + "step": 15 + }, + { + "epoch": 0.006126747080222095, + "grad_norm": 1.330793857574463, + "learning_rate": 8.000000000000001e-07, + "loss": 0.965, + "step": 16 + }, + { + "epoch": 0.006509668772735976, + "grad_norm": 1.1742899417877197, + "learning_rate": 8.500000000000001e-07, + "loss": 0.9486, + "step": 17 + }, + { + "epoch": 0.0068925904652498565, + "grad_norm": 1.1290744543075562, + "learning_rate": 9.000000000000001e-07, + "loss": 0.9652, + "step": 18 + }, + { + "epoch": 0.0072755121577637374, + "grad_norm": 1.0592819452285767, + "learning_rate": 9.500000000000001e-07, + "loss": 0.9177, + "step": 19 + }, + { + "epoch": 0.007658433850277618, + "grad_norm": 0.9967736005783081, + "learning_rate": 1.0000000000000002e-06, + "loss": 0.9321, + "step": 20 + }, + { + "epoch": 0.0080413555427915, + "grad_norm": 0.9203951358795166, + "learning_rate": 1.0500000000000001e-06, + "loss": 0.869, + "step": 21 + }, + { + "epoch": 0.00842427723530538, + "grad_norm": 0.8977105021476746, + "learning_rate": 1.1e-06, + "loss": 0.9206, + "step": 22 + }, + { + "epoch": 0.008807198927819261, + "grad_norm": 0.8208876252174377, + "learning_rate": 1.1500000000000002e-06, + "loss": 0.8671, + "step": 23 + }, + { + "epoch": 0.009190120620333142, + "grad_norm": 0.8222814202308655, + "learning_rate": 1.2000000000000002e-06, + "loss": 0.9111, + "step": 24 + }, + { + "epoch": 0.009573042312847023, + "grad_norm": 0.7808475494384766, + "learning_rate": 1.25e-06, + "loss": 0.9222, + "step": 25 + }, + { + "epoch": 0.009955964005360904, + "grad_norm": 0.7570652961730957, + "learning_rate": 1.3e-06, + "loss": 0.9151, + "step": 26 + }, + { + "epoch": 0.010338885697874785, + "grad_norm": 0.7235645651817322, + "learning_rate": 1.3500000000000002e-06, + "loss": 0.9068, + "step": 27 + }, + { + "epoch": 0.010721807390388666, + "grad_norm": 0.6783406138420105, + "learning_rate": 1.4000000000000001e-06, + "loss": 0.863, + "step": 28 + }, + { + "epoch": 0.011104729082902547, + "grad_norm": 0.6298149228096008, + "learning_rate": 1.45e-06, + "loss": 0.8956, + "step": 29 + }, + { + "epoch": 0.011487650775416428, + "grad_norm": 0.6101565957069397, + "learning_rate": 1.5e-06, + "loss": 0.8933, + "step": 30 + }, + { + "epoch": 0.011870572467930308, + "grad_norm": 0.5784317255020142, + "learning_rate": 1.5500000000000002e-06, + "loss": 0.8739, + "step": 31 + }, + { + "epoch": 0.01225349416044419, + "grad_norm": 0.5533347725868225, + "learning_rate": 1.6000000000000001e-06, + "loss": 0.8753, + "step": 32 + }, + { + "epoch": 0.01263641585295807, + "grad_norm": 0.5175893306732178, + "learning_rate": 1.6500000000000003e-06, + "loss": 0.8757, + "step": 33 + }, + { + "epoch": 0.013019337545471951, + "grad_norm": 0.4827329218387604, + "learning_rate": 1.7000000000000002e-06, + "loss": 0.8784, + "step": 34 + }, + { + "epoch": 0.013402259237985832, + "grad_norm": 0.4603060781955719, + "learning_rate": 1.75e-06, + "loss": 0.8488, + "step": 35 + }, + { + "epoch": 0.013785180930499713, + "grad_norm": 0.429062157869339, + "learning_rate": 1.8000000000000001e-06, + "loss": 0.8601, + "step": 36 + }, + { + "epoch": 0.014168102623013594, + "grad_norm": 0.4230949878692627, + "learning_rate": 1.85e-06, + "loss": 0.8592, + "step": 37 + }, + { + "epoch": 0.014551024315527475, + "grad_norm": 0.3980255126953125, + "learning_rate": 1.9000000000000002e-06, + "loss": 0.8387, + "step": 38 + }, + { + "epoch": 0.014933946008041356, + "grad_norm": 0.39369726181030273, + "learning_rate": 1.9500000000000004e-06, + "loss": 0.8565, + "step": 39 + }, + { + "epoch": 0.015316867700555237, + "grad_norm": 0.3655986189842224, + "learning_rate": 2.0000000000000003e-06, + "loss": 0.8545, + "step": 40 + }, + { + "epoch": 0.015699789393069118, + "grad_norm": 0.3628844916820526, + "learning_rate": 2.05e-06, + "loss": 0.8191, + "step": 41 + }, + { + "epoch": 0.016082711085583, + "grad_norm": 0.3656681180000305, + "learning_rate": 2.1000000000000002e-06, + "loss": 0.8506, + "step": 42 + }, + { + "epoch": 0.01646563277809688, + "grad_norm": 0.35703447461128235, + "learning_rate": 2.15e-06, + "loss": 0.8395, + "step": 43 + }, + { + "epoch": 0.01684855447061076, + "grad_norm": 0.34360966086387634, + "learning_rate": 2.2e-06, + "loss": 0.8379, + "step": 44 + }, + { + "epoch": 0.01723147616312464, + "grad_norm": 0.3302212357521057, + "learning_rate": 2.25e-06, + "loss": 0.8214, + "step": 45 + }, + { + "epoch": 0.017614397855638522, + "grad_norm": 0.32497096061706543, + "learning_rate": 2.3000000000000004e-06, + "loss": 0.8152, + "step": 46 + }, + { + "epoch": 0.017997319548152403, + "grad_norm": 0.3161090910434723, + "learning_rate": 2.35e-06, + "loss": 0.8113, + "step": 47 + }, + { + "epoch": 0.018380241240666284, + "grad_norm": 0.30970415472984314, + "learning_rate": 2.4000000000000003e-06, + "loss": 0.8126, + "step": 48 + }, + { + "epoch": 0.018763162933180165, + "grad_norm": 0.3097487986087799, + "learning_rate": 2.4500000000000003e-06, + "loss": 0.8378, + "step": 49 + }, + { + "epoch": 0.019146084625694046, + "grad_norm": 0.28779536485671997, + "learning_rate": 2.5e-06, + "loss": 0.8102, + "step": 50 + }, + { + "epoch": 0.019529006318207927, + "grad_norm": 0.2880657911300659, + "learning_rate": 2.55e-06, + "loss": 0.8065, + "step": 51 + }, + { + "epoch": 0.019911928010721808, + "grad_norm": 0.283342182636261, + "learning_rate": 2.6e-06, + "loss": 0.8131, + "step": 52 + }, + { + "epoch": 0.02029484970323569, + "grad_norm": 0.26725032925605774, + "learning_rate": 2.6500000000000005e-06, + "loss": 0.8193, + "step": 53 + }, + { + "epoch": 0.02067777139574957, + "grad_norm": 0.2662472128868103, + "learning_rate": 2.7000000000000004e-06, + "loss": 0.8171, + "step": 54 + }, + { + "epoch": 0.02106069308826345, + "grad_norm": 0.25971609354019165, + "learning_rate": 2.7500000000000004e-06, + "loss": 0.8067, + "step": 55 + }, + { + "epoch": 0.02144361478077733, + "grad_norm": 0.2470993995666504, + "learning_rate": 2.8000000000000003e-06, + "loss": 0.7906, + "step": 56 + }, + { + "epoch": 0.021826536473291212, + "grad_norm": 0.24413476884365082, + "learning_rate": 2.85e-06, + "loss": 0.813, + "step": 57 + }, + { + "epoch": 0.022209458165805093, + "grad_norm": 0.2281661480665207, + "learning_rate": 2.9e-06, + "loss": 0.8004, + "step": 58 + }, + { + "epoch": 0.022592379858318974, + "grad_norm": 0.2510649263858795, + "learning_rate": 2.95e-06, + "loss": 0.8144, + "step": 59 + }, + { + "epoch": 0.022975301550832855, + "grad_norm": 0.22505076229572296, + "learning_rate": 3e-06, + "loss": 0.7959, + "step": 60 + }, + { + "epoch": 0.023358223243346736, + "grad_norm": 0.22958475351333618, + "learning_rate": 3.05e-06, + "loss": 0.7918, + "step": 61 + }, + { + "epoch": 0.023741144935860617, + "grad_norm": 0.2242196947336197, + "learning_rate": 3.1000000000000004e-06, + "loss": 0.8004, + "step": 62 + }, + { + "epoch": 0.024124066628374498, + "grad_norm": 0.23278626799583435, + "learning_rate": 3.1500000000000003e-06, + "loss": 0.8133, + "step": 63 + }, + { + "epoch": 0.02450698832088838, + "grad_norm": 0.230966255068779, + "learning_rate": 3.2000000000000003e-06, + "loss": 0.8119, + "step": 64 + }, + { + "epoch": 0.02488991001340226, + "grad_norm": 0.21737371385097504, + "learning_rate": 3.2500000000000002e-06, + "loss": 0.8, + "step": 65 + }, + { + "epoch": 0.02527283170591614, + "grad_norm": 0.21784932911396027, + "learning_rate": 3.3000000000000006e-06, + "loss": 0.813, + "step": 66 + }, + { + "epoch": 0.02565575339843002, + "grad_norm": 0.21986863017082214, + "learning_rate": 3.3500000000000005e-06, + "loss": 0.7739, + "step": 67 + }, + { + "epoch": 0.026038675090943902, + "grad_norm": 0.2227390855550766, + "learning_rate": 3.4000000000000005e-06, + "loss": 0.7865, + "step": 68 + }, + { + "epoch": 0.026421596783457783, + "grad_norm": 0.2248571366071701, + "learning_rate": 3.45e-06, + "loss": 0.7959, + "step": 69 + }, + { + "epoch": 0.026804518475971664, + "grad_norm": 0.22709882259368896, + "learning_rate": 3.5e-06, + "loss": 0.8137, + "step": 70 + }, + { + "epoch": 0.027187440168485545, + "grad_norm": 0.2120588719844818, + "learning_rate": 3.5500000000000003e-06, + "loss": 0.7869, + "step": 71 + }, + { + "epoch": 0.027570361860999426, + "grad_norm": 0.20743295550346375, + "learning_rate": 3.6000000000000003e-06, + "loss": 0.7929, + "step": 72 + }, + { + "epoch": 0.027953283553513307, + "grad_norm": 0.20250606536865234, + "learning_rate": 3.65e-06, + "loss": 0.7967, + "step": 73 + }, + { + "epoch": 0.028336205246027188, + "grad_norm": 0.200389102101326, + "learning_rate": 3.7e-06, + "loss": 0.7764, + "step": 74 + }, + { + "epoch": 0.02871912693854107, + "grad_norm": 0.19319339096546173, + "learning_rate": 3.7500000000000005e-06, + "loss": 0.7723, + "step": 75 + }, + { + "epoch": 0.02910204863105495, + "grad_norm": 0.19402000308036804, + "learning_rate": 3.8000000000000005e-06, + "loss": 0.8049, + "step": 76 + }, + { + "epoch": 0.02948497032356883, + "grad_norm": 0.1952792853116989, + "learning_rate": 3.85e-06, + "loss": 0.7721, + "step": 77 + }, + { + "epoch": 0.02986789201608271, + "grad_norm": 0.18888434767723083, + "learning_rate": 3.900000000000001e-06, + "loss": 0.7846, + "step": 78 + }, + { + "epoch": 0.030250813708596593, + "grad_norm": 0.1821848452091217, + "learning_rate": 3.95e-06, + "loss": 0.7753, + "step": 79 + }, + { + "epoch": 0.030633735401110473, + "grad_norm": 0.1820901334285736, + "learning_rate": 4.000000000000001e-06, + "loss": 0.7826, + "step": 80 + }, + { + "epoch": 0.031016657093624354, + "grad_norm": 0.19272948801517487, + "learning_rate": 4.05e-06, + "loss": 0.7989, + "step": 81 + }, + { + "epoch": 0.031399578786138235, + "grad_norm": 0.184035524725914, + "learning_rate": 4.1e-06, + "loss": 0.791, + "step": 82 + }, + { + "epoch": 0.031782500478652116, + "grad_norm": 0.18398649990558624, + "learning_rate": 4.15e-06, + "loss": 0.7768, + "step": 83 + }, + { + "epoch": 0.032165422171166, + "grad_norm": 0.18076884746551514, + "learning_rate": 4.2000000000000004e-06, + "loss": 0.7729, + "step": 84 + }, + { + "epoch": 0.03254834386367988, + "grad_norm": 0.18133755028247833, + "learning_rate": 4.25e-06, + "loss": 0.7854, + "step": 85 + }, + { + "epoch": 0.03293126555619376, + "grad_norm": 0.18535292148590088, + "learning_rate": 4.3e-06, + "loss": 0.774, + "step": 86 + }, + { + "epoch": 0.03331418724870764, + "grad_norm": 0.1858111023902893, + "learning_rate": 4.350000000000001e-06, + "loss": 0.7942, + "step": 87 + }, + { + "epoch": 0.03369710894122152, + "grad_norm": 0.1836749166250229, + "learning_rate": 4.4e-06, + "loss": 0.7923, + "step": 88 + }, + { + "epoch": 0.0340800306337354, + "grad_norm": 0.1837168186903, + "learning_rate": 4.450000000000001e-06, + "loss": 0.7763, + "step": 89 + }, + { + "epoch": 0.03446295232624928, + "grad_norm": 0.17813347280025482, + "learning_rate": 4.5e-06, + "loss": 0.7667, + "step": 90 + }, + { + "epoch": 0.034845874018763164, + "grad_norm": 0.1816832423210144, + "learning_rate": 4.5500000000000005e-06, + "loss": 0.7816, + "step": 91 + }, + { + "epoch": 0.035228795711277044, + "grad_norm": 0.1764446496963501, + "learning_rate": 4.600000000000001e-06, + "loss": 0.755, + "step": 92 + }, + { + "epoch": 0.035611717403790925, + "grad_norm": 0.18107102811336517, + "learning_rate": 4.65e-06, + "loss": 0.7808, + "step": 93 + }, + { + "epoch": 0.035994639096304806, + "grad_norm": 0.17818239331245422, + "learning_rate": 4.7e-06, + "loss": 0.7682, + "step": 94 + }, + { + "epoch": 0.03637756078881869, + "grad_norm": 0.1828058362007141, + "learning_rate": 4.75e-06, + "loss": 0.7313, + "step": 95 + }, + { + "epoch": 0.03676048248133257, + "grad_norm": 0.18249830603599548, + "learning_rate": 4.800000000000001e-06, + "loss": 0.7704, + "step": 96 + }, + { + "epoch": 0.03714340417384645, + "grad_norm": 0.177218496799469, + "learning_rate": 4.85e-06, + "loss": 0.7691, + "step": 97 + }, + { + "epoch": 0.03752632586636033, + "grad_norm": 0.17561639845371246, + "learning_rate": 4.9000000000000005e-06, + "loss": 0.7679, + "step": 98 + }, + { + "epoch": 0.03790924755887421, + "grad_norm": 0.18068744242191315, + "learning_rate": 4.95e-06, + "loss": 0.7807, + "step": 99 + }, + { + "epoch": 0.03829216925138809, + "grad_norm": 0.1780369132757187, + "learning_rate": 5e-06, + "loss": 0.7685, + "step": 100 + }, + { + "epoch": 0.03867509094390197, + "grad_norm": 0.1759374588727951, + "learning_rate": 4.99999979369311e-06, + "loss": 0.7534, + "step": 101 + }, + { + "epoch": 0.039058012636415854, + "grad_norm": 0.1728181093931198, + "learning_rate": 4.999999174772476e-06, + "loss": 0.7606, + "step": 102 + }, + { + "epoch": 0.039440934328929735, + "grad_norm": 0.17361775040626526, + "learning_rate": 4.999998143238196e-06, + "loss": 0.7729, + "step": 103 + }, + { + "epoch": 0.039823856021443615, + "grad_norm": 0.17070014774799347, + "learning_rate": 4.999996699090443e-06, + "loss": 0.7979, + "step": 104 + }, + { + "epoch": 0.040206777713957496, + "grad_norm": 0.19051475822925568, + "learning_rate": 4.999994842329455e-06, + "loss": 0.7766, + "step": 105 + }, + { + "epoch": 0.04058969940647138, + "grad_norm": 0.17698678374290466, + "learning_rate": 4.99999257295554e-06, + "loss": 0.7636, + "step": 106 + }, + { + "epoch": 0.04097262109898526, + "grad_norm": 0.17492914199829102, + "learning_rate": 4.999989890969069e-06, + "loss": 0.7558, + "step": 107 + }, + { + "epoch": 0.04135554279149914, + "grad_norm": 0.17513518035411835, + "learning_rate": 4.9999867963704875e-06, + "loss": 0.7501, + "step": 108 + }, + { + "epoch": 0.04173846448401302, + "grad_norm": 0.1762644648551941, + "learning_rate": 4.9999832891603055e-06, + "loss": 0.7677, + "step": 109 + }, + { + "epoch": 0.0421213861765269, + "grad_norm": 0.1700354665517807, + "learning_rate": 4.999979369339102e-06, + "loss": 0.7462, + "step": 110 + }, + { + "epoch": 0.04250430786904078, + "grad_norm": 0.17454922199249268, + "learning_rate": 4.999975036907523e-06, + "loss": 0.7501, + "step": 111 + }, + { + "epoch": 0.04288722956155466, + "grad_norm": 0.1729852706193924, + "learning_rate": 4.9999702918662855e-06, + "loss": 0.7917, + "step": 112 + }, + { + "epoch": 0.043270151254068544, + "grad_norm": 0.171586275100708, + "learning_rate": 4.99996513421617e-06, + "loss": 0.7846, + "step": 113 + }, + { + "epoch": 0.043653072946582425, + "grad_norm": 0.17341765761375427, + "learning_rate": 4.99995956395803e-06, + "loss": 0.7652, + "step": 114 + }, + { + "epoch": 0.044035994639096306, + "grad_norm": 0.1708601862192154, + "learning_rate": 4.999953581092784e-06, + "loss": 0.7605, + "step": 115 + }, + { + "epoch": 0.044418916331610186, + "grad_norm": 0.17076106369495392, + "learning_rate": 4.999947185621419e-06, + "loss": 0.7434, + "step": 116 + }, + { + "epoch": 0.04480183802412407, + "grad_norm": 0.1779450923204422, + "learning_rate": 4.999940377544991e-06, + "loss": 0.7754, + "step": 117 + }, + { + "epoch": 0.04518475971663795, + "grad_norm": 0.17263121902942657, + "learning_rate": 4.999933156864624e-06, + "loss": 0.7504, + "step": 118 + }, + { + "epoch": 0.04556768140915183, + "grad_norm": 0.17754054069519043, + "learning_rate": 4.999925523581508e-06, + "loss": 0.7669, + "step": 119 + }, + { + "epoch": 0.04595060310166571, + "grad_norm": 0.1822337955236435, + "learning_rate": 4.999917477696906e-06, + "loss": 0.769, + "step": 120 + }, + { + "epoch": 0.04633352479417959, + "grad_norm": 0.17491310834884644, + "learning_rate": 4.999909019212144e-06, + "loss": 0.7679, + "step": 121 + }, + { + "epoch": 0.04671644648669347, + "grad_norm": 0.17979766428470612, + "learning_rate": 4.999900148128617e-06, + "loss": 0.7584, + "step": 122 + }, + { + "epoch": 0.04709936817920735, + "grad_norm": 0.17815561592578888, + "learning_rate": 4.999890864447791e-06, + "loss": 0.7806, + "step": 123 + }, + { + "epoch": 0.047482289871721234, + "grad_norm": 0.18036983907222748, + "learning_rate": 4.999881168171197e-06, + "loss": 0.7639, + "step": 124 + }, + { + "epoch": 0.047865211564235115, + "grad_norm": 0.1740260273218155, + "learning_rate": 4.999871059300436e-06, + "loss": 0.7672, + "step": 125 + }, + { + "epoch": 0.048248133256748996, + "grad_norm": 0.17238426208496094, + "learning_rate": 4.9998605378371765e-06, + "loss": 0.762, + "step": 126 + }, + { + "epoch": 0.04863105494926288, + "grad_norm": 0.17532694339752197, + "learning_rate": 4.999849603783155e-06, + "loss": 0.7559, + "step": 127 + }, + { + "epoch": 0.04901397664177676, + "grad_norm": 0.18054932355880737, + "learning_rate": 4.999838257140176e-06, + "loss": 0.7727, + "step": 128 + }, + { + "epoch": 0.04939689833429064, + "grad_norm": 0.18592335283756256, + "learning_rate": 4.999826497910112e-06, + "loss": 0.741, + "step": 129 + }, + { + "epoch": 0.04977982002680452, + "grad_norm": 0.1814471334218979, + "learning_rate": 4.999814326094904e-06, + "loss": 0.7467, + "step": 130 + }, + { + "epoch": 0.0501627417193184, + "grad_norm": 0.17422868311405182, + "learning_rate": 4.999801741696561e-06, + "loss": 0.7415, + "step": 131 + }, + { + "epoch": 0.05054566341183228, + "grad_norm": 0.17902833223342896, + "learning_rate": 4.99978874471716e-06, + "loss": 0.7771, + "step": 132 + }, + { + "epoch": 0.05092858510434616, + "grad_norm": 0.17519019544124603, + "learning_rate": 4.999775335158845e-06, + "loss": 0.769, + "step": 133 + }, + { + "epoch": 0.05131150679686004, + "grad_norm": 0.17666171491146088, + "learning_rate": 4.999761513023831e-06, + "loss": 0.7368, + "step": 134 + }, + { + "epoch": 0.051694428489373924, + "grad_norm": 0.17427121102809906, + "learning_rate": 4.999747278314398e-06, + "loss": 0.743, + "step": 135 + }, + { + "epoch": 0.052077350181887805, + "grad_norm": 0.17496870458126068, + "learning_rate": 4.9997326310328954e-06, + "loss": 0.7399, + "step": 136 + }, + { + "epoch": 0.052460271874401686, + "grad_norm": 0.17835672199726105, + "learning_rate": 4.999717571181742e-06, + "loss": 0.7491, + "step": 137 + }, + { + "epoch": 0.05284319356691557, + "grad_norm": 0.17227713763713837, + "learning_rate": 4.999702098763422e-06, + "loss": 0.7351, + "step": 138 + }, + { + "epoch": 0.05322611525942945, + "grad_norm": 0.17822884023189545, + "learning_rate": 4.999686213780489e-06, + "loss": 0.765, + "step": 139 + }, + { + "epoch": 0.05360903695194333, + "grad_norm": 0.17335212230682373, + "learning_rate": 4.999669916235566e-06, + "loss": 0.7558, + "step": 140 + }, + { + "epoch": 0.05399195864445721, + "grad_norm": 0.18416562676429749, + "learning_rate": 4.999653206131342e-06, + "loss": 0.7746, + "step": 141 + }, + { + "epoch": 0.05437488033697109, + "grad_norm": 0.1746789515018463, + "learning_rate": 4.999636083470574e-06, + "loss": 0.7386, + "step": 142 + }, + { + "epoch": 0.05475780202948497, + "grad_norm": 0.17718607187271118, + "learning_rate": 4.99961854825609e-06, + "loss": 0.7565, + "step": 143 + }, + { + "epoch": 0.05514072372199885, + "grad_norm": 0.17604880034923553, + "learning_rate": 4.999600600490783e-06, + "loss": 0.7601, + "step": 144 + }, + { + "epoch": 0.05552364541451273, + "grad_norm": 0.18366846442222595, + "learning_rate": 4.999582240177615e-06, + "loss": 0.7615, + "step": 145 + }, + { + "epoch": 0.055906567107026614, + "grad_norm": 0.18398283421993256, + "learning_rate": 4.999563467319616e-06, + "loss": 0.7554, + "step": 146 + }, + { + "epoch": 0.056289488799540495, + "grad_norm": 0.1746983826160431, + "learning_rate": 4.9995442819198856e-06, + "loss": 0.7382, + "step": 147 + }, + { + "epoch": 0.056672410492054376, + "grad_norm": 0.17418019473552704, + "learning_rate": 4.99952468398159e-06, + "loss": 0.7298, + "step": 148 + }, + { + "epoch": 0.05705533218456826, + "grad_norm": 0.17950575053691864, + "learning_rate": 4.999504673507964e-06, + "loss": 0.7603, + "step": 149 + }, + { + "epoch": 0.05743825387708214, + "grad_norm": 0.18124008178710938, + "learning_rate": 4.999484250502308e-06, + "loss": 0.7541, + "step": 150 + }, + { + "epoch": 0.05782117556959602, + "grad_norm": 0.1779160499572754, + "learning_rate": 4.999463414967994e-06, + "loss": 0.7268, + "step": 151 + }, + { + "epoch": 0.0582040972621099, + "grad_norm": 0.17948390543460846, + "learning_rate": 4.999442166908462e-06, + "loss": 0.7427, + "step": 152 + }, + { + "epoch": 0.05858701895462378, + "grad_norm": 0.17860113084316254, + "learning_rate": 4.999420506327219e-06, + "loss": 0.7661, + "step": 153 + }, + { + "epoch": 0.05896994064713766, + "grad_norm": 0.18167608976364136, + "learning_rate": 4.999398433227838e-06, + "loss": 0.7534, + "step": 154 + }, + { + "epoch": 0.05935286233965154, + "grad_norm": 0.1777266561985016, + "learning_rate": 4.999375947613963e-06, + "loss": 0.7631, + "step": 155 + }, + { + "epoch": 0.05973578403216542, + "grad_norm": 0.18054281175136566, + "learning_rate": 4.999353049489306e-06, + "loss": 0.75, + "step": 156 + }, + { + "epoch": 0.060118705724679304, + "grad_norm": 0.1803414523601532, + "learning_rate": 4.999329738857645e-06, + "loss": 0.7527, + "step": 157 + }, + { + "epoch": 0.060501627417193185, + "grad_norm": 0.18072110414505005, + "learning_rate": 4.999306015722828e-06, + "loss": 0.7661, + "step": 158 + }, + { + "epoch": 0.060884549109707066, + "grad_norm": 0.18042054772377014, + "learning_rate": 4.99928188008877e-06, + "loss": 0.7686, + "step": 159 + }, + { + "epoch": 0.06126747080222095, + "grad_norm": 0.17903763055801392, + "learning_rate": 4.9992573319594554e-06, + "loss": 0.7239, + "step": 160 + }, + { + "epoch": 0.06165039249473483, + "grad_norm": 0.17662514746189117, + "learning_rate": 4.999232371338934e-06, + "loss": 0.7362, + "step": 161 + }, + { + "epoch": 0.06203331418724871, + "grad_norm": 0.1817161589860916, + "learning_rate": 4.999206998231326e-06, + "loss": 0.7553, + "step": 162 + }, + { + "epoch": 0.06241623587976259, + "grad_norm": 0.1797408014535904, + "learning_rate": 4.9991812126408215e-06, + "loss": 0.7302, + "step": 163 + }, + { + "epoch": 0.06279915757227647, + "grad_norm": 0.19102029502391815, + "learning_rate": 4.999155014571673e-06, + "loss": 0.7447, + "step": 164 + }, + { + "epoch": 0.06318207926479034, + "grad_norm": 0.18745839595794678, + "learning_rate": 4.9991284040282064e-06, + "loss": 0.76, + "step": 165 + }, + { + "epoch": 0.06356500095730423, + "grad_norm": 0.18368470668792725, + "learning_rate": 4.999101381014813e-06, + "loss": 0.7423, + "step": 166 + }, + { + "epoch": 0.0639479226498181, + "grad_norm": 0.18375608325004578, + "learning_rate": 4.999073945535952e-06, + "loss": 0.7235, + "step": 167 + }, + { + "epoch": 0.064330844342332, + "grad_norm": 0.17745068669319153, + "learning_rate": 4.9990460975961525e-06, + "loss": 0.7538, + "step": 168 + }, + { + "epoch": 0.06471376603484587, + "grad_norm": 0.18004310131072998, + "learning_rate": 4.999017837200011e-06, + "loss": 0.7487, + "step": 169 + }, + { + "epoch": 0.06509668772735976, + "grad_norm": 0.1802518218755722, + "learning_rate": 4.998989164352191e-06, + "loss": 0.7311, + "step": 170 + }, + { + "epoch": 0.06547960941987363, + "grad_norm": 0.18313384056091309, + "learning_rate": 4.998960079057426e-06, + "loss": 0.7379, + "step": 171 + }, + { + "epoch": 0.06586253111238752, + "grad_norm": 0.1821993589401245, + "learning_rate": 4.9989305813205134e-06, + "loss": 0.7465, + "step": 172 + }, + { + "epoch": 0.06624545280490139, + "grad_norm": 0.18214775621891022, + "learning_rate": 4.998900671146326e-06, + "loss": 0.7282, + "step": 173 + }, + { + "epoch": 0.06662837449741528, + "grad_norm": 0.18141110241413116, + "learning_rate": 4.998870348539797e-06, + "loss": 0.7435, + "step": 174 + }, + { + "epoch": 0.06701129618992915, + "grad_norm": 0.18263347446918488, + "learning_rate": 4.998839613505932e-06, + "loss": 0.7287, + "step": 175 + }, + { + "epoch": 0.06739421788244304, + "grad_norm": 0.1850353181362152, + "learning_rate": 4.998808466049803e-06, + "loss": 0.7257, + "step": 176 + }, + { + "epoch": 0.06777713957495692, + "grad_norm": 0.18254730105400085, + "learning_rate": 4.998776906176553e-06, + "loss": 0.7411, + "step": 177 + }, + { + "epoch": 0.0681600612674708, + "grad_norm": 0.18580451607704163, + "learning_rate": 4.99874493389139e-06, + "loss": 0.757, + "step": 178 + }, + { + "epoch": 0.06854298295998468, + "grad_norm": 0.18148383498191833, + "learning_rate": 4.998712549199589e-06, + "loss": 0.7234, + "step": 179 + }, + { + "epoch": 0.06892590465249857, + "grad_norm": 0.18495115637779236, + "learning_rate": 4.998679752106496e-06, + "loss": 0.7551, + "step": 180 + }, + { + "epoch": 0.06930882634501244, + "grad_norm": 0.18330596387386322, + "learning_rate": 4.998646542617525e-06, + "loss": 0.7452, + "step": 181 + }, + { + "epoch": 0.06969174803752633, + "grad_norm": 0.17899492383003235, + "learning_rate": 4.9986129207381555e-06, + "loss": 0.7054, + "step": 182 + }, + { + "epoch": 0.0700746697300402, + "grad_norm": 0.1845427006483078, + "learning_rate": 4.9985788864739375e-06, + "loss": 0.7393, + "step": 183 + }, + { + "epoch": 0.07045759142255409, + "grad_norm": 0.18768930435180664, + "learning_rate": 4.998544439830488e-06, + "loss": 0.7473, + "step": 184 + }, + { + "epoch": 0.07084051311506796, + "grad_norm": 0.18853271007537842, + "learning_rate": 4.998509580813493e-06, + "loss": 0.7192, + "step": 185 + }, + { + "epoch": 0.07122343480758185, + "grad_norm": 0.19321995973587036, + "learning_rate": 4.998474309428705e-06, + "loss": 0.7518, + "step": 186 + }, + { + "epoch": 0.07160635650009572, + "grad_norm": 0.18516433238983154, + "learning_rate": 4.9984386256819455e-06, + "loss": 0.7253, + "step": 187 + }, + { + "epoch": 0.07198927819260961, + "grad_norm": 0.18672487139701843, + "learning_rate": 4.998402529579104e-06, + "loss": 0.75, + "step": 188 + }, + { + "epoch": 0.07237219988512349, + "grad_norm": 0.19649949669837952, + "learning_rate": 4.998366021126139e-06, + "loss": 0.7248, + "step": 189 + }, + { + "epoch": 0.07275512157763737, + "grad_norm": 0.1804923266172409, + "learning_rate": 4.998329100329073e-06, + "loss": 0.7222, + "step": 190 + }, + { + "epoch": 0.07313804327015125, + "grad_norm": 0.19032728672027588, + "learning_rate": 4.998291767194003e-06, + "loss": 0.7319, + "step": 191 + }, + { + "epoch": 0.07352096496266514, + "grad_norm": 0.1883210986852646, + "learning_rate": 4.998254021727089e-06, + "loss": 0.7417, + "step": 192 + }, + { + "epoch": 0.07390388665517901, + "grad_norm": 0.17950430512428284, + "learning_rate": 4.998215863934561e-06, + "loss": 0.7145, + "step": 193 + }, + { + "epoch": 0.0742868083476929, + "grad_norm": 0.1965007483959198, + "learning_rate": 4.998177293822718e-06, + "loss": 0.7482, + "step": 194 + }, + { + "epoch": 0.07466973004020677, + "grad_norm": 0.18340559303760529, + "learning_rate": 4.998138311397924e-06, + "loss": 0.7101, + "step": 195 + }, + { + "epoch": 0.07505265173272066, + "grad_norm": 0.18560108542442322, + "learning_rate": 4.998098916666613e-06, + "loss": 0.7384, + "step": 196 + }, + { + "epoch": 0.07543557342523453, + "grad_norm": 0.20068591833114624, + "learning_rate": 4.9980591096352885e-06, + "loss": 0.7537, + "step": 197 + }, + { + "epoch": 0.07581849511774842, + "grad_norm": 0.18525315821170807, + "learning_rate": 4.998018890310518e-06, + "loss": 0.7359, + "step": 198 + }, + { + "epoch": 0.0762014168102623, + "grad_norm": 0.18367959558963776, + "learning_rate": 4.997978258698942e-06, + "loss": 0.6975, + "step": 199 + }, + { + "epoch": 0.07658433850277618, + "grad_norm": 0.18198728561401367, + "learning_rate": 4.997937214807265e-06, + "loss": 0.7338, + "step": 200 + }, + { + "epoch": 0.07696726019529006, + "grad_norm": 0.18427884578704834, + "learning_rate": 4.997895758642262e-06, + "loss": 0.7367, + "step": 201 + }, + { + "epoch": 0.07735018188780395, + "grad_norm": 0.18311342597007751, + "learning_rate": 4.997853890210775e-06, + "loss": 0.7326, + "step": 202 + }, + { + "epoch": 0.07773310358031782, + "grad_norm": 0.18621565401554108, + "learning_rate": 4.997811609519714e-06, + "loss": 0.737, + "step": 203 + }, + { + "epoch": 0.07811602527283171, + "grad_norm": 0.1909012496471405, + "learning_rate": 4.997768916576057e-06, + "loss": 0.7481, + "step": 204 + }, + { + "epoch": 0.07849894696534558, + "grad_norm": 0.1837858110666275, + "learning_rate": 4.99772581138685e-06, + "loss": 0.7266, + "step": 205 + }, + { + "epoch": 0.07888186865785947, + "grad_norm": 0.1804484724998474, + "learning_rate": 4.9976822939592085e-06, + "loss": 0.7187, + "step": 206 + }, + { + "epoch": 0.07926479035037334, + "grad_norm": 0.1876893788576126, + "learning_rate": 4.997638364300314e-06, + "loss": 0.7199, + "step": 207 + }, + { + "epoch": 0.07964771204288723, + "grad_norm": 0.1820950210094452, + "learning_rate": 4.997594022417417e-06, + "loss": 0.7137, + "step": 208 + }, + { + "epoch": 0.0800306337354011, + "grad_norm": 0.18648962676525116, + "learning_rate": 4.997549268317836e-06, + "loss": 0.7309, + "step": 209 + }, + { + "epoch": 0.08041355542791499, + "grad_norm": 0.200729101896286, + "learning_rate": 4.997504102008957e-06, + "loss": 0.7374, + "step": 210 + }, + { + "epoch": 0.08079647712042887, + "grad_norm": 0.18724213540554047, + "learning_rate": 4.997458523498236e-06, + "loss": 0.7415, + "step": 211 + }, + { + "epoch": 0.08117939881294275, + "grad_norm": 0.1862158477306366, + "learning_rate": 4.997412532793195e-06, + "loss": 0.7286, + "step": 212 + }, + { + "epoch": 0.08156232050545663, + "grad_norm": 0.206449493765831, + "learning_rate": 4.997366129901422e-06, + "loss": 0.7608, + "step": 213 + }, + { + "epoch": 0.08194524219797052, + "grad_norm": 0.18451204895973206, + "learning_rate": 4.9973193148305784e-06, + "loss": 0.7334, + "step": 214 + }, + { + "epoch": 0.08232816389048439, + "grad_norm": 0.1887936145067215, + "learning_rate": 4.997272087588391e-06, + "loss": 0.7387, + "step": 215 + }, + { + "epoch": 0.08271108558299828, + "grad_norm": 0.1905495524406433, + "learning_rate": 4.9972244481826524e-06, + "loss": 0.7284, + "step": 216 + }, + { + "epoch": 0.08309400727551215, + "grad_norm": 0.18892931938171387, + "learning_rate": 4.997176396621227e-06, + "loss": 0.7268, + "step": 217 + }, + { + "epoch": 0.08347692896802604, + "grad_norm": 0.1919863075017929, + "learning_rate": 4.9971279329120445e-06, + "loss": 0.7433, + "step": 218 + }, + { + "epoch": 0.08385985066053991, + "grad_norm": 0.192040354013443, + "learning_rate": 4.9970790570631045e-06, + "loss": 0.7529, + "step": 219 + }, + { + "epoch": 0.0842427723530538, + "grad_norm": 0.19068577885627747, + "learning_rate": 4.9970297690824735e-06, + "loss": 0.74, + "step": 220 + }, + { + "epoch": 0.08462569404556768, + "grad_norm": 0.19420921802520752, + "learning_rate": 4.996980068978285e-06, + "loss": 0.7388, + "step": 221 + }, + { + "epoch": 0.08500861573808156, + "grad_norm": 0.19175250828266144, + "learning_rate": 4.9969299567587435e-06, + "loss": 0.7227, + "step": 222 + }, + { + "epoch": 0.08539153743059544, + "grad_norm": 0.19421349465847015, + "learning_rate": 4.996879432432118e-06, + "loss": 0.7348, + "step": 223 + }, + { + "epoch": 0.08577445912310933, + "grad_norm": 0.18917426466941833, + "learning_rate": 4.996828496006749e-06, + "loss": 0.7273, + "step": 224 + }, + { + "epoch": 0.0861573808156232, + "grad_norm": 0.1845076084136963, + "learning_rate": 4.996777147491043e-06, + "loss": 0.7105, + "step": 225 + }, + { + "epoch": 0.08654030250813709, + "grad_norm": 0.18641623854637146, + "learning_rate": 4.996725386893475e-06, + "loss": 0.7244, + "step": 226 + }, + { + "epoch": 0.08692322420065096, + "grad_norm": 0.19008800387382507, + "learning_rate": 4.996673214222586e-06, + "loss": 0.7281, + "step": 227 + }, + { + "epoch": 0.08730614589316485, + "grad_norm": 0.19295676052570343, + "learning_rate": 4.99662062948699e-06, + "loss": 0.7342, + "step": 228 + }, + { + "epoch": 0.08768906758567872, + "grad_norm": 0.1977388709783554, + "learning_rate": 4.9965676326953624e-06, + "loss": 0.7277, + "step": 229 + }, + { + "epoch": 0.08807198927819261, + "grad_norm": 0.18937358260154724, + "learning_rate": 4.996514223856453e-06, + "loss": 0.7153, + "step": 230 + }, + { + "epoch": 0.08845491097070649, + "grad_norm": 0.19146166741847992, + "learning_rate": 4.996460402979076e-06, + "loss": 0.7255, + "step": 231 + }, + { + "epoch": 0.08883783266322037, + "grad_norm": 0.20314456522464752, + "learning_rate": 4.9964061700721124e-06, + "loss": 0.7245, + "step": 232 + }, + { + "epoch": 0.08922075435573425, + "grad_norm": 0.19764558970928192, + "learning_rate": 4.996351525144515e-06, + "loss": 0.7496, + "step": 233 + }, + { + "epoch": 0.08960367604824813, + "grad_norm": 0.19225457310676575, + "learning_rate": 4.9962964682053015e-06, + "loss": 0.7286, + "step": 234 + }, + { + "epoch": 0.08998659774076201, + "grad_norm": 0.19454413652420044, + "learning_rate": 4.996240999263561e-06, + "loss": 0.7274, + "step": 235 + }, + { + "epoch": 0.0903695194332759, + "grad_norm": 0.19086669385433197, + "learning_rate": 4.996185118328446e-06, + "loss": 0.7282, + "step": 236 + }, + { + "epoch": 0.09075244112578977, + "grad_norm": 0.19874230027198792, + "learning_rate": 4.99612882540918e-06, + "loss": 0.7324, + "step": 237 + }, + { + "epoch": 0.09113536281830366, + "grad_norm": 0.19608773291110992, + "learning_rate": 4.9960721205150545e-06, + "loss": 0.7302, + "step": 238 + }, + { + "epoch": 0.09151828451081753, + "grad_norm": 0.19033324718475342, + "learning_rate": 4.996015003655429e-06, + "loss": 0.7208, + "step": 239 + }, + { + "epoch": 0.09190120620333142, + "grad_norm": 0.19876620173454285, + "learning_rate": 4.995957474839728e-06, + "loss": 0.7351, + "step": 240 + }, + { + "epoch": 0.0922841278958453, + "grad_norm": 0.19782209396362305, + "learning_rate": 4.9958995340774486e-06, + "loss": 0.7312, + "step": 241 + }, + { + "epoch": 0.09266704958835918, + "grad_norm": 0.19374790787696838, + "learning_rate": 4.995841181378153e-06, + "loss": 0.7439, + "step": 242 + }, + { + "epoch": 0.09304997128087306, + "grad_norm": 0.19944606721401215, + "learning_rate": 4.995782416751472e-06, + "loss": 0.7373, + "step": 243 + }, + { + "epoch": 0.09343289297338694, + "grad_norm": 0.19650287926197052, + "learning_rate": 4.9957232402071045e-06, + "loss": 0.71, + "step": 244 + }, + { + "epoch": 0.09381581466590082, + "grad_norm": 0.19653984904289246, + "learning_rate": 4.995663651754816e-06, + "loss": 0.7128, + "step": 245 + }, + { + "epoch": 0.0941987363584147, + "grad_norm": 0.19898127019405365, + "learning_rate": 4.995603651404444e-06, + "loss": 0.7288, + "step": 246 + }, + { + "epoch": 0.09458165805092858, + "grad_norm": 0.19377675652503967, + "learning_rate": 4.9955432391658886e-06, + "loss": 0.7256, + "step": 247 + }, + { + "epoch": 0.09496457974344247, + "grad_norm": 0.2016996592283249, + "learning_rate": 4.995482415049123e-06, + "loss": 0.7465, + "step": 248 + }, + { + "epoch": 0.09534750143595634, + "grad_norm": 0.19421467185020447, + "learning_rate": 4.995421179064183e-06, + "loss": 0.7215, + "step": 249 + }, + { + "epoch": 0.09573042312847023, + "grad_norm": 0.19672785699367523, + "learning_rate": 4.995359531221179e-06, + "loss": 0.7218, + "step": 250 + }, + { + "epoch": 0.0961133448209841, + "grad_norm": 0.20065009593963623, + "learning_rate": 4.995297471530283e-06, + "loss": 0.7325, + "step": 251 + }, + { + "epoch": 0.09649626651349799, + "grad_norm": 0.1908031553030014, + "learning_rate": 4.995235000001739e-06, + "loss": 0.7283, + "step": 252 + }, + { + "epoch": 0.09687918820601187, + "grad_norm": 0.19145245850086212, + "learning_rate": 4.995172116645856e-06, + "loss": 0.7255, + "step": 253 + }, + { + "epoch": 0.09726210989852575, + "grad_norm": 0.1942753791809082, + "learning_rate": 4.995108821473014e-06, + "loss": 0.7038, + "step": 254 + }, + { + "epoch": 0.09764503159103963, + "grad_norm": 0.19058293104171753, + "learning_rate": 4.99504511449366e-06, + "loss": 0.7134, + "step": 255 + }, + { + "epoch": 0.09802795328355352, + "grad_norm": 0.19463270902633667, + "learning_rate": 4.994980995718306e-06, + "loss": 0.7274, + "step": 256 + }, + { + "epoch": 0.09841087497606739, + "grad_norm": 0.19273309409618378, + "learning_rate": 4.994916465157538e-06, + "loss": 0.7141, + "step": 257 + }, + { + "epoch": 0.09879379666858128, + "grad_norm": 0.18812306225299835, + "learning_rate": 4.994851522822004e-06, + "loss": 0.7033, + "step": 258 + }, + { + "epoch": 0.09917671836109515, + "grad_norm": 0.1985722929239273, + "learning_rate": 4.994786168722424e-06, + "loss": 0.7086, + "step": 259 + }, + { + "epoch": 0.09955964005360904, + "grad_norm": 0.2013116478919983, + "learning_rate": 4.994720402869583e-06, + "loss": 0.7115, + "step": 260 + }, + { + "epoch": 0.09994256174612291, + "grad_norm": 0.20156681537628174, + "learning_rate": 4.994654225274337e-06, + "loss": 0.7357, + "step": 261 + }, + { + "epoch": 0.1003254834386368, + "grad_norm": 0.20074698328971863, + "learning_rate": 4.994587635947607e-06, + "loss": 0.7185, + "step": 262 + }, + { + "epoch": 0.10070840513115067, + "grad_norm": 0.20263032615184784, + "learning_rate": 4.994520634900383e-06, + "loss": 0.7151, + "step": 263 + }, + { + "epoch": 0.10109132682366456, + "grad_norm": 0.2002752721309662, + "learning_rate": 4.994453222143724e-06, + "loss": 0.7333, + "step": 264 + }, + { + "epoch": 0.10147424851617844, + "grad_norm": 0.19758231937885284, + "learning_rate": 4.994385397688756e-06, + "loss": 0.7337, + "step": 265 + }, + { + "epoch": 0.10185717020869232, + "grad_norm": 0.19643495976924896, + "learning_rate": 4.994317161546672e-06, + "loss": 0.756, + "step": 266 + }, + { + "epoch": 0.1022400919012062, + "grad_norm": 0.19787892699241638, + "learning_rate": 4.994248513728737e-06, + "loss": 0.7251, + "step": 267 + }, + { + "epoch": 0.10262301359372009, + "grad_norm": 0.2025986909866333, + "learning_rate": 4.994179454246279e-06, + "loss": 0.7428, + "step": 268 + }, + { + "epoch": 0.10300593528623396, + "grad_norm": 0.1986268311738968, + "learning_rate": 4.994109983110695e-06, + "loss": 0.7141, + "step": 269 + }, + { + "epoch": 0.10338885697874785, + "grad_norm": 0.1972547024488449, + "learning_rate": 4.9940401003334525e-06, + "loss": 0.717, + "step": 270 + }, + { + "epoch": 0.10377177867126172, + "grad_norm": 0.20430248975753784, + "learning_rate": 4.993969805926085e-06, + "loss": 0.7417, + "step": 271 + }, + { + "epoch": 0.10415470036377561, + "grad_norm": 0.198611319065094, + "learning_rate": 4.993899099900194e-06, + "loss": 0.7337, + "step": 272 + }, + { + "epoch": 0.10453762205628948, + "grad_norm": 0.2080761045217514, + "learning_rate": 4.993827982267451e-06, + "loss": 0.6972, + "step": 273 + }, + { + "epoch": 0.10492054374880337, + "grad_norm": 0.2025095522403717, + "learning_rate": 4.99375645303959e-06, + "loss": 0.7337, + "step": 274 + }, + { + "epoch": 0.10530346544131725, + "grad_norm": 0.20767749845981598, + "learning_rate": 4.993684512228419e-06, + "loss": 0.708, + "step": 275 + }, + { + "epoch": 0.10568638713383113, + "grad_norm": 0.20261400938034058, + "learning_rate": 4.993612159845812e-06, + "loss": 0.7095, + "step": 276 + }, + { + "epoch": 0.10606930882634501, + "grad_norm": 0.211074098944664, + "learning_rate": 4.993539395903709e-06, + "loss": 0.7349, + "step": 277 + }, + { + "epoch": 0.1064522305188589, + "grad_norm": 0.20446090400218964, + "learning_rate": 4.99346622041412e-06, + "loss": 0.7142, + "step": 278 + }, + { + "epoch": 0.10683515221137277, + "grad_norm": 0.1996392160654068, + "learning_rate": 4.993392633389123e-06, + "loss": 0.7298, + "step": 279 + }, + { + "epoch": 0.10721807390388666, + "grad_norm": 0.20801635086536407, + "learning_rate": 4.99331863484086e-06, + "loss": 0.7335, + "step": 280 + }, + { + "epoch": 0.10760099559640053, + "grad_norm": 0.20142178237438202, + "learning_rate": 4.993244224781549e-06, + "loss": 0.7336, + "step": 281 + }, + { + "epoch": 0.10798391728891442, + "grad_norm": 0.20546960830688477, + "learning_rate": 4.993169403223468e-06, + "loss": 0.7472, + "step": 282 + }, + { + "epoch": 0.10836683898142829, + "grad_norm": 0.20113351941108704, + "learning_rate": 4.993094170178966e-06, + "loss": 0.7222, + "step": 283 + }, + { + "epoch": 0.10874976067394218, + "grad_norm": 0.20306220650672913, + "learning_rate": 4.99301852566046e-06, + "loss": 0.7059, + "step": 284 + }, + { + "epoch": 0.10913268236645605, + "grad_norm": 0.2000206559896469, + "learning_rate": 4.992942469680437e-06, + "loss": 0.7042, + "step": 285 + }, + { + "epoch": 0.10951560405896994, + "grad_norm": 0.20378772914409637, + "learning_rate": 4.992866002251446e-06, + "loss": 0.7338, + "step": 286 + }, + { + "epoch": 0.10989852575148382, + "grad_norm": 0.19586919248104095, + "learning_rate": 4.99278912338611e-06, + "loss": 0.7108, + "step": 287 + }, + { + "epoch": 0.1102814474439977, + "grad_norm": 0.19886094331741333, + "learning_rate": 4.992711833097117e-06, + "loss": 0.7137, + "step": 288 + }, + { + "epoch": 0.11066436913651158, + "grad_norm": 0.20048004388809204, + "learning_rate": 4.992634131397223e-06, + "loss": 0.7281, + "step": 289 + }, + { + "epoch": 0.11104729082902547, + "grad_norm": 0.2020483762025833, + "learning_rate": 4.992556018299255e-06, + "loss": 0.7294, + "step": 290 + }, + { + "epoch": 0.11143021252153934, + "grad_norm": 0.1963844746351242, + "learning_rate": 4.9924774938161015e-06, + "loss": 0.7267, + "step": 291 + }, + { + "epoch": 0.11181313421405323, + "grad_norm": 0.1980431079864502, + "learning_rate": 4.992398557960724e-06, + "loss": 0.7172, + "step": 292 + }, + { + "epoch": 0.1121960559065671, + "grad_norm": 0.19689813256263733, + "learning_rate": 4.99231921074615e-06, + "loss": 0.6962, + "step": 293 + }, + { + "epoch": 0.11257897759908099, + "grad_norm": 0.20420525968074799, + "learning_rate": 4.992239452185476e-06, + "loss": 0.7157, + "step": 294 + }, + { + "epoch": 0.11296189929159486, + "grad_norm": 0.2101951539516449, + "learning_rate": 4.992159282291868e-06, + "loss": 0.7417, + "step": 295 + }, + { + "epoch": 0.11334482098410875, + "grad_norm": 0.2151687741279602, + "learning_rate": 4.9920787010785536e-06, + "loss": 0.7279, + "step": 296 + }, + { + "epoch": 0.11372774267662263, + "grad_norm": 0.20385341346263885, + "learning_rate": 4.991997708558835e-06, + "loss": 0.7436, + "step": 297 + }, + { + "epoch": 0.11411066436913651, + "grad_norm": 0.21158473193645477, + "learning_rate": 4.991916304746077e-06, + "loss": 0.7322, + "step": 298 + }, + { + "epoch": 0.11449358606165039, + "grad_norm": 0.20609347522258759, + "learning_rate": 4.991834489653718e-06, + "loss": 0.734, + "step": 299 + }, + { + "epoch": 0.11487650775416428, + "grad_norm": 0.2112123966217041, + "learning_rate": 4.991752263295261e-06, + "loss": 0.7063, + "step": 300 + }, + { + "epoch": 0.11525942944667815, + "grad_norm": 0.2099299132823944, + "learning_rate": 4.991669625684276e-06, + "loss": 0.7216, + "step": 301 + }, + { + "epoch": 0.11564235113919204, + "grad_norm": 0.20389284193515778, + "learning_rate": 4.9915865768344005e-06, + "loss": 0.7186, + "step": 302 + }, + { + "epoch": 0.11602527283170591, + "grad_norm": 0.2019105851650238, + "learning_rate": 4.991503116759345e-06, + "loss": 0.703, + "step": 303 + }, + { + "epoch": 0.1164081945242198, + "grad_norm": 0.20238110423088074, + "learning_rate": 4.99141924547288e-06, + "loss": 0.7024, + "step": 304 + }, + { + "epoch": 0.11679111621673367, + "grad_norm": 0.2012929469347, + "learning_rate": 4.991334962988852e-06, + "loss": 0.7171, + "step": 305 + }, + { + "epoch": 0.11717403790924756, + "grad_norm": 0.21306589245796204, + "learning_rate": 4.991250269321169e-06, + "loss": 0.714, + "step": 306 + }, + { + "epoch": 0.11755695960176143, + "grad_norm": 0.2003168761730194, + "learning_rate": 4.99116516448381e-06, + "loss": 0.7159, + "step": 307 + }, + { + "epoch": 0.11793988129427532, + "grad_norm": 0.20672893524169922, + "learning_rate": 4.991079648490821e-06, + "loss": 0.7351, + "step": 308 + }, + { + "epoch": 0.1183228029867892, + "grad_norm": 0.2076590359210968, + "learning_rate": 4.990993721356317e-06, + "loss": 0.7316, + "step": 309 + }, + { + "epoch": 0.11870572467930308, + "grad_norm": 0.2050575613975525, + "learning_rate": 4.990907383094478e-06, + "loss": 0.7133, + "step": 310 + }, + { + "epoch": 0.11908864637181696, + "grad_norm": 0.20803672075271606, + "learning_rate": 4.990820633719556e-06, + "loss": 0.7173, + "step": 311 + }, + { + "epoch": 0.11947156806433085, + "grad_norm": 0.20378495752811432, + "learning_rate": 4.990733473245866e-06, + "loss": 0.721, + "step": 312 + }, + { + "epoch": 0.11985448975684472, + "grad_norm": 0.20986348390579224, + "learning_rate": 4.9906459016877955e-06, + "loss": 0.7366, + "step": 313 + }, + { + "epoch": 0.12023741144935861, + "grad_norm": 0.202503502368927, + "learning_rate": 4.990557919059798e-06, + "loss": 0.7066, + "step": 314 + }, + { + "epoch": 0.12062033314187248, + "grad_norm": 0.21007058024406433, + "learning_rate": 4.990469525376392e-06, + "loss": 0.7503, + "step": 315 + }, + { + "epoch": 0.12100325483438637, + "grad_norm": 0.21536998450756073, + "learning_rate": 4.99038072065217e-06, + "loss": 0.7204, + "step": 316 + }, + { + "epoch": 0.12138617652690024, + "grad_norm": 0.20570923388004303, + "learning_rate": 4.990291504901785e-06, + "loss": 0.7302, + "step": 317 + }, + { + "epoch": 0.12176909821941413, + "grad_norm": 0.20915095508098602, + "learning_rate": 4.990201878139966e-06, + "loss": 0.7294, + "step": 318 + }, + { + "epoch": 0.122152019911928, + "grad_norm": 0.205043226480484, + "learning_rate": 4.9901118403815016e-06, + "loss": 0.7169, + "step": 319 + }, + { + "epoch": 0.1225349416044419, + "grad_norm": 0.20671123266220093, + "learning_rate": 4.990021391641255e-06, + "loss": 0.7065, + "step": 320 + }, + { + "epoch": 0.12291786329695577, + "grad_norm": 0.21355882287025452, + "learning_rate": 4.989930531934152e-06, + "loss": 0.7336, + "step": 321 + }, + { + "epoch": 0.12330078498946966, + "grad_norm": 0.21913346648216248, + "learning_rate": 4.989839261275191e-06, + "loss": 0.7168, + "step": 322 + }, + { + "epoch": 0.12368370668198353, + "grad_norm": 0.20357254147529602, + "learning_rate": 4.989747579679434e-06, + "loss": 0.6956, + "step": 323 + }, + { + "epoch": 0.12406662837449742, + "grad_norm": 0.20099161565303802, + "learning_rate": 4.989655487162013e-06, + "loss": 0.7097, + "step": 324 + }, + { + "epoch": 0.12444955006701129, + "grad_norm": 0.20629535615444183, + "learning_rate": 4.989562983738128e-06, + "loss": 0.7182, + "step": 325 + }, + { + "epoch": 0.12483247175952518, + "grad_norm": 0.20825421810150146, + "learning_rate": 4.989470069423045e-06, + "loss": 0.7101, + "step": 326 + }, + { + "epoch": 0.12521539345203905, + "grad_norm": 0.2155250906944275, + "learning_rate": 4.989376744232101e-06, + "loss": 0.714, + "step": 327 + }, + { + "epoch": 0.12559831514455294, + "grad_norm": 0.2070736289024353, + "learning_rate": 4.989283008180697e-06, + "loss": 0.7107, + "step": 328 + }, + { + "epoch": 0.12598123683706683, + "grad_norm": 0.20970770716667175, + "learning_rate": 4.989188861284305e-06, + "loss": 0.7054, + "step": 329 + }, + { + "epoch": 0.1263641585295807, + "grad_norm": 0.2124643325805664, + "learning_rate": 4.989094303558463e-06, + "loss": 0.7188, + "step": 330 + }, + { + "epoch": 0.12674708022209458, + "grad_norm": 0.2089206874370575, + "learning_rate": 4.988999335018777e-06, + "loss": 0.704, + "step": 331 + }, + { + "epoch": 0.12713000191460846, + "grad_norm": 0.2029915302991867, + "learning_rate": 4.988903955680923e-06, + "loss": 0.7125, + "step": 332 + }, + { + "epoch": 0.12751292360712235, + "grad_norm": 0.20648616552352905, + "learning_rate": 4.98880816556064e-06, + "loss": 0.7043, + "step": 333 + }, + { + "epoch": 0.1278958452996362, + "grad_norm": 0.20721794664859772, + "learning_rate": 4.98871196467374e-06, + "loss": 0.714, + "step": 334 + }, + { + "epoch": 0.1282787669921501, + "grad_norm": 0.2074366956949234, + "learning_rate": 4.988615353036101e-06, + "loss": 0.7005, + "step": 335 + }, + { + "epoch": 0.128661688684664, + "grad_norm": 0.2147573083639145, + "learning_rate": 4.988518330663665e-06, + "loss": 0.7176, + "step": 336 + }, + { + "epoch": 0.12904461037717788, + "grad_norm": 0.22254632413387299, + "learning_rate": 4.988420897572448e-06, + "loss": 0.7276, + "step": 337 + }, + { + "epoch": 0.12942753206969174, + "grad_norm": 0.21455398201942444, + "learning_rate": 4.98832305377853e-06, + "loss": 0.6962, + "step": 338 + }, + { + "epoch": 0.12981045376220562, + "grad_norm": 0.214791938662529, + "learning_rate": 4.98822479929806e-06, + "loss": 0.7168, + "step": 339 + }, + { + "epoch": 0.1301933754547195, + "grad_norm": 0.2137991189956665, + "learning_rate": 4.988126134147254e-06, + "loss": 0.7236, + "step": 340 + }, + { + "epoch": 0.1305762971472334, + "grad_norm": 0.21773158013820648, + "learning_rate": 4.988027058342396e-06, + "loss": 0.7225, + "step": 341 + }, + { + "epoch": 0.13095921883974726, + "grad_norm": 0.21584519743919373, + "learning_rate": 4.98792757189984e-06, + "loss": 0.7368, + "step": 342 + }, + { + "epoch": 0.13134214053226115, + "grad_norm": 0.21187016367912292, + "learning_rate": 4.987827674836003e-06, + "loss": 0.7163, + "step": 343 + }, + { + "epoch": 0.13172506222477504, + "grad_norm": 0.21678388118743896, + "learning_rate": 4.987727367167374e-06, + "loss": 0.721, + "step": 344 + }, + { + "epoch": 0.13210798391728892, + "grad_norm": 0.2196904569864273, + "learning_rate": 4.987626648910508e-06, + "loss": 0.7357, + "step": 345 + }, + { + "epoch": 0.13249090560980278, + "grad_norm": 0.22697031497955322, + "learning_rate": 4.9875255200820296e-06, + "loss": 0.7419, + "step": 346 + }, + { + "epoch": 0.13287382730231667, + "grad_norm": 0.21749354898929596, + "learning_rate": 4.987423980698627e-06, + "loss": 0.724, + "step": 347 + }, + { + "epoch": 0.13325674899483056, + "grad_norm": 0.20972779393196106, + "learning_rate": 4.9873220307770606e-06, + "loss": 0.7208, + "step": 348 + }, + { + "epoch": 0.13363967068734445, + "grad_norm": 0.2108878791332245, + "learning_rate": 4.987219670334156e-06, + "loss": 0.705, + "step": 349 + }, + { + "epoch": 0.1340225923798583, + "grad_norm": 0.21607382595539093, + "learning_rate": 4.987116899386808e-06, + "loss": 0.7033, + "step": 350 + }, + { + "epoch": 0.1344055140723722, + "grad_norm": 0.22202174365520477, + "learning_rate": 4.987013717951978e-06, + "loss": 0.7083, + "step": 351 + }, + { + "epoch": 0.13478843576488608, + "grad_norm": 0.21639248728752136, + "learning_rate": 4.986910126046695e-06, + "loss": 0.7111, + "step": 352 + }, + { + "epoch": 0.13517135745739997, + "grad_norm": 0.2141568511724472, + "learning_rate": 4.986806123688059e-06, + "loss": 0.7111, + "step": 353 + }, + { + "epoch": 0.13555427914991383, + "grad_norm": 0.21743625402450562, + "learning_rate": 4.986701710893232e-06, + "loss": 0.7262, + "step": 354 + }, + { + "epoch": 0.13593720084242772, + "grad_norm": 0.21381889283657074, + "learning_rate": 4.9865968876794476e-06, + "loss": 0.7068, + "step": 355 + }, + { + "epoch": 0.1363201225349416, + "grad_norm": 0.2151729315519333, + "learning_rate": 4.986491654064009e-06, + "loss": 0.7174, + "step": 356 + }, + { + "epoch": 0.1367030442274555, + "grad_norm": 0.21748758852481842, + "learning_rate": 4.986386010064281e-06, + "loss": 0.7201, + "step": 357 + }, + { + "epoch": 0.13708596591996935, + "grad_norm": 0.21864524483680725, + "learning_rate": 4.986279955697701e-06, + "loss": 0.7119, + "step": 358 + }, + { + "epoch": 0.13746888761248324, + "grad_norm": 0.21429306268692017, + "learning_rate": 4.986173490981773e-06, + "loss": 0.7173, + "step": 359 + }, + { + "epoch": 0.13785180930499713, + "grad_norm": 0.21339617669582367, + "learning_rate": 4.986066615934069e-06, + "loss": 0.7323, + "step": 360 + }, + { + "epoch": 0.13823473099751102, + "grad_norm": 0.21178525686264038, + "learning_rate": 4.985959330572227e-06, + "loss": 0.7143, + "step": 361 + }, + { + "epoch": 0.13861765269002488, + "grad_norm": 0.21623355150222778, + "learning_rate": 4.985851634913955e-06, + "loss": 0.6892, + "step": 362 + }, + { + "epoch": 0.13900057438253877, + "grad_norm": 0.21810880303382874, + "learning_rate": 4.985743528977029e-06, + "loss": 0.7135, + "step": 363 + }, + { + "epoch": 0.13938349607505265, + "grad_norm": 0.2134767770767212, + "learning_rate": 4.985635012779288e-06, + "loss": 0.7034, + "step": 364 + }, + { + "epoch": 0.13976641776756654, + "grad_norm": 0.21651697158813477, + "learning_rate": 4.985526086338644e-06, + "loss": 0.7017, + "step": 365 + }, + { + "epoch": 0.1401493394600804, + "grad_norm": 0.21995492279529572, + "learning_rate": 4.985416749673075e-06, + "loss": 0.7278, + "step": 366 + }, + { + "epoch": 0.1405322611525943, + "grad_norm": 0.2165203094482422, + "learning_rate": 4.985307002800626e-06, + "loss": 0.7089, + "step": 367 + }, + { + "epoch": 0.14091518284510818, + "grad_norm": 0.21220526099205017, + "learning_rate": 4.985196845739409e-06, + "loss": 0.7092, + "step": 368 + }, + { + "epoch": 0.14129810453762207, + "grad_norm": 0.22091609239578247, + "learning_rate": 4.9850862785076085e-06, + "loss": 0.7389, + "step": 369 + }, + { + "epoch": 0.14168102623013593, + "grad_norm": 0.21460163593292236, + "learning_rate": 4.98497530112347e-06, + "loss": 0.6937, + "step": 370 + }, + { + "epoch": 0.1420639479226498, + "grad_norm": 0.21762219071388245, + "learning_rate": 4.984863913605312e-06, + "loss": 0.7161, + "step": 371 + }, + { + "epoch": 0.1424468696151637, + "grad_norm": 0.21648332476615906, + "learning_rate": 4.9847521159715165e-06, + "loss": 0.7175, + "step": 372 + }, + { + "epoch": 0.1428297913076776, + "grad_norm": 0.21821698546409607, + "learning_rate": 4.984639908240536e-06, + "loss": 0.7006, + "step": 373 + }, + { + "epoch": 0.14321271300019145, + "grad_norm": 0.22117845714092255, + "learning_rate": 4.984527290430889e-06, + "loss": 0.7221, + "step": 374 + }, + { + "epoch": 0.14359563469270534, + "grad_norm": 0.22102953493595123, + "learning_rate": 4.984414262561166e-06, + "loss": 0.7028, + "step": 375 + }, + { + "epoch": 0.14397855638521923, + "grad_norm": 0.21901941299438477, + "learning_rate": 4.984300824650017e-06, + "loss": 0.7126, + "step": 376 + }, + { + "epoch": 0.1443614780777331, + "grad_norm": 0.22398801147937775, + "learning_rate": 4.984186976716167e-06, + "loss": 0.7104, + "step": 377 + }, + { + "epoch": 0.14474439977024697, + "grad_norm": 0.21393615007400513, + "learning_rate": 4.984072718778406e-06, + "loss": 0.7071, + "step": 378 + }, + { + "epoch": 0.14512732146276086, + "grad_norm": 0.219665989279747, + "learning_rate": 4.9839580508555915e-06, + "loss": 0.7232, + "step": 379 + }, + { + "epoch": 0.14551024315527475, + "grad_norm": 0.21342062950134277, + "learning_rate": 4.983842972966649e-06, + "loss": 0.6934, + "step": 380 + }, + { + "epoch": 0.14589316484778864, + "grad_norm": 0.21543823182582855, + "learning_rate": 4.983727485130572e-06, + "loss": 0.7114, + "step": 381 + }, + { + "epoch": 0.1462760865403025, + "grad_norm": 0.2211451679468155, + "learning_rate": 4.9836115873664195e-06, + "loss": 0.6972, + "step": 382 + }, + { + "epoch": 0.14665900823281638, + "grad_norm": 0.21137243509292603, + "learning_rate": 4.983495279693322e-06, + "loss": 0.7089, + "step": 383 + }, + { + "epoch": 0.14704192992533027, + "grad_norm": 0.22880102694034576, + "learning_rate": 4.983378562130474e-06, + "loss": 0.7372, + "step": 384 + }, + { + "epoch": 0.14742485161784416, + "grad_norm": 0.2202749252319336, + "learning_rate": 4.983261434697141e-06, + "loss": 0.6957, + "step": 385 + }, + { + "epoch": 0.14780777331035802, + "grad_norm": 0.22220788896083832, + "learning_rate": 4.9831438974126525e-06, + "loss": 0.7208, + "step": 386 + }, + { + "epoch": 0.1481906950028719, + "grad_norm": 0.2227896898984909, + "learning_rate": 4.983025950296408e-06, + "loss": 0.7037, + "step": 387 + }, + { + "epoch": 0.1485736166953858, + "grad_norm": 0.2159932553768158, + "learning_rate": 4.982907593367876e-06, + "loss": 0.7111, + "step": 388 + }, + { + "epoch": 0.14895653838789968, + "grad_norm": 0.220118910074234, + "learning_rate": 4.9827888266465876e-06, + "loss": 0.7127, + "step": 389 + }, + { + "epoch": 0.14933946008041354, + "grad_norm": 0.22082901000976562, + "learning_rate": 4.982669650152148e-06, + "loss": 0.6958, + "step": 390 + }, + { + "epoch": 0.14972238177292743, + "grad_norm": 0.22335520386695862, + "learning_rate": 4.982550063904223e-06, + "loss": 0.7102, + "step": 391 + }, + { + "epoch": 0.15010530346544132, + "grad_norm": 0.22340475022792816, + "learning_rate": 4.982430067922553e-06, + "loss": 0.7178, + "step": 392 + }, + { + "epoch": 0.1504882251579552, + "grad_norm": 0.22858190536499023, + "learning_rate": 4.9823096622269415e-06, + "loss": 0.7196, + "step": 393 + }, + { + "epoch": 0.15087114685046907, + "grad_norm": 0.21896734833717346, + "learning_rate": 4.9821888468372615e-06, + "loss": 0.6883, + "step": 394 + }, + { + "epoch": 0.15125406854298296, + "grad_norm": 0.23210509121418, + "learning_rate": 4.982067621773452e-06, + "loss": 0.7139, + "step": 395 + }, + { + "epoch": 0.15163699023549684, + "grad_norm": 0.22944042086601257, + "learning_rate": 4.981945987055521e-06, + "loss": 0.7194, + "step": 396 + }, + { + "epoch": 0.15201991192801073, + "grad_norm": 0.2184358388185501, + "learning_rate": 4.981823942703544e-06, + "loss": 0.7202, + "step": 397 + }, + { + "epoch": 0.1524028336205246, + "grad_norm": 0.22763623297214508, + "learning_rate": 4.9817014887376645e-06, + "loss": 0.7241, + "step": 398 + }, + { + "epoch": 0.15278575531303848, + "grad_norm": 0.22321753203868866, + "learning_rate": 4.981578625178093e-06, + "loss": 0.7001, + "step": 399 + }, + { + "epoch": 0.15316867700555237, + "grad_norm": 0.23028865456581116, + "learning_rate": 4.981455352045106e-06, + "loss": 0.7091, + "step": 400 + }, + { + "epoch": 0.15355159869806626, + "grad_norm": 0.22734788060188293, + "learning_rate": 4.981331669359049e-06, + "loss": 0.7291, + "step": 401 + }, + { + "epoch": 0.15393452039058012, + "grad_norm": 0.2209152728319168, + "learning_rate": 4.981207577140338e-06, + "loss": 0.722, + "step": 402 + }, + { + "epoch": 0.154317442083094, + "grad_norm": 0.216696634888649, + "learning_rate": 4.981083075409452e-06, + "loss": 0.7181, + "step": 403 + }, + { + "epoch": 0.1547003637756079, + "grad_norm": 0.2153381109237671, + "learning_rate": 4.98095816418694e-06, + "loss": 0.7228, + "step": 404 + }, + { + "epoch": 0.15508328546812178, + "grad_norm": 0.2197411209344864, + "learning_rate": 4.980832843493417e-06, + "loss": 0.7071, + "step": 405 + }, + { + "epoch": 0.15546620716063564, + "grad_norm": 0.21317681670188904, + "learning_rate": 4.980707113349567e-06, + "loss": 0.6979, + "step": 406 + }, + { + "epoch": 0.15584912885314953, + "grad_norm": 0.21619658172130585, + "learning_rate": 4.980580973776143e-06, + "loss": 0.7002, + "step": 407 + }, + { + "epoch": 0.15623205054566341, + "grad_norm": 0.21794861555099487, + "learning_rate": 4.980454424793961e-06, + "loss": 0.716, + "step": 408 + }, + { + "epoch": 0.1566149722381773, + "grad_norm": 0.2209479808807373, + "learning_rate": 4.980327466423909e-06, + "loss": 0.7106, + "step": 409 + }, + { + "epoch": 0.15699789393069116, + "grad_norm": 0.2195109724998474, + "learning_rate": 4.9802000986869415e-06, + "loss": 0.7041, + "step": 410 + }, + { + "epoch": 0.15738081562320505, + "grad_norm": 0.21919915080070496, + "learning_rate": 4.980072321604078e-06, + "loss": 0.7141, + "step": 411 + }, + { + "epoch": 0.15776373731571894, + "grad_norm": 0.21942178905010223, + "learning_rate": 4.979944135196409e-06, + "loss": 0.7016, + "step": 412 + }, + { + "epoch": 0.15814665900823283, + "grad_norm": 0.22026552259922028, + "learning_rate": 4.97981553948509e-06, + "loss": 0.7176, + "step": 413 + }, + { + "epoch": 0.1585295807007467, + "grad_norm": 0.22086086869239807, + "learning_rate": 4.979686534491347e-06, + "loss": 0.7022, + "step": 414 + }, + { + "epoch": 0.15891250239326057, + "grad_norm": 0.22359570860862732, + "learning_rate": 4.9795571202364705e-06, + "loss": 0.7176, + "step": 415 + }, + { + "epoch": 0.15929542408577446, + "grad_norm": 0.22810761630535126, + "learning_rate": 4.979427296741818e-06, + "loss": 0.7182, + "step": 416 + }, + { + "epoch": 0.15967834577828835, + "grad_norm": 0.2195035070180893, + "learning_rate": 4.97929706402882e-06, + "loss": 0.7047, + "step": 417 + }, + { + "epoch": 0.1600612674708022, + "grad_norm": 0.21832919120788574, + "learning_rate": 4.9791664221189675e-06, + "loss": 0.7105, + "step": 418 + }, + { + "epoch": 0.1604441891633161, + "grad_norm": 0.2244974821805954, + "learning_rate": 4.979035371033824e-06, + "loss": 0.7345, + "step": 419 + }, + { + "epoch": 0.16082711085582999, + "grad_norm": 0.2344287633895874, + "learning_rate": 4.97890391079502e-06, + "loss": 0.7368, + "step": 420 + }, + { + "epoch": 0.16121003254834387, + "grad_norm": 0.2236819863319397, + "learning_rate": 4.978772041424249e-06, + "loss": 0.7133, + "step": 421 + }, + { + "epoch": 0.16159295424085773, + "grad_norm": 0.22271811962127686, + "learning_rate": 4.978639762943279e-06, + "loss": 0.7214, + "step": 422 + }, + { + "epoch": 0.16197587593337162, + "grad_norm": 0.2247050553560257, + "learning_rate": 4.97850707537394e-06, + "loss": 0.6983, + "step": 423 + }, + { + "epoch": 0.1623587976258855, + "grad_norm": 0.22323517501354218, + "learning_rate": 4.978373978738132e-06, + "loss": 0.7194, + "step": 424 + }, + { + "epoch": 0.1627417193183994, + "grad_norm": 0.2268957793712616, + "learning_rate": 4.9782404730578225e-06, + "loss": 0.719, + "step": 425 + }, + { + "epoch": 0.16312464101091326, + "grad_norm": 0.22057485580444336, + "learning_rate": 4.978106558355044e-06, + "loss": 0.6976, + "step": 426 + }, + { + "epoch": 0.16350756270342715, + "grad_norm": 0.2155829668045044, + "learning_rate": 4.9779722346519014e-06, + "loss": 0.7051, + "step": 427 + }, + { + "epoch": 0.16389048439594103, + "grad_norm": 0.21671557426452637, + "learning_rate": 4.977837501970562e-06, + "loss": 0.69, + "step": 428 + }, + { + "epoch": 0.16427340608845492, + "grad_norm": 0.22859551012516022, + "learning_rate": 4.977702360333264e-06, + "loss": 0.6963, + "step": 429 + }, + { + "epoch": 0.16465632778096878, + "grad_norm": 0.2260478436946869, + "learning_rate": 4.977566809762312e-06, + "loss": 0.6866, + "step": 430 + }, + { + "epoch": 0.16503924947348267, + "grad_norm": 0.22191403806209564, + "learning_rate": 4.977430850280076e-06, + "loss": 0.7066, + "step": 431 + }, + { + "epoch": 0.16542217116599656, + "grad_norm": 0.22449170053005219, + "learning_rate": 4.977294481908999e-06, + "loss": 0.7283, + "step": 432 + }, + { + "epoch": 0.16580509285851044, + "grad_norm": 0.2276216596364975, + "learning_rate": 4.977157704671585e-06, + "loss": 0.7271, + "step": 433 + }, + { + "epoch": 0.1661880145510243, + "grad_norm": 0.23174940049648285, + "learning_rate": 4.97702051859041e-06, + "loss": 0.7095, + "step": 434 + }, + { + "epoch": 0.1665709362435382, + "grad_norm": 0.23013226687908173, + "learning_rate": 4.976882923688114e-06, + "loss": 0.7124, + "step": 435 + }, + { + "epoch": 0.16695385793605208, + "grad_norm": 0.23638951778411865, + "learning_rate": 4.976744919987409e-06, + "loss": 0.7311, + "step": 436 + }, + { + "epoch": 0.16733677962856597, + "grad_norm": 0.2298298478126526, + "learning_rate": 4.976606507511072e-06, + "loss": 0.6905, + "step": 437 + }, + { + "epoch": 0.16771970132107983, + "grad_norm": 0.23514661192893982, + "learning_rate": 4.976467686281945e-06, + "loss": 0.7382, + "step": 438 + }, + { + "epoch": 0.16810262301359372, + "grad_norm": 0.2360498607158661, + "learning_rate": 4.976328456322941e-06, + "loss": 0.708, + "step": 439 + }, + { + "epoch": 0.1684855447061076, + "grad_norm": 0.23474329710006714, + "learning_rate": 4.976188817657039e-06, + "loss": 0.717, + "step": 440 + }, + { + "epoch": 0.1688684663986215, + "grad_norm": 0.21756932139396667, + "learning_rate": 4.976048770307286e-06, + "loss": 0.6842, + "step": 441 + }, + { + "epoch": 0.16925138809113535, + "grad_norm": 0.22361315786838531, + "learning_rate": 4.9759083142967965e-06, + "loss": 0.7273, + "step": 442 + }, + { + "epoch": 0.16963430978364924, + "grad_norm": 0.22651907801628113, + "learning_rate": 4.975767449648751e-06, + "loss": 0.7157, + "step": 443 + }, + { + "epoch": 0.17001723147616313, + "grad_norm": 0.22782133519649506, + "learning_rate": 4.9756261763864e-06, + "loss": 0.7122, + "step": 444 + }, + { + "epoch": 0.17040015316867702, + "grad_norm": 0.2352352738380432, + "learning_rate": 4.97548449453306e-06, + "loss": 0.7224, + "step": 445 + }, + { + "epoch": 0.17078307486119088, + "grad_norm": 0.22195972502231598, + "learning_rate": 4.975342404112114e-06, + "loss": 0.7092, + "step": 446 + }, + { + "epoch": 0.17116599655370476, + "grad_norm": 0.21784667670726776, + "learning_rate": 4.975199905147014e-06, + "loss": 0.6812, + "step": 447 + }, + { + "epoch": 0.17154891824621865, + "grad_norm": 0.21954038739204407, + "learning_rate": 4.9750569976612776e-06, + "loss": 0.6989, + "step": 448 + }, + { + "epoch": 0.17193183993873254, + "grad_norm": 0.2203446477651596, + "learning_rate": 4.9749136816784924e-06, + "loss": 0.7066, + "step": 449 + }, + { + "epoch": 0.1723147616312464, + "grad_norm": 0.22727733850479126, + "learning_rate": 4.974769957222312e-06, + "loss": 0.6922, + "step": 450 + }, + { + "epoch": 0.1726976833237603, + "grad_norm": 0.22717690467834473, + "learning_rate": 4.974625824316456e-06, + "loss": 0.6975, + "step": 451 + }, + { + "epoch": 0.17308060501627417, + "grad_norm": 0.2288667857646942, + "learning_rate": 4.974481282984716e-06, + "loss": 0.7036, + "step": 452 + }, + { + "epoch": 0.17346352670878806, + "grad_norm": 0.22407589852809906, + "learning_rate": 4.974336333250944e-06, + "loss": 0.6969, + "step": 453 + }, + { + "epoch": 0.17384644840130192, + "grad_norm": 0.2228279709815979, + "learning_rate": 4.9741909751390664e-06, + "loss": 0.7024, + "step": 454 + }, + { + "epoch": 0.1742293700938158, + "grad_norm": 0.2224264144897461, + "learning_rate": 4.9740452086730725e-06, + "loss": 0.6967, + "step": 455 + }, + { + "epoch": 0.1746122917863297, + "grad_norm": 0.22775305807590485, + "learning_rate": 4.97389903387702e-06, + "loss": 0.6995, + "step": 456 + }, + { + "epoch": 0.1749952134788436, + "grad_norm": 0.22632873058319092, + "learning_rate": 4.973752450775036e-06, + "loss": 0.6902, + "step": 457 + }, + { + "epoch": 0.17537813517135745, + "grad_norm": 0.22588998079299927, + "learning_rate": 4.973605459391312e-06, + "loss": 0.7133, + "step": 458 + }, + { + "epoch": 0.17576105686387133, + "grad_norm": 0.22500504553318024, + "learning_rate": 4.973458059750108e-06, + "loss": 0.712, + "step": 459 + }, + { + "epoch": 0.17614397855638522, + "grad_norm": 0.22925838828086853, + "learning_rate": 4.973310251875753e-06, + "loss": 0.696, + "step": 460 + }, + { + "epoch": 0.1765269002488991, + "grad_norm": 0.22367170453071594, + "learning_rate": 4.973162035792641e-06, + "loss": 0.681, + "step": 461 + }, + { + "epoch": 0.17690982194141297, + "grad_norm": 0.22455467283725739, + "learning_rate": 4.973013411525235e-06, + "loss": 0.7047, + "step": 462 + }, + { + "epoch": 0.17729274363392686, + "grad_norm": 0.23515203595161438, + "learning_rate": 4.972864379098064e-06, + "loss": 0.7032, + "step": 463 + }, + { + "epoch": 0.17767566532644075, + "grad_norm": 0.230363667011261, + "learning_rate": 4.972714938535727e-06, + "loss": 0.698, + "step": 464 + }, + { + "epoch": 0.17805858701895463, + "grad_norm": 0.2247180938720703, + "learning_rate": 4.972565089862886e-06, + "loss": 0.7066, + "step": 465 + }, + { + "epoch": 0.1784415087114685, + "grad_norm": 0.23525093495845795, + "learning_rate": 4.972414833104274e-06, + "loss": 0.7219, + "step": 466 + }, + { + "epoch": 0.17882443040398238, + "grad_norm": 0.22985999286174774, + "learning_rate": 4.9722641682846914e-06, + "loss": 0.7186, + "step": 467 + }, + { + "epoch": 0.17920735209649627, + "grad_norm": 0.22825634479522705, + "learning_rate": 4.972113095429003e-06, + "loss": 0.7191, + "step": 468 + }, + { + "epoch": 0.17959027378901016, + "grad_norm": 0.22180423140525818, + "learning_rate": 4.971961614562143e-06, + "loss": 0.6829, + "step": 469 + }, + { + "epoch": 0.17997319548152402, + "grad_norm": 0.22781048715114594, + "learning_rate": 4.971809725709112e-06, + "loss": 0.6898, + "step": 470 + }, + { + "epoch": 0.1803561171740379, + "grad_norm": 0.23069560527801514, + "learning_rate": 4.9716574288949804e-06, + "loss": 0.7253, + "step": 471 + }, + { + "epoch": 0.1807390388665518, + "grad_norm": 0.22806665301322937, + "learning_rate": 4.971504724144884e-06, + "loss": 0.6897, + "step": 472 + }, + { + "epoch": 0.18112196055906568, + "grad_norm": 0.23605108261108398, + "learning_rate": 4.971351611484023e-06, + "loss": 0.7297, + "step": 473 + }, + { + "epoch": 0.18150488225157954, + "grad_norm": 0.23285695910453796, + "learning_rate": 4.971198090937671e-06, + "loss": 0.7113, + "step": 474 + }, + { + "epoch": 0.18188780394409343, + "grad_norm": 0.23309025168418884, + "learning_rate": 4.971044162531165e-06, + "loss": 0.7037, + "step": 475 + }, + { + "epoch": 0.18227072563660732, + "grad_norm": 0.23858731985092163, + "learning_rate": 4.970889826289912e-06, + "loss": 0.7117, + "step": 476 + }, + { + "epoch": 0.1826536473291212, + "grad_norm": 0.23126953840255737, + "learning_rate": 4.9707350822393805e-06, + "loss": 0.7114, + "step": 477 + }, + { + "epoch": 0.18303656902163506, + "grad_norm": 0.23670174181461334, + "learning_rate": 4.970579930405114e-06, + "loss": 0.7077, + "step": 478 + }, + { + "epoch": 0.18341949071414895, + "grad_norm": 0.2424033135175705, + "learning_rate": 4.970424370812718e-06, + "loss": 0.6897, + "step": 479 + }, + { + "epoch": 0.18380241240666284, + "grad_norm": 0.2186141163110733, + "learning_rate": 4.970268403487866e-06, + "loss": 0.7031, + "step": 480 + }, + { + "epoch": 0.18418533409917673, + "grad_norm": 0.22593295574188232, + "learning_rate": 4.970112028456301e-06, + "loss": 0.6798, + "step": 481 + }, + { + "epoch": 0.1845682557916906, + "grad_norm": 0.23915334045886993, + "learning_rate": 4.969955245743832e-06, + "loss": 0.6964, + "step": 482 + }, + { + "epoch": 0.18495117748420448, + "grad_norm": 0.23232129216194153, + "learning_rate": 4.9697980553763356e-06, + "loss": 0.7196, + "step": 483 + }, + { + "epoch": 0.18533409917671836, + "grad_norm": 0.22518081963062286, + "learning_rate": 4.969640457379754e-06, + "loss": 0.6879, + "step": 484 + }, + { + "epoch": 0.18571702086923225, + "grad_norm": 0.22891272604465485, + "learning_rate": 4.969482451780099e-06, + "loss": 0.6841, + "step": 485 + }, + { + "epoch": 0.1860999425617461, + "grad_norm": 0.23179848492145538, + "learning_rate": 4.969324038603449e-06, + "loss": 0.7019, + "step": 486 + }, + { + "epoch": 0.18648286425426, + "grad_norm": 0.23586530983448029, + "learning_rate": 4.969165217875948e-06, + "loss": 0.6975, + "step": 487 + }, + { + "epoch": 0.1868657859467739, + "grad_norm": 0.23328320682048798, + "learning_rate": 4.969005989623812e-06, + "loss": 0.7116, + "step": 488 + }, + { + "epoch": 0.18724870763928778, + "grad_norm": 0.22832241654396057, + "learning_rate": 4.968846353873316e-06, + "loss": 0.7058, + "step": 489 + }, + { + "epoch": 0.18763162933180164, + "grad_norm": 0.23435759544372559, + "learning_rate": 4.968686310650811e-06, + "loss": 0.7225, + "step": 490 + }, + { + "epoch": 0.18801455102431552, + "grad_norm": 0.23468558490276337, + "learning_rate": 4.968525859982709e-06, + "loss": 0.6828, + "step": 491 + }, + { + "epoch": 0.1883974727168294, + "grad_norm": 0.2350325882434845, + "learning_rate": 4.9683650018954935e-06, + "loss": 0.6831, + "step": 492 + }, + { + "epoch": 0.1887803944093433, + "grad_norm": 0.22212567925453186, + "learning_rate": 4.968203736415713e-06, + "loss": 0.6908, + "step": 493 + }, + { + "epoch": 0.18916331610185716, + "grad_norm": 0.2288699448108673, + "learning_rate": 4.968042063569982e-06, + "loss": 0.6914, + "step": 494 + }, + { + "epoch": 0.18954623779437105, + "grad_norm": 0.24049511551856995, + "learning_rate": 4.967879983384987e-06, + "loss": 0.7103, + "step": 495 + }, + { + "epoch": 0.18992915948688494, + "grad_norm": 0.23875749111175537, + "learning_rate": 4.967717495887476e-06, + "loss": 0.6892, + "step": 496 + }, + { + "epoch": 0.19031208117939882, + "grad_norm": 0.230393648147583, + "learning_rate": 4.967554601104268e-06, + "loss": 0.6968, + "step": 497 + }, + { + "epoch": 0.19069500287191268, + "grad_norm": 0.22483550012111664, + "learning_rate": 4.967391299062247e-06, + "loss": 0.683, + "step": 498 + }, + { + "epoch": 0.19107792456442657, + "grad_norm": 0.2363535761833191, + "learning_rate": 4.967227589788365e-06, + "loss": 0.7108, + "step": 499 + }, + { + "epoch": 0.19146084625694046, + "grad_norm": 0.23204447329044342, + "learning_rate": 4.967063473309644e-06, + "loss": 0.6833, + "step": 500 + }, + { + "epoch": 0.19184376794945435, + "grad_norm": 0.23585835099220276, + "learning_rate": 4.9668989496531684e-06, + "loss": 0.6713, + "step": 501 + }, + { + "epoch": 0.1922266896419682, + "grad_norm": 0.24668626487255096, + "learning_rate": 4.966734018846093e-06, + "loss": 0.6959, + "step": 502 + }, + { + "epoch": 0.1926096113344821, + "grad_norm": 0.24632133543491364, + "learning_rate": 4.966568680915639e-06, + "loss": 0.7179, + "step": 503 + }, + { + "epoch": 0.19299253302699598, + "grad_norm": 0.2375093400478363, + "learning_rate": 4.966402935889094e-06, + "loss": 0.6956, + "step": 504 + }, + { + "epoch": 0.19337545471950987, + "grad_norm": 0.2281908541917801, + "learning_rate": 4.966236783793814e-06, + "loss": 0.7169, + "step": 505 + }, + { + "epoch": 0.19375837641202373, + "grad_norm": 0.23612891137599945, + "learning_rate": 4.9660702246572215e-06, + "loss": 0.6921, + "step": 506 + }, + { + "epoch": 0.19414129810453762, + "grad_norm": 0.2388295829296112, + "learning_rate": 4.965903258506806e-06, + "loss": 0.7058, + "step": 507 + }, + { + "epoch": 0.1945242197970515, + "grad_norm": 0.23923897743225098, + "learning_rate": 4.965735885370126e-06, + "loss": 0.6878, + "step": 508 + }, + { + "epoch": 0.1949071414895654, + "grad_norm": 0.24728970229625702, + "learning_rate": 4.965568105274803e-06, + "loss": 0.7118, + "step": 509 + }, + { + "epoch": 0.19529006318207925, + "grad_norm": 0.24075351655483246, + "learning_rate": 4.96539991824853e-06, + "loss": 0.7101, + "step": 510 + }, + { + "epoch": 0.19567298487459314, + "grad_norm": 0.23857763409614563, + "learning_rate": 4.965231324319067e-06, + "loss": 0.7122, + "step": 511 + }, + { + "epoch": 0.19605590656710703, + "grad_norm": 0.24270926415920258, + "learning_rate": 4.965062323514237e-06, + "loss": 0.6893, + "step": 512 + }, + { + "epoch": 0.19643882825962092, + "grad_norm": 0.23443444073200226, + "learning_rate": 4.964892915861934e-06, + "loss": 0.6913, + "step": 513 + }, + { + "epoch": 0.19682174995213478, + "grad_norm": 0.23783984780311584, + "learning_rate": 4.964723101390118e-06, + "loss": 0.7059, + "step": 514 + }, + { + "epoch": 0.19720467164464867, + "grad_norm": 0.23205767571926117, + "learning_rate": 4.964552880126816e-06, + "loss": 0.6983, + "step": 515 + }, + { + "epoch": 0.19758759333716255, + "grad_norm": 0.22749286890029907, + "learning_rate": 4.964382252100122e-06, + "loss": 0.7025, + "step": 516 + }, + { + "epoch": 0.19797051502967644, + "grad_norm": 0.23700077831745148, + "learning_rate": 4.9642112173381976e-06, + "loss": 0.6903, + "step": 517 + }, + { + "epoch": 0.1983534367221903, + "grad_norm": 0.2351311892271042, + "learning_rate": 4.964039775869271e-06, + "loss": 0.6945, + "step": 518 + }, + { + "epoch": 0.1987363584147042, + "grad_norm": 0.24296842515468597, + "learning_rate": 4.963867927721639e-06, + "loss": 0.7023, + "step": 519 + }, + { + "epoch": 0.19911928010721808, + "grad_norm": 0.2292814999818802, + "learning_rate": 4.963695672923663e-06, + "loss": 0.7018, + "step": 520 + }, + { + "epoch": 0.19950220179973197, + "grad_norm": 0.23721252381801605, + "learning_rate": 4.963523011503775e-06, + "loss": 0.6805, + "step": 521 + }, + { + "epoch": 0.19988512349224583, + "grad_norm": 0.24078230559825897, + "learning_rate": 4.9633499434904685e-06, + "loss": 0.6988, + "step": 522 + }, + { + "epoch": 0.2002680451847597, + "grad_norm": 0.2441691756248474, + "learning_rate": 4.96317646891231e-06, + "loss": 0.6765, + "step": 523 + }, + { + "epoch": 0.2006509668772736, + "grad_norm": 0.23363593220710754, + "learning_rate": 4.96300258779793e-06, + "loss": 0.6955, + "step": 524 + }, + { + "epoch": 0.2010338885697875, + "grad_norm": 0.24009943008422852, + "learning_rate": 4.962828300176028e-06, + "loss": 0.7068, + "step": 525 + }, + { + "epoch": 0.20141681026230135, + "grad_norm": 0.236362487077713, + "learning_rate": 4.9626536060753686e-06, + "loss": 0.6873, + "step": 526 + }, + { + "epoch": 0.20179973195481524, + "grad_norm": 0.2265264391899109, + "learning_rate": 4.962478505524783e-06, + "loss": 0.6931, + "step": 527 + }, + { + "epoch": 0.20218265364732912, + "grad_norm": 0.23028504848480225, + "learning_rate": 4.962302998553173e-06, + "loss": 0.6955, + "step": 528 + }, + { + "epoch": 0.202565575339843, + "grad_norm": 0.2343558669090271, + "learning_rate": 4.9621270851895035e-06, + "loss": 0.7112, + "step": 529 + }, + { + "epoch": 0.20294849703235687, + "grad_norm": 0.2429768443107605, + "learning_rate": 4.961950765462809e-06, + "loss": 0.7068, + "step": 530 + }, + { + "epoch": 0.20333141872487076, + "grad_norm": 0.23821046948432922, + "learning_rate": 4.961774039402189e-06, + "loss": 0.7064, + "step": 531 + }, + { + "epoch": 0.20371434041738465, + "grad_norm": 0.23160751163959503, + "learning_rate": 4.961596907036813e-06, + "loss": 0.7006, + "step": 532 + }, + { + "epoch": 0.20409726210989854, + "grad_norm": 0.2379685938358307, + "learning_rate": 4.961419368395916e-06, + "loss": 0.6992, + "step": 533 + }, + { + "epoch": 0.2044801838024124, + "grad_norm": 0.23077930510044098, + "learning_rate": 4.961241423508799e-06, + "loss": 0.6935, + "step": 534 + }, + { + "epoch": 0.20486310549492628, + "grad_norm": 0.23924565315246582, + "learning_rate": 4.9610630724048315e-06, + "loss": 0.6875, + "step": 535 + }, + { + "epoch": 0.20524602718744017, + "grad_norm": 0.24769453704357147, + "learning_rate": 4.960884315113449e-06, + "loss": 0.6924, + "step": 536 + }, + { + "epoch": 0.20562894887995406, + "grad_norm": 0.23660869896411896, + "learning_rate": 4.960705151664155e-06, + "loss": 0.6824, + "step": 537 + }, + { + "epoch": 0.20601187057246792, + "grad_norm": 0.24358759820461273, + "learning_rate": 4.960525582086519e-06, + "loss": 0.7247, + "step": 538 + }, + { + "epoch": 0.2063947922649818, + "grad_norm": 0.23302821815013885, + "learning_rate": 4.960345606410179e-06, + "loss": 0.7004, + "step": 539 + }, + { + "epoch": 0.2067777139574957, + "grad_norm": 0.2414388209581375, + "learning_rate": 4.96016522466484e-06, + "loss": 0.708, + "step": 540 + }, + { + "epoch": 0.20716063565000958, + "grad_norm": 0.23602741956710815, + "learning_rate": 4.959984436880271e-06, + "loss": 0.6991, + "step": 541 + }, + { + "epoch": 0.20754355734252344, + "grad_norm": 0.23501873016357422, + "learning_rate": 4.959803243086312e-06, + "loss": 0.6956, + "step": 542 + }, + { + "epoch": 0.20792647903503733, + "grad_norm": 0.2384674996137619, + "learning_rate": 4.959621643312867e-06, + "loss": 0.691, + "step": 543 + }, + { + "epoch": 0.20830940072755122, + "grad_norm": 0.24334029853343964, + "learning_rate": 4.959439637589909e-06, + "loss": 0.7004, + "step": 544 + }, + { + "epoch": 0.2086923224200651, + "grad_norm": 0.2337769865989685, + "learning_rate": 4.959257225947477e-06, + "loss": 0.7001, + "step": 545 + }, + { + "epoch": 0.20907524411257897, + "grad_norm": 0.24959228932857513, + "learning_rate": 4.959074408415676e-06, + "loss": 0.6854, + "step": 546 + }, + { + "epoch": 0.20945816580509286, + "grad_norm": 0.23717932403087616, + "learning_rate": 4.958891185024682e-06, + "loss": 0.6883, + "step": 547 + }, + { + "epoch": 0.20984108749760674, + "grad_norm": 0.22805754840373993, + "learning_rate": 4.958707555804734e-06, + "loss": 0.6875, + "step": 548 + }, + { + "epoch": 0.21022400919012063, + "grad_norm": 0.23835137486457825, + "learning_rate": 4.958523520786137e-06, + "loss": 0.6849, + "step": 549 + }, + { + "epoch": 0.2106069308826345, + "grad_norm": 0.2451052963733673, + "learning_rate": 4.958339079999268e-06, + "loss": 0.6947, + "step": 550 + }, + { + "epoch": 0.21098985257514838, + "grad_norm": 0.23878198862075806, + "learning_rate": 4.958154233474567e-06, + "loss": 0.6897, + "step": 551 + }, + { + "epoch": 0.21137277426766227, + "grad_norm": 0.243367537856102, + "learning_rate": 4.957968981242543e-06, + "loss": 0.7118, + "step": 552 + }, + { + "epoch": 0.21175569596017615, + "grad_norm": 0.2516103684902191, + "learning_rate": 4.957783323333768e-06, + "loss": 0.7118, + "step": 553 + }, + { + "epoch": 0.21213861765269001, + "grad_norm": 0.24193982779979706, + "learning_rate": 4.9575972597788875e-06, + "loss": 0.6816, + "step": 554 + }, + { + "epoch": 0.2125215393452039, + "grad_norm": 0.2456333190202713, + "learning_rate": 4.95741079060861e-06, + "loss": 0.7218, + "step": 555 + }, + { + "epoch": 0.2129044610377178, + "grad_norm": 0.23211970925331116, + "learning_rate": 4.9572239158537095e-06, + "loss": 0.6877, + "step": 556 + }, + { + "epoch": 0.21328738273023168, + "grad_norm": 0.2383391559123993, + "learning_rate": 4.95703663554503e-06, + "loss": 0.6862, + "step": 557 + }, + { + "epoch": 0.21367030442274554, + "grad_norm": 0.2499326914548874, + "learning_rate": 4.956848949713481e-06, + "loss": 0.7077, + "step": 558 + }, + { + "epoch": 0.21405322611525943, + "grad_norm": 0.24644067883491516, + "learning_rate": 4.95666085839004e-06, + "loss": 0.7048, + "step": 559 + }, + { + "epoch": 0.21443614780777331, + "grad_norm": 0.24420374631881714, + "learning_rate": 4.95647236160575e-06, + "loss": 0.6893, + "step": 560 + }, + { + "epoch": 0.2148190695002872, + "grad_norm": 0.23663920164108276, + "learning_rate": 4.956283459391721e-06, + "loss": 0.6912, + "step": 561 + }, + { + "epoch": 0.21520199119280106, + "grad_norm": 0.2413814216852188, + "learning_rate": 4.956094151779132e-06, + "loss": 0.6999, + "step": 562 + }, + { + "epoch": 0.21558491288531495, + "grad_norm": 0.25189071893692017, + "learning_rate": 4.955904438799225e-06, + "loss": 0.7002, + "step": 563 + }, + { + "epoch": 0.21596783457782884, + "grad_norm": 0.24884158372879028, + "learning_rate": 4.955714320483314e-06, + "loss": 0.698, + "step": 564 + }, + { + "epoch": 0.21635075627034273, + "grad_norm": 0.23854660987854004, + "learning_rate": 4.955523796862775e-06, + "loss": 0.6886, + "step": 565 + }, + { + "epoch": 0.21673367796285659, + "grad_norm": 0.24080811440944672, + "learning_rate": 4.9553328679690545e-06, + "loss": 0.704, + "step": 566 + }, + { + "epoch": 0.21711659965537047, + "grad_norm": 0.24664145708084106, + "learning_rate": 4.955141533833663e-06, + "loss": 0.6978, + "step": 567 + }, + { + "epoch": 0.21749952134788436, + "grad_norm": 0.240121990442276, + "learning_rate": 4.954949794488181e-06, + "loss": 0.7007, + "step": 568 + }, + { + "epoch": 0.21788244304039825, + "grad_norm": 0.2344406694173813, + "learning_rate": 4.954757649964253e-06, + "loss": 0.7014, + "step": 569 + }, + { + "epoch": 0.2182653647329121, + "grad_norm": 0.23647718131542206, + "learning_rate": 4.954565100293593e-06, + "loss": 0.6952, + "step": 570 + }, + { + "epoch": 0.218648286425426, + "grad_norm": 0.25824204087257385, + "learning_rate": 4.954372145507979e-06, + "loss": 0.7084, + "step": 571 + }, + { + "epoch": 0.21903120811793989, + "grad_norm": 0.23864921927452087, + "learning_rate": 4.954178785639257e-06, + "loss": 0.7012, + "step": 572 + }, + { + "epoch": 0.21941412981045377, + "grad_norm": 0.23659467697143555, + "learning_rate": 4.953985020719342e-06, + "loss": 0.7097, + "step": 573 + }, + { + "epoch": 0.21979705150296763, + "grad_norm": 0.23632408678531647, + "learning_rate": 4.953790850780213e-06, + "loss": 0.7185, + "step": 574 + }, + { + "epoch": 0.22017997319548152, + "grad_norm": 0.23713268339633942, + "learning_rate": 4.9535962758539155e-06, + "loss": 0.702, + "step": 575 + }, + { + "epoch": 0.2205628948879954, + "grad_norm": 0.23817777633666992, + "learning_rate": 4.953401295972565e-06, + "loss": 0.6957, + "step": 576 + }, + { + "epoch": 0.2209458165805093, + "grad_norm": 0.24862009286880493, + "learning_rate": 4.953205911168343e-06, + "loss": 0.6936, + "step": 577 + }, + { + "epoch": 0.22132873827302316, + "grad_norm": 0.24382250010967255, + "learning_rate": 4.953010121473495e-06, + "loss": 0.7066, + "step": 578 + }, + { + "epoch": 0.22171165996553704, + "grad_norm": 0.23864789307117462, + "learning_rate": 4.952813926920335e-06, + "loss": 0.6881, + "step": 579 + }, + { + "epoch": 0.22209458165805093, + "grad_norm": 0.24444547295570374, + "learning_rate": 4.952617327541246e-06, + "loss": 0.7082, + "step": 580 + }, + { + "epoch": 0.22247750335056482, + "grad_norm": 0.24924702942371368, + "learning_rate": 4.952420323368673e-06, + "loss": 0.732, + "step": 581 + }, + { + "epoch": 0.22286042504307868, + "grad_norm": 0.24768982827663422, + "learning_rate": 4.952222914435134e-06, + "loss": 0.7057, + "step": 582 + }, + { + "epoch": 0.22324334673559257, + "grad_norm": 0.23804764449596405, + "learning_rate": 4.952025100773208e-06, + "loss": 0.7024, + "step": 583 + }, + { + "epoch": 0.22362626842810646, + "grad_norm": 0.23817269504070282, + "learning_rate": 4.951826882415544e-06, + "loss": 0.7052, + "step": 584 + }, + { + "epoch": 0.22400919012062034, + "grad_norm": 0.2470194399356842, + "learning_rate": 4.9516282593948566e-06, + "loss": 0.6916, + "step": 585 + }, + { + "epoch": 0.2243921118131342, + "grad_norm": 0.252437949180603, + "learning_rate": 4.951429231743929e-06, + "loss": 0.6999, + "step": 586 + }, + { + "epoch": 0.2247750335056481, + "grad_norm": 0.23925770819187164, + "learning_rate": 4.951229799495609e-06, + "loss": 0.6975, + "step": 587 + }, + { + "epoch": 0.22515795519816198, + "grad_norm": 0.23760266602039337, + "learning_rate": 4.951029962682811e-06, + "loss": 0.6877, + "step": 588 + }, + { + "epoch": 0.22554087689067587, + "grad_norm": 0.25361642241477966, + "learning_rate": 4.950829721338518e-06, + "loss": 0.6848, + "step": 589 + }, + { + "epoch": 0.22592379858318973, + "grad_norm": 0.25293734669685364, + "learning_rate": 4.95062907549578e-06, + "loss": 0.6876, + "step": 590 + }, + { + "epoch": 0.22630672027570362, + "grad_norm": 0.24178287386894226, + "learning_rate": 4.9504280251877105e-06, + "loss": 0.6681, + "step": 591 + }, + { + "epoch": 0.2266896419682175, + "grad_norm": 0.2571343779563904, + "learning_rate": 4.9502265704474935e-06, + "loss": 0.6889, + "step": 592 + }, + { + "epoch": 0.2270725636607314, + "grad_norm": 0.244108647108078, + "learning_rate": 4.950024711308379e-06, + "loss": 0.6934, + "step": 593 + }, + { + "epoch": 0.22745548535324525, + "grad_norm": 0.2604058086872101, + "learning_rate": 4.949822447803681e-06, + "loss": 0.7133, + "step": 594 + }, + { + "epoch": 0.22783840704575914, + "grad_norm": 0.24538995325565338, + "learning_rate": 4.949619779966783e-06, + "loss": 0.6937, + "step": 595 + }, + { + "epoch": 0.22822132873827303, + "grad_norm": 0.2400553971529007, + "learning_rate": 4.949416707831134e-06, + "loss": 0.6809, + "step": 596 + }, + { + "epoch": 0.22860425043078691, + "grad_norm": 0.2515188157558441, + "learning_rate": 4.9492132314302515e-06, + "loss": 0.6999, + "step": 597 + }, + { + "epoch": 0.22898717212330078, + "grad_norm": 0.23625919222831726, + "learning_rate": 4.949009350797717e-06, + "loss": 0.6963, + "step": 598 + }, + { + "epoch": 0.22937009381581466, + "grad_norm": 0.24061603844165802, + "learning_rate": 4.94880506596718e-06, + "loss": 0.6925, + "step": 599 + }, + { + "epoch": 0.22975301550832855, + "grad_norm": 0.24639979004859924, + "learning_rate": 4.9486003769723574e-06, + "loss": 0.6729, + "step": 600 + }, + { + "epoch": 0.23013593720084244, + "grad_norm": 0.2444503903388977, + "learning_rate": 4.948395283847033e-06, + "loss": 0.7179, + "step": 601 + }, + { + "epoch": 0.2305188588933563, + "grad_norm": 0.2438957691192627, + "learning_rate": 4.948189786625055e-06, + "loss": 0.6953, + "step": 602 + }, + { + "epoch": 0.2309017805858702, + "grad_norm": 0.24474462866783142, + "learning_rate": 4.947983885340341e-06, + "loss": 0.6761, + "step": 603 + }, + { + "epoch": 0.23128470227838407, + "grad_norm": 0.23572909832000732, + "learning_rate": 4.947777580026872e-06, + "loss": 0.7164, + "step": 604 + }, + { + "epoch": 0.23166762397089796, + "grad_norm": 0.24651245772838593, + "learning_rate": 4.947570870718701e-06, + "loss": 0.6898, + "step": 605 + }, + { + "epoch": 0.23205054566341182, + "grad_norm": 0.24207252264022827, + "learning_rate": 4.947363757449941e-06, + "loss": 0.6962, + "step": 606 + }, + { + "epoch": 0.2324334673559257, + "grad_norm": 0.2437017560005188, + "learning_rate": 4.947156240254778e-06, + "loss": 0.6916, + "step": 607 + }, + { + "epoch": 0.2328163890484396, + "grad_norm": 0.23833702504634857, + "learning_rate": 4.946948319167461e-06, + "loss": 0.6734, + "step": 608 + }, + { + "epoch": 0.23319931074095349, + "grad_norm": 0.2502235770225525, + "learning_rate": 4.946739994222305e-06, + "loss": 0.6793, + "step": 609 + }, + { + "epoch": 0.23358223243346735, + "grad_norm": 0.24891985952854156, + "learning_rate": 4.946531265453694e-06, + "loss": 0.6727, + "step": 610 + }, + { + "epoch": 0.23396515412598123, + "grad_norm": 0.24262681603431702, + "learning_rate": 4.9463221328960785e-06, + "loss": 0.6662, + "step": 611 + }, + { + "epoch": 0.23434807581849512, + "grad_norm": 0.2385813146829605, + "learning_rate": 4.946112596583974e-06, + "loss": 0.6895, + "step": 612 + }, + { + "epoch": 0.234730997511009, + "grad_norm": 0.24113699793815613, + "learning_rate": 4.945902656551964e-06, + "loss": 0.7058, + "step": 613 + }, + { + "epoch": 0.23511391920352287, + "grad_norm": 0.24542860686779022, + "learning_rate": 4.945692312834698e-06, + "loss": 0.6796, + "step": 614 + }, + { + "epoch": 0.23549684089603676, + "grad_norm": 0.23808449506759644, + "learning_rate": 4.945481565466891e-06, + "loss": 0.7038, + "step": 615 + }, + { + "epoch": 0.23587976258855065, + "grad_norm": 0.23834019899368286, + "learning_rate": 4.945270414483328e-06, + "loss": 0.676, + "step": 616 + }, + { + "epoch": 0.23626268428106453, + "grad_norm": 0.23723448812961578, + "learning_rate": 4.945058859918858e-06, + "loss": 0.6925, + "step": 617 + }, + { + "epoch": 0.2366456059735784, + "grad_norm": 0.24539196491241455, + "learning_rate": 4.944846901808397e-06, + "loss": 0.6928, + "step": 618 + }, + { + "epoch": 0.23702852766609228, + "grad_norm": 0.24122965335845947, + "learning_rate": 4.944634540186927e-06, + "loss": 0.6637, + "step": 619 + }, + { + "epoch": 0.23741144935860617, + "grad_norm": 0.2460906058549881, + "learning_rate": 4.944421775089497e-06, + "loss": 0.7034, + "step": 620 + }, + { + "epoch": 0.23779437105112006, + "grad_norm": 0.2400168925523758, + "learning_rate": 4.9442086065512254e-06, + "loss": 0.688, + "step": 621 + }, + { + "epoch": 0.23817729274363392, + "grad_norm": 0.2557614743709564, + "learning_rate": 4.943995034607292e-06, + "loss": 0.7188, + "step": 622 + }, + { + "epoch": 0.2385602144361478, + "grad_norm": 0.2443343997001648, + "learning_rate": 4.943781059292948e-06, + "loss": 0.6951, + "step": 623 + }, + { + "epoch": 0.2389431361286617, + "grad_norm": 0.2577515244483948, + "learning_rate": 4.943566680643508e-06, + "loss": 0.7086, + "step": 624 + }, + { + "epoch": 0.23932605782117558, + "grad_norm": 0.2645318806171417, + "learning_rate": 4.943351898694353e-06, + "loss": 0.6923, + "step": 625 + }, + { + "epoch": 0.23970897951368944, + "grad_norm": 0.24452121555805206, + "learning_rate": 4.943136713480934e-06, + "loss": 0.6806, + "step": 626 + }, + { + "epoch": 0.24009190120620333, + "grad_norm": 0.2464938759803772, + "learning_rate": 4.942921125038765e-06, + "loss": 0.7121, + "step": 627 + }, + { + "epoch": 0.24047482289871722, + "grad_norm": 0.24740079045295715, + "learning_rate": 4.942705133403429e-06, + "loss": 0.6819, + "step": 628 + }, + { + "epoch": 0.2408577445912311, + "grad_norm": 0.24974393844604492, + "learning_rate": 4.942488738610574e-06, + "loss": 0.6604, + "step": 629 + }, + { + "epoch": 0.24124066628374496, + "grad_norm": 0.24305953085422516, + "learning_rate": 4.942271940695913e-06, + "loss": 0.6738, + "step": 630 + }, + { + "epoch": 0.24162358797625885, + "grad_norm": 0.2633149027824402, + "learning_rate": 4.942054739695231e-06, + "loss": 0.6646, + "step": 631 + }, + { + "epoch": 0.24200650966877274, + "grad_norm": 0.24853935837745667, + "learning_rate": 4.941837135644374e-06, + "loss": 0.687, + "step": 632 + }, + { + "epoch": 0.24238943136128663, + "grad_norm": 0.24881082773208618, + "learning_rate": 4.941619128579257e-06, + "loss": 0.6794, + "step": 633 + }, + { + "epoch": 0.2427723530538005, + "grad_norm": 0.26114174723625183, + "learning_rate": 4.94140071853586e-06, + "loss": 0.708, + "step": 634 + }, + { + "epoch": 0.24315527474631438, + "grad_norm": 0.2673856317996979, + "learning_rate": 4.941181905550233e-06, + "loss": 0.6966, + "step": 635 + }, + { + "epoch": 0.24353819643882826, + "grad_norm": 0.25460389256477356, + "learning_rate": 4.940962689658488e-06, + "loss": 0.7009, + "step": 636 + }, + { + "epoch": 0.24392111813134215, + "grad_norm": 0.25717636942863464, + "learning_rate": 4.940743070896806e-06, + "loss": 0.7, + "step": 637 + }, + { + "epoch": 0.244304039823856, + "grad_norm": 0.25204774737358093, + "learning_rate": 4.940523049301436e-06, + "loss": 0.6667, + "step": 638 + }, + { + "epoch": 0.2446869615163699, + "grad_norm": 0.24111990630626678, + "learning_rate": 4.940302624908689e-06, + "loss": 0.6826, + "step": 639 + }, + { + "epoch": 0.2450698832088838, + "grad_norm": 0.2586113512516022, + "learning_rate": 4.9400817977549455e-06, + "loss": 0.691, + "step": 640 + }, + { + "epoch": 0.24545280490139768, + "grad_norm": 0.24710243940353394, + "learning_rate": 4.939860567876652e-06, + "loss": 0.7184, + "step": 641 + }, + { + "epoch": 0.24583572659391154, + "grad_norm": 0.2547800838947296, + "learning_rate": 4.939638935310324e-06, + "loss": 0.6763, + "step": 642 + }, + { + "epoch": 0.24621864828642542, + "grad_norm": 0.25336694717407227, + "learning_rate": 4.939416900092537e-06, + "loss": 0.6925, + "step": 643 + }, + { + "epoch": 0.2466015699789393, + "grad_norm": 0.24745817482471466, + "learning_rate": 4.9391944622599405e-06, + "loss": 0.6996, + "step": 644 + }, + { + "epoch": 0.2469844916714532, + "grad_norm": 0.2559024691581726, + "learning_rate": 4.9389716218492445e-06, + "loss": 0.6987, + "step": 645 + }, + { + "epoch": 0.24736741336396706, + "grad_norm": 0.2510485053062439, + "learning_rate": 4.938748378897229e-06, + "loss": 0.6981, + "step": 646 + }, + { + "epoch": 0.24775033505648095, + "grad_norm": 0.24602986872196198, + "learning_rate": 4.938524733440738e-06, + "loss": 0.6833, + "step": 647 + }, + { + "epoch": 0.24813325674899483, + "grad_norm": 0.2407660037279129, + "learning_rate": 4.938300685516686e-06, + "loss": 0.6933, + "step": 648 + }, + { + "epoch": 0.24851617844150872, + "grad_norm": 0.2540413439273834, + "learning_rate": 4.938076235162047e-06, + "loss": 0.6903, + "step": 649 + }, + { + "epoch": 0.24889910013402258, + "grad_norm": 0.26849615573883057, + "learning_rate": 4.937851382413869e-06, + "loss": 0.6988, + "step": 650 + }, + { + "epoch": 0.24928202182653647, + "grad_norm": 0.248337984085083, + "learning_rate": 4.9376261273092614e-06, + "loss": 0.6965, + "step": 651 + }, + { + "epoch": 0.24966494351905036, + "grad_norm": 0.24852517247200012, + "learning_rate": 4.937400469885401e-06, + "loss": 0.6975, + "step": 652 + }, + { + "epoch": 0.25004786521156425, + "grad_norm": 0.24815134704113007, + "learning_rate": 4.937174410179534e-06, + "loss": 0.7148, + "step": 653 + }, + { + "epoch": 0.2504307869040781, + "grad_norm": 0.242036372423172, + "learning_rate": 4.936947948228967e-06, + "loss": 0.6911, + "step": 654 + }, + { + "epoch": 0.250813708596592, + "grad_norm": 0.24787138402462006, + "learning_rate": 4.936721084071079e-06, + "loss": 0.708, + "step": 655 + }, + { + "epoch": 0.2511966302891059, + "grad_norm": 0.2459169179201126, + "learning_rate": 4.936493817743313e-06, + "loss": 0.6897, + "step": 656 + }, + { + "epoch": 0.25157955198161974, + "grad_norm": 0.2538677453994751, + "learning_rate": 4.936266149283176e-06, + "loss": 0.7044, + "step": 657 + }, + { + "epoch": 0.25196247367413366, + "grad_norm": 0.2510700821876526, + "learning_rate": 4.936038078728247e-06, + "loss": 0.6953, + "step": 658 + }, + { + "epoch": 0.2523453953666475, + "grad_norm": 0.24351082742214203, + "learning_rate": 4.935809606116165e-06, + "loss": 0.6808, + "step": 659 + }, + { + "epoch": 0.2527283170591614, + "grad_norm": 0.25081127882003784, + "learning_rate": 4.93558073148464e-06, + "loss": 0.702, + "step": 660 + }, + { + "epoch": 0.2531112387516753, + "grad_norm": 0.2425825446844101, + "learning_rate": 4.935351454871447e-06, + "loss": 0.7019, + "step": 661 + }, + { + "epoch": 0.25349416044418915, + "grad_norm": 0.24126264452934265, + "learning_rate": 4.935121776314425e-06, + "loss": 0.721, + "step": 662 + }, + { + "epoch": 0.25387708213670307, + "grad_norm": 0.2510846257209778, + "learning_rate": 4.934891695851483e-06, + "loss": 0.6832, + "step": 663 + }, + { + "epoch": 0.25426000382921693, + "grad_norm": 0.2648461163043976, + "learning_rate": 4.934661213520596e-06, + "loss": 0.693, + "step": 664 + }, + { + "epoch": 0.2546429255217308, + "grad_norm": 0.24153569340705872, + "learning_rate": 4.934430329359803e-06, + "loss": 0.6797, + "step": 665 + }, + { + "epoch": 0.2550258472142447, + "grad_norm": 0.25208204984664917, + "learning_rate": 4.934199043407209e-06, + "loss": 0.6902, + "step": 666 + }, + { + "epoch": 0.25540876890675857, + "grad_norm": 0.2638026177883148, + "learning_rate": 4.9339673557009875e-06, + "loss": 0.686, + "step": 667 + }, + { + "epoch": 0.2557916905992724, + "grad_norm": 0.2548025846481323, + "learning_rate": 4.933735266279377e-06, + "loss": 0.6804, + "step": 668 + }, + { + "epoch": 0.25617461229178634, + "grad_norm": 0.2605096101760864, + "learning_rate": 4.9335027751806855e-06, + "loss": 0.7069, + "step": 669 + }, + { + "epoch": 0.2565575339843002, + "grad_norm": 0.24284419417381287, + "learning_rate": 4.933269882443281e-06, + "loss": 0.6813, + "step": 670 + }, + { + "epoch": 0.2569404556768141, + "grad_norm": 0.2466583251953125, + "learning_rate": 4.933036588105604e-06, + "loss": 0.7137, + "step": 671 + }, + { + "epoch": 0.257323377369328, + "grad_norm": 0.2555830478668213, + "learning_rate": 4.932802892206158e-06, + "loss": 0.6957, + "step": 672 + }, + { + "epoch": 0.25770629906184184, + "grad_norm": 0.24138735234737396, + "learning_rate": 4.932568794783513e-06, + "loss": 0.6662, + "step": 673 + }, + { + "epoch": 0.25808922075435575, + "grad_norm": 0.2518264651298523, + "learning_rate": 4.932334295876306e-06, + "loss": 0.6965, + "step": 674 + }, + { + "epoch": 0.2584721424468696, + "grad_norm": 0.2566470801830292, + "learning_rate": 4.9320993955232396e-06, + "loss": 0.7018, + "step": 675 + }, + { + "epoch": 0.2588550641393835, + "grad_norm": 0.24778035283088684, + "learning_rate": 4.931864093763085e-06, + "loss": 0.6782, + "step": 676 + }, + { + "epoch": 0.2592379858318974, + "grad_norm": 0.24779430031776428, + "learning_rate": 4.931628390634674e-06, + "loss": 0.674, + "step": 677 + }, + { + "epoch": 0.25962090752441125, + "grad_norm": 0.2474948763847351, + "learning_rate": 4.931392286176911e-06, + "loss": 0.6987, + "step": 678 + }, + { + "epoch": 0.26000382921692516, + "grad_norm": 0.24874965846538544, + "learning_rate": 4.931155780428764e-06, + "loss": 0.7017, + "step": 679 + }, + { + "epoch": 0.260386750909439, + "grad_norm": 0.2501877248287201, + "learning_rate": 4.930918873429268e-06, + "loss": 0.6959, + "step": 680 + }, + { + "epoch": 0.2607696726019529, + "grad_norm": 0.24506975710391998, + "learning_rate": 4.93068156521752e-06, + "loss": 0.6941, + "step": 681 + }, + { + "epoch": 0.2611525942944668, + "grad_norm": 0.24200353026390076, + "learning_rate": 4.930443855832689e-06, + "loss": 0.6911, + "step": 682 + }, + { + "epoch": 0.26153551598698066, + "grad_norm": 0.24502404034137726, + "learning_rate": 4.9302057453140085e-06, + "loss": 0.6985, + "step": 683 + }, + { + "epoch": 0.2619184376794945, + "grad_norm": 0.24491210281848907, + "learning_rate": 4.929967233700776e-06, + "loss": 0.6831, + "step": 684 + }, + { + "epoch": 0.26230135937200844, + "grad_norm": 0.24369360506534576, + "learning_rate": 4.929728321032358e-06, + "loss": 0.6796, + "step": 685 + }, + { + "epoch": 0.2626842810645223, + "grad_norm": 0.24483689665794373, + "learning_rate": 4.929489007348185e-06, + "loss": 0.6918, + "step": 686 + }, + { + "epoch": 0.2630672027570362, + "grad_norm": 0.24488209187984467, + "learning_rate": 4.929249292687756e-06, + "loss": 0.6828, + "step": 687 + }, + { + "epoch": 0.26345012444955007, + "grad_norm": 0.2576925456523895, + "learning_rate": 4.929009177090632e-06, + "loss": 0.6968, + "step": 688 + }, + { + "epoch": 0.26383304614206393, + "grad_norm": 0.25123924016952515, + "learning_rate": 4.928768660596446e-06, + "loss": 0.6976, + "step": 689 + }, + { + "epoch": 0.26421596783457785, + "grad_norm": 0.2571817636489868, + "learning_rate": 4.928527743244893e-06, + "loss": 0.6797, + "step": 690 + }, + { + "epoch": 0.2645988895270917, + "grad_norm": 0.25800999999046326, + "learning_rate": 4.928286425075736e-06, + "loss": 0.6964, + "step": 691 + }, + { + "epoch": 0.26498181121960557, + "grad_norm": 0.2471861094236374, + "learning_rate": 4.928044706128803e-06, + "loss": 0.6927, + "step": 692 + }, + { + "epoch": 0.2653647329121195, + "grad_norm": 0.24648526310920715, + "learning_rate": 4.927802586443987e-06, + "loss": 0.6979, + "step": 693 + }, + { + "epoch": 0.26574765460463334, + "grad_norm": 0.25215795636177063, + "learning_rate": 4.927560066061251e-06, + "loss": 0.7, + "step": 694 + }, + { + "epoch": 0.26613057629714726, + "grad_norm": 0.2663182020187378, + "learning_rate": 4.927317145020622e-06, + "loss": 0.693, + "step": 695 + }, + { + "epoch": 0.2665134979896611, + "grad_norm": 0.25089675188064575, + "learning_rate": 4.9270738233621915e-06, + "loss": 0.6702, + "step": 696 + }, + { + "epoch": 0.266896419682175, + "grad_norm": 0.24895228445529938, + "learning_rate": 4.92683010112612e-06, + "loss": 0.6962, + "step": 697 + }, + { + "epoch": 0.2672793413746889, + "grad_norm": 0.2562544345855713, + "learning_rate": 4.926585978352632e-06, + "loss": 0.7017, + "step": 698 + }, + { + "epoch": 0.26766226306720275, + "grad_norm": 0.25019386410713196, + "learning_rate": 4.926341455082018e-06, + "loss": 0.6772, + "step": 699 + }, + { + "epoch": 0.2680451847597166, + "grad_norm": 0.2618144750595093, + "learning_rate": 4.926096531354637e-06, + "loss": 0.6831, + "step": 700 + }, + { + "epoch": 0.26842810645223053, + "grad_norm": 0.2517690062522888, + "learning_rate": 4.925851207210912e-06, + "loss": 0.675, + "step": 701 + }, + { + "epoch": 0.2688110281447444, + "grad_norm": 0.2592850923538208, + "learning_rate": 4.925605482691332e-06, + "loss": 0.6971, + "step": 702 + }, + { + "epoch": 0.2691939498372583, + "grad_norm": 0.26684650778770447, + "learning_rate": 4.9253593578364545e-06, + "loss": 0.701, + "step": 703 + }, + { + "epoch": 0.26957687152977217, + "grad_norm": 0.25434690713882446, + "learning_rate": 4.9251128326869e-06, + "loss": 0.6915, + "step": 704 + }, + { + "epoch": 0.269959793222286, + "grad_norm": 0.2535107731819153, + "learning_rate": 4.924865907283356e-06, + "loss": 0.6853, + "step": 705 + }, + { + "epoch": 0.27034271491479994, + "grad_norm": 0.25544658303260803, + "learning_rate": 4.924618581666578e-06, + "loss": 0.6827, + "step": 706 + }, + { + "epoch": 0.2707256366073138, + "grad_norm": 0.26514726877212524, + "learning_rate": 4.924370855877384e-06, + "loss": 0.7171, + "step": 707 + }, + { + "epoch": 0.27110855829982766, + "grad_norm": 0.25380080938339233, + "learning_rate": 4.924122729956662e-06, + "loss": 0.6958, + "step": 708 + }, + { + "epoch": 0.2714914799923416, + "grad_norm": 0.25360482931137085, + "learning_rate": 4.9238742039453626e-06, + "loss": 0.7015, + "step": 709 + }, + { + "epoch": 0.27187440168485544, + "grad_norm": 0.25626882910728455, + "learning_rate": 4.923625277884504e-06, + "loss": 0.6868, + "step": 710 + }, + { + "epoch": 0.27225732337736935, + "grad_norm": 0.26761749386787415, + "learning_rate": 4.923375951815171e-06, + "loss": 0.6919, + "step": 711 + }, + { + "epoch": 0.2726402450698832, + "grad_norm": 0.263866126537323, + "learning_rate": 4.923126225778513e-06, + "loss": 0.7029, + "step": 712 + }, + { + "epoch": 0.2730231667623971, + "grad_norm": 0.2864058315753937, + "learning_rate": 4.9228760998157475e-06, + "loss": 0.6957, + "step": 713 + }, + { + "epoch": 0.273406088454911, + "grad_norm": 0.2648506760597229, + "learning_rate": 4.922625573968155e-06, + "loss": 0.6834, + "step": 714 + }, + { + "epoch": 0.27378901014742485, + "grad_norm": 0.2676790952682495, + "learning_rate": 4.922374648277085e-06, + "loss": 0.7012, + "step": 715 + }, + { + "epoch": 0.2741719318399387, + "grad_norm": 0.25565361976623535, + "learning_rate": 4.922123322783951e-06, + "loss": 0.7191, + "step": 716 + }, + { + "epoch": 0.2745548535324526, + "grad_norm": 0.2526341676712036, + "learning_rate": 4.921871597530234e-06, + "loss": 0.6957, + "step": 717 + }, + { + "epoch": 0.2749377752249665, + "grad_norm": 0.2454516589641571, + "learning_rate": 4.921619472557478e-06, + "loss": 0.6863, + "step": 718 + }, + { + "epoch": 0.2753206969174804, + "grad_norm": 0.2534882426261902, + "learning_rate": 4.9213669479072964e-06, + "loss": 0.6813, + "step": 719 + }, + { + "epoch": 0.27570361860999426, + "grad_norm": 0.2647958993911743, + "learning_rate": 4.921114023621368e-06, + "loss": 0.6723, + "step": 720 + }, + { + "epoch": 0.2760865403025081, + "grad_norm": 0.2482769936323166, + "learning_rate": 4.920860699741436e-06, + "loss": 0.6816, + "step": 721 + }, + { + "epoch": 0.27646946199502204, + "grad_norm": 0.25155022740364075, + "learning_rate": 4.920606976309311e-06, + "loss": 0.6829, + "step": 722 + }, + { + "epoch": 0.2768523836875359, + "grad_norm": 0.25965702533721924, + "learning_rate": 4.920352853366869e-06, + "loss": 0.6881, + "step": 723 + }, + { + "epoch": 0.27723530538004976, + "grad_norm": 0.26698043942451477, + "learning_rate": 4.9200983309560505e-06, + "loss": 0.6947, + "step": 724 + }, + { + "epoch": 0.2776182270725637, + "grad_norm": 0.2601257562637329, + "learning_rate": 4.919843409118863e-06, + "loss": 0.6816, + "step": 725 + }, + { + "epoch": 0.27800114876507753, + "grad_norm": 0.24915724992752075, + "learning_rate": 4.919588087897383e-06, + "loss": 0.6667, + "step": 726 + }, + { + "epoch": 0.27838407045759145, + "grad_norm": 0.2661370635032654, + "learning_rate": 4.919332367333748e-06, + "loss": 0.7011, + "step": 727 + }, + { + "epoch": 0.2787669921501053, + "grad_norm": 0.24617598950862885, + "learning_rate": 4.919076247470163e-06, + "loss": 0.6789, + "step": 728 + }, + { + "epoch": 0.27914991384261917, + "grad_norm": 0.259954571723938, + "learning_rate": 4.918819728348901e-06, + "loss": 0.6858, + "step": 729 + }, + { + "epoch": 0.2795328355351331, + "grad_norm": 0.2742285430431366, + "learning_rate": 4.918562810012299e-06, + "loss": 0.6576, + "step": 730 + }, + { + "epoch": 0.27991575722764694, + "grad_norm": 0.2507971525192261, + "learning_rate": 4.91830549250276e-06, + "loss": 0.7046, + "step": 731 + }, + { + "epoch": 0.2802986789201608, + "grad_norm": 0.25771722197532654, + "learning_rate": 4.918047775862752e-06, + "loss": 0.6869, + "step": 732 + }, + { + "epoch": 0.2806816006126747, + "grad_norm": 0.2926652431488037, + "learning_rate": 4.917789660134812e-06, + "loss": 0.6852, + "step": 733 + }, + { + "epoch": 0.2810645223051886, + "grad_norm": 0.2552712559700012, + "learning_rate": 4.91753114536154e-06, + "loss": 0.6691, + "step": 734 + }, + { + "epoch": 0.2814474439977025, + "grad_norm": 0.2440652698278427, + "learning_rate": 4.917272231585601e-06, + "loss": 0.665, + "step": 735 + }, + { + "epoch": 0.28183036569021636, + "grad_norm": 0.286358505487442, + "learning_rate": 4.91701291884973e-06, + "loss": 0.693, + "step": 736 + }, + { + "epoch": 0.2822132873827302, + "grad_norm": 0.2688429653644562, + "learning_rate": 4.916753207196725e-06, + "loss": 0.6837, + "step": 737 + }, + { + "epoch": 0.28259620907524413, + "grad_norm": 0.25828817486763, + "learning_rate": 4.916493096669449e-06, + "loss": 0.6796, + "step": 738 + }, + { + "epoch": 0.282979130767758, + "grad_norm": 0.2633633017539978, + "learning_rate": 4.916232587310833e-06, + "loss": 0.7175, + "step": 739 + }, + { + "epoch": 0.28336205246027185, + "grad_norm": 0.26460349559783936, + "learning_rate": 4.915971679163873e-06, + "loss": 0.6979, + "step": 740 + }, + { + "epoch": 0.28374497415278577, + "grad_norm": 0.25029146671295166, + "learning_rate": 4.91571037227163e-06, + "loss": 0.6978, + "step": 741 + }, + { + "epoch": 0.2841278958452996, + "grad_norm": 0.2536710202693939, + "learning_rate": 4.915448666677233e-06, + "loss": 0.6759, + "step": 742 + }, + { + "epoch": 0.28451081753781354, + "grad_norm": 0.252708375453949, + "learning_rate": 4.915186562423872e-06, + "loss": 0.7197, + "step": 743 + }, + { + "epoch": 0.2848937392303274, + "grad_norm": 0.2493618130683899, + "learning_rate": 4.914924059554811e-06, + "loss": 0.6846, + "step": 744 + }, + { + "epoch": 0.28527666092284126, + "grad_norm": 0.25273728370666504, + "learning_rate": 4.914661158113372e-06, + "loss": 0.708, + "step": 745 + }, + { + "epoch": 0.2856595826153552, + "grad_norm": 0.25201109051704407, + "learning_rate": 4.914397858142945e-06, + "loss": 0.6955, + "step": 746 + }, + { + "epoch": 0.28604250430786904, + "grad_norm": 0.2539828419685364, + "learning_rate": 4.914134159686988e-06, + "loss": 0.6937, + "step": 747 + }, + { + "epoch": 0.2864254260003829, + "grad_norm": 0.24846617877483368, + "learning_rate": 4.913870062789023e-06, + "loss": 0.672, + "step": 748 + }, + { + "epoch": 0.2868083476928968, + "grad_norm": 0.25717365741729736, + "learning_rate": 4.913605567492636e-06, + "loss": 0.6838, + "step": 749 + }, + { + "epoch": 0.2871912693854107, + "grad_norm": 0.2502661347389221, + "learning_rate": 4.9133406738414845e-06, + "loss": 0.6803, + "step": 750 + }, + { + "epoch": 0.2875741910779246, + "grad_norm": 0.25706034898757935, + "learning_rate": 4.913075381879285e-06, + "loss": 0.6741, + "step": 751 + }, + { + "epoch": 0.28795711277043845, + "grad_norm": 0.26075178384780884, + "learning_rate": 4.9128096916498245e-06, + "loss": 0.6923, + "step": 752 + }, + { + "epoch": 0.2883400344629523, + "grad_norm": 0.2518405318260193, + "learning_rate": 4.912543603196954e-06, + "loss": 0.6902, + "step": 753 + }, + { + "epoch": 0.2887229561554662, + "grad_norm": 0.26775062084198, + "learning_rate": 4.912277116564588e-06, + "loss": 0.6924, + "step": 754 + }, + { + "epoch": 0.2891058778479801, + "grad_norm": 0.254680871963501, + "learning_rate": 4.912010231796711e-06, + "loss": 0.6906, + "step": 755 + }, + { + "epoch": 0.28948879954049395, + "grad_norm": 0.2532293200492859, + "learning_rate": 4.911742948937371e-06, + "loss": 0.6899, + "step": 756 + }, + { + "epoch": 0.28987172123300786, + "grad_norm": 0.2509908676147461, + "learning_rate": 4.911475268030681e-06, + "loss": 0.6785, + "step": 757 + }, + { + "epoch": 0.2902546429255217, + "grad_norm": 0.2599368989467621, + "learning_rate": 4.91120718912082e-06, + "loss": 0.7066, + "step": 758 + }, + { + "epoch": 0.29063756461803564, + "grad_norm": 0.24985715746879578, + "learning_rate": 4.910938712252036e-06, + "loss": 0.6531, + "step": 759 + }, + { + "epoch": 0.2910204863105495, + "grad_norm": 0.2594451904296875, + "learning_rate": 4.910669837468637e-06, + "loss": 0.6768, + "step": 760 + }, + { + "epoch": 0.29140340800306336, + "grad_norm": 0.24278171360492706, + "learning_rate": 4.910400564815e-06, + "loss": 0.6756, + "step": 761 + }, + { + "epoch": 0.2917863296955773, + "grad_norm": 0.26272347569465637, + "learning_rate": 4.9101308943355685e-06, + "loss": 0.682, + "step": 762 + }, + { + "epoch": 0.29216925138809113, + "grad_norm": 0.25561460852622986, + "learning_rate": 4.90986082607485e-06, + "loss": 0.7127, + "step": 763 + }, + { + "epoch": 0.292552173080605, + "grad_norm": 0.26740762591362, + "learning_rate": 4.909590360077417e-06, + "loss": 0.6828, + "step": 764 + }, + { + "epoch": 0.2929350947731189, + "grad_norm": 0.25563833117485046, + "learning_rate": 4.909319496387911e-06, + "loss": 0.699, + "step": 765 + }, + { + "epoch": 0.29331801646563277, + "grad_norm": 0.25876274704933167, + "learning_rate": 4.909048235051033e-06, + "loss": 0.6922, + "step": 766 + }, + { + "epoch": 0.2937009381581467, + "grad_norm": 0.2549036741256714, + "learning_rate": 4.908776576111558e-06, + "loss": 0.6704, + "step": 767 + }, + { + "epoch": 0.29408385985066055, + "grad_norm": 0.25628861784935, + "learning_rate": 4.908504519614319e-06, + "loss": 0.6954, + "step": 768 + }, + { + "epoch": 0.2944667815431744, + "grad_norm": 0.2500770688056946, + "learning_rate": 4.908232065604219e-06, + "loss": 0.6795, + "step": 769 + }, + { + "epoch": 0.2948497032356883, + "grad_norm": 0.2657643258571625, + "learning_rate": 4.907959214126225e-06, + "loss": 0.712, + "step": 770 + }, + { + "epoch": 0.2952326249282022, + "grad_norm": 0.26558786630630493, + "learning_rate": 4.907685965225369e-06, + "loss": 0.7003, + "step": 771 + }, + { + "epoch": 0.29561554662071604, + "grad_norm": 0.2577509582042694, + "learning_rate": 4.9074123189467506e-06, + "loss": 0.6976, + "step": 772 + }, + { + "epoch": 0.29599846831322996, + "grad_norm": 0.2677637040615082, + "learning_rate": 4.9071382753355335e-06, + "loss": 0.6844, + "step": 773 + }, + { + "epoch": 0.2963813900057438, + "grad_norm": 0.2563629448413849, + "learning_rate": 4.906863834436948e-06, + "loss": 0.6913, + "step": 774 + }, + { + "epoch": 0.29676431169825773, + "grad_norm": 0.26639705896377563, + "learning_rate": 4.906588996296289e-06, + "loss": 0.6813, + "step": 775 + }, + { + "epoch": 0.2971472333907716, + "grad_norm": 0.26096269488334656, + "learning_rate": 4.9063137609589166e-06, + "loss": 0.6939, + "step": 776 + }, + { + "epoch": 0.29753015508328545, + "grad_norm": 0.250601202249527, + "learning_rate": 4.906038128470258e-06, + "loss": 0.6461, + "step": 777 + }, + { + "epoch": 0.29791307677579937, + "grad_norm": 0.25932541489601135, + "learning_rate": 4.905762098875804e-06, + "loss": 0.6873, + "step": 778 + }, + { + "epoch": 0.29829599846831323, + "grad_norm": 0.256270170211792, + "learning_rate": 4.905485672221113e-06, + "loss": 0.6746, + "step": 779 + }, + { + "epoch": 0.2986789201608271, + "grad_norm": 0.2550566494464874, + "learning_rate": 4.905208848551809e-06, + "loss": 0.6768, + "step": 780 + }, + { + "epoch": 0.299061841853341, + "grad_norm": 0.24824075400829315, + "learning_rate": 4.904931627913578e-06, + "loss": 0.6764, + "step": 781 + }, + { + "epoch": 0.29944476354585486, + "grad_norm": 0.2658909261226654, + "learning_rate": 4.9046540103521765e-06, + "loss": 0.6801, + "step": 782 + }, + { + "epoch": 0.2998276852383688, + "grad_norm": 0.25569969415664673, + "learning_rate": 4.904375995913422e-06, + "loss": 0.696, + "step": 783 + }, + { + "epoch": 0.30021060693088264, + "grad_norm": 0.2601824104785919, + "learning_rate": 4.904097584643201e-06, + "loss": 0.707, + "step": 784 + }, + { + "epoch": 0.3005935286233965, + "grad_norm": 0.2594698667526245, + "learning_rate": 4.903818776587464e-06, + "loss": 0.705, + "step": 785 + }, + { + "epoch": 0.3009764503159104, + "grad_norm": 0.24596643447875977, + "learning_rate": 4.903539571792224e-06, + "loss": 0.6957, + "step": 786 + }, + { + "epoch": 0.3013593720084243, + "grad_norm": 0.25608620047569275, + "learning_rate": 4.903259970303567e-06, + "loss": 0.6978, + "step": 787 + }, + { + "epoch": 0.30174229370093814, + "grad_norm": 0.25500133633613586, + "learning_rate": 4.902979972167637e-06, + "loss": 0.6831, + "step": 788 + }, + { + "epoch": 0.30212521539345205, + "grad_norm": 0.2699365019798279, + "learning_rate": 4.902699577430649e-06, + "loss": 0.6903, + "step": 789 + }, + { + "epoch": 0.3025081370859659, + "grad_norm": 0.25621193647384644, + "learning_rate": 4.902418786138878e-06, + "loss": 0.6923, + "step": 790 + }, + { + "epoch": 0.3028910587784798, + "grad_norm": 0.2602829039096832, + "learning_rate": 4.902137598338669e-06, + "loss": 0.6556, + "step": 791 + }, + { + "epoch": 0.3032739804709937, + "grad_norm": 0.25946536660194397, + "learning_rate": 4.901856014076429e-06, + "loss": 0.6696, + "step": 792 + }, + { + "epoch": 0.30365690216350755, + "grad_norm": 0.2651613652706146, + "learning_rate": 4.901574033398635e-06, + "loss": 0.6842, + "step": 793 + }, + { + "epoch": 0.30403982385602146, + "grad_norm": 0.2593366801738739, + "learning_rate": 4.9012916563518255e-06, + "loss": 0.6854, + "step": 794 + }, + { + "epoch": 0.3044227455485353, + "grad_norm": 0.26688313484191895, + "learning_rate": 4.901008882982605e-06, + "loss": 0.6859, + "step": 795 + }, + { + "epoch": 0.3048056672410492, + "grad_norm": 0.254645973443985, + "learning_rate": 4.900725713337643e-06, + "loss": 0.672, + "step": 796 + }, + { + "epoch": 0.3051885889335631, + "grad_norm": 0.27753981947898865, + "learning_rate": 4.900442147463679e-06, + "loss": 0.6849, + "step": 797 + }, + { + "epoch": 0.30557151062607696, + "grad_norm": 0.25524258613586426, + "learning_rate": 4.90015818540751e-06, + "loss": 0.6982, + "step": 798 + }, + { + "epoch": 0.3059544323185909, + "grad_norm": 0.2585286796092987, + "learning_rate": 4.899873827216005e-06, + "loss": 0.6935, + "step": 799 + }, + { + "epoch": 0.30633735401110473, + "grad_norm": 0.2776353061199188, + "learning_rate": 4.899589072936097e-06, + "loss": 0.7047, + "step": 800 + }, + { + "epoch": 0.3067202757036186, + "grad_norm": 0.26528456807136536, + "learning_rate": 4.89930392261478e-06, + "loss": 0.716, + "step": 801 + }, + { + "epoch": 0.3071031973961325, + "grad_norm": 0.2629239559173584, + "learning_rate": 4.89901837629912e-06, + "loss": 0.6856, + "step": 802 + }, + { + "epoch": 0.30748611908864637, + "grad_norm": 0.25472939014434814, + "learning_rate": 4.8987324340362445e-06, + "loss": 0.6658, + "step": 803 + }, + { + "epoch": 0.30786904078116023, + "grad_norm": 0.25384870171546936, + "learning_rate": 4.898446095873345e-06, + "loss": 0.6897, + "step": 804 + }, + { + "epoch": 0.30825196247367415, + "grad_norm": 0.26003652811050415, + "learning_rate": 4.898159361857683e-06, + "loss": 0.6742, + "step": 805 + }, + { + "epoch": 0.308634884166188, + "grad_norm": 0.26269611716270447, + "learning_rate": 4.897872232036581e-06, + "loss": 0.685, + "step": 806 + }, + { + "epoch": 0.3090178058587019, + "grad_norm": 0.2580546736717224, + "learning_rate": 4.897584706457429e-06, + "loss": 0.6812, + "step": 807 + }, + { + "epoch": 0.3094007275512158, + "grad_norm": 0.25622835755348206, + "learning_rate": 4.897296785167681e-06, + "loss": 0.6828, + "step": 808 + }, + { + "epoch": 0.30978364924372964, + "grad_norm": 0.2555811107158661, + "learning_rate": 4.897008468214859e-06, + "loss": 0.6909, + "step": 809 + }, + { + "epoch": 0.31016657093624356, + "grad_norm": 0.2533404231071472, + "learning_rate": 4.896719755646548e-06, + "loss": 0.6968, + "step": 810 + }, + { + "epoch": 0.3105494926287574, + "grad_norm": 0.27069544792175293, + "learning_rate": 4.896430647510396e-06, + "loss": 0.6924, + "step": 811 + }, + { + "epoch": 0.3109324143212713, + "grad_norm": 0.2559230923652649, + "learning_rate": 4.896141143854122e-06, + "loss": 0.6983, + "step": 812 + }, + { + "epoch": 0.3113153360137852, + "grad_norm": 0.24792268872261047, + "learning_rate": 4.8958512447255055e-06, + "loss": 0.6721, + "step": 813 + }, + { + "epoch": 0.31169825770629905, + "grad_norm": 0.27516400814056396, + "learning_rate": 4.895560950172395e-06, + "loss": 0.6672, + "step": 814 + }, + { + "epoch": 0.31208117939881297, + "grad_norm": 0.25589725375175476, + "learning_rate": 4.895270260242701e-06, + "loss": 0.6861, + "step": 815 + }, + { + "epoch": 0.31246410109132683, + "grad_norm": 0.2611221373081207, + "learning_rate": 4.8949791749844e-06, + "loss": 0.6716, + "step": 816 + }, + { + "epoch": 0.3128470227838407, + "grad_norm": 0.26100388169288635, + "learning_rate": 4.8946876944455355e-06, + "loss": 0.6676, + "step": 817 + }, + { + "epoch": 0.3132299444763546, + "grad_norm": 0.27761179208755493, + "learning_rate": 4.8943958186742145e-06, + "loss": 0.6845, + "step": 818 + }, + { + "epoch": 0.31361286616886846, + "grad_norm": 0.26875218749046326, + "learning_rate": 4.894103547718611e-06, + "loss": 0.6902, + "step": 819 + }, + { + "epoch": 0.3139957878613823, + "grad_norm": 0.26297008991241455, + "learning_rate": 4.893810881626961e-06, + "loss": 0.7205, + "step": 820 + }, + { + "epoch": 0.31437870955389624, + "grad_norm": 0.2671011686325073, + "learning_rate": 4.893517820447568e-06, + "loss": 0.6782, + "step": 821 + }, + { + "epoch": 0.3147616312464101, + "grad_norm": 0.271348774433136, + "learning_rate": 4.8932243642288e-06, + "loss": 0.6993, + "step": 822 + }, + { + "epoch": 0.315144552938924, + "grad_norm": 0.271992564201355, + "learning_rate": 4.892930513019094e-06, + "loss": 0.682, + "step": 823 + }, + { + "epoch": 0.3155274746314379, + "grad_norm": 0.2696378827095032, + "learning_rate": 4.8926362668669454e-06, + "loss": 0.6805, + "step": 824 + }, + { + "epoch": 0.31591039632395174, + "grad_norm": 0.25821250677108765, + "learning_rate": 4.89234162582092e-06, + "loss": 0.6712, + "step": 825 + }, + { + "epoch": 0.31629331801646565, + "grad_norm": 0.2636469602584839, + "learning_rate": 4.892046589929645e-06, + "loss": 0.6956, + "step": 826 + }, + { + "epoch": 0.3166762397089795, + "grad_norm": 0.2635691165924072, + "learning_rate": 4.891751159241817e-06, + "loss": 0.6714, + "step": 827 + }, + { + "epoch": 0.3170591614014934, + "grad_norm": 0.26654869318008423, + "learning_rate": 4.891455333806194e-06, + "loss": 0.6818, + "step": 828 + }, + { + "epoch": 0.3174420830940073, + "grad_norm": 0.26420891284942627, + "learning_rate": 4.8911591136716015e-06, + "loss": 0.6848, + "step": 829 + }, + { + "epoch": 0.31782500478652115, + "grad_norm": 0.27182644605636597, + "learning_rate": 4.89086249888693e-06, + "loss": 0.7063, + "step": 830 + }, + { + "epoch": 0.31820792647903506, + "grad_norm": 0.26235124468803406, + "learning_rate": 4.8905654895011325e-06, + "loss": 0.6772, + "step": 831 + }, + { + "epoch": 0.3185908481715489, + "grad_norm": 0.27484792470932007, + "learning_rate": 4.89026808556323e-06, + "loss": 0.6824, + "step": 832 + }, + { + "epoch": 0.3189737698640628, + "grad_norm": 0.26560086011886597, + "learning_rate": 4.889970287122307e-06, + "loss": 0.6896, + "step": 833 + }, + { + "epoch": 0.3193566915565767, + "grad_norm": 0.2708161771297455, + "learning_rate": 4.889672094227515e-06, + "loss": 0.6785, + "step": 834 + }, + { + "epoch": 0.31973961324909056, + "grad_norm": 0.2682739794254303, + "learning_rate": 4.88937350692807e-06, + "loss": 0.7022, + "step": 835 + }, + { + "epoch": 0.3201225349416044, + "grad_norm": 0.26009342074394226, + "learning_rate": 4.8890745252732504e-06, + "loss": 0.6895, + "step": 836 + }, + { + "epoch": 0.32050545663411834, + "grad_norm": 0.2639565169811249, + "learning_rate": 4.888775149312403e-06, + "loss": 0.6939, + "step": 837 + }, + { + "epoch": 0.3208883783266322, + "grad_norm": 0.2597087025642395, + "learning_rate": 4.888475379094938e-06, + "loss": 0.6914, + "step": 838 + }, + { + "epoch": 0.3212713000191461, + "grad_norm": 0.2735402584075928, + "learning_rate": 4.888175214670331e-06, + "loss": 0.6842, + "step": 839 + }, + { + "epoch": 0.32165422171165997, + "grad_norm": 0.24896885454654694, + "learning_rate": 4.887874656088124e-06, + "loss": 0.6972, + "step": 840 + }, + { + "epoch": 0.32203714340417383, + "grad_norm": 0.25625696778297424, + "learning_rate": 4.887573703397921e-06, + "loss": 0.6811, + "step": 841 + }, + { + "epoch": 0.32242006509668775, + "grad_norm": 0.2628985047340393, + "learning_rate": 4.887272356649395e-06, + "loss": 0.6847, + "step": 842 + }, + { + "epoch": 0.3228029867892016, + "grad_norm": 0.26313072443008423, + "learning_rate": 4.88697061589228e-06, + "loss": 0.6871, + "step": 843 + }, + { + "epoch": 0.32318590848171547, + "grad_norm": 0.27291008830070496, + "learning_rate": 4.886668481176377e-06, + "loss": 0.6791, + "step": 844 + }, + { + "epoch": 0.3235688301742294, + "grad_norm": 0.26743587851524353, + "learning_rate": 4.886365952551555e-06, + "loss": 0.6861, + "step": 845 + }, + { + "epoch": 0.32395175186674324, + "grad_norm": 0.25605612993240356, + "learning_rate": 4.886063030067741e-06, + "loss": 0.6695, + "step": 846 + }, + { + "epoch": 0.32433467355925716, + "grad_norm": 0.2629465162754059, + "learning_rate": 4.885759713774933e-06, + "loss": 0.6802, + "step": 847 + }, + { + "epoch": 0.324717595251771, + "grad_norm": 0.26355117559432983, + "learning_rate": 4.885456003723192e-06, + "loss": 0.657, + "step": 848 + }, + { + "epoch": 0.3251005169442849, + "grad_norm": 0.258645623922348, + "learning_rate": 4.885151899962644e-06, + "loss": 0.6602, + "step": 849 + }, + { + "epoch": 0.3254834386367988, + "grad_norm": 0.2570325434207916, + "learning_rate": 4.884847402543479e-06, + "loss": 0.6907, + "step": 850 + }, + { + "epoch": 0.32586636032931265, + "grad_norm": 0.2724635899066925, + "learning_rate": 4.884542511515954e-06, + "loss": 0.6648, + "step": 851 + }, + { + "epoch": 0.3262492820218265, + "grad_norm": 0.2683566212654114, + "learning_rate": 4.884237226930389e-06, + "loss": 0.6936, + "step": 852 + }, + { + "epoch": 0.32663220371434043, + "grad_norm": 0.27550438046455383, + "learning_rate": 4.883931548837172e-06, + "loss": 0.6915, + "step": 853 + }, + { + "epoch": 0.3270151254068543, + "grad_norm": 0.26245197653770447, + "learning_rate": 4.883625477286751e-06, + "loss": 0.6805, + "step": 854 + }, + { + "epoch": 0.3273980470993682, + "grad_norm": 0.26423540711402893, + "learning_rate": 4.883319012329643e-06, + "loss": 0.6811, + "step": 855 + }, + { + "epoch": 0.32778096879188207, + "grad_norm": 0.28096291422843933, + "learning_rate": 4.883012154016428e-06, + "loss": 0.6846, + "step": 856 + }, + { + "epoch": 0.3281638904843959, + "grad_norm": 0.2643588185310364, + "learning_rate": 4.882704902397752e-06, + "loss": 0.7193, + "step": 857 + }, + { + "epoch": 0.32854681217690984, + "grad_norm": 0.2602258026599884, + "learning_rate": 4.882397257524326e-06, + "loss": 0.6753, + "step": 858 + }, + { + "epoch": 0.3289297338694237, + "grad_norm": 0.27310651540756226, + "learning_rate": 4.882089219446925e-06, + "loss": 0.6663, + "step": 859 + }, + { + "epoch": 0.32931265556193756, + "grad_norm": 0.26811903715133667, + "learning_rate": 4.8817807882163904e-06, + "loss": 0.6825, + "step": 860 + }, + { + "epoch": 0.3296955772544515, + "grad_norm": 0.266361266374588, + "learning_rate": 4.881471963883625e-06, + "loss": 0.694, + "step": 861 + }, + { + "epoch": 0.33007849894696534, + "grad_norm": 0.27039119601249695, + "learning_rate": 4.8811627464996005e-06, + "loss": 0.6773, + "step": 862 + }, + { + "epoch": 0.33046142063947925, + "grad_norm": 0.28884178400039673, + "learning_rate": 4.880853136115352e-06, + "loss": 0.6791, + "step": 863 + }, + { + "epoch": 0.3308443423319931, + "grad_norm": 0.2610810399055481, + "learning_rate": 4.8805431327819786e-06, + "loss": 0.6783, + "step": 864 + }, + { + "epoch": 0.331227264024507, + "grad_norm": 0.27807381749153137, + "learning_rate": 4.8802327365506456e-06, + "loss": 0.6904, + "step": 865 + }, + { + "epoch": 0.3316101857170209, + "grad_norm": 0.2709735333919525, + "learning_rate": 4.8799219474725814e-06, + "loss": 0.6842, + "step": 866 + }, + { + "epoch": 0.33199310740953475, + "grad_norm": 0.26760631799697876, + "learning_rate": 4.879610765599081e-06, + "loss": 0.6987, + "step": 867 + }, + { + "epoch": 0.3323760291020486, + "grad_norm": 0.26859745383262634, + "learning_rate": 4.8792991909815035e-06, + "loss": 0.6839, + "step": 868 + }, + { + "epoch": 0.3327589507945625, + "grad_norm": 0.27805396914482117, + "learning_rate": 4.878987223671274e-06, + "loss": 0.6883, + "step": 869 + }, + { + "epoch": 0.3331418724870764, + "grad_norm": 0.2685054540634155, + "learning_rate": 4.878674863719879e-06, + "loss": 0.6723, + "step": 870 + }, + { + "epoch": 0.3335247941795903, + "grad_norm": 0.2590447664260864, + "learning_rate": 4.878362111178875e-06, + "loss": 0.6792, + "step": 871 + }, + { + "epoch": 0.33390771587210416, + "grad_norm": 0.2719431519508362, + "learning_rate": 4.8780489660998776e-06, + "loss": 0.6908, + "step": 872 + }, + { + "epoch": 0.334290637564618, + "grad_norm": 0.27068865299224854, + "learning_rate": 4.877735428534572e-06, + "loss": 0.6869, + "step": 873 + }, + { + "epoch": 0.33467355925713194, + "grad_norm": 0.263995885848999, + "learning_rate": 4.877421498534705e-06, + "loss": 0.6813, + "step": 874 + }, + { + "epoch": 0.3350564809496458, + "grad_norm": 0.27139508724212646, + "learning_rate": 4.877107176152089e-06, + "loss": 0.6733, + "step": 875 + }, + { + "epoch": 0.33543940264215966, + "grad_norm": 0.2608059346675873, + "learning_rate": 4.876792461438604e-06, + "loss": 0.6744, + "step": 876 + }, + { + "epoch": 0.33582232433467357, + "grad_norm": 0.26705706119537354, + "learning_rate": 4.8764773544461895e-06, + "loss": 0.6919, + "step": 877 + }, + { + "epoch": 0.33620524602718743, + "grad_norm": 0.2600170373916626, + "learning_rate": 4.8761618552268534e-06, + "loss": 0.6843, + "step": 878 + }, + { + "epoch": 0.33658816771970135, + "grad_norm": 0.2667454481124878, + "learning_rate": 4.875845963832667e-06, + "loss": 0.6843, + "step": 879 + }, + { + "epoch": 0.3369710894122152, + "grad_norm": 0.2666393220424652, + "learning_rate": 4.875529680315768e-06, + "loss": 0.6843, + "step": 880 + }, + { + "epoch": 0.33735401110472907, + "grad_norm": 0.2725430727005005, + "learning_rate": 4.8752130047283565e-06, + "loss": 0.678, + "step": 881 + }, + { + "epoch": 0.337736932797243, + "grad_norm": 0.2890147864818573, + "learning_rate": 4.874895937122699e-06, + "loss": 0.696, + "step": 882 + }, + { + "epoch": 0.33811985448975684, + "grad_norm": 0.2604501247406006, + "learning_rate": 4.874578477551126e-06, + "loss": 0.6901, + "step": 883 + }, + { + "epoch": 0.3385027761822707, + "grad_norm": 0.26774969696998596, + "learning_rate": 4.8742606260660315e-06, + "loss": 0.6895, + "step": 884 + }, + { + "epoch": 0.3388856978747846, + "grad_norm": 0.2798639237880707, + "learning_rate": 4.873942382719878e-06, + "loss": 0.6654, + "step": 885 + }, + { + "epoch": 0.3392686195672985, + "grad_norm": 0.27278244495391846, + "learning_rate": 4.873623747565188e-06, + "loss": 0.6967, + "step": 886 + }, + { + "epoch": 0.3396515412598124, + "grad_norm": 0.25627878308296204, + "learning_rate": 4.873304720654551e-06, + "loss": 0.6756, + "step": 887 + }, + { + "epoch": 0.34003446295232626, + "grad_norm": 0.26743650436401367, + "learning_rate": 4.872985302040622e-06, + "loss": 0.7111, + "step": 888 + }, + { + "epoch": 0.3404173846448401, + "grad_norm": 0.27230358123779297, + "learning_rate": 4.872665491776119e-06, + "loss": 0.6911, + "step": 889 + }, + { + "epoch": 0.34080030633735403, + "grad_norm": 0.2792832553386688, + "learning_rate": 4.872345289913826e-06, + "loss": 0.6947, + "step": 890 + }, + { + "epoch": 0.3411832280298679, + "grad_norm": 0.27925628423690796, + "learning_rate": 4.872024696506588e-06, + "loss": 0.6942, + "step": 891 + }, + { + "epoch": 0.34156614972238175, + "grad_norm": 0.2760670781135559, + "learning_rate": 4.871703711607322e-06, + "loss": 0.6865, + "step": 892 + }, + { + "epoch": 0.34194907141489567, + "grad_norm": 0.27710941433906555, + "learning_rate": 4.871382335269001e-06, + "loss": 0.6918, + "step": 893 + }, + { + "epoch": 0.3423319931074095, + "grad_norm": 0.2548944056034088, + "learning_rate": 4.871060567544671e-06, + "loss": 0.6917, + "step": 894 + }, + { + "epoch": 0.34271491479992344, + "grad_norm": 0.26869043707847595, + "learning_rate": 4.870738408487433e-06, + "loss": 0.699, + "step": 895 + }, + { + "epoch": 0.3430978364924373, + "grad_norm": 0.2725992500782013, + "learning_rate": 4.870415858150462e-06, + "loss": 0.6977, + "step": 896 + }, + { + "epoch": 0.34348075818495116, + "grad_norm": 0.2812182605266571, + "learning_rate": 4.8700929165869925e-06, + "loss": 0.6617, + "step": 897 + }, + { + "epoch": 0.3438636798774651, + "grad_norm": 0.257763534784317, + "learning_rate": 4.869769583850324e-06, + "loss": 0.7007, + "step": 898 + }, + { + "epoch": 0.34424660156997894, + "grad_norm": 0.26617327332496643, + "learning_rate": 4.869445859993822e-06, + "loss": 0.6779, + "step": 899 + }, + { + "epoch": 0.3446295232624928, + "grad_norm": 0.2741871178150177, + "learning_rate": 4.8691217450709145e-06, + "loss": 0.6832, + "step": 900 + }, + { + "epoch": 0.3450124449550067, + "grad_norm": 0.2675487697124481, + "learning_rate": 4.8687972391350955e-06, + "loss": 0.6533, + "step": 901 + }, + { + "epoch": 0.3453953666475206, + "grad_norm": 0.2638150453567505, + "learning_rate": 4.868472342239924e-06, + "loss": 0.6685, + "step": 902 + }, + { + "epoch": 0.3457782883400345, + "grad_norm": 0.29722779989242554, + "learning_rate": 4.868147054439022e-06, + "loss": 0.687, + "step": 903 + }, + { + "epoch": 0.34616121003254835, + "grad_norm": 0.2719815969467163, + "learning_rate": 4.8678213757860774e-06, + "loss": 0.6888, + "step": 904 + }, + { + "epoch": 0.3465441317250622, + "grad_norm": 0.27905941009521484, + "learning_rate": 4.8674953063348415e-06, + "loss": 0.7021, + "step": 905 + }, + { + "epoch": 0.3469270534175761, + "grad_norm": 0.28070658445358276, + "learning_rate": 4.867168846139131e-06, + "loss": 0.6817, + "step": 906 + }, + { + "epoch": 0.34730997511009, + "grad_norm": 0.2584090828895569, + "learning_rate": 4.866841995252826e-06, + "loss": 0.6798, + "step": 907 + }, + { + "epoch": 0.34769289680260385, + "grad_norm": 0.2600197196006775, + "learning_rate": 4.8665147537298726e-06, + "loss": 0.6869, + "step": 908 + }, + { + "epoch": 0.34807581849511776, + "grad_norm": 0.2670934796333313, + "learning_rate": 4.8661871216242805e-06, + "loss": 0.6887, + "step": 909 + }, + { + "epoch": 0.3484587401876316, + "grad_norm": 0.2869112193584442, + "learning_rate": 4.865859098990123e-06, + "loss": 0.675, + "step": 910 + }, + { + "epoch": 0.34884166188014554, + "grad_norm": 0.300727903842926, + "learning_rate": 4.865530685881541e-06, + "loss": 0.6701, + "step": 911 + }, + { + "epoch": 0.3492245835726594, + "grad_norm": 0.2582904100418091, + "learning_rate": 4.8652018823527345e-06, + "loss": 0.6827, + "step": 912 + }, + { + "epoch": 0.34960750526517326, + "grad_norm": 0.27412885427474976, + "learning_rate": 4.864872688457972e-06, + "loss": 0.6772, + "step": 913 + }, + { + "epoch": 0.3499904269576872, + "grad_norm": 0.323688268661499, + "learning_rate": 4.864543104251587e-06, + "loss": 0.6837, + "step": 914 + }, + { + "epoch": 0.35037334865020103, + "grad_norm": 0.27548015117645264, + "learning_rate": 4.864213129787975e-06, + "loss": 0.6885, + "step": 915 + }, + { + "epoch": 0.3507562703427149, + "grad_norm": 0.2752622067928314, + "learning_rate": 4.863882765121596e-06, + "loss": 0.7021, + "step": 916 + }, + { + "epoch": 0.3511391920352288, + "grad_norm": 0.2829137444496155, + "learning_rate": 4.863552010306976e-06, + "loss": 0.7023, + "step": 917 + }, + { + "epoch": 0.35152211372774267, + "grad_norm": 0.30739355087280273, + "learning_rate": 4.863220865398705e-06, + "loss": 0.6655, + "step": 918 + }, + { + "epoch": 0.3519050354202566, + "grad_norm": 0.26492998003959656, + "learning_rate": 4.862889330451437e-06, + "loss": 0.6986, + "step": 919 + }, + { + "epoch": 0.35228795711277044, + "grad_norm": 0.3324214518070221, + "learning_rate": 4.862557405519889e-06, + "loss": 0.6975, + "step": 920 + }, + { + "epoch": 0.3526708788052843, + "grad_norm": 0.3313480615615845, + "learning_rate": 4.862225090658844e-06, + "loss": 0.6782, + "step": 921 + }, + { + "epoch": 0.3530538004977982, + "grad_norm": 0.26132631301879883, + "learning_rate": 4.861892385923151e-06, + "loss": 0.6698, + "step": 922 + }, + { + "epoch": 0.3534367221903121, + "grad_norm": 0.2920131981372833, + "learning_rate": 4.86155929136772e-06, + "loss": 0.6999, + "step": 923 + }, + { + "epoch": 0.35381964388282594, + "grad_norm": 0.30363237857818604, + "learning_rate": 4.861225807047526e-06, + "loss": 0.6962, + "step": 924 + }, + { + "epoch": 0.35420256557533986, + "grad_norm": 0.28039583563804626, + "learning_rate": 4.86089193301761e-06, + "loss": 0.6997, + "step": 925 + }, + { + "epoch": 0.3545854872678537, + "grad_norm": 0.2662922739982605, + "learning_rate": 4.860557669333076e-06, + "loss": 0.6465, + "step": 926 + }, + { + "epoch": 0.35496840896036763, + "grad_norm": 0.26688340306282043, + "learning_rate": 4.860223016049094e-06, + "loss": 0.6846, + "step": 927 + }, + { + "epoch": 0.3553513306528815, + "grad_norm": 0.2974063754081726, + "learning_rate": 4.859887973220896e-06, + "loss": 0.6984, + "step": 928 + }, + { + "epoch": 0.35573425234539535, + "grad_norm": 0.2696240246295929, + "learning_rate": 4.859552540903779e-06, + "loss": 0.6646, + "step": 929 + }, + { + "epoch": 0.35611717403790927, + "grad_norm": 0.26638904213905334, + "learning_rate": 4.859216719153105e-06, + "loss": 0.6651, + "step": 930 + }, + { + "epoch": 0.3565000957304231, + "grad_norm": 0.2707373797893524, + "learning_rate": 4.8588805080243e-06, + "loss": 0.7, + "step": 931 + }, + { + "epoch": 0.356883017422937, + "grad_norm": 0.2757275700569153, + "learning_rate": 4.858543907572854e-06, + "loss": 0.6929, + "step": 932 + }, + { + "epoch": 0.3572659391154509, + "grad_norm": 0.26656290888786316, + "learning_rate": 4.858206917854322e-06, + "loss": 0.6706, + "step": 933 + }, + { + "epoch": 0.35764886080796476, + "grad_norm": 0.2593183219432831, + "learning_rate": 4.85786953892432e-06, + "loss": 0.6753, + "step": 934 + }, + { + "epoch": 0.3580317825004787, + "grad_norm": 0.26421022415161133, + "learning_rate": 4.857531770838534e-06, + "loss": 0.6742, + "step": 935 + }, + { + "epoch": 0.35841470419299254, + "grad_norm": 0.2740086317062378, + "learning_rate": 4.857193613652711e-06, + "loss": 0.6766, + "step": 936 + }, + { + "epoch": 0.3587976258855064, + "grad_norm": 0.270923912525177, + "learning_rate": 4.856855067422661e-06, + "loss": 0.6986, + "step": 937 + }, + { + "epoch": 0.3591805475780203, + "grad_norm": 0.2739553153514862, + "learning_rate": 4.856516132204258e-06, + "loss": 0.6595, + "step": 938 + }, + { + "epoch": 0.3595634692705342, + "grad_norm": 0.2649962902069092, + "learning_rate": 4.856176808053446e-06, + "loss": 0.696, + "step": 939 + }, + { + "epoch": 0.35994639096304804, + "grad_norm": 0.27668675780296326, + "learning_rate": 4.8558370950262256e-06, + "loss": 0.6849, + "step": 940 + }, + { + "epoch": 0.36032931265556195, + "grad_norm": 0.2732677757740021, + "learning_rate": 4.855496993178666e-06, + "loss": 0.6809, + "step": 941 + }, + { + "epoch": 0.3607122343480758, + "grad_norm": 0.27247822284698486, + "learning_rate": 4.855156502566899e-06, + "loss": 0.6912, + "step": 942 + }, + { + "epoch": 0.3610951560405897, + "grad_norm": 0.2725856900215149, + "learning_rate": 4.854815623247122e-06, + "loss": 0.6715, + "step": 943 + }, + { + "epoch": 0.3614780777331036, + "grad_norm": 0.25335705280303955, + "learning_rate": 4.854474355275595e-06, + "loss": 0.6646, + "step": 944 + }, + { + "epoch": 0.36186099942561745, + "grad_norm": 0.2773241400718689, + "learning_rate": 4.854132698708642e-06, + "loss": 0.6806, + "step": 945 + }, + { + "epoch": 0.36224392111813136, + "grad_norm": 0.26226091384887695, + "learning_rate": 4.853790653602653e-06, + "loss": 0.6695, + "step": 946 + }, + { + "epoch": 0.3626268428106452, + "grad_norm": 0.27693572640419006, + "learning_rate": 4.85344822001408e-06, + "loss": 0.6796, + "step": 947 + }, + { + "epoch": 0.3630097645031591, + "grad_norm": 0.26777103543281555, + "learning_rate": 4.853105397999442e-06, + "loss": 0.6876, + "step": 948 + }, + { + "epoch": 0.363392686195673, + "grad_norm": 0.2739417850971222, + "learning_rate": 4.852762187615318e-06, + "loss": 0.6838, + "step": 949 + }, + { + "epoch": 0.36377560788818686, + "grad_norm": 0.2750835418701172, + "learning_rate": 4.852418588918355e-06, + "loss": 0.6644, + "step": 950 + }, + { + "epoch": 0.3641585295807008, + "grad_norm": 0.2623859941959381, + "learning_rate": 4.852074601965261e-06, + "loss": 0.6554, + "step": 951 + }, + { + "epoch": 0.36454145127321463, + "grad_norm": 0.26994043588638306, + "learning_rate": 4.851730226812811e-06, + "loss": 0.6794, + "step": 952 + }, + { + "epoch": 0.3649243729657285, + "grad_norm": 0.25890594720840454, + "learning_rate": 4.851385463517841e-06, + "loss": 0.6914, + "step": 953 + }, + { + "epoch": 0.3653072946582424, + "grad_norm": 0.26775312423706055, + "learning_rate": 4.851040312137254e-06, + "loss": 0.6853, + "step": 954 + }, + { + "epoch": 0.36569021635075627, + "grad_norm": 0.26391908526420593, + "learning_rate": 4.850694772728015e-06, + "loss": 0.679, + "step": 955 + }, + { + "epoch": 0.36607313804327013, + "grad_norm": 0.25619781017303467, + "learning_rate": 4.850348845347155e-06, + "loss": 0.6732, + "step": 956 + }, + { + "epoch": 0.36645605973578405, + "grad_norm": 0.26470351219177246, + "learning_rate": 4.850002530051765e-06, + "loss": 0.6927, + "step": 957 + }, + { + "epoch": 0.3668389814282979, + "grad_norm": 0.2638048529624939, + "learning_rate": 4.849655826899005e-06, + "loss": 0.6904, + "step": 958 + }, + { + "epoch": 0.3672219031208118, + "grad_norm": 0.25979623198509216, + "learning_rate": 4.849308735946096e-06, + "loss": 0.6736, + "step": 959 + }, + { + "epoch": 0.3676048248133257, + "grad_norm": 0.2671094834804535, + "learning_rate": 4.8489612572503245e-06, + "loss": 0.662, + "step": 960 + }, + { + "epoch": 0.36798774650583954, + "grad_norm": 0.26025474071502686, + "learning_rate": 4.84861339086904e-06, + "loss": 0.6571, + "step": 961 + }, + { + "epoch": 0.36837066819835346, + "grad_norm": 0.26459765434265137, + "learning_rate": 4.848265136859656e-06, + "loss": 0.6853, + "step": 962 + }, + { + "epoch": 0.3687535898908673, + "grad_norm": 0.2781910300254822, + "learning_rate": 4.84791649527965e-06, + "loss": 0.6968, + "step": 963 + }, + { + "epoch": 0.3691365115833812, + "grad_norm": 0.2741437256336212, + "learning_rate": 4.8475674661865636e-06, + "loss": 0.6897, + "step": 964 + }, + { + "epoch": 0.3695194332758951, + "grad_norm": 0.2687261402606964, + "learning_rate": 4.8472180496380035e-06, + "loss": 0.6884, + "step": 965 + }, + { + "epoch": 0.36990235496840895, + "grad_norm": 0.2683657705783844, + "learning_rate": 4.846868245691639e-06, + "loss": 0.6919, + "step": 966 + }, + { + "epoch": 0.37028527666092287, + "grad_norm": 0.28037914633750916, + "learning_rate": 4.846518054405203e-06, + "loss": 0.6939, + "step": 967 + }, + { + "epoch": 0.37066819835343673, + "grad_norm": 0.2678395211696625, + "learning_rate": 4.846167475836494e-06, + "loss": 0.6784, + "step": 968 + }, + { + "epoch": 0.3710511200459506, + "grad_norm": 0.263564795255661, + "learning_rate": 4.8458165100433725e-06, + "loss": 0.6776, + "step": 969 + }, + { + "epoch": 0.3714340417384645, + "grad_norm": 0.26803597807884216, + "learning_rate": 4.845465157083765e-06, + "loss": 0.6867, + "step": 970 + }, + { + "epoch": 0.37181696343097836, + "grad_norm": 0.27506449818611145, + "learning_rate": 4.8451134170156585e-06, + "loss": 0.6862, + "step": 971 + }, + { + "epoch": 0.3721998851234922, + "grad_norm": 0.26313772797584534, + "learning_rate": 4.844761289897109e-06, + "loss": 0.6739, + "step": 972 + }, + { + "epoch": 0.37258280681600614, + "grad_norm": 0.26042506098747253, + "learning_rate": 4.844408775786231e-06, + "loss": 0.686, + "step": 973 + }, + { + "epoch": 0.37296572850852, + "grad_norm": 0.2822672426700592, + "learning_rate": 4.844055874741208e-06, + "loss": 0.6853, + "step": 974 + }, + { + "epoch": 0.3733486502010339, + "grad_norm": 0.26979514956474304, + "learning_rate": 4.843702586820282e-06, + "loss": 0.6936, + "step": 975 + }, + { + "epoch": 0.3737315718935478, + "grad_norm": 0.27148738503456116, + "learning_rate": 4.843348912081763e-06, + "loss": 0.6817, + "step": 976 + }, + { + "epoch": 0.37411449358606164, + "grad_norm": 0.2711876332759857, + "learning_rate": 4.842994850584023e-06, + "loss": 0.6781, + "step": 977 + }, + { + "epoch": 0.37449741527857555, + "grad_norm": 0.25822556018829346, + "learning_rate": 4.842640402385499e-06, + "loss": 0.6595, + "step": 978 + }, + { + "epoch": 0.3748803369710894, + "grad_norm": 0.27487432956695557, + "learning_rate": 4.84228556754469e-06, + "loss": 0.6633, + "step": 979 + }, + { + "epoch": 0.37526325866360327, + "grad_norm": 0.27101749181747437, + "learning_rate": 4.841930346120161e-06, + "loss": 0.6827, + "step": 980 + }, + { + "epoch": 0.3756461803561172, + "grad_norm": 0.26800671219825745, + "learning_rate": 4.841574738170538e-06, + "loss": 0.6703, + "step": 981 + }, + { + "epoch": 0.37602910204863105, + "grad_norm": 0.2811592221260071, + "learning_rate": 4.841218743754515e-06, + "loss": 0.7082, + "step": 982 + }, + { + "epoch": 0.37641202374114496, + "grad_norm": 0.2691783010959625, + "learning_rate": 4.840862362930845e-06, + "loss": 0.7007, + "step": 983 + }, + { + "epoch": 0.3767949454336588, + "grad_norm": 0.2679787576198578, + "learning_rate": 4.840505595758348e-06, + "loss": 0.6866, + "step": 984 + }, + { + "epoch": 0.3771778671261727, + "grad_norm": 0.2782638669013977, + "learning_rate": 4.840148442295907e-06, + "loss": 0.6829, + "step": 985 + }, + { + "epoch": 0.3775607888186866, + "grad_norm": 0.25745078921318054, + "learning_rate": 4.839790902602469e-06, + "loss": 0.6601, + "step": 986 + }, + { + "epoch": 0.37794371051120046, + "grad_norm": 0.2637532353401184, + "learning_rate": 4.839432976737043e-06, + "loss": 0.6565, + "step": 987 + }, + { + "epoch": 0.3783266322037143, + "grad_norm": 0.269976943731308, + "learning_rate": 4.839074664758705e-06, + "loss": 0.6715, + "step": 988 + }, + { + "epoch": 0.37870955389622823, + "grad_norm": 0.2653595209121704, + "learning_rate": 4.83871596672659e-06, + "loss": 0.6675, + "step": 989 + }, + { + "epoch": 0.3790924755887421, + "grad_norm": 0.26211804151535034, + "learning_rate": 4.838356882699902e-06, + "loss": 0.6545, + "step": 990 + }, + { + "epoch": 0.379475397281256, + "grad_norm": 0.2690889537334442, + "learning_rate": 4.837997412737905e-06, + "loss": 0.6998, + "step": 991 + }, + { + "epoch": 0.37985831897376987, + "grad_norm": 0.26827001571655273, + "learning_rate": 4.837637556899928e-06, + "loss": 0.6807, + "step": 992 + }, + { + "epoch": 0.38024124066628373, + "grad_norm": 0.2662839889526367, + "learning_rate": 4.837277315245364e-06, + "loss": 0.6842, + "step": 993 + }, + { + "epoch": 0.38062416235879765, + "grad_norm": 0.26887455582618713, + "learning_rate": 4.836916687833668e-06, + "loss": 0.6807, + "step": 994 + }, + { + "epoch": 0.3810070840513115, + "grad_norm": 0.2758978009223938, + "learning_rate": 4.836555674724362e-06, + "loss": 0.6982, + "step": 995 + }, + { + "epoch": 0.38139000574382537, + "grad_norm": 0.2752774953842163, + "learning_rate": 4.836194275977027e-06, + "loss": 0.6831, + "step": 996 + }, + { + "epoch": 0.3817729274363393, + "grad_norm": 0.3027510344982147, + "learning_rate": 4.835832491651314e-06, + "loss": 0.7033, + "step": 997 + }, + { + "epoch": 0.38215584912885314, + "grad_norm": 0.26609593629837036, + "learning_rate": 4.83547032180693e-06, + "loss": 0.6718, + "step": 998 + }, + { + "epoch": 0.38253877082136706, + "grad_norm": 0.2735961675643921, + "learning_rate": 4.835107766503652e-06, + "loss": 0.6791, + "step": 999 + }, + { + "epoch": 0.3829216925138809, + "grad_norm": 0.26820945739746094, + "learning_rate": 4.834744825801318e-06, + "loss": 0.6805, + "step": 1000 + }, + { + "epoch": 0.3833046142063948, + "grad_norm": 0.2665892541408539, + "learning_rate": 4.834381499759827e-06, + "loss": 0.6672, + "step": 1001 + }, + { + "epoch": 0.3836875358989087, + "grad_norm": 0.2774675786495209, + "learning_rate": 4.834017788439148e-06, + "loss": 0.6774, + "step": 1002 + }, + { + "epoch": 0.38407045759142255, + "grad_norm": 0.27913185954093933, + "learning_rate": 4.833653691899308e-06, + "loss": 0.6656, + "step": 1003 + }, + { + "epoch": 0.3844533792839364, + "grad_norm": 0.2730454206466675, + "learning_rate": 4.833289210200399e-06, + "loss": 0.691, + "step": 1004 + }, + { + "epoch": 0.38483630097645033, + "grad_norm": 0.27499866485595703, + "learning_rate": 4.8329243434025785e-06, + "loss": 0.688, + "step": 1005 + }, + { + "epoch": 0.3852192226689642, + "grad_norm": 0.26650702953338623, + "learning_rate": 4.832559091566066e-06, + "loss": 0.6836, + "step": 1006 + }, + { + "epoch": 0.3856021443614781, + "grad_norm": 0.2753099501132965, + "learning_rate": 4.832193454751143e-06, + "loss": 0.6721, + "step": 1007 + }, + { + "epoch": 0.38598506605399197, + "grad_norm": 0.275810569524765, + "learning_rate": 4.831827433018158e-06, + "loss": 0.6722, + "step": 1008 + }, + { + "epoch": 0.3863679877465058, + "grad_norm": 0.27451977133750916, + "learning_rate": 4.8314610264275205e-06, + "loss": 0.6783, + "step": 1009 + }, + { + "epoch": 0.38675090943901974, + "grad_norm": 0.26779070496559143, + "learning_rate": 4.831094235039704e-06, + "loss": 0.6762, + "step": 1010 + }, + { + "epoch": 0.3871338311315336, + "grad_norm": 0.2588435113430023, + "learning_rate": 4.830727058915247e-06, + "loss": 0.6666, + "step": 1011 + }, + { + "epoch": 0.38751675282404746, + "grad_norm": 0.2793857157230377, + "learning_rate": 4.830359498114749e-06, + "loss": 0.6658, + "step": 1012 + }, + { + "epoch": 0.3878996745165614, + "grad_norm": 0.2771088182926178, + "learning_rate": 4.829991552698875e-06, + "loss": 0.674, + "step": 1013 + }, + { + "epoch": 0.38828259620907524, + "grad_norm": 0.2702658772468567, + "learning_rate": 4.829623222728352e-06, + "loss": 0.6715, + "step": 1014 + }, + { + "epoch": 0.38866551790158915, + "grad_norm": 0.28128373622894287, + "learning_rate": 4.829254508263971e-06, + "loss": 0.6675, + "step": 1015 + }, + { + "epoch": 0.389048439594103, + "grad_norm": 0.26931461691856384, + "learning_rate": 4.828885409366589e-06, + "loss": 0.6567, + "step": 1016 + }, + { + "epoch": 0.3894313612866169, + "grad_norm": 0.26983270049095154, + "learning_rate": 4.828515926097121e-06, + "loss": 0.6727, + "step": 1017 + }, + { + "epoch": 0.3898142829791308, + "grad_norm": 0.2659534215927124, + "learning_rate": 4.82814605851655e-06, + "loss": 0.6587, + "step": 1018 + }, + { + "epoch": 0.39019720467164465, + "grad_norm": 0.26131471991539, + "learning_rate": 4.827775806685922e-06, + "loss": 0.6813, + "step": 1019 + }, + { + "epoch": 0.3905801263641585, + "grad_norm": 0.28075239062309265, + "learning_rate": 4.8274051706663435e-06, + "loss": 0.672, + "step": 1020 + }, + { + "epoch": 0.3909630480566724, + "grad_norm": 0.2587306499481201, + "learning_rate": 4.827034150518987e-06, + "loss": 0.6643, + "step": 1021 + }, + { + "epoch": 0.3913459697491863, + "grad_norm": 0.2753465473651886, + "learning_rate": 4.826662746305089e-06, + "loss": 0.656, + "step": 1022 + }, + { + "epoch": 0.3917288914417002, + "grad_norm": 0.27708113193511963, + "learning_rate": 4.826290958085945e-06, + "loss": 0.6798, + "step": 1023 + }, + { + "epoch": 0.39211181313421406, + "grad_norm": 0.2657153308391571, + "learning_rate": 4.825918785922921e-06, + "loss": 0.6881, + "step": 1024 + }, + { + "epoch": 0.3924947348267279, + "grad_norm": 0.26796260476112366, + "learning_rate": 4.825546229877439e-06, + "loss": 0.6856, + "step": 1025 + }, + { + "epoch": 0.39287765651924184, + "grad_norm": 0.2698305547237396, + "learning_rate": 4.82517329001099e-06, + "loss": 0.6748, + "step": 1026 + }, + { + "epoch": 0.3932605782117557, + "grad_norm": 0.2884180545806885, + "learning_rate": 4.824799966385124e-06, + "loss": 0.683, + "step": 1027 + }, + { + "epoch": 0.39364349990426956, + "grad_norm": 0.269398033618927, + "learning_rate": 4.824426259061458e-06, + "loss": 0.6904, + "step": 1028 + }, + { + "epoch": 0.39402642159678347, + "grad_norm": 0.2623305320739746, + "learning_rate": 4.82405216810167e-06, + "loss": 0.6731, + "step": 1029 + }, + { + "epoch": 0.39440934328929733, + "grad_norm": 0.2715912461280823, + "learning_rate": 4.823677693567502e-06, + "loss": 0.6661, + "step": 1030 + }, + { + "epoch": 0.39479226498181125, + "grad_norm": 0.27043530344963074, + "learning_rate": 4.82330283552076e-06, + "loss": 0.6994, + "step": 1031 + }, + { + "epoch": 0.3951751866743251, + "grad_norm": 0.27699190378189087, + "learning_rate": 4.822927594023312e-06, + "loss": 0.6645, + "step": 1032 + }, + { + "epoch": 0.39555810836683897, + "grad_norm": 0.27182725071907043, + "learning_rate": 4.8225519691370905e-06, + "loss": 0.679, + "step": 1033 + }, + { + "epoch": 0.3959410300593529, + "grad_norm": 0.27347299456596375, + "learning_rate": 4.8221759609240894e-06, + "loss": 0.678, + "step": 1034 + }, + { + "epoch": 0.39632395175186674, + "grad_norm": 0.2720860242843628, + "learning_rate": 4.821799569446368e-06, + "loss": 0.7019, + "step": 1035 + }, + { + "epoch": 0.3967068734443806, + "grad_norm": 0.27360793948173523, + "learning_rate": 4.821422794766049e-06, + "loss": 0.6719, + "step": 1036 + }, + { + "epoch": 0.3970897951368945, + "grad_norm": 0.2907519042491913, + "learning_rate": 4.821045636945316e-06, + "loss": 0.6768, + "step": 1037 + }, + { + "epoch": 0.3974727168294084, + "grad_norm": 0.2736169397830963, + "learning_rate": 4.820668096046418e-06, + "loss": 0.6717, + "step": 1038 + }, + { + "epoch": 0.3978556385219223, + "grad_norm": 0.270683616399765, + "learning_rate": 4.820290172131667e-06, + "loss": 0.6808, + "step": 1039 + }, + { + "epoch": 0.39823856021443615, + "grad_norm": 0.2760835587978363, + "learning_rate": 4.819911865263434e-06, + "loss": 0.6861, + "step": 1040 + }, + { + "epoch": 0.39862148190695, + "grad_norm": 0.2689175605773926, + "learning_rate": 4.8195331755041616e-06, + "loss": 0.6642, + "step": 1041 + }, + { + "epoch": 0.39900440359946393, + "grad_norm": 0.2810250222682953, + "learning_rate": 4.819154102916349e-06, + "loss": 0.6601, + "step": 1042 + }, + { + "epoch": 0.3993873252919778, + "grad_norm": 0.2765415608882904, + "learning_rate": 4.81877464756256e-06, + "loss": 0.6694, + "step": 1043 + }, + { + "epoch": 0.39977024698449165, + "grad_norm": 0.2762261927127838, + "learning_rate": 4.818394809505421e-06, + "loss": 0.6962, + "step": 1044 + }, + { + "epoch": 0.40015316867700557, + "grad_norm": 0.28600603342056274, + "learning_rate": 4.818014588807625e-06, + "loss": 0.6734, + "step": 1045 + }, + { + "epoch": 0.4005360903695194, + "grad_norm": 0.26933541893959045, + "learning_rate": 4.817633985531924e-06, + "loss": 0.6775, + "step": 1046 + }, + { + "epoch": 0.4009190120620333, + "grad_norm": 0.2707444131374359, + "learning_rate": 4.817252999741137e-06, + "loss": 0.6924, + "step": 1047 + }, + { + "epoch": 0.4013019337545472, + "grad_norm": 0.27176135778427124, + "learning_rate": 4.816871631498141e-06, + "loss": 0.6689, + "step": 1048 + }, + { + "epoch": 0.40168485544706106, + "grad_norm": 0.28387269377708435, + "learning_rate": 4.81648988086588e-06, + "loss": 0.6879, + "step": 1049 + }, + { + "epoch": 0.402067777139575, + "grad_norm": 0.2803856432437897, + "learning_rate": 4.816107747907362e-06, + "loss": 0.6779, + "step": 1050 + }, + { + "epoch": 0.40245069883208884, + "grad_norm": 0.26472827792167664, + "learning_rate": 4.815725232685654e-06, + "loss": 0.6671, + "step": 1051 + }, + { + "epoch": 0.4028336205246027, + "grad_norm": 0.2867729663848877, + "learning_rate": 4.815342335263891e-06, + "loss": 0.6849, + "step": 1052 + }, + { + "epoch": 0.4032165422171166, + "grad_norm": 0.2667827904224396, + "learning_rate": 4.814959055705266e-06, + "loss": 0.6759, + "step": 1053 + }, + { + "epoch": 0.4035994639096305, + "grad_norm": 0.2698298394680023, + "learning_rate": 4.814575394073039e-06, + "loss": 0.6777, + "step": 1054 + }, + { + "epoch": 0.40398238560214433, + "grad_norm": 0.27925172448158264, + "learning_rate": 4.8141913504305305e-06, + "loss": 0.6872, + "step": 1055 + }, + { + "epoch": 0.40436530729465825, + "grad_norm": 0.26620960235595703, + "learning_rate": 4.813806924841125e-06, + "loss": 0.6848, + "step": 1056 + }, + { + "epoch": 0.4047482289871721, + "grad_norm": 0.2703430950641632, + "learning_rate": 4.813422117368273e-06, + "loss": 0.6508, + "step": 1057 + }, + { + "epoch": 0.405131150679686, + "grad_norm": 0.2778773903846741, + "learning_rate": 4.813036928075483e-06, + "loss": 0.6802, + "step": 1058 + }, + { + "epoch": 0.4055140723721999, + "grad_norm": 0.268136590719223, + "learning_rate": 4.812651357026329e-06, + "loss": 0.6894, + "step": 1059 + }, + { + "epoch": 0.40589699406471375, + "grad_norm": 0.2591809332370758, + "learning_rate": 4.812265404284447e-06, + "loss": 0.6751, + "step": 1060 + }, + { + "epoch": 0.40627991575722766, + "grad_norm": 0.26853811740875244, + "learning_rate": 4.811879069913539e-06, + "loss": 0.7125, + "step": 1061 + }, + { + "epoch": 0.4066628374497415, + "grad_norm": 0.2730932831764221, + "learning_rate": 4.811492353977366e-06, + "loss": 0.6852, + "step": 1062 + }, + { + "epoch": 0.4070457591422554, + "grad_norm": 0.26328590512275696, + "learning_rate": 4.811105256539754e-06, + "loss": 0.6585, + "step": 1063 + }, + { + "epoch": 0.4074286808347693, + "grad_norm": 0.27260056138038635, + "learning_rate": 4.810717777664592e-06, + "loss": 0.677, + "step": 1064 + }, + { + "epoch": 0.40781160252728316, + "grad_norm": 0.2737332880496979, + "learning_rate": 4.810329917415832e-06, + "loss": 0.6636, + "step": 1065 + }, + { + "epoch": 0.4081945242197971, + "grad_norm": 0.2796006500720978, + "learning_rate": 4.809941675857487e-06, + "loss": 0.6968, + "step": 1066 + }, + { + "epoch": 0.40857744591231093, + "grad_norm": 0.2870083153247833, + "learning_rate": 4.809553053053635e-06, + "loss": 0.6814, + "step": 1067 + }, + { + "epoch": 0.4089603676048248, + "grad_norm": 0.2642041742801666, + "learning_rate": 4.809164049068419e-06, + "loss": 0.6851, + "step": 1068 + }, + { + "epoch": 0.4093432892973387, + "grad_norm": 0.27570584416389465, + "learning_rate": 4.808774663966039e-06, + "loss": 0.6857, + "step": 1069 + }, + { + "epoch": 0.40972621098985257, + "grad_norm": 0.2671034038066864, + "learning_rate": 4.808384897810763e-06, + "loss": 0.663, + "step": 1070 + }, + { + "epoch": 0.41010913268236643, + "grad_norm": 0.2757396996021271, + "learning_rate": 4.80799475066692e-06, + "loss": 0.6729, + "step": 1071 + }, + { + "epoch": 0.41049205437488034, + "grad_norm": 0.28543686866760254, + "learning_rate": 4.807604222598902e-06, + "loss": 0.7029, + "step": 1072 + }, + { + "epoch": 0.4108749760673942, + "grad_norm": 0.28299906849861145, + "learning_rate": 4.807213313671163e-06, + "loss": 0.6813, + "step": 1073 + }, + { + "epoch": 0.4112578977599081, + "grad_norm": 0.27811309695243835, + "learning_rate": 4.806822023948223e-06, + "loss": 0.6832, + "step": 1074 + }, + { + "epoch": 0.411640819452422, + "grad_norm": 0.2794390320777893, + "learning_rate": 4.80643035349466e-06, + "loss": 0.6993, + "step": 1075 + }, + { + "epoch": 0.41202374114493584, + "grad_norm": 0.2820933163166046, + "learning_rate": 4.806038302375119e-06, + "loss": 0.6751, + "step": 1076 + }, + { + "epoch": 0.41240666283744976, + "grad_norm": 0.28078168630599976, + "learning_rate": 4.805645870654306e-06, + "loss": 0.6661, + "step": 1077 + }, + { + "epoch": 0.4127895845299636, + "grad_norm": 0.2717399001121521, + "learning_rate": 4.805253058396989e-06, + "loss": 0.6955, + "step": 1078 + }, + { + "epoch": 0.4131725062224775, + "grad_norm": 0.2717928886413574, + "learning_rate": 4.804859865668002e-06, + "loss": 0.6648, + "step": 1079 + }, + { + "epoch": 0.4135554279149914, + "grad_norm": 0.28002336621284485, + "learning_rate": 4.804466292532239e-06, + "loss": 0.6931, + "step": 1080 + }, + { + "epoch": 0.41393834960750525, + "grad_norm": 0.28147804737091064, + "learning_rate": 4.804072339054656e-06, + "loss": 0.6844, + "step": 1081 + }, + { + "epoch": 0.41432127130001917, + "grad_norm": 0.2660662829875946, + "learning_rate": 4.803678005300274e-06, + "loss": 0.6701, + "step": 1082 + }, + { + "epoch": 0.414704192992533, + "grad_norm": 0.26961514353752136, + "learning_rate": 4.803283291334177e-06, + "loss": 0.6874, + "step": 1083 + }, + { + "epoch": 0.4150871146850469, + "grad_norm": 0.2822124660015106, + "learning_rate": 4.802888197221511e-06, + "loss": 0.6702, + "step": 1084 + }, + { + "epoch": 0.4154700363775608, + "grad_norm": 0.264700710773468, + "learning_rate": 4.802492723027482e-06, + "loss": 0.6755, + "step": 1085 + }, + { + "epoch": 0.41585295807007466, + "grad_norm": 0.2882467210292816, + "learning_rate": 4.802096868817363e-06, + "loss": 0.6995, + "step": 1086 + }, + { + "epoch": 0.4162358797625885, + "grad_norm": 0.2640404999256134, + "learning_rate": 4.801700634656488e-06, + "loss": 0.6657, + "step": 1087 + }, + { + "epoch": 0.41661880145510244, + "grad_norm": 0.2711586654186249, + "learning_rate": 4.801304020610255e-06, + "loss": 0.6831, + "step": 1088 + }, + { + "epoch": 0.4170017231476163, + "grad_norm": 0.27783191204071045, + "learning_rate": 4.80090702674412e-06, + "loss": 0.6816, + "step": 1089 + }, + { + "epoch": 0.4173846448401302, + "grad_norm": 0.28974947333335876, + "learning_rate": 4.800509653123607e-06, + "loss": 0.6523, + "step": 1090 + }, + { + "epoch": 0.4177675665326441, + "grad_norm": 0.27213558554649353, + "learning_rate": 4.800111899814301e-06, + "loss": 0.6873, + "step": 1091 + }, + { + "epoch": 0.41815048822515793, + "grad_norm": 0.26818469166755676, + "learning_rate": 4.7997137668818495e-06, + "loss": 0.6833, + "step": 1092 + }, + { + "epoch": 0.41853340991767185, + "grad_norm": 0.2849678099155426, + "learning_rate": 4.7993152543919615e-06, + "loss": 0.676, + "step": 1093 + }, + { + "epoch": 0.4189163316101857, + "grad_norm": 0.2663874328136444, + "learning_rate": 4.798916362410411e-06, + "loss": 0.6421, + "step": 1094 + }, + { + "epoch": 0.41929925330269957, + "grad_norm": 0.28957879543304443, + "learning_rate": 4.798517091003032e-06, + "loss": 0.691, + "step": 1095 + }, + { + "epoch": 0.4196821749952135, + "grad_norm": 0.2755984663963318, + "learning_rate": 4.798117440235723e-06, + "loss": 0.6774, + "step": 1096 + }, + { + "epoch": 0.42006509668772735, + "grad_norm": 0.2807273268699646, + "learning_rate": 4.797717410174445e-06, + "loss": 0.6792, + "step": 1097 + }, + { + "epoch": 0.42044801838024126, + "grad_norm": 0.2845119833946228, + "learning_rate": 4.79731700088522e-06, + "loss": 0.6551, + "step": 1098 + }, + { + "epoch": 0.4208309400727551, + "grad_norm": 0.28206029534339905, + "learning_rate": 4.7969162124341354e-06, + "loss": 0.6965, + "step": 1099 + }, + { + "epoch": 0.421213861765269, + "grad_norm": 0.276937872171402, + "learning_rate": 4.796515044887338e-06, + "loss": 0.6652, + "step": 1100 + }, + { + "epoch": 0.4215967834577829, + "grad_norm": 0.2648085653781891, + "learning_rate": 4.79611349831104e-06, + "loss": 0.6611, + "step": 1101 + }, + { + "epoch": 0.42197970515029676, + "grad_norm": 0.28414374589920044, + "learning_rate": 4.795711572771515e-06, + "loss": 0.6802, + "step": 1102 + }, + { + "epoch": 0.4223626268428106, + "grad_norm": 0.27629321813583374, + "learning_rate": 4.795309268335096e-06, + "loss": 0.6681, + "step": 1103 + }, + { + "epoch": 0.42274554853532453, + "grad_norm": 0.27043506503105164, + "learning_rate": 4.794906585068185e-06, + "loss": 0.6861, + "step": 1104 + }, + { + "epoch": 0.4231284702278384, + "grad_norm": 0.2690626382827759, + "learning_rate": 4.794503523037242e-06, + "loss": 0.6826, + "step": 1105 + }, + { + "epoch": 0.4235113919203523, + "grad_norm": 0.2778787910938263, + "learning_rate": 4.794100082308791e-06, + "loss": 0.6767, + "step": 1106 + }, + { + "epoch": 0.42389431361286617, + "grad_norm": 0.27799922227859497, + "learning_rate": 4.793696262949417e-06, + "loss": 0.6542, + "step": 1107 + }, + { + "epoch": 0.42427723530538003, + "grad_norm": 0.2779891788959503, + "learning_rate": 4.793292065025768e-06, + "loss": 0.6926, + "step": 1108 + }, + { + "epoch": 0.42466015699789394, + "grad_norm": 0.2670944631099701, + "learning_rate": 4.792887488604558e-06, + "loss": 0.6685, + "step": 1109 + }, + { + "epoch": 0.4250430786904078, + "grad_norm": 0.2732831537723541, + "learning_rate": 4.792482533752557e-06, + "loss": 0.6636, + "step": 1110 + }, + { + "epoch": 0.42542600038292167, + "grad_norm": 0.2689459025859833, + "learning_rate": 4.7920772005366035e-06, + "loss": 0.6914, + "step": 1111 + }, + { + "epoch": 0.4258089220754356, + "grad_norm": 0.27064675092697144, + "learning_rate": 4.791671489023595e-06, + "loss": 0.6674, + "step": 1112 + }, + { + "epoch": 0.42619184376794944, + "grad_norm": 0.2776367962360382, + "learning_rate": 4.791265399280493e-06, + "loss": 0.6828, + "step": 1113 + }, + { + "epoch": 0.42657476546046336, + "grad_norm": 0.2685474157333374, + "learning_rate": 4.7908589313743195e-06, + "loss": 0.6682, + "step": 1114 + }, + { + "epoch": 0.4269576871529772, + "grad_norm": 0.2742334306240082, + "learning_rate": 4.7904520853721616e-06, + "loss": 0.6943, + "step": 1115 + }, + { + "epoch": 0.4273406088454911, + "grad_norm": 0.2855285704135895, + "learning_rate": 4.790044861341166e-06, + "loss": 0.6797, + "step": 1116 + }, + { + "epoch": 0.427723530538005, + "grad_norm": 0.2690899074077606, + "learning_rate": 4.789637259348544e-06, + "loss": 0.6776, + "step": 1117 + }, + { + "epoch": 0.42810645223051885, + "grad_norm": 0.2653801143169403, + "learning_rate": 4.789229279461569e-06, + "loss": 0.6502, + "step": 1118 + }, + { + "epoch": 0.4284893739230327, + "grad_norm": 0.2711450755596161, + "learning_rate": 4.788820921747576e-06, + "loss": 0.6854, + "step": 1119 + }, + { + "epoch": 0.42887229561554663, + "grad_norm": 0.27033019065856934, + "learning_rate": 4.788412186273961e-06, + "loss": 0.6678, + "step": 1120 + }, + { + "epoch": 0.4292552173080605, + "grad_norm": 0.26320865750312805, + "learning_rate": 4.788003073108186e-06, + "loss": 0.6555, + "step": 1121 + }, + { + "epoch": 0.4296381390005744, + "grad_norm": 0.2914315462112427, + "learning_rate": 4.787593582317772e-06, + "loss": 0.6776, + "step": 1122 + }, + { + "epoch": 0.43002106069308826, + "grad_norm": 0.269633412361145, + "learning_rate": 4.787183713970304e-06, + "loss": 0.6617, + "step": 1123 + }, + { + "epoch": 0.4304039823856021, + "grad_norm": 0.2707017660140991, + "learning_rate": 4.786773468133431e-06, + "loss": 0.6623, + "step": 1124 + }, + { + "epoch": 0.43078690407811604, + "grad_norm": 0.2661648392677307, + "learning_rate": 4.786362844874859e-06, + "loss": 0.6749, + "step": 1125 + }, + { + "epoch": 0.4311698257706299, + "grad_norm": 0.2712387144565582, + "learning_rate": 4.785951844262361e-06, + "loss": 0.6721, + "step": 1126 + }, + { + "epoch": 0.43155274746314376, + "grad_norm": 0.2739746868610382, + "learning_rate": 4.7855404663637695e-06, + "loss": 0.6413, + "step": 1127 + }, + { + "epoch": 0.4319356691556577, + "grad_norm": 0.27242210507392883, + "learning_rate": 4.785128711246983e-06, + "loss": 0.6718, + "step": 1128 + }, + { + "epoch": 0.43231859084817154, + "grad_norm": 0.2739618718624115, + "learning_rate": 4.7847165789799586e-06, + "loss": 0.6632, + "step": 1129 + }, + { + "epoch": 0.43270151254068545, + "grad_norm": 0.272318571805954, + "learning_rate": 4.784304069630716e-06, + "loss": 0.6604, + "step": 1130 + }, + { + "epoch": 0.4330844342331993, + "grad_norm": 0.2719273865222931, + "learning_rate": 4.78389118326734e-06, + "loss": 0.6635, + "step": 1131 + }, + { + "epoch": 0.43346735592571317, + "grad_norm": 0.28921982645988464, + "learning_rate": 4.783477919957973e-06, + "loss": 0.6716, + "step": 1132 + }, + { + "epoch": 0.4338502776182271, + "grad_norm": 0.26451462507247925, + "learning_rate": 4.783064279770825e-06, + "loss": 0.6699, + "step": 1133 + }, + { + "epoch": 0.43423319931074095, + "grad_norm": 0.2656538784503937, + "learning_rate": 4.782650262774164e-06, + "loss": 0.6748, + "step": 1134 + }, + { + "epoch": 0.4346161210032548, + "grad_norm": 0.27099308371543884, + "learning_rate": 4.782235869036321e-06, + "loss": 0.6716, + "step": 1135 + }, + { + "epoch": 0.4349990426957687, + "grad_norm": 0.28783848881721497, + "learning_rate": 4.781821098625691e-06, + "loss": 0.673, + "step": 1136 + }, + { + "epoch": 0.4353819643882826, + "grad_norm": 0.27346324920654297, + "learning_rate": 4.78140595161073e-06, + "loss": 0.6527, + "step": 1137 + }, + { + "epoch": 0.4357648860807965, + "grad_norm": 0.27585262060165405, + "learning_rate": 4.7809904280599564e-06, + "loss": 0.6528, + "step": 1138 + }, + { + "epoch": 0.43614780777331036, + "grad_norm": 0.28196394443511963, + "learning_rate": 4.780574528041949e-06, + "loss": 0.6911, + "step": 1139 + }, + { + "epoch": 0.4365307294658242, + "grad_norm": 0.27955520153045654, + "learning_rate": 4.780158251625351e-06, + "loss": 0.6759, + "step": 1140 + }, + { + "epoch": 0.43691365115833813, + "grad_norm": 0.2854354679584503, + "learning_rate": 4.7797415988788675e-06, + "loss": 0.6759, + "step": 1141 + }, + { + "epoch": 0.437296572850852, + "grad_norm": 0.2818031311035156, + "learning_rate": 4.7793245698712655e-06, + "loss": 0.677, + "step": 1142 + }, + { + "epoch": 0.43767949454336585, + "grad_norm": 0.2737523913383484, + "learning_rate": 4.778907164671372e-06, + "loss": 0.7044, + "step": 1143 + }, + { + "epoch": 0.43806241623587977, + "grad_norm": 0.27465736865997314, + "learning_rate": 4.778489383348079e-06, + "loss": 0.6752, + "step": 1144 + }, + { + "epoch": 0.43844533792839363, + "grad_norm": 0.27898678183555603, + "learning_rate": 4.77807122597034e-06, + "loss": 0.6794, + "step": 1145 + }, + { + "epoch": 0.43882825962090755, + "grad_norm": 0.27347368001937866, + "learning_rate": 4.777652692607168e-06, + "loss": 0.6559, + "step": 1146 + }, + { + "epoch": 0.4392111813134214, + "grad_norm": 0.2761583924293518, + "learning_rate": 4.7772337833276425e-06, + "loss": 0.679, + "step": 1147 + }, + { + "epoch": 0.43959410300593527, + "grad_norm": 0.285239577293396, + "learning_rate": 4.776814498200901e-06, + "loss": 0.6658, + "step": 1148 + }, + { + "epoch": 0.4399770246984492, + "grad_norm": 0.2704262435436249, + "learning_rate": 4.776394837296145e-06, + "loss": 0.6773, + "step": 1149 + }, + { + "epoch": 0.44035994639096304, + "grad_norm": 0.2713160514831543, + "learning_rate": 4.775974800682639e-06, + "loss": 0.6722, + "step": 1150 + }, + { + "epoch": 0.4407428680834769, + "grad_norm": 0.2755696177482605, + "learning_rate": 4.775554388429706e-06, + "loss": 0.6703, + "step": 1151 + }, + { + "epoch": 0.4411257897759908, + "grad_norm": 0.2795194685459137, + "learning_rate": 4.775133600606734e-06, + "loss": 0.6668, + "step": 1152 + }, + { + "epoch": 0.4415087114685047, + "grad_norm": 0.2720242440700531, + "learning_rate": 4.774712437283172e-06, + "loss": 0.6758, + "step": 1153 + }, + { + "epoch": 0.4418916331610186, + "grad_norm": 0.27648648619651794, + "learning_rate": 4.774290898528532e-06, + "loss": 0.6724, + "step": 1154 + }, + { + "epoch": 0.44227455485353245, + "grad_norm": 0.29117751121520996, + "learning_rate": 4.773868984412386e-06, + "loss": 0.6599, + "step": 1155 + }, + { + "epoch": 0.4426574765460463, + "grad_norm": 0.2748560607433319, + "learning_rate": 4.7734466950043686e-06, + "loss": 0.6716, + "step": 1156 + }, + { + "epoch": 0.44304039823856023, + "grad_norm": 0.29462385177612305, + "learning_rate": 4.773024030374179e-06, + "loss": 0.6465, + "step": 1157 + }, + { + "epoch": 0.4434233199310741, + "grad_norm": 0.27808573842048645, + "learning_rate": 4.772600990591574e-06, + "loss": 0.6567, + "step": 1158 + }, + { + "epoch": 0.44380624162358795, + "grad_norm": 0.2768598198890686, + "learning_rate": 4.772177575726375e-06, + "loss": 0.6814, + "step": 1159 + }, + { + "epoch": 0.44418916331610186, + "grad_norm": 0.2841338813304901, + "learning_rate": 4.7717537858484654e-06, + "loss": 0.666, + "step": 1160 + }, + { + "epoch": 0.4445720850086157, + "grad_norm": 0.2649598717689514, + "learning_rate": 4.77132962102779e-06, + "loss": 0.6514, + "step": 1161 + }, + { + "epoch": 0.44495500670112964, + "grad_norm": 0.27885758876800537, + "learning_rate": 4.770905081334352e-06, + "loss": 0.6653, + "step": 1162 + }, + { + "epoch": 0.4453379283936435, + "grad_norm": 0.2814213037490845, + "learning_rate": 4.770480166838224e-06, + "loss": 0.6548, + "step": 1163 + }, + { + "epoch": 0.44572085008615736, + "grad_norm": 0.2747483551502228, + "learning_rate": 4.770054877609534e-06, + "loss": 0.6671, + "step": 1164 + }, + { + "epoch": 0.4461037717786713, + "grad_norm": 0.26971474289894104, + "learning_rate": 4.769629213718476e-06, + "loss": 0.6826, + "step": 1165 + }, + { + "epoch": 0.44648669347118514, + "grad_norm": 0.27830854058265686, + "learning_rate": 4.769203175235302e-06, + "loss": 0.6828, + "step": 1166 + }, + { + "epoch": 0.446869615163699, + "grad_norm": 0.27223125100135803, + "learning_rate": 4.768776762230328e-06, + "loss": 0.6711, + "step": 1167 + }, + { + "epoch": 0.4472525368562129, + "grad_norm": 0.2749093174934387, + "learning_rate": 4.768349974773931e-06, + "loss": 0.6877, + "step": 1168 + }, + { + "epoch": 0.4476354585487268, + "grad_norm": 0.2665577530860901, + "learning_rate": 4.767922812936552e-06, + "loss": 0.6861, + "step": 1169 + }, + { + "epoch": 0.4480183802412407, + "grad_norm": 0.2772512137889862, + "learning_rate": 4.767495276788692e-06, + "loss": 0.6654, + "step": 1170 + }, + { + "epoch": 0.44840130193375455, + "grad_norm": 0.2787632942199707, + "learning_rate": 4.767067366400913e-06, + "loss": 0.6702, + "step": 1171 + }, + { + "epoch": 0.4487842236262684, + "grad_norm": 0.2907279431819916, + "learning_rate": 4.766639081843839e-06, + "loss": 0.6818, + "step": 1172 + }, + { + "epoch": 0.4491671453187823, + "grad_norm": 0.27912425994873047, + "learning_rate": 4.766210423188158e-06, + "loss": 0.6535, + "step": 1173 + }, + { + "epoch": 0.4495500670112962, + "grad_norm": 0.29031383991241455, + "learning_rate": 4.765781390504617e-06, + "loss": 0.6915, + "step": 1174 + }, + { + "epoch": 0.44993298870381004, + "grad_norm": 0.2793537676334381, + "learning_rate": 4.765351983864027e-06, + "loss": 0.6703, + "step": 1175 + }, + { + "epoch": 0.45031591039632396, + "grad_norm": 0.28281429409980774, + "learning_rate": 4.76492220333726e-06, + "loss": 0.6725, + "step": 1176 + }, + { + "epoch": 0.4506988320888378, + "grad_norm": 0.2810090482234955, + "learning_rate": 4.7644920489952474e-06, + "loss": 0.6859, + "step": 1177 + }, + { + "epoch": 0.45108175378135174, + "grad_norm": 0.28927022218704224, + "learning_rate": 4.764061520908986e-06, + "loss": 0.6828, + "step": 1178 + }, + { + "epoch": 0.4514646754738656, + "grad_norm": 0.29566314816474915, + "learning_rate": 4.763630619149531e-06, + "loss": 0.6731, + "step": 1179 + }, + { + "epoch": 0.45184759716637946, + "grad_norm": 0.2778233289718628, + "learning_rate": 4.763199343788002e-06, + "loss": 0.6833, + "step": 1180 + }, + { + "epoch": 0.45223051885889337, + "grad_norm": 0.27039292454719543, + "learning_rate": 4.762767694895579e-06, + "loss": 0.661, + "step": 1181 + }, + { + "epoch": 0.45261344055140723, + "grad_norm": 0.2723767161369324, + "learning_rate": 4.762335672543503e-06, + "loss": 0.6646, + "step": 1182 + }, + { + "epoch": 0.4529963622439211, + "grad_norm": 0.2879144847393036, + "learning_rate": 4.761903276803078e-06, + "loss": 0.6699, + "step": 1183 + }, + { + "epoch": 0.453379283936435, + "grad_norm": 0.2802518904209137, + "learning_rate": 4.761470507745669e-06, + "loss": 0.6644, + "step": 1184 + }, + { + "epoch": 0.45376220562894887, + "grad_norm": 0.2758473753929138, + "learning_rate": 4.761037365442703e-06, + "loss": 0.6555, + "step": 1185 + }, + { + "epoch": 0.4541451273214628, + "grad_norm": 0.282187819480896, + "learning_rate": 4.760603849965666e-06, + "loss": 0.6723, + "step": 1186 + }, + { + "epoch": 0.45452804901397664, + "grad_norm": 0.275897741317749, + "learning_rate": 4.760169961386111e-06, + "loss": 0.6669, + "step": 1187 + }, + { + "epoch": 0.4549109707064905, + "grad_norm": 0.2768770754337311, + "learning_rate": 4.759735699775647e-06, + "loss": 0.6624, + "step": 1188 + }, + { + "epoch": 0.4552938923990044, + "grad_norm": 0.26994645595550537, + "learning_rate": 4.759301065205947e-06, + "loss": 0.667, + "step": 1189 + }, + { + "epoch": 0.4556768140915183, + "grad_norm": 0.27651163935661316, + "learning_rate": 4.758866057748747e-06, + "loss": 0.6694, + "step": 1190 + }, + { + "epoch": 0.45605973578403214, + "grad_norm": 0.2905432879924774, + "learning_rate": 4.758430677475841e-06, + "loss": 0.6865, + "step": 1191 + }, + { + "epoch": 0.45644265747654605, + "grad_norm": 0.27474573254585266, + "learning_rate": 4.757994924459089e-06, + "loss": 0.6821, + "step": 1192 + }, + { + "epoch": 0.4568255791690599, + "grad_norm": 0.27210354804992676, + "learning_rate": 4.7575587987704085e-06, + "loss": 0.6592, + "step": 1193 + }, + { + "epoch": 0.45720850086157383, + "grad_norm": 0.26745250821113586, + "learning_rate": 4.75712230048178e-06, + "loss": 0.6641, + "step": 1194 + }, + { + "epoch": 0.4575914225540877, + "grad_norm": 0.271131306886673, + "learning_rate": 4.756685429665246e-06, + "loss": 0.6684, + "step": 1195 + }, + { + "epoch": 0.45797434424660155, + "grad_norm": 0.2732897102832794, + "learning_rate": 4.75624818639291e-06, + "loss": 0.6835, + "step": 1196 + }, + { + "epoch": 0.45835726593911547, + "grad_norm": 0.2688729465007782, + "learning_rate": 4.755810570736938e-06, + "loss": 0.6834, + "step": 1197 + }, + { + "epoch": 0.4587401876316293, + "grad_norm": 0.2814828157424927, + "learning_rate": 4.755372582769554e-06, + "loss": 0.6643, + "step": 1198 + }, + { + "epoch": 0.4591231093241432, + "grad_norm": 0.2694145441055298, + "learning_rate": 4.754934222563048e-06, + "loss": 0.6753, + "step": 1199 + }, + { + "epoch": 0.4595060310166571, + "grad_norm": 0.2745696008205414, + "learning_rate": 4.754495490189769e-06, + "loss": 0.6885, + "step": 1200 + }, + { + "epoch": 0.45988895270917096, + "grad_norm": 0.2827391028404236, + "learning_rate": 4.754056385722128e-06, + "loss": 0.6745, + "step": 1201 + }, + { + "epoch": 0.4602718744016849, + "grad_norm": 0.2724725306034088, + "learning_rate": 4.753616909232597e-06, + "loss": 0.6589, + "step": 1202 + }, + { + "epoch": 0.46065479609419874, + "grad_norm": 0.2880653142929077, + "learning_rate": 4.753177060793709e-06, + "loss": 0.6796, + "step": 1203 + }, + { + "epoch": 0.4610377177867126, + "grad_norm": 0.27427205443382263, + "learning_rate": 4.75273684047806e-06, + "loss": 0.6961, + "step": 1204 + }, + { + "epoch": 0.4614206394792265, + "grad_norm": 0.2762588858604431, + "learning_rate": 4.752296248358306e-06, + "loss": 0.6542, + "step": 1205 + }, + { + "epoch": 0.4618035611717404, + "grad_norm": 0.2718510925769806, + "learning_rate": 4.7518552845071646e-06, + "loss": 0.676, + "step": 1206 + }, + { + "epoch": 0.46218648286425423, + "grad_norm": 0.27809596061706543, + "learning_rate": 4.7514139489974145e-06, + "loss": 0.6526, + "step": 1207 + }, + { + "epoch": 0.46256940455676815, + "grad_norm": 0.2677210867404938, + "learning_rate": 4.750972241901897e-06, + "loss": 0.6587, + "step": 1208 + }, + { + "epoch": 0.462952326249282, + "grad_norm": 0.2862596809864044, + "learning_rate": 4.750530163293514e-06, + "loss": 0.6818, + "step": 1209 + }, + { + "epoch": 0.4633352479417959, + "grad_norm": 0.27850595116615295, + "learning_rate": 4.750087713245227e-06, + "loss": 0.6662, + "step": 1210 + }, + { + "epoch": 0.4637181696343098, + "grad_norm": 0.27618399262428284, + "learning_rate": 4.749644891830063e-06, + "loss": 0.6614, + "step": 1211 + }, + { + "epoch": 0.46410109132682364, + "grad_norm": 0.2669837176799774, + "learning_rate": 4.749201699121105e-06, + "loss": 0.6749, + "step": 1212 + }, + { + "epoch": 0.46448401301933756, + "grad_norm": 0.27776703238487244, + "learning_rate": 4.748758135191502e-06, + "loss": 0.6695, + "step": 1213 + }, + { + "epoch": 0.4648669347118514, + "grad_norm": 0.26955899596214294, + "learning_rate": 4.74831420011446e-06, + "loss": 0.6766, + "step": 1214 + }, + { + "epoch": 0.4652498564043653, + "grad_norm": 0.2800334393978119, + "learning_rate": 4.747869893963252e-06, + "loss": 0.6707, + "step": 1215 + }, + { + "epoch": 0.4656327780968792, + "grad_norm": 0.3023186922073364, + "learning_rate": 4.747425216811205e-06, + "loss": 0.669, + "step": 1216 + }, + { + "epoch": 0.46601569978939306, + "grad_norm": 0.2796546518802643, + "learning_rate": 4.746980168731713e-06, + "loss": 0.6911, + "step": 1217 + }, + { + "epoch": 0.46639862148190697, + "grad_norm": 0.26911649107933044, + "learning_rate": 4.7465347497982286e-06, + "loss": 0.6709, + "step": 1218 + }, + { + "epoch": 0.46678154317442083, + "grad_norm": 0.2805267870426178, + "learning_rate": 4.746088960084267e-06, + "loss": 0.6772, + "step": 1219 + }, + { + "epoch": 0.4671644648669347, + "grad_norm": 0.26144716143608093, + "learning_rate": 4.745642799663403e-06, + "loss": 0.6947, + "step": 1220 + }, + { + "epoch": 0.4675473865594486, + "grad_norm": 0.28579649329185486, + "learning_rate": 4.745196268609274e-06, + "loss": 0.66, + "step": 1221 + }, + { + "epoch": 0.46793030825196247, + "grad_norm": 0.27812519669532776, + "learning_rate": 4.744749366995576e-06, + "loss": 0.6628, + "step": 1222 + }, + { + "epoch": 0.46831322994447633, + "grad_norm": 0.26414886116981506, + "learning_rate": 4.744302094896071e-06, + "loss": 0.653, + "step": 1223 + }, + { + "epoch": 0.46869615163699024, + "grad_norm": 0.2737218737602234, + "learning_rate": 4.743854452384577e-06, + "loss": 0.6817, + "step": 1224 + }, + { + "epoch": 0.4690790733295041, + "grad_norm": 0.2919729948043823, + "learning_rate": 4.743406439534976e-06, + "loss": 0.6666, + "step": 1225 + }, + { + "epoch": 0.469461995022018, + "grad_norm": 0.2719593644142151, + "learning_rate": 4.742958056421211e-06, + "loss": 0.674, + "step": 1226 + }, + { + "epoch": 0.4698449167145319, + "grad_norm": 0.27731117606163025, + "learning_rate": 4.742509303117286e-06, + "loss": 0.6695, + "step": 1227 + }, + { + "epoch": 0.47022783840704574, + "grad_norm": 0.2854015529155731, + "learning_rate": 4.742060179697264e-06, + "loss": 0.6684, + "step": 1228 + }, + { + "epoch": 0.47061076009955966, + "grad_norm": 0.275790810585022, + "learning_rate": 4.741610686235273e-06, + "loss": 0.673, + "step": 1229 + }, + { + "epoch": 0.4709936817920735, + "grad_norm": 0.27415552735328674, + "learning_rate": 4.741160822805499e-06, + "loss": 0.6658, + "step": 1230 + }, + { + "epoch": 0.4713766034845874, + "grad_norm": 0.2823668122291565, + "learning_rate": 4.740710589482189e-06, + "loss": 0.681, + "step": 1231 + }, + { + "epoch": 0.4717595251771013, + "grad_norm": 0.27480632066726685, + "learning_rate": 4.740259986339652e-06, + "loss": 0.6796, + "step": 1232 + }, + { + "epoch": 0.47214244686961515, + "grad_norm": 0.28374215960502625, + "learning_rate": 4.73980901345226e-06, + "loss": 0.6785, + "step": 1233 + }, + { + "epoch": 0.47252536856212907, + "grad_norm": 0.28118008375167847, + "learning_rate": 4.739357670894442e-06, + "loss": 0.6686, + "step": 1234 + }, + { + "epoch": 0.4729082902546429, + "grad_norm": 0.28157249093055725, + "learning_rate": 4.738905958740692e-06, + "loss": 0.6779, + "step": 1235 + }, + { + "epoch": 0.4732912119471568, + "grad_norm": 0.2744462192058563, + "learning_rate": 4.738453877065561e-06, + "loss": 0.6666, + "step": 1236 + }, + { + "epoch": 0.4736741336396707, + "grad_norm": 0.28920602798461914, + "learning_rate": 4.738001425943664e-06, + "loss": 0.7009, + "step": 1237 + }, + { + "epoch": 0.47405705533218456, + "grad_norm": 0.261415034532547, + "learning_rate": 4.737548605449675e-06, + "loss": 0.6634, + "step": 1238 + }, + { + "epoch": 0.4744399770246984, + "grad_norm": 0.2748316526412964, + "learning_rate": 4.7370954156583326e-06, + "loss": 0.6804, + "step": 1239 + }, + { + "epoch": 0.47482289871721234, + "grad_norm": 0.2750059962272644, + "learning_rate": 4.736641856644431e-06, + "loss": 0.6934, + "step": 1240 + }, + { + "epoch": 0.4752058204097262, + "grad_norm": 0.28661346435546875, + "learning_rate": 4.73618792848283e-06, + "loss": 0.698, + "step": 1241 + }, + { + "epoch": 0.4755887421022401, + "grad_norm": 0.2804570496082306, + "learning_rate": 4.7357336312484466e-06, + "loss": 0.6702, + "step": 1242 + }, + { + "epoch": 0.475971663794754, + "grad_norm": 0.2865023612976074, + "learning_rate": 4.735278965016262e-06, + "loss": 0.6707, + "step": 1243 + }, + { + "epoch": 0.47635458548726783, + "grad_norm": 0.30556708574295044, + "learning_rate": 4.734823929861317e-06, + "loss": 0.6755, + "step": 1244 + }, + { + "epoch": 0.47673750717978175, + "grad_norm": 0.28410694003105164, + "learning_rate": 4.734368525858713e-06, + "loss": 0.6764, + "step": 1245 + }, + { + "epoch": 0.4771204288722956, + "grad_norm": 0.2751966416835785, + "learning_rate": 4.7339127530836095e-06, + "loss": 0.6598, + "step": 1246 + }, + { + "epoch": 0.47750335056480947, + "grad_norm": 0.2905697226524353, + "learning_rate": 4.733456611611233e-06, + "loss": 0.6876, + "step": 1247 + }, + { + "epoch": 0.4778862722573234, + "grad_norm": 0.30000394582748413, + "learning_rate": 4.733000101516868e-06, + "loss": 0.6747, + "step": 1248 + }, + { + "epoch": 0.47826919394983725, + "grad_norm": 0.26689612865448, + "learning_rate": 4.732543222875857e-06, + "loss": 0.6729, + "step": 1249 + }, + { + "epoch": 0.47865211564235116, + "grad_norm": 0.2899976372718811, + "learning_rate": 4.7320859757636074e-06, + "loss": 0.6758, + "step": 1250 + }, + { + "epoch": 0.479035037334865, + "grad_norm": 0.3092990815639496, + "learning_rate": 4.731628360255585e-06, + "loss": 0.6826, + "step": 1251 + }, + { + "epoch": 0.4794179590273789, + "grad_norm": 0.2779189348220825, + "learning_rate": 4.731170376427317e-06, + "loss": 0.667, + "step": 1252 + }, + { + "epoch": 0.4798008807198928, + "grad_norm": 0.29295408725738525, + "learning_rate": 4.7307120243543925e-06, + "loss": 0.6593, + "step": 1253 + }, + { + "epoch": 0.48018380241240666, + "grad_norm": 0.29263874888420105, + "learning_rate": 4.73025330411246e-06, + "loss": 0.6545, + "step": 1254 + }, + { + "epoch": 0.4805667241049205, + "grad_norm": 0.26484522223472595, + "learning_rate": 4.729794215777229e-06, + "loss": 0.6459, + "step": 1255 + }, + { + "epoch": 0.48094964579743443, + "grad_norm": 0.277542382478714, + "learning_rate": 4.729334759424471e-06, + "loss": 0.6722, + "step": 1256 + }, + { + "epoch": 0.4813325674899483, + "grad_norm": 0.31448233127593994, + "learning_rate": 4.728874935130015e-06, + "loss": 0.689, + "step": 1257 + }, + { + "epoch": 0.4817154891824622, + "grad_norm": 0.2727276384830475, + "learning_rate": 4.728414742969756e-06, + "loss": 0.6703, + "step": 1258 + }, + { + "epoch": 0.48209841087497607, + "grad_norm": 0.29821106791496277, + "learning_rate": 4.727954183019644e-06, + "loss": 0.6801, + "step": 1259 + }, + { + "epoch": 0.48248133256748993, + "grad_norm": 0.28618431091308594, + "learning_rate": 4.727493255355693e-06, + "loss": 0.6812, + "step": 1260 + }, + { + "epoch": 0.48286425426000384, + "grad_norm": 0.26692572236061096, + "learning_rate": 4.727031960053978e-06, + "loss": 0.6776, + "step": 1261 + }, + { + "epoch": 0.4832471759525177, + "grad_norm": 0.27408280968666077, + "learning_rate": 4.726570297190632e-06, + "loss": 0.6467, + "step": 1262 + }, + { + "epoch": 0.48363009764503156, + "grad_norm": 0.3020334839820862, + "learning_rate": 4.726108266841854e-06, + "loss": 0.657, + "step": 1263 + }, + { + "epoch": 0.4840130193375455, + "grad_norm": 0.2789212465286255, + "learning_rate": 4.725645869083895e-06, + "loss": 0.6625, + "step": 1264 + }, + { + "epoch": 0.48439594103005934, + "grad_norm": 0.27989277243614197, + "learning_rate": 4.725183103993075e-06, + "loss": 0.6611, + "step": 1265 + }, + { + "epoch": 0.48477886272257326, + "grad_norm": 0.2788684368133545, + "learning_rate": 4.724719971645771e-06, + "loss": 0.6794, + "step": 1266 + }, + { + "epoch": 0.4851617844150871, + "grad_norm": 0.2791091203689575, + "learning_rate": 4.724256472118419e-06, + "loss": 0.6688, + "step": 1267 + }, + { + "epoch": 0.485544706107601, + "grad_norm": 0.27072206139564514, + "learning_rate": 4.7237926054875206e-06, + "loss": 0.6753, + "step": 1268 + }, + { + "epoch": 0.4859276278001149, + "grad_norm": 0.27124497294425964, + "learning_rate": 4.723328371829632e-06, + "loss": 0.6501, + "step": 1269 + }, + { + "epoch": 0.48631054949262875, + "grad_norm": 0.27444323897361755, + "learning_rate": 4.7228637712213756e-06, + "loss": 0.6664, + "step": 1270 + }, + { + "epoch": 0.4866934711851426, + "grad_norm": 0.2776550054550171, + "learning_rate": 4.722398803739428e-06, + "loss": 0.6751, + "step": 1271 + }, + { + "epoch": 0.4870763928776565, + "grad_norm": 0.27999234199523926, + "learning_rate": 4.721933469460534e-06, + "loss": 0.6773, + "step": 1272 + }, + { + "epoch": 0.4874593145701704, + "grad_norm": 0.28608107566833496, + "learning_rate": 4.721467768461493e-06, + "loss": 0.6839, + "step": 1273 + }, + { + "epoch": 0.4878422362626843, + "grad_norm": 0.2772696614265442, + "learning_rate": 4.721001700819166e-06, + "loss": 0.6797, + "step": 1274 + }, + { + "epoch": 0.48822515795519816, + "grad_norm": 0.27554333209991455, + "learning_rate": 4.720535266610477e-06, + "loss": 0.7028, + "step": 1275 + }, + { + "epoch": 0.488608079647712, + "grad_norm": 0.2786296606063843, + "learning_rate": 4.720068465912407e-06, + "loss": 0.672, + "step": 1276 + }, + { + "epoch": 0.48899100134022594, + "grad_norm": 0.28190577030181885, + "learning_rate": 4.719601298802002e-06, + "loss": 0.6731, + "step": 1277 + }, + { + "epoch": 0.4893739230327398, + "grad_norm": 0.27765437960624695, + "learning_rate": 4.7191337653563635e-06, + "loss": 0.6686, + "step": 1278 + }, + { + "epoch": 0.48975684472525366, + "grad_norm": 0.28349748253822327, + "learning_rate": 4.718665865652657e-06, + "loss": 0.6928, + "step": 1279 + }, + { + "epoch": 0.4901397664177676, + "grad_norm": 0.2807316482067108, + "learning_rate": 4.7181975997681075e-06, + "loss": 0.6791, + "step": 1280 + }, + { + "epoch": 0.49052268811028144, + "grad_norm": 0.27356651425361633, + "learning_rate": 4.717728967779999e-06, + "loss": 0.6482, + "step": 1281 + }, + { + "epoch": 0.49090560980279535, + "grad_norm": 0.2867351472377777, + "learning_rate": 4.717259969765678e-06, + "loss": 0.6438, + "step": 1282 + }, + { + "epoch": 0.4912885314953092, + "grad_norm": 0.2828361392021179, + "learning_rate": 4.716790605802549e-06, + "loss": 0.6748, + "step": 1283 + }, + { + "epoch": 0.49167145318782307, + "grad_norm": 0.28004953265190125, + "learning_rate": 4.716320875968081e-06, + "loss": 0.6665, + "step": 1284 + }, + { + "epoch": 0.492054374880337, + "grad_norm": 0.27554139494895935, + "learning_rate": 4.715850780339799e-06, + "loss": 0.6589, + "step": 1285 + }, + { + "epoch": 0.49243729657285085, + "grad_norm": 0.2844911217689514, + "learning_rate": 4.7153803189952906e-06, + "loss": 0.6761, + "step": 1286 + }, + { + "epoch": 0.4928202182653647, + "grad_norm": 0.2917320132255554, + "learning_rate": 4.714909492012204e-06, + "loss": 0.6466, + "step": 1287 + }, + { + "epoch": 0.4932031399578786, + "grad_norm": 0.28606316447257996, + "learning_rate": 4.714438299468245e-06, + "loss": 0.6543, + "step": 1288 + }, + { + "epoch": 0.4935860616503925, + "grad_norm": 0.27791330218315125, + "learning_rate": 4.713966741441185e-06, + "loss": 0.6776, + "step": 1289 + }, + { + "epoch": 0.4939689833429064, + "grad_norm": 0.279191792011261, + "learning_rate": 4.713494818008849e-06, + "loss": 0.6945, + "step": 1290 + }, + { + "epoch": 0.49435190503542026, + "grad_norm": 0.28241732716560364, + "learning_rate": 4.71302252924913e-06, + "loss": 0.6591, + "step": 1291 + }, + { + "epoch": 0.4947348267279341, + "grad_norm": 0.2747727334499359, + "learning_rate": 4.712549875239972e-06, + "loss": 0.6614, + "step": 1292 + }, + { + "epoch": 0.49511774842044803, + "grad_norm": 0.2691855728626251, + "learning_rate": 4.712076856059389e-06, + "loss": 0.6235, + "step": 1293 + }, + { + "epoch": 0.4955006701129619, + "grad_norm": 0.2748763859272003, + "learning_rate": 4.711603471785449e-06, + "loss": 0.6372, + "step": 1294 + }, + { + "epoch": 0.49588359180547575, + "grad_norm": 0.28962263464927673, + "learning_rate": 4.7111297224962805e-06, + "loss": 0.648, + "step": 1295 + }, + { + "epoch": 0.49626651349798967, + "grad_norm": 0.27451568841934204, + "learning_rate": 4.710655608270076e-06, + "loss": 0.6611, + "step": 1296 + }, + { + "epoch": 0.49664943519050353, + "grad_norm": 0.28200188279151917, + "learning_rate": 4.710181129185085e-06, + "loss": 0.6583, + "step": 1297 + }, + { + "epoch": 0.49703235688301745, + "grad_norm": 0.26564377546310425, + "learning_rate": 4.7097062853196175e-06, + "loss": 0.6768, + "step": 1298 + }, + { + "epoch": 0.4974152785755313, + "grad_norm": 0.28046175837516785, + "learning_rate": 4.709231076752045e-06, + "loss": 0.6771, + "step": 1299 + }, + { + "epoch": 0.49779820026804517, + "grad_norm": 0.27810144424438477, + "learning_rate": 4.708755503560799e-06, + "loss": 0.6817, + "step": 1300 + }, + { + "epoch": 0.4981811219605591, + "grad_norm": 0.28252771496772766, + "learning_rate": 4.70827956582437e-06, + "loss": 0.6763, + "step": 1301 + }, + { + "epoch": 0.49856404365307294, + "grad_norm": 0.2781440019607544, + "learning_rate": 4.70780326362131e-06, + "loss": 0.6608, + "step": 1302 + }, + { + "epoch": 0.4989469653455868, + "grad_norm": 0.2670237123966217, + "learning_rate": 4.707326597030229e-06, + "loss": 0.6565, + "step": 1303 + }, + { + "epoch": 0.4993298870381007, + "grad_norm": 0.2824726104736328, + "learning_rate": 4.706849566129801e-06, + "loss": 0.6763, + "step": 1304 + }, + { + "epoch": 0.4997128087306146, + "grad_norm": 0.28362149000167847, + "learning_rate": 4.706372170998756e-06, + "loss": 0.668, + "step": 1305 + }, + { + "epoch": 0.5000957304231285, + "grad_norm": 0.27518826723098755, + "learning_rate": 4.705894411715887e-06, + "loss": 0.677, + "step": 1306 + }, + { + "epoch": 0.5004786521156424, + "grad_norm": 0.2708577513694763, + "learning_rate": 4.705416288360045e-06, + "loss": 0.6512, + "step": 1307 + }, + { + "epoch": 0.5008615738081562, + "grad_norm": 0.27491506934165955, + "learning_rate": 4.704937801010144e-06, + "loss": 0.6766, + "step": 1308 + }, + { + "epoch": 0.5012444955006701, + "grad_norm": 0.3019718527793884, + "learning_rate": 4.704458949745154e-06, + "loss": 0.6728, + "step": 1309 + }, + { + "epoch": 0.501627417193184, + "grad_norm": 0.2824283838272095, + "learning_rate": 4.703979734644107e-06, + "loss": 0.6695, + "step": 1310 + }, + { + "epoch": 0.5020103388856979, + "grad_norm": 0.2763907015323639, + "learning_rate": 4.703500155786098e-06, + "loss": 0.6738, + "step": 1311 + }, + { + "epoch": 0.5023932605782118, + "grad_norm": 0.27073514461517334, + "learning_rate": 4.703020213250278e-06, + "loss": 0.6664, + "step": 1312 + }, + { + "epoch": 0.5027761822707256, + "grad_norm": 0.2841322720050812, + "learning_rate": 4.7025399071158575e-06, + "loss": 0.6603, + "step": 1313 + }, + { + "epoch": 0.5031591039632395, + "grad_norm": 0.28222954273223877, + "learning_rate": 4.702059237462112e-06, + "loss": 0.6768, + "step": 1314 + }, + { + "epoch": 0.5035420256557533, + "grad_norm": 0.28418296575546265, + "learning_rate": 4.701578204368371e-06, + "loss": 0.677, + "step": 1315 + }, + { + "epoch": 0.5039249473482673, + "grad_norm": 0.2761109173297882, + "learning_rate": 4.70109680791403e-06, + "loss": 0.6829, + "step": 1316 + }, + { + "epoch": 0.5043078690407812, + "grad_norm": 0.2744296193122864, + "learning_rate": 4.700615048178538e-06, + "loss": 0.6865, + "step": 1317 + }, + { + "epoch": 0.504690790733295, + "grad_norm": 0.2845577299594879, + "learning_rate": 4.7001329252414095e-06, + "loss": 0.6632, + "step": 1318 + }, + { + "epoch": 0.5050737124258089, + "grad_norm": 0.28562062978744507, + "learning_rate": 4.699650439182216e-06, + "loss": 0.6737, + "step": 1319 + }, + { + "epoch": 0.5054566341183228, + "grad_norm": 0.28627490997314453, + "learning_rate": 4.69916759008059e-06, + "loss": 0.655, + "step": 1320 + }, + { + "epoch": 0.5058395558108367, + "grad_norm": 0.27140292525291443, + "learning_rate": 4.698684378016223e-06, + "loss": 0.6518, + "step": 1321 + }, + { + "epoch": 0.5062224775033506, + "grad_norm": 0.2792279124259949, + "learning_rate": 4.698200803068868e-06, + "loss": 0.6855, + "step": 1322 + }, + { + "epoch": 0.5066053991958644, + "grad_norm": 0.2705386281013489, + "learning_rate": 4.697716865318335e-06, + "loss": 0.6755, + "step": 1323 + }, + { + "epoch": 0.5069883208883783, + "grad_norm": 0.2843645215034485, + "learning_rate": 4.6972325648444974e-06, + "loss": 0.6692, + "step": 1324 + }, + { + "epoch": 0.5073712425808922, + "grad_norm": 0.28340429067611694, + "learning_rate": 4.6967479017272865e-06, + "loss": 0.6628, + "step": 1325 + }, + { + "epoch": 0.5077541642734061, + "grad_norm": 0.2885495722293854, + "learning_rate": 4.696262876046693e-06, + "loss": 0.6676, + "step": 1326 + }, + { + "epoch": 0.50813708596592, + "grad_norm": 0.2888467609882355, + "learning_rate": 4.6957774878827694e-06, + "loss": 0.678, + "step": 1327 + }, + { + "epoch": 0.5085200076584339, + "grad_norm": 0.28116628527641296, + "learning_rate": 4.695291737315626e-06, + "loss": 0.6849, + "step": 1328 + }, + { + "epoch": 0.5089029293509477, + "grad_norm": 0.29132142663002014, + "learning_rate": 4.694805624425435e-06, + "loss": 0.6769, + "step": 1329 + }, + { + "epoch": 0.5092858510434616, + "grad_norm": 0.2741115391254425, + "learning_rate": 4.6943191492924246e-06, + "loss": 0.6757, + "step": 1330 + }, + { + "epoch": 0.5096687727359754, + "grad_norm": 0.28962674736976624, + "learning_rate": 4.693832311996887e-06, + "loss": 0.6501, + "step": 1331 + }, + { + "epoch": 0.5100516944284894, + "grad_norm": 0.2865285873413086, + "learning_rate": 4.693345112619173e-06, + "loss": 0.6683, + "step": 1332 + }, + { + "epoch": 0.5104346161210033, + "grad_norm": 0.2814137935638428, + "learning_rate": 4.692857551239692e-06, + "loss": 0.6767, + "step": 1333 + }, + { + "epoch": 0.5108175378135171, + "grad_norm": 0.2837572991847992, + "learning_rate": 4.692369627938914e-06, + "loss": 0.6554, + "step": 1334 + }, + { + "epoch": 0.511200459506031, + "grad_norm": 0.2752167284488678, + "learning_rate": 4.691881342797368e-06, + "loss": 0.6835, + "step": 1335 + }, + { + "epoch": 0.5115833811985449, + "grad_norm": 0.2947768270969391, + "learning_rate": 4.691392695895645e-06, + "loss": 0.6759, + "step": 1336 + }, + { + "epoch": 0.5119663028910588, + "grad_norm": 0.30000224709510803, + "learning_rate": 4.690903687314391e-06, + "loss": 0.6697, + "step": 1337 + }, + { + "epoch": 0.5123492245835727, + "grad_norm": 0.2816091477870941, + "learning_rate": 4.690414317134318e-06, + "loss": 0.6646, + "step": 1338 + }, + { + "epoch": 0.5127321462760865, + "grad_norm": 0.2815537750720978, + "learning_rate": 4.689924585436192e-06, + "loss": 0.6631, + "step": 1339 + }, + { + "epoch": 0.5131150679686004, + "grad_norm": 0.29559898376464844, + "learning_rate": 4.689434492300841e-06, + "loss": 0.669, + "step": 1340 + }, + { + "epoch": 0.5134979896611143, + "grad_norm": 0.28763526678085327, + "learning_rate": 4.688944037809154e-06, + "loss": 0.6675, + "step": 1341 + }, + { + "epoch": 0.5138809113536282, + "grad_norm": 0.2929317355155945, + "learning_rate": 4.688453222042077e-06, + "loss": 0.7009, + "step": 1342 + }, + { + "epoch": 0.5142638330461421, + "grad_norm": 0.2719643712043762, + "learning_rate": 4.687962045080619e-06, + "loss": 0.6697, + "step": 1343 + }, + { + "epoch": 0.514646754738656, + "grad_norm": 0.299638569355011, + "learning_rate": 4.687470507005845e-06, + "loss": 0.6644, + "step": 1344 + }, + { + "epoch": 0.5150296764311698, + "grad_norm": 0.2654574513435364, + "learning_rate": 4.686978607898881e-06, + "loss": 0.6745, + "step": 1345 + }, + { + "epoch": 0.5154125981236837, + "grad_norm": 0.2910325527191162, + "learning_rate": 4.686486347840913e-06, + "loss": 0.6711, + "step": 1346 + }, + { + "epoch": 0.5157955198161975, + "grad_norm": 0.3001216650009155, + "learning_rate": 4.685993726913186e-06, + "loss": 0.6611, + "step": 1347 + }, + { + "epoch": 0.5161784415087115, + "grad_norm": 0.31864020228385925, + "learning_rate": 4.685500745197006e-06, + "loss": 0.6557, + "step": 1348 + }, + { + "epoch": 0.5165613632012254, + "grad_norm": 0.31896722316741943, + "learning_rate": 4.685007402773737e-06, + "loss": 0.6679, + "step": 1349 + }, + { + "epoch": 0.5169442848937392, + "grad_norm": 0.28124380111694336, + "learning_rate": 4.684513699724802e-06, + "loss": 0.6614, + "step": 1350 + }, + { + "epoch": 0.5173272065862531, + "grad_norm": 0.28782278299331665, + "learning_rate": 4.684019636131685e-06, + "loss": 0.6586, + "step": 1351 + }, + { + "epoch": 0.517710128278767, + "grad_norm": 0.2993348240852356, + "learning_rate": 4.6835252120759286e-06, + "loss": 0.662, + "step": 1352 + }, + { + "epoch": 0.5180930499712809, + "grad_norm": 0.2770722806453705, + "learning_rate": 4.683030427639137e-06, + "loss": 0.652, + "step": 1353 + }, + { + "epoch": 0.5184759716637948, + "grad_norm": 0.2981843054294586, + "learning_rate": 4.6825352829029705e-06, + "loss": 0.6851, + "step": 1354 + }, + { + "epoch": 0.5188588933563086, + "grad_norm": 0.2957566976547241, + "learning_rate": 4.682039777949151e-06, + "loss": 0.6664, + "step": 1355 + }, + { + "epoch": 0.5192418150488225, + "grad_norm": 0.293414831161499, + "learning_rate": 4.68154391285946e-06, + "loss": 0.684, + "step": 1356 + }, + { + "epoch": 0.5196247367413364, + "grad_norm": 0.2782313823699951, + "learning_rate": 4.6810476877157365e-06, + "loss": 0.6713, + "step": 1357 + }, + { + "epoch": 0.5200076584338503, + "grad_norm": 0.286119282245636, + "learning_rate": 4.680551102599881e-06, + "loss": 0.6933, + "step": 1358 + }, + { + "epoch": 0.5203905801263642, + "grad_norm": 0.3102959096431732, + "learning_rate": 4.680054157593852e-06, + "loss": 0.6777, + "step": 1359 + }, + { + "epoch": 0.520773501818878, + "grad_norm": 0.2751920521259308, + "learning_rate": 4.679556852779668e-06, + "loss": 0.6723, + "step": 1360 + }, + { + "epoch": 0.5211564235113919, + "grad_norm": 0.28185367584228516, + "learning_rate": 4.679059188239408e-06, + "loss": 0.6609, + "step": 1361 + }, + { + "epoch": 0.5215393452039058, + "grad_norm": 0.31509310007095337, + "learning_rate": 4.67856116405521e-06, + "loss": 0.6618, + "step": 1362 + }, + { + "epoch": 0.5219222668964196, + "grad_norm": 0.27555981278419495, + "learning_rate": 4.678062780309268e-06, + "loss": 0.6678, + "step": 1363 + }, + { + "epoch": 0.5223051885889336, + "grad_norm": 0.2945083677768707, + "learning_rate": 4.677564037083841e-06, + "loss": 0.6553, + "step": 1364 + }, + { + "epoch": 0.5226881102814475, + "grad_norm": 0.28949201107025146, + "learning_rate": 4.677064934461241e-06, + "loss": 0.6602, + "step": 1365 + }, + { + "epoch": 0.5230710319739613, + "grad_norm": 0.3095218241214752, + "learning_rate": 4.676565472523845e-06, + "loss": 0.6551, + "step": 1366 + }, + { + "epoch": 0.5234539536664752, + "grad_norm": 0.2740354537963867, + "learning_rate": 4.676065651354087e-06, + "loss": 0.6821, + "step": 1367 + }, + { + "epoch": 0.523836875358989, + "grad_norm": 0.29293930530548096, + "learning_rate": 4.6755654710344585e-06, + "loss": 0.6524, + "step": 1368 + }, + { + "epoch": 0.524219797051503, + "grad_norm": 0.289955198764801, + "learning_rate": 4.675064931647514e-06, + "loss": 0.6827, + "step": 1369 + }, + { + "epoch": 0.5246027187440169, + "grad_norm": 0.29977336525917053, + "learning_rate": 4.674564033275864e-06, + "loss": 0.6796, + "step": 1370 + }, + { + "epoch": 0.5249856404365307, + "grad_norm": 0.2801453471183777, + "learning_rate": 4.674062776002181e-06, + "loss": 0.6512, + "step": 1371 + }, + { + "epoch": 0.5253685621290446, + "grad_norm": 0.3363115191459656, + "learning_rate": 4.673561159909192e-06, + "loss": 0.6616, + "step": 1372 + }, + { + "epoch": 0.5257514838215585, + "grad_norm": 0.2854152321815491, + "learning_rate": 4.67305918507969e-06, + "loss": 0.6748, + "step": 1373 + }, + { + "epoch": 0.5261344055140724, + "grad_norm": 0.2665503919124603, + "learning_rate": 4.672556851596522e-06, + "loss": 0.6601, + "step": 1374 + }, + { + "epoch": 0.5265173272065863, + "grad_norm": 0.31339603662490845, + "learning_rate": 4.672054159542597e-06, + "loss": 0.661, + "step": 1375 + }, + { + "epoch": 0.5269002488991001, + "grad_norm": 0.2875712513923645, + "learning_rate": 4.671551109000882e-06, + "loss": 0.6817, + "step": 1376 + }, + { + "epoch": 0.527283170591614, + "grad_norm": 0.27537474036216736, + "learning_rate": 4.671047700054401e-06, + "loss": 0.6596, + "step": 1377 + }, + { + "epoch": 0.5276660922841279, + "grad_norm": 0.29016685485839844, + "learning_rate": 4.6705439327862415e-06, + "loss": 0.6781, + "step": 1378 + }, + { + "epoch": 0.5280490139766417, + "grad_norm": 0.3029150664806366, + "learning_rate": 4.670039807279547e-06, + "loss": 0.6677, + "step": 1379 + }, + { + "epoch": 0.5284319356691557, + "grad_norm": 0.28253310918807983, + "learning_rate": 4.669535323617522e-06, + "loss": 0.6905, + "step": 1380 + }, + { + "epoch": 0.5288148573616696, + "grad_norm": 0.29214736819267273, + "learning_rate": 4.669030481883429e-06, + "loss": 0.6606, + "step": 1381 + }, + { + "epoch": 0.5291977790541834, + "grad_norm": 0.28440895676612854, + "learning_rate": 4.66852528216059e-06, + "loss": 0.6699, + "step": 1382 + }, + { + "epoch": 0.5295807007466973, + "grad_norm": 0.295693963766098, + "learning_rate": 4.668019724532385e-06, + "loss": 0.6803, + "step": 1383 + }, + { + "epoch": 0.5299636224392111, + "grad_norm": 0.28351619839668274, + "learning_rate": 4.667513809082255e-06, + "loss": 0.6652, + "step": 1384 + }, + { + "epoch": 0.5303465441317251, + "grad_norm": 0.2787420153617859, + "learning_rate": 4.667007535893698e-06, + "loss": 0.6753, + "step": 1385 + }, + { + "epoch": 0.530729465824239, + "grad_norm": 0.2919510006904602, + "learning_rate": 4.666500905050274e-06, + "loss": 0.6764, + "step": 1386 + }, + { + "epoch": 0.5311123875167528, + "grad_norm": 0.2776666283607483, + "learning_rate": 4.665993916635599e-06, + "loss": 0.6651, + "step": 1387 + }, + { + "epoch": 0.5314953092092667, + "grad_norm": 0.2779915928840637, + "learning_rate": 4.66548657073335e-06, + "loss": 0.6683, + "step": 1388 + }, + { + "epoch": 0.5318782309017805, + "grad_norm": 0.28630954027175903, + "learning_rate": 4.66497886742726e-06, + "loss": 0.6743, + "step": 1389 + }, + { + "epoch": 0.5322611525942945, + "grad_norm": 0.2745286822319031, + "learning_rate": 4.664470806801125e-06, + "loss": 0.6704, + "step": 1390 + }, + { + "epoch": 0.5326440742868084, + "grad_norm": 0.28013500571250916, + "learning_rate": 4.663962388938797e-06, + "loss": 0.6716, + "step": 1391 + }, + { + "epoch": 0.5330269959793222, + "grad_norm": 0.2824057936668396, + "learning_rate": 4.66345361392419e-06, + "loss": 0.6646, + "step": 1392 + }, + { + "epoch": 0.5334099176718361, + "grad_norm": 0.2962542474269867, + "learning_rate": 4.662944481841273e-06, + "loss": 0.6611, + "step": 1393 + }, + { + "epoch": 0.53379283936435, + "grad_norm": 0.28005772829055786, + "learning_rate": 4.662434992774077e-06, + "loss": 0.6657, + "step": 1394 + }, + { + "epoch": 0.5341757610568638, + "grad_norm": 0.2697499394416809, + "learning_rate": 4.66192514680669e-06, + "loss": 0.6613, + "step": 1395 + }, + { + "epoch": 0.5345586827493778, + "grad_norm": 0.29074200987815857, + "learning_rate": 4.661414944023261e-06, + "loss": 0.6743, + "step": 1396 + }, + { + "epoch": 0.5349416044418916, + "grad_norm": 0.28916990756988525, + "learning_rate": 4.660904384507996e-06, + "loss": 0.6538, + "step": 1397 + }, + { + "epoch": 0.5353245261344055, + "grad_norm": 0.27587875723838806, + "learning_rate": 4.66039346834516e-06, + "loss": 0.6739, + "step": 1398 + }, + { + "epoch": 0.5357074478269194, + "grad_norm": 0.27750566601753235, + "learning_rate": 4.659882195619079e-06, + "loss": 0.6638, + "step": 1399 + }, + { + "epoch": 0.5360903695194332, + "grad_norm": 0.2798866927623749, + "learning_rate": 4.6593705664141356e-06, + "loss": 0.6735, + "step": 1400 + }, + { + "epoch": 0.5364732912119472, + "grad_norm": 0.27637436985969543, + "learning_rate": 4.658858580814772e-06, + "loss": 0.6794, + "step": 1401 + }, + { + "epoch": 0.5368562129044611, + "grad_norm": 0.27356451749801636, + "learning_rate": 4.658346238905487e-06, + "loss": 0.6762, + "step": 1402 + }, + { + "epoch": 0.5372391345969749, + "grad_norm": 0.2700641453266144, + "learning_rate": 4.657833540770843e-06, + "loss": 0.6445, + "step": 1403 + }, + { + "epoch": 0.5376220562894888, + "grad_norm": 0.286549836397171, + "learning_rate": 4.657320486495458e-06, + "loss": 0.6895, + "step": 1404 + }, + { + "epoch": 0.5380049779820026, + "grad_norm": 0.29119592905044556, + "learning_rate": 4.656807076164008e-06, + "loss": 0.6788, + "step": 1405 + }, + { + "epoch": 0.5383878996745166, + "grad_norm": 0.2834751307964325, + "learning_rate": 4.656293309861231e-06, + "loss": 0.6681, + "step": 1406 + }, + { + "epoch": 0.5387708213670305, + "grad_norm": 0.27905791997909546, + "learning_rate": 4.65577918767192e-06, + "loss": 0.6561, + "step": 1407 + }, + { + "epoch": 0.5391537430595443, + "grad_norm": 0.26963064074516296, + "learning_rate": 4.65526470968093e-06, + "loss": 0.6637, + "step": 1408 + }, + { + "epoch": 0.5395366647520582, + "grad_norm": 0.29435136914253235, + "learning_rate": 4.6547498759731725e-06, + "loss": 0.6698, + "step": 1409 + }, + { + "epoch": 0.539919586444572, + "grad_norm": 0.2793678045272827, + "learning_rate": 4.654234686633619e-06, + "loss": 0.6593, + "step": 1410 + }, + { + "epoch": 0.5403025081370859, + "grad_norm": 0.29543066024780273, + "learning_rate": 4.653719141747297e-06, + "loss": 0.6729, + "step": 1411 + }, + { + "epoch": 0.5406854298295999, + "grad_norm": 0.2844732105731964, + "learning_rate": 4.653203241399299e-06, + "loss": 0.6711, + "step": 1412 + }, + { + "epoch": 0.5410683515221137, + "grad_norm": 0.29418832063674927, + "learning_rate": 4.652686985674769e-06, + "loss": 0.6463, + "step": 1413 + }, + { + "epoch": 0.5414512732146276, + "grad_norm": 0.2700132131576538, + "learning_rate": 4.652170374658913e-06, + "loss": 0.6678, + "step": 1414 + }, + { + "epoch": 0.5418341949071415, + "grad_norm": 0.2833193242549896, + "learning_rate": 4.651653408436996e-06, + "loss": 0.662, + "step": 1415 + }, + { + "epoch": 0.5422171165996553, + "grad_norm": 0.28621354699134827, + "learning_rate": 4.651136087094341e-06, + "loss": 0.6552, + "step": 1416 + }, + { + "epoch": 0.5426000382921693, + "grad_norm": 0.28285256028175354, + "learning_rate": 4.65061841071633e-06, + "loss": 0.6707, + "step": 1417 + }, + { + "epoch": 0.5429829599846832, + "grad_norm": 0.29332616925239563, + "learning_rate": 4.650100379388401e-06, + "loss": 0.6639, + "step": 1418 + }, + { + "epoch": 0.543365881677197, + "grad_norm": 0.27814438939094543, + "learning_rate": 4.649581993196055e-06, + "loss": 0.6575, + "step": 1419 + }, + { + "epoch": 0.5437488033697109, + "grad_norm": 0.2910020649433136, + "learning_rate": 4.6490632522248485e-06, + "loss": 0.6731, + "step": 1420 + }, + { + "epoch": 0.5441317250622247, + "grad_norm": 0.2723335325717926, + "learning_rate": 4.648544156560397e-06, + "loss": 0.6646, + "step": 1421 + }, + { + "epoch": 0.5445146467547387, + "grad_norm": 0.2737279534339905, + "learning_rate": 4.648024706288376e-06, + "loss": 0.6533, + "step": 1422 + }, + { + "epoch": 0.5448975684472526, + "grad_norm": 0.2696785032749176, + "learning_rate": 4.647504901494517e-06, + "loss": 0.6515, + "step": 1423 + }, + { + "epoch": 0.5452804901397664, + "grad_norm": 0.280367374420166, + "learning_rate": 4.646984742264613e-06, + "loss": 0.6497, + "step": 1424 + }, + { + "epoch": 0.5456634118322803, + "grad_norm": 0.2739644944667816, + "learning_rate": 4.646464228684512e-06, + "loss": 0.6475, + "step": 1425 + }, + { + "epoch": 0.5460463335247941, + "grad_norm": 0.27324730157852173, + "learning_rate": 4.645943360840124e-06, + "loss": 0.6874, + "step": 1426 + }, + { + "epoch": 0.546429255217308, + "grad_norm": 0.28439703583717346, + "learning_rate": 4.6454221388174165e-06, + "loss": 0.6622, + "step": 1427 + }, + { + "epoch": 0.546812176909822, + "grad_norm": 0.2706713378429413, + "learning_rate": 4.6449005627024126e-06, + "loss": 0.6481, + "step": 1428 + }, + { + "epoch": 0.5471950986023358, + "grad_norm": 0.28453972935676575, + "learning_rate": 4.644378632581198e-06, + "loss": 0.668, + "step": 1429 + }, + { + "epoch": 0.5475780202948497, + "grad_norm": 0.28511837124824524, + "learning_rate": 4.643856348539913e-06, + "loss": 0.6736, + "step": 1430 + }, + { + "epoch": 0.5479609419873636, + "grad_norm": 0.2903471291065216, + "learning_rate": 4.643333710664761e-06, + "loss": 0.659, + "step": 1431 + }, + { + "epoch": 0.5483438636798774, + "grad_norm": 0.28712940216064453, + "learning_rate": 4.642810719041999e-06, + "loss": 0.6664, + "step": 1432 + }, + { + "epoch": 0.5487267853723914, + "grad_norm": 0.27602365612983704, + "learning_rate": 4.642287373757946e-06, + "loss": 0.675, + "step": 1433 + }, + { + "epoch": 0.5491097070649053, + "grad_norm": 0.29832765460014343, + "learning_rate": 4.641763674898975e-06, + "loss": 0.6763, + "step": 1434 + }, + { + "epoch": 0.5494926287574191, + "grad_norm": 0.2785443663597107, + "learning_rate": 4.641239622551522e-06, + "loss": 0.6493, + "step": 1435 + }, + { + "epoch": 0.549875550449933, + "grad_norm": 0.28630906343460083, + "learning_rate": 4.64071521680208e-06, + "loss": 0.6475, + "step": 1436 + }, + { + "epoch": 0.5502584721424468, + "grad_norm": 0.27429166436195374, + "learning_rate": 4.6401904577371995e-06, + "loss": 0.6491, + "step": 1437 + }, + { + "epoch": 0.5506413938349608, + "grad_norm": 0.2856921851634979, + "learning_rate": 4.6396653454434884e-06, + "loss": 0.6831, + "step": 1438 + }, + { + "epoch": 0.5510243155274747, + "grad_norm": 0.2848499119281769, + "learning_rate": 4.639139880007615e-06, + "loss": 0.6703, + "step": 1439 + }, + { + "epoch": 0.5514072372199885, + "grad_norm": 0.2933283746242523, + "learning_rate": 4.6386140615163056e-06, + "loss": 0.6472, + "step": 1440 + }, + { + "epoch": 0.5517901589125024, + "grad_norm": 0.28132501244544983, + "learning_rate": 4.638087890056343e-06, + "loss": 0.6666, + "step": 1441 + }, + { + "epoch": 0.5521730806050162, + "grad_norm": 0.29337993264198303, + "learning_rate": 4.637561365714571e-06, + "loss": 0.6824, + "step": 1442 + }, + { + "epoch": 0.5525560022975301, + "grad_norm": 0.2802111804485321, + "learning_rate": 4.637034488577889e-06, + "loss": 0.6486, + "step": 1443 + }, + { + "epoch": 0.5529389239900441, + "grad_norm": 0.27604779601097107, + "learning_rate": 4.636507258733255e-06, + "loss": 0.655, + "step": 1444 + }, + { + "epoch": 0.5533218456825579, + "grad_norm": 0.2832767069339752, + "learning_rate": 4.6359796762676875e-06, + "loss": 0.6719, + "step": 1445 + }, + { + "epoch": 0.5537047673750718, + "grad_norm": 0.2880399227142334, + "learning_rate": 4.6354517412682605e-06, + "loss": 0.6646, + "step": 1446 + }, + { + "epoch": 0.5540876890675857, + "grad_norm": 0.2932654619216919, + "learning_rate": 4.634923453822108e-06, + "loss": 0.6746, + "step": 1447 + }, + { + "epoch": 0.5544706107600995, + "grad_norm": 0.2653713822364807, + "learning_rate": 4.634394814016421e-06, + "loss": 0.6555, + "step": 1448 + }, + { + "epoch": 0.5548535324526135, + "grad_norm": 0.2792087495326996, + "learning_rate": 4.633865821938449e-06, + "loss": 0.6799, + "step": 1449 + }, + { + "epoch": 0.5552364541451273, + "grad_norm": 0.28352558612823486, + "learning_rate": 4.6333364776755006e-06, + "loss": 0.6632, + "step": 1450 + }, + { + "epoch": 0.5556193758376412, + "grad_norm": 0.2973845303058624, + "learning_rate": 4.632806781314941e-06, + "loss": 0.664, + "step": 1451 + }, + { + "epoch": 0.5560022975301551, + "grad_norm": 0.2806834578514099, + "learning_rate": 4.632276732944194e-06, + "loss": 0.6655, + "step": 1452 + }, + { + "epoch": 0.5563852192226689, + "grad_norm": 0.2828749120235443, + "learning_rate": 4.631746332650743e-06, + "loss": 0.6776, + "step": 1453 + }, + { + "epoch": 0.5567681409151829, + "grad_norm": 0.28335845470428467, + "learning_rate": 4.631215580522126e-06, + "loss": 0.6798, + "step": 1454 + }, + { + "epoch": 0.5571510626076968, + "grad_norm": 0.2798022925853729, + "learning_rate": 4.6306844766459435e-06, + "loss": 0.6655, + "step": 1455 + }, + { + "epoch": 0.5575339843002106, + "grad_norm": 0.284993976354599, + "learning_rate": 4.63015302110985e-06, + "loss": 0.668, + "step": 1456 + }, + { + "epoch": 0.5579169059927245, + "grad_norm": 0.2809392213821411, + "learning_rate": 4.629621214001562e-06, + "loss": 0.6799, + "step": 1457 + }, + { + "epoch": 0.5582998276852383, + "grad_norm": 0.2810289263725281, + "learning_rate": 4.6290890554088485e-06, + "loss": 0.6751, + "step": 1458 + }, + { + "epoch": 0.5586827493777522, + "grad_norm": 0.2804025709629059, + "learning_rate": 4.628556545419543e-06, + "loss": 0.6704, + "step": 1459 + }, + { + "epoch": 0.5590656710702662, + "grad_norm": 0.2831740379333496, + "learning_rate": 4.6280236841215335e-06, + "loss": 0.6615, + "step": 1460 + }, + { + "epoch": 0.55944859276278, + "grad_norm": 0.2875085175037384, + "learning_rate": 4.627490471602765e-06, + "loss": 0.6644, + "step": 1461 + }, + { + "epoch": 0.5598315144552939, + "grad_norm": 0.27246806025505066, + "learning_rate": 4.626956907951242e-06, + "loss": 0.642, + "step": 1462 + }, + { + "epoch": 0.5602144361478077, + "grad_norm": 0.2839241325855255, + "learning_rate": 4.626422993255028e-06, + "loss": 0.6532, + "step": 1463 + }, + { + "epoch": 0.5605973578403216, + "grad_norm": 0.276871919631958, + "learning_rate": 4.6258887276022425e-06, + "loss": 0.663, + "step": 1464 + }, + { + "epoch": 0.5609802795328356, + "grad_norm": 0.2767009139060974, + "learning_rate": 4.6253541110810635e-06, + "loss": 0.6735, + "step": 1465 + }, + { + "epoch": 0.5613632012253494, + "grad_norm": 0.28140896558761597, + "learning_rate": 4.6248191437797275e-06, + "loss": 0.6337, + "step": 1466 + }, + { + "epoch": 0.5617461229178633, + "grad_norm": 0.2746315896511078, + "learning_rate": 4.624283825786528e-06, + "loss": 0.6402, + "step": 1467 + }, + { + "epoch": 0.5621290446103772, + "grad_norm": 0.2852254807949066, + "learning_rate": 4.623748157189817e-06, + "loss": 0.6781, + "step": 1468 + }, + { + "epoch": 0.562511966302891, + "grad_norm": 0.2737531363964081, + "learning_rate": 4.623212138078004e-06, + "loss": 0.6339, + "step": 1469 + }, + { + "epoch": 0.562894887995405, + "grad_norm": 0.2868412137031555, + "learning_rate": 4.622675768539557e-06, + "loss": 0.6779, + "step": 1470 + }, + { + "epoch": 0.5632778096879189, + "grad_norm": 0.2785370647907257, + "learning_rate": 4.622139048663001e-06, + "loss": 0.676, + "step": 1471 + }, + { + "epoch": 0.5636607313804327, + "grad_norm": 0.29825419187545776, + "learning_rate": 4.621601978536919e-06, + "loss": 0.679, + "step": 1472 + }, + { + "epoch": 0.5640436530729466, + "grad_norm": 0.2933095693588257, + "learning_rate": 4.6210645582499536e-06, + "loss": 0.651, + "step": 1473 + }, + { + "epoch": 0.5644265747654604, + "grad_norm": 0.27432429790496826, + "learning_rate": 4.620526787890801e-06, + "loss": 0.6734, + "step": 1474 + }, + { + "epoch": 0.5648094964579743, + "grad_norm": 0.29858604073524475, + "learning_rate": 4.6199886675482195e-06, + "loss": 0.6762, + "step": 1475 + }, + { + "epoch": 0.5651924181504883, + "grad_norm": 0.2854273021221161, + "learning_rate": 4.619450197311024e-06, + "loss": 0.6666, + "step": 1476 + }, + { + "epoch": 0.5655753398430021, + "grad_norm": 0.2858935296535492, + "learning_rate": 4.618911377268085e-06, + "loss": 0.656, + "step": 1477 + }, + { + "epoch": 0.565958261535516, + "grad_norm": 0.2800264358520508, + "learning_rate": 4.618372207508333e-06, + "loss": 0.6727, + "step": 1478 + }, + { + "epoch": 0.5663411832280298, + "grad_norm": 0.2903977334499359, + "learning_rate": 4.617832688120756e-06, + "loss": 0.6712, + "step": 1479 + }, + { + "epoch": 0.5667241049205437, + "grad_norm": 0.30259647965431213, + "learning_rate": 4.617292819194398e-06, + "loss": 0.6776, + "step": 1480 + }, + { + "epoch": 0.5671070266130577, + "grad_norm": 0.2795424461364746, + "learning_rate": 4.616752600818363e-06, + "loss": 0.6726, + "step": 1481 + }, + { + "epoch": 0.5674899483055715, + "grad_norm": 0.2867182195186615, + "learning_rate": 4.616212033081811e-06, + "loss": 0.6604, + "step": 1482 + }, + { + "epoch": 0.5678728699980854, + "grad_norm": 0.2813325822353363, + "learning_rate": 4.615671116073962e-06, + "loss": 0.6661, + "step": 1483 + }, + { + "epoch": 0.5682557916905993, + "grad_norm": 0.29366227984428406, + "learning_rate": 4.61512984988409e-06, + "loss": 0.6698, + "step": 1484 + }, + { + "epoch": 0.5686387133831131, + "grad_norm": 0.271113783121109, + "learning_rate": 4.614588234601529e-06, + "loss": 0.6711, + "step": 1485 + }, + { + "epoch": 0.5690216350756271, + "grad_norm": 0.2789374887943268, + "learning_rate": 4.614046270315671e-06, + "loss": 0.6735, + "step": 1486 + }, + { + "epoch": 0.569404556768141, + "grad_norm": 0.2933051884174347, + "learning_rate": 4.613503957115963e-06, + "loss": 0.6819, + "step": 1487 + }, + { + "epoch": 0.5697874784606548, + "grad_norm": 0.2969438135623932, + "learning_rate": 4.612961295091913e-06, + "loss": 0.6719, + "step": 1488 + }, + { + "epoch": 0.5701704001531687, + "grad_norm": 0.2947516143321991, + "learning_rate": 4.612418284333085e-06, + "loss": 0.6706, + "step": 1489 + }, + { + "epoch": 0.5705533218456825, + "grad_norm": 0.30186280608177185, + "learning_rate": 4.6118749249290985e-06, + "loss": 0.6443, + "step": 1490 + }, + { + "epoch": 0.5709362435381964, + "grad_norm": 0.2903065085411072, + "learning_rate": 4.611331216969634e-06, + "loss": 0.6672, + "step": 1491 + }, + { + "epoch": 0.5713191652307104, + "grad_norm": 0.2867312431335449, + "learning_rate": 4.610787160544429e-06, + "loss": 0.6658, + "step": 1492 + }, + { + "epoch": 0.5717020869232242, + "grad_norm": 0.2932186424732208, + "learning_rate": 4.610242755743276e-06, + "loss": 0.6628, + "step": 1493 + }, + { + "epoch": 0.5720850086157381, + "grad_norm": 0.2865498960018158, + "learning_rate": 4.609698002656027e-06, + "loss": 0.6605, + "step": 1494 + }, + { + "epoch": 0.5724679303082519, + "grad_norm": 0.2915932238101959, + "learning_rate": 4.609152901372592e-06, + "loss": 0.6669, + "step": 1495 + }, + { + "epoch": 0.5728508520007658, + "grad_norm": 0.291251540184021, + "learning_rate": 4.608607451982936e-06, + "loss": 0.6642, + "step": 1496 + }, + { + "epoch": 0.5732337736932798, + "grad_norm": 0.28860780596733093, + "learning_rate": 4.608061654577084e-06, + "loss": 0.6716, + "step": 1497 + }, + { + "epoch": 0.5736166953857936, + "grad_norm": 0.2796918749809265, + "learning_rate": 4.607515509245116e-06, + "loss": 0.6678, + "step": 1498 + }, + { + "epoch": 0.5739996170783075, + "grad_norm": 0.2800701856613159, + "learning_rate": 4.606969016077173e-06, + "loss": 0.66, + "step": 1499 + }, + { + "epoch": 0.5743825387708213, + "grad_norm": 0.2912375032901764, + "learning_rate": 4.6064221751634495e-06, + "loss": 0.657, + "step": 1500 + }, + { + "epoch": 0.5747654604633352, + "grad_norm": 0.29719606041908264, + "learning_rate": 4.6058749865942e-06, + "loss": 0.6662, + "step": 1501 + }, + { + "epoch": 0.5751483821558492, + "grad_norm": 0.2698865532875061, + "learning_rate": 4.605327450459736e-06, + "loss": 0.6733, + "step": 1502 + }, + { + "epoch": 0.575531303848363, + "grad_norm": 0.29630282521247864, + "learning_rate": 4.604779566850425e-06, + "loss": 0.6667, + "step": 1503 + }, + { + "epoch": 0.5759142255408769, + "grad_norm": 0.291897714138031, + "learning_rate": 4.604231335856692e-06, + "loss": 0.6604, + "step": 1504 + }, + { + "epoch": 0.5762971472333908, + "grad_norm": 0.278290718793869, + "learning_rate": 4.603682757569021e-06, + "loss": 0.6448, + "step": 1505 + }, + { + "epoch": 0.5766800689259046, + "grad_norm": 0.2939246594905853, + "learning_rate": 4.603133832077953e-06, + "loss": 0.6592, + "step": 1506 + }, + { + "epoch": 0.5770629906184185, + "grad_norm": 0.2818027138710022, + "learning_rate": 4.602584559474085e-06, + "loss": 0.6665, + "step": 1507 + }, + { + "epoch": 0.5774459123109325, + "grad_norm": 0.2802014946937561, + "learning_rate": 4.602034939848072e-06, + "loss": 0.6741, + "step": 1508 + }, + { + "epoch": 0.5778288340034463, + "grad_norm": 0.2793828547000885, + "learning_rate": 4.601484973290627e-06, + "loss": 0.6701, + "step": 1509 + }, + { + "epoch": 0.5782117556959602, + "grad_norm": 0.29030466079711914, + "learning_rate": 4.600934659892518e-06, + "loss": 0.6631, + "step": 1510 + }, + { + "epoch": 0.578594677388474, + "grad_norm": 0.2780533730983734, + "learning_rate": 4.600383999744572e-06, + "loss": 0.6601, + "step": 1511 + }, + { + "epoch": 0.5789775990809879, + "grad_norm": 0.2862532436847687, + "learning_rate": 4.599832992937674e-06, + "loss": 0.6734, + "step": 1512 + }, + { + "epoch": 0.5793605207735019, + "grad_norm": 0.2872373163700104, + "learning_rate": 4.599281639562765e-06, + "loss": 0.6543, + "step": 1513 + }, + { + "epoch": 0.5797434424660157, + "grad_norm": 0.2874769866466522, + "learning_rate": 4.598729939710843e-06, + "loss": 0.6712, + "step": 1514 + }, + { + "epoch": 0.5801263641585296, + "grad_norm": 0.2845470607280731, + "learning_rate": 4.598177893472963e-06, + "loss": 0.6732, + "step": 1515 + }, + { + "epoch": 0.5805092858510434, + "grad_norm": 0.2877166271209717, + "learning_rate": 4.59762550094024e-06, + "loss": 0.6772, + "step": 1516 + }, + { + "epoch": 0.5808922075435573, + "grad_norm": 0.2804192900657654, + "learning_rate": 4.5970727622038415e-06, + "loss": 0.6493, + "step": 1517 + }, + { + "epoch": 0.5812751292360713, + "grad_norm": 0.29686498641967773, + "learning_rate": 4.596519677354995e-06, + "loss": 0.6717, + "step": 1518 + }, + { + "epoch": 0.5816580509285851, + "grad_norm": 0.2753806710243225, + "learning_rate": 4.595966246484986e-06, + "loss": 0.6572, + "step": 1519 + }, + { + "epoch": 0.582040972621099, + "grad_norm": 0.2850971221923828, + "learning_rate": 4.595412469685154e-06, + "loss": 0.647, + "step": 1520 + }, + { + "epoch": 0.5824238943136129, + "grad_norm": 0.3031977713108063, + "learning_rate": 4.594858347046899e-06, + "loss": 0.6634, + "step": 1521 + }, + { + "epoch": 0.5828068160061267, + "grad_norm": 0.2768973112106323, + "learning_rate": 4.594303878661676e-06, + "loss": 0.6569, + "step": 1522 + }, + { + "epoch": 0.5831897376986406, + "grad_norm": 0.29982253909111023, + "learning_rate": 4.593749064620996e-06, + "loss": 0.6503, + "step": 1523 + }, + { + "epoch": 0.5835726593911545, + "grad_norm": 0.3226918876171112, + "learning_rate": 4.593193905016431e-06, + "loss": 0.6593, + "step": 1524 + }, + { + "epoch": 0.5839555810836684, + "grad_norm": 0.28621339797973633, + "learning_rate": 4.592638399939606e-06, + "loss": 0.6677, + "step": 1525 + }, + { + "epoch": 0.5843385027761823, + "grad_norm": 0.31656476855278015, + "learning_rate": 4.592082549482205e-06, + "loss": 0.6422, + "step": 1526 + }, + { + "epoch": 0.5847214244686961, + "grad_norm": 0.31038275361061096, + "learning_rate": 4.5915263537359685e-06, + "loss": 0.6521, + "step": 1527 + }, + { + "epoch": 0.58510434616121, + "grad_norm": 0.2947365641593933, + "learning_rate": 4.590969812792694e-06, + "loss": 0.6655, + "step": 1528 + }, + { + "epoch": 0.585487267853724, + "grad_norm": 0.3024202883243561, + "learning_rate": 4.590412926744237e-06, + "loss": 0.6577, + "step": 1529 + }, + { + "epoch": 0.5858701895462378, + "grad_norm": 0.2764289081096649, + "learning_rate": 4.589855695682508e-06, + "loss": 0.64, + "step": 1530 + }, + { + "epoch": 0.5862531112387517, + "grad_norm": 0.28834182024002075, + "learning_rate": 4.5892981196994755e-06, + "loss": 0.6613, + "step": 1531 + }, + { + "epoch": 0.5866360329312655, + "grad_norm": 0.29963499307632446, + "learning_rate": 4.588740198887166e-06, + "loss": 0.6848, + "step": 1532 + }, + { + "epoch": 0.5870189546237794, + "grad_norm": 0.2901003360748291, + "learning_rate": 4.588181933337661e-06, + "loss": 0.6728, + "step": 1533 + }, + { + "epoch": 0.5874018763162934, + "grad_norm": 0.28126630187034607, + "learning_rate": 4.587623323143099e-06, + "loss": 0.6958, + "step": 1534 + }, + { + "epoch": 0.5877847980088072, + "grad_norm": 0.2929370105266571, + "learning_rate": 4.587064368395679e-06, + "loss": 0.6556, + "step": 1535 + }, + { + "epoch": 0.5881677197013211, + "grad_norm": 0.2871021330356598, + "learning_rate": 4.58650506918765e-06, + "loss": 0.6726, + "step": 1536 + }, + { + "epoch": 0.588550641393835, + "grad_norm": 0.280502587556839, + "learning_rate": 4.5859454256113235e-06, + "loss": 0.686, + "step": 1537 + }, + { + "epoch": 0.5889335630863488, + "grad_norm": 0.28914710879325867, + "learning_rate": 4.5853854377590675e-06, + "loss": 0.6656, + "step": 1538 + }, + { + "epoch": 0.5893164847788627, + "grad_norm": 0.2821902930736542, + "learning_rate": 4.584825105723304e-06, + "loss": 0.6686, + "step": 1539 + }, + { + "epoch": 0.5896994064713766, + "grad_norm": 0.29660072922706604, + "learning_rate": 4.5842644295965136e-06, + "loss": 0.6291, + "step": 1540 + }, + { + "epoch": 0.5900823281638905, + "grad_norm": 0.28437337279319763, + "learning_rate": 4.583703409471234e-06, + "loss": 0.6716, + "step": 1541 + }, + { + "epoch": 0.5904652498564044, + "grad_norm": 0.27621564269065857, + "learning_rate": 4.5831420454400575e-06, + "loss": 0.653, + "step": 1542 + }, + { + "epoch": 0.5908481715489182, + "grad_norm": 0.2817482054233551, + "learning_rate": 4.582580337595636e-06, + "loss": 0.6565, + "step": 1543 + }, + { + "epoch": 0.5912310932414321, + "grad_norm": 0.289776474237442, + "learning_rate": 4.582018286030677e-06, + "loss": 0.6541, + "step": 1544 + }, + { + "epoch": 0.591614014933946, + "grad_norm": 0.2802707850933075, + "learning_rate": 4.581455890837943e-06, + "loss": 0.6412, + "step": 1545 + }, + { + "epoch": 0.5919969366264599, + "grad_norm": 0.28509002923965454, + "learning_rate": 4.5808931521102585e-06, + "loss": 0.669, + "step": 1546 + }, + { + "epoch": 0.5923798583189738, + "grad_norm": 0.2864404022693634, + "learning_rate": 4.580330069940496e-06, + "loss": 0.6604, + "step": 1547 + }, + { + "epoch": 0.5927627800114876, + "grad_norm": 0.28585776686668396, + "learning_rate": 4.579766644421593e-06, + "loss": 0.6552, + "step": 1548 + }, + { + "epoch": 0.5931457017040015, + "grad_norm": 0.29642269015312195, + "learning_rate": 4.57920287564654e-06, + "loss": 0.6626, + "step": 1549 + }, + { + "epoch": 0.5935286233965155, + "grad_norm": 0.3006981313228607, + "learning_rate": 4.578638763708384e-06, + "loss": 0.6478, + "step": 1550 + }, + { + "epoch": 0.5939115450890293, + "grad_norm": 0.28076109290122986, + "learning_rate": 4.578074308700228e-06, + "loss": 0.6551, + "step": 1551 + }, + { + "epoch": 0.5942944667815432, + "grad_norm": 0.2876197397708893, + "learning_rate": 4.577509510715234e-06, + "loss": 0.6683, + "step": 1552 + }, + { + "epoch": 0.594677388474057, + "grad_norm": 0.285412460565567, + "learning_rate": 4.576944369846621e-06, + "loss": 0.6697, + "step": 1553 + }, + { + "epoch": 0.5950603101665709, + "grad_norm": 0.2955072820186615, + "learning_rate": 4.57637888618766e-06, + "loss": 0.6512, + "step": 1554 + }, + { + "epoch": 0.5954432318590848, + "grad_norm": 0.2828025221824646, + "learning_rate": 4.575813059831683e-06, + "loss": 0.6519, + "step": 1555 + }, + { + "epoch": 0.5958261535515987, + "grad_norm": 0.29677820205688477, + "learning_rate": 4.575246890872077e-06, + "loss": 0.6749, + "step": 1556 + }, + { + "epoch": 0.5962090752441126, + "grad_norm": 0.28829675912857056, + "learning_rate": 4.574680379402285e-06, + "loss": 0.663, + "step": 1557 + }, + { + "epoch": 0.5965919969366265, + "grad_norm": 0.28253594040870667, + "learning_rate": 4.574113525515809e-06, + "loss": 0.6689, + "step": 1558 + }, + { + "epoch": 0.5969749186291403, + "grad_norm": 0.29308921098709106, + "learning_rate": 4.5735463293062026e-06, + "loss": 0.6837, + "step": 1559 + }, + { + "epoch": 0.5973578403216542, + "grad_norm": 0.3044282793998718, + "learning_rate": 4.572978790867082e-06, + "loss": 0.6541, + "step": 1560 + }, + { + "epoch": 0.5977407620141681, + "grad_norm": 0.2985263764858246, + "learning_rate": 4.572410910292116e-06, + "loss": 0.6708, + "step": 1561 + }, + { + "epoch": 0.598123683706682, + "grad_norm": 0.3009042739868164, + "learning_rate": 4.571842687675029e-06, + "loss": 0.6658, + "step": 1562 + }, + { + "epoch": 0.5985066053991959, + "grad_norm": 0.2898600101470947, + "learning_rate": 4.571274123109606e-06, + "loss": 0.6702, + "step": 1563 + }, + { + "epoch": 0.5988895270917097, + "grad_norm": 0.28798240423202515, + "learning_rate": 4.570705216689685e-06, + "loss": 0.668, + "step": 1564 + }, + { + "epoch": 0.5992724487842236, + "grad_norm": 0.2966676354408264, + "learning_rate": 4.5701359685091605e-06, + "loss": 0.6587, + "step": 1565 + }, + { + "epoch": 0.5996553704767376, + "grad_norm": 0.28760483860969543, + "learning_rate": 4.569566378661987e-06, + "loss": 0.6412, + "step": 1566 + }, + { + "epoch": 0.6000382921692514, + "grad_norm": 0.290778785943985, + "learning_rate": 4.56899644724217e-06, + "loss": 0.6698, + "step": 1567 + }, + { + "epoch": 0.6004212138617653, + "grad_norm": 0.284491628408432, + "learning_rate": 4.5684261743437756e-06, + "loss": 0.6863, + "step": 1568 + }, + { + "epoch": 0.6008041355542791, + "grad_norm": 0.28828680515289307, + "learning_rate": 4.567855560060923e-06, + "loss": 0.656, + "step": 1569 + }, + { + "epoch": 0.601187057246793, + "grad_norm": 0.2792324721813202, + "learning_rate": 4.567284604487793e-06, + "loss": 0.6665, + "step": 1570 + }, + { + "epoch": 0.6015699789393069, + "grad_norm": 0.2900835871696472, + "learning_rate": 4.566713307718616e-06, + "loss": 0.6789, + "step": 1571 + }, + { + "epoch": 0.6019529006318208, + "grad_norm": 0.2878018915653229, + "learning_rate": 4.566141669847683e-06, + "loss": 0.6603, + "step": 1572 + }, + { + "epoch": 0.6023358223243347, + "grad_norm": 0.28778883814811707, + "learning_rate": 4.565569690969341e-06, + "loss": 0.6343, + "step": 1573 + }, + { + "epoch": 0.6027187440168486, + "grad_norm": 0.2889214754104614, + "learning_rate": 4.564997371177992e-06, + "loss": 0.6672, + "step": 1574 + }, + { + "epoch": 0.6031016657093624, + "grad_norm": 0.2788236737251282, + "learning_rate": 4.564424710568095e-06, + "loss": 0.6515, + "step": 1575 + }, + { + "epoch": 0.6034845874018763, + "grad_norm": 0.27206557989120483, + "learning_rate": 4.563851709234164e-06, + "loss": 0.6483, + "step": 1576 + }, + { + "epoch": 0.6038675090943902, + "grad_norm": 0.28735655546188354, + "learning_rate": 4.563278367270772e-06, + "loss": 0.6597, + "step": 1577 + }, + { + "epoch": 0.6042504307869041, + "grad_norm": 0.29162538051605225, + "learning_rate": 4.562704684772545e-06, + "loss": 0.6685, + "step": 1578 + }, + { + "epoch": 0.604633352479418, + "grad_norm": 0.2774061858654022, + "learning_rate": 4.5621306618341674e-06, + "loss": 0.658, + "step": 1579 + }, + { + "epoch": 0.6050162741719318, + "grad_norm": 0.2785862386226654, + "learning_rate": 4.561556298550379e-06, + "loss": 0.6621, + "step": 1580 + }, + { + "epoch": 0.6053991958644457, + "grad_norm": 0.29351580142974854, + "learning_rate": 4.560981595015976e-06, + "loss": 0.6577, + "step": 1581 + }, + { + "epoch": 0.6057821175569597, + "grad_norm": 0.2703281044960022, + "learning_rate": 4.560406551325811e-06, + "loss": 0.6528, + "step": 1582 + }, + { + "epoch": 0.6061650392494735, + "grad_norm": 0.28498125076293945, + "learning_rate": 4.559831167574792e-06, + "loss": 0.6598, + "step": 1583 + }, + { + "epoch": 0.6065479609419874, + "grad_norm": 0.27280595898628235, + "learning_rate": 4.5592554438578825e-06, + "loss": 0.662, + "step": 1584 + }, + { + "epoch": 0.6069308826345012, + "grad_norm": 0.2835381031036377, + "learning_rate": 4.5586793802701055e-06, + "loss": 0.6476, + "step": 1585 + }, + { + "epoch": 0.6073138043270151, + "grad_norm": 0.2844344675540924, + "learning_rate": 4.5581029769065356e-06, + "loss": 0.6461, + "step": 1586 + }, + { + "epoch": 0.607696726019529, + "grad_norm": 0.29839304089546204, + "learning_rate": 4.557526233862306e-06, + "loss": 0.6476, + "step": 1587 + }, + { + "epoch": 0.6080796477120429, + "grad_norm": 0.288681298494339, + "learning_rate": 4.556949151232606e-06, + "loss": 0.6656, + "step": 1588 + }, + { + "epoch": 0.6084625694045568, + "grad_norm": 0.28268155455589294, + "learning_rate": 4.556371729112681e-06, + "loss": 0.6607, + "step": 1589 + }, + { + "epoch": 0.6088454910970706, + "grad_norm": 0.28057920932769775, + "learning_rate": 4.55579396759783e-06, + "loss": 0.666, + "step": 1590 + }, + { + "epoch": 0.6092284127895845, + "grad_norm": 0.2781616747379303, + "learning_rate": 4.555215866783412e-06, + "loss": 0.6482, + "step": 1591 + }, + { + "epoch": 0.6096113344820984, + "grad_norm": 0.27866655588150024, + "learning_rate": 4.55463742676484e-06, + "loss": 0.6714, + "step": 1592 + }, + { + "epoch": 0.6099942561746123, + "grad_norm": 0.2757575213909149, + "learning_rate": 4.55405864763758e-06, + "loss": 0.6692, + "step": 1593 + }, + { + "epoch": 0.6103771778671262, + "grad_norm": 0.2932291030883789, + "learning_rate": 4.553479529497161e-06, + "loss": 0.6728, + "step": 1594 + }, + { + "epoch": 0.6107600995596401, + "grad_norm": 0.2893361747264862, + "learning_rate": 4.552900072439161e-06, + "loss": 0.6583, + "step": 1595 + }, + { + "epoch": 0.6111430212521539, + "grad_norm": 0.28445225954055786, + "learning_rate": 4.552320276559218e-06, + "loss": 0.651, + "step": 1596 + }, + { + "epoch": 0.6115259429446678, + "grad_norm": 0.2812953591346741, + "learning_rate": 4.551740141953024e-06, + "loss": 0.6565, + "step": 1597 + }, + { + "epoch": 0.6119088646371817, + "grad_norm": 0.2781602442264557, + "learning_rate": 4.551159668716328e-06, + "loss": 0.6742, + "step": 1598 + }, + { + "epoch": 0.6122917863296956, + "grad_norm": 0.29508906602859497, + "learning_rate": 4.550578856944934e-06, + "loss": 0.664, + "step": 1599 + }, + { + "epoch": 0.6126747080222095, + "grad_norm": 0.27833086252212524, + "learning_rate": 4.549997706734703e-06, + "loss": 0.6435, + "step": 1600 + }, + { + "epoch": 0.6130576297147233, + "grad_norm": 0.2886672616004944, + "learning_rate": 4.5494162181815515e-06, + "loss": 0.6747, + "step": 1601 + }, + { + "epoch": 0.6134405514072372, + "grad_norm": 0.2867382764816284, + "learning_rate": 4.5488343913814514e-06, + "loss": 0.6662, + "step": 1602 + }, + { + "epoch": 0.613823473099751, + "grad_norm": 0.2865198850631714, + "learning_rate": 4.548252226430429e-06, + "loss": 0.6753, + "step": 1603 + }, + { + "epoch": 0.614206394792265, + "grad_norm": 0.27338707447052, + "learning_rate": 4.547669723424571e-06, + "loss": 0.6728, + "step": 1604 + }, + { + "epoch": 0.6145893164847789, + "grad_norm": 0.2886753976345062, + "learning_rate": 4.547086882460015e-06, + "loss": 0.6351, + "step": 1605 + }, + { + "epoch": 0.6149722381772927, + "grad_norm": 0.2924022674560547, + "learning_rate": 4.5465037036329555e-06, + "loss": 0.6796, + "step": 1606 + }, + { + "epoch": 0.6153551598698066, + "grad_norm": 0.29398876428604126, + "learning_rate": 4.545920187039645e-06, + "loss": 0.6784, + "step": 1607 + }, + { + "epoch": 0.6157380815623205, + "grad_norm": 0.29421985149383545, + "learning_rate": 4.545336332776391e-06, + "loss": 0.6648, + "step": 1608 + }, + { + "epoch": 0.6161210032548344, + "grad_norm": 0.268839031457901, + "learning_rate": 4.544752140939553e-06, + "loss": 0.6481, + "step": 1609 + }, + { + "epoch": 0.6165039249473483, + "grad_norm": 0.2773914337158203, + "learning_rate": 4.5441676116255524e-06, + "loss": 0.6589, + "step": 1610 + }, + { + "epoch": 0.6168868466398622, + "grad_norm": 0.3153362572193146, + "learning_rate": 4.543582744930861e-06, + "loss": 0.6506, + "step": 1611 + }, + { + "epoch": 0.617269768332376, + "grad_norm": 0.28729212284088135, + "learning_rate": 4.54299754095201e-06, + "loss": 0.6676, + "step": 1612 + }, + { + "epoch": 0.6176526900248899, + "grad_norm": 0.2928646206855774, + "learning_rate": 4.542411999785584e-06, + "loss": 0.6375, + "step": 1613 + }, + { + "epoch": 0.6180356117174038, + "grad_norm": 0.28571853041648865, + "learning_rate": 4.541826121528225e-06, + "loss": 0.6607, + "step": 1614 + }, + { + "epoch": 0.6184185334099177, + "grad_norm": 0.2773991525173187, + "learning_rate": 4.541239906276627e-06, + "loss": 0.6351, + "step": 1615 + }, + { + "epoch": 0.6188014551024316, + "grad_norm": 0.28088879585266113, + "learning_rate": 4.540653354127544e-06, + "loss": 0.6465, + "step": 1616 + }, + { + "epoch": 0.6191843767949454, + "grad_norm": 0.288987934589386, + "learning_rate": 4.5400664651777835e-06, + "loss": 0.6516, + "step": 1617 + }, + { + "epoch": 0.6195672984874593, + "grad_norm": 0.28885790705680847, + "learning_rate": 4.539479239524209e-06, + "loss": 0.6507, + "step": 1618 + }, + { + "epoch": 0.6199502201799731, + "grad_norm": 0.3078453540802002, + "learning_rate": 4.538891677263739e-06, + "loss": 0.6939, + "step": 1619 + }, + { + "epoch": 0.6203331418724871, + "grad_norm": 0.28978171944618225, + "learning_rate": 4.5383037784933494e-06, + "loss": 0.6672, + "step": 1620 + }, + { + "epoch": 0.620716063565001, + "grad_norm": 0.28613409399986267, + "learning_rate": 4.537715543310068e-06, + "loss": 0.6395, + "step": 1621 + }, + { + "epoch": 0.6210989852575148, + "grad_norm": 0.27480873465538025, + "learning_rate": 4.537126971810982e-06, + "loss": 0.6531, + "step": 1622 + }, + { + "epoch": 0.6214819069500287, + "grad_norm": 0.29575228691101074, + "learning_rate": 4.5365380640932335e-06, + "loss": 0.6687, + "step": 1623 + }, + { + "epoch": 0.6218648286425426, + "grad_norm": 0.29107657074928284, + "learning_rate": 4.5359488202540156e-06, + "loss": 0.6475, + "step": 1624 + }, + { + "epoch": 0.6222477503350565, + "grad_norm": 0.2802061438560486, + "learning_rate": 4.535359240390584e-06, + "loss": 0.6574, + "step": 1625 + }, + { + "epoch": 0.6226306720275704, + "grad_norm": 0.30288583040237427, + "learning_rate": 4.534769324600243e-06, + "loss": 0.6612, + "step": 1626 + }, + { + "epoch": 0.6230135937200842, + "grad_norm": 0.27718931436538696, + "learning_rate": 4.534179072980359e-06, + "loss": 0.6539, + "step": 1627 + }, + { + "epoch": 0.6233965154125981, + "grad_norm": 0.2963017225265503, + "learning_rate": 4.533588485628347e-06, + "loss": 0.6723, + "step": 1628 + }, + { + "epoch": 0.623779437105112, + "grad_norm": 0.29171425104141235, + "learning_rate": 4.532997562641683e-06, + "loss": 0.6616, + "step": 1629 + }, + { + "epoch": 0.6241623587976259, + "grad_norm": 0.2799478769302368, + "learning_rate": 4.532406304117896e-06, + "loss": 0.6597, + "step": 1630 + }, + { + "epoch": 0.6245452804901398, + "grad_norm": 0.28800514340400696, + "learning_rate": 4.53181471015457e-06, + "loss": 0.6743, + "step": 1631 + }, + { + "epoch": 0.6249282021826537, + "grad_norm": 0.284427672624588, + "learning_rate": 4.5312227808493445e-06, + "loss": 0.6608, + "step": 1632 + }, + { + "epoch": 0.6253111238751675, + "grad_norm": 0.2893275022506714, + "learning_rate": 4.530630516299915e-06, + "loss": 0.668, + "step": 1633 + }, + { + "epoch": 0.6256940455676814, + "grad_norm": 0.2894516885280609, + "learning_rate": 4.530037916604033e-06, + "loss": 0.6694, + "step": 1634 + }, + { + "epoch": 0.6260769672601952, + "grad_norm": 0.2836352288722992, + "learning_rate": 4.5294449818595045e-06, + "loss": 0.6578, + "step": 1635 + }, + { + "epoch": 0.6264598889527092, + "grad_norm": 0.30977001786231995, + "learning_rate": 4.52885171216419e-06, + "loss": 0.6582, + "step": 1636 + }, + { + "epoch": 0.6268428106452231, + "grad_norm": 0.2887343764305115, + "learning_rate": 4.528258107616005e-06, + "loss": 0.6636, + "step": 1637 + }, + { + "epoch": 0.6272257323377369, + "grad_norm": 0.2849438190460205, + "learning_rate": 4.527664168312923e-06, + "loss": 0.642, + "step": 1638 + }, + { + "epoch": 0.6276086540302508, + "grad_norm": 0.28767022490501404, + "learning_rate": 4.52706989435297e-06, + "loss": 0.6772, + "step": 1639 + }, + { + "epoch": 0.6279915757227647, + "grad_norm": 0.30522409081459045, + "learning_rate": 4.526475285834229e-06, + "loss": 0.6664, + "step": 1640 + }, + { + "epoch": 0.6283744974152786, + "grad_norm": 0.2825944423675537, + "learning_rate": 4.525880342854836e-06, + "loss": 0.6559, + "step": 1641 + }, + { + "epoch": 0.6287574191077925, + "grad_norm": 0.289946049451828, + "learning_rate": 4.525285065512986e-06, + "loss": 0.6728, + "step": 1642 + }, + { + "epoch": 0.6291403408003063, + "grad_norm": 0.2922952473163605, + "learning_rate": 4.524689453906926e-06, + "loss": 0.6641, + "step": 1643 + }, + { + "epoch": 0.6295232624928202, + "grad_norm": 0.2942145764827728, + "learning_rate": 4.524093508134957e-06, + "loss": 0.6946, + "step": 1644 + }, + { + "epoch": 0.6299061841853341, + "grad_norm": 0.271501362323761, + "learning_rate": 4.52349722829544e-06, + "loss": 0.6785, + "step": 1645 + }, + { + "epoch": 0.630289105877848, + "grad_norm": 0.27138328552246094, + "learning_rate": 4.522900614486786e-06, + "loss": 0.653, + "step": 1646 + }, + { + "epoch": 0.6306720275703619, + "grad_norm": 0.2834060788154602, + "learning_rate": 4.522303666807465e-06, + "loss": 0.6615, + "step": 1647 + }, + { + "epoch": 0.6310549492628758, + "grad_norm": 0.28810012340545654, + "learning_rate": 4.521706385356e-06, + "loss": 0.6805, + "step": 1648 + }, + { + "epoch": 0.6314378709553896, + "grad_norm": 0.3068544864654541, + "learning_rate": 4.52110877023097e-06, + "loss": 0.6554, + "step": 1649 + }, + { + "epoch": 0.6318207926479035, + "grad_norm": 0.2793084979057312, + "learning_rate": 4.520510821531008e-06, + "loss": 0.6598, + "step": 1650 + }, + { + "epoch": 0.6322037143404173, + "grad_norm": 0.28801456093788147, + "learning_rate": 4.519912539354803e-06, + "loss": 0.655, + "step": 1651 + }, + { + "epoch": 0.6325866360329313, + "grad_norm": 0.28550705313682556, + "learning_rate": 4.5193139238011e-06, + "loss": 0.6442, + "step": 1652 + }, + { + "epoch": 0.6329695577254452, + "grad_norm": 0.2896331548690796, + "learning_rate": 4.518714974968696e-06, + "loss": 0.654, + "step": 1653 + }, + { + "epoch": 0.633352479417959, + "grad_norm": 0.2918282747268677, + "learning_rate": 4.518115692956445e-06, + "loss": 0.6489, + "step": 1654 + }, + { + "epoch": 0.6337354011104729, + "grad_norm": 0.2840275168418884, + "learning_rate": 4.517516077863256e-06, + "loss": 0.6819, + "step": 1655 + }, + { + "epoch": 0.6341183228029867, + "grad_norm": 0.28906315565109253, + "learning_rate": 4.516916129788095e-06, + "loss": 0.6513, + "step": 1656 + }, + { + "epoch": 0.6345012444955007, + "grad_norm": 0.281761109828949, + "learning_rate": 4.516315848829977e-06, + "loss": 0.6519, + "step": 1657 + }, + { + "epoch": 0.6348841661880146, + "grad_norm": 0.28358712792396545, + "learning_rate": 4.515715235087979e-06, + "loss": 0.6615, + "step": 1658 + }, + { + "epoch": 0.6352670878805284, + "grad_norm": 0.27831947803497314, + "learning_rate": 4.515114288661227e-06, + "loss": 0.6573, + "step": 1659 + }, + { + "epoch": 0.6356500095730423, + "grad_norm": 0.29940974712371826, + "learning_rate": 4.514513009648907e-06, + "loss": 0.6503, + "step": 1660 + }, + { + "epoch": 0.6360329312655562, + "grad_norm": 0.28754907846450806, + "learning_rate": 4.5139113981502545e-06, + "loss": 0.6643, + "step": 1661 + }, + { + "epoch": 0.6364158529580701, + "grad_norm": 0.2905157506465912, + "learning_rate": 4.513309454264566e-06, + "loss": 0.6751, + "step": 1662 + }, + { + "epoch": 0.636798774650584, + "grad_norm": 0.29055503010749817, + "learning_rate": 4.512707178091187e-06, + "loss": 0.6474, + "step": 1663 + }, + { + "epoch": 0.6371816963430978, + "grad_norm": 0.2892800569534302, + "learning_rate": 4.512104569729521e-06, + "loss": 0.6609, + "step": 1664 + }, + { + "epoch": 0.6375646180356117, + "grad_norm": 0.28348493576049805, + "learning_rate": 4.511501629279027e-06, + "loss": 0.6618, + "step": 1665 + }, + { + "epoch": 0.6379475397281256, + "grad_norm": 0.27873194217681885, + "learning_rate": 4.510898356839216e-06, + "loss": 0.6448, + "step": 1666 + }, + { + "epoch": 0.6383304614206394, + "grad_norm": 0.28761720657348633, + "learning_rate": 4.5102947525096575e-06, + "loss": 0.6746, + "step": 1667 + }, + { + "epoch": 0.6387133831131534, + "grad_norm": 0.3056875169277191, + "learning_rate": 4.509690816389971e-06, + "loss": 0.6671, + "step": 1668 + }, + { + "epoch": 0.6390963048056673, + "grad_norm": 0.27585890889167786, + "learning_rate": 4.509086548579836e-06, + "loss": 0.6602, + "step": 1669 + }, + { + "epoch": 0.6394792264981811, + "grad_norm": 0.2888655364513397, + "learning_rate": 4.508481949178983e-06, + "loss": 0.6606, + "step": 1670 + }, + { + "epoch": 0.639862148190695, + "grad_norm": 0.2705753445625305, + "learning_rate": 4.507877018287199e-06, + "loss": 0.6507, + "step": 1671 + }, + { + "epoch": 0.6402450698832088, + "grad_norm": 0.2835851013660431, + "learning_rate": 4.507271756004323e-06, + "loss": 0.6638, + "step": 1672 + }, + { + "epoch": 0.6406279915757228, + "grad_norm": 0.28667914867401123, + "learning_rate": 4.506666162430254e-06, + "loss": 0.6861, + "step": 1673 + }, + { + "epoch": 0.6410109132682367, + "grad_norm": 0.2852957844734192, + "learning_rate": 4.506060237664941e-06, + "loss": 0.6582, + "step": 1674 + }, + { + "epoch": 0.6413938349607505, + "grad_norm": 0.287737101316452, + "learning_rate": 4.505453981808389e-06, + "loss": 0.6758, + "step": 1675 + }, + { + "epoch": 0.6417767566532644, + "grad_norm": 0.2849019765853882, + "learning_rate": 4.504847394960657e-06, + "loss": 0.6753, + "step": 1676 + }, + { + "epoch": 0.6421596783457783, + "grad_norm": 0.28787752985954285, + "learning_rate": 4.504240477221861e-06, + "loss": 0.6547, + "step": 1677 + }, + { + "epoch": 0.6425426000382922, + "grad_norm": 0.2829058766365051, + "learning_rate": 4.503633228692169e-06, + "loss": 0.643, + "step": 1678 + }, + { + "epoch": 0.6429255217308061, + "grad_norm": 0.2887209951877594, + "learning_rate": 4.503025649471806e-06, + "loss": 0.6785, + "step": 1679 + }, + { + "epoch": 0.6433084434233199, + "grad_norm": 0.28281310200691223, + "learning_rate": 4.502417739661049e-06, + "loss": 0.6697, + "step": 1680 + }, + { + "epoch": 0.6436913651158338, + "grad_norm": 0.277035117149353, + "learning_rate": 4.50180949936023e-06, + "loss": 0.6516, + "step": 1681 + }, + { + "epoch": 0.6440742868083477, + "grad_norm": 0.28803032636642456, + "learning_rate": 4.501200928669739e-06, + "loss": 0.6869, + "step": 1682 + }, + { + "epoch": 0.6444572085008615, + "grad_norm": 0.27366191148757935, + "learning_rate": 4.500592027690015e-06, + "loss": 0.6537, + "step": 1683 + }, + { + "epoch": 0.6448401301933755, + "grad_norm": 0.2732759416103363, + "learning_rate": 4.499982796521556e-06, + "loss": 0.6628, + "step": 1684 + }, + { + "epoch": 0.6452230518858894, + "grad_norm": 0.28367921710014343, + "learning_rate": 4.499373235264913e-06, + "loss": 0.7002, + "step": 1685 + }, + { + "epoch": 0.6456059735784032, + "grad_norm": 0.2750464379787445, + "learning_rate": 4.4987633440206905e-06, + "loss": 0.6538, + "step": 1686 + }, + { + "epoch": 0.6459888952709171, + "grad_norm": 0.286153644323349, + "learning_rate": 4.498153122889549e-06, + "loss": 0.6687, + "step": 1687 + }, + { + "epoch": 0.6463718169634309, + "grad_norm": 0.28421804308891296, + "learning_rate": 4.4975425719722025e-06, + "loss": 0.659, + "step": 1688 + }, + { + "epoch": 0.6467547386559449, + "grad_norm": 0.2871510684490204, + "learning_rate": 4.496931691369419e-06, + "loss": 0.6614, + "step": 1689 + }, + { + "epoch": 0.6471376603484588, + "grad_norm": 0.2869611978530884, + "learning_rate": 4.496320481182023e-06, + "loss": 0.6705, + "step": 1690 + }, + { + "epoch": 0.6475205820409726, + "grad_norm": 0.28083720803260803, + "learning_rate": 4.49570894151089e-06, + "loss": 0.662, + "step": 1691 + }, + { + "epoch": 0.6479035037334865, + "grad_norm": 0.27947762608528137, + "learning_rate": 4.495097072456954e-06, + "loss": 0.6655, + "step": 1692 + }, + { + "epoch": 0.6482864254260003, + "grad_norm": 0.28797227144241333, + "learning_rate": 4.4944848741212e-06, + "loss": 0.659, + "step": 1693 + }, + { + "epoch": 0.6486693471185143, + "grad_norm": 0.2906720042228699, + "learning_rate": 4.4938723466046694e-06, + "loss": 0.6401, + "step": 1694 + }, + { + "epoch": 0.6490522688110282, + "grad_norm": 0.2948593199253082, + "learning_rate": 4.493259490008457e-06, + "loss": 0.6663, + "step": 1695 + }, + { + "epoch": 0.649435190503542, + "grad_norm": 0.29138094186782837, + "learning_rate": 4.492646304433711e-06, + "loss": 0.6919, + "step": 1696 + }, + { + "epoch": 0.6498181121960559, + "grad_norm": 0.28296101093292236, + "learning_rate": 4.4920327899816356e-06, + "loss": 0.6407, + "step": 1697 + }, + { + "epoch": 0.6502010338885698, + "grad_norm": 0.2952142357826233, + "learning_rate": 4.491418946753489e-06, + "loss": 0.6761, + "step": 1698 + }, + { + "epoch": 0.6505839555810836, + "grad_norm": 0.2904972434043884, + "learning_rate": 4.490804774850582e-06, + "loss": 0.6652, + "step": 1699 + }, + { + "epoch": 0.6509668772735976, + "grad_norm": 0.29449400305747986, + "learning_rate": 4.4901902743742826e-06, + "loss": 0.6747, + "step": 1700 + }, + { + "epoch": 0.6513497989661114, + "grad_norm": 0.2727620303630829, + "learning_rate": 4.48957544542601e-06, + "loss": 0.6389, + "step": 1701 + }, + { + "epoch": 0.6517327206586253, + "grad_norm": 0.29242339730262756, + "learning_rate": 4.48896028810724e-06, + "loss": 0.6538, + "step": 1702 + }, + { + "epoch": 0.6521156423511392, + "grad_norm": 0.2878548502922058, + "learning_rate": 4.488344802519501e-06, + "loss": 0.6608, + "step": 1703 + }, + { + "epoch": 0.652498564043653, + "grad_norm": 0.2867409884929657, + "learning_rate": 4.487728988764377e-06, + "loss": 0.6515, + "step": 1704 + }, + { + "epoch": 0.652881485736167, + "grad_norm": 0.28430137038230896, + "learning_rate": 4.487112846943503e-06, + "loss": 0.6466, + "step": 1705 + }, + { + "epoch": 0.6532644074286809, + "grad_norm": 0.28649213910102844, + "learning_rate": 4.486496377158572e-06, + "loss": 0.6495, + "step": 1706 + }, + { + "epoch": 0.6536473291211947, + "grad_norm": 0.2940782308578491, + "learning_rate": 4.485879579511331e-06, + "loss": 0.6508, + "step": 1707 + }, + { + "epoch": 0.6540302508137086, + "grad_norm": 0.28731769323349, + "learning_rate": 4.485262454103577e-06, + "loss": 0.6755, + "step": 1708 + }, + { + "epoch": 0.6544131725062224, + "grad_norm": 0.28093746304512024, + "learning_rate": 4.4846450010371656e-06, + "loss": 0.6552, + "step": 1709 + }, + { + "epoch": 0.6547960941987364, + "grad_norm": 0.2853204905986786, + "learning_rate": 4.484027220414003e-06, + "loss": 0.6542, + "step": 1710 + }, + { + "epoch": 0.6551790158912503, + "grad_norm": 0.28559935092926025, + "learning_rate": 4.483409112336054e-06, + "loss": 0.6569, + "step": 1711 + }, + { + "epoch": 0.6555619375837641, + "grad_norm": 0.2773648500442505, + "learning_rate": 4.482790676905331e-06, + "loss": 0.6666, + "step": 1712 + }, + { + "epoch": 0.655944859276278, + "grad_norm": 0.28084635734558105, + "learning_rate": 4.482171914223906e-06, + "loss": 0.665, + "step": 1713 + }, + { + "epoch": 0.6563277809687919, + "grad_norm": 0.28929758071899414, + "learning_rate": 4.481552824393903e-06, + "loss": 0.6717, + "step": 1714 + }, + { + "epoch": 0.6567107026613057, + "grad_norm": 0.28908833861351013, + "learning_rate": 4.4809334075175e-06, + "loss": 0.6682, + "step": 1715 + }, + { + "epoch": 0.6570936243538197, + "grad_norm": 0.28981640934944153, + "learning_rate": 4.480313663696928e-06, + "loss": 0.6601, + "step": 1716 + }, + { + "epoch": 0.6574765460463335, + "grad_norm": 0.30396929383277893, + "learning_rate": 4.479693593034475e-06, + "loss": 0.6716, + "step": 1717 + }, + { + "epoch": 0.6578594677388474, + "grad_norm": 0.27771109342575073, + "learning_rate": 4.479073195632478e-06, + "loss": 0.642, + "step": 1718 + }, + { + "epoch": 0.6582423894313613, + "grad_norm": 0.305408239364624, + "learning_rate": 4.478452471593333e-06, + "loss": 0.652, + "step": 1719 + }, + { + "epoch": 0.6586253111238751, + "grad_norm": 0.28755393624305725, + "learning_rate": 4.477831421019486e-06, + "loss": 0.6782, + "step": 1720 + }, + { + "epoch": 0.6590082328163891, + "grad_norm": 0.2777462303638458, + "learning_rate": 4.477210044013441e-06, + "loss": 0.6638, + "step": 1721 + }, + { + "epoch": 0.659391154508903, + "grad_norm": 0.2919204533100128, + "learning_rate": 4.476588340677751e-06, + "loss": 0.6705, + "step": 1722 + }, + { + "epoch": 0.6597740762014168, + "grad_norm": 0.2931578457355499, + "learning_rate": 4.475966311115027e-06, + "loss": 0.663, + "step": 1723 + }, + { + "epoch": 0.6601569978939307, + "grad_norm": 0.28356921672821045, + "learning_rate": 4.475343955427932e-06, + "loss": 0.6698, + "step": 1724 + }, + { + "epoch": 0.6605399195864445, + "grad_norm": 0.2742938995361328, + "learning_rate": 4.474721273719183e-06, + "loss": 0.6678, + "step": 1725 + }, + { + "epoch": 0.6609228412789585, + "grad_norm": 0.2960493862628937, + "learning_rate": 4.4740982660915485e-06, + "loss": 0.6545, + "step": 1726 + }, + { + "epoch": 0.6613057629714724, + "grad_norm": 0.2709237337112427, + "learning_rate": 4.473474932647856e-06, + "loss": 0.663, + "step": 1727 + }, + { + "epoch": 0.6616886846639862, + "grad_norm": 0.2795345187187195, + "learning_rate": 4.472851273490985e-06, + "loss": 0.6517, + "step": 1728 + }, + { + "epoch": 0.6620716063565001, + "grad_norm": 0.2873730957508087, + "learning_rate": 4.472227288723863e-06, + "loss": 0.6635, + "step": 1729 + }, + { + "epoch": 0.662454528049014, + "grad_norm": 0.2867303788661957, + "learning_rate": 4.47160297844948e-06, + "loss": 0.6524, + "step": 1730 + }, + { + "epoch": 0.6628374497415278, + "grad_norm": 0.2772071361541748, + "learning_rate": 4.470978342770873e-06, + "loss": 0.6543, + "step": 1731 + }, + { + "epoch": 0.6632203714340418, + "grad_norm": 0.2774072587490082, + "learning_rate": 4.470353381791138e-06, + "loss": 0.6521, + "step": 1732 + }, + { + "epoch": 0.6636032931265556, + "grad_norm": 0.285992294549942, + "learning_rate": 4.469728095613419e-06, + "loss": 0.666, + "step": 1733 + }, + { + "epoch": 0.6639862148190695, + "grad_norm": 0.2970382571220398, + "learning_rate": 4.469102484340919e-06, + "loss": 0.6812, + "step": 1734 + }, + { + "epoch": 0.6643691365115834, + "grad_norm": 0.28140783309936523, + "learning_rate": 4.468476548076891e-06, + "loss": 0.6624, + "step": 1735 + }, + { + "epoch": 0.6647520582040972, + "grad_norm": 0.28314921259880066, + "learning_rate": 4.4678502869246435e-06, + "loss": 0.649, + "step": 1736 + }, + { + "epoch": 0.6651349798966112, + "grad_norm": 0.2802729308605194, + "learning_rate": 4.467223700987538e-06, + "loss": 0.6639, + "step": 1737 + }, + { + "epoch": 0.665517901589125, + "grad_norm": 0.2866319417953491, + "learning_rate": 4.46659679036899e-06, + "loss": 0.6548, + "step": 1738 + }, + { + "epoch": 0.6659008232816389, + "grad_norm": 0.2863485515117645, + "learning_rate": 4.465969555172468e-06, + "loss": 0.6637, + "step": 1739 + }, + { + "epoch": 0.6662837449741528, + "grad_norm": 0.2826560437679291, + "learning_rate": 4.465341995501493e-06, + "loss": 0.6571, + "step": 1740 + }, + { + "epoch": 0.6666666666666666, + "grad_norm": 0.2898571789264679, + "learning_rate": 4.464714111459644e-06, + "loss": 0.6683, + "step": 1741 + }, + { + "epoch": 0.6670495883591806, + "grad_norm": 0.29692378640174866, + "learning_rate": 4.464085903150548e-06, + "loss": 0.6669, + "step": 1742 + }, + { + "epoch": 0.6674325100516945, + "grad_norm": 0.2766851782798767, + "learning_rate": 4.463457370677889e-06, + "loss": 0.6526, + "step": 1743 + }, + { + "epoch": 0.6678154317442083, + "grad_norm": 0.2955440580844879, + "learning_rate": 4.462828514145402e-06, + "loss": 0.6662, + "step": 1744 + }, + { + "epoch": 0.6681983534367222, + "grad_norm": 0.3091248571872711, + "learning_rate": 4.462199333656879e-06, + "loss": 0.6773, + "step": 1745 + }, + { + "epoch": 0.668581275129236, + "grad_norm": 0.29684820771217346, + "learning_rate": 4.461569829316163e-06, + "loss": 0.6738, + "step": 1746 + }, + { + "epoch": 0.6689641968217499, + "grad_norm": 0.28915321826934814, + "learning_rate": 4.46094000122715e-06, + "loss": 0.6582, + "step": 1747 + }, + { + "epoch": 0.6693471185142639, + "grad_norm": 0.2869669198989868, + "learning_rate": 4.4603098494937905e-06, + "loss": 0.6486, + "step": 1748 + }, + { + "epoch": 0.6697300402067777, + "grad_norm": 0.28078657388687134, + "learning_rate": 4.4596793742200886e-06, + "loss": 0.6572, + "step": 1749 + }, + { + "epoch": 0.6701129618992916, + "grad_norm": 0.275255024433136, + "learning_rate": 4.4590485755101e-06, + "loss": 0.6426, + "step": 1750 + }, + { + "epoch": 0.6704958835918055, + "grad_norm": 0.2922937273979187, + "learning_rate": 4.4584174534679375e-06, + "loss": 0.6681, + "step": 1751 + }, + { + "epoch": 0.6708788052843193, + "grad_norm": 0.2934999167919159, + "learning_rate": 4.4577860081977635e-06, + "loss": 0.6581, + "step": 1752 + }, + { + "epoch": 0.6712617269768333, + "grad_norm": 0.29109132289886475, + "learning_rate": 4.457154239803796e-06, + "loss": 0.6644, + "step": 1753 + }, + { + "epoch": 0.6716446486693471, + "grad_norm": 0.28353652358055115, + "learning_rate": 4.456522148390304e-06, + "loss": 0.6681, + "step": 1754 + }, + { + "epoch": 0.672027570361861, + "grad_norm": 0.28220492601394653, + "learning_rate": 4.455889734061613e-06, + "loss": 0.6493, + "step": 1755 + }, + { + "epoch": 0.6724104920543749, + "grad_norm": 0.2920376658439636, + "learning_rate": 4.4552569969221e-06, + "loss": 0.6581, + "step": 1756 + }, + { + "epoch": 0.6727934137468887, + "grad_norm": 0.3047402799129486, + "learning_rate": 4.454623937076195e-06, + "loss": 0.6537, + "step": 1757 + }, + { + "epoch": 0.6731763354394027, + "grad_norm": 0.29887130856513977, + "learning_rate": 4.453990554628382e-06, + "loss": 0.6724, + "step": 1758 + }, + { + "epoch": 0.6735592571319166, + "grad_norm": 0.29401645064353943, + "learning_rate": 4.4533568496831965e-06, + "loss": 0.6733, + "step": 1759 + }, + { + "epoch": 0.6739421788244304, + "grad_norm": 0.2885599136352539, + "learning_rate": 4.452722822345231e-06, + "loss": 0.6693, + "step": 1760 + }, + { + "epoch": 0.6743251005169443, + "grad_norm": 0.28618887066841125, + "learning_rate": 4.452088472719127e-06, + "loss": 0.6704, + "step": 1761 + }, + { + "epoch": 0.6747080222094581, + "grad_norm": 0.2804790437221527, + "learning_rate": 4.451453800909582e-06, + "loss": 0.6591, + "step": 1762 + }, + { + "epoch": 0.675090943901972, + "grad_norm": 0.29983004927635193, + "learning_rate": 4.450818807021345e-06, + "loss": 0.6763, + "step": 1763 + }, + { + "epoch": 0.675473865594486, + "grad_norm": 0.2811990976333618, + "learning_rate": 4.45018349115922e-06, + "loss": 0.6881, + "step": 1764 + }, + { + "epoch": 0.6758567872869998, + "grad_norm": 0.2975471317768097, + "learning_rate": 4.449547853428061e-06, + "loss": 0.6758, + "step": 1765 + }, + { + "epoch": 0.6762397089795137, + "grad_norm": 0.2989833652973175, + "learning_rate": 4.4489118939327795e-06, + "loss": 0.6724, + "step": 1766 + }, + { + "epoch": 0.6766226306720275, + "grad_norm": 0.2850721776485443, + "learning_rate": 4.448275612778336e-06, + "loss": 0.6687, + "step": 1767 + }, + { + "epoch": 0.6770055523645414, + "grad_norm": 0.2916548550128937, + "learning_rate": 4.447639010069748e-06, + "loss": 0.6719, + "step": 1768 + }, + { + "epoch": 0.6773884740570554, + "grad_norm": 0.28292280435562134, + "learning_rate": 4.447002085912082e-06, + "loss": 0.6601, + "step": 1769 + }, + { + "epoch": 0.6777713957495692, + "grad_norm": 0.2910638153553009, + "learning_rate": 4.44636484041046e-06, + "loss": 0.6704, + "step": 1770 + }, + { + "epoch": 0.6781543174420831, + "grad_norm": 0.2982378900051117, + "learning_rate": 4.445727273670057e-06, + "loss": 0.6322, + "step": 1771 + }, + { + "epoch": 0.678537239134597, + "grad_norm": 0.2990197539329529, + "learning_rate": 4.445089385796099e-06, + "loss": 0.6653, + "step": 1772 + }, + { + "epoch": 0.6789201608271108, + "grad_norm": 0.2974478304386139, + "learning_rate": 4.444451176893868e-06, + "loss": 0.6725, + "step": 1773 + }, + { + "epoch": 0.6793030825196248, + "grad_norm": 0.2938443720340729, + "learning_rate": 4.443812647068697e-06, + "loss": 0.65, + "step": 1774 + }, + { + "epoch": 0.6796860042121387, + "grad_norm": 0.28330492973327637, + "learning_rate": 4.443173796425973e-06, + "loss": 0.6685, + "step": 1775 + }, + { + "epoch": 0.6800689259046525, + "grad_norm": 0.28686612844467163, + "learning_rate": 4.4425346250711355e-06, + "loss": 0.647, + "step": 1776 + }, + { + "epoch": 0.6804518475971664, + "grad_norm": 0.28971579670906067, + "learning_rate": 4.441895133109675e-06, + "loss": 0.669, + "step": 1777 + }, + { + "epoch": 0.6808347692896802, + "grad_norm": 0.2774415910243988, + "learning_rate": 4.441255320647139e-06, + "loss": 0.6576, + "step": 1778 + }, + { + "epoch": 0.6812176909821941, + "grad_norm": 0.2871474027633667, + "learning_rate": 4.440615187789124e-06, + "loss": 0.6644, + "step": 1779 + }, + { + "epoch": 0.6816006126747081, + "grad_norm": 0.28074896335601807, + "learning_rate": 4.4399747346412834e-06, + "loss": 0.6573, + "step": 1780 + }, + { + "epoch": 0.6819835343672219, + "grad_norm": 0.2852190434932709, + "learning_rate": 4.439333961309319e-06, + "loss": 0.6534, + "step": 1781 + }, + { + "epoch": 0.6823664560597358, + "grad_norm": 0.3023935556411743, + "learning_rate": 4.438692867898988e-06, + "loss": 0.6626, + "step": 1782 + }, + { + "epoch": 0.6827493777522496, + "grad_norm": 0.29996219277381897, + "learning_rate": 4.4380514545161e-06, + "loss": 0.6467, + "step": 1783 + }, + { + "epoch": 0.6831322994447635, + "grad_norm": 0.2838921546936035, + "learning_rate": 4.437409721266518e-06, + "loss": 0.665, + "step": 1784 + }, + { + "epoch": 0.6835152211372775, + "grad_norm": 0.2941323220729828, + "learning_rate": 4.436767668256157e-06, + "loss": 0.6541, + "step": 1785 + }, + { + "epoch": 0.6838981428297913, + "grad_norm": 0.2942572832107544, + "learning_rate": 4.436125295590984e-06, + "loss": 0.6671, + "step": 1786 + }, + { + "epoch": 0.6842810645223052, + "grad_norm": 0.28804951906204224, + "learning_rate": 4.43548260337702e-06, + "loss": 0.6837, + "step": 1787 + }, + { + "epoch": 0.684663986214819, + "grad_norm": 0.28608110547065735, + "learning_rate": 4.43483959172034e-06, + "loss": 0.6569, + "step": 1788 + }, + { + "epoch": 0.6850469079073329, + "grad_norm": 0.2884553074836731, + "learning_rate": 4.434196260727069e-06, + "loss": 0.6325, + "step": 1789 + }, + { + "epoch": 0.6854298295998469, + "grad_norm": 0.28959155082702637, + "learning_rate": 4.433552610503386e-06, + "loss": 0.6596, + "step": 1790 + }, + { + "epoch": 0.6858127512923607, + "grad_norm": 0.28355011343955994, + "learning_rate": 4.432908641155522e-06, + "loss": 0.6754, + "step": 1791 + }, + { + "epoch": 0.6861956729848746, + "grad_norm": 0.30101996660232544, + "learning_rate": 4.432264352789763e-06, + "loss": 0.6523, + "step": 1792 + }, + { + "epoch": 0.6865785946773885, + "grad_norm": 0.2937653064727783, + "learning_rate": 4.431619745512443e-06, + "loss": 0.6745, + "step": 1793 + }, + { + "epoch": 0.6869615163699023, + "grad_norm": 0.2968035340309143, + "learning_rate": 4.430974819429954e-06, + "loss": 0.6867, + "step": 1794 + }, + { + "epoch": 0.6873444380624162, + "grad_norm": 0.28310543298721313, + "learning_rate": 4.430329574648737e-06, + "loss": 0.6451, + "step": 1795 + }, + { + "epoch": 0.6877273597549302, + "grad_norm": 0.2911510467529297, + "learning_rate": 4.429684011275288e-06, + "loss": 0.669, + "step": 1796 + }, + { + "epoch": 0.688110281447444, + "grad_norm": 0.29060813784599304, + "learning_rate": 4.429038129416153e-06, + "loss": 0.6443, + "step": 1797 + }, + { + "epoch": 0.6884932031399579, + "grad_norm": 0.2903493642807007, + "learning_rate": 4.428391929177933e-06, + "loss": 0.6735, + "step": 1798 + }, + { + "epoch": 0.6888761248324717, + "grad_norm": 0.2872551381587982, + "learning_rate": 4.42774541066728e-06, + "loss": 0.6419, + "step": 1799 + }, + { + "epoch": 0.6892590465249856, + "grad_norm": 0.27953702211380005, + "learning_rate": 4.427098573990897e-06, + "loss": 0.6544, + "step": 1800 + }, + { + "epoch": 0.6896419682174996, + "grad_norm": 0.2751811444759369, + "learning_rate": 4.4264514192555456e-06, + "loss": 0.6413, + "step": 1801 + }, + { + "epoch": 0.6900248899100134, + "grad_norm": 0.280831515789032, + "learning_rate": 4.425803946568033e-06, + "loss": 0.6744, + "step": 1802 + }, + { + "epoch": 0.6904078116025273, + "grad_norm": 0.2886102497577667, + "learning_rate": 4.4251561560352215e-06, + "loss": 0.6643, + "step": 1803 + }, + { + "epoch": 0.6907907332950411, + "grad_norm": 0.27799156308174133, + "learning_rate": 4.424508047764027e-06, + "loss": 0.6584, + "step": 1804 + }, + { + "epoch": 0.691173654987555, + "grad_norm": 0.28802868723869324, + "learning_rate": 4.423859621861417e-06, + "loss": 0.6397, + "step": 1805 + }, + { + "epoch": 0.691556576680069, + "grad_norm": 0.2986770570278168, + "learning_rate": 4.42321087843441e-06, + "loss": 0.6442, + "step": 1806 + }, + { + "epoch": 0.6919394983725828, + "grad_norm": 0.2985818088054657, + "learning_rate": 4.42256181759008e-06, + "loss": 0.6433, + "step": 1807 + }, + { + "epoch": 0.6923224200650967, + "grad_norm": 0.2903055250644684, + "learning_rate": 4.421912439435549e-06, + "loss": 0.6487, + "step": 1808 + }, + { + "epoch": 0.6927053417576106, + "grad_norm": 0.28338488936424255, + "learning_rate": 4.421262744077997e-06, + "loss": 0.6602, + "step": 1809 + }, + { + "epoch": 0.6930882634501244, + "grad_norm": 0.2961269021034241, + "learning_rate": 4.420612731624652e-06, + "loss": 0.6797, + "step": 1810 + }, + { + "epoch": 0.6934711851426383, + "grad_norm": 0.2834930121898651, + "learning_rate": 4.4199624021827954e-06, + "loss": 0.6852, + "step": 1811 + }, + { + "epoch": 0.6938541068351523, + "grad_norm": 0.2932717204093933, + "learning_rate": 4.419311755859761e-06, + "loss": 0.6547, + "step": 1812 + }, + { + "epoch": 0.6942370285276661, + "grad_norm": 0.2781160771846771, + "learning_rate": 4.418660792762935e-06, + "loss": 0.6625, + "step": 1813 + }, + { + "epoch": 0.69461995022018, + "grad_norm": 0.2849212884902954, + "learning_rate": 4.418009512999757e-06, + "loss": 0.6531, + "step": 1814 + }, + { + "epoch": 0.6950028719126938, + "grad_norm": 0.2946004867553711, + "learning_rate": 4.417357916677716e-06, + "loss": 0.6722, + "step": 1815 + }, + { + "epoch": 0.6953857936052077, + "grad_norm": 0.2866877317428589, + "learning_rate": 4.416706003904357e-06, + "loss": 0.6451, + "step": 1816 + }, + { + "epoch": 0.6957687152977217, + "grad_norm": 0.29456812143325806, + "learning_rate": 4.416053774787274e-06, + "loss": 0.6667, + "step": 1817 + }, + { + "epoch": 0.6961516369902355, + "grad_norm": 0.30627280473709106, + "learning_rate": 4.415401229434115e-06, + "loss": 0.6727, + "step": 1818 + }, + { + "epoch": 0.6965345586827494, + "grad_norm": 0.2828829288482666, + "learning_rate": 4.414748367952579e-06, + "loss": 0.6611, + "step": 1819 + }, + { + "epoch": 0.6969174803752632, + "grad_norm": 0.28239619731903076, + "learning_rate": 4.4140951904504194e-06, + "loss": 0.6568, + "step": 1820 + }, + { + "epoch": 0.6973004020677771, + "grad_norm": 0.32551002502441406, + "learning_rate": 4.413441697035439e-06, + "loss": 0.6598, + "step": 1821 + }, + { + "epoch": 0.6976833237602911, + "grad_norm": 0.281393438577652, + "learning_rate": 4.412787887815494e-06, + "loss": 0.6662, + "step": 1822 + }, + { + "epoch": 0.6980662454528049, + "grad_norm": 0.288491427898407, + "learning_rate": 4.412133762898493e-06, + "loss": 0.6675, + "step": 1823 + }, + { + "epoch": 0.6984491671453188, + "grad_norm": 0.289437472820282, + "learning_rate": 4.411479322392396e-06, + "loss": 0.6629, + "step": 1824 + }, + { + "epoch": 0.6988320888378327, + "grad_norm": 0.2796823978424072, + "learning_rate": 4.410824566405215e-06, + "loss": 0.6668, + "step": 1825 + }, + { + "epoch": 0.6992150105303465, + "grad_norm": 0.29910945892333984, + "learning_rate": 4.410169495045017e-06, + "loss": 0.6625, + "step": 1826 + }, + { + "epoch": 0.6995979322228604, + "grad_norm": 0.30837470293045044, + "learning_rate": 4.4095141084199155e-06, + "loss": 0.6756, + "step": 1827 + }, + { + "epoch": 0.6999808539153743, + "grad_norm": 0.29551079869270325, + "learning_rate": 4.4088584066380804e-06, + "loss": 0.654, + "step": 1828 + }, + { + "epoch": 0.7003637756078882, + "grad_norm": 0.293451726436615, + "learning_rate": 4.408202389807733e-06, + "loss": 0.6499, + "step": 1829 + }, + { + "epoch": 0.7007466973004021, + "grad_norm": 0.3022894263267517, + "learning_rate": 4.4075460580371455e-06, + "loss": 0.6477, + "step": 1830 + }, + { + "epoch": 0.7011296189929159, + "grad_norm": 0.29170361161231995, + "learning_rate": 4.406889411434641e-06, + "loss": 0.6666, + "step": 1831 + }, + { + "epoch": 0.7015125406854298, + "grad_norm": 0.27581146359443665, + "learning_rate": 4.4062324501086e-06, + "loss": 0.6501, + "step": 1832 + }, + { + "epoch": 0.7018954623779438, + "grad_norm": 0.3292374014854431, + "learning_rate": 4.4055751741674454e-06, + "loss": 0.6692, + "step": 1833 + }, + { + "epoch": 0.7022783840704576, + "grad_norm": 0.28566980361938477, + "learning_rate": 4.404917583719663e-06, + "loss": 0.6558, + "step": 1834 + }, + { + "epoch": 0.7026613057629715, + "grad_norm": 0.2983189523220062, + "learning_rate": 4.404259678873782e-06, + "loss": 0.6573, + "step": 1835 + }, + { + "epoch": 0.7030442274554853, + "grad_norm": 0.28240758180618286, + "learning_rate": 4.403601459738387e-06, + "loss": 0.6597, + "step": 1836 + }, + { + "epoch": 0.7034271491479992, + "grad_norm": 0.2784506380558014, + "learning_rate": 4.402942926422115e-06, + "loss": 0.6627, + "step": 1837 + }, + { + "epoch": 0.7038100708405132, + "grad_norm": 0.28692924976348877, + "learning_rate": 4.402284079033654e-06, + "loss": 0.6611, + "step": 1838 + }, + { + "epoch": 0.704192992533027, + "grad_norm": 0.28517448902130127, + "learning_rate": 4.401624917681743e-06, + "loss": 0.6547, + "step": 1839 + }, + { + "epoch": 0.7045759142255409, + "grad_norm": 0.28843966126441956, + "learning_rate": 4.400965442475173e-06, + "loss": 0.6786, + "step": 1840 + }, + { + "epoch": 0.7049588359180547, + "grad_norm": 0.2878416180610657, + "learning_rate": 4.4003056535227905e-06, + "loss": 0.6551, + "step": 1841 + }, + { + "epoch": 0.7053417576105686, + "grad_norm": 0.2929476797580719, + "learning_rate": 4.3996455509334875e-06, + "loss": 0.6737, + "step": 1842 + }, + { + "epoch": 0.7057246793030825, + "grad_norm": 0.317920058965683, + "learning_rate": 4.398985134816211e-06, + "loss": 0.6606, + "step": 1843 + }, + { + "epoch": 0.7061076009955964, + "grad_norm": 0.2867056131362915, + "learning_rate": 4.398324405279963e-06, + "loss": 0.6522, + "step": 1844 + }, + { + "epoch": 0.7064905226881103, + "grad_norm": 0.3179545998573303, + "learning_rate": 4.39766336243379e-06, + "loss": 0.6614, + "step": 1845 + }, + { + "epoch": 0.7068734443806242, + "grad_norm": 0.27060508728027344, + "learning_rate": 4.397002006386797e-06, + "loss": 0.6424, + "step": 1846 + }, + { + "epoch": 0.707256366073138, + "grad_norm": 0.2941526770591736, + "learning_rate": 4.3963403372481364e-06, + "loss": 0.6826, + "step": 1847 + }, + { + "epoch": 0.7076392877656519, + "grad_norm": 0.28471407294273376, + "learning_rate": 4.395678355127013e-06, + "loss": 0.6682, + "step": 1848 + }, + { + "epoch": 0.7080222094581659, + "grad_norm": 0.2784045338630676, + "learning_rate": 4.3950160601326865e-06, + "loss": 0.6473, + "step": 1849 + }, + { + "epoch": 0.7084051311506797, + "grad_norm": 0.29154446721076965, + "learning_rate": 4.394353452374464e-06, + "loss": 0.6697, + "step": 1850 + }, + { + "epoch": 0.7087880528431936, + "grad_norm": 0.3087870180606842, + "learning_rate": 4.3936905319617056e-06, + "loss": 0.6606, + "step": 1851 + }, + { + "epoch": 0.7091709745357074, + "grad_norm": 0.290182888507843, + "learning_rate": 4.3930272990038245e-06, + "loss": 0.6679, + "step": 1852 + }, + { + "epoch": 0.7095538962282213, + "grad_norm": 0.2789461612701416, + "learning_rate": 4.392363753610284e-06, + "loss": 0.6768, + "step": 1853 + }, + { + "epoch": 0.7099368179207353, + "grad_norm": 0.2986856698989868, + "learning_rate": 4.3916998958905975e-06, + "loss": 0.6807, + "step": 1854 + }, + { + "epoch": 0.7103197396132491, + "grad_norm": 0.31243690848350525, + "learning_rate": 4.391035725954335e-06, + "loss": 0.6561, + "step": 1855 + }, + { + "epoch": 0.710702661305763, + "grad_norm": 0.28528085350990295, + "learning_rate": 4.390371243911112e-06, + "loss": 0.6574, + "step": 1856 + }, + { + "epoch": 0.7110855829982768, + "grad_norm": 0.29143473505973816, + "learning_rate": 4.3897064498706e-06, + "loss": 0.6511, + "step": 1857 + }, + { + "epoch": 0.7114685046907907, + "grad_norm": 0.29057714343070984, + "learning_rate": 4.389041343942519e-06, + "loss": 0.6707, + "step": 1858 + }, + { + "epoch": 0.7118514263833046, + "grad_norm": 0.2966049611568451, + "learning_rate": 4.3883759262366435e-06, + "loss": 0.6547, + "step": 1859 + }, + { + "epoch": 0.7122343480758185, + "grad_norm": 0.283877432346344, + "learning_rate": 4.387710196862796e-06, + "loss": 0.6563, + "step": 1860 + }, + { + "epoch": 0.7126172697683324, + "grad_norm": 0.28457963466644287, + "learning_rate": 4.387044155930852e-06, + "loss": 0.6565, + "step": 1861 + }, + { + "epoch": 0.7130001914608463, + "grad_norm": 0.3141806721687317, + "learning_rate": 4.38637780355074e-06, + "loss": 0.6428, + "step": 1862 + }, + { + "epoch": 0.7133831131533601, + "grad_norm": 0.2856174409389496, + "learning_rate": 4.385711139832437e-06, + "loss": 0.6543, + "step": 1863 + }, + { + "epoch": 0.713766034845874, + "grad_norm": 0.2827475666999817, + "learning_rate": 4.3850441648859745e-06, + "loss": 0.6714, + "step": 1864 + }, + { + "epoch": 0.714148956538388, + "grad_norm": 0.32685548067092896, + "learning_rate": 4.384376878821432e-06, + "loss": 0.6511, + "step": 1865 + }, + { + "epoch": 0.7145318782309018, + "grad_norm": 0.3012218177318573, + "learning_rate": 4.383709281748943e-06, + "loss": 0.6505, + "step": 1866 + }, + { + "epoch": 0.7149147999234157, + "grad_norm": 0.28398776054382324, + "learning_rate": 4.383041373778691e-06, + "loss": 0.6643, + "step": 1867 + }, + { + "epoch": 0.7152977216159295, + "grad_norm": 0.30807650089263916, + "learning_rate": 4.382373155020913e-06, + "loss": 0.6628, + "step": 1868 + }, + { + "epoch": 0.7156806433084434, + "grad_norm": 0.30395251512527466, + "learning_rate": 4.381704625585893e-06, + "loss": 0.6634, + "step": 1869 + }, + { + "epoch": 0.7160635650009574, + "grad_norm": 0.2886289060115814, + "learning_rate": 4.3810357855839705e-06, + "loss": 0.647, + "step": 1870 + }, + { + "epoch": 0.7164464866934712, + "grad_norm": 0.2849857211112976, + "learning_rate": 4.380366635125533e-06, + "loss": 0.6356, + "step": 1871 + }, + { + "epoch": 0.7168294083859851, + "grad_norm": 0.3216630816459656, + "learning_rate": 4.379697174321021e-06, + "loss": 0.6569, + "step": 1872 + }, + { + "epoch": 0.7172123300784989, + "grad_norm": 0.3093796372413635, + "learning_rate": 4.379027403280927e-06, + "loss": 0.6558, + "step": 1873 + }, + { + "epoch": 0.7175952517710128, + "grad_norm": 0.30553147196769714, + "learning_rate": 4.378357322115795e-06, + "loss": 0.6695, + "step": 1874 + }, + { + "epoch": 0.7179781734635267, + "grad_norm": 0.3528289496898651, + "learning_rate": 4.377686930936216e-06, + "loss": 0.6471, + "step": 1875 + }, + { + "epoch": 0.7183610951560406, + "grad_norm": 0.3141665756702423, + "learning_rate": 4.377016229852836e-06, + "loss": 0.6499, + "step": 1876 + }, + { + "epoch": 0.7187440168485545, + "grad_norm": 0.30015069246292114, + "learning_rate": 4.3763452189763514e-06, + "loss": 0.6564, + "step": 1877 + }, + { + "epoch": 0.7191269385410684, + "grad_norm": 0.3349354565143585, + "learning_rate": 4.3756738984175104e-06, + "loss": 0.6503, + "step": 1878 + }, + { + "epoch": 0.7195098602335822, + "grad_norm": 0.338704913854599, + "learning_rate": 4.375002268287109e-06, + "loss": 0.6371, + "step": 1879 + }, + { + "epoch": 0.7198927819260961, + "grad_norm": 0.3038701117038727, + "learning_rate": 4.374330328696001e-06, + "loss": 0.6472, + "step": 1880 + }, + { + "epoch": 0.72027570361861, + "grad_norm": 0.30665767192840576, + "learning_rate": 4.373658079755082e-06, + "loss": 0.6557, + "step": 1881 + }, + { + "epoch": 0.7206586253111239, + "grad_norm": 0.36826714873313904, + "learning_rate": 4.372985521575307e-06, + "loss": 0.6397, + "step": 1882 + }, + { + "epoch": 0.7210415470036378, + "grad_norm": 0.2825036942958832, + "learning_rate": 4.372312654267678e-06, + "loss": 0.6817, + "step": 1883 + }, + { + "epoch": 0.7214244686961516, + "grad_norm": 0.3002195656299591, + "learning_rate": 4.371639477943248e-06, + "loss": 0.6813, + "step": 1884 + }, + { + "epoch": 0.7218073903886655, + "grad_norm": 0.315328449010849, + "learning_rate": 4.370965992713123e-06, + "loss": 0.668, + "step": 1885 + }, + { + "epoch": 0.7221903120811795, + "grad_norm": 0.326954185962677, + "learning_rate": 4.3702921986884576e-06, + "loss": 0.6608, + "step": 1886 + }, + { + "epoch": 0.7225732337736933, + "grad_norm": 0.30041590332984924, + "learning_rate": 4.369618095980458e-06, + "loss": 0.6568, + "step": 1887 + }, + { + "epoch": 0.7229561554662072, + "grad_norm": 0.28943854570388794, + "learning_rate": 4.368943684700384e-06, + "loss": 0.6685, + "step": 1888 + }, + { + "epoch": 0.723339077158721, + "grad_norm": 0.3098449110984802, + "learning_rate": 4.368268964959542e-06, + "loss": 0.6878, + "step": 1889 + }, + { + "epoch": 0.7237219988512349, + "grad_norm": 0.34474292397499084, + "learning_rate": 4.3675939368692925e-06, + "loss": 0.6672, + "step": 1890 + }, + { + "epoch": 0.7241049205437488, + "grad_norm": 0.28440067172050476, + "learning_rate": 4.366918600541045e-06, + "loss": 0.666, + "step": 1891 + }, + { + "epoch": 0.7244878422362627, + "grad_norm": 0.29537713527679443, + "learning_rate": 4.366242956086262e-06, + "loss": 0.6772, + "step": 1892 + }, + { + "epoch": 0.7248707639287766, + "grad_norm": 0.3174888491630554, + "learning_rate": 4.365567003616455e-06, + "loss": 0.6373, + "step": 1893 + }, + { + "epoch": 0.7252536856212904, + "grad_norm": 0.31100913882255554, + "learning_rate": 4.364890743243187e-06, + "loss": 0.6606, + "step": 1894 + }, + { + "epoch": 0.7256366073138043, + "grad_norm": 0.29556766152381897, + "learning_rate": 4.364214175078071e-06, + "loss": 0.6574, + "step": 1895 + }, + { + "epoch": 0.7260195290063182, + "grad_norm": 0.2948823869228363, + "learning_rate": 4.363537299232773e-06, + "loss": 0.6535, + "step": 1896 + }, + { + "epoch": 0.7264024506988321, + "grad_norm": 0.3056114912033081, + "learning_rate": 4.362860115819007e-06, + "loss": 0.6618, + "step": 1897 + }, + { + "epoch": 0.726785372391346, + "grad_norm": 0.30672791600227356, + "learning_rate": 4.36218262494854e-06, + "loss": 0.6685, + "step": 1898 + }, + { + "epoch": 0.7271682940838599, + "grad_norm": 0.2834531366825104, + "learning_rate": 4.3615048267331875e-06, + "loss": 0.6522, + "step": 1899 + }, + { + "epoch": 0.7275512157763737, + "grad_norm": 0.28408098220825195, + "learning_rate": 4.360826721284818e-06, + "loss": 0.624, + "step": 1900 + }, + { + "epoch": 0.7279341374688876, + "grad_norm": 0.2917988896369934, + "learning_rate": 4.360148308715349e-06, + "loss": 0.6591, + "step": 1901 + }, + { + "epoch": 0.7283170591614015, + "grad_norm": 0.2883460223674774, + "learning_rate": 4.359469589136752e-06, + "loss": 0.6486, + "step": 1902 + }, + { + "epoch": 0.7286999808539154, + "grad_norm": 0.284496933221817, + "learning_rate": 4.358790562661044e-06, + "loss": 0.6544, + "step": 1903 + }, + { + "epoch": 0.7290829025464293, + "grad_norm": 0.29910603165626526, + "learning_rate": 4.358111229400296e-06, + "loss": 0.65, + "step": 1904 + }, + { + "epoch": 0.7294658242389431, + "grad_norm": 0.29271793365478516, + "learning_rate": 4.357431589466629e-06, + "loss": 0.6547, + "step": 1905 + }, + { + "epoch": 0.729848745931457, + "grad_norm": 0.2830300033092499, + "learning_rate": 4.356751642972214e-06, + "loss": 0.6682, + "step": 1906 + }, + { + "epoch": 0.7302316676239708, + "grad_norm": 0.2894238829612732, + "learning_rate": 4.356071390029275e-06, + "loss": 0.6732, + "step": 1907 + }, + { + "epoch": 0.7306145893164848, + "grad_norm": 0.2848636209964752, + "learning_rate": 4.355390830750081e-06, + "loss": 0.6644, + "step": 1908 + }, + { + "epoch": 0.7309975110089987, + "grad_norm": 0.27900969982147217, + "learning_rate": 4.35470996524696e-06, + "loss": 0.6324, + "step": 1909 + }, + { + "epoch": 0.7313804327015125, + "grad_norm": 0.2817847430706024, + "learning_rate": 4.354028793632281e-06, + "loss": 0.6676, + "step": 1910 + }, + { + "epoch": 0.7317633543940264, + "grad_norm": 0.28212547302246094, + "learning_rate": 4.3533473160184725e-06, + "loss": 0.6704, + "step": 1911 + }, + { + "epoch": 0.7321462760865403, + "grad_norm": 0.28725746273994446, + "learning_rate": 4.352665532518007e-06, + "loss": 0.6529, + "step": 1912 + }, + { + "epoch": 0.7325291977790542, + "grad_norm": 0.297559916973114, + "learning_rate": 4.3519834432434095e-06, + "loss": 0.6707, + "step": 1913 + }, + { + "epoch": 0.7329121194715681, + "grad_norm": 0.2973473072052002, + "learning_rate": 4.351301048307257e-06, + "loss": 0.6776, + "step": 1914 + }, + { + "epoch": 0.733295041164082, + "grad_norm": 0.2814374566078186, + "learning_rate": 4.350618347822175e-06, + "loss": 0.6393, + "step": 1915 + }, + { + "epoch": 0.7336779628565958, + "grad_norm": 0.2936658263206482, + "learning_rate": 4.349935341900841e-06, + "loss": 0.6767, + "step": 1916 + }, + { + "epoch": 0.7340608845491097, + "grad_norm": 0.29175832867622375, + "learning_rate": 4.349252030655982e-06, + "loss": 0.6656, + "step": 1917 + }, + { + "epoch": 0.7344438062416236, + "grad_norm": 0.2962329387664795, + "learning_rate": 4.3485684142003734e-06, + "loss": 0.6584, + "step": 1918 + }, + { + "epoch": 0.7348267279341375, + "grad_norm": 0.2833195924758911, + "learning_rate": 4.347884492646846e-06, + "loss": 0.6468, + "step": 1919 + }, + { + "epoch": 0.7352096496266514, + "grad_norm": 0.2913295030593872, + "learning_rate": 4.347200266108275e-06, + "loss": 0.6416, + "step": 1920 + }, + { + "epoch": 0.7355925713191652, + "grad_norm": 0.29444143176078796, + "learning_rate": 4.346515734697592e-06, + "loss": 0.6737, + "step": 1921 + }, + { + "epoch": 0.7359754930116791, + "grad_norm": 0.2952899932861328, + "learning_rate": 4.345830898527773e-06, + "loss": 0.656, + "step": 1922 + }, + { + "epoch": 0.7363584147041929, + "grad_norm": 0.2749316990375519, + "learning_rate": 4.345145757711849e-06, + "loss": 0.629, + "step": 1923 + }, + { + "epoch": 0.7367413363967069, + "grad_norm": 0.2893662750720978, + "learning_rate": 4.344460312362899e-06, + "loss": 0.6492, + "step": 1924 + }, + { + "epoch": 0.7371242580892208, + "grad_norm": 0.28754866123199463, + "learning_rate": 4.343774562594053e-06, + "loss": 0.6572, + "step": 1925 + }, + { + "epoch": 0.7375071797817346, + "grad_norm": 0.2948099970817566, + "learning_rate": 4.343088508518489e-06, + "loss": 0.6643, + "step": 1926 + }, + { + "epoch": 0.7378901014742485, + "grad_norm": 0.28945931792259216, + "learning_rate": 4.34240215024944e-06, + "loss": 0.6538, + "step": 1927 + }, + { + "epoch": 0.7382730231667624, + "grad_norm": 0.29446089267730713, + "learning_rate": 4.3417154879001845e-06, + "loss": 0.6541, + "step": 1928 + }, + { + "epoch": 0.7386559448592763, + "grad_norm": 0.28317776322364807, + "learning_rate": 4.341028521584053e-06, + "loss": 0.6502, + "step": 1929 + }, + { + "epoch": 0.7390388665517902, + "grad_norm": 0.28600841760635376, + "learning_rate": 4.340341251414427e-06, + "loss": 0.6562, + "step": 1930 + }, + { + "epoch": 0.739421788244304, + "grad_norm": 0.2847815155982971, + "learning_rate": 4.339653677504737e-06, + "loss": 0.6656, + "step": 1931 + }, + { + "epoch": 0.7398047099368179, + "grad_norm": 0.29913073778152466, + "learning_rate": 4.338965799968465e-06, + "loss": 0.6739, + "step": 1932 + }, + { + "epoch": 0.7401876316293318, + "grad_norm": 0.276906818151474, + "learning_rate": 4.33827761891914e-06, + "loss": 0.6686, + "step": 1933 + }, + { + "epoch": 0.7405705533218457, + "grad_norm": 0.27966615557670593, + "learning_rate": 4.337589134470345e-06, + "loss": 0.6497, + "step": 1934 + }, + { + "epoch": 0.7409534750143596, + "grad_norm": 0.28722241520881653, + "learning_rate": 4.33690034673571e-06, + "loss": 0.6552, + "step": 1935 + }, + { + "epoch": 0.7413363967068735, + "grad_norm": 0.28933677077293396, + "learning_rate": 4.336211255828917e-06, + "loss": 0.6609, + "step": 1936 + }, + { + "epoch": 0.7417193183993873, + "grad_norm": 0.2806028723716736, + "learning_rate": 4.335521861863698e-06, + "loss": 0.646, + "step": 1937 + }, + { + "epoch": 0.7421022400919012, + "grad_norm": 0.2815571129322052, + "learning_rate": 4.334832164953833e-06, + "loss": 0.6587, + "step": 1938 + }, + { + "epoch": 0.742485161784415, + "grad_norm": 0.279063880443573, + "learning_rate": 4.334142165213154e-06, + "loss": 0.6603, + "step": 1939 + }, + { + "epoch": 0.742868083476929, + "grad_norm": 0.28541961312294006, + "learning_rate": 4.333451862755543e-06, + "loss": 0.6857, + "step": 1940 + }, + { + "epoch": 0.7432510051694429, + "grad_norm": 0.2963019907474518, + "learning_rate": 4.332761257694931e-06, + "loss": 0.6595, + "step": 1941 + }, + { + "epoch": 0.7436339268619567, + "grad_norm": 0.27836063504219055, + "learning_rate": 4.332070350145297e-06, + "loss": 0.6561, + "step": 1942 + }, + { + "epoch": 0.7440168485544706, + "grad_norm": 0.28990232944488525, + "learning_rate": 4.3313791402206765e-06, + "loss": 0.6476, + "step": 1943 + }, + { + "epoch": 0.7443997702469844, + "grad_norm": 0.29161038994789124, + "learning_rate": 4.330687628035147e-06, + "loss": 0.6857, + "step": 1944 + }, + { + "epoch": 0.7447826919394984, + "grad_norm": 0.291381299495697, + "learning_rate": 4.329995813702842e-06, + "loss": 0.6229, + "step": 1945 + }, + { + "epoch": 0.7451656136320123, + "grad_norm": 0.28754791617393494, + "learning_rate": 4.32930369733794e-06, + "loss": 0.6511, + "step": 1946 + }, + { + "epoch": 0.7455485353245261, + "grad_norm": 0.28677695989608765, + "learning_rate": 4.328611279054673e-06, + "loss": 0.6583, + "step": 1947 + }, + { + "epoch": 0.74593145701704, + "grad_norm": 0.29700377583503723, + "learning_rate": 4.327918558967321e-06, + "loss": 0.6356, + "step": 1948 + }, + { + "epoch": 0.7463143787095539, + "grad_norm": 0.2861420214176178, + "learning_rate": 4.327225537190215e-06, + "loss": 0.6599, + "step": 1949 + }, + { + "epoch": 0.7466973004020678, + "grad_norm": 0.29210418462753296, + "learning_rate": 4.326532213837735e-06, + "loss": 0.6394, + "step": 1950 + }, + { + "epoch": 0.7470802220945817, + "grad_norm": 0.28130602836608887, + "learning_rate": 4.325838589024312e-06, + "loss": 0.6691, + "step": 1951 + }, + { + "epoch": 0.7474631437870956, + "grad_norm": 0.28725385665893555, + "learning_rate": 4.3251446628644224e-06, + "loss": 0.6825, + "step": 1952 + }, + { + "epoch": 0.7478460654796094, + "grad_norm": 0.29483655095100403, + "learning_rate": 4.324450435472598e-06, + "loss": 0.661, + "step": 1953 + }, + { + "epoch": 0.7482289871721233, + "grad_norm": 0.2880934476852417, + "learning_rate": 4.323755906963419e-06, + "loss": 0.6563, + "step": 1954 + }, + { + "epoch": 0.7486119088646371, + "grad_norm": 0.29493528604507446, + "learning_rate": 4.32306107745151e-06, + "loss": 0.6526, + "step": 1955 + }, + { + "epoch": 0.7489948305571511, + "grad_norm": 0.2936811149120331, + "learning_rate": 4.322365947051554e-06, + "loss": 0.6359, + "step": 1956 + }, + { + "epoch": 0.749377752249665, + "grad_norm": 0.282199889421463, + "learning_rate": 4.321670515878277e-06, + "loss": 0.6715, + "step": 1957 + }, + { + "epoch": 0.7497606739421788, + "grad_norm": 0.2862069606781006, + "learning_rate": 4.320974784046458e-06, + "loss": 0.6765, + "step": 1958 + }, + { + "epoch": 0.7501435956346927, + "grad_norm": 0.2906152904033661, + "learning_rate": 4.320278751670922e-06, + "loss": 0.6715, + "step": 1959 + }, + { + "epoch": 0.7505265173272065, + "grad_norm": 0.2825862169265747, + "learning_rate": 4.319582418866549e-06, + "loss": 0.6532, + "step": 1960 + }, + { + "epoch": 0.7509094390197205, + "grad_norm": 0.28862181305885315, + "learning_rate": 4.318885785748264e-06, + "loss": 0.6526, + "step": 1961 + }, + { + "epoch": 0.7512923607122344, + "grad_norm": 0.28897467255592346, + "learning_rate": 4.318188852431043e-06, + "loss": 0.648, + "step": 1962 + }, + { + "epoch": 0.7516752824047482, + "grad_norm": 0.28821879625320435, + "learning_rate": 4.317491619029911e-06, + "loss": 0.6303, + "step": 1963 + }, + { + "epoch": 0.7520582040972621, + "grad_norm": 0.290291428565979, + "learning_rate": 4.316794085659946e-06, + "loss": 0.6506, + "step": 1964 + }, + { + "epoch": 0.752441125789776, + "grad_norm": 0.29138660430908203, + "learning_rate": 4.316096252436271e-06, + "loss": 0.6614, + "step": 1965 + }, + { + "epoch": 0.7528240474822899, + "grad_norm": 0.3001135587692261, + "learning_rate": 4.315398119474059e-06, + "loss": 0.653, + "step": 1966 + }, + { + "epoch": 0.7532069691748038, + "grad_norm": 0.29022935032844543, + "learning_rate": 4.314699686888536e-06, + "loss": 0.6664, + "step": 1967 + }, + { + "epoch": 0.7535898908673176, + "grad_norm": 0.30858826637268066, + "learning_rate": 4.314000954794974e-06, + "loss": 0.6643, + "step": 1968 + }, + { + "epoch": 0.7539728125598315, + "grad_norm": 0.29346123337745667, + "learning_rate": 4.313301923308696e-06, + "loss": 0.6671, + "step": 1969 + }, + { + "epoch": 0.7543557342523454, + "grad_norm": 0.28771576285362244, + "learning_rate": 4.312602592545073e-06, + "loss": 0.6453, + "step": 1970 + }, + { + "epoch": 0.7547386559448592, + "grad_norm": 0.28616222739219666, + "learning_rate": 4.311902962619529e-06, + "loss": 0.6666, + "step": 1971 + }, + { + "epoch": 0.7551215776373732, + "grad_norm": 0.2930017113685608, + "learning_rate": 4.311203033647532e-06, + "loss": 0.6458, + "step": 1972 + }, + { + "epoch": 0.7555044993298871, + "grad_norm": 0.2986257076263428, + "learning_rate": 4.310502805744603e-06, + "loss": 0.6402, + "step": 1973 + }, + { + "epoch": 0.7558874210224009, + "grad_norm": 0.29714515805244446, + "learning_rate": 4.3098022790263115e-06, + "loss": 0.6673, + "step": 1974 + }, + { + "epoch": 0.7562703427149148, + "grad_norm": 0.2817387580871582, + "learning_rate": 4.309101453608276e-06, + "loss": 0.6604, + "step": 1975 + }, + { + "epoch": 0.7566532644074286, + "grad_norm": 0.29210686683654785, + "learning_rate": 4.3084003296061655e-06, + "loss": 0.6518, + "step": 1976 + }, + { + "epoch": 0.7570361860999426, + "grad_norm": 0.29889750480651855, + "learning_rate": 4.307698907135697e-06, + "loss": 0.6293, + "step": 1977 + }, + { + "epoch": 0.7574191077924565, + "grad_norm": 0.28968730568885803, + "learning_rate": 4.306997186312637e-06, + "loss": 0.662, + "step": 1978 + }, + { + "epoch": 0.7578020294849703, + "grad_norm": 0.29846274852752686, + "learning_rate": 4.306295167252801e-06, + "loss": 0.6515, + "step": 1979 + }, + { + "epoch": 0.7581849511774842, + "grad_norm": 0.3046862483024597, + "learning_rate": 4.3055928500720546e-06, + "loss": 0.656, + "step": 1980 + }, + { + "epoch": 0.758567872869998, + "grad_norm": 0.2888549566268921, + "learning_rate": 4.3048902348863116e-06, + "loss": 0.6594, + "step": 1981 + }, + { + "epoch": 0.758950794562512, + "grad_norm": 0.29400166869163513, + "learning_rate": 4.304187321811536e-06, + "loss": 0.6515, + "step": 1982 + }, + { + "epoch": 0.7593337162550259, + "grad_norm": 0.28630971908569336, + "learning_rate": 4.30348411096374e-06, + "loss": 0.6469, + "step": 1983 + }, + { + "epoch": 0.7597166379475397, + "grad_norm": 0.2840445935726166, + "learning_rate": 4.302780602458987e-06, + "loss": 0.6764, + "step": 1984 + }, + { + "epoch": 0.7600995596400536, + "grad_norm": 0.2886156141757965, + "learning_rate": 4.302076796413385e-06, + "loss": 0.6644, + "step": 1985 + }, + { + "epoch": 0.7604824813325675, + "grad_norm": 0.2923414409160614, + "learning_rate": 4.301372692943096e-06, + "loss": 0.6501, + "step": 1986 + }, + { + "epoch": 0.7608654030250813, + "grad_norm": 0.2898506820201874, + "learning_rate": 4.300668292164329e-06, + "loss": 0.6712, + "step": 1987 + }, + { + "epoch": 0.7612483247175953, + "grad_norm": 0.2883935868740082, + "learning_rate": 4.299963594193342e-06, + "loss": 0.6675, + "step": 1988 + }, + { + "epoch": 0.7616312464101092, + "grad_norm": 0.2911122739315033, + "learning_rate": 4.2992585991464434e-06, + "loss": 0.666, + "step": 1989 + }, + { + "epoch": 0.762014168102623, + "grad_norm": 0.28685012459754944, + "learning_rate": 4.2985533071399865e-06, + "loss": 0.6582, + "step": 1990 + }, + { + "epoch": 0.7623970897951369, + "grad_norm": 0.28739434480667114, + "learning_rate": 4.297847718290378e-06, + "loss": 0.6596, + "step": 1991 + }, + { + "epoch": 0.7627800114876507, + "grad_norm": 0.2994686961174011, + "learning_rate": 4.297141832714073e-06, + "loss": 0.644, + "step": 1992 + }, + { + "epoch": 0.7631629331801647, + "grad_norm": 0.31029248237609863, + "learning_rate": 4.296435650527575e-06, + "loss": 0.6405, + "step": 1993 + }, + { + "epoch": 0.7635458548726786, + "grad_norm": 0.2902706563472748, + "learning_rate": 4.2957291718474356e-06, + "loss": 0.6474, + "step": 1994 + }, + { + "epoch": 0.7639287765651924, + "grad_norm": 0.2979045808315277, + "learning_rate": 4.2950223967902545e-06, + "loss": 0.6488, + "step": 1995 + }, + { + "epoch": 0.7643116982577063, + "grad_norm": 0.3013862371444702, + "learning_rate": 4.2943153254726835e-06, + "loss": 0.6656, + "step": 1996 + }, + { + "epoch": 0.7646946199502201, + "grad_norm": 0.30337440967559814, + "learning_rate": 4.2936079580114215e-06, + "loss": 0.6534, + "step": 1997 + }, + { + "epoch": 0.7650775416427341, + "grad_norm": 0.3107244074344635, + "learning_rate": 4.292900294523216e-06, + "loss": 0.6734, + "step": 1998 + }, + { + "epoch": 0.765460463335248, + "grad_norm": 0.29626262187957764, + "learning_rate": 4.292192335124863e-06, + "loss": 0.6505, + "step": 1999 + }, + { + "epoch": 0.7658433850277618, + "grad_norm": 0.28887301683425903, + "learning_rate": 4.291484079933208e-06, + "loss": 0.6537, + "step": 2000 + }, + { + "epoch": 0.7662263067202757, + "grad_norm": 0.28143438696861267, + "learning_rate": 4.290775529065146e-06, + "loss": 0.6745, + "step": 2001 + }, + { + "epoch": 0.7666092284127896, + "grad_norm": 0.2881353199481964, + "learning_rate": 4.290066682637622e-06, + "loss": 0.6453, + "step": 2002 + }, + { + "epoch": 0.7669921501053034, + "grad_norm": 0.2832963168621063, + "learning_rate": 4.289357540767625e-06, + "loss": 0.652, + "step": 2003 + }, + { + "epoch": 0.7673750717978174, + "grad_norm": 0.29034432768821716, + "learning_rate": 4.288648103572196e-06, + "loss": 0.6567, + "step": 2004 + }, + { + "epoch": 0.7677579934903312, + "grad_norm": 0.28968292474746704, + "learning_rate": 4.287938371168425e-06, + "loss": 0.6488, + "step": 2005 + }, + { + "epoch": 0.7681409151828451, + "grad_norm": 0.3006720244884491, + "learning_rate": 4.287228343673451e-06, + "loss": 0.6778, + "step": 2006 + }, + { + "epoch": 0.768523836875359, + "grad_norm": 0.28195542097091675, + "learning_rate": 4.28651802120446e-06, + "loss": 0.6434, + "step": 2007 + }, + { + "epoch": 0.7689067585678728, + "grad_norm": 0.2804199457168579, + "learning_rate": 4.285807403878687e-06, + "loss": 0.6645, + "step": 2008 + }, + { + "epoch": 0.7692896802603868, + "grad_norm": 0.2874583899974823, + "learning_rate": 4.2850964918134175e-06, + "loss": 0.6535, + "step": 2009 + }, + { + "epoch": 0.7696726019529007, + "grad_norm": 0.29446691274642944, + "learning_rate": 4.284385285125983e-06, + "loss": 0.6552, + "step": 2010 + }, + { + "epoch": 0.7700555236454145, + "grad_norm": 0.28751301765441895, + "learning_rate": 4.283673783933767e-06, + "loss": 0.6768, + "step": 2011 + }, + { + "epoch": 0.7704384453379284, + "grad_norm": 0.30472174286842346, + "learning_rate": 4.282961988354197e-06, + "loss": 0.6275, + "step": 2012 + }, + { + "epoch": 0.7708213670304422, + "grad_norm": 0.2874471843242645, + "learning_rate": 4.2822498985047535e-06, + "loss": 0.6547, + "step": 2013 + }, + { + "epoch": 0.7712042887229562, + "grad_norm": 0.2831214666366577, + "learning_rate": 4.281537514502962e-06, + "loss": 0.6268, + "step": 2014 + }, + { + "epoch": 0.7715872104154701, + "grad_norm": 0.28203102946281433, + "learning_rate": 4.280824836466401e-06, + "loss": 0.6709, + "step": 2015 + }, + { + "epoch": 0.7719701321079839, + "grad_norm": 0.2814365029335022, + "learning_rate": 4.280111864512693e-06, + "loss": 0.6597, + "step": 2016 + }, + { + "epoch": 0.7723530538004978, + "grad_norm": 0.29994678497314453, + "learning_rate": 4.279398598759511e-06, + "loss": 0.6542, + "step": 2017 + }, + { + "epoch": 0.7727359754930117, + "grad_norm": 0.2828892767429352, + "learning_rate": 4.278685039324576e-06, + "loss": 0.644, + "step": 2018 + }, + { + "epoch": 0.7731188971855255, + "grad_norm": 0.2878169119358063, + "learning_rate": 4.277971186325658e-06, + "loss": 0.6564, + "step": 2019 + }, + { + "epoch": 0.7735018188780395, + "grad_norm": 0.28668612241744995, + "learning_rate": 4.277257039880576e-06, + "loss": 0.6468, + "step": 2020 + }, + { + "epoch": 0.7738847405705533, + "grad_norm": 0.2916262149810791, + "learning_rate": 4.276542600107195e-06, + "loss": 0.6558, + "step": 2021 + }, + { + "epoch": 0.7742676622630672, + "grad_norm": 0.29401564598083496, + "learning_rate": 4.2758278671234315e-06, + "loss": 0.6591, + "step": 2022 + }, + { + "epoch": 0.7746505839555811, + "grad_norm": 0.3051813244819641, + "learning_rate": 4.275112841047249e-06, + "loss": 0.6466, + "step": 2023 + }, + { + "epoch": 0.7750335056480949, + "grad_norm": 0.2905217707157135, + "learning_rate": 4.274397521996658e-06, + "loss": 0.6694, + "step": 2024 + }, + { + "epoch": 0.7754164273406089, + "grad_norm": 0.2871151864528656, + "learning_rate": 4.273681910089721e-06, + "loss": 0.6376, + "step": 2025 + }, + { + "epoch": 0.7757993490331228, + "grad_norm": 0.28331542015075684, + "learning_rate": 4.272966005444545e-06, + "loss": 0.6481, + "step": 2026 + }, + { + "epoch": 0.7761822707256366, + "grad_norm": 0.28714999556541443, + "learning_rate": 4.272249808179286e-06, + "loss": 0.6762, + "step": 2027 + }, + { + "epoch": 0.7765651924181505, + "grad_norm": 0.29239681363105774, + "learning_rate": 4.271533318412151e-06, + "loss": 0.6519, + "step": 2028 + }, + { + "epoch": 0.7769481141106643, + "grad_norm": 0.28685858845710754, + "learning_rate": 4.2708165362613915e-06, + "loss": 0.6599, + "step": 2029 + }, + { + "epoch": 0.7773310358031783, + "grad_norm": 0.30236196517944336, + "learning_rate": 4.270099461845311e-06, + "loss": 0.6368, + "step": 2030 + }, + { + "epoch": 0.7777139574956922, + "grad_norm": 0.2790868282318115, + "learning_rate": 4.2693820952822594e-06, + "loss": 0.6438, + "step": 2031 + }, + { + "epoch": 0.778096879188206, + "grad_norm": 0.2869567275047302, + "learning_rate": 4.268664436690633e-06, + "loss": 0.6599, + "step": 2032 + }, + { + "epoch": 0.7784798008807199, + "grad_norm": 0.29027190804481506, + "learning_rate": 4.2679464861888796e-06, + "loss": 0.6568, + "step": 2033 + }, + { + "epoch": 0.7788627225732337, + "grad_norm": 0.28916051983833313, + "learning_rate": 4.267228243895494e-06, + "loss": 0.6513, + "step": 2034 + }, + { + "epoch": 0.7792456442657476, + "grad_norm": 0.2791849672794342, + "learning_rate": 4.266509709929017e-06, + "loss": 0.6332, + "step": 2035 + }, + { + "epoch": 0.7796285659582616, + "grad_norm": 0.2916945219039917, + "learning_rate": 4.265790884408042e-06, + "loss": 0.6468, + "step": 2036 + }, + { + "epoch": 0.7800114876507754, + "grad_norm": 0.30323725938796997, + "learning_rate": 4.265071767451206e-06, + "loss": 0.6513, + "step": 2037 + }, + { + "epoch": 0.7803944093432893, + "grad_norm": 0.2838743031024933, + "learning_rate": 4.264352359177196e-06, + "loss": 0.6575, + "step": 2038 + }, + { + "epoch": 0.7807773310358032, + "grad_norm": 0.2878491282463074, + "learning_rate": 4.263632659704748e-06, + "loss": 0.6424, + "step": 2039 + }, + { + "epoch": 0.781160252728317, + "grad_norm": 0.29993826150894165, + "learning_rate": 4.262912669152644e-06, + "loss": 0.6447, + "step": 2040 + }, + { + "epoch": 0.781543174420831, + "grad_norm": 0.29164037108421326, + "learning_rate": 4.2621923876397175e-06, + "loss": 0.6704, + "step": 2041 + }, + { + "epoch": 0.7819260961133448, + "grad_norm": 0.27584871649742126, + "learning_rate": 4.261471815284846e-06, + "loss": 0.6518, + "step": 2042 + }, + { + "epoch": 0.7823090178058587, + "grad_norm": 0.2810426354408264, + "learning_rate": 4.260750952206957e-06, + "loss": 0.6738, + "step": 2043 + }, + { + "epoch": 0.7826919394983726, + "grad_norm": 0.2827095687389374, + "learning_rate": 4.260029798525025e-06, + "loss": 0.6357, + "step": 2044 + }, + { + "epoch": 0.7830748611908864, + "grad_norm": 0.3064935505390167, + "learning_rate": 4.2593083543580735e-06, + "loss": 0.6453, + "step": 2045 + }, + { + "epoch": 0.7834577828834004, + "grad_norm": 0.2863493263721466, + "learning_rate": 4.258586619825176e-06, + "loss": 0.66, + "step": 2046 + }, + { + "epoch": 0.7838407045759143, + "grad_norm": 0.294557124376297, + "learning_rate": 4.2578645950454475e-06, + "loss": 0.6747, + "step": 2047 + }, + { + "epoch": 0.7842236262684281, + "grad_norm": 0.2933705747127533, + "learning_rate": 4.257142280138058e-06, + "loss": 0.6598, + "step": 2048 + }, + { + "epoch": 0.784606547960942, + "grad_norm": 0.2848224639892578, + "learning_rate": 4.256419675222222e-06, + "loss": 0.66, + "step": 2049 + }, + { + "epoch": 0.7849894696534558, + "grad_norm": 0.28834375739097595, + "learning_rate": 4.255696780417201e-06, + "loss": 0.6373, + "step": 2050 + }, + { + "epoch": 0.7853723913459697, + "grad_norm": 0.28696176409721375, + "learning_rate": 4.2549735958423065e-06, + "loss": 0.6574, + "step": 2051 + }, + { + "epoch": 0.7857553130384837, + "grad_norm": 0.2905980944633484, + "learning_rate": 4.254250121616895e-06, + "loss": 0.635, + "step": 2052 + }, + { + "epoch": 0.7861382347309975, + "grad_norm": 0.2829233407974243, + "learning_rate": 4.253526357860376e-06, + "loss": 0.6393, + "step": 2053 + }, + { + "epoch": 0.7865211564235114, + "grad_norm": 0.2907569408416748, + "learning_rate": 4.2528023046922005e-06, + "loss": 0.6564, + "step": 2054 + }, + { + "epoch": 0.7869040781160253, + "grad_norm": 0.27671337127685547, + "learning_rate": 4.252077962231872e-06, + "loss": 0.6582, + "step": 2055 + }, + { + "epoch": 0.7872869998085391, + "grad_norm": 0.2857098877429962, + "learning_rate": 4.251353330598938e-06, + "loss": 0.6529, + "step": 2056 + }, + { + "epoch": 0.7876699215010531, + "grad_norm": 0.28680211305618286, + "learning_rate": 4.250628409912998e-06, + "loss": 0.6568, + "step": 2057 + }, + { + "epoch": 0.7880528431935669, + "grad_norm": 0.28168386220932007, + "learning_rate": 4.2499032002936955e-06, + "loss": 0.6642, + "step": 2058 + }, + { + "epoch": 0.7884357648860808, + "grad_norm": 0.27457278966903687, + "learning_rate": 4.249177701860724e-06, + "loss": 0.6267, + "step": 2059 + }, + { + "epoch": 0.7888186865785947, + "grad_norm": 0.2845628261566162, + "learning_rate": 4.2484519147338235e-06, + "loss": 0.6384, + "step": 2060 + }, + { + "epoch": 0.7892016082711085, + "grad_norm": 0.28178882598876953, + "learning_rate": 4.247725839032781e-06, + "loss": 0.6455, + "step": 2061 + }, + { + "epoch": 0.7895845299636225, + "grad_norm": 0.29785293340682983, + "learning_rate": 4.246999474877433e-06, + "loss": 0.6516, + "step": 2062 + }, + { + "epoch": 0.7899674516561364, + "grad_norm": 0.30443844199180603, + "learning_rate": 4.246272822387662e-06, + "loss": 0.6749, + "step": 2063 + }, + { + "epoch": 0.7903503733486502, + "grad_norm": 0.2878495454788208, + "learning_rate": 4.245545881683399e-06, + "loss": 0.6285, + "step": 2064 + }, + { + "epoch": 0.7907332950411641, + "grad_norm": 0.280362069606781, + "learning_rate": 4.244818652884623e-06, + "loss": 0.6376, + "step": 2065 + }, + { + "epoch": 0.7911162167336779, + "grad_norm": 0.2950364351272583, + "learning_rate": 4.244091136111359e-06, + "loss": 0.6311, + "step": 2066 + }, + { + "epoch": 0.7914991384261918, + "grad_norm": 0.2914651036262512, + "learning_rate": 4.243363331483681e-06, + "loss": 0.6459, + "step": 2067 + }, + { + "epoch": 0.7918820601187058, + "grad_norm": 0.28748568892478943, + "learning_rate": 4.242635239121709e-06, + "loss": 0.6404, + "step": 2068 + }, + { + "epoch": 0.7922649818112196, + "grad_norm": 0.29014062881469727, + "learning_rate": 4.241906859145611e-06, + "loss": 0.6436, + "step": 2069 + }, + { + "epoch": 0.7926479035037335, + "grad_norm": 0.29581183195114136, + "learning_rate": 4.241178191675606e-06, + "loss": 0.6611, + "step": 2070 + }, + { + "epoch": 0.7930308251962473, + "grad_norm": 0.285098671913147, + "learning_rate": 4.240449236831954e-06, + "loss": 0.6412, + "step": 2071 + }, + { + "epoch": 0.7934137468887612, + "grad_norm": 0.2723717987537384, + "learning_rate": 4.239719994734965e-06, + "loss": 0.6238, + "step": 2072 + }, + { + "epoch": 0.7937966685812752, + "grad_norm": 0.2964610159397125, + "learning_rate": 4.238990465505001e-06, + "loss": 0.6535, + "step": 2073 + }, + { + "epoch": 0.794179590273789, + "grad_norm": 0.2931128740310669, + "learning_rate": 4.238260649262465e-06, + "loss": 0.6777, + "step": 2074 + }, + { + "epoch": 0.7945625119663029, + "grad_norm": 0.2829414904117584, + "learning_rate": 4.237530546127809e-06, + "loss": 0.6667, + "step": 2075 + }, + { + "epoch": 0.7949454336588168, + "grad_norm": 0.29371336102485657, + "learning_rate": 4.236800156221536e-06, + "loss": 0.6654, + "step": 2076 + }, + { + "epoch": 0.7953283553513306, + "grad_norm": 0.29698020219802856, + "learning_rate": 4.2360694796641915e-06, + "loss": 0.6351, + "step": 2077 + }, + { + "epoch": 0.7957112770438446, + "grad_norm": 0.29009485244750977, + "learning_rate": 4.2353385165763715e-06, + "loss": 0.6565, + "step": 2078 + }, + { + "epoch": 0.7960941987363584, + "grad_norm": 0.2830968201160431, + "learning_rate": 4.234607267078718e-06, + "loss": 0.6514, + "step": 2079 + }, + { + "epoch": 0.7964771204288723, + "grad_norm": 0.2878909409046173, + "learning_rate": 4.2338757312919184e-06, + "loss": 0.6565, + "step": 2080 + }, + { + "epoch": 0.7968600421213862, + "grad_norm": 0.28284013271331787, + "learning_rate": 4.233143909336712e-06, + "loss": 0.637, + "step": 2081 + }, + { + "epoch": 0.7972429638139, + "grad_norm": 0.2860836386680603, + "learning_rate": 4.232411801333882e-06, + "loss": 0.6633, + "step": 2082 + }, + { + "epoch": 0.7976258855064139, + "grad_norm": 0.2868812680244446, + "learning_rate": 4.23167940740426e-06, + "loss": 0.654, + "step": 2083 + }, + { + "epoch": 0.7980088071989279, + "grad_norm": 0.2891739308834076, + "learning_rate": 4.230946727668723e-06, + "loss": 0.6667, + "step": 2084 + }, + { + "epoch": 0.7983917288914417, + "grad_norm": 0.3002691864967346, + "learning_rate": 4.230213762248198e-06, + "loss": 0.6595, + "step": 2085 + }, + { + "epoch": 0.7987746505839556, + "grad_norm": 0.2895083725452423, + "learning_rate": 4.229480511263655e-06, + "loss": 0.6366, + "step": 2086 + }, + { + "epoch": 0.7991575722764694, + "grad_norm": 0.29998913407325745, + "learning_rate": 4.2287469748361174e-06, + "loss": 0.6609, + "step": 2087 + }, + { + "epoch": 0.7995404939689833, + "grad_norm": 0.30361121892929077, + "learning_rate": 4.2280131530866495e-06, + "loss": 0.6884, + "step": 2088 + }, + { + "epoch": 0.7999234156614973, + "grad_norm": 0.2897234857082367, + "learning_rate": 4.227279046136367e-06, + "loss": 0.6719, + "step": 2089 + }, + { + "epoch": 0.8003063373540111, + "grad_norm": 0.28361034393310547, + "learning_rate": 4.226544654106429e-06, + "loss": 0.6422, + "step": 2090 + }, + { + "epoch": 0.800689259046525, + "grad_norm": 0.2946014702320099, + "learning_rate": 4.225809977118046e-06, + "loss": 0.639, + "step": 2091 + }, + { + "epoch": 0.8010721807390389, + "grad_norm": 0.2925739288330078, + "learning_rate": 4.225075015292471e-06, + "loss": 0.646, + "step": 2092 + }, + { + "epoch": 0.8014551024315527, + "grad_norm": 0.296834796667099, + "learning_rate": 4.224339768751008e-06, + "loss": 0.6659, + "step": 2093 + }, + { + "epoch": 0.8018380241240666, + "grad_norm": 0.30825576186180115, + "learning_rate": 4.223604237615004e-06, + "loss": 0.6568, + "step": 2094 + }, + { + "epoch": 0.8022209458165805, + "grad_norm": 0.30276602506637573, + "learning_rate": 4.222868422005856e-06, + "loss": 0.6322, + "step": 2095 + }, + { + "epoch": 0.8026038675090944, + "grad_norm": 0.2918483018875122, + "learning_rate": 4.222132322045009e-06, + "loss": 0.6462, + "step": 2096 + }, + { + "epoch": 0.8029867892016083, + "grad_norm": 0.2914141118526459, + "learning_rate": 4.221395937853949e-06, + "loss": 0.6463, + "step": 2097 + }, + { + "epoch": 0.8033697108941221, + "grad_norm": 0.2930675148963928, + "learning_rate": 4.220659269554217e-06, + "loss": 0.6689, + "step": 2098 + }, + { + "epoch": 0.803752632586636, + "grad_norm": 0.2935783565044403, + "learning_rate": 4.219922317267395e-06, + "loss": 0.6522, + "step": 2099 + }, + { + "epoch": 0.80413555427915, + "grad_norm": 0.28854966163635254, + "learning_rate": 4.219185081115113e-06, + "loss": 0.6435, + "step": 2100 + }, + { + "epoch": 0.8045184759716638, + "grad_norm": 0.29615625739097595, + "learning_rate": 4.218447561219049e-06, + "loss": 0.6402, + "step": 2101 + }, + { + "epoch": 0.8049013976641777, + "grad_norm": 0.2831020951271057, + "learning_rate": 4.217709757700928e-06, + "loss": 0.6549, + "step": 2102 + }, + { + "epoch": 0.8052843193566915, + "grad_norm": 0.3040040135383606, + "learning_rate": 4.216971670682522e-06, + "loss": 0.6589, + "step": 2103 + }, + { + "epoch": 0.8056672410492054, + "grad_norm": 0.2909078299999237, + "learning_rate": 4.216233300285645e-06, + "loss": 0.6438, + "step": 2104 + }, + { + "epoch": 0.8060501627417194, + "grad_norm": 0.2976897954940796, + "learning_rate": 4.215494646632165e-06, + "loss": 0.6488, + "step": 2105 + }, + { + "epoch": 0.8064330844342332, + "grad_norm": 0.2799261212348938, + "learning_rate": 4.214755709843995e-06, + "loss": 0.6438, + "step": 2106 + }, + { + "epoch": 0.8068160061267471, + "grad_norm": 0.3053889274597168, + "learning_rate": 4.214016490043089e-06, + "loss": 0.6533, + "step": 2107 + }, + { + "epoch": 0.807198927819261, + "grad_norm": 0.30414000153541565, + "learning_rate": 4.213276987351455e-06, + "loss": 0.6334, + "step": 2108 + }, + { + "epoch": 0.8075818495117748, + "grad_norm": 0.3047749996185303, + "learning_rate": 4.212537201891143e-06, + "loss": 0.6575, + "step": 2109 + }, + { + "epoch": 0.8079647712042887, + "grad_norm": 0.29599082469940186, + "learning_rate": 4.211797133784252e-06, + "loss": 0.6378, + "step": 2110 + }, + { + "epoch": 0.8083476928968026, + "grad_norm": 0.31033197045326233, + "learning_rate": 4.211056783152927e-06, + "loss": 0.6599, + "step": 2111 + }, + { + "epoch": 0.8087306145893165, + "grad_norm": 0.3023338317871094, + "learning_rate": 4.210316150119359e-06, + "loss": 0.6323, + "step": 2112 + }, + { + "epoch": 0.8091135362818304, + "grad_norm": 0.29808109998703003, + "learning_rate": 4.209575234805786e-06, + "loss": 0.6549, + "step": 2113 + }, + { + "epoch": 0.8094964579743442, + "grad_norm": 0.28547993302345276, + "learning_rate": 4.208834037334494e-06, + "loss": 0.6489, + "step": 2114 + }, + { + "epoch": 0.8098793796668581, + "grad_norm": 0.2818564176559448, + "learning_rate": 4.208092557827813e-06, + "loss": 0.6689, + "step": 2115 + }, + { + "epoch": 0.810262301359372, + "grad_norm": 0.28998851776123047, + "learning_rate": 4.207350796408122e-06, + "loss": 0.6408, + "step": 2116 + }, + { + "epoch": 0.8106452230518859, + "grad_norm": 0.28835082054138184, + "learning_rate": 4.2066087531978444e-06, + "loss": 0.6675, + "step": 2117 + }, + { + "epoch": 0.8110281447443998, + "grad_norm": 0.30105793476104736, + "learning_rate": 4.205866428319451e-06, + "loss": 0.6485, + "step": 2118 + }, + { + "epoch": 0.8114110664369136, + "grad_norm": 0.2862924337387085, + "learning_rate": 4.20512382189546e-06, + "loss": 0.6579, + "step": 2119 + }, + { + "epoch": 0.8117939881294275, + "grad_norm": 0.29366791248321533, + "learning_rate": 4.2043809340484355e-06, + "loss": 0.6453, + "step": 2120 + }, + { + "epoch": 0.8121769098219415, + "grad_norm": 0.2950849235057831, + "learning_rate": 4.203637764900987e-06, + "loss": 0.6704, + "step": 2121 + }, + { + "epoch": 0.8125598315144553, + "grad_norm": 0.2939912974834442, + "learning_rate": 4.202894314575771e-06, + "loss": 0.6549, + "step": 2122 + }, + { + "epoch": 0.8129427532069692, + "grad_norm": 0.2888442873954773, + "learning_rate": 4.202150583195491e-06, + "loss": 0.6425, + "step": 2123 + }, + { + "epoch": 0.813325674899483, + "grad_norm": 0.29482021927833557, + "learning_rate": 4.201406570882898e-06, + "loss": 0.6466, + "step": 2124 + }, + { + "epoch": 0.8137085965919969, + "grad_norm": 0.2820267379283905, + "learning_rate": 4.200662277760785e-06, + "loss": 0.6533, + "step": 2125 + }, + { + "epoch": 0.8140915182845108, + "grad_norm": 0.29276299476623535, + "learning_rate": 4.199917703951997e-06, + "loss": 0.6764, + "step": 2126 + }, + { + "epoch": 0.8144744399770247, + "grad_norm": 0.2967347502708435, + "learning_rate": 4.19917284957942e-06, + "loss": 0.6613, + "step": 2127 + }, + { + "epoch": 0.8148573616695386, + "grad_norm": 0.2874256670475006, + "learning_rate": 4.198427714765991e-06, + "loss": 0.6733, + "step": 2128 + }, + { + "epoch": 0.8152402833620525, + "grad_norm": 0.2876199185848236, + "learning_rate": 4.19768229963469e-06, + "loss": 0.667, + "step": 2129 + }, + { + "epoch": 0.8156232050545663, + "grad_norm": 0.293805867433548, + "learning_rate": 4.196936604308546e-06, + "loss": 0.6846, + "step": 2130 + }, + { + "epoch": 0.8160061267470802, + "grad_norm": 0.29738301038742065, + "learning_rate": 4.19619062891063e-06, + "loss": 0.6382, + "step": 2131 + }, + { + "epoch": 0.8163890484395941, + "grad_norm": 0.2883049249649048, + "learning_rate": 4.195444373564063e-06, + "loss": 0.6522, + "step": 2132 + }, + { + "epoch": 0.816771970132108, + "grad_norm": 0.2838800251483917, + "learning_rate": 4.194697838392013e-06, + "loss": 0.6569, + "step": 2133 + }, + { + "epoch": 0.8171548918246219, + "grad_norm": 0.2773458659648895, + "learning_rate": 4.1939510235176895e-06, + "loss": 0.6588, + "step": 2134 + }, + { + "epoch": 0.8175378135171357, + "grad_norm": 0.2932501435279846, + "learning_rate": 4.1932039290643534e-06, + "loss": 0.6636, + "step": 2135 + }, + { + "epoch": 0.8179207352096496, + "grad_norm": 0.29236772656440735, + "learning_rate": 4.192456555155306e-06, + "loss": 0.6298, + "step": 2136 + }, + { + "epoch": 0.8183036569021636, + "grad_norm": 0.29943132400512695, + "learning_rate": 4.191708901913902e-06, + "loss": 0.6605, + "step": 2137 + }, + { + "epoch": 0.8186865785946774, + "grad_norm": 0.2825752794742584, + "learning_rate": 4.190960969463536e-06, + "loss": 0.6612, + "step": 2138 + }, + { + "epoch": 0.8190695002871913, + "grad_norm": 0.29188987612724304, + "learning_rate": 4.190212757927651e-06, + "loss": 0.651, + "step": 2139 + }, + { + "epoch": 0.8194524219797051, + "grad_norm": 0.2977637052536011, + "learning_rate": 4.189464267429736e-06, + "loss": 0.6714, + "step": 2140 + }, + { + "epoch": 0.819835343672219, + "grad_norm": 0.2999652922153473, + "learning_rate": 4.188715498093326e-06, + "loss": 0.6558, + "step": 2141 + }, + { + "epoch": 0.8202182653647329, + "grad_norm": 0.28152018785476685, + "learning_rate": 4.1879664500420014e-06, + "loss": 0.6545, + "step": 2142 + }, + { + "epoch": 0.8206011870572468, + "grad_norm": 0.2833508253097534, + "learning_rate": 4.187217123399391e-06, + "loss": 0.667, + "step": 2143 + }, + { + "epoch": 0.8209841087497607, + "grad_norm": 0.2977939248085022, + "learning_rate": 4.186467518289166e-06, + "loss": 0.6519, + "step": 2144 + }, + { + "epoch": 0.8213670304422745, + "grad_norm": 0.29274484515190125, + "learning_rate": 4.185717634835047e-06, + "loss": 0.6487, + "step": 2145 + }, + { + "epoch": 0.8217499521347884, + "grad_norm": 0.2835085988044739, + "learning_rate": 4.184967473160797e-06, + "loss": 0.6472, + "step": 2146 + }, + { + "epoch": 0.8221328738273023, + "grad_norm": 0.29186347126960754, + "learning_rate": 4.184217033390229e-06, + "loss": 0.6438, + "step": 2147 + }, + { + "epoch": 0.8225157955198162, + "grad_norm": 0.2994496524333954, + "learning_rate": 4.183466315647198e-06, + "loss": 0.6501, + "step": 2148 + }, + { + "epoch": 0.8228987172123301, + "grad_norm": 0.298163503408432, + "learning_rate": 4.182715320055607e-06, + "loss": 0.6686, + "step": 2149 + }, + { + "epoch": 0.823281638904844, + "grad_norm": 0.29106858372688293, + "learning_rate": 4.181964046739404e-06, + "loss": 0.6421, + "step": 2150 + }, + { + "epoch": 0.8236645605973578, + "grad_norm": 0.3015703856945038, + "learning_rate": 4.181212495822585e-06, + "loss": 0.6518, + "step": 2151 + }, + { + "epoch": 0.8240474822898717, + "grad_norm": 0.28766122460365295, + "learning_rate": 4.180460667429188e-06, + "loss": 0.6584, + "step": 2152 + }, + { + "epoch": 0.8244304039823857, + "grad_norm": 0.28938916325569153, + "learning_rate": 4.1797085616833015e-06, + "loss": 0.6476, + "step": 2153 + }, + { + "epoch": 0.8248133256748995, + "grad_norm": 0.2908332049846649, + "learning_rate": 4.178956178709054e-06, + "loss": 0.6345, + "step": 2154 + }, + { + "epoch": 0.8251962473674134, + "grad_norm": 0.3065876066684723, + "learning_rate": 4.178203518630625e-06, + "loss": 0.6461, + "step": 2155 + }, + { + "epoch": 0.8255791690599272, + "grad_norm": 0.2874744236469269, + "learning_rate": 4.177450581572236e-06, + "loss": 0.6561, + "step": 2156 + }, + { + "epoch": 0.8259620907524411, + "grad_norm": 0.3040878176689148, + "learning_rate": 4.176697367658159e-06, + "loss": 0.6527, + "step": 2157 + }, + { + "epoch": 0.826345012444955, + "grad_norm": 0.313347190618515, + "learning_rate": 4.175943877012706e-06, + "loss": 0.6602, + "step": 2158 + }, + { + "epoch": 0.8267279341374689, + "grad_norm": 0.28679871559143066, + "learning_rate": 4.175190109760237e-06, + "loss": 0.6439, + "step": 2159 + }, + { + "epoch": 0.8271108558299828, + "grad_norm": 0.3119087517261505, + "learning_rate": 4.17443606602516e-06, + "loss": 0.639, + "step": 2160 + }, + { + "epoch": 0.8274937775224966, + "grad_norm": 0.2842283844947815, + "learning_rate": 4.173681745931925e-06, + "loss": 0.6531, + "step": 2161 + }, + { + "epoch": 0.8278766992150105, + "grad_norm": 0.28919950127601624, + "learning_rate": 4.1729271496050284e-06, + "loss": 0.6349, + "step": 2162 + }, + { + "epoch": 0.8282596209075244, + "grad_norm": 0.29278960824012756, + "learning_rate": 4.172172277169015e-06, + "loss": 0.6223, + "step": 2163 + }, + { + "epoch": 0.8286425426000383, + "grad_norm": 0.291665643453598, + "learning_rate": 4.171417128748472e-06, + "loss": 0.6451, + "step": 2164 + }, + { + "epoch": 0.8290254642925522, + "grad_norm": 0.2924218773841858, + "learning_rate": 4.170661704468033e-06, + "loss": 0.6545, + "step": 2165 + }, + { + "epoch": 0.829408385985066, + "grad_norm": 0.2932129204273224, + "learning_rate": 4.169906004452378e-06, + "loss": 0.64, + "step": 2166 + }, + { + "epoch": 0.8297913076775799, + "grad_norm": 0.30480051040649414, + "learning_rate": 4.169150028826232e-06, + "loss": 0.6519, + "step": 2167 + }, + { + "epoch": 0.8301742293700938, + "grad_norm": 0.2939715087413788, + "learning_rate": 4.168393777714364e-06, + "loss": 0.66, + "step": 2168 + }, + { + "epoch": 0.8305571510626077, + "grad_norm": 0.2925496995449066, + "learning_rate": 4.167637251241592e-06, + "loss": 0.6571, + "step": 2169 + }, + { + "epoch": 0.8309400727551216, + "grad_norm": 0.3134564757347107, + "learning_rate": 4.166880449532776e-06, + "loss": 0.6606, + "step": 2170 + }, + { + "epoch": 0.8313229944476355, + "grad_norm": 0.29317769408226013, + "learning_rate": 4.166123372712823e-06, + "loss": 0.6538, + "step": 2171 + }, + { + "epoch": 0.8317059161401493, + "grad_norm": 0.28427040576934814, + "learning_rate": 4.1653660209066835e-06, + "loss": 0.6629, + "step": 2172 + }, + { + "epoch": 0.8320888378326632, + "grad_norm": 0.30414286255836487, + "learning_rate": 4.164608394239358e-06, + "loss": 0.6428, + "step": 2173 + }, + { + "epoch": 0.832471759525177, + "grad_norm": 0.2859039902687073, + "learning_rate": 4.163850492835888e-06, + "loss": 0.6483, + "step": 2174 + }, + { + "epoch": 0.832854681217691, + "grad_norm": 0.30278804898262024, + "learning_rate": 4.163092316821361e-06, + "loss": 0.6487, + "step": 2175 + }, + { + "epoch": 0.8332376029102049, + "grad_norm": 0.3010164201259613, + "learning_rate": 4.162333866320911e-06, + "loss": 0.6645, + "step": 2176 + }, + { + "epoch": 0.8336205246027187, + "grad_norm": 0.3262551426887512, + "learning_rate": 4.161575141459718e-06, + "loss": 0.6419, + "step": 2177 + }, + { + "epoch": 0.8340034462952326, + "grad_norm": 0.3013157248497009, + "learning_rate": 4.160816142363005e-06, + "loss": 0.6385, + "step": 2178 + }, + { + "epoch": 0.8343863679877465, + "grad_norm": 0.28078892827033997, + "learning_rate": 4.160056869156041e-06, + "loss": 0.6637, + "step": 2179 + }, + { + "epoch": 0.8347692896802604, + "grad_norm": 0.34636321663856506, + "learning_rate": 4.159297321964141e-06, + "loss": 0.6199, + "step": 2180 + }, + { + "epoch": 0.8351522113727743, + "grad_norm": 0.30984798073768616, + "learning_rate": 4.158537500912666e-06, + "loss": 0.6431, + "step": 2181 + }, + { + "epoch": 0.8355351330652881, + "grad_norm": 0.29280543327331543, + "learning_rate": 4.15777740612702e-06, + "loss": 0.6798, + "step": 2182 + }, + { + "epoch": 0.835918054757802, + "grad_norm": 0.3050188422203064, + "learning_rate": 4.157017037732652e-06, + "loss": 0.6618, + "step": 2183 + }, + { + "epoch": 0.8363009764503159, + "grad_norm": 0.32322168350219727, + "learning_rate": 4.156256395855061e-06, + "loss": 0.6472, + "step": 2184 + }, + { + "epoch": 0.8366838981428298, + "grad_norm": 0.29877689480781555, + "learning_rate": 4.155495480619784e-06, + "loss": 0.6285, + "step": 2185 + }, + { + "epoch": 0.8370668198353437, + "grad_norm": 0.29478585720062256, + "learning_rate": 4.154734292152408e-06, + "loss": 0.6351, + "step": 2186 + }, + { + "epoch": 0.8374497415278576, + "grad_norm": 0.28708431124687195, + "learning_rate": 4.153972830578563e-06, + "loss": 0.6482, + "step": 2187 + }, + { + "epoch": 0.8378326632203714, + "grad_norm": 0.3226141035556793, + "learning_rate": 4.153211096023927e-06, + "loss": 0.6394, + "step": 2188 + }, + { + "epoch": 0.8382155849128853, + "grad_norm": 0.3001896142959595, + "learning_rate": 4.152449088614218e-06, + "loss": 0.6491, + "step": 2189 + }, + { + "epoch": 0.8385985066053991, + "grad_norm": 0.2948077917098999, + "learning_rate": 4.151686808475204e-06, + "loss": 0.6552, + "step": 2190 + }, + { + "epoch": 0.8389814282979131, + "grad_norm": 0.30342164635658264, + "learning_rate": 4.150924255732695e-06, + "loss": 0.6501, + "step": 2191 + }, + { + "epoch": 0.839364349990427, + "grad_norm": 0.2866464853286743, + "learning_rate": 4.150161430512547e-06, + "loss": 0.6526, + "step": 2192 + }, + { + "epoch": 0.8397472716829408, + "grad_norm": 0.2831853926181793, + "learning_rate": 4.1493983329406605e-06, + "loss": 0.6462, + "step": 2193 + }, + { + "epoch": 0.8401301933754547, + "grad_norm": 0.3006346523761749, + "learning_rate": 4.148634963142982e-06, + "loss": 0.6494, + "step": 2194 + }, + { + "epoch": 0.8405131150679686, + "grad_norm": 0.30146533250808716, + "learning_rate": 4.147871321245503e-06, + "loss": 0.6524, + "step": 2195 + }, + { + "epoch": 0.8408960367604825, + "grad_norm": 0.30787044763565063, + "learning_rate": 4.147107407374258e-06, + "loss": 0.6353, + "step": 2196 + }, + { + "epoch": 0.8412789584529964, + "grad_norm": 0.30071893334388733, + "learning_rate": 4.146343221655327e-06, + "loss": 0.6527, + "step": 2197 + }, + { + "epoch": 0.8416618801455102, + "grad_norm": 0.2882702648639679, + "learning_rate": 4.145578764214836e-06, + "loss": 0.6407, + "step": 2198 + }, + { + "epoch": 0.8420448018380241, + "grad_norm": 0.2802189886569977, + "learning_rate": 4.144814035178956e-06, + "loss": 0.6425, + "step": 2199 + }, + { + "epoch": 0.842427723530538, + "grad_norm": 0.29071342945098877, + "learning_rate": 4.144049034673901e-06, + "loss": 0.6593, + "step": 2200 + }, + { + "epoch": 0.8428106452230519, + "grad_norm": 0.2915160357952118, + "learning_rate": 4.1432837628259325e-06, + "loss": 0.6523, + "step": 2201 + }, + { + "epoch": 0.8431935669155658, + "grad_norm": 0.2865479290485382, + "learning_rate": 4.142518219761354e-06, + "loss": 0.639, + "step": 2202 + }, + { + "epoch": 0.8435764886080797, + "grad_norm": 0.289639949798584, + "learning_rate": 4.141752405606514e-06, + "loss": 0.6538, + "step": 2203 + }, + { + "epoch": 0.8439594103005935, + "grad_norm": 0.29012227058410645, + "learning_rate": 4.140986320487809e-06, + "loss": 0.6522, + "step": 2204 + }, + { + "epoch": 0.8443423319931074, + "grad_norm": 0.2918872833251953, + "learning_rate": 4.1402199645316756e-06, + "loss": 0.6591, + "step": 2205 + }, + { + "epoch": 0.8447252536856212, + "grad_norm": 0.28388962149620056, + "learning_rate": 4.139453337864599e-06, + "loss": 0.653, + "step": 2206 + }, + { + "epoch": 0.8451081753781352, + "grad_norm": 0.293449342250824, + "learning_rate": 4.138686440613107e-06, + "loss": 0.6742, + "step": 2207 + }, + { + "epoch": 0.8454910970706491, + "grad_norm": 0.29294872283935547, + "learning_rate": 4.137919272903772e-06, + "loss": 0.6322, + "step": 2208 + }, + { + "epoch": 0.8458740187631629, + "grad_norm": 0.29742491245269775, + "learning_rate": 4.137151834863213e-06, + "loss": 0.6551, + "step": 2209 + }, + { + "epoch": 0.8462569404556768, + "grad_norm": 0.2894884943962097, + "learning_rate": 4.136384126618091e-06, + "loss": 0.6372, + "step": 2210 + }, + { + "epoch": 0.8466398621481906, + "grad_norm": 0.2900755703449249, + "learning_rate": 4.135616148295113e-06, + "loss": 0.6525, + "step": 2211 + }, + { + "epoch": 0.8470227838407046, + "grad_norm": 0.28828489780426025, + "learning_rate": 4.1348479000210305e-06, + "loss": 0.6503, + "step": 2212 + }, + { + "epoch": 0.8474057055332185, + "grad_norm": 0.2918212115764618, + "learning_rate": 4.1340793819226395e-06, + "loss": 0.6751, + "step": 2213 + }, + { + "epoch": 0.8477886272257323, + "grad_norm": 0.29055020213127136, + "learning_rate": 4.1333105941267805e-06, + "loss": 0.6618, + "step": 2214 + }, + { + "epoch": 0.8481715489182462, + "grad_norm": 0.2884266674518585, + "learning_rate": 4.132541536760338e-06, + "loss": 0.6556, + "step": 2215 + }, + { + "epoch": 0.8485544706107601, + "grad_norm": 0.2972971498966217, + "learning_rate": 4.1317722099502425e-06, + "loss": 0.6481, + "step": 2216 + }, + { + "epoch": 0.848937392303274, + "grad_norm": 0.2842160761356354, + "learning_rate": 4.131002613823467e-06, + "loss": 0.6648, + "step": 2217 + }, + { + "epoch": 0.8493203139957879, + "grad_norm": 0.2807370126247406, + "learning_rate": 4.13023274850703e-06, + "loss": 0.6519, + "step": 2218 + }, + { + "epoch": 0.8497032356883018, + "grad_norm": 0.282664954662323, + "learning_rate": 4.1294626141279944e-06, + "loss": 0.662, + "step": 2219 + }, + { + "epoch": 0.8500861573808156, + "grad_norm": 0.2957598567008972, + "learning_rate": 4.128692210813468e-06, + "loss": 0.672, + "step": 2220 + }, + { + "epoch": 0.8504690790733295, + "grad_norm": 0.2845847010612488, + "learning_rate": 4.127921538690601e-06, + "loss": 0.6555, + "step": 2221 + }, + { + "epoch": 0.8508520007658433, + "grad_norm": 0.286517471075058, + "learning_rate": 4.127150597886591e-06, + "loss": 0.6611, + "step": 2222 + }, + { + "epoch": 0.8512349224583573, + "grad_norm": 0.29605650901794434, + "learning_rate": 4.1263793885286775e-06, + "loss": 0.6256, + "step": 2223 + }, + { + "epoch": 0.8516178441508712, + "grad_norm": 0.2793962061405182, + "learning_rate": 4.125607910744146e-06, + "loss": 0.6548, + "step": 2224 + }, + { + "epoch": 0.852000765843385, + "grad_norm": 0.2938767075538635, + "learning_rate": 4.124836164660323e-06, + "loss": 0.672, + "step": 2225 + }, + { + "epoch": 0.8523836875358989, + "grad_norm": 0.30065855383872986, + "learning_rate": 4.124064150404584e-06, + "loss": 0.6595, + "step": 2226 + }, + { + "epoch": 0.8527666092284127, + "grad_norm": 0.2833341360092163, + "learning_rate": 4.1232918681043465e-06, + "loss": 0.6237, + "step": 2227 + }, + { + "epoch": 0.8531495309209267, + "grad_norm": 0.2910335659980774, + "learning_rate": 4.122519317887072e-06, + "loss": 0.6479, + "step": 2228 + }, + { + "epoch": 0.8535324526134406, + "grad_norm": 0.2925763726234436, + "learning_rate": 4.121746499880265e-06, + "loss": 0.6394, + "step": 2229 + }, + { + "epoch": 0.8539153743059544, + "grad_norm": 0.29192501306533813, + "learning_rate": 4.120973414211477e-06, + "loss": 0.6658, + "step": 2230 + }, + { + "epoch": 0.8542982959984683, + "grad_norm": 0.2963842749595642, + "learning_rate": 4.120200061008303e-06, + "loss": 0.6319, + "step": 2231 + }, + { + "epoch": 0.8546812176909822, + "grad_norm": 0.3151523768901825, + "learning_rate": 4.119426440398378e-06, + "loss": 0.6558, + "step": 2232 + }, + { + "epoch": 0.8550641393834961, + "grad_norm": 0.29208821058273315, + "learning_rate": 4.11865255250939e-06, + "loss": 0.6442, + "step": 2233 + }, + { + "epoch": 0.85544706107601, + "grad_norm": 0.296401709318161, + "learning_rate": 4.117878397469062e-06, + "loss": 0.6463, + "step": 2234 + }, + { + "epoch": 0.8558299827685238, + "grad_norm": 0.2935371398925781, + "learning_rate": 4.117103975405165e-06, + "loss": 0.6702, + "step": 2235 + }, + { + "epoch": 0.8562129044610377, + "grad_norm": 0.30976298451423645, + "learning_rate": 4.116329286445515e-06, + "loss": 0.6891, + "step": 2236 + }, + { + "epoch": 0.8565958261535516, + "grad_norm": 0.29204466938972473, + "learning_rate": 4.11555433071797e-06, + "loss": 0.6513, + "step": 2237 + }, + { + "epoch": 0.8569787478460654, + "grad_norm": 0.2978476583957672, + "learning_rate": 4.114779108350434e-06, + "loss": 0.6617, + "step": 2238 + }, + { + "epoch": 0.8573616695385794, + "grad_norm": 0.32826146483421326, + "learning_rate": 4.114003619470853e-06, + "loss": 0.6763, + "step": 2239 + }, + { + "epoch": 0.8577445912310933, + "grad_norm": 0.30818793177604675, + "learning_rate": 4.113227864207219e-06, + "loss": 0.6575, + "step": 2240 + }, + { + "epoch": 0.8581275129236071, + "grad_norm": 0.29063844680786133, + "learning_rate": 4.112451842687567e-06, + "loss": 0.6495, + "step": 2241 + }, + { + "epoch": 0.858510434616121, + "grad_norm": 0.29370319843292236, + "learning_rate": 4.1116755550399735e-06, + "loss": 0.6713, + "step": 2242 + }, + { + "epoch": 0.8588933563086348, + "grad_norm": 0.3022923767566681, + "learning_rate": 4.110899001392563e-06, + "loss": 0.6339, + "step": 2243 + }, + { + "epoch": 0.8592762780011488, + "grad_norm": 0.28490132093429565, + "learning_rate": 4.110122181873504e-06, + "loss": 0.6433, + "step": 2244 + }, + { + "epoch": 0.8596591996936627, + "grad_norm": 0.2953353822231293, + "learning_rate": 4.109345096611004e-06, + "loss": 0.6681, + "step": 2245 + }, + { + "epoch": 0.8600421213861765, + "grad_norm": 0.3136865198612213, + "learning_rate": 4.108567745733318e-06, + "loss": 0.649, + "step": 2246 + }, + { + "epoch": 0.8604250430786904, + "grad_norm": 0.31371670961380005, + "learning_rate": 4.1077901293687464e-06, + "loss": 0.6384, + "step": 2247 + }, + { + "epoch": 0.8608079647712042, + "grad_norm": 0.2912449538707733, + "learning_rate": 4.107012247645629e-06, + "loss": 0.6313, + "step": 2248 + }, + { + "epoch": 0.8611908864637182, + "grad_norm": 0.3179062306880951, + "learning_rate": 4.106234100692352e-06, + "loss": 0.656, + "step": 2249 + }, + { + "epoch": 0.8615738081562321, + "grad_norm": 0.3314395844936371, + "learning_rate": 4.105455688637347e-06, + "loss": 0.6606, + "step": 2250 + }, + { + "epoch": 0.8619567298487459, + "grad_norm": 0.29691487550735474, + "learning_rate": 4.104677011609085e-06, + "loss": 0.6442, + "step": 2251 + }, + { + "epoch": 0.8623396515412598, + "grad_norm": 0.3266996145248413, + "learning_rate": 4.103898069736084e-06, + "loss": 0.6441, + "step": 2252 + }, + { + "epoch": 0.8627225732337737, + "grad_norm": 0.3615022599697113, + "learning_rate": 4.103118863146905e-06, + "loss": 0.6529, + "step": 2253 + }, + { + "epoch": 0.8631054949262875, + "grad_norm": 0.28917160630226135, + "learning_rate": 4.102339391970153e-06, + "loss": 0.6371, + "step": 2254 + }, + { + "epoch": 0.8634884166188015, + "grad_norm": 0.31351974606513977, + "learning_rate": 4.101559656334476e-06, + "loss": 0.634, + "step": 2255 + }, + { + "epoch": 0.8638713383113154, + "grad_norm": 0.3009597063064575, + "learning_rate": 4.100779656368566e-06, + "loss": 0.6495, + "step": 2256 + }, + { + "epoch": 0.8642542600038292, + "grad_norm": 0.2938322424888611, + "learning_rate": 4.099999392201157e-06, + "loss": 0.6509, + "step": 2257 + }, + { + "epoch": 0.8646371816963431, + "grad_norm": 0.2903949022293091, + "learning_rate": 4.09921886396103e-06, + "loss": 0.6459, + "step": 2258 + }, + { + "epoch": 0.8650201033888569, + "grad_norm": 0.30466920137405396, + "learning_rate": 4.098438071777006e-06, + "loss": 0.6632, + "step": 2259 + }, + { + "epoch": 0.8654030250813709, + "grad_norm": 0.29187896847724915, + "learning_rate": 4.097657015777954e-06, + "loss": 0.6704, + "step": 2260 + }, + { + "epoch": 0.8657859467738848, + "grad_norm": 0.298870325088501, + "learning_rate": 4.09687569609278e-06, + "loss": 0.644, + "step": 2261 + }, + { + "epoch": 0.8661688684663986, + "grad_norm": 0.29359111189842224, + "learning_rate": 4.096094112850441e-06, + "loss": 0.6666, + "step": 2262 + }, + { + "epoch": 0.8665517901589125, + "grad_norm": 0.3013385832309723, + "learning_rate": 4.095312266179931e-06, + "loss": 0.6528, + "step": 2263 + }, + { + "epoch": 0.8669347118514263, + "grad_norm": 0.29920855164527893, + "learning_rate": 4.094530156210292e-06, + "loss": 0.6637, + "step": 2264 + }, + { + "epoch": 0.8673176335439403, + "grad_norm": 0.2906420826911926, + "learning_rate": 4.093747783070607e-06, + "loss": 0.6572, + "step": 2265 + }, + { + "epoch": 0.8677005552364542, + "grad_norm": 0.28690263628959656, + "learning_rate": 4.092965146890002e-06, + "loss": 0.6644, + "step": 2266 + }, + { + "epoch": 0.868083476928968, + "grad_norm": 0.30393892526626587, + "learning_rate": 4.09218224779765e-06, + "loss": 0.651, + "step": 2267 + }, + { + "epoch": 0.8684663986214819, + "grad_norm": 0.29441913962364197, + "learning_rate": 4.091399085922764e-06, + "loss": 0.6748, + "step": 2268 + }, + { + "epoch": 0.8688493203139958, + "grad_norm": 0.28890660405158997, + "learning_rate": 4.090615661394601e-06, + "loss": 0.6493, + "step": 2269 + }, + { + "epoch": 0.8692322420065096, + "grad_norm": 0.28217312693595886, + "learning_rate": 4.089831974342462e-06, + "loss": 0.6621, + "step": 2270 + }, + { + "epoch": 0.8696151636990236, + "grad_norm": 0.28720107674598694, + "learning_rate": 4.08904802489569e-06, + "loss": 0.629, + "step": 2271 + }, + { + "epoch": 0.8699980853915374, + "grad_norm": 0.29145902395248413, + "learning_rate": 4.088263813183674e-06, + "loss": 0.6307, + "step": 2272 + }, + { + "epoch": 0.8703810070840513, + "grad_norm": 0.2882422208786011, + "learning_rate": 4.087479339335844e-06, + "loss": 0.6531, + "step": 2273 + }, + { + "epoch": 0.8707639287765652, + "grad_norm": 0.27108249068260193, + "learning_rate": 4.086694603481673e-06, + "loss": 0.6418, + "step": 2274 + }, + { + "epoch": 0.871146850469079, + "grad_norm": 0.28425681591033936, + "learning_rate": 4.0859096057506805e-06, + "loss": 0.6432, + "step": 2275 + }, + { + "epoch": 0.871529772161593, + "grad_norm": 0.29180005192756653, + "learning_rate": 4.0851243462724245e-06, + "loss": 0.6494, + "step": 2276 + }, + { + "epoch": 0.8719126938541069, + "grad_norm": 0.2760147452354431, + "learning_rate": 4.084338825176509e-06, + "loss": 0.6551, + "step": 2277 + }, + { + "epoch": 0.8722956155466207, + "grad_norm": 0.27937087416648865, + "learning_rate": 4.083553042592581e-06, + "loss": 0.6519, + "step": 2278 + }, + { + "epoch": 0.8726785372391346, + "grad_norm": 0.28661325573921204, + "learning_rate": 4.082766998650331e-06, + "loss": 0.6527, + "step": 2279 + }, + { + "epoch": 0.8730614589316484, + "grad_norm": 0.29728376865386963, + "learning_rate": 4.08198069347949e-06, + "loss": 0.618, + "step": 2280 + }, + { + "epoch": 0.8734443806241624, + "grad_norm": 0.3050868809223175, + "learning_rate": 4.081194127209837e-06, + "loss": 0.6618, + "step": 2281 + }, + { + "epoch": 0.8738273023166763, + "grad_norm": 0.2857801914215088, + "learning_rate": 4.08040729997119e-06, + "loss": 0.6464, + "step": 2282 + }, + { + "epoch": 0.8742102240091901, + "grad_norm": 0.29985737800598145, + "learning_rate": 4.07962021189341e-06, + "loss": 0.6489, + "step": 2283 + }, + { + "epoch": 0.874593145701704, + "grad_norm": 0.2908843457698822, + "learning_rate": 4.078832863106404e-06, + "loss": 0.657, + "step": 2284 + }, + { + "epoch": 0.8749760673942178, + "grad_norm": 0.2829788625240326, + "learning_rate": 4.078045253740121e-06, + "loss": 0.6667, + "step": 2285 + }, + { + "epoch": 0.8753589890867317, + "grad_norm": 0.28448620438575745, + "learning_rate": 4.07725738392455e-06, + "loss": 0.6482, + "step": 2286 + }, + { + "epoch": 0.8757419107792457, + "grad_norm": 0.2935240864753723, + "learning_rate": 4.076469253789727e-06, + "loss": 0.6479, + "step": 2287 + }, + { + "epoch": 0.8761248324717595, + "grad_norm": 0.28987401723861694, + "learning_rate": 4.075680863465729e-06, + "loss": 0.659, + "step": 2288 + }, + { + "epoch": 0.8765077541642734, + "grad_norm": 0.2843879461288452, + "learning_rate": 4.074892213082676e-06, + "loss": 0.6618, + "step": 2289 + }, + { + "epoch": 0.8768906758567873, + "grad_norm": 0.2928538918495178, + "learning_rate": 4.074103302770731e-06, + "loss": 0.6552, + "step": 2290 + }, + { + "epoch": 0.8772735975493011, + "grad_norm": 0.28846338391304016, + "learning_rate": 4.073314132660102e-06, + "loss": 0.6462, + "step": 2291 + }, + { + "epoch": 0.8776565192418151, + "grad_norm": 0.2852533757686615, + "learning_rate": 4.072524702881036e-06, + "loss": 0.6578, + "step": 2292 + }, + { + "epoch": 0.878039440934329, + "grad_norm": 0.28771287202835083, + "learning_rate": 4.071735013563824e-06, + "loss": 0.6648, + "step": 2293 + }, + { + "epoch": 0.8784223626268428, + "grad_norm": 0.28862497210502625, + "learning_rate": 4.0709450648388034e-06, + "loss": 0.6423, + "step": 2294 + }, + { + "epoch": 0.8788052843193567, + "grad_norm": 0.2900797724723816, + "learning_rate": 4.070154856836351e-06, + "loss": 0.6323, + "step": 2295 + }, + { + "epoch": 0.8791882060118705, + "grad_norm": 0.30567964911460876, + "learning_rate": 4.069364389686886e-06, + "loss": 0.6732, + "step": 2296 + }, + { + "epoch": 0.8795711277043845, + "grad_norm": 0.297863632440567, + "learning_rate": 4.068573663520871e-06, + "loss": 0.6706, + "step": 2297 + }, + { + "epoch": 0.8799540493968984, + "grad_norm": 0.29229220747947693, + "learning_rate": 4.0677826784688136e-06, + "loss": 0.6686, + "step": 2298 + }, + { + "epoch": 0.8803369710894122, + "grad_norm": 0.28343459963798523, + "learning_rate": 4.066991434661262e-06, + "loss": 0.6478, + "step": 2299 + }, + { + "epoch": 0.8807198927819261, + "grad_norm": 0.29238826036453247, + "learning_rate": 4.066199932228805e-06, + "loss": 0.658, + "step": 2300 + }, + { + "epoch": 0.8811028144744399, + "grad_norm": 0.3182104527950287, + "learning_rate": 4.06540817130208e-06, + "loss": 0.6373, + "step": 2301 + }, + { + "epoch": 0.8814857361669538, + "grad_norm": 0.2971557080745697, + "learning_rate": 4.064616152011761e-06, + "loss": 0.644, + "step": 2302 + }, + { + "epoch": 0.8818686578594678, + "grad_norm": 0.3003726303577423, + "learning_rate": 4.063823874488569e-06, + "loss": 0.6362, + "step": 2303 + }, + { + "epoch": 0.8822515795519816, + "grad_norm": 0.28735241293907166, + "learning_rate": 4.0630313388632645e-06, + "loss": 0.6592, + "step": 2304 + }, + { + "epoch": 0.8826345012444955, + "grad_norm": 0.2918720841407776, + "learning_rate": 4.062238545266654e-06, + "loss": 0.6378, + "step": 2305 + }, + { + "epoch": 0.8830174229370094, + "grad_norm": 0.30211561918258667, + "learning_rate": 4.061445493829582e-06, + "loss": 0.6392, + "step": 2306 + }, + { + "epoch": 0.8834003446295232, + "grad_norm": 0.2915424108505249, + "learning_rate": 4.060652184682938e-06, + "loss": 0.6617, + "step": 2307 + }, + { + "epoch": 0.8837832663220372, + "grad_norm": 0.29661399126052856, + "learning_rate": 4.059858617957657e-06, + "loss": 0.6405, + "step": 2308 + }, + { + "epoch": 0.884166188014551, + "grad_norm": 0.28728193044662476, + "learning_rate": 4.059064793784711e-06, + "loss": 0.64, + "step": 2309 + }, + { + "epoch": 0.8845491097070649, + "grad_norm": 0.3088046610355377, + "learning_rate": 4.058270712295118e-06, + "loss": 0.6753, + "step": 2310 + }, + { + "epoch": 0.8849320313995788, + "grad_norm": 0.28920331597328186, + "learning_rate": 4.057476373619938e-06, + "loss": 0.6485, + "step": 2311 + }, + { + "epoch": 0.8853149530920926, + "grad_norm": 0.29318201541900635, + "learning_rate": 4.056681777890273e-06, + "loss": 0.6602, + "step": 2312 + }, + { + "epoch": 0.8856978747846066, + "grad_norm": 0.2955826222896576, + "learning_rate": 4.055886925237266e-06, + "loss": 0.6645, + "step": 2313 + }, + { + "epoch": 0.8860807964771205, + "grad_norm": 0.2962397634983063, + "learning_rate": 4.055091815792106e-06, + "loss": 0.6318, + "step": 2314 + }, + { + "epoch": 0.8864637181696343, + "grad_norm": 0.2946590483188629, + "learning_rate": 4.05429644968602e-06, + "loss": 0.6649, + "step": 2315 + }, + { + "epoch": 0.8868466398621482, + "grad_norm": 0.291528582572937, + "learning_rate": 4.0535008270502815e-06, + "loss": 0.6452, + "step": 2316 + }, + { + "epoch": 0.887229561554662, + "grad_norm": 0.28559184074401855, + "learning_rate": 4.052704948016204e-06, + "loss": 0.6632, + "step": 2317 + }, + { + "epoch": 0.8876124832471759, + "grad_norm": 0.3063846230506897, + "learning_rate": 4.051908812715143e-06, + "loss": 0.6577, + "step": 2318 + }, + { + "epoch": 0.8879954049396899, + "grad_norm": 0.29996639490127563, + "learning_rate": 4.051112421278498e-06, + "loss": 0.661, + "step": 2319 + }, + { + "epoch": 0.8883783266322037, + "grad_norm": 0.2953406572341919, + "learning_rate": 4.050315773837708e-06, + "loss": 0.646, + "step": 2320 + }, + { + "epoch": 0.8887612483247176, + "grad_norm": 0.3012860119342804, + "learning_rate": 4.049518870524259e-06, + "loss": 0.6583, + "step": 2321 + }, + { + "epoch": 0.8891441700172314, + "grad_norm": 0.2990375757217407, + "learning_rate": 4.048721711469675e-06, + "loss": 0.6394, + "step": 2322 + }, + { + "epoch": 0.8895270917097453, + "grad_norm": 0.2893383800983429, + "learning_rate": 4.047924296805522e-06, + "loss": 0.6623, + "step": 2323 + }, + { + "epoch": 0.8899100134022593, + "grad_norm": 0.29396453499794006, + "learning_rate": 4.047126626663413e-06, + "loss": 0.6473, + "step": 2324 + }, + { + "epoch": 0.8902929350947731, + "grad_norm": 0.28445231914520264, + "learning_rate": 4.046328701174995e-06, + "loss": 0.6367, + "step": 2325 + }, + { + "epoch": 0.890675856787287, + "grad_norm": 0.29325002431869507, + "learning_rate": 4.045530520471968e-06, + "loss": 0.6418, + "step": 2326 + }, + { + "epoch": 0.8910587784798009, + "grad_norm": 0.28435030579566956, + "learning_rate": 4.044732084686062e-06, + "loss": 0.6629, + "step": 2327 + }, + { + "epoch": 0.8914417001723147, + "grad_norm": 0.2926534116268158, + "learning_rate": 4.0439333939490604e-06, + "loss": 0.6488, + "step": 2328 + }, + { + "epoch": 0.8918246218648287, + "grad_norm": 0.2846304476261139, + "learning_rate": 4.043134448392782e-06, + "loss": 0.6469, + "step": 2329 + }, + { + "epoch": 0.8922075435573426, + "grad_norm": 0.29454225301742554, + "learning_rate": 4.0423352481490865e-06, + "loss": 0.663, + "step": 2330 + }, + { + "epoch": 0.8925904652498564, + "grad_norm": 0.30551332235336304, + "learning_rate": 4.041535793349881e-06, + "loss": 0.6511, + "step": 2331 + }, + { + "epoch": 0.8929733869423703, + "grad_norm": 0.30226585268974304, + "learning_rate": 4.040736084127113e-06, + "loss": 0.6418, + "step": 2332 + }, + { + "epoch": 0.8933563086348841, + "grad_norm": 0.29474323987960815, + "learning_rate": 4.039936120612767e-06, + "loss": 0.6412, + "step": 2333 + }, + { + "epoch": 0.893739230327398, + "grad_norm": 0.30319032073020935, + "learning_rate": 4.039135902938877e-06, + "loss": 0.6713, + "step": 2334 + }, + { + "epoch": 0.894122152019912, + "grad_norm": 0.3011849522590637, + "learning_rate": 4.038335431237513e-06, + "loss": 0.648, + "step": 2335 + }, + { + "epoch": 0.8945050737124258, + "grad_norm": 0.2926468551158905, + "learning_rate": 4.037534705640792e-06, + "loss": 0.65, + "step": 2336 + }, + { + "epoch": 0.8948879954049397, + "grad_norm": 0.2955300211906433, + "learning_rate": 4.036733726280867e-06, + "loss": 0.6275, + "step": 2337 + }, + { + "epoch": 0.8952709170974535, + "grad_norm": 0.2946912348270416, + "learning_rate": 4.035932493289938e-06, + "loss": 0.6608, + "step": 2338 + }, + { + "epoch": 0.8956538387899674, + "grad_norm": 0.29266324639320374, + "learning_rate": 4.0351310068002434e-06, + "loss": 0.6578, + "step": 2339 + }, + { + "epoch": 0.8960367604824814, + "grad_norm": 0.2917732298374176, + "learning_rate": 4.034329266944068e-06, + "loss": 0.651, + "step": 2340 + }, + { + "epoch": 0.8964196821749952, + "grad_norm": 0.2940225303173065, + "learning_rate": 4.033527273853731e-06, + "loss": 0.647, + "step": 2341 + }, + { + "epoch": 0.8968026038675091, + "grad_norm": 0.29845407605171204, + "learning_rate": 4.032725027661601e-06, + "loss": 0.6519, + "step": 2342 + }, + { + "epoch": 0.897185525560023, + "grad_norm": 0.29687100648880005, + "learning_rate": 4.031922528500085e-06, + "loss": 0.6442, + "step": 2343 + }, + { + "epoch": 0.8975684472525368, + "grad_norm": 0.29440659284591675, + "learning_rate": 4.03111977650163e-06, + "loss": 0.6609, + "step": 2344 + }, + { + "epoch": 0.8979513689450508, + "grad_norm": 0.2876416742801666, + "learning_rate": 4.030316771798728e-06, + "loss": 0.6568, + "step": 2345 + }, + { + "epoch": 0.8983342906375646, + "grad_norm": 0.2890982925891876, + "learning_rate": 4.029513514523911e-06, + "loss": 0.6433, + "step": 2346 + }, + { + "epoch": 0.8987172123300785, + "grad_norm": 0.2900437116622925, + "learning_rate": 4.028710004809753e-06, + "loss": 0.6384, + "step": 2347 + }, + { + "epoch": 0.8991001340225924, + "grad_norm": 0.2865522801876068, + "learning_rate": 4.027906242788869e-06, + "loss": 0.662, + "step": 2348 + }, + { + "epoch": 0.8994830557151062, + "grad_norm": 0.2898208498954773, + "learning_rate": 4.0271022285939175e-06, + "loss": 0.648, + "step": 2349 + }, + { + "epoch": 0.8998659774076201, + "grad_norm": 0.2918628454208374, + "learning_rate": 4.026297962357598e-06, + "loss": 0.6657, + "step": 2350 + }, + { + "epoch": 0.9002488991001341, + "grad_norm": 0.28812891244888306, + "learning_rate": 4.0254934442126495e-06, + "loss": 0.672, + "step": 2351 + }, + { + "epoch": 0.9006318207926479, + "grad_norm": 0.2973209023475647, + "learning_rate": 4.024688674291854e-06, + "loss": 0.6431, + "step": 2352 + }, + { + "epoch": 0.9010147424851618, + "grad_norm": 0.3002658486366272, + "learning_rate": 4.023883652728036e-06, + "loss": 0.6493, + "step": 2353 + }, + { + "epoch": 0.9013976641776756, + "grad_norm": 0.2987782061100006, + "learning_rate": 4.0230783796540595e-06, + "loss": 0.6589, + "step": 2354 + }, + { + "epoch": 0.9017805858701895, + "grad_norm": 0.3019258975982666, + "learning_rate": 4.022272855202833e-06, + "loss": 0.6543, + "step": 2355 + }, + { + "epoch": 0.9021635075627035, + "grad_norm": 0.30396032333374023, + "learning_rate": 4.021467079507303e-06, + "loss": 0.6438, + "step": 2356 + }, + { + "epoch": 0.9025464292552173, + "grad_norm": 0.2919063866138458, + "learning_rate": 4.020661052700462e-06, + "loss": 0.6235, + "step": 2357 + }, + { + "epoch": 0.9029293509477312, + "grad_norm": 0.2995392382144928, + "learning_rate": 4.0198547749153376e-06, + "loss": 0.6538, + "step": 2358 + }, + { + "epoch": 0.903312272640245, + "grad_norm": 0.2881810963153839, + "learning_rate": 4.019048246285004e-06, + "loss": 0.6459, + "step": 2359 + }, + { + "epoch": 0.9036951943327589, + "grad_norm": 0.27683988213539124, + "learning_rate": 4.018241466942575e-06, + "loss": 0.6679, + "step": 2360 + }, + { + "epoch": 0.9040781160252729, + "grad_norm": 0.3057282865047455, + "learning_rate": 4.017434437021206e-06, + "loss": 0.6717, + "step": 2361 + }, + { + "epoch": 0.9044610377177867, + "grad_norm": 0.2937275767326355, + "learning_rate": 4.016627156654094e-06, + "loss": 0.6344, + "step": 2362 + }, + { + "epoch": 0.9048439594103006, + "grad_norm": 0.28433698415756226, + "learning_rate": 4.015819625974475e-06, + "loss": 0.6257, + "step": 2363 + }, + { + "epoch": 0.9052268811028145, + "grad_norm": 0.30163976550102234, + "learning_rate": 4.015011845115631e-06, + "loss": 0.6455, + "step": 2364 + }, + { + "epoch": 0.9056098027953283, + "grad_norm": 0.2907462418079376, + "learning_rate": 4.014203814210882e-06, + "loss": 0.6772, + "step": 2365 + }, + { + "epoch": 0.9059927244878422, + "grad_norm": 0.30592411756515503, + "learning_rate": 4.0133955333935884e-06, + "loss": 0.6606, + "step": 2366 + }, + { + "epoch": 0.9063756461803562, + "grad_norm": 0.2915680706501007, + "learning_rate": 4.012587002797154e-06, + "loss": 0.6497, + "step": 2367 + }, + { + "epoch": 0.90675856787287, + "grad_norm": 0.31376343965530396, + "learning_rate": 4.0117782225550245e-06, + "loss": 0.6459, + "step": 2368 + }, + { + "epoch": 0.9071414895653839, + "grad_norm": 0.29236075282096863, + "learning_rate": 4.010969192800684e-06, + "loss": 0.6547, + "step": 2369 + }, + { + "epoch": 0.9075244112578977, + "grad_norm": 0.2886476516723633, + "learning_rate": 4.010159913667659e-06, + "loss": 0.664, + "step": 2370 + }, + { + "epoch": 0.9079073329504116, + "grad_norm": 0.30533117055892944, + "learning_rate": 4.0093503852895176e-06, + "loss": 0.6572, + "step": 2371 + }, + { + "epoch": 0.9082902546429256, + "grad_norm": 0.3040546178817749, + "learning_rate": 4.00854060779987e-06, + "loss": 0.6462, + "step": 2372 + }, + { + "epoch": 0.9086731763354394, + "grad_norm": 0.28597450256347656, + "learning_rate": 4.007730581332366e-06, + "loss": 0.6433, + "step": 2373 + }, + { + "epoch": 0.9090560980279533, + "grad_norm": 0.27774664759635925, + "learning_rate": 4.006920306020696e-06, + "loss": 0.6539, + "step": 2374 + }, + { + "epoch": 0.9094390197204671, + "grad_norm": 0.29287946224212646, + "learning_rate": 4.006109781998593e-06, + "loss": 0.662, + "step": 2375 + }, + { + "epoch": 0.909821941412981, + "grad_norm": 0.3047502636909485, + "learning_rate": 4.005299009399829e-06, + "loss": 0.6461, + "step": 2376 + }, + { + "epoch": 0.910204863105495, + "grad_norm": 0.28814053535461426, + "learning_rate": 4.00448798835822e-06, + "loss": 0.6602, + "step": 2377 + }, + { + "epoch": 0.9105877847980088, + "grad_norm": 0.28405508399009705, + "learning_rate": 4.00367671900762e-06, + "loss": 0.6353, + "step": 2378 + }, + { + "epoch": 0.9109707064905227, + "grad_norm": 0.29485368728637695, + "learning_rate": 4.0028652014819276e-06, + "loss": 0.6547, + "step": 2379 + }, + { + "epoch": 0.9113536281830366, + "grad_norm": 0.28571370244026184, + "learning_rate": 4.002053435915078e-06, + "loss": 0.6387, + "step": 2380 + }, + { + "epoch": 0.9117365498755504, + "grad_norm": 0.2909572124481201, + "learning_rate": 4.00124142244105e-06, + "loss": 0.6237, + "step": 2381 + }, + { + "epoch": 0.9121194715680643, + "grad_norm": 0.29019856452941895, + "learning_rate": 4.000429161193863e-06, + "loss": 0.6357, + "step": 2382 + }, + { + "epoch": 0.9125023932605782, + "grad_norm": 0.30379518866539, + "learning_rate": 3.999616652307576e-06, + "loss": 0.6513, + "step": 2383 + }, + { + "epoch": 0.9128853149530921, + "grad_norm": 0.279227077960968, + "learning_rate": 3.998803895916292e-06, + "loss": 0.6447, + "step": 2384 + }, + { + "epoch": 0.913268236645606, + "grad_norm": 0.2874331772327423, + "learning_rate": 3.9979908921541525e-06, + "loss": 0.6471, + "step": 2385 + }, + { + "epoch": 0.9136511583381198, + "grad_norm": 0.298424631357193, + "learning_rate": 3.997177641155338e-06, + "loss": 0.6438, + "step": 2386 + }, + { + "epoch": 0.9140340800306337, + "grad_norm": 0.291726678609848, + "learning_rate": 3.996364143054074e-06, + "loss": 0.6654, + "step": 2387 + }, + { + "epoch": 0.9144170017231477, + "grad_norm": 0.29213789105415344, + "learning_rate": 3.995550397984624e-06, + "loss": 0.6605, + "step": 2388 + }, + { + "epoch": 0.9147999234156615, + "grad_norm": 0.290400892496109, + "learning_rate": 3.9947364060812925e-06, + "loss": 0.6615, + "step": 2389 + }, + { + "epoch": 0.9151828451081754, + "grad_norm": 0.3002352714538574, + "learning_rate": 3.993922167478426e-06, + "loss": 0.6687, + "step": 2390 + }, + { + "epoch": 0.9155657668006892, + "grad_norm": 0.2931500971317291, + "learning_rate": 3.993107682310411e-06, + "loss": 0.6399, + "step": 2391 + }, + { + "epoch": 0.9159486884932031, + "grad_norm": 0.30621516704559326, + "learning_rate": 3.992292950711674e-06, + "loss": 0.6578, + "step": 2392 + }, + { + "epoch": 0.9163316101857171, + "grad_norm": 0.304832398891449, + "learning_rate": 3.991477972816682e-06, + "loss": 0.6457, + "step": 2393 + }, + { + "epoch": 0.9167145318782309, + "grad_norm": 0.2896204888820648, + "learning_rate": 3.990662748759946e-06, + "loss": 0.665, + "step": 2394 + }, + { + "epoch": 0.9170974535707448, + "grad_norm": 0.3004479706287384, + "learning_rate": 3.989847278676012e-06, + "loss": 0.6534, + "step": 2395 + }, + { + "epoch": 0.9174803752632587, + "grad_norm": 0.2964184284210205, + "learning_rate": 3.989031562699472e-06, + "loss": 0.6652, + "step": 2396 + }, + { + "epoch": 0.9178632969557725, + "grad_norm": 0.30814141035079956, + "learning_rate": 3.988215600964956e-06, + "loss": 0.6453, + "step": 2397 + }, + { + "epoch": 0.9182462186482864, + "grad_norm": 0.297821968793869, + "learning_rate": 3.987399393607134e-06, + "loss": 0.6634, + "step": 2398 + }, + { + "epoch": 0.9186291403408003, + "grad_norm": 0.307809978723526, + "learning_rate": 3.986582940760717e-06, + "loss": 0.6391, + "step": 2399 + }, + { + "epoch": 0.9190120620333142, + "grad_norm": 0.2985922396183014, + "learning_rate": 3.985766242560458e-06, + "loss": 0.6593, + "step": 2400 + }, + { + "epoch": 0.9193949837258281, + "grad_norm": 0.2958591878414154, + "learning_rate": 3.984949299141149e-06, + "loss": 0.6563, + "step": 2401 + }, + { + "epoch": 0.9197779054183419, + "grad_norm": 0.29857268929481506, + "learning_rate": 3.984132110637622e-06, + "loss": 0.6599, + "step": 2402 + }, + { + "epoch": 0.9201608271108558, + "grad_norm": 0.297341525554657, + "learning_rate": 3.983314677184752e-06, + "loss": 0.6493, + "step": 2403 + }, + { + "epoch": 0.9205437488033698, + "grad_norm": 0.30079755187034607, + "learning_rate": 3.982496998917451e-06, + "loss": 0.6528, + "step": 2404 + }, + { + "epoch": 0.9209266704958836, + "grad_norm": 0.29261571168899536, + "learning_rate": 3.981679075970675e-06, + "loss": 0.6587, + "step": 2405 + }, + { + "epoch": 0.9213095921883975, + "grad_norm": 0.2840799391269684, + "learning_rate": 3.980860908479417e-06, + "loss": 0.6337, + "step": 2406 + }, + { + "epoch": 0.9216925138809113, + "grad_norm": 0.29934579133987427, + "learning_rate": 3.9800424965787115e-06, + "loss": 0.6602, + "step": 2407 + }, + { + "epoch": 0.9220754355734252, + "grad_norm": 0.2924461364746094, + "learning_rate": 3.979223840403636e-06, + "loss": 0.6252, + "step": 2408 + }, + { + "epoch": 0.9224583572659392, + "grad_norm": 0.2996419668197632, + "learning_rate": 3.978404940089302e-06, + "loss": 0.6351, + "step": 2409 + }, + { + "epoch": 0.922841278958453, + "grad_norm": 0.28820133209228516, + "learning_rate": 3.9775857957708695e-06, + "loss": 0.6527, + "step": 2410 + }, + { + "epoch": 0.9232242006509669, + "grad_norm": 0.2895001471042633, + "learning_rate": 3.976766407583533e-06, + "loss": 0.6568, + "step": 2411 + }, + { + "epoch": 0.9236071223434807, + "grad_norm": 0.2845630943775177, + "learning_rate": 3.975946775662528e-06, + "loss": 0.6405, + "step": 2412 + }, + { + "epoch": 0.9239900440359946, + "grad_norm": 0.2985185384750366, + "learning_rate": 3.975126900143133e-06, + "loss": 0.6695, + "step": 2413 + }, + { + "epoch": 0.9243729657285085, + "grad_norm": 0.29201918840408325, + "learning_rate": 3.974306781160664e-06, + "loss": 0.6376, + "step": 2414 + }, + { + "epoch": 0.9247558874210224, + "grad_norm": 0.2998596429824829, + "learning_rate": 3.973486418850476e-06, + "loss": 0.6619, + "step": 2415 + }, + { + "epoch": 0.9251388091135363, + "grad_norm": 0.2949308156967163, + "learning_rate": 3.9726658133479676e-06, + "loss": 0.6558, + "step": 2416 + }, + { + "epoch": 0.9255217308060502, + "grad_norm": 0.28473496437072754, + "learning_rate": 3.971844964788577e-06, + "loss": 0.6446, + "step": 2417 + }, + { + "epoch": 0.925904652498564, + "grad_norm": 0.29481175541877747, + "learning_rate": 3.971023873307781e-06, + "loss": 0.66, + "step": 2418 + }, + { + "epoch": 0.9262875741910779, + "grad_norm": 0.30271342396736145, + "learning_rate": 3.970202539041096e-06, + "loss": 0.6628, + "step": 2419 + }, + { + "epoch": 0.9266704958835918, + "grad_norm": 0.30491283535957336, + "learning_rate": 3.969380962124081e-06, + "loss": 0.6553, + "step": 2420 + }, + { + "epoch": 0.9270534175761057, + "grad_norm": 0.2750245928764343, + "learning_rate": 3.968559142692333e-06, + "loss": 0.6516, + "step": 2421 + }, + { + "epoch": 0.9274363392686196, + "grad_norm": 0.2851044535636902, + "learning_rate": 3.9677370808814885e-06, + "loss": 0.6303, + "step": 2422 + }, + { + "epoch": 0.9278192609611334, + "grad_norm": 0.29597732424736023, + "learning_rate": 3.966914776827227e-06, + "loss": 0.6486, + "step": 2423 + }, + { + "epoch": 0.9282021826536473, + "grad_norm": 0.29492276906967163, + "learning_rate": 3.966092230665265e-06, + "loss": 0.6677, + "step": 2424 + }, + { + "epoch": 0.9285851043461613, + "grad_norm": 0.2950708866119385, + "learning_rate": 3.9652694425313606e-06, + "loss": 0.6562, + "step": 2425 + }, + { + "epoch": 0.9289680260386751, + "grad_norm": 0.29869431257247925, + "learning_rate": 3.96444641256131e-06, + "loss": 0.6403, + "step": 2426 + }, + { + "epoch": 0.929350947731189, + "grad_norm": 0.29527124762535095, + "learning_rate": 3.963623140890952e-06, + "loss": 0.6691, + "step": 2427 + }, + { + "epoch": 0.9297338694237028, + "grad_norm": 0.2984407842159271, + "learning_rate": 3.962799627656163e-06, + "loss": 0.6479, + "step": 2428 + }, + { + "epoch": 0.9301167911162167, + "grad_norm": 0.3068842887878418, + "learning_rate": 3.9619758729928624e-06, + "loss": 0.6341, + "step": 2429 + }, + { + "epoch": 0.9304997128087306, + "grad_norm": 0.2958669364452362, + "learning_rate": 3.961151877037004e-06, + "loss": 0.6526, + "step": 2430 + }, + { + "epoch": 0.9308826345012445, + "grad_norm": 0.30777838826179504, + "learning_rate": 3.9603276399245864e-06, + "loss": 0.6617, + "step": 2431 + }, + { + "epoch": 0.9312655561937584, + "grad_norm": 0.29739493131637573, + "learning_rate": 3.959503161791645e-06, + "loss": 0.6562, + "step": 2432 + }, + { + "epoch": 0.9316484778862723, + "grad_norm": 0.29062512516975403, + "learning_rate": 3.958678442774259e-06, + "loss": 0.6568, + "step": 2433 + }, + { + "epoch": 0.9320313995787861, + "grad_norm": 0.2967102825641632, + "learning_rate": 3.957853483008541e-06, + "loss": 0.6532, + "step": 2434 + }, + { + "epoch": 0.9324143212713, + "grad_norm": 0.3010478615760803, + "learning_rate": 3.957028282630649e-06, + "loss": 0.6368, + "step": 2435 + }, + { + "epoch": 0.9327972429638139, + "grad_norm": 0.3022819459438324, + "learning_rate": 3.956202841776779e-06, + "loss": 0.6658, + "step": 2436 + }, + { + "epoch": 0.9331801646563278, + "grad_norm": 0.29249346256256104, + "learning_rate": 3.955377160583165e-06, + "loss": 0.647, + "step": 2437 + }, + { + "epoch": 0.9335630863488417, + "grad_norm": 0.29361385107040405, + "learning_rate": 3.954551239186082e-06, + "loss": 0.6237, + "step": 2438 + }, + { + "epoch": 0.9339460080413555, + "grad_norm": 0.2925169765949249, + "learning_rate": 3.953725077721846e-06, + "loss": 0.6477, + "step": 2439 + }, + { + "epoch": 0.9343289297338694, + "grad_norm": 0.2952122688293457, + "learning_rate": 3.95289867632681e-06, + "loss": 0.639, + "step": 2440 + }, + { + "epoch": 0.9347118514263834, + "grad_norm": 0.2906058728694916, + "learning_rate": 3.952072035137368e-06, + "loss": 0.6342, + "step": 2441 + }, + { + "epoch": 0.9350947731188972, + "grad_norm": 0.28912121057510376, + "learning_rate": 3.9512451542899534e-06, + "loss": 0.6512, + "step": 2442 + }, + { + "epoch": 0.9354776948114111, + "grad_norm": 0.28959429264068604, + "learning_rate": 3.95041803392104e-06, + "loss": 0.6387, + "step": 2443 + }, + { + "epoch": 0.9358606165039249, + "grad_norm": 0.2835792899131775, + "learning_rate": 3.94959067416714e-06, + "loss": 0.6477, + "step": 2444 + }, + { + "epoch": 0.9362435381964388, + "grad_norm": 0.3021864593029022, + "learning_rate": 3.948763075164804e-06, + "loss": 0.6412, + "step": 2445 + }, + { + "epoch": 0.9366264598889527, + "grad_norm": 0.3023189902305603, + "learning_rate": 3.947935237050625e-06, + "loss": 0.6636, + "step": 2446 + }, + { + "epoch": 0.9370093815814666, + "grad_norm": 0.299590140581131, + "learning_rate": 3.947107159961234e-06, + "loss": 0.641, + "step": 2447 + }, + { + "epoch": 0.9373923032739805, + "grad_norm": 0.2920341193675995, + "learning_rate": 3.946278844033301e-06, + "loss": 0.6586, + "step": 2448 + }, + { + "epoch": 0.9377752249664943, + "grad_norm": 0.3043041527271271, + "learning_rate": 3.945450289403535e-06, + "loss": 0.6569, + "step": 2449 + }, + { + "epoch": 0.9381581466590082, + "grad_norm": 0.3054685592651367, + "learning_rate": 3.944621496208687e-06, + "loss": 0.6608, + "step": 2450 + }, + { + "epoch": 0.9385410683515221, + "grad_norm": 0.2984093725681305, + "learning_rate": 3.943792464585544e-06, + "loss": 0.643, + "step": 2451 + }, + { + "epoch": 0.938923990044036, + "grad_norm": 0.2824691832065582, + "learning_rate": 3.942963194670934e-06, + "loss": 0.6291, + "step": 2452 + }, + { + "epoch": 0.9393069117365499, + "grad_norm": 0.3064553439617157, + "learning_rate": 3.942133686601727e-06, + "loss": 0.6465, + "step": 2453 + }, + { + "epoch": 0.9396898334290638, + "grad_norm": 0.299063503742218, + "learning_rate": 3.941303940514826e-06, + "loss": 0.6309, + "step": 2454 + }, + { + "epoch": 0.9400727551215776, + "grad_norm": 0.29257938265800476, + "learning_rate": 3.940473956547177e-06, + "loss": 0.627, + "step": 2455 + }, + { + "epoch": 0.9404556768140915, + "grad_norm": 0.2964414656162262, + "learning_rate": 3.939643734835768e-06, + "loss": 0.649, + "step": 2456 + }, + { + "epoch": 0.9408385985066055, + "grad_norm": 0.28775110840797424, + "learning_rate": 3.938813275517622e-06, + "loss": 0.6679, + "step": 2457 + }, + { + "epoch": 0.9412215201991193, + "grad_norm": 0.28418123722076416, + "learning_rate": 3.937982578729802e-06, + "loss": 0.659, + "step": 2458 + }, + { + "epoch": 0.9416044418916332, + "grad_norm": 0.2941417694091797, + "learning_rate": 3.937151644609412e-06, + "loss": 0.6513, + "step": 2459 + }, + { + "epoch": 0.941987363584147, + "grad_norm": 0.29704391956329346, + "learning_rate": 3.936320473293592e-06, + "loss": 0.6446, + "step": 2460 + }, + { + "epoch": 0.9423702852766609, + "grad_norm": 0.2998020350933075, + "learning_rate": 3.9354890649195246e-06, + "loss": 0.658, + "step": 2461 + }, + { + "epoch": 0.9427532069691748, + "grad_norm": 0.28896981477737427, + "learning_rate": 3.934657419624429e-06, + "loss": 0.6569, + "step": 2462 + }, + { + "epoch": 0.9431361286616887, + "grad_norm": 0.2883463203907013, + "learning_rate": 3.933825537545566e-06, + "loss": 0.6553, + "step": 2463 + }, + { + "epoch": 0.9435190503542026, + "grad_norm": 0.29482680559158325, + "learning_rate": 3.932993418820233e-06, + "loss": 0.6549, + "step": 2464 + }, + { + "epoch": 0.9439019720467164, + "grad_norm": 0.2945811450481415, + "learning_rate": 3.9321610635857675e-06, + "loss": 0.6426, + "step": 2465 + }, + { + "epoch": 0.9442848937392303, + "grad_norm": 0.29945310950279236, + "learning_rate": 3.931328471979545e-06, + "loss": 0.6511, + "step": 2466 + }, + { + "epoch": 0.9446678154317442, + "grad_norm": 0.30127647519111633, + "learning_rate": 3.930495644138984e-06, + "loss": 0.6652, + "step": 2467 + }, + { + "epoch": 0.9450507371242581, + "grad_norm": 0.29690733551979065, + "learning_rate": 3.929662580201536e-06, + "loss": 0.6691, + "step": 2468 + }, + { + "epoch": 0.945433658816772, + "grad_norm": 0.2941737473011017, + "learning_rate": 3.9288292803046954e-06, + "loss": 0.6687, + "step": 2469 + }, + { + "epoch": 0.9458165805092859, + "grad_norm": 0.29528361558914185, + "learning_rate": 3.927995744585995e-06, + "loss": 0.6419, + "step": 2470 + }, + { + "epoch": 0.9461995022017997, + "grad_norm": 0.2993655502796173, + "learning_rate": 3.927161973183007e-06, + "loss": 0.6709, + "step": 2471 + }, + { + "epoch": 0.9465824238943136, + "grad_norm": 0.2861931025981903, + "learning_rate": 3.92632796623334e-06, + "loss": 0.6465, + "step": 2472 + }, + { + "epoch": 0.9469653455868275, + "grad_norm": 0.2920815944671631, + "learning_rate": 3.925493723874643e-06, + "loss": 0.6495, + "step": 2473 + }, + { + "epoch": 0.9473482672793414, + "grad_norm": 0.28802213072776794, + "learning_rate": 3.924659246244604e-06, + "loss": 0.6422, + "step": 2474 + }, + { + "epoch": 0.9477311889718553, + "grad_norm": 0.28447744250297546, + "learning_rate": 3.9238245334809525e-06, + "loss": 0.6422, + "step": 2475 + }, + { + "epoch": 0.9481141106643691, + "grad_norm": 0.3059292137622833, + "learning_rate": 3.92298958572145e-06, + "loss": 0.6537, + "step": 2476 + }, + { + "epoch": 0.948497032356883, + "grad_norm": 0.2921289801597595, + "learning_rate": 3.922154403103904e-06, + "loss": 0.6334, + "step": 2477 + }, + { + "epoch": 0.9488799540493968, + "grad_norm": 0.2936064600944519, + "learning_rate": 3.921318985766156e-06, + "loss": 0.6615, + "step": 2478 + }, + { + "epoch": 0.9492628757419108, + "grad_norm": 0.285410076379776, + "learning_rate": 3.920483333846089e-06, + "loss": 0.63, + "step": 2479 + }, + { + "epoch": 0.9496457974344247, + "grad_norm": 0.2900061011314392, + "learning_rate": 3.9196474474816225e-06, + "loss": 0.6462, + "step": 2480 + }, + { + "epoch": 0.9500287191269385, + "grad_norm": 0.2814337909221649, + "learning_rate": 3.918811326810717e-06, + "loss": 0.6745, + "step": 2481 + }, + { + "epoch": 0.9504116408194524, + "grad_norm": 0.29532554745674133, + "learning_rate": 3.917974971971369e-06, + "loss": 0.6406, + "step": 2482 + }, + { + "epoch": 0.9507945625119663, + "grad_norm": 0.2889082729816437, + "learning_rate": 3.917138383101616e-06, + "loss": 0.6555, + "step": 2483 + }, + { + "epoch": 0.9511774842044802, + "grad_norm": 0.29341015219688416, + "learning_rate": 3.916301560339533e-06, + "loss": 0.6534, + "step": 2484 + }, + { + "epoch": 0.9515604058969941, + "grad_norm": 0.2957211136817932, + "learning_rate": 3.915464503823234e-06, + "loss": 0.6656, + "step": 2485 + }, + { + "epoch": 0.951943327589508, + "grad_norm": 0.30043575167655945, + "learning_rate": 3.9146272136908715e-06, + "loss": 0.6657, + "step": 2486 + }, + { + "epoch": 0.9523262492820218, + "grad_norm": 0.2976532578468323, + "learning_rate": 3.913789690080636e-06, + "loss": 0.6578, + "step": 2487 + }, + { + "epoch": 0.9527091709745357, + "grad_norm": 0.27795541286468506, + "learning_rate": 3.912951933130758e-06, + "loss": 0.6348, + "step": 2488 + }, + { + "epoch": 0.9530920926670496, + "grad_norm": 0.29307544231414795, + "learning_rate": 3.912113942979504e-06, + "loss": 0.6661, + "step": 2489 + }, + { + "epoch": 0.9534750143595635, + "grad_norm": 0.2948361933231354, + "learning_rate": 3.911275719765182e-06, + "loss": 0.6552, + "step": 2490 + }, + { + "epoch": 0.9538579360520774, + "grad_norm": 0.29154619574546814, + "learning_rate": 3.910437263626137e-06, + "loss": 0.635, + "step": 2491 + }, + { + "epoch": 0.9542408577445912, + "grad_norm": 0.2864062488079071, + "learning_rate": 3.909598574700751e-06, + "loss": 0.6311, + "step": 2492 + }, + { + "epoch": 0.9546237794371051, + "grad_norm": 0.3037509322166443, + "learning_rate": 3.908759653127446e-06, + "loss": 0.6497, + "step": 2493 + }, + { + "epoch": 0.9550067011296189, + "grad_norm": 0.3685162663459778, + "learning_rate": 3.907920499044684e-06, + "loss": 0.6597, + "step": 2494 + }, + { + "epoch": 0.9553896228221329, + "grad_norm": 0.3015262186527252, + "learning_rate": 3.907081112590962e-06, + "loss": 0.6539, + "step": 2495 + }, + { + "epoch": 0.9557725445146468, + "grad_norm": 0.29515382647514343, + "learning_rate": 3.906241493904818e-06, + "loss": 0.6462, + "step": 2496 + }, + { + "epoch": 0.9561554662071606, + "grad_norm": 0.29523947834968567, + "learning_rate": 3.905401643124826e-06, + "loss": 0.6477, + "step": 2497 + }, + { + "epoch": 0.9565383878996745, + "grad_norm": 0.3039727509021759, + "learning_rate": 3.904561560389601e-06, + "loss": 0.6656, + "step": 2498 + }, + { + "epoch": 0.9569213095921884, + "grad_norm": 0.29031190276145935, + "learning_rate": 3.903721245837794e-06, + "loss": 0.6619, + "step": 2499 + }, + { + "epoch": 0.9573042312847023, + "grad_norm": 0.28732046484947205, + "learning_rate": 3.902880699608095e-06, + "loss": 0.6615, + "step": 2500 + }, + { + "epoch": 0.9576871529772162, + "grad_norm": 0.292327344417572, + "learning_rate": 3.902039921839233e-06, + "loss": 0.6526, + "step": 2501 + }, + { + "epoch": 0.95807007466973, + "grad_norm": 0.30777719616889954, + "learning_rate": 3.9011989126699746e-06, + "loss": 0.6652, + "step": 2502 + }, + { + "epoch": 0.9584529963622439, + "grad_norm": 0.286577969789505, + "learning_rate": 3.900357672239125e-06, + "loss": 0.6277, + "step": 2503 + }, + { + "epoch": 0.9588359180547578, + "grad_norm": 0.30643603205680847, + "learning_rate": 3.899516200685525e-06, + "loss": 0.6704, + "step": 2504 + }, + { + "epoch": 0.9592188397472717, + "grad_norm": 0.3036949932575226, + "learning_rate": 3.898674498148058e-06, + "loss": 0.6674, + "step": 2505 + }, + { + "epoch": 0.9596017614397856, + "grad_norm": 0.29391995072364807, + "learning_rate": 3.8978325647656426e-06, + "loss": 0.6631, + "step": 2506 + }, + { + "epoch": 0.9599846831322995, + "grad_norm": 0.2952329218387604, + "learning_rate": 3.896990400677235e-06, + "loss": 0.6674, + "step": 2507 + }, + { + "epoch": 0.9603676048248133, + "grad_norm": 0.3082561492919922, + "learning_rate": 3.896148006021833e-06, + "loss": 0.6497, + "step": 2508 + }, + { + "epoch": 0.9607505265173272, + "grad_norm": 0.2941219210624695, + "learning_rate": 3.895305380938468e-06, + "loss": 0.6625, + "step": 2509 + }, + { + "epoch": 0.961133448209841, + "grad_norm": 0.2878792881965637, + "learning_rate": 3.894462525566212e-06, + "loss": 0.6405, + "step": 2510 + }, + { + "epoch": 0.961516369902355, + "grad_norm": 0.2885337471961975, + "learning_rate": 3.893619440044174e-06, + "loss": 0.6448, + "step": 2511 + }, + { + "epoch": 0.9618992915948689, + "grad_norm": 0.29196929931640625, + "learning_rate": 3.892776124511503e-06, + "loss": 0.6434, + "step": 2512 + }, + { + "epoch": 0.9622822132873827, + "grad_norm": 0.3003067970275879, + "learning_rate": 3.891932579107384e-06, + "loss": 0.6387, + "step": 2513 + }, + { + "epoch": 0.9626651349798966, + "grad_norm": 0.2982931435108185, + "learning_rate": 3.8910888039710395e-06, + "loss": 0.6637, + "step": 2514 + }, + { + "epoch": 0.9630480566724104, + "grad_norm": 0.3193811774253845, + "learning_rate": 3.8902447992417305e-06, + "loss": 0.6677, + "step": 2515 + }, + { + "epoch": 0.9634309783649244, + "grad_norm": 0.29214462637901306, + "learning_rate": 3.889400565058758e-06, + "loss": 0.6428, + "step": 2516 + }, + { + "epoch": 0.9638139000574383, + "grad_norm": 0.3077375590801239, + "learning_rate": 3.888556101561457e-06, + "loss": 0.6399, + "step": 2517 + }, + { + "epoch": 0.9641968217499521, + "grad_norm": 0.30510959029197693, + "learning_rate": 3.887711408889204e-06, + "loss": 0.643, + "step": 2518 + }, + { + "epoch": 0.964579743442466, + "grad_norm": 0.2839040160179138, + "learning_rate": 3.886866487181411e-06, + "loss": 0.6301, + "step": 2519 + }, + { + "epoch": 0.9649626651349799, + "grad_norm": 0.28605926036834717, + "learning_rate": 3.886021336577528e-06, + "loss": 0.655, + "step": 2520 + }, + { + "epoch": 0.9653455868274938, + "grad_norm": 0.3065052628517151, + "learning_rate": 3.885175957217045e-06, + "loss": 0.6483, + "step": 2521 + }, + { + "epoch": 0.9657285085200077, + "grad_norm": 0.282900333404541, + "learning_rate": 3.884330349239486e-06, + "loss": 0.6417, + "step": 2522 + }, + { + "epoch": 0.9661114302125215, + "grad_norm": 0.3028445541858673, + "learning_rate": 3.883484512784417e-06, + "loss": 0.6474, + "step": 2523 + }, + { + "epoch": 0.9664943519050354, + "grad_norm": 0.2901262044906616, + "learning_rate": 3.882638447991438e-06, + "loss": 0.6533, + "step": 2524 + }, + { + "epoch": 0.9668772735975493, + "grad_norm": 0.3081863820552826, + "learning_rate": 3.881792155000188e-06, + "loss": 0.6544, + "step": 2525 + }, + { + "epoch": 0.9672601952900631, + "grad_norm": 0.2975146174430847, + "learning_rate": 3.880945633950345e-06, + "loss": 0.6488, + "step": 2526 + }, + { + "epoch": 0.9676431169825771, + "grad_norm": 0.3060862720012665, + "learning_rate": 3.880098884981622e-06, + "loss": 0.645, + "step": 2527 + }, + { + "epoch": 0.968026038675091, + "grad_norm": 0.2798968255519867, + "learning_rate": 3.879251908233773e-06, + "loss": 0.6332, + "step": 2528 + }, + { + "epoch": 0.9684089603676048, + "grad_norm": 0.30223652720451355, + "learning_rate": 3.878404703846586e-06, + "loss": 0.6519, + "step": 2529 + }, + { + "epoch": 0.9687918820601187, + "grad_norm": 0.30503448843955994, + "learning_rate": 3.8775572719598896e-06, + "loss": 0.6669, + "step": 2530 + }, + { + "epoch": 0.9691748037526325, + "grad_norm": 0.29106512665748596, + "learning_rate": 3.876709612713548e-06, + "loss": 0.6663, + "step": 2531 + }, + { + "epoch": 0.9695577254451465, + "grad_norm": 0.28600332140922546, + "learning_rate": 3.875861726247464e-06, + "loss": 0.6219, + "step": 2532 + }, + { + "epoch": 0.9699406471376604, + "grad_norm": 0.29538947343826294, + "learning_rate": 3.875013612701577e-06, + "loss": 0.6668, + "step": 2533 + }, + { + "epoch": 0.9703235688301742, + "grad_norm": 0.2874445617198944, + "learning_rate": 3.874165272215863e-06, + "loss": 0.6328, + "step": 2534 + }, + { + "epoch": 0.9707064905226881, + "grad_norm": 0.29347294569015503, + "learning_rate": 3.87331670493034e-06, + "loss": 0.6546, + "step": 2535 + }, + { + "epoch": 0.971089412215202, + "grad_norm": 0.29179924726486206, + "learning_rate": 3.8724679109850574e-06, + "loss": 0.6458, + "step": 2536 + }, + { + "epoch": 0.9714723339077159, + "grad_norm": 0.2879701256752014, + "learning_rate": 3.871618890520105e-06, + "loss": 0.6203, + "step": 2537 + }, + { + "epoch": 0.9718552556002298, + "grad_norm": 0.2921457886695862, + "learning_rate": 3.8707696436756125e-06, + "loss": 0.6372, + "step": 2538 + }, + { + "epoch": 0.9722381772927436, + "grad_norm": 0.294516921043396, + "learning_rate": 3.8699201705917416e-06, + "loss": 0.6376, + "step": 2539 + }, + { + "epoch": 0.9726210989852575, + "grad_norm": 0.28862977027893066, + "learning_rate": 3.869070471408694e-06, + "loss": 0.6449, + "step": 2540 + }, + { + "epoch": 0.9730040206777714, + "grad_norm": 0.308726966381073, + "learning_rate": 3.8682205462667115e-06, + "loss": 0.6462, + "step": 2541 + }, + { + "epoch": 0.9733869423702852, + "grad_norm": 0.28768783807754517, + "learning_rate": 3.8673703953060685e-06, + "loss": 0.6473, + "step": 2542 + }, + { + "epoch": 0.9737698640627992, + "grad_norm": 0.29856985807418823, + "learning_rate": 3.866520018667077e-06, + "loss": 0.6494, + "step": 2543 + }, + { + "epoch": 0.974152785755313, + "grad_norm": 0.295842707157135, + "learning_rate": 3.86566941649009e-06, + "loss": 0.6598, + "step": 2544 + }, + { + "epoch": 0.9745357074478269, + "grad_norm": 0.2955799102783203, + "learning_rate": 3.864818588915495e-06, + "loss": 0.6566, + "step": 2545 + }, + { + "epoch": 0.9749186291403408, + "grad_norm": 0.29329386353492737, + "learning_rate": 3.863967536083718e-06, + "loss": 0.6511, + "step": 2546 + }, + { + "epoch": 0.9753015508328546, + "grad_norm": 0.28482767939567566, + "learning_rate": 3.86311625813522e-06, + "loss": 0.6419, + "step": 2547 + }, + { + "epoch": 0.9756844725253686, + "grad_norm": 0.2978568971157074, + "learning_rate": 3.862264755210501e-06, + "loss": 0.6295, + "step": 2548 + }, + { + "epoch": 0.9760673942178825, + "grad_norm": 0.2871286869049072, + "learning_rate": 3.8614130274501e-06, + "loss": 0.6611, + "step": 2549 + }, + { + "epoch": 0.9764503159103963, + "grad_norm": 0.2937026917934418, + "learning_rate": 3.860561074994587e-06, + "loss": 0.6614, + "step": 2550 + }, + { + "epoch": 0.9768332376029102, + "grad_norm": 0.295292466878891, + "learning_rate": 3.859708897984575e-06, + "loss": 0.6334, + "step": 2551 + }, + { + "epoch": 0.977216159295424, + "grad_norm": 0.2879801094532013, + "learning_rate": 3.858856496560713e-06, + "loss": 0.6299, + "step": 2552 + }, + { + "epoch": 0.977599080987938, + "grad_norm": 0.2882961928844452, + "learning_rate": 3.858003870863684e-06, + "loss": 0.644, + "step": 2553 + }, + { + "epoch": 0.9779820026804519, + "grad_norm": 0.2908332943916321, + "learning_rate": 3.857151021034211e-06, + "loss": 0.63, + "step": 2554 + }, + { + "epoch": 0.9783649243729657, + "grad_norm": 0.29038068652153015, + "learning_rate": 3.8562979472130535e-06, + "loss": 0.6295, + "step": 2555 + }, + { + "epoch": 0.9787478460654796, + "grad_norm": 0.2989266812801361, + "learning_rate": 3.855444649541007e-06, + "loss": 0.6527, + "step": 2556 + }, + { + "epoch": 0.9791307677579935, + "grad_norm": 0.3030693829059601, + "learning_rate": 3.8545911281589046e-06, + "loss": 0.6557, + "step": 2557 + }, + { + "epoch": 0.9795136894505073, + "grad_norm": 0.2984217405319214, + "learning_rate": 3.853737383207615e-06, + "loss": 0.6572, + "step": 2558 + }, + { + "epoch": 0.9798966111430213, + "grad_norm": 0.3122209906578064, + "learning_rate": 3.852883414828046e-06, + "loss": 0.6373, + "step": 2559 + }, + { + "epoch": 0.9802795328355352, + "grad_norm": 0.2848958671092987, + "learning_rate": 3.8520292231611425e-06, + "loss": 0.6531, + "step": 2560 + }, + { + "epoch": 0.980662454528049, + "grad_norm": 0.3094291090965271, + "learning_rate": 3.851174808347883e-06, + "loss": 0.651, + "step": 2561 + }, + { + "epoch": 0.9810453762205629, + "grad_norm": 0.28635185956954956, + "learning_rate": 3.850320170529285e-06, + "loss": 0.63, + "step": 2562 + }, + { + "epoch": 0.9814282979130767, + "grad_norm": 0.28737953305244446, + "learning_rate": 3.849465309846404e-06, + "loss": 0.644, + "step": 2563 + }, + { + "epoch": 0.9818112196055907, + "grad_norm": 0.31178387999534607, + "learning_rate": 3.84861022644033e-06, + "loss": 0.6624, + "step": 2564 + }, + { + "epoch": 0.9821941412981046, + "grad_norm": 0.2955792248249054, + "learning_rate": 3.847754920452191e-06, + "loss": 0.6539, + "step": 2565 + }, + { + "epoch": 0.9825770629906184, + "grad_norm": 0.2956996560096741, + "learning_rate": 3.846899392023151e-06, + "loss": 0.659, + "step": 2566 + }, + { + "epoch": 0.9829599846831323, + "grad_norm": 0.3170894384384155, + "learning_rate": 3.846043641294411e-06, + "loss": 0.6562, + "step": 2567 + }, + { + "epoch": 0.9833429063756461, + "grad_norm": 0.3065207302570343, + "learning_rate": 3.84518766840721e-06, + "loss": 0.658, + "step": 2568 + }, + { + "epoch": 0.9837258280681601, + "grad_norm": 0.3012702465057373, + "learning_rate": 3.844331473502821e-06, + "loss": 0.6451, + "step": 2569 + }, + { + "epoch": 0.984108749760674, + "grad_norm": 0.3071759045124054, + "learning_rate": 3.843475056722555e-06, + "loss": 0.6721, + "step": 2570 + }, + { + "epoch": 0.9844916714531878, + "grad_norm": 0.29533374309539795, + "learning_rate": 3.842618418207762e-06, + "loss": 0.6434, + "step": 2571 + }, + { + "epoch": 0.9848745931457017, + "grad_norm": 0.29023638367652893, + "learning_rate": 3.841761558099824e-06, + "loss": 0.6353, + "step": 2572 + }, + { + "epoch": 0.9852575148382156, + "grad_norm": 0.3099059760570526, + "learning_rate": 3.840904476540164e-06, + "loss": 0.6499, + "step": 2573 + }, + { + "epoch": 0.9856404365307294, + "grad_norm": 0.305209219455719, + "learning_rate": 3.840047173670236e-06, + "loss": 0.6497, + "step": 2574 + }, + { + "epoch": 0.9860233582232434, + "grad_norm": 0.307308167219162, + "learning_rate": 3.839189649631538e-06, + "loss": 0.658, + "step": 2575 + }, + { + "epoch": 0.9864062799157572, + "grad_norm": 0.31121161580085754, + "learning_rate": 3.8383319045655985e-06, + "loss": 0.6529, + "step": 2576 + }, + { + "epoch": 0.9867892016082711, + "grad_norm": 0.2935795187950134, + "learning_rate": 3.837473938613984e-06, + "loss": 0.6491, + "step": 2577 + }, + { + "epoch": 0.987172123300785, + "grad_norm": 0.2984502911567688, + "learning_rate": 3.8366157519183e-06, + "loss": 0.636, + "step": 2578 + }, + { + "epoch": 0.9875550449932988, + "grad_norm": 0.3134182393550873, + "learning_rate": 3.835757344620183e-06, + "loss": 0.6476, + "step": 2579 + }, + { + "epoch": 0.9879379666858128, + "grad_norm": 0.3170115649700165, + "learning_rate": 3.8348987168613126e-06, + "loss": 0.6368, + "step": 2580 + }, + { + "epoch": 0.9883208883783267, + "grad_norm": 0.3063246011734009, + "learning_rate": 3.8340398687834e-06, + "loss": 0.6595, + "step": 2581 + }, + { + "epoch": 0.9887038100708405, + "grad_norm": 0.294207900762558, + "learning_rate": 3.8331808005281935e-06, + "loss": 0.6539, + "step": 2582 + }, + { + "epoch": 0.9890867317633544, + "grad_norm": 0.3054911494255066, + "learning_rate": 3.83232151223748e-06, + "loss": 0.6492, + "step": 2583 + }, + { + "epoch": 0.9894696534558682, + "grad_norm": 0.29723116755485535, + "learning_rate": 3.8314620040530805e-06, + "loss": 0.6324, + "step": 2584 + }, + { + "epoch": 0.9898525751483822, + "grad_norm": 0.2975267767906189, + "learning_rate": 3.830602276116853e-06, + "loss": 0.6455, + "step": 2585 + }, + { + "epoch": 0.9902354968408961, + "grad_norm": 0.31881144642829895, + "learning_rate": 3.829742328570691e-06, + "loss": 0.6548, + "step": 2586 + }, + { + "epoch": 0.9906184185334099, + "grad_norm": 0.30828386545181274, + "learning_rate": 3.828882161556527e-06, + "loss": 0.6653, + "step": 2587 + }, + { + "epoch": 0.9910013402259238, + "grad_norm": 0.30030110478401184, + "learning_rate": 3.828021775216325e-06, + "loss": 0.6333, + "step": 2588 + }, + { + "epoch": 0.9913842619184376, + "grad_norm": 0.298301100730896, + "learning_rate": 3.82716116969209e-06, + "loss": 0.6538, + "step": 2589 + }, + { + "epoch": 0.9917671836109515, + "grad_norm": 0.2987930476665497, + "learning_rate": 3.82630034512586e-06, + "loss": 0.6413, + "step": 2590 + }, + { + "epoch": 0.9921501053034655, + "grad_norm": 0.2933206856250763, + "learning_rate": 3.825439301659711e-06, + "loss": 0.6419, + "step": 2591 + }, + { + "epoch": 0.9925330269959793, + "grad_norm": 0.3178972005844116, + "learning_rate": 3.824578039435754e-06, + "loss": 0.6697, + "step": 2592 + }, + { + "epoch": 0.9929159486884932, + "grad_norm": 0.2987993657588959, + "learning_rate": 3.8237165585961355e-06, + "loss": 0.6468, + "step": 2593 + }, + { + "epoch": 0.9932988703810071, + "grad_norm": 0.31328508257865906, + "learning_rate": 3.82285485928304e-06, + "loss": 0.6411, + "step": 2594 + }, + { + "epoch": 0.9936817920735209, + "grad_norm": 0.306226909160614, + "learning_rate": 3.821992941638688e-06, + "loss": 0.6516, + "step": 2595 + }, + { + "epoch": 0.9940647137660349, + "grad_norm": 0.29100140929222107, + "learning_rate": 3.821130805805334e-06, + "loss": 0.6319, + "step": 2596 + }, + { + "epoch": 0.9944476354585488, + "grad_norm": 0.30655136704444885, + "learning_rate": 3.82026845192527e-06, + "loss": 0.6527, + "step": 2597 + }, + { + "epoch": 0.9948305571510626, + "grad_norm": 0.2994900345802307, + "learning_rate": 3.819405880140823e-06, + "loss": 0.6568, + "step": 2598 + }, + { + "epoch": 0.9952134788435765, + "grad_norm": 0.2904807925224304, + "learning_rate": 3.8185430905943564e-06, + "loss": 0.6687, + "step": 2599 + }, + { + "epoch": 0.9955964005360903, + "grad_norm": 0.2897905111312866, + "learning_rate": 3.8176800834282715e-06, + "loss": 0.6394, + "step": 2600 + }, + { + "epoch": 0.9959793222286043, + "grad_norm": 0.30152979493141174, + "learning_rate": 3.816816858785002e-06, + "loss": 0.6519, + "step": 2601 + }, + { + "epoch": 0.9963622439211182, + "grad_norm": 0.28953397274017334, + "learning_rate": 3.815953416807021e-06, + "loss": 0.6339, + "step": 2602 + }, + { + "epoch": 0.996745165613632, + "grad_norm": 0.2889082729816437, + "learning_rate": 3.815089757636834e-06, + "loss": 0.6486, + "step": 2603 + }, + { + "epoch": 0.9971280873061459, + "grad_norm": 0.29194390773773193, + "learning_rate": 3.814225881416984e-06, + "loss": 0.6397, + "step": 2604 + }, + { + "epoch": 0.9975110089986597, + "grad_norm": 0.28397029638290405, + "learning_rate": 3.813361788290052e-06, + "loss": 0.6517, + "step": 2605 + }, + { + "epoch": 0.9978939306911736, + "grad_norm": 0.2952829897403717, + "learning_rate": 3.812497478398651e-06, + "loss": 0.6401, + "step": 2606 + }, + { + "epoch": 0.9982768523836876, + "grad_norm": 0.2944144904613495, + "learning_rate": 3.8116329518854313e-06, + "loss": 0.6656, + "step": 2607 + }, + { + "epoch": 0.9986597740762014, + "grad_norm": 0.29345735907554626, + "learning_rate": 3.8107682088930797e-06, + "loss": 0.6343, + "step": 2608 + }, + { + "epoch": 0.9990426957687153, + "grad_norm": 0.2927607297897339, + "learning_rate": 3.809903249564318e-06, + "loss": 0.6371, + "step": 2609 + }, + { + "epoch": 0.9994256174612292, + "grad_norm": 0.288296639919281, + "learning_rate": 3.809038074041904e-06, + "loss": 0.6506, + "step": 2610 + }, + { + "epoch": 0.999808539153743, + "grad_norm": 0.29800692200660706, + "learning_rate": 3.8081726824686303e-06, + "loss": 0.6114, + "step": 2611 + }, + { + "epoch": 1.000382921692514, + "grad_norm": 0.6201800107955933, + "learning_rate": 3.8073070749873283e-06, + "loss": 1.2979, + "step": 2612 + }, + { + "epoch": 1.0007658433850277, + "grad_norm": 0.2835237979888916, + "learning_rate": 3.80644125174086e-06, + "loss": 0.6401, + "step": 2613 + }, + { + "epoch": 1.0011487650775417, + "grad_norm": 0.29264310002326965, + "learning_rate": 3.8055752128721258e-06, + "loss": 0.6178, + "step": 2614 + }, + { + "epoch": 1.0015316867700554, + "grad_norm": 0.2921473979949951, + "learning_rate": 3.8047089585240625e-06, + "loss": 0.6333, + "step": 2615 + }, + { + "epoch": 1.0019146084625694, + "grad_norm": 0.29367759823799133, + "learning_rate": 3.803842488839642e-06, + "loss": 0.6562, + "step": 2616 + }, + { + "epoch": 1.0022975301550834, + "grad_norm": 0.29014942049980164, + "learning_rate": 3.80297580396187e-06, + "loss": 0.6335, + "step": 2617 + }, + { + "epoch": 1.0026804518475971, + "grad_norm": 0.30432870984077454, + "learning_rate": 3.8021089040337887e-06, + "loss": 0.6343, + "step": 2618 + }, + { + "epoch": 1.003063373540111, + "grad_norm": 0.30501067638397217, + "learning_rate": 3.8012417891984776e-06, + "loss": 0.6613, + "step": 2619 + }, + { + "epoch": 1.0034462952326249, + "grad_norm": 0.298501193523407, + "learning_rate": 3.80037445959905e-06, + "loss": 0.6555, + "step": 2620 + }, + { + "epoch": 1.0038292169251388, + "grad_norm": 0.30998656153678894, + "learning_rate": 3.7995069153786525e-06, + "loss": 0.6335, + "step": 2621 + }, + { + "epoch": 1.0042121386176528, + "grad_norm": 0.2906511127948761, + "learning_rate": 3.7986391566804716e-06, + "loss": 0.6372, + "step": 2622 + }, + { + "epoch": 1.0045950603101665, + "grad_norm": 0.3090764582157135, + "learning_rate": 3.7977711836477266e-06, + "loss": 0.6439, + "step": 2623 + }, + { + "epoch": 1.0049779820026805, + "grad_norm": 0.2993593215942383, + "learning_rate": 3.796902996423672e-06, + "loss": 0.6517, + "step": 2624 + }, + { + "epoch": 1.0053609036951943, + "grad_norm": 0.3124483525753021, + "learning_rate": 3.796034595151598e-06, + "loss": 0.6391, + "step": 2625 + }, + { + "epoch": 1.0057438253877082, + "grad_norm": 0.31987473368644714, + "learning_rate": 3.795165979974831e-06, + "loss": 0.6437, + "step": 2626 + }, + { + "epoch": 1.0061267470802222, + "grad_norm": 0.2922317385673523, + "learning_rate": 3.794297151036732e-06, + "loss": 0.6539, + "step": 2627 + }, + { + "epoch": 1.006509668772736, + "grad_norm": 0.2869653105735779, + "learning_rate": 3.7934281084806974e-06, + "loss": 0.6262, + "step": 2628 + }, + { + "epoch": 1.00689259046525, + "grad_norm": 0.28647735714912415, + "learning_rate": 3.792558852450158e-06, + "loss": 0.6464, + "step": 2629 + }, + { + "epoch": 1.0072755121577637, + "grad_norm": 0.29125460982322693, + "learning_rate": 3.7916893830885816e-06, + "loss": 0.6348, + "step": 2630 + }, + { + "epoch": 1.0076584338502776, + "grad_norm": 0.29443812370300293, + "learning_rate": 3.7908197005394686e-06, + "loss": 0.6249, + "step": 2631 + }, + { + "epoch": 1.0080413555427914, + "grad_norm": 0.28413528203964233, + "learning_rate": 3.789949804946359e-06, + "loss": 0.6229, + "step": 2632 + }, + { + "epoch": 1.0084242772353054, + "grad_norm": 0.3106536269187927, + "learning_rate": 3.789079696452822e-06, + "loss": 0.6451, + "step": 2633 + }, + { + "epoch": 1.0088071989278193, + "grad_norm": 0.3159669041633606, + "learning_rate": 3.788209375202468e-06, + "loss": 0.6454, + "step": 2634 + }, + { + "epoch": 1.009190120620333, + "grad_norm": 0.2946128249168396, + "learning_rate": 3.7873388413389367e-06, + "loss": 0.6473, + "step": 2635 + }, + { + "epoch": 1.009573042312847, + "grad_norm": 0.3580203950405121, + "learning_rate": 3.7864680950059086e-06, + "loss": 0.6416, + "step": 2636 + }, + { + "epoch": 1.0099559640053608, + "grad_norm": 0.30298998951911926, + "learning_rate": 3.7855971363470946e-06, + "loss": 0.6325, + "step": 2637 + }, + { + "epoch": 1.0103388856978748, + "grad_norm": 0.29925474524497986, + "learning_rate": 3.784725965506243e-06, + "loss": 0.642, + "step": 2638 + }, + { + "epoch": 1.0107218073903887, + "grad_norm": 0.3114474415779114, + "learning_rate": 3.7838545826271374e-06, + "loss": 0.6441, + "step": 2639 + }, + { + "epoch": 1.0111047290829025, + "grad_norm": 0.35924795269966125, + "learning_rate": 3.782982987853594e-06, + "loss": 0.632, + "step": 2640 + }, + { + "epoch": 1.0114876507754165, + "grad_norm": 0.2931000292301178, + "learning_rate": 3.7821111813294674e-06, + "loss": 0.6502, + "step": 2641 + }, + { + "epoch": 1.0118705724679302, + "grad_norm": 0.33271512389183044, + "learning_rate": 3.7812391631986447e-06, + "loss": 0.6386, + "step": 2642 + }, + { + "epoch": 1.0122534941604442, + "grad_norm": 0.30961841344833374, + "learning_rate": 3.7803669336050476e-06, + "loss": 0.6486, + "step": 2643 + }, + { + "epoch": 1.0126364158529582, + "grad_norm": 0.30921536684036255, + "learning_rate": 3.779494492692635e-06, + "loss": 0.6484, + "step": 2644 + }, + { + "epoch": 1.013019337545472, + "grad_norm": 0.2950277626514435, + "learning_rate": 3.7786218406053987e-06, + "loss": 0.6395, + "step": 2645 + }, + { + "epoch": 1.0134022592379859, + "grad_norm": 0.319284051656723, + "learning_rate": 3.777748977487366e-06, + "loss": 0.6518, + "step": 2646 + }, + { + "epoch": 1.0137851809304996, + "grad_norm": 0.31476452946662903, + "learning_rate": 3.7768759034825993e-06, + "loss": 0.6583, + "step": 2647 + }, + { + "epoch": 1.0141681026230136, + "grad_norm": 0.29135146737098694, + "learning_rate": 3.7760026187351957e-06, + "loss": 0.6321, + "step": 2648 + }, + { + "epoch": 1.0145510243155276, + "grad_norm": 0.31046757102012634, + "learning_rate": 3.775129123389286e-06, + "loss": 0.6548, + "step": 2649 + }, + { + "epoch": 1.0149339460080413, + "grad_norm": 0.30898699164390564, + "learning_rate": 3.7742554175890377e-06, + "loss": 0.6416, + "step": 2650 + }, + { + "epoch": 1.0153168677005553, + "grad_norm": 0.3126746118068695, + "learning_rate": 3.7733815014786523e-06, + "loss": 0.6643, + "step": 2651 + }, + { + "epoch": 1.015699789393069, + "grad_norm": 0.28995442390441895, + "learning_rate": 3.772507375202365e-06, + "loss": 0.6212, + "step": 2652 + }, + { + "epoch": 1.016082711085583, + "grad_norm": 0.2873072922229767, + "learning_rate": 3.7716330389044463e-06, + "loss": 0.6446, + "step": 2653 + }, + { + "epoch": 1.016465632778097, + "grad_norm": 0.29912304878234863, + "learning_rate": 3.770758492729202e-06, + "loss": 0.6404, + "step": 2654 + }, + { + "epoch": 1.0168485544706107, + "grad_norm": 0.2976933419704437, + "learning_rate": 3.7698837368209706e-06, + "loss": 0.6328, + "step": 2655 + }, + { + "epoch": 1.0172314761631247, + "grad_norm": 0.29689842462539673, + "learning_rate": 3.769008771324129e-06, + "loss": 0.6408, + "step": 2656 + }, + { + "epoch": 1.0176143978556385, + "grad_norm": 0.30077067017555237, + "learning_rate": 3.7681335963830845e-06, + "loss": 0.6496, + "step": 2657 + }, + { + "epoch": 1.0179973195481524, + "grad_norm": 0.3138182461261749, + "learning_rate": 3.767258212142282e-06, + "loss": 0.6536, + "step": 2658 + }, + { + "epoch": 1.0183802412406662, + "grad_norm": 0.3145606219768524, + "learning_rate": 3.7663826187461982e-06, + "loss": 0.6359, + "step": 2659 + }, + { + "epoch": 1.0187631629331801, + "grad_norm": 0.30564919114112854, + "learning_rate": 3.765506816339347e-06, + "loss": 0.6481, + "step": 2660 + }, + { + "epoch": 1.0191460846256941, + "grad_norm": 0.28875479102134705, + "learning_rate": 3.7646308050662763e-06, + "loss": 0.6182, + "step": 2661 + }, + { + "epoch": 1.0195290063182079, + "grad_norm": 0.30726489424705505, + "learning_rate": 3.763754585071567e-06, + "loss": 0.6385, + "step": 2662 + }, + { + "epoch": 1.0199119280107218, + "grad_norm": 0.30254071950912476, + "learning_rate": 3.762878156499835e-06, + "loss": 0.6521, + "step": 2663 + }, + { + "epoch": 1.0202948497032356, + "grad_norm": 0.2987937033176422, + "learning_rate": 3.7620015194957305e-06, + "loss": 0.6324, + "step": 2664 + }, + { + "epoch": 1.0206777713957496, + "grad_norm": 0.2970277965068817, + "learning_rate": 3.76112467420394e-06, + "loss": 0.6361, + "step": 2665 + }, + { + "epoch": 1.0210606930882635, + "grad_norm": 0.29615169763565063, + "learning_rate": 3.7602476207691817e-06, + "loss": 0.6106, + "step": 2666 + }, + { + "epoch": 1.0214436147807773, + "grad_norm": 0.29066285490989685, + "learning_rate": 3.759370359336209e-06, + "loss": 0.624, + "step": 2667 + }, + { + "epoch": 1.0218265364732912, + "grad_norm": 0.2924114763736725, + "learning_rate": 3.758492890049812e-06, + "loss": 0.6384, + "step": 2668 + }, + { + "epoch": 1.022209458165805, + "grad_norm": 0.2987139821052551, + "learning_rate": 3.7576152130548105e-06, + "loss": 0.6534, + "step": 2669 + }, + { + "epoch": 1.022592379858319, + "grad_norm": 0.29018595814704895, + "learning_rate": 3.756737328496063e-06, + "loss": 0.6434, + "step": 2670 + }, + { + "epoch": 1.022975301550833, + "grad_norm": 0.3101225197315216, + "learning_rate": 3.75585923651846e-06, + "loss": 0.6419, + "step": 2671 + }, + { + "epoch": 1.0233582232433467, + "grad_norm": 0.3001331090927124, + "learning_rate": 3.7549809372669253e-06, + "loss": 0.627, + "step": 2672 + }, + { + "epoch": 1.0237411449358607, + "grad_norm": 0.3136843740940094, + "learning_rate": 3.7541024308864203e-06, + "loss": 0.638, + "step": 2673 + }, + { + "epoch": 1.0241240666283744, + "grad_norm": 0.30513647198677063, + "learning_rate": 3.7532237175219378e-06, + "loss": 0.658, + "step": 2674 + }, + { + "epoch": 1.0245069883208884, + "grad_norm": 0.29506388306617737, + "learning_rate": 3.7523447973185045e-06, + "loss": 0.6452, + "step": 2675 + }, + { + "epoch": 1.0248899100134023, + "grad_norm": 0.2931237518787384, + "learning_rate": 3.751465670421184e-06, + "loss": 0.6311, + "step": 2676 + }, + { + "epoch": 1.025272831705916, + "grad_norm": 0.30337023735046387, + "learning_rate": 3.7505863369750706e-06, + "loss": 0.6565, + "step": 2677 + }, + { + "epoch": 1.02565575339843, + "grad_norm": 0.2908066511154175, + "learning_rate": 3.7497067971252955e-06, + "loss": 0.6524, + "step": 2678 + }, + { + "epoch": 1.0260386750909438, + "grad_norm": 0.3014826476573944, + "learning_rate": 3.748827051017021e-06, + "loss": 0.6317, + "step": 2679 + }, + { + "epoch": 1.0264215967834578, + "grad_norm": 0.2967444360256195, + "learning_rate": 3.7479470987954487e-06, + "loss": 0.6347, + "step": 2680 + }, + { + "epoch": 1.0268045184759718, + "grad_norm": 0.3050481975078583, + "learning_rate": 3.747066940605807e-06, + "loss": 0.65, + "step": 2681 + }, + { + "epoch": 1.0271874401684855, + "grad_norm": 0.3011174201965332, + "learning_rate": 3.7461865765933638e-06, + "loss": 0.6352, + "step": 2682 + }, + { + "epoch": 1.0275703618609995, + "grad_norm": 0.298149973154068, + "learning_rate": 3.74530600690342e-06, + "loss": 0.6428, + "step": 2683 + }, + { + "epoch": 1.0279532835535132, + "grad_norm": 0.30855464935302734, + "learning_rate": 3.744425231681308e-06, + "loss": 0.6456, + "step": 2684 + }, + { + "epoch": 1.0283362052460272, + "grad_norm": 0.2995569109916687, + "learning_rate": 3.7435442510723967e-06, + "loss": 0.6387, + "step": 2685 + }, + { + "epoch": 1.0287191269385412, + "grad_norm": 0.3239957094192505, + "learning_rate": 3.742663065222088e-06, + "loss": 0.6421, + "step": 2686 + }, + { + "epoch": 1.029102048631055, + "grad_norm": 0.2976495623588562, + "learning_rate": 3.7417816742758178e-06, + "loss": 0.6379, + "step": 2687 + }, + { + "epoch": 1.029484970323569, + "grad_norm": 0.30536195635795593, + "learning_rate": 3.7409000783790554e-06, + "loss": 0.6598, + "step": 2688 + }, + { + "epoch": 1.0298678920160826, + "grad_norm": 0.29449784755706787, + "learning_rate": 3.7400182776773044e-06, + "loss": 0.6481, + "step": 2689 + }, + { + "epoch": 1.0302508137085966, + "grad_norm": 0.2924303710460663, + "learning_rate": 3.739136272316102e-06, + "loss": 0.6541, + "step": 2690 + }, + { + "epoch": 1.0306337354011106, + "grad_norm": 0.3128041923046112, + "learning_rate": 3.7382540624410186e-06, + "loss": 0.6591, + "step": 2691 + }, + { + "epoch": 1.0310166570936243, + "grad_norm": 0.3099387586116791, + "learning_rate": 3.7373716481976603e-06, + "loss": 0.6517, + "step": 2692 + }, + { + "epoch": 1.0313995787861383, + "grad_norm": 0.28915831446647644, + "learning_rate": 3.736489029731664e-06, + "loss": 0.6295, + "step": 2693 + }, + { + "epoch": 1.031782500478652, + "grad_norm": 0.317484050989151, + "learning_rate": 3.7356062071887033e-06, + "loss": 0.6275, + "step": 2694 + }, + { + "epoch": 1.032165422171166, + "grad_norm": 0.31788381934165955, + "learning_rate": 3.734723180714484e-06, + "loss": 0.6413, + "step": 2695 + }, + { + "epoch": 1.0325483438636798, + "grad_norm": 0.3102053105831146, + "learning_rate": 3.7338399504547445e-06, + "loss": 0.6445, + "step": 2696 + }, + { + "epoch": 1.0329312655561937, + "grad_norm": 0.2990587055683136, + "learning_rate": 3.7329565165552596e-06, + "loss": 0.6463, + "step": 2697 + }, + { + "epoch": 1.0333141872487077, + "grad_norm": 0.30451837182044983, + "learning_rate": 3.7320728791618348e-06, + "loss": 0.6315, + "step": 2698 + }, + { + "epoch": 1.0336971089412215, + "grad_norm": 0.30733415484428406, + "learning_rate": 3.731189038420311e-06, + "loss": 0.6221, + "step": 2699 + }, + { + "epoch": 1.0340800306337354, + "grad_norm": 0.298105388879776, + "learning_rate": 3.7303049944765617e-06, + "loss": 0.6475, + "step": 2700 + }, + { + "epoch": 1.0344629523262492, + "grad_norm": 0.2926783859729767, + "learning_rate": 3.7294207474764952e-06, + "loss": 0.6449, + "step": 2701 + }, + { + "epoch": 1.0348458740187632, + "grad_norm": 0.29328179359436035, + "learning_rate": 3.728536297566053e-06, + "loss": 0.6591, + "step": 2702 + }, + { + "epoch": 1.0352287957112771, + "grad_norm": 0.315210223197937, + "learning_rate": 3.727651644891207e-06, + "loss": 0.6484, + "step": 2703 + }, + { + "epoch": 1.0356117174037909, + "grad_norm": 0.3034343123435974, + "learning_rate": 3.7267667895979677e-06, + "loss": 0.6414, + "step": 2704 + }, + { + "epoch": 1.0359946390963048, + "grad_norm": 0.3080562353134155, + "learning_rate": 3.725881731832375e-06, + "loss": 0.641, + "step": 2705 + }, + { + "epoch": 1.0363775607888186, + "grad_norm": 0.29416999220848083, + "learning_rate": 3.7249964717405045e-06, + "loss": 0.6412, + "step": 2706 + }, + { + "epoch": 1.0367604824813326, + "grad_norm": 0.2919902503490448, + "learning_rate": 3.7241110094684648e-06, + "loss": 0.6419, + "step": 2707 + }, + { + "epoch": 1.0371434041738465, + "grad_norm": 0.2956090569496155, + "learning_rate": 3.7232253451623967e-06, + "loss": 0.6436, + "step": 2708 + }, + { + "epoch": 1.0375263258663603, + "grad_norm": 0.2998788058757782, + "learning_rate": 3.7223394789684752e-06, + "loss": 0.6141, + "step": 2709 + }, + { + "epoch": 1.0379092475588743, + "grad_norm": 0.2903580367565155, + "learning_rate": 3.721453411032908e-06, + "loss": 0.6355, + "step": 2710 + }, + { + "epoch": 1.038292169251388, + "grad_norm": 0.293494313955307, + "learning_rate": 3.720567141501938e-06, + "loss": 0.6416, + "step": 2711 + }, + { + "epoch": 1.038675090943902, + "grad_norm": 0.28456395864486694, + "learning_rate": 3.7196806705218398e-06, + "loss": 0.6311, + "step": 2712 + }, + { + "epoch": 1.039058012636416, + "grad_norm": 0.3049861788749695, + "learning_rate": 3.71879399823892e-06, + "loss": 0.6453, + "step": 2713 + }, + { + "epoch": 1.0394409343289297, + "grad_norm": 0.29754766821861267, + "learning_rate": 3.717907124799521e-06, + "loss": 0.6418, + "step": 2714 + }, + { + "epoch": 1.0398238560214437, + "grad_norm": 0.2878425717353821, + "learning_rate": 3.7170200503500176e-06, + "loss": 0.6467, + "step": 2715 + }, + { + "epoch": 1.0402067777139574, + "grad_norm": 0.30091592669487, + "learning_rate": 3.7161327750368166e-06, + "loss": 0.6401, + "step": 2716 + }, + { + "epoch": 1.0405896994064714, + "grad_norm": 0.3048005998134613, + "learning_rate": 3.7152452990063592e-06, + "loss": 0.6536, + "step": 2717 + }, + { + "epoch": 1.0409726210989854, + "grad_norm": 0.30043476819992065, + "learning_rate": 3.714357622405119e-06, + "loss": 0.6327, + "step": 2718 + }, + { + "epoch": 1.0413555427914991, + "grad_norm": 0.3011834919452667, + "learning_rate": 3.713469745379604e-06, + "loss": 0.646, + "step": 2719 + }, + { + "epoch": 1.041738464484013, + "grad_norm": 0.29110977053642273, + "learning_rate": 3.712581668076354e-06, + "loss": 0.6365, + "step": 2720 + }, + { + "epoch": 1.0421213861765268, + "grad_norm": 0.29209911823272705, + "learning_rate": 3.711693390641941e-06, + "loss": 0.644, + "step": 2721 + }, + { + "epoch": 1.0425043078690408, + "grad_norm": 0.30214056372642517, + "learning_rate": 3.710804913222972e-06, + "loss": 0.6281, + "step": 2722 + }, + { + "epoch": 1.0428872295615546, + "grad_norm": 0.2980204224586487, + "learning_rate": 3.7099162359660868e-06, + "loss": 0.6147, + "step": 2723 + }, + { + "epoch": 1.0432701512540685, + "grad_norm": 0.30293184518814087, + "learning_rate": 3.709027359017957e-06, + "loss": 0.632, + "step": 2724 + }, + { + "epoch": 1.0436530729465825, + "grad_norm": 0.3033329248428345, + "learning_rate": 3.708138282525287e-06, + "loss": 0.6492, + "step": 2725 + }, + { + "epoch": 1.0440359946390962, + "grad_norm": 0.296652615070343, + "learning_rate": 3.7072490066348164e-06, + "loss": 0.6473, + "step": 2726 + }, + { + "epoch": 1.0444189163316102, + "grad_norm": 0.2859066426753998, + "learning_rate": 3.706359531493316e-06, + "loss": 0.6164, + "step": 2727 + }, + { + "epoch": 1.044801838024124, + "grad_norm": 0.28973037004470825, + "learning_rate": 3.7054698572475883e-06, + "loss": 0.6785, + "step": 2728 + }, + { + "epoch": 1.045184759716638, + "grad_norm": 0.29713091254234314, + "learning_rate": 3.7045799840444712e-06, + "loss": 0.6117, + "step": 2729 + }, + { + "epoch": 1.045567681409152, + "grad_norm": 0.28935956954956055, + "learning_rate": 3.7036899120308346e-06, + "loss": 0.6456, + "step": 2730 + }, + { + "epoch": 1.0459506031016657, + "grad_norm": 0.30467647314071655, + "learning_rate": 3.7027996413535793e-06, + "loss": 0.6586, + "step": 2731 + }, + { + "epoch": 1.0463335247941796, + "grad_norm": 0.2993554174900055, + "learning_rate": 3.701909172159642e-06, + "loss": 0.6215, + "step": 2732 + }, + { + "epoch": 1.0467164464866934, + "grad_norm": 0.2934916913509369, + "learning_rate": 3.7010185045959898e-06, + "loss": 0.6303, + "step": 2733 + }, + { + "epoch": 1.0470993681792073, + "grad_norm": 0.2955273985862732, + "learning_rate": 3.700127638809624e-06, + "loss": 0.6541, + "step": 2734 + }, + { + "epoch": 1.0474822898717213, + "grad_norm": 0.28426697850227356, + "learning_rate": 3.6992365749475766e-06, + "loss": 0.6524, + "step": 2735 + }, + { + "epoch": 1.047865211564235, + "grad_norm": 0.2950335443019867, + "learning_rate": 3.6983453131569164e-06, + "loss": 0.638, + "step": 2736 + }, + { + "epoch": 1.048248133256749, + "grad_norm": 0.30433130264282227, + "learning_rate": 3.69745385358474e-06, + "loss": 0.6469, + "step": 2737 + }, + { + "epoch": 1.0486310549492628, + "grad_norm": 0.2949274778366089, + "learning_rate": 3.696562196378178e-06, + "loss": 0.636, + "step": 2738 + }, + { + "epoch": 1.0490139766417768, + "grad_norm": 0.2984357178211212, + "learning_rate": 3.6956703416843975e-06, + "loss": 0.652, + "step": 2739 + }, + { + "epoch": 1.0493968983342907, + "grad_norm": 0.288908451795578, + "learning_rate": 3.694778289650593e-06, + "loss": 0.6479, + "step": 2740 + }, + { + "epoch": 1.0497798200268045, + "grad_norm": 0.29162946343421936, + "learning_rate": 3.693886040423994e-06, + "loss": 0.6131, + "step": 2741 + }, + { + "epoch": 1.0501627417193184, + "grad_norm": 0.2965639531612396, + "learning_rate": 3.692993594151862e-06, + "loss": 0.6511, + "step": 2742 + }, + { + "epoch": 1.0505456634118322, + "grad_norm": 0.3004387617111206, + "learning_rate": 3.692100950981492e-06, + "loss": 0.6568, + "step": 2743 + }, + { + "epoch": 1.0509285851043462, + "grad_norm": 0.29793962836265564, + "learning_rate": 3.69120811106021e-06, + "loss": 0.6502, + "step": 2744 + }, + { + "epoch": 1.0513115067968601, + "grad_norm": 0.3005545139312744, + "learning_rate": 3.690315074535375e-06, + "loss": 0.6514, + "step": 2745 + }, + { + "epoch": 1.051694428489374, + "grad_norm": 0.31490692496299744, + "learning_rate": 3.68942184155438e-06, + "loss": 0.6478, + "step": 2746 + }, + { + "epoch": 1.0520773501818879, + "grad_norm": 0.3004230260848999, + "learning_rate": 3.688528412264648e-06, + "loss": 0.6192, + "step": 2747 + }, + { + "epoch": 1.0524602718744016, + "grad_norm": 0.3077643811702728, + "learning_rate": 3.687634786813635e-06, + "loss": 0.6554, + "step": 2748 + }, + { + "epoch": 1.0528431935669156, + "grad_norm": 0.31804051995277405, + "learning_rate": 3.686740965348832e-06, + "loss": 0.6408, + "step": 2749 + }, + { + "epoch": 1.0532261152594296, + "grad_norm": 0.2889087498188019, + "learning_rate": 3.685846948017757e-06, + "loss": 0.6445, + "step": 2750 + }, + { + "epoch": 1.0536090369519433, + "grad_norm": 0.3044969141483307, + "learning_rate": 3.6849527349679674e-06, + "loss": 0.6209, + "step": 2751 + }, + { + "epoch": 1.0539919586444573, + "grad_norm": 0.3219917416572571, + "learning_rate": 3.6840583263470466e-06, + "loss": 0.6357, + "step": 2752 + }, + { + "epoch": 1.054374880336971, + "grad_norm": 0.3113216459751129, + "learning_rate": 3.6831637223026123e-06, + "loss": 0.6248, + "step": 2753 + }, + { + "epoch": 1.054757802029485, + "grad_norm": 0.2890131175518036, + "learning_rate": 3.682268922982317e-06, + "loss": 0.6324, + "step": 2754 + }, + { + "epoch": 1.055140723721999, + "grad_norm": 0.30434489250183105, + "learning_rate": 3.6813739285338407e-06, + "loss": 0.6364, + "step": 2755 + }, + { + "epoch": 1.0555236454145127, + "grad_norm": 0.3029651641845703, + "learning_rate": 3.6804787391049e-06, + "loss": 0.6283, + "step": 2756 + }, + { + "epoch": 1.0559065671070267, + "grad_norm": 0.2993582785129547, + "learning_rate": 3.679583354843242e-06, + "loss": 0.662, + "step": 2757 + }, + { + "epoch": 1.0562894887995404, + "grad_norm": 0.295393168926239, + "learning_rate": 3.6786877758966456e-06, + "loss": 0.6308, + "step": 2758 + }, + { + "epoch": 1.0566724104920544, + "grad_norm": 0.2970457673072815, + "learning_rate": 3.677792002412921e-06, + "loss": 0.6396, + "step": 2759 + }, + { + "epoch": 1.0570553321845682, + "grad_norm": 0.3003216087818146, + "learning_rate": 3.676896034539913e-06, + "loss": 0.6383, + "step": 2760 + }, + { + "epoch": 1.0574382538770821, + "grad_norm": 0.2983080744743347, + "learning_rate": 3.6759998724254964e-06, + "loss": 0.6262, + "step": 2761 + }, + { + "epoch": 1.057821175569596, + "grad_norm": 0.29837164282798767, + "learning_rate": 3.6751035162175784e-06, + "loss": 0.6448, + "step": 2762 + }, + { + "epoch": 1.0582040972621098, + "grad_norm": 0.29891669750213623, + "learning_rate": 3.6742069660640996e-06, + "loss": 0.6106, + "step": 2763 + }, + { + "epoch": 1.0585870189546238, + "grad_norm": 0.3070757985115051, + "learning_rate": 3.6733102221130303e-06, + "loss": 0.6615, + "step": 2764 + }, + { + "epoch": 1.0589699406471376, + "grad_norm": 0.32574987411499023, + "learning_rate": 3.6724132845123754e-06, + "loss": 0.6299, + "step": 2765 + }, + { + "epoch": 1.0593528623396515, + "grad_norm": 0.30031269788742065, + "learning_rate": 3.6715161534101697e-06, + "loss": 0.6333, + "step": 2766 + }, + { + "epoch": 1.0597357840321655, + "grad_norm": 0.3040315806865692, + "learning_rate": 3.67061882895448e-06, + "loss": 0.635, + "step": 2767 + }, + { + "epoch": 1.0601187057246793, + "grad_norm": 0.29626211524009705, + "learning_rate": 3.669721311293408e-06, + "loss": 0.647, + "step": 2768 + }, + { + "epoch": 1.0605016274171932, + "grad_norm": 0.3075331151485443, + "learning_rate": 3.668823600575082e-06, + "loss": 0.6283, + "step": 2769 + }, + { + "epoch": 1.060884549109707, + "grad_norm": 0.3068351745605469, + "learning_rate": 3.6679256969476673e-06, + "loss": 0.6427, + "step": 2770 + }, + { + "epoch": 1.061267470802221, + "grad_norm": 0.29604947566986084, + "learning_rate": 3.667027600559357e-06, + "loss": 0.6133, + "step": 2771 + }, + { + "epoch": 1.061650392494735, + "grad_norm": 0.30618706345558167, + "learning_rate": 3.6661293115583794e-06, + "loss": 0.6378, + "step": 2772 + }, + { + "epoch": 1.0620333141872487, + "grad_norm": 0.31926074624061584, + "learning_rate": 3.6652308300929935e-06, + "loss": 0.6399, + "step": 2773 + }, + { + "epoch": 1.0624162358797626, + "grad_norm": 0.30740901827812195, + "learning_rate": 3.6643321563114874e-06, + "loss": 0.6193, + "step": 2774 + }, + { + "epoch": 1.0627991575722764, + "grad_norm": 0.3151159882545471, + "learning_rate": 3.6634332903621856e-06, + "loss": 0.6301, + "step": 2775 + }, + { + "epoch": 1.0631820792647904, + "grad_norm": 0.2917773127555847, + "learning_rate": 3.6625342323934406e-06, + "loss": 0.6516, + "step": 2776 + }, + { + "epoch": 1.0635650009573043, + "grad_norm": 0.3029305040836334, + "learning_rate": 3.6616349825536375e-06, + "loss": 0.6375, + "step": 2777 + }, + { + "epoch": 1.063947922649818, + "grad_norm": 0.29742011427879333, + "learning_rate": 3.6607355409911943e-06, + "loss": 0.6693, + "step": 2778 + }, + { + "epoch": 1.064330844342332, + "grad_norm": 0.3148004114627838, + "learning_rate": 3.6598359078545597e-06, + "loss": 0.6293, + "step": 2779 + }, + { + "epoch": 1.0647137660348458, + "grad_norm": 0.29934290051460266, + "learning_rate": 3.6589360832922145e-06, + "loss": 0.6306, + "step": 2780 + }, + { + "epoch": 1.0650966877273598, + "grad_norm": 0.29205870628356934, + "learning_rate": 3.6580360674526693e-06, + "loss": 0.6418, + "step": 2781 + }, + { + "epoch": 1.0654796094198735, + "grad_norm": 0.2927928566932678, + "learning_rate": 3.6571358604844687e-06, + "loss": 0.6385, + "step": 2782 + }, + { + "epoch": 1.0658625311123875, + "grad_norm": 0.30552494525909424, + "learning_rate": 3.656235462536188e-06, + "loss": 0.6219, + "step": 2783 + }, + { + "epoch": 1.0662454528049015, + "grad_norm": 0.297821044921875, + "learning_rate": 3.6553348737564328e-06, + "loss": 0.6452, + "step": 2784 + }, + { + "epoch": 1.0666283744974152, + "grad_norm": 0.30037564039230347, + "learning_rate": 3.654434094293843e-06, + "loss": 0.6504, + "step": 2785 + }, + { + "epoch": 1.0670112961899292, + "grad_norm": 0.33273208141326904, + "learning_rate": 3.6535331242970864e-06, + "loss": 0.6168, + "step": 2786 + }, + { + "epoch": 1.067394217882443, + "grad_norm": 0.2969847023487091, + "learning_rate": 3.6526319639148646e-06, + "loss": 0.6378, + "step": 2787 + }, + { + "epoch": 1.067777139574957, + "grad_norm": 0.3362911343574524, + "learning_rate": 3.6517306132959102e-06, + "loss": 0.6386, + "step": 2788 + }, + { + "epoch": 1.0681600612674709, + "grad_norm": 0.29880666732788086, + "learning_rate": 3.650829072588987e-06, + "loss": 0.6434, + "step": 2789 + }, + { + "epoch": 1.0685429829599846, + "grad_norm": 0.2969112992286682, + "learning_rate": 3.6499273419428916e-06, + "loss": 0.6282, + "step": 2790 + }, + { + "epoch": 1.0689259046524986, + "grad_norm": 0.3212871849536896, + "learning_rate": 3.6490254215064485e-06, + "loss": 0.6394, + "step": 2791 + }, + { + "epoch": 1.0693088263450123, + "grad_norm": 0.29578179121017456, + "learning_rate": 3.6481233114285163e-06, + "loss": 0.6245, + "step": 2792 + }, + { + "epoch": 1.0696917480375263, + "grad_norm": 0.30621272325515747, + "learning_rate": 3.6472210118579856e-06, + "loss": 0.6337, + "step": 2793 + }, + { + "epoch": 1.0700746697300403, + "grad_norm": 0.30127787590026855, + "learning_rate": 3.646318522943774e-06, + "loss": 0.6323, + "step": 2794 + }, + { + "epoch": 1.070457591422554, + "grad_norm": 0.2997731566429138, + "learning_rate": 3.6454158448348363e-06, + "loss": 0.6366, + "step": 2795 + }, + { + "epoch": 1.070840513115068, + "grad_norm": 0.31373804807662964, + "learning_rate": 3.644512977680153e-06, + "loss": 0.6399, + "step": 2796 + }, + { + "epoch": 1.0712234348075818, + "grad_norm": 0.3003597557544708, + "learning_rate": 3.6436099216287407e-06, + "loss": 0.6137, + "step": 2797 + }, + { + "epoch": 1.0716063565000957, + "grad_norm": 0.2922988533973694, + "learning_rate": 3.6427066768296425e-06, + "loss": 0.6313, + "step": 2798 + }, + { + "epoch": 1.0719892781926097, + "grad_norm": 0.2972053587436676, + "learning_rate": 3.641803243431936e-06, + "loss": 0.6407, + "step": 2799 + }, + { + "epoch": 1.0723721998851234, + "grad_norm": 0.29849258065223694, + "learning_rate": 3.6408996215847292e-06, + "loss": 0.6436, + "step": 2800 + }, + { + "epoch": 1.0727551215776374, + "grad_norm": 0.29572793841362, + "learning_rate": 3.6399958114371597e-06, + "loss": 0.6261, + "step": 2801 + }, + { + "epoch": 1.0731380432701512, + "grad_norm": 0.29902389645576477, + "learning_rate": 3.6390918131383983e-06, + "loss": 0.6352, + "step": 2802 + }, + { + "epoch": 1.0735209649626651, + "grad_norm": 0.29438233375549316, + "learning_rate": 3.638187626837645e-06, + "loss": 0.6642, + "step": 2803 + }, + { + "epoch": 1.073903886655179, + "grad_norm": 0.2985974848270416, + "learning_rate": 3.637283252684133e-06, + "loss": 0.6488, + "step": 2804 + }, + { + "epoch": 1.0742868083476929, + "grad_norm": 0.3008054792881012, + "learning_rate": 3.636378690827124e-06, + "loss": 0.6352, + "step": 2805 + }, + { + "epoch": 1.0746697300402068, + "grad_norm": 0.28973904252052307, + "learning_rate": 3.6354739414159114e-06, + "loss": 0.6513, + "step": 2806 + }, + { + "epoch": 1.0750526517327206, + "grad_norm": 0.3014800548553467, + "learning_rate": 3.6345690045998223e-06, + "loss": 0.6581, + "step": 2807 + }, + { + "epoch": 1.0754355734252345, + "grad_norm": 0.30003228783607483, + "learning_rate": 3.63366388052821e-06, + "loss": 0.617, + "step": 2808 + }, + { + "epoch": 1.0758184951177485, + "grad_norm": 0.29623013734817505, + "learning_rate": 3.632758569350463e-06, + "loss": 0.6413, + "step": 2809 + }, + { + "epoch": 1.0762014168102623, + "grad_norm": 0.29669392108917236, + "learning_rate": 3.6318530712159972e-06, + "loss": 0.6369, + "step": 2810 + }, + { + "epoch": 1.0765843385027762, + "grad_norm": 0.30288803577423096, + "learning_rate": 3.6309473862742616e-06, + "loss": 0.6409, + "step": 2811 + }, + { + "epoch": 1.07696726019529, + "grad_norm": 0.29814502596855164, + "learning_rate": 3.630041514674737e-06, + "loss": 0.6325, + "step": 2812 + }, + { + "epoch": 1.077350181887804, + "grad_norm": 0.3065796196460724, + "learning_rate": 3.6291354565669314e-06, + "loss": 0.6329, + "step": 2813 + }, + { + "epoch": 1.077733103580318, + "grad_norm": 0.29496321082115173, + "learning_rate": 3.6282292121003865e-06, + "loss": 0.6521, + "step": 2814 + }, + { + "epoch": 1.0781160252728317, + "grad_norm": 0.3122825026512146, + "learning_rate": 3.627322781424673e-06, + "loss": 0.6477, + "step": 2815 + }, + { + "epoch": 1.0784989469653457, + "grad_norm": 0.30609387159347534, + "learning_rate": 3.626416164689394e-06, + "loss": 0.6334, + "step": 2816 + }, + { + "epoch": 1.0788818686578594, + "grad_norm": 0.3035013973712921, + "learning_rate": 3.6255093620441835e-06, + "loss": 0.629, + "step": 2817 + }, + { + "epoch": 1.0792647903503734, + "grad_norm": 0.3003818094730377, + "learning_rate": 3.6246023736387035e-06, + "loss": 0.6471, + "step": 2818 + }, + { + "epoch": 1.0796477120428873, + "grad_norm": 0.30532458424568176, + "learning_rate": 3.62369519962265e-06, + "loss": 0.642, + "step": 2819 + }, + { + "epoch": 1.080030633735401, + "grad_norm": 0.30444273352622986, + "learning_rate": 3.6227878401457458e-06, + "loss": 0.659, + "step": 2820 + }, + { + "epoch": 1.080413555427915, + "grad_norm": 0.31010353565216064, + "learning_rate": 3.621880295357748e-06, + "loss": 0.6233, + "step": 2821 + }, + { + "epoch": 1.0807964771204288, + "grad_norm": 0.3040667474269867, + "learning_rate": 3.6209725654084427e-06, + "loss": 0.6457, + "step": 2822 + }, + { + "epoch": 1.0811793988129428, + "grad_norm": 0.30253347754478455, + "learning_rate": 3.6200646504476465e-06, + "loss": 0.6664, + "step": 2823 + }, + { + "epoch": 1.0815623205054565, + "grad_norm": 0.3174254298210144, + "learning_rate": 3.619156550625207e-06, + "loss": 0.6399, + "step": 2824 + }, + { + "epoch": 1.0819452421979705, + "grad_norm": 0.2984876036643982, + "learning_rate": 3.6182482660910013e-06, + "loss": 0.6426, + "step": 2825 + }, + { + "epoch": 1.0823281638904845, + "grad_norm": 0.30349209904670715, + "learning_rate": 3.617339796994938e-06, + "loss": 0.6476, + "step": 2826 + }, + { + "epoch": 1.0827110855829982, + "grad_norm": 0.2968597114086151, + "learning_rate": 3.616431143486956e-06, + "loss": 0.6424, + "step": 2827 + }, + { + "epoch": 1.0830940072755122, + "grad_norm": 0.2884227931499481, + "learning_rate": 3.6155223057170247e-06, + "loss": 0.6467, + "step": 2828 + }, + { + "epoch": 1.083476928968026, + "grad_norm": 0.2893022894859314, + "learning_rate": 3.6146132838351432e-06, + "loss": 0.6284, + "step": 2829 + }, + { + "epoch": 1.08385985066054, + "grad_norm": 0.29164621233940125, + "learning_rate": 3.613704077991341e-06, + "loss": 0.6323, + "step": 2830 + }, + { + "epoch": 1.0842427723530539, + "grad_norm": 0.28902071714401245, + "learning_rate": 3.61279468833568e-06, + "loss": 0.6315, + "step": 2831 + }, + { + "epoch": 1.0846256940455676, + "grad_norm": 0.3045511543750763, + "learning_rate": 3.6118851150182497e-06, + "loss": 0.6402, + "step": 2832 + }, + { + "epoch": 1.0850086157380816, + "grad_norm": 0.29124462604522705, + "learning_rate": 3.6109753581891717e-06, + "loss": 0.6492, + "step": 2833 + }, + { + "epoch": 1.0853915374305954, + "grad_norm": 0.30100876092910767, + "learning_rate": 3.6100654179985966e-06, + "loss": 0.6379, + "step": 2834 + }, + { + "epoch": 1.0857744591231093, + "grad_norm": 0.30462414026260376, + "learning_rate": 3.6091552945967063e-06, + "loss": 0.631, + "step": 2835 + }, + { + "epoch": 1.0861573808156233, + "grad_norm": 0.2924759089946747, + "learning_rate": 3.6082449881337132e-06, + "loss": 0.6305, + "step": 2836 + }, + { + "epoch": 1.086540302508137, + "grad_norm": 0.2975121736526489, + "learning_rate": 3.607334498759858e-06, + "loss": 0.6176, + "step": 2837 + }, + { + "epoch": 1.086923224200651, + "grad_norm": 0.3037700355052948, + "learning_rate": 3.606423826625414e-06, + "loss": 0.634, + "step": 2838 + }, + { + "epoch": 1.0873061458931648, + "grad_norm": 0.29970163106918335, + "learning_rate": 3.6055129718806836e-06, + "loss": 0.619, + "step": 2839 + }, + { + "epoch": 1.0876890675856787, + "grad_norm": 0.3016922175884247, + "learning_rate": 3.6046019346759973e-06, + "loss": 0.6319, + "step": 2840 + }, + { + "epoch": 1.0880719892781927, + "grad_norm": 0.32059445977211, + "learning_rate": 3.603690715161721e-06, + "loss": 0.6135, + "step": 2841 + }, + { + "epoch": 1.0884549109707065, + "grad_norm": 0.29288583993911743, + "learning_rate": 3.6027793134882436e-06, + "loss": 0.642, + "step": 2842 + }, + { + "epoch": 1.0888378326632204, + "grad_norm": 0.2927685081958771, + "learning_rate": 3.6018677298059924e-06, + "loss": 0.6451, + "step": 2843 + }, + { + "epoch": 1.0892207543557342, + "grad_norm": 0.2864687740802765, + "learning_rate": 3.600955964265416e-06, + "loss": 0.6275, + "step": 2844 + }, + { + "epoch": 1.0896036760482481, + "grad_norm": 0.3134166896343231, + "learning_rate": 3.600044017016999e-06, + "loss": 0.6291, + "step": 2845 + }, + { + "epoch": 1.089986597740762, + "grad_norm": 0.30348706245422363, + "learning_rate": 3.5991318882112547e-06, + "loss": 0.6305, + "step": 2846 + }, + { + "epoch": 1.0903695194332759, + "grad_norm": 0.3097226619720459, + "learning_rate": 3.5982195779987246e-06, + "loss": 0.6292, + "step": 2847 + }, + { + "epoch": 1.0907524411257898, + "grad_norm": 0.29113224148750305, + "learning_rate": 3.597307086529982e-06, + "loss": 0.6424, + "step": 2848 + }, + { + "epoch": 1.0911353628183036, + "grad_norm": 0.3042992353439331, + "learning_rate": 3.5963944139556293e-06, + "loss": 0.6597, + "step": 2849 + }, + { + "epoch": 1.0915182845108176, + "grad_norm": 0.29799994826316833, + "learning_rate": 3.5954815604263004e-06, + "loss": 0.6449, + "step": 2850 + }, + { + "epoch": 1.0919012062033313, + "grad_norm": 0.30045947432518005, + "learning_rate": 3.594568526092656e-06, + "loss": 0.6541, + "step": 2851 + }, + { + "epoch": 1.0922841278958453, + "grad_norm": 0.30128735303878784, + "learning_rate": 3.5936553111053885e-06, + "loss": 0.6451, + "step": 2852 + }, + { + "epoch": 1.0926670495883593, + "grad_norm": 0.30706992745399475, + "learning_rate": 3.5927419156152217e-06, + "loss": 0.6414, + "step": 2853 + }, + { + "epoch": 1.093049971280873, + "grad_norm": 0.30036231875419617, + "learning_rate": 3.591828339772904e-06, + "loss": 0.6481, + "step": 2854 + }, + { + "epoch": 1.093432892973387, + "grad_norm": 0.3107207715511322, + "learning_rate": 3.5909145837292207e-06, + "loss": 0.6448, + "step": 2855 + }, + { + "epoch": 1.0938158146659007, + "grad_norm": 0.31839489936828613, + "learning_rate": 3.59000064763498e-06, + "loss": 0.6335, + "step": 2856 + }, + { + "epoch": 1.0941987363584147, + "grad_norm": 0.29632407426834106, + "learning_rate": 3.589086531641026e-06, + "loss": 0.6258, + "step": 2857 + }, + { + "epoch": 1.0945816580509287, + "grad_norm": 0.30535823106765747, + "learning_rate": 3.588172235898227e-06, + "loss": 0.6426, + "step": 2858 + }, + { + "epoch": 1.0949645797434424, + "grad_norm": 0.3217841386795044, + "learning_rate": 3.5872577605574845e-06, + "loss": 0.6213, + "step": 2859 + }, + { + "epoch": 1.0953475014359564, + "grad_norm": 0.2977196276187897, + "learning_rate": 3.586343105769728e-06, + "loss": 0.6357, + "step": 2860 + }, + { + "epoch": 1.0957304231284701, + "grad_norm": 0.3202294707298279, + "learning_rate": 3.5854282716859177e-06, + "loss": 0.632, + "step": 2861 + }, + { + "epoch": 1.096113344820984, + "grad_norm": 0.30617111921310425, + "learning_rate": 3.5845132584570426e-06, + "loss": 0.641, + "step": 2862 + }, + { + "epoch": 1.096496266513498, + "grad_norm": 0.29896652698516846, + "learning_rate": 3.583598066234122e-06, + "loss": 0.6253, + "step": 2863 + }, + { + "epoch": 1.0968791882060118, + "grad_norm": 0.3173781931400299, + "learning_rate": 3.582682695168203e-06, + "loss": 0.6336, + "step": 2864 + }, + { + "epoch": 1.0972621098985258, + "grad_norm": 0.3127101957798004, + "learning_rate": 3.5817671454103653e-06, + "loss": 0.6368, + "step": 2865 + }, + { + "epoch": 1.0976450315910395, + "grad_norm": 0.30277255177497864, + "learning_rate": 3.580851417111714e-06, + "loss": 0.6157, + "step": 2866 + }, + { + "epoch": 1.0980279532835535, + "grad_norm": 0.3053516745567322, + "learning_rate": 3.5799355104233887e-06, + "loss": 0.6243, + "step": 2867 + }, + { + "epoch": 1.0984108749760675, + "grad_norm": 0.29536378383636475, + "learning_rate": 3.5790194254965536e-06, + "loss": 0.6448, + "step": 2868 + }, + { + "epoch": 1.0987937966685812, + "grad_norm": 0.32280904054641724, + "learning_rate": 3.5781031624824047e-06, + "loss": 0.6381, + "step": 2869 + }, + { + "epoch": 1.0991767183610952, + "grad_norm": 0.30329760909080505, + "learning_rate": 3.577186721532168e-06, + "loss": 0.6455, + "step": 2870 + }, + { + "epoch": 1.099559640053609, + "grad_norm": 0.3000885844230652, + "learning_rate": 3.5762701027970975e-06, + "loss": 0.6348, + "step": 2871 + }, + { + "epoch": 1.099942561746123, + "grad_norm": 0.30269700288772583, + "learning_rate": 3.575353306428476e-06, + "loss": 0.6281, + "step": 2872 + }, + { + "epoch": 1.100325483438637, + "grad_norm": 0.31272220611572266, + "learning_rate": 3.574436332577619e-06, + "loss": 0.6369, + "step": 2873 + }, + { + "epoch": 1.1007084051311506, + "grad_norm": 0.29924291372299194, + "learning_rate": 3.5735191813958657e-06, + "loss": 0.6585, + "step": 2874 + }, + { + "epoch": 1.1010913268236646, + "grad_norm": 0.30852073431015015, + "learning_rate": 3.5726018530345913e-06, + "loss": 0.6235, + "step": 2875 + }, + { + "epoch": 1.1014742485161784, + "grad_norm": 0.2975667119026184, + "learning_rate": 3.571684347645194e-06, + "loss": 0.6476, + "step": 2876 + }, + { + "epoch": 1.1018571702086923, + "grad_norm": 0.30046606063842773, + "learning_rate": 3.570766665379105e-06, + "loss": 0.6513, + "step": 2877 + }, + { + "epoch": 1.1022400919012063, + "grad_norm": 0.30504873394966125, + "learning_rate": 3.569848806387784e-06, + "loss": 0.6461, + "step": 2878 + }, + { + "epoch": 1.10262301359372, + "grad_norm": 0.308785617351532, + "learning_rate": 3.5689307708227183e-06, + "loss": 0.6427, + "step": 2879 + }, + { + "epoch": 1.103005935286234, + "grad_norm": 0.30854329466819763, + "learning_rate": 3.568012558835427e-06, + "loss": 0.6478, + "step": 2880 + }, + { + "epoch": 1.1033888569787478, + "grad_norm": 0.3001681864261627, + "learning_rate": 3.5670941705774553e-06, + "loss": 0.6355, + "step": 2881 + }, + { + "epoch": 1.1037717786712617, + "grad_norm": 0.2990501821041107, + "learning_rate": 3.566175606200382e-06, + "loss": 0.6345, + "step": 2882 + }, + { + "epoch": 1.1041547003637757, + "grad_norm": 0.2952714264392853, + "learning_rate": 3.565256865855808e-06, + "loss": 0.6488, + "step": 2883 + }, + { + "epoch": 1.1045376220562895, + "grad_norm": 0.3039375841617584, + "learning_rate": 3.5643379496953694e-06, + "loss": 0.6354, + "step": 2884 + }, + { + "epoch": 1.1049205437488034, + "grad_norm": 0.3012932538986206, + "learning_rate": 3.56341885787073e-06, + "loss": 0.656, + "step": 2885 + }, + { + "epoch": 1.1053034654413172, + "grad_norm": 0.31279462575912476, + "learning_rate": 3.56249959053358e-06, + "loss": 0.6239, + "step": 2886 + }, + { + "epoch": 1.1056863871338312, + "grad_norm": 0.29985374212265015, + "learning_rate": 3.5615801478356416e-06, + "loss": 0.6191, + "step": 2887 + }, + { + "epoch": 1.106069308826345, + "grad_norm": 0.30841219425201416, + "learning_rate": 3.560660529928664e-06, + "loss": 0.6512, + "step": 2888 + }, + { + "epoch": 1.1064522305188589, + "grad_norm": 0.30682387948036194, + "learning_rate": 3.5597407369644266e-06, + "loss": 0.599, + "step": 2889 + }, + { + "epoch": 1.1068351522113729, + "grad_norm": 0.302079975605011, + "learning_rate": 3.558820769094737e-06, + "loss": 0.6415, + "step": 2890 + }, + { + "epoch": 1.1072180739038866, + "grad_norm": 0.29730144143104553, + "learning_rate": 3.5579006264714304e-06, + "loss": 0.6367, + "step": 2891 + }, + { + "epoch": 1.1076009955964006, + "grad_norm": 0.30473604798316956, + "learning_rate": 3.5569803092463752e-06, + "loss": 0.6449, + "step": 2892 + }, + { + "epoch": 1.1079839172889143, + "grad_norm": 0.2956167757511139, + "learning_rate": 3.5560598175714617e-06, + "loss": 0.6392, + "step": 2893 + }, + { + "epoch": 1.1083668389814283, + "grad_norm": 0.300261914730072, + "learning_rate": 3.5551391515986163e-06, + "loss": 0.6423, + "step": 2894 + }, + { + "epoch": 1.1087497606739423, + "grad_norm": 0.3110640048980713, + "learning_rate": 3.5542183114797885e-06, + "loss": 0.6391, + "step": 2895 + }, + { + "epoch": 1.109132682366456, + "grad_norm": 0.2972400188446045, + "learning_rate": 3.5532972973669607e-06, + "loss": 0.6293, + "step": 2896 + }, + { + "epoch": 1.10951560405897, + "grad_norm": 0.2988949716091156, + "learning_rate": 3.5523761094121414e-06, + "loss": 0.6355, + "step": 2897 + }, + { + "epoch": 1.1098985257514837, + "grad_norm": 0.31272605061531067, + "learning_rate": 3.551454747767367e-06, + "loss": 0.6539, + "step": 2898 + }, + { + "epoch": 1.1102814474439977, + "grad_norm": 0.29314085841178894, + "learning_rate": 3.5505332125847068e-06, + "loss": 0.6263, + "step": 2899 + }, + { + "epoch": 1.1106643691365117, + "grad_norm": 0.30604884028434753, + "learning_rate": 3.549611504016255e-06, + "loss": 0.6436, + "step": 2900 + }, + { + "epoch": 1.1110472908290254, + "grad_norm": 0.29860472679138184, + "learning_rate": 3.5486896222141336e-06, + "loss": 0.6524, + "step": 2901 + }, + { + "epoch": 1.1114302125215394, + "grad_norm": 0.30778035521507263, + "learning_rate": 3.5477675673304978e-06, + "loss": 0.6126, + "step": 2902 + }, + { + "epoch": 1.1118131342140531, + "grad_norm": 0.2969551980495453, + "learning_rate": 3.5468453395175277e-06, + "loss": 0.6329, + "step": 2903 + }, + { + "epoch": 1.1121960559065671, + "grad_norm": 0.310872882604599, + "learning_rate": 3.5459229389274323e-06, + "loss": 0.6522, + "step": 2904 + }, + { + "epoch": 1.112578977599081, + "grad_norm": 0.3107014298439026, + "learning_rate": 3.54500036571245e-06, + "loss": 0.6389, + "step": 2905 + }, + { + "epoch": 1.1129618992915948, + "grad_norm": 0.3006989061832428, + "learning_rate": 3.544077620024847e-06, + "loss": 0.6596, + "step": 2906 + }, + { + "epoch": 1.1133448209841088, + "grad_norm": 0.29612088203430176, + "learning_rate": 3.54315470201692e-06, + "loss": 0.6349, + "step": 2907 + }, + { + "epoch": 1.1137277426766226, + "grad_norm": 0.3026573359966278, + "learning_rate": 3.5422316118409895e-06, + "loss": 0.6385, + "step": 2908 + }, + { + "epoch": 1.1141106643691365, + "grad_norm": 0.2943755090236664, + "learning_rate": 3.5413083496494104e-06, + "loss": 0.6214, + "step": 2909 + }, + { + "epoch": 1.1144935860616503, + "grad_norm": 0.301417738199234, + "learning_rate": 3.5403849155945613e-06, + "loss": 0.6492, + "step": 2910 + }, + { + "epoch": 1.1148765077541642, + "grad_norm": 0.29453423619270325, + "learning_rate": 3.5394613098288504e-06, + "loss": 0.6365, + "step": 2911 + }, + { + "epoch": 1.1152594294466782, + "grad_norm": 0.29185009002685547, + "learning_rate": 3.5385375325047167e-06, + "loss": 0.6315, + "step": 2912 + }, + { + "epoch": 1.115642351139192, + "grad_norm": 0.30309614539146423, + "learning_rate": 3.5376135837746227e-06, + "loss": 0.6468, + "step": 2913 + }, + { + "epoch": 1.116025272831706, + "grad_norm": 0.29688191413879395, + "learning_rate": 3.536689463791066e-06, + "loss": 0.6257, + "step": 2914 + }, + { + "epoch": 1.1164081945242197, + "grad_norm": 0.2886456251144409, + "learning_rate": 3.535765172706564e-06, + "loss": 0.6164, + "step": 2915 + }, + { + "epoch": 1.1167911162167337, + "grad_norm": 0.29280808568000793, + "learning_rate": 3.53484071067367e-06, + "loss": 0.6312, + "step": 2916 + }, + { + "epoch": 1.1171740379092476, + "grad_norm": 0.29417091608047485, + "learning_rate": 3.5339160778449606e-06, + "loss": 0.6446, + "step": 2917 + }, + { + "epoch": 1.1175569596017614, + "grad_norm": 0.30296099185943604, + "learning_rate": 3.532991274373043e-06, + "loss": 0.6407, + "step": 2918 + }, + { + "epoch": 1.1179398812942754, + "grad_norm": 0.2997060716152191, + "learning_rate": 3.5320663004105515e-06, + "loss": 0.6375, + "step": 2919 + }, + { + "epoch": 1.118322802986789, + "grad_norm": 0.29506009817123413, + "learning_rate": 3.5311411561101488e-06, + "loss": 0.6281, + "step": 2920 + }, + { + "epoch": 1.118705724679303, + "grad_norm": 0.29762738943099976, + "learning_rate": 3.5302158416245274e-06, + "loss": 0.6384, + "step": 2921 + }, + { + "epoch": 1.119088646371817, + "grad_norm": 0.30581051111221313, + "learning_rate": 3.529290357106404e-06, + "loss": 0.6436, + "step": 2922 + }, + { + "epoch": 1.1194715680643308, + "grad_norm": 0.3248584270477295, + "learning_rate": 3.5283647027085266e-06, + "loss": 0.6256, + "step": 2923 + }, + { + "epoch": 1.1198544897568448, + "grad_norm": 0.3132539987564087, + "learning_rate": 3.527438878583671e-06, + "loss": 0.6422, + "step": 2924 + }, + { + "epoch": 1.1202374114493585, + "grad_norm": 0.296010285615921, + "learning_rate": 3.5265128848846397e-06, + "loss": 0.6006, + "step": 2925 + }, + { + "epoch": 1.1206203331418725, + "grad_norm": 0.32502424716949463, + "learning_rate": 3.5255867217642636e-06, + "loss": 0.6513, + "step": 2926 + }, + { + "epoch": 1.1210032548343865, + "grad_norm": 0.30820217728614807, + "learning_rate": 3.5246603893754017e-06, + "loss": 0.6328, + "step": 2927 + }, + { + "epoch": 1.1213861765269002, + "grad_norm": 0.303303062915802, + "learning_rate": 3.5237338878709413e-06, + "loss": 0.6434, + "step": 2928 + }, + { + "epoch": 1.1217690982194142, + "grad_norm": 0.2993101477622986, + "learning_rate": 3.522807217403798e-06, + "loss": 0.6535, + "step": 2929 + }, + { + "epoch": 1.122152019911928, + "grad_norm": 0.30999955534935, + "learning_rate": 3.521880378126913e-06, + "loss": 0.6302, + "step": 2930 + }, + { + "epoch": 1.122534941604442, + "grad_norm": 0.3004814088344574, + "learning_rate": 3.520953370193259e-06, + "loss": 0.6333, + "step": 2931 + }, + { + "epoch": 1.1229178632969559, + "grad_norm": 0.30235061049461365, + "learning_rate": 3.5200261937558323e-06, + "loss": 0.6616, + "step": 2932 + }, + { + "epoch": 1.1233007849894696, + "grad_norm": 0.2931925654411316, + "learning_rate": 3.5190988489676614e-06, + "loss": 0.6628, + "step": 2933 + }, + { + "epoch": 1.1236837066819836, + "grad_norm": 0.31069281697273254, + "learning_rate": 3.5181713359817984e-06, + "loss": 0.6575, + "step": 2934 + }, + { + "epoch": 1.1240666283744973, + "grad_norm": 0.2995961606502533, + "learning_rate": 3.517243654951326e-06, + "loss": 0.6473, + "step": 2935 + }, + { + "epoch": 1.1244495500670113, + "grad_norm": 0.3044862747192383, + "learning_rate": 3.516315806029354e-06, + "loss": 0.6302, + "step": 2936 + }, + { + "epoch": 1.1248324717595253, + "grad_norm": 0.30701902508735657, + "learning_rate": 3.515387789369019e-06, + "loss": 0.6594, + "step": 2937 + }, + { + "epoch": 1.125215393452039, + "grad_norm": 0.300830215215683, + "learning_rate": 3.5144596051234875e-06, + "loss": 0.6336, + "step": 2938 + }, + { + "epoch": 1.125598315144553, + "grad_norm": 0.30521881580352783, + "learning_rate": 3.513531253445951e-06, + "loss": 0.6519, + "step": 2939 + }, + { + "epoch": 1.1259812368370667, + "grad_norm": 0.3063434064388275, + "learning_rate": 3.5126027344896295e-06, + "loss": 0.6314, + "step": 2940 + }, + { + "epoch": 1.1263641585295807, + "grad_norm": 0.31025150418281555, + "learning_rate": 3.5116740484077715e-06, + "loss": 0.6245, + "step": 2941 + }, + { + "epoch": 1.1267470802220947, + "grad_norm": 0.30604809522628784, + "learning_rate": 3.5107451953536526e-06, + "loss": 0.6416, + "step": 2942 + }, + { + "epoch": 1.1271300019146084, + "grad_norm": 0.3047782778739929, + "learning_rate": 3.509816175480575e-06, + "loss": 0.638, + "step": 2943 + }, + { + "epoch": 1.1275129236071224, + "grad_norm": 0.29616108536720276, + "learning_rate": 3.50888698894187e-06, + "loss": 0.6148, + "step": 2944 + }, + { + "epoch": 1.1278958452996362, + "grad_norm": 0.2975909411907196, + "learning_rate": 3.507957635890895e-06, + "loss": 0.6585, + "step": 2945 + }, + { + "epoch": 1.1282787669921501, + "grad_norm": 0.3045974373817444, + "learning_rate": 3.507028116481037e-06, + "loss": 0.6534, + "step": 2946 + }, + { + "epoch": 1.128661688684664, + "grad_norm": 0.30258604884147644, + "learning_rate": 3.506098430865707e-06, + "loss": 0.6428, + "step": 2947 + }, + { + "epoch": 1.1290446103771778, + "grad_norm": 0.30882376432418823, + "learning_rate": 3.505168579198347e-06, + "loss": 0.6433, + "step": 2948 + }, + { + "epoch": 1.1294275320696918, + "grad_norm": 0.2948397099971771, + "learning_rate": 3.5042385616324243e-06, + "loss": 0.632, + "step": 2949 + }, + { + "epoch": 1.1298104537622056, + "grad_norm": 0.2953971028327942, + "learning_rate": 3.5033083783214334e-06, + "loss": 0.6572, + "step": 2950 + }, + { + "epoch": 1.1301933754547195, + "grad_norm": 0.3047292232513428, + "learning_rate": 3.502378029418898e-06, + "loss": 0.6166, + "step": 2951 + }, + { + "epoch": 1.1305762971472335, + "grad_norm": 0.3040366768836975, + "learning_rate": 3.5014475150783676e-06, + "loss": 0.6374, + "step": 2952 + }, + { + "epoch": 1.1309592188397473, + "grad_norm": 0.30481603741645813, + "learning_rate": 3.50051683545342e-06, + "loss": 0.6513, + "step": 2953 + }, + { + "epoch": 1.1313421405322612, + "grad_norm": 0.28988370299339294, + "learning_rate": 3.499585990697659e-06, + "loss": 0.6455, + "step": 2954 + }, + { + "epoch": 1.131725062224775, + "grad_norm": 0.30798378586769104, + "learning_rate": 3.4986549809647157e-06, + "loss": 0.6576, + "step": 2955 + }, + { + "epoch": 1.132107983917289, + "grad_norm": 0.296202689409256, + "learning_rate": 3.49772380640825e-06, + "loss": 0.6152, + "step": 2956 + }, + { + "epoch": 1.1324909056098027, + "grad_norm": 0.31315869092941284, + "learning_rate": 3.496792467181948e-06, + "loss": 0.6355, + "step": 2957 + }, + { + "epoch": 1.1328738273023167, + "grad_norm": 0.3043880760669708, + "learning_rate": 3.4958609634395235e-06, + "loss": 0.6318, + "step": 2958 + }, + { + "epoch": 1.1332567489948306, + "grad_norm": 0.30076542496681213, + "learning_rate": 3.494929295334716e-06, + "loss": 0.6405, + "step": 2959 + }, + { + "epoch": 1.1336396706873444, + "grad_norm": 0.3020724058151245, + "learning_rate": 3.493997463021294e-06, + "loss": 0.6523, + "step": 2960 + }, + { + "epoch": 1.1340225923798584, + "grad_norm": 0.30913645029067993, + "learning_rate": 3.4930654666530518e-06, + "loss": 0.6614, + "step": 2961 + }, + { + "epoch": 1.1344055140723721, + "grad_norm": 0.3013768792152405, + "learning_rate": 3.492133306383811e-06, + "loss": 0.6153, + "step": 2962 + }, + { + "epoch": 1.134788435764886, + "grad_norm": 0.3051517903804779, + "learning_rate": 3.491200982367421e-06, + "loss": 0.623, + "step": 2963 + }, + { + "epoch": 1.1351713574574, + "grad_norm": 0.30079856514930725, + "learning_rate": 3.490268494757757e-06, + "loss": 0.6422, + "step": 2964 + }, + { + "epoch": 1.1355542791499138, + "grad_norm": 0.297507643699646, + "learning_rate": 3.4893358437087225e-06, + "loss": 0.6495, + "step": 2965 + }, + { + "epoch": 1.1359372008424278, + "grad_norm": 0.2935941815376282, + "learning_rate": 3.4884030293742473e-06, + "loss": 0.617, + "step": 2966 + }, + { + "epoch": 1.1363201225349415, + "grad_norm": 0.3005307614803314, + "learning_rate": 3.487470051908288e-06, + "loss": 0.6411, + "step": 2967 + }, + { + "epoch": 1.1367030442274555, + "grad_norm": 0.2922939360141754, + "learning_rate": 3.486536911464829e-06, + "loss": 0.6219, + "step": 2968 + }, + { + "epoch": 1.1370859659199692, + "grad_norm": 0.2942897081375122, + "learning_rate": 3.48560360819788e-06, + "loss": 0.6354, + "step": 2969 + }, + { + "epoch": 1.1374688876124832, + "grad_norm": 0.29832935333251953, + "learning_rate": 3.4846701422614797e-06, + "loss": 0.6276, + "step": 2970 + }, + { + "epoch": 1.1378518093049972, + "grad_norm": 0.2939033806324005, + "learning_rate": 3.4837365138096906e-06, + "loss": 0.6285, + "step": 2971 + }, + { + "epoch": 1.138234730997511, + "grad_norm": 0.31000685691833496, + "learning_rate": 3.4828027229966057e-06, + "loss": 0.632, + "step": 2972 + }, + { + "epoch": 1.138617652690025, + "grad_norm": 0.29513347148895264, + "learning_rate": 3.4818687699763416e-06, + "loss": 0.618, + "step": 2973 + }, + { + "epoch": 1.1390005743825387, + "grad_norm": 0.2842392325401306, + "learning_rate": 3.4809346549030444e-06, + "loss": 0.6431, + "step": 2974 + }, + { + "epoch": 1.1393834960750526, + "grad_norm": 0.3002409040927887, + "learning_rate": 3.4800003779308843e-06, + "loss": 0.6655, + "step": 2975 + }, + { + "epoch": 1.1397664177675666, + "grad_norm": 0.30052199959754944, + "learning_rate": 3.47906593921406e-06, + "loss": 0.6472, + "step": 2976 + }, + { + "epoch": 1.1401493394600803, + "grad_norm": 0.3065297305583954, + "learning_rate": 3.4781313389067985e-06, + "loss": 0.6541, + "step": 2977 + }, + { + "epoch": 1.1405322611525943, + "grad_norm": 0.28881242871284485, + "learning_rate": 3.4771965771633475e-06, + "loss": 0.6084, + "step": 2978 + }, + { + "epoch": 1.140915182845108, + "grad_norm": 0.29813912510871887, + "learning_rate": 3.4762616541379873e-06, + "loss": 0.6247, + "step": 2979 + }, + { + "epoch": 1.141298104537622, + "grad_norm": 0.3057692050933838, + "learning_rate": 3.4753265699850232e-06, + "loss": 0.6318, + "step": 2980 + }, + { + "epoch": 1.141681026230136, + "grad_norm": 0.296878844499588, + "learning_rate": 3.4743913248587855e-06, + "loss": 0.6428, + "step": 2981 + }, + { + "epoch": 1.1420639479226498, + "grad_norm": 0.29651567339897156, + "learning_rate": 3.4734559189136337e-06, + "loss": 0.6406, + "step": 2982 + }, + { + "epoch": 1.1424468696151637, + "grad_norm": 0.2973792850971222, + "learning_rate": 3.472520352303951e-06, + "loss": 0.6363, + "step": 2983 + }, + { + "epoch": 1.1428297913076775, + "grad_norm": 0.2996625006198883, + "learning_rate": 3.4715846251841483e-06, + "loss": 0.6419, + "step": 2984 + }, + { + "epoch": 1.1432127130001914, + "grad_norm": 0.2939167022705078, + "learning_rate": 3.4706487377086645e-06, + "loss": 0.6248, + "step": 2985 + }, + { + "epoch": 1.1435956346927054, + "grad_norm": 0.299533873796463, + "learning_rate": 3.469712690031962e-06, + "loss": 0.6611, + "step": 2986 + }, + { + "epoch": 1.1439785563852192, + "grad_norm": 0.29775145649909973, + "learning_rate": 3.4687764823085336e-06, + "loss": 0.6158, + "step": 2987 + }, + { + "epoch": 1.1443614780777331, + "grad_norm": 0.2970540225505829, + "learning_rate": 3.467840114692894e-06, + "loss": 0.6512, + "step": 2988 + }, + { + "epoch": 1.144744399770247, + "grad_norm": 0.3062463104724884, + "learning_rate": 3.466903587339587e-06, + "loss": 0.6576, + "step": 2989 + }, + { + "epoch": 1.1451273214627609, + "grad_norm": 0.2947791516780853, + "learning_rate": 3.4659669004031827e-06, + "loss": 0.6453, + "step": 2990 + }, + { + "epoch": 1.1455102431552748, + "grad_norm": 0.29653674364089966, + "learning_rate": 3.4650300540382764e-06, + "loss": 0.6293, + "step": 2991 + }, + { + "epoch": 1.1458931648477886, + "grad_norm": 0.3094513416290283, + "learning_rate": 3.464093048399492e-06, + "loss": 0.6407, + "step": 2992 + }, + { + "epoch": 1.1462760865403026, + "grad_norm": 0.3101944923400879, + "learning_rate": 3.4631558836414753e-06, + "loss": 0.6508, + "step": 2993 + }, + { + "epoch": 1.1466590082328163, + "grad_norm": 0.29986560344696045, + "learning_rate": 3.4622185599189034e-06, + "loss": 0.6325, + "step": 2994 + }, + { + "epoch": 1.1470419299253303, + "grad_norm": 0.3093869686126709, + "learning_rate": 3.461281077386477e-06, + "loss": 0.6352, + "step": 2995 + }, + { + "epoch": 1.1474248516178442, + "grad_norm": 0.308662474155426, + "learning_rate": 3.4603434361989225e-06, + "loss": 0.6354, + "step": 2996 + }, + { + "epoch": 1.147807773310358, + "grad_norm": 0.3035331964492798, + "learning_rate": 3.4594056365109946e-06, + "loss": 0.6175, + "step": 2997 + }, + { + "epoch": 1.148190695002872, + "grad_norm": 0.30759233236312866, + "learning_rate": 3.458467678477472e-06, + "loss": 0.6446, + "step": 2998 + }, + { + "epoch": 1.1485736166953857, + "grad_norm": 0.3018249571323395, + "learning_rate": 3.4575295622531614e-06, + "loss": 0.6466, + "step": 2999 + }, + { + "epoch": 1.1489565383878997, + "grad_norm": 0.297177255153656, + "learning_rate": 3.4565912879928924e-06, + "loss": 0.6365, + "step": 3000 + }, + { + "epoch": 1.1493394600804137, + "grad_norm": 0.3035942614078522, + "learning_rate": 3.4556528558515256e-06, + "loss": 0.632, + "step": 3001 + }, + { + "epoch": 1.1497223817729274, + "grad_norm": 0.30431827902793884, + "learning_rate": 3.4547142659839445e-06, + "loss": 0.6559, + "step": 3002 + }, + { + "epoch": 1.1501053034654414, + "grad_norm": 0.29237306118011475, + "learning_rate": 3.453775518545058e-06, + "loss": 0.651, + "step": 3003 + }, + { + "epoch": 1.1504882251579551, + "grad_norm": 0.3006161153316498, + "learning_rate": 3.452836613689803e-06, + "loss": 0.6158, + "step": 3004 + }, + { + "epoch": 1.150871146850469, + "grad_norm": 0.29211270809173584, + "learning_rate": 3.4518975515731407e-06, + "loss": 0.6419, + "step": 3005 + }, + { + "epoch": 1.151254068542983, + "grad_norm": 0.2835102081298828, + "learning_rate": 3.45095833235006e-06, + "loss": 0.6285, + "step": 3006 + }, + { + "epoch": 1.1516369902354968, + "grad_norm": 0.2961466312408447, + "learning_rate": 3.450018956175575e-06, + "loss": 0.6457, + "step": 3007 + }, + { + "epoch": 1.1520199119280108, + "grad_norm": 0.30123957991600037, + "learning_rate": 3.449079423204724e-06, + "loss": 0.6525, + "step": 3008 + }, + { + "epoch": 1.1524028336205245, + "grad_norm": 0.29967620968818665, + "learning_rate": 3.448139733592576e-06, + "loss": 0.6273, + "step": 3009 + }, + { + "epoch": 1.1527857553130385, + "grad_norm": 0.2959441542625427, + "learning_rate": 3.4471998874942182e-06, + "loss": 0.6166, + "step": 3010 + }, + { + "epoch": 1.1531686770055525, + "grad_norm": 0.29656732082366943, + "learning_rate": 3.4462598850647705e-06, + "loss": 0.622, + "step": 3011 + }, + { + "epoch": 1.1535515986980662, + "grad_norm": 0.2923170328140259, + "learning_rate": 3.4453197264593757e-06, + "loss": 0.6427, + "step": 3012 + }, + { + "epoch": 1.1539345203905802, + "grad_norm": 0.3024528920650482, + "learning_rate": 3.444379411833203e-06, + "loss": 0.6417, + "step": 3013 + }, + { + "epoch": 1.154317442083094, + "grad_norm": 0.3019734025001526, + "learning_rate": 3.443438941341447e-06, + "loss": 0.6419, + "step": 3014 + }, + { + "epoch": 1.154700363775608, + "grad_norm": 0.30900803208351135, + "learning_rate": 3.442498315139327e-06, + "loss": 0.635, + "step": 3015 + }, + { + "epoch": 1.155083285468122, + "grad_norm": 0.30060309171676636, + "learning_rate": 3.4415575333820917e-06, + "loss": 0.6025, + "step": 3016 + }, + { + "epoch": 1.1554662071606356, + "grad_norm": 0.3014119565486908, + "learning_rate": 3.44061659622501e-06, + "loss": 0.6326, + "step": 3017 + }, + { + "epoch": 1.1558491288531496, + "grad_norm": 0.3086121380329132, + "learning_rate": 3.4396755038233804e-06, + "loss": 0.6489, + "step": 3018 + }, + { + "epoch": 1.1562320505456634, + "grad_norm": 0.29982659220695496, + "learning_rate": 3.4387342563325273e-06, + "loss": 0.6174, + "step": 3019 + }, + { + "epoch": 1.1566149722381773, + "grad_norm": 0.29232150316238403, + "learning_rate": 3.4377928539077976e-06, + "loss": 0.6161, + "step": 3020 + }, + { + "epoch": 1.156997893930691, + "grad_norm": 0.30171093344688416, + "learning_rate": 3.436851296704567e-06, + "loss": 0.6625, + "step": 3021 + }, + { + "epoch": 1.157380815623205, + "grad_norm": 0.298186331987381, + "learning_rate": 3.435909584878234e-06, + "loss": 0.6213, + "step": 3022 + }, + { + "epoch": 1.157763737315719, + "grad_norm": 0.29324406385421753, + "learning_rate": 3.4349677185842246e-06, + "loss": 0.6468, + "step": 3023 + }, + { + "epoch": 1.1581466590082328, + "grad_norm": 0.2957988679409027, + "learning_rate": 3.434025697977989e-06, + "loss": 0.6453, + "step": 3024 + }, + { + "epoch": 1.1585295807007467, + "grad_norm": 0.314351886510849, + "learning_rate": 3.433083523215004e-06, + "loss": 0.6574, + "step": 3025 + }, + { + "epoch": 1.1589125023932605, + "grad_norm": 0.2999173104763031, + "learning_rate": 3.432141194450772e-06, + "loss": 0.6529, + "step": 3026 + }, + { + "epoch": 1.1592954240857745, + "grad_norm": 0.30799034237861633, + "learning_rate": 3.431198711840819e-06, + "loss": 0.657, + "step": 3027 + }, + { + "epoch": 1.1596783457782884, + "grad_norm": 0.31757575273513794, + "learning_rate": 3.4302560755406976e-06, + "loss": 0.6629, + "step": 3028 + }, + { + "epoch": 1.1600612674708022, + "grad_norm": 0.2985464036464691, + "learning_rate": 3.4293132857059864e-06, + "loss": 0.6368, + "step": 3029 + }, + { + "epoch": 1.1604441891633162, + "grad_norm": 0.29600659012794495, + "learning_rate": 3.428370342492288e-06, + "loss": 0.6504, + "step": 3030 + }, + { + "epoch": 1.16082711085583, + "grad_norm": 0.2982555031776428, + "learning_rate": 3.4274272460552316e-06, + "loss": 0.6289, + "step": 3031 + }, + { + "epoch": 1.1612100325483439, + "grad_norm": 0.30543068051338196, + "learning_rate": 3.42648399655047e-06, + "loss": 0.6252, + "step": 3032 + }, + { + "epoch": 1.1615929542408576, + "grad_norm": 0.29676687717437744, + "learning_rate": 3.4255405941336827e-06, + "loss": 0.5993, + "step": 3033 + }, + { + "epoch": 1.1619758759333716, + "grad_norm": 0.31045234203338623, + "learning_rate": 3.4245970389605754e-06, + "loss": 0.6383, + "step": 3034 + }, + { + "epoch": 1.1623587976258856, + "grad_norm": 0.3006237745285034, + "learning_rate": 3.423653331186875e-06, + "loss": 0.6124, + "step": 3035 + }, + { + "epoch": 1.1627417193183993, + "grad_norm": 0.298446387052536, + "learning_rate": 3.4227094709683388e-06, + "loss": 0.643, + "step": 3036 + }, + { + "epoch": 1.1631246410109133, + "grad_norm": 0.3041348159313202, + "learning_rate": 3.421765458460745e-06, + "loss": 0.6399, + "step": 3037 + }, + { + "epoch": 1.163507562703427, + "grad_norm": 0.30076003074645996, + "learning_rate": 3.4208212938199003e-06, + "loss": 0.6251, + "step": 3038 + }, + { + "epoch": 1.163890484395941, + "grad_norm": 0.3052762746810913, + "learning_rate": 3.4198769772016327e-06, + "loss": 0.6501, + "step": 3039 + }, + { + "epoch": 1.164273406088455, + "grad_norm": 0.30287763476371765, + "learning_rate": 3.4189325087617987e-06, + "loss": 0.6161, + "step": 3040 + }, + { + "epoch": 1.1646563277809687, + "grad_norm": 0.31583505868911743, + "learning_rate": 3.417987888656279e-06, + "loss": 0.6297, + "step": 3041 + }, + { + "epoch": 1.1650392494734827, + "grad_norm": 0.307576984167099, + "learning_rate": 3.4170431170409774e-06, + "loss": 0.6566, + "step": 3042 + }, + { + "epoch": 1.1654221711659964, + "grad_norm": 0.31491604447364807, + "learning_rate": 3.4160981940718252e-06, + "loss": 0.6139, + "step": 3043 + }, + { + "epoch": 1.1658050928585104, + "grad_norm": 0.30323418974876404, + "learning_rate": 3.4151531199047775e-06, + "loss": 0.6247, + "step": 3044 + }, + { + "epoch": 1.1661880145510244, + "grad_norm": 0.29866665601730347, + "learning_rate": 3.414207894695816e-06, + "loss": 0.6317, + "step": 3045 + }, + { + "epoch": 1.1665709362435381, + "grad_norm": 0.30115482211112976, + "learning_rate": 3.4132625186009434e-06, + "loss": 0.6364, + "step": 3046 + }, + { + "epoch": 1.166953857936052, + "grad_norm": 0.29922133684158325, + "learning_rate": 3.4123169917761908e-06, + "loss": 0.6486, + "step": 3047 + }, + { + "epoch": 1.1673367796285659, + "grad_norm": 0.29645836353302, + "learning_rate": 3.411371314377614e-06, + "loss": 0.6282, + "step": 3048 + }, + { + "epoch": 1.1677197013210798, + "grad_norm": 0.29277947545051575, + "learning_rate": 3.4104254865612905e-06, + "loss": 0.6484, + "step": 3049 + }, + { + "epoch": 1.1681026230135938, + "grad_norm": 0.3051757216453552, + "learning_rate": 3.4094795084833284e-06, + "loss": 0.6175, + "step": 3050 + }, + { + "epoch": 1.1684855447061075, + "grad_norm": 0.29768556356430054, + "learning_rate": 3.408533380299854e-06, + "loss": 0.6335, + "step": 3051 + }, + { + "epoch": 1.1688684663986215, + "grad_norm": 0.312002032995224, + "learning_rate": 3.4075871021670233e-06, + "loss": 0.6255, + "step": 3052 + }, + { + "epoch": 1.1692513880911353, + "grad_norm": 0.29926207661628723, + "learning_rate": 3.406640674241015e-06, + "loss": 0.6186, + "step": 3053 + }, + { + "epoch": 1.1696343097836492, + "grad_norm": 0.3107825219631195, + "learning_rate": 3.4056940966780317e-06, + "loss": 0.6368, + "step": 3054 + }, + { + "epoch": 1.1700172314761632, + "grad_norm": 0.3175870478153229, + "learning_rate": 3.404747369634303e-06, + "loss": 0.6367, + "step": 3055 + }, + { + "epoch": 1.170400153168677, + "grad_norm": 0.29598942399024963, + "learning_rate": 3.4038004932660813e-06, + "loss": 0.6264, + "step": 3056 + }, + { + "epoch": 1.170783074861191, + "grad_norm": 0.30206790566444397, + "learning_rate": 3.4028534677296443e-06, + "loss": 0.6233, + "step": 3057 + }, + { + "epoch": 1.1711659965537047, + "grad_norm": 0.30035173892974854, + "learning_rate": 3.401906293181295e-06, + "loss": 0.6476, + "step": 3058 + }, + { + "epoch": 1.1715489182462187, + "grad_norm": 0.3083301782608032, + "learning_rate": 3.4009589697773605e-06, + "loss": 0.6285, + "step": 3059 + }, + { + "epoch": 1.1719318399387326, + "grad_norm": 0.30406492948532104, + "learning_rate": 3.4000114976741905e-06, + "loss": 0.6332, + "step": 3060 + }, + { + "epoch": 1.1723147616312464, + "grad_norm": 0.2946275770664215, + "learning_rate": 3.3990638770281625e-06, + "loss": 0.6305, + "step": 3061 + }, + { + "epoch": 1.1726976833237603, + "grad_norm": 0.3064177930355072, + "learning_rate": 3.3981161079956764e-06, + "loss": 0.6262, + "step": 3062 + }, + { + "epoch": 1.173080605016274, + "grad_norm": 0.29940205812454224, + "learning_rate": 3.397168190733158e-06, + "loss": 0.637, + "step": 3063 + }, + { + "epoch": 1.173463526708788, + "grad_norm": 0.3060838282108307, + "learning_rate": 3.3962201253970563e-06, + "loss": 0.6486, + "step": 3064 + }, + { + "epoch": 1.173846448401302, + "grad_norm": 0.3096790015697479, + "learning_rate": 3.395271912143845e-06, + "loss": 0.6451, + "step": 3065 + }, + { + "epoch": 1.1742293700938158, + "grad_norm": 0.29914188385009766, + "learning_rate": 3.3943235511300228e-06, + "loss": 0.6367, + "step": 3066 + }, + { + "epoch": 1.1746122917863298, + "grad_norm": 0.29788118600845337, + "learning_rate": 3.393375042512112e-06, + "loss": 0.644, + "step": 3067 + }, + { + "epoch": 1.1749952134788435, + "grad_norm": 0.29346421360969543, + "learning_rate": 3.3924263864466597e-06, + "loss": 0.6316, + "step": 3068 + }, + { + "epoch": 1.1753781351713575, + "grad_norm": 0.29294195771217346, + "learning_rate": 3.391477583090238e-06, + "loss": 0.6294, + "step": 3069 + }, + { + "epoch": 1.1757610568638714, + "grad_norm": 0.30053451657295227, + "learning_rate": 3.3905286325994423e-06, + "loss": 0.6325, + "step": 3070 + }, + { + "epoch": 1.1761439785563852, + "grad_norm": 0.3131263554096222, + "learning_rate": 3.3895795351308925e-06, + "loss": 0.6199, + "step": 3071 + }, + { + "epoch": 1.1765269002488992, + "grad_norm": 0.30711281299591064, + "learning_rate": 3.388630290841233e-06, + "loss": 0.6208, + "step": 3072 + }, + { + "epoch": 1.176909821941413, + "grad_norm": 0.298085480928421, + "learning_rate": 3.387680899887132e-06, + "loss": 0.6448, + "step": 3073 + }, + { + "epoch": 1.1772927436339269, + "grad_norm": 0.310481458902359, + "learning_rate": 3.3867313624252823e-06, + "loss": 0.6227, + "step": 3074 + }, + { + "epoch": 1.1776756653264409, + "grad_norm": 0.3187938928604126, + "learning_rate": 3.3857816786124015e-06, + "loss": 0.6309, + "step": 3075 + }, + { + "epoch": 1.1780585870189546, + "grad_norm": 0.29171696305274963, + "learning_rate": 3.3848318486052297e-06, + "loss": 0.6386, + "step": 3076 + }, + { + "epoch": 1.1784415087114686, + "grad_norm": 0.3059884309768677, + "learning_rate": 3.383881872560533e-06, + "loss": 0.6451, + "step": 3077 + }, + { + "epoch": 1.1788244304039823, + "grad_norm": 0.30924007296562195, + "learning_rate": 3.3829317506350994e-06, + "loss": 0.6351, + "step": 3078 + }, + { + "epoch": 1.1792073520964963, + "grad_norm": 0.29961124062538147, + "learning_rate": 3.381981482985744e-06, + "loss": 0.619, + "step": 3079 + }, + { + "epoch": 1.1795902737890103, + "grad_norm": 0.30279725790023804, + "learning_rate": 3.3810310697693027e-06, + "loss": 0.6359, + "step": 3080 + }, + { + "epoch": 1.179973195481524, + "grad_norm": 0.30474141240119934, + "learning_rate": 3.380080511142637e-06, + "loss": 0.6221, + "step": 3081 + }, + { + "epoch": 1.180356117174038, + "grad_norm": 0.3026995062828064, + "learning_rate": 3.3791298072626334e-06, + "loss": 0.641, + "step": 3082 + }, + { + "epoch": 1.1807390388665517, + "grad_norm": 0.2905854880809784, + "learning_rate": 3.3781789582862e-06, + "loss": 0.6403, + "step": 3083 + }, + { + "epoch": 1.1811219605590657, + "grad_norm": 0.2954099476337433, + "learning_rate": 3.3772279643702722e-06, + "loss": 0.6179, + "step": 3084 + }, + { + "epoch": 1.1815048822515795, + "grad_norm": 0.30013203620910645, + "learning_rate": 3.376276825671805e-06, + "loss": 0.6331, + "step": 3085 + }, + { + "epoch": 1.1818878039440934, + "grad_norm": 0.31417974829673767, + "learning_rate": 3.3753255423477805e-06, + "loss": 0.6326, + "step": 3086 + }, + { + "epoch": 1.1822707256366074, + "grad_norm": 0.30636584758758545, + "learning_rate": 3.3743741145552044e-06, + "loss": 0.6415, + "step": 3087 + }, + { + "epoch": 1.1826536473291211, + "grad_norm": 0.3032277524471283, + "learning_rate": 3.373422542451104e-06, + "loss": 0.6342, + "step": 3088 + }, + { + "epoch": 1.1830365690216351, + "grad_norm": 0.30162492394447327, + "learning_rate": 3.372470826192534e-06, + "loss": 0.627, + "step": 3089 + }, + { + "epoch": 1.1834194907141489, + "grad_norm": 0.3018966615200043, + "learning_rate": 3.371518965936569e-06, + "loss": 0.652, + "step": 3090 + }, + { + "epoch": 1.1838024124066628, + "grad_norm": 0.2957569360733032, + "learning_rate": 3.3705669618403108e-06, + "loss": 0.6532, + "step": 3091 + }, + { + "epoch": 1.1841853340991768, + "grad_norm": 0.3041404187679291, + "learning_rate": 3.3696148140608827e-06, + "loss": 0.6456, + "step": 3092 + }, + { + "epoch": 1.1845682557916906, + "grad_norm": 0.3030630350112915, + "learning_rate": 3.3686625227554317e-06, + "loss": 0.6293, + "step": 3093 + }, + { + "epoch": 1.1849511774842045, + "grad_norm": 0.2925577163696289, + "learning_rate": 3.367710088081131e-06, + "loss": 0.6303, + "step": 3094 + }, + { + "epoch": 1.1853340991767183, + "grad_norm": 0.2941893935203552, + "learning_rate": 3.3667575101951733e-06, + "loss": 0.6273, + "step": 3095 + }, + { + "epoch": 1.1857170208692323, + "grad_norm": 0.30883273482322693, + "learning_rate": 3.3658047892547794e-06, + "loss": 0.6441, + "step": 3096 + }, + { + "epoch": 1.186099942561746, + "grad_norm": 0.2927250564098358, + "learning_rate": 3.3648519254171906e-06, + "loss": 0.6453, + "step": 3097 + }, + { + "epoch": 1.18648286425426, + "grad_norm": 0.30045342445373535, + "learning_rate": 3.3638989188396737e-06, + "loss": 0.6572, + "step": 3098 + }, + { + "epoch": 1.186865785946774, + "grad_norm": 0.2903536856174469, + "learning_rate": 3.362945769679517e-06, + "loss": 0.6299, + "step": 3099 + }, + { + "epoch": 1.1872487076392877, + "grad_norm": 0.3260241746902466, + "learning_rate": 3.361992478094034e-06, + "loss": 0.6252, + "step": 3100 + }, + { + "epoch": 1.1876316293318017, + "grad_norm": 0.30903589725494385, + "learning_rate": 3.361039044240561e-06, + "loss": 0.6317, + "step": 3101 + }, + { + "epoch": 1.1880145510243154, + "grad_norm": 0.341140478849411, + "learning_rate": 3.3600854682764585e-06, + "loss": 0.63, + "step": 3102 + }, + { + "epoch": 1.1883974727168294, + "grad_norm": 0.30233532190322876, + "learning_rate": 3.3591317503591093e-06, + "loss": 0.645, + "step": 3103 + }, + { + "epoch": 1.1887803944093434, + "grad_norm": 0.3063500225543976, + "learning_rate": 3.3581778906459206e-06, + "loss": 0.6477, + "step": 3104 + }, + { + "epoch": 1.189163316101857, + "grad_norm": 0.3095088601112366, + "learning_rate": 3.357223889294323e-06, + "loss": 0.6455, + "step": 3105 + }, + { + "epoch": 1.189546237794371, + "grad_norm": 0.3073406517505646, + "learning_rate": 3.3562697464617694e-06, + "loss": 0.6382, + "step": 3106 + }, + { + "epoch": 1.1899291594868848, + "grad_norm": 0.31049203872680664, + "learning_rate": 3.3553154623057372e-06, + "loss": 0.6614, + "step": 3107 + }, + { + "epoch": 1.1903120811793988, + "grad_norm": 0.2952621877193451, + "learning_rate": 3.3543610369837264e-06, + "loss": 0.6323, + "step": 3108 + }, + { + "epoch": 1.1906950028719128, + "grad_norm": 0.31357645988464355, + "learning_rate": 3.3534064706532614e-06, + "loss": 0.643, + "step": 3109 + }, + { + "epoch": 1.1910779245644265, + "grad_norm": 0.3169551193714142, + "learning_rate": 3.3524517634718885e-06, + "loss": 0.63, + "step": 3110 + }, + { + "epoch": 1.1914608462569405, + "grad_norm": 0.2969547510147095, + "learning_rate": 3.3514969155971777e-06, + "loss": 0.6169, + "step": 3111 + }, + { + "epoch": 1.1918437679494542, + "grad_norm": 0.3073836863040924, + "learning_rate": 3.350541927186723e-06, + "loss": 0.6399, + "step": 3112 + }, + { + "epoch": 1.1922266896419682, + "grad_norm": 0.32103368639945984, + "learning_rate": 3.3495867983981396e-06, + "loss": 0.6327, + "step": 3113 + }, + { + "epoch": 1.1926096113344822, + "grad_norm": 0.31980183720588684, + "learning_rate": 3.3486315293890693e-06, + "loss": 0.6184, + "step": 3114 + }, + { + "epoch": 1.192992533026996, + "grad_norm": 0.3071334660053253, + "learning_rate": 3.347676120317173e-06, + "loss": 0.6332, + "step": 3115 + }, + { + "epoch": 1.19337545471951, + "grad_norm": 0.3262181878089905, + "learning_rate": 3.3467205713401378e-06, + "loss": 0.6381, + "step": 3116 + }, + { + "epoch": 1.1937583764120236, + "grad_norm": 0.31094563007354736, + "learning_rate": 3.345764882615672e-06, + "loss": 0.6535, + "step": 3117 + }, + { + "epoch": 1.1941412981045376, + "grad_norm": 0.3051980435848236, + "learning_rate": 3.3448090543015087e-06, + "loss": 0.6428, + "step": 3118 + }, + { + "epoch": 1.1945242197970516, + "grad_norm": 0.3073912560939789, + "learning_rate": 3.3438530865554026e-06, + "loss": 0.6359, + "step": 3119 + }, + { + "epoch": 1.1949071414895653, + "grad_norm": 0.30801281332969666, + "learning_rate": 3.3428969795351312e-06, + "loss": 0.6394, + "step": 3120 + }, + { + "epoch": 1.1952900631820793, + "grad_norm": 0.30035340785980225, + "learning_rate": 3.3419407333984972e-06, + "loss": 0.6296, + "step": 3121 + }, + { + "epoch": 1.195672984874593, + "grad_norm": 0.3098042905330658, + "learning_rate": 3.3409843483033228e-06, + "loss": 0.6476, + "step": 3122 + }, + { + "epoch": 1.196055906567107, + "grad_norm": 0.30431973934173584, + "learning_rate": 3.3400278244074575e-06, + "loss": 0.6019, + "step": 3123 + }, + { + "epoch": 1.196438828259621, + "grad_norm": 0.311336487531662, + "learning_rate": 3.339071161868769e-06, + "loss": 0.6466, + "step": 3124 + }, + { + "epoch": 1.1968217499521348, + "grad_norm": 0.304923415184021, + "learning_rate": 3.3381143608451507e-06, + "loss": 0.6369, + "step": 3125 + }, + { + "epoch": 1.1972046716446487, + "grad_norm": 0.31816065311431885, + "learning_rate": 3.3371574214945195e-06, + "loss": 0.6342, + "step": 3126 + }, + { + "epoch": 1.1975875933371625, + "grad_norm": 0.30785971879959106, + "learning_rate": 3.3362003439748125e-06, + "loss": 0.6426, + "step": 3127 + }, + { + "epoch": 1.1979705150296764, + "grad_norm": 0.3149166703224182, + "learning_rate": 3.3352431284439922e-06, + "loss": 0.6401, + "step": 3128 + }, + { + "epoch": 1.1983534367221904, + "grad_norm": 0.3231953978538513, + "learning_rate": 3.3342857750600417e-06, + "loss": 0.6134, + "step": 3129 + }, + { + "epoch": 1.1987363584147042, + "grad_norm": 0.3046688139438629, + "learning_rate": 3.3333282839809685e-06, + "loss": 0.6444, + "step": 3130 + }, + { + "epoch": 1.1991192801072181, + "grad_norm": 0.30938521027565, + "learning_rate": 3.3323706553648023e-06, + "loss": 0.6611, + "step": 3131 + }, + { + "epoch": 1.1995022017997319, + "grad_norm": 0.3077024519443512, + "learning_rate": 3.3314128893695944e-06, + "loss": 0.6375, + "step": 3132 + }, + { + "epoch": 1.1998851234922459, + "grad_norm": 0.31264743208885193, + "learning_rate": 3.330454986153422e-06, + "loss": 0.6472, + "step": 3133 + }, + { + "epoch": 1.2002680451847598, + "grad_norm": 0.32338380813598633, + "learning_rate": 3.32949694587438e-06, + "loss": 0.6459, + "step": 3134 + }, + { + "epoch": 1.2006509668772736, + "grad_norm": 0.3111104369163513, + "learning_rate": 3.3285387686905906e-06, + "loss": 0.6191, + "step": 3135 + }, + { + "epoch": 1.2010338885697875, + "grad_norm": 0.3032006025314331, + "learning_rate": 3.3275804547601953e-06, + "loss": 0.6272, + "step": 3136 + }, + { + "epoch": 1.2014168102623013, + "grad_norm": 0.3343944847583771, + "learning_rate": 3.326622004241361e-06, + "loss": 0.638, + "step": 3137 + }, + { + "epoch": 1.2017997319548153, + "grad_norm": 0.3289186656475067, + "learning_rate": 3.3256634172922746e-06, + "loss": 0.6295, + "step": 3138 + }, + { + "epoch": 1.2021826536473292, + "grad_norm": 0.3102615773677826, + "learning_rate": 3.324704694071147e-06, + "loss": 0.6428, + "step": 3139 + }, + { + "epoch": 1.202565575339843, + "grad_norm": 0.3457825481891632, + "learning_rate": 3.3237458347362106e-06, + "loss": 0.6409, + "step": 3140 + }, + { + "epoch": 1.202948497032357, + "grad_norm": 0.30256426334381104, + "learning_rate": 3.3227868394457214e-06, + "loss": 0.6332, + "step": 3141 + }, + { + "epoch": 1.2033314187248707, + "grad_norm": 0.3069865107536316, + "learning_rate": 3.3218277083579564e-06, + "loss": 0.6522, + "step": 3142 + }, + { + "epoch": 1.2037143404173847, + "grad_norm": 0.31416645646095276, + "learning_rate": 3.3208684416312174e-06, + "loss": 0.6352, + "step": 3143 + }, + { + "epoch": 1.2040972621098986, + "grad_norm": 0.3245389461517334, + "learning_rate": 3.3199090394238265e-06, + "loss": 0.6702, + "step": 3144 + }, + { + "epoch": 1.2044801838024124, + "grad_norm": 0.3271872401237488, + "learning_rate": 3.318949501894128e-06, + "loss": 0.6174, + "step": 3145 + }, + { + "epoch": 1.2048631054949264, + "grad_norm": 0.30836963653564453, + "learning_rate": 3.31798982920049e-06, + "loss": 0.64, + "step": 3146 + }, + { + "epoch": 1.2052460271874401, + "grad_norm": 0.30794307589530945, + "learning_rate": 3.3170300215013013e-06, + "loss": 0.643, + "step": 3147 + }, + { + "epoch": 1.205628948879954, + "grad_norm": 0.3118075430393219, + "learning_rate": 3.316070078954975e-06, + "loss": 0.6315, + "step": 3148 + }, + { + "epoch": 1.2060118705724678, + "grad_norm": 0.30457648634910583, + "learning_rate": 3.3151100017199437e-06, + "loss": 0.624, + "step": 3149 + }, + { + "epoch": 1.2063947922649818, + "grad_norm": 0.30681130290031433, + "learning_rate": 3.314149789954666e-06, + "loss": 0.6271, + "step": 3150 + }, + { + "epoch": 1.2067777139574958, + "grad_norm": 0.3112008571624756, + "learning_rate": 3.313189443817619e-06, + "loss": 0.6285, + "step": 3151 + }, + { + "epoch": 1.2071606356500095, + "grad_norm": 0.30933505296707153, + "learning_rate": 3.3122289634673042e-06, + "loss": 0.6453, + "step": 3152 + }, + { + "epoch": 1.2075435573425235, + "grad_norm": 0.30465900897979736, + "learning_rate": 3.3112683490622445e-06, + "loss": 0.6256, + "step": 3153 + }, + { + "epoch": 1.2079264790350372, + "grad_norm": 0.2949281632900238, + "learning_rate": 3.3103076007609845e-06, + "loss": 0.6633, + "step": 3154 + }, + { + "epoch": 1.2083094007275512, + "grad_norm": 0.31145644187927246, + "learning_rate": 3.3093467187220923e-06, + "loss": 0.6332, + "step": 3155 + }, + { + "epoch": 1.2086923224200652, + "grad_norm": 0.31175538897514343, + "learning_rate": 3.3083857031041557e-06, + "loss": 0.6553, + "step": 3156 + }, + { + "epoch": 1.209075244112579, + "grad_norm": 0.30894431471824646, + "learning_rate": 3.307424554065787e-06, + "loss": 0.624, + "step": 3157 + }, + { + "epoch": 1.209458165805093, + "grad_norm": 0.30649998784065247, + "learning_rate": 3.3064632717656203e-06, + "loss": 0.6463, + "step": 3158 + }, + { + "epoch": 1.2098410874976067, + "grad_norm": 0.3122660219669342, + "learning_rate": 3.3055018563623085e-06, + "loss": 0.649, + "step": 3159 + }, + { + "epoch": 1.2102240091901206, + "grad_norm": 0.300945520401001, + "learning_rate": 3.304540308014532e-06, + "loss": 0.6165, + "step": 3160 + }, + { + "epoch": 1.2106069308826344, + "grad_norm": 0.2981884181499481, + "learning_rate": 3.303578626880988e-06, + "loss": 0.6388, + "step": 3161 + }, + { + "epoch": 1.2109898525751484, + "grad_norm": 0.3044055700302124, + "learning_rate": 3.3026168131203984e-06, + "loss": 0.6406, + "step": 3162 + }, + { + "epoch": 1.2113727742676623, + "grad_norm": 0.3004518449306488, + "learning_rate": 3.3016548668915056e-06, + "loss": 0.6288, + "step": 3163 + }, + { + "epoch": 1.211755695960176, + "grad_norm": 0.3121183514595032, + "learning_rate": 3.3006927883530748e-06, + "loss": 0.6338, + "step": 3164 + }, + { + "epoch": 1.21213861765269, + "grad_norm": 0.29893359541893005, + "learning_rate": 3.299730577663893e-06, + "loss": 0.6404, + "step": 3165 + }, + { + "epoch": 1.2125215393452038, + "grad_norm": 0.29656678438186646, + "learning_rate": 3.298768234982769e-06, + "loss": 0.6574, + "step": 3166 + }, + { + "epoch": 1.2129044610377178, + "grad_norm": 0.3049081563949585, + "learning_rate": 3.2978057604685326e-06, + "loss": 0.6424, + "step": 3167 + }, + { + "epoch": 1.2132873827302317, + "grad_norm": 0.29937291145324707, + "learning_rate": 3.2968431542800354e-06, + "loss": 0.6286, + "step": 3168 + }, + { + "epoch": 1.2136703044227455, + "grad_norm": 0.30627888441085815, + "learning_rate": 3.295880416576153e-06, + "loss": 0.6406, + "step": 3169 + }, + { + "epoch": 1.2140532261152595, + "grad_norm": 0.29770469665527344, + "learning_rate": 3.294917547515779e-06, + "loss": 0.6185, + "step": 3170 + }, + { + "epoch": 1.2144361478077732, + "grad_norm": 0.31077998876571655, + "learning_rate": 3.2939545472578314e-06, + "loss": 0.6184, + "step": 3171 + }, + { + "epoch": 1.2148190695002872, + "grad_norm": 0.305830180644989, + "learning_rate": 3.2929914159612498e-06, + "loss": 0.6478, + "step": 3172 + }, + { + "epoch": 1.2152019911928011, + "grad_norm": 0.29923877120018005, + "learning_rate": 3.2920281537849937e-06, + "loss": 0.6367, + "step": 3173 + }, + { + "epoch": 1.215584912885315, + "grad_norm": 0.30622598528862, + "learning_rate": 3.2910647608880452e-06, + "loss": 0.6585, + "step": 3174 + }, + { + "epoch": 1.2159678345778289, + "grad_norm": 0.29234644770622253, + "learning_rate": 3.2901012374294082e-06, + "loss": 0.6359, + "step": 3175 + }, + { + "epoch": 1.2163507562703426, + "grad_norm": 0.3031441271305084, + "learning_rate": 3.2891375835681084e-06, + "loss": 0.6392, + "step": 3176 + }, + { + "epoch": 1.2167336779628566, + "grad_norm": 0.31722012162208557, + "learning_rate": 3.2881737994631924e-06, + "loss": 0.6273, + "step": 3177 + }, + { + "epoch": 1.2171165996553706, + "grad_norm": 0.30615052580833435, + "learning_rate": 3.2872098852737274e-06, + "loss": 0.6079, + "step": 3178 + }, + { + "epoch": 1.2174995213478843, + "grad_norm": 0.2979710102081299, + "learning_rate": 3.286245841158804e-06, + "loss": 0.6312, + "step": 3179 + }, + { + "epoch": 1.2178824430403983, + "grad_norm": 0.2990206778049469, + "learning_rate": 3.285281667277534e-06, + "loss": 0.6352, + "step": 3180 + }, + { + "epoch": 1.218265364732912, + "grad_norm": 0.3113645017147064, + "learning_rate": 3.284317363789048e-06, + "loss": 0.6452, + "step": 3181 + }, + { + "epoch": 1.218648286425426, + "grad_norm": 0.30548733472824097, + "learning_rate": 3.2833529308525026e-06, + "loss": 0.6325, + "step": 3182 + }, + { + "epoch": 1.21903120811794, + "grad_norm": 0.2985915541648865, + "learning_rate": 3.2823883686270706e-06, + "loss": 0.6409, + "step": 3183 + }, + { + "epoch": 1.2194141298104537, + "grad_norm": 0.3036695122718811, + "learning_rate": 3.281423677271951e-06, + "loss": 0.641, + "step": 3184 + }, + { + "epoch": 1.2197970515029677, + "grad_norm": 0.30887073278427124, + "learning_rate": 3.280458856946359e-06, + "loss": 0.6213, + "step": 3185 + }, + { + "epoch": 1.2201799731954814, + "grad_norm": 0.30206504464149475, + "learning_rate": 3.2794939078095366e-06, + "loss": 0.6397, + "step": 3186 + }, + { + "epoch": 1.2205628948879954, + "grad_norm": 0.3033526837825775, + "learning_rate": 3.2785288300207428e-06, + "loss": 0.6563, + "step": 3187 + }, + { + "epoch": 1.2209458165805094, + "grad_norm": 0.34493187069892883, + "learning_rate": 3.2775636237392595e-06, + "loss": 0.6416, + "step": 3188 + }, + { + "epoch": 1.2213287382730231, + "grad_norm": 0.3178871273994446, + "learning_rate": 3.2765982891243903e-06, + "loss": 0.6369, + "step": 3189 + }, + { + "epoch": 1.221711659965537, + "grad_norm": 0.30893343687057495, + "learning_rate": 3.275632826335459e-06, + "loss": 0.6152, + "step": 3190 + }, + { + "epoch": 1.2220945816580508, + "grad_norm": 0.31709998846054077, + "learning_rate": 3.27466723553181e-06, + "loss": 0.6477, + "step": 3191 + }, + { + "epoch": 1.2224775033505648, + "grad_norm": 0.3091951906681061, + "learning_rate": 3.273701516872811e-06, + "loss": 0.6505, + "step": 3192 + }, + { + "epoch": 1.2228604250430788, + "grad_norm": 0.308051735162735, + "learning_rate": 3.2727356705178493e-06, + "loss": 0.6165, + "step": 3193 + }, + { + "epoch": 1.2232433467355925, + "grad_norm": 0.29437336325645447, + "learning_rate": 3.2717696966263334e-06, + "loss": 0.6524, + "step": 3194 + }, + { + "epoch": 1.2236262684281065, + "grad_norm": 0.3080149292945862, + "learning_rate": 3.2708035953576924e-06, + "loss": 0.6419, + "step": 3195 + }, + { + "epoch": 1.2240091901206203, + "grad_norm": 0.3170960247516632, + "learning_rate": 3.2698373668713773e-06, + "loss": 0.6309, + "step": 3196 + }, + { + "epoch": 1.2243921118131342, + "grad_norm": 0.3092925548553467, + "learning_rate": 3.268871011326861e-06, + "loss": 0.6477, + "step": 3197 + }, + { + "epoch": 1.2247750335056482, + "grad_norm": 0.31064000725746155, + "learning_rate": 3.2679045288836332e-06, + "loss": 0.6531, + "step": 3198 + }, + { + "epoch": 1.225157955198162, + "grad_norm": 0.30165842175483704, + "learning_rate": 3.2669379197012107e-06, + "loss": 0.6484, + "step": 3199 + }, + { + "epoch": 1.225540876890676, + "grad_norm": 0.30968889594078064, + "learning_rate": 3.265971183939126e-06, + "loss": 0.6367, + "step": 3200 + }, + { + "epoch": 1.2259237985831897, + "grad_norm": 0.2960895895957947, + "learning_rate": 3.2650043217569353e-06, + "loss": 0.6206, + "step": 3201 + }, + { + "epoch": 1.2263067202757036, + "grad_norm": 0.30400365591049194, + "learning_rate": 3.264037333314215e-06, + "loss": 0.6495, + "step": 3202 + }, + { + "epoch": 1.2266896419682176, + "grad_norm": 0.2982025146484375, + "learning_rate": 3.2630702187705614e-06, + "loss": 0.6394, + "step": 3203 + }, + { + "epoch": 1.2270725636607314, + "grad_norm": 0.30492743849754333, + "learning_rate": 3.2621029782855933e-06, + "loss": 0.6649, + "step": 3204 + }, + { + "epoch": 1.2274554853532453, + "grad_norm": 0.3069272041320801, + "learning_rate": 3.2611356120189486e-06, + "loss": 0.6408, + "step": 3205 + }, + { + "epoch": 1.227838407045759, + "grad_norm": 0.31136906147003174, + "learning_rate": 3.2601681201302878e-06, + "loss": 0.6195, + "step": 3206 + }, + { + "epoch": 1.228221328738273, + "grad_norm": 0.29804372787475586, + "learning_rate": 3.2592005027792894e-06, + "loss": 0.6374, + "step": 3207 + }, + { + "epoch": 1.228604250430787, + "grad_norm": 0.29439473152160645, + "learning_rate": 3.2582327601256567e-06, + "loss": 0.6476, + "step": 3208 + }, + { + "epoch": 1.2289871721233008, + "grad_norm": 0.304158478975296, + "learning_rate": 3.2572648923291094e-06, + "loss": 0.6344, + "step": 3209 + }, + { + "epoch": 1.2293700938158147, + "grad_norm": 0.293686181306839, + "learning_rate": 3.2562968995493897e-06, + "loss": 0.6109, + "step": 3210 + }, + { + "epoch": 1.2297530155083285, + "grad_norm": 0.3041535019874573, + "learning_rate": 3.2553287819462625e-06, + "loss": 0.6554, + "step": 3211 + }, + { + "epoch": 1.2301359372008425, + "grad_norm": 0.3087673485279083, + "learning_rate": 3.2543605396795087e-06, + "loss": 0.6416, + "step": 3212 + }, + { + "epoch": 1.2305188588933562, + "grad_norm": 0.2950975000858307, + "learning_rate": 3.2533921729089333e-06, + "loss": 0.6225, + "step": 3213 + }, + { + "epoch": 1.2309017805858702, + "grad_norm": 0.2994769513607025, + "learning_rate": 3.2524236817943617e-06, + "loss": 0.6236, + "step": 3214 + }, + { + "epoch": 1.2312847022783842, + "grad_norm": 0.3098432123661041, + "learning_rate": 3.251455066495638e-06, + "loss": 0.6093, + "step": 3215 + }, + { + "epoch": 1.231667623970898, + "grad_norm": 0.3064119219779968, + "learning_rate": 3.2504863271726284e-06, + "loss": 0.6498, + "step": 3216 + }, + { + "epoch": 1.2320505456634119, + "grad_norm": 0.3000611364841461, + "learning_rate": 3.2495174639852182e-06, + "loss": 0.6125, + "step": 3217 + }, + { + "epoch": 1.2324334673559256, + "grad_norm": 0.31257539987564087, + "learning_rate": 3.248548477093316e-06, + "loss": 0.6331, + "step": 3218 + }, + { + "epoch": 1.2328163890484396, + "grad_norm": 0.31372538208961487, + "learning_rate": 3.247579366656845e-06, + "loss": 0.6479, + "step": 3219 + }, + { + "epoch": 1.2331993107409536, + "grad_norm": 0.30366671085357666, + "learning_rate": 3.246610132835756e-06, + "loss": 0.6328, + "step": 3220 + }, + { + "epoch": 1.2335822324334673, + "grad_norm": 0.3083849549293518, + "learning_rate": 3.245640775790015e-06, + "loss": 0.6277, + "step": 3221 + }, + { + "epoch": 1.2339651541259813, + "grad_norm": 0.3108576834201813, + "learning_rate": 3.244671295679611e-06, + "loss": 0.6261, + "step": 3222 + }, + { + "epoch": 1.234348075818495, + "grad_norm": 0.30852648615837097, + "learning_rate": 3.2437016926645513e-06, + "loss": 0.6242, + "step": 3223 + }, + { + "epoch": 1.234730997511009, + "grad_norm": 0.29925426840782166, + "learning_rate": 3.242731966904865e-06, + "loss": 0.647, + "step": 3224 + }, + { + "epoch": 1.2351139192035228, + "grad_norm": 0.31132084131240845, + "learning_rate": 3.241762118560601e-06, + "loss": 0.6161, + "step": 3225 + }, + { + "epoch": 1.2354968408960367, + "grad_norm": 0.297787070274353, + "learning_rate": 3.2407921477918292e-06, + "loss": 0.622, + "step": 3226 + }, + { + "epoch": 1.2358797625885507, + "grad_norm": 0.314738005399704, + "learning_rate": 3.2398220547586366e-06, + "loss": 0.6371, + "step": 3227 + }, + { + "epoch": 1.2362626842810645, + "grad_norm": 0.32148897647857666, + "learning_rate": 3.2388518396211355e-06, + "loss": 0.6484, + "step": 3228 + }, + { + "epoch": 1.2366456059735784, + "grad_norm": 0.3126406967639923, + "learning_rate": 3.2378815025394545e-06, + "loss": 0.6288, + "step": 3229 + }, + { + "epoch": 1.2370285276660922, + "grad_norm": 0.2967692017555237, + "learning_rate": 3.236911043673743e-06, + "loss": 0.6404, + "step": 3230 + }, + { + "epoch": 1.2374114493586061, + "grad_norm": 0.310128390789032, + "learning_rate": 3.23594046318417e-06, + "loss": 0.6497, + "step": 3231 + }, + { + "epoch": 1.2377943710511201, + "grad_norm": 0.3275293707847595, + "learning_rate": 3.2349697612309273e-06, + "loss": 0.6414, + "step": 3232 + }, + { + "epoch": 1.2381772927436339, + "grad_norm": 0.3017575144767761, + "learning_rate": 3.2339989379742244e-06, + "loss": 0.632, + "step": 3233 + }, + { + "epoch": 1.2385602144361478, + "grad_norm": 0.30867525935173035, + "learning_rate": 3.2330279935742905e-06, + "loss": 0.6298, + "step": 3234 + }, + { + "epoch": 1.2389431361286616, + "grad_norm": 0.3130476176738739, + "learning_rate": 3.232056928191376e-06, + "loss": 0.632, + "step": 3235 + }, + { + "epoch": 1.2393260578211756, + "grad_norm": 0.3017469644546509, + "learning_rate": 3.231085741985752e-06, + "loss": 0.6482, + "step": 3236 + }, + { + "epoch": 1.2397089795136895, + "grad_norm": 0.30170688033103943, + "learning_rate": 3.230114435117706e-06, + "loss": 0.6298, + "step": 3237 + }, + { + "epoch": 1.2400919012062033, + "grad_norm": 0.31267961859703064, + "learning_rate": 3.2291430077475506e-06, + "loss": 0.635, + "step": 3238 + }, + { + "epoch": 1.2404748228987172, + "grad_norm": 0.2961098849773407, + "learning_rate": 3.228171460035614e-06, + "loss": 0.6353, + "step": 3239 + }, + { + "epoch": 1.240857744591231, + "grad_norm": 0.3017948269844055, + "learning_rate": 3.227199792142246e-06, + "loss": 0.6441, + "step": 3240 + }, + { + "epoch": 1.241240666283745, + "grad_norm": 0.3013845384120941, + "learning_rate": 3.2262280042278155e-06, + "loss": 0.6269, + "step": 3241 + }, + { + "epoch": 1.241623587976259, + "grad_norm": 0.2992444634437561, + "learning_rate": 3.2252560964527123e-06, + "loss": 0.6483, + "step": 3242 + }, + { + "epoch": 1.2420065096687727, + "grad_norm": 0.3005957007408142, + "learning_rate": 3.224284068977346e-06, + "loss": 0.6192, + "step": 3243 + }, + { + "epoch": 1.2423894313612867, + "grad_norm": 0.29909175634384155, + "learning_rate": 3.223311921962144e-06, + "loss": 0.6172, + "step": 3244 + }, + { + "epoch": 1.2427723530538004, + "grad_norm": 0.30320921540260315, + "learning_rate": 3.222339655567556e-06, + "loss": 0.6229, + "step": 3245 + }, + { + "epoch": 1.2431552747463144, + "grad_norm": 0.3046598434448242, + "learning_rate": 3.22136726995405e-06, + "loss": 0.629, + "step": 3246 + }, + { + "epoch": 1.2435381964388283, + "grad_norm": 0.30688151717185974, + "learning_rate": 3.2203947652821133e-06, + "loss": 0.6559, + "step": 3247 + }, + { + "epoch": 1.243921118131342, + "grad_norm": 0.30767685174942017, + "learning_rate": 3.219422141712254e-06, + "loss": 0.6471, + "step": 3248 + }, + { + "epoch": 1.244304039823856, + "grad_norm": 0.3162931501865387, + "learning_rate": 3.2184493994049988e-06, + "loss": 0.6288, + "step": 3249 + }, + { + "epoch": 1.2446869615163698, + "grad_norm": 0.3018272817134857, + "learning_rate": 3.217476538520895e-06, + "loss": 0.6221, + "step": 3250 + }, + { + "epoch": 1.2450698832088838, + "grad_norm": 0.2930925190448761, + "learning_rate": 3.2165035592205084e-06, + "loss": 0.6567, + "step": 3251 + }, + { + "epoch": 1.2454528049013978, + "grad_norm": 0.30316081643104553, + "learning_rate": 3.215530461664425e-06, + "loss": 0.6314, + "step": 3252 + }, + { + "epoch": 1.2458357265939115, + "grad_norm": 0.29480230808258057, + "learning_rate": 3.21455724601325e-06, + "loss": 0.6299, + "step": 3253 + }, + { + "epoch": 1.2462186482864255, + "grad_norm": 0.3178538978099823, + "learning_rate": 3.213583912427609e-06, + "loss": 0.6254, + "step": 3254 + }, + { + "epoch": 1.2466015699789392, + "grad_norm": 0.2982751131057739, + "learning_rate": 3.2126104610681464e-06, + "loss": 0.6442, + "step": 3255 + }, + { + "epoch": 1.2469844916714532, + "grad_norm": 0.2966442108154297, + "learning_rate": 3.2116368920955244e-06, + "loss": 0.6648, + "step": 3256 + }, + { + "epoch": 1.2473674133639672, + "grad_norm": 0.3126610815525055, + "learning_rate": 3.2106632056704277e-06, + "loss": 0.6423, + "step": 3257 + }, + { + "epoch": 1.247750335056481, + "grad_norm": 0.31517601013183594, + "learning_rate": 3.2096894019535583e-06, + "loss": 0.638, + "step": 3258 + }, + { + "epoch": 1.248133256748995, + "grad_norm": 0.3176659643650055, + "learning_rate": 3.208715481105638e-06, + "loss": 0.6219, + "step": 3259 + }, + { + "epoch": 1.2485161784415086, + "grad_norm": 0.3133630156517029, + "learning_rate": 3.2077414432874088e-06, + "loss": 0.6297, + "step": 3260 + }, + { + "epoch": 1.2488991001340226, + "grad_norm": 0.3311651349067688, + "learning_rate": 3.2067672886596314e-06, + "loss": 0.639, + "step": 3261 + }, + { + "epoch": 1.2492820218265366, + "grad_norm": 0.3203081488609314, + "learning_rate": 3.205793017383084e-06, + "loss": 0.6337, + "step": 3262 + }, + { + "epoch": 1.2496649435190503, + "grad_norm": 0.3108779489994049, + "learning_rate": 3.204818629618567e-06, + "loss": 0.6278, + "step": 3263 + }, + { + "epoch": 1.2500478652115643, + "grad_norm": 0.30094388127326965, + "learning_rate": 3.203844125526898e-06, + "loss": 0.6291, + "step": 3264 + }, + { + "epoch": 1.250430786904078, + "grad_norm": 0.3072316348552704, + "learning_rate": 3.202869505268916e-06, + "loss": 0.6364, + "step": 3265 + }, + { + "epoch": 1.250813708596592, + "grad_norm": 0.29571276903152466, + "learning_rate": 3.201894769005476e-06, + "loss": 0.6242, + "step": 3266 + }, + { + "epoch": 1.251196630289106, + "grad_norm": 0.3029182553291321, + "learning_rate": 3.2009199168974547e-06, + "loss": 0.6283, + "step": 3267 + }, + { + "epoch": 1.2515795519816197, + "grad_norm": 0.30267104506492615, + "learning_rate": 3.1999449491057474e-06, + "loss": 0.6416, + "step": 3268 + }, + { + "epoch": 1.2519624736741337, + "grad_norm": 0.29282158613204956, + "learning_rate": 3.1989698657912673e-06, + "loss": 0.6203, + "step": 3269 + }, + { + "epoch": 1.2523453953666475, + "grad_norm": 0.3109108507633209, + "learning_rate": 3.197994667114947e-06, + "loss": 0.6507, + "step": 3270 + }, + { + "epoch": 1.2527283170591614, + "grad_norm": 0.30930864810943604, + "learning_rate": 3.1970193532377404e-06, + "loss": 0.6499, + "step": 3271 + }, + { + "epoch": 1.2531112387516754, + "grad_norm": 0.30899038910865784, + "learning_rate": 3.1960439243206177e-06, + "loss": 0.644, + "step": 3272 + }, + { + "epoch": 1.2534941604441892, + "grad_norm": 0.303649365901947, + "learning_rate": 3.195068380524569e-06, + "loss": 0.6342, + "step": 3273 + }, + { + "epoch": 1.2538770821367031, + "grad_norm": 0.30613043904304504, + "learning_rate": 3.1940927220106045e-06, + "loss": 0.6108, + "step": 3274 + }, + { + "epoch": 1.2542600038292169, + "grad_norm": 0.3070712685585022, + "learning_rate": 3.1931169489397505e-06, + "loss": 0.6474, + "step": 3275 + }, + { + "epoch": 1.2546429255217308, + "grad_norm": 0.3077320158481598, + "learning_rate": 3.192141061473055e-06, + "loss": 0.63, + "step": 3276 + }, + { + "epoch": 1.2550258472142448, + "grad_norm": 0.29825252294540405, + "learning_rate": 3.191165059771584e-06, + "loss": 0.6285, + "step": 3277 + }, + { + "epoch": 1.2554087689067586, + "grad_norm": 0.29719796776771545, + "learning_rate": 3.1901889439964214e-06, + "loss": 0.629, + "step": 3278 + }, + { + "epoch": 1.2557916905992723, + "grad_norm": 0.2951110601425171, + "learning_rate": 3.1892127143086716e-06, + "loss": 0.6108, + "step": 3279 + }, + { + "epoch": 1.2561746122917863, + "grad_norm": 0.2983551025390625, + "learning_rate": 3.1882363708694563e-06, + "loss": 0.652, + "step": 3280 + }, + { + "epoch": 1.2565575339843003, + "grad_norm": 0.2987157702445984, + "learning_rate": 3.1872599138399167e-06, + "loss": 0.6393, + "step": 3281 + }, + { + "epoch": 1.2569404556768142, + "grad_norm": 0.3053284287452698, + "learning_rate": 3.1862833433812137e-06, + "loss": 0.632, + "step": 3282 + }, + { + "epoch": 1.257323377369328, + "grad_norm": 0.30585920810699463, + "learning_rate": 3.1853066596545235e-06, + "loss": 0.639, + "step": 3283 + }, + { + "epoch": 1.2577062990618417, + "grad_norm": 0.29770398139953613, + "learning_rate": 3.1843298628210463e-06, + "loss": 0.6397, + "step": 3284 + }, + { + "epoch": 1.2580892207543557, + "grad_norm": 0.29046571254730225, + "learning_rate": 3.183352953041996e-06, + "loss": 0.6327, + "step": 3285 + }, + { + "epoch": 1.2584721424468697, + "grad_norm": 0.28972792625427246, + "learning_rate": 3.1823759304786074e-06, + "loss": 0.6173, + "step": 3286 + }, + { + "epoch": 1.2588550641393834, + "grad_norm": 0.30661410093307495, + "learning_rate": 3.181398795292134e-06, + "loss": 0.6351, + "step": 3287 + }, + { + "epoch": 1.2592379858318974, + "grad_norm": 0.3089430332183838, + "learning_rate": 3.1804215476438475e-06, + "loss": 0.6456, + "step": 3288 + }, + { + "epoch": 1.2596209075244111, + "grad_norm": 0.3143659830093384, + "learning_rate": 3.17944418769504e-06, + "loss": 0.6393, + "step": 3289 + }, + { + "epoch": 1.260003829216925, + "grad_norm": 0.2929110527038574, + "learning_rate": 3.178466715607017e-06, + "loss": 0.6324, + "step": 3290 + }, + { + "epoch": 1.260386750909439, + "grad_norm": 0.3113831877708435, + "learning_rate": 3.177489131541108e-06, + "loss": 0.6461, + "step": 3291 + }, + { + "epoch": 1.2607696726019528, + "grad_norm": 0.2998027205467224, + "learning_rate": 3.1765114356586592e-06, + "loss": 0.6504, + "step": 3292 + }, + { + "epoch": 1.2611525942944668, + "grad_norm": 0.2909545302391052, + "learning_rate": 3.1755336281210325e-06, + "loss": 0.6366, + "step": 3293 + }, + { + "epoch": 1.2615355159869805, + "grad_norm": 0.3028958737850189, + "learning_rate": 3.1745557090896138e-06, + "loss": 0.6374, + "step": 3294 + }, + { + "epoch": 1.2619184376794945, + "grad_norm": 0.30873608589172363, + "learning_rate": 3.1735776787258023e-06, + "loss": 0.6473, + "step": 3295 + }, + { + "epoch": 1.2623013593720085, + "grad_norm": 0.30712270736694336, + "learning_rate": 3.1725995371910185e-06, + "loss": 0.6384, + "step": 3296 + }, + { + "epoch": 1.2626842810645222, + "grad_norm": 0.2883552312850952, + "learning_rate": 3.1716212846466986e-06, + "loss": 0.6336, + "step": 3297 + }, + { + "epoch": 1.2630672027570362, + "grad_norm": 0.30539336800575256, + "learning_rate": 3.1706429212543004e-06, + "loss": 0.6497, + "step": 3298 + }, + { + "epoch": 1.26345012444955, + "grad_norm": 0.3000241816043854, + "learning_rate": 3.1696644471752983e-06, + "loss": 0.6449, + "step": 3299 + }, + { + "epoch": 1.263833046142064, + "grad_norm": 0.31833916902542114, + "learning_rate": 3.168685862571185e-06, + "loss": 0.6593, + "step": 3300 + }, + { + "epoch": 1.264215967834578, + "grad_norm": 0.3133701682090759, + "learning_rate": 3.1677071676034708e-06, + "loss": 0.6382, + "step": 3301 + }, + { + "epoch": 1.2645988895270917, + "grad_norm": 0.3049345314502716, + "learning_rate": 3.1667283624336854e-06, + "loss": 0.6271, + "step": 3302 + }, + { + "epoch": 1.2649818112196056, + "grad_norm": 0.2950793504714966, + "learning_rate": 3.1657494472233765e-06, + "loss": 0.6463, + "step": 3303 + }, + { + "epoch": 1.2653647329121194, + "grad_norm": 0.3062182366847992, + "learning_rate": 3.1647704221341084e-06, + "loss": 0.6221, + "step": 3304 + }, + { + "epoch": 1.2657476546046333, + "grad_norm": 0.31563371419906616, + "learning_rate": 3.163791287327466e-06, + "loss": 0.6315, + "step": 3305 + }, + { + "epoch": 1.2661305762971473, + "grad_norm": 0.2946052849292755, + "learning_rate": 3.162812042965051e-06, + "loss": 0.638, + "step": 3306 + }, + { + "epoch": 1.266513497989661, + "grad_norm": 0.29777106642723083, + "learning_rate": 3.1618326892084826e-06, + "loss": 0.6464, + "step": 3307 + }, + { + "epoch": 1.266896419682175, + "grad_norm": 0.30252063274383545, + "learning_rate": 3.1608532262194e-06, + "loss": 0.625, + "step": 3308 + }, + { + "epoch": 1.2672793413746888, + "grad_norm": 0.3034707307815552, + "learning_rate": 3.1598736541594577e-06, + "loss": 0.6376, + "step": 3309 + }, + { + "epoch": 1.2676622630672028, + "grad_norm": 0.3028099238872528, + "learning_rate": 3.158893973190331e-06, + "loss": 0.6468, + "step": 3310 + }, + { + "epoch": 1.2680451847597167, + "grad_norm": 0.30121949315071106, + "learning_rate": 3.1579141834737106e-06, + "loss": 0.6349, + "step": 3311 + }, + { + "epoch": 1.2684281064522305, + "grad_norm": 0.3000284433364868, + "learning_rate": 3.156934285171307e-06, + "loss": 0.6444, + "step": 3312 + }, + { + "epoch": 1.2688110281447444, + "grad_norm": 0.3178223669528961, + "learning_rate": 3.1559542784448485e-06, + "loss": 0.6416, + "step": 3313 + }, + { + "epoch": 1.2691939498372582, + "grad_norm": 0.3054428994655609, + "learning_rate": 3.1549741634560792e-06, + "loss": 0.6312, + "step": 3314 + }, + { + "epoch": 1.2695768715297722, + "grad_norm": 0.301340788602829, + "learning_rate": 3.1539939403667647e-06, + "loss": 0.6297, + "step": 3315 + }, + { + "epoch": 1.2699597932222861, + "grad_norm": 0.2968747615814209, + "learning_rate": 3.1530136093386855e-06, + "loss": 0.6364, + "step": 3316 + }, + { + "epoch": 1.2703427149147999, + "grad_norm": 0.3132340610027313, + "learning_rate": 3.152033170533641e-06, + "loss": 0.6198, + "step": 3317 + }, + { + "epoch": 1.2707256366073139, + "grad_norm": 0.3089751899242401, + "learning_rate": 3.1510526241134477e-06, + "loss": 0.6563, + "step": 3318 + }, + { + "epoch": 1.2711085582998276, + "grad_norm": 0.3130260407924652, + "learning_rate": 3.1500719702399406e-06, + "loss": 0.6373, + "step": 3319 + }, + { + "epoch": 1.2714914799923416, + "grad_norm": 0.3054065406322479, + "learning_rate": 3.1490912090749733e-06, + "loss": 0.6286, + "step": 3320 + }, + { + "epoch": 1.2718744016848555, + "grad_norm": 0.3106709122657776, + "learning_rate": 3.148110340780414e-06, + "loss": 0.6293, + "step": 3321 + }, + { + "epoch": 1.2722573233773693, + "grad_norm": 0.31249696016311646, + "learning_rate": 3.147129365518152e-06, + "loss": 0.6303, + "step": 3322 + }, + { + "epoch": 1.2726402450698833, + "grad_norm": 0.3025418817996979, + "learning_rate": 3.1461482834500936e-06, + "loss": 0.6275, + "step": 3323 + }, + { + "epoch": 1.273023166762397, + "grad_norm": 0.29861536622047424, + "learning_rate": 3.14516709473816e-06, + "loss": 0.6294, + "step": 3324 + }, + { + "epoch": 1.273406088454911, + "grad_norm": 0.3068664073944092, + "learning_rate": 3.1441857995442925e-06, + "loss": 0.6034, + "step": 3325 + }, + { + "epoch": 1.273789010147425, + "grad_norm": 0.29913750290870667, + "learning_rate": 3.143204398030451e-06, + "loss": 0.6299, + "step": 3326 + }, + { + "epoch": 1.2741719318399387, + "grad_norm": 0.3043437898159027, + "learning_rate": 3.1422228903586096e-06, + "loss": 0.6221, + "step": 3327 + }, + { + "epoch": 1.2745548535324527, + "grad_norm": 0.30663397908210754, + "learning_rate": 3.1412412766907633e-06, + "loss": 0.6555, + "step": 3328 + }, + { + "epoch": 1.2749377752249664, + "grad_norm": 0.3039126694202423, + "learning_rate": 3.1402595571889218e-06, + "loss": 0.639, + "step": 3329 + }, + { + "epoch": 1.2753206969174804, + "grad_norm": 0.3154139220714569, + "learning_rate": 3.139277732015114e-06, + "loss": 0.6398, + "step": 3330 + }, + { + "epoch": 1.2757036186099944, + "grad_norm": 0.3114401698112488, + "learning_rate": 3.138295801331386e-06, + "loss": 0.6551, + "step": 3331 + }, + { + "epoch": 1.2760865403025081, + "grad_norm": 0.30008989572525024, + "learning_rate": 3.1373137652998e-06, + "loss": 0.6235, + "step": 3332 + }, + { + "epoch": 1.276469461995022, + "grad_norm": 0.31326979398727417, + "learning_rate": 3.1363316240824378e-06, + "loss": 0.6222, + "step": 3333 + }, + { + "epoch": 1.2768523836875358, + "grad_norm": 0.3011954426765442, + "learning_rate": 3.135349377841396e-06, + "loss": 0.6342, + "step": 3334 + }, + { + "epoch": 1.2772353053800498, + "grad_norm": 0.3118431270122528, + "learning_rate": 3.1343670267387928e-06, + "loss": 0.6126, + "step": 3335 + }, + { + "epoch": 1.2776182270725638, + "grad_norm": 0.31113603711128235, + "learning_rate": 3.133384570936757e-06, + "loss": 0.6464, + "step": 3336 + }, + { + "epoch": 1.2780011487650775, + "grad_norm": 0.31071075797080994, + "learning_rate": 3.1324020105974417e-06, + "loss": 0.6249, + "step": 3337 + }, + { + "epoch": 1.2783840704575915, + "grad_norm": 0.3089357018470764, + "learning_rate": 3.1314193458830123e-06, + "loss": 0.6449, + "step": 3338 + }, + { + "epoch": 1.2787669921501053, + "grad_norm": 0.30263274908065796, + "learning_rate": 3.1304365769556537e-06, + "loss": 0.6542, + "step": 3339 + }, + { + "epoch": 1.2791499138426192, + "grad_norm": 0.28739720582962036, + "learning_rate": 3.129453703977568e-06, + "loss": 0.612, + "step": 3340 + }, + { + "epoch": 1.2795328355351332, + "grad_norm": 0.2991805076599121, + "learning_rate": 3.1284707271109727e-06, + "loss": 0.6434, + "step": 3341 + }, + { + "epoch": 1.279915757227647, + "grad_norm": 0.3000507056713104, + "learning_rate": 3.1274876465181046e-06, + "loss": 0.622, + "step": 3342 + }, + { + "epoch": 1.2802986789201607, + "grad_norm": 0.3076266050338745, + "learning_rate": 3.126504462361217e-06, + "loss": 0.6197, + "step": 3343 + }, + { + "epoch": 1.2806816006126747, + "grad_norm": 0.31797125935554504, + "learning_rate": 3.1255211748025797e-06, + "loss": 0.6265, + "step": 3344 + }, + { + "epoch": 1.2810645223051886, + "grad_norm": 0.3036264181137085, + "learning_rate": 3.1245377840044805e-06, + "loss": 0.6286, + "step": 3345 + }, + { + "epoch": 1.2814474439977026, + "grad_norm": 0.30863913893699646, + "learning_rate": 3.1235542901292225e-06, + "loss": 0.6415, + "step": 3346 + }, + { + "epoch": 1.2818303656902164, + "grad_norm": 0.3076283037662506, + "learning_rate": 3.1225706933391277e-06, + "loss": 0.6215, + "step": 3347 + }, + { + "epoch": 1.28221328738273, + "grad_norm": 0.2931717336177826, + "learning_rate": 3.1215869937965333e-06, + "loss": 0.6345, + "step": 3348 + }, + { + "epoch": 1.282596209075244, + "grad_norm": 0.29722750186920166, + "learning_rate": 3.1206031916637956e-06, + "loss": 0.6209, + "step": 3349 + }, + { + "epoch": 1.282979130767758, + "grad_norm": 0.2920721173286438, + "learning_rate": 3.119619287103287e-06, + "loss": 0.6313, + "step": 3350 + }, + { + "epoch": 1.2833620524602718, + "grad_norm": 0.3051886260509491, + "learning_rate": 3.1186352802773954e-06, + "loss": 0.6274, + "step": 3351 + }, + { + "epoch": 1.2837449741527858, + "grad_norm": 0.308017760515213, + "learning_rate": 3.1176511713485276e-06, + "loss": 0.6345, + "step": 3352 + }, + { + "epoch": 1.2841278958452995, + "grad_norm": 0.30479660630226135, + "learning_rate": 3.116666960479106e-06, + "loss": 0.6092, + "step": 3353 + }, + { + "epoch": 1.2845108175378135, + "grad_norm": 0.3014877140522003, + "learning_rate": 3.11568264783157e-06, + "loss": 0.6513, + "step": 3354 + }, + { + "epoch": 1.2848937392303275, + "grad_norm": 0.30998489260673523, + "learning_rate": 3.1146982335683763e-06, + "loss": 0.6393, + "step": 3355 + }, + { + "epoch": 1.2852766609228412, + "grad_norm": 0.2925424575805664, + "learning_rate": 3.1137137178519983e-06, + "loss": 0.6557, + "step": 3356 + }, + { + "epoch": 1.2856595826153552, + "grad_norm": 0.3085787296295166, + "learning_rate": 3.1127291008449255e-06, + "loss": 0.6283, + "step": 3357 + }, + { + "epoch": 1.286042504307869, + "grad_norm": 0.3091817796230316, + "learning_rate": 3.1117443827096637e-06, + "loss": 0.643, + "step": 3358 + }, + { + "epoch": 1.286425426000383, + "grad_norm": 0.30707457661628723, + "learning_rate": 3.1107595636087383e-06, + "loss": 0.6411, + "step": 3359 + }, + { + "epoch": 1.2868083476928969, + "grad_norm": 0.30366799235343933, + "learning_rate": 3.1097746437046877e-06, + "loss": 0.6348, + "step": 3360 + }, + { + "epoch": 1.2871912693854106, + "grad_norm": 0.29969143867492676, + "learning_rate": 3.1087896231600684e-06, + "loss": 0.6224, + "step": 3361 + }, + { + "epoch": 1.2875741910779246, + "grad_norm": 0.30460187792778015, + "learning_rate": 3.107804502137455e-06, + "loss": 0.6605, + "step": 3362 + }, + { + "epoch": 1.2879571127704383, + "grad_norm": 0.3067445158958435, + "learning_rate": 3.1068192807994357e-06, + "loss": 0.6046, + "step": 3363 + }, + { + "epoch": 1.2883400344629523, + "grad_norm": 0.3058575391769409, + "learning_rate": 3.1058339593086173e-06, + "loss": 0.6113, + "step": 3364 + }, + { + "epoch": 1.2887229561554663, + "grad_norm": 0.30179211497306824, + "learning_rate": 3.1048485378276233e-06, + "loss": 0.621, + "step": 3365 + }, + { + "epoch": 1.28910587784798, + "grad_norm": 0.300685316324234, + "learning_rate": 3.1038630165190924e-06, + "loss": 0.6362, + "step": 3366 + }, + { + "epoch": 1.289488799540494, + "grad_norm": 0.32143479585647583, + "learning_rate": 3.102877395545682e-06, + "loss": 0.644, + "step": 3367 + }, + { + "epoch": 1.2898717212330078, + "grad_norm": 0.2975754737854004, + "learning_rate": 3.1018916750700618e-06, + "loss": 0.6357, + "step": 3368 + }, + { + "epoch": 1.2902546429255217, + "grad_norm": 0.3113241195678711, + "learning_rate": 3.1009058552549226e-06, + "loss": 0.6383, + "step": 3369 + }, + { + "epoch": 1.2906375646180357, + "grad_norm": 0.3050669729709625, + "learning_rate": 3.0999199362629695e-06, + "loss": 0.6358, + "step": 3370 + }, + { + "epoch": 1.2910204863105494, + "grad_norm": 0.30290520191192627, + "learning_rate": 3.098933918256922e-06, + "loss": 0.6378, + "step": 3371 + }, + { + "epoch": 1.2914034080030634, + "grad_norm": 0.30451831221580505, + "learning_rate": 3.0979478013995205e-06, + "loss": 0.6314, + "step": 3372 + }, + { + "epoch": 1.2917863296955772, + "grad_norm": 0.29953572154045105, + "learning_rate": 3.0969615858535175e-06, + "loss": 0.6517, + "step": 3373 + }, + { + "epoch": 1.2921692513880911, + "grad_norm": 0.30300819873809814, + "learning_rate": 3.095975271781684e-06, + "loss": 0.6569, + "step": 3374 + }, + { + "epoch": 1.292552173080605, + "grad_norm": 0.3085183799266815, + "learning_rate": 3.094988859346807e-06, + "loss": 0.6328, + "step": 3375 + }, + { + "epoch": 1.2929350947731189, + "grad_norm": 0.3153502345085144, + "learning_rate": 3.094002348711689e-06, + "loss": 0.6505, + "step": 3376 + }, + { + "epoch": 1.2933180164656328, + "grad_norm": 0.29932931065559387, + "learning_rate": 3.0930157400391492e-06, + "loss": 0.6469, + "step": 3377 + }, + { + "epoch": 1.2937009381581466, + "grad_norm": 0.3013656437397003, + "learning_rate": 3.0920290334920228e-06, + "loss": 0.6341, + "step": 3378 + }, + { + "epoch": 1.2940838598506605, + "grad_norm": 0.3102124035358429, + "learning_rate": 3.0910422292331622e-06, + "loss": 0.6492, + "step": 3379 + }, + { + "epoch": 1.2944667815431745, + "grad_norm": 0.31404250860214233, + "learning_rate": 3.090055327425433e-06, + "loss": 0.6274, + "step": 3380 + }, + { + "epoch": 1.2948497032356883, + "grad_norm": 0.3045584261417389, + "learning_rate": 3.089068328231721e-06, + "loss": 0.6305, + "step": 3381 + }, + { + "epoch": 1.2952326249282022, + "grad_norm": 0.3016815185546875, + "learning_rate": 3.0880812318149257e-06, + "loss": 0.6287, + "step": 3382 + }, + { + "epoch": 1.295615546620716, + "grad_norm": 0.32292407751083374, + "learning_rate": 3.0870940383379617e-06, + "loss": 0.6551, + "step": 3383 + }, + { + "epoch": 1.29599846831323, + "grad_norm": 0.3142809271812439, + "learning_rate": 3.0861067479637617e-06, + "loss": 0.6302, + "step": 3384 + }, + { + "epoch": 1.296381390005744, + "grad_norm": 0.2990870773792267, + "learning_rate": 3.0851193608552744e-06, + "loss": 0.6312, + "step": 3385 + }, + { + "epoch": 1.2967643116982577, + "grad_norm": 0.2978760600090027, + "learning_rate": 3.0841318771754618e-06, + "loss": 0.6353, + "step": 3386 + }, + { + "epoch": 1.2971472333907716, + "grad_norm": 0.30675920844078064, + "learning_rate": 3.0831442970873044e-06, + "loss": 0.624, + "step": 3387 + }, + { + "epoch": 1.2975301550832854, + "grad_norm": 0.3023224174976349, + "learning_rate": 3.0821566207537984e-06, + "loss": 0.6544, + "step": 3388 + }, + { + "epoch": 1.2979130767757994, + "grad_norm": 0.29305171966552734, + "learning_rate": 3.0811688483379546e-06, + "loss": 0.648, + "step": 3389 + }, + { + "epoch": 1.2982959984683133, + "grad_norm": 0.29888835549354553, + "learning_rate": 3.0801809800028005e-06, + "loss": 0.628, + "step": 3390 + }, + { + "epoch": 1.298678920160827, + "grad_norm": 0.30469363927841187, + "learning_rate": 3.07919301591138e-06, + "loss": 0.6375, + "step": 3391 + }, + { + "epoch": 1.299061841853341, + "grad_norm": 0.29228898882865906, + "learning_rate": 3.0782049562267518e-06, + "loss": 0.623, + "step": 3392 + }, + { + "epoch": 1.2994447635458548, + "grad_norm": 0.30720311403274536, + "learning_rate": 3.0772168011119894e-06, + "loss": 0.6322, + "step": 3393 + }, + { + "epoch": 1.2998276852383688, + "grad_norm": 0.3039799630641937, + "learning_rate": 3.076228550730186e-06, + "loss": 0.6391, + "step": 3394 + }, + { + "epoch": 1.3002106069308828, + "grad_norm": 0.313266783952713, + "learning_rate": 3.0752402052444464e-06, + "loss": 0.648, + "step": 3395 + }, + { + "epoch": 1.3005935286233965, + "grad_norm": 0.30656468868255615, + "learning_rate": 3.074251764817892e-06, + "loss": 0.653, + "step": 3396 + }, + { + "epoch": 1.3009764503159105, + "grad_norm": 0.29536452889442444, + "learning_rate": 3.073263229613661e-06, + "loss": 0.6258, + "step": 3397 + }, + { + "epoch": 1.3013593720084242, + "grad_norm": 0.3054569661617279, + "learning_rate": 3.072274599794908e-06, + "loss": 0.6357, + "step": 3398 + }, + { + "epoch": 1.3017422937009382, + "grad_norm": 0.29739561676979065, + "learning_rate": 3.0712858755248e-06, + "loss": 0.6138, + "step": 3399 + }, + { + "epoch": 1.3021252153934522, + "grad_norm": 0.30128034949302673, + "learning_rate": 3.0702970569665213e-06, + "loss": 0.65, + "step": 3400 + }, + { + "epoch": 1.302508137085966, + "grad_norm": 0.30430489778518677, + "learning_rate": 3.069308144283274e-06, + "loss": 0.6261, + "step": 3401 + }, + { + "epoch": 1.3028910587784799, + "grad_norm": 0.31417909264564514, + "learning_rate": 3.068319137638272e-06, + "loss": 0.6284, + "step": 3402 + }, + { + "epoch": 1.3032739804709936, + "grad_norm": 0.31440991163253784, + "learning_rate": 3.067330037194748e-06, + "loss": 0.6279, + "step": 3403 + }, + { + "epoch": 1.3036569021635076, + "grad_norm": 0.3081521987915039, + "learning_rate": 3.0663408431159463e-06, + "loss": 0.6507, + "step": 3404 + }, + { + "epoch": 1.3040398238560216, + "grad_norm": 0.30417507886886597, + "learning_rate": 3.0653515555651303e-06, + "loss": 0.6521, + "step": 3405 + }, + { + "epoch": 1.3044227455485353, + "grad_norm": 0.3088940680027008, + "learning_rate": 3.0643621747055785e-06, + "loss": 0.6371, + "step": 3406 + }, + { + "epoch": 1.304805667241049, + "grad_norm": 0.29372096061706543, + "learning_rate": 3.0633727007005816e-06, + "loss": 0.6226, + "step": 3407 + }, + { + "epoch": 1.305188588933563, + "grad_norm": 0.3101111948490143, + "learning_rate": 3.0623831337134497e-06, + "loss": 0.6354, + "step": 3408 + }, + { + "epoch": 1.305571510626077, + "grad_norm": 0.31358519196510315, + "learning_rate": 3.061393473907505e-06, + "loss": 0.6332, + "step": 3409 + }, + { + "epoch": 1.305954432318591, + "grad_norm": 0.2956601679325104, + "learning_rate": 3.0604037214460867e-06, + "loss": 0.6309, + "step": 3410 + }, + { + "epoch": 1.3063373540111047, + "grad_norm": 0.2942085862159729, + "learning_rate": 3.0594138764925507e-06, + "loss": 0.6181, + "step": 3411 + }, + { + "epoch": 1.3067202757036185, + "grad_norm": 0.31667640805244446, + "learning_rate": 3.058423939210264e-06, + "loss": 0.6574, + "step": 3412 + }, + { + "epoch": 1.3071031973961325, + "grad_norm": 0.30363062024116516, + "learning_rate": 3.057433909762613e-06, + "loss": 0.6269, + "step": 3413 + }, + { + "epoch": 1.3074861190886464, + "grad_norm": 0.3038477897644043, + "learning_rate": 3.056443788312996e-06, + "loss": 0.6401, + "step": 3414 + }, + { + "epoch": 1.3078690407811602, + "grad_norm": 0.31721600890159607, + "learning_rate": 3.0554535750248304e-06, + "loss": 0.6262, + "step": 3415 + }, + { + "epoch": 1.3082519624736741, + "grad_norm": 0.3201966881752014, + "learning_rate": 3.0544632700615447e-06, + "loss": 0.632, + "step": 3416 + }, + { + "epoch": 1.308634884166188, + "grad_norm": 0.29836297035217285, + "learning_rate": 3.0534728735865837e-06, + "loss": 0.6363, + "step": 3417 + }, + { + "epoch": 1.3090178058587019, + "grad_norm": 0.31272047758102417, + "learning_rate": 3.05248238576341e-06, + "loss": 0.6271, + "step": 3418 + }, + { + "epoch": 1.3094007275512158, + "grad_norm": 0.30757108330726624, + "learning_rate": 3.0514918067554975e-06, + "loss": 0.649, + "step": 3419 + }, + { + "epoch": 1.3097836492437296, + "grad_norm": 0.31474122405052185, + "learning_rate": 3.050501136726338e-06, + "loss": 0.6425, + "step": 3420 + }, + { + "epoch": 1.3101665709362436, + "grad_norm": 0.30411624908447266, + "learning_rate": 3.0495103758394366e-06, + "loss": 0.636, + "step": 3421 + }, + { + "epoch": 1.3105494926287573, + "grad_norm": 0.32310327887535095, + "learning_rate": 3.048519524258314e-06, + "loss": 0.6279, + "step": 3422 + }, + { + "epoch": 1.3109324143212713, + "grad_norm": 0.32532253861427307, + "learning_rate": 3.0475285821465055e-06, + "loss": 0.6281, + "step": 3423 + }, + { + "epoch": 1.3113153360137852, + "grad_norm": 0.3010132312774658, + "learning_rate": 3.046537549667562e-06, + "loss": 0.6364, + "step": 3424 + }, + { + "epoch": 1.311698257706299, + "grad_norm": 0.3065738379955292, + "learning_rate": 3.045546426985049e-06, + "loss": 0.6433, + "step": 3425 + }, + { + "epoch": 1.312081179398813, + "grad_norm": 0.30610325932502747, + "learning_rate": 3.0445552142625455e-06, + "loss": 0.6402, + "step": 3426 + }, + { + "epoch": 1.3124641010913267, + "grad_norm": 0.3185786306858063, + "learning_rate": 3.043563911663649e-06, + "loss": 0.6484, + "step": 3427 + }, + { + "epoch": 1.3128470227838407, + "grad_norm": 0.30856701731681824, + "learning_rate": 3.0425725193519683e-06, + "loss": 0.6452, + "step": 3428 + }, + { + "epoch": 1.3132299444763547, + "grad_norm": 0.3027724623680115, + "learning_rate": 3.0415810374911283e-06, + "loss": 0.6334, + "step": 3429 + }, + { + "epoch": 1.3136128661688684, + "grad_norm": 0.3149544894695282, + "learning_rate": 3.0405894662447682e-06, + "loss": 0.6365, + "step": 3430 + }, + { + "epoch": 1.3139957878613824, + "grad_norm": 0.34518617391586304, + "learning_rate": 3.039597805776544e-06, + "loss": 0.6264, + "step": 3431 + }, + { + "epoch": 1.3143787095538961, + "grad_norm": 0.2973020374774933, + "learning_rate": 3.0386060562501224e-06, + "loss": 0.6042, + "step": 3432 + }, + { + "epoch": 1.31476163124641, + "grad_norm": 0.30090364813804626, + "learning_rate": 3.0376142178291895e-06, + "loss": 0.6407, + "step": 3433 + }, + { + "epoch": 1.315144552938924, + "grad_norm": 0.3288208842277527, + "learning_rate": 3.036622290677443e-06, + "loss": 0.6408, + "step": 3434 + }, + { + "epoch": 1.3155274746314378, + "grad_norm": 0.31786048412323, + "learning_rate": 3.0356302749585954e-06, + "loss": 0.6616, + "step": 3435 + }, + { + "epoch": 1.3159103963239518, + "grad_norm": 0.304740309715271, + "learning_rate": 3.034638170836374e-06, + "loss": 0.6281, + "step": 3436 + }, + { + "epoch": 1.3162933180164655, + "grad_norm": 0.3104275166988373, + "learning_rate": 3.0336459784745233e-06, + "loss": 0.6439, + "step": 3437 + }, + { + "epoch": 1.3166762397089795, + "grad_norm": 0.31141525506973267, + "learning_rate": 3.0326536980367987e-06, + "loss": 0.6322, + "step": 3438 + }, + { + "epoch": 1.3170591614014935, + "grad_norm": 0.3017503619194031, + "learning_rate": 3.031661329686971e-06, + "loss": 0.6349, + "step": 3439 + }, + { + "epoch": 1.3174420830940072, + "grad_norm": 0.3082047402858734, + "learning_rate": 3.0306688735888278e-06, + "loss": 0.6403, + "step": 3440 + }, + { + "epoch": 1.3178250047865212, + "grad_norm": 0.3151637315750122, + "learning_rate": 3.029676329906169e-06, + "loss": 0.6184, + "step": 3441 + }, + { + "epoch": 1.318207926479035, + "grad_norm": 0.30428728461265564, + "learning_rate": 3.0286836988028084e-06, + "loss": 0.6114, + "step": 3442 + }, + { + "epoch": 1.318590848171549, + "grad_norm": 0.29638904333114624, + "learning_rate": 3.0276909804425766e-06, + "loss": 0.6458, + "step": 3443 + }, + { + "epoch": 1.318973769864063, + "grad_norm": 0.291727215051651, + "learning_rate": 3.026698174989316e-06, + "loss": 0.622, + "step": 3444 + }, + { + "epoch": 1.3193566915565766, + "grad_norm": 0.30981624126434326, + "learning_rate": 3.025705282606887e-06, + "loss": 0.6483, + "step": 3445 + }, + { + "epoch": 1.3197396132490906, + "grad_norm": 0.2957691550254822, + "learning_rate": 3.0247123034591586e-06, + "loss": 0.6482, + "step": 3446 + }, + { + "epoch": 1.3201225349416044, + "grad_norm": 0.3026467263698578, + "learning_rate": 3.0237192377100205e-06, + "loss": 0.6501, + "step": 3447 + }, + { + "epoch": 1.3205054566341183, + "grad_norm": 0.2967451214790344, + "learning_rate": 3.022726085523373e-06, + "loss": 0.6227, + "step": 3448 + }, + { + "epoch": 1.3208883783266323, + "grad_norm": 0.31251364946365356, + "learning_rate": 3.0217328470631303e-06, + "loss": 0.6405, + "step": 3449 + }, + { + "epoch": 1.321271300019146, + "grad_norm": 0.29531168937683105, + "learning_rate": 3.0207395224932236e-06, + "loss": 0.6152, + "step": 3450 + }, + { + "epoch": 1.32165422171166, + "grad_norm": 0.3061845302581787, + "learning_rate": 3.0197461119775943e-06, + "loss": 0.6405, + "step": 3451 + }, + { + "epoch": 1.3220371434041738, + "grad_norm": 0.30625516176223755, + "learning_rate": 3.018752615680204e-06, + "loss": 0.6558, + "step": 3452 + }, + { + "epoch": 1.3224200650966877, + "grad_norm": 0.30157727003097534, + "learning_rate": 3.0177590337650216e-06, + "loss": 0.6454, + "step": 3453 + }, + { + "epoch": 1.3228029867892017, + "grad_norm": 0.31994837522506714, + "learning_rate": 3.0167653663960345e-06, + "loss": 0.6509, + "step": 3454 + }, + { + "epoch": 1.3231859084817155, + "grad_norm": 0.31176507472991943, + "learning_rate": 3.015771613737243e-06, + "loss": 0.6535, + "step": 3455 + }, + { + "epoch": 1.3235688301742294, + "grad_norm": 0.30533549189567566, + "learning_rate": 3.0147777759526612e-06, + "loss": 0.6162, + "step": 3456 + }, + { + "epoch": 1.3239517518667432, + "grad_norm": 0.30813509225845337, + "learning_rate": 3.0137838532063178e-06, + "loss": 0.6568, + "step": 3457 + }, + { + "epoch": 1.3243346735592572, + "grad_norm": 0.32369309663772583, + "learning_rate": 3.0127898456622558e-06, + "loss": 0.635, + "step": 3458 + }, + { + "epoch": 1.3247175952517711, + "grad_norm": 0.3016875982284546, + "learning_rate": 3.011795753484531e-06, + "loss": 0.6552, + "step": 3459 + }, + { + "epoch": 1.3251005169442849, + "grad_norm": 0.3015488088130951, + "learning_rate": 3.0108015768372144e-06, + "loss": 0.6385, + "step": 3460 + }, + { + "epoch": 1.3254834386367988, + "grad_norm": 0.3006605803966522, + "learning_rate": 3.009807315884389e-06, + "loss": 0.6477, + "step": 3461 + }, + { + "epoch": 1.3258663603293126, + "grad_norm": 0.326983243227005, + "learning_rate": 3.008812970790156e-06, + "loss": 0.6454, + "step": 3462 + }, + { + "epoch": 1.3262492820218266, + "grad_norm": 0.3204255998134613, + "learning_rate": 3.0078185417186245e-06, + "loss": 0.6291, + "step": 3463 + }, + { + "epoch": 1.3266322037143405, + "grad_norm": 0.3035261332988739, + "learning_rate": 3.0068240288339222e-06, + "loss": 0.6227, + "step": 3464 + }, + { + "epoch": 1.3270151254068543, + "grad_norm": 0.31438133120536804, + "learning_rate": 3.005829432300188e-06, + "loss": 0.6392, + "step": 3465 + }, + { + "epoch": 1.3273980470993683, + "grad_norm": 0.3038402795791626, + "learning_rate": 3.004834752281577e-06, + "loss": 0.63, + "step": 3466 + }, + { + "epoch": 1.327780968791882, + "grad_norm": 0.31573766469955444, + "learning_rate": 3.0038399889422553e-06, + "loss": 0.6325, + "step": 3467 + }, + { + "epoch": 1.328163890484396, + "grad_norm": 0.303928017616272, + "learning_rate": 3.0028451424464045e-06, + "loss": 0.6336, + "step": 3468 + }, + { + "epoch": 1.32854681217691, + "grad_norm": 0.313711017370224, + "learning_rate": 3.0018502129582207e-06, + "loss": 0.6114, + "step": 3469 + }, + { + "epoch": 1.3289297338694237, + "grad_norm": 0.3077126145362854, + "learning_rate": 3.0008552006419113e-06, + "loss": 0.6138, + "step": 3470 + }, + { + "epoch": 1.3293126555619375, + "grad_norm": 0.3379300832748413, + "learning_rate": 2.9998601056616986e-06, + "loss": 0.6346, + "step": 3471 + }, + { + "epoch": 1.3296955772544514, + "grad_norm": 0.29071134328842163, + "learning_rate": 2.998864928181818e-06, + "loss": 0.6415, + "step": 3472 + }, + { + "epoch": 1.3300784989469654, + "grad_norm": 0.3412402868270874, + "learning_rate": 2.9978696683665215e-06, + "loss": 0.6423, + "step": 3473 + }, + { + "epoch": 1.3304614206394794, + "grad_norm": 0.3114089071750641, + "learning_rate": 2.9968743263800703e-06, + "loss": 0.6117, + "step": 3474 + }, + { + "epoch": 1.3308443423319931, + "grad_norm": 0.3122973144054413, + "learning_rate": 2.995878902386741e-06, + "loss": 0.6201, + "step": 3475 + }, + { + "epoch": 1.3312272640245069, + "grad_norm": 0.3368566930294037, + "learning_rate": 2.994883396550825e-06, + "loss": 0.6082, + "step": 3476 + }, + { + "epoch": 1.3316101857170208, + "grad_norm": 0.3157362937927246, + "learning_rate": 2.9938878090366253e-06, + "loss": 0.6458, + "step": 3477 + }, + { + "epoch": 1.3319931074095348, + "grad_norm": 0.31080082058906555, + "learning_rate": 2.9928921400084594e-06, + "loss": 0.631, + "step": 3478 + }, + { + "epoch": 1.3323760291020486, + "grad_norm": 0.31414538621902466, + "learning_rate": 2.9918963896306576e-06, + "loss": 0.6488, + "step": 3479 + }, + { + "epoch": 1.3327589507945625, + "grad_norm": 0.33237704634666443, + "learning_rate": 2.990900558067565e-06, + "loss": 0.627, + "step": 3480 + }, + { + "epoch": 1.3331418724870763, + "grad_norm": 0.30415868759155273, + "learning_rate": 2.989904645483538e-06, + "loss": 0.6401, + "step": 3481 + }, + { + "epoch": 1.3335247941795902, + "grad_norm": 0.3034331500530243, + "learning_rate": 2.988908652042948e-06, + "loss": 0.6359, + "step": 3482 + }, + { + "epoch": 1.3339077158721042, + "grad_norm": 0.29974913597106934, + "learning_rate": 2.9879125779101793e-06, + "loss": 0.6357, + "step": 3483 + }, + { + "epoch": 1.334290637564618, + "grad_norm": 0.31873369216918945, + "learning_rate": 2.9869164232496296e-06, + "loss": 0.6426, + "step": 3484 + }, + { + "epoch": 1.334673559257132, + "grad_norm": 0.3236675560474396, + "learning_rate": 2.985920188225709e-06, + "loss": 0.6114, + "step": 3485 + }, + { + "epoch": 1.3350564809496457, + "grad_norm": 0.3190526068210602, + "learning_rate": 2.9849238730028423e-06, + "loss": 0.6097, + "step": 3486 + }, + { + "epoch": 1.3354394026421597, + "grad_norm": 0.30243387818336487, + "learning_rate": 2.9839274777454673e-06, + "loss": 0.6307, + "step": 3487 + }, + { + "epoch": 1.3358223243346736, + "grad_norm": 0.31402692198753357, + "learning_rate": 2.9829310026180336e-06, + "loss": 0.6344, + "step": 3488 + }, + { + "epoch": 1.3362052460271874, + "grad_norm": 0.3246251344680786, + "learning_rate": 2.9819344477850055e-06, + "loss": 0.6246, + "step": 3489 + }, + { + "epoch": 1.3365881677197013, + "grad_norm": 0.3181582987308502, + "learning_rate": 2.9809378134108592e-06, + "loss": 0.6433, + "step": 3490 + }, + { + "epoch": 1.336971089412215, + "grad_norm": 0.3021993339061737, + "learning_rate": 2.9799410996600866e-06, + "loss": 0.645, + "step": 3491 + }, + { + "epoch": 1.337354011104729, + "grad_norm": 0.3065638542175293, + "learning_rate": 2.9789443066971886e-06, + "loss": 0.6112, + "step": 3492 + }, + { + "epoch": 1.337736932797243, + "grad_norm": 0.31228554248809814, + "learning_rate": 2.9779474346866822e-06, + "loss": 0.6288, + "step": 3493 + }, + { + "epoch": 1.3381198544897568, + "grad_norm": 0.3105200529098511, + "learning_rate": 2.9769504837930974e-06, + "loss": 0.6236, + "step": 3494 + }, + { + "epoch": 1.3385027761822708, + "grad_norm": 0.32057079672813416, + "learning_rate": 2.975953454180975e-06, + "loss": 0.6329, + "step": 3495 + }, + { + "epoch": 1.3388856978747845, + "grad_norm": 0.30296993255615234, + "learning_rate": 2.9749563460148715e-06, + "loss": 0.6342, + "step": 3496 + }, + { + "epoch": 1.3392686195672985, + "grad_norm": 0.2965753674507141, + "learning_rate": 2.973959159459354e-06, + "loss": 0.6311, + "step": 3497 + }, + { + "epoch": 1.3396515412598125, + "grad_norm": 0.3148595690727234, + "learning_rate": 2.9729618946790064e-06, + "loss": 0.6548, + "step": 3498 + }, + { + "epoch": 1.3400344629523262, + "grad_norm": 0.30869805812835693, + "learning_rate": 2.9719645518384194e-06, + "loss": 0.6228, + "step": 3499 + }, + { + "epoch": 1.3404173846448402, + "grad_norm": 0.31261077523231506, + "learning_rate": 2.9709671311022008e-06, + "loss": 0.6084, + "step": 3500 + }, + { + "epoch": 1.340800306337354, + "grad_norm": 0.2979696989059448, + "learning_rate": 2.9699696326349726e-06, + "loss": 0.6537, + "step": 3501 + }, + { + "epoch": 1.341183228029868, + "grad_norm": 0.3083201050758362, + "learning_rate": 2.9689720566013643e-06, + "loss": 0.6517, + "step": 3502 + }, + { + "epoch": 1.3415661497223819, + "grad_norm": 0.3180858790874481, + "learning_rate": 2.9679744031660233e-06, + "loss": 0.6338, + "step": 3503 + }, + { + "epoch": 1.3419490714148956, + "grad_norm": 0.31594112515449524, + "learning_rate": 2.9669766724936074e-06, + "loss": 0.6299, + "step": 3504 + }, + { + "epoch": 1.3423319931074096, + "grad_norm": 0.29814061522483826, + "learning_rate": 2.9659788647487873e-06, + "loss": 0.6297, + "step": 3505 + }, + { + "epoch": 1.3427149147999233, + "grad_norm": 0.31089481711387634, + "learning_rate": 2.9649809800962475e-06, + "loss": 0.6459, + "step": 3506 + }, + { + "epoch": 1.3430978364924373, + "grad_norm": 0.3138785660266876, + "learning_rate": 2.9639830187006834e-06, + "loss": 0.6593, + "step": 3507 + }, + { + "epoch": 1.3434807581849513, + "grad_norm": 0.31171032786369324, + "learning_rate": 2.962984980726805e-06, + "loss": 0.6358, + "step": 3508 + }, + { + "epoch": 1.343863679877465, + "grad_norm": 0.3212209939956665, + "learning_rate": 2.9619868663393326e-06, + "loss": 0.6288, + "step": 3509 + }, + { + "epoch": 1.344246601569979, + "grad_norm": 0.32481661438941956, + "learning_rate": 2.960988675703002e-06, + "loss": 0.6408, + "step": 3510 + }, + { + "epoch": 1.3446295232624927, + "grad_norm": 0.30457279086112976, + "learning_rate": 2.9599904089825586e-06, + "loss": 0.6438, + "step": 3511 + }, + { + "epoch": 1.3450124449550067, + "grad_norm": 0.29482752084732056, + "learning_rate": 2.958992066342764e-06, + "loss": 0.6147, + "step": 3512 + }, + { + "epoch": 1.3453953666475207, + "grad_norm": 0.3114534020423889, + "learning_rate": 2.957993647948388e-06, + "loss": 0.6368, + "step": 3513 + }, + { + "epoch": 1.3457782883400344, + "grad_norm": 0.31687167286872864, + "learning_rate": 2.9569951539642156e-06, + "loss": 0.6234, + "step": 3514 + }, + { + "epoch": 1.3461612100325484, + "grad_norm": 0.3141995668411255, + "learning_rate": 2.9559965845550446e-06, + "loss": 0.635, + "step": 3515 + }, + { + "epoch": 1.3465441317250622, + "grad_norm": 0.30248358845710754, + "learning_rate": 2.9549979398856842e-06, + "loss": 0.5936, + "step": 3516 + }, + { + "epoch": 1.3469270534175761, + "grad_norm": 0.3059993386268616, + "learning_rate": 2.953999220120955e-06, + "loss": 0.6133, + "step": 3517 + }, + { + "epoch": 1.34730997511009, + "grad_norm": 0.3149002492427826, + "learning_rate": 2.9530004254256926e-06, + "loss": 0.6363, + "step": 3518 + }, + { + "epoch": 1.3476928968026038, + "grad_norm": 0.3106870651245117, + "learning_rate": 2.952001555964743e-06, + "loss": 0.6395, + "step": 3519 + }, + { + "epoch": 1.3480758184951178, + "grad_norm": 0.29037734866142273, + "learning_rate": 2.9510026119029655e-06, + "loss": 0.6511, + "step": 3520 + }, + { + "epoch": 1.3484587401876316, + "grad_norm": 0.3048921525478363, + "learning_rate": 2.9500035934052307e-06, + "loss": 0.6457, + "step": 3521 + }, + { + "epoch": 1.3488416618801455, + "grad_norm": 0.3078567683696747, + "learning_rate": 2.9490045006364227e-06, + "loss": 0.6379, + "step": 3522 + }, + { + "epoch": 1.3492245835726595, + "grad_norm": 0.3131403923034668, + "learning_rate": 2.9480053337614366e-06, + "loss": 0.6296, + "step": 3523 + }, + { + "epoch": 1.3496075052651733, + "grad_norm": 0.2994149327278137, + "learning_rate": 2.947006092945181e-06, + "loss": 0.6256, + "step": 3524 + }, + { + "epoch": 1.3499904269576872, + "grad_norm": 0.3043414354324341, + "learning_rate": 2.9460067783525763e-06, + "loss": 0.6248, + "step": 3525 + }, + { + "epoch": 1.350373348650201, + "grad_norm": 0.3111502230167389, + "learning_rate": 2.945007390148555e-06, + "loss": 0.627, + "step": 3526 + }, + { + "epoch": 1.350756270342715, + "grad_norm": 0.2965335249900818, + "learning_rate": 2.94400792849806e-06, + "loss": 0.6068, + "step": 3527 + }, + { + "epoch": 1.351139192035229, + "grad_norm": 0.3032225966453552, + "learning_rate": 2.9430083935660495e-06, + "loss": 0.6432, + "step": 3528 + }, + { + "epoch": 1.3515221137277427, + "grad_norm": 0.29998916387557983, + "learning_rate": 2.9420087855174918e-06, + "loss": 0.6299, + "step": 3529 + }, + { + "epoch": 1.3519050354202566, + "grad_norm": 0.3105623722076416, + "learning_rate": 2.941009104517369e-06, + "loss": 0.6335, + "step": 3530 + }, + { + "epoch": 1.3522879571127704, + "grad_norm": 0.30252307653427124, + "learning_rate": 2.9400093507306714e-06, + "loss": 0.6291, + "step": 3531 + }, + { + "epoch": 1.3526708788052844, + "grad_norm": 0.30943647027015686, + "learning_rate": 2.939009524322406e-06, + "loss": 0.6282, + "step": 3532 + }, + { + "epoch": 1.3530538004977983, + "grad_norm": 0.3124632239341736, + "learning_rate": 2.9380096254575885e-06, + "loss": 0.6529, + "step": 3533 + }, + { + "epoch": 1.353436722190312, + "grad_norm": 0.3047025203704834, + "learning_rate": 2.937009654301248e-06, + "loss": 0.6283, + "step": 3534 + }, + { + "epoch": 1.3538196438828258, + "grad_norm": 0.3021862208843231, + "learning_rate": 2.936009611018426e-06, + "loss": 0.6392, + "step": 3535 + }, + { + "epoch": 1.3542025655753398, + "grad_norm": 0.31318697333335876, + "learning_rate": 2.935009495774174e-06, + "loss": 0.6022, + "step": 3536 + }, + { + "epoch": 1.3545854872678538, + "grad_norm": 0.3080444633960724, + "learning_rate": 2.934009308733558e-06, + "loss": 0.6195, + "step": 3537 + }, + { + "epoch": 1.3549684089603677, + "grad_norm": 0.32029610872268677, + "learning_rate": 2.933009050061652e-06, + "loss": 0.6274, + "step": 3538 + }, + { + "epoch": 1.3553513306528815, + "grad_norm": 0.3241141438484192, + "learning_rate": 2.9320087199235463e-06, + "loss": 0.6248, + "step": 3539 + }, + { + "epoch": 1.3557342523453952, + "grad_norm": 0.29844725131988525, + "learning_rate": 2.9310083184843414e-06, + "loss": 0.6463, + "step": 3540 + }, + { + "epoch": 1.3561171740379092, + "grad_norm": 0.3161974549293518, + "learning_rate": 2.930007845909146e-06, + "loss": 0.6365, + "step": 3541 + }, + { + "epoch": 1.3565000957304232, + "grad_norm": 0.3106662929058075, + "learning_rate": 2.9290073023630873e-06, + "loss": 0.5992, + "step": 3542 + }, + { + "epoch": 1.356883017422937, + "grad_norm": 0.3189105689525604, + "learning_rate": 2.928006688011297e-06, + "loss": 0.6449, + "step": 3543 + }, + { + "epoch": 1.357265939115451, + "grad_norm": 0.3028534948825836, + "learning_rate": 2.9270060030189252e-06, + "loss": 0.6388, + "step": 3544 + }, + { + "epoch": 1.3576488608079647, + "grad_norm": 0.3208237588405609, + "learning_rate": 2.9260052475511284e-06, + "loss": 0.6459, + "step": 3545 + }, + { + "epoch": 1.3580317825004786, + "grad_norm": 0.32957223057746887, + "learning_rate": 2.925004421773078e-06, + "loss": 0.6183, + "step": 3546 + }, + { + "epoch": 1.3584147041929926, + "grad_norm": 0.30632898211479187, + "learning_rate": 2.924003525849955e-06, + "loss": 0.6297, + "step": 3547 + }, + { + "epoch": 1.3587976258855063, + "grad_norm": 0.3172092139720917, + "learning_rate": 2.9230025599469526e-06, + "loss": 0.6143, + "step": 3548 + }, + { + "epoch": 1.3591805475780203, + "grad_norm": 0.320317804813385, + "learning_rate": 2.9220015242292763e-06, + "loss": 0.6333, + "step": 3549 + }, + { + "epoch": 1.359563469270534, + "grad_norm": 0.337410569190979, + "learning_rate": 2.921000418862143e-06, + "loss": 0.6288, + "step": 3550 + }, + { + "epoch": 1.359946390963048, + "grad_norm": 0.3029799163341522, + "learning_rate": 2.9199992440107796e-06, + "loss": 0.6405, + "step": 3551 + }, + { + "epoch": 1.360329312655562, + "grad_norm": 0.31723785400390625, + "learning_rate": 2.918997999840426e-06, + "loss": 0.6335, + "step": 3552 + }, + { + "epoch": 1.3607122343480758, + "grad_norm": 0.324891597032547, + "learning_rate": 2.917996686516332e-06, + "loss": 0.6196, + "step": 3553 + }, + { + "epoch": 1.3610951560405897, + "grad_norm": 0.34273794293403625, + "learning_rate": 2.9169953042037623e-06, + "loss": 0.6329, + "step": 3554 + }, + { + "epoch": 1.3614780777331035, + "grad_norm": 0.30912837386131287, + "learning_rate": 2.915993853067989e-06, + "loss": 0.6181, + "step": 3555 + }, + { + "epoch": 1.3618609994256174, + "grad_norm": 0.3377562463283539, + "learning_rate": 2.914992333274297e-06, + "loss": 0.6278, + "step": 3556 + }, + { + "epoch": 1.3622439211181314, + "grad_norm": 0.31978172063827515, + "learning_rate": 2.913990744987983e-06, + "loss": 0.6435, + "step": 3557 + }, + { + "epoch": 1.3626268428106452, + "grad_norm": 0.3117522597312927, + "learning_rate": 2.9129890883743544e-06, + "loss": 0.6194, + "step": 3558 + }, + { + "epoch": 1.3630097645031591, + "grad_norm": 0.3061492443084717, + "learning_rate": 2.9119873635987306e-06, + "loss": 0.6437, + "step": 3559 + }, + { + "epoch": 1.3633926861956729, + "grad_norm": 0.3146529197692871, + "learning_rate": 2.9109855708264404e-06, + "loss": 0.6461, + "step": 3560 + }, + { + "epoch": 1.3637756078881869, + "grad_norm": 0.33041971921920776, + "learning_rate": 2.9099837102228273e-06, + "loss": 0.6145, + "step": 3561 + }, + { + "epoch": 1.3641585295807008, + "grad_norm": 0.31546491384506226, + "learning_rate": 2.9089817819532427e-06, + "loss": 0.6376, + "step": 3562 + }, + { + "epoch": 1.3645414512732146, + "grad_norm": 0.2955082356929779, + "learning_rate": 2.9079797861830507e-06, + "loss": 0.6415, + "step": 3563 + }, + { + "epoch": 1.3649243729657285, + "grad_norm": 0.30190083384513855, + "learning_rate": 2.906977723077626e-06, + "loss": 0.6157, + "step": 3564 + }, + { + "epoch": 1.3653072946582423, + "grad_norm": 0.3112003207206726, + "learning_rate": 2.9059755928023547e-06, + "loss": 0.6469, + "step": 3565 + }, + { + "epoch": 1.3656902163507563, + "grad_norm": 0.2988673448562622, + "learning_rate": 2.904973395522633e-06, + "loss": 0.643, + "step": 3566 + }, + { + "epoch": 1.3660731380432702, + "grad_norm": 0.3124767541885376, + "learning_rate": 2.903971131403871e-06, + "loss": 0.6502, + "step": 3567 + }, + { + "epoch": 1.366456059735784, + "grad_norm": 0.314493328332901, + "learning_rate": 2.902968800611486e-06, + "loss": 0.6491, + "step": 3568 + }, + { + "epoch": 1.366838981428298, + "grad_norm": 0.2988538146018982, + "learning_rate": 2.9019664033109095e-06, + "loss": 0.6266, + "step": 3569 + }, + { + "epoch": 1.3672219031208117, + "grad_norm": 0.3087799549102783, + "learning_rate": 2.9009639396675822e-06, + "loss": 0.6302, + "step": 3570 + }, + { + "epoch": 1.3676048248133257, + "grad_norm": 0.3033668100833893, + "learning_rate": 2.8999614098469566e-06, + "loss": 0.6379, + "step": 3571 + }, + { + "epoch": 1.3679877465058397, + "grad_norm": 0.31477120518684387, + "learning_rate": 2.898958814014495e-06, + "loss": 0.6358, + "step": 3572 + }, + { + "epoch": 1.3683706681983534, + "grad_norm": 0.309085875749588, + "learning_rate": 2.897956152335671e-06, + "loss": 0.6375, + "step": 3573 + }, + { + "epoch": 1.3687535898908674, + "grad_norm": 0.3145656883716583, + "learning_rate": 2.896953424975971e-06, + "loss": 0.641, + "step": 3574 + }, + { + "epoch": 1.3691365115833811, + "grad_norm": 0.31623488664627075, + "learning_rate": 2.895950632100889e-06, + "loss": 0.6383, + "step": 3575 + }, + { + "epoch": 1.369519433275895, + "grad_norm": 0.325952410697937, + "learning_rate": 2.8949477738759337e-06, + "loss": 0.6458, + "step": 3576 + }, + { + "epoch": 1.369902354968409, + "grad_norm": 0.30940940976142883, + "learning_rate": 2.893944850466619e-06, + "loss": 0.6372, + "step": 3577 + }, + { + "epoch": 1.3702852766609228, + "grad_norm": 0.3286595344543457, + "learning_rate": 2.892941862038475e-06, + "loss": 0.6358, + "step": 3578 + }, + { + "epoch": 1.3706681983534368, + "grad_norm": 0.30766287446022034, + "learning_rate": 2.891938808757042e-06, + "loss": 0.6278, + "step": 3579 + }, + { + "epoch": 1.3710511200459505, + "grad_norm": 0.31462857127189636, + "learning_rate": 2.8909356907878662e-06, + "loss": 0.6265, + "step": 3580 + }, + { + "epoch": 1.3714340417384645, + "grad_norm": 0.31746456027030945, + "learning_rate": 2.8899325082965096e-06, + "loss": 0.6434, + "step": 3581 + }, + { + "epoch": 1.3718169634309785, + "grad_norm": 0.3038647770881653, + "learning_rate": 2.888929261448542e-06, + "loss": 0.6372, + "step": 3582 + }, + { + "epoch": 1.3721998851234922, + "grad_norm": 0.30422812700271606, + "learning_rate": 2.887925950409546e-06, + "loss": 0.6187, + "step": 3583 + }, + { + "epoch": 1.3725828068160062, + "grad_norm": 0.3123907148838043, + "learning_rate": 2.886922575345113e-06, + "loss": 0.6271, + "step": 3584 + }, + { + "epoch": 1.37296572850852, + "grad_norm": 0.30144569277763367, + "learning_rate": 2.885919136420845e-06, + "loss": 0.6357, + "step": 3585 + }, + { + "epoch": 1.373348650201034, + "grad_norm": 0.331651508808136, + "learning_rate": 2.8849156338023566e-06, + "loss": 0.6277, + "step": 3586 + }, + { + "epoch": 1.3737315718935479, + "grad_norm": 0.3100121021270752, + "learning_rate": 2.883912067655269e-06, + "loss": 0.6318, + "step": 3587 + }, + { + "epoch": 1.3741144935860616, + "grad_norm": 0.3198723793029785, + "learning_rate": 2.882908438145219e-06, + "loss": 0.6358, + "step": 3588 + }, + { + "epoch": 1.3744974152785756, + "grad_norm": 0.3102874159812927, + "learning_rate": 2.8819047454378483e-06, + "loss": 0.6341, + "step": 3589 + }, + { + "epoch": 1.3748803369710894, + "grad_norm": 0.3200148344039917, + "learning_rate": 2.8809009896988144e-06, + "loss": 0.6367, + "step": 3590 + }, + { + "epoch": 1.3752632586636033, + "grad_norm": 0.3058067858219147, + "learning_rate": 2.879897171093782e-06, + "loss": 0.6281, + "step": 3591 + }, + { + "epoch": 1.3756461803561173, + "grad_norm": 0.32124942541122437, + "learning_rate": 2.878893289788425e-06, + "loss": 0.6275, + "step": 3592 + }, + { + "epoch": 1.376029102048631, + "grad_norm": 0.3092166781425476, + "learning_rate": 2.877889345948432e-06, + "loss": 0.6605, + "step": 3593 + }, + { + "epoch": 1.376412023741145, + "grad_norm": 0.31537073850631714, + "learning_rate": 2.8768853397394985e-06, + "loss": 0.6388, + "step": 3594 + }, + { + "epoch": 1.3767949454336588, + "grad_norm": 0.3219885230064392, + "learning_rate": 2.875881271327331e-06, + "loss": 0.6484, + "step": 3595 + }, + { + "epoch": 1.3771778671261727, + "grad_norm": 0.3089297413825989, + "learning_rate": 2.8748771408776467e-06, + "loss": 0.6276, + "step": 3596 + }, + { + "epoch": 1.3775607888186867, + "grad_norm": 0.31109708547592163, + "learning_rate": 2.8738729485561735e-06, + "loss": 0.6005, + "step": 3597 + }, + { + "epoch": 1.3779437105112005, + "grad_norm": 0.3075871169567108, + "learning_rate": 2.872868694528647e-06, + "loss": 0.6389, + "step": 3598 + }, + { + "epoch": 1.3783266322037142, + "grad_norm": 0.30176427960395813, + "learning_rate": 2.8718643789608163e-06, + "loss": 0.6393, + "step": 3599 + }, + { + "epoch": 1.3787095538962282, + "grad_norm": 0.3020285367965698, + "learning_rate": 2.870860002018439e-06, + "loss": 0.6641, + "step": 3600 + }, + { + "epoch": 1.3790924755887422, + "grad_norm": 0.302559494972229, + "learning_rate": 2.8698555638672826e-06, + "loss": 0.6176, + "step": 3601 + }, + { + "epoch": 1.3794753972812561, + "grad_norm": 0.3106479048728943, + "learning_rate": 2.8688510646731253e-06, + "loss": 0.6323, + "step": 3602 + }, + { + "epoch": 1.3798583189737699, + "grad_norm": 0.3018491268157959, + "learning_rate": 2.8678465046017555e-06, + "loss": 0.6212, + "step": 3603 + }, + { + "epoch": 1.3802412406662836, + "grad_norm": 0.3049074709415436, + "learning_rate": 2.866841883818971e-06, + "loss": 0.635, + "step": 3604 + }, + { + "epoch": 1.3806241623587976, + "grad_norm": 0.30644792318344116, + "learning_rate": 2.86583720249058e-06, + "loss": 0.6325, + "step": 3605 + }, + { + "epoch": 1.3810070840513116, + "grad_norm": 0.31348636746406555, + "learning_rate": 2.8648324607824e-06, + "loss": 0.6221, + "step": 3606 + }, + { + "epoch": 1.3813900057438253, + "grad_norm": 0.3000335693359375, + "learning_rate": 2.86382765886026e-06, + "loss": 0.6328, + "step": 3607 + }, + { + "epoch": 1.3817729274363393, + "grad_norm": 0.3158789277076721, + "learning_rate": 2.8628227968899987e-06, + "loss": 0.6333, + "step": 3608 + }, + { + "epoch": 1.382155849128853, + "grad_norm": 0.29982367157936096, + "learning_rate": 2.861817875037462e-06, + "loss": 0.6395, + "step": 3609 + }, + { + "epoch": 1.382538770821367, + "grad_norm": 0.30386263132095337, + "learning_rate": 2.86081289346851e-06, + "loss": 0.6038, + "step": 3610 + }, + { + "epoch": 1.382921692513881, + "grad_norm": 0.32542338967323303, + "learning_rate": 2.8598078523490087e-06, + "loss": 0.6105, + "step": 3611 + }, + { + "epoch": 1.3833046142063947, + "grad_norm": 0.3099135160446167, + "learning_rate": 2.8588027518448357e-06, + "loss": 0.6375, + "step": 3612 + }, + { + "epoch": 1.3836875358989087, + "grad_norm": 0.2956951856613159, + "learning_rate": 2.8577975921218793e-06, + "loss": 0.6096, + "step": 3613 + }, + { + "epoch": 1.3840704575914224, + "grad_norm": 0.3071860074996948, + "learning_rate": 2.8567923733460355e-06, + "loss": 0.6018, + "step": 3614 + }, + { + "epoch": 1.3844533792839364, + "grad_norm": 0.3014329969882965, + "learning_rate": 2.8557870956832135e-06, + "loss": 0.6322, + "step": 3615 + }, + { + "epoch": 1.3848363009764504, + "grad_norm": 0.3095967769622803, + "learning_rate": 2.8547817592993264e-06, + "loss": 0.625, + "step": 3616 + }, + { + "epoch": 1.3852192226689641, + "grad_norm": 0.29387566447257996, + "learning_rate": 2.853776364360303e-06, + "loss": 0.6234, + "step": 3617 + }, + { + "epoch": 1.385602144361478, + "grad_norm": 0.3122255504131317, + "learning_rate": 2.852770911032079e-06, + "loss": 0.6313, + "step": 3618 + }, + { + "epoch": 1.3859850660539919, + "grad_norm": 0.30043843388557434, + "learning_rate": 2.8517653994805984e-06, + "loss": 0.6347, + "step": 3619 + }, + { + "epoch": 1.3863679877465058, + "grad_norm": 0.3086322247982025, + "learning_rate": 2.850759829871818e-06, + "loss": 0.6398, + "step": 3620 + }, + { + "epoch": 1.3867509094390198, + "grad_norm": 0.30852675437927246, + "learning_rate": 2.8497542023717013e-06, + "loss": 0.6317, + "step": 3621 + }, + { + "epoch": 1.3871338311315335, + "grad_norm": 0.3191261887550354, + "learning_rate": 2.8487485171462238e-06, + "loss": 0.6389, + "step": 3622 + }, + { + "epoch": 1.3875167528240475, + "grad_norm": 0.3085120618343353, + "learning_rate": 2.847742774361368e-06, + "loss": 0.6318, + "step": 3623 + }, + { + "epoch": 1.3878996745165613, + "grad_norm": 0.3133058249950409, + "learning_rate": 2.8467369741831284e-06, + "loss": 0.6202, + "step": 3624 + }, + { + "epoch": 1.3882825962090752, + "grad_norm": 0.2965955436229706, + "learning_rate": 2.845731116777509e-06, + "loss": 0.6093, + "step": 3625 + }, + { + "epoch": 1.3886655179015892, + "grad_norm": 0.2997977137565613, + "learning_rate": 2.8447252023105183e-06, + "loss": 0.6123, + "step": 3626 + }, + { + "epoch": 1.389048439594103, + "grad_norm": 0.31416189670562744, + "learning_rate": 2.843719230948181e-06, + "loss": 0.6441, + "step": 3627 + }, + { + "epoch": 1.389431361286617, + "grad_norm": 0.32525116205215454, + "learning_rate": 2.8427132028565262e-06, + "loss": 0.6255, + "step": 3628 + }, + { + "epoch": 1.3898142829791307, + "grad_norm": 0.3140980303287506, + "learning_rate": 2.8417071182015973e-06, + "loss": 0.6277, + "step": 3629 + }, + { + "epoch": 1.3901972046716446, + "grad_norm": 0.313581645488739, + "learning_rate": 2.840700977149441e-06, + "loss": 0.6342, + "step": 3630 + }, + { + "epoch": 1.3905801263641586, + "grad_norm": 0.32624226808547974, + "learning_rate": 2.839694779866117e-06, + "loss": 0.6612, + "step": 3631 + }, + { + "epoch": 1.3909630480566724, + "grad_norm": 0.34412962198257446, + "learning_rate": 2.8386885265176956e-06, + "loss": 0.6423, + "step": 3632 + }, + { + "epoch": 1.3913459697491863, + "grad_norm": 0.34408631920814514, + "learning_rate": 2.837682217270252e-06, + "loss": 0.6204, + "step": 3633 + }, + { + "epoch": 1.3917288914417, + "grad_norm": 0.30787500739097595, + "learning_rate": 2.8366758522898737e-06, + "loss": 0.6415, + "step": 3634 + }, + { + "epoch": 1.392111813134214, + "grad_norm": 0.3133569657802582, + "learning_rate": 2.8356694317426577e-06, + "loss": 0.6473, + "step": 3635 + }, + { + "epoch": 1.392494734826728, + "grad_norm": 0.3492833971977234, + "learning_rate": 2.834662955794708e-06, + "loss": 0.6279, + "step": 3636 + }, + { + "epoch": 1.3928776565192418, + "grad_norm": 0.3177170753479004, + "learning_rate": 2.83365642461214e-06, + "loss": 0.5995, + "step": 3637 + }, + { + "epoch": 1.3932605782117558, + "grad_norm": 0.31068146228790283, + "learning_rate": 2.8326498383610757e-06, + "loss": 0.6459, + "step": 3638 + }, + { + "epoch": 1.3936434999042695, + "grad_norm": 0.31614750623703003, + "learning_rate": 2.8316431972076497e-06, + "loss": 0.6196, + "step": 3639 + }, + { + "epoch": 1.3940264215967835, + "grad_norm": 0.44080477952957153, + "learning_rate": 2.8306365013180014e-06, + "loss": 0.6059, + "step": 3640 + }, + { + "epoch": 1.3944093432892974, + "grad_norm": 0.3008778989315033, + "learning_rate": 2.8296297508582825e-06, + "loss": 0.6001, + "step": 3641 + }, + { + "epoch": 1.3947922649818112, + "grad_norm": 0.31189730763435364, + "learning_rate": 2.8286229459946534e-06, + "loss": 0.6656, + "step": 3642 + }, + { + "epoch": 1.3951751866743252, + "grad_norm": 0.3171970248222351, + "learning_rate": 2.827616086893282e-06, + "loss": 0.6302, + "step": 3643 + }, + { + "epoch": 1.395558108366839, + "grad_norm": 0.32453832030296326, + "learning_rate": 2.8266091737203455e-06, + "loss": 0.6315, + "step": 3644 + }, + { + "epoch": 1.3959410300593529, + "grad_norm": 0.32535889744758606, + "learning_rate": 2.82560220664203e-06, + "loss": 0.6264, + "step": 3645 + }, + { + "epoch": 1.3963239517518669, + "grad_norm": 0.30627521872520447, + "learning_rate": 2.8245951858245326e-06, + "loss": 0.6497, + "step": 3646 + }, + { + "epoch": 1.3967068734443806, + "grad_norm": 0.3225005567073822, + "learning_rate": 2.8235881114340567e-06, + "loss": 0.6307, + "step": 3647 + }, + { + "epoch": 1.3970897951368946, + "grad_norm": 0.34354519844055176, + "learning_rate": 2.8225809836368146e-06, + "loss": 0.6249, + "step": 3648 + }, + { + "epoch": 1.3974727168294083, + "grad_norm": 0.324582040309906, + "learning_rate": 2.821573802599029e-06, + "loss": 0.6093, + "step": 3649 + }, + { + "epoch": 1.3978556385219223, + "grad_norm": 0.3119044005870819, + "learning_rate": 2.820566568486931e-06, + "loss": 0.631, + "step": 3650 + }, + { + "epoch": 1.3982385602144363, + "grad_norm": 0.2936842143535614, + "learning_rate": 2.8195592814667593e-06, + "loss": 0.6264, + "step": 3651 + }, + { + "epoch": 1.39862148190695, + "grad_norm": 0.33680272102355957, + "learning_rate": 2.8185519417047624e-06, + "loss": 0.6306, + "step": 3652 + }, + { + "epoch": 1.399004403599464, + "grad_norm": 0.31987708806991577, + "learning_rate": 2.817544549367197e-06, + "loss": 0.6385, + "step": 3653 + }, + { + "epoch": 1.3993873252919777, + "grad_norm": 0.32836776971817017, + "learning_rate": 2.81653710462033e-06, + "loss": 0.628, + "step": 3654 + }, + { + "epoch": 1.3997702469844917, + "grad_norm": 0.2968779504299164, + "learning_rate": 2.815529607630433e-06, + "loss": 0.6246, + "step": 3655 + }, + { + "epoch": 1.4001531686770057, + "grad_norm": 0.31670016050338745, + "learning_rate": 2.814522058563791e-06, + "loss": 0.6338, + "step": 3656 + }, + { + "epoch": 1.4005360903695194, + "grad_norm": 0.34047868847846985, + "learning_rate": 2.8135144575866953e-06, + "loss": 0.6199, + "step": 3657 + }, + { + "epoch": 1.4009190120620332, + "grad_norm": 0.2917149066925049, + "learning_rate": 2.8125068048654448e-06, + "loss": 0.6224, + "step": 3658 + }, + { + "epoch": 1.4013019337545471, + "grad_norm": 0.3035246431827545, + "learning_rate": 2.811499100566349e-06, + "loss": 0.6381, + "step": 3659 + }, + { + "epoch": 1.4016848554470611, + "grad_norm": 0.32511311769485474, + "learning_rate": 2.8104913448557243e-06, + "loss": 0.6086, + "step": 3660 + }, + { + "epoch": 1.402067777139575, + "grad_norm": 0.3061164617538452, + "learning_rate": 2.809483537899897e-06, + "loss": 0.6352, + "step": 3661 + }, + { + "epoch": 1.4024506988320888, + "grad_norm": 0.30962204933166504, + "learning_rate": 2.808475679865201e-06, + "loss": 0.6303, + "step": 3662 + }, + { + "epoch": 1.4028336205246026, + "grad_norm": 0.30887526273727417, + "learning_rate": 2.8074677709179775e-06, + "loss": 0.6348, + "step": 3663 + }, + { + "epoch": 1.4032165422171166, + "grad_norm": 0.30785197019577026, + "learning_rate": 2.80645981122458e-06, + "loss": 0.6488, + "step": 3664 + }, + { + "epoch": 1.4035994639096305, + "grad_norm": 0.3148612082004547, + "learning_rate": 2.8054518009513647e-06, + "loss": 0.6466, + "step": 3665 + }, + { + "epoch": 1.4039823856021443, + "grad_norm": 0.31565338373184204, + "learning_rate": 2.8044437402647012e-06, + "loss": 0.6363, + "step": 3666 + }, + { + "epoch": 1.4043653072946582, + "grad_norm": 0.3060840368270874, + "learning_rate": 2.8034356293309647e-06, + "loss": 0.6393, + "step": 3667 + }, + { + "epoch": 1.404748228987172, + "grad_norm": 0.29679277539253235, + "learning_rate": 2.8024274683165385e-06, + "loss": 0.6189, + "step": 3668 + }, + { + "epoch": 1.405131150679686, + "grad_norm": 0.3113795220851898, + "learning_rate": 2.8014192573878173e-06, + "loss": 0.6391, + "step": 3669 + }, + { + "epoch": 1.4055140723722, + "grad_norm": 0.3047578036785126, + "learning_rate": 2.8004109967111998e-06, + "loss": 0.6368, + "step": 3670 + }, + { + "epoch": 1.4058969940647137, + "grad_norm": 0.30446192622184753, + "learning_rate": 2.799402686453096e-06, + "loss": 0.623, + "step": 3671 + }, + { + "epoch": 1.4062799157572277, + "grad_norm": 0.31138619780540466, + "learning_rate": 2.798394326779922e-06, + "loss": 0.6289, + "step": 3672 + }, + { + "epoch": 1.4066628374497414, + "grad_norm": 0.3020646572113037, + "learning_rate": 2.797385917858103e-06, + "loss": 0.6186, + "step": 3673 + }, + { + "epoch": 1.4070457591422554, + "grad_norm": 0.3022039830684662, + "learning_rate": 2.7963774598540744e-06, + "loss": 0.609, + "step": 3674 + }, + { + "epoch": 1.4074286808347694, + "grad_norm": 0.2991883456707001, + "learning_rate": 2.795368952934275e-06, + "loss": 0.6183, + "step": 3675 + }, + { + "epoch": 1.407811602527283, + "grad_norm": 0.30065852403640747, + "learning_rate": 2.7943603972651566e-06, + "loss": 0.628, + "step": 3676 + }, + { + "epoch": 1.408194524219797, + "grad_norm": 0.3056091070175171, + "learning_rate": 2.7933517930131753e-06, + "loss": 0.6517, + "step": 3677 + }, + { + "epoch": 1.4085774459123108, + "grad_norm": 0.29916757345199585, + "learning_rate": 2.792343140344797e-06, + "loss": 0.636, + "step": 3678 + }, + { + "epoch": 1.4089603676048248, + "grad_norm": 0.2996805012226105, + "learning_rate": 2.791334439426496e-06, + "loss": 0.6259, + "step": 3679 + }, + { + "epoch": 1.4093432892973388, + "grad_norm": 0.29919224977493286, + "learning_rate": 2.7903256904247523e-06, + "loss": 0.6499, + "step": 3680 + }, + { + "epoch": 1.4097262109898525, + "grad_norm": 0.31093913316726685, + "learning_rate": 2.789316893506057e-06, + "loss": 0.6319, + "step": 3681 + }, + { + "epoch": 1.4101091326823665, + "grad_norm": 0.3069301247596741, + "learning_rate": 2.7883080488369075e-06, + "loss": 0.6161, + "step": 3682 + }, + { + "epoch": 1.4104920543748802, + "grad_norm": 0.3003811240196228, + "learning_rate": 2.7872991565838076e-06, + "loss": 0.616, + "step": 3683 + }, + { + "epoch": 1.4108749760673942, + "grad_norm": 0.2942151129245758, + "learning_rate": 2.786290216913271e-06, + "loss": 0.6267, + "step": 3684 + }, + { + "epoch": 1.4112578977599082, + "grad_norm": 0.30369260907173157, + "learning_rate": 2.78528122999182e-06, + "loss": 0.6348, + "step": 3685 + }, + { + "epoch": 1.411640819452422, + "grad_norm": 0.30487385392189026, + "learning_rate": 2.784272195985982e-06, + "loss": 0.6212, + "step": 3686 + }, + { + "epoch": 1.412023741144936, + "grad_norm": 0.3080061674118042, + "learning_rate": 2.7832631150622936e-06, + "loss": 0.6433, + "step": 3687 + }, + { + "epoch": 1.4124066628374496, + "grad_norm": 0.3034511208534241, + "learning_rate": 2.7822539873873e-06, + "loss": 0.6541, + "step": 3688 + }, + { + "epoch": 1.4127895845299636, + "grad_norm": 0.30441346764564514, + "learning_rate": 2.781244813127552e-06, + "loss": 0.6327, + "step": 3689 + }, + { + "epoch": 1.4131725062224776, + "grad_norm": 0.30653446912765503, + "learning_rate": 2.7802355924496098e-06, + "loss": 0.6464, + "step": 3690 + }, + { + "epoch": 1.4135554279149913, + "grad_norm": 0.30547547340393066, + "learning_rate": 2.7792263255200406e-06, + "loss": 0.6231, + "step": 3691 + }, + { + "epoch": 1.4139383496075053, + "grad_norm": 0.3045615255832672, + "learning_rate": 2.7782170125054197e-06, + "loss": 0.6205, + "step": 3692 + }, + { + "epoch": 1.414321271300019, + "grad_norm": 0.30261164903640747, + "learning_rate": 2.777207653572331e-06, + "loss": 0.6339, + "step": 3693 + }, + { + "epoch": 1.414704192992533, + "grad_norm": 0.30321750044822693, + "learning_rate": 2.7761982488873612e-06, + "loss": 0.6164, + "step": 3694 + }, + { + "epoch": 1.415087114685047, + "grad_norm": 0.30724939703941345, + "learning_rate": 2.7751887986171113e-06, + "loss": 0.6309, + "step": 3695 + }, + { + "epoch": 1.4154700363775607, + "grad_norm": 0.3157292306423187, + "learning_rate": 2.7741793029281848e-06, + "loss": 0.6431, + "step": 3696 + }, + { + "epoch": 1.4158529580700747, + "grad_norm": 0.3035392463207245, + "learning_rate": 2.7731697619871943e-06, + "loss": 0.646, + "step": 3697 + }, + { + "epoch": 1.4162358797625885, + "grad_norm": 0.31166937947273254, + "learning_rate": 2.7721601759607615e-06, + "loss": 0.6319, + "step": 3698 + }, + { + "epoch": 1.4166188014551024, + "grad_norm": 0.31186115741729736, + "learning_rate": 2.7711505450155125e-06, + "loss": 0.6303, + "step": 3699 + }, + { + "epoch": 1.4170017231476164, + "grad_norm": 0.3044099509716034, + "learning_rate": 2.7701408693180836e-06, + "loss": 0.6298, + "step": 3700 + }, + { + "epoch": 1.4173846448401302, + "grad_norm": 0.30870771408081055, + "learning_rate": 2.7691311490351164e-06, + "loss": 0.6404, + "step": 3701 + }, + { + "epoch": 1.4177675665326441, + "grad_norm": 0.31779536604881287, + "learning_rate": 2.7681213843332603e-06, + "loss": 0.6219, + "step": 3702 + }, + { + "epoch": 1.4181504882251579, + "grad_norm": 0.3054141402244568, + "learning_rate": 2.7671115753791746e-06, + "loss": 0.6228, + "step": 3703 + }, + { + "epoch": 1.4185334099176719, + "grad_norm": 0.3040907084941864, + "learning_rate": 2.766101722339521e-06, + "loss": 0.6205, + "step": 3704 + }, + { + "epoch": 1.4189163316101858, + "grad_norm": 0.29814836382865906, + "learning_rate": 2.765091825380972e-06, + "loss": 0.6223, + "step": 3705 + }, + { + "epoch": 1.4192992533026996, + "grad_norm": 0.3175666928291321, + "learning_rate": 2.764081884670208e-06, + "loss": 0.6132, + "step": 3706 + }, + { + "epoch": 1.4196821749952135, + "grad_norm": 0.31555232405662537, + "learning_rate": 2.763071900373914e-06, + "loss": 0.637, + "step": 3707 + }, + { + "epoch": 1.4200650966877273, + "grad_norm": 0.31552639603614807, + "learning_rate": 2.7620618726587832e-06, + "loss": 0.6478, + "step": 3708 + }, + { + "epoch": 1.4204480183802413, + "grad_norm": 0.33246198296546936, + "learning_rate": 2.761051801691516e-06, + "loss": 0.6172, + "step": 3709 + }, + { + "epoch": 1.4208309400727552, + "grad_norm": 0.317823588848114, + "learning_rate": 2.760041687638822e-06, + "loss": 0.6438, + "step": 3710 + }, + { + "epoch": 1.421213861765269, + "grad_norm": 0.3148415982723236, + "learning_rate": 2.7590315306674133e-06, + "loss": 0.6267, + "step": 3711 + }, + { + "epoch": 1.421596783457783, + "grad_norm": 0.3049493134021759, + "learning_rate": 2.7580213309440133e-06, + "loss": 0.6373, + "step": 3712 + }, + { + "epoch": 1.4219797051502967, + "grad_norm": 0.31461092829704285, + "learning_rate": 2.7570110886353505e-06, + "loss": 0.6324, + "step": 3713 + }, + { + "epoch": 1.4223626268428107, + "grad_norm": 0.31016841530799866, + "learning_rate": 2.7560008039081616e-06, + "loss": 0.6225, + "step": 3714 + }, + { + "epoch": 1.4227455485353246, + "grad_norm": 0.301126092672348, + "learning_rate": 2.7549904769291884e-06, + "loss": 0.6428, + "step": 3715 + }, + { + "epoch": 1.4231284702278384, + "grad_norm": 0.3044571876525879, + "learning_rate": 2.7539801078651812e-06, + "loss": 0.6197, + "step": 3716 + }, + { + "epoch": 1.4235113919203524, + "grad_norm": 0.3122325837612152, + "learning_rate": 2.752969696882898e-06, + "loss": 0.6215, + "step": 3717 + }, + { + "epoch": 1.4238943136128661, + "grad_norm": 0.31531381607055664, + "learning_rate": 2.751959244149101e-06, + "loss": 0.627, + "step": 3718 + }, + { + "epoch": 1.42427723530538, + "grad_norm": 0.30565524101257324, + "learning_rate": 2.7509487498305615e-06, + "loss": 0.6454, + "step": 3719 + }, + { + "epoch": 1.424660156997894, + "grad_norm": 0.3023408055305481, + "learning_rate": 2.749938214094058e-06, + "loss": 0.6124, + "step": 3720 + }, + { + "epoch": 1.4250430786904078, + "grad_norm": 0.31364941596984863, + "learning_rate": 2.748927637106373e-06, + "loss": 0.635, + "step": 3721 + }, + { + "epoch": 1.4254260003829216, + "grad_norm": 0.30580320954322815, + "learning_rate": 2.7479170190342997e-06, + "loss": 0.6205, + "step": 3722 + }, + { + "epoch": 1.4258089220754355, + "grad_norm": 0.308299720287323, + "learning_rate": 2.7469063600446337e-06, + "loss": 0.6349, + "step": 3723 + }, + { + "epoch": 1.4261918437679495, + "grad_norm": 0.31228312849998474, + "learning_rate": 2.745895660304182e-06, + "loss": 0.6484, + "step": 3724 + }, + { + "epoch": 1.4265747654604635, + "grad_norm": 0.3123389482498169, + "learning_rate": 2.744884919979755e-06, + "loss": 0.6389, + "step": 3725 + }, + { + "epoch": 1.4269576871529772, + "grad_norm": 0.32960495352745056, + "learning_rate": 2.743874139238171e-06, + "loss": 0.6336, + "step": 3726 + }, + { + "epoch": 1.427340608845491, + "grad_norm": 0.30544236302375793, + "learning_rate": 2.742863318246255e-06, + "loss": 0.629, + "step": 3727 + }, + { + "epoch": 1.427723530538005, + "grad_norm": 0.3022859990596771, + "learning_rate": 2.7418524571708383e-06, + "loss": 0.6492, + "step": 3728 + }, + { + "epoch": 1.428106452230519, + "grad_norm": 0.32605376839637756, + "learning_rate": 2.7408415561787587e-06, + "loss": 0.6303, + "step": 3729 + }, + { + "epoch": 1.4284893739230327, + "grad_norm": 0.3320281505584717, + "learning_rate": 2.7398306154368613e-06, + "loss": 0.6261, + "step": 3730 + }, + { + "epoch": 1.4288722956155466, + "grad_norm": 0.312447190284729, + "learning_rate": 2.7388196351119974e-06, + "loss": 0.6282, + "step": 3731 + }, + { + "epoch": 1.4292552173080604, + "grad_norm": 0.3128534257411957, + "learning_rate": 2.737808615371025e-06, + "loss": 0.6319, + "step": 3732 + }, + { + "epoch": 1.4296381390005743, + "grad_norm": 0.30351343750953674, + "learning_rate": 2.7367975563808068e-06, + "loss": 0.625, + "step": 3733 + }, + { + "epoch": 1.4300210606930883, + "grad_norm": 0.31596967577934265, + "learning_rate": 2.7357864583082157e-06, + "loss": 0.6369, + "step": 3734 + }, + { + "epoch": 1.430403982385602, + "grad_norm": 0.33387261629104614, + "learning_rate": 2.7347753213201274e-06, + "loss": 0.6326, + "step": 3735 + }, + { + "epoch": 1.430786904078116, + "grad_norm": 0.3074447214603424, + "learning_rate": 2.7337641455834255e-06, + "loss": 0.638, + "step": 3736 + }, + { + "epoch": 1.4311698257706298, + "grad_norm": 0.3196054697036743, + "learning_rate": 2.7327529312650014e-06, + "loss": 0.6348, + "step": 3737 + }, + { + "epoch": 1.4315527474631438, + "grad_norm": 0.31072962284088135, + "learning_rate": 2.7317416785317497e-06, + "loss": 0.6541, + "step": 3738 + }, + { + "epoch": 1.4319356691556577, + "grad_norm": 0.31064823269844055, + "learning_rate": 2.7307303875505754e-06, + "loss": 0.6445, + "step": 3739 + }, + { + "epoch": 1.4323185908481715, + "grad_norm": 0.3014507293701172, + "learning_rate": 2.729719058488385e-06, + "loss": 0.6149, + "step": 3740 + }, + { + "epoch": 1.4327015125406855, + "grad_norm": 0.30648329854011536, + "learning_rate": 2.7287076915120952e-06, + "loss": 0.6398, + "step": 3741 + }, + { + "epoch": 1.4330844342331992, + "grad_norm": 0.3069413900375366, + "learning_rate": 2.727696286788628e-06, + "loss": 0.6313, + "step": 3742 + }, + { + "epoch": 1.4334673559257132, + "grad_norm": 0.30560827255249023, + "learning_rate": 2.7266848444849096e-06, + "loss": 0.6334, + "step": 3743 + }, + { + "epoch": 1.4338502776182271, + "grad_norm": 0.31910544633865356, + "learning_rate": 2.7256733647678764e-06, + "loss": 0.6363, + "step": 3744 + }, + { + "epoch": 1.434233199310741, + "grad_norm": 0.29997876286506653, + "learning_rate": 2.7246618478044654e-06, + "loss": 0.6366, + "step": 3745 + }, + { + "epoch": 1.4346161210032549, + "grad_norm": 0.3046404719352722, + "learning_rate": 2.723650293761626e-06, + "loss": 0.6379, + "step": 3746 + }, + { + "epoch": 1.4349990426957686, + "grad_norm": 0.3062083125114441, + "learning_rate": 2.722638702806309e-06, + "loss": 0.6323, + "step": 3747 + }, + { + "epoch": 1.4353819643882826, + "grad_norm": 0.3127845227718353, + "learning_rate": 2.721627075105473e-06, + "loss": 0.6357, + "step": 3748 + }, + { + "epoch": 1.4357648860807966, + "grad_norm": 0.3066936731338501, + "learning_rate": 2.720615410826084e-06, + "loss": 0.6347, + "step": 3749 + }, + { + "epoch": 1.4361478077733103, + "grad_norm": 0.31431758403778076, + "learning_rate": 2.719603710135111e-06, + "loss": 0.6159, + "step": 3750 + }, + { + "epoch": 1.4365307294658243, + "grad_norm": 0.3121020197868347, + "learning_rate": 2.718591973199531e-06, + "loss": 0.6314, + "step": 3751 + }, + { + "epoch": 1.436913651158338, + "grad_norm": 0.3162870705127716, + "learning_rate": 2.7175802001863267e-06, + "loss": 0.6223, + "step": 3752 + }, + { + "epoch": 1.437296572850852, + "grad_norm": 0.30316686630249023, + "learning_rate": 2.7165683912624872e-06, + "loss": 0.6184, + "step": 3753 + }, + { + "epoch": 1.437679494543366, + "grad_norm": 0.29638609290122986, + "learning_rate": 2.7155565465950068e-06, + "loss": 0.6356, + "step": 3754 + }, + { + "epoch": 1.4380624162358797, + "grad_norm": 0.30067428946495056, + "learning_rate": 2.714544666350885e-06, + "loss": 0.6244, + "step": 3755 + }, + { + "epoch": 1.4384453379283937, + "grad_norm": 0.2970254421234131, + "learning_rate": 2.7135327506971297e-06, + "loss": 0.6242, + "step": 3756 + }, + { + "epoch": 1.4388282596209074, + "grad_norm": 0.30594199895858765, + "learning_rate": 2.7125207998007524e-06, + "loss": 0.6545, + "step": 3757 + }, + { + "epoch": 1.4392111813134214, + "grad_norm": 0.3075084686279297, + "learning_rate": 2.7115088138287704e-06, + "loss": 0.6316, + "step": 3758 + }, + { + "epoch": 1.4395941030059354, + "grad_norm": 0.3194836378097534, + "learning_rate": 2.7104967929482084e-06, + "loss": 0.6415, + "step": 3759 + }, + { + "epoch": 1.4399770246984491, + "grad_norm": 0.3027834892272949, + "learning_rate": 2.709484737326096e-06, + "loss": 0.6275, + "step": 3760 + }, + { + "epoch": 1.440359946390963, + "grad_norm": 0.3120865225791931, + "learning_rate": 2.708472647129467e-06, + "loss": 0.6587, + "step": 3761 + }, + { + "epoch": 1.4407428680834768, + "grad_norm": 0.31038805842399597, + "learning_rate": 2.707460522525363e-06, + "loss": 0.607, + "step": 3762 + }, + { + "epoch": 1.4411257897759908, + "grad_norm": 0.3079681694507599, + "learning_rate": 2.7064483636808314e-06, + "loss": 0.6399, + "step": 3763 + }, + { + "epoch": 1.4415087114685048, + "grad_norm": 0.3164820969104767, + "learning_rate": 2.705436170762924e-06, + "loss": 0.6344, + "step": 3764 + }, + { + "epoch": 1.4418916331610185, + "grad_norm": 0.3042416572570801, + "learning_rate": 2.7044239439386976e-06, + "loss": 0.6337, + "step": 3765 + }, + { + "epoch": 1.4422745548535325, + "grad_norm": 0.2997894287109375, + "learning_rate": 2.7034116833752178e-06, + "loss": 0.645, + "step": 3766 + }, + { + "epoch": 1.4426574765460463, + "grad_norm": 0.30265265703201294, + "learning_rate": 2.7023993892395523e-06, + "loss": 0.6347, + "step": 3767 + }, + { + "epoch": 1.4430403982385602, + "grad_norm": 0.31644943356513977, + "learning_rate": 2.701387061698776e-06, + "loss": 0.6222, + "step": 3768 + }, + { + "epoch": 1.4434233199310742, + "grad_norm": 0.30674806237220764, + "learning_rate": 2.700374700919969e-06, + "loss": 0.627, + "step": 3769 + }, + { + "epoch": 1.443806241623588, + "grad_norm": 0.3116602897644043, + "learning_rate": 2.699362307070217e-06, + "loss": 0.6347, + "step": 3770 + }, + { + "epoch": 1.444189163316102, + "grad_norm": 0.3028624355792999, + "learning_rate": 2.698349880316611e-06, + "loss": 0.6488, + "step": 3771 + }, + { + "epoch": 1.4445720850086157, + "grad_norm": 0.3009147346019745, + "learning_rate": 2.697337420826247e-06, + "loss": 0.6535, + "step": 3772 + }, + { + "epoch": 1.4449550067011296, + "grad_norm": 0.29710546135902405, + "learning_rate": 2.696324928766228e-06, + "loss": 0.6165, + "step": 3773 + }, + { + "epoch": 1.4453379283936436, + "grad_norm": 0.3117179274559021, + "learning_rate": 2.6953124043036604e-06, + "loss": 0.638, + "step": 3774 + }, + { + "epoch": 1.4457208500861574, + "grad_norm": 0.29840564727783203, + "learning_rate": 2.6942998476056563e-06, + "loss": 0.6153, + "step": 3775 + }, + { + "epoch": 1.4461037717786713, + "grad_norm": 0.3000156283378601, + "learning_rate": 2.6932872588393354e-06, + "loss": 0.6201, + "step": 3776 + }, + { + "epoch": 1.446486693471185, + "grad_norm": 0.307381808757782, + "learning_rate": 2.6922746381718195e-06, + "loss": 0.6502, + "step": 3777 + }, + { + "epoch": 1.446869615163699, + "grad_norm": 0.30113545060157776, + "learning_rate": 2.6912619857702384e-06, + "loss": 0.6316, + "step": 3778 + }, + { + "epoch": 1.447252536856213, + "grad_norm": 0.3162819743156433, + "learning_rate": 2.6902493018017236e-06, + "loss": 0.6313, + "step": 3779 + }, + { + "epoch": 1.4476354585487268, + "grad_norm": 0.29214099049568176, + "learning_rate": 2.6892365864334157e-06, + "loss": 0.6171, + "step": 3780 + }, + { + "epoch": 1.4480183802412407, + "grad_norm": 0.3060738444328308, + "learning_rate": 2.6882238398324595e-06, + "loss": 0.6252, + "step": 3781 + }, + { + "epoch": 1.4484013019337545, + "grad_norm": 0.3071605861186981, + "learning_rate": 2.687211062166002e-06, + "loss": 0.6298, + "step": 3782 + }, + { + "epoch": 1.4487842236262685, + "grad_norm": 0.3142364025115967, + "learning_rate": 2.6861982536012e-06, + "loss": 0.6388, + "step": 3783 + }, + { + "epoch": 1.4491671453187824, + "grad_norm": 0.3041272461414337, + "learning_rate": 2.685185414305211e-06, + "loss": 0.6376, + "step": 3784 + }, + { + "epoch": 1.4495500670112962, + "grad_norm": 0.31094521284103394, + "learning_rate": 2.6841725444452005e-06, + "loss": 0.6087, + "step": 3785 + }, + { + "epoch": 1.44993298870381, + "grad_norm": 0.30806612968444824, + "learning_rate": 2.6831596441883388e-06, + "loss": 0.6336, + "step": 3786 + }, + { + "epoch": 1.450315910396324, + "grad_norm": 0.30039775371551514, + "learning_rate": 2.682146713701799e-06, + "loss": 0.6184, + "step": 3787 + }, + { + "epoch": 1.4506988320888379, + "grad_norm": 0.31023460626602173, + "learning_rate": 2.6811337531527632e-06, + "loss": 0.6197, + "step": 3788 + }, + { + "epoch": 1.4510817537813518, + "grad_norm": 0.31085115671157837, + "learning_rate": 2.680120762708413e-06, + "loss": 0.6158, + "step": 3789 + }, + { + "epoch": 1.4514646754738656, + "grad_norm": 0.3125998079776764, + "learning_rate": 2.6791077425359396e-06, + "loss": 0.6251, + "step": 3790 + }, + { + "epoch": 1.4518475971663793, + "grad_norm": 0.31331029534339905, + "learning_rate": 2.678094692802537e-06, + "loss": 0.5992, + "step": 3791 + }, + { + "epoch": 1.4522305188588933, + "grad_norm": 0.30863407254219055, + "learning_rate": 2.6770816136754053e-06, + "loss": 0.6135, + "step": 3792 + }, + { + "epoch": 1.4526134405514073, + "grad_norm": 0.31204918026924133, + "learning_rate": 2.6760685053217472e-06, + "loss": 0.6272, + "step": 3793 + }, + { + "epoch": 1.452996362243921, + "grad_norm": 0.3123208284378052, + "learning_rate": 2.6750553679087728e-06, + "loss": 0.6422, + "step": 3794 + }, + { + "epoch": 1.453379283936435, + "grad_norm": 0.30645641684532166, + "learning_rate": 2.674042201603696e-06, + "loss": 0.6395, + "step": 3795 + }, + { + "epoch": 1.4537622056289488, + "grad_norm": 0.3000721037387848, + "learning_rate": 2.673029006573735e-06, + "loss": 0.6215, + "step": 3796 + }, + { + "epoch": 1.4541451273214627, + "grad_norm": 0.3077699840068817, + "learning_rate": 2.6720157829861114e-06, + "loss": 0.6358, + "step": 3797 + }, + { + "epoch": 1.4545280490139767, + "grad_norm": 0.30700695514678955, + "learning_rate": 2.671002531008057e-06, + "loss": 0.6337, + "step": 3798 + }, + { + "epoch": 1.4549109707064904, + "grad_norm": 0.2980685830116272, + "learning_rate": 2.6699892508068015e-06, + "loss": 0.6376, + "step": 3799 + }, + { + "epoch": 1.4552938923990044, + "grad_norm": 0.3070155680179596, + "learning_rate": 2.6689759425495833e-06, + "loss": 0.6438, + "step": 3800 + }, + { + "epoch": 1.4556768140915182, + "grad_norm": 0.30220094323158264, + "learning_rate": 2.6679626064036428e-06, + "loss": 0.6208, + "step": 3801 + }, + { + "epoch": 1.4560597357840321, + "grad_norm": 0.314609169960022, + "learning_rate": 2.6669492425362294e-06, + "loss": 0.6347, + "step": 3802 + }, + { + "epoch": 1.456442657476546, + "grad_norm": 0.30625787377357483, + "learning_rate": 2.665935851114593e-06, + "loss": 0.6369, + "step": 3803 + }, + { + "epoch": 1.4568255791690599, + "grad_norm": 0.2979176938533783, + "learning_rate": 2.6649224323059874e-06, + "loss": 0.6294, + "step": 3804 + }, + { + "epoch": 1.4572085008615738, + "grad_norm": 0.3120006024837494, + "learning_rate": 2.6639089862776763e-06, + "loss": 0.6248, + "step": 3805 + }, + { + "epoch": 1.4575914225540876, + "grad_norm": 0.30231842398643494, + "learning_rate": 2.662895513196922e-06, + "loss": 0.6277, + "step": 3806 + }, + { + "epoch": 1.4579743442466016, + "grad_norm": 0.30227622389793396, + "learning_rate": 2.661882013230994e-06, + "loss": 0.6201, + "step": 3807 + }, + { + "epoch": 1.4583572659391155, + "grad_norm": 0.31949928402900696, + "learning_rate": 2.660868486547167e-06, + "loss": 0.6519, + "step": 3808 + }, + { + "epoch": 1.4587401876316293, + "grad_norm": 0.3098791241645813, + "learning_rate": 2.659854933312718e-06, + "loss": 0.6337, + "step": 3809 + }, + { + "epoch": 1.4591231093241432, + "grad_norm": 0.31932637095451355, + "learning_rate": 2.6588413536949294e-06, + "loss": 0.6244, + "step": 3810 + }, + { + "epoch": 1.459506031016657, + "grad_norm": 0.32003653049468994, + "learning_rate": 2.657827747861088e-06, + "loss": 0.6323, + "step": 3811 + }, + { + "epoch": 1.459888952709171, + "grad_norm": 0.3152216970920563, + "learning_rate": 2.656814115978486e-06, + "loss": 0.6314, + "step": 3812 + }, + { + "epoch": 1.460271874401685, + "grad_norm": 0.30677735805511475, + "learning_rate": 2.6558004582144177e-06, + "loss": 0.6475, + "step": 3813 + }, + { + "epoch": 1.4606547960941987, + "grad_norm": 0.2988449037075043, + "learning_rate": 2.6547867747361822e-06, + "loss": 0.6154, + "step": 3814 + }, + { + "epoch": 1.4610377177867127, + "grad_norm": 0.2977429926395416, + "learning_rate": 2.6537730657110854e-06, + "loss": 0.6473, + "step": 3815 + }, + { + "epoch": 1.4614206394792264, + "grad_norm": 0.3141452968120575, + "learning_rate": 2.652759331306433e-06, + "loss": 0.6421, + "step": 3816 + }, + { + "epoch": 1.4618035611717404, + "grad_norm": 0.3057606518268585, + "learning_rate": 2.65174557168954e-06, + "loss": 0.6266, + "step": 3817 + }, + { + "epoch": 1.4621864828642543, + "grad_norm": 0.32949110865592957, + "learning_rate": 2.6507317870277205e-06, + "loss": 0.6173, + "step": 3818 + }, + { + "epoch": 1.462569404556768, + "grad_norm": 0.316135048866272, + "learning_rate": 2.649717977488296e-06, + "loss": 0.6502, + "step": 3819 + }, + { + "epoch": 1.462952326249282, + "grad_norm": 0.3084282875061035, + "learning_rate": 2.6487041432385923e-06, + "loss": 0.6369, + "step": 3820 + }, + { + "epoch": 1.4633352479417958, + "grad_norm": 0.31082332134246826, + "learning_rate": 2.647690284445936e-06, + "loss": 0.6181, + "step": 3821 + }, + { + "epoch": 1.4637181696343098, + "grad_norm": 0.30825600028038025, + "learning_rate": 2.646676401277662e-06, + "loss": 0.6475, + "step": 3822 + }, + { + "epoch": 1.4641010913268238, + "grad_norm": 0.3085956573486328, + "learning_rate": 2.645662493901105e-06, + "loss": 0.6407, + "step": 3823 + }, + { + "epoch": 1.4644840130193375, + "grad_norm": 0.32856547832489014, + "learning_rate": 2.644648562483608e-06, + "loss": 0.6183, + "step": 3824 + }, + { + "epoch": 1.4648669347118515, + "grad_norm": 0.3092578947544098, + "learning_rate": 2.643634607192515e-06, + "loss": 0.6378, + "step": 3825 + }, + { + "epoch": 1.4652498564043652, + "grad_norm": 0.32308244705200195, + "learning_rate": 2.6426206281951744e-06, + "loss": 0.6232, + "step": 3826 + }, + { + "epoch": 1.4656327780968792, + "grad_norm": 0.3269934356212616, + "learning_rate": 2.64160662565894e-06, + "loss": 0.644, + "step": 3827 + }, + { + "epoch": 1.4660156997893932, + "grad_norm": 0.31093406677246094, + "learning_rate": 2.6405925997511667e-06, + "loss": 0.636, + "step": 3828 + }, + { + "epoch": 1.466398621481907, + "grad_norm": 0.3049316704273224, + "learning_rate": 2.6395785506392164e-06, + "loss": 0.6379, + "step": 3829 + }, + { + "epoch": 1.4667815431744209, + "grad_norm": 0.3016248643398285, + "learning_rate": 2.6385644784904517e-06, + "loss": 0.6212, + "step": 3830 + }, + { + "epoch": 1.4671644648669346, + "grad_norm": 0.3490169644355774, + "learning_rate": 2.6375503834722425e-06, + "loss": 0.642, + "step": 3831 + }, + { + "epoch": 1.4675473865594486, + "grad_norm": 0.30553266406059265, + "learning_rate": 2.63653626575196e-06, + "loss": 0.6205, + "step": 3832 + }, + { + "epoch": 1.4679303082519626, + "grad_norm": 0.30912256240844727, + "learning_rate": 2.6355221254969793e-06, + "loss": 0.6388, + "step": 3833 + }, + { + "epoch": 1.4683132299444763, + "grad_norm": 0.31346631050109863, + "learning_rate": 2.63450796287468e-06, + "loss": 0.6468, + "step": 3834 + }, + { + "epoch": 1.4686961516369903, + "grad_norm": 0.31064870953559875, + "learning_rate": 2.6334937780524457e-06, + "loss": 0.6201, + "step": 3835 + }, + { + "epoch": 1.469079073329504, + "grad_norm": 0.30559617280960083, + "learning_rate": 2.6324795711976614e-06, + "loss": 0.6175, + "step": 3836 + }, + { + "epoch": 1.469461995022018, + "grad_norm": 0.30293774604797363, + "learning_rate": 2.6314653424777194e-06, + "loss": 0.6362, + "step": 3837 + }, + { + "epoch": 1.469844916714532, + "grad_norm": 0.3056025505065918, + "learning_rate": 2.6304510920600124e-06, + "loss": 0.5998, + "step": 3838 + }, + { + "epoch": 1.4702278384070457, + "grad_norm": 0.3071589767932892, + "learning_rate": 2.629436820111938e-06, + "loss": 0.6329, + "step": 3839 + }, + { + "epoch": 1.4706107600995597, + "grad_norm": 0.31483227014541626, + "learning_rate": 2.6284225268008975e-06, + "loss": 0.6161, + "step": 3840 + }, + { + "epoch": 1.4709936817920735, + "grad_norm": 0.2940903902053833, + "learning_rate": 2.627408212294295e-06, + "loss": 0.6275, + "step": 3841 + }, + { + "epoch": 1.4713766034845874, + "grad_norm": 0.3113257884979248, + "learning_rate": 2.6263938767595393e-06, + "loss": 0.6351, + "step": 3842 + }, + { + "epoch": 1.4717595251771014, + "grad_norm": 0.3128410279750824, + "learning_rate": 2.625379520364041e-06, + "loss": 0.6373, + "step": 3843 + }, + { + "epoch": 1.4721424468696152, + "grad_norm": 0.3038770854473114, + "learning_rate": 2.624365143275216e-06, + "loss": 0.6287, + "step": 3844 + }, + { + "epoch": 1.4725253685621291, + "grad_norm": 0.307320237159729, + "learning_rate": 2.6233507456604825e-06, + "loss": 0.628, + "step": 3845 + }, + { + "epoch": 1.4729082902546429, + "grad_norm": 0.3144170343875885, + "learning_rate": 2.6223363276872614e-06, + "loss": 0.6486, + "step": 3846 + }, + { + "epoch": 1.4732912119471568, + "grad_norm": 0.31333404779434204, + "learning_rate": 2.6213218895229788e-06, + "loss": 0.6416, + "step": 3847 + }, + { + "epoch": 1.4736741336396708, + "grad_norm": 0.3031613826751709, + "learning_rate": 2.620307431335063e-06, + "loss": 0.6378, + "step": 3848 + }, + { + "epoch": 1.4740570553321846, + "grad_norm": 0.30111441016197205, + "learning_rate": 2.6192929532909455e-06, + "loss": 0.631, + "step": 3849 + }, + { + "epoch": 1.4744399770246983, + "grad_norm": 0.30961307883262634, + "learning_rate": 2.6182784555580607e-06, + "loss": 0.6275, + "step": 3850 + }, + { + "epoch": 1.4748228987172123, + "grad_norm": 0.3124462366104126, + "learning_rate": 2.617263938303849e-06, + "loss": 0.6575, + "step": 3851 + }, + { + "epoch": 1.4752058204097263, + "grad_norm": 0.30343881249427795, + "learning_rate": 2.6162494016957495e-06, + "loss": 0.6329, + "step": 3852 + }, + { + "epoch": 1.4755887421022402, + "grad_norm": 0.30654168128967285, + "learning_rate": 2.6152348459012084e-06, + "loss": 0.6148, + "step": 3853 + }, + { + "epoch": 1.475971663794754, + "grad_norm": 0.3114983141422272, + "learning_rate": 2.614220271087673e-06, + "loss": 0.6294, + "step": 3854 + }, + { + "epoch": 1.4763545854872677, + "grad_norm": 0.3270221948623657, + "learning_rate": 2.613205677422595e-06, + "loss": 0.613, + "step": 3855 + }, + { + "epoch": 1.4767375071797817, + "grad_norm": 0.31351593136787415, + "learning_rate": 2.612191065073428e-06, + "loss": 0.615, + "step": 3856 + }, + { + "epoch": 1.4771204288722957, + "grad_norm": 0.30691540241241455, + "learning_rate": 2.6111764342076285e-06, + "loss": 0.623, + "step": 3857 + }, + { + "epoch": 1.4775033505648094, + "grad_norm": 0.30246594548225403, + "learning_rate": 2.610161784992658e-06, + "loss": 0.6301, + "step": 3858 + }, + { + "epoch": 1.4778862722573234, + "grad_norm": 0.3045431077480316, + "learning_rate": 2.6091471175959794e-06, + "loss": 0.6341, + "step": 3859 + }, + { + "epoch": 1.4782691939498371, + "grad_norm": 0.3167795240879059, + "learning_rate": 2.608132432185058e-06, + "loss": 0.633, + "step": 3860 + }, + { + "epoch": 1.478652115642351, + "grad_norm": 0.31391459703445435, + "learning_rate": 2.6071177289273646e-06, + "loss": 0.6076, + "step": 3861 + }, + { + "epoch": 1.479035037334865, + "grad_norm": 0.31789183616638184, + "learning_rate": 2.606103007990371e-06, + "loss": 0.6354, + "step": 3862 + }, + { + "epoch": 1.4794179590273788, + "grad_norm": 0.32087624073028564, + "learning_rate": 2.6050882695415517e-06, + "loss": 0.6431, + "step": 3863 + }, + { + "epoch": 1.4798008807198928, + "grad_norm": 0.3151323199272156, + "learning_rate": 2.604073513748386e-06, + "loss": 0.6363, + "step": 3864 + }, + { + "epoch": 1.4801838024124065, + "grad_norm": 0.30652421712875366, + "learning_rate": 2.603058740778353e-06, + "loss": 0.6451, + "step": 3865 + }, + { + "epoch": 1.4805667241049205, + "grad_norm": 0.31759896874427795, + "learning_rate": 2.602043950798938e-06, + "loss": 0.6269, + "step": 3866 + }, + { + "epoch": 1.4809496457974345, + "grad_norm": 0.3134854733943939, + "learning_rate": 2.6010291439776265e-06, + "loss": 0.6373, + "step": 3867 + }, + { + "epoch": 1.4813325674899482, + "grad_norm": 0.30470889806747437, + "learning_rate": 2.6000143204819088e-06, + "loss": 0.6244, + "step": 3868 + }, + { + "epoch": 1.4817154891824622, + "grad_norm": 0.3090627193450928, + "learning_rate": 2.5989994804792756e-06, + "loss": 0.6344, + "step": 3869 + }, + { + "epoch": 1.482098410874976, + "grad_norm": 0.312086820602417, + "learning_rate": 2.5979846241372232e-06, + "loss": 0.6226, + "step": 3870 + }, + { + "epoch": 1.48248133256749, + "grad_norm": 0.29910963773727417, + "learning_rate": 2.5969697516232476e-06, + "loss": 0.6348, + "step": 3871 + }, + { + "epoch": 1.482864254260004, + "grad_norm": 0.3147103488445282, + "learning_rate": 2.5959548631048496e-06, + "loss": 0.6265, + "step": 3872 + }, + { + "epoch": 1.4832471759525176, + "grad_norm": 0.32287973165512085, + "learning_rate": 2.5949399587495333e-06, + "loss": 0.6453, + "step": 3873 + }, + { + "epoch": 1.4836300976450316, + "grad_norm": 0.2961839437484741, + "learning_rate": 2.593925038724802e-06, + "loss": 0.6221, + "step": 3874 + }, + { + "epoch": 1.4840130193375454, + "grad_norm": 0.30762919783592224, + "learning_rate": 2.5929101031981647e-06, + "loss": 0.6283, + "step": 3875 + }, + { + "epoch": 1.4843959410300593, + "grad_norm": 0.30955198407173157, + "learning_rate": 2.5918951523371317e-06, + "loss": 0.6273, + "step": 3876 + }, + { + "epoch": 1.4847788627225733, + "grad_norm": 0.31446561217308044, + "learning_rate": 2.590880186309217e-06, + "loss": 0.6209, + "step": 3877 + }, + { + "epoch": 1.485161784415087, + "grad_norm": 0.30787530541419983, + "learning_rate": 2.589865205281935e-06, + "loss": 0.6378, + "step": 3878 + }, + { + "epoch": 1.485544706107601, + "grad_norm": 0.30394428968429565, + "learning_rate": 2.5888502094228037e-06, + "loss": 0.6247, + "step": 3879 + }, + { + "epoch": 1.4859276278001148, + "grad_norm": 0.312267929315567, + "learning_rate": 2.5878351988993444e-06, + "loss": 0.63, + "step": 3880 + }, + { + "epoch": 1.4863105494926288, + "grad_norm": 0.3163042664527893, + "learning_rate": 2.5868201738790798e-06, + "loss": 0.6549, + "step": 3881 + }, + { + "epoch": 1.4866934711851427, + "grad_norm": 0.3131655156612396, + "learning_rate": 2.585805134529535e-06, + "loss": 0.6181, + "step": 3882 + }, + { + "epoch": 1.4870763928776565, + "grad_norm": 0.3064081072807312, + "learning_rate": 2.584790081018238e-06, + "loss": 0.6375, + "step": 3883 + }, + { + "epoch": 1.4874593145701704, + "grad_norm": 0.316826194524765, + "learning_rate": 2.5837750135127192e-06, + "loss": 0.647, + "step": 3884 + }, + { + "epoch": 1.4878422362626842, + "grad_norm": 0.30689767003059387, + "learning_rate": 2.5827599321805103e-06, + "loss": 0.6253, + "step": 3885 + }, + { + "epoch": 1.4882251579551982, + "grad_norm": 0.2950907349586487, + "learning_rate": 2.581744837189145e-06, + "loss": 0.6352, + "step": 3886 + }, + { + "epoch": 1.4886080796477121, + "grad_norm": 0.3090571165084839, + "learning_rate": 2.5807297287061616e-06, + "loss": 0.613, + "step": 3887 + }, + { + "epoch": 1.4889910013402259, + "grad_norm": 0.30332788825035095, + "learning_rate": 2.5797146068990995e-06, + "loss": 0.629, + "step": 3888 + }, + { + "epoch": 1.4893739230327399, + "grad_norm": 0.3034944534301758, + "learning_rate": 2.5786994719354986e-06, + "loss": 0.6357, + "step": 3889 + }, + { + "epoch": 1.4897568447252536, + "grad_norm": 0.3056902289390564, + "learning_rate": 2.5776843239829037e-06, + "loss": 0.6174, + "step": 3890 + }, + { + "epoch": 1.4901397664177676, + "grad_norm": 0.3162563443183899, + "learning_rate": 2.576669163208859e-06, + "loss": 0.6505, + "step": 3891 + }, + { + "epoch": 1.4905226881102815, + "grad_norm": 0.3121894896030426, + "learning_rate": 2.575653989780913e-06, + "loss": 0.6303, + "step": 3892 + }, + { + "epoch": 1.4909056098027953, + "grad_norm": 0.2926870882511139, + "learning_rate": 2.5746388038666165e-06, + "loss": 0.6221, + "step": 3893 + }, + { + "epoch": 1.4912885314953093, + "grad_norm": 0.30174705386161804, + "learning_rate": 2.5736236056335194e-06, + "loss": 0.6145, + "step": 3894 + }, + { + "epoch": 1.491671453187823, + "grad_norm": 0.309876412153244, + "learning_rate": 2.5726083952491774e-06, + "loss": 0.6309, + "step": 3895 + }, + { + "epoch": 1.492054374880337, + "grad_norm": 0.3110174536705017, + "learning_rate": 2.571593172881145e-06, + "loss": 0.6286, + "step": 3896 + }, + { + "epoch": 1.492437296572851, + "grad_norm": 0.3040805459022522, + "learning_rate": 2.5705779386969813e-06, + "loss": 0.6505, + "step": 3897 + }, + { + "epoch": 1.4928202182653647, + "grad_norm": 0.3062208294868469, + "learning_rate": 2.569562692864245e-06, + "loss": 0.6237, + "step": 3898 + }, + { + "epoch": 1.4932031399578787, + "grad_norm": 0.31146174669265747, + "learning_rate": 2.5685474355504984e-06, + "loss": 0.6382, + "step": 3899 + }, + { + "epoch": 1.4935860616503924, + "grad_norm": 0.31083017587661743, + "learning_rate": 2.567532166923306e-06, + "loss": 0.6395, + "step": 3900 + }, + { + "epoch": 1.4939689833429064, + "grad_norm": 0.30292627215385437, + "learning_rate": 2.566516887150232e-06, + "loss": 0.6155, + "step": 3901 + }, + { + "epoch": 1.4943519050354204, + "grad_norm": 0.30320677161216736, + "learning_rate": 2.5655015963988446e-06, + "loss": 0.6302, + "step": 3902 + }, + { + "epoch": 1.4947348267279341, + "grad_norm": 0.3084265887737274, + "learning_rate": 2.564486294836712e-06, + "loss": 0.6397, + "step": 3903 + }, + { + "epoch": 1.495117748420448, + "grad_norm": 0.29505085945129395, + "learning_rate": 2.5634709826314064e-06, + "loss": 0.6144, + "step": 3904 + }, + { + "epoch": 1.4955006701129618, + "grad_norm": 0.30430760979652405, + "learning_rate": 2.5624556599505008e-06, + "loss": 0.6273, + "step": 3905 + }, + { + "epoch": 1.4958835918054758, + "grad_norm": 0.30163511633872986, + "learning_rate": 2.561440326961567e-06, + "loss": 0.6305, + "step": 3906 + }, + { + "epoch": 1.4962665134979898, + "grad_norm": 0.2967780828475952, + "learning_rate": 2.5604249838321842e-06, + "loss": 0.6427, + "step": 3907 + }, + { + "epoch": 1.4966494351905035, + "grad_norm": 0.31595125794410706, + "learning_rate": 2.559409630729929e-06, + "loss": 0.6173, + "step": 3908 + }, + { + "epoch": 1.4970323568830175, + "grad_norm": 0.31400901079177856, + "learning_rate": 2.5583942678223805e-06, + "loss": 0.6406, + "step": 3909 + }, + { + "epoch": 1.4974152785755313, + "grad_norm": 0.311427503824234, + "learning_rate": 2.5573788952771204e-06, + "loss": 0.6192, + "step": 3910 + }, + { + "epoch": 1.4977982002680452, + "grad_norm": 0.3010925054550171, + "learning_rate": 2.5563635132617305e-06, + "loss": 0.6351, + "step": 3911 + }, + { + "epoch": 1.4981811219605592, + "grad_norm": 0.31229111552238464, + "learning_rate": 2.555348121943797e-06, + "loss": 0.6325, + "step": 3912 + }, + { + "epoch": 1.498564043653073, + "grad_norm": 0.31278619170188904, + "learning_rate": 2.554332721490903e-06, + "loss": 0.616, + "step": 3913 + }, + { + "epoch": 1.4989469653455867, + "grad_norm": 0.29633843898773193, + "learning_rate": 2.5533173120706378e-06, + "loss": 0.6303, + "step": 3914 + }, + { + "epoch": 1.4993298870381007, + "grad_norm": 0.3136017322540283, + "learning_rate": 2.5523018938505893e-06, + "loss": 0.6389, + "step": 3915 + }, + { + "epoch": 1.4997128087306146, + "grad_norm": 0.3160909116268158, + "learning_rate": 2.551286466998349e-06, + "loss": 0.6283, + "step": 3916 + }, + { + "epoch": 1.5000957304231286, + "grad_norm": 0.30139702558517456, + "learning_rate": 2.5502710316815066e-06, + "loss": 0.6322, + "step": 3917 + }, + { + "epoch": 1.5004786521156424, + "grad_norm": 0.31308290362358093, + "learning_rate": 2.5492555880676556e-06, + "loss": 0.6607, + "step": 3918 + }, + { + "epoch": 1.500861573808156, + "grad_norm": 0.312312513589859, + "learning_rate": 2.548240136324392e-06, + "loss": 0.6074, + "step": 3919 + }, + { + "epoch": 1.50124449550067, + "grad_norm": 0.30416786670684814, + "learning_rate": 2.54722467661931e-06, + "loss": 0.6223, + "step": 3920 + }, + { + "epoch": 1.501627417193184, + "grad_norm": 0.3039172887802124, + "learning_rate": 2.5462092091200065e-06, + "loss": 0.6177, + "step": 3921 + }, + { + "epoch": 1.502010338885698, + "grad_norm": 0.30854976177215576, + "learning_rate": 2.5451937339940807e-06, + "loss": 0.6081, + "step": 3922 + }, + { + "epoch": 1.5023932605782118, + "grad_norm": 0.30484291911125183, + "learning_rate": 2.5441782514091323e-06, + "loss": 0.6513, + "step": 3923 + }, + { + "epoch": 1.5027761822707255, + "grad_norm": 0.30959928035736084, + "learning_rate": 2.5431627615327613e-06, + "loss": 0.6462, + "step": 3924 + }, + { + "epoch": 1.5031591039632395, + "grad_norm": 0.3062193989753723, + "learning_rate": 2.54214726453257e-06, + "loss": 0.6194, + "step": 3925 + }, + { + "epoch": 1.5035420256557535, + "grad_norm": 0.3023938238620758, + "learning_rate": 2.5411317605761625e-06, + "loss": 0.6178, + "step": 3926 + }, + { + "epoch": 1.5039249473482674, + "grad_norm": 0.3154454231262207, + "learning_rate": 2.5401162498311426e-06, + "loss": 0.6461, + "step": 3927 + }, + { + "epoch": 1.5043078690407812, + "grad_norm": 0.30025947093963623, + "learning_rate": 2.539100732465115e-06, + "loss": 0.6069, + "step": 3928 + }, + { + "epoch": 1.504690790733295, + "grad_norm": 0.31008902192115784, + "learning_rate": 2.5380852086456877e-06, + "loss": 0.6368, + "step": 3929 + }, + { + "epoch": 1.505073712425809, + "grad_norm": 0.32286205887794495, + "learning_rate": 2.5370696785404673e-06, + "loss": 0.6248, + "step": 3930 + }, + { + "epoch": 1.5054566341183229, + "grad_norm": 0.30671799182891846, + "learning_rate": 2.5360541423170627e-06, + "loss": 0.6408, + "step": 3931 + }, + { + "epoch": 1.5058395558108368, + "grad_norm": 0.3128107190132141, + "learning_rate": 2.535038600143084e-06, + "loss": 0.6352, + "step": 3932 + }, + { + "epoch": 1.5062224775033506, + "grad_norm": 0.3279794752597809, + "learning_rate": 2.534023052186142e-06, + "loss": 0.6427, + "step": 3933 + }, + { + "epoch": 1.5066053991958643, + "grad_norm": 0.3025140166282654, + "learning_rate": 2.5330074986138473e-06, + "loss": 0.6434, + "step": 3934 + }, + { + "epoch": 1.5069883208883783, + "grad_norm": 0.3161628842353821, + "learning_rate": 2.531991939593813e-06, + "loss": 0.6245, + "step": 3935 + }, + { + "epoch": 1.5073712425808923, + "grad_norm": 0.31464841961860657, + "learning_rate": 2.5309763752936527e-06, + "loss": 0.649, + "step": 3936 + }, + { + "epoch": 1.5077541642734062, + "grad_norm": 0.3108835518360138, + "learning_rate": 2.529960805880981e-06, + "loss": 0.6201, + "step": 3937 + }, + { + "epoch": 1.50813708596592, + "grad_norm": 0.33190426230430603, + "learning_rate": 2.5289452315234125e-06, + "loss": 0.6119, + "step": 3938 + }, + { + "epoch": 1.5085200076584337, + "grad_norm": 0.3119288682937622, + "learning_rate": 2.5279296523885636e-06, + "loss": 0.6402, + "step": 3939 + }, + { + "epoch": 1.5089029293509477, + "grad_norm": 0.3182007670402527, + "learning_rate": 2.5269140686440513e-06, + "loss": 0.6181, + "step": 3940 + }, + { + "epoch": 1.5092858510434617, + "grad_norm": 0.3082406520843506, + "learning_rate": 2.525898480457492e-06, + "loss": 0.6308, + "step": 3941 + }, + { + "epoch": 1.5096687727359754, + "grad_norm": 0.3078022599220276, + "learning_rate": 2.5248828879965055e-06, + "loss": 0.6449, + "step": 3942 + }, + { + "epoch": 1.5100516944284894, + "grad_norm": 0.3034479022026062, + "learning_rate": 2.523867291428709e-06, + "loss": 0.6342, + "step": 3943 + }, + { + "epoch": 1.5104346161210032, + "grad_norm": 0.3038896322250366, + "learning_rate": 2.522851690921724e-06, + "loss": 0.6488, + "step": 3944 + }, + { + "epoch": 1.5108175378135171, + "grad_norm": 0.30930837988853455, + "learning_rate": 2.5218360866431695e-06, + "loss": 0.636, + "step": 3945 + }, + { + "epoch": 1.511200459506031, + "grad_norm": 0.30886512994766235, + "learning_rate": 2.520820478760668e-06, + "loss": 0.6098, + "step": 3946 + }, + { + "epoch": 1.5115833811985449, + "grad_norm": 0.3085118234157562, + "learning_rate": 2.5198048674418386e-06, + "loss": 0.6439, + "step": 3947 + }, + { + "epoch": 1.5119663028910588, + "grad_norm": 0.31368038058280945, + "learning_rate": 2.518789252854305e-06, + "loss": 0.6427, + "step": 3948 + }, + { + "epoch": 1.5123492245835726, + "grad_norm": 0.30791816115379333, + "learning_rate": 2.5177736351656903e-06, + "loss": 0.6363, + "step": 3949 + }, + { + "epoch": 1.5127321462760865, + "grad_norm": 0.3083215057849884, + "learning_rate": 2.5167580145436153e-06, + "loss": 0.633, + "step": 3950 + }, + { + "epoch": 1.5131150679686005, + "grad_norm": 0.3036786913871765, + "learning_rate": 2.515742391155707e-06, + "loss": 0.6127, + "step": 3951 + }, + { + "epoch": 1.5134979896611143, + "grad_norm": 0.3127679228782654, + "learning_rate": 2.514726765169587e-06, + "loss": 0.6329, + "step": 3952 + }, + { + "epoch": 1.5138809113536282, + "grad_norm": 0.31523454189300537, + "learning_rate": 2.51371113675288e-06, + "loss": 0.6348, + "step": 3953 + }, + { + "epoch": 1.514263833046142, + "grad_norm": 0.30303141474723816, + "learning_rate": 2.5126955060732117e-06, + "loss": 0.6528, + "step": 3954 + }, + { + "epoch": 1.514646754738656, + "grad_norm": 0.30544713139533997, + "learning_rate": 2.5116798732982074e-06, + "loss": 0.6353, + "step": 3955 + }, + { + "epoch": 1.51502967643117, + "grad_norm": 0.30993223190307617, + "learning_rate": 2.5106642385954926e-06, + "loss": 0.6213, + "step": 3956 + }, + { + "epoch": 1.5154125981236837, + "grad_norm": 0.3063317835330963, + "learning_rate": 2.509648602132692e-06, + "loss": 0.6298, + "step": 3957 + }, + { + "epoch": 1.5157955198161974, + "grad_norm": 0.30958086252212524, + "learning_rate": 2.508632964077434e-06, + "loss": 0.6268, + "step": 3958 + }, + { + "epoch": 1.5161784415087114, + "grad_norm": 0.30552297830581665, + "learning_rate": 2.507617324597344e-06, + "loss": 0.6233, + "step": 3959 + }, + { + "epoch": 1.5165613632012254, + "grad_norm": 0.3074049949645996, + "learning_rate": 2.506601683860048e-06, + "loss": 0.6378, + "step": 3960 + }, + { + "epoch": 1.5169442848937393, + "grad_norm": 0.2998192012310028, + "learning_rate": 2.5055860420331746e-06, + "loss": 0.6479, + "step": 3961 + }, + { + "epoch": 1.517327206586253, + "grad_norm": 0.307524710893631, + "learning_rate": 2.5045703992843494e-06, + "loss": 0.6394, + "step": 3962 + }, + { + "epoch": 1.5177101282787668, + "grad_norm": 0.3119560480117798, + "learning_rate": 2.5035547557812006e-06, + "loss": 0.6295, + "step": 3963 + }, + { + "epoch": 1.5180930499712808, + "grad_norm": 0.3066055178642273, + "learning_rate": 2.502539111691355e-06, + "loss": 0.6327, + "step": 3964 + }, + { + "epoch": 1.5184759716637948, + "grad_norm": 0.3096063733100891, + "learning_rate": 2.5015234671824414e-06, + "loss": 0.64, + "step": 3965 + }, + { + "epoch": 1.5188588933563087, + "grad_norm": 0.30414167046546936, + "learning_rate": 2.5005078224220852e-06, + "loss": 0.6352, + "step": 3966 + }, + { + "epoch": 1.5192418150488225, + "grad_norm": 0.315701961517334, + "learning_rate": 2.4994921775779156e-06, + "loss": 0.5995, + "step": 3967 + }, + { + "epoch": 1.5196247367413362, + "grad_norm": 0.30007249116897583, + "learning_rate": 2.49847653281756e-06, + "loss": 0.6256, + "step": 3968 + }, + { + "epoch": 1.5200076584338502, + "grad_norm": 0.3042781949043274, + "learning_rate": 2.497460888308645e-06, + "loss": 0.636, + "step": 3969 + }, + { + "epoch": 1.5203905801263642, + "grad_norm": 0.29259464144706726, + "learning_rate": 2.4964452442187998e-06, + "loss": 0.6268, + "step": 3970 + }, + { + "epoch": 1.5207735018188782, + "grad_norm": 0.31706172227859497, + "learning_rate": 2.495429600715652e-06, + "loss": 0.6458, + "step": 3971 + }, + { + "epoch": 1.521156423511392, + "grad_norm": 0.3141801953315735, + "learning_rate": 2.494413957966826e-06, + "loss": 0.619, + "step": 3972 + }, + { + "epoch": 1.5215393452039057, + "grad_norm": 0.30911046266555786, + "learning_rate": 2.4933983161399527e-06, + "loss": 0.6076, + "step": 3973 + }, + { + "epoch": 1.5219222668964196, + "grad_norm": 0.3140821158885956, + "learning_rate": 2.4923826754026566e-06, + "loss": 0.6523, + "step": 3974 + }, + { + "epoch": 1.5223051885889336, + "grad_norm": 0.3216916024684906, + "learning_rate": 2.4913670359225666e-06, + "loss": 0.6295, + "step": 3975 + }, + { + "epoch": 1.5226881102814476, + "grad_norm": 0.30086466670036316, + "learning_rate": 2.4903513978673076e-06, + "loss": 0.6539, + "step": 3976 + }, + { + "epoch": 1.5230710319739613, + "grad_norm": 0.3012973964214325, + "learning_rate": 2.4893357614045083e-06, + "loss": 0.6205, + "step": 3977 + }, + { + "epoch": 1.523453953666475, + "grad_norm": 0.30558377504348755, + "learning_rate": 2.488320126701794e-06, + "loss": 0.608, + "step": 3978 + }, + { + "epoch": 1.523836875358989, + "grad_norm": 0.3005669116973877, + "learning_rate": 2.487304493926789e-06, + "loss": 0.6148, + "step": 3979 + }, + { + "epoch": 1.524219797051503, + "grad_norm": 0.3060648441314697, + "learning_rate": 2.486288863247121e-06, + "loss": 0.6457, + "step": 3980 + }, + { + "epoch": 1.524602718744017, + "grad_norm": 0.31762269139289856, + "learning_rate": 2.485273234830414e-06, + "loss": 0.6353, + "step": 3981 + }, + { + "epoch": 1.5249856404365307, + "grad_norm": 0.3052898943424225, + "learning_rate": 2.4842576088442937e-06, + "loss": 0.6205, + "step": 3982 + }, + { + "epoch": 1.5253685621290445, + "grad_norm": 0.3068496882915497, + "learning_rate": 2.4832419854563847e-06, + "loss": 0.6458, + "step": 3983 + }, + { + "epoch": 1.5257514838215585, + "grad_norm": 0.3196716606616974, + "learning_rate": 2.48222636483431e-06, + "loss": 0.6262, + "step": 3984 + }, + { + "epoch": 1.5261344055140724, + "grad_norm": 0.3056877553462982, + "learning_rate": 2.4812107471456958e-06, + "loss": 0.6366, + "step": 3985 + }, + { + "epoch": 1.5265173272065864, + "grad_norm": 0.3014051616191864, + "learning_rate": 2.4801951325581626e-06, + "loss": 0.611, + "step": 3986 + }, + { + "epoch": 1.5269002488991001, + "grad_norm": 0.2991178631782532, + "learning_rate": 2.4791795212393335e-06, + "loss": 0.6237, + "step": 3987 + }, + { + "epoch": 1.527283170591614, + "grad_norm": 0.3264990448951721, + "learning_rate": 2.478163913356831e-06, + "loss": 0.6504, + "step": 3988 + }, + { + "epoch": 1.5276660922841279, + "grad_norm": 0.31263795495033264, + "learning_rate": 2.477148309078276e-06, + "loss": 0.6158, + "step": 3989 + }, + { + "epoch": 1.5280490139766418, + "grad_norm": 0.29732510447502136, + "learning_rate": 2.4761327085712915e-06, + "loss": 0.6365, + "step": 3990 + }, + { + "epoch": 1.5284319356691558, + "grad_norm": 0.29962366819381714, + "learning_rate": 2.475117112003495e-06, + "loss": 0.6307, + "step": 3991 + }, + { + "epoch": 1.5288148573616696, + "grad_norm": 0.31728267669677734, + "learning_rate": 2.4741015195425088e-06, + "loss": 0.6345, + "step": 3992 + }, + { + "epoch": 1.5291977790541833, + "grad_norm": 0.3012188971042633, + "learning_rate": 2.47308593135595e-06, + "loss": 0.6352, + "step": 3993 + }, + { + "epoch": 1.5295807007466973, + "grad_norm": 0.3233160972595215, + "learning_rate": 2.472070347611437e-06, + "loss": 0.6412, + "step": 3994 + }, + { + "epoch": 1.5299636224392112, + "grad_norm": 0.3041386008262634, + "learning_rate": 2.4710547684765884e-06, + "loss": 0.6059, + "step": 3995 + }, + { + "epoch": 1.5303465441317252, + "grad_norm": 0.32570070028305054, + "learning_rate": 2.4700391941190193e-06, + "loss": 0.6236, + "step": 3996 + }, + { + "epoch": 1.530729465824239, + "grad_norm": 0.311078280210495, + "learning_rate": 2.4690236247063477e-06, + "loss": 0.6143, + "step": 3997 + }, + { + "epoch": 1.5311123875167527, + "grad_norm": 0.3136025667190552, + "learning_rate": 2.4680080604061874e-06, + "loss": 0.626, + "step": 3998 + }, + { + "epoch": 1.5314953092092667, + "grad_norm": 0.31025800108909607, + "learning_rate": 2.466992501386154e-06, + "loss": 0.6347, + "step": 3999 + }, + { + "epoch": 1.5318782309017807, + "grad_norm": 0.2996547520160675, + "learning_rate": 2.4659769478138594e-06, + "loss": 0.6249, + "step": 4000 + }, + { + "epoch": 1.5322611525942946, + "grad_norm": 0.30275166034698486, + "learning_rate": 2.4649613998569165e-06, + "loss": 0.6376, + "step": 4001 + }, + { + "epoch": 1.5326440742868084, + "grad_norm": 0.3069632649421692, + "learning_rate": 2.463945857682938e-06, + "loss": 0.6455, + "step": 4002 + }, + { + "epoch": 1.5330269959793221, + "grad_norm": 0.3208339512348175, + "learning_rate": 2.462930321459533e-06, + "loss": 0.6199, + "step": 4003 + }, + { + "epoch": 1.533409917671836, + "grad_norm": 0.3333430290222168, + "learning_rate": 2.461914791354313e-06, + "loss": 0.614, + "step": 4004 + }, + { + "epoch": 1.53379283936435, + "grad_norm": 0.3028113842010498, + "learning_rate": 2.460899267534885e-06, + "loss": 0.6258, + "step": 4005 + }, + { + "epoch": 1.5341757610568638, + "grad_norm": 0.30521437525749207, + "learning_rate": 2.4598837501688586e-06, + "loss": 0.6364, + "step": 4006 + }, + { + "epoch": 1.5345586827493778, + "grad_norm": 0.3160460591316223, + "learning_rate": 2.4588682394238387e-06, + "loss": 0.6337, + "step": 4007 + }, + { + "epoch": 1.5349416044418915, + "grad_norm": 0.2983297109603882, + "learning_rate": 2.4578527354674304e-06, + "loss": 0.6169, + "step": 4008 + }, + { + "epoch": 1.5353245261344055, + "grad_norm": 0.3187613785266876, + "learning_rate": 2.456837238467239e-06, + "loss": 0.6159, + "step": 4009 + }, + { + "epoch": 1.5357074478269195, + "grad_norm": 0.30170896649360657, + "learning_rate": 2.4558217485908685e-06, + "loss": 0.5955, + "step": 4010 + }, + { + "epoch": 1.5360903695194332, + "grad_norm": 0.2984427809715271, + "learning_rate": 2.4548062660059198e-06, + "loss": 0.6279, + "step": 4011 + }, + { + "epoch": 1.5364732912119472, + "grad_norm": 0.3179510235786438, + "learning_rate": 2.453790790879994e-06, + "loss": 0.6149, + "step": 4012 + }, + { + "epoch": 1.536856212904461, + "grad_norm": 0.3076147437095642, + "learning_rate": 2.4527753233806918e-06, + "loss": 0.6373, + "step": 4013 + }, + { + "epoch": 1.537239134596975, + "grad_norm": 0.3226023018360138, + "learning_rate": 2.451759863675609e-06, + "loss": 0.6455, + "step": 4014 + }, + { + "epoch": 1.537622056289489, + "grad_norm": 0.3032069802284241, + "learning_rate": 2.450744411932345e-06, + "loss": 0.6067, + "step": 4015 + }, + { + "epoch": 1.5380049779820026, + "grad_norm": 0.3086148500442505, + "learning_rate": 2.449728968318494e-06, + "loss": 0.6269, + "step": 4016 + }, + { + "epoch": 1.5383878996745166, + "grad_norm": 0.3078845739364624, + "learning_rate": 2.448713533001652e-06, + "loss": 0.6236, + "step": 4017 + }, + { + "epoch": 1.5387708213670304, + "grad_norm": 0.31064140796661377, + "learning_rate": 2.4476981061494103e-06, + "loss": 0.6274, + "step": 4018 + }, + { + "epoch": 1.5391537430595443, + "grad_norm": 0.29724809527397156, + "learning_rate": 2.4466826879293622e-06, + "loss": 0.6155, + "step": 4019 + }, + { + "epoch": 1.5395366647520583, + "grad_norm": 0.3223803639411926, + "learning_rate": 2.445667278509098e-06, + "loss": 0.6383, + "step": 4020 + }, + { + "epoch": 1.539919586444572, + "grad_norm": 0.3130844533443451, + "learning_rate": 2.4446518780562042e-06, + "loss": 0.6249, + "step": 4021 + }, + { + "epoch": 1.5403025081370858, + "grad_norm": 0.29881101846694946, + "learning_rate": 2.44363648673827e-06, + "loss": 0.631, + "step": 4022 + }, + { + "epoch": 1.5406854298295998, + "grad_norm": 0.3078271448612213, + "learning_rate": 2.4426211047228804e-06, + "loss": 0.6219, + "step": 4023 + }, + { + "epoch": 1.5410683515221137, + "grad_norm": 0.3044681251049042, + "learning_rate": 2.4416057321776204e-06, + "loss": 0.6194, + "step": 4024 + }, + { + "epoch": 1.5414512732146277, + "grad_norm": 0.31333568692207336, + "learning_rate": 2.4405903692700715e-06, + "loss": 0.6602, + "step": 4025 + }, + { + "epoch": 1.5418341949071415, + "grad_norm": 0.30425846576690674, + "learning_rate": 2.439575016167816e-06, + "loss": 0.6378, + "step": 4026 + }, + { + "epoch": 1.5422171165996552, + "grad_norm": 0.31169235706329346, + "learning_rate": 2.438559673038434e-06, + "loss": 0.6396, + "step": 4027 + }, + { + "epoch": 1.5426000382921692, + "grad_norm": 0.3142582178115845, + "learning_rate": 2.437544340049501e-06, + "loss": 0.6416, + "step": 4028 + }, + { + "epoch": 1.5429829599846832, + "grad_norm": 0.30906203389167786, + "learning_rate": 2.4365290173685944e-06, + "loss": 0.6462, + "step": 4029 + }, + { + "epoch": 1.5433658816771971, + "grad_norm": 0.3062228858470917, + "learning_rate": 2.4355137051632883e-06, + "loss": 0.6381, + "step": 4030 + }, + { + "epoch": 1.5437488033697109, + "grad_norm": 0.30716004967689514, + "learning_rate": 2.4344984036011566e-06, + "loss": 0.6334, + "step": 4031 + }, + { + "epoch": 1.5441317250622246, + "grad_norm": 0.3089611828327179, + "learning_rate": 2.4334831128497687e-06, + "loss": 0.631, + "step": 4032 + }, + { + "epoch": 1.5445146467547386, + "grad_norm": 0.303734689950943, + "learning_rate": 2.432467833076694e-06, + "loss": 0.6371, + "step": 4033 + }, + { + "epoch": 1.5448975684472526, + "grad_norm": 0.30264773964881897, + "learning_rate": 2.4314525644495024e-06, + "loss": 0.612, + "step": 4034 + }, + { + "epoch": 1.5452804901397665, + "grad_norm": 0.31619319319725037, + "learning_rate": 2.430437307135756e-06, + "loss": 0.6167, + "step": 4035 + }, + { + "epoch": 1.5456634118322803, + "grad_norm": 0.3128945231437683, + "learning_rate": 2.4294220613030195e-06, + "loss": 0.628, + "step": 4036 + }, + { + "epoch": 1.546046333524794, + "grad_norm": 0.31108132004737854, + "learning_rate": 2.428406827118856e-06, + "loss": 0.6121, + "step": 4037 + }, + { + "epoch": 1.546429255217308, + "grad_norm": 0.3100755214691162, + "learning_rate": 2.427391604750823e-06, + "loss": 0.656, + "step": 4038 + }, + { + "epoch": 1.546812176909822, + "grad_norm": 0.3169742524623871, + "learning_rate": 2.426376394366481e-06, + "loss": 0.6442, + "step": 4039 + }, + { + "epoch": 1.547195098602336, + "grad_norm": 0.30973201990127563, + "learning_rate": 2.425361196133384e-06, + "loss": 0.6183, + "step": 4040 + }, + { + "epoch": 1.5475780202948497, + "grad_norm": 0.3088168799877167, + "learning_rate": 2.4243460102190872e-06, + "loss": 0.6457, + "step": 4041 + }, + { + "epoch": 1.5479609419873634, + "grad_norm": 0.30473074316978455, + "learning_rate": 2.423330836791142e-06, + "loss": 0.6188, + "step": 4042 + }, + { + "epoch": 1.5483438636798774, + "grad_norm": 0.30431094765663147, + "learning_rate": 2.4223156760170976e-06, + "loss": 0.637, + "step": 4043 + }, + { + "epoch": 1.5487267853723914, + "grad_norm": 0.3165663182735443, + "learning_rate": 2.4213005280645022e-06, + "loss": 0.6383, + "step": 4044 + }, + { + "epoch": 1.5491097070649054, + "grad_norm": 0.3053666055202484, + "learning_rate": 2.4202853931009013e-06, + "loss": 0.6426, + "step": 4045 + }, + { + "epoch": 1.549492628757419, + "grad_norm": 0.3089866638183594, + "learning_rate": 2.419270271293839e-06, + "loss": 0.6252, + "step": 4046 + }, + { + "epoch": 1.5498755504499329, + "grad_norm": 0.31108561158180237, + "learning_rate": 2.4182551628108553e-06, + "loss": 0.622, + "step": 4047 + }, + { + "epoch": 1.5502584721424468, + "grad_norm": 0.2920471429824829, + "learning_rate": 2.4172400678194913e-06, + "loss": 0.6388, + "step": 4048 + }, + { + "epoch": 1.5506413938349608, + "grad_norm": 0.32072851061820984, + "learning_rate": 2.416224986487282e-06, + "loss": 0.6212, + "step": 4049 + }, + { + "epoch": 1.5510243155274748, + "grad_norm": 0.3059203624725342, + "learning_rate": 2.4152099189817624e-06, + "loss": 0.635, + "step": 4050 + }, + { + "epoch": 1.5514072372199885, + "grad_norm": 0.299489825963974, + "learning_rate": 2.4141948654704654e-06, + "loss": 0.6252, + "step": 4051 + }, + { + "epoch": 1.5517901589125023, + "grad_norm": 0.30650538206100464, + "learning_rate": 2.4131798261209206e-06, + "loss": 0.6201, + "step": 4052 + }, + { + "epoch": 1.5521730806050162, + "grad_norm": 0.3110271990299225, + "learning_rate": 2.4121648011006564e-06, + "loss": 0.6459, + "step": 4053 + }, + { + "epoch": 1.5525560022975302, + "grad_norm": 0.3129662275314331, + "learning_rate": 2.4111497905771967e-06, + "loss": 0.6356, + "step": 4054 + }, + { + "epoch": 1.5529389239900442, + "grad_norm": 0.2955263555049896, + "learning_rate": 2.410134794718066e-06, + "loss": 0.6242, + "step": 4055 + }, + { + "epoch": 1.553321845682558, + "grad_norm": 0.3057519197463989, + "learning_rate": 2.4091198136907847e-06, + "loss": 0.6261, + "step": 4056 + }, + { + "epoch": 1.5537047673750717, + "grad_norm": 0.31974682211875916, + "learning_rate": 2.4081048476628687e-06, + "loss": 0.6237, + "step": 4057 + }, + { + "epoch": 1.5540876890675857, + "grad_norm": 0.313350111246109, + "learning_rate": 2.4070898968018365e-06, + "loss": 0.6241, + "step": 4058 + }, + { + "epoch": 1.5544706107600996, + "grad_norm": 0.307029664516449, + "learning_rate": 2.4060749612751987e-06, + "loss": 0.6108, + "step": 4059 + }, + { + "epoch": 1.5548535324526136, + "grad_norm": 0.31785449385643005, + "learning_rate": 2.405060041250467e-06, + "loss": 0.618, + "step": 4060 + }, + { + "epoch": 1.5552364541451273, + "grad_norm": 0.313344269990921, + "learning_rate": 2.4040451368951504e-06, + "loss": 0.6694, + "step": 4061 + }, + { + "epoch": 1.555619375837641, + "grad_norm": 0.31993594765663147, + "learning_rate": 2.4030302483767524e-06, + "loss": 0.6412, + "step": 4062 + }, + { + "epoch": 1.556002297530155, + "grad_norm": 0.3105093240737915, + "learning_rate": 2.402015375862778e-06, + "loss": 0.6502, + "step": 4063 + }, + { + "epoch": 1.556385219222669, + "grad_norm": 0.3041859567165375, + "learning_rate": 2.401000519520725e-06, + "loss": 0.6478, + "step": 4064 + }, + { + "epoch": 1.556768140915183, + "grad_norm": 0.29936468601226807, + "learning_rate": 2.399985679518092e-06, + "loss": 0.6301, + "step": 4065 + }, + { + "epoch": 1.5571510626076968, + "grad_norm": 0.30676236748695374, + "learning_rate": 2.398970856022374e-06, + "loss": 0.6121, + "step": 4066 + }, + { + "epoch": 1.5575339843002105, + "grad_norm": 0.3220148980617523, + "learning_rate": 2.397956049201062e-06, + "loss": 0.6377, + "step": 4067 + }, + { + "epoch": 1.5579169059927245, + "grad_norm": 0.29830971360206604, + "learning_rate": 2.3969412592216475e-06, + "loss": 0.6204, + "step": 4068 + }, + { + "epoch": 1.5582998276852384, + "grad_norm": 0.3051277697086334, + "learning_rate": 2.3959264862516145e-06, + "loss": 0.6269, + "step": 4069 + }, + { + "epoch": 1.5586827493777522, + "grad_norm": 0.3069190979003906, + "learning_rate": 2.3949117304584487e-06, + "loss": 0.6385, + "step": 4070 + }, + { + "epoch": 1.5590656710702662, + "grad_norm": 0.3124160170555115, + "learning_rate": 2.39389699200963e-06, + "loss": 0.6221, + "step": 4071 + }, + { + "epoch": 1.55944859276278, + "grad_norm": 0.3104918301105499, + "learning_rate": 2.3928822710726358e-06, + "loss": 0.6367, + "step": 4072 + }, + { + "epoch": 1.5598315144552939, + "grad_norm": 0.309496134519577, + "learning_rate": 2.3918675678149424e-06, + "loss": 0.6437, + "step": 4073 + }, + { + "epoch": 1.5602144361478079, + "grad_norm": 0.30972275137901306, + "learning_rate": 2.3908528824040214e-06, + "loss": 0.6396, + "step": 4074 + }, + { + "epoch": 1.5605973578403216, + "grad_norm": 0.30616679787635803, + "learning_rate": 2.3898382150073427e-06, + "loss": 0.6515, + "step": 4075 + }, + { + "epoch": 1.5609802795328356, + "grad_norm": 0.30103224515914917, + "learning_rate": 2.388823565792372e-06, + "loss": 0.6198, + "step": 4076 + }, + { + "epoch": 1.5613632012253493, + "grad_norm": 0.30729299783706665, + "learning_rate": 2.3878089349265734e-06, + "loss": 0.6177, + "step": 4077 + }, + { + "epoch": 1.5617461229178633, + "grad_norm": 0.301432341337204, + "learning_rate": 2.3867943225774064e-06, + "loss": 0.6444, + "step": 4078 + }, + { + "epoch": 1.5621290446103773, + "grad_norm": 0.3121088445186615, + "learning_rate": 2.3857797289123276e-06, + "loss": 0.6234, + "step": 4079 + }, + { + "epoch": 1.562511966302891, + "grad_norm": 0.3122314512729645, + "learning_rate": 2.3847651540987925e-06, + "loss": 0.6433, + "step": 4080 + }, + { + "epoch": 1.562894887995405, + "grad_norm": 0.31055715680122375, + "learning_rate": 2.383750598304251e-06, + "loss": 0.6306, + "step": 4081 + }, + { + "epoch": 1.5632778096879187, + "grad_norm": 0.3161414861679077, + "learning_rate": 2.382736061696152e-06, + "loss": 0.6291, + "step": 4082 + }, + { + "epoch": 1.5636607313804327, + "grad_norm": 0.3138085901737213, + "learning_rate": 2.3817215444419392e-06, + "loss": 0.6643, + "step": 4083 + }, + { + "epoch": 1.5640436530729467, + "grad_norm": 0.3187800943851471, + "learning_rate": 2.380707046709056e-06, + "loss": 0.6368, + "step": 4084 + }, + { + "epoch": 1.5644265747654604, + "grad_norm": 0.31199902296066284, + "learning_rate": 2.3796925686649382e-06, + "loss": 0.6272, + "step": 4085 + }, + { + "epoch": 1.5648094964579742, + "grad_norm": 0.31432831287384033, + "learning_rate": 2.378678110477022e-06, + "loss": 0.6197, + "step": 4086 + }, + { + "epoch": 1.5651924181504882, + "grad_norm": 0.3184930086135864, + "learning_rate": 2.377663672312739e-06, + "loss": 0.6193, + "step": 4087 + }, + { + "epoch": 1.5655753398430021, + "grad_norm": 0.3189293146133423, + "learning_rate": 2.3766492543395183e-06, + "loss": 0.6242, + "step": 4088 + }, + { + "epoch": 1.565958261535516, + "grad_norm": 0.31775569915771484, + "learning_rate": 2.375634856724784e-06, + "loss": 0.6321, + "step": 4089 + }, + { + "epoch": 1.5663411832280298, + "grad_norm": 0.3047056198120117, + "learning_rate": 2.374620479635959e-06, + "loss": 0.62, + "step": 4090 + }, + { + "epoch": 1.5667241049205436, + "grad_norm": 0.31037959456443787, + "learning_rate": 2.373606123240462e-06, + "loss": 0.6295, + "step": 4091 + }, + { + "epoch": 1.5671070266130576, + "grad_norm": 0.31016772985458374, + "learning_rate": 2.3725917877057054e-06, + "loss": 0.639, + "step": 4092 + }, + { + "epoch": 1.5674899483055715, + "grad_norm": 0.30299362540245056, + "learning_rate": 2.3715774731991034e-06, + "loss": 0.6289, + "step": 4093 + }, + { + "epoch": 1.5678728699980855, + "grad_norm": 0.301461786031723, + "learning_rate": 2.3705631798880625e-06, + "loss": 0.6226, + "step": 4094 + }, + { + "epoch": 1.5682557916905993, + "grad_norm": 0.3076412081718445, + "learning_rate": 2.3695489079399885e-06, + "loss": 0.6225, + "step": 4095 + }, + { + "epoch": 1.568638713383113, + "grad_norm": 0.29231682419776917, + "learning_rate": 2.368534657522281e-06, + "loss": 0.6258, + "step": 4096 + }, + { + "epoch": 1.569021635075627, + "grad_norm": 0.3074251711368561, + "learning_rate": 2.3675204288023386e-06, + "loss": 0.6217, + "step": 4097 + }, + { + "epoch": 1.569404556768141, + "grad_norm": 0.31576603651046753, + "learning_rate": 2.3665062219475556e-06, + "loss": 0.6265, + "step": 4098 + }, + { + "epoch": 1.569787478460655, + "grad_norm": 0.30854710936546326, + "learning_rate": 2.3654920371253205e-06, + "loss": 0.6261, + "step": 4099 + }, + { + "epoch": 1.5701704001531687, + "grad_norm": 0.3100806176662445, + "learning_rate": 2.3644778745030215e-06, + "loss": 0.6432, + "step": 4100 + }, + { + "epoch": 1.5705533218456824, + "grad_norm": 0.30547037720680237, + "learning_rate": 2.3634637342480403e-06, + "loss": 0.6443, + "step": 4101 + }, + { + "epoch": 1.5709362435381964, + "grad_norm": 0.31237226724624634, + "learning_rate": 2.362449616527758e-06, + "loss": 0.6226, + "step": 4102 + }, + { + "epoch": 1.5713191652307104, + "grad_norm": 0.3147776424884796, + "learning_rate": 2.3614355215095483e-06, + "loss": 0.6452, + "step": 4103 + }, + { + "epoch": 1.5717020869232243, + "grad_norm": 0.3154314160346985, + "learning_rate": 2.3604214493607844e-06, + "loss": 0.6351, + "step": 4104 + }, + { + "epoch": 1.572085008615738, + "grad_norm": 0.30704769492149353, + "learning_rate": 2.3594074002488345e-06, + "loss": 0.6212, + "step": 4105 + }, + { + "epoch": 1.5724679303082518, + "grad_norm": 0.30162665247917175, + "learning_rate": 2.3583933743410613e-06, + "loss": 0.6157, + "step": 4106 + }, + { + "epoch": 1.5728508520007658, + "grad_norm": 0.30751460790634155, + "learning_rate": 2.3573793718048265e-06, + "loss": 0.6433, + "step": 4107 + }, + { + "epoch": 1.5732337736932798, + "grad_norm": 0.3191346526145935, + "learning_rate": 2.3563653928074857e-06, + "loss": 0.6251, + "step": 4108 + }, + { + "epoch": 1.5736166953857937, + "grad_norm": 0.3138386607170105, + "learning_rate": 2.3553514375163926e-06, + "loss": 0.6121, + "step": 4109 + }, + { + "epoch": 1.5739996170783075, + "grad_norm": 0.30303409695625305, + "learning_rate": 2.3543375060988953e-06, + "loss": 0.6234, + "step": 4110 + }, + { + "epoch": 1.5743825387708212, + "grad_norm": 0.3079211711883545, + "learning_rate": 2.353323598722339e-06, + "loss": 0.6171, + "step": 4111 + }, + { + "epoch": 1.5747654604633352, + "grad_norm": 0.2996062934398651, + "learning_rate": 2.352309715554065e-06, + "loss": 0.6217, + "step": 4112 + }, + { + "epoch": 1.5751483821558492, + "grad_norm": 0.3040037453174591, + "learning_rate": 2.351295856761409e-06, + "loss": 0.6145, + "step": 4113 + }, + { + "epoch": 1.5755313038483632, + "grad_norm": 0.30515366792678833, + "learning_rate": 2.350282022511705e-06, + "loss": 0.6235, + "step": 4114 + }, + { + "epoch": 1.575914225540877, + "grad_norm": 0.3041256070137024, + "learning_rate": 2.3492682129722803e-06, + "loss": 0.643, + "step": 4115 + }, + { + "epoch": 1.5762971472333906, + "grad_norm": 0.31586289405822754, + "learning_rate": 2.3482544283104603e-06, + "loss": 0.6196, + "step": 4116 + }, + { + "epoch": 1.5766800689259046, + "grad_norm": 0.3158038854598999, + "learning_rate": 2.347240668693567e-06, + "loss": 0.6365, + "step": 4117 + }, + { + "epoch": 1.5770629906184186, + "grad_norm": 0.29749569296836853, + "learning_rate": 2.346226934288915e-06, + "loss": 0.6127, + "step": 4118 + }, + { + "epoch": 1.5774459123109326, + "grad_norm": 0.3016461133956909, + "learning_rate": 2.345213225263818e-06, + "loss": 0.6323, + "step": 4119 + }, + { + "epoch": 1.5778288340034463, + "grad_norm": 0.31081822514533997, + "learning_rate": 2.3441995417855835e-06, + "loss": 0.6338, + "step": 4120 + }, + { + "epoch": 1.57821175569596, + "grad_norm": 0.30540379881858826, + "learning_rate": 2.343185884021515e-06, + "loss": 0.6469, + "step": 4121 + }, + { + "epoch": 1.578594677388474, + "grad_norm": 0.29972654581069946, + "learning_rate": 2.3421722521389125e-06, + "loss": 0.624, + "step": 4122 + }, + { + "epoch": 1.578977599080988, + "grad_norm": 0.2996334135532379, + "learning_rate": 2.341158646305071e-06, + "loss": 0.6167, + "step": 4123 + }, + { + "epoch": 1.579360520773502, + "grad_norm": 0.2999216616153717, + "learning_rate": 2.340145066687283e-06, + "loss": 0.6224, + "step": 4124 + }, + { + "epoch": 1.5797434424660157, + "grad_norm": 0.31695669889450073, + "learning_rate": 2.339131513452833e-06, + "loss": 0.6136, + "step": 4125 + }, + { + "epoch": 1.5801263641585295, + "grad_norm": 0.2997908592224121, + "learning_rate": 2.3381179867690067e-06, + "loss": 0.6345, + "step": 4126 + }, + { + "epoch": 1.5805092858510434, + "grad_norm": 0.3049721419811249, + "learning_rate": 2.337104486803079e-06, + "loss": 0.6251, + "step": 4127 + }, + { + "epoch": 1.5808922075435574, + "grad_norm": 0.3121729791164398, + "learning_rate": 2.3360910137223246e-06, + "loss": 0.6404, + "step": 4128 + }, + { + "epoch": 1.5812751292360714, + "grad_norm": 0.3162428140640259, + "learning_rate": 2.335077567694013e-06, + "loss": 0.6303, + "step": 4129 + }, + { + "epoch": 1.5816580509285851, + "grad_norm": 0.32083892822265625, + "learning_rate": 2.3340641488854084e-06, + "loss": 0.6192, + "step": 4130 + }, + { + "epoch": 1.5820409726210989, + "grad_norm": 0.31646478176116943, + "learning_rate": 2.333050757463771e-06, + "loss": 0.6287, + "step": 4131 + }, + { + "epoch": 1.5824238943136129, + "grad_norm": 0.3102489411830902, + "learning_rate": 2.332037393596357e-06, + "loss": 0.6407, + "step": 4132 + }, + { + "epoch": 1.5828068160061268, + "grad_norm": 0.3095809817314148, + "learning_rate": 2.3310240574504184e-06, + "loss": 0.6295, + "step": 4133 + }, + { + "epoch": 1.5831897376986406, + "grad_norm": 0.3293621838092804, + "learning_rate": 2.3300107491932002e-06, + "loss": 0.6403, + "step": 4134 + }, + { + "epoch": 1.5835726593911545, + "grad_norm": 0.32825222611427307, + "learning_rate": 2.328997468991944e-06, + "loss": 0.6044, + "step": 4135 + }, + { + "epoch": 1.5839555810836683, + "grad_norm": 0.31198835372924805, + "learning_rate": 2.327984217013889e-06, + "loss": 0.6462, + "step": 4136 + }, + { + "epoch": 1.5843385027761823, + "grad_norm": 0.31242629885673523, + "learning_rate": 2.3269709934262664e-06, + "loss": 0.6349, + "step": 4137 + }, + { + "epoch": 1.5847214244686962, + "grad_norm": 0.3098580241203308, + "learning_rate": 2.325957798396305e-06, + "loss": 0.6212, + "step": 4138 + }, + { + "epoch": 1.58510434616121, + "grad_norm": 0.31811580061912537, + "learning_rate": 2.3249446320912277e-06, + "loss": 0.636, + "step": 4139 + }, + { + "epoch": 1.585487267853724, + "grad_norm": 0.325848251581192, + "learning_rate": 2.3239314946782536e-06, + "loss": 0.6335, + "step": 4140 + }, + { + "epoch": 1.5858701895462377, + "grad_norm": 0.3145314157009125, + "learning_rate": 2.3229183863245964e-06, + "loss": 0.6482, + "step": 4141 + }, + { + "epoch": 1.5862531112387517, + "grad_norm": 0.31943273544311523, + "learning_rate": 2.3219053071974637e-06, + "loss": 0.6298, + "step": 4142 + }, + { + "epoch": 1.5866360329312656, + "grad_norm": 0.3052866756916046, + "learning_rate": 2.3208922574640612e-06, + "loss": 0.6408, + "step": 4143 + }, + { + "epoch": 1.5870189546237794, + "grad_norm": 0.3176013231277466, + "learning_rate": 2.319879237291588e-06, + "loss": 0.6207, + "step": 4144 + }, + { + "epoch": 1.5874018763162934, + "grad_norm": 0.32713839411735535, + "learning_rate": 2.3188662468472376e-06, + "loss": 0.6386, + "step": 4145 + }, + { + "epoch": 1.5877847980088071, + "grad_norm": 0.31876373291015625, + "learning_rate": 2.317853286298201e-06, + "loss": 0.6321, + "step": 4146 + }, + { + "epoch": 1.588167719701321, + "grad_norm": 0.3036886751651764, + "learning_rate": 2.3168403558116612e-06, + "loss": 0.6205, + "step": 4147 + }, + { + "epoch": 1.588550641393835, + "grad_norm": 0.30306875705718994, + "learning_rate": 2.3158274555548e-06, + "loss": 0.6282, + "step": 4148 + }, + { + "epoch": 1.5889335630863488, + "grad_norm": 0.3102256953716278, + "learning_rate": 2.31481458569479e-06, + "loss": 0.6317, + "step": 4149 + }, + { + "epoch": 1.5893164847788626, + "grad_norm": 0.3065396547317505, + "learning_rate": 2.313801746398801e-06, + "loss": 0.635, + "step": 4150 + }, + { + "epoch": 1.5896994064713765, + "grad_norm": 0.3208254277706146, + "learning_rate": 2.312788937833999e-06, + "loss": 0.6282, + "step": 4151 + }, + { + "epoch": 1.5900823281638905, + "grad_norm": 0.2942972779273987, + "learning_rate": 2.3117761601675414e-06, + "loss": 0.6176, + "step": 4152 + }, + { + "epoch": 1.5904652498564045, + "grad_norm": 0.3133404850959778, + "learning_rate": 2.3107634135665847e-06, + "loss": 0.6377, + "step": 4153 + }, + { + "epoch": 1.5908481715489182, + "grad_norm": 0.3113650381565094, + "learning_rate": 2.3097506981982764e-06, + "loss": 0.6236, + "step": 4154 + }, + { + "epoch": 1.591231093241432, + "grad_norm": 0.2996228337287903, + "learning_rate": 2.308738014229763e-06, + "loss": 0.6138, + "step": 4155 + }, + { + "epoch": 1.591614014933946, + "grad_norm": 0.3065096437931061, + "learning_rate": 2.3077253618281813e-06, + "loss": 0.6379, + "step": 4156 + }, + { + "epoch": 1.59199693662646, + "grad_norm": 0.29608598351478577, + "learning_rate": 2.306712741160665e-06, + "loss": 0.6167, + "step": 4157 + }, + { + "epoch": 1.5923798583189739, + "grad_norm": 0.3045905530452728, + "learning_rate": 2.305700152394344e-06, + "loss": 0.6234, + "step": 4158 + }, + { + "epoch": 1.5927627800114876, + "grad_norm": 0.31149375438690186, + "learning_rate": 2.30468759569634e-06, + "loss": 0.6259, + "step": 4159 + }, + { + "epoch": 1.5931457017040014, + "grad_norm": 0.3019504249095917, + "learning_rate": 2.3036750712337726e-06, + "loss": 0.6316, + "step": 4160 + }, + { + "epoch": 1.5935286233965154, + "grad_norm": 0.3150195777416229, + "learning_rate": 2.302662579173753e-06, + "loss": 0.6274, + "step": 4161 + }, + { + "epoch": 1.5939115450890293, + "grad_norm": 0.3074125051498413, + "learning_rate": 2.3016501196833903e-06, + "loss": 0.6297, + "step": 4162 + }, + { + "epoch": 1.5942944667815433, + "grad_norm": 0.3255399763584137, + "learning_rate": 2.3006376929297842e-06, + "loss": 0.6418, + "step": 4163 + }, + { + "epoch": 1.594677388474057, + "grad_norm": 0.30653679370880127, + "learning_rate": 2.2996252990800314e-06, + "loss": 0.6385, + "step": 4164 + }, + { + "epoch": 1.5950603101665708, + "grad_norm": 0.30491143465042114, + "learning_rate": 2.298612938301225e-06, + "loss": 0.6375, + "step": 4165 + }, + { + "epoch": 1.5954432318590848, + "grad_norm": 0.3082655966281891, + "learning_rate": 2.297600610760448e-06, + "loss": 0.6077, + "step": 4166 + }, + { + "epoch": 1.5958261535515987, + "grad_norm": 0.316263347864151, + "learning_rate": 2.2965883166247822e-06, + "loss": 0.6282, + "step": 4167 + }, + { + "epoch": 1.5962090752441127, + "grad_norm": 0.3127671778202057, + "learning_rate": 2.2955760560613024e-06, + "loss": 0.6257, + "step": 4168 + }, + { + "epoch": 1.5965919969366265, + "grad_norm": 0.312476247549057, + "learning_rate": 2.2945638292370777e-06, + "loss": 0.6401, + "step": 4169 + }, + { + "epoch": 1.5969749186291402, + "grad_norm": 0.3150080442428589, + "learning_rate": 2.2935516363191695e-06, + "loss": 0.6334, + "step": 4170 + }, + { + "epoch": 1.5973578403216542, + "grad_norm": 0.3099198043346405, + "learning_rate": 2.292539477474638e-06, + "loss": 0.6157, + "step": 4171 + }, + { + "epoch": 1.5977407620141681, + "grad_norm": 0.3027891516685486, + "learning_rate": 2.291527352870534e-06, + "loss": 0.6221, + "step": 4172 + }, + { + "epoch": 1.5981236837066821, + "grad_norm": 0.30542635917663574, + "learning_rate": 2.2905152626739054e-06, + "loss": 0.6122, + "step": 4173 + }, + { + "epoch": 1.5985066053991959, + "grad_norm": 0.30441731214523315, + "learning_rate": 2.2895032070517916e-06, + "loss": 0.6419, + "step": 4174 + }, + { + "epoch": 1.5988895270917096, + "grad_norm": 0.29802995920181274, + "learning_rate": 2.2884911861712296e-06, + "loss": 0.6356, + "step": 4175 + }, + { + "epoch": 1.5992724487842236, + "grad_norm": 0.3023606538772583, + "learning_rate": 2.2874792001992485e-06, + "loss": 0.6142, + "step": 4176 + }, + { + "epoch": 1.5996553704767376, + "grad_norm": 0.3082006871700287, + "learning_rate": 2.2864672493028708e-06, + "loss": 0.6217, + "step": 4177 + }, + { + "epoch": 1.6000382921692515, + "grad_norm": 0.301119863986969, + "learning_rate": 2.2854553336491153e-06, + "loss": 0.6475, + "step": 4178 + }, + { + "epoch": 1.6004212138617653, + "grad_norm": 0.31710493564605713, + "learning_rate": 2.284443453404994e-06, + "loss": 0.6322, + "step": 4179 + }, + { + "epoch": 1.600804135554279, + "grad_norm": 0.31864047050476074, + "learning_rate": 2.2834316087375136e-06, + "loss": 0.6306, + "step": 4180 + }, + { + "epoch": 1.601187057246793, + "grad_norm": 0.31082892417907715, + "learning_rate": 2.2824197998136737e-06, + "loss": 0.6271, + "step": 4181 + }, + { + "epoch": 1.601569978939307, + "grad_norm": 0.3050023913383484, + "learning_rate": 2.28140802680047e-06, + "loss": 0.6381, + "step": 4182 + }, + { + "epoch": 1.601952900631821, + "grad_norm": 0.30371639132499695, + "learning_rate": 2.2803962898648907e-06, + "loss": 0.6142, + "step": 4183 + }, + { + "epoch": 1.6023358223243347, + "grad_norm": 0.3214907646179199, + "learning_rate": 2.279384589173917e-06, + "loss": 0.6379, + "step": 4184 + }, + { + "epoch": 1.6027187440168484, + "grad_norm": 0.3213587999343872, + "learning_rate": 2.2783729248945275e-06, + "loss": 0.6296, + "step": 4185 + }, + { + "epoch": 1.6031016657093624, + "grad_norm": 0.31715714931488037, + "learning_rate": 2.2773612971936916e-06, + "loss": 0.6391, + "step": 4186 + }, + { + "epoch": 1.6034845874018764, + "grad_norm": 0.3020121455192566, + "learning_rate": 2.2763497062383748e-06, + "loss": 0.6327, + "step": 4187 + }, + { + "epoch": 1.6038675090943904, + "grad_norm": 0.3249233365058899, + "learning_rate": 2.2753381521955345e-06, + "loss": 0.6472, + "step": 4188 + }, + { + "epoch": 1.604250430786904, + "grad_norm": 0.34253984689712524, + "learning_rate": 2.274326635232125e-06, + "loss": 0.6376, + "step": 4189 + }, + { + "epoch": 1.6046333524794179, + "grad_norm": 0.3286554515361786, + "learning_rate": 2.2733151555150917e-06, + "loss": 0.6116, + "step": 4190 + }, + { + "epoch": 1.6050162741719318, + "grad_norm": 0.30143681168556213, + "learning_rate": 2.2723037132113733e-06, + "loss": 0.6432, + "step": 4191 + }, + { + "epoch": 1.6053991958644458, + "grad_norm": 0.30886906385421753, + "learning_rate": 2.2712923084879056e-06, + "loss": 0.6524, + "step": 4192 + }, + { + "epoch": 1.6057821175569598, + "grad_norm": 0.3125979006290436, + "learning_rate": 2.2702809415116154e-06, + "loss": 0.6429, + "step": 4193 + }, + { + "epoch": 1.6061650392494735, + "grad_norm": 0.329543799161911, + "learning_rate": 2.2692696124494254e-06, + "loss": 0.6163, + "step": 4194 + }, + { + "epoch": 1.6065479609419873, + "grad_norm": 0.3624904751777649, + "learning_rate": 2.2682583214682507e-06, + "loss": 0.6133, + "step": 4195 + }, + { + "epoch": 1.6069308826345012, + "grad_norm": 0.30061206221580505, + "learning_rate": 2.267247068734999e-06, + "loss": 0.6161, + "step": 4196 + }, + { + "epoch": 1.6073138043270152, + "grad_norm": 0.3322734236717224, + "learning_rate": 2.2662358544165753e-06, + "loss": 0.6435, + "step": 4197 + }, + { + "epoch": 1.607696726019529, + "grad_norm": 0.30932679772377014, + "learning_rate": 2.265224678679874e-06, + "loss": 0.6528, + "step": 4198 + }, + { + "epoch": 1.608079647712043, + "grad_norm": 0.3287763297557831, + "learning_rate": 2.2642135416917855e-06, + "loss": 0.6161, + "step": 4199 + }, + { + "epoch": 1.6084625694045567, + "grad_norm": 0.33302974700927734, + "learning_rate": 2.263202443619194e-06, + "loss": 0.6139, + "step": 4200 + }, + { + "epoch": 1.6088454910970706, + "grad_norm": 0.3082435727119446, + "learning_rate": 2.262191384628976e-06, + "loss": 0.628, + "step": 4201 + }, + { + "epoch": 1.6092284127895846, + "grad_norm": 0.31206557154655457, + "learning_rate": 2.261180364888003e-06, + "loss": 0.6279, + "step": 4202 + }, + { + "epoch": 1.6096113344820984, + "grad_norm": 0.36720749735832214, + "learning_rate": 2.2601693845631387e-06, + "loss": 0.6113, + "step": 4203 + }, + { + "epoch": 1.6099942561746123, + "grad_norm": 0.3202786445617676, + "learning_rate": 2.259158443821242e-06, + "loss": 0.6505, + "step": 4204 + }, + { + "epoch": 1.610377177867126, + "grad_norm": 0.31435686349868774, + "learning_rate": 2.258147542829163e-06, + "loss": 0.6506, + "step": 4205 + }, + { + "epoch": 1.61076009955964, + "grad_norm": 0.3013192415237427, + "learning_rate": 2.2571366817537457e-06, + "loss": 0.6222, + "step": 4206 + }, + { + "epoch": 1.611143021252154, + "grad_norm": 0.304572194814682, + "learning_rate": 2.2561258607618296e-06, + "loss": 0.6101, + "step": 4207 + }, + { + "epoch": 1.6115259429446678, + "grad_norm": 0.3269023299217224, + "learning_rate": 2.2551150800202452e-06, + "loss": 0.6215, + "step": 4208 + }, + { + "epoch": 1.6119088646371817, + "grad_norm": 0.31622642278671265, + "learning_rate": 2.254104339695819e-06, + "loss": 0.6217, + "step": 4209 + }, + { + "epoch": 1.6122917863296955, + "grad_norm": 0.3095210790634155, + "learning_rate": 2.2530936399553667e-06, + "loss": 0.6424, + "step": 4210 + }, + { + "epoch": 1.6126747080222095, + "grad_norm": 0.318411648273468, + "learning_rate": 2.252082980965702e-06, + "loss": 0.6319, + "step": 4211 + }, + { + "epoch": 1.6130576297147234, + "grad_norm": 0.30553948879241943, + "learning_rate": 2.251072362893628e-06, + "loss": 0.6341, + "step": 4212 + }, + { + "epoch": 1.6134405514072372, + "grad_norm": 0.3127796947956085, + "learning_rate": 2.2500617859059433e-06, + "loss": 0.6129, + "step": 4213 + }, + { + "epoch": 1.613823473099751, + "grad_norm": 0.3015971779823303, + "learning_rate": 2.2490512501694394e-06, + "loss": 0.6141, + "step": 4214 + }, + { + "epoch": 1.614206394792265, + "grad_norm": 0.30390268564224243, + "learning_rate": 2.2480407558508997e-06, + "loss": 0.6261, + "step": 4215 + }, + { + "epoch": 1.6145893164847789, + "grad_norm": 0.3014738857746124, + "learning_rate": 2.247030303117103e-06, + "loss": 0.6194, + "step": 4216 + }, + { + "epoch": 1.6149722381772929, + "grad_norm": 0.3104712963104248, + "learning_rate": 2.2460198921348188e-06, + "loss": 0.6238, + "step": 4217 + }, + { + "epoch": 1.6153551598698066, + "grad_norm": 0.3029942512512207, + "learning_rate": 2.245009523070813e-06, + "loss": 0.6226, + "step": 4218 + }, + { + "epoch": 1.6157380815623203, + "grad_norm": 0.3158225119113922, + "learning_rate": 2.2439991960918396e-06, + "loss": 0.6113, + "step": 4219 + }, + { + "epoch": 1.6161210032548343, + "grad_norm": 0.30687978863716125, + "learning_rate": 2.2429889113646503e-06, + "loss": 0.6398, + "step": 4220 + }, + { + "epoch": 1.6165039249473483, + "grad_norm": 0.30504411458969116, + "learning_rate": 2.241978669055987e-06, + "loss": 0.6257, + "step": 4221 + }, + { + "epoch": 1.6168868466398623, + "grad_norm": 0.3155345916748047, + "learning_rate": 2.2409684693325876e-06, + "loss": 0.6382, + "step": 4222 + }, + { + "epoch": 1.617269768332376, + "grad_norm": 0.2990438938140869, + "learning_rate": 2.2399583123611788e-06, + "loss": 0.6123, + "step": 4223 + }, + { + "epoch": 1.6176526900248898, + "grad_norm": 0.33353888988494873, + "learning_rate": 2.238948198308484e-06, + "loss": 0.6413, + "step": 4224 + }, + { + "epoch": 1.6180356117174037, + "grad_norm": 0.30949491262435913, + "learning_rate": 2.237938127341218e-06, + "loss": 0.6491, + "step": 4225 + }, + { + "epoch": 1.6184185334099177, + "grad_norm": 0.3024851679801941, + "learning_rate": 2.236928099626087e-06, + "loss": 0.6194, + "step": 4226 + }, + { + "epoch": 1.6188014551024317, + "grad_norm": 0.3197498917579651, + "learning_rate": 2.235918115329793e-06, + "loss": 0.6205, + "step": 4227 + }, + { + "epoch": 1.6191843767949454, + "grad_norm": 0.31193023920059204, + "learning_rate": 2.2349081746190284e-06, + "loss": 0.6292, + "step": 4228 + }, + { + "epoch": 1.6195672984874592, + "grad_norm": 0.3097589910030365, + "learning_rate": 2.23389827766048e-06, + "loss": 0.6332, + "step": 4229 + }, + { + "epoch": 1.6199502201799731, + "grad_norm": 0.30342933535575867, + "learning_rate": 2.2328884246208262e-06, + "loss": 0.6388, + "step": 4230 + }, + { + "epoch": 1.6203331418724871, + "grad_norm": 0.3145466446876526, + "learning_rate": 2.23187861566674e-06, + "loss": 0.6208, + "step": 4231 + }, + { + "epoch": 1.620716063565001, + "grad_norm": 0.3174145817756653, + "learning_rate": 2.2308688509648836e-06, + "loss": 0.6059, + "step": 4232 + }, + { + "epoch": 1.6210989852575148, + "grad_norm": 0.29978421330451965, + "learning_rate": 2.2298591306819172e-06, + "loss": 0.6512, + "step": 4233 + }, + { + "epoch": 1.6214819069500286, + "grad_norm": 0.3045039474964142, + "learning_rate": 2.2288494549844884e-06, + "loss": 0.6181, + "step": 4234 + }, + { + "epoch": 1.6218648286425426, + "grad_norm": 0.3016273081302643, + "learning_rate": 2.227839824039239e-06, + "loss": 0.6293, + "step": 4235 + }, + { + "epoch": 1.6222477503350565, + "grad_norm": 0.2966153621673584, + "learning_rate": 2.226830238012806e-06, + "loss": 0.6316, + "step": 4236 + }, + { + "epoch": 1.6226306720275705, + "grad_norm": 0.31372833251953125, + "learning_rate": 2.2258206970718156e-06, + "loss": 0.6042, + "step": 4237 + }, + { + "epoch": 1.6230135937200842, + "grad_norm": 0.312318354845047, + "learning_rate": 2.2248112013828895e-06, + "loss": 0.6141, + "step": 4238 + }, + { + "epoch": 1.623396515412598, + "grad_norm": 0.30682802200317383, + "learning_rate": 2.2238017511126388e-06, + "loss": 0.6172, + "step": 4239 + }, + { + "epoch": 1.623779437105112, + "grad_norm": 0.31137460470199585, + "learning_rate": 2.2227923464276704e-06, + "loss": 0.6281, + "step": 4240 + }, + { + "epoch": 1.624162358797626, + "grad_norm": 0.3106917440891266, + "learning_rate": 2.2217829874945807e-06, + "loss": 0.6163, + "step": 4241 + }, + { + "epoch": 1.62454528049014, + "grad_norm": 0.29638421535491943, + "learning_rate": 2.2207736744799598e-06, + "loss": 0.6153, + "step": 4242 + }, + { + "epoch": 1.6249282021826537, + "grad_norm": 0.3052218556404114, + "learning_rate": 2.219764407550391e-06, + "loss": 0.6182, + "step": 4243 + }, + { + "epoch": 1.6253111238751674, + "grad_norm": 0.3224520981311798, + "learning_rate": 2.2187551868724487e-06, + "loss": 0.6379, + "step": 4244 + }, + { + "epoch": 1.6256940455676814, + "grad_norm": 0.3082274794578552, + "learning_rate": 2.217746012612701e-06, + "loss": 0.627, + "step": 4245 + }, + { + "epoch": 1.6260769672601953, + "grad_norm": 0.3059547543525696, + "learning_rate": 2.2167368849377064e-06, + "loss": 0.6149, + "step": 4246 + }, + { + "epoch": 1.6264598889527093, + "grad_norm": 0.3203917145729065, + "learning_rate": 2.215727804014019e-06, + "loss": 0.6268, + "step": 4247 + }, + { + "epoch": 1.626842810645223, + "grad_norm": 0.3056049644947052, + "learning_rate": 2.2147187700081813e-06, + "loss": 0.6105, + "step": 4248 + }, + { + "epoch": 1.6272257323377368, + "grad_norm": 0.3108392655849457, + "learning_rate": 2.2137097830867295e-06, + "loss": 0.6213, + "step": 4249 + }, + { + "epoch": 1.6276086540302508, + "grad_norm": 0.30117565393447876, + "learning_rate": 2.212700843416193e-06, + "loss": 0.6247, + "step": 4250 + }, + { + "epoch": 1.6279915757227648, + "grad_norm": 0.3060005307197571, + "learning_rate": 2.2116919511630937e-06, + "loss": 0.6206, + "step": 4251 + }, + { + "epoch": 1.6283744974152787, + "grad_norm": 0.31629350781440735, + "learning_rate": 2.210683106493943e-06, + "loss": 0.6359, + "step": 4252 + }, + { + "epoch": 1.6287574191077925, + "grad_norm": 0.3114181458950043, + "learning_rate": 2.209674309575248e-06, + "loss": 0.6208, + "step": 4253 + }, + { + "epoch": 1.6291403408003062, + "grad_norm": 0.30304333567619324, + "learning_rate": 2.2086655605735054e-06, + "loss": 0.6249, + "step": 4254 + }, + { + "epoch": 1.6295232624928202, + "grad_norm": 0.31625843048095703, + "learning_rate": 2.2076568596552036e-06, + "loss": 0.6143, + "step": 4255 + }, + { + "epoch": 1.6299061841853342, + "grad_norm": 0.3006319999694824, + "learning_rate": 2.2066482069868255e-06, + "loss": 0.6253, + "step": 4256 + }, + { + "epoch": 1.6302891058778481, + "grad_norm": 0.3161226212978363, + "learning_rate": 2.205639602734844e-06, + "loss": 0.6087, + "step": 4257 + }, + { + "epoch": 1.630672027570362, + "grad_norm": 0.30929625034332275, + "learning_rate": 2.204631047065725e-06, + "loss": 0.6148, + "step": 4258 + }, + { + "epoch": 1.6310549492628756, + "grad_norm": 0.3107418715953827, + "learning_rate": 2.2036225401459255e-06, + "loss": 0.6163, + "step": 4259 + }, + { + "epoch": 1.6314378709553896, + "grad_norm": 0.3194058835506439, + "learning_rate": 2.2026140821418966e-06, + "loss": 0.6347, + "step": 4260 + }, + { + "epoch": 1.6318207926479036, + "grad_norm": 0.3140994906425476, + "learning_rate": 2.2016056732200794e-06, + "loss": 0.605, + "step": 4261 + }, + { + "epoch": 1.6322037143404173, + "grad_norm": 0.3139664828777313, + "learning_rate": 2.200597313546905e-06, + "loss": 0.6358, + "step": 4262 + }, + { + "epoch": 1.6325866360329313, + "grad_norm": 0.29905828833580017, + "learning_rate": 2.199589003288801e-06, + "loss": 0.6163, + "step": 4263 + }, + { + "epoch": 1.632969557725445, + "grad_norm": 0.31295666098594666, + "learning_rate": 2.198580742612183e-06, + "loss": 0.6094, + "step": 4264 + }, + { + "epoch": 1.633352479417959, + "grad_norm": 0.30487769842147827, + "learning_rate": 2.197572531683462e-06, + "loss": 0.6388, + "step": 4265 + }, + { + "epoch": 1.633735401110473, + "grad_norm": 0.30184340476989746, + "learning_rate": 2.1965643706690357e-06, + "loss": 0.6444, + "step": 4266 + }, + { + "epoch": 1.6341183228029867, + "grad_norm": 0.3045661151409149, + "learning_rate": 2.1955562597352996e-06, + "loss": 0.6431, + "step": 4267 + }, + { + "epoch": 1.6345012444955007, + "grad_norm": 0.30941420793533325, + "learning_rate": 2.1945481990486366e-06, + "loss": 0.6314, + "step": 4268 + }, + { + "epoch": 1.6348841661880145, + "grad_norm": 0.31236281991004944, + "learning_rate": 2.1935401887754213e-06, + "loss": 0.641, + "step": 4269 + }, + { + "epoch": 1.6352670878805284, + "grad_norm": 0.30340245366096497, + "learning_rate": 2.192532229082023e-06, + "loss": 0.6389, + "step": 4270 + }, + { + "epoch": 1.6356500095730424, + "grad_norm": 0.31498265266418457, + "learning_rate": 2.1915243201348e-06, + "loss": 0.625, + "step": 4271 + }, + { + "epoch": 1.6360329312655562, + "grad_norm": 0.293312668800354, + "learning_rate": 2.190516462100104e-06, + "loss": 0.617, + "step": 4272 + }, + { + "epoch": 1.6364158529580701, + "grad_norm": 0.30899563431739807, + "learning_rate": 2.189508655144276e-06, + "loss": 0.6255, + "step": 4273 + }, + { + "epoch": 1.6367987746505839, + "grad_norm": 0.3097638487815857, + "learning_rate": 2.188500899433651e-06, + "loss": 0.6138, + "step": 4274 + }, + { + "epoch": 1.6371816963430978, + "grad_norm": 0.3064741790294647, + "learning_rate": 2.187493195134557e-06, + "loss": 0.646, + "step": 4275 + }, + { + "epoch": 1.6375646180356118, + "grad_norm": 0.3157450258731842, + "learning_rate": 2.186485542413306e-06, + "loss": 0.625, + "step": 4276 + }, + { + "epoch": 1.6379475397281256, + "grad_norm": 0.3201080560684204, + "learning_rate": 2.1854779414362097e-06, + "loss": 0.6339, + "step": 4277 + }, + { + "epoch": 1.6383304614206393, + "grad_norm": 0.30533677339553833, + "learning_rate": 2.1844703923695674e-06, + "loss": 0.6204, + "step": 4278 + }, + { + "epoch": 1.6387133831131533, + "grad_norm": 0.3024739623069763, + "learning_rate": 2.183462895379671e-06, + "loss": 0.6264, + "step": 4279 + }, + { + "epoch": 1.6390963048056673, + "grad_norm": 0.3194553852081299, + "learning_rate": 2.1824554506328033e-06, + "loss": 0.6188, + "step": 4280 + }, + { + "epoch": 1.6394792264981812, + "grad_norm": 0.3232802450656891, + "learning_rate": 2.1814480582952376e-06, + "loss": 0.628, + "step": 4281 + }, + { + "epoch": 1.639862148190695, + "grad_norm": 0.3055027723312378, + "learning_rate": 2.1804407185332416e-06, + "loss": 0.6281, + "step": 4282 + }, + { + "epoch": 1.6402450698832087, + "grad_norm": 0.31738170981407166, + "learning_rate": 2.17943343151307e-06, + "loss": 0.621, + "step": 4283 + }, + { + "epoch": 1.6406279915757227, + "grad_norm": 0.31120410561561584, + "learning_rate": 2.1784261974009717e-06, + "loss": 0.6385, + "step": 4284 + }, + { + "epoch": 1.6410109132682367, + "grad_norm": 0.313363641500473, + "learning_rate": 2.177419016363186e-06, + "loss": 0.6511, + "step": 4285 + }, + { + "epoch": 1.6413938349607506, + "grad_norm": 0.3335781991481781, + "learning_rate": 2.176411888565944e-06, + "loss": 0.6419, + "step": 4286 + }, + { + "epoch": 1.6417767566532644, + "grad_norm": 0.31257423758506775, + "learning_rate": 2.175404814175468e-06, + "loss": 0.6312, + "step": 4287 + }, + { + "epoch": 1.6421596783457781, + "grad_norm": 0.3190150260925293, + "learning_rate": 2.1743977933579695e-06, + "loss": 0.6223, + "step": 4288 + }, + { + "epoch": 1.642542600038292, + "grad_norm": 0.3033623695373535, + "learning_rate": 2.1733908262796557e-06, + "loss": 0.6492, + "step": 4289 + }, + { + "epoch": 1.642925521730806, + "grad_norm": 0.30958983302116394, + "learning_rate": 2.172383913106719e-06, + "loss": 0.6397, + "step": 4290 + }, + { + "epoch": 1.64330844342332, + "grad_norm": 0.3213809132575989, + "learning_rate": 2.1713770540053474e-06, + "loss": 0.6387, + "step": 4291 + }, + { + "epoch": 1.6436913651158338, + "grad_norm": 0.313606321811676, + "learning_rate": 2.170370249141718e-06, + "loss": 0.6311, + "step": 4292 + }, + { + "epoch": 1.6440742868083476, + "grad_norm": 0.3188220262527466, + "learning_rate": 2.169363498681999e-06, + "loss": 0.6239, + "step": 4293 + }, + { + "epoch": 1.6444572085008615, + "grad_norm": 0.30684661865234375, + "learning_rate": 2.168356802792351e-06, + "loss": 0.6269, + "step": 4294 + }, + { + "epoch": 1.6448401301933755, + "grad_norm": 0.30644869804382324, + "learning_rate": 2.1673501616389243e-06, + "loss": 0.6219, + "step": 4295 + }, + { + "epoch": 1.6452230518858895, + "grad_norm": 0.3077220916748047, + "learning_rate": 2.1663435753878616e-06, + "loss": 0.6339, + "step": 4296 + }, + { + "epoch": 1.6456059735784032, + "grad_norm": 0.29906147718429565, + "learning_rate": 2.1653370442052932e-06, + "loss": 0.6377, + "step": 4297 + }, + { + "epoch": 1.645988895270917, + "grad_norm": 0.3120008111000061, + "learning_rate": 2.1643305682573435e-06, + "loss": 0.6419, + "step": 4298 + }, + { + "epoch": 1.646371816963431, + "grad_norm": 0.3241884708404541, + "learning_rate": 2.163324147710127e-06, + "loss": 0.6425, + "step": 4299 + }, + { + "epoch": 1.646754738655945, + "grad_norm": 0.3280892074108124, + "learning_rate": 2.162317782729749e-06, + "loss": 0.626, + "step": 4300 + }, + { + "epoch": 1.6471376603484589, + "grad_norm": 0.3079608380794525, + "learning_rate": 2.161311473482305e-06, + "loss": 0.641, + "step": 4301 + }, + { + "epoch": 1.6475205820409726, + "grad_norm": 0.30884504318237305, + "learning_rate": 2.160305220133883e-06, + "loss": 0.6267, + "step": 4302 + }, + { + "epoch": 1.6479035037334864, + "grad_norm": 0.3278655707836151, + "learning_rate": 2.1592990228505607e-06, + "loss": 0.6252, + "step": 4303 + }, + { + "epoch": 1.6482864254260003, + "grad_norm": 0.30765900015830994, + "learning_rate": 2.158292881798404e-06, + "loss": 0.6282, + "step": 4304 + }, + { + "epoch": 1.6486693471185143, + "grad_norm": 0.31709206104278564, + "learning_rate": 2.157286797143474e-06, + "loss": 0.6552, + "step": 4305 + }, + { + "epoch": 1.6490522688110283, + "grad_norm": 0.31247928738594055, + "learning_rate": 2.15628076905182e-06, + "loss": 0.6357, + "step": 4306 + }, + { + "epoch": 1.649435190503542, + "grad_norm": 0.3151305615901947, + "learning_rate": 2.1552747976894826e-06, + "loss": 0.6123, + "step": 4307 + }, + { + "epoch": 1.6498181121960558, + "grad_norm": 0.3128744959831238, + "learning_rate": 2.154268883222492e-06, + "loss": 0.6298, + "step": 4308 + }, + { + "epoch": 1.6502010338885698, + "grad_norm": 0.32046979665756226, + "learning_rate": 2.153263025816871e-06, + "loss": 0.6265, + "step": 4309 + }, + { + "epoch": 1.6505839555810837, + "grad_norm": 0.3119691014289856, + "learning_rate": 2.1522572256386326e-06, + "loss": 0.6334, + "step": 4310 + }, + { + "epoch": 1.6509668772735977, + "grad_norm": 0.30906015634536743, + "learning_rate": 2.151251482853777e-06, + "loss": 0.6196, + "step": 4311 + }, + { + "epoch": 1.6513497989661114, + "grad_norm": 0.3093790113925934, + "learning_rate": 2.1502457976282996e-06, + "loss": 0.631, + "step": 4312 + }, + { + "epoch": 1.6517327206586252, + "grad_norm": 0.33127719163894653, + "learning_rate": 2.1492401701281827e-06, + "loss": 0.6402, + "step": 4313 + }, + { + "epoch": 1.6521156423511392, + "grad_norm": 0.3117757737636566, + "learning_rate": 2.148234600519402e-06, + "loss": 0.624, + "step": 4314 + }, + { + "epoch": 1.6524985640436531, + "grad_norm": 0.30520617961883545, + "learning_rate": 2.1472290889679215e-06, + "loss": 0.6233, + "step": 4315 + }, + { + "epoch": 1.652881485736167, + "grad_norm": 0.32380589842796326, + "learning_rate": 2.1462236356396972e-06, + "loss": 0.6441, + "step": 4316 + }, + { + "epoch": 1.6532644074286809, + "grad_norm": 0.3164161145687103, + "learning_rate": 2.145218240700673e-06, + "loss": 0.6252, + "step": 4317 + }, + { + "epoch": 1.6536473291211946, + "grad_norm": 0.30982843041419983, + "learning_rate": 2.1442129043167877e-06, + "loss": 0.6298, + "step": 4318 + }, + { + "epoch": 1.6540302508137086, + "grad_norm": 0.32967594265937805, + "learning_rate": 2.143207626653965e-06, + "loss": 0.6171, + "step": 4319 + }, + { + "epoch": 1.6544131725062226, + "grad_norm": 0.3048552870750427, + "learning_rate": 2.142202407878121e-06, + "loss": 0.6349, + "step": 4320 + }, + { + "epoch": 1.6547960941987365, + "grad_norm": 0.3223854899406433, + "learning_rate": 2.141197248155165e-06, + "loss": 0.6295, + "step": 4321 + }, + { + "epoch": 1.6551790158912503, + "grad_norm": 0.32866907119750977, + "learning_rate": 2.1401921476509925e-06, + "loss": 0.6014, + "step": 4322 + }, + { + "epoch": 1.655561937583764, + "grad_norm": 0.3206460475921631, + "learning_rate": 2.1391871065314914e-06, + "loss": 0.6364, + "step": 4323 + }, + { + "epoch": 1.655944859276278, + "grad_norm": 0.3125913739204407, + "learning_rate": 2.1381821249625383e-06, + "loss": 0.6093, + "step": 4324 + }, + { + "epoch": 1.656327780968792, + "grad_norm": 0.3181209862232208, + "learning_rate": 2.1371772031100025e-06, + "loss": 0.6371, + "step": 4325 + }, + { + "epoch": 1.6567107026613057, + "grad_norm": 0.30531325936317444, + "learning_rate": 2.1361723411397404e-06, + "loss": 0.6383, + "step": 4326 + }, + { + "epoch": 1.6570936243538197, + "grad_norm": 0.30637025833129883, + "learning_rate": 2.1351675392176006e-06, + "loss": 0.6498, + "step": 4327 + }, + { + "epoch": 1.6574765460463334, + "grad_norm": 0.3116320073604584, + "learning_rate": 2.1341627975094207e-06, + "loss": 0.6404, + "step": 4328 + }, + { + "epoch": 1.6578594677388474, + "grad_norm": 0.3063797354698181, + "learning_rate": 2.13315811618103e-06, + "loss": 0.6266, + "step": 4329 + }, + { + "epoch": 1.6582423894313614, + "grad_norm": 0.306890070438385, + "learning_rate": 2.132153495398245e-06, + "loss": 0.6191, + "step": 4330 + }, + { + "epoch": 1.6586253111238751, + "grad_norm": 0.31118789315223694, + "learning_rate": 2.131148935326875e-06, + "loss": 0.6367, + "step": 4331 + }, + { + "epoch": 1.659008232816389, + "grad_norm": 0.3025604486465454, + "learning_rate": 2.130144436132718e-06, + "loss": 0.6296, + "step": 4332 + }, + { + "epoch": 1.6593911545089028, + "grad_norm": 0.3147870898246765, + "learning_rate": 2.129139997981562e-06, + "loss": 0.636, + "step": 4333 + }, + { + "epoch": 1.6597740762014168, + "grad_norm": 0.29525622725486755, + "learning_rate": 2.1281356210391845e-06, + "loss": 0.608, + "step": 4334 + }, + { + "epoch": 1.6601569978939308, + "grad_norm": 0.304660826921463, + "learning_rate": 2.1271313054713534e-06, + "loss": 0.6327, + "step": 4335 + }, + { + "epoch": 1.6605399195864445, + "grad_norm": 0.3029381334781647, + "learning_rate": 2.1261270514438277e-06, + "loss": 0.6231, + "step": 4336 + }, + { + "epoch": 1.6609228412789585, + "grad_norm": 0.30060598254203796, + "learning_rate": 2.1251228591223533e-06, + "loss": 0.6229, + "step": 4337 + }, + { + "epoch": 1.6613057629714723, + "grad_norm": 0.3122200071811676, + "learning_rate": 2.124118728672669e-06, + "loss": 0.6382, + "step": 4338 + }, + { + "epoch": 1.6616886846639862, + "grad_norm": 0.33768031001091003, + "learning_rate": 2.1231146602605023e-06, + "loss": 0.6412, + "step": 4339 + }, + { + "epoch": 1.6620716063565002, + "grad_norm": 0.31210458278656006, + "learning_rate": 2.1221106540515685e-06, + "loss": 0.6197, + "step": 4340 + }, + { + "epoch": 1.662454528049014, + "grad_norm": 0.31238964200019836, + "learning_rate": 2.1211067102115756e-06, + "loss": 0.6208, + "step": 4341 + }, + { + "epoch": 1.6628374497415277, + "grad_norm": 0.30566897988319397, + "learning_rate": 2.1201028289062193e-06, + "loss": 0.6441, + "step": 4342 + }, + { + "epoch": 1.6632203714340417, + "grad_norm": 0.3068365454673767, + "learning_rate": 2.1190990103011865e-06, + "loss": 0.6414, + "step": 4343 + }, + { + "epoch": 1.6636032931265556, + "grad_norm": 0.314069539308548, + "learning_rate": 2.118095254562152e-06, + "loss": 0.6303, + "step": 4344 + }, + { + "epoch": 1.6639862148190696, + "grad_norm": 0.31680190563201904, + "learning_rate": 2.1170915618547823e-06, + "loss": 0.6302, + "step": 4345 + }, + { + "epoch": 1.6643691365115834, + "grad_norm": 0.3049539029598236, + "learning_rate": 2.116087932344732e-06, + "loss": 0.627, + "step": 4346 + }, + { + "epoch": 1.664752058204097, + "grad_norm": 0.3042599558830261, + "learning_rate": 2.1150843661976447e-06, + "loss": 0.6421, + "step": 4347 + }, + { + "epoch": 1.665134979896611, + "grad_norm": 0.33182117342948914, + "learning_rate": 2.1140808635791558e-06, + "loss": 0.6259, + "step": 4348 + }, + { + "epoch": 1.665517901589125, + "grad_norm": 0.31740811467170715, + "learning_rate": 2.1130774246548878e-06, + "loss": 0.6364, + "step": 4349 + }, + { + "epoch": 1.665900823281639, + "grad_norm": 0.3056165277957916, + "learning_rate": 2.1120740495904547e-06, + "loss": 0.6466, + "step": 4350 + }, + { + "epoch": 1.6662837449741528, + "grad_norm": 0.3141825199127197, + "learning_rate": 2.1110707385514578e-06, + "loss": 0.6252, + "step": 4351 + }, + { + "epoch": 1.6666666666666665, + "grad_norm": 0.33872079849243164, + "learning_rate": 2.110067491703491e-06, + "loss": 0.6082, + "step": 4352 + }, + { + "epoch": 1.6670495883591805, + "grad_norm": 0.3108309507369995, + "learning_rate": 2.109064309212135e-06, + "loss": 0.643, + "step": 4353 + }, + { + "epoch": 1.6674325100516945, + "grad_norm": 0.3081771433353424, + "learning_rate": 2.1080611912429592e-06, + "loss": 0.6186, + "step": 4354 + }, + { + "epoch": 1.6678154317442084, + "grad_norm": 0.313401997089386, + "learning_rate": 2.1070581379615253e-06, + "loss": 0.6169, + "step": 4355 + }, + { + "epoch": 1.6681983534367222, + "grad_norm": 0.31228187680244446, + "learning_rate": 2.1060551495333816e-06, + "loss": 0.639, + "step": 4356 + }, + { + "epoch": 1.668581275129236, + "grad_norm": 0.3136352598667145, + "learning_rate": 2.1050522261240676e-06, + "loss": 0.6258, + "step": 4357 + }, + { + "epoch": 1.66896419682175, + "grad_norm": 0.3131379187107086, + "learning_rate": 2.104049367899111e-06, + "loss": 0.6267, + "step": 4358 + }, + { + "epoch": 1.6693471185142639, + "grad_norm": 0.31512364745140076, + "learning_rate": 2.1030465750240296e-06, + "loss": 0.6327, + "step": 4359 + }, + { + "epoch": 1.6697300402067778, + "grad_norm": 0.31374433636665344, + "learning_rate": 2.10204384766433e-06, + "loss": 0.6261, + "step": 4360 + }, + { + "epoch": 1.6701129618992916, + "grad_norm": 0.3361717760562897, + "learning_rate": 2.1010411859855064e-06, + "loss": 0.6366, + "step": 4361 + }, + { + "epoch": 1.6704958835918053, + "grad_norm": 0.31218063831329346, + "learning_rate": 2.1000385901530447e-06, + "loss": 0.6368, + "step": 4362 + }, + { + "epoch": 1.6708788052843193, + "grad_norm": 0.3117731213569641, + "learning_rate": 2.0990360603324186e-06, + "loss": 0.6543, + "step": 4363 + }, + { + "epoch": 1.6712617269768333, + "grad_norm": 0.31845220923423767, + "learning_rate": 2.098033596689091e-06, + "loss": 0.6422, + "step": 4364 + }, + { + "epoch": 1.6716446486693473, + "grad_norm": 0.34159037470817566, + "learning_rate": 2.0970311993885143e-06, + "loss": 0.6202, + "step": 4365 + }, + { + "epoch": 1.672027570361861, + "grad_norm": 0.3347354531288147, + "learning_rate": 2.0960288685961295e-06, + "loss": 0.6245, + "step": 4366 + }, + { + "epoch": 1.6724104920543748, + "grad_norm": 0.3165040612220764, + "learning_rate": 2.0950266044773677e-06, + "loss": 0.6198, + "step": 4367 + }, + { + "epoch": 1.6727934137468887, + "grad_norm": 0.3160633444786072, + "learning_rate": 2.0940244071976466e-06, + "loss": 0.6235, + "step": 4368 + }, + { + "epoch": 1.6731763354394027, + "grad_norm": 0.32191571593284607, + "learning_rate": 2.093022276922375e-06, + "loss": 0.6178, + "step": 4369 + }, + { + "epoch": 1.6735592571319167, + "grad_norm": 0.31007322669029236, + "learning_rate": 2.09202021381695e-06, + "loss": 0.62, + "step": 4370 + }, + { + "epoch": 1.6739421788244304, + "grad_norm": 0.3105906844139099, + "learning_rate": 2.0910182180467577e-06, + "loss": 0.6382, + "step": 4371 + }, + { + "epoch": 1.6743251005169442, + "grad_norm": 0.3185935914516449, + "learning_rate": 2.090016289777173e-06, + "loss": 0.6441, + "step": 4372 + }, + { + "epoch": 1.6747080222094581, + "grad_norm": 0.31556636095046997, + "learning_rate": 2.0890144291735596e-06, + "loss": 0.6321, + "step": 4373 + }, + { + "epoch": 1.675090943901972, + "grad_norm": 0.32077932357788086, + "learning_rate": 2.088012636401271e-06, + "loss": 0.6202, + "step": 4374 + }, + { + "epoch": 1.675473865594486, + "grad_norm": 0.3444354236125946, + "learning_rate": 2.087010911625647e-06, + "loss": 0.634, + "step": 4375 + }, + { + "epoch": 1.6758567872869998, + "grad_norm": 0.3140331208705902, + "learning_rate": 2.086009255012018e-06, + "loss": 0.6322, + "step": 4376 + }, + { + "epoch": 1.6762397089795136, + "grad_norm": 0.3038775324821472, + "learning_rate": 2.085007666725704e-06, + "loss": 0.6197, + "step": 4377 + }, + { + "epoch": 1.6766226306720275, + "grad_norm": 0.32823866605758667, + "learning_rate": 2.0840061469320115e-06, + "loss": 0.6582, + "step": 4378 + }, + { + "epoch": 1.6770055523645415, + "grad_norm": 0.320975661277771, + "learning_rate": 2.083004695796238e-06, + "loss": 0.6258, + "step": 4379 + }, + { + "epoch": 1.6773884740570555, + "grad_norm": 0.3170107305049896, + "learning_rate": 2.082003313483668e-06, + "loss": 0.6145, + "step": 4380 + }, + { + "epoch": 1.6777713957495692, + "grad_norm": 0.312732070684433, + "learning_rate": 2.0810020001595756e-06, + "loss": 0.6422, + "step": 4381 + }, + { + "epoch": 1.678154317442083, + "grad_norm": 0.311092346906662, + "learning_rate": 2.080000755989222e-06, + "loss": 0.6269, + "step": 4382 + }, + { + "epoch": 1.678537239134597, + "grad_norm": 0.3137769103050232, + "learning_rate": 2.0789995811378584e-06, + "loss": 0.6378, + "step": 4383 + }, + { + "epoch": 1.678920160827111, + "grad_norm": 0.3164680004119873, + "learning_rate": 2.077998475770724e-06, + "loss": 0.6437, + "step": 4384 + }, + { + "epoch": 1.679303082519625, + "grad_norm": 0.31407684087753296, + "learning_rate": 2.0769974400530478e-06, + "loss": 0.6413, + "step": 4385 + }, + { + "epoch": 1.6796860042121387, + "grad_norm": 0.31553417444229126, + "learning_rate": 2.0759964741500454e-06, + "loss": 0.6286, + "step": 4386 + }, + { + "epoch": 1.6800689259046524, + "grad_norm": 0.3197893798351288, + "learning_rate": 2.0749955782269225e-06, + "loss": 0.6205, + "step": 4387 + }, + { + "epoch": 1.6804518475971664, + "grad_norm": 0.3159908354282379, + "learning_rate": 2.0739947524488724e-06, + "loss": 0.6253, + "step": 4388 + }, + { + "epoch": 1.6808347692896803, + "grad_norm": 0.3068619668483734, + "learning_rate": 2.0729939969810756e-06, + "loss": 0.6314, + "step": 4389 + }, + { + "epoch": 1.681217690982194, + "grad_norm": 0.31634727120399475, + "learning_rate": 2.0719933119887032e-06, + "loss": 0.6447, + "step": 4390 + }, + { + "epoch": 1.681600612674708, + "grad_norm": 0.3197683095932007, + "learning_rate": 2.070992697636914e-06, + "loss": 0.6338, + "step": 4391 + }, + { + "epoch": 1.6819835343672218, + "grad_norm": 0.3305831253528595, + "learning_rate": 2.0699921540908542e-06, + "loss": 0.6262, + "step": 4392 + }, + { + "epoch": 1.6823664560597358, + "grad_norm": 0.32171621918678284, + "learning_rate": 2.0689916815156595e-06, + "loss": 0.6294, + "step": 4393 + }, + { + "epoch": 1.6827493777522498, + "grad_norm": 0.29958459734916687, + "learning_rate": 2.0679912800764537e-06, + "loss": 0.6172, + "step": 4394 + }, + { + "epoch": 1.6831322994447635, + "grad_norm": 0.31191137433052063, + "learning_rate": 2.066990949938349e-06, + "loss": 0.6179, + "step": 4395 + }, + { + "epoch": 1.6835152211372775, + "grad_norm": 0.30481216311454773, + "learning_rate": 2.0659906912664435e-06, + "loss": 0.609, + "step": 4396 + }, + { + "epoch": 1.6838981428297912, + "grad_norm": 0.31317660212516785, + "learning_rate": 2.064990504225827e-06, + "loss": 0.6305, + "step": 4397 + }, + { + "epoch": 1.6842810645223052, + "grad_norm": 0.31992727518081665, + "learning_rate": 2.0639903889815747e-06, + "loss": 0.6381, + "step": 4398 + }, + { + "epoch": 1.6846639862148192, + "grad_norm": 0.3109089434146881, + "learning_rate": 2.0629903456987526e-06, + "loss": 0.6364, + "step": 4399 + }, + { + "epoch": 1.685046907907333, + "grad_norm": 0.3140223026275635, + "learning_rate": 2.061990374542412e-06, + "loss": 0.628, + "step": 4400 + }, + { + "epoch": 1.6854298295998469, + "grad_norm": 0.3099183738231659, + "learning_rate": 2.060990475677595e-06, + "loss": 0.6313, + "step": 4401 + }, + { + "epoch": 1.6858127512923606, + "grad_norm": 0.313439279794693, + "learning_rate": 2.059990649269329e-06, + "loss": 0.6349, + "step": 4402 + }, + { + "epoch": 1.6861956729848746, + "grad_norm": 0.3116743862628937, + "learning_rate": 2.0589908954826324e-06, + "loss": 0.6312, + "step": 4403 + }, + { + "epoch": 1.6865785946773886, + "grad_norm": 0.32145965099334717, + "learning_rate": 2.057991214482509e-06, + "loss": 0.6386, + "step": 4404 + }, + { + "epoch": 1.6869615163699023, + "grad_norm": 0.31221234798431396, + "learning_rate": 2.0569916064339513e-06, + "loss": 0.6265, + "step": 4405 + }, + { + "epoch": 1.687344438062416, + "grad_norm": 0.3237135410308838, + "learning_rate": 2.055992071501941e-06, + "loss": 0.6411, + "step": 4406 + }, + { + "epoch": 1.68772735975493, + "grad_norm": 0.31138432025909424, + "learning_rate": 2.054992609851446e-06, + "loss": 0.629, + "step": 4407 + }, + { + "epoch": 1.688110281447444, + "grad_norm": 0.29925647377967834, + "learning_rate": 2.0539932216474237e-06, + "loss": 0.6405, + "step": 4408 + }, + { + "epoch": 1.688493203139958, + "grad_norm": 0.2968170940876007, + "learning_rate": 2.0529939070548192e-06, + "loss": 0.6375, + "step": 4409 + }, + { + "epoch": 1.6888761248324717, + "grad_norm": 0.31298890709877014, + "learning_rate": 2.0519946662385642e-06, + "loss": 0.6352, + "step": 4410 + }, + { + "epoch": 1.6892590465249855, + "grad_norm": 0.3137513995170593, + "learning_rate": 2.050995499363578e-06, + "loss": 0.6185, + "step": 4411 + }, + { + "epoch": 1.6896419682174995, + "grad_norm": 0.29807335138320923, + "learning_rate": 2.04999640659477e-06, + "loss": 0.6181, + "step": 4412 + }, + { + "epoch": 1.6900248899100134, + "grad_norm": 0.3114708364009857, + "learning_rate": 2.048997388097035e-06, + "loss": 0.6287, + "step": 4413 + }, + { + "epoch": 1.6904078116025274, + "grad_norm": 0.3136223554611206, + "learning_rate": 2.0479984440352573e-06, + "loss": 0.6128, + "step": 4414 + }, + { + "epoch": 1.6907907332950411, + "grad_norm": 0.3096925616264343, + "learning_rate": 2.046999574574308e-06, + "loss": 0.6346, + "step": 4415 + }, + { + "epoch": 1.691173654987555, + "grad_norm": 0.3173500597476959, + "learning_rate": 2.0460007798790452e-06, + "loss": 0.6229, + "step": 4416 + }, + { + "epoch": 1.6915565766800689, + "grad_norm": 0.31414884328842163, + "learning_rate": 2.0450020601143174e-06, + "loss": 0.6202, + "step": 4417 + }, + { + "epoch": 1.6919394983725828, + "grad_norm": 0.3200772702693939, + "learning_rate": 2.0440034154449558e-06, + "loss": 0.6287, + "step": 4418 + }, + { + "epoch": 1.6923224200650968, + "grad_norm": 0.31523358821868896, + "learning_rate": 2.0430048460357848e-06, + "loss": 0.6088, + "step": 4419 + }, + { + "epoch": 1.6927053417576106, + "grad_norm": 0.319668710231781, + "learning_rate": 2.0420063520516125e-06, + "loss": 0.6173, + "step": 4420 + }, + { + "epoch": 1.6930882634501243, + "grad_norm": 0.31046798825263977, + "learning_rate": 2.041007933657237e-06, + "loss": 0.6119, + "step": 4421 + }, + { + "epoch": 1.6934711851426383, + "grad_norm": 0.30751627683639526, + "learning_rate": 2.040009591017441e-06, + "loss": 0.6068, + "step": 4422 + }, + { + "epoch": 1.6938541068351523, + "grad_norm": 0.3115965723991394, + "learning_rate": 2.0390113242969984e-06, + "loss": 0.6391, + "step": 4423 + }, + { + "epoch": 1.6942370285276662, + "grad_norm": 0.31340187788009644, + "learning_rate": 2.0380131336606682e-06, + "loss": 0.6453, + "step": 4424 + }, + { + "epoch": 1.69461995022018, + "grad_norm": 0.3054353892803192, + "learning_rate": 2.037015019273196e-06, + "loss": 0.6207, + "step": 4425 + }, + { + "epoch": 1.6950028719126937, + "grad_norm": 0.317080557346344, + "learning_rate": 2.0360169812993175e-06, + "loss": 0.6347, + "step": 4426 + }, + { + "epoch": 1.6953857936052077, + "grad_norm": 0.3091543912887573, + "learning_rate": 2.0350190199037533e-06, + "loss": 0.6238, + "step": 4427 + }, + { + "epoch": 1.6957687152977217, + "grad_norm": 0.3062279224395752, + "learning_rate": 2.034021135251213e-06, + "loss": 0.6279, + "step": 4428 + }, + { + "epoch": 1.6961516369902356, + "grad_norm": 0.31509119272232056, + "learning_rate": 2.033023327506393e-06, + "loss": 0.6275, + "step": 4429 + }, + { + "epoch": 1.6965345586827494, + "grad_norm": 0.3042357861995697, + "learning_rate": 2.032025596833977e-06, + "loss": 0.6119, + "step": 4430 + }, + { + "epoch": 1.6969174803752631, + "grad_norm": 0.30593517422676086, + "learning_rate": 2.031027943398637e-06, + "loss": 0.6141, + "step": 4431 + }, + { + "epoch": 1.697300402067777, + "grad_norm": 0.3061474859714508, + "learning_rate": 2.0300303673650286e-06, + "loss": 0.617, + "step": 4432 + }, + { + "epoch": 1.697683323760291, + "grad_norm": 0.3200860619544983, + "learning_rate": 2.0290328688977996e-06, + "loss": 0.6175, + "step": 4433 + }, + { + "epoch": 1.698066245452805, + "grad_norm": 0.31428831815719604, + "learning_rate": 2.0280354481615814e-06, + "loss": 0.6585, + "step": 4434 + }, + { + "epoch": 1.6984491671453188, + "grad_norm": 0.30109256505966187, + "learning_rate": 2.027038105320994e-06, + "loss": 0.6062, + "step": 4435 + }, + { + "epoch": 1.6988320888378325, + "grad_norm": 0.3174463212490082, + "learning_rate": 2.0260408405406455e-06, + "loss": 0.6133, + "step": 4436 + }, + { + "epoch": 1.6992150105303465, + "grad_norm": 0.3162376880645752, + "learning_rate": 2.025043653985129e-06, + "loss": 0.6363, + "step": 4437 + }, + { + "epoch": 1.6995979322228605, + "grad_norm": 0.3020956516265869, + "learning_rate": 2.024046545819026e-06, + "loss": 0.6387, + "step": 4438 + }, + { + "epoch": 1.6999808539153745, + "grad_norm": 0.3217693865299225, + "learning_rate": 2.0230495162069043e-06, + "loss": 0.6291, + "step": 4439 + }, + { + "epoch": 1.7003637756078882, + "grad_norm": 0.3088738024234772, + "learning_rate": 2.022052565313318e-06, + "loss": 0.629, + "step": 4440 + }, + { + "epoch": 1.700746697300402, + "grad_norm": 0.29854699969291687, + "learning_rate": 2.0210556933028123e-06, + "loss": 0.6199, + "step": 4441 + }, + { + "epoch": 1.701129618992916, + "grad_norm": 0.3101494014263153, + "learning_rate": 2.020058900339914e-06, + "loss": 0.6267, + "step": 4442 + }, + { + "epoch": 1.70151254068543, + "grad_norm": 0.3157016634941101, + "learning_rate": 2.0190621865891408e-06, + "loss": 0.6129, + "step": 4443 + }, + { + "epoch": 1.7018954623779439, + "grad_norm": 0.3054714500904083, + "learning_rate": 2.018065552214995e-06, + "loss": 0.6349, + "step": 4444 + }, + { + "epoch": 1.7022783840704576, + "grad_norm": 0.3130691647529602, + "learning_rate": 2.017068997381967e-06, + "loss": 0.6259, + "step": 4445 + }, + { + "epoch": 1.7026613057629714, + "grad_norm": 0.3341308534145355, + "learning_rate": 2.016072522254534e-06, + "loss": 0.6197, + "step": 4446 + }, + { + "epoch": 1.7030442274554853, + "grad_norm": 0.31293797492980957, + "learning_rate": 2.015076126997158e-06, + "loss": 0.6321, + "step": 4447 + }, + { + "epoch": 1.7034271491479993, + "grad_norm": 0.30630922317504883, + "learning_rate": 2.014079811774292e-06, + "loss": 0.6203, + "step": 4448 + }, + { + "epoch": 1.7038100708405133, + "grad_norm": 0.3240378797054291, + "learning_rate": 2.0130835767503712e-06, + "loss": 0.6295, + "step": 4449 + }, + { + "epoch": 1.704192992533027, + "grad_norm": 0.32213348150253296, + "learning_rate": 2.0120874220898216e-06, + "loss": 0.6388, + "step": 4450 + }, + { + "epoch": 1.7045759142255408, + "grad_norm": 0.31212958693504333, + "learning_rate": 2.0110913479570525e-06, + "loss": 0.6385, + "step": 4451 + }, + { + "epoch": 1.7049588359180547, + "grad_norm": 0.3041835427284241, + "learning_rate": 2.010095354516463e-06, + "loss": 0.6246, + "step": 4452 + }, + { + "epoch": 1.7053417576105687, + "grad_norm": 0.3237878084182739, + "learning_rate": 2.0090994419324363e-06, + "loss": 0.6289, + "step": 4453 + }, + { + "epoch": 1.7057246793030825, + "grad_norm": 0.32107916474342346, + "learning_rate": 2.008103610369343e-06, + "loss": 0.6298, + "step": 4454 + }, + { + "epoch": 1.7061076009955964, + "grad_norm": 0.3182363510131836, + "learning_rate": 2.0071078599915415e-06, + "loss": 0.6121, + "step": 4455 + }, + { + "epoch": 1.7064905226881102, + "grad_norm": 0.29557475447654724, + "learning_rate": 2.006112190963375e-06, + "loss": 0.6225, + "step": 4456 + }, + { + "epoch": 1.7068734443806242, + "grad_norm": 0.30374157428741455, + "learning_rate": 2.0051166034491755e-06, + "loss": 0.6063, + "step": 4457 + }, + { + "epoch": 1.7072563660731381, + "grad_norm": 0.3119162619113922, + "learning_rate": 2.0041210976132592e-06, + "loss": 0.6244, + "step": 4458 + }, + { + "epoch": 1.7076392877656519, + "grad_norm": 0.3261125385761261, + "learning_rate": 2.003125673619931e-06, + "loss": 0.6096, + "step": 4459 + }, + { + "epoch": 1.7080222094581659, + "grad_norm": 0.3153318762779236, + "learning_rate": 2.0021303316334793e-06, + "loss": 0.6436, + "step": 4460 + }, + { + "epoch": 1.7084051311506796, + "grad_norm": 0.3133585751056671, + "learning_rate": 2.001135071818182e-06, + "loss": 0.6294, + "step": 4461 + }, + { + "epoch": 1.7087880528431936, + "grad_norm": 0.3087129294872284, + "learning_rate": 2.000139894338302e-06, + "loss": 0.6319, + "step": 4462 + }, + { + "epoch": 1.7091709745357075, + "grad_norm": 0.32635125517845154, + "learning_rate": 1.9991447993580896e-06, + "loss": 0.6412, + "step": 4463 + }, + { + "epoch": 1.7095538962282213, + "grad_norm": 0.30250707268714905, + "learning_rate": 1.9981497870417797e-06, + "loss": 0.6283, + "step": 4464 + }, + { + "epoch": 1.7099368179207353, + "grad_norm": 0.3032040596008301, + "learning_rate": 1.9971548575535955e-06, + "loss": 0.627, + "step": 4465 + }, + { + "epoch": 1.710319739613249, + "grad_norm": 0.30283063650131226, + "learning_rate": 1.996160011057746e-06, + "loss": 0.642, + "step": 4466 + }, + { + "epoch": 1.710702661305763, + "grad_norm": 0.3109924793243408, + "learning_rate": 1.995165247718424e-06, + "loss": 0.631, + "step": 4467 + }, + { + "epoch": 1.711085582998277, + "grad_norm": 0.321865051984787, + "learning_rate": 1.9941705676998127e-06, + "loss": 0.6188, + "step": 4468 + }, + { + "epoch": 1.7114685046907907, + "grad_norm": 0.3180679678916931, + "learning_rate": 1.9931759711660786e-06, + "loss": 0.6382, + "step": 4469 + }, + { + "epoch": 1.7118514263833045, + "grad_norm": 0.30101141333580017, + "learning_rate": 1.9921814582813763e-06, + "loss": 0.6351, + "step": 4470 + }, + { + "epoch": 1.7122343480758184, + "grad_norm": 0.30683138966560364, + "learning_rate": 1.9911870292098443e-06, + "loss": 0.6269, + "step": 4471 + }, + { + "epoch": 1.7126172697683324, + "grad_norm": 0.31139999628067017, + "learning_rate": 1.990192684115611e-06, + "loss": 0.632, + "step": 4472 + }, + { + "epoch": 1.7130001914608464, + "grad_norm": 0.3126366138458252, + "learning_rate": 1.989198423162787e-06, + "loss": 0.6097, + "step": 4473 + }, + { + "epoch": 1.7133831131533601, + "grad_norm": 0.32761138677597046, + "learning_rate": 1.9882042465154698e-06, + "loss": 0.6233, + "step": 4474 + }, + { + "epoch": 1.7137660348458739, + "grad_norm": 0.3112486004829407, + "learning_rate": 1.987210154337745e-06, + "loss": 0.6268, + "step": 4475 + }, + { + "epoch": 1.7141489565383878, + "grad_norm": 0.3098362982273102, + "learning_rate": 1.9862161467936826e-06, + "loss": 0.6188, + "step": 4476 + }, + { + "epoch": 1.7145318782309018, + "grad_norm": 0.3085300922393799, + "learning_rate": 1.9852222240473396e-06, + "loss": 0.6084, + "step": 4477 + }, + { + "epoch": 1.7149147999234158, + "grad_norm": 0.31969887018203735, + "learning_rate": 1.9842283862627575e-06, + "loss": 0.632, + "step": 4478 + }, + { + "epoch": 1.7152977216159295, + "grad_norm": 0.3093487024307251, + "learning_rate": 1.9832346336039663e-06, + "loss": 0.617, + "step": 4479 + }, + { + "epoch": 1.7156806433084433, + "grad_norm": 0.3053451180458069, + "learning_rate": 1.9822409662349796e-06, + "loss": 0.6174, + "step": 4480 + }, + { + "epoch": 1.7160635650009572, + "grad_norm": 0.31284230947494507, + "learning_rate": 1.981247384319797e-06, + "loss": 0.6383, + "step": 4481 + }, + { + "epoch": 1.7164464866934712, + "grad_norm": 0.31671953201293945, + "learning_rate": 1.980253888022406e-06, + "loss": 0.6343, + "step": 4482 + }, + { + "epoch": 1.7168294083859852, + "grad_norm": 0.31300127506256104, + "learning_rate": 1.9792604775067776e-06, + "loss": 0.6219, + "step": 4483 + }, + { + "epoch": 1.717212330078499, + "grad_norm": 0.3066330850124359, + "learning_rate": 1.9782671529368706e-06, + "loss": 0.623, + "step": 4484 + }, + { + "epoch": 1.7175952517710127, + "grad_norm": 0.3153098225593567, + "learning_rate": 1.9772739144766277e-06, + "loss": 0.637, + "step": 4485 + }, + { + "epoch": 1.7179781734635267, + "grad_norm": 0.32286545634269714, + "learning_rate": 1.9762807622899804e-06, + "loss": 0.6485, + "step": 4486 + }, + { + "epoch": 1.7183610951560406, + "grad_norm": 0.3081876337528229, + "learning_rate": 1.9752876965408426e-06, + "loss": 0.6044, + "step": 4487 + }, + { + "epoch": 1.7187440168485546, + "grad_norm": 0.31541839241981506, + "learning_rate": 1.974294717393115e-06, + "loss": 0.6398, + "step": 4488 + }, + { + "epoch": 1.7191269385410684, + "grad_norm": 0.3054082691669464, + "learning_rate": 1.973301825010685e-06, + "loss": 0.6121, + "step": 4489 + }, + { + "epoch": 1.719509860233582, + "grad_norm": 0.3280740976333618, + "learning_rate": 1.9723090195574247e-06, + "loss": 0.641, + "step": 4490 + }, + { + "epoch": 1.719892781926096, + "grad_norm": 0.33849048614501953, + "learning_rate": 1.971316301197192e-06, + "loss": 0.6014, + "step": 4491 + }, + { + "epoch": 1.72027570361861, + "grad_norm": 0.31578579545021057, + "learning_rate": 1.970323670093832e-06, + "loss": 0.6203, + "step": 4492 + }, + { + "epoch": 1.720658625311124, + "grad_norm": 0.31407028436660767, + "learning_rate": 1.9693311264111722e-06, + "loss": 0.6203, + "step": 4493 + }, + { + "epoch": 1.7210415470036378, + "grad_norm": 0.3059990108013153, + "learning_rate": 1.968338670313029e-06, + "loss": 0.6099, + "step": 4494 + }, + { + "epoch": 1.7214244686961515, + "grad_norm": 0.3124779462814331, + "learning_rate": 1.9673463019632026e-06, + "loss": 0.6178, + "step": 4495 + }, + { + "epoch": 1.7218073903886655, + "grad_norm": 0.3244715929031372, + "learning_rate": 1.966354021525477e-06, + "loss": 0.6458, + "step": 4496 + }, + { + "epoch": 1.7221903120811795, + "grad_norm": 0.31624624133110046, + "learning_rate": 1.965361829163626e-06, + "loss": 0.6267, + "step": 4497 + }, + { + "epoch": 1.7225732337736934, + "grad_norm": 0.30919909477233887, + "learning_rate": 1.9643697250414055e-06, + "loss": 0.6332, + "step": 4498 + }, + { + "epoch": 1.7229561554662072, + "grad_norm": 0.3125494718551636, + "learning_rate": 1.963377709322558e-06, + "loss": 0.6374, + "step": 4499 + }, + { + "epoch": 1.723339077158721, + "grad_norm": 0.30337321758270264, + "learning_rate": 1.9623857821708105e-06, + "loss": 0.6312, + "step": 4500 + }, + { + "epoch": 1.723721998851235, + "grad_norm": 0.32289382815361023, + "learning_rate": 1.9613939437498776e-06, + "loss": 0.6572, + "step": 4501 + }, + { + "epoch": 1.7241049205437489, + "grad_norm": 0.31457316875457764, + "learning_rate": 1.9604021942234575e-06, + "loss": 0.6406, + "step": 4502 + }, + { + "epoch": 1.7244878422362628, + "grad_norm": 0.30481842160224915, + "learning_rate": 1.959410533755232e-06, + "loss": 0.6128, + "step": 4503 + }, + { + "epoch": 1.7248707639287766, + "grad_norm": 0.31026169657707214, + "learning_rate": 1.958418962508873e-06, + "loss": 0.6215, + "step": 4504 + }, + { + "epoch": 1.7252536856212903, + "grad_norm": 0.3148886263370514, + "learning_rate": 1.9574274806480326e-06, + "loss": 0.6131, + "step": 4505 + }, + { + "epoch": 1.7256366073138043, + "grad_norm": 0.30842316150665283, + "learning_rate": 1.9564360883363518e-06, + "loss": 0.613, + "step": 4506 + }, + { + "epoch": 1.7260195290063183, + "grad_norm": 0.30964067578315735, + "learning_rate": 1.955444785737455e-06, + "loss": 0.6396, + "step": 4507 + }, + { + "epoch": 1.7264024506988322, + "grad_norm": 0.3092633783817291, + "learning_rate": 1.9544535730149524e-06, + "loss": 0.6394, + "step": 4508 + }, + { + "epoch": 1.726785372391346, + "grad_norm": 0.31294724345207214, + "learning_rate": 1.9534624503324396e-06, + "loss": 0.6444, + "step": 4509 + }, + { + "epoch": 1.7271682940838597, + "grad_norm": 0.3154677152633667, + "learning_rate": 1.9524714178534953e-06, + "loss": 0.6284, + "step": 4510 + }, + { + "epoch": 1.7275512157763737, + "grad_norm": 0.3131828010082245, + "learning_rate": 1.9514804757416873e-06, + "loss": 0.6261, + "step": 4511 + }, + { + "epoch": 1.7279341374688877, + "grad_norm": 0.31505709886550903, + "learning_rate": 1.950489624160564e-06, + "loss": 0.6251, + "step": 4512 + }, + { + "epoch": 1.7283170591614017, + "grad_norm": 0.3162958025932312, + "learning_rate": 1.9494988632736623e-06, + "loss": 0.6312, + "step": 4513 + }, + { + "epoch": 1.7286999808539154, + "grad_norm": 0.31222274899482727, + "learning_rate": 1.9485081932445024e-06, + "loss": 0.6204, + "step": 4514 + }, + { + "epoch": 1.7290829025464292, + "grad_norm": 0.3095482289791107, + "learning_rate": 1.94751761423659e-06, + "loss": 0.6405, + "step": 4515 + }, + { + "epoch": 1.7294658242389431, + "grad_norm": 0.3148365318775177, + "learning_rate": 1.946527126413417e-06, + "loss": 0.629, + "step": 4516 + }, + { + "epoch": 1.729848745931457, + "grad_norm": 0.31101712584495544, + "learning_rate": 1.945536729938457e-06, + "loss": 0.6278, + "step": 4517 + }, + { + "epoch": 1.7302316676239708, + "grad_norm": 0.32112812995910645, + "learning_rate": 1.9445464249751704e-06, + "loss": 0.6382, + "step": 4518 + }, + { + "epoch": 1.7306145893164848, + "grad_norm": 0.31279852986335754, + "learning_rate": 1.9435562116870043e-06, + "loss": 0.6261, + "step": 4519 + }, + { + "epoch": 1.7309975110089986, + "grad_norm": 0.31799688935279846, + "learning_rate": 1.9425660902373876e-06, + "loss": 0.625, + "step": 4520 + }, + { + "epoch": 1.7313804327015125, + "grad_norm": 0.31216466426849365, + "learning_rate": 1.9415760607897364e-06, + "loss": 0.6157, + "step": 4521 + }, + { + "epoch": 1.7317633543940265, + "grad_norm": 0.31968051195144653, + "learning_rate": 1.9405861235074497e-06, + "loss": 0.6047, + "step": 4522 + }, + { + "epoch": 1.7321462760865403, + "grad_norm": 0.30351918935775757, + "learning_rate": 1.9395962785539137e-06, + "loss": 0.6357, + "step": 4523 + }, + { + "epoch": 1.7325291977790542, + "grad_norm": 0.3046968877315521, + "learning_rate": 1.938606526092496e-06, + "loss": 0.6216, + "step": 4524 + }, + { + "epoch": 1.732912119471568, + "grad_norm": 0.3088667094707489, + "learning_rate": 1.9376168662865515e-06, + "loss": 0.6319, + "step": 4525 + }, + { + "epoch": 1.733295041164082, + "grad_norm": 0.313653826713562, + "learning_rate": 1.9366272992994188e-06, + "loss": 0.6438, + "step": 4526 + }, + { + "epoch": 1.733677962856596, + "grad_norm": 0.3079594373703003, + "learning_rate": 1.9356378252944223e-06, + "loss": 0.648, + "step": 4527 + }, + { + "epoch": 1.7340608845491097, + "grad_norm": 0.3188839554786682, + "learning_rate": 1.93464844443487e-06, + "loss": 0.6372, + "step": 4528 + }, + { + "epoch": 1.7344438062416236, + "grad_norm": 0.3080839514732361, + "learning_rate": 1.933659156884054e-06, + "loss": 0.6084, + "step": 4529 + }, + { + "epoch": 1.7348267279341374, + "grad_norm": 0.3141629993915558, + "learning_rate": 1.9326699628052537e-06, + "loss": 0.639, + "step": 4530 + }, + { + "epoch": 1.7352096496266514, + "grad_norm": 0.3089374005794525, + "learning_rate": 1.931680862361729e-06, + "loss": 0.6, + "step": 4531 + }, + { + "epoch": 1.7355925713191653, + "grad_norm": 0.3107096552848816, + "learning_rate": 1.930691855716727e-06, + "loss": 0.6309, + "step": 4532 + }, + { + "epoch": 1.735975493011679, + "grad_norm": 0.32387420535087585, + "learning_rate": 1.9297029430334795e-06, + "loss": 0.6235, + "step": 4533 + }, + { + "epoch": 1.7363584147041928, + "grad_norm": 0.3147973120212555, + "learning_rate": 1.928714124475201e-06, + "loss": 0.6384, + "step": 4534 + }, + { + "epoch": 1.7367413363967068, + "grad_norm": 0.3061930239200592, + "learning_rate": 1.9277254002050934e-06, + "loss": 0.6141, + "step": 4535 + }, + { + "epoch": 1.7371242580892208, + "grad_norm": 0.31188297271728516, + "learning_rate": 1.926736770386339e-06, + "loss": 0.6174, + "step": 4536 + }, + { + "epoch": 1.7375071797817347, + "grad_norm": 0.3251532018184662, + "learning_rate": 1.925748235182109e-06, + "loss": 0.6426, + "step": 4537 + }, + { + "epoch": 1.7378901014742485, + "grad_norm": 0.31116345524787903, + "learning_rate": 1.924759794755555e-06, + "loss": 0.6123, + "step": 4538 + }, + { + "epoch": 1.7382730231667622, + "grad_norm": 0.3077068030834198, + "learning_rate": 1.9237714492698145e-06, + "loss": 0.6213, + "step": 4539 + }, + { + "epoch": 1.7386559448592762, + "grad_norm": 0.3042605519294739, + "learning_rate": 1.922783198888011e-06, + "loss": 0.6154, + "step": 4540 + }, + { + "epoch": 1.7390388665517902, + "grad_norm": 0.3136458992958069, + "learning_rate": 1.921795043773249e-06, + "loss": 0.6412, + "step": 4541 + }, + { + "epoch": 1.7394217882443042, + "grad_norm": 0.31796056032180786, + "learning_rate": 1.92080698408862e-06, + "loss": 0.6399, + "step": 4542 + }, + { + "epoch": 1.739804709936818, + "grad_norm": 0.3173500597476959, + "learning_rate": 1.9198190199972e-06, + "loss": 0.6162, + "step": 4543 + }, + { + "epoch": 1.7401876316293317, + "grad_norm": 0.3163227438926697, + "learning_rate": 1.9188311516620466e-06, + "loss": 0.6306, + "step": 4544 + }, + { + "epoch": 1.7405705533218456, + "grad_norm": 0.3224067986011505, + "learning_rate": 1.9178433792462024e-06, + "loss": 0.6349, + "step": 4545 + }, + { + "epoch": 1.7409534750143596, + "grad_norm": 0.3077217638492584, + "learning_rate": 1.9168557029126965e-06, + "loss": 0.6388, + "step": 4546 + }, + { + "epoch": 1.7413363967068736, + "grad_norm": 0.30990180373191833, + "learning_rate": 1.9158681228245386e-06, + "loss": 0.6356, + "step": 4547 + }, + { + "epoch": 1.7417193183993873, + "grad_norm": 0.31467992067337036, + "learning_rate": 1.9148806391447264e-06, + "loss": 0.6211, + "step": 4548 + }, + { + "epoch": 1.742102240091901, + "grad_norm": 0.3162992000579834, + "learning_rate": 1.913893252036238e-06, + "loss": 0.632, + "step": 4549 + }, + { + "epoch": 1.742485161784415, + "grad_norm": 0.3085786998271942, + "learning_rate": 1.9129059616620383e-06, + "loss": 0.6536, + "step": 4550 + }, + { + "epoch": 1.742868083476929, + "grad_norm": 0.3210308253765106, + "learning_rate": 1.911918768185075e-06, + "loss": 0.6156, + "step": 4551 + }, + { + "epoch": 1.743251005169443, + "grad_norm": 0.31203943490982056, + "learning_rate": 1.910931671768279e-06, + "loss": 0.6258, + "step": 4552 + }, + { + "epoch": 1.7436339268619567, + "grad_norm": 0.2993856370449066, + "learning_rate": 1.9099446725745674e-06, + "loss": 0.6296, + "step": 4553 + }, + { + "epoch": 1.7440168485544705, + "grad_norm": 0.3104995787143707, + "learning_rate": 1.908957770766839e-06, + "loss": 0.6233, + "step": 4554 + }, + { + "epoch": 1.7443997702469844, + "grad_norm": 0.3162441849708557, + "learning_rate": 1.907970966507978e-06, + "loss": 0.6449, + "step": 4555 + }, + { + "epoch": 1.7447826919394984, + "grad_norm": 0.3416912257671356, + "learning_rate": 1.9069842599608512e-06, + "loss": 0.6163, + "step": 4556 + }, + { + "epoch": 1.7451656136320124, + "grad_norm": 0.3166864514350891, + "learning_rate": 1.9059976512883117e-06, + "loss": 0.65, + "step": 4557 + }, + { + "epoch": 1.7455485353245261, + "grad_norm": 0.3170754909515381, + "learning_rate": 1.9050111406531943e-06, + "loss": 0.6133, + "step": 4558 + }, + { + "epoch": 1.74593145701704, + "grad_norm": 0.3131784200668335, + "learning_rate": 1.9040247282183166e-06, + "loss": 0.634, + "step": 4559 + }, + { + "epoch": 1.7463143787095539, + "grad_norm": 0.3132355511188507, + "learning_rate": 1.9030384141464836e-06, + "loss": 0.6233, + "step": 4560 + }, + { + "epoch": 1.7466973004020678, + "grad_norm": 0.3269815146923065, + "learning_rate": 1.9020521986004803e-06, + "loss": 0.5892, + "step": 4561 + }, + { + "epoch": 1.7470802220945818, + "grad_norm": 0.3122226893901825, + "learning_rate": 1.9010660817430787e-06, + "loss": 0.6111, + "step": 4562 + }, + { + "epoch": 1.7474631437870956, + "grad_norm": 0.30711492896080017, + "learning_rate": 1.9000800637370316e-06, + "loss": 0.6093, + "step": 4563 + }, + { + "epoch": 1.7478460654796093, + "grad_norm": 0.31391143798828125, + "learning_rate": 1.8990941447450778e-06, + "loss": 0.6417, + "step": 4564 + }, + { + "epoch": 1.7482289871721233, + "grad_norm": 0.30937373638153076, + "learning_rate": 1.8981083249299393e-06, + "loss": 0.6337, + "step": 4565 + }, + { + "epoch": 1.7486119088646372, + "grad_norm": 0.31283625960350037, + "learning_rate": 1.8971226044543195e-06, + "loss": 0.6365, + "step": 4566 + }, + { + "epoch": 1.7489948305571512, + "grad_norm": 0.32945746183395386, + "learning_rate": 1.8961369834809084e-06, + "loss": 0.5974, + "step": 4567 + }, + { + "epoch": 1.749377752249665, + "grad_norm": 0.30858030915260315, + "learning_rate": 1.8951514621723773e-06, + "loss": 0.6204, + "step": 4568 + }, + { + "epoch": 1.7497606739421787, + "grad_norm": 0.3131084740161896, + "learning_rate": 1.894166040691383e-06, + "loss": 0.6373, + "step": 4569 + }, + { + "epoch": 1.7501435956346927, + "grad_norm": 0.32645314931869507, + "learning_rate": 1.8931807192005653e-06, + "loss": 0.6432, + "step": 4570 + }, + { + "epoch": 1.7505265173272067, + "grad_norm": 0.3030444085597992, + "learning_rate": 1.8921954978625457e-06, + "loss": 0.5938, + "step": 4571 + }, + { + "epoch": 1.7509094390197206, + "grad_norm": 0.3105182945728302, + "learning_rate": 1.8912103768399322e-06, + "loss": 0.6354, + "step": 4572 + }, + { + "epoch": 1.7512923607122344, + "grad_norm": 0.3232519030570984, + "learning_rate": 1.8902253562953136e-06, + "loss": 0.6106, + "step": 4573 + }, + { + "epoch": 1.7516752824047481, + "grad_norm": 0.3053780794143677, + "learning_rate": 1.8892404363912625e-06, + "loss": 0.6371, + "step": 4574 + }, + { + "epoch": 1.752058204097262, + "grad_norm": 0.30347755551338196, + "learning_rate": 1.8882556172903367e-06, + "loss": 0.6424, + "step": 4575 + }, + { + "epoch": 1.752441125789776, + "grad_norm": 0.3090442717075348, + "learning_rate": 1.8872708991550753e-06, + "loss": 0.6483, + "step": 4576 + }, + { + "epoch": 1.75282404748229, + "grad_norm": 0.3154149055480957, + "learning_rate": 1.8862862821480023e-06, + "loss": 0.6406, + "step": 4577 + }, + { + "epoch": 1.7532069691748038, + "grad_norm": 0.3122040927410126, + "learning_rate": 1.8853017664316237e-06, + "loss": 0.6273, + "step": 4578 + }, + { + "epoch": 1.7535898908673175, + "grad_norm": 0.32033830881118774, + "learning_rate": 1.8843173521684301e-06, + "loss": 0.6191, + "step": 4579 + }, + { + "epoch": 1.7539728125598315, + "grad_norm": 0.31428325176239014, + "learning_rate": 1.8833330395208951e-06, + "loss": 0.6287, + "step": 4580 + }, + { + "epoch": 1.7543557342523455, + "grad_norm": 0.3213519752025604, + "learning_rate": 1.882348828651473e-06, + "loss": 0.6396, + "step": 4581 + }, + { + "epoch": 1.7547386559448592, + "grad_norm": 0.3059098422527313, + "learning_rate": 1.8813647197226054e-06, + "loss": 0.6105, + "step": 4582 + }, + { + "epoch": 1.7551215776373732, + "grad_norm": 0.30547255277633667, + "learning_rate": 1.8803807128967135e-06, + "loss": 0.6133, + "step": 4583 + }, + { + "epoch": 1.755504499329887, + "grad_norm": 0.30515623092651367, + "learning_rate": 1.879396808336205e-06, + "loss": 0.6034, + "step": 4584 + }, + { + "epoch": 1.755887421022401, + "grad_norm": 0.31945493817329407, + "learning_rate": 1.8784130062034669e-06, + "loss": 0.608, + "step": 4585 + }, + { + "epoch": 1.756270342714915, + "grad_norm": 0.3135055899620056, + "learning_rate": 1.8774293066608733e-06, + "loss": 0.6417, + "step": 4586 + }, + { + "epoch": 1.7566532644074286, + "grad_norm": 0.3201060891151428, + "learning_rate": 1.876445709870779e-06, + "loss": 0.6209, + "step": 4587 + }, + { + "epoch": 1.7570361860999426, + "grad_norm": 0.30848217010498047, + "learning_rate": 1.8754622159955205e-06, + "loss": 0.6311, + "step": 4588 + }, + { + "epoch": 1.7574191077924564, + "grad_norm": 0.31547772884368896, + "learning_rate": 1.8744788251974209e-06, + "loss": 0.6332, + "step": 4589 + }, + { + "epoch": 1.7578020294849703, + "grad_norm": 0.31692028045654297, + "learning_rate": 1.873495537638783e-06, + "loss": 0.6195, + "step": 4590 + }, + { + "epoch": 1.7581849511774843, + "grad_norm": 0.3061498999595642, + "learning_rate": 1.8725123534818956e-06, + "loss": 0.6156, + "step": 4591 + }, + { + "epoch": 1.758567872869998, + "grad_norm": 0.3049583435058594, + "learning_rate": 1.8715292728890277e-06, + "loss": 0.6227, + "step": 4592 + }, + { + "epoch": 1.758950794562512, + "grad_norm": 0.3138638734817505, + "learning_rate": 1.870546296022433e-06, + "loss": 0.5749, + "step": 4593 + }, + { + "epoch": 1.7593337162550258, + "grad_norm": 0.32688939571380615, + "learning_rate": 1.8695634230443477e-06, + "loss": 0.5947, + "step": 4594 + }, + { + "epoch": 1.7597166379475397, + "grad_norm": 0.30960240960121155, + "learning_rate": 1.8685806541169887e-06, + "loss": 0.6111, + "step": 4595 + }, + { + "epoch": 1.7600995596400537, + "grad_norm": 0.3151426315307617, + "learning_rate": 1.8675979894025591e-06, + "loss": 0.6238, + "step": 4596 + }, + { + "epoch": 1.7604824813325675, + "grad_norm": 0.30396804213523865, + "learning_rate": 1.8666154290632435e-06, + "loss": 0.6354, + "step": 4597 + }, + { + "epoch": 1.7608654030250812, + "grad_norm": 0.3128806948661804, + "learning_rate": 1.8656329732612083e-06, + "loss": 0.6235, + "step": 4598 + }, + { + "epoch": 1.7612483247175952, + "grad_norm": 0.3088151216506958, + "learning_rate": 1.864650622158604e-06, + "loss": 0.6298, + "step": 4599 + }, + { + "epoch": 1.7616312464101092, + "grad_norm": 0.3104528486728668, + "learning_rate": 1.8636683759175628e-06, + "loss": 0.6298, + "step": 4600 + }, + { + "epoch": 1.7620141681026231, + "grad_norm": 0.31444376707077026, + "learning_rate": 1.862686234700201e-06, + "loss": 0.6434, + "step": 4601 + }, + { + "epoch": 1.7623970897951369, + "grad_norm": 0.3105086088180542, + "learning_rate": 1.8617041986686155e-06, + "loss": 0.6154, + "step": 4602 + }, + { + "epoch": 1.7627800114876506, + "grad_norm": 0.3143225312232971, + "learning_rate": 1.860722267984887e-06, + "loss": 0.6275, + "step": 4603 + }, + { + "epoch": 1.7631629331801646, + "grad_norm": 0.3214721977710724, + "learning_rate": 1.8597404428110789e-06, + "loss": 0.6526, + "step": 4604 + }, + { + "epoch": 1.7635458548726786, + "grad_norm": 0.3079444169998169, + "learning_rate": 1.8587587233092369e-06, + "loss": 0.6161, + "step": 4605 + }, + { + "epoch": 1.7639287765651925, + "grad_norm": 0.30815768241882324, + "learning_rate": 1.8577771096413904e-06, + "loss": 0.6522, + "step": 4606 + }, + { + "epoch": 1.7643116982577063, + "grad_norm": 0.3067401945590973, + "learning_rate": 1.8567956019695493e-06, + "loss": 0.6238, + "step": 4607 + }, + { + "epoch": 1.76469461995022, + "grad_norm": 0.3075568974018097, + "learning_rate": 1.8558142004557079e-06, + "loss": 0.6333, + "step": 4608 + }, + { + "epoch": 1.765077541642734, + "grad_norm": 0.30141159892082214, + "learning_rate": 1.8548329052618414e-06, + "loss": 0.6151, + "step": 4609 + }, + { + "epoch": 1.765460463335248, + "grad_norm": 0.324932336807251, + "learning_rate": 1.8538517165499077e-06, + "loss": 0.6339, + "step": 4610 + }, + { + "epoch": 1.765843385027762, + "grad_norm": 0.317778617143631, + "learning_rate": 1.8528706344818487e-06, + "loss": 0.6302, + "step": 4611 + }, + { + "epoch": 1.7662263067202757, + "grad_norm": 0.3201824426651001, + "learning_rate": 1.8518896592195862e-06, + "loss": 0.6059, + "step": 4612 + }, + { + "epoch": 1.7666092284127894, + "grad_norm": 0.31012022495269775, + "learning_rate": 1.8509087909250278e-06, + "loss": 0.6004, + "step": 4613 + }, + { + "epoch": 1.7669921501053034, + "grad_norm": 0.3073303699493408, + "learning_rate": 1.8499280297600594e-06, + "loss": 0.6064, + "step": 4614 + }, + { + "epoch": 1.7673750717978174, + "grad_norm": 0.3317025899887085, + "learning_rate": 1.8489473758865534e-06, + "loss": 0.626, + "step": 4615 + }, + { + "epoch": 1.7677579934903314, + "grad_norm": 0.3154352307319641, + "learning_rate": 1.8479668294663604e-06, + "loss": 0.6208, + "step": 4616 + }, + { + "epoch": 1.768140915182845, + "grad_norm": 0.319514662027359, + "learning_rate": 1.8469863906613153e-06, + "loss": 0.6222, + "step": 4617 + }, + { + "epoch": 1.7685238368753589, + "grad_norm": 0.30301085114479065, + "learning_rate": 1.846006059633236e-06, + "loss": 0.6272, + "step": 4618 + }, + { + "epoch": 1.7689067585678728, + "grad_norm": 0.3042938709259033, + "learning_rate": 1.8450258365439212e-06, + "loss": 0.6405, + "step": 4619 + }, + { + "epoch": 1.7692896802603868, + "grad_norm": 0.3157995045185089, + "learning_rate": 1.8440457215551528e-06, + "loss": 0.6397, + "step": 4620 + }, + { + "epoch": 1.7696726019529008, + "grad_norm": 0.3252478539943695, + "learning_rate": 1.8430657148286935e-06, + "loss": 0.6115, + "step": 4621 + }, + { + "epoch": 1.7700555236454145, + "grad_norm": 0.3055269718170166, + "learning_rate": 1.8420858165262905e-06, + "loss": 0.6102, + "step": 4622 + }, + { + "epoch": 1.7704384453379283, + "grad_norm": 0.3062238097190857, + "learning_rate": 1.8411060268096708e-06, + "loss": 0.6194, + "step": 4623 + }, + { + "epoch": 1.7708213670304422, + "grad_norm": 0.3059941530227661, + "learning_rate": 1.8401263458405433e-06, + "loss": 0.6269, + "step": 4624 + }, + { + "epoch": 1.7712042887229562, + "grad_norm": 0.32513895630836487, + "learning_rate": 1.8391467737806006e-06, + "loss": 0.6435, + "step": 4625 + }, + { + "epoch": 1.7715872104154702, + "grad_norm": 0.31075945496559143, + "learning_rate": 1.8381673107915176e-06, + "loss": 0.6415, + "step": 4626 + }, + { + "epoch": 1.771970132107984, + "grad_norm": 0.3136698603630066, + "learning_rate": 1.8371879570349494e-06, + "loss": 0.6221, + "step": 4627 + }, + { + "epoch": 1.7723530538004977, + "grad_norm": 0.31427696347236633, + "learning_rate": 1.8362087126725346e-06, + "loss": 0.6278, + "step": 4628 + }, + { + "epoch": 1.7727359754930117, + "grad_norm": 0.30843719840049744, + "learning_rate": 1.835229577865893e-06, + "loss": 0.6185, + "step": 4629 + }, + { + "epoch": 1.7731188971855256, + "grad_norm": 0.3056352734565735, + "learning_rate": 1.8342505527766248e-06, + "loss": 0.6384, + "step": 4630 + }, + { + "epoch": 1.7735018188780396, + "grad_norm": 0.3102870285511017, + "learning_rate": 1.8332716375663157e-06, + "loss": 0.6138, + "step": 4631 + }, + { + "epoch": 1.7738847405705533, + "grad_norm": 0.30550795793533325, + "learning_rate": 1.8322928323965296e-06, + "loss": 0.6427, + "step": 4632 + }, + { + "epoch": 1.774267662263067, + "grad_norm": 0.31388863921165466, + "learning_rate": 1.8313141374288156e-06, + "loss": 0.6226, + "step": 4633 + }, + { + "epoch": 1.774650583955581, + "grad_norm": 0.3262490928173065, + "learning_rate": 1.8303355528247019e-06, + "loss": 0.6399, + "step": 4634 + }, + { + "epoch": 1.775033505648095, + "grad_norm": 0.3205210864543915, + "learning_rate": 1.8293570787456996e-06, + "loss": 0.6222, + "step": 4635 + }, + { + "epoch": 1.775416427340609, + "grad_norm": 0.3001057505607605, + "learning_rate": 1.8283787153533022e-06, + "loss": 0.6282, + "step": 4636 + }, + { + "epoch": 1.7757993490331228, + "grad_norm": 0.32256650924682617, + "learning_rate": 1.8274004628089828e-06, + "loss": 0.6382, + "step": 4637 + }, + { + "epoch": 1.7761822707256365, + "grad_norm": 0.31250858306884766, + "learning_rate": 1.826422321274199e-06, + "loss": 0.6427, + "step": 4638 + }, + { + "epoch": 1.7765651924181505, + "grad_norm": 0.31545165181159973, + "learning_rate": 1.825444290910387e-06, + "loss": 0.6338, + "step": 4639 + }, + { + "epoch": 1.7769481141106644, + "grad_norm": 0.31707993149757385, + "learning_rate": 1.824466371878968e-06, + "loss": 0.6381, + "step": 4640 + }, + { + "epoch": 1.7773310358031784, + "grad_norm": 0.3038886487483978, + "learning_rate": 1.823488564341342e-06, + "loss": 0.6339, + "step": 4641 + }, + { + "epoch": 1.7777139574956922, + "grad_norm": 0.3052169680595398, + "learning_rate": 1.8225108684588928e-06, + "loss": 0.6206, + "step": 4642 + }, + { + "epoch": 1.778096879188206, + "grad_norm": 0.3171171545982361, + "learning_rate": 1.8215332843929844e-06, + "loss": 0.6242, + "step": 4643 + }, + { + "epoch": 1.7784798008807199, + "grad_norm": 0.30828073620796204, + "learning_rate": 1.8205558123049615e-06, + "loss": 0.6221, + "step": 4644 + }, + { + "epoch": 1.7788627225732339, + "grad_norm": 0.3012714385986328, + "learning_rate": 1.819578452356153e-06, + "loss": 0.6193, + "step": 4645 + }, + { + "epoch": 1.7792456442657476, + "grad_norm": 0.3061826527118683, + "learning_rate": 1.8186012047078667e-06, + "loss": 0.6454, + "step": 4646 + }, + { + "epoch": 1.7796285659582616, + "grad_norm": 0.30667340755462646, + "learning_rate": 1.8176240695213936e-06, + "loss": 0.6065, + "step": 4647 + }, + { + "epoch": 1.7800114876507753, + "grad_norm": 0.3174270689487457, + "learning_rate": 1.816647046958005e-06, + "loss": 0.6324, + "step": 4648 + }, + { + "epoch": 1.7803944093432893, + "grad_norm": 0.310070276260376, + "learning_rate": 1.8156701371789542e-06, + "loss": 0.6435, + "step": 4649 + }, + { + "epoch": 1.7807773310358033, + "grad_norm": 0.3176279664039612, + "learning_rate": 1.8146933403454776e-06, + "loss": 0.6143, + "step": 4650 + }, + { + "epoch": 1.781160252728317, + "grad_norm": 0.30817142128944397, + "learning_rate": 1.813716656618788e-06, + "loss": 0.6191, + "step": 4651 + }, + { + "epoch": 1.781543174420831, + "grad_norm": 0.30590665340423584, + "learning_rate": 1.8127400861600839e-06, + "loss": 0.6491, + "step": 4652 + }, + { + "epoch": 1.7819260961133447, + "grad_norm": 0.3099651634693146, + "learning_rate": 1.8117636291305446e-06, + "loss": 0.6105, + "step": 4653 + }, + { + "epoch": 1.7823090178058587, + "grad_norm": 0.319184809923172, + "learning_rate": 1.8107872856913293e-06, + "loss": 0.6201, + "step": 4654 + }, + { + "epoch": 1.7826919394983727, + "grad_norm": 0.3098507821559906, + "learning_rate": 1.8098110560035792e-06, + "loss": 0.6012, + "step": 4655 + }, + { + "epoch": 1.7830748611908864, + "grad_norm": 0.31123754382133484, + "learning_rate": 1.8088349402284167e-06, + "loss": 0.6291, + "step": 4656 + }, + { + "epoch": 1.7834577828834004, + "grad_norm": 0.31197497248649597, + "learning_rate": 1.807858938526946e-06, + "loss": 0.6346, + "step": 4657 + }, + { + "epoch": 1.7838407045759141, + "grad_norm": 0.3135431408882141, + "learning_rate": 1.8068830510602508e-06, + "loss": 0.6136, + "step": 4658 + }, + { + "epoch": 1.7842236262684281, + "grad_norm": 0.315276563167572, + "learning_rate": 1.8059072779893965e-06, + "loss": 0.6369, + "step": 4659 + }, + { + "epoch": 1.784606547960942, + "grad_norm": 0.30775246024131775, + "learning_rate": 1.8049316194754319e-06, + "loss": 0.6458, + "step": 4660 + }, + { + "epoch": 1.7849894696534558, + "grad_norm": 0.31935280561447144, + "learning_rate": 1.8039560756793827e-06, + "loss": 0.6407, + "step": 4661 + }, + { + "epoch": 1.7853723913459696, + "grad_norm": 0.3279930651187897, + "learning_rate": 1.8029806467622602e-06, + "loss": 0.5971, + "step": 4662 + }, + { + "epoch": 1.7857553130384836, + "grad_norm": 0.3154400587081909, + "learning_rate": 1.8020053328850532e-06, + "loss": 0.621, + "step": 4663 + }, + { + "epoch": 1.7861382347309975, + "grad_norm": 0.30977264046669006, + "learning_rate": 1.801030134208734e-06, + "loss": 0.6146, + "step": 4664 + }, + { + "epoch": 1.7865211564235115, + "grad_norm": 0.31872597336769104, + "learning_rate": 1.8000550508942543e-06, + "loss": 0.6297, + "step": 4665 + }, + { + "epoch": 1.7869040781160253, + "grad_norm": 0.3409903645515442, + "learning_rate": 1.7990800831025462e-06, + "loss": 0.6295, + "step": 4666 + }, + { + "epoch": 1.787286999808539, + "grad_norm": 0.3065347969532013, + "learning_rate": 1.7981052309945252e-06, + "loss": 0.6406, + "step": 4667 + }, + { + "epoch": 1.787669921501053, + "grad_norm": 0.3076355755329132, + "learning_rate": 1.7971304947310847e-06, + "loss": 0.6385, + "step": 4668 + }, + { + "epoch": 1.788052843193567, + "grad_norm": 0.3237532079219818, + "learning_rate": 1.7961558744731023e-06, + "loss": 0.6311, + "step": 4669 + }, + { + "epoch": 1.788435764886081, + "grad_norm": 0.30718499422073364, + "learning_rate": 1.7951813703814335e-06, + "loss": 0.6192, + "step": 4670 + }, + { + "epoch": 1.7888186865785947, + "grad_norm": 0.30459973216056824, + "learning_rate": 1.7942069826169168e-06, + "loss": 0.6347, + "step": 4671 + }, + { + "epoch": 1.7892016082711084, + "grad_norm": 0.3048368990421295, + "learning_rate": 1.7932327113403702e-06, + "loss": 0.6191, + "step": 4672 + }, + { + "epoch": 1.7895845299636224, + "grad_norm": 0.3117280900478363, + "learning_rate": 1.7922585567125917e-06, + "loss": 0.6385, + "step": 4673 + }, + { + "epoch": 1.7899674516561364, + "grad_norm": 0.3069230020046234, + "learning_rate": 1.7912845188943625e-06, + "loss": 0.6116, + "step": 4674 + }, + { + "epoch": 1.7903503733486503, + "grad_norm": 0.30421894788742065, + "learning_rate": 1.7903105980464421e-06, + "loss": 0.6138, + "step": 4675 + }, + { + "epoch": 1.790733295041164, + "grad_norm": 0.32558903098106384, + "learning_rate": 1.7893367943295725e-06, + "loss": 0.6317, + "step": 4676 + }, + { + "epoch": 1.7911162167336778, + "grad_norm": 0.31072917580604553, + "learning_rate": 1.7883631079044762e-06, + "loss": 0.6297, + "step": 4677 + }, + { + "epoch": 1.7914991384261918, + "grad_norm": 0.3178878426551819, + "learning_rate": 1.7873895389318542e-06, + "loss": 0.6299, + "step": 4678 + }, + { + "epoch": 1.7918820601187058, + "grad_norm": 0.3085070848464966, + "learning_rate": 1.7864160875723916e-06, + "loss": 0.6315, + "step": 4679 + }, + { + "epoch": 1.7922649818112197, + "grad_norm": 0.30670878291130066, + "learning_rate": 1.7854427539867505e-06, + "loss": 0.6419, + "step": 4680 + }, + { + "epoch": 1.7926479035037335, + "grad_norm": 0.3098073899745941, + "learning_rate": 1.7844695383355757e-06, + "loss": 0.6169, + "step": 4681 + }, + { + "epoch": 1.7930308251962472, + "grad_norm": 0.3142548203468323, + "learning_rate": 1.7834964407794924e-06, + "loss": 0.6241, + "step": 4682 + }, + { + "epoch": 1.7934137468887612, + "grad_norm": 0.3162960410118103, + "learning_rate": 1.7825234614791053e-06, + "loss": 0.6121, + "step": 4683 + }, + { + "epoch": 1.7937966685812752, + "grad_norm": 0.31699326634407043, + "learning_rate": 1.7815506005950016e-06, + "loss": 0.6375, + "step": 4684 + }, + { + "epoch": 1.7941795902737891, + "grad_norm": 0.3036661148071289, + "learning_rate": 1.7805778582877465e-06, + "loss": 0.612, + "step": 4685 + }, + { + "epoch": 1.794562511966303, + "grad_norm": 0.31154710054397583, + "learning_rate": 1.7796052347178877e-06, + "loss": 0.5879, + "step": 4686 + }, + { + "epoch": 1.7949454336588166, + "grad_norm": 0.30076220631599426, + "learning_rate": 1.778632730045951e-06, + "loss": 0.6297, + "step": 4687 + }, + { + "epoch": 1.7953283553513306, + "grad_norm": 0.3055517375469208, + "learning_rate": 1.7776603444324445e-06, + "loss": 0.6161, + "step": 4688 + }, + { + "epoch": 1.7957112770438446, + "grad_norm": 0.3152478039264679, + "learning_rate": 1.7766880780378565e-06, + "loss": 0.635, + "step": 4689 + }, + { + "epoch": 1.7960941987363586, + "grad_norm": 0.31451210379600525, + "learning_rate": 1.7757159310226546e-06, + "loss": 0.6454, + "step": 4690 + }, + { + "epoch": 1.7964771204288723, + "grad_norm": 0.315701425075531, + "learning_rate": 1.774743903547288e-06, + "loss": 0.6286, + "step": 4691 + }, + { + "epoch": 1.796860042121386, + "grad_norm": 0.30651500821113586, + "learning_rate": 1.773771995772185e-06, + "loss": 0.638, + "step": 4692 + }, + { + "epoch": 1.7972429638139, + "grad_norm": 0.31515610218048096, + "learning_rate": 1.7728002078577555e-06, + "loss": 0.6159, + "step": 4693 + }, + { + "epoch": 1.797625885506414, + "grad_norm": 0.32298529148101807, + "learning_rate": 1.7718285399643872e-06, + "loss": 0.6399, + "step": 4694 + }, + { + "epoch": 1.798008807198928, + "grad_norm": 0.3087218701839447, + "learning_rate": 1.7708569922524498e-06, + "loss": 0.6338, + "step": 4695 + }, + { + "epoch": 1.7983917288914417, + "grad_norm": 0.3153378367424011, + "learning_rate": 1.7698855648822944e-06, + "loss": 0.6207, + "step": 4696 + }, + { + "epoch": 1.7987746505839555, + "grad_norm": 0.30247360467910767, + "learning_rate": 1.7689142580142488e-06, + "loss": 0.6268, + "step": 4697 + }, + { + "epoch": 1.7991575722764694, + "grad_norm": 0.3253176510334015, + "learning_rate": 1.7679430718086244e-06, + "loss": 0.621, + "step": 4698 + }, + { + "epoch": 1.7995404939689834, + "grad_norm": 0.31461814045906067, + "learning_rate": 1.7669720064257097e-06, + "loss": 0.6317, + "step": 4699 + }, + { + "epoch": 1.7999234156614974, + "grad_norm": 0.31463271379470825, + "learning_rate": 1.7660010620257766e-06, + "loss": 0.6327, + "step": 4700 + }, + { + "epoch": 1.8003063373540111, + "grad_norm": 0.3101353347301483, + "learning_rate": 1.7650302387690738e-06, + "loss": 0.6155, + "step": 4701 + }, + { + "epoch": 1.8006892590465249, + "grad_norm": 0.29936864972114563, + "learning_rate": 1.7640595368158305e-06, + "loss": 0.629, + "step": 4702 + }, + { + "epoch": 1.8010721807390389, + "grad_norm": 0.31403782963752747, + "learning_rate": 1.7630889563262579e-06, + "loss": 0.6227, + "step": 4703 + }, + { + "epoch": 1.8014551024315528, + "grad_norm": 0.30827516317367554, + "learning_rate": 1.7621184974605463e-06, + "loss": 0.6397, + "step": 4704 + }, + { + "epoch": 1.8018380241240666, + "grad_norm": 0.3079252243041992, + "learning_rate": 1.7611481603788645e-06, + "loss": 0.6264, + "step": 4705 + }, + { + "epoch": 1.8022209458165805, + "grad_norm": 0.3050033152103424, + "learning_rate": 1.7601779452413631e-06, + "loss": 0.642, + "step": 4706 + }, + { + "epoch": 1.8026038675090943, + "grad_norm": 0.31098079681396484, + "learning_rate": 1.7592078522081725e-06, + "loss": 0.6213, + "step": 4707 + }, + { + "epoch": 1.8029867892016083, + "grad_norm": 0.3117855191230774, + "learning_rate": 1.7582378814393994e-06, + "loss": 0.6458, + "step": 4708 + }, + { + "epoch": 1.8033697108941222, + "grad_norm": 0.3195698857307434, + "learning_rate": 1.7572680330951359e-06, + "loss": 0.6158, + "step": 4709 + }, + { + "epoch": 1.803752632586636, + "grad_norm": 0.30917593836784363, + "learning_rate": 1.756298307335449e-06, + "loss": 0.6342, + "step": 4710 + }, + { + "epoch": 1.80413555427915, + "grad_norm": 0.3022180199623108, + "learning_rate": 1.7553287043203899e-06, + "loss": 0.6268, + "step": 4711 + }, + { + "epoch": 1.8045184759716637, + "grad_norm": 0.30801355838775635, + "learning_rate": 1.754359224209985e-06, + "loss": 0.6173, + "step": 4712 + }, + { + "epoch": 1.8049013976641777, + "grad_norm": 0.3128814697265625, + "learning_rate": 1.7533898671642446e-06, + "loss": 0.6469, + "step": 4713 + }, + { + "epoch": 1.8052843193566916, + "grad_norm": 0.3275573253631592, + "learning_rate": 1.7524206333431558e-06, + "loss": 0.6214, + "step": 4714 + }, + { + "epoch": 1.8056672410492054, + "grad_norm": 0.3207276463508606, + "learning_rate": 1.7514515229066859e-06, + "loss": 0.6201, + "step": 4715 + }, + { + "epoch": 1.8060501627417194, + "grad_norm": 0.31348294019699097, + "learning_rate": 1.7504825360147826e-06, + "loss": 0.6136, + "step": 4716 + }, + { + "epoch": 1.8064330844342331, + "grad_norm": 0.3114925026893616, + "learning_rate": 1.7495136728273722e-06, + "loss": 0.6215, + "step": 4717 + }, + { + "epoch": 1.806816006126747, + "grad_norm": 0.3073063790798187, + "learning_rate": 1.7485449335043627e-06, + "loss": 0.6362, + "step": 4718 + }, + { + "epoch": 1.807198927819261, + "grad_norm": 0.3013920783996582, + "learning_rate": 1.7475763182056387e-06, + "loss": 0.6302, + "step": 4719 + }, + { + "epoch": 1.8075818495117748, + "grad_norm": 0.30515918135643005, + "learning_rate": 1.746607827091067e-06, + "loss": 0.64, + "step": 4720 + }, + { + "epoch": 1.8079647712042886, + "grad_norm": 0.3112282156944275, + "learning_rate": 1.7456394603204926e-06, + "loss": 0.6241, + "step": 4721 + }, + { + "epoch": 1.8083476928968025, + "grad_norm": 0.3089345097541809, + "learning_rate": 1.7446712180537387e-06, + "loss": 0.6276, + "step": 4722 + }, + { + "epoch": 1.8087306145893165, + "grad_norm": 0.3245830535888672, + "learning_rate": 1.7437031004506109e-06, + "loss": 0.6402, + "step": 4723 + }, + { + "epoch": 1.8091135362818305, + "grad_norm": 0.32121482491493225, + "learning_rate": 1.7427351076708912e-06, + "loss": 0.616, + "step": 4724 + }, + { + "epoch": 1.8094964579743442, + "grad_norm": 0.303287535905838, + "learning_rate": 1.741767239874344e-06, + "loss": 0.6104, + "step": 4725 + }, + { + "epoch": 1.809879379666858, + "grad_norm": 0.3017740249633789, + "learning_rate": 1.7407994972207104e-06, + "loss": 0.6022, + "step": 4726 + }, + { + "epoch": 1.810262301359372, + "grad_norm": 0.3129803240299225, + "learning_rate": 1.7398318798697129e-06, + "loss": 0.6223, + "step": 4727 + }, + { + "epoch": 1.810645223051886, + "grad_norm": 0.293645977973938, + "learning_rate": 1.7388643879810525e-06, + "loss": 0.6183, + "step": 4728 + }, + { + "epoch": 1.8110281447443999, + "grad_norm": 0.3059869408607483, + "learning_rate": 1.7378970217144075e-06, + "loss": 0.6217, + "step": 4729 + }, + { + "epoch": 1.8114110664369136, + "grad_norm": 0.31755349040031433, + "learning_rate": 1.736929781229439e-06, + "loss": 0.6371, + "step": 4730 + }, + { + "epoch": 1.8117939881294274, + "grad_norm": 0.3127236068248749, + "learning_rate": 1.735962666685786e-06, + "loss": 0.6199, + "step": 4731 + }, + { + "epoch": 1.8121769098219414, + "grad_norm": 0.3164852261543274, + "learning_rate": 1.734995678243065e-06, + "loss": 0.63, + "step": 4732 + }, + { + "epoch": 1.8125598315144553, + "grad_norm": 0.305999755859375, + "learning_rate": 1.7340288160608745e-06, + "loss": 0.6161, + "step": 4733 + }, + { + "epoch": 1.8129427532069693, + "grad_norm": 0.31158092617988586, + "learning_rate": 1.7330620802987897e-06, + "loss": 0.6387, + "step": 4734 + }, + { + "epoch": 1.813325674899483, + "grad_norm": 0.309907466173172, + "learning_rate": 1.7320954711163674e-06, + "loss": 0.6105, + "step": 4735 + }, + { + "epoch": 1.8137085965919968, + "grad_norm": 0.3084656298160553, + "learning_rate": 1.7311289886731408e-06, + "loss": 0.6025, + "step": 4736 + }, + { + "epoch": 1.8140915182845108, + "grad_norm": 0.3069480359554291, + "learning_rate": 1.7301626331286233e-06, + "loss": 0.6351, + "step": 4737 + }, + { + "epoch": 1.8144744399770247, + "grad_norm": 0.316021203994751, + "learning_rate": 1.7291964046423087e-06, + "loss": 0.6382, + "step": 4738 + }, + { + "epoch": 1.8148573616695387, + "grad_norm": 0.3153451979160309, + "learning_rate": 1.7282303033736675e-06, + "loss": 0.6306, + "step": 4739 + }, + { + "epoch": 1.8152402833620525, + "grad_norm": 0.3227040469646454, + "learning_rate": 1.7272643294821511e-06, + "loss": 0.6265, + "step": 4740 + }, + { + "epoch": 1.8156232050545662, + "grad_norm": 0.3143640458583832, + "learning_rate": 1.726298483127189e-06, + "loss": 0.6265, + "step": 4741 + }, + { + "epoch": 1.8160061267470802, + "grad_norm": 0.321386456489563, + "learning_rate": 1.7253327644681908e-06, + "loss": 0.6157, + "step": 4742 + }, + { + "epoch": 1.8163890484395941, + "grad_norm": 0.33368605375289917, + "learning_rate": 1.7243671736645422e-06, + "loss": 0.6402, + "step": 4743 + }, + { + "epoch": 1.8167719701321081, + "grad_norm": 0.32025715708732605, + "learning_rate": 1.7234017108756105e-06, + "loss": 0.6347, + "step": 4744 + }, + { + "epoch": 1.8171548918246219, + "grad_norm": 0.32035475969314575, + "learning_rate": 1.722436376260741e-06, + "loss": 0.6391, + "step": 4745 + }, + { + "epoch": 1.8175378135171356, + "grad_norm": 0.3064672648906708, + "learning_rate": 1.7214711699792576e-06, + "loss": 0.6281, + "step": 4746 + }, + { + "epoch": 1.8179207352096496, + "grad_norm": 0.3056202232837677, + "learning_rate": 1.720506092190464e-06, + "loss": 0.633, + "step": 4747 + }, + { + "epoch": 1.8183036569021636, + "grad_norm": 0.3017079830169678, + "learning_rate": 1.7195411430536408e-06, + "loss": 0.6353, + "step": 4748 + }, + { + "epoch": 1.8186865785946775, + "grad_norm": 0.32286104559898376, + "learning_rate": 1.71857632272805e-06, + "loss": 0.6211, + "step": 4749 + }, + { + "epoch": 1.8190695002871913, + "grad_norm": 0.3177458345890045, + "learning_rate": 1.7176116313729302e-06, + "loss": 0.6001, + "step": 4750 + }, + { + "epoch": 1.819452421979705, + "grad_norm": 0.3031606674194336, + "learning_rate": 1.7166470691474985e-06, + "loss": 0.6349, + "step": 4751 + }, + { + "epoch": 1.819835343672219, + "grad_norm": 0.3121035397052765, + "learning_rate": 1.7156826362109523e-06, + "loss": 0.636, + "step": 4752 + }, + { + "epoch": 1.820218265364733, + "grad_norm": 0.3169090449810028, + "learning_rate": 1.714718332722467e-06, + "loss": 0.6119, + "step": 4753 + }, + { + "epoch": 1.820601187057247, + "grad_norm": 0.31004953384399414, + "learning_rate": 1.7137541588411966e-06, + "loss": 0.6082, + "step": 4754 + }, + { + "epoch": 1.8209841087497607, + "grad_norm": 0.3123345375061035, + "learning_rate": 1.7127901147262732e-06, + "loss": 0.6399, + "step": 4755 + }, + { + "epoch": 1.8213670304422744, + "grad_norm": 0.3082245886325836, + "learning_rate": 1.711826200536808e-06, + "loss": 0.6285, + "step": 4756 + }, + { + "epoch": 1.8217499521347884, + "grad_norm": 0.3051731586456299, + "learning_rate": 1.710862416431893e-06, + "loss": 0.6239, + "step": 4757 + }, + { + "epoch": 1.8221328738273024, + "grad_norm": 0.31963053345680237, + "learning_rate": 1.7098987625705926e-06, + "loss": 0.635, + "step": 4758 + }, + { + "epoch": 1.8225157955198164, + "grad_norm": 0.321990430355072, + "learning_rate": 1.7089352391119554e-06, + "loss": 0.6162, + "step": 4759 + }, + { + "epoch": 1.82289871721233, + "grad_norm": 0.31493422389030457, + "learning_rate": 1.7079718462150074e-06, + "loss": 0.6221, + "step": 4760 + }, + { + "epoch": 1.8232816389048438, + "grad_norm": 0.31402111053466797, + "learning_rate": 1.7070085840387506e-06, + "loss": 0.6429, + "step": 4761 + }, + { + "epoch": 1.8236645605973578, + "grad_norm": 0.31030428409576416, + "learning_rate": 1.7060454527421688e-06, + "loss": 0.6333, + "step": 4762 + }, + { + "epoch": 1.8240474822898718, + "grad_norm": 0.3204543888568878, + "learning_rate": 1.7050824524842213e-06, + "loss": 0.6466, + "step": 4763 + }, + { + "epoch": 1.8244304039823858, + "grad_norm": 0.32027482986450195, + "learning_rate": 1.704119583423848e-06, + "loss": 0.597, + "step": 4764 + }, + { + "epoch": 1.8248133256748995, + "grad_norm": 0.31141820549964905, + "learning_rate": 1.7031568457199652e-06, + "loss": 0.6178, + "step": 4765 + }, + { + "epoch": 1.8251962473674133, + "grad_norm": 0.3087855875492096, + "learning_rate": 1.7021942395314684e-06, + "loss": 0.6039, + "step": 4766 + }, + { + "epoch": 1.8255791690599272, + "grad_norm": 0.32316136360168457, + "learning_rate": 1.7012317650172317e-06, + "loss": 0.6286, + "step": 4767 + }, + { + "epoch": 1.8259620907524412, + "grad_norm": 0.3089243471622467, + "learning_rate": 1.7002694223361072e-06, + "loss": 0.6276, + "step": 4768 + }, + { + "epoch": 1.826345012444955, + "grad_norm": 0.3147084414958954, + "learning_rate": 1.6993072116469256e-06, + "loss": 0.6099, + "step": 4769 + }, + { + "epoch": 1.826727934137469, + "grad_norm": 0.31519395112991333, + "learning_rate": 1.6983451331084946e-06, + "loss": 0.6148, + "step": 4770 + }, + { + "epoch": 1.8271108558299827, + "grad_norm": 0.3141353726387024, + "learning_rate": 1.6973831868796027e-06, + "loss": 0.6229, + "step": 4771 + }, + { + "epoch": 1.8274937775224966, + "grad_norm": 0.3250616192817688, + "learning_rate": 1.6964213731190129e-06, + "loss": 0.638, + "step": 4772 + }, + { + "epoch": 1.8278766992150106, + "grad_norm": 0.31556153297424316, + "learning_rate": 1.6954596919854685e-06, + "loss": 0.6269, + "step": 4773 + }, + { + "epoch": 1.8282596209075244, + "grad_norm": 0.31914612650871277, + "learning_rate": 1.6944981436376917e-06, + "loss": 0.6391, + "step": 4774 + }, + { + "epoch": 1.8286425426000383, + "grad_norm": 0.3087437152862549, + "learning_rate": 1.6935367282343806e-06, + "loss": 0.6288, + "step": 4775 + }, + { + "epoch": 1.829025464292552, + "grad_norm": 0.3086366653442383, + "learning_rate": 1.6925754459342133e-06, + "loss": 0.6099, + "step": 4776 + }, + { + "epoch": 1.829408385985066, + "grad_norm": 0.32459673285484314, + "learning_rate": 1.6916142968958447e-06, + "loss": 0.6112, + "step": 4777 + }, + { + "epoch": 1.82979130767758, + "grad_norm": 0.3065108060836792, + "learning_rate": 1.6906532812779092e-06, + "loss": 0.6318, + "step": 4778 + }, + { + "epoch": 1.8301742293700938, + "grad_norm": 0.30731749534606934, + "learning_rate": 1.6896923992390166e-06, + "loss": 0.6163, + "step": 4779 + }, + { + "epoch": 1.8305571510626077, + "grad_norm": 0.311744749546051, + "learning_rate": 1.688731650937756e-06, + "loss": 0.625, + "step": 4780 + }, + { + "epoch": 1.8309400727551215, + "grad_norm": 0.31356215476989746, + "learning_rate": 1.6877710365326966e-06, + "loss": 0.6397, + "step": 4781 + }, + { + "epoch": 1.8313229944476355, + "grad_norm": 0.30476561188697815, + "learning_rate": 1.686810556182381e-06, + "loss": 0.6292, + "step": 4782 + }, + { + "epoch": 1.8317059161401494, + "grad_norm": 0.3198491930961609, + "learning_rate": 1.6858502100453339e-06, + "loss": 0.6271, + "step": 4783 + }, + { + "epoch": 1.8320888378326632, + "grad_norm": 0.31545236706733704, + "learning_rate": 1.6848899982800559e-06, + "loss": 0.6269, + "step": 4784 + }, + { + "epoch": 1.832471759525177, + "grad_norm": 0.3098365068435669, + "learning_rate": 1.6839299210450263e-06, + "loss": 0.63, + "step": 4785 + }, + { + "epoch": 1.832854681217691, + "grad_norm": 0.3042409420013428, + "learning_rate": 1.6829699784986995e-06, + "loss": 0.6109, + "step": 4786 + }, + { + "epoch": 1.8332376029102049, + "grad_norm": 0.3153615891933441, + "learning_rate": 1.6820101707995112e-06, + "loss": 0.6309, + "step": 4787 + }, + { + "epoch": 1.8336205246027188, + "grad_norm": 0.3054594397544861, + "learning_rate": 1.6810504981058726e-06, + "loss": 0.6407, + "step": 4788 + }, + { + "epoch": 1.8340034462952326, + "grad_norm": 0.3064599335193634, + "learning_rate": 1.6800909605761743e-06, + "loss": 0.6313, + "step": 4789 + }, + { + "epoch": 1.8343863679877463, + "grad_norm": 0.3137023448944092, + "learning_rate": 1.6791315583687825e-06, + "loss": 0.5975, + "step": 4790 + }, + { + "epoch": 1.8347692896802603, + "grad_norm": 0.31386834383010864, + "learning_rate": 1.6781722916420434e-06, + "loss": 0.6053, + "step": 4791 + }, + { + "epoch": 1.8351522113727743, + "grad_norm": 0.3080231249332428, + "learning_rate": 1.6772131605542796e-06, + "loss": 0.6195, + "step": 4792 + }, + { + "epoch": 1.8355351330652883, + "grad_norm": 0.31506994366645813, + "learning_rate": 1.6762541652637904e-06, + "loss": 0.6335, + "step": 4793 + }, + { + "epoch": 1.835918054757802, + "grad_norm": 0.3188265562057495, + "learning_rate": 1.675295305928854e-06, + "loss": 0.6268, + "step": 4794 + }, + { + "epoch": 1.8363009764503158, + "grad_norm": 0.310051828622818, + "learning_rate": 1.6743365827077256e-06, + "loss": 0.623, + "step": 4795 + }, + { + "epoch": 1.8366838981428297, + "grad_norm": 0.3102946877479553, + "learning_rate": 1.6733779957586394e-06, + "loss": 0.6084, + "step": 4796 + }, + { + "epoch": 1.8370668198353437, + "grad_norm": 0.3095152676105499, + "learning_rate": 1.6724195452398045e-06, + "loss": 0.6432, + "step": 4797 + }, + { + "epoch": 1.8374497415278577, + "grad_norm": 0.31101375818252563, + "learning_rate": 1.6714612313094098e-06, + "loss": 0.6066, + "step": 4798 + }, + { + "epoch": 1.8378326632203714, + "grad_norm": 0.30710574984550476, + "learning_rate": 1.6705030541256211e-06, + "loss": 0.6248, + "step": 4799 + }, + { + "epoch": 1.8382155849128852, + "grad_norm": 0.30553314089775085, + "learning_rate": 1.6695450138465791e-06, + "loss": 0.6463, + "step": 4800 + }, + { + "epoch": 1.8385985066053991, + "grad_norm": 0.3097374737262726, + "learning_rate": 1.668587110630406e-06, + "loss": 0.6269, + "step": 4801 + }, + { + "epoch": 1.8389814282979131, + "grad_norm": 0.3098433315753937, + "learning_rate": 1.6676293446351985e-06, + "loss": 0.6221, + "step": 4802 + }, + { + "epoch": 1.839364349990427, + "grad_norm": 0.3106970489025116, + "learning_rate": 1.6666717160190322e-06, + "loss": 0.6201, + "step": 4803 + }, + { + "epoch": 1.8397472716829408, + "grad_norm": 0.3140954077243805, + "learning_rate": 1.6657142249399587e-06, + "loss": 0.6382, + "step": 4804 + }, + { + "epoch": 1.8401301933754546, + "grad_norm": 0.31580713391304016, + "learning_rate": 1.6647568715560086e-06, + "loss": 0.6123, + "step": 4805 + }, + { + "epoch": 1.8405131150679686, + "grad_norm": 0.3152233362197876, + "learning_rate": 1.6637996560251887e-06, + "loss": 0.613, + "step": 4806 + }, + { + "epoch": 1.8408960367604825, + "grad_norm": 0.3162561357021332, + "learning_rate": 1.6628425785054816e-06, + "loss": 0.622, + "step": 4807 + }, + { + "epoch": 1.8412789584529965, + "grad_norm": 0.32152849435806274, + "learning_rate": 1.6618856391548501e-06, + "loss": 0.619, + "step": 4808 + }, + { + "epoch": 1.8416618801455102, + "grad_norm": 0.32317015528678894, + "learning_rate": 1.6609288381312321e-06, + "loss": 0.6332, + "step": 4809 + }, + { + "epoch": 1.842044801838024, + "grad_norm": 0.3195847272872925, + "learning_rate": 1.6599721755925433e-06, + "loss": 0.6327, + "step": 4810 + }, + { + "epoch": 1.842427723530538, + "grad_norm": 0.3263967037200928, + "learning_rate": 1.6590156516966774e-06, + "loss": 0.6163, + "step": 4811 + }, + { + "epoch": 1.842810645223052, + "grad_norm": 0.31183549761772156, + "learning_rate": 1.6580592666015034e-06, + "loss": 0.617, + "step": 4812 + }, + { + "epoch": 1.843193566915566, + "grad_norm": 0.3224901854991913, + "learning_rate": 1.6571030204648694e-06, + "loss": 0.6213, + "step": 4813 + }, + { + "epoch": 1.8435764886080797, + "grad_norm": 0.30693385004997253, + "learning_rate": 1.6561469134445986e-06, + "loss": 0.6279, + "step": 4814 + }, + { + "epoch": 1.8439594103005934, + "grad_norm": 0.3099835216999054, + "learning_rate": 1.655190945698492e-06, + "loss": 0.6072, + "step": 4815 + }, + { + "epoch": 1.8443423319931074, + "grad_norm": 0.3195720911026001, + "learning_rate": 1.6542351173843285e-06, + "loss": 0.6224, + "step": 4816 + }, + { + "epoch": 1.8447252536856213, + "grad_norm": 0.30748066306114197, + "learning_rate": 1.653279428659863e-06, + "loss": 0.618, + "step": 4817 + }, + { + "epoch": 1.8451081753781353, + "grad_norm": 0.3036177456378937, + "learning_rate": 1.6523238796828277e-06, + "loss": 0.6323, + "step": 4818 + }, + { + "epoch": 1.845491097070649, + "grad_norm": 0.318185418844223, + "learning_rate": 1.6513684706109311e-06, + "loss": 0.6175, + "step": 4819 + }, + { + "epoch": 1.8458740187631628, + "grad_norm": 0.3074016869068146, + "learning_rate": 1.650413201601861e-06, + "loss": 0.6236, + "step": 4820 + }, + { + "epoch": 1.8462569404556768, + "grad_norm": 0.29662662744522095, + "learning_rate": 1.6494580728132783e-06, + "loss": 0.628, + "step": 4821 + }, + { + "epoch": 1.8466398621481908, + "grad_norm": 0.3125719726085663, + "learning_rate": 1.648503084402823e-06, + "loss": 0.6285, + "step": 4822 + }, + { + "epoch": 1.8470227838407047, + "grad_norm": 0.312045693397522, + "learning_rate": 1.6475482365281125e-06, + "loss": 0.6206, + "step": 4823 + }, + { + "epoch": 1.8474057055332185, + "grad_norm": 0.3120470345020294, + "learning_rate": 1.6465935293467388e-06, + "loss": 0.6459, + "step": 4824 + }, + { + "epoch": 1.8477886272257322, + "grad_norm": 0.3133928179740906, + "learning_rate": 1.645638963016274e-06, + "loss": 0.6301, + "step": 4825 + }, + { + "epoch": 1.8481715489182462, + "grad_norm": 0.3183332085609436, + "learning_rate": 1.6446845376942634e-06, + "loss": 0.6055, + "step": 4826 + }, + { + "epoch": 1.8485544706107602, + "grad_norm": 0.3145903944969177, + "learning_rate": 1.6437302535382316e-06, + "loss": 0.635, + "step": 4827 + }, + { + "epoch": 1.8489373923032741, + "grad_norm": 0.3021862506866455, + "learning_rate": 1.6427761107056783e-06, + "loss": 0.6097, + "step": 4828 + }, + { + "epoch": 1.849320313995788, + "grad_norm": 0.30783361196517944, + "learning_rate": 1.64182210935408e-06, + "loss": 0.6302, + "step": 4829 + }, + { + "epoch": 1.8497032356883016, + "grad_norm": 0.3158133327960968, + "learning_rate": 1.6408682496408918e-06, + "loss": 0.6216, + "step": 4830 + }, + { + "epoch": 1.8500861573808156, + "grad_norm": 0.31632304191589355, + "learning_rate": 1.6399145317235422e-06, + "loss": 0.6299, + "step": 4831 + }, + { + "epoch": 1.8504690790733296, + "grad_norm": 0.30702826380729675, + "learning_rate": 1.6389609557594396e-06, + "loss": 0.6312, + "step": 4832 + }, + { + "epoch": 1.8508520007658433, + "grad_norm": 0.3139340877532959, + "learning_rate": 1.6380075219059666e-06, + "loss": 0.6372, + "step": 4833 + }, + { + "epoch": 1.8512349224583573, + "grad_norm": 0.3229789435863495, + "learning_rate": 1.6370542303204842e-06, + "loss": 0.6018, + "step": 4834 + }, + { + "epoch": 1.851617844150871, + "grad_norm": 0.3158475160598755, + "learning_rate": 1.6361010811603276e-06, + "loss": 0.6106, + "step": 4835 + }, + { + "epoch": 1.852000765843385, + "grad_norm": 0.3098587691783905, + "learning_rate": 1.6351480745828098e-06, + "loss": 0.6191, + "step": 4836 + }, + { + "epoch": 1.852383687535899, + "grad_norm": 0.3104322552680969, + "learning_rate": 1.6341952107452208e-06, + "loss": 0.6325, + "step": 4837 + }, + { + "epoch": 1.8527666092284127, + "grad_norm": 0.335178017616272, + "learning_rate": 1.6332424898048272e-06, + "loss": 0.6313, + "step": 4838 + }, + { + "epoch": 1.8531495309209267, + "grad_norm": 0.30757004022598267, + "learning_rate": 1.6322899119188698e-06, + "loss": 0.6214, + "step": 4839 + }, + { + "epoch": 1.8535324526134405, + "grad_norm": 0.31948840618133545, + "learning_rate": 1.6313374772445687e-06, + "loss": 0.6382, + "step": 4840 + }, + { + "epoch": 1.8539153743059544, + "grad_norm": 0.3080410063266754, + "learning_rate": 1.630385185939118e-06, + "loss": 0.6398, + "step": 4841 + }, + { + "epoch": 1.8542982959984684, + "grad_norm": 0.31119096279144287, + "learning_rate": 1.62943303815969e-06, + "loss": 0.6516, + "step": 4842 + }, + { + "epoch": 1.8546812176909822, + "grad_norm": 0.31177544593811035, + "learning_rate": 1.6284810340634315e-06, + "loss": 0.6082, + "step": 4843 + }, + { + "epoch": 1.8550641393834961, + "grad_norm": 0.3069167733192444, + "learning_rate": 1.6275291738074667e-06, + "loss": 0.6355, + "step": 4844 + }, + { + "epoch": 1.8554470610760099, + "grad_norm": 0.3097560703754425, + "learning_rate": 1.6265774575488963e-06, + "loss": 0.6289, + "step": 4845 + }, + { + "epoch": 1.8558299827685238, + "grad_norm": 0.30980032682418823, + "learning_rate": 1.6256258854447962e-06, + "loss": 0.6341, + "step": 4846 + }, + { + "epoch": 1.8562129044610378, + "grad_norm": 0.3172157108783722, + "learning_rate": 1.6246744576522199e-06, + "loss": 0.6356, + "step": 4847 + }, + { + "epoch": 1.8565958261535516, + "grad_norm": 0.3076455593109131, + "learning_rate": 1.623723174328195e-06, + "loss": 0.6387, + "step": 4848 + }, + { + "epoch": 1.8569787478460653, + "grad_norm": 0.31170299649238586, + "learning_rate": 1.6227720356297288e-06, + "loss": 0.6277, + "step": 4849 + }, + { + "epoch": 1.8573616695385793, + "grad_norm": 0.3150955140590668, + "learning_rate": 1.6218210417138007e-06, + "loss": 0.6333, + "step": 4850 + }, + { + "epoch": 1.8577445912310933, + "grad_norm": 0.31612396240234375, + "learning_rate": 1.620870192737367e-06, + "loss": 0.6316, + "step": 4851 + }, + { + "epoch": 1.8581275129236072, + "grad_norm": 0.31290173530578613, + "learning_rate": 1.619919488857364e-06, + "loss": 0.6476, + "step": 4852 + }, + { + "epoch": 1.858510434616121, + "grad_norm": 0.3076207935810089, + "learning_rate": 1.6189689302306981e-06, + "loss": 0.6347, + "step": 4853 + }, + { + "epoch": 1.8588933563086347, + "grad_norm": 0.31849369406700134, + "learning_rate": 1.618018517014257e-06, + "loss": 0.646, + "step": 4854 + }, + { + "epoch": 1.8592762780011487, + "grad_norm": 0.3237910270690918, + "learning_rate": 1.6170682493649006e-06, + "loss": 0.6249, + "step": 4855 + }, + { + "epoch": 1.8596591996936627, + "grad_norm": 0.31276920437812805, + "learning_rate": 1.6161181274394683e-06, + "loss": 0.6157, + "step": 4856 + }, + { + "epoch": 1.8600421213861766, + "grad_norm": 0.31406912207603455, + "learning_rate": 1.6151681513947711e-06, + "loss": 0.6318, + "step": 4857 + }, + { + "epoch": 1.8604250430786904, + "grad_norm": 0.31612223386764526, + "learning_rate": 1.6142183213875991e-06, + "loss": 0.6311, + "step": 4858 + }, + { + "epoch": 1.8608079647712041, + "grad_norm": 0.32318368554115295, + "learning_rate": 1.6132686375747183e-06, + "loss": 0.6278, + "step": 4859 + }, + { + "epoch": 1.861190886463718, + "grad_norm": 0.3099827468395233, + "learning_rate": 1.6123191001128685e-06, + "loss": 0.6151, + "step": 4860 + }, + { + "epoch": 1.861573808156232, + "grad_norm": 0.3158944547176361, + "learning_rate": 1.611369709158768e-06, + "loss": 0.633, + "step": 4861 + }, + { + "epoch": 1.861956729848746, + "grad_norm": 0.32171231508255005, + "learning_rate": 1.6104204648691079e-06, + "loss": 0.6333, + "step": 4862 + }, + { + "epoch": 1.8623396515412598, + "grad_norm": 0.3010120987892151, + "learning_rate": 1.6094713674005587e-06, + "loss": 0.6292, + "step": 4863 + }, + { + "epoch": 1.8627225732337735, + "grad_norm": 0.30932122468948364, + "learning_rate": 1.6085224169097627e-06, + "loss": 0.6294, + "step": 4864 + }, + { + "epoch": 1.8631054949262875, + "grad_norm": 0.31684213876724243, + "learning_rate": 1.607573613553341e-06, + "loss": 0.6279, + "step": 4865 + }, + { + "epoch": 1.8634884166188015, + "grad_norm": 0.31360581517219543, + "learning_rate": 1.6066249574878889e-06, + "loss": 0.5963, + "step": 4866 + }, + { + "epoch": 1.8638713383113155, + "grad_norm": 0.3036286532878876, + "learning_rate": 1.6056764488699783e-06, + "loss": 0.6071, + "step": 4867 + }, + { + "epoch": 1.8642542600038292, + "grad_norm": 0.30579817295074463, + "learning_rate": 1.6047280878561555e-06, + "loss": 0.6196, + "step": 4868 + }, + { + "epoch": 1.864637181696343, + "grad_norm": 0.31151944398880005, + "learning_rate": 1.6037798746029444e-06, + "loss": 0.6192, + "step": 4869 + }, + { + "epoch": 1.865020103388857, + "grad_norm": 0.32098183035850525, + "learning_rate": 1.6028318092668431e-06, + "loss": 0.6344, + "step": 4870 + }, + { + "epoch": 1.865403025081371, + "grad_norm": 0.3084013760089874, + "learning_rate": 1.601883892004324e-06, + "loss": 0.637, + "step": 4871 + }, + { + "epoch": 1.8657859467738849, + "grad_norm": 0.3207344114780426, + "learning_rate": 1.6009361229718383e-06, + "loss": 0.6447, + "step": 4872 + }, + { + "epoch": 1.8661688684663986, + "grad_norm": 0.31773841381073, + "learning_rate": 1.5999885023258099e-06, + "loss": 0.6409, + "step": 4873 + }, + { + "epoch": 1.8665517901589124, + "grad_norm": 0.30972927808761597, + "learning_rate": 1.5990410302226405e-06, + "loss": 0.6334, + "step": 4874 + }, + { + "epoch": 1.8669347118514263, + "grad_norm": 0.30722248554229736, + "learning_rate": 1.598093706818705e-06, + "loss": 0.627, + "step": 4875 + }, + { + "epoch": 1.8673176335439403, + "grad_norm": 0.3157768249511719, + "learning_rate": 1.597146532270356e-06, + "loss": 0.6384, + "step": 4876 + }, + { + "epoch": 1.8677005552364543, + "grad_norm": 0.30441293120384216, + "learning_rate": 1.5961995067339197e-06, + "loss": 0.6275, + "step": 4877 + }, + { + "epoch": 1.868083476928968, + "grad_norm": 0.3194660246372223, + "learning_rate": 1.595252630365698e-06, + "loss": 0.6209, + "step": 4878 + }, + { + "epoch": 1.8684663986214818, + "grad_norm": 0.3171878457069397, + "learning_rate": 1.5943059033219693e-06, + "loss": 0.6503, + "step": 4879 + }, + { + "epoch": 1.8688493203139958, + "grad_norm": 0.3193030059337616, + "learning_rate": 1.593359325758986e-06, + "loss": 0.6329, + "step": 4880 + }, + { + "epoch": 1.8692322420065097, + "grad_norm": 0.3145703077316284, + "learning_rate": 1.5924128978329776e-06, + "loss": 0.6406, + "step": 4881 + }, + { + "epoch": 1.8696151636990237, + "grad_norm": 0.31330347061157227, + "learning_rate": 1.5914666197001463e-06, + "loss": 0.634, + "step": 4882 + }, + { + "epoch": 1.8699980853915374, + "grad_norm": 0.30764734745025635, + "learning_rate": 1.5905204915166725e-06, + "loss": 0.6256, + "step": 4883 + }, + { + "epoch": 1.8703810070840512, + "grad_norm": 0.31126853823661804, + "learning_rate": 1.5895745134387103e-06, + "loss": 0.6294, + "step": 4884 + }, + { + "epoch": 1.8707639287765652, + "grad_norm": 0.31320720911026, + "learning_rate": 1.5886286856223876e-06, + "loss": 0.6103, + "step": 4885 + }, + { + "epoch": 1.8711468504690791, + "grad_norm": 0.314548134803772, + "learning_rate": 1.5876830082238103e-06, + "loss": 0.625, + "step": 4886 + }, + { + "epoch": 1.871529772161593, + "grad_norm": 0.3105063736438751, + "learning_rate": 1.5867374813990577e-06, + "loss": 0.6172, + "step": 4887 + }, + { + "epoch": 1.8719126938541069, + "grad_norm": 0.3220457136631012, + "learning_rate": 1.5857921053041852e-06, + "loss": 0.6338, + "step": 4888 + }, + { + "epoch": 1.8722956155466206, + "grad_norm": 0.3126189410686493, + "learning_rate": 1.5848468800952225e-06, + "loss": 0.6307, + "step": 4889 + }, + { + "epoch": 1.8726785372391346, + "grad_norm": 0.3096812665462494, + "learning_rate": 1.5839018059281748e-06, + "loss": 0.6305, + "step": 4890 + }, + { + "epoch": 1.8730614589316485, + "grad_norm": 0.3056377172470093, + "learning_rate": 1.5829568829590238e-06, + "loss": 0.6394, + "step": 4891 + }, + { + "epoch": 1.8734443806241625, + "grad_norm": 0.3061021566390991, + "learning_rate": 1.5820121113437226e-06, + "loss": 0.6181, + "step": 4892 + }, + { + "epoch": 1.8738273023166763, + "grad_norm": 0.315096378326416, + "learning_rate": 1.5810674912382019e-06, + "loss": 0.6175, + "step": 4893 + }, + { + "epoch": 1.87421022400919, + "grad_norm": 0.30474328994750977, + "learning_rate": 1.5801230227983684e-06, + "loss": 0.6223, + "step": 4894 + }, + { + "epoch": 1.874593145701704, + "grad_norm": 0.293841153383255, + "learning_rate": 1.5791787061801007e-06, + "loss": 0.612, + "step": 4895 + }, + { + "epoch": 1.874976067394218, + "grad_norm": 0.3173324465751648, + "learning_rate": 1.5782345415392553e-06, + "loss": 0.6248, + "step": 4896 + }, + { + "epoch": 1.8753589890867317, + "grad_norm": 0.31970033049583435, + "learning_rate": 1.5772905290316614e-06, + "loss": 0.6424, + "step": 4897 + }, + { + "epoch": 1.8757419107792457, + "grad_norm": 0.3099396824836731, + "learning_rate": 1.5763466688131255e-06, + "loss": 0.636, + "step": 4898 + }, + { + "epoch": 1.8761248324717594, + "grad_norm": 0.316292405128479, + "learning_rate": 1.5754029610394262e-06, + "loss": 0.6288, + "step": 4899 + }, + { + "epoch": 1.8765077541642734, + "grad_norm": 0.31429773569107056, + "learning_rate": 1.5744594058663177e-06, + "loss": 0.6415, + "step": 4900 + }, + { + "epoch": 1.8768906758567874, + "grad_norm": 0.3028872311115265, + "learning_rate": 1.5735160034495312e-06, + "loss": 0.6234, + "step": 4901 + }, + { + "epoch": 1.8772735975493011, + "grad_norm": 0.31573495268821716, + "learning_rate": 1.5725727539447695e-06, + "loss": 0.6266, + "step": 4902 + }, + { + "epoch": 1.877656519241815, + "grad_norm": 0.3072626292705536, + "learning_rate": 1.5716296575077128e-06, + "loss": 0.6379, + "step": 4903 + }, + { + "epoch": 1.8780394409343288, + "grad_norm": 0.3030337691307068, + "learning_rate": 1.5706867142940142e-06, + "loss": 0.5968, + "step": 4904 + }, + { + "epoch": 1.8784223626268428, + "grad_norm": 0.3235377073287964, + "learning_rate": 1.5697439244593033e-06, + "loss": 0.6262, + "step": 4905 + }, + { + "epoch": 1.8788052843193568, + "grad_norm": 0.3054574429988861, + "learning_rate": 1.5688012881591824e-06, + "loss": 0.5989, + "step": 4906 + }, + { + "epoch": 1.8791882060118705, + "grad_norm": 0.31811076402664185, + "learning_rate": 1.5678588055492289e-06, + "loss": 0.6418, + "step": 4907 + }, + { + "epoch": 1.8795711277043845, + "grad_norm": 0.32263267040252686, + "learning_rate": 1.5669164767849965e-06, + "loss": 0.6276, + "step": 4908 + }, + { + "epoch": 1.8799540493968983, + "grad_norm": 0.3117421269416809, + "learning_rate": 1.5659743020220113e-06, + "loss": 0.6381, + "step": 4909 + }, + { + "epoch": 1.8803369710894122, + "grad_norm": 0.3199598789215088, + "learning_rate": 1.5650322814157764e-06, + "loss": 0.6126, + "step": 4910 + }, + { + "epoch": 1.8807198927819262, + "grad_norm": 0.31634506583213806, + "learning_rate": 1.5640904151217668e-06, + "loss": 0.6335, + "step": 4911 + }, + { + "epoch": 1.88110281447444, + "grad_norm": 0.31946253776550293, + "learning_rate": 1.5631487032954344e-06, + "loss": 0.6446, + "step": 4912 + }, + { + "epoch": 1.8814857361669537, + "grad_norm": 0.3141351640224457, + "learning_rate": 1.5622071460922033e-06, + "loss": 0.6334, + "step": 4913 + }, + { + "epoch": 1.8818686578594677, + "grad_norm": 0.30748021602630615, + "learning_rate": 1.5612657436674735e-06, + "loss": 0.6167, + "step": 4914 + }, + { + "epoch": 1.8822515795519816, + "grad_norm": 0.32144254446029663, + "learning_rate": 1.5603244961766202e-06, + "loss": 0.6173, + "step": 4915 + }, + { + "epoch": 1.8826345012444956, + "grad_norm": 0.3066026568412781, + "learning_rate": 1.5593834037749908e-06, + "loss": 0.6287, + "step": 4916 + }, + { + "epoch": 1.8830174229370094, + "grad_norm": 0.31411874294281006, + "learning_rate": 1.5584424666179093e-06, + "loss": 0.6332, + "step": 4917 + }, + { + "epoch": 1.883400344629523, + "grad_norm": 0.31867849826812744, + "learning_rate": 1.557501684860673e-06, + "loss": 0.6061, + "step": 4918 + }, + { + "epoch": 1.883783266322037, + "grad_norm": 0.31233203411102295, + "learning_rate": 1.5565610586585544e-06, + "loss": 0.6148, + "step": 4919 + }, + { + "epoch": 1.884166188014551, + "grad_norm": 0.3101940453052521, + "learning_rate": 1.5556205881667976e-06, + "loss": 0.648, + "step": 4920 + }, + { + "epoch": 1.884549109707065, + "grad_norm": 0.30805471539497375, + "learning_rate": 1.5546802735406252e-06, + "loss": 0.6173, + "step": 4921 + }, + { + "epoch": 1.8849320313995788, + "grad_norm": 0.3187432587146759, + "learning_rate": 1.55374011493523e-06, + "loss": 0.6211, + "step": 4922 + }, + { + "epoch": 1.8853149530920925, + "grad_norm": 0.3143121600151062, + "learning_rate": 1.5528001125057826e-06, + "loss": 0.648, + "step": 4923 + }, + { + "epoch": 1.8856978747846065, + "grad_norm": 0.3121349513530731, + "learning_rate": 1.551860266407425e-06, + "loss": 0.6357, + "step": 4924 + }, + { + "epoch": 1.8860807964771205, + "grad_norm": 0.31299149990081787, + "learning_rate": 1.5509205767952756e-06, + "loss": 0.6131, + "step": 4925 + }, + { + "epoch": 1.8864637181696344, + "grad_norm": 0.3228021562099457, + "learning_rate": 1.5499810438244251e-06, + "loss": 0.6218, + "step": 4926 + }, + { + "epoch": 1.8868466398621482, + "grad_norm": 0.32011085748672485, + "learning_rate": 1.5490416676499404e-06, + "loss": 0.6239, + "step": 4927 + }, + { + "epoch": 1.887229561554662, + "grad_norm": 0.32844439148902893, + "learning_rate": 1.5481024484268604e-06, + "loss": 0.6456, + "step": 4928 + }, + { + "epoch": 1.887612483247176, + "grad_norm": 0.31263241171836853, + "learning_rate": 1.5471633863101982e-06, + "loss": 0.6145, + "step": 4929 + }, + { + "epoch": 1.8879954049396899, + "grad_norm": 0.3137678802013397, + "learning_rate": 1.5462244814549432e-06, + "loss": 0.6317, + "step": 4930 + }, + { + "epoch": 1.8883783266322038, + "grad_norm": 0.3138011693954468, + "learning_rate": 1.5452857340160563e-06, + "loss": 0.6408, + "step": 4931 + }, + { + "epoch": 1.8887612483247176, + "grad_norm": 0.31230077147483826, + "learning_rate": 1.5443471441484748e-06, + "loss": 0.5976, + "step": 4932 + }, + { + "epoch": 1.8891441700172313, + "grad_norm": 0.3096688985824585, + "learning_rate": 1.5434087120071076e-06, + "loss": 0.6234, + "step": 4933 + }, + { + "epoch": 1.8895270917097453, + "grad_norm": 0.30034953355789185, + "learning_rate": 1.5424704377468402e-06, + "loss": 0.6136, + "step": 4934 + }, + { + "epoch": 1.8899100134022593, + "grad_norm": 0.31016042828559875, + "learning_rate": 1.5415323215225291e-06, + "loss": 0.6142, + "step": 4935 + }, + { + "epoch": 1.8902929350947733, + "grad_norm": 0.3132694363594055, + "learning_rate": 1.540594363489006e-06, + "loss": 0.6457, + "step": 4936 + }, + { + "epoch": 1.890675856787287, + "grad_norm": 0.31245484948158264, + "learning_rate": 1.539656563801078e-06, + "loss": 0.6359, + "step": 4937 + }, + { + "epoch": 1.8910587784798008, + "grad_norm": 0.3222920894622803, + "learning_rate": 1.5387189226135236e-06, + "loss": 0.6583, + "step": 4938 + }, + { + "epoch": 1.8914417001723147, + "grad_norm": 0.31842246651649475, + "learning_rate": 1.537781440081097e-06, + "loss": 0.6194, + "step": 4939 + }, + { + "epoch": 1.8918246218648287, + "grad_norm": 0.3101430833339691, + "learning_rate": 1.5368441163585251e-06, + "loss": 0.6178, + "step": 4940 + }, + { + "epoch": 1.8922075435573427, + "grad_norm": 0.30781808495521545, + "learning_rate": 1.5359069516005098e-06, + "loss": 0.6296, + "step": 4941 + }, + { + "epoch": 1.8925904652498564, + "grad_norm": 0.3146223723888397, + "learning_rate": 1.5349699459617246e-06, + "loss": 0.6305, + "step": 4942 + }, + { + "epoch": 1.8929733869423702, + "grad_norm": 0.310890793800354, + "learning_rate": 1.5340330995968183e-06, + "loss": 0.6301, + "step": 4943 + }, + { + "epoch": 1.8933563086348841, + "grad_norm": 0.3235717713832855, + "learning_rate": 1.5330964126604137e-06, + "loss": 0.624, + "step": 4944 + }, + { + "epoch": 1.893739230327398, + "grad_norm": 0.3217226564884186, + "learning_rate": 1.5321598853071068e-06, + "loss": 0.63, + "step": 4945 + }, + { + "epoch": 1.894122152019912, + "grad_norm": 0.3114512264728546, + "learning_rate": 1.531223517691467e-06, + "loss": 0.6238, + "step": 4946 + }, + { + "epoch": 1.8945050737124258, + "grad_norm": 0.31336596608161926, + "learning_rate": 1.5302873099680378e-06, + "loss": 0.6088, + "step": 4947 + }, + { + "epoch": 1.8948879954049396, + "grad_norm": 0.3313237130641937, + "learning_rate": 1.5293512622913365e-06, + "loss": 0.6545, + "step": 4948 + }, + { + "epoch": 1.8952709170974535, + "grad_norm": 0.30452772974967957, + "learning_rate": 1.5284153748158523e-06, + "loss": 0.6268, + "step": 4949 + }, + { + "epoch": 1.8956538387899675, + "grad_norm": 0.3215401768684387, + "learning_rate": 1.5274796476960501e-06, + "loss": 0.6152, + "step": 4950 + }, + { + "epoch": 1.8960367604824815, + "grad_norm": 0.31210559606552124, + "learning_rate": 1.526544081086367e-06, + "loss": 0.6057, + "step": 4951 + }, + { + "epoch": 1.8964196821749952, + "grad_norm": 0.3118828535079956, + "learning_rate": 1.5256086751412147e-06, + "loss": 0.6292, + "step": 4952 + }, + { + "epoch": 1.896802603867509, + "grad_norm": 0.3220480978488922, + "learning_rate": 1.5246734300149774e-06, + "loss": 0.6253, + "step": 4953 + }, + { + "epoch": 1.897185525560023, + "grad_norm": 0.3106321394443512, + "learning_rate": 1.5237383458620131e-06, + "loss": 0.6106, + "step": 4954 + }, + { + "epoch": 1.897568447252537, + "grad_norm": 0.3280981779098511, + "learning_rate": 1.522803422836654e-06, + "loss": 0.6374, + "step": 4955 + }, + { + "epoch": 1.897951368945051, + "grad_norm": 0.3003244698047638, + "learning_rate": 1.5218686610932032e-06, + "loss": 0.6235, + "step": 4956 + }, + { + "epoch": 1.8983342906375646, + "grad_norm": 0.3088221549987793, + "learning_rate": 1.5209340607859403e-06, + "loss": 0.6346, + "step": 4957 + }, + { + "epoch": 1.8987172123300784, + "grad_norm": 0.32254278659820557, + "learning_rate": 1.519999622069116e-06, + "loss": 0.6379, + "step": 4958 + }, + { + "epoch": 1.8991001340225924, + "grad_norm": 0.31922048330307007, + "learning_rate": 1.5190653450969564e-06, + "loss": 0.6144, + "step": 4959 + }, + { + "epoch": 1.8994830557151063, + "grad_norm": 0.3055124878883362, + "learning_rate": 1.5181312300236586e-06, + "loss": 0.6244, + "step": 4960 + }, + { + "epoch": 1.89986597740762, + "grad_norm": 0.3249337077140808, + "learning_rate": 1.517197277003395e-06, + "loss": 0.6234, + "step": 4961 + }, + { + "epoch": 1.900248899100134, + "grad_norm": 0.3082480728626251, + "learning_rate": 1.5162634861903106e-06, + "loss": 0.6466, + "step": 4962 + }, + { + "epoch": 1.9006318207926478, + "grad_norm": 0.31758198142051697, + "learning_rate": 1.5153298577385218e-06, + "loss": 0.6065, + "step": 4963 + }, + { + "epoch": 1.9010147424851618, + "grad_norm": 0.3239452540874481, + "learning_rate": 1.514396391802121e-06, + "loss": 0.6156, + "step": 4964 + }, + { + "epoch": 1.9013976641776758, + "grad_norm": 0.30908533930778503, + "learning_rate": 1.5134630885351714e-06, + "loss": 0.6133, + "step": 4965 + }, + { + "epoch": 1.9017805858701895, + "grad_norm": 0.3242105543613434, + "learning_rate": 1.5125299480917124e-06, + "loss": 0.6395, + "step": 4966 + }, + { + "epoch": 1.9021635075627035, + "grad_norm": 0.3262992799282074, + "learning_rate": 1.511596970625753e-06, + "loss": 0.6319, + "step": 4967 + }, + { + "epoch": 1.9025464292552172, + "grad_norm": 0.3257356286048889, + "learning_rate": 1.5106641562912779e-06, + "loss": 0.6192, + "step": 4968 + }, + { + "epoch": 1.9029293509477312, + "grad_norm": 0.31665587425231934, + "learning_rate": 1.5097315052422442e-06, + "loss": 0.64, + "step": 4969 + }, + { + "epoch": 1.9033122726402452, + "grad_norm": 0.31195348501205444, + "learning_rate": 1.50879901763258e-06, + "loss": 0.6379, + "step": 4970 + }, + { + "epoch": 1.903695194332759, + "grad_norm": 0.31476885080337524, + "learning_rate": 1.5078666936161898e-06, + "loss": 0.6244, + "step": 4971 + }, + { + "epoch": 1.9040781160252729, + "grad_norm": 0.30885955691337585, + "learning_rate": 1.5069345333469493e-06, + "loss": 0.6201, + "step": 4972 + }, + { + "epoch": 1.9044610377177866, + "grad_norm": 0.3229456841945648, + "learning_rate": 1.5060025369787065e-06, + "loss": 0.645, + "step": 4973 + }, + { + "epoch": 1.9048439594103006, + "grad_norm": 0.31829920411109924, + "learning_rate": 1.5050707046652843e-06, + "loss": 0.6292, + "step": 4974 + }, + { + "epoch": 1.9052268811028146, + "grad_norm": 0.31123149394989014, + "learning_rate": 1.5041390365604767e-06, + "loss": 0.6383, + "step": 4975 + }, + { + "epoch": 1.9056098027953283, + "grad_norm": 0.31189393997192383, + "learning_rate": 1.5032075328180524e-06, + "loss": 0.634, + "step": 4976 + }, + { + "epoch": 1.905992724487842, + "grad_norm": 0.30978330969810486, + "learning_rate": 1.5022761935917507e-06, + "loss": 0.6217, + "step": 4977 + }, + { + "epoch": 1.906375646180356, + "grad_norm": 0.3082875609397888, + "learning_rate": 1.501345019035285e-06, + "loss": 0.6127, + "step": 4978 + }, + { + "epoch": 1.90675856787287, + "grad_norm": 0.3121907413005829, + "learning_rate": 1.5004140093023422e-06, + "loss": 0.6301, + "step": 4979 + }, + { + "epoch": 1.907141489565384, + "grad_norm": 0.3200100064277649, + "learning_rate": 1.4994831645465805e-06, + "loss": 0.6194, + "step": 4980 + }, + { + "epoch": 1.9075244112578977, + "grad_norm": 0.3028210401535034, + "learning_rate": 1.4985524849216326e-06, + "loss": 0.6293, + "step": 4981 + }, + { + "epoch": 1.9079073329504115, + "grad_norm": 0.3064923882484436, + "learning_rate": 1.4976219705811018e-06, + "loss": 0.6325, + "step": 4982 + }, + { + "epoch": 1.9082902546429255, + "grad_norm": 0.30606594681739807, + "learning_rate": 1.496691621678567e-06, + "loss": 0.6286, + "step": 4983 + }, + { + "epoch": 1.9086731763354394, + "grad_norm": 0.3103726804256439, + "learning_rate": 1.495761438367577e-06, + "loss": 0.6213, + "step": 4984 + }, + { + "epoch": 1.9090560980279534, + "grad_norm": 0.3157011866569519, + "learning_rate": 1.494831420801654e-06, + "loss": 0.6551, + "step": 4985 + }, + { + "epoch": 1.9094390197204671, + "grad_norm": 0.3113675117492676, + "learning_rate": 1.493901569134294e-06, + "loss": 0.6317, + "step": 4986 + }, + { + "epoch": 1.909821941412981, + "grad_norm": 0.32216528058052063, + "learning_rate": 1.4929718835189639e-06, + "loss": 0.6062, + "step": 4987 + }, + { + "epoch": 1.9102048631054949, + "grad_norm": 0.3082331717014313, + "learning_rate": 1.4920423641091053e-06, + "loss": 0.6218, + "step": 4988 + }, + { + "epoch": 1.9105877847980088, + "grad_norm": 0.31743526458740234, + "learning_rate": 1.4911130110581305e-06, + "loss": 0.6198, + "step": 4989 + }, + { + "epoch": 1.9109707064905228, + "grad_norm": 0.3126719892024994, + "learning_rate": 1.4901838245194261e-06, + "loss": 0.6193, + "step": 4990 + }, + { + "epoch": 1.9113536281830366, + "grad_norm": 0.3097245693206787, + "learning_rate": 1.489254804646349e-06, + "loss": 0.6329, + "step": 4991 + }, + { + "epoch": 1.9117365498755503, + "grad_norm": 0.3085971772670746, + "learning_rate": 1.4883259515922294e-06, + "loss": 0.623, + "step": 4992 + }, + { + "epoch": 1.9121194715680643, + "grad_norm": 0.3141256868839264, + "learning_rate": 1.4873972655103713e-06, + "loss": 0.6143, + "step": 4993 + }, + { + "epoch": 1.9125023932605782, + "grad_norm": 0.3154732584953308, + "learning_rate": 1.4864687465540495e-06, + "loss": 0.6005, + "step": 4994 + }, + { + "epoch": 1.9128853149530922, + "grad_norm": 0.3230641186237335, + "learning_rate": 1.4855403948765129e-06, + "loss": 0.6541, + "step": 4995 + }, + { + "epoch": 1.913268236645606, + "grad_norm": 0.30840691924095154, + "learning_rate": 1.4846122106309807e-06, + "loss": 0.6435, + "step": 4996 + }, + { + "epoch": 1.9136511583381197, + "grad_norm": 0.3141586482524872, + "learning_rate": 1.483684193970647e-06, + "loss": 0.6422, + "step": 4997 + }, + { + "epoch": 1.9140340800306337, + "grad_norm": 0.30769744515419006, + "learning_rate": 1.482756345048675e-06, + "loss": 0.6192, + "step": 4998 + }, + { + "epoch": 1.9144170017231477, + "grad_norm": 0.3096294403076172, + "learning_rate": 1.481828664018203e-06, + "loss": 0.6175, + "step": 4999 + }, + { + "epoch": 1.9147999234156616, + "grad_norm": 0.3157786726951599, + "learning_rate": 1.4809011510323396e-06, + "loss": 0.627, + "step": 5000 + }, + { + "epoch": 1.9151828451081754, + "grad_norm": 0.30312833189964294, + "learning_rate": 1.4799738062441681e-06, + "loss": 0.6412, + "step": 5001 + }, + { + "epoch": 1.9155657668006891, + "grad_norm": 0.3127349317073822, + "learning_rate": 1.4790466298067415e-06, + "loss": 0.6219, + "step": 5002 + }, + { + "epoch": 1.915948688493203, + "grad_norm": 0.3163009285926819, + "learning_rate": 1.478119621873087e-06, + "loss": 0.6204, + "step": 5003 + }, + { + "epoch": 1.916331610185717, + "grad_norm": 0.3228578567504883, + "learning_rate": 1.4771927825962033e-06, + "loss": 0.6296, + "step": 5004 + }, + { + "epoch": 1.916714531878231, + "grad_norm": 0.320868581533432, + "learning_rate": 1.476266112129059e-06, + "loss": 0.6514, + "step": 5005 + }, + { + "epoch": 1.9170974535707448, + "grad_norm": 0.31178754568099976, + "learning_rate": 1.4753396106245993e-06, + "loss": 0.6351, + "step": 5006 + }, + { + "epoch": 1.9174803752632585, + "grad_norm": 0.30461645126342773, + "learning_rate": 1.4744132782357372e-06, + "loss": 0.6191, + "step": 5007 + }, + { + "epoch": 1.9178632969557725, + "grad_norm": 0.30451157689094543, + "learning_rate": 1.4734871151153613e-06, + "loss": 0.6359, + "step": 5008 + }, + { + "epoch": 1.9182462186482865, + "grad_norm": 0.3176251947879791, + "learning_rate": 1.4725611214163294e-06, + "loss": 0.6191, + "step": 5009 + }, + { + "epoch": 1.9186291403408005, + "grad_norm": 0.30785226821899414, + "learning_rate": 1.4716352972914738e-06, + "loss": 0.6291, + "step": 5010 + }, + { + "epoch": 1.9190120620333142, + "grad_norm": 0.30682864785194397, + "learning_rate": 1.4707096428935962e-06, + "loss": 0.6143, + "step": 5011 + }, + { + "epoch": 1.919394983725828, + "grad_norm": 0.32344210147857666, + "learning_rate": 1.4697841583754738e-06, + "loss": 0.6191, + "step": 5012 + }, + { + "epoch": 1.919777905418342, + "grad_norm": 0.314776748418808, + "learning_rate": 1.4688588438898517e-06, + "loss": 0.6192, + "step": 5013 + }, + { + "epoch": 1.920160827110856, + "grad_norm": 0.3072172999382019, + "learning_rate": 1.467933699589449e-06, + "loss": 0.6307, + "step": 5014 + }, + { + "epoch": 1.9205437488033699, + "grad_norm": 0.3153126835823059, + "learning_rate": 1.467008725626958e-06, + "loss": 0.6261, + "step": 5015 + }, + { + "epoch": 1.9209266704958836, + "grad_norm": 0.3088264763355255, + "learning_rate": 1.46608392215504e-06, + "loss": 0.6013, + "step": 5016 + }, + { + "epoch": 1.9213095921883974, + "grad_norm": 0.30542412400245667, + "learning_rate": 1.4651592893263309e-06, + "loss": 0.6302, + "step": 5017 + }, + { + "epoch": 1.9216925138809113, + "grad_norm": 0.3155474066734314, + "learning_rate": 1.4642348272934358e-06, + "loss": 0.6309, + "step": 5018 + }, + { + "epoch": 1.9220754355734253, + "grad_norm": 0.31215447187423706, + "learning_rate": 1.4633105362089356e-06, + "loss": 0.6374, + "step": 5019 + }, + { + "epoch": 1.9224583572659393, + "grad_norm": 0.32837310433387756, + "learning_rate": 1.4623864162253775e-06, + "loss": 0.6356, + "step": 5020 + }, + { + "epoch": 1.922841278958453, + "grad_norm": 0.31222036480903625, + "learning_rate": 1.4614624674952843e-06, + "loss": 0.6263, + "step": 5021 + }, + { + "epoch": 1.9232242006509668, + "grad_norm": 0.31274116039276123, + "learning_rate": 1.4605386901711502e-06, + "loss": 0.6273, + "step": 5022 + }, + { + "epoch": 1.9236071223434807, + "grad_norm": 0.3057968020439148, + "learning_rate": 1.4596150844054396e-06, + "loss": 0.6204, + "step": 5023 + }, + { + "epoch": 1.9239900440359947, + "grad_norm": 0.31626540422439575, + "learning_rate": 1.45869165035059e-06, + "loss": 0.6278, + "step": 5024 + }, + { + "epoch": 1.9243729657285085, + "grad_norm": 0.3160731792449951, + "learning_rate": 1.4577683881590098e-06, + "loss": 0.6321, + "step": 5025 + }, + { + "epoch": 1.9247558874210224, + "grad_norm": 0.30913159251213074, + "learning_rate": 1.4568452979830811e-06, + "loss": 0.6257, + "step": 5026 + }, + { + "epoch": 1.9251388091135362, + "grad_norm": 0.31254154443740845, + "learning_rate": 1.4559223799751538e-06, + "loss": 0.6154, + "step": 5027 + }, + { + "epoch": 1.9255217308060502, + "grad_norm": 0.31256604194641113, + "learning_rate": 1.4549996342875507e-06, + "loss": 0.6314, + "step": 5028 + }, + { + "epoch": 1.9259046524985641, + "grad_norm": 0.3066786527633667, + "learning_rate": 1.454077061072568e-06, + "loss": 0.6246, + "step": 5029 + }, + { + "epoch": 1.9262875741910779, + "grad_norm": 0.312098890542984, + "learning_rate": 1.4531546604824725e-06, + "loss": 0.6142, + "step": 5030 + }, + { + "epoch": 1.9266704958835918, + "grad_norm": 0.3071805536746979, + "learning_rate": 1.4522324326695026e-06, + "loss": 0.6194, + "step": 5031 + }, + { + "epoch": 1.9270534175761056, + "grad_norm": 0.31440040469169617, + "learning_rate": 1.4513103777858662e-06, + "loss": 0.6367, + "step": 5032 + }, + { + "epoch": 1.9274363392686196, + "grad_norm": 0.31206777691841125, + "learning_rate": 1.4503884959837466e-06, + "loss": 0.624, + "step": 5033 + }, + { + "epoch": 1.9278192609611335, + "grad_norm": 0.31250983476638794, + "learning_rate": 1.4494667874152936e-06, + "loss": 0.6118, + "step": 5034 + }, + { + "epoch": 1.9282021826536473, + "grad_norm": 0.3171892464160919, + "learning_rate": 1.448545252232633e-06, + "loss": 0.6223, + "step": 5035 + }, + { + "epoch": 1.9285851043461613, + "grad_norm": 0.3081512153148651, + "learning_rate": 1.44762389058786e-06, + "loss": 0.6276, + "step": 5036 + }, + { + "epoch": 1.928968026038675, + "grad_norm": 0.3081750273704529, + "learning_rate": 1.44670270263304e-06, + "loss": 0.6191, + "step": 5037 + }, + { + "epoch": 1.929350947731189, + "grad_norm": 0.3247116804122925, + "learning_rate": 1.4457816885202115e-06, + "loss": 0.617, + "step": 5038 + }, + { + "epoch": 1.929733869423703, + "grad_norm": 0.3031737208366394, + "learning_rate": 1.444860848401384e-06, + "loss": 0.616, + "step": 5039 + }, + { + "epoch": 1.9301167911162167, + "grad_norm": 0.31412363052368164, + "learning_rate": 1.4439401824285387e-06, + "loss": 0.6316, + "step": 5040 + }, + { + "epoch": 1.9304997128087305, + "grad_norm": 0.3078151047229767, + "learning_rate": 1.443019690753627e-06, + "loss": 0.6384, + "step": 5041 + }, + { + "epoch": 1.9308826345012444, + "grad_norm": 0.3162228465080261, + "learning_rate": 1.4420993735285705e-06, + "loss": 0.6196, + "step": 5042 + }, + { + "epoch": 1.9312655561937584, + "grad_norm": 0.3039613366127014, + "learning_rate": 1.441179230905264e-06, + "loss": 0.6377, + "step": 5043 + }, + { + "epoch": 1.9316484778862724, + "grad_norm": 0.3221755921840668, + "learning_rate": 1.4402592630355738e-06, + "loss": 0.6306, + "step": 5044 + }, + { + "epoch": 1.9320313995787861, + "grad_norm": 0.32170745730400085, + "learning_rate": 1.4393394700713366e-06, + "loss": 0.6093, + "step": 5045 + }, + { + "epoch": 1.9324143212712999, + "grad_norm": 0.31589117646217346, + "learning_rate": 1.4384198521643588e-06, + "loss": 0.6328, + "step": 5046 + }, + { + "epoch": 1.9327972429638138, + "grad_norm": 0.31844598054885864, + "learning_rate": 1.4375004094664211e-06, + "loss": 0.6324, + "step": 5047 + }, + { + "epoch": 1.9331801646563278, + "grad_norm": 0.31353527307510376, + "learning_rate": 1.4365811421292711e-06, + "loss": 0.6431, + "step": 5048 + }, + { + "epoch": 1.9335630863488418, + "grad_norm": 0.3179960548877716, + "learning_rate": 1.435662050304631e-06, + "loss": 0.6317, + "step": 5049 + }, + { + "epoch": 1.9339460080413555, + "grad_norm": 0.32347482442855835, + "learning_rate": 1.4347431341441934e-06, + "loss": 0.6166, + "step": 5050 + }, + { + "epoch": 1.9343289297338693, + "grad_norm": 0.31788742542266846, + "learning_rate": 1.4338243937996197e-06, + "loss": 0.6271, + "step": 5051 + }, + { + "epoch": 1.9347118514263832, + "grad_norm": 0.3072623014450073, + "learning_rate": 1.4329058294225444e-06, + "loss": 0.6234, + "step": 5052 + }, + { + "epoch": 1.9350947731188972, + "grad_norm": 0.3398687541484833, + "learning_rate": 1.431987441164573e-06, + "loss": 0.6215, + "step": 5053 + }, + { + "epoch": 1.9354776948114112, + "grad_norm": 0.3236002027988434, + "learning_rate": 1.4310692291772823e-06, + "loss": 0.628, + "step": 5054 + }, + { + "epoch": 1.935860616503925, + "grad_norm": 0.322128027677536, + "learning_rate": 1.4301511936122167e-06, + "loss": 0.6299, + "step": 5055 + }, + { + "epoch": 1.9362435381964387, + "grad_norm": 0.3177052140235901, + "learning_rate": 1.429233334620896e-06, + "loss": 0.6208, + "step": 5056 + }, + { + "epoch": 1.9366264598889527, + "grad_norm": 0.3149329125881195, + "learning_rate": 1.428315652354807e-06, + "loss": 0.6429, + "step": 5057 + }, + { + "epoch": 1.9370093815814666, + "grad_norm": 0.32320544123649597, + "learning_rate": 1.4273981469654093e-06, + "loss": 0.6246, + "step": 5058 + }, + { + "epoch": 1.9373923032739806, + "grad_norm": 0.3201727867126465, + "learning_rate": 1.426480818604134e-06, + "loss": 0.6294, + "step": 5059 + }, + { + "epoch": 1.9377752249664943, + "grad_norm": 0.31700408458709717, + "learning_rate": 1.4255636674223823e-06, + "loss": 0.6351, + "step": 5060 + }, + { + "epoch": 1.938158146659008, + "grad_norm": 0.3189130127429962, + "learning_rate": 1.424646693571525e-06, + "loss": 0.627, + "step": 5061 + }, + { + "epoch": 1.938541068351522, + "grad_norm": 0.31046104431152344, + "learning_rate": 1.4237298972029037e-06, + "loss": 0.628, + "step": 5062 + }, + { + "epoch": 1.938923990044036, + "grad_norm": 0.3095782995223999, + "learning_rate": 1.4228132784678329e-06, + "loss": 0.6126, + "step": 5063 + }, + { + "epoch": 1.93930691173655, + "grad_norm": 0.31647706031799316, + "learning_rate": 1.4218968375175953e-06, + "loss": 0.6014, + "step": 5064 + }, + { + "epoch": 1.9396898334290638, + "grad_norm": 0.31570008397102356, + "learning_rate": 1.4209805745034475e-06, + "loss": 0.6369, + "step": 5065 + }, + { + "epoch": 1.9400727551215775, + "grad_norm": 0.32132506370544434, + "learning_rate": 1.420064489576612e-06, + "loss": 0.6367, + "step": 5066 + }, + { + "epoch": 1.9404556768140915, + "grad_norm": 0.30725210905075073, + "learning_rate": 1.4191485828882855e-06, + "loss": 0.6329, + "step": 5067 + }, + { + "epoch": 1.9408385985066055, + "grad_norm": 0.3084251284599304, + "learning_rate": 1.418232854589636e-06, + "loss": 0.6228, + "step": 5068 + }, + { + "epoch": 1.9412215201991194, + "grad_norm": 0.32739898562431335, + "learning_rate": 1.4173173048317972e-06, + "loss": 0.65, + "step": 5069 + }, + { + "epoch": 1.9416044418916332, + "grad_norm": 0.31981804966926575, + "learning_rate": 1.4164019337658796e-06, + "loss": 0.6296, + "step": 5070 + }, + { + "epoch": 1.941987363584147, + "grad_norm": 0.32027503848075867, + "learning_rate": 1.4154867415429582e-06, + "loss": 0.6147, + "step": 5071 + }, + { + "epoch": 1.942370285276661, + "grad_norm": 0.3184222877025604, + "learning_rate": 1.4145717283140827e-06, + "loss": 0.618, + "step": 5072 + }, + { + "epoch": 1.9427532069691749, + "grad_norm": 0.3189840316772461, + "learning_rate": 1.4136568942302719e-06, + "loss": 0.6162, + "step": 5073 + }, + { + "epoch": 1.9431361286616888, + "grad_norm": 0.31721383333206177, + "learning_rate": 1.4127422394425168e-06, + "loss": 0.6302, + "step": 5074 + }, + { + "epoch": 1.9435190503542026, + "grad_norm": 0.3193591833114624, + "learning_rate": 1.4118277641017747e-06, + "loss": 0.6157, + "step": 5075 + }, + { + "epoch": 1.9439019720467163, + "grad_norm": 0.30782389640808105, + "learning_rate": 1.4109134683589754e-06, + "loss": 0.6192, + "step": 5076 + }, + { + "epoch": 1.9442848937392303, + "grad_norm": 0.3104211091995239, + "learning_rate": 1.4099993523650202e-06, + "loss": 0.614, + "step": 5077 + }, + { + "epoch": 1.9446678154317443, + "grad_norm": 0.3127465844154358, + "learning_rate": 1.4090854162707801e-06, + "loss": 0.6225, + "step": 5078 + }, + { + "epoch": 1.9450507371242582, + "grad_norm": 0.3277418613433838, + "learning_rate": 1.4081716602270957e-06, + "loss": 0.6221, + "step": 5079 + }, + { + "epoch": 1.945433658816772, + "grad_norm": 0.31940293312072754, + "learning_rate": 1.4072580843847796e-06, + "loss": 0.6194, + "step": 5080 + }, + { + "epoch": 1.9458165805092857, + "grad_norm": 0.3107777535915375, + "learning_rate": 1.4063446888946113e-06, + "loss": 0.6316, + "step": 5081 + }, + { + "epoch": 1.9461995022017997, + "grad_norm": 0.308743953704834, + "learning_rate": 1.405431473907345e-06, + "loss": 0.6546, + "step": 5082 + }, + { + "epoch": 1.9465824238943137, + "grad_norm": 0.3127445578575134, + "learning_rate": 1.4045184395737004e-06, + "loss": 0.6409, + "step": 5083 + }, + { + "epoch": 1.9469653455868277, + "grad_norm": 0.3191375136375427, + "learning_rate": 1.4036055860443703e-06, + "loss": 0.6445, + "step": 5084 + }, + { + "epoch": 1.9473482672793414, + "grad_norm": 0.32737991213798523, + "learning_rate": 1.402692913470019e-06, + "loss": 0.6313, + "step": 5085 + }, + { + "epoch": 1.9477311889718552, + "grad_norm": 0.3125200867652893, + "learning_rate": 1.4017804220012765e-06, + "loss": 0.6189, + "step": 5086 + }, + { + "epoch": 1.9481141106643691, + "grad_norm": 0.31971582770347595, + "learning_rate": 1.4008681117887463e-06, + "loss": 0.6324, + "step": 5087 + }, + { + "epoch": 1.948497032356883, + "grad_norm": 0.3165053725242615, + "learning_rate": 1.3999559829830011e-06, + "loss": 0.6095, + "step": 5088 + }, + { + "epoch": 1.9488799540493968, + "grad_norm": 0.3131575286388397, + "learning_rate": 1.3990440357345847e-06, + "loss": 0.6183, + "step": 5089 + }, + { + "epoch": 1.9492628757419108, + "grad_norm": 0.3167624771595001, + "learning_rate": 1.3981322701940095e-06, + "loss": 0.6458, + "step": 5090 + }, + { + "epoch": 1.9496457974344246, + "grad_norm": 0.3244912624359131, + "learning_rate": 1.3972206865117566e-06, + "loss": 0.6107, + "step": 5091 + }, + { + "epoch": 1.9500287191269385, + "grad_norm": 0.30777913331985474, + "learning_rate": 1.3963092848382798e-06, + "loss": 0.634, + "step": 5092 + }, + { + "epoch": 1.9504116408194525, + "grad_norm": 0.3076857328414917, + "learning_rate": 1.3953980653240023e-06, + "loss": 0.6224, + "step": 5093 + }, + { + "epoch": 1.9507945625119663, + "grad_norm": 0.3117028474807739, + "learning_rate": 1.3944870281193178e-06, + "loss": 0.6321, + "step": 5094 + }, + { + "epoch": 1.9511774842044802, + "grad_norm": 0.32702213525772095, + "learning_rate": 1.3935761733745865e-06, + "loss": 0.6225, + "step": 5095 + }, + { + "epoch": 1.951560405896994, + "grad_norm": 0.30659353733062744, + "learning_rate": 1.392665501240143e-06, + "loss": 0.6079, + "step": 5096 + }, + { + "epoch": 1.951943327589508, + "grad_norm": 0.3014058768749237, + "learning_rate": 1.3917550118662876e-06, + "loss": 0.6222, + "step": 5097 + }, + { + "epoch": 1.952326249282022, + "grad_norm": 0.300886332988739, + "learning_rate": 1.3908447054032937e-06, + "loss": 0.6238, + "step": 5098 + }, + { + "epoch": 1.9527091709745357, + "grad_norm": 0.30490171909332275, + "learning_rate": 1.3899345820014043e-06, + "loss": 0.6161, + "step": 5099 + }, + { + "epoch": 1.9530920926670496, + "grad_norm": 0.32077914476394653, + "learning_rate": 1.389024641810829e-06, + "loss": 0.6599, + "step": 5100 + }, + { + "epoch": 1.9534750143595634, + "grad_norm": 0.32109999656677246, + "learning_rate": 1.3881148849817505e-06, + "loss": 0.6194, + "step": 5101 + }, + { + "epoch": 1.9538579360520774, + "grad_norm": 0.310263454914093, + "learning_rate": 1.38720531166432e-06, + "loss": 0.6218, + "step": 5102 + }, + { + "epoch": 1.9542408577445913, + "grad_norm": 0.306298166513443, + "learning_rate": 1.3862959220086584e-06, + "loss": 0.62, + "step": 5103 + }, + { + "epoch": 1.954623779437105, + "grad_norm": 0.31048673391342163, + "learning_rate": 1.3853867161648584e-06, + "loss": 0.6143, + "step": 5104 + }, + { + "epoch": 1.9550067011296188, + "grad_norm": 0.3021582067012787, + "learning_rate": 1.3844776942829766e-06, + "loss": 0.622, + "step": 5105 + }, + { + "epoch": 1.9553896228221328, + "grad_norm": 0.3093564808368683, + "learning_rate": 1.3835688565130445e-06, + "loss": 0.6177, + "step": 5106 + }, + { + "epoch": 1.9557725445146468, + "grad_norm": 0.31135401129722595, + "learning_rate": 1.3826602030050624e-06, + "loss": 0.602, + "step": 5107 + }, + { + "epoch": 1.9561554662071607, + "grad_norm": 0.3070765733718872, + "learning_rate": 1.381751733908999e-06, + "loss": 0.5993, + "step": 5108 + }, + { + "epoch": 1.9565383878996745, + "grad_norm": 0.3140491247177124, + "learning_rate": 1.380843449374794e-06, + "loss": 0.6291, + "step": 5109 + }, + { + "epoch": 1.9569213095921882, + "grad_norm": 0.31392115354537964, + "learning_rate": 1.3799353495523537e-06, + "loss": 0.6201, + "step": 5110 + }, + { + "epoch": 1.9573042312847022, + "grad_norm": 0.308012992143631, + "learning_rate": 1.3790274345915583e-06, + "loss": 0.637, + "step": 5111 + }, + { + "epoch": 1.9576871529772162, + "grad_norm": 0.32702088356018066, + "learning_rate": 1.3781197046422527e-06, + "loss": 0.6165, + "step": 5112 + }, + { + "epoch": 1.9580700746697302, + "grad_norm": 0.31054800748825073, + "learning_rate": 1.3772121598542548e-06, + "loss": 0.6177, + "step": 5113 + }, + { + "epoch": 1.958452996362244, + "grad_norm": 0.3131434917449951, + "learning_rate": 1.3763048003773516e-06, + "loss": 0.6086, + "step": 5114 + }, + { + "epoch": 1.9588359180547577, + "grad_norm": 0.30912819504737854, + "learning_rate": 1.3753976263612967e-06, + "loss": 0.6249, + "step": 5115 + }, + { + "epoch": 1.9592188397472716, + "grad_norm": 0.32189348340034485, + "learning_rate": 1.3744906379558165e-06, + "loss": 0.6038, + "step": 5116 + }, + { + "epoch": 1.9596017614397856, + "grad_norm": 0.31051790714263916, + "learning_rate": 1.373583835310605e-06, + "loss": 0.6194, + "step": 5117 + }, + { + "epoch": 1.9599846831322996, + "grad_norm": 0.32205265760421753, + "learning_rate": 1.3726772185753274e-06, + "loss": 0.6202, + "step": 5118 + }, + { + "epoch": 1.9603676048248133, + "grad_norm": 0.3300328850746155, + "learning_rate": 1.371770787899615e-06, + "loss": 0.6346, + "step": 5119 + }, + { + "epoch": 1.960750526517327, + "grad_norm": 0.31199911236763, + "learning_rate": 1.3708645434330698e-06, + "loss": 0.6212, + "step": 5120 + }, + { + "epoch": 1.961133448209841, + "grad_norm": 0.3036895990371704, + "learning_rate": 1.369958485325264e-06, + "loss": 0.611, + "step": 5121 + }, + { + "epoch": 1.961516369902355, + "grad_norm": 0.30478569865226746, + "learning_rate": 1.3690526137257381e-06, + "loss": 0.6318, + "step": 5122 + }, + { + "epoch": 1.961899291594869, + "grad_norm": 0.3192116320133209, + "learning_rate": 1.3681469287840038e-06, + "loss": 0.6147, + "step": 5123 + }, + { + "epoch": 1.9622822132873827, + "grad_norm": 0.31774240732192993, + "learning_rate": 1.367241430649538e-06, + "loss": 0.628, + "step": 5124 + }, + { + "epoch": 1.9626651349798965, + "grad_norm": 0.32279643416404724, + "learning_rate": 1.3663361194717912e-06, + "loss": 0.6418, + "step": 5125 + }, + { + "epoch": 1.9630480566724104, + "grad_norm": 0.3127553164958954, + "learning_rate": 1.3654309954001788e-06, + "loss": 0.6227, + "step": 5126 + }, + { + "epoch": 1.9634309783649244, + "grad_norm": 0.30828770995140076, + "learning_rate": 1.3645260585840886e-06, + "loss": 0.6152, + "step": 5127 + }, + { + "epoch": 1.9638139000574384, + "grad_norm": 0.33443233370780945, + "learning_rate": 1.3636213091728774e-06, + "loss": 0.6338, + "step": 5128 + }, + { + "epoch": 1.9641968217499521, + "grad_norm": 0.32175934314727783, + "learning_rate": 1.362716747315868e-06, + "loss": 0.6336, + "step": 5129 + }, + { + "epoch": 1.9645797434424659, + "grad_norm": 0.33465608954429626, + "learning_rate": 1.361812373162355e-06, + "loss": 0.6289, + "step": 5130 + }, + { + "epoch": 1.9649626651349799, + "grad_norm": 0.3117779493331909, + "learning_rate": 1.360908186861602e-06, + "loss": 0.6485, + "step": 5131 + }, + { + "epoch": 1.9653455868274938, + "grad_norm": 0.32846710085868835, + "learning_rate": 1.360004188562841e-06, + "loss": 0.6164, + "step": 5132 + }, + { + "epoch": 1.9657285085200078, + "grad_norm": 0.3181144595146179, + "learning_rate": 1.3591003784152712e-06, + "loss": 0.6273, + "step": 5133 + }, + { + "epoch": 1.9661114302125215, + "grad_norm": 0.31782659888267517, + "learning_rate": 1.3581967565680648e-06, + "loss": 0.6181, + "step": 5134 + }, + { + "epoch": 1.9664943519050353, + "grad_norm": 0.31081587076187134, + "learning_rate": 1.3572933231703582e-06, + "loss": 0.6208, + "step": 5135 + }, + { + "epoch": 1.9668772735975493, + "grad_norm": 0.3121805787086487, + "learning_rate": 1.3563900783712597e-06, + "loss": 0.6313, + "step": 5136 + }, + { + "epoch": 1.9672601952900632, + "grad_norm": 0.34626999497413635, + "learning_rate": 1.3554870223198465e-06, + "loss": 0.6108, + "step": 5137 + }, + { + "epoch": 1.9676431169825772, + "grad_norm": 0.30951911211013794, + "learning_rate": 1.3545841551651646e-06, + "loss": 0.615, + "step": 5138 + }, + { + "epoch": 1.968026038675091, + "grad_norm": 0.31300225853919983, + "learning_rate": 1.353681477056227e-06, + "loss": 0.6163, + "step": 5139 + }, + { + "epoch": 1.9684089603676047, + "grad_norm": 0.316706120967865, + "learning_rate": 1.3527789881420159e-06, + "loss": 0.6175, + "step": 5140 + }, + { + "epoch": 1.9687918820601187, + "grad_norm": 0.30724725127220154, + "learning_rate": 1.351876688571484e-06, + "loss": 0.6185, + "step": 5141 + }, + { + "epoch": 1.9691748037526327, + "grad_norm": 0.30257532000541687, + "learning_rate": 1.3509745784935519e-06, + "loss": 0.6254, + "step": 5142 + }, + { + "epoch": 1.9695577254451466, + "grad_norm": 0.3238000273704529, + "learning_rate": 1.3500726580571094e-06, + "loss": 0.6349, + "step": 5143 + }, + { + "epoch": 1.9699406471376604, + "grad_norm": 0.31432944536209106, + "learning_rate": 1.3491709274110132e-06, + "loss": 0.6236, + "step": 5144 + }, + { + "epoch": 1.9703235688301741, + "grad_norm": 0.31099751591682434, + "learning_rate": 1.34826938670409e-06, + "loss": 0.6345, + "step": 5145 + }, + { + "epoch": 1.970706490522688, + "grad_norm": 0.30358022451400757, + "learning_rate": 1.3473680360851367e-06, + "loss": 0.6296, + "step": 5146 + }, + { + "epoch": 1.971089412215202, + "grad_norm": 0.31128251552581787, + "learning_rate": 1.3464668757029147e-06, + "loss": 0.6165, + "step": 5147 + }, + { + "epoch": 1.971472333907716, + "grad_norm": 0.3069625198841095, + "learning_rate": 1.3455659057061588e-06, + "loss": 0.6172, + "step": 5148 + }, + { + "epoch": 1.9718552556002298, + "grad_norm": 0.3083273768424988, + "learning_rate": 1.3446651262435679e-06, + "loss": 0.6295, + "step": 5149 + }, + { + "epoch": 1.9722381772927435, + "grad_norm": 0.3068319261074066, + "learning_rate": 1.3437645374638127e-06, + "loss": 0.617, + "step": 5150 + }, + { + "epoch": 1.9726210989852575, + "grad_norm": 0.3190390467643738, + "learning_rate": 1.3428641395155315e-06, + "loss": 0.6175, + "step": 5151 + }, + { + "epoch": 1.9730040206777715, + "grad_norm": 0.3331483006477356, + "learning_rate": 1.3419639325473316e-06, + "loss": 0.6311, + "step": 5152 + }, + { + "epoch": 1.9733869423702852, + "grad_norm": 0.30718666315078735, + "learning_rate": 1.3410639167077872e-06, + "loss": 0.6275, + "step": 5153 + }, + { + "epoch": 1.9737698640627992, + "grad_norm": 0.32706186175346375, + "learning_rate": 1.3401640921454411e-06, + "loss": 0.628, + "step": 5154 + }, + { + "epoch": 1.974152785755313, + "grad_norm": 0.32413598895072937, + "learning_rate": 1.3392644590088061e-06, + "loss": 0.5923, + "step": 5155 + }, + { + "epoch": 1.974535707447827, + "grad_norm": 0.32903605699539185, + "learning_rate": 1.3383650174463629e-06, + "loss": 0.6281, + "step": 5156 + }, + { + "epoch": 1.9749186291403409, + "grad_norm": 0.3228914737701416, + "learning_rate": 1.3374657676065598e-06, + "loss": 0.6279, + "step": 5157 + }, + { + "epoch": 1.9753015508328546, + "grad_norm": 0.3164849281311035, + "learning_rate": 1.336566709637815e-06, + "loss": 0.6245, + "step": 5158 + }, + { + "epoch": 1.9756844725253686, + "grad_norm": 0.32843542098999023, + "learning_rate": 1.3356678436885125e-06, + "loss": 0.6138, + "step": 5159 + }, + { + "epoch": 1.9760673942178824, + "grad_norm": 0.31241267919540405, + "learning_rate": 1.3347691699070077e-06, + "loss": 0.6116, + "step": 5160 + }, + { + "epoch": 1.9764503159103963, + "grad_norm": 0.3158017694950104, + "learning_rate": 1.3338706884416208e-06, + "loss": 0.6276, + "step": 5161 + }, + { + "epoch": 1.9768332376029103, + "grad_norm": 0.3219403624534607, + "learning_rate": 1.332972399440643e-06, + "loss": 0.6274, + "step": 5162 + }, + { + "epoch": 1.977216159295424, + "grad_norm": 0.31369057297706604, + "learning_rate": 1.3320743030523341e-06, + "loss": 0.6132, + "step": 5163 + }, + { + "epoch": 1.977599080987938, + "grad_norm": 0.3202962875366211, + "learning_rate": 1.3311763994249188e-06, + "loss": 0.6041, + "step": 5164 + }, + { + "epoch": 1.9779820026804518, + "grad_norm": 0.31284499168395996, + "learning_rate": 1.3302786887065927e-06, + "loss": 0.6257, + "step": 5165 + }, + { + "epoch": 1.9783649243729657, + "grad_norm": 0.3128538727760315, + "learning_rate": 1.3293811710455195e-06, + "loss": 0.6151, + "step": 5166 + }, + { + "epoch": 1.9787478460654797, + "grad_norm": 0.3180862367153168, + "learning_rate": 1.3284838465898309e-06, + "loss": 0.6241, + "step": 5167 + }, + { + "epoch": 1.9791307677579935, + "grad_norm": 0.3294867277145386, + "learning_rate": 1.327586715487626e-06, + "loss": 0.6122, + "step": 5168 + }, + { + "epoch": 1.9795136894505072, + "grad_norm": 0.3150072991847992, + "learning_rate": 1.3266897778869704e-06, + "loss": 0.6182, + "step": 5169 + }, + { + "epoch": 1.9798966111430212, + "grad_norm": 0.3125482499599457, + "learning_rate": 1.325793033935901e-06, + "loss": 0.6245, + "step": 5170 + }, + { + "epoch": 1.9802795328355352, + "grad_norm": 0.31183719635009766, + "learning_rate": 1.3248964837824218e-06, + "loss": 0.635, + "step": 5171 + }, + { + "epoch": 1.9806624545280491, + "grad_norm": 0.31182435154914856, + "learning_rate": 1.3240001275745046e-06, + "loss": 0.5993, + "step": 5172 + }, + { + "epoch": 1.9810453762205629, + "grad_norm": 0.30675509572029114, + "learning_rate": 1.3231039654600875e-06, + "loss": 0.6473, + "step": 5173 + }, + { + "epoch": 1.9814282979130766, + "grad_norm": 0.30671364068984985, + "learning_rate": 1.32220799758708e-06, + "loss": 0.6389, + "step": 5174 + }, + { + "epoch": 1.9818112196055906, + "grad_norm": 0.31439539790153503, + "learning_rate": 1.3213122241033554e-06, + "loss": 0.6275, + "step": 5175 + }, + { + "epoch": 1.9821941412981046, + "grad_norm": 0.3124629557132721, + "learning_rate": 1.3204166451567585e-06, + "loss": 0.6241, + "step": 5176 + }, + { + "epoch": 1.9825770629906185, + "grad_norm": 0.31686821579933167, + "learning_rate": 1.3195212608951009e-06, + "loss": 0.6323, + "step": 5177 + }, + { + "epoch": 1.9829599846831323, + "grad_norm": 0.31350550055503845, + "learning_rate": 1.31862607146616e-06, + "loss": 0.6181, + "step": 5178 + }, + { + "epoch": 1.983342906375646, + "grad_norm": 0.29978227615356445, + "learning_rate": 1.317731077017684e-06, + "loss": 0.6349, + "step": 5179 + }, + { + "epoch": 1.98372582806816, + "grad_norm": 0.30903324484825134, + "learning_rate": 1.3168362776973875e-06, + "loss": 0.6234, + "step": 5180 + }, + { + "epoch": 1.984108749760674, + "grad_norm": 0.313711941242218, + "learning_rate": 1.3159416736529545e-06, + "loss": 0.6182, + "step": 5181 + }, + { + "epoch": 1.984491671453188, + "grad_norm": 0.31516361236572266, + "learning_rate": 1.3150472650320339e-06, + "loss": 0.6214, + "step": 5182 + }, + { + "epoch": 1.9848745931457017, + "grad_norm": 0.31742000579833984, + "learning_rate": 1.3141530519822432e-06, + "loss": 0.6292, + "step": 5183 + }, + { + "epoch": 1.9852575148382154, + "grad_norm": 0.3144228756427765, + "learning_rate": 1.3132590346511693e-06, + "loss": 0.6029, + "step": 5184 + }, + { + "epoch": 1.9856404365307294, + "grad_norm": 0.31334105134010315, + "learning_rate": 1.312365213186365e-06, + "loss": 0.6079, + "step": 5185 + }, + { + "epoch": 1.9860233582232434, + "grad_norm": 0.31280407309532166, + "learning_rate": 1.3114715877353523e-06, + "loss": 0.6218, + "step": 5186 + }, + { + "epoch": 1.9864062799157574, + "grad_norm": 0.3055483102798462, + "learning_rate": 1.3105781584456208e-06, + "loss": 0.6143, + "step": 5187 + }, + { + "epoch": 1.986789201608271, + "grad_norm": 0.3104936182498932, + "learning_rate": 1.3096849254646255e-06, + "loss": 0.6073, + "step": 5188 + }, + { + "epoch": 1.9871721233007849, + "grad_norm": 0.315992534160614, + "learning_rate": 1.3087918889397914e-06, + "loss": 0.6135, + "step": 5189 + }, + { + "epoch": 1.9875550449932988, + "grad_norm": 0.3113788068294525, + "learning_rate": 1.307899049018509e-06, + "loss": 0.6302, + "step": 5190 + }, + { + "epoch": 1.9879379666858128, + "grad_norm": 0.3093588054180145, + "learning_rate": 1.3070064058481383e-06, + "loss": 0.6164, + "step": 5191 + }, + { + "epoch": 1.9883208883783268, + "grad_norm": 0.31237781047821045, + "learning_rate": 1.306113959576007e-06, + "loss": 0.6247, + "step": 5192 + }, + { + "epoch": 1.9887038100708405, + "grad_norm": 0.3112935721874237, + "learning_rate": 1.3052217103494074e-06, + "loss": 0.6278, + "step": 5193 + }, + { + "epoch": 1.9890867317633543, + "grad_norm": 0.3184419572353363, + "learning_rate": 1.3043296583156023e-06, + "loss": 0.614, + "step": 5194 + }, + { + "epoch": 1.9894696534558682, + "grad_norm": 0.31362298130989075, + "learning_rate": 1.303437803621821e-06, + "loss": 0.6463, + "step": 5195 + }, + { + "epoch": 1.9898525751483822, + "grad_norm": 0.3274880349636078, + "learning_rate": 1.3025461464152608e-06, + "loss": 0.6455, + "step": 5196 + }, + { + "epoch": 1.9902354968408962, + "grad_norm": 0.313505619764328, + "learning_rate": 1.3016546868430848e-06, + "loss": 0.6316, + "step": 5197 + }, + { + "epoch": 1.99061841853341, + "grad_norm": 0.312380850315094, + "learning_rate": 1.3007634250524236e-06, + "loss": 0.6284, + "step": 5198 + }, + { + "epoch": 1.9910013402259237, + "grad_norm": 0.3117523193359375, + "learning_rate": 1.2998723611903768e-06, + "loss": 0.6205, + "step": 5199 + }, + { + "epoch": 1.9913842619184376, + "grad_norm": 0.3151082992553711, + "learning_rate": 1.2989814954040107e-06, + "loss": 0.6409, + "step": 5200 + }, + { + "epoch": 1.9917671836109516, + "grad_norm": 0.344458669424057, + "learning_rate": 1.2980908278403592e-06, + "loss": 0.5952, + "step": 5201 + }, + { + "epoch": 1.9921501053034656, + "grad_norm": 0.3198493421077728, + "learning_rate": 1.2972003586464213e-06, + "loss": 0.6223, + "step": 5202 + }, + { + "epoch": 1.9925330269959793, + "grad_norm": 0.3215101957321167, + "learning_rate": 1.296310087969167e-06, + "loss": 0.6157, + "step": 5203 + }, + { + "epoch": 1.992915948688493, + "grad_norm": 0.31487035751342773, + "learning_rate": 1.2954200159555294e-06, + "loss": 0.6127, + "step": 5204 + }, + { + "epoch": 1.993298870381007, + "grad_norm": 0.3286244869232178, + "learning_rate": 1.294530142752412e-06, + "loss": 0.621, + "step": 5205 + }, + { + "epoch": 1.993681792073521, + "grad_norm": 0.3406181037425995, + "learning_rate": 1.2936404685066852e-06, + "loss": 0.6102, + "step": 5206 + }, + { + "epoch": 1.994064713766035, + "grad_norm": 0.3176896572113037, + "learning_rate": 1.2927509933651838e-06, + "loss": 0.6307, + "step": 5207 + }, + { + "epoch": 1.9944476354585488, + "grad_norm": 0.311084508895874, + "learning_rate": 1.2918617174747128e-06, + "loss": 0.6281, + "step": 5208 + }, + { + "epoch": 1.9948305571510625, + "grad_norm": 0.31288790702819824, + "learning_rate": 1.2909726409820433e-06, + "loss": 0.6171, + "step": 5209 + }, + { + "epoch": 1.9952134788435765, + "grad_norm": 0.31540316343307495, + "learning_rate": 1.2900837640339138e-06, + "loss": 0.6287, + "step": 5210 + }, + { + "epoch": 1.9955964005360904, + "grad_norm": 0.3272678852081299, + "learning_rate": 1.2891950867770291e-06, + "loss": 0.6346, + "step": 5211 + }, + { + "epoch": 1.9959793222286044, + "grad_norm": 0.32290002703666687, + "learning_rate": 1.28830660935806e-06, + "loss": 0.6298, + "step": 5212 + }, + { + "epoch": 1.9963622439211182, + "grad_norm": 0.3152669370174408, + "learning_rate": 1.2874183319236471e-06, + "loss": 0.6329, + "step": 5213 + }, + { + "epoch": 1.996745165613632, + "grad_norm": 0.306155800819397, + "learning_rate": 1.286530254620396e-06, + "loss": 0.6001, + "step": 5214 + }, + { + "epoch": 1.9971280873061459, + "grad_norm": 0.3182489573955536, + "learning_rate": 1.2856423775948807e-06, + "loss": 0.6388, + "step": 5215 + }, + { + "epoch": 1.9975110089986599, + "grad_norm": 0.3153967261314392, + "learning_rate": 1.2847547009936414e-06, + "loss": 0.6255, + "step": 5216 + }, + { + "epoch": 1.9978939306911736, + "grad_norm": 0.3209621012210846, + "learning_rate": 1.2838672249631845e-06, + "loss": 0.638, + "step": 5217 + }, + { + "epoch": 1.9982768523836876, + "grad_norm": 0.30864375829696655, + "learning_rate": 1.2829799496499835e-06, + "loss": 0.6151, + "step": 5218 + }, + { + "epoch": 1.9986597740762013, + "grad_norm": 0.32462942600250244, + "learning_rate": 1.2820928752004794e-06, + "loss": 0.6502, + "step": 5219 + }, + { + "epoch": 1.9990426957687153, + "grad_norm": 0.31418466567993164, + "learning_rate": 1.2812060017610802e-06, + "loss": 0.6191, + "step": 5220 + }, + { + "epoch": 1.9994256174612293, + "grad_norm": 0.3115984797477722, + "learning_rate": 1.2803193294781613e-06, + "loss": 0.6295, + "step": 5221 + }, + { + "epoch": 1.999808539153743, + "grad_norm": 0.31889766454696655, + "learning_rate": 1.2794328584980623e-06, + "loss": 0.6195, + "step": 5222 + } + ], + "logging_steps": 1, + "max_steps": 7833, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 2611, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 4.6794347211000906e+19, + "train_batch_size": 12, + "trial_name": null, + "trial_params": null +}