{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 3.0, "eval_steps": 500, "global_step": 1053, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.002849002849002849, "grad_norm": 56.0523474942412, "learning_rate": 0.0, "loss": 11.2375, "step": 1 }, { "epoch": 0.005698005698005698, "grad_norm": 56.692305048985084, "learning_rate": 4.7169811320754717e-07, "loss": 11.1498, "step": 2 }, { "epoch": 0.008547008547008548, "grad_norm": 56.20671215302259, "learning_rate": 9.433962264150943e-07, "loss": 11.1995, "step": 3 }, { "epoch": 0.011396011396011397, "grad_norm": 57.07016050341919, "learning_rate": 1.4150943396226415e-06, "loss": 11.17, "step": 4 }, { "epoch": 0.014245014245014245, "grad_norm": 55.67070270531962, "learning_rate": 1.8867924528301887e-06, "loss": 11.2141, "step": 5 }, { "epoch": 0.017094017094017096, "grad_norm": 67.41256511207831, "learning_rate": 2.358490566037736e-06, "loss": 10.8094, "step": 6 }, { "epoch": 0.019943019943019943, "grad_norm": 66.98704800790512, "learning_rate": 2.830188679245283e-06, "loss": 10.6191, "step": 7 }, { "epoch": 0.022792022792022793, "grad_norm": 99.53214132413315, "learning_rate": 3.30188679245283e-06, "loss": 9.2224, "step": 8 }, { "epoch": 0.02564102564102564, "grad_norm": 107.3638343426545, "learning_rate": 3.7735849056603773e-06, "loss": 8.9747, "step": 9 }, { "epoch": 0.02849002849002849, "grad_norm": 65.50392578974085, "learning_rate": 4.245283018867925e-06, "loss": 3.7694, "step": 10 }, { "epoch": 0.03133903133903134, "grad_norm": 52.753132297048275, "learning_rate": 4.716981132075472e-06, "loss": 3.1517, "step": 11 }, { "epoch": 0.03418803418803419, "grad_norm": 34.22470483345867, "learning_rate": 5.188679245283019e-06, "loss": 2.3701, "step": 12 }, { "epoch": 0.037037037037037035, "grad_norm": 8.289999698115077, "learning_rate": 5.660377358490566e-06, "loss": 1.5114, "step": 13 }, { "epoch": 0.039886039886039885, "grad_norm": 4.907971366358463, "learning_rate": 6.132075471698113e-06, "loss": 1.3167, "step": 14 }, { "epoch": 0.042735042735042736, "grad_norm": 3.878187177104665, "learning_rate": 6.60377358490566e-06, "loss": 1.1878, "step": 15 }, { "epoch": 0.045584045584045586, "grad_norm": 2.614057256962741, "learning_rate": 7.0754716981132075e-06, "loss": 1.134, "step": 16 }, { "epoch": 0.04843304843304843, "grad_norm": 2.1563070672973015, "learning_rate": 7.547169811320755e-06, "loss": 1.0503, "step": 17 }, { "epoch": 0.05128205128205128, "grad_norm": 5.014628550733802, "learning_rate": 8.018867924528302e-06, "loss": 0.968, "step": 18 }, { "epoch": 0.05413105413105413, "grad_norm": 5.938244849443368, "learning_rate": 8.49056603773585e-06, "loss": 0.919, "step": 19 }, { "epoch": 0.05698005698005698, "grad_norm": 1.2998505975680303, "learning_rate": 8.962264150943396e-06, "loss": 0.9074, "step": 20 }, { "epoch": 0.05982905982905983, "grad_norm": 1.0231800381234573, "learning_rate": 9.433962264150944e-06, "loss": 0.864, "step": 21 }, { "epoch": 0.06267806267806268, "grad_norm": 0.791900948928139, "learning_rate": 9.905660377358492e-06, "loss": 0.8147, "step": 22 }, { "epoch": 0.06552706552706553, "grad_norm": 0.7091561135310548, "learning_rate": 1.0377358490566038e-05, "loss": 0.8188, "step": 23 }, { "epoch": 0.06837606837606838, "grad_norm": 0.7166203845304848, "learning_rate": 1.0849056603773586e-05, "loss": 0.7601, "step": 24 }, { "epoch": 0.07122507122507123, "grad_norm": 0.8012651357824443, "learning_rate": 1.1320754716981132e-05, "loss": 0.7598, "step": 25 }, { "epoch": 0.07407407407407407, "grad_norm": 0.6383142924108046, "learning_rate": 1.179245283018868e-05, "loss": 0.7514, "step": 26 }, { "epoch": 0.07692307692307693, "grad_norm": 0.5725581821662165, "learning_rate": 1.2264150943396227e-05, "loss": 0.7071, "step": 27 }, { "epoch": 0.07977207977207977, "grad_norm": 0.6425822567656265, "learning_rate": 1.2735849056603775e-05, "loss": 0.7076, "step": 28 }, { "epoch": 0.08262108262108261, "grad_norm": 0.6708098823752852, "learning_rate": 1.320754716981132e-05, "loss": 0.6887, "step": 29 }, { "epoch": 0.08547008547008547, "grad_norm": 0.5685632358364251, "learning_rate": 1.3679245283018869e-05, "loss": 0.6724, "step": 30 }, { "epoch": 0.08831908831908832, "grad_norm": 0.44124658499047564, "learning_rate": 1.4150943396226415e-05, "loss": 0.6856, "step": 31 }, { "epoch": 0.09116809116809117, "grad_norm": 0.46800011010506615, "learning_rate": 1.4622641509433963e-05, "loss": 0.6644, "step": 32 }, { "epoch": 0.09401709401709402, "grad_norm": 0.5235273680352841, "learning_rate": 1.509433962264151e-05, "loss": 0.6491, "step": 33 }, { "epoch": 0.09686609686609686, "grad_norm": 0.46201236009771707, "learning_rate": 1.5566037735849056e-05, "loss": 0.6403, "step": 34 }, { "epoch": 0.09971509971509972, "grad_norm": 0.38923522274855954, "learning_rate": 1.6037735849056604e-05, "loss": 0.6325, "step": 35 }, { "epoch": 0.10256410256410256, "grad_norm": 0.38662678714338583, "learning_rate": 1.650943396226415e-05, "loss": 0.6432, "step": 36 }, { "epoch": 0.10541310541310542, "grad_norm": 0.40192513620663517, "learning_rate": 1.69811320754717e-05, "loss": 0.6126, "step": 37 }, { "epoch": 0.10826210826210826, "grad_norm": 0.3948475707755113, "learning_rate": 1.7452830188679244e-05, "loss": 0.647, "step": 38 }, { "epoch": 0.1111111111111111, "grad_norm": 0.3356078989163562, "learning_rate": 1.7924528301886792e-05, "loss": 0.6174, "step": 39 }, { "epoch": 0.11396011396011396, "grad_norm": 0.30046759356107233, "learning_rate": 1.839622641509434e-05, "loss": 0.6072, "step": 40 }, { "epoch": 0.1168091168091168, "grad_norm": 0.30888114755625345, "learning_rate": 1.8867924528301888e-05, "loss": 0.5935, "step": 41 }, { "epoch": 0.11965811965811966, "grad_norm": 0.30715993049476403, "learning_rate": 1.9339622641509436e-05, "loss": 0.6021, "step": 42 }, { "epoch": 0.1225071225071225, "grad_norm": 0.29429080617445386, "learning_rate": 1.9811320754716984e-05, "loss": 0.5897, "step": 43 }, { "epoch": 0.12535612535612536, "grad_norm": 0.2806609748601335, "learning_rate": 2.0283018867924532e-05, "loss": 0.5712, "step": 44 }, { "epoch": 0.1282051282051282, "grad_norm": 0.3035543416731401, "learning_rate": 2.0754716981132076e-05, "loss": 0.5879, "step": 45 }, { "epoch": 0.13105413105413105, "grad_norm": 0.2757782348628876, "learning_rate": 2.1226415094339624e-05, "loss": 0.5777, "step": 46 }, { "epoch": 0.1339031339031339, "grad_norm": 0.2639801457025701, "learning_rate": 2.1698113207547172e-05, "loss": 0.5605, "step": 47 }, { "epoch": 0.13675213675213677, "grad_norm": 0.2712816187794645, "learning_rate": 2.216981132075472e-05, "loss": 0.5873, "step": 48 }, { "epoch": 0.1396011396011396, "grad_norm": 0.28871963465497535, "learning_rate": 2.2641509433962265e-05, "loss": 0.5727, "step": 49 }, { "epoch": 0.14245014245014245, "grad_norm": 0.28060397078353755, "learning_rate": 2.3113207547169813e-05, "loss": 0.5602, "step": 50 }, { "epoch": 0.1452991452991453, "grad_norm": 0.2771839138643321, "learning_rate": 2.358490566037736e-05, "loss": 0.5543, "step": 51 }, { "epoch": 0.14814814814814814, "grad_norm": 0.25817545404116393, "learning_rate": 2.405660377358491e-05, "loss": 0.5495, "step": 52 }, { "epoch": 0.150997150997151, "grad_norm": 0.2715327298990826, "learning_rate": 2.4528301886792453e-05, "loss": 0.5386, "step": 53 }, { "epoch": 0.15384615384615385, "grad_norm": 0.3009269107284128, "learning_rate": 2.5e-05, "loss": 0.54, "step": 54 }, { "epoch": 0.15669515669515668, "grad_norm": 0.26603937399372396, "learning_rate": 2.547169811320755e-05, "loss": 0.5688, "step": 55 }, { "epoch": 0.15954415954415954, "grad_norm": 0.23331692549766264, "learning_rate": 2.5943396226415094e-05, "loss": 0.5408, "step": 56 }, { "epoch": 0.1623931623931624, "grad_norm": 0.257528511255327, "learning_rate": 2.641509433962264e-05, "loss": 0.5451, "step": 57 }, { "epoch": 0.16524216524216523, "grad_norm": 0.24892299069207233, "learning_rate": 2.688679245283019e-05, "loss": 0.5488, "step": 58 }, { "epoch": 0.16809116809116809, "grad_norm": 0.22893417370533484, "learning_rate": 2.7358490566037738e-05, "loss": 0.5427, "step": 59 }, { "epoch": 0.17094017094017094, "grad_norm": 0.2529033488136545, "learning_rate": 2.7830188679245282e-05, "loss": 0.5667, "step": 60 }, { "epoch": 0.1737891737891738, "grad_norm": 0.2363533954890101, "learning_rate": 2.830188679245283e-05, "loss": 0.5535, "step": 61 }, { "epoch": 0.17663817663817663, "grad_norm": 0.22892978820463591, "learning_rate": 2.8773584905660378e-05, "loss": 0.5482, "step": 62 }, { "epoch": 0.1794871794871795, "grad_norm": 0.23882495609332313, "learning_rate": 2.9245283018867926e-05, "loss": 0.535, "step": 63 }, { "epoch": 0.18233618233618235, "grad_norm": 0.2506420779209896, "learning_rate": 2.971698113207547e-05, "loss": 0.5454, "step": 64 }, { "epoch": 0.18518518518518517, "grad_norm": 0.3063275398024107, "learning_rate": 3.018867924528302e-05, "loss": 0.5444, "step": 65 }, { "epoch": 0.18803418803418803, "grad_norm": 0.23091797227336122, "learning_rate": 3.0660377358490567e-05, "loss": 0.5524, "step": 66 }, { "epoch": 0.1908831908831909, "grad_norm": 0.26395478944072415, "learning_rate": 3.113207547169811e-05, "loss": 0.5063, "step": 67 }, { "epoch": 0.19373219373219372, "grad_norm": 0.24299930461895702, "learning_rate": 3.160377358490566e-05, "loss": 0.5369, "step": 68 }, { "epoch": 0.19658119658119658, "grad_norm": 0.24673885752288957, "learning_rate": 3.207547169811321e-05, "loss": 0.51, "step": 69 }, { "epoch": 0.19943019943019943, "grad_norm": 0.2706880306834431, "learning_rate": 3.254716981132075e-05, "loss": 0.5264, "step": 70 }, { "epoch": 0.2022792022792023, "grad_norm": 0.26249806371875767, "learning_rate": 3.30188679245283e-05, "loss": 0.5451, "step": 71 }, { "epoch": 0.20512820512820512, "grad_norm": 0.3079867132956617, "learning_rate": 3.349056603773585e-05, "loss": 0.5639, "step": 72 }, { "epoch": 0.20797720797720798, "grad_norm": 0.2677449528243656, "learning_rate": 3.39622641509434e-05, "loss": 0.5236, "step": 73 }, { "epoch": 0.21082621082621084, "grad_norm": 0.29313189476756446, "learning_rate": 3.4433962264150943e-05, "loss": 0.5238, "step": 74 }, { "epoch": 0.21367521367521367, "grad_norm": 0.3077114867531552, "learning_rate": 3.490566037735849e-05, "loss": 0.5184, "step": 75 }, { "epoch": 0.21652421652421652, "grad_norm": 0.24841288456262042, "learning_rate": 3.537735849056604e-05, "loss": 0.5164, "step": 76 }, { "epoch": 0.21937321937321938, "grad_norm": 0.28963777073614116, "learning_rate": 3.5849056603773584e-05, "loss": 0.5073, "step": 77 }, { "epoch": 0.2222222222222222, "grad_norm": 0.27823065639727734, "learning_rate": 3.632075471698113e-05, "loss": 0.5269, "step": 78 }, { "epoch": 0.22507122507122507, "grad_norm": 0.32870802313426584, "learning_rate": 3.679245283018868e-05, "loss": 0.5043, "step": 79 }, { "epoch": 0.22792022792022792, "grad_norm": 0.27025584380220774, "learning_rate": 3.7264150943396224e-05, "loss": 0.5024, "step": 80 }, { "epoch": 0.23076923076923078, "grad_norm": 0.33251614434676685, "learning_rate": 3.7735849056603776e-05, "loss": 0.5305, "step": 81 }, { "epoch": 0.2336182336182336, "grad_norm": 0.33641696003904814, "learning_rate": 3.820754716981133e-05, "loss": 0.5161, "step": 82 }, { "epoch": 0.23646723646723647, "grad_norm": 0.27164595148241766, "learning_rate": 3.867924528301887e-05, "loss": 0.4933, "step": 83 }, { "epoch": 0.23931623931623933, "grad_norm": 0.293519778964147, "learning_rate": 3.9150943396226416e-05, "loss": 0.5106, "step": 84 }, { "epoch": 0.24216524216524216, "grad_norm": 0.2711786679654629, "learning_rate": 3.962264150943397e-05, "loss": 0.5056, "step": 85 }, { "epoch": 0.245014245014245, "grad_norm": 0.2909911239118421, "learning_rate": 4.009433962264151e-05, "loss": 0.4906, "step": 86 }, { "epoch": 0.24786324786324787, "grad_norm": 0.2749435561722525, "learning_rate": 4.0566037735849064e-05, "loss": 0.5239, "step": 87 }, { "epoch": 0.25071225071225073, "grad_norm": 0.2966697733485334, "learning_rate": 4.103773584905661e-05, "loss": 0.4857, "step": 88 }, { "epoch": 0.2535612535612536, "grad_norm": 0.2577003407624272, "learning_rate": 4.150943396226415e-05, "loss": 0.5017, "step": 89 }, { "epoch": 0.2564102564102564, "grad_norm": 0.2509274192908608, "learning_rate": 4.1981132075471704e-05, "loss": 0.5094, "step": 90 }, { "epoch": 0.25925925925925924, "grad_norm": 0.26220476456285075, "learning_rate": 4.245283018867925e-05, "loss": 0.521, "step": 91 }, { "epoch": 0.2621082621082621, "grad_norm": 0.26233472271330177, "learning_rate": 4.292452830188679e-05, "loss": 0.5039, "step": 92 }, { "epoch": 0.26495726495726496, "grad_norm": 0.24761878468553983, "learning_rate": 4.3396226415094345e-05, "loss": 0.4948, "step": 93 }, { "epoch": 0.2678062678062678, "grad_norm": 0.2525575556034209, "learning_rate": 4.386792452830189e-05, "loss": 0.5044, "step": 94 }, { "epoch": 0.2706552706552707, "grad_norm": 0.27089466017971453, "learning_rate": 4.433962264150944e-05, "loss": 0.5008, "step": 95 }, { "epoch": 0.27350427350427353, "grad_norm": 0.25465583266476105, "learning_rate": 4.4811320754716985e-05, "loss": 0.4925, "step": 96 }, { "epoch": 0.27635327635327633, "grad_norm": 0.2459248256786745, "learning_rate": 4.528301886792453e-05, "loss": 0.4716, "step": 97 }, { "epoch": 0.2792022792022792, "grad_norm": 0.27690227321171584, "learning_rate": 4.575471698113208e-05, "loss": 0.5057, "step": 98 }, { "epoch": 0.28205128205128205, "grad_norm": 0.31505919622587614, "learning_rate": 4.6226415094339625e-05, "loss": 0.5196, "step": 99 }, { "epoch": 0.2849002849002849, "grad_norm": 0.2645004241178192, "learning_rate": 4.669811320754717e-05, "loss": 0.4741, "step": 100 }, { "epoch": 0.28774928774928776, "grad_norm": 0.28420887135800427, "learning_rate": 4.716981132075472e-05, "loss": 0.489, "step": 101 }, { "epoch": 0.2905982905982906, "grad_norm": 0.2979244733987184, "learning_rate": 4.7641509433962266e-05, "loss": 0.5049, "step": 102 }, { "epoch": 0.2934472934472934, "grad_norm": 0.31699020934728167, "learning_rate": 4.811320754716982e-05, "loss": 0.4846, "step": 103 }, { "epoch": 0.2962962962962963, "grad_norm": 0.2958211929457519, "learning_rate": 4.858490566037736e-05, "loss": 0.4904, "step": 104 }, { "epoch": 0.29914529914529914, "grad_norm": 0.31430371098933885, "learning_rate": 4.9056603773584906e-05, "loss": 0.4864, "step": 105 }, { "epoch": 0.301994301994302, "grad_norm": 0.31807223918157157, "learning_rate": 4.952830188679246e-05, "loss": 0.496, "step": 106 }, { "epoch": 0.30484330484330485, "grad_norm": 0.3012030394747298, "learning_rate": 5e-05, "loss": 0.4889, "step": 107 }, { "epoch": 0.3076923076923077, "grad_norm": 0.26530051400289184, "learning_rate": 4.994720168954594e-05, "loss": 0.4933, "step": 108 }, { "epoch": 0.31054131054131057, "grad_norm": 0.3230552696656243, "learning_rate": 4.989440337909187e-05, "loss": 0.5068, "step": 109 }, { "epoch": 0.31339031339031337, "grad_norm": 0.25705267366009066, "learning_rate": 4.9841605068637805e-05, "loss": 0.499, "step": 110 }, { "epoch": 0.3162393162393162, "grad_norm": 0.3143313651222358, "learning_rate": 4.978880675818374e-05, "loss": 0.4784, "step": 111 }, { "epoch": 0.3190883190883191, "grad_norm": 0.3376761893109501, "learning_rate": 4.973600844772968e-05, "loss": 0.482, "step": 112 }, { "epoch": 0.32193732193732194, "grad_norm": 0.28882046798653027, "learning_rate": 4.968321013727561e-05, "loss": 0.4961, "step": 113 }, { "epoch": 0.3247863247863248, "grad_norm": 0.34419030637706083, "learning_rate": 4.9630411826821544e-05, "loss": 0.4971, "step": 114 }, { "epoch": 0.32763532763532766, "grad_norm": 0.29426005659261695, "learning_rate": 4.957761351636748e-05, "loss": 0.4903, "step": 115 }, { "epoch": 0.33048433048433046, "grad_norm": 0.29019064666682737, "learning_rate": 4.952481520591341e-05, "loss": 0.4897, "step": 116 }, { "epoch": 0.3333333333333333, "grad_norm": 0.33297662577690634, "learning_rate": 4.947201689545935e-05, "loss": 0.4792, "step": 117 }, { "epoch": 0.33618233618233617, "grad_norm": 0.35409595270488886, "learning_rate": 4.941921858500528e-05, "loss": 0.4925, "step": 118 }, { "epoch": 0.33903133903133903, "grad_norm": 0.292433952733655, "learning_rate": 4.936642027455122e-05, "loss": 0.5069, "step": 119 }, { "epoch": 0.3418803418803419, "grad_norm": 0.3320923132157635, "learning_rate": 4.931362196409715e-05, "loss": 0.4946, "step": 120 }, { "epoch": 0.34472934472934474, "grad_norm": 0.27067630597401476, "learning_rate": 4.9260823653643085e-05, "loss": 0.4871, "step": 121 }, { "epoch": 0.3475783475783476, "grad_norm": 0.30304447173190324, "learning_rate": 4.920802534318902e-05, "loss": 0.487, "step": 122 }, { "epoch": 0.3504273504273504, "grad_norm": 0.3039645539285113, "learning_rate": 4.915522703273496e-05, "loss": 0.4867, "step": 123 }, { "epoch": 0.35327635327635326, "grad_norm": 0.2824982694185181, "learning_rate": 4.9102428722280894e-05, "loss": 0.4611, "step": 124 }, { "epoch": 0.3561253561253561, "grad_norm": 0.28260131702992686, "learning_rate": 4.9049630411826823e-05, "loss": 0.4652, "step": 125 }, { "epoch": 0.358974358974359, "grad_norm": 0.34703216534596504, "learning_rate": 4.899683210137276e-05, "loss": 0.4865, "step": 126 }, { "epoch": 0.36182336182336183, "grad_norm": 0.3162837076477058, "learning_rate": 4.894403379091869e-05, "loss": 0.4954, "step": 127 }, { "epoch": 0.3646723646723647, "grad_norm": 0.3293203697511355, "learning_rate": 4.8891235480464626e-05, "loss": 0.4995, "step": 128 }, { "epoch": 0.36752136752136755, "grad_norm": 0.27295379955106025, "learning_rate": 4.883843717001056e-05, "loss": 0.4872, "step": 129 }, { "epoch": 0.37037037037037035, "grad_norm": 0.3102376623883744, "learning_rate": 4.87856388595565e-05, "loss": 0.481, "step": 130 }, { "epoch": 0.3732193732193732, "grad_norm": 0.31091911179304077, "learning_rate": 4.8732840549102435e-05, "loss": 0.4891, "step": 131 }, { "epoch": 0.37606837606837606, "grad_norm": 0.29399727700411593, "learning_rate": 4.8680042238648365e-05, "loss": 0.4678, "step": 132 }, { "epoch": 0.3789173789173789, "grad_norm": 0.3409889899938261, "learning_rate": 4.86272439281943e-05, "loss": 0.4881, "step": 133 }, { "epoch": 0.3817663817663818, "grad_norm": 0.3046568772014337, "learning_rate": 4.857444561774023e-05, "loss": 0.4788, "step": 134 }, { "epoch": 0.38461538461538464, "grad_norm": 0.24665708555682755, "learning_rate": 4.852164730728617e-05, "loss": 0.4572, "step": 135 }, { "epoch": 0.38746438746438744, "grad_norm": 0.3016729598811574, "learning_rate": 4.8468848996832103e-05, "loss": 0.4804, "step": 136 }, { "epoch": 0.3903133903133903, "grad_norm": 0.3530427507474638, "learning_rate": 4.841605068637804e-05, "loss": 0.4843, "step": 137 }, { "epoch": 0.39316239316239315, "grad_norm": 0.2856167787643274, "learning_rate": 4.8363252375923976e-05, "loss": 0.4715, "step": 138 }, { "epoch": 0.396011396011396, "grad_norm": 0.37847572959684284, "learning_rate": 4.8310454065469906e-05, "loss": 0.4968, "step": 139 }, { "epoch": 0.39886039886039887, "grad_norm": 0.28801978777804754, "learning_rate": 4.825765575501584e-05, "loss": 0.4632, "step": 140 }, { "epoch": 0.4017094017094017, "grad_norm": 0.3525074681127096, "learning_rate": 4.820485744456177e-05, "loss": 0.4827, "step": 141 }, { "epoch": 0.4045584045584046, "grad_norm": 0.3526537528509632, "learning_rate": 4.8152059134107715e-05, "loss": 0.4732, "step": 142 }, { "epoch": 0.4074074074074074, "grad_norm": 0.34857652018932117, "learning_rate": 4.8099260823653645e-05, "loss": 0.4867, "step": 143 }, { "epoch": 0.41025641025641024, "grad_norm": 0.3727176334678898, "learning_rate": 4.804646251319958e-05, "loss": 0.4834, "step": 144 }, { "epoch": 0.4131054131054131, "grad_norm": 0.3140976089416083, "learning_rate": 4.799366420274552e-05, "loss": 0.4675, "step": 145 }, { "epoch": 0.41595441595441596, "grad_norm": 0.44474986509411835, "learning_rate": 4.794086589229145e-05, "loss": 0.4914, "step": 146 }, { "epoch": 0.4188034188034188, "grad_norm": 0.3390407952306162, "learning_rate": 4.788806758183738e-05, "loss": 0.4731, "step": 147 }, { "epoch": 0.42165242165242167, "grad_norm": 0.48836879750115086, "learning_rate": 4.783526927138332e-05, "loss": 0.4921, "step": 148 }, { "epoch": 0.42450142450142453, "grad_norm": 0.33589080513403735, "learning_rate": 4.7782470960929256e-05, "loss": 0.4912, "step": 149 }, { "epoch": 0.42735042735042733, "grad_norm": 0.36236282963696415, "learning_rate": 4.7729672650475186e-05, "loss": 0.4601, "step": 150 }, { "epoch": 0.4301994301994302, "grad_norm": 0.2797897413436219, "learning_rate": 4.767687434002112e-05, "loss": 0.4672, "step": 151 }, { "epoch": 0.43304843304843305, "grad_norm": 0.4148496937386019, "learning_rate": 4.762407602956706e-05, "loss": 0.4819, "step": 152 }, { "epoch": 0.4358974358974359, "grad_norm": 0.3395999932409137, "learning_rate": 4.757127771911299e-05, "loss": 0.4701, "step": 153 }, { "epoch": 0.43874643874643876, "grad_norm": 0.31394240245627125, "learning_rate": 4.7518479408658925e-05, "loss": 0.4702, "step": 154 }, { "epoch": 0.4415954415954416, "grad_norm": 0.3976902607497981, "learning_rate": 4.746568109820486e-05, "loss": 0.4754, "step": 155 }, { "epoch": 0.4444444444444444, "grad_norm": 0.26007701688937257, "learning_rate": 4.74128827877508e-05, "loss": 0.4719, "step": 156 }, { "epoch": 0.4472934472934473, "grad_norm": 0.3615746631327949, "learning_rate": 4.736008447729673e-05, "loss": 0.483, "step": 157 }, { "epoch": 0.45014245014245013, "grad_norm": 0.3262559663383385, "learning_rate": 4.730728616684266e-05, "loss": 0.4701, "step": 158 }, { "epoch": 0.452991452991453, "grad_norm": 0.3054913734906384, "learning_rate": 4.725448785638859e-05, "loss": 0.4755, "step": 159 }, { "epoch": 0.45584045584045585, "grad_norm": 0.28714899368154406, "learning_rate": 4.720168954593453e-05, "loss": 0.4715, "step": 160 }, { "epoch": 0.4586894586894587, "grad_norm": 0.3094277794993974, "learning_rate": 4.7148891235480466e-05, "loss": 0.4621, "step": 161 }, { "epoch": 0.46153846153846156, "grad_norm": 0.3440166736205401, "learning_rate": 4.70960929250264e-05, "loss": 0.485, "step": 162 }, { "epoch": 0.46438746438746437, "grad_norm": 0.32050130614453004, "learning_rate": 4.704329461457234e-05, "loss": 0.4699, "step": 163 }, { "epoch": 0.4672364672364672, "grad_norm": 0.3462845175709499, "learning_rate": 4.699049630411827e-05, "loss": 0.467, "step": 164 }, { "epoch": 0.4700854700854701, "grad_norm": 0.25813062925512487, "learning_rate": 4.6937697993664204e-05, "loss": 0.4678, "step": 165 }, { "epoch": 0.47293447293447294, "grad_norm": 0.2774578719211831, "learning_rate": 4.6884899683210134e-05, "loss": 0.4517, "step": 166 }, { "epoch": 0.4757834757834758, "grad_norm": 0.27947314173878857, "learning_rate": 4.683210137275608e-05, "loss": 0.465, "step": 167 }, { "epoch": 0.47863247863247865, "grad_norm": 0.22733723587768467, "learning_rate": 4.677930306230201e-05, "loss": 0.4537, "step": 168 }, { "epoch": 0.48148148148148145, "grad_norm": 0.29661252845278596, "learning_rate": 4.672650475184794e-05, "loss": 0.4846, "step": 169 }, { "epoch": 0.4843304843304843, "grad_norm": 0.3089830748079026, "learning_rate": 4.667370644139388e-05, "loss": 0.4557, "step": 170 }, { "epoch": 0.48717948717948717, "grad_norm": 0.24997311130541303, "learning_rate": 4.662090813093981e-05, "loss": 0.45, "step": 171 }, { "epoch": 0.49002849002849, "grad_norm": 0.33175156279149914, "learning_rate": 4.6568109820485746e-05, "loss": 0.4689, "step": 172 }, { "epoch": 0.4928774928774929, "grad_norm": 0.2602987055448183, "learning_rate": 4.651531151003168e-05, "loss": 0.4659, "step": 173 }, { "epoch": 0.49572649572649574, "grad_norm": 0.3460341458612977, "learning_rate": 4.646251319957762e-05, "loss": 0.4765, "step": 174 }, { "epoch": 0.4985754985754986, "grad_norm": 0.2961420307282209, "learning_rate": 4.640971488912355e-05, "loss": 0.4924, "step": 175 }, { "epoch": 0.5014245014245015, "grad_norm": 0.300448359282463, "learning_rate": 4.6356916578669484e-05, "loss": 0.4772, "step": 176 }, { "epoch": 0.5042735042735043, "grad_norm": 0.27326740002063005, "learning_rate": 4.630411826821542e-05, "loss": 0.4481, "step": 177 }, { "epoch": 0.5071225071225072, "grad_norm": 0.3260005365533567, "learning_rate": 4.625131995776135e-05, "loss": 0.4757, "step": 178 }, { "epoch": 0.50997150997151, "grad_norm": 0.3072044287066991, "learning_rate": 4.619852164730729e-05, "loss": 0.4749, "step": 179 }, { "epoch": 0.5128205128205128, "grad_norm": 0.29686452489417353, "learning_rate": 4.614572333685322e-05, "loss": 0.4599, "step": 180 }, { "epoch": 0.5156695156695157, "grad_norm": 0.294950820175362, "learning_rate": 4.609292502639916e-05, "loss": 0.4821, "step": 181 }, { "epoch": 0.5185185185185185, "grad_norm": 0.30491136720622375, "learning_rate": 4.604012671594509e-05, "loss": 0.4537, "step": 182 }, { "epoch": 0.5213675213675214, "grad_norm": 0.30581018423213313, "learning_rate": 4.5987328405491026e-05, "loss": 0.4641, "step": 183 }, { "epoch": 0.5242165242165242, "grad_norm": 0.35426832507345146, "learning_rate": 4.593453009503696e-05, "loss": 0.4671, "step": 184 }, { "epoch": 0.5270655270655271, "grad_norm": 0.33464439271326646, "learning_rate": 4.588173178458289e-05, "loss": 0.4692, "step": 185 }, { "epoch": 0.5299145299145299, "grad_norm": 0.30593521291345677, "learning_rate": 4.5828933474128835e-05, "loss": 0.4744, "step": 186 }, { "epoch": 0.5327635327635327, "grad_norm": 0.27674703467670597, "learning_rate": 4.5776135163674764e-05, "loss": 0.4437, "step": 187 }, { "epoch": 0.5356125356125356, "grad_norm": 0.280964484952, "learning_rate": 4.57233368532207e-05, "loss": 0.467, "step": 188 }, { "epoch": 0.5384615384615384, "grad_norm": 0.26185851570092444, "learning_rate": 4.567053854276663e-05, "loss": 0.4426, "step": 189 }, { "epoch": 0.5413105413105413, "grad_norm": 0.2584808882693542, "learning_rate": 4.561774023231257e-05, "loss": 0.4611, "step": 190 }, { "epoch": 0.5441595441595442, "grad_norm": 0.26904667580690866, "learning_rate": 4.55649419218585e-05, "loss": 0.4575, "step": 191 }, { "epoch": 0.5470085470085471, "grad_norm": 0.2487167605386707, "learning_rate": 4.551214361140444e-05, "loss": 0.457, "step": 192 }, { "epoch": 0.5498575498575499, "grad_norm": 0.28142301387763136, "learning_rate": 4.5459345300950376e-05, "loss": 0.4586, "step": 193 }, { "epoch": 0.5527065527065527, "grad_norm": 0.31342474806414733, "learning_rate": 4.5406546990496306e-05, "loss": 0.4866, "step": 194 }, { "epoch": 0.5555555555555556, "grad_norm": 0.2847923191286588, "learning_rate": 4.535374868004224e-05, "loss": 0.4643, "step": 195 }, { "epoch": 0.5584045584045584, "grad_norm": 0.3095635932106011, "learning_rate": 4.530095036958817e-05, "loss": 0.4645, "step": 196 }, { "epoch": 0.5612535612535613, "grad_norm": 0.24639525285599803, "learning_rate": 4.524815205913411e-05, "loss": 0.4577, "step": 197 }, { "epoch": 0.5641025641025641, "grad_norm": 0.26955963685848316, "learning_rate": 4.5195353748680044e-05, "loss": 0.4626, "step": 198 }, { "epoch": 0.5669515669515669, "grad_norm": 0.24214389870117736, "learning_rate": 4.514255543822598e-05, "loss": 0.4535, "step": 199 }, { "epoch": 0.5698005698005698, "grad_norm": 0.2610341258337823, "learning_rate": 4.508975712777192e-05, "loss": 0.4583, "step": 200 }, { "epoch": 0.5726495726495726, "grad_norm": 0.25907163228127456, "learning_rate": 4.503695881731785e-05, "loss": 0.4535, "step": 201 }, { "epoch": 0.5754985754985755, "grad_norm": 0.2761897395293128, "learning_rate": 4.498416050686378e-05, "loss": 0.4543, "step": 202 }, { "epoch": 0.5783475783475783, "grad_norm": 0.31059563404737606, "learning_rate": 4.493136219640971e-05, "loss": 0.4523, "step": 203 }, { "epoch": 0.5811965811965812, "grad_norm": 0.286738470928712, "learning_rate": 4.487856388595565e-05, "loss": 0.4836, "step": 204 }, { "epoch": 0.584045584045584, "grad_norm": 0.2967602486544982, "learning_rate": 4.4825765575501585e-05, "loss": 0.4637, "step": 205 }, { "epoch": 0.5868945868945868, "grad_norm": 0.2708153047318558, "learning_rate": 4.477296726504752e-05, "loss": 0.4497, "step": 206 }, { "epoch": 0.5897435897435898, "grad_norm": 0.2746307872171804, "learning_rate": 4.472016895459346e-05, "loss": 0.4512, "step": 207 }, { "epoch": 0.5925925925925926, "grad_norm": 0.3328817061733048, "learning_rate": 4.466737064413939e-05, "loss": 0.4613, "step": 208 }, { "epoch": 0.5954415954415955, "grad_norm": 0.2759545614945111, "learning_rate": 4.4614572333685324e-05, "loss": 0.4727, "step": 209 }, { "epoch": 0.5982905982905983, "grad_norm": 0.318009089876618, "learning_rate": 4.4561774023231254e-05, "loss": 0.4747, "step": 210 }, { "epoch": 0.6011396011396012, "grad_norm": 0.25640815318642596, "learning_rate": 4.45089757127772e-05, "loss": 0.4719, "step": 211 }, { "epoch": 0.603988603988604, "grad_norm": 0.34436494976967913, "learning_rate": 4.445617740232313e-05, "loss": 0.4646, "step": 212 }, { "epoch": 0.6068376068376068, "grad_norm": 1.7399032191303827, "learning_rate": 4.440337909186906e-05, "loss": 0.5284, "step": 213 }, { "epoch": 0.6096866096866097, "grad_norm": 0.26447171131249575, "learning_rate": 4.4350580781415e-05, "loss": 0.4598, "step": 214 }, { "epoch": 0.6125356125356125, "grad_norm": 0.2571865578633, "learning_rate": 4.429778247096093e-05, "loss": 0.4491, "step": 215 }, { "epoch": 0.6153846153846154, "grad_norm": 0.28498607832783857, "learning_rate": 4.4244984160506865e-05, "loss": 0.4572, "step": 216 }, { "epoch": 0.6182336182336182, "grad_norm": 0.23557920258321058, "learning_rate": 4.41921858500528e-05, "loss": 0.4489, "step": 217 }, { "epoch": 0.6210826210826211, "grad_norm": 0.3049575913824341, "learning_rate": 4.413938753959874e-05, "loss": 0.4505, "step": 218 }, { "epoch": 0.6239316239316239, "grad_norm": 0.2582271333688809, "learning_rate": 4.408658922914467e-05, "loss": 0.4608, "step": 219 }, { "epoch": 0.6267806267806267, "grad_norm": 0.2823458887788208, "learning_rate": 4.4033790918690604e-05, "loss": 0.4533, "step": 220 }, { "epoch": 0.6296296296296297, "grad_norm": 0.2386332818361518, "learning_rate": 4.398099260823654e-05, "loss": 0.4559, "step": 221 }, { "epoch": 0.6324786324786325, "grad_norm": 0.28795366606672157, "learning_rate": 4.392819429778247e-05, "loss": 0.4566, "step": 222 }, { "epoch": 0.6353276353276354, "grad_norm": 0.24919391725324347, "learning_rate": 4.3875395987328407e-05, "loss": 0.4716, "step": 223 }, { "epoch": 0.6381766381766382, "grad_norm": 0.2818278431774406, "learning_rate": 4.382259767687434e-05, "loss": 0.4504, "step": 224 }, { "epoch": 0.6410256410256411, "grad_norm": 0.2814493678841764, "learning_rate": 4.376979936642028e-05, "loss": 0.4434, "step": 225 }, { "epoch": 0.6438746438746439, "grad_norm": 0.2666133240725091, "learning_rate": 4.371700105596621e-05, "loss": 0.4552, "step": 226 }, { "epoch": 0.6467236467236467, "grad_norm": 0.2685467121594498, "learning_rate": 4.3664202745512145e-05, "loss": 0.4726, "step": 227 }, { "epoch": 0.6495726495726496, "grad_norm": 0.299879211553589, "learning_rate": 4.361140443505808e-05, "loss": 0.4471, "step": 228 }, { "epoch": 0.6524216524216524, "grad_norm": 0.23994886539791888, "learning_rate": 4.355860612460401e-05, "loss": 0.4588, "step": 229 }, { "epoch": 0.6552706552706553, "grad_norm": 0.32262816469344513, "learning_rate": 4.3505807814149955e-05, "loss": 0.4518, "step": 230 }, { "epoch": 0.6581196581196581, "grad_norm": 0.26072493194234536, "learning_rate": 4.3453009503695884e-05, "loss": 0.4659, "step": 231 }, { "epoch": 0.6609686609686609, "grad_norm": 0.31586564589968463, "learning_rate": 4.340021119324182e-05, "loss": 0.456, "step": 232 }, { "epoch": 0.6638176638176638, "grad_norm": 0.253201215778207, "learning_rate": 4.334741288278775e-05, "loss": 0.4519, "step": 233 }, { "epoch": 0.6666666666666666, "grad_norm": 0.33113742268580176, "learning_rate": 4.3294614572333687e-05, "loss": 0.4533, "step": 234 }, { "epoch": 0.6695156695156695, "grad_norm": 0.2760668240736258, "learning_rate": 4.324181626187962e-05, "loss": 0.4663, "step": 235 }, { "epoch": 0.6723646723646723, "grad_norm": 0.344650634136037, "learning_rate": 4.318901795142556e-05, "loss": 0.4161, "step": 236 }, { "epoch": 0.6752136752136753, "grad_norm": 0.26467354227622525, "learning_rate": 4.3136219640971496e-05, "loss": 0.4558, "step": 237 }, { "epoch": 0.6780626780626781, "grad_norm": 0.310387292035022, "learning_rate": 4.3083421330517425e-05, "loss": 0.4507, "step": 238 }, { "epoch": 0.6809116809116809, "grad_norm": 0.26104826670851183, "learning_rate": 4.303062302006336e-05, "loss": 0.4418, "step": 239 }, { "epoch": 0.6837606837606838, "grad_norm": 0.33335191034296224, "learning_rate": 4.297782470960929e-05, "loss": 0.4696, "step": 240 }, { "epoch": 0.6866096866096866, "grad_norm": 0.23349617986863688, "learning_rate": 4.292502639915523e-05, "loss": 0.4354, "step": 241 }, { "epoch": 0.6894586894586895, "grad_norm": 0.3264954704665932, "learning_rate": 4.2872228088701164e-05, "loss": 0.4495, "step": 242 }, { "epoch": 0.6923076923076923, "grad_norm": 0.23219348921049893, "learning_rate": 4.28194297782471e-05, "loss": 0.4368, "step": 243 }, { "epoch": 0.6951566951566952, "grad_norm": 0.30581540788611133, "learning_rate": 4.276663146779304e-05, "loss": 0.4632, "step": 244 }, { "epoch": 0.698005698005698, "grad_norm": 0.2544190208608847, "learning_rate": 4.2713833157338966e-05, "loss": 0.4289, "step": 245 }, { "epoch": 0.7008547008547008, "grad_norm": 0.28161765276112943, "learning_rate": 4.26610348468849e-05, "loss": 0.4581, "step": 246 }, { "epoch": 0.7037037037037037, "grad_norm": 0.22266448187906657, "learning_rate": 4.260823653643083e-05, "loss": 0.4442, "step": 247 }, { "epoch": 0.7065527065527065, "grad_norm": 0.2845469466464988, "learning_rate": 4.255543822597677e-05, "loss": 0.4574, "step": 248 }, { "epoch": 0.7094017094017094, "grad_norm": 0.27607235956300574, "learning_rate": 4.2502639915522705e-05, "loss": 0.4486, "step": 249 }, { "epoch": 0.7122507122507122, "grad_norm": 0.2771911103653057, "learning_rate": 4.244984160506864e-05, "loss": 0.4502, "step": 250 }, { "epoch": 0.7150997150997151, "grad_norm": 0.30336246911969705, "learning_rate": 4.239704329461457e-05, "loss": 0.4521, "step": 251 }, { "epoch": 0.717948717948718, "grad_norm": 0.2530206241923089, "learning_rate": 4.234424498416051e-05, "loss": 0.4548, "step": 252 }, { "epoch": 0.7207977207977208, "grad_norm": 0.3101890920570867, "learning_rate": 4.2291446673706444e-05, "loss": 0.4477, "step": 253 }, { "epoch": 0.7236467236467237, "grad_norm": 0.255737006138244, "learning_rate": 4.2238648363252374e-05, "loss": 0.4494, "step": 254 }, { "epoch": 0.7264957264957265, "grad_norm": 0.3158292287331288, "learning_rate": 4.218585005279832e-05, "loss": 0.4486, "step": 255 }, { "epoch": 0.7293447293447294, "grad_norm": 0.2874817334553929, "learning_rate": 4.2133051742344246e-05, "loss": 0.4492, "step": 256 }, { "epoch": 0.7321937321937322, "grad_norm": 0.2993674086081719, "learning_rate": 4.208025343189018e-05, "loss": 0.446, "step": 257 }, { "epoch": 0.7350427350427351, "grad_norm": 0.26020269412885194, "learning_rate": 4.202745512143611e-05, "loss": 0.4416, "step": 258 }, { "epoch": 0.7378917378917379, "grad_norm": 0.29543250701797663, "learning_rate": 4.197465681098205e-05, "loss": 0.4481, "step": 259 }, { "epoch": 0.7407407407407407, "grad_norm": 0.24853929821482904, "learning_rate": 4.1921858500527985e-05, "loss": 0.46, "step": 260 }, { "epoch": 0.7435897435897436, "grad_norm": 0.3195779365953625, "learning_rate": 4.186906019007392e-05, "loss": 0.4508, "step": 261 }, { "epoch": 0.7464387464387464, "grad_norm": 0.2326123884289747, "learning_rate": 4.181626187961986e-05, "loss": 0.446, "step": 262 }, { "epoch": 0.7492877492877493, "grad_norm": 0.2848262848249219, "learning_rate": 4.176346356916579e-05, "loss": 0.4558, "step": 263 }, { "epoch": 0.7521367521367521, "grad_norm": 0.2499757501835736, "learning_rate": 4.1710665258711724e-05, "loss": 0.4426, "step": 264 }, { "epoch": 0.7549857549857549, "grad_norm": 0.2914547474394867, "learning_rate": 4.1657866948257654e-05, "loss": 0.4563, "step": 265 }, { "epoch": 0.7578347578347578, "grad_norm": 0.2576885537277797, "learning_rate": 4.160506863780359e-05, "loss": 0.4396, "step": 266 }, { "epoch": 0.7606837606837606, "grad_norm": 0.26204392829857903, "learning_rate": 4.1552270327349526e-05, "loss": 0.4472, "step": 267 }, { "epoch": 0.7635327635327636, "grad_norm": 0.3265158503996716, "learning_rate": 4.149947201689546e-05, "loss": 0.451, "step": 268 }, { "epoch": 0.7663817663817664, "grad_norm": 0.2609942596958091, "learning_rate": 4.14466737064414e-05, "loss": 0.4595, "step": 269 }, { "epoch": 0.7692307692307693, "grad_norm": 0.268707951552568, "learning_rate": 4.139387539598733e-05, "loss": 0.4447, "step": 270 }, { "epoch": 0.7720797720797721, "grad_norm": 0.28137031045693933, "learning_rate": 4.1341077085533265e-05, "loss": 0.4549, "step": 271 }, { "epoch": 0.7749287749287749, "grad_norm": 0.29190078199597813, "learning_rate": 4.1288278775079195e-05, "loss": 0.4477, "step": 272 }, { "epoch": 0.7777777777777778, "grad_norm": 0.2454715815920555, "learning_rate": 4.123548046462513e-05, "loss": 0.4324, "step": 273 }, { "epoch": 0.7806267806267806, "grad_norm": 0.3094193045635567, "learning_rate": 4.118268215417107e-05, "loss": 0.427, "step": 274 }, { "epoch": 0.7834757834757835, "grad_norm": 0.24434281879639438, "learning_rate": 4.1129883843717004e-05, "loss": 0.4446, "step": 275 }, { "epoch": 0.7863247863247863, "grad_norm": 0.29785841511085437, "learning_rate": 4.107708553326294e-05, "loss": 0.4316, "step": 276 }, { "epoch": 0.7891737891737892, "grad_norm": 0.24685554556974273, "learning_rate": 4.102428722280887e-05, "loss": 0.4376, "step": 277 }, { "epoch": 0.792022792022792, "grad_norm": 0.24898713039939369, "learning_rate": 4.0971488912354806e-05, "loss": 0.4403, "step": 278 }, { "epoch": 0.7948717948717948, "grad_norm": 0.30241270134797515, "learning_rate": 4.0918690601900736e-05, "loss": 0.4416, "step": 279 }, { "epoch": 0.7977207977207977, "grad_norm": 0.24307748159009482, "learning_rate": 4.086589229144668e-05, "loss": 0.4398, "step": 280 }, { "epoch": 0.8005698005698005, "grad_norm": 0.26394622572849324, "learning_rate": 4.081309398099261e-05, "loss": 0.4421, "step": 281 }, { "epoch": 0.8034188034188035, "grad_norm": 0.24299948905266622, "learning_rate": 4.0760295670538545e-05, "loss": 0.456, "step": 282 }, { "epoch": 0.8062678062678063, "grad_norm": 0.2538995235843933, "learning_rate": 4.070749736008448e-05, "loss": 0.4411, "step": 283 }, { "epoch": 0.8091168091168092, "grad_norm": 0.24352793364283998, "learning_rate": 4.065469904963041e-05, "loss": 0.4457, "step": 284 }, { "epoch": 0.811965811965812, "grad_norm": 0.754328353026504, "learning_rate": 4.060190073917635e-05, "loss": 0.4749, "step": 285 }, { "epoch": 0.8148148148148148, "grad_norm": 0.2608557320949828, "learning_rate": 4.0549102428722284e-05, "loss": 0.4366, "step": 286 }, { "epoch": 0.8176638176638177, "grad_norm": 0.25680855446240963, "learning_rate": 4.049630411826822e-05, "loss": 0.4435, "step": 287 }, { "epoch": 0.8205128205128205, "grad_norm": 0.23235073760106573, "learning_rate": 4.044350580781415e-05, "loss": 0.4507, "step": 288 }, { "epoch": 0.8233618233618234, "grad_norm": 0.2503349096849241, "learning_rate": 4.0390707497360086e-05, "loss": 0.4474, "step": 289 }, { "epoch": 0.8262108262108262, "grad_norm": 0.24087457888823843, "learning_rate": 4.033790918690602e-05, "loss": 0.4312, "step": 290 }, { "epoch": 0.8290598290598291, "grad_norm": 0.25448968083447765, "learning_rate": 4.028511087645195e-05, "loss": 0.4466, "step": 291 }, { "epoch": 0.8319088319088319, "grad_norm": 0.24829756385015966, "learning_rate": 4.023231256599789e-05, "loss": 0.4545, "step": 292 }, { "epoch": 0.8347578347578347, "grad_norm": 0.25607715347273097, "learning_rate": 4.0179514255543825e-05, "loss": 0.4565, "step": 293 }, { "epoch": 0.8376068376068376, "grad_norm": 0.3091527511324264, "learning_rate": 4.012671594508976e-05, "loss": 0.4561, "step": 294 }, { "epoch": 0.8404558404558404, "grad_norm": 0.24780329551924565, "learning_rate": 4.007391763463569e-05, "loss": 0.4442, "step": 295 }, { "epoch": 0.8433048433048433, "grad_norm": 0.2931917392412753, "learning_rate": 4.002111932418163e-05, "loss": 0.4562, "step": 296 }, { "epoch": 0.8461538461538461, "grad_norm": 0.23945446840779216, "learning_rate": 3.9968321013727564e-05, "loss": 0.4407, "step": 297 }, { "epoch": 0.8490028490028491, "grad_norm": 0.27070479647529605, "learning_rate": 3.991552270327349e-05, "loss": 0.4371, "step": 298 }, { "epoch": 0.8518518518518519, "grad_norm": 0.2692423182003942, "learning_rate": 3.9862724392819437e-05, "loss": 0.4461, "step": 299 }, { "epoch": 0.8547008547008547, "grad_norm": 0.2707107693436041, "learning_rate": 3.9809926082365366e-05, "loss": 0.4435, "step": 300 }, { "epoch": 0.8575498575498576, "grad_norm": 0.2500219756684885, "learning_rate": 3.97571277719113e-05, "loss": 0.4288, "step": 301 }, { "epoch": 0.8603988603988604, "grad_norm": 0.2783790061607216, "learning_rate": 3.970432946145723e-05, "loss": 0.4486, "step": 302 }, { "epoch": 0.8632478632478633, "grad_norm": 0.2407111058958332, "learning_rate": 3.965153115100317e-05, "loss": 0.4514, "step": 303 }, { "epoch": 0.8660968660968661, "grad_norm": 0.30074552592413345, "learning_rate": 3.9598732840549105e-05, "loss": 0.4408, "step": 304 }, { "epoch": 0.8689458689458689, "grad_norm": 0.21341465024781467, "learning_rate": 3.954593453009504e-05, "loss": 0.4216, "step": 305 }, { "epoch": 0.8717948717948718, "grad_norm": 0.2938618051639469, "learning_rate": 3.949313621964098e-05, "loss": 0.4416, "step": 306 }, { "epoch": 0.8746438746438746, "grad_norm": 0.48732469986333266, "learning_rate": 3.944033790918691e-05, "loss": 0.4516, "step": 307 }, { "epoch": 0.8774928774928775, "grad_norm": 0.2735554566382607, "learning_rate": 3.9387539598732844e-05, "loss": 0.4383, "step": 308 }, { "epoch": 0.8803418803418803, "grad_norm": 0.28479664265954874, "learning_rate": 3.933474128827877e-05, "loss": 0.4458, "step": 309 }, { "epoch": 0.8831908831908832, "grad_norm": 0.22798637141219172, "learning_rate": 3.928194297782471e-05, "loss": 0.442, "step": 310 }, { "epoch": 0.886039886039886, "grad_norm": 0.24677969245732811, "learning_rate": 3.9229144667370646e-05, "loss": 0.4481, "step": 311 }, { "epoch": 0.8888888888888888, "grad_norm": 0.29202684528549283, "learning_rate": 3.917634635691658e-05, "loss": 0.4412, "step": 312 }, { "epoch": 0.8917378917378918, "grad_norm": 0.2606008473665173, "learning_rate": 3.912354804646252e-05, "loss": 0.435, "step": 313 }, { "epoch": 0.8945868945868946, "grad_norm": 0.26749609172100086, "learning_rate": 3.907074973600845e-05, "loss": 0.4516, "step": 314 }, { "epoch": 0.8974358974358975, "grad_norm": 0.22520943451800887, "learning_rate": 3.9017951425554385e-05, "loss": 0.4241, "step": 315 }, { "epoch": 0.9002849002849003, "grad_norm": 0.22909468804333016, "learning_rate": 3.8965153115100314e-05, "loss": 0.4379, "step": 316 }, { "epoch": 0.9031339031339032, "grad_norm": 0.2836733485621368, "learning_rate": 3.891235480464625e-05, "loss": 0.4476, "step": 317 }, { "epoch": 0.905982905982906, "grad_norm": 0.244148163521808, "learning_rate": 3.885955649419219e-05, "loss": 0.4546, "step": 318 }, { "epoch": 0.9088319088319088, "grad_norm": 0.24672915009941915, "learning_rate": 3.8806758183738124e-05, "loss": 0.4368, "step": 319 }, { "epoch": 0.9116809116809117, "grad_norm": 0.4272615119652992, "learning_rate": 3.875395987328406e-05, "loss": 0.4334, "step": 320 }, { "epoch": 0.9145299145299145, "grad_norm": 0.24895020848554578, "learning_rate": 3.870116156282999e-05, "loss": 0.4442, "step": 321 }, { "epoch": 0.9173789173789174, "grad_norm": 0.2658053036431955, "learning_rate": 3.8648363252375926e-05, "loss": 0.4383, "step": 322 }, { "epoch": 0.9202279202279202, "grad_norm": 0.2643219649864309, "learning_rate": 3.8595564941921856e-05, "loss": 0.4251, "step": 323 }, { "epoch": 0.9230769230769231, "grad_norm": 0.2365954433037453, "learning_rate": 3.85427666314678e-05, "loss": 0.4499, "step": 324 }, { "epoch": 0.9259259259259259, "grad_norm": 0.2633023483060389, "learning_rate": 3.848996832101373e-05, "loss": 0.4384, "step": 325 }, { "epoch": 0.9287749287749287, "grad_norm": 0.2956130036845057, "learning_rate": 3.8437170010559665e-05, "loss": 0.4396, "step": 326 }, { "epoch": 0.9316239316239316, "grad_norm": 0.24882987462496686, "learning_rate": 3.83843717001056e-05, "loss": 0.45, "step": 327 }, { "epoch": 0.9344729344729344, "grad_norm": 1.5891449688358956, "learning_rate": 3.833157338965153e-05, "loss": 0.4415, "step": 328 }, { "epoch": 0.9373219373219374, "grad_norm": 0.3349715325263592, "learning_rate": 3.827877507919747e-05, "loss": 0.4457, "step": 329 }, { "epoch": 0.9401709401709402, "grad_norm": 0.2848178852794376, "learning_rate": 3.8225976768743404e-05, "loss": 0.456, "step": 330 }, { "epoch": 0.9430199430199431, "grad_norm": 0.302751156256454, "learning_rate": 3.817317845828934e-05, "loss": 0.4387, "step": 331 }, { "epoch": 0.9458689458689459, "grad_norm": 0.25955277850770286, "learning_rate": 3.812038014783527e-05, "loss": 0.44, "step": 332 }, { "epoch": 0.9487179487179487, "grad_norm": 0.2502011385796011, "learning_rate": 3.8067581837381206e-05, "loss": 0.4515, "step": 333 }, { "epoch": 0.9515669515669516, "grad_norm": 0.2650885248812146, "learning_rate": 3.801478352692714e-05, "loss": 0.4475, "step": 334 }, { "epoch": 0.9544159544159544, "grad_norm": 0.24800530026732448, "learning_rate": 3.796198521647307e-05, "loss": 0.4589, "step": 335 }, { "epoch": 0.9572649572649573, "grad_norm": 0.24826963064202007, "learning_rate": 3.7909186906019015e-05, "loss": 0.4338, "step": 336 }, { "epoch": 0.9601139601139601, "grad_norm": 0.23027863612091976, "learning_rate": 3.7856388595564945e-05, "loss": 0.4301, "step": 337 }, { "epoch": 0.9629629629629629, "grad_norm": 0.22579121856289947, "learning_rate": 3.780359028511088e-05, "loss": 0.4592, "step": 338 }, { "epoch": 0.9658119658119658, "grad_norm": 0.2548134083345822, "learning_rate": 3.775079197465681e-05, "loss": 0.442, "step": 339 }, { "epoch": 0.9686609686609686, "grad_norm": 0.24004323105561068, "learning_rate": 3.769799366420275e-05, "loss": 0.4379, "step": 340 }, { "epoch": 0.9715099715099715, "grad_norm": 0.22500518319359325, "learning_rate": 3.764519535374868e-05, "loss": 0.4383, "step": 341 }, { "epoch": 0.9743589743589743, "grad_norm": 0.23655137638652632, "learning_rate": 3.759239704329461e-05, "loss": 0.4304, "step": 342 }, { "epoch": 0.9772079772079773, "grad_norm": 0.23186063099612125, "learning_rate": 3.7539598732840556e-05, "loss": 0.457, "step": 343 }, { "epoch": 0.98005698005698, "grad_norm": 0.22345800936056795, "learning_rate": 3.7486800422386486e-05, "loss": 0.4233, "step": 344 }, { "epoch": 0.9829059829059829, "grad_norm": 0.22901659678419103, "learning_rate": 3.743400211193242e-05, "loss": 0.4575, "step": 345 }, { "epoch": 0.9857549857549858, "grad_norm": 0.34541054663294773, "learning_rate": 3.738120380147835e-05, "loss": 0.4539, "step": 346 }, { "epoch": 0.9886039886039886, "grad_norm": 0.24563352556354082, "learning_rate": 3.732840549102429e-05, "loss": 0.4326, "step": 347 }, { "epoch": 0.9914529914529915, "grad_norm": 0.28378321755449876, "learning_rate": 3.727560718057022e-05, "loss": 0.4527, "step": 348 }, { "epoch": 0.9943019943019943, "grad_norm": 0.2376023328816585, "learning_rate": 3.722280887011616e-05, "loss": 0.465, "step": 349 }, { "epoch": 0.9971509971509972, "grad_norm": 0.23913998831345243, "learning_rate": 3.717001055966209e-05, "loss": 0.4502, "step": 350 }, { "epoch": 1.0, "grad_norm": 0.2446661739630635, "learning_rate": 3.711721224920803e-05, "loss": 0.436, "step": 351 }, { "epoch": 1.002849002849003, "grad_norm": 0.30212623378795767, "learning_rate": 3.7064413938753963e-05, "loss": 0.3773, "step": 352 }, { "epoch": 1.0056980056980056, "grad_norm": 0.2727102306693993, "learning_rate": 3.701161562829989e-05, "loss": 0.3721, "step": 353 }, { "epoch": 1.0085470085470085, "grad_norm": 0.24519118257200442, "learning_rate": 3.695881731784583e-05, "loss": 0.3556, "step": 354 }, { "epoch": 1.0113960113960114, "grad_norm": 0.2708532250345628, "learning_rate": 3.6906019007391766e-05, "loss": 0.3631, "step": 355 }, { "epoch": 1.0142450142450143, "grad_norm": 0.2848868957990222, "learning_rate": 3.68532206969377e-05, "loss": 0.3647, "step": 356 }, { "epoch": 1.017094017094017, "grad_norm": 0.2609789117933919, "learning_rate": 3.680042238648363e-05, "loss": 0.3598, "step": 357 }, { "epoch": 1.01994301994302, "grad_norm": 0.30731469194417627, "learning_rate": 3.674762407602957e-05, "loss": 0.3905, "step": 358 }, { "epoch": 1.0227920227920229, "grad_norm": 0.2751477526094167, "learning_rate": 3.6694825765575505e-05, "loss": 0.3766, "step": 359 }, { "epoch": 1.0256410256410255, "grad_norm": 0.3041300782163382, "learning_rate": 3.6642027455121434e-05, "loss": 0.3598, "step": 360 }, { "epoch": 1.0284900284900285, "grad_norm": 0.29025593237743824, "learning_rate": 3.658922914466738e-05, "loss": 0.3656, "step": 361 }, { "epoch": 1.0313390313390314, "grad_norm": 0.26450155835177075, "learning_rate": 3.653643083421331e-05, "loss": 0.3765, "step": 362 }, { "epoch": 1.0341880341880343, "grad_norm": 0.29763564022742284, "learning_rate": 3.648363252375924e-05, "loss": 0.3677, "step": 363 }, { "epoch": 1.037037037037037, "grad_norm": 0.2531649793647418, "learning_rate": 3.643083421330517e-05, "loss": 0.3687, "step": 364 }, { "epoch": 1.03988603988604, "grad_norm": 0.278947460622146, "learning_rate": 3.637803590285111e-05, "loss": 0.3988, "step": 365 }, { "epoch": 1.0427350427350428, "grad_norm": 0.28829274848081526, "learning_rate": 3.6325237592397046e-05, "loss": 0.3724, "step": 366 }, { "epoch": 1.0455840455840455, "grad_norm": 0.2618749521918147, "learning_rate": 3.6272439281942975e-05, "loss": 0.3542, "step": 367 }, { "epoch": 1.0484330484330484, "grad_norm": 2.0062858158052452, "learning_rate": 3.621964097148892e-05, "loss": 0.3987, "step": 368 }, { "epoch": 1.0512820512820513, "grad_norm": 0.28445975801774975, "learning_rate": 3.616684266103485e-05, "loss": 0.3653, "step": 369 }, { "epoch": 1.0541310541310542, "grad_norm": 0.2462050134870002, "learning_rate": 3.6114044350580785e-05, "loss": 0.3609, "step": 370 }, { "epoch": 1.056980056980057, "grad_norm": 0.28062644181944496, "learning_rate": 3.6061246040126714e-05, "loss": 0.3865, "step": 371 }, { "epoch": 1.0598290598290598, "grad_norm": 0.24606665304462752, "learning_rate": 3.600844772967265e-05, "loss": 0.3502, "step": 372 }, { "epoch": 1.0626780626780628, "grad_norm": 0.3129397581417704, "learning_rate": 3.595564941921859e-05, "loss": 0.3373, "step": 373 }, { "epoch": 1.0655270655270654, "grad_norm": 0.2957891434648334, "learning_rate": 3.590285110876452e-05, "loss": 0.3591, "step": 374 }, { "epoch": 1.0683760683760684, "grad_norm": 0.2570065060278186, "learning_rate": 3.585005279831046e-05, "loss": 0.3861, "step": 375 }, { "epoch": 1.0712250712250713, "grad_norm": 0.25605383657422265, "learning_rate": 3.579725448785639e-05, "loss": 0.3834, "step": 376 }, { "epoch": 1.074074074074074, "grad_norm": 0.26319733021304453, "learning_rate": 3.5744456177402326e-05, "loss": 0.3973, "step": 377 }, { "epoch": 1.0769230769230769, "grad_norm": 0.26823999119784286, "learning_rate": 3.5691657866948255e-05, "loss": 0.3843, "step": 378 }, { "epoch": 1.0797720797720798, "grad_norm": 0.2801739766417507, "learning_rate": 3.563885955649419e-05, "loss": 0.3613, "step": 379 }, { "epoch": 1.0826210826210827, "grad_norm": 0.2752489664634236, "learning_rate": 3.558606124604013e-05, "loss": 0.3608, "step": 380 }, { "epoch": 1.0854700854700854, "grad_norm": 0.253420517582108, "learning_rate": 3.5533262935586064e-05, "loss": 0.3654, "step": 381 }, { "epoch": 1.0883190883190883, "grad_norm": 0.2352861368854798, "learning_rate": 3.5480464625132e-05, "loss": 0.3567, "step": 382 }, { "epoch": 1.0911680911680912, "grad_norm": 0.24009458895765762, "learning_rate": 3.542766631467793e-05, "loss": 0.3622, "step": 383 }, { "epoch": 1.0940170940170941, "grad_norm": 0.26857315547032595, "learning_rate": 3.537486800422387e-05, "loss": 0.3654, "step": 384 }, { "epoch": 1.0968660968660968, "grad_norm": 0.3755228777720786, "learning_rate": 3.5322069693769796e-05, "loss": 0.3785, "step": 385 }, { "epoch": 1.0997150997150997, "grad_norm": 0.25523099197348387, "learning_rate": 3.526927138331573e-05, "loss": 0.3703, "step": 386 }, { "epoch": 1.1025641025641026, "grad_norm": 0.25215901520376743, "learning_rate": 3.521647307286167e-05, "loss": 0.3717, "step": 387 }, { "epoch": 1.1054131054131053, "grad_norm": 0.21795121913627405, "learning_rate": 3.5163674762407606e-05, "loss": 0.3613, "step": 388 }, { "epoch": 1.1082621082621082, "grad_norm": 0.21930209265394288, "learning_rate": 3.511087645195354e-05, "loss": 0.3656, "step": 389 }, { "epoch": 1.1111111111111112, "grad_norm": 0.2443453451193516, "learning_rate": 3.505807814149947e-05, "loss": 0.3569, "step": 390 }, { "epoch": 1.1139601139601139, "grad_norm": 0.23464081224607158, "learning_rate": 3.500527983104541e-05, "loss": 0.3733, "step": 391 }, { "epoch": 1.1168091168091168, "grad_norm": 0.23312948395717237, "learning_rate": 3.495248152059134e-05, "loss": 0.3695, "step": 392 }, { "epoch": 1.1196581196581197, "grad_norm": 0.20794997263096304, "learning_rate": 3.489968321013728e-05, "loss": 0.3573, "step": 393 }, { "epoch": 1.1225071225071226, "grad_norm": 0.26627834231921077, "learning_rate": 3.484688489968321e-05, "loss": 0.3801, "step": 394 }, { "epoch": 1.1253561253561253, "grad_norm": 0.210991273677547, "learning_rate": 3.479408658922915e-05, "loss": 0.3447, "step": 395 }, { "epoch": 1.1282051282051282, "grad_norm": 0.23724562353300943, "learning_rate": 3.474128827877508e-05, "loss": 0.356, "step": 396 }, { "epoch": 1.131054131054131, "grad_norm": 0.2629540478309352, "learning_rate": 3.468848996832101e-05, "loss": 0.3808, "step": 397 }, { "epoch": 1.133903133903134, "grad_norm": 0.21190850006514703, "learning_rate": 3.463569165786695e-05, "loss": 0.3606, "step": 398 }, { "epoch": 1.1367521367521367, "grad_norm": 0.22884348327135656, "learning_rate": 3.4582893347412886e-05, "loss": 0.3612, "step": 399 }, { "epoch": 1.1396011396011396, "grad_norm": 0.2387204757429875, "learning_rate": 3.453009503695882e-05, "loss": 0.3663, "step": 400 }, { "epoch": 1.1424501424501425, "grad_norm": 0.2531987312790015, "learning_rate": 3.447729672650475e-05, "loss": 0.3708, "step": 401 }, { "epoch": 1.1452991452991452, "grad_norm": 0.23529785297450195, "learning_rate": 3.442449841605069e-05, "loss": 0.3908, "step": 402 }, { "epoch": 1.1481481481481481, "grad_norm": 0.21187095337907066, "learning_rate": 3.4371700105596624e-05, "loss": 0.3746, "step": 403 }, { "epoch": 1.150997150997151, "grad_norm": 0.20927824341881443, "learning_rate": 3.4318901795142554e-05, "loss": 0.3744, "step": 404 }, { "epoch": 1.1538461538461537, "grad_norm": 0.23470612919324702, "learning_rate": 3.42661034846885e-05, "loss": 0.3546, "step": 405 }, { "epoch": 1.1566951566951567, "grad_norm": 0.20792179111583126, "learning_rate": 3.421330517423443e-05, "loss": 0.3711, "step": 406 }, { "epoch": 1.1595441595441596, "grad_norm": 0.2426450060857107, "learning_rate": 3.416050686378036e-05, "loss": 0.3446, "step": 407 }, { "epoch": 1.1623931623931625, "grad_norm": 0.2583692183849906, "learning_rate": 3.410770855332629e-05, "loss": 0.3725, "step": 408 }, { "epoch": 1.1652421652421652, "grad_norm": 0.24566054945662688, "learning_rate": 3.405491024287223e-05, "loss": 0.3802, "step": 409 }, { "epoch": 1.168091168091168, "grad_norm": 0.26319005374456306, "learning_rate": 3.4002111932418166e-05, "loss": 0.3579, "step": 410 }, { "epoch": 1.170940170940171, "grad_norm": 0.2371042478251275, "learning_rate": 3.3949313621964095e-05, "loss": 0.3713, "step": 411 }, { "epoch": 1.173789173789174, "grad_norm": 0.21150424231515216, "learning_rate": 3.389651531151004e-05, "loss": 0.3617, "step": 412 }, { "epoch": 1.1766381766381766, "grad_norm": 0.2660573468294059, "learning_rate": 3.384371700105597e-05, "loss": 0.3619, "step": 413 }, { "epoch": 1.1794871794871795, "grad_norm": 0.2629889818118525, "learning_rate": 3.3790918690601904e-05, "loss": 0.3729, "step": 414 }, { "epoch": 1.1823361823361824, "grad_norm": 0.21963192309042254, "learning_rate": 3.3738120380147834e-05, "loss": 0.3649, "step": 415 }, { "epoch": 1.1851851851851851, "grad_norm": 0.2676340624788005, "learning_rate": 3.368532206969377e-05, "loss": 0.3557, "step": 416 }, { "epoch": 1.188034188034188, "grad_norm": 0.21995558755670147, "learning_rate": 3.363252375923971e-05, "loss": 0.3645, "step": 417 }, { "epoch": 1.190883190883191, "grad_norm": 0.21570771890834206, "learning_rate": 3.357972544878564e-05, "loss": 0.3576, "step": 418 }, { "epoch": 1.1937321937321936, "grad_norm": 0.26463658898885334, "learning_rate": 3.352692713833158e-05, "loss": 0.3567, "step": 419 }, { "epoch": 1.1965811965811965, "grad_norm": 0.21514849821934176, "learning_rate": 3.347412882787751e-05, "loss": 0.3745, "step": 420 }, { "epoch": 1.1994301994301995, "grad_norm": 0.24477901943499678, "learning_rate": 3.3421330517423445e-05, "loss": 0.3733, "step": 421 }, { "epoch": 1.2022792022792024, "grad_norm": 0.21967390180786867, "learning_rate": 3.3368532206969375e-05, "loss": 0.3848, "step": 422 }, { "epoch": 1.205128205128205, "grad_norm": 0.22210804416610816, "learning_rate": 3.331573389651531e-05, "loss": 0.3502, "step": 423 }, { "epoch": 1.207977207977208, "grad_norm": 0.23341712719478797, "learning_rate": 3.326293558606125e-05, "loss": 0.3788, "step": 424 }, { "epoch": 1.210826210826211, "grad_norm": 0.22474183284157034, "learning_rate": 3.3210137275607184e-05, "loss": 0.3739, "step": 425 }, { "epoch": 1.2136752136752136, "grad_norm": 0.20810663859008569, "learning_rate": 3.315733896515312e-05, "loss": 0.3575, "step": 426 }, { "epoch": 1.2165242165242165, "grad_norm": 0.23332852349670008, "learning_rate": 3.310454065469905e-05, "loss": 0.3431, "step": 427 }, { "epoch": 1.2193732193732194, "grad_norm": 0.2158486368866276, "learning_rate": 3.305174234424499e-05, "loss": 0.3579, "step": 428 }, { "epoch": 1.2222222222222223, "grad_norm": 0.23876477976217, "learning_rate": 3.2998944033790916e-05, "loss": 0.3748, "step": 429 }, { "epoch": 1.225071225071225, "grad_norm": 0.22671754097370178, "learning_rate": 3.294614572333686e-05, "loss": 0.3733, "step": 430 }, { "epoch": 1.227920227920228, "grad_norm": 0.20882908615979884, "learning_rate": 3.289334741288279e-05, "loss": 0.3478, "step": 431 }, { "epoch": 1.2307692307692308, "grad_norm": 0.22921490176990594, "learning_rate": 3.2840549102428725e-05, "loss": 0.3947, "step": 432 }, { "epoch": 1.2336182336182335, "grad_norm": 0.24877033092993325, "learning_rate": 3.278775079197466e-05, "loss": 0.3587, "step": 433 }, { "epoch": 1.2364672364672364, "grad_norm": 0.21983363052940397, "learning_rate": 3.273495248152059e-05, "loss": 0.3621, "step": 434 }, { "epoch": 1.2393162393162394, "grad_norm": 0.23244050435128452, "learning_rate": 3.268215417106653e-05, "loss": 0.3414, "step": 435 }, { "epoch": 1.242165242165242, "grad_norm": 0.3093091073113401, "learning_rate": 3.262935586061246e-05, "loss": 0.3708, "step": 436 }, { "epoch": 1.245014245014245, "grad_norm": 0.21362305859740863, "learning_rate": 3.25765575501584e-05, "loss": 0.3516, "step": 437 }, { "epoch": 1.2478632478632479, "grad_norm": 0.2454599928955601, "learning_rate": 3.252375923970433e-05, "loss": 0.3692, "step": 438 }, { "epoch": 1.2507122507122508, "grad_norm": 0.25257783308579096, "learning_rate": 3.2470960929250267e-05, "loss": 0.3522, "step": 439 }, { "epoch": 1.2535612535612537, "grad_norm": 0.24797928773992586, "learning_rate": 3.2418162618796196e-05, "loss": 0.3681, "step": 440 }, { "epoch": 1.2564102564102564, "grad_norm": 0.2276797048697118, "learning_rate": 3.236536430834213e-05, "loss": 0.3419, "step": 441 }, { "epoch": 1.2592592592592593, "grad_norm": 0.2646126048006256, "learning_rate": 3.231256599788807e-05, "loss": 0.3834, "step": 442 }, { "epoch": 1.2621082621082622, "grad_norm": 0.23971069233365036, "learning_rate": 3.2259767687434005e-05, "loss": 0.3496, "step": 443 }, { "epoch": 1.264957264957265, "grad_norm": 0.21565481165037637, "learning_rate": 3.220696937697994e-05, "loss": 0.3621, "step": 444 }, { "epoch": 1.2678062678062678, "grad_norm": 0.2429214837642576, "learning_rate": 3.215417106652587e-05, "loss": 0.3653, "step": 445 }, { "epoch": 1.2706552706552707, "grad_norm": 0.23114362643819203, "learning_rate": 3.210137275607181e-05, "loss": 0.3563, "step": 446 }, { "epoch": 1.2735042735042734, "grad_norm": 0.22101693642272935, "learning_rate": 3.204857444561774e-05, "loss": 0.3485, "step": 447 }, { "epoch": 1.2763532763532763, "grad_norm": 0.2483813722230642, "learning_rate": 3.1995776135163674e-05, "loss": 0.3679, "step": 448 }, { "epoch": 1.2792022792022792, "grad_norm": 0.21814879797928116, "learning_rate": 3.194297782470961e-05, "loss": 0.3699, "step": 449 }, { "epoch": 1.282051282051282, "grad_norm": 0.22150552594897266, "learning_rate": 3.1890179514255547e-05, "loss": 0.3595, "step": 450 }, { "epoch": 1.2849002849002849, "grad_norm": 0.22366186353090892, "learning_rate": 3.183738120380148e-05, "loss": 0.3599, "step": 451 }, { "epoch": 1.2877492877492878, "grad_norm": 0.23116335927041887, "learning_rate": 3.178458289334741e-05, "loss": 0.3676, "step": 452 }, { "epoch": 1.2905982905982907, "grad_norm": 0.2033308734127439, "learning_rate": 3.173178458289335e-05, "loss": 0.3521, "step": 453 }, { "epoch": 1.2934472934472934, "grad_norm": 0.2165082929329912, "learning_rate": 3.167898627243928e-05, "loss": 0.3542, "step": 454 }, { "epoch": 1.2962962962962963, "grad_norm": 0.19988858957987832, "learning_rate": 3.1626187961985215e-05, "loss": 0.3498, "step": 455 }, { "epoch": 1.2991452991452992, "grad_norm": 0.2292675192834246, "learning_rate": 3.157338965153115e-05, "loss": 0.3739, "step": 456 }, { "epoch": 1.301994301994302, "grad_norm": 0.2110292978902313, "learning_rate": 3.152059134107709e-05, "loss": 0.3585, "step": 457 }, { "epoch": 1.3048433048433048, "grad_norm": 0.22404470257022518, "learning_rate": 3.1467793030623024e-05, "loss": 0.3448, "step": 458 }, { "epoch": 1.3076923076923077, "grad_norm": 0.244584760285099, "learning_rate": 3.1414994720168954e-05, "loss": 0.371, "step": 459 }, { "epoch": 1.3105413105413106, "grad_norm": 0.2014664128466252, "learning_rate": 3.136219640971489e-05, "loss": 0.3568, "step": 460 }, { "epoch": 1.3133903133903133, "grad_norm": 0.23086767835485536, "learning_rate": 3.130939809926082e-05, "loss": 0.359, "step": 461 }, { "epoch": 1.3162393162393162, "grad_norm": 0.23970941434834442, "learning_rate": 3.125659978880676e-05, "loss": 0.3865, "step": 462 }, { "epoch": 1.3190883190883191, "grad_norm": 0.22027324796314413, "learning_rate": 3.120380147835269e-05, "loss": 0.3814, "step": 463 }, { "epoch": 1.3219373219373218, "grad_norm": 0.1997543044951977, "learning_rate": 3.115100316789863e-05, "loss": 0.3612, "step": 464 }, { "epoch": 1.3247863247863247, "grad_norm": 0.2298484065968129, "learning_rate": 3.1098204857444565e-05, "loss": 0.3486, "step": 465 }, { "epoch": 1.3276353276353277, "grad_norm": 0.23459882831026682, "learning_rate": 3.1045406546990495e-05, "loss": 0.3531, "step": 466 }, { "epoch": 1.3304843304843303, "grad_norm": 0.19872573469259944, "learning_rate": 3.099260823653643e-05, "loss": 0.3572, "step": 467 }, { "epoch": 1.3333333333333333, "grad_norm": 0.22072697065816743, "learning_rate": 3.093980992608237e-05, "loss": 0.3722, "step": 468 }, { "epoch": 1.3361823361823362, "grad_norm": 0.24096874308813088, "learning_rate": 3.0887011615628304e-05, "loss": 0.3596, "step": 469 }, { "epoch": 1.339031339031339, "grad_norm": 0.19914935516155746, "learning_rate": 3.0834213305174234e-05, "loss": 0.3515, "step": 470 }, { "epoch": 1.341880341880342, "grad_norm": 0.2064034362417422, "learning_rate": 3.078141499472017e-05, "loss": 0.3683, "step": 471 }, { "epoch": 1.3447293447293447, "grad_norm": 0.21371146142366845, "learning_rate": 3.0728616684266106e-05, "loss": 0.3517, "step": 472 }, { "epoch": 1.3475783475783476, "grad_norm": 0.21765447741613048, "learning_rate": 3.0675818373812036e-05, "loss": 0.3602, "step": 473 }, { "epoch": 1.3504273504273505, "grad_norm": 0.20977178211422376, "learning_rate": 3.062302006335798e-05, "loss": 0.356, "step": 474 }, { "epoch": 1.3532763532763532, "grad_norm": 0.2204679501825524, "learning_rate": 3.057022175290391e-05, "loss": 0.3638, "step": 475 }, { "epoch": 1.3561253561253561, "grad_norm": 0.2351375455226534, "learning_rate": 3.0517423442449845e-05, "loss": 0.3771, "step": 476 }, { "epoch": 1.358974358974359, "grad_norm": 0.32620489900955313, "learning_rate": 3.0464625131995778e-05, "loss": 0.3753, "step": 477 }, { "epoch": 1.3618233618233617, "grad_norm": 0.222755197443211, "learning_rate": 3.041182682154171e-05, "loss": 0.3674, "step": 478 }, { "epoch": 1.3646723646723646, "grad_norm": 0.22289274030678055, "learning_rate": 3.0359028511087644e-05, "loss": 0.3573, "step": 479 }, { "epoch": 1.3675213675213675, "grad_norm": 0.23379350169681865, "learning_rate": 3.0306230200633577e-05, "loss": 0.3653, "step": 480 }, { "epoch": 1.3703703703703702, "grad_norm": 0.18992629897318228, "learning_rate": 3.0253431890179517e-05, "loss": 0.3364, "step": 481 }, { "epoch": 1.3732193732193732, "grad_norm": 0.21790569721651318, "learning_rate": 3.020063357972545e-05, "loss": 0.3583, "step": 482 }, { "epoch": 1.376068376068376, "grad_norm": 0.21174281018764105, "learning_rate": 3.0147835269271386e-05, "loss": 0.3525, "step": 483 }, { "epoch": 1.378917378917379, "grad_norm": 0.23195477568030867, "learning_rate": 3.009503695881732e-05, "loss": 0.3603, "step": 484 }, { "epoch": 1.381766381766382, "grad_norm": 0.2185865276436869, "learning_rate": 3.0042238648363252e-05, "loss": 0.3751, "step": 485 }, { "epoch": 1.3846153846153846, "grad_norm": 0.21213207655408392, "learning_rate": 2.9989440337909185e-05, "loss": 0.3695, "step": 486 }, { "epoch": 1.3874643874643875, "grad_norm": 0.21005915264245484, "learning_rate": 2.9936642027455125e-05, "loss": 0.356, "step": 487 }, { "epoch": 1.3903133903133904, "grad_norm": 0.21163776861595562, "learning_rate": 2.9883843717001058e-05, "loss": 0.3709, "step": 488 }, { "epoch": 1.393162393162393, "grad_norm": 0.20028548407646818, "learning_rate": 2.983104540654699e-05, "loss": 0.3677, "step": 489 }, { "epoch": 1.396011396011396, "grad_norm": 0.18880981649360443, "learning_rate": 2.9778247096092927e-05, "loss": 0.349, "step": 490 }, { "epoch": 1.398860398860399, "grad_norm": 0.2234668654648058, "learning_rate": 2.972544878563886e-05, "loss": 0.3663, "step": 491 }, { "epoch": 1.4017094017094016, "grad_norm": 0.21853579314131166, "learning_rate": 2.9672650475184793e-05, "loss": 0.3613, "step": 492 }, { "epoch": 1.4045584045584045, "grad_norm": 0.22246500288700016, "learning_rate": 2.9619852164730733e-05, "loss": 0.3803, "step": 493 }, { "epoch": 1.4074074074074074, "grad_norm": 0.20429826743140633, "learning_rate": 2.9567053854276666e-05, "loss": 0.3658, "step": 494 }, { "epoch": 1.4102564102564101, "grad_norm": 0.2539252714453901, "learning_rate": 2.95142555438226e-05, "loss": 0.3734, "step": 495 }, { "epoch": 1.413105413105413, "grad_norm": 0.2176197906573368, "learning_rate": 2.9461457233368532e-05, "loss": 0.3697, "step": 496 }, { "epoch": 1.415954415954416, "grad_norm": 0.20262652133163678, "learning_rate": 2.940865892291447e-05, "loss": 0.3552, "step": 497 }, { "epoch": 1.4188034188034189, "grad_norm": 0.2307771718731254, "learning_rate": 2.93558606124604e-05, "loss": 0.3483, "step": 498 }, { "epoch": 1.4216524216524218, "grad_norm": 0.19342225282003825, "learning_rate": 2.930306230200634e-05, "loss": 0.3689, "step": 499 }, { "epoch": 1.4245014245014245, "grad_norm": 0.20142188241082173, "learning_rate": 2.9250263991552274e-05, "loss": 0.3515, "step": 500 }, { "epoch": 1.4273504273504274, "grad_norm": 0.21412132531144384, "learning_rate": 2.9197465681098207e-05, "loss": 0.3507, "step": 501 }, { "epoch": 1.4301994301994303, "grad_norm": 0.25094985245186724, "learning_rate": 2.914466737064414e-05, "loss": 0.383, "step": 502 }, { "epoch": 1.433048433048433, "grad_norm": 0.21232864199916715, "learning_rate": 2.9091869060190073e-05, "loss": 0.3764, "step": 503 }, { "epoch": 1.435897435897436, "grad_norm": 0.22049102509543148, "learning_rate": 2.903907074973601e-05, "loss": 0.3595, "step": 504 }, { "epoch": 1.4387464387464388, "grad_norm": 0.21682319020739224, "learning_rate": 2.8986272439281943e-05, "loss": 0.3547, "step": 505 }, { "epoch": 1.4415954415954415, "grad_norm": 0.23709929673053004, "learning_rate": 2.8933474128827883e-05, "loss": 0.3687, "step": 506 }, { "epoch": 1.4444444444444444, "grad_norm": 0.2101951752079666, "learning_rate": 2.8880675818373816e-05, "loss": 0.359, "step": 507 }, { "epoch": 1.4472934472934473, "grad_norm": 0.2474601153654613, "learning_rate": 2.882787750791975e-05, "loss": 0.3705, "step": 508 }, { "epoch": 1.45014245014245, "grad_norm": 0.21254139978772235, "learning_rate": 2.877507919746568e-05, "loss": 0.3564, "step": 509 }, { "epoch": 1.452991452991453, "grad_norm": 0.24070048810217534, "learning_rate": 2.8722280887011615e-05, "loss": 0.3706, "step": 510 }, { "epoch": 1.4558404558404558, "grad_norm": 0.2987871246913325, "learning_rate": 2.8669482576557548e-05, "loss": 0.3687, "step": 511 }, { "epoch": 1.4586894586894588, "grad_norm": 0.2300676722177236, "learning_rate": 2.8616684266103487e-05, "loss": 0.3647, "step": 512 }, { "epoch": 1.4615384615384617, "grad_norm": 0.26581166943990137, "learning_rate": 2.856388595564942e-05, "loss": 0.372, "step": 513 }, { "epoch": 1.4643874643874644, "grad_norm": 0.3655749618941395, "learning_rate": 2.8511087645195357e-05, "loss": 0.3776, "step": 514 }, { "epoch": 1.4672364672364673, "grad_norm": 0.26115893033445464, "learning_rate": 2.845828933474129e-05, "loss": 0.3645, "step": 515 }, { "epoch": 1.4700854700854702, "grad_norm": 0.21588849574942176, "learning_rate": 2.8405491024287223e-05, "loss": 0.3554, "step": 516 }, { "epoch": 1.4729344729344729, "grad_norm": 0.22519123294647936, "learning_rate": 2.8352692713833156e-05, "loss": 0.3378, "step": 517 }, { "epoch": 1.4757834757834758, "grad_norm": 0.24778294747275895, "learning_rate": 2.8299894403379096e-05, "loss": 0.3738, "step": 518 }, { "epoch": 1.4786324786324787, "grad_norm": 0.2265365912202125, "learning_rate": 2.824709609292503e-05, "loss": 0.3691, "step": 519 }, { "epoch": 1.4814814814814814, "grad_norm": 0.22460772397621678, "learning_rate": 2.819429778247096e-05, "loss": 0.3691, "step": 520 }, { "epoch": 1.4843304843304843, "grad_norm": 0.21608141296269037, "learning_rate": 2.8141499472016898e-05, "loss": 0.3467, "step": 521 }, { "epoch": 1.4871794871794872, "grad_norm": 0.2213366031457417, "learning_rate": 2.808870116156283e-05, "loss": 0.3622, "step": 522 }, { "epoch": 1.49002849002849, "grad_norm": 0.21175641658464092, "learning_rate": 2.8035902851108764e-05, "loss": 0.3667, "step": 523 }, { "epoch": 1.4928774928774928, "grad_norm": 0.2543379886177831, "learning_rate": 2.7983104540654697e-05, "loss": 0.3909, "step": 524 }, { "epoch": 1.4957264957264957, "grad_norm": 0.2394839478000981, "learning_rate": 2.7930306230200637e-05, "loss": 0.3694, "step": 525 }, { "epoch": 1.4985754985754987, "grad_norm": 0.2275501360298665, "learning_rate": 2.787750791974657e-05, "loss": 0.3559, "step": 526 }, { "epoch": 1.5014245014245016, "grad_norm": 0.2763856292468043, "learning_rate": 2.7824709609292503e-05, "loss": 0.3508, "step": 527 }, { "epoch": 1.5042735042735043, "grad_norm": 0.23957120121880873, "learning_rate": 2.777191129883844e-05, "loss": 0.3589, "step": 528 }, { "epoch": 1.5071225071225072, "grad_norm": 0.23319500553009573, "learning_rate": 2.7719112988384372e-05, "loss": 0.3703, "step": 529 }, { "epoch": 1.50997150997151, "grad_norm": 0.24921375732341508, "learning_rate": 2.7666314677930305e-05, "loss": 0.359, "step": 530 }, { "epoch": 1.5128205128205128, "grad_norm": 0.21008771333534176, "learning_rate": 2.7613516367476245e-05, "loss": 0.3572, "step": 531 }, { "epoch": 1.5156695156695157, "grad_norm": 0.21839097000403634, "learning_rate": 2.7560718057022178e-05, "loss": 0.3657, "step": 532 }, { "epoch": 1.5185185185185186, "grad_norm": 0.22689335860669002, "learning_rate": 2.750791974656811e-05, "loss": 0.3707, "step": 533 }, { "epoch": 1.5213675213675213, "grad_norm": 0.2454238635949199, "learning_rate": 2.7455121436114044e-05, "loss": 0.3665, "step": 534 }, { "epoch": 1.5242165242165242, "grad_norm": 0.21117182341715873, "learning_rate": 2.740232312565998e-05, "loss": 0.3543, "step": 535 }, { "epoch": 1.5270655270655271, "grad_norm": 0.2167859209575041, "learning_rate": 2.7349524815205913e-05, "loss": 0.3648, "step": 536 }, { "epoch": 1.5299145299145298, "grad_norm": 0.25424621395457675, "learning_rate": 2.7296726504751853e-05, "loss": 0.3783, "step": 537 }, { "epoch": 1.5327635327635327, "grad_norm": 0.21880596935336596, "learning_rate": 2.7243928194297786e-05, "loss": 0.3628, "step": 538 }, { "epoch": 1.5356125356125356, "grad_norm": 0.21692887185646115, "learning_rate": 2.719112988384372e-05, "loss": 0.363, "step": 539 }, { "epoch": 1.5384615384615383, "grad_norm": 0.2575849502562457, "learning_rate": 2.7138331573389652e-05, "loss": 0.369, "step": 540 }, { "epoch": 1.5413105413105415, "grad_norm": 0.23885268332551002, "learning_rate": 2.7085533262935585e-05, "loss": 0.3501, "step": 541 }, { "epoch": 1.5441595441595442, "grad_norm": 0.21686275578065198, "learning_rate": 2.703273495248152e-05, "loss": 0.3601, "step": 542 }, { "epoch": 1.547008547008547, "grad_norm": 0.233651783121331, "learning_rate": 2.6979936642027458e-05, "loss": 0.3629, "step": 543 }, { "epoch": 1.54985754985755, "grad_norm": 0.23142549801608311, "learning_rate": 2.6927138331573394e-05, "loss": 0.3756, "step": 544 }, { "epoch": 1.5527065527065527, "grad_norm": 0.23654232452295837, "learning_rate": 2.6874340021119327e-05, "loss": 0.3836, "step": 545 }, { "epoch": 1.5555555555555556, "grad_norm": 0.23955118714830956, "learning_rate": 2.682154171066526e-05, "loss": 0.3599, "step": 546 }, { "epoch": 1.5584045584045585, "grad_norm": 0.22135467807774808, "learning_rate": 2.6768743400211193e-05, "loss": 0.3747, "step": 547 }, { "epoch": 1.5612535612535612, "grad_norm": 0.21590468574896746, "learning_rate": 2.6715945089757126e-05, "loss": 0.3469, "step": 548 }, { "epoch": 1.564102564102564, "grad_norm": 0.24081611853474838, "learning_rate": 2.6663146779303063e-05, "loss": 0.3448, "step": 549 }, { "epoch": 1.566951566951567, "grad_norm": 0.22050273169428228, "learning_rate": 2.6610348468849e-05, "loss": 0.3665, "step": 550 }, { "epoch": 1.5698005698005697, "grad_norm": 0.2358630867918145, "learning_rate": 2.6557550158394935e-05, "loss": 0.3693, "step": 551 }, { "epoch": 1.5726495726495726, "grad_norm": 0.2256231077467112, "learning_rate": 2.650475184794087e-05, "loss": 0.3575, "step": 552 }, { "epoch": 1.5754985754985755, "grad_norm": 0.2193287720903084, "learning_rate": 2.64519535374868e-05, "loss": 0.3602, "step": 553 }, { "epoch": 1.5783475783475782, "grad_norm": 0.23352787405471498, "learning_rate": 2.6399155227032734e-05, "loss": 0.3697, "step": 554 }, { "epoch": 1.5811965811965814, "grad_norm": 0.2395424874218671, "learning_rate": 2.6346356916578667e-05, "loss": 0.3635, "step": 555 }, { "epoch": 1.584045584045584, "grad_norm": 0.43346665073783924, "learning_rate": 2.6293558606124607e-05, "loss": 0.3691, "step": 556 }, { "epoch": 1.5868945868945867, "grad_norm": 0.23220070707086665, "learning_rate": 2.624076029567054e-05, "loss": 0.3731, "step": 557 }, { "epoch": 1.5897435897435899, "grad_norm": 0.20943796833119865, "learning_rate": 2.6187961985216473e-05, "loss": 0.3698, "step": 558 }, { "epoch": 1.5925925925925926, "grad_norm": 0.2257633484876588, "learning_rate": 2.613516367476241e-05, "loss": 0.3635, "step": 559 }, { "epoch": 1.5954415954415955, "grad_norm": 0.22794583966236925, "learning_rate": 2.6082365364308343e-05, "loss": 0.3619, "step": 560 }, { "epoch": 1.5982905982905984, "grad_norm": 0.21844307238312177, "learning_rate": 2.6029567053854276e-05, "loss": 0.3735, "step": 561 }, { "epoch": 1.601139601139601, "grad_norm": 0.25182987504775967, "learning_rate": 2.5976768743400215e-05, "loss": 0.3657, "step": 562 }, { "epoch": 1.603988603988604, "grad_norm": 0.2328339817261743, "learning_rate": 2.5923970432946148e-05, "loss": 0.3617, "step": 563 }, { "epoch": 1.606837606837607, "grad_norm": 0.2198747745202425, "learning_rate": 2.587117212249208e-05, "loss": 0.3567, "step": 564 }, { "epoch": 1.6096866096866096, "grad_norm": 0.22458365808076017, "learning_rate": 2.5818373812038014e-05, "loss": 0.3633, "step": 565 }, { "epoch": 1.6125356125356125, "grad_norm": 0.2050810486479959, "learning_rate": 2.576557550158395e-05, "loss": 0.3662, "step": 566 }, { "epoch": 1.6153846153846154, "grad_norm": 0.23109142394514665, "learning_rate": 2.5712777191129884e-05, "loss": 0.377, "step": 567 }, { "epoch": 1.618233618233618, "grad_norm": 0.20488685566526976, "learning_rate": 2.5659978880675823e-05, "loss": 0.365, "step": 568 }, { "epoch": 1.6210826210826212, "grad_norm": 0.22434853437066676, "learning_rate": 2.5607180570221756e-05, "loss": 0.359, "step": 569 }, { "epoch": 1.623931623931624, "grad_norm": 0.22035628663691662, "learning_rate": 2.555438225976769e-05, "loss": 0.3613, "step": 570 }, { "epoch": 1.6267806267806266, "grad_norm": 0.2296131512377323, "learning_rate": 2.5501583949313622e-05, "loss": 0.3583, "step": 571 }, { "epoch": 1.6296296296296298, "grad_norm": 0.22699499325399286, "learning_rate": 2.5448785638859555e-05, "loss": 0.3615, "step": 572 }, { "epoch": 1.6324786324786325, "grad_norm": 0.23137613302922347, "learning_rate": 2.5395987328405492e-05, "loss": 0.3493, "step": 573 }, { "epoch": 1.6353276353276354, "grad_norm": 0.23513034060758498, "learning_rate": 2.5343189017951425e-05, "loss": 0.3536, "step": 574 }, { "epoch": 1.6381766381766383, "grad_norm": 0.20556474209048517, "learning_rate": 2.5290390707497365e-05, "loss": 0.3494, "step": 575 }, { "epoch": 1.641025641025641, "grad_norm": 0.20951535308942912, "learning_rate": 2.5237592397043298e-05, "loss": 0.3608, "step": 576 }, { "epoch": 1.6438746438746439, "grad_norm": 0.21051061589323755, "learning_rate": 2.518479408658923e-05, "loss": 0.343, "step": 577 }, { "epoch": 1.6467236467236468, "grad_norm": 0.20389495741565292, "learning_rate": 2.5131995776135164e-05, "loss": 0.3515, "step": 578 }, { "epoch": 1.6495726495726495, "grad_norm": 0.20666357754178133, "learning_rate": 2.5079197465681097e-05, "loss": 0.3637, "step": 579 }, { "epoch": 1.6524216524216524, "grad_norm": 0.2124678103193664, "learning_rate": 2.5026399155227033e-05, "loss": 0.3519, "step": 580 }, { "epoch": 1.6552706552706553, "grad_norm": 0.21745017494419377, "learning_rate": 2.497360084477297e-05, "loss": 0.3594, "step": 581 }, { "epoch": 1.658119658119658, "grad_norm": 0.1903592780911563, "learning_rate": 2.4920802534318902e-05, "loss": 0.336, "step": 582 }, { "epoch": 1.660968660968661, "grad_norm": 0.23741085468942869, "learning_rate": 2.486800422386484e-05, "loss": 0.3543, "step": 583 }, { "epoch": 1.6638176638176638, "grad_norm": 0.20157828540341033, "learning_rate": 2.4815205913410772e-05, "loss": 0.3342, "step": 584 }, { "epoch": 1.6666666666666665, "grad_norm": 0.2204911302874124, "learning_rate": 2.4762407602956705e-05, "loss": 0.3586, "step": 585 }, { "epoch": 1.6695156695156697, "grad_norm": 0.21109017199979024, "learning_rate": 2.470960929250264e-05, "loss": 0.3646, "step": 586 }, { "epoch": 1.6723646723646723, "grad_norm": 0.20646510827582915, "learning_rate": 2.4656810982048574e-05, "loss": 0.3563, "step": 587 }, { "epoch": 1.6752136752136753, "grad_norm": 0.21584768343148578, "learning_rate": 2.460401267159451e-05, "loss": 0.3535, "step": 588 }, { "epoch": 1.6780626780626782, "grad_norm": 0.23185019886698816, "learning_rate": 2.4551214361140447e-05, "loss": 0.3697, "step": 589 }, { "epoch": 1.6809116809116809, "grad_norm": 0.21268814828277405, "learning_rate": 2.449841605068638e-05, "loss": 0.3551, "step": 590 }, { "epoch": 1.6837606837606838, "grad_norm": 0.21604754434232787, "learning_rate": 2.4445617740232313e-05, "loss": 0.3655, "step": 591 }, { "epoch": 1.6866096866096867, "grad_norm": 0.22278622809811766, "learning_rate": 2.439281942977825e-05, "loss": 0.3538, "step": 592 }, { "epoch": 1.6894586894586894, "grad_norm": 0.2052145744246248, "learning_rate": 2.4340021119324182e-05, "loss": 0.356, "step": 593 }, { "epoch": 1.6923076923076923, "grad_norm": 0.19837577056881597, "learning_rate": 2.4287222808870115e-05, "loss": 0.356, "step": 594 }, { "epoch": 1.6951566951566952, "grad_norm": 0.2605768702541718, "learning_rate": 2.4234424498416052e-05, "loss": 0.376, "step": 595 }, { "epoch": 1.698005698005698, "grad_norm": 0.2150615912164544, "learning_rate": 2.4181626187961988e-05, "loss": 0.3442, "step": 596 }, { "epoch": 1.7008547008547008, "grad_norm": 0.20499211512651233, "learning_rate": 2.412882787750792e-05, "loss": 0.3699, "step": 597 }, { "epoch": 1.7037037037037037, "grad_norm": 0.21382275219015714, "learning_rate": 2.4076029567053857e-05, "loss": 0.3565, "step": 598 }, { "epoch": 1.7065527065527064, "grad_norm": 0.20994705780047032, "learning_rate": 2.402323125659979e-05, "loss": 0.3598, "step": 599 }, { "epoch": 1.7094017094017095, "grad_norm": 0.20135559643209863, "learning_rate": 2.3970432946145723e-05, "loss": 0.3596, "step": 600 }, { "epoch": 1.7122507122507122, "grad_norm": 0.20036899067341168, "learning_rate": 2.391763463569166e-05, "loss": 0.3569, "step": 601 }, { "epoch": 1.7150997150997151, "grad_norm": 0.21002141879928451, "learning_rate": 2.3864836325237593e-05, "loss": 0.3654, "step": 602 }, { "epoch": 1.717948717948718, "grad_norm": 0.22764067995733198, "learning_rate": 2.381203801478353e-05, "loss": 0.3625, "step": 603 }, { "epoch": 1.7207977207977208, "grad_norm": 0.19291409387702949, "learning_rate": 2.3759239704329462e-05, "loss": 0.3712, "step": 604 }, { "epoch": 1.7236467236467237, "grad_norm": 0.19990638382401998, "learning_rate": 2.37064413938754e-05, "loss": 0.3466, "step": 605 }, { "epoch": 1.7264957264957266, "grad_norm": 0.21172503078878085, "learning_rate": 2.365364308342133e-05, "loss": 0.373, "step": 606 }, { "epoch": 1.7293447293447293, "grad_norm": 0.21340989500402005, "learning_rate": 2.3600844772967265e-05, "loss": 0.3607, "step": 607 }, { "epoch": 1.7321937321937322, "grad_norm": 0.21348773962215495, "learning_rate": 2.35480464625132e-05, "loss": 0.3618, "step": 608 }, { "epoch": 1.735042735042735, "grad_norm": 0.21866328753111697, "learning_rate": 2.3495248152059134e-05, "loss": 0.3556, "step": 609 }, { "epoch": 1.7378917378917378, "grad_norm": 0.19763848284771224, "learning_rate": 2.3442449841605067e-05, "loss": 0.3607, "step": 610 }, { "epoch": 1.7407407407407407, "grad_norm": 0.2152332174618033, "learning_rate": 2.3389651531151003e-05, "loss": 0.3704, "step": 611 }, { "epoch": 1.7435897435897436, "grad_norm": 0.19269940169730435, "learning_rate": 2.333685322069694e-05, "loss": 0.3638, "step": 612 }, { "epoch": 1.7464387464387463, "grad_norm": 0.1962642279399981, "learning_rate": 2.3284054910242873e-05, "loss": 0.3482, "step": 613 }, { "epoch": 1.7492877492877494, "grad_norm": 0.2364856796721012, "learning_rate": 2.323125659978881e-05, "loss": 0.347, "step": 614 }, { "epoch": 1.7521367521367521, "grad_norm": 0.22509925838340694, "learning_rate": 2.3178458289334742e-05, "loss": 0.3604, "step": 615 }, { "epoch": 1.7549857549857548, "grad_norm": 0.2488709063431033, "learning_rate": 2.3125659978880675e-05, "loss": 0.3657, "step": 616 }, { "epoch": 1.757834757834758, "grad_norm": 0.2398318070372559, "learning_rate": 2.307286166842661e-05, "loss": 0.3629, "step": 617 }, { "epoch": 1.7606837606837606, "grad_norm": 0.22323781791119499, "learning_rate": 2.3020063357972545e-05, "loss": 0.359, "step": 618 }, { "epoch": 1.7635327635327636, "grad_norm": 0.2022321694186785, "learning_rate": 2.296726504751848e-05, "loss": 0.3604, "step": 619 }, { "epoch": 1.7663817663817665, "grad_norm": 0.21110134786679816, "learning_rate": 2.2914466737064417e-05, "loss": 0.3687, "step": 620 }, { "epoch": 1.7692307692307692, "grad_norm": 0.19625112710435982, "learning_rate": 2.286166842661035e-05, "loss": 0.3599, "step": 621 }, { "epoch": 1.772079772079772, "grad_norm": 0.20619851901285047, "learning_rate": 2.2808870116156283e-05, "loss": 0.3723, "step": 622 }, { "epoch": 1.774928774928775, "grad_norm": 0.20469389194337959, "learning_rate": 2.275607180570222e-05, "loss": 0.3661, "step": 623 }, { "epoch": 1.7777777777777777, "grad_norm": 0.20778072555494564, "learning_rate": 2.2703273495248153e-05, "loss": 0.3585, "step": 624 }, { "epoch": 1.7806267806267806, "grad_norm": 0.19870626339902325, "learning_rate": 2.2650475184794086e-05, "loss": 0.3441, "step": 625 }, { "epoch": 1.7834757834757835, "grad_norm": 0.19114151393977027, "learning_rate": 2.2597676874340022e-05, "loss": 0.3591, "step": 626 }, { "epoch": 1.7863247863247862, "grad_norm": 0.1982172563967693, "learning_rate": 2.254487856388596e-05, "loss": 0.3854, "step": 627 }, { "epoch": 1.7891737891737893, "grad_norm": 0.20351344941450986, "learning_rate": 2.249208025343189e-05, "loss": 0.3822, "step": 628 }, { "epoch": 1.792022792022792, "grad_norm": 0.19819562909909946, "learning_rate": 2.2439281942977825e-05, "loss": 0.3645, "step": 629 }, { "epoch": 1.7948717948717947, "grad_norm": 0.20300969334790384, "learning_rate": 2.238648363252376e-05, "loss": 0.3631, "step": 630 }, { "epoch": 1.7977207977207978, "grad_norm": 0.20862867991682074, "learning_rate": 2.2333685322069694e-05, "loss": 0.3619, "step": 631 }, { "epoch": 1.8005698005698005, "grad_norm": 0.21833179546935874, "learning_rate": 2.2280887011615627e-05, "loss": 0.3608, "step": 632 }, { "epoch": 1.8034188034188035, "grad_norm": 0.6005070390326573, "learning_rate": 2.2228088701161563e-05, "loss": 0.3965, "step": 633 }, { "epoch": 1.8062678062678064, "grad_norm": 0.35943818916514836, "learning_rate": 2.21752903907075e-05, "loss": 0.3594, "step": 634 }, { "epoch": 1.809116809116809, "grad_norm": 0.2159038665602852, "learning_rate": 2.2122492080253433e-05, "loss": 0.3755, "step": 635 }, { "epoch": 1.811965811965812, "grad_norm": 0.20549883442788655, "learning_rate": 2.206969376979937e-05, "loss": 0.3451, "step": 636 }, { "epoch": 1.8148148148148149, "grad_norm": 0.20440135853564803, "learning_rate": 2.2016895459345302e-05, "loss": 0.3497, "step": 637 }, { "epoch": 1.8176638176638176, "grad_norm": 0.25722335101398636, "learning_rate": 2.1964097148891235e-05, "loss": 0.3639, "step": 638 }, { "epoch": 1.8205128205128205, "grad_norm": 0.19953052748781105, "learning_rate": 2.191129883843717e-05, "loss": 0.3481, "step": 639 }, { "epoch": 1.8233618233618234, "grad_norm": 0.2341568009909009, "learning_rate": 2.1858500527983104e-05, "loss": 0.3593, "step": 640 }, { "epoch": 1.826210826210826, "grad_norm": 0.21325737340547804, "learning_rate": 2.180570221752904e-05, "loss": 0.3636, "step": 641 }, { "epoch": 1.8290598290598292, "grad_norm": 0.22284714887271365, "learning_rate": 2.1752903907074977e-05, "loss": 0.3797, "step": 642 }, { "epoch": 1.831908831908832, "grad_norm": 0.20422527863661058, "learning_rate": 2.170010559662091e-05, "loss": 0.3527, "step": 643 }, { "epoch": 1.8347578347578346, "grad_norm": 0.22652999934815843, "learning_rate": 2.1647307286166843e-05, "loss": 0.3717, "step": 644 }, { "epoch": 1.8376068376068377, "grad_norm": 0.24779116320854475, "learning_rate": 2.159450897571278e-05, "loss": 0.3706, "step": 645 }, { "epoch": 1.8404558404558404, "grad_norm": 0.21671225191720797, "learning_rate": 2.1541710665258713e-05, "loss": 0.3386, "step": 646 }, { "epoch": 1.8433048433048433, "grad_norm": 0.22555743360580618, "learning_rate": 2.1488912354804646e-05, "loss": 0.3804, "step": 647 }, { "epoch": 1.8461538461538463, "grad_norm": 0.23617998499869738, "learning_rate": 2.1436114044350582e-05, "loss": 0.3723, "step": 648 }, { "epoch": 1.849002849002849, "grad_norm": 0.23014596171154433, "learning_rate": 2.138331573389652e-05, "loss": 0.3526, "step": 649 }, { "epoch": 1.8518518518518519, "grad_norm": 0.22916111003439332, "learning_rate": 2.133051742344245e-05, "loss": 0.3451, "step": 650 }, { "epoch": 1.8547008547008548, "grad_norm": 0.19810306429063848, "learning_rate": 2.1277719112988384e-05, "loss": 0.349, "step": 651 }, { "epoch": 1.8575498575498575, "grad_norm": 0.2202335756589606, "learning_rate": 2.122492080253432e-05, "loss": 0.3653, "step": 652 }, { "epoch": 1.8603988603988604, "grad_norm": 0.19827409095956613, "learning_rate": 2.1172122492080254e-05, "loss": 0.3457, "step": 653 }, { "epoch": 1.8632478632478633, "grad_norm": 0.19935795905418485, "learning_rate": 2.1119324181626187e-05, "loss": 0.3554, "step": 654 }, { "epoch": 1.866096866096866, "grad_norm": 0.20847219512744983, "learning_rate": 2.1066525871172123e-05, "loss": 0.351, "step": 655 }, { "epoch": 1.868945868945869, "grad_norm": 0.2156361373347463, "learning_rate": 2.1013727560718056e-05, "loss": 0.3618, "step": 656 }, { "epoch": 1.8717948717948718, "grad_norm": 0.20100877639931164, "learning_rate": 2.0960929250263993e-05, "loss": 0.3527, "step": 657 }, { "epoch": 1.8746438746438745, "grad_norm": 0.21961338320882828, "learning_rate": 2.090813093980993e-05, "loss": 0.3655, "step": 658 }, { "epoch": 1.8774928774928776, "grad_norm": 0.22970963785057988, "learning_rate": 2.0855332629355862e-05, "loss": 0.3578, "step": 659 }, { "epoch": 1.8803418803418803, "grad_norm": 0.1985917394092499, "learning_rate": 2.0802534318901795e-05, "loss": 0.3555, "step": 660 }, { "epoch": 1.8831908831908832, "grad_norm": 0.19782225715456844, "learning_rate": 2.074973600844773e-05, "loss": 0.3544, "step": 661 }, { "epoch": 1.8860398860398861, "grad_norm": 0.24471883420488305, "learning_rate": 2.0696937697993664e-05, "loss": 0.3689, "step": 662 }, { "epoch": 1.8888888888888888, "grad_norm": 0.22271236910553516, "learning_rate": 2.0644139387539597e-05, "loss": 0.3871, "step": 663 }, { "epoch": 1.8917378917378918, "grad_norm": 0.19560513729734122, "learning_rate": 2.0591341077085534e-05, "loss": 0.3561, "step": 664 }, { "epoch": 1.8945868945868947, "grad_norm": 0.21617290509605866, "learning_rate": 2.053854276663147e-05, "loss": 0.3286, "step": 665 }, { "epoch": 1.8974358974358974, "grad_norm": 0.22798617788173128, "learning_rate": 2.0485744456177403e-05, "loss": 0.3688, "step": 666 }, { "epoch": 1.9002849002849003, "grad_norm": 0.21021348614356922, "learning_rate": 2.043294614572334e-05, "loss": 0.3562, "step": 667 }, { "epoch": 1.9031339031339032, "grad_norm": 0.20996089592223216, "learning_rate": 2.0380147835269273e-05, "loss": 0.361, "step": 668 }, { "epoch": 1.9059829059829059, "grad_norm": 0.23565900871607817, "learning_rate": 2.0327349524815206e-05, "loss": 0.3754, "step": 669 }, { "epoch": 1.9088319088319088, "grad_norm": 0.18858686585228232, "learning_rate": 2.0274551214361142e-05, "loss": 0.3516, "step": 670 }, { "epoch": 1.9116809116809117, "grad_norm": 0.1893165779352585, "learning_rate": 2.0221752903907075e-05, "loss": 0.3366, "step": 671 }, { "epoch": 1.9145299145299144, "grad_norm": 0.21668943186182438, "learning_rate": 2.016895459345301e-05, "loss": 0.3502, "step": 672 }, { "epoch": 1.9173789173789175, "grad_norm": 0.20149473922057556, "learning_rate": 2.0116156282998944e-05, "loss": 0.3615, "step": 673 }, { "epoch": 1.9202279202279202, "grad_norm": 0.25869865465945596, "learning_rate": 2.006335797254488e-05, "loss": 0.3691, "step": 674 }, { "epoch": 1.9230769230769231, "grad_norm": 0.2036066715728498, "learning_rate": 2.0010559662090814e-05, "loss": 0.3502, "step": 675 }, { "epoch": 1.925925925925926, "grad_norm": 0.2004788360097017, "learning_rate": 1.9957761351636747e-05, "loss": 0.3594, "step": 676 }, { "epoch": 1.9287749287749287, "grad_norm": 0.21360843600832238, "learning_rate": 1.9904963041182683e-05, "loss": 0.3585, "step": 677 }, { "epoch": 1.9316239316239316, "grad_norm": 0.19762168727225163, "learning_rate": 1.9852164730728616e-05, "loss": 0.3675, "step": 678 }, { "epoch": 1.9344729344729346, "grad_norm": 0.20229203944486152, "learning_rate": 1.9799366420274552e-05, "loss": 0.355, "step": 679 }, { "epoch": 1.9373219373219372, "grad_norm": 0.2022552654861909, "learning_rate": 1.974656810982049e-05, "loss": 0.3547, "step": 680 }, { "epoch": 1.9401709401709402, "grad_norm": 0.2217881214201902, "learning_rate": 1.9693769799366422e-05, "loss": 0.3662, "step": 681 }, { "epoch": 1.943019943019943, "grad_norm": 0.21089292714234747, "learning_rate": 1.9640971488912355e-05, "loss": 0.3632, "step": 682 }, { "epoch": 1.9458689458689458, "grad_norm": 0.21183549408551403, "learning_rate": 1.958817317845829e-05, "loss": 0.3685, "step": 683 }, { "epoch": 1.9487179487179487, "grad_norm": 0.22399399219360525, "learning_rate": 1.9535374868004224e-05, "loss": 0.389, "step": 684 }, { "epoch": 1.9515669515669516, "grad_norm": 0.23478634085970435, "learning_rate": 1.9482576557550157e-05, "loss": 0.349, "step": 685 }, { "epoch": 1.9544159544159543, "grad_norm": 0.18663083025931687, "learning_rate": 1.9429778247096094e-05, "loss": 0.3638, "step": 686 }, { "epoch": 1.9572649572649574, "grad_norm": 0.20785309716673614, "learning_rate": 1.937697993664203e-05, "loss": 0.3529, "step": 687 }, { "epoch": 1.96011396011396, "grad_norm": 0.22463804321134564, "learning_rate": 1.9324181626187963e-05, "loss": 0.3594, "step": 688 }, { "epoch": 1.9629629629629628, "grad_norm": 0.1906326731349908, "learning_rate": 1.92713833157339e-05, "loss": 0.3502, "step": 689 }, { "epoch": 1.965811965811966, "grad_norm": 0.2066700222331773, "learning_rate": 1.9218585005279832e-05, "loss": 0.3672, "step": 690 }, { "epoch": 1.9686609686609686, "grad_norm": 0.1998642686060524, "learning_rate": 1.9165786694825765e-05, "loss": 0.364, "step": 691 }, { "epoch": 1.9715099715099715, "grad_norm": 0.1887608555452163, "learning_rate": 1.9112988384371702e-05, "loss": 0.3509, "step": 692 }, { "epoch": 1.9743589743589745, "grad_norm": 0.196768422003315, "learning_rate": 1.9060190073917635e-05, "loss": 0.342, "step": 693 }, { "epoch": 1.9772079772079771, "grad_norm": 0.19163494415125434, "learning_rate": 1.900739176346357e-05, "loss": 0.3461, "step": 694 }, { "epoch": 1.98005698005698, "grad_norm": 0.22754534461400272, "learning_rate": 1.8954593453009508e-05, "loss": 0.3764, "step": 695 }, { "epoch": 1.982905982905983, "grad_norm": 0.23834935791341716, "learning_rate": 1.890179514255544e-05, "loss": 0.3484, "step": 696 }, { "epoch": 1.9857549857549857, "grad_norm": 0.21264794974633236, "learning_rate": 1.8848996832101374e-05, "loss": 0.368, "step": 697 }, { "epoch": 1.9886039886039886, "grad_norm": 0.21453842577290422, "learning_rate": 1.8796198521647307e-05, "loss": 0.3593, "step": 698 }, { "epoch": 1.9914529914529915, "grad_norm": 0.20799231966171178, "learning_rate": 1.8743400211193243e-05, "loss": 0.3523, "step": 699 }, { "epoch": 1.9943019943019942, "grad_norm": 0.2080913590301314, "learning_rate": 1.8690601900739176e-05, "loss": 0.3768, "step": 700 }, { "epoch": 1.9971509971509973, "grad_norm": 0.20083753257614662, "learning_rate": 1.863780359028511e-05, "loss": 0.373, "step": 701 }, { "epoch": 2.0, "grad_norm": 0.25026072457519927, "learning_rate": 1.8585005279831045e-05, "loss": 0.3476, "step": 702 }, { "epoch": 2.0028490028490027, "grad_norm": 0.31749430934031786, "learning_rate": 1.8532206969376982e-05, "loss": 0.2939, "step": 703 }, { "epoch": 2.005698005698006, "grad_norm": 0.24903718426554183, "learning_rate": 1.8479408658922915e-05, "loss": 0.2786, "step": 704 }, { "epoch": 2.0085470085470085, "grad_norm": 0.32095802283737634, "learning_rate": 1.842661034846885e-05, "loss": 0.2838, "step": 705 }, { "epoch": 2.011396011396011, "grad_norm": 0.29328887995626335, "learning_rate": 1.8373812038014784e-05, "loss": 0.2801, "step": 706 }, { "epoch": 2.0142450142450143, "grad_norm": 0.20925973424050445, "learning_rate": 1.8321013727560717e-05, "loss": 0.2631, "step": 707 }, { "epoch": 2.017094017094017, "grad_norm": 0.28435426480239495, "learning_rate": 1.8268215417106653e-05, "loss": 0.2843, "step": 708 }, { "epoch": 2.0199430199430197, "grad_norm": 0.2500899473941078, "learning_rate": 1.8215417106652586e-05, "loss": 0.2874, "step": 709 }, { "epoch": 2.022792022792023, "grad_norm": 0.2642215470553276, "learning_rate": 1.8162618796198523e-05, "loss": 0.2812, "step": 710 }, { "epoch": 2.0256410256410255, "grad_norm": 0.2787623192910837, "learning_rate": 1.810982048574446e-05, "loss": 0.2852, "step": 711 }, { "epoch": 2.0284900284900287, "grad_norm": 0.2532708996249064, "learning_rate": 1.8057022175290392e-05, "loss": 0.2767, "step": 712 }, { "epoch": 2.0313390313390314, "grad_norm": 0.24900158666379366, "learning_rate": 1.8004223864836325e-05, "loss": 0.2737, "step": 713 }, { "epoch": 2.034188034188034, "grad_norm": 0.2635713749224828, "learning_rate": 1.795142555438226e-05, "loss": 0.2705, "step": 714 }, { "epoch": 2.037037037037037, "grad_norm": 0.20086792181553093, "learning_rate": 1.7898627243928195e-05, "loss": 0.2798, "step": 715 }, { "epoch": 2.03988603988604, "grad_norm": 0.2432885568665745, "learning_rate": 1.7845828933474128e-05, "loss": 0.281, "step": 716 }, { "epoch": 2.0427350427350426, "grad_norm": 0.21507703672726577, "learning_rate": 1.7793030623020064e-05, "loss": 0.2762, "step": 717 }, { "epoch": 2.0455840455840457, "grad_norm": 0.21933858359248162, "learning_rate": 1.7740232312566e-05, "loss": 0.2729, "step": 718 }, { "epoch": 2.0484330484330484, "grad_norm": 0.22017588537800314, "learning_rate": 1.7687434002111933e-05, "loss": 0.2855, "step": 719 }, { "epoch": 2.051282051282051, "grad_norm": 0.20337914754232117, "learning_rate": 1.7634635691657866e-05, "loss": 0.2715, "step": 720 }, { "epoch": 2.0541310541310542, "grad_norm": 0.19719485220504646, "learning_rate": 1.7581837381203803e-05, "loss": 0.2611, "step": 721 }, { "epoch": 2.056980056980057, "grad_norm": 0.2031620606156548, "learning_rate": 1.7529039070749736e-05, "loss": 0.2853, "step": 722 }, { "epoch": 2.0598290598290596, "grad_norm": 0.22411987580889667, "learning_rate": 1.747624076029567e-05, "loss": 0.2807, "step": 723 }, { "epoch": 2.0626780626780628, "grad_norm": 0.1997733419374099, "learning_rate": 1.7423442449841605e-05, "loss": 0.2749, "step": 724 }, { "epoch": 2.0655270655270654, "grad_norm": 0.20442777603565418, "learning_rate": 1.737064413938754e-05, "loss": 0.2709, "step": 725 }, { "epoch": 2.0683760683760686, "grad_norm": 0.2053651992461902, "learning_rate": 1.7317845828933475e-05, "loss": 0.2739, "step": 726 }, { "epoch": 2.0712250712250713, "grad_norm": 0.2041941282028662, "learning_rate": 1.726504751847941e-05, "loss": 0.2787, "step": 727 }, { "epoch": 2.074074074074074, "grad_norm": 0.2289119309925019, "learning_rate": 1.7212249208025344e-05, "loss": 0.276, "step": 728 }, { "epoch": 2.076923076923077, "grad_norm": 0.20075289027919085, "learning_rate": 1.7159450897571277e-05, "loss": 0.2754, "step": 729 }, { "epoch": 2.07977207977208, "grad_norm": 0.1910485513723653, "learning_rate": 1.7106652587117213e-05, "loss": 0.274, "step": 730 }, { "epoch": 2.0826210826210825, "grad_norm": 0.2075379358650543, "learning_rate": 1.7053854276663146e-05, "loss": 0.2724, "step": 731 }, { "epoch": 2.0854700854700856, "grad_norm": 0.21528112160170912, "learning_rate": 1.7001055966209083e-05, "loss": 0.2726, "step": 732 }, { "epoch": 2.0883190883190883, "grad_norm": 0.2110995481791853, "learning_rate": 1.694825765575502e-05, "loss": 0.2772, "step": 733 }, { "epoch": 2.091168091168091, "grad_norm": 0.1829351870654865, "learning_rate": 1.6895459345300952e-05, "loss": 0.2704, "step": 734 }, { "epoch": 2.094017094017094, "grad_norm": 0.19540194440796962, "learning_rate": 1.6842661034846885e-05, "loss": 0.2658, "step": 735 }, { "epoch": 2.096866096866097, "grad_norm": 0.21657896164501325, "learning_rate": 1.678986272439282e-05, "loss": 0.2746, "step": 736 }, { "epoch": 2.0997150997150995, "grad_norm": 0.19563365784205655, "learning_rate": 1.6737064413938755e-05, "loss": 0.2652, "step": 737 }, { "epoch": 2.1025641025641026, "grad_norm": 0.19212356110155093, "learning_rate": 1.6684266103484688e-05, "loss": 0.263, "step": 738 }, { "epoch": 2.1054131054131053, "grad_norm": 0.20460727383155605, "learning_rate": 1.6631467793030624e-05, "loss": 0.2742, "step": 739 }, { "epoch": 2.1082621082621085, "grad_norm": 0.21427650674256282, "learning_rate": 1.657866948257656e-05, "loss": 0.2825, "step": 740 }, { "epoch": 2.111111111111111, "grad_norm": 0.20672190281174416, "learning_rate": 1.6525871172122493e-05, "loss": 0.2839, "step": 741 }, { "epoch": 2.113960113960114, "grad_norm": 0.20211237414090671, "learning_rate": 1.647307286166843e-05, "loss": 0.2826, "step": 742 }, { "epoch": 2.116809116809117, "grad_norm": 0.19195099076895822, "learning_rate": 1.6420274551214363e-05, "loss": 0.2759, "step": 743 }, { "epoch": 2.1196581196581197, "grad_norm": 0.20560314548089362, "learning_rate": 1.6367476240760296e-05, "loss": 0.2681, "step": 744 }, { "epoch": 2.1225071225071224, "grad_norm": 0.21253013838923676, "learning_rate": 1.631467793030623e-05, "loss": 0.2691, "step": 745 }, { "epoch": 2.1253561253561255, "grad_norm": 0.19213504270638337, "learning_rate": 1.6261879619852165e-05, "loss": 0.2709, "step": 746 }, { "epoch": 2.128205128205128, "grad_norm": 0.19810360169754043, "learning_rate": 1.6209081309398098e-05, "loss": 0.2681, "step": 747 }, { "epoch": 2.131054131054131, "grad_norm": 0.8925954469518941, "learning_rate": 1.6156282998944034e-05, "loss": 0.3358, "step": 748 }, { "epoch": 2.133903133903134, "grad_norm": 0.21118384990812827, "learning_rate": 1.610348468848997e-05, "loss": 0.2834, "step": 749 }, { "epoch": 2.1367521367521367, "grad_norm": 0.22061422429768562, "learning_rate": 1.6050686378035904e-05, "loss": 0.2782, "step": 750 }, { "epoch": 2.1396011396011394, "grad_norm": 0.2124786481646926, "learning_rate": 1.5997888067581837e-05, "loss": 0.2998, "step": 751 }, { "epoch": 2.1424501424501425, "grad_norm": 0.20503289997556293, "learning_rate": 1.5945089757127773e-05, "loss": 0.2807, "step": 752 }, { "epoch": 2.1452991452991452, "grad_norm": 0.22522249962157764, "learning_rate": 1.5892291446673706e-05, "loss": 0.2806, "step": 753 }, { "epoch": 2.148148148148148, "grad_norm": 0.22270532746805854, "learning_rate": 1.583949313621964e-05, "loss": 0.273, "step": 754 }, { "epoch": 2.150997150997151, "grad_norm": 0.19772734044545245, "learning_rate": 1.5786694825765576e-05, "loss": 0.272, "step": 755 }, { "epoch": 2.1538461538461537, "grad_norm": 0.1967943644277297, "learning_rate": 1.5733896515311512e-05, "loss": 0.2753, "step": 756 }, { "epoch": 2.156695156695157, "grad_norm": 0.19154676700453233, "learning_rate": 1.5681098204857445e-05, "loss": 0.2709, "step": 757 }, { "epoch": 2.1595441595441596, "grad_norm": 0.2221016196227244, "learning_rate": 1.562829989440338e-05, "loss": 0.2823, "step": 758 }, { "epoch": 2.1623931623931623, "grad_norm": 0.1937702666618624, "learning_rate": 1.5575501583949314e-05, "loss": 0.2645, "step": 759 }, { "epoch": 2.1652421652421654, "grad_norm": 0.1969269538576539, "learning_rate": 1.5522703273495247e-05, "loss": 0.2761, "step": 760 }, { "epoch": 2.168091168091168, "grad_norm": 0.20395502697309204, "learning_rate": 1.5469904963041184e-05, "loss": 0.2792, "step": 761 }, { "epoch": 2.1709401709401708, "grad_norm": 0.20751492020213136, "learning_rate": 1.5417106652587117e-05, "loss": 0.2623, "step": 762 }, { "epoch": 2.173789173789174, "grad_norm": 0.19379484649828319, "learning_rate": 1.5364308342133053e-05, "loss": 0.2606, "step": 763 }, { "epoch": 2.1766381766381766, "grad_norm": 0.18718035792100948, "learning_rate": 1.531151003167899e-05, "loss": 0.2751, "step": 764 }, { "epoch": 2.1794871794871793, "grad_norm": 0.19066401480212206, "learning_rate": 1.5258711721224923e-05, "loss": 0.2732, "step": 765 }, { "epoch": 2.1823361823361824, "grad_norm": 0.20925353974218774, "learning_rate": 1.5205913410770856e-05, "loss": 0.2881, "step": 766 }, { "epoch": 2.185185185185185, "grad_norm": 0.20547467764539895, "learning_rate": 1.5153115100316789e-05, "loss": 0.2638, "step": 767 }, { "epoch": 2.1880341880341883, "grad_norm": 0.19527197604782076, "learning_rate": 1.5100316789862725e-05, "loss": 0.2783, "step": 768 }, { "epoch": 2.190883190883191, "grad_norm": 0.1942190592627724, "learning_rate": 1.504751847940866e-05, "loss": 0.2709, "step": 769 }, { "epoch": 2.1937321937321936, "grad_norm": 0.2103546914669498, "learning_rate": 1.4994720168954593e-05, "loss": 0.2703, "step": 770 }, { "epoch": 2.1965811965811968, "grad_norm": 0.20158953401192725, "learning_rate": 1.4941921858500529e-05, "loss": 0.2786, "step": 771 }, { "epoch": 2.1994301994301995, "grad_norm": 0.18874398829424063, "learning_rate": 1.4889123548046464e-05, "loss": 0.2625, "step": 772 }, { "epoch": 2.202279202279202, "grad_norm": 0.1919940823226646, "learning_rate": 1.4836325237592397e-05, "loss": 0.2819, "step": 773 }, { "epoch": 2.2051282051282053, "grad_norm": 0.20270992368788238, "learning_rate": 1.4783526927138333e-05, "loss": 0.2746, "step": 774 }, { "epoch": 2.207977207977208, "grad_norm": 0.19972156431692997, "learning_rate": 1.4730728616684266e-05, "loss": 0.277, "step": 775 }, { "epoch": 2.2108262108262107, "grad_norm": 0.2060428169408646, "learning_rate": 1.46779303062302e-05, "loss": 0.2866, "step": 776 }, { "epoch": 2.213675213675214, "grad_norm": 0.2180940293343127, "learning_rate": 1.4625131995776137e-05, "loss": 0.2664, "step": 777 }, { "epoch": 2.2165242165242165, "grad_norm": 0.19764083611553848, "learning_rate": 1.457233368532207e-05, "loss": 0.2698, "step": 778 }, { "epoch": 2.219373219373219, "grad_norm": 0.18295950495897126, "learning_rate": 1.4519535374868005e-05, "loss": 0.266, "step": 779 }, { "epoch": 2.2222222222222223, "grad_norm": 0.20138778009514838, "learning_rate": 1.4466737064413941e-05, "loss": 0.2732, "step": 780 }, { "epoch": 2.225071225071225, "grad_norm": 0.21789436891588496, "learning_rate": 1.4413938753959874e-05, "loss": 0.2842, "step": 781 }, { "epoch": 2.2279202279202277, "grad_norm": 0.19845024094048266, "learning_rate": 1.4361140443505807e-05, "loss": 0.2617, "step": 782 }, { "epoch": 2.230769230769231, "grad_norm": 0.20136450371041287, "learning_rate": 1.4308342133051744e-05, "loss": 0.2673, "step": 783 }, { "epoch": 2.2336182336182335, "grad_norm": 0.20064270708774173, "learning_rate": 1.4255543822597678e-05, "loss": 0.2635, "step": 784 }, { "epoch": 2.2364672364672367, "grad_norm": 0.20230154595496835, "learning_rate": 1.4202745512143611e-05, "loss": 0.2929, "step": 785 }, { "epoch": 2.2393162393162394, "grad_norm": 0.201820653024005, "learning_rate": 1.4149947201689548e-05, "loss": 0.2834, "step": 786 }, { "epoch": 2.242165242165242, "grad_norm": 0.19121793530935455, "learning_rate": 1.409714889123548e-05, "loss": 0.2625, "step": 787 }, { "epoch": 2.245014245014245, "grad_norm": 0.19330950900844268, "learning_rate": 1.4044350580781415e-05, "loss": 0.2772, "step": 788 }, { "epoch": 2.247863247863248, "grad_norm": 0.19306556715969658, "learning_rate": 1.3991552270327348e-05, "loss": 0.2627, "step": 789 }, { "epoch": 2.2507122507122506, "grad_norm": 0.19901233397623183, "learning_rate": 1.3938753959873285e-05, "loss": 0.2696, "step": 790 }, { "epoch": 2.2535612535612537, "grad_norm": 0.19925310664104037, "learning_rate": 1.388595564941922e-05, "loss": 0.2584, "step": 791 }, { "epoch": 2.2564102564102564, "grad_norm": 0.2034999029639149, "learning_rate": 1.3833157338965153e-05, "loss": 0.2966, "step": 792 }, { "epoch": 2.259259259259259, "grad_norm": 0.19015133424936778, "learning_rate": 1.3780359028511089e-05, "loss": 0.2865, "step": 793 }, { "epoch": 2.262108262108262, "grad_norm": 0.22328712353887642, "learning_rate": 1.3727560718057022e-05, "loss": 0.2772, "step": 794 }, { "epoch": 2.264957264957265, "grad_norm": 0.19798853546327766, "learning_rate": 1.3674762407602957e-05, "loss": 0.2632, "step": 795 }, { "epoch": 2.267806267806268, "grad_norm": 0.18631225104292679, "learning_rate": 1.3621964097148893e-05, "loss": 0.2775, "step": 796 }, { "epoch": 2.2706552706552707, "grad_norm": 0.1909878525568524, "learning_rate": 1.3569165786694826e-05, "loss": 0.2848, "step": 797 }, { "epoch": 2.2735042735042734, "grad_norm": 0.2068674181031461, "learning_rate": 1.351636747624076e-05, "loss": 0.2754, "step": 798 }, { "epoch": 2.2763532763532766, "grad_norm": 0.20247748116558922, "learning_rate": 1.3463569165786697e-05, "loss": 0.2681, "step": 799 }, { "epoch": 2.2792022792022792, "grad_norm": 0.19577250906503782, "learning_rate": 1.341077085533263e-05, "loss": 0.2698, "step": 800 }, { "epoch": 2.282051282051282, "grad_norm": 0.192461914460849, "learning_rate": 1.3357972544878563e-05, "loss": 0.2624, "step": 801 }, { "epoch": 2.284900284900285, "grad_norm": 0.21973379890078096, "learning_rate": 1.33051742344245e-05, "loss": 0.2754, "step": 802 }, { "epoch": 2.2877492877492878, "grad_norm": 0.19582967491065492, "learning_rate": 1.3252375923970434e-05, "loss": 0.277, "step": 803 }, { "epoch": 2.2905982905982905, "grad_norm": 0.19580304220995717, "learning_rate": 1.3199577613516367e-05, "loss": 0.2716, "step": 804 }, { "epoch": 2.2934472934472936, "grad_norm": 0.2588055907574961, "learning_rate": 1.3146779303062304e-05, "loss": 0.2772, "step": 805 }, { "epoch": 2.2962962962962963, "grad_norm": 0.1985796106477907, "learning_rate": 1.3093980992608237e-05, "loss": 0.2725, "step": 806 }, { "epoch": 2.299145299145299, "grad_norm": 0.21599405795203994, "learning_rate": 1.3041182682154171e-05, "loss": 0.2924, "step": 807 }, { "epoch": 2.301994301994302, "grad_norm": 0.20158022481963558, "learning_rate": 1.2988384371700108e-05, "loss": 0.2885, "step": 808 }, { "epoch": 2.304843304843305, "grad_norm": 0.18745667148509224, "learning_rate": 1.293558606124604e-05, "loss": 0.2943, "step": 809 }, { "epoch": 2.3076923076923075, "grad_norm": 0.2065772682516858, "learning_rate": 1.2882787750791975e-05, "loss": 0.2778, "step": 810 }, { "epoch": 2.3105413105413106, "grad_norm": 0.2170049193233547, "learning_rate": 1.2829989440337912e-05, "loss": 0.2842, "step": 811 }, { "epoch": 2.3133903133903133, "grad_norm": 0.20199308409143069, "learning_rate": 1.2777191129883845e-05, "loss": 0.2717, "step": 812 }, { "epoch": 2.316239316239316, "grad_norm": 0.20699440291292984, "learning_rate": 1.2724392819429778e-05, "loss": 0.2925, "step": 813 }, { "epoch": 2.319088319088319, "grad_norm": 0.19508346223108827, "learning_rate": 1.2671594508975712e-05, "loss": 0.277, "step": 814 }, { "epoch": 2.321937321937322, "grad_norm": 0.19079264766629775, "learning_rate": 1.2618796198521649e-05, "loss": 0.2778, "step": 815 }, { "epoch": 2.324786324786325, "grad_norm": 0.18964264325936672, "learning_rate": 1.2565997888067582e-05, "loss": 0.2782, "step": 816 }, { "epoch": 2.3276353276353277, "grad_norm": 0.20280215258566556, "learning_rate": 1.2513199577613517e-05, "loss": 0.2893, "step": 817 }, { "epoch": 2.3304843304843303, "grad_norm": 0.2156167038301731, "learning_rate": 1.2460401267159451e-05, "loss": 0.2854, "step": 818 }, { "epoch": 2.3333333333333335, "grad_norm": 0.1964934145520702, "learning_rate": 1.2407602956705386e-05, "loss": 0.2789, "step": 819 }, { "epoch": 2.336182336182336, "grad_norm": 0.20506845724808467, "learning_rate": 1.235480464625132e-05, "loss": 0.2837, "step": 820 }, { "epoch": 2.339031339031339, "grad_norm": 0.20608756180445, "learning_rate": 1.2302006335797255e-05, "loss": 0.2867, "step": 821 }, { "epoch": 2.341880341880342, "grad_norm": 0.22728463316725367, "learning_rate": 1.224920802534319e-05, "loss": 0.2879, "step": 822 }, { "epoch": 2.3447293447293447, "grad_norm": 0.21134582426808612, "learning_rate": 1.2196409714889125e-05, "loss": 0.2706, "step": 823 }, { "epoch": 2.347578347578348, "grad_norm": 0.2117202491888567, "learning_rate": 1.2143611404435058e-05, "loss": 0.2901, "step": 824 }, { "epoch": 2.3504273504273505, "grad_norm": 0.20417464138588973, "learning_rate": 1.2090813093980994e-05, "loss": 0.2838, "step": 825 }, { "epoch": 2.353276353276353, "grad_norm": 0.1996078258358169, "learning_rate": 1.2038014783526929e-05, "loss": 0.2669, "step": 826 }, { "epoch": 2.3561253561253563, "grad_norm": 0.22247317080848442, "learning_rate": 1.1985216473072862e-05, "loss": 0.2855, "step": 827 }, { "epoch": 2.358974358974359, "grad_norm": 0.20677945710130288, "learning_rate": 1.1932418162618796e-05, "loss": 0.2617, "step": 828 }, { "epoch": 2.3618233618233617, "grad_norm": 0.22160179441363054, "learning_rate": 1.1879619852164731e-05, "loss": 0.2756, "step": 829 }, { "epoch": 2.364672364672365, "grad_norm": 0.22587987413436983, "learning_rate": 1.1826821541710666e-05, "loss": 0.2817, "step": 830 }, { "epoch": 2.3675213675213675, "grad_norm": 0.20154794971028572, "learning_rate": 1.17740232312566e-05, "loss": 0.2651, "step": 831 }, { "epoch": 2.3703703703703702, "grad_norm": 0.20403481294874842, "learning_rate": 1.1721224920802534e-05, "loss": 0.2775, "step": 832 }, { "epoch": 2.3732193732193734, "grad_norm": 0.22073919271992984, "learning_rate": 1.166842661034847e-05, "loss": 0.2778, "step": 833 }, { "epoch": 2.376068376068376, "grad_norm": 0.22085826451951224, "learning_rate": 1.1615628299894405e-05, "loss": 0.2716, "step": 834 }, { "epoch": 2.3789173789173788, "grad_norm": 0.20286442068216853, "learning_rate": 1.1562829989440338e-05, "loss": 0.2725, "step": 835 }, { "epoch": 2.381766381766382, "grad_norm": 0.20066512198655337, "learning_rate": 1.1510031678986272e-05, "loss": 0.2791, "step": 836 }, { "epoch": 2.3846153846153846, "grad_norm": 0.2138581117365153, "learning_rate": 1.1457233368532209e-05, "loss": 0.2748, "step": 837 }, { "epoch": 2.3874643874643873, "grad_norm": 0.21527795194888655, "learning_rate": 1.1404435058078142e-05, "loss": 0.2742, "step": 838 }, { "epoch": 2.3903133903133904, "grad_norm": 0.20801036949744942, "learning_rate": 1.1351636747624076e-05, "loss": 0.2797, "step": 839 }, { "epoch": 2.393162393162393, "grad_norm": 0.20792622413478573, "learning_rate": 1.1298838437170011e-05, "loss": 0.282, "step": 840 }, { "epoch": 2.396011396011396, "grad_norm": 0.22033246310576216, "learning_rate": 1.1246040126715946e-05, "loss": 0.279, "step": 841 }, { "epoch": 2.398860398860399, "grad_norm": 0.1967422885160002, "learning_rate": 1.119324181626188e-05, "loss": 0.2811, "step": 842 }, { "epoch": 2.4017094017094016, "grad_norm": 0.1814858766804614, "learning_rate": 1.1140443505807813e-05, "loss": 0.2608, "step": 843 }, { "epoch": 2.4045584045584047, "grad_norm": 0.19499452238873044, "learning_rate": 1.108764519535375e-05, "loss": 0.2678, "step": 844 }, { "epoch": 2.4074074074074074, "grad_norm": 0.2102124248647648, "learning_rate": 1.1034846884899685e-05, "loss": 0.2818, "step": 845 }, { "epoch": 2.41025641025641, "grad_norm": 0.19814574523643397, "learning_rate": 1.0982048574445618e-05, "loss": 0.2653, "step": 846 }, { "epoch": 2.4131054131054133, "grad_norm": 0.18054162144741087, "learning_rate": 1.0929250263991552e-05, "loss": 0.2695, "step": 847 }, { "epoch": 2.415954415954416, "grad_norm": 0.1808318357229325, "learning_rate": 1.0876451953537489e-05, "loss": 0.2801, "step": 848 }, { "epoch": 2.4188034188034186, "grad_norm": 0.19987403185338987, "learning_rate": 1.0823653643083422e-05, "loss": 0.2721, "step": 849 }, { "epoch": 2.421652421652422, "grad_norm": 0.2248142211042877, "learning_rate": 1.0770855332629356e-05, "loss": 0.2843, "step": 850 }, { "epoch": 2.4245014245014245, "grad_norm": 0.1964174663784782, "learning_rate": 1.0718057022175291e-05, "loss": 0.2834, "step": 851 }, { "epoch": 2.427350427350427, "grad_norm": 0.18315937428538173, "learning_rate": 1.0665258711721226e-05, "loss": 0.267, "step": 852 }, { "epoch": 2.4301994301994303, "grad_norm": 0.1926077433772735, "learning_rate": 1.061246040126716e-05, "loss": 0.2748, "step": 853 }, { "epoch": 2.433048433048433, "grad_norm": 0.20408583222643678, "learning_rate": 1.0559662090813093e-05, "loss": 0.2776, "step": 854 }, { "epoch": 2.435897435897436, "grad_norm": 0.1973007081934127, "learning_rate": 1.0506863780359028e-05, "loss": 0.2748, "step": 855 }, { "epoch": 2.438746438746439, "grad_norm": 0.18467705527546302, "learning_rate": 1.0454065469904964e-05, "loss": 0.2767, "step": 856 }, { "epoch": 2.4415954415954415, "grad_norm": 0.18392804629969942, "learning_rate": 1.0401267159450897e-05, "loss": 0.2803, "step": 857 }, { "epoch": 2.4444444444444446, "grad_norm": 0.19726145099036824, "learning_rate": 1.0348468848996832e-05, "loss": 0.2763, "step": 858 }, { "epoch": 2.4472934472934473, "grad_norm": 0.19595567470607117, "learning_rate": 1.0295670538542767e-05, "loss": 0.288, "step": 859 }, { "epoch": 2.45014245014245, "grad_norm": 0.19049627669313532, "learning_rate": 1.0242872228088702e-05, "loss": 0.2804, "step": 860 }, { "epoch": 2.452991452991453, "grad_norm": 0.19465038042198368, "learning_rate": 1.0190073917634636e-05, "loss": 0.2853, "step": 861 }, { "epoch": 2.455840455840456, "grad_norm": 0.18907018318647903, "learning_rate": 1.0137275607180571e-05, "loss": 0.262, "step": 862 }, { "epoch": 2.4586894586894585, "grad_norm": 0.2117775134978185, "learning_rate": 1.0084477296726506e-05, "loss": 0.2832, "step": 863 }, { "epoch": 2.4615384615384617, "grad_norm": 0.18369057037488082, "learning_rate": 1.003167898627244e-05, "loss": 0.2818, "step": 864 }, { "epoch": 2.4643874643874644, "grad_norm": 0.1938139353246623, "learning_rate": 9.978880675818373e-06, "loss": 0.2687, "step": 865 }, { "epoch": 2.467236467236467, "grad_norm": 0.21926317040236148, "learning_rate": 9.926082365364308e-06, "loss": 0.2762, "step": 866 }, { "epoch": 2.47008547008547, "grad_norm": 0.20036724393686023, "learning_rate": 9.873284054910244e-06, "loss": 0.2788, "step": 867 }, { "epoch": 2.472934472934473, "grad_norm": 0.195455632630333, "learning_rate": 9.820485744456177e-06, "loss": 0.2712, "step": 868 }, { "epoch": 2.4757834757834756, "grad_norm": 0.19688632965075772, "learning_rate": 9.767687434002112e-06, "loss": 0.2928, "step": 869 }, { "epoch": 2.4786324786324787, "grad_norm": 0.196362275701545, "learning_rate": 9.714889123548047e-06, "loss": 0.2856, "step": 870 }, { "epoch": 2.4814814814814814, "grad_norm": 0.19446432785612933, "learning_rate": 9.662090813093982e-06, "loss": 0.279, "step": 871 }, { "epoch": 2.484330484330484, "grad_norm": 0.1879688895780496, "learning_rate": 9.609292502639916e-06, "loss": 0.2854, "step": 872 }, { "epoch": 2.4871794871794872, "grad_norm": 0.19437610153561674, "learning_rate": 9.556494192185851e-06, "loss": 0.2738, "step": 873 }, { "epoch": 2.49002849002849, "grad_norm": 0.20143365503846905, "learning_rate": 9.503695881731786e-06, "loss": 0.2759, "step": 874 }, { "epoch": 2.492877492877493, "grad_norm": 0.1935287651459583, "learning_rate": 9.45089757127772e-06, "loss": 0.2835, "step": 875 }, { "epoch": 2.4957264957264957, "grad_norm": 0.19317522985073163, "learning_rate": 9.398099260823653e-06, "loss": 0.2858, "step": 876 }, { "epoch": 2.4985754985754984, "grad_norm": 0.2012616552901727, "learning_rate": 9.345300950369588e-06, "loss": 0.2686, "step": 877 }, { "epoch": 2.5014245014245016, "grad_norm": 0.18496934505004578, "learning_rate": 9.292502639915523e-06, "loss": 0.2649, "step": 878 }, { "epoch": 2.5042735042735043, "grad_norm": 0.1952709023463477, "learning_rate": 9.239704329461457e-06, "loss": 0.2765, "step": 879 }, { "epoch": 2.5071225071225074, "grad_norm": 0.1879189333509987, "learning_rate": 9.186906019007392e-06, "loss": 0.2832, "step": 880 }, { "epoch": 2.50997150997151, "grad_norm": 0.18950447218698951, "learning_rate": 9.134107708553327e-06, "loss": 0.2843, "step": 881 }, { "epoch": 2.5128205128205128, "grad_norm": 0.18653191496871754, "learning_rate": 9.081309398099261e-06, "loss": 0.277, "step": 882 }, { "epoch": 2.515669515669516, "grad_norm": 0.17645739186127074, "learning_rate": 9.028511087645196e-06, "loss": 0.2686, "step": 883 }, { "epoch": 2.5185185185185186, "grad_norm": 0.18764217708119302, "learning_rate": 8.97571277719113e-06, "loss": 0.284, "step": 884 }, { "epoch": 2.5213675213675213, "grad_norm": 0.19467937140297095, "learning_rate": 8.922914466737064e-06, "loss": 0.2717, "step": 885 }, { "epoch": 2.5242165242165244, "grad_norm": 0.19792747219286208, "learning_rate": 8.870116156283e-06, "loss": 0.2855, "step": 886 }, { "epoch": 2.527065527065527, "grad_norm": 0.18656641235510737, "learning_rate": 8.817317845828933e-06, "loss": 0.274, "step": 887 }, { "epoch": 2.52991452991453, "grad_norm": 0.19826113067880127, "learning_rate": 8.764519535374868e-06, "loss": 0.2973, "step": 888 }, { "epoch": 2.532763532763533, "grad_norm": 0.19491004642921803, "learning_rate": 8.711721224920803e-06, "loss": 0.2755, "step": 889 }, { "epoch": 2.5356125356125356, "grad_norm": 0.2012953347472746, "learning_rate": 8.658922914466737e-06, "loss": 0.2836, "step": 890 }, { "epoch": 2.5384615384615383, "grad_norm": 0.19357544104347177, "learning_rate": 8.606124604012672e-06, "loss": 0.2671, "step": 891 }, { "epoch": 2.5413105413105415, "grad_norm": 0.18892143464854252, "learning_rate": 8.553326293558607e-06, "loss": 0.2733, "step": 892 }, { "epoch": 2.544159544159544, "grad_norm": 0.6265508514551001, "learning_rate": 8.500527983104541e-06, "loss": 0.3024, "step": 893 }, { "epoch": 2.547008547008547, "grad_norm": 0.180843319457107, "learning_rate": 8.447729672650476e-06, "loss": 0.2758, "step": 894 }, { "epoch": 2.54985754985755, "grad_norm": 0.18824519820336835, "learning_rate": 8.39493136219641e-06, "loss": 0.2777, "step": 895 }, { "epoch": 2.5527065527065527, "grad_norm": 0.18679719864151892, "learning_rate": 8.342133051742344e-06, "loss": 0.273, "step": 896 }, { "epoch": 2.5555555555555554, "grad_norm": 0.18954916069648725, "learning_rate": 8.28933474128828e-06, "loss": 0.271, "step": 897 }, { "epoch": 2.5584045584045585, "grad_norm": 0.18840475356548164, "learning_rate": 8.236536430834215e-06, "loss": 0.2634, "step": 898 }, { "epoch": 2.561253561253561, "grad_norm": 0.19100598305938737, "learning_rate": 8.183738120380148e-06, "loss": 0.281, "step": 899 }, { "epoch": 2.564102564102564, "grad_norm": 0.1759624339368046, "learning_rate": 8.130939809926083e-06, "loss": 0.2651, "step": 900 }, { "epoch": 2.566951566951567, "grad_norm": 0.19734582647437332, "learning_rate": 8.078141499472017e-06, "loss": 0.2638, "step": 901 }, { "epoch": 2.5698005698005697, "grad_norm": 0.19411668207145158, "learning_rate": 8.025343189017952e-06, "loss": 0.2736, "step": 902 }, { "epoch": 2.5726495726495724, "grad_norm": 0.19879584464405567, "learning_rate": 7.972544878563887e-06, "loss": 0.2769, "step": 903 }, { "epoch": 2.5754985754985755, "grad_norm": 0.1927813426522461, "learning_rate": 7.91974656810982e-06, "loss": 0.2764, "step": 904 }, { "epoch": 2.578347578347578, "grad_norm": 0.19856911323488508, "learning_rate": 7.866948257655756e-06, "loss": 0.292, "step": 905 }, { "epoch": 2.5811965811965814, "grad_norm": 0.17679245135236604, "learning_rate": 7.81414994720169e-06, "loss": 0.2785, "step": 906 }, { "epoch": 2.584045584045584, "grad_norm": 0.19567448364102336, "learning_rate": 7.761351636747624e-06, "loss": 0.2867, "step": 907 }, { "epoch": 2.5868945868945867, "grad_norm": 0.20115854063114447, "learning_rate": 7.708553326293558e-06, "loss": 0.2713, "step": 908 }, { "epoch": 2.58974358974359, "grad_norm": 0.18545206529638672, "learning_rate": 7.655755015839495e-06, "loss": 0.2634, "step": 909 }, { "epoch": 2.5925925925925926, "grad_norm": 0.19495758015732068, "learning_rate": 7.602956705385428e-06, "loss": 0.2723, "step": 910 }, { "epoch": 2.5954415954415957, "grad_norm": 0.18613675067205288, "learning_rate": 7.5501583949313625e-06, "loss": 0.29, "step": 911 }, { "epoch": 2.5982905982905984, "grad_norm": 0.19754545334787374, "learning_rate": 7.497360084477296e-06, "loss": 0.2716, "step": 912 }, { "epoch": 2.601139601139601, "grad_norm": 0.1949333474464277, "learning_rate": 7.444561774023232e-06, "loss": 0.2786, "step": 913 }, { "epoch": 2.603988603988604, "grad_norm": 0.1866184206464396, "learning_rate": 7.3917634635691666e-06, "loss": 0.2845, "step": 914 }, { "epoch": 2.606837606837607, "grad_norm": 0.195676886907894, "learning_rate": 7.3389651531151e-06, "loss": 0.2769, "step": 915 }, { "epoch": 2.6096866096866096, "grad_norm": 0.18429715351385098, "learning_rate": 7.286166842661035e-06, "loss": 0.2756, "step": 916 }, { "epoch": 2.6125356125356127, "grad_norm": 0.18772973110071034, "learning_rate": 7.233368532206971e-06, "loss": 0.2559, "step": 917 }, { "epoch": 2.6153846153846154, "grad_norm": 0.1836610271469343, "learning_rate": 7.180570221752904e-06, "loss": 0.2681, "step": 918 }, { "epoch": 2.618233618233618, "grad_norm": 0.1754155052969751, "learning_rate": 7.127771911298839e-06, "loss": 0.272, "step": 919 }, { "epoch": 2.6210826210826212, "grad_norm": 0.19652318272004426, "learning_rate": 7.074973600844774e-06, "loss": 0.2686, "step": 920 }, { "epoch": 2.623931623931624, "grad_norm": 0.17933290326611778, "learning_rate": 7.022175290390708e-06, "loss": 0.2759, "step": 921 }, { "epoch": 2.6267806267806266, "grad_norm": 0.18155759050319922, "learning_rate": 6.9693769799366424e-06, "loss": 0.2701, "step": 922 }, { "epoch": 2.6296296296296298, "grad_norm": 0.18892897612236667, "learning_rate": 6.916578669482576e-06, "loss": 0.2769, "step": 923 }, { "epoch": 2.6324786324786325, "grad_norm": 0.18477414784333626, "learning_rate": 6.863780359028511e-06, "loss": 0.2696, "step": 924 }, { "epoch": 2.635327635327635, "grad_norm": 0.1797448137751472, "learning_rate": 6.8109820485744465e-06, "loss": 0.2721, "step": 925 }, { "epoch": 2.6381766381766383, "grad_norm": 0.18868102200145584, "learning_rate": 6.75818373812038e-06, "loss": 0.2654, "step": 926 }, { "epoch": 2.641025641025641, "grad_norm": 0.18760249991357172, "learning_rate": 6.705385427666315e-06, "loss": 0.2883, "step": 927 }, { "epoch": 2.6438746438746437, "grad_norm": 0.1792089439057089, "learning_rate": 6.65258711721225e-06, "loss": 0.2631, "step": 928 }, { "epoch": 2.646723646723647, "grad_norm": 0.18382428762667938, "learning_rate": 6.599788806758184e-06, "loss": 0.2832, "step": 929 }, { "epoch": 2.6495726495726495, "grad_norm": 0.18732937470339528, "learning_rate": 6.546990496304118e-06, "loss": 0.2885, "step": 930 }, { "epoch": 2.652421652421652, "grad_norm": 0.18238422692769168, "learning_rate": 6.494192185850054e-06, "loss": 0.2714, "step": 931 }, { "epoch": 2.6552706552706553, "grad_norm": 0.6796049281047194, "learning_rate": 6.441393875395988e-06, "loss": 0.2736, "step": 932 }, { "epoch": 2.658119658119658, "grad_norm": 0.1843964132149436, "learning_rate": 6.388595564941922e-06, "loss": 0.2812, "step": 933 }, { "epoch": 2.6609686609686607, "grad_norm": 0.1883869870506764, "learning_rate": 6.335797254487856e-06, "loss": 0.2876, "step": 934 }, { "epoch": 2.663817663817664, "grad_norm": 0.19372865429873587, "learning_rate": 6.282998944033791e-06, "loss": 0.2721, "step": 935 }, { "epoch": 2.6666666666666665, "grad_norm": 0.17475184508734368, "learning_rate": 6.230200633579726e-06, "loss": 0.2678, "step": 936 }, { "epoch": 2.6695156695156697, "grad_norm": 0.19140736568353386, "learning_rate": 6.17740232312566e-06, "loss": 0.2694, "step": 937 }, { "epoch": 2.6723646723646723, "grad_norm": 0.19095212031796968, "learning_rate": 6.124604012671595e-06, "loss": 0.2714, "step": 938 }, { "epoch": 2.6752136752136755, "grad_norm": 0.18688967708133333, "learning_rate": 6.071805702217529e-06, "loss": 0.2777, "step": 939 }, { "epoch": 2.678062678062678, "grad_norm": 0.1808730509793612, "learning_rate": 6.019007391763464e-06, "loss": 0.2867, "step": 940 }, { "epoch": 2.680911680911681, "grad_norm": 0.1914570882813849, "learning_rate": 5.966209081309398e-06, "loss": 0.2752, "step": 941 }, { "epoch": 2.683760683760684, "grad_norm": 0.18796118324257888, "learning_rate": 5.913410770855333e-06, "loss": 0.282, "step": 942 }, { "epoch": 2.6866096866096867, "grad_norm": 0.19242846471973257, "learning_rate": 5.860612460401267e-06, "loss": 0.2815, "step": 943 }, { "epoch": 2.6894586894586894, "grad_norm": 0.19514299446898078, "learning_rate": 5.807814149947202e-06, "loss": 0.2655, "step": 944 }, { "epoch": 2.6923076923076925, "grad_norm": 0.1879893908540662, "learning_rate": 5.755015839493136e-06, "loss": 0.2857, "step": 945 }, { "epoch": 2.695156695156695, "grad_norm": 0.19017667637491129, "learning_rate": 5.702217529039071e-06, "loss": 0.2807, "step": 946 }, { "epoch": 2.698005698005698, "grad_norm": 0.18886407972543068, "learning_rate": 5.6494192185850055e-06, "loss": 0.2822, "step": 947 }, { "epoch": 2.700854700854701, "grad_norm": 0.2030341917370533, "learning_rate": 5.59662090813094e-06, "loss": 0.2979, "step": 948 }, { "epoch": 2.7037037037037037, "grad_norm": 0.18801018868503175, "learning_rate": 5.543822597676875e-06, "loss": 0.2754, "step": 949 }, { "epoch": 2.7065527065527064, "grad_norm": 0.18722591061825838, "learning_rate": 5.491024287222809e-06, "loss": 0.291, "step": 950 }, { "epoch": 2.7094017094017095, "grad_norm": 0.18419313321990788, "learning_rate": 5.438225976768744e-06, "loss": 0.2786, "step": 951 }, { "epoch": 2.7122507122507122, "grad_norm": 0.1795053639312283, "learning_rate": 5.385427666314678e-06, "loss": 0.2616, "step": 952 }, { "epoch": 2.715099715099715, "grad_norm": 0.19092668654205353, "learning_rate": 5.332629355860613e-06, "loss": 0.266, "step": 953 }, { "epoch": 2.717948717948718, "grad_norm": 0.18515878208885198, "learning_rate": 5.279831045406547e-06, "loss": 0.2683, "step": 954 }, { "epoch": 2.7207977207977208, "grad_norm": 0.18512141190512552, "learning_rate": 5.227032734952482e-06, "loss": 0.2613, "step": 955 }, { "epoch": 2.7236467236467234, "grad_norm": 0.17864087041783647, "learning_rate": 5.174234424498416e-06, "loss": 0.2728, "step": 956 }, { "epoch": 2.7264957264957266, "grad_norm": 0.18397723270469116, "learning_rate": 5.121436114044351e-06, "loss": 0.2878, "step": 957 }, { "epoch": 2.7293447293447293, "grad_norm": 0.19206914593194782, "learning_rate": 5.0686378035902855e-06, "loss": 0.2898, "step": 958 }, { "epoch": 2.732193732193732, "grad_norm": 0.18145822655495156, "learning_rate": 5.01583949313622e-06, "loss": 0.2811, "step": 959 }, { "epoch": 2.735042735042735, "grad_norm": 0.18983105326320673, "learning_rate": 4.963041182682154e-06, "loss": 0.2961, "step": 960 }, { "epoch": 2.737891737891738, "grad_norm": 0.18514943684352905, "learning_rate": 4.910242872228089e-06, "loss": 0.2687, "step": 961 }, { "epoch": 2.7407407407407405, "grad_norm": 0.19048747070950345, "learning_rate": 4.857444561774023e-06, "loss": 0.2742, "step": 962 }, { "epoch": 2.7435897435897436, "grad_norm": 0.18667552629580247, "learning_rate": 4.804646251319958e-06, "loss": 0.2778, "step": 963 }, { "epoch": 2.7464387464387463, "grad_norm": 0.34574568987357907, "learning_rate": 4.751847940865893e-06, "loss": 0.2956, "step": 964 }, { "epoch": 2.7492877492877494, "grad_norm": 0.18370077080612704, "learning_rate": 4.699049630411827e-06, "loss": 0.2873, "step": 965 }, { "epoch": 2.752136752136752, "grad_norm": 0.18353727774729006, "learning_rate": 4.646251319957761e-06, "loss": 0.2729, "step": 966 }, { "epoch": 2.754985754985755, "grad_norm": 0.19726651604825912, "learning_rate": 4.593453009503696e-06, "loss": 0.2774, "step": 967 }, { "epoch": 2.757834757834758, "grad_norm": 0.18831766661505295, "learning_rate": 4.540654699049631e-06, "loss": 0.2918, "step": 968 }, { "epoch": 2.7606837606837606, "grad_norm": 0.1822011598025739, "learning_rate": 4.487856388595565e-06, "loss": 0.2686, "step": 969 }, { "epoch": 2.763532763532764, "grad_norm": 0.1845843568198463, "learning_rate": 4.4350580781415e-06, "loss": 0.2835, "step": 970 }, { "epoch": 2.7663817663817665, "grad_norm": 0.1801611398952323, "learning_rate": 4.382259767687434e-06, "loss": 0.2738, "step": 971 }, { "epoch": 2.769230769230769, "grad_norm": 0.1798470399085861, "learning_rate": 4.329461457233369e-06, "loss": 0.2668, "step": 972 }, { "epoch": 2.7720797720797723, "grad_norm": 0.18404634330655054, "learning_rate": 4.276663146779303e-06, "loss": 0.2685, "step": 973 }, { "epoch": 2.774928774928775, "grad_norm": 0.19542597110088794, "learning_rate": 4.223864836325238e-06, "loss": 0.2854, "step": 974 }, { "epoch": 2.7777777777777777, "grad_norm": 0.17947114319794663, "learning_rate": 4.171066525871172e-06, "loss": 0.2757, "step": 975 }, { "epoch": 2.780626780626781, "grad_norm": 0.17988682607436737, "learning_rate": 4.1182682154171074e-06, "loss": 0.2623, "step": 976 }, { "epoch": 2.7834757834757835, "grad_norm": 0.18466891151924947, "learning_rate": 4.065469904963041e-06, "loss": 0.2673, "step": 977 }, { "epoch": 2.786324786324786, "grad_norm": 0.1803744986096463, "learning_rate": 4.012671594508976e-06, "loss": 0.2769, "step": 978 }, { "epoch": 2.7891737891737893, "grad_norm": 0.18375712425725002, "learning_rate": 3.95987328405491e-06, "loss": 0.2736, "step": 979 }, { "epoch": 2.792022792022792, "grad_norm": 0.1823276203279524, "learning_rate": 3.907074973600845e-06, "loss": 0.2689, "step": 980 }, { "epoch": 2.7948717948717947, "grad_norm": 0.1896182512304451, "learning_rate": 3.854276663146779e-06, "loss": 0.29, "step": 981 }, { "epoch": 2.797720797720798, "grad_norm": 0.20688431780425465, "learning_rate": 3.801478352692714e-06, "loss": 0.2788, "step": 982 }, { "epoch": 2.8005698005698005, "grad_norm": 0.1758754882132954, "learning_rate": 3.748680042238648e-06, "loss": 0.264, "step": 983 }, { "epoch": 2.8034188034188032, "grad_norm": 0.17784331145573304, "learning_rate": 3.6958817317845833e-06, "loss": 0.266, "step": 984 }, { "epoch": 2.8062678062678064, "grad_norm": 0.2198776065570171, "learning_rate": 3.6430834213305176e-06, "loss": 0.2898, "step": 985 }, { "epoch": 2.809116809116809, "grad_norm": 0.21195194560926758, "learning_rate": 3.590285110876452e-06, "loss": 0.2743, "step": 986 }, { "epoch": 2.8119658119658117, "grad_norm": 0.17653709190418312, "learning_rate": 3.537486800422387e-06, "loss": 0.2684, "step": 987 }, { "epoch": 2.814814814814815, "grad_norm": 0.17452382412070946, "learning_rate": 3.4846884899683212e-06, "loss": 0.2672, "step": 988 }, { "epoch": 2.8176638176638176, "grad_norm": 0.1710375478613801, "learning_rate": 3.4318901795142555e-06, "loss": 0.2704, "step": 989 }, { "epoch": 2.8205128205128203, "grad_norm": 0.17775988353031527, "learning_rate": 3.37909186906019e-06, "loss": 0.277, "step": 990 }, { "epoch": 2.8233618233618234, "grad_norm": 0.17938796003091406, "learning_rate": 3.326293558606125e-06, "loss": 0.2844, "step": 991 }, { "epoch": 2.826210826210826, "grad_norm": 0.1762865102421063, "learning_rate": 3.273495248152059e-06, "loss": 0.2761, "step": 992 }, { "epoch": 2.8290598290598292, "grad_norm": 0.18155573837120068, "learning_rate": 3.220696937697994e-06, "loss": 0.2783, "step": 993 }, { "epoch": 2.831908831908832, "grad_norm": 0.1815146018987946, "learning_rate": 3.167898627243928e-06, "loss": 0.264, "step": 994 }, { "epoch": 2.8347578347578346, "grad_norm": 0.17387563307830614, "learning_rate": 3.115100316789863e-06, "loss": 0.2672, "step": 995 }, { "epoch": 2.8376068376068377, "grad_norm": 0.17754589749967897, "learning_rate": 3.0623020063357975e-06, "loss": 0.2604, "step": 996 }, { "epoch": 2.8404558404558404, "grad_norm": 0.1788982942306512, "learning_rate": 3.009503695881732e-06, "loss": 0.2779, "step": 997 }, { "epoch": 2.8433048433048436, "grad_norm": 0.17594751406364015, "learning_rate": 2.9567053854276665e-06, "loss": 0.279, "step": 998 }, { "epoch": 2.8461538461538463, "grad_norm": 0.17837671634444546, "learning_rate": 2.903907074973601e-06, "loss": 0.2701, "step": 999 }, { "epoch": 2.849002849002849, "grad_norm": 0.17682311169747053, "learning_rate": 2.8511087645195354e-06, "loss": 0.2823, "step": 1000 }, { "epoch": 2.851851851851852, "grad_norm": 0.18269585079770145, "learning_rate": 2.79831045406547e-06, "loss": 0.2813, "step": 1001 }, { "epoch": 2.8547008547008548, "grad_norm": 0.17761877789216213, "learning_rate": 2.7455121436114044e-06, "loss": 0.2776, "step": 1002 }, { "epoch": 2.8575498575498575, "grad_norm": 0.1827216995625995, "learning_rate": 2.692713833157339e-06, "loss": 0.2763, "step": 1003 }, { "epoch": 2.8603988603988606, "grad_norm": 0.17763354260264558, "learning_rate": 2.6399155227032734e-06, "loss": 0.2717, "step": 1004 }, { "epoch": 2.8632478632478633, "grad_norm": 0.17915576491895774, "learning_rate": 2.587117212249208e-06, "loss": 0.2806, "step": 1005 }, { "epoch": 2.866096866096866, "grad_norm": 0.18114152761979074, "learning_rate": 2.5343189017951427e-06, "loss": 0.2666, "step": 1006 }, { "epoch": 2.868945868945869, "grad_norm": 0.18225553283170726, "learning_rate": 2.481520591341077e-06, "loss": 0.2624, "step": 1007 }, { "epoch": 2.871794871794872, "grad_norm": 0.1728604793697752, "learning_rate": 2.4287222808870117e-06, "loss": 0.2737, "step": 1008 }, { "epoch": 2.8746438746438745, "grad_norm": 0.18389395414319457, "learning_rate": 2.3759239704329464e-06, "loss": 0.2775, "step": 1009 }, { "epoch": 2.8774928774928776, "grad_norm": 0.1796113765481641, "learning_rate": 2.3231256599788807e-06, "loss": 0.2817, "step": 1010 }, { "epoch": 2.8803418803418803, "grad_norm": 0.18478152892147015, "learning_rate": 2.2703273495248154e-06, "loss": 0.2618, "step": 1011 }, { "epoch": 2.883190883190883, "grad_norm": 0.17453347268398672, "learning_rate": 2.21752903907075e-06, "loss": 0.2749, "step": 1012 }, { "epoch": 2.886039886039886, "grad_norm": 0.1839452873289296, "learning_rate": 2.1647307286166843e-06, "loss": 0.2713, "step": 1013 }, { "epoch": 2.888888888888889, "grad_norm": 0.2010428027513303, "learning_rate": 2.111932418162619e-06, "loss": 0.2704, "step": 1014 }, { "epoch": 2.8917378917378915, "grad_norm": 0.18767478898568274, "learning_rate": 2.0591341077085537e-06, "loss": 0.2806, "step": 1015 }, { "epoch": 2.8945868945868947, "grad_norm": 0.1752592201926614, "learning_rate": 2.006335797254488e-06, "loss": 0.2781, "step": 1016 }, { "epoch": 2.8974358974358974, "grad_norm": 0.18503198686977332, "learning_rate": 1.9535374868004227e-06, "loss": 0.2829, "step": 1017 }, { "epoch": 2.9002849002849, "grad_norm": 0.17491938775507743, "learning_rate": 1.900739176346357e-06, "loss": 0.2735, "step": 1018 }, { "epoch": 2.903133903133903, "grad_norm": 0.18604953590453358, "learning_rate": 1.8479408658922916e-06, "loss": 0.2822, "step": 1019 }, { "epoch": 2.905982905982906, "grad_norm": 0.17680948230251936, "learning_rate": 1.795142555438226e-06, "loss": 0.2654, "step": 1020 }, { "epoch": 2.9088319088319086, "grad_norm": 0.17779057298202855, "learning_rate": 1.7423442449841606e-06, "loss": 0.2809, "step": 1021 }, { "epoch": 2.9116809116809117, "grad_norm": 0.17367317767184776, "learning_rate": 1.689545934530095e-06, "loss": 0.2553, "step": 1022 }, { "epoch": 2.9145299145299144, "grad_norm": 0.1849086381558193, "learning_rate": 1.6367476240760296e-06, "loss": 0.2704, "step": 1023 }, { "epoch": 2.9173789173789175, "grad_norm": 0.2504975572538142, "learning_rate": 1.583949313621964e-06, "loss": 0.2855, "step": 1024 }, { "epoch": 2.92022792022792, "grad_norm": 0.17838279902935134, "learning_rate": 1.5311510031678987e-06, "loss": 0.2633, "step": 1025 }, { "epoch": 2.9230769230769234, "grad_norm": 0.187133440325405, "learning_rate": 1.4783526927138332e-06, "loss": 0.2681, "step": 1026 }, { "epoch": 2.925925925925926, "grad_norm": 0.17495304668665154, "learning_rate": 1.4255543822597677e-06, "loss": 0.2747, "step": 1027 }, { "epoch": 2.9287749287749287, "grad_norm": 0.1768663958665078, "learning_rate": 1.3727560718057022e-06, "loss": 0.2646, "step": 1028 }, { "epoch": 2.931623931623932, "grad_norm": 0.17508470828848635, "learning_rate": 1.3199577613516367e-06, "loss": 0.2664, "step": 1029 }, { "epoch": 2.9344729344729346, "grad_norm": 0.17563477250519552, "learning_rate": 1.2671594508975714e-06, "loss": 0.276, "step": 1030 }, { "epoch": 2.9373219373219372, "grad_norm": 0.18333810882390714, "learning_rate": 1.2143611404435059e-06, "loss": 0.2724, "step": 1031 }, { "epoch": 2.9401709401709404, "grad_norm": 0.18258423976716023, "learning_rate": 1.1615628299894403e-06, "loss": 0.2688, "step": 1032 }, { "epoch": 2.943019943019943, "grad_norm": 0.1740940942606421, "learning_rate": 1.108764519535375e-06, "loss": 0.2628, "step": 1033 }, { "epoch": 2.9458689458689458, "grad_norm": 0.18677299556807853, "learning_rate": 1.0559662090813095e-06, "loss": 0.2727, "step": 1034 }, { "epoch": 2.948717948717949, "grad_norm": 0.17102730623765852, "learning_rate": 1.003167898627244e-06, "loss": 0.2601, "step": 1035 }, { "epoch": 2.9515669515669516, "grad_norm": 0.17386263200587151, "learning_rate": 9.503695881731785e-07, "loss": 0.2657, "step": 1036 }, { "epoch": 2.9544159544159543, "grad_norm": 0.1693040202829608, "learning_rate": 8.97571277719113e-07, "loss": 0.2639, "step": 1037 }, { "epoch": 2.9572649572649574, "grad_norm": 0.1823134515198666, "learning_rate": 8.447729672650475e-07, "loss": 0.2809, "step": 1038 }, { "epoch": 2.96011396011396, "grad_norm": 0.18713660780573363, "learning_rate": 7.91974656810982e-07, "loss": 0.2985, "step": 1039 }, { "epoch": 2.962962962962963, "grad_norm": 0.17859164741168496, "learning_rate": 7.391763463569166e-07, "loss": 0.2728, "step": 1040 }, { "epoch": 2.965811965811966, "grad_norm": 0.17765435111212308, "learning_rate": 6.863780359028511e-07, "loss": 0.273, "step": 1041 }, { "epoch": 2.9686609686609686, "grad_norm": 0.172193961059641, "learning_rate": 6.335797254487857e-07, "loss": 0.2777, "step": 1042 }, { "epoch": 2.9715099715099713, "grad_norm": 0.17703694074644172, "learning_rate": 5.807814149947202e-07, "loss": 0.2729, "step": 1043 }, { "epoch": 2.9743589743589745, "grad_norm": 0.17094716083563843, "learning_rate": 5.279831045406548e-07, "loss": 0.2565, "step": 1044 }, { "epoch": 2.977207977207977, "grad_norm": 0.17381058656887102, "learning_rate": 4.7518479408658924e-07, "loss": 0.2676, "step": 1045 }, { "epoch": 2.98005698005698, "grad_norm": 0.17011520350387407, "learning_rate": 4.2238648363252377e-07, "loss": 0.2711, "step": 1046 }, { "epoch": 2.982905982905983, "grad_norm": 0.1693608759086481, "learning_rate": 3.695881731784583e-07, "loss": 0.2777, "step": 1047 }, { "epoch": 2.9857549857549857, "grad_norm": 0.1691539977152749, "learning_rate": 3.1678986272439284e-07, "loss": 0.258, "step": 1048 }, { "epoch": 2.9886039886039883, "grad_norm": 0.1725795825235304, "learning_rate": 2.639915522703274e-07, "loss": 0.2698, "step": 1049 }, { "epoch": 2.9914529914529915, "grad_norm": 0.17564106791013198, "learning_rate": 2.1119324181626189e-07, "loss": 0.2727, "step": 1050 }, { "epoch": 2.994301994301994, "grad_norm": 0.17354288175962146, "learning_rate": 1.5839493136219642e-07, "loss": 0.2743, "step": 1051 }, { "epoch": 2.9971509971509973, "grad_norm": 0.17083245194414953, "learning_rate": 1.0559662090813094e-07, "loss": 0.2757, "step": 1052 }, { "epoch": 3.0, "grad_norm": 0.17469201616735772, "learning_rate": 5.279831045406547e-08, "loss": 0.2663, "step": 1053 }, { "epoch": 3.0, "step": 1053, "total_flos": 1.1708723554781495e+19, "train_loss": 0.4741822677856384, "train_runtime": 33666.7653, "train_samples_per_second": 0.5, "train_steps_per_second": 0.031 } ], "logging_steps": 1, "max_steps": 1053, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.1708723554781495e+19, "train_batch_size": 2, "trial_name": null, "trial_params": null }