| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 3.0, |
| "eval_steps": 3289, |
| "global_step": 16443, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.0, |
| "grad_norm": 24.78779624251123, |
| "learning_rate": 6.079027355623101e-10, |
| "loss": 2.6939, |
| "step": 1 |
| }, |
| { |
| "epoch": 0.01, |
| "grad_norm": 35.873109917628135, |
| "learning_rate": 1.9452887537993922e-08, |
| "loss": 3.4616, |
| "step": 32 |
| }, |
| { |
| "epoch": 0.01, |
| "grad_norm": 54.36320326793996, |
| "learning_rate": 3.8905775075987844e-08, |
| "loss": 3.544, |
| "step": 64 |
| }, |
| { |
| "epoch": 0.02, |
| "grad_norm": 26.680119386381374, |
| "learning_rate": 5.8358662613981756e-08, |
| "loss": 3.3751, |
| "step": 96 |
| }, |
| { |
| "epoch": 0.02, |
| "grad_norm": 21.12845447174909, |
| "learning_rate": 7.781155015197569e-08, |
| "loss": 3.5534, |
| "step": 128 |
| }, |
| { |
| "epoch": 0.03, |
| "grad_norm": 14.326825936063585, |
| "learning_rate": 9.72644376899696e-08, |
| "loss": 3.2669, |
| "step": 160 |
| }, |
| { |
| "epoch": 0.04, |
| "grad_norm": 25.878315201522145, |
| "learning_rate": 1.1671732522796351e-07, |
| "loss": 2.9778, |
| "step": 192 |
| }, |
| { |
| "epoch": 0.04, |
| "grad_norm": 36.45201708374618, |
| "learning_rate": 1.3617021276595745e-07, |
| "loss": 2.5704, |
| "step": 224 |
| }, |
| { |
| "epoch": 0.05, |
| "grad_norm": 23.711566820577637, |
| "learning_rate": 1.5562310030395138e-07, |
| "loss": 1.3894, |
| "step": 256 |
| }, |
| { |
| "epoch": 0.05, |
| "grad_norm": 4.875938314628448, |
| "learning_rate": 1.7507598784194527e-07, |
| "loss": 0.6738, |
| "step": 288 |
| }, |
| { |
| "epoch": 0.06, |
| "grad_norm": 6.241944245269492, |
| "learning_rate": 1.945288753799392e-07, |
| "loss": 0.5629, |
| "step": 320 |
| }, |
| { |
| "epoch": 0.06, |
| "grad_norm": 2.8778836099761524, |
| "learning_rate": 2.1398176291793313e-07, |
| "loss": 0.4872, |
| "step": 352 |
| }, |
| { |
| "epoch": 0.07, |
| "grad_norm": 1.5436867519527993, |
| "learning_rate": 2.3343465045592702e-07, |
| "loss": 0.4714, |
| "step": 384 |
| }, |
| { |
| "epoch": 0.08, |
| "grad_norm": 4.3096231574697645, |
| "learning_rate": 2.52887537993921e-07, |
| "loss": 0.4378, |
| "step": 416 |
| }, |
| { |
| "epoch": 0.08, |
| "grad_norm": 1.9666937909371047, |
| "learning_rate": 2.723404255319149e-07, |
| "loss": 0.4291, |
| "step": 448 |
| }, |
| { |
| "epoch": 0.09, |
| "grad_norm": 2.2115158499878724, |
| "learning_rate": 2.917933130699088e-07, |
| "loss": 0.3954, |
| "step": 480 |
| }, |
| { |
| "epoch": 0.09, |
| "grad_norm": 1.7998977018189959, |
| "learning_rate": 3.1124620060790275e-07, |
| "loss": 0.4086, |
| "step": 512 |
| }, |
| { |
| "epoch": 0.1, |
| "grad_norm": 1.8295831973352625, |
| "learning_rate": 3.3069908814589665e-07, |
| "loss": 0.3926, |
| "step": 544 |
| }, |
| { |
| "epoch": 0.11, |
| "grad_norm": 2.14598569285712, |
| "learning_rate": 3.5015197568389055e-07, |
| "loss": 0.4026, |
| "step": 576 |
| }, |
| { |
| "epoch": 0.11, |
| "grad_norm": 1.5109663276292344, |
| "learning_rate": 3.696048632218845e-07, |
| "loss": 0.3688, |
| "step": 608 |
| }, |
| { |
| "epoch": 0.12, |
| "grad_norm": 2.453154442640789, |
| "learning_rate": 3.890577507598784e-07, |
| "loss": 0.3919, |
| "step": 640 |
| }, |
| { |
| "epoch": 0.12, |
| "grad_norm": 1.6423681780897064, |
| "learning_rate": 4.085106382978723e-07, |
| "loss": 0.364, |
| "step": 672 |
| }, |
| { |
| "epoch": 0.13, |
| "grad_norm": 1.8702893489598191, |
| "learning_rate": 4.2796352583586625e-07, |
| "loss": 0.3543, |
| "step": 704 |
| }, |
| { |
| "epoch": 0.13, |
| "grad_norm": 1.8401881440614818, |
| "learning_rate": 4.4741641337386015e-07, |
| "loss": 0.3601, |
| "step": 736 |
| }, |
| { |
| "epoch": 0.14, |
| "grad_norm": 1.850566734809788, |
| "learning_rate": 4.6686930091185405e-07, |
| "loss": 0.3566, |
| "step": 768 |
| }, |
| { |
| "epoch": 0.15, |
| "grad_norm": 1.6994281914890326, |
| "learning_rate": 4.86322188449848e-07, |
| "loss": 0.3505, |
| "step": 800 |
| }, |
| { |
| "epoch": 0.15, |
| "grad_norm": 1.8680937366461803, |
| "learning_rate": 5.05775075987842e-07, |
| "loss": 0.3358, |
| "step": 832 |
| }, |
| { |
| "epoch": 0.16, |
| "grad_norm": 1.660364495604376, |
| "learning_rate": 5.252279635258359e-07, |
| "loss": 0.3312, |
| "step": 864 |
| }, |
| { |
| "epoch": 0.16, |
| "grad_norm": 2.562680158872079, |
| "learning_rate": 5.446808510638298e-07, |
| "loss": 0.3346, |
| "step": 896 |
| }, |
| { |
| "epoch": 0.17, |
| "grad_norm": 1.6948446357514406, |
| "learning_rate": 5.641337386018237e-07, |
| "loss": 0.3356, |
| "step": 928 |
| }, |
| { |
| "epoch": 0.18, |
| "grad_norm": 1.209363365641442, |
| "learning_rate": 5.835866261398176e-07, |
| "loss": 0.3279, |
| "step": 960 |
| }, |
| { |
| "epoch": 0.18, |
| "grad_norm": 1.6925812671610025, |
| "learning_rate": 6.030395136778115e-07, |
| "loss": 0.3367, |
| "step": 992 |
| }, |
| { |
| "epoch": 0.19, |
| "grad_norm": 1.5391435373855369, |
| "learning_rate": 6.224924012158055e-07, |
| "loss": 0.3356, |
| "step": 1024 |
| }, |
| { |
| "epoch": 0.19, |
| "grad_norm": 1.4744316930455266, |
| "learning_rate": 6.419452887537994e-07, |
| "loss": 0.3277, |
| "step": 1056 |
| }, |
| { |
| "epoch": 0.2, |
| "grad_norm": 1.7620056653926688, |
| "learning_rate": 6.613981762917933e-07, |
| "loss": 0.3307, |
| "step": 1088 |
| }, |
| { |
| "epoch": 0.2, |
| "grad_norm": 1.4025350626099162, |
| "learning_rate": 6.808510638297872e-07, |
| "loss": 0.3197, |
| "step": 1120 |
| }, |
| { |
| "epoch": 0.21, |
| "grad_norm": 1.3405530061214819, |
| "learning_rate": 7.003039513677811e-07, |
| "loss": 0.3218, |
| "step": 1152 |
| }, |
| { |
| "epoch": 0.22, |
| "grad_norm": 1.4535529320190042, |
| "learning_rate": 7.197568389057751e-07, |
| "loss": 0.3067, |
| "step": 1184 |
| }, |
| { |
| "epoch": 0.22, |
| "grad_norm": 1.445157014748904, |
| "learning_rate": 7.39209726443769e-07, |
| "loss": 0.3087, |
| "step": 1216 |
| }, |
| { |
| "epoch": 0.23, |
| "grad_norm": 1.5871727851863946, |
| "learning_rate": 7.586626139817629e-07, |
| "loss": 0.3252, |
| "step": 1248 |
| }, |
| { |
| "epoch": 0.23, |
| "grad_norm": 1.3848383016325827, |
| "learning_rate": 7.781155015197568e-07, |
| "loss": 0.3113, |
| "step": 1280 |
| }, |
| { |
| "epoch": 0.24, |
| "grad_norm": 1.7455453453163468, |
| "learning_rate": 7.975683890577507e-07, |
| "loss": 0.297, |
| "step": 1312 |
| }, |
| { |
| "epoch": 0.25, |
| "grad_norm": 1.3001668041428838, |
| "learning_rate": 8.170212765957446e-07, |
| "loss": 0.3223, |
| "step": 1344 |
| }, |
| { |
| "epoch": 0.25, |
| "grad_norm": 1.2132377152476534, |
| "learning_rate": 8.364741641337386e-07, |
| "loss": 0.315, |
| "step": 1376 |
| }, |
| { |
| "epoch": 0.26, |
| "grad_norm": 1.4384605625322182, |
| "learning_rate": 8.559270516717325e-07, |
| "loss": 0.3105, |
| "step": 1408 |
| }, |
| { |
| "epoch": 0.26, |
| "grad_norm": 1.1957556077083222, |
| "learning_rate": 8.753799392097264e-07, |
| "loss": 0.3007, |
| "step": 1440 |
| }, |
| { |
| "epoch": 0.27, |
| "grad_norm": 1.466257939308454, |
| "learning_rate": 8.948328267477203e-07, |
| "loss": 0.3034, |
| "step": 1472 |
| }, |
| { |
| "epoch": 0.27, |
| "grad_norm": 1.348037306034132, |
| "learning_rate": 9.142857142857142e-07, |
| "loss": 0.3044, |
| "step": 1504 |
| }, |
| { |
| "epoch": 0.28, |
| "grad_norm": 1.3312009756781122, |
| "learning_rate": 9.337386018237081e-07, |
| "loss": 0.3124, |
| "step": 1536 |
| }, |
| { |
| "epoch": 0.29, |
| "grad_norm": 1.5165497951579907, |
| "learning_rate": 9.531914893617021e-07, |
| "loss": 0.3064, |
| "step": 1568 |
| }, |
| { |
| "epoch": 0.29, |
| "grad_norm": 1.2308464212993413, |
| "learning_rate": 9.72644376899696e-07, |
| "loss": 0.3096, |
| "step": 1600 |
| }, |
| { |
| "epoch": 0.3, |
| "grad_norm": 1.4477927464029368, |
| "learning_rate": 9.9209726443769e-07, |
| "loss": 0.2967, |
| "step": 1632 |
| }, |
| { |
| "epoch": 0.3, |
| "grad_norm": 1.6161236994381232, |
| "learning_rate": 9.999959323804955e-07, |
| "loss": 0.307, |
| "step": 1664 |
| }, |
| { |
| "epoch": 0.31, |
| "grad_norm": 1.408597466262752, |
| "learning_rate": 9.999706931043734e-07, |
| "loss": 0.3045, |
| "step": 1696 |
| }, |
| { |
| "epoch": 0.32, |
| "grad_norm": 1.2262804987989535, |
| "learning_rate": 9.999223791032621e-07, |
| "loss": 0.2982, |
| "step": 1728 |
| }, |
| { |
| "epoch": 0.32, |
| "grad_norm": 1.4772857240369206, |
| "learning_rate": 9.998509926069568e-07, |
| "loss": 0.3085, |
| "step": 1760 |
| }, |
| { |
| "epoch": 0.33, |
| "grad_norm": 1.1705510022471537, |
| "learning_rate": 9.997565369100982e-07, |
| "loss": 0.3102, |
| "step": 1792 |
| }, |
| { |
| "epoch": 0.33, |
| "grad_norm": 1.354193914272892, |
| "learning_rate": 9.996390163720203e-07, |
| "loss": 0.3079, |
| "step": 1824 |
| }, |
| { |
| "epoch": 0.34, |
| "grad_norm": 1.6499740915478518, |
| "learning_rate": 9.99498436416549e-07, |
| "loss": 0.3058, |
| "step": 1856 |
| }, |
| { |
| "epoch": 0.34, |
| "grad_norm": 1.4176797632580183, |
| "learning_rate": 9.993348035317523e-07, |
| "loss": 0.2898, |
| "step": 1888 |
| }, |
| { |
| "epoch": 0.35, |
| "grad_norm": 1.221569587599335, |
| "learning_rate": 9.991481252696405e-07, |
| "loss": 0.2971, |
| "step": 1920 |
| }, |
| { |
| "epoch": 0.36, |
| "grad_norm": 1.6858523944684565, |
| "learning_rate": 9.98938410245818e-07, |
| "loss": 0.3015, |
| "step": 1952 |
| }, |
| { |
| "epoch": 0.36, |
| "grad_norm": 1.3071072285598695, |
| "learning_rate": 9.987056681390846e-07, |
| "loss": 0.3017, |
| "step": 1984 |
| }, |
| { |
| "epoch": 0.37, |
| "grad_norm": 1.3559736389856838, |
| "learning_rate": 9.984499096909904e-07, |
| "loss": 0.2996, |
| "step": 2016 |
| }, |
| { |
| "epoch": 0.37, |
| "grad_norm": 1.6577354687629797, |
| "learning_rate": 9.981711467053391e-07, |
| "loss": 0.2899, |
| "step": 2048 |
| }, |
| { |
| "epoch": 0.38, |
| "grad_norm": 1.3911817831183464, |
| "learning_rate": 9.97869392047643e-07, |
| "loss": 0.2916, |
| "step": 2080 |
| }, |
| { |
| "epoch": 0.39, |
| "grad_norm": 1.4068725629362402, |
| "learning_rate": 9.9754465964453e-07, |
| "loss": 0.3042, |
| "step": 2112 |
| }, |
| { |
| "epoch": 0.39, |
| "grad_norm": 1.2845003819775453, |
| "learning_rate": 9.971969644831001e-07, |
| "loss": 0.2953, |
| "step": 2144 |
| }, |
| { |
| "epoch": 0.4, |
| "grad_norm": 1.3034821415736415, |
| "learning_rate": 9.968263226102348e-07, |
| "loss": 0.2947, |
| "step": 2176 |
| }, |
| { |
| "epoch": 0.4, |
| "grad_norm": 1.297243001736353, |
| "learning_rate": 9.96432751131855e-07, |
| "loss": 0.2849, |
| "step": 2208 |
| }, |
| { |
| "epoch": 0.41, |
| "grad_norm": 1.411186926138334, |
| "learning_rate": 9.960162682121326e-07, |
| "loss": 0.2974, |
| "step": 2240 |
| }, |
| { |
| "epoch": 0.41, |
| "grad_norm": 1.2374941457414885, |
| "learning_rate": 9.955768930726523e-07, |
| "loss": 0.2824, |
| "step": 2272 |
| }, |
| { |
| "epoch": 0.42, |
| "grad_norm": 1.2627652146626807, |
| "learning_rate": 9.95114645991523e-07, |
| "loss": 0.2865, |
| "step": 2304 |
| }, |
| { |
| "epoch": 0.43, |
| "grad_norm": 1.3134029095546937, |
| "learning_rate": 9.946295483024446e-07, |
| "loss": 0.2867, |
| "step": 2336 |
| }, |
| { |
| "epoch": 0.43, |
| "grad_norm": 1.125719093465591, |
| "learning_rate": 9.9412162239372e-07, |
| "loss": 0.292, |
| "step": 2368 |
| }, |
| { |
| "epoch": 0.44, |
| "grad_norm": 1.4224011133495338, |
| "learning_rate": 9.93590891707225e-07, |
| "loss": 0.309, |
| "step": 2400 |
| }, |
| { |
| "epoch": 0.44, |
| "grad_norm": 1.558277250619359, |
| "learning_rate": 9.930373807373245e-07, |
| "loss": 0.289, |
| "step": 2432 |
| }, |
| { |
| "epoch": 0.45, |
| "grad_norm": 1.3208871305799181, |
| "learning_rate": 9.924611150297428e-07, |
| "loss": 0.2883, |
| "step": 2464 |
| }, |
| { |
| "epoch": 0.46, |
| "grad_norm": 1.3373963931205746, |
| "learning_rate": 9.91862121180384e-07, |
| "loss": 0.2808, |
| "step": 2496 |
| }, |
| { |
| "epoch": 0.46, |
| "grad_norm": 1.3910055395902359, |
| "learning_rate": 9.912404268341051e-07, |
| "loss": 0.2932, |
| "step": 2528 |
| }, |
| { |
| "epoch": 0.47, |
| "grad_norm": 1.4574028412197861, |
| "learning_rate": 9.905960606834404e-07, |
| "loss": 0.2947, |
| "step": 2560 |
| }, |
| { |
| "epoch": 0.47, |
| "grad_norm": 1.203878457595175, |
| "learning_rate": 9.899290524672763e-07, |
| "loss": 0.2867, |
| "step": 2592 |
| }, |
| { |
| "epoch": 0.48, |
| "grad_norm": 1.4673267738670204, |
| "learning_rate": 9.892394329694793e-07, |
| "loss": 0.2837, |
| "step": 2624 |
| }, |
| { |
| "epoch": 0.48, |
| "grad_norm": 1.6281520652734605, |
| "learning_rate": 9.885272340174754e-07, |
| "loss": 0.3032, |
| "step": 2656 |
| }, |
| { |
| "epoch": 0.49, |
| "grad_norm": 1.2522157670744074, |
| "learning_rate": 9.877924884807814e-07, |
| "loss": 0.2963, |
| "step": 2688 |
| }, |
| { |
| "epoch": 0.5, |
| "grad_norm": 1.4632043067121088, |
| "learning_rate": 9.870352302694869e-07, |
| "loss": 0.2898, |
| "step": 2720 |
| }, |
| { |
| "epoch": 0.5, |
| "grad_norm": 1.398722594166713, |
| "learning_rate": 9.86255494332691e-07, |
| "loss": 0.2896, |
| "step": 2752 |
| }, |
| { |
| "epoch": 0.51, |
| "grad_norm": 1.3916529397644415, |
| "learning_rate": 9.854533166568867e-07, |
| "loss": 0.2967, |
| "step": 2784 |
| }, |
| { |
| "epoch": 0.51, |
| "grad_norm": 1.3208718268032766, |
| "learning_rate": 9.846287342643032e-07, |
| "loss": 0.2838, |
| "step": 2816 |
| }, |
| { |
| "epoch": 0.52, |
| "grad_norm": 1.0718990960806063, |
| "learning_rate": 9.837817852111949e-07, |
| "loss": 0.2942, |
| "step": 2848 |
| }, |
| { |
| "epoch": 0.53, |
| "grad_norm": 1.3984255388696125, |
| "learning_rate": 9.829125085860858e-07, |
| "loss": 0.2832, |
| "step": 2880 |
| }, |
| { |
| "epoch": 0.53, |
| "grad_norm": 1.258839102452435, |
| "learning_rate": 9.820209445079655e-07, |
| "loss": 0.2964, |
| "step": 2912 |
| }, |
| { |
| "epoch": 0.54, |
| "grad_norm": 1.1619142104791713, |
| "learning_rate": 9.811071341244379e-07, |
| "loss": 0.2727, |
| "step": 2944 |
| }, |
| { |
| "epoch": 0.54, |
| "grad_norm": 1.3494984294429655, |
| "learning_rate": 9.801711196098213e-07, |
| "loss": 0.2848, |
| "step": 2976 |
| }, |
| { |
| "epoch": 0.55, |
| "grad_norm": 1.2501831577892621, |
| "learning_rate": 9.792129441632027e-07, |
| "loss": 0.2749, |
| "step": 3008 |
| }, |
| { |
| "epoch": 0.55, |
| "grad_norm": 1.2580708434458154, |
| "learning_rate": 9.782326520064443e-07, |
| "loss": 0.2868, |
| "step": 3040 |
| }, |
| { |
| "epoch": 0.56, |
| "grad_norm": 1.5601918050638766, |
| "learning_rate": 9.772302883821418e-07, |
| "loss": 0.292, |
| "step": 3072 |
| }, |
| { |
| "epoch": 0.57, |
| "grad_norm": 1.3545732270149806, |
| "learning_rate": 9.762058995515362e-07, |
| "loss": 0.2895, |
| "step": 3104 |
| }, |
| { |
| "epoch": 0.57, |
| "grad_norm": 1.4796071501521761, |
| "learning_rate": 9.751595327923802e-07, |
| "loss": 0.291, |
| "step": 3136 |
| }, |
| { |
| "epoch": 0.58, |
| "grad_norm": 1.3020677021992657, |
| "learning_rate": 9.740912363967546e-07, |
| "loss": 0.294, |
| "step": 3168 |
| }, |
| { |
| "epoch": 0.58, |
| "grad_norm": 1.2903648809439663, |
| "learning_rate": 9.730010596688405e-07, |
| "loss": 0.2971, |
| "step": 3200 |
| }, |
| { |
| "epoch": 0.59, |
| "grad_norm": 1.4322652904369149, |
| "learning_rate": 9.718890529226432e-07, |
| "loss": 0.2842, |
| "step": 3232 |
| }, |
| { |
| "epoch": 0.6, |
| "grad_norm": 1.4918014229559349, |
| "learning_rate": 9.707552674796704e-07, |
| "loss": 0.2855, |
| "step": 3264 |
| }, |
| { |
| "epoch": 0.6, |
| "grad_norm": 1.495160093783148, |
| "learning_rate": 9.695997556665642e-07, |
| "loss": 0.2961, |
| "step": 3296 |
| }, |
| { |
| "epoch": 0.61, |
| "grad_norm": 1.2299397480418435, |
| "learning_rate": 9.68422570812685e-07, |
| "loss": 0.275, |
| "step": 3328 |
| }, |
| { |
| "epoch": 0.61, |
| "grad_norm": 1.4595570985088122, |
| "learning_rate": 9.672237672476504e-07, |
| "loss": 0.2816, |
| "step": 3360 |
| }, |
| { |
| "epoch": 0.62, |
| "grad_norm": 1.2523764393563719, |
| "learning_rate": 9.660034002988288e-07, |
| "loss": 0.2826, |
| "step": 3392 |
| }, |
| { |
| "epoch": 0.62, |
| "grad_norm": 1.1556114989117734, |
| "learning_rate": 9.64761526288785e-07, |
| "loss": 0.2979, |
| "step": 3424 |
| }, |
| { |
| "epoch": 0.63, |
| "grad_norm": 1.2760487310976472, |
| "learning_rate": 9.634982025326808e-07, |
| "loss": 0.2703, |
| "step": 3456 |
| }, |
| { |
| "epoch": 0.64, |
| "grad_norm": 1.1268494754892229, |
| "learning_rate": 9.622134873356302e-07, |
| "loss": 0.2888, |
| "step": 3488 |
| }, |
| { |
| "epoch": 0.64, |
| "grad_norm": 1.5177691734350403, |
| "learning_rate": 9.60907439990008e-07, |
| "loss": 0.2854, |
| "step": 3520 |
| }, |
| { |
| "epoch": 0.65, |
| "grad_norm": 1.4479572539864909, |
| "learning_rate": 9.595801207727145e-07, |
| "loss": 0.2871, |
| "step": 3552 |
| }, |
| { |
| "epoch": 0.65, |
| "grad_norm": 1.3320405553461194, |
| "learning_rate": 9.58231590942392e-07, |
| "loss": 0.2722, |
| "step": 3584 |
| }, |
| { |
| "epoch": 0.66, |
| "grad_norm": 1.2684704306064236, |
| "learning_rate": 9.568619127365979e-07, |
| "loss": 0.2918, |
| "step": 3616 |
| }, |
| { |
| "epoch": 0.67, |
| "grad_norm": 1.153157224662708, |
| "learning_rate": 9.554711493689337e-07, |
| "loss": 0.2727, |
| "step": 3648 |
| }, |
| { |
| "epoch": 0.67, |
| "grad_norm": 1.2053424871803355, |
| "learning_rate": 9.540593650261259e-07, |
| "loss": 0.2727, |
| "step": 3680 |
| }, |
| { |
| "epoch": 0.68, |
| "grad_norm": 1.5070136801134912, |
| "learning_rate": 9.526266248650647e-07, |
| "loss": 0.2829, |
| "step": 3712 |
| }, |
| { |
| "epoch": 0.68, |
| "grad_norm": 1.3626256868915094, |
| "learning_rate": 9.511729950097962e-07, |
| "loss": 0.2948, |
| "step": 3744 |
| }, |
| { |
| "epoch": 0.69, |
| "grad_norm": 1.3413698467022634, |
| "learning_rate": 9.496985425484708e-07, |
| "loss": 0.2804, |
| "step": 3776 |
| }, |
| { |
| "epoch": 0.69, |
| "grad_norm": 1.1666360657022152, |
| "learning_rate": 9.482033355302474e-07, |
| "loss": 0.2779, |
| "step": 3808 |
| }, |
| { |
| "epoch": 0.7, |
| "grad_norm": 1.2681212059953142, |
| "learning_rate": 9.466874429621523e-07, |
| "loss": 0.2766, |
| "step": 3840 |
| }, |
| { |
| "epoch": 0.71, |
| "grad_norm": 1.570024716133246, |
| "learning_rate": 9.451509348058943e-07, |
| "loss": 0.278, |
| "step": 3872 |
| }, |
| { |
| "epoch": 0.71, |
| "grad_norm": 1.348864102054253, |
| "learning_rate": 9.435938819746363e-07, |
| "loss": 0.2814, |
| "step": 3904 |
| }, |
| { |
| "epoch": 0.72, |
| "grad_norm": 1.479630195353394, |
| "learning_rate": 9.420163563297221e-07, |
| "loss": 0.2768, |
| "step": 3936 |
| }, |
| { |
| "epoch": 0.72, |
| "grad_norm": 1.3437829697550214, |
| "learning_rate": 9.4041843067736e-07, |
| "loss": 0.2796, |
| "step": 3968 |
| }, |
| { |
| "epoch": 0.73, |
| "grad_norm": 1.7505808749964085, |
| "learning_rate": 9.388001787652626e-07, |
| "loss": 0.2804, |
| "step": 4000 |
| }, |
| { |
| "epoch": 0.74, |
| "grad_norm": 1.2804732688993112, |
| "learning_rate": 9.37161675279243e-07, |
| "loss": 0.2671, |
| "step": 4032 |
| }, |
| { |
| "epoch": 0.74, |
| "grad_norm": 1.5599945996293794, |
| "learning_rate": 9.355029958397686e-07, |
| "loss": 0.2778, |
| "step": 4064 |
| }, |
| { |
| "epoch": 0.75, |
| "grad_norm": 1.140562030900848, |
| "learning_rate": 9.338242169984701e-07, |
| "loss": 0.2779, |
| "step": 4096 |
| }, |
| { |
| "epoch": 0.75, |
| "grad_norm": 1.1891602382042474, |
| "learning_rate": 9.321254162346089e-07, |
| "loss": 0.2779, |
| "step": 4128 |
| }, |
| { |
| "epoch": 0.76, |
| "grad_norm": 1.6066027745264384, |
| "learning_rate": 9.304066719515013e-07, |
| "loss": 0.2831, |
| "step": 4160 |
| }, |
| { |
| "epoch": 0.76, |
| "grad_norm": 1.3535261095585958, |
| "learning_rate": 9.286680634729005e-07, |
| "loss": 0.2869, |
| "step": 4192 |
| }, |
| { |
| "epoch": 0.77, |
| "grad_norm": 1.449372671449822, |
| "learning_rate": 9.269096710393343e-07, |
| "loss": 0.2859, |
| "step": 4224 |
| }, |
| { |
| "epoch": 0.78, |
| "grad_norm": 1.2209167474794183, |
| "learning_rate": 9.251315758044032e-07, |
| "loss": 0.2908, |
| "step": 4256 |
| }, |
| { |
| "epoch": 0.78, |
| "grad_norm": 1.4925119088076193, |
| "learning_rate": 9.233338598310343e-07, |
| "loss": 0.2821, |
| "step": 4288 |
| }, |
| { |
| "epoch": 0.79, |
| "grad_norm": 1.2663637068625269, |
| "learning_rate": 9.215166060876943e-07, |
| "loss": 0.2839, |
| "step": 4320 |
| }, |
| { |
| "epoch": 0.79, |
| "grad_norm": 1.335605475288019, |
| "learning_rate": 9.196798984445595e-07, |
| "loss": 0.2767, |
| "step": 4352 |
| }, |
| { |
| "epoch": 0.8, |
| "grad_norm": 1.1999064453898192, |
| "learning_rate": 9.178238216696463e-07, |
| "loss": 0.2813, |
| "step": 4384 |
| }, |
| { |
| "epoch": 0.81, |
| "grad_norm": 1.2005680882302527, |
| "learning_rate": 9.159484614248978e-07, |
| "loss": 0.2776, |
| "step": 4416 |
| }, |
| { |
| "epoch": 0.81, |
| "grad_norm": 1.2022881962586156, |
| "learning_rate": 9.140539042622311e-07, |
| "loss": 0.2812, |
| "step": 4448 |
| }, |
| { |
| "epoch": 0.82, |
| "grad_norm": 1.4918926856382582, |
| "learning_rate": 9.121402376195421e-07, |
| "loss": 0.2763, |
| "step": 4480 |
| }, |
| { |
| "epoch": 0.82, |
| "grad_norm": 1.452212648334464, |
| "learning_rate": 9.102075498166705e-07, |
| "loss": 0.2756, |
| "step": 4512 |
| }, |
| { |
| "epoch": 0.83, |
| "grad_norm": 1.4610868927687688, |
| "learning_rate": 9.082559300513237e-07, |
| "loss": 0.2755, |
| "step": 4544 |
| }, |
| { |
| "epoch": 0.83, |
| "grad_norm": 1.2140564103776683, |
| "learning_rate": 9.062854683949602e-07, |
| "loss": 0.266, |
| "step": 4576 |
| }, |
| { |
| "epoch": 0.84, |
| "grad_norm": 1.23274371490023, |
| "learning_rate": 9.042962557886313e-07, |
| "loss": 0.2856, |
| "step": 4608 |
| }, |
| { |
| "epoch": 0.85, |
| "grad_norm": 1.3898063562012308, |
| "learning_rate": 9.022883840387865e-07, |
| "loss": 0.2815, |
| "step": 4640 |
| }, |
| { |
| "epoch": 0.85, |
| "grad_norm": 1.330139097581169, |
| "learning_rate": 9.002619458130339e-07, |
| "loss": 0.28, |
| "step": 4672 |
| }, |
| { |
| "epoch": 0.86, |
| "grad_norm": 1.2874922030319198, |
| "learning_rate": 8.982170346358651e-07, |
| "loss": 0.2669, |
| "step": 4704 |
| }, |
| { |
| "epoch": 0.86, |
| "grad_norm": 1.107586259057668, |
| "learning_rate": 8.961537448843377e-07, |
| "loss": 0.2853, |
| "step": 4736 |
| }, |
| { |
| "epoch": 0.87, |
| "grad_norm": 1.448957134616539, |
| "learning_rate": 8.940721717837205e-07, |
| "loss": 0.2652, |
| "step": 4768 |
| }, |
| { |
| "epoch": 0.88, |
| "grad_norm": 1.210724298652212, |
| "learning_rate": 8.919724114030984e-07, |
| "loss": 0.2795, |
| "step": 4800 |
| }, |
| { |
| "epoch": 0.88, |
| "grad_norm": 1.3698632505550181, |
| "learning_rate": 8.898545606509378e-07, |
| "loss": 0.284, |
| "step": 4832 |
| }, |
| { |
| "epoch": 0.89, |
| "grad_norm": 1.1911981203106947, |
| "learning_rate": 8.877187172706155e-07, |
| "loss": 0.2746, |
| "step": 4864 |
| }, |
| { |
| "epoch": 0.89, |
| "grad_norm": 1.3333516374957606, |
| "learning_rate": 8.855649798359064e-07, |
| "loss": 0.2888, |
| "step": 4896 |
| }, |
| { |
| "epoch": 0.9, |
| "grad_norm": 1.2351888608354114, |
| "learning_rate": 8.833934477464347e-07, |
| "loss": 0.2831, |
| "step": 4928 |
| }, |
| { |
| "epoch": 0.9, |
| "grad_norm": 1.3038819344536112, |
| "learning_rate": 8.812042212230862e-07, |
| "loss": 0.2753, |
| "step": 4960 |
| }, |
| { |
| "epoch": 0.91, |
| "grad_norm": 1.5707271650867898, |
| "learning_rate": 8.789974013033831e-07, |
| "loss": 0.2759, |
| "step": 4992 |
| }, |
| { |
| "epoch": 0.92, |
| "grad_norm": 1.2820047239265864, |
| "learning_rate": 8.767730898368208e-07, |
| "loss": 0.2746, |
| "step": 5024 |
| }, |
| { |
| "epoch": 0.92, |
| "grad_norm": 1.1744722963290781, |
| "learning_rate": 8.745313894801669e-07, |
| "loss": 0.291, |
| "step": 5056 |
| }, |
| { |
| "epoch": 0.93, |
| "grad_norm": 1.0913528023957022, |
| "learning_rate": 8.722724036927242e-07, |
| "loss": 0.2837, |
| "step": 5088 |
| }, |
| { |
| "epoch": 0.93, |
| "grad_norm": 1.3404827415629068, |
| "learning_rate": 8.699962367315552e-07, |
| "loss": 0.2676, |
| "step": 5120 |
| }, |
| { |
| "epoch": 0.94, |
| "grad_norm": 1.163041143181648, |
| "learning_rate": 8.677029936466706e-07, |
| "loss": 0.2832, |
| "step": 5152 |
| }, |
| { |
| "epoch": 0.95, |
| "grad_norm": 1.436779107568984, |
| "learning_rate": 8.653927802761809e-07, |
| "loss": 0.2728, |
| "step": 5184 |
| }, |
| { |
| "epoch": 0.95, |
| "grad_norm": 1.0080451953033644, |
| "learning_rate": 8.630657032414121e-07, |
| "loss": 0.2617, |
| "step": 5216 |
| }, |
| { |
| "epoch": 0.96, |
| "grad_norm": 1.1100819749235487, |
| "learning_rate": 8.607218699419843e-07, |
| "loss": 0.2942, |
| "step": 5248 |
| }, |
| { |
| "epoch": 0.96, |
| "grad_norm": 1.2992077116418177, |
| "learning_rate": 8.583613885508556e-07, |
| "loss": 0.2672, |
| "step": 5280 |
| }, |
| { |
| "epoch": 0.97, |
| "grad_norm": 1.223919929869006, |
| "learning_rate": 8.559843680093296e-07, |
| "loss": 0.277, |
| "step": 5312 |
| }, |
| { |
| "epoch": 0.98, |
| "grad_norm": 1.1391386556761285, |
| "learning_rate": 8.535909180220268e-07, |
| "loss": 0.2806, |
| "step": 5344 |
| }, |
| { |
| "epoch": 0.98, |
| "grad_norm": 1.2127293368764673, |
| "learning_rate": 8.511811490518225e-07, |
| "loss": 0.2619, |
| "step": 5376 |
| }, |
| { |
| "epoch": 0.99, |
| "grad_norm": 1.140029182839281, |
| "learning_rate": 8.48755172314748e-07, |
| "loss": 0.2718, |
| "step": 5408 |
| }, |
| { |
| "epoch": 0.99, |
| "grad_norm": 1.4671327877043736, |
| "learning_rate": 8.463130997748578e-07, |
| "loss": 0.272, |
| "step": 5440 |
| }, |
| { |
| "epoch": 1.0, |
| "grad_norm": 1.1390017949538402, |
| "learning_rate": 8.43855044139063e-07, |
| "loss": 0.2737, |
| "step": 5472 |
| }, |
| { |
| "epoch": 1.0, |
| "grad_norm": 1.2142876905398077, |
| "learning_rate": 8.413811188519282e-07, |
| "loss": 0.2599, |
| "step": 5504 |
| }, |
| { |
| "epoch": 1.01, |
| "grad_norm": 1.266246650515794, |
| "learning_rate": 8.388914380904373e-07, |
| "loss": 0.2465, |
| "step": 5536 |
| }, |
| { |
| "epoch": 1.02, |
| "grad_norm": 1.173280963516161, |
| "learning_rate": 8.36386116758723e-07, |
| "loss": 0.2483, |
| "step": 5568 |
| }, |
| { |
| "epoch": 1.02, |
| "grad_norm": 1.1617564369184084, |
| "learning_rate": 8.33865270482764e-07, |
| "loss": 0.2568, |
| "step": 5600 |
| }, |
| { |
| "epoch": 1.03, |
| "grad_norm": 1.420351934994977, |
| "learning_rate": 8.313290156050487e-07, |
| "loss": 0.2502, |
| "step": 5632 |
| }, |
| { |
| "epoch": 1.03, |
| "grad_norm": 1.3547439002353128, |
| "learning_rate": 8.287774691792057e-07, |
| "loss": 0.246, |
| "step": 5664 |
| }, |
| { |
| "epoch": 1.04, |
| "grad_norm": 1.1455753710552021, |
| "learning_rate": 8.262107489646014e-07, |
| "loss": 0.2554, |
| "step": 5696 |
| }, |
| { |
| "epoch": 1.05, |
| "grad_norm": 1.3859247572490137, |
| "learning_rate": 8.236289734209054e-07, |
| "loss": 0.2466, |
| "step": 5728 |
| }, |
| { |
| "epoch": 1.05, |
| "grad_norm": 1.1343048895326515, |
| "learning_rate": 8.210322617026232e-07, |
| "loss": 0.2467, |
| "step": 5760 |
| }, |
| { |
| "epoch": 1.06, |
| "grad_norm": 1.2217575468560562, |
| "learning_rate": 8.184207336535967e-07, |
| "loss": 0.2574, |
| "step": 5792 |
| }, |
| { |
| "epoch": 1.06, |
| "grad_norm": 1.3780997994944977, |
| "learning_rate": 8.157945098014733e-07, |
| "loss": 0.2481, |
| "step": 5824 |
| }, |
| { |
| "epoch": 1.07, |
| "grad_norm": 1.0870131072717084, |
| "learning_rate": 8.131537113521443e-07, |
| "loss": 0.2588, |
| "step": 5856 |
| }, |
| { |
| "epoch": 1.07, |
| "grad_norm": 1.0490087402219708, |
| "learning_rate": 8.104984601841489e-07, |
| "loss": 0.246, |
| "step": 5888 |
| }, |
| { |
| "epoch": 1.08, |
| "grad_norm": 1.2965280317141281, |
| "learning_rate": 8.07828878843051e-07, |
| "loss": 0.249, |
| "step": 5920 |
| }, |
| { |
| "epoch": 1.09, |
| "grad_norm": 1.2861279001372838, |
| "learning_rate": 8.051450905357829e-07, |
| "loss": 0.2412, |
| "step": 5952 |
| }, |
| { |
| "epoch": 1.09, |
| "grad_norm": 1.409445153252456, |
| "learning_rate": 8.024472191249588e-07, |
| "loss": 0.2431, |
| "step": 5984 |
| }, |
| { |
| "epoch": 1.1, |
| "grad_norm": 1.3691683733505722, |
| "learning_rate": 7.997353891231584e-07, |
| "loss": 0.2517, |
| "step": 6016 |
| }, |
| { |
| "epoch": 1.1, |
| "grad_norm": 1.4906649500848184, |
| "learning_rate": 7.970097256871811e-07, |
| "loss": 0.2517, |
| "step": 6048 |
| }, |
| { |
| "epoch": 1.11, |
| "grad_norm": 1.4438858223954047, |
| "learning_rate": 7.942703546122682e-07, |
| "loss": 0.2584, |
| "step": 6080 |
| }, |
| { |
| "epoch": 1.12, |
| "grad_norm": 1.268958450317541, |
| "learning_rate": 7.915174023262988e-07, |
| "loss": 0.2522, |
| "step": 6112 |
| }, |
| { |
| "epoch": 1.12, |
| "grad_norm": 1.4781129605275483, |
| "learning_rate": 7.887509958839538e-07, |
| "loss": 0.243, |
| "step": 6144 |
| }, |
| { |
| "epoch": 1.13, |
| "grad_norm": 1.331949493045608, |
| "learning_rate": 7.859712629608524e-07, |
| "loss": 0.2429, |
| "step": 6176 |
| }, |
| { |
| "epoch": 1.13, |
| "grad_norm": 1.306062446109076, |
| "learning_rate": 7.831783318476596e-07, |
| "loss": 0.2446, |
| "step": 6208 |
| }, |
| { |
| "epoch": 1.14, |
| "grad_norm": 1.242511236714611, |
| "learning_rate": 7.803723314441656e-07, |
| "loss": 0.2451, |
| "step": 6240 |
| }, |
| { |
| "epoch": 1.14, |
| "grad_norm": 1.4482758408678764, |
| "learning_rate": 7.775533912533363e-07, |
| "loss": 0.2412, |
| "step": 6272 |
| }, |
| { |
| "epoch": 1.15, |
| "grad_norm": 1.3719989834400954, |
| "learning_rate": 7.747216413753366e-07, |
| "loss": 0.2421, |
| "step": 6304 |
| }, |
| { |
| "epoch": 1.16, |
| "grad_norm": 1.1916507119961282, |
| "learning_rate": 7.718772125015263e-07, |
| "loss": 0.2536, |
| "step": 6336 |
| }, |
| { |
| "epoch": 1.16, |
| "grad_norm": 1.2009787273487618, |
| "learning_rate": 7.690202359084278e-07, |
| "loss": 0.2501, |
| "step": 6368 |
| }, |
| { |
| "epoch": 1.17, |
| "grad_norm": 1.2823116236203336, |
| "learning_rate": 7.661508434516682e-07, |
| "loss": 0.2524, |
| "step": 6400 |
| }, |
| { |
| "epoch": 1.17, |
| "grad_norm": 1.3430035950312693, |
| "learning_rate": 7.632691675598935e-07, |
| "loss": 0.2577, |
| "step": 6432 |
| }, |
| { |
| "epoch": 1.18, |
| "grad_norm": 1.4985924089749454, |
| "learning_rate": 7.60375341228656e-07, |
| "loss": 0.2444, |
| "step": 6464 |
| }, |
| { |
| "epoch": 1.19, |
| "grad_norm": 1.2984272436604025, |
| "learning_rate": 7.574694980142779e-07, |
| "loss": 0.2615, |
| "step": 6496 |
| }, |
| { |
| "epoch": 1.19, |
| "grad_norm": 1.3354550623218928, |
| "learning_rate": 7.545517720276857e-07, |
| "loss": 0.2389, |
| "step": 6528 |
| }, |
| { |
| "epoch": 1.2, |
| "grad_norm": 1.4301768294716657, |
| "learning_rate": 7.51622297928222e-07, |
| "loss": 0.2449, |
| "step": 6560 |
| }, |
| { |
| "epoch": 1.2, |
| "grad_norm": 1.1719113606920948, |
| "learning_rate": 7.4868121091743e-07, |
| "loss": 0.2423, |
| "step": 6592 |
| }, |
| { |
| "epoch": 1.21, |
| "grad_norm": 1.4204627356681534, |
| "learning_rate": 7.457286467328135e-07, |
| "loss": 0.2445, |
| "step": 6624 |
| }, |
| { |
| "epoch": 1.21, |
| "grad_norm": 1.4900172188399152, |
| "learning_rate": 7.427647416415725e-07, |
| "loss": 0.2483, |
| "step": 6656 |
| }, |
| { |
| "epoch": 1.22, |
| "grad_norm": 1.3961859599185324, |
| "learning_rate": 7.397896324343151e-07, |
| "loss": 0.2359, |
| "step": 6688 |
| }, |
| { |
| "epoch": 1.23, |
| "grad_norm": 1.3110300825940397, |
| "learning_rate": 7.368034564187424e-07, |
| "loss": 0.2512, |
| "step": 6720 |
| }, |
| { |
| "epoch": 1.23, |
| "grad_norm": 1.2757069881397953, |
| "learning_rate": 7.338063514133136e-07, |
| "loss": 0.2547, |
| "step": 6752 |
| }, |
| { |
| "epoch": 1.24, |
| "grad_norm": 1.4074893254437457, |
| "learning_rate": 7.307984557408837e-07, |
| "loss": 0.2496, |
| "step": 6784 |
| }, |
| { |
| "epoch": 1.24, |
| "grad_norm": 1.351839197713545, |
| "learning_rate": 7.277799082223204e-07, |
| "loss": 0.2353, |
| "step": 6816 |
| }, |
| { |
| "epoch": 1.25, |
| "grad_norm": 1.2030842443572731, |
| "learning_rate": 7.24750848170097e-07, |
| "loss": 0.2559, |
| "step": 6848 |
| }, |
| { |
| "epoch": 1.26, |
| "grad_norm": 1.324374188137303, |
| "learning_rate": 7.217114153818629e-07, |
| "loss": 0.2448, |
| "step": 6880 |
| }, |
| { |
| "epoch": 1.26, |
| "grad_norm": 1.2343518311233508, |
| "learning_rate": 7.186617501339917e-07, |
| "loss": 0.2475, |
| "step": 6912 |
| }, |
| { |
| "epoch": 1.27, |
| "grad_norm": 1.257477506253048, |
| "learning_rate": 7.156019931751071e-07, |
| "loss": 0.2529, |
| "step": 6944 |
| }, |
| { |
| "epoch": 1.27, |
| "grad_norm": 1.3065578571620988, |
| "learning_rate": 7.125322857195867e-07, |
| "loss": 0.2439, |
| "step": 6976 |
| }, |
| { |
| "epoch": 1.28, |
| "grad_norm": 1.491754680548748, |
| "learning_rate": 7.094527694410455e-07, |
| "loss": 0.2581, |
| "step": 7008 |
| }, |
| { |
| "epoch": 1.28, |
| "grad_norm": 1.2313699652632502, |
| "learning_rate": 7.063635864657964e-07, |
| "loss": 0.2462, |
| "step": 7040 |
| }, |
| { |
| "epoch": 1.29, |
| "grad_norm": 1.200487097478514, |
| "learning_rate": 7.032648793662912e-07, |
| "loss": 0.2422, |
| "step": 7072 |
| }, |
| { |
| "epoch": 1.3, |
| "grad_norm": 1.3322942461823077, |
| "learning_rate": 7.001567911545406e-07, |
| "loss": 0.2456, |
| "step": 7104 |
| }, |
| { |
| "epoch": 1.3, |
| "grad_norm": 1.4431169758144065, |
| "learning_rate": 6.970394652755144e-07, |
| "loss": 0.2482, |
| "step": 7136 |
| }, |
| { |
| "epoch": 1.31, |
| "grad_norm": 1.261334251262187, |
| "learning_rate": 6.939130456005196e-07, |
| "loss": 0.2456, |
| "step": 7168 |
| }, |
| { |
| "epoch": 1.31, |
| "grad_norm": 1.4803616635583337, |
| "learning_rate": 6.907776764205622e-07, |
| "loss": 0.2628, |
| "step": 7200 |
| }, |
| { |
| "epoch": 1.32, |
| "grad_norm": 1.318319361538514, |
| "learning_rate": 6.876335024396871e-07, |
| "loss": 0.2408, |
| "step": 7232 |
| }, |
| { |
| "epoch": 1.33, |
| "grad_norm": 1.2799133546143286, |
| "learning_rate": 6.844806687682996e-07, |
| "loss": 0.2445, |
| "step": 7264 |
| }, |
| { |
| "epoch": 1.33, |
| "grad_norm": 1.2877655780208916, |
| "learning_rate": 6.813193209164683e-07, |
| "loss": 0.2406, |
| "step": 7296 |
| }, |
| { |
| "epoch": 1.34, |
| "grad_norm": 1.344111445915826, |
| "learning_rate": 6.781496047872098e-07, |
| "loss": 0.2408, |
| "step": 7328 |
| }, |
| { |
| "epoch": 1.34, |
| "grad_norm": 1.5226804349220626, |
| "learning_rate": 6.749716666697545e-07, |
| "loss": 0.2469, |
| "step": 7360 |
| }, |
| { |
| "epoch": 1.35, |
| "grad_norm": 1.3170301342228112, |
| "learning_rate": 6.717856532327956e-07, |
| "loss": 0.2407, |
| "step": 7392 |
| }, |
| { |
| "epoch": 1.35, |
| "grad_norm": 1.185336007297627, |
| "learning_rate": 6.685917115177193e-07, |
| "loss": 0.2445, |
| "step": 7424 |
| }, |
| { |
| "epoch": 1.36, |
| "grad_norm": 1.3491930633425604, |
| "learning_rate": 6.653899889318192e-07, |
| "loss": 0.2562, |
| "step": 7456 |
| }, |
| { |
| "epoch": 1.37, |
| "grad_norm": 1.160879383772545, |
| "learning_rate": 6.621806332414925e-07, |
| "loss": 0.2427, |
| "step": 7488 |
| }, |
| { |
| "epoch": 1.37, |
| "grad_norm": 1.5333857134961049, |
| "learning_rate": 6.589637925654215e-07, |
| "loss": 0.251, |
| "step": 7520 |
| }, |
| { |
| "epoch": 1.38, |
| "grad_norm": 1.5263236812468681, |
| "learning_rate": 6.557396153677356e-07, |
| "loss": 0.2543, |
| "step": 7552 |
| }, |
| { |
| "epoch": 1.38, |
| "grad_norm": 1.0995345813700304, |
| "learning_rate": 6.525082504511612e-07, |
| "loss": 0.2493, |
| "step": 7584 |
| }, |
| { |
| "epoch": 1.39, |
| "grad_norm": 1.4740647676869196, |
| "learning_rate": 6.492698469501532e-07, |
| "loss": 0.2414, |
| "step": 7616 |
| }, |
| { |
| "epoch": 1.4, |
| "grad_norm": 1.4036985523166017, |
| "learning_rate": 6.460245543240123e-07, |
| "loss": 0.2373, |
| "step": 7648 |
| }, |
| { |
| "epoch": 1.4, |
| "grad_norm": 1.4125143998119496, |
| "learning_rate": 6.427725223499871e-07, |
| "loss": 0.2427, |
| "step": 7680 |
| }, |
| { |
| "epoch": 1.41, |
| "grad_norm": 1.195620693310162, |
| "learning_rate": 6.39513901116362e-07, |
| "loss": 0.2522, |
| "step": 7712 |
| }, |
| { |
| "epoch": 1.41, |
| "grad_norm": 1.2679060414661505, |
| "learning_rate": 6.362488410155298e-07, |
| "loss": 0.2448, |
| "step": 7744 |
| }, |
| { |
| "epoch": 1.42, |
| "grad_norm": 1.4675997129467317, |
| "learning_rate": 6.329774927370504e-07, |
| "loss": 0.2459, |
| "step": 7776 |
| }, |
| { |
| "epoch": 1.42, |
| "grad_norm": 1.3511806562135458, |
| "learning_rate": 6.297000072606972e-07, |
| "loss": 0.257, |
| "step": 7808 |
| }, |
| { |
| "epoch": 1.43, |
| "grad_norm": 1.272848715104825, |
| "learning_rate": 6.264165358494884e-07, |
| "loss": 0.246, |
| "step": 7840 |
| }, |
| { |
| "epoch": 1.44, |
| "grad_norm": 1.4081103274783833, |
| "learning_rate": 6.231272300427059e-07, |
| "loss": 0.2485, |
| "step": 7872 |
| }, |
| { |
| "epoch": 1.44, |
| "grad_norm": 1.2073750932804077, |
| "learning_rate": 6.198322416489016e-07, |
| "loss": 0.2354, |
| "step": 7904 |
| }, |
| { |
| "epoch": 1.45, |
| "grad_norm": 1.3509938246652688, |
| "learning_rate": 6.165317227388911e-07, |
| "loss": 0.2423, |
| "step": 7936 |
| }, |
| { |
| "epoch": 1.45, |
| "grad_norm": 1.8975854309012892, |
| "learning_rate": 6.132258256387348e-07, |
| "loss": 0.2422, |
| "step": 7968 |
| }, |
| { |
| "epoch": 1.46, |
| "grad_norm": 1.616263805539082, |
| "learning_rate": 6.099147029227088e-07, |
| "loss": 0.2452, |
| "step": 8000 |
| }, |
| { |
| "epoch": 1.47, |
| "grad_norm": 1.4581239370531887, |
| "learning_rate": 6.065985074062624e-07, |
| "loss": 0.2564, |
| "step": 8032 |
| }, |
| { |
| "epoch": 1.47, |
| "grad_norm": 1.4107792940118344, |
| "learning_rate": 6.032773921389654e-07, |
| "loss": 0.2613, |
| "step": 8064 |
| }, |
| { |
| "epoch": 1.48, |
| "grad_norm": 1.4254738015511745, |
| "learning_rate": 5.999515103974447e-07, |
| "loss": 0.2441, |
| "step": 8096 |
| }, |
| { |
| "epoch": 1.48, |
| "grad_norm": 1.2694180719111068, |
| "learning_rate": 5.966210156783108e-07, |
| "loss": 0.2559, |
| "step": 8128 |
| }, |
| { |
| "epoch": 1.49, |
| "grad_norm": 1.3375582532247023, |
| "learning_rate": 5.932860616910721e-07, |
| "loss": 0.2488, |
| "step": 8160 |
| }, |
| { |
| "epoch": 1.49, |
| "grad_norm": 1.2586037041521065, |
| "learning_rate": 5.899468023510428e-07, |
| "loss": 0.2427, |
| "step": 8192 |
| }, |
| { |
| "epoch": 1.5, |
| "grad_norm": 1.2896167720703278, |
| "learning_rate": 5.866033917722379e-07, |
| "loss": 0.2567, |
| "step": 8224 |
| }, |
| { |
| "epoch": 1.51, |
| "grad_norm": 1.101067530433089, |
| "learning_rate": 5.832559842602608e-07, |
| "loss": 0.2342, |
| "step": 8256 |
| }, |
| { |
| "epoch": 1.51, |
| "grad_norm": 1.1591181439961709, |
| "learning_rate": 5.799047343051827e-07, |
| "loss": 0.2411, |
| "step": 8288 |
| }, |
| { |
| "epoch": 1.52, |
| "grad_norm": 1.2887740481533372, |
| "learning_rate": 5.765497965744111e-07, |
| "loss": 0.2433, |
| "step": 8320 |
| }, |
| { |
| "epoch": 1.52, |
| "grad_norm": 1.1116096883871758, |
| "learning_rate": 5.731913259055527e-07, |
| "loss": 0.2435, |
| "step": 8352 |
| }, |
| { |
| "epoch": 1.53, |
| "grad_norm": 1.3978304499588927, |
| "learning_rate": 5.698294772992669e-07, |
| "loss": 0.2485, |
| "step": 8384 |
| }, |
| { |
| "epoch": 1.54, |
| "grad_norm": 1.1353060264434358, |
| "learning_rate": 5.664644059121121e-07, |
| "loss": 0.2508, |
| "step": 8416 |
| }, |
| { |
| "epoch": 1.54, |
| "grad_norm": 1.337326494552789, |
| "learning_rate": 5.630962670493848e-07, |
| "loss": 0.2446, |
| "step": 8448 |
| }, |
| { |
| "epoch": 1.55, |
| "grad_norm": 1.5497857023189168, |
| "learning_rate": 5.597252161579525e-07, |
| "loss": 0.2498, |
| "step": 8480 |
| }, |
| { |
| "epoch": 1.55, |
| "grad_norm": 1.6735926212275043, |
| "learning_rate": 5.563514088190788e-07, |
| "loss": 0.2382, |
| "step": 8512 |
| }, |
| { |
| "epoch": 1.56, |
| "grad_norm": 1.5802941502096313, |
| "learning_rate": 5.529750007412435e-07, |
| "loss": 0.24, |
| "step": 8544 |
| }, |
| { |
| "epoch": 1.56, |
| "grad_norm": 1.434775976642666, |
| "learning_rate": 5.495961477529559e-07, |
| "loss": 0.2508, |
| "step": 8576 |
| }, |
| { |
| "epoch": 1.57, |
| "grad_norm": 1.3871703235640929, |
| "learning_rate": 5.462150057955633e-07, |
| "loss": 0.24, |
| "step": 8608 |
| }, |
| { |
| "epoch": 1.58, |
| "grad_norm": 1.3448291807441222, |
| "learning_rate": 5.428317309160538e-07, |
| "loss": 0.2484, |
| "step": 8640 |
| }, |
| { |
| "epoch": 1.58, |
| "grad_norm": 1.3157038644549268, |
| "learning_rate": 5.394464792598545e-07, |
| "loss": 0.2409, |
| "step": 8672 |
| }, |
| { |
| "epoch": 1.59, |
| "grad_norm": 1.2008640772377304, |
| "learning_rate": 5.360594070636248e-07, |
| "loss": 0.2454, |
| "step": 8704 |
| }, |
| { |
| "epoch": 1.59, |
| "grad_norm": 1.3073489027432128, |
| "learning_rate": 5.326706706480467e-07, |
| "loss": 0.2446, |
| "step": 8736 |
| }, |
| { |
| "epoch": 1.6, |
| "grad_norm": 1.5955373808274311, |
| "learning_rate": 5.292804264106083e-07, |
| "loss": 0.2433, |
| "step": 8768 |
| }, |
| { |
| "epoch": 1.61, |
| "grad_norm": 1.2901907823242207, |
| "learning_rate": 5.25888830818388e-07, |
| "loss": 0.2426, |
| "step": 8800 |
| }, |
| { |
| "epoch": 1.61, |
| "grad_norm": 1.4454540634426136, |
| "learning_rate": 5.224960404008319e-07, |
| "loss": 0.2487, |
| "step": 8832 |
| }, |
| { |
| "epoch": 1.62, |
| "grad_norm": 1.3252857725269103, |
| "learning_rate": 5.19102211742529e-07, |
| "loss": 0.242, |
| "step": 8864 |
| }, |
| { |
| "epoch": 1.62, |
| "grad_norm": 1.3963560963709287, |
| "learning_rate": 5.157075014759866e-07, |
| "loss": 0.2333, |
| "step": 8896 |
| }, |
| { |
| "epoch": 1.63, |
| "grad_norm": 1.7474842902023573, |
| "learning_rate": 5.123120662743993e-07, |
| "loss": 0.2397, |
| "step": 8928 |
| }, |
| { |
| "epoch": 1.63, |
| "grad_norm": 1.463552241271405, |
| "learning_rate": 5.089160628444193e-07, |
| "loss": 0.2475, |
| "step": 8960 |
| }, |
| { |
| "epoch": 1.64, |
| "grad_norm": 1.4711904387458636, |
| "learning_rate": 5.055196479189237e-07, |
| "loss": 0.2468, |
| "step": 8992 |
| }, |
| { |
| "epoch": 1.65, |
| "grad_norm": 1.1486741329450094, |
| "learning_rate": 5.021229782497811e-07, |
| "loss": 0.2474, |
| "step": 9024 |
| }, |
| { |
| "epoch": 1.65, |
| "grad_norm": 1.465014768579321, |
| "learning_rate": 4.987262106006171e-07, |
| "loss": 0.2437, |
| "step": 9056 |
| }, |
| { |
| "epoch": 1.66, |
| "grad_norm": 1.6572146980799467, |
| "learning_rate": 4.953295017395788e-07, |
| "loss": 0.2358, |
| "step": 9088 |
| }, |
| { |
| "epoch": 1.66, |
| "grad_norm": 1.1497400773353195, |
| "learning_rate": 4.919330084321009e-07, |
| "loss": 0.2459, |
| "step": 9120 |
| }, |
| { |
| "epoch": 1.67, |
| "grad_norm": 1.4973289494569486, |
| "learning_rate": 4.885368874336694e-07, |
| "loss": 0.2497, |
| "step": 9152 |
| }, |
| { |
| "epoch": 1.68, |
| "grad_norm": 1.375218554438502, |
| "learning_rate": 4.851412954825874e-07, |
| "loss": 0.2508, |
| "step": 9184 |
| }, |
| { |
| "epoch": 1.68, |
| "grad_norm": 1.3780517501267755, |
| "learning_rate": 4.817463892927411e-07, |
| "loss": 0.2567, |
| "step": 9216 |
| }, |
| { |
| "epoch": 1.69, |
| "grad_norm": 1.4504412670024611, |
| "learning_rate": 4.783523255463679e-07, |
| "loss": 0.2409, |
| "step": 9248 |
| }, |
| { |
| "epoch": 1.69, |
| "grad_norm": 1.2870854505712077, |
| "learning_rate": 4.7495926088682436e-07, |
| "loss": 0.2423, |
| "step": 9280 |
| }, |
| { |
| "epoch": 1.7, |
| "grad_norm": 1.3985153319949002, |
| "learning_rate": 4.71567351911357e-07, |
| "loss": 0.251, |
| "step": 9312 |
| }, |
| { |
| "epoch": 1.7, |
| "grad_norm": 1.0349716694291753, |
| "learning_rate": 4.681767551638751e-07, |
| "loss": 0.2404, |
| "step": 9344 |
| }, |
| { |
| "epoch": 1.71, |
| "grad_norm": 1.428114124678113, |
| "learning_rate": 4.647876271277257e-07, |
| "loss": 0.2391, |
| "step": 9376 |
| }, |
| { |
| "epoch": 1.72, |
| "grad_norm": 1.3200951440851763, |
| "learning_rate": 4.6140012421847136e-07, |
| "loss": 0.2535, |
| "step": 9408 |
| }, |
| { |
| "epoch": 1.72, |
| "grad_norm": 1.3868492152964986, |
| "learning_rate": 4.5801440277667235e-07, |
| "loss": 0.2483, |
| "step": 9440 |
| }, |
| { |
| "epoch": 1.73, |
| "grad_norm": 1.2729900579595428, |
| "learning_rate": 4.5463061906066965e-07, |
| "loss": 0.2512, |
| "step": 9472 |
| }, |
| { |
| "epoch": 1.73, |
| "grad_norm": 1.315198823460812, |
| "learning_rate": 4.5124892923937416e-07, |
| "loss": 0.2417, |
| "step": 9504 |
| }, |
| { |
| "epoch": 1.74, |
| "grad_norm": 1.7120982404617975, |
| "learning_rate": 4.478694893850593e-07, |
| "loss": 0.2362, |
| "step": 9536 |
| }, |
| { |
| "epoch": 1.75, |
| "grad_norm": 1.412290074923546, |
| "learning_rate": 4.4449245546615753e-07, |
| "loss": 0.2483, |
| "step": 9568 |
| }, |
| { |
| "epoch": 1.75, |
| "grad_norm": 1.2470698026793279, |
| "learning_rate": 4.411179833400619e-07, |
| "loss": 0.2432, |
| "step": 9600 |
| }, |
| { |
| "epoch": 1.76, |
| "grad_norm": 1.268429893461802, |
| "learning_rate": 4.377462287459337e-07, |
| "loss": 0.2428, |
| "step": 9632 |
| }, |
| { |
| "epoch": 1.76, |
| "grad_norm": 1.2455579800684395, |
| "learning_rate": 4.343773472975139e-07, |
| "loss": 0.2508, |
| "step": 9664 |
| }, |
| { |
| "epoch": 1.77, |
| "grad_norm": 1.228328713009981, |
| "learning_rate": 4.3101149447594163e-07, |
| "loss": 0.2551, |
| "step": 9696 |
| }, |
| { |
| "epoch": 1.77, |
| "grad_norm": 1.2042503922389745, |
| "learning_rate": 4.276488256225787e-07, |
| "loss": 0.2541, |
| "step": 9728 |
| }, |
| { |
| "epoch": 1.78, |
| "grad_norm": 1.2346699498129847, |
| "learning_rate": 4.242894959318395e-07, |
| "loss": 0.2461, |
| "step": 9760 |
| }, |
| { |
| "epoch": 1.79, |
| "grad_norm": 1.3489561185627454, |
| "learning_rate": 4.209336604440294e-07, |
| "loss": 0.2477, |
| "step": 9792 |
| }, |
| { |
| "epoch": 1.79, |
| "grad_norm": 1.6713827739937024, |
| "learning_rate": 4.175814740381879e-07, |
| "loss": 0.2458, |
| "step": 9824 |
| }, |
| { |
| "epoch": 1.8, |
| "grad_norm": 1.3796544902936816, |
| "learning_rate": 4.1423309142494236e-07, |
| "loss": 0.2517, |
| "step": 9856 |
| }, |
| { |
| "epoch": 1.8, |
| "grad_norm": 1.6289046211602507, |
| "learning_rate": 4.1088866713936633e-07, |
| "loss": 0.2411, |
| "step": 9888 |
| }, |
| { |
| "epoch": 1.81, |
| "grad_norm": 1.4906278025262827, |
| "learning_rate": 4.0754835553384864e-07, |
| "loss": 0.2425, |
| "step": 9920 |
| }, |
| { |
| "epoch": 1.82, |
| "grad_norm": 1.1884900865492198, |
| "learning_rate": 4.0421231077096844e-07, |
| "loss": 0.2587, |
| "step": 9952 |
| }, |
| { |
| "epoch": 1.82, |
| "grad_norm": 1.329339394262938, |
| "learning_rate": 4.008806868163814e-07, |
| "loss": 0.2391, |
| "step": 9984 |
| }, |
| { |
| "epoch": 1.83, |
| "grad_norm": 1.2186411235918218, |
| "learning_rate": 3.9755363743171265e-07, |
| "loss": 0.253, |
| "step": 10016 |
| }, |
| { |
| "epoch": 1.83, |
| "grad_norm": 2.0559660633265047, |
| "learning_rate": 3.9423131616746187e-07, |
| "loss": 0.2566, |
| "step": 10048 |
| }, |
| { |
| "epoch": 1.84, |
| "grad_norm": 1.255138693515569, |
| "learning_rate": 3.9091387635591533e-07, |
| "loss": 0.2404, |
| "step": 10080 |
| }, |
| { |
| "epoch": 1.84, |
| "grad_norm": 1.5524707973995229, |
| "learning_rate": 3.8760147110406995e-07, |
| "loss": 0.2519, |
| "step": 10112 |
| }, |
| { |
| "epoch": 1.85, |
| "grad_norm": 1.2244273291645265, |
| "learning_rate": 3.8429425328656674e-07, |
| "loss": 0.2388, |
| "step": 10144 |
| }, |
| { |
| "epoch": 1.86, |
| "grad_norm": 1.438386617275495, |
| "learning_rate": 3.809923755386355e-07, |
| "loss": 0.25, |
| "step": 10176 |
| }, |
| { |
| "epoch": 1.86, |
| "grad_norm": 1.5262712808405177, |
| "learning_rate": 3.7769599024905004e-07, |
| "loss": 0.2299, |
| "step": 10208 |
| }, |
| { |
| "epoch": 1.87, |
| "grad_norm": 1.0721875435698476, |
| "learning_rate": 3.744052495530959e-07, |
| "loss": 0.2375, |
| "step": 10240 |
| }, |
| { |
| "epoch": 1.87, |
| "grad_norm": 1.1626282358875817, |
| "learning_rate": 3.7112030532554806e-07, |
| "loss": 0.2497, |
| "step": 10272 |
| }, |
| { |
| "epoch": 1.88, |
| "grad_norm": 1.3954184625846942, |
| "learning_rate": 3.6784130917366195e-07, |
| "loss": 0.242, |
| "step": 10304 |
| }, |
| { |
| "epoch": 1.89, |
| "grad_norm": 1.3259939974154864, |
| "learning_rate": 3.64568412430177e-07, |
| "loss": 0.2516, |
| "step": 10336 |
| }, |
| { |
| "epoch": 1.89, |
| "grad_norm": 1.73086784272557, |
| "learning_rate": 3.613017661463313e-07, |
| "loss": 0.2483, |
| "step": 10368 |
| }, |
| { |
| "epoch": 1.9, |
| "grad_norm": 1.188961288223144, |
| "learning_rate": 3.5804152108489065e-07, |
| "loss": 0.2391, |
| "step": 10400 |
| }, |
| { |
| "epoch": 1.9, |
| "grad_norm": 1.1246493355621157, |
| "learning_rate": 3.547878277131913e-07, |
| "loss": 0.2383, |
| "step": 10432 |
| }, |
| { |
| "epoch": 1.91, |
| "grad_norm": 1.059964023169643, |
| "learning_rate": 3.515408361961941e-07, |
| "loss": 0.2404, |
| "step": 10464 |
| }, |
| { |
| "epoch": 1.91, |
| "grad_norm": 1.70561149747635, |
| "learning_rate": 3.483006963895555e-07, |
| "loss": 0.2452, |
| "step": 10496 |
| }, |
| { |
| "epoch": 1.92, |
| "grad_norm": 1.2969563700400484, |
| "learning_rate": 3.4506755783271045e-07, |
| "loss": 0.2324, |
| "step": 10528 |
| }, |
| { |
| "epoch": 1.93, |
| "grad_norm": 1.2846755096320104, |
| "learning_rate": 3.418415697419712e-07, |
| "loss": 0.244, |
| "step": 10560 |
| }, |
| { |
| "epoch": 1.93, |
| "grad_norm": 1.7188757741977305, |
| "learning_rate": 3.386228810036408e-07, |
| "loss": 0.2369, |
| "step": 10592 |
| }, |
| { |
| "epoch": 1.94, |
| "grad_norm": 1.362077199933714, |
| "learning_rate": 3.3541164016714065e-07, |
| "loss": 0.2458, |
| "step": 10624 |
| }, |
| { |
| "epoch": 1.94, |
| "grad_norm": 1.354641295840275, |
| "learning_rate": 3.3220799543815634e-07, |
| "loss": 0.2397, |
| "step": 10656 |
| }, |
| { |
| "epoch": 1.95, |
| "grad_norm": 1.3277536016021265, |
| "learning_rate": 3.2901209467179637e-07, |
| "loss": 0.2458, |
| "step": 10688 |
| }, |
| { |
| "epoch": 1.96, |
| "grad_norm": 1.3886060047726545, |
| "learning_rate": 3.2582408536576877e-07, |
| "loss": 0.2344, |
| "step": 10720 |
| }, |
| { |
| "epoch": 1.96, |
| "grad_norm": 1.3601591408784948, |
| "learning_rate": 3.2264411465357335e-07, |
| "loss": 0.2343, |
| "step": 10752 |
| }, |
| { |
| "epoch": 1.97, |
| "grad_norm": 1.1017573737076742, |
| "learning_rate": 3.194723292977123e-07, |
| "loss": 0.2442, |
| "step": 10784 |
| }, |
| { |
| "epoch": 1.97, |
| "grad_norm": 1.4213916209193176, |
| "learning_rate": 3.1630887568291465e-07, |
| "loss": 0.2364, |
| "step": 10816 |
| }, |
| { |
| "epoch": 1.98, |
| "grad_norm": 1.3035008616357522, |
| "learning_rate": 3.131538998093828e-07, |
| "loss": 0.2458, |
| "step": 10848 |
| }, |
| { |
| "epoch": 1.99, |
| "grad_norm": 1.218412032577656, |
| "learning_rate": 3.1000754728605256e-07, |
| "loss": 0.2413, |
| "step": 10880 |
| }, |
| { |
| "epoch": 1.99, |
| "grad_norm": 1.2325820939927188, |
| "learning_rate": 3.068699633238738e-07, |
| "loss": 0.2511, |
| "step": 10912 |
| }, |
| { |
| "epoch": 2.0, |
| "grad_norm": 1.5401850018534633, |
| "learning_rate": 3.037412927291083e-07, |
| "loss": 0.2461, |
| "step": 10944 |
| }, |
| { |
| "epoch": 2.0, |
| "grad_norm": 1.3586152389713673, |
| "learning_rate": 3.006216798966468e-07, |
| "loss": 0.2358, |
| "step": 10976 |
| }, |
| { |
| "epoch": 2.01, |
| "grad_norm": 1.4011774529220307, |
| "learning_rate": 2.975112688033442e-07, |
| "loss": 0.2169, |
| "step": 11008 |
| }, |
| { |
| "epoch": 2.01, |
| "grad_norm": 1.2241897731555507, |
| "learning_rate": 2.944102030013763e-07, |
| "loss": 0.2108, |
| "step": 11040 |
| }, |
| { |
| "epoch": 2.02, |
| "grad_norm": 1.4027474158874935, |
| "learning_rate": 2.9131862561161293e-07, |
| "loss": 0.2293, |
| "step": 11072 |
| }, |
| { |
| "epoch": 2.03, |
| "grad_norm": 1.4670817911133367, |
| "learning_rate": 2.882366793170133e-07, |
| "loss": 0.2244, |
| "step": 11104 |
| }, |
| { |
| "epoch": 2.03, |
| "grad_norm": 1.3153219602513937, |
| "learning_rate": 2.8516450635604086e-07, |
| "loss": 0.22, |
| "step": 11136 |
| }, |
| { |
| "epoch": 2.04, |
| "grad_norm": 1.2976761125367386, |
| "learning_rate": 2.821022485160991e-07, |
| "loss": 0.2097, |
| "step": 11168 |
| }, |
| { |
| "epoch": 2.04, |
| "grad_norm": 1.3625162799781199, |
| "learning_rate": 2.7905004712698645e-07, |
| "loss": 0.2199, |
| "step": 11200 |
| }, |
| { |
| "epoch": 2.05, |
| "grad_norm": 1.232550115651448, |
| "learning_rate": 2.7600804305437474e-07, |
| "loss": 0.2239, |
| "step": 11232 |
| }, |
| { |
| "epoch": 2.06, |
| "grad_norm": 1.4796643786771864, |
| "learning_rate": 2.7297637669330847e-07, |
| "loss": 0.222, |
| "step": 11264 |
| }, |
| { |
| "epoch": 2.06, |
| "grad_norm": 1.2946736681294901, |
| "learning_rate": 2.699551879617235e-07, |
| "loss": 0.2246, |
| "step": 11296 |
| }, |
| { |
| "epoch": 2.07, |
| "grad_norm": 1.2068473533560813, |
| "learning_rate": 2.669446162939909e-07, |
| "loss": 0.2229, |
| "step": 11328 |
| }, |
| { |
| "epoch": 2.07, |
| "grad_norm": 1.4805315280685087, |
| "learning_rate": 2.639448006344812e-07, |
| "loss": 0.2131, |
| "step": 11360 |
| }, |
| { |
| "epoch": 2.08, |
| "grad_norm": 1.1871137311489093, |
| "learning_rate": 2.6095587943115227e-07, |
| "loss": 0.2235, |
| "step": 11392 |
| }, |
| { |
| "epoch": 2.08, |
| "grad_norm": 1.088896036136479, |
| "learning_rate": 2.57977990629159e-07, |
| "loss": 0.2073, |
| "step": 11424 |
| }, |
| { |
| "epoch": 2.09, |
| "grad_norm": 1.2960712344910867, |
| "learning_rate": 2.550112716644875e-07, |
| "loss": 0.2197, |
| "step": 11456 |
| }, |
| { |
| "epoch": 2.1, |
| "grad_norm": 1.2204092718307542, |
| "learning_rate": 2.520558594576115e-07, |
| "loss": 0.2282, |
| "step": 11488 |
| }, |
| { |
| "epoch": 2.1, |
| "grad_norm": 1.2693877027082432, |
| "learning_rate": 2.4911189040717357e-07, |
| "loss": 0.225, |
| "step": 11520 |
| }, |
| { |
| "epoch": 2.11, |
| "grad_norm": 1.2846435895566266, |
| "learning_rate": 2.4617950038369035e-07, |
| "loss": 0.2234, |
| "step": 11552 |
| }, |
| { |
| "epoch": 2.11, |
| "grad_norm": 1.2777271226205853, |
| "learning_rate": 2.432588247232811e-07, |
| "loss": 0.2211, |
| "step": 11584 |
| }, |
| { |
| "epoch": 2.12, |
| "grad_norm": 1.528372242529096, |
| "learning_rate": 2.4034999822142166e-07, |
| "loss": 0.22, |
| "step": 11616 |
| }, |
| { |
| "epoch": 2.13, |
| "grad_norm": 1.3339829276713877, |
| "learning_rate": 2.3745315512672397e-07, |
| "loss": 0.2195, |
| "step": 11648 |
| }, |
| { |
| "epoch": 2.13, |
| "grad_norm": 1.3101509979598953, |
| "learning_rate": 2.345684291347403e-07, |
| "loss": 0.2159, |
| "step": 11680 |
| }, |
| { |
| "epoch": 2.14, |
| "grad_norm": 1.3071577674707573, |
| "learning_rate": 2.3169595338179192e-07, |
| "loss": 0.2295, |
| "step": 11712 |
| }, |
| { |
| "epoch": 2.14, |
| "grad_norm": 1.3186809416500118, |
| "learning_rate": 2.2883586043882548e-07, |
| "loss": 0.2133, |
| "step": 11744 |
| }, |
| { |
| "epoch": 2.15, |
| "grad_norm": 1.5867314938269408, |
| "learning_rate": 2.259882823052941e-07, |
| "loss": 0.2316, |
| "step": 11776 |
| }, |
| { |
| "epoch": 2.15, |
| "grad_norm": 1.1557431014242034, |
| "learning_rate": 2.2315335040306554e-07, |
| "loss": 0.2189, |
| "step": 11808 |
| }, |
| { |
| "epoch": 2.16, |
| "grad_norm": 1.2736258434323458, |
| "learning_rate": 2.203311955703568e-07, |
| "loss": 0.2199, |
| "step": 11840 |
| }, |
| { |
| "epoch": 2.17, |
| "grad_norm": 1.302216300423092, |
| "learning_rate": 2.1752194805569552e-07, |
| "loss": 0.229, |
| "step": 11872 |
| }, |
| { |
| "epoch": 2.17, |
| "grad_norm": 1.3128201457087052, |
| "learning_rate": 2.147257375119087e-07, |
| "loss": 0.236, |
| "step": 11904 |
| }, |
| { |
| "epoch": 2.18, |
| "grad_norm": 1.3748881212985564, |
| "learning_rate": 2.119426929901391e-07, |
| "loss": 0.2206, |
| "step": 11936 |
| }, |
| { |
| "epoch": 2.18, |
| "grad_norm": 2.0921207372413435, |
| "learning_rate": 2.0917294293388966e-07, |
| "loss": 0.2117, |
| "step": 11968 |
| }, |
| { |
| "epoch": 2.19, |
| "grad_norm": 1.2949118526726222, |
| "learning_rate": 2.0641661517309417e-07, |
| "loss": 0.2135, |
| "step": 12000 |
| }, |
| { |
| "epoch": 2.2, |
| "grad_norm": 1.3415735498836403, |
| "learning_rate": 2.03673836918219e-07, |
| "loss": 0.2259, |
| "step": 12032 |
| }, |
| { |
| "epoch": 2.2, |
| "grad_norm": 1.55429350265861, |
| "learning_rate": 2.00944734754392e-07, |
| "loss": 0.2279, |
| "step": 12064 |
| }, |
| { |
| "epoch": 2.21, |
| "grad_norm": 1.2935837103909433, |
| "learning_rate": 1.9822943463555948e-07, |
| "loss": 0.2308, |
| "step": 12096 |
| }, |
| { |
| "epoch": 2.21, |
| "grad_norm": 1.642310611966749, |
| "learning_rate": 1.9552806187867372e-07, |
| "loss": 0.2144, |
| "step": 12128 |
| }, |
| { |
| "epoch": 2.22, |
| "grad_norm": 1.2465970815408216, |
| "learning_rate": 1.9284074115790932e-07, |
| "loss": 0.2198, |
| "step": 12160 |
| }, |
| { |
| "epoch": 2.22, |
| "grad_norm": 1.6894912323452993, |
| "learning_rate": 1.9016759649890895e-07, |
| "loss": 0.2147, |
| "step": 12192 |
| }, |
| { |
| "epoch": 2.23, |
| "grad_norm": 1.1571001402696899, |
| "learning_rate": 1.8750875127305977e-07, |
| "loss": 0.2149, |
| "step": 12224 |
| }, |
| { |
| "epoch": 2.24, |
| "grad_norm": 1.311070700271489, |
| "learning_rate": 1.8486432819179892e-07, |
| "loss": 0.2135, |
| "step": 12256 |
| }, |
| { |
| "epoch": 2.24, |
| "grad_norm": 1.358941411906666, |
| "learning_rate": 1.822344493009509e-07, |
| "loss": 0.2291, |
| "step": 12288 |
| }, |
| { |
| "epoch": 2.25, |
| "grad_norm": 1.2572126595123219, |
| "learning_rate": 1.7961923597509387e-07, |
| "loss": 0.2253, |
| "step": 12320 |
| }, |
| { |
| "epoch": 2.25, |
| "grad_norm": 1.4204877616087515, |
| "learning_rate": 1.7701880891195942e-07, |
| "loss": 0.212, |
| "step": 12352 |
| }, |
| { |
| "epoch": 2.26, |
| "grad_norm": 1.3082239260119037, |
| "learning_rate": 1.744332881268607e-07, |
| "loss": 0.2258, |
| "step": 12384 |
| }, |
| { |
| "epoch": 2.27, |
| "grad_norm": 1.3818606595114133, |
| "learning_rate": 1.7186279294715382e-07, |
| "loss": 0.2186, |
| "step": 12416 |
| }, |
| { |
| "epoch": 2.27, |
| "grad_norm": 1.344313917268325, |
| "learning_rate": 1.6930744200673104e-07, |
| "loss": 0.2309, |
| "step": 12448 |
| }, |
| { |
| "epoch": 2.28, |
| "grad_norm": 1.2849788562306779, |
| "learning_rate": 1.667673532405457e-07, |
| "loss": 0.2119, |
| "step": 12480 |
| }, |
| { |
| "epoch": 2.28, |
| "grad_norm": 1.048703241818117, |
| "learning_rate": 1.6424264387916842e-07, |
| "loss": 0.2175, |
| "step": 12512 |
| }, |
| { |
| "epoch": 2.29, |
| "grad_norm": 1.643116827546219, |
| "learning_rate": 1.6173343044337734e-07, |
| "loss": 0.2264, |
| "step": 12544 |
| }, |
| { |
| "epoch": 2.29, |
| "grad_norm": 1.2754388516106783, |
| "learning_rate": 1.5923982873878012e-07, |
| "loss": 0.2296, |
| "step": 12576 |
| }, |
| { |
| "epoch": 2.3, |
| "grad_norm": 1.2217192268295975, |
| "learning_rate": 1.567619538504696e-07, |
| "loss": 0.2281, |
| "step": 12608 |
| }, |
| { |
| "epoch": 2.31, |
| "grad_norm": 1.4770776525365248, |
| "learning_rate": 1.542999201377119e-07, |
| "loss": 0.2277, |
| "step": 12640 |
| }, |
| { |
| "epoch": 2.31, |
| "grad_norm": 1.642064071179986, |
| "learning_rate": 1.5185384122866897e-07, |
| "loss": 0.218, |
| "step": 12672 |
| }, |
| { |
| "epoch": 2.32, |
| "grad_norm": 1.237204280837322, |
| "learning_rate": 1.4942383001515386e-07, |
| "loss": 0.2079, |
| "step": 12704 |
| }, |
| { |
| "epoch": 2.32, |
| "grad_norm": 1.1715182674707414, |
| "learning_rate": 1.4700999864742136e-07, |
| "loss": 0.2124, |
| "step": 12736 |
| }, |
| { |
| "epoch": 2.33, |
| "grad_norm": 1.770494939875621, |
| "learning_rate": 1.4461245852899128e-07, |
| "loss": 0.2046, |
| "step": 12768 |
| }, |
| { |
| "epoch": 2.34, |
| "grad_norm": 1.2181121499687988, |
| "learning_rate": 1.4223132031150676e-07, |
| "loss": 0.2179, |
| "step": 12800 |
| }, |
| { |
| "epoch": 2.34, |
| "grad_norm": 1.2327845797316905, |
| "learning_rate": 1.3986669388962819e-07, |
| "loss": 0.2202, |
| "step": 12832 |
| }, |
| { |
| "epoch": 2.35, |
| "grad_norm": 1.5192364095540325, |
| "learning_rate": 1.375186883959613e-07, |
| "loss": 0.2187, |
| "step": 12864 |
| }, |
| { |
| "epoch": 2.35, |
| "grad_norm": 1.1447367117093976, |
| "learning_rate": 1.3518741219601988e-07, |
| "loss": 0.2267, |
| "step": 12896 |
| }, |
| { |
| "epoch": 2.36, |
| "grad_norm": 1.374708110513197, |
| "learning_rate": 1.3287297288322458e-07, |
| "loss": 0.2192, |
| "step": 12928 |
| }, |
| { |
| "epoch": 2.36, |
| "grad_norm": 1.1547706075025737, |
| "learning_rate": 1.305754772739377e-07, |
| "loss": 0.2101, |
| "step": 12960 |
| }, |
| { |
| "epoch": 2.37, |
| "grad_norm": 1.350330342693555, |
| "learning_rate": 1.2829503140253296e-07, |
| "loss": 0.2169, |
| "step": 12992 |
| }, |
| { |
| "epoch": 2.38, |
| "grad_norm": 1.416668233872815, |
| "learning_rate": 1.2603174051650207e-07, |
| "loss": 0.2181, |
| "step": 13024 |
| }, |
| { |
| "epoch": 2.38, |
| "grad_norm": 1.2796831245124296, |
| "learning_rate": 1.2378570907159697e-07, |
| "loss": 0.2264, |
| "step": 13056 |
| }, |
| { |
| "epoch": 2.39, |
| "grad_norm": 1.5438983104227264, |
| "learning_rate": 1.215570407270095e-07, |
| "loss": 0.2233, |
| "step": 13088 |
| }, |
| { |
| "epoch": 2.39, |
| "grad_norm": 1.7803269270190754, |
| "learning_rate": 1.1934583834058658e-07, |
| "loss": 0.2241, |
| "step": 13120 |
| }, |
| { |
| "epoch": 2.4, |
| "grad_norm": 1.5471141725466007, |
| "learning_rate": 1.1715220396408426e-07, |
| "loss": 0.2124, |
| "step": 13152 |
| }, |
| { |
| "epoch": 2.41, |
| "grad_norm": 1.1206628903783873, |
| "learning_rate": 1.1497623883845614e-07, |
| "loss": 0.2127, |
| "step": 13184 |
| }, |
| { |
| "epoch": 2.41, |
| "grad_norm": 2.5095477739496967, |
| "learning_rate": 1.1281804338918239e-07, |
| "loss": 0.2282, |
| "step": 13216 |
| }, |
| { |
| "epoch": 2.42, |
| "grad_norm": 1.3128197458740387, |
| "learning_rate": 1.1067771722163404e-07, |
| "loss": 0.2185, |
| "step": 13248 |
| }, |
| { |
| "epoch": 2.42, |
| "grad_norm": 1.2471858850247612, |
| "learning_rate": 1.0855535911647673e-07, |
| "loss": 0.2185, |
| "step": 13280 |
| }, |
| { |
| "epoch": 2.43, |
| "grad_norm": 1.655671236852773, |
| "learning_rate": 1.064510670251108e-07, |
| "loss": 0.2246, |
| "step": 13312 |
| }, |
| { |
| "epoch": 2.43, |
| "grad_norm": 1.2165519973679901, |
| "learning_rate": 1.0436493806515128e-07, |
| "loss": 0.223, |
| "step": 13344 |
| }, |
| { |
| "epoch": 2.44, |
| "grad_norm": 1.1834223459074134, |
| "learning_rate": 1.0229706851594561e-07, |
| "loss": 0.2188, |
| "step": 13376 |
| }, |
| { |
| "epoch": 2.45, |
| "grad_norm": 1.5035939672617133, |
| "learning_rate": 1.0024755381412997e-07, |
| "loss": 0.2204, |
| "step": 13408 |
| }, |
| { |
| "epoch": 2.45, |
| "grad_norm": 1.362765323113097, |
| "learning_rate": 9.821648854922481e-08, |
| "loss": 0.2165, |
| "step": 13440 |
| }, |
| { |
| "epoch": 2.46, |
| "grad_norm": 1.4329165569522946, |
| "learning_rate": 9.620396645926937e-08, |
| "loss": 0.2104, |
| "step": 13472 |
| }, |
| { |
| "epoch": 2.46, |
| "grad_norm": 1.2794493917175387, |
| "learning_rate": 9.421008042649514e-08, |
| "loss": 0.2225, |
| "step": 13504 |
| }, |
| { |
| "epoch": 2.47, |
| "grad_norm": 1.365469705607083, |
| "learning_rate": 9.223492247303999e-08, |
| "loss": 0.2166, |
| "step": 13536 |
| }, |
| { |
| "epoch": 2.48, |
| "grad_norm": 1.3658393916273586, |
| "learning_rate": 9.027858375670005e-08, |
| "loss": 0.2244, |
| "step": 13568 |
| }, |
| { |
| "epoch": 2.48, |
| "grad_norm": 1.6134388977737084, |
| "learning_rate": 8.834115456672314e-08, |
| "loss": 0.2238, |
| "step": 13600 |
| }, |
| { |
| "epoch": 2.49, |
| "grad_norm": 1.2747001650429752, |
| "learning_rate": 8.642272431964171e-08, |
| "loss": 0.2117, |
| "step": 13632 |
| }, |
| { |
| "epoch": 2.49, |
| "grad_norm": 1.4495915191793898, |
| "learning_rate": 8.452338155514644e-08, |
| "loss": 0.2153, |
| "step": 13664 |
| }, |
| { |
| "epoch": 2.5, |
| "grad_norm": 1.704239796998396, |
| "learning_rate": 8.264321393199902e-08, |
| "loss": 0.214, |
| "step": 13696 |
| }, |
| { |
| "epoch": 2.5, |
| "grad_norm": 1.1540455435637955, |
| "learning_rate": 8.078230822398735e-08, |
| "loss": 0.2189, |
| "step": 13728 |
| }, |
| { |
| "epoch": 2.51, |
| "grad_norm": 1.4264481560916344, |
| "learning_rate": 7.894075031592029e-08, |
| "loss": 0.2142, |
| "step": 13760 |
| }, |
| { |
| "epoch": 2.52, |
| "grad_norm": 1.2771622178109003, |
| "learning_rate": 7.711862519966416e-08, |
| "loss": 0.2238, |
| "step": 13792 |
| }, |
| { |
| "epoch": 2.52, |
| "grad_norm": 1.4176109439678235, |
| "learning_rate": 7.531601697022e-08, |
| "loss": 0.2141, |
| "step": 13824 |
| }, |
| { |
| "epoch": 2.53, |
| "grad_norm": 1.211872373307708, |
| "learning_rate": 7.353300882184238e-08, |
| "loss": 0.2094, |
| "step": 13856 |
| }, |
| { |
| "epoch": 2.53, |
| "grad_norm": 1.6208090587491066, |
| "learning_rate": 7.176968304420006e-08, |
| "loss": 0.213, |
| "step": 13888 |
| }, |
| { |
| "epoch": 2.54, |
| "grad_norm": 1.4294182508163742, |
| "learning_rate": 7.002612101857769e-08, |
| "loss": 0.2206, |
| "step": 13920 |
| }, |
| { |
| "epoch": 2.55, |
| "grad_norm": 1.5238939549044208, |
| "learning_rate": 6.830240321412067e-08, |
| "loss": 0.2187, |
| "step": 13952 |
| }, |
| { |
| "epoch": 2.55, |
| "grad_norm": 1.4817796496161193, |
| "learning_rate": 6.65986091841203e-08, |
| "loss": 0.223, |
| "step": 13984 |
| }, |
| { |
| "epoch": 2.56, |
| "grad_norm": 1.4428543645745906, |
| "learning_rate": 6.491481756234296e-08, |
| "loss": 0.2178, |
| "step": 14016 |
| }, |
| { |
| "epoch": 2.56, |
| "grad_norm": 1.3986608111205634, |
| "learning_rate": 6.32511060594007e-08, |
| "loss": 0.2244, |
| "step": 14048 |
| }, |
| { |
| "epoch": 2.57, |
| "grad_norm": 1.1779478103336172, |
| "learning_rate": 6.160755145916518e-08, |
| "loss": 0.2326, |
| "step": 14080 |
| }, |
| { |
| "epoch": 2.57, |
| "grad_norm": 1.4661784823721757, |
| "learning_rate": 5.99842296152231e-08, |
| "loss": 0.2257, |
| "step": 14112 |
| }, |
| { |
| "epoch": 2.58, |
| "grad_norm": 1.3997989298044267, |
| "learning_rate": 5.8381215447376006e-08, |
| "loss": 0.218, |
| "step": 14144 |
| }, |
| { |
| "epoch": 2.59, |
| "grad_norm": 1.1966150734513934, |
| "learning_rate": 5.679858293818257e-08, |
| "loss": 0.2139, |
| "step": 14176 |
| }, |
| { |
| "epoch": 2.59, |
| "grad_norm": 1.2975301799355876, |
| "learning_rate": 5.523640512954375e-08, |
| "loss": 0.2298, |
| "step": 14208 |
| }, |
| { |
| "epoch": 2.6, |
| "grad_norm": 1.1520772784392557, |
| "learning_rate": 5.369475411933222e-08, |
| "loss": 0.2086, |
| "step": 14240 |
| }, |
| { |
| "epoch": 2.6, |
| "grad_norm": 1.4920063917973507, |
| "learning_rate": 5.2173701058064454e-08, |
| "loss": 0.215, |
| "step": 14272 |
| }, |
| { |
| "epoch": 2.61, |
| "grad_norm": 1.4418059613129284, |
| "learning_rate": 5.067331614561726e-08, |
| "loss": 0.2108, |
| "step": 14304 |
| }, |
| { |
| "epoch": 2.62, |
| "grad_norm": 1.122573143556837, |
| "learning_rate": 4.9193668627988073e-08, |
| "loss": 0.2156, |
| "step": 14336 |
| }, |
| { |
| "epoch": 2.62, |
| "grad_norm": 1.169213395405146, |
| "learning_rate": 4.7734826794098624e-08, |
| "loss": 0.2155, |
| "step": 14368 |
| }, |
| { |
| "epoch": 2.63, |
| "grad_norm": 1.4739891927117594, |
| "learning_rate": 4.629685797264338e-08, |
| "loss": 0.2197, |
| "step": 14400 |
| }, |
| { |
| "epoch": 2.63, |
| "grad_norm": 1.1061529994285666, |
| "learning_rate": 4.487982852898248e-08, |
| "loss": 0.2161, |
| "step": 14432 |
| }, |
| { |
| "epoch": 2.64, |
| "grad_norm": 1.2665429377042383, |
| "learning_rate": 4.348380386207884e-08, |
| "loss": 0.2293, |
| "step": 14464 |
| }, |
| { |
| "epoch": 2.64, |
| "grad_norm": 1.4251977400136069, |
| "learning_rate": 4.210884840147927e-08, |
| "loss": 0.219, |
| "step": 14496 |
| }, |
| { |
| "epoch": 2.65, |
| "grad_norm": 1.3125295629014255, |
| "learning_rate": 4.075502560434158e-08, |
| "loss": 0.2215, |
| "step": 14528 |
| }, |
| { |
| "epoch": 2.66, |
| "grad_norm": 1.6114888474605462, |
| "learning_rate": 3.942239795250546e-08, |
| "loss": 0.2141, |
| "step": 14560 |
| }, |
| { |
| "epoch": 2.66, |
| "grad_norm": 1.3415946209746852, |
| "learning_rate": 3.8111026949609045e-08, |
| "loss": 0.2145, |
| "step": 14592 |
| }, |
| { |
| "epoch": 2.67, |
| "grad_norm": 1.5169187816349006, |
| "learning_rate": 3.682097311825033e-08, |
| "loss": 0.2241, |
| "step": 14624 |
| }, |
| { |
| "epoch": 2.67, |
| "grad_norm": 1.3176284198519965, |
| "learning_rate": 3.5552295997193834e-08, |
| "loss": 0.2238, |
| "step": 14656 |
| }, |
| { |
| "epoch": 2.68, |
| "grad_norm": 1.5611137018278904, |
| "learning_rate": 3.430505413862284e-08, |
| "loss": 0.2029, |
| "step": 14688 |
| }, |
| { |
| "epoch": 2.69, |
| "grad_norm": 1.4814939062116772, |
| "learning_rate": 3.307930510543705e-08, |
| "loss": 0.2073, |
| "step": 14720 |
| }, |
| { |
| "epoch": 2.69, |
| "grad_norm": 1.4585161010377092, |
| "learning_rate": 3.187510546859623e-08, |
| "loss": 0.2158, |
| "step": 14752 |
| }, |
| { |
| "epoch": 2.7, |
| "grad_norm": 1.3111001801067392, |
| "learning_rate": 3.069251080450863e-08, |
| "loss": 0.2114, |
| "step": 14784 |
| }, |
| { |
| "epoch": 2.7, |
| "grad_norm": 1.3748879882896838, |
| "learning_rate": 2.953157569246656e-08, |
| "loss": 0.2191, |
| "step": 14816 |
| }, |
| { |
| "epoch": 2.71, |
| "grad_norm": 1.2458399054987248, |
| "learning_rate": 2.8392353712127615e-08, |
| "loss": 0.2226, |
| "step": 14848 |
| }, |
| { |
| "epoch": 2.71, |
| "grad_norm": 1.183457902413645, |
| "learning_rate": 2.7274897441041223e-08, |
| "loss": 0.2276, |
| "step": 14880 |
| }, |
| { |
| "epoch": 2.72, |
| "grad_norm": 1.3996675489557702, |
| "learning_rate": 2.6179258452222687e-08, |
| "loss": 0.2223, |
| "step": 14912 |
| }, |
| { |
| "epoch": 2.73, |
| "grad_norm": 1.3797514575768821, |
| "learning_rate": 2.5105487311772478e-08, |
| "loss": 0.2346, |
| "step": 14944 |
| }, |
| { |
| "epoch": 2.73, |
| "grad_norm": 1.3453274139513625, |
| "learning_rate": 2.4053633576542833e-08, |
| "loss": 0.2148, |
| "step": 14976 |
| }, |
| { |
| "epoch": 2.74, |
| "grad_norm": 1.4296834231597417, |
| "learning_rate": 2.3023745791850623e-08, |
| "loss": 0.2252, |
| "step": 15008 |
| }, |
| { |
| "epoch": 2.74, |
| "grad_norm": 1.5820604044938296, |
| "learning_rate": 2.201587148923656e-08, |
| "loss": 0.2203, |
| "step": 15040 |
| }, |
| { |
| "epoch": 2.75, |
| "grad_norm": 1.3679161695840207, |
| "learning_rate": 2.103005718427181e-08, |
| "loss": 0.2096, |
| "step": 15072 |
| }, |
| { |
| "epoch": 2.76, |
| "grad_norm": 1.2454498587173481, |
| "learning_rate": 2.0066348374411122e-08, |
| "loss": 0.228, |
| "step": 15104 |
| }, |
| { |
| "epoch": 2.76, |
| "grad_norm": 1.094395180196629, |
| "learning_rate": 1.912478953689306e-08, |
| "loss": 0.2183, |
| "step": 15136 |
| }, |
| { |
| "epoch": 2.77, |
| "grad_norm": 1.2797588189588773, |
| "learning_rate": 1.8205424126687187e-08, |
| "loss": 0.2218, |
| "step": 15168 |
| }, |
| { |
| "epoch": 2.77, |
| "grad_norm": 1.4533835000058788, |
| "learning_rate": 1.7308294574488347e-08, |
| "loss": 0.2226, |
| "step": 15200 |
| }, |
| { |
| "epoch": 2.78, |
| "grad_norm": 1.4492805126695452, |
| "learning_rate": 1.6433442284758902e-08, |
| "loss": 0.2274, |
| "step": 15232 |
| }, |
| { |
| "epoch": 2.78, |
| "grad_norm": 1.4350039162599688, |
| "learning_rate": 1.5580907633817475e-08, |
| "loss": 0.2282, |
| "step": 15264 |
| }, |
| { |
| "epoch": 2.79, |
| "grad_norm": 1.1810113931037964, |
| "learning_rate": 1.475072996797555e-08, |
| "loss": 0.2224, |
| "step": 15296 |
| }, |
| { |
| "epoch": 2.8, |
| "grad_norm": 1.1286561070757017, |
| "learning_rate": 1.3942947601721488e-08, |
| "loss": 0.215, |
| "step": 15328 |
| }, |
| { |
| "epoch": 2.8, |
| "grad_norm": 1.1871769686911617, |
| "learning_rate": 1.3157597815952382e-08, |
| "loss": 0.2214, |
| "step": 15360 |
| }, |
| { |
| "epoch": 2.81, |
| "grad_norm": 1.2452370435442608, |
| "learning_rate": 1.2394716856253374e-08, |
| "loss": 0.2122, |
| "step": 15392 |
| }, |
| { |
| "epoch": 2.81, |
| "grad_norm": 1.2518679025330413, |
| "learning_rate": 1.1654339931224887e-08, |
| "loss": 0.217, |
| "step": 15424 |
| }, |
| { |
| "epoch": 2.82, |
| "grad_norm": 1.252731324308369, |
| "learning_rate": 1.0936501210857651e-08, |
| "loss": 0.2317, |
| "step": 15456 |
| }, |
| { |
| "epoch": 2.83, |
| "grad_norm": 1.3386257390126015, |
| "learning_rate": 1.0241233824955509e-08, |
| "loss": 0.2256, |
| "step": 15488 |
| }, |
| { |
| "epoch": 2.83, |
| "grad_norm": 1.1313351017583362, |
| "learning_rate": 9.568569861606868e-09, |
| "loss": 0.2341, |
| "step": 15520 |
| }, |
| { |
| "epoch": 2.84, |
| "grad_norm": 1.4449940727667887, |
| "learning_rate": 8.918540365703331e-09, |
| "loss": 0.2166, |
| "step": 15552 |
| }, |
| { |
| "epoch": 2.84, |
| "grad_norm": 1.4530241900991612, |
| "learning_rate": 8.291175337506894e-09, |
| "loss": 0.2408, |
| "step": 15584 |
| }, |
| { |
| "epoch": 2.85, |
| "grad_norm": 1.2609793071770357, |
| "learning_rate": 7.68650373126567e-09, |
| "loss": 0.2046, |
| "step": 15616 |
| }, |
| { |
| "epoch": 2.85, |
| "grad_norm": 1.1344728672758062, |
| "learning_rate": 7.10455345387756e-09, |
| "loss": 0.2196, |
| "step": 15648 |
| }, |
| { |
| "epoch": 2.86, |
| "grad_norm": 1.5048112445610096, |
| "learning_rate": 6.545351363601959e-09, |
| "loss": 0.2214, |
| "step": 15680 |
| }, |
| { |
| "epoch": 2.87, |
| "grad_norm": 1.4050748695357576, |
| "learning_rate": 6.00892326882052e-09, |
| "loss": 0.2265, |
| "step": 15712 |
| }, |
| { |
| "epoch": 2.87, |
| "grad_norm": 1.1997477163250476, |
| "learning_rate": 5.495293926845834e-09, |
| "loss": 0.2163, |
| "step": 15744 |
| }, |
| { |
| "epoch": 2.88, |
| "grad_norm": 1.3144891186853807, |
| "learning_rate": 5.00448704277906e-09, |
| "loss": 0.224, |
| "step": 15776 |
| }, |
| { |
| "epoch": 2.88, |
| "grad_norm": 1.353851004486482, |
| "learning_rate": 4.536525268415581e-09, |
| "loss": 0.2127, |
| "step": 15808 |
| }, |
| { |
| "epoch": 2.89, |
| "grad_norm": 1.4192328708826423, |
| "learning_rate": 4.091430201199841e-09, |
| "loss": 0.221, |
| "step": 15840 |
| }, |
| { |
| "epoch": 2.9, |
| "grad_norm": 1.3274920629279923, |
| "learning_rate": 3.6692223832283653e-09, |
| "loss": 0.2137, |
| "step": 15872 |
| }, |
| { |
| "epoch": 2.9, |
| "grad_norm": 1.2096820045248131, |
| "learning_rate": 3.2699213003019588e-09, |
| "loss": 0.2164, |
| "step": 15904 |
| }, |
| { |
| "epoch": 2.91, |
| "grad_norm": 1.4949483678866275, |
| "learning_rate": 2.8935453810260413e-09, |
| "loss": 0.211, |
| "step": 15936 |
| }, |
| { |
| "epoch": 2.91, |
| "grad_norm": 1.4480264990764071, |
| "learning_rate": 2.5401119959606033e-09, |
| "loss": 0.2219, |
| "step": 15968 |
| }, |
| { |
| "epoch": 2.92, |
| "grad_norm": 1.3984591809802007, |
| "learning_rate": 2.209637456817848e-09, |
| "loss": 0.2187, |
| "step": 16000 |
| }, |
| { |
| "epoch": 2.93, |
| "grad_norm": 1.2138235012481149, |
| "learning_rate": 1.9021370157100703e-09, |
| "loss": 0.2044, |
| "step": 16032 |
| }, |
| { |
| "epoch": 2.93, |
| "grad_norm": 1.4196275071259965, |
| "learning_rate": 1.6176248644453883e-09, |
| "loss": 0.2206, |
| "step": 16064 |
| }, |
| { |
| "epoch": 2.94, |
| "grad_norm": 1.4494340877185368, |
| "learning_rate": 1.3561141338726545e-09, |
| "loss": 0.2155, |
| "step": 16096 |
| }, |
| { |
| "epoch": 2.94, |
| "grad_norm": 1.3684818624338524, |
| "learning_rate": 1.117616893275719e-09, |
| "loss": 0.228, |
| "step": 16128 |
| }, |
| { |
| "epoch": 2.95, |
| "grad_norm": 4.300044658083074, |
| "learning_rate": 9.02144149816153e-10, |
| "loss": 0.2138, |
| "step": 16160 |
| }, |
| { |
| "epoch": 2.95, |
| "grad_norm": 1.4879340606023383, |
| "learning_rate": 7.097058480255991e-10, |
| "loss": 0.2135, |
| "step": 16192 |
| }, |
| { |
| "epoch": 2.96, |
| "grad_norm": 1.628757008031087, |
| "learning_rate": 5.403108693462499e-10, |
| "loss": 0.2169, |
| "step": 16224 |
| }, |
| { |
| "epoch": 2.97, |
| "grad_norm": 1.355596318247962, |
| "learning_rate": 3.9396703172150936e-10, |
| "loss": 0.212, |
| "step": 16256 |
| }, |
| { |
| "epoch": 2.97, |
| "grad_norm": 1.1638938575418796, |
| "learning_rate": 2.706810892348921e-10, |
| "loss": 0.2177, |
| "step": 16288 |
| }, |
| { |
| "epoch": 2.98, |
| "grad_norm": 1.2545697493468788, |
| "learning_rate": 1.704587317983286e-10, |
| "loss": 0.2226, |
| "step": 16320 |
| }, |
| { |
| "epoch": 2.98, |
| "grad_norm": 1.2958637026668016, |
| "learning_rate": 9.330458488959746e-11, |
| "loss": 0.2206, |
| "step": 16352 |
| }, |
| { |
| "epoch": 2.99, |
| "grad_norm": 1.1646489691129969, |
| "learning_rate": 3.922220933882947e-11, |
| "loss": 0.2219, |
| "step": 16384 |
| }, |
| { |
| "epoch": 3.0, |
| "grad_norm": 1.126290495132723, |
| "learning_rate": 8.214101164305543e-12, |
| "loss": 0.2393, |
| "step": 16416 |
| } |
| ], |
| "logging_steps": 32, |
| "max_steps": 16443, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 3, |
| "save_steps": 3289, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 0.0, |
| "train_batch_size": 1, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|