| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 1.0, |
| "eval_steps": 500, |
| "global_step": 3859, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.0002591344908007256, |
| "grad_norm": 111.01161666281861, |
| "learning_rate": 5.181347150259068e-07, |
| "loss": 12.3076, |
| "step": 1 |
| }, |
| { |
| "epoch": 0.001295672454003628, |
| "grad_norm": 128.99309976149883, |
| "learning_rate": 2.5906735751295338e-06, |
| "loss": 11.8519, |
| "step": 5 |
| }, |
| { |
| "epoch": 0.002591344908007256, |
| "grad_norm": 117.2928726132405, |
| "learning_rate": 5.1813471502590676e-06, |
| "loss": 12.3366, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.0038870173620108835, |
| "grad_norm": 89.9905121675766, |
| "learning_rate": 7.772020725388602e-06, |
| "loss": 11.7319, |
| "step": 15 |
| }, |
| { |
| "epoch": 0.005182689816014512, |
| "grad_norm": 52.319681120581, |
| "learning_rate": 1.0362694300518135e-05, |
| "loss": 8.7838, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.0064783622700181395, |
| "grad_norm": 38.24088557602701, |
| "learning_rate": 1.2953367875647668e-05, |
| "loss": 7.3055, |
| "step": 25 |
| }, |
| { |
| "epoch": 0.007774034724021767, |
| "grad_norm": 25.83067627606677, |
| "learning_rate": 1.5544041450777204e-05, |
| "loss": 5.7221, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.009069707178025395, |
| "grad_norm": 7.762285591416621, |
| "learning_rate": 1.813471502590674e-05, |
| "loss": 4.3566, |
| "step": 35 |
| }, |
| { |
| "epoch": 0.010365379632029024, |
| "grad_norm": 4.409457180338291, |
| "learning_rate": 2.072538860103627e-05, |
| "loss": 3.6268, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.011661052086032651, |
| "grad_norm": 2.601317803014797, |
| "learning_rate": 2.3316062176165805e-05, |
| "loss": 3.3055, |
| "step": 45 |
| }, |
| { |
| "epoch": 0.012956724540036279, |
| "grad_norm": 1.8400291034971308, |
| "learning_rate": 2.5906735751295337e-05, |
| "loss": 3.0095, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.014252396994039906, |
| "grad_norm": 1.4104791862031891, |
| "learning_rate": 2.8497409326424872e-05, |
| "loss": 2.7918, |
| "step": 55 |
| }, |
| { |
| "epoch": 0.015548069448043534, |
| "grad_norm": 1.4652220793947772, |
| "learning_rate": 3.108808290155441e-05, |
| "loss": 2.7691, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.01684374190204716, |
| "grad_norm": 2.1149598221315955, |
| "learning_rate": 3.367875647668394e-05, |
| "loss": 2.6915, |
| "step": 65 |
| }, |
| { |
| "epoch": 0.01813941435605079, |
| "grad_norm": 2.508562821895131, |
| "learning_rate": 3.626943005181348e-05, |
| "loss": 2.5864, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.019435086810054417, |
| "grad_norm": 5.312266133287117, |
| "learning_rate": 3.886010362694301e-05, |
| "loss": 2.3482, |
| "step": 75 |
| }, |
| { |
| "epoch": 0.020730759264058048, |
| "grad_norm": 6.635762805683308, |
| "learning_rate": 4.145077720207254e-05, |
| "loss": 2.0166, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.022026431718061675, |
| "grad_norm": 1.6009933261373699, |
| "learning_rate": 4.404145077720208e-05, |
| "loss": 1.5391, |
| "step": 85 |
| }, |
| { |
| "epoch": 0.023322104172065303, |
| "grad_norm": 1.390865987104561, |
| "learning_rate": 4.663212435233161e-05, |
| "loss": 1.3383, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.02461777662606893, |
| "grad_norm": 1.0505820893868818, |
| "learning_rate": 4.922279792746114e-05, |
| "loss": 1.2646, |
| "step": 95 |
| }, |
| { |
| "epoch": 0.025913449080072558, |
| "grad_norm": 1.0086607725155037, |
| "learning_rate": 5.1813471502590674e-05, |
| "loss": 1.3202, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.027209121534076185, |
| "grad_norm": 5.464697038037798, |
| "learning_rate": 5.440414507772021e-05, |
| "loss": 1.2451, |
| "step": 105 |
| }, |
| { |
| "epoch": 0.028504793988079813, |
| "grad_norm": 0.830545367098088, |
| "learning_rate": 5.6994818652849744e-05, |
| "loss": 1.2681, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.02980046644208344, |
| "grad_norm": 0.7132598923355695, |
| "learning_rate": 5.9585492227979276e-05, |
| "loss": 1.2245, |
| "step": 115 |
| }, |
| { |
| "epoch": 0.031096138896087068, |
| "grad_norm": 0.8138882694476683, |
| "learning_rate": 6.217616580310881e-05, |
| "loss": 1.2305, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.0323918113500907, |
| "grad_norm": 0.8807133122197233, |
| "learning_rate": 6.476683937823834e-05, |
| "loss": 1.2634, |
| "step": 125 |
| }, |
| { |
| "epoch": 0.03368748380409432, |
| "grad_norm": 0.62485091325362, |
| "learning_rate": 6.735751295336788e-05, |
| "loss": 1.3386, |
| "step": 130 |
| }, |
| { |
| "epoch": 0.034983156258097954, |
| "grad_norm": 0.7549667435061143, |
| "learning_rate": 6.994818652849742e-05, |
| "loss": 1.2513, |
| "step": 135 |
| }, |
| { |
| "epoch": 0.03627882871210158, |
| "grad_norm": 0.6420487687876172, |
| "learning_rate": 7.253886010362695e-05, |
| "loss": 1.2101, |
| "step": 140 |
| }, |
| { |
| "epoch": 0.03757450116610521, |
| "grad_norm": 0.7398721962367694, |
| "learning_rate": 7.512953367875648e-05, |
| "loss": 1.2776, |
| "step": 145 |
| }, |
| { |
| "epoch": 0.03887017362010883, |
| "grad_norm": 0.6450506158667421, |
| "learning_rate": 7.772020725388602e-05, |
| "loss": 1.2096, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.040165846074112464, |
| "grad_norm": 0.6506841169994266, |
| "learning_rate": 8.031088082901554e-05, |
| "loss": 1.228, |
| "step": 155 |
| }, |
| { |
| "epoch": 0.041461518528116095, |
| "grad_norm": 0.711359178146542, |
| "learning_rate": 8.290155440414508e-05, |
| "loss": 1.2474, |
| "step": 160 |
| }, |
| { |
| "epoch": 0.04275719098211972, |
| "grad_norm": 1.459357949005984, |
| "learning_rate": 8.549222797927462e-05, |
| "loss": 1.2251, |
| "step": 165 |
| }, |
| { |
| "epoch": 0.04405286343612335, |
| "grad_norm": 0.8207401278130507, |
| "learning_rate": 8.808290155440416e-05, |
| "loss": 1.2625, |
| "step": 170 |
| }, |
| { |
| "epoch": 0.045348535890126974, |
| "grad_norm": 0.6271337633480082, |
| "learning_rate": 9.067357512953368e-05, |
| "loss": 1.2222, |
| "step": 175 |
| }, |
| { |
| "epoch": 0.046644208344130605, |
| "grad_norm": 0.6465301721957318, |
| "learning_rate": 9.326424870466322e-05, |
| "loss": 1.2357, |
| "step": 180 |
| }, |
| { |
| "epoch": 0.04793988079813423, |
| "grad_norm": 0.5864341660869246, |
| "learning_rate": 9.585492227979275e-05, |
| "loss": 1.2422, |
| "step": 185 |
| }, |
| { |
| "epoch": 0.04923555325213786, |
| "grad_norm": 0.5644516524286168, |
| "learning_rate": 9.844559585492228e-05, |
| "loss": 1.1901, |
| "step": 190 |
| }, |
| { |
| "epoch": 0.050531225706141485, |
| "grad_norm": 0.654016063568062, |
| "learning_rate": 0.00010103626943005182, |
| "loss": 1.182, |
| "step": 195 |
| }, |
| { |
| "epoch": 0.051826898160145116, |
| "grad_norm": 0.569272168627719, |
| "learning_rate": 0.00010362694300518135, |
| "loss": 1.1958, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.05312257061414875, |
| "grad_norm": 0.6368680847682564, |
| "learning_rate": 0.00010621761658031089, |
| "loss": 1.2694, |
| "step": 205 |
| }, |
| { |
| "epoch": 0.05441824306815237, |
| "grad_norm": 0.5948244637123268, |
| "learning_rate": 0.00010880829015544042, |
| "loss": 1.1967, |
| "step": 210 |
| }, |
| { |
| "epoch": 0.055713915522156, |
| "grad_norm": 0.5541951524654064, |
| "learning_rate": 0.00011139896373056995, |
| "loss": 1.1941, |
| "step": 215 |
| }, |
| { |
| "epoch": 0.057009587976159626, |
| "grad_norm": 0.5701648418604164, |
| "learning_rate": 0.00011398963730569949, |
| "loss": 1.2443, |
| "step": 220 |
| }, |
| { |
| "epoch": 0.05830526043016326, |
| "grad_norm": 0.4865998338825169, |
| "learning_rate": 0.00011658031088082901, |
| "loss": 1.2106, |
| "step": 225 |
| }, |
| { |
| "epoch": 0.05960093288416688, |
| "grad_norm": 0.5131095922403084, |
| "learning_rate": 0.00011917098445595855, |
| "loss": 1.1278, |
| "step": 230 |
| }, |
| { |
| "epoch": 0.06089660533817051, |
| "grad_norm": 0.6563988864261227, |
| "learning_rate": 0.0001217616580310881, |
| "loss": 1.1991, |
| "step": 235 |
| }, |
| { |
| "epoch": 0.062192277792174136, |
| "grad_norm": 0.5769507376758363, |
| "learning_rate": 0.00012435233160621763, |
| "loss": 1.1674, |
| "step": 240 |
| }, |
| { |
| "epoch": 0.06348795024617776, |
| "grad_norm": 0.5283246008008212, |
| "learning_rate": 0.00012694300518134715, |
| "loss": 1.2367, |
| "step": 245 |
| }, |
| { |
| "epoch": 0.0647836227001814, |
| "grad_norm": 0.5536285028186603, |
| "learning_rate": 0.00012953367875647668, |
| "loss": 1.1649, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.06607929515418502, |
| "grad_norm": 0.5163473385172975, |
| "learning_rate": 0.00013212435233160623, |
| "loss": 1.2867, |
| "step": 255 |
| }, |
| { |
| "epoch": 0.06737496760818865, |
| "grad_norm": 0.5475813868100357, |
| "learning_rate": 0.00013471502590673575, |
| "loss": 1.2152, |
| "step": 260 |
| }, |
| { |
| "epoch": 0.06867064006219228, |
| "grad_norm": 0.511381640001362, |
| "learning_rate": 0.00013730569948186528, |
| "loss": 1.2065, |
| "step": 265 |
| }, |
| { |
| "epoch": 0.06996631251619591, |
| "grad_norm": 0.4982475298525502, |
| "learning_rate": 0.00013989637305699483, |
| "loss": 1.2142, |
| "step": 270 |
| }, |
| { |
| "epoch": 0.07126198497019953, |
| "grad_norm": 0.5433832176160214, |
| "learning_rate": 0.00014248704663212436, |
| "loss": 1.2276, |
| "step": 275 |
| }, |
| { |
| "epoch": 0.07255765742420316, |
| "grad_norm": 0.47834421596861043, |
| "learning_rate": 0.0001450777202072539, |
| "loss": 1.1183, |
| "step": 280 |
| }, |
| { |
| "epoch": 0.0738533298782068, |
| "grad_norm": 2.276620321866631, |
| "learning_rate": 0.0001476683937823834, |
| "loss": 1.2225, |
| "step": 285 |
| }, |
| { |
| "epoch": 0.07514900233221042, |
| "grad_norm": 0.5622928799874418, |
| "learning_rate": 0.00015025906735751296, |
| "loss": 1.142, |
| "step": 290 |
| }, |
| { |
| "epoch": 0.07644467478621404, |
| "grad_norm": 0.48315577966120404, |
| "learning_rate": 0.0001528497409326425, |
| "loss": 1.1481, |
| "step": 295 |
| }, |
| { |
| "epoch": 0.07774034724021767, |
| "grad_norm": 0.5054965652645259, |
| "learning_rate": 0.00015544041450777204, |
| "loss": 1.1334, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.0790360196942213, |
| "grad_norm": 0.5412094676646951, |
| "learning_rate": 0.00015803108808290156, |
| "loss": 1.2215, |
| "step": 305 |
| }, |
| { |
| "epoch": 0.08033169214822493, |
| "grad_norm": 0.489058036506764, |
| "learning_rate": 0.00016062176165803108, |
| "loss": 1.1651, |
| "step": 310 |
| }, |
| { |
| "epoch": 0.08162736460222855, |
| "grad_norm": 0.5225993706207775, |
| "learning_rate": 0.00016321243523316064, |
| "loss": 1.173, |
| "step": 315 |
| }, |
| { |
| "epoch": 0.08292303705623219, |
| "grad_norm": 0.4941052544991184, |
| "learning_rate": 0.00016580310880829016, |
| "loss": 1.1537, |
| "step": 320 |
| }, |
| { |
| "epoch": 0.08421870951023581, |
| "grad_norm": 0.5317455067919804, |
| "learning_rate": 0.0001683937823834197, |
| "loss": 1.1702, |
| "step": 325 |
| }, |
| { |
| "epoch": 0.08551438196423944, |
| "grad_norm": 0.5382129937077951, |
| "learning_rate": 0.00017098445595854924, |
| "loss": 1.1954, |
| "step": 330 |
| }, |
| { |
| "epoch": 0.08681005441824306, |
| "grad_norm": 0.4873382544458068, |
| "learning_rate": 0.00017357512953367876, |
| "loss": 1.1292, |
| "step": 335 |
| }, |
| { |
| "epoch": 0.0881057268722467, |
| "grad_norm": 0.869634547296972, |
| "learning_rate": 0.00017616580310880832, |
| "loss": 1.2342, |
| "step": 340 |
| }, |
| { |
| "epoch": 0.08940139932625032, |
| "grad_norm": 0.47105707267021146, |
| "learning_rate": 0.0001787564766839378, |
| "loss": 1.1903, |
| "step": 345 |
| }, |
| { |
| "epoch": 0.09069707178025395, |
| "grad_norm": 0.483761714583125, |
| "learning_rate": 0.00018134715025906737, |
| "loss": 1.1753, |
| "step": 350 |
| }, |
| { |
| "epoch": 0.09199274423425757, |
| "grad_norm": 0.5094212322859152, |
| "learning_rate": 0.00018393782383419692, |
| "loss": 1.1736, |
| "step": 355 |
| }, |
| { |
| "epoch": 0.09328841668826121, |
| "grad_norm": 0.49665577393117427, |
| "learning_rate": 0.00018652849740932644, |
| "loss": 1.19, |
| "step": 360 |
| }, |
| { |
| "epoch": 0.09458408914226483, |
| "grad_norm": 0.4634306042294334, |
| "learning_rate": 0.00018911917098445597, |
| "loss": 1.2068, |
| "step": 365 |
| }, |
| { |
| "epoch": 0.09587976159626846, |
| "grad_norm": 0.4771955076541271, |
| "learning_rate": 0.0001917098445595855, |
| "loss": 1.1746, |
| "step": 370 |
| }, |
| { |
| "epoch": 0.0971754340502721, |
| "grad_norm": 0.49488098544084536, |
| "learning_rate": 0.00019430051813471504, |
| "loss": 1.1781, |
| "step": 375 |
| }, |
| { |
| "epoch": 0.09847110650427572, |
| "grad_norm": 0.45559447847276796, |
| "learning_rate": 0.00019689119170984457, |
| "loss": 1.207, |
| "step": 380 |
| }, |
| { |
| "epoch": 0.09976677895827935, |
| "grad_norm": 0.530644388105199, |
| "learning_rate": 0.0001994818652849741, |
| "loss": 1.2059, |
| "step": 385 |
| }, |
| { |
| "epoch": 0.10106245141228297, |
| "grad_norm": 0.4262101606219483, |
| "learning_rate": 0.0001999993453944367, |
| "loss": 1.1728, |
| "step": 390 |
| }, |
| { |
| "epoch": 0.10235812386628661, |
| "grad_norm": 0.5019930710713301, |
| "learning_rate": 0.00019999668607402385, |
| "loss": 1.1566, |
| "step": 395 |
| }, |
| { |
| "epoch": 0.10365379632029023, |
| "grad_norm": 0.41212142912399047, |
| "learning_rate": 0.00019999198118027207, |
| "loss": 1.1367, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.10494946877429386, |
| "grad_norm": 0.4880776387603866, |
| "learning_rate": 0.00019998523080942663, |
| "loss": 1.1654, |
| "step": 405 |
| }, |
| { |
| "epoch": 0.1062451412282975, |
| "grad_norm": 0.46286032679323585, |
| "learning_rate": 0.00019997643509957582, |
| "loss": 1.1944, |
| "step": 410 |
| }, |
| { |
| "epoch": 0.10754081368230112, |
| "grad_norm": 0.5382364265983718, |
| "learning_rate": 0.00019996559423064838, |
| "loss": 1.1871, |
| "step": 415 |
| }, |
| { |
| "epoch": 0.10883648613630474, |
| "grad_norm": 0.5437136001752437, |
| "learning_rate": 0.0001999527084244095, |
| "loss": 1.1211, |
| "step": 420 |
| }, |
| { |
| "epoch": 0.11013215859030837, |
| "grad_norm": 0.4683209703143848, |
| "learning_rate": 0.00019993777794445662, |
| "loss": 1.2111, |
| "step": 425 |
| }, |
| { |
| "epoch": 0.111427831044312, |
| "grad_norm": 0.43839848429469386, |
| "learning_rate": 0.00019992080309621371, |
| "loss": 1.1655, |
| "step": 430 |
| }, |
| { |
| "epoch": 0.11272350349831563, |
| "grad_norm": 0.4352884542992481, |
| "learning_rate": 0.00019990178422692528, |
| "loss": 1.1674, |
| "step": 435 |
| }, |
| { |
| "epoch": 0.11401917595231925, |
| "grad_norm": 0.45552967972039005, |
| "learning_rate": 0.00019988072172564918, |
| "loss": 1.2005, |
| "step": 440 |
| }, |
| { |
| "epoch": 0.11531484840632288, |
| "grad_norm": 0.46569070576531457, |
| "learning_rate": 0.0001998576160232485, |
| "loss": 1.1611, |
| "step": 445 |
| }, |
| { |
| "epoch": 0.11661052086032651, |
| "grad_norm": 0.4344558621258368, |
| "learning_rate": 0.00019983246759238305, |
| "loss": 1.1632, |
| "step": 450 |
| }, |
| { |
| "epoch": 0.11790619331433014, |
| "grad_norm": 0.4524631212974695, |
| "learning_rate": 0.00019980527694749952, |
| "loss": 1.2003, |
| "step": 455 |
| }, |
| { |
| "epoch": 0.11920186576833376, |
| "grad_norm": 0.692041740038434, |
| "learning_rate": 0.00019977604464482083, |
| "loss": 1.233, |
| "step": 460 |
| }, |
| { |
| "epoch": 0.1204975382223374, |
| "grad_norm": 0.4499116414766127, |
| "learning_rate": 0.00019974477128233505, |
| "loss": 1.1431, |
| "step": 465 |
| }, |
| { |
| "epoch": 0.12179321067634102, |
| "grad_norm": 0.4755303917497374, |
| "learning_rate": 0.00019971145749978294, |
| "loss": 1.186, |
| "step": 470 |
| }, |
| { |
| "epoch": 0.12308888313034465, |
| "grad_norm": 0.4587427439073901, |
| "learning_rate": 0.00019967610397864493, |
| "loss": 1.1772, |
| "step": 475 |
| }, |
| { |
| "epoch": 0.12438455558434827, |
| "grad_norm": 0.43004396404922146, |
| "learning_rate": 0.0001996387114421272, |
| "loss": 1.1613, |
| "step": 480 |
| }, |
| { |
| "epoch": 0.1256802280383519, |
| "grad_norm": 0.4457323824719563, |
| "learning_rate": 0.0001995992806551468, |
| "loss": 1.2043, |
| "step": 485 |
| }, |
| { |
| "epoch": 0.12697590049235552, |
| "grad_norm": 0.46784030959656325, |
| "learning_rate": 0.00019955781242431622, |
| "loss": 1.1678, |
| "step": 490 |
| }, |
| { |
| "epoch": 0.12827157294635916, |
| "grad_norm": 0.4610982688278206, |
| "learning_rate": 0.00019951430759792654, |
| "loss": 1.2353, |
| "step": 495 |
| }, |
| { |
| "epoch": 0.1295672454003628, |
| "grad_norm": 0.4259792086854904, |
| "learning_rate": 0.0001994687670659305, |
| "loss": 1.1612, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.1308629178543664, |
| "grad_norm": 0.45520840895219594, |
| "learning_rate": 0.00019942119175992383, |
| "loss": 1.099, |
| "step": 505 |
| }, |
| { |
| "epoch": 0.13215859030837004, |
| "grad_norm": 0.4473123797320602, |
| "learning_rate": 0.00019937158265312667, |
| "loss": 1.2141, |
| "step": 510 |
| }, |
| { |
| "epoch": 0.13345426276237368, |
| "grad_norm": 0.4713319110562635, |
| "learning_rate": 0.00019931994076036324, |
| "loss": 1.1454, |
| "step": 515 |
| }, |
| { |
| "epoch": 0.1347499352163773, |
| "grad_norm": 0.5136825790567691, |
| "learning_rate": 0.00019926626713804137, |
| "loss": 1.2187, |
| "step": 520 |
| }, |
| { |
| "epoch": 0.13604560767038093, |
| "grad_norm": 0.4727966110942116, |
| "learning_rate": 0.00019921056288413076, |
| "loss": 1.1988, |
| "step": 525 |
| }, |
| { |
| "epoch": 0.13734128012438457, |
| "grad_norm": 0.4364604236158943, |
| "learning_rate": 0.00019915282913814052, |
| "loss": 1.1685, |
| "step": 530 |
| }, |
| { |
| "epoch": 0.13863695257838818, |
| "grad_norm": 0.48078362848188844, |
| "learning_rate": 0.00019909306708109585, |
| "loss": 1.1773, |
| "step": 535 |
| }, |
| { |
| "epoch": 0.13993262503239182, |
| "grad_norm": 0.45186495929915804, |
| "learning_rate": 0.00019903127793551408, |
| "loss": 1.1409, |
| "step": 540 |
| }, |
| { |
| "epoch": 0.14122829748639543, |
| "grad_norm": 0.4557721517395367, |
| "learning_rate": 0.0001989674629653793, |
| "loss": 1.1243, |
| "step": 545 |
| }, |
| { |
| "epoch": 0.14252396994039906, |
| "grad_norm": 0.47938538143807347, |
| "learning_rate": 0.00019890162347611687, |
| "loss": 1.1644, |
| "step": 550 |
| }, |
| { |
| "epoch": 0.1438196423944027, |
| "grad_norm": 0.4541147395460418, |
| "learning_rate": 0.00019883376081456636, |
| "loss": 1.2578, |
| "step": 555 |
| }, |
| { |
| "epoch": 0.1451153148484063, |
| "grad_norm": 0.4780871852039313, |
| "learning_rate": 0.00019876387636895437, |
| "loss": 1.206, |
| "step": 560 |
| }, |
| { |
| "epoch": 0.14641098730240995, |
| "grad_norm": 0.42628778420968705, |
| "learning_rate": 0.00019869197156886586, |
| "loss": 1.1728, |
| "step": 565 |
| }, |
| { |
| "epoch": 0.1477066597564136, |
| "grad_norm": 0.4303201965277035, |
| "learning_rate": 0.00019861804788521493, |
| "loss": 1.1807, |
| "step": 570 |
| }, |
| { |
| "epoch": 0.1490023322104172, |
| "grad_norm": 0.48120812526739415, |
| "learning_rate": 0.00019854210683021485, |
| "loss": 1.1825, |
| "step": 575 |
| }, |
| { |
| "epoch": 0.15029800466442084, |
| "grad_norm": 0.479571534871543, |
| "learning_rate": 0.0001984641499573472, |
| "loss": 1.1213, |
| "step": 580 |
| }, |
| { |
| "epoch": 0.15159367711842447, |
| "grad_norm": 0.4517330978325847, |
| "learning_rate": 0.0001983841788613297, |
| "loss": 1.1799, |
| "step": 585 |
| }, |
| { |
| "epoch": 0.15288934957242808, |
| "grad_norm": 0.45662716170150547, |
| "learning_rate": 0.00019830219517808404, |
| "loss": 1.2154, |
| "step": 590 |
| }, |
| { |
| "epoch": 0.15418502202643172, |
| "grad_norm": 0.46419741725194397, |
| "learning_rate": 0.00019821820058470215, |
| "loss": 1.2244, |
| "step": 595 |
| }, |
| { |
| "epoch": 0.15548069448043533, |
| "grad_norm": 0.48643943236229664, |
| "learning_rate": 0.00019813219679941203, |
| "loss": 1.1773, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.15677636693443897, |
| "grad_norm": 0.4213189862656036, |
| "learning_rate": 0.00019804418558154243, |
| "loss": 1.1918, |
| "step": 605 |
| }, |
| { |
| "epoch": 0.1580720393884426, |
| "grad_norm": 0.4039792149279076, |
| "learning_rate": 0.00019795416873148703, |
| "loss": 1.1937, |
| "step": 610 |
| }, |
| { |
| "epoch": 0.15936771184244622, |
| "grad_norm": 0.42408763347068035, |
| "learning_rate": 0.00019786214809066753, |
| "loss": 1.1792, |
| "step": 615 |
| }, |
| { |
| "epoch": 0.16066338429644986, |
| "grad_norm": 0.4418941392259797, |
| "learning_rate": 0.00019776812554149603, |
| "loss": 1.0983, |
| "step": 620 |
| }, |
| { |
| "epoch": 0.1619590567504535, |
| "grad_norm": 0.44778137756258257, |
| "learning_rate": 0.00019767210300733647, |
| "loss": 1.1441, |
| "step": 625 |
| }, |
| { |
| "epoch": 0.1632547292044571, |
| "grad_norm": 0.4214693702892699, |
| "learning_rate": 0.0001975740824524653, |
| "loss": 1.126, |
| "step": 630 |
| }, |
| { |
| "epoch": 0.16455040165846074, |
| "grad_norm": 0.431868681760054, |
| "learning_rate": 0.00019747406588203128, |
| "loss": 1.2244, |
| "step": 635 |
| }, |
| { |
| "epoch": 0.16584607411246438, |
| "grad_norm": 0.41521488211670027, |
| "learning_rate": 0.00019737205534201455, |
| "loss": 1.1443, |
| "step": 640 |
| }, |
| { |
| "epoch": 0.167141746566468, |
| "grad_norm": 0.4141916874944073, |
| "learning_rate": 0.00019726805291918464, |
| "loss": 1.2012, |
| "step": 645 |
| }, |
| { |
| "epoch": 0.16843741902047163, |
| "grad_norm": 0.8019720904500388, |
| "learning_rate": 0.0001971620607410579, |
| "loss": 1.1871, |
| "step": 650 |
| }, |
| { |
| "epoch": 0.16973309147447524, |
| "grad_norm": 0.44302123796014203, |
| "learning_rate": 0.00019705408097585393, |
| "loss": 1.217, |
| "step": 655 |
| }, |
| { |
| "epoch": 0.17102876392847888, |
| "grad_norm": 0.44266067140812565, |
| "learning_rate": 0.0001969441158324512, |
| "loss": 1.159, |
| "step": 660 |
| }, |
| { |
| "epoch": 0.17232443638248252, |
| "grad_norm": 0.44647716533971304, |
| "learning_rate": 0.0001968321675603419, |
| "loss": 1.1653, |
| "step": 665 |
| }, |
| { |
| "epoch": 0.17362010883648613, |
| "grad_norm": 0.44623301312377617, |
| "learning_rate": 0.000196718238449586, |
| "loss": 1.2078, |
| "step": 670 |
| }, |
| { |
| "epoch": 0.17491578129048976, |
| "grad_norm": 0.41505140877535734, |
| "learning_rate": 0.00019660233083076416, |
| "loss": 1.1727, |
| "step": 675 |
| }, |
| { |
| "epoch": 0.1762114537444934, |
| "grad_norm": 0.4603289406918021, |
| "learning_rate": 0.00019648444707493035, |
| "loss": 1.2151, |
| "step": 680 |
| }, |
| { |
| "epoch": 0.177507126198497, |
| "grad_norm": 0.4185225256437822, |
| "learning_rate": 0.00019636458959356316, |
| "loss": 1.1708, |
| "step": 685 |
| }, |
| { |
| "epoch": 0.17880279865250065, |
| "grad_norm": 0.4207300329188764, |
| "learning_rate": 0.00019624276083851655, |
| "loss": 1.1821, |
| "step": 690 |
| }, |
| { |
| "epoch": 0.1800984711065043, |
| "grad_norm": 0.44173498052158916, |
| "learning_rate": 0.00019611896330196956, |
| "loss": 1.175, |
| "step": 695 |
| }, |
| { |
| "epoch": 0.1813941435605079, |
| "grad_norm": 0.43648924778909787, |
| "learning_rate": 0.0001959931995163756, |
| "loss": 1.1409, |
| "step": 700 |
| }, |
| { |
| "epoch": 0.18268981601451154, |
| "grad_norm": 0.41607673192874983, |
| "learning_rate": 0.00019586547205441038, |
| "loss": 1.1903, |
| "step": 705 |
| }, |
| { |
| "epoch": 0.18398548846851515, |
| "grad_norm": 0.40413543802190666, |
| "learning_rate": 0.00019573578352891937, |
| "loss": 1.1395, |
| "step": 710 |
| }, |
| { |
| "epoch": 0.18528116092251878, |
| "grad_norm": 0.45031653314971887, |
| "learning_rate": 0.00019560413659286437, |
| "loss": 1.1551, |
| "step": 715 |
| }, |
| { |
| "epoch": 0.18657683337652242, |
| "grad_norm": 0.4308169448866955, |
| "learning_rate": 0.00019547053393926934, |
| "loss": 1.1875, |
| "step": 720 |
| }, |
| { |
| "epoch": 0.18787250583052603, |
| "grad_norm": 0.42706739288180534, |
| "learning_rate": 0.000195334978301165, |
| "loss": 1.1828, |
| "step": 725 |
| }, |
| { |
| "epoch": 0.18916817828452967, |
| "grad_norm": 0.43201723960587324, |
| "learning_rate": 0.00019519747245153333, |
| "loss": 1.1269, |
| "step": 730 |
| }, |
| { |
| "epoch": 0.1904638507385333, |
| "grad_norm": 0.45113312626512714, |
| "learning_rate": 0.0001950580192032505, |
| "loss": 1.228, |
| "step": 735 |
| }, |
| { |
| "epoch": 0.19175952319253692, |
| "grad_norm": 0.44161985380394386, |
| "learning_rate": 0.0001949166214090295, |
| "loss": 1.1242, |
| "step": 740 |
| }, |
| { |
| "epoch": 0.19305519564654056, |
| "grad_norm": 0.46536800518090093, |
| "learning_rate": 0.00019477328196136178, |
| "loss": 1.205, |
| "step": 745 |
| }, |
| { |
| "epoch": 0.1943508681005442, |
| "grad_norm": 0.4304870726759157, |
| "learning_rate": 0.00019462800379245807, |
| "loss": 1.1677, |
| "step": 750 |
| }, |
| { |
| "epoch": 0.1956465405545478, |
| "grad_norm": 0.4500472206954455, |
| "learning_rate": 0.0001944807898741883, |
| "loss": 1.2031, |
| "step": 755 |
| }, |
| { |
| "epoch": 0.19694221300855144, |
| "grad_norm": 0.4270521213996621, |
| "learning_rate": 0.00019433164321802095, |
| "loss": 1.221, |
| "step": 760 |
| }, |
| { |
| "epoch": 0.19823788546255505, |
| "grad_norm": 0.4500335364742716, |
| "learning_rate": 0.00019418056687496135, |
| "loss": 1.2206, |
| "step": 765 |
| }, |
| { |
| "epoch": 0.1995335579165587, |
| "grad_norm": 0.46760662103580936, |
| "learning_rate": 0.00019402756393548936, |
| "loss": 1.2628, |
| "step": 770 |
| }, |
| { |
| "epoch": 0.20082923037056233, |
| "grad_norm": 0.4374227196524263, |
| "learning_rate": 0.00019387263752949598, |
| "loss": 1.1895, |
| "step": 775 |
| }, |
| { |
| "epoch": 0.20212490282456594, |
| "grad_norm": 0.44423980739831237, |
| "learning_rate": 0.00019371579082621952, |
| "loss": 1.1915, |
| "step": 780 |
| }, |
| { |
| "epoch": 0.20342057527856958, |
| "grad_norm": 0.5067585361087695, |
| "learning_rate": 0.00019355702703418063, |
| "loss": 1.1396, |
| "step": 785 |
| }, |
| { |
| "epoch": 0.20471624773257321, |
| "grad_norm": 0.4397713592175906, |
| "learning_rate": 0.0001933963494011168, |
| "loss": 1.1829, |
| "step": 790 |
| }, |
| { |
| "epoch": 0.20601192018657682, |
| "grad_norm": 0.41142670533283804, |
| "learning_rate": 0.0001932337612139157, |
| "loss": 1.1403, |
| "step": 795 |
| }, |
| { |
| "epoch": 0.20730759264058046, |
| "grad_norm": 0.431969506585272, |
| "learning_rate": 0.00019306926579854821, |
| "loss": 1.1219, |
| "step": 800 |
| }, |
| { |
| "epoch": 0.2086032650945841, |
| "grad_norm": 0.44265249198309475, |
| "learning_rate": 0.00019290286652000018, |
| "loss": 1.1897, |
| "step": 805 |
| }, |
| { |
| "epoch": 0.2098989375485877, |
| "grad_norm": 0.4277657316516301, |
| "learning_rate": 0.0001927345667822037, |
| "loss": 1.1746, |
| "step": 810 |
| }, |
| { |
| "epoch": 0.21119461000259135, |
| "grad_norm": 0.40123037504170683, |
| "learning_rate": 0.00019256437002796744, |
| "loss": 1.1775, |
| "step": 815 |
| }, |
| { |
| "epoch": 0.212490282456595, |
| "grad_norm": 0.4250319228234625, |
| "learning_rate": 0.00019239227973890622, |
| "loss": 1.1357, |
| "step": 820 |
| }, |
| { |
| "epoch": 0.2137859549105986, |
| "grad_norm": 0.469142216595277, |
| "learning_rate": 0.0001922182994353697, |
| "loss": 1.1709, |
| "step": 825 |
| }, |
| { |
| "epoch": 0.21508162736460223, |
| "grad_norm": 0.4205186026441914, |
| "learning_rate": 0.0001920424326763706, |
| "loss": 1.1846, |
| "step": 830 |
| }, |
| { |
| "epoch": 0.21637729981860584, |
| "grad_norm": 0.4364780340562057, |
| "learning_rate": 0.00019186468305951165, |
| "loss": 1.1611, |
| "step": 835 |
| }, |
| { |
| "epoch": 0.21767297227260948, |
| "grad_norm": 0.42862545463693674, |
| "learning_rate": 0.00019168505422091214, |
| "loss": 1.2604, |
| "step": 840 |
| }, |
| { |
| "epoch": 0.21896864472661312, |
| "grad_norm": 0.4298577312506422, |
| "learning_rate": 0.00019150354983513346, |
| "loss": 1.1584, |
| "step": 845 |
| }, |
| { |
| "epoch": 0.22026431718061673, |
| "grad_norm": 0.4188255164345959, |
| "learning_rate": 0.00019132017361510396, |
| "loss": 1.1681, |
| "step": 850 |
| }, |
| { |
| "epoch": 0.22155998963462037, |
| "grad_norm": 0.42610008321144127, |
| "learning_rate": 0.00019113492931204304, |
| "loss": 1.2502, |
| "step": 855 |
| }, |
| { |
| "epoch": 0.222855662088624, |
| "grad_norm": 0.41917044829804134, |
| "learning_rate": 0.00019094782071538434, |
| "loss": 1.1441, |
| "step": 860 |
| }, |
| { |
| "epoch": 0.22415133454262762, |
| "grad_norm": 0.41937866972534293, |
| "learning_rate": 0.0001907588516526983, |
| "loss": 1.2056, |
| "step": 865 |
| }, |
| { |
| "epoch": 0.22544700699663126, |
| "grad_norm": 0.47221544899406714, |
| "learning_rate": 0.00019056802598961376, |
| "loss": 1.1754, |
| "step": 870 |
| }, |
| { |
| "epoch": 0.2267426794506349, |
| "grad_norm": 0.45128209262811003, |
| "learning_rate": 0.000190375347629739, |
| "loss": 1.1963, |
| "step": 875 |
| }, |
| { |
| "epoch": 0.2280383519046385, |
| "grad_norm": 0.41254887942116875, |
| "learning_rate": 0.00019018082051458176, |
| "loss": 1.1696, |
| "step": 880 |
| }, |
| { |
| "epoch": 0.22933402435864214, |
| "grad_norm": 0.4336535335825639, |
| "learning_rate": 0.00018998444862346873, |
| "loss": 1.1942, |
| "step": 885 |
| }, |
| { |
| "epoch": 0.23062969681264575, |
| "grad_norm": 0.4203068249880088, |
| "learning_rate": 0.00018978623597346408, |
| "loss": 1.1571, |
| "step": 890 |
| }, |
| { |
| "epoch": 0.2319253692666494, |
| "grad_norm": 0.4504420384371143, |
| "learning_rate": 0.00018958618661928732, |
| "loss": 1.2145, |
| "step": 895 |
| }, |
| { |
| "epoch": 0.23322104172065303, |
| "grad_norm": 1.2852982602777483, |
| "learning_rate": 0.00018938430465323034, |
| "loss": 1.1602, |
| "step": 900 |
| }, |
| { |
| "epoch": 0.23451671417465664, |
| "grad_norm": 0.4456499671770626, |
| "learning_rate": 0.0001891805942050736, |
| "loss": 1.2208, |
| "step": 905 |
| }, |
| { |
| "epoch": 0.23581238662866028, |
| "grad_norm": 0.4218488577208689, |
| "learning_rate": 0.00018897505944200186, |
| "loss": 1.1544, |
| "step": 910 |
| }, |
| { |
| "epoch": 0.2371080590826639, |
| "grad_norm": 0.4509293570117875, |
| "learning_rate": 0.00018876770456851877, |
| "loss": 1.2167, |
| "step": 915 |
| }, |
| { |
| "epoch": 0.23840373153666752, |
| "grad_norm": 0.4354264628463739, |
| "learning_rate": 0.00018855853382636093, |
| "loss": 1.2103, |
| "step": 920 |
| }, |
| { |
| "epoch": 0.23969940399067116, |
| "grad_norm": 0.41330759489370955, |
| "learning_rate": 0.00018834755149441104, |
| "loss": 1.1713, |
| "step": 925 |
| }, |
| { |
| "epoch": 0.2409950764446748, |
| "grad_norm": 0.43531782144147513, |
| "learning_rate": 0.00018813476188861043, |
| "loss": 1.1783, |
| "step": 930 |
| }, |
| { |
| "epoch": 0.2422907488986784, |
| "grad_norm": 0.4260507531562403, |
| "learning_rate": 0.00018792016936187086, |
| "loss": 1.1473, |
| "step": 935 |
| }, |
| { |
| "epoch": 0.24358642135268205, |
| "grad_norm": 0.46161844432613697, |
| "learning_rate": 0.00018770377830398525, |
| "loss": 1.1719, |
| "step": 940 |
| }, |
| { |
| "epoch": 0.24488209380668566, |
| "grad_norm": 0.44108410378556556, |
| "learning_rate": 0.00018748559314153818, |
| "loss": 1.2101, |
| "step": 945 |
| }, |
| { |
| "epoch": 0.2461777662606893, |
| "grad_norm": 0.43144293315888466, |
| "learning_rate": 0.00018726561833781497, |
| "loss": 1.1933, |
| "step": 950 |
| }, |
| { |
| "epoch": 0.24747343871469293, |
| "grad_norm": 0.437185606075459, |
| "learning_rate": 0.00018704385839271074, |
| "loss": 1.1563, |
| "step": 955 |
| }, |
| { |
| "epoch": 0.24876911116869654, |
| "grad_norm": 0.44187547375302677, |
| "learning_rate": 0.00018682031784263814, |
| "loss": 1.1219, |
| "step": 960 |
| }, |
| { |
| "epoch": 0.2500647836227002, |
| "grad_norm": 0.4286003185934626, |
| "learning_rate": 0.00018659500126043456, |
| "loss": 1.1347, |
| "step": 965 |
| }, |
| { |
| "epoch": 0.2513604560767038, |
| "grad_norm": 0.5193197943854574, |
| "learning_rate": 0.00018636791325526872, |
| "loss": 1.2118, |
| "step": 970 |
| }, |
| { |
| "epoch": 0.25265612853070746, |
| "grad_norm": 0.4243801823353042, |
| "learning_rate": 0.00018613905847254624, |
| "loss": 1.2015, |
| "step": 975 |
| }, |
| { |
| "epoch": 0.25395180098471104, |
| "grad_norm": 0.4017619629276402, |
| "learning_rate": 0.0001859084415938147, |
| "loss": 1.1405, |
| "step": 980 |
| }, |
| { |
| "epoch": 0.2552474734387147, |
| "grad_norm": 0.46780432324474447, |
| "learning_rate": 0.00018567606733666775, |
| "loss": 1.1803, |
| "step": 985 |
| }, |
| { |
| "epoch": 0.2565431458927183, |
| "grad_norm": 0.41305152773119796, |
| "learning_rate": 0.00018544194045464886, |
| "loss": 1.138, |
| "step": 990 |
| }, |
| { |
| "epoch": 0.25783881834672195, |
| "grad_norm": 0.4090708833399176, |
| "learning_rate": 0.0001852060657371538, |
| "loss": 1.2036, |
| "step": 995 |
| }, |
| { |
| "epoch": 0.2591344908007256, |
| "grad_norm": 0.43431398541973126, |
| "learning_rate": 0.00018496844800933277, |
| "loss": 1.1328, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.26043016325472923, |
| "grad_norm": 0.4473287652132595, |
| "learning_rate": 0.0001847290921319918, |
| "loss": 1.183, |
| "step": 1005 |
| }, |
| { |
| "epoch": 0.2617258357087328, |
| "grad_norm": 0.42139497138720455, |
| "learning_rate": 0.00018448800300149314, |
| "loss": 1.1885, |
| "step": 1010 |
| }, |
| { |
| "epoch": 0.26302150816273645, |
| "grad_norm": 0.4192742651155342, |
| "learning_rate": 0.00018424518554965516, |
| "loss": 1.1673, |
| "step": 1015 |
| }, |
| { |
| "epoch": 0.2643171806167401, |
| "grad_norm": 0.4276967282946324, |
| "learning_rate": 0.00018400064474365156, |
| "loss": 1.1442, |
| "step": 1020 |
| }, |
| { |
| "epoch": 0.2656128530707437, |
| "grad_norm": 0.4279959695376702, |
| "learning_rate": 0.00018375438558590967, |
| "loss": 1.1713, |
| "step": 1025 |
| }, |
| { |
| "epoch": 0.26690852552474736, |
| "grad_norm": 0.4138904922875282, |
| "learning_rate": 0.00018350641311400812, |
| "loss": 1.1294, |
| "step": 1030 |
| }, |
| { |
| "epoch": 0.26820419797875095, |
| "grad_norm": 0.43337813322172525, |
| "learning_rate": 0.0001832567324005737, |
| "loss": 1.1421, |
| "step": 1035 |
| }, |
| { |
| "epoch": 0.2694998704327546, |
| "grad_norm": 0.4418444877860619, |
| "learning_rate": 0.00018300534855317783, |
| "loss": 1.1775, |
| "step": 1040 |
| }, |
| { |
| "epoch": 0.2707955428867582, |
| "grad_norm": 0.46191806443855654, |
| "learning_rate": 0.00018275226671423195, |
| "loss": 1.1188, |
| "step": 1045 |
| }, |
| { |
| "epoch": 0.27209121534076186, |
| "grad_norm": 0.4449373885557454, |
| "learning_rate": 0.0001824974920608821, |
| "loss": 1.1445, |
| "step": 1050 |
| }, |
| { |
| "epoch": 0.2733868877947655, |
| "grad_norm": 0.4138240087897747, |
| "learning_rate": 0.0001822410298049035, |
| "loss": 1.1403, |
| "step": 1055 |
| }, |
| { |
| "epoch": 0.27468256024876914, |
| "grad_norm": 0.4410582459345502, |
| "learning_rate": 0.00018198288519259353, |
| "loss": 1.1835, |
| "step": 1060 |
| }, |
| { |
| "epoch": 0.2759782327027727, |
| "grad_norm": 0.4419431606975578, |
| "learning_rate": 0.0001817230635046645, |
| "loss": 1.2308, |
| "step": 1065 |
| }, |
| { |
| "epoch": 0.27727390515677636, |
| "grad_norm": 0.41612794192020086, |
| "learning_rate": 0.0001814615700561358, |
| "loss": 1.21, |
| "step": 1070 |
| }, |
| { |
| "epoch": 0.27856957761078, |
| "grad_norm": 0.4728861496942093, |
| "learning_rate": 0.00018119841019622487, |
| "loss": 1.1687, |
| "step": 1075 |
| }, |
| { |
| "epoch": 0.27986525006478363, |
| "grad_norm": 0.39483035476368555, |
| "learning_rate": 0.0001809335893082381, |
| "loss": 1.2031, |
| "step": 1080 |
| }, |
| { |
| "epoch": 0.28116092251878727, |
| "grad_norm": 0.40167341751112157, |
| "learning_rate": 0.0001806671128094605, |
| "loss": 1.1495, |
| "step": 1085 |
| }, |
| { |
| "epoch": 0.28245659497279085, |
| "grad_norm": 0.42702634032994197, |
| "learning_rate": 0.0001803989861510449, |
| "loss": 1.1885, |
| "step": 1090 |
| }, |
| { |
| "epoch": 0.2837522674267945, |
| "grad_norm": 0.3970503569404702, |
| "learning_rate": 0.00018012921481790054, |
| "loss": 1.1608, |
| "step": 1095 |
| }, |
| { |
| "epoch": 0.28504793988079813, |
| "grad_norm": 0.42637878811973734, |
| "learning_rate": 0.0001798578043285807, |
| "loss": 1.1386, |
| "step": 1100 |
| }, |
| { |
| "epoch": 0.28634361233480177, |
| "grad_norm": 0.39626715470331225, |
| "learning_rate": 0.00017958476023517008, |
| "loss": 1.1355, |
| "step": 1105 |
| }, |
| { |
| "epoch": 0.2876392847888054, |
| "grad_norm": 0.4158526138591002, |
| "learning_rate": 0.00017931008812317089, |
| "loss": 1.1844, |
| "step": 1110 |
| }, |
| { |
| "epoch": 0.28893495724280904, |
| "grad_norm": 0.4110580138866959, |
| "learning_rate": 0.00017903379361138884, |
| "loss": 1.179, |
| "step": 1115 |
| }, |
| { |
| "epoch": 0.2902306296968126, |
| "grad_norm": 0.39558148991812647, |
| "learning_rate": 0.0001787558823518181, |
| "loss": 1.1529, |
| "step": 1120 |
| }, |
| { |
| "epoch": 0.29152630215081626, |
| "grad_norm": 0.43235047146868255, |
| "learning_rate": 0.0001784763600295257, |
| "loss": 1.1253, |
| "step": 1125 |
| }, |
| { |
| "epoch": 0.2928219746048199, |
| "grad_norm": 0.4029217795192826, |
| "learning_rate": 0.00017819523236253524, |
| "loss": 1.1775, |
| "step": 1130 |
| }, |
| { |
| "epoch": 0.29411764705882354, |
| "grad_norm": 0.43737260937528, |
| "learning_rate": 0.0001779125051017099, |
| "loss": 1.1658, |
| "step": 1135 |
| }, |
| { |
| "epoch": 0.2954133195128272, |
| "grad_norm": 0.4223589589109796, |
| "learning_rate": 0.00017762818403063485, |
| "loss": 1.2193, |
| "step": 1140 |
| }, |
| { |
| "epoch": 0.29670899196683076, |
| "grad_norm": 0.42153072570772854, |
| "learning_rate": 0.0001773422749654988, |
| "loss": 1.1381, |
| "step": 1145 |
| }, |
| { |
| "epoch": 0.2980046644208344, |
| "grad_norm": 0.44333235395249593, |
| "learning_rate": 0.0001770547837549752, |
| "loss": 1.2025, |
| "step": 1150 |
| }, |
| { |
| "epoch": 0.29930033687483804, |
| "grad_norm": 0.390430213600016, |
| "learning_rate": 0.0001767657162801025, |
| "loss": 1.2179, |
| "step": 1155 |
| }, |
| { |
| "epoch": 0.3005960093288417, |
| "grad_norm": 0.39596734484692325, |
| "learning_rate": 0.00017647507845416392, |
| "loss": 1.1443, |
| "step": 1160 |
| }, |
| { |
| "epoch": 0.3018916817828453, |
| "grad_norm": 0.3964805626892672, |
| "learning_rate": 0.00017618287622256625, |
| "loss": 1.0953, |
| "step": 1165 |
| }, |
| { |
| "epoch": 0.30318735423684895, |
| "grad_norm": 0.397165519182473, |
| "learning_rate": 0.00017588911556271858, |
| "loss": 1.2468, |
| "step": 1170 |
| }, |
| { |
| "epoch": 0.30448302669085253, |
| "grad_norm": 0.40738383895968944, |
| "learning_rate": 0.00017559380248390982, |
| "loss": 1.2279, |
| "step": 1175 |
| }, |
| { |
| "epoch": 0.30577869914485617, |
| "grad_norm": 0.4322556749554571, |
| "learning_rate": 0.00017529694302718574, |
| "loss": 1.1614, |
| "step": 1180 |
| }, |
| { |
| "epoch": 0.3070743715988598, |
| "grad_norm": 0.3795921376766387, |
| "learning_rate": 0.0001749985432652255, |
| "loss": 1.1514, |
| "step": 1185 |
| }, |
| { |
| "epoch": 0.30837004405286345, |
| "grad_norm": 0.3984980575708157, |
| "learning_rate": 0.00017469860930221734, |
| "loss": 1.2262, |
| "step": 1190 |
| }, |
| { |
| "epoch": 0.3096657165068671, |
| "grad_norm": 0.4239852292999169, |
| "learning_rate": 0.00017439714727373378, |
| "loss": 1.1279, |
| "step": 1195 |
| }, |
| { |
| "epoch": 0.31096138896087067, |
| "grad_norm": 0.4243568518066299, |
| "learning_rate": 0.00017409416334660606, |
| "loss": 1.1482, |
| "step": 1200 |
| }, |
| { |
| "epoch": 0.3122570614148743, |
| "grad_norm": 0.39261780158026205, |
| "learning_rate": 0.00017378966371879803, |
| "loss": 1.1245, |
| "step": 1205 |
| }, |
| { |
| "epoch": 0.31355273386887794, |
| "grad_norm": 0.4129200331115499, |
| "learning_rate": 0.00017348365461927932, |
| "loss": 1.164, |
| "step": 1210 |
| }, |
| { |
| "epoch": 0.3148484063228816, |
| "grad_norm": 0.46767017383779913, |
| "learning_rate": 0.00017317614230789792, |
| "loss": 1.2038, |
| "step": 1215 |
| }, |
| { |
| "epoch": 0.3161440787768852, |
| "grad_norm": 0.4636682224687034, |
| "learning_rate": 0.00017286713307525212, |
| "loss": 1.1798, |
| "step": 1220 |
| }, |
| { |
| "epoch": 0.31743975123088886, |
| "grad_norm": 0.464530514696448, |
| "learning_rate": 0.00017255663324256194, |
| "loss": 1.1569, |
| "step": 1225 |
| }, |
| { |
| "epoch": 0.31873542368489244, |
| "grad_norm": 0.49392901042147025, |
| "learning_rate": 0.00017224464916153963, |
| "loss": 1.1722, |
| "step": 1230 |
| }, |
| { |
| "epoch": 0.3200310961388961, |
| "grad_norm": 0.437281931716071, |
| "learning_rate": 0.00017193118721425986, |
| "loss": 1.1472, |
| "step": 1235 |
| }, |
| { |
| "epoch": 0.3213267685928997, |
| "grad_norm": 0.49973298936543603, |
| "learning_rate": 0.00017161625381302914, |
| "loss": 1.1662, |
| "step": 1240 |
| }, |
| { |
| "epoch": 0.32262244104690335, |
| "grad_norm": 0.9030290188031315, |
| "learning_rate": 0.00017129985540025473, |
| "loss": 1.1751, |
| "step": 1245 |
| }, |
| { |
| "epoch": 0.323918113500907, |
| "grad_norm": 0.46932404302432773, |
| "learning_rate": 0.00017098199844831262, |
| "loss": 1.1753, |
| "step": 1250 |
| }, |
| { |
| "epoch": 0.3252137859549106, |
| "grad_norm": 0.43737397519531773, |
| "learning_rate": 0.0001706626894594154, |
| "loss": 1.1539, |
| "step": 1255 |
| }, |
| { |
| "epoch": 0.3265094584089142, |
| "grad_norm": 0.47678251178263875, |
| "learning_rate": 0.00017034193496547902, |
| "loss": 1.1572, |
| "step": 1260 |
| }, |
| { |
| "epoch": 0.32780513086291785, |
| "grad_norm": 0.48005221081289434, |
| "learning_rate": 0.00017001974152798942, |
| "loss": 1.2371, |
| "step": 1265 |
| }, |
| { |
| "epoch": 0.3291008033169215, |
| "grad_norm": 0.4005197274939468, |
| "learning_rate": 0.000169696115737868, |
| "loss": 1.1375, |
| "step": 1270 |
| }, |
| { |
| "epoch": 0.3303964757709251, |
| "grad_norm": 0.520470386619454, |
| "learning_rate": 0.00016937106421533707, |
| "loss": 1.2201, |
| "step": 1275 |
| }, |
| { |
| "epoch": 0.33169214822492876, |
| "grad_norm": 0.4024597397089223, |
| "learning_rate": 0.00016904459360978427, |
| "loss": 1.16, |
| "step": 1280 |
| }, |
| { |
| "epoch": 0.33298782067893234, |
| "grad_norm": 0.43690159855763916, |
| "learning_rate": 0.00016871671059962655, |
| "loss": 1.2303, |
| "step": 1285 |
| }, |
| { |
| "epoch": 0.334283493132936, |
| "grad_norm": 0.42841787432183803, |
| "learning_rate": 0.00016838742189217366, |
| "loss": 1.1601, |
| "step": 1290 |
| }, |
| { |
| "epoch": 0.3355791655869396, |
| "grad_norm": 0.41150940120312113, |
| "learning_rate": 0.00016805673422349082, |
| "loss": 1.1542, |
| "step": 1295 |
| }, |
| { |
| "epoch": 0.33687483804094326, |
| "grad_norm": 0.4108846561673421, |
| "learning_rate": 0.000167724654358261, |
| "loss": 1.1497, |
| "step": 1300 |
| }, |
| { |
| "epoch": 0.3381705104949469, |
| "grad_norm": 0.4485077675503156, |
| "learning_rate": 0.00016739118908964647, |
| "loss": 1.1979, |
| "step": 1305 |
| }, |
| { |
| "epoch": 0.3394661829489505, |
| "grad_norm": 0.40380367544471263, |
| "learning_rate": 0.00016705634523915, |
| "loss": 1.156, |
| "step": 1310 |
| }, |
| { |
| "epoch": 0.3407618554029541, |
| "grad_norm": 0.4315676369718417, |
| "learning_rate": 0.000166720129656475, |
| "loss": 1.1876, |
| "step": 1315 |
| }, |
| { |
| "epoch": 0.34205752785695775, |
| "grad_norm": 0.42298698967693915, |
| "learning_rate": 0.00016638254921938587, |
| "loss": 1.226, |
| "step": 1320 |
| }, |
| { |
| "epoch": 0.3433532003109614, |
| "grad_norm": 0.40711175836921143, |
| "learning_rate": 0.00016604361083356675, |
| "loss": 1.1865, |
| "step": 1325 |
| }, |
| { |
| "epoch": 0.34464887276496503, |
| "grad_norm": 0.4209373436004817, |
| "learning_rate": 0.0001657033214324807, |
| "loss": 1.1485, |
| "step": 1330 |
| }, |
| { |
| "epoch": 0.34594454521896867, |
| "grad_norm": 0.3973628247221751, |
| "learning_rate": 0.0001653616879772277, |
| "loss": 1.1419, |
| "step": 1335 |
| }, |
| { |
| "epoch": 0.34724021767297225, |
| "grad_norm": 0.40714636825697015, |
| "learning_rate": 0.00016501871745640213, |
| "loss": 1.1607, |
| "step": 1340 |
| }, |
| { |
| "epoch": 0.3485358901269759, |
| "grad_norm": 0.4091289111560899, |
| "learning_rate": 0.00016467441688595015, |
| "loss": 1.169, |
| "step": 1345 |
| }, |
| { |
| "epoch": 0.3498315625809795, |
| "grad_norm": 0.39294191213880614, |
| "learning_rate": 0.0001643287933090258, |
| "loss": 1.1856, |
| "step": 1350 |
| }, |
| { |
| "epoch": 0.35112723503498317, |
| "grad_norm": 0.4153756492643378, |
| "learning_rate": 0.00016398185379584707, |
| "loss": 1.1601, |
| "step": 1355 |
| }, |
| { |
| "epoch": 0.3524229074889868, |
| "grad_norm": 0.39772148869790713, |
| "learning_rate": 0.0001636336054435514, |
| "loss": 1.1402, |
| "step": 1360 |
| }, |
| { |
| "epoch": 0.3537185799429904, |
| "grad_norm": 0.4071165220940827, |
| "learning_rate": 0.00016328405537605032, |
| "loss": 1.1333, |
| "step": 1365 |
| }, |
| { |
| "epoch": 0.355014252396994, |
| "grad_norm": 0.4269465126269822, |
| "learning_rate": 0.00016293321074388375, |
| "loss": 1.1948, |
| "step": 1370 |
| }, |
| { |
| "epoch": 0.35630992485099766, |
| "grad_norm": 0.41865801202656294, |
| "learning_rate": 0.00016258107872407375, |
| "loss": 1.1465, |
| "step": 1375 |
| }, |
| { |
| "epoch": 0.3576055973050013, |
| "grad_norm": 0.40838786927024157, |
| "learning_rate": 0.00016222766651997789, |
| "loss": 1.1695, |
| "step": 1380 |
| }, |
| { |
| "epoch": 0.35890126975900494, |
| "grad_norm": 0.4096176785342024, |
| "learning_rate": 0.0001618729813611414, |
| "loss": 1.1447, |
| "step": 1385 |
| }, |
| { |
| "epoch": 0.3601969422130086, |
| "grad_norm": 0.41309116576243843, |
| "learning_rate": 0.00016151703050314986, |
| "loss": 1.1804, |
| "step": 1390 |
| }, |
| { |
| "epoch": 0.36149261466701216, |
| "grad_norm": 0.4053926841703832, |
| "learning_rate": 0.00016115982122748043, |
| "loss": 1.1471, |
| "step": 1395 |
| }, |
| { |
| "epoch": 0.3627882871210158, |
| "grad_norm": 0.4058497805268484, |
| "learning_rate": 0.00016080136084135297, |
| "loss": 1.1494, |
| "step": 1400 |
| }, |
| { |
| "epoch": 0.36408395957501943, |
| "grad_norm": 0.4216552491807955, |
| "learning_rate": 0.00016044165667758055, |
| "loss": 1.1928, |
| "step": 1405 |
| }, |
| { |
| "epoch": 0.36537963202902307, |
| "grad_norm": 0.4336798036895343, |
| "learning_rate": 0.0001600807160944195, |
| "loss": 1.19, |
| "step": 1410 |
| }, |
| { |
| "epoch": 0.3666753044830267, |
| "grad_norm": 0.42380600181334993, |
| "learning_rate": 0.00015971854647541884, |
| "loss": 1.1674, |
| "step": 1415 |
| }, |
| { |
| "epoch": 0.3679709769370303, |
| "grad_norm": 0.4116489897844781, |
| "learning_rate": 0.00015935515522926927, |
| "loss": 1.1407, |
| "step": 1420 |
| }, |
| { |
| "epoch": 0.36926664939103393, |
| "grad_norm": 0.4051816422975703, |
| "learning_rate": 0.00015899054978965157, |
| "loss": 1.1861, |
| "step": 1425 |
| }, |
| { |
| "epoch": 0.37056232184503757, |
| "grad_norm": 0.4374412539295244, |
| "learning_rate": 0.0001586247376150846, |
| "loss": 1.2273, |
| "step": 1430 |
| }, |
| { |
| "epoch": 0.3718579942990412, |
| "grad_norm": 0.4225494310104097, |
| "learning_rate": 0.00015825772618877263, |
| "loss": 1.2218, |
| "step": 1435 |
| }, |
| { |
| "epoch": 0.37315366675304484, |
| "grad_norm": 0.4202018672382752, |
| "learning_rate": 0.00015788952301845237, |
| "loss": 1.1155, |
| "step": 1440 |
| }, |
| { |
| "epoch": 0.3744493392070485, |
| "grad_norm": 0.4313070413558118, |
| "learning_rate": 0.0001575201356362393, |
| "loss": 1.1551, |
| "step": 1445 |
| }, |
| { |
| "epoch": 0.37574501166105206, |
| "grad_norm": 0.40304574068983823, |
| "learning_rate": 0.00015714957159847367, |
| "loss": 1.1491, |
| "step": 1450 |
| }, |
| { |
| "epoch": 0.3770406841150557, |
| "grad_norm": 0.4413697670634288, |
| "learning_rate": 0.00015677783848556576, |
| "loss": 1.1631, |
| "step": 1455 |
| }, |
| { |
| "epoch": 0.37833635656905934, |
| "grad_norm": 0.4021843049770175, |
| "learning_rate": 0.00015640494390184112, |
| "loss": 1.1304, |
| "step": 1460 |
| }, |
| { |
| "epoch": 0.379632029023063, |
| "grad_norm": 0.41280223940193794, |
| "learning_rate": 0.0001560308954753847, |
| "loss": 1.1458, |
| "step": 1465 |
| }, |
| { |
| "epoch": 0.3809277014770666, |
| "grad_norm": 0.44403414056325813, |
| "learning_rate": 0.00015565570085788495, |
| "loss": 1.2007, |
| "step": 1470 |
| }, |
| { |
| "epoch": 0.3822233739310702, |
| "grad_norm": 0.42327265730452945, |
| "learning_rate": 0.00015527936772447725, |
| "loss": 1.1168, |
| "step": 1475 |
| }, |
| { |
| "epoch": 0.38351904638507384, |
| "grad_norm": 0.41872242471581905, |
| "learning_rate": 0.00015490190377358704, |
| "loss": 1.1551, |
| "step": 1480 |
| }, |
| { |
| "epoch": 0.3848147188390775, |
| "grad_norm": 0.4076213102610575, |
| "learning_rate": 0.00015452331672677206, |
| "loss": 1.0902, |
| "step": 1485 |
| }, |
| { |
| "epoch": 0.3861103912930811, |
| "grad_norm": 0.39807246728274887, |
| "learning_rate": 0.00015414361432856475, |
| "loss": 1.1598, |
| "step": 1490 |
| }, |
| { |
| "epoch": 0.38740606374708475, |
| "grad_norm": 0.40854063250285516, |
| "learning_rate": 0.00015376280434631345, |
| "loss": 1.1806, |
| "step": 1495 |
| }, |
| { |
| "epoch": 0.3887017362010884, |
| "grad_norm": 0.40837052057355316, |
| "learning_rate": 0.00015338089457002382, |
| "loss": 1.0829, |
| "step": 1500 |
| }, |
| { |
| "epoch": 0.38999740865509197, |
| "grad_norm": 0.4123609619777807, |
| "learning_rate": 0.00015299789281219935, |
| "loss": 1.1688, |
| "step": 1505 |
| }, |
| { |
| "epoch": 0.3912930811090956, |
| "grad_norm": 0.4316199953541538, |
| "learning_rate": 0.00015261380690768144, |
| "loss": 1.1543, |
| "step": 1510 |
| }, |
| { |
| "epoch": 0.39258875356309925, |
| "grad_norm": 0.4398633371362292, |
| "learning_rate": 0.00015222864471348943, |
| "loss": 1.1724, |
| "step": 1515 |
| }, |
| { |
| "epoch": 0.3938844260171029, |
| "grad_norm": 0.4081954480961544, |
| "learning_rate": 0.00015184241410865954, |
| "loss": 1.1269, |
| "step": 1520 |
| }, |
| { |
| "epoch": 0.3951800984711065, |
| "grad_norm": 0.3994337234690745, |
| "learning_rate": 0.00015145512299408388, |
| "loss": 1.1598, |
| "step": 1525 |
| }, |
| { |
| "epoch": 0.3964757709251101, |
| "grad_norm": 0.4415117779921472, |
| "learning_rate": 0.00015106677929234877, |
| "loss": 1.0969, |
| "step": 1530 |
| }, |
| { |
| "epoch": 0.39777144337911374, |
| "grad_norm": 0.4033267536569039, |
| "learning_rate": 0.0001506773909475727, |
| "loss": 1.1955, |
| "step": 1535 |
| }, |
| { |
| "epoch": 0.3990671158331174, |
| "grad_norm": 0.39024517506575723, |
| "learning_rate": 0.00015028696592524386, |
| "loss": 1.1562, |
| "step": 1540 |
| }, |
| { |
| "epoch": 0.400362788287121, |
| "grad_norm": 0.4359801226192451, |
| "learning_rate": 0.000149895512212057, |
| "loss": 1.1695, |
| "step": 1545 |
| }, |
| { |
| "epoch": 0.40165846074112466, |
| "grad_norm": 0.3996337334749442, |
| "learning_rate": 0.00014950303781575034, |
| "loss": 1.0982, |
| "step": 1550 |
| }, |
| { |
| "epoch": 0.4029541331951283, |
| "grad_norm": 0.39816310192557475, |
| "learning_rate": 0.00014910955076494152, |
| "loss": 1.1223, |
| "step": 1555 |
| }, |
| { |
| "epoch": 0.4042498056491319, |
| "grad_norm": 0.37894092829518466, |
| "learning_rate": 0.00014871505910896352, |
| "loss": 1.1217, |
| "step": 1560 |
| }, |
| { |
| "epoch": 0.4055454781031355, |
| "grad_norm": 0.4319425134415384, |
| "learning_rate": 0.0001483195709176999, |
| "loss": 1.1966, |
| "step": 1565 |
| }, |
| { |
| "epoch": 0.40684115055713915, |
| "grad_norm": 0.3996173231205214, |
| "learning_rate": 0.00014792309428141978, |
| "loss": 1.1039, |
| "step": 1570 |
| }, |
| { |
| "epoch": 0.4081368230111428, |
| "grad_norm": 0.40718491453474887, |
| "learning_rate": 0.0001475256373106123, |
| "loss": 1.1725, |
| "step": 1575 |
| }, |
| { |
| "epoch": 0.40943249546514643, |
| "grad_norm": 0.4286665586760103, |
| "learning_rate": 0.00014712720813582066, |
| "loss": 1.1443, |
| "step": 1580 |
| }, |
| { |
| "epoch": 0.41072816791915, |
| "grad_norm": 0.4033525905918133, |
| "learning_rate": 0.00014672781490747606, |
| "loss": 1.1742, |
| "step": 1585 |
| }, |
| { |
| "epoch": 0.41202384037315365, |
| "grad_norm": 0.40558326017198004, |
| "learning_rate": 0.00014632746579573052, |
| "loss": 1.2117, |
| "step": 1590 |
| }, |
| { |
| "epoch": 0.4133195128271573, |
| "grad_norm": 0.4056524890982348, |
| "learning_rate": 0.0001459261689902902, |
| "loss": 1.2034, |
| "step": 1595 |
| }, |
| { |
| "epoch": 0.4146151852811609, |
| "grad_norm": 0.44100980335173223, |
| "learning_rate": 0.00014552393270024765, |
| "loss": 1.1544, |
| "step": 1600 |
| }, |
| { |
| "epoch": 0.41591085773516456, |
| "grad_norm": 0.39891755340287666, |
| "learning_rate": 0.00014512076515391375, |
| "loss": 1.1256, |
| "step": 1605 |
| }, |
| { |
| "epoch": 0.4172065301891682, |
| "grad_norm": 0.40682040920269774, |
| "learning_rate": 0.00014471667459864973, |
| "loss": 1.1499, |
| "step": 1610 |
| }, |
| { |
| "epoch": 0.4185022026431718, |
| "grad_norm": 0.44312416626229856, |
| "learning_rate": 0.00014431166930069816, |
| "loss": 1.0977, |
| "step": 1615 |
| }, |
| { |
| "epoch": 0.4197978750971754, |
| "grad_norm": 0.4301392634897964, |
| "learning_rate": 0.00014390575754501402, |
| "loss": 1.139, |
| "step": 1620 |
| }, |
| { |
| "epoch": 0.42109354755117906, |
| "grad_norm": 0.418724699659618, |
| "learning_rate": 0.0001434989476350951, |
| "loss": 1.1592, |
| "step": 1625 |
| }, |
| { |
| "epoch": 0.4223892200051827, |
| "grad_norm": 0.4134020107191313, |
| "learning_rate": 0.00014309124789281226, |
| "loss": 1.2104, |
| "step": 1630 |
| }, |
| { |
| "epoch": 0.42368489245918634, |
| "grad_norm": 0.39580446880805387, |
| "learning_rate": 0.00014268266665823912, |
| "loss": 1.1757, |
| "step": 1635 |
| }, |
| { |
| "epoch": 0.42498056491319, |
| "grad_norm": 0.41472381643526246, |
| "learning_rate": 0.00014227321228948146, |
| "loss": 1.1322, |
| "step": 1640 |
| }, |
| { |
| "epoch": 0.42627623736719356, |
| "grad_norm": 0.39275693329437805, |
| "learning_rate": 0.0001418628931625062, |
| "loss": 1.1742, |
| "step": 1645 |
| }, |
| { |
| "epoch": 0.4275719098211972, |
| "grad_norm": 0.38619824420216714, |
| "learning_rate": 0.00014145171767097027, |
| "loss": 1.1511, |
| "step": 1650 |
| }, |
| { |
| "epoch": 0.42886758227520083, |
| "grad_norm": 0.4254507618246464, |
| "learning_rate": 0.00014103969422604856, |
| "loss": 1.1231, |
| "step": 1655 |
| }, |
| { |
| "epoch": 0.43016325472920447, |
| "grad_norm": 0.40459401347151125, |
| "learning_rate": 0.00014062683125626218, |
| "loss": 1.1384, |
| "step": 1660 |
| }, |
| { |
| "epoch": 0.4314589271832081, |
| "grad_norm": 0.409140169024848, |
| "learning_rate": 0.0001402131372073058, |
| "loss": 1.191, |
| "step": 1665 |
| }, |
| { |
| "epoch": 0.4327545996372117, |
| "grad_norm": 0.389695812394897, |
| "learning_rate": 0.00013979862054187505, |
| "loss": 1.1543, |
| "step": 1670 |
| }, |
| { |
| "epoch": 0.43405027209121533, |
| "grad_norm": 0.42112717245729503, |
| "learning_rate": 0.00013938328973949336, |
| "loss": 1.1715, |
| "step": 1675 |
| }, |
| { |
| "epoch": 0.43534594454521897, |
| "grad_norm": 0.3864671779444786, |
| "learning_rate": 0.0001389671532963384, |
| "loss": 1.1508, |
| "step": 1680 |
| }, |
| { |
| "epoch": 0.4366416169992226, |
| "grad_norm": 0.39052429534303434, |
| "learning_rate": 0.00013855021972506844, |
| "loss": 1.1054, |
| "step": 1685 |
| }, |
| { |
| "epoch": 0.43793728945322624, |
| "grad_norm": 0.404117930960366, |
| "learning_rate": 0.0001381324975546481, |
| "loss": 1.1623, |
| "step": 1690 |
| }, |
| { |
| "epoch": 0.4392329619072299, |
| "grad_norm": 0.45134279442398223, |
| "learning_rate": 0.0001377139953301739, |
| "loss": 1.1301, |
| "step": 1695 |
| }, |
| { |
| "epoch": 0.44052863436123346, |
| "grad_norm": 0.42206784378450607, |
| "learning_rate": 0.00013729472161269946, |
| "loss": 1.18, |
| "step": 1700 |
| }, |
| { |
| "epoch": 0.4418243068152371, |
| "grad_norm": 0.43048779695832334, |
| "learning_rate": 0.00013687468497906044, |
| "loss": 1.1413, |
| "step": 1705 |
| }, |
| { |
| "epoch": 0.44311997926924074, |
| "grad_norm": 0.3902243379067639, |
| "learning_rate": 0.00013645389402169893, |
| "loss": 1.1473, |
| "step": 1710 |
| }, |
| { |
| "epoch": 0.4444156517232444, |
| "grad_norm": 0.3790071383777115, |
| "learning_rate": 0.00013603235734848784, |
| "loss": 1.2114, |
| "step": 1715 |
| }, |
| { |
| "epoch": 0.445711324177248, |
| "grad_norm": 0.39998613746463574, |
| "learning_rate": 0.00013561008358255468, |
| "loss": 1.1193, |
| "step": 1720 |
| }, |
| { |
| "epoch": 0.4470069966312516, |
| "grad_norm": 0.4136780934178077, |
| "learning_rate": 0.0001351870813621054, |
| "loss": 1.1417, |
| "step": 1725 |
| }, |
| { |
| "epoch": 0.44830266908525523, |
| "grad_norm": 0.39313685685153266, |
| "learning_rate": 0.00013476335934024735, |
| "loss": 1.1437, |
| "step": 1730 |
| }, |
| { |
| "epoch": 0.44959834153925887, |
| "grad_norm": 0.41028501570471904, |
| "learning_rate": 0.00013433892618481248, |
| "loss": 1.1032, |
| "step": 1735 |
| }, |
| { |
| "epoch": 0.4508940139932625, |
| "grad_norm": 0.43704478597398877, |
| "learning_rate": 0.00013391379057817995, |
| "loss": 1.1933, |
| "step": 1740 |
| }, |
| { |
| "epoch": 0.45218968644726615, |
| "grad_norm": 0.37937862506331654, |
| "learning_rate": 0.00013348796121709862, |
| "loss": 1.187, |
| "step": 1745 |
| }, |
| { |
| "epoch": 0.4534853589012698, |
| "grad_norm": 0.40181873506259413, |
| "learning_rate": 0.00013306144681250908, |
| "loss": 1.1625, |
| "step": 1750 |
| }, |
| { |
| "epoch": 0.45478103135527337, |
| "grad_norm": 0.4117326535807377, |
| "learning_rate": 0.00013263425608936536, |
| "loss": 1.1875, |
| "step": 1755 |
| }, |
| { |
| "epoch": 0.456076703809277, |
| "grad_norm": 0.39234377730019654, |
| "learning_rate": 0.00013220639778645663, |
| "loss": 1.1888, |
| "step": 1760 |
| }, |
| { |
| "epoch": 0.45737237626328064, |
| "grad_norm": 0.42622878050330665, |
| "learning_rate": 0.0001317778806562283, |
| "loss": 1.151, |
| "step": 1765 |
| }, |
| { |
| "epoch": 0.4586680487172843, |
| "grad_norm": 0.3997439152224071, |
| "learning_rate": 0.000131348713464603, |
| "loss": 1.1271, |
| "step": 1770 |
| }, |
| { |
| "epoch": 0.4599637211712879, |
| "grad_norm": 0.4151543746885238, |
| "learning_rate": 0.0001309189049908014, |
| "loss": 1.1745, |
| "step": 1775 |
| }, |
| { |
| "epoch": 0.4612593936252915, |
| "grad_norm": 0.4136123970052988, |
| "learning_rate": 0.00013048846402716237, |
| "loss": 1.1446, |
| "step": 1780 |
| }, |
| { |
| "epoch": 0.46255506607929514, |
| "grad_norm": 0.3925741414110596, |
| "learning_rate": 0.0001300573993789633, |
| "loss": 1.1186, |
| "step": 1785 |
| }, |
| { |
| "epoch": 0.4638507385332988, |
| "grad_norm": 0.43074300043773284, |
| "learning_rate": 0.00012962571986423993, |
| "loss": 1.2004, |
| "step": 1790 |
| }, |
| { |
| "epoch": 0.4651464109873024, |
| "grad_norm": 0.3948645456920277, |
| "learning_rate": 0.00012919343431360596, |
| "loss": 1.1534, |
| "step": 1795 |
| }, |
| { |
| "epoch": 0.46644208344130605, |
| "grad_norm": 0.4423236206923003, |
| "learning_rate": 0.00012876055157007242, |
| "loss": 1.1509, |
| "step": 1800 |
| }, |
| { |
| "epoch": 0.4677377558953097, |
| "grad_norm": 0.39371508060725335, |
| "learning_rate": 0.00012832708048886679, |
| "loss": 1.1941, |
| "step": 1805 |
| }, |
| { |
| "epoch": 0.4690334283493133, |
| "grad_norm": 0.4156050706970669, |
| "learning_rate": 0.00012789302993725175, |
| "loss": 1.2233, |
| "step": 1810 |
| }, |
| { |
| "epoch": 0.4703291008033169, |
| "grad_norm": 0.3844699731827056, |
| "learning_rate": 0.0001274584087943439, |
| "loss": 1.172, |
| "step": 1815 |
| }, |
| { |
| "epoch": 0.47162477325732055, |
| "grad_norm": 0.392439037665497, |
| "learning_rate": 0.00012702322595093212, |
| "loss": 1.1935, |
| "step": 1820 |
| }, |
| { |
| "epoch": 0.4729204457113242, |
| "grad_norm": 0.42109447627306007, |
| "learning_rate": 0.00012658749030929566, |
| "loss": 1.0821, |
| "step": 1825 |
| }, |
| { |
| "epoch": 0.4742161181653278, |
| "grad_norm": 0.3985407382924968, |
| "learning_rate": 0.00012615121078302202, |
| "loss": 1.1564, |
| "step": 1830 |
| }, |
| { |
| "epoch": 0.4755117906193314, |
| "grad_norm": 0.40190944372077425, |
| "learning_rate": 0.0001257143962968246, |
| "loss": 1.1858, |
| "step": 1835 |
| }, |
| { |
| "epoch": 0.47680746307333505, |
| "grad_norm": 0.45076787480964914, |
| "learning_rate": 0.00012527705578636023, |
| "loss": 1.1514, |
| "step": 1840 |
| }, |
| { |
| "epoch": 0.4781031355273387, |
| "grad_norm": 0.4363861711704788, |
| "learning_rate": 0.0001248391981980462, |
| "loss": 1.133, |
| "step": 1845 |
| }, |
| { |
| "epoch": 0.4793988079813423, |
| "grad_norm": 0.43776230767541446, |
| "learning_rate": 0.00012440083248887754, |
| "loss": 1.2082, |
| "step": 1850 |
| }, |
| { |
| "epoch": 0.48069448043534596, |
| "grad_norm": 0.4127859770012284, |
| "learning_rate": 0.00012396196762624341, |
| "loss": 1.1613, |
| "step": 1855 |
| }, |
| { |
| "epoch": 0.4819901528893496, |
| "grad_norm": 0.3940447846474964, |
| "learning_rate": 0.00012352261258774395, |
| "loss": 1.15, |
| "step": 1860 |
| }, |
| { |
| "epoch": 0.4832858253433532, |
| "grad_norm": 0.4193523558931992, |
| "learning_rate": 0.0001230827763610066, |
| "loss": 1.1382, |
| "step": 1865 |
| }, |
| { |
| "epoch": 0.4845814977973568, |
| "grad_norm": 0.395557633792394, |
| "learning_rate": 0.00012264246794350202, |
| "loss": 1.1678, |
| "step": 1870 |
| }, |
| { |
| "epoch": 0.48587717025136046, |
| "grad_norm": 0.40423444904954847, |
| "learning_rate": 0.00012220169634236038, |
| "loss": 1.157, |
| "step": 1875 |
| }, |
| { |
| "epoch": 0.4871728427053641, |
| "grad_norm": 0.3870746898688869, |
| "learning_rate": 0.00012176047057418682, |
| "loss": 1.1439, |
| "step": 1880 |
| }, |
| { |
| "epoch": 0.48846851515936773, |
| "grad_norm": 0.4458893388021195, |
| "learning_rate": 0.00012131879966487709, |
| "loss": 1.1513, |
| "step": 1885 |
| }, |
| { |
| "epoch": 0.4897641876133713, |
| "grad_norm": 0.38478400725184214, |
| "learning_rate": 0.00012087669264943302, |
| "loss": 1.1333, |
| "step": 1890 |
| }, |
| { |
| "epoch": 0.49105986006737495, |
| "grad_norm": 0.37819480972862285, |
| "learning_rate": 0.00012043415857177751, |
| "loss": 1.1663, |
| "step": 1895 |
| }, |
| { |
| "epoch": 0.4923555325213786, |
| "grad_norm": 0.4125210415708117, |
| "learning_rate": 0.00011999120648456974, |
| "loss": 1.1457, |
| "step": 1900 |
| }, |
| { |
| "epoch": 0.49365120497538223, |
| "grad_norm": 0.3869780886696955, |
| "learning_rate": 0.00011954784544901971, |
| "loss": 1.1536, |
| "step": 1905 |
| }, |
| { |
| "epoch": 0.49494687742938587, |
| "grad_norm": 0.40907576777936727, |
| "learning_rate": 0.00011910408453470316, |
| "loss": 1.1361, |
| "step": 1910 |
| }, |
| { |
| "epoch": 0.4962425498833895, |
| "grad_norm": 0.3737647497815572, |
| "learning_rate": 0.00011865993281937589, |
| "loss": 1.1006, |
| "step": 1915 |
| }, |
| { |
| "epoch": 0.4975382223373931, |
| "grad_norm": 0.41125742423172207, |
| "learning_rate": 0.00011821539938878801, |
| "loss": 1.1641, |
| "step": 1920 |
| }, |
| { |
| "epoch": 0.4988338947913967, |
| "grad_norm": 0.38132649216646763, |
| "learning_rate": 0.00011777049333649826, |
| "loss": 1.1525, |
| "step": 1925 |
| }, |
| { |
| "epoch": 0.5001295672454004, |
| "grad_norm": 0.38834910320923277, |
| "learning_rate": 0.00011732522376368781, |
| "loss": 1.1531, |
| "step": 1930 |
| }, |
| { |
| "epoch": 0.501425239699404, |
| "grad_norm": 0.40314094443067383, |
| "learning_rate": 0.0001168795997789742, |
| "loss": 1.1592, |
| "step": 1935 |
| }, |
| { |
| "epoch": 0.5027209121534076, |
| "grad_norm": 0.4040865102971838, |
| "learning_rate": 0.00011643363049822496, |
| "loss": 1.2547, |
| "step": 1940 |
| }, |
| { |
| "epoch": 0.5040165846074113, |
| "grad_norm": 0.38562943372958325, |
| "learning_rate": 0.00011598732504437107, |
| "loss": 1.1109, |
| "step": 1945 |
| }, |
| { |
| "epoch": 0.5053122570614149, |
| "grad_norm": 0.4250211446665384, |
| "learning_rate": 0.00011554069254722051, |
| "loss": 1.1941, |
| "step": 1950 |
| }, |
| { |
| "epoch": 0.5066079295154186, |
| "grad_norm": 0.4032090808510551, |
| "learning_rate": 0.00011509374214327131, |
| "loss": 1.1572, |
| "step": 1955 |
| }, |
| { |
| "epoch": 0.5079036019694221, |
| "grad_norm": 0.40575976251729745, |
| "learning_rate": 0.00011464648297552478, |
| "loss": 1.1657, |
| "step": 1960 |
| }, |
| { |
| "epoch": 0.5091992744234257, |
| "grad_norm": 0.41052107415559824, |
| "learning_rate": 0.00011419892419329844, |
| "loss": 1.1642, |
| "step": 1965 |
| }, |
| { |
| "epoch": 0.5104949468774294, |
| "grad_norm": 0.3958628275091308, |
| "learning_rate": 0.00011375107495203873, |
| "loss": 1.2116, |
| "step": 1970 |
| }, |
| { |
| "epoch": 0.511790619331433, |
| "grad_norm": 0.38478613819779206, |
| "learning_rate": 0.00011330294441313402, |
| "loss": 1.1451, |
| "step": 1975 |
| }, |
| { |
| "epoch": 0.5130862917854366, |
| "grad_norm": 0.43128674114012594, |
| "learning_rate": 0.00011285454174372692, |
| "loss": 1.1494, |
| "step": 1980 |
| }, |
| { |
| "epoch": 0.5143819642394403, |
| "grad_norm": 0.4424953381805022, |
| "learning_rate": 0.0001124058761165268, |
| "loss": 1.0932, |
| "step": 1985 |
| }, |
| { |
| "epoch": 0.5156776366934439, |
| "grad_norm": 0.4063559419481111, |
| "learning_rate": 0.00011195695670962234, |
| "loss": 1.1137, |
| "step": 1990 |
| }, |
| { |
| "epoch": 0.5169733091474475, |
| "grad_norm": 0.3975356665709952, |
| "learning_rate": 0.00011150779270629353, |
| "loss": 1.1873, |
| "step": 1995 |
| }, |
| { |
| "epoch": 0.5182689816014512, |
| "grad_norm": 0.4185740854748383, |
| "learning_rate": 0.00011105839329482397, |
| "loss": 1.1108, |
| "step": 2000 |
| }, |
| { |
| "epoch": 0.5195646540554548, |
| "grad_norm": 0.38989666093808334, |
| "learning_rate": 0.00011060876766831285, |
| "loss": 1.1663, |
| "step": 2005 |
| }, |
| { |
| "epoch": 0.5208603265094585, |
| "grad_norm": 0.40596909372428913, |
| "learning_rate": 0.00011015892502448692, |
| "loss": 1.1382, |
| "step": 2010 |
| }, |
| { |
| "epoch": 0.522155998963462, |
| "grad_norm": 0.37956800433559007, |
| "learning_rate": 0.00010970887456551234, |
| "loss": 1.104, |
| "step": 2015 |
| }, |
| { |
| "epoch": 0.5234516714174656, |
| "grad_norm": 0.3972204538731939, |
| "learning_rate": 0.00010925862549780637, |
| "loss": 1.1414, |
| "step": 2020 |
| }, |
| { |
| "epoch": 0.5247473438714693, |
| "grad_norm": 0.4266101329792826, |
| "learning_rate": 0.00010880818703184919, |
| "loss": 1.2131, |
| "step": 2025 |
| }, |
| { |
| "epoch": 0.5260430163254729, |
| "grad_norm": 0.4059164191077208, |
| "learning_rate": 0.00010835756838199524, |
| "loss": 1.1005, |
| "step": 2030 |
| }, |
| { |
| "epoch": 0.5273386887794765, |
| "grad_norm": 0.37690318544316265, |
| "learning_rate": 0.00010790677876628501, |
| "loss": 1.1448, |
| "step": 2035 |
| }, |
| { |
| "epoch": 0.5286343612334802, |
| "grad_norm": 0.39147947369844754, |
| "learning_rate": 0.00010745582740625631, |
| "loss": 1.161, |
| "step": 2040 |
| }, |
| { |
| "epoch": 0.5299300336874838, |
| "grad_norm": 0.4061256597825514, |
| "learning_rate": 0.00010700472352675556, |
| "loss": 1.1861, |
| "step": 2045 |
| }, |
| { |
| "epoch": 0.5312257061414875, |
| "grad_norm": 0.3877742069695207, |
| "learning_rate": 0.00010655347635574937, |
| "loss": 1.1816, |
| "step": 2050 |
| }, |
| { |
| "epoch": 0.5325213785954911, |
| "grad_norm": 0.3971133244937986, |
| "learning_rate": 0.00010610209512413536, |
| "loss": 1.2017, |
| "step": 2055 |
| }, |
| { |
| "epoch": 0.5338170510494947, |
| "grad_norm": 0.41551450058475425, |
| "learning_rate": 0.0001056505890655537, |
| "loss": 1.1692, |
| "step": 2060 |
| }, |
| { |
| "epoch": 0.5351127235034984, |
| "grad_norm": 0.43646326285085413, |
| "learning_rate": 0.00010519896741619803, |
| "loss": 1.0993, |
| "step": 2065 |
| }, |
| { |
| "epoch": 0.5364083959575019, |
| "grad_norm": 0.38667765038581386, |
| "learning_rate": 0.00010474723941462658, |
| "loss": 1.1151, |
| "step": 2070 |
| }, |
| { |
| "epoch": 0.5377040684115055, |
| "grad_norm": 0.39156927596261365, |
| "learning_rate": 0.00010429541430157313, |
| "loss": 1.1641, |
| "step": 2075 |
| }, |
| { |
| "epoch": 0.5389997408655092, |
| "grad_norm": 0.39506144218425854, |
| "learning_rate": 0.00010384350131975802, |
| "loss": 1.2444, |
| "step": 2080 |
| }, |
| { |
| "epoch": 0.5402954133195128, |
| "grad_norm": 0.3881326675938924, |
| "learning_rate": 0.00010339150971369906, |
| "loss": 1.129, |
| "step": 2085 |
| }, |
| { |
| "epoch": 0.5415910857735164, |
| "grad_norm": 0.3901794452377504, |
| "learning_rate": 0.00010293944872952248, |
| "loss": 1.1536, |
| "step": 2090 |
| }, |
| { |
| "epoch": 0.5428867582275201, |
| "grad_norm": 0.3852428804649106, |
| "learning_rate": 0.00010248732761477371, |
| "loss": 1.1365, |
| "step": 2095 |
| }, |
| { |
| "epoch": 0.5441824306815237, |
| "grad_norm": 0.4261491575765494, |
| "learning_rate": 0.0001020351556182282, |
| "loss": 1.089, |
| "step": 2100 |
| }, |
| { |
| "epoch": 0.5454781031355274, |
| "grad_norm": 0.4070655687819442, |
| "learning_rate": 0.00010158294198970232, |
| "loss": 1.1591, |
| "step": 2105 |
| }, |
| { |
| "epoch": 0.546773775589531, |
| "grad_norm": 0.4142505000419484, |
| "learning_rate": 0.00010113069597986402, |
| "loss": 1.1403, |
| "step": 2110 |
| }, |
| { |
| "epoch": 0.5480694480435346, |
| "grad_norm": 0.3988427337375973, |
| "learning_rate": 0.0001006784268400437, |
| "loss": 1.1043, |
| "step": 2115 |
| }, |
| { |
| "epoch": 0.5493651204975383, |
| "grad_norm": 0.37926295342271793, |
| "learning_rate": 0.00010022614382204492, |
| "loss": 1.1694, |
| "step": 2120 |
| }, |
| { |
| "epoch": 0.5506607929515418, |
| "grad_norm": 0.3854042883512697, |
| "learning_rate": 9.97738561779551e-05, |
| "loss": 1.1891, |
| "step": 2125 |
| }, |
| { |
| "epoch": 0.5519564654055454, |
| "grad_norm": 0.4035084032954952, |
| "learning_rate": 9.932157315995631e-05, |
| "loss": 1.1563, |
| "step": 2130 |
| }, |
| { |
| "epoch": 0.5532521378595491, |
| "grad_norm": 0.4083098281929601, |
| "learning_rate": 9.8869304020136e-05, |
| "loss": 1.2097, |
| "step": 2135 |
| }, |
| { |
| "epoch": 0.5545478103135527, |
| "grad_norm": 0.3884516778316842, |
| "learning_rate": 9.841705801029769e-05, |
| "loss": 1.1932, |
| "step": 2140 |
| }, |
| { |
| "epoch": 0.5558434827675564, |
| "grad_norm": 0.38224759440627476, |
| "learning_rate": 9.79648443817718e-05, |
| "loss": 1.0978, |
| "step": 2145 |
| }, |
| { |
| "epoch": 0.55713915522156, |
| "grad_norm": 0.46057864945040095, |
| "learning_rate": 9.751267238522631e-05, |
| "loss": 1.1465, |
| "step": 2150 |
| }, |
| { |
| "epoch": 0.5584348276755636, |
| "grad_norm": 0.38627135619303393, |
| "learning_rate": 9.706055127047755e-05, |
| "loss": 1.1533, |
| "step": 2155 |
| }, |
| { |
| "epoch": 0.5597305001295673, |
| "grad_norm": 0.3891179649712941, |
| "learning_rate": 9.660849028630096e-05, |
| "loss": 1.1984, |
| "step": 2160 |
| }, |
| { |
| "epoch": 0.5610261725835709, |
| "grad_norm": 0.39223924604212307, |
| "learning_rate": 9.615649868024199e-05, |
| "loss": 1.1463, |
| "step": 2165 |
| }, |
| { |
| "epoch": 0.5623218450375745, |
| "grad_norm": 0.39471896059848777, |
| "learning_rate": 9.570458569842688e-05, |
| "loss": 1.1487, |
| "step": 2170 |
| }, |
| { |
| "epoch": 0.5636175174915782, |
| "grad_norm": 0.3928874290731547, |
| "learning_rate": 9.525276058537344e-05, |
| "loss": 1.1315, |
| "step": 2175 |
| }, |
| { |
| "epoch": 0.5649131899455817, |
| "grad_norm": 0.4064990206731607, |
| "learning_rate": 9.480103258380198e-05, |
| "loss": 1.1562, |
| "step": 2180 |
| }, |
| { |
| "epoch": 0.5662088623995853, |
| "grad_norm": 0.38415170548201955, |
| "learning_rate": 9.434941093444632e-05, |
| "loss": 1.1811, |
| "step": 2185 |
| }, |
| { |
| "epoch": 0.567504534853589, |
| "grad_norm": 0.3859240132285136, |
| "learning_rate": 9.389790487586465e-05, |
| "loss": 1.2216, |
| "step": 2190 |
| }, |
| { |
| "epoch": 0.5688002073075926, |
| "grad_norm": 0.40615716009377495, |
| "learning_rate": 9.344652364425065e-05, |
| "loss": 1.2206, |
| "step": 2195 |
| }, |
| { |
| "epoch": 0.5700958797615963, |
| "grad_norm": 0.3919524007695092, |
| "learning_rate": 9.299527647324444e-05, |
| "loss": 1.1929, |
| "step": 2200 |
| }, |
| { |
| "epoch": 0.5713915522155999, |
| "grad_norm": 0.3821208159345121, |
| "learning_rate": 9.254417259374374e-05, |
| "loss": 1.0911, |
| "step": 2205 |
| }, |
| { |
| "epoch": 0.5726872246696035, |
| "grad_norm": 0.42022947326779125, |
| "learning_rate": 9.2093221233715e-05, |
| "loss": 1.1504, |
| "step": 2210 |
| }, |
| { |
| "epoch": 0.5739828971236072, |
| "grad_norm": 0.4544847994091584, |
| "learning_rate": 9.164243161800477e-05, |
| "loss": 1.1582, |
| "step": 2215 |
| }, |
| { |
| "epoch": 0.5752785695776108, |
| "grad_norm": 0.41094059546634143, |
| "learning_rate": 9.119181296815085e-05, |
| "loss": 1.1297, |
| "step": 2220 |
| }, |
| { |
| "epoch": 0.5765742420316144, |
| "grad_norm": 0.3805641014764856, |
| "learning_rate": 9.074137450219364e-05, |
| "loss": 1.185, |
| "step": 2225 |
| }, |
| { |
| "epoch": 0.5778699144856181, |
| "grad_norm": 0.39504497212816847, |
| "learning_rate": 9.02911254344877e-05, |
| "loss": 1.1875, |
| "step": 2230 |
| }, |
| { |
| "epoch": 0.5791655869396216, |
| "grad_norm": 0.40451192253045454, |
| "learning_rate": 8.984107497551311e-05, |
| "loss": 1.1368, |
| "step": 2235 |
| }, |
| { |
| "epoch": 0.5804612593936253, |
| "grad_norm": 0.3769782624908954, |
| "learning_rate": 8.939123233168717e-05, |
| "loss": 1.1391, |
| "step": 2240 |
| }, |
| { |
| "epoch": 0.5817569318476289, |
| "grad_norm": 0.4058748960564205, |
| "learning_rate": 8.894160670517606e-05, |
| "loss": 1.1751, |
| "step": 2245 |
| }, |
| { |
| "epoch": 0.5830526043016325, |
| "grad_norm": 0.3881661677768996, |
| "learning_rate": 8.849220729370651e-05, |
| "loss": 1.1215, |
| "step": 2250 |
| }, |
| { |
| "epoch": 0.5843482767556362, |
| "grad_norm": 0.3807361487895863, |
| "learning_rate": 8.804304329037771e-05, |
| "loss": 1.1604, |
| "step": 2255 |
| }, |
| { |
| "epoch": 0.5856439492096398, |
| "grad_norm": 0.3780073193073684, |
| "learning_rate": 8.759412388347321e-05, |
| "loss": 1.1159, |
| "step": 2260 |
| }, |
| { |
| "epoch": 0.5869396216636434, |
| "grad_norm": 0.3910762750334457, |
| "learning_rate": 8.71454582562731e-05, |
| "loss": 1.1283, |
| "step": 2265 |
| }, |
| { |
| "epoch": 0.5882352941176471, |
| "grad_norm": 0.3577377830833829, |
| "learning_rate": 8.669705558686599e-05, |
| "loss": 1.0881, |
| "step": 2270 |
| }, |
| { |
| "epoch": 0.5895309665716507, |
| "grad_norm": 0.38870803678431104, |
| "learning_rate": 8.624892504796128e-05, |
| "loss": 1.1302, |
| "step": 2275 |
| }, |
| { |
| "epoch": 0.5908266390256544, |
| "grad_norm": 0.43127992879460963, |
| "learning_rate": 8.580107580670163e-05, |
| "loss": 1.183, |
| "step": 2280 |
| }, |
| { |
| "epoch": 0.592122311479658, |
| "grad_norm": 0.37408901643032355, |
| "learning_rate": 8.535351702447524e-05, |
| "loss": 1.1624, |
| "step": 2285 |
| }, |
| { |
| "epoch": 0.5934179839336615, |
| "grad_norm": 0.4003666810623533, |
| "learning_rate": 8.49062578567287e-05, |
| "loss": 1.1193, |
| "step": 2290 |
| }, |
| { |
| "epoch": 0.5947136563876652, |
| "grad_norm": 0.3723417367446364, |
| "learning_rate": 8.445930745277953e-05, |
| "loss": 1.146, |
| "step": 2295 |
| }, |
| { |
| "epoch": 0.5960093288416688, |
| "grad_norm": 0.35036357860872624, |
| "learning_rate": 8.401267495562894e-05, |
| "loss": 1.0963, |
| "step": 2300 |
| }, |
| { |
| "epoch": 0.5973050012956724, |
| "grad_norm": 0.3969850486600124, |
| "learning_rate": 8.356636950177509e-05, |
| "loss": 1.1645, |
| "step": 2305 |
| }, |
| { |
| "epoch": 0.5986006737496761, |
| "grad_norm": 0.3908066674173964, |
| "learning_rate": 8.312040022102581e-05, |
| "loss": 1.1573, |
| "step": 2310 |
| }, |
| { |
| "epoch": 0.5998963462036797, |
| "grad_norm": 0.3970139193288793, |
| "learning_rate": 8.26747762363122e-05, |
| "loss": 1.1475, |
| "step": 2315 |
| }, |
| { |
| "epoch": 0.6011920186576833, |
| "grad_norm": 0.3809149209886327, |
| "learning_rate": 8.222950666350176e-05, |
| "loss": 1.0996, |
| "step": 2320 |
| }, |
| { |
| "epoch": 0.602487691111687, |
| "grad_norm": 0.3831688926512033, |
| "learning_rate": 8.1784600611212e-05, |
| "loss": 1.1586, |
| "step": 2325 |
| }, |
| { |
| "epoch": 0.6037833635656906, |
| "grad_norm": 0.37892711696966613, |
| "learning_rate": 8.134006718062417e-05, |
| "loss": 1.1394, |
| "step": 2330 |
| }, |
| { |
| "epoch": 0.6050790360196943, |
| "grad_norm": 0.3796251460883534, |
| "learning_rate": 8.089591546529686e-05, |
| "loss": 1.1342, |
| "step": 2335 |
| }, |
| { |
| "epoch": 0.6063747084736979, |
| "grad_norm": 0.3854947360629934, |
| "learning_rate": 8.04521545509803e-05, |
| "loss": 1.1234, |
| "step": 2340 |
| }, |
| { |
| "epoch": 0.6076703809277014, |
| "grad_norm": 0.39157237980540016, |
| "learning_rate": 8.00087935154303e-05, |
| "loss": 1.1789, |
| "step": 2345 |
| }, |
| { |
| "epoch": 0.6089660533817051, |
| "grad_norm": 0.37973961502373016, |
| "learning_rate": 7.956584142822248e-05, |
| "loss": 1.1725, |
| "step": 2350 |
| }, |
| { |
| "epoch": 0.6102617258357087, |
| "grad_norm": 0.4002316998884078, |
| "learning_rate": 7.912330735056702e-05, |
| "loss": 1.1761, |
| "step": 2355 |
| }, |
| { |
| "epoch": 0.6115573982897123, |
| "grad_norm": 0.3933317900533284, |
| "learning_rate": 7.868120033512294e-05, |
| "loss": 1.1279, |
| "step": 2360 |
| }, |
| { |
| "epoch": 0.612853070743716, |
| "grad_norm": 0.3994518364881897, |
| "learning_rate": 7.82395294258132e-05, |
| "loss": 1.1292, |
| "step": 2365 |
| }, |
| { |
| "epoch": 0.6141487431977196, |
| "grad_norm": 0.3774270174393722, |
| "learning_rate": 7.779830365763963e-05, |
| "loss": 1.1699, |
| "step": 2370 |
| }, |
| { |
| "epoch": 0.6154444156517233, |
| "grad_norm": 0.3936443342897645, |
| "learning_rate": 7.735753205649798e-05, |
| "loss": 1.1691, |
| "step": 2375 |
| }, |
| { |
| "epoch": 0.6167400881057269, |
| "grad_norm": 0.3958990632556032, |
| "learning_rate": 7.691722363899346e-05, |
| "loss": 1.1053, |
| "step": 2380 |
| }, |
| { |
| "epoch": 0.6180357605597305, |
| "grad_norm": 0.40113089490170173, |
| "learning_rate": 7.647738741225605e-05, |
| "loss": 1.0826, |
| "step": 2385 |
| }, |
| { |
| "epoch": 0.6193314330137342, |
| "grad_norm": 0.39930864888851864, |
| "learning_rate": 7.60380323737566e-05, |
| "loss": 1.199, |
| "step": 2390 |
| }, |
| { |
| "epoch": 0.6206271054677378, |
| "grad_norm": 0.3962521144461154, |
| "learning_rate": 7.559916751112248e-05, |
| "loss": 1.1933, |
| "step": 2395 |
| }, |
| { |
| "epoch": 0.6219227779217413, |
| "grad_norm": 0.37888926408124746, |
| "learning_rate": 7.516080180195379e-05, |
| "loss": 1.1713, |
| "step": 2400 |
| }, |
| { |
| "epoch": 0.623218450375745, |
| "grad_norm": 0.3708133442805237, |
| "learning_rate": 7.472294421363982e-05, |
| "loss": 1.2177, |
| "step": 2405 |
| }, |
| { |
| "epoch": 0.6245141228297486, |
| "grad_norm": 0.41709292403420495, |
| "learning_rate": 7.428560370317542e-05, |
| "loss": 1.1678, |
| "step": 2410 |
| }, |
| { |
| "epoch": 0.6258097952837522, |
| "grad_norm": 0.38822425697798224, |
| "learning_rate": 7.3848789216978e-05, |
| "loss": 1.1628, |
| "step": 2415 |
| }, |
| { |
| "epoch": 0.6271054677377559, |
| "grad_norm": 0.38293720555111366, |
| "learning_rate": 7.341250969070435e-05, |
| "loss": 1.1668, |
| "step": 2420 |
| }, |
| { |
| "epoch": 0.6284011401917595, |
| "grad_norm": 0.3982879136320185, |
| "learning_rate": 7.297677404906787e-05, |
| "loss": 1.1808, |
| "step": 2425 |
| }, |
| { |
| "epoch": 0.6296968126457632, |
| "grad_norm": 0.4970736508224388, |
| "learning_rate": 7.254159120565614e-05, |
| "loss": 1.1915, |
| "step": 2430 |
| }, |
| { |
| "epoch": 0.6309924850997668, |
| "grad_norm": 0.3859126967679283, |
| "learning_rate": 7.210697006274829e-05, |
| "loss": 1.1032, |
| "step": 2435 |
| }, |
| { |
| "epoch": 0.6322881575537704, |
| "grad_norm": 0.3804578357151546, |
| "learning_rate": 7.167291951113322e-05, |
| "loss": 1.1567, |
| "step": 2440 |
| }, |
| { |
| "epoch": 0.6335838300077741, |
| "grad_norm": 0.38190993552219266, |
| "learning_rate": 7.123944842992759e-05, |
| "loss": 1.091, |
| "step": 2445 |
| }, |
| { |
| "epoch": 0.6348795024617777, |
| "grad_norm": 0.38091685226387034, |
| "learning_rate": 7.080656568639406e-05, |
| "loss": 1.1457, |
| "step": 2450 |
| }, |
| { |
| "epoch": 0.6361751749157812, |
| "grad_norm": 0.3853362546606581, |
| "learning_rate": 7.037428013576013e-05, |
| "loss": 1.1463, |
| "step": 2455 |
| }, |
| { |
| "epoch": 0.6374708473697849, |
| "grad_norm": 0.36970236476000445, |
| "learning_rate": 6.994260062103674e-05, |
| "loss": 1.1393, |
| "step": 2460 |
| }, |
| { |
| "epoch": 0.6387665198237885, |
| "grad_norm": 0.40400326202643533, |
| "learning_rate": 6.951153597283766e-05, |
| "loss": 1.1295, |
| "step": 2465 |
| }, |
| { |
| "epoch": 0.6400621922777922, |
| "grad_norm": 0.3816874564299913, |
| "learning_rate": 6.908109500919861e-05, |
| "loss": 1.1662, |
| "step": 2470 |
| }, |
| { |
| "epoch": 0.6413578647317958, |
| "grad_norm": 0.3799563658685613, |
| "learning_rate": 6.865128653539699e-05, |
| "loss": 1.1876, |
| "step": 2475 |
| }, |
| { |
| "epoch": 0.6426535371857994, |
| "grad_norm": 0.379580454677419, |
| "learning_rate": 6.822211934377176e-05, |
| "loss": 1.1005, |
| "step": 2480 |
| }, |
| { |
| "epoch": 0.6439492096398031, |
| "grad_norm": 0.36403149625971076, |
| "learning_rate": 6.77936022135434e-05, |
| "loss": 1.0924, |
| "step": 2485 |
| }, |
| { |
| "epoch": 0.6452448820938067, |
| "grad_norm": 0.3864263981944349, |
| "learning_rate": 6.736574391063466e-05, |
| "loss": 1.166, |
| "step": 2490 |
| }, |
| { |
| "epoch": 0.6465405545478103, |
| "grad_norm": 0.4278354770987037, |
| "learning_rate": 6.693855318749096e-05, |
| "loss": 1.1463, |
| "step": 2495 |
| }, |
| { |
| "epoch": 0.647836227001814, |
| "grad_norm": 0.3881716691518243, |
| "learning_rate": 6.651203878290139e-05, |
| "loss": 1.1658, |
| "step": 2500 |
| }, |
| { |
| "epoch": 0.6491318994558176, |
| "grad_norm": 0.3773480140369342, |
| "learning_rate": 6.608620942182011e-05, |
| "loss": 1.1561, |
| "step": 2505 |
| }, |
| { |
| "epoch": 0.6504275719098211, |
| "grad_norm": 0.38362919406181467, |
| "learning_rate": 6.566107381518758e-05, |
| "loss": 1.1553, |
| "step": 2510 |
| }, |
| { |
| "epoch": 0.6517232443638248, |
| "grad_norm": 0.3889927393341413, |
| "learning_rate": 6.523664065975268e-05, |
| "loss": 1.0927, |
| "step": 2515 |
| }, |
| { |
| "epoch": 0.6530189168178284, |
| "grad_norm": 0.3770322611579158, |
| "learning_rate": 6.481291863789461e-05, |
| "loss": 1.1621, |
| "step": 2520 |
| }, |
| { |
| "epoch": 0.6543145892718321, |
| "grad_norm": 0.3583870387014761, |
| "learning_rate": 6.43899164174453e-05, |
| "loss": 1.1688, |
| "step": 2525 |
| }, |
| { |
| "epoch": 0.6556102617258357, |
| "grad_norm": 0.36282391266493386, |
| "learning_rate": 6.396764265151221e-05, |
| "loss": 1.0909, |
| "step": 2530 |
| }, |
| { |
| "epoch": 0.6569059341798393, |
| "grad_norm": 0.40358089248833323, |
| "learning_rate": 6.35461059783011e-05, |
| "loss": 1.185, |
| "step": 2535 |
| }, |
| { |
| "epoch": 0.658201606633843, |
| "grad_norm": 0.3622883465103346, |
| "learning_rate": 6.312531502093958e-05, |
| "loss": 1.1586, |
| "step": 2540 |
| }, |
| { |
| "epoch": 0.6594972790878466, |
| "grad_norm": 0.39189718464338913, |
| "learning_rate": 6.270527838730053e-05, |
| "loss": 1.1592, |
| "step": 2545 |
| }, |
| { |
| "epoch": 0.6607929515418502, |
| "grad_norm": 0.36779638890391525, |
| "learning_rate": 6.228600466982611e-05, |
| "loss": 1.156, |
| "step": 2550 |
| }, |
| { |
| "epoch": 0.6620886239958539, |
| "grad_norm": 0.37597682220271855, |
| "learning_rate": 6.186750244535194e-05, |
| "loss": 1.1353, |
| "step": 2555 |
| }, |
| { |
| "epoch": 0.6633842964498575, |
| "grad_norm": 0.3886787309603002, |
| "learning_rate": 6.144978027493158e-05, |
| "loss": 1.1341, |
| "step": 2560 |
| }, |
| { |
| "epoch": 0.664679968903861, |
| "grad_norm": 0.35942315304405126, |
| "learning_rate": 6.103284670366162e-05, |
| "loss": 1.1597, |
| "step": 2565 |
| }, |
| { |
| "epoch": 0.6659756413578647, |
| "grad_norm": 0.41146010175047676, |
| "learning_rate": 6.061671026050668e-05, |
| "loss": 1.18, |
| "step": 2570 |
| }, |
| { |
| "epoch": 0.6672713138118683, |
| "grad_norm": 0.38590903226923434, |
| "learning_rate": 6.0201379458124964e-05, |
| "loss": 1.1639, |
| "step": 2575 |
| }, |
| { |
| "epoch": 0.668566986265872, |
| "grad_norm": 0.386535230930437, |
| "learning_rate": 5.978686279269421e-05, |
| "loss": 1.0866, |
| "step": 2580 |
| }, |
| { |
| "epoch": 0.6698626587198756, |
| "grad_norm": 0.3877441934910258, |
| "learning_rate": 5.9373168743737864e-05, |
| "loss": 1.1325, |
| "step": 2585 |
| }, |
| { |
| "epoch": 0.6711583311738792, |
| "grad_norm": 0.3990007152806006, |
| "learning_rate": 5.896030577395144e-05, |
| "loss": 1.1188, |
| "step": 2590 |
| }, |
| { |
| "epoch": 0.6724540036278829, |
| "grad_norm": 0.3970387094956777, |
| "learning_rate": 5.854828232902976e-05, |
| "loss": 1.1939, |
| "step": 2595 |
| }, |
| { |
| "epoch": 0.6737496760818865, |
| "grad_norm": 0.39839395255304766, |
| "learning_rate": 5.813710683749379e-05, |
| "loss": 1.2078, |
| "step": 2600 |
| }, |
| { |
| "epoch": 0.6750453485358902, |
| "grad_norm": 0.39807324424891044, |
| "learning_rate": 5.772678771051858e-05, |
| "loss": 1.1513, |
| "step": 2605 |
| }, |
| { |
| "epoch": 0.6763410209898938, |
| "grad_norm": 0.371233691663641, |
| "learning_rate": 5.7317333341760906e-05, |
| "loss": 1.1896, |
| "step": 2610 |
| }, |
| { |
| "epoch": 0.6776366934438974, |
| "grad_norm": 0.3914320864664493, |
| "learning_rate": 5.690875210718778e-05, |
| "loss": 1.144, |
| "step": 2615 |
| }, |
| { |
| "epoch": 0.678932365897901, |
| "grad_norm": 0.38952180289646227, |
| "learning_rate": 5.6501052364904906e-05, |
| "loss": 1.1669, |
| "step": 2620 |
| }, |
| { |
| "epoch": 0.6802280383519046, |
| "grad_norm": 0.3623267889899647, |
| "learning_rate": 5.6094242454986e-05, |
| "loss": 1.179, |
| "step": 2625 |
| }, |
| { |
| "epoch": 0.6815237108059082, |
| "grad_norm": 0.3637529461983289, |
| "learning_rate": 5.568833069930186e-05, |
| "loss": 1.1244, |
| "step": 2630 |
| }, |
| { |
| "epoch": 0.6828193832599119, |
| "grad_norm": 0.387937514120572, |
| "learning_rate": 5.528332540135031e-05, |
| "loss": 1.1699, |
| "step": 2635 |
| }, |
| { |
| "epoch": 0.6841150557139155, |
| "grad_norm": 0.37878126929029915, |
| "learning_rate": 5.487923484608629e-05, |
| "loss": 1.131, |
| "step": 2640 |
| }, |
| { |
| "epoch": 0.6854107281679191, |
| "grad_norm": 0.3903173526956151, |
| "learning_rate": 5.4476067299752385e-05, |
| "loss": 1.1165, |
| "step": 2645 |
| }, |
| { |
| "epoch": 0.6867064006219228, |
| "grad_norm": 0.3857342653067374, |
| "learning_rate": 5.4073831009709805e-05, |
| "loss": 1.0854, |
| "step": 2650 |
| }, |
| { |
| "epoch": 0.6880020730759264, |
| "grad_norm": 0.3661818366505174, |
| "learning_rate": 5.367253420426952e-05, |
| "loss": 1.173, |
| "step": 2655 |
| }, |
| { |
| "epoch": 0.6892977455299301, |
| "grad_norm": 0.3695899768899783, |
| "learning_rate": 5.3272185092524004e-05, |
| "loss": 1.1917, |
| "step": 2660 |
| }, |
| { |
| "epoch": 0.6905934179839337, |
| "grad_norm": 0.3709048127868108, |
| "learning_rate": 5.287279186417938e-05, |
| "loss": 1.1595, |
| "step": 2665 |
| }, |
| { |
| "epoch": 0.6918890904379373, |
| "grad_norm": 0.388715290668864, |
| "learning_rate": 5.2474362689387745e-05, |
| "loss": 1.1681, |
| "step": 2670 |
| }, |
| { |
| "epoch": 0.6931847628919409, |
| "grad_norm": 0.41572003898003435, |
| "learning_rate": 5.207690571858025e-05, |
| "loss": 1.1344, |
| "step": 2675 |
| }, |
| { |
| "epoch": 0.6944804353459445, |
| "grad_norm": 0.3688637831571822, |
| "learning_rate": 5.1680429082300134e-05, |
| "loss": 1.1911, |
| "step": 2680 |
| }, |
| { |
| "epoch": 0.6957761077999481, |
| "grad_norm": 0.3793170468762598, |
| "learning_rate": 5.128494089103652e-05, |
| "loss": 1.1127, |
| "step": 2685 |
| }, |
| { |
| "epoch": 0.6970717802539518, |
| "grad_norm": 0.43134296541622413, |
| "learning_rate": 5.0890449235058525e-05, |
| "loss": 1.1784, |
| "step": 2690 |
| }, |
| { |
| "epoch": 0.6983674527079554, |
| "grad_norm": 0.3635096529171273, |
| "learning_rate": 5.049696218424969e-05, |
| "loss": 1.1428, |
| "step": 2695 |
| }, |
| { |
| "epoch": 0.699663125161959, |
| "grad_norm": 0.3878749986675832, |
| "learning_rate": 5.010448778794303e-05, |
| "loss": 1.1741, |
| "step": 2700 |
| }, |
| { |
| "epoch": 0.7009587976159627, |
| "grad_norm": 0.40169342624996973, |
| "learning_rate": 4.971303407475618e-05, |
| "loss": 1.1599, |
| "step": 2705 |
| }, |
| { |
| "epoch": 0.7022544700699663, |
| "grad_norm": 0.3694514584240007, |
| "learning_rate": 4.932260905242731e-05, |
| "loss": 1.1187, |
| "step": 2710 |
| }, |
| { |
| "epoch": 0.70355014252397, |
| "grad_norm": 0.40107470744868656, |
| "learning_rate": 4.893322070765126e-05, |
| "loss": 1.1087, |
| "step": 2715 |
| }, |
| { |
| "epoch": 0.7048458149779736, |
| "grad_norm": 0.3973309016316779, |
| "learning_rate": 4.8544877005916126e-05, |
| "loss": 1.1353, |
| "step": 2720 |
| }, |
| { |
| "epoch": 0.7061414874319772, |
| "grad_norm": 0.36945482581363814, |
| "learning_rate": 4.815758589134046e-05, |
| "loss": 1.1537, |
| "step": 2725 |
| }, |
| { |
| "epoch": 0.7074371598859808, |
| "grad_norm": 0.39405014852767495, |
| "learning_rate": 4.777135528651058e-05, |
| "loss": 1.1864, |
| "step": 2730 |
| }, |
| { |
| "epoch": 0.7087328323399844, |
| "grad_norm": 0.36438744458294, |
| "learning_rate": 4.738619309231857e-05, |
| "loss": 1.1571, |
| "step": 2735 |
| }, |
| { |
| "epoch": 0.710028504793988, |
| "grad_norm": 0.3718851263620155, |
| "learning_rate": 4.700210718780072e-05, |
| "loss": 1.1375, |
| "step": 2740 |
| }, |
| { |
| "epoch": 0.7113241772479917, |
| "grad_norm": 0.3968100064213781, |
| "learning_rate": 4.6619105429976193e-05, |
| "loss": 1.0992, |
| "step": 2745 |
| }, |
| { |
| "epoch": 0.7126198497019953, |
| "grad_norm": 0.3735407865710053, |
| "learning_rate": 4.623719565368657e-05, |
| "loss": 1.154, |
| "step": 2750 |
| }, |
| { |
| "epoch": 0.713915522155999, |
| "grad_norm": 0.3855769481718261, |
| "learning_rate": 4.585638567143529e-05, |
| "loss": 1.1087, |
| "step": 2755 |
| }, |
| { |
| "epoch": 0.7152111946100026, |
| "grad_norm": 0.39904660177632, |
| "learning_rate": 4.547668327322796e-05, |
| "loss": 1.1557, |
| "step": 2760 |
| }, |
| { |
| "epoch": 0.7165068670640062, |
| "grad_norm": 0.37552561838186715, |
| "learning_rate": 4.5098096226413e-05, |
| "loss": 1.1788, |
| "step": 2765 |
| }, |
| { |
| "epoch": 0.7178025395180099, |
| "grad_norm": 0.39324730815800873, |
| "learning_rate": 4.472063227552274e-05, |
| "loss": 1.1688, |
| "step": 2770 |
| }, |
| { |
| "epoch": 0.7190982119720135, |
| "grad_norm": 0.38301296634310333, |
| "learning_rate": 4.434429914211508e-05, |
| "loss": 1.1711, |
| "step": 2775 |
| }, |
| { |
| "epoch": 0.7203938844260172, |
| "grad_norm": 0.3709443062095975, |
| "learning_rate": 4.396910452461532e-05, |
| "loss": 1.1718, |
| "step": 2780 |
| }, |
| { |
| "epoch": 0.7216895568800207, |
| "grad_norm": 0.39239048686699185, |
| "learning_rate": 4.3595056098158906e-05, |
| "loss": 1.1574, |
| "step": 2785 |
| }, |
| { |
| "epoch": 0.7229852293340243, |
| "grad_norm": 0.41357484414155377, |
| "learning_rate": 4.322216151443428e-05, |
| "loss": 1.1074, |
| "step": 2790 |
| }, |
| { |
| "epoch": 0.724280901788028, |
| "grad_norm": 0.3672227895875153, |
| "learning_rate": 4.2850428401526376e-05, |
| "loss": 1.1028, |
| "step": 2795 |
| }, |
| { |
| "epoch": 0.7255765742420316, |
| "grad_norm": 0.3664872416848814, |
| "learning_rate": 4.2479864363760726e-05, |
| "loss": 1.1508, |
| "step": 2800 |
| }, |
| { |
| "epoch": 0.7268722466960352, |
| "grad_norm": 0.3710822782612391, |
| "learning_rate": 4.211047698154765e-05, |
| "loss": 1.1369, |
| "step": 2805 |
| }, |
| { |
| "epoch": 0.7281679191500389, |
| "grad_norm": 0.38199563385579133, |
| "learning_rate": 4.1742273811227395e-05, |
| "loss": 1.1716, |
| "step": 2810 |
| }, |
| { |
| "epoch": 0.7294635916040425, |
| "grad_norm": 0.36978932404422615, |
| "learning_rate": 4.1375262384915433e-05, |
| "loss": 1.1613, |
| "step": 2815 |
| }, |
| { |
| "epoch": 0.7307592640580461, |
| "grad_norm": 0.37568319756030377, |
| "learning_rate": 4.100945021034843e-05, |
| "loss": 1.1396, |
| "step": 2820 |
| }, |
| { |
| "epoch": 0.7320549365120498, |
| "grad_norm": 0.3802598263999195, |
| "learning_rate": 4.064484477073074e-05, |
| "loss": 1.172, |
| "step": 2825 |
| }, |
| { |
| "epoch": 0.7333506089660534, |
| "grad_norm": 0.3690252573947318, |
| "learning_rate": 4.028145352458118e-05, |
| "loss": 1.1413, |
| "step": 2830 |
| }, |
| { |
| "epoch": 0.7346462814200571, |
| "grad_norm": 0.3942754823686454, |
| "learning_rate": 3.991928390558054e-05, |
| "loss": 1.1069, |
| "step": 2835 |
| }, |
| { |
| "epoch": 0.7359419538740606, |
| "grad_norm": 0.3720821371337478, |
| "learning_rate": 3.95583433224195e-05, |
| "loss": 1.0835, |
| "step": 2840 |
| }, |
| { |
| "epoch": 0.7372376263280642, |
| "grad_norm": 0.3807303246565132, |
| "learning_rate": 3.9198639158647056e-05, |
| "loss": 1.1501, |
| "step": 2845 |
| }, |
| { |
| "epoch": 0.7385332987820679, |
| "grad_norm": 0.36231750839777666, |
| "learning_rate": 3.884017877251959e-05, |
| "loss": 1.0563, |
| "step": 2850 |
| }, |
| { |
| "epoch": 0.7398289712360715, |
| "grad_norm": 0.388874708381432, |
| "learning_rate": 3.8482969496850166e-05, |
| "loss": 1.0889, |
| "step": 2855 |
| }, |
| { |
| "epoch": 0.7411246436900751, |
| "grad_norm": 0.38013479771110215, |
| "learning_rate": 3.812701863885865e-05, |
| "loss": 1.1145, |
| "step": 2860 |
| }, |
| { |
| "epoch": 0.7424203161440788, |
| "grad_norm": 0.3717583300327742, |
| "learning_rate": 3.7772333480022185e-05, |
| "loss": 1.0663, |
| "step": 2865 |
| }, |
| { |
| "epoch": 0.7437159885980824, |
| "grad_norm": 0.36405208418925755, |
| "learning_rate": 3.741892127592625e-05, |
| "loss": 1.1102, |
| "step": 2870 |
| }, |
| { |
| "epoch": 0.745011661052086, |
| "grad_norm": 0.3723066034084609, |
| "learning_rate": 3.706678925611629e-05, |
| "loss": 1.1602, |
| "step": 2875 |
| }, |
| { |
| "epoch": 0.7463073335060897, |
| "grad_norm": 0.3783951434167031, |
| "learning_rate": 3.67159446239497e-05, |
| "loss": 1.1233, |
| "step": 2880 |
| }, |
| { |
| "epoch": 0.7476030059600933, |
| "grad_norm": 0.37951067018123125, |
| "learning_rate": 3.636639455644858e-05, |
| "loss": 1.0545, |
| "step": 2885 |
| }, |
| { |
| "epoch": 0.748898678414097, |
| "grad_norm": 0.37421338215744115, |
| "learning_rate": 3.601814620415296e-05, |
| "loss": 1.1367, |
| "step": 2890 |
| }, |
| { |
| "epoch": 0.7501943508681005, |
| "grad_norm": 0.3603159123651838, |
| "learning_rate": 3.567120669097422e-05, |
| "loss": 1.0909, |
| "step": 2895 |
| }, |
| { |
| "epoch": 0.7514900233221041, |
| "grad_norm": 0.3788231837374063, |
| "learning_rate": 3.532558311404986e-05, |
| "loss": 1.1441, |
| "step": 2900 |
| }, |
| { |
| "epoch": 0.7527856957761078, |
| "grad_norm": 0.3840910679589615, |
| "learning_rate": 3.498128254359788e-05, |
| "loss": 1.1403, |
| "step": 2905 |
| }, |
| { |
| "epoch": 0.7540813682301114, |
| "grad_norm": 0.3753865209395276, |
| "learning_rate": 3.4638312022772335e-05, |
| "loss": 1.1004, |
| "step": 2910 |
| }, |
| { |
| "epoch": 0.755377040684115, |
| "grad_norm": 0.3954242259257984, |
| "learning_rate": 3.4296678567519345e-05, |
| "loss": 1.1347, |
| "step": 2915 |
| }, |
| { |
| "epoch": 0.7566727131381187, |
| "grad_norm": 0.3741932569564154, |
| "learning_rate": 3.3956389166433276e-05, |
| "loss": 1.0908, |
| "step": 2920 |
| }, |
| { |
| "epoch": 0.7579683855921223, |
| "grad_norm": 0.3930950040373874, |
| "learning_rate": 3.361745078061416e-05, |
| "loss": 1.1988, |
| "step": 2925 |
| }, |
| { |
| "epoch": 0.759264058046126, |
| "grad_norm": 0.3685363645101882, |
| "learning_rate": 3.327987034352499e-05, |
| "loss": 1.2067, |
| "step": 2930 |
| }, |
| { |
| "epoch": 0.7605597305001296, |
| "grad_norm": 0.3595878774367224, |
| "learning_rate": 3.294365476085001e-05, |
| "loss": 1.1128, |
| "step": 2935 |
| }, |
| { |
| "epoch": 0.7618554029541332, |
| "grad_norm": 0.3823858203204671, |
| "learning_rate": 3.260881091035356e-05, |
| "loss": 1.1894, |
| "step": 2940 |
| }, |
| { |
| "epoch": 0.7631510754081369, |
| "grad_norm": 0.35550275725905045, |
| "learning_rate": 3.227534564173903e-05, |
| "loss": 1.0744, |
| "step": 2945 |
| }, |
| { |
| "epoch": 0.7644467478621404, |
| "grad_norm": 0.38923682484253824, |
| "learning_rate": 3.1943265776509215e-05, |
| "loss": 1.1599, |
| "step": 2950 |
| }, |
| { |
| "epoch": 0.765742420316144, |
| "grad_norm": 0.3789384247163881, |
| "learning_rate": 3.1612578107826375e-05, |
| "loss": 1.1504, |
| "step": 2955 |
| }, |
| { |
| "epoch": 0.7670380927701477, |
| "grad_norm": 0.3620246203990784, |
| "learning_rate": 3.128328940037345e-05, |
| "loss": 1.0777, |
| "step": 2960 |
| }, |
| { |
| "epoch": 0.7683337652241513, |
| "grad_norm": 0.36815557923240855, |
| "learning_rate": 3.095540639021578e-05, |
| "loss": 1.1174, |
| "step": 2965 |
| }, |
| { |
| "epoch": 0.769629437678155, |
| "grad_norm": 0.3547702710213197, |
| "learning_rate": 3.0628935784662947e-05, |
| "loss": 1.1485, |
| "step": 2970 |
| }, |
| { |
| "epoch": 0.7709251101321586, |
| "grad_norm": 0.3781571691176777, |
| "learning_rate": 3.030388426213202e-05, |
| "loss": 1.1883, |
| "step": 2975 |
| }, |
| { |
| "epoch": 0.7722207825861622, |
| "grad_norm": 0.40077870265330406, |
| "learning_rate": 2.9980258472010624e-05, |
| "loss": 1.1932, |
| "step": 2980 |
| }, |
| { |
| "epoch": 0.7735164550401659, |
| "grad_norm": 0.4044666107500387, |
| "learning_rate": 2.9658065034520978e-05, |
| "loss": 1.1241, |
| "step": 2985 |
| }, |
| { |
| "epoch": 0.7748121274941695, |
| "grad_norm": 0.38599774202480436, |
| "learning_rate": 2.9337310540584662e-05, |
| "loss": 1.1289, |
| "step": 2990 |
| }, |
| { |
| "epoch": 0.7761077999481731, |
| "grad_norm": 0.3937959908311204, |
| "learning_rate": 2.90180015516874e-05, |
| "loss": 1.124, |
| "step": 2995 |
| }, |
| { |
| "epoch": 0.7774034724021768, |
| "grad_norm": 0.3632550787528134, |
| "learning_rate": 2.8700144599745304e-05, |
| "loss": 1.054, |
| "step": 3000 |
| }, |
| { |
| "epoch": 0.7786991448561803, |
| "grad_norm": 0.3622147641579977, |
| "learning_rate": 2.8383746186970885e-05, |
| "loss": 1.1541, |
| "step": 3005 |
| }, |
| { |
| "epoch": 0.7799948173101839, |
| "grad_norm": 0.37112400149549435, |
| "learning_rate": 2.806881278574016e-05, |
| "loss": 1.0959, |
| "step": 3010 |
| }, |
| { |
| "epoch": 0.7812904897641876, |
| "grad_norm": 0.35896620554471, |
| "learning_rate": 2.7755350838460437e-05, |
| "loss": 1.1097, |
| "step": 3015 |
| }, |
| { |
| "epoch": 0.7825861622181912, |
| "grad_norm": 0.3722645573928852, |
| "learning_rate": 2.7443366757438084e-05, |
| "loss": 1.1536, |
| "step": 3020 |
| }, |
| { |
| "epoch": 0.7838818346721949, |
| "grad_norm": 0.3733728824194641, |
| "learning_rate": 2.71328669247479e-05, |
| "loss": 1.0988, |
| "step": 3025 |
| }, |
| { |
| "epoch": 0.7851775071261985, |
| "grad_norm": 0.39726305827044567, |
| "learning_rate": 2.6823857692102115e-05, |
| "loss": 1.1347, |
| "step": 3030 |
| }, |
| { |
| "epoch": 0.7864731795802021, |
| "grad_norm": 0.3757164187600116, |
| "learning_rate": 2.6516345380720685e-05, |
| "loss": 1.1498, |
| "step": 3035 |
| }, |
| { |
| "epoch": 0.7877688520342058, |
| "grad_norm": 0.3795959004315457, |
| "learning_rate": 2.6210336281201996e-05, |
| "loss": 1.1716, |
| "step": 3040 |
| }, |
| { |
| "epoch": 0.7890645244882094, |
| "grad_norm": 0.35714774223397056, |
| "learning_rate": 2.5905836653393955e-05, |
| "loss": 1.1829, |
| "step": 3045 |
| }, |
| { |
| "epoch": 0.790360196942213, |
| "grad_norm": 0.3654895953286083, |
| "learning_rate": 2.5602852726266246e-05, |
| "loss": 1.1138, |
| "step": 3050 |
| }, |
| { |
| "epoch": 0.7916558693962167, |
| "grad_norm": 0.3790228967067868, |
| "learning_rate": 2.53013906977827e-05, |
| "loss": 1.1361, |
| "step": 3055 |
| }, |
| { |
| "epoch": 0.7929515418502202, |
| "grad_norm": 0.3766214688716055, |
| "learning_rate": 2.500145673477452e-05, |
| "loss": 1.1386, |
| "step": 3060 |
| }, |
| { |
| "epoch": 0.7942472143042238, |
| "grad_norm": 0.35689403644408657, |
| "learning_rate": 2.4703056972814298e-05, |
| "loss": 1.1716, |
| "step": 3065 |
| }, |
| { |
| "epoch": 0.7955428867582275, |
| "grad_norm": 0.3837337630770634, |
| "learning_rate": 2.44061975160902e-05, |
| "loss": 1.1253, |
| "step": 3070 |
| }, |
| { |
| "epoch": 0.7968385592122311, |
| "grad_norm": 0.3617908256654209, |
| "learning_rate": 2.4110884437281433e-05, |
| "loss": 1.1669, |
| "step": 3075 |
| }, |
| { |
| "epoch": 0.7981342316662348, |
| "grad_norm": 0.35978473851190906, |
| "learning_rate": 2.381712377743379e-05, |
| "loss": 1.1195, |
| "step": 3080 |
| }, |
| { |
| "epoch": 0.7994299041202384, |
| "grad_norm": 0.36747445883447966, |
| "learning_rate": 2.352492154583611e-05, |
| "loss": 1.1494, |
| "step": 3085 |
| }, |
| { |
| "epoch": 0.800725576574242, |
| "grad_norm": 0.38717115547376924, |
| "learning_rate": 2.323428371989752e-05, |
| "loss": 1.1342, |
| "step": 3090 |
| }, |
| { |
| "epoch": 0.8020212490282457, |
| "grad_norm": 0.39093054065016897, |
| "learning_rate": 2.2945216245024804e-05, |
| "loss": 1.1304, |
| "step": 3095 |
| }, |
| { |
| "epoch": 0.8033169214822493, |
| "grad_norm": 0.368756140981371, |
| "learning_rate": 2.265772503450122e-05, |
| "loss": 1.1902, |
| "step": 3100 |
| }, |
| { |
| "epoch": 0.804612593936253, |
| "grad_norm": 0.3877482016759796, |
| "learning_rate": 2.237181596936515e-05, |
| "loss": 1.2071, |
| "step": 3105 |
| }, |
| { |
| "epoch": 0.8059082663902566, |
| "grad_norm": 0.37792584298436904, |
| "learning_rate": 2.2087494898290084e-05, |
| "loss": 1.1263, |
| "step": 3110 |
| }, |
| { |
| "epoch": 0.8072039388442601, |
| "grad_norm": 0.377175581709425, |
| "learning_rate": 2.1804767637464783e-05, |
| "loss": 1.1409, |
| "step": 3115 |
| }, |
| { |
| "epoch": 0.8084996112982638, |
| "grad_norm": 0.3675212675701305, |
| "learning_rate": 2.152363997047432e-05, |
| "loss": 1.1222, |
| "step": 3120 |
| }, |
| { |
| "epoch": 0.8097952837522674, |
| "grad_norm": 0.38222618034509426, |
| "learning_rate": 2.1244117648181926e-05, |
| "loss": 1.1282, |
| "step": 3125 |
| }, |
| { |
| "epoch": 0.811090956206271, |
| "grad_norm": 0.36558116292433085, |
| "learning_rate": 2.0966206388611177e-05, |
| "loss": 1.1599, |
| "step": 3130 |
| }, |
| { |
| "epoch": 0.8123866286602747, |
| "grad_norm": 0.3727524846387591, |
| "learning_rate": 2.0689911876829127e-05, |
| "loss": 1.1496, |
| "step": 3135 |
| }, |
| { |
| "epoch": 0.8136823011142783, |
| "grad_norm": 0.3628037534722254, |
| "learning_rate": 2.0415239764829976e-05, |
| "loss": 1.1416, |
| "step": 3140 |
| }, |
| { |
| "epoch": 0.8149779735682819, |
| "grad_norm": 0.3761023568840631, |
| "learning_rate": 2.014219567141932e-05, |
| "loss": 1.0828, |
| "step": 3145 |
| }, |
| { |
| "epoch": 0.8162736460222856, |
| "grad_norm": 0.3504341651004141, |
| "learning_rate": 1.9870785182099505e-05, |
| "loss": 1.0956, |
| "step": 3150 |
| }, |
| { |
| "epoch": 0.8175693184762892, |
| "grad_norm": 0.3734885666536526, |
| "learning_rate": 1.960101384895511e-05, |
| "loss": 1.1044, |
| "step": 3155 |
| }, |
| { |
| "epoch": 0.8188649909302929, |
| "grad_norm": 0.386535956936462, |
| "learning_rate": 1.9332887190539516e-05, |
| "loss": 1.2004, |
| "step": 3160 |
| }, |
| { |
| "epoch": 0.8201606633842965, |
| "grad_norm": 0.3646373120550804, |
| "learning_rate": 1.9066410691761937e-05, |
| "loss": 1.1377, |
| "step": 3165 |
| }, |
| { |
| "epoch": 0.8214563358383, |
| "grad_norm": 0.3715486267985764, |
| "learning_rate": 1.8801589803775154e-05, |
| "loss": 1.1128, |
| "step": 3170 |
| }, |
| { |
| "epoch": 0.8227520082923037, |
| "grad_norm": 0.3789134249781072, |
| "learning_rate": 1.8538429943864244e-05, |
| "loss": 1.1429, |
| "step": 3175 |
| }, |
| { |
| "epoch": 0.8240476807463073, |
| "grad_norm": 0.4352073640959862, |
| "learning_rate": 1.8276936495335485e-05, |
| "loss": 1.1167, |
| "step": 3180 |
| }, |
| { |
| "epoch": 0.8253433532003109, |
| "grad_norm": 0.3681435590124036, |
| "learning_rate": 1.8017114807406478e-05, |
| "loss": 1.1063, |
| "step": 3185 |
| }, |
| { |
| "epoch": 0.8266390256543146, |
| "grad_norm": 0.3694513270515317, |
| "learning_rate": 1.775897019509649e-05, |
| "loss": 1.1836, |
| "step": 3190 |
| }, |
| { |
| "epoch": 0.8279346981083182, |
| "grad_norm": 0.388666401561133, |
| "learning_rate": 1.7502507939117897e-05, |
| "loss": 1.1396, |
| "step": 3195 |
| }, |
| { |
| "epoch": 0.8292303705623219, |
| "grad_norm": 0.3687281974613644, |
| "learning_rate": 1.7247733285768098e-05, |
| "loss": 1.1352, |
| "step": 3200 |
| }, |
| { |
| "epoch": 0.8305260430163255, |
| "grad_norm": 0.3733462546424826, |
| "learning_rate": 1.6994651446822153e-05, |
| "loss": 1.125, |
| "step": 3205 |
| }, |
| { |
| "epoch": 0.8318217154703291, |
| "grad_norm": 0.3745095547011446, |
| "learning_rate": 1.6743267599426303e-05, |
| "loss": 1.1645, |
| "step": 3210 |
| }, |
| { |
| "epoch": 0.8331173879243328, |
| "grad_norm": 0.36579848745894494, |
| "learning_rate": 1.649358688599191e-05, |
| "loss": 1.1373, |
| "step": 3215 |
| }, |
| { |
| "epoch": 0.8344130603783364, |
| "grad_norm": 0.37060439276258045, |
| "learning_rate": 1.624561441409034e-05, |
| "loss": 1.0775, |
| "step": 3220 |
| }, |
| { |
| "epoch": 0.83570873283234, |
| "grad_norm": 0.38418584872461575, |
| "learning_rate": 1.5999355256348448e-05, |
| "loss": 1.1678, |
| "step": 3225 |
| }, |
| { |
| "epoch": 0.8370044052863436, |
| "grad_norm": 0.3814911739387983, |
| "learning_rate": 1.5754814450344845e-05, |
| "loss": 1.1772, |
| "step": 3230 |
| }, |
| { |
| "epoch": 0.8383000777403472, |
| "grad_norm": 0.3642771788771785, |
| "learning_rate": 1.5511996998506883e-05, |
| "loss": 1.1279, |
| "step": 3235 |
| }, |
| { |
| "epoch": 0.8395957501943508, |
| "grad_norm": 0.3798091083229991, |
| "learning_rate": 1.527090786800821e-05, |
| "loss": 1.1663, |
| "step": 3240 |
| }, |
| { |
| "epoch": 0.8408914226483545, |
| "grad_norm": 0.3621301040720381, |
| "learning_rate": 1.5031551990667236e-05, |
| "loss": 1.0961, |
| "step": 3245 |
| }, |
| { |
| "epoch": 0.8421870951023581, |
| "grad_norm": 0.3716900347036682, |
| "learning_rate": 1.4793934262846232e-05, |
| "loss": 1.1246, |
| "step": 3250 |
| }, |
| { |
| "epoch": 0.8434827675563618, |
| "grad_norm": 0.3431347363725958, |
| "learning_rate": 1.4558059545351143e-05, |
| "loss": 1.126, |
| "step": 3255 |
| }, |
| { |
| "epoch": 0.8447784400103654, |
| "grad_norm": 0.38516161930376763, |
| "learning_rate": 1.4323932663332251e-05, |
| "loss": 1.1487, |
| "step": 3260 |
| }, |
| { |
| "epoch": 0.846074112464369, |
| "grad_norm": 0.36582214467557017, |
| "learning_rate": 1.4091558406185335e-05, |
| "loss": 1.1464, |
| "step": 3265 |
| }, |
| { |
| "epoch": 0.8473697849183727, |
| "grad_norm": 0.3600375664520519, |
| "learning_rate": 1.3860941527453786e-05, |
| "loss": 1.1331, |
| "step": 3270 |
| }, |
| { |
| "epoch": 0.8486654573723763, |
| "grad_norm": 0.36937163487487523, |
| "learning_rate": 1.3632086744731299e-05, |
| "loss": 1.1, |
| "step": 3275 |
| }, |
| { |
| "epoch": 0.84996112982638, |
| "grad_norm": 0.37811170010742284, |
| "learning_rate": 1.3404998739565433e-05, |
| "loss": 1.1278, |
| "step": 3280 |
| }, |
| { |
| "epoch": 0.8512568022803835, |
| "grad_norm": 0.36422197248748484, |
| "learning_rate": 1.3179682157361872e-05, |
| "loss": 1.1193, |
| "step": 3285 |
| }, |
| { |
| "epoch": 0.8525524747343871, |
| "grad_norm": 0.37165480607350426, |
| "learning_rate": 1.2956141607289262e-05, |
| "loss": 1.1615, |
| "step": 3290 |
| }, |
| { |
| "epoch": 0.8538481471883907, |
| "grad_norm": 0.37701311154160283, |
| "learning_rate": 1.2734381662185035e-05, |
| "loss": 1.1036, |
| "step": 3295 |
| }, |
| { |
| "epoch": 0.8551438196423944, |
| "grad_norm": 0.3815170151959314, |
| "learning_rate": 1.2514406858461847e-05, |
| "loss": 1.1735, |
| "step": 3300 |
| }, |
| { |
| "epoch": 0.856439492096398, |
| "grad_norm": 0.37919701821847407, |
| "learning_rate": 1.2296221696014732e-05, |
| "loss": 1.1837, |
| "step": 3305 |
| }, |
| { |
| "epoch": 0.8577351645504017, |
| "grad_norm": 0.36743928128733283, |
| "learning_rate": 1.2079830638129164e-05, |
| "loss": 1.1148, |
| "step": 3310 |
| }, |
| { |
| "epoch": 0.8590308370044053, |
| "grad_norm": 0.3738619978563127, |
| "learning_rate": 1.1865238111389588e-05, |
| "loss": 1.1604, |
| "step": 3315 |
| }, |
| { |
| "epoch": 0.8603265094584089, |
| "grad_norm": 0.3541904420302459, |
| "learning_rate": 1.1652448505588998e-05, |
| "loss": 1.1544, |
| "step": 3320 |
| }, |
| { |
| "epoch": 0.8616221819124126, |
| "grad_norm": 0.36684430135825935, |
| "learning_rate": 1.1441466173639092e-05, |
| "loss": 1.156, |
| "step": 3325 |
| }, |
| { |
| "epoch": 0.8629178543664162, |
| "grad_norm": 0.37365304474041194, |
| "learning_rate": 1.1232295431481222e-05, |
| "loss": 1.1084, |
| "step": 3330 |
| }, |
| { |
| "epoch": 0.8642135268204199, |
| "grad_norm": 0.3810375689519541, |
| "learning_rate": 1.1024940557998143e-05, |
| "loss": 1.1386, |
| "step": 3335 |
| }, |
| { |
| "epoch": 0.8655091992744234, |
| "grad_norm": 0.3673719354100514, |
| "learning_rate": 1.0819405794926418e-05, |
| "loss": 1.1629, |
| "step": 3340 |
| }, |
| { |
| "epoch": 0.866804871728427, |
| "grad_norm": 0.37342554741406375, |
| "learning_rate": 1.0615695346769693e-05, |
| "loss": 1.1134, |
| "step": 3345 |
| }, |
| { |
| "epoch": 0.8681005441824307, |
| "grad_norm": 0.3635218597663001, |
| "learning_rate": 1.0413813380712701e-05, |
| "loss": 1.1147, |
| "step": 3350 |
| }, |
| { |
| "epoch": 0.8693962166364343, |
| "grad_norm": 0.3570736451827848, |
| "learning_rate": 1.0213764026535921e-05, |
| "loss": 1.1321, |
| "step": 3355 |
| }, |
| { |
| "epoch": 0.8706918890904379, |
| "grad_norm": 0.3891720605828237, |
| "learning_rate": 1.0015551376531296e-05, |
| "loss": 1.1001, |
| "step": 3360 |
| }, |
| { |
| "epoch": 0.8719875615444416, |
| "grad_norm": 0.37097051363645606, |
| "learning_rate": 9.819179485418273e-06, |
| "loss": 1.1241, |
| "step": 3365 |
| }, |
| { |
| "epoch": 0.8732832339984452, |
| "grad_norm": 0.3917978913676709, |
| "learning_rate": 9.624652370261034e-06, |
| "loss": 1.1304, |
| "step": 3370 |
| }, |
| { |
| "epoch": 0.8745789064524488, |
| "grad_norm": 0.37226262222810586, |
| "learning_rate": 9.431974010386258e-06, |
| "loss": 1.1421, |
| "step": 3375 |
| }, |
| { |
| "epoch": 0.8758745789064525, |
| "grad_norm": 0.3751233666218166, |
| "learning_rate": 9.24114834730171e-06, |
| "loss": 1.1022, |
| "step": 3380 |
| }, |
| { |
| "epoch": 0.8771702513604561, |
| "grad_norm": 0.36555054148949856, |
| "learning_rate": 9.052179284615658e-06, |
| "loss": 1.13, |
| "step": 3385 |
| }, |
| { |
| "epoch": 0.8784659238144598, |
| "grad_norm": 0.38148359927958986, |
| "learning_rate": 8.865070687956977e-06, |
| "loss": 1.1894, |
| "step": 3390 |
| }, |
| { |
| "epoch": 0.8797615962684633, |
| "grad_norm": 0.416874595298691, |
| "learning_rate": 8.679826384896061e-06, |
| "loss": 1.1736, |
| "step": 3395 |
| }, |
| { |
| "epoch": 0.8810572687224669, |
| "grad_norm": 0.387504501992165, |
| "learning_rate": 8.496450164866565e-06, |
| "loss": 1.1511, |
| "step": 3400 |
| }, |
| { |
| "epoch": 0.8823529411764706, |
| "grad_norm": 0.34801518476142257, |
| "learning_rate": 8.314945779087858e-06, |
| "loss": 1.1224, |
| "step": 3405 |
| }, |
| { |
| "epoch": 0.8836486136304742, |
| "grad_norm": 0.3758149676256447, |
| "learning_rate": 8.135316940488347e-06, |
| "loss": 1.1448, |
| "step": 3410 |
| }, |
| { |
| "epoch": 0.8849442860844778, |
| "grad_norm": 0.36726096701274225, |
| "learning_rate": 7.9575673236294e-06, |
| "loss": 1.1719, |
| "step": 3415 |
| }, |
| { |
| "epoch": 0.8862399585384815, |
| "grad_norm": 0.3578469295777726, |
| "learning_rate": 7.781700564630311e-06, |
| "loss": 1.1385, |
| "step": 3420 |
| }, |
| { |
| "epoch": 0.8875356309924851, |
| "grad_norm": 0.38263190643853745, |
| "learning_rate": 7.6077202610938205e-06, |
| "loss": 1.177, |
| "step": 3425 |
| }, |
| { |
| "epoch": 0.8888313034464888, |
| "grad_norm": 0.38665333786102024, |
| "learning_rate": 7.435629972032565e-06, |
| "loss": 1.1358, |
| "step": 3430 |
| }, |
| { |
| "epoch": 0.8901269759004924, |
| "grad_norm": 0.36097779607915553, |
| "learning_rate": 7.265433217796302e-06, |
| "loss": 1.1055, |
| "step": 3435 |
| }, |
| { |
| "epoch": 0.891422648354496, |
| "grad_norm": 0.4052378204511826, |
| "learning_rate": 7.0971334799998404e-06, |
| "loss": 1.0912, |
| "step": 3440 |
| }, |
| { |
| "epoch": 0.8927183208084997, |
| "grad_norm": 0.3704412054341248, |
| "learning_rate": 6.930734201451816e-06, |
| "loss": 1.1591, |
| "step": 3445 |
| }, |
| { |
| "epoch": 0.8940139932625032, |
| "grad_norm": 0.37555016415939446, |
| "learning_rate": 6.7662387860843225e-06, |
| "loss": 1.1149, |
| "step": 3450 |
| }, |
| { |
| "epoch": 0.8953096657165068, |
| "grad_norm": 0.3798653467067839, |
| "learning_rate": 6.603650598883226e-06, |
| "loss": 1.0958, |
| "step": 3455 |
| }, |
| { |
| "epoch": 0.8966053381705105, |
| "grad_norm": 0.34342261742501184, |
| "learning_rate": 6.4429729658193714e-06, |
| "loss": 1.0718, |
| "step": 3460 |
| }, |
| { |
| "epoch": 0.8979010106245141, |
| "grad_norm": 0.38053269516746213, |
| "learning_rate": 6.284209173780498e-06, |
| "loss": 1.1133, |
| "step": 3465 |
| }, |
| { |
| "epoch": 0.8991966830785177, |
| "grad_norm": 0.38564284280238426, |
| "learning_rate": 6.127362470504027e-06, |
| "loss": 1.1239, |
| "step": 3470 |
| }, |
| { |
| "epoch": 0.9004923555325214, |
| "grad_norm": 0.3624213790138937, |
| "learning_rate": 5.972436064510667e-06, |
| "loss": 1.1766, |
| "step": 3475 |
| }, |
| { |
| "epoch": 0.901788027986525, |
| "grad_norm": 0.36614157981464224, |
| "learning_rate": 5.819433125038643e-06, |
| "loss": 1.146, |
| "step": 3480 |
| }, |
| { |
| "epoch": 0.9030837004405287, |
| "grad_norm": 0.39479288263797235, |
| "learning_rate": 5.668356781979056e-06, |
| "loss": 1.1669, |
| "step": 3485 |
| }, |
| { |
| "epoch": 0.9043793728945323, |
| "grad_norm": 0.35385937176801885, |
| "learning_rate": 5.519210125811713e-06, |
| "loss": 1.1224, |
| "step": 3490 |
| }, |
| { |
| "epoch": 0.9056750453485359, |
| "grad_norm": 0.3778900407632985, |
| "learning_rate": 5.371996207541941e-06, |
| "loss": 1.1287, |
| "step": 3495 |
| }, |
| { |
| "epoch": 0.9069707178025396, |
| "grad_norm": 0.35939889684365445, |
| "learning_rate": 5.226718038638234e-06, |
| "loss": 1.1274, |
| "step": 3500 |
| }, |
| { |
| "epoch": 0.9082663902565431, |
| "grad_norm": 0.38058001351217496, |
| "learning_rate": 5.083378590970511e-06, |
| "loss": 1.154, |
| "step": 3505 |
| }, |
| { |
| "epoch": 0.9095620627105467, |
| "grad_norm": 0.36339105009188977, |
| "learning_rate": 4.941980796749524e-06, |
| "loss": 1.1234, |
| "step": 3510 |
| }, |
| { |
| "epoch": 0.9108577351645504, |
| "grad_norm": 0.3753439481173232, |
| "learning_rate": 4.802527548466684e-06, |
| "loss": 1.0982, |
| "step": 3515 |
| }, |
| { |
| "epoch": 0.912153407618554, |
| "grad_norm": 0.3547606097647797, |
| "learning_rate": 4.665021698834981e-06, |
| "loss": 1.1405, |
| "step": 3520 |
| }, |
| { |
| "epoch": 0.9134490800725577, |
| "grad_norm": 0.3804573132062168, |
| "learning_rate": 4.529466060730681e-06, |
| "loss": 1.1529, |
| "step": 3525 |
| }, |
| { |
| "epoch": 0.9147447525265613, |
| "grad_norm": 0.3766860837932323, |
| "learning_rate": 4.395863407135614e-06, |
| "loss": 1.0821, |
| "step": 3530 |
| }, |
| { |
| "epoch": 0.9160404249805649, |
| "grad_norm": 0.3760175631598513, |
| "learning_rate": 4.264216471080651e-06, |
| "loss": 1.1451, |
| "step": 3535 |
| }, |
| { |
| "epoch": 0.9173360974345686, |
| "grad_norm": 0.3901693827353405, |
| "learning_rate": 4.134527945589639e-06, |
| "loss": 1.1148, |
| "step": 3540 |
| }, |
| { |
| "epoch": 0.9186317698885722, |
| "grad_norm": 0.4488796382773471, |
| "learning_rate": 4.006800483624396e-06, |
| "loss": 1.1614, |
| "step": 3545 |
| }, |
| { |
| "epoch": 0.9199274423425758, |
| "grad_norm": 0.40897274265302724, |
| "learning_rate": 3.881036698030449e-06, |
| "loss": 1.1744, |
| "step": 3550 |
| }, |
| { |
| "epoch": 0.9212231147965795, |
| "grad_norm": 0.3690887838305789, |
| "learning_rate": 3.7572391614834833e-06, |
| "loss": 1.136, |
| "step": 3555 |
| }, |
| { |
| "epoch": 0.922518787250583, |
| "grad_norm": 0.3688927093580592, |
| "learning_rate": 3.6354104064368566e-06, |
| "loss": 1.0947, |
| "step": 3560 |
| }, |
| { |
| "epoch": 0.9238144597045866, |
| "grad_norm": 0.3826643203341735, |
| "learning_rate": 3.515552925069676e-06, |
| "loss": 1.1757, |
| "step": 3565 |
| }, |
| { |
| "epoch": 0.9251101321585903, |
| "grad_norm": 0.3684225941093978, |
| "learning_rate": 3.397669169235862e-06, |
| "loss": 1.1574, |
| "step": 3570 |
| }, |
| { |
| "epoch": 0.9264058046125939, |
| "grad_norm": 0.3447005114527934, |
| "learning_rate": 3.281761550414042e-06, |
| "loss": 1.0847, |
| "step": 3575 |
| }, |
| { |
| "epoch": 0.9277014770665976, |
| "grad_norm": 0.38643242398136646, |
| "learning_rate": 3.1678324396581137e-06, |
| "loss": 1.2221, |
| "step": 3580 |
| }, |
| { |
| "epoch": 0.9289971495206012, |
| "grad_norm": 0.37779005091381196, |
| "learning_rate": 3.0558841675488393e-06, |
| "loss": 1.1612, |
| "step": 3585 |
| }, |
| { |
| "epoch": 0.9302928219746048, |
| "grad_norm": 0.38097475759248384, |
| "learning_rate": 2.9459190241461043e-06, |
| "loss": 1.1267, |
| "step": 3590 |
| }, |
| { |
| "epoch": 0.9315884944286085, |
| "grad_norm": 0.37618918488273334, |
| "learning_rate": 2.8379392589421237e-06, |
| "loss": 1.1563, |
| "step": 3595 |
| }, |
| { |
| "epoch": 0.9328841668826121, |
| "grad_norm": 0.3621451372937871, |
| "learning_rate": 2.7319470808153892e-06, |
| "loss": 1.1197, |
| "step": 3600 |
| }, |
| { |
| "epoch": 0.9341798393366157, |
| "grad_norm": 0.35758689530964316, |
| "learning_rate": 2.6279446579854707e-06, |
| "loss": 1.1446, |
| "step": 3605 |
| }, |
| { |
| "epoch": 0.9354755117906194, |
| "grad_norm": 0.3803519145314357, |
| "learning_rate": 2.525934117968731e-06, |
| "loss": 1.1823, |
| "step": 3610 |
| }, |
| { |
| "epoch": 0.9367711842446229, |
| "grad_norm": 0.36540749180301496, |
| "learning_rate": 2.4259175475347172e-06, |
| "loss": 1.1748, |
| "step": 3615 |
| }, |
| { |
| "epoch": 0.9380668566986265, |
| "grad_norm": 0.34819215016667493, |
| "learning_rate": 2.3278969926635252e-06, |
| "loss": 1.1036, |
| "step": 3620 |
| }, |
| { |
| "epoch": 0.9393625291526302, |
| "grad_norm": 0.3586884150884424, |
| "learning_rate": 2.2318744585039796e-06, |
| "loss": 1.1192, |
| "step": 3625 |
| }, |
| { |
| "epoch": 0.9406582016066338, |
| "grad_norm": 0.37877938188955895, |
| "learning_rate": 2.1378519093324776e-06, |
| "loss": 1.128, |
| "step": 3630 |
| }, |
| { |
| "epoch": 0.9419538740606375, |
| "grad_norm": 0.3781105512240613, |
| "learning_rate": 2.0458312685129876e-06, |
| "loss": 1.1612, |
| "step": 3635 |
| }, |
| { |
| "epoch": 0.9432495465146411, |
| "grad_norm": 0.37487413050848234, |
| "learning_rate": 1.955814418457591e-06, |
| "loss": 1.1396, |
| "step": 3640 |
| }, |
| { |
| "epoch": 0.9445452189686447, |
| "grad_norm": 0.3840114910436269, |
| "learning_rate": 1.867803200587981e-06, |
| "loss": 1.1129, |
| "step": 3645 |
| }, |
| { |
| "epoch": 0.9458408914226484, |
| "grad_norm": 0.3719048343830175, |
| "learning_rate": 1.7817994152978468e-06, |
| "loss": 1.1157, |
| "step": 3650 |
| }, |
| { |
| "epoch": 0.947136563876652, |
| "grad_norm": 0.3772856287062892, |
| "learning_rate": 1.6978048219159714e-06, |
| "loss": 1.1194, |
| "step": 3655 |
| }, |
| { |
| "epoch": 0.9484322363306557, |
| "grad_norm": 0.3598408324385285, |
| "learning_rate": 1.6158211386703259e-06, |
| "loss": 1.1284, |
| "step": 3660 |
| }, |
| { |
| "epoch": 0.9497279087846593, |
| "grad_norm": 0.365397297934618, |
| "learning_rate": 1.5358500426528422e-06, |
| "loss": 1.1282, |
| "step": 3665 |
| }, |
| { |
| "epoch": 0.9510235812386628, |
| "grad_norm": 0.3659664905751428, |
| "learning_rate": 1.4578931697851406e-06, |
| "loss": 1.1051, |
| "step": 3670 |
| }, |
| { |
| "epoch": 0.9523192536926665, |
| "grad_norm": 0.3549703867179501, |
| "learning_rate": 1.3819521147851123e-06, |
| "loss": 1.1292, |
| "step": 3675 |
| }, |
| { |
| "epoch": 0.9536149261466701, |
| "grad_norm": 0.36856057544037374, |
| "learning_rate": 1.3080284311341674e-06, |
| "loss": 1.141, |
| "step": 3680 |
| }, |
| { |
| "epoch": 0.9549105986006737, |
| "grad_norm": 0.3869169065095209, |
| "learning_rate": 1.236123631045627e-06, |
| "loss": 1.1622, |
| "step": 3685 |
| }, |
| { |
| "epoch": 0.9562062710546774, |
| "grad_norm": 0.35481859423446044, |
| "learning_rate": 1.1662391854336263e-06, |
| "loss": 1.1223, |
| "step": 3690 |
| }, |
| { |
| "epoch": 0.957501943508681, |
| "grad_norm": 0.38700062266195656, |
| "learning_rate": 1.0983765238831377e-06, |
| "loss": 1.1501, |
| "step": 3695 |
| }, |
| { |
| "epoch": 0.9587976159626846, |
| "grad_norm": 0.3672922868247709, |
| "learning_rate": 1.032537034620684e-06, |
| "loss": 1.1419, |
| "step": 3700 |
| }, |
| { |
| "epoch": 0.9600932884166883, |
| "grad_norm": 0.37605980694280633, |
| "learning_rate": 9.687220644859275e-07, |
| "loss": 1.133, |
| "step": 3705 |
| }, |
| { |
| "epoch": 0.9613889608706919, |
| "grad_norm": 0.36877920673269027, |
| "learning_rate": 9.069329189041464e-07, |
| "loss": 1.1608, |
| "step": 3710 |
| }, |
| { |
| "epoch": 0.9626846333246956, |
| "grad_norm": 0.381237649400668, |
| "learning_rate": 8.471708618595142e-07, |
| "loss": 1.132, |
| "step": 3715 |
| }, |
| { |
| "epoch": 0.9639803057786992, |
| "grad_norm": 1.8685424500994194, |
| "learning_rate": 7.894371158692627e-07, |
| "loss": 1.1656, |
| "step": 3720 |
| }, |
| { |
| "epoch": 0.9652759782327027, |
| "grad_norm": 0.38287612615523475, |
| "learning_rate": 7.337328619586359e-07, |
| "loss": 1.1992, |
| "step": 3725 |
| }, |
| { |
| "epoch": 0.9665716506867064, |
| "grad_norm": 0.3838324502538139, |
| "learning_rate": 6.800592396367545e-07, |
| "loss": 1.1816, |
| "step": 3730 |
| }, |
| { |
| "epoch": 0.96786732314071, |
| "grad_norm": 0.36218594513869495, |
| "learning_rate": 6.284173468733334e-07, |
| "loss": 1.1188, |
| "step": 3735 |
| }, |
| { |
| "epoch": 0.9691629955947136, |
| "grad_norm": 0.3605898736094485, |
| "learning_rate": 5.788082400761563e-07, |
| "loss": 1.1342, |
| "step": 3740 |
| }, |
| { |
| "epoch": 0.9704586680487173, |
| "grad_norm": 0.3819284101383782, |
| "learning_rate": 5.312329340695143e-07, |
| "loss": 1.1341, |
| "step": 3745 |
| }, |
| { |
| "epoch": 0.9717543405027209, |
| "grad_norm": 0.371840798865179, |
| "learning_rate": 4.856924020734565e-07, |
| "loss": 1.1315, |
| "step": 3750 |
| }, |
| { |
| "epoch": 0.9730500129567246, |
| "grad_norm": 0.37381875278150484, |
| "learning_rate": 4.4218757568380563e-07, |
| "loss": 1.1129, |
| "step": 3755 |
| }, |
| { |
| "epoch": 0.9743456854107282, |
| "grad_norm": 0.37126771253731844, |
| "learning_rate": 4.007193448532065e-07, |
| "loss": 1.1666, |
| "step": 3760 |
| }, |
| { |
| "epoch": 0.9756413578647318, |
| "grad_norm": 0.36988126308198593, |
| "learning_rate": 3.612885578728298e-07, |
| "loss": 1.1429, |
| "step": 3765 |
| }, |
| { |
| "epoch": 0.9769370303187355, |
| "grad_norm": 0.3778202136491613, |
| "learning_rate": 3.2389602135507457e-07, |
| "loss": 1.1985, |
| "step": 3770 |
| }, |
| { |
| "epoch": 0.9782327027727391, |
| "grad_norm": 0.3544356950249304, |
| "learning_rate": 2.885425002170594e-07, |
| "loss": 1.1469, |
| "step": 3775 |
| }, |
| { |
| "epoch": 0.9795283752267426, |
| "grad_norm": 0.37874103417633515, |
| "learning_rate": 2.5522871766494595e-07, |
| "loss": 1.1101, |
| "step": 3780 |
| }, |
| { |
| "epoch": 0.9808240476807463, |
| "grad_norm": 0.3788020452535586, |
| "learning_rate": 2.2395535517917287e-07, |
| "loss": 1.1895, |
| "step": 3785 |
| }, |
| { |
| "epoch": 0.9821197201347499, |
| "grad_norm": 0.37493085740769366, |
| "learning_rate": 1.947230525005006e-07, |
| "loss": 1.1213, |
| "step": 3790 |
| }, |
| { |
| "epoch": 0.9834153925887535, |
| "grad_norm": 0.36932806609876156, |
| "learning_rate": 1.6753240761693268e-07, |
| "loss": 1.1141, |
| "step": 3795 |
| }, |
| { |
| "epoch": 0.9847110650427572, |
| "grad_norm": 0.3675515005536323, |
| "learning_rate": 1.4238397675150339e-07, |
| "loss": 1.1292, |
| "step": 3800 |
| }, |
| { |
| "epoch": 0.9860067374967608, |
| "grad_norm": 0.3595954681947242, |
| "learning_rate": 1.1927827435084248e-07, |
| "loss": 1.1044, |
| "step": 3805 |
| }, |
| { |
| "epoch": 0.9873024099507645, |
| "grad_norm": 0.3750795386829234, |
| "learning_rate": 9.821577307470575e-08, |
| "loss": 1.1441, |
| "step": 3810 |
| }, |
| { |
| "epoch": 0.9885980824047681, |
| "grad_norm": 0.36749997185699007, |
| "learning_rate": 7.919690378629385e-08, |
| "loss": 1.171, |
| "step": 3815 |
| }, |
| { |
| "epoch": 0.9898937548587717, |
| "grad_norm": 0.3622505687301632, |
| "learning_rate": 6.222205554339277e-08, |
| "loss": 1.1675, |
| "step": 3820 |
| }, |
| { |
| "epoch": 0.9911894273127754, |
| "grad_norm": 0.3589951765733612, |
| "learning_rate": 4.729157559049124e-08, |
| "loss": 1.1451, |
| "step": 3825 |
| }, |
| { |
| "epoch": 0.992485099766779, |
| "grad_norm": 0.372014869272016, |
| "learning_rate": 3.440576935164197e-08, |
| "loss": 1.1383, |
| "step": 3830 |
| }, |
| { |
| "epoch": 0.9937807722207825, |
| "grad_norm": 0.36261244538328924, |
| "learning_rate": 2.3564900424188906e-08, |
| "loss": 1.1706, |
| "step": 3835 |
| }, |
| { |
| "epoch": 0.9950764446747862, |
| "grad_norm": 0.37064889363490416, |
| "learning_rate": 1.4769190573393765e-08, |
| "loss": 1.1344, |
| "step": 3840 |
| }, |
| { |
| "epoch": 0.9963721171287898, |
| "grad_norm": 0.4019334217280328, |
| "learning_rate": 8.018819727928505e-09, |
| "loss": 1.19, |
| "step": 3845 |
| }, |
| { |
| "epoch": 0.9976677895827935, |
| "grad_norm": 0.3532944069166521, |
| "learning_rate": 3.313925976156096e-09, |
| "loss": 1.1704, |
| "step": 3850 |
| }, |
| { |
| "epoch": 0.9989634620367971, |
| "grad_norm": 0.3954932358396543, |
| "learning_rate": 6.546055633105397e-10, |
| "loss": 1.0999, |
| "step": 3855 |
| }, |
| { |
| "epoch": 1.0, |
| "eval_runtime": 3.6908, |
| "eval_samples_per_second": 2.709, |
| "eval_steps_per_second": 0.813, |
| "step": 3859 |
| }, |
| { |
| "epoch": 1.0, |
| "step": 3859, |
| "total_flos": 1.3011067827388416e+16, |
| "train_loss": 1.2484874595598594, |
| "train_runtime": 22644.1998, |
| "train_samples_per_second": 2.726, |
| "train_steps_per_second": 0.17 |
| } |
| ], |
| "logging_steps": 5, |
| "max_steps": 3859, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 1, |
| "save_steps": 100, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 1.3011067827388416e+16, |
| "train_batch_size": 4, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|