| { |
| "best_global_step": null, |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 1.0, |
| "eval_steps": 500, |
| "global_step": 4954, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.002018927444794953, |
| "grad_norm": 2.3389563305266345, |
| "learning_rate": 1.8145161290322583e-07, |
| "loss": 0.7881, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.004037854889589906, |
| "grad_norm": 2.024167058240825, |
| "learning_rate": 3.830645161290323e-07, |
| "loss": 0.8121, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.006056782334384858, |
| "grad_norm": 1.4351269819866417, |
| "learning_rate": 5.846774193548388e-07, |
| "loss": 0.7797, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.008075709779179811, |
| "grad_norm": 0.9310252821124614, |
| "learning_rate": 7.862903225806453e-07, |
| "loss": 0.7318, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.010094637223974764, |
| "grad_norm": 0.9557517206875442, |
| "learning_rate": 9.879032258064516e-07, |
| "loss": 0.7018, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.012113564668769715, |
| "grad_norm": 0.5829217778568054, |
| "learning_rate": 1.1895161290322582e-06, |
| "loss": 0.6708, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.014132492113564668, |
| "grad_norm": 0.5112881851342524, |
| "learning_rate": 1.3911290322580644e-06, |
| "loss": 0.6417, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.016151419558359623, |
| "grad_norm": 0.5140187941233466, |
| "learning_rate": 1.5927419354838712e-06, |
| "loss": 0.6227, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.018170347003154574, |
| "grad_norm": 0.4872446144280347, |
| "learning_rate": 1.7943548387096777e-06, |
| "loss": 0.615, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.02018927444794953, |
| "grad_norm": 0.5191969395080858, |
| "learning_rate": 1.9959677419354837e-06, |
| "loss": 0.6105, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.02220820189274448, |
| "grad_norm": 0.451802382492337, |
| "learning_rate": 2.1975806451612907e-06, |
| "loss": 0.5911, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.02422712933753943, |
| "grad_norm": 0.42163928465445194, |
| "learning_rate": 2.399193548387097e-06, |
| "loss": 0.5921, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.026246056782334385, |
| "grad_norm": 0.4552027864136731, |
| "learning_rate": 2.6008064516129032e-06, |
| "loss": 0.5818, |
| "step": 130 |
| }, |
| { |
| "epoch": 0.028264984227129336, |
| "grad_norm": 0.44985963500335385, |
| "learning_rate": 2.8024193548387097e-06, |
| "loss": 0.5876, |
| "step": 140 |
| }, |
| { |
| "epoch": 0.03028391167192429, |
| "grad_norm": 0.5111630527751778, |
| "learning_rate": 3.0040322580645167e-06, |
| "loss": 0.5728, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.032302839116719245, |
| "grad_norm": 0.44240607593167414, |
| "learning_rate": 3.2056451612903228e-06, |
| "loss": 0.5643, |
| "step": 160 |
| }, |
| { |
| "epoch": 0.03432176656151419, |
| "grad_norm": 0.4865981695157576, |
| "learning_rate": 3.4072580645161293e-06, |
| "loss": 0.5617, |
| "step": 170 |
| }, |
| { |
| "epoch": 0.03634069400630915, |
| "grad_norm": 0.6085288533687223, |
| "learning_rate": 3.6088709677419358e-06, |
| "loss": 0.5501, |
| "step": 180 |
| }, |
| { |
| "epoch": 0.0383596214511041, |
| "grad_norm": 0.47289717909858964, |
| "learning_rate": 3.8104838709677423e-06, |
| "loss": 0.559, |
| "step": 190 |
| }, |
| { |
| "epoch": 0.04037854889589906, |
| "grad_norm": 0.5389683685989423, |
| "learning_rate": 4.012096774193548e-06, |
| "loss": 0.5601, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.042397476340694004, |
| "grad_norm": 0.44905068010794785, |
| "learning_rate": 4.213709677419355e-06, |
| "loss": 0.546, |
| "step": 210 |
| }, |
| { |
| "epoch": 0.04441640378548896, |
| "grad_norm": 0.4738265544351665, |
| "learning_rate": 4.415322580645161e-06, |
| "loss": 0.5527, |
| "step": 220 |
| }, |
| { |
| "epoch": 0.04643533123028391, |
| "grad_norm": 0.48116551901846, |
| "learning_rate": 4.616935483870968e-06, |
| "loss": 0.5428, |
| "step": 230 |
| }, |
| { |
| "epoch": 0.04845425867507886, |
| "grad_norm": 0.4514773717137442, |
| "learning_rate": 4.818548387096774e-06, |
| "loss": 0.5338, |
| "step": 240 |
| }, |
| { |
| "epoch": 0.050473186119873815, |
| "grad_norm": 0.5092040454559039, |
| "learning_rate": 5.020161290322581e-06, |
| "loss": 0.5546, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.05249211356466877, |
| "grad_norm": 0.45858657596184155, |
| "learning_rate": 5.221774193548388e-06, |
| "loss": 0.5408, |
| "step": 260 |
| }, |
| { |
| "epoch": 0.054511041009463725, |
| "grad_norm": 0.47766211942842146, |
| "learning_rate": 5.423387096774194e-06, |
| "loss": 0.5427, |
| "step": 270 |
| }, |
| { |
| "epoch": 0.05652996845425867, |
| "grad_norm": 0.5108229913280311, |
| "learning_rate": 5.625e-06, |
| "loss": 0.5357, |
| "step": 280 |
| }, |
| { |
| "epoch": 0.05854889589905363, |
| "grad_norm": 0.5232937676969737, |
| "learning_rate": 5.8266129032258064e-06, |
| "loss": 0.5316, |
| "step": 290 |
| }, |
| { |
| "epoch": 0.06056782334384858, |
| "grad_norm": 0.5241114539289579, |
| "learning_rate": 6.028225806451613e-06, |
| "loss": 0.5444, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.06258675078864354, |
| "grad_norm": 0.9566252867340465, |
| "learning_rate": 6.2298387096774194e-06, |
| "loss": 0.5256, |
| "step": 310 |
| }, |
| { |
| "epoch": 0.06460567823343849, |
| "grad_norm": 0.4700069264487592, |
| "learning_rate": 6.431451612903226e-06, |
| "loss": 0.535, |
| "step": 320 |
| }, |
| { |
| "epoch": 0.06662460567823345, |
| "grad_norm": 0.46343674018411934, |
| "learning_rate": 6.633064516129033e-06, |
| "loss": 0.5281, |
| "step": 330 |
| }, |
| { |
| "epoch": 0.06864353312302839, |
| "grad_norm": 0.46610206235853135, |
| "learning_rate": 6.834677419354839e-06, |
| "loss": 0.5324, |
| "step": 340 |
| }, |
| { |
| "epoch": 0.07066246056782334, |
| "grad_norm": 0.5217091961421014, |
| "learning_rate": 7.0362903225806454e-06, |
| "loss": 0.5321, |
| "step": 350 |
| }, |
| { |
| "epoch": 0.0726813880126183, |
| "grad_norm": 0.4532872443875366, |
| "learning_rate": 7.2379032258064515e-06, |
| "loss": 0.5235, |
| "step": 360 |
| }, |
| { |
| "epoch": 0.07470031545741325, |
| "grad_norm": 0.5481191485070994, |
| "learning_rate": 7.4395161290322585e-06, |
| "loss": 0.5203, |
| "step": 370 |
| }, |
| { |
| "epoch": 0.0767192429022082, |
| "grad_norm": 0.4817884817896886, |
| "learning_rate": 7.641129032258065e-06, |
| "loss": 0.5289, |
| "step": 380 |
| }, |
| { |
| "epoch": 0.07873817034700316, |
| "grad_norm": 0.463062633118143, |
| "learning_rate": 7.842741935483872e-06, |
| "loss": 0.5199, |
| "step": 390 |
| }, |
| { |
| "epoch": 0.08075709779179811, |
| "grad_norm": 0.5267129255916911, |
| "learning_rate": 8.044354838709678e-06, |
| "loss": 0.5239, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.08277602523659305, |
| "grad_norm": 0.4703183395271576, |
| "learning_rate": 8.245967741935484e-06, |
| "loss": 0.5217, |
| "step": 410 |
| }, |
| { |
| "epoch": 0.08479495268138801, |
| "grad_norm": 0.4559407675951848, |
| "learning_rate": 8.447580645161291e-06, |
| "loss": 0.5187, |
| "step": 420 |
| }, |
| { |
| "epoch": 0.08681388012618296, |
| "grad_norm": 0.4638358181971901, |
| "learning_rate": 8.649193548387097e-06, |
| "loss": 0.5194, |
| "step": 430 |
| }, |
| { |
| "epoch": 0.08883280757097792, |
| "grad_norm": 0.47768712570505284, |
| "learning_rate": 8.850806451612905e-06, |
| "loss": 0.513, |
| "step": 440 |
| }, |
| { |
| "epoch": 0.09085173501577287, |
| "grad_norm": 0.5140723294282747, |
| "learning_rate": 9.05241935483871e-06, |
| "loss": 0.5161, |
| "step": 450 |
| }, |
| { |
| "epoch": 0.09287066246056783, |
| "grad_norm": 0.4719590441343577, |
| "learning_rate": 9.254032258064517e-06, |
| "loss": 0.5033, |
| "step": 460 |
| }, |
| { |
| "epoch": 0.09488958990536278, |
| "grad_norm": 0.5031363320062582, |
| "learning_rate": 9.455645161290323e-06, |
| "loss": 0.5145, |
| "step": 470 |
| }, |
| { |
| "epoch": 0.09690851735015772, |
| "grad_norm": 0.5027908604602157, |
| "learning_rate": 9.65725806451613e-06, |
| "loss": 0.5178, |
| "step": 480 |
| }, |
| { |
| "epoch": 0.09892744479495268, |
| "grad_norm": 0.5046628095943158, |
| "learning_rate": 9.858870967741936e-06, |
| "loss": 0.5069, |
| "step": 490 |
| }, |
| { |
| "epoch": 0.10094637223974763, |
| "grad_norm": 0.5522780839548007, |
| "learning_rate": 9.99998882617222e-06, |
| "loss": 0.5169, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.10296529968454259, |
| "grad_norm": 0.48260445699138166, |
| "learning_rate": 9.999790181734308e-06, |
| "loss": 0.5067, |
| "step": 510 |
| }, |
| { |
| "epoch": 0.10498422712933754, |
| "grad_norm": 0.4875096921972213, |
| "learning_rate": 9.999343241367322e-06, |
| "loss": 0.5152, |
| "step": 520 |
| }, |
| { |
| "epoch": 0.1070031545741325, |
| "grad_norm": 0.5080882476390746, |
| "learning_rate": 9.998648027266884e-06, |
| "loss": 0.5131, |
| "step": 530 |
| }, |
| { |
| "epoch": 0.10902208201892745, |
| "grad_norm": 0.5107719615388011, |
| "learning_rate": 9.997704573958209e-06, |
| "loss": 0.5053, |
| "step": 540 |
| }, |
| { |
| "epoch": 0.1110410094637224, |
| "grad_norm": 0.5361602322829081, |
| "learning_rate": 9.996512928294389e-06, |
| "loss": 0.5083, |
| "step": 550 |
| }, |
| { |
| "epoch": 0.11305993690851734, |
| "grad_norm": 0.4851941991561818, |
| "learning_rate": 9.995073149454057e-06, |
| "loss": 0.4997, |
| "step": 560 |
| }, |
| { |
| "epoch": 0.1150788643533123, |
| "grad_norm": 0.49258962377904003, |
| "learning_rate": 9.99338530893846e-06, |
| "loss": 0.516, |
| "step": 570 |
| }, |
| { |
| "epoch": 0.11709779179810725, |
| "grad_norm": 0.4835880405802275, |
| "learning_rate": 9.991449490567901e-06, |
| "loss": 0.5073, |
| "step": 580 |
| }, |
| { |
| "epoch": 0.11911671924290221, |
| "grad_norm": 0.507245673357914, |
| "learning_rate": 9.989265790477581e-06, |
| "loss": 0.5014, |
| "step": 590 |
| }, |
| { |
| "epoch": 0.12113564668769716, |
| "grad_norm": 0.5146575464998254, |
| "learning_rate": 9.986834317112817e-06, |
| "loss": 0.5177, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.12315457413249212, |
| "grad_norm": 0.4662328996304102, |
| "learning_rate": 9.984155191223663e-06, |
| "loss": 0.5054, |
| "step": 610 |
| }, |
| { |
| "epoch": 0.12517350157728707, |
| "grad_norm": 0.4571893765964666, |
| "learning_rate": 9.981228545858913e-06, |
| "loss": 0.5086, |
| "step": 620 |
| }, |
| { |
| "epoch": 0.12719242902208203, |
| "grad_norm": 0.4697827268486345, |
| "learning_rate": 9.978054526359493e-06, |
| "loss": 0.5029, |
| "step": 630 |
| }, |
| { |
| "epoch": 0.12921135646687698, |
| "grad_norm": 0.5031062337107917, |
| "learning_rate": 9.97463329035124e-06, |
| "loss": 0.5021, |
| "step": 640 |
| }, |
| { |
| "epoch": 0.13123028391167194, |
| "grad_norm": 0.571029993720401, |
| "learning_rate": 9.970965007737081e-06, |
| "loss": 0.5077, |
| "step": 650 |
| }, |
| { |
| "epoch": 0.1332492113564669, |
| "grad_norm": 0.45753840087765046, |
| "learning_rate": 9.967049860688587e-06, |
| "loss": 0.4961, |
| "step": 660 |
| }, |
| { |
| "epoch": 0.13526813880126182, |
| "grad_norm": 0.5189371422618692, |
| "learning_rate": 9.962888043636931e-06, |
| "loss": 0.5036, |
| "step": 670 |
| }, |
| { |
| "epoch": 0.13728706624605677, |
| "grad_norm": 0.47066047930695143, |
| "learning_rate": 9.958479763263234e-06, |
| "loss": 0.4972, |
| "step": 680 |
| }, |
| { |
| "epoch": 0.13930599369085173, |
| "grad_norm": 0.49380951082560154, |
| "learning_rate": 9.953825238488296e-06, |
| "loss": 0.4955, |
| "step": 690 |
| }, |
| { |
| "epoch": 0.14132492113564668, |
| "grad_norm": 0.49487805704511495, |
| "learning_rate": 9.948924700461727e-06, |
| "loss": 0.5013, |
| "step": 700 |
| }, |
| { |
| "epoch": 0.14334384858044164, |
| "grad_norm": 0.4768179474157311, |
| "learning_rate": 9.94377839255047e-06, |
| "loss": 0.4945, |
| "step": 710 |
| }, |
| { |
| "epoch": 0.1453627760252366, |
| "grad_norm": 0.5086022999656893, |
| "learning_rate": 9.938386570326707e-06, |
| "loss": 0.5033, |
| "step": 720 |
| }, |
| { |
| "epoch": 0.14738170347003154, |
| "grad_norm": 0.4947936431118982, |
| "learning_rate": 9.93274950155518e-06, |
| "loss": 0.5013, |
| "step": 730 |
| }, |
| { |
| "epoch": 0.1494006309148265, |
| "grad_norm": 0.49370518434763616, |
| "learning_rate": 9.926867466179883e-06, |
| "loss": 0.4923, |
| "step": 740 |
| }, |
| { |
| "epoch": 0.15141955835962145, |
| "grad_norm": 0.49665569470371385, |
| "learning_rate": 9.920740756310156e-06, |
| "loss": 0.5038, |
| "step": 750 |
| }, |
| { |
| "epoch": 0.1534384858044164, |
| "grad_norm": 0.46813454785234254, |
| "learning_rate": 9.914369676206195e-06, |
| "loss": 0.4875, |
| "step": 760 |
| }, |
| { |
| "epoch": 0.15545741324921136, |
| "grad_norm": 0.49814972796278695, |
| "learning_rate": 9.907754542263926e-06, |
| "loss": 0.4939, |
| "step": 770 |
| }, |
| { |
| "epoch": 0.15747634069400632, |
| "grad_norm": 0.5218798292518447, |
| "learning_rate": 9.900895682999301e-06, |
| "loss": 0.501, |
| "step": 780 |
| }, |
| { |
| "epoch": 0.15949526813880127, |
| "grad_norm": 0.5913103279447773, |
| "learning_rate": 9.893793439031982e-06, |
| "loss": 0.4927, |
| "step": 790 |
| }, |
| { |
| "epoch": 0.16151419558359623, |
| "grad_norm": 0.5125446596453367, |
| "learning_rate": 9.88644816306842e-06, |
| "loss": 0.4971, |
| "step": 800 |
| }, |
| { |
| "epoch": 0.16353312302839118, |
| "grad_norm": 0.4483762668624327, |
| "learning_rate": 9.878860219884347e-06, |
| "loss": 0.4981, |
| "step": 810 |
| }, |
| { |
| "epoch": 0.1655520504731861, |
| "grad_norm": 0.4787860866647759, |
| "learning_rate": 9.871029986306658e-06, |
| "loss": 0.4964, |
| "step": 820 |
| }, |
| { |
| "epoch": 0.16757097791798106, |
| "grad_norm": 0.5030394247546482, |
| "learning_rate": 9.862957851194694e-06, |
| "loss": 0.4891, |
| "step": 830 |
| }, |
| { |
| "epoch": 0.16958990536277602, |
| "grad_norm": 0.5239985323894606, |
| "learning_rate": 9.854644215420933e-06, |
| "loss": 0.501, |
| "step": 840 |
| }, |
| { |
| "epoch": 0.17160883280757097, |
| "grad_norm": 0.4934736601735223, |
| "learning_rate": 9.846089491851089e-06, |
| "loss": 0.5012, |
| "step": 850 |
| }, |
| { |
| "epoch": 0.17362776025236593, |
| "grad_norm": 0.5649851363661309, |
| "learning_rate": 9.837294105323592e-06, |
| "loss": 0.501, |
| "step": 860 |
| }, |
| { |
| "epoch": 0.17564668769716088, |
| "grad_norm": 0.5032183289408247, |
| "learning_rate": 9.828258492628506e-06, |
| "loss": 0.5019, |
| "step": 870 |
| }, |
| { |
| "epoch": 0.17766561514195583, |
| "grad_norm": 0.5408394773542722, |
| "learning_rate": 9.818983102485832e-06, |
| "loss": 0.4914, |
| "step": 880 |
| }, |
| { |
| "epoch": 0.1796845425867508, |
| "grad_norm": 0.47080262308715565, |
| "learning_rate": 9.809468395523225e-06, |
| "loss": 0.4874, |
| "step": 890 |
| }, |
| { |
| "epoch": 0.18170347003154574, |
| "grad_norm": 0.5707749159791521, |
| "learning_rate": 9.799714844253109e-06, |
| "loss": 0.4995, |
| "step": 900 |
| }, |
| { |
| "epoch": 0.1837223974763407, |
| "grad_norm": 0.48122327046673424, |
| "learning_rate": 9.789722933049227e-06, |
| "loss": 0.4892, |
| "step": 910 |
| }, |
| { |
| "epoch": 0.18574132492113565, |
| "grad_norm": 0.45082444705324104, |
| "learning_rate": 9.779493158122578e-06, |
| "loss": 0.4909, |
| "step": 920 |
| }, |
| { |
| "epoch": 0.1877602523659306, |
| "grad_norm": 0.5068314998947344, |
| "learning_rate": 9.769026027496773e-06, |
| "loss": 0.4881, |
| "step": 930 |
| }, |
| { |
| "epoch": 0.18977917981072556, |
| "grad_norm": 0.44089908003969003, |
| "learning_rate": 9.758322060982813e-06, |
| "loss": 0.4923, |
| "step": 940 |
| }, |
| { |
| "epoch": 0.19179810725552052, |
| "grad_norm": 0.4786006068183123, |
| "learning_rate": 9.747381790153262e-06, |
| "loss": 0.4912, |
| "step": 950 |
| }, |
| { |
| "epoch": 0.19381703470031544, |
| "grad_norm": 0.4531421936277408, |
| "learning_rate": 9.736205758315864e-06, |
| "loss": 0.4744, |
| "step": 960 |
| }, |
| { |
| "epoch": 0.1958359621451104, |
| "grad_norm": 0.4492650045467347, |
| "learning_rate": 9.724794520486551e-06, |
| "loss": 0.485, |
| "step": 970 |
| }, |
| { |
| "epoch": 0.19785488958990535, |
| "grad_norm": 0.4616235423638888, |
| "learning_rate": 9.713148643361884e-06, |
| "loss": 0.488, |
| "step": 980 |
| }, |
| { |
| "epoch": 0.1998738170347003, |
| "grad_norm": 0.5078516721505932, |
| "learning_rate": 9.701268705290907e-06, |
| "loss": 0.4886, |
| "step": 990 |
| }, |
| { |
| "epoch": 0.20189274447949526, |
| "grad_norm": 0.5292871585531458, |
| "learning_rate": 9.689155296246431e-06, |
| "loss": 0.4897, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.20391167192429022, |
| "grad_norm": 0.4614297555636878, |
| "learning_rate": 9.67680901779573e-06, |
| "loss": 0.4809, |
| "step": 1010 |
| }, |
| { |
| "epoch": 0.20593059936908517, |
| "grad_norm": 0.4853770956696164, |
| "learning_rate": 9.664230483070669e-06, |
| "loss": 0.4848, |
| "step": 1020 |
| }, |
| { |
| "epoch": 0.20794952681388013, |
| "grad_norm": 0.45954875393494693, |
| "learning_rate": 9.651420316737254e-06, |
| "loss": 0.4817, |
| "step": 1030 |
| }, |
| { |
| "epoch": 0.20996845425867508, |
| "grad_norm": 0.4539790244117171, |
| "learning_rate": 9.638379154964615e-06, |
| "loss": 0.4927, |
| "step": 1040 |
| }, |
| { |
| "epoch": 0.21198738170347003, |
| "grad_norm": 0.4999117191134038, |
| "learning_rate": 9.625107645393398e-06, |
| "loss": 0.4906, |
| "step": 1050 |
| }, |
| { |
| "epoch": 0.214006309148265, |
| "grad_norm": 0.5315614747354604, |
| "learning_rate": 9.611606447103625e-06, |
| "loss": 0.4701, |
| "step": 1060 |
| }, |
| { |
| "epoch": 0.21602523659305994, |
| "grad_norm": 0.4465543761029164, |
| "learning_rate": 9.597876230581946e-06, |
| "loss": 0.4844, |
| "step": 1070 |
| }, |
| { |
| "epoch": 0.2180441640378549, |
| "grad_norm": 0.5240713468337838, |
| "learning_rate": 9.583917677688348e-06, |
| "loss": 0.4802, |
| "step": 1080 |
| }, |
| { |
| "epoch": 0.22006309148264985, |
| "grad_norm": 0.4797396408083518, |
| "learning_rate": 9.56973148162229e-06, |
| "loss": 0.4822, |
| "step": 1090 |
| }, |
| { |
| "epoch": 0.2220820189274448, |
| "grad_norm": 0.5124904557581254, |
| "learning_rate": 9.555318346888287e-06, |
| "loss": 0.4887, |
| "step": 1100 |
| }, |
| { |
| "epoch": 0.22410094637223973, |
| "grad_norm": 0.448720333545146, |
| "learning_rate": 9.540678989260907e-06, |
| "loss": 0.4804, |
| "step": 1110 |
| }, |
| { |
| "epoch": 0.2261198738170347, |
| "grad_norm": 0.4479157257263002, |
| "learning_rate": 9.525814135749238e-06, |
| "loss": 0.4776, |
| "step": 1120 |
| }, |
| { |
| "epoch": 0.22813880126182964, |
| "grad_norm": 0.4630993156519356, |
| "learning_rate": 9.510724524560783e-06, |
| "loss": 0.4753, |
| "step": 1130 |
| }, |
| { |
| "epoch": 0.2301577287066246, |
| "grad_norm": 0.46574160333254905, |
| "learning_rate": 9.495410905064795e-06, |
| "loss": 0.4802, |
| "step": 1140 |
| }, |
| { |
| "epoch": 0.23217665615141955, |
| "grad_norm": 0.532351729710861, |
| "learning_rate": 9.479874037755061e-06, |
| "loss": 0.4826, |
| "step": 1150 |
| }, |
| { |
| "epoch": 0.2341955835962145, |
| "grad_norm": 0.4629220502691385, |
| "learning_rate": 9.46411469421214e-06, |
| "loss": 0.4802, |
| "step": 1160 |
| }, |
| { |
| "epoch": 0.23621451104100946, |
| "grad_norm": 0.459458980682449, |
| "learning_rate": 9.448133657065045e-06, |
| "loss": 0.4863, |
| "step": 1170 |
| }, |
| { |
| "epoch": 0.23823343848580442, |
| "grad_norm": 0.4652002127480942, |
| "learning_rate": 9.431931719952372e-06, |
| "loss": 0.4785, |
| "step": 1180 |
| }, |
| { |
| "epoch": 0.24025236593059937, |
| "grad_norm": 0.466540922054408, |
| "learning_rate": 9.415509687482892e-06, |
| "loss": 0.4719, |
| "step": 1190 |
| }, |
| { |
| "epoch": 0.24227129337539433, |
| "grad_norm": 0.49341259273484933, |
| "learning_rate": 9.398868375195587e-06, |
| "loss": 0.4926, |
| "step": 1200 |
| }, |
| { |
| "epoch": 0.24429022082018928, |
| "grad_norm": 0.4728385366149605, |
| "learning_rate": 9.382008609519159e-06, |
| "loss": 0.4741, |
| "step": 1210 |
| }, |
| { |
| "epoch": 0.24630914826498423, |
| "grad_norm": 0.5027959432042377, |
| "learning_rate": 9.364931227730978e-06, |
| "loss": 0.4826, |
| "step": 1220 |
| }, |
| { |
| "epoch": 0.2483280757097792, |
| "grad_norm": 0.4883346568815589, |
| "learning_rate": 9.34763707791551e-06, |
| "loss": 0.4837, |
| "step": 1230 |
| }, |
| { |
| "epoch": 0.25034700315457414, |
| "grad_norm": 0.4689919758806841, |
| "learning_rate": 9.330127018922195e-06, |
| "loss": 0.4834, |
| "step": 1240 |
| }, |
| { |
| "epoch": 0.25236593059936907, |
| "grad_norm": 0.5054719137784303, |
| "learning_rate": 9.312401920322793e-06, |
| "loss": 0.4858, |
| "step": 1250 |
| }, |
| { |
| "epoch": 0.25438485804416405, |
| "grad_norm": 0.4829012063372003, |
| "learning_rate": 9.294462662368211e-06, |
| "loss": 0.4757, |
| "step": 1260 |
| }, |
| { |
| "epoch": 0.256403785488959, |
| "grad_norm": 0.4627206332424085, |
| "learning_rate": 9.276310135944777e-06, |
| "loss": 0.4783, |
| "step": 1270 |
| }, |
| { |
| "epoch": 0.25842271293375396, |
| "grad_norm": 0.4497460790448332, |
| "learning_rate": 9.257945242530002e-06, |
| "loss": 0.4785, |
| "step": 1280 |
| }, |
| { |
| "epoch": 0.2604416403785489, |
| "grad_norm": 0.44985051561838785, |
| "learning_rate": 9.239368894147812e-06, |
| "loss": 0.4786, |
| "step": 1290 |
| }, |
| { |
| "epoch": 0.26246056782334387, |
| "grad_norm": 0.48240095905903635, |
| "learning_rate": 9.220582013323258e-06, |
| "loss": 0.4807, |
| "step": 1300 |
| }, |
| { |
| "epoch": 0.2644794952681388, |
| "grad_norm": 0.48132105969556493, |
| "learning_rate": 9.201585533036696e-06, |
| "loss": 0.4706, |
| "step": 1310 |
| }, |
| { |
| "epoch": 0.2664984227129338, |
| "grad_norm": 0.46675857398369297, |
| "learning_rate": 9.18238039667746e-06, |
| "loss": 0.4879, |
| "step": 1320 |
| }, |
| { |
| "epoch": 0.2685173501577287, |
| "grad_norm": 0.4665634722250575, |
| "learning_rate": 9.162967557997003e-06, |
| "loss": 0.4799, |
| "step": 1330 |
| }, |
| { |
| "epoch": 0.27053627760252363, |
| "grad_norm": 0.4432315628923224, |
| "learning_rate": 9.14334798106155e-06, |
| "loss": 0.4757, |
| "step": 1340 |
| }, |
| { |
| "epoch": 0.2725552050473186, |
| "grad_norm": 0.5000831797296084, |
| "learning_rate": 9.123522640204198e-06, |
| "loss": 0.4805, |
| "step": 1350 |
| }, |
| { |
| "epoch": 0.27457413249211354, |
| "grad_norm": 0.4441933490286196, |
| "learning_rate": 9.103492519976555e-06, |
| "loss": 0.4769, |
| "step": 1360 |
| }, |
| { |
| "epoch": 0.2765930599369085, |
| "grad_norm": 0.4805005293661477, |
| "learning_rate": 9.083258615099819e-06, |
| "loss": 0.4777, |
| "step": 1370 |
| }, |
| { |
| "epoch": 0.27861198738170345, |
| "grad_norm": 0.4445460559749252, |
| "learning_rate": 9.062821930415404e-06, |
| "loss": 0.4728, |
| "step": 1380 |
| }, |
| { |
| "epoch": 0.28063091482649843, |
| "grad_norm": 0.4333540273890526, |
| "learning_rate": 9.042183480835018e-06, |
| "loss": 0.4705, |
| "step": 1390 |
| }, |
| { |
| "epoch": 0.28264984227129336, |
| "grad_norm": 0.48259999792627817, |
| "learning_rate": 9.021344291290275e-06, |
| "loss": 0.4782, |
| "step": 1400 |
| }, |
| { |
| "epoch": 0.28466876971608834, |
| "grad_norm": 0.439829603884435, |
| "learning_rate": 9.000305396681788e-06, |
| "loss": 0.4705, |
| "step": 1410 |
| }, |
| { |
| "epoch": 0.28668769716088327, |
| "grad_norm": 0.4776753948778827, |
| "learning_rate": 8.97906784182778e-06, |
| "loss": 0.4694, |
| "step": 1420 |
| }, |
| { |
| "epoch": 0.28870662460567825, |
| "grad_norm": 0.49821375314109334, |
| "learning_rate": 8.95763268141219e-06, |
| "loss": 0.4698, |
| "step": 1430 |
| }, |
| { |
| "epoch": 0.2907255520504732, |
| "grad_norm": 0.4626475218122892, |
| "learning_rate": 8.936000979932304e-06, |
| "loss": 0.4766, |
| "step": 1440 |
| }, |
| { |
| "epoch": 0.29274447949526816, |
| "grad_norm": 0.4797061046529278, |
| "learning_rate": 8.914173811645884e-06, |
| "loss": 0.4773, |
| "step": 1450 |
| }, |
| { |
| "epoch": 0.2947634069400631, |
| "grad_norm": 0.5067089149884695, |
| "learning_rate": 8.892152260517824e-06, |
| "loss": 0.4652, |
| "step": 1460 |
| }, |
| { |
| "epoch": 0.29678233438485807, |
| "grad_norm": 0.5216910139465747, |
| "learning_rate": 8.869937420166312e-06, |
| "loss": 0.475, |
| "step": 1470 |
| }, |
| { |
| "epoch": 0.298801261829653, |
| "grad_norm": 0.4655884354714463, |
| "learning_rate": 8.847530393808534e-06, |
| "loss": 0.4811, |
| "step": 1480 |
| }, |
| { |
| "epoch": 0.3008201892744479, |
| "grad_norm": 0.47533704121654896, |
| "learning_rate": 8.824932294205868e-06, |
| "loss": 0.4723, |
| "step": 1490 |
| }, |
| { |
| "epoch": 0.3028391167192429, |
| "grad_norm": 0.479585079565894, |
| "learning_rate": 8.802144243608633e-06, |
| "loss": 0.4728, |
| "step": 1500 |
| }, |
| { |
| "epoch": 0.30485804416403783, |
| "grad_norm": 0.4338897481002412, |
| "learning_rate": 8.779167373700359e-06, |
| "loss": 0.4687, |
| "step": 1510 |
| }, |
| { |
| "epoch": 0.3068769716088328, |
| "grad_norm": 0.51284609503485, |
| "learning_rate": 8.756002825541583e-06, |
| "loss": 0.4795, |
| "step": 1520 |
| }, |
| { |
| "epoch": 0.30889589905362774, |
| "grad_norm": 0.5019515223957067, |
| "learning_rate": 8.73265174951318e-06, |
| "loss": 0.4782, |
| "step": 1530 |
| }, |
| { |
| "epoch": 0.3109148264984227, |
| "grad_norm": 0.5764514697725889, |
| "learning_rate": 8.709115305259233e-06, |
| "loss": 0.4722, |
| "step": 1540 |
| }, |
| { |
| "epoch": 0.31293375394321765, |
| "grad_norm": 0.48602580093856246, |
| "learning_rate": 8.685394661629453e-06, |
| "loss": 0.4764, |
| "step": 1550 |
| }, |
| { |
| "epoch": 0.31495268138801263, |
| "grad_norm": 0.43106495742778833, |
| "learning_rate": 8.661490996621125e-06, |
| "loss": 0.4749, |
| "step": 1560 |
| }, |
| { |
| "epoch": 0.31697160883280756, |
| "grad_norm": 0.4879400234119546, |
| "learning_rate": 8.637405497320599e-06, |
| "loss": 0.4685, |
| "step": 1570 |
| }, |
| { |
| "epoch": 0.31899053627760254, |
| "grad_norm": 0.48144632266967335, |
| "learning_rate": 8.61313935984436e-06, |
| "loss": 0.473, |
| "step": 1580 |
| }, |
| { |
| "epoch": 0.32100946372239747, |
| "grad_norm": 0.4824945838927638, |
| "learning_rate": 8.588693789279605e-06, |
| "loss": 0.467, |
| "step": 1590 |
| }, |
| { |
| "epoch": 0.32302839116719245, |
| "grad_norm": 0.48854234491790616, |
| "learning_rate": 8.56406999962441e-06, |
| "loss": 0.4715, |
| "step": 1600 |
| }, |
| { |
| "epoch": 0.3250473186119874, |
| "grad_norm": 0.43356661154496373, |
| "learning_rate": 8.539269213727434e-06, |
| "loss": 0.4643, |
| "step": 1610 |
| }, |
| { |
| "epoch": 0.32706624605678236, |
| "grad_norm": 0.45038524364573934, |
| "learning_rate": 8.514292663227196e-06, |
| "loss": 0.4753, |
| "step": 1620 |
| }, |
| { |
| "epoch": 0.3290851735015773, |
| "grad_norm": 0.5390076730569227, |
| "learning_rate": 8.489141588490914e-06, |
| "loss": 0.4692, |
| "step": 1630 |
| }, |
| { |
| "epoch": 0.3311041009463722, |
| "grad_norm": 0.5166665541948204, |
| "learning_rate": 8.46381723855289e-06, |
| "loss": 0.4626, |
| "step": 1640 |
| }, |
| { |
| "epoch": 0.3331230283911672, |
| "grad_norm": 0.4973742178832253, |
| "learning_rate": 8.438320871052501e-06, |
| "loss": 0.4722, |
| "step": 1650 |
| }, |
| { |
| "epoch": 0.3351419558359621, |
| "grad_norm": 0.4625137579015689, |
| "learning_rate": 8.412653752171732e-06, |
| "loss": 0.4591, |
| "step": 1660 |
| }, |
| { |
| "epoch": 0.3371608832807571, |
| "grad_norm": 0.48424933592035946, |
| "learning_rate": 8.386817156572298e-06, |
| "loss": 0.48, |
| "step": 1670 |
| }, |
| { |
| "epoch": 0.33917981072555203, |
| "grad_norm": 0.4532582849486311, |
| "learning_rate": 8.360812367332342e-06, |
| "loss": 0.469, |
| "step": 1680 |
| }, |
| { |
| "epoch": 0.341198738170347, |
| "grad_norm": 0.4456128521040015, |
| "learning_rate": 8.33464067588272e-06, |
| "loss": 0.4701, |
| "step": 1690 |
| }, |
| { |
| "epoch": 0.34321766561514194, |
| "grad_norm": 0.47975070454468693, |
| "learning_rate": 8.30830338194286e-06, |
| "loss": 0.4742, |
| "step": 1700 |
| }, |
| { |
| "epoch": 0.3452365930599369, |
| "grad_norm": 0.44758135515111036, |
| "learning_rate": 8.281801793456223e-06, |
| "loss": 0.4563, |
| "step": 1710 |
| }, |
| { |
| "epoch": 0.34725552050473185, |
| "grad_norm": 0.539530429343652, |
| "learning_rate": 8.255137226525346e-06, |
| "loss": 0.4678, |
| "step": 1720 |
| }, |
| { |
| "epoch": 0.34927444794952683, |
| "grad_norm": 0.44634229380095736, |
| "learning_rate": 8.22831100534648e-06, |
| "loss": 0.4688, |
| "step": 1730 |
| }, |
| { |
| "epoch": 0.35129337539432176, |
| "grad_norm": 0.41960408937669463, |
| "learning_rate": 8.201324462143827e-06, |
| "loss": 0.4542, |
| "step": 1740 |
| }, |
| { |
| "epoch": 0.35331230283911674, |
| "grad_norm": 0.4791531544406239, |
| "learning_rate": 8.174178937103397e-06, |
| "loss": 0.4743, |
| "step": 1750 |
| }, |
| { |
| "epoch": 0.35533123028391167, |
| "grad_norm": 0.4352818312435405, |
| "learning_rate": 8.14687577830643e-06, |
| "loss": 0.4604, |
| "step": 1760 |
| }, |
| { |
| "epoch": 0.35735015772870665, |
| "grad_norm": 0.4672874879206997, |
| "learning_rate": 8.11941634166246e-06, |
| "loss": 0.4712, |
| "step": 1770 |
| }, |
| { |
| "epoch": 0.3593690851735016, |
| "grad_norm": 0.4637056308756207, |
| "learning_rate": 8.09180199084198e-06, |
| "loss": 0.4668, |
| "step": 1780 |
| }, |
| { |
| "epoch": 0.3613880126182965, |
| "grad_norm": 0.44264459219210456, |
| "learning_rate": 8.064034097208712e-06, |
| "loss": 0.4596, |
| "step": 1790 |
| }, |
| { |
| "epoch": 0.3634069400630915, |
| "grad_norm": 0.49462230789807315, |
| "learning_rate": 8.036114039751516e-06, |
| "loss": 0.4667, |
| "step": 1800 |
| }, |
| { |
| "epoch": 0.3654258675078864, |
| "grad_norm": 0.4634309345247295, |
| "learning_rate": 8.0080432050159e-06, |
| "loss": 0.4604, |
| "step": 1810 |
| }, |
| { |
| "epoch": 0.3674447949526814, |
| "grad_norm": 0.4999816402185204, |
| "learning_rate": 7.979822987035157e-06, |
| "loss": 0.4644, |
| "step": 1820 |
| }, |
| { |
| "epoch": 0.3694637223974763, |
| "grad_norm": 0.47642327436037435, |
| "learning_rate": 7.951454787261148e-06, |
| "loss": 0.4677, |
| "step": 1830 |
| }, |
| { |
| "epoch": 0.3714826498422713, |
| "grad_norm": 0.4953986824691425, |
| "learning_rate": 7.922940014494693e-06, |
| "loss": 0.4593, |
| "step": 1840 |
| }, |
| { |
| "epoch": 0.37350157728706623, |
| "grad_norm": 0.4817430868786167, |
| "learning_rate": 7.894280084815618e-06, |
| "loss": 0.4689, |
| "step": 1850 |
| }, |
| { |
| "epoch": 0.3755205047318612, |
| "grad_norm": 0.4356746961237441, |
| "learning_rate": 7.86547642151242e-06, |
| "loss": 0.4611, |
| "step": 1860 |
| }, |
| { |
| "epoch": 0.37753943217665614, |
| "grad_norm": 0.4905696785040258, |
| "learning_rate": 7.836530455011594e-06, |
| "loss": 0.4647, |
| "step": 1870 |
| }, |
| { |
| "epoch": 0.3795583596214511, |
| "grad_norm": 0.4730705035791906, |
| "learning_rate": 7.807443622806592e-06, |
| "loss": 0.4643, |
| "step": 1880 |
| }, |
| { |
| "epoch": 0.38157728706624605, |
| "grad_norm": 0.4626721574097504, |
| "learning_rate": 7.778217369386434e-06, |
| "loss": 0.4621, |
| "step": 1890 |
| }, |
| { |
| "epoch": 0.38359621451104103, |
| "grad_norm": 0.5095748907639542, |
| "learning_rate": 7.748853146163978e-06, |
| "loss": 0.4702, |
| "step": 1900 |
| }, |
| { |
| "epoch": 0.38561514195583596, |
| "grad_norm": 0.45666908318697785, |
| "learning_rate": 7.719352411403836e-06, |
| "loss": 0.4584, |
| "step": 1910 |
| }, |
| { |
| "epoch": 0.3876340694006309, |
| "grad_norm": 0.5831960484166556, |
| "learning_rate": 7.68971663014995e-06, |
| "loss": 0.4581, |
| "step": 1920 |
| }, |
| { |
| "epoch": 0.38965299684542587, |
| "grad_norm": 0.43083136772614855, |
| "learning_rate": 7.659947274152857e-06, |
| "loss": 0.4595, |
| "step": 1930 |
| }, |
| { |
| "epoch": 0.3916719242902208, |
| "grad_norm": 0.44527597409724967, |
| "learning_rate": 7.630045821796567e-06, |
| "loss": 0.4603, |
| "step": 1940 |
| }, |
| { |
| "epoch": 0.3936908517350158, |
| "grad_norm": 0.5085813030698595, |
| "learning_rate": 7.60001375802518e-06, |
| "loss": 0.4726, |
| "step": 1950 |
| }, |
| { |
| "epoch": 0.3957097791798107, |
| "grad_norm": 0.5516780680340057, |
| "learning_rate": 7.5698525742691145e-06, |
| "loss": 0.4553, |
| "step": 1960 |
| }, |
| { |
| "epoch": 0.3977287066246057, |
| "grad_norm": 0.4473799858576117, |
| "learning_rate": 7.5395637683710534e-06, |
| "loss": 0.4652, |
| "step": 1970 |
| }, |
| { |
| "epoch": 0.3997476340694006, |
| "grad_norm": 0.4594068744960154, |
| "learning_rate": 7.509148844511556e-06, |
| "loss": 0.4569, |
| "step": 1980 |
| }, |
| { |
| "epoch": 0.4017665615141956, |
| "grad_norm": 0.4195303708920141, |
| "learning_rate": 7.478609313134367e-06, |
| "loss": 0.4608, |
| "step": 1990 |
| }, |
| { |
| "epoch": 0.4037854889589905, |
| "grad_norm": 0.5224611450690858, |
| "learning_rate": 7.447946690871387e-06, |
| "loss": 0.4618, |
| "step": 2000 |
| }, |
| { |
| "epoch": 0.4058044164037855, |
| "grad_norm": 0.5037868351130359, |
| "learning_rate": 7.4171625004673765e-06, |
| "loss": 0.4549, |
| "step": 2010 |
| }, |
| { |
| "epoch": 0.40782334384858043, |
| "grad_norm": 0.6617292048565804, |
| "learning_rate": 7.386258270704318e-06, |
| "loss": 0.4648, |
| "step": 2020 |
| }, |
| { |
| "epoch": 0.4098422712933754, |
| "grad_norm": 0.46778481767707863, |
| "learning_rate": 7.355235536325507e-06, |
| "loss": 0.4618, |
| "step": 2030 |
| }, |
| { |
| "epoch": 0.41186119873817034, |
| "grad_norm": 0.4467025867029599, |
| "learning_rate": 7.32409583795932e-06, |
| "loss": 0.4556, |
| "step": 2040 |
| }, |
| { |
| "epoch": 0.4138801261829653, |
| "grad_norm": 0.4864826536631601, |
| "learning_rate": 7.292840722042721e-06, |
| "loss": 0.4649, |
| "step": 2050 |
| }, |
| { |
| "epoch": 0.41589905362776025, |
| "grad_norm": 0.44674303304698915, |
| "learning_rate": 7.261471740744453e-06, |
| "loss": 0.4546, |
| "step": 2060 |
| }, |
| { |
| "epoch": 0.4179179810725552, |
| "grad_norm": 0.4638264920705765, |
| "learning_rate": 7.229990451887961e-06, |
| "loss": 0.459, |
| "step": 2070 |
| }, |
| { |
| "epoch": 0.41993690851735016, |
| "grad_norm": 0.44062835097290626, |
| "learning_rate": 7.1983984188740215e-06, |
| "loss": 0.4627, |
| "step": 2080 |
| }, |
| { |
| "epoch": 0.4219558359621451, |
| "grad_norm": 0.4596889143174662, |
| "learning_rate": 7.1666972106031105e-06, |
| "loss": 0.4584, |
| "step": 2090 |
| }, |
| { |
| "epoch": 0.42397476340694007, |
| "grad_norm": 0.4866244769293159, |
| "learning_rate": 7.134888401397484e-06, |
| "loss": 0.4629, |
| "step": 2100 |
| }, |
| { |
| "epoch": 0.425993690851735, |
| "grad_norm": 0.5201676761862537, |
| "learning_rate": 7.102973570922997e-06, |
| "loss": 0.4549, |
| "step": 2110 |
| }, |
| { |
| "epoch": 0.42801261829653, |
| "grad_norm": 0.4790871394995968, |
| "learning_rate": 7.070954304110656e-06, |
| "loss": 0.4632, |
| "step": 2120 |
| }, |
| { |
| "epoch": 0.4300315457413249, |
| "grad_norm": 0.4659786396553266, |
| "learning_rate": 7.03883219107791e-06, |
| "loss": 0.4581, |
| "step": 2130 |
| }, |
| { |
| "epoch": 0.4320504731861199, |
| "grad_norm": 0.4104121258537053, |
| "learning_rate": 7.0066088270496775e-06, |
| "loss": 0.4596, |
| "step": 2140 |
| }, |
| { |
| "epoch": 0.4340694006309148, |
| "grad_norm": 0.46988504036618683, |
| "learning_rate": 6.9742858122791355e-06, |
| "loss": 0.4648, |
| "step": 2150 |
| }, |
| { |
| "epoch": 0.4360883280757098, |
| "grad_norm": 0.6098242782979109, |
| "learning_rate": 6.941864751968239e-06, |
| "loss": 0.4548, |
| "step": 2160 |
| }, |
| { |
| "epoch": 0.4381072555205047, |
| "grad_norm": 0.7557069029520974, |
| "learning_rate": 6.90934725618801e-06, |
| "loss": 0.461, |
| "step": 2170 |
| }, |
| { |
| "epoch": 0.4401261829652997, |
| "grad_norm": 0.46196984209426056, |
| "learning_rate": 6.876734939798584e-06, |
| "loss": 0.4631, |
| "step": 2180 |
| }, |
| { |
| "epoch": 0.44214511041009463, |
| "grad_norm": 0.45367052105694244, |
| "learning_rate": 6.844029422368998e-06, |
| "loss": 0.4567, |
| "step": 2190 |
| }, |
| { |
| "epoch": 0.4441640378548896, |
| "grad_norm": 0.4923825492028847, |
| "learning_rate": 6.811232328096778e-06, |
| "loss": 0.4619, |
| "step": 2200 |
| }, |
| { |
| "epoch": 0.44618296529968454, |
| "grad_norm": 0.4412882669817311, |
| "learning_rate": 6.7783452857272725e-06, |
| "loss": 0.4472, |
| "step": 2210 |
| }, |
| { |
| "epoch": 0.44820189274447947, |
| "grad_norm": 0.4804805684578094, |
| "learning_rate": 6.745369928472764e-06, |
| "loss": 0.4634, |
| "step": 2220 |
| }, |
| { |
| "epoch": 0.45022082018927445, |
| "grad_norm": 0.46765384977250785, |
| "learning_rate": 6.712307893931364e-06, |
| "loss": 0.4595, |
| "step": 2230 |
| }, |
| { |
| "epoch": 0.4522397476340694, |
| "grad_norm": 1.0068029764875739, |
| "learning_rate": 6.67916082400569e-06, |
| "loss": 0.457, |
| "step": 2240 |
| }, |
| { |
| "epoch": 0.45425867507886436, |
| "grad_norm": 0.4828558935559972, |
| "learning_rate": 6.645930364821327e-06, |
| "loss": 0.465, |
| "step": 2250 |
| }, |
| { |
| "epoch": 0.4562776025236593, |
| "grad_norm": 0.4571716607551568, |
| "learning_rate": 6.612618166645068e-06, |
| "loss": 0.4521, |
| "step": 2260 |
| }, |
| { |
| "epoch": 0.45829652996845427, |
| "grad_norm": 0.49891810921169605, |
| "learning_rate": 6.579225883802981e-06, |
| "loss": 0.4621, |
| "step": 2270 |
| }, |
| { |
| "epoch": 0.4603154574132492, |
| "grad_norm": 0.6812817990255348, |
| "learning_rate": 6.5457551745982275e-06, |
| "loss": 0.4553, |
| "step": 2280 |
| }, |
| { |
| "epoch": 0.4623343848580442, |
| "grad_norm": 0.466296009966539, |
| "learning_rate": 6.512207701228727e-06, |
| "loss": 0.4585, |
| "step": 2290 |
| }, |
| { |
| "epoch": 0.4643533123028391, |
| "grad_norm": 0.48144166164862184, |
| "learning_rate": 6.478585129704608e-06, |
| "loss": 0.4579, |
| "step": 2300 |
| }, |
| { |
| "epoch": 0.4663722397476341, |
| "grad_norm": 0.4575454070862058, |
| "learning_rate": 6.444889129765465e-06, |
| "loss": 0.453, |
| "step": 2310 |
| }, |
| { |
| "epoch": 0.468391167192429, |
| "grad_norm": 0.5049212463992705, |
| "learning_rate": 6.4111213747974385e-06, |
| "loss": 0.4606, |
| "step": 2320 |
| }, |
| { |
| "epoch": 0.470410094637224, |
| "grad_norm": 0.44196403700173553, |
| "learning_rate": 6.377283541750123e-06, |
| "loss": 0.4528, |
| "step": 2330 |
| }, |
| { |
| "epoch": 0.4724290220820189, |
| "grad_norm": 0.44557725863878417, |
| "learning_rate": 6.343377311053271e-06, |
| "loss": 0.4614, |
| "step": 2340 |
| }, |
| { |
| "epoch": 0.4744479495268139, |
| "grad_norm": 0.485138333868983, |
| "learning_rate": 6.309404366533355e-06, |
| "loss": 0.4662, |
| "step": 2350 |
| }, |
| { |
| "epoch": 0.47646687697160883, |
| "grad_norm": 0.44703181568848005, |
| "learning_rate": 6.275366395329934e-06, |
| "loss": 0.4484, |
| "step": 2360 |
| }, |
| { |
| "epoch": 0.47848580441640376, |
| "grad_norm": 0.46684239360571517, |
| "learning_rate": 6.241265087811883e-06, |
| "loss": 0.461, |
| "step": 2370 |
| }, |
| { |
| "epoch": 0.48050473186119874, |
| "grad_norm": 0.4308686375763305, |
| "learning_rate": 6.207102137493433e-06, |
| "loss": 0.4572, |
| "step": 2380 |
| }, |
| { |
| "epoch": 0.48252365930599367, |
| "grad_norm": 0.4349332573681559, |
| "learning_rate": 6.172879240950075e-06, |
| "loss": 0.4586, |
| "step": 2390 |
| }, |
| { |
| "epoch": 0.48454258675078865, |
| "grad_norm": 0.4946893193107334, |
| "learning_rate": 6.138598097734308e-06, |
| "loss": 0.4613, |
| "step": 2400 |
| }, |
| { |
| "epoch": 0.4865615141955836, |
| "grad_norm": 0.4278600145969151, |
| "learning_rate": 6.104260410291237e-06, |
| "loss": 0.4524, |
| "step": 2410 |
| }, |
| { |
| "epoch": 0.48858044164037856, |
| "grad_norm": 0.438261930357501, |
| "learning_rate": 6.069867883874017e-06, |
| "loss": 0.4546, |
| "step": 2420 |
| }, |
| { |
| "epoch": 0.4905993690851735, |
| "grad_norm": 0.4834328335496326, |
| "learning_rate": 6.035422226459183e-06, |
| "loss": 0.4601, |
| "step": 2430 |
| }, |
| { |
| "epoch": 0.49261829652996847, |
| "grad_norm": 0.4560202224259225, |
| "learning_rate": 6.000925148661822e-06, |
| "loss": 0.4552, |
| "step": 2440 |
| }, |
| { |
| "epoch": 0.4946372239747634, |
| "grad_norm": 0.46242134458299394, |
| "learning_rate": 5.966378363650618e-06, |
| "loss": 0.4534, |
| "step": 2450 |
| }, |
| { |
| "epoch": 0.4966561514195584, |
| "grad_norm": 0.509798379798209, |
| "learning_rate": 5.931783587062785e-06, |
| "loss": 0.4478, |
| "step": 2460 |
| }, |
| { |
| "epoch": 0.4986750788643533, |
| "grad_norm": 0.4916270732208669, |
| "learning_rate": 5.897142536918854e-06, |
| "loss": 0.4606, |
| "step": 2470 |
| }, |
| { |
| "epoch": 0.5006940063091483, |
| "grad_norm": 0.4493237090311729, |
| "learning_rate": 5.8624569335373625e-06, |
| "loss": 0.4512, |
| "step": 2480 |
| }, |
| { |
| "epoch": 0.5027129337539432, |
| "grad_norm": 0.4570244860395198, |
| "learning_rate": 5.827728499449417e-06, |
| "loss": 0.4503, |
| "step": 2490 |
| }, |
| { |
| "epoch": 0.5047318611987381, |
| "grad_norm": 0.5044698787582333, |
| "learning_rate": 5.792958959313155e-06, |
| "loss": 0.4586, |
| "step": 2500 |
| }, |
| { |
| "epoch": 0.5067507886435331, |
| "grad_norm": 0.4567238497826546, |
| "learning_rate": 5.758150039828084e-06, |
| "loss": 0.442, |
| "step": 2510 |
| }, |
| { |
| "epoch": 0.5087697160883281, |
| "grad_norm": 0.49456514738525703, |
| "learning_rate": 5.723303469649352e-06, |
| "loss": 0.4524, |
| "step": 2520 |
| }, |
| { |
| "epoch": 0.510788643533123, |
| "grad_norm": 0.46129063291343286, |
| "learning_rate": 5.688420979301876e-06, |
| "loss": 0.4533, |
| "step": 2530 |
| }, |
| { |
| "epoch": 0.512807570977918, |
| "grad_norm": 0.46780830492612324, |
| "learning_rate": 5.6535043010944226e-06, |
| "loss": 0.4537, |
| "step": 2540 |
| }, |
| { |
| "epoch": 0.5148264984227129, |
| "grad_norm": 0.49663726459668367, |
| "learning_rate": 5.618555169033567e-06, |
| "loss": 0.4624, |
| "step": 2550 |
| }, |
| { |
| "epoch": 0.5168454258675079, |
| "grad_norm": 0.4323698197626643, |
| "learning_rate": 5.583575318737588e-06, |
| "loss": 0.4394, |
| "step": 2560 |
| }, |
| { |
| "epoch": 0.5188643533123029, |
| "grad_norm": 0.433967581020834, |
| "learning_rate": 5.548566487350264e-06, |
| "loss": 0.4482, |
| "step": 2570 |
| }, |
| { |
| "epoch": 0.5208832807570978, |
| "grad_norm": 0.5093937708469708, |
| "learning_rate": 5.513530413454618e-06, |
| "loss": 0.4488, |
| "step": 2580 |
| }, |
| { |
| "epoch": 0.5229022082018927, |
| "grad_norm": 0.4479313579903872, |
| "learning_rate": 5.478468836986568e-06, |
| "loss": 0.4547, |
| "step": 2590 |
| }, |
| { |
| "epoch": 0.5249211356466877, |
| "grad_norm": 0.4784953987086218, |
| "learning_rate": 5.443383499148519e-06, |
| "loss": 0.4536, |
| "step": 2600 |
| }, |
| { |
| "epoch": 0.5269400630914827, |
| "grad_norm": 0.4311672354723456, |
| "learning_rate": 5.408276142322904e-06, |
| "loss": 0.4489, |
| "step": 2610 |
| }, |
| { |
| "epoch": 0.5289589905362776, |
| "grad_norm": 0.4320377418884208, |
| "learning_rate": 5.373148509985636e-06, |
| "loss": 0.457, |
| "step": 2620 |
| }, |
| { |
| "epoch": 0.5309779179810725, |
| "grad_norm": 0.4475608921551904, |
| "learning_rate": 5.3380023466195455e-06, |
| "loss": 0.4491, |
| "step": 2630 |
| }, |
| { |
| "epoch": 0.5329968454258676, |
| "grad_norm": 0.4150312768857391, |
| "learning_rate": 5.302839397627732e-06, |
| "loss": 0.4509, |
| "step": 2640 |
| }, |
| { |
| "epoch": 0.5350157728706625, |
| "grad_norm": 0.4842663915075593, |
| "learning_rate": 5.267661409246893e-06, |
| "loss": 0.4497, |
| "step": 2650 |
| }, |
| { |
| "epoch": 0.5370347003154574, |
| "grad_norm": 0.48020608359904937, |
| "learning_rate": 5.232470128460597e-06, |
| "loss": 0.4455, |
| "step": 2660 |
| }, |
| { |
| "epoch": 0.5390536277602523, |
| "grad_norm": 0.460861374110654, |
| "learning_rate": 5.197267302912535e-06, |
| "loss": 0.4633, |
| "step": 2670 |
| }, |
| { |
| "epoch": 0.5410725552050473, |
| "grad_norm": 0.4861656655295904, |
| "learning_rate": 5.162054680819724e-06, |
| "loss": 0.455, |
| "step": 2680 |
| }, |
| { |
| "epoch": 0.5430914826498423, |
| "grad_norm": 0.4280331150720691, |
| "learning_rate": 5.12683401088569e-06, |
| "loss": 0.4527, |
| "step": 2690 |
| }, |
| { |
| "epoch": 0.5451104100946372, |
| "grad_norm": 0.4803934844859794, |
| "learning_rate": 5.091607042213621e-06, |
| "loss": 0.4507, |
| "step": 2700 |
| }, |
| { |
| "epoch": 0.5471293375394322, |
| "grad_norm": 0.44679887041219724, |
| "learning_rate": 5.056375524219516e-06, |
| "loss": 0.4467, |
| "step": 2710 |
| }, |
| { |
| "epoch": 0.5491482649842271, |
| "grad_norm": 0.4562693718725788, |
| "learning_rate": 5.021141206545293e-06, |
| "loss": 0.4481, |
| "step": 2720 |
| }, |
| { |
| "epoch": 0.5511671924290221, |
| "grad_norm": 0.4568391909726876, |
| "learning_rate": 4.9859058389719095e-06, |
| "loss": 0.4467, |
| "step": 2730 |
| }, |
| { |
| "epoch": 0.553186119873817, |
| "grad_norm": 0.4402113900197009, |
| "learning_rate": 4.950671171332458e-06, |
| "loss": 0.4438, |
| "step": 2740 |
| }, |
| { |
| "epoch": 0.555205047318612, |
| "grad_norm": 0.4694901686129437, |
| "learning_rate": 4.915438953425277e-06, |
| "loss": 0.455, |
| "step": 2750 |
| }, |
| { |
| "epoch": 0.5572239747634069, |
| "grad_norm": 0.5092377388071306, |
| "learning_rate": 4.8802109349270415e-06, |
| "loss": 0.4431, |
| "step": 2760 |
| }, |
| { |
| "epoch": 0.5592429022082019, |
| "grad_norm": 0.5334478668234227, |
| "learning_rate": 4.844988865305883e-06, |
| "loss": 0.4457, |
| "step": 2770 |
| }, |
| { |
| "epoch": 0.5612618296529969, |
| "grad_norm": 0.450190134489615, |
| "learning_rate": 4.809774493734507e-06, |
| "loss": 0.4468, |
| "step": 2780 |
| }, |
| { |
| "epoch": 0.5632807570977918, |
| "grad_norm": 0.6136460189487958, |
| "learning_rate": 4.774569569003314e-06, |
| "loss": 0.4531, |
| "step": 2790 |
| }, |
| { |
| "epoch": 0.5652996845425867, |
| "grad_norm": 0.47669716943494894, |
| "learning_rate": 4.739375839433575e-06, |
| "loss": 0.4559, |
| "step": 2800 |
| }, |
| { |
| "epoch": 0.5673186119873816, |
| "grad_norm": 0.47396268716032475, |
| "learning_rate": 4.704195052790584e-06, |
| "loss": 0.4463, |
| "step": 2810 |
| }, |
| { |
| "epoch": 0.5693375394321767, |
| "grad_norm": 0.5115114860936024, |
| "learning_rate": 4.66902895619688e-06, |
| "loss": 0.4494, |
| "step": 2820 |
| }, |
| { |
| "epoch": 0.5713564668769716, |
| "grad_norm": 0.4561358589231192, |
| "learning_rate": 4.633879296045474e-06, |
| "loss": 0.4477, |
| "step": 2830 |
| }, |
| { |
| "epoch": 0.5733753943217665, |
| "grad_norm": 0.4547946639578122, |
| "learning_rate": 4.59874781791312e-06, |
| "loss": 0.4502, |
| "step": 2840 |
| }, |
| { |
| "epoch": 0.5753943217665615, |
| "grad_norm": 0.5036899795994354, |
| "learning_rate": 4.56363626647363e-06, |
| "loss": 0.4553, |
| "step": 2850 |
| }, |
| { |
| "epoch": 0.5774132492113565, |
| "grad_norm": 0.46515739613875146, |
| "learning_rate": 4.528546385411239e-06, |
| "loss": 0.4373, |
| "step": 2860 |
| }, |
| { |
| "epoch": 0.5794321766561514, |
| "grad_norm": 0.4834989187992637, |
| "learning_rate": 4.493479917333989e-06, |
| "loss": 0.453, |
| "step": 2870 |
| }, |
| { |
| "epoch": 0.5814511041009464, |
| "grad_norm": 0.45922181466381384, |
| "learning_rate": 4.4584386036872155e-06, |
| "loss": 0.4527, |
| "step": 2880 |
| }, |
| { |
| "epoch": 0.5834700315457413, |
| "grad_norm": 0.46803871925708684, |
| "learning_rate": 4.42342418466705e-06, |
| "loss": 0.4426, |
| "step": 2890 |
| }, |
| { |
| "epoch": 0.5854889589905363, |
| "grad_norm": 0.47319916436776355, |
| "learning_rate": 4.388438399134e-06, |
| "loss": 0.4547, |
| "step": 2900 |
| }, |
| { |
| "epoch": 0.5875078864353313, |
| "grad_norm": 0.4598569796278221, |
| "learning_rate": 4.3534829845265996e-06, |
| "loss": 0.4441, |
| "step": 2910 |
| }, |
| { |
| "epoch": 0.5895268138801262, |
| "grad_norm": 0.47100191118178447, |
| "learning_rate": 4.318559676775126e-06, |
| "loss": 0.4472, |
| "step": 2920 |
| }, |
| { |
| "epoch": 0.5915457413249211, |
| "grad_norm": 0.495923691841689, |
| "learning_rate": 4.283670210215384e-06, |
| "loss": 0.4516, |
| "step": 2930 |
| }, |
| { |
| "epoch": 0.5935646687697161, |
| "grad_norm": 0.4391764043906036, |
| "learning_rate": 4.248816317502585e-06, |
| "loss": 0.4468, |
| "step": 2940 |
| }, |
| { |
| "epoch": 0.5955835962145111, |
| "grad_norm": 0.47900304182889214, |
| "learning_rate": 4.213999729525304e-06, |
| "loss": 0.452, |
| "step": 2950 |
| }, |
| { |
| "epoch": 0.597602523659306, |
| "grad_norm": 0.4894376668991494, |
| "learning_rate": 4.179222175319502e-06, |
| "loss": 0.4418, |
| "step": 2960 |
| }, |
| { |
| "epoch": 0.5996214511041009, |
| "grad_norm": 0.4698266233405637, |
| "learning_rate": 4.1444853819826845e-06, |
| "loss": 0.4503, |
| "step": 2970 |
| }, |
| { |
| "epoch": 0.6016403785488958, |
| "grad_norm": 0.4601994675920038, |
| "learning_rate": 4.1097910745881175e-06, |
| "loss": 0.4421, |
| "step": 2980 |
| }, |
| { |
| "epoch": 0.6036593059936909, |
| "grad_norm": 0.6167906848346414, |
| "learning_rate": 4.0751409760991585e-06, |
| "loss": 0.4469, |
| "step": 2990 |
| }, |
| { |
| "epoch": 0.6056782334384858, |
| "grad_norm": 0.5051790895352023, |
| "learning_rate": 4.040536807283696e-06, |
| "loss": 0.4502, |
| "step": 3000 |
| }, |
| { |
| "epoch": 0.6076971608832807, |
| "grad_norm": 0.45322450050363433, |
| "learning_rate": 4.005980286628698e-06, |
| "loss": 0.4505, |
| "step": 3010 |
| }, |
| { |
| "epoch": 0.6097160883280757, |
| "grad_norm": 0.4910869337898713, |
| "learning_rate": 3.9714731302548525e-06, |
| "loss": 0.4486, |
| "step": 3020 |
| }, |
| { |
| "epoch": 0.6117350157728707, |
| "grad_norm": 0.447631013542978, |
| "learning_rate": 3.937017051831367e-06, |
| "loss": 0.4478, |
| "step": 3030 |
| }, |
| { |
| "epoch": 0.6137539432176656, |
| "grad_norm": 0.7018069369756337, |
| "learning_rate": 3.9026137624908466e-06, |
| "loss": 0.4492, |
| "step": 3040 |
| }, |
| { |
| "epoch": 0.6157728706624606, |
| "grad_norm": 0.4762076148803905, |
| "learning_rate": 3.868264970744324e-06, |
| "loss": 0.4513, |
| "step": 3050 |
| }, |
| { |
| "epoch": 0.6177917981072555, |
| "grad_norm": 0.5020509623933802, |
| "learning_rate": 3.833972382396412e-06, |
| "loss": 0.4482, |
| "step": 3060 |
| }, |
| { |
| "epoch": 0.6198107255520505, |
| "grad_norm": 0.44347011131653674, |
| "learning_rate": 3.799737700460594e-06, |
| "loss": 0.4441, |
| "step": 3070 |
| }, |
| { |
| "epoch": 0.6218296529968455, |
| "grad_norm": 0.45353690385677753, |
| "learning_rate": 3.7655626250746446e-06, |
| "loss": 0.4443, |
| "step": 3080 |
| }, |
| { |
| "epoch": 0.6238485804416404, |
| "grad_norm": 0.44404552147051596, |
| "learning_rate": 3.7314488534162007e-06, |
| "loss": 0.4509, |
| "step": 3090 |
| }, |
| { |
| "epoch": 0.6258675078864353, |
| "grad_norm": 0.4759693579595236, |
| "learning_rate": 3.6973980796184843e-06, |
| "loss": 0.4573, |
| "step": 3100 |
| }, |
| { |
| "epoch": 0.6278864353312302, |
| "grad_norm": 0.45248182104327295, |
| "learning_rate": 3.6634119946861556e-06, |
| "loss": 0.4442, |
| "step": 3110 |
| }, |
| { |
| "epoch": 0.6299053627760253, |
| "grad_norm": 0.44605725901650006, |
| "learning_rate": 3.629492286411349e-06, |
| "loss": 0.4515, |
| "step": 3120 |
| }, |
| { |
| "epoch": 0.6319242902208202, |
| "grad_norm": 0.4770842161429305, |
| "learning_rate": 3.5956406392898484e-06, |
| "loss": 0.4481, |
| "step": 3130 |
| }, |
| { |
| "epoch": 0.6339432176656151, |
| "grad_norm": 0.4611511904942963, |
| "learning_rate": 3.561858734437432e-06, |
| "loss": 0.4553, |
| "step": 3140 |
| }, |
| { |
| "epoch": 0.63596214511041, |
| "grad_norm": 0.4886490480648605, |
| "learning_rate": 3.528148249506391e-06, |
| "loss": 0.4541, |
| "step": 3150 |
| }, |
| { |
| "epoch": 0.6379810725552051, |
| "grad_norm": 0.5063011038652176, |
| "learning_rate": 3.494510858602217e-06, |
| "loss": 0.448, |
| "step": 3160 |
| }, |
| { |
| "epoch": 0.64, |
| "grad_norm": 0.5297665998286369, |
| "learning_rate": 3.4609482322004463e-06, |
| "loss": 0.45, |
| "step": 3170 |
| }, |
| { |
| "epoch": 0.6420189274447949, |
| "grad_norm": 0.789602090268308, |
| "learning_rate": 3.427462037063729e-06, |
| "loss": 0.4499, |
| "step": 3180 |
| }, |
| { |
| "epoch": 0.6440378548895899, |
| "grad_norm": 0.4595511607863831, |
| "learning_rate": 3.394053936159035e-06, |
| "loss": 0.4451, |
| "step": 3190 |
| }, |
| { |
| "epoch": 0.6460567823343849, |
| "grad_norm": 0.5053416927018038, |
| "learning_rate": 3.3607255885750756e-06, |
| "loss": 0.4488, |
| "step": 3200 |
| }, |
| { |
| "epoch": 0.6480757097791798, |
| "grad_norm": 0.47050871567377006, |
| "learning_rate": 3.3274786494399102e-06, |
| "loss": 0.4461, |
| "step": 3210 |
| }, |
| { |
| "epoch": 0.6500946372239748, |
| "grad_norm": 0.46433780403203007, |
| "learning_rate": 3.2943147698387562e-06, |
| "loss": 0.4497, |
| "step": 3220 |
| }, |
| { |
| "epoch": 0.6521135646687697, |
| "grad_norm": 0.45409078162209787, |
| "learning_rate": 3.261235596731982e-06, |
| "loss": 0.4457, |
| "step": 3230 |
| }, |
| { |
| "epoch": 0.6541324921135647, |
| "grad_norm": 0.4218245687222197, |
| "learning_rate": 3.228242772873328e-06, |
| "loss": 0.4437, |
| "step": 3240 |
| }, |
| { |
| "epoch": 0.6561514195583596, |
| "grad_norm": 0.47127385494827945, |
| "learning_rate": 3.195337936728323e-06, |
| "loss": 0.4488, |
| "step": 3250 |
| }, |
| { |
| "epoch": 0.6581703470031546, |
| "grad_norm": 0.43568607166498186, |
| "learning_rate": 3.1625227223929105e-06, |
| "loss": 0.4415, |
| "step": 3260 |
| }, |
| { |
| "epoch": 0.6601892744479495, |
| "grad_norm": 0.4200623699882891, |
| "learning_rate": 3.1297987595123024e-06, |
| "loss": 0.4423, |
| "step": 3270 |
| }, |
| { |
| "epoch": 0.6622082018927444, |
| "grad_norm": 0.45902924831849934, |
| "learning_rate": 3.0971676732000494e-06, |
| "loss": 0.4488, |
| "step": 3280 |
| }, |
| { |
| "epoch": 0.6642271293375395, |
| "grad_norm": 0.48610535709106045, |
| "learning_rate": 3.0646310839573302e-06, |
| "loss": 0.441, |
| "step": 3290 |
| }, |
| { |
| "epoch": 0.6662460567823344, |
| "grad_norm": 0.5237594196219926, |
| "learning_rate": 3.03219060759248e-06, |
| "loss": 0.454, |
| "step": 3300 |
| }, |
| { |
| "epoch": 0.6682649842271293, |
| "grad_norm": 0.4228282553579445, |
| "learning_rate": 2.9998478551407524e-06, |
| "loss": 0.4378, |
| "step": 3310 |
| }, |
| { |
| "epoch": 0.6702839116719242, |
| "grad_norm": 0.46926328208349977, |
| "learning_rate": 2.9676044327842963e-06, |
| "loss": 0.4475, |
| "step": 3320 |
| }, |
| { |
| "epoch": 0.6723028391167193, |
| "grad_norm": 0.47029450291352376, |
| "learning_rate": 2.935461941772412e-06, |
| "loss": 0.4418, |
| "step": 3330 |
| }, |
| { |
| "epoch": 0.6743217665615142, |
| "grad_norm": 0.461936788764074, |
| "learning_rate": 2.9034219783420183e-06, |
| "loss": 0.438, |
| "step": 3340 |
| }, |
| { |
| "epoch": 0.6763406940063091, |
| "grad_norm": 0.4879384251763818, |
| "learning_rate": 2.871486133638378e-06, |
| "loss": 0.4498, |
| "step": 3350 |
| }, |
| { |
| "epoch": 0.6783596214511041, |
| "grad_norm": 0.47390435742643133, |
| "learning_rate": 2.8396559936360934e-06, |
| "loss": 0.4444, |
| "step": 3360 |
| }, |
| { |
| "epoch": 0.6803785488958991, |
| "grad_norm": 0.43781719559966065, |
| "learning_rate": 2.807933139060337e-06, |
| "loss": 0.4481, |
| "step": 3370 |
| }, |
| { |
| "epoch": 0.682397476340694, |
| "grad_norm": 0.452166890113836, |
| "learning_rate": 2.776319145308345e-06, |
| "loss": 0.4397, |
| "step": 3380 |
| }, |
| { |
| "epoch": 0.684416403785489, |
| "grad_norm": 0.4480696892989974, |
| "learning_rate": 2.7448155823711897e-06, |
| "loss": 0.4415, |
| "step": 3390 |
| }, |
| { |
| "epoch": 0.6864353312302839, |
| "grad_norm": 0.482631983733039, |
| "learning_rate": 2.7134240147558124e-06, |
| "loss": 0.4494, |
| "step": 3400 |
| }, |
| { |
| "epoch": 0.6884542586750788, |
| "grad_norm": 0.4215351521880855, |
| "learning_rate": 2.6821460014073187e-06, |
| "loss": 0.4359, |
| "step": 3410 |
| }, |
| { |
| "epoch": 0.6904731861198738, |
| "grad_norm": 0.44170545621316426, |
| "learning_rate": 2.6509830956315684e-06, |
| "loss": 0.4516, |
| "step": 3420 |
| }, |
| { |
| "epoch": 0.6924921135646688, |
| "grad_norm": 0.43846533524679027, |
| "learning_rate": 2.619936845018031e-06, |
| "loss": 0.4357, |
| "step": 3430 |
| }, |
| { |
| "epoch": 0.6945110410094637, |
| "grad_norm": 0.5008934216933979, |
| "learning_rate": 2.5890087913629358e-06, |
| "loss": 0.4467, |
| "step": 3440 |
| }, |
| { |
| "epoch": 0.6965299684542586, |
| "grad_norm": 0.5250213600953301, |
| "learning_rate": 2.558200470592701e-06, |
| "loss": 0.449, |
| "step": 3450 |
| }, |
| { |
| "epoch": 0.6985488958990537, |
| "grad_norm": 0.5362972001626155, |
| "learning_rate": 2.527513412687654e-06, |
| "loss": 0.436, |
| "step": 3460 |
| }, |
| { |
| "epoch": 0.7005678233438486, |
| "grad_norm": 0.4447368579357635, |
| "learning_rate": 2.4969491416060542e-06, |
| "loss": 0.4393, |
| "step": 3470 |
| }, |
| { |
| "epoch": 0.7025867507886435, |
| "grad_norm": 0.42761346324330285, |
| "learning_rate": 2.4665091752084175e-06, |
| "loss": 0.4423, |
| "step": 3480 |
| }, |
| { |
| "epoch": 0.7046056782334384, |
| "grad_norm": 0.45180644732668995, |
| "learning_rate": 2.4361950251821253e-06, |
| "loss": 0.4351, |
| "step": 3490 |
| }, |
| { |
| "epoch": 0.7066246056782335, |
| "grad_norm": 0.4655597774869552, |
| "learning_rate": 2.406008196966359e-06, |
| "loss": 0.445, |
| "step": 3500 |
| }, |
| { |
| "epoch": 0.7086435331230284, |
| "grad_norm": 0.44607134328593223, |
| "learning_rate": 2.375950189677342e-06, |
| "loss": 0.4322, |
| "step": 3510 |
| }, |
| { |
| "epoch": 0.7106624605678233, |
| "grad_norm": 0.45886523824853853, |
| "learning_rate": 2.346022496033879e-06, |
| "loss": 0.439, |
| "step": 3520 |
| }, |
| { |
| "epoch": 0.7126813880126183, |
| "grad_norm": 0.439206301308698, |
| "learning_rate": 2.316226602283243e-06, |
| "loss": 0.445, |
| "step": 3530 |
| }, |
| { |
| "epoch": 0.7147003154574133, |
| "grad_norm": 0.42444173885080955, |
| "learning_rate": 2.2865639881273494e-06, |
| "loss": 0.4402, |
| "step": 3540 |
| }, |
| { |
| "epoch": 0.7167192429022082, |
| "grad_norm": 0.48281990057391616, |
| "learning_rate": 2.25703612664928e-06, |
| "loss": 0.445, |
| "step": 3550 |
| }, |
| { |
| "epoch": 0.7187381703470032, |
| "grad_norm": 0.46127829547820787, |
| "learning_rate": 2.227644484240135e-06, |
| "loss": 0.4356, |
| "step": 3560 |
| }, |
| { |
| "epoch": 0.7207570977917981, |
| "grad_norm": 0.4610423355631118, |
| "learning_rate": 2.198390520526194e-06, |
| "loss": 0.4438, |
| "step": 3570 |
| }, |
| { |
| "epoch": 0.722776025236593, |
| "grad_norm": 0.4711885745663891, |
| "learning_rate": 2.1692756882964404e-06, |
| "loss": 0.4464, |
| "step": 3580 |
| }, |
| { |
| "epoch": 0.724794952681388, |
| "grad_norm": 0.5420237747030388, |
| "learning_rate": 2.140301433430414e-06, |
| "loss": 0.4436, |
| "step": 3590 |
| }, |
| { |
| "epoch": 0.726813880126183, |
| "grad_norm": 0.47745493287406926, |
| "learning_rate": 2.1114691948264038e-06, |
| "loss": 0.4433, |
| "step": 3600 |
| }, |
| { |
| "epoch": 0.7288328075709779, |
| "grad_norm": 0.43683824200923127, |
| "learning_rate": 2.082780404329988e-06, |
| "loss": 0.4429, |
| "step": 3610 |
| }, |
| { |
| "epoch": 0.7308517350157728, |
| "grad_norm": 0.42141576682539833, |
| "learning_rate": 2.0542364866629292e-06, |
| "loss": 0.4402, |
| "step": 3620 |
| }, |
| { |
| "epoch": 0.7328706624605679, |
| "grad_norm": 0.4304526228372951, |
| "learning_rate": 2.025838859352428e-06, |
| "loss": 0.4462, |
| "step": 3630 |
| }, |
| { |
| "epoch": 0.7348895899053628, |
| "grad_norm": 0.4960972537231228, |
| "learning_rate": 1.997588932660713e-06, |
| "loss": 0.4298, |
| "step": 3640 |
| }, |
| { |
| "epoch": 0.7369085173501577, |
| "grad_norm": 0.5055917828329974, |
| "learning_rate": 1.9694881095150125e-06, |
| "loss": 0.4429, |
| "step": 3650 |
| }, |
| { |
| "epoch": 0.7389274447949526, |
| "grad_norm": 0.4756073151541865, |
| "learning_rate": 1.9415377854378903e-06, |
| "loss": 0.4423, |
| "step": 3660 |
| }, |
| { |
| "epoch": 0.7409463722397477, |
| "grad_norm": 0.4293238953341605, |
| "learning_rate": 1.913739348477927e-06, |
| "loss": 0.444, |
| "step": 3670 |
| }, |
| { |
| "epoch": 0.7429652996845426, |
| "grad_norm": 0.4476184732471472, |
| "learning_rate": 1.8860941791408038e-06, |
| "loss": 0.4432, |
| "step": 3680 |
| }, |
| { |
| "epoch": 0.7449842271293375, |
| "grad_norm": 0.4453341449071723, |
| "learning_rate": 1.8586036503207316e-06, |
| "loss": 0.4351, |
| "step": 3690 |
| }, |
| { |
| "epoch": 0.7470031545741325, |
| "grad_norm": 0.4741404781356218, |
| "learning_rate": 1.8312691272322768e-06, |
| "loss": 0.4445, |
| "step": 3700 |
| }, |
| { |
| "epoch": 0.7490220820189274, |
| "grad_norm": 0.4314595245056323, |
| "learning_rate": 1.804091967342569e-06, |
| "loss": 0.4392, |
| "step": 3710 |
| }, |
| { |
| "epoch": 0.7510410094637224, |
| "grad_norm": 0.4960471983973586, |
| "learning_rate": 1.777073520303874e-06, |
| "loss": 0.4406, |
| "step": 3720 |
| }, |
| { |
| "epoch": 0.7530599369085174, |
| "grad_norm": 0.47724657720097824, |
| "learning_rate": 1.7502151278865798e-06, |
| "loss": 0.4391, |
| "step": 3730 |
| }, |
| { |
| "epoch": 0.7550788643533123, |
| "grad_norm": 0.45566327509907684, |
| "learning_rate": 1.7235181239125587e-06, |
| "loss": 0.4411, |
| "step": 3740 |
| }, |
| { |
| "epoch": 0.7570977917981072, |
| "grad_norm": 0.47859031492685, |
| "learning_rate": 1.6969838341889288e-06, |
| "loss": 0.4412, |
| "step": 3750 |
| }, |
| { |
| "epoch": 0.7591167192429022, |
| "grad_norm": 0.4239970850626806, |
| "learning_rate": 1.67061357644221e-06, |
| "loss": 0.4327, |
| "step": 3760 |
| }, |
| { |
| "epoch": 0.7611356466876972, |
| "grad_norm": 0.41439508221708543, |
| "learning_rate": 1.6444086602528835e-06, |
| "loss": 0.4403, |
| "step": 3770 |
| }, |
| { |
| "epoch": 0.7631545741324921, |
| "grad_norm": 0.4724942386548484, |
| "learning_rate": 1.6183703869903666e-06, |
| "loss": 0.4399, |
| "step": 3780 |
| }, |
| { |
| "epoch": 0.765173501577287, |
| "grad_norm": 0.5068083200686946, |
| "learning_rate": 1.59250004974837e-06, |
| "loss": 0.4484, |
| "step": 3790 |
| }, |
| { |
| "epoch": 0.7671924290220821, |
| "grad_norm": 0.47623913375530136, |
| "learning_rate": 1.5667989332806894e-06, |
| "loss": 0.442, |
| "step": 3800 |
| }, |
| { |
| "epoch": 0.769211356466877, |
| "grad_norm": 0.5448383844625657, |
| "learning_rate": 1.5412683139374068e-06, |
| "loss": 0.4339, |
| "step": 3810 |
| }, |
| { |
| "epoch": 0.7712302839116719, |
| "grad_norm": 0.4819612793724808, |
| "learning_rate": 1.5159094596014934e-06, |
| "loss": 0.4478, |
| "step": 3820 |
| }, |
| { |
| "epoch": 0.7732492113564668, |
| "grad_norm": 0.4456007417706316, |
| "learning_rate": 1.4907236296258598e-06, |
| "loss": 0.4423, |
| "step": 3830 |
| }, |
| { |
| "epoch": 0.7752681388012618, |
| "grad_norm": 0.43979298393219013, |
| "learning_rate": 1.4657120747708014e-06, |
| "loss": 0.4417, |
| "step": 3840 |
| }, |
| { |
| "epoch": 0.7772870662460568, |
| "grad_norm": 0.5066920888924865, |
| "learning_rate": 1.4408760371418895e-06, |
| "loss": 0.4397, |
| "step": 3850 |
| }, |
| { |
| "epoch": 0.7793059936908517, |
| "grad_norm": 0.47046124228822955, |
| "learning_rate": 1.4162167501282925e-06, |
| "loss": 0.4416, |
| "step": 3860 |
| }, |
| { |
| "epoch": 0.7813249211356467, |
| "grad_norm": 0.4558898604297212, |
| "learning_rate": 1.391735438341514e-06, |
| "loss": 0.437, |
| "step": 3870 |
| }, |
| { |
| "epoch": 0.7833438485804416, |
| "grad_norm": 0.4635342780752439, |
| "learning_rate": 1.3674333175545818e-06, |
| "loss": 0.4425, |
| "step": 3880 |
| }, |
| { |
| "epoch": 0.7853627760252366, |
| "grad_norm": 0.450134125379297, |
| "learning_rate": 1.343311594641674e-06, |
| "loss": 0.4371, |
| "step": 3890 |
| }, |
| { |
| "epoch": 0.7873817034700316, |
| "grad_norm": 0.48501901287598154, |
| "learning_rate": 1.3193714675181813e-06, |
| "loss": 0.4428, |
| "step": 3900 |
| }, |
| { |
| "epoch": 0.7894006309148265, |
| "grad_norm": 0.41955405495588977, |
| "learning_rate": 1.2956141250812144e-06, |
| "loss": 0.4307, |
| "step": 3910 |
| }, |
| { |
| "epoch": 0.7914195583596214, |
| "grad_norm": 0.41659445986128424, |
| "learning_rate": 1.272040747150563e-06, |
| "loss": 0.447, |
| "step": 3920 |
| }, |
| { |
| "epoch": 0.7934384858044164, |
| "grad_norm": 0.44707673152149113, |
| "learning_rate": 1.2486525044101123e-06, |
| "loss": 0.4333, |
| "step": 3930 |
| }, |
| { |
| "epoch": 0.7954574132492114, |
| "grad_norm": 0.4698766268506151, |
| "learning_rate": 1.2254505583496922e-06, |
| "loss": 0.4403, |
| "step": 3940 |
| }, |
| { |
| "epoch": 0.7974763406940063, |
| "grad_norm": 0.47512471796296707, |
| "learning_rate": 1.2024360612074055e-06, |
| "loss": 0.4454, |
| "step": 3950 |
| }, |
| { |
| "epoch": 0.7994952681388012, |
| "grad_norm": 0.4585910400457452, |
| "learning_rate": 1.179610155912405e-06, |
| "loss": 0.436, |
| "step": 3960 |
| }, |
| { |
| "epoch": 0.8015141955835963, |
| "grad_norm": 0.4557389746705491, |
| "learning_rate": 1.1569739760281295e-06, |
| "loss": 0.4393, |
| "step": 3970 |
| }, |
| { |
| "epoch": 0.8035331230283912, |
| "grad_norm": 0.4553106102818403, |
| "learning_rate": 1.1345286456960186e-06, |
| "loss": 0.4439, |
| "step": 3980 |
| }, |
| { |
| "epoch": 0.8055520504731861, |
| "grad_norm": 0.4443887135897382, |
| "learning_rate": 1.1122752795796748e-06, |
| "loss": 0.4413, |
| "step": 3990 |
| }, |
| { |
| "epoch": 0.807570977917981, |
| "grad_norm": 0.47272003456758904, |
| "learning_rate": 1.0902149828095154e-06, |
| "loss": 0.4422, |
| "step": 4000 |
| }, |
| { |
| "epoch": 0.809589905362776, |
| "grad_norm": 0.43703108025040205, |
| "learning_rate": 1.0683488509278928e-06, |
| "loss": 0.4386, |
| "step": 4010 |
| }, |
| { |
| "epoch": 0.811608832807571, |
| "grad_norm": 0.4441202945224273, |
| "learning_rate": 1.0466779698346797e-06, |
| "loss": 0.4363, |
| "step": 4020 |
| }, |
| { |
| "epoch": 0.8136277602523659, |
| "grad_norm": 0.42622028909468485, |
| "learning_rate": 1.025203415733348e-06, |
| "loss": 0.4365, |
| "step": 4030 |
| }, |
| { |
| "epoch": 0.8156466876971609, |
| "grad_norm": 0.440821020950604, |
| "learning_rate": 1.0039262550775235e-06, |
| "loss": 0.4345, |
| "step": 4040 |
| }, |
| { |
| "epoch": 0.8176656151419558, |
| "grad_norm": 0.4670601139166515, |
| "learning_rate": 9.828475445180212e-07, |
| "loss": 0.4404, |
| "step": 4050 |
| }, |
| { |
| "epoch": 0.8196845425867508, |
| "grad_norm": 0.4363041969561276, |
| "learning_rate": 9.61968330850374e-07, |
| "loss": 0.4369, |
| "step": 4060 |
| }, |
| { |
| "epoch": 0.8217034700315458, |
| "grad_norm": 0.4866069797752605, |
| "learning_rate": 9.412896509628439e-07, |
| "loss": 0.4434, |
| "step": 4070 |
| }, |
| { |
| "epoch": 0.8237223974763407, |
| "grad_norm": 0.4360948594973755, |
| "learning_rate": 9.208125317849304e-07, |
| "loss": 0.4395, |
| "step": 4080 |
| }, |
| { |
| "epoch": 0.8257413249211356, |
| "grad_norm": 0.4490621547393745, |
| "learning_rate": 9.005379902363753e-07, |
| "loss": 0.4405, |
| "step": 4090 |
| }, |
| { |
| "epoch": 0.8277602523659306, |
| "grad_norm": 0.4809678412101839, |
| "learning_rate": 8.804670331766557e-07, |
| "loss": 0.4449, |
| "step": 4100 |
| }, |
| { |
| "epoch": 0.8297791798107256, |
| "grad_norm": 0.4626290445825763, |
| "learning_rate": 8.606006573549836e-07, |
| "loss": 0.4359, |
| "step": 4110 |
| }, |
| { |
| "epoch": 0.8317981072555205, |
| "grad_norm": 0.5187638972410576, |
| "learning_rate": 8.409398493608089e-07, |
| "loss": 0.4413, |
| "step": 4120 |
| }, |
| { |
| "epoch": 0.8338170347003154, |
| "grad_norm": 0.46587792642977804, |
| "learning_rate": 8.214855855748249e-07, |
| "loss": 0.4409, |
| "step": 4130 |
| }, |
| { |
| "epoch": 0.8358359621451104, |
| "grad_norm": 0.4706901825283824, |
| "learning_rate": 8.022388321204688e-07, |
| "loss": 0.4354, |
| "step": 4140 |
| }, |
| { |
| "epoch": 0.8378548895899054, |
| "grad_norm": 0.562791948402191, |
| "learning_rate": 7.832005448159585e-07, |
| "loss": 0.4445, |
| "step": 4150 |
| }, |
| { |
| "epoch": 0.8398738170347003, |
| "grad_norm": 0.4543738469403552, |
| "learning_rate": 7.643716691268171e-07, |
| "loss": 0.4374, |
| "step": 4160 |
| }, |
| { |
| "epoch": 0.8418927444794952, |
| "grad_norm": 1.6174532184269235, |
| "learning_rate": 7.457531401189188e-07, |
| "loss": 0.4408, |
| "step": 4170 |
| }, |
| { |
| "epoch": 0.8439116719242902, |
| "grad_norm": 0.46922997370194347, |
| "learning_rate": 7.273458824120555e-07, |
| "loss": 0.4376, |
| "step": 4180 |
| }, |
| { |
| "epoch": 0.8459305993690852, |
| "grad_norm": 0.44210687899012985, |
| "learning_rate": 7.091508101340189e-07, |
| "loss": 0.4407, |
| "step": 4190 |
| }, |
| { |
| "epoch": 0.8479495268138801, |
| "grad_norm": 0.488809217492704, |
| "learning_rate": 6.91168826875201e-07, |
| "loss": 0.4489, |
| "step": 4200 |
| }, |
| { |
| "epoch": 0.8499684542586751, |
| "grad_norm": 0.4335482219101724, |
| "learning_rate": 6.734008256437253e-07, |
| "loss": 0.435, |
| "step": 4210 |
| }, |
| { |
| "epoch": 0.85198738170347, |
| "grad_norm": 0.4549494452467777, |
| "learning_rate": 6.558476888210935e-07, |
| "loss": 0.4375, |
| "step": 4220 |
| }, |
| { |
| "epoch": 0.854006309148265, |
| "grad_norm": 0.44896647048001603, |
| "learning_rate": 6.38510288118368e-07, |
| "loss": 0.4409, |
| "step": 4230 |
| }, |
| { |
| "epoch": 0.85602523659306, |
| "grad_norm": 0.43685823117883865, |
| "learning_rate": 6.213894845328844e-07, |
| "loss": 0.4418, |
| "step": 4240 |
| }, |
| { |
| "epoch": 0.8580441640378549, |
| "grad_norm": 0.47792243578520655, |
| "learning_rate": 6.044861283054876e-07, |
| "loss": 0.4452, |
| "step": 4250 |
| }, |
| { |
| "epoch": 0.8600630914826498, |
| "grad_norm": 0.43885235740502504, |
| "learning_rate": 5.878010588783101e-07, |
| "loss": 0.4423, |
| "step": 4260 |
| }, |
| { |
| "epoch": 0.8620820189274448, |
| "grad_norm": 0.5121668333108534, |
| "learning_rate": 5.713351048530874e-07, |
| "loss": 0.4429, |
| "step": 4270 |
| }, |
| { |
| "epoch": 0.8641009463722398, |
| "grad_norm": 0.44403508398564756, |
| "learning_rate": 5.550890839500067e-07, |
| "loss": 0.4511, |
| "step": 4280 |
| }, |
| { |
| "epoch": 0.8661198738170347, |
| "grad_norm": 0.42769598938785636, |
| "learning_rate": 5.390638029670908e-07, |
| "loss": 0.4322, |
| "step": 4290 |
| }, |
| { |
| "epoch": 0.8681388012618296, |
| "grad_norm": 0.4949663561795469, |
| "learning_rate": 5.232600577401436e-07, |
| "loss": 0.4457, |
| "step": 4300 |
| }, |
| { |
| "epoch": 0.8701577287066246, |
| "grad_norm": 0.4489314407451146, |
| "learning_rate": 5.076786331032224e-07, |
| "loss": 0.4327, |
| "step": 4310 |
| }, |
| { |
| "epoch": 0.8721766561514196, |
| "grad_norm": 0.48025463446625855, |
| "learning_rate": 4.923203028496576e-07, |
| "loss": 0.4447, |
| "step": 4320 |
| }, |
| { |
| "epoch": 0.8741955835962145, |
| "grad_norm": 0.5743773868003562, |
| "learning_rate": 4.771858296936311e-07, |
| "loss": 0.4402, |
| "step": 4330 |
| }, |
| { |
| "epoch": 0.8762145110410094, |
| "grad_norm": 0.4329672101913256, |
| "learning_rate": 4.622759652322989e-07, |
| "loss": 0.4455, |
| "step": 4340 |
| }, |
| { |
| "epoch": 0.8782334384858044, |
| "grad_norm": 0.49048963795844497, |
| "learning_rate": 4.475914499084605e-07, |
| "loss": 0.4467, |
| "step": 4350 |
| }, |
| { |
| "epoch": 0.8802523659305994, |
| "grad_norm": 0.4829611101804599, |
| "learning_rate": 4.331330129737954e-07, |
| "loss": 0.4287, |
| "step": 4360 |
| }, |
| { |
| "epoch": 0.8822712933753943, |
| "grad_norm": 0.44140910883095635, |
| "learning_rate": 4.189013724526392e-07, |
| "loss": 0.4446, |
| "step": 4370 |
| }, |
| { |
| "epoch": 0.8842902208201893, |
| "grad_norm": 0.45272062627566334, |
| "learning_rate": 4.048972351063313e-07, |
| "loss": 0.4354, |
| "step": 4380 |
| }, |
| { |
| "epoch": 0.8863091482649842, |
| "grad_norm": 0.7049959167129851, |
| "learning_rate": 3.9112129639811535e-07, |
| "loss": 0.4346, |
| "step": 4390 |
| }, |
| { |
| "epoch": 0.8883280757097792, |
| "grad_norm": 0.4811238411002169, |
| "learning_rate": 3.7757424045860025e-07, |
| "loss": 0.4372, |
| "step": 4400 |
| }, |
| { |
| "epoch": 0.8903470031545742, |
| "grad_norm": 0.4540483251684103, |
| "learning_rate": 3.642567400517838e-07, |
| "loss": 0.4314, |
| "step": 4410 |
| }, |
| { |
| "epoch": 0.8923659305993691, |
| "grad_norm": 0.47603601308214405, |
| "learning_rate": 3.511694565416485e-07, |
| "loss": 0.4382, |
| "step": 4420 |
| }, |
| { |
| "epoch": 0.894384858044164, |
| "grad_norm": 0.4286142580967842, |
| "learning_rate": 3.383130398593121e-07, |
| "loss": 0.4399, |
| "step": 4430 |
| }, |
| { |
| "epoch": 0.8964037854889589, |
| "grad_norm": 0.4337017987281163, |
| "learning_rate": 3.2568812847074913e-07, |
| "loss": 0.4437, |
| "step": 4440 |
| }, |
| { |
| "epoch": 0.898422712933754, |
| "grad_norm": 0.48386529314250487, |
| "learning_rate": 3.1329534934509186e-07, |
| "loss": 0.4415, |
| "step": 4450 |
| }, |
| { |
| "epoch": 0.9004416403785489, |
| "grad_norm": 0.46990087367597777, |
| "learning_rate": 3.011353179234888e-07, |
| "loss": 0.4317, |
| "step": 4460 |
| }, |
| { |
| "epoch": 0.9024605678233438, |
| "grad_norm": 0.42506034975479673, |
| "learning_rate": 2.8920863808853947e-07, |
| "loss": 0.444, |
| "step": 4470 |
| }, |
| { |
| "epoch": 0.9044794952681388, |
| "grad_norm": 0.6216117310744336, |
| "learning_rate": 2.775159021343082e-07, |
| "loss": 0.4367, |
| "step": 4480 |
| }, |
| { |
| "epoch": 0.9064984227129338, |
| "grad_norm": 0.4583995655944248, |
| "learning_rate": 2.660576907369111e-07, |
| "loss": 0.4412, |
| "step": 4490 |
| }, |
| { |
| "epoch": 0.9085173501577287, |
| "grad_norm": 0.4889160958013681, |
| "learning_rate": 2.5483457292567517e-07, |
| "loss": 0.4403, |
| "step": 4500 |
| }, |
| { |
| "epoch": 0.9105362776025236, |
| "grad_norm": 0.4267838945833425, |
| "learning_rate": 2.4384710605488195e-07, |
| "loss": 0.4402, |
| "step": 4510 |
| }, |
| { |
| "epoch": 0.9125552050473186, |
| "grad_norm": 0.4986776395611631, |
| "learning_rate": 2.330958357760882e-07, |
| "loss": 0.4399, |
| "step": 4520 |
| }, |
| { |
| "epoch": 0.9145741324921136, |
| "grad_norm": 0.459784020601065, |
| "learning_rate": 2.225812960110263e-07, |
| "loss": 0.4316, |
| "step": 4530 |
| }, |
| { |
| "epoch": 0.9165930599369085, |
| "grad_norm": 0.44426443349474987, |
| "learning_rate": 2.123040089250944e-07, |
| "loss": 0.4402, |
| "step": 4540 |
| }, |
| { |
| "epoch": 0.9186119873817035, |
| "grad_norm": 0.49235316049109906, |
| "learning_rate": 2.0226448490141715e-07, |
| "loss": 0.4481, |
| "step": 4550 |
| }, |
| { |
| "epoch": 0.9206309148264984, |
| "grad_norm": 0.43058790479448, |
| "learning_rate": 1.9246322251550653e-07, |
| "loss": 0.4317, |
| "step": 4560 |
| }, |
| { |
| "epoch": 0.9226498422712934, |
| "grad_norm": 0.4966838385984267, |
| "learning_rate": 1.829007085104978e-07, |
| "loss": 0.4427, |
| "step": 4570 |
| }, |
| { |
| "epoch": 0.9246687697160884, |
| "grad_norm": 0.4666275183152453, |
| "learning_rate": 1.7357741777297987e-07, |
| "loss": 0.4485, |
| "step": 4580 |
| }, |
| { |
| "epoch": 0.9266876971608833, |
| "grad_norm": 0.4412830191687912, |
| "learning_rate": 1.644938133094076e-07, |
| "loss": 0.4312, |
| "step": 4590 |
| }, |
| { |
| "epoch": 0.9287066246056782, |
| "grad_norm": 0.4886548137136921, |
| "learning_rate": 1.5565034622311349e-07, |
| "loss": 0.4461, |
| "step": 4600 |
| }, |
| { |
| "epoch": 0.9307255520504731, |
| "grad_norm": 0.4652701957357695, |
| "learning_rate": 1.4704745569190327e-07, |
| "loss": 0.4375, |
| "step": 4610 |
| }, |
| { |
| "epoch": 0.9327444794952682, |
| "grad_norm": 0.5298674695548953, |
| "learning_rate": 1.3868556894624353e-07, |
| "loss": 0.4399, |
| "step": 4620 |
| }, |
| { |
| "epoch": 0.9347634069400631, |
| "grad_norm": 0.4562802443840814, |
| "learning_rate": 1.30565101248048e-07, |
| "loss": 0.4385, |
| "step": 4630 |
| }, |
| { |
| "epoch": 0.936782334384858, |
| "grad_norm": 0.4578530629768795, |
| "learning_rate": 1.226864558700541e-07, |
| "loss": 0.4416, |
| "step": 4640 |
| }, |
| { |
| "epoch": 0.938801261829653, |
| "grad_norm": 0.4951660779551572, |
| "learning_rate": 1.1505002407579512e-07, |
| "loss": 0.4424, |
| "step": 4650 |
| }, |
| { |
| "epoch": 0.940820189274448, |
| "grad_norm": 0.44757841438265406, |
| "learning_rate": 1.0765618510017118e-07, |
| "loss": 0.4368, |
| "step": 4660 |
| }, |
| { |
| "epoch": 0.9428391167192429, |
| "grad_norm": 1.5035499500846865, |
| "learning_rate": 1.0050530613061327e-07, |
| "loss": 0.4353, |
| "step": 4670 |
| }, |
| { |
| "epoch": 0.9448580441640378, |
| "grad_norm": 0.4938389329390479, |
| "learning_rate": 9.359774228885177e-08, |
| "loss": 0.4391, |
| "step": 4680 |
| }, |
| { |
| "epoch": 0.9468769716088328, |
| "grad_norm": 0.4711850612472287, |
| "learning_rate": 8.693383661327882e-08, |
| "loss": 0.4388, |
| "step": 4690 |
| }, |
| { |
| "epoch": 0.9488958990536278, |
| "grad_norm": 0.48652374519922575, |
| "learning_rate": 8.051392004191139e-08, |
| "loss": 0.4486, |
| "step": 4700 |
| }, |
| { |
| "epoch": 0.9509148264984227, |
| "grad_norm": 0.4473704324045684, |
| "learning_rate": 7.433831139595727e-08, |
| "loss": 0.4385, |
| "step": 4710 |
| }, |
| { |
| "epoch": 0.9529337539432177, |
| "grad_norm": 0.509944701555344, |
| "learning_rate": 6.840731736398542e-08, |
| "loss": 0.4362, |
| "step": 4720 |
| }, |
| { |
| "epoch": 0.9549526813880126, |
| "grad_norm": 0.4468123798191573, |
| "learning_rate": 6.272123248668993e-08, |
| "loss": 0.4356, |
| "step": 4730 |
| }, |
| { |
| "epoch": 0.9569716088328075, |
| "grad_norm": 0.4432856047100861, |
| "learning_rate": 5.728033914226494e-08, |
| "loss": 0.4406, |
| "step": 4740 |
| }, |
| { |
| "epoch": 0.9589905362776026, |
| "grad_norm": 0.8312641118207356, |
| "learning_rate": 5.2084907532383156e-08, |
| "loss": 0.4457, |
| "step": 4750 |
| }, |
| { |
| "epoch": 0.9610094637223975, |
| "grad_norm": 0.4308428180533281, |
| "learning_rate": 4.7135195668776e-08, |
| "loss": 0.4337, |
| "step": 4760 |
| }, |
| { |
| "epoch": 0.9630283911671924, |
| "grad_norm": 0.47706748470749133, |
| "learning_rate": 4.243144936041943e-08, |
| "loss": 0.4413, |
| "step": 4770 |
| }, |
| { |
| "epoch": 0.9650473186119873, |
| "grad_norm": 0.46426542825118344, |
| "learning_rate": 3.797390220132757e-08, |
| "loss": 0.4325, |
| "step": 4780 |
| }, |
| { |
| "epoch": 0.9670662460567824, |
| "grad_norm": 0.4324982289321672, |
| "learning_rate": 3.3762775558952576e-08, |
| "loss": 0.4377, |
| "step": 4790 |
| }, |
| { |
| "epoch": 0.9690851735015773, |
| "grad_norm": 0.47719223462129745, |
| "learning_rate": 2.9798278563190618e-08, |
| "loss": 0.4439, |
| "step": 4800 |
| }, |
| { |
| "epoch": 0.9711041009463722, |
| "grad_norm": 0.4540530080947163, |
| "learning_rate": 2.6080608095997438e-08, |
| "loss": 0.4377, |
| "step": 4810 |
| }, |
| { |
| "epoch": 0.9731230283911672, |
| "grad_norm": 0.4413909184803219, |
| "learning_rate": 2.2609948781607827e-08, |
| "loss": 0.4424, |
| "step": 4820 |
| }, |
| { |
| "epoch": 0.9751419558359622, |
| "grad_norm": 0.4712177319951425, |
| "learning_rate": 1.9386472977370728e-08, |
| "loss": 0.4412, |
| "step": 4830 |
| }, |
| { |
| "epoch": 0.9771608832807571, |
| "grad_norm": 0.443140708468747, |
| "learning_rate": 1.6410340765187772e-08, |
| "loss": 0.4382, |
| "step": 4840 |
| }, |
| { |
| "epoch": 0.979179810725552, |
| "grad_norm": 0.4782858905945508, |
| "learning_rate": 1.3681699943564053e-08, |
| "loss": 0.4387, |
| "step": 4850 |
| }, |
| { |
| "epoch": 0.981198738170347, |
| "grad_norm": 0.6206662222549734, |
| "learning_rate": 1.1200686020266805e-08, |
| "loss": 0.4431, |
| "step": 4860 |
| }, |
| { |
| "epoch": 0.983217665615142, |
| "grad_norm": 2.8460570799996985, |
| "learning_rate": 8.967422205598541e-09, |
| "loss": 0.4393, |
| "step": 4870 |
| }, |
| { |
| "epoch": 0.9852365930599369, |
| "grad_norm": 0.4435087118229933, |
| "learning_rate": 6.982019406275853e-09, |
| "loss": 0.436, |
| "step": 4880 |
| }, |
| { |
| "epoch": 0.9872555205047319, |
| "grad_norm": 0.4300508595922509, |
| "learning_rate": 5.244576219922693e-09, |
| "loss": 0.4316, |
| "step": 4890 |
| }, |
| { |
| "epoch": 0.9892744479495268, |
| "grad_norm": 0.49712300151356087, |
| "learning_rate": 3.755178930173742e-09, |
| "loss": 0.4463, |
| "step": 4900 |
| }, |
| { |
| "epoch": 0.9912933753943217, |
| "grad_norm": 0.4798533947795922, |
| "learning_rate": 2.513901502389504e-09, |
| "loss": 0.4327, |
| "step": 4910 |
| }, |
| { |
| "epoch": 0.9933123028391168, |
| "grad_norm": 0.4936501161499046, |
| "learning_rate": 1.520805579983131e-09, |
| "loss": 0.4403, |
| "step": 4920 |
| }, |
| { |
| "epoch": 0.9953312302839117, |
| "grad_norm": 0.4638813307919199, |
| "learning_rate": 7.759404813600935e-10, |
| "loss": 0.4385, |
| "step": 4930 |
| }, |
| { |
| "epoch": 0.9973501577287066, |
| "grad_norm": 0.43789683220104403, |
| "learning_rate": 2.7934319746625306e-10, |
| "loss": 0.4387, |
| "step": 4940 |
| }, |
| { |
| "epoch": 0.9993690851735015, |
| "grad_norm": 0.4980004114526748, |
| "learning_rate": 3.103838995321962e-11, |
| "loss": 0.4391, |
| "step": 4950 |
| }, |
| { |
| "epoch": 1.0, |
| "step": 4954, |
| "total_flos": 1.1092292323919266e+19, |
| "train_loss": 0.46906891030845593, |
| "train_runtime": 81648.0087, |
| "train_samples_per_second": 7.765, |
| "train_steps_per_second": 0.061 |
| } |
| ], |
| "logging_steps": 10, |
| "max_steps": 4954, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 1, |
| "save_steps": 50000, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 1.1092292323919266e+19, |
| "train_batch_size": 1, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|