| { |
| "best_global_step": null, |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 4.402861860209136, |
| "eval_steps": 500, |
| "global_step": 40000, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.00550357732526142, |
| "grad_norm": 5.828019142150879, |
| "learning_rate": 6.47007042253521e-06, |
| "loss": 8.235, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.01100715465052284, |
| "grad_norm": 15.358248710632324, |
| "learning_rate": 1.3072183098591547e-05, |
| "loss": 6.0857, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.01651073197578426, |
| "grad_norm": 7.99273681640625, |
| "learning_rate": 1.9674295774647885e-05, |
| "loss": 4.4315, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.02201430930104568, |
| "grad_norm": 2.8550803661346436, |
| "learning_rate": 2.6276408450704222e-05, |
| "loss": 3.681, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.0275178866263071, |
| "grad_norm": 2.3769114017486572, |
| "learning_rate": 3.2878521126760565e-05, |
| "loss": 3.4729, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.03302146395156852, |
| "grad_norm": 5.98811149597168, |
| "learning_rate": 3.94806338028169e-05, |
| "loss": 3.3829, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.03852504127682994, |
| "grad_norm": 3.616163492202759, |
| "learning_rate": 4.608274647887324e-05, |
| "loss": 3.4119, |
| "step": 350 |
| }, |
| { |
| "epoch": 0.04402861860209136, |
| "grad_norm": 2.1746344566345215, |
| "learning_rate": 5.268485915492957e-05, |
| "loss": 3.3955, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.04953219592735278, |
| "grad_norm": 1.3189276456832886, |
| "learning_rate": 5.928697183098591e-05, |
| "loss": 3.3177, |
| "step": 450 |
| }, |
| { |
| "epoch": 0.0550357732526142, |
| "grad_norm": 1.2459770441055298, |
| "learning_rate": 6.588908450704225e-05, |
| "loss": 3.3095, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.06053935057787562, |
| "grad_norm": 1.2855902910232544, |
| "learning_rate": 7.249119718309858e-05, |
| "loss": 3.2437, |
| "step": 550 |
| }, |
| { |
| "epoch": 0.06604292790313704, |
| "grad_norm": 1.2871235609054565, |
| "learning_rate": 7.909330985915493e-05, |
| "loss": 3.2069, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.07154650522839846, |
| "grad_norm": 1.0846847295761108, |
| "learning_rate": 8.569542253521127e-05, |
| "loss": 3.2512, |
| "step": 650 |
| }, |
| { |
| "epoch": 0.07705008255365987, |
| "grad_norm": 1.5505499839782715, |
| "learning_rate": 9.22975352112676e-05, |
| "loss": 3.1704, |
| "step": 700 |
| }, |
| { |
| "epoch": 0.0825536598789213, |
| "grad_norm": 1.178614616394043, |
| "learning_rate": 9.889964788732394e-05, |
| "loss": 3.1747, |
| "step": 750 |
| }, |
| { |
| "epoch": 0.08805723720418272, |
| "grad_norm": 1.1504727602005005, |
| "learning_rate": 0.00010550176056338028, |
| "loss": 3.1241, |
| "step": 800 |
| }, |
| { |
| "epoch": 0.09356081452944413, |
| "grad_norm": 1.02865469455719, |
| "learning_rate": 0.00011210387323943662, |
| "loss": 3.0481, |
| "step": 850 |
| }, |
| { |
| "epoch": 0.09906439185470556, |
| "grad_norm": 1.1366077661514282, |
| "learning_rate": 0.00011870598591549295, |
| "loss": 3.0201, |
| "step": 900 |
| }, |
| { |
| "epoch": 0.10456796917996698, |
| "grad_norm": 0.9753648042678833, |
| "learning_rate": 0.00012530809859154929, |
| "loss": 3.0145, |
| "step": 950 |
| }, |
| { |
| "epoch": 0.1100715465052284, |
| "grad_norm": 0.6859256625175476, |
| "learning_rate": 0.00013191021126760563, |
| "loss": 2.9671, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.11557512383048982, |
| "grad_norm": 0.8368203043937683, |
| "learning_rate": 0.00013851232394366197, |
| "loss": 2.992, |
| "step": 1050 |
| }, |
| { |
| "epoch": 0.12107870115575124, |
| "grad_norm": 1.1109174489974976, |
| "learning_rate": 0.0001451144366197183, |
| "loss": 2.967, |
| "step": 1100 |
| }, |
| { |
| "epoch": 0.12658227848101267, |
| "grad_norm": 0.6072912812232971, |
| "learning_rate": 0.00015171654929577465, |
| "loss": 2.9501, |
| "step": 1150 |
| }, |
| { |
| "epoch": 0.13208585580627408, |
| "grad_norm": 0.7659889459609985, |
| "learning_rate": 0.00015831866197183099, |
| "loss": 2.9026, |
| "step": 1200 |
| }, |
| { |
| "epoch": 0.1375894331315355, |
| "grad_norm": 0.5841110348701477, |
| "learning_rate": 0.0001649207746478873, |
| "loss": 2.8972, |
| "step": 1250 |
| }, |
| { |
| "epoch": 0.14309301045679693, |
| "grad_norm": 0.5997458100318909, |
| "learning_rate": 0.00017152288732394364, |
| "loss": 2.9189, |
| "step": 1300 |
| }, |
| { |
| "epoch": 0.14859658778205834, |
| "grad_norm": 0.6082264184951782, |
| "learning_rate": 0.00017812499999999998, |
| "loss": 2.8877, |
| "step": 1350 |
| }, |
| { |
| "epoch": 0.15410016510731975, |
| "grad_norm": 0.696685254573822, |
| "learning_rate": 0.00018472711267605632, |
| "loss": 2.8752, |
| "step": 1400 |
| }, |
| { |
| "epoch": 0.15960374243258119, |
| "grad_norm": 0.6795832514762878, |
| "learning_rate": 0.00019132922535211266, |
| "loss": 2.851, |
| "step": 1450 |
| }, |
| { |
| "epoch": 0.1651073197578426, |
| "grad_norm": 0.5588585734367371, |
| "learning_rate": 0.000197931338028169, |
| "loss": 2.8546, |
| "step": 1500 |
| }, |
| { |
| "epoch": 0.170610897083104, |
| "grad_norm": 0.5267760157585144, |
| "learning_rate": 0.00020453345070422534, |
| "loss": 2.8049, |
| "step": 1550 |
| }, |
| { |
| "epoch": 0.17611447440836545, |
| "grad_norm": 0.591826856136322, |
| "learning_rate": 0.00021113556338028168, |
| "loss": 2.8071, |
| "step": 1600 |
| }, |
| { |
| "epoch": 0.18161805173362686, |
| "grad_norm": 0.5463298559188843, |
| "learning_rate": 0.00021773767605633802, |
| "loss": 2.7667, |
| "step": 1650 |
| }, |
| { |
| "epoch": 0.18712162905888827, |
| "grad_norm": 0.5745858550071716, |
| "learning_rate": 0.00022433978873239433, |
| "loss": 2.7968, |
| "step": 1700 |
| }, |
| { |
| "epoch": 0.1926252063841497, |
| "grad_norm": 0.6239858865737915, |
| "learning_rate": 0.00023094190140845067, |
| "loss": 2.8206, |
| "step": 1750 |
| }, |
| { |
| "epoch": 0.19812878370941112, |
| "grad_norm": 0.46496015787124634, |
| "learning_rate": 0.000237544014084507, |
| "loss": 2.7944, |
| "step": 1800 |
| }, |
| { |
| "epoch": 0.20363236103467253, |
| "grad_norm": 0.5664075016975403, |
| "learning_rate": 0.00024414612676056335, |
| "loss": 2.7837, |
| "step": 1850 |
| }, |
| { |
| "epoch": 0.20913593835993396, |
| "grad_norm": 0.537627100944519, |
| "learning_rate": 0.0002507482394366197, |
| "loss": 2.7179, |
| "step": 1900 |
| }, |
| { |
| "epoch": 0.21463951568519538, |
| "grad_norm": 0.544585645198822, |
| "learning_rate": 0.00025735035211267603, |
| "loss": 2.7552, |
| "step": 1950 |
| }, |
| { |
| "epoch": 0.2201430930104568, |
| "grad_norm": 0.5067969560623169, |
| "learning_rate": 0.0002639524647887324, |
| "loss": 2.7309, |
| "step": 2000 |
| }, |
| { |
| "epoch": 0.22564667033571822, |
| "grad_norm": 0.462003618478775, |
| "learning_rate": 0.0002705545774647887, |
| "loss": 2.716, |
| "step": 2050 |
| }, |
| { |
| "epoch": 0.23115024766097964, |
| "grad_norm": 0.44174736738204956, |
| "learning_rate": 0.00027715669014084505, |
| "loss": 2.7319, |
| "step": 2100 |
| }, |
| { |
| "epoch": 0.23665382498624105, |
| "grad_norm": 0.5010894536972046, |
| "learning_rate": 0.0002837588028169014, |
| "loss": 2.7199, |
| "step": 2150 |
| }, |
| { |
| "epoch": 0.24215740231150248, |
| "grad_norm": 0.44113022089004517, |
| "learning_rate": 0.00029036091549295773, |
| "loss": 2.7146, |
| "step": 2200 |
| }, |
| { |
| "epoch": 0.2476609796367639, |
| "grad_norm": 0.4972345232963562, |
| "learning_rate": 0.0002969630281690141, |
| "loss": 2.6471, |
| "step": 2250 |
| }, |
| { |
| "epoch": 0.25316455696202533, |
| "grad_norm": 0.5614003539085388, |
| "learning_rate": 0.0002999997102213327, |
| "loss": 2.687, |
| "step": 2300 |
| }, |
| { |
| "epoch": 0.25866813428728674, |
| "grad_norm": 0.5223066806793213, |
| "learning_rate": 0.0002999976432184194, |
| "loss": 2.6279, |
| "step": 2350 |
| }, |
| { |
| "epoch": 0.26417171161254815, |
| "grad_norm": 0.4381965100765228, |
| "learning_rate": 0.0002999935887402823, |
| "loss": 2.6889, |
| "step": 2400 |
| }, |
| { |
| "epoch": 0.26967528893780957, |
| "grad_norm": 0.3969985842704773, |
| "learning_rate": 0.00029998754684064345, |
| "loss": 2.6565, |
| "step": 2450 |
| }, |
| { |
| "epoch": 0.275178866263071, |
| "grad_norm": 0.44786953926086426, |
| "learning_rate": 0.00029997951759955823, |
| "loss": 2.6736, |
| "step": 2500 |
| }, |
| { |
| "epoch": 0.28068244358833244, |
| "grad_norm": 0.3951723873615265, |
| "learning_rate": 0.0002999695011234145, |
| "loss": 2.6173, |
| "step": 2550 |
| }, |
| { |
| "epoch": 0.28618602091359385, |
| "grad_norm": 0.4258750081062317, |
| "learning_rate": 0.00029995749754493093, |
| "loss": 2.5992, |
| "step": 2600 |
| }, |
| { |
| "epoch": 0.29168959823885526, |
| "grad_norm": 0.41892707347869873, |
| "learning_rate": 0.0002999435070231555, |
| "loss": 2.6194, |
| "step": 2650 |
| }, |
| { |
| "epoch": 0.2971931755641167, |
| "grad_norm": 0.4064221978187561, |
| "learning_rate": 0.0002999275297434632, |
| "loss": 2.6265, |
| "step": 2700 |
| }, |
| { |
| "epoch": 0.3026967528893781, |
| "grad_norm": 0.3920956254005432, |
| "learning_rate": 0.00029990956591755365, |
| "loss": 2.6035, |
| "step": 2750 |
| }, |
| { |
| "epoch": 0.3082003302146395, |
| "grad_norm": 0.366318941116333, |
| "learning_rate": 0.0002998896157834484, |
| "loss": 2.6352, |
| "step": 2800 |
| }, |
| { |
| "epoch": 0.31370390753990096, |
| "grad_norm": 0.40867865085601807, |
| "learning_rate": 0.0002998676796054875, |
| "loss": 2.5674, |
| "step": 2850 |
| }, |
| { |
| "epoch": 0.31920748486516237, |
| "grad_norm": 0.3685750365257263, |
| "learning_rate": 0.00029984375767432627, |
| "loss": 2.5748, |
| "step": 2900 |
| }, |
| { |
| "epoch": 0.3247110621904238, |
| "grad_norm": 0.38134968280792236, |
| "learning_rate": 0.0002998178503069314, |
| "loss": 2.5703, |
| "step": 2950 |
| }, |
| { |
| "epoch": 0.3302146395156852, |
| "grad_norm": 0.36195963621139526, |
| "learning_rate": 0.00029978995784657643, |
| "loss": 2.5367, |
| "step": 3000 |
| }, |
| { |
| "epoch": 0.3357182168409466, |
| "grad_norm": 0.3696858584880829, |
| "learning_rate": 0.0002997600806628379, |
| "loss": 2.5693, |
| "step": 3050 |
| }, |
| { |
| "epoch": 0.341221794166208, |
| "grad_norm": 0.3404608964920044, |
| "learning_rate": 0.00029972821915158964, |
| "loss": 2.5454, |
| "step": 3100 |
| }, |
| { |
| "epoch": 0.3467253714914695, |
| "grad_norm": 0.33220136165618896, |
| "learning_rate": 0.0002996943737349981, |
| "loss": 2.5675, |
| "step": 3150 |
| }, |
| { |
| "epoch": 0.3522289488167309, |
| "grad_norm": 0.3744208812713623, |
| "learning_rate": 0.00029965854486151643, |
| "loss": 2.5453, |
| "step": 3200 |
| }, |
| { |
| "epoch": 0.3577325261419923, |
| "grad_norm": 0.36920756101608276, |
| "learning_rate": 0.0002996207330058788, |
| "loss": 2.5314, |
| "step": 3250 |
| }, |
| { |
| "epoch": 0.3632361034672537, |
| "grad_norm": 0.36301830410957336, |
| "learning_rate": 0.00029958093866909403, |
| "loss": 2.5361, |
| "step": 3300 |
| }, |
| { |
| "epoch": 0.3687396807925151, |
| "grad_norm": 0.3135142922401428, |
| "learning_rate": 0.00029953916237843853, |
| "loss": 2.5222, |
| "step": 3350 |
| }, |
| { |
| "epoch": 0.37424325811777653, |
| "grad_norm": 0.3634016513824463, |
| "learning_rate": 0.00029949540468745015, |
| "loss": 2.5179, |
| "step": 3400 |
| }, |
| { |
| "epoch": 0.379746835443038, |
| "grad_norm": 0.370914101600647, |
| "learning_rate": 0.00029944966617592017, |
| "loss": 2.5077, |
| "step": 3450 |
| }, |
| { |
| "epoch": 0.3852504127682994, |
| "grad_norm": 0.32538744807243347, |
| "learning_rate": 0.0002994019474498858, |
| "loss": 2.4873, |
| "step": 3500 |
| }, |
| { |
| "epoch": 0.3907539900935608, |
| "grad_norm": 0.3077157139778137, |
| "learning_rate": 0.0002993522491416221, |
| "loss": 2.5149, |
| "step": 3550 |
| }, |
| { |
| "epoch": 0.39625756741882223, |
| "grad_norm": 0.32142704725265503, |
| "learning_rate": 0.000299300571909634, |
| "loss": 2.4989, |
| "step": 3600 |
| }, |
| { |
| "epoch": 0.40176114474408364, |
| "grad_norm": 0.3485497832298279, |
| "learning_rate": 0.00029924691643864684, |
| "loss": 2.5114, |
| "step": 3650 |
| }, |
| { |
| "epoch": 0.40726472206934505, |
| "grad_norm": 0.34483280777931213, |
| "learning_rate": 0.0002991912834395981, |
| "loss": 2.4863, |
| "step": 3700 |
| }, |
| { |
| "epoch": 0.4127682993946065, |
| "grad_norm": 0.30801650881767273, |
| "learning_rate": 0.00029913367364962733, |
| "loss": 2.4935, |
| "step": 3750 |
| }, |
| { |
| "epoch": 0.41827187671986793, |
| "grad_norm": 0.3099140524864197, |
| "learning_rate": 0.00029907408783206674, |
| "loss": 2.5122, |
| "step": 3800 |
| }, |
| { |
| "epoch": 0.42377545404512934, |
| "grad_norm": 0.33073538541793823, |
| "learning_rate": 0.0002990125267764309, |
| "loss": 2.4613, |
| "step": 3850 |
| }, |
| { |
| "epoch": 0.42927903137039075, |
| "grad_norm": 0.3025052845478058, |
| "learning_rate": 0.00029894899129840653, |
| "loss": 2.4855, |
| "step": 3900 |
| }, |
| { |
| "epoch": 0.43478260869565216, |
| "grad_norm": 0.3200172483921051, |
| "learning_rate": 0.00029888348223984143, |
| "loss": 2.4542, |
| "step": 3950 |
| }, |
| { |
| "epoch": 0.4402861860209136, |
| "grad_norm": 0.31750544905662537, |
| "learning_rate": 0.0002988160004687335, |
| "loss": 2.4714, |
| "step": 4000 |
| }, |
| { |
| "epoch": 0.44578976334617504, |
| "grad_norm": 0.28732484579086304, |
| "learning_rate": 0.00029874654687921895, |
| "loss": 2.4595, |
| "step": 4050 |
| }, |
| { |
| "epoch": 0.45129334067143645, |
| "grad_norm": 0.33312809467315674, |
| "learning_rate": 0.0002986751223915609, |
| "loss": 2.4797, |
| "step": 4100 |
| }, |
| { |
| "epoch": 0.45679691799669786, |
| "grad_norm": 0.3109307289123535, |
| "learning_rate": 0.00029860172795213695, |
| "loss": 2.4506, |
| "step": 4150 |
| }, |
| { |
| "epoch": 0.46230049532195927, |
| "grad_norm": 0.31127694249153137, |
| "learning_rate": 0.0002985263645334266, |
| "loss": 2.4544, |
| "step": 4200 |
| }, |
| { |
| "epoch": 0.4678040726472207, |
| "grad_norm": 0.34132882952690125, |
| "learning_rate": 0.0002984490331339982, |
| "loss": 2.455, |
| "step": 4250 |
| }, |
| { |
| "epoch": 0.4733076499724821, |
| "grad_norm": 0.3107958436012268, |
| "learning_rate": 0.00029836973477849634, |
| "loss": 2.4733, |
| "step": 4300 |
| }, |
| { |
| "epoch": 0.47881122729774356, |
| "grad_norm": 0.27321889996528625, |
| "learning_rate": 0.00029828847051762753, |
| "loss": 2.4381, |
| "step": 4350 |
| }, |
| { |
| "epoch": 0.48431480462300497, |
| "grad_norm": 0.3114171326160431, |
| "learning_rate": 0.0002982052414281467, |
| "loss": 2.4612, |
| "step": 4400 |
| }, |
| { |
| "epoch": 0.4898183819482664, |
| "grad_norm": 0.2926501929759979, |
| "learning_rate": 0.00029812004861284294, |
| "loss": 2.4364, |
| "step": 4450 |
| }, |
| { |
| "epoch": 0.4953219592735278, |
| "grad_norm": 0.2824021279811859, |
| "learning_rate": 0.00029803289320052466, |
| "loss": 2.4453, |
| "step": 4500 |
| }, |
| { |
| "epoch": 0.5008255365987893, |
| "grad_norm": 0.2710123062133789, |
| "learning_rate": 0.00029794377634600485, |
| "loss": 2.4207, |
| "step": 4550 |
| }, |
| { |
| "epoch": 0.5063291139240507, |
| "grad_norm": 0.28232479095458984, |
| "learning_rate": 0.0002978526992300856, |
| "loss": 2.4332, |
| "step": 4600 |
| }, |
| { |
| "epoch": 0.5118326912493121, |
| "grad_norm": 0.2661692798137665, |
| "learning_rate": 0.0002977596630595427, |
| "loss": 2.4469, |
| "step": 4650 |
| }, |
| { |
| "epoch": 0.5173362685745735, |
| "grad_norm": 0.317094087600708, |
| "learning_rate": 0.0002976646690671094, |
| "loss": 2.4486, |
| "step": 4700 |
| }, |
| { |
| "epoch": 0.5228398458998349, |
| "grad_norm": 0.268915593624115, |
| "learning_rate": 0.0002975677185114602, |
| "loss": 2.4447, |
| "step": 4750 |
| }, |
| { |
| "epoch": 0.5283434232250963, |
| "grad_norm": 0.2714874744415283, |
| "learning_rate": 0.000297468812677194, |
| "loss": 2.4065, |
| "step": 4800 |
| }, |
| { |
| "epoch": 0.5338470005503577, |
| "grad_norm": 0.2872399389743805, |
| "learning_rate": 0.0002973679528748175, |
| "loss": 2.3943, |
| "step": 4850 |
| }, |
| { |
| "epoch": 0.5393505778756191, |
| "grad_norm": 0.24350598454475403, |
| "learning_rate": 0.00029726514044072736, |
| "loss": 2.3888, |
| "step": 4900 |
| }, |
| { |
| "epoch": 0.5448541552008805, |
| "grad_norm": 0.2634688913822174, |
| "learning_rate": 0.00029716037673719275, |
| "loss": 2.3953, |
| "step": 4950 |
| }, |
| { |
| "epoch": 0.550357732526142, |
| "grad_norm": 0.2842876613140106, |
| "learning_rate": 0.0002970536631523373, |
| "loss": 2.403, |
| "step": 5000 |
| }, |
| { |
| "epoch": 0.5558613098514034, |
| "grad_norm": 0.30130457878112793, |
| "learning_rate": 0.00029694500110012055, |
| "loss": 2.4393, |
| "step": 5050 |
| }, |
| { |
| "epoch": 0.5613648871766649, |
| "grad_norm": 0.2594560980796814, |
| "learning_rate": 0.00029683439202031936, |
| "loss": 2.3827, |
| "step": 5100 |
| }, |
| { |
| "epoch": 0.5668684645019263, |
| "grad_norm": 0.25417500734329224, |
| "learning_rate": 0.0002967218373785088, |
| "loss": 2.3789, |
| "step": 5150 |
| }, |
| { |
| "epoch": 0.5723720418271877, |
| "grad_norm": 0.2533874213695526, |
| "learning_rate": 0.0002966073386660428, |
| "loss": 2.4211, |
| "step": 5200 |
| }, |
| { |
| "epoch": 0.5778756191524491, |
| "grad_norm": 0.2343342900276184, |
| "learning_rate": 0.0002964908974000341, |
| "loss": 2.4332, |
| "step": 5250 |
| }, |
| { |
| "epoch": 0.5833791964777105, |
| "grad_norm": 0.25453534722328186, |
| "learning_rate": 0.0002963725151233345, |
| "loss": 2.3984, |
| "step": 5300 |
| }, |
| { |
| "epoch": 0.5888827738029719, |
| "grad_norm": 0.2670257091522217, |
| "learning_rate": 0.00029625219340451435, |
| "loss": 2.4126, |
| "step": 5350 |
| }, |
| { |
| "epoch": 0.5943863511282333, |
| "grad_norm": 0.2728422284126282, |
| "learning_rate": 0.00029612993383784154, |
| "loss": 2.4152, |
| "step": 5400 |
| }, |
| { |
| "epoch": 0.5998899284534948, |
| "grad_norm": 0.27964428067207336, |
| "learning_rate": 0.0002960057380432606, |
| "loss": 2.3675, |
| "step": 5450 |
| }, |
| { |
| "epoch": 0.6053935057787562, |
| "grad_norm": 0.27354755997657776, |
| "learning_rate": 0.00029587960766637103, |
| "loss": 2.4109, |
| "step": 5500 |
| }, |
| { |
| "epoch": 0.6108970831040176, |
| "grad_norm": 0.26175597310066223, |
| "learning_rate": 0.000295751544378406, |
| "loss": 2.3953, |
| "step": 5550 |
| }, |
| { |
| "epoch": 0.616400660429279, |
| "grad_norm": 0.2590219974517822, |
| "learning_rate": 0.0002956215498762093, |
| "loss": 2.3699, |
| "step": 5600 |
| }, |
| { |
| "epoch": 0.6219042377545404, |
| "grad_norm": 0.2495882660150528, |
| "learning_rate": 0.0002954896258822139, |
| "loss": 2.3949, |
| "step": 5650 |
| }, |
| { |
| "epoch": 0.6274078150798019, |
| "grad_norm": 0.26052239537239075, |
| "learning_rate": 0.0002953557741444183, |
| "loss": 2.3642, |
| "step": 5700 |
| }, |
| { |
| "epoch": 0.6329113924050633, |
| "grad_norm": 0.2627600431442261, |
| "learning_rate": 0.0002952199964363638, |
| "loss": 2.3895, |
| "step": 5750 |
| }, |
| { |
| "epoch": 0.6384149697303247, |
| "grad_norm": 0.2417898327112198, |
| "learning_rate": 0.00029508229455711086, |
| "loss": 2.4087, |
| "step": 5800 |
| }, |
| { |
| "epoch": 0.6439185470555862, |
| "grad_norm": 0.24891215562820435, |
| "learning_rate": 0.00029494267033121525, |
| "loss": 2.3642, |
| "step": 5850 |
| }, |
| { |
| "epoch": 0.6494221243808476, |
| "grad_norm": 0.2453078180551529, |
| "learning_rate": 0.0002948011256087041, |
| "loss": 2.3761, |
| "step": 5900 |
| }, |
| { |
| "epoch": 0.654925701706109, |
| "grad_norm": 0.24607980251312256, |
| "learning_rate": 0.0002946576622650509, |
| "loss": 2.3596, |
| "step": 5950 |
| }, |
| { |
| "epoch": 0.6604292790313704, |
| "grad_norm": 0.2540590763092041, |
| "learning_rate": 0.0002945122822011513, |
| "loss": 2.3666, |
| "step": 6000 |
| }, |
| { |
| "epoch": 0.6659328563566318, |
| "grad_norm": 0.25918814539909363, |
| "learning_rate": 0.00029436498734329727, |
| "loss": 2.3494, |
| "step": 6050 |
| }, |
| { |
| "epoch": 0.6714364336818932, |
| "grad_norm": 0.27554330229759216, |
| "learning_rate": 0.0002942157796431521, |
| "loss": 2.3793, |
| "step": 6100 |
| }, |
| { |
| "epoch": 0.6769400110071546, |
| "grad_norm": 0.22879928350448608, |
| "learning_rate": 0.00029406466107772416, |
| "loss": 2.3898, |
| "step": 6150 |
| }, |
| { |
| "epoch": 0.682443588332416, |
| "grad_norm": 0.2851306200027466, |
| "learning_rate": 0.00029391163364934095, |
| "loss": 2.3493, |
| "step": 6200 |
| }, |
| { |
| "epoch": 0.6879471656576774, |
| "grad_norm": 0.255014568567276, |
| "learning_rate": 0.0002937566993856225, |
| "loss": 2.3959, |
| "step": 6250 |
| }, |
| { |
| "epoch": 0.693450742982939, |
| "grad_norm": 0.23992781341075897, |
| "learning_rate": 0.00029359986033945454, |
| "loss": 2.3588, |
| "step": 6300 |
| }, |
| { |
| "epoch": 0.6989543203082004, |
| "grad_norm": 0.24902793765068054, |
| "learning_rate": 0.000293441118588961, |
| "loss": 2.3908, |
| "step": 6350 |
| }, |
| { |
| "epoch": 0.7044578976334618, |
| "grad_norm": 0.2294873744249344, |
| "learning_rate": 0.0002932804762374771, |
| "loss": 2.398, |
| "step": 6400 |
| }, |
| { |
| "epoch": 0.7099614749587232, |
| "grad_norm": 0.23560309410095215, |
| "learning_rate": 0.00029311793541352075, |
| "loss": 2.3481, |
| "step": 6450 |
| }, |
| { |
| "epoch": 0.7154650522839846, |
| "grad_norm": 0.27493488788604736, |
| "learning_rate": 0.00029295349827076497, |
| "loss": 2.3248, |
| "step": 6500 |
| }, |
| { |
| "epoch": 0.720968629609246, |
| "grad_norm": 0.22731854021549225, |
| "learning_rate": 0.0002927871669880089, |
| "loss": 2.3444, |
| "step": 6550 |
| }, |
| { |
| "epoch": 0.7264722069345074, |
| "grad_norm": 0.22198539972305298, |
| "learning_rate": 0.0002926189437691492, |
| "loss": 2.3148, |
| "step": 6600 |
| }, |
| { |
| "epoch": 0.7319757842597688, |
| "grad_norm": 0.25049686431884766, |
| "learning_rate": 0.0002924488308431508, |
| "loss": 2.3134, |
| "step": 6650 |
| }, |
| { |
| "epoch": 0.7374793615850302, |
| "grad_norm": 0.2354484349489212, |
| "learning_rate": 0.0002922768304640172, |
| "loss": 2.348, |
| "step": 6700 |
| }, |
| { |
| "epoch": 0.7429829389102917, |
| "grad_norm": 0.23955915868282318, |
| "learning_rate": 0.00029210294491076094, |
| "loss": 2.3172, |
| "step": 6750 |
| }, |
| { |
| "epoch": 0.7484865162355531, |
| "grad_norm": 0.2476130574941635, |
| "learning_rate": 0.000291927176487373, |
| "loss": 2.3842, |
| "step": 6800 |
| }, |
| { |
| "epoch": 0.7539900935608145, |
| "grad_norm": 0.2247840166091919, |
| "learning_rate": 0.00029174952752279254, |
| "loss": 2.3302, |
| "step": 6850 |
| }, |
| { |
| "epoch": 0.759493670886076, |
| "grad_norm": 0.22530816495418549, |
| "learning_rate": 0.000291570000370876, |
| "loss": 2.3707, |
| "step": 6900 |
| }, |
| { |
| "epoch": 0.7649972482113374, |
| "grad_norm": 0.2242722362279892, |
| "learning_rate": 0.00029138859741036587, |
| "loss": 2.312, |
| "step": 6950 |
| }, |
| { |
| "epoch": 0.7705008255365988, |
| "grad_norm": 0.23376357555389404, |
| "learning_rate": 0.0002912053210448592, |
| "loss": 2.3334, |
| "step": 7000 |
| }, |
| { |
| "epoch": 0.7760044028618602, |
| "grad_norm": 0.23131632804870605, |
| "learning_rate": 0.0002910201737027757, |
| "loss": 2.3293, |
| "step": 7050 |
| }, |
| { |
| "epoch": 0.7815079801871216, |
| "grad_norm": 0.21440596878528595, |
| "learning_rate": 0.0002908331578373256, |
| "loss": 2.3593, |
| "step": 7100 |
| }, |
| { |
| "epoch": 0.787011557512383, |
| "grad_norm": 0.225584477186203, |
| "learning_rate": 0.00029064427592647715, |
| "loss": 2.3464, |
| "step": 7150 |
| }, |
| { |
| "epoch": 0.7925151348376445, |
| "grad_norm": 0.22819140553474426, |
| "learning_rate": 0.0002904535304729238, |
| "loss": 2.3145, |
| "step": 7200 |
| }, |
| { |
| "epoch": 0.7980187121629059, |
| "grad_norm": 0.2503032982349396, |
| "learning_rate": 0.00029026092400405115, |
| "loss": 2.3376, |
| "step": 7250 |
| }, |
| { |
| "epoch": 0.8035222894881673, |
| "grad_norm": 0.2061757743358612, |
| "learning_rate": 0.00029006645907190295, |
| "loss": 2.3252, |
| "step": 7300 |
| }, |
| { |
| "epoch": 0.8090258668134287, |
| "grad_norm": 0.2099982351064682, |
| "learning_rate": 0.00028987013825314804, |
| "loss": 2.35, |
| "step": 7350 |
| }, |
| { |
| "epoch": 0.8145294441386901, |
| "grad_norm": 0.212050199508667, |
| "learning_rate": 0.00028967196414904573, |
| "loss": 2.3304, |
| "step": 7400 |
| }, |
| { |
| "epoch": 0.8200330214639515, |
| "grad_norm": 0.22283074259757996, |
| "learning_rate": 0.0002894719393854114, |
| "loss": 2.3389, |
| "step": 7450 |
| }, |
| { |
| "epoch": 0.825536598789213, |
| "grad_norm": 0.2060365378856659, |
| "learning_rate": 0.0002892700666125817, |
| "loss": 2.3215, |
| "step": 7500 |
| }, |
| { |
| "epoch": 0.8310401761144744, |
| "grad_norm": 0.22510799765586853, |
| "learning_rate": 0.0002890663485053797, |
| "loss": 2.3295, |
| "step": 7550 |
| }, |
| { |
| "epoch": 0.8365437534397359, |
| "grad_norm": 0.21860332787036896, |
| "learning_rate": 0.00028886078776307905, |
| "loss": 2.3286, |
| "step": 7600 |
| }, |
| { |
| "epoch": 0.8420473307649973, |
| "grad_norm": 0.20771068334579468, |
| "learning_rate": 0.00028865338710936826, |
| "loss": 2.3143, |
| "step": 7650 |
| }, |
| { |
| "epoch": 0.8475509080902587, |
| "grad_norm": 0.23098550736904144, |
| "learning_rate": 0.0002884441492923151, |
| "loss": 2.3583, |
| "step": 7700 |
| }, |
| { |
| "epoch": 0.8530544854155201, |
| "grad_norm": 0.204328715801239, |
| "learning_rate": 0.00028823307708432963, |
| "loss": 2.3425, |
| "step": 7750 |
| }, |
| { |
| "epoch": 0.8585580627407815, |
| "grad_norm": 0.20610789954662323, |
| "learning_rate": 0.0002880201732821275, |
| "loss": 2.3269, |
| "step": 7800 |
| }, |
| { |
| "epoch": 0.8640616400660429, |
| "grad_norm": 0.22584037482738495, |
| "learning_rate": 0.0002878054407066935, |
| "loss": 2.3061, |
| "step": 7850 |
| }, |
| { |
| "epoch": 0.8695652173913043, |
| "grad_norm": 0.22317476570606232, |
| "learning_rate": 0.0002875888822032433, |
| "loss": 2.3047, |
| "step": 7900 |
| }, |
| { |
| "epoch": 0.8750687947165657, |
| "grad_norm": 0.2132745087146759, |
| "learning_rate": 0.00028737050064118645, |
| "loss": 2.2925, |
| "step": 7950 |
| }, |
| { |
| "epoch": 0.8805723720418271, |
| "grad_norm": 0.21960894763469696, |
| "learning_rate": 0.00028715029891408805, |
| "loss": 2.3365, |
| "step": 8000 |
| }, |
| { |
| "epoch": 0.8860759493670886, |
| "grad_norm": 0.2113264799118042, |
| "learning_rate": 0.0002869282799396305, |
| "loss": 2.2814, |
| "step": 8050 |
| }, |
| { |
| "epoch": 0.8915795266923501, |
| "grad_norm": 0.19959582388401031, |
| "learning_rate": 0.00028670444665957465, |
| "loss": 2.2787, |
| "step": 8100 |
| }, |
| { |
| "epoch": 0.8970831040176115, |
| "grad_norm": 0.24770890176296234, |
| "learning_rate": 0.00028647880203972115, |
| "loss": 2.3312, |
| "step": 8150 |
| }, |
| { |
| "epoch": 0.9025866813428729, |
| "grad_norm": 0.2229924350976944, |
| "learning_rate": 0.0002862513490698709, |
| "loss": 2.2825, |
| "step": 8200 |
| }, |
| { |
| "epoch": 0.9080902586681343, |
| "grad_norm": 0.23118580877780914, |
| "learning_rate": 0.0002860220907637856, |
| "loss": 2.2973, |
| "step": 8250 |
| }, |
| { |
| "epoch": 0.9135938359933957, |
| "grad_norm": 0.1978590488433838, |
| "learning_rate": 0.0002857910301591475, |
| "loss": 2.3106, |
| "step": 8300 |
| }, |
| { |
| "epoch": 0.9190974133186571, |
| "grad_norm": 0.20972274243831635, |
| "learning_rate": 0.0002855581703175198, |
| "loss": 2.3489, |
| "step": 8350 |
| }, |
| { |
| "epoch": 0.9246009906439185, |
| "grad_norm": 0.20369485020637512, |
| "learning_rate": 0.0002853235143243052, |
| "loss": 2.3135, |
| "step": 8400 |
| }, |
| { |
| "epoch": 0.93010456796918, |
| "grad_norm": 0.19085292518138885, |
| "learning_rate": 0.00028508706528870576, |
| "loss": 2.2969, |
| "step": 8450 |
| }, |
| { |
| "epoch": 0.9356081452944414, |
| "grad_norm": 0.21080902218818665, |
| "learning_rate": 0.0002848488263436814, |
| "loss": 2.2979, |
| "step": 8500 |
| }, |
| { |
| "epoch": 0.9411117226197028, |
| "grad_norm": 0.1939728856086731, |
| "learning_rate": 0.00028460880064590835, |
| "loss": 2.31, |
| "step": 8550 |
| }, |
| { |
| "epoch": 0.9466152999449642, |
| "grad_norm": 0.23909969627857208, |
| "learning_rate": 0.0002843669913757375, |
| "loss": 2.2585, |
| "step": 8600 |
| }, |
| { |
| "epoch": 0.9521188772702256, |
| "grad_norm": 0.22809971868991852, |
| "learning_rate": 0.000284123401737152, |
| "loss": 2.3058, |
| "step": 8650 |
| }, |
| { |
| "epoch": 0.9576224545954871, |
| "grad_norm": 0.2066497504711151, |
| "learning_rate": 0.00028387803495772513, |
| "loss": 2.2736, |
| "step": 8700 |
| }, |
| { |
| "epoch": 0.9631260319207485, |
| "grad_norm": 0.1921154260635376, |
| "learning_rate": 0.00028363089428857716, |
| "loss": 2.2813, |
| "step": 8750 |
| }, |
| { |
| "epoch": 0.9686296092460099, |
| "grad_norm": 0.19423851370811462, |
| "learning_rate": 0.00028338198300433264, |
| "loss": 2.314, |
| "step": 8800 |
| }, |
| { |
| "epoch": 0.9741331865712713, |
| "grad_norm": 0.224751815199852, |
| "learning_rate": 0.0002831313044030768, |
| "loss": 2.3335, |
| "step": 8850 |
| }, |
| { |
| "epoch": 0.9796367638965328, |
| "grad_norm": 0.21634523570537567, |
| "learning_rate": 0.00028287886180631183, |
| "loss": 2.3021, |
| "step": 8900 |
| }, |
| { |
| "epoch": 0.9851403412217942, |
| "grad_norm": 0.19320085644721985, |
| "learning_rate": 0.00028262465855891306, |
| "loss": 2.2635, |
| "step": 8950 |
| }, |
| { |
| "epoch": 0.9906439185470556, |
| "grad_norm": 0.20169596374034882, |
| "learning_rate": 0.00028236869802908453, |
| "loss": 2.3013, |
| "step": 9000 |
| }, |
| { |
| "epoch": 0.996147495872317, |
| "grad_norm": 0.2045382857322693, |
| "learning_rate": 0.00028211098360831407, |
| "loss": 2.2576, |
| "step": 9050 |
| }, |
| { |
| "epoch": 1.0016510731975785, |
| "grad_norm": 0.2171631157398224, |
| "learning_rate": 0.00028185151871132894, |
| "loss": 2.3136, |
| "step": 9100 |
| }, |
| { |
| "epoch": 1.00715465052284, |
| "grad_norm": 0.19541560113430023, |
| "learning_rate": 0.00028159030677605017, |
| "loss": 2.2547, |
| "step": 9150 |
| }, |
| { |
| "epoch": 1.0126582278481013, |
| "grad_norm": 0.19888882339000702, |
| "learning_rate": 0.0002813273512635472, |
| "loss": 2.2275, |
| "step": 9200 |
| }, |
| { |
| "epoch": 1.0181618051733627, |
| "grad_norm": 0.24578270316123962, |
| "learning_rate": 0.0002810626556579918, |
| "loss": 2.2775, |
| "step": 9250 |
| }, |
| { |
| "epoch": 1.0236653824986242, |
| "grad_norm": 0.19902436435222626, |
| "learning_rate": 0.00028079622346661217, |
| "loss": 2.2353, |
| "step": 9300 |
| }, |
| { |
| "epoch": 1.0291689598238856, |
| "grad_norm": 0.19234856963157654, |
| "learning_rate": 0.00028052805821964633, |
| "loss": 2.2769, |
| "step": 9350 |
| }, |
| { |
| "epoch": 1.034672537149147, |
| "grad_norm": 0.19912928342819214, |
| "learning_rate": 0.0002802581634702952, |
| "loss": 2.2625, |
| "step": 9400 |
| }, |
| { |
| "epoch": 1.0401761144744084, |
| "grad_norm": 0.1918436884880066, |
| "learning_rate": 0.00027998654279467604, |
| "loss": 2.2335, |
| "step": 9450 |
| }, |
| { |
| "epoch": 1.0456796917996698, |
| "grad_norm": 0.20422010123729706, |
| "learning_rate": 0.0002797131997917743, |
| "loss": 2.3022, |
| "step": 9500 |
| }, |
| { |
| "epoch": 1.0511832691249312, |
| "grad_norm": 0.19621697068214417, |
| "learning_rate": 0.00027943813808339657, |
| "loss": 2.2378, |
| "step": 9550 |
| }, |
| { |
| "epoch": 1.0566868464501926, |
| "grad_norm": 0.18403683602809906, |
| "learning_rate": 0.0002791613613141223, |
| "loss": 2.2502, |
| "step": 9600 |
| }, |
| { |
| "epoch": 1.062190423775454, |
| "grad_norm": 0.2140798568725586, |
| "learning_rate": 0.0002788828731512556, |
| "loss": 2.2722, |
| "step": 9650 |
| }, |
| { |
| "epoch": 1.0676940011007154, |
| "grad_norm": 0.19459135830402374, |
| "learning_rate": 0.0002786026772847767, |
| "loss": 2.2361, |
| "step": 9700 |
| }, |
| { |
| "epoch": 1.0731975784259769, |
| "grad_norm": 0.20383061468601227, |
| "learning_rate": 0.00027832077742729277, |
| "loss": 2.2471, |
| "step": 9750 |
| }, |
| { |
| "epoch": 1.0787011557512383, |
| "grad_norm": 0.1901775300502777, |
| "learning_rate": 0.0002780371773139891, |
| "loss": 2.2968, |
| "step": 9800 |
| }, |
| { |
| "epoch": 1.0842047330764997, |
| "grad_norm": 0.20100219547748566, |
| "learning_rate": 0.0002777518807025793, |
| "loss": 2.2568, |
| "step": 9850 |
| }, |
| { |
| "epoch": 1.089708310401761, |
| "grad_norm": 0.2127520740032196, |
| "learning_rate": 0.00027746489137325586, |
| "loss": 2.28, |
| "step": 9900 |
| }, |
| { |
| "epoch": 1.0952118877270225, |
| "grad_norm": 0.19059176743030548, |
| "learning_rate": 0.00027717621312863965, |
| "loss": 2.284, |
| "step": 9950 |
| }, |
| { |
| "epoch": 1.100715465052284, |
| "grad_norm": 0.20845113694667816, |
| "learning_rate": 0.0002768858497937299, |
| "loss": 2.2409, |
| "step": 10000 |
| }, |
| { |
| "epoch": 1.1062190423775453, |
| "grad_norm": 0.18969906866550446, |
| "learning_rate": 0.00027659380521585325, |
| "loss": 2.2725, |
| "step": 10050 |
| }, |
| { |
| "epoch": 1.1117226197028067, |
| "grad_norm": 0.19071218371391296, |
| "learning_rate": 0.0002763000832646129, |
| "loss": 2.2624, |
| "step": 10100 |
| }, |
| { |
| "epoch": 1.1172261970280681, |
| "grad_norm": 0.21723783016204834, |
| "learning_rate": 0.0002760046878318373, |
| "loss": 2.2453, |
| "step": 10150 |
| }, |
| { |
| "epoch": 1.1227297743533298, |
| "grad_norm": 0.19336852431297302, |
| "learning_rate": 0.00027570762283152874, |
| "loss": 2.2322, |
| "step": 10200 |
| }, |
| { |
| "epoch": 1.1282333516785912, |
| "grad_norm": 0.19500704109668732, |
| "learning_rate": 0.0002754088921998112, |
| "loss": 2.1858, |
| "step": 10250 |
| }, |
| { |
| "epoch": 1.1337369290038526, |
| "grad_norm": 0.19920021295547485, |
| "learning_rate": 0.0002751084998948784, |
| "loss": 2.2286, |
| "step": 10300 |
| }, |
| { |
| "epoch": 1.139240506329114, |
| "grad_norm": 0.18863485753536224, |
| "learning_rate": 0.00027480644989694126, |
| "loss": 2.2623, |
| "step": 10350 |
| }, |
| { |
| "epoch": 1.1447440836543754, |
| "grad_norm": 0.18571630120277405, |
| "learning_rate": 0.0002745027462081753, |
| "loss": 2.2616, |
| "step": 10400 |
| }, |
| { |
| "epoch": 1.1502476609796368, |
| "grad_norm": 0.187313050031662, |
| "learning_rate": 0.00027419739285266745, |
| "loss": 2.2608, |
| "step": 10450 |
| }, |
| { |
| "epoch": 1.1557512383048982, |
| "grad_norm": 0.19139795005321503, |
| "learning_rate": 0.0002738903938763628, |
| "loss": 2.2204, |
| "step": 10500 |
| }, |
| { |
| "epoch": 1.1612548156301596, |
| "grad_norm": 0.188013955950737, |
| "learning_rate": 0.0002735817533470109, |
| "loss": 2.2596, |
| "step": 10550 |
| }, |
| { |
| "epoch": 1.166758392955421, |
| "grad_norm": 0.1884177178144455, |
| "learning_rate": 0.0002732714753541122, |
| "loss": 2.2618, |
| "step": 10600 |
| }, |
| { |
| "epoch": 1.1722619702806825, |
| "grad_norm": 0.19187124073505402, |
| "learning_rate": 0.00027295956400886335, |
| "loss": 2.2075, |
| "step": 10650 |
| }, |
| { |
| "epoch": 1.1777655476059439, |
| "grad_norm": 0.19815993309020996, |
| "learning_rate": 0.0002726460234441031, |
| "loss": 2.2677, |
| "step": 10700 |
| }, |
| { |
| "epoch": 1.1832691249312053, |
| "grad_norm": 0.18218094110488892, |
| "learning_rate": 0.00027233085781425744, |
| "loss": 2.2697, |
| "step": 10750 |
| }, |
| { |
| "epoch": 1.1887727022564667, |
| "grad_norm": 0.1923886239528656, |
| "learning_rate": 0.0002720140712952845, |
| "loss": 2.2127, |
| "step": 10800 |
| }, |
| { |
| "epoch": 1.194276279581728, |
| "grad_norm": 0.18907909095287323, |
| "learning_rate": 0.0002716956680846193, |
| "loss": 2.1989, |
| "step": 10850 |
| }, |
| { |
| "epoch": 1.1997798569069895, |
| "grad_norm": 0.1902381181716919, |
| "learning_rate": 0.000271375652401118, |
| "loss": 2.2211, |
| "step": 10900 |
| }, |
| { |
| "epoch": 1.205283434232251, |
| "grad_norm": 0.20184637606143951, |
| "learning_rate": 0.0002710540284850023, |
| "loss": 2.2451, |
| "step": 10950 |
| }, |
| { |
| "epoch": 1.2107870115575123, |
| "grad_norm": 0.19903656840324402, |
| "learning_rate": 0.00027073080059780285, |
| "loss": 2.2233, |
| "step": 11000 |
| }, |
| { |
| "epoch": 1.2162905888827737, |
| "grad_norm": 0.19023925065994263, |
| "learning_rate": 0.000270405973022303, |
| "loss": 2.2438, |
| "step": 11050 |
| }, |
| { |
| "epoch": 1.2217941662080352, |
| "grad_norm": 0.17722088098526, |
| "learning_rate": 0.0002700795500624822, |
| "loss": 2.259, |
| "step": 11100 |
| }, |
| { |
| "epoch": 1.2272977435332966, |
| "grad_norm": 0.19715473055839539, |
| "learning_rate": 0.0002697515360434587, |
| "loss": 2.2509, |
| "step": 11150 |
| }, |
| { |
| "epoch": 1.232801320858558, |
| "grad_norm": 0.18065761029720306, |
| "learning_rate": 0.00026942193531143225, |
| "loss": 2.2006, |
| "step": 11200 |
| }, |
| { |
| "epoch": 1.2383048981838194, |
| "grad_norm": 0.18757817149162292, |
| "learning_rate": 0.00026909075223362683, |
| "loss": 2.2215, |
| "step": 11250 |
| }, |
| { |
| "epoch": 1.243808475509081, |
| "grad_norm": 0.18816334009170532, |
| "learning_rate": 0.00026875799119823243, |
| "loss": 2.2314, |
| "step": 11300 |
| }, |
| { |
| "epoch": 1.2493120528343424, |
| "grad_norm": 0.19772210717201233, |
| "learning_rate": 0.0002684236566143471, |
| "loss": 2.2256, |
| "step": 11350 |
| }, |
| { |
| "epoch": 1.2548156301596038, |
| "grad_norm": 0.19510945677757263, |
| "learning_rate": 0.0002680877529119184, |
| "loss": 2.2427, |
| "step": 11400 |
| }, |
| { |
| "epoch": 1.2603192074848653, |
| "grad_norm": 0.19960781931877136, |
| "learning_rate": 0.0002677502845416849, |
| "loss": 2.2324, |
| "step": 11450 |
| }, |
| { |
| "epoch": 1.2658227848101267, |
| "grad_norm": 0.1882307380437851, |
| "learning_rate": 0.0002674112559751169, |
| "loss": 2.2432, |
| "step": 11500 |
| }, |
| { |
| "epoch": 1.271326362135388, |
| "grad_norm": 0.18959587812423706, |
| "learning_rate": 0.00026707067170435767, |
| "loss": 2.19, |
| "step": 11550 |
| }, |
| { |
| "epoch": 1.2768299394606495, |
| "grad_norm": 0.18768323957920074, |
| "learning_rate": 0.0002667285362421634, |
| "loss": 2.2294, |
| "step": 11600 |
| }, |
| { |
| "epoch": 1.282333516785911, |
| "grad_norm": 0.20077955722808838, |
| "learning_rate": 0.00026638485412184355, |
| "loss": 2.2518, |
| "step": 11650 |
| }, |
| { |
| "epoch": 1.2878370941111723, |
| "grad_norm": 0.1956523358821869, |
| "learning_rate": 0.00026603962989720105, |
| "loss": 2.2662, |
| "step": 11700 |
| }, |
| { |
| "epoch": 1.2933406714364337, |
| "grad_norm": 0.19226029515266418, |
| "learning_rate": 0.0002656928681424718, |
| "loss": 2.2133, |
| "step": 11750 |
| }, |
| { |
| "epoch": 1.2988442487616951, |
| "grad_norm": 0.16987648606300354, |
| "learning_rate": 0.0002653445734522639, |
| "loss": 2.2226, |
| "step": 11800 |
| }, |
| { |
| "epoch": 1.3043478260869565, |
| "grad_norm": 0.17524655163288116, |
| "learning_rate": 0.00026499475044149703, |
| "loss": 2.2555, |
| "step": 11850 |
| }, |
| { |
| "epoch": 1.309851403412218, |
| "grad_norm": 0.19188162684440613, |
| "learning_rate": 0.00026464340374534104, |
| "loss": 2.2234, |
| "step": 11900 |
| }, |
| { |
| "epoch": 1.3153549807374794, |
| "grad_norm": 0.17676417529582977, |
| "learning_rate": 0.00026429053801915487, |
| "loss": 2.2188, |
| "step": 11950 |
| }, |
| { |
| "epoch": 1.3208585580627408, |
| "grad_norm": 0.20071938633918762, |
| "learning_rate": 0.0002639361579384245, |
| "loss": 2.234, |
| "step": 12000 |
| }, |
| { |
| "epoch": 1.3263621353880022, |
| "grad_norm": 0.19267459213733673, |
| "learning_rate": 0.0002635802681987012, |
| "loss": 2.235, |
| "step": 12050 |
| }, |
| { |
| "epoch": 1.3318657127132636, |
| "grad_norm": 0.1724405139684677, |
| "learning_rate": 0.00026322287351553944, |
| "loss": 2.2388, |
| "step": 12100 |
| }, |
| { |
| "epoch": 1.337369290038525, |
| "grad_norm": 0.1907806396484375, |
| "learning_rate": 0.000262863978624434, |
| "loss": 2.1937, |
| "step": 12150 |
| }, |
| { |
| "epoch": 1.3428728673637864, |
| "grad_norm": 0.18452796339988708, |
| "learning_rate": 0.0002625035882807578, |
| "loss": 2.2429, |
| "step": 12200 |
| }, |
| { |
| "epoch": 1.3483764446890478, |
| "grad_norm": 0.17565739154815674, |
| "learning_rate": 0.0002621417072596982, |
| "loss": 2.2367, |
| "step": 12250 |
| }, |
| { |
| "epoch": 1.3538800220143092, |
| "grad_norm": 0.1761862337589264, |
| "learning_rate": 0.00026177834035619445, |
| "loss": 2.2258, |
| "step": 12300 |
| }, |
| { |
| "epoch": 1.3593835993395706, |
| "grad_norm": 0.18616576492786407, |
| "learning_rate": 0.0002614134923848736, |
| "loss": 2.2207, |
| "step": 12350 |
| }, |
| { |
| "epoch": 1.364887176664832, |
| "grad_norm": 0.18103408813476562, |
| "learning_rate": 0.00026104716817998696, |
| "loss": 2.2346, |
| "step": 12400 |
| }, |
| { |
| "epoch": 1.3703907539900935, |
| "grad_norm": 0.18511448800563812, |
| "learning_rate": 0.00026067937259534595, |
| "loss": 2.2363, |
| "step": 12450 |
| }, |
| { |
| "epoch": 1.3758943313153549, |
| "grad_norm": 0.1737717241048813, |
| "learning_rate": 0.00026031011050425796, |
| "loss": 2.2199, |
| "step": 12500 |
| }, |
| { |
| "epoch": 1.3813979086406163, |
| "grad_norm": 0.18761631846427917, |
| "learning_rate": 0.00025993938679946154, |
| "loss": 2.228, |
| "step": 12550 |
| }, |
| { |
| "epoch": 1.3869014859658777, |
| "grad_norm": 0.18815304338932037, |
| "learning_rate": 0.0002595672063930617, |
| "loss": 2.2034, |
| "step": 12600 |
| }, |
| { |
| "epoch": 1.3924050632911391, |
| "grad_norm": 0.1868327558040619, |
| "learning_rate": 0.00025919357421646486, |
| "loss": 2.197, |
| "step": 12650 |
| }, |
| { |
| "epoch": 1.3979086406164005, |
| "grad_norm": 0.17544154822826385, |
| "learning_rate": 0.00025881849522031345, |
| "loss": 2.1956, |
| "step": 12700 |
| }, |
| { |
| "epoch": 1.4034122179416622, |
| "grad_norm": 0.18338458240032196, |
| "learning_rate": 0.00025844197437442033, |
| "loss": 2.2291, |
| "step": 12750 |
| }, |
| { |
| "epoch": 1.4089157952669236, |
| "grad_norm": 0.17127631604671478, |
| "learning_rate": 0.00025806401666770277, |
| "loss": 2.2205, |
| "step": 12800 |
| }, |
| { |
| "epoch": 1.414419372592185, |
| "grad_norm": 0.17331157624721527, |
| "learning_rate": 0.00025768462710811673, |
| "loss": 2.2356, |
| "step": 12850 |
| }, |
| { |
| "epoch": 1.4199229499174464, |
| "grad_norm": 0.16098715364933014, |
| "learning_rate": 0.00025730381072259026, |
| "loss": 2.2065, |
| "step": 12900 |
| }, |
| { |
| "epoch": 1.4254265272427078, |
| "grad_norm": 0.17098484933376312, |
| "learning_rate": 0.00025692157255695667, |
| "loss": 2.2323, |
| "step": 12950 |
| }, |
| { |
| "epoch": 1.4309301045679692, |
| "grad_norm": 0.15763573348522186, |
| "learning_rate": 0.00025653791767588823, |
| "loss": 2.2064, |
| "step": 13000 |
| }, |
| { |
| "epoch": 1.4364336818932306, |
| "grad_norm": 0.17587807774543762, |
| "learning_rate": 0.0002561528511628286, |
| "loss": 2.2085, |
| "step": 13050 |
| }, |
| { |
| "epoch": 1.441937259218492, |
| "grad_norm": 0.17770230770111084, |
| "learning_rate": 0.00025576637811992555, |
| "loss": 2.2204, |
| "step": 13100 |
| }, |
| { |
| "epoch": 1.4474408365437534, |
| "grad_norm": 0.18236953020095825, |
| "learning_rate": 0.0002553785036679636, |
| "loss": 2.2388, |
| "step": 13150 |
| }, |
| { |
| "epoch": 1.4529444138690149, |
| "grad_norm": 0.16505366563796997, |
| "learning_rate": 0.00025498923294629594, |
| "loss": 2.2345, |
| "step": 13200 |
| }, |
| { |
| "epoch": 1.4584479911942763, |
| "grad_norm": 0.1687227189540863, |
| "learning_rate": 0.0002545985711127763, |
| "loss": 2.1652, |
| "step": 13250 |
| }, |
| { |
| "epoch": 1.4639515685195377, |
| "grad_norm": 0.17849218845367432, |
| "learning_rate": 0.00025420652334369085, |
| "loss": 2.2115, |
| "step": 13300 |
| }, |
| { |
| "epoch": 1.469455145844799, |
| "grad_norm": 0.17708005011081696, |
| "learning_rate": 0.0002538130948336894, |
| "loss": 2.2017, |
| "step": 13350 |
| }, |
| { |
| "epoch": 1.4749587231700605, |
| "grad_norm": 0.17772626876831055, |
| "learning_rate": 0.0002534182907957165, |
| "loss": 2.1926, |
| "step": 13400 |
| }, |
| { |
| "epoch": 1.480462300495322, |
| "grad_norm": 0.16764992475509644, |
| "learning_rate": 0.00025302211646094277, |
| "loss": 2.1986, |
| "step": 13450 |
| }, |
| { |
| "epoch": 1.4859658778205833, |
| "grad_norm": 0.19117778539657593, |
| "learning_rate": 0.00025262457707869506, |
| "loss": 2.2118, |
| "step": 13500 |
| }, |
| { |
| "epoch": 1.4914694551458447, |
| "grad_norm": 0.17632248997688293, |
| "learning_rate": 0.00025222567791638743, |
| "loss": 2.1897, |
| "step": 13550 |
| }, |
| { |
| "epoch": 1.4969730324711064, |
| "grad_norm": 0.19731369614601135, |
| "learning_rate": 0.0002518254242594508, |
| "loss": 2.1945, |
| "step": 13600 |
| }, |
| { |
| "epoch": 1.5024766097963678, |
| "grad_norm": 0.16187229752540588, |
| "learning_rate": 0.0002514238214112635, |
| "loss": 2.1959, |
| "step": 13650 |
| }, |
| { |
| "epoch": 1.5079801871216292, |
| "grad_norm": 0.1828337460756302, |
| "learning_rate": 0.00025102087469308036, |
| "loss": 2.1994, |
| "step": 13700 |
| }, |
| { |
| "epoch": 1.5134837644468906, |
| "grad_norm": 0.17188578844070435, |
| "learning_rate": 0.0002506165894439628, |
| "loss": 2.2087, |
| "step": 13750 |
| }, |
| { |
| "epoch": 1.518987341772152, |
| "grad_norm": 0.17914508283138275, |
| "learning_rate": 0.00025021097102070786, |
| "loss": 2.2266, |
| "step": 13800 |
| }, |
| { |
| "epoch": 1.5244909190974134, |
| "grad_norm": 0.17482663691043854, |
| "learning_rate": 0.0002498040247977769, |
| "loss": 2.2431, |
| "step": 13850 |
| }, |
| { |
| "epoch": 1.5299944964226748, |
| "grad_norm": 0.17342698574066162, |
| "learning_rate": 0.00024939575616722505, |
| "loss": 2.2045, |
| "step": 13900 |
| }, |
| { |
| "epoch": 1.5354980737479362, |
| "grad_norm": 0.18258632719516754, |
| "learning_rate": 0.00024898617053862904, |
| "loss": 2.1843, |
| "step": 13950 |
| }, |
| { |
| "epoch": 1.5410016510731976, |
| "grad_norm": 0.16960634291172028, |
| "learning_rate": 0.00024857527333901625, |
| "loss": 2.2096, |
| "step": 14000 |
| }, |
| { |
| "epoch": 1.546505228398459, |
| "grad_norm": 0.17008773982524872, |
| "learning_rate": 0.00024816307001279214, |
| "loss": 2.1522, |
| "step": 14050 |
| }, |
| { |
| "epoch": 1.5520088057237205, |
| "grad_norm": 0.16970837116241455, |
| "learning_rate": 0.00024774956602166857, |
| "loss": 2.2055, |
| "step": 14100 |
| }, |
| { |
| "epoch": 1.5575123830489819, |
| "grad_norm": 0.17486636340618134, |
| "learning_rate": 0.0002473347668445912, |
| "loss": 2.1906, |
| "step": 14150 |
| }, |
| { |
| "epoch": 1.5630159603742433, |
| "grad_norm": 0.19200065732002258, |
| "learning_rate": 0.00024691867797766684, |
| "loss": 2.1839, |
| "step": 14200 |
| }, |
| { |
| "epoch": 1.5685195376995047, |
| "grad_norm": 0.19441814720630646, |
| "learning_rate": 0.00024650130493409104, |
| "loss": 2.2121, |
| "step": 14250 |
| }, |
| { |
| "epoch": 1.574023115024766, |
| "grad_norm": 0.16507557034492493, |
| "learning_rate": 0.0002460826532440745, |
| "loss": 2.2016, |
| "step": 14300 |
| }, |
| { |
| "epoch": 1.5795266923500275, |
| "grad_norm": 0.16732299327850342, |
| "learning_rate": 0.00024566272845477014, |
| "loss": 2.2362, |
| "step": 14350 |
| }, |
| { |
| "epoch": 1.585030269675289, |
| "grad_norm": 0.18136867880821228, |
| "learning_rate": 0.00024524153613019947, |
| "loss": 2.193, |
| "step": 14400 |
| }, |
| { |
| "epoch": 1.5905338470005503, |
| "grad_norm": 0.16409073770046234, |
| "learning_rate": 0.0002448190818511789, |
| "loss": 2.1589, |
| "step": 14450 |
| }, |
| { |
| "epoch": 1.5960374243258117, |
| "grad_norm": 0.17381897568702698, |
| "learning_rate": 0.00024439537121524583, |
| "loss": 2.1834, |
| "step": 14500 |
| }, |
| { |
| "epoch": 1.6015410016510732, |
| "grad_norm": 0.1628628373146057, |
| "learning_rate": 0.00024397040983658436, |
| "loss": 2.172, |
| "step": 14550 |
| }, |
| { |
| "epoch": 1.6070445789763346, |
| "grad_norm": 0.16793246567249298, |
| "learning_rate": 0.00024354420334595105, |
| "loss": 2.176, |
| "step": 14600 |
| }, |
| { |
| "epoch": 1.612548156301596, |
| "grad_norm": 0.17394863069057465, |
| "learning_rate": 0.0002431167573906002, |
| "loss": 2.199, |
| "step": 14650 |
| }, |
| { |
| "epoch": 1.6180517336268574, |
| "grad_norm": 0.1554850935935974, |
| "learning_rate": 0.00024268807763420914, |
| "loss": 2.2042, |
| "step": 14700 |
| }, |
| { |
| "epoch": 1.6235553109521188, |
| "grad_norm": 0.17003560066223145, |
| "learning_rate": 0.00024225816975680312, |
| "loss": 2.1717, |
| "step": 14750 |
| }, |
| { |
| "epoch": 1.6290588882773802, |
| "grad_norm": 0.17207174003124237, |
| "learning_rate": 0.0002418270394546799, |
| "loss": 2.1894, |
| "step": 14800 |
| }, |
| { |
| "epoch": 1.6345624656026416, |
| "grad_norm": 0.1507563292980194, |
| "learning_rate": 0.00024139469244033467, |
| "loss": 2.2101, |
| "step": 14850 |
| }, |
| { |
| "epoch": 1.640066042927903, |
| "grad_norm": 0.17063497006893158, |
| "learning_rate": 0.0002409611344423838, |
| "loss": 2.1854, |
| "step": 14900 |
| }, |
| { |
| "epoch": 1.6455696202531644, |
| "grad_norm": 0.17093567550182343, |
| "learning_rate": 0.00024052637120548954, |
| "loss": 2.2028, |
| "step": 14950 |
| }, |
| { |
| "epoch": 1.6510731975784259, |
| "grad_norm": 0.17256326973438263, |
| "learning_rate": 0.00024009040849028341, |
| "loss": 2.2102, |
| "step": 15000 |
| }, |
| { |
| "epoch": 1.6565767749036873, |
| "grad_norm": 0.17648595571517944, |
| "learning_rate": 0.00023965325207329016, |
| "loss": 2.2094, |
| "step": 15050 |
| }, |
| { |
| "epoch": 1.6620803522289487, |
| "grad_norm": 0.17140266299247742, |
| "learning_rate": 0.0002392149077468511, |
| "loss": 2.1936, |
| "step": 15100 |
| }, |
| { |
| "epoch": 1.66758392955421, |
| "grad_norm": 0.16277456283569336, |
| "learning_rate": 0.00023877538131904743, |
| "loss": 2.1705, |
| "step": 15150 |
| }, |
| { |
| "epoch": 1.6730875068794715, |
| "grad_norm": 0.16912780702114105, |
| "learning_rate": 0.00023833467861362323, |
| "loss": 2.1834, |
| "step": 15200 |
| }, |
| { |
| "epoch": 1.6785910842047331, |
| "grad_norm": 0.1684638261795044, |
| "learning_rate": 0.00023789280546990824, |
| "loss": 2.2044, |
| "step": 15250 |
| }, |
| { |
| "epoch": 1.6840946615299945, |
| "grad_norm": 0.16793768107891083, |
| "learning_rate": 0.00023744976774274065, |
| "loss": 2.188, |
| "step": 15300 |
| }, |
| { |
| "epoch": 1.689598238855256, |
| "grad_norm": 0.15743038058280945, |
| "learning_rate": 0.00023700557130238944, |
| "loss": 2.183, |
| "step": 15350 |
| }, |
| { |
| "epoch": 1.6951018161805174, |
| "grad_norm": 0.1788504421710968, |
| "learning_rate": 0.0002365602220344765, |
| "loss": 2.1923, |
| "step": 15400 |
| }, |
| { |
| "epoch": 1.7006053935057788, |
| "grad_norm": 0.1614784598350525, |
| "learning_rate": 0.00023611372583989883, |
| "loss": 2.196, |
| "step": 15450 |
| }, |
| { |
| "epoch": 1.7061089708310402, |
| "grad_norm": 0.16596226394176483, |
| "learning_rate": 0.0002356660886347502, |
| "loss": 2.1952, |
| "step": 15500 |
| }, |
| { |
| "epoch": 1.7116125481563016, |
| "grad_norm": 0.18096621334552765, |
| "learning_rate": 0.00023521731635024274, |
| "loss": 2.1858, |
| "step": 15550 |
| }, |
| { |
| "epoch": 1.717116125481563, |
| "grad_norm": 0.1885058581829071, |
| "learning_rate": 0.0002347674149326286, |
| "loss": 2.1894, |
| "step": 15600 |
| }, |
| { |
| "epoch": 1.7226197028068244, |
| "grad_norm": 0.15273383259773254, |
| "learning_rate": 0.00023431639034312096, |
| "loss": 2.1896, |
| "step": 15650 |
| }, |
| { |
| "epoch": 1.7281232801320858, |
| "grad_norm": 0.17447490990161896, |
| "learning_rate": 0.00023386424855781495, |
| "loss": 2.1896, |
| "step": 15700 |
| }, |
| { |
| "epoch": 1.7336268574573472, |
| "grad_norm": 0.16734521090984344, |
| "learning_rate": 0.00023341099556760866, |
| "loss": 2.1886, |
| "step": 15750 |
| }, |
| { |
| "epoch": 1.7391304347826086, |
| "grad_norm": 0.17426900565624237, |
| "learning_rate": 0.0002329566373781238, |
| "loss": 2.1838, |
| "step": 15800 |
| }, |
| { |
| "epoch": 1.7446340121078703, |
| "grad_norm": 0.1640431135892868, |
| "learning_rate": 0.00023250118000962583, |
| "loss": 2.2026, |
| "step": 15850 |
| }, |
| { |
| "epoch": 1.7501375894331317, |
| "grad_norm": 0.16575609147548676, |
| "learning_rate": 0.00023204462949694447, |
| "loss": 2.1721, |
| "step": 15900 |
| }, |
| { |
| "epoch": 1.755641166758393, |
| "grad_norm": 0.17576967179775238, |
| "learning_rate": 0.00023158699188939371, |
| "loss": 2.1517, |
| "step": 15950 |
| }, |
| { |
| "epoch": 1.7611447440836545, |
| "grad_norm": 0.1661233901977539, |
| "learning_rate": 0.00023112827325069154, |
| "loss": 2.1594, |
| "step": 16000 |
| }, |
| { |
| "epoch": 1.766648321408916, |
| "grad_norm": 0.17433176934719086, |
| "learning_rate": 0.00023066847965887954, |
| "loss": 2.2069, |
| "step": 16050 |
| }, |
| { |
| "epoch": 1.7721518987341773, |
| "grad_norm": 0.164288729429245, |
| "learning_rate": 0.0002302076172062427, |
| "loss": 2.1945, |
| "step": 16100 |
| }, |
| { |
| "epoch": 1.7776554760594387, |
| "grad_norm": 0.16619046032428741, |
| "learning_rate": 0.00022974569199922827, |
| "loss": 2.1678, |
| "step": 16150 |
| }, |
| { |
| "epoch": 1.7831590533847002, |
| "grad_norm": 0.1572977751493454, |
| "learning_rate": 0.00022928271015836514, |
| "loss": 2.1834, |
| "step": 16200 |
| }, |
| { |
| "epoch": 1.7886626307099616, |
| "grad_norm": 0.15041351318359375, |
| "learning_rate": 0.0002288186778181826, |
| "loss": 2.1372, |
| "step": 16250 |
| }, |
| { |
| "epoch": 1.794166208035223, |
| "grad_norm": 0.17850619554519653, |
| "learning_rate": 0.00022835360112712913, |
| "loss": 2.2064, |
| "step": 16300 |
| }, |
| { |
| "epoch": 1.7996697853604844, |
| "grad_norm": 0.1657785028219223, |
| "learning_rate": 0.00022788748624749083, |
| "loss": 2.1948, |
| "step": 16350 |
| }, |
| { |
| "epoch": 1.8051733626857458, |
| "grad_norm": 0.16514721512794495, |
| "learning_rate": 0.00022742033935530995, |
| "loss": 2.209, |
| "step": 16400 |
| }, |
| { |
| "epoch": 1.8106769400110072, |
| "grad_norm": 0.1726425439119339, |
| "learning_rate": 0.00022695216664030297, |
| "loss": 2.1992, |
| "step": 16450 |
| }, |
| { |
| "epoch": 1.8161805173362686, |
| "grad_norm": 0.1753402054309845, |
| "learning_rate": 0.00022648297430577835, |
| "loss": 2.1929, |
| "step": 16500 |
| }, |
| { |
| "epoch": 1.82168409466153, |
| "grad_norm": 0.18259774148464203, |
| "learning_rate": 0.00022601276856855482, |
| "loss": 2.2083, |
| "step": 16550 |
| }, |
| { |
| "epoch": 1.8271876719867914, |
| "grad_norm": 0.16030113399028778, |
| "learning_rate": 0.0002255415556588786, |
| "loss": 2.1571, |
| "step": 16600 |
| }, |
| { |
| "epoch": 1.8326912493120529, |
| "grad_norm": 0.1431453675031662, |
| "learning_rate": 0.000225069341820341, |
| "loss": 2.149, |
| "step": 16650 |
| }, |
| { |
| "epoch": 1.8381948266373143, |
| "grad_norm": 0.1664692461490631, |
| "learning_rate": 0.0002245961333097956, |
| "loss": 2.1864, |
| "step": 16700 |
| }, |
| { |
| "epoch": 1.8436984039625757, |
| "grad_norm": 0.17130005359649658, |
| "learning_rate": 0.00022412193639727564, |
| "loss": 2.18, |
| "step": 16750 |
| }, |
| { |
| "epoch": 1.849201981287837, |
| "grad_norm": 0.15278129279613495, |
| "learning_rate": 0.00022364675736591058, |
| "loss": 2.1648, |
| "step": 16800 |
| }, |
| { |
| "epoch": 1.8547055586130985, |
| "grad_norm": 0.176371768116951, |
| "learning_rate": 0.000223170602511843, |
| "loss": 2.1582, |
| "step": 16850 |
| }, |
| { |
| "epoch": 1.86020913593836, |
| "grad_norm": 0.15902136266231537, |
| "learning_rate": 0.0002226934781441452, |
| "loss": 2.183, |
| "step": 16900 |
| }, |
| { |
| "epoch": 1.8657127132636213, |
| "grad_norm": 0.14821839332580566, |
| "learning_rate": 0.00022221539058473553, |
| "loss": 2.172, |
| "step": 16950 |
| }, |
| { |
| "epoch": 1.8712162905888827, |
| "grad_norm": 0.15611177682876587, |
| "learning_rate": 0.00022173634616829481, |
| "loss": 2.1715, |
| "step": 17000 |
| }, |
| { |
| "epoch": 1.8767198679141441, |
| "grad_norm": 0.1633993536233902, |
| "learning_rate": 0.00022125635124218207, |
| "loss": 2.1961, |
| "step": 17050 |
| }, |
| { |
| "epoch": 1.8822234452394055, |
| "grad_norm": 0.16087235510349274, |
| "learning_rate": 0.00022077541216635076, |
| "loss": 2.1963, |
| "step": 17100 |
| }, |
| { |
| "epoch": 1.887727022564667, |
| "grad_norm": 0.16733358800411224, |
| "learning_rate": 0.00022029353531326432, |
| "loss": 2.1605, |
| "step": 17150 |
| }, |
| { |
| "epoch": 1.8932305998899284, |
| "grad_norm": 0.16267576813697815, |
| "learning_rate": 0.00021981072706781185, |
| "loss": 2.1554, |
| "step": 17200 |
| }, |
| { |
| "epoch": 1.8987341772151898, |
| "grad_norm": 0.16028568148612976, |
| "learning_rate": 0.0002193269938272232, |
| "loss": 2.1793, |
| "step": 17250 |
| }, |
| { |
| "epoch": 1.9042377545404512, |
| "grad_norm": 0.15961337089538574, |
| "learning_rate": 0.00021884234200098473, |
| "loss": 2.1673, |
| "step": 17300 |
| }, |
| { |
| "epoch": 1.9097413318657126, |
| "grad_norm": 0.1637711226940155, |
| "learning_rate": 0.00021835677801075397, |
| "loss": 2.1676, |
| "step": 17350 |
| }, |
| { |
| "epoch": 1.915244909190974, |
| "grad_norm": 0.16490189731121063, |
| "learning_rate": 0.00021787030829027468, |
| "loss": 2.2119, |
| "step": 17400 |
| }, |
| { |
| "epoch": 1.9207484865162354, |
| "grad_norm": 0.17354600131511688, |
| "learning_rate": 0.00021738293928529154, |
| "loss": 2.2001, |
| "step": 17450 |
| }, |
| { |
| "epoch": 1.9262520638414968, |
| "grad_norm": 0.16645126044750214, |
| "learning_rate": 0.00021689467745346483, |
| "loss": 2.1915, |
| "step": 17500 |
| }, |
| { |
| "epoch": 1.9317556411667582, |
| "grad_norm": 0.16187366843223572, |
| "learning_rate": 0.0002164055292642849, |
| "loss": 2.1829, |
| "step": 17550 |
| }, |
| { |
| "epoch": 1.9372592184920197, |
| "grad_norm": 0.17859427630901337, |
| "learning_rate": 0.00021591550119898616, |
| "loss": 2.1203, |
| "step": 17600 |
| }, |
| { |
| "epoch": 1.9427627958172813, |
| "grad_norm": 0.15773439407348633, |
| "learning_rate": 0.00021542459975046164, |
| "loss": 2.1535, |
| "step": 17650 |
| }, |
| { |
| "epoch": 1.9482663731425427, |
| "grad_norm": 0.1696433126926422, |
| "learning_rate": 0.00021493283142317668, |
| "loss": 2.1271, |
| "step": 17700 |
| }, |
| { |
| "epoch": 1.953769950467804, |
| "grad_norm": 0.16615457832813263, |
| "learning_rate": 0.00021444020273308267, |
| "loss": 2.1688, |
| "step": 17750 |
| }, |
| { |
| "epoch": 1.9592735277930655, |
| "grad_norm": 0.16561925411224365, |
| "learning_rate": 0.00021394672020753102, |
| "loss": 2.1228, |
| "step": 17800 |
| }, |
| { |
| "epoch": 1.964777105118327, |
| "grad_norm": 0.1655699759721756, |
| "learning_rate": 0.00021345239038518638, |
| "loss": 2.1614, |
| "step": 17850 |
| }, |
| { |
| "epoch": 1.9702806824435883, |
| "grad_norm": 0.16106578707695007, |
| "learning_rate": 0.0002129572198159401, |
| "loss": 2.1471, |
| "step": 17900 |
| }, |
| { |
| "epoch": 1.9757842597688497, |
| "grad_norm": 0.1635727882385254, |
| "learning_rate": 0.00021246121506082367, |
| "loss": 2.184, |
| "step": 17950 |
| }, |
| { |
| "epoch": 1.9812878370941112, |
| "grad_norm": 0.15699493885040283, |
| "learning_rate": 0.0002119643826919213, |
| "loss": 2.1602, |
| "step": 18000 |
| }, |
| { |
| "epoch": 1.9867914144193726, |
| "grad_norm": 0.15784280002117157, |
| "learning_rate": 0.00021146672929228327, |
| "loss": 2.1781, |
| "step": 18050 |
| }, |
| { |
| "epoch": 1.992294991744634, |
| "grad_norm": 0.15157456696033478, |
| "learning_rate": 0.00021096826145583852, |
| "loss": 2.1532, |
| "step": 18100 |
| }, |
| { |
| "epoch": 1.9977985690698954, |
| "grad_norm": 0.15633012354373932, |
| "learning_rate": 0.0002104689857873074, |
| "loss": 2.1676, |
| "step": 18150 |
| }, |
| { |
| "epoch": 2.003302146395157, |
| "grad_norm": 0.16586118936538696, |
| "learning_rate": 0.00020996890890211386, |
| "loss": 2.1335, |
| "step": 18200 |
| }, |
| { |
| "epoch": 2.0088057237204184, |
| "grad_norm": 0.1573173552751541, |
| "learning_rate": 0.00020946803742629826, |
| "loss": 2.11, |
| "step": 18250 |
| }, |
| { |
| "epoch": 2.01430930104568, |
| "grad_norm": 0.18040034174919128, |
| "learning_rate": 0.00020896637799642915, |
| "loss": 2.1128, |
| "step": 18300 |
| }, |
| { |
| "epoch": 2.0198128783709413, |
| "grad_norm": 0.1664518266916275, |
| "learning_rate": 0.00020846393725951556, |
| "loss": 2.1384, |
| "step": 18350 |
| }, |
| { |
| "epoch": 2.0253164556962027, |
| "grad_norm": 0.17078042030334473, |
| "learning_rate": 0.0002079607218729188, |
| "loss": 2.0909, |
| "step": 18400 |
| }, |
| { |
| "epoch": 2.030820033021464, |
| "grad_norm": 0.15759077668190002, |
| "learning_rate": 0.00020745673850426446, |
| "loss": 2.1015, |
| "step": 18450 |
| }, |
| { |
| "epoch": 2.0363236103467255, |
| "grad_norm": 0.17311988770961761, |
| "learning_rate": 0.00020695199383135386, |
| "loss": 2.1436, |
| "step": 18500 |
| }, |
| { |
| "epoch": 2.041827187671987, |
| "grad_norm": 0.1612362265586853, |
| "learning_rate": 0.00020644649454207553, |
| "loss": 2.0868, |
| "step": 18550 |
| }, |
| { |
| "epoch": 2.0473307649972483, |
| "grad_norm": 0.16166241466999054, |
| "learning_rate": 0.0002059402473343169, |
| "loss": 2.0683, |
| "step": 18600 |
| }, |
| { |
| "epoch": 2.0528343423225097, |
| "grad_norm": 0.17205455899238586, |
| "learning_rate": 0.00020543325891587516, |
| "loss": 2.0618, |
| "step": 18650 |
| }, |
| { |
| "epoch": 2.058337919647771, |
| "grad_norm": 0.17534597218036652, |
| "learning_rate": 0.00020492553600436872, |
| "loss": 2.1173, |
| "step": 18700 |
| }, |
| { |
| "epoch": 2.0638414969730325, |
| "grad_norm": 0.15081949532032013, |
| "learning_rate": 0.00020441708532714797, |
| "loss": 2.0991, |
| "step": 18750 |
| }, |
| { |
| "epoch": 2.069345074298294, |
| "grad_norm": 0.15837198495864868, |
| "learning_rate": 0.00020390791362120625, |
| "loss": 2.1332, |
| "step": 18800 |
| }, |
| { |
| "epoch": 2.0748486516235554, |
| "grad_norm": 0.1598348468542099, |
| "learning_rate": 0.00020339802763309054, |
| "loss": 2.1164, |
| "step": 18850 |
| }, |
| { |
| "epoch": 2.0803522289488168, |
| "grad_norm": 0.1536133736371994, |
| "learning_rate": 0.0002028874341188122, |
| "loss": 2.1169, |
| "step": 18900 |
| }, |
| { |
| "epoch": 2.085855806274078, |
| "grad_norm": 0.1654282808303833, |
| "learning_rate": 0.00020237613984375711, |
| "loss": 2.1253, |
| "step": 18950 |
| }, |
| { |
| "epoch": 2.0913593835993396, |
| "grad_norm": 0.16015826165676117, |
| "learning_rate": 0.00020186415158259652, |
| "loss": 2.1196, |
| "step": 19000 |
| }, |
| { |
| "epoch": 2.096862960924601, |
| "grad_norm": 0.16205786168575287, |
| "learning_rate": 0.00020135147611919686, |
| "loss": 2.0872, |
| "step": 19050 |
| }, |
| { |
| "epoch": 2.1023665382498624, |
| "grad_norm": 0.16814254224300385, |
| "learning_rate": 0.00020083812024653013, |
| "loss": 2.1569, |
| "step": 19100 |
| }, |
| { |
| "epoch": 2.107870115575124, |
| "grad_norm": 0.15433181822299957, |
| "learning_rate": 0.00020032409076658364, |
| "loss": 2.0967, |
| "step": 19150 |
| }, |
| { |
| "epoch": 2.1133736929003852, |
| "grad_norm": 0.16261817514896393, |
| "learning_rate": 0.00019980939449027016, |
| "loss": 2.0938, |
| "step": 19200 |
| }, |
| { |
| "epoch": 2.1188772702256466, |
| "grad_norm": 0.17877838015556335, |
| "learning_rate": 0.00019929403823733748, |
| "loss": 2.1399, |
| "step": 19250 |
| }, |
| { |
| "epoch": 2.124380847550908, |
| "grad_norm": 0.178861603140831, |
| "learning_rate": 0.00019877802883627808, |
| "loss": 2.1362, |
| "step": 19300 |
| }, |
| { |
| "epoch": 2.1298844248761695, |
| "grad_norm": 0.1621016561985016, |
| "learning_rate": 0.00019826137312423873, |
| "loss": 2.1089, |
| "step": 19350 |
| }, |
| { |
| "epoch": 2.135388002201431, |
| "grad_norm": 0.17772269248962402, |
| "learning_rate": 0.00019774407794692986, |
| "loss": 2.1278, |
| "step": 19400 |
| }, |
| { |
| "epoch": 2.1408915795266923, |
| "grad_norm": 0.16339927911758423, |
| "learning_rate": 0.00019722615015853483, |
| "loss": 2.1244, |
| "step": 19450 |
| }, |
| { |
| "epoch": 2.1463951568519537, |
| "grad_norm": 0.16879414021968842, |
| "learning_rate": 0.0001967075966216191, |
| "loss": 2.106, |
| "step": 19500 |
| }, |
| { |
| "epoch": 2.151898734177215, |
| "grad_norm": 0.15002557635307312, |
| "learning_rate": 0.00019618842420703923, |
| "loss": 2.1007, |
| "step": 19550 |
| }, |
| { |
| "epoch": 2.1574023115024765, |
| "grad_norm": 0.17130738496780396, |
| "learning_rate": 0.0001956686397938522, |
| "loss": 2.1104, |
| "step": 19600 |
| }, |
| { |
| "epoch": 2.162905888827738, |
| "grad_norm": 0.16796913743019104, |
| "learning_rate": 0.00019514825026922377, |
| "loss": 2.1278, |
| "step": 19650 |
| }, |
| { |
| "epoch": 2.1684094661529993, |
| "grad_norm": 0.19111864268779755, |
| "learning_rate": 0.0001946272625283375, |
| "loss": 2.1402, |
| "step": 19700 |
| }, |
| { |
| "epoch": 2.1739130434782608, |
| "grad_norm": 0.17543087899684906, |
| "learning_rate": 0.00019410568347430344, |
| "loss": 2.0984, |
| "step": 19750 |
| }, |
| { |
| "epoch": 2.179416620803522, |
| "grad_norm": 0.15795323252677917, |
| "learning_rate": 0.00019358352001806642, |
| "loss": 2.1125, |
| "step": 19800 |
| }, |
| { |
| "epoch": 2.1849201981287836, |
| "grad_norm": 0.15581749379634857, |
| "learning_rate": 0.0001930607790783147, |
| "loss": 2.119, |
| "step": 19850 |
| }, |
| { |
| "epoch": 2.190423775454045, |
| "grad_norm": 0.17086704075336456, |
| "learning_rate": 0.00019253746758138822, |
| "loss": 2.1029, |
| "step": 19900 |
| }, |
| { |
| "epoch": 2.1959273527793064, |
| "grad_norm": 0.15870486199855804, |
| "learning_rate": 0.00019201359246118682, |
| "loss": 2.1133, |
| "step": 19950 |
| }, |
| { |
| "epoch": 2.201430930104568, |
| "grad_norm": 0.16840116679668427, |
| "learning_rate": 0.0001914891606590784, |
| "loss": 2.1196, |
| "step": 20000 |
| }, |
| { |
| "epoch": 2.206934507429829, |
| "grad_norm": 0.1643972396850586, |
| "learning_rate": 0.00019096417912380686, |
| "loss": 2.1173, |
| "step": 20050 |
| }, |
| { |
| "epoch": 2.2124380847550906, |
| "grad_norm": 0.1598280966281891, |
| "learning_rate": 0.0001904386548114001, |
| "loss": 2.1338, |
| "step": 20100 |
| }, |
| { |
| "epoch": 2.217941662080352, |
| "grad_norm": 0.16909152269363403, |
| "learning_rate": 0.0001899125946850779, |
| "loss": 2.0987, |
| "step": 20150 |
| }, |
| { |
| "epoch": 2.2234452394056134, |
| "grad_norm": 0.15597747266292572, |
| "learning_rate": 0.00018938600571515962, |
| "loss": 2.1295, |
| "step": 20200 |
| }, |
| { |
| "epoch": 2.228948816730875, |
| "grad_norm": 0.16759732365608215, |
| "learning_rate": 0.00018885889487897162, |
| "loss": 2.1374, |
| "step": 20250 |
| }, |
| { |
| "epoch": 2.2344523940561363, |
| "grad_norm": 0.15526509284973145, |
| "learning_rate": 0.00018833126916075526, |
| "loss": 2.1287, |
| "step": 20300 |
| }, |
| { |
| "epoch": 2.2399559713813977, |
| "grad_norm": 0.16562257707118988, |
| "learning_rate": 0.00018780313555157393, |
| "loss": 2.126, |
| "step": 20350 |
| }, |
| { |
| "epoch": 2.2454595487066595, |
| "grad_norm": 0.17471948266029358, |
| "learning_rate": 0.00018727450104922075, |
| "loss": 2.1175, |
| "step": 20400 |
| }, |
| { |
| "epoch": 2.2509631260319205, |
| "grad_norm": 0.16391079127788544, |
| "learning_rate": 0.00018674537265812548, |
| "loss": 2.0965, |
| "step": 20450 |
| }, |
| { |
| "epoch": 2.2564667033571824, |
| "grad_norm": 0.1693831831216812, |
| "learning_rate": 0.00018621575738926217, |
| "loss": 2.1124, |
| "step": 20500 |
| }, |
| { |
| "epoch": 2.2619702806824438, |
| "grad_norm": 0.16163313388824463, |
| "learning_rate": 0.00018568566226005587, |
| "loss": 2.1034, |
| "step": 20550 |
| }, |
| { |
| "epoch": 2.267473858007705, |
| "grad_norm": 0.1676999181509018, |
| "learning_rate": 0.00018515509429428988, |
| "loss": 2.0991, |
| "step": 20600 |
| }, |
| { |
| "epoch": 2.2729774353329666, |
| "grad_norm": 0.16531667113304138, |
| "learning_rate": 0.00018462406052201252, |
| "loss": 2.1183, |
| "step": 20650 |
| }, |
| { |
| "epoch": 2.278481012658228, |
| "grad_norm": 0.16208726167678833, |
| "learning_rate": 0.00018409256797944412, |
| "loss": 2.0981, |
| "step": 20700 |
| }, |
| { |
| "epoch": 2.2839845899834894, |
| "grad_norm": 0.1707569658756256, |
| "learning_rate": 0.0001835606237088837, |
| "loss": 2.0926, |
| "step": 20750 |
| }, |
| { |
| "epoch": 2.289488167308751, |
| "grad_norm": 0.16191871464252472, |
| "learning_rate": 0.00018302823475861582, |
| "loss": 2.0895, |
| "step": 20800 |
| }, |
| { |
| "epoch": 2.2949917446340122, |
| "grad_norm": 0.154753640294075, |
| "learning_rate": 0.0001824954081828168, |
| "loss": 2.0972, |
| "step": 20850 |
| }, |
| { |
| "epoch": 2.3004953219592736, |
| "grad_norm": 0.16868236660957336, |
| "learning_rate": 0.00018196215104146187, |
| "loss": 2.1009, |
| "step": 20900 |
| }, |
| { |
| "epoch": 2.305998899284535, |
| "grad_norm": 0.1602269858121872, |
| "learning_rate": 0.00018142847040023097, |
| "loss": 2.1063, |
| "step": 20950 |
| }, |
| { |
| "epoch": 2.3115024766097965, |
| "grad_norm": 0.17220260202884674, |
| "learning_rate": 0.00018089437333041562, |
| "loss": 2.1049, |
| "step": 21000 |
| }, |
| { |
| "epoch": 2.317006053935058, |
| "grad_norm": 0.14858907461166382, |
| "learning_rate": 0.0001803598669088249, |
| "loss": 2.1129, |
| "step": 21050 |
| }, |
| { |
| "epoch": 2.3225096312603193, |
| "grad_norm": 0.17586296796798706, |
| "learning_rate": 0.0001798249582176919, |
| "loss": 2.0961, |
| "step": 21100 |
| }, |
| { |
| "epoch": 2.3280132085855807, |
| "grad_norm": 0.14931316673755646, |
| "learning_rate": 0.00017928965434457983, |
| "loss": 2.1108, |
| "step": 21150 |
| }, |
| { |
| "epoch": 2.333516785910842, |
| "grad_norm": 0.17148137092590332, |
| "learning_rate": 0.00017875396238228798, |
| "loss": 2.1029, |
| "step": 21200 |
| }, |
| { |
| "epoch": 2.3390203632361035, |
| "grad_norm": 0.15888847410678864, |
| "learning_rate": 0.00017821788942875793, |
| "loss": 2.1258, |
| "step": 21250 |
| }, |
| { |
| "epoch": 2.344523940561365, |
| "grad_norm": 0.1749711036682129, |
| "learning_rate": 0.0001776814425869793, |
| "loss": 2.1176, |
| "step": 21300 |
| }, |
| { |
| "epoch": 2.3500275178866263, |
| "grad_norm": 0.16511616110801697, |
| "learning_rate": 0.0001771446289648959, |
| "loss": 2.0992, |
| "step": 21350 |
| }, |
| { |
| "epoch": 2.3555310952118877, |
| "grad_norm": 0.16869331896305084, |
| "learning_rate": 0.00017660745567531124, |
| "loss": 2.1274, |
| "step": 21400 |
| }, |
| { |
| "epoch": 2.361034672537149, |
| "grad_norm": 0.16024701297283173, |
| "learning_rate": 0.00017606992983579457, |
| "loss": 2.1379, |
| "step": 21450 |
| }, |
| { |
| "epoch": 2.3665382498624106, |
| "grad_norm": 0.1646272987127304, |
| "learning_rate": 0.00017553205856858637, |
| "loss": 2.1365, |
| "step": 21500 |
| }, |
| { |
| "epoch": 2.372041827187672, |
| "grad_norm": 0.1765459179878235, |
| "learning_rate": 0.00017499384900050404, |
| "loss": 2.1125, |
| "step": 21550 |
| }, |
| { |
| "epoch": 2.3775454045129334, |
| "grad_norm": 0.15617190301418304, |
| "learning_rate": 0.0001744553082628475, |
| "loss": 2.0985, |
| "step": 21600 |
| }, |
| { |
| "epoch": 2.383048981838195, |
| "grad_norm": 0.14687784016132355, |
| "learning_rate": 0.00017391644349130467, |
| "loss": 2.0771, |
| "step": 21650 |
| }, |
| { |
| "epoch": 2.388552559163456, |
| "grad_norm": 0.16840553283691406, |
| "learning_rate": 0.00017337726182585698, |
| "loss": 2.0909, |
| "step": 21700 |
| }, |
| { |
| "epoch": 2.3940561364887176, |
| "grad_norm": 0.16857478022575378, |
| "learning_rate": 0.00017283777041068472, |
| "loss": 2.0968, |
| "step": 21750 |
| }, |
| { |
| "epoch": 2.399559713813979, |
| "grad_norm": 0.1799498200416565, |
| "learning_rate": 0.00017229797639407233, |
| "loss": 2.1023, |
| "step": 21800 |
| }, |
| { |
| "epoch": 2.4050632911392404, |
| "grad_norm": 0.15700192749500275, |
| "learning_rate": 0.00017175788692831367, |
| "loss": 2.0986, |
| "step": 21850 |
| }, |
| { |
| "epoch": 2.410566868464502, |
| "grad_norm": 0.16778048872947693, |
| "learning_rate": 0.00017121750916961742, |
| "loss": 2.0914, |
| "step": 21900 |
| }, |
| { |
| "epoch": 2.4160704457897633, |
| "grad_norm": 0.15119898319244385, |
| "learning_rate": 0.0001706768502780122, |
| "loss": 2.0857, |
| "step": 21950 |
| }, |
| { |
| "epoch": 2.4215740231150247, |
| "grad_norm": 0.17003406584262848, |
| "learning_rate": 0.0001701359174172515, |
| "loss": 2.1366, |
| "step": 22000 |
| }, |
| { |
| "epoch": 2.427077600440286, |
| "grad_norm": 0.17182987928390503, |
| "learning_rate": 0.000169594717754719, |
| "loss": 2.1126, |
| "step": 22050 |
| }, |
| { |
| "epoch": 2.4325811777655475, |
| "grad_norm": 0.16026781499385834, |
| "learning_rate": 0.00016905325846133354, |
| "loss": 2.076, |
| "step": 22100 |
| }, |
| { |
| "epoch": 2.438084755090809, |
| "grad_norm": 0.16470612585544586, |
| "learning_rate": 0.00016851154671145403, |
| "loss": 2.0907, |
| "step": 22150 |
| }, |
| { |
| "epoch": 2.4435883324160703, |
| "grad_norm": 0.14621587097644806, |
| "learning_rate": 0.0001679695896827844, |
| "loss": 2.1073, |
| "step": 22200 |
| }, |
| { |
| "epoch": 2.4490919097413317, |
| "grad_norm": 0.1563323438167572, |
| "learning_rate": 0.00016742739455627862, |
| "loss": 2.0757, |
| "step": 22250 |
| }, |
| { |
| "epoch": 2.454595487066593, |
| "grad_norm": 0.17763760685920715, |
| "learning_rate": 0.0001668849685160455, |
| "loss": 2.0705, |
| "step": 22300 |
| }, |
| { |
| "epoch": 2.4600990643918546, |
| "grad_norm": 0.16651757061481476, |
| "learning_rate": 0.00016634231874925335, |
| "loss": 2.0881, |
| "step": 22350 |
| }, |
| { |
| "epoch": 2.465602641717116, |
| "grad_norm": 0.16752833127975464, |
| "learning_rate": 0.0001657994524460349, |
| "loss": 2.0639, |
| "step": 22400 |
| }, |
| { |
| "epoch": 2.4711062190423774, |
| "grad_norm": 0.14574387669563293, |
| "learning_rate": 0.000165256376799392, |
| "loss": 2.087, |
| "step": 22450 |
| }, |
| { |
| "epoch": 2.476609796367639, |
| "grad_norm": 0.15595164895057678, |
| "learning_rate": 0.00016471309900510043, |
| "loss": 2.0965, |
| "step": 22500 |
| }, |
| { |
| "epoch": 2.4821133736929, |
| "grad_norm": 0.17218752205371857, |
| "learning_rate": 0.0001641696262616142, |
| "loss": 2.0943, |
| "step": 22550 |
| }, |
| { |
| "epoch": 2.487616951018162, |
| "grad_norm": 0.1563566029071808, |
| "learning_rate": 0.00016362596576997057, |
| "loss": 2.0946, |
| "step": 22600 |
| }, |
| { |
| "epoch": 2.4931205283434235, |
| "grad_norm": 0.16190436482429504, |
| "learning_rate": 0.00016308212473369453, |
| "loss": 2.0924, |
| "step": 22650 |
| }, |
| { |
| "epoch": 2.498624105668685, |
| "grad_norm": 0.16890835762023926, |
| "learning_rate": 0.00016253811035870313, |
| "loss": 2.0954, |
| "step": 22700 |
| }, |
| { |
| "epoch": 2.5041276829939463, |
| "grad_norm": 0.16371208429336548, |
| "learning_rate": 0.00016199392985321017, |
| "loss": 2.0954, |
| "step": 22750 |
| }, |
| { |
| "epoch": 2.5096312603192077, |
| "grad_norm": 0.15898428857326508, |
| "learning_rate": 0.00016144959042763084, |
| "loss": 2.0766, |
| "step": 22800 |
| }, |
| { |
| "epoch": 2.515134837644469, |
| "grad_norm": 0.15516646206378937, |
| "learning_rate": 0.00016090509929448594, |
| "loss": 2.0891, |
| "step": 22850 |
| }, |
| { |
| "epoch": 2.5206384149697305, |
| "grad_norm": 0.15007217228412628, |
| "learning_rate": 0.00016036046366830635, |
| "loss": 2.078, |
| "step": 22900 |
| }, |
| { |
| "epoch": 2.526141992294992, |
| "grad_norm": 0.17989104986190796, |
| "learning_rate": 0.00015981569076553765, |
| "loss": 2.0888, |
| "step": 22950 |
| }, |
| { |
| "epoch": 2.5316455696202533, |
| "grad_norm": 0.16754071414470673, |
| "learning_rate": 0.00015927078780444403, |
| "loss": 2.0882, |
| "step": 23000 |
| }, |
| { |
| "epoch": 2.5371491469455147, |
| "grad_norm": 0.15897449851036072, |
| "learning_rate": 0.0001587257620050134, |
| "loss": 2.0933, |
| "step": 23050 |
| }, |
| { |
| "epoch": 2.542652724270776, |
| "grad_norm": 0.16594423353672028, |
| "learning_rate": 0.00015818062058886082, |
| "loss": 2.0968, |
| "step": 23100 |
| }, |
| { |
| "epoch": 2.5481563015960376, |
| "grad_norm": 0.1700371652841568, |
| "learning_rate": 0.00015763537077913365, |
| "loss": 2.096, |
| "step": 23150 |
| }, |
| { |
| "epoch": 2.553659878921299, |
| "grad_norm": 0.16013766825199127, |
| "learning_rate": 0.00015709001980041527, |
| "loss": 2.1134, |
| "step": 23200 |
| }, |
| { |
| "epoch": 2.5591634562465604, |
| "grad_norm": 0.1601455807685852, |
| "learning_rate": 0.00015654457487862961, |
| "loss": 2.0983, |
| "step": 23250 |
| }, |
| { |
| "epoch": 2.564667033571822, |
| "grad_norm": 0.16373537480831146, |
| "learning_rate": 0.00015599904324094534, |
| "loss": 2.0866, |
| "step": 23300 |
| }, |
| { |
| "epoch": 2.570170610897083, |
| "grad_norm": 0.15919940173625946, |
| "learning_rate": 0.0001554534321156801, |
| "loss": 2.1147, |
| "step": 23350 |
| }, |
| { |
| "epoch": 2.5756741882223446, |
| "grad_norm": 0.1523171216249466, |
| "learning_rate": 0.0001549077487322048, |
| "loss": 2.0887, |
| "step": 23400 |
| }, |
| { |
| "epoch": 2.581177765547606, |
| "grad_norm": 0.16425256431102753, |
| "learning_rate": 0.00015436200032084774, |
| "loss": 2.0918, |
| "step": 23450 |
| }, |
| { |
| "epoch": 2.5866813428728674, |
| "grad_norm": 0.16046775877475739, |
| "learning_rate": 0.0001538161941127988, |
| "loss": 2.0816, |
| "step": 23500 |
| }, |
| { |
| "epoch": 2.592184920198129, |
| "grad_norm": 0.1673419028520584, |
| "learning_rate": 0.00015327033734001375, |
| "loss": 2.1186, |
| "step": 23550 |
| }, |
| { |
| "epoch": 2.5976884975233903, |
| "grad_norm": 0.15573158860206604, |
| "learning_rate": 0.00015272443723511825, |
| "loss": 2.0721, |
| "step": 23600 |
| }, |
| { |
| "epoch": 2.6031920748486517, |
| "grad_norm": 0.16536468267440796, |
| "learning_rate": 0.00015217850103131227, |
| "loss": 2.0991, |
| "step": 23650 |
| }, |
| { |
| "epoch": 2.608695652173913, |
| "grad_norm": 0.15281961858272552, |
| "learning_rate": 0.00015163253596227393, |
| "loss": 2.0713, |
| "step": 23700 |
| }, |
| { |
| "epoch": 2.6141992294991745, |
| "grad_norm": 0.1582518070936203, |
| "learning_rate": 0.0001510865492620638, |
| "loss": 2.0923, |
| "step": 23750 |
| }, |
| { |
| "epoch": 2.619702806824436, |
| "grad_norm": 0.1581014096736908, |
| "learning_rate": 0.0001505405481650293, |
| "loss": 2.0895, |
| "step": 23800 |
| }, |
| { |
| "epoch": 2.6252063841496973, |
| "grad_norm": 0.16346633434295654, |
| "learning_rate": 0.00014999453990570839, |
| "loss": 2.1011, |
| "step": 23850 |
| }, |
| { |
| "epoch": 2.6307099614749587, |
| "grad_norm": 0.16536889970302582, |
| "learning_rate": 0.00014944853171873387, |
| "loss": 2.113, |
| "step": 23900 |
| }, |
| { |
| "epoch": 2.63621353880022, |
| "grad_norm": 0.170758455991745, |
| "learning_rate": 0.00014890253083873784, |
| "loss": 2.1124, |
| "step": 23950 |
| }, |
| { |
| "epoch": 2.6417171161254815, |
| "grad_norm": 0.15669451653957367, |
| "learning_rate": 0.00014835654450025542, |
| "loss": 2.0894, |
| "step": 24000 |
| }, |
| { |
| "epoch": 2.647220693450743, |
| "grad_norm": 0.16937313973903656, |
| "learning_rate": 0.00014781057993762902, |
| "loss": 2.127, |
| "step": 24050 |
| }, |
| { |
| "epoch": 2.6527242707760044, |
| "grad_norm": 0.15754981338977814, |
| "learning_rate": 0.0001472646443849126, |
| "loss": 2.0959, |
| "step": 24100 |
| }, |
| { |
| "epoch": 2.6582278481012658, |
| "grad_norm": 0.16420376300811768, |
| "learning_rate": 0.00014671874507577574, |
| "loss": 2.0742, |
| "step": 24150 |
| }, |
| { |
| "epoch": 2.663731425426527, |
| "grad_norm": 0.16400571167469025, |
| "learning_rate": 0.00014617288924340775, |
| "loss": 2.0992, |
| "step": 24200 |
| }, |
| { |
| "epoch": 2.6692350027517886, |
| "grad_norm": 0.1658066064119339, |
| "learning_rate": 0.00014562708412042193, |
| "loss": 2.0925, |
| "step": 24250 |
| }, |
| { |
| "epoch": 2.67473858007705, |
| "grad_norm": 0.16872134804725647, |
| "learning_rate": 0.00014508133693875953, |
| "loss": 2.0899, |
| "step": 24300 |
| }, |
| { |
| "epoch": 2.6802421574023114, |
| "grad_norm": 0.17002440989017487, |
| "learning_rate": 0.00014453565492959422, |
| "loss": 2.1252, |
| "step": 24350 |
| }, |
| { |
| "epoch": 2.685745734727573, |
| "grad_norm": 0.16417178511619568, |
| "learning_rate": 0.00014399004532323614, |
| "loss": 2.112, |
| "step": 24400 |
| }, |
| { |
| "epoch": 2.6912493120528342, |
| "grad_norm": 0.172217458486557, |
| "learning_rate": 0.000143444515349036, |
| "loss": 2.0784, |
| "step": 24450 |
| }, |
| { |
| "epoch": 2.6967528893780957, |
| "grad_norm": 0.16687516868114471, |
| "learning_rate": 0.0001428990722352894, |
| "loss": 2.0775, |
| "step": 24500 |
| }, |
| { |
| "epoch": 2.702256466703357, |
| "grad_norm": 0.162861630320549, |
| "learning_rate": 0.00014235372320914108, |
| "loss": 2.0928, |
| "step": 24550 |
| }, |
| { |
| "epoch": 2.7077600440286185, |
| "grad_norm": 0.1640874743461609, |
| "learning_rate": 0.00014180847549648913, |
| "loss": 2.1071, |
| "step": 24600 |
| }, |
| { |
| "epoch": 2.71326362135388, |
| "grad_norm": 0.1654275506734848, |
| "learning_rate": 0.0001412633363218891, |
| "loss": 2.0615, |
| "step": 24650 |
| }, |
| { |
| "epoch": 2.7187671986791413, |
| "grad_norm": 0.14277833700180054, |
| "learning_rate": 0.00014071831290845857, |
| "loss": 2.1078, |
| "step": 24700 |
| }, |
| { |
| "epoch": 2.7242707760044027, |
| "grad_norm": 0.1705639362335205, |
| "learning_rate": 0.00014017341247778123, |
| "loss": 2.0858, |
| "step": 24750 |
| }, |
| { |
| "epoch": 2.729774353329664, |
| "grad_norm": 0.16390259563922882, |
| "learning_rate": 0.0001396286422498112, |
| "loss": 2.0508, |
| "step": 24800 |
| }, |
| { |
| "epoch": 2.7352779306549255, |
| "grad_norm": 0.1526353508234024, |
| "learning_rate": 0.00013908400944277738, |
| "loss": 2.0784, |
| "step": 24850 |
| }, |
| { |
| "epoch": 2.740781507980187, |
| "grad_norm": 0.1581793874502182, |
| "learning_rate": 0.0001385395212730879, |
| "loss": 2.0596, |
| "step": 24900 |
| }, |
| { |
| "epoch": 2.7462850853054483, |
| "grad_norm": 0.1692277193069458, |
| "learning_rate": 0.00013799518495523445, |
| "loss": 2.0838, |
| "step": 24950 |
| }, |
| { |
| "epoch": 2.7517886626307098, |
| "grad_norm": 0.1614445298910141, |
| "learning_rate": 0.0001374510077016967, |
| "loss": 2.0928, |
| "step": 25000 |
| }, |
| { |
| "epoch": 2.757292239955971, |
| "grad_norm": 0.15747620165348053, |
| "learning_rate": 0.00013690699672284657, |
| "loss": 2.0686, |
| "step": 25050 |
| }, |
| { |
| "epoch": 2.7627958172812326, |
| "grad_norm": 0.17565257847309113, |
| "learning_rate": 0.00013636315922685294, |
| "loss": 2.0971, |
| "step": 25100 |
| }, |
| { |
| "epoch": 2.768299394606494, |
| "grad_norm": 0.16367433965206146, |
| "learning_rate": 0.0001358195024195861, |
| "loss": 2.1052, |
| "step": 25150 |
| }, |
| { |
| "epoch": 2.7738029719317554, |
| "grad_norm": 0.1564227044582367, |
| "learning_rate": 0.00013527603350452206, |
| "loss": 2.0902, |
| "step": 25200 |
| }, |
| { |
| "epoch": 2.779306549257017, |
| "grad_norm": 0.16555672883987427, |
| "learning_rate": 0.0001347327596826473, |
| "loss": 2.0683, |
| "step": 25250 |
| }, |
| { |
| "epoch": 2.7848101265822782, |
| "grad_norm": 0.16134041547775269, |
| "learning_rate": 0.00013418968815236346, |
| "loss": 2.0627, |
| "step": 25300 |
| }, |
| { |
| "epoch": 2.7903137039075396, |
| "grad_norm": 0.17097671329975128, |
| "learning_rate": 0.0001336468261093915, |
| "loss": 2.0724, |
| "step": 25350 |
| }, |
| { |
| "epoch": 2.795817281232801, |
| "grad_norm": 0.1656900942325592, |
| "learning_rate": 0.00013310418074667698, |
| "loss": 2.1266, |
| "step": 25400 |
| }, |
| { |
| "epoch": 2.801320858558063, |
| "grad_norm": 0.1745171695947647, |
| "learning_rate": 0.00013256175925429415, |
| "loss": 2.0983, |
| "step": 25450 |
| }, |
| { |
| "epoch": 2.8068244358833243, |
| "grad_norm": 0.15649184584617615, |
| "learning_rate": 0.0001320195688193512, |
| "loss": 2.0672, |
| "step": 25500 |
| }, |
| { |
| "epoch": 2.8123280132085857, |
| "grad_norm": 0.16284936666488647, |
| "learning_rate": 0.00013147761662589473, |
| "loss": 2.0865, |
| "step": 25550 |
| }, |
| { |
| "epoch": 2.817831590533847, |
| "grad_norm": 0.15236727893352509, |
| "learning_rate": 0.00013093590985481456, |
| "loss": 2.0595, |
| "step": 25600 |
| }, |
| { |
| "epoch": 2.8233351678591085, |
| "grad_norm": 0.166220560669899, |
| "learning_rate": 0.00013039445568374872, |
| "loss": 2.0478, |
| "step": 25650 |
| }, |
| { |
| "epoch": 2.82883874518437, |
| "grad_norm": 0.16826435923576355, |
| "learning_rate": 0.0001298532612869883, |
| "loss": 2.0756, |
| "step": 25700 |
| }, |
| { |
| "epoch": 2.8343423225096314, |
| "grad_norm": 0.16447168588638306, |
| "learning_rate": 0.0001293123338353823, |
| "loss": 2.0742, |
| "step": 25750 |
| }, |
| { |
| "epoch": 2.8398458998348928, |
| "grad_norm": 0.157373309135437, |
| "learning_rate": 0.00012877168049624268, |
| "loss": 2.0728, |
| "step": 25800 |
| }, |
| { |
| "epoch": 2.845349477160154, |
| "grad_norm": 0.1642802357673645, |
| "learning_rate": 0.00012823130843324948, |
| "loss": 2.0643, |
| "step": 25850 |
| }, |
| { |
| "epoch": 2.8508530544854156, |
| "grad_norm": 0.15527865290641785, |
| "learning_rate": 0.00012769122480635583, |
| "loss": 2.1088, |
| "step": 25900 |
| }, |
| { |
| "epoch": 2.856356631810677, |
| "grad_norm": 0.175029918551445, |
| "learning_rate": 0.00012715143677169297, |
| "loss": 2.0764, |
| "step": 25950 |
| }, |
| { |
| "epoch": 2.8618602091359384, |
| "grad_norm": 0.17664563655853271, |
| "learning_rate": 0.0001266119514814755, |
| "loss": 2.0527, |
| "step": 26000 |
| }, |
| { |
| "epoch": 2.8673637864612, |
| "grad_norm": 0.16839756071567535, |
| "learning_rate": 0.00012607277608390677, |
| "loss": 2.0708, |
| "step": 26050 |
| }, |
| { |
| "epoch": 2.8728673637864612, |
| "grad_norm": 0.162934809923172, |
| "learning_rate": 0.00012553391772308407, |
| "loss": 2.0984, |
| "step": 26100 |
| }, |
| { |
| "epoch": 2.8783709411117226, |
| "grad_norm": 0.1836494356393814, |
| "learning_rate": 0.0001249953835389037, |
| "loss": 2.1118, |
| "step": 26150 |
| }, |
| { |
| "epoch": 2.883874518436984, |
| "grad_norm": 0.15794949233531952, |
| "learning_rate": 0.00012445718066696687, |
| "loss": 2.0541, |
| "step": 26200 |
| }, |
| { |
| "epoch": 2.8893780957622455, |
| "grad_norm": 0.16560156643390656, |
| "learning_rate": 0.00012391931623848476, |
| "loss": 2.1111, |
| "step": 26250 |
| }, |
| { |
| "epoch": 2.894881673087507, |
| "grad_norm": 0.16835126280784607, |
| "learning_rate": 0.0001233817973801842, |
| "loss": 2.0635, |
| "step": 26300 |
| }, |
| { |
| "epoch": 2.9003852504127683, |
| "grad_norm": 0.16151325404644012, |
| "learning_rate": 0.0001228446312142131, |
| "loss": 2.0949, |
| "step": 26350 |
| }, |
| { |
| "epoch": 2.9058888277380297, |
| "grad_norm": 0.16751372814178467, |
| "learning_rate": 0.00012230782485804633, |
| "loss": 2.0896, |
| "step": 26400 |
| }, |
| { |
| "epoch": 2.911392405063291, |
| "grad_norm": 0.18384750187397003, |
| "learning_rate": 0.00012177138542439122, |
| "loss": 2.047, |
| "step": 26450 |
| }, |
| { |
| "epoch": 2.9168959823885525, |
| "grad_norm": 0.15618419647216797, |
| "learning_rate": 0.00012123532002109328, |
| "loss": 2.0787, |
| "step": 26500 |
| }, |
| { |
| "epoch": 2.922399559713814, |
| "grad_norm": 0.15803013741970062, |
| "learning_rate": 0.00012069963575104223, |
| "loss": 2.0735, |
| "step": 26550 |
| }, |
| { |
| "epoch": 2.9279031370390753, |
| "grad_norm": 0.1603049784898758, |
| "learning_rate": 0.00012016433971207757, |
| "loss": 2.0802, |
| "step": 26600 |
| }, |
| { |
| "epoch": 2.9334067143643368, |
| "grad_norm": 0.1752861887216568, |
| "learning_rate": 0.0001196294389968949, |
| "loss": 2.0969, |
| "step": 26650 |
| }, |
| { |
| "epoch": 2.938910291689598, |
| "grad_norm": 0.16382624208927155, |
| "learning_rate": 0.00011909494069295175, |
| "loss": 2.0933, |
| "step": 26700 |
| }, |
| { |
| "epoch": 2.9444138690148596, |
| "grad_norm": 0.16802163422107697, |
| "learning_rate": 0.00011856085188237357, |
| "loss": 2.1042, |
| "step": 26750 |
| }, |
| { |
| "epoch": 2.949917446340121, |
| "grad_norm": 0.15907612442970276, |
| "learning_rate": 0.00011802717964186006, |
| "loss": 2.0575, |
| "step": 26800 |
| }, |
| { |
| "epoch": 2.9554210236653824, |
| "grad_norm": 0.16981548070907593, |
| "learning_rate": 0.00011749393104259143, |
| "loss": 2.0951, |
| "step": 26850 |
| }, |
| { |
| "epoch": 2.960924600990644, |
| "grad_norm": 0.15169958770275116, |
| "learning_rate": 0.0001169611131501345, |
| "loss": 2.0778, |
| "step": 26900 |
| }, |
| { |
| "epoch": 2.966428178315905, |
| "grad_norm": 0.1601976752281189, |
| "learning_rate": 0.00011642873302434926, |
| "loss": 2.0621, |
| "step": 26950 |
| }, |
| { |
| "epoch": 2.9719317556411666, |
| "grad_norm": 0.16251477599143982, |
| "learning_rate": 0.00011589679771929529, |
| "loss": 2.08, |
| "step": 27000 |
| }, |
| { |
| "epoch": 2.977435332966428, |
| "grad_norm": 0.15681184828281403, |
| "learning_rate": 0.00011536531428313828, |
| "loss": 2.0518, |
| "step": 27050 |
| }, |
| { |
| "epoch": 2.9829389102916894, |
| "grad_norm": 0.17260442674160004, |
| "learning_rate": 0.00011483428975805664, |
| "loss": 2.0899, |
| "step": 27100 |
| }, |
| { |
| "epoch": 2.988442487616951, |
| "grad_norm": 0.16616973280906677, |
| "learning_rate": 0.00011430373118014808, |
| "loss": 2.0787, |
| "step": 27150 |
| }, |
| { |
| "epoch": 2.9939460649422127, |
| "grad_norm": 0.1666904091835022, |
| "learning_rate": 0.00011377364557933658, |
| "loss": 2.0555, |
| "step": 27200 |
| }, |
| { |
| "epoch": 2.999449642267474, |
| "grad_norm": 0.15540987253189087, |
| "learning_rate": 0.00011324403997927922, |
| "loss": 2.0676, |
| "step": 27250 |
| }, |
| { |
| "epoch": 3.004953219592735, |
| "grad_norm": 0.16125087440013885, |
| "learning_rate": 0.00011271492139727294, |
| "loss": 2.0144, |
| "step": 27300 |
| }, |
| { |
| "epoch": 3.0104567969179965, |
| "grad_norm": 0.17550891637802124, |
| "learning_rate": 0.00011218629684416168, |
| "loss": 2.0102, |
| "step": 27350 |
| }, |
| { |
| "epoch": 3.015960374243258, |
| "grad_norm": 0.16927887499332428, |
| "learning_rate": 0.00011165817332424356, |
| "loss": 1.9816, |
| "step": 27400 |
| }, |
| { |
| "epoch": 3.0214639515685193, |
| "grad_norm": 0.17522330582141876, |
| "learning_rate": 0.00011113055783517793, |
| "loss": 2.0066, |
| "step": 27450 |
| }, |
| { |
| "epoch": 3.026967528893781, |
| "grad_norm": 0.160496324300766, |
| "learning_rate": 0.00011060345736789265, |
| "loss": 2.0003, |
| "step": 27500 |
| }, |
| { |
| "epoch": 3.0324711062190426, |
| "grad_norm": 0.17563533782958984, |
| "learning_rate": 0.00011007687890649163, |
| "loss": 2.0272, |
| "step": 27550 |
| }, |
| { |
| "epoch": 3.037974683544304, |
| "grad_norm": 0.17000171542167664, |
| "learning_rate": 0.00010955082942816215, |
| "loss": 2.0447, |
| "step": 27600 |
| }, |
| { |
| "epoch": 3.0434782608695654, |
| "grad_norm": 0.17253416776657104, |
| "learning_rate": 0.00010902531590308236, |
| "loss": 1.9968, |
| "step": 27650 |
| }, |
| { |
| "epoch": 3.048981838194827, |
| "grad_norm": 0.16862879693508148, |
| "learning_rate": 0.00010850034529432919, |
| "loss": 1.9952, |
| "step": 27700 |
| }, |
| { |
| "epoch": 3.0544854155200882, |
| "grad_norm": 0.166275292634964, |
| "learning_rate": 0.00010797592455778562, |
| "loss": 2.0258, |
| "step": 27750 |
| }, |
| { |
| "epoch": 3.0599889928453496, |
| "grad_norm": 0.1734013706445694, |
| "learning_rate": 0.00010745206064204903, |
| "loss": 2.0047, |
| "step": 27800 |
| }, |
| { |
| "epoch": 3.065492570170611, |
| "grad_norm": 0.1760258674621582, |
| "learning_rate": 0.00010692876048833889, |
| "loss": 1.997, |
| "step": 27850 |
| }, |
| { |
| "epoch": 3.0709961474958725, |
| "grad_norm": 0.1616799384355545, |
| "learning_rate": 0.00010640603103040464, |
| "loss": 1.9817, |
| "step": 27900 |
| }, |
| { |
| "epoch": 3.076499724821134, |
| "grad_norm": 0.1775207221508026, |
| "learning_rate": 0.00010588387919443416, |
| "loss": 2.0391, |
| "step": 27950 |
| }, |
| { |
| "epoch": 3.0820033021463953, |
| "grad_norm": 0.17136207222938538, |
| "learning_rate": 0.00010536231189896175, |
| "loss": 1.999, |
| "step": 28000 |
| }, |
| { |
| "epoch": 3.0875068794716567, |
| "grad_norm": 0.1694386601448059, |
| "learning_rate": 0.00010484133605477644, |
| "loss": 2.0153, |
| "step": 28050 |
| }, |
| { |
| "epoch": 3.093010456796918, |
| "grad_norm": 0.17021538317203522, |
| "learning_rate": 0.00010432095856483057, |
| "loss": 2.0411, |
| "step": 28100 |
| }, |
| { |
| "epoch": 3.0985140341221795, |
| "grad_norm": 0.17548739910125732, |
| "learning_rate": 0.00010380118632414833, |
| "loss": 2.0416, |
| "step": 28150 |
| }, |
| { |
| "epoch": 3.104017611447441, |
| "grad_norm": 0.1785077303647995, |
| "learning_rate": 0.0001032820262197342, |
| "loss": 2.032, |
| "step": 28200 |
| }, |
| { |
| "epoch": 3.1095211887727023, |
| "grad_norm": 0.17226973176002502, |
| "learning_rate": 0.00010276348513048193, |
| "loss": 1.986, |
| "step": 28250 |
| }, |
| { |
| "epoch": 3.1150247660979637, |
| "grad_norm": 0.18131954967975616, |
| "learning_rate": 0.00010224556992708328, |
| "loss": 2.0149, |
| "step": 28300 |
| }, |
| { |
| "epoch": 3.120528343423225, |
| "grad_norm": 0.16980765759944916, |
| "learning_rate": 0.00010172828747193689, |
| "loss": 2.0309, |
| "step": 28350 |
| }, |
| { |
| "epoch": 3.1260319207484866, |
| "grad_norm": 0.16887415945529938, |
| "learning_rate": 0.00010121164461905759, |
| "loss": 1.9814, |
| "step": 28400 |
| }, |
| { |
| "epoch": 3.131535498073748, |
| "grad_norm": 0.18381917476654053, |
| "learning_rate": 0.00010069564821398536, |
| "loss": 2.0228, |
| "step": 28450 |
| }, |
| { |
| "epoch": 3.1370390753990094, |
| "grad_norm": 0.18095500767230988, |
| "learning_rate": 0.0001001803050936948, |
| "loss": 2.0197, |
| "step": 28500 |
| }, |
| { |
| "epoch": 3.142542652724271, |
| "grad_norm": 0.15869130194187164, |
| "learning_rate": 9.966562208650441e-05, |
| "loss": 2.0114, |
| "step": 28550 |
| }, |
| { |
| "epoch": 3.148046230049532, |
| "grad_norm": 0.17333059012889862, |
| "learning_rate": 9.915160601198616e-05, |
| "loss": 2.0025, |
| "step": 28600 |
| }, |
| { |
| "epoch": 3.1535498073747936, |
| "grad_norm": 0.161941796541214, |
| "learning_rate": 9.863826368087516e-05, |
| "loss": 2.0273, |
| "step": 28650 |
| }, |
| { |
| "epoch": 3.159053384700055, |
| "grad_norm": 0.16833285987377167, |
| "learning_rate": 9.812560189497935e-05, |
| "loss": 2.0108, |
| "step": 28700 |
| }, |
| { |
| "epoch": 3.1645569620253164, |
| "grad_norm": 0.15908978879451752, |
| "learning_rate": 9.76136274470895e-05, |
| "loss": 1.9852, |
| "step": 28750 |
| }, |
| { |
| "epoch": 3.170060539350578, |
| "grad_norm": 0.1724586933851242, |
| "learning_rate": 9.710234712088896e-05, |
| "loss": 2.0297, |
| "step": 28800 |
| }, |
| { |
| "epoch": 3.1755641166758393, |
| "grad_norm": 0.17113400995731354, |
| "learning_rate": 9.659176769086419e-05, |
| "loss": 1.9969, |
| "step": 28850 |
| }, |
| { |
| "epoch": 3.1810676940011007, |
| "grad_norm": 0.16423186659812927, |
| "learning_rate": 9.608189592221446e-05, |
| "loss": 2.0106, |
| "step": 28900 |
| }, |
| { |
| "epoch": 3.186571271326362, |
| "grad_norm": 0.16140978038311005, |
| "learning_rate": 9.557273857076271e-05, |
| "loss": 2.0339, |
| "step": 28950 |
| }, |
| { |
| "epoch": 3.1920748486516235, |
| "grad_norm": 0.1718360334634781, |
| "learning_rate": 9.506430238286566e-05, |
| "loss": 2.04, |
| "step": 29000 |
| }, |
| { |
| "epoch": 3.197578425976885, |
| "grad_norm": 0.164507195353508, |
| "learning_rate": 9.455659409532474e-05, |
| "loss": 2.0168, |
| "step": 29050 |
| }, |
| { |
| "epoch": 3.2030820033021463, |
| "grad_norm": 0.17989708483219147, |
| "learning_rate": 9.40496204352966e-05, |
| "loss": 2.0273, |
| "step": 29100 |
| }, |
| { |
| "epoch": 3.2085855806274077, |
| "grad_norm": 0.17319265007972717, |
| "learning_rate": 9.354338812020402e-05, |
| "loss": 1.9911, |
| "step": 29150 |
| }, |
| { |
| "epoch": 3.214089157952669, |
| "grad_norm": 0.17870402336120605, |
| "learning_rate": 9.303790385764694e-05, |
| "loss": 2.0103, |
| "step": 29200 |
| }, |
| { |
| "epoch": 3.2195927352779306, |
| "grad_norm": 0.1797982007265091, |
| "learning_rate": 9.253317434531357e-05, |
| "loss": 2.0169, |
| "step": 29250 |
| }, |
| { |
| "epoch": 3.225096312603192, |
| "grad_norm": 0.15193064510822296, |
| "learning_rate": 9.202920627089171e-05, |
| "loss": 2.0116, |
| "step": 29300 |
| }, |
| { |
| "epoch": 3.2305998899284534, |
| "grad_norm": 0.16679364442825317, |
| "learning_rate": 9.152600631197996e-05, |
| "loss": 2.0362, |
| "step": 29350 |
| }, |
| { |
| "epoch": 3.236103467253715, |
| "grad_norm": 0.17053711414337158, |
| "learning_rate": 9.102358113599942e-05, |
| "loss": 1.9829, |
| "step": 29400 |
| }, |
| { |
| "epoch": 3.241607044578976, |
| "grad_norm": 0.16923174262046814, |
| "learning_rate": 9.052193740010536e-05, |
| "loss": 2.0011, |
| "step": 29450 |
| }, |
| { |
| "epoch": 3.2471106219042376, |
| "grad_norm": 0.18323257565498352, |
| "learning_rate": 9.002108175109877e-05, |
| "loss": 1.9784, |
| "step": 29500 |
| }, |
| { |
| "epoch": 3.252614199229499, |
| "grad_norm": 0.16737329959869385, |
| "learning_rate": 8.952102082533855e-05, |
| "loss": 2.0276, |
| "step": 29550 |
| }, |
| { |
| "epoch": 3.2581177765547604, |
| "grad_norm": 0.17038094997406006, |
| "learning_rate": 8.902176124865348e-05, |
| "loss": 2.0315, |
| "step": 29600 |
| }, |
| { |
| "epoch": 3.263621353880022, |
| "grad_norm": 0.16778729856014252, |
| "learning_rate": 8.852330963625444e-05, |
| "loss": 2.0139, |
| "step": 29650 |
| }, |
| { |
| "epoch": 3.2691249312052832, |
| "grad_norm": 0.18061307072639465, |
| "learning_rate": 8.802567259264674e-05, |
| "loss": 2.002, |
| "step": 29700 |
| }, |
| { |
| "epoch": 3.2746285085305447, |
| "grad_norm": 0.17191821336746216, |
| "learning_rate": 8.752885671154253e-05, |
| "loss": 2.0218, |
| "step": 29750 |
| }, |
| { |
| "epoch": 3.280132085855806, |
| "grad_norm": 0.17957885563373566, |
| "learning_rate": 8.703286857577378e-05, |
| "loss": 2.0217, |
| "step": 29800 |
| }, |
| { |
| "epoch": 3.2856356631810675, |
| "grad_norm": 0.16803069412708282, |
| "learning_rate": 8.653771475720442e-05, |
| "loss": 2.0185, |
| "step": 29850 |
| }, |
| { |
| "epoch": 3.291139240506329, |
| "grad_norm": 0.1788654327392578, |
| "learning_rate": 8.604340181664395e-05, |
| "loss": 1.9973, |
| "step": 29900 |
| }, |
| { |
| "epoch": 3.2966428178315903, |
| "grad_norm": 0.17996376752853394, |
| "learning_rate": 8.554993630376006e-05, |
| "loss": 2.0161, |
| "step": 29950 |
| }, |
| { |
| "epoch": 3.3021463951568517, |
| "grad_norm": 0.18701910972595215, |
| "learning_rate": 8.50573247569921e-05, |
| "loss": 2.0156, |
| "step": 30000 |
| }, |
| { |
| "epoch": 3.3076499724821136, |
| "grad_norm": 0.1713022142648697, |
| "learning_rate": 8.45655737034641e-05, |
| "loss": 2.0144, |
| "step": 30050 |
| }, |
| { |
| "epoch": 3.313153549807375, |
| "grad_norm": 0.17666327953338623, |
| "learning_rate": 8.407468965889878e-05, |
| "loss": 2.0116, |
| "step": 30100 |
| }, |
| { |
| "epoch": 3.3186571271326364, |
| "grad_norm": 0.18481209874153137, |
| "learning_rate": 8.358467912753079e-05, |
| "loss": 2.0149, |
| "step": 30150 |
| }, |
| { |
| "epoch": 3.324160704457898, |
| "grad_norm": 0.17895914614200592, |
| "learning_rate": 8.309554860202082e-05, |
| "loss": 1.9913, |
| "step": 30200 |
| }, |
| { |
| "epoch": 3.329664281783159, |
| "grad_norm": 0.18427938222885132, |
| "learning_rate": 8.260730456336925e-05, |
| "loss": 1.9913, |
| "step": 30250 |
| }, |
| { |
| "epoch": 3.3351678591084206, |
| "grad_norm": 0.1806970238685608, |
| "learning_rate": 8.211995348083067e-05, |
| "loss": 2.0177, |
| "step": 30300 |
| }, |
| { |
| "epoch": 3.340671436433682, |
| "grad_norm": 0.1735043078660965, |
| "learning_rate": 8.163350181182787e-05, |
| "loss": 2.0275, |
| "step": 30350 |
| }, |
| { |
| "epoch": 3.3461750137589434, |
| "grad_norm": 0.17593735456466675, |
| "learning_rate": 8.11479560018664e-05, |
| "loss": 2.0525, |
| "step": 30400 |
| }, |
| { |
| "epoch": 3.351678591084205, |
| "grad_norm": 0.1755988448858261, |
| "learning_rate": 8.066332248444908e-05, |
| "loss": 2.0182, |
| "step": 30450 |
| }, |
| { |
| "epoch": 3.3571821684094663, |
| "grad_norm": 0.15805041790008545, |
| "learning_rate": 8.017960768099091e-05, |
| "loss": 1.9748, |
| "step": 30500 |
| }, |
| { |
| "epoch": 3.3626857457347277, |
| "grad_norm": 0.18487168848514557, |
| "learning_rate": 7.969681800073386e-05, |
| "loss": 2.021, |
| "step": 30550 |
| }, |
| { |
| "epoch": 3.368189323059989, |
| "grad_norm": 0.17120800912380219, |
| "learning_rate": 7.921495984066194e-05, |
| "loss": 2.0113, |
| "step": 30600 |
| }, |
| { |
| "epoch": 3.3736929003852505, |
| "grad_norm": 0.1721143126487732, |
| "learning_rate": 7.87340395854164e-05, |
| "loss": 1.9936, |
| "step": 30650 |
| }, |
| { |
| "epoch": 3.379196477710512, |
| "grad_norm": 0.17595061659812927, |
| "learning_rate": 7.825406360721139e-05, |
| "loss": 2.0044, |
| "step": 30700 |
| }, |
| { |
| "epoch": 3.3847000550357733, |
| "grad_norm": 0.16751642525196075, |
| "learning_rate": 7.777503826574925e-05, |
| "loss": 1.9988, |
| "step": 30750 |
| }, |
| { |
| "epoch": 3.3902036323610347, |
| "grad_norm": 0.17864260077476501, |
| "learning_rate": 7.729696990813634e-05, |
| "loss": 2.0189, |
| "step": 30800 |
| }, |
| { |
| "epoch": 3.395707209686296, |
| "grad_norm": 0.1791946142911911, |
| "learning_rate": 7.681986486879898e-05, |
| "loss": 2.0038, |
| "step": 30850 |
| }, |
| { |
| "epoch": 3.4012107870115575, |
| "grad_norm": 0.1704343855381012, |
| "learning_rate": 7.634372946939945e-05, |
| "loss": 1.9625, |
| "step": 30900 |
| }, |
| { |
| "epoch": 3.406714364336819, |
| "grad_norm": 0.17794691026210785, |
| "learning_rate": 7.586857001875235e-05, |
| "loss": 2.0017, |
| "step": 30950 |
| }, |
| { |
| "epoch": 3.4122179416620804, |
| "grad_norm": 0.17922665178775787, |
| "learning_rate": 7.539439281274071e-05, |
| "loss": 2.054, |
| "step": 31000 |
| }, |
| { |
| "epoch": 3.4177215189873418, |
| "grad_norm": 0.174763485789299, |
| "learning_rate": 7.492120413423296e-05, |
| "loss": 2.0185, |
| "step": 31050 |
| }, |
| { |
| "epoch": 3.423225096312603, |
| "grad_norm": 0.1687181144952774, |
| "learning_rate": 7.444901025299941e-05, |
| "loss": 2.0463, |
| "step": 31100 |
| }, |
| { |
| "epoch": 3.4287286736378646, |
| "grad_norm": 0.17793361842632294, |
| "learning_rate": 7.397781742562941e-05, |
| "loss": 2.0296, |
| "step": 31150 |
| }, |
| { |
| "epoch": 3.434232250963126, |
| "grad_norm": 0.16858656704425812, |
| "learning_rate": 7.350763189544803e-05, |
| "loss": 1.9916, |
| "step": 31200 |
| }, |
| { |
| "epoch": 3.4397358282883874, |
| "grad_norm": 0.16184011101722717, |
| "learning_rate": 7.303845989243384e-05, |
| "loss": 2.0314, |
| "step": 31250 |
| }, |
| { |
| "epoch": 3.445239405613649, |
| "grad_norm": 0.1733245998620987, |
| "learning_rate": 7.257030763313607e-05, |
| "loss": 2.0125, |
| "step": 31300 |
| }, |
| { |
| "epoch": 3.4507429829389102, |
| "grad_norm": 0.18471761047840118, |
| "learning_rate": 7.210318132059231e-05, |
| "loss": 1.9691, |
| "step": 31350 |
| }, |
| { |
| "epoch": 3.4562465602641717, |
| "grad_norm": 0.17260605096817017, |
| "learning_rate": 7.16370871442462e-05, |
| "loss": 2.0113, |
| "step": 31400 |
| }, |
| { |
| "epoch": 3.461750137589433, |
| "grad_norm": 0.1758509874343872, |
| "learning_rate": 7.117203127986569e-05, |
| "loss": 2.0239, |
| "step": 31450 |
| }, |
| { |
| "epoch": 3.4672537149146945, |
| "grad_norm": 0.18456335365772247, |
| "learning_rate": 7.070801988946098e-05, |
| "loss": 1.999, |
| "step": 31500 |
| }, |
| { |
| "epoch": 3.472757292239956, |
| "grad_norm": 0.17995281517505646, |
| "learning_rate": 7.024505912120293e-05, |
| "loss": 2.0141, |
| "step": 31550 |
| }, |
| { |
| "epoch": 3.4782608695652173, |
| "grad_norm": 0.16747882962226868, |
| "learning_rate": 6.978315510934164e-05, |
| "loss": 2.0073, |
| "step": 31600 |
| }, |
| { |
| "epoch": 3.4837644468904787, |
| "grad_norm": 0.1658480316400528, |
| "learning_rate": 6.932231397412522e-05, |
| "loss": 2.0079, |
| "step": 31650 |
| }, |
| { |
| "epoch": 3.48926802421574, |
| "grad_norm": 0.17987202107906342, |
| "learning_rate": 6.886254182171836e-05, |
| "loss": 2.0439, |
| "step": 31700 |
| }, |
| { |
| "epoch": 3.4947716015410015, |
| "grad_norm": 0.1672779768705368, |
| "learning_rate": 6.840384474412201e-05, |
| "loss": 1.9878, |
| "step": 31750 |
| }, |
| { |
| "epoch": 3.500275178866263, |
| "grad_norm": 0.1716667115688324, |
| "learning_rate": 6.794622881909207e-05, |
| "loss": 2.0248, |
| "step": 31800 |
| }, |
| { |
| "epoch": 3.5057787561915243, |
| "grad_norm": 0.17986662685871124, |
| "learning_rate": 6.748970011005924e-05, |
| "loss": 1.99, |
| "step": 31850 |
| }, |
| { |
| "epoch": 3.5112823335167858, |
| "grad_norm": 0.1808827668428421, |
| "learning_rate": 6.703426466604858e-05, |
| "loss": 1.9692, |
| "step": 31900 |
| }, |
| { |
| "epoch": 3.516785910842047, |
| "grad_norm": 0.17262905836105347, |
| "learning_rate": 6.65799285215993e-05, |
| "loss": 1.9636, |
| "step": 31950 |
| }, |
| { |
| "epoch": 3.522289488167309, |
| "grad_norm": 0.16857630014419556, |
| "learning_rate": 6.612669769668489e-05, |
| "loss": 1.9923, |
| "step": 32000 |
| }, |
| { |
| "epoch": 3.5277930654925704, |
| "grad_norm": 0.17498353123664856, |
| "learning_rate": 6.567457819663327e-05, |
| "loss": 2.0248, |
| "step": 32050 |
| }, |
| { |
| "epoch": 3.533296642817832, |
| "grad_norm": 0.16948480904102325, |
| "learning_rate": 6.522357601204719e-05, |
| "loss": 1.9966, |
| "step": 32100 |
| }, |
| { |
| "epoch": 3.5388002201430933, |
| "grad_norm": 0.1695474237203598, |
| "learning_rate": 6.477369711872502e-05, |
| "loss": 1.994, |
| "step": 32150 |
| }, |
| { |
| "epoch": 3.5443037974683547, |
| "grad_norm": 0.1761193871498108, |
| "learning_rate": 6.432494747758146e-05, |
| "loss": 1.9827, |
| "step": 32200 |
| }, |
| { |
| "epoch": 3.549807374793616, |
| "grad_norm": 0.17206500470638275, |
| "learning_rate": 6.387733303456858e-05, |
| "loss": 1.9944, |
| "step": 32250 |
| }, |
| { |
| "epoch": 3.5553109521188775, |
| "grad_norm": 0.19551701843738556, |
| "learning_rate": 6.3430859720597e-05, |
| "loss": 2.0118, |
| "step": 32300 |
| }, |
| { |
| "epoch": 3.560814529444139, |
| "grad_norm": 0.17827004194259644, |
| "learning_rate": 6.298553345145728e-05, |
| "loss": 1.9953, |
| "step": 32350 |
| }, |
| { |
| "epoch": 3.5663181067694003, |
| "grad_norm": 0.17359542846679688, |
| "learning_rate": 6.254136012774166e-05, |
| "loss": 2.0098, |
| "step": 32400 |
| }, |
| { |
| "epoch": 3.5718216840946617, |
| "grad_norm": 0.16225001215934753, |
| "learning_rate": 6.209834563476578e-05, |
| "loss": 2.008, |
| "step": 32450 |
| }, |
| { |
| "epoch": 3.577325261419923, |
| "grad_norm": 0.17888249456882477, |
| "learning_rate": 6.165649584249079e-05, |
| "loss": 1.9962, |
| "step": 32500 |
| }, |
| { |
| "epoch": 3.5828288387451845, |
| "grad_norm": 0.16299229860305786, |
| "learning_rate": 6.121581660544532e-05, |
| "loss": 2.0387, |
| "step": 32550 |
| }, |
| { |
| "epoch": 3.588332416070446, |
| "grad_norm": 0.16753186285495758, |
| "learning_rate": 6.077631376264832e-05, |
| "loss": 1.9662, |
| "step": 32600 |
| }, |
| { |
| "epoch": 3.5938359933957074, |
| "grad_norm": 0.18166500329971313, |
| "learning_rate": 6.033799313753129e-05, |
| "loss": 1.997, |
| "step": 32650 |
| }, |
| { |
| "epoch": 3.5993395707209688, |
| "grad_norm": 0.16691668331623077, |
| "learning_rate": 5.990086053786139e-05, |
| "loss": 2.0072, |
| "step": 32700 |
| }, |
| { |
| "epoch": 3.60484314804623, |
| "grad_norm": 0.18671706318855286, |
| "learning_rate": 5.946492175566438e-05, |
| "loss": 2.0217, |
| "step": 32750 |
| }, |
| { |
| "epoch": 3.6103467253714916, |
| "grad_norm": 0.16831424832344055, |
| "learning_rate": 5.9030182567147905e-05, |
| "loss": 2.0024, |
| "step": 32800 |
| }, |
| { |
| "epoch": 3.615850302696753, |
| "grad_norm": 0.17562079429626465, |
| "learning_rate": 5.859664873262483e-05, |
| "loss": 2.0351, |
| "step": 32850 |
| }, |
| { |
| "epoch": 3.6213538800220144, |
| "grad_norm": 0.18190255761146545, |
| "learning_rate": 5.816432599643719e-05, |
| "loss": 2.0117, |
| "step": 32900 |
| }, |
| { |
| "epoch": 3.626857457347276, |
| "grad_norm": 0.1830035150051117, |
| "learning_rate": 5.773322008687972e-05, |
| "loss": 2.012, |
| "step": 32950 |
| }, |
| { |
| "epoch": 3.6323610346725372, |
| "grad_norm": 0.17998354136943817, |
| "learning_rate": 5.730333671612435e-05, |
| "loss": 1.9977, |
| "step": 33000 |
| }, |
| { |
| "epoch": 3.6378646119977986, |
| "grad_norm": 0.1839103251695633, |
| "learning_rate": 5.687468158014421e-05, |
| "loss": 2.003, |
| "step": 33050 |
| }, |
| { |
| "epoch": 3.64336818932306, |
| "grad_norm": 0.18631495535373688, |
| "learning_rate": 5.6447260358638285e-05, |
| "loss": 2.0072, |
| "step": 33100 |
| }, |
| { |
| "epoch": 3.6488717666483215, |
| "grad_norm": 0.19609171152114868, |
| "learning_rate": 5.602107871495615e-05, |
| "loss": 1.9912, |
| "step": 33150 |
| }, |
| { |
| "epoch": 3.654375343973583, |
| "grad_norm": 0.16935598850250244, |
| "learning_rate": 5.5596142296022954e-05, |
| "loss": 2.0097, |
| "step": 33200 |
| }, |
| { |
| "epoch": 3.6598789212988443, |
| "grad_norm": 0.16973347961902618, |
| "learning_rate": 5.5172456732264435e-05, |
| "loss": 2.0151, |
| "step": 33250 |
| }, |
| { |
| "epoch": 3.6653824986241057, |
| "grad_norm": 0.17289844155311584, |
| "learning_rate": 5.475002763753257e-05, |
| "loss": 2.0393, |
| "step": 33300 |
| }, |
| { |
| "epoch": 3.670886075949367, |
| "grad_norm": 0.1864277720451355, |
| "learning_rate": 5.432886060903105e-05, |
| "loss": 2.0069, |
| "step": 33350 |
| }, |
| { |
| "epoch": 3.6763896532746285, |
| "grad_norm": 0.17381645739078522, |
| "learning_rate": 5.390896122724111e-05, |
| "loss": 2.0323, |
| "step": 33400 |
| }, |
| { |
| "epoch": 3.68189323059989, |
| "grad_norm": 0.18182097375392914, |
| "learning_rate": 5.349033505584767e-05, |
| "loss": 2.0344, |
| "step": 33450 |
| }, |
| { |
| "epoch": 3.6873968079251513, |
| "grad_norm": 0.17326848208904266, |
| "learning_rate": 5.3072987641665373e-05, |
| "loss": 1.9888, |
| "step": 33500 |
| }, |
| { |
| "epoch": 3.6929003852504128, |
| "grad_norm": 0.17132443189620972, |
| "learning_rate": 5.2656924514565515e-05, |
| "loss": 1.9741, |
| "step": 33550 |
| }, |
| { |
| "epoch": 3.698403962575674, |
| "grad_norm": 0.17236609756946564, |
| "learning_rate": 5.224215118740243e-05, |
| "loss": 1.9967, |
| "step": 33600 |
| }, |
| { |
| "epoch": 3.7039075399009356, |
| "grad_norm": 0.16628102958202362, |
| "learning_rate": 5.182867315594046e-05, |
| "loss": 2.0036, |
| "step": 33650 |
| }, |
| { |
| "epoch": 3.709411117226197, |
| "grad_norm": 0.17410129308700562, |
| "learning_rate": 5.141649589878134e-05, |
| "loss": 2.0173, |
| "step": 33700 |
| }, |
| { |
| "epoch": 3.7149146945514584, |
| "grad_norm": 0.1739521622657776, |
| "learning_rate": 5.100562487729148e-05, |
| "loss": 1.9977, |
| "step": 33750 |
| }, |
| { |
| "epoch": 3.72041827187672, |
| "grad_norm": 0.17988687753677368, |
| "learning_rate": 5.059606553552956e-05, |
| "loss": 1.9964, |
| "step": 33800 |
| }, |
| { |
| "epoch": 3.725921849201981, |
| "grad_norm": 0.17351609468460083, |
| "learning_rate": 5.018782330017448e-05, |
| "loss": 2.0078, |
| "step": 33850 |
| }, |
| { |
| "epoch": 3.7314254265272426, |
| "grad_norm": 0.176718607544899, |
| "learning_rate": 4.9780903580453435e-05, |
| "loss": 2.0036, |
| "step": 33900 |
| }, |
| { |
| "epoch": 3.736929003852504, |
| "grad_norm": 0.17504329979419708, |
| "learning_rate": 4.937531176807023e-05, |
| "loss": 1.9837, |
| "step": 33950 |
| }, |
| { |
| "epoch": 3.7424325811777654, |
| "grad_norm": 0.179831862449646, |
| "learning_rate": 4.897105323713374e-05, |
| "loss": 1.999, |
| "step": 34000 |
| }, |
| { |
| "epoch": 3.747936158503027, |
| "grad_norm": 0.17752991616725922, |
| "learning_rate": 4.8568133344086986e-05, |
| "loss": 2.0194, |
| "step": 34050 |
| }, |
| { |
| "epoch": 3.7534397358282883, |
| "grad_norm": 0.17888140678405762, |
| "learning_rate": 4.8166557427635774e-05, |
| "loss": 2.0108, |
| "step": 34100 |
| }, |
| { |
| "epoch": 3.7589433131535497, |
| "grad_norm": 0.17585836350917816, |
| "learning_rate": 4.776633080867834e-05, |
| "loss": 2.0421, |
| "step": 34150 |
| }, |
| { |
| "epoch": 3.764446890478811, |
| "grad_norm": 0.16860149800777435, |
| "learning_rate": 4.736745879023457e-05, |
| "loss": 2.0126, |
| "step": 34200 |
| }, |
| { |
| "epoch": 3.7699504678040725, |
| "grad_norm": 0.17601364850997925, |
| "learning_rate": 4.6969946657375865e-05, |
| "loss": 2.0127, |
| "step": 34250 |
| }, |
| { |
| "epoch": 3.775454045129334, |
| "grad_norm": 0.19177961349487305, |
| "learning_rate": 4.657379967715511e-05, |
| "loss": 2.0234, |
| "step": 34300 |
| }, |
| { |
| "epoch": 3.7809576224545953, |
| "grad_norm": 0.19012008607387543, |
| "learning_rate": 4.617902309853685e-05, |
| "loss": 1.976, |
| "step": 34350 |
| }, |
| { |
| "epoch": 3.7864611997798567, |
| "grad_norm": 0.18448469042778015, |
| "learning_rate": 4.578562215232766e-05, |
| "loss": 1.9897, |
| "step": 34400 |
| }, |
| { |
| "epoch": 3.791964777105118, |
| "grad_norm": 0.18167300522327423, |
| "learning_rate": 4.539360205110701e-05, |
| "loss": 2.0088, |
| "step": 34450 |
| }, |
| { |
| "epoch": 3.7974683544303796, |
| "grad_norm": 0.1806318610906601, |
| "learning_rate": 4.500296798915813e-05, |
| "loss": 1.9949, |
| "step": 34500 |
| }, |
| { |
| "epoch": 3.802971931755641, |
| "grad_norm": 0.17676065862178802, |
| "learning_rate": 4.461372514239911e-05, |
| "loss": 1.9966, |
| "step": 34550 |
| }, |
| { |
| "epoch": 3.8084755090809024, |
| "grad_norm": 0.16621781885623932, |
| "learning_rate": 4.422587866831446e-05, |
| "loss": 1.9952, |
| "step": 34600 |
| }, |
| { |
| "epoch": 3.813979086406164, |
| "grad_norm": 0.16662217676639557, |
| "learning_rate": 4.3839433705886646e-05, |
| "loss": 2.0032, |
| "step": 34650 |
| }, |
| { |
| "epoch": 3.819482663731425, |
| "grad_norm": 0.15986546874046326, |
| "learning_rate": 4.3454395375527966e-05, |
| "loss": 1.9974, |
| "step": 34700 |
| }, |
| { |
| "epoch": 3.8249862410566866, |
| "grad_norm": 0.18636077642440796, |
| "learning_rate": 4.307076877901293e-05, |
| "loss": 1.9959, |
| "step": 34750 |
| }, |
| { |
| "epoch": 3.830489818381948, |
| "grad_norm": 0.17939697206020355, |
| "learning_rate": 4.268855899941039e-05, |
| "loss": 1.9923, |
| "step": 34800 |
| }, |
| { |
| "epoch": 3.8359933957072094, |
| "grad_norm": 0.1743779331445694, |
| "learning_rate": 4.2307771101016366e-05, |
| "loss": 2.0055, |
| "step": 34850 |
| }, |
| { |
| "epoch": 3.841496973032471, |
| "grad_norm": 0.1767256259918213, |
| "learning_rate": 4.192841012928691e-05, |
| "loss": 1.9834, |
| "step": 34900 |
| }, |
| { |
| "epoch": 3.8470005503577323, |
| "grad_norm": 0.1775825172662735, |
| "learning_rate": 4.1550481110771224e-05, |
| "loss": 2.01, |
| "step": 34950 |
| }, |
| { |
| "epoch": 3.852504127682994, |
| "grad_norm": 0.1638861745595932, |
| "learning_rate": 4.1173989053045024e-05, |
| "loss": 2.0001, |
| "step": 35000 |
| }, |
| { |
| "epoch": 3.8580077050082555, |
| "grad_norm": 0.1761082410812378, |
| "learning_rate": 4.079893894464429e-05, |
| "loss": 2.0041, |
| "step": 35050 |
| }, |
| { |
| "epoch": 3.863511282333517, |
| "grad_norm": 0.17102594673633575, |
| "learning_rate": 4.042533575499914e-05, |
| "loss": 2.003, |
| "step": 35100 |
| }, |
| { |
| "epoch": 3.8690148596587783, |
| "grad_norm": 0.19267229735851288, |
| "learning_rate": 4.005318443436779e-05, |
| "loss": 1.973, |
| "step": 35150 |
| }, |
| { |
| "epoch": 3.8745184369840397, |
| "grad_norm": 0.16570790112018585, |
| "learning_rate": 3.968248991377137e-05, |
| "loss": 1.9834, |
| "step": 35200 |
| }, |
| { |
| "epoch": 3.880022014309301, |
| "grad_norm": 0.16816405951976776, |
| "learning_rate": 3.9313257104928094e-05, |
| "loss": 2.0011, |
| "step": 35250 |
| }, |
| { |
| "epoch": 3.8855255916345626, |
| "grad_norm": 0.16395322978496552, |
| "learning_rate": 3.8945490900188616e-05, |
| "loss": 1.9771, |
| "step": 35300 |
| }, |
| { |
| "epoch": 3.891029168959824, |
| "grad_norm": 0.17911794781684875, |
| "learning_rate": 3.857919617247097e-05, |
| "loss": 2.0, |
| "step": 35350 |
| }, |
| { |
| "epoch": 3.8965327462850854, |
| "grad_norm": 0.18344633281230927, |
| "learning_rate": 3.821437777519607e-05, |
| "loss": 1.9952, |
| "step": 35400 |
| }, |
| { |
| "epoch": 3.902036323610347, |
| "grad_norm": 0.1782706081867218, |
| "learning_rate": 3.785104054222334e-05, |
| "loss": 1.9952, |
| "step": 35450 |
| }, |
| { |
| "epoch": 3.907539900935608, |
| "grad_norm": 0.17840951681137085, |
| "learning_rate": 3.74891892877868e-05, |
| "loss": 1.9994, |
| "step": 35500 |
| }, |
| { |
| "epoch": 3.9130434782608696, |
| "grad_norm": 0.19014611840248108, |
| "learning_rate": 3.7128828806431046e-05, |
| "loss": 2.0023, |
| "step": 35550 |
| }, |
| { |
| "epoch": 3.918547055586131, |
| "grad_norm": 0.17002084851264954, |
| "learning_rate": 3.6769963872947996e-05, |
| "loss": 2.0068, |
| "step": 35600 |
| }, |
| { |
| "epoch": 3.9240506329113924, |
| "grad_norm": 0.16855107247829437, |
| "learning_rate": 3.641259924231345e-05, |
| "loss": 1.9847, |
| "step": 35650 |
| }, |
| { |
| "epoch": 3.929554210236654, |
| "grad_norm": 0.190132737159729, |
| "learning_rate": 3.605673964962414e-05, |
| "loss": 1.9955, |
| "step": 35700 |
| }, |
| { |
| "epoch": 3.9350577875619153, |
| "grad_norm": 0.1558249294757843, |
| "learning_rate": 3.570238981003498e-05, |
| "loss": 1.9861, |
| "step": 35750 |
| }, |
| { |
| "epoch": 3.9405613648871767, |
| "grad_norm": 0.18199962377548218, |
| "learning_rate": 3.53495544186966e-05, |
| "loss": 2.0136, |
| "step": 35800 |
| }, |
| { |
| "epoch": 3.946064942212438, |
| "grad_norm": 0.17813271284103394, |
| "learning_rate": 3.499823815069306e-05, |
| "loss": 2.0012, |
| "step": 35850 |
| }, |
| { |
| "epoch": 3.9515685195376995, |
| "grad_norm": 0.18256065249443054, |
| "learning_rate": 3.464844566098008e-05, |
| "loss": 1.9892, |
| "step": 35900 |
| }, |
| { |
| "epoch": 3.957072096862961, |
| "grad_norm": 0.1784873753786087, |
| "learning_rate": 3.4300181584323126e-05, |
| "loss": 1.9983, |
| "step": 35950 |
| }, |
| { |
| "epoch": 3.9625756741882223, |
| "grad_norm": 0.18530665338039398, |
| "learning_rate": 3.3953450535236226e-05, |
| "loss": 2.0357, |
| "step": 36000 |
| }, |
| { |
| "epoch": 3.9680792515134837, |
| "grad_norm": 0.17063435912132263, |
| "learning_rate": 3.360825710792068e-05, |
| "loss": 2.0029, |
| "step": 36050 |
| }, |
| { |
| "epoch": 3.973582828838745, |
| "grad_norm": 0.1718268245458603, |
| "learning_rate": 3.326460587620427e-05, |
| "loss": 1.9762, |
| "step": 36100 |
| }, |
| { |
| "epoch": 3.9790864061640066, |
| "grad_norm": 0.15901614725589752, |
| "learning_rate": 3.292250139348057e-05, |
| "loss": 1.9952, |
| "step": 36150 |
| }, |
| { |
| "epoch": 3.984589983489268, |
| "grad_norm": 0.18968409299850464, |
| "learning_rate": 3.258194819264871e-05, |
| "loss": 2.0127, |
| "step": 36200 |
| }, |
| { |
| "epoch": 3.9900935608145294, |
| "grad_norm": 0.18861430883407593, |
| "learning_rate": 3.2242950786053296e-05, |
| "loss": 1.9981, |
| "step": 36250 |
| }, |
| { |
| "epoch": 3.995597138139791, |
| "grad_norm": 0.1702040284872055, |
| "learning_rate": 3.1905513665424445e-05, |
| "loss": 2.0023, |
| "step": 36300 |
| }, |
| { |
| "epoch": 4.001100715465053, |
| "grad_norm": 0.18254053592681885, |
| "learning_rate": 3.156964130181859e-05, |
| "loss": 2.0038, |
| "step": 36350 |
| }, |
| { |
| "epoch": 4.006604292790314, |
| "grad_norm": 0.18867474794387817, |
| "learning_rate": 3.1235338145558925e-05, |
| "loss": 1.9139, |
| "step": 36400 |
| }, |
| { |
| "epoch": 4.0121078701155755, |
| "grad_norm": 0.1872694045305252, |
| "learning_rate": 3.090260862617671e-05, |
| "loss": 1.9484, |
| "step": 36450 |
| }, |
| { |
| "epoch": 4.017611447440837, |
| "grad_norm": 0.17595593631267548, |
| "learning_rate": 3.057145715235236e-05, |
| "loss": 1.9543, |
| "step": 36500 |
| }, |
| { |
| "epoch": 4.023115024766098, |
| "grad_norm": 0.17811767756938934, |
| "learning_rate": 3.0241888111857145e-05, |
| "loss": 1.9268, |
| "step": 36550 |
| }, |
| { |
| "epoch": 4.02861860209136, |
| "grad_norm": 0.17598308622837067, |
| "learning_rate": 2.9913905871495037e-05, |
| "loss": 1.9439, |
| "step": 36600 |
| }, |
| { |
| "epoch": 4.034122179416621, |
| "grad_norm": 0.19237573444843292, |
| "learning_rate": 2.9587514777044874e-05, |
| "loss": 1.9044, |
| "step": 36650 |
| }, |
| { |
| "epoch": 4.0396257567418825, |
| "grad_norm": 0.1873626410961151, |
| "learning_rate": 2.926271915320259e-05, |
| "loss": 1.943, |
| "step": 36700 |
| }, |
| { |
| "epoch": 4.045129334067144, |
| "grad_norm": 0.2001618891954422, |
| "learning_rate": 2.893952330352423e-05, |
| "loss": 1.9311, |
| "step": 36750 |
| }, |
| { |
| "epoch": 4.050632911392405, |
| "grad_norm": 0.1838996410369873, |
| "learning_rate": 2.86179315103687e-05, |
| "loss": 1.9373, |
| "step": 36800 |
| }, |
| { |
| "epoch": 4.056136488717667, |
| "grad_norm": 0.16863055527210236, |
| "learning_rate": 2.8297948034841074e-05, |
| "loss": 1.9289, |
| "step": 36850 |
| }, |
| { |
| "epoch": 4.061640066042928, |
| "grad_norm": 0.19162894785404205, |
| "learning_rate": 2.797957711673619e-05, |
| "loss": 1.9293, |
| "step": 36900 |
| }, |
| { |
| "epoch": 4.06714364336819, |
| "grad_norm": 0.1854933500289917, |
| "learning_rate": 2.7662822974482423e-05, |
| "loss": 1.9336, |
| "step": 36950 |
| }, |
| { |
| "epoch": 4.072647220693451, |
| "grad_norm": 0.18568897247314453, |
| "learning_rate": 2.7347689805085733e-05, |
| "loss": 1.9431, |
| "step": 37000 |
| }, |
| { |
| "epoch": 4.078150798018712, |
| "grad_norm": 0.1947336494922638, |
| "learning_rate": 2.7034181784074205e-05, |
| "loss": 1.9345, |
| "step": 37050 |
| }, |
| { |
| "epoch": 4.083654375343974, |
| "grad_norm": 0.18774710595607758, |
| "learning_rate": 2.672230306544254e-05, |
| "loss": 1.9364, |
| "step": 37100 |
| }, |
| { |
| "epoch": 4.089157952669235, |
| "grad_norm": 0.1984746754169464, |
| "learning_rate": 2.6412057781597223e-05, |
| "loss": 1.954, |
| "step": 37150 |
| }, |
| { |
| "epoch": 4.094661529994497, |
| "grad_norm": 0.1908876746892929, |
| "learning_rate": 2.61034500433016e-05, |
| "loss": 1.9422, |
| "step": 37200 |
| }, |
| { |
| "epoch": 4.100165107319758, |
| "grad_norm": 0.18860432505607605, |
| "learning_rate": 2.579648393962147e-05, |
| "loss": 1.9649, |
| "step": 37250 |
| }, |
| { |
| "epoch": 4.105668684645019, |
| "grad_norm": 0.18678942322731018, |
| "learning_rate": 2.5491163537870922e-05, |
| "loss": 1.9544, |
| "step": 37300 |
| }, |
| { |
| "epoch": 4.111172261970281, |
| "grad_norm": 0.19560185074806213, |
| "learning_rate": 2.5187492883558435e-05, |
| "loss": 1.9396, |
| "step": 37350 |
| }, |
| { |
| "epoch": 4.116675839295542, |
| "grad_norm": 0.1825692355632782, |
| "learning_rate": 2.488547600033317e-05, |
| "loss": 1.949, |
| "step": 37400 |
| }, |
| { |
| "epoch": 4.122179416620804, |
| "grad_norm": 0.1951771229505539, |
| "learning_rate": 2.4585116889931873e-05, |
| "loss": 1.9482, |
| "step": 37450 |
| }, |
| { |
| "epoch": 4.127682993946065, |
| "grad_norm": 0.1859281212091446, |
| "learning_rate": 2.4286419532125668e-05, |
| "loss": 1.9512, |
| "step": 37500 |
| }, |
| { |
| "epoch": 4.1331865712713265, |
| "grad_norm": 0.19174982607364655, |
| "learning_rate": 2.3989387884667365e-05, |
| "loss": 1.9358, |
| "step": 37550 |
| }, |
| { |
| "epoch": 4.138690148596588, |
| "grad_norm": 0.187012180685997, |
| "learning_rate": 2.369402588323908e-05, |
| "loss": 1.9114, |
| "step": 37600 |
| }, |
| { |
| "epoch": 4.144193725921849, |
| "grad_norm": 0.20616699755191803, |
| "learning_rate": 2.340033744140004e-05, |
| "loss": 1.8965, |
| "step": 37650 |
| }, |
| { |
| "epoch": 4.149697303247111, |
| "grad_norm": 0.17636051774024963, |
| "learning_rate": 2.3108326450534725e-05, |
| "loss": 1.9549, |
| "step": 37700 |
| }, |
| { |
| "epoch": 4.155200880572372, |
| "grad_norm": 0.19219453632831573, |
| "learning_rate": 2.281799677980136e-05, |
| "loss": 1.8911, |
| "step": 37750 |
| }, |
| { |
| "epoch": 4.1607044578976335, |
| "grad_norm": 0.19742678105831146, |
| "learning_rate": 2.2529352276080513e-05, |
| "loss": 1.9562, |
| "step": 37800 |
| }, |
| { |
| "epoch": 4.166208035222895, |
| "grad_norm": 0.18049471080303192, |
| "learning_rate": 2.2242396763924292e-05, |
| "loss": 1.9169, |
| "step": 37850 |
| }, |
| { |
| "epoch": 4.171711612548156, |
| "grad_norm": 0.2108810693025589, |
| "learning_rate": 2.1957134045505588e-05, |
| "loss": 1.9363, |
| "step": 37900 |
| }, |
| { |
| "epoch": 4.177215189873418, |
| "grad_norm": 0.20711076259613037, |
| "learning_rate": 2.1673567900567667e-05, |
| "loss": 1.9685, |
| "step": 37950 |
| }, |
| { |
| "epoch": 4.182718767198679, |
| "grad_norm": 0.18670059740543365, |
| "learning_rate": 2.139170208637415e-05, |
| "loss": 1.9537, |
| "step": 38000 |
| }, |
| { |
| "epoch": 4.188222344523941, |
| "grad_norm": 0.17974555492401123, |
| "learning_rate": 2.1111540337659227e-05, |
| "loss": 1.9476, |
| "step": 38050 |
| }, |
| { |
| "epoch": 4.193725921849202, |
| "grad_norm": 0.18271256983280182, |
| "learning_rate": 2.083308636657811e-05, |
| "loss": 1.9535, |
| "step": 38100 |
| }, |
| { |
| "epoch": 4.199229499174463, |
| "grad_norm": 0.19260470569133759, |
| "learning_rate": 2.0556343862657855e-05, |
| "loss": 1.9477, |
| "step": 38150 |
| }, |
| { |
| "epoch": 4.204733076499725, |
| "grad_norm": 0.19317923486232758, |
| "learning_rate": 2.0281316492748595e-05, |
| "loss": 1.9518, |
| "step": 38200 |
| }, |
| { |
| "epoch": 4.210236653824986, |
| "grad_norm": 0.1854427605867386, |
| "learning_rate": 2.0008007900974738e-05, |
| "loss": 1.9245, |
| "step": 38250 |
| }, |
| { |
| "epoch": 4.215740231150248, |
| "grad_norm": 0.1781974732875824, |
| "learning_rate": 1.9736421708686934e-05, |
| "loss": 1.9387, |
| "step": 38300 |
| }, |
| { |
| "epoch": 4.221243808475509, |
| "grad_norm": 0.18503354489803314, |
| "learning_rate": 1.946656151441389e-05, |
| "loss": 1.9683, |
| "step": 38350 |
| }, |
| { |
| "epoch": 4.2267473858007705, |
| "grad_norm": 0.19728437066078186, |
| "learning_rate": 1.9198430893814798e-05, |
| "loss": 1.9546, |
| "step": 38400 |
| }, |
| { |
| "epoch": 4.232250963126032, |
| "grad_norm": 0.1957421451807022, |
| "learning_rate": 1.89320333996319e-05, |
| "loss": 1.9526, |
| "step": 38450 |
| }, |
| { |
| "epoch": 4.237754540451293, |
| "grad_norm": 0.18944865465164185, |
| "learning_rate": 1.866737256164349e-05, |
| "loss": 1.9467, |
| "step": 38500 |
| }, |
| { |
| "epoch": 4.243258117776555, |
| "grad_norm": 0.1760573536157608, |
| "learning_rate": 1.8404451886617005e-05, |
| "loss": 1.9142, |
| "step": 38550 |
| }, |
| { |
| "epoch": 4.248761695101816, |
| "grad_norm": 0.19618487358093262, |
| "learning_rate": 1.8143274858262702e-05, |
| "loss": 1.9491, |
| "step": 38600 |
| }, |
| { |
| "epoch": 4.2542652724270775, |
| "grad_norm": 0.19029423594474792, |
| "learning_rate": 1.7883844937187495e-05, |
| "loss": 1.9553, |
| "step": 38650 |
| }, |
| { |
| "epoch": 4.259768849752339, |
| "grad_norm": 0.19688698649406433, |
| "learning_rate": 1.762616556084891e-05, |
| "loss": 1.9698, |
| "step": 38700 |
| }, |
| { |
| "epoch": 4.2652724270776, |
| "grad_norm": 0.16535942256450653, |
| "learning_rate": 1.737024014350983e-05, |
| "loss": 1.9395, |
| "step": 38750 |
| }, |
| { |
| "epoch": 4.270776004402862, |
| "grad_norm": 0.19671253859996796, |
| "learning_rate": 1.7116072076193e-05, |
| "loss": 1.9343, |
| "step": 38800 |
| }, |
| { |
| "epoch": 4.276279581728123, |
| "grad_norm": 0.18672049045562744, |
| "learning_rate": 1.6863664726636278e-05, |
| "loss": 1.939, |
| "step": 38850 |
| }, |
| { |
| "epoch": 4.281783159053385, |
| "grad_norm": 0.19714199006557465, |
| "learning_rate": 1.6613021439247914e-05, |
| "loss": 1.9455, |
| "step": 38900 |
| }, |
| { |
| "epoch": 4.287286736378646, |
| "grad_norm": 0.2018548846244812, |
| "learning_rate": 1.6364145535062172e-05, |
| "loss": 1.9169, |
| "step": 38950 |
| }, |
| { |
| "epoch": 4.292790313703907, |
| "grad_norm": 0.19057820737361908, |
| "learning_rate": 1.611704031169555e-05, |
| "loss": 1.956, |
| "step": 39000 |
| }, |
| { |
| "epoch": 4.298293891029169, |
| "grad_norm": 0.18719470500946045, |
| "learning_rate": 1.5871709043302876e-05, |
| "loss": 1.9483, |
| "step": 39050 |
| }, |
| { |
| "epoch": 4.30379746835443, |
| "grad_norm": 0.19479648768901825, |
| "learning_rate": 1.5628154980533984e-05, |
| "loss": 1.9507, |
| "step": 39100 |
| }, |
| { |
| "epoch": 4.309301045679692, |
| "grad_norm": 0.2034020870923996, |
| "learning_rate": 1.538638135049071e-05, |
| "loss": 1.9343, |
| "step": 39150 |
| }, |
| { |
| "epoch": 4.314804623004953, |
| "grad_norm": 0.18177463114261627, |
| "learning_rate": 1.5146391356684029e-05, |
| "loss": 1.9481, |
| "step": 39200 |
| }, |
| { |
| "epoch": 4.3203082003302145, |
| "grad_norm": 0.1850568652153015, |
| "learning_rate": 1.4908188178991714e-05, |
| "loss": 1.9324, |
| "step": 39250 |
| }, |
| { |
| "epoch": 4.325811777655476, |
| "grad_norm": 0.1880549043416977, |
| "learning_rate": 1.4671774973616102e-05, |
| "loss": 1.9406, |
| "step": 39300 |
| }, |
| { |
| "epoch": 4.331315354980737, |
| "grad_norm": 0.18556420505046844, |
| "learning_rate": 1.4437154873042279e-05, |
| "loss": 1.9516, |
| "step": 39350 |
| }, |
| { |
| "epoch": 4.336818932305999, |
| "grad_norm": 0.19034932553768158, |
| "learning_rate": 1.420433098599672e-05, |
| "loss": 1.9477, |
| "step": 39400 |
| }, |
| { |
| "epoch": 4.34232250963126, |
| "grad_norm": 0.18247225880622864, |
| "learning_rate": 1.3973306397405909e-05, |
| "loss": 1.9552, |
| "step": 39450 |
| }, |
| { |
| "epoch": 4.3478260869565215, |
| "grad_norm": 0.19472351670265198, |
| "learning_rate": 1.3744084168355612e-05, |
| "loss": 1.9543, |
| "step": 39500 |
| }, |
| { |
| "epoch": 4.353329664281783, |
| "grad_norm": 0.18814486265182495, |
| "learning_rate": 1.3516667336050219e-05, |
| "loss": 1.956, |
| "step": 39550 |
| }, |
| { |
| "epoch": 4.358833241607044, |
| "grad_norm": 0.18639877438545227, |
| "learning_rate": 1.3291058913772517e-05, |
| "loss": 1.9262, |
| "step": 39600 |
| }, |
| { |
| "epoch": 4.364336818932306, |
| "grad_norm": 0.19503189623355865, |
| "learning_rate": 1.306726189084385e-05, |
| "loss": 1.9768, |
| "step": 39650 |
| }, |
| { |
| "epoch": 4.369840396257567, |
| "grad_norm": 0.1823161244392395, |
| "learning_rate": 1.2845279232584354e-05, |
| "loss": 1.9588, |
| "step": 39700 |
| }, |
| { |
| "epoch": 4.375343973582829, |
| "grad_norm": 0.19661200046539307, |
| "learning_rate": 1.2625113880273786e-05, |
| "loss": 1.9576, |
| "step": 39750 |
| }, |
| { |
| "epoch": 4.38084755090809, |
| "grad_norm": 0.17353294789791107, |
| "learning_rate": 1.2406768751112572e-05, |
| "loss": 1.9249, |
| "step": 39800 |
| }, |
| { |
| "epoch": 4.386351128233351, |
| "grad_norm": 0.1832091063261032, |
| "learning_rate": 1.2190246738183074e-05, |
| "loss": 1.9169, |
| "step": 39850 |
| }, |
| { |
| "epoch": 4.391854705558613, |
| "grad_norm": 0.18823185563087463, |
| "learning_rate": 1.197555071041122e-05, |
| "loss": 1.9368, |
| "step": 39900 |
| }, |
| { |
| "epoch": 4.397358282883874, |
| "grad_norm": 0.1941109150648117, |
| "learning_rate": 1.1762683512528642e-05, |
| "loss": 1.96, |
| "step": 39950 |
| }, |
| { |
| "epoch": 4.402861860209136, |
| "grad_norm": 0.1973351240158081, |
| "learning_rate": 1.155164796503486e-05, |
| "loss": 1.9304, |
| "step": 40000 |
| } |
| ], |
| "logging_steps": 50, |
| "max_steps": 45425, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 5, |
| "save_steps": 10000, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": false |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 5.91746903834624e+17, |
| "train_batch_size": 16, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|