{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.125, "eval_steps": 250, "global_step": 1000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.000125, "grad_norm": 2.537714958190918, "learning_rate": 1.0000000000000002e-06, "loss": 0.5468, "loss/crossentropy": 2.2066214084625244, "loss/hidden": 0.248046875, "loss/logits": 0.03443578630685806, "loss/reg": 0.026429571211338043, "step": 1 }, { "epoch": 0.00025, "grad_norm": 2.4728448390960693, "learning_rate": 2.0000000000000003e-06, "loss": 0.6642, "loss/crossentropy": 2.132329225540161, "loss/hidden": 0.345703125, "loss/logits": 0.05424630641937256, "loss/reg": 0.026429571211338043, "step": 2 }, { "epoch": 0.000375, "grad_norm": 2.773984670639038, "learning_rate": 3e-06, "loss": 0.5822, "loss/crossentropy": 2.3457791805267334, "loss/hidden": 0.2734375, "loss/logits": 0.044443465769290924, "loss/reg": 0.02642953023314476, "step": 3 }, { "epoch": 0.0005, "grad_norm": 4.14040470123291, "learning_rate": 4.000000000000001e-06, "loss": 0.7192, "loss/crossentropy": 2.7209200859069824, "loss/hidden": 0.35546875, "loss/logits": 0.09940779209136963, "loss/reg": 0.026429466903209686, "step": 4 }, { "epoch": 0.000625, "grad_norm": 1.9164764881134033, "learning_rate": 5e-06, "loss": 0.5467, "loss/crossentropy": 2.4304752349853516, "loss/hidden": 0.244140625, "loss/logits": 0.03826362267136574, "loss/reg": 0.02642936445772648, "step": 5 }, { "epoch": 0.00075, "grad_norm": 1.9878246784210205, "learning_rate": 6e-06, "loss": 0.517, "loss/crossentropy": 2.472181797027588, "loss/hidden": 0.2255859375, "loss/logits": 0.027161670848727226, "loss/reg": 0.02642924338579178, "step": 6 }, { "epoch": 0.000875, "grad_norm": 2.1939733028411865, "learning_rate": 7.000000000000001e-06, "loss": 0.6043, "loss/crossentropy": 2.241501808166504, "loss/hidden": 0.298828125, "loss/logits": 0.04118040204048157, "loss/reg": 0.02642909064888954, "step": 7 }, { "epoch": 0.001, "grad_norm": 3.516223907470703, "learning_rate": 8.000000000000001e-06, "loss": 0.5199, "loss/crossentropy": 2.409766912460327, "loss/hidden": 0.2236328125, "loss/logits": 0.032000549137592316, "loss/reg": 0.02642889879643917, "step": 8 }, { "epoch": 0.001125, "grad_norm": 1.9335486888885498, "learning_rate": 9e-06, "loss": 0.5575, "loss/crossentropy": 2.6256861686706543, "loss/hidden": 0.255859375, "loss/logits": 0.037392452359199524, "loss/reg": 0.02642873302102089, "step": 9 }, { "epoch": 0.00125, "grad_norm": 1.6782876253128052, "learning_rate": 1e-05, "loss": 0.5162, "loss/crossentropy": 2.1947107315063477, "loss/hidden": 0.2255859375, "loss/logits": 0.026354767382144928, "loss/reg": 0.026428483426570892, "step": 10 }, { "epoch": 0.001375, "grad_norm": 10.848552703857422, "learning_rate": 1.1000000000000001e-05, "loss": 1.0046, "loss/crossentropy": 3.0539069175720215, "loss/hidden": 0.640625, "loss/logits": 0.09970991313457489, "loss/reg": 0.026428230106830597, "step": 11 }, { "epoch": 0.0015, "grad_norm": 2.237061023712158, "learning_rate": 1.2e-05, "loss": 0.563, "loss/crossentropy": 2.5601325035095215, "loss/hidden": 0.248046875, "loss/logits": 0.050660137087106705, "loss/reg": 0.02642793208360672, "step": 12 }, { "epoch": 0.001625, "grad_norm": 1.4406346082687378, "learning_rate": 1.3000000000000001e-05, "loss": 0.507, "loss/crossentropy": 1.965380311012268, "loss/hidden": 0.2197265625, "loss/logits": 0.02295786701142788, "loss/reg": 0.02642756886780262, "step": 13 }, { "epoch": 0.00175, "grad_norm": 3.0757036209106445, "learning_rate": 1.4000000000000001e-05, "loss": 0.7761, "loss/crossentropy": 2.15138840675354, "loss/hidden": 0.4375, "loss/logits": 0.07431840896606445, "loss/reg": 0.026427194476127625, "step": 14 }, { "epoch": 0.001875, "grad_norm": 2.8731143474578857, "learning_rate": 1.5e-05, "loss": 0.4684, "loss/crossentropy": 2.5530812740325928, "loss/hidden": 0.1845703125, "loss/logits": 0.019558344036340714, "loss/reg": 0.026426764205098152, "step": 15 }, { "epoch": 0.002, "grad_norm": 2.5288755893707275, "grad_norm_var": 4.846526347105633, "learning_rate": 1.6000000000000003e-05, "loss": 0.5781, "loss/crossentropy": 2.5096747875213623, "loss/hidden": 0.275390625, "loss/logits": 0.0384209081530571, "loss/reg": 0.026426298543810844, "step": 16 }, { "epoch": 0.002125, "grad_norm": 2.0281474590301514, "grad_norm_var": 4.89482291121508, "learning_rate": 1.7000000000000003e-05, "loss": 0.5318, "loss/crossentropy": 2.396097421646118, "loss/hidden": 0.232421875, "loss/logits": 0.03516196087002754, "loss/reg": 0.026425909250974655, "step": 17 }, { "epoch": 0.00225, "grad_norm": 2.4487411975860596, "grad_norm_var": 4.896482229626747, "learning_rate": 1.8e-05, "loss": 0.5984, "loss/crossentropy": 2.3916616439819336, "loss/hidden": 0.28125, "loss/logits": 0.052923329174518585, "loss/reg": 0.026425503194332123, "step": 18 }, { "epoch": 0.002375, "grad_norm": 1.986022710800171, "grad_norm_var": 4.956548008938709, "learning_rate": 1.9e-05, "loss": 0.5504, "loss/crossentropy": 2.4791200160980225, "loss/hidden": 0.2421875, "loss/logits": 0.04391499236226082, "loss/reg": 0.026425078511238098, "step": 19 }, { "epoch": 0.0025, "grad_norm": 2.0934784412384033, "grad_norm_var": 4.887277710992484, "learning_rate": 2e-05, "loss": 0.5369, "loss/crossentropy": 2.1741297245025635, "loss/hidden": 0.2392578125, "loss/logits": 0.03335873782634735, "loss/reg": 0.026424556970596313, "step": 20 }, { "epoch": 0.002625, "grad_norm": 1.9445254802703857, "grad_norm_var": 4.884025740026245, "learning_rate": 2.1e-05, "loss": 0.4865, "loss/crossentropy": 2.45112943649292, "loss/hidden": 0.1962890625, "loss/logits": 0.025960583239793777, "loss/reg": 0.02642405778169632, "step": 21 }, { "epoch": 0.00275, "grad_norm": 3.070704221725464, "grad_norm_var": 4.8399171328504895, "learning_rate": 2.2000000000000003e-05, "loss": 0.5887, "loss/crossentropy": 2.1550512313842773, "loss/hidden": 0.275390625, "loss/logits": 0.049097511917352676, "loss/reg": 0.02642347477376461, "step": 22 }, { "epoch": 0.002875, "grad_norm": 2.2452821731567383, "grad_norm_var": 4.835466428034186, "learning_rate": 2.3000000000000003e-05, "loss": 0.5482, "loss/crossentropy": 2.1640255451202393, "loss/hidden": 0.2470703125, "loss/logits": 0.036868080496788025, "loss/reg": 0.026422815397381783, "step": 23 }, { "epoch": 0.003, "grad_norm": 2.032148838043213, "grad_norm_var": 4.84560617678243, "learning_rate": 2.4e-05, "loss": 0.5756, "loss/crossentropy": 2.323482036590576, "loss/hidden": 0.271484375, "loss/logits": 0.03993295133113861, "loss/reg": 0.02642211876809597, "step": 24 }, { "epoch": 0.003125, "grad_norm": 1.763465404510498, "grad_norm_var": 4.8665883230545335, "learning_rate": 2.5e-05, "loss": 0.5139, "loss/crossentropy": 2.33661150932312, "loss/hidden": 0.22265625, "loss/logits": 0.02701444923877716, "loss/reg": 0.026421383023262024, "step": 25 }, { "epoch": 0.00325, "grad_norm": 1.7001625299453735, "grad_norm_var": 4.863438686484135, "learning_rate": 2.6000000000000002e-05, "loss": 0.5696, "loss/crossentropy": 2.2305383682250977, "loss/hidden": 0.275390625, "loss/logits": 0.02997000887989998, "loss/reg": 0.026420695707201958, "step": 26 }, { "epoch": 0.003375, "grad_norm": 3.474130392074585, "grad_norm_var": 0.318824614624526, "learning_rate": 2.7000000000000002e-05, "loss": 0.5456, "loss/crossentropy": 2.1680750846862793, "loss/hidden": 0.2451171875, "loss/logits": 0.036261945962905884, "loss/reg": 0.0264199897646904, "step": 27 }, { "epoch": 0.0035, "grad_norm": 3.8201987743377686, "grad_norm_var": 0.46030846745106163, "learning_rate": 2.8000000000000003e-05, "loss": 0.5181, "loss/crossentropy": 2.418672800064087, "loss/hidden": 0.224609375, "loss/logits": 0.029263213276863098, "loss/reg": 0.02641921117901802, "step": 28 }, { "epoch": 0.003625, "grad_norm": 1.9781090021133423, "grad_norm_var": 0.40905077024170067, "learning_rate": 2.9e-05, "loss": 0.5379, "loss/crossentropy": 2.390868663787842, "loss/hidden": 0.2392578125, "loss/logits": 0.034465983510017395, "loss/reg": 0.026418352499604225, "step": 29 }, { "epoch": 0.00375, "grad_norm": 1.6551319360733032, "grad_norm_var": 0.41503895204729413, "learning_rate": 3e-05, "loss": 0.496, "loss/crossentropy": 2.5960400104522705, "loss/hidden": 0.205078125, "loss/logits": 0.02678578905761242, "loss/reg": 0.026417305693030357, "step": 30 }, { "epoch": 0.003875, "grad_norm": 1.7136921882629395, "grad_norm_var": 0.4185952392532807, "learning_rate": 3.1e-05, "loss": 0.5235, "loss/crossentropy": 2.349839687347412, "loss/hidden": 0.2275390625, "loss/logits": 0.031792763620615005, "loss/reg": 0.026416433975100517, "step": 31 }, { "epoch": 0.004, "grad_norm": 1.9992157220840454, "grad_norm_var": 0.41856547198587274, "learning_rate": 3.2000000000000005e-05, "loss": 0.4928, "loss/crossentropy": 2.3164803981781006, "loss/hidden": 0.2041015625, "loss/logits": 0.024579893797636032, "loss/reg": 0.02641524001955986, "step": 32 }, { "epoch": 0.004125, "grad_norm": 2.705052614212036, "grad_norm_var": 0.42744416353313663, "learning_rate": 3.3e-05, "loss": 0.5732, "loss/crossentropy": 2.42107892036438, "loss/hidden": 0.275390625, "loss/logits": 0.03370767831802368, "loss/reg": 0.02641397900879383, "step": 33 }, { "epoch": 0.00425, "grad_norm": 1.8898464441299438, "grad_norm_var": 0.4350913020843951, "learning_rate": 3.4000000000000007e-05, "loss": 0.5531, "loss/crossentropy": 2.4147770404815674, "loss/hidden": 0.25390625, "loss/logits": 0.03504405915737152, "loss/reg": 0.026412710547447205, "step": 34 }, { "epoch": 0.004375, "grad_norm": 4.9570159912109375, "grad_norm_var": 0.8804344329355491, "learning_rate": 3.5e-05, "loss": 0.6763, "loss/crossentropy": 1.6753497123718262, "loss/hidden": 0.376953125, "loss/logits": 0.03519564867019653, "loss/reg": 0.0264116358011961, "step": 35 }, { "epoch": 0.0045, "grad_norm": 4.928956508636475, "grad_norm_var": 1.2518721453244992, "learning_rate": 3.6e-05, "loss": 0.7329, "loss/crossentropy": 2.6104867458343506, "loss/hidden": 0.400390625, "loss/logits": 0.06845290958881378, "loss/reg": 0.02641039527952671, "step": 36 }, { "epoch": 0.004625, "grad_norm": 7.503647327423096, "grad_norm_var": 2.684651641752033, "learning_rate": 3.7e-05, "loss": 0.6258, "loss/crossentropy": 2.2158656120300293, "loss/hidden": 0.318359375, "loss/logits": 0.043342188000679016, "loss/reg": 0.026409219950437546, "step": 37 }, { "epoch": 0.00475, "grad_norm": 2.6838622093200684, "grad_norm_var": 2.6885420074665602, "learning_rate": 3.8e-05, "loss": 0.5939, "loss/crossentropy": 2.344879627227783, "loss/hidden": 0.28515625, "loss/logits": 0.04461552947759628, "loss/reg": 0.026408080011606216, "step": 38 }, { "epoch": 0.004875, "grad_norm": 3.357893705368042, "grad_norm_var": 2.662758933855309, "learning_rate": 3.9000000000000006e-05, "loss": 0.5729, "loss/crossentropy": 2.6759543418884277, "loss/hidden": 0.275390625, "loss/logits": 0.033413954079151154, "loss/reg": 0.026406895369291306, "step": 39 }, { "epoch": 0.005, "grad_norm": 3.0177316665649414, "grad_norm_var": 2.5949485604856193, "learning_rate": 4e-05, "loss": 0.7498, "loss/crossentropy": 2.2261273860931396, "loss/hidden": 0.408203125, "loss/logits": 0.07758316397666931, "loss/reg": 0.026405224576592445, "step": 40 }, { "epoch": 0.005125, "grad_norm": 2.1196699142456055, "grad_norm_var": 2.54074274703229, "learning_rate": 4.1e-05, "loss": 0.6396, "loss/crossentropy": 2.193378448486328, "loss/hidden": 0.30859375, "loss/logits": 0.06692355871200562, "loss/reg": 0.026403924450278282, "step": 41 }, { "epoch": 0.00525, "grad_norm": 2.456051826477051, "grad_norm_var": 2.435973046683167, "learning_rate": 4.2e-05, "loss": 0.5571, "loss/crossentropy": 1.9526888132095337, "loss/hidden": 0.26171875, "loss/logits": 0.03133418411016464, "loss/reg": 0.026402529329061508, "step": 42 }, { "epoch": 0.005375, "grad_norm": 2.257375717163086, "grad_norm_var": 2.474501380785125, "learning_rate": 4.3e-05, "loss": 0.5544, "loss/crossentropy": 2.3284847736358643, "loss/hidden": 0.25390625, "loss/logits": 0.03650724142789841, "loss/reg": 0.026400938630104065, "step": 43 }, { "epoch": 0.0055, "grad_norm": 2.9145264625549316, "grad_norm_var": 2.4345975605903694, "learning_rate": 4.4000000000000006e-05, "loss": 0.5175, "loss/crossentropy": 2.295241594314575, "loss/hidden": 0.220703125, "loss/logits": 0.03284794092178345, "loss/reg": 0.026399515569210052, "step": 44 }, { "epoch": 0.005625, "grad_norm": 3.1294264793395996, "grad_norm_var": 2.3592519473156615, "learning_rate": 4.5e-05, "loss": 0.5567, "loss/crossentropy": 2.660597085952759, "loss/hidden": 0.255859375, "loss/logits": 0.03686758130788803, "loss/reg": 0.026398126035928726, "step": 45 }, { "epoch": 0.00575, "grad_norm": 2.197265863418579, "grad_norm_var": 2.2745842657817748, "learning_rate": 4.600000000000001e-05, "loss": 0.5512, "loss/crossentropy": 2.3832643032073975, "loss/hidden": 0.2490234375, "loss/logits": 0.038256023079156876, "loss/reg": 0.02639671601355076, "step": 46 }, { "epoch": 0.005875, "grad_norm": 2.883378744125366, "grad_norm_var": 2.141634704665381, "learning_rate": 4.7e-05, "loss": 0.5298, "loss/crossentropy": 2.6035244464874268, "loss/hidden": 0.2333984375, "loss/logits": 0.03240815922617912, "loss/reg": 0.026395246386528015, "step": 47 }, { "epoch": 0.006, "grad_norm": 3.1519744396209717, "grad_norm_var": 2.0420385103816727, "learning_rate": 4.8e-05, "loss": 0.5385, "loss/crossentropy": 2.250037908554077, "loss/hidden": 0.244140625, "loss/logits": 0.03043752908706665, "loss/reg": 0.026393571868538857, "step": 48 }, { "epoch": 0.006125, "grad_norm": 3.187680244445801, "grad_norm_var": 2.0209109756516943, "learning_rate": 4.9e-05, "loss": 0.5614, "loss/crossentropy": 2.366483688354492, "loss/hidden": 0.263671875, "loss/logits": 0.033859170973300934, "loss/reg": 0.02639181725680828, "step": 49 }, { "epoch": 0.00625, "grad_norm": 2.3717658519744873, "grad_norm_var": 1.9454730589909708, "learning_rate": 5e-05, "loss": 0.5865, "loss/crossentropy": 2.007732391357422, "loss/hidden": 0.2890625, "loss/logits": 0.0335388109087944, "loss/reg": 0.02638987824320793, "step": 50 }, { "epoch": 0.006375, "grad_norm": 3.658735990524292, "grad_norm_var": 1.767425501826429, "learning_rate": 5.1000000000000006e-05, "loss": 0.5028, "loss/crossentropy": 2.511072874069214, "loss/hidden": 0.2099609375, "loss/logits": 0.02893088385462761, "loss/reg": 0.026387827470898628, "step": 51 }, { "epoch": 0.0065, "grad_norm": 2.5912654399871826, "grad_norm_var": 1.582150273328572, "learning_rate": 5.2000000000000004e-05, "loss": 0.5619, "loss/crossentropy": 2.3280093669891357, "loss/hidden": 0.263671875, "loss/logits": 0.03436018154025078, "loss/reg": 0.026385735720396042, "step": 52 }, { "epoch": 0.006625, "grad_norm": 2.0419421195983887, "grad_norm_var": 0.23432357463535497, "learning_rate": 5.300000000000001e-05, "loss": 0.5674, "loss/crossentropy": 2.3851194381713867, "loss/hidden": 0.263671875, "loss/logits": 0.039869021624326706, "loss/reg": 0.026383817195892334, "step": 53 }, { "epoch": 0.00675, "grad_norm": 2.4164810180664062, "grad_norm_var": 0.24119551692934707, "learning_rate": 5.4000000000000005e-05, "loss": 0.6087, "loss/crossentropy": 2.6006996631622314, "loss/hidden": 0.296875, "loss/logits": 0.04797635227441788, "loss/reg": 0.026381801813840866, "step": 54 }, { "epoch": 0.006875, "grad_norm": 2.697831153869629, "grad_norm_var": 0.2135682431387058, "learning_rate": 5.500000000000001e-05, "loss": 0.523, "loss/crossentropy": 2.472208261489868, "loss/hidden": 0.2275390625, "loss/logits": 0.031705208122730255, "loss/reg": 0.026379752904176712, "step": 55 }, { "epoch": 0.007, "grad_norm": 4.182509422302246, "grad_norm_var": 0.34874494246430365, "learning_rate": 5.6000000000000006e-05, "loss": 0.6766, "loss/crossentropy": 2.693652868270874, "loss/hidden": 0.3671875, "loss/logits": 0.04566050320863724, "loss/reg": 0.02637762948870659, "step": 56 }, { "epoch": 0.007125, "grad_norm": 2.231238842010498, "grad_norm_var": 0.33990645656106155, "learning_rate": 5.6999999999999996e-05, "loss": 0.5811, "loss/crossentropy": 2.4935543537139893, "loss/hidden": 0.28125, "loss/logits": 0.03613065183162689, "loss/reg": 0.026375366374850273, "step": 57 }, { "epoch": 0.00725, "grad_norm": 2.0192184448242188, "grad_norm_var": 0.37029866859904437, "learning_rate": 5.8e-05, "loss": 0.5285, "loss/crossentropy": 2.192227840423584, "loss/hidden": 0.2294921875, "loss/logits": 0.03531934320926666, "loss/reg": 0.02637314423918724, "step": 58 }, { "epoch": 0.007375, "grad_norm": 2.3108532428741455, "grad_norm_var": 0.36699486123436575, "learning_rate": 5.9e-05, "loss": 0.507, "loss/crossentropy": 2.39101243019104, "loss/hidden": 0.2177734375, "loss/logits": 0.025539016351103783, "loss/reg": 0.02637065388262272, "step": 59 }, { "epoch": 0.0075, "grad_norm": 2.049551486968994, "grad_norm_var": 0.3946811437011318, "learning_rate": 6e-05, "loss": 0.5351, "loss/crossentropy": 2.7062017917633057, "loss/hidden": 0.2353515625, "loss/logits": 0.03605186939239502, "loss/reg": 0.026368385180830956, "step": 60 }, { "epoch": 0.007625, "grad_norm": 2.6327223777770996, "grad_norm_var": 0.38133460463904284, "learning_rate": 6.1e-05, "loss": 0.5344, "loss/crossentropy": 2.0810956954956055, "loss/hidden": 0.23828125, "loss/logits": 0.03246723860502243, "loss/reg": 0.02636607363820076, "step": 61 }, { "epoch": 0.00775, "grad_norm": 2.06585955619812, "grad_norm_var": 0.39059185941555535, "learning_rate": 6.2e-05, "loss": 0.5202, "loss/crossentropy": 2.6240835189819336, "loss/hidden": 0.224609375, "loss/logits": 0.03200242295861244, "loss/reg": 0.026363445445895195, "step": 62 }, { "epoch": 0.007875, "grad_norm": 2.109790563583374, "grad_norm_var": 0.40452198957435875, "learning_rate": 6.3e-05, "loss": 0.5249, "loss/crossentropy": 2.587536573410034, "loss/hidden": 0.2294921875, "loss/logits": 0.03178905323147774, "loss/reg": 0.02636083774268627, "step": 63 }, { "epoch": 0.008, "grad_norm": 3.818783760070801, "grad_norm_var": 0.48072296241430573, "learning_rate": 6.400000000000001e-05, "loss": 0.6632, "loss/crossentropy": 2.011171817779541, "loss/hidden": 0.353515625, "loss/logits": 0.04613731801509857, "loss/reg": 0.02635800838470459, "step": 64 }, { "epoch": 0.008125, "grad_norm": 2.4136369228363037, "grad_norm_var": 0.46258887231494605, "learning_rate": 6.500000000000001e-05, "loss": 0.522, "loss/crossentropy": 2.600787401199341, "loss/hidden": 0.224609375, "loss/logits": 0.03384025767445564, "loss/reg": 0.026355121284723282, "step": 65 }, { "epoch": 0.00825, "grad_norm": 2.3908252716064453, "grad_norm_var": 0.46202963925557394, "learning_rate": 6.6e-05, "loss": 0.5859, "loss/crossentropy": 2.1056201457977295, "loss/hidden": 0.279296875, "loss/logits": 0.04309317469596863, "loss/reg": 0.02635251171886921, "step": 66 }, { "epoch": 0.008375, "grad_norm": 2.5190653800964355, "grad_norm_var": 0.38262308323354144, "learning_rate": 6.7e-05, "loss": 0.5391, "loss/crossentropy": 2.5527184009552, "loss/hidden": 0.2421875, "loss/logits": 0.0334152951836586, "loss/reg": 0.02634957991540432, "step": 67 }, { "epoch": 0.0085, "grad_norm": 2.500368595123291, "grad_norm_var": 0.3824057294099087, "learning_rate": 6.800000000000001e-05, "loss": 0.5958, "loss/crossentropy": 2.30499267578125, "loss/hidden": 0.28125, "loss/logits": 0.05108712613582611, "loss/reg": 0.026346800848841667, "step": 68 }, { "epoch": 0.008625, "grad_norm": 2.7905988693237305, "grad_norm_var": 0.3692126592154902, "learning_rate": 6.9e-05, "loss": 0.6049, "loss/crossentropy": 2.3807146549224854, "loss/hidden": 0.294921875, "loss/logits": 0.04648623988032341, "loss/reg": 0.026344334706664085, "step": 69 }, { "epoch": 0.00875, "grad_norm": 2.147470235824585, "grad_norm_var": 0.3793077808516782, "learning_rate": 7e-05, "loss": 0.5345, "loss/crossentropy": 2.627505302429199, "loss/hidden": 0.236328125, "loss/logits": 0.034769318997859955, "loss/reg": 0.026341637596488, "step": 70 }, { "epoch": 0.008875, "grad_norm": 2.6987268924713135, "grad_norm_var": 0.3793248871627156, "learning_rate": 7.1e-05, "loss": 0.5722, "loss/crossentropy": 2.382685899734497, "loss/hidden": 0.271484375, "loss/logits": 0.03735022246837616, "loss/reg": 0.02633870206773281, "step": 71 }, { "epoch": 0.009, "grad_norm": 3.085085153579712, "grad_norm_var": 0.2164648496489896, "learning_rate": 7.2e-05, "loss": 0.5882, "loss/crossentropy": 2.371429681777954, "loss/hidden": 0.27734375, "loss/logits": 0.04748620092868805, "loss/reg": 0.02633603662252426, "step": 72 }, { "epoch": 0.009125, "grad_norm": 4.158353328704834, "grad_norm_var": 0.3829897758196862, "learning_rate": 7.3e-05, "loss": 0.8663, "loss/crossentropy": 2.29622745513916, "loss/hidden": 0.5234375, "loss/logits": 0.07955377548933029, "loss/reg": 0.026333071291446686, "step": 73 }, { "epoch": 0.00925, "grad_norm": 2.111111879348755, "grad_norm_var": 0.37631661688178514, "learning_rate": 7.4e-05, "loss": 0.5468, "loss/crossentropy": 2.29744815826416, "loss/hidden": 0.24609375, "loss/logits": 0.037446070462465286, "loss/reg": 0.026330096647143364, "step": 74 }, { "epoch": 0.009375, "grad_norm": 2.545919179916382, "grad_norm_var": 0.37031037444480336, "learning_rate": 7.500000000000001e-05, "loss": 0.625, "loss/crossentropy": 2.4376375675201416, "loss/hidden": 0.306640625, "loss/logits": 0.05505819618701935, "loss/reg": 0.026326792314648628, "step": 75 }, { "epoch": 0.0095, "grad_norm": 2.362215042114258, "grad_norm_var": 0.35233204024674003, "learning_rate": 7.6e-05, "loss": 0.5747, "loss/crossentropy": 2.677924156188965, "loss/hidden": 0.267578125, "loss/logits": 0.04392882436513901, "loss/reg": 0.026323769241571426, "step": 76 }, { "epoch": 0.009625, "grad_norm": 3.135709762573242, "grad_norm_var": 0.3671929300455114, "learning_rate": 7.7e-05, "loss": 0.7113, "loss/crossentropy": 1.972798466682434, "loss/hidden": 0.384765625, "loss/logits": 0.06329117715358734, "loss/reg": 0.026320943608880043, "step": 77 }, { "epoch": 0.00975, "grad_norm": 4.418634414672852, "grad_norm_var": 0.5210260544686395, "learning_rate": 7.800000000000001e-05, "loss": 0.6851, "loss/crossentropy": 2.558809518814087, "loss/hidden": 0.357421875, "loss/logits": 0.06447892636060715, "loss/reg": 0.02631756290793419, "step": 78 }, { "epoch": 0.009875, "grad_norm": 3.9261293411254883, "grad_norm_var": 0.55391532710314, "learning_rate": 7.900000000000001e-05, "loss": 0.5968, "loss/crossentropy": 2.6102137565612793, "loss/hidden": 0.28515625, "loss/logits": 0.04846350848674774, "loss/reg": 0.026314500719308853, "step": 79 }, { "epoch": 0.01, "grad_norm": 3.1532020568847656, "grad_norm_var": 0.5035194586586452, "learning_rate": 8e-05, "loss": 0.7066, "loss/crossentropy": 2.36220121383667, "loss/hidden": 0.38671875, "loss/logits": 0.05672474205493927, "loss/reg": 0.026311254128813744, "step": 80 }, { "epoch": 0.010125, "grad_norm": 3.557161808013916, "grad_norm_var": 0.5115010248662256, "learning_rate": 8.1e-05, "loss": 0.6647, "loss/crossentropy": 2.412325859069824, "loss/hidden": 0.345703125, "loss/logits": 0.05596970394253731, "loss/reg": 0.026307715103030205, "step": 81 }, { "epoch": 0.01025, "grad_norm": 2.2158408164978027, "grad_norm_var": 0.5268993015208875, "learning_rate": 8.2e-05, "loss": 0.5575, "loss/crossentropy": 2.3789050579071045, "loss/hidden": 0.255859375, "loss/logits": 0.03855578228831291, "loss/reg": 0.0263040903955698, "step": 82 }, { "epoch": 0.010375, "grad_norm": 3.2140979766845703, "grad_norm_var": 0.5164286227091848, "learning_rate": 8.3e-05, "loss": 0.5548, "loss/crossentropy": 2.401925802230835, "loss/hidden": 0.2578125, "loss/logits": 0.03401009738445282, "loss/reg": 0.026300577446818352, "step": 83 }, { "epoch": 0.0105, "grad_norm": 2.4155867099761963, "grad_norm_var": 0.5225404018326155, "learning_rate": 8.4e-05, "loss": 0.5432, "loss/crossentropy": 2.6546974182128906, "loss/hidden": 0.244140625, "loss/logits": 0.03612750768661499, "loss/reg": 0.02629682794213295, "step": 84 }, { "epoch": 0.010625, "grad_norm": 4.232295036315918, "grad_norm_var": 0.6129643025970267, "learning_rate": 8.5e-05, "loss": 0.7199, "loss/crossentropy": 2.2300524711608887, "loss/hidden": 0.412109375, "loss/logits": 0.044823646545410156, "loss/reg": 0.0262930728495121, "step": 85 }, { "epoch": 0.01075, "grad_norm": 2.7160282135009766, "grad_norm_var": 0.5620128324129702, "learning_rate": 8.6e-05, "loss": 0.6973, "loss/crossentropy": 2.2953288555145264, "loss/hidden": 0.37109375, "loss/logits": 0.06333646178245544, "loss/reg": 0.02628917805850506, "step": 86 }, { "epoch": 0.010875, "grad_norm": 3.0872819423675537, "grad_norm_var": 0.5495392294868544, "learning_rate": 8.7e-05, "loss": 0.636, "loss/crossentropy": 2.449223756790161, "loss/hidden": 0.328125, "loss/logits": 0.045011188834905624, "loss/reg": 0.02628495544195175, "step": 87 }, { "epoch": 0.011, "grad_norm": 2.6966607570648193, "grad_norm_var": 0.5621192378124493, "learning_rate": 8.800000000000001e-05, "loss": 0.6272, "loss/crossentropy": 2.5449047088623047, "loss/hidden": 0.31640625, "loss/logits": 0.04802623763680458, "loss/reg": 0.026280568912625313, "step": 88 }, { "epoch": 0.011125, "grad_norm": 2.9160921573638916, "grad_norm_var": 0.48685408890983506, "learning_rate": 8.900000000000001e-05, "loss": 0.6278, "loss/crossentropy": 2.1442108154296875, "loss/hidden": 0.310546875, "loss/logits": 0.054462507367134094, "loss/reg": 0.02627684734761715, "step": 89 }, { "epoch": 0.01125, "grad_norm": 3.3378944396972656, "grad_norm_var": 0.4283231906687052, "learning_rate": 9e-05, "loss": 0.6536, "loss/crossentropy": 2.4361062049865723, "loss/hidden": 0.349609375, "loss/logits": 0.04130454361438751, "loss/reg": 0.026273205876350403, "step": 90 }, { "epoch": 0.011375, "grad_norm": 2.597607374191284, "grad_norm_var": 0.42452911296148627, "learning_rate": 9.1e-05, "loss": 0.6391, "loss/crossentropy": 2.0079762935638428, "loss/hidden": 0.3203125, "loss/logits": 0.056107863783836365, "loss/reg": 0.02626909501850605, "step": 91 }, { "epoch": 0.0115, "grad_norm": 4.960971355438232, "grad_norm_var": 0.5826997127177641, "learning_rate": 9.200000000000001e-05, "loss": 0.6735, "loss/crossentropy": 2.708275079727173, "loss/hidden": 0.349609375, "loss/logits": 0.06119866296648979, "loss/reg": 0.026265164837241173, "step": 92 }, { "epoch": 0.011625, "grad_norm": 3.8193323612213135, "grad_norm_var": 0.5981799563928756, "learning_rate": 9.300000000000001e-05, "loss": 0.8429, "loss/crossentropy": 2.4117016792297363, "loss/hidden": 0.498046875, "loss/logits": 0.08221981674432755, "loss/reg": 0.026260720565915108, "step": 93 }, { "epoch": 0.01175, "grad_norm": 3.434213638305664, "grad_norm_var": 0.5157332557296352, "learning_rate": 9.4e-05, "loss": 0.6738, "loss/crossentropy": 2.62178111076355, "loss/hidden": 0.36328125, "loss/logits": 0.04794853553175926, "loss/reg": 0.02625615894794464, "step": 94 }, { "epoch": 0.011875, "grad_norm": 3.0944480895996094, "grad_norm_var": 0.4859309000509171, "learning_rate": 9.5e-05, "loss": 0.6829, "loss/crossentropy": 2.582462787628174, "loss/hidden": 0.353515625, "loss/logits": 0.0669020414352417, "loss/reg": 0.026251958683133125, "step": 95 }, { "epoch": 0.012, "grad_norm": 3.9548256397247314, "grad_norm_var": 0.5194300484800329, "learning_rate": 9.6e-05, "loss": 0.8508, "loss/crossentropy": 2.3243165016174316, "loss/hidden": 0.49609375, "loss/logits": 0.09221720695495605, "loss/reg": 0.026247689500451088, "step": 96 }, { "epoch": 0.012125, "grad_norm": 8.949115753173828, "grad_norm_var": 2.5460815450669125, "learning_rate": 9.7e-05, "loss": 0.9668, "loss/crossentropy": 2.3593697547912598, "loss/hidden": 0.62109375, "loss/logits": 0.08330727368593216, "loss/reg": 0.026243869215250015, "step": 97 }, { "epoch": 0.01225, "grad_norm": 3.874511957168579, "grad_norm_var": 2.411331023611861, "learning_rate": 9.8e-05, "loss": 0.7272, "loss/crossentropy": 1.9933784008026123, "loss/hidden": 0.41015625, "loss/logits": 0.05464401841163635, "loss/reg": 0.026239972561597824, "step": 98 }, { "epoch": 0.012375, "grad_norm": 5.088143825531006, "grad_norm_var": 2.507843574169987, "learning_rate": 9.900000000000001e-05, "loss": 0.6761, "loss/crossentropy": 2.578767776489258, "loss/hidden": 0.3515625, "loss/logits": 0.0621890164911747, "loss/reg": 0.026235179975628853, "step": 99 }, { "epoch": 0.0125, "grad_norm": 3.9010627269744873, "grad_norm_var": 2.366914585761602, "learning_rate": 0.0001, "loss": 0.7051, "loss/crossentropy": 2.4717133045196533, "loss/hidden": 0.375, "loss/logits": 0.0677795261144638, "loss/reg": 0.026230769231915474, "step": 100 }, { "epoch": 0.012625, "grad_norm": 5.50706148147583, "grad_norm_var": 2.522191588171327, "learning_rate": 0.0001, "loss": 0.7765, "loss/crossentropy": 2.3764612674713135, "loss/hidden": 0.44921875, "loss/logits": 0.06503438949584961, "loss/reg": 0.02622627653181553, "step": 101 }, { "epoch": 0.01275, "grad_norm": 5.103200435638428, "grad_norm_var": 2.470966679212188, "learning_rate": 0.0001, "loss": 0.7008, "loss/crossentropy": 2.5796542167663574, "loss/hidden": 0.38671875, "loss/logits": 0.05191829800605774, "loss/reg": 0.026221245527267456, "step": 102 }, { "epoch": 0.012875, "grad_norm": 18.05303192138672, "grad_norm_var": 14.358413039824521, "learning_rate": 0.0001, "loss": 1.0008, "loss/crossentropy": 1.927337646484375, "loss/hidden": 0.6796875, "loss/logits": 0.05896752327680588, "loss/reg": 0.02621658518910408, "step": 103 }, { "epoch": 0.013, "grad_norm": 3.410438299179077, "grad_norm_var": 14.16338361533652, "learning_rate": 0.0001, "loss": 0.735, "loss/crossentropy": 2.290928363800049, "loss/hidden": 0.40625, "loss/logits": 0.06666909158229828, "loss/reg": 0.026211561635136604, "step": 104 }, { "epoch": 0.013125, "grad_norm": 3.117622137069702, "grad_norm_var": 14.10656391346422, "learning_rate": 0.0001, "loss": 0.6665, "loss/crossentropy": 2.6549246311187744, "loss/hidden": 0.353515625, "loss/logits": 0.05095440149307251, "loss/reg": 0.026206739246845245, "step": 105 }, { "epoch": 0.01325, "grad_norm": 3.9999241828918457, "grad_norm_var": 13.97508509706009, "learning_rate": 0.0001, "loss": 0.8082, "loss/crossentropy": 2.460174798965454, "loss/hidden": 0.46484375, "loss/logits": 0.08136071264743805, "loss/reg": 0.02620157040655613, "step": 106 }, { "epoch": 0.013375, "grad_norm": 3.405712842941284, "grad_norm_var": 13.73775124044489, "learning_rate": 0.0001, "loss": 0.6518, "loss/crossentropy": 2.521803855895996, "loss/hidden": 0.3359375, "loss/logits": 0.053909383714199066, "loss/reg": 0.02619684301316738, "step": 107 }, { "epoch": 0.0135, "grad_norm": 3.615098237991333, "grad_norm_var": 13.899167673014775, "learning_rate": 0.0001, "loss": 0.7068, "loss/crossentropy": 2.510159969329834, "loss/hidden": 0.37890625, "loss/logits": 0.06601101160049438, "loss/reg": 0.02619197592139244, "step": 108 }, { "epoch": 0.013625, "grad_norm": 4.2520599365234375, "grad_norm_var": 13.834356012442765, "learning_rate": 0.0001, "loss": 0.735, "loss/crossentropy": 2.508683681488037, "loss/hidden": 0.41015625, "loss/logits": 0.0629870742559433, "loss/reg": 0.0261868704110384, "step": 109 }, { "epoch": 0.01375, "grad_norm": 3.215749979019165, "grad_norm_var": 13.887973421518442, "learning_rate": 0.0001, "loss": 0.8268, "loss/crossentropy": 2.3564252853393555, "loss/hidden": 0.48046875, "loss/logits": 0.08449074625968933, "loss/reg": 0.026181429624557495, "step": 110 }, { "epoch": 0.013875, "grad_norm": 4.598328590393066, "grad_norm_var": 13.615373346458785, "learning_rate": 0.0001, "loss": 0.7398, "loss/crossentropy": 2.366943359375, "loss/hidden": 0.41796875, "loss/logits": 0.060083672404289246, "loss/reg": 0.026176555082201958, "step": 111 }, { "epoch": 0.014, "grad_norm": 2.758070707321167, "grad_norm_var": 13.912012390229316, "learning_rate": 0.0001, "loss": 0.6839, "loss/crossentropy": 2.3351521492004395, "loss/hidden": 0.365234375, "loss/logits": 0.056987129151821136, "loss/reg": 0.0261719711124897, "step": 112 }, { "epoch": 0.014125, "grad_norm": 2.9389584064483643, "grad_norm_var": 13.147693721903025, "learning_rate": 0.0001, "loss": 0.8964, "loss/crossentropy": 2.188626766204834, "loss/hidden": 0.54296875, "loss/logits": 0.09171397984027863, "loss/reg": 0.026167072355747223, "step": 113 }, { "epoch": 0.01425, "grad_norm": 2.8545026779174805, "grad_norm_var": 13.338918720074266, "learning_rate": 0.0001, "loss": 0.6652, "loss/crossentropy": 2.488462448120117, "loss/hidden": 0.341796875, "loss/logits": 0.06177069991827011, "loss/reg": 0.026162149384617805, "step": 114 }, { "epoch": 0.014375, "grad_norm": 3.343590497970581, "grad_norm_var": 13.447848849906688, "learning_rate": 0.0001, "loss": 0.7317, "loss/crossentropy": 2.4826672077178955, "loss/hidden": 0.396484375, "loss/logits": 0.07369040697813034, "loss/reg": 0.026156950742006302, "step": 115 }, { "epoch": 0.0145, "grad_norm": 5.309541702270508, "grad_norm_var": 13.435010363164546, "learning_rate": 0.0001, "loss": 0.6918, "loss/crossentropy": 2.715517282485962, "loss/hidden": 0.376953125, "loss/logits": 0.05337735265493393, "loss/reg": 0.026151426136493683, "step": 116 }, { "epoch": 0.014625, "grad_norm": 3.413027763366699, "grad_norm_var": 13.488672790501717, "learning_rate": 0.0001, "loss": 0.7942, "loss/crossentropy": 2.3932089805603027, "loss/hidden": 0.451171875, "loss/logits": 0.08154396712779999, "loss/reg": 0.026146216318011284, "step": 117 }, { "epoch": 0.01475, "grad_norm": 2.735275983810425, "grad_norm_var": 13.676075950246783, "learning_rate": 0.0001, "loss": 0.7606, "loss/crossentropy": 2.2933082580566406, "loss/hidden": 0.4296875, "loss/logits": 0.06949938833713531, "loss/reg": 0.02614082768559456, "step": 118 }, { "epoch": 0.014875, "grad_norm": 3.1346964836120605, "grad_norm_var": 0.5056645529599865, "learning_rate": 0.0001, "loss": 0.7759, "loss/crossentropy": 2.331713914871216, "loss/hidden": 0.4375, "loss/logits": 0.07704727351665497, "loss/reg": 0.026135168969631195, "step": 119 }, { "epoch": 0.015, "grad_norm": 3.8077635765075684, "grad_norm_var": 0.5104468723684629, "learning_rate": 0.0001, "loss": 0.6882, "loss/crossentropy": 2.331220865249634, "loss/hidden": 0.373046875, "loss/logits": 0.05386776477098465, "loss/reg": 0.02612963318824768, "step": 120 }, { "epoch": 0.015125, "grad_norm": 3.209914445877075, "grad_norm_var": 0.5058893418769751, "learning_rate": 0.0001, "loss": 0.7536, "loss/crossentropy": 2.352771759033203, "loss/hidden": 0.4140625, "loss/logits": 0.07825946807861328, "loss/reg": 0.026124266907572746, "step": 121 }, { "epoch": 0.01525, "grad_norm": 3.3548500537872314, "grad_norm_var": 0.49208198737674874, "learning_rate": 0.0001, "loss": 0.7088, "loss/crossentropy": 2.483644723892212, "loss/hidden": 0.384765625, "loss/logits": 0.06287863850593567, "loss/reg": 0.0261182002723217, "step": 122 }, { "epoch": 0.015375, "grad_norm": 3.9953765869140625, "grad_norm_var": 0.5066601541023895, "learning_rate": 0.0001, "loss": 0.7792, "loss/crossentropy": 2.6117637157440186, "loss/hidden": 0.435546875, "loss/logits": 0.08250629901885986, "loss/reg": 0.026112213730812073, "step": 123 }, { "epoch": 0.0155, "grad_norm": 3.1783852577209473, "grad_norm_var": 0.5138316405800327, "learning_rate": 0.0001, "loss": 0.7559, "loss/crossentropy": 2.401679754257202, "loss/hidden": 0.423828125, "loss/logits": 0.07104581594467163, "loss/reg": 0.02610679157078266, "step": 124 }, { "epoch": 0.015625, "grad_norm": 3.2759885787963867, "grad_norm_var": 0.47631527116517774, "learning_rate": 0.0001, "loss": 0.8262, "loss/crossentropy": 2.3979361057281494, "loss/hidden": 0.486328125, "loss/logits": 0.0788530558347702, "loss/reg": 0.026101654395461082, "step": 125 }, { "epoch": 0.01575, "grad_norm": 4.0768632888793945, "grad_norm_var": 0.4963098069624029, "learning_rate": 0.0001, "loss": 0.6983, "loss/crossentropy": 2.5287113189697266, "loss/hidden": 0.375, "loss/logits": 0.062308911234140396, "loss/reg": 0.026096193119883537, "step": 126 }, { "epoch": 0.015875, "grad_norm": 4.300101280212402, "grad_norm_var": 0.45815803943682926, "learning_rate": 0.0001, "loss": 0.858, "loss/crossentropy": 2.255234956741333, "loss/hidden": 0.50390625, "loss/logits": 0.0932290330529213, "loss/reg": 0.026090849190950394, "step": 127 }, { "epoch": 0.016, "grad_norm": 3.303663492202759, "grad_norm_var": 0.42421384752866137, "learning_rate": 0.0001, "loss": 0.7284, "loss/crossentropy": 2.528862476348877, "loss/hidden": 0.40234375, "loss/logits": 0.06523742526769638, "loss/reg": 0.026085302233695984, "step": 128 }, { "epoch": 0.016125, "grad_norm": 6.868241310119629, "grad_norm_var": 1.0876227157335427, "learning_rate": 0.0001, "loss": 0.8999, "loss/crossentropy": 2.6554996967315674, "loss/hidden": 0.5390625, "loss/logits": 0.1000661626458168, "loss/reg": 0.02607985958456993, "step": 129 }, { "epoch": 0.01625, "grad_norm": 3.3035075664520264, "grad_norm_var": 1.046006684658701, "learning_rate": 0.0001, "loss": 0.7367, "loss/crossentropy": 2.293642282485962, "loss/hidden": 0.40625, "loss/logits": 0.06973426043987274, "loss/reg": 0.026074659079313278, "step": 130 }, { "epoch": 0.016375, "grad_norm": 4.2563276290893555, "grad_norm_var": 1.0439696727840135, "learning_rate": 0.0001, "loss": 0.7754, "loss/crossentropy": 2.2192564010620117, "loss/hidden": 0.447265625, "loss/logits": 0.06743350625038147, "loss/reg": 0.026069074869155884, "step": 131 }, { "epoch": 0.0165, "grad_norm": 4.646778583526611, "grad_norm_var": 0.9420233457711596, "learning_rate": 0.0001, "loss": 0.7496, "loss/crossentropy": 2.430368423461914, "loss/hidden": 0.4140625, "loss/logits": 0.07494455575942993, "loss/reg": 0.026063458994030952, "step": 132 }, { "epoch": 0.016625, "grad_norm": 7.465832233428955, "grad_norm_var": 1.7574380087301391, "learning_rate": 0.0001, "loss": 0.9607, "loss/crossentropy": 2.6182897090911865, "loss/hidden": 0.57421875, "loss/logits": 0.1258610337972641, "loss/reg": 0.026057813316583633, "step": 133 }, { "epoch": 0.01675, "grad_norm": 4.5479936599731445, "grad_norm_var": 1.6433309350178509, "learning_rate": 0.0001, "loss": 0.9529, "loss/crossentropy": 2.238551378250122, "loss/hidden": 0.58984375, "loss/logits": 0.10252824425697327, "loss/reg": 0.02605200558900833, "step": 134 }, { "epoch": 0.016875, "grad_norm": 3.4055774211883545, "grad_norm_var": 1.6105102483452751, "learning_rate": 0.0001, "loss": 0.8407, "loss/crossentropy": 2.4216599464416504, "loss/hidden": 0.498046875, "loss/logits": 0.08216647803783417, "loss/reg": 0.026046328246593475, "step": 135 }, { "epoch": 0.017, "grad_norm": 8.839641571044922, "grad_norm_var": 2.938344740358995, "learning_rate": 0.0001, "loss": 1.1023, "loss/crossentropy": 2.534442901611328, "loss/hidden": 0.7109375, "loss/logits": 0.1310025304555893, "loss/reg": 0.026040658354759216, "step": 136 }, { "epoch": 0.017125, "grad_norm": 4.6421589851379395, "grad_norm_var": 2.819843479450094, "learning_rate": 0.0001, "loss": 0.7418, "loss/crossentropy": 2.605559825897217, "loss/hidden": 0.416015625, "loss/logits": 0.06548085808753967, "loss/reg": 0.02603481523692608, "step": 137 }, { "epoch": 0.01725, "grad_norm": 3.1547701358795166, "grad_norm_var": 2.8553314644504555, "learning_rate": 0.0001, "loss": 0.731, "loss/crossentropy": 2.5905325412750244, "loss/hidden": 0.40234375, "loss/logits": 0.06835847347974777, "loss/reg": 0.02602926455438137, "step": 138 }, { "epoch": 0.017375, "grad_norm": 4.074351787567139, "grad_norm_var": 2.849577549707145, "learning_rate": 0.0001, "loss": 0.9582, "loss/crossentropy": 2.1483733654022217, "loss/hidden": 0.60546875, "loss/logits": 0.0924658477306366, "loss/reg": 0.026023706421256065, "step": 139 }, { "epoch": 0.0175, "grad_norm": 3.758636713027954, "grad_norm_var": 2.7618912420839155, "learning_rate": 0.0001, "loss": 0.9996, "loss/crossentropy": 2.335742473602295, "loss/hidden": 0.6328125, "loss/logits": 0.10663188993930817, "loss/reg": 0.026017924770712852, "step": 140 }, { "epoch": 0.017625, "grad_norm": 4.186927795410156, "grad_norm_var": 2.6505093919275544, "learning_rate": 0.0001, "loss": 0.8414, "loss/crossentropy": 2.281843423843384, "loss/hidden": 0.498046875, "loss/logits": 0.08321215212345123, "loss/reg": 0.026011699810624123, "step": 141 }, { "epoch": 0.01775, "grad_norm": 3.1666276454925537, "grad_norm_var": 2.775123140671519, "learning_rate": 0.0001, "loss": 0.7768, "loss/crossentropy": 2.4267935752868652, "loss/hidden": 0.44140625, "loss/logits": 0.07538889348506927, "loss/reg": 0.02600528486073017, "step": 142 }, { "epoch": 0.017875, "grad_norm": 6.386529445648193, "grad_norm_var": 2.9581845034072245, "learning_rate": 0.0001, "loss": 0.8401, "loss/crossentropy": 2.7848174571990967, "loss/hidden": 0.50390625, "loss/logits": 0.07621172070503235, "loss/reg": 0.025999369099736214, "step": 143 }, { "epoch": 0.018, "grad_norm": 3.8083512783050537, "grad_norm_var": 2.876745593692829, "learning_rate": 0.0001, "loss": 0.9046, "loss/crossentropy": 2.341048002243042, "loss/hidden": 0.56640625, "loss/logits": 0.0782276839017868, "loss/reg": 0.0259928647428751, "step": 144 }, { "epoch": 0.018125, "grad_norm": 4.083465576171875, "grad_norm_var": 2.5868089188751657, "learning_rate": 0.0001, "loss": 0.9449, "loss/crossentropy": 2.3942995071411133, "loss/hidden": 0.59375, "loss/logits": 0.09130540490150452, "loss/reg": 0.025986921042203903, "step": 145 }, { "epoch": 0.01825, "grad_norm": 3.654815673828125, "grad_norm_var": 2.533420197907486, "learning_rate": 0.0001, "loss": 0.9459, "loss/crossentropy": 2.2414467334747314, "loss/hidden": 0.5859375, "loss/logits": 0.10016702860593796, "loss/reg": 0.02598092146217823, "step": 146 }, { "epoch": 0.018375, "grad_norm": 5.4976935386657715, "grad_norm_var": 2.567896035243579, "learning_rate": 0.0001, "loss": 0.9079, "loss/crossentropy": 2.1185100078582764, "loss/hidden": 0.58203125, "loss/logits": 0.06617112457752228, "loss/reg": 0.025974513962864876, "step": 147 }, { "epoch": 0.0185, "grad_norm": 3.4107933044433594, "grad_norm_var": 2.673383097164577, "learning_rate": 0.0001, "loss": 0.7982, "loss/crossentropy": 2.417313575744629, "loss/hidden": 0.462890625, "loss/logits": 0.07567355036735535, "loss/reg": 0.025968506932258606, "step": 148 }, { "epoch": 0.018625, "grad_norm": 3.589749574661255, "grad_norm_var": 2.1469293827286418, "learning_rate": 0.0001, "loss": 0.8088, "loss/crossentropy": 2.2554194927215576, "loss/hidden": 0.478515625, "loss/logits": 0.07071521133184433, "loss/reg": 0.025961775332689285, "step": 149 }, { "epoch": 0.01875, "grad_norm": 4.003805160522461, "grad_norm_var": 2.1538296896943887, "learning_rate": 0.0001, "loss": 0.7856, "loss/crossentropy": 2.7128918170928955, "loss/hidden": 0.453125, "loss/logits": 0.07290500402450562, "loss/reg": 0.025955306366086006, "step": 150 }, { "epoch": 0.018875, "grad_norm": 3.536449432373047, "grad_norm_var": 2.1383506752067736, "learning_rate": 0.0001, "loss": 0.7436, "loss/crossentropy": 2.509995460510254, "loss/hidden": 0.421875, "loss/logits": 0.062191903591156006, "loss/reg": 0.02594931609928608, "step": 151 }, { "epoch": 0.019, "grad_norm": 3.434654951095581, "grad_norm_var": 0.7374638182579057, "learning_rate": 0.0001, "loss": 0.9118, "loss/crossentropy": 2.3447437286376953, "loss/hidden": 0.55859375, "loss/logits": 0.0937500149011612, "loss/reg": 0.025943227112293243, "step": 152 }, { "epoch": 0.019125, "grad_norm": 8.066261291503906, "grad_norm_var": 1.7522972641868202, "learning_rate": 0.0001, "loss": 0.8648, "loss/crossentropy": 2.4481232166290283, "loss/hidden": 0.52734375, "loss/logits": 0.07811163365840912, "loss/reg": 0.025936946272850037, "step": 153 }, { "epoch": 0.01925, "grad_norm": 3.2214348316192627, "grad_norm_var": 1.7429433318934864, "learning_rate": 0.0001, "loss": 0.7673, "loss/crossentropy": 2.3142549991607666, "loss/hidden": 0.43359375, "loss/logits": 0.07435894012451172, "loss/reg": 0.02593095973134041, "step": 154 }, { "epoch": 0.019375, "grad_norm": 2.854038715362549, "grad_norm_var": 1.8633807825236384, "learning_rate": 0.0001, "loss": 0.7601, "loss/crossentropy": 2.7022647857666016, "loss/hidden": 0.43359375, "loss/logits": 0.06724615395069122, "loss/reg": 0.025924943387508392, "step": 155 }, { "epoch": 0.0195, "grad_norm": 3.1763484477996826, "grad_norm_var": 1.9162196068123232, "learning_rate": 0.0001, "loss": 0.9, "loss/crossentropy": 2.574676036834717, "loss/hidden": 0.54296875, "loss/logits": 0.09785018861293793, "loss/reg": 0.025919148698449135, "step": 156 }, { "epoch": 0.019625, "grad_norm": 3.999523162841797, "grad_norm_var": 1.9169889601133074, "learning_rate": 0.0001, "loss": 0.858, "loss/crossentropy": 2.480532646179199, "loss/hidden": 0.515625, "loss/logits": 0.08323468267917633, "loss/reg": 0.025913061574101448, "step": 157 }, { "epoch": 0.01975, "grad_norm": 2.9662230014801025, "grad_norm_var": 1.944924590139983, "learning_rate": 0.0001, "loss": 0.818, "loss/crossentropy": 2.459967613220215, "loss/hidden": 0.48046875, "loss/logits": 0.07847169041633606, "loss/reg": 0.025906959548592567, "step": 158 }, { "epoch": 0.019875, "grad_norm": 7.074191093444824, "grad_norm_var": 2.1836107796529065, "learning_rate": 0.0001, "loss": 0.9801, "loss/crossentropy": 2.2858352661132812, "loss/hidden": 0.6484375, "loss/logits": 0.07268328964710236, "loss/reg": 0.025901462882757187, "step": 159 }, { "epoch": 0.02, "grad_norm": 3.6333370208740234, "grad_norm_var": 2.193465227977892, "learning_rate": 0.0001, "loss": 0.9048, "loss/crossentropy": 2.4626388549804688, "loss/hidden": 0.5546875, "loss/logits": 0.09115847945213318, "loss/reg": 0.02589540183544159, "step": 160 }, { "epoch": 0.020125, "grad_norm": 4.283749103546143, "grad_norm_var": 2.1945247126451437, "learning_rate": 0.0001, "loss": 0.9251, "loss/crossentropy": 2.5746121406555176, "loss/hidden": 0.58203125, "loss/logits": 0.0841851755976677, "loss/reg": 0.025888830423355103, "step": 161 }, { "epoch": 0.02025, "grad_norm": 3.628138542175293, "grad_norm_var": 2.196331220420876, "learning_rate": 0.0001, "loss": 0.9408, "loss/crossentropy": 2.398010730743408, "loss/hidden": 0.58203125, "loss/logits": 0.09991887211799622, "loss/reg": 0.025882074609398842, "step": 162 }, { "epoch": 0.020375, "grad_norm": 3.5412559509277344, "grad_norm_var": 2.0836172065033507, "learning_rate": 0.0001, "loss": 0.857, "loss/crossentropy": 2.5064220428466797, "loss/hidden": 0.50390625, "loss/logits": 0.0943569540977478, "loss/reg": 0.025876009836792946, "step": 163 }, { "epoch": 0.0205, "grad_norm": 3.498668670654297, "grad_norm_var": 2.0768887394910185, "learning_rate": 0.0001, "loss": 0.8917, "loss/crossentropy": 2.535529375076294, "loss/hidden": 0.52734375, "loss/logits": 0.10565976053476334, "loss/reg": 0.02587023191154003, "step": 164 }, { "epoch": 0.020625, "grad_norm": 3.14025616645813, "grad_norm_var": 2.116006039378421, "learning_rate": 0.0001, "loss": 0.7816, "loss/crossentropy": 2.7250640392303467, "loss/hidden": 0.453125, "loss/logits": 0.06980661302804947, "loss/reg": 0.025863803923130035, "step": 165 }, { "epoch": 0.02075, "grad_norm": 2.6821706295013428, "grad_norm_var": 2.2251478520018773, "learning_rate": 0.0001, "loss": 0.7558, "loss/crossentropy": 2.230567693710327, "loss/hidden": 0.43359375, "loss/logits": 0.06358660757541656, "loss/reg": 0.02585742622613907, "step": 166 }, { "epoch": 0.020875, "grad_norm": 3.8048856258392334, "grad_norm_var": 2.2158862694911607, "learning_rate": 0.0001, "loss": 0.7965, "loss/crossentropy": 2.69014310836792, "loss/hidden": 0.453125, "loss/logits": 0.08488957583904266, "loss/reg": 0.025850988924503326, "step": 167 }, { "epoch": 0.021, "grad_norm": 4.305826187133789, "grad_norm_var": 2.2048741298976515, "learning_rate": 0.0001, "loss": 1.1009, "loss/crossentropy": 2.2925851345062256, "loss/hidden": 0.73828125, "loss/logits": 0.10412566363811493, "loss/reg": 0.025845136493444443, "step": 168 }, { "epoch": 0.021125, "grad_norm": 4.051706790924072, "grad_norm_var": 1.0314628897999674, "learning_rate": 0.0001, "loss": 0.7756, "loss/crossentropy": 2.4515578746795654, "loss/hidden": 0.4453125, "loss/logits": 0.07186460494995117, "loss/reg": 0.02583896555006504, "step": 169 }, { "epoch": 0.02125, "grad_norm": 3.3800575733184814, "grad_norm_var": 1.022039210437893, "learning_rate": 0.0001, "loss": 0.798, "loss/crossentropy": 2.3387436866760254, "loss/hidden": 0.455078125, "loss/logits": 0.08458675444126129, "loss/reg": 0.025832952931523323, "step": 170 }, { "epoch": 0.021375, "grad_norm": 3.069735527038574, "grad_norm_var": 0.9991429378891439, "learning_rate": 0.0001, "loss": 0.7416, "loss/crossentropy": 2.660727024078369, "loss/hidden": 0.412109375, "loss/logits": 0.07125158607959747, "loss/reg": 0.025827286764979362, "step": 171 }, { "epoch": 0.0215, "grad_norm": 3.5827748775482178, "grad_norm_var": 0.9775809993657352, "learning_rate": 0.0001, "loss": 0.9301, "loss/crossentropy": 2.160230875015259, "loss/hidden": 0.5703125, "loss/logits": 0.10158500075340271, "loss/reg": 0.025821613147854805, "step": 172 }, { "epoch": 0.021625, "grad_norm": 3.0348143577575684, "grad_norm_var": 1.008817027257092, "learning_rate": 0.0001, "loss": 0.9059, "loss/crossentropy": 2.5519795417785645, "loss/hidden": 0.5390625, "loss/logits": 0.10867513716220856, "loss/reg": 0.025815250352025032, "step": 173 }, { "epoch": 0.02175, "grad_norm": 3.325514316558838, "grad_norm_var": 0.980302655794393, "learning_rate": 0.0001, "loss": 0.7651, "loss/crossentropy": 2.3060855865478516, "loss/hidden": 0.43359375, "loss/logits": 0.0733788013458252, "loss/reg": 0.02580902725458145, "step": 174 }, { "epoch": 0.021875, "grad_norm": 2.9402523040771484, "grad_norm_var": 0.21740374576502078, "learning_rate": 0.0001, "loss": 0.859, "loss/crossentropy": 2.4109151363372803, "loss/hidden": 0.515625, "loss/logits": 0.0853327289223671, "loss/reg": 0.025802936404943466, "step": 175 }, { "epoch": 0.022, "grad_norm": 4.970669746398926, "grad_norm_var": 0.354037293260262, "learning_rate": 0.0001, "loss": 0.9835, "loss/crossentropy": 2.9366648197174072, "loss/hidden": 0.6171875, "loss/logits": 0.10840301960706711, "loss/reg": 0.02579565905034542, "step": 176 }, { "epoch": 0.022125, "grad_norm": 4.438536643981934, "grad_norm_var": 0.37010993593279384, "learning_rate": 0.0001, "loss": 0.8574, "loss/crossentropy": 2.4177615642547607, "loss/hidden": 0.51171875, "loss/logits": 0.08775197714567184, "loss/reg": 0.025788920000195503, "step": 177 }, { "epoch": 0.02225, "grad_norm": 4.3905863761901855, "grad_norm_var": 0.41060424896311526, "learning_rate": 0.0001, "loss": 0.8796, "loss/crossentropy": 2.320418119430542, "loss/hidden": 0.53125, "loss/logits": 0.09051363915205002, "loss/reg": 0.02578234300017357, "step": 178 }, { "epoch": 0.022375, "grad_norm": 3.2044453620910645, "grad_norm_var": 0.42189777730298467, "learning_rate": 0.0001, "loss": 0.785, "loss/crossentropy": 2.327108383178711, "loss/hidden": 0.447265625, "loss/logits": 0.08003242313861847, "loss/reg": 0.025775177404284477, "step": 179 }, { "epoch": 0.0225, "grad_norm": 3.2016260623931885, "grad_norm_var": 0.4319725268587102, "learning_rate": 0.0001, "loss": 0.8076, "loss/crossentropy": 2.6913022994995117, "loss/hidden": 0.466796875, "loss/logits": 0.08316424489021301, "loss/reg": 0.02576799876987934, "step": 180 }, { "epoch": 0.022625, "grad_norm": 3.216141939163208, "grad_norm_var": 0.42772885748243633, "learning_rate": 0.0001, "loss": 0.7782, "loss/crossentropy": 2.4444668292999268, "loss/hidden": 0.447265625, "loss/logits": 0.07332297414541245, "loss/reg": 0.025760415941476822, "step": 181 }, { "epoch": 0.02275, "grad_norm": 3.6005637645721436, "grad_norm_var": 0.3680557604441513, "learning_rate": 0.0001, "loss": 0.8578, "loss/crossentropy": 2.2084226608276367, "loss/hidden": 0.51171875, "loss/logits": 0.0885147675871849, "loss/reg": 0.025753989815711975, "step": 182 }, { "epoch": 0.022875, "grad_norm": 4.19577693939209, "grad_norm_var": 0.3852931468556484, "learning_rate": 0.0001, "loss": 0.9394, "loss/crossentropy": 2.557783842086792, "loss/hidden": 0.56640625, "loss/logits": 0.11552520841360092, "loss/reg": 0.025747526437044144, "step": 183 }, { "epoch": 0.023, "grad_norm": 3.024552822113037, "grad_norm_var": 0.38129301153876227, "learning_rate": 0.0001, "loss": 0.8858, "loss/crossentropy": 2.994615316390991, "loss/hidden": 0.52734375, "loss/logits": 0.10108112543821335, "loss/reg": 0.02574075385928154, "step": 184 }, { "epoch": 0.023125, "grad_norm": 3.3255319595336914, "grad_norm_var": 0.3706833429951111, "learning_rate": 0.0001, "loss": 0.8407, "loss/crossentropy": 2.677245855331421, "loss/hidden": 0.4765625, "loss/logits": 0.10678394883871078, "loss/reg": 0.02573317475616932, "step": 185 }, { "epoch": 0.02325, "grad_norm": 3.341599464416504, "grad_norm_var": 0.3716797590150757, "learning_rate": 0.0001, "loss": 0.9127, "loss/crossentropy": 2.156444549560547, "loss/hidden": 0.56640625, "loss/logits": 0.0890708938241005, "loss/reg": 0.02572541870176792, "step": 186 }, { "epoch": 0.023375, "grad_norm": 2.925915241241455, "grad_norm_var": 0.3822577484351124, "learning_rate": 0.0001, "loss": 0.8815, "loss/crossentropy": 2.2716755867004395, "loss/hidden": 0.53125, "loss/logits": 0.09304732084274292, "loss/reg": 0.02571748197078705, "step": 187 }, { "epoch": 0.0235, "grad_norm": 4.192226886749268, "grad_norm_var": 0.40854537365233884, "learning_rate": 0.0001, "loss": 0.8547, "loss/crossentropy": 2.378901720046997, "loss/hidden": 0.5078125, "loss/logits": 0.08977752178907394, "loss/reg": 0.025709524750709534, "step": 188 }, { "epoch": 0.023625, "grad_norm": 5.648179054260254, "grad_norm_var": 0.6443691048121629, "learning_rate": 0.0001, "loss": 1.0032, "loss/crossentropy": 2.4307687282562256, "loss/hidden": 0.63671875, "loss/logits": 0.10948194563388824, "loss/reg": 0.025701580569148064, "step": 189 }, { "epoch": 0.02375, "grad_norm": 6.345841884613037, "grad_norm_var": 1.045029826307897, "learning_rate": 0.0001, "loss": 1.0995, "loss/crossentropy": 2.279203414916992, "loss/hidden": 0.74609375, "loss/logits": 0.09642117470502853, "loss/reg": 0.025694590061903, "step": 190 }, { "epoch": 0.023875, "grad_norm": 4.242865085601807, "grad_norm_var": 0.9782837984014707, "learning_rate": 0.0001, "loss": 1.155, "loss/crossentropy": 2.235325574874878, "loss/hidden": 0.7890625, "loss/logits": 0.10906486213207245, "loss/reg": 0.025686509907245636, "step": 191 }, { "epoch": 0.024, "grad_norm": 6.347895622253418, "grad_norm_var": 1.272032888242258, "learning_rate": 0.0001, "loss": 1.0789, "loss/crossentropy": 2.5386898517608643, "loss/hidden": 0.67578125, "loss/logits": 0.1463102549314499, "loss/reg": 0.025679145008325577, "step": 192 }, { "epoch": 0.024125, "grad_norm": 3.077846050262451, "grad_norm_var": 1.3268106432816444, "learning_rate": 0.0001, "loss": 0.855, "loss/crossentropy": 2.5889694690704346, "loss/hidden": 0.515625, "loss/logits": 0.08266487717628479, "loss/reg": 0.025671878829598427, "step": 193 }, { "epoch": 0.02425, "grad_norm": 3.672849416732788, "grad_norm_var": 1.323313109234428, "learning_rate": 0.0001, "loss": 0.8213, "loss/crossentropy": 2.269009590148926, "loss/hidden": 0.4921875, "loss/logits": 0.07244250178337097, "loss/reg": 0.02566472254693508, "step": 194 }, { "epoch": 0.024375, "grad_norm": 3.13712215423584, "grad_norm_var": 1.330492936258482, "learning_rate": 0.0001, "loss": 0.9044, "loss/crossentropy": 2.3420536518096924, "loss/hidden": 0.5390625, "loss/logits": 0.10874692350625992, "loss/reg": 0.02565707452595234, "step": 195 }, { "epoch": 0.0245, "grad_norm": 5.941372871398926, "grad_norm_var": 1.5194802994125645, "learning_rate": 0.0001, "loss": 1.0268, "loss/crossentropy": 2.245668649673462, "loss/hidden": 0.63671875, "loss/logits": 0.13362175226211548, "loss/reg": 0.02564912661910057, "step": 196 }, { "epoch": 0.024625, "grad_norm": 2.8778631687164307, "grad_norm_var": 1.5682913914576866, "learning_rate": 0.0001, "loss": 0.933, "loss/crossentropy": 2.4744086265563965, "loss/hidden": 0.57421875, "loss/logits": 0.10236240178346634, "loss/reg": 0.025642510503530502, "step": 197 }, { "epoch": 0.02475, "grad_norm": 5.235295295715332, "grad_norm_var": 1.6223942527523605, "learning_rate": 0.0001, "loss": 0.993, "loss/crossentropy": 2.3120462894439697, "loss/hidden": 0.61328125, "loss/logits": 0.12334179133176804, "loss/reg": 0.025634463876485825, "step": 198 }, { "epoch": 0.024875, "grad_norm": 3.2772397994995117, "grad_norm_var": 1.6781902664948347, "learning_rate": 0.0001, "loss": 0.9134, "loss/crossentropy": 2.2896904945373535, "loss/hidden": 0.5625, "loss/logits": 0.0946369469165802, "loss/reg": 0.025627706199884415, "step": 199 }, { "epoch": 0.025, "grad_norm": 4.021130084991455, "grad_norm_var": 1.5889382838258257, "learning_rate": 0.0001, "loss": 0.9467, "loss/crossentropy": 2.3281702995300293, "loss/hidden": 0.5859375, "loss/logits": 0.10459813475608826, "loss/reg": 0.025620225816965103, "step": 200 }, { "epoch": 0.025125, "grad_norm": 3.3705508708953857, "grad_norm_var": 1.5836618344967157, "learning_rate": 0.0001, "loss": 1.0247, "loss/crossentropy": 2.3704278469085693, "loss/hidden": 0.640625, "loss/logits": 0.12795141339302063, "loss/reg": 0.025613589212298393, "step": 201 }, { "epoch": 0.02525, "grad_norm": 5.586423397064209, "grad_norm_var": 1.6331597901730046, "learning_rate": 0.0001, "loss": 1.2703, "loss/crossentropy": 2.3346736431121826, "loss/hidden": 0.83984375, "loss/logits": 0.17438159883022308, "loss/reg": 0.025606893002986908, "step": 202 }, { "epoch": 0.025375, "grad_norm": 6.558523654937744, "grad_norm_var": 1.7590475344041898, "learning_rate": 0.0001, "loss": 1.0711, "loss/crossentropy": 2.264883518218994, "loss/hidden": 0.71484375, "loss/logits": 0.10028564184904099, "loss/reg": 0.025599893182516098, "step": 203 }, { "epoch": 0.0255, "grad_norm": 3.024080991744995, "grad_norm_var": 1.9071946132324447, "learning_rate": 0.0001, "loss": 0.9349, "loss/crossentropy": 2.505457878112793, "loss/hidden": 0.58203125, "loss/logits": 0.09690214693546295, "loss/reg": 0.025592036545276642, "step": 204 }, { "epoch": 0.025625, "grad_norm": 3.268216133117676, "grad_norm_var": 1.9040994009139534, "learning_rate": 0.0001, "loss": 0.8433, "loss/crossentropy": 2.629786968231201, "loss/hidden": 0.5, "loss/logits": 0.08743932843208313, "loss/reg": 0.025583887472748756, "step": 205 }, { "epoch": 0.02575, "grad_norm": 5.203437328338623, "grad_norm_var": 1.6853258867355625, "learning_rate": 0.0001, "loss": 0.9278, "loss/crossentropy": 2.368603229522705, "loss/hidden": 0.5625, "loss/logits": 0.10955986380577087, "loss/reg": 0.025574835017323494, "step": 206 }, { "epoch": 0.025875, "grad_norm": 5.106112480163574, "grad_norm_var": 1.725017173963382, "learning_rate": 0.0001, "loss": 0.9881, "loss/crossentropy": 2.554746389389038, "loss/hidden": 0.6171875, "loss/logits": 0.11524944007396698, "loss/reg": 0.02556804195046425, "step": 207 }, { "epoch": 0.026, "grad_norm": 8.258187294006348, "grad_norm_var": 2.4602814049521387, "learning_rate": 0.0001, "loss": 1.3675, "loss/crossentropy": 2.4391205310821533, "loss/hidden": 0.94140625, "loss/logits": 0.17047560214996338, "loss/reg": 0.025560656562447548, "step": 208 }, { "epoch": 0.026125, "grad_norm": 3.8223764896392822, "grad_norm_var": 2.3561294395388566, "learning_rate": 0.0001, "loss": 1.02, "loss/crossentropy": 2.5300426483154297, "loss/hidden": 0.62890625, "loss/logits": 0.13556598126888275, "loss/reg": 0.025551345199346542, "step": 209 }, { "epoch": 0.02625, "grad_norm": 3.6237666606903076, "grad_norm_var": 2.36184075461094, "learning_rate": 0.0001, "loss": 1.0218, "loss/crossentropy": 2.3368213176727295, "loss/hidden": 0.6640625, "loss/logits": 0.10230866074562073, "loss/reg": 0.025544527918100357, "step": 210 }, { "epoch": 0.026375, "grad_norm": 3.1394050121307373, "grad_norm_var": 2.3614203164344407, "learning_rate": 0.0001, "loss": 0.964, "loss/crossentropy": 2.338050603866577, "loss/hidden": 0.609375, "loss/logits": 0.09931059181690216, "loss/reg": 0.025535617023706436, "step": 211 }, { "epoch": 0.0265, "grad_norm": 3.1498186588287354, "grad_norm_var": 2.319283484830568, "learning_rate": 0.0001, "loss": 0.8039, "loss/crossentropy": 2.5108940601348877, "loss/hidden": 0.46875, "loss/logits": 0.07982419431209564, "loss/reg": 0.025528721511363983, "step": 212 }, { "epoch": 0.026625, "grad_norm": 3.334510326385498, "grad_norm_var": 2.2429786468962374, "learning_rate": 0.0001, "loss": 1.1105, "loss/crossentropy": 2.458519697189331, "loss/hidden": 0.71875, "loss/logits": 0.1365831047296524, "loss/reg": 0.02552017569541931, "step": 213 }, { "epoch": 0.02675, "grad_norm": 3.3243789672851562, "grad_norm_var": 2.2516768547287977, "learning_rate": 0.0001, "loss": 1.0016, "loss/crossentropy": 2.2530109882354736, "loss/hidden": 0.62890625, "loss/logits": 0.11759582161903381, "loss/reg": 0.025511734187602997, "step": 214 }, { "epoch": 0.026875, "grad_norm": 3.3768937587738037, "grad_norm_var": 2.2393156807375245, "learning_rate": 0.0001, "loss": 1.0452, "loss/crossentropy": 2.3267643451690674, "loss/hidden": 0.65625, "loss/logits": 0.13390058279037476, "loss/reg": 0.0255054272711277, "step": 215 }, { "epoch": 0.027, "grad_norm": 7.391561031341553, "grad_norm_var": 2.841738119885866, "learning_rate": 0.0001, "loss": 1.2263, "loss/crossentropy": 2.285508394241333, "loss/hidden": 0.8203125, "loss/logits": 0.1510239541530609, "loss/reg": 0.025499247014522552, "step": 216 }, { "epoch": 0.027125, "grad_norm": 3.143969774246216, "grad_norm_var": 2.8781965049841705, "learning_rate": 0.0001, "loss": 0.9421, "loss/crossentropy": 2.6111316680908203, "loss/hidden": 0.5703125, "loss/logits": 0.11691074818372726, "loss/reg": 0.025491848587989807, "step": 217 }, { "epoch": 0.02725, "grad_norm": 4.989267826080322, "grad_norm_var": 2.8105564664802234, "learning_rate": 0.0001, "loss": 0.9104, "loss/crossentropy": 2.7856109142303467, "loss/hidden": 0.546875, "loss/logits": 0.10870229452848434, "loss/reg": 0.025485411286354065, "step": 218 }, { "epoch": 0.027375, "grad_norm": 8.867380142211914, "grad_norm_var": 3.802177537112387, "learning_rate": 0.0001, "loss": 1.2976, "loss/crossentropy": 2.414778470993042, "loss/hidden": 0.84375, "loss/logits": 0.199102982878685, "loss/reg": 0.025478005409240723, "step": 219 }, { "epoch": 0.0275, "grad_norm": 3.949193239212036, "grad_norm_var": 3.665725599495321, "learning_rate": 0.0001, "loss": 1.0781, "loss/crossentropy": 2.3898277282714844, "loss/hidden": 0.6796875, "loss/logits": 0.14368270337581635, "loss/reg": 0.02547168917953968, "step": 220 }, { "epoch": 0.027625, "grad_norm": 7.980153560638428, "grad_norm_var": 4.202985170085262, "learning_rate": 0.0001, "loss": 1.4005, "loss/crossentropy": 2.0598959922790527, "loss/hidden": 0.9921875, "loss/logits": 0.15361803770065308, "loss/reg": 0.02546495571732521, "step": 221 }, { "epoch": 0.02775, "grad_norm": 4.118736743927002, "grad_norm_var": 4.234989890674561, "learning_rate": 0.0001, "loss": 0.9084, "loss/crossentropy": 2.2568750381469727, "loss/hidden": 0.55859375, "loss/logits": 0.09525588899850845, "loss/reg": 0.02545757219195366, "step": 222 }, { "epoch": 0.027875, "grad_norm": 3.730299711227417, "grad_norm_var": 4.306033514824207, "learning_rate": 0.0001, "loss": 0.9394, "loss/crossentropy": 2.7180914878845215, "loss/hidden": 0.57421875, "loss/logits": 0.11064038425683975, "loss/reg": 0.025450890883803368, "step": 223 }, { "epoch": 0.028, "grad_norm": 3.138925552368164, "grad_norm_var": 3.557911666554559, "learning_rate": 0.0001, "loss": 0.8406, "loss/crossentropy": 2.3719072341918945, "loss/hidden": 0.5078125, "loss/logits": 0.07834647595882416, "loss/reg": 0.02544352412223816, "step": 224 }, { "epoch": 0.028125, "grad_norm": 3.478994846343994, "grad_norm_var": 3.593674795871404, "learning_rate": 0.0001, "loss": 0.9072, "loss/crossentropy": 2.4101219177246094, "loss/hidden": 0.56640625, "loss/logits": 0.0863800048828125, "loss/reg": 0.025437019765377045, "step": 225 }, { "epoch": 0.02825, "grad_norm": 5.091615676879883, "grad_norm_var": 3.572291640880083, "learning_rate": 0.0001, "loss": 0.9864, "loss/crossentropy": 2.2258856296539307, "loss/hidden": 0.640625, "loss/logits": 0.09143185615539551, "loss/reg": 0.025430168956518173, "step": 226 }, { "epoch": 0.028375, "grad_norm": 3.616190195083618, "grad_norm_var": 3.4991896025782796, "learning_rate": 0.0001, "loss": 0.9246, "loss/crossentropy": 2.765105724334717, "loss/hidden": 0.5703125, "loss/logits": 0.10003923624753952, "loss/reg": 0.025423482060432434, "step": 227 }, { "epoch": 0.0285, "grad_norm": 4.406581878662109, "grad_norm_var": 3.364516245493509, "learning_rate": 0.0001, "loss": 0.98, "loss/crossentropy": 2.41001033782959, "loss/hidden": 0.6171875, "loss/logits": 0.10860306769609451, "loss/reg": 0.025416266173124313, "step": 228 }, { "epoch": 0.028625, "grad_norm": 3.003995418548584, "grad_norm_var": 3.4280449285692023, "learning_rate": 0.0001, "loss": 0.9759, "loss/crossentropy": 2.518749952316284, "loss/hidden": 0.62109375, "loss/logits": 0.10075733810663223, "loss/reg": 0.025409165769815445, "step": 229 }, { "epoch": 0.02875, "grad_norm": 4.073727130889893, "grad_norm_var": 3.3356380380428576, "learning_rate": 0.0001, "loss": 0.8881, "loss/crossentropy": 2.6824846267700195, "loss/hidden": 0.5390625, "loss/logits": 0.09498724341392517, "loss/reg": 0.025402268394827843, "step": 230 }, { "epoch": 0.028875, "grad_norm": 3.8958635330200195, "grad_norm_var": 3.2645611787952435, "learning_rate": 0.0001, "loss": 0.8794, "loss/crossentropy": 2.684971332550049, "loss/hidden": 0.53515625, "loss/logits": 0.09024453163146973, "loss/reg": 0.025395380333065987, "step": 231 }, { "epoch": 0.029, "grad_norm": 3.5619406700134277, "grad_norm_var": 2.796506014438906, "learning_rate": 0.0001, "loss": 0.957, "loss/crossentropy": 2.7883284091949463, "loss/hidden": 0.6015625, "loss/logits": 0.10159540176391602, "loss/reg": 0.02538810484111309, "step": 232 }, { "epoch": 0.029125, "grad_norm": 12.658771514892578, "grad_norm_var": 6.809983669727001, "learning_rate": 0.0001, "loss": 1.1843, "loss/crossentropy": 2.537827253341675, "loss/hidden": 0.7890625, "loss/logits": 0.14144758880138397, "loss/reg": 0.02538110502064228, "step": 233 }, { "epoch": 0.02925, "grad_norm": 8.465475082397461, "grad_norm_var": 7.543990683504188, "learning_rate": 0.0001, "loss": 1.2958, "loss/crossentropy": 2.6715972423553467, "loss/hidden": 0.875, "loss/logits": 0.16701380908489227, "loss/reg": 0.02537420578300953, "step": 234 }, { "epoch": 0.029375, "grad_norm": 6.49767541885376, "grad_norm_var": 6.75275709893707, "learning_rate": 0.0001, "loss": 1.0826, "loss/crossentropy": 2.316210985183716, "loss/hidden": 0.7109375, "loss/logits": 0.11794352531433105, "loss/reg": 0.025367144495248795, "step": 235 }, { "epoch": 0.0295, "grad_norm": 4.668674945831299, "grad_norm_var": 6.674304000957216, "learning_rate": 0.0001, "loss": 1.0642, "loss/crossentropy": 1.956210732460022, "loss/hidden": 0.69921875, "loss/logits": 0.11135473847389221, "loss/reg": 0.02535996399819851, "step": 236 }, { "epoch": 0.029625, "grad_norm": 6.132915019989014, "grad_norm_var": 6.19031909782113, "learning_rate": 0.0001, "loss": 1.0462, "loss/crossentropy": 2.6552460193634033, "loss/hidden": 0.6640625, "loss/logits": 0.12862679362297058, "loss/reg": 0.025352442637085915, "step": 237 }, { "epoch": 0.02975, "grad_norm": 4.94125509262085, "grad_norm_var": 6.132251305092321, "learning_rate": 0.0001, "loss": 1.2795, "loss/crossentropy": 2.204523801803589, "loss/hidden": 0.84765625, "loss/logits": 0.17835499346256256, "loss/reg": 0.025344664230942726, "step": 238 }, { "epoch": 0.029875, "grad_norm": 3.7683677673339844, "grad_norm_var": 6.125464850588167, "learning_rate": 0.0001, "loss": 0.985, "loss/crossentropy": 2.652397632598877, "loss/hidden": 0.625, "loss/logits": 0.10665792226791382, "loss/reg": 0.0253366157412529, "step": 239 }, { "epoch": 0.03, "grad_norm": 4.732015609741211, "grad_norm_var": 5.870172361717241, "learning_rate": 0.0001, "loss": 1.1018, "loss/crossentropy": 2.544055938720703, "loss/hidden": 0.7109375, "loss/logits": 0.13754525780677795, "loss/reg": 0.02532930299639702, "step": 240 }, { "epoch": 0.030125, "grad_norm": 4.046056270599365, "grad_norm_var": 5.761120866273571, "learning_rate": 0.0001, "loss": 0.9396, "loss/crossentropy": 2.6015427112579346, "loss/hidden": 0.5859375, "loss/logits": 0.10041546076536179, "loss/reg": 0.025322062894701958, "step": 241 }, { "epoch": 0.03025, "grad_norm": 3.066027879714966, "grad_norm_var": 6.052926687728684, "learning_rate": 0.0001, "loss": 1.0668, "loss/crossentropy": 2.4402151107788086, "loss/hidden": 0.6875, "loss/logits": 0.12614867091178894, "loss/reg": 0.0253145694732666, "step": 242 }, { "epoch": 0.030375, "grad_norm": 3.168888807296753, "grad_norm_var": 6.1536859873832555, "learning_rate": 0.0001, "loss": 0.9413, "loss/crossentropy": 2.5689940452575684, "loss/hidden": 0.58984375, "loss/logits": 0.09841729700565338, "loss/reg": 0.02530776336789131, "step": 243 }, { "epoch": 0.0305, "grad_norm": 3.806450366973877, "grad_norm_var": 6.229122059899357, "learning_rate": 0.0001, "loss": 1.021, "loss/crossentropy": 2.2438809871673584, "loss/hidden": 0.6328125, "loss/logits": 0.13515594601631165, "loss/reg": 0.02530042454600334, "step": 244 }, { "epoch": 0.030625, "grad_norm": 4.313736915588379, "grad_norm_var": 5.982441934425083, "learning_rate": 0.0001, "loss": 1.0307, "loss/crossentropy": 2.5067784786224365, "loss/hidden": 0.65234375, "loss/logits": 0.1253766119480133, "loss/reg": 0.025293124839663506, "step": 245 }, { "epoch": 0.03075, "grad_norm": 4.419394016265869, "grad_norm_var": 5.942040082684442, "learning_rate": 0.0001, "loss": 0.8957, "loss/crossentropy": 2.4725239276885986, "loss/hidden": 0.55078125, "loss/logits": 0.09201550483703613, "loss/reg": 0.02528616413474083, "step": 246 }, { "epoch": 0.030875, "grad_norm": 3.709151268005371, "grad_norm_var": 5.975041529003943, "learning_rate": 0.0001, "loss": 0.9152, "loss/crossentropy": 2.413250207901001, "loss/hidden": 0.56640625, "loss/logits": 0.09596529603004456, "loss/reg": 0.025278838351368904, "step": 247 }, { "epoch": 0.031, "grad_norm": 3.4139935970306396, "grad_norm_var": 6.007189625317282, "learning_rate": 0.0001, "loss": 1.0268, "loss/crossentropy": 2.1580560207366943, "loss/hidden": 0.65625, "loss/logits": 0.11786328256130219, "loss/reg": 0.025271562859416008, "step": 248 }, { "epoch": 0.031125, "grad_norm": 6.089378833770752, "grad_norm_var": 2.0950588257899443, "learning_rate": 0.0001, "loss": 1.0999, "loss/crossentropy": 2.5466785430908203, "loss/hidden": 0.703125, "loss/logits": 0.1441519856452942, "loss/reg": 0.025264522060751915, "step": 249 }, { "epoch": 0.03125, "grad_norm": 4.001245498657227, "grad_norm_var": 1.1007847740596406, "learning_rate": 0.0001, "loss": 0.9263, "loss/crossentropy": 2.433518171310425, "loss/hidden": 0.578125, "loss/logits": 0.09559185057878494, "loss/reg": 0.025257611647248268, "step": 250 }, { "epoch": 0.031375, "grad_norm": 4.982790946960449, "grad_norm_var": 0.8252532202355399, "learning_rate": 0.0001, "loss": 1.1439, "loss/crossentropy": 2.3542721271514893, "loss/hidden": 0.73046875, "loss/logits": 0.16091328859329224, "loss/reg": 0.025250321254134178, "step": 251 }, { "epoch": 0.0315, "grad_norm": 3.6227378845214844, "grad_norm_var": 0.8462248829388517, "learning_rate": 0.0001, "loss": 1.1789, "loss/crossentropy": 2.1333043575286865, "loss/hidden": 0.7890625, "loss/logits": 0.1374010145664215, "loss/reg": 0.025242896750569344, "step": 252 }, { "epoch": 0.031625, "grad_norm": 3.415194511413574, "grad_norm_var": 0.6304077366129337, "learning_rate": 0.0001, "loss": 1.0916, "loss/crossentropy": 2.23525071144104, "loss/hidden": 0.71484375, "loss/logits": 0.1243971735239029, "loss/reg": 0.025235962122678757, "step": 253 }, { "epoch": 0.03175, "grad_norm": 3.7671289443969727, "grad_norm_var": 0.5838590152321442, "learning_rate": 0.0001, "loss": 0.908, "loss/crossentropy": 2.254054546356201, "loss/hidden": 0.56640625, "loss/logits": 0.0893106684088707, "loss/reg": 0.025228681042790413, "step": 254 }, { "epoch": 0.031875, "grad_norm": 4.164376735687256, "grad_norm_var": 0.5803655311074387, "learning_rate": 0.0001, "loss": 0.9548, "loss/crossentropy": 2.4128520488739014, "loss/hidden": 0.5859375, "loss/logits": 0.11668873578310013, "loss/reg": 0.025221774354577065, "step": 255 }, { "epoch": 0.032, "grad_norm": 3.6412672996520996, "grad_norm_var": 0.5547959425019464, "learning_rate": 0.0001, "loss": 1.0098, "loss/crossentropy": 2.7745320796966553, "loss/hidden": 0.62890625, "loss/logits": 0.12875014543533325, "loss/reg": 0.025214577093720436, "step": 256 }, { "epoch": 0.032125, "grad_norm": 2.900871515274048, "grad_norm_var": 0.6261772657264061, "learning_rate": 0.0001, "loss": 0.9348, "loss/crossentropy": 2.286968469619751, "loss/hidden": 0.578125, "loss/logits": 0.10455667227506638, "loss/reg": 0.02520710788667202, "step": 257 }, { "epoch": 0.03225, "grad_norm": 3.6853647232055664, "grad_norm_var": 0.5808564529014478, "learning_rate": 0.0001, "loss": 0.9013, "loss/crossentropy": 2.4515562057495117, "loss/hidden": 0.55078125, "loss/logits": 0.09847953915596008, "loss/reg": 0.025199349969625473, "step": 258 }, { "epoch": 0.032375, "grad_norm": 2.9091956615448, "grad_norm_var": 0.6119059054418109, "learning_rate": 0.0001, "loss": 0.8784, "loss/crossentropy": 2.163628339767456, "loss/hidden": 0.5390625, "loss/logits": 0.08743810653686523, "loss/reg": 0.025191258639097214, "step": 259 }, { "epoch": 0.0325, "grad_norm": 2.9488673210144043, "grad_norm_var": 0.6717290813098125, "learning_rate": 0.0001, "loss": 0.9195, "loss/crossentropy": 2.3084585666656494, "loss/hidden": 0.5546875, "loss/logits": 0.11298206448554993, "loss/reg": 0.025183765217661858, "step": 260 }, { "epoch": 0.032625, "grad_norm": 3.4345953464508057, "grad_norm_var": 0.6684943296663647, "learning_rate": 0.0001, "loss": 0.8851, "loss/crossentropy": 2.19558048248291, "loss/hidden": 0.546875, "loss/logits": 0.08649411797523499, "loss/reg": 0.02517561800777912, "step": 261 }, { "epoch": 0.03275, "grad_norm": 28.159074783325195, "grad_norm_var": 37.79188620028727, "learning_rate": 0.0001, "loss": 1.1781, "loss/crossentropy": 2.839057207107544, "loss/hidden": 0.7890625, "loss/logits": 0.13732999563217163, "loss/reg": 0.025168145075440407, "step": 262 }, { "epoch": 0.032875, "grad_norm": 2.734104871749878, "grad_norm_var": 38.05849364469687, "learning_rate": 0.0001, "loss": 0.8892, "loss/crossentropy": 2.4754340648651123, "loss/hidden": 0.54296875, "loss/logits": 0.09462890028953552, "loss/reg": 0.025160137563943863, "step": 263 }, { "epoch": 0.033, "grad_norm": 4.089654922485352, "grad_norm_var": 37.922354469790704, "learning_rate": 0.0001, "loss": 0.9326, "loss/crossentropy": 2.476450204849243, "loss/hidden": 0.57421875, "loss/logits": 0.1068512499332428, "loss/reg": 0.025151856243610382, "step": 264 }, { "epoch": 0.033125, "grad_norm": 3.838066577911377, "grad_norm_var": 37.99741003814723, "learning_rate": 0.0001, "loss": 1.0909, "loss/crossentropy": 2.114243268966675, "loss/hidden": 0.734375, "loss/logits": 0.10510236769914627, "loss/reg": 0.025143325328826904, "step": 265 }, { "epoch": 0.03325, "grad_norm": 3.0674874782562256, "grad_norm_var": 38.19410456778619, "learning_rate": 0.0001, "loss": 0.8898, "loss/crossentropy": 2.6426563262939453, "loss/hidden": 0.54296875, "loss/logits": 0.09550425410270691, "loss/reg": 0.025135278701782227, "step": 266 }, { "epoch": 0.033375, "grad_norm": 7.777696132659912, "grad_norm_var": 38.64421623432597, "learning_rate": 0.0001, "loss": 1.0482, "loss/crossentropy": 2.640554666519165, "loss/hidden": 0.6796875, "loss/logits": 0.11728060245513916, "loss/reg": 0.025126684457063675, "step": 267 }, { "epoch": 0.0335, "grad_norm": 3.6375653743743896, "grad_norm_var": 38.64099364344996, "learning_rate": 0.0001, "loss": 0.9888, "loss/crossentropy": 2.227538824081421, "loss/hidden": 0.63671875, "loss/logits": 0.1008879542350769, "loss/reg": 0.0251180287450552, "step": 268 }, { "epoch": 0.033625, "grad_norm": 4.305724620819092, "grad_norm_var": 38.4714335626231, "learning_rate": 0.0001, "loss": 1.1268, "loss/crossentropy": 2.556248903274536, "loss/hidden": 0.73828125, "loss/logits": 0.1374487727880478, "loss/reg": 0.02511041797697544, "step": 269 }, { "epoch": 0.03375, "grad_norm": 3.658052921295166, "grad_norm_var": 38.4947077039297, "learning_rate": 0.0001, "loss": 0.9616, "loss/crossentropy": 2.437307596206665, "loss/hidden": 0.60546875, "loss/logits": 0.1050904244184494, "loss/reg": 0.025102900341153145, "step": 270 }, { "epoch": 0.033875, "grad_norm": 6.534849643707275, "grad_norm_var": 38.483973576312195, "learning_rate": 0.0001, "loss": 1.2119, "loss/crossentropy": 2.36796236038208, "loss/hidden": 0.81640625, "loss/logits": 0.1445278525352478, "loss/reg": 0.025095317512750626, "step": 271 }, { "epoch": 0.034, "grad_norm": 5.486751079559326, "grad_norm_var": 38.24988881420663, "learning_rate": 0.0001, "loss": 1.0337, "loss/crossentropy": 2.517333984375, "loss/hidden": 0.66015625, "loss/logits": 0.12263435125350952, "loss/reg": 0.025087665766477585, "step": 272 }, { "epoch": 0.034125, "grad_norm": 3.231940746307373, "grad_norm_var": 38.13878485092763, "learning_rate": 0.0001, "loss": 1.0586, "loss/crossentropy": 2.6812305450439453, "loss/hidden": 0.67578125, "loss/logits": 0.13204006850719452, "loss/reg": 0.025080092251300812, "step": 273 }, { "epoch": 0.03425, "grad_norm": 4.71685791015625, "grad_norm_var": 37.942827296069424, "learning_rate": 0.0001, "loss": 1.0409, "loss/crossentropy": 2.093151807785034, "loss/hidden": 0.6796875, "loss/logits": 0.11049959808588028, "loss/reg": 0.02507280558347702, "step": 274 }, { "epoch": 0.034375, "grad_norm": 3.7086901664733887, "grad_norm_var": 37.68973967522894, "learning_rate": 0.0001, "loss": 0.9144, "loss/crossentropy": 2.2392518520355225, "loss/hidden": 0.57421875, "loss/logits": 0.08954055607318878, "loss/reg": 0.02506582997739315, "step": 275 }, { "epoch": 0.0345, "grad_norm": 5.537543773651123, "grad_norm_var": 37.1561912525544, "learning_rate": 0.0001, "loss": 1.3586, "loss/crossentropy": 1.8622019290924072, "loss/hidden": 0.96484375, "loss/logits": 0.14318415522575378, "loss/reg": 0.025058824568986893, "step": 276 }, { "epoch": 0.034625, "grad_norm": 2.898383617401123, "grad_norm_var": 37.34827444255352, "learning_rate": 0.0001, "loss": 0.9413, "loss/crossentropy": 2.4155869483947754, "loss/hidden": 0.578125, "loss/logits": 0.11263684928417206, "loss/reg": 0.025052132084965706, "step": 277 }, { "epoch": 0.03475, "grad_norm": 2.803711175918579, "grad_norm_var": 2.0625830733920933, "learning_rate": 0.0001, "loss": 0.9404, "loss/crossentropy": 2.2058048248291016, "loss/hidden": 0.59765625, "loss/logits": 0.09232598543167114, "loss/reg": 0.02504453808069229, "step": 278 }, { "epoch": 0.034875, "grad_norm": 3.180114507675171, "grad_norm_var": 1.9847680294286107, "learning_rate": 0.0001, "loss": 0.9311, "loss/crossentropy": 2.6069722175598145, "loss/hidden": 0.57421875, "loss/logits": 0.10655493289232254, "loss/reg": 0.02503693662583828, "step": 279 }, { "epoch": 0.035, "grad_norm": 5.609209060668945, "grad_norm_var": 2.0906055341906256, "learning_rate": 0.0001, "loss": 1.0675, "loss/crossentropy": 2.5023670196533203, "loss/hidden": 0.68359375, "loss/logits": 0.13361594080924988, "loss/reg": 0.025029515847563744, "step": 280 }, { "epoch": 0.035125, "grad_norm": 5.165370464324951, "grad_norm_var": 2.105772188928517, "learning_rate": 0.0001, "loss": 1.1103, "loss/crossentropy": 2.5107007026672363, "loss/hidden": 0.73828125, "loss/logits": 0.12183347344398499, "loss/reg": 0.025022249668836594, "step": 281 }, { "epoch": 0.03525, "grad_norm": 4.322115421295166, "grad_norm_var": 1.9716269568170388, "learning_rate": 0.0001, "loss": 0.95, "loss/crossentropy": 2.483142137527466, "loss/hidden": 0.5859375, "loss/logits": 0.11390361189842224, "loss/reg": 0.025014575570821762, "step": 282 }, { "epoch": 0.035375, "grad_norm": 4.432065486907959, "grad_norm_var": 1.2250959918754403, "learning_rate": 0.0001, "loss": 0.9886, "loss/crossentropy": 2.1206424236297607, "loss/hidden": 0.6328125, "loss/logits": 0.10568062961101532, "loss/reg": 0.025007015094161034, "step": 283 }, { "epoch": 0.0355, "grad_norm": 4.0429558753967285, "grad_norm_var": 1.198112283867575, "learning_rate": 0.0001, "loss": 1.0004, "loss/crossentropy": 2.6465060710906982, "loss/hidden": 0.6328125, "loss/logits": 0.11765305697917938, "loss/reg": 0.024997249245643616, "step": 284 }, { "epoch": 0.035625, "grad_norm": 3.551006555557251, "grad_norm_var": 1.23838358717468, "learning_rate": 0.0001, "loss": 0.9764, "loss/crossentropy": 2.4274182319641113, "loss/hidden": 0.61328125, "loss/logits": 0.11320526152849197, "loss/reg": 0.02498740889132023, "step": 285 }, { "epoch": 0.03575, "grad_norm": 2.874697685241699, "grad_norm_var": 1.344305852802292, "learning_rate": 0.0001, "loss": 1.0133, "loss/crossentropy": 2.512948751449585, "loss/hidden": 0.6484375, "loss/logits": 0.11510799080133438, "loss/reg": 0.02497740648686886, "step": 286 }, { "epoch": 0.035875, "grad_norm": 3.4613194465637207, "grad_norm_var": 1.0008425760722102, "learning_rate": 0.0001, "loss": 0.9274, "loss/crossentropy": 2.4450602531433105, "loss/hidden": 0.5703125, "loss/logits": 0.10743874311447144, "loss/reg": 0.024968616664409637, "step": 287 }, { "epoch": 0.036, "grad_norm": 3.452150344848633, "grad_norm_var": 0.8735820507410946, "learning_rate": 0.0001, "loss": 0.952, "loss/crossentropy": 2.304729700088501, "loss/hidden": 0.578125, "loss/logits": 0.1242954432964325, "loss/reg": 0.024959923699498177, "step": 288 }, { "epoch": 0.036125, "grad_norm": 3.3603436946868896, "grad_norm_var": 0.8625457550688983, "learning_rate": 0.0001, "loss": 0.9447, "loss/crossentropy": 2.425712823867798, "loss/hidden": 0.58984375, "loss/logits": 0.1053181067109108, "loss/reg": 0.024950530380010605, "step": 289 }, { "epoch": 0.03625, "grad_norm": 3.3667662143707275, "grad_norm_var": 0.8374846368179986, "learning_rate": 0.0001, "loss": 0.9067, "loss/crossentropy": 2.2623162269592285, "loss/hidden": 0.55859375, "loss/logits": 0.098650723695755, "loss/reg": 0.02494119666516781, "step": 290 }, { "epoch": 0.036375, "grad_norm": 2.7996857166290283, "grad_norm_var": 0.9075153562133816, "learning_rate": 0.0001, "loss": 0.9594, "loss/crossentropy": 2.370417356491089, "loss/hidden": 0.59765625, "loss/logits": 0.11243726313114166, "loss/reg": 0.024933209642767906, "step": 291 }, { "epoch": 0.0365, "grad_norm": 2.9999380111694336, "grad_norm_var": 0.7233017120786666, "learning_rate": 0.0001, "loss": 1.0396, "loss/crossentropy": 2.286776542663574, "loss/hidden": 0.6640625, "loss/logits": 0.12626898288726807, "loss/reg": 0.02492516115307808, "step": 292 }, { "epoch": 0.036625, "grad_norm": 3.880777597427368, "grad_norm_var": 0.685825505757979, "learning_rate": 0.0001, "loss": 0.9764, "loss/crossentropy": 2.613823652267456, "loss/hidden": 0.61328125, "loss/logits": 0.11395551264286041, "loss/reg": 0.02491726726293564, "step": 293 }, { "epoch": 0.03675, "grad_norm": 4.2950568199157715, "grad_norm_var": 0.6453385025094112, "learning_rate": 0.0001, "loss": 1.0535, "loss/crossentropy": 2.6882171630859375, "loss/hidden": 0.66796875, "loss/logits": 0.13644324243068695, "loss/reg": 0.02490835078060627, "step": 294 }, { "epoch": 0.036875, "grad_norm": 3.61423659324646, "grad_norm_var": 0.6212598300915251, "learning_rate": 0.0001, "loss": 1.155, "loss/crossentropy": 2.0867068767547607, "loss/hidden": 0.76953125, "loss/logits": 0.13649392127990723, "loss/reg": 0.024899456650018692, "step": 295 }, { "epoch": 0.037, "grad_norm": 3.9245386123657227, "grad_norm_var": 0.3982568915415859, "learning_rate": 0.0001, "loss": 0.951, "loss/crossentropy": 2.703791618347168, "loss/hidden": 0.5859375, "loss/logits": 0.11614765971899033, "loss/reg": 0.02489159069955349, "step": 296 }, { "epoch": 0.037125, "grad_norm": 4.020047187805176, "grad_norm_var": 0.2597397925732442, "learning_rate": 0.0001, "loss": 0.9707, "loss/crossentropy": 2.429636001586914, "loss/hidden": 0.61328125, "loss/logits": 0.1086028665304184, "loss/reg": 0.024883201345801353, "step": 297 }, { "epoch": 0.03725, "grad_norm": 4.086516857147217, "grad_norm_var": 0.24209119003572066, "learning_rate": 0.0001, "loss": 1.0769, "loss/crossentropy": 2.8581056594848633, "loss/hidden": 0.6796875, "loss/logits": 0.14844708144664764, "loss/reg": 0.024874594062566757, "step": 298 }, { "epoch": 0.037375, "grad_norm": 2.956085443496704, "grad_norm_var": 0.22141400399237127, "learning_rate": 0.0001, "loss": 0.9903, "loss/crossentropy": 2.4802041053771973, "loss/hidden": 0.609375, "loss/logits": 0.13230839371681213, "loss/reg": 0.024866636842489243, "step": 299 }, { "epoch": 0.0375, "grad_norm": 3.165804386138916, "grad_norm_var": 0.2110158468875736, "learning_rate": 0.0001, "loss": 1.0347, "loss/crossentropy": 2.4370639324188232, "loss/hidden": 0.66015625, "loss/logits": 0.1259341686964035, "loss/reg": 0.024857714772224426, "step": 300 }, { "epoch": 0.037625, "grad_norm": 3.0942695140838623, "grad_norm_var": 0.22022059823099174, "learning_rate": 0.0001, "loss": 0.9747, "loss/crossentropy": 2.674190044403076, "loss/hidden": 0.6015625, "loss/logits": 0.12467385828495026, "loss/reg": 0.024848785251379013, "step": 301 }, { "epoch": 0.03775, "grad_norm": 2.949205160140991, "grad_norm_var": 0.21475779393049, "learning_rate": 0.0001, "loss": 0.8898, "loss/crossentropy": 2.491746664047241, "loss/hidden": 0.55078125, "loss/logits": 0.09065801650285721, "loss/reg": 0.02483983524143696, "step": 302 }, { "epoch": 0.037875, "grad_norm": 3.549431324005127, "grad_norm_var": 0.21520952048913009, "learning_rate": 0.0001, "loss": 0.9197, "loss/crossentropy": 2.7578208446502686, "loss/hidden": 0.57421875, "loss/logits": 0.09712585806846619, "loss/reg": 0.024831857532262802, "step": 303 }, { "epoch": 0.038, "grad_norm": 3.0621931552886963, "grad_norm_var": 0.22562503941355938, "learning_rate": 0.0001, "loss": 0.894, "loss/crossentropy": 2.440351724624634, "loss/hidden": 0.546875, "loss/logits": 0.09888219833374023, "loss/reg": 0.02482294850051403, "step": 304 }, { "epoch": 0.038125, "grad_norm": 3.334942579269409, "grad_norm_var": 0.22595311715915212, "learning_rate": 0.0001, "loss": 1.1509, "loss/crossentropy": 2.378168821334839, "loss/hidden": 0.74609375, "loss/logits": 0.15670377016067505, "loss/reg": 0.024814244359731674, "step": 305 }, { "epoch": 0.03825, "grad_norm": 3.3253893852233887, "grad_norm_var": 0.22648465837488155, "learning_rate": 0.0001, "loss": 0.9744, "loss/crossentropy": 2.4409470558166504, "loss/hidden": 0.6015625, "loss/logits": 0.12483348697423935, "loss/reg": 0.024805361405014992, "step": 306 }, { "epoch": 0.038375, "grad_norm": 3.1687541007995605, "grad_norm_var": 0.20343285009947346, "learning_rate": 0.0001, "loss": 0.8998, "loss/crossentropy": 2.3339829444885254, "loss/hidden": 0.55078125, "loss/logits": 0.10110392421483994, "loss/reg": 0.02479635737836361, "step": 307 }, { "epoch": 0.0385, "grad_norm": 4.096661567687988, "grad_norm_var": 0.21071919009248533, "learning_rate": 0.0001, "loss": 1.0154, "loss/crossentropy": 2.468813180923462, "loss/hidden": 0.64453125, "loss/logits": 0.12295258045196533, "loss/reg": 0.024787236005067825, "step": 308 }, { "epoch": 0.038625, "grad_norm": 3.097642660140991, "grad_norm_var": 0.2127095324618587, "learning_rate": 0.0001, "loss": 0.8772, "loss/crossentropy": 2.5380513668060303, "loss/hidden": 0.53125, "loss/logits": 0.09821398556232452, "loss/reg": 0.024777989834547043, "step": 309 }, { "epoch": 0.03875, "grad_norm": 4.647727012634277, "grad_norm_var": 0.25863060133758775, "learning_rate": 0.0001, "loss": 0.9162, "loss/crossentropy": 2.0996336936950684, "loss/hidden": 0.5625, "loss/logits": 0.10599061101675034, "loss/reg": 0.024768849834799767, "step": 310 }, { "epoch": 0.038875, "grad_norm": 3.627246141433716, "grad_norm_var": 0.25882920418562944, "learning_rate": 0.0001, "loss": 1.0161, "loss/crossentropy": 2.2958993911743164, "loss/hidden": 0.65234375, "loss/logits": 0.11611323803663254, "loss/reg": 0.024759870022535324, "step": 311 }, { "epoch": 0.039, "grad_norm": 3.7236764430999756, "grad_norm_var": 0.2501591619921593, "learning_rate": 0.0001, "loss": 1.1308, "loss/crossentropy": 2.4322752952575684, "loss/hidden": 0.7421875, "loss/logits": 0.1410805881023407, "loss/reg": 0.024751078337430954, "step": 312 }, { "epoch": 0.039125, "grad_norm": 3.998807430267334, "grad_norm_var": 0.24869789076210413, "learning_rate": 0.0001, "loss": 1.0749, "loss/crossentropy": 2.374758720397949, "loss/hidden": 0.703125, "loss/logits": 0.1243831142783165, "loss/reg": 0.024742012843489647, "step": 313 }, { "epoch": 0.03925, "grad_norm": 4.315618515014648, "grad_norm_var": 0.2701154191311919, "learning_rate": 0.0001, "loss": 0.9542, "loss/crossentropy": 2.290766477584839, "loss/hidden": 0.60546875, "loss/logits": 0.10138154029846191, "loss/reg": 0.02473386563360691, "step": 314 }, { "epoch": 0.039375, "grad_norm": 11.789870262145996, "grad_norm_var": 4.498354875640615, "learning_rate": 0.0001, "loss": 1.6276, "loss/crossentropy": 2.391585350036621, "loss/hidden": 1.25, "loss/logits": 0.13034726679325104, "loss/reg": 0.02472575195133686, "step": 315 }, { "epoch": 0.0395, "grad_norm": 3.499288320541382, "grad_norm_var": 4.465581075155145, "learning_rate": 0.0001, "loss": 0.9478, "loss/crossentropy": 2.4527926445007324, "loss/hidden": 0.5703125, "loss/logits": 0.13031822443008423, "loss/reg": 0.024717645719647408, "step": 316 }, { "epoch": 0.039625, "grad_norm": 3.0139851570129395, "grad_norm_var": 4.476536239649594, "learning_rate": 0.0001, "loss": 0.9555, "loss/crossentropy": 2.506457805633545, "loss/hidden": 0.58203125, "loss/logits": 0.12640972435474396, "loss/reg": 0.02470862865447998, "step": 317 }, { "epoch": 0.03975, "grad_norm": 3.671523094177246, "grad_norm_var": 4.4007183053584145, "learning_rate": 0.0001, "loss": 1.0902, "loss/crossentropy": 2.1730313301086426, "loss/hidden": 0.7109375, "loss/logits": 0.13228294253349304, "loss/reg": 0.024700626730918884, "step": 318 }, { "epoch": 0.039875, "grad_norm": 3.453159809112549, "grad_norm_var": 4.408623714873802, "learning_rate": 0.0001, "loss": 0.9518, "loss/crossentropy": 2.7452075481414795, "loss/hidden": 0.58984375, "loss/logits": 0.11500866711139679, "loss/reg": 0.02469259686768055, "step": 319 }, { "epoch": 0.04, "grad_norm": 8.18558406829834, "grad_norm_var": 5.330579476502794, "learning_rate": 0.0001, "loss": 1.2593, "loss/crossentropy": 2.295231819152832, "loss/hidden": 0.78125, "loss/logits": 0.2312176525592804, "loss/reg": 0.024684064090251923, "step": 320 }, { "epoch": 0.040125, "grad_norm": 4.58513879776001, "grad_norm_var": 5.245000173569268, "learning_rate": 0.0001, "loss": 1.0428, "loss/crossentropy": 2.3456668853759766, "loss/hidden": 0.67578125, "loss/logits": 0.12027865648269653, "loss/reg": 0.024676108732819557, "step": 321 }, { "epoch": 0.04025, "grad_norm": 3.6148953437805176, "grad_norm_var": 5.20441494141252, "learning_rate": 0.0001, "loss": 0.885, "loss/crossentropy": 2.3615198135375977, "loss/hidden": 0.53515625, "loss/logits": 0.10311760008335114, "loss/reg": 0.02466769702732563, "step": 322 }, { "epoch": 0.040375, "grad_norm": 4.596187114715576, "grad_norm_var": 5.072570501388933, "learning_rate": 0.0001, "loss": 1.3642, "loss/crossentropy": 1.9407044649124146, "loss/hidden": 0.9140625, "loss/logits": 0.20351722836494446, "loss/reg": 0.02465960383415222, "step": 323 }, { "epoch": 0.0405, "grad_norm": 4.305622577667236, "grad_norm_var": 5.060723771971758, "learning_rate": 0.0001, "loss": 1.2854, "loss/crossentropy": 2.1887283325195312, "loss/hidden": 0.83203125, "loss/logits": 0.20682212710380554, "loss/reg": 0.024651024490594864, "step": 324 }, { "epoch": 0.040625, "grad_norm": 3.1052956581115723, "grad_norm_var": 5.059160883569213, "learning_rate": 0.0001, "loss": 0.9277, "loss/crossentropy": 2.7552340030670166, "loss/hidden": 0.56640625, "loss/logits": 0.11491444706916809, "loss/reg": 0.024642454460263252, "step": 325 }, { "epoch": 0.04075, "grad_norm": 3.902174234390259, "grad_norm_var": 5.092472426369553, "learning_rate": 0.0001, "loss": 1.0922, "loss/crossentropy": 2.328878402709961, "loss/hidden": 0.70703125, "loss/logits": 0.13887017965316772, "loss/reg": 0.024634363129734993, "step": 326 }, { "epoch": 0.040875, "grad_norm": 3.144326686859131, "grad_norm_var": 5.168830163995767, "learning_rate": 0.0001, "loss": 0.9028, "loss/crossentropy": 2.765711545944214, "loss/hidden": 0.55078125, "loss/logits": 0.10574661940336227, "loss/reg": 0.02462565153837204, "step": 327 }, { "epoch": 0.041, "grad_norm": 4.085259914398193, "grad_norm_var": 5.136846736797656, "learning_rate": 0.0001, "loss": 1.1366, "loss/crossentropy": 2.2278430461883545, "loss/hidden": 0.75390625, "loss/logits": 0.13655412197113037, "loss/reg": 0.0246175117790699, "step": 328 }, { "epoch": 0.041125, "grad_norm": 3.578554391860962, "grad_norm_var": 5.180404969237465, "learning_rate": 0.0001, "loss": 1.0227, "loss/crossentropy": 2.7795305252075195, "loss/hidden": 0.640625, "loss/logits": 0.13600832223892212, "loss/reg": 0.024609515443444252, "step": 329 }, { "epoch": 0.04125, "grad_norm": 3.7418434619903564, "grad_norm_var": 5.219134310055354, "learning_rate": 0.0001, "loss": 1.0244, "loss/crossentropy": 2.636258840560913, "loss/hidden": 0.63671875, "loss/logits": 0.1417178213596344, "loss/reg": 0.024600951001048088, "step": 330 }, { "epoch": 0.041375, "grad_norm": 3.369840383529663, "grad_norm_var": 1.4852025101017772, "learning_rate": 0.0001, "loss": 1.0208, "loss/crossentropy": 2.448961019515991, "loss/hidden": 0.6640625, "loss/logits": 0.1107659712433815, "loss/reg": 0.024592852219939232, "step": 331 }, { "epoch": 0.0415, "grad_norm": 2.90629506111145, "grad_norm_var": 1.5460412234751968, "learning_rate": 0.0001, "loss": 0.9411, "loss/crossentropy": 2.595834732055664, "loss/hidden": 0.59375, "loss/logits": 0.10147438943386078, "loss/reg": 0.024584442377090454, "step": 332 }, { "epoch": 0.041625, "grad_norm": 3.402386426925659, "grad_norm_var": 1.5068032644485272, "learning_rate": 0.0001, "loss": 1.0144, "loss/crossentropy": 2.363499402999878, "loss/hidden": 0.62890625, "loss/logits": 0.1397327035665512, "loss/reg": 0.02457563206553459, "step": 333 }, { "epoch": 0.04175, "grad_norm": 3.597465753555298, "grad_norm_var": 1.5101723473758888, "learning_rate": 0.0001, "loss": 1.1917, "loss/crossentropy": 2.3256943225860596, "loss/hidden": 0.7890625, "loss/logits": 0.15691371262073517, "loss/reg": 0.02456764318048954, "step": 334 }, { "epoch": 0.041875, "grad_norm": 3.512730121612549, "grad_norm_var": 1.506262203991567, "learning_rate": 0.0001, "loss": 1.0631, "loss/crossentropy": 2.383385181427002, "loss/hidden": 0.69140625, "loss/logits": 0.12608817219734192, "loss/reg": 0.024559510871767998, "step": 335 }, { "epoch": 0.042, "grad_norm": 5.968827247619629, "grad_norm_var": 0.5694964439723395, "learning_rate": 0.0001, "loss": 1.258, "loss/crossentropy": 2.212991952896118, "loss/hidden": 0.828125, "loss/logits": 0.18438610434532166, "loss/reg": 0.02455153875052929, "step": 336 }, { "epoch": 0.042125, "grad_norm": 3.644115924835205, "grad_norm_var": 0.5311677507970897, "learning_rate": 0.0001, "loss": 1.1203, "loss/crossentropy": 2.3620948791503906, "loss/hidden": 0.72265625, "loss/logits": 0.15219135582447052, "loss/reg": 0.02454366721212864, "step": 337 }, { "epoch": 0.04225, "grad_norm": 2.716367244720459, "grad_norm_var": 0.6013761572733584, "learning_rate": 0.0001, "loss": 0.919, "loss/crossentropy": 2.3863117694854736, "loss/hidden": 0.578125, "loss/logits": 0.0955524817109108, "loss/reg": 0.024536145851016045, "step": 338 }, { "epoch": 0.042375, "grad_norm": 3.8117287158966064, "grad_norm_var": 0.5485673092684531, "learning_rate": 0.0001, "loss": 1.0529, "loss/crossentropy": 2.4951469898223877, "loss/hidden": 0.67578125, "loss/logits": 0.1318708062171936, "loss/reg": 0.024528514593839645, "step": 339 }, { "epoch": 0.0425, "grad_norm": 2.5218894481658936, "grad_norm_var": 0.5973356289049185, "learning_rate": 0.0001, "loss": 0.9673, "loss/crossentropy": 2.370839834213257, "loss/hidden": 0.6015625, "loss/logits": 0.12049120664596558, "loss/reg": 0.02452007494866848, "step": 340 }, { "epoch": 0.042625, "grad_norm": 2.7343108654022217, "grad_norm_var": 0.6285810690168129, "learning_rate": 0.0001, "loss": 1.1201, "loss/crossentropy": 2.2990617752075195, "loss/hidden": 0.7109375, "loss/logits": 0.16403642296791077, "loss/reg": 0.024511409923434258, "step": 341 }, { "epoch": 0.04275, "grad_norm": 2.5267186164855957, "grad_norm_var": 0.6803812464423664, "learning_rate": 0.0001, "loss": 0.9056, "loss/crossentropy": 2.5629754066467285, "loss/hidden": 0.5546875, "loss/logits": 0.10588675737380981, "loss/reg": 0.024503152817487717, "step": 342 }, { "epoch": 0.042875, "grad_norm": 2.2731988430023193, "grad_norm_var": 0.7637691760365584, "learning_rate": 0.0001, "loss": 0.8154, "loss/crossentropy": 2.4123644828796387, "loss/hidden": 0.490234375, "loss/logits": 0.08022630214691162, "loss/reg": 0.024494923651218414, "step": 343 }, { "epoch": 0.043, "grad_norm": 3.2791945934295654, "grad_norm_var": 0.7306725618305314, "learning_rate": 0.0001, "loss": 0.9879, "loss/crossentropy": 2.7643816471099854, "loss/hidden": 0.6171875, "loss/logits": 0.12586694955825806, "loss/reg": 0.024486759677529335, "step": 344 }, { "epoch": 0.043125, "grad_norm": 3.6740994453430176, "grad_norm_var": 0.7341663188433093, "learning_rate": 0.0001, "loss": 0.8947, "loss/crossentropy": 2.6802780628204346, "loss/hidden": 0.5546875, "loss/logits": 0.0951782613992691, "loss/reg": 0.024478696286678314, "step": 345 }, { "epoch": 0.04325, "grad_norm": 3.234722137451172, "grad_norm_var": 0.7240869727338347, "learning_rate": 0.0001, "loss": 0.8517, "loss/crossentropy": 2.577921152114868, "loss/hidden": 0.515625, "loss/logits": 0.09137749671936035, "loss/reg": 0.024470962584018707, "step": 346 }, { "epoch": 0.043375, "grad_norm": 4.024074554443359, "grad_norm_var": 0.7548921970504276, "learning_rate": 0.0001, "loss": 1.0017, "loss/crossentropy": 2.8129634857177734, "loss/hidden": 0.63671875, "loss/logits": 0.120377317070961, "loss/reg": 0.02446298860013485, "step": 347 }, { "epoch": 0.0435, "grad_norm": 3.4327027797698975, "grad_norm_var": 0.7400679146500114, "learning_rate": 0.0001, "loss": 0.9864, "loss/crossentropy": 2.4310224056243896, "loss/hidden": 0.640625, "loss/logits": 0.10121208429336548, "loss/reg": 0.0244552381336689, "step": 348 }, { "epoch": 0.043625, "grad_norm": 3.2115890979766846, "grad_norm_var": 0.7422101391295163, "learning_rate": 0.0001, "loss": 1.0594, "loss/crossentropy": 2.3658318519592285, "loss/hidden": 0.6796875, "loss/logits": 0.13528358936309814, "loss/reg": 0.02444704994559288, "step": 349 }, { "epoch": 0.04375, "grad_norm": 3.4005136489868164, "grad_norm_var": 0.739061242813568, "learning_rate": 0.0001, "loss": 1.0154, "loss/crossentropy": 2.9171154499053955, "loss/hidden": 0.6171875, "loss/logits": 0.15386469662189484, "loss/reg": 0.024438532069325447, "step": 350 }, { "epoch": 0.043875, "grad_norm": 4.1175127029418945, "grad_norm_var": 0.7731950105324129, "learning_rate": 0.0001, "loss": 0.949, "loss/crossentropy": 2.3405494689941406, "loss/hidden": 0.59765625, "loss/logits": 0.10704682767391205, "loss/reg": 0.024429937824606895, "step": 351 }, { "epoch": 0.044, "grad_norm": 3.9496090412139893, "grad_norm_var": 0.33930652052577653, "learning_rate": 0.0001, "loss": 0.889, "loss/crossentropy": 2.9135613441467285, "loss/hidden": 0.54296875, "loss/logits": 0.1018136739730835, "loss/reg": 0.024421829730272293, "step": 352 }, { "epoch": 0.044125, "grad_norm": 3.794520139694214, "grad_norm_var": 0.347931624130162, "learning_rate": 0.0001, "loss": 1.1771, "loss/crossentropy": 2.5529844760894775, "loss/hidden": 0.75390625, "loss/logits": 0.17909468710422516, "loss/reg": 0.02441396936774254, "step": 353 }, { "epoch": 0.04425, "grad_norm": 3.164349317550659, "grad_norm_var": 0.3259767305032634, "learning_rate": 0.0001, "loss": 1.0804, "loss/crossentropy": 2.3843231201171875, "loss/hidden": 0.71484375, "loss/logits": 0.12148329615592957, "loss/reg": 0.02440580353140831, "step": 354 }, { "epoch": 0.044375, "grad_norm": 2.7847914695739746, "grad_norm_var": 0.32482231475126633, "learning_rate": 0.0001, "loss": 1.0463, "loss/crossentropy": 2.521649122238159, "loss/hidden": 0.68359375, "loss/logits": 0.1187494546175003, "loss/reg": 0.024397339671850204, "step": 355 }, { "epoch": 0.0445, "grad_norm": 3.2391860485076904, "grad_norm_var": 0.28660331114573767, "learning_rate": 0.0001, "loss": 1.0225, "loss/crossentropy": 2.4073832035064697, "loss/hidden": 0.640625, "loss/logits": 0.13795481622219086, "loss/reg": 0.024389205500483513, "step": 356 }, { "epoch": 0.044625, "grad_norm": 5.839352130889893, "grad_norm_var": 0.6539216724231817, "learning_rate": 0.0001, "loss": 1.1058, "loss/crossentropy": 2.6860220432281494, "loss/hidden": 0.7265625, "loss/logits": 0.13543812930583954, "loss/reg": 0.02438061311841011, "step": 357 }, { "epoch": 0.04475, "grad_norm": 4.340758800506592, "grad_norm_var": 0.6249977794062299, "learning_rate": 0.0001, "loss": 1.1, "loss/crossentropy": 2.770012378692627, "loss/hidden": 0.703125, "loss/logits": 0.1531440019607544, "loss/reg": 0.024372335523366928, "step": 358 }, { "epoch": 0.044875, "grad_norm": 8.795727729797363, "grad_norm_var": 2.1213731683569863, "learning_rate": 0.0001, "loss": 1.3149, "loss/crossentropy": 2.2893428802490234, "loss/hidden": 0.91796875, "loss/logits": 0.15326841175556183, "loss/reg": 0.02436378225684166, "step": 359 }, { "epoch": 0.045, "grad_norm": 4.104447841644287, "grad_norm_var": 2.0826812332104394, "learning_rate": 0.0001, "loss": 1.15, "loss/crossentropy": 2.625378370285034, "loss/hidden": 0.7578125, "loss/logits": 0.14858195185661316, "loss/reg": 0.024355949833989143, "step": 360 }, { "epoch": 0.045125, "grad_norm": 3.354828357696533, "grad_norm_var": 2.105873348197963, "learning_rate": 0.0001, "loss": 0.9494, "loss/crossentropy": 2.4916443824768066, "loss/hidden": 0.5859375, "loss/logits": 0.11993111670017242, "loss/reg": 0.024348480626940727, "step": 361 }, { "epoch": 0.04525, "grad_norm": 3.3247056007385254, "grad_norm_var": 2.096606359520402, "learning_rate": 0.0001, "loss": 0.9604, "loss/crossentropy": 2.282543420791626, "loss/hidden": 0.609375, "loss/logits": 0.10765975713729858, "loss/reg": 0.02434113249182701, "step": 362 }, { "epoch": 0.045375, "grad_norm": 4.314214706420898, "grad_norm_var": 2.1006745469652883, "learning_rate": 0.0001, "loss": 1.0684, "loss/crossentropy": 2.447218179702759, "loss/hidden": 0.6953125, "loss/logits": 0.12974585592746735, "loss/reg": 0.024332784116268158, "step": 363 }, { "epoch": 0.0455, "grad_norm": 3.6783666610717773, "grad_norm_var": 2.083471757970481, "learning_rate": 0.0001, "loss": 1.1509, "loss/crossentropy": 2.6020665168762207, "loss/hidden": 0.75, "loss/logits": 0.1576002687215805, "loss/reg": 0.024325383827090263, "step": 364 }, { "epoch": 0.045625, "grad_norm": 3.535550832748413, "grad_norm_var": 2.0521572529950523, "learning_rate": 0.0001, "loss": 0.9672, "loss/crossentropy": 2.2514841556549072, "loss/hidden": 0.61328125, "loss/logits": 0.110772505402565, "loss/reg": 0.024317855015397072, "step": 365 }, { "epoch": 0.04575, "grad_norm": 6.190865993499756, "grad_norm_var": 2.275325586048537, "learning_rate": 0.0001, "loss": 1.2696, "loss/crossentropy": 2.5250887870788574, "loss/hidden": 0.87890625, "loss/logits": 0.1475657969713211, "loss/reg": 0.02430957928299904, "step": 366 }, { "epoch": 0.045875, "grad_norm": 6.9109907150268555, "grad_norm_var": 2.701389202772653, "learning_rate": 0.0001, "loss": 1.42, "loss/crossentropy": 2.0714285373687744, "loss/hidden": 0.98828125, "loss/logits": 0.18868008255958557, "loss/reg": 0.024301210418343544, "step": 367 }, { "epoch": 0.046, "grad_norm": 3.939924955368042, "grad_norm_var": 2.702051041555256, "learning_rate": 0.0001, "loss": 1.0163, "loss/crossentropy": 2.677281379699707, "loss/hidden": 0.640625, "loss/logits": 0.1327013224363327, "loss/reg": 0.024292904883623123, "step": 368 }, { "epoch": 0.046125, "grad_norm": 2.942032814025879, "grad_norm_var": 2.822776844100568, "learning_rate": 0.0001, "loss": 0.9582, "loss/crossentropy": 2.3630495071411133, "loss/hidden": 0.6015625, "loss/logits": 0.11380600929260254, "loss/reg": 0.02428455464541912, "step": 369 }, { "epoch": 0.04625, "grad_norm": 3.125349521636963, "grad_norm_var": 2.8293167859701627, "learning_rate": 0.0001, "loss": 1.0463, "loss/crossentropy": 2.5113635063171387, "loss/hidden": 0.67578125, "loss/logits": 0.12774598598480225, "loss/reg": 0.02427608147263527, "step": 370 }, { "epoch": 0.046375, "grad_norm": 3.15159273147583, "grad_norm_var": 2.758666518773039, "learning_rate": 0.0001, "loss": 0.9198, "loss/crossentropy": 2.31132435798645, "loss/hidden": 0.5703125, "loss/logits": 0.10678299516439438, "loss/reg": 0.024267377331852913, "step": 371 }, { "epoch": 0.0465, "grad_norm": 4.394677639007568, "grad_norm_var": 2.6595375525429406, "learning_rate": 0.0001, "loss": 0.9954, "loss/crossentropy": 2.4987642765045166, "loss/hidden": 0.6328125, "loss/logits": 0.11996030062437057, "loss/reg": 0.0242580845952034, "step": 372 }, { "epoch": 0.046625, "grad_norm": 3.7915477752685547, "grad_norm_var": 2.5549678839666425, "learning_rate": 0.0001, "loss": 0.9193, "loss/crossentropy": 2.3996403217315674, "loss/hidden": 0.578125, "loss/logits": 0.09870465099811554, "loss/reg": 0.02424856275320053, "step": 373 }, { "epoch": 0.04675, "grad_norm": 2.833364725112915, "grad_norm_var": 2.7025530371611066, "learning_rate": 0.0001, "loss": 0.9417, "loss/crossentropy": 2.336583137512207, "loss/hidden": 0.59765625, "loss/logits": 0.10166990756988525, "loss/reg": 0.024240419268608093, "step": 374 }, { "epoch": 0.046875, "grad_norm": 5.302695274353027, "grad_norm_var": 1.359315799583595, "learning_rate": 0.0001, "loss": 0.9495, "loss/crossentropy": 2.5733697414398193, "loss/hidden": 0.578125, "loss/logits": 0.1290503740310669, "loss/reg": 0.0242319293320179, "step": 375 }, { "epoch": 0.047, "grad_norm": 5.087683200836182, "grad_norm_var": 1.426096117003745, "learning_rate": 0.0001, "loss": 0.9067, "loss/crossentropy": 2.548323631286621, "loss/hidden": 0.56640625, "loss/logits": 0.09806410223245621, "loss/reg": 0.02422359585762024, "step": 376 }, { "epoch": 0.047125, "grad_norm": 2.840883255004883, "grad_norm_var": 1.4948607984557458, "learning_rate": 0.0001, "loss": 1.0234, "loss/crossentropy": 2.484494209289551, "loss/hidden": 0.65234375, "loss/logits": 0.12894591689109802, "loss/reg": 0.024214565753936768, "step": 377 }, { "epoch": 0.04725, "grad_norm": 6.647733688354492, "grad_norm_var": 1.848030946106532, "learning_rate": 0.0001, "loss": 1.131, "loss/crossentropy": 2.832048177719116, "loss/hidden": 0.7421875, "loss/logits": 0.14676988124847412, "loss/reg": 0.024205682799220085, "step": 378 }, { "epoch": 0.047375, "grad_norm": 3.465564727783203, "grad_norm_var": 1.8906396391038638, "learning_rate": 0.0001, "loss": 1.0051, "loss/crossentropy": 2.5129964351654053, "loss/hidden": 0.62890625, "loss/logits": 0.134174644947052, "loss/reg": 0.024197354912757874, "step": 379 }, { "epoch": 0.0475, "grad_norm": 3.6985855102539062, "grad_norm_var": 1.8891513099755568, "learning_rate": 0.0001, "loss": 1.4422, "loss/crossentropy": 2.2384328842163086, "loss/hidden": 1.0, "loss/logits": 0.20027095079421997, "loss/reg": 0.02418883889913559, "step": 380 }, { "epoch": 0.047625, "grad_norm": 3.4343388080596924, "grad_norm_var": 1.8993141107729303, "learning_rate": 0.0001, "loss": 0.8648, "loss/crossentropy": 2.753706932067871, "loss/hidden": 0.5234375, "loss/logits": 0.09959565848112106, "loss/reg": 0.024180689826607704, "step": 381 }, { "epoch": 0.04775, "grad_norm": 3.418168067932129, "grad_norm_var": 1.6566847859375398, "learning_rate": 0.0001, "loss": 1.0532, "loss/crossentropy": 2.1205835342407227, "loss/hidden": 0.69921875, "loss/logits": 0.11221310496330261, "loss/reg": 0.02417258359491825, "step": 382 }, { "epoch": 0.047875, "grad_norm": 3.3881125450134277, "grad_norm_var": 1.093930487598201, "learning_rate": 0.0001, "loss": 0.9968, "loss/crossentropy": 2.540379285812378, "loss/hidden": 0.62890625, "loss/logits": 0.1262589991092682, "loss/reg": 0.024164721369743347, "step": 383 }, { "epoch": 0.048, "grad_norm": 3.142486095428467, "grad_norm_var": 1.1231981378319982, "learning_rate": 0.0001, "loss": 0.9186, "loss/crossentropy": 2.5265657901763916, "loss/hidden": 0.58203125, "loss/logits": 0.0949624702334404, "loss/reg": 0.02415630966424942, "step": 384 }, { "epoch": 0.048125, "grad_norm": 7.363763332366943, "grad_norm_var": 1.8443340238906671, "learning_rate": 0.0001, "loss": 1.1703, "loss/crossentropy": 2.442409038543701, "loss/hidden": 0.8046875, "loss/logits": 0.12408800423145294, "loss/reg": 0.024148130789399147, "step": 385 }, { "epoch": 0.04825, "grad_norm": 3.801536798477173, "grad_norm_var": 1.7879312710551798, "learning_rate": 0.0001, "loss": 1.0129, "loss/crossentropy": 2.461942434310913, "loss/hidden": 0.640625, "loss/logits": 0.13092291355133057, "loss/reg": 0.024139659479260445, "step": 386 }, { "epoch": 0.048375, "grad_norm": 4.24082612991333, "grad_norm_var": 1.7228677295443293, "learning_rate": 0.0001, "loss": 1.1161, "loss/crossentropy": 2.2889018058776855, "loss/hidden": 0.74609375, "loss/logits": 0.12870003283023834, "loss/reg": 0.024130841717123985, "step": 387 }, { "epoch": 0.0485, "grad_norm": 4.190494060516357, "grad_norm_var": 1.7195812284180172, "learning_rate": 0.0001, "loss": 1.1649, "loss/crossentropy": 2.4475483894348145, "loss/hidden": 0.76171875, "loss/logits": 0.16198021173477173, "loss/reg": 0.024122456088662148, "step": 388 }, { "epoch": 0.048625, "grad_norm": 15.206843376159668, "grad_norm_var": 9.294742605150057, "learning_rate": 0.0001, "loss": 1.0606, "loss/crossentropy": 2.3164589405059814, "loss/hidden": 0.703125, "loss/logits": 0.116313636302948, "loss/reg": 0.024113710969686508, "step": 389 }, { "epoch": 0.04875, "grad_norm": 10.421854972839355, "grad_norm_var": 10.824103712955843, "learning_rate": 0.0001, "loss": 1.0638, "loss/crossentropy": 2.1764075756073, "loss/hidden": 0.70703125, "loss/logits": 0.11567908525466919, "loss/reg": 0.024104835465550423, "step": 390 }, { "epoch": 0.048875, "grad_norm": 4.923640727996826, "grad_norm_var": 10.83563756748113, "learning_rate": 0.0001, "loss": 0.9465, "loss/crossentropy": 2.6762068271636963, "loss/hidden": 0.6015625, "loss/logits": 0.10399520397186279, "loss/reg": 0.02409605123102665, "step": 391 }, { "epoch": 0.049, "grad_norm": 4.840577125549316, "grad_norm_var": 10.847422220224528, "learning_rate": 0.0001, "loss": 1.2037, "loss/crossentropy": 2.482144832611084, "loss/hidden": 0.8125, "loss/logits": 0.15031081438064575, "loss/reg": 0.02408732660114765, "step": 392 }, { "epoch": 0.049125, "grad_norm": 2.9535470008850098, "grad_norm_var": 10.811063470934846, "learning_rate": 0.0001, "loss": 1.0811, "loss/crossentropy": 2.5998470783233643, "loss/hidden": 0.71484375, "loss/logits": 0.12545132637023926, "loss/reg": 0.024078134447336197, "step": 393 }, { "epoch": 0.04925, "grad_norm": 4.08555793762207, "grad_norm_var": 10.768160950066, "learning_rate": 0.0001, "loss": 1.3211, "loss/crossentropy": 2.611746072769165, "loss/hidden": 0.90625, "loss/logits": 0.17412351071834564, "loss/reg": 0.02406897209584713, "step": 394 }, { "epoch": 0.049375, "grad_norm": 4.730119705200195, "grad_norm_var": 10.582242923890295, "learning_rate": 0.0001, "loss": 0.9463, "loss/crossentropy": 2.5379066467285156, "loss/hidden": 0.59765625, "loss/logits": 0.10806328058242798, "loss/reg": 0.024059420451521873, "step": 395 }, { "epoch": 0.0495, "grad_norm": 4.012064456939697, "grad_norm_var": 10.523956759484621, "learning_rate": 0.0001, "loss": 0.9181, "loss/crossentropy": 2.567375898361206, "loss/hidden": 0.57421875, "loss/logits": 0.10337453335523605, "loss/reg": 0.024049852043390274, "step": 396 }, { "epoch": 0.049625, "grad_norm": 4.57706880569458, "grad_norm_var": 10.32746400090784, "learning_rate": 0.0001, "loss": 1.051, "loss/crossentropy": 2.6955151557922363, "loss/hidden": 0.6875, "loss/logits": 0.12309969961643219, "loss/reg": 0.02403969317674637, "step": 397 }, { "epoch": 0.04975, "grad_norm": 2.7931346893310547, "grad_norm_var": 10.511295288820635, "learning_rate": 0.0001, "loss": 1.0373, "loss/crossentropy": 2.34134578704834, "loss/hidden": 0.66796875, "loss/logits": 0.12901151180267334, "loss/reg": 0.02403116784989834, "step": 398 }, { "epoch": 0.049875, "grad_norm": 2.4789700508117676, "grad_norm_var": 10.793738555266915, "learning_rate": 0.0001, "loss": 1.1463, "loss/crossentropy": 2.1852893829345703, "loss/hidden": 0.75390625, "loss/logits": 0.15216518938541412, "loss/reg": 0.024022690951824188, "step": 399 }, { "epoch": 0.05, "grad_norm": 5.340605735778809, "grad_norm_var": 10.482396698239997, "learning_rate": 0.0001, "loss": 1.2874, "loss/crossentropy": 2.4223577976226807, "loss/hidden": 0.91015625, "loss/logits": 0.13714072108268738, "loss/reg": 0.024014031514525414, "step": 400 }, { "epoch": 0.050125, "grad_norm": 5.600348949432373, "grad_norm_var": 10.208567826877847, "learning_rate": 0.0001, "loss": 0.9905, "loss/crossentropy": 2.494180917739868, "loss/hidden": 0.64453125, "loss/logits": 0.10594967007637024, "loss/reg": 0.024006787687540054, "step": 401 }, { "epoch": 0.05025, "grad_norm": 5.094600677490234, "grad_norm_var": 10.061216488426146, "learning_rate": 0.0001, "loss": 1.2828, "loss/crossentropy": 2.1899051666259766, "loss/hidden": 0.89453125, "loss/logits": 0.1482730507850647, "loss/reg": 0.02399739809334278, "step": 402 }, { "epoch": 0.050375, "grad_norm": 3.915607452392578, "grad_norm_var": 10.115626051260246, "learning_rate": 0.0001, "loss": 1.0375, "loss/crossentropy": 2.790240526199341, "loss/hidden": 0.65625, "loss/logits": 0.14136558771133423, "loss/reg": 0.02398892305791378, "step": 403 }, { "epoch": 0.0505, "grad_norm": 70.49039459228516, "grad_norm_var": 274.8357269833382, "learning_rate": 0.0001, "loss": 1.1534, "loss/crossentropy": 2.2490358352661133, "loss/hidden": 0.79296875, "loss/logits": 0.12061962485313416, "loss/reg": 0.02398114837706089, "step": 404 }, { "epoch": 0.050625, "grad_norm": 6.5985307693481445, "grad_norm_var": 272.87861181728414, "learning_rate": 0.0001, "loss": 1.1479, "loss/crossentropy": 2.856632709503174, "loss/hidden": 0.78515625, "loss/logits": 0.12297768890857697, "loss/reg": 0.0239717997610569, "step": 405 }, { "epoch": 0.05075, "grad_norm": 4.041878700256348, "grad_norm_var": 274.1523084795167, "learning_rate": 0.0001, "loss": 1.1408, "loss/crossentropy": 2.5744457244873047, "loss/hidden": 0.76171875, "loss/logits": 0.13939061760902405, "loss/reg": 0.02396412193775177, "step": 406 }, { "epoch": 0.050875, "grad_norm": 3.5284969806671143, "grad_norm_var": 274.94477307618513, "learning_rate": 0.0001, "loss": 1.0838, "loss/crossentropy": 2.7705729007720947, "loss/hidden": 0.72265625, "loss/logits": 0.12160193920135498, "loss/reg": 0.023956267163157463, "step": 407 }, { "epoch": 0.051, "grad_norm": 3.867558240890503, "grad_norm_var": 275.4712566581114, "learning_rate": 0.0001, "loss": 1.1238, "loss/crossentropy": 2.2807302474975586, "loss/hidden": 0.75390625, "loss/logits": 0.1304117739200592, "loss/reg": 0.023948216810822487, "step": 408 }, { "epoch": 0.051125, "grad_norm": 4.9265217781066895, "grad_norm_var": 274.2865770164501, "learning_rate": 0.0001, "loss": 0.955, "loss/crossentropy": 2.6176843643188477, "loss/hidden": 0.62890625, "loss/logits": 0.08674542605876923, "loss/reg": 0.02393944188952446, "step": 409 }, { "epoch": 0.05125, "grad_norm": 3.1925463676452637, "grad_norm_var": 274.86264478448203, "learning_rate": 0.0001, "loss": 1.0329, "loss/crossentropy": 2.790286064147949, "loss/hidden": 0.6796875, "loss/logits": 0.11385629326105118, "loss/reg": 0.02393159456551075, "step": 410 }, { "epoch": 0.051375, "grad_norm": 2.6993377208709717, "grad_norm_var": 276.12743945534294, "learning_rate": 0.0001, "loss": 0.9268, "loss/crossentropy": 2.5643763542175293, "loss/hidden": 0.58203125, "loss/logits": 0.10557639598846436, "loss/reg": 0.0239238403737545, "step": 411 }, { "epoch": 0.0515, "grad_norm": 4.7841315269470215, "grad_norm_var": 275.72098389943244, "learning_rate": 0.0001, "loss": 1.3399, "loss/crossentropy": 2.3737518787384033, "loss/hidden": 0.89453125, "loss/logits": 0.2062419056892395, "loss/reg": 0.023915138095617294, "step": 412 }, { "epoch": 0.051625, "grad_norm": 2.8743667602539062, "grad_norm_var": 276.7634192046356, "learning_rate": 0.0001, "loss": 1.105, "loss/crossentropy": 2.225154399871826, "loss/hidden": 0.734375, "loss/logits": 0.13158033788204193, "loss/reg": 0.023907041177153587, "step": 413 }, { "epoch": 0.05175, "grad_norm": 4.873403072357178, "grad_norm_var": 275.51638736026473, "learning_rate": 0.0001, "loss": 1.1328, "loss/crossentropy": 2.444675922393799, "loss/hidden": 0.734375, "loss/logits": 0.15947584807872772, "loss/reg": 0.023899447172880173, "step": 414 }, { "epoch": 0.051875, "grad_norm": 3.6676955223083496, "grad_norm_var": 274.6671585398764, "learning_rate": 0.0001, "loss": 0.9504, "loss/crossentropy": 2.6828341484069824, "loss/hidden": 0.6015625, "loss/logits": 0.10996139049530029, "loss/reg": 0.023892199620604515, "step": 415 }, { "epoch": 0.052, "grad_norm": 2.703623056411743, "grad_norm_var": 276.2015243387772, "learning_rate": 0.0001, "loss": 1.049, "loss/crossentropy": 2.4529809951782227, "loss/hidden": 0.6796875, "loss/logits": 0.1305149495601654, "loss/reg": 0.023883724585175514, "step": 416 }, { "epoch": 0.052125, "grad_norm": 2.9288852214813232, "grad_norm_var": 277.61048629826035, "learning_rate": 0.0001, "loss": 1.0151, "loss/crossentropy": 2.6102705001831055, "loss/hidden": 0.640625, "loss/logits": 0.13567429780960083, "loss/reg": 0.023875238373875618, "step": 417 }, { "epoch": 0.05225, "grad_norm": 2.5006144046783447, "grad_norm_var": 279.0831974622093, "learning_rate": 0.0001, "loss": 1.0571, "loss/crossentropy": 2.457202672958374, "loss/hidden": 0.69140625, "loss/logits": 0.12700514495372772, "loss/reg": 0.023867420852184296, "step": 418 }, { "epoch": 0.052375, "grad_norm": 5.180849552154541, "grad_norm_var": 278.49850212580674, "learning_rate": 0.0001, "loss": 1.1214, "loss/crossentropy": 2.7064497470855713, "loss/hidden": 0.74609375, "loss/logits": 0.13676053285598755, "loss/reg": 0.02385888434946537, "step": 419 }, { "epoch": 0.0525, "grad_norm": 3.2954249382019043, "grad_norm_var": 1.3051375489769337, "learning_rate": 0.0001, "loss": 1.0717, "loss/crossentropy": 2.31532621383667, "loss/hidden": 0.703125, "loss/logits": 0.13004590570926666, "loss/reg": 0.023851698264479637, "step": 420 }, { "epoch": 0.052625, "grad_norm": 2.8389148712158203, "grad_norm_var": 0.8127685868335741, "learning_rate": 0.0001, "loss": 1.0781, "loss/crossentropy": 2.3562088012695312, "loss/hidden": 0.70703125, "loss/logits": 0.13266143202781677, "loss/reg": 0.023843195289373398, "step": 421 }, { "epoch": 0.05275, "grad_norm": 4.581103324890137, "grad_norm_var": 0.8613437167520175, "learning_rate": 0.0001, "loss": 1.4299, "loss/crossentropy": 2.1004514694213867, "loss/hidden": 1.0546875, "loss/logits": 0.13683074712753296, "loss/reg": 0.023835282772779465, "step": 422 }, { "epoch": 0.052875, "grad_norm": 4.119724750518799, "grad_norm_var": 0.8733982923945914, "learning_rate": 0.0001, "loss": 1.2152, "loss/crossentropy": 2.640662670135498, "loss/hidden": 0.8515625, "loss/logits": 0.12537327408790588, "loss/reg": 0.02382684126496315, "step": 423 }, { "epoch": 0.053, "grad_norm": 4.544858455657959, "grad_norm_var": 0.918133871994677, "learning_rate": 0.0001, "loss": 1.0957, "loss/crossentropy": 2.8237485885620117, "loss/hidden": 0.7265625, "loss/logits": 0.1309557855129242, "loss/reg": 0.023818302899599075, "step": 424 }, { "epoch": 0.053125, "grad_norm": 2.757870674133301, "grad_norm_var": 0.8666742418813403, "learning_rate": 0.0001, "loss": 0.962, "loss/crossentropy": 2.5533342361450195, "loss/hidden": 0.62109375, "loss/logits": 0.10283903032541275, "loss/reg": 0.023810207843780518, "step": 425 }, { "epoch": 0.05325, "grad_norm": 2.8035690784454346, "grad_norm_var": 0.8970790990362765, "learning_rate": 0.0001, "loss": 1.0877, "loss/crossentropy": 2.610870361328125, "loss/hidden": 0.6953125, "loss/logits": 0.1544169783592224, "loss/reg": 0.02380150742828846, "step": 426 }, { "epoch": 0.053375, "grad_norm": 3.5559544563293457, "grad_norm_var": 0.8432525593756196, "learning_rate": 0.0001, "loss": 1.1169, "loss/crossentropy": 2.537252187728882, "loss/hidden": 0.74609375, "loss/logits": 0.1328693926334381, "loss/reg": 0.02379263937473297, "step": 427 }, { "epoch": 0.0535, "grad_norm": 3.58505916595459, "grad_norm_var": 0.7479056021171611, "learning_rate": 0.0001, "loss": 1.0427, "loss/crossentropy": 2.240241289138794, "loss/hidden": 0.68359375, "loss/logits": 0.1212427169084549, "loss/reg": 0.02378367818892002, "step": 428 }, { "epoch": 0.053625, "grad_norm": 3.6646056175231934, "grad_norm_var": 0.7156687449512967, "learning_rate": 0.0001, "loss": 1.0166, "loss/crossentropy": 2.765550374984741, "loss/hidden": 0.6328125, "loss/logits": 0.14600837230682373, "loss/reg": 0.023774517700076103, "step": 429 }, { "epoch": 0.05375, "grad_norm": 3.2596421241760254, "grad_norm_var": 0.6044660126437359, "learning_rate": 0.0001, "loss": 1.13, "loss/crossentropy": 2.1970372200012207, "loss/hidden": 0.7578125, "loss/logits": 0.13451042771339417, "loss/reg": 0.023765094578266144, "step": 430 }, { "epoch": 0.053875, "grad_norm": 4.039770603179932, "grad_norm_var": 0.6214738630237046, "learning_rate": 0.0001, "loss": 1.2708, "loss/crossentropy": 2.5254645347595215, "loss/hidden": 0.83984375, "loss/logits": 0.19336232542991638, "loss/reg": 0.023756500333547592, "step": 431 }, { "epoch": 0.054, "grad_norm": 2.9176530838012695, "grad_norm_var": 0.6009675102137348, "learning_rate": 0.0001, "loss": 0.9399, "loss/crossentropy": 2.611178398132324, "loss/hidden": 0.59375, "loss/logits": 0.10868553817272186, "loss/reg": 0.023747922852635384, "step": 432 }, { "epoch": 0.054125, "grad_norm": 3.540189027786255, "grad_norm_var": 0.5748467113480954, "learning_rate": 0.0001, "loss": 1.164, "loss/crossentropy": 2.4766628742218018, "loss/hidden": 0.80078125, "loss/logits": 0.1258353739976883, "loss/reg": 0.023739352822303772, "step": 433 }, { "epoch": 0.05425, "grad_norm": 3.7053966522216797, "grad_norm_var": 0.4931212433281331, "learning_rate": 0.0001, "loss": 1.0522, "loss/crossentropy": 2.491478681564331, "loss/hidden": 0.67578125, "loss/logits": 0.13909485936164856, "loss/reg": 0.023731039837002754, "step": 434 }, { "epoch": 0.054375, "grad_norm": 3.4101648330688477, "grad_norm_var": 0.32751985750053336, "learning_rate": 0.0001, "loss": 0.9527, "loss/crossentropy": 2.5495922565460205, "loss/hidden": 0.59765625, "loss/logits": 0.1178436130285263, "loss/reg": 0.023722674697637558, "step": 435 }, { "epoch": 0.0545, "grad_norm": 3.923790693283081, "grad_norm_var": 0.33181180777142816, "learning_rate": 0.0001, "loss": 1.1217, "loss/crossentropy": 2.350161552429199, "loss/hidden": 0.75, "loss/logits": 0.13453412055969238, "loss/reg": 0.02371453307569027, "step": 436 }, { "epoch": 0.054625, "grad_norm": 3.2052547931671143, "grad_norm_var": 0.3040979482718351, "learning_rate": 0.0001, "loss": 1.0641, "loss/crossentropy": 2.5055789947509766, "loss/hidden": 0.71484375, "loss/logits": 0.11222882568836212, "loss/reg": 0.023706616833806038, "step": 437 }, { "epoch": 0.05475, "grad_norm": 2.820063591003418, "grad_norm_var": 0.26777286633351405, "learning_rate": 0.0001, "loss": 1.1811, "loss/crossentropy": 2.535069704055786, "loss/hidden": 0.78125, "loss/logits": 0.1628357172012329, "loss/reg": 0.023698095232248306, "step": 438 }, { "epoch": 0.054875, "grad_norm": 3.095841884613037, "grad_norm_var": 0.24744105333314317, "learning_rate": 0.0001, "loss": 1.1936, "loss/crossentropy": 2.313276767730713, "loss/hidden": 0.78125, "loss/logits": 0.17543524503707886, "loss/reg": 0.02368931844830513, "step": 439 }, { "epoch": 0.055, "grad_norm": 2.955540895462036, "grad_norm_var": 0.1683967569747227, "learning_rate": 0.0001, "loss": 0.991, "loss/crossentropy": 2.747128486633301, "loss/hidden": 0.6484375, "loss/logits": 0.10580303519964218, "loss/reg": 0.023680580779910088, "step": 440 }, { "epoch": 0.055125, "grad_norm": 2.866759777069092, "grad_norm_var": 0.16086728592038804, "learning_rate": 0.0001, "loss": 0.9014, "loss/crossentropy": 2.4465224742889404, "loss/hidden": 0.5703125, "loss/logits": 0.09433356672525406, "loss/reg": 0.023672088980674744, "step": 441 }, { "epoch": 0.05525, "grad_norm": 3.1793012619018555, "grad_norm_var": 0.14310091597801508, "learning_rate": 0.0001, "loss": 1.0742, "loss/crossentropy": 2.364579677581787, "loss/hidden": 0.6875, "loss/logits": 0.15003816783428192, "loss/reg": 0.023663459345698357, "step": 442 }, { "epoch": 0.055375, "grad_norm": 4.3040900230407715, "grad_norm_var": 0.19784760386154687, "learning_rate": 0.0001, "loss": 1.1406, "loss/crossentropy": 2.955233573913574, "loss/hidden": 0.74609375, "loss/logits": 0.1579587459564209, "loss/reg": 0.023654496297240257, "step": 443 }, { "epoch": 0.0555, "grad_norm": 3.6495676040649414, "grad_norm_var": 0.19966009525053988, "learning_rate": 0.0001, "loss": 1.0678, "loss/crossentropy": 2.184626340866089, "loss/hidden": 0.71484375, "loss/logits": 0.11654888093471527, "loss/reg": 0.023645464330911636, "step": 444 }, { "epoch": 0.055625, "grad_norm": 2.7992637157440186, "grad_norm_var": 0.21692371557562992, "learning_rate": 0.0001, "loss": 1.0746, "loss/crossentropy": 1.9573417901992798, "loss/hidden": 0.71484375, "loss/logits": 0.1234317272901535, "loss/reg": 0.023636594414711, "step": 445 }, { "epoch": 0.05575, "grad_norm": 3.4131267070770264, "grad_norm_var": 0.21645445922388262, "learning_rate": 0.0001, "loss": 1.0938, "loss/crossentropy": 2.3615517616271973, "loss/hidden": 0.71875, "loss/logits": 0.1387380063533783, "loss/reg": 0.023627731949090958, "step": 446 }, { "epoch": 0.055875, "grad_norm": 2.890436887741089, "grad_norm_var": 0.19547383544341201, "learning_rate": 0.0001, "loss": 0.8903, "loss/crossentropy": 2.2940785884857178, "loss/hidden": 0.5546875, "loss/logits": 0.09945578873157501, "loss/reg": 0.023619333282113075, "step": 447 }, { "epoch": 0.056, "grad_norm": 3.9202187061309814, "grad_norm_var": 0.20821686288428035, "learning_rate": 0.0001, "loss": 1.1124, "loss/crossentropy": 2.561680316925049, "loss/hidden": 0.73828125, "loss/logits": 0.13801740109920502, "loss/reg": 0.023610329255461693, "step": 448 }, { "epoch": 0.056125, "grad_norm": 3.0585479736328125, "grad_norm_var": 0.21081889060941586, "learning_rate": 0.0001, "loss": 0.9545, "loss/crossentropy": 2.4095382690429688, "loss/hidden": 0.609375, "loss/logits": 0.10911431908607483, "loss/reg": 0.02360081672668457, "step": 449 }, { "epoch": 0.05625, "grad_norm": 3.3532161712646484, "grad_norm_var": 0.2007006666523369, "learning_rate": 0.0001, "loss": 1.1045, "loss/crossentropy": 1.8727831840515137, "loss/hidden": 0.7421875, "loss/logits": 0.1264159381389618, "loss/reg": 0.023592744022607803, "step": 450 }, { "epoch": 0.056375, "grad_norm": 3.1456074714660645, "grad_norm_var": 0.20128870800301848, "learning_rate": 0.0001, "loss": 0.9574, "loss/crossentropy": 2.4792492389678955, "loss/hidden": 0.609375, "loss/logits": 0.11215440928936005, "loss/reg": 0.02358343079686165, "step": 451 }, { "epoch": 0.0565, "grad_norm": 3.402637004852295, "grad_norm_var": 0.1739656178124496, "learning_rate": 0.0001, "loss": 1.0743, "loss/crossentropy": 2.3384110927581787, "loss/hidden": 0.6875, "loss/logits": 0.15106429159641266, "loss/reg": 0.02357417158782482, "step": 452 }, { "epoch": 0.056625, "grad_norm": 3.259817361831665, "grad_norm_var": 0.17379912081048493, "learning_rate": 0.0001, "loss": 1.1221, "loss/crossentropy": 2.456613779067993, "loss/hidden": 0.734375, "loss/logits": 0.15212517976760864, "loss/reg": 0.023564757779240608, "step": 453 }, { "epoch": 0.05675, "grad_norm": 2.531987190246582, "grad_norm_var": 0.195773570863138, "learning_rate": 0.0001, "loss": 0.9605, "loss/crossentropy": 2.394343376159668, "loss/hidden": 0.609375, "loss/logits": 0.11555634438991547, "loss/reg": 0.023556271567940712, "step": 454 }, { "epoch": 0.056875, "grad_norm": 2.594336986541748, "grad_norm_var": 0.22107356191806862, "learning_rate": 0.0001, "loss": 0.9207, "loss/crossentropy": 2.781731367111206, "loss/hidden": 0.58203125, "loss/logits": 0.10317155718803406, "loss/reg": 0.023547139018774033, "step": 455 }, { "epoch": 0.057, "grad_norm": 3.1260924339294434, "grad_norm_var": 0.21715561662650557, "learning_rate": 0.0001, "loss": 1.1637, "loss/crossentropy": 2.5018725395202637, "loss/hidden": 0.78125, "loss/logits": 0.14706720411777496, "loss/reg": 0.02353852428495884, "step": 456 }, { "epoch": 0.057125, "grad_norm": 3.159911632537842, "grad_norm_var": 0.20878072756432645, "learning_rate": 0.0001, "loss": 1.0676, "loss/crossentropy": 2.668788433074951, "loss/hidden": 0.68359375, "loss/logits": 0.1487644910812378, "loss/reg": 0.02352879010140896, "step": 457 }, { "epoch": 0.05725, "grad_norm": 3.0937418937683105, "grad_norm_var": 0.20989373673105333, "learning_rate": 0.0001, "loss": 1.0188, "loss/crossentropy": 2.498286008834839, "loss/hidden": 0.6640625, "loss/logits": 0.11953231692314148, "loss/reg": 0.023519227281212807, "step": 458 }, { "epoch": 0.057375, "grad_norm": 2.4465949535369873, "grad_norm_var": 0.15987096754079838, "learning_rate": 0.0001, "loss": 1.0333, "loss/crossentropy": 2.401630401611328, "loss/hidden": 0.671875, "loss/logits": 0.12631601095199585, "loss/reg": 0.023509083315730095, "step": 459 }, { "epoch": 0.0575, "grad_norm": 3.8185830116271973, "grad_norm_var": 0.17369585396981316, "learning_rate": 0.0001, "loss": 1.2185, "loss/crossentropy": 2.3343100547790527, "loss/hidden": 0.84375, "loss/logits": 0.13973468542099, "loss/reg": 0.023500461131334305, "step": 460 }, { "epoch": 0.057625, "grad_norm": 2.769894599914551, "grad_norm_var": 0.1750287637092998, "learning_rate": 0.0001, "loss": 1.0478, "loss/crossentropy": 2.594421148300171, "loss/hidden": 0.66796875, "loss/logits": 0.14495806396007538, "loss/reg": 0.02349086105823517, "step": 461 }, { "epoch": 0.05775, "grad_norm": 3.924386501312256, "grad_norm_var": 0.21107140664515026, "learning_rate": 0.0001, "loss": 1.1201, "loss/crossentropy": 1.8543033599853516, "loss/hidden": 0.71875, "loss/logits": 0.16654378175735474, "loss/reg": 0.023482073098421097, "step": 462 }, { "epoch": 0.057875, "grad_norm": 2.719325304031372, "grad_norm_var": 0.21896016035892957, "learning_rate": 0.0001, "loss": 0.9973, "loss/crossentropy": 2.7110376358032227, "loss/hidden": 0.6328125, "loss/logits": 0.1297917366027832, "loss/reg": 0.02347267046570778, "step": 463 }, { "epoch": 0.058, "grad_norm": 3.110532522201538, "grad_norm_var": 0.17627651595159174, "learning_rate": 0.0001, "loss": 0.9503, "loss/crossentropy": 2.424137830734253, "loss/hidden": 0.61328125, "loss/logits": 0.10241679847240448, "loss/reg": 0.023463333025574684, "step": 464 }, { "epoch": 0.058125, "grad_norm": 3.2945988178253174, "grad_norm_var": 0.17862116157392785, "learning_rate": 0.0001, "loss": 1.1634, "loss/crossentropy": 2.8107378482818604, "loss/hidden": 0.75, "loss/logits": 0.17888996005058289, "loss/reg": 0.0234534852206707, "step": 465 }, { "epoch": 0.05825, "grad_norm": 4.864523887634277, "grad_norm_var": 0.37049430510921844, "learning_rate": 0.0001, "loss": 1.0812, "loss/crossentropy": 2.5031394958496094, "loss/hidden": 0.71484375, "loss/logits": 0.13192051649093628, "loss/reg": 0.023444540798664093, "step": 466 }, { "epoch": 0.058375, "grad_norm": 3.8722984790802, "grad_norm_var": 0.3978501673810001, "learning_rate": 0.0001, "loss": 1.2316, "loss/crossentropy": 1.9772007465362549, "loss/hidden": 0.8515625, "loss/logits": 0.14563970267772675, "loss/reg": 0.023435747250914574, "step": 467 }, { "epoch": 0.0585, "grad_norm": 4.621346473693848, "grad_norm_var": 0.5155902021721573, "learning_rate": 0.0001, "loss": 1.1725, "loss/crossentropy": 2.4977753162384033, "loss/hidden": 0.7890625, "loss/logits": 0.149122953414917, "loss/reg": 0.023427119478583336, "step": 468 }, { "epoch": 0.058625, "grad_norm": 3.36370849609375, "grad_norm_var": 0.5153549660191058, "learning_rate": 0.0001, "loss": 1.2411, "loss/crossentropy": 2.4750683307647705, "loss/hidden": 0.8203125, "loss/logits": 0.18663738667964935, "loss/reg": 0.023417862132191658, "step": 469 }, { "epoch": 0.05875, "grad_norm": 3.391871690750122, "grad_norm_var": 0.46984604899865395, "learning_rate": 0.0001, "loss": 1.1066, "loss/crossentropy": 2.5189239978790283, "loss/hidden": 0.71875, "loss/logits": 0.1537722945213318, "loss/reg": 0.023408619686961174, "step": 470 }, { "epoch": 0.058875, "grad_norm": 3.903122901916504, "grad_norm_var": 0.43880097595690587, "learning_rate": 0.0001, "loss": 0.9258, "loss/crossentropy": 2.548199415206909, "loss/hidden": 0.59375, "loss/logits": 0.09802491217851639, "loss/reg": 0.02339930646121502, "step": 471 }, { "epoch": 0.059, "grad_norm": 5.576291084289551, "grad_norm_var": 0.7024716555345464, "learning_rate": 0.0001, "loss": 1.2843, "loss/crossentropy": 2.682227849960327, "loss/hidden": 0.859375, "loss/logits": 0.19106051325798035, "loss/reg": 0.023390140384435654, "step": 472 }, { "epoch": 0.059125, "grad_norm": 2.4201464653015137, "grad_norm_var": 0.7821220779043936, "learning_rate": 0.0001, "loss": 0.9341, "loss/crossentropy": 2.5251269340515137, "loss/hidden": 0.59375, "loss/logits": 0.10658347606658936, "loss/reg": 0.023381320759654045, "step": 473 }, { "epoch": 0.05925, "grad_norm": 4.325901985168457, "grad_norm_var": 0.7980385459591806, "learning_rate": 0.0001, "loss": 0.9798, "loss/crossentropy": 2.3592865467071533, "loss/hidden": 0.640625, "loss/logits": 0.10548710823059082, "loss/reg": 0.023372096940875053, "step": 474 }, { "epoch": 0.059375, "grad_norm": 3.1382455825805664, "grad_norm_var": 0.7168259193102716, "learning_rate": 0.0001, "loss": 0.9839, "loss/crossentropy": 2.6056907176971436, "loss/hidden": 0.62890625, "loss/logits": 0.121395543217659, "loss/reg": 0.023363398388028145, "step": 475 }, { "epoch": 0.0595, "grad_norm": 3.0779755115509033, "grad_norm_var": 0.7388713721113239, "learning_rate": 0.0001, "loss": 1.0798, "loss/crossentropy": 2.320854663848877, "loss/hidden": 0.72265625, "loss/logits": 0.12356055527925491, "loss/reg": 0.023354284465312958, "step": 476 }, { "epoch": 0.059625, "grad_norm": 3.293567657470703, "grad_norm_var": 0.6946720185858983, "learning_rate": 0.0001, "loss": 0.9685, "loss/crossentropy": 2.7064430713653564, "loss/hidden": 0.625, "loss/logits": 0.11008161306381226, "loss/reg": 0.023345019668340683, "step": 477 }, { "epoch": 0.05975, "grad_norm": 2.664088249206543, "grad_norm_var": 0.7530647477645431, "learning_rate": 0.0001, "loss": 0.9101, "loss/crossentropy": 2.6147005558013916, "loss/hidden": 0.5703125, "loss/logits": 0.10643748193979263, "loss/reg": 0.02333623729646206, "step": 478 }, { "epoch": 0.059875, "grad_norm": 2.9174795150756836, "grad_norm_var": 0.7321888983535926, "learning_rate": 0.0001, "loss": 0.9342, "loss/crossentropy": 2.566849946975708, "loss/hidden": 0.58984375, "loss/logits": 0.11108069121837616, "loss/reg": 0.023327510803937912, "step": 479 }, { "epoch": 0.06, "grad_norm": 3.3596463203430176, "grad_norm_var": 0.71932045702877, "learning_rate": 0.0001, "loss": 1.0961, "loss/crossentropy": 2.645395040512085, "loss/hidden": 0.703125, "loss/logits": 0.15979796648025513, "loss/reg": 0.023318573832511902, "step": 480 }, { "epoch": 0.060125, "grad_norm": 4.0903096199035645, "grad_norm_var": 0.7232764591548666, "learning_rate": 0.0001, "loss": 1.1686, "loss/crossentropy": 2.077467918395996, "loss/hidden": 0.79296875, "loss/logits": 0.14252969622612, "loss/reg": 0.02330981194972992, "step": 481 }, { "epoch": 0.06025, "grad_norm": 3.277656078338623, "grad_norm_var": 0.6300433507978689, "learning_rate": 0.0001, "loss": 1.1136, "loss/crossentropy": 2.588385581970215, "loss/hidden": 0.7421875, "loss/logits": 0.13843819499015808, "loss/reg": 0.02330118976533413, "step": 482 }, { "epoch": 0.060375, "grad_norm": 4.117528915405273, "grad_norm_var": 0.6433314640873874, "learning_rate": 0.0001, "loss": 1.0239, "loss/crossentropy": 2.766176700592041, "loss/hidden": 0.6640625, "loss/logits": 0.12686920166015625, "loss/reg": 0.023292165249586105, "step": 483 }, { "epoch": 0.0605, "grad_norm": 11.29445743560791, "grad_norm_var": 4.338621670503842, "learning_rate": 0.0001, "loss": 1.9614, "loss/crossentropy": 2.5279059410095215, "loss/hidden": 1.2109375, "loss/logits": 0.5176718235015869, "loss/reg": 0.02328311838209629, "step": 484 }, { "epoch": 0.060625, "grad_norm": 3.5054728984832764, "grad_norm_var": 4.327600163307723, "learning_rate": 0.0001, "loss": 1.0719, "loss/crossentropy": 2.726095199584961, "loss/hidden": 0.7109375, "loss/logits": 0.12821289896965027, "loss/reg": 0.023273879662156105, "step": 485 }, { "epoch": 0.06075, "grad_norm": 3.157118558883667, "grad_norm_var": 4.350771203860321, "learning_rate": 0.0001, "loss": 0.9326, "loss/crossentropy": 2.207131862640381, "loss/hidden": 0.60546875, "loss/logits": 0.09451892971992493, "loss/reg": 0.02326469123363495, "step": 486 }, { "epoch": 0.060875, "grad_norm": 5.006563186645508, "grad_norm_var": 4.411522578027558, "learning_rate": 0.0001, "loss": 1.1884, "loss/crossentropy": 3.005479335784912, "loss/hidden": 0.78515625, "loss/logits": 0.17073442041873932, "loss/reg": 0.02325539104640484, "step": 487 }, { "epoch": 0.061, "grad_norm": 8.253157615661621, "grad_norm_var": 5.3947068177792, "learning_rate": 0.0001, "loss": 1.4073, "loss/crossentropy": 2.329857349395752, "loss/hidden": 0.984375, "loss/logits": 0.19045758247375488, "loss/reg": 0.02324584126472473, "step": 488 }, { "epoch": 0.061125, "grad_norm": 4.795626640319824, "grad_norm_var": 5.169810789054156, "learning_rate": 0.0001, "loss": 1.2274, "loss/crossentropy": 2.354893207550049, "loss/hidden": 0.8515625, "loss/logits": 0.14352190494537354, "loss/reg": 0.023236218839883804, "step": 489 }, { "epoch": 0.06125, "grad_norm": 3.1360483169555664, "grad_norm_var": 5.268809256909969, "learning_rate": 0.0001, "loss": 1.0242, "loss/crossentropy": 2.2810122966766357, "loss/hidden": 0.671875, "loss/logits": 0.12006018310785294, "loss/reg": 0.02322734333574772, "step": 490 }, { "epoch": 0.061375, "grad_norm": 3.931467056274414, "grad_norm_var": 5.1833802842946906, "learning_rate": 0.0001, "loss": 1.1688, "loss/crossentropy": 2.3372247219085693, "loss/hidden": 0.78515625, "loss/logits": 0.15147234499454498, "loss/reg": 0.023217879235744476, "step": 491 }, { "epoch": 0.0615, "grad_norm": 3.8858096599578857, "grad_norm_var": 5.085283642122107, "learning_rate": 0.0001, "loss": 1.0309, "loss/crossentropy": 2.597487688064575, "loss/hidden": 0.6796875, "loss/logits": 0.11914543062448502, "loss/reg": 0.02320869080722332, "step": 492 }, { "epoch": 0.061625, "grad_norm": 6.116000175476074, "grad_norm_var": 5.160062314221837, "learning_rate": 0.0001, "loss": 1.2295, "loss/crossentropy": 2.4041733741760254, "loss/hidden": 0.8515625, "loss/logits": 0.14599129557609558, "loss/reg": 0.023199014365673065, "step": 493 }, { "epoch": 0.06175, "grad_norm": 3.4635026454925537, "grad_norm_var": 4.994267697000357, "learning_rate": 0.0001, "loss": 1.0957, "loss/crossentropy": 2.692230224609375, "loss/hidden": 0.73046875, "loss/logits": 0.13329669833183289, "loss/reg": 0.023189352825284004, "step": 494 }, { "epoch": 0.061875, "grad_norm": 6.833379745483398, "grad_norm_var": 5.051083471594066, "learning_rate": 0.0001, "loss": 1.1437, "loss/crossentropy": 2.470660448074341, "loss/hidden": 0.7890625, "loss/logits": 0.12278926372528076, "loss/reg": 0.023179946467280388, "step": 495 }, { "epoch": 0.062, "grad_norm": 3.2948107719421387, "grad_norm_var": 5.064566926371495, "learning_rate": 0.0001, "loss": 1.109, "loss/crossentropy": 2.3682029247283936, "loss/hidden": 0.7421875, "loss/logits": 0.13507232069969177, "loss/reg": 0.02317013218998909, "step": 496 }, { "epoch": 0.062125, "grad_norm": 4.057919025421143, "grad_norm_var": 5.068064269732227, "learning_rate": 0.0001, "loss": 1.1648, "loss/crossentropy": 2.4153223037719727, "loss/hidden": 0.765625, "loss/logits": 0.16754823923110962, "loss/reg": 0.02316114492714405, "step": 497 }, { "epoch": 0.06225, "grad_norm": 3.3985679149627686, "grad_norm_var": 5.043098814178749, "learning_rate": 0.0001, "loss": 1.1033, "loss/crossentropy": 2.6683197021484375, "loss/hidden": 0.73828125, "loss/logits": 0.13353273272514343, "loss/reg": 0.023151271045207977, "step": 498 }, { "epoch": 0.062375, "grad_norm": 2.6574859619140625, "grad_norm_var": 5.326800856327217, "learning_rate": 0.0001, "loss": 0.9413, "loss/crossentropy": 2.490849733352661, "loss/hidden": 0.60546875, "loss/logits": 0.1043705940246582, "loss/reg": 0.023141290992498398, "step": 499 }, { "epoch": 0.0625, "grad_norm": 3.2627739906311035, "grad_norm_var": 2.402846049322009, "learning_rate": 0.0001, "loss": 1.0233, "loss/crossentropy": 2.3429057598114014, "loss/hidden": 0.63671875, "loss/logits": 0.1552838534116745, "loss/reg": 0.02313125506043434, "step": 500 }, { "epoch": 0.062625, "grad_norm": 3.2453906536102295, "grad_norm_var": 2.4345300369903553, "learning_rate": 0.0001, "loss": 1.1719, "loss/crossentropy": 2.497878313064575, "loss/hidden": 0.76953125, "loss/logits": 0.17116406559944153, "loss/reg": 0.02312229759991169, "step": 501 }, { "epoch": 0.06275, "grad_norm": 4.79340934753418, "grad_norm_var": 2.3566760840149366, "learning_rate": 0.0001, "loss": 1.0001, "loss/crossentropy": 2.360431671142578, "loss/hidden": 0.6484375, "loss/logits": 0.12051868438720703, "loss/reg": 0.02311263047158718, "step": 502 }, { "epoch": 0.062875, "grad_norm": 3.1595826148986816, "grad_norm_var": 2.4163836713766425, "learning_rate": 0.0001, "loss": 0.9158, "loss/crossentropy": 2.266618490219116, "loss/hidden": 0.58203125, "loss/logits": 0.1027822494506836, "loss/reg": 0.023102767765522003, "step": 503 }, { "epoch": 0.063, "grad_norm": 3.59019136428833, "grad_norm_var": 1.2975304557542653, "learning_rate": 0.0001, "loss": 0.9685, "loss/crossentropy": 2.6386334896087646, "loss/hidden": 0.625, "loss/logits": 0.11253425478935242, "loss/reg": 0.023093828931450844, "step": 504 }, { "epoch": 0.063125, "grad_norm": 3.1218326091766357, "grad_norm_var": 1.2897946661694502, "learning_rate": 0.0001, "loss": 0.9493, "loss/crossentropy": 2.2931203842163086, "loss/hidden": 0.6171875, "loss/logits": 0.10128200799226761, "loss/reg": 0.0230838842689991, "step": 505 }, { "epoch": 0.06325, "grad_norm": 7.243019104003906, "grad_norm_var": 1.941121973828988, "learning_rate": 0.0001, "loss": 1.1363, "loss/crossentropy": 2.510519504547119, "loss/hidden": 0.7890625, "loss/logits": 0.11650878190994263, "loss/reg": 0.023074399679899216, "step": 506 }, { "epoch": 0.063375, "grad_norm": 2.7458910942077637, "grad_norm_var": 2.0601092371510408, "learning_rate": 0.0001, "loss": 0.8688, "loss/crossentropy": 2.504798173904419, "loss/hidden": 0.54296875, "loss/logits": 0.09517204016447067, "loss/reg": 0.023064618930220604, "step": 507 }, { "epoch": 0.0635, "grad_norm": 3.834894895553589, "grad_norm_var": 2.061415401484546, "learning_rate": 0.0001, "loss": 1.0614, "loss/crossentropy": 2.504178285598755, "loss/hidden": 0.69140625, "loss/logits": 0.13941214978694916, "loss/reg": 0.023055192083120346, "step": 508 }, { "epoch": 0.063625, "grad_norm": 3.0524418354034424, "grad_norm_var": 1.8045701590720038, "learning_rate": 0.0001, "loss": 0.9395, "loss/crossentropy": 2.6670830249786377, "loss/hidden": 0.6015625, "loss/logits": 0.10751838982105255, "loss/reg": 0.023046277463436127, "step": 509 }, { "epoch": 0.06375, "grad_norm": 2.638979196548462, "grad_norm_var": 1.8906158947457992, "learning_rate": 0.0001, "loss": 0.9649, "loss/crossentropy": 2.5606770515441895, "loss/hidden": 0.62109375, "loss/logits": 0.11341118812561035, "loss/reg": 0.02303677424788475, "step": 510 }, { "epoch": 0.063875, "grad_norm": 4.029105186462402, "grad_norm_var": 1.2509737999906814, "learning_rate": 0.0001, "loss": 1.0378, "loss/crossentropy": 2.446560859680176, "loss/hidden": 0.7109375, "loss/logits": 0.09661944955587387, "loss/reg": 0.023027852177619934, "step": 511 }, { "epoch": 0.064, "grad_norm": 3.203378438949585, "grad_norm_var": 1.255617850639648, "learning_rate": 0.0001, "loss": 0.9705, "loss/crossentropy": 2.08353328704834, "loss/hidden": 0.63671875, "loss/logits": 0.10362571477890015, "loss/reg": 0.02301831543445587, "step": 512 }, { "epoch": 0.064125, "grad_norm": 2.5737931728363037, "grad_norm_var": 1.3080458668089554, "learning_rate": 0.0001, "loss": 1.0597, "loss/crossentropy": 2.3520448207855225, "loss/hidden": 0.7109375, "loss/logits": 0.11864635348320007, "loss/reg": 0.023009376600384712, "step": 513 }, { "epoch": 0.06425, "grad_norm": 3.6731107234954834, "grad_norm_var": 1.307783724921588, "learning_rate": 0.0001, "loss": 1.1013, "loss/crossentropy": 2.77113938331604, "loss/hidden": 0.73046875, "loss/logits": 0.1408485472202301, "loss/reg": 0.022999830543994904, "step": 514 }, { "epoch": 0.064375, "grad_norm": 3.06605863571167, "grad_norm_var": 1.269509965568249, "learning_rate": 0.0001, "loss": 0.9224, "loss/crossentropy": 2.3147072792053223, "loss/hidden": 0.59375, "loss/logits": 0.0987289547920227, "loss/reg": 0.022990131750702858, "step": 515 }, { "epoch": 0.0645, "grad_norm": 3.462446689605713, "grad_norm_var": 1.2636330593023397, "learning_rate": 0.0001, "loss": 1.0201, "loss/crossentropy": 2.4042327404022217, "loss/hidden": 0.66796875, "loss/logits": 0.12231434136629105, "loss/reg": 0.02298046089708805, "step": 516 }, { "epoch": 0.064625, "grad_norm": 4.2029500007629395, "grad_norm_var": 1.2769943636458339, "learning_rate": 0.0001, "loss": 1.222, "loss/crossentropy": 2.514112710952759, "loss/hidden": 0.82421875, "loss/logits": 0.16808617115020752, "loss/reg": 0.022970519959926605, "step": 517 }, { "epoch": 0.06475, "grad_norm": 3.433554172515869, "grad_norm_var": 1.1851525686550586, "learning_rate": 0.0001, "loss": 1.2444, "loss/crossentropy": 2.493224859237671, "loss/hidden": 0.84375, "loss/logits": 0.17107471823692322, "loss/reg": 0.022961357608437538, "step": 518 }, { "epoch": 0.064875, "grad_norm": 4.310100078582764, "grad_norm_var": 1.2057753361074104, "learning_rate": 0.0001, "loss": 1.0108, "loss/crossentropy": 2.5606930255889893, "loss/hidden": 0.67578125, "loss/logits": 0.10546360909938812, "loss/reg": 0.022952331230044365, "step": 519 }, { "epoch": 0.065, "grad_norm": 3.674527883529663, "grad_norm_var": 1.2057007253632908, "learning_rate": 0.0001, "loss": 0.9803, "loss/crossentropy": 2.4196624755859375, "loss/hidden": 0.64453125, "loss/logits": 0.10631287097930908, "loss/reg": 0.02294265851378441, "step": 520 }, { "epoch": 0.065125, "grad_norm": 3.101484775543213, "grad_norm_var": 1.20713683658368, "learning_rate": 0.0001, "loss": 1.2375, "loss/crossentropy": 2.1448891162872314, "loss/hidden": 0.88671875, "loss/logits": 0.12145140767097473, "loss/reg": 0.02293260022997856, "step": 521 }, { "epoch": 0.06525, "grad_norm": 3.2564265727996826, "grad_norm_var": 0.2854656858181736, "learning_rate": 0.0001, "loss": 0.9783, "loss/crossentropy": 2.3986809253692627, "loss/hidden": 0.609375, "loss/logits": 0.1396813988685608, "loss/reg": 0.022922798991203308, "step": 522 }, { "epoch": 0.065375, "grad_norm": 3.3007333278656006, "grad_norm_var": 0.2569672821287893, "learning_rate": 0.0001, "loss": 0.9586, "loss/crossentropy": 2.7955212593078613, "loss/hidden": 0.62109375, "loss/logits": 0.10833179950714111, "loss/reg": 0.02291307970881462, "step": 523 }, { "epoch": 0.0655, "grad_norm": 2.9546499252319336, "grad_norm_var": 0.25738909944094907, "learning_rate": 0.0001, "loss": 1.0937, "loss/crossentropy": 2.409029006958008, "loss/hidden": 0.71875, "loss/logits": 0.14591118693351746, "loss/reg": 0.022904111072421074, "step": 524 }, { "epoch": 0.065625, "grad_norm": 3.2193830013275146, "grad_norm_var": 0.25204334767618, "learning_rate": 0.0001, "loss": 1.1221, "loss/crossentropy": 2.491401195526123, "loss/hidden": 0.75390625, "loss/logits": 0.13920898735523224, "loss/reg": 0.02289445698261261, "step": 525 }, { "epoch": 0.06575, "grad_norm": 40.478694915771484, "grad_norm_var": 85.9971082258447, "learning_rate": 0.0001, "loss": 1.1342, "loss/crossentropy": 2.4311652183532715, "loss/hidden": 0.796875, "loss/logits": 0.10845671594142914, "loss/reg": 0.022885650396347046, "step": 526 }, { "epoch": 0.065875, "grad_norm": 3.2905385494232178, "grad_norm_var": 86.20029999738617, "learning_rate": 0.0001, "loss": 1.0101, "loss/crossentropy": 2.1737422943115234, "loss/hidden": 0.671875, "loss/logits": 0.10944204032421112, "loss/reg": 0.022876843810081482, "step": 527 }, { "epoch": 0.066, "grad_norm": 4.666721343994141, "grad_norm_var": 85.84699165642114, "learning_rate": 0.0001, "loss": 1.2142, "loss/crossentropy": 2.462200880050659, "loss/hidden": 0.86328125, "loss/logits": 0.12223749607801437, "loss/reg": 0.02286742813885212, "step": 528 }, { "epoch": 0.066125, "grad_norm": 3.0347273349761963, "grad_norm_var": 85.66251244998139, "learning_rate": 0.0001, "loss": 1.0285, "loss/crossentropy": 2.38840651512146, "loss/hidden": 0.6640625, "loss/logits": 0.13581448793411255, "loss/reg": 0.022858494892716408, "step": 529 }, { "epoch": 0.06625, "grad_norm": 2.5890092849731445, "grad_norm_var": 86.04634847608627, "learning_rate": 0.0001, "loss": 0.967, "loss/crossentropy": 2.5042731761932373, "loss/hidden": 0.62890625, "loss/logits": 0.10961504280567169, "loss/reg": 0.02284966967999935, "step": 530 }, { "epoch": 0.066375, "grad_norm": 3.3963401317596436, "grad_norm_var": 85.93485657047692, "learning_rate": 0.0001, "loss": 1.072, "loss/crossentropy": 2.7259116172790527, "loss/hidden": 0.7109375, "loss/logits": 0.13268503546714783, "loss/reg": 0.02284088172018528, "step": 531 }, { "epoch": 0.0665, "grad_norm": 3.731293201446533, "grad_norm_var": 85.85653980693046, "learning_rate": 0.0001, "loss": 1.0647, "loss/crossentropy": 2.277968168258667, "loss/hidden": 0.71484375, "loss/logits": 0.12150134146213531, "loss/reg": 0.02283208817243576, "step": 532 }, { "epoch": 0.066625, "grad_norm": 4.581428050994873, "grad_norm_var": 85.78540060231343, "learning_rate": 0.0001, "loss": 1.2486, "loss/crossentropy": 2.067720890045166, "loss/hidden": 0.84375, "loss/logits": 0.17660680413246155, "loss/reg": 0.02282322198152542, "step": 533 }, { "epoch": 0.06675, "grad_norm": 3.0526421070098877, "grad_norm_var": 85.91535378874303, "learning_rate": 0.0001, "loss": 1.1318, "loss/crossentropy": 2.4441521167755127, "loss/hidden": 0.7734375, "loss/logits": 0.13023720681667328, "loss/reg": 0.022814445197582245, "step": 534 }, { "epoch": 0.066875, "grad_norm": 3.4852664470672607, "grad_norm_var": 86.12062292738887, "learning_rate": 0.0001, "loss": 0.9936, "loss/crossentropy": 2.418733596801758, "loss/hidden": 0.66015625, "loss/logits": 0.10543158650398254, "loss/reg": 0.022805610671639442, "step": 535 }, { "epoch": 0.067, "grad_norm": 2.7321274280548096, "grad_norm_var": 86.43545869041343, "learning_rate": 0.0001, "loss": 1.0096, "loss/crossentropy": 2.3275394439697266, "loss/hidden": 0.65625, "loss/logits": 0.12536926567554474, "loss/reg": 0.022797243669629097, "step": 536 }, { "epoch": 0.067125, "grad_norm": 3.029811382293701, "grad_norm_var": 86.46041611877568, "learning_rate": 0.0001, "loss": 1.1049, "loss/crossentropy": 2.2477617263793945, "loss/hidden": 0.74609375, "loss/logits": 0.1309519112110138, "loss/reg": 0.022788099944591522, "step": 537 }, { "epoch": 0.06725, "grad_norm": 2.7345895767211914, "grad_norm_var": 86.64571497988939, "learning_rate": 0.0001, "loss": 1.1272, "loss/crossentropy": 2.4289236068725586, "loss/hidden": 0.7578125, "loss/logits": 0.141631618142128, "loss/reg": 0.022778736427426338, "step": 538 }, { "epoch": 0.067375, "grad_norm": 3.2934482097625732, "grad_norm_var": 86.6479928457627, "learning_rate": 0.0001, "loss": 1.0286, "loss/crossentropy": 2.538973093032837, "loss/hidden": 0.67578125, "loss/logits": 0.12509004771709442, "loss/reg": 0.02277030609548092, "step": 539 }, { "epoch": 0.0675, "grad_norm": 3.833656072616577, "grad_norm_var": 86.38133368804911, "learning_rate": 0.0001, "loss": 0.999, "loss/crossentropy": 2.429593563079834, "loss/hidden": 0.64453125, "loss/logits": 0.12682604789733887, "loss/reg": 0.022761952131986618, "step": 540 }, { "epoch": 0.067625, "grad_norm": 3.544104814529419, "grad_norm_var": 86.28065873545309, "learning_rate": 0.0001, "loss": 1.0002, "loss/crossentropy": 2.166292905807495, "loss/hidden": 0.66015625, "loss/logits": 0.11254848539829254, "loss/reg": 0.02275264821946621, "step": 541 }, { "epoch": 0.06775, "grad_norm": 4.461411952972412, "grad_norm_var": 0.4229304647166086, "learning_rate": 0.0001, "loss": 1.6203, "loss/crossentropy": 2.6347737312316895, "loss/hidden": 1.1328125, "loss/logits": 0.2600440979003906, "loss/reg": 0.022744029760360718, "step": 542 }, { "epoch": 0.067875, "grad_norm": 2.6814959049224854, "grad_norm_var": 0.46036790462302574, "learning_rate": 0.0001, "loss": 0.9923, "loss/crossentropy": 2.580264091491699, "loss/hidden": 0.65234375, "loss/logits": 0.11260214447975159, "loss/reg": 0.022734828293323517, "step": 543 }, { "epoch": 0.068, "grad_norm": 3.685408353805542, "grad_norm_var": 0.35847800648438505, "learning_rate": 0.0001, "loss": 1.1452, "loss/crossentropy": 2.7129862308502197, "loss/hidden": 0.78515625, "loss/logits": 0.13279825448989868, "loss/reg": 0.022726204246282578, "step": 544 }, { "epoch": 0.068125, "grad_norm": 6.349724292755127, "grad_norm_var": 0.8985836730566762, "learning_rate": 0.0001, "loss": 1.1353, "loss/crossentropy": 2.7293214797973633, "loss/hidden": 0.65625, "loss/logits": 0.2518823742866516, "loss/reg": 0.022716930136084557, "step": 545 }, { "epoch": 0.06825, "grad_norm": 3.681774616241455, "grad_norm_var": 0.829722440393675, "learning_rate": 0.0001, "loss": 1.1759, "loss/crossentropy": 2.278223991394043, "loss/hidden": 0.8046875, "loss/logits": 0.1441642791032791, "loss/reg": 0.022707859054207802, "step": 546 }, { "epoch": 0.068375, "grad_norm": 3.84778094291687, "grad_norm_var": 0.8276635905853821, "learning_rate": 0.0001, "loss": 1.3398, "loss/crossentropy": 2.1253304481506348, "loss/hidden": 0.94921875, "loss/logits": 0.16354900598526, "loss/reg": 0.022699227556586266, "step": 547 }, { "epoch": 0.0685, "grad_norm": 5.178676605224609, "grad_norm_var": 0.9703527182714744, "learning_rate": 0.0001, "loss": 1.022, "loss/crossentropy": 2.7832319736480713, "loss/hidden": 0.67578125, "loss/logits": 0.11932960152626038, "loss/reg": 0.02269013226032257, "step": 548 }, { "epoch": 0.068625, "grad_norm": 3.0284383296966553, "grad_norm_var": 0.9511722709094016, "learning_rate": 0.0001, "loss": 1.0412, "loss/crossentropy": 1.9599052667617798, "loss/hidden": 0.703125, "loss/logits": 0.11125050485134125, "loss/reg": 0.022681355476379395, "step": 549 }, { "epoch": 0.06875, "grad_norm": 3.167809247970581, "grad_norm_var": 0.942616955302132, "learning_rate": 0.0001, "loss": 1.0545, "loss/crossentropy": 2.5844783782958984, "loss/hidden": 0.69140625, "loss/logits": 0.13634443283081055, "loss/reg": 0.02267223782837391, "step": 550 }, { "epoch": 0.068875, "grad_norm": 3.2949278354644775, "grad_norm_var": 0.9495941353113788, "learning_rate": 0.0001, "loss": 1.0619, "loss/crossentropy": 2.314173460006714, "loss/hidden": 0.71484375, "loss/logits": 0.12046810984611511, "loss/reg": 0.022662866860628128, "step": 551 }, { "epoch": 0.069, "grad_norm": 2.8322856426239014, "grad_norm_var": 0.9378422714313653, "learning_rate": 0.0001, "loss": 1.0272, "loss/crossentropy": 2.768298387527466, "loss/hidden": 0.66796875, "loss/logits": 0.1326732635498047, "loss/reg": 0.022653890773653984, "step": 552 }, { "epoch": 0.069125, "grad_norm": 6.815005779266357, "grad_norm_var": 1.5125797637252996, "learning_rate": 0.0001, "loss": 1.1457, "loss/crossentropy": 2.8011491298675537, "loss/hidden": 0.78515625, "loss/logits": 0.13409599661827087, "loss/reg": 0.022644398733973503, "step": 553 }, { "epoch": 0.06925, "grad_norm": 4.219581127166748, "grad_norm_var": 1.4192768991356985, "learning_rate": 0.0001, "loss": 1.2504, "loss/crossentropy": 2.5874292850494385, "loss/hidden": 0.8515625, "loss/logits": 0.17252284288406372, "loss/reg": 0.02263464592397213, "step": 554 }, { "epoch": 0.069375, "grad_norm": 3.8279531002044678, "grad_norm_var": 1.3871550629864857, "learning_rate": 0.0001, "loss": 1.1814, "loss/crossentropy": 2.5384669303894043, "loss/hidden": 0.80078125, "loss/logits": 0.1543978452682495, "loss/reg": 0.022625621408224106, "step": 555 }, { "epoch": 0.0695, "grad_norm": 3.563680648803711, "grad_norm_var": 1.3987108056073487, "learning_rate": 0.0001, "loss": 1.0727, "loss/crossentropy": 2.111318349838257, "loss/hidden": 0.73828125, "loss/logits": 0.10826431214809418, "loss/reg": 0.022616824135184288, "step": 556 }, { "epoch": 0.069625, "grad_norm": 3.9599223136901855, "grad_norm_var": 1.3836174934919199, "learning_rate": 0.0001, "loss": 1.0309, "loss/crossentropy": 2.429032325744629, "loss/hidden": 0.67578125, "loss/logits": 0.12908612191677094, "loss/reg": 0.022607678547501564, "step": 557 }, { "epoch": 0.06975, "grad_norm": 2.8072519302368164, "grad_norm_var": 1.4610802306207225, "learning_rate": 0.0001, "loss": 1.0694, "loss/crossentropy": 2.5817906856536865, "loss/hidden": 0.6875, "loss/logits": 0.15589380264282227, "loss/reg": 0.022598396986722946, "step": 558 }, { "epoch": 0.069875, "grad_norm": 2.764833927154541, "grad_norm_var": 1.4475983977607596, "learning_rate": 0.0001, "loss": 1.0233, "loss/crossentropy": 2.202237844467163, "loss/hidden": 0.66015625, "loss/logits": 0.13721255958080292, "loss/reg": 0.02258932963013649, "step": 559 }, { "epoch": 0.07, "grad_norm": 5.113494396209717, "grad_norm_var": 1.5267634464669517, "learning_rate": 0.0001, "loss": 1.1672, "loss/crossentropy": 2.006186008453369, "loss/hidden": 0.80078125, "loss/logits": 0.14061376452445984, "loss/reg": 0.022580046206712723, "step": 560 }, { "epoch": 0.070125, "grad_norm": 3.8133227825164795, "grad_norm_var": 1.1437787263680603, "learning_rate": 0.0001, "loss": 1.0148, "loss/crossentropy": 2.37170672416687, "loss/hidden": 0.67578125, "loss/logits": 0.11329137533903122, "loss/reg": 0.02257111482322216, "step": 561 }, { "epoch": 0.07025, "grad_norm": 2.7208938598632812, "grad_norm_var": 1.2255733087022764, "learning_rate": 0.0001, "loss": 1.0621, "loss/crossentropy": 2.271667242050171, "loss/hidden": 0.703125, "loss/logits": 0.1333208978176117, "loss/reg": 0.02256210334599018, "step": 562 }, { "epoch": 0.070375, "grad_norm": 53.38179016113281, "grad_norm_var": 154.8279377341773, "learning_rate": 0.0001, "loss": 0.9618, "loss/crossentropy": 2.2667322158813477, "loss/hidden": 0.640625, "loss/logits": 0.09567096829414368, "loss/reg": 0.02255306765437126, "step": 563 }, { "epoch": 0.0705, "grad_norm": 5.538954257965088, "grad_norm_var": 154.75309317540348, "learning_rate": 0.0001, "loss": 1.782, "loss/crossentropy": 2.5168824195861816, "loss/hidden": 1.3671875, "loss/logits": 0.1893981695175171, "loss/reg": 0.02254408597946167, "step": 564 }, { "epoch": 0.070625, "grad_norm": 2.8311338424682617, "grad_norm_var": 154.85811657117597, "learning_rate": 0.0001, "loss": 1.1104, "loss/crossentropy": 2.1938281059265137, "loss/hidden": 0.75, "loss/logits": 0.13508498668670654, "loss/reg": 0.02253509685397148, "step": 565 }, { "epoch": 0.07075, "grad_norm": 4.604408264160156, "grad_norm_var": 154.26918998432618, "learning_rate": 0.0001, "loss": 1.2995, "loss/crossentropy": 2.103062391281128, "loss/hidden": 0.91015625, "loss/logits": 0.16405051946640015, "loss/reg": 0.02252543345093727, "step": 566 }, { "epoch": 0.070875, "grad_norm": 4.917901992797852, "grad_norm_var": 153.63084329919155, "learning_rate": 0.0001, "loss": 1.2945, "loss/crossentropy": 2.6932315826416016, "loss/hidden": 0.87890625, "loss/logits": 0.1904207468032837, "loss/reg": 0.022515632212162018, "step": 567 }, { "epoch": 0.071, "grad_norm": 4.055351257324219, "grad_norm_var": 153.02723135387427, "learning_rate": 0.0001, "loss": 1.2253, "loss/crossentropy": 2.3849422931671143, "loss/hidden": 0.8359375, "loss/logits": 0.1642536073923111, "loss/reg": 0.02250652387738228, "step": 568 }, { "epoch": 0.071125, "grad_norm": 2.851072072982788, "grad_norm_var": 154.20402053832456, "learning_rate": 0.0001, "loss": 1.0724, "loss/crossentropy": 2.580428123474121, "loss/hidden": 0.703125, "loss/logits": 0.144349604845047, "loss/reg": 0.022497190162539482, "step": 569 }, { "epoch": 0.07125, "grad_norm": 3.257493495941162, "grad_norm_var": 154.6102933496205, "learning_rate": 0.0001, "loss": 0.9328, "loss/crossentropy": 2.4263837337493896, "loss/hidden": 0.59375, "loss/logits": 0.11414404958486557, "loss/reg": 0.02248740941286087, "step": 570 }, { "epoch": 0.071375, "grad_norm": 3.6194608211517334, "grad_norm_var": 154.69773136421824, "learning_rate": 0.0001, "loss": 1.0417, "loss/crossentropy": 2.7035937309265137, "loss/hidden": 0.69140625, "loss/logits": 0.12549756467342377, "loss/reg": 0.022477447986602783, "step": 571 }, { "epoch": 0.0715, "grad_norm": 4.193033695220947, "grad_norm_var": 154.44566535859553, "learning_rate": 0.0001, "loss": 1.6505, "loss/crossentropy": 2.274057149887085, "loss/hidden": 1.1640625, "loss/logits": 0.26177138090133667, "loss/reg": 0.022467276081442833, "step": 572 }, { "epoch": 0.071625, "grad_norm": 2.7127623558044434, "grad_norm_var": 155.03209308401435, "learning_rate": 0.0001, "loss": 0.9118, "loss/crossentropy": 2.4076664447784424, "loss/hidden": 0.58203125, "loss/logits": 0.10521911084651947, "loss/reg": 0.022456735372543335, "step": 573 }, { "epoch": 0.07175, "grad_norm": 3.5871214866638184, "grad_norm_var": 154.65243889362196, "learning_rate": 0.0001, "loss": 0.9755, "loss/crossentropy": 2.8088905811309814, "loss/hidden": 0.62890625, "loss/logits": 0.12216061353683472, "loss/reg": 0.022446416318416595, "step": 574 }, { "epoch": 0.071875, "grad_norm": 20.843276977539062, "grad_norm_var": 165.17750310305152, "learning_rate": 0.0001, "loss": 1.1707, "loss/crossentropy": 2.569218635559082, "loss/hidden": 0.82421875, "loss/logits": 0.1221412867307663, "loss/reg": 0.02243630215525627, "step": 575 }, { "epoch": 0.072, "grad_norm": 3.375251293182373, "grad_norm_var": 166.03594003131977, "learning_rate": 0.0001, "loss": 0.9923, "loss/crossentropy": 2.7540183067321777, "loss/hidden": 0.640625, "loss/logits": 0.12738245725631714, "loss/reg": 0.022425668314099312, "step": 576 }, { "epoch": 0.072125, "grad_norm": 3.457340717315674, "grad_norm_var": 166.23754433202586, "learning_rate": 0.0001, "loss": 1.0371, "loss/crossentropy": 2.2943716049194336, "loss/hidden": 0.66796875, "loss/logits": 0.14497271180152893, "loss/reg": 0.022415172308683395, "step": 577 }, { "epoch": 0.07225, "grad_norm": 2.966371774673462, "grad_norm_var": 166.07272256293106, "learning_rate": 0.0001, "loss": 1.2044, "loss/crossentropy": 2.760359048843384, "loss/hidden": 0.8125, "loss/logits": 0.16780292987823486, "loss/reg": 0.022406071424484253, "step": 578 }, { "epoch": 0.072375, "grad_norm": 2.5532329082489014, "grad_norm_var": 19.219812763276813, "learning_rate": 0.0001, "loss": 0.9996, "loss/crossentropy": 2.427797794342041, "loss/hidden": 0.66796875, "loss/logits": 0.10767525434494019, "loss/reg": 0.02239692024886608, "step": 579 }, { "epoch": 0.0725, "grad_norm": 3.250906229019165, "grad_norm_var": 19.294198335433887, "learning_rate": 0.0001, "loss": 1.1148, "loss/crossentropy": 2.29978084564209, "loss/hidden": 0.75, "loss/logits": 0.1408957540988922, "loss/reg": 0.0223868228495121, "step": 580 }, { "epoch": 0.072625, "grad_norm": 3.8051562309265137, "grad_norm_var": 19.12802354300362, "learning_rate": 0.0001, "loss": 1.2329, "loss/crossentropy": 2.402350664138794, "loss/hidden": 0.83203125, "loss/logits": 0.17713911831378937, "loss/reg": 0.022376833483576775, "step": 581 }, { "epoch": 0.07275, "grad_norm": 4.466115951538086, "grad_norm_var": 19.129656316190147, "learning_rate": 0.0001, "loss": 1.1846, "loss/crossentropy": 2.5129504203796387, "loss/hidden": 0.8125, "loss/logits": 0.14841441810131073, "loss/reg": 0.02236761339008808, "step": 582 }, { "epoch": 0.072875, "grad_norm": 3.1465752124786377, "grad_norm_var": 19.255278342461487, "learning_rate": 0.0001, "loss": 1.1857, "loss/crossentropy": 2.5782837867736816, "loss/hidden": 0.8203125, "loss/logits": 0.14178822934627533, "loss/reg": 0.022358402609825134, "step": 583 }, { "epoch": 0.073, "grad_norm": 2.7474803924560547, "grad_norm_var": 19.441256858474677, "learning_rate": 0.0001, "loss": 1.0532, "loss/crossentropy": 2.4901747703552246, "loss/hidden": 0.703125, "loss/logits": 0.12661507725715637, "loss/reg": 0.02234930731356144, "step": 584 }, { "epoch": 0.073125, "grad_norm": 2.7027359008789062, "grad_norm_var": 19.47380183903334, "learning_rate": 0.0001, "loss": 0.8469, "loss/crossentropy": 2.694582223892212, "loss/hidden": 0.5390625, "loss/logits": 0.08438676595687866, "loss/reg": 0.022340187802910805, "step": 585 }, { "epoch": 0.07325, "grad_norm": 2.847770929336548, "grad_norm_var": 19.547679388785195, "learning_rate": 0.0001, "loss": 0.9885, "loss/crossentropy": 2.5558888912200928, "loss/hidden": 0.6484375, "loss/logits": 0.11670757830142975, "loss/reg": 0.022330984473228455, "step": 586 }, { "epoch": 0.073375, "grad_norm": 3.20444917678833, "grad_norm_var": 19.601201389954156, "learning_rate": 0.0001, "loss": 1.2216, "loss/crossentropy": 2.0031001567840576, "loss/hidden": 0.84375, "loss/logits": 0.1546502560377121, "loss/reg": 0.022321749478578568, "step": 587 }, { "epoch": 0.0735, "grad_norm": 4.657837390899658, "grad_norm_var": 19.6039707895662, "learning_rate": 0.0001, "loss": 1.1377, "loss/crossentropy": 2.598100185394287, "loss/hidden": 0.78125, "loss/logits": 0.13329939544200897, "loss/reg": 0.02231265790760517, "step": 588 }, { "epoch": 0.073625, "grad_norm": 3.1166787147521973, "grad_norm_var": 19.52355503271277, "learning_rate": 0.0001, "loss": 1.304, "loss/crossentropy": 2.2209084033966064, "loss/hidden": 0.89453125, "loss/logits": 0.18648099899291992, "loss/reg": 0.022303014993667603, "step": 589 }, { "epoch": 0.07375, "grad_norm": 2.988344669342041, "grad_norm_var": 19.61249925539728, "learning_rate": 0.0001, "loss": 0.9412, "loss/crossentropy": 2.4533708095550537, "loss/hidden": 0.60546875, "loss/logits": 0.11275988817214966, "loss/reg": 0.022293319925665855, "step": 590 }, { "epoch": 0.073875, "grad_norm": 3.0553901195526123, "grad_norm_var": 0.3491433903254536, "learning_rate": 0.0001, "loss": 1.0883, "loss/crossentropy": 2.3986990451812744, "loss/hidden": 0.73046875, "loss/logits": 0.13498544692993164, "loss/reg": 0.022283662110567093, "step": 591 }, { "epoch": 0.074, "grad_norm": 4.445681095123291, "grad_norm_var": 0.43558600780207446, "learning_rate": 0.0001, "loss": 1.3195, "loss/crossentropy": 2.335937976837158, "loss/hidden": 0.90625, "loss/logits": 0.19050292670726776, "loss/reg": 0.022273709997534752, "step": 592 }, { "epoch": 0.074125, "grad_norm": 41.39726638793945, "grad_norm_var": 91.00287624901027, "learning_rate": 0.0001, "loss": 1.2527, "loss/crossentropy": 2.3441669940948486, "loss/hidden": 0.86328125, "loss/logits": 0.1667976826429367, "loss/reg": 0.02226419560611248, "step": 593 }, { "epoch": 0.07425, "grad_norm": 2.8557939529418945, "grad_norm_var": 91.04408434440504, "learning_rate": 0.0001, "loss": 1.0612, "loss/crossentropy": 2.380885362625122, "loss/hidden": 0.703125, "loss/logits": 0.1354852318763733, "loss/reg": 0.022254258394241333, "step": 594 }, { "epoch": 0.074375, "grad_norm": 6.565737724304199, "grad_norm_var": 90.36543928633743, "learning_rate": 0.0001, "loss": 1.2519, "loss/crossentropy": 2.3877017498016357, "loss/hidden": 0.90625, "loss/logits": 0.1231798455119133, "loss/reg": 0.022244345396757126, "step": 595 }, { "epoch": 0.0745, "grad_norm": 3.6746349334716797, "grad_norm_var": 90.22397938232942, "learning_rate": 0.0001, "loss": 1.1718, "loss/crossentropy": 2.397080898284912, "loss/hidden": 0.80078125, "loss/logits": 0.1486300826072693, "loss/reg": 0.022234413772821426, "step": 596 }, { "epoch": 0.074625, "grad_norm": 2.981614589691162, "grad_norm_var": 90.50516196939815, "learning_rate": 0.0001, "loss": 1.043, "loss/crossentropy": 2.37412428855896, "loss/hidden": 0.69140625, "loss/logits": 0.12931303679943085, "loss/reg": 0.022224588319659233, "step": 597 }, { "epoch": 0.07475, "grad_norm": 3.6531307697296143, "grad_norm_var": 90.70497774366554, "learning_rate": 0.0001, "loss": 1.1272, "loss/crossentropy": 2.6906161308288574, "loss/hidden": 0.76953125, "loss/logits": 0.13550975918769836, "loss/reg": 0.02221417799592018, "step": 598 }, { "epoch": 0.074875, "grad_norm": 3.751652240753174, "grad_norm_var": 90.50753182721607, "learning_rate": 0.0001, "loss": 1.1618, "loss/crossentropy": 2.6506831645965576, "loss/hidden": 0.79296875, "loss/logits": 0.1468081772327423, "loss/reg": 0.022204989567399025, "step": 599 }, { "epoch": 0.075, "grad_norm": 3.228563070297241, "grad_norm_var": 90.31879350061254, "learning_rate": 0.0001, "loss": 1.0791, "loss/crossentropy": 2.381368398666382, "loss/hidden": 0.71875, "loss/logits": 0.13835087418556213, "loss/reg": 0.022195899859070778, "step": 600 }, { "epoch": 0.075125, "grad_norm": 5.392886638641357, "grad_norm_var": 89.60797997668053, "learning_rate": 0.0001, "loss": 1.2309, "loss/crossentropy": 2.52596378326416, "loss/hidden": 0.8359375, "loss/logits": 0.17307403683662415, "loss/reg": 0.022185994312167168, "step": 601 }, { "epoch": 0.07525, "grad_norm": 3.2832775115966797, "grad_norm_var": 89.43019603463321, "learning_rate": 0.0001, "loss": 1.1462, "loss/crossentropy": 2.4442057609558105, "loss/hidden": 0.76953125, "loss/logits": 0.15487736463546753, "loss/reg": 0.02217610739171505, "step": 602 }, { "epoch": 0.075375, "grad_norm": 4.1014084815979, "grad_norm_var": 89.1293068696746, "learning_rate": 0.0001, "loss": 1.0847, "loss/crossentropy": 2.531064987182617, "loss/hidden": 0.7265625, "loss/logits": 0.13643184304237366, "loss/reg": 0.02216634899377823, "step": 603 }, { "epoch": 0.0755, "grad_norm": 3.656172037124634, "grad_norm_var": 89.39756111673695, "learning_rate": 0.0001, "loss": 1.0161, "loss/crossentropy": 2.470468044281006, "loss/hidden": 0.68359375, "loss/logits": 0.11097002029418945, "loss/reg": 0.02215682342648506, "step": 604 }, { "epoch": 0.075625, "grad_norm": 3.8061492443084717, "grad_norm_var": 89.1498668494714, "learning_rate": 0.0001, "loss": 1.1195, "loss/crossentropy": 2.469788074493408, "loss/hidden": 0.7734375, "loss/logits": 0.12457874417304993, "loss/reg": 0.022147687152028084, "step": 605 }, { "epoch": 0.07575, "grad_norm": 2.658135414123535, "grad_norm_var": 89.29708722871554, "learning_rate": 0.0001, "loss": 1.0657, "loss/crossentropy": 2.2552011013031006, "loss/hidden": 0.71484375, "loss/logits": 0.12950366735458374, "loss/reg": 0.022138802334666252, "step": 606 }, { "epoch": 0.075875, "grad_norm": 3.291750907897949, "grad_norm_var": 89.20284122750789, "learning_rate": 0.0001, "loss": 1.133, "loss/crossentropy": 2.225055456161499, "loss/hidden": 0.73828125, "loss/logits": 0.1734064519405365, "loss/reg": 0.022129878401756287, "step": 607 }, { "epoch": 0.076, "grad_norm": 4.001339912414551, "grad_norm_var": 89.31742762195415, "learning_rate": 0.0001, "loss": 0.9658, "loss/crossentropy": 2.3826169967651367, "loss/hidden": 0.64453125, "loss/logits": 0.10007497668266296, "loss/reg": 0.022121025249361992, "step": 608 }, { "epoch": 0.076125, "grad_norm": 2.9729602336883545, "grad_norm_var": 0.9817241823775095, "learning_rate": 0.0001, "loss": 1.1962, "loss/crossentropy": 2.4926881790161133, "loss/hidden": 0.82421875, "loss/logits": 0.1508128046989441, "loss/reg": 0.022111859172582626, "step": 609 }, { "epoch": 0.07625, "grad_norm": 2.5144121646881104, "grad_norm_var": 1.0293551003726749, "learning_rate": 0.0001, "loss": 0.9883, "loss/crossentropy": 2.478576898574829, "loss/hidden": 0.65234375, "loss/logits": 0.11488830298185349, "loss/reg": 0.02210419811308384, "step": 610 }, { "epoch": 0.076375, "grad_norm": 3.7777743339538574, "grad_norm_var": 0.4576308797359, "learning_rate": 0.0001, "loss": 1.0059, "loss/crossentropy": 2.4356257915496826, "loss/hidden": 0.6484375, "loss/logits": 0.13651525974273682, "loss/reg": 0.022095149382948875, "step": 611 }, { "epoch": 0.0765, "grad_norm": 3.232398748397827, "grad_norm_var": 0.4623055923756754, "learning_rate": 0.0001, "loss": 1.0108, "loss/crossentropy": 2.4058218002319336, "loss/hidden": 0.67578125, "loss/logits": 0.11410736292600632, "loss/reg": 0.022087210789322853, "step": 612 }, { "epoch": 0.076625, "grad_norm": 2.61934757232666, "grad_norm_var": 0.49646373584008807, "learning_rate": 0.0001, "loss": 0.9466, "loss/crossentropy": 2.3981077671051025, "loss/hidden": 0.59765625, "loss/logits": 0.1281721591949463, "loss/reg": 0.022078126668930054, "step": 613 }, { "epoch": 0.07675, "grad_norm": 4.228268146514893, "grad_norm_var": 0.5291615579452734, "learning_rate": 0.0001, "loss": 1.0729, "loss/crossentropy": 2.664149761199951, "loss/hidden": 0.73046875, "loss/logits": 0.12176868319511414, "loss/reg": 0.02207016758620739, "step": 614 }, { "epoch": 0.076875, "grad_norm": 2.9137685298919678, "grad_norm_var": 0.5485319535320492, "learning_rate": 0.0001, "loss": 0.8708, "loss/crossentropy": 2.5457890033721924, "loss/hidden": 0.55078125, "loss/logits": 0.09941907972097397, "loss/reg": 0.022060981020331383, "step": 615 }, { "epoch": 0.077, "grad_norm": 3.4499685764312744, "grad_norm_var": 0.5441756848342263, "learning_rate": 0.0001, "loss": 1.1439, "loss/crossentropy": 2.3822786808013916, "loss/hidden": 0.78125, "loss/logits": 0.14211627840995789, "loss/reg": 0.02205180749297142, "step": 616 }, { "epoch": 0.077125, "grad_norm": 2.271921157836914, "grad_norm_var": 0.36266744154546565, "learning_rate": 0.0001, "loss": 0.9061, "loss/crossentropy": 2.707848072052002, "loss/hidden": 0.58203125, "loss/logits": 0.10360105335712433, "loss/reg": 0.022042402997612953, "step": 617 }, { "epoch": 0.07725, "grad_norm": 3.16980242729187, "grad_norm_var": 0.36370543210803513, "learning_rate": 0.0001, "loss": 1.2733, "loss/crossentropy": 2.1980466842651367, "loss/hidden": 0.86328125, "loss/logits": 0.18972548842430115, "loss/reg": 0.02203306369483471, "step": 618 }, { "epoch": 0.077375, "grad_norm": 2.4408295154571533, "grad_norm_var": 0.3567501583972517, "learning_rate": 0.0001, "loss": 1.1352, "loss/crossentropy": 2.3986692428588867, "loss/hidden": 0.76953125, "loss/logits": 0.1454274207353592, "loss/reg": 0.022023871541023254, "step": 619 }, { "epoch": 0.0775, "grad_norm": 2.558687925338745, "grad_norm_var": 0.36349398943808237, "learning_rate": 0.0001, "loss": 0.9776, "loss/crossentropy": 2.4290616512298584, "loss/hidden": 0.65625, "loss/logits": 0.1012360006570816, "loss/reg": 0.0220141913741827, "step": 620 }, { "epoch": 0.077625, "grad_norm": 3.7468619346618652, "grad_norm_var": 0.3582835152003213, "learning_rate": 0.0001, "loss": 1.1635, "loss/crossentropy": 2.4566454887390137, "loss/hidden": 0.78515625, "loss/logits": 0.15830263495445251, "loss/reg": 0.02200442925095558, "step": 621 }, { "epoch": 0.07775, "grad_norm": 3.8364651203155518, "grad_norm_var": 0.3732032502257139, "learning_rate": 0.0001, "loss": 1.3213, "loss/crossentropy": 2.5934860706329346, "loss/hidden": 0.9296875, "loss/logits": 0.17167437076568604, "loss/reg": 0.021995313465595245, "step": 622 }, { "epoch": 0.077875, "grad_norm": 2.9136173725128174, "grad_norm_var": 0.37696739372618043, "learning_rate": 0.0001, "loss": 1.1191, "loss/crossentropy": 2.1625287532806396, "loss/hidden": 0.75, "loss/logits": 0.1492336541414261, "loss/reg": 0.021986283361911774, "step": 623 }, { "epoch": 0.078, "grad_norm": 3.3136067390441895, "grad_norm_var": 0.3298862344756988, "learning_rate": 0.0001, "loss": 1.1487, "loss/crossentropy": 2.3915464878082275, "loss/hidden": 0.7890625, "loss/logits": 0.1398892104625702, "loss/reg": 0.021976841613650322, "step": 624 }, { "epoch": 0.078125, "grad_norm": 3.1731748580932617, "grad_norm_var": 0.3283984444798058, "learning_rate": 0.0001, "loss": 1.0945, "loss/crossentropy": 2.53873610496521, "loss/hidden": 0.7109375, "loss/logits": 0.16392138600349426, "loss/reg": 0.021966535598039627, "step": 625 }, { "epoch": 0.07825, "grad_norm": 4.172556400299072, "grad_norm_var": 0.3630228628347131, "learning_rate": 0.0001, "loss": 1.1888, "loss/crossentropy": 2.4219956398010254, "loss/hidden": 0.79296875, "loss/logits": 0.17627781629562378, "loss/reg": 0.021956363692879677, "step": 626 }, { "epoch": 0.078375, "grad_norm": 3.99817156791687, "grad_norm_var": 0.3819004722530487, "learning_rate": 0.0001, "loss": 1.3074, "loss/crossentropy": 2.2407050132751465, "loss/hidden": 0.8828125, "loss/logits": 0.20513707399368286, "loss/reg": 0.02194611169397831, "step": 627 }, { "epoch": 0.0785, "grad_norm": 3.2869558334350586, "grad_norm_var": 0.3819405314837089, "learning_rate": 0.0001, "loss": 0.923, "loss/crossentropy": 2.5824477672576904, "loss/hidden": 0.59375, "loss/logits": 0.10985252261161804, "loss/reg": 0.021935785189270973, "step": 628 }, { "epoch": 0.078625, "grad_norm": 3.0037360191345215, "grad_norm_var": 0.35855200267849247, "learning_rate": 0.0001, "loss": 1.0636, "loss/crossentropy": 2.4916722774505615, "loss/hidden": 0.7109375, "loss/logits": 0.13343587517738342, "loss/reg": 0.021925168111920357, "step": 629 }, { "epoch": 0.07875, "grad_norm": 3.5274088382720947, "grad_norm_var": 0.3006291732182412, "learning_rate": 0.0001, "loss": 1.1441, "loss/crossentropy": 1.7279773950576782, "loss/hidden": 0.78515625, "loss/logits": 0.13975940644741058, "loss/reg": 0.02191445231437683, "step": 630 }, { "epoch": 0.078875, "grad_norm": 3.2895989418029785, "grad_norm_var": 0.29330515223301745, "learning_rate": 0.0001, "loss": 1.0394, "loss/crossentropy": 2.370850086212158, "loss/hidden": 0.68359375, "loss/logits": 0.13678821921348572, "loss/reg": 0.02190525084733963, "step": 631 }, { "epoch": 0.079, "grad_norm": 3.741135597229004, "grad_norm_var": 0.30599490652712474, "learning_rate": 0.0001, "loss": 1.2655, "loss/crossentropy": 2.365464210510254, "loss/hidden": 0.87890625, "loss/logits": 0.16766133904457092, "loss/reg": 0.02189476415514946, "step": 632 }, { "epoch": 0.079125, "grad_norm": 4.110158443450928, "grad_norm_var": 0.2706546096444164, "learning_rate": 0.0001, "loss": 1.1804, "loss/crossentropy": 2.658433437347412, "loss/hidden": 0.796875, "loss/logits": 0.1646716445684433, "loss/reg": 0.021885616704821587, "step": 633 }, { "epoch": 0.07925, "grad_norm": 3.7503654956817627, "grad_norm_var": 0.27446839769864156, "learning_rate": 0.0001, "loss": 1.1, "loss/crossentropy": 2.4457831382751465, "loss/hidden": 0.74609375, "loss/logits": 0.13510534167289734, "loss/reg": 0.021875550970435143, "step": 634 }, { "epoch": 0.079375, "grad_norm": 2.917835235595703, "grad_norm_var": 0.22584356567046956, "learning_rate": 0.0001, "loss": 1.1498, "loss/crossentropy": 2.3328001499176025, "loss/hidden": 0.77734375, "loss/logits": 0.1537725031375885, "loss/reg": 0.021865583956241608, "step": 635 }, { "epoch": 0.0795, "grad_norm": 2.907240152359009, "grad_norm_var": 0.19160647764443486, "learning_rate": 0.0001, "loss": 0.9073, "loss/crossentropy": 2.556042432785034, "loss/hidden": 0.59375, "loss/logits": 0.09503352642059326, "loss/reg": 0.02185530960559845, "step": 636 }, { "epoch": 0.079625, "grad_norm": 2.5937955379486084, "grad_norm_var": 0.2337615816576618, "learning_rate": 0.0001, "loss": 0.9198, "loss/crossentropy": 2.363370180130005, "loss/hidden": 0.5859375, "loss/logits": 0.11537887156009674, "loss/reg": 0.021845519542694092, "step": 637 }, { "epoch": 0.07975, "grad_norm": 3.9080708026885986, "grad_norm_var": 0.2381681132369368, "learning_rate": 0.0001, "loss": 1.1552, "loss/crossentropy": 2.5054194927215576, "loss/hidden": 0.81640625, "loss/logits": 0.12043754756450653, "loss/reg": 0.021835271269083023, "step": 638 }, { "epoch": 0.079875, "grad_norm": 2.8192946910858154, "grad_norm_var": 0.24500412598165416, "learning_rate": 0.0001, "loss": 0.9375, "loss/crossentropy": 2.6611719131469727, "loss/hidden": 0.60546875, "loss/logits": 0.11382012814283371, "loss/reg": 0.021824965253472328, "step": 639 }, { "epoch": 0.08, "grad_norm": 4.111716270446777, "grad_norm_var": 0.27486954530744445, "learning_rate": 0.0001, "loss": 1.2972, "loss/crossentropy": 2.472568988800049, "loss/hidden": 0.91015625, "loss/logits": 0.16890740394592285, "loss/reg": 0.02181575633585453, "step": 640 }, { "epoch": 0.080125, "grad_norm": 5.715061187744141, "grad_norm_var": 0.5825168124345791, "learning_rate": 0.0001, "loss": 1.0307, "loss/crossentropy": 2.2123966217041016, "loss/hidden": 0.68359375, "loss/logits": 0.12899596989154816, "loss/reg": 0.021806620061397552, "step": 641 }, { "epoch": 0.08025, "grad_norm": 3.8680315017700195, "grad_norm_var": 0.5657073815126407, "learning_rate": 0.0001, "loss": 1.053, "loss/crossentropy": 2.4526453018188477, "loss/hidden": 0.703125, "loss/logits": 0.13190723955631256, "loss/reg": 0.021797508001327515, "step": 642 }, { "epoch": 0.080375, "grad_norm": 4.992859363555908, "grad_norm_var": 0.680778895488037, "learning_rate": 0.0001, "loss": 1.1683, "loss/crossentropy": 2.3445968627929688, "loss/hidden": 0.82421875, "loss/logits": 0.12617573142051697, "loss/reg": 0.02178841643035412, "step": 643 }, { "epoch": 0.0805, "grad_norm": 4.339104175567627, "grad_norm_var": 0.6977811040599325, "learning_rate": 0.0001, "loss": 1.3433, "loss/crossentropy": 1.8577097654342651, "loss/hidden": 0.9453125, "loss/logits": 0.18017446994781494, "loss/reg": 0.02177964523434639, "step": 644 }, { "epoch": 0.080625, "grad_norm": 4.026562213897705, "grad_norm_var": 0.6648423545945282, "learning_rate": 0.0001, "loss": 1.3266, "loss/crossentropy": 2.415410041809082, "loss/hidden": 0.859375, "loss/logits": 0.24951061606407166, "loss/reg": 0.021770501509308815, "step": 645 }, { "epoch": 0.08075, "grad_norm": 2.8517775535583496, "grad_norm_var": 0.7169049906385166, "learning_rate": 0.0001, "loss": 0.9283, "loss/crossentropy": 2.6426851749420166, "loss/hidden": 0.6015625, "loss/logits": 0.10916159301996231, "loss/reg": 0.021761184558272362, "step": 646 }, { "epoch": 0.080875, "grad_norm": 3.3171377182006836, "grad_norm_var": 0.7152750431492562, "learning_rate": 0.0001, "loss": 1.0088, "loss/crossentropy": 2.638533115386963, "loss/hidden": 0.62890625, "loss/logits": 0.1623907834291458, "loss/reg": 0.02175196446478367, "step": 647 }, { "epoch": 0.081, "grad_norm": 3.653881311416626, "grad_norm_var": 0.7158322952113887, "learning_rate": 0.0001, "loss": 1.1132, "loss/crossentropy": 2.745229482650757, "loss/hidden": 0.765625, "loss/logits": 0.1301097273826599, "loss/reg": 0.021742329001426697, "step": 648 }, { "epoch": 0.081125, "grad_norm": 4.31439733505249, "grad_norm_var": 0.728446489341123, "learning_rate": 0.0001, "loss": 1.2706, "loss/crossentropy": 2.7004082202911377, "loss/hidden": 0.80078125, "loss/logits": 0.25250470638275146, "loss/reg": 0.021732579916715622, "step": 649 }, { "epoch": 0.08125, "grad_norm": 4.621687889099121, "grad_norm_var": 0.7753064642270201, "learning_rate": 0.0001, "loss": 1.4091, "loss/crossentropy": 2.38405179977417, "loss/hidden": 0.89453125, "loss/logits": 0.29737186431884766, "loss/reg": 0.02172265760600567, "step": 650 }, { "epoch": 0.081375, "grad_norm": 2.87424898147583, "grad_norm_var": 0.7806094534209419, "learning_rate": 0.0001, "loss": 1.2103, "loss/crossentropy": 2.0905954837799072, "loss/hidden": 0.8125, "loss/logits": 0.18066659569740295, "loss/reg": 0.021712414920330048, "step": 651 }, { "epoch": 0.0815, "grad_norm": 4.724517345428467, "grad_norm_var": 0.7689569917943563, "learning_rate": 0.0001, "loss": 1.2928, "loss/crossentropy": 2.500594139099121, "loss/hidden": 0.89453125, "loss/logits": 0.1812204122543335, "loss/reg": 0.021703310310840607, "step": 652 }, { "epoch": 0.081625, "grad_norm": 2.8673081398010254, "grad_norm_var": 0.7252403996552116, "learning_rate": 0.0001, "loss": 1.0588, "loss/crossentropy": 2.3599724769592285, "loss/hidden": 0.703125, "loss/logits": 0.13873916864395142, "loss/reg": 0.021693557500839233, "step": 653 }, { "epoch": 0.08175, "grad_norm": 3.099315643310547, "grad_norm_var": 0.7693322976491117, "learning_rate": 0.0001, "loss": 1.1961, "loss/crossentropy": 2.3167355060577393, "loss/hidden": 0.83203125, "loss/logits": 0.14727652072906494, "loss/reg": 0.021683741360902786, "step": 654 }, { "epoch": 0.081875, "grad_norm": 3.287602186203003, "grad_norm_var": 0.7163515778113151, "learning_rate": 0.0001, "loss": 1.0429, "loss/crossentropy": 2.7967636585235596, "loss/hidden": 0.6875, "loss/logits": 0.13863959908485413, "loss/reg": 0.021674364805221558, "step": 655 }, { "epoch": 0.082, "grad_norm": 3.487874746322632, "grad_norm_var": 0.7244436337536264, "learning_rate": 0.0001, "loss": 1.0883, "loss/crossentropy": 2.5409998893737793, "loss/hidden": 0.71875, "loss/logits": 0.1528949737548828, "loss/reg": 0.02166520059108734, "step": 656 }, { "epoch": 0.082125, "grad_norm": 3.6202752590179443, "grad_norm_var": 0.4854858648423857, "learning_rate": 0.0001, "loss": 1.2516, "loss/crossentropy": 2.3633673191070557, "loss/hidden": 0.88671875, "loss/logits": 0.14828172326087952, "loss/reg": 0.021655315533280373, "step": 657 }, { "epoch": 0.08225, "grad_norm": 3.7921783924102783, "grad_norm_var": 0.484617963461113, "learning_rate": 0.0001, "loss": 1.07, "loss/crossentropy": 2.497013807296753, "loss/hidden": 0.71484375, "loss/logits": 0.13874131441116333, "loss/reg": 0.021646033972501755, "step": 658 }, { "epoch": 0.082375, "grad_norm": 5.818857669830322, "grad_norm_var": 0.6650298211735747, "learning_rate": 0.0001, "loss": 1.2961, "loss/crossentropy": 2.6599833965301514, "loss/hidden": 0.91015625, "loss/logits": 0.16955840587615967, "loss/reg": 0.021636882796883583, "step": 659 }, { "epoch": 0.0825, "grad_norm": 3.465527057647705, "grad_norm_var": 0.6491808619433999, "learning_rate": 0.0001, "loss": 1.1365, "loss/crossentropy": 2.5972306728363037, "loss/hidden": 0.7734375, "loss/logits": 0.14674408733844757, "loss/reg": 0.02162766456604004, "step": 660 }, { "epoch": 0.082625, "grad_norm": 3.143159866333008, "grad_norm_var": 0.664078497493661, "learning_rate": 0.0001, "loss": 1.2241, "loss/crossentropy": 2.5945792198181152, "loss/hidden": 0.83203125, "loss/logits": 0.17588043212890625, "loss/reg": 0.021618474274873734, "step": 661 }, { "epoch": 0.08275, "grad_norm": 3.7091660499572754, "grad_norm_var": 0.6149151800980365, "learning_rate": 0.0001, "loss": 1.1901, "loss/crossentropy": 2.631565809249878, "loss/hidden": 0.8125, "loss/logits": 0.16149985790252686, "loss/reg": 0.021609637886285782, "step": 662 }, { "epoch": 0.082875, "grad_norm": 3.1449224948883057, "grad_norm_var": 0.6264170707357067, "learning_rate": 0.0001, "loss": 1.1491, "loss/crossentropy": 2.2890501022338867, "loss/hidden": 0.78515625, "loss/logits": 0.14797118306159973, "loss/reg": 0.021600957959890366, "step": 663 }, { "epoch": 0.083, "grad_norm": 3.096752405166626, "grad_norm_var": 0.6512152784754629, "learning_rate": 0.0001, "loss": 1.0925, "loss/crossentropy": 2.767069101333618, "loss/hidden": 0.7421875, "loss/logits": 0.1343650072813034, "loss/reg": 0.021592585369944572, "step": 664 }, { "epoch": 0.083125, "grad_norm": 3.8935883045196533, "grad_norm_var": 0.6273466460071402, "learning_rate": 0.0001, "loss": 1.0996, "loss/crossentropy": 2.4779109954833984, "loss/hidden": 0.7578125, "loss/logits": 0.12599951028823853, "loss/reg": 0.021583350375294685, "step": 665 }, { "epoch": 0.08325, "grad_norm": 3.6881141662597656, "grad_norm_var": 0.5627883047301658, "learning_rate": 0.0001, "loss": 1.1184, "loss/crossentropy": 2.1603968143463135, "loss/hidden": 0.76171875, "loss/logits": 0.14095276594161987, "loss/reg": 0.021574225276708603, "step": 666 }, { "epoch": 0.083375, "grad_norm": 2.9750800132751465, "grad_norm_var": 0.5535713466115603, "learning_rate": 0.0001, "loss": 1.1215, "loss/crossentropy": 2.4817895889282227, "loss/hidden": 0.75390625, "loss/logits": 0.15194162726402283, "loss/reg": 0.02156493254005909, "step": 667 }, { "epoch": 0.0835, "grad_norm": 2.649543046951294, "grad_norm_var": 0.5152581471177806, "learning_rate": 0.0001, "loss": 0.918, "loss/crossentropy": 2.334226608276367, "loss/hidden": 0.59375, "loss/logits": 0.10870292782783508, "loss/reg": 0.02155502513051033, "step": 668 }, { "epoch": 0.083625, "grad_norm": 4.535495281219482, "grad_norm_var": 0.5520843285134825, "learning_rate": 0.0001, "loss": 1.168, "loss/crossentropy": 2.4633467197418213, "loss/hidden": 0.796875, "loss/logits": 0.15570059418678284, "loss/reg": 0.021545063704252243, "step": 669 }, { "epoch": 0.08375, "grad_norm": 3.3291449546813965, "grad_norm_var": 0.5404115229162234, "learning_rate": 0.0001, "loss": 1.1241, "loss/crossentropy": 2.317607879638672, "loss/hidden": 0.78515625, "loss/logits": 0.12363065779209137, "loss/reg": 0.021535001695156097, "step": 670 }, { "epoch": 0.083875, "grad_norm": 4.071917533874512, "grad_norm_var": 0.5459456401930327, "learning_rate": 0.0001, "loss": 1.1652, "loss/crossentropy": 2.4951958656311035, "loss/hidden": 0.83203125, "loss/logits": 0.11796893179416656, "loss/reg": 0.021524924784898758, "step": 671 }, { "epoch": 0.084, "grad_norm": 4.647782802581787, "grad_norm_var": 0.6047501670354971, "learning_rate": 0.0001, "loss": 1.157, "loss/crossentropy": 2.6994447708129883, "loss/hidden": 0.78125, "loss/logits": 0.16059228777885437, "loss/reg": 0.021514689549803734, "step": 672 }, { "epoch": 0.084125, "grad_norm": 4.240062713623047, "grad_norm_var": 0.6201999433703328, "learning_rate": 0.0001, "loss": 1.0984, "loss/crossentropy": 2.592426300048828, "loss/hidden": 0.7734375, "loss/logits": 0.10989370942115784, "loss/reg": 0.021505359560251236, "step": 673 }, { "epoch": 0.08425, "grad_norm": 15.647865295410156, "grad_norm_var": 9.451818582857973, "learning_rate": 0.0001, "loss": 1.6418, "loss/crossentropy": 2.723198413848877, "loss/hidden": 1.078125, "loss/logits": 0.3486996293067932, "loss/reg": 0.02149534970521927, "step": 674 }, { "epoch": 0.084375, "grad_norm": 3.6480963230133057, "grad_norm_var": 9.365638761154676, "learning_rate": 0.0001, "loss": 1.1052, "loss/crossentropy": 2.618795871734619, "loss/hidden": 0.734375, "loss/logits": 0.1559816300868988, "loss/reg": 0.02148519456386566, "step": 675 }, { "epoch": 0.0845, "grad_norm": 3.5789527893066406, "grad_norm_var": 9.352796045350159, "learning_rate": 0.0001, "loss": 1.1092, "loss/crossentropy": 2.4817607402801514, "loss/hidden": 0.75, "loss/logits": 0.14444658160209656, "loss/reg": 0.021475963294506073, "step": 676 }, { "epoch": 0.084625, "grad_norm": 3.143718719482422, "grad_norm_var": 9.352704277495931, "learning_rate": 0.0001, "loss": 1.0394, "loss/crossentropy": 2.28035044670105, "loss/hidden": 0.70703125, "loss/logits": 0.11767329275608063, "loss/reg": 0.02146601676940918, "step": 677 }, { "epoch": 0.08475, "grad_norm": 4.2889580726623535, "grad_norm_var": 9.322240526517621, "learning_rate": 0.0001, "loss": 1.1067, "loss/crossentropy": 2.28658390045166, "loss/hidden": 0.7109375, "loss/logits": 0.18121060729026794, "loss/reg": 0.021456118673086166, "step": 678 }, { "epoch": 0.084875, "grad_norm": 2.800516128540039, "grad_norm_var": 9.387804829955044, "learning_rate": 0.0001, "loss": 0.975, "loss/crossentropy": 2.699044704437256, "loss/hidden": 0.63671875, "loss/logits": 0.12379397451877594, "loss/reg": 0.02144702896475792, "step": 679 }, { "epoch": 0.085, "grad_norm": 3.064215898513794, "grad_norm_var": 9.393480165725077, "learning_rate": 0.0001, "loss": 1.1186, "loss/crossentropy": 2.547557830810547, "loss/hidden": 0.76171875, "loss/logits": 0.14249341189861298, "loss/reg": 0.02143782004714012, "step": 680 }, { "epoch": 0.085125, "grad_norm": 3.7314915657043457, "grad_norm_var": 9.405801361337378, "learning_rate": 0.0001, "loss": 1.1376, "loss/crossentropy": 2.705601930618286, "loss/hidden": 0.7734375, "loss/logits": 0.14988452196121216, "loss/reg": 0.02142806351184845, "step": 681 }, { "epoch": 0.08525, "grad_norm": 2.256387948989868, "grad_norm_var": 9.665529326305519, "learning_rate": 0.0001, "loss": 0.945, "loss/crossentropy": 2.464989423751831, "loss/hidden": 0.62109375, "loss/logits": 0.10975323617458344, "loss/reg": 0.02141808532178402, "step": 682 }, { "epoch": 0.085375, "grad_norm": 3.150348424911499, "grad_norm_var": 9.636765682886749, "learning_rate": 0.0001, "loss": 1.0573, "loss/crossentropy": 2.2520275115966797, "loss/hidden": 0.72265625, "loss/logits": 0.12059713900089264, "loss/reg": 0.021408328786492348, "step": 683 }, { "epoch": 0.0855, "grad_norm": 2.973684549331665, "grad_norm_var": 9.572043410499656, "learning_rate": 0.0001, "loss": 1.0663, "loss/crossentropy": 2.4148662090301514, "loss/hidden": 0.73828125, "loss/logits": 0.11397817730903625, "loss/reg": 0.021399127319455147, "step": 684 }, { "epoch": 0.085625, "grad_norm": 4.39288330078125, "grad_norm_var": 9.56920341692891, "learning_rate": 0.0001, "loss": 1.206, "loss/crossentropy": 2.5025415420532227, "loss/hidden": 0.84765625, "loss/logits": 0.1444375216960907, "loss/reg": 0.021389208734035492, "step": 685 }, { "epoch": 0.08575, "grad_norm": 2.815019369125366, "grad_norm_var": 9.652987248771876, "learning_rate": 0.0001, "loss": 1.0631, "loss/crossentropy": 2.236358165740967, "loss/hidden": 0.72265625, "loss/logits": 0.1266387552022934, "loss/reg": 0.021379247307777405, "step": 686 }, { "epoch": 0.085875, "grad_norm": 3.6744375228881836, "grad_norm_var": 9.673796390527396, "learning_rate": 0.0001, "loss": 1.2523, "loss/crossentropy": 2.1673667430877686, "loss/hidden": 0.8671875, "loss/logits": 0.1714528650045395, "loss/reg": 0.02136901021003723, "step": 687 }, { "epoch": 0.086, "grad_norm": 3.169265031814575, "grad_norm_var": 9.732675648460468, "learning_rate": 0.0001, "loss": 1.1389, "loss/crossentropy": 2.45385479927063, "loss/hidden": 0.79296875, "loss/logits": 0.13236522674560547, "loss/reg": 0.021359853446483612, "step": 688 }, { "epoch": 0.086125, "grad_norm": 7.017651557922363, "grad_norm_var": 10.2441458601344, "learning_rate": 0.0001, "loss": 1.4071, "loss/crossentropy": 2.344740629196167, "loss/hidden": 1.0625, "loss/logits": 0.13107535243034363, "loss/reg": 0.021349839866161346, "step": 689 }, { "epoch": 0.08625, "grad_norm": 3.5523998737335205, "grad_norm_var": 1.142674868302701, "learning_rate": 0.0001, "loss": 1.1408, "loss/crossentropy": 2.323866367340088, "loss/hidden": 0.79296875, "loss/logits": 0.13442741334438324, "loss/reg": 0.021340306848287582, "step": 690 }, { "epoch": 0.086375, "grad_norm": 3.2477400302886963, "grad_norm_var": 1.1489843436981233, "learning_rate": 0.0001, "loss": 1.0067, "loss/crossentropy": 2.672182083129883, "loss/hidden": 0.67578125, "loss/logits": 0.11757320165634155, "loss/reg": 0.021331045776605606, "step": 691 }, { "epoch": 0.0865, "grad_norm": 3.1238696575164795, "grad_norm_var": 1.1603900529546722, "learning_rate": 0.0001, "loss": 1.1122, "loss/crossentropy": 2.510279893875122, "loss/hidden": 0.765625, "loss/logits": 0.13339656591415405, "loss/reg": 0.021321000531315804, "step": 692 }, { "epoch": 0.086625, "grad_norm": 2.552560329437256, "grad_norm_var": 1.2122975827491755, "learning_rate": 0.0001, "loss": 1.0396, "loss/crossentropy": 2.5303587913513184, "loss/hidden": 0.6953125, "loss/logits": 0.13117440044879913, "loss/reg": 0.021313220262527466, "step": 693 }, { "epoch": 0.08675, "grad_norm": 2.7075002193450928, "grad_norm_var": 1.1997649773340213, "learning_rate": 0.0001, "loss": 1.0787, "loss/crossentropy": 2.4859445095062256, "loss/hidden": 0.73828125, "loss/logits": 0.12738223373889923, "loss/reg": 0.021303845569491386, "step": 694 }, { "epoch": 0.086875, "grad_norm": 3.2640225887298584, "grad_norm_var": 1.1768004922088529, "learning_rate": 0.0001, "loss": 1.3383, "loss/crossentropy": 2.4017555713653564, "loss/hidden": 0.96875, "loss/logits": 0.15659019351005554, "loss/reg": 0.021294469013810158, "step": 695 }, { "epoch": 0.087, "grad_norm": 3.5289857387542725, "grad_norm_var": 1.1683562063707291, "learning_rate": 0.0001, "loss": 1.2607, "loss/crossentropy": 2.2716236114501953, "loss/hidden": 0.87109375, "loss/logits": 0.17679363489151, "loss/reg": 0.021284854039549828, "step": 696 }, { "epoch": 0.087125, "grad_norm": 8.924003601074219, "grad_norm_var": 3.0501856400161764, "learning_rate": 0.0001, "loss": 1.474, "loss/crossentropy": 2.3684239387512207, "loss/hidden": 1.0546875, "loss/logits": 0.2065410166978836, "loss/reg": 0.02127666585147381, "step": 697 }, { "epoch": 0.08725, "grad_norm": 4.2463788986206055, "grad_norm_var": 2.8955696375948605, "learning_rate": 0.0001, "loss": 1.1835, "loss/crossentropy": 2.234137535095215, "loss/hidden": 0.828125, "loss/logits": 0.1427059769630432, "loss/reg": 0.021268585696816444, "step": 698 }, { "epoch": 0.087375, "grad_norm": 3.0776546001434326, "grad_norm_var": 2.903130025314302, "learning_rate": 0.0001, "loss": 1.0559, "loss/crossentropy": 2.4638020992279053, "loss/hidden": 0.7265625, "loss/logits": 0.1167045459151268, "loss/reg": 0.02125934511423111, "step": 699 }, { "epoch": 0.0875, "grad_norm": 3.9374022483825684, "grad_norm_var": 2.843209099820052, "learning_rate": 0.0001, "loss": 0.9629, "loss/crossentropy": 2.88035249710083, "loss/hidden": 0.63671875, "loss/logits": 0.11369955539703369, "loss/reg": 0.021249722689390182, "step": 700 }, { "epoch": 0.087625, "grad_norm": 4.235450744628906, "grad_norm_var": 2.8355032825089417, "learning_rate": 0.0001, "loss": 1.278, "loss/crossentropy": 2.2560718059539795, "loss/hidden": 0.8828125, "loss/logits": 0.1828073114156723, "loss/reg": 0.0212401133030653, "step": 701 }, { "epoch": 0.08775, "grad_norm": 2.3705031871795654, "grad_norm_var": 2.9146564397348773, "learning_rate": 0.0001, "loss": 0.9899, "loss/crossentropy": 2.531632661819458, "loss/hidden": 0.64453125, "loss/logits": 0.13305732607841492, "loss/reg": 0.021231388673186302, "step": 702 }, { "epoch": 0.087875, "grad_norm": 3.5063636302948, "grad_norm_var": 2.921798711310286, "learning_rate": 0.0001, "loss": 1.0524, "loss/crossentropy": 2.548243999481201, "loss/hidden": 0.72265625, "loss/logits": 0.11747653782367706, "loss/reg": 0.02122276835143566, "step": 703 }, { "epoch": 0.088, "grad_norm": 3.141618013381958, "grad_norm_var": 2.924554396554724, "learning_rate": 0.0001, "loss": 0.9087, "loss/crossentropy": 2.3723928928375244, "loss/hidden": 0.59375, "loss/logits": 0.10285645723342896, "loss/reg": 0.021213354542851448, "step": 704 }, { "epoch": 0.088125, "grad_norm": 4.195517063140869, "grad_norm_var": 2.2500098957229806, "learning_rate": 0.0001, "loss": 1.1263, "loss/crossentropy": 2.383502244949341, "loss/hidden": 0.765625, "loss/logits": 0.14867368340492249, "loss/reg": 0.021204529330134392, "step": 705 }, { "epoch": 0.08825, "grad_norm": 3.646482467651367, "grad_norm_var": 2.2483885758775686, "learning_rate": 0.0001, "loss": 1.1459, "loss/crossentropy": 2.239716053009033, "loss/hidden": 0.79296875, "loss/logits": 0.14099720120429993, "loss/reg": 0.021195242181420326, "step": 706 }, { "epoch": 0.088375, "grad_norm": 3.3426883220672607, "grad_norm_var": 2.242826109053838, "learning_rate": 0.0001, "loss": 1.0323, "loss/crossentropy": 2.270444869995117, "loss/hidden": 0.7109375, "loss/logits": 0.10951918363571167, "loss/reg": 0.021186839789152145, "step": 707 }, { "epoch": 0.0885, "grad_norm": 3.083806037902832, "grad_norm_var": 2.2462046620558, "learning_rate": 0.0001, "loss": 1.086, "loss/crossentropy": 2.5013253688812256, "loss/hidden": 0.75390625, "loss/logits": 0.12026840448379517, "loss/reg": 0.021178435534238815, "step": 708 }, { "epoch": 0.088625, "grad_norm": 3.71588134765625, "grad_norm_var": 2.147370219186798, "learning_rate": 0.0001, "loss": 1.1125, "loss/crossentropy": 2.342475175857544, "loss/hidden": 0.76953125, "loss/logits": 0.13124068081378937, "loss/reg": 0.021169869229197502, "step": 709 }, { "epoch": 0.08875, "grad_norm": 3.4402713775634766, "grad_norm_var": 2.0734307300644255, "learning_rate": 0.0001, "loss": 1.1229, "loss/crossentropy": 2.4113690853118896, "loss/hidden": 0.76171875, "loss/logits": 0.14960095286369324, "loss/reg": 0.021161576732993126, "step": 710 }, { "epoch": 0.088875, "grad_norm": 10.640914916992188, "grad_norm_var": 4.894724677276531, "learning_rate": 0.0001, "loss": 1.7804, "loss/crossentropy": 2.6555004119873047, "loss/hidden": 1.328125, "loss/logits": 0.240725576877594, "loss/reg": 0.02115357480943203, "step": 711 }, { "epoch": 0.089, "grad_norm": 3.23919415473938, "grad_norm_var": 4.930329406483421, "learning_rate": 0.0001, "loss": 1.1368, "loss/crossentropy": 2.4652106761932373, "loss/hidden": 0.77734375, "loss/logits": 0.14798855781555176, "loss/reg": 0.021144360303878784, "step": 712 }, { "epoch": 0.089125, "grad_norm": 2.8362622261047363, "grad_norm_var": 3.4904838717428524, "learning_rate": 0.0001, "loss": 1.0668, "loss/crossentropy": 2.216999053955078, "loss/hidden": 0.7265625, "loss/logits": 0.1288391500711441, "loss/reg": 0.021136239171028137, "step": 713 }, { "epoch": 0.08925, "grad_norm": 5.395811080932617, "grad_norm_var": 3.623687874884585, "learning_rate": 0.0001, "loss": 1.2022, "loss/crossentropy": 2.463663101196289, "loss/hidden": 0.82421875, "loss/logits": 0.1667168289422989, "loss/reg": 0.02112707309424877, "step": 714 }, { "epoch": 0.089375, "grad_norm": 3.475656032562256, "grad_norm_var": 3.585286252049487, "learning_rate": 0.0001, "loss": 1.1524, "loss/crossentropy": 2.4688560962677, "loss/hidden": 0.796875, "loss/logits": 0.14434993267059326, "loss/reg": 0.02111782319843769, "step": 715 }, { "epoch": 0.0895, "grad_norm": 4.29054594039917, "grad_norm_var": 3.5895333664829043, "learning_rate": 0.0001, "loss": 1.1214, "loss/crossentropy": 2.4820876121520996, "loss/hidden": 0.7578125, "loss/logits": 0.15247562527656555, "loss/reg": 0.021108638495206833, "step": 716 }, { "epoch": 0.089625, "grad_norm": 2.927002429962158, "grad_norm_var": 3.661532010616088, "learning_rate": 0.0001, "loss": 1.0904, "loss/crossentropy": 2.499390125274658, "loss/hidden": 0.734375, "loss/logits": 0.14508014917373657, "loss/reg": 0.021099381148815155, "step": 717 }, { "epoch": 0.08975, "grad_norm": 2.529557943344116, "grad_norm_var": 3.629551988733732, "learning_rate": 0.0001, "loss": 1.1194, "loss/crossentropy": 2.395061492919922, "loss/hidden": 0.76953125, "loss/logits": 0.13898837566375732, "loss/reg": 0.021089982241392136, "step": 718 }, { "epoch": 0.089875, "grad_norm": 3.2681446075439453, "grad_norm_var": 3.6476018392648397, "learning_rate": 0.0001, "loss": 1.0036, "loss/crossentropy": 2.65079927444458, "loss/hidden": 0.6640625, "loss/logits": 0.12875254452228546, "loss/reg": 0.021080130711197853, "step": 719 }, { "epoch": 0.09, "grad_norm": 2.6700246334075928, "grad_norm_var": 3.7122117675621022, "learning_rate": 0.0001, "loss": 0.9929, "loss/crossentropy": 2.428039789199829, "loss/hidden": 0.65234375, "loss/logits": 0.12986770272254944, "loss/reg": 0.021070368587970734, "step": 720 }, { "epoch": 0.090125, "grad_norm": 3.296482563018799, "grad_norm_var": 3.7295350110356558, "learning_rate": 0.0001, "loss": 1.1571, "loss/crossentropy": 2.317190408706665, "loss/hidden": 0.80078125, "loss/logits": 0.14570315182209015, "loss/reg": 0.02106117643415928, "step": 721 }, { "epoch": 0.09025, "grad_norm": 4.451722145080566, "grad_norm_var": 3.74687645800365, "learning_rate": 0.0001, "loss": 1.1385, "loss/crossentropy": 2.2958080768585205, "loss/hidden": 0.78125, "loss/logits": 0.146757572889328, "loss/reg": 0.021051928400993347, "step": 722 }, { "epoch": 0.090375, "grad_norm": 2.4288058280944824, "grad_norm_var": 3.8685376080960414, "learning_rate": 0.0001, "loss": 0.9788, "loss/crossentropy": 2.4692001342773438, "loss/hidden": 0.65234375, "loss/logits": 0.11606692522764206, "loss/reg": 0.021042969077825546, "step": 723 }, { "epoch": 0.0905, "grad_norm": 45.21147918701172, "grad_norm_var": 110.45448625779909, "learning_rate": 0.0001, "loss": 1.0823, "loss/crossentropy": 2.4428551197052, "loss/hidden": 0.7421875, "loss/logits": 0.12983539700508118, "loss/reg": 0.021031970158219337, "step": 724 }, { "epoch": 0.090625, "grad_norm": 3.7111551761627197, "grad_norm_var": 110.45623490585022, "learning_rate": 0.0001, "loss": 1.4512, "loss/crossentropy": 2.1149277687072754, "loss/hidden": 1.046875, "loss/logits": 0.1941366195678711, "loss/reg": 0.021020574495196342, "step": 725 }, { "epoch": 0.09075, "grad_norm": 3.3139920234680176, "grad_norm_var": 110.50855221427315, "learning_rate": 0.0001, "loss": 1.0686, "loss/crossentropy": 2.798910140991211, "loss/hidden": 0.71875, "loss/logits": 0.13971024751663208, "loss/reg": 0.021009519696235657, "step": 726 }, { "epoch": 0.090875, "grad_norm": 2.5206596851348877, "grad_norm_var": 110.12514261997971, "learning_rate": 0.0001, "loss": 1.0391, "loss/crossentropy": 2.400813579559326, "loss/hidden": 0.6953125, "loss/logits": 0.13384617865085602, "loss/reg": 0.020998528227210045, "step": 727 }, { "epoch": 0.091, "grad_norm": 2.698018789291382, "grad_norm_var": 110.34070270952832, "learning_rate": 0.0001, "loss": 0.928, "loss/crossentropy": 2.505309820175171, "loss/hidden": 0.61328125, "loss/logits": 0.10482652485370636, "loss/reg": 0.020989248529076576, "step": 728 }, { "epoch": 0.091125, "grad_norm": 6.086211204528809, "grad_norm_var": 109.65630388036335, "learning_rate": 0.0001, "loss": 1.4543, "loss/crossentropy": 2.065880537033081, "loss/hidden": 1.1171875, "loss/logits": 0.12730881571769714, "loss/reg": 0.020978538319468498, "step": 729 }, { "epoch": 0.09125, "grad_norm": 2.527377128601074, "grad_norm_var": 110.45601242879228, "learning_rate": 0.0001, "loss": 0.9281, "loss/crossentropy": 2.575125217437744, "loss/hidden": 0.60546875, "loss/logits": 0.11296658217906952, "loss/reg": 0.020967954769730568, "step": 730 }, { "epoch": 0.091375, "grad_norm": 2.405383825302124, "grad_norm_var": 110.88254605251709, "learning_rate": 0.0001, "loss": 1.0005, "loss/crossentropy": 2.1985392570495605, "loss/hidden": 0.671875, "loss/logits": 0.1190965548157692, "loss/reg": 0.02095715142786503, "step": 731 }, { "epoch": 0.0915, "grad_norm": 5.095945835113525, "grad_norm_var": 110.75067974759948, "learning_rate": 0.0001, "loss": 1.2535, "loss/crossentropy": 2.451446056365967, "loss/hidden": 0.88671875, "loss/logits": 0.15732397139072418, "loss/reg": 0.020946422591805458, "step": 732 }, { "epoch": 0.091625, "grad_norm": 2.497173547744751, "grad_norm_var": 110.93526847423995, "learning_rate": 0.0001, "loss": 1.1301, "loss/crossentropy": 2.4981319904327393, "loss/hidden": 0.76953125, "loss/logits": 0.15116086602210999, "loss/reg": 0.020936597138643265, "step": 733 }, { "epoch": 0.09175, "grad_norm": 4.977592468261719, "grad_norm_var": 110.20332761050602, "learning_rate": 0.0001, "loss": 1.0306, "loss/crossentropy": 2.37424898147583, "loss/hidden": 0.703125, "loss/logits": 0.11818103492259979, "loss/reg": 0.020926134660840034, "step": 734 }, { "epoch": 0.091875, "grad_norm": 2.8039772510528564, "grad_norm_var": 110.39035266849663, "learning_rate": 0.0001, "loss": 0.9389, "loss/crossentropy": 2.5503880977630615, "loss/hidden": 0.62109375, "loss/logits": 0.1085958182811737, "loss/reg": 0.02091672271490097, "step": 735 }, { "epoch": 0.092, "grad_norm": 3.418717861175537, "grad_norm_var": 110.08862675247053, "learning_rate": 0.0001, "loss": 0.9544, "loss/crossentropy": 2.6695592403411865, "loss/hidden": 0.62890625, "loss/logits": 0.11641087383031845, "loss/reg": 0.02090657874941826, "step": 736 }, { "epoch": 0.092125, "grad_norm": 2.807041883468628, "grad_norm_var": 110.28591938740921, "learning_rate": 0.0001, "loss": 1.1135, "loss/crossentropy": 2.4328622817993164, "loss/hidden": 0.76953125, "loss/logits": 0.13500146567821503, "loss/reg": 0.02089635282754898, "step": 737 }, { "epoch": 0.09225, "grad_norm": 3.713057518005371, "grad_norm_var": 110.4783888232826, "learning_rate": 0.0001, "loss": 1.1755, "loss/crossentropy": 2.0374624729156494, "loss/hidden": 0.83984375, "loss/logits": 0.12679257988929749, "loss/reg": 0.02088700234889984, "step": 738 }, { "epoch": 0.092375, "grad_norm": 3.618948459625244, "grad_norm_var": 109.99807079993953, "learning_rate": 0.0001, "loss": 0.9255, "loss/crossentropy": 2.536527395248413, "loss/hidden": 0.609375, "loss/logits": 0.10737244784832001, "loss/reg": 0.020877836272120476, "step": 739 }, { "epoch": 0.0925, "grad_norm": 2.815113067626953, "grad_norm_var": 1.1792510177041378, "learning_rate": 0.0001, "loss": 0.9755, "loss/crossentropy": 2.464996337890625, "loss/hidden": 0.640625, "loss/logits": 0.12616491317749023, "loss/reg": 0.020868681371212006, "step": 740 }, { "epoch": 0.092625, "grad_norm": 3.194117546081543, "grad_norm_var": 1.1771383378848062, "learning_rate": 0.0001, "loss": 1.1989, "loss/crossentropy": 2.325310707092285, "loss/hidden": 0.84375, "loss/logits": 0.14658081531524658, "loss/reg": 0.020859118551015854, "step": 741 }, { "epoch": 0.09275, "grad_norm": 2.970301628112793, "grad_norm_var": 1.1887296793511715, "learning_rate": 0.0001, "loss": 0.9939, "loss/crossentropy": 2.4747390747070312, "loss/hidden": 0.66015625, "loss/logits": 0.1252739280462265, "loss/reg": 0.020850006490945816, "step": 742 }, { "epoch": 0.092875, "grad_norm": 3.7745604515075684, "grad_norm_var": 1.1425983881417718, "learning_rate": 0.0001, "loss": 0.9835, "loss/crossentropy": 2.4191653728485107, "loss/hidden": 0.67578125, "loss/logits": 0.09930374473333359, "loss/reg": 0.020840618759393692, "step": 743 }, { "epoch": 0.093, "grad_norm": 2.8048055171966553, "grad_norm_var": 1.132423092522494, "learning_rate": 0.0001, "loss": 1.083, "loss/crossentropy": 2.764143228530884, "loss/hidden": 0.75390625, "loss/logits": 0.12075912207365036, "loss/reg": 0.020830942317843437, "step": 744 }, { "epoch": 0.093125, "grad_norm": 9.04628849029541, "grad_norm_var": 2.712848654929009, "learning_rate": 0.0001, "loss": 1.4779, "loss/crossentropy": 2.2979469299316406, "loss/hidden": 1.125, "loss/logits": 0.1446615606546402, "loss/reg": 0.02082117274403572, "step": 745 }, { "epoch": 0.09325, "grad_norm": 3.1999406814575195, "grad_norm_var": 2.6400540651182967, "learning_rate": 0.0001, "loss": 1.0424, "loss/crossentropy": 2.6157429218292236, "loss/hidden": 0.71875, "loss/logits": 0.11555971205234528, "loss/reg": 0.020811092108488083, "step": 746 }, { "epoch": 0.093375, "grad_norm": 3.6833674907684326, "grad_norm_var": 2.522139333113606, "learning_rate": 0.0001, "loss": 1.1046, "loss/crossentropy": 2.6110498905181885, "loss/hidden": 0.7578125, "loss/logits": 0.13881847262382507, "loss/reg": 0.020801017060875893, "step": 747 }, { "epoch": 0.0935, "grad_norm": 3.073922872543335, "grad_norm_var": 2.4218973518929827, "learning_rate": 0.0001, "loss": 0.9672, "loss/crossentropy": 2.0898826122283936, "loss/hidden": 0.65625, "loss/logits": 0.10303568840026855, "loss/reg": 0.020791731774806976, "step": 748 }, { "epoch": 0.093625, "grad_norm": 2.9481358528137207, "grad_norm_var": 2.365294319547022, "learning_rate": 0.0001, "loss": 0.9589, "loss/crossentropy": 2.477987766265869, "loss/hidden": 0.640625, "loss/logits": 0.11042475700378418, "loss/reg": 0.02078239433467388, "step": 749 }, { "epoch": 0.09375, "grad_norm": 5.792114734649658, "grad_norm_var": 2.5478865053407236, "learning_rate": 0.0001, "loss": 1.2395, "loss/crossentropy": 2.0092225074768066, "loss/hidden": 0.87109375, "loss/logits": 0.16063663363456726, "loss/reg": 0.020772725343704224, "step": 750 }, { "epoch": 0.093875, "grad_norm": 3.148350954055786, "grad_norm_var": 2.512823601683457, "learning_rate": 0.0001, "loss": 1.14, "loss/crossentropy": 2.669532537460327, "loss/hidden": 0.78125, "loss/logits": 0.15109741687774658, "loss/reg": 0.020762871950864792, "step": 751 }, { "epoch": 0.094, "grad_norm": 3.2779595851898193, "grad_norm_var": 2.5202896391695124, "learning_rate": 0.0001, "loss": 1.1492, "loss/crossentropy": 2.477522373199463, "loss/hidden": 0.8125, "loss/logits": 0.1291784942150116, "loss/reg": 0.020753389224410057, "step": 752 }, { "epoch": 0.094125, "grad_norm": 2.3718457221984863, "grad_norm_var": 2.586364485192132, "learning_rate": 0.0001, "loss": 1.0399, "loss/crossentropy": 2.497688055038452, "loss/hidden": 0.69921875, "loss/logits": 0.13327988982200623, "loss/reg": 0.02074403502047062, "step": 753 }, { "epoch": 0.09425, "grad_norm": 4.972538948059082, "grad_norm_var": 2.6852568725766104, "learning_rate": 0.0001, "loss": 1.2511, "loss/crossentropy": 2.232241630554199, "loss/hidden": 0.90234375, "loss/logits": 0.1414394974708557, "loss/reg": 0.020734604448080063, "step": 754 }, { "epoch": 0.094375, "grad_norm": 2.6991426944732666, "grad_norm_var": 2.7595134043336267, "learning_rate": 0.0001, "loss": 1.001, "loss/crossentropy": 2.596348285675049, "loss/hidden": 0.6640625, "loss/logits": 0.1297152191400528, "loss/reg": 0.020725268870592117, "step": 755 }, { "epoch": 0.0945, "grad_norm": 2.8633017539978027, "grad_norm_var": 2.753743097466793, "learning_rate": 0.0001, "loss": 1.065, "loss/crossentropy": 2.3198070526123047, "loss/hidden": 0.74609375, "loss/logits": 0.11169925332069397, "loss/reg": 0.020716087892651558, "step": 756 }, { "epoch": 0.094625, "grad_norm": 3.628239154815674, "grad_norm_var": 2.733994536045853, "learning_rate": 0.0001, "loss": 1.3141, "loss/crossentropy": 2.1950411796569824, "loss/hidden": 0.92578125, "loss/logits": 0.1812862753868103, "loss/reg": 0.020707255229353905, "step": 757 }, { "epoch": 0.09475, "grad_norm": 2.805727958679199, "grad_norm_var": 2.753145827217212, "learning_rate": 0.0001, "loss": 1.109, "loss/crossentropy": 2.40027117729187, "loss/hidden": 0.76953125, "loss/logits": 0.13251210749149323, "loss/reg": 0.02069801278412342, "step": 758 }, { "epoch": 0.094875, "grad_norm": 3.1892051696777344, "grad_norm_var": 2.7730842000576295, "learning_rate": 0.0001, "loss": 1.1224, "loss/crossentropy": 2.2343788146972656, "loss/hidden": 0.77734375, "loss/logits": 0.1382053792476654, "loss/reg": 0.020689615979790688, "step": 759 }, { "epoch": 0.095, "grad_norm": 2.925913095474243, "grad_norm_var": 2.759237877311041, "learning_rate": 0.0001, "loss": 0.9857, "loss/crossentropy": 2.5283889770507812, "loss/hidden": 0.6640625, "loss/logits": 0.11487281322479248, "loss/reg": 0.02068025805056095, "step": 760 }, { "epoch": 0.095125, "grad_norm": 4.760406970977783, "grad_norm_var": 0.8673601536624308, "learning_rate": 0.0001, "loss": 1.3807, "loss/crossentropy": 2.8629026412963867, "loss/hidden": 0.9765625, "loss/logits": 0.1973596215248108, "loss/reg": 0.020673030987381935, "step": 761 }, { "epoch": 0.09525, "grad_norm": 2.6182596683502197, "grad_norm_var": 0.9085803501248163, "learning_rate": 0.0001, "loss": 0.9554, "loss/crossentropy": 2.6935925483703613, "loss/hidden": 0.6328125, "loss/logits": 0.11598716676235199, "loss/reg": 0.020663931965827942, "step": 762 }, { "epoch": 0.095375, "grad_norm": 3.404240131378174, "grad_norm_var": 0.9037375089777697, "learning_rate": 0.0001, "loss": 1.0173, "loss/crossentropy": 2.5958027839660645, "loss/hidden": 0.6953125, "loss/logits": 0.11542315781116486, "loss/reg": 0.020654823631048203, "step": 763 }, { "epoch": 0.0955, "grad_norm": 3.0021934509277344, "grad_norm_var": 0.9072250591900151, "learning_rate": 0.0001, "loss": 0.9811, "loss/crossentropy": 2.4611666202545166, "loss/hidden": 0.6640625, "loss/logits": 0.11060373485088348, "loss/reg": 0.020645687356591225, "step": 764 }, { "epoch": 0.095625, "grad_norm": 4.123987674713135, "grad_norm_var": 0.9227216736856043, "learning_rate": 0.0001, "loss": 1.2112, "loss/crossentropy": 2.327146530151367, "loss/hidden": 0.87109375, "loss/logits": 0.13372407853603363, "loss/reg": 0.020637821406126022, "step": 765 }, { "epoch": 0.09575, "grad_norm": 3.835116147994995, "grad_norm_var": 0.5572045887439032, "learning_rate": 0.0001, "loss": 1.0062, "loss/crossentropy": 2.5438876152038574, "loss/hidden": 0.67578125, "loss/logits": 0.12415439635515213, "loss/reg": 0.02063015103340149, "step": 766 }, { "epoch": 0.095875, "grad_norm": 3.2649309635162354, "grad_norm_var": 0.5548939110280107, "learning_rate": 0.0001, "loss": 1.1664, "loss/crossentropy": 2.6621615886688232, "loss/hidden": 0.80078125, "loss/logits": 0.15938454866409302, "loss/reg": 0.020622732117772102, "step": 767 }, { "epoch": 0.096, "grad_norm": 3.398061752319336, "grad_norm_var": 0.554498685348062, "learning_rate": 0.0001, "loss": 0.852, "loss/crossentropy": 2.3231897354125977, "loss/hidden": 0.5625, "loss/logits": 0.08329755067825317, "loss/reg": 0.0206154715269804, "step": 768 }, { "epoch": 0.096125, "grad_norm": 2.8744146823883057, "grad_norm_var": 0.5036373977995244, "learning_rate": 0.0001, "loss": 0.9094, "loss/crossentropy": 2.5140726566314697, "loss/hidden": 0.6171875, "loss/logits": 0.0860939770936966, "loss/reg": 0.020607706159353256, "step": 769 }, { "epoch": 0.09625, "grad_norm": 8.404803276062012, "grad_norm_var": 1.9605456650252857, "learning_rate": 0.0001, "loss": 1.614, "loss/crossentropy": 2.4072842597961426, "loss/hidden": 1.1328125, "loss/logits": 0.2751774787902832, "loss/reg": 0.020600339397788048, "step": 770 }, { "epoch": 0.096375, "grad_norm": 2.581511974334717, "grad_norm_var": 1.975733645477993, "learning_rate": 0.0001, "loss": 1.08, "loss/crossentropy": 2.3724427223205566, "loss/hidden": 0.73828125, "loss/logits": 0.1358003169298172, "loss/reg": 0.02059323526918888, "step": 771 }, { "epoch": 0.0965, "grad_norm": 3.9817845821380615, "grad_norm_var": 1.9433082266341482, "learning_rate": 0.0001, "loss": 1.0926, "loss/crossentropy": 2.459740400314331, "loss/hidden": 0.765625, "loss/logits": 0.12109124660491943, "loss/reg": 0.02058546058833599, "step": 772 }, { "epoch": 0.096625, "grad_norm": 7.001491069793701, "grad_norm_var": 2.633487351928299, "learning_rate": 0.0001, "loss": 1.1253, "loss/crossentropy": 2.6342828273773193, "loss/hidden": 0.8125, "loss/logits": 0.10705184936523438, "loss/reg": 0.020576275885105133, "step": 773 }, { "epoch": 0.09675, "grad_norm": 2.4826319217681885, "grad_norm_var": 2.6865387021083524, "learning_rate": 0.0001, "loss": 1.1239, "loss/crossentropy": 2.537883996963501, "loss/hidden": 0.7734375, "loss/logits": 0.1447601616382599, "loss/reg": 0.02056770585477352, "step": 774 }, { "epoch": 0.096875, "grad_norm": 3.5470290184020996, "grad_norm_var": 2.6622723084153237, "learning_rate": 0.0001, "loss": 1.1572, "loss/crossentropy": 2.572312831878662, "loss/hidden": 0.80078125, "loss/logits": 0.15083444118499756, "loss/reg": 0.020558428019285202, "step": 775 }, { "epoch": 0.097, "grad_norm": 2.8388142585754395, "grad_norm_var": 2.673918444962514, "learning_rate": 0.0001, "loss": 1.1426, "loss/crossentropy": 2.2697503566741943, "loss/hidden": 0.79296875, "loss/logits": 0.14412574470043182, "loss/reg": 0.020549749955534935, "step": 776 }, { "epoch": 0.097125, "grad_norm": 2.9166338443756104, "grad_norm_var": 2.670560695292122, "learning_rate": 0.0001, "loss": 0.8926, "loss/crossentropy": 2.4288480281829834, "loss/hidden": 0.58203125, "loss/logits": 0.10516718029975891, "loss/reg": 0.02054043672978878, "step": 777 }, { "epoch": 0.09725, "grad_norm": 3.565744161605835, "grad_norm_var": 2.58151597609506, "learning_rate": 0.0001, "loss": 1.1299, "loss/crossentropy": 2.444842576980591, "loss/hidden": 0.77734375, "loss/logits": 0.1472683846950531, "loss/reg": 0.020531047135591507, "step": 778 }, { "epoch": 0.097375, "grad_norm": 2.8829805850982666, "grad_norm_var": 2.627842889624617, "learning_rate": 0.0001, "loss": 1.1382, "loss/crossentropy": 2.548234462738037, "loss/hidden": 0.78515625, "loss/logits": 0.1478501409292221, "loss/reg": 0.02052178978919983, "step": 779 }, { "epoch": 0.0975, "grad_norm": 2.908299207687378, "grad_norm_var": 2.6383052442278556, "learning_rate": 0.0001, "loss": 1.069, "loss/crossentropy": 2.3388137817382812, "loss/hidden": 0.734375, "loss/logits": 0.12949424982070923, "loss/reg": 0.020512979477643967, "step": 780 }, { "epoch": 0.097625, "grad_norm": 2.5852208137512207, "grad_norm_var": 2.717361748364273, "learning_rate": 0.0001, "loss": 0.9782, "loss/crossentropy": 2.4337894916534424, "loss/hidden": 0.65625, "loss/logits": 0.11689651757478714, "loss/reg": 0.020503604784607887, "step": 781 }, { "epoch": 0.09775, "grad_norm": 5.328097820281982, "grad_norm_var": 2.885194693951976, "learning_rate": 0.0001, "loss": 1.2053, "loss/crossentropy": 2.5564181804656982, "loss/hidden": 0.80078125, "loss/logits": 0.19961750507354736, "loss/reg": 0.020494818687438965, "step": 782 }, { "epoch": 0.097875, "grad_norm": 3.647054672241211, "grad_norm_var": 2.8678156226553613, "learning_rate": 0.0001, "loss": 1.0039, "loss/crossentropy": 2.5830650329589844, "loss/hidden": 0.66796875, "loss/logits": 0.13103139400482178, "loss/reg": 0.020485466346144676, "step": 783 }, { "epoch": 0.098, "grad_norm": 4.480771541595459, "grad_norm_var": 2.8817531456144723, "learning_rate": 0.0001, "loss": 1.3811, "loss/crossentropy": 2.4550743103027344, "loss/hidden": 0.9765625, "loss/logits": 0.1997724324464798, "loss/reg": 0.020477164536714554, "step": 784 }, { "epoch": 0.098125, "grad_norm": 3.1495656967163086, "grad_norm_var": 2.8497140664534366, "learning_rate": 0.0001, "loss": 1.1386, "loss/crossentropy": 2.846536636352539, "loss/hidden": 0.76171875, "loss/logits": 0.1722334325313568, "loss/reg": 0.020469149574637413, "step": 785 }, { "epoch": 0.09825, "grad_norm": 3.611919641494751, "grad_norm_var": 1.4027508562338329, "learning_rate": 0.0001, "loss": 1.0783, "loss/crossentropy": 2.7328994274139404, "loss/hidden": 0.73046875, "loss/logits": 0.14321856200695038, "loss/reg": 0.020459884777665138, "step": 786 }, { "epoch": 0.098375, "grad_norm": 4.535567283630371, "grad_norm_var": 1.3775118805215696, "learning_rate": 0.0001, "loss": 1.3035, "loss/crossentropy": 2.616641044616699, "loss/hidden": 0.8515625, "loss/logits": 0.24746158719062805, "loss/reg": 0.020450593903660774, "step": 787 }, { "epoch": 0.0985, "grad_norm": 2.7010657787323, "grad_norm_var": 1.4347220572574786, "learning_rate": 0.0001, "loss": 0.9439, "loss/crossentropy": 2.7851712703704834, "loss/hidden": 0.62890625, "loss/logits": 0.11055716872215271, "loss/reg": 0.020441319793462753, "step": 788 }, { "epoch": 0.098625, "grad_norm": 3.3690221309661865, "grad_norm_var": 0.6296018822433316, "learning_rate": 0.0001, "loss": 0.9858, "loss/crossentropy": 2.5199151039123535, "loss/hidden": 0.6640625, "loss/logits": 0.11739970743656158, "loss/reg": 0.020432572811841965, "step": 789 }, { "epoch": 0.09875, "grad_norm": 3.5332016944885254, "grad_norm_var": 0.5687648370759459, "learning_rate": 0.0001, "loss": 1.2114, "loss/crossentropy": 2.3861191272735596, "loss/hidden": 0.80859375, "loss/logits": 0.19853942096233368, "loss/reg": 0.020423252135515213, "step": 790 }, { "epoch": 0.098875, "grad_norm": 3.4778192043304443, "grad_norm_var": 0.5684000998912779, "learning_rate": 0.0001, "loss": 1.3337, "loss/crossentropy": 2.177149772644043, "loss/hidden": 0.94921875, "loss/logits": 0.18037351965904236, "loss/reg": 0.020413951948285103, "step": 791 }, { "epoch": 0.099, "grad_norm": 3.1103479862213135, "grad_norm_var": 0.5501298461305394, "learning_rate": 0.0001, "loss": 1.1811, "loss/crossentropy": 2.381289005279541, "loss/hidden": 0.8125, "loss/logits": 0.16453775763511658, "loss/reg": 0.020404649898409843, "step": 792 }, { "epoch": 0.099125, "grad_norm": 2.7654690742492676, "grad_norm_var": 0.563068172749172, "learning_rate": 0.0001, "loss": 0.9788, "loss/crossentropy": 2.6645448207855225, "loss/hidden": 0.65234375, "loss/logits": 0.12248219549655914, "loss/reg": 0.020395854488015175, "step": 793 }, { "epoch": 0.09925, "grad_norm": 2.9942195415496826, "grad_norm_var": 0.5768165563916947, "learning_rate": 0.0001, "loss": 1.1054, "loss/crossentropy": 2.388904571533203, "loss/hidden": 0.73828125, "loss/logits": 0.16325941681861877, "loss/reg": 0.020387381315231323, "step": 794 }, { "epoch": 0.099375, "grad_norm": 3.1070520877838135, "grad_norm_var": 0.5632370819485725, "learning_rate": 0.0001, "loss": 1.0819, "loss/crossentropy": 2.4645302295684814, "loss/hidden": 0.74609375, "loss/logits": 0.13203924894332886, "loss/reg": 0.020378144457936287, "step": 795 }, { "epoch": 0.0995, "grad_norm": 3.5182595252990723, "grad_norm_var": 0.5419026805153273, "learning_rate": 0.0001, "loss": 1.3242, "loss/crossentropy": 2.337388038635254, "loss/hidden": 0.96875, "loss/logits": 0.15180304646492004, "loss/reg": 0.020368557423353195, "step": 796 }, { "epoch": 0.099625, "grad_norm": 2.4217121601104736, "grad_norm_var": 0.5634005753459926, "learning_rate": 0.0001, "loss": 0.976, "loss/crossentropy": 2.520176887512207, "loss/hidden": 0.6640625, "loss/logits": 0.1083153486251831, "loss/reg": 0.020359758287668228, "step": 797 }, { "epoch": 0.09975, "grad_norm": 3.818143129348755, "grad_norm_var": 0.33472096860269646, "learning_rate": 0.0001, "loss": 1.2273, "loss/crossentropy": 2.452592134475708, "loss/hidden": 0.86328125, "loss/logits": 0.1604856252670288, "loss/reg": 0.02035023830831051, "step": 798 }, { "epoch": 0.099875, "grad_norm": 2.8298075199127197, "grad_norm_var": 0.3484620943588491, "learning_rate": 0.0001, "loss": 1.094, "loss/crossentropy": 2.6919286251068115, "loss/hidden": 0.73828125, "loss/logits": 0.15227138996124268, "loss/reg": 0.020340625196695328, "step": 799 }, { "epoch": 0.1, "grad_norm": 45.50175094604492, "grad_norm_var": 111.76340644728633, "learning_rate": 0.0001, "loss": 1.6213, "loss/crossentropy": 2.2076241970062256, "loss/hidden": 1.125, "loss/logits": 0.2930048406124115, "loss/reg": 0.020331410691142082, "step": 800 }, { "epoch": 0.100125, "grad_norm": 3.073981523513794, "grad_norm_var": 111.7915103772579, "learning_rate": 0.0001, "loss": 1.1146, "loss/crossentropy": 2.5723347663879395, "loss/hidden": 0.77734375, "loss/logits": 0.13403840363025665, "loss/reg": 0.020322071388363838, "step": 801 }, { "epoch": 0.10025, "grad_norm": 4.337286472320557, "grad_norm_var": 111.60328751499571, "learning_rate": 0.0001, "loss": 1.2228, "loss/crossentropy": 2.7784037590026855, "loss/hidden": 0.83984375, "loss/logits": 0.17983263731002808, "loss/reg": 0.020312372595071793, "step": 802 }, { "epoch": 0.100375, "grad_norm": 3.656367778778076, "grad_norm_var": 111.81663718658596, "learning_rate": 0.0001, "loss": 1.1003, "loss/crossentropy": 2.4638426303863525, "loss/hidden": 0.76953125, "loss/logits": 0.12776704132556915, "loss/reg": 0.020302986726164818, "step": 803 }, { "epoch": 0.1005, "grad_norm": 3.38935923576355, "grad_norm_var": 111.55373057700994, "learning_rate": 0.0001, "loss": 0.9749, "loss/crossentropy": 2.4947969913482666, "loss/hidden": 0.6640625, "loss/logits": 0.10794391483068466, "loss/reg": 0.020293867215514183, "step": 804 }, { "epoch": 0.100625, "grad_norm": 3.2775380611419678, "grad_norm_var": 111.58551029522327, "learning_rate": 0.0001, "loss": 1.133, "loss/crossentropy": 2.4812419414520264, "loss/hidden": 0.77734375, "loss/logits": 0.15277597308158875, "loss/reg": 0.020284701138734818, "step": 805 }, { "epoch": 0.10075, "grad_norm": 2.996157646179199, "grad_norm_var": 111.77485823890761, "learning_rate": 0.0001, "loss": 1.0998, "loss/crossentropy": 2.516984224319458, "loss/hidden": 0.75, "loss/logits": 0.14709463715553284, "loss/reg": 0.020275365561246872, "step": 806 }, { "epoch": 0.100875, "grad_norm": 5.575490951538086, "grad_norm_var": 111.37459403701224, "learning_rate": 0.0001, "loss": 1.0629, "loss/crossentropy": 2.6265506744384766, "loss/hidden": 0.73828125, "loss/logits": 0.12198775261640549, "loss/reg": 0.02026602067053318, "step": 807 }, { "epoch": 0.101, "grad_norm": 2.653712749481201, "grad_norm_var": 111.56498102164149, "learning_rate": 0.0001, "loss": 0.957, "loss/crossentropy": 2.755121946334839, "loss/hidden": 0.6328125, "loss/logits": 0.12166983634233475, "loss/reg": 0.02025618776679039, "step": 808 }, { "epoch": 0.101125, "grad_norm": 2.5454790592193604, "grad_norm_var": 111.66272758702647, "learning_rate": 0.0001, "loss": 1.0209, "loss/crossentropy": 2.475360870361328, "loss/hidden": 0.6953125, "loss/logits": 0.1231006383895874, "loss/reg": 0.02024705521762371, "step": 809 }, { "epoch": 0.10125, "grad_norm": 12.002355575561523, "grad_norm_var": 113.1469842386682, "learning_rate": 0.0001, "loss": 1.1289, "loss/crossentropy": 2.485873222351074, "loss/hidden": 0.77734375, "loss/logits": 0.1492297500371933, "loss/reg": 0.020237451419234276, "step": 810 }, { "epoch": 0.101375, "grad_norm": 3.0118675231933594, "grad_norm_var": 113.19117010752396, "learning_rate": 0.0001, "loss": 1.0673, "loss/crossentropy": 2.610520124435425, "loss/hidden": 0.7265625, "loss/logits": 0.13847574591636658, "loss/reg": 0.020227529108524323, "step": 811 }, { "epoch": 0.1015, "grad_norm": 37.33988952636719, "grad_norm_var": 171.06705552642327, "learning_rate": 0.0001, "loss": 4.5438, "loss/crossentropy": 3.938170909881592, "loss/hidden": 2.203125, "loss/logits": 2.1385304927825928, "loss/reg": 0.02021711878478527, "step": 812 }, { "epoch": 0.101625, "grad_norm": 2.306735038757324, "grad_norm_var": 171.16339278078715, "learning_rate": 0.0001, "loss": 0.928, "loss/crossentropy": 2.3164937496185303, "loss/hidden": 0.625, "loss/logits": 0.10087703168392181, "loss/reg": 0.020207591354846954, "step": 813 }, { "epoch": 0.10175, "grad_norm": 3.00903058052063, "grad_norm_var": 171.72501112959992, "learning_rate": 0.0001, "loss": 1.1753, "loss/crossentropy": 2.4386990070343018, "loss/hidden": 0.828125, "loss/logits": 0.1452336609363556, "loss/reg": 0.02019745111465454, "step": 814 }, { "epoch": 0.101875, "grad_norm": 2.680140972137451, "grad_norm_var": 171.84144221114087, "learning_rate": 0.0001, "loss": 1.157, "loss/crossentropy": 2.531343460083008, "loss/hidden": 0.796875, "loss/logits": 0.15824541449546814, "loss/reg": 0.020187031477689743, "step": 815 }, { "epoch": 0.102, "grad_norm": 2.582740068435669, "grad_norm_var": 75.71062264039246, "learning_rate": 0.0001, "loss": 1.0065, "loss/crossentropy": 2.5258586406707764, "loss/hidden": 0.69140625, "loss/logits": 0.11330173909664154, "loss/reg": 0.02017681486904621, "step": 816 }, { "epoch": 0.102125, "grad_norm": 3.5809519290924072, "grad_norm_var": 75.53549752812218, "learning_rate": 0.0001, "loss": 1.2756, "loss/crossentropy": 2.080437183380127, "loss/hidden": 0.90625, "loss/logits": 0.16763544082641602, "loss/reg": 0.02016652189195156, "step": 817 }, { "epoch": 0.10225, "grad_norm": 3.093592405319214, "grad_norm_var": 75.89695881356819, "learning_rate": 0.0001, "loss": 1.1484, "loss/crossentropy": 2.630833625793457, "loss/hidden": 0.80078125, "loss/logits": 0.14600682258605957, "loss/reg": 0.020157409831881523, "step": 818 }, { "epoch": 0.102375, "grad_norm": 3.1120946407318115, "grad_norm_var": 76.0751246894024, "learning_rate": 0.0001, "loss": 1.038, "loss/crossentropy": 2.5411901473999023, "loss/hidden": 0.69140625, "loss/logits": 0.14511412382125854, "loss/reg": 0.020148303359746933, "step": 819 }, { "epoch": 0.1025, "grad_norm": 2.9903945922851562, "grad_norm_var": 76.21449508483448, "learning_rate": 0.0001, "loss": 1.0769, "loss/crossentropy": 2.556246042251587, "loss/hidden": 0.73828125, "loss/logits": 0.13721047341823578, "loss/reg": 0.020139139145612717, "step": 820 }, { "epoch": 0.102625, "grad_norm": 3.297760248184204, "grad_norm_var": 76.2077263993312, "learning_rate": 0.0001, "loss": 1.025, "loss/crossentropy": 2.5704898834228516, "loss/hidden": 0.6796875, "loss/logits": 0.14399868249893188, "loss/reg": 0.020129989832639694, "step": 821 }, { "epoch": 0.10275, "grad_norm": 3.810601234436035, "grad_norm_var": 75.94485425030823, "learning_rate": 0.0001, "loss": 1.2142, "loss/crossentropy": 2.4684388637542725, "loss/hidden": 0.83984375, "loss/logits": 0.17318351566791534, "loss/reg": 0.020120643079280853, "step": 822 }, { "epoch": 0.102875, "grad_norm": 2.798835515975952, "grad_norm_var": 76.52818091118122, "learning_rate": 0.0001, "loss": 0.9558, "loss/crossentropy": 2.476940393447876, "loss/hidden": 0.6484375, "loss/logits": 0.10626688599586487, "loss/reg": 0.020111503079533577, "step": 823 }, { "epoch": 0.103, "grad_norm": 4.541812419891357, "grad_norm_var": 75.99013496754353, "learning_rate": 0.0001, "loss": 1.1311, "loss/crossentropy": 2.339744806289673, "loss/hidden": 0.7265625, "loss/logits": 0.2034740447998047, "loss/reg": 0.020102351903915405, "step": 824 }, { "epoch": 0.103125, "grad_norm": 2.6360268592834473, "grad_norm_var": 75.95142766348106, "learning_rate": 0.0001, "loss": 1.0849, "loss/crossentropy": 2.3795337677001953, "loss/hidden": 0.74609375, "loss/logits": 0.1378898024559021, "loss/reg": 0.02009383775293827, "step": 825 }, { "epoch": 0.10325, "grad_norm": 3.2206642627716064, "grad_norm_var": 73.50864103963063, "learning_rate": 0.0001, "loss": 1.09, "loss/crossentropy": 2.645775556564331, "loss/hidden": 0.7578125, "loss/logits": 0.13137856125831604, "loss/reg": 0.020084405317902565, "step": 826 }, { "epoch": 0.103375, "grad_norm": 3.2071568965911865, "grad_norm_var": 73.45272548167614, "learning_rate": 0.0001, "loss": 0.9383, "loss/crossentropy": 2.7224762439727783, "loss/hidden": 0.625, "loss/logits": 0.11251779645681381, "loss/reg": 0.020074598491191864, "step": 827 }, { "epoch": 0.1035, "grad_norm": 3.2850613594055176, "grad_norm_var": 0.2863261800221416, "learning_rate": 0.0001, "loss": 1.1194, "loss/crossentropy": 2.522642135620117, "loss/hidden": 0.7421875, "loss/logits": 0.17652815580368042, "loss/reg": 0.02006435953080654, "step": 828 }, { "epoch": 0.103625, "grad_norm": 3.5068256855010986, "grad_norm_var": 0.24387138774257647, "learning_rate": 0.0001, "loss": 0.9457, "loss/crossentropy": 2.6463444232940674, "loss/hidden": 0.6328125, "loss/logits": 0.11229754984378815, "loss/reg": 0.020054515451192856, "step": 829 }, { "epoch": 0.10375, "grad_norm": 4.16710090637207, "grad_norm_var": 0.29672115328222404, "learning_rate": 0.0001, "loss": 1.301, "loss/crossentropy": 2.519270420074463, "loss/hidden": 0.921875, "loss/logits": 0.1786651611328125, "loss/reg": 0.020044928416609764, "step": 830 }, { "epoch": 0.103875, "grad_norm": 3.5839273929595947, "grad_norm_var": 0.27524789373512704, "learning_rate": 0.0001, "loss": 0.9949, "loss/crossentropy": 2.0604701042175293, "loss/hidden": 0.68359375, "loss/logits": 0.1109282597899437, "loss/reg": 0.02003585919737816, "step": 831 }, { "epoch": 0.104, "grad_norm": 2.8121650218963623, "grad_norm_var": 0.25541980739101955, "learning_rate": 0.0001, "loss": 1.1057, "loss/crossentropy": 2.670191764831543, "loss/hidden": 0.75390625, "loss/logits": 0.1515616476535797, "loss/reg": 0.020026110112667084, "step": 832 }, { "epoch": 0.104125, "grad_norm": 3.116018295288086, "grad_norm_var": 0.2547872758708628, "learning_rate": 0.0001, "loss": 1.0301, "loss/crossentropy": 3.016143321990967, "loss/hidden": 0.6796875, "loss/logits": 0.1502160131931305, "loss/reg": 0.020016156136989594, "step": 833 }, { "epoch": 0.10425, "grad_norm": 6.202447414398193, "grad_norm_var": 0.7634439694535559, "learning_rate": 0.0001, "loss": 1.2045, "loss/crossentropy": 2.8383448123931885, "loss/hidden": 0.84765625, "loss/logits": 0.15678462386131287, "loss/reg": 0.020007088780403137, "step": 834 }, { "epoch": 0.104375, "grad_norm": 3.4904069900512695, "grad_norm_var": 0.7519116349075012, "learning_rate": 0.0001, "loss": 1.2658, "loss/crossentropy": 2.597175359725952, "loss/hidden": 0.84375, "loss/logits": 0.22209098935127258, "loss/reg": 0.019998185336589813, "step": 835 }, { "epoch": 0.1045, "grad_norm": 4.6242594718933105, "grad_norm_var": 0.7986550791863064, "learning_rate": 0.0001, "loss": 1.1411, "loss/crossentropy": 2.6017813682556152, "loss/hidden": 0.796875, "loss/logits": 0.14435435831546783, "loss/reg": 0.019988389685750008, "step": 836 }, { "epoch": 0.104625, "grad_norm": 3.2676827907562256, "grad_norm_var": 0.800099420481778, "learning_rate": 0.0001, "loss": 1.1908, "loss/crossentropy": 2.454334259033203, "loss/hidden": 0.82421875, "loss/logits": 0.16681547462940216, "loss/reg": 0.019979091361165047, "step": 837 }, { "epoch": 0.10475, "grad_norm": 3.4931089878082275, "grad_norm_var": 0.7992595598721048, "learning_rate": 0.0001, "loss": 1.1803, "loss/crossentropy": 2.2890915870666504, "loss/hidden": 0.828125, "loss/logits": 0.15247204899787903, "loss/reg": 0.019969170913100243, "step": 838 }, { "epoch": 0.104875, "grad_norm": 2.7106313705444336, "grad_norm_var": 0.8094277801425143, "learning_rate": 0.0001, "loss": 1.097, "loss/crossentropy": 2.56706166267395, "loss/hidden": 0.76953125, "loss/logits": 0.12783397734165192, "loss/reg": 0.019960079342126846, "step": 839 }, { "epoch": 0.105, "grad_norm": 3.2825114727020264, "grad_norm_var": 0.7531900707246374, "learning_rate": 0.0001, "loss": 1.2726, "loss/crossentropy": 2.324145555496216, "loss/hidden": 0.88671875, "loss/logits": 0.18636029958724976, "loss/reg": 0.019951237365603447, "step": 840 }, { "epoch": 0.105125, "grad_norm": 2.752570152282715, "grad_norm_var": 0.7400250579900473, "learning_rate": 0.0001, "loss": 1.0885, "loss/crossentropy": 2.4273452758789062, "loss/hidden": 0.7578125, "loss/logits": 0.1312946379184723, "loss/reg": 0.01994233950972557, "step": 841 }, { "epoch": 0.10525, "grad_norm": 3.451720714569092, "grad_norm_var": 0.733364881032247, "learning_rate": 0.0001, "loss": 1.1042, "loss/crossentropy": 2.5271100997924805, "loss/hidden": 0.76171875, "loss/logits": 0.1431449055671692, "loss/reg": 0.019932815805077553, "step": 842 }, { "epoch": 0.105375, "grad_norm": 3.0586564540863037, "grad_norm_var": 0.741721542830341, "learning_rate": 0.0001, "loss": 1.0412, "loss/crossentropy": 2.9447247982025146, "loss/hidden": 0.7109375, "loss/logits": 0.13100013136863708, "loss/reg": 0.019923273473978043, "step": 843 }, { "epoch": 0.1055, "grad_norm": 3.4214084148406982, "grad_norm_var": 0.7380611813534226, "learning_rate": 0.0001, "loss": 1.2281, "loss/crossentropy": 2.3005340099334717, "loss/hidden": 0.86328125, "loss/logits": 0.1657242327928543, "loss/reg": 0.01991339959204197, "step": 844 }, { "epoch": 0.105625, "grad_norm": 2.6607367992401123, "grad_norm_var": 0.7886706735221035, "learning_rate": 0.0001, "loss": 1.1345, "loss/crossentropy": 2.5435853004455566, "loss/hidden": 0.80078125, "loss/logits": 0.1347101330757141, "loss/reg": 0.019903138279914856, "step": 845 }, { "epoch": 0.10575, "grad_norm": 3.0965163707733154, "grad_norm_var": 0.7659307635756494, "learning_rate": 0.0001, "loss": 1.0956, "loss/crossentropy": 2.6083555221557617, "loss/hidden": 0.74609375, "loss/logits": 0.1505199819803238, "loss/reg": 0.019893797114491463, "step": 846 }, { "epoch": 0.105875, "grad_norm": 4.309004306793213, "grad_norm_var": 0.8127957898222188, "learning_rate": 0.0001, "loss": 1.078, "loss/crossentropy": 2.515653371810913, "loss/hidden": 0.76171875, "loss/logits": 0.11742238700389862, "loss/reg": 0.019883660599589348, "step": 847 }, { "epoch": 0.106, "grad_norm": 3.248624563217163, "grad_norm_var": 0.7855834171862691, "learning_rate": 0.0001, "loss": 1.0838, "loss/crossentropy": 2.593562126159668, "loss/hidden": 0.75390625, "loss/logits": 0.13119123876094818, "loss/reg": 0.019874349236488342, "step": 848 }, { "epoch": 0.106125, "grad_norm": 3.347926378250122, "grad_norm_var": 0.7767115778534122, "learning_rate": 0.0001, "loss": 1.0816, "loss/crossentropy": 2.3947036266326904, "loss/hidden": 0.75, "loss/logits": 0.13295108079910278, "loss/reg": 0.019864298403263092, "step": 849 }, { "epoch": 0.10625, "grad_norm": 4.369658946990967, "grad_norm_var": 0.33264170947598637, "learning_rate": 0.0001, "loss": 1.0708, "loss/crossentropy": 2.449446201324463, "loss/hidden": 0.7421875, "loss/logits": 0.13007891178131104, "loss/reg": 0.019854165613651276, "step": 850 }, { "epoch": 0.106375, "grad_norm": 3.361130475997925, "grad_norm_var": 0.332327660409797, "learning_rate": 0.0001, "loss": 0.9761, "loss/crossentropy": 2.566408157348633, "loss/hidden": 0.66796875, "loss/logits": 0.10966208577156067, "loss/reg": 0.019843947142362595, "step": 851 }, { "epoch": 0.1065, "grad_norm": 2.6885204315185547, "grad_norm_var": 0.25144665871681204, "learning_rate": 0.0001, "loss": 0.9331, "loss/crossentropy": 2.3620035648345947, "loss/hidden": 0.625, "loss/logits": 0.10973039269447327, "loss/reg": 0.019834715873003006, "step": 852 }, { "epoch": 0.106625, "grad_norm": 2.615734338760376, "grad_norm_var": 0.2793016853206145, "learning_rate": 0.0001, "loss": 1.0168, "loss/crossentropy": 2.605545997619629, "loss/hidden": 0.69140625, "loss/logits": 0.12715153396129608, "loss/reg": 0.01982559822499752, "step": 853 }, { "epoch": 0.10675, "grad_norm": 3.073012113571167, "grad_norm_var": 0.276254032788457, "learning_rate": 0.0001, "loss": 0.9855, "loss/crossentropy": 2.634042739868164, "loss/hidden": 0.671875, "loss/logits": 0.11542729288339615, "loss/reg": 0.019816165789961815, "step": 854 }, { "epoch": 0.106875, "grad_norm": 2.759152412414551, "grad_norm_var": 0.27313479552050995, "learning_rate": 0.0001, "loss": 0.9145, "loss/crossentropy": 2.4920616149902344, "loss/hidden": 0.61328125, "loss/logits": 0.10316716134548187, "loss/reg": 0.019806833937764168, "step": 855 }, { "epoch": 0.107, "grad_norm": 3.647084951400757, "grad_norm_var": 0.28455080731760823, "learning_rate": 0.0001, "loss": 1.1336, "loss/crossentropy": 2.34993314743042, "loss/hidden": 0.78515625, "loss/logits": 0.150485098361969, "loss/reg": 0.019797123968601227, "step": 856 }, { "epoch": 0.107125, "grad_norm": 5.309474468231201, "grad_norm_var": 0.5265287098230971, "learning_rate": 0.0001, "loss": 1.3513, "loss/crossentropy": 2.4511518478393555, "loss/hidden": 0.94140625, "loss/logits": 0.21205759048461914, "loss/reg": 0.01978708617389202, "step": 857 }, { "epoch": 0.10725, "grad_norm": 3.1982548236846924, "grad_norm_var": 0.5288348795583182, "learning_rate": 0.0001, "loss": 1.1847, "loss/crossentropy": 2.2767844200134277, "loss/hidden": 0.8359375, "loss/logits": 0.15097308158874512, "loss/reg": 0.01977648213505745, "step": 858 }, { "epoch": 0.107375, "grad_norm": 3.7261335849761963, "grad_norm_var": 0.5276094221235986, "learning_rate": 0.0001, "loss": 1.1138, "loss/crossentropy": 2.660844087600708, "loss/hidden": 0.78515625, "loss/logits": 0.13101539015769958, "loss/reg": 0.01976662687957287, "step": 859 }, { "epoch": 0.1075, "grad_norm": 3.3435637950897217, "grad_norm_var": 0.5280464375318101, "learning_rate": 0.0001, "loss": 1.2877, "loss/crossentropy": 2.4701263904571533, "loss/hidden": 0.90625, "loss/logits": 0.18388135731220245, "loss/reg": 0.019757471978664398, "step": 860 }, { "epoch": 0.107625, "grad_norm": 2.630688190460205, "grad_norm_var": 0.5311534898567278, "learning_rate": 0.0001, "loss": 1.063, "loss/crossentropy": 2.5359132289886475, "loss/hidden": 0.734375, "loss/logits": 0.1311398446559906, "loss/reg": 0.019748201593756676, "step": 861 }, { "epoch": 0.10775, "grad_norm": 2.5543456077575684, "grad_norm_var": 0.5729300014134933, "learning_rate": 0.0001, "loss": 0.987, "loss/crossentropy": 2.3763418197631836, "loss/hidden": 0.671875, "loss/logits": 0.11773102730512619, "loss/reg": 0.01973855495452881, "step": 862 }, { "epoch": 0.107875, "grad_norm": 2.8768351078033447, "grad_norm_var": 0.5249464789316676, "learning_rate": 0.0001, "loss": 1.11, "loss/crossentropy": 2.467682361602783, "loss/hidden": 0.7734375, "loss/logits": 0.13924749195575714, "loss/reg": 0.019729435443878174, "step": 863 }, { "epoch": 0.108, "grad_norm": 9.608988761901855, "grad_norm_var": 3.012409881249372, "learning_rate": 0.0001, "loss": 1.8371, "loss/crossentropy": 2.4410533905029297, "loss/hidden": 1.359375, "loss/logits": 0.280517578125, "loss/reg": 0.019719891250133514, "step": 864 }, { "epoch": 0.108125, "grad_norm": 2.6161296367645264, "grad_norm_var": 3.079687357926654, "learning_rate": 0.0001, "loss": 0.9254, "loss/crossentropy": 2.656090497970581, "loss/hidden": 0.6171875, "loss/logits": 0.11112320423126221, "loss/reg": 0.01971041038632393, "step": 865 }, { "epoch": 0.10825, "grad_norm": 6.282230377197266, "grad_norm_var": 3.4921671952336317, "learning_rate": 0.0001, "loss": 1.1269, "loss/crossentropy": 2.4799087047576904, "loss/hidden": 0.80078125, "loss/logits": 0.12914146482944489, "loss/reg": 0.01970127783715725, "step": 866 }, { "epoch": 0.108375, "grad_norm": 3.052783727645874, "grad_norm_var": 3.514845564297898, "learning_rate": 0.0001, "loss": 1.1383, "loss/crossentropy": 2.4700145721435547, "loss/hidden": 0.78515625, "loss/logits": 0.15620394051074982, "loss/reg": 0.019691679626703262, "step": 867 }, { "epoch": 0.1085, "grad_norm": 3.2105300426483154, "grad_norm_var": 3.458070348929603, "learning_rate": 0.0001, "loss": 1.067, "loss/crossentropy": 2.135777473449707, "loss/hidden": 0.7578125, "loss/logits": 0.11233663558959961, "loss/reg": 0.019682079553604126, "step": 868 }, { "epoch": 0.108625, "grad_norm": 2.7341737747192383, "grad_norm_var": 3.4405364793380135, "learning_rate": 0.0001, "loss": 1.0495, "loss/crossentropy": 2.4592809677124023, "loss/hidden": 0.73046875, "loss/logits": 0.1223248764872551, "loss/reg": 0.019672293215990067, "step": 869 }, { "epoch": 0.10875, "grad_norm": 4.007035732269287, "grad_norm_var": 3.4058996890488658, "learning_rate": 0.0001, "loss": 1.1341, "loss/crossentropy": 2.5008175373077393, "loss/hidden": 0.80078125, "loss/logits": 0.13670845329761505, "loss/reg": 0.01966211199760437, "step": 870 }, { "epoch": 0.108875, "grad_norm": 3.146348476409912, "grad_norm_var": 3.359090924722083, "learning_rate": 0.0001, "loss": 1.2459, "loss/crossentropy": 2.0440711975097656, "loss/hidden": 0.8984375, "loss/logits": 0.15090960264205933, "loss/reg": 0.019652366638183594, "step": 871 }, { "epoch": 0.109, "grad_norm": 4.632335662841797, "grad_norm_var": 3.3902752468766453, "learning_rate": 0.0001, "loss": 1.1446, "loss/crossentropy": 2.4574947357177734, "loss/hidden": 0.80859375, "loss/logits": 0.13958214223384857, "loss/reg": 0.01964336633682251, "step": 872 }, { "epoch": 0.109125, "grad_norm": 5.440892219543457, "grad_norm_var": 3.415471723579617, "learning_rate": 0.0001, "loss": 1.1098, "loss/crossentropy": 2.565347671508789, "loss/hidden": 0.8046875, "loss/logits": 0.10876456648111343, "loss/reg": 0.01963435485959053, "step": 873 }, { "epoch": 0.10925, "grad_norm": 2.873284101486206, "grad_norm_var": 3.454269091905695, "learning_rate": 0.0001, "loss": 1.0607, "loss/crossentropy": 2.561459541320801, "loss/hidden": 0.72265625, "loss/logits": 0.14177533984184265, "loss/reg": 0.01962495781481266, "step": 874 }, { "epoch": 0.109375, "grad_norm": 3.0819554328918457, "grad_norm_var": 3.496943197417559, "learning_rate": 0.0001, "loss": 1.2422, "loss/crossentropy": 2.2944936752319336, "loss/hidden": 0.8828125, "loss/logits": 0.16319133341312408, "loss/reg": 0.019616009667515755, "step": 875 }, { "epoch": 0.1095, "grad_norm": 4.361453533172607, "grad_norm_var": 3.488792217244569, "learning_rate": 0.0001, "loss": 1.3828, "loss/crossentropy": 2.4624311923980713, "loss/hidden": 1.0078125, "loss/logits": 0.1789003610610962, "loss/reg": 0.019607286900281906, "step": 876 }, { "epoch": 0.109625, "grad_norm": 3.34078049659729, "grad_norm_var": 3.3959280790074216, "learning_rate": 0.0001, "loss": 0.9589, "loss/crossentropy": 2.4549739360809326, "loss/hidden": 0.64453125, "loss/logits": 0.11839590966701508, "loss/reg": 0.019598115235567093, "step": 877 }, { "epoch": 0.10975, "grad_norm": 3.434715986251831, "grad_norm_var": 3.2759937907981884, "learning_rate": 0.0001, "loss": 1.1192, "loss/crossentropy": 2.571798086166382, "loss/hidden": 0.7578125, "loss/logits": 0.16550706326961517, "loss/reg": 0.01958884485065937, "step": 878 }, { "epoch": 0.109875, "grad_norm": 3.593993663787842, "grad_norm_var": 3.196554005024426, "learning_rate": 0.0001, "loss": 1.1792, "loss/crossentropy": 2.447843551635742, "loss/hidden": 0.83203125, "loss/logits": 0.15137381851673126, "loss/reg": 0.019579457119107246, "step": 879 }, { "epoch": 0.11, "grad_norm": 3.3654563426971436, "grad_norm_var": 1.0373482238282006, "learning_rate": 0.0001, "loss": 1.0326, "loss/crossentropy": 2.6210246086120605, "loss/hidden": 0.71484375, "loss/logits": 0.12209475785493851, "loss/reg": 0.019569827243685722, "step": 880 }, { "epoch": 0.110125, "grad_norm": 2.6370863914489746, "grad_norm_var": 1.0343516088559113, "learning_rate": 0.0001, "loss": 1.1442, "loss/crossentropy": 2.2681713104248047, "loss/hidden": 0.81640625, "loss/logits": 0.132216215133667, "loss/reg": 0.019560784101486206, "step": 881 }, { "epoch": 0.11025, "grad_norm": 2.152343273162842, "grad_norm_var": 0.6782700998492743, "learning_rate": 0.0001, "loss": 1.0136, "loss/crossentropy": 2.437594175338745, "loss/hidden": 0.70703125, "loss/logits": 0.11108942329883575, "loss/reg": 0.019551947712898254, "step": 882 }, { "epoch": 0.110375, "grad_norm": 3.2755074501037598, "grad_norm_var": 0.6698247850929626, "learning_rate": 0.0001, "loss": 0.9972, "loss/crossentropy": 2.481198310852051, "loss/hidden": 0.6796875, "loss/logits": 0.12207823246717453, "loss/reg": 0.01954270713031292, "step": 883 }, { "epoch": 0.1105, "grad_norm": 3.770535707473755, "grad_norm_var": 0.6711344077538037, "learning_rate": 0.0001, "loss": 1.0475, "loss/crossentropy": 2.434851884841919, "loss/hidden": 0.71875, "loss/logits": 0.13345816731452942, "loss/reg": 0.019533507525920868, "step": 884 }, { "epoch": 0.110625, "grad_norm": 2.7666234970092773, "grad_norm_var": 0.6679279033368438, "learning_rate": 0.0001, "loss": 1.1166, "loss/crossentropy": 2.433852434158325, "loss/hidden": 0.78125, "loss/logits": 0.1401294767856598, "loss/reg": 0.01952442154288292, "step": 885 }, { "epoch": 0.11075, "grad_norm": 2.7105534076690674, "grad_norm_var": 0.6840409496040507, "learning_rate": 0.0001, "loss": 1.0924, "loss/crossentropy": 2.219019889831543, "loss/hidden": 0.765625, "loss/logits": 0.13160449266433716, "loss/reg": 0.019515201449394226, "step": 886 }, { "epoch": 0.110875, "grad_norm": 2.8931920528411865, "grad_norm_var": 0.6969961519386968, "learning_rate": 0.0001, "loss": 1.1134, "loss/crossentropy": 2.511371612548828, "loss/hidden": 0.78125, "loss/logits": 0.13708999752998352, "loss/reg": 0.019505700096488, "step": 887 }, { "epoch": 0.111, "grad_norm": 2.821718215942383, "grad_norm_var": 0.6033415037749854, "learning_rate": 0.0001, "loss": 1.2446, "loss/crossentropy": 2.389580011367798, "loss/hidden": 0.8828125, "loss/logits": 0.16678079962730408, "loss/reg": 0.019496839493513107, "step": 888 }, { "epoch": 0.111125, "grad_norm": 2.833461284637451, "grad_norm_var": 0.2778808504856213, "learning_rate": 0.0001, "loss": 1.029, "loss/crossentropy": 2.1618683338165283, "loss/hidden": 0.70703125, "loss/logits": 0.12710769474506378, "loss/reg": 0.019487854093313217, "step": 889 }, { "epoch": 0.11125, "grad_norm": 3.3806753158569336, "grad_norm_var": 0.2773113837378702, "learning_rate": 0.0001, "loss": 1.1349, "loss/crossentropy": 2.656121253967285, "loss/hidden": 0.79296875, "loss/logits": 0.1471368372440338, "loss/reg": 0.019478676840662956, "step": 890 }, { "epoch": 0.111375, "grad_norm": 2.932758092880249, "grad_norm_var": 0.2800811641911004, "learning_rate": 0.0001, "loss": 1.1, "loss/crossentropy": 2.5325851440429688, "loss/hidden": 0.76171875, "loss/logits": 0.14360320568084717, "loss/reg": 0.019469575956463814, "step": 891 }, { "epoch": 0.1115, "grad_norm": 2.3603265285491943, "grad_norm_var": 0.20497304301798067, "learning_rate": 0.0001, "loss": 1.0739, "loss/crossentropy": 2.3697421550750732, "loss/hidden": 0.75, "loss/logits": 0.1292482614517212, "loss/reg": 0.01946048066020012, "step": 892 }, { "epoch": 0.111625, "grad_norm": 4.056443214416504, "grad_norm_var": 0.2678930990243863, "learning_rate": 0.0001, "loss": 1.0008, "loss/crossentropy": 2.7553114891052246, "loss/hidden": 0.68359375, "loss/logits": 0.12268239259719849, "loss/reg": 0.019451187923550606, "step": 893 }, { "epoch": 0.11175, "grad_norm": 3.346064329147339, "grad_norm_var": 0.2639738255705176, "learning_rate": 0.0001, "loss": 1.1791, "loss/crossentropy": 2.4348855018615723, "loss/hidden": 0.7890625, "loss/logits": 0.19565626978874207, "loss/reg": 0.019442636519670486, "step": 894 }, { "epoch": 0.111875, "grad_norm": 6.878971576690674, "grad_norm_var": 1.1740357353356365, "learning_rate": 0.0001, "loss": 1.7085, "loss/crossentropy": 2.9299001693725586, "loss/hidden": 1.1640625, "loss/logits": 0.3501082956790924, "loss/reg": 0.019433531910181046, "step": 895 }, { "epoch": 0.112, "grad_norm": 3.1834142208099365, "grad_norm_var": 1.1735802221223497, "learning_rate": 0.0001, "loss": 1.2953, "loss/crossentropy": 2.4165709018707275, "loss/hidden": 0.92578125, "loss/logits": 0.17523059248924255, "loss/reg": 0.019424354657530785, "step": 896 }, { "epoch": 0.112125, "grad_norm": 2.5551555156707764, "grad_norm_var": 1.180695081530242, "learning_rate": 0.0001, "loss": 1.0171, "loss/crossentropy": 2.4345083236694336, "loss/hidden": 0.6953125, "loss/logits": 0.1276446282863617, "loss/reg": 0.019415004178881645, "step": 897 }, { "epoch": 0.11225, "grad_norm": 3.5786855220794678, "grad_norm_var": 1.1000748366509354, "learning_rate": 0.0001, "loss": 0.9925, "loss/crossentropy": 2.464219808578491, "loss/hidden": 0.69140625, "loss/logits": 0.10706112533807755, "loss/reg": 0.01940576173365116, "step": 898 }, { "epoch": 0.112375, "grad_norm": 8.784457206726074, "grad_norm_var": 2.9538895197120096, "learning_rate": 0.0001, "loss": 1.6799, "loss/crossentropy": 2.3273468017578125, "loss/hidden": 1.2265625, "loss/logits": 0.25934016704559326, "loss/reg": 0.019396713003516197, "step": 899 }, { "epoch": 0.1125, "grad_norm": 4.44765567779541, "grad_norm_var": 2.990871190956643, "learning_rate": 0.0001, "loss": 1.0685, "loss/crossentropy": 2.5760438442230225, "loss/hidden": 0.73046875, "loss/logits": 0.14414295554161072, "loss/reg": 0.019387517124414444, "step": 900 }, { "epoch": 0.112625, "grad_norm": 4.9651265144348145, "grad_norm_var": 3.013306784613239, "learning_rate": 0.0001, "loss": 1.1917, "loss/crossentropy": 2.5148589611053467, "loss/hidden": 0.828125, "loss/logits": 0.1697455495595932, "loss/reg": 0.01937839388847351, "step": 901 }, { "epoch": 0.11275, "grad_norm": 3.8322513103485107, "grad_norm_var": 2.9203267227302745, "learning_rate": 0.0001, "loss": 1.2309, "loss/crossentropy": 2.536973476409912, "loss/hidden": 0.8828125, "loss/logits": 0.15436521172523499, "loss/reg": 0.019369108602404594, "step": 902 }, { "epoch": 0.112875, "grad_norm": 3.5994656085968018, "grad_norm_var": 2.8540415836766555, "learning_rate": 0.0001, "loss": 1.0761, "loss/crossentropy": 2.3160693645477295, "loss/hidden": 0.76171875, "loss/logits": 0.12077254056930542, "loss/reg": 0.019359666854143143, "step": 903 }, { "epoch": 0.113, "grad_norm": 4.031139373779297, "grad_norm_var": 2.7599236229360753, "learning_rate": 0.0001, "loss": 1.0988, "loss/crossentropy": 2.5636203289031982, "loss/hidden": 0.78515625, "loss/logits": 0.12018904089927673, "loss/reg": 0.019350115209817886, "step": 904 }, { "epoch": 0.113125, "grad_norm": 3.2178378105163574, "grad_norm_var": 2.7069185907568794, "learning_rate": 0.0001, "loss": 1.0472, "loss/crossentropy": 2.3457703590393066, "loss/hidden": 0.72265625, "loss/logits": 0.13114379346370697, "loss/reg": 0.019340479746460915, "step": 905 }, { "epoch": 0.11325, "grad_norm": 3.096679925918579, "grad_norm_var": 2.7381334427643536, "learning_rate": 0.0001, "loss": 0.9662, "loss/crossentropy": 2.3510820865631104, "loss/hidden": 0.671875, "loss/logits": 0.10106582939624786, "loss/reg": 0.019330844283103943, "step": 906 }, { "epoch": 0.113375, "grad_norm": 7.590158462524414, "grad_norm_var": 3.3974738441650887, "learning_rate": 0.0001, "loss": 1.6666, "loss/crossentropy": 2.3520147800445557, "loss/hidden": 1.2265625, "loss/logits": 0.2468346357345581, "loss/reg": 0.019321195781230927, "step": 907 }, { "epoch": 0.1135, "grad_norm": 3.3216323852539062, "grad_norm_var": 3.20081618522182, "learning_rate": 0.0001, "loss": 1.2612, "loss/crossentropy": 2.330040216445923, "loss/hidden": 0.90234375, "loss/logits": 0.16571447253227234, "loss/reg": 0.019311606884002686, "step": 908 }, { "epoch": 0.113625, "grad_norm": 3.4920501708984375, "grad_norm_var": 3.246978809627046, "learning_rate": 0.0001, "loss": 1.1161, "loss/crossentropy": 2.598465919494629, "loss/hidden": 0.76953125, "loss/logits": 0.15353354811668396, "loss/reg": 0.019302019849419594, "step": 909 }, { "epoch": 0.11375, "grad_norm": 3.0414321422576904, "grad_norm_var": 3.294370585536838, "learning_rate": 0.0001, "loss": 1.0693, "loss/crossentropy": 2.459712028503418, "loss/hidden": 0.7421875, "loss/logits": 0.13418710231781006, "loss/reg": 0.019292324781417847, "step": 910 }, { "epoch": 0.113875, "grad_norm": 3.131361722946167, "grad_norm_var": 2.908980195007492, "learning_rate": 0.0001, "loss": 1.2301, "loss/crossentropy": 2.268738269805908, "loss/hidden": 0.875, "loss/logits": 0.16223490238189697, "loss/reg": 0.019283456727862358, "step": 911 }, { "epoch": 0.114, "grad_norm": 3.160707950592041, "grad_norm_var": 2.9118381902992776, "learning_rate": 0.0001, "loss": 1.3173, "loss/crossentropy": 2.2185730934143066, "loss/hidden": 0.9453125, "loss/logits": 0.17927365005016327, "loss/reg": 0.019274268299341202, "step": 912 }, { "epoch": 0.114125, "grad_norm": 14.604296684265137, "grad_norm_var": 9.479147248477696, "learning_rate": 0.0001, "loss": 1.4278, "loss/crossentropy": 2.2313976287841797, "loss/hidden": 1.078125, "loss/logits": 0.15699920058250427, "loss/reg": 0.019266733899712563, "step": 913 }, { "epoch": 0.11425, "grad_norm": 3.744290590286255, "grad_norm_var": 9.452382803070204, "learning_rate": 0.0001, "loss": 1.1776, "loss/crossentropy": 2.689699172973633, "loss/hidden": 0.83203125, "loss/logits": 0.15302729606628418, "loss/reg": 0.019258547574281693, "step": 914 }, { "epoch": 0.114375, "grad_norm": 2.999354839324951, "grad_norm_var": 8.531466626401157, "learning_rate": 0.0001, "loss": 1.3107, "loss/crossentropy": 1.8664510250091553, "loss/hidden": 0.9453125, "loss/logits": 0.1728517711162567, "loss/reg": 0.01925109326839447, "step": 915 }, { "epoch": 0.1145, "grad_norm": 3.194913148880005, "grad_norm_var": 8.64117053500835, "learning_rate": 0.0001, "loss": 1.3016, "loss/crossentropy": 2.25707745552063, "loss/hidden": 0.95703125, "loss/logits": 0.15214993059635162, "loss/reg": 0.019243914633989334, "step": 916 }, { "epoch": 0.114625, "grad_norm": 2.335845708847046, "grad_norm_var": 8.88876728908843, "learning_rate": 0.0001, "loss": 0.9484, "loss/crossentropy": 2.6386678218841553, "loss/hidden": 0.63671875, "loss/logits": 0.11931365728378296, "loss/reg": 0.019234785810112953, "step": 917 }, { "epoch": 0.11475, "grad_norm": 3.6922056674957275, "grad_norm_var": 8.898252742921356, "learning_rate": 0.0001, "loss": 1.0667, "loss/crossentropy": 2.5539944171905518, "loss/hidden": 0.73046875, "loss/logits": 0.14395025372505188, "loss/reg": 0.019226964563131332, "step": 918 }, { "epoch": 0.114875, "grad_norm": 2.946389675140381, "grad_norm_var": 8.982934878513694, "learning_rate": 0.0001, "loss": 1.0911, "loss/crossentropy": 2.3651351928710938, "loss/hidden": 0.76171875, "loss/logits": 0.1371677815914154, "loss/reg": 0.019219111651182175, "step": 919 }, { "epoch": 0.115, "grad_norm": 3.050600051879883, "grad_norm_var": 9.068373446668678, "learning_rate": 0.0001, "loss": 1.0419, "loss/crossentropy": 2.3772196769714355, "loss/hidden": 0.734375, "loss/logits": 0.11545050889253616, "loss/reg": 0.019211286678910255, "step": 920 }, { "epoch": 0.115125, "grad_norm": 2.954526424407959, "grad_norm_var": 9.105915478669973, "learning_rate": 0.0001, "loss": 1.0807, "loss/crossentropy": 2.2472290992736816, "loss/hidden": 0.75390625, "loss/logits": 0.13482055068016052, "loss/reg": 0.019202249124646187, "step": 921 }, { "epoch": 0.11525, "grad_norm": 6.214420795440674, "grad_norm_var": 9.27670245999236, "learning_rate": 0.0001, "loss": 1.8562, "loss/crossentropy": 2.9924590587615967, "loss/hidden": 1.2734375, "loss/logits": 0.3907894492149353, "loss/reg": 0.019193273037672043, "step": 922 }, { "epoch": 0.115375, "grad_norm": 4.534095287322998, "grad_norm_var": 8.536934613221737, "learning_rate": 0.0001, "loss": 1.4885, "loss/crossentropy": 2.394090414047241, "loss/hidden": 1.078125, "loss/logits": 0.2185368537902832, "loss/reg": 0.019184142351150513, "step": 923 }, { "epoch": 0.1155, "grad_norm": 3.6761627197265625, "grad_norm_var": 8.505579278095963, "learning_rate": 0.0001, "loss": 1.1306, "loss/crossentropy": 2.3566806316375732, "loss/hidden": 0.7890625, "loss/logits": 0.14977289736270905, "loss/reg": 0.019175738096237183, "step": 924 }, { "epoch": 0.115625, "grad_norm": 4.592898368835449, "grad_norm_var": 8.481328607270026, "learning_rate": 0.0001, "loss": 1.2942, "loss/crossentropy": 2.7458250522613525, "loss/hidden": 0.953125, "loss/logits": 0.14940449595451355, "loss/reg": 0.01916695386171341, "step": 925 }, { "epoch": 0.11575, "grad_norm": 2.7483372688293457, "grad_norm_var": 8.533618684340597, "learning_rate": 0.0001, "loss": 1.0727, "loss/crossentropy": 2.4450652599334717, "loss/hidden": 0.75390625, "loss/logits": 0.12724441289901733, "loss/reg": 0.019158538430929184, "step": 926 }, { "epoch": 0.115875, "grad_norm": 3.075195074081421, "grad_norm_var": 8.541996814909611, "learning_rate": 0.0001, "loss": 1.1251, "loss/crossentropy": 2.3960814476013184, "loss/hidden": 0.76171875, "loss/logits": 0.1718631386756897, "loss/reg": 0.019150495529174805, "step": 927 }, { "epoch": 0.116, "grad_norm": 3.313359022140503, "grad_norm_var": 8.521887542243068, "learning_rate": 0.0001, "loss": 1.1316, "loss/crossentropy": 2.3464736938476562, "loss/hidden": 0.8125, "loss/logits": 0.1276848018169403, "loss/reg": 0.019141457974910736, "step": 928 }, { "epoch": 0.116125, "grad_norm": 4.252639293670654, "grad_norm_var": 0.9000980544993648, "learning_rate": 0.0001, "loss": 1.1818, "loss/crossentropy": 2.523444414138794, "loss/hidden": 0.83203125, "loss/logits": 0.15845248103141785, "loss/reg": 0.01913331262767315, "step": 929 }, { "epoch": 0.11625, "grad_norm": 3.097456455230713, "grad_norm_var": 0.9123223599265882, "learning_rate": 0.0001, "loss": 1.1264, "loss/crossentropy": 2.595120668411255, "loss/hidden": 0.7734375, "loss/logits": 0.16165900230407715, "loss/reg": 0.019125619903206825, "step": 930 }, { "epoch": 0.116375, "grad_norm": 3.292982816696167, "grad_norm_var": 0.8964505136113113, "learning_rate": 0.0001, "loss": 1.1132, "loss/crossentropy": 2.4093523025512695, "loss/hidden": 0.77734375, "loss/logits": 0.14468303322792053, "loss/reg": 0.019116582348942757, "step": 931 }, { "epoch": 0.1165, "grad_norm": 2.559980869293213, "grad_norm_var": 0.952617731514821, "learning_rate": 0.0001, "loss": 0.9697, "loss/crossentropy": 2.5491104125976562, "loss/hidden": 0.6484375, "loss/logits": 0.13018551468849182, "loss/reg": 0.019108334556221962, "step": 932 }, { "epoch": 0.116625, "grad_norm": 3.058579683303833, "grad_norm_var": 0.871050822267735, "learning_rate": 0.0001, "loss": 1.2919, "loss/crossentropy": 2.303837537765503, "loss/hidden": 0.9296875, "loss/logits": 0.17123734951019287, "loss/reg": 0.019099365919828415, "step": 933 }, { "epoch": 0.11675, "grad_norm": 4.866446495056152, "grad_norm_var": 0.9769503909617764, "learning_rate": 0.0001, "loss": 1.1876, "loss/crossentropy": 2.364741325378418, "loss/hidden": 0.8359375, "loss/logits": 0.1607905626296997, "loss/reg": 0.019090238958597183, "step": 934 }, { "epoch": 0.116875, "grad_norm": 5.912527561187744, "grad_norm_var": 1.252657817578581, "learning_rate": 0.0001, "loss": 1.4831, "loss/crossentropy": 2.641162633895874, "loss/hidden": 1.09375, "loss/logits": 0.19855040311813354, "loss/reg": 0.019081177189946175, "step": 935 }, { "epoch": 0.117, "grad_norm": 2.8301663398742676, "grad_norm_var": 1.2784556528629765, "learning_rate": 0.0001, "loss": 1.0878, "loss/crossentropy": 2.277157783508301, "loss/hidden": 0.765625, "loss/logits": 0.13142800331115723, "loss/reg": 0.019072722643613815, "step": 936 }, { "epoch": 0.117125, "grad_norm": 2.9044833183288574, "grad_norm_var": 1.2843284928455583, "learning_rate": 0.0001, "loss": 1.1381, "loss/crossentropy": 2.3311665058135986, "loss/hidden": 0.796875, "loss/logits": 0.1506001055240631, "loss/reg": 0.01906409114599228, "step": 937 }, { "epoch": 0.11725, "grad_norm": 3.78202223777771, "grad_norm_var": 0.8736988295376342, "learning_rate": 0.0001, "loss": 1.2869, "loss/crossentropy": 2.4287989139556885, "loss/hidden": 0.8984375, "loss/logits": 0.197871595621109, "loss/reg": 0.019055521115660667, "step": 938 }, { "epoch": 0.117375, "grad_norm": 3.1086387634277344, "grad_norm_var": 0.8338185014655397, "learning_rate": 0.0001, "loss": 1.0576, "loss/crossentropy": 2.335935354232788, "loss/hidden": 0.73828125, "loss/logits": 0.1288556456565857, "loss/reg": 0.019047552719712257, "step": 939 }, { "epoch": 0.1175, "grad_norm": 4.869723320007324, "grad_norm_var": 0.9402287231159085, "learning_rate": 0.0001, "loss": 1.0378, "loss/crossentropy": 2.9672138690948486, "loss/hidden": 0.7109375, "loss/logits": 0.13644808530807495, "loss/reg": 0.01903851516544819, "step": 940 }, { "epoch": 0.117625, "grad_norm": 2.8750882148742676, "grad_norm_var": 0.9067692046415797, "learning_rate": 0.0001, "loss": 1.0037, "loss/crossentropy": 2.3138959407806396, "loss/hidden": 0.69140625, "loss/logits": 0.12202918529510498, "loss/reg": 0.019029438495635986, "step": 941 }, { "epoch": 0.11775, "grad_norm": 3.044121742248535, "grad_norm_var": 0.8812433820018912, "learning_rate": 0.0001, "loss": 1.0173, "loss/crossentropy": 2.4543912410736084, "loss/hidden": 0.70703125, "loss/logits": 0.12005805224180222, "loss/reg": 0.01902030035853386, "step": 942 }, { "epoch": 0.117875, "grad_norm": 2.945160150527954, "grad_norm_var": 0.8905794039935603, "learning_rate": 0.0001, "loss": 1.1688, "loss/crossentropy": 2.3482954502105713, "loss/hidden": 0.828125, "loss/logits": 0.15057498216629028, "loss/reg": 0.019011201336979866, "step": 943 }, { "epoch": 0.118, "grad_norm": 3.0109965801239014, "grad_norm_var": 0.9056152589294107, "learning_rate": 0.0001, "loss": 1.3721, "loss/crossentropy": 2.353516101837158, "loss/hidden": 0.984375, "loss/logits": 0.1977054923772812, "loss/reg": 0.0190016757696867, "step": 944 }, { "epoch": 0.118125, "grad_norm": 3.424039363861084, "grad_norm_var": 0.8682128423744849, "learning_rate": 0.0001, "loss": 1.1647, "loss/crossentropy": 2.513850450515747, "loss/hidden": 0.81640625, "loss/logits": 0.15838034451007843, "loss/reg": 0.018992552533745766, "step": 945 }, { "epoch": 0.11825, "grad_norm": 2.7656776905059814, "grad_norm_var": 0.8917454992029661, "learning_rate": 0.0001, "loss": 1.0819, "loss/crossentropy": 2.650881767272949, "loss/hidden": 0.75, "loss/logits": 0.1420745849609375, "loss/reg": 0.01898341253399849, "step": 946 }, { "epoch": 0.118375, "grad_norm": 4.2130560874938965, "grad_norm_var": 0.9250033835133629, "learning_rate": 0.0001, "loss": 1.2469, "loss/crossentropy": 2.683312177658081, "loss/hidden": 0.90625, "loss/logits": 0.15088841319084167, "loss/reg": 0.018973875790834427, "step": 947 }, { "epoch": 0.1185, "grad_norm": 3.834226608276367, "grad_norm_var": 0.8649633510209883, "learning_rate": 0.0001, "loss": 0.9767, "loss/crossentropy": 2.5706145763397217, "loss/hidden": 0.6640625, "loss/logits": 0.12295837700366974, "loss/reg": 0.018963845446705818, "step": 948 }, { "epoch": 0.118625, "grad_norm": 3.3009235858917236, "grad_norm_var": 0.8514524765901378, "learning_rate": 0.0001, "loss": 1.1857, "loss/crossentropy": 2.6797893047332764, "loss/hidden": 0.82421875, "loss/logits": 0.17191748321056366, "loss/reg": 0.018954817205667496, "step": 949 }, { "epoch": 0.11875, "grad_norm": 6.452317237854004, "grad_norm_var": 1.2752747995844325, "learning_rate": 0.0001, "loss": 1.4528, "loss/crossentropy": 2.7505640983581543, "loss/hidden": 1.0546875, "loss/logits": 0.20869939029216766, "loss/reg": 0.018945740535855293, "step": 950 }, { "epoch": 0.118875, "grad_norm": 3.405714273452759, "grad_norm_var": 0.9300412257064863, "learning_rate": 0.0001, "loss": 1.0922, "loss/crossentropy": 2.802401065826416, "loss/hidden": 0.7578125, "loss/logits": 0.14506830275058746, "loss/reg": 0.018936749547719955, "step": 951 }, { "epoch": 0.119, "grad_norm": 9.662891387939453, "grad_norm_var": 3.194050081601077, "learning_rate": 0.0001, "loss": 1.5003, "loss/crossentropy": 2.700052499771118, "loss/hidden": 1.1171875, "loss/logits": 0.19381779432296753, "loss/reg": 0.01892753876745701, "step": 952 }, { "epoch": 0.119125, "grad_norm": 2.8670711517333984, "grad_norm_var": 3.199477320796336, "learning_rate": 0.0001, "loss": 1.0153, "loss/crossentropy": 2.490739107131958, "loss/hidden": 0.69921875, "loss/logits": 0.12687504291534424, "loss/reg": 0.01891852729022503, "step": 953 }, { "epoch": 0.11925, "grad_norm": 4.066836357116699, "grad_norm_var": 3.1973098694543274, "learning_rate": 0.0001, "loss": 1.2704, "loss/crossentropy": 2.4122891426086426, "loss/hidden": 0.9140625, "loss/logits": 0.1672634333372116, "loss/reg": 0.018909232690930367, "step": 954 }, { "epoch": 0.119375, "grad_norm": 2.8044497966766357, "grad_norm_var": 3.2388562001879793, "learning_rate": 0.0001, "loss": 1.0385, "loss/crossentropy": 2.473945379257202, "loss/hidden": 0.71484375, "loss/logits": 0.13464638590812683, "loss/reg": 0.018900100141763687, "step": 955 }, { "epoch": 0.1195, "grad_norm": 3.7214255332946777, "grad_norm_var": 3.1837278954586044, "learning_rate": 0.0001, "loss": 1.0832, "loss/crossentropy": 2.6148641109466553, "loss/hidden": 0.75390625, "loss/logits": 0.1403505802154541, "loss/reg": 0.01889113523066044, "step": 956 }, { "epoch": 0.119625, "grad_norm": 3.0579674243927, "grad_norm_var": 3.160836005262268, "learning_rate": 0.0001, "loss": 1.1761, "loss/crossentropy": 2.5016398429870605, "loss/hidden": 0.83984375, "loss/logits": 0.1474056839942932, "loss/reg": 0.0188821442425251, "step": 957 }, { "epoch": 0.11975, "grad_norm": 2.429112672805786, "grad_norm_var": 3.255565314691306, "learning_rate": 0.0001, "loss": 0.9853, "loss/crossentropy": 2.3262507915496826, "loss/hidden": 0.66796875, "loss/logits": 0.1286056935787201, "loss/reg": 0.018873048946261406, "step": 958 }, { "epoch": 0.119875, "grad_norm": 2.9561123847961426, "grad_norm_var": 3.2542184489207098, "learning_rate": 0.0001, "loss": 0.9739, "loss/crossentropy": 2.5669143199920654, "loss/hidden": 0.65234375, "loss/logits": 0.13291960954666138, "loss/reg": 0.018863873556256294, "step": 959 }, { "epoch": 0.12, "grad_norm": 3.1904752254486084, "grad_norm_var": 3.2355963683487268, "learning_rate": 0.0001, "loss": 1.074, "loss/crossentropy": 2.5522408485412598, "loss/hidden": 0.75390625, "loss/logits": 0.1315881609916687, "loss/reg": 0.018855126574635506, "step": 960 }, { "epoch": 0.120125, "grad_norm": 7.435352802276611, "grad_norm_var": 3.9949775747956586, "learning_rate": 0.0001, "loss": 1.6007, "loss/crossentropy": 2.7422168254852295, "loss/hidden": 1.1875, "loss/logits": 0.2246999442577362, "loss/reg": 0.018846556544303894, "step": 961 }, { "epoch": 0.12025, "grad_norm": 4.614249229431152, "grad_norm_var": 3.8709926395951384, "learning_rate": 0.0001, "loss": 1.3908, "loss/crossentropy": 2.356748104095459, "loss/hidden": 0.953125, "loss/logits": 0.24928607046604156, "loss/reg": 0.018837420269846916, "step": 962 }, { "epoch": 0.120375, "grad_norm": 4.978577136993408, "grad_norm_var": 3.903770487124885, "learning_rate": 0.0001, "loss": 1.4333, "loss/crossentropy": 2.2456307411193848, "loss/hidden": 1.0703125, "loss/logits": 0.17467612028121948, "loss/reg": 0.01882883533835411, "step": 963 }, { "epoch": 0.1205, "grad_norm": 5.51161003112793, "grad_norm_var": 3.97576236618078, "learning_rate": 0.0001, "loss": 1.2552, "loss/crossentropy": 2.7595887184143066, "loss/hidden": 0.8828125, "loss/logits": 0.18415382504463196, "loss/reg": 0.01882052607834339, "step": 964 }, { "epoch": 0.120625, "grad_norm": 3.2307283878326416, "grad_norm_var": 3.9863892013288393, "learning_rate": 0.0001, "loss": 1.2728, "loss/crossentropy": 2.48502779006958, "loss/hidden": 0.88671875, "loss/logits": 0.19795754551887512, "loss/reg": 0.01881156861782074, "step": 965 }, { "epoch": 0.12075, "grad_norm": 3.441767454147339, "grad_norm_var": 3.7286595116638916, "learning_rate": 0.0001, "loss": 1.182, "loss/crossentropy": 2.508680582046509, "loss/hidden": 0.8359375, "loss/logits": 0.15802894532680511, "loss/reg": 0.018803071230649948, "step": 966 }, { "epoch": 0.120875, "grad_norm": 4.063246726989746, "grad_norm_var": 3.685090208705704, "learning_rate": 0.0001, "loss": 1.3354, "loss/crossentropy": 2.47729229927063, "loss/hidden": 0.984375, "loss/logits": 0.16310644149780273, "loss/reg": 0.018794314935803413, "step": 967 }, { "epoch": 0.121, "grad_norm": 5.8381171226501465, "grad_norm_var": 1.8400005684538645, "learning_rate": 0.0001, "loss": 1.1025, "loss/crossentropy": 2.5666096210479736, "loss/hidden": 0.77734375, "loss/logits": 0.1373094618320465, "loss/reg": 0.018785255029797554, "step": 968 }, { "epoch": 0.121125, "grad_norm": 2.7132527828216553, "grad_norm_var": 1.8649801572693356, "learning_rate": 0.0001, "loss": 1.0912, "loss/crossentropy": 2.4335010051727295, "loss/hidden": 0.76953125, "loss/logits": 0.13391214609146118, "loss/reg": 0.018776265904307365, "step": 969 }, { "epoch": 0.12125, "grad_norm": 2.9154088497161865, "grad_norm_var": 1.938092020210782, "learning_rate": 0.0001, "loss": 1.1314, "loss/crossentropy": 2.523137331008911, "loss/hidden": 0.80078125, "loss/logits": 0.1429774910211563, "loss/reg": 0.01876768097281456, "step": 970 }, { "epoch": 0.121375, "grad_norm": 3.3754584789276123, "grad_norm_var": 1.8726730225127388, "learning_rate": 0.0001, "loss": 1.0747, "loss/crossentropy": 2.6033432483673096, "loss/hidden": 0.74609375, "loss/logits": 0.14106187224388123, "loss/reg": 0.018759164959192276, "step": 971 }, { "epoch": 0.1215, "grad_norm": 2.7394940853118896, "grad_norm_var": 1.9650935524715956, "learning_rate": 0.0001, "loss": 1.0871, "loss/crossentropy": 2.2976291179656982, "loss/hidden": 0.76953125, "loss/logits": 0.1300249695777893, "loss/reg": 0.01874978095293045, "step": 972 }, { "epoch": 0.121625, "grad_norm": 2.9742624759674072, "grad_norm_var": 1.9749925269591906, "learning_rate": 0.0001, "loss": 1.1221, "loss/crossentropy": 2.7049221992492676, "loss/hidden": 0.79296875, "loss/logits": 0.1416921317577362, "loss/reg": 0.018740687519311905, "step": 973 }, { "epoch": 0.12175, "grad_norm": 2.927666187286377, "grad_norm_var": 1.892721758937742, "learning_rate": 0.0001, "loss": 1.0252, "loss/crossentropy": 2.6492226123809814, "loss/hidden": 0.71875, "loss/logits": 0.11909263581037521, "loss/reg": 0.018731672316789627, "step": 974 }, { "epoch": 0.121875, "grad_norm": 16.453693389892578, "grad_norm_var": 11.523681815422924, "learning_rate": 0.0001, "loss": 1.5788, "loss/crossentropy": 2.375779867172241, "loss/hidden": 1.171875, "loss/logits": 0.2197396457195282, "loss/reg": 0.01872306317090988, "step": 975 }, { "epoch": 0.122, "grad_norm": 2.928443193435669, "grad_norm_var": 11.583339951760102, "learning_rate": 0.0001, "loss": 1.0995, "loss/crossentropy": 2.8136162757873535, "loss/hidden": 0.77734375, "loss/logits": 0.13505280017852783, "loss/reg": 0.018714020028710365, "step": 976 }, { "epoch": 0.122125, "grad_norm": 4.540535926818848, "grad_norm_var": 11.07401646408874, "learning_rate": 0.0001, "loss": 1.1806, "loss/crossentropy": 2.5312135219573975, "loss/hidden": 0.83984375, "loss/logits": 0.15369677543640137, "loss/reg": 0.01870504766702652, "step": 977 }, { "epoch": 0.12225, "grad_norm": 2.4433298110961914, "grad_norm_var": 11.358052675820652, "learning_rate": 0.0001, "loss": 1.0167, "loss/crossentropy": 2.578800678253174, "loss/hidden": 0.6953125, "loss/logits": 0.13442841172218323, "loss/reg": 0.018696293234825134, "step": 978 }, { "epoch": 0.122375, "grad_norm": 3.341182231903076, "grad_norm_var": 11.408522912728658, "learning_rate": 0.0001, "loss": 1.1346, "loss/crossentropy": 2.4482579231262207, "loss/hidden": 0.8046875, "loss/logits": 0.14306378364562988, "loss/reg": 0.018688105046749115, "step": 979 }, { "epoch": 0.1225, "grad_norm": 2.433337450027466, "grad_norm_var": 11.51984045745032, "learning_rate": 0.0001, "loss": 0.9764, "loss/crossentropy": 2.321781873703003, "loss/hidden": 0.65625, "loss/logits": 0.13331879675388336, "loss/reg": 0.018679112195968628, "step": 980 }, { "epoch": 0.122625, "grad_norm": 3.6000678539276123, "grad_norm_var": 11.483219758864427, "learning_rate": 0.0001, "loss": 1.1731, "loss/crossentropy": 2.5233397483825684, "loss/hidden": 0.8203125, "loss/logits": 0.16610421240329742, "loss/reg": 0.018670594319701195, "step": 981 }, { "epoch": 0.12275, "grad_norm": 2.8748631477355957, "grad_norm_var": 11.558394893606714, "learning_rate": 0.0001, "loss": 1.0362, "loss/crossentropy": 2.4155187606811523, "loss/hidden": 0.7265625, "loss/logits": 0.12304510176181793, "loss/reg": 0.018661517649888992, "step": 982 }, { "epoch": 0.122875, "grad_norm": 2.2293035984039307, "grad_norm_var": 11.786185692154314, "learning_rate": 0.0001, "loss": 1.0077, "loss/crossentropy": 2.618363618850708, "loss/hidden": 0.6875, "loss/logits": 0.13366106152534485, "loss/reg": 0.01865258812904358, "step": 983 }, { "epoch": 0.123, "grad_norm": 3.615424633026123, "grad_norm_var": 11.556298836968558, "learning_rate": 0.0001, "loss": 1.0544, "loss/crossentropy": 2.224966526031494, "loss/hidden": 0.734375, "loss/logits": 0.1335442066192627, "loss/reg": 0.01864360086619854, "step": 984 }, { "epoch": 0.123125, "grad_norm": 2.93835186958313, "grad_norm_var": 11.524399601900045, "learning_rate": 0.0001, "loss": 1.232, "loss/crossentropy": 2.348299026489258, "loss/hidden": 0.87890625, "loss/logits": 0.1667439341545105, "loss/reg": 0.018634630367159843, "step": 985 }, { "epoch": 0.12325, "grad_norm": 3.2750635147094727, "grad_norm_var": 11.485476360611186, "learning_rate": 0.0001, "loss": 1.0639, "loss/crossentropy": 2.5712990760803223, "loss/hidden": 0.75, "loss/logits": 0.12762659788131714, "loss/reg": 0.018625380471348763, "step": 986 }, { "epoch": 0.123375, "grad_norm": 3.81864595413208, "grad_norm_var": 11.465683474564775, "learning_rate": 0.0001, "loss": 1.1284, "loss/crossentropy": 2.4274511337280273, "loss/hidden": 0.7890625, "loss/logits": 0.15321871638298035, "loss/reg": 0.01861615665256977, "step": 987 }, { "epoch": 0.1235, "grad_norm": 3.2677698135375977, "grad_norm_var": 11.398153583229371, "learning_rate": 0.0001, "loss": 1.1315, "loss/crossentropy": 2.656604290008545, "loss/hidden": 0.8046875, "loss/logits": 0.14076048135757446, "loss/reg": 0.018606893718242645, "step": 988 }, { "epoch": 0.123625, "grad_norm": 3.561713457107544, "grad_norm_var": 11.341034456038914, "learning_rate": 0.0001, "loss": 1.0152, "loss/crossentropy": 2.6205251216888428, "loss/hidden": 0.70703125, "loss/logits": 0.12218683958053589, "loss/reg": 0.01859763078391552, "step": 989 }, { "epoch": 0.12375, "grad_norm": 2.146240472793579, "grad_norm_var": 11.49254916357394, "learning_rate": 0.0001, "loss": 1.0449, "loss/crossentropy": 2.4804821014404297, "loss/hidden": 0.71484375, "loss/logits": 0.14422178268432617, "loss/reg": 0.018588390201330185, "step": 990 }, { "epoch": 0.123875, "grad_norm": 2.6142220497131348, "grad_norm_var": 0.4215380256098586, "learning_rate": 0.0001, "loss": 1.0371, "loss/crossentropy": 2.3303298950195312, "loss/hidden": 0.73828125, "loss/logits": 0.11297546327114105, "loss/reg": 0.018579507246613503, "step": 991 }, { "epoch": 0.124, "grad_norm": 4.17028284072876, "grad_norm_var": 0.4892223582938262, "learning_rate": 0.0001, "loss": 1.6898, "loss/crossentropy": 2.250098943710327, "loss/hidden": 1.203125, "loss/logits": 0.30099910497665405, "loss/reg": 0.018569782376289368, "step": 992 }, { "epoch": 0.124125, "grad_norm": 2.3814339637756348, "grad_norm_var": 0.3887345955884323, "learning_rate": 0.0001, "loss": 1.0837, "loss/crossentropy": 2.6447057723999023, "loss/hidden": 0.74609375, "loss/logits": 0.15198630094528198, "loss/reg": 0.01855996623635292, "step": 993 }, { "epoch": 0.12425, "grad_norm": 4.670334815979004, "grad_norm_var": 0.5202129226035586, "learning_rate": 0.0001, "loss": 1.1908, "loss/crossentropy": 2.3941245079040527, "loss/hidden": 0.85546875, "loss/logits": 0.1498216688632965, "loss/reg": 0.018550006672739983, "step": 994 }, { "epoch": 0.124375, "grad_norm": 2.498332977294922, "grad_norm_var": 0.5469080049785059, "learning_rate": 0.0001, "loss": 0.9226, "loss/crossentropy": 2.489830255508423, "loss/hidden": 0.63671875, "loss/logits": 0.10051175206899643, "loss/reg": 0.01854090392589569, "step": 995 }, { "epoch": 0.1245, "grad_norm": 2.987192392349243, "grad_norm_var": 0.5145625202891589, "learning_rate": 0.0001, "loss": 1.1638, "loss/crossentropy": 2.094637632369995, "loss/hidden": 0.8515625, "loss/logits": 0.12690997123718262, "loss/reg": 0.018531804904341698, "step": 996 }, { "epoch": 0.124625, "grad_norm": 2.738851547241211, "grad_norm_var": 0.5110263660772335, "learning_rate": 0.0001, "loss": 1.1959, "loss/crossentropy": 2.1406211853027344, "loss/hidden": 0.84375, "loss/logits": 0.16688010096549988, "loss/reg": 0.018522722646594048, "step": 997 }, { "epoch": 0.12475, "grad_norm": 5.02874231338501, "grad_norm_var": 0.732945509426732, "learning_rate": 0.0001, "loss": 1.089, "loss/crossentropy": 2.442840337753296, "loss/hidden": 0.76171875, "loss/logits": 0.142162024974823, "loss/reg": 0.018513953313231468, "step": 998 }, { "epoch": 0.124875, "grad_norm": 2.947786331176758, "grad_norm_var": 0.6677765621166297, "learning_rate": 0.0001, "loss": 1.3459, "loss/crossentropy": 2.483854055404663, "loss/hidden": 0.96875, "loss/logits": 0.19213837385177612, "loss/reg": 0.018505612388253212, "step": 999 }, { "epoch": 0.125, "grad_norm": 4.166240692138672, "grad_norm_var": 0.7105452516630361, "learning_rate": 0.0001, "loss": 1.2402, "loss/crossentropy": 2.454993963241577, "loss/hidden": 0.87890625, "loss/logits": 0.17627671360969543, "loss/reg": 0.018497284501791, "step": 1000 } ], "logging_steps": 1, "max_steps": 8000, "num_input_tokens_seen": 0, "num_train_epochs": 9223372036854775807, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": true, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 6.4405861564416e+16, "train_batch_size": 2, "trial_name": null, "trial_params": null }