| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 3.541703559412077, |
| "eval_steps": 500, |
| "global_step": 2500, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.0, |
| "grad_norm": 54.43408203125, |
| "learning_rate": 0.0001, |
| "loss": 6.3724, |
| "step": 1 |
| }, |
| { |
| "epoch": 0.0, |
| "grad_norm": 54.14735794067383, |
| "learning_rate": 0.0002, |
| "loss": 6.3055, |
| "step": 2 |
| }, |
| { |
| "epoch": 0.0, |
| "grad_norm": 19.2130184173584, |
| "learning_rate": 0.00019992902767920512, |
| "loss": 3.739, |
| "step": 3 |
| }, |
| { |
| "epoch": 0.01, |
| "grad_norm": 12.525617599487305, |
| "learning_rate": 0.00019985805535841024, |
| "loss": 2.6998, |
| "step": 4 |
| }, |
| { |
| "epoch": 0.01, |
| "grad_norm": 10.059170722961426, |
| "learning_rate": 0.00019978708303761532, |
| "loss": 2.136, |
| "step": 5 |
| }, |
| { |
| "epoch": 0.01, |
| "grad_norm": 6.011496067047119, |
| "learning_rate": 0.00019971611071682046, |
| "loss": 1.8854, |
| "step": 6 |
| }, |
| { |
| "epoch": 0.01, |
| "grad_norm": 2.9094812870025635, |
| "learning_rate": 0.00019964513839602557, |
| "loss": 1.8689, |
| "step": 7 |
| }, |
| { |
| "epoch": 0.01, |
| "grad_norm": 1.9098069667816162, |
| "learning_rate": 0.00019957416607523069, |
| "loss": 1.762, |
| "step": 8 |
| }, |
| { |
| "epoch": 0.01, |
| "grad_norm": 1.3670941591262817, |
| "learning_rate": 0.00019950319375443577, |
| "loss": 1.6628, |
| "step": 9 |
| }, |
| { |
| "epoch": 0.01, |
| "grad_norm": 1.1179559230804443, |
| "learning_rate": 0.00019943222143364088, |
| "loss": 1.6198, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.02, |
| "grad_norm": 0.896386444568634, |
| "learning_rate": 0.000199361249112846, |
| "loss": 1.5714, |
| "step": 11 |
| }, |
| { |
| "epoch": 0.02, |
| "grad_norm": 0.9485528469085693, |
| "learning_rate": 0.0001992902767920511, |
| "loss": 1.5582, |
| "step": 12 |
| }, |
| { |
| "epoch": 0.02, |
| "grad_norm": 0.9695132970809937, |
| "learning_rate": 0.00019921930447125622, |
| "loss": 1.5308, |
| "step": 13 |
| }, |
| { |
| "epoch": 0.02, |
| "grad_norm": 0.7802467346191406, |
| "learning_rate": 0.00019914833215046134, |
| "loss": 1.5607, |
| "step": 14 |
| }, |
| { |
| "epoch": 0.02, |
| "grad_norm": 0.7041746377944946, |
| "learning_rate": 0.00019907735982966645, |
| "loss": 1.5013, |
| "step": 15 |
| }, |
| { |
| "epoch": 0.02, |
| "grad_norm": 0.6526293158531189, |
| "learning_rate": 0.00019900638750887156, |
| "loss": 1.4903, |
| "step": 16 |
| }, |
| { |
| "epoch": 0.02, |
| "grad_norm": 0.7452432513237, |
| "learning_rate": 0.00019893541518807665, |
| "loss": 1.495, |
| "step": 17 |
| }, |
| { |
| "epoch": 0.03, |
| "grad_norm": 0.642892062664032, |
| "learning_rate": 0.00019886444286728176, |
| "loss": 1.3996, |
| "step": 18 |
| }, |
| { |
| "epoch": 0.03, |
| "grad_norm": 0.608568012714386, |
| "learning_rate": 0.00019879347054648687, |
| "loss": 1.4793, |
| "step": 19 |
| }, |
| { |
| "epoch": 0.03, |
| "grad_norm": 0.6717165112495422, |
| "learning_rate": 0.000198722498225692, |
| "loss": 1.4569, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.03, |
| "grad_norm": 0.6507506370544434, |
| "learning_rate": 0.0001986515259048971, |
| "loss": 1.4318, |
| "step": 21 |
| }, |
| { |
| "epoch": 0.03, |
| "grad_norm": 0.6129789352416992, |
| "learning_rate": 0.0001985805535841022, |
| "loss": 1.4275, |
| "step": 22 |
| }, |
| { |
| "epoch": 0.03, |
| "grad_norm": 0.606242835521698, |
| "learning_rate": 0.00019850958126330732, |
| "loss": 1.3986, |
| "step": 23 |
| }, |
| { |
| "epoch": 0.03, |
| "grad_norm": 0.656890869140625, |
| "learning_rate": 0.00019843860894251244, |
| "loss": 1.4376, |
| "step": 24 |
| }, |
| { |
| "epoch": 0.04, |
| "grad_norm": 0.6069300174713135, |
| "learning_rate": 0.00019836763662171752, |
| "loss": 1.4592, |
| "step": 25 |
| }, |
| { |
| "epoch": 0.04, |
| "grad_norm": 0.6284168362617493, |
| "learning_rate": 0.00019829666430092263, |
| "loss": 1.4375, |
| "step": 26 |
| }, |
| { |
| "epoch": 0.04, |
| "grad_norm": 0.6825727820396423, |
| "learning_rate": 0.00019822569198012777, |
| "loss": 1.4902, |
| "step": 27 |
| }, |
| { |
| "epoch": 0.04, |
| "grad_norm": 0.5566306114196777, |
| "learning_rate": 0.00019815471965933289, |
| "loss": 1.3259, |
| "step": 28 |
| }, |
| { |
| "epoch": 0.04, |
| "grad_norm": 0.5704931020736694, |
| "learning_rate": 0.00019808374733853797, |
| "loss": 1.4229, |
| "step": 29 |
| }, |
| { |
| "epoch": 0.04, |
| "grad_norm": 0.6081483364105225, |
| "learning_rate": 0.00019801277501774308, |
| "loss": 1.3713, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.04, |
| "grad_norm": 0.6242185235023499, |
| "learning_rate": 0.0001979418026969482, |
| "loss": 1.3438, |
| "step": 31 |
| }, |
| { |
| "epoch": 0.05, |
| "grad_norm": 0.5827323198318481, |
| "learning_rate": 0.0001978708303761533, |
| "loss": 1.4987, |
| "step": 32 |
| }, |
| { |
| "epoch": 0.05, |
| "grad_norm": 0.5230266451835632, |
| "learning_rate": 0.00019779985805535842, |
| "loss": 1.3116, |
| "step": 33 |
| }, |
| { |
| "epoch": 0.05, |
| "grad_norm": 0.5829095244407654, |
| "learning_rate": 0.00019772888573456354, |
| "loss": 1.5099, |
| "step": 34 |
| }, |
| { |
| "epoch": 0.05, |
| "grad_norm": 0.61451655626297, |
| "learning_rate": 0.00019765791341376865, |
| "loss": 1.4576, |
| "step": 35 |
| }, |
| { |
| "epoch": 0.05, |
| "grad_norm": 0.5359353423118591, |
| "learning_rate": 0.00019758694109297376, |
| "loss": 1.4095, |
| "step": 36 |
| }, |
| { |
| "epoch": 0.05, |
| "grad_norm": 0.5368149876594543, |
| "learning_rate": 0.00019751596877217887, |
| "loss": 1.3962, |
| "step": 37 |
| }, |
| { |
| "epoch": 0.05, |
| "grad_norm": 0.5719636678695679, |
| "learning_rate": 0.00019744499645138396, |
| "loss": 1.3462, |
| "step": 38 |
| }, |
| { |
| "epoch": 0.06, |
| "grad_norm": 0.5419232845306396, |
| "learning_rate": 0.00019737402413058907, |
| "loss": 1.3416, |
| "step": 39 |
| }, |
| { |
| "epoch": 0.06, |
| "grad_norm": 0.530976414680481, |
| "learning_rate": 0.00019730305180979418, |
| "loss": 1.4029, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.06, |
| "grad_norm": 0.5526696443557739, |
| "learning_rate": 0.0001972320794889993, |
| "loss": 1.3693, |
| "step": 41 |
| }, |
| { |
| "epoch": 0.06, |
| "grad_norm": 0.5667834281921387, |
| "learning_rate": 0.0001971611071682044, |
| "loss": 1.3882, |
| "step": 42 |
| }, |
| { |
| "epoch": 0.06, |
| "grad_norm": 0.5483084917068481, |
| "learning_rate": 0.00019709013484740952, |
| "loss": 1.4203, |
| "step": 43 |
| }, |
| { |
| "epoch": 0.06, |
| "grad_norm": 0.5592283606529236, |
| "learning_rate": 0.00019701916252661464, |
| "loss": 1.366, |
| "step": 44 |
| }, |
| { |
| "epoch": 0.06, |
| "grad_norm": 0.5499030351638794, |
| "learning_rate": 0.00019694819020581975, |
| "loss": 1.3449, |
| "step": 45 |
| }, |
| { |
| "epoch": 0.07, |
| "grad_norm": 0.6002295613288879, |
| "learning_rate": 0.00019687721788502483, |
| "loss": 1.3919, |
| "step": 46 |
| }, |
| { |
| "epoch": 0.07, |
| "grad_norm": 0.5805856585502625, |
| "learning_rate": 0.00019680624556422995, |
| "loss": 1.317, |
| "step": 47 |
| }, |
| { |
| "epoch": 0.07, |
| "grad_norm": 0.5594860911369324, |
| "learning_rate": 0.00019673527324343506, |
| "loss": 1.4043, |
| "step": 48 |
| }, |
| { |
| "epoch": 0.07, |
| "grad_norm": 0.5409876108169556, |
| "learning_rate": 0.0001966643009226402, |
| "loss": 1.411, |
| "step": 49 |
| }, |
| { |
| "epoch": 0.07, |
| "grad_norm": 0.5239064693450928, |
| "learning_rate": 0.00019659332860184528, |
| "loss": 1.4308, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.07, |
| "grad_norm": 0.5057370662689209, |
| "learning_rate": 0.0001965223562810504, |
| "loss": 1.3265, |
| "step": 51 |
| }, |
| { |
| "epoch": 0.07, |
| "grad_norm": 0.5035980939865112, |
| "learning_rate": 0.0001964513839602555, |
| "loss": 1.3382, |
| "step": 52 |
| }, |
| { |
| "epoch": 0.08, |
| "grad_norm": 0.5412520170211792, |
| "learning_rate": 0.00019638041163946062, |
| "loss": 1.3436, |
| "step": 53 |
| }, |
| { |
| "epoch": 0.08, |
| "grad_norm": 0.5295203328132629, |
| "learning_rate": 0.0001963094393186657, |
| "loss": 1.4424, |
| "step": 54 |
| }, |
| { |
| "epoch": 0.08, |
| "grad_norm": 0.5216801166534424, |
| "learning_rate": 0.00019623846699787082, |
| "loss": 1.3575, |
| "step": 55 |
| }, |
| { |
| "epoch": 0.08, |
| "grad_norm": 0.5418757200241089, |
| "learning_rate": 0.00019616749467707596, |
| "loss": 1.2319, |
| "step": 56 |
| }, |
| { |
| "epoch": 0.08, |
| "grad_norm": 0.5106856822967529, |
| "learning_rate": 0.00019609652235628107, |
| "loss": 1.3407, |
| "step": 57 |
| }, |
| { |
| "epoch": 0.08, |
| "grad_norm": 0.5250461101531982, |
| "learning_rate": 0.00019602555003548616, |
| "loss": 1.3413, |
| "step": 58 |
| }, |
| { |
| "epoch": 0.08, |
| "grad_norm": 0.5034674406051636, |
| "learning_rate": 0.00019595457771469127, |
| "loss": 1.3602, |
| "step": 59 |
| }, |
| { |
| "epoch": 0.09, |
| "grad_norm": 0.5334136486053467, |
| "learning_rate": 0.00019588360539389639, |
| "loss": 1.3282, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.09, |
| "grad_norm": 0.5571975708007812, |
| "learning_rate": 0.0001958126330731015, |
| "loss": 1.4221, |
| "step": 61 |
| }, |
| { |
| "epoch": 0.09, |
| "grad_norm": 0.48554927110671997, |
| "learning_rate": 0.0001957416607523066, |
| "loss": 1.3636, |
| "step": 62 |
| }, |
| { |
| "epoch": 0.09, |
| "grad_norm": 0.5140034556388855, |
| "learning_rate": 0.00019567068843151172, |
| "loss": 1.3278, |
| "step": 63 |
| }, |
| { |
| "epoch": 0.09, |
| "grad_norm": 0.5018851161003113, |
| "learning_rate": 0.00019559971611071684, |
| "loss": 1.3048, |
| "step": 64 |
| }, |
| { |
| "epoch": 0.09, |
| "grad_norm": 0.543761670589447, |
| "learning_rate": 0.00019552874378992195, |
| "loss": 1.3423, |
| "step": 65 |
| }, |
| { |
| "epoch": 0.09, |
| "grad_norm": 0.5353609919548035, |
| "learning_rate": 0.00019545777146912706, |
| "loss": 1.3029, |
| "step": 66 |
| }, |
| { |
| "epoch": 0.09, |
| "grad_norm": 0.5764626264572144, |
| "learning_rate": 0.00019538679914833215, |
| "loss": 1.409, |
| "step": 67 |
| }, |
| { |
| "epoch": 0.1, |
| "grad_norm": 0.4783156216144562, |
| "learning_rate": 0.00019531582682753726, |
| "loss": 1.3306, |
| "step": 68 |
| }, |
| { |
| "epoch": 0.1, |
| "grad_norm": 0.48478803038597107, |
| "learning_rate": 0.00019524485450674237, |
| "loss": 1.4423, |
| "step": 69 |
| }, |
| { |
| "epoch": 0.1, |
| "grad_norm": 0.5065445899963379, |
| "learning_rate": 0.00019517388218594749, |
| "loss": 1.2997, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.1, |
| "grad_norm": 0.4838726222515106, |
| "learning_rate": 0.0001951029098651526, |
| "loss": 1.3767, |
| "step": 71 |
| }, |
| { |
| "epoch": 0.1, |
| "grad_norm": 0.6124043464660645, |
| "learning_rate": 0.0001950319375443577, |
| "loss": 1.305, |
| "step": 72 |
| }, |
| { |
| "epoch": 0.1, |
| "grad_norm": 0.5889860391616821, |
| "learning_rate": 0.00019496096522356282, |
| "loss": 1.3477, |
| "step": 73 |
| }, |
| { |
| "epoch": 0.1, |
| "grad_norm": 0.5476070046424866, |
| "learning_rate": 0.00019488999290276794, |
| "loss": 1.3736, |
| "step": 74 |
| }, |
| { |
| "epoch": 0.11, |
| "grad_norm": 0.4919136166572571, |
| "learning_rate": 0.00019481902058197302, |
| "loss": 1.3518, |
| "step": 75 |
| }, |
| { |
| "epoch": 0.11, |
| "grad_norm": 0.5444558262825012, |
| "learning_rate": 0.00019474804826117813, |
| "loss": 1.4115, |
| "step": 76 |
| }, |
| { |
| "epoch": 0.11, |
| "grad_norm": 0.51210618019104, |
| "learning_rate": 0.00019467707594038327, |
| "loss": 1.3419, |
| "step": 77 |
| }, |
| { |
| "epoch": 0.11, |
| "grad_norm": 0.5465154647827148, |
| "learning_rate": 0.0001946061036195884, |
| "loss": 1.3435, |
| "step": 78 |
| }, |
| { |
| "epoch": 0.11, |
| "grad_norm": 0.4859498143196106, |
| "learning_rate": 0.00019453513129879347, |
| "loss": 1.3392, |
| "step": 79 |
| }, |
| { |
| "epoch": 0.11, |
| "grad_norm": 0.48886799812316895, |
| "learning_rate": 0.00019446415897799859, |
| "loss": 1.3239, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.11, |
| "grad_norm": 0.514495313167572, |
| "learning_rate": 0.0001943931866572037, |
| "loss": 1.3007, |
| "step": 81 |
| }, |
| { |
| "epoch": 0.12, |
| "grad_norm": 0.5247107148170471, |
| "learning_rate": 0.0001943222143364088, |
| "loss": 1.3192, |
| "step": 82 |
| }, |
| { |
| "epoch": 0.12, |
| "grad_norm": 0.5341958999633789, |
| "learning_rate": 0.00019425124201561392, |
| "loss": 1.3269, |
| "step": 83 |
| }, |
| { |
| "epoch": 0.12, |
| "grad_norm": 0.5308558940887451, |
| "learning_rate": 0.00019418026969481904, |
| "loss": 1.3739, |
| "step": 84 |
| }, |
| { |
| "epoch": 0.12, |
| "grad_norm": 0.5067780613899231, |
| "learning_rate": 0.00019410929737402415, |
| "loss": 1.2969, |
| "step": 85 |
| }, |
| { |
| "epoch": 0.12, |
| "grad_norm": 0.5011096596717834, |
| "learning_rate": 0.00019403832505322926, |
| "loss": 1.3132, |
| "step": 86 |
| }, |
| { |
| "epoch": 0.12, |
| "grad_norm": 0.49596866965293884, |
| "learning_rate": 0.00019396735273243435, |
| "loss": 1.3118, |
| "step": 87 |
| }, |
| { |
| "epoch": 0.12, |
| "grad_norm": 0.5615769624710083, |
| "learning_rate": 0.00019389638041163946, |
| "loss": 1.4271, |
| "step": 88 |
| }, |
| { |
| "epoch": 0.13, |
| "grad_norm": 0.5133160352706909, |
| "learning_rate": 0.00019382540809084457, |
| "loss": 1.4121, |
| "step": 89 |
| }, |
| { |
| "epoch": 0.13, |
| "grad_norm": 0.5177298784255981, |
| "learning_rate": 0.00019375443577004969, |
| "loss": 1.3674, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.13, |
| "grad_norm": 0.5397670269012451, |
| "learning_rate": 0.0001936834634492548, |
| "loss": 1.3749, |
| "step": 91 |
| }, |
| { |
| "epoch": 0.13, |
| "grad_norm": 0.510032057762146, |
| "learning_rate": 0.0001936124911284599, |
| "loss": 1.4079, |
| "step": 92 |
| }, |
| { |
| "epoch": 0.13, |
| "grad_norm": 0.49133965373039246, |
| "learning_rate": 0.00019354151880766502, |
| "loss": 1.3387, |
| "step": 93 |
| }, |
| { |
| "epoch": 0.13, |
| "grad_norm": 0.5044322609901428, |
| "learning_rate": 0.00019347054648687014, |
| "loss": 1.3763, |
| "step": 94 |
| }, |
| { |
| "epoch": 0.13, |
| "grad_norm": 0.5398576259613037, |
| "learning_rate": 0.00019339957416607525, |
| "loss": 1.342, |
| "step": 95 |
| }, |
| { |
| "epoch": 0.14, |
| "grad_norm": 0.5276138782501221, |
| "learning_rate": 0.00019332860184528033, |
| "loss": 1.3398, |
| "step": 96 |
| }, |
| { |
| "epoch": 0.14, |
| "grad_norm": 0.49148228764533997, |
| "learning_rate": 0.00019325762952448545, |
| "loss": 1.3678, |
| "step": 97 |
| }, |
| { |
| "epoch": 0.14, |
| "grad_norm": 0.5225054025650024, |
| "learning_rate": 0.0001931866572036906, |
| "loss": 1.297, |
| "step": 98 |
| }, |
| { |
| "epoch": 0.14, |
| "grad_norm": 0.5107095241546631, |
| "learning_rate": 0.0001931156848828957, |
| "loss": 1.2915, |
| "step": 99 |
| }, |
| { |
| "epoch": 0.14, |
| "grad_norm": 0.5660765171051025, |
| "learning_rate": 0.00019304471256210079, |
| "loss": 1.3034, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.14, |
| "grad_norm": 0.5086053013801575, |
| "learning_rate": 0.0001929737402413059, |
| "loss": 1.2538, |
| "step": 101 |
| }, |
| { |
| "epoch": 0.14, |
| "grad_norm": 0.5296218991279602, |
| "learning_rate": 0.000192902767920511, |
| "loss": 1.3653, |
| "step": 102 |
| }, |
| { |
| "epoch": 0.15, |
| "grad_norm": 0.505626380443573, |
| "learning_rate": 0.00019283179559971612, |
| "loss": 1.3468, |
| "step": 103 |
| }, |
| { |
| "epoch": 0.15, |
| "grad_norm": 0.48313605785369873, |
| "learning_rate": 0.0001927608232789212, |
| "loss": 1.3689, |
| "step": 104 |
| }, |
| { |
| "epoch": 0.15, |
| "grad_norm": 0.5230417847633362, |
| "learning_rate": 0.00019268985095812635, |
| "loss": 1.3385, |
| "step": 105 |
| }, |
| { |
| "epoch": 0.15, |
| "grad_norm": 0.548197865486145, |
| "learning_rate": 0.00019261887863733146, |
| "loss": 1.4077, |
| "step": 106 |
| }, |
| { |
| "epoch": 0.15, |
| "grad_norm": 0.49077072739601135, |
| "learning_rate": 0.00019254790631653657, |
| "loss": 1.3333, |
| "step": 107 |
| }, |
| { |
| "epoch": 0.15, |
| "grad_norm": 0.49506676197052, |
| "learning_rate": 0.00019247693399574166, |
| "loss": 1.3019, |
| "step": 108 |
| }, |
| { |
| "epoch": 0.15, |
| "grad_norm": 0.5120548009872437, |
| "learning_rate": 0.00019240596167494677, |
| "loss": 1.2955, |
| "step": 109 |
| }, |
| { |
| "epoch": 0.16, |
| "grad_norm": 0.5009211301803589, |
| "learning_rate": 0.00019233498935415189, |
| "loss": 1.33, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.16, |
| "grad_norm": 0.4983401894569397, |
| "learning_rate": 0.000192264017033357, |
| "loss": 1.2762, |
| "step": 111 |
| }, |
| { |
| "epoch": 0.16, |
| "grad_norm": 0.5030891299247742, |
| "learning_rate": 0.0001921930447125621, |
| "loss": 1.2902, |
| "step": 112 |
| }, |
| { |
| "epoch": 0.16, |
| "grad_norm": 0.5501623749732971, |
| "learning_rate": 0.00019212207239176722, |
| "loss": 1.3797, |
| "step": 113 |
| }, |
| { |
| "epoch": 0.16, |
| "grad_norm": 0.5063967704772949, |
| "learning_rate": 0.00019205110007097234, |
| "loss": 1.2997, |
| "step": 114 |
| }, |
| { |
| "epoch": 0.16, |
| "grad_norm": 0.5631483793258667, |
| "learning_rate": 0.00019198012775017745, |
| "loss": 1.274, |
| "step": 115 |
| }, |
| { |
| "epoch": 0.16, |
| "grad_norm": 0.537670373916626, |
| "learning_rate": 0.00019190915542938254, |
| "loss": 1.3115, |
| "step": 116 |
| }, |
| { |
| "epoch": 0.17, |
| "grad_norm": 0.541203498840332, |
| "learning_rate": 0.00019183818310858765, |
| "loss": 1.3092, |
| "step": 117 |
| }, |
| { |
| "epoch": 0.17, |
| "grad_norm": 0.5539008975028992, |
| "learning_rate": 0.00019176721078779276, |
| "loss": 1.3223, |
| "step": 118 |
| }, |
| { |
| "epoch": 0.17, |
| "grad_norm": 0.5102487802505493, |
| "learning_rate": 0.0001916962384669979, |
| "loss": 1.3181, |
| "step": 119 |
| }, |
| { |
| "epoch": 0.17, |
| "grad_norm": 0.502543032169342, |
| "learning_rate": 0.00019162526614620299, |
| "loss": 1.4029, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.17, |
| "grad_norm": 0.5096634030342102, |
| "learning_rate": 0.0001915542938254081, |
| "loss": 1.308, |
| "step": 121 |
| }, |
| { |
| "epoch": 0.17, |
| "grad_norm": 0.547661542892456, |
| "learning_rate": 0.0001914833215046132, |
| "loss": 1.3109, |
| "step": 122 |
| }, |
| { |
| "epoch": 0.17, |
| "grad_norm": 0.4794137179851532, |
| "learning_rate": 0.00019141234918381832, |
| "loss": 1.3134, |
| "step": 123 |
| }, |
| { |
| "epoch": 0.18, |
| "grad_norm": 0.542648434638977, |
| "learning_rate": 0.00019134137686302344, |
| "loss": 1.3346, |
| "step": 124 |
| }, |
| { |
| "epoch": 0.18, |
| "grad_norm": 0.5438342690467834, |
| "learning_rate": 0.00019127040454222852, |
| "loss": 1.2907, |
| "step": 125 |
| }, |
| { |
| "epoch": 0.18, |
| "grad_norm": 0.5569918155670166, |
| "learning_rate": 0.00019119943222143366, |
| "loss": 1.348, |
| "step": 126 |
| }, |
| { |
| "epoch": 0.18, |
| "grad_norm": 0.5183106064796448, |
| "learning_rate": 0.00019112845990063878, |
| "loss": 1.3846, |
| "step": 127 |
| }, |
| { |
| "epoch": 0.18, |
| "grad_norm": 0.5425500869750977, |
| "learning_rate": 0.0001910574875798439, |
| "loss": 1.2643, |
| "step": 128 |
| }, |
| { |
| "epoch": 0.18, |
| "grad_norm": 0.6269490718841553, |
| "learning_rate": 0.00019098651525904897, |
| "loss": 1.2857, |
| "step": 129 |
| }, |
| { |
| "epoch": 0.18, |
| "grad_norm": 0.5118828415870667, |
| "learning_rate": 0.00019091554293825409, |
| "loss": 1.3618, |
| "step": 130 |
| }, |
| { |
| "epoch": 0.19, |
| "grad_norm": 0.48022595047950745, |
| "learning_rate": 0.0001908445706174592, |
| "loss": 1.3721, |
| "step": 131 |
| }, |
| { |
| "epoch": 0.19, |
| "grad_norm": 0.5537993907928467, |
| "learning_rate": 0.0001907735982966643, |
| "loss": 1.2895, |
| "step": 132 |
| }, |
| { |
| "epoch": 0.19, |
| "grad_norm": 0.5127624869346619, |
| "learning_rate": 0.00019070262597586942, |
| "loss": 1.3227, |
| "step": 133 |
| }, |
| { |
| "epoch": 0.19, |
| "grad_norm": 0.49052900075912476, |
| "learning_rate": 0.00019063165365507454, |
| "loss": 1.2652, |
| "step": 134 |
| }, |
| { |
| "epoch": 0.19, |
| "grad_norm": 0.499605268239975, |
| "learning_rate": 0.00019056068133427965, |
| "loss": 1.2807, |
| "step": 135 |
| }, |
| { |
| "epoch": 0.19, |
| "grad_norm": 0.514790952205658, |
| "learning_rate": 0.00019048970901348476, |
| "loss": 1.2387, |
| "step": 136 |
| }, |
| { |
| "epoch": 0.19, |
| "grad_norm": 0.5162804126739502, |
| "learning_rate": 0.00019041873669268985, |
| "loss": 1.3021, |
| "step": 137 |
| }, |
| { |
| "epoch": 0.2, |
| "grad_norm": 0.5387127995491028, |
| "learning_rate": 0.00019034776437189496, |
| "loss": 1.3652, |
| "step": 138 |
| }, |
| { |
| "epoch": 0.2, |
| "grad_norm": 0.5182953476905823, |
| "learning_rate": 0.00019027679205110007, |
| "loss": 1.3704, |
| "step": 139 |
| }, |
| { |
| "epoch": 0.2, |
| "grad_norm": 0.5148733854293823, |
| "learning_rate": 0.00019020581973030519, |
| "loss": 1.2932, |
| "step": 140 |
| }, |
| { |
| "epoch": 0.2, |
| "grad_norm": 0.5453186631202698, |
| "learning_rate": 0.0001901348474095103, |
| "loss": 1.3298, |
| "step": 141 |
| }, |
| { |
| "epoch": 0.2, |
| "grad_norm": 0.49780547618865967, |
| "learning_rate": 0.0001900638750887154, |
| "loss": 1.3241, |
| "step": 142 |
| }, |
| { |
| "epoch": 0.2, |
| "grad_norm": 0.563477635383606, |
| "learning_rate": 0.00018999290276792052, |
| "loss": 1.3003, |
| "step": 143 |
| }, |
| { |
| "epoch": 0.2, |
| "grad_norm": 0.6826238036155701, |
| "learning_rate": 0.00018992193044712564, |
| "loss": 1.2782, |
| "step": 144 |
| }, |
| { |
| "epoch": 0.21, |
| "grad_norm": 0.5311530828475952, |
| "learning_rate": 0.00018985095812633072, |
| "loss": 1.2525, |
| "step": 145 |
| }, |
| { |
| "epoch": 0.21, |
| "grad_norm": 0.5208378434181213, |
| "learning_rate": 0.00018977998580553584, |
| "loss": 1.2451, |
| "step": 146 |
| }, |
| { |
| "epoch": 0.21, |
| "grad_norm": 0.5175229907035828, |
| "learning_rate": 0.00018970901348474095, |
| "loss": 1.2783, |
| "step": 147 |
| }, |
| { |
| "epoch": 0.21, |
| "grad_norm": 0.5157881379127502, |
| "learning_rate": 0.0001896380411639461, |
| "loss": 1.3475, |
| "step": 148 |
| }, |
| { |
| "epoch": 0.21, |
| "grad_norm": 0.510566234588623, |
| "learning_rate": 0.00018956706884315117, |
| "loss": 1.3182, |
| "step": 149 |
| }, |
| { |
| "epoch": 0.21, |
| "grad_norm": 0.520010232925415, |
| "learning_rate": 0.00018949609652235629, |
| "loss": 1.3495, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.21, |
| "grad_norm": 0.5622514486312866, |
| "learning_rate": 0.0001894251242015614, |
| "loss": 1.3128, |
| "step": 151 |
| }, |
| { |
| "epoch": 0.22, |
| "grad_norm": 0.5312150120735168, |
| "learning_rate": 0.0001893541518807665, |
| "loss": 1.3203, |
| "step": 152 |
| }, |
| { |
| "epoch": 0.22, |
| "grad_norm": 0.5105769634246826, |
| "learning_rate": 0.00018928317955997162, |
| "loss": 1.3155, |
| "step": 153 |
| }, |
| { |
| "epoch": 0.22, |
| "grad_norm": 0.5248401165008545, |
| "learning_rate": 0.0001892122072391767, |
| "loss": 1.2616, |
| "step": 154 |
| }, |
| { |
| "epoch": 0.22, |
| "grad_norm": 0.5388246774673462, |
| "learning_rate": 0.00018914123491838185, |
| "loss": 1.2817, |
| "step": 155 |
| }, |
| { |
| "epoch": 0.22, |
| "grad_norm": 0.5068986415863037, |
| "learning_rate": 0.00018907026259758696, |
| "loss": 1.2676, |
| "step": 156 |
| }, |
| { |
| "epoch": 0.22, |
| "grad_norm": 0.5076289176940918, |
| "learning_rate": 0.00018899929027679208, |
| "loss": 1.3285, |
| "step": 157 |
| }, |
| { |
| "epoch": 0.22, |
| "grad_norm": 0.5198759436607361, |
| "learning_rate": 0.00018892831795599716, |
| "loss": 1.2746, |
| "step": 158 |
| }, |
| { |
| "epoch": 0.23, |
| "grad_norm": 0.5820220112800598, |
| "learning_rate": 0.00018885734563520227, |
| "loss": 1.3505, |
| "step": 159 |
| }, |
| { |
| "epoch": 0.23, |
| "grad_norm": 0.5358129143714905, |
| "learning_rate": 0.0001887863733144074, |
| "loss": 1.2784, |
| "step": 160 |
| }, |
| { |
| "epoch": 0.23, |
| "grad_norm": 0.5665339827537537, |
| "learning_rate": 0.0001887154009936125, |
| "loss": 1.3348, |
| "step": 161 |
| }, |
| { |
| "epoch": 0.23, |
| "grad_norm": 0.5267711877822876, |
| "learning_rate": 0.0001886444286728176, |
| "loss": 1.3123, |
| "step": 162 |
| }, |
| { |
| "epoch": 0.23, |
| "grad_norm": 0.5009903907775879, |
| "learning_rate": 0.00018857345635202272, |
| "loss": 1.274, |
| "step": 163 |
| }, |
| { |
| "epoch": 0.23, |
| "grad_norm": 0.6048489212989807, |
| "learning_rate": 0.00018850248403122784, |
| "loss": 1.3001, |
| "step": 164 |
| }, |
| { |
| "epoch": 0.23, |
| "grad_norm": 0.5325021147727966, |
| "learning_rate": 0.00018843151171043295, |
| "loss": 1.2895, |
| "step": 165 |
| }, |
| { |
| "epoch": 0.24, |
| "grad_norm": 0.6042301654815674, |
| "learning_rate": 0.00018836053938963804, |
| "loss": 1.315, |
| "step": 166 |
| }, |
| { |
| "epoch": 0.24, |
| "grad_norm": 0.567377507686615, |
| "learning_rate": 0.00018828956706884315, |
| "loss": 1.2663, |
| "step": 167 |
| }, |
| { |
| "epoch": 0.24, |
| "grad_norm": 0.5558852553367615, |
| "learning_rate": 0.00018821859474804826, |
| "loss": 1.3094, |
| "step": 168 |
| }, |
| { |
| "epoch": 0.24, |
| "grad_norm": 0.5398116707801819, |
| "learning_rate": 0.0001881476224272534, |
| "loss": 1.2863, |
| "step": 169 |
| }, |
| { |
| "epoch": 0.24, |
| "grad_norm": 0.5448858141899109, |
| "learning_rate": 0.0001880766501064585, |
| "loss": 1.2846, |
| "step": 170 |
| }, |
| { |
| "epoch": 0.24, |
| "grad_norm": 0.5054113268852234, |
| "learning_rate": 0.0001880056777856636, |
| "loss": 1.2498, |
| "step": 171 |
| }, |
| { |
| "epoch": 0.24, |
| "grad_norm": 0.5151749849319458, |
| "learning_rate": 0.0001879347054648687, |
| "loss": 1.2958, |
| "step": 172 |
| }, |
| { |
| "epoch": 0.25, |
| "grad_norm": 0.5271678566932678, |
| "learning_rate": 0.00018786373314407382, |
| "loss": 1.3349, |
| "step": 173 |
| }, |
| { |
| "epoch": 0.25, |
| "grad_norm": 0.5065703392028809, |
| "learning_rate": 0.0001877927608232789, |
| "loss": 1.2965, |
| "step": 174 |
| }, |
| { |
| "epoch": 0.25, |
| "grad_norm": 0.5131989121437073, |
| "learning_rate": 0.00018772178850248402, |
| "loss": 1.3322, |
| "step": 175 |
| }, |
| { |
| "epoch": 0.25, |
| "grad_norm": 0.5324966907501221, |
| "learning_rate": 0.00018765081618168916, |
| "loss": 1.2557, |
| "step": 176 |
| }, |
| { |
| "epoch": 0.25, |
| "grad_norm": 0.5682156085968018, |
| "learning_rate": 0.00018757984386089428, |
| "loss": 1.352, |
| "step": 177 |
| }, |
| { |
| "epoch": 0.25, |
| "grad_norm": 0.5812451243400574, |
| "learning_rate": 0.00018750887154009936, |
| "loss": 1.3167, |
| "step": 178 |
| }, |
| { |
| "epoch": 0.25, |
| "grad_norm": 0.5220502018928528, |
| "learning_rate": 0.00018743789921930447, |
| "loss": 1.3982, |
| "step": 179 |
| }, |
| { |
| "epoch": 0.26, |
| "grad_norm": 0.519579291343689, |
| "learning_rate": 0.0001873669268985096, |
| "loss": 1.2545, |
| "step": 180 |
| }, |
| { |
| "epoch": 0.26, |
| "grad_norm": 0.5515694618225098, |
| "learning_rate": 0.0001872959545777147, |
| "loss": 1.2739, |
| "step": 181 |
| }, |
| { |
| "epoch": 0.26, |
| "grad_norm": 0.494443416595459, |
| "learning_rate": 0.0001872249822569198, |
| "loss": 1.3223, |
| "step": 182 |
| }, |
| { |
| "epoch": 0.26, |
| "grad_norm": 0.5370269417762756, |
| "learning_rate": 0.00018715400993612493, |
| "loss": 1.2643, |
| "step": 183 |
| }, |
| { |
| "epoch": 0.26, |
| "grad_norm": 0.5089952945709229, |
| "learning_rate": 0.00018708303761533004, |
| "loss": 1.2543, |
| "step": 184 |
| }, |
| { |
| "epoch": 0.26, |
| "grad_norm": 0.5908558964729309, |
| "learning_rate": 0.00018701206529453515, |
| "loss": 1.3621, |
| "step": 185 |
| }, |
| { |
| "epoch": 0.26, |
| "grad_norm": 0.5236672163009644, |
| "learning_rate": 0.00018694109297374026, |
| "loss": 1.2589, |
| "step": 186 |
| }, |
| { |
| "epoch": 0.26, |
| "grad_norm": 0.5075586438179016, |
| "learning_rate": 0.00018687012065294535, |
| "loss": 1.2567, |
| "step": 187 |
| }, |
| { |
| "epoch": 0.27, |
| "grad_norm": 0.484017938375473, |
| "learning_rate": 0.00018679914833215046, |
| "loss": 1.253, |
| "step": 188 |
| }, |
| { |
| "epoch": 0.27, |
| "grad_norm": 0.49491867423057556, |
| "learning_rate": 0.00018672817601135557, |
| "loss": 1.2903, |
| "step": 189 |
| }, |
| { |
| "epoch": 0.27, |
| "grad_norm": 0.5130958557128906, |
| "learning_rate": 0.0001866572036905607, |
| "loss": 1.3576, |
| "step": 190 |
| }, |
| { |
| "epoch": 0.27, |
| "grad_norm": 0.5242413282394409, |
| "learning_rate": 0.0001865862313697658, |
| "loss": 1.259, |
| "step": 191 |
| }, |
| { |
| "epoch": 0.27, |
| "grad_norm": 0.5665105581283569, |
| "learning_rate": 0.0001865152590489709, |
| "loss": 1.2892, |
| "step": 192 |
| }, |
| { |
| "epoch": 0.27, |
| "grad_norm": 0.5331645011901855, |
| "learning_rate": 0.00018644428672817603, |
| "loss": 1.3102, |
| "step": 193 |
| }, |
| { |
| "epoch": 0.27, |
| "grad_norm": 0.4966602921485901, |
| "learning_rate": 0.00018637331440738114, |
| "loss": 1.2915, |
| "step": 194 |
| }, |
| { |
| "epoch": 0.28, |
| "grad_norm": 0.5544999837875366, |
| "learning_rate": 0.00018630234208658622, |
| "loss": 1.258, |
| "step": 195 |
| }, |
| { |
| "epoch": 0.28, |
| "grad_norm": 0.520144522190094, |
| "learning_rate": 0.00018623136976579134, |
| "loss": 1.3164, |
| "step": 196 |
| }, |
| { |
| "epoch": 0.28, |
| "grad_norm": 0.5174679160118103, |
| "learning_rate": 0.00018616039744499648, |
| "loss": 1.344, |
| "step": 197 |
| }, |
| { |
| "epoch": 0.28, |
| "grad_norm": 0.4990881383419037, |
| "learning_rate": 0.0001860894251242016, |
| "loss": 1.3103, |
| "step": 198 |
| }, |
| { |
| "epoch": 0.28, |
| "grad_norm": 0.5118770003318787, |
| "learning_rate": 0.00018601845280340667, |
| "loss": 1.2647, |
| "step": 199 |
| }, |
| { |
| "epoch": 0.28, |
| "grad_norm": 0.5196303725242615, |
| "learning_rate": 0.0001859474804826118, |
| "loss": 1.336, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.28, |
| "grad_norm": 0.5092043280601501, |
| "learning_rate": 0.0001858765081618169, |
| "loss": 1.2411, |
| "step": 201 |
| }, |
| { |
| "epoch": 0.29, |
| "grad_norm": 0.5203234553337097, |
| "learning_rate": 0.000185805535841022, |
| "loss": 1.2752, |
| "step": 202 |
| }, |
| { |
| "epoch": 0.29, |
| "grad_norm": 0.5390545725822449, |
| "learning_rate": 0.0001857345635202271, |
| "loss": 1.2102, |
| "step": 203 |
| }, |
| { |
| "epoch": 0.29, |
| "grad_norm": 0.5045953392982483, |
| "learning_rate": 0.00018566359119943224, |
| "loss": 1.3237, |
| "step": 204 |
| }, |
| { |
| "epoch": 0.29, |
| "grad_norm": 0.5172394514083862, |
| "learning_rate": 0.00018559261887863735, |
| "loss": 1.2755, |
| "step": 205 |
| }, |
| { |
| "epoch": 0.29, |
| "grad_norm": 0.5241258144378662, |
| "learning_rate": 0.00018552164655784246, |
| "loss": 1.3713, |
| "step": 206 |
| }, |
| { |
| "epoch": 0.29, |
| "grad_norm": 0.541176438331604, |
| "learning_rate": 0.00018545067423704755, |
| "loss": 1.2301, |
| "step": 207 |
| }, |
| { |
| "epoch": 0.29, |
| "grad_norm": 0.500068187713623, |
| "learning_rate": 0.00018537970191625266, |
| "loss": 1.2682, |
| "step": 208 |
| }, |
| { |
| "epoch": 0.3, |
| "grad_norm": 0.5059691071510315, |
| "learning_rate": 0.00018530872959545777, |
| "loss": 1.2505, |
| "step": 209 |
| }, |
| { |
| "epoch": 0.3, |
| "grad_norm": 0.49532344937324524, |
| "learning_rate": 0.0001852377572746629, |
| "loss": 1.2387, |
| "step": 210 |
| }, |
| { |
| "epoch": 0.3, |
| "grad_norm": 0.6257210373878479, |
| "learning_rate": 0.000185166784953868, |
| "loss": 1.3048, |
| "step": 211 |
| }, |
| { |
| "epoch": 0.3, |
| "grad_norm": 0.5715426206588745, |
| "learning_rate": 0.0001850958126330731, |
| "loss": 1.3693, |
| "step": 212 |
| }, |
| { |
| "epoch": 0.3, |
| "grad_norm": 0.5583224296569824, |
| "learning_rate": 0.00018502484031227823, |
| "loss": 1.2703, |
| "step": 213 |
| }, |
| { |
| "epoch": 0.3, |
| "grad_norm": 0.6376559138298035, |
| "learning_rate": 0.00018495386799148334, |
| "loss": 1.3016, |
| "step": 214 |
| }, |
| { |
| "epoch": 0.3, |
| "grad_norm": 0.535058856010437, |
| "learning_rate": 0.00018488289567068845, |
| "loss": 1.3551, |
| "step": 215 |
| }, |
| { |
| "epoch": 0.31, |
| "grad_norm": 0.5314571857452393, |
| "learning_rate": 0.00018481192334989354, |
| "loss": 1.2762, |
| "step": 216 |
| }, |
| { |
| "epoch": 0.31, |
| "grad_norm": 0.5577436089515686, |
| "learning_rate": 0.00018474095102909865, |
| "loss": 1.2742, |
| "step": 217 |
| }, |
| { |
| "epoch": 0.31, |
| "grad_norm": 0.5784611701965332, |
| "learning_rate": 0.0001846699787083038, |
| "loss": 1.3804, |
| "step": 218 |
| }, |
| { |
| "epoch": 0.31, |
| "grad_norm": 0.5190976858139038, |
| "learning_rate": 0.00018459900638750887, |
| "loss": 1.2978, |
| "step": 219 |
| }, |
| { |
| "epoch": 0.31, |
| "grad_norm": 0.5705554485321045, |
| "learning_rate": 0.000184528034066714, |
| "loss": 1.3462, |
| "step": 220 |
| }, |
| { |
| "epoch": 0.31, |
| "grad_norm": 0.5286474227905273, |
| "learning_rate": 0.0001844570617459191, |
| "loss": 1.3275, |
| "step": 221 |
| }, |
| { |
| "epoch": 0.31, |
| "grad_norm": 0.5056504607200623, |
| "learning_rate": 0.0001843860894251242, |
| "loss": 1.2601, |
| "step": 222 |
| }, |
| { |
| "epoch": 0.32, |
| "grad_norm": 0.5417407751083374, |
| "learning_rate": 0.00018431511710432933, |
| "loss": 1.2604, |
| "step": 223 |
| }, |
| { |
| "epoch": 0.32, |
| "grad_norm": 0.5213840007781982, |
| "learning_rate": 0.0001842441447835344, |
| "loss": 1.2505, |
| "step": 224 |
| }, |
| { |
| "epoch": 0.32, |
| "grad_norm": 0.5501871109008789, |
| "learning_rate": 0.00018417317246273952, |
| "loss": 1.2911, |
| "step": 225 |
| }, |
| { |
| "epoch": 0.32, |
| "grad_norm": 0.561215341091156, |
| "learning_rate": 0.00018410220014194466, |
| "loss": 1.2829, |
| "step": 226 |
| }, |
| { |
| "epoch": 0.32, |
| "grad_norm": 0.5201948285102844, |
| "learning_rate": 0.00018403122782114978, |
| "loss": 1.2713, |
| "step": 227 |
| }, |
| { |
| "epoch": 0.32, |
| "grad_norm": 0.5538742542266846, |
| "learning_rate": 0.00018396025550035486, |
| "loss": 1.2533, |
| "step": 228 |
| }, |
| { |
| "epoch": 0.32, |
| "grad_norm": 0.5249223113059998, |
| "learning_rate": 0.00018388928317955997, |
| "loss": 1.3244, |
| "step": 229 |
| }, |
| { |
| "epoch": 0.33, |
| "grad_norm": 0.5145981311798096, |
| "learning_rate": 0.0001838183108587651, |
| "loss": 1.2735, |
| "step": 230 |
| }, |
| { |
| "epoch": 0.33, |
| "grad_norm": 0.5346860885620117, |
| "learning_rate": 0.0001837473385379702, |
| "loss": 1.2238, |
| "step": 231 |
| }, |
| { |
| "epoch": 0.33, |
| "grad_norm": 0.5297681093215942, |
| "learning_rate": 0.0001836763662171753, |
| "loss": 1.3163, |
| "step": 232 |
| }, |
| { |
| "epoch": 0.33, |
| "grad_norm": 0.5075122117996216, |
| "learning_rate": 0.00018360539389638043, |
| "loss": 1.3422, |
| "step": 233 |
| }, |
| { |
| "epoch": 0.33, |
| "grad_norm": 0.5438109636306763, |
| "learning_rate": 0.00018353442157558554, |
| "loss": 1.3367, |
| "step": 234 |
| }, |
| { |
| "epoch": 0.33, |
| "grad_norm": 0.5246129035949707, |
| "learning_rate": 0.00018346344925479065, |
| "loss": 1.2883, |
| "step": 235 |
| }, |
| { |
| "epoch": 0.33, |
| "grad_norm": 0.5447835326194763, |
| "learning_rate": 0.00018339247693399574, |
| "loss": 1.2409, |
| "step": 236 |
| }, |
| { |
| "epoch": 0.34, |
| "grad_norm": 0.5204280018806458, |
| "learning_rate": 0.00018332150461320085, |
| "loss": 1.26, |
| "step": 237 |
| }, |
| { |
| "epoch": 0.34, |
| "grad_norm": 0.5784415602684021, |
| "learning_rate": 0.00018325053229240596, |
| "loss": 1.2953, |
| "step": 238 |
| }, |
| { |
| "epoch": 0.34, |
| "grad_norm": 0.5427743196487427, |
| "learning_rate": 0.00018317955997161108, |
| "loss": 1.273, |
| "step": 239 |
| }, |
| { |
| "epoch": 0.34, |
| "grad_norm": 0.6766988635063171, |
| "learning_rate": 0.0001831085876508162, |
| "loss": 1.2985, |
| "step": 240 |
| }, |
| { |
| "epoch": 0.34, |
| "grad_norm": 0.5197088122367859, |
| "learning_rate": 0.0001830376153300213, |
| "loss": 1.331, |
| "step": 241 |
| }, |
| { |
| "epoch": 0.34, |
| "grad_norm": 0.5441769957542419, |
| "learning_rate": 0.0001829666430092264, |
| "loss": 1.2786, |
| "step": 242 |
| }, |
| { |
| "epoch": 0.34, |
| "grad_norm": 0.5796704888343811, |
| "learning_rate": 0.00018289567068843153, |
| "loss": 1.3317, |
| "step": 243 |
| }, |
| { |
| "epoch": 0.35, |
| "grad_norm": 0.5018938183784485, |
| "learning_rate": 0.00018282469836763664, |
| "loss": 1.1423, |
| "step": 244 |
| }, |
| { |
| "epoch": 0.35, |
| "grad_norm": 0.5162649154663086, |
| "learning_rate": 0.00018275372604684172, |
| "loss": 1.2544, |
| "step": 245 |
| }, |
| { |
| "epoch": 0.35, |
| "grad_norm": 0.5415024161338806, |
| "learning_rate": 0.00018268275372604684, |
| "loss": 1.2748, |
| "step": 246 |
| }, |
| { |
| "epoch": 0.35, |
| "grad_norm": 0.5413600206375122, |
| "learning_rate": 0.00018261178140525198, |
| "loss": 1.3206, |
| "step": 247 |
| }, |
| { |
| "epoch": 0.35, |
| "grad_norm": 0.5510334968566895, |
| "learning_rate": 0.0001825408090844571, |
| "loss": 1.2939, |
| "step": 248 |
| }, |
| { |
| "epoch": 0.35, |
| "grad_norm": 0.5214188694953918, |
| "learning_rate": 0.00018246983676366218, |
| "loss": 1.3322, |
| "step": 249 |
| }, |
| { |
| "epoch": 0.35, |
| "grad_norm": 0.5412008166313171, |
| "learning_rate": 0.0001823988644428673, |
| "loss": 1.2781, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.36, |
| "grad_norm": 0.565697431564331, |
| "learning_rate": 0.0001823278921220724, |
| "loss": 1.3404, |
| "step": 251 |
| }, |
| { |
| "epoch": 0.36, |
| "grad_norm": 0.5353158712387085, |
| "learning_rate": 0.0001822569198012775, |
| "loss": 1.3575, |
| "step": 252 |
| }, |
| { |
| "epoch": 0.36, |
| "grad_norm": 0.5572609305381775, |
| "learning_rate": 0.0001821859474804826, |
| "loss": 1.2713, |
| "step": 253 |
| }, |
| { |
| "epoch": 0.36, |
| "grad_norm": 0.5632525682449341, |
| "learning_rate": 0.00018211497515968774, |
| "loss": 1.2556, |
| "step": 254 |
| }, |
| { |
| "epoch": 0.36, |
| "grad_norm": 0.5715951323509216, |
| "learning_rate": 0.00018204400283889285, |
| "loss": 1.3299, |
| "step": 255 |
| }, |
| { |
| "epoch": 0.36, |
| "grad_norm": 0.5702312588691711, |
| "learning_rate": 0.00018197303051809796, |
| "loss": 1.1876, |
| "step": 256 |
| }, |
| { |
| "epoch": 0.36, |
| "grad_norm": 0.5902386903762817, |
| "learning_rate": 0.00018190205819730305, |
| "loss": 1.3058, |
| "step": 257 |
| }, |
| { |
| "epoch": 0.37, |
| "grad_norm": 0.5512535572052002, |
| "learning_rate": 0.00018183108587650816, |
| "loss": 1.2739, |
| "step": 258 |
| }, |
| { |
| "epoch": 0.37, |
| "grad_norm": 0.5071843266487122, |
| "learning_rate": 0.00018176011355571328, |
| "loss": 1.2882, |
| "step": 259 |
| }, |
| { |
| "epoch": 0.37, |
| "grad_norm": 0.5268561840057373, |
| "learning_rate": 0.0001816891412349184, |
| "loss": 1.2187, |
| "step": 260 |
| }, |
| { |
| "epoch": 0.37, |
| "grad_norm": 0.5496917963027954, |
| "learning_rate": 0.0001816181689141235, |
| "loss": 1.3007, |
| "step": 261 |
| }, |
| { |
| "epoch": 0.37, |
| "grad_norm": 0.5308434367179871, |
| "learning_rate": 0.0001815471965933286, |
| "loss": 1.2778, |
| "step": 262 |
| }, |
| { |
| "epoch": 0.37, |
| "grad_norm": 0.5531402230262756, |
| "learning_rate": 0.00018147622427253373, |
| "loss": 1.2411, |
| "step": 263 |
| }, |
| { |
| "epoch": 0.37, |
| "grad_norm": 0.5323889255523682, |
| "learning_rate": 0.00018140525195173884, |
| "loss": 1.2655, |
| "step": 264 |
| }, |
| { |
| "epoch": 0.38, |
| "grad_norm": 0.50742506980896, |
| "learning_rate": 0.00018133427963094392, |
| "loss": 1.2429, |
| "step": 265 |
| }, |
| { |
| "epoch": 0.38, |
| "grad_norm": 0.5636353492736816, |
| "learning_rate": 0.00018126330731014904, |
| "loss": 1.2079, |
| "step": 266 |
| }, |
| { |
| "epoch": 0.38, |
| "grad_norm": 0.5328524708747864, |
| "learning_rate": 0.00018119233498935415, |
| "loss": 1.2601, |
| "step": 267 |
| }, |
| { |
| "epoch": 0.38, |
| "grad_norm": 0.5404384136199951, |
| "learning_rate": 0.0001811213626685593, |
| "loss": 1.2461, |
| "step": 268 |
| }, |
| { |
| "epoch": 0.38, |
| "grad_norm": 0.5366867184638977, |
| "learning_rate": 0.00018105039034776438, |
| "loss": 1.2189, |
| "step": 269 |
| }, |
| { |
| "epoch": 0.38, |
| "grad_norm": 0.548911988735199, |
| "learning_rate": 0.0001809794180269695, |
| "loss": 1.3085, |
| "step": 270 |
| }, |
| { |
| "epoch": 0.38, |
| "grad_norm": 0.48217761516571045, |
| "learning_rate": 0.0001809084457061746, |
| "loss": 1.1768, |
| "step": 271 |
| }, |
| { |
| "epoch": 0.39, |
| "grad_norm": 0.5497826933860779, |
| "learning_rate": 0.00018083747338537971, |
| "loss": 1.2676, |
| "step": 272 |
| }, |
| { |
| "epoch": 0.39, |
| "grad_norm": 0.5193854570388794, |
| "learning_rate": 0.00018076650106458483, |
| "loss": 1.2323, |
| "step": 273 |
| }, |
| { |
| "epoch": 0.39, |
| "grad_norm": 0.5157699584960938, |
| "learning_rate": 0.0001806955287437899, |
| "loss": 1.2549, |
| "step": 274 |
| }, |
| { |
| "epoch": 0.39, |
| "grad_norm": 0.5180855989456177, |
| "learning_rate": 0.00018062455642299505, |
| "loss": 1.2495, |
| "step": 275 |
| }, |
| { |
| "epoch": 0.39, |
| "grad_norm": 0.5382695198059082, |
| "learning_rate": 0.00018055358410220016, |
| "loss": 1.2594, |
| "step": 276 |
| }, |
| { |
| "epoch": 0.39, |
| "grad_norm": 0.525709331035614, |
| "learning_rate": 0.00018048261178140528, |
| "loss": 1.3069, |
| "step": 277 |
| }, |
| { |
| "epoch": 0.39, |
| "grad_norm": 0.5481734275817871, |
| "learning_rate": 0.00018041163946061036, |
| "loss": 1.2754, |
| "step": 278 |
| }, |
| { |
| "epoch": 0.4, |
| "grad_norm": 0.5247664451599121, |
| "learning_rate": 0.00018034066713981548, |
| "loss": 1.2433, |
| "step": 279 |
| }, |
| { |
| "epoch": 0.4, |
| "grad_norm": 0.5272576808929443, |
| "learning_rate": 0.0001802696948190206, |
| "loss": 1.3306, |
| "step": 280 |
| }, |
| { |
| "epoch": 0.4, |
| "grad_norm": 0.5923863649368286, |
| "learning_rate": 0.0001801987224982257, |
| "loss": 1.2251, |
| "step": 281 |
| }, |
| { |
| "epoch": 0.4, |
| "grad_norm": 0.5423222780227661, |
| "learning_rate": 0.00018012775017743081, |
| "loss": 1.3083, |
| "step": 282 |
| }, |
| { |
| "epoch": 0.4, |
| "grad_norm": 0.5334257483482361, |
| "learning_rate": 0.00018005677785663593, |
| "loss": 1.2794, |
| "step": 283 |
| }, |
| { |
| "epoch": 0.4, |
| "grad_norm": 0.5238288044929504, |
| "learning_rate": 0.00017998580553584104, |
| "loss": 1.2115, |
| "step": 284 |
| }, |
| { |
| "epoch": 0.4, |
| "grad_norm": 0.5420286059379578, |
| "learning_rate": 0.00017991483321504615, |
| "loss": 1.1612, |
| "step": 285 |
| }, |
| { |
| "epoch": 0.41, |
| "grad_norm": 0.511910080909729, |
| "learning_rate": 0.00017984386089425124, |
| "loss": 1.2526, |
| "step": 286 |
| }, |
| { |
| "epoch": 0.41, |
| "grad_norm": 0.5181018710136414, |
| "learning_rate": 0.00017977288857345635, |
| "loss": 1.3193, |
| "step": 287 |
| }, |
| { |
| "epoch": 0.41, |
| "grad_norm": 0.5733616352081299, |
| "learning_rate": 0.00017970191625266146, |
| "loss": 1.3032, |
| "step": 288 |
| }, |
| { |
| "epoch": 0.41, |
| "grad_norm": 0.5532722473144531, |
| "learning_rate": 0.0001796309439318666, |
| "loss": 1.2011, |
| "step": 289 |
| }, |
| { |
| "epoch": 0.41, |
| "grad_norm": 0.5308284163475037, |
| "learning_rate": 0.0001795599716110717, |
| "loss": 1.3084, |
| "step": 290 |
| }, |
| { |
| "epoch": 0.41, |
| "grad_norm": 0.5570117831230164, |
| "learning_rate": 0.0001794889992902768, |
| "loss": 1.2663, |
| "step": 291 |
| }, |
| { |
| "epoch": 0.41, |
| "grad_norm": 0.5518406629562378, |
| "learning_rate": 0.00017941802696948191, |
| "loss": 1.2782, |
| "step": 292 |
| }, |
| { |
| "epoch": 0.42, |
| "grad_norm": 0.5204089283943176, |
| "learning_rate": 0.00017934705464868703, |
| "loss": 1.2532, |
| "step": 293 |
| }, |
| { |
| "epoch": 0.42, |
| "grad_norm": 0.5458951592445374, |
| "learning_rate": 0.0001792760823278921, |
| "loss": 1.2507, |
| "step": 294 |
| }, |
| { |
| "epoch": 0.42, |
| "grad_norm": 0.6208042502403259, |
| "learning_rate": 0.00017920511000709723, |
| "loss": 1.2569, |
| "step": 295 |
| }, |
| { |
| "epoch": 0.42, |
| "grad_norm": 0.5473780632019043, |
| "learning_rate": 0.00017913413768630236, |
| "loss": 1.3277, |
| "step": 296 |
| }, |
| { |
| "epoch": 0.42, |
| "grad_norm": 0.5519291162490845, |
| "learning_rate": 0.00017906316536550748, |
| "loss": 1.2588, |
| "step": 297 |
| }, |
| { |
| "epoch": 0.42, |
| "grad_norm": 0.5681371092796326, |
| "learning_rate": 0.00017899219304471256, |
| "loss": 1.2454, |
| "step": 298 |
| }, |
| { |
| "epoch": 0.42, |
| "grad_norm": 0.5682107210159302, |
| "learning_rate": 0.00017892122072391768, |
| "loss": 1.2342, |
| "step": 299 |
| }, |
| { |
| "epoch": 0.43, |
| "grad_norm": 0.5326645374298096, |
| "learning_rate": 0.0001788502484031228, |
| "loss": 1.2052, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.43, |
| "grad_norm": 0.6371567249298096, |
| "learning_rate": 0.0001787792760823279, |
| "loss": 1.1617, |
| "step": 301 |
| }, |
| { |
| "epoch": 0.43, |
| "grad_norm": 0.6094531416893005, |
| "learning_rate": 0.00017870830376153301, |
| "loss": 1.218, |
| "step": 302 |
| }, |
| { |
| "epoch": 0.43, |
| "grad_norm": 0.5426862835884094, |
| "learning_rate": 0.00017863733144073813, |
| "loss": 1.313, |
| "step": 303 |
| }, |
| { |
| "epoch": 0.43, |
| "grad_norm": 0.5334256291389465, |
| "learning_rate": 0.00017856635911994324, |
| "loss": 1.2978, |
| "step": 304 |
| }, |
| { |
| "epoch": 0.43, |
| "grad_norm": 0.5406014919281006, |
| "learning_rate": 0.00017849538679914835, |
| "loss": 1.3094, |
| "step": 305 |
| }, |
| { |
| "epoch": 0.43, |
| "grad_norm": 0.5107502341270447, |
| "learning_rate": 0.00017842441447835347, |
| "loss": 1.1634, |
| "step": 306 |
| }, |
| { |
| "epoch": 0.43, |
| "grad_norm": 0.5104424357414246, |
| "learning_rate": 0.00017835344215755855, |
| "loss": 1.2576, |
| "step": 307 |
| }, |
| { |
| "epoch": 0.44, |
| "grad_norm": 0.5558850169181824, |
| "learning_rate": 0.00017828246983676366, |
| "loss": 1.2696, |
| "step": 308 |
| }, |
| { |
| "epoch": 0.44, |
| "grad_norm": 0.5559511184692383, |
| "learning_rate": 0.00017821149751596878, |
| "loss": 1.292, |
| "step": 309 |
| }, |
| { |
| "epoch": 0.44, |
| "grad_norm": 0.5495506525039673, |
| "learning_rate": 0.0001781405251951739, |
| "loss": 1.2363, |
| "step": 310 |
| }, |
| { |
| "epoch": 0.44, |
| "grad_norm": 0.5591257214546204, |
| "learning_rate": 0.000178069552874379, |
| "loss": 1.2425, |
| "step": 311 |
| }, |
| { |
| "epoch": 0.44, |
| "grad_norm": 0.5381730198860168, |
| "learning_rate": 0.00017799858055358411, |
| "loss": 1.1903, |
| "step": 312 |
| }, |
| { |
| "epoch": 0.44, |
| "grad_norm": 0.5501212477684021, |
| "learning_rate": 0.00017792760823278923, |
| "loss": 1.2598, |
| "step": 313 |
| }, |
| { |
| "epoch": 0.44, |
| "grad_norm": 0.6151654124259949, |
| "learning_rate": 0.00017785663591199434, |
| "loss": 1.1703, |
| "step": 314 |
| }, |
| { |
| "epoch": 0.45, |
| "grad_norm": 0.5302771925926208, |
| "learning_rate": 0.00017778566359119943, |
| "loss": 1.268, |
| "step": 315 |
| }, |
| { |
| "epoch": 0.45, |
| "grad_norm": 0.5328003764152527, |
| "learning_rate": 0.00017771469127040454, |
| "loss": 1.259, |
| "step": 316 |
| }, |
| { |
| "epoch": 0.45, |
| "grad_norm": 0.5223655700683594, |
| "learning_rate": 0.00017764371894960965, |
| "loss": 1.2631, |
| "step": 317 |
| }, |
| { |
| "epoch": 0.45, |
| "grad_norm": 0.5435971021652222, |
| "learning_rate": 0.0001775727466288148, |
| "loss": 1.2645, |
| "step": 318 |
| }, |
| { |
| "epoch": 0.45, |
| "grad_norm": 0.5286442041397095, |
| "learning_rate": 0.00017750177430801988, |
| "loss": 1.2613, |
| "step": 319 |
| }, |
| { |
| "epoch": 0.45, |
| "grad_norm": 0.5257332921028137, |
| "learning_rate": 0.000177430801987225, |
| "loss": 1.2435, |
| "step": 320 |
| }, |
| { |
| "epoch": 0.45, |
| "grad_norm": 0.5707381963729858, |
| "learning_rate": 0.0001773598296664301, |
| "loss": 1.2085, |
| "step": 321 |
| }, |
| { |
| "epoch": 0.46, |
| "grad_norm": 0.5447467565536499, |
| "learning_rate": 0.00017728885734563521, |
| "loss": 1.2266, |
| "step": 322 |
| }, |
| { |
| "epoch": 0.46, |
| "grad_norm": 0.5707226991653442, |
| "learning_rate": 0.0001772178850248403, |
| "loss": 1.2284, |
| "step": 323 |
| }, |
| { |
| "epoch": 0.46, |
| "grad_norm": 0.5743198394775391, |
| "learning_rate": 0.0001771469127040454, |
| "loss": 1.1793, |
| "step": 324 |
| }, |
| { |
| "epoch": 0.46, |
| "grad_norm": 0.5705985426902771, |
| "learning_rate": 0.00017707594038325055, |
| "loss": 1.2348, |
| "step": 325 |
| }, |
| { |
| "epoch": 0.46, |
| "grad_norm": 0.5277898907661438, |
| "learning_rate": 0.00017700496806245567, |
| "loss": 1.1603, |
| "step": 326 |
| }, |
| { |
| "epoch": 0.46, |
| "grad_norm": 0.5345696806907654, |
| "learning_rate": 0.00017693399574166075, |
| "loss": 1.2799, |
| "step": 327 |
| }, |
| { |
| "epoch": 0.46, |
| "grad_norm": 0.5616069436073303, |
| "learning_rate": 0.00017686302342086586, |
| "loss": 1.2574, |
| "step": 328 |
| }, |
| { |
| "epoch": 0.47, |
| "grad_norm": 0.5110874772071838, |
| "learning_rate": 0.00017679205110007098, |
| "loss": 1.2796, |
| "step": 329 |
| }, |
| { |
| "epoch": 0.47, |
| "grad_norm": 0.5312853455543518, |
| "learning_rate": 0.0001767210787792761, |
| "loss": 1.294, |
| "step": 330 |
| }, |
| { |
| "epoch": 0.47, |
| "grad_norm": 0.576275110244751, |
| "learning_rate": 0.0001766501064584812, |
| "loss": 1.2761, |
| "step": 331 |
| }, |
| { |
| "epoch": 0.47, |
| "grad_norm": 0.5550281405448914, |
| "learning_rate": 0.00017657913413768631, |
| "loss": 1.2489, |
| "step": 332 |
| }, |
| { |
| "epoch": 0.47, |
| "grad_norm": 0.55446457862854, |
| "learning_rate": 0.00017650816181689143, |
| "loss": 1.3195, |
| "step": 333 |
| }, |
| { |
| "epoch": 0.47, |
| "grad_norm": 0.5434035658836365, |
| "learning_rate": 0.00017643718949609654, |
| "loss": 1.2927, |
| "step": 334 |
| }, |
| { |
| "epoch": 0.47, |
| "grad_norm": 0.536832869052887, |
| "learning_rate": 0.00017636621717530165, |
| "loss": 1.2448, |
| "step": 335 |
| }, |
| { |
| "epoch": 0.48, |
| "grad_norm": 0.5537269711494446, |
| "learning_rate": 0.00017629524485450674, |
| "loss": 1.2461, |
| "step": 336 |
| }, |
| { |
| "epoch": 0.48, |
| "grad_norm": 0.5346850156784058, |
| "learning_rate": 0.00017622427253371185, |
| "loss": 1.2614, |
| "step": 337 |
| }, |
| { |
| "epoch": 0.48, |
| "grad_norm": 0.5255287289619446, |
| "learning_rate": 0.00017615330021291696, |
| "loss": 1.2376, |
| "step": 338 |
| }, |
| { |
| "epoch": 0.48, |
| "grad_norm": 0.5487991571426392, |
| "learning_rate": 0.00017608232789212208, |
| "loss": 1.2523, |
| "step": 339 |
| }, |
| { |
| "epoch": 0.48, |
| "grad_norm": 0.554786205291748, |
| "learning_rate": 0.0001760113555713272, |
| "loss": 1.2265, |
| "step": 340 |
| }, |
| { |
| "epoch": 0.48, |
| "grad_norm": 0.8868914246559143, |
| "learning_rate": 0.0001759403832505323, |
| "loss": 1.2401, |
| "step": 341 |
| }, |
| { |
| "epoch": 0.48, |
| "grad_norm": 0.5412053465843201, |
| "learning_rate": 0.00017586941092973741, |
| "loss": 1.2027, |
| "step": 342 |
| }, |
| { |
| "epoch": 0.49, |
| "grad_norm": 0.5590038895606995, |
| "learning_rate": 0.00017579843860894253, |
| "loss": 1.1883, |
| "step": 343 |
| }, |
| { |
| "epoch": 0.49, |
| "grad_norm": 0.5496480464935303, |
| "learning_rate": 0.0001757274662881476, |
| "loss": 1.2329, |
| "step": 344 |
| }, |
| { |
| "epoch": 0.49, |
| "grad_norm": 0.5843716263771057, |
| "learning_rate": 0.00017565649396735273, |
| "loss": 1.249, |
| "step": 345 |
| }, |
| { |
| "epoch": 0.49, |
| "grad_norm": 0.6055741906166077, |
| "learning_rate": 0.00017558552164655787, |
| "loss": 1.1477, |
| "step": 346 |
| }, |
| { |
| "epoch": 0.49, |
| "grad_norm": 0.5583242177963257, |
| "learning_rate": 0.00017551454932576298, |
| "loss": 1.2862, |
| "step": 347 |
| }, |
| { |
| "epoch": 0.49, |
| "grad_norm": 2.0163357257843018, |
| "learning_rate": 0.00017544357700496806, |
| "loss": 1.2014, |
| "step": 348 |
| }, |
| { |
| "epoch": 0.49, |
| "grad_norm": 0.6089881658554077, |
| "learning_rate": 0.00017537260468417318, |
| "loss": 1.3267, |
| "step": 349 |
| }, |
| { |
| "epoch": 0.5, |
| "grad_norm": 0.5367897748947144, |
| "learning_rate": 0.0001753016323633783, |
| "loss": 1.2379, |
| "step": 350 |
| }, |
| { |
| "epoch": 0.5, |
| "grad_norm": 0.5732371807098389, |
| "learning_rate": 0.0001752306600425834, |
| "loss": 1.2557, |
| "step": 351 |
| }, |
| { |
| "epoch": 0.5, |
| "grad_norm": 0.6220402717590332, |
| "learning_rate": 0.0001751596877217885, |
| "loss": 1.2415, |
| "step": 352 |
| }, |
| { |
| "epoch": 0.5, |
| "grad_norm": 0.5300910472869873, |
| "learning_rate": 0.00017508871540099363, |
| "loss": 1.2322, |
| "step": 353 |
| }, |
| { |
| "epoch": 0.5, |
| "grad_norm": 0.5924064517021179, |
| "learning_rate": 0.00017501774308019874, |
| "loss": 1.2216, |
| "step": 354 |
| }, |
| { |
| "epoch": 0.5, |
| "grad_norm": 0.5625554323196411, |
| "learning_rate": 0.00017494677075940385, |
| "loss": 1.1887, |
| "step": 355 |
| }, |
| { |
| "epoch": 0.5, |
| "grad_norm": 0.5225299000740051, |
| "learning_rate": 0.00017487579843860894, |
| "loss": 1.2274, |
| "step": 356 |
| }, |
| { |
| "epoch": 0.51, |
| "grad_norm": 0.5613044500350952, |
| "learning_rate": 0.00017480482611781405, |
| "loss": 1.2268, |
| "step": 357 |
| }, |
| { |
| "epoch": 0.51, |
| "grad_norm": 0.5904155373573303, |
| "learning_rate": 0.00017473385379701916, |
| "loss": 1.266, |
| "step": 358 |
| }, |
| { |
| "epoch": 0.51, |
| "grad_norm": 0.5112659335136414, |
| "learning_rate": 0.00017466288147622428, |
| "loss": 1.2891, |
| "step": 359 |
| }, |
| { |
| "epoch": 0.51, |
| "grad_norm": 0.5344319343566895, |
| "learning_rate": 0.0001745919091554294, |
| "loss": 1.2561, |
| "step": 360 |
| }, |
| { |
| "epoch": 0.51, |
| "grad_norm": 0.5556378364562988, |
| "learning_rate": 0.0001745209368346345, |
| "loss": 1.263, |
| "step": 361 |
| }, |
| { |
| "epoch": 0.51, |
| "grad_norm": 0.5675813555717468, |
| "learning_rate": 0.00017444996451383962, |
| "loss": 1.3235, |
| "step": 362 |
| }, |
| { |
| "epoch": 0.51, |
| "grad_norm": 0.5817593932151794, |
| "learning_rate": 0.00017437899219304473, |
| "loss": 1.3471, |
| "step": 363 |
| }, |
| { |
| "epoch": 0.52, |
| "grad_norm": 0.5670614242553711, |
| "learning_rate": 0.00017430801987224984, |
| "loss": 1.3019, |
| "step": 364 |
| }, |
| { |
| "epoch": 0.52, |
| "grad_norm": 0.5784518122673035, |
| "learning_rate": 0.00017423704755145493, |
| "loss": 1.2122, |
| "step": 365 |
| }, |
| { |
| "epoch": 0.52, |
| "grad_norm": 0.5680185556411743, |
| "learning_rate": 0.00017416607523066004, |
| "loss": 1.2042, |
| "step": 366 |
| }, |
| { |
| "epoch": 0.52, |
| "grad_norm": 0.5474495887756348, |
| "learning_rate": 0.00017409510290986518, |
| "loss": 1.1273, |
| "step": 367 |
| }, |
| { |
| "epoch": 0.52, |
| "grad_norm": 0.5586380362510681, |
| "learning_rate": 0.00017402413058907026, |
| "loss": 1.2425, |
| "step": 368 |
| }, |
| { |
| "epoch": 0.52, |
| "grad_norm": 0.5608516335487366, |
| "learning_rate": 0.00017395315826827538, |
| "loss": 1.228, |
| "step": 369 |
| }, |
| { |
| "epoch": 0.52, |
| "grad_norm": 0.5339083075523376, |
| "learning_rate": 0.0001738821859474805, |
| "loss": 1.2171, |
| "step": 370 |
| }, |
| { |
| "epoch": 0.53, |
| "grad_norm": 0.5438979268074036, |
| "learning_rate": 0.0001738112136266856, |
| "loss": 1.2474, |
| "step": 371 |
| }, |
| { |
| "epoch": 0.53, |
| "grad_norm": 0.534762978553772, |
| "learning_rate": 0.00017374024130589072, |
| "loss": 1.2844, |
| "step": 372 |
| }, |
| { |
| "epoch": 0.53, |
| "grad_norm": 0.5463000535964966, |
| "learning_rate": 0.0001736692689850958, |
| "loss": 1.2389, |
| "step": 373 |
| }, |
| { |
| "epoch": 0.53, |
| "grad_norm": 0.5256685018539429, |
| "learning_rate": 0.00017359829666430094, |
| "loss": 1.2425, |
| "step": 374 |
| }, |
| { |
| "epoch": 0.53, |
| "grad_norm": 0.509695827960968, |
| "learning_rate": 0.00017352732434350605, |
| "loss": 1.2685, |
| "step": 375 |
| }, |
| { |
| "epoch": 0.53, |
| "grad_norm": 0.532911479473114, |
| "learning_rate": 0.00017345635202271117, |
| "loss": 1.237, |
| "step": 376 |
| }, |
| { |
| "epoch": 0.53, |
| "grad_norm": 0.5342946648597717, |
| "learning_rate": 0.00017338537970191625, |
| "loss": 1.2334, |
| "step": 377 |
| }, |
| { |
| "epoch": 0.54, |
| "grad_norm": 0.5660805702209473, |
| "learning_rate": 0.00017331440738112136, |
| "loss": 1.1716, |
| "step": 378 |
| }, |
| { |
| "epoch": 0.54, |
| "grad_norm": 0.5654976963996887, |
| "learning_rate": 0.00017324343506032648, |
| "loss": 1.2343, |
| "step": 379 |
| }, |
| { |
| "epoch": 0.54, |
| "grad_norm": 0.5845637917518616, |
| "learning_rate": 0.0001731724627395316, |
| "loss": 1.2037, |
| "step": 380 |
| }, |
| { |
| "epoch": 0.54, |
| "grad_norm": 0.570573091506958, |
| "learning_rate": 0.0001731014904187367, |
| "loss": 1.2255, |
| "step": 381 |
| }, |
| { |
| "epoch": 0.54, |
| "grad_norm": 0.567817747592926, |
| "learning_rate": 0.00017303051809794182, |
| "loss": 1.2238, |
| "step": 382 |
| }, |
| { |
| "epoch": 0.54, |
| "grad_norm": 0.5502271056175232, |
| "learning_rate": 0.00017295954577714693, |
| "loss": 1.2237, |
| "step": 383 |
| }, |
| { |
| "epoch": 0.54, |
| "grad_norm": 0.5540646314620972, |
| "learning_rate": 0.00017288857345635204, |
| "loss": 1.2192, |
| "step": 384 |
| }, |
| { |
| "epoch": 0.55, |
| "grad_norm": 0.5388753414154053, |
| "learning_rate": 0.00017281760113555713, |
| "loss": 1.2181, |
| "step": 385 |
| }, |
| { |
| "epoch": 0.55, |
| "grad_norm": 0.5435704588890076, |
| "learning_rate": 0.00017274662881476224, |
| "loss": 1.1588, |
| "step": 386 |
| }, |
| { |
| "epoch": 0.55, |
| "grad_norm": 0.5815427899360657, |
| "learning_rate": 0.00017267565649396735, |
| "loss": 1.1784, |
| "step": 387 |
| }, |
| { |
| "epoch": 0.55, |
| "grad_norm": 0.5804842710494995, |
| "learning_rate": 0.0001726046841731725, |
| "loss": 1.274, |
| "step": 388 |
| }, |
| { |
| "epoch": 0.55, |
| "grad_norm": 0.5361278057098389, |
| "learning_rate": 0.00017253371185237758, |
| "loss": 1.1567, |
| "step": 389 |
| }, |
| { |
| "epoch": 0.55, |
| "grad_norm": 0.5804944038391113, |
| "learning_rate": 0.0001724627395315827, |
| "loss": 1.2138, |
| "step": 390 |
| }, |
| { |
| "epoch": 0.55, |
| "grad_norm": 0.5311663150787354, |
| "learning_rate": 0.0001723917672107878, |
| "loss": 1.2318, |
| "step": 391 |
| }, |
| { |
| "epoch": 0.56, |
| "grad_norm": 0.5714538097381592, |
| "learning_rate": 0.00017232079488999292, |
| "loss": 1.1928, |
| "step": 392 |
| }, |
| { |
| "epoch": 0.56, |
| "grad_norm": 0.5973841547966003, |
| "learning_rate": 0.00017224982256919803, |
| "loss": 1.208, |
| "step": 393 |
| }, |
| { |
| "epoch": 0.56, |
| "grad_norm": 0.644917368888855, |
| "learning_rate": 0.00017217885024840311, |
| "loss": 1.2642, |
| "step": 394 |
| }, |
| { |
| "epoch": 0.56, |
| "grad_norm": 0.586104691028595, |
| "learning_rate": 0.00017210787792760825, |
| "loss": 1.2574, |
| "step": 395 |
| }, |
| { |
| "epoch": 0.56, |
| "grad_norm": 0.583574652671814, |
| "learning_rate": 0.00017203690560681337, |
| "loss": 1.327, |
| "step": 396 |
| }, |
| { |
| "epoch": 0.56, |
| "grad_norm": 0.5617191195487976, |
| "learning_rate": 0.00017196593328601848, |
| "loss": 1.1605, |
| "step": 397 |
| }, |
| { |
| "epoch": 0.56, |
| "grad_norm": 0.5635659098625183, |
| "learning_rate": 0.00017189496096522356, |
| "loss": 1.2222, |
| "step": 398 |
| }, |
| { |
| "epoch": 0.57, |
| "grad_norm": 0.6253629326820374, |
| "learning_rate": 0.00017182398864442868, |
| "loss": 1.2162, |
| "step": 399 |
| }, |
| { |
| "epoch": 0.57, |
| "grad_norm": 0.6384497880935669, |
| "learning_rate": 0.0001717530163236338, |
| "loss": 1.2878, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.57, |
| "grad_norm": 0.5805338621139526, |
| "learning_rate": 0.0001716820440028389, |
| "loss": 1.2664, |
| "step": 401 |
| }, |
| { |
| "epoch": 0.57, |
| "grad_norm": 0.5705095529556274, |
| "learning_rate": 0.000171611071682044, |
| "loss": 1.2205, |
| "step": 402 |
| }, |
| { |
| "epoch": 0.57, |
| "grad_norm": 0.5843649506568909, |
| "learning_rate": 0.00017154009936124913, |
| "loss": 1.2003, |
| "step": 403 |
| }, |
| { |
| "epoch": 0.57, |
| "grad_norm": 0.5798406600952148, |
| "learning_rate": 0.00017146912704045424, |
| "loss": 1.228, |
| "step": 404 |
| }, |
| { |
| "epoch": 0.57, |
| "grad_norm": 0.60892254114151, |
| "learning_rate": 0.00017139815471965935, |
| "loss": 1.1853, |
| "step": 405 |
| }, |
| { |
| "epoch": 0.58, |
| "grad_norm": 0.592847466468811, |
| "learning_rate": 0.00017132718239886444, |
| "loss": 1.2062, |
| "step": 406 |
| }, |
| { |
| "epoch": 0.58, |
| "grad_norm": 0.5821055769920349, |
| "learning_rate": 0.00017125621007806955, |
| "loss": 1.1844, |
| "step": 407 |
| }, |
| { |
| "epoch": 0.58, |
| "grad_norm": 0.5680034160614014, |
| "learning_rate": 0.00017118523775727466, |
| "loss": 1.235, |
| "step": 408 |
| }, |
| { |
| "epoch": 0.58, |
| "grad_norm": 0.5684024095535278, |
| "learning_rate": 0.00017111426543647978, |
| "loss": 1.209, |
| "step": 409 |
| }, |
| { |
| "epoch": 0.58, |
| "grad_norm": 0.5451253056526184, |
| "learning_rate": 0.0001710432931156849, |
| "loss": 1.2026, |
| "step": 410 |
| }, |
| { |
| "epoch": 0.58, |
| "grad_norm": 0.5240196585655212, |
| "learning_rate": 0.00017097232079489, |
| "loss": 1.1733, |
| "step": 411 |
| }, |
| { |
| "epoch": 0.58, |
| "grad_norm": 0.5445153713226318, |
| "learning_rate": 0.00017090134847409512, |
| "loss": 1.2643, |
| "step": 412 |
| }, |
| { |
| "epoch": 0.59, |
| "grad_norm": 0.5214688181877136, |
| "learning_rate": 0.00017083037615330023, |
| "loss": 1.2381, |
| "step": 413 |
| }, |
| { |
| "epoch": 0.59, |
| "grad_norm": 0.5414560437202454, |
| "learning_rate": 0.00017075940383250531, |
| "loss": 1.2705, |
| "step": 414 |
| }, |
| { |
| "epoch": 0.59, |
| "grad_norm": 0.567355215549469, |
| "learning_rate": 0.00017068843151171043, |
| "loss": 1.2956, |
| "step": 415 |
| }, |
| { |
| "epoch": 0.59, |
| "grad_norm": 0.5429472327232361, |
| "learning_rate": 0.00017061745919091554, |
| "loss": 1.2057, |
| "step": 416 |
| }, |
| { |
| "epoch": 0.59, |
| "grad_norm": 0.5171405673027039, |
| "learning_rate": 0.00017054648687012068, |
| "loss": 1.1803, |
| "step": 417 |
| }, |
| { |
| "epoch": 0.59, |
| "grad_norm": 0.5672470331192017, |
| "learning_rate": 0.00017047551454932577, |
| "loss": 1.2566, |
| "step": 418 |
| }, |
| { |
| "epoch": 0.59, |
| "grad_norm": 0.5415303707122803, |
| "learning_rate": 0.00017040454222853088, |
| "loss": 1.2364, |
| "step": 419 |
| }, |
| { |
| "epoch": 0.6, |
| "grad_norm": 0.546066164970398, |
| "learning_rate": 0.000170333569907736, |
| "loss": 1.2458, |
| "step": 420 |
| }, |
| { |
| "epoch": 0.6, |
| "grad_norm": 0.532200813293457, |
| "learning_rate": 0.0001702625975869411, |
| "loss": 1.227, |
| "step": 421 |
| }, |
| { |
| "epoch": 0.6, |
| "grad_norm": 0.5418232679367065, |
| "learning_rate": 0.00017019162526614622, |
| "loss": 1.1896, |
| "step": 422 |
| }, |
| { |
| "epoch": 0.6, |
| "grad_norm": 0.5557692646980286, |
| "learning_rate": 0.0001701206529453513, |
| "loss": 1.213, |
| "step": 423 |
| }, |
| { |
| "epoch": 0.6, |
| "grad_norm": 0.5730442404747009, |
| "learning_rate": 0.00017004968062455644, |
| "loss": 1.1846, |
| "step": 424 |
| }, |
| { |
| "epoch": 0.6, |
| "grad_norm": 0.5887081027030945, |
| "learning_rate": 0.00016997870830376155, |
| "loss": 1.2557, |
| "step": 425 |
| }, |
| { |
| "epoch": 0.6, |
| "grad_norm": 0.5441513061523438, |
| "learning_rate": 0.00016990773598296667, |
| "loss": 1.2167, |
| "step": 426 |
| }, |
| { |
| "epoch": 0.6, |
| "grad_norm": 0.5610032081604004, |
| "learning_rate": 0.00016983676366217175, |
| "loss": 1.2331, |
| "step": 427 |
| }, |
| { |
| "epoch": 0.61, |
| "grad_norm": 0.5748931765556335, |
| "learning_rate": 0.00016976579134137687, |
| "loss": 1.1176, |
| "step": 428 |
| }, |
| { |
| "epoch": 0.61, |
| "grad_norm": 0.5597652196884155, |
| "learning_rate": 0.00016969481902058198, |
| "loss": 1.2147, |
| "step": 429 |
| }, |
| { |
| "epoch": 0.61, |
| "grad_norm": 0.5750680565834045, |
| "learning_rate": 0.0001696238466997871, |
| "loss": 1.2943, |
| "step": 430 |
| }, |
| { |
| "epoch": 0.61, |
| "grad_norm": 0.5501390695571899, |
| "learning_rate": 0.0001695528743789922, |
| "loss": 1.242, |
| "step": 431 |
| }, |
| { |
| "epoch": 0.61, |
| "grad_norm": 0.5521880984306335, |
| "learning_rate": 0.00016948190205819732, |
| "loss": 1.1144, |
| "step": 432 |
| }, |
| { |
| "epoch": 0.61, |
| "grad_norm": 0.563677191734314, |
| "learning_rate": 0.00016941092973740243, |
| "loss": 1.2875, |
| "step": 433 |
| }, |
| { |
| "epoch": 0.61, |
| "grad_norm": 0.5883916616439819, |
| "learning_rate": 0.00016933995741660754, |
| "loss": 1.2823, |
| "step": 434 |
| }, |
| { |
| "epoch": 0.62, |
| "grad_norm": 0.5612311363220215, |
| "learning_rate": 0.00016926898509581263, |
| "loss": 1.1956, |
| "step": 435 |
| }, |
| { |
| "epoch": 0.62, |
| "grad_norm": 0.5482412576675415, |
| "learning_rate": 0.00016919801277501774, |
| "loss": 1.2513, |
| "step": 436 |
| }, |
| { |
| "epoch": 0.62, |
| "grad_norm": 0.519913911819458, |
| "learning_rate": 0.00016912704045422285, |
| "loss": 1.2337, |
| "step": 437 |
| }, |
| { |
| "epoch": 0.62, |
| "grad_norm": 0.5395796895027161, |
| "learning_rate": 0.000169056068133428, |
| "loss": 1.2148, |
| "step": 438 |
| }, |
| { |
| "epoch": 0.62, |
| "grad_norm": 0.5586941242218018, |
| "learning_rate": 0.00016898509581263308, |
| "loss": 1.1486, |
| "step": 439 |
| }, |
| { |
| "epoch": 0.62, |
| "grad_norm": 0.5568752884864807, |
| "learning_rate": 0.0001689141234918382, |
| "loss": 1.2541, |
| "step": 440 |
| }, |
| { |
| "epoch": 0.62, |
| "grad_norm": 0.5618448853492737, |
| "learning_rate": 0.0001688431511710433, |
| "loss": 1.2332, |
| "step": 441 |
| }, |
| { |
| "epoch": 0.63, |
| "grad_norm": 0.5413042902946472, |
| "learning_rate": 0.00016877217885024842, |
| "loss": 1.2234, |
| "step": 442 |
| }, |
| { |
| "epoch": 0.63, |
| "grad_norm": 0.5402277112007141, |
| "learning_rate": 0.0001687012065294535, |
| "loss": 1.2421, |
| "step": 443 |
| }, |
| { |
| "epoch": 0.63, |
| "grad_norm": 0.5068058967590332, |
| "learning_rate": 0.00016863023420865861, |
| "loss": 1.2887, |
| "step": 444 |
| }, |
| { |
| "epoch": 0.63, |
| "grad_norm": 0.5408315658569336, |
| "learning_rate": 0.00016855926188786375, |
| "loss": 1.2649, |
| "step": 445 |
| }, |
| { |
| "epoch": 0.63, |
| "grad_norm": 0.5410880446434021, |
| "learning_rate": 0.00016848828956706887, |
| "loss": 1.2779, |
| "step": 446 |
| }, |
| { |
| "epoch": 0.63, |
| "grad_norm": 0.5637301206588745, |
| "learning_rate": 0.00016841731724627395, |
| "loss": 1.1927, |
| "step": 447 |
| }, |
| { |
| "epoch": 0.63, |
| "grad_norm": 0.5596689581871033, |
| "learning_rate": 0.00016834634492547907, |
| "loss": 1.246, |
| "step": 448 |
| }, |
| { |
| "epoch": 0.64, |
| "grad_norm": 0.5656019449234009, |
| "learning_rate": 0.00016827537260468418, |
| "loss": 1.2089, |
| "step": 449 |
| }, |
| { |
| "epoch": 0.64, |
| "grad_norm": 0.5710034966468811, |
| "learning_rate": 0.0001682044002838893, |
| "loss": 1.2327, |
| "step": 450 |
| }, |
| { |
| "epoch": 0.64, |
| "grad_norm": 0.5633525848388672, |
| "learning_rate": 0.0001681334279630944, |
| "loss": 1.1619, |
| "step": 451 |
| }, |
| { |
| "epoch": 0.64, |
| "grad_norm": 0.5352773070335388, |
| "learning_rate": 0.00016806245564229952, |
| "loss": 1.2318, |
| "step": 452 |
| }, |
| { |
| "epoch": 0.64, |
| "grad_norm": 0.5729439854621887, |
| "learning_rate": 0.00016799148332150463, |
| "loss": 1.208, |
| "step": 453 |
| }, |
| { |
| "epoch": 0.64, |
| "grad_norm": 0.5703145861625671, |
| "learning_rate": 0.00016792051100070974, |
| "loss": 1.2752, |
| "step": 454 |
| }, |
| { |
| "epoch": 0.64, |
| "grad_norm": 0.5441925525665283, |
| "learning_rate": 0.00016784953867991485, |
| "loss": 1.2427, |
| "step": 455 |
| }, |
| { |
| "epoch": 0.65, |
| "grad_norm": 0.5500312447547913, |
| "learning_rate": 0.00016777856635911994, |
| "loss": 1.1833, |
| "step": 456 |
| }, |
| { |
| "epoch": 0.65, |
| "grad_norm": 0.538560688495636, |
| "learning_rate": 0.00016770759403832505, |
| "loss": 1.1809, |
| "step": 457 |
| }, |
| { |
| "epoch": 0.65, |
| "grad_norm": 0.5843518972396851, |
| "learning_rate": 0.00016763662171753017, |
| "loss": 1.2804, |
| "step": 458 |
| }, |
| { |
| "epoch": 0.65, |
| "grad_norm": 0.5413994789123535, |
| "learning_rate": 0.00016756564939673528, |
| "loss": 1.2109, |
| "step": 459 |
| }, |
| { |
| "epoch": 0.65, |
| "grad_norm": 0.5343271493911743, |
| "learning_rate": 0.0001674946770759404, |
| "loss": 1.2236, |
| "step": 460 |
| }, |
| { |
| "epoch": 0.65, |
| "grad_norm": 0.5626118183135986, |
| "learning_rate": 0.0001674237047551455, |
| "loss": 1.1744, |
| "step": 461 |
| }, |
| { |
| "epoch": 0.65, |
| "grad_norm": 0.5811516642570496, |
| "learning_rate": 0.00016735273243435062, |
| "loss": 1.2039, |
| "step": 462 |
| }, |
| { |
| "epoch": 0.66, |
| "grad_norm": 0.5644727945327759, |
| "learning_rate": 0.00016728176011355573, |
| "loss": 1.2103, |
| "step": 463 |
| }, |
| { |
| "epoch": 0.66, |
| "grad_norm": 0.560869574546814, |
| "learning_rate": 0.00016721078779276081, |
| "loss": 1.1694, |
| "step": 464 |
| }, |
| { |
| "epoch": 0.66, |
| "grad_norm": 0.5319089293479919, |
| "learning_rate": 0.00016713981547196593, |
| "loss": 1.1809, |
| "step": 465 |
| }, |
| { |
| "epoch": 0.66, |
| "grad_norm": 0.5337138175964355, |
| "learning_rate": 0.00016706884315117107, |
| "loss": 1.1381, |
| "step": 466 |
| }, |
| { |
| "epoch": 0.66, |
| "grad_norm": 0.565428614616394, |
| "learning_rate": 0.00016699787083037618, |
| "loss": 1.2766, |
| "step": 467 |
| }, |
| { |
| "epoch": 0.66, |
| "grad_norm": 0.5666337013244629, |
| "learning_rate": 0.00016692689850958127, |
| "loss": 1.2236, |
| "step": 468 |
| }, |
| { |
| "epoch": 0.66, |
| "grad_norm": 0.5685385465621948, |
| "learning_rate": 0.00016685592618878638, |
| "loss": 1.1651, |
| "step": 469 |
| }, |
| { |
| "epoch": 0.67, |
| "grad_norm": 0.5501173734664917, |
| "learning_rate": 0.0001667849538679915, |
| "loss": 1.2433, |
| "step": 470 |
| }, |
| { |
| "epoch": 0.67, |
| "grad_norm": 0.5514901280403137, |
| "learning_rate": 0.0001667139815471966, |
| "loss": 1.1912, |
| "step": 471 |
| }, |
| { |
| "epoch": 0.67, |
| "grad_norm": 0.5558573603630066, |
| "learning_rate": 0.0001666430092264017, |
| "loss": 1.2777, |
| "step": 472 |
| }, |
| { |
| "epoch": 0.67, |
| "grad_norm": 0.5633502006530762, |
| "learning_rate": 0.00016657203690560683, |
| "loss": 1.2223, |
| "step": 473 |
| }, |
| { |
| "epoch": 0.67, |
| "grad_norm": 0.5611795783042908, |
| "learning_rate": 0.00016650106458481194, |
| "loss": 1.156, |
| "step": 474 |
| }, |
| { |
| "epoch": 0.67, |
| "grad_norm": 0.5518985986709595, |
| "learning_rate": 0.00016643009226401705, |
| "loss": 1.2571, |
| "step": 475 |
| }, |
| { |
| "epoch": 0.67, |
| "grad_norm": 0.5667358040809631, |
| "learning_rate": 0.00016635911994322214, |
| "loss": 1.2574, |
| "step": 476 |
| }, |
| { |
| "epoch": 0.68, |
| "grad_norm": 0.5549746751785278, |
| "learning_rate": 0.00016628814762242725, |
| "loss": 1.2252, |
| "step": 477 |
| }, |
| { |
| "epoch": 0.68, |
| "grad_norm": 0.5677898526191711, |
| "learning_rate": 0.00016621717530163237, |
| "loss": 1.2288, |
| "step": 478 |
| }, |
| { |
| "epoch": 0.68, |
| "grad_norm": 0.5517006516456604, |
| "learning_rate": 0.00016614620298083748, |
| "loss": 1.2153, |
| "step": 479 |
| }, |
| { |
| "epoch": 0.68, |
| "grad_norm": 0.5561969876289368, |
| "learning_rate": 0.0001660752306600426, |
| "loss": 1.2308, |
| "step": 480 |
| }, |
| { |
| "epoch": 0.68, |
| "grad_norm": 0.5672180652618408, |
| "learning_rate": 0.0001660042583392477, |
| "loss": 1.2362, |
| "step": 481 |
| }, |
| { |
| "epoch": 0.68, |
| "grad_norm": 0.5887801647186279, |
| "learning_rate": 0.00016593328601845282, |
| "loss": 1.2541, |
| "step": 482 |
| }, |
| { |
| "epoch": 0.68, |
| "grad_norm": 0.5410957336425781, |
| "learning_rate": 0.00016586231369765793, |
| "loss": 1.198, |
| "step": 483 |
| }, |
| { |
| "epoch": 0.69, |
| "grad_norm": 0.5757806897163391, |
| "learning_rate": 0.00016579134137686304, |
| "loss": 1.2226, |
| "step": 484 |
| }, |
| { |
| "epoch": 0.69, |
| "grad_norm": 0.5371631979942322, |
| "learning_rate": 0.00016572036905606813, |
| "loss": 1.1952, |
| "step": 485 |
| }, |
| { |
| "epoch": 0.69, |
| "grad_norm": 0.528356671333313, |
| "learning_rate": 0.00016564939673527324, |
| "loss": 1.2062, |
| "step": 486 |
| }, |
| { |
| "epoch": 0.69, |
| "grad_norm": 0.5555338859558105, |
| "learning_rate": 0.00016557842441447835, |
| "loss": 1.283, |
| "step": 487 |
| }, |
| { |
| "epoch": 0.69, |
| "grad_norm": 0.5706049799919128, |
| "learning_rate": 0.00016550745209368347, |
| "loss": 1.1803, |
| "step": 488 |
| }, |
| { |
| "epoch": 0.69, |
| "grad_norm": 0.5852779150009155, |
| "learning_rate": 0.00016543647977288858, |
| "loss": 1.2062, |
| "step": 489 |
| }, |
| { |
| "epoch": 0.69, |
| "grad_norm": 0.5755358934402466, |
| "learning_rate": 0.0001653655074520937, |
| "loss": 1.3143, |
| "step": 490 |
| }, |
| { |
| "epoch": 0.7, |
| "grad_norm": 0.5640034079551697, |
| "learning_rate": 0.0001652945351312988, |
| "loss": 1.1585, |
| "step": 491 |
| }, |
| { |
| "epoch": 0.7, |
| "grad_norm": 0.576317310333252, |
| "learning_rate": 0.00016522356281050392, |
| "loss": 1.332, |
| "step": 492 |
| }, |
| { |
| "epoch": 0.7, |
| "grad_norm": 0.5770896077156067, |
| "learning_rate": 0.000165152590489709, |
| "loss": 1.2555, |
| "step": 493 |
| }, |
| { |
| "epoch": 0.7, |
| "grad_norm": 0.576921284198761, |
| "learning_rate": 0.00016508161816891412, |
| "loss": 1.3109, |
| "step": 494 |
| }, |
| { |
| "epoch": 0.7, |
| "grad_norm": 0.5501707792282104, |
| "learning_rate": 0.00016501064584811926, |
| "loss": 1.2497, |
| "step": 495 |
| }, |
| { |
| "epoch": 0.7, |
| "grad_norm": 0.5693079829216003, |
| "learning_rate": 0.00016493967352732437, |
| "loss": 1.2341, |
| "step": 496 |
| }, |
| { |
| "epoch": 0.7, |
| "grad_norm": 0.5721957087516785, |
| "learning_rate": 0.00016486870120652945, |
| "loss": 1.1374, |
| "step": 497 |
| }, |
| { |
| "epoch": 0.71, |
| "grad_norm": 0.5555086135864258, |
| "learning_rate": 0.00016479772888573457, |
| "loss": 1.1855, |
| "step": 498 |
| }, |
| { |
| "epoch": 0.71, |
| "grad_norm": 0.5749430060386658, |
| "learning_rate": 0.00016472675656493968, |
| "loss": 1.1561, |
| "step": 499 |
| }, |
| { |
| "epoch": 0.71, |
| "grad_norm": 0.5816613435745239, |
| "learning_rate": 0.0001646557842441448, |
| "loss": 1.2336, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.71, |
| "grad_norm": 0.5536159873008728, |
| "learning_rate": 0.00016458481192334988, |
| "loss": 1.2006, |
| "step": 501 |
| }, |
| { |
| "epoch": 0.71, |
| "grad_norm": 0.5615043044090271, |
| "learning_rate": 0.00016451383960255502, |
| "loss": 1.2464, |
| "step": 502 |
| }, |
| { |
| "epoch": 0.71, |
| "grad_norm": 0.5301942825317383, |
| "learning_rate": 0.00016444286728176013, |
| "loss": 1.0993, |
| "step": 503 |
| }, |
| { |
| "epoch": 0.71, |
| "grad_norm": 0.5370047092437744, |
| "learning_rate": 0.00016437189496096524, |
| "loss": 1.2876, |
| "step": 504 |
| }, |
| { |
| "epoch": 0.72, |
| "grad_norm": 0.549637496471405, |
| "learning_rate": 0.00016430092264017033, |
| "loss": 1.2099, |
| "step": 505 |
| }, |
| { |
| "epoch": 0.72, |
| "grad_norm": 0.5674926042556763, |
| "learning_rate": 0.00016422995031937544, |
| "loss": 1.2473, |
| "step": 506 |
| }, |
| { |
| "epoch": 0.72, |
| "grad_norm": 0.5429442524909973, |
| "learning_rate": 0.00016415897799858055, |
| "loss": 1.1994, |
| "step": 507 |
| }, |
| { |
| "epoch": 0.72, |
| "grad_norm": 0.5792283415794373, |
| "learning_rate": 0.00016408800567778567, |
| "loss": 1.1834, |
| "step": 508 |
| }, |
| { |
| "epoch": 0.72, |
| "grad_norm": 0.5722516179084778, |
| "learning_rate": 0.00016401703335699078, |
| "loss": 1.1922, |
| "step": 509 |
| }, |
| { |
| "epoch": 0.72, |
| "grad_norm": 0.6014196276664734, |
| "learning_rate": 0.0001639460610361959, |
| "loss": 1.1813, |
| "step": 510 |
| }, |
| { |
| "epoch": 0.72, |
| "grad_norm": 0.5406440496444702, |
| "learning_rate": 0.000163875088715401, |
| "loss": 1.1248, |
| "step": 511 |
| }, |
| { |
| "epoch": 0.73, |
| "grad_norm": 0.5606802701950073, |
| "learning_rate": 0.00016380411639460612, |
| "loss": 1.2143, |
| "step": 512 |
| }, |
| { |
| "epoch": 0.73, |
| "grad_norm": 0.5731990933418274, |
| "learning_rate": 0.00016373314407381123, |
| "loss": 1.2202, |
| "step": 513 |
| }, |
| { |
| "epoch": 0.73, |
| "grad_norm": 0.5451279282569885, |
| "learning_rate": 0.00016366217175301632, |
| "loss": 1.2338, |
| "step": 514 |
| }, |
| { |
| "epoch": 0.73, |
| "grad_norm": 0.5566918849945068, |
| "learning_rate": 0.00016359119943222143, |
| "loss": 1.2574, |
| "step": 515 |
| }, |
| { |
| "epoch": 0.73, |
| "grad_norm": 0.5759538412094116, |
| "learning_rate": 0.00016352022711142657, |
| "loss": 1.2868, |
| "step": 516 |
| }, |
| { |
| "epoch": 0.73, |
| "grad_norm": 0.5528054237365723, |
| "learning_rate": 0.00016344925479063165, |
| "loss": 1.1986, |
| "step": 517 |
| }, |
| { |
| "epoch": 0.73, |
| "grad_norm": 0.5640310645103455, |
| "learning_rate": 0.00016337828246983677, |
| "loss": 1.2458, |
| "step": 518 |
| }, |
| { |
| "epoch": 0.74, |
| "grad_norm": 0.5853886604309082, |
| "learning_rate": 0.00016330731014904188, |
| "loss": 1.1772, |
| "step": 519 |
| }, |
| { |
| "epoch": 0.74, |
| "grad_norm": 0.5519319176673889, |
| "learning_rate": 0.000163236337828247, |
| "loss": 1.1771, |
| "step": 520 |
| }, |
| { |
| "epoch": 0.74, |
| "grad_norm": 0.571728527545929, |
| "learning_rate": 0.0001631653655074521, |
| "loss": 1.1939, |
| "step": 521 |
| }, |
| { |
| "epoch": 0.74, |
| "grad_norm": 0.5816935896873474, |
| "learning_rate": 0.0001630943931866572, |
| "loss": 1.1395, |
| "step": 522 |
| }, |
| { |
| "epoch": 0.74, |
| "grad_norm": 0.5645069479942322, |
| "learning_rate": 0.00016302342086586233, |
| "loss": 1.2277, |
| "step": 523 |
| }, |
| { |
| "epoch": 0.74, |
| "grad_norm": 0.5603705048561096, |
| "learning_rate": 0.00016295244854506744, |
| "loss": 1.2194, |
| "step": 524 |
| }, |
| { |
| "epoch": 0.74, |
| "grad_norm": 0.5555822253227234, |
| "learning_rate": 0.00016288147622427256, |
| "loss": 1.166, |
| "step": 525 |
| }, |
| { |
| "epoch": 0.75, |
| "grad_norm": 0.5700538158416748, |
| "learning_rate": 0.00016281050390347764, |
| "loss": 1.2747, |
| "step": 526 |
| }, |
| { |
| "epoch": 0.75, |
| "grad_norm": 0.5572327375411987, |
| "learning_rate": 0.00016273953158268275, |
| "loss": 1.1742, |
| "step": 527 |
| }, |
| { |
| "epoch": 0.75, |
| "grad_norm": 0.5874422192573547, |
| "learning_rate": 0.00016266855926188787, |
| "loss": 1.2637, |
| "step": 528 |
| }, |
| { |
| "epoch": 0.75, |
| "grad_norm": 0.61215740442276, |
| "learning_rate": 0.00016259758694109298, |
| "loss": 1.2082, |
| "step": 529 |
| }, |
| { |
| "epoch": 0.75, |
| "grad_norm": 0.5912395119667053, |
| "learning_rate": 0.0001625266146202981, |
| "loss": 1.2397, |
| "step": 530 |
| }, |
| { |
| "epoch": 0.75, |
| "grad_norm": 0.5844393968582153, |
| "learning_rate": 0.0001624556422995032, |
| "loss": 1.2337, |
| "step": 531 |
| }, |
| { |
| "epoch": 0.75, |
| "grad_norm": 0.5393280982971191, |
| "learning_rate": 0.00016238466997870832, |
| "loss": 1.1741, |
| "step": 532 |
| }, |
| { |
| "epoch": 0.76, |
| "grad_norm": 0.5622760653495789, |
| "learning_rate": 0.00016231369765791343, |
| "loss": 1.1932, |
| "step": 533 |
| }, |
| { |
| "epoch": 0.76, |
| "grad_norm": 0.5254821181297302, |
| "learning_rate": 0.00016224272533711852, |
| "loss": 1.1992, |
| "step": 534 |
| }, |
| { |
| "epoch": 0.76, |
| "grad_norm": 0.5416676998138428, |
| "learning_rate": 0.00016217175301632363, |
| "loss": 1.2038, |
| "step": 535 |
| }, |
| { |
| "epoch": 0.76, |
| "grad_norm": 0.5292723178863525, |
| "learning_rate": 0.00016210078069552874, |
| "loss": 1.2346, |
| "step": 536 |
| }, |
| { |
| "epoch": 0.76, |
| "grad_norm": 0.5630815029144287, |
| "learning_rate": 0.00016202980837473388, |
| "loss": 1.1923, |
| "step": 537 |
| }, |
| { |
| "epoch": 0.76, |
| "grad_norm": 0.5653424263000488, |
| "learning_rate": 0.00016195883605393897, |
| "loss": 1.2701, |
| "step": 538 |
| }, |
| { |
| "epoch": 0.76, |
| "grad_norm": 0.5717771053314209, |
| "learning_rate": 0.00016188786373314408, |
| "loss": 1.2556, |
| "step": 539 |
| }, |
| { |
| "epoch": 0.77, |
| "grad_norm": 0.5919432640075684, |
| "learning_rate": 0.0001618168914123492, |
| "loss": 1.2069, |
| "step": 540 |
| }, |
| { |
| "epoch": 0.77, |
| "grad_norm": 0.5614644885063171, |
| "learning_rate": 0.0001617459190915543, |
| "loss": 1.24, |
| "step": 541 |
| }, |
| { |
| "epoch": 0.77, |
| "grad_norm": 0.5731059908866882, |
| "learning_rate": 0.00016167494677075942, |
| "loss": 1.1578, |
| "step": 542 |
| }, |
| { |
| "epoch": 0.77, |
| "grad_norm": 0.5611764192581177, |
| "learning_rate": 0.0001616039744499645, |
| "loss": 1.1595, |
| "step": 543 |
| }, |
| { |
| "epoch": 0.77, |
| "grad_norm": 0.5955880880355835, |
| "learning_rate": 0.00016153300212916964, |
| "loss": 1.1857, |
| "step": 544 |
| }, |
| { |
| "epoch": 0.77, |
| "grad_norm": 0.5684686303138733, |
| "learning_rate": 0.00016146202980837476, |
| "loss": 1.1648, |
| "step": 545 |
| }, |
| { |
| "epoch": 0.77, |
| "grad_norm": 0.5427114367485046, |
| "learning_rate": 0.00016139105748757987, |
| "loss": 1.2056, |
| "step": 546 |
| }, |
| { |
| "epoch": 0.77, |
| "grad_norm": 0.5569632053375244, |
| "learning_rate": 0.00016132008516678495, |
| "loss": 1.1554, |
| "step": 547 |
| }, |
| { |
| "epoch": 0.78, |
| "grad_norm": 0.5450757145881653, |
| "learning_rate": 0.00016124911284599007, |
| "loss": 1.1602, |
| "step": 548 |
| }, |
| { |
| "epoch": 0.78, |
| "grad_norm": 0.5493456125259399, |
| "learning_rate": 0.00016117814052519518, |
| "loss": 1.1994, |
| "step": 549 |
| }, |
| { |
| "epoch": 0.78, |
| "grad_norm": 0.5915331244468689, |
| "learning_rate": 0.0001611071682044003, |
| "loss": 1.2192, |
| "step": 550 |
| }, |
| { |
| "epoch": 0.78, |
| "grad_norm": 0.5696746110916138, |
| "learning_rate": 0.0001610361958836054, |
| "loss": 1.1858, |
| "step": 551 |
| }, |
| { |
| "epoch": 0.78, |
| "grad_norm": 0.5555996894836426, |
| "learning_rate": 0.00016096522356281052, |
| "loss": 1.2077, |
| "step": 552 |
| }, |
| { |
| "epoch": 0.78, |
| "grad_norm": 0.5695299506187439, |
| "learning_rate": 0.00016089425124201563, |
| "loss": 1.2805, |
| "step": 553 |
| }, |
| { |
| "epoch": 0.78, |
| "grad_norm": 0.5820121169090271, |
| "learning_rate": 0.00016082327892122074, |
| "loss": 1.1841, |
| "step": 554 |
| }, |
| { |
| "epoch": 0.79, |
| "grad_norm": 0.5951303243637085, |
| "learning_rate": 0.00016075230660042583, |
| "loss": 1.2728, |
| "step": 555 |
| }, |
| { |
| "epoch": 0.79, |
| "grad_norm": 0.5539563298225403, |
| "learning_rate": 0.00016068133427963094, |
| "loss": 1.2861, |
| "step": 556 |
| }, |
| { |
| "epoch": 0.79, |
| "grad_norm": 0.5624521970748901, |
| "learning_rate": 0.00016061036195883605, |
| "loss": 1.1379, |
| "step": 557 |
| }, |
| { |
| "epoch": 0.79, |
| "grad_norm": 0.5960082411766052, |
| "learning_rate": 0.0001605393896380412, |
| "loss": 1.2268, |
| "step": 558 |
| }, |
| { |
| "epoch": 0.79, |
| "grad_norm": 0.5910595059394836, |
| "learning_rate": 0.00016046841731724628, |
| "loss": 1.1165, |
| "step": 559 |
| }, |
| { |
| "epoch": 0.79, |
| "grad_norm": 0.6063516139984131, |
| "learning_rate": 0.0001603974449964514, |
| "loss": 1.2154, |
| "step": 560 |
| }, |
| { |
| "epoch": 0.79, |
| "grad_norm": 0.5970266461372375, |
| "learning_rate": 0.0001603264726756565, |
| "loss": 1.2419, |
| "step": 561 |
| }, |
| { |
| "epoch": 0.8, |
| "grad_norm": 0.5765990018844604, |
| "learning_rate": 0.00016025550035486162, |
| "loss": 1.1821, |
| "step": 562 |
| }, |
| { |
| "epoch": 0.8, |
| "grad_norm": 0.556180477142334, |
| "learning_rate": 0.0001601845280340667, |
| "loss": 1.206, |
| "step": 563 |
| }, |
| { |
| "epoch": 0.8, |
| "grad_norm": 0.5440099835395813, |
| "learning_rate": 0.00016011355571327182, |
| "loss": 1.2757, |
| "step": 564 |
| }, |
| { |
| "epoch": 0.8, |
| "grad_norm": 0.5394881963729858, |
| "learning_rate": 0.00016004258339247696, |
| "loss": 1.1814, |
| "step": 565 |
| }, |
| { |
| "epoch": 0.8, |
| "grad_norm": 0.5824431777000427, |
| "learning_rate": 0.00015997161107168207, |
| "loss": 1.1821, |
| "step": 566 |
| }, |
| { |
| "epoch": 0.8, |
| "grad_norm": 0.6307337880134583, |
| "learning_rate": 0.00015990063875088715, |
| "loss": 1.2679, |
| "step": 567 |
| }, |
| { |
| "epoch": 0.8, |
| "grad_norm": 0.5689281225204468, |
| "learning_rate": 0.00015982966643009227, |
| "loss": 1.2582, |
| "step": 568 |
| }, |
| { |
| "epoch": 0.81, |
| "grad_norm": 0.5547791123390198, |
| "learning_rate": 0.00015975869410929738, |
| "loss": 1.2367, |
| "step": 569 |
| }, |
| { |
| "epoch": 0.81, |
| "grad_norm": 0.6274536848068237, |
| "learning_rate": 0.0001596877217885025, |
| "loss": 1.1979, |
| "step": 570 |
| }, |
| { |
| "epoch": 0.81, |
| "grad_norm": 0.5949268341064453, |
| "learning_rate": 0.0001596167494677076, |
| "loss": 1.2517, |
| "step": 571 |
| }, |
| { |
| "epoch": 0.81, |
| "grad_norm": 0.5719754099845886, |
| "learning_rate": 0.00015954577714691272, |
| "loss": 1.2743, |
| "step": 572 |
| }, |
| { |
| "epoch": 0.81, |
| "grad_norm": 0.5814628601074219, |
| "learning_rate": 0.00015947480482611783, |
| "loss": 1.1369, |
| "step": 573 |
| }, |
| { |
| "epoch": 0.81, |
| "grad_norm": 0.5738956928253174, |
| "learning_rate": 0.00015940383250532294, |
| "loss": 1.1537, |
| "step": 574 |
| }, |
| { |
| "epoch": 0.81, |
| "grad_norm": 0.6162692904472351, |
| "learning_rate": 0.00015933286018452806, |
| "loss": 1.139, |
| "step": 575 |
| }, |
| { |
| "epoch": 0.82, |
| "grad_norm": 0.5667981505393982, |
| "learning_rate": 0.00015926188786373314, |
| "loss": 1.2606, |
| "step": 576 |
| }, |
| { |
| "epoch": 0.82, |
| "grad_norm": 0.6261587142944336, |
| "learning_rate": 0.00015919091554293825, |
| "loss": 1.2971, |
| "step": 577 |
| }, |
| { |
| "epoch": 0.82, |
| "grad_norm": 0.5734363794326782, |
| "learning_rate": 0.00015911994322214337, |
| "loss": 1.248, |
| "step": 578 |
| }, |
| { |
| "epoch": 0.82, |
| "grad_norm": 0.5452428460121155, |
| "learning_rate": 0.00015904897090134848, |
| "loss": 1.1698, |
| "step": 579 |
| }, |
| { |
| "epoch": 0.82, |
| "grad_norm": 0.5635668635368347, |
| "learning_rate": 0.0001589779985805536, |
| "loss": 1.2266, |
| "step": 580 |
| }, |
| { |
| "epoch": 0.82, |
| "grad_norm": 0.5433382987976074, |
| "learning_rate": 0.0001589070262597587, |
| "loss": 1.1104, |
| "step": 581 |
| }, |
| { |
| "epoch": 0.82, |
| "grad_norm": 0.5507457256317139, |
| "learning_rate": 0.00015883605393896382, |
| "loss": 1.2375, |
| "step": 582 |
| }, |
| { |
| "epoch": 0.83, |
| "grad_norm": 0.5394136309623718, |
| "learning_rate": 0.00015876508161816893, |
| "loss": 1.1949, |
| "step": 583 |
| }, |
| { |
| "epoch": 0.83, |
| "grad_norm": 0.5387130975723267, |
| "learning_rate": 0.00015869410929737402, |
| "loss": 1.1563, |
| "step": 584 |
| }, |
| { |
| "epoch": 0.83, |
| "grad_norm": 0.5782540440559387, |
| "learning_rate": 0.00015862313697657913, |
| "loss": 1.2099, |
| "step": 585 |
| }, |
| { |
| "epoch": 0.83, |
| "grad_norm": 0.5795274972915649, |
| "learning_rate": 0.00015855216465578424, |
| "loss": 1.1736, |
| "step": 586 |
| }, |
| { |
| "epoch": 0.83, |
| "grad_norm": 0.5905880331993103, |
| "learning_rate": 0.00015848119233498938, |
| "loss": 1.1577, |
| "step": 587 |
| }, |
| { |
| "epoch": 0.83, |
| "grad_norm": 0.6062709093093872, |
| "learning_rate": 0.00015841022001419447, |
| "loss": 1.1655, |
| "step": 588 |
| }, |
| { |
| "epoch": 0.83, |
| "grad_norm": 0.6285403370857239, |
| "learning_rate": 0.00015833924769339958, |
| "loss": 1.1556, |
| "step": 589 |
| }, |
| { |
| "epoch": 0.84, |
| "grad_norm": 0.5971753001213074, |
| "learning_rate": 0.0001582682753726047, |
| "loss": 1.2227, |
| "step": 590 |
| }, |
| { |
| "epoch": 0.84, |
| "grad_norm": 0.5791364312171936, |
| "learning_rate": 0.0001581973030518098, |
| "loss": 1.2093, |
| "step": 591 |
| }, |
| { |
| "epoch": 0.84, |
| "grad_norm": 0.5568805932998657, |
| "learning_rate": 0.0001581263307310149, |
| "loss": 1.2346, |
| "step": 592 |
| }, |
| { |
| "epoch": 0.84, |
| "grad_norm": 0.5391885042190552, |
| "learning_rate": 0.00015805535841022, |
| "loss": 1.1697, |
| "step": 593 |
| }, |
| { |
| "epoch": 0.84, |
| "grad_norm": 0.567952036857605, |
| "learning_rate": 0.00015798438608942514, |
| "loss": 1.2303, |
| "step": 594 |
| }, |
| { |
| "epoch": 0.84, |
| "grad_norm": 0.5382447242736816, |
| "learning_rate": 0.00015791341376863026, |
| "loss": 1.1986, |
| "step": 595 |
| }, |
| { |
| "epoch": 0.84, |
| "grad_norm": 0.5621328949928284, |
| "learning_rate": 0.00015784244144783534, |
| "loss": 1.1619, |
| "step": 596 |
| }, |
| { |
| "epoch": 0.85, |
| "grad_norm": 0.6134306788444519, |
| "learning_rate": 0.00015777146912704046, |
| "loss": 1.1331, |
| "step": 597 |
| }, |
| { |
| "epoch": 0.85, |
| "grad_norm": 0.57140052318573, |
| "learning_rate": 0.00015770049680624557, |
| "loss": 1.1858, |
| "step": 598 |
| }, |
| { |
| "epoch": 0.85, |
| "grad_norm": 0.5893669128417969, |
| "learning_rate": 0.00015762952448545068, |
| "loss": 1.2289, |
| "step": 599 |
| }, |
| { |
| "epoch": 0.85, |
| "grad_norm": 0.5760409832000732, |
| "learning_rate": 0.0001575585521646558, |
| "loss": 1.2426, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.85, |
| "grad_norm": 0.591411828994751, |
| "learning_rate": 0.0001574875798438609, |
| "loss": 1.195, |
| "step": 601 |
| }, |
| { |
| "epoch": 0.85, |
| "grad_norm": 0.5461949706077576, |
| "learning_rate": 0.00015741660752306602, |
| "loss": 1.1627, |
| "step": 602 |
| }, |
| { |
| "epoch": 0.85, |
| "grad_norm": 0.5686529874801636, |
| "learning_rate": 0.00015734563520227113, |
| "loss": 1.1048, |
| "step": 603 |
| }, |
| { |
| "epoch": 0.86, |
| "grad_norm": 0.5683251023292542, |
| "learning_rate": 0.00015727466288147624, |
| "loss": 1.2594, |
| "step": 604 |
| }, |
| { |
| "epoch": 0.86, |
| "grad_norm": 0.5748125314712524, |
| "learning_rate": 0.00015720369056068133, |
| "loss": 1.151, |
| "step": 605 |
| }, |
| { |
| "epoch": 0.86, |
| "grad_norm": 0.5953657031059265, |
| "learning_rate": 0.00015713271823988644, |
| "loss": 1.2148, |
| "step": 606 |
| }, |
| { |
| "epoch": 0.86, |
| "grad_norm": 0.5915418863296509, |
| "learning_rate": 0.00015706174591909156, |
| "loss": 1.2249, |
| "step": 607 |
| }, |
| { |
| "epoch": 0.86, |
| "grad_norm": 0.5777379274368286, |
| "learning_rate": 0.00015699077359829667, |
| "loss": 1.18, |
| "step": 608 |
| }, |
| { |
| "epoch": 0.86, |
| "grad_norm": 0.592585563659668, |
| "learning_rate": 0.00015691980127750178, |
| "loss": 1.2306, |
| "step": 609 |
| }, |
| { |
| "epoch": 0.86, |
| "grad_norm": 0.5442766547203064, |
| "learning_rate": 0.0001568488289567069, |
| "loss": 1.1789, |
| "step": 610 |
| }, |
| { |
| "epoch": 0.87, |
| "grad_norm": 0.5644369721412659, |
| "learning_rate": 0.000156777856635912, |
| "loss": 1.2501, |
| "step": 611 |
| }, |
| { |
| "epoch": 0.87, |
| "grad_norm": 0.55912846326828, |
| "learning_rate": 0.00015670688431511712, |
| "loss": 1.2893, |
| "step": 612 |
| }, |
| { |
| "epoch": 0.87, |
| "grad_norm": 0.5613831877708435, |
| "learning_rate": 0.0001566359119943222, |
| "loss": 1.1141, |
| "step": 613 |
| }, |
| { |
| "epoch": 0.87, |
| "grad_norm": 0.5890555381774902, |
| "learning_rate": 0.00015656493967352732, |
| "loss": 1.2149, |
| "step": 614 |
| }, |
| { |
| "epoch": 0.87, |
| "grad_norm": 0.5675429105758667, |
| "learning_rate": 0.00015649396735273246, |
| "loss": 1.2001, |
| "step": 615 |
| }, |
| { |
| "epoch": 0.87, |
| "grad_norm": 0.5950176119804382, |
| "learning_rate": 0.00015642299503193757, |
| "loss": 1.2019, |
| "step": 616 |
| }, |
| { |
| "epoch": 0.87, |
| "grad_norm": 0.5831938982009888, |
| "learning_rate": 0.00015635202271114266, |
| "loss": 1.1673, |
| "step": 617 |
| }, |
| { |
| "epoch": 0.88, |
| "grad_norm": 0.5806068181991577, |
| "learning_rate": 0.00015628105039034777, |
| "loss": 1.2013, |
| "step": 618 |
| }, |
| { |
| "epoch": 0.88, |
| "grad_norm": 0.5464046001434326, |
| "learning_rate": 0.00015621007806955288, |
| "loss": 1.1557, |
| "step": 619 |
| }, |
| { |
| "epoch": 0.88, |
| "grad_norm": 0.6087239980697632, |
| "learning_rate": 0.000156139105748758, |
| "loss": 1.2024, |
| "step": 620 |
| }, |
| { |
| "epoch": 0.88, |
| "grad_norm": 0.544928252696991, |
| "learning_rate": 0.00015606813342796308, |
| "loss": 1.1968, |
| "step": 621 |
| }, |
| { |
| "epoch": 0.88, |
| "grad_norm": 0.5821595788002014, |
| "learning_rate": 0.00015599716110716822, |
| "loss": 1.3009, |
| "step": 622 |
| }, |
| { |
| "epoch": 0.88, |
| "grad_norm": 0.5546013116836548, |
| "learning_rate": 0.00015592618878637333, |
| "loss": 1.1127, |
| "step": 623 |
| }, |
| { |
| "epoch": 0.88, |
| "grad_norm": 0.5370335578918457, |
| "learning_rate": 0.00015585521646557844, |
| "loss": 1.1543, |
| "step": 624 |
| }, |
| { |
| "epoch": 0.89, |
| "grad_norm": 0.5727035999298096, |
| "learning_rate": 0.00015578424414478353, |
| "loss": 1.2239, |
| "step": 625 |
| }, |
| { |
| "epoch": 0.89, |
| "grad_norm": 0.5680972933769226, |
| "learning_rate": 0.00015571327182398864, |
| "loss": 1.2741, |
| "step": 626 |
| }, |
| { |
| "epoch": 0.89, |
| "grad_norm": 0.5887457728385925, |
| "learning_rate": 0.00015564229950319376, |
| "loss": 1.2175, |
| "step": 627 |
| }, |
| { |
| "epoch": 0.89, |
| "grad_norm": 0.5596421957015991, |
| "learning_rate": 0.00015557132718239887, |
| "loss": 1.199, |
| "step": 628 |
| }, |
| { |
| "epoch": 0.89, |
| "grad_norm": 0.5555490851402283, |
| "learning_rate": 0.00015550035486160398, |
| "loss": 1.125, |
| "step": 629 |
| }, |
| { |
| "epoch": 0.89, |
| "grad_norm": 0.5999779105186462, |
| "learning_rate": 0.0001554293825408091, |
| "loss": 1.2344, |
| "step": 630 |
| }, |
| { |
| "epoch": 0.89, |
| "grad_norm": 0.6120169162750244, |
| "learning_rate": 0.0001553584102200142, |
| "loss": 1.1822, |
| "step": 631 |
| }, |
| { |
| "epoch": 0.9, |
| "grad_norm": 0.5653102993965149, |
| "learning_rate": 0.00015528743789921932, |
| "loss": 1.2004, |
| "step": 632 |
| }, |
| { |
| "epoch": 0.9, |
| "grad_norm": 0.5853379964828491, |
| "learning_rate": 0.00015521646557842443, |
| "loss": 1.1797, |
| "step": 633 |
| }, |
| { |
| "epoch": 0.9, |
| "grad_norm": 0.563293993473053, |
| "learning_rate": 0.00015514549325762952, |
| "loss": 1.1991, |
| "step": 634 |
| }, |
| { |
| "epoch": 0.9, |
| "grad_norm": 0.5557974576950073, |
| "learning_rate": 0.00015507452093683463, |
| "loss": 1.1843, |
| "step": 635 |
| }, |
| { |
| "epoch": 0.9, |
| "grad_norm": 0.5982331037521362, |
| "learning_rate": 0.00015500354861603977, |
| "loss": 1.1827, |
| "step": 636 |
| }, |
| { |
| "epoch": 0.9, |
| "grad_norm": 0.5909957885742188, |
| "learning_rate": 0.00015493257629524486, |
| "loss": 1.1338, |
| "step": 637 |
| }, |
| { |
| "epoch": 0.9, |
| "grad_norm": 0.5773521661758423, |
| "learning_rate": 0.00015486160397444997, |
| "loss": 1.2313, |
| "step": 638 |
| }, |
| { |
| "epoch": 0.91, |
| "grad_norm": 0.5929623246192932, |
| "learning_rate": 0.00015479063165365508, |
| "loss": 1.144, |
| "step": 639 |
| }, |
| { |
| "epoch": 0.91, |
| "grad_norm": 0.5677060484886169, |
| "learning_rate": 0.0001547196593328602, |
| "loss": 1.1656, |
| "step": 640 |
| }, |
| { |
| "epoch": 0.91, |
| "grad_norm": 0.552845299243927, |
| "learning_rate": 0.0001546486870120653, |
| "loss": 1.1576, |
| "step": 641 |
| }, |
| { |
| "epoch": 0.91, |
| "grad_norm": 0.5725761651992798, |
| "learning_rate": 0.0001545777146912704, |
| "loss": 1.1562, |
| "step": 642 |
| }, |
| { |
| "epoch": 0.91, |
| "grad_norm": 0.5661845803260803, |
| "learning_rate": 0.00015450674237047553, |
| "loss": 1.1172, |
| "step": 643 |
| }, |
| { |
| "epoch": 0.91, |
| "grad_norm": 0.6051177382469177, |
| "learning_rate": 0.00015443577004968064, |
| "loss": 1.2199, |
| "step": 644 |
| }, |
| { |
| "epoch": 0.91, |
| "grad_norm": 0.575823187828064, |
| "learning_rate": 0.00015436479772888576, |
| "loss": 1.2077, |
| "step": 645 |
| }, |
| { |
| "epoch": 0.92, |
| "grad_norm": 0.5754451751708984, |
| "learning_rate": 0.00015429382540809084, |
| "loss": 1.204, |
| "step": 646 |
| }, |
| { |
| "epoch": 0.92, |
| "grad_norm": 0.5662161111831665, |
| "learning_rate": 0.00015422285308729596, |
| "loss": 1.2517, |
| "step": 647 |
| }, |
| { |
| "epoch": 0.92, |
| "grad_norm": 0.62320476770401, |
| "learning_rate": 0.00015415188076650107, |
| "loss": 1.2593, |
| "step": 648 |
| }, |
| { |
| "epoch": 0.92, |
| "grad_norm": 0.5987563729286194, |
| "learning_rate": 0.00015408090844570618, |
| "loss": 1.1732, |
| "step": 649 |
| }, |
| { |
| "epoch": 0.92, |
| "grad_norm": 0.5595415234565735, |
| "learning_rate": 0.0001540099361249113, |
| "loss": 1.2089, |
| "step": 650 |
| }, |
| { |
| "epoch": 0.92, |
| "grad_norm": 0.5797716379165649, |
| "learning_rate": 0.0001539389638041164, |
| "loss": 1.143, |
| "step": 651 |
| }, |
| { |
| "epoch": 0.92, |
| "grad_norm": 0.5579701662063599, |
| "learning_rate": 0.00015386799148332152, |
| "loss": 1.131, |
| "step": 652 |
| }, |
| { |
| "epoch": 0.93, |
| "grad_norm": 0.6035544276237488, |
| "learning_rate": 0.00015379701916252663, |
| "loss": 1.1589, |
| "step": 653 |
| }, |
| { |
| "epoch": 0.93, |
| "grad_norm": 0.6274803876876831, |
| "learning_rate": 0.00015372604684173172, |
| "loss": 1.2097, |
| "step": 654 |
| }, |
| { |
| "epoch": 0.93, |
| "grad_norm": 0.5816579461097717, |
| "learning_rate": 0.00015365507452093683, |
| "loss": 1.112, |
| "step": 655 |
| }, |
| { |
| "epoch": 0.93, |
| "grad_norm": 0.5723366737365723, |
| "learning_rate": 0.00015358410220014194, |
| "loss": 1.1616, |
| "step": 656 |
| }, |
| { |
| "epoch": 0.93, |
| "grad_norm": 0.5701184272766113, |
| "learning_rate": 0.00015351312987934708, |
| "loss": 1.1869, |
| "step": 657 |
| }, |
| { |
| "epoch": 0.93, |
| "grad_norm": 0.5675951838493347, |
| "learning_rate": 0.00015344215755855217, |
| "loss": 1.1766, |
| "step": 658 |
| }, |
| { |
| "epoch": 0.93, |
| "grad_norm": 0.5601312518119812, |
| "learning_rate": 0.00015337118523775728, |
| "loss": 1.1863, |
| "step": 659 |
| }, |
| { |
| "epoch": 0.94, |
| "grad_norm": 0.5662368535995483, |
| "learning_rate": 0.0001533002129169624, |
| "loss": 1.2104, |
| "step": 660 |
| }, |
| { |
| "epoch": 0.94, |
| "grad_norm": 0.5629364252090454, |
| "learning_rate": 0.0001532292405961675, |
| "loss": 1.172, |
| "step": 661 |
| }, |
| { |
| "epoch": 0.94, |
| "grad_norm": 0.5831137895584106, |
| "learning_rate": 0.00015315826827537262, |
| "loss": 1.1191, |
| "step": 662 |
| }, |
| { |
| "epoch": 0.94, |
| "grad_norm": 0.5520037412643433, |
| "learning_rate": 0.0001530872959545777, |
| "loss": 1.2139, |
| "step": 663 |
| }, |
| { |
| "epoch": 0.94, |
| "grad_norm": 0.5635483264923096, |
| "learning_rate": 0.00015301632363378282, |
| "loss": 1.2279, |
| "step": 664 |
| }, |
| { |
| "epoch": 0.94, |
| "grad_norm": 0.635055422782898, |
| "learning_rate": 0.00015294535131298796, |
| "loss": 1.3509, |
| "step": 665 |
| }, |
| { |
| "epoch": 0.94, |
| "grad_norm": 0.5712576508522034, |
| "learning_rate": 0.00015287437899219304, |
| "loss": 1.1376, |
| "step": 666 |
| }, |
| { |
| "epoch": 0.94, |
| "grad_norm": 0.6059345602989197, |
| "learning_rate": 0.00015280340667139816, |
| "loss": 1.1995, |
| "step": 667 |
| }, |
| { |
| "epoch": 0.95, |
| "grad_norm": 0.5608190298080444, |
| "learning_rate": 0.00015273243435060327, |
| "loss": 1.2018, |
| "step": 668 |
| }, |
| { |
| "epoch": 0.95, |
| "grad_norm": 0.583979606628418, |
| "learning_rate": 0.00015266146202980838, |
| "loss": 1.2649, |
| "step": 669 |
| }, |
| { |
| "epoch": 0.95, |
| "grad_norm": 0.5718328356742859, |
| "learning_rate": 0.0001525904897090135, |
| "loss": 1.2384, |
| "step": 670 |
| }, |
| { |
| "epoch": 0.95, |
| "grad_norm": 0.5428913235664368, |
| "learning_rate": 0.00015251951738821858, |
| "loss": 1.2453, |
| "step": 671 |
| }, |
| { |
| "epoch": 0.95, |
| "grad_norm": 0.5979950428009033, |
| "learning_rate": 0.00015244854506742372, |
| "loss": 1.191, |
| "step": 672 |
| }, |
| { |
| "epoch": 0.95, |
| "grad_norm": 0.5848691463470459, |
| "learning_rate": 0.00015237757274662883, |
| "loss": 1.2098, |
| "step": 673 |
| }, |
| { |
| "epoch": 0.95, |
| "grad_norm": 0.5947878360748291, |
| "learning_rate": 0.00015230660042583395, |
| "loss": 1.2174, |
| "step": 674 |
| }, |
| { |
| "epoch": 0.96, |
| "grad_norm": 0.5705249309539795, |
| "learning_rate": 0.00015223562810503903, |
| "loss": 1.1701, |
| "step": 675 |
| }, |
| { |
| "epoch": 0.96, |
| "grad_norm": 0.5978186726570129, |
| "learning_rate": 0.00015216465578424414, |
| "loss": 1.1468, |
| "step": 676 |
| }, |
| { |
| "epoch": 0.96, |
| "grad_norm": 0.5883394479751587, |
| "learning_rate": 0.00015209368346344926, |
| "loss": 1.1998, |
| "step": 677 |
| }, |
| { |
| "epoch": 0.96, |
| "grad_norm": 0.551814079284668, |
| "learning_rate": 0.00015202271114265437, |
| "loss": 1.1576, |
| "step": 678 |
| }, |
| { |
| "epoch": 0.96, |
| "grad_norm": 0.5914448499679565, |
| "learning_rate": 0.00015195173882185948, |
| "loss": 1.2328, |
| "step": 679 |
| }, |
| { |
| "epoch": 0.96, |
| "grad_norm": 0.6165048480033875, |
| "learning_rate": 0.0001518807665010646, |
| "loss": 1.203, |
| "step": 680 |
| }, |
| { |
| "epoch": 0.96, |
| "grad_norm": 0.5738925933837891, |
| "learning_rate": 0.0001518097941802697, |
| "loss": 1.1468, |
| "step": 681 |
| }, |
| { |
| "epoch": 0.97, |
| "grad_norm": 0.5782356262207031, |
| "learning_rate": 0.00015173882185947482, |
| "loss": 1.165, |
| "step": 682 |
| }, |
| { |
| "epoch": 0.97, |
| "grad_norm": 0.6173332333564758, |
| "learning_rate": 0.0001516678495386799, |
| "loss": 1.2598, |
| "step": 683 |
| }, |
| { |
| "epoch": 0.97, |
| "grad_norm": 0.5931529998779297, |
| "learning_rate": 0.00015159687721788502, |
| "loss": 1.1117, |
| "step": 684 |
| }, |
| { |
| "epoch": 0.97, |
| "grad_norm": 0.5708393454551697, |
| "learning_rate": 0.00015152590489709013, |
| "loss": 1.2371, |
| "step": 685 |
| }, |
| { |
| "epoch": 0.97, |
| "grad_norm": 0.5764403939247131, |
| "learning_rate": 0.00015145493257629527, |
| "loss": 1.1984, |
| "step": 686 |
| }, |
| { |
| "epoch": 0.97, |
| "grad_norm": 0.5749918818473816, |
| "learning_rate": 0.00015138396025550036, |
| "loss": 1.1347, |
| "step": 687 |
| }, |
| { |
| "epoch": 0.97, |
| "grad_norm": 0.5802225470542908, |
| "learning_rate": 0.00015131298793470547, |
| "loss": 1.2307, |
| "step": 688 |
| }, |
| { |
| "epoch": 0.98, |
| "grad_norm": 0.5560474991798401, |
| "learning_rate": 0.00015124201561391058, |
| "loss": 1.1978, |
| "step": 689 |
| }, |
| { |
| "epoch": 0.98, |
| "grad_norm": 0.5951694250106812, |
| "learning_rate": 0.0001511710432931157, |
| "loss": 1.2, |
| "step": 690 |
| }, |
| { |
| "epoch": 0.98, |
| "grad_norm": 0.5568058490753174, |
| "learning_rate": 0.0001511000709723208, |
| "loss": 1.154, |
| "step": 691 |
| }, |
| { |
| "epoch": 0.98, |
| "grad_norm": 0.5962051153182983, |
| "learning_rate": 0.0001510290986515259, |
| "loss": 1.1111, |
| "step": 692 |
| }, |
| { |
| "epoch": 0.98, |
| "grad_norm": 0.590822160243988, |
| "learning_rate": 0.00015095812633073103, |
| "loss": 1.2304, |
| "step": 693 |
| }, |
| { |
| "epoch": 0.98, |
| "grad_norm": 0.6002994775772095, |
| "learning_rate": 0.00015088715400993615, |
| "loss": 1.2463, |
| "step": 694 |
| }, |
| { |
| "epoch": 0.98, |
| "grad_norm": 0.5498103499412537, |
| "learning_rate": 0.00015081618168914126, |
| "loss": 1.1797, |
| "step": 695 |
| }, |
| { |
| "epoch": 0.99, |
| "grad_norm": 0.5912280678749084, |
| "learning_rate": 0.00015074520936834634, |
| "loss": 1.1483, |
| "step": 696 |
| }, |
| { |
| "epoch": 0.99, |
| "grad_norm": 0.5918639302253723, |
| "learning_rate": 0.00015067423704755146, |
| "loss": 1.2741, |
| "step": 697 |
| }, |
| { |
| "epoch": 0.99, |
| "grad_norm": 0.5741661190986633, |
| "learning_rate": 0.00015060326472675657, |
| "loss": 1.1968, |
| "step": 698 |
| }, |
| { |
| "epoch": 0.99, |
| "grad_norm": 0.5830884575843811, |
| "learning_rate": 0.00015053229240596168, |
| "loss": 1.1663, |
| "step": 699 |
| }, |
| { |
| "epoch": 0.99, |
| "grad_norm": 0.5797367095947266, |
| "learning_rate": 0.0001504613200851668, |
| "loss": 1.1406, |
| "step": 700 |
| }, |
| { |
| "epoch": 0.99, |
| "grad_norm": 0.5564558506011963, |
| "learning_rate": 0.0001503903477643719, |
| "loss": 1.1447, |
| "step": 701 |
| }, |
| { |
| "epoch": 0.99, |
| "grad_norm": 0.5836336612701416, |
| "learning_rate": 0.00015031937544357702, |
| "loss": 1.192, |
| "step": 702 |
| }, |
| { |
| "epoch": 1.0, |
| "grad_norm": 0.6179781556129456, |
| "learning_rate": 0.00015024840312278213, |
| "loss": 1.2181, |
| "step": 703 |
| }, |
| { |
| "epoch": 1.0, |
| "grad_norm": 0.6290362477302551, |
| "learning_rate": 0.00015017743080198722, |
| "loss": 1.234, |
| "step": 704 |
| }, |
| { |
| "epoch": 1.0, |
| "grad_norm": 0.6224949359893799, |
| "learning_rate": 0.00015010645848119233, |
| "loss": 1.2052, |
| "step": 705 |
| }, |
| { |
| "epoch": 1.0, |
| "grad_norm": 0.584284782409668, |
| "learning_rate": 0.00015003548616039744, |
| "loss": 1.198, |
| "step": 706 |
| }, |
| { |
| "epoch": 1.0, |
| "grad_norm": 0.5489341020584106, |
| "learning_rate": 0.00014996451383960258, |
| "loss": 0.9271, |
| "step": 707 |
| }, |
| { |
| "epoch": 1.0, |
| "grad_norm": 0.5513188242912292, |
| "learning_rate": 0.00014989354151880767, |
| "loss": 1.0042, |
| "step": 708 |
| }, |
| { |
| "epoch": 1.0, |
| "grad_norm": 0.5592027306556702, |
| "learning_rate": 0.00014982256919801278, |
| "loss": 0.9466, |
| "step": 709 |
| }, |
| { |
| "epoch": 1.01, |
| "grad_norm": 0.6773933172225952, |
| "learning_rate": 0.0001497515968772179, |
| "loss": 0.9967, |
| "step": 710 |
| }, |
| { |
| "epoch": 1.01, |
| "grad_norm": 0.7364052534103394, |
| "learning_rate": 0.000149680624556423, |
| "loss": 0.9311, |
| "step": 711 |
| }, |
| { |
| "epoch": 1.01, |
| "grad_norm": 0.7685378789901733, |
| "learning_rate": 0.0001496096522356281, |
| "loss": 1.0459, |
| "step": 712 |
| }, |
| { |
| "epoch": 1.01, |
| "grad_norm": 0.6515752673149109, |
| "learning_rate": 0.0001495386799148332, |
| "loss": 0.9924, |
| "step": 713 |
| }, |
| { |
| "epoch": 1.01, |
| "grad_norm": 0.5922237634658813, |
| "learning_rate": 0.00014946770759403835, |
| "loss": 0.9159, |
| "step": 714 |
| }, |
| { |
| "epoch": 1.01, |
| "grad_norm": 0.6064658761024475, |
| "learning_rate": 0.00014939673527324346, |
| "loss": 0.9521, |
| "step": 715 |
| }, |
| { |
| "epoch": 1.01, |
| "grad_norm": 0.6219831109046936, |
| "learning_rate": 0.00014932576295244854, |
| "loss": 1.0295, |
| "step": 716 |
| }, |
| { |
| "epoch": 1.02, |
| "grad_norm": 0.5930375456809998, |
| "learning_rate": 0.00014925479063165366, |
| "loss": 1.0015, |
| "step": 717 |
| }, |
| { |
| "epoch": 1.02, |
| "grad_norm": 0.5987579822540283, |
| "learning_rate": 0.00014918381831085877, |
| "loss": 0.937, |
| "step": 718 |
| }, |
| { |
| "epoch": 1.02, |
| "grad_norm": 0.6563482880592346, |
| "learning_rate": 0.00014911284599006388, |
| "loss": 0.9396, |
| "step": 719 |
| }, |
| { |
| "epoch": 1.02, |
| "grad_norm": 0.6364894509315491, |
| "learning_rate": 0.000149041873669269, |
| "loss": 0.9889, |
| "step": 720 |
| }, |
| { |
| "epoch": 1.02, |
| "grad_norm": 0.6609119772911072, |
| "learning_rate": 0.0001489709013484741, |
| "loss": 0.9447, |
| "step": 721 |
| }, |
| { |
| "epoch": 1.02, |
| "grad_norm": 0.6506882309913635, |
| "learning_rate": 0.00014889992902767922, |
| "loss": 0.957, |
| "step": 722 |
| }, |
| { |
| "epoch": 1.02, |
| "grad_norm": 0.7169485092163086, |
| "learning_rate": 0.00014882895670688433, |
| "loss": 1.0139, |
| "step": 723 |
| }, |
| { |
| "epoch": 1.03, |
| "grad_norm": 0.623009443283081, |
| "learning_rate": 0.00014875798438608945, |
| "loss": 0.9973, |
| "step": 724 |
| }, |
| { |
| "epoch": 1.03, |
| "grad_norm": 0.6281269788742065, |
| "learning_rate": 0.00014868701206529453, |
| "loss": 0.9378, |
| "step": 725 |
| }, |
| { |
| "epoch": 1.03, |
| "grad_norm": 0.6167576313018799, |
| "learning_rate": 0.00014861603974449964, |
| "loss": 1.0196, |
| "step": 726 |
| }, |
| { |
| "epoch": 1.03, |
| "grad_norm": 0.6171091198921204, |
| "learning_rate": 0.00014854506742370476, |
| "loss": 0.9586, |
| "step": 727 |
| }, |
| { |
| "epoch": 1.03, |
| "grad_norm": 0.6009536385536194, |
| "learning_rate": 0.00014847409510290987, |
| "loss": 0.9412, |
| "step": 728 |
| }, |
| { |
| "epoch": 1.03, |
| "grad_norm": 0.6541954278945923, |
| "learning_rate": 0.00014840312278211498, |
| "loss": 0.9836, |
| "step": 729 |
| }, |
| { |
| "epoch": 1.03, |
| "grad_norm": 0.6337497234344482, |
| "learning_rate": 0.0001483321504613201, |
| "loss": 0.9975, |
| "step": 730 |
| }, |
| { |
| "epoch": 1.04, |
| "grad_norm": 0.625019907951355, |
| "learning_rate": 0.0001482611781405252, |
| "loss": 0.9526, |
| "step": 731 |
| }, |
| { |
| "epoch": 1.04, |
| "grad_norm": 0.6225535273551941, |
| "learning_rate": 0.00014819020581973032, |
| "loss": 1.0116, |
| "step": 732 |
| }, |
| { |
| "epoch": 1.04, |
| "grad_norm": 0.6529517769813538, |
| "learning_rate": 0.0001481192334989354, |
| "loss": 1.0537, |
| "step": 733 |
| }, |
| { |
| "epoch": 1.04, |
| "grad_norm": 0.6694484949111938, |
| "learning_rate": 0.00014804826117814052, |
| "loss": 0.9945, |
| "step": 734 |
| }, |
| { |
| "epoch": 1.04, |
| "grad_norm": 0.6434985995292664, |
| "learning_rate": 0.00014797728885734566, |
| "loss": 1.0104, |
| "step": 735 |
| }, |
| { |
| "epoch": 1.04, |
| "grad_norm": 0.643535852432251, |
| "learning_rate": 0.00014790631653655077, |
| "loss": 0.9848, |
| "step": 736 |
| }, |
| { |
| "epoch": 1.04, |
| "grad_norm": 0.6169139742851257, |
| "learning_rate": 0.00014783534421575586, |
| "loss": 0.8907, |
| "step": 737 |
| }, |
| { |
| "epoch": 1.05, |
| "grad_norm": 0.6531071662902832, |
| "learning_rate": 0.00014776437189496097, |
| "loss": 1.0544, |
| "step": 738 |
| }, |
| { |
| "epoch": 1.05, |
| "grad_norm": 0.6137780547142029, |
| "learning_rate": 0.00014769339957416608, |
| "loss": 0.919, |
| "step": 739 |
| }, |
| { |
| "epoch": 1.05, |
| "grad_norm": 0.685459554195404, |
| "learning_rate": 0.0001476224272533712, |
| "loss": 1.0618, |
| "step": 740 |
| }, |
| { |
| "epoch": 1.05, |
| "grad_norm": 0.6708681583404541, |
| "learning_rate": 0.00014755145493257628, |
| "loss": 0.9946, |
| "step": 741 |
| }, |
| { |
| "epoch": 1.05, |
| "grad_norm": 0.6316698789596558, |
| "learning_rate": 0.00014748048261178142, |
| "loss": 0.9819, |
| "step": 742 |
| }, |
| { |
| "epoch": 1.05, |
| "grad_norm": 0.622658371925354, |
| "learning_rate": 0.00014740951029098653, |
| "loss": 0.9845, |
| "step": 743 |
| }, |
| { |
| "epoch": 1.05, |
| "grad_norm": 0.6192933917045593, |
| "learning_rate": 0.00014733853797019165, |
| "loss": 0.9678, |
| "step": 744 |
| }, |
| { |
| "epoch": 1.06, |
| "grad_norm": 0.651993453502655, |
| "learning_rate": 0.00014726756564939673, |
| "loss": 0.9554, |
| "step": 745 |
| }, |
| { |
| "epoch": 1.06, |
| "grad_norm": 0.6194689869880676, |
| "learning_rate": 0.00014719659332860184, |
| "loss": 0.9167, |
| "step": 746 |
| }, |
| { |
| "epoch": 1.06, |
| "grad_norm": 0.6849738955497742, |
| "learning_rate": 0.00014712562100780696, |
| "loss": 0.9528, |
| "step": 747 |
| }, |
| { |
| "epoch": 1.06, |
| "grad_norm": 0.6835150122642517, |
| "learning_rate": 0.00014705464868701207, |
| "loss": 0.9799, |
| "step": 748 |
| }, |
| { |
| "epoch": 1.06, |
| "grad_norm": 0.6851332783699036, |
| "learning_rate": 0.00014698367636621718, |
| "loss": 1.0472, |
| "step": 749 |
| }, |
| { |
| "epoch": 1.06, |
| "grad_norm": 0.6524817943572998, |
| "learning_rate": 0.0001469127040454223, |
| "loss": 0.9258, |
| "step": 750 |
| }, |
| { |
| "epoch": 1.06, |
| "grad_norm": 0.6759579181671143, |
| "learning_rate": 0.0001468417317246274, |
| "loss": 0.9545, |
| "step": 751 |
| }, |
| { |
| "epoch": 1.07, |
| "grad_norm": 0.6385304927825928, |
| "learning_rate": 0.00014677075940383252, |
| "loss": 0.9751, |
| "step": 752 |
| }, |
| { |
| "epoch": 1.07, |
| "grad_norm": 0.6409848928451538, |
| "learning_rate": 0.00014669978708303763, |
| "loss": 0.9349, |
| "step": 753 |
| }, |
| { |
| "epoch": 1.07, |
| "grad_norm": 0.6459067463874817, |
| "learning_rate": 0.00014662881476224272, |
| "loss": 1.0341, |
| "step": 754 |
| }, |
| { |
| "epoch": 1.07, |
| "grad_norm": 0.6660894155502319, |
| "learning_rate": 0.00014655784244144783, |
| "loss": 1.0151, |
| "step": 755 |
| }, |
| { |
| "epoch": 1.07, |
| "grad_norm": 0.6208553910255432, |
| "learning_rate": 0.00014648687012065294, |
| "loss": 0.9547, |
| "step": 756 |
| }, |
| { |
| "epoch": 1.07, |
| "grad_norm": 0.6216368675231934, |
| "learning_rate": 0.00014641589779985806, |
| "loss": 1.0178, |
| "step": 757 |
| }, |
| { |
| "epoch": 1.07, |
| "grad_norm": 0.6845688819885254, |
| "learning_rate": 0.00014634492547906317, |
| "loss": 0.9976, |
| "step": 758 |
| }, |
| { |
| "epoch": 1.08, |
| "grad_norm": 0.6555722951889038, |
| "learning_rate": 0.00014627395315826828, |
| "loss": 0.9684, |
| "step": 759 |
| }, |
| { |
| "epoch": 1.08, |
| "grad_norm": 0.6671058535575867, |
| "learning_rate": 0.0001462029808374734, |
| "loss": 0.9691, |
| "step": 760 |
| }, |
| { |
| "epoch": 1.08, |
| "grad_norm": 0.6752671003341675, |
| "learning_rate": 0.0001461320085166785, |
| "loss": 0.9492, |
| "step": 761 |
| }, |
| { |
| "epoch": 1.08, |
| "grad_norm": 0.7133360505104065, |
| "learning_rate": 0.0001460610361958836, |
| "loss": 0.9867, |
| "step": 762 |
| }, |
| { |
| "epoch": 1.08, |
| "grad_norm": 0.6610892415046692, |
| "learning_rate": 0.0001459900638750887, |
| "loss": 0.9809, |
| "step": 763 |
| }, |
| { |
| "epoch": 1.08, |
| "grad_norm": 0.6392022371292114, |
| "learning_rate": 0.00014591909155429385, |
| "loss": 0.9485, |
| "step": 764 |
| }, |
| { |
| "epoch": 1.08, |
| "grad_norm": 0.6533021330833435, |
| "learning_rate": 0.00014584811923349896, |
| "loss": 1.0081, |
| "step": 765 |
| }, |
| { |
| "epoch": 1.09, |
| "grad_norm": 0.6720760464668274, |
| "learning_rate": 0.00014577714691270404, |
| "loss": 0.9421, |
| "step": 766 |
| }, |
| { |
| "epoch": 1.09, |
| "grad_norm": 0.6587719321250916, |
| "learning_rate": 0.00014570617459190916, |
| "loss": 0.953, |
| "step": 767 |
| }, |
| { |
| "epoch": 1.09, |
| "grad_norm": 0.6704784631729126, |
| "learning_rate": 0.00014563520227111427, |
| "loss": 0.9633, |
| "step": 768 |
| }, |
| { |
| "epoch": 1.09, |
| "grad_norm": 0.6440714597702026, |
| "learning_rate": 0.00014556422995031938, |
| "loss": 0.9655, |
| "step": 769 |
| }, |
| { |
| "epoch": 1.09, |
| "grad_norm": 0.6508157849311829, |
| "learning_rate": 0.00014549325762952447, |
| "loss": 0.9552, |
| "step": 770 |
| }, |
| { |
| "epoch": 1.09, |
| "grad_norm": 0.6560282707214355, |
| "learning_rate": 0.0001454222853087296, |
| "loss": 1.0196, |
| "step": 771 |
| }, |
| { |
| "epoch": 1.09, |
| "grad_norm": 0.6468494534492493, |
| "learning_rate": 0.00014535131298793472, |
| "loss": 0.9734, |
| "step": 772 |
| }, |
| { |
| "epoch": 1.1, |
| "grad_norm": 0.6644648313522339, |
| "learning_rate": 0.00014528034066713983, |
| "loss": 1.011, |
| "step": 773 |
| }, |
| { |
| "epoch": 1.1, |
| "grad_norm": 0.6273561120033264, |
| "learning_rate": 0.00014520936834634492, |
| "loss": 0.9456, |
| "step": 774 |
| }, |
| { |
| "epoch": 1.1, |
| "grad_norm": 0.6624634265899658, |
| "learning_rate": 0.00014513839602555003, |
| "loss": 1.0313, |
| "step": 775 |
| }, |
| { |
| "epoch": 1.1, |
| "grad_norm": 0.6488678455352783, |
| "learning_rate": 0.00014506742370475515, |
| "loss": 0.9081, |
| "step": 776 |
| }, |
| { |
| "epoch": 1.1, |
| "grad_norm": 0.6563031077384949, |
| "learning_rate": 0.00014499645138396026, |
| "loss": 0.9934, |
| "step": 777 |
| }, |
| { |
| "epoch": 1.1, |
| "grad_norm": 0.6557429432868958, |
| "learning_rate": 0.00014492547906316537, |
| "loss": 0.9944, |
| "step": 778 |
| }, |
| { |
| "epoch": 1.1, |
| "grad_norm": 0.6796723008155823, |
| "learning_rate": 0.00014485450674237048, |
| "loss": 0.9943, |
| "step": 779 |
| }, |
| { |
| "epoch": 1.11, |
| "grad_norm": 0.6730267405509949, |
| "learning_rate": 0.0001447835344215756, |
| "loss": 0.974, |
| "step": 780 |
| }, |
| { |
| "epoch": 1.11, |
| "grad_norm": 0.670569658279419, |
| "learning_rate": 0.0001447125621007807, |
| "loss": 0.9909, |
| "step": 781 |
| }, |
| { |
| "epoch": 1.11, |
| "grad_norm": 0.6504651308059692, |
| "learning_rate": 0.00014464158977998582, |
| "loss": 1.0146, |
| "step": 782 |
| }, |
| { |
| "epoch": 1.11, |
| "grad_norm": 0.6637928485870361, |
| "learning_rate": 0.0001445706174591909, |
| "loss": 0.9802, |
| "step": 783 |
| }, |
| { |
| "epoch": 1.11, |
| "grad_norm": 0.6541115045547485, |
| "learning_rate": 0.00014449964513839602, |
| "loss": 1.017, |
| "step": 784 |
| }, |
| { |
| "epoch": 1.11, |
| "grad_norm": 0.6643988490104675, |
| "learning_rate": 0.00014442867281760116, |
| "loss": 0.9983, |
| "step": 785 |
| }, |
| { |
| "epoch": 1.11, |
| "grad_norm": 0.6713914275169373, |
| "learning_rate": 0.00014435770049680625, |
| "loss": 0.9617, |
| "step": 786 |
| }, |
| { |
| "epoch": 1.11, |
| "grad_norm": 0.6338568329811096, |
| "learning_rate": 0.00014428672817601136, |
| "loss": 0.8769, |
| "step": 787 |
| }, |
| { |
| "epoch": 1.12, |
| "grad_norm": 0.6811724305152893, |
| "learning_rate": 0.00014421575585521647, |
| "loss": 1.0026, |
| "step": 788 |
| }, |
| { |
| "epoch": 1.12, |
| "grad_norm": 0.7284192442893982, |
| "learning_rate": 0.00014414478353442158, |
| "loss": 0.9643, |
| "step": 789 |
| }, |
| { |
| "epoch": 1.12, |
| "grad_norm": 0.6587553024291992, |
| "learning_rate": 0.0001440738112136267, |
| "loss": 0.9988, |
| "step": 790 |
| }, |
| { |
| "epoch": 1.12, |
| "grad_norm": 0.6484501957893372, |
| "learning_rate": 0.00014400283889283178, |
| "loss": 0.9761, |
| "step": 791 |
| }, |
| { |
| "epoch": 1.12, |
| "grad_norm": 0.6753117442131042, |
| "learning_rate": 0.00014393186657203692, |
| "loss": 1.0066, |
| "step": 792 |
| }, |
| { |
| "epoch": 1.12, |
| "grad_norm": 0.6668440699577332, |
| "learning_rate": 0.00014386089425124203, |
| "loss": 0.9439, |
| "step": 793 |
| }, |
| { |
| "epoch": 1.12, |
| "grad_norm": 0.6919518709182739, |
| "learning_rate": 0.00014378992193044715, |
| "loss": 0.9829, |
| "step": 794 |
| }, |
| { |
| "epoch": 1.13, |
| "grad_norm": 0.6882046461105347, |
| "learning_rate": 0.00014371894960965223, |
| "loss": 1.0086, |
| "step": 795 |
| }, |
| { |
| "epoch": 1.13, |
| "grad_norm": 0.641126811504364, |
| "learning_rate": 0.00014364797728885735, |
| "loss": 0.9792, |
| "step": 796 |
| }, |
| { |
| "epoch": 1.13, |
| "grad_norm": 0.6557233333587646, |
| "learning_rate": 0.00014357700496806246, |
| "loss": 0.8754, |
| "step": 797 |
| }, |
| { |
| "epoch": 1.13, |
| "grad_norm": 0.6546401381492615, |
| "learning_rate": 0.00014350603264726757, |
| "loss": 0.9407, |
| "step": 798 |
| }, |
| { |
| "epoch": 1.13, |
| "grad_norm": 0.6610774397850037, |
| "learning_rate": 0.00014343506032647268, |
| "loss": 1.0165, |
| "step": 799 |
| }, |
| { |
| "epoch": 1.13, |
| "grad_norm": 0.692677915096283, |
| "learning_rate": 0.0001433640880056778, |
| "loss": 0.931, |
| "step": 800 |
| }, |
| { |
| "epoch": 1.13, |
| "grad_norm": 0.6578258872032166, |
| "learning_rate": 0.0001432931156848829, |
| "loss": 0.9976, |
| "step": 801 |
| }, |
| { |
| "epoch": 1.14, |
| "grad_norm": 0.6596624851226807, |
| "learning_rate": 0.00014322214336408802, |
| "loss": 0.9312, |
| "step": 802 |
| }, |
| { |
| "epoch": 1.14, |
| "grad_norm": 0.6386009454727173, |
| "learning_rate": 0.0001431511710432931, |
| "loss": 0.992, |
| "step": 803 |
| }, |
| { |
| "epoch": 1.14, |
| "grad_norm": 0.6921107769012451, |
| "learning_rate": 0.00014308019872249822, |
| "loss": 0.9981, |
| "step": 804 |
| }, |
| { |
| "epoch": 1.14, |
| "grad_norm": 0.6416789293289185, |
| "learning_rate": 0.00014300922640170333, |
| "loss": 1.0192, |
| "step": 805 |
| }, |
| { |
| "epoch": 1.14, |
| "grad_norm": 0.6847473382949829, |
| "learning_rate": 0.00014293825408090847, |
| "loss": 0.945, |
| "step": 806 |
| }, |
| { |
| "epoch": 1.14, |
| "grad_norm": 0.6717972755432129, |
| "learning_rate": 0.00014286728176011356, |
| "loss": 0.9767, |
| "step": 807 |
| }, |
| { |
| "epoch": 1.14, |
| "grad_norm": 0.6813983917236328, |
| "learning_rate": 0.00014279630943931867, |
| "loss": 0.9056, |
| "step": 808 |
| }, |
| { |
| "epoch": 1.15, |
| "grad_norm": 0.6662610769271851, |
| "learning_rate": 0.00014272533711852378, |
| "loss": 1.1045, |
| "step": 809 |
| }, |
| { |
| "epoch": 1.15, |
| "grad_norm": 0.6779379844665527, |
| "learning_rate": 0.0001426543647977289, |
| "loss": 0.9158, |
| "step": 810 |
| }, |
| { |
| "epoch": 1.15, |
| "grad_norm": 0.6368980407714844, |
| "learning_rate": 0.000142583392476934, |
| "loss": 0.8638, |
| "step": 811 |
| }, |
| { |
| "epoch": 1.15, |
| "grad_norm": 0.648033857345581, |
| "learning_rate": 0.0001425124201561391, |
| "loss": 1.0107, |
| "step": 812 |
| }, |
| { |
| "epoch": 1.15, |
| "grad_norm": 0.6696658134460449, |
| "learning_rate": 0.00014244144783534423, |
| "loss": 0.9926, |
| "step": 813 |
| }, |
| { |
| "epoch": 1.15, |
| "grad_norm": 0.6661670804023743, |
| "learning_rate": 0.00014237047551454935, |
| "loss": 0.9805, |
| "step": 814 |
| }, |
| { |
| "epoch": 1.15, |
| "grad_norm": 0.6772279739379883, |
| "learning_rate": 0.00014229950319375443, |
| "loss": 0.9443, |
| "step": 815 |
| }, |
| { |
| "epoch": 1.16, |
| "grad_norm": 0.7234826683998108, |
| "learning_rate": 0.00014222853087295955, |
| "loss": 1.0605, |
| "step": 816 |
| }, |
| { |
| "epoch": 1.16, |
| "grad_norm": 0.6951813101768494, |
| "learning_rate": 0.00014215755855216466, |
| "loss": 1.0043, |
| "step": 817 |
| }, |
| { |
| "epoch": 1.16, |
| "grad_norm": 0.6815015077590942, |
| "learning_rate": 0.00014208658623136977, |
| "loss": 0.9496, |
| "step": 818 |
| }, |
| { |
| "epoch": 1.16, |
| "grad_norm": 0.6754716634750366, |
| "learning_rate": 0.00014201561391057488, |
| "loss": 0.954, |
| "step": 819 |
| }, |
| { |
| "epoch": 1.16, |
| "grad_norm": 0.6836742758750916, |
| "learning_rate": 0.00014194464158978, |
| "loss": 0.9787, |
| "step": 820 |
| }, |
| { |
| "epoch": 1.16, |
| "grad_norm": 0.6898475885391235, |
| "learning_rate": 0.0001418736692689851, |
| "loss": 1.0429, |
| "step": 821 |
| }, |
| { |
| "epoch": 1.16, |
| "grad_norm": 0.6403341889381409, |
| "learning_rate": 0.00014180269694819022, |
| "loss": 0.9681, |
| "step": 822 |
| }, |
| { |
| "epoch": 1.17, |
| "grad_norm": 0.6508347392082214, |
| "learning_rate": 0.00014173172462739533, |
| "loss": 1.0251, |
| "step": 823 |
| }, |
| { |
| "epoch": 1.17, |
| "grad_norm": 0.6511325836181641, |
| "learning_rate": 0.00014166075230660042, |
| "loss": 0.934, |
| "step": 824 |
| }, |
| { |
| "epoch": 1.17, |
| "grad_norm": 0.6893059611320496, |
| "learning_rate": 0.00014158977998580553, |
| "loss": 0.9481, |
| "step": 825 |
| }, |
| { |
| "epoch": 1.17, |
| "grad_norm": 0.7065815329551697, |
| "learning_rate": 0.00014151880766501065, |
| "loss": 0.9531, |
| "step": 826 |
| }, |
| { |
| "epoch": 1.17, |
| "grad_norm": 0.6699268221855164, |
| "learning_rate": 0.00014144783534421579, |
| "loss": 0.9593, |
| "step": 827 |
| }, |
| { |
| "epoch": 1.17, |
| "grad_norm": 0.691835343837738, |
| "learning_rate": 0.00014137686302342087, |
| "loss": 1.0152, |
| "step": 828 |
| }, |
| { |
| "epoch": 1.17, |
| "grad_norm": 0.6948768496513367, |
| "learning_rate": 0.00014130589070262598, |
| "loss": 0.9809, |
| "step": 829 |
| }, |
| { |
| "epoch": 1.18, |
| "grad_norm": 0.685308039188385, |
| "learning_rate": 0.0001412349183818311, |
| "loss": 0.9709, |
| "step": 830 |
| }, |
| { |
| "epoch": 1.18, |
| "grad_norm": 0.7141802310943604, |
| "learning_rate": 0.0001411639460610362, |
| "loss": 1.0286, |
| "step": 831 |
| }, |
| { |
| "epoch": 1.18, |
| "grad_norm": 0.6938399076461792, |
| "learning_rate": 0.0001410929737402413, |
| "loss": 1.0431, |
| "step": 832 |
| }, |
| { |
| "epoch": 1.18, |
| "grad_norm": 0.6528292894363403, |
| "learning_rate": 0.0001410220014194464, |
| "loss": 0.9173, |
| "step": 833 |
| }, |
| { |
| "epoch": 1.18, |
| "grad_norm": 0.675507664680481, |
| "learning_rate": 0.00014095102909865155, |
| "loss": 0.958, |
| "step": 834 |
| }, |
| { |
| "epoch": 1.18, |
| "grad_norm": 0.668474555015564, |
| "learning_rate": 0.00014088005677785666, |
| "loss": 1.0094, |
| "step": 835 |
| }, |
| { |
| "epoch": 1.18, |
| "grad_norm": 0.6946351528167725, |
| "learning_rate": 0.00014080908445706175, |
| "loss": 0.986, |
| "step": 836 |
| }, |
| { |
| "epoch": 1.19, |
| "grad_norm": 0.701863706111908, |
| "learning_rate": 0.00014073811213626686, |
| "loss": 0.9987, |
| "step": 837 |
| }, |
| { |
| "epoch": 1.19, |
| "grad_norm": 0.722443163394928, |
| "learning_rate": 0.00014066713981547197, |
| "loss": 0.9995, |
| "step": 838 |
| }, |
| { |
| "epoch": 1.19, |
| "grad_norm": 0.675977885723114, |
| "learning_rate": 0.00014059616749467708, |
| "loss": 0.9929, |
| "step": 839 |
| }, |
| { |
| "epoch": 1.19, |
| "grad_norm": 0.7431498765945435, |
| "learning_rate": 0.0001405251951738822, |
| "loss": 1.0413, |
| "step": 840 |
| }, |
| { |
| "epoch": 1.19, |
| "grad_norm": 0.6718546152114868, |
| "learning_rate": 0.00014045422285308728, |
| "loss": 0.9265, |
| "step": 841 |
| }, |
| { |
| "epoch": 1.19, |
| "grad_norm": 0.6670490503311157, |
| "learning_rate": 0.00014038325053229242, |
| "loss": 0.9314, |
| "step": 842 |
| }, |
| { |
| "epoch": 1.19, |
| "grad_norm": 0.6900096535682678, |
| "learning_rate": 0.00014031227821149754, |
| "loss": 1.0211, |
| "step": 843 |
| }, |
| { |
| "epoch": 1.2, |
| "grad_norm": 0.6813675761222839, |
| "learning_rate": 0.00014024130589070265, |
| "loss": 1.0462, |
| "step": 844 |
| }, |
| { |
| "epoch": 1.2, |
| "grad_norm": 0.6931719183921814, |
| "learning_rate": 0.00014017033356990773, |
| "loss": 0.9741, |
| "step": 845 |
| }, |
| { |
| "epoch": 1.2, |
| "grad_norm": 0.683791995048523, |
| "learning_rate": 0.00014009936124911285, |
| "loss": 1.054, |
| "step": 846 |
| }, |
| { |
| "epoch": 1.2, |
| "grad_norm": 0.7232967615127563, |
| "learning_rate": 0.00014002838892831796, |
| "loss": 0.9365, |
| "step": 847 |
| }, |
| { |
| "epoch": 1.2, |
| "grad_norm": 0.6782677173614502, |
| "learning_rate": 0.00013995741660752307, |
| "loss": 0.9311, |
| "step": 848 |
| }, |
| { |
| "epoch": 1.2, |
| "grad_norm": 0.6695821285247803, |
| "learning_rate": 0.00013988644428672818, |
| "loss": 0.993, |
| "step": 849 |
| }, |
| { |
| "epoch": 1.2, |
| "grad_norm": 0.6955466866493225, |
| "learning_rate": 0.0001398154719659333, |
| "loss": 0.985, |
| "step": 850 |
| }, |
| { |
| "epoch": 1.21, |
| "grad_norm": 0.692164421081543, |
| "learning_rate": 0.0001397444996451384, |
| "loss": 0.9836, |
| "step": 851 |
| }, |
| { |
| "epoch": 1.21, |
| "grad_norm": 0.6746982932090759, |
| "learning_rate": 0.00013967352732434352, |
| "loss": 0.9096, |
| "step": 852 |
| }, |
| { |
| "epoch": 1.21, |
| "grad_norm": 0.6445276141166687, |
| "learning_rate": 0.0001396025550035486, |
| "loss": 1.0032, |
| "step": 853 |
| }, |
| { |
| "epoch": 1.21, |
| "grad_norm": 0.7133185863494873, |
| "learning_rate": 0.00013953158268275372, |
| "loss": 1.009, |
| "step": 854 |
| }, |
| { |
| "epoch": 1.21, |
| "grad_norm": 0.6872855424880981, |
| "learning_rate": 0.00013946061036195883, |
| "loss": 1.0236, |
| "step": 855 |
| }, |
| { |
| "epoch": 1.21, |
| "grad_norm": 0.7020975351333618, |
| "learning_rate": 0.00013938963804116397, |
| "loss": 1.0142, |
| "step": 856 |
| }, |
| { |
| "epoch": 1.21, |
| "grad_norm": 0.663386881351471, |
| "learning_rate": 0.00013931866572036906, |
| "loss": 0.9976, |
| "step": 857 |
| }, |
| { |
| "epoch": 1.22, |
| "grad_norm": 0.7051934599876404, |
| "learning_rate": 0.00013924769339957417, |
| "loss": 1.0646, |
| "step": 858 |
| }, |
| { |
| "epoch": 1.22, |
| "grad_norm": 0.6661969423294067, |
| "learning_rate": 0.00013917672107877928, |
| "loss": 0.9287, |
| "step": 859 |
| }, |
| { |
| "epoch": 1.22, |
| "grad_norm": 0.6907512545585632, |
| "learning_rate": 0.0001391057487579844, |
| "loss": 1.0021, |
| "step": 860 |
| }, |
| { |
| "epoch": 1.22, |
| "grad_norm": 0.66982501745224, |
| "learning_rate": 0.00013903477643718948, |
| "loss": 0.9768, |
| "step": 861 |
| }, |
| { |
| "epoch": 1.22, |
| "grad_norm": 0.6967639923095703, |
| "learning_rate": 0.0001389638041163946, |
| "loss": 0.994, |
| "step": 862 |
| }, |
| { |
| "epoch": 1.22, |
| "grad_norm": 0.7178871035575867, |
| "learning_rate": 0.00013889283179559974, |
| "loss": 0.9517, |
| "step": 863 |
| }, |
| { |
| "epoch": 1.22, |
| "grad_norm": 0.6952171921730042, |
| "learning_rate": 0.00013882185947480485, |
| "loss": 1.0229, |
| "step": 864 |
| }, |
| { |
| "epoch": 1.23, |
| "grad_norm": 0.6783744096755981, |
| "learning_rate": 0.00013875088715400993, |
| "loss": 0.9575, |
| "step": 865 |
| }, |
| { |
| "epoch": 1.23, |
| "grad_norm": 0.6978974938392639, |
| "learning_rate": 0.00013867991483321505, |
| "loss": 0.9912, |
| "step": 866 |
| }, |
| { |
| "epoch": 1.23, |
| "grad_norm": 0.6688209772109985, |
| "learning_rate": 0.00013860894251242016, |
| "loss": 0.9916, |
| "step": 867 |
| }, |
| { |
| "epoch": 1.23, |
| "grad_norm": 0.6768879890441895, |
| "learning_rate": 0.00013853797019162527, |
| "loss": 1.0178, |
| "step": 868 |
| }, |
| { |
| "epoch": 1.23, |
| "grad_norm": 0.6849448084831238, |
| "learning_rate": 0.00013846699787083038, |
| "loss": 1.003, |
| "step": 869 |
| }, |
| { |
| "epoch": 1.23, |
| "grad_norm": 0.6396516561508179, |
| "learning_rate": 0.0001383960255500355, |
| "loss": 0.8896, |
| "step": 870 |
| }, |
| { |
| "epoch": 1.23, |
| "grad_norm": 0.6769018769264221, |
| "learning_rate": 0.0001383250532292406, |
| "loss": 1.0025, |
| "step": 871 |
| }, |
| { |
| "epoch": 1.24, |
| "grad_norm": 0.7078860998153687, |
| "learning_rate": 0.00013825408090844572, |
| "loss": 0.9968, |
| "step": 872 |
| }, |
| { |
| "epoch": 1.24, |
| "grad_norm": 0.6808391213417053, |
| "learning_rate": 0.00013818310858765084, |
| "loss": 0.9811, |
| "step": 873 |
| }, |
| { |
| "epoch": 1.24, |
| "grad_norm": 0.7080936431884766, |
| "learning_rate": 0.00013811213626685592, |
| "loss": 0.9996, |
| "step": 874 |
| }, |
| { |
| "epoch": 1.24, |
| "grad_norm": 0.7115461230278015, |
| "learning_rate": 0.00013804116394606103, |
| "loss": 0.9898, |
| "step": 875 |
| }, |
| { |
| "epoch": 1.24, |
| "grad_norm": 0.6735612154006958, |
| "learning_rate": 0.00013797019162526615, |
| "loss": 0.9852, |
| "step": 876 |
| }, |
| { |
| "epoch": 1.24, |
| "grad_norm": 0.7107511162757874, |
| "learning_rate": 0.00013789921930447126, |
| "loss": 1.0363, |
| "step": 877 |
| }, |
| { |
| "epoch": 1.24, |
| "grad_norm": 0.7112343311309814, |
| "learning_rate": 0.00013782824698367637, |
| "loss": 1.0529, |
| "step": 878 |
| }, |
| { |
| "epoch": 1.25, |
| "grad_norm": 0.6874762773513794, |
| "learning_rate": 0.00013775727466288148, |
| "loss": 0.9774, |
| "step": 879 |
| }, |
| { |
| "epoch": 1.25, |
| "grad_norm": 0.6808403134346008, |
| "learning_rate": 0.0001376863023420866, |
| "loss": 1.0055, |
| "step": 880 |
| }, |
| { |
| "epoch": 1.25, |
| "grad_norm": 0.700853705406189, |
| "learning_rate": 0.0001376153300212917, |
| "loss": 1.0177, |
| "step": 881 |
| }, |
| { |
| "epoch": 1.25, |
| "grad_norm": 0.7160459160804749, |
| "learning_rate": 0.0001375443577004968, |
| "loss": 0.9944, |
| "step": 882 |
| }, |
| { |
| "epoch": 1.25, |
| "grad_norm": 0.7002721428871155, |
| "learning_rate": 0.0001374733853797019, |
| "loss": 0.9271, |
| "step": 883 |
| }, |
| { |
| "epoch": 1.25, |
| "grad_norm": 0.7237516045570374, |
| "learning_rate": 0.00013740241305890705, |
| "loss": 0.9932, |
| "step": 884 |
| }, |
| { |
| "epoch": 1.25, |
| "grad_norm": 0.7104388475418091, |
| "learning_rate": 0.00013733144073811216, |
| "loss": 1.044, |
| "step": 885 |
| }, |
| { |
| "epoch": 1.26, |
| "grad_norm": 0.6789724826812744, |
| "learning_rate": 0.00013726046841731725, |
| "loss": 0.9781, |
| "step": 886 |
| }, |
| { |
| "epoch": 1.26, |
| "grad_norm": 0.6558204293251038, |
| "learning_rate": 0.00013718949609652236, |
| "loss": 0.9708, |
| "step": 887 |
| }, |
| { |
| "epoch": 1.26, |
| "grad_norm": 0.6882026791572571, |
| "learning_rate": 0.00013711852377572747, |
| "loss": 0.9466, |
| "step": 888 |
| }, |
| { |
| "epoch": 1.26, |
| "grad_norm": 0.6949423551559448, |
| "learning_rate": 0.00013704755145493258, |
| "loss": 1.0056, |
| "step": 889 |
| }, |
| { |
| "epoch": 1.26, |
| "grad_norm": 0.7023829221725464, |
| "learning_rate": 0.00013697657913413767, |
| "loss": 0.982, |
| "step": 890 |
| }, |
| { |
| "epoch": 1.26, |
| "grad_norm": 0.7513879537582397, |
| "learning_rate": 0.0001369056068133428, |
| "loss": 0.987, |
| "step": 891 |
| }, |
| { |
| "epoch": 1.26, |
| "grad_norm": 0.6762917041778564, |
| "learning_rate": 0.00013683463449254792, |
| "loss": 0.9202, |
| "step": 892 |
| }, |
| { |
| "epoch": 1.27, |
| "grad_norm": 0.7018841505050659, |
| "learning_rate": 0.00013676366217175304, |
| "loss": 1.0525, |
| "step": 893 |
| }, |
| { |
| "epoch": 1.27, |
| "grad_norm": 0.6917586922645569, |
| "learning_rate": 0.00013669268985095812, |
| "loss": 0.9392, |
| "step": 894 |
| }, |
| { |
| "epoch": 1.27, |
| "grad_norm": 0.7120482325553894, |
| "learning_rate": 0.00013662171753016323, |
| "loss": 1.0218, |
| "step": 895 |
| }, |
| { |
| "epoch": 1.27, |
| "grad_norm": 0.6614890098571777, |
| "learning_rate": 0.00013655074520936835, |
| "loss": 0.9586, |
| "step": 896 |
| }, |
| { |
| "epoch": 1.27, |
| "grad_norm": 0.713033139705658, |
| "learning_rate": 0.00013647977288857346, |
| "loss": 0.9867, |
| "step": 897 |
| }, |
| { |
| "epoch": 1.27, |
| "grad_norm": 0.6614014506340027, |
| "learning_rate": 0.00013640880056777857, |
| "loss": 0.9747, |
| "step": 898 |
| }, |
| { |
| "epoch": 1.27, |
| "grad_norm": 0.6631972193717957, |
| "learning_rate": 0.00013633782824698369, |
| "loss": 1.0244, |
| "step": 899 |
| }, |
| { |
| "epoch": 1.28, |
| "grad_norm": 0.6855646967887878, |
| "learning_rate": 0.0001362668559261888, |
| "loss": 1.0053, |
| "step": 900 |
| }, |
| { |
| "epoch": 1.28, |
| "grad_norm": 0.6771067976951599, |
| "learning_rate": 0.0001361958836053939, |
| "loss": 0.956, |
| "step": 901 |
| }, |
| { |
| "epoch": 1.28, |
| "grad_norm": 0.7038958668708801, |
| "learning_rate": 0.00013612491128459902, |
| "loss": 1.0221, |
| "step": 902 |
| }, |
| { |
| "epoch": 1.28, |
| "grad_norm": 0.7114011645317078, |
| "learning_rate": 0.0001360539389638041, |
| "loss": 1.0298, |
| "step": 903 |
| }, |
| { |
| "epoch": 1.28, |
| "grad_norm": 0.707399308681488, |
| "learning_rate": 0.00013598296664300922, |
| "loss": 0.9623, |
| "step": 904 |
| }, |
| { |
| "epoch": 1.28, |
| "grad_norm": 0.7122653722763062, |
| "learning_rate": 0.00013591199432221436, |
| "loss": 1.0042, |
| "step": 905 |
| }, |
| { |
| "epoch": 1.28, |
| "grad_norm": 0.699234664440155, |
| "learning_rate": 0.00013584102200141945, |
| "loss": 0.9806, |
| "step": 906 |
| }, |
| { |
| "epoch": 1.28, |
| "grad_norm": 0.7016271352767944, |
| "learning_rate": 0.00013577004968062456, |
| "loss": 0.9817, |
| "step": 907 |
| }, |
| { |
| "epoch": 1.29, |
| "grad_norm": 0.7013638019561768, |
| "learning_rate": 0.00013569907735982967, |
| "loss": 1.0045, |
| "step": 908 |
| }, |
| { |
| "epoch": 1.29, |
| "grad_norm": 0.7025014162063599, |
| "learning_rate": 0.00013562810503903479, |
| "loss": 1.0147, |
| "step": 909 |
| }, |
| { |
| "epoch": 1.29, |
| "grad_norm": 0.718924880027771, |
| "learning_rate": 0.0001355571327182399, |
| "loss": 1.0292, |
| "step": 910 |
| }, |
| { |
| "epoch": 1.29, |
| "grad_norm": 0.6805405020713806, |
| "learning_rate": 0.00013548616039744498, |
| "loss": 0.9774, |
| "step": 911 |
| }, |
| { |
| "epoch": 1.29, |
| "grad_norm": 0.6721410155296326, |
| "learning_rate": 0.00013541518807665012, |
| "loss": 0.9475, |
| "step": 912 |
| }, |
| { |
| "epoch": 1.29, |
| "grad_norm": 0.6628838181495667, |
| "learning_rate": 0.00013534421575585524, |
| "loss": 0.9408, |
| "step": 913 |
| }, |
| { |
| "epoch": 1.29, |
| "grad_norm": 0.7161340713500977, |
| "learning_rate": 0.00013527324343506035, |
| "loss": 0.9819, |
| "step": 914 |
| }, |
| { |
| "epoch": 1.3, |
| "grad_norm": 0.7012698650360107, |
| "learning_rate": 0.00013520227111426543, |
| "loss": 1.0074, |
| "step": 915 |
| }, |
| { |
| "epoch": 1.3, |
| "grad_norm": 0.6986689567565918, |
| "learning_rate": 0.00013513129879347055, |
| "loss": 0.9736, |
| "step": 916 |
| }, |
| { |
| "epoch": 1.3, |
| "grad_norm": 0.7003822922706604, |
| "learning_rate": 0.00013506032647267566, |
| "loss": 1.0369, |
| "step": 917 |
| }, |
| { |
| "epoch": 1.3, |
| "grad_norm": 0.723176896572113, |
| "learning_rate": 0.00013498935415188077, |
| "loss": 0.9936, |
| "step": 918 |
| }, |
| { |
| "epoch": 1.3, |
| "grad_norm": 0.7104766368865967, |
| "learning_rate": 0.00013491838183108589, |
| "loss": 0.9297, |
| "step": 919 |
| }, |
| { |
| "epoch": 1.3, |
| "grad_norm": 0.7126396894454956, |
| "learning_rate": 0.000134847409510291, |
| "loss": 1.0051, |
| "step": 920 |
| }, |
| { |
| "epoch": 1.3, |
| "grad_norm": 0.6854078769683838, |
| "learning_rate": 0.0001347764371894961, |
| "loss": 0.8957, |
| "step": 921 |
| }, |
| { |
| "epoch": 1.31, |
| "grad_norm": 0.7038775086402893, |
| "learning_rate": 0.00013470546486870122, |
| "loss": 1.0806, |
| "step": 922 |
| }, |
| { |
| "epoch": 1.31, |
| "grad_norm": 0.6905360817909241, |
| "learning_rate": 0.0001346344925479063, |
| "loss": 1.0533, |
| "step": 923 |
| }, |
| { |
| "epoch": 1.31, |
| "grad_norm": 0.6846379637718201, |
| "learning_rate": 0.00013456352022711142, |
| "loss": 0.9419, |
| "step": 924 |
| }, |
| { |
| "epoch": 1.31, |
| "grad_norm": 0.6838957667350769, |
| "learning_rate": 0.00013449254790631653, |
| "loss": 1.0041, |
| "step": 925 |
| }, |
| { |
| "epoch": 1.31, |
| "grad_norm": 0.6939088106155396, |
| "learning_rate": 0.00013442157558552167, |
| "loss": 1.0256, |
| "step": 926 |
| }, |
| { |
| "epoch": 1.31, |
| "grad_norm": 0.705324649810791, |
| "learning_rate": 0.00013435060326472676, |
| "loss": 1.0321, |
| "step": 927 |
| }, |
| { |
| "epoch": 1.31, |
| "grad_norm": 0.680023729801178, |
| "learning_rate": 0.00013427963094393187, |
| "loss": 0.9367, |
| "step": 928 |
| }, |
| { |
| "epoch": 1.32, |
| "grad_norm": 0.6934013962745667, |
| "learning_rate": 0.00013420865862313699, |
| "loss": 0.9717, |
| "step": 929 |
| }, |
| { |
| "epoch": 1.32, |
| "grad_norm": 0.6807674765586853, |
| "learning_rate": 0.0001341376863023421, |
| "loss": 0.9661, |
| "step": 930 |
| }, |
| { |
| "epoch": 1.32, |
| "grad_norm": 0.695918083190918, |
| "learning_rate": 0.0001340667139815472, |
| "loss": 0.9762, |
| "step": 931 |
| }, |
| { |
| "epoch": 1.32, |
| "grad_norm": 0.6872202754020691, |
| "learning_rate": 0.0001339957416607523, |
| "loss": 0.9487, |
| "step": 932 |
| }, |
| { |
| "epoch": 1.32, |
| "grad_norm": 0.7057836651802063, |
| "learning_rate": 0.0001339247693399574, |
| "loss": 0.9665, |
| "step": 933 |
| }, |
| { |
| "epoch": 1.32, |
| "grad_norm": 0.688558042049408, |
| "learning_rate": 0.00013385379701916255, |
| "loss": 0.9639, |
| "step": 934 |
| }, |
| { |
| "epoch": 1.32, |
| "grad_norm": 0.7189505100250244, |
| "learning_rate": 0.00013378282469836763, |
| "loss": 1.0014, |
| "step": 935 |
| }, |
| { |
| "epoch": 1.33, |
| "grad_norm": 0.6903408169746399, |
| "learning_rate": 0.00013371185237757275, |
| "loss": 1.0319, |
| "step": 936 |
| }, |
| { |
| "epoch": 1.33, |
| "grad_norm": 0.6907972097396851, |
| "learning_rate": 0.00013364088005677786, |
| "loss": 0.9705, |
| "step": 937 |
| }, |
| { |
| "epoch": 1.33, |
| "grad_norm": 0.7174811363220215, |
| "learning_rate": 0.00013356990773598297, |
| "loss": 0.9992, |
| "step": 938 |
| }, |
| { |
| "epoch": 1.33, |
| "grad_norm": 0.6912071108818054, |
| "learning_rate": 0.00013349893541518809, |
| "loss": 0.9241, |
| "step": 939 |
| }, |
| { |
| "epoch": 1.33, |
| "grad_norm": 0.7143810391426086, |
| "learning_rate": 0.00013342796309439317, |
| "loss": 1.0302, |
| "step": 940 |
| }, |
| { |
| "epoch": 1.33, |
| "grad_norm": 0.6805667281150818, |
| "learning_rate": 0.0001333569907735983, |
| "loss": 0.9711, |
| "step": 941 |
| }, |
| { |
| "epoch": 1.33, |
| "grad_norm": 0.6745904088020325, |
| "learning_rate": 0.00013328601845280342, |
| "loss": 0.9788, |
| "step": 942 |
| }, |
| { |
| "epoch": 1.34, |
| "grad_norm": 0.7004431486129761, |
| "learning_rate": 0.00013321504613200854, |
| "loss": 0.913, |
| "step": 943 |
| }, |
| { |
| "epoch": 1.34, |
| "grad_norm": 0.709544837474823, |
| "learning_rate": 0.00013314407381121362, |
| "loss": 1.0007, |
| "step": 944 |
| }, |
| { |
| "epoch": 1.34, |
| "grad_norm": 0.7254632115364075, |
| "learning_rate": 0.00013307310149041873, |
| "loss": 0.9366, |
| "step": 945 |
| }, |
| { |
| "epoch": 1.34, |
| "grad_norm": 0.7147793173789978, |
| "learning_rate": 0.00013300212916962385, |
| "loss": 0.9834, |
| "step": 946 |
| }, |
| { |
| "epoch": 1.34, |
| "grad_norm": 0.733207106590271, |
| "learning_rate": 0.00013293115684882896, |
| "loss": 1.0158, |
| "step": 947 |
| }, |
| { |
| "epoch": 1.34, |
| "grad_norm": 0.7627166509628296, |
| "learning_rate": 0.00013286018452803407, |
| "loss": 1.0266, |
| "step": 948 |
| }, |
| { |
| "epoch": 1.34, |
| "grad_norm": 0.7074815034866333, |
| "learning_rate": 0.00013278921220723919, |
| "loss": 1.012, |
| "step": 949 |
| }, |
| { |
| "epoch": 1.35, |
| "grad_norm": 0.7292148470878601, |
| "learning_rate": 0.0001327182398864443, |
| "loss": 1.0539, |
| "step": 950 |
| }, |
| { |
| "epoch": 1.35, |
| "grad_norm": 0.6858687400817871, |
| "learning_rate": 0.0001326472675656494, |
| "loss": 1.0065, |
| "step": 951 |
| }, |
| { |
| "epoch": 1.35, |
| "grad_norm": 0.6861886978149414, |
| "learning_rate": 0.0001325762952448545, |
| "loss": 0.9988, |
| "step": 952 |
| }, |
| { |
| "epoch": 1.35, |
| "grad_norm": 0.6918312311172485, |
| "learning_rate": 0.0001325053229240596, |
| "loss": 1.0447, |
| "step": 953 |
| }, |
| { |
| "epoch": 1.35, |
| "grad_norm": 0.6733079552650452, |
| "learning_rate": 0.00013243435060326472, |
| "loss": 0.9746, |
| "step": 954 |
| }, |
| { |
| "epoch": 1.35, |
| "grad_norm": 0.7106968760490417, |
| "learning_rate": 0.00013236337828246986, |
| "loss": 1.0219, |
| "step": 955 |
| }, |
| { |
| "epoch": 1.35, |
| "grad_norm": 0.7458065748214722, |
| "learning_rate": 0.00013229240596167495, |
| "loss": 0.9056, |
| "step": 956 |
| }, |
| { |
| "epoch": 1.36, |
| "grad_norm": 0.7352179884910583, |
| "learning_rate": 0.00013222143364088006, |
| "loss": 0.9863, |
| "step": 957 |
| }, |
| { |
| "epoch": 1.36, |
| "grad_norm": 0.6810528635978699, |
| "learning_rate": 0.00013215046132008517, |
| "loss": 0.9686, |
| "step": 958 |
| }, |
| { |
| "epoch": 1.36, |
| "grad_norm": 0.6850380301475525, |
| "learning_rate": 0.00013207948899929029, |
| "loss": 0.9124, |
| "step": 959 |
| }, |
| { |
| "epoch": 1.36, |
| "grad_norm": 0.7115997672080994, |
| "learning_rate": 0.0001320085166784954, |
| "loss": 0.9862, |
| "step": 960 |
| }, |
| { |
| "epoch": 1.36, |
| "grad_norm": 0.7444871664047241, |
| "learning_rate": 0.00013193754435770048, |
| "loss": 0.9958, |
| "step": 961 |
| }, |
| { |
| "epoch": 1.36, |
| "grad_norm": 0.7253167629241943, |
| "learning_rate": 0.00013186657203690562, |
| "loss": 0.9751, |
| "step": 962 |
| }, |
| { |
| "epoch": 1.36, |
| "grad_norm": 0.7287327647209167, |
| "learning_rate": 0.00013179559971611074, |
| "loss": 0.9621, |
| "step": 963 |
| }, |
| { |
| "epoch": 1.37, |
| "grad_norm": 0.7266460061073303, |
| "learning_rate": 0.00013172462739531582, |
| "loss": 1.0085, |
| "step": 964 |
| }, |
| { |
| "epoch": 1.37, |
| "grad_norm": 0.6825550198554993, |
| "learning_rate": 0.00013165365507452094, |
| "loss": 0.9696, |
| "step": 965 |
| }, |
| { |
| "epoch": 1.37, |
| "grad_norm": 0.7191211581230164, |
| "learning_rate": 0.00013158268275372605, |
| "loss": 1.0167, |
| "step": 966 |
| }, |
| { |
| "epoch": 1.37, |
| "grad_norm": 0.6877723336219788, |
| "learning_rate": 0.00013151171043293116, |
| "loss": 0.9224, |
| "step": 967 |
| }, |
| { |
| "epoch": 1.37, |
| "grad_norm": 0.6919003129005432, |
| "learning_rate": 0.00013144073811213627, |
| "loss": 0.9628, |
| "step": 968 |
| }, |
| { |
| "epoch": 1.37, |
| "grad_norm": 0.7358181476593018, |
| "learning_rate": 0.00013136976579134139, |
| "loss": 1.0684, |
| "step": 969 |
| }, |
| { |
| "epoch": 1.37, |
| "grad_norm": 0.7027356624603271, |
| "learning_rate": 0.0001312987934705465, |
| "loss": 1.0333, |
| "step": 970 |
| }, |
| { |
| "epoch": 1.38, |
| "grad_norm": 0.6980712413787842, |
| "learning_rate": 0.0001312278211497516, |
| "loss": 0.9823, |
| "step": 971 |
| }, |
| { |
| "epoch": 1.38, |
| "grad_norm": 0.7023440599441528, |
| "learning_rate": 0.00013115684882895672, |
| "loss": 0.9497, |
| "step": 972 |
| }, |
| { |
| "epoch": 1.38, |
| "grad_norm": 0.7038329839706421, |
| "learning_rate": 0.0001310858765081618, |
| "loss": 0.9915, |
| "step": 973 |
| }, |
| { |
| "epoch": 1.38, |
| "grad_norm": 0.7077039480209351, |
| "learning_rate": 0.00013101490418736692, |
| "loss": 0.947, |
| "step": 974 |
| }, |
| { |
| "epoch": 1.38, |
| "grad_norm": 0.7472418546676636, |
| "learning_rate": 0.00013094393186657204, |
| "loss": 0.9416, |
| "step": 975 |
| }, |
| { |
| "epoch": 1.38, |
| "grad_norm": 0.6974193453788757, |
| "learning_rate": 0.00013087295954577718, |
| "loss": 0.9788, |
| "step": 976 |
| }, |
| { |
| "epoch": 1.38, |
| "grad_norm": 0.6888452768325806, |
| "learning_rate": 0.00013080198722498226, |
| "loss": 0.9453, |
| "step": 977 |
| }, |
| { |
| "epoch": 1.39, |
| "grad_norm": 0.6868741512298584, |
| "learning_rate": 0.00013073101490418737, |
| "loss": 0.9096, |
| "step": 978 |
| }, |
| { |
| "epoch": 1.39, |
| "grad_norm": 0.701124370098114, |
| "learning_rate": 0.00013066004258339249, |
| "loss": 1.0178, |
| "step": 979 |
| }, |
| { |
| "epoch": 1.39, |
| "grad_norm": 0.7294360399246216, |
| "learning_rate": 0.0001305890702625976, |
| "loss": 1.0384, |
| "step": 980 |
| }, |
| { |
| "epoch": 1.39, |
| "grad_norm": 0.719120442867279, |
| "learning_rate": 0.00013051809794180268, |
| "loss": 0.9209, |
| "step": 981 |
| }, |
| { |
| "epoch": 1.39, |
| "grad_norm": 0.718369722366333, |
| "learning_rate": 0.0001304471256210078, |
| "loss": 0.9457, |
| "step": 982 |
| }, |
| { |
| "epoch": 1.39, |
| "grad_norm": 0.6928309202194214, |
| "learning_rate": 0.00013037615330021294, |
| "loss": 0.9093, |
| "step": 983 |
| }, |
| { |
| "epoch": 1.39, |
| "grad_norm": 0.7082747220993042, |
| "learning_rate": 0.00013030518097941805, |
| "loss": 0.9639, |
| "step": 984 |
| }, |
| { |
| "epoch": 1.4, |
| "grad_norm": 0.747394323348999, |
| "learning_rate": 0.00013023420865862314, |
| "loss": 1.0332, |
| "step": 985 |
| }, |
| { |
| "epoch": 1.4, |
| "grad_norm": 0.6965970993041992, |
| "learning_rate": 0.00013016323633782825, |
| "loss": 1.011, |
| "step": 986 |
| }, |
| { |
| "epoch": 1.4, |
| "grad_norm": 0.6957747340202332, |
| "learning_rate": 0.00013009226401703336, |
| "loss": 0.9917, |
| "step": 987 |
| }, |
| { |
| "epoch": 1.4, |
| "grad_norm": 0.7042578458786011, |
| "learning_rate": 0.00013002129169623847, |
| "loss": 0.9221, |
| "step": 988 |
| }, |
| { |
| "epoch": 1.4, |
| "grad_norm": 0.709457278251648, |
| "learning_rate": 0.0001299503193754436, |
| "loss": 0.9621, |
| "step": 989 |
| }, |
| { |
| "epoch": 1.4, |
| "grad_norm": 0.7135986685752869, |
| "learning_rate": 0.0001298793470546487, |
| "loss": 0.9894, |
| "step": 990 |
| }, |
| { |
| "epoch": 1.4, |
| "grad_norm": 0.7359340190887451, |
| "learning_rate": 0.0001298083747338538, |
| "loss": 0.9508, |
| "step": 991 |
| }, |
| { |
| "epoch": 1.41, |
| "grad_norm": 0.7456433176994324, |
| "learning_rate": 0.00012973740241305892, |
| "loss": 1.0153, |
| "step": 992 |
| }, |
| { |
| "epoch": 1.41, |
| "grad_norm": 0.7097339034080505, |
| "learning_rate": 0.00012966643009226404, |
| "loss": 0.9799, |
| "step": 993 |
| }, |
| { |
| "epoch": 1.41, |
| "grad_norm": 0.7416273951530457, |
| "learning_rate": 0.00012959545777146912, |
| "loss": 1.0445, |
| "step": 994 |
| }, |
| { |
| "epoch": 1.41, |
| "grad_norm": 0.7126497030258179, |
| "learning_rate": 0.00012952448545067424, |
| "loss": 0.9907, |
| "step": 995 |
| }, |
| { |
| "epoch": 1.41, |
| "grad_norm": 0.7022689580917358, |
| "learning_rate": 0.00012945351312987935, |
| "loss": 0.9576, |
| "step": 996 |
| }, |
| { |
| "epoch": 1.41, |
| "grad_norm": 0.728905439376831, |
| "learning_rate": 0.00012938254080908446, |
| "loss": 0.9537, |
| "step": 997 |
| }, |
| { |
| "epoch": 1.41, |
| "grad_norm": 0.7206951975822449, |
| "learning_rate": 0.00012931156848828957, |
| "loss": 1.0048, |
| "step": 998 |
| }, |
| { |
| "epoch": 1.42, |
| "grad_norm": 0.6895846724510193, |
| "learning_rate": 0.0001292405961674947, |
| "loss": 1.0182, |
| "step": 999 |
| }, |
| { |
| "epoch": 1.42, |
| "grad_norm": 0.7300769090652466, |
| "learning_rate": 0.0001291696238466998, |
| "loss": 1.0212, |
| "step": 1000 |
| }, |
| { |
| "epoch": 1.42, |
| "grad_norm": 0.7058583498001099, |
| "learning_rate": 0.0001290986515259049, |
| "loss": 0.9982, |
| "step": 1001 |
| }, |
| { |
| "epoch": 1.42, |
| "grad_norm": 0.681090235710144, |
| "learning_rate": 0.00012902767920511, |
| "loss": 0.9898, |
| "step": 1002 |
| }, |
| { |
| "epoch": 1.42, |
| "grad_norm": 0.6858307123184204, |
| "learning_rate": 0.0001289567068843151, |
| "loss": 0.961, |
| "step": 1003 |
| }, |
| { |
| "epoch": 1.42, |
| "grad_norm": 0.6954583525657654, |
| "learning_rate": 0.00012888573456352025, |
| "loss": 1.0346, |
| "step": 1004 |
| }, |
| { |
| "epoch": 1.42, |
| "grad_norm": 0.7096951007843018, |
| "learning_rate": 0.00012881476224272536, |
| "loss": 0.9599, |
| "step": 1005 |
| }, |
| { |
| "epoch": 1.43, |
| "grad_norm": 0.7118542194366455, |
| "learning_rate": 0.00012874378992193045, |
| "loss": 0.9677, |
| "step": 1006 |
| }, |
| { |
| "epoch": 1.43, |
| "grad_norm": 0.7197057604789734, |
| "learning_rate": 0.00012867281760113556, |
| "loss": 0.937, |
| "step": 1007 |
| }, |
| { |
| "epoch": 1.43, |
| "grad_norm": 0.720856785774231, |
| "learning_rate": 0.00012860184528034067, |
| "loss": 0.9606, |
| "step": 1008 |
| }, |
| { |
| "epoch": 1.43, |
| "grad_norm": 0.6895302534103394, |
| "learning_rate": 0.0001285308729595458, |
| "loss": 0.9052, |
| "step": 1009 |
| }, |
| { |
| "epoch": 1.43, |
| "grad_norm": 0.7161164879798889, |
| "learning_rate": 0.00012845990063875087, |
| "loss": 0.9784, |
| "step": 1010 |
| }, |
| { |
| "epoch": 1.43, |
| "grad_norm": 0.7340686321258545, |
| "learning_rate": 0.000128388928317956, |
| "loss": 0.9954, |
| "step": 1011 |
| }, |
| { |
| "epoch": 1.43, |
| "grad_norm": 0.7402980923652649, |
| "learning_rate": 0.00012831795599716112, |
| "loss": 0.9905, |
| "step": 1012 |
| }, |
| { |
| "epoch": 1.44, |
| "grad_norm": 0.6780891418457031, |
| "learning_rate": 0.00012824698367636624, |
| "loss": 0.8967, |
| "step": 1013 |
| }, |
| { |
| "epoch": 1.44, |
| "grad_norm": 0.6960120797157288, |
| "learning_rate": 0.00012817601135557132, |
| "loss": 0.8983, |
| "step": 1014 |
| }, |
| { |
| "epoch": 1.44, |
| "grad_norm": 0.7087661623954773, |
| "learning_rate": 0.00012810503903477644, |
| "loss": 0.927, |
| "step": 1015 |
| }, |
| { |
| "epoch": 1.44, |
| "grad_norm": 0.747851550579071, |
| "learning_rate": 0.00012803406671398155, |
| "loss": 1.0698, |
| "step": 1016 |
| }, |
| { |
| "epoch": 1.44, |
| "grad_norm": 0.7293078899383545, |
| "learning_rate": 0.00012796309439318666, |
| "loss": 1.007, |
| "step": 1017 |
| }, |
| { |
| "epoch": 1.44, |
| "grad_norm": 0.7039139270782471, |
| "learning_rate": 0.00012789212207239177, |
| "loss": 0.8827, |
| "step": 1018 |
| }, |
| { |
| "epoch": 1.44, |
| "grad_norm": 0.7060878276824951, |
| "learning_rate": 0.0001278211497515969, |
| "loss": 0.9078, |
| "step": 1019 |
| }, |
| { |
| "epoch": 1.45, |
| "grad_norm": 0.7231726050376892, |
| "learning_rate": 0.000127750177430802, |
| "loss": 0.9605, |
| "step": 1020 |
| }, |
| { |
| "epoch": 1.45, |
| "grad_norm": 0.7151123285293579, |
| "learning_rate": 0.0001276792051100071, |
| "loss": 0.9377, |
| "step": 1021 |
| }, |
| { |
| "epoch": 1.45, |
| "grad_norm": 0.7285674810409546, |
| "learning_rate": 0.00012760823278921223, |
| "loss": 0.959, |
| "step": 1022 |
| }, |
| { |
| "epoch": 1.45, |
| "grad_norm": 0.7382057309150696, |
| "learning_rate": 0.0001275372604684173, |
| "loss": 0.9733, |
| "step": 1023 |
| }, |
| { |
| "epoch": 1.45, |
| "grad_norm": 0.7037867307662964, |
| "learning_rate": 0.00012746628814762242, |
| "loss": 0.9261, |
| "step": 1024 |
| }, |
| { |
| "epoch": 1.45, |
| "grad_norm": 0.7247341275215149, |
| "learning_rate": 0.00012739531582682754, |
| "loss": 0.9765, |
| "step": 1025 |
| }, |
| { |
| "epoch": 1.45, |
| "grad_norm": 0.73041170835495, |
| "learning_rate": 0.00012732434350603265, |
| "loss": 0.956, |
| "step": 1026 |
| }, |
| { |
| "epoch": 1.45, |
| "grad_norm": 0.6923772096633911, |
| "learning_rate": 0.00012725337118523776, |
| "loss": 1.0133, |
| "step": 1027 |
| }, |
| { |
| "epoch": 1.46, |
| "grad_norm": 0.6898492574691772, |
| "learning_rate": 0.00012718239886444287, |
| "loss": 1.0216, |
| "step": 1028 |
| }, |
| { |
| "epoch": 1.46, |
| "grad_norm": 0.7031357288360596, |
| "learning_rate": 0.000127111426543648, |
| "loss": 0.9667, |
| "step": 1029 |
| }, |
| { |
| "epoch": 1.46, |
| "grad_norm": 0.7023702263832092, |
| "learning_rate": 0.0001270404542228531, |
| "loss": 0.9699, |
| "step": 1030 |
| }, |
| { |
| "epoch": 1.46, |
| "grad_norm": 0.6999973654747009, |
| "learning_rate": 0.00012696948190205819, |
| "loss": 0.9927, |
| "step": 1031 |
| }, |
| { |
| "epoch": 1.46, |
| "grad_norm": 0.7087770104408264, |
| "learning_rate": 0.0001268985095812633, |
| "loss": 0.8973, |
| "step": 1032 |
| }, |
| { |
| "epoch": 1.46, |
| "grad_norm": 0.7190529704093933, |
| "learning_rate": 0.00012682753726046844, |
| "loss": 0.9662, |
| "step": 1033 |
| }, |
| { |
| "epoch": 1.46, |
| "grad_norm": 0.7416104674339294, |
| "learning_rate": 0.00012675656493967355, |
| "loss": 1.0066, |
| "step": 1034 |
| }, |
| { |
| "epoch": 1.47, |
| "grad_norm": 0.7032634615898132, |
| "learning_rate": 0.00012668559261887864, |
| "loss": 0.9999, |
| "step": 1035 |
| }, |
| { |
| "epoch": 1.47, |
| "grad_norm": 0.7097875475883484, |
| "learning_rate": 0.00012661462029808375, |
| "loss": 0.9751, |
| "step": 1036 |
| }, |
| { |
| "epoch": 1.47, |
| "grad_norm": 0.6958321928977966, |
| "learning_rate": 0.00012654364797728886, |
| "loss": 0.9407, |
| "step": 1037 |
| }, |
| { |
| "epoch": 1.47, |
| "grad_norm": 0.7388033270835876, |
| "learning_rate": 0.00012647267565649397, |
| "loss": 1.0527, |
| "step": 1038 |
| }, |
| { |
| "epoch": 1.47, |
| "grad_norm": 0.6906083226203918, |
| "learning_rate": 0.00012640170333569906, |
| "loss": 0.9368, |
| "step": 1039 |
| }, |
| { |
| "epoch": 1.47, |
| "grad_norm": 0.7116028666496277, |
| "learning_rate": 0.0001263307310149042, |
| "loss": 0.9618, |
| "step": 1040 |
| }, |
| { |
| "epoch": 1.47, |
| "grad_norm": 0.7041985392570496, |
| "learning_rate": 0.0001262597586941093, |
| "loss": 0.9927, |
| "step": 1041 |
| }, |
| { |
| "epoch": 1.48, |
| "grad_norm": 0.7428408861160278, |
| "learning_rate": 0.00012618878637331443, |
| "loss": 1.0282, |
| "step": 1042 |
| }, |
| { |
| "epoch": 1.48, |
| "grad_norm": 0.6752524375915527, |
| "learning_rate": 0.0001261178140525195, |
| "loss": 0.927, |
| "step": 1043 |
| }, |
| { |
| "epoch": 1.48, |
| "grad_norm": 0.7071763277053833, |
| "learning_rate": 0.00012604684173172462, |
| "loss": 0.9534, |
| "step": 1044 |
| }, |
| { |
| "epoch": 1.48, |
| "grad_norm": 0.710394561290741, |
| "learning_rate": 0.00012597586941092974, |
| "loss": 0.9324, |
| "step": 1045 |
| }, |
| { |
| "epoch": 1.48, |
| "grad_norm": 0.7077054381370544, |
| "learning_rate": 0.00012590489709013485, |
| "loss": 0.9826, |
| "step": 1046 |
| }, |
| { |
| "epoch": 1.48, |
| "grad_norm": 0.7074711918830872, |
| "learning_rate": 0.00012583392476933996, |
| "loss": 0.9878, |
| "step": 1047 |
| }, |
| { |
| "epoch": 1.48, |
| "grad_norm": 0.6914960145950317, |
| "learning_rate": 0.00012576295244854507, |
| "loss": 1.0357, |
| "step": 1048 |
| }, |
| { |
| "epoch": 1.49, |
| "grad_norm": 0.6889827251434326, |
| "learning_rate": 0.0001256919801277502, |
| "loss": 0.982, |
| "step": 1049 |
| }, |
| { |
| "epoch": 1.49, |
| "grad_norm": 0.6742181777954102, |
| "learning_rate": 0.0001256210078069553, |
| "loss": 0.9513, |
| "step": 1050 |
| }, |
| { |
| "epoch": 1.49, |
| "grad_norm": 0.6830121874809265, |
| "learning_rate": 0.0001255500354861604, |
| "loss": 0.9961, |
| "step": 1051 |
| }, |
| { |
| "epoch": 1.49, |
| "grad_norm": 0.7067031264305115, |
| "learning_rate": 0.0001254790631653655, |
| "loss": 0.9623, |
| "step": 1052 |
| }, |
| { |
| "epoch": 1.49, |
| "grad_norm": 0.6789557933807373, |
| "learning_rate": 0.0001254080908445706, |
| "loss": 0.9556, |
| "step": 1053 |
| }, |
| { |
| "epoch": 1.49, |
| "grad_norm": 0.7340845465660095, |
| "learning_rate": 0.00012533711852377575, |
| "loss": 1.0122, |
| "step": 1054 |
| }, |
| { |
| "epoch": 1.49, |
| "grad_norm": 0.7131407260894775, |
| "learning_rate": 0.00012526614620298084, |
| "loss": 0.9998, |
| "step": 1055 |
| }, |
| { |
| "epoch": 1.5, |
| "grad_norm": 0.7129774689674377, |
| "learning_rate": 0.00012519517388218595, |
| "loss": 0.9348, |
| "step": 1056 |
| }, |
| { |
| "epoch": 1.5, |
| "grad_norm": 0.7226494550704956, |
| "learning_rate": 0.00012512420156139106, |
| "loss": 0.9792, |
| "step": 1057 |
| }, |
| { |
| "epoch": 1.5, |
| "grad_norm": 0.7202288508415222, |
| "learning_rate": 0.00012505322924059617, |
| "loss": 0.9546, |
| "step": 1058 |
| }, |
| { |
| "epoch": 1.5, |
| "grad_norm": 0.7222435474395752, |
| "learning_rate": 0.0001249822569198013, |
| "loss": 0.9995, |
| "step": 1059 |
| }, |
| { |
| "epoch": 1.5, |
| "grad_norm": 0.7435225248336792, |
| "learning_rate": 0.00012491128459900637, |
| "loss": 1.033, |
| "step": 1060 |
| }, |
| { |
| "epoch": 1.5, |
| "grad_norm": 0.699711799621582, |
| "learning_rate": 0.0001248403122782115, |
| "loss": 0.9639, |
| "step": 1061 |
| }, |
| { |
| "epoch": 1.5, |
| "grad_norm": 0.7038230895996094, |
| "learning_rate": 0.00012476933995741663, |
| "loss": 0.9198, |
| "step": 1062 |
| }, |
| { |
| "epoch": 1.51, |
| "grad_norm": 0.7003630995750427, |
| "learning_rate": 0.00012469836763662174, |
| "loss": 1.0208, |
| "step": 1063 |
| }, |
| { |
| "epoch": 1.51, |
| "grad_norm": 0.7104167938232422, |
| "learning_rate": 0.00012462739531582682, |
| "loss": 0.986, |
| "step": 1064 |
| }, |
| { |
| "epoch": 1.51, |
| "grad_norm": 0.7140862941741943, |
| "learning_rate": 0.00012455642299503194, |
| "loss": 0.9822, |
| "step": 1065 |
| }, |
| { |
| "epoch": 1.51, |
| "grad_norm": 0.7276419401168823, |
| "learning_rate": 0.00012448545067423705, |
| "loss": 1.0231, |
| "step": 1066 |
| }, |
| { |
| "epoch": 1.51, |
| "grad_norm": 0.7366431355476379, |
| "learning_rate": 0.00012441447835344216, |
| "loss": 1.0071, |
| "step": 1067 |
| }, |
| { |
| "epoch": 1.51, |
| "grad_norm": 0.7350658178329468, |
| "learning_rate": 0.00012434350603264727, |
| "loss": 0.9918, |
| "step": 1068 |
| }, |
| { |
| "epoch": 1.51, |
| "grad_norm": 0.6769629716873169, |
| "learning_rate": 0.0001242725337118524, |
| "loss": 0.9973, |
| "step": 1069 |
| }, |
| { |
| "epoch": 1.52, |
| "grad_norm": 0.7047297358512878, |
| "learning_rate": 0.0001242015613910575, |
| "loss": 1.0167, |
| "step": 1070 |
| }, |
| { |
| "epoch": 1.52, |
| "grad_norm": 0.6941127181053162, |
| "learning_rate": 0.0001241305890702626, |
| "loss": 0.9861, |
| "step": 1071 |
| }, |
| { |
| "epoch": 1.52, |
| "grad_norm": 0.727709949016571, |
| "learning_rate": 0.0001240596167494677, |
| "loss": 1.0325, |
| "step": 1072 |
| }, |
| { |
| "epoch": 1.52, |
| "grad_norm": 0.7129109501838684, |
| "learning_rate": 0.0001239886444286728, |
| "loss": 0.9249, |
| "step": 1073 |
| }, |
| { |
| "epoch": 1.52, |
| "grad_norm": 0.6976962089538574, |
| "learning_rate": 0.00012391767210787792, |
| "loss": 0.9472, |
| "step": 1074 |
| }, |
| { |
| "epoch": 1.52, |
| "grad_norm": 0.7473706603050232, |
| "learning_rate": 0.00012384669978708306, |
| "loss": 0.9843, |
| "step": 1075 |
| }, |
| { |
| "epoch": 1.52, |
| "grad_norm": 0.7451398968696594, |
| "learning_rate": 0.00012377572746628815, |
| "loss": 0.9926, |
| "step": 1076 |
| }, |
| { |
| "epoch": 1.53, |
| "grad_norm": 0.7221840023994446, |
| "learning_rate": 0.00012370475514549326, |
| "loss": 0.9513, |
| "step": 1077 |
| }, |
| { |
| "epoch": 1.53, |
| "grad_norm": 0.7023710012435913, |
| "learning_rate": 0.00012363378282469838, |
| "loss": 0.8927, |
| "step": 1078 |
| }, |
| { |
| "epoch": 1.53, |
| "grad_norm": 0.6784040331840515, |
| "learning_rate": 0.0001235628105039035, |
| "loss": 0.9059, |
| "step": 1079 |
| }, |
| { |
| "epoch": 1.53, |
| "grad_norm": 0.7393279671669006, |
| "learning_rate": 0.0001234918381831086, |
| "loss": 1.0369, |
| "step": 1080 |
| }, |
| { |
| "epoch": 1.53, |
| "grad_norm": 0.6987690329551697, |
| "learning_rate": 0.00012342086586231369, |
| "loss": 0.961, |
| "step": 1081 |
| }, |
| { |
| "epoch": 1.53, |
| "grad_norm": 0.7038789987564087, |
| "learning_rate": 0.00012334989354151883, |
| "loss": 1.0484, |
| "step": 1082 |
| }, |
| { |
| "epoch": 1.53, |
| "grad_norm": 0.684845507144928, |
| "learning_rate": 0.00012327892122072394, |
| "loss": 0.9604, |
| "step": 1083 |
| }, |
| { |
| "epoch": 1.54, |
| "grad_norm": 0.7157427072525024, |
| "learning_rate": 0.00012320794889992902, |
| "loss": 0.9702, |
| "step": 1084 |
| }, |
| { |
| "epoch": 1.54, |
| "grad_norm": 0.7225085496902466, |
| "learning_rate": 0.00012313697657913414, |
| "loss": 1.0421, |
| "step": 1085 |
| }, |
| { |
| "epoch": 1.54, |
| "grad_norm": 0.7261961698532104, |
| "learning_rate": 0.00012306600425833925, |
| "loss": 1.0224, |
| "step": 1086 |
| }, |
| { |
| "epoch": 1.54, |
| "grad_norm": 0.6956989765167236, |
| "learning_rate": 0.00012299503193754436, |
| "loss": 0.9708, |
| "step": 1087 |
| }, |
| { |
| "epoch": 1.54, |
| "grad_norm": 0.7316460609436035, |
| "learning_rate": 0.00012292405961674948, |
| "loss": 1.0543, |
| "step": 1088 |
| }, |
| { |
| "epoch": 1.54, |
| "grad_norm": 0.7196297645568848, |
| "learning_rate": 0.0001228530872959546, |
| "loss": 1.0021, |
| "step": 1089 |
| }, |
| { |
| "epoch": 1.54, |
| "grad_norm": 0.7039411664009094, |
| "learning_rate": 0.0001227821149751597, |
| "loss": 0.9563, |
| "step": 1090 |
| }, |
| { |
| "epoch": 1.55, |
| "grad_norm": 0.719147801399231, |
| "learning_rate": 0.0001227111426543648, |
| "loss": 1.0067, |
| "step": 1091 |
| }, |
| { |
| "epoch": 1.55, |
| "grad_norm": 0.6809403300285339, |
| "learning_rate": 0.00012264017033356993, |
| "loss": 0.995, |
| "step": 1092 |
| }, |
| { |
| "epoch": 1.55, |
| "grad_norm": 0.7004520297050476, |
| "learning_rate": 0.000122569198012775, |
| "loss": 0.9479, |
| "step": 1093 |
| }, |
| { |
| "epoch": 1.55, |
| "grad_norm": 0.7003827691078186, |
| "learning_rate": 0.00012249822569198012, |
| "loss": 0.9724, |
| "step": 1094 |
| }, |
| { |
| "epoch": 1.55, |
| "grad_norm": 0.7075307965278625, |
| "learning_rate": 0.00012242725337118524, |
| "loss": 0.9856, |
| "step": 1095 |
| }, |
| { |
| "epoch": 1.55, |
| "grad_norm": 0.723334789276123, |
| "learning_rate": 0.00012235628105039038, |
| "loss": 0.9768, |
| "step": 1096 |
| }, |
| { |
| "epoch": 1.55, |
| "grad_norm": 0.7285984754562378, |
| "learning_rate": 0.00012228530872959546, |
| "loss": 1.005, |
| "step": 1097 |
| }, |
| { |
| "epoch": 1.56, |
| "grad_norm": 0.7101818919181824, |
| "learning_rate": 0.00012221433640880058, |
| "loss": 0.9543, |
| "step": 1098 |
| }, |
| { |
| "epoch": 1.56, |
| "grad_norm": 0.749718427658081, |
| "learning_rate": 0.0001221433640880057, |
| "loss": 1.0076, |
| "step": 1099 |
| }, |
| { |
| "epoch": 1.56, |
| "grad_norm": 0.7114327549934387, |
| "learning_rate": 0.0001220723917672108, |
| "loss": 0.9289, |
| "step": 1100 |
| }, |
| { |
| "epoch": 1.56, |
| "grad_norm": 0.7335402965545654, |
| "learning_rate": 0.0001220014194464159, |
| "loss": 0.9761, |
| "step": 1101 |
| }, |
| { |
| "epoch": 1.56, |
| "grad_norm": 0.7315895557403564, |
| "learning_rate": 0.000121930447125621, |
| "loss": 0.9498, |
| "step": 1102 |
| }, |
| { |
| "epoch": 1.56, |
| "grad_norm": 0.7099215984344482, |
| "learning_rate": 0.00012185947480482611, |
| "loss": 0.9575, |
| "step": 1103 |
| }, |
| { |
| "epoch": 1.56, |
| "grad_norm": 0.7005743384361267, |
| "learning_rate": 0.00012178850248403124, |
| "loss": 0.9605, |
| "step": 1104 |
| }, |
| { |
| "epoch": 1.57, |
| "grad_norm": 0.7421797513961792, |
| "learning_rate": 0.00012171753016323635, |
| "loss": 1.067, |
| "step": 1105 |
| }, |
| { |
| "epoch": 1.57, |
| "grad_norm": 0.7173910737037659, |
| "learning_rate": 0.00012164655784244145, |
| "loss": 0.9843, |
| "step": 1106 |
| }, |
| { |
| "epoch": 1.57, |
| "grad_norm": 0.6898432374000549, |
| "learning_rate": 0.00012157558552164656, |
| "loss": 0.9735, |
| "step": 1107 |
| }, |
| { |
| "epoch": 1.57, |
| "grad_norm": 0.708389937877655, |
| "learning_rate": 0.00012150461320085168, |
| "loss": 0.9497, |
| "step": 1108 |
| }, |
| { |
| "epoch": 1.57, |
| "grad_norm": 0.7111474275588989, |
| "learning_rate": 0.00012143364088005677, |
| "loss": 0.8981, |
| "step": 1109 |
| }, |
| { |
| "epoch": 1.57, |
| "grad_norm": 0.6907044053077698, |
| "learning_rate": 0.00012136266855926189, |
| "loss": 0.9711, |
| "step": 1110 |
| }, |
| { |
| "epoch": 1.57, |
| "grad_norm": 0.744263231754303, |
| "learning_rate": 0.00012129169623846701, |
| "loss": 0.9874, |
| "step": 1111 |
| }, |
| { |
| "epoch": 1.58, |
| "grad_norm": 0.7199329733848572, |
| "learning_rate": 0.00012122072391767211, |
| "loss": 0.9704, |
| "step": 1112 |
| }, |
| { |
| "epoch": 1.58, |
| "grad_norm": 0.7198705077171326, |
| "learning_rate": 0.00012114975159687723, |
| "loss": 0.9508, |
| "step": 1113 |
| }, |
| { |
| "epoch": 1.58, |
| "grad_norm": 0.7411786913871765, |
| "learning_rate": 0.00012107877927608234, |
| "loss": 0.9998, |
| "step": 1114 |
| }, |
| { |
| "epoch": 1.58, |
| "grad_norm": 0.675391435623169, |
| "learning_rate": 0.00012100780695528744, |
| "loss": 0.9317, |
| "step": 1115 |
| }, |
| { |
| "epoch": 1.58, |
| "grad_norm": 0.6812769174575806, |
| "learning_rate": 0.00012093683463449255, |
| "loss": 0.889, |
| "step": 1116 |
| }, |
| { |
| "epoch": 1.58, |
| "grad_norm": 0.7100497484207153, |
| "learning_rate": 0.00012086586231369765, |
| "loss": 1.0161, |
| "step": 1117 |
| }, |
| { |
| "epoch": 1.58, |
| "grad_norm": 0.7073559761047363, |
| "learning_rate": 0.00012079488999290278, |
| "loss": 0.9373, |
| "step": 1118 |
| }, |
| { |
| "epoch": 1.59, |
| "grad_norm": 0.7181984186172485, |
| "learning_rate": 0.00012072391767210789, |
| "loss": 1.0138, |
| "step": 1119 |
| }, |
| { |
| "epoch": 1.59, |
| "grad_norm": 0.7441282272338867, |
| "learning_rate": 0.000120652945351313, |
| "loss": 1.0439, |
| "step": 1120 |
| }, |
| { |
| "epoch": 1.59, |
| "grad_norm": 0.7104730606079102, |
| "learning_rate": 0.0001205819730305181, |
| "loss": 1.0055, |
| "step": 1121 |
| }, |
| { |
| "epoch": 1.59, |
| "grad_norm": 0.7464049458503723, |
| "learning_rate": 0.00012051100070972321, |
| "loss": 1.011, |
| "step": 1122 |
| }, |
| { |
| "epoch": 1.59, |
| "grad_norm": 0.7502149939537048, |
| "learning_rate": 0.00012044002838892831, |
| "loss": 0.9707, |
| "step": 1123 |
| }, |
| { |
| "epoch": 1.59, |
| "grad_norm": 0.7351347804069519, |
| "learning_rate": 0.00012036905606813342, |
| "loss": 0.9797, |
| "step": 1124 |
| }, |
| { |
| "epoch": 1.59, |
| "grad_norm": 0.7235199809074402, |
| "learning_rate": 0.00012029808374733855, |
| "loss": 1.0525, |
| "step": 1125 |
| }, |
| { |
| "epoch": 1.6, |
| "grad_norm": 0.6959457993507385, |
| "learning_rate": 0.00012022711142654366, |
| "loss": 0.9818, |
| "step": 1126 |
| }, |
| { |
| "epoch": 1.6, |
| "grad_norm": 0.7070314288139343, |
| "learning_rate": 0.00012015613910574876, |
| "loss": 1.0983, |
| "step": 1127 |
| }, |
| { |
| "epoch": 1.6, |
| "grad_norm": 0.7202725410461426, |
| "learning_rate": 0.00012008516678495388, |
| "loss": 1.046, |
| "step": 1128 |
| }, |
| { |
| "epoch": 1.6, |
| "grad_norm": 0.6858893632888794, |
| "learning_rate": 0.00012001419446415898, |
| "loss": 0.8999, |
| "step": 1129 |
| }, |
| { |
| "epoch": 1.6, |
| "grad_norm": 0.7225475907325745, |
| "learning_rate": 0.00011994322214336409, |
| "loss": 1.0171, |
| "step": 1130 |
| }, |
| { |
| "epoch": 1.6, |
| "grad_norm": 0.7360539436340332, |
| "learning_rate": 0.00011987224982256919, |
| "loss": 1.047, |
| "step": 1131 |
| }, |
| { |
| "epoch": 1.6, |
| "grad_norm": 0.7139810919761658, |
| "learning_rate": 0.00011980127750177433, |
| "loss": 1.008, |
| "step": 1132 |
| }, |
| { |
| "epoch": 1.61, |
| "grad_norm": 0.7204037308692932, |
| "learning_rate": 0.00011973030518097943, |
| "loss": 0.9262, |
| "step": 1133 |
| }, |
| { |
| "epoch": 1.61, |
| "grad_norm": 0.7312289476394653, |
| "learning_rate": 0.00011965933286018454, |
| "loss": 1.0079, |
| "step": 1134 |
| }, |
| { |
| "epoch": 1.61, |
| "grad_norm": 0.7065761685371399, |
| "learning_rate": 0.00011958836053938964, |
| "loss": 0.9956, |
| "step": 1135 |
| }, |
| { |
| "epoch": 1.61, |
| "grad_norm": 0.743116557598114, |
| "learning_rate": 0.00011951738821859475, |
| "loss": 0.9679, |
| "step": 1136 |
| }, |
| { |
| "epoch": 1.61, |
| "grad_norm": 0.7406318783760071, |
| "learning_rate": 0.00011944641589779986, |
| "loss": 1.0024, |
| "step": 1137 |
| }, |
| { |
| "epoch": 1.61, |
| "grad_norm": 0.7344107031822205, |
| "learning_rate": 0.00011937544357700496, |
| "loss": 0.9899, |
| "step": 1138 |
| }, |
| { |
| "epoch": 1.61, |
| "grad_norm": 0.7228212356567383, |
| "learning_rate": 0.00011930447125621009, |
| "loss": 0.9902, |
| "step": 1139 |
| }, |
| { |
| "epoch": 1.62, |
| "grad_norm": 0.7286756634712219, |
| "learning_rate": 0.0001192334989354152, |
| "loss": 1.0239, |
| "step": 1140 |
| }, |
| { |
| "epoch": 1.62, |
| "grad_norm": 0.7067569494247437, |
| "learning_rate": 0.0001191625266146203, |
| "loss": 0.95, |
| "step": 1141 |
| }, |
| { |
| "epoch": 1.62, |
| "grad_norm": 0.7505934238433838, |
| "learning_rate": 0.00011909155429382541, |
| "loss": 1.0528, |
| "step": 1142 |
| }, |
| { |
| "epoch": 1.62, |
| "grad_norm": 0.6888457536697388, |
| "learning_rate": 0.00011902058197303053, |
| "loss": 0.9544, |
| "step": 1143 |
| }, |
| { |
| "epoch": 1.62, |
| "grad_norm": 0.6858482956886292, |
| "learning_rate": 0.00011894960965223563, |
| "loss": 0.9525, |
| "step": 1144 |
| }, |
| { |
| "epoch": 1.62, |
| "grad_norm": 0.722373902797699, |
| "learning_rate": 0.00011887863733144074, |
| "loss": 1.047, |
| "step": 1145 |
| }, |
| { |
| "epoch": 1.62, |
| "grad_norm": 0.7318529486656189, |
| "learning_rate": 0.00011880766501064586, |
| "loss": 0.9955, |
| "step": 1146 |
| }, |
| { |
| "epoch": 1.62, |
| "grad_norm": 0.7528226375579834, |
| "learning_rate": 0.00011873669268985096, |
| "loss": 0.926, |
| "step": 1147 |
| }, |
| { |
| "epoch": 1.63, |
| "grad_norm": 0.7202761173248291, |
| "learning_rate": 0.00011866572036905608, |
| "loss": 0.9099, |
| "step": 1148 |
| }, |
| { |
| "epoch": 1.63, |
| "grad_norm": 0.7594791054725647, |
| "learning_rate": 0.00011859474804826119, |
| "loss": 1.0122, |
| "step": 1149 |
| }, |
| { |
| "epoch": 1.63, |
| "grad_norm": 0.7222248315811157, |
| "learning_rate": 0.00011852377572746629, |
| "loss": 0.9161, |
| "step": 1150 |
| }, |
| { |
| "epoch": 1.63, |
| "grad_norm": 0.7424381971359253, |
| "learning_rate": 0.0001184528034066714, |
| "loss": 0.9357, |
| "step": 1151 |
| }, |
| { |
| "epoch": 1.63, |
| "grad_norm": 0.7474575638771057, |
| "learning_rate": 0.0001183818310858765, |
| "loss": 0.9426, |
| "step": 1152 |
| }, |
| { |
| "epoch": 1.63, |
| "grad_norm": 0.7351663112640381, |
| "learning_rate": 0.00011831085876508164, |
| "loss": 0.954, |
| "step": 1153 |
| }, |
| { |
| "epoch": 1.63, |
| "grad_norm": 0.6973652839660645, |
| "learning_rate": 0.00011823988644428674, |
| "loss": 0.9697, |
| "step": 1154 |
| }, |
| { |
| "epoch": 1.64, |
| "grad_norm": 0.7030205726623535, |
| "learning_rate": 0.00011816891412349185, |
| "loss": 0.9262, |
| "step": 1155 |
| }, |
| { |
| "epoch": 1.64, |
| "grad_norm": 0.6983229517936707, |
| "learning_rate": 0.00011809794180269695, |
| "loss": 0.9387, |
| "step": 1156 |
| }, |
| { |
| "epoch": 1.64, |
| "grad_norm": 0.7123252153396606, |
| "learning_rate": 0.00011802696948190206, |
| "loss": 1.0255, |
| "step": 1157 |
| }, |
| { |
| "epoch": 1.64, |
| "grad_norm": 0.7124939560890198, |
| "learning_rate": 0.00011795599716110716, |
| "loss": 0.953, |
| "step": 1158 |
| }, |
| { |
| "epoch": 1.64, |
| "grad_norm": 0.6996143460273743, |
| "learning_rate": 0.00011788502484031228, |
| "loss": 0.9785, |
| "step": 1159 |
| }, |
| { |
| "epoch": 1.64, |
| "grad_norm": 0.7493870854377747, |
| "learning_rate": 0.0001178140525195174, |
| "loss": 0.9201, |
| "step": 1160 |
| }, |
| { |
| "epoch": 1.64, |
| "grad_norm": 0.7479891777038574, |
| "learning_rate": 0.00011774308019872251, |
| "loss": 0.9566, |
| "step": 1161 |
| }, |
| { |
| "epoch": 1.65, |
| "grad_norm": 0.7288089394569397, |
| "learning_rate": 0.00011767210787792761, |
| "loss": 1.004, |
| "step": 1162 |
| }, |
| { |
| "epoch": 1.65, |
| "grad_norm": 0.7156316041946411, |
| "learning_rate": 0.00011760113555713273, |
| "loss": 0.9416, |
| "step": 1163 |
| }, |
| { |
| "epoch": 1.65, |
| "grad_norm": 0.7374414801597595, |
| "learning_rate": 0.00011753016323633783, |
| "loss": 1.0313, |
| "step": 1164 |
| }, |
| { |
| "epoch": 1.65, |
| "grad_norm": 0.6988320350646973, |
| "learning_rate": 0.00011745919091554294, |
| "loss": 0.9442, |
| "step": 1165 |
| }, |
| { |
| "epoch": 1.65, |
| "grad_norm": 0.7388788461685181, |
| "learning_rate": 0.00011738821859474805, |
| "loss": 1.0254, |
| "step": 1166 |
| }, |
| { |
| "epoch": 1.65, |
| "grad_norm": 0.6990705132484436, |
| "learning_rate": 0.00011731724627395318, |
| "loss": 0.9383, |
| "step": 1167 |
| }, |
| { |
| "epoch": 1.65, |
| "grad_norm": 0.698650598526001, |
| "learning_rate": 0.00011724627395315828, |
| "loss": 0.9462, |
| "step": 1168 |
| }, |
| { |
| "epoch": 1.66, |
| "grad_norm": 0.7219197154045105, |
| "learning_rate": 0.00011717530163236339, |
| "loss": 0.9926, |
| "step": 1169 |
| }, |
| { |
| "epoch": 1.66, |
| "grad_norm": 0.7224728465080261, |
| "learning_rate": 0.00011710432931156849, |
| "loss": 0.9425, |
| "step": 1170 |
| }, |
| { |
| "epoch": 1.66, |
| "grad_norm": 0.7238569259643555, |
| "learning_rate": 0.0001170333569907736, |
| "loss": 0.9901, |
| "step": 1171 |
| }, |
| { |
| "epoch": 1.66, |
| "grad_norm": 0.7193832993507385, |
| "learning_rate": 0.00011696238466997871, |
| "loss": 0.9147, |
| "step": 1172 |
| }, |
| { |
| "epoch": 1.66, |
| "grad_norm": 0.7369004487991333, |
| "learning_rate": 0.00011689141234918381, |
| "loss": 0.9768, |
| "step": 1173 |
| }, |
| { |
| "epoch": 1.66, |
| "grad_norm": 0.7380697131156921, |
| "learning_rate": 0.00011682044002838894, |
| "loss": 0.9593, |
| "step": 1174 |
| }, |
| { |
| "epoch": 1.66, |
| "grad_norm": 0.737091064453125, |
| "learning_rate": 0.00011674946770759405, |
| "loss": 1.022, |
| "step": 1175 |
| }, |
| { |
| "epoch": 1.67, |
| "grad_norm": 0.7183380722999573, |
| "learning_rate": 0.00011667849538679916, |
| "loss": 0.9548, |
| "step": 1176 |
| }, |
| { |
| "epoch": 1.67, |
| "grad_norm": 0.7147498726844788, |
| "learning_rate": 0.00011660752306600426, |
| "loss": 0.9307, |
| "step": 1177 |
| }, |
| { |
| "epoch": 1.67, |
| "grad_norm": 0.7256461381912231, |
| "learning_rate": 0.00011653655074520938, |
| "loss": 0.9884, |
| "step": 1178 |
| }, |
| { |
| "epoch": 1.67, |
| "grad_norm": 0.701350212097168, |
| "learning_rate": 0.00011646557842441448, |
| "loss": 0.9209, |
| "step": 1179 |
| }, |
| { |
| "epoch": 1.67, |
| "grad_norm": 0.7127419710159302, |
| "learning_rate": 0.00011639460610361959, |
| "loss": 0.9475, |
| "step": 1180 |
| }, |
| { |
| "epoch": 1.67, |
| "grad_norm": 0.7424458265304565, |
| "learning_rate": 0.00011632363378282471, |
| "loss": 1.0199, |
| "step": 1181 |
| }, |
| { |
| "epoch": 1.67, |
| "grad_norm": 0.7015305757522583, |
| "learning_rate": 0.00011625266146202983, |
| "loss": 0.9896, |
| "step": 1182 |
| }, |
| { |
| "epoch": 1.68, |
| "grad_norm": 0.7185436487197876, |
| "learning_rate": 0.00011618168914123493, |
| "loss": 0.9676, |
| "step": 1183 |
| }, |
| { |
| "epoch": 1.68, |
| "grad_norm": 0.7177726030349731, |
| "learning_rate": 0.00011611071682044004, |
| "loss": 0.973, |
| "step": 1184 |
| }, |
| { |
| "epoch": 1.68, |
| "grad_norm": 0.7331061959266663, |
| "learning_rate": 0.00011603974449964514, |
| "loss": 0.9516, |
| "step": 1185 |
| }, |
| { |
| "epoch": 1.68, |
| "grad_norm": 0.7594481706619263, |
| "learning_rate": 0.00011596877217885025, |
| "loss": 1.011, |
| "step": 1186 |
| }, |
| { |
| "epoch": 1.68, |
| "grad_norm": 0.723116397857666, |
| "learning_rate": 0.00011589779985805535, |
| "loss": 0.9492, |
| "step": 1187 |
| }, |
| { |
| "epoch": 1.68, |
| "grad_norm": 0.773428738117218, |
| "learning_rate": 0.00011582682753726049, |
| "loss": 0.9604, |
| "step": 1188 |
| }, |
| { |
| "epoch": 1.68, |
| "grad_norm": 0.7496587634086609, |
| "learning_rate": 0.00011575585521646559, |
| "loss": 1.0742, |
| "step": 1189 |
| }, |
| { |
| "epoch": 1.69, |
| "grad_norm": 0.7052460312843323, |
| "learning_rate": 0.0001156848828956707, |
| "loss": 0.9653, |
| "step": 1190 |
| }, |
| { |
| "epoch": 1.69, |
| "grad_norm": 0.716080904006958, |
| "learning_rate": 0.0001156139105748758, |
| "loss": 1.0207, |
| "step": 1191 |
| }, |
| { |
| "epoch": 1.69, |
| "grad_norm": 0.7135874032974243, |
| "learning_rate": 0.00011554293825408091, |
| "loss": 0.9421, |
| "step": 1192 |
| }, |
| { |
| "epoch": 1.69, |
| "grad_norm": 0.7133678197860718, |
| "learning_rate": 0.00011547196593328601, |
| "loss": 0.9889, |
| "step": 1193 |
| }, |
| { |
| "epoch": 1.69, |
| "grad_norm": 0.711086630821228, |
| "learning_rate": 0.00011540099361249113, |
| "loss": 0.9795, |
| "step": 1194 |
| }, |
| { |
| "epoch": 1.69, |
| "grad_norm": 0.716465413570404, |
| "learning_rate": 0.00011533002129169624, |
| "loss": 1.0216, |
| "step": 1195 |
| }, |
| { |
| "epoch": 1.69, |
| "grad_norm": 0.7101050615310669, |
| "learning_rate": 0.00011525904897090137, |
| "loss": 0.965, |
| "step": 1196 |
| }, |
| { |
| "epoch": 1.7, |
| "grad_norm": 0.7278016209602356, |
| "learning_rate": 0.00011518807665010646, |
| "loss": 0.9736, |
| "step": 1197 |
| }, |
| { |
| "epoch": 1.7, |
| "grad_norm": 0.6872366666793823, |
| "learning_rate": 0.00011511710432931158, |
| "loss": 0.9184, |
| "step": 1198 |
| }, |
| { |
| "epoch": 1.7, |
| "grad_norm": 0.6575968861579895, |
| "learning_rate": 0.00011504613200851668, |
| "loss": 0.8861, |
| "step": 1199 |
| }, |
| { |
| "epoch": 1.7, |
| "grad_norm": 0.7604357004165649, |
| "learning_rate": 0.00011497515968772179, |
| "loss": 0.9599, |
| "step": 1200 |
| }, |
| { |
| "epoch": 1.7, |
| "grad_norm": 0.7085766792297363, |
| "learning_rate": 0.0001149041873669269, |
| "loss": 0.9557, |
| "step": 1201 |
| }, |
| { |
| "epoch": 1.7, |
| "grad_norm": 0.7655233144760132, |
| "learning_rate": 0.000114833215046132, |
| "loss": 0.969, |
| "step": 1202 |
| }, |
| { |
| "epoch": 1.7, |
| "grad_norm": 0.7449689507484436, |
| "learning_rate": 0.00011476224272533713, |
| "loss": 1.0241, |
| "step": 1203 |
| }, |
| { |
| "epoch": 1.71, |
| "grad_norm": 0.7180941700935364, |
| "learning_rate": 0.00011469127040454224, |
| "loss": 1.0062, |
| "step": 1204 |
| }, |
| { |
| "epoch": 1.71, |
| "grad_norm": 0.7065799236297607, |
| "learning_rate": 0.00011462029808374735, |
| "loss": 0.8982, |
| "step": 1205 |
| }, |
| { |
| "epoch": 1.71, |
| "grad_norm": 0.7072206735610962, |
| "learning_rate": 0.00011454932576295245, |
| "loss": 0.9872, |
| "step": 1206 |
| }, |
| { |
| "epoch": 1.71, |
| "grad_norm": 0.7332132458686829, |
| "learning_rate": 0.00011447835344215756, |
| "loss": 1.0426, |
| "step": 1207 |
| }, |
| { |
| "epoch": 1.71, |
| "grad_norm": 0.6793428063392639, |
| "learning_rate": 0.00011440738112136266, |
| "loss": 1.003, |
| "step": 1208 |
| }, |
| { |
| "epoch": 1.71, |
| "grad_norm": 0.6779395341873169, |
| "learning_rate": 0.00011433640880056778, |
| "loss": 0.8887, |
| "step": 1209 |
| }, |
| { |
| "epoch": 1.71, |
| "grad_norm": 0.7192417979240417, |
| "learning_rate": 0.0001142654364797729, |
| "loss": 1.0065, |
| "step": 1210 |
| }, |
| { |
| "epoch": 1.72, |
| "grad_norm": 0.7286444902420044, |
| "learning_rate": 0.00011419446415897802, |
| "loss": 0.9387, |
| "step": 1211 |
| }, |
| { |
| "epoch": 1.72, |
| "grad_norm": 0.7276972532272339, |
| "learning_rate": 0.00011412349183818311, |
| "loss": 0.9405, |
| "step": 1212 |
| }, |
| { |
| "epoch": 1.72, |
| "grad_norm": 0.7740305066108704, |
| "learning_rate": 0.00011405251951738823, |
| "loss": 1.0154, |
| "step": 1213 |
| }, |
| { |
| "epoch": 1.72, |
| "grad_norm": 0.7302063703536987, |
| "learning_rate": 0.00011398154719659333, |
| "loss": 0.9258, |
| "step": 1214 |
| }, |
| { |
| "epoch": 1.72, |
| "grad_norm": 0.7404993772506714, |
| "learning_rate": 0.00011391057487579844, |
| "loss": 0.9767, |
| "step": 1215 |
| }, |
| { |
| "epoch": 1.72, |
| "grad_norm": 0.7460538744926453, |
| "learning_rate": 0.00011383960255500354, |
| "loss": 0.9811, |
| "step": 1216 |
| }, |
| { |
| "epoch": 1.72, |
| "grad_norm": 0.7067095041275024, |
| "learning_rate": 0.00011376863023420868, |
| "loss": 0.923, |
| "step": 1217 |
| }, |
| { |
| "epoch": 1.73, |
| "grad_norm": 0.7123129963874817, |
| "learning_rate": 0.00011369765791341378, |
| "loss": 1.0037, |
| "step": 1218 |
| }, |
| { |
| "epoch": 1.73, |
| "grad_norm": 0.6522353291511536, |
| "learning_rate": 0.00011362668559261889, |
| "loss": 0.9337, |
| "step": 1219 |
| }, |
| { |
| "epoch": 1.73, |
| "grad_norm": 0.7282575368881226, |
| "learning_rate": 0.00011355571327182399, |
| "loss": 1.0617, |
| "step": 1220 |
| }, |
| { |
| "epoch": 1.73, |
| "grad_norm": 0.6829158067703247, |
| "learning_rate": 0.0001134847409510291, |
| "loss": 0.8669, |
| "step": 1221 |
| }, |
| { |
| "epoch": 1.73, |
| "grad_norm": 0.7063448429107666, |
| "learning_rate": 0.0001134137686302342, |
| "loss": 0.931, |
| "step": 1222 |
| }, |
| { |
| "epoch": 1.73, |
| "grad_norm": 0.7179009318351746, |
| "learning_rate": 0.00011334279630943931, |
| "loss": 1.0027, |
| "step": 1223 |
| }, |
| { |
| "epoch": 1.73, |
| "grad_norm": 0.7192047834396362, |
| "learning_rate": 0.00011327182398864444, |
| "loss": 0.9713, |
| "step": 1224 |
| }, |
| { |
| "epoch": 1.74, |
| "grad_norm": 0.7517329454421997, |
| "learning_rate": 0.00011320085166784955, |
| "loss": 1.0261, |
| "step": 1225 |
| }, |
| { |
| "epoch": 1.74, |
| "grad_norm": 0.7560898661613464, |
| "learning_rate": 0.00011312987934705465, |
| "loss": 0.9925, |
| "step": 1226 |
| }, |
| { |
| "epoch": 1.74, |
| "grad_norm": 0.7354062795639038, |
| "learning_rate": 0.00011305890702625976, |
| "loss": 0.9544, |
| "step": 1227 |
| }, |
| { |
| "epoch": 1.74, |
| "grad_norm": 0.7111262679100037, |
| "learning_rate": 0.00011298793470546488, |
| "loss": 1.0094, |
| "step": 1228 |
| }, |
| { |
| "epoch": 1.74, |
| "grad_norm": 0.7077128291130066, |
| "learning_rate": 0.00011291696238466998, |
| "loss": 0.9743, |
| "step": 1229 |
| }, |
| { |
| "epoch": 1.74, |
| "grad_norm": 0.7007378339767456, |
| "learning_rate": 0.00011284599006387509, |
| "loss": 0.9428, |
| "step": 1230 |
| }, |
| { |
| "epoch": 1.74, |
| "grad_norm": 0.7295317053794861, |
| "learning_rate": 0.00011277501774308022, |
| "loss": 0.96, |
| "step": 1231 |
| }, |
| { |
| "epoch": 1.75, |
| "grad_norm": 0.7037018537521362, |
| "learning_rate": 0.00011270404542228531, |
| "loss": 0.939, |
| "step": 1232 |
| }, |
| { |
| "epoch": 1.75, |
| "grad_norm": 0.7029048204421997, |
| "learning_rate": 0.00011263307310149043, |
| "loss": 0.932, |
| "step": 1233 |
| }, |
| { |
| "epoch": 1.75, |
| "grad_norm": 0.711932361125946, |
| "learning_rate": 0.00011256210078069554, |
| "loss": 0.9009, |
| "step": 1234 |
| }, |
| { |
| "epoch": 1.75, |
| "grad_norm": 0.7566869854927063, |
| "learning_rate": 0.00011249112845990064, |
| "loss": 0.9681, |
| "step": 1235 |
| }, |
| { |
| "epoch": 1.75, |
| "grad_norm": 0.7486440539360046, |
| "learning_rate": 0.00011242015613910575, |
| "loss": 1.0618, |
| "step": 1236 |
| }, |
| { |
| "epoch": 1.75, |
| "grad_norm": 0.7305347323417664, |
| "learning_rate": 0.00011234918381831085, |
| "loss": 0.9537, |
| "step": 1237 |
| }, |
| { |
| "epoch": 1.75, |
| "grad_norm": 0.7390864491462708, |
| "learning_rate": 0.00011227821149751598, |
| "loss": 1.0045, |
| "step": 1238 |
| }, |
| { |
| "epoch": 1.76, |
| "grad_norm": 0.7143562436103821, |
| "learning_rate": 0.00011220723917672109, |
| "loss": 0.9666, |
| "step": 1239 |
| }, |
| { |
| "epoch": 1.76, |
| "grad_norm": 0.7177388668060303, |
| "learning_rate": 0.0001121362668559262, |
| "loss": 1.0068, |
| "step": 1240 |
| }, |
| { |
| "epoch": 1.76, |
| "grad_norm": 0.6928455233573914, |
| "learning_rate": 0.0001120652945351313, |
| "loss": 0.9245, |
| "step": 1241 |
| }, |
| { |
| "epoch": 1.76, |
| "grad_norm": 0.6976019144058228, |
| "learning_rate": 0.00011199432221433641, |
| "loss": 0.9042, |
| "step": 1242 |
| }, |
| { |
| "epoch": 1.76, |
| "grad_norm": 0.7117376327514648, |
| "learning_rate": 0.00011192334989354151, |
| "loss": 0.9605, |
| "step": 1243 |
| }, |
| { |
| "epoch": 1.76, |
| "grad_norm": 0.7036189436912537, |
| "learning_rate": 0.00011185237757274663, |
| "loss": 0.8431, |
| "step": 1244 |
| }, |
| { |
| "epoch": 1.76, |
| "grad_norm": 0.7574560046195984, |
| "learning_rate": 0.00011178140525195175, |
| "loss": 1.0275, |
| "step": 1245 |
| }, |
| { |
| "epoch": 1.77, |
| "grad_norm": 0.7326035499572754, |
| "learning_rate": 0.00011171043293115687, |
| "loss": 0.9939, |
| "step": 1246 |
| }, |
| { |
| "epoch": 1.77, |
| "grad_norm": 0.7244006395339966, |
| "learning_rate": 0.00011163946061036197, |
| "loss": 0.9466, |
| "step": 1247 |
| }, |
| { |
| "epoch": 1.77, |
| "grad_norm": 0.7348788380622864, |
| "learning_rate": 0.00011156848828956708, |
| "loss": 0.942, |
| "step": 1248 |
| }, |
| { |
| "epoch": 1.77, |
| "grad_norm": 0.7395305037498474, |
| "learning_rate": 0.00011149751596877218, |
| "loss": 1.0074, |
| "step": 1249 |
| }, |
| { |
| "epoch": 1.77, |
| "grad_norm": 0.7284185886383057, |
| "learning_rate": 0.00011142654364797729, |
| "loss": 0.9612, |
| "step": 1250 |
| }, |
| { |
| "epoch": 1.77, |
| "grad_norm": 0.6985504031181335, |
| "learning_rate": 0.00011135557132718239, |
| "loss": 0.9172, |
| "step": 1251 |
| }, |
| { |
| "epoch": 1.77, |
| "grad_norm": 0.7086307406425476, |
| "learning_rate": 0.00011128459900638753, |
| "loss": 1.0036, |
| "step": 1252 |
| }, |
| { |
| "epoch": 1.78, |
| "grad_norm": 0.7150089144706726, |
| "learning_rate": 0.00011121362668559263, |
| "loss": 0.9534, |
| "step": 1253 |
| }, |
| { |
| "epoch": 1.78, |
| "grad_norm": 0.7261559963226318, |
| "learning_rate": 0.00011114265436479774, |
| "loss": 0.9956, |
| "step": 1254 |
| }, |
| { |
| "epoch": 1.78, |
| "grad_norm": 0.7401515245437622, |
| "learning_rate": 0.00011107168204400284, |
| "loss": 0.959, |
| "step": 1255 |
| }, |
| { |
| "epoch": 1.78, |
| "grad_norm": 0.732208788394928, |
| "learning_rate": 0.00011100070972320795, |
| "loss": 1.0415, |
| "step": 1256 |
| }, |
| { |
| "epoch": 1.78, |
| "grad_norm": 0.7246513366699219, |
| "learning_rate": 0.00011092973740241307, |
| "loss": 0.9867, |
| "step": 1257 |
| }, |
| { |
| "epoch": 1.78, |
| "grad_norm": 0.7390275001525879, |
| "learning_rate": 0.00011085876508161816, |
| "loss": 0.9737, |
| "step": 1258 |
| }, |
| { |
| "epoch": 1.78, |
| "grad_norm": 0.7073413133621216, |
| "learning_rate": 0.00011078779276082329, |
| "loss": 1.011, |
| "step": 1259 |
| }, |
| { |
| "epoch": 1.79, |
| "grad_norm": 0.7103257775306702, |
| "learning_rate": 0.0001107168204400284, |
| "loss": 0.9457, |
| "step": 1260 |
| }, |
| { |
| "epoch": 1.79, |
| "grad_norm": 0.7341899871826172, |
| "learning_rate": 0.0001106458481192335, |
| "loss": 0.9708, |
| "step": 1261 |
| }, |
| { |
| "epoch": 1.79, |
| "grad_norm": 0.743951141834259, |
| "learning_rate": 0.00011057487579843862, |
| "loss": 0.9338, |
| "step": 1262 |
| }, |
| { |
| "epoch": 1.79, |
| "grad_norm": 0.7139862775802612, |
| "learning_rate": 0.00011050390347764373, |
| "loss": 0.9012, |
| "step": 1263 |
| }, |
| { |
| "epoch": 1.79, |
| "grad_norm": 0.802229642868042, |
| "learning_rate": 0.00011043293115684883, |
| "loss": 0.9703, |
| "step": 1264 |
| }, |
| { |
| "epoch": 1.79, |
| "grad_norm": 0.7592867612838745, |
| "learning_rate": 0.00011036195883605394, |
| "loss": 1.0162, |
| "step": 1265 |
| }, |
| { |
| "epoch": 1.79, |
| "grad_norm": 0.7308145761489868, |
| "learning_rate": 0.00011029098651525907, |
| "loss": 0.9908, |
| "step": 1266 |
| }, |
| { |
| "epoch": 1.79, |
| "grad_norm": 0.764416515827179, |
| "learning_rate": 0.00011022001419446417, |
| "loss": 1.0165, |
| "step": 1267 |
| }, |
| { |
| "epoch": 1.8, |
| "grad_norm": 0.7167296409606934, |
| "learning_rate": 0.00011014904187366928, |
| "loss": 1.0095, |
| "step": 1268 |
| }, |
| { |
| "epoch": 1.8, |
| "grad_norm": 0.6738303899765015, |
| "learning_rate": 0.00011007806955287439, |
| "loss": 0.8957, |
| "step": 1269 |
| }, |
| { |
| "epoch": 1.8, |
| "grad_norm": 0.7134904265403748, |
| "learning_rate": 0.00011000709723207949, |
| "loss": 0.9553, |
| "step": 1270 |
| }, |
| { |
| "epoch": 1.8, |
| "grad_norm": 0.6910573840141296, |
| "learning_rate": 0.0001099361249112846, |
| "loss": 0.9147, |
| "step": 1271 |
| }, |
| { |
| "epoch": 1.8, |
| "grad_norm": 0.7500548958778381, |
| "learning_rate": 0.0001098651525904897, |
| "loss": 1.0135, |
| "step": 1272 |
| }, |
| { |
| "epoch": 1.8, |
| "grad_norm": 0.7269773483276367, |
| "learning_rate": 0.00010979418026969484, |
| "loss": 1.0022, |
| "step": 1273 |
| }, |
| { |
| "epoch": 1.8, |
| "grad_norm": 0.7309441566467285, |
| "learning_rate": 0.00010972320794889994, |
| "loss": 0.9536, |
| "step": 1274 |
| }, |
| { |
| "epoch": 1.81, |
| "grad_norm": 0.7545249462127686, |
| "learning_rate": 0.00010965223562810505, |
| "loss": 0.9793, |
| "step": 1275 |
| }, |
| { |
| "epoch": 1.81, |
| "grad_norm": 0.7531616687774658, |
| "learning_rate": 0.00010958126330731015, |
| "loss": 0.8926, |
| "step": 1276 |
| }, |
| { |
| "epoch": 1.81, |
| "grad_norm": 0.7523400187492371, |
| "learning_rate": 0.00010951029098651527, |
| "loss": 0.9868, |
| "step": 1277 |
| }, |
| { |
| "epoch": 1.81, |
| "grad_norm": 0.7421838045120239, |
| "learning_rate": 0.00010943931866572036, |
| "loss": 0.9743, |
| "step": 1278 |
| }, |
| { |
| "epoch": 1.81, |
| "grad_norm": 0.7826520204544067, |
| "learning_rate": 0.00010936834634492548, |
| "loss": 1.042, |
| "step": 1279 |
| }, |
| { |
| "epoch": 1.81, |
| "grad_norm": 0.7460055351257324, |
| "learning_rate": 0.00010929737402413058, |
| "loss": 0.9605, |
| "step": 1280 |
| }, |
| { |
| "epoch": 1.81, |
| "grad_norm": 0.723533570766449, |
| "learning_rate": 0.00010922640170333572, |
| "loss": 1.0004, |
| "step": 1281 |
| }, |
| { |
| "epoch": 1.82, |
| "grad_norm": 0.7450777888298035, |
| "learning_rate": 0.00010915542938254082, |
| "loss": 1.058, |
| "step": 1282 |
| }, |
| { |
| "epoch": 1.82, |
| "grad_norm": 0.7143047451972961, |
| "learning_rate": 0.00010908445706174593, |
| "loss": 1.0019, |
| "step": 1283 |
| }, |
| { |
| "epoch": 1.82, |
| "grad_norm": 0.6794545650482178, |
| "learning_rate": 0.00010901348474095103, |
| "loss": 0.9604, |
| "step": 1284 |
| }, |
| { |
| "epoch": 1.82, |
| "grad_norm": 0.6825183629989624, |
| "learning_rate": 0.00010894251242015614, |
| "loss": 0.935, |
| "step": 1285 |
| }, |
| { |
| "epoch": 1.82, |
| "grad_norm": 0.6939642429351807, |
| "learning_rate": 0.00010887154009936125, |
| "loss": 1.0092, |
| "step": 1286 |
| }, |
| { |
| "epoch": 1.82, |
| "grad_norm": 0.7390894293785095, |
| "learning_rate": 0.00010880056777856635, |
| "loss": 1.0106, |
| "step": 1287 |
| }, |
| { |
| "epoch": 1.82, |
| "grad_norm": 0.740071177482605, |
| "learning_rate": 0.00010872959545777148, |
| "loss": 0.9024, |
| "step": 1288 |
| }, |
| { |
| "epoch": 1.83, |
| "grad_norm": 0.7248451709747314, |
| "learning_rate": 0.00010865862313697659, |
| "loss": 0.9534, |
| "step": 1289 |
| }, |
| { |
| "epoch": 1.83, |
| "grad_norm": 0.7937654852867126, |
| "learning_rate": 0.00010858765081618169, |
| "loss": 1.022, |
| "step": 1290 |
| }, |
| { |
| "epoch": 1.83, |
| "grad_norm": 0.7495038509368896, |
| "learning_rate": 0.0001085166784953868, |
| "loss": 0.8344, |
| "step": 1291 |
| }, |
| { |
| "epoch": 1.83, |
| "grad_norm": 0.7504763603210449, |
| "learning_rate": 0.00010844570617459192, |
| "loss": 0.9882, |
| "step": 1292 |
| }, |
| { |
| "epoch": 1.83, |
| "grad_norm": 0.7662994861602783, |
| "learning_rate": 0.00010837473385379701, |
| "loss": 0.9283, |
| "step": 1293 |
| }, |
| { |
| "epoch": 1.83, |
| "grad_norm": 0.7336791157722473, |
| "learning_rate": 0.00010830376153300213, |
| "loss": 0.9955, |
| "step": 1294 |
| }, |
| { |
| "epoch": 1.83, |
| "grad_norm": 0.7345770597457886, |
| "learning_rate": 0.00010823278921220725, |
| "loss": 0.9991, |
| "step": 1295 |
| }, |
| { |
| "epoch": 1.84, |
| "grad_norm": 0.6934396624565125, |
| "learning_rate": 0.00010816181689141235, |
| "loss": 0.8826, |
| "step": 1296 |
| }, |
| { |
| "epoch": 1.84, |
| "grad_norm": 0.7034685611724854, |
| "learning_rate": 0.00010809084457061747, |
| "loss": 0.9665, |
| "step": 1297 |
| }, |
| { |
| "epoch": 1.84, |
| "grad_norm": 0.6718670725822449, |
| "learning_rate": 0.00010801987224982258, |
| "loss": 0.8808, |
| "step": 1298 |
| }, |
| { |
| "epoch": 1.84, |
| "grad_norm": 0.7105706334114075, |
| "learning_rate": 0.00010794889992902768, |
| "loss": 0.9399, |
| "step": 1299 |
| }, |
| { |
| "epoch": 1.84, |
| "grad_norm": 0.7107860445976257, |
| "learning_rate": 0.00010787792760823279, |
| "loss": 0.9766, |
| "step": 1300 |
| }, |
| { |
| "epoch": 1.84, |
| "grad_norm": 0.7587505578994751, |
| "learning_rate": 0.00010780695528743789, |
| "loss": 0.993, |
| "step": 1301 |
| }, |
| { |
| "epoch": 1.84, |
| "grad_norm": 0.7261898517608643, |
| "learning_rate": 0.00010773598296664303, |
| "loss": 0.9075, |
| "step": 1302 |
| }, |
| { |
| "epoch": 1.85, |
| "grad_norm": 0.712391197681427, |
| "learning_rate": 0.00010766501064584813, |
| "loss": 0.991, |
| "step": 1303 |
| }, |
| { |
| "epoch": 1.85, |
| "grad_norm": 0.7336340546607971, |
| "learning_rate": 0.00010759403832505324, |
| "loss": 1.036, |
| "step": 1304 |
| }, |
| { |
| "epoch": 1.85, |
| "grad_norm": 0.7203515768051147, |
| "learning_rate": 0.00010752306600425834, |
| "loss": 1.0046, |
| "step": 1305 |
| }, |
| { |
| "epoch": 1.85, |
| "grad_norm": 0.723571240901947, |
| "learning_rate": 0.00010745209368346345, |
| "loss": 0.9865, |
| "step": 1306 |
| }, |
| { |
| "epoch": 1.85, |
| "grad_norm": 0.7366999983787537, |
| "learning_rate": 0.00010738112136266855, |
| "loss": 0.9563, |
| "step": 1307 |
| }, |
| { |
| "epoch": 1.85, |
| "grad_norm": 0.7270590662956238, |
| "learning_rate": 0.00010731014904187367, |
| "loss": 0.9038, |
| "step": 1308 |
| }, |
| { |
| "epoch": 1.85, |
| "grad_norm": 0.7648587226867676, |
| "learning_rate": 0.00010723917672107879, |
| "loss": 1.0028, |
| "step": 1309 |
| }, |
| { |
| "epoch": 1.86, |
| "grad_norm": 0.6916332244873047, |
| "learning_rate": 0.0001071682044002839, |
| "loss": 0.8921, |
| "step": 1310 |
| }, |
| { |
| "epoch": 1.86, |
| "grad_norm": 0.7208322882652283, |
| "learning_rate": 0.000107097232079489, |
| "loss": 0.9974, |
| "step": 1311 |
| }, |
| { |
| "epoch": 1.86, |
| "grad_norm": 0.7530882358551025, |
| "learning_rate": 0.00010702625975869412, |
| "loss": 0.9533, |
| "step": 1312 |
| }, |
| { |
| "epoch": 1.86, |
| "grad_norm": 0.7088322639465332, |
| "learning_rate": 0.00010695528743789922, |
| "loss": 0.9691, |
| "step": 1313 |
| }, |
| { |
| "epoch": 1.86, |
| "grad_norm": 0.7041926383972168, |
| "learning_rate": 0.00010688431511710433, |
| "loss": 0.9699, |
| "step": 1314 |
| }, |
| { |
| "epoch": 1.86, |
| "grad_norm": 0.7026035785675049, |
| "learning_rate": 0.00010681334279630944, |
| "loss": 1.0027, |
| "step": 1315 |
| }, |
| { |
| "epoch": 1.86, |
| "grad_norm": 0.7356511354446411, |
| "learning_rate": 0.00010674237047551457, |
| "loss": 0.9931, |
| "step": 1316 |
| }, |
| { |
| "epoch": 1.87, |
| "grad_norm": 0.7051964402198792, |
| "learning_rate": 0.00010667139815471967, |
| "loss": 0.9606, |
| "step": 1317 |
| }, |
| { |
| "epoch": 1.87, |
| "grad_norm": 0.7253301739692688, |
| "learning_rate": 0.00010660042583392478, |
| "loss": 1.0027, |
| "step": 1318 |
| }, |
| { |
| "epoch": 1.87, |
| "grad_norm": 0.7091968059539795, |
| "learning_rate": 0.00010652945351312988, |
| "loss": 0.9177, |
| "step": 1319 |
| }, |
| { |
| "epoch": 1.87, |
| "grad_norm": 0.7086256146430969, |
| "learning_rate": 0.00010645848119233499, |
| "loss": 0.9631, |
| "step": 1320 |
| }, |
| { |
| "epoch": 1.87, |
| "grad_norm": 0.7075737714767456, |
| "learning_rate": 0.0001063875088715401, |
| "loss": 0.9245, |
| "step": 1321 |
| }, |
| { |
| "epoch": 1.87, |
| "grad_norm": 0.733659565448761, |
| "learning_rate": 0.0001063165365507452, |
| "loss": 0.9595, |
| "step": 1322 |
| }, |
| { |
| "epoch": 1.87, |
| "grad_norm": 0.7235379815101624, |
| "learning_rate": 0.00010624556422995033, |
| "loss": 0.8914, |
| "step": 1323 |
| }, |
| { |
| "epoch": 1.88, |
| "grad_norm": 0.8256986141204834, |
| "learning_rate": 0.00010617459190915544, |
| "loss": 1.031, |
| "step": 1324 |
| }, |
| { |
| "epoch": 1.88, |
| "grad_norm": 0.7533237934112549, |
| "learning_rate": 0.00010610361958836055, |
| "loss": 0.977, |
| "step": 1325 |
| }, |
| { |
| "epoch": 1.88, |
| "grad_norm": 0.7689191699028015, |
| "learning_rate": 0.00010603264726756565, |
| "loss": 0.926, |
| "step": 1326 |
| }, |
| { |
| "epoch": 1.88, |
| "grad_norm": 0.7381359934806824, |
| "learning_rate": 0.00010596167494677077, |
| "loss": 0.9478, |
| "step": 1327 |
| }, |
| { |
| "epoch": 1.88, |
| "grad_norm": 0.7638797163963318, |
| "learning_rate": 0.00010589070262597587, |
| "loss": 0.9694, |
| "step": 1328 |
| }, |
| { |
| "epoch": 1.88, |
| "grad_norm": 0.7295525074005127, |
| "learning_rate": 0.00010581973030518098, |
| "loss": 1.0069, |
| "step": 1329 |
| }, |
| { |
| "epoch": 1.88, |
| "grad_norm": 0.7049389481544495, |
| "learning_rate": 0.0001057487579843861, |
| "loss": 0.9564, |
| "step": 1330 |
| }, |
| { |
| "epoch": 1.89, |
| "grad_norm": 0.7222816944122314, |
| "learning_rate": 0.00010567778566359122, |
| "loss": 0.9999, |
| "step": 1331 |
| }, |
| { |
| "epoch": 1.89, |
| "grad_norm": 0.7252581715583801, |
| "learning_rate": 0.00010560681334279632, |
| "loss": 0.9455, |
| "step": 1332 |
| }, |
| { |
| "epoch": 1.89, |
| "grad_norm": 0.7246282696723938, |
| "learning_rate": 0.00010553584102200143, |
| "loss": 0.9012, |
| "step": 1333 |
| }, |
| { |
| "epoch": 1.89, |
| "grad_norm": 0.6909769177436829, |
| "learning_rate": 0.00010546486870120653, |
| "loss": 0.9417, |
| "step": 1334 |
| }, |
| { |
| "epoch": 1.89, |
| "grad_norm": 0.7185254693031311, |
| "learning_rate": 0.00010539389638041164, |
| "loss": 0.9911, |
| "step": 1335 |
| }, |
| { |
| "epoch": 1.89, |
| "grad_norm": 0.7959221005439758, |
| "learning_rate": 0.00010532292405961674, |
| "loss": 0.9653, |
| "step": 1336 |
| }, |
| { |
| "epoch": 1.89, |
| "grad_norm": 0.7081154584884644, |
| "learning_rate": 0.00010525195173882188, |
| "loss": 0.8621, |
| "step": 1337 |
| }, |
| { |
| "epoch": 1.9, |
| "grad_norm": 0.7107864618301392, |
| "learning_rate": 0.00010518097941802698, |
| "loss": 0.994, |
| "step": 1338 |
| }, |
| { |
| "epoch": 1.9, |
| "grad_norm": 0.7213383316993713, |
| "learning_rate": 0.00010511000709723209, |
| "loss": 0.9479, |
| "step": 1339 |
| }, |
| { |
| "epoch": 1.9, |
| "grad_norm": 0.7340273261070251, |
| "learning_rate": 0.00010503903477643719, |
| "loss": 0.9278, |
| "step": 1340 |
| }, |
| { |
| "epoch": 1.9, |
| "grad_norm": 0.7713385820388794, |
| "learning_rate": 0.0001049680624556423, |
| "loss": 1.0469, |
| "step": 1341 |
| }, |
| { |
| "epoch": 1.9, |
| "grad_norm": 0.6916841268539429, |
| "learning_rate": 0.0001048970901348474, |
| "loss": 0.9564, |
| "step": 1342 |
| }, |
| { |
| "epoch": 1.9, |
| "grad_norm": 0.7080590128898621, |
| "learning_rate": 0.00010482611781405252, |
| "loss": 0.9239, |
| "step": 1343 |
| }, |
| { |
| "epoch": 1.9, |
| "grad_norm": 0.7093188762664795, |
| "learning_rate": 0.00010475514549325764, |
| "loss": 0.9249, |
| "step": 1344 |
| }, |
| { |
| "epoch": 1.91, |
| "grad_norm": 0.7387856841087341, |
| "learning_rate": 0.00010468417317246275, |
| "loss": 1.0172, |
| "step": 1345 |
| }, |
| { |
| "epoch": 1.91, |
| "grad_norm": 0.6952479481697083, |
| "learning_rate": 0.00010461320085166785, |
| "loss": 0.9636, |
| "step": 1346 |
| }, |
| { |
| "epoch": 1.91, |
| "grad_norm": 0.7166315913200378, |
| "learning_rate": 0.00010454222853087297, |
| "loss": 0.9475, |
| "step": 1347 |
| }, |
| { |
| "epoch": 1.91, |
| "grad_norm": 0.7198975682258606, |
| "learning_rate": 0.00010447125621007807, |
| "loss": 0.9229, |
| "step": 1348 |
| }, |
| { |
| "epoch": 1.91, |
| "grad_norm": 0.7340227961540222, |
| "learning_rate": 0.00010440028388928318, |
| "loss": 1.0101, |
| "step": 1349 |
| }, |
| { |
| "epoch": 1.91, |
| "grad_norm": 0.7560737133026123, |
| "learning_rate": 0.00010432931156848829, |
| "loss": 1.0183, |
| "step": 1350 |
| }, |
| { |
| "epoch": 1.91, |
| "grad_norm": 0.7338775396347046, |
| "learning_rate": 0.00010425833924769342, |
| "loss": 0.9869, |
| "step": 1351 |
| }, |
| { |
| "epoch": 1.92, |
| "grad_norm": 0.7228880524635315, |
| "learning_rate": 0.00010418736692689852, |
| "loss": 0.9771, |
| "step": 1352 |
| }, |
| { |
| "epoch": 1.92, |
| "grad_norm": 0.7613531351089478, |
| "learning_rate": 0.00010411639460610363, |
| "loss": 0.9756, |
| "step": 1353 |
| }, |
| { |
| "epoch": 1.92, |
| "grad_norm": 0.7517747282981873, |
| "learning_rate": 0.00010404542228530874, |
| "loss": 0.9873, |
| "step": 1354 |
| }, |
| { |
| "epoch": 1.92, |
| "grad_norm": 0.7492972612380981, |
| "learning_rate": 0.00010397444996451384, |
| "loss": 1.0389, |
| "step": 1355 |
| }, |
| { |
| "epoch": 1.92, |
| "grad_norm": 0.7404029965400696, |
| "learning_rate": 0.00010390347764371895, |
| "loss": 1.0034, |
| "step": 1356 |
| }, |
| { |
| "epoch": 1.92, |
| "grad_norm": 0.6921476125717163, |
| "learning_rate": 0.00010383250532292405, |
| "loss": 0.9379, |
| "step": 1357 |
| }, |
| { |
| "epoch": 1.92, |
| "grad_norm": 0.7141453623771667, |
| "learning_rate": 0.00010376153300212918, |
| "loss": 0.8902, |
| "step": 1358 |
| }, |
| { |
| "epoch": 1.93, |
| "grad_norm": 0.7601017355918884, |
| "learning_rate": 0.00010369056068133429, |
| "loss": 0.9661, |
| "step": 1359 |
| }, |
| { |
| "epoch": 1.93, |
| "grad_norm": 0.7252589464187622, |
| "learning_rate": 0.0001036195883605394, |
| "loss": 0.9692, |
| "step": 1360 |
| }, |
| { |
| "epoch": 1.93, |
| "grad_norm": 0.7792155742645264, |
| "learning_rate": 0.0001035486160397445, |
| "loss": 0.9614, |
| "step": 1361 |
| }, |
| { |
| "epoch": 1.93, |
| "grad_norm": 0.7573446035385132, |
| "learning_rate": 0.00010347764371894962, |
| "loss": 0.8806, |
| "step": 1362 |
| }, |
| { |
| "epoch": 1.93, |
| "grad_norm": 0.7885202169418335, |
| "learning_rate": 0.00010340667139815472, |
| "loss": 0.9797, |
| "step": 1363 |
| }, |
| { |
| "epoch": 1.93, |
| "grad_norm": 0.7697455286979675, |
| "learning_rate": 0.00010333569907735983, |
| "loss": 0.9303, |
| "step": 1364 |
| }, |
| { |
| "epoch": 1.93, |
| "grad_norm": 0.7955215573310852, |
| "learning_rate": 0.00010326472675656495, |
| "loss": 1.0014, |
| "step": 1365 |
| }, |
| { |
| "epoch": 1.94, |
| "grad_norm": 0.7274532318115234, |
| "learning_rate": 0.00010319375443577007, |
| "loss": 0.9392, |
| "step": 1366 |
| }, |
| { |
| "epoch": 1.94, |
| "grad_norm": 0.7063309550285339, |
| "learning_rate": 0.00010312278211497517, |
| "loss": 0.9254, |
| "step": 1367 |
| }, |
| { |
| "epoch": 1.94, |
| "grad_norm": 0.6960466504096985, |
| "learning_rate": 0.00010305180979418028, |
| "loss": 0.9474, |
| "step": 1368 |
| }, |
| { |
| "epoch": 1.94, |
| "grad_norm": 0.6692228317260742, |
| "learning_rate": 0.00010298083747338538, |
| "loss": 0.9358, |
| "step": 1369 |
| }, |
| { |
| "epoch": 1.94, |
| "grad_norm": 0.7357027530670166, |
| "learning_rate": 0.00010290986515259049, |
| "loss": 1.055, |
| "step": 1370 |
| }, |
| { |
| "epoch": 1.94, |
| "grad_norm": 0.7488313317298889, |
| "learning_rate": 0.00010283889283179559, |
| "loss": 0.8772, |
| "step": 1371 |
| }, |
| { |
| "epoch": 1.94, |
| "grad_norm": 0.7428348660469055, |
| "learning_rate": 0.0001027679205110007, |
| "loss": 1.0286, |
| "step": 1372 |
| }, |
| { |
| "epoch": 1.95, |
| "grad_norm": 0.7692130208015442, |
| "learning_rate": 0.00010269694819020583, |
| "loss": 0.9619, |
| "step": 1373 |
| }, |
| { |
| "epoch": 1.95, |
| "grad_norm": 0.748082160949707, |
| "learning_rate": 0.00010262597586941094, |
| "loss": 0.9234, |
| "step": 1374 |
| }, |
| { |
| "epoch": 1.95, |
| "grad_norm": 0.7470769882202148, |
| "learning_rate": 0.00010255500354861604, |
| "loss": 0.9904, |
| "step": 1375 |
| }, |
| { |
| "epoch": 1.95, |
| "grad_norm": 0.8034458160400391, |
| "learning_rate": 0.00010248403122782115, |
| "loss": 1.0794, |
| "step": 1376 |
| }, |
| { |
| "epoch": 1.95, |
| "grad_norm": 0.7386705875396729, |
| "learning_rate": 0.00010241305890702627, |
| "loss": 0.9268, |
| "step": 1377 |
| }, |
| { |
| "epoch": 1.95, |
| "grad_norm": 0.7536996603012085, |
| "learning_rate": 0.00010234208658623137, |
| "loss": 1.0428, |
| "step": 1378 |
| }, |
| { |
| "epoch": 1.95, |
| "grad_norm": 0.753163754940033, |
| "learning_rate": 0.00010227111426543648, |
| "loss": 1.0112, |
| "step": 1379 |
| }, |
| { |
| "epoch": 1.96, |
| "grad_norm": 0.7234356999397278, |
| "learning_rate": 0.0001022001419446416, |
| "loss": 0.9551, |
| "step": 1380 |
| }, |
| { |
| "epoch": 1.96, |
| "grad_norm": 0.7315410375595093, |
| "learning_rate": 0.0001021291696238467, |
| "loss": 0.9997, |
| "step": 1381 |
| }, |
| { |
| "epoch": 1.96, |
| "grad_norm": 0.7085781097412109, |
| "learning_rate": 0.00010205819730305182, |
| "loss": 0.987, |
| "step": 1382 |
| }, |
| { |
| "epoch": 1.96, |
| "grad_norm": 0.7303948998451233, |
| "learning_rate": 0.00010198722498225693, |
| "loss": 0.9763, |
| "step": 1383 |
| }, |
| { |
| "epoch": 1.96, |
| "grad_norm": 0.7143583297729492, |
| "learning_rate": 0.00010191625266146203, |
| "loss": 0.8577, |
| "step": 1384 |
| }, |
| { |
| "epoch": 1.96, |
| "grad_norm": 0.7335658073425293, |
| "learning_rate": 0.00010184528034066714, |
| "loss": 0.9503, |
| "step": 1385 |
| }, |
| { |
| "epoch": 1.96, |
| "grad_norm": 0.7341468334197998, |
| "learning_rate": 0.00010177430801987224, |
| "loss": 0.9232, |
| "step": 1386 |
| }, |
| { |
| "epoch": 1.96, |
| "grad_norm": 0.7755731344223022, |
| "learning_rate": 0.00010170333569907737, |
| "loss": 1.0562, |
| "step": 1387 |
| }, |
| { |
| "epoch": 1.97, |
| "grad_norm": 0.7738588452339172, |
| "learning_rate": 0.00010163236337828248, |
| "loss": 0.9736, |
| "step": 1388 |
| }, |
| { |
| "epoch": 1.97, |
| "grad_norm": 0.7616381645202637, |
| "learning_rate": 0.00010156139105748759, |
| "loss": 0.9838, |
| "step": 1389 |
| }, |
| { |
| "epoch": 1.97, |
| "grad_norm": 0.7650443315505981, |
| "learning_rate": 0.00010149041873669269, |
| "loss": 0.9184, |
| "step": 1390 |
| }, |
| { |
| "epoch": 1.97, |
| "grad_norm": 0.7394657135009766, |
| "learning_rate": 0.0001014194464158978, |
| "loss": 0.8795, |
| "step": 1391 |
| }, |
| { |
| "epoch": 1.97, |
| "grad_norm": 0.6988821625709534, |
| "learning_rate": 0.0001013484740951029, |
| "loss": 0.9545, |
| "step": 1392 |
| }, |
| { |
| "epoch": 1.97, |
| "grad_norm": 0.7445807456970215, |
| "learning_rate": 0.00010127750177430802, |
| "loss": 0.9847, |
| "step": 1393 |
| }, |
| { |
| "epoch": 1.97, |
| "grad_norm": 0.7342961430549622, |
| "learning_rate": 0.00010120652945351314, |
| "loss": 0.9654, |
| "step": 1394 |
| }, |
| { |
| "epoch": 1.98, |
| "grad_norm": 0.726063072681427, |
| "learning_rate": 0.00010113555713271826, |
| "loss": 0.9411, |
| "step": 1395 |
| }, |
| { |
| "epoch": 1.98, |
| "grad_norm": 0.6817072629928589, |
| "learning_rate": 0.00010106458481192335, |
| "loss": 0.9266, |
| "step": 1396 |
| }, |
| { |
| "epoch": 1.98, |
| "grad_norm": 0.717892050743103, |
| "learning_rate": 0.00010099361249112847, |
| "loss": 0.9439, |
| "step": 1397 |
| }, |
| { |
| "epoch": 1.98, |
| "grad_norm": 0.7288515567779541, |
| "learning_rate": 0.00010092264017033357, |
| "loss": 1.0106, |
| "step": 1398 |
| }, |
| { |
| "epoch": 1.98, |
| "grad_norm": 0.7539445757865906, |
| "learning_rate": 0.00010085166784953868, |
| "loss": 1.0806, |
| "step": 1399 |
| }, |
| { |
| "epoch": 1.98, |
| "grad_norm": 0.7196104526519775, |
| "learning_rate": 0.00010078069552874378, |
| "loss": 0.8989, |
| "step": 1400 |
| }, |
| { |
| "epoch": 1.98, |
| "grad_norm": 0.7768797278404236, |
| "learning_rate": 0.00010070972320794892, |
| "loss": 0.8933, |
| "step": 1401 |
| }, |
| { |
| "epoch": 1.99, |
| "grad_norm": 0.7455546855926514, |
| "learning_rate": 0.00010063875088715402, |
| "loss": 1.0315, |
| "step": 1402 |
| }, |
| { |
| "epoch": 1.99, |
| "grad_norm": 0.7538041472434998, |
| "learning_rate": 0.00010056777856635913, |
| "loss": 1.001, |
| "step": 1403 |
| }, |
| { |
| "epoch": 1.99, |
| "grad_norm": 0.7568172216415405, |
| "learning_rate": 0.00010049680624556423, |
| "loss": 1.0073, |
| "step": 1404 |
| }, |
| { |
| "epoch": 1.99, |
| "grad_norm": 0.761991024017334, |
| "learning_rate": 0.00010042583392476934, |
| "loss": 0.9716, |
| "step": 1405 |
| }, |
| { |
| "epoch": 1.99, |
| "grad_norm": 0.7252638339996338, |
| "learning_rate": 0.00010035486160397445, |
| "loss": 0.9636, |
| "step": 1406 |
| }, |
| { |
| "epoch": 1.99, |
| "grad_norm": 0.7555544972419739, |
| "learning_rate": 0.00010028388928317955, |
| "loss": 0.9955, |
| "step": 1407 |
| }, |
| { |
| "epoch": 1.99, |
| "grad_norm": 0.7491613030433655, |
| "learning_rate": 0.00010021291696238468, |
| "loss": 1.0634, |
| "step": 1408 |
| }, |
| { |
| "epoch": 2.0, |
| "grad_norm": 0.7551286220550537, |
| "learning_rate": 0.00010014194464158979, |
| "loss": 0.9722, |
| "step": 1409 |
| }, |
| { |
| "epoch": 2.0, |
| "grad_norm": 0.7532618045806885, |
| "learning_rate": 0.00010007097232079489, |
| "loss": 0.9702, |
| "step": 1410 |
| }, |
| { |
| "epoch": 2.0, |
| "grad_norm": 0.7265384793281555, |
| "learning_rate": 0.0001, |
| "loss": 0.9758, |
| "step": 1411 |
| }, |
| { |
| "epoch": 2.0, |
| "grad_norm": 0.6909193992614746, |
| "learning_rate": 9.992902767920512e-05, |
| "loss": 0.8897, |
| "step": 1412 |
| }, |
| { |
| "epoch": 2.0, |
| "grad_norm": 0.6946807503700256, |
| "learning_rate": 9.985805535841023e-05, |
| "loss": 0.7733, |
| "step": 1413 |
| }, |
| { |
| "epoch": 2.0, |
| "grad_norm": 0.6953454613685608, |
| "learning_rate": 9.978708303761534e-05, |
| "loss": 0.7321, |
| "step": 1414 |
| }, |
| { |
| "epoch": 2.0, |
| "grad_norm": 0.7531546950340271, |
| "learning_rate": 9.971611071682044e-05, |
| "loss": 0.788, |
| "step": 1415 |
| }, |
| { |
| "epoch": 2.01, |
| "grad_norm": 0.8027557730674744, |
| "learning_rate": 9.964513839602555e-05, |
| "loss": 0.6882, |
| "step": 1416 |
| }, |
| { |
| "epoch": 2.01, |
| "grad_norm": 0.9294509291648865, |
| "learning_rate": 9.957416607523067e-05, |
| "loss": 0.7098, |
| "step": 1417 |
| }, |
| { |
| "epoch": 2.01, |
| "grad_norm": 1.0754622220993042, |
| "learning_rate": 9.950319375443578e-05, |
| "loss": 0.716, |
| "step": 1418 |
| }, |
| { |
| "epoch": 2.01, |
| "grad_norm": 0.9674079418182373, |
| "learning_rate": 9.943222143364088e-05, |
| "loss": 0.6679, |
| "step": 1419 |
| }, |
| { |
| "epoch": 2.01, |
| "grad_norm": 0.9014368057250977, |
| "learning_rate": 9.9361249112846e-05, |
| "loss": 0.713, |
| "step": 1420 |
| }, |
| { |
| "epoch": 2.01, |
| "grad_norm": 0.8310858607292175, |
| "learning_rate": 9.92902767920511e-05, |
| "loss": 0.6925, |
| "step": 1421 |
| }, |
| { |
| "epoch": 2.01, |
| "grad_norm": 0.8001790642738342, |
| "learning_rate": 9.921930447125622e-05, |
| "loss": 0.7517, |
| "step": 1422 |
| }, |
| { |
| "epoch": 2.02, |
| "grad_norm": 0.7294822335243225, |
| "learning_rate": 9.914833215046132e-05, |
| "loss": 0.6376, |
| "step": 1423 |
| }, |
| { |
| "epoch": 2.02, |
| "grad_norm": 0.7423235177993774, |
| "learning_rate": 9.907735982966644e-05, |
| "loss": 0.7022, |
| "step": 1424 |
| }, |
| { |
| "epoch": 2.02, |
| "grad_norm": 0.7563339471817017, |
| "learning_rate": 9.900638750887154e-05, |
| "loss": 0.6896, |
| "step": 1425 |
| }, |
| { |
| "epoch": 2.02, |
| "grad_norm": 0.8096082210540771, |
| "learning_rate": 9.893541518807666e-05, |
| "loss": 0.6826, |
| "step": 1426 |
| }, |
| { |
| "epoch": 2.02, |
| "grad_norm": 0.8095359802246094, |
| "learning_rate": 9.886444286728177e-05, |
| "loss": 0.6946, |
| "step": 1427 |
| }, |
| { |
| "epoch": 2.02, |
| "grad_norm": 0.8249964118003845, |
| "learning_rate": 9.879347054648688e-05, |
| "loss": 0.6902, |
| "step": 1428 |
| }, |
| { |
| "epoch": 2.02, |
| "grad_norm": 0.8830640316009521, |
| "learning_rate": 9.872249822569198e-05, |
| "loss": 0.6451, |
| "step": 1429 |
| }, |
| { |
| "epoch": 2.03, |
| "grad_norm": 0.9052359461784363, |
| "learning_rate": 9.865152590489709e-05, |
| "loss": 0.7589, |
| "step": 1430 |
| }, |
| { |
| "epoch": 2.03, |
| "grad_norm": 0.8446089029312134, |
| "learning_rate": 9.85805535841022e-05, |
| "loss": 0.6526, |
| "step": 1431 |
| }, |
| { |
| "epoch": 2.03, |
| "grad_norm": 0.84389728307724, |
| "learning_rate": 9.850958126330732e-05, |
| "loss": 0.7144, |
| "step": 1432 |
| }, |
| { |
| "epoch": 2.03, |
| "grad_norm": 0.8531579971313477, |
| "learning_rate": 9.843860894251242e-05, |
| "loss": 0.7598, |
| "step": 1433 |
| }, |
| { |
| "epoch": 2.03, |
| "grad_norm": 0.7820523977279663, |
| "learning_rate": 9.836763662171753e-05, |
| "loss": 0.666, |
| "step": 1434 |
| }, |
| { |
| "epoch": 2.03, |
| "grad_norm": 0.8026299476623535, |
| "learning_rate": 9.829666430092264e-05, |
| "loss": 0.689, |
| "step": 1435 |
| }, |
| { |
| "epoch": 2.03, |
| "grad_norm": 0.8456664681434631, |
| "learning_rate": 9.822569198012776e-05, |
| "loss": 0.7466, |
| "step": 1436 |
| }, |
| { |
| "epoch": 2.04, |
| "grad_norm": 0.81344074010849, |
| "learning_rate": 9.815471965933285e-05, |
| "loss": 0.6924, |
| "step": 1437 |
| }, |
| { |
| "epoch": 2.04, |
| "grad_norm": 0.8156119585037231, |
| "learning_rate": 9.808374733853798e-05, |
| "loss": 0.6812, |
| "step": 1438 |
| }, |
| { |
| "epoch": 2.04, |
| "grad_norm": 0.8709864020347595, |
| "learning_rate": 9.801277501774308e-05, |
| "loss": 0.7375, |
| "step": 1439 |
| }, |
| { |
| "epoch": 2.04, |
| "grad_norm": 0.8173854947090149, |
| "learning_rate": 9.794180269694819e-05, |
| "loss": 0.7124, |
| "step": 1440 |
| }, |
| { |
| "epoch": 2.04, |
| "grad_norm": 0.8464558124542236, |
| "learning_rate": 9.78708303761533e-05, |
| "loss": 0.7316, |
| "step": 1441 |
| }, |
| { |
| "epoch": 2.04, |
| "grad_norm": 0.8382313251495361, |
| "learning_rate": 9.779985805535842e-05, |
| "loss": 0.6897, |
| "step": 1442 |
| }, |
| { |
| "epoch": 2.04, |
| "grad_norm": 0.8649709820747375, |
| "learning_rate": 9.772888573456353e-05, |
| "loss": 0.7518, |
| "step": 1443 |
| }, |
| { |
| "epoch": 2.05, |
| "grad_norm": 0.7927874326705933, |
| "learning_rate": 9.765791341376863e-05, |
| "loss": 0.7071, |
| "step": 1444 |
| }, |
| { |
| "epoch": 2.05, |
| "grad_norm": 0.8561937808990479, |
| "learning_rate": 9.758694109297374e-05, |
| "loss": 0.7027, |
| "step": 1445 |
| }, |
| { |
| "epoch": 2.05, |
| "grad_norm": 0.8343231678009033, |
| "learning_rate": 9.751596877217886e-05, |
| "loss": 0.7474, |
| "step": 1446 |
| }, |
| { |
| "epoch": 2.05, |
| "grad_norm": 0.8136131763458252, |
| "learning_rate": 9.744499645138397e-05, |
| "loss": 0.7358, |
| "step": 1447 |
| }, |
| { |
| "epoch": 2.05, |
| "grad_norm": 0.8063412308692932, |
| "learning_rate": 9.737402413058907e-05, |
| "loss": 0.7115, |
| "step": 1448 |
| }, |
| { |
| "epoch": 2.05, |
| "grad_norm": 0.8385120630264282, |
| "learning_rate": 9.73030518097942e-05, |
| "loss": 0.7042, |
| "step": 1449 |
| }, |
| { |
| "epoch": 2.05, |
| "grad_norm": 0.8496181964874268, |
| "learning_rate": 9.723207948899929e-05, |
| "loss": 0.7378, |
| "step": 1450 |
| }, |
| { |
| "epoch": 2.06, |
| "grad_norm": 0.8625522255897522, |
| "learning_rate": 9.71611071682044e-05, |
| "loss": 0.69, |
| "step": 1451 |
| }, |
| { |
| "epoch": 2.06, |
| "grad_norm": 0.8408932089805603, |
| "learning_rate": 9.709013484740952e-05, |
| "loss": 0.6602, |
| "step": 1452 |
| }, |
| { |
| "epoch": 2.06, |
| "grad_norm": 0.82053142786026, |
| "learning_rate": 9.701916252661463e-05, |
| "loss": 0.6759, |
| "step": 1453 |
| }, |
| { |
| "epoch": 2.06, |
| "grad_norm": 0.8382329344749451, |
| "learning_rate": 9.694819020581973e-05, |
| "loss": 0.6741, |
| "step": 1454 |
| }, |
| { |
| "epoch": 2.06, |
| "grad_norm": 0.8781008720397949, |
| "learning_rate": 9.687721788502484e-05, |
| "loss": 0.7876, |
| "step": 1455 |
| }, |
| { |
| "epoch": 2.06, |
| "grad_norm": 0.8545266389846802, |
| "learning_rate": 9.680624556422996e-05, |
| "loss": 0.7504, |
| "step": 1456 |
| }, |
| { |
| "epoch": 2.06, |
| "grad_norm": 0.8554444909095764, |
| "learning_rate": 9.673527324343507e-05, |
| "loss": 0.7148, |
| "step": 1457 |
| }, |
| { |
| "epoch": 2.07, |
| "grad_norm": 0.8741387128829956, |
| "learning_rate": 9.666430092264017e-05, |
| "loss": 0.7527, |
| "step": 1458 |
| }, |
| { |
| "epoch": 2.07, |
| "grad_norm": 0.8928270936012268, |
| "learning_rate": 9.65933286018453e-05, |
| "loss": 0.7233, |
| "step": 1459 |
| }, |
| { |
| "epoch": 2.07, |
| "grad_norm": 0.8724638223648071, |
| "learning_rate": 9.652235628105039e-05, |
| "loss": 0.7014, |
| "step": 1460 |
| }, |
| { |
| "epoch": 2.07, |
| "grad_norm": 0.8374599814414978, |
| "learning_rate": 9.64513839602555e-05, |
| "loss": 0.6646, |
| "step": 1461 |
| }, |
| { |
| "epoch": 2.07, |
| "grad_norm": 0.8397142291069031, |
| "learning_rate": 9.63804116394606e-05, |
| "loss": 0.6829, |
| "step": 1462 |
| }, |
| { |
| "epoch": 2.07, |
| "grad_norm": 0.7984825372695923, |
| "learning_rate": 9.630943931866573e-05, |
| "loss": 0.748, |
| "step": 1463 |
| }, |
| { |
| "epoch": 2.07, |
| "grad_norm": 0.8543975353240967, |
| "learning_rate": 9.623846699787083e-05, |
| "loss": 0.6743, |
| "step": 1464 |
| }, |
| { |
| "epoch": 2.08, |
| "grad_norm": 0.8055514693260193, |
| "learning_rate": 9.616749467707594e-05, |
| "loss": 0.6725, |
| "step": 1465 |
| }, |
| { |
| "epoch": 2.08, |
| "grad_norm": 0.8549181818962097, |
| "learning_rate": 9.609652235628106e-05, |
| "loss": 0.7744, |
| "step": 1466 |
| }, |
| { |
| "epoch": 2.08, |
| "grad_norm": 0.8540986180305481, |
| "learning_rate": 9.602555003548617e-05, |
| "loss": 0.779, |
| "step": 1467 |
| }, |
| { |
| "epoch": 2.08, |
| "grad_norm": 0.8481354117393494, |
| "learning_rate": 9.595457771469127e-05, |
| "loss": 0.706, |
| "step": 1468 |
| }, |
| { |
| "epoch": 2.08, |
| "grad_norm": 0.8669779896736145, |
| "learning_rate": 9.588360539389638e-05, |
| "loss": 0.6772, |
| "step": 1469 |
| }, |
| { |
| "epoch": 2.08, |
| "grad_norm": 0.9160670638084412, |
| "learning_rate": 9.581263307310149e-05, |
| "loss": 0.7225, |
| "step": 1470 |
| }, |
| { |
| "epoch": 2.08, |
| "grad_norm": 0.9088666439056396, |
| "learning_rate": 9.57416607523066e-05, |
| "loss": 0.7003, |
| "step": 1471 |
| }, |
| { |
| "epoch": 2.09, |
| "grad_norm": 0.906645655632019, |
| "learning_rate": 9.567068843151172e-05, |
| "loss": 0.7324, |
| "step": 1472 |
| }, |
| { |
| "epoch": 2.09, |
| "grad_norm": 0.8714525103569031, |
| "learning_rate": 9.559971611071683e-05, |
| "loss": 0.7198, |
| "step": 1473 |
| }, |
| { |
| "epoch": 2.09, |
| "grad_norm": 0.8619925379753113, |
| "learning_rate": 9.552874378992194e-05, |
| "loss": 0.7415, |
| "step": 1474 |
| }, |
| { |
| "epoch": 2.09, |
| "grad_norm": 0.8662142157554626, |
| "learning_rate": 9.545777146912704e-05, |
| "loss": 0.7169, |
| "step": 1475 |
| }, |
| { |
| "epoch": 2.09, |
| "grad_norm": 0.8264342546463013, |
| "learning_rate": 9.538679914833216e-05, |
| "loss": 0.7027, |
| "step": 1476 |
| }, |
| { |
| "epoch": 2.09, |
| "grad_norm": 0.849160373210907, |
| "learning_rate": 9.531582682753727e-05, |
| "loss": 0.7204, |
| "step": 1477 |
| }, |
| { |
| "epoch": 2.09, |
| "grad_norm": 0.8109767436981201, |
| "learning_rate": 9.524485450674238e-05, |
| "loss": 0.7396, |
| "step": 1478 |
| }, |
| { |
| "epoch": 2.1, |
| "grad_norm": 0.8427298069000244, |
| "learning_rate": 9.517388218594748e-05, |
| "loss": 0.7111, |
| "step": 1479 |
| }, |
| { |
| "epoch": 2.1, |
| "grad_norm": 0.8257899284362793, |
| "learning_rate": 9.510290986515259e-05, |
| "loss": 0.6849, |
| "step": 1480 |
| }, |
| { |
| "epoch": 2.1, |
| "grad_norm": 0.8619416356086731, |
| "learning_rate": 9.50319375443577e-05, |
| "loss": 0.6701, |
| "step": 1481 |
| }, |
| { |
| "epoch": 2.1, |
| "grad_norm": 0.9152725338935852, |
| "learning_rate": 9.496096522356282e-05, |
| "loss": 0.7793, |
| "step": 1482 |
| }, |
| { |
| "epoch": 2.1, |
| "grad_norm": 0.9197123050689697, |
| "learning_rate": 9.488999290276792e-05, |
| "loss": 0.7291, |
| "step": 1483 |
| }, |
| { |
| "epoch": 2.1, |
| "grad_norm": 0.8989003896713257, |
| "learning_rate": 9.481902058197304e-05, |
| "loss": 0.733, |
| "step": 1484 |
| }, |
| { |
| "epoch": 2.1, |
| "grad_norm": 0.8624677062034607, |
| "learning_rate": 9.474804826117814e-05, |
| "loss": 0.7366, |
| "step": 1485 |
| }, |
| { |
| "epoch": 2.11, |
| "grad_norm": 0.8718342185020447, |
| "learning_rate": 9.467707594038326e-05, |
| "loss": 0.7399, |
| "step": 1486 |
| }, |
| { |
| "epoch": 2.11, |
| "grad_norm": 0.868829071521759, |
| "learning_rate": 9.460610361958836e-05, |
| "loss": 0.7323, |
| "step": 1487 |
| }, |
| { |
| "epoch": 2.11, |
| "grad_norm": 0.8415536284446716, |
| "learning_rate": 9.453513129879348e-05, |
| "loss": 0.712, |
| "step": 1488 |
| }, |
| { |
| "epoch": 2.11, |
| "grad_norm": 0.8465087413787842, |
| "learning_rate": 9.446415897799858e-05, |
| "loss": 0.7202, |
| "step": 1489 |
| }, |
| { |
| "epoch": 2.11, |
| "grad_norm": 0.8721869587898254, |
| "learning_rate": 9.43931866572037e-05, |
| "loss": 0.7269, |
| "step": 1490 |
| }, |
| { |
| "epoch": 2.11, |
| "grad_norm": 0.8698458671569824, |
| "learning_rate": 9.43222143364088e-05, |
| "loss": 0.7442, |
| "step": 1491 |
| }, |
| { |
| "epoch": 2.11, |
| "grad_norm": 0.8857313394546509, |
| "learning_rate": 9.425124201561392e-05, |
| "loss": 0.7525, |
| "step": 1492 |
| }, |
| { |
| "epoch": 2.12, |
| "grad_norm": 0.8833426833152771, |
| "learning_rate": 9.418026969481902e-05, |
| "loss": 0.7618, |
| "step": 1493 |
| }, |
| { |
| "epoch": 2.12, |
| "grad_norm": 0.8640385866165161, |
| "learning_rate": 9.410929737402413e-05, |
| "loss": 0.7094, |
| "step": 1494 |
| }, |
| { |
| "epoch": 2.12, |
| "grad_norm": 0.8465377688407898, |
| "learning_rate": 9.403832505322924e-05, |
| "loss": 0.7406, |
| "step": 1495 |
| }, |
| { |
| "epoch": 2.12, |
| "grad_norm": 0.8516689538955688, |
| "learning_rate": 9.396735273243436e-05, |
| "loss": 0.7149, |
| "step": 1496 |
| }, |
| { |
| "epoch": 2.12, |
| "grad_norm": 0.8319384455680847, |
| "learning_rate": 9.389638041163946e-05, |
| "loss": 0.6586, |
| "step": 1497 |
| }, |
| { |
| "epoch": 2.12, |
| "grad_norm": 0.8527852892875671, |
| "learning_rate": 9.382540809084458e-05, |
| "loss": 0.6629, |
| "step": 1498 |
| }, |
| { |
| "epoch": 2.12, |
| "grad_norm": 0.8378131985664368, |
| "learning_rate": 9.375443577004968e-05, |
| "loss": 0.704, |
| "step": 1499 |
| }, |
| { |
| "epoch": 2.13, |
| "grad_norm": 0.8492695689201355, |
| "learning_rate": 9.36834634492548e-05, |
| "loss": 0.6833, |
| "step": 1500 |
| }, |
| { |
| "epoch": 2.13, |
| "grad_norm": 0.8611241579055786, |
| "learning_rate": 9.36124911284599e-05, |
| "loss": 0.6794, |
| "step": 1501 |
| }, |
| { |
| "epoch": 2.13, |
| "grad_norm": 0.914234459400177, |
| "learning_rate": 9.354151880766502e-05, |
| "loss": 0.7751, |
| "step": 1502 |
| }, |
| { |
| "epoch": 2.13, |
| "grad_norm": 0.8970404863357544, |
| "learning_rate": 9.347054648687013e-05, |
| "loss": 0.7603, |
| "step": 1503 |
| }, |
| { |
| "epoch": 2.13, |
| "grad_norm": 0.8614270687103271, |
| "learning_rate": 9.339957416607523e-05, |
| "loss": 0.7352, |
| "step": 1504 |
| }, |
| { |
| "epoch": 2.13, |
| "grad_norm": 0.8698607683181763, |
| "learning_rate": 9.332860184528034e-05, |
| "loss": 0.7116, |
| "step": 1505 |
| }, |
| { |
| "epoch": 2.13, |
| "grad_norm": 0.8466779589653015, |
| "learning_rate": 9.325762952448546e-05, |
| "loss": 0.695, |
| "step": 1506 |
| }, |
| { |
| "epoch": 2.13, |
| "grad_norm": 0.8445769548416138, |
| "learning_rate": 9.318665720369057e-05, |
| "loss": 0.7525, |
| "step": 1507 |
| }, |
| { |
| "epoch": 2.14, |
| "grad_norm": 0.8695936799049377, |
| "learning_rate": 9.311568488289567e-05, |
| "loss": 0.7352, |
| "step": 1508 |
| }, |
| { |
| "epoch": 2.14, |
| "grad_norm": 0.8878143429756165, |
| "learning_rate": 9.30447125621008e-05, |
| "loss": 0.713, |
| "step": 1509 |
| }, |
| { |
| "epoch": 2.14, |
| "grad_norm": 0.8836119174957275, |
| "learning_rate": 9.29737402413059e-05, |
| "loss": 0.7132, |
| "step": 1510 |
| }, |
| { |
| "epoch": 2.14, |
| "grad_norm": 0.8474823236465454, |
| "learning_rate": 9.2902767920511e-05, |
| "loss": 0.6765, |
| "step": 1511 |
| }, |
| { |
| "epoch": 2.14, |
| "grad_norm": 0.8740355968475342, |
| "learning_rate": 9.283179559971612e-05, |
| "loss": 0.7044, |
| "step": 1512 |
| }, |
| { |
| "epoch": 2.14, |
| "grad_norm": 0.8775798082351685, |
| "learning_rate": 9.276082327892123e-05, |
| "loss": 0.7338, |
| "step": 1513 |
| }, |
| { |
| "epoch": 2.14, |
| "grad_norm": 0.8951154351234436, |
| "learning_rate": 9.268985095812633e-05, |
| "loss": 0.723, |
| "step": 1514 |
| }, |
| { |
| "epoch": 2.15, |
| "grad_norm": 0.891923189163208, |
| "learning_rate": 9.261887863733144e-05, |
| "loss": 0.7056, |
| "step": 1515 |
| }, |
| { |
| "epoch": 2.15, |
| "grad_norm": 0.8443423509597778, |
| "learning_rate": 9.254790631653656e-05, |
| "loss": 0.6884, |
| "step": 1516 |
| }, |
| { |
| "epoch": 2.15, |
| "grad_norm": 0.8721277713775635, |
| "learning_rate": 9.247693399574167e-05, |
| "loss": 0.6926, |
| "step": 1517 |
| }, |
| { |
| "epoch": 2.15, |
| "grad_norm": 0.9046376943588257, |
| "learning_rate": 9.240596167494677e-05, |
| "loss": 0.7471, |
| "step": 1518 |
| }, |
| { |
| "epoch": 2.15, |
| "grad_norm": 0.8587860465049744, |
| "learning_rate": 9.23349893541519e-05, |
| "loss": 0.726, |
| "step": 1519 |
| }, |
| { |
| "epoch": 2.15, |
| "grad_norm": 0.8803458213806152, |
| "learning_rate": 9.2264017033357e-05, |
| "loss": 0.7576, |
| "step": 1520 |
| }, |
| { |
| "epoch": 2.15, |
| "grad_norm": 0.8257414102554321, |
| "learning_rate": 9.21930447125621e-05, |
| "loss": 0.6852, |
| "step": 1521 |
| }, |
| { |
| "epoch": 2.16, |
| "grad_norm": 0.8487634658813477, |
| "learning_rate": 9.21220723917672e-05, |
| "loss": 0.7674, |
| "step": 1522 |
| }, |
| { |
| "epoch": 2.16, |
| "grad_norm": 0.8407074213027954, |
| "learning_rate": 9.205110007097233e-05, |
| "loss": 0.6811, |
| "step": 1523 |
| }, |
| { |
| "epoch": 2.16, |
| "grad_norm": 0.8839500546455383, |
| "learning_rate": 9.198012775017743e-05, |
| "loss": 0.7953, |
| "step": 1524 |
| }, |
| { |
| "epoch": 2.16, |
| "grad_norm": 0.875453531742096, |
| "learning_rate": 9.190915542938254e-05, |
| "loss": 0.7141, |
| "step": 1525 |
| }, |
| { |
| "epoch": 2.16, |
| "grad_norm": 0.8481379747390747, |
| "learning_rate": 9.183818310858766e-05, |
| "loss": 0.6424, |
| "step": 1526 |
| }, |
| { |
| "epoch": 2.16, |
| "grad_norm": 0.8793319463729858, |
| "learning_rate": 9.176721078779277e-05, |
| "loss": 0.6961, |
| "step": 1527 |
| }, |
| { |
| "epoch": 2.16, |
| "grad_norm": 0.8625543117523193, |
| "learning_rate": 9.169623846699787e-05, |
| "loss": 0.7493, |
| "step": 1528 |
| }, |
| { |
| "epoch": 2.17, |
| "grad_norm": 0.8540360927581787, |
| "learning_rate": 9.162526614620298e-05, |
| "loss": 0.7261, |
| "step": 1529 |
| }, |
| { |
| "epoch": 2.17, |
| "grad_norm": 0.8655250072479248, |
| "learning_rate": 9.15542938254081e-05, |
| "loss": 0.7124, |
| "step": 1530 |
| }, |
| { |
| "epoch": 2.17, |
| "grad_norm": 0.8972883224487305, |
| "learning_rate": 9.14833215046132e-05, |
| "loss": 0.7118, |
| "step": 1531 |
| }, |
| { |
| "epoch": 2.17, |
| "grad_norm": 0.8736087679862976, |
| "learning_rate": 9.141234918381832e-05, |
| "loss": 0.7169, |
| "step": 1532 |
| }, |
| { |
| "epoch": 2.17, |
| "grad_norm": 0.8996816873550415, |
| "learning_rate": 9.134137686302342e-05, |
| "loss": 0.7213, |
| "step": 1533 |
| }, |
| { |
| "epoch": 2.17, |
| "grad_norm": 0.8447397351264954, |
| "learning_rate": 9.127040454222854e-05, |
| "loss": 0.7168, |
| "step": 1534 |
| }, |
| { |
| "epoch": 2.17, |
| "grad_norm": 0.837806224822998, |
| "learning_rate": 9.119943222143364e-05, |
| "loss": 0.71, |
| "step": 1535 |
| }, |
| { |
| "epoch": 2.18, |
| "grad_norm": 0.8344355821609497, |
| "learning_rate": 9.112845990063876e-05, |
| "loss": 0.6721, |
| "step": 1536 |
| }, |
| { |
| "epoch": 2.18, |
| "grad_norm": 0.8420239090919495, |
| "learning_rate": 9.105748757984387e-05, |
| "loss": 0.6764, |
| "step": 1537 |
| }, |
| { |
| "epoch": 2.18, |
| "grad_norm": 0.8825182914733887, |
| "learning_rate": 9.098651525904898e-05, |
| "loss": 0.709, |
| "step": 1538 |
| }, |
| { |
| "epoch": 2.18, |
| "grad_norm": 0.9009390473365784, |
| "learning_rate": 9.091554293825408e-05, |
| "loss": 0.7687, |
| "step": 1539 |
| }, |
| { |
| "epoch": 2.18, |
| "grad_norm": 0.854705810546875, |
| "learning_rate": 9.08445706174592e-05, |
| "loss": 0.7118, |
| "step": 1540 |
| }, |
| { |
| "epoch": 2.18, |
| "grad_norm": 0.9011061191558838, |
| "learning_rate": 9.07735982966643e-05, |
| "loss": 0.716, |
| "step": 1541 |
| }, |
| { |
| "epoch": 2.18, |
| "grad_norm": 0.8979266285896301, |
| "learning_rate": 9.070262597586942e-05, |
| "loss": 0.7841, |
| "step": 1542 |
| }, |
| { |
| "epoch": 2.19, |
| "grad_norm": 0.9013125896453857, |
| "learning_rate": 9.063165365507452e-05, |
| "loss": 0.7402, |
| "step": 1543 |
| }, |
| { |
| "epoch": 2.19, |
| "grad_norm": 0.8848702907562256, |
| "learning_rate": 9.056068133427964e-05, |
| "loss": 0.7384, |
| "step": 1544 |
| }, |
| { |
| "epoch": 2.19, |
| "grad_norm": 0.9167879819869995, |
| "learning_rate": 9.048970901348474e-05, |
| "loss": 0.8391, |
| "step": 1545 |
| }, |
| { |
| "epoch": 2.19, |
| "grad_norm": 0.8656322360038757, |
| "learning_rate": 9.041873669268986e-05, |
| "loss": 0.7492, |
| "step": 1546 |
| }, |
| { |
| "epoch": 2.19, |
| "grad_norm": 0.842842698097229, |
| "learning_rate": 9.034776437189496e-05, |
| "loss": 0.7487, |
| "step": 1547 |
| }, |
| { |
| "epoch": 2.19, |
| "grad_norm": 0.8449062705039978, |
| "learning_rate": 9.027679205110008e-05, |
| "loss": 0.761, |
| "step": 1548 |
| }, |
| { |
| "epoch": 2.19, |
| "grad_norm": 0.8444206118583679, |
| "learning_rate": 9.020581973030518e-05, |
| "loss": 0.6761, |
| "step": 1549 |
| }, |
| { |
| "epoch": 2.2, |
| "grad_norm": 0.8564352989196777, |
| "learning_rate": 9.01348474095103e-05, |
| "loss": 0.693, |
| "step": 1550 |
| }, |
| { |
| "epoch": 2.2, |
| "grad_norm": 0.8907448053359985, |
| "learning_rate": 9.006387508871541e-05, |
| "loss": 0.7221, |
| "step": 1551 |
| }, |
| { |
| "epoch": 2.2, |
| "grad_norm": 0.8845468163490295, |
| "learning_rate": 8.999290276792052e-05, |
| "loss": 0.7229, |
| "step": 1552 |
| }, |
| { |
| "epoch": 2.2, |
| "grad_norm": 0.8686636090278625, |
| "learning_rate": 8.992193044712562e-05, |
| "loss": 0.6997, |
| "step": 1553 |
| }, |
| { |
| "epoch": 2.2, |
| "grad_norm": 0.9412215948104858, |
| "learning_rate": 8.985095812633073e-05, |
| "loss": 0.7125, |
| "step": 1554 |
| }, |
| { |
| "epoch": 2.2, |
| "grad_norm": 0.8855116963386536, |
| "learning_rate": 8.977998580553584e-05, |
| "loss": 0.7206, |
| "step": 1555 |
| }, |
| { |
| "epoch": 2.2, |
| "grad_norm": 0.8830530643463135, |
| "learning_rate": 8.970901348474096e-05, |
| "loss": 0.6728, |
| "step": 1556 |
| }, |
| { |
| "epoch": 2.21, |
| "grad_norm": 0.892914891242981, |
| "learning_rate": 8.963804116394606e-05, |
| "loss": 0.6458, |
| "step": 1557 |
| }, |
| { |
| "epoch": 2.21, |
| "grad_norm": 0.8819034099578857, |
| "learning_rate": 8.956706884315118e-05, |
| "loss": 0.7155, |
| "step": 1558 |
| }, |
| { |
| "epoch": 2.21, |
| "grad_norm": 0.8601769208908081, |
| "learning_rate": 8.949609652235628e-05, |
| "loss": 0.7252, |
| "step": 1559 |
| }, |
| { |
| "epoch": 2.21, |
| "grad_norm": 0.8619159460067749, |
| "learning_rate": 8.94251242015614e-05, |
| "loss": 0.7115, |
| "step": 1560 |
| }, |
| { |
| "epoch": 2.21, |
| "grad_norm": 0.8631980419158936, |
| "learning_rate": 8.935415188076651e-05, |
| "loss": 0.7112, |
| "step": 1561 |
| }, |
| { |
| "epoch": 2.21, |
| "grad_norm": 0.888449490070343, |
| "learning_rate": 8.928317955997162e-05, |
| "loss": 0.7384, |
| "step": 1562 |
| }, |
| { |
| "epoch": 2.21, |
| "grad_norm": 0.8786005973815918, |
| "learning_rate": 8.921220723917673e-05, |
| "loss": 0.6962, |
| "step": 1563 |
| }, |
| { |
| "epoch": 2.22, |
| "grad_norm": 0.891455352306366, |
| "learning_rate": 8.914123491838183e-05, |
| "loss": 0.7471, |
| "step": 1564 |
| }, |
| { |
| "epoch": 2.22, |
| "grad_norm": 0.8893013596534729, |
| "learning_rate": 8.907026259758694e-05, |
| "loss": 0.7321, |
| "step": 1565 |
| }, |
| { |
| "epoch": 2.22, |
| "grad_norm": 0.8937901854515076, |
| "learning_rate": 8.899929027679206e-05, |
| "loss": 0.712, |
| "step": 1566 |
| }, |
| { |
| "epoch": 2.22, |
| "grad_norm": 0.9010066986083984, |
| "learning_rate": 8.892831795599717e-05, |
| "loss": 0.742, |
| "step": 1567 |
| }, |
| { |
| "epoch": 2.22, |
| "grad_norm": 0.8523016571998596, |
| "learning_rate": 8.885734563520227e-05, |
| "loss": 0.6838, |
| "step": 1568 |
| }, |
| { |
| "epoch": 2.22, |
| "grad_norm": 0.8775461912155151, |
| "learning_rate": 8.87863733144074e-05, |
| "loss": 0.7036, |
| "step": 1569 |
| }, |
| { |
| "epoch": 2.22, |
| "grad_norm": 0.8605543375015259, |
| "learning_rate": 8.87154009936125e-05, |
| "loss": 0.6835, |
| "step": 1570 |
| }, |
| { |
| "epoch": 2.23, |
| "grad_norm": 0.8719514012336731, |
| "learning_rate": 8.864442867281761e-05, |
| "loss": 0.747, |
| "step": 1571 |
| }, |
| { |
| "epoch": 2.23, |
| "grad_norm": 0.8636995553970337, |
| "learning_rate": 8.85734563520227e-05, |
| "loss": 0.6874, |
| "step": 1572 |
| }, |
| { |
| "epoch": 2.23, |
| "grad_norm": 0.856977641582489, |
| "learning_rate": 8.850248403122783e-05, |
| "loss": 0.7234, |
| "step": 1573 |
| }, |
| { |
| "epoch": 2.23, |
| "grad_norm": 0.8813896775245667, |
| "learning_rate": 8.843151171043293e-05, |
| "loss": 0.6992, |
| "step": 1574 |
| }, |
| { |
| "epoch": 2.23, |
| "grad_norm": 0.8693534135818481, |
| "learning_rate": 8.836053938963804e-05, |
| "loss": 0.7058, |
| "step": 1575 |
| }, |
| { |
| "epoch": 2.23, |
| "grad_norm": 0.8914763331413269, |
| "learning_rate": 8.828956706884316e-05, |
| "loss": 0.7247, |
| "step": 1576 |
| }, |
| { |
| "epoch": 2.23, |
| "grad_norm": 0.8628402948379517, |
| "learning_rate": 8.821859474804827e-05, |
| "loss": 0.7046, |
| "step": 1577 |
| }, |
| { |
| "epoch": 2.24, |
| "grad_norm": 0.8618187308311462, |
| "learning_rate": 8.814762242725337e-05, |
| "loss": 0.7106, |
| "step": 1578 |
| }, |
| { |
| "epoch": 2.24, |
| "grad_norm": 0.8622534871101379, |
| "learning_rate": 8.807665010645848e-05, |
| "loss": 0.674, |
| "step": 1579 |
| }, |
| { |
| "epoch": 2.24, |
| "grad_norm": 0.8331353068351746, |
| "learning_rate": 8.80056777856636e-05, |
| "loss": 0.6985, |
| "step": 1580 |
| }, |
| { |
| "epoch": 2.24, |
| "grad_norm": 0.8842141032218933, |
| "learning_rate": 8.793470546486871e-05, |
| "loss": 0.7563, |
| "step": 1581 |
| }, |
| { |
| "epoch": 2.24, |
| "grad_norm": 0.8404679894447327, |
| "learning_rate": 8.78637331440738e-05, |
| "loss": 0.6946, |
| "step": 1582 |
| }, |
| { |
| "epoch": 2.24, |
| "grad_norm": 0.8643913269042969, |
| "learning_rate": 8.779276082327893e-05, |
| "loss": 0.6439, |
| "step": 1583 |
| }, |
| { |
| "epoch": 2.24, |
| "grad_norm": 0.8534083366394043, |
| "learning_rate": 8.772178850248403e-05, |
| "loss": 0.6621, |
| "step": 1584 |
| }, |
| { |
| "epoch": 2.25, |
| "grad_norm": 0.9061523079872131, |
| "learning_rate": 8.765081618168914e-05, |
| "loss": 0.7496, |
| "step": 1585 |
| }, |
| { |
| "epoch": 2.25, |
| "grad_norm": 0.881190299987793, |
| "learning_rate": 8.757984386089424e-05, |
| "loss": 0.7319, |
| "step": 1586 |
| }, |
| { |
| "epoch": 2.25, |
| "grad_norm": 0.8903520107269287, |
| "learning_rate": 8.750887154009937e-05, |
| "loss": 0.7292, |
| "step": 1587 |
| }, |
| { |
| "epoch": 2.25, |
| "grad_norm": 0.8949286341667175, |
| "learning_rate": 8.743789921930447e-05, |
| "loss": 0.7834, |
| "step": 1588 |
| }, |
| { |
| "epoch": 2.25, |
| "grad_norm": 0.8609988689422607, |
| "learning_rate": 8.736692689850958e-05, |
| "loss": 0.7096, |
| "step": 1589 |
| }, |
| { |
| "epoch": 2.25, |
| "grad_norm": 0.8585081100463867, |
| "learning_rate": 8.72959545777147e-05, |
| "loss": 0.7309, |
| "step": 1590 |
| }, |
| { |
| "epoch": 2.25, |
| "grad_norm": 0.8561607599258423, |
| "learning_rate": 8.722498225691981e-05, |
| "loss": 0.7206, |
| "step": 1591 |
| }, |
| { |
| "epoch": 2.26, |
| "grad_norm": 0.8617041707038879, |
| "learning_rate": 8.715400993612492e-05, |
| "loss": 0.7539, |
| "step": 1592 |
| }, |
| { |
| "epoch": 2.26, |
| "grad_norm": 0.8783171772956848, |
| "learning_rate": 8.708303761533002e-05, |
| "loss": 0.6923, |
| "step": 1593 |
| }, |
| { |
| "epoch": 2.26, |
| "grad_norm": 0.8528505563735962, |
| "learning_rate": 8.701206529453513e-05, |
| "loss": 0.7053, |
| "step": 1594 |
| }, |
| { |
| "epoch": 2.26, |
| "grad_norm": 0.8965705633163452, |
| "learning_rate": 8.694109297374024e-05, |
| "loss": 0.7004, |
| "step": 1595 |
| }, |
| { |
| "epoch": 2.26, |
| "grad_norm": 0.8787754774093628, |
| "learning_rate": 8.687012065294536e-05, |
| "loss": 0.7189, |
| "step": 1596 |
| }, |
| { |
| "epoch": 2.26, |
| "grad_norm": 0.9146087169647217, |
| "learning_rate": 8.679914833215047e-05, |
| "loss": 0.6916, |
| "step": 1597 |
| }, |
| { |
| "epoch": 2.26, |
| "grad_norm": 0.931847870349884, |
| "learning_rate": 8.672817601135558e-05, |
| "loss": 0.7567, |
| "step": 1598 |
| }, |
| { |
| "epoch": 2.27, |
| "grad_norm": 0.9045712351799011, |
| "learning_rate": 8.665720369056068e-05, |
| "loss": 0.755, |
| "step": 1599 |
| }, |
| { |
| "epoch": 2.27, |
| "grad_norm": 0.8855098485946655, |
| "learning_rate": 8.65862313697658e-05, |
| "loss": 0.7277, |
| "step": 1600 |
| }, |
| { |
| "epoch": 2.27, |
| "grad_norm": 0.8935732245445251, |
| "learning_rate": 8.651525904897091e-05, |
| "loss": 0.7153, |
| "step": 1601 |
| }, |
| { |
| "epoch": 2.27, |
| "grad_norm": 0.9022087454795837, |
| "learning_rate": 8.644428672817602e-05, |
| "loss": 0.735, |
| "step": 1602 |
| }, |
| { |
| "epoch": 2.27, |
| "grad_norm": 0.8693438172340393, |
| "learning_rate": 8.637331440738112e-05, |
| "loss": 0.7232, |
| "step": 1603 |
| }, |
| { |
| "epoch": 2.27, |
| "grad_norm": 0.85861736536026, |
| "learning_rate": 8.630234208658625e-05, |
| "loss": 0.7183, |
| "step": 1604 |
| }, |
| { |
| "epoch": 2.27, |
| "grad_norm": 0.8743335604667664, |
| "learning_rate": 8.623136976579135e-05, |
| "loss": 0.7973, |
| "step": 1605 |
| }, |
| { |
| "epoch": 2.28, |
| "grad_norm": 0.8646829128265381, |
| "learning_rate": 8.616039744499646e-05, |
| "loss": 0.7171, |
| "step": 1606 |
| }, |
| { |
| "epoch": 2.28, |
| "grad_norm": 0.8387510180473328, |
| "learning_rate": 8.608942512420156e-05, |
| "loss": 0.7389, |
| "step": 1607 |
| }, |
| { |
| "epoch": 2.28, |
| "grad_norm": 0.8693937063217163, |
| "learning_rate": 8.601845280340668e-05, |
| "loss": 0.7036, |
| "step": 1608 |
| }, |
| { |
| "epoch": 2.28, |
| "grad_norm": 0.8225006461143494, |
| "learning_rate": 8.594748048261178e-05, |
| "loss": 0.7298, |
| "step": 1609 |
| }, |
| { |
| "epoch": 2.28, |
| "grad_norm": 0.9259347319602966, |
| "learning_rate": 8.58765081618169e-05, |
| "loss": 0.7549, |
| "step": 1610 |
| }, |
| { |
| "epoch": 2.28, |
| "grad_norm": 0.9134079813957214, |
| "learning_rate": 8.5805535841022e-05, |
| "loss": 0.7127, |
| "step": 1611 |
| }, |
| { |
| "epoch": 2.28, |
| "grad_norm": 0.9061034917831421, |
| "learning_rate": 8.573456352022712e-05, |
| "loss": 0.7162, |
| "step": 1612 |
| }, |
| { |
| "epoch": 2.29, |
| "grad_norm": 0.9124405384063721, |
| "learning_rate": 8.566359119943222e-05, |
| "loss": 0.7308, |
| "step": 1613 |
| }, |
| { |
| "epoch": 2.29, |
| "grad_norm": 0.8995393514633179, |
| "learning_rate": 8.559261887863733e-05, |
| "loss": 0.7131, |
| "step": 1614 |
| }, |
| { |
| "epoch": 2.29, |
| "grad_norm": 0.9184269905090332, |
| "learning_rate": 8.552164655784245e-05, |
| "loss": 0.7235, |
| "step": 1615 |
| }, |
| { |
| "epoch": 2.29, |
| "grad_norm": 0.8481336832046509, |
| "learning_rate": 8.545067423704756e-05, |
| "loss": 0.6814, |
| "step": 1616 |
| }, |
| { |
| "epoch": 2.29, |
| "grad_norm": 0.9017961025238037, |
| "learning_rate": 8.537970191625266e-05, |
| "loss": 0.6856, |
| "step": 1617 |
| }, |
| { |
| "epoch": 2.29, |
| "grad_norm": 0.8780509829521179, |
| "learning_rate": 8.530872959545777e-05, |
| "loss": 0.7255, |
| "step": 1618 |
| }, |
| { |
| "epoch": 2.29, |
| "grad_norm": 0.8859030604362488, |
| "learning_rate": 8.523775727466288e-05, |
| "loss": 0.7343, |
| "step": 1619 |
| }, |
| { |
| "epoch": 2.3, |
| "grad_norm": 0.8603264689445496, |
| "learning_rate": 8.5166784953868e-05, |
| "loss": 0.7088, |
| "step": 1620 |
| }, |
| { |
| "epoch": 2.3, |
| "grad_norm": 0.8859502077102661, |
| "learning_rate": 8.509581263307311e-05, |
| "loss": 0.7264, |
| "step": 1621 |
| }, |
| { |
| "epoch": 2.3, |
| "grad_norm": 0.8135156631469727, |
| "learning_rate": 8.502484031227822e-05, |
| "loss": 0.6724, |
| "step": 1622 |
| }, |
| { |
| "epoch": 2.3, |
| "grad_norm": 0.8489874005317688, |
| "learning_rate": 8.495386799148333e-05, |
| "loss": 0.6948, |
| "step": 1623 |
| }, |
| { |
| "epoch": 2.3, |
| "grad_norm": 0.8707240223884583, |
| "learning_rate": 8.488289567068843e-05, |
| "loss": 0.742, |
| "step": 1624 |
| }, |
| { |
| "epoch": 2.3, |
| "grad_norm": 0.8624871373176575, |
| "learning_rate": 8.481192334989355e-05, |
| "loss": 0.6816, |
| "step": 1625 |
| }, |
| { |
| "epoch": 2.3, |
| "grad_norm": 0.8828703165054321, |
| "learning_rate": 8.474095102909866e-05, |
| "loss": 0.7385, |
| "step": 1626 |
| }, |
| { |
| "epoch": 2.3, |
| "grad_norm": 0.8730260133743286, |
| "learning_rate": 8.466997870830377e-05, |
| "loss": 0.7126, |
| "step": 1627 |
| }, |
| { |
| "epoch": 2.31, |
| "grad_norm": 0.8839560747146606, |
| "learning_rate": 8.459900638750887e-05, |
| "loss": 0.7272, |
| "step": 1628 |
| }, |
| { |
| "epoch": 2.31, |
| "grad_norm": 0.8795418739318848, |
| "learning_rate": 8.4528034066714e-05, |
| "loss": 0.6953, |
| "step": 1629 |
| }, |
| { |
| "epoch": 2.31, |
| "grad_norm": 0.8948290348052979, |
| "learning_rate": 8.44570617459191e-05, |
| "loss": 0.7241, |
| "step": 1630 |
| }, |
| { |
| "epoch": 2.31, |
| "grad_norm": 0.8926292061805725, |
| "learning_rate": 8.438608942512421e-05, |
| "loss": 0.6998, |
| "step": 1631 |
| }, |
| { |
| "epoch": 2.31, |
| "grad_norm": 0.9082595705986023, |
| "learning_rate": 8.431511710432931e-05, |
| "loss": 0.7715, |
| "step": 1632 |
| }, |
| { |
| "epoch": 2.31, |
| "grad_norm": 0.9164185523986816, |
| "learning_rate": 8.424414478353443e-05, |
| "loss": 0.7412, |
| "step": 1633 |
| }, |
| { |
| "epoch": 2.31, |
| "grad_norm": 0.8599331378936768, |
| "learning_rate": 8.417317246273953e-05, |
| "loss": 0.7327, |
| "step": 1634 |
| }, |
| { |
| "epoch": 2.32, |
| "grad_norm": 0.8912233114242554, |
| "learning_rate": 8.410220014194465e-05, |
| "loss": 0.7073, |
| "step": 1635 |
| }, |
| { |
| "epoch": 2.32, |
| "grad_norm": 0.878819465637207, |
| "learning_rate": 8.403122782114976e-05, |
| "loss": 0.6973, |
| "step": 1636 |
| }, |
| { |
| "epoch": 2.32, |
| "grad_norm": 0.8971090912818909, |
| "learning_rate": 8.396025550035487e-05, |
| "loss": 0.7356, |
| "step": 1637 |
| }, |
| { |
| "epoch": 2.32, |
| "grad_norm": 0.9173135757446289, |
| "learning_rate": 8.388928317955997e-05, |
| "loss": 0.7359, |
| "step": 1638 |
| }, |
| { |
| "epoch": 2.32, |
| "grad_norm": 0.8811411261558533, |
| "learning_rate": 8.381831085876508e-05, |
| "loss": 0.6925, |
| "step": 1639 |
| }, |
| { |
| "epoch": 2.32, |
| "grad_norm": 0.8744279146194458, |
| "learning_rate": 8.37473385379702e-05, |
| "loss": 0.6386, |
| "step": 1640 |
| }, |
| { |
| "epoch": 2.32, |
| "grad_norm": 0.8834649324417114, |
| "learning_rate": 8.367636621717531e-05, |
| "loss": 0.748, |
| "step": 1641 |
| }, |
| { |
| "epoch": 2.33, |
| "grad_norm": 0.840594470500946, |
| "learning_rate": 8.360539389638041e-05, |
| "loss": 0.6875, |
| "step": 1642 |
| }, |
| { |
| "epoch": 2.33, |
| "grad_norm": 0.896140992641449, |
| "learning_rate": 8.353442157558553e-05, |
| "loss": 0.7505, |
| "step": 1643 |
| }, |
| { |
| "epoch": 2.33, |
| "grad_norm": 0.8742374181747437, |
| "learning_rate": 8.346344925479063e-05, |
| "loss": 0.7154, |
| "step": 1644 |
| }, |
| { |
| "epoch": 2.33, |
| "grad_norm": 0.8660913705825806, |
| "learning_rate": 8.339247693399575e-05, |
| "loss": 0.7332, |
| "step": 1645 |
| }, |
| { |
| "epoch": 2.33, |
| "grad_norm": 0.8654317855834961, |
| "learning_rate": 8.332150461320084e-05, |
| "loss": 0.7588, |
| "step": 1646 |
| }, |
| { |
| "epoch": 2.33, |
| "grad_norm": 0.8597747087478638, |
| "learning_rate": 8.325053229240597e-05, |
| "loss": 0.7629, |
| "step": 1647 |
| }, |
| { |
| "epoch": 2.33, |
| "grad_norm": 0.9062843322753906, |
| "learning_rate": 8.317955997161107e-05, |
| "loss": 0.8117, |
| "step": 1648 |
| }, |
| { |
| "epoch": 2.34, |
| "grad_norm": 0.8763964176177979, |
| "learning_rate": 8.310858765081618e-05, |
| "loss": 0.6865, |
| "step": 1649 |
| }, |
| { |
| "epoch": 2.34, |
| "grad_norm": 0.8947573304176331, |
| "learning_rate": 8.30376153300213e-05, |
| "loss": 0.7502, |
| "step": 1650 |
| }, |
| { |
| "epoch": 2.34, |
| "grad_norm": 0.9027217030525208, |
| "learning_rate": 8.296664300922641e-05, |
| "loss": 0.7362, |
| "step": 1651 |
| }, |
| { |
| "epoch": 2.34, |
| "grad_norm": 0.9003787040710449, |
| "learning_rate": 8.289567068843152e-05, |
| "loss": 0.7266, |
| "step": 1652 |
| }, |
| { |
| "epoch": 2.34, |
| "grad_norm": 0.8940318822860718, |
| "learning_rate": 8.282469836763662e-05, |
| "loss": 0.6746, |
| "step": 1653 |
| }, |
| { |
| "epoch": 2.34, |
| "grad_norm": 0.9138883352279663, |
| "learning_rate": 8.275372604684173e-05, |
| "loss": 0.7553, |
| "step": 1654 |
| }, |
| { |
| "epoch": 2.34, |
| "grad_norm": 0.8982981443405151, |
| "learning_rate": 8.268275372604685e-05, |
| "loss": 0.7519, |
| "step": 1655 |
| }, |
| { |
| "epoch": 2.35, |
| "grad_norm": 0.8930261731147766, |
| "learning_rate": 8.261178140525196e-05, |
| "loss": 0.7216, |
| "step": 1656 |
| }, |
| { |
| "epoch": 2.35, |
| "grad_norm": 0.8809303045272827, |
| "learning_rate": 8.254080908445706e-05, |
| "loss": 0.7515, |
| "step": 1657 |
| }, |
| { |
| "epoch": 2.35, |
| "grad_norm": 0.8962745070457458, |
| "learning_rate": 8.246983676366218e-05, |
| "loss": 0.7273, |
| "step": 1658 |
| }, |
| { |
| "epoch": 2.35, |
| "grad_norm": 0.8883169889450073, |
| "learning_rate": 8.239886444286728e-05, |
| "loss": 0.7156, |
| "step": 1659 |
| }, |
| { |
| "epoch": 2.35, |
| "grad_norm": 0.9205072522163391, |
| "learning_rate": 8.23278921220724e-05, |
| "loss": 0.7682, |
| "step": 1660 |
| }, |
| { |
| "epoch": 2.35, |
| "grad_norm": 0.8944506645202637, |
| "learning_rate": 8.225691980127751e-05, |
| "loss": 0.7306, |
| "step": 1661 |
| }, |
| { |
| "epoch": 2.35, |
| "grad_norm": 0.866554319858551, |
| "learning_rate": 8.218594748048262e-05, |
| "loss": 0.6766, |
| "step": 1662 |
| }, |
| { |
| "epoch": 2.36, |
| "grad_norm": 0.8790395855903625, |
| "learning_rate": 8.211497515968772e-05, |
| "loss": 0.7525, |
| "step": 1663 |
| }, |
| { |
| "epoch": 2.36, |
| "grad_norm": 0.8888643980026245, |
| "learning_rate": 8.204400283889283e-05, |
| "loss": 0.7373, |
| "step": 1664 |
| }, |
| { |
| "epoch": 2.36, |
| "grad_norm": 0.9181011915206909, |
| "learning_rate": 8.197303051809795e-05, |
| "loss": 0.7151, |
| "step": 1665 |
| }, |
| { |
| "epoch": 2.36, |
| "grad_norm": 0.8800945281982422, |
| "learning_rate": 8.190205819730306e-05, |
| "loss": 0.7438, |
| "step": 1666 |
| }, |
| { |
| "epoch": 2.36, |
| "grad_norm": 0.9163196086883545, |
| "learning_rate": 8.183108587650816e-05, |
| "loss": 0.7177, |
| "step": 1667 |
| }, |
| { |
| "epoch": 2.36, |
| "grad_norm": 0.8852064609527588, |
| "learning_rate": 8.176011355571328e-05, |
| "loss": 0.7321, |
| "step": 1668 |
| }, |
| { |
| "epoch": 2.36, |
| "grad_norm": 0.8783362507820129, |
| "learning_rate": 8.168914123491838e-05, |
| "loss": 0.7223, |
| "step": 1669 |
| }, |
| { |
| "epoch": 2.37, |
| "grad_norm": 0.8974859714508057, |
| "learning_rate": 8.16181689141235e-05, |
| "loss": 0.6787, |
| "step": 1670 |
| }, |
| { |
| "epoch": 2.37, |
| "grad_norm": 0.9229791760444641, |
| "learning_rate": 8.15471965933286e-05, |
| "loss": 0.76, |
| "step": 1671 |
| }, |
| { |
| "epoch": 2.37, |
| "grad_norm": 0.8831940293312073, |
| "learning_rate": 8.147622427253372e-05, |
| "loss": 0.6959, |
| "step": 1672 |
| }, |
| { |
| "epoch": 2.37, |
| "grad_norm": 0.9193075895309448, |
| "learning_rate": 8.140525195173882e-05, |
| "loss": 0.6838, |
| "step": 1673 |
| }, |
| { |
| "epoch": 2.37, |
| "grad_norm": 0.9039551615715027, |
| "learning_rate": 8.133427963094393e-05, |
| "loss": 0.7676, |
| "step": 1674 |
| }, |
| { |
| "epoch": 2.37, |
| "grad_norm": 0.8817799091339111, |
| "learning_rate": 8.126330731014905e-05, |
| "loss": 0.7223, |
| "step": 1675 |
| }, |
| { |
| "epoch": 2.37, |
| "grad_norm": 0.8838679194450378, |
| "learning_rate": 8.119233498935416e-05, |
| "loss": 0.7361, |
| "step": 1676 |
| }, |
| { |
| "epoch": 2.38, |
| "grad_norm": 0.9286802411079407, |
| "learning_rate": 8.112136266855926e-05, |
| "loss": 0.8229, |
| "step": 1677 |
| }, |
| { |
| "epoch": 2.38, |
| "grad_norm": 0.877018392086029, |
| "learning_rate": 8.105039034776437e-05, |
| "loss": 0.7307, |
| "step": 1678 |
| }, |
| { |
| "epoch": 2.38, |
| "grad_norm": 0.8911111950874329, |
| "learning_rate": 8.097941802696948e-05, |
| "loss": 0.7706, |
| "step": 1679 |
| }, |
| { |
| "epoch": 2.38, |
| "grad_norm": 0.8774330615997314, |
| "learning_rate": 8.09084457061746e-05, |
| "loss": 0.7347, |
| "step": 1680 |
| }, |
| { |
| "epoch": 2.38, |
| "grad_norm": 0.8946451544761658, |
| "learning_rate": 8.083747338537971e-05, |
| "loss": 0.7148, |
| "step": 1681 |
| }, |
| { |
| "epoch": 2.38, |
| "grad_norm": 0.9081972241401672, |
| "learning_rate": 8.076650106458482e-05, |
| "loss": 0.7809, |
| "step": 1682 |
| }, |
| { |
| "epoch": 2.38, |
| "grad_norm": 0.8362038731575012, |
| "learning_rate": 8.069552874378993e-05, |
| "loss": 0.7048, |
| "step": 1683 |
| }, |
| { |
| "epoch": 2.39, |
| "grad_norm": 0.8865955471992493, |
| "learning_rate": 8.062455642299503e-05, |
| "loss": 0.7437, |
| "step": 1684 |
| }, |
| { |
| "epoch": 2.39, |
| "grad_norm": 0.9173011779785156, |
| "learning_rate": 8.055358410220015e-05, |
| "loss": 0.7158, |
| "step": 1685 |
| }, |
| { |
| "epoch": 2.39, |
| "grad_norm": 0.8824982047080994, |
| "learning_rate": 8.048261178140526e-05, |
| "loss": 0.7485, |
| "step": 1686 |
| }, |
| { |
| "epoch": 2.39, |
| "grad_norm": 0.8951696753501892, |
| "learning_rate": 8.041163946061037e-05, |
| "loss": 0.7133, |
| "step": 1687 |
| }, |
| { |
| "epoch": 2.39, |
| "grad_norm": 0.8572990298271179, |
| "learning_rate": 8.034066713981547e-05, |
| "loss": 0.7082, |
| "step": 1688 |
| }, |
| { |
| "epoch": 2.39, |
| "grad_norm": 0.8780161738395691, |
| "learning_rate": 8.02696948190206e-05, |
| "loss": 0.6906, |
| "step": 1689 |
| }, |
| { |
| "epoch": 2.39, |
| "grad_norm": 0.8812801241874695, |
| "learning_rate": 8.01987224982257e-05, |
| "loss": 0.7179, |
| "step": 1690 |
| }, |
| { |
| "epoch": 2.4, |
| "grad_norm": 0.8629257082939148, |
| "learning_rate": 8.012775017743081e-05, |
| "loss": 0.6859, |
| "step": 1691 |
| }, |
| { |
| "epoch": 2.4, |
| "grad_norm": 0.8868771195411682, |
| "learning_rate": 8.005677785663591e-05, |
| "loss": 0.7254, |
| "step": 1692 |
| }, |
| { |
| "epoch": 2.4, |
| "grad_norm": 0.8882796764373779, |
| "learning_rate": 7.998580553584103e-05, |
| "loss": 0.7425, |
| "step": 1693 |
| }, |
| { |
| "epoch": 2.4, |
| "grad_norm": 0.9046310782432556, |
| "learning_rate": 7.991483321504613e-05, |
| "loss": 0.7593, |
| "step": 1694 |
| }, |
| { |
| "epoch": 2.4, |
| "grad_norm": 0.9165465831756592, |
| "learning_rate": 7.984386089425125e-05, |
| "loss": 0.7138, |
| "step": 1695 |
| }, |
| { |
| "epoch": 2.4, |
| "grad_norm": 0.8907314538955688, |
| "learning_rate": 7.977288857345636e-05, |
| "loss": 0.734, |
| "step": 1696 |
| }, |
| { |
| "epoch": 2.4, |
| "grad_norm": 0.9012109041213989, |
| "learning_rate": 7.970191625266147e-05, |
| "loss": 0.7136, |
| "step": 1697 |
| }, |
| { |
| "epoch": 2.41, |
| "grad_norm": 0.8788906335830688, |
| "learning_rate": 7.963094393186657e-05, |
| "loss": 0.7005, |
| "step": 1698 |
| }, |
| { |
| "epoch": 2.41, |
| "grad_norm": 0.9376303553581238, |
| "learning_rate": 7.955997161107168e-05, |
| "loss": 0.7704, |
| "step": 1699 |
| }, |
| { |
| "epoch": 2.41, |
| "grad_norm": 0.8864246606826782, |
| "learning_rate": 7.94889992902768e-05, |
| "loss": 0.723, |
| "step": 1700 |
| }, |
| { |
| "epoch": 2.41, |
| "grad_norm": 0.9175934791564941, |
| "learning_rate": 7.941802696948191e-05, |
| "loss": 0.744, |
| "step": 1701 |
| }, |
| { |
| "epoch": 2.41, |
| "grad_norm": 0.8643361926078796, |
| "learning_rate": 7.934705464868701e-05, |
| "loss": 0.7499, |
| "step": 1702 |
| }, |
| { |
| "epoch": 2.41, |
| "grad_norm": 0.901351809501648, |
| "learning_rate": 7.927608232789212e-05, |
| "loss": 0.7667, |
| "step": 1703 |
| }, |
| { |
| "epoch": 2.41, |
| "grad_norm": 0.8338555693626404, |
| "learning_rate": 7.920511000709723e-05, |
| "loss": 0.6274, |
| "step": 1704 |
| }, |
| { |
| "epoch": 2.42, |
| "grad_norm": 0.8955431580543518, |
| "learning_rate": 7.913413768630235e-05, |
| "loss": 0.7669, |
| "step": 1705 |
| }, |
| { |
| "epoch": 2.42, |
| "grad_norm": 0.8964597582817078, |
| "learning_rate": 7.906316536550745e-05, |
| "loss": 0.699, |
| "step": 1706 |
| }, |
| { |
| "epoch": 2.42, |
| "grad_norm": 0.8782421350479126, |
| "learning_rate": 7.899219304471257e-05, |
| "loss": 0.7406, |
| "step": 1707 |
| }, |
| { |
| "epoch": 2.42, |
| "grad_norm": 0.9423942565917969, |
| "learning_rate": 7.892122072391767e-05, |
| "loss": 0.7557, |
| "step": 1708 |
| }, |
| { |
| "epoch": 2.42, |
| "grad_norm": 0.8742157816886902, |
| "learning_rate": 7.885024840312278e-05, |
| "loss": 0.6471, |
| "step": 1709 |
| }, |
| { |
| "epoch": 2.42, |
| "grad_norm": 0.9231082201004028, |
| "learning_rate": 7.87792760823279e-05, |
| "loss": 0.7049, |
| "step": 1710 |
| }, |
| { |
| "epoch": 2.42, |
| "grad_norm": 0.9250949025154114, |
| "learning_rate": 7.870830376153301e-05, |
| "loss": 0.7393, |
| "step": 1711 |
| }, |
| { |
| "epoch": 2.43, |
| "grad_norm": 0.8757654428482056, |
| "learning_rate": 7.863733144073812e-05, |
| "loss": 0.7057, |
| "step": 1712 |
| }, |
| { |
| "epoch": 2.43, |
| "grad_norm": 0.8953589200973511, |
| "learning_rate": 7.856635911994322e-05, |
| "loss": 0.7286, |
| "step": 1713 |
| }, |
| { |
| "epoch": 2.43, |
| "grad_norm": 0.9081451296806335, |
| "learning_rate": 7.849538679914833e-05, |
| "loss": 0.7699, |
| "step": 1714 |
| }, |
| { |
| "epoch": 2.43, |
| "grad_norm": 0.8773189783096313, |
| "learning_rate": 7.842441447835345e-05, |
| "loss": 0.763, |
| "step": 1715 |
| }, |
| { |
| "epoch": 2.43, |
| "grad_norm": 0.8770549297332764, |
| "learning_rate": 7.835344215755856e-05, |
| "loss": 0.7245, |
| "step": 1716 |
| }, |
| { |
| "epoch": 2.43, |
| "grad_norm": 0.9066706895828247, |
| "learning_rate": 7.828246983676366e-05, |
| "loss": 0.7746, |
| "step": 1717 |
| }, |
| { |
| "epoch": 2.43, |
| "grad_norm": 0.901549756526947, |
| "learning_rate": 7.821149751596878e-05, |
| "loss": 0.7526, |
| "step": 1718 |
| }, |
| { |
| "epoch": 2.44, |
| "grad_norm": 0.894270122051239, |
| "learning_rate": 7.814052519517388e-05, |
| "loss": 0.7267, |
| "step": 1719 |
| }, |
| { |
| "epoch": 2.44, |
| "grad_norm": 0.9308816194534302, |
| "learning_rate": 7.8069552874379e-05, |
| "loss": 0.7116, |
| "step": 1720 |
| }, |
| { |
| "epoch": 2.44, |
| "grad_norm": 0.8835683465003967, |
| "learning_rate": 7.799858055358411e-05, |
| "loss": 0.726, |
| "step": 1721 |
| }, |
| { |
| "epoch": 2.44, |
| "grad_norm": 0.8589925169944763, |
| "learning_rate": 7.792760823278922e-05, |
| "loss": 0.7438, |
| "step": 1722 |
| }, |
| { |
| "epoch": 2.44, |
| "grad_norm": 0.910952091217041, |
| "learning_rate": 7.785663591199432e-05, |
| "loss": 0.7553, |
| "step": 1723 |
| }, |
| { |
| "epoch": 2.44, |
| "grad_norm": 0.8601446747779846, |
| "learning_rate": 7.778566359119943e-05, |
| "loss": 0.7107, |
| "step": 1724 |
| }, |
| { |
| "epoch": 2.44, |
| "grad_norm": 0.8741260170936584, |
| "learning_rate": 7.771469127040455e-05, |
| "loss": 0.7428, |
| "step": 1725 |
| }, |
| { |
| "epoch": 2.45, |
| "grad_norm": 0.9290738701820374, |
| "learning_rate": 7.764371894960966e-05, |
| "loss": 0.7675, |
| "step": 1726 |
| }, |
| { |
| "epoch": 2.45, |
| "grad_norm": 0.8710927963256836, |
| "learning_rate": 7.757274662881476e-05, |
| "loss": 0.6856, |
| "step": 1727 |
| }, |
| { |
| "epoch": 2.45, |
| "grad_norm": 0.9226647615432739, |
| "learning_rate": 7.750177430801989e-05, |
| "loss": 0.8229, |
| "step": 1728 |
| }, |
| { |
| "epoch": 2.45, |
| "grad_norm": 0.8699370622634888, |
| "learning_rate": 7.743080198722498e-05, |
| "loss": 0.7268, |
| "step": 1729 |
| }, |
| { |
| "epoch": 2.45, |
| "grad_norm": 0.8959556221961975, |
| "learning_rate": 7.73598296664301e-05, |
| "loss": 0.7698, |
| "step": 1730 |
| }, |
| { |
| "epoch": 2.45, |
| "grad_norm": 0.8745953440666199, |
| "learning_rate": 7.72888573456352e-05, |
| "loss": 0.7181, |
| "step": 1731 |
| }, |
| { |
| "epoch": 2.45, |
| "grad_norm": 0.8749833703041077, |
| "learning_rate": 7.721788502484032e-05, |
| "loss": 0.792, |
| "step": 1732 |
| }, |
| { |
| "epoch": 2.46, |
| "grad_norm": 0.9342421889305115, |
| "learning_rate": 7.714691270404542e-05, |
| "loss": 0.7907, |
| "step": 1733 |
| }, |
| { |
| "epoch": 2.46, |
| "grad_norm": 0.8828331232070923, |
| "learning_rate": 7.707594038325053e-05, |
| "loss": 0.708, |
| "step": 1734 |
| }, |
| { |
| "epoch": 2.46, |
| "grad_norm": 0.8733470439910889, |
| "learning_rate": 7.700496806245565e-05, |
| "loss": 0.7326, |
| "step": 1735 |
| }, |
| { |
| "epoch": 2.46, |
| "grad_norm": 0.9026939272880554, |
| "learning_rate": 7.693399574166076e-05, |
| "loss": 0.7149, |
| "step": 1736 |
| }, |
| { |
| "epoch": 2.46, |
| "grad_norm": 0.870780348777771, |
| "learning_rate": 7.686302342086586e-05, |
| "loss": 0.745, |
| "step": 1737 |
| }, |
| { |
| "epoch": 2.46, |
| "grad_norm": 0.8790858387947083, |
| "learning_rate": 7.679205110007097e-05, |
| "loss": 0.7532, |
| "step": 1738 |
| }, |
| { |
| "epoch": 2.46, |
| "grad_norm": 0.8772610425949097, |
| "learning_rate": 7.672107877927608e-05, |
| "loss": 0.6456, |
| "step": 1739 |
| }, |
| { |
| "epoch": 2.47, |
| "grad_norm": 0.9049152731895447, |
| "learning_rate": 7.66501064584812e-05, |
| "loss": 0.7901, |
| "step": 1740 |
| }, |
| { |
| "epoch": 2.47, |
| "grad_norm": 0.8748606443405151, |
| "learning_rate": 7.657913413768631e-05, |
| "loss": 0.7347, |
| "step": 1741 |
| }, |
| { |
| "epoch": 2.47, |
| "grad_norm": 0.8891382813453674, |
| "learning_rate": 7.650816181689141e-05, |
| "loss": 0.6857, |
| "step": 1742 |
| }, |
| { |
| "epoch": 2.47, |
| "grad_norm": 0.9277533292770386, |
| "learning_rate": 7.643718949609652e-05, |
| "loss": 0.7585, |
| "step": 1743 |
| }, |
| { |
| "epoch": 2.47, |
| "grad_norm": 0.8965739011764526, |
| "learning_rate": 7.636621717530163e-05, |
| "loss": 0.7177, |
| "step": 1744 |
| }, |
| { |
| "epoch": 2.47, |
| "grad_norm": 0.8912781476974487, |
| "learning_rate": 7.629524485450675e-05, |
| "loss": 0.7499, |
| "step": 1745 |
| }, |
| { |
| "epoch": 2.47, |
| "grad_norm": 0.925321102142334, |
| "learning_rate": 7.622427253371186e-05, |
| "loss": 0.7524, |
| "step": 1746 |
| }, |
| { |
| "epoch": 2.47, |
| "grad_norm": 0.888886570930481, |
| "learning_rate": 7.615330021291697e-05, |
| "loss": 0.7251, |
| "step": 1747 |
| }, |
| { |
| "epoch": 2.48, |
| "grad_norm": 0.8739018440246582, |
| "learning_rate": 7.608232789212207e-05, |
| "loss": 0.7421, |
| "step": 1748 |
| }, |
| { |
| "epoch": 2.48, |
| "grad_norm": 0.8653773665428162, |
| "learning_rate": 7.601135557132718e-05, |
| "loss": 0.7121, |
| "step": 1749 |
| }, |
| { |
| "epoch": 2.48, |
| "grad_norm": 0.9399885535240173, |
| "learning_rate": 7.59403832505323e-05, |
| "loss": 0.8104, |
| "step": 1750 |
| }, |
| { |
| "epoch": 2.48, |
| "grad_norm": 0.8935320377349854, |
| "learning_rate": 7.586941092973741e-05, |
| "loss": 0.7285, |
| "step": 1751 |
| }, |
| { |
| "epoch": 2.48, |
| "grad_norm": 0.8893241286277771, |
| "learning_rate": 7.579843860894251e-05, |
| "loss": 0.711, |
| "step": 1752 |
| }, |
| { |
| "epoch": 2.48, |
| "grad_norm": 0.8903502225875854, |
| "learning_rate": 7.572746628814764e-05, |
| "loss": 0.74, |
| "step": 1753 |
| }, |
| { |
| "epoch": 2.48, |
| "grad_norm": 0.8928572535514832, |
| "learning_rate": 7.565649396735273e-05, |
| "loss": 0.7333, |
| "step": 1754 |
| }, |
| { |
| "epoch": 2.49, |
| "grad_norm": 0.8857137560844421, |
| "learning_rate": 7.558552164655785e-05, |
| "loss": 0.7382, |
| "step": 1755 |
| }, |
| { |
| "epoch": 2.49, |
| "grad_norm": 0.9179412722587585, |
| "learning_rate": 7.551454932576295e-05, |
| "loss": 0.7226, |
| "step": 1756 |
| }, |
| { |
| "epoch": 2.49, |
| "grad_norm": 0.9129919409751892, |
| "learning_rate": 7.544357700496807e-05, |
| "loss": 0.7687, |
| "step": 1757 |
| }, |
| { |
| "epoch": 2.49, |
| "grad_norm": 0.9090712666511536, |
| "learning_rate": 7.537260468417317e-05, |
| "loss": 0.7881, |
| "step": 1758 |
| }, |
| { |
| "epoch": 2.49, |
| "grad_norm": 0.9000692963600159, |
| "learning_rate": 7.530163236337828e-05, |
| "loss": 0.7255, |
| "step": 1759 |
| }, |
| { |
| "epoch": 2.49, |
| "grad_norm": 0.8983575701713562, |
| "learning_rate": 7.52306600425834e-05, |
| "loss": 0.7276, |
| "step": 1760 |
| }, |
| { |
| "epoch": 2.49, |
| "grad_norm": 0.9295850396156311, |
| "learning_rate": 7.515968772178851e-05, |
| "loss": 0.7576, |
| "step": 1761 |
| }, |
| { |
| "epoch": 2.5, |
| "grad_norm": 0.8923923373222351, |
| "learning_rate": 7.508871540099361e-05, |
| "loss": 0.7553, |
| "step": 1762 |
| }, |
| { |
| "epoch": 2.5, |
| "grad_norm": 0.9035564064979553, |
| "learning_rate": 7.501774308019872e-05, |
| "loss": 0.757, |
| "step": 1763 |
| }, |
| { |
| "epoch": 2.5, |
| "grad_norm": 0.8981541991233826, |
| "learning_rate": 7.494677075940383e-05, |
| "loss": 0.7188, |
| "step": 1764 |
| }, |
| { |
| "epoch": 2.5, |
| "grad_norm": 0.9119907021522522, |
| "learning_rate": 7.487579843860895e-05, |
| "loss": 0.73, |
| "step": 1765 |
| }, |
| { |
| "epoch": 2.5, |
| "grad_norm": 0.9041171669960022, |
| "learning_rate": 7.480482611781405e-05, |
| "loss": 0.7081, |
| "step": 1766 |
| }, |
| { |
| "epoch": 2.5, |
| "grad_norm": 0.8893494606018066, |
| "learning_rate": 7.473385379701917e-05, |
| "loss": 0.7115, |
| "step": 1767 |
| }, |
| { |
| "epoch": 2.5, |
| "grad_norm": 0.9286924004554749, |
| "learning_rate": 7.466288147622427e-05, |
| "loss": 0.7761, |
| "step": 1768 |
| }, |
| { |
| "epoch": 2.51, |
| "grad_norm": 0.9019243121147156, |
| "learning_rate": 7.459190915542938e-05, |
| "loss": 0.6966, |
| "step": 1769 |
| }, |
| { |
| "epoch": 2.51, |
| "grad_norm": 0.88730788230896, |
| "learning_rate": 7.45209368346345e-05, |
| "loss": 0.6791, |
| "step": 1770 |
| }, |
| { |
| "epoch": 2.51, |
| "grad_norm": 0.9363870620727539, |
| "learning_rate": 7.444996451383961e-05, |
| "loss": 0.7713, |
| "step": 1771 |
| }, |
| { |
| "epoch": 2.51, |
| "grad_norm": 0.8667078614234924, |
| "learning_rate": 7.437899219304472e-05, |
| "loss": 0.7231, |
| "step": 1772 |
| }, |
| { |
| "epoch": 2.51, |
| "grad_norm": 0.91669100522995, |
| "learning_rate": 7.430801987224982e-05, |
| "loss": 0.7306, |
| "step": 1773 |
| }, |
| { |
| "epoch": 2.51, |
| "grad_norm": 0.9173486828804016, |
| "learning_rate": 7.423704755145493e-05, |
| "loss": 0.714, |
| "step": 1774 |
| }, |
| { |
| "epoch": 2.51, |
| "grad_norm": 0.931074321269989, |
| "learning_rate": 7.416607523066005e-05, |
| "loss": 0.7594, |
| "step": 1775 |
| }, |
| { |
| "epoch": 2.52, |
| "grad_norm": 0.8770849704742432, |
| "learning_rate": 7.409510290986516e-05, |
| "loss": 0.7178, |
| "step": 1776 |
| }, |
| { |
| "epoch": 2.52, |
| "grad_norm": 0.9106402397155762, |
| "learning_rate": 7.402413058907026e-05, |
| "loss": 0.732, |
| "step": 1777 |
| }, |
| { |
| "epoch": 2.52, |
| "grad_norm": 0.9002892374992371, |
| "learning_rate": 7.395315826827539e-05, |
| "loss": 0.7282, |
| "step": 1778 |
| }, |
| { |
| "epoch": 2.52, |
| "grad_norm": 0.9161092638969421, |
| "learning_rate": 7.388218594748048e-05, |
| "loss": 0.7786, |
| "step": 1779 |
| }, |
| { |
| "epoch": 2.52, |
| "grad_norm": 0.9152305126190186, |
| "learning_rate": 7.38112136266856e-05, |
| "loss": 0.7226, |
| "step": 1780 |
| }, |
| { |
| "epoch": 2.52, |
| "grad_norm": 0.8772377371788025, |
| "learning_rate": 7.374024130589071e-05, |
| "loss": 0.7233, |
| "step": 1781 |
| }, |
| { |
| "epoch": 2.52, |
| "grad_norm": 0.9021669030189514, |
| "learning_rate": 7.366926898509582e-05, |
| "loss": 0.7336, |
| "step": 1782 |
| }, |
| { |
| "epoch": 2.53, |
| "grad_norm": 0.8629419207572937, |
| "learning_rate": 7.359829666430092e-05, |
| "loss": 0.7086, |
| "step": 1783 |
| }, |
| { |
| "epoch": 2.53, |
| "grad_norm": 0.8985278010368347, |
| "learning_rate": 7.352732434350604e-05, |
| "loss": 0.7495, |
| "step": 1784 |
| }, |
| { |
| "epoch": 2.53, |
| "grad_norm": 0.8771053552627563, |
| "learning_rate": 7.345635202271115e-05, |
| "loss": 0.7, |
| "step": 1785 |
| }, |
| { |
| "epoch": 2.53, |
| "grad_norm": 0.8981669545173645, |
| "learning_rate": 7.338537970191626e-05, |
| "loss": 0.6897, |
| "step": 1786 |
| }, |
| { |
| "epoch": 2.53, |
| "grad_norm": 0.9112007021903992, |
| "learning_rate": 7.331440738112136e-05, |
| "loss": 0.7408, |
| "step": 1787 |
| }, |
| { |
| "epoch": 2.53, |
| "grad_norm": 0.9054397940635681, |
| "learning_rate": 7.324343506032647e-05, |
| "loss": 0.7054, |
| "step": 1788 |
| }, |
| { |
| "epoch": 2.53, |
| "grad_norm": 0.869867205619812, |
| "learning_rate": 7.317246273953159e-05, |
| "loss": 0.6891, |
| "step": 1789 |
| }, |
| { |
| "epoch": 2.54, |
| "grad_norm": 0.90545654296875, |
| "learning_rate": 7.31014904187367e-05, |
| "loss": 0.7293, |
| "step": 1790 |
| }, |
| { |
| "epoch": 2.54, |
| "grad_norm": 0.9081699848175049, |
| "learning_rate": 7.30305180979418e-05, |
| "loss": 0.7445, |
| "step": 1791 |
| }, |
| { |
| "epoch": 2.54, |
| "grad_norm": 0.9117109775543213, |
| "learning_rate": 7.295954577714692e-05, |
| "loss": 0.7802, |
| "step": 1792 |
| }, |
| { |
| "epoch": 2.54, |
| "grad_norm": 0.9705706834793091, |
| "learning_rate": 7.288857345635202e-05, |
| "loss": 0.7802, |
| "step": 1793 |
| }, |
| { |
| "epoch": 2.54, |
| "grad_norm": 0.9153757095336914, |
| "learning_rate": 7.281760113555714e-05, |
| "loss": 0.7444, |
| "step": 1794 |
| }, |
| { |
| "epoch": 2.54, |
| "grad_norm": 0.8785500526428223, |
| "learning_rate": 7.274662881476223e-05, |
| "loss": 0.6899, |
| "step": 1795 |
| }, |
| { |
| "epoch": 2.54, |
| "grad_norm": 0.9095218777656555, |
| "learning_rate": 7.267565649396736e-05, |
| "loss": 0.7234, |
| "step": 1796 |
| }, |
| { |
| "epoch": 2.55, |
| "grad_norm": 0.8905136585235596, |
| "learning_rate": 7.260468417317246e-05, |
| "loss": 0.7169, |
| "step": 1797 |
| }, |
| { |
| "epoch": 2.55, |
| "grad_norm": 0.9118934869766235, |
| "learning_rate": 7.253371185237757e-05, |
| "loss": 0.7301, |
| "step": 1798 |
| }, |
| { |
| "epoch": 2.55, |
| "grad_norm": 0.9225103259086609, |
| "learning_rate": 7.246273953158269e-05, |
| "loss": 0.798, |
| "step": 1799 |
| }, |
| { |
| "epoch": 2.55, |
| "grad_norm": 0.8873489499092102, |
| "learning_rate": 7.23917672107878e-05, |
| "loss": 0.7378, |
| "step": 1800 |
| }, |
| { |
| "epoch": 2.55, |
| "grad_norm": 0.8982245922088623, |
| "learning_rate": 7.232079488999291e-05, |
| "loss": 0.7738, |
| "step": 1801 |
| }, |
| { |
| "epoch": 2.55, |
| "grad_norm": 0.8934197425842285, |
| "learning_rate": 7.224982256919801e-05, |
| "loss": 0.6641, |
| "step": 1802 |
| }, |
| { |
| "epoch": 2.55, |
| "grad_norm": 0.9242858290672302, |
| "learning_rate": 7.217885024840312e-05, |
| "loss": 0.7361, |
| "step": 1803 |
| }, |
| { |
| "epoch": 2.56, |
| "grad_norm": 0.8764644861221313, |
| "learning_rate": 7.210787792760824e-05, |
| "loss": 0.7311, |
| "step": 1804 |
| }, |
| { |
| "epoch": 2.56, |
| "grad_norm": 0.8844506144523621, |
| "learning_rate": 7.203690560681335e-05, |
| "loss": 0.7216, |
| "step": 1805 |
| }, |
| { |
| "epoch": 2.56, |
| "grad_norm": 0.8994446396827698, |
| "learning_rate": 7.196593328601846e-05, |
| "loss": 0.7009, |
| "step": 1806 |
| }, |
| { |
| "epoch": 2.56, |
| "grad_norm": 0.9672061204910278, |
| "learning_rate": 7.189496096522357e-05, |
| "loss": 0.8095, |
| "step": 1807 |
| }, |
| { |
| "epoch": 2.56, |
| "grad_norm": 0.929573655128479, |
| "learning_rate": 7.182398864442867e-05, |
| "loss": 0.7629, |
| "step": 1808 |
| }, |
| { |
| "epoch": 2.56, |
| "grad_norm": 0.8954296708106995, |
| "learning_rate": 7.175301632363379e-05, |
| "loss": 0.7393, |
| "step": 1809 |
| }, |
| { |
| "epoch": 2.56, |
| "grad_norm": 0.9068436026573181, |
| "learning_rate": 7.16820440028389e-05, |
| "loss": 0.7749, |
| "step": 1810 |
| }, |
| { |
| "epoch": 2.57, |
| "grad_norm": 0.8931055665016174, |
| "learning_rate": 7.161107168204401e-05, |
| "loss": 0.7452, |
| "step": 1811 |
| }, |
| { |
| "epoch": 2.57, |
| "grad_norm": 0.9035736918449402, |
| "learning_rate": 7.154009936124911e-05, |
| "loss": 0.7888, |
| "step": 1812 |
| }, |
| { |
| "epoch": 2.57, |
| "grad_norm": 0.9025890231132507, |
| "learning_rate": 7.146912704045424e-05, |
| "loss": 0.7025, |
| "step": 1813 |
| }, |
| { |
| "epoch": 2.57, |
| "grad_norm": 0.8863286972045898, |
| "learning_rate": 7.139815471965934e-05, |
| "loss": 0.7307, |
| "step": 1814 |
| }, |
| { |
| "epoch": 2.57, |
| "grad_norm": 0.8847543597221375, |
| "learning_rate": 7.132718239886445e-05, |
| "loss": 0.7458, |
| "step": 1815 |
| }, |
| { |
| "epoch": 2.57, |
| "grad_norm": 0.8659204840660095, |
| "learning_rate": 7.125621007806955e-05, |
| "loss": 0.753, |
| "step": 1816 |
| }, |
| { |
| "epoch": 2.57, |
| "grad_norm": 0.9039578437805176, |
| "learning_rate": 7.118523775727467e-05, |
| "loss": 0.7161, |
| "step": 1817 |
| }, |
| { |
| "epoch": 2.58, |
| "grad_norm": 0.935804545879364, |
| "learning_rate": 7.111426543647977e-05, |
| "loss": 0.7696, |
| "step": 1818 |
| }, |
| { |
| "epoch": 2.58, |
| "grad_norm": 0.8946291208267212, |
| "learning_rate": 7.104329311568489e-05, |
| "loss": 0.743, |
| "step": 1819 |
| }, |
| { |
| "epoch": 2.58, |
| "grad_norm": 0.9393401145935059, |
| "learning_rate": 7.097232079489e-05, |
| "loss": 0.737, |
| "step": 1820 |
| }, |
| { |
| "epoch": 2.58, |
| "grad_norm": 0.8611619472503662, |
| "learning_rate": 7.090134847409511e-05, |
| "loss": 0.6716, |
| "step": 1821 |
| }, |
| { |
| "epoch": 2.58, |
| "grad_norm": 0.8942080140113831, |
| "learning_rate": 7.083037615330021e-05, |
| "loss": 0.7581, |
| "step": 1822 |
| }, |
| { |
| "epoch": 2.58, |
| "grad_norm": 0.9117892980575562, |
| "learning_rate": 7.075940383250532e-05, |
| "loss": 0.7086, |
| "step": 1823 |
| }, |
| { |
| "epoch": 2.58, |
| "grad_norm": 0.8960133790969849, |
| "learning_rate": 7.068843151171044e-05, |
| "loss": 0.7807, |
| "step": 1824 |
| }, |
| { |
| "epoch": 2.59, |
| "grad_norm": 0.9127882719039917, |
| "learning_rate": 7.061745919091555e-05, |
| "loss": 0.7258, |
| "step": 1825 |
| }, |
| { |
| "epoch": 2.59, |
| "grad_norm": 0.9065718650817871, |
| "learning_rate": 7.054648687012065e-05, |
| "loss": 0.7771, |
| "step": 1826 |
| }, |
| { |
| "epoch": 2.59, |
| "grad_norm": 0.9114099144935608, |
| "learning_rate": 7.047551454932577e-05, |
| "loss": 0.7564, |
| "step": 1827 |
| }, |
| { |
| "epoch": 2.59, |
| "grad_norm": 0.8710154294967651, |
| "learning_rate": 7.040454222853087e-05, |
| "loss": 0.691, |
| "step": 1828 |
| }, |
| { |
| "epoch": 2.59, |
| "grad_norm": 0.8883103132247925, |
| "learning_rate": 7.033356990773599e-05, |
| "loss": 0.797, |
| "step": 1829 |
| }, |
| { |
| "epoch": 2.59, |
| "grad_norm": 0.8676865100860596, |
| "learning_rate": 7.02625975869411e-05, |
| "loss": 0.707, |
| "step": 1830 |
| }, |
| { |
| "epoch": 2.59, |
| "grad_norm": 0.8942274451255798, |
| "learning_rate": 7.019162526614621e-05, |
| "loss": 0.7155, |
| "step": 1831 |
| }, |
| { |
| "epoch": 2.6, |
| "grad_norm": 0.9445174336433411, |
| "learning_rate": 7.012065294535132e-05, |
| "loss": 0.7967, |
| "step": 1832 |
| }, |
| { |
| "epoch": 2.6, |
| "grad_norm": 0.9001713991165161, |
| "learning_rate": 7.004968062455642e-05, |
| "loss": 0.7332, |
| "step": 1833 |
| }, |
| { |
| "epoch": 2.6, |
| "grad_norm": 0.9131384491920471, |
| "learning_rate": 6.997870830376154e-05, |
| "loss": 0.719, |
| "step": 1834 |
| }, |
| { |
| "epoch": 2.6, |
| "grad_norm": 0.9131073355674744, |
| "learning_rate": 6.990773598296665e-05, |
| "loss": 0.7155, |
| "step": 1835 |
| }, |
| { |
| "epoch": 2.6, |
| "grad_norm": 0.9039396643638611, |
| "learning_rate": 6.983676366217176e-05, |
| "loss": 0.7428, |
| "step": 1836 |
| }, |
| { |
| "epoch": 2.6, |
| "grad_norm": 0.9080092906951904, |
| "learning_rate": 6.976579134137686e-05, |
| "loss": 0.7161, |
| "step": 1837 |
| }, |
| { |
| "epoch": 2.6, |
| "grad_norm": 0.8788297772407532, |
| "learning_rate": 6.969481902058199e-05, |
| "loss": 0.7092, |
| "step": 1838 |
| }, |
| { |
| "epoch": 2.61, |
| "grad_norm": 0.8493945598602295, |
| "learning_rate": 6.962384669978709e-05, |
| "loss": 0.6627, |
| "step": 1839 |
| }, |
| { |
| "epoch": 2.61, |
| "grad_norm": 0.9463279843330383, |
| "learning_rate": 6.95528743789922e-05, |
| "loss": 0.7828, |
| "step": 1840 |
| }, |
| { |
| "epoch": 2.61, |
| "grad_norm": 0.9215744733810425, |
| "learning_rate": 6.94819020581973e-05, |
| "loss": 0.7036, |
| "step": 1841 |
| }, |
| { |
| "epoch": 2.61, |
| "grad_norm": 0.882713258266449, |
| "learning_rate": 6.941092973740242e-05, |
| "loss": 0.7174, |
| "step": 1842 |
| }, |
| { |
| "epoch": 2.61, |
| "grad_norm": 0.8597182631492615, |
| "learning_rate": 6.933995741660752e-05, |
| "loss": 0.7197, |
| "step": 1843 |
| }, |
| { |
| "epoch": 2.61, |
| "grad_norm": 0.8798161745071411, |
| "learning_rate": 6.926898509581264e-05, |
| "loss": 0.7363, |
| "step": 1844 |
| }, |
| { |
| "epoch": 2.61, |
| "grad_norm": 0.8880274891853333, |
| "learning_rate": 6.919801277501775e-05, |
| "loss": 0.7668, |
| "step": 1845 |
| }, |
| { |
| "epoch": 2.62, |
| "grad_norm": 0.9019714593887329, |
| "learning_rate": 6.912704045422286e-05, |
| "loss": 0.7559, |
| "step": 1846 |
| }, |
| { |
| "epoch": 2.62, |
| "grad_norm": 0.9086313843727112, |
| "learning_rate": 6.905606813342796e-05, |
| "loss": 0.747, |
| "step": 1847 |
| }, |
| { |
| "epoch": 2.62, |
| "grad_norm": 0.9187281131744385, |
| "learning_rate": 6.898509581263307e-05, |
| "loss": 0.7866, |
| "step": 1848 |
| }, |
| { |
| "epoch": 2.62, |
| "grad_norm": 0.9017612338066101, |
| "learning_rate": 6.891412349183819e-05, |
| "loss": 0.7581, |
| "step": 1849 |
| }, |
| { |
| "epoch": 2.62, |
| "grad_norm": 0.9309147000312805, |
| "learning_rate": 6.88431511710433e-05, |
| "loss": 0.7512, |
| "step": 1850 |
| }, |
| { |
| "epoch": 2.62, |
| "grad_norm": 0.8845265507698059, |
| "learning_rate": 6.87721788502484e-05, |
| "loss": 0.7758, |
| "step": 1851 |
| }, |
| { |
| "epoch": 2.62, |
| "grad_norm": 0.8595403432846069, |
| "learning_rate": 6.870120652945352e-05, |
| "loss": 0.665, |
| "step": 1852 |
| }, |
| { |
| "epoch": 2.63, |
| "grad_norm": 0.9455280303955078, |
| "learning_rate": 6.863023420865862e-05, |
| "loss": 0.7181, |
| "step": 1853 |
| }, |
| { |
| "epoch": 2.63, |
| "grad_norm": 0.8818469047546387, |
| "learning_rate": 6.855926188786374e-05, |
| "loss": 0.6943, |
| "step": 1854 |
| }, |
| { |
| "epoch": 2.63, |
| "grad_norm": 0.8975241780281067, |
| "learning_rate": 6.848828956706884e-05, |
| "loss": 0.7303, |
| "step": 1855 |
| }, |
| { |
| "epoch": 2.63, |
| "grad_norm": 0.8924829363822937, |
| "learning_rate": 6.841731724627396e-05, |
| "loss": 0.7318, |
| "step": 1856 |
| }, |
| { |
| "epoch": 2.63, |
| "grad_norm": 0.9156109690666199, |
| "learning_rate": 6.834634492547906e-05, |
| "loss": 0.7303, |
| "step": 1857 |
| }, |
| { |
| "epoch": 2.63, |
| "grad_norm": 0.8979274034500122, |
| "learning_rate": 6.827537260468417e-05, |
| "loss": 0.7247, |
| "step": 1858 |
| }, |
| { |
| "epoch": 2.63, |
| "grad_norm": 0.9190047979354858, |
| "learning_rate": 6.820440028388929e-05, |
| "loss": 0.7229, |
| "step": 1859 |
| }, |
| { |
| "epoch": 2.64, |
| "grad_norm": 0.8945170640945435, |
| "learning_rate": 6.81334279630944e-05, |
| "loss": 0.7494, |
| "step": 1860 |
| }, |
| { |
| "epoch": 2.64, |
| "grad_norm": 0.9074508547782898, |
| "learning_rate": 6.806245564229951e-05, |
| "loss": 0.7074, |
| "step": 1861 |
| }, |
| { |
| "epoch": 2.64, |
| "grad_norm": 0.884390115737915, |
| "learning_rate": 6.799148332150461e-05, |
| "loss": 0.6879, |
| "step": 1862 |
| }, |
| { |
| "epoch": 2.64, |
| "grad_norm": 0.8993882536888123, |
| "learning_rate": 6.792051100070972e-05, |
| "loss": 0.7678, |
| "step": 1863 |
| }, |
| { |
| "epoch": 2.64, |
| "grad_norm": 0.9103842377662659, |
| "learning_rate": 6.784953867991484e-05, |
| "loss": 0.6994, |
| "step": 1864 |
| }, |
| { |
| "epoch": 2.64, |
| "grad_norm": 0.8733007311820984, |
| "learning_rate": 6.777856635911995e-05, |
| "loss": 0.6775, |
| "step": 1865 |
| }, |
| { |
| "epoch": 2.64, |
| "grad_norm": 0.9129227995872498, |
| "learning_rate": 6.770759403832506e-05, |
| "loss": 0.7798, |
| "step": 1866 |
| }, |
| { |
| "epoch": 2.64, |
| "grad_norm": 0.8990234732627869, |
| "learning_rate": 6.763662171753017e-05, |
| "loss": 0.7268, |
| "step": 1867 |
| }, |
| { |
| "epoch": 2.65, |
| "grad_norm": 0.92862868309021, |
| "learning_rate": 6.756564939673527e-05, |
| "loss": 0.7577, |
| "step": 1868 |
| }, |
| { |
| "epoch": 2.65, |
| "grad_norm": 0.901336133480072, |
| "learning_rate": 6.749467707594039e-05, |
| "loss": 0.7165, |
| "step": 1869 |
| }, |
| { |
| "epoch": 2.65, |
| "grad_norm": 0.8477864861488342, |
| "learning_rate": 6.74237047551455e-05, |
| "loss": 0.6478, |
| "step": 1870 |
| }, |
| { |
| "epoch": 2.65, |
| "grad_norm": 0.8649226427078247, |
| "learning_rate": 6.735273243435061e-05, |
| "loss": 0.7218, |
| "step": 1871 |
| }, |
| { |
| "epoch": 2.65, |
| "grad_norm": 0.895254373550415, |
| "learning_rate": 6.728176011355571e-05, |
| "loss": 0.7311, |
| "step": 1872 |
| }, |
| { |
| "epoch": 2.65, |
| "grad_norm": 0.8805977702140808, |
| "learning_rate": 6.721078779276084e-05, |
| "loss": 0.6869, |
| "step": 1873 |
| }, |
| { |
| "epoch": 2.65, |
| "grad_norm": 0.8643150329589844, |
| "learning_rate": 6.713981547196594e-05, |
| "loss": 0.6596, |
| "step": 1874 |
| }, |
| { |
| "epoch": 2.66, |
| "grad_norm": 0.9333701133728027, |
| "learning_rate": 6.706884315117105e-05, |
| "loss": 0.732, |
| "step": 1875 |
| }, |
| { |
| "epoch": 2.66, |
| "grad_norm": 0.9489626884460449, |
| "learning_rate": 6.699787083037615e-05, |
| "loss": 0.7647, |
| "step": 1876 |
| }, |
| { |
| "epoch": 2.66, |
| "grad_norm": 0.8754850029945374, |
| "learning_rate": 6.692689850958127e-05, |
| "loss": 0.7136, |
| "step": 1877 |
| }, |
| { |
| "epoch": 2.66, |
| "grad_norm": 0.95391845703125, |
| "learning_rate": 6.685592618878637e-05, |
| "loss": 0.7988, |
| "step": 1878 |
| }, |
| { |
| "epoch": 2.66, |
| "grad_norm": 0.9592831134796143, |
| "learning_rate": 6.678495386799149e-05, |
| "loss": 0.7275, |
| "step": 1879 |
| }, |
| { |
| "epoch": 2.66, |
| "grad_norm": 0.9160040616989136, |
| "learning_rate": 6.671398154719659e-05, |
| "loss": 0.7331, |
| "step": 1880 |
| }, |
| { |
| "epoch": 2.66, |
| "grad_norm": 0.9226792454719543, |
| "learning_rate": 6.664300922640171e-05, |
| "loss": 0.7482, |
| "step": 1881 |
| }, |
| { |
| "epoch": 2.67, |
| "grad_norm": 0.9255938529968262, |
| "learning_rate": 6.657203690560681e-05, |
| "loss": 0.7231, |
| "step": 1882 |
| }, |
| { |
| "epoch": 2.67, |
| "grad_norm": 0.8794947862625122, |
| "learning_rate": 6.650106458481192e-05, |
| "loss": 0.6833, |
| "step": 1883 |
| }, |
| { |
| "epoch": 2.67, |
| "grad_norm": 0.9187234044075012, |
| "learning_rate": 6.643009226401704e-05, |
| "loss": 0.7916, |
| "step": 1884 |
| }, |
| { |
| "epoch": 2.67, |
| "grad_norm": 0.933605432510376, |
| "learning_rate": 6.635911994322215e-05, |
| "loss": 0.7747, |
| "step": 1885 |
| }, |
| { |
| "epoch": 2.67, |
| "grad_norm": 0.904399037361145, |
| "learning_rate": 6.628814762242725e-05, |
| "loss": 0.7911, |
| "step": 1886 |
| }, |
| { |
| "epoch": 2.67, |
| "grad_norm": 0.8907902240753174, |
| "learning_rate": 6.621717530163236e-05, |
| "loss": 0.7408, |
| "step": 1887 |
| }, |
| { |
| "epoch": 2.67, |
| "grad_norm": 0.865777313709259, |
| "learning_rate": 6.614620298083747e-05, |
| "loss": 0.6625, |
| "step": 1888 |
| }, |
| { |
| "epoch": 2.68, |
| "grad_norm": 0.9031429886817932, |
| "learning_rate": 6.607523066004259e-05, |
| "loss": 0.7341, |
| "step": 1889 |
| }, |
| { |
| "epoch": 2.68, |
| "grad_norm": 0.9169644713401794, |
| "learning_rate": 6.60042583392477e-05, |
| "loss": 0.7321, |
| "step": 1890 |
| }, |
| { |
| "epoch": 2.68, |
| "grad_norm": 0.9195284247398376, |
| "learning_rate": 6.593328601845281e-05, |
| "loss": 0.7736, |
| "step": 1891 |
| }, |
| { |
| "epoch": 2.68, |
| "grad_norm": 0.9451195001602173, |
| "learning_rate": 6.586231369765791e-05, |
| "loss": 0.8059, |
| "step": 1892 |
| }, |
| { |
| "epoch": 2.68, |
| "grad_norm": 0.8922215700149536, |
| "learning_rate": 6.579134137686302e-05, |
| "loss": 0.6853, |
| "step": 1893 |
| }, |
| { |
| "epoch": 2.68, |
| "grad_norm": 0.9502759575843811, |
| "learning_rate": 6.572036905606814e-05, |
| "loss": 0.719, |
| "step": 1894 |
| }, |
| { |
| "epoch": 2.68, |
| "grad_norm": 0.9210609793663025, |
| "learning_rate": 6.564939673527325e-05, |
| "loss": 0.726, |
| "step": 1895 |
| }, |
| { |
| "epoch": 2.69, |
| "grad_norm": 0.9551956653594971, |
| "learning_rate": 6.557842441447836e-05, |
| "loss": 0.8149, |
| "step": 1896 |
| }, |
| { |
| "epoch": 2.69, |
| "grad_norm": 0.8991343379020691, |
| "learning_rate": 6.550745209368346e-05, |
| "loss": 0.7282, |
| "step": 1897 |
| }, |
| { |
| "epoch": 2.69, |
| "grad_norm": 0.8829290866851807, |
| "learning_rate": 6.543647977288859e-05, |
| "loss": 0.7694, |
| "step": 1898 |
| }, |
| { |
| "epoch": 2.69, |
| "grad_norm": 0.9269784092903137, |
| "learning_rate": 6.536550745209369e-05, |
| "loss": 0.7331, |
| "step": 1899 |
| }, |
| { |
| "epoch": 2.69, |
| "grad_norm": 0.8731749057769775, |
| "learning_rate": 6.52945351312988e-05, |
| "loss": 0.6979, |
| "step": 1900 |
| }, |
| { |
| "epoch": 2.69, |
| "grad_norm": 0.8528725504875183, |
| "learning_rate": 6.52235628105039e-05, |
| "loss": 0.7293, |
| "step": 1901 |
| }, |
| { |
| "epoch": 2.69, |
| "grad_norm": 0.8936817049980164, |
| "learning_rate": 6.515259048970902e-05, |
| "loss": 0.7219, |
| "step": 1902 |
| }, |
| { |
| "epoch": 2.7, |
| "grad_norm": 0.8615530133247375, |
| "learning_rate": 6.508161816891412e-05, |
| "loss": 0.7127, |
| "step": 1903 |
| }, |
| { |
| "epoch": 2.7, |
| "grad_norm": 0.9263240694999695, |
| "learning_rate": 6.501064584811924e-05, |
| "loss": 0.7366, |
| "step": 1904 |
| }, |
| { |
| "epoch": 2.7, |
| "grad_norm": 0.894145667552948, |
| "learning_rate": 6.493967352732435e-05, |
| "loss": 0.7331, |
| "step": 1905 |
| }, |
| { |
| "epoch": 2.7, |
| "grad_norm": 0.8919436931610107, |
| "learning_rate": 6.486870120652946e-05, |
| "loss": 0.7227, |
| "step": 1906 |
| }, |
| { |
| "epoch": 2.7, |
| "grad_norm": 0.9151416420936584, |
| "learning_rate": 6.479772888573456e-05, |
| "loss": 0.749, |
| "step": 1907 |
| }, |
| { |
| "epoch": 2.7, |
| "grad_norm": 0.9217795729637146, |
| "learning_rate": 6.472675656493967e-05, |
| "loss": 0.7824, |
| "step": 1908 |
| }, |
| { |
| "epoch": 2.7, |
| "grad_norm": 0.9153746366500854, |
| "learning_rate": 6.465578424414479e-05, |
| "loss": 0.7191, |
| "step": 1909 |
| }, |
| { |
| "epoch": 2.71, |
| "grad_norm": 0.8637219667434692, |
| "learning_rate": 6.45848119233499e-05, |
| "loss": 0.7216, |
| "step": 1910 |
| }, |
| { |
| "epoch": 2.71, |
| "grad_norm": 0.9067671895027161, |
| "learning_rate": 6.4513839602555e-05, |
| "loss": 0.7083, |
| "step": 1911 |
| }, |
| { |
| "epoch": 2.71, |
| "grad_norm": 0.9326914548873901, |
| "learning_rate": 6.444286728176013e-05, |
| "loss": 0.7884, |
| "step": 1912 |
| }, |
| { |
| "epoch": 2.71, |
| "grad_norm": 0.8889709711074829, |
| "learning_rate": 6.437189496096522e-05, |
| "loss": 0.7329, |
| "step": 1913 |
| }, |
| { |
| "epoch": 2.71, |
| "grad_norm": 0.8617497086524963, |
| "learning_rate": 6.430092264017034e-05, |
| "loss": 0.7025, |
| "step": 1914 |
| }, |
| { |
| "epoch": 2.71, |
| "grad_norm": 0.9106520414352417, |
| "learning_rate": 6.422995031937544e-05, |
| "loss": 0.7113, |
| "step": 1915 |
| }, |
| { |
| "epoch": 2.71, |
| "grad_norm": 0.9133102893829346, |
| "learning_rate": 6.415897799858056e-05, |
| "loss": 0.7495, |
| "step": 1916 |
| }, |
| { |
| "epoch": 2.72, |
| "grad_norm": 0.8778348565101624, |
| "learning_rate": 6.408800567778566e-05, |
| "loss": 0.691, |
| "step": 1917 |
| }, |
| { |
| "epoch": 2.72, |
| "grad_norm": 0.8651662468910217, |
| "learning_rate": 6.401703335699077e-05, |
| "loss": 0.6919, |
| "step": 1918 |
| }, |
| { |
| "epoch": 2.72, |
| "grad_norm": 0.9208177924156189, |
| "learning_rate": 6.394606103619589e-05, |
| "loss": 0.7614, |
| "step": 1919 |
| }, |
| { |
| "epoch": 2.72, |
| "grad_norm": 0.8430038094520569, |
| "learning_rate": 6.3875088715401e-05, |
| "loss": 0.6678, |
| "step": 1920 |
| }, |
| { |
| "epoch": 2.72, |
| "grad_norm": 0.8798860907554626, |
| "learning_rate": 6.380411639460611e-05, |
| "loss": 0.7001, |
| "step": 1921 |
| }, |
| { |
| "epoch": 2.72, |
| "grad_norm": 0.9283832311630249, |
| "learning_rate": 6.373314407381121e-05, |
| "loss": 0.6932, |
| "step": 1922 |
| }, |
| { |
| "epoch": 2.72, |
| "grad_norm": 0.9014415740966797, |
| "learning_rate": 6.366217175301632e-05, |
| "loss": 0.7526, |
| "step": 1923 |
| }, |
| { |
| "epoch": 2.73, |
| "grad_norm": 0.9172340035438538, |
| "learning_rate": 6.359119943222144e-05, |
| "loss": 0.7518, |
| "step": 1924 |
| }, |
| { |
| "epoch": 2.73, |
| "grad_norm": 0.8508337736129761, |
| "learning_rate": 6.352022711142655e-05, |
| "loss": 0.7273, |
| "step": 1925 |
| }, |
| { |
| "epoch": 2.73, |
| "grad_norm": 0.909666895866394, |
| "learning_rate": 6.344925479063165e-05, |
| "loss": 0.7415, |
| "step": 1926 |
| }, |
| { |
| "epoch": 2.73, |
| "grad_norm": 0.8876848220825195, |
| "learning_rate": 6.337828246983678e-05, |
| "loss": 0.7353, |
| "step": 1927 |
| }, |
| { |
| "epoch": 2.73, |
| "grad_norm": 0.9056611657142639, |
| "learning_rate": 6.330731014904187e-05, |
| "loss": 0.7776, |
| "step": 1928 |
| }, |
| { |
| "epoch": 2.73, |
| "grad_norm": 0.9231218099594116, |
| "learning_rate": 6.323633782824699e-05, |
| "loss": 0.8055, |
| "step": 1929 |
| }, |
| { |
| "epoch": 2.73, |
| "grad_norm": 0.9067128896713257, |
| "learning_rate": 6.31653655074521e-05, |
| "loss": 0.7259, |
| "step": 1930 |
| }, |
| { |
| "epoch": 2.74, |
| "grad_norm": 0.8786343336105347, |
| "learning_rate": 6.309439318665721e-05, |
| "loss": 0.7008, |
| "step": 1931 |
| }, |
| { |
| "epoch": 2.74, |
| "grad_norm": 0.9066187143325806, |
| "learning_rate": 6.302342086586231e-05, |
| "loss": 0.7225, |
| "step": 1932 |
| }, |
| { |
| "epoch": 2.74, |
| "grad_norm": 0.8969742059707642, |
| "learning_rate": 6.295244854506742e-05, |
| "loss": 0.733, |
| "step": 1933 |
| }, |
| { |
| "epoch": 2.74, |
| "grad_norm": 0.9022447466850281, |
| "learning_rate": 6.288147622427254e-05, |
| "loss": 0.7418, |
| "step": 1934 |
| }, |
| { |
| "epoch": 2.74, |
| "grad_norm": 0.9096084833145142, |
| "learning_rate": 6.281050390347765e-05, |
| "loss": 0.7125, |
| "step": 1935 |
| }, |
| { |
| "epoch": 2.74, |
| "grad_norm": 0.9495100378990173, |
| "learning_rate": 6.273953158268275e-05, |
| "loss": 0.7232, |
| "step": 1936 |
| }, |
| { |
| "epoch": 2.74, |
| "grad_norm": 0.9231953024864197, |
| "learning_rate": 6.266855926188788e-05, |
| "loss": 0.7536, |
| "step": 1937 |
| }, |
| { |
| "epoch": 2.75, |
| "grad_norm": 0.9400760531425476, |
| "learning_rate": 6.259758694109297e-05, |
| "loss": 0.7545, |
| "step": 1938 |
| }, |
| { |
| "epoch": 2.75, |
| "grad_norm": 0.899880051612854, |
| "learning_rate": 6.252661462029809e-05, |
| "loss": 0.7119, |
| "step": 1939 |
| }, |
| { |
| "epoch": 2.75, |
| "grad_norm": 0.9516352415084839, |
| "learning_rate": 6.245564229950319e-05, |
| "loss": 0.7314, |
| "step": 1940 |
| }, |
| { |
| "epoch": 2.75, |
| "grad_norm": 0.9273751378059387, |
| "learning_rate": 6.238466997870831e-05, |
| "loss": 0.7458, |
| "step": 1941 |
| }, |
| { |
| "epoch": 2.75, |
| "grad_norm": 0.8848811984062195, |
| "learning_rate": 6.231369765791341e-05, |
| "loss": 0.6936, |
| "step": 1942 |
| }, |
| { |
| "epoch": 2.75, |
| "grad_norm": 0.9142954349517822, |
| "learning_rate": 6.224272533711852e-05, |
| "loss": 0.7136, |
| "step": 1943 |
| }, |
| { |
| "epoch": 2.75, |
| "grad_norm": 0.8864548802375793, |
| "learning_rate": 6.217175301632364e-05, |
| "loss": 0.7343, |
| "step": 1944 |
| }, |
| { |
| "epoch": 2.76, |
| "grad_norm": 0.8726581931114197, |
| "learning_rate": 6.210078069552875e-05, |
| "loss": 0.7012, |
| "step": 1945 |
| }, |
| { |
| "epoch": 2.76, |
| "grad_norm": 0.8997368812561035, |
| "learning_rate": 6.202980837473385e-05, |
| "loss": 0.7389, |
| "step": 1946 |
| }, |
| { |
| "epoch": 2.76, |
| "grad_norm": 0.87708979845047, |
| "learning_rate": 6.195883605393896e-05, |
| "loss": 0.6718, |
| "step": 1947 |
| }, |
| { |
| "epoch": 2.76, |
| "grad_norm": 0.8872997164726257, |
| "learning_rate": 6.188786373314407e-05, |
| "loss": 0.7301, |
| "step": 1948 |
| }, |
| { |
| "epoch": 2.76, |
| "grad_norm": 0.9232304692268372, |
| "learning_rate": 6.181689141234919e-05, |
| "loss": 0.7711, |
| "step": 1949 |
| }, |
| { |
| "epoch": 2.76, |
| "grad_norm": 0.8790326714515686, |
| "learning_rate": 6.17459190915543e-05, |
| "loss": 0.6659, |
| "step": 1950 |
| }, |
| { |
| "epoch": 2.76, |
| "grad_norm": 0.8775014281272888, |
| "learning_rate": 6.167494677075941e-05, |
| "loss": 0.6797, |
| "step": 1951 |
| }, |
| { |
| "epoch": 2.77, |
| "grad_norm": 0.9247701168060303, |
| "learning_rate": 6.160397444996451e-05, |
| "loss": 0.7555, |
| "step": 1952 |
| }, |
| { |
| "epoch": 2.77, |
| "grad_norm": 0.9140119552612305, |
| "learning_rate": 6.153300212916962e-05, |
| "loss": 0.7253, |
| "step": 1953 |
| }, |
| { |
| "epoch": 2.77, |
| "grad_norm": 0.9283299446105957, |
| "learning_rate": 6.146202980837474e-05, |
| "loss": 0.7503, |
| "step": 1954 |
| }, |
| { |
| "epoch": 2.77, |
| "grad_norm": 0.9792236685752869, |
| "learning_rate": 6.139105748757985e-05, |
| "loss": 0.8229, |
| "step": 1955 |
| }, |
| { |
| "epoch": 2.77, |
| "grad_norm": 0.9352570176124573, |
| "learning_rate": 6.132008516678496e-05, |
| "loss": 0.7465, |
| "step": 1956 |
| }, |
| { |
| "epoch": 2.77, |
| "grad_norm": 0.9299400448799133, |
| "learning_rate": 6.124911284599006e-05, |
| "loss": 0.7143, |
| "step": 1957 |
| }, |
| { |
| "epoch": 2.77, |
| "grad_norm": 0.9018230438232422, |
| "learning_rate": 6.117814052519519e-05, |
| "loss": 0.7457, |
| "step": 1958 |
| }, |
| { |
| "epoch": 2.78, |
| "grad_norm": 0.9120910167694092, |
| "learning_rate": 6.110716820440029e-05, |
| "loss": 0.718, |
| "step": 1959 |
| }, |
| { |
| "epoch": 2.78, |
| "grad_norm": 0.9172790050506592, |
| "learning_rate": 6.10361958836054e-05, |
| "loss": 0.7895, |
| "step": 1960 |
| }, |
| { |
| "epoch": 2.78, |
| "grad_norm": 0.9235164523124695, |
| "learning_rate": 6.09652235628105e-05, |
| "loss": 0.7568, |
| "step": 1961 |
| }, |
| { |
| "epoch": 2.78, |
| "grad_norm": 0.915084183216095, |
| "learning_rate": 6.089425124201562e-05, |
| "loss": 0.7685, |
| "step": 1962 |
| }, |
| { |
| "epoch": 2.78, |
| "grad_norm": 0.8504372835159302, |
| "learning_rate": 6.0823278921220725e-05, |
| "loss": 0.7283, |
| "step": 1963 |
| }, |
| { |
| "epoch": 2.78, |
| "grad_norm": 0.8520084619522095, |
| "learning_rate": 6.075230660042584e-05, |
| "loss": 0.6907, |
| "step": 1964 |
| }, |
| { |
| "epoch": 2.78, |
| "grad_norm": 0.8992055654525757, |
| "learning_rate": 6.0681334279630944e-05, |
| "loss": 0.7292, |
| "step": 1965 |
| }, |
| { |
| "epoch": 2.79, |
| "grad_norm": 0.9049320220947266, |
| "learning_rate": 6.0610361958836056e-05, |
| "loss": 0.7567, |
| "step": 1966 |
| }, |
| { |
| "epoch": 2.79, |
| "grad_norm": 0.9057159423828125, |
| "learning_rate": 6.053938963804117e-05, |
| "loss": 0.6883, |
| "step": 1967 |
| }, |
| { |
| "epoch": 2.79, |
| "grad_norm": 0.9200868010520935, |
| "learning_rate": 6.0468417317246275e-05, |
| "loss": 0.7741, |
| "step": 1968 |
| }, |
| { |
| "epoch": 2.79, |
| "grad_norm": 0.9224002957344055, |
| "learning_rate": 6.039744499645139e-05, |
| "loss": 0.7559, |
| "step": 1969 |
| }, |
| { |
| "epoch": 2.79, |
| "grad_norm": 0.9786500930786133, |
| "learning_rate": 6.03264726756565e-05, |
| "loss": 0.7951, |
| "step": 1970 |
| }, |
| { |
| "epoch": 2.79, |
| "grad_norm": 0.9218242168426514, |
| "learning_rate": 6.0255500354861607e-05, |
| "loss": 0.741, |
| "step": 1971 |
| }, |
| { |
| "epoch": 2.79, |
| "grad_norm": 0.9101900458335876, |
| "learning_rate": 6.018452803406671e-05, |
| "loss": 0.7705, |
| "step": 1972 |
| }, |
| { |
| "epoch": 2.8, |
| "grad_norm": 0.8758111596107483, |
| "learning_rate": 6.011355571327183e-05, |
| "loss": 0.705, |
| "step": 1973 |
| }, |
| { |
| "epoch": 2.8, |
| "grad_norm": 0.9413437247276306, |
| "learning_rate": 6.004258339247694e-05, |
| "loss": 0.705, |
| "step": 1974 |
| }, |
| { |
| "epoch": 2.8, |
| "grad_norm": 0.9429991841316223, |
| "learning_rate": 5.9971611071682044e-05, |
| "loss": 0.7461, |
| "step": 1975 |
| }, |
| { |
| "epoch": 2.8, |
| "grad_norm": 0.8979993462562561, |
| "learning_rate": 5.9900638750887163e-05, |
| "loss": 0.6894, |
| "step": 1976 |
| }, |
| { |
| "epoch": 2.8, |
| "grad_norm": 0.9537708163261414, |
| "learning_rate": 5.982966643009227e-05, |
| "loss": 0.7837, |
| "step": 1977 |
| }, |
| { |
| "epoch": 2.8, |
| "grad_norm": 0.9004853367805481, |
| "learning_rate": 5.9758694109297375e-05, |
| "loss": 0.7106, |
| "step": 1978 |
| }, |
| { |
| "epoch": 2.8, |
| "grad_norm": 0.9031293988227844, |
| "learning_rate": 5.968772178850248e-05, |
| "loss": 0.7171, |
| "step": 1979 |
| }, |
| { |
| "epoch": 2.81, |
| "grad_norm": 0.8935607671737671, |
| "learning_rate": 5.96167494677076e-05, |
| "loss": 0.7401, |
| "step": 1980 |
| }, |
| { |
| "epoch": 2.81, |
| "grad_norm": 0.8977299928665161, |
| "learning_rate": 5.954577714691271e-05, |
| "loss": 0.7532, |
| "step": 1981 |
| }, |
| { |
| "epoch": 2.81, |
| "grad_norm": 0.8968639373779297, |
| "learning_rate": 5.947480482611781e-05, |
| "loss": 0.7306, |
| "step": 1982 |
| }, |
| { |
| "epoch": 2.81, |
| "grad_norm": 0.9397892951965332, |
| "learning_rate": 5.940383250532293e-05, |
| "loss": 0.8076, |
| "step": 1983 |
| }, |
| { |
| "epoch": 2.81, |
| "grad_norm": 0.9595673084259033, |
| "learning_rate": 5.933286018452804e-05, |
| "loss": 0.7778, |
| "step": 1984 |
| }, |
| { |
| "epoch": 2.81, |
| "grad_norm": 0.9051293134689331, |
| "learning_rate": 5.9261887863733144e-05, |
| "loss": 0.7372, |
| "step": 1985 |
| }, |
| { |
| "epoch": 2.81, |
| "grad_norm": 0.919461190700531, |
| "learning_rate": 5.919091554293825e-05, |
| "loss": 0.68, |
| "step": 1986 |
| }, |
| { |
| "epoch": 2.81, |
| "grad_norm": 0.947431743144989, |
| "learning_rate": 5.911994322214337e-05, |
| "loss": 0.7676, |
| "step": 1987 |
| }, |
| { |
| "epoch": 2.82, |
| "grad_norm": 0.9109405875205994, |
| "learning_rate": 5.9048970901348475e-05, |
| "loss": 0.7342, |
| "step": 1988 |
| }, |
| { |
| "epoch": 2.82, |
| "grad_norm": 0.9157077074050903, |
| "learning_rate": 5.897799858055358e-05, |
| "loss": 0.7254, |
| "step": 1989 |
| }, |
| { |
| "epoch": 2.82, |
| "grad_norm": 0.9192777276039124, |
| "learning_rate": 5.89070262597587e-05, |
| "loss": 0.816, |
| "step": 1990 |
| }, |
| { |
| "epoch": 2.82, |
| "grad_norm": 0.8862888813018799, |
| "learning_rate": 5.883605393896381e-05, |
| "loss": 0.693, |
| "step": 1991 |
| }, |
| { |
| "epoch": 2.82, |
| "grad_norm": 0.8700743317604065, |
| "learning_rate": 5.876508161816891e-05, |
| "loss": 0.7199, |
| "step": 1992 |
| }, |
| { |
| "epoch": 2.82, |
| "grad_norm": 0.861030638217926, |
| "learning_rate": 5.8694109297374026e-05, |
| "loss": 0.7043, |
| "step": 1993 |
| }, |
| { |
| "epoch": 2.82, |
| "grad_norm": 0.865680456161499, |
| "learning_rate": 5.862313697657914e-05, |
| "loss": 0.7142, |
| "step": 1994 |
| }, |
| { |
| "epoch": 2.83, |
| "grad_norm": 0.8904520869255066, |
| "learning_rate": 5.8552164655784244e-05, |
| "loss": 0.6742, |
| "step": 1995 |
| }, |
| { |
| "epoch": 2.83, |
| "grad_norm": 0.906788170337677, |
| "learning_rate": 5.848119233498936e-05, |
| "loss": 0.745, |
| "step": 1996 |
| }, |
| { |
| "epoch": 2.83, |
| "grad_norm": 0.8844560980796814, |
| "learning_rate": 5.841022001419447e-05, |
| "loss": 0.6675, |
| "step": 1997 |
| }, |
| { |
| "epoch": 2.83, |
| "grad_norm": 0.9228169918060303, |
| "learning_rate": 5.833924769339958e-05, |
| "loss": 0.7112, |
| "step": 1998 |
| }, |
| { |
| "epoch": 2.83, |
| "grad_norm": 0.9667858481407166, |
| "learning_rate": 5.826827537260469e-05, |
| "loss": 0.76, |
| "step": 1999 |
| }, |
| { |
| "epoch": 2.83, |
| "grad_norm": 0.9407635927200317, |
| "learning_rate": 5.8197303051809794e-05, |
| "loss": 0.7387, |
| "step": 2000 |
| }, |
| { |
| "epoch": 2.83, |
| "grad_norm": 0.9030634164810181, |
| "learning_rate": 5.8126330731014914e-05, |
| "loss": 0.6959, |
| "step": 2001 |
| }, |
| { |
| "epoch": 2.84, |
| "grad_norm": 0.9199838638305664, |
| "learning_rate": 5.805535841022002e-05, |
| "loss": 0.7235, |
| "step": 2002 |
| }, |
| { |
| "epoch": 2.84, |
| "grad_norm": 0.9523055553436279, |
| "learning_rate": 5.7984386089425126e-05, |
| "loss": 0.7482, |
| "step": 2003 |
| }, |
| { |
| "epoch": 2.84, |
| "grad_norm": 0.8901258707046509, |
| "learning_rate": 5.7913413768630245e-05, |
| "loss": 0.7052, |
| "step": 2004 |
| }, |
| { |
| "epoch": 2.84, |
| "grad_norm": 0.919878363609314, |
| "learning_rate": 5.784244144783535e-05, |
| "loss": 0.7624, |
| "step": 2005 |
| }, |
| { |
| "epoch": 2.84, |
| "grad_norm": 0.9046053886413574, |
| "learning_rate": 5.777146912704046e-05, |
| "loss": 0.6738, |
| "step": 2006 |
| }, |
| { |
| "epoch": 2.84, |
| "grad_norm": 0.9026862978935242, |
| "learning_rate": 5.770049680624556e-05, |
| "loss": 0.7391, |
| "step": 2007 |
| }, |
| { |
| "epoch": 2.84, |
| "grad_norm": 0.8612751960754395, |
| "learning_rate": 5.762952448545068e-05, |
| "loss": 0.6809, |
| "step": 2008 |
| }, |
| { |
| "epoch": 2.85, |
| "grad_norm": 0.909647524356842, |
| "learning_rate": 5.755855216465579e-05, |
| "loss": 0.7777, |
| "step": 2009 |
| }, |
| { |
| "epoch": 2.85, |
| "grad_norm": 0.9045459032058716, |
| "learning_rate": 5.7487579843860894e-05, |
| "loss": 0.7579, |
| "step": 2010 |
| }, |
| { |
| "epoch": 2.85, |
| "grad_norm": 0.8565319180488586, |
| "learning_rate": 5.7416607523066e-05, |
| "loss": 0.6686, |
| "step": 2011 |
| }, |
| { |
| "epoch": 2.85, |
| "grad_norm": 0.8806788325309753, |
| "learning_rate": 5.734563520227112e-05, |
| "loss": 0.7081, |
| "step": 2012 |
| }, |
| { |
| "epoch": 2.85, |
| "grad_norm": 0.9146614074707031, |
| "learning_rate": 5.7274662881476226e-05, |
| "loss": 0.7461, |
| "step": 2013 |
| }, |
| { |
| "epoch": 2.85, |
| "grad_norm": 0.9521096348762512, |
| "learning_rate": 5.720369056068133e-05, |
| "loss": 0.7745, |
| "step": 2014 |
| }, |
| { |
| "epoch": 2.85, |
| "grad_norm": 0.9208850264549255, |
| "learning_rate": 5.713271823988645e-05, |
| "loss": 0.7343, |
| "step": 2015 |
| }, |
| { |
| "epoch": 2.86, |
| "grad_norm": 0.9413381814956665, |
| "learning_rate": 5.706174591909156e-05, |
| "loss": 0.744, |
| "step": 2016 |
| }, |
| { |
| "epoch": 2.86, |
| "grad_norm": 0.9284963011741638, |
| "learning_rate": 5.699077359829666e-05, |
| "loss": 0.7354, |
| "step": 2017 |
| }, |
| { |
| "epoch": 2.86, |
| "grad_norm": 0.9209182262420654, |
| "learning_rate": 5.691980127750177e-05, |
| "loss": 0.7992, |
| "step": 2018 |
| }, |
| { |
| "epoch": 2.86, |
| "grad_norm": 0.8778393268585205, |
| "learning_rate": 5.684882895670689e-05, |
| "loss": 0.7211, |
| "step": 2019 |
| }, |
| { |
| "epoch": 2.86, |
| "grad_norm": 0.9171314239501953, |
| "learning_rate": 5.6777856635911995e-05, |
| "loss": 0.7308, |
| "step": 2020 |
| }, |
| { |
| "epoch": 2.86, |
| "grad_norm": 0.9178740382194519, |
| "learning_rate": 5.67068843151171e-05, |
| "loss": 0.7791, |
| "step": 2021 |
| }, |
| { |
| "epoch": 2.86, |
| "grad_norm": 0.9107670187950134, |
| "learning_rate": 5.663591199432222e-05, |
| "loss": 0.7673, |
| "step": 2022 |
| }, |
| { |
| "epoch": 2.87, |
| "grad_norm": 0.8879327178001404, |
| "learning_rate": 5.6564939673527326e-05, |
| "loss": 0.7244, |
| "step": 2023 |
| }, |
| { |
| "epoch": 2.87, |
| "grad_norm": 0.8985375165939331, |
| "learning_rate": 5.649396735273244e-05, |
| "loss": 0.735, |
| "step": 2024 |
| }, |
| { |
| "epoch": 2.87, |
| "grad_norm": 0.8995974659919739, |
| "learning_rate": 5.6422995031937545e-05, |
| "loss": 0.7454, |
| "step": 2025 |
| }, |
| { |
| "epoch": 2.87, |
| "grad_norm": 0.9525470733642578, |
| "learning_rate": 5.635202271114266e-05, |
| "loss": 0.7669, |
| "step": 2026 |
| }, |
| { |
| "epoch": 2.87, |
| "grad_norm": 0.9079092144966125, |
| "learning_rate": 5.628105039034777e-05, |
| "loss": 0.7018, |
| "step": 2027 |
| }, |
| { |
| "epoch": 2.87, |
| "grad_norm": 0.9271257519721985, |
| "learning_rate": 5.6210078069552876e-05, |
| "loss": 0.7295, |
| "step": 2028 |
| }, |
| { |
| "epoch": 2.87, |
| "grad_norm": 0.943726122379303, |
| "learning_rate": 5.613910574875799e-05, |
| "loss": 0.7472, |
| "step": 2029 |
| }, |
| { |
| "epoch": 2.88, |
| "grad_norm": 0.9313754439353943, |
| "learning_rate": 5.60681334279631e-05, |
| "loss": 0.7358, |
| "step": 2030 |
| }, |
| { |
| "epoch": 2.88, |
| "grad_norm": 0.9185075163841248, |
| "learning_rate": 5.599716110716821e-05, |
| "loss": 0.7313, |
| "step": 2031 |
| }, |
| { |
| "epoch": 2.88, |
| "grad_norm": 0.9241396188735962, |
| "learning_rate": 5.5926188786373313e-05, |
| "loss": 0.7114, |
| "step": 2032 |
| }, |
| { |
| "epoch": 2.88, |
| "grad_norm": 0.884719967842102, |
| "learning_rate": 5.585521646557843e-05, |
| "loss": 0.727, |
| "step": 2033 |
| }, |
| { |
| "epoch": 2.88, |
| "grad_norm": 0.8894933462142944, |
| "learning_rate": 5.578424414478354e-05, |
| "loss": 0.7598, |
| "step": 2034 |
| }, |
| { |
| "epoch": 2.88, |
| "grad_norm": 0.8764230608940125, |
| "learning_rate": 5.5713271823988645e-05, |
| "loss": 0.7433, |
| "step": 2035 |
| }, |
| { |
| "epoch": 2.88, |
| "grad_norm": 0.8621907234191895, |
| "learning_rate": 5.5642299503193764e-05, |
| "loss": 0.7148, |
| "step": 2036 |
| }, |
| { |
| "epoch": 2.89, |
| "grad_norm": 0.8671466112136841, |
| "learning_rate": 5.557132718239887e-05, |
| "loss": 0.7362, |
| "step": 2037 |
| }, |
| { |
| "epoch": 2.89, |
| "grad_norm": 0.875935435295105, |
| "learning_rate": 5.5500354861603976e-05, |
| "loss": 0.7166, |
| "step": 2038 |
| }, |
| { |
| "epoch": 2.89, |
| "grad_norm": 0.9075550436973572, |
| "learning_rate": 5.542938254080908e-05, |
| "loss": 0.7744, |
| "step": 2039 |
| }, |
| { |
| "epoch": 2.89, |
| "grad_norm": 0.9083443880081177, |
| "learning_rate": 5.53584102200142e-05, |
| "loss": 0.709, |
| "step": 2040 |
| }, |
| { |
| "epoch": 2.89, |
| "grad_norm": 0.8900015950202942, |
| "learning_rate": 5.528743789921931e-05, |
| "loss": 0.7487, |
| "step": 2041 |
| }, |
| { |
| "epoch": 2.89, |
| "grad_norm": 0.9004586338996887, |
| "learning_rate": 5.5216465578424414e-05, |
| "loss": 0.7373, |
| "step": 2042 |
| }, |
| { |
| "epoch": 2.89, |
| "grad_norm": 0.9120739698410034, |
| "learning_rate": 5.514549325762953e-05, |
| "loss": 0.7307, |
| "step": 2043 |
| }, |
| { |
| "epoch": 2.9, |
| "grad_norm": 0.9269658327102661, |
| "learning_rate": 5.507452093683464e-05, |
| "loss": 0.7593, |
| "step": 2044 |
| }, |
| { |
| "epoch": 2.9, |
| "grad_norm": 0.9356985688209534, |
| "learning_rate": 5.5003548616039745e-05, |
| "loss": 0.7623, |
| "step": 2045 |
| }, |
| { |
| "epoch": 2.9, |
| "grad_norm": 0.9285359382629395, |
| "learning_rate": 5.493257629524485e-05, |
| "loss": 0.7215, |
| "step": 2046 |
| }, |
| { |
| "epoch": 2.9, |
| "grad_norm": 0.892284095287323, |
| "learning_rate": 5.486160397444997e-05, |
| "loss": 0.7151, |
| "step": 2047 |
| }, |
| { |
| "epoch": 2.9, |
| "grad_norm": 0.9217097759246826, |
| "learning_rate": 5.4790631653655076e-05, |
| "loss": 0.7354, |
| "step": 2048 |
| }, |
| { |
| "epoch": 2.9, |
| "grad_norm": 0.9596689939498901, |
| "learning_rate": 5.471965933286018e-05, |
| "loss": 0.7735, |
| "step": 2049 |
| }, |
| { |
| "epoch": 2.9, |
| "grad_norm": 0.9006109237670898, |
| "learning_rate": 5.464868701206529e-05, |
| "loss": 0.6733, |
| "step": 2050 |
| }, |
| { |
| "epoch": 2.91, |
| "grad_norm": 0.9174147248268127, |
| "learning_rate": 5.457771469127041e-05, |
| "loss": 0.7423, |
| "step": 2051 |
| }, |
| { |
| "epoch": 2.91, |
| "grad_norm": 0.944945752620697, |
| "learning_rate": 5.4506742370475514e-05, |
| "loss": 0.7928, |
| "step": 2052 |
| }, |
| { |
| "epoch": 2.91, |
| "grad_norm": 0.8980944156646729, |
| "learning_rate": 5.4435770049680626e-05, |
| "loss": 0.6827, |
| "step": 2053 |
| }, |
| { |
| "epoch": 2.91, |
| "grad_norm": 0.9077931642532349, |
| "learning_rate": 5.436479772888574e-05, |
| "loss": 0.7515, |
| "step": 2054 |
| }, |
| { |
| "epoch": 2.91, |
| "grad_norm": 0.8934177756309509, |
| "learning_rate": 5.4293825408090845e-05, |
| "loss": 0.7314, |
| "step": 2055 |
| }, |
| { |
| "epoch": 2.91, |
| "grad_norm": 0.8950974345207214, |
| "learning_rate": 5.422285308729596e-05, |
| "loss": 0.7644, |
| "step": 2056 |
| }, |
| { |
| "epoch": 2.91, |
| "grad_norm": 0.9017379283905029, |
| "learning_rate": 5.4151880766501064e-05, |
| "loss": 0.7599, |
| "step": 2057 |
| }, |
| { |
| "epoch": 2.92, |
| "grad_norm": 0.8867719173431396, |
| "learning_rate": 5.4080908445706177e-05, |
| "loss": 0.7247, |
| "step": 2058 |
| }, |
| { |
| "epoch": 2.92, |
| "grad_norm": 0.8819170594215393, |
| "learning_rate": 5.400993612491129e-05, |
| "loss": 0.7039, |
| "step": 2059 |
| }, |
| { |
| "epoch": 2.92, |
| "grad_norm": 0.9082474708557129, |
| "learning_rate": 5.3938963804116395e-05, |
| "loss": 0.7261, |
| "step": 2060 |
| }, |
| { |
| "epoch": 2.92, |
| "grad_norm": 0.8912190198898315, |
| "learning_rate": 5.3867991483321515e-05, |
| "loss": 0.758, |
| "step": 2061 |
| }, |
| { |
| "epoch": 2.92, |
| "grad_norm": 1.0254679918289185, |
| "learning_rate": 5.379701916252662e-05, |
| "loss": 0.7901, |
| "step": 2062 |
| }, |
| { |
| "epoch": 2.92, |
| "grad_norm": 0.8879702687263489, |
| "learning_rate": 5.3726046841731727e-05, |
| "loss": 0.7172, |
| "step": 2063 |
| }, |
| { |
| "epoch": 2.92, |
| "grad_norm": 0.8472481966018677, |
| "learning_rate": 5.365507452093683e-05, |
| "loss": 0.7248, |
| "step": 2064 |
| }, |
| { |
| "epoch": 2.93, |
| "grad_norm": 0.9291859865188599, |
| "learning_rate": 5.358410220014195e-05, |
| "loss": 0.7221, |
| "step": 2065 |
| }, |
| { |
| "epoch": 2.93, |
| "grad_norm": 0.8945410251617432, |
| "learning_rate": 5.351312987934706e-05, |
| "loss": 0.6759, |
| "step": 2066 |
| }, |
| { |
| "epoch": 2.93, |
| "grad_norm": 0.8901339769363403, |
| "learning_rate": 5.3442157558552164e-05, |
| "loss": 0.7342, |
| "step": 2067 |
| }, |
| { |
| "epoch": 2.93, |
| "grad_norm": 0.8966627717018127, |
| "learning_rate": 5.3371185237757283e-05, |
| "loss": 0.6834, |
| "step": 2068 |
| }, |
| { |
| "epoch": 2.93, |
| "grad_norm": 0.9248061180114746, |
| "learning_rate": 5.330021291696239e-05, |
| "loss": 0.7561, |
| "step": 2069 |
| }, |
| { |
| "epoch": 2.93, |
| "grad_norm": 0.9252817630767822, |
| "learning_rate": 5.3229240596167495e-05, |
| "loss": 0.7436, |
| "step": 2070 |
| }, |
| { |
| "epoch": 2.93, |
| "grad_norm": 0.888176441192627, |
| "learning_rate": 5.31582682753726e-05, |
| "loss": 0.7017, |
| "step": 2071 |
| }, |
| { |
| "epoch": 2.94, |
| "grad_norm": 0.894523561000824, |
| "learning_rate": 5.308729595457772e-05, |
| "loss": 0.7224, |
| "step": 2072 |
| }, |
| { |
| "epoch": 2.94, |
| "grad_norm": 0.9010726809501648, |
| "learning_rate": 5.301632363378283e-05, |
| "loss": 0.7276, |
| "step": 2073 |
| }, |
| { |
| "epoch": 2.94, |
| "grad_norm": 0.9328283071517944, |
| "learning_rate": 5.294535131298793e-05, |
| "loss": 0.7599, |
| "step": 2074 |
| }, |
| { |
| "epoch": 2.94, |
| "grad_norm": 0.8821666836738586, |
| "learning_rate": 5.287437899219305e-05, |
| "loss": 0.7245, |
| "step": 2075 |
| }, |
| { |
| "epoch": 2.94, |
| "grad_norm": 0.9441255927085876, |
| "learning_rate": 5.280340667139816e-05, |
| "loss": 0.7344, |
| "step": 2076 |
| }, |
| { |
| "epoch": 2.94, |
| "grad_norm": 0.9148538708686829, |
| "learning_rate": 5.2732434350603264e-05, |
| "loss": 0.7172, |
| "step": 2077 |
| }, |
| { |
| "epoch": 2.94, |
| "grad_norm": 0.895046055316925, |
| "learning_rate": 5.266146202980837e-05, |
| "loss": 0.7425, |
| "step": 2078 |
| }, |
| { |
| "epoch": 2.95, |
| "grad_norm": 0.8581205606460571, |
| "learning_rate": 5.259048970901349e-05, |
| "loss": 0.703, |
| "step": 2079 |
| }, |
| { |
| "epoch": 2.95, |
| "grad_norm": 0.8967068195343018, |
| "learning_rate": 5.2519517388218595e-05, |
| "loss": 0.7363, |
| "step": 2080 |
| }, |
| { |
| "epoch": 2.95, |
| "grad_norm": 0.8934385180473328, |
| "learning_rate": 5.24485450674237e-05, |
| "loss": 0.7344, |
| "step": 2081 |
| }, |
| { |
| "epoch": 2.95, |
| "grad_norm": 0.9267292618751526, |
| "learning_rate": 5.237757274662882e-05, |
| "loss": 0.7429, |
| "step": 2082 |
| }, |
| { |
| "epoch": 2.95, |
| "grad_norm": 0.8986665606498718, |
| "learning_rate": 5.230660042583393e-05, |
| "loss": 0.7341, |
| "step": 2083 |
| }, |
| { |
| "epoch": 2.95, |
| "grad_norm": 0.9249792098999023, |
| "learning_rate": 5.223562810503903e-05, |
| "loss": 0.723, |
| "step": 2084 |
| }, |
| { |
| "epoch": 2.95, |
| "grad_norm": 0.9367862343788147, |
| "learning_rate": 5.2164655784244146e-05, |
| "loss": 0.7556, |
| "step": 2085 |
| }, |
| { |
| "epoch": 2.96, |
| "grad_norm": 0.9350747466087341, |
| "learning_rate": 5.209368346344926e-05, |
| "loss": 0.7687, |
| "step": 2086 |
| }, |
| { |
| "epoch": 2.96, |
| "grad_norm": 0.9194963574409485, |
| "learning_rate": 5.202271114265437e-05, |
| "loss": 0.7667, |
| "step": 2087 |
| }, |
| { |
| "epoch": 2.96, |
| "grad_norm": 0.910021185874939, |
| "learning_rate": 5.195173882185948e-05, |
| "loss": 0.7181, |
| "step": 2088 |
| }, |
| { |
| "epoch": 2.96, |
| "grad_norm": 0.9334052205085754, |
| "learning_rate": 5.188076650106459e-05, |
| "loss": 0.7773, |
| "step": 2089 |
| }, |
| { |
| "epoch": 2.96, |
| "grad_norm": 0.9095138311386108, |
| "learning_rate": 5.18097941802697e-05, |
| "loss": 0.7153, |
| "step": 2090 |
| }, |
| { |
| "epoch": 2.96, |
| "grad_norm": 0.8687496781349182, |
| "learning_rate": 5.173882185947481e-05, |
| "loss": 0.7251, |
| "step": 2091 |
| }, |
| { |
| "epoch": 2.96, |
| "grad_norm": 0.889177143573761, |
| "learning_rate": 5.1667849538679914e-05, |
| "loss": 0.7084, |
| "step": 2092 |
| }, |
| { |
| "epoch": 2.97, |
| "grad_norm": 0.8826084733009338, |
| "learning_rate": 5.1596877217885034e-05, |
| "loss": 0.6988, |
| "step": 2093 |
| }, |
| { |
| "epoch": 2.97, |
| "grad_norm": 0.891457736492157, |
| "learning_rate": 5.152590489709014e-05, |
| "loss": 0.7394, |
| "step": 2094 |
| }, |
| { |
| "epoch": 2.97, |
| "grad_norm": 0.8934881091117859, |
| "learning_rate": 5.1454932576295246e-05, |
| "loss": 0.6861, |
| "step": 2095 |
| }, |
| { |
| "epoch": 2.97, |
| "grad_norm": 0.9134221076965332, |
| "learning_rate": 5.138396025550035e-05, |
| "loss": 0.7227, |
| "step": 2096 |
| }, |
| { |
| "epoch": 2.97, |
| "grad_norm": 0.8815884590148926, |
| "learning_rate": 5.131298793470547e-05, |
| "loss": 0.675, |
| "step": 2097 |
| }, |
| { |
| "epoch": 2.97, |
| "grad_norm": 0.9253513813018799, |
| "learning_rate": 5.124201561391058e-05, |
| "loss": 0.723, |
| "step": 2098 |
| }, |
| { |
| "epoch": 2.97, |
| "grad_norm": 0.9489588141441345, |
| "learning_rate": 5.117104329311568e-05, |
| "loss": 0.7302, |
| "step": 2099 |
| }, |
| { |
| "epoch": 2.98, |
| "grad_norm": 0.9437028169631958, |
| "learning_rate": 5.11000709723208e-05, |
| "loss": 0.7355, |
| "step": 2100 |
| }, |
| { |
| "epoch": 2.98, |
| "grad_norm": 0.9252711534500122, |
| "learning_rate": 5.102909865152591e-05, |
| "loss": 0.7317, |
| "step": 2101 |
| }, |
| { |
| "epoch": 2.98, |
| "grad_norm": 0.9146744608879089, |
| "learning_rate": 5.0958126330731014e-05, |
| "loss": 0.7059, |
| "step": 2102 |
| }, |
| { |
| "epoch": 2.98, |
| "grad_norm": 0.9322413206100464, |
| "learning_rate": 5.088715400993612e-05, |
| "loss": 0.7326, |
| "step": 2103 |
| }, |
| { |
| "epoch": 2.98, |
| "grad_norm": 0.9023984670639038, |
| "learning_rate": 5.081618168914124e-05, |
| "loss": 0.7999, |
| "step": 2104 |
| }, |
| { |
| "epoch": 2.98, |
| "grad_norm": 0.8638840913772583, |
| "learning_rate": 5.0745209368346346e-05, |
| "loss": 0.7179, |
| "step": 2105 |
| }, |
| { |
| "epoch": 2.98, |
| "grad_norm": 0.8496398329734802, |
| "learning_rate": 5.067423704755145e-05, |
| "loss": 0.6619, |
| "step": 2106 |
| }, |
| { |
| "epoch": 2.98, |
| "grad_norm": 0.9351739287376404, |
| "learning_rate": 5.060326472675657e-05, |
| "loss": 0.7478, |
| "step": 2107 |
| }, |
| { |
| "epoch": 2.99, |
| "grad_norm": 0.9056074619293213, |
| "learning_rate": 5.053229240596168e-05, |
| "loss": 0.7747, |
| "step": 2108 |
| }, |
| { |
| "epoch": 2.99, |
| "grad_norm": 0.8992380499839783, |
| "learning_rate": 5.046132008516678e-05, |
| "loss": 0.733, |
| "step": 2109 |
| }, |
| { |
| "epoch": 2.99, |
| "grad_norm": 0.9041831493377686, |
| "learning_rate": 5.039034776437189e-05, |
| "loss": 0.7573, |
| "step": 2110 |
| }, |
| { |
| "epoch": 2.99, |
| "grad_norm": 0.9171251058578491, |
| "learning_rate": 5.031937544357701e-05, |
| "loss": 0.7364, |
| "step": 2111 |
| }, |
| { |
| "epoch": 2.99, |
| "grad_norm": 0.9151423573493958, |
| "learning_rate": 5.0248403122782115e-05, |
| "loss": 0.7923, |
| "step": 2112 |
| }, |
| { |
| "epoch": 2.99, |
| "grad_norm": 0.8989052176475525, |
| "learning_rate": 5.017743080198723e-05, |
| "loss": 0.7144, |
| "step": 2113 |
| }, |
| { |
| "epoch": 2.99, |
| "grad_norm": 0.9141291379928589, |
| "learning_rate": 5.010645848119234e-05, |
| "loss": 0.6852, |
| "step": 2114 |
| }, |
| { |
| "epoch": 3.0, |
| "grad_norm": 0.9194472432136536, |
| "learning_rate": 5.0035486160397446e-05, |
| "loss": 0.7559, |
| "step": 2115 |
| }, |
| { |
| "epoch": 3.0, |
| "grad_norm": 0.867283284664154, |
| "learning_rate": 4.996451383960256e-05, |
| "loss": 0.7188, |
| "step": 2116 |
| }, |
| { |
| "epoch": 3.0, |
| "grad_norm": 0.9092669486999512, |
| "learning_rate": 4.989354151880767e-05, |
| "loss": 0.7543, |
| "step": 2117 |
| }, |
| { |
| "epoch": 3.0, |
| "grad_norm": 0.8685262203216553, |
| "learning_rate": 4.982256919801278e-05, |
| "loss": 0.6828, |
| "step": 2118 |
| }, |
| { |
| "epoch": 3.0, |
| "grad_norm": 0.7743335962295532, |
| "learning_rate": 4.975159687721789e-05, |
| "loss": 0.5045, |
| "step": 2119 |
| }, |
| { |
| "epoch": 3.0, |
| "grad_norm": 0.8006496429443359, |
| "learning_rate": 4.9680624556423e-05, |
| "loss": 0.5517, |
| "step": 2120 |
| }, |
| { |
| "epoch": 3.0, |
| "grad_norm": 0.8102491497993469, |
| "learning_rate": 4.960965223562811e-05, |
| "loss": 0.504, |
| "step": 2121 |
| }, |
| { |
| "epoch": 3.01, |
| "grad_norm": 0.7949885129928589, |
| "learning_rate": 4.953867991483322e-05, |
| "loss": 0.4944, |
| "step": 2122 |
| }, |
| { |
| "epoch": 3.01, |
| "grad_norm": 0.8948567509651184, |
| "learning_rate": 4.946770759403833e-05, |
| "loss": 0.4933, |
| "step": 2123 |
| }, |
| { |
| "epoch": 3.01, |
| "grad_norm": 0.9919142723083496, |
| "learning_rate": 4.939673527324344e-05, |
| "loss": 0.558, |
| "step": 2124 |
| }, |
| { |
| "epoch": 3.01, |
| "grad_norm": 1.0344558954238892, |
| "learning_rate": 4.9325762952448546e-05, |
| "loss": 0.5067, |
| "step": 2125 |
| }, |
| { |
| "epoch": 3.01, |
| "grad_norm": 1.1019103527069092, |
| "learning_rate": 4.925479063165366e-05, |
| "loss": 0.5355, |
| "step": 2126 |
| }, |
| { |
| "epoch": 3.01, |
| "grad_norm": 1.170531153678894, |
| "learning_rate": 4.9183818310858765e-05, |
| "loss": 0.5802, |
| "step": 2127 |
| }, |
| { |
| "epoch": 3.01, |
| "grad_norm": 1.1433138847351074, |
| "learning_rate": 4.911284599006388e-05, |
| "loss": 0.5597, |
| "step": 2128 |
| }, |
| { |
| "epoch": 3.02, |
| "grad_norm": 1.1154450178146362, |
| "learning_rate": 4.904187366926899e-05, |
| "loss": 0.542, |
| "step": 2129 |
| }, |
| { |
| "epoch": 3.02, |
| "grad_norm": 1.0891841650009155, |
| "learning_rate": 4.8970901348474096e-05, |
| "loss": 0.559, |
| "step": 2130 |
| }, |
| { |
| "epoch": 3.02, |
| "grad_norm": 0.9962801337242126, |
| "learning_rate": 4.889992902767921e-05, |
| "loss": 0.5171, |
| "step": 2131 |
| }, |
| { |
| "epoch": 3.02, |
| "grad_norm": 0.9601651430130005, |
| "learning_rate": 4.8828956706884315e-05, |
| "loss": 0.5508, |
| "step": 2132 |
| }, |
| { |
| "epoch": 3.02, |
| "grad_norm": 0.8857892751693726, |
| "learning_rate": 4.875798438608943e-05, |
| "loss": 0.4933, |
| "step": 2133 |
| }, |
| { |
| "epoch": 3.02, |
| "grad_norm": 0.8972840905189514, |
| "learning_rate": 4.8687012065294534e-05, |
| "loss": 0.5005, |
| "step": 2134 |
| }, |
| { |
| "epoch": 3.02, |
| "grad_norm": 0.922217071056366, |
| "learning_rate": 4.8616039744499646e-05, |
| "loss": 0.5266, |
| "step": 2135 |
| }, |
| { |
| "epoch": 3.03, |
| "grad_norm": 0.8686092495918274, |
| "learning_rate": 4.854506742370476e-05, |
| "loss": 0.5273, |
| "step": 2136 |
| }, |
| { |
| "epoch": 3.03, |
| "grad_norm": 0.901107132434845, |
| "learning_rate": 4.8474095102909865e-05, |
| "loss": 0.49, |
| "step": 2137 |
| }, |
| { |
| "epoch": 3.03, |
| "grad_norm": 0.9138848185539246, |
| "learning_rate": 4.840312278211498e-05, |
| "loss": 0.5071, |
| "step": 2138 |
| }, |
| { |
| "epoch": 3.03, |
| "grad_norm": 0.9287373423576355, |
| "learning_rate": 4.8332150461320084e-05, |
| "loss": 0.5132, |
| "step": 2139 |
| }, |
| { |
| "epoch": 3.03, |
| "grad_norm": 0.9281260371208191, |
| "learning_rate": 4.8261178140525196e-05, |
| "loss": 0.5068, |
| "step": 2140 |
| }, |
| { |
| "epoch": 3.03, |
| "grad_norm": 0.9996431469917297, |
| "learning_rate": 4.81902058197303e-05, |
| "loss": 0.497, |
| "step": 2141 |
| }, |
| { |
| "epoch": 3.03, |
| "grad_norm": 1.0466423034667969, |
| "learning_rate": 4.8119233498935415e-05, |
| "loss": 0.5328, |
| "step": 2142 |
| }, |
| { |
| "epoch": 3.04, |
| "grad_norm": 1.0136125087738037, |
| "learning_rate": 4.804826117814053e-05, |
| "loss": 0.4828, |
| "step": 2143 |
| }, |
| { |
| "epoch": 3.04, |
| "grad_norm": 1.0378624200820923, |
| "learning_rate": 4.7977288857345634e-05, |
| "loss": 0.5387, |
| "step": 2144 |
| }, |
| { |
| "epoch": 3.04, |
| "grad_norm": 1.0223605632781982, |
| "learning_rate": 4.7906316536550746e-05, |
| "loss": 0.5476, |
| "step": 2145 |
| }, |
| { |
| "epoch": 3.04, |
| "grad_norm": 1.0363613367080688, |
| "learning_rate": 4.783534421575586e-05, |
| "loss": 0.5516, |
| "step": 2146 |
| }, |
| { |
| "epoch": 3.04, |
| "grad_norm": 0.9881285429000854, |
| "learning_rate": 4.776437189496097e-05, |
| "loss": 0.5342, |
| "step": 2147 |
| }, |
| { |
| "epoch": 3.04, |
| "grad_norm": 0.989923894405365, |
| "learning_rate": 4.769339957416608e-05, |
| "loss": 0.5694, |
| "step": 2148 |
| }, |
| { |
| "epoch": 3.04, |
| "grad_norm": 0.9041772484779358, |
| "learning_rate": 4.762242725337119e-05, |
| "loss": 0.4411, |
| "step": 2149 |
| }, |
| { |
| "epoch": 3.05, |
| "grad_norm": 0.9580135941505432, |
| "learning_rate": 4.7551454932576297e-05, |
| "loss": 0.5294, |
| "step": 2150 |
| }, |
| { |
| "epoch": 3.05, |
| "grad_norm": 0.9735754132270813, |
| "learning_rate": 4.748048261178141e-05, |
| "loss": 0.5518, |
| "step": 2151 |
| }, |
| { |
| "epoch": 3.05, |
| "grad_norm": 0.9906786680221558, |
| "learning_rate": 4.740951029098652e-05, |
| "loss": 0.5644, |
| "step": 2152 |
| }, |
| { |
| "epoch": 3.05, |
| "grad_norm": 0.9299677610397339, |
| "learning_rate": 4.733853797019163e-05, |
| "loss": 0.536, |
| "step": 2153 |
| }, |
| { |
| "epoch": 3.05, |
| "grad_norm": 0.9068963527679443, |
| "learning_rate": 4.726756564939674e-05, |
| "loss": 0.4836, |
| "step": 2154 |
| }, |
| { |
| "epoch": 3.05, |
| "grad_norm": 0.9534969925880432, |
| "learning_rate": 4.719659332860185e-05, |
| "loss": 0.5291, |
| "step": 2155 |
| }, |
| { |
| "epoch": 3.05, |
| "grad_norm": 1.0046902894973755, |
| "learning_rate": 4.712562100780696e-05, |
| "loss": 0.5388, |
| "step": 2156 |
| }, |
| { |
| "epoch": 3.06, |
| "grad_norm": 0.9631446003913879, |
| "learning_rate": 4.7054648687012065e-05, |
| "loss": 0.5067, |
| "step": 2157 |
| }, |
| { |
| "epoch": 3.06, |
| "grad_norm": 1.0535892248153687, |
| "learning_rate": 4.698367636621718e-05, |
| "loss": 0.5609, |
| "step": 2158 |
| }, |
| { |
| "epoch": 3.06, |
| "grad_norm": 0.9783676266670227, |
| "learning_rate": 4.691270404542229e-05, |
| "loss": 0.5228, |
| "step": 2159 |
| }, |
| { |
| "epoch": 3.06, |
| "grad_norm": 0.9673240184783936, |
| "learning_rate": 4.68417317246274e-05, |
| "loss": 0.5479, |
| "step": 2160 |
| }, |
| { |
| "epoch": 3.06, |
| "grad_norm": 0.9668963551521301, |
| "learning_rate": 4.677075940383251e-05, |
| "loss": 0.477, |
| "step": 2161 |
| }, |
| { |
| "epoch": 3.06, |
| "grad_norm": 0.971133291721344, |
| "learning_rate": 4.6699787083037615e-05, |
| "loss": 0.4775, |
| "step": 2162 |
| }, |
| { |
| "epoch": 3.06, |
| "grad_norm": 0.9844760298728943, |
| "learning_rate": 4.662881476224273e-05, |
| "loss": 0.5123, |
| "step": 2163 |
| }, |
| { |
| "epoch": 3.07, |
| "grad_norm": 0.9454901814460754, |
| "learning_rate": 4.6557842441447834e-05, |
| "loss": 0.5169, |
| "step": 2164 |
| }, |
| { |
| "epoch": 3.07, |
| "grad_norm": 0.9617669582366943, |
| "learning_rate": 4.648687012065295e-05, |
| "loss": 0.5141, |
| "step": 2165 |
| }, |
| { |
| "epoch": 3.07, |
| "grad_norm": 0.9806207418441772, |
| "learning_rate": 4.641589779985806e-05, |
| "loss": 0.4989, |
| "step": 2166 |
| }, |
| { |
| "epoch": 3.07, |
| "grad_norm": 0.929340124130249, |
| "learning_rate": 4.6344925479063165e-05, |
| "loss": 0.5034, |
| "step": 2167 |
| }, |
| { |
| "epoch": 3.07, |
| "grad_norm": 0.9709641933441162, |
| "learning_rate": 4.627395315826828e-05, |
| "loss": 0.4999, |
| "step": 2168 |
| }, |
| { |
| "epoch": 3.07, |
| "grad_norm": 0.9410818219184875, |
| "learning_rate": 4.6202980837473384e-05, |
| "loss": 0.4847, |
| "step": 2169 |
| }, |
| { |
| "epoch": 3.07, |
| "grad_norm": 0.9140300154685974, |
| "learning_rate": 4.61320085166785e-05, |
| "loss": 0.4953, |
| "step": 2170 |
| }, |
| { |
| "epoch": 3.08, |
| "grad_norm": 0.9406004548072815, |
| "learning_rate": 4.60610361958836e-05, |
| "loss": 0.5043, |
| "step": 2171 |
| }, |
| { |
| "epoch": 3.08, |
| "grad_norm": 0.9244046211242676, |
| "learning_rate": 4.5990063875088716e-05, |
| "loss": 0.4742, |
| "step": 2172 |
| }, |
| { |
| "epoch": 3.08, |
| "grad_norm": 0.9861319661140442, |
| "learning_rate": 4.591909155429383e-05, |
| "loss": 0.5383, |
| "step": 2173 |
| }, |
| { |
| "epoch": 3.08, |
| "grad_norm": 0.9648599028587341, |
| "learning_rate": 4.5848119233498934e-05, |
| "loss": 0.4916, |
| "step": 2174 |
| }, |
| { |
| "epoch": 3.08, |
| "grad_norm": 0.9462875127792358, |
| "learning_rate": 4.577714691270405e-05, |
| "loss": 0.4533, |
| "step": 2175 |
| }, |
| { |
| "epoch": 3.08, |
| "grad_norm": 1.0142507553100586, |
| "learning_rate": 4.570617459190916e-05, |
| "loss": 0.5145, |
| "step": 2176 |
| }, |
| { |
| "epoch": 3.08, |
| "grad_norm": 0.9575462341308594, |
| "learning_rate": 4.563520227111427e-05, |
| "loss": 0.471, |
| "step": 2177 |
| }, |
| { |
| "epoch": 3.09, |
| "grad_norm": 1.0417957305908203, |
| "learning_rate": 4.556422995031938e-05, |
| "loss": 0.5373, |
| "step": 2178 |
| }, |
| { |
| "epoch": 3.09, |
| "grad_norm": 0.9894844889640808, |
| "learning_rate": 4.549325762952449e-05, |
| "loss": 0.5464, |
| "step": 2179 |
| }, |
| { |
| "epoch": 3.09, |
| "grad_norm": 0.9838773608207703, |
| "learning_rate": 4.54222853087296e-05, |
| "loss": 0.5221, |
| "step": 2180 |
| }, |
| { |
| "epoch": 3.09, |
| "grad_norm": 0.9868409633636475, |
| "learning_rate": 4.535131298793471e-05, |
| "loss": 0.5311, |
| "step": 2181 |
| }, |
| { |
| "epoch": 3.09, |
| "grad_norm": 0.944324791431427, |
| "learning_rate": 4.528034066713982e-05, |
| "loss": 0.5446, |
| "step": 2182 |
| }, |
| { |
| "epoch": 3.09, |
| "grad_norm": 0.9434787034988403, |
| "learning_rate": 4.520936834634493e-05, |
| "loss": 0.5004, |
| "step": 2183 |
| }, |
| { |
| "epoch": 3.09, |
| "grad_norm": 0.9857025146484375, |
| "learning_rate": 4.513839602555004e-05, |
| "loss": 0.5373, |
| "step": 2184 |
| }, |
| { |
| "epoch": 3.1, |
| "grad_norm": 0.9952114224433899, |
| "learning_rate": 4.506742370475515e-05, |
| "loss": 0.5013, |
| "step": 2185 |
| }, |
| { |
| "epoch": 3.1, |
| "grad_norm": 0.9452676177024841, |
| "learning_rate": 4.499645138396026e-05, |
| "loss": 0.5079, |
| "step": 2186 |
| }, |
| { |
| "epoch": 3.1, |
| "grad_norm": 1.0157325267791748, |
| "learning_rate": 4.4925479063165366e-05, |
| "loss": 0.5197, |
| "step": 2187 |
| }, |
| { |
| "epoch": 3.1, |
| "grad_norm": 0.9407671093940735, |
| "learning_rate": 4.485450674237048e-05, |
| "loss": 0.4895, |
| "step": 2188 |
| }, |
| { |
| "epoch": 3.1, |
| "grad_norm": 0.9767752289772034, |
| "learning_rate": 4.478353442157559e-05, |
| "loss": 0.5002, |
| "step": 2189 |
| }, |
| { |
| "epoch": 3.1, |
| "grad_norm": 0.9899623990058899, |
| "learning_rate": 4.47125621007807e-05, |
| "loss": 0.4964, |
| "step": 2190 |
| }, |
| { |
| "epoch": 3.1, |
| "grad_norm": 1.0107783079147339, |
| "learning_rate": 4.464158977998581e-05, |
| "loss": 0.4909, |
| "step": 2191 |
| }, |
| { |
| "epoch": 3.11, |
| "grad_norm": 1.00058114528656, |
| "learning_rate": 4.4570617459190916e-05, |
| "loss": 0.5369, |
| "step": 2192 |
| }, |
| { |
| "epoch": 3.11, |
| "grad_norm": 0.9973490238189697, |
| "learning_rate": 4.449964513839603e-05, |
| "loss": 0.5219, |
| "step": 2193 |
| }, |
| { |
| "epoch": 3.11, |
| "grad_norm": 0.9762208461761475, |
| "learning_rate": 4.4428672817601135e-05, |
| "loss": 0.5218, |
| "step": 2194 |
| }, |
| { |
| "epoch": 3.11, |
| "grad_norm": 0.994667112827301, |
| "learning_rate": 4.435770049680625e-05, |
| "loss": 0.5097, |
| "step": 2195 |
| }, |
| { |
| "epoch": 3.11, |
| "grad_norm": 0.9990437030792236, |
| "learning_rate": 4.428672817601135e-05, |
| "loss": 0.5359, |
| "step": 2196 |
| }, |
| { |
| "epoch": 3.11, |
| "grad_norm": 0.9898905754089355, |
| "learning_rate": 4.4215755855216466e-05, |
| "loss": 0.4987, |
| "step": 2197 |
| }, |
| { |
| "epoch": 3.11, |
| "grad_norm": 0.9653964638710022, |
| "learning_rate": 4.414478353442158e-05, |
| "loss": 0.4742, |
| "step": 2198 |
| }, |
| { |
| "epoch": 3.12, |
| "grad_norm": 0.9734763503074646, |
| "learning_rate": 4.4073811213626685e-05, |
| "loss": 0.5236, |
| "step": 2199 |
| }, |
| { |
| "epoch": 3.12, |
| "grad_norm": 0.9727082848548889, |
| "learning_rate": 4.40028388928318e-05, |
| "loss": 0.5181, |
| "step": 2200 |
| }, |
| { |
| "epoch": 3.12, |
| "grad_norm": 1.0015164613723755, |
| "learning_rate": 4.39318665720369e-05, |
| "loss": 0.502, |
| "step": 2201 |
| }, |
| { |
| "epoch": 3.12, |
| "grad_norm": 0.965710461139679, |
| "learning_rate": 4.3860894251242016e-05, |
| "loss": 0.5046, |
| "step": 2202 |
| }, |
| { |
| "epoch": 3.12, |
| "grad_norm": 0.9759023189544678, |
| "learning_rate": 4.378992193044712e-05, |
| "loss": 0.5203, |
| "step": 2203 |
| }, |
| { |
| "epoch": 3.12, |
| "grad_norm": 0.9397246837615967, |
| "learning_rate": 4.3718949609652235e-05, |
| "loss": 0.4816, |
| "step": 2204 |
| }, |
| { |
| "epoch": 3.12, |
| "grad_norm": 0.979550302028656, |
| "learning_rate": 4.364797728885735e-05, |
| "loss": 0.5214, |
| "step": 2205 |
| }, |
| { |
| "epoch": 3.13, |
| "grad_norm": 0.9628053307533264, |
| "learning_rate": 4.357700496806246e-05, |
| "loss": 0.5117, |
| "step": 2206 |
| }, |
| { |
| "epoch": 3.13, |
| "grad_norm": 1.0112930536270142, |
| "learning_rate": 4.3506032647267566e-05, |
| "loss": 0.5711, |
| "step": 2207 |
| }, |
| { |
| "epoch": 3.13, |
| "grad_norm": 1.0025320053100586, |
| "learning_rate": 4.343506032647268e-05, |
| "loss": 0.5474, |
| "step": 2208 |
| }, |
| { |
| "epoch": 3.13, |
| "grad_norm": 1.0063894987106323, |
| "learning_rate": 4.336408800567779e-05, |
| "loss": 0.5618, |
| "step": 2209 |
| }, |
| { |
| "epoch": 3.13, |
| "grad_norm": 1.0305397510528564, |
| "learning_rate": 4.32931156848829e-05, |
| "loss": 0.5585, |
| "step": 2210 |
| }, |
| { |
| "epoch": 3.13, |
| "grad_norm": 0.9876142144203186, |
| "learning_rate": 4.322214336408801e-05, |
| "loss": 0.517, |
| "step": 2211 |
| }, |
| { |
| "epoch": 3.13, |
| "grad_norm": 0.9939770102500916, |
| "learning_rate": 4.315117104329312e-05, |
| "loss": 0.4898, |
| "step": 2212 |
| }, |
| { |
| "epoch": 3.14, |
| "grad_norm": 0.9995191097259521, |
| "learning_rate": 4.308019872249823e-05, |
| "loss": 0.5408, |
| "step": 2213 |
| }, |
| { |
| "epoch": 3.14, |
| "grad_norm": 1.0050206184387207, |
| "learning_rate": 4.300922640170334e-05, |
| "loss": 0.5386, |
| "step": 2214 |
| }, |
| { |
| "epoch": 3.14, |
| "grad_norm": 0.9904208183288574, |
| "learning_rate": 4.293825408090845e-05, |
| "loss": 0.5028, |
| "step": 2215 |
| }, |
| { |
| "epoch": 3.14, |
| "grad_norm": 1.012172818183899, |
| "learning_rate": 4.286728176011356e-05, |
| "loss": 0.5367, |
| "step": 2216 |
| }, |
| { |
| "epoch": 3.14, |
| "grad_norm": 0.981663703918457, |
| "learning_rate": 4.2796309439318666e-05, |
| "loss": 0.5185, |
| "step": 2217 |
| }, |
| { |
| "epoch": 3.14, |
| "grad_norm": 0.9844231009483337, |
| "learning_rate": 4.272533711852378e-05, |
| "loss": 0.5212, |
| "step": 2218 |
| }, |
| { |
| "epoch": 3.14, |
| "grad_norm": 0.9846050143241882, |
| "learning_rate": 4.2654364797728885e-05, |
| "loss": 0.5427, |
| "step": 2219 |
| }, |
| { |
| "epoch": 3.15, |
| "grad_norm": 0.9734468460083008, |
| "learning_rate": 4.2583392476934e-05, |
| "loss": 0.5248, |
| "step": 2220 |
| }, |
| { |
| "epoch": 3.15, |
| "grad_norm": 0.9388629794120789, |
| "learning_rate": 4.251242015613911e-05, |
| "loss": 0.4593, |
| "step": 2221 |
| }, |
| { |
| "epoch": 3.15, |
| "grad_norm": 0.9485184550285339, |
| "learning_rate": 4.2441447835344216e-05, |
| "loss": 0.4847, |
| "step": 2222 |
| }, |
| { |
| "epoch": 3.15, |
| "grad_norm": 1.0168582201004028, |
| "learning_rate": 4.237047551454933e-05, |
| "loss": 0.5204, |
| "step": 2223 |
| }, |
| { |
| "epoch": 3.15, |
| "grad_norm": 0.9456900358200073, |
| "learning_rate": 4.2299503193754435e-05, |
| "loss": 0.4953, |
| "step": 2224 |
| }, |
| { |
| "epoch": 3.15, |
| "grad_norm": 1.0003365278244019, |
| "learning_rate": 4.222853087295955e-05, |
| "loss": 0.5383, |
| "step": 2225 |
| }, |
| { |
| "epoch": 3.15, |
| "grad_norm": 1.0175565481185913, |
| "learning_rate": 4.2157558552164654e-05, |
| "loss": 0.5451, |
| "step": 2226 |
| }, |
| { |
| "epoch": 3.15, |
| "grad_norm": 1.0304280519485474, |
| "learning_rate": 4.2086586231369766e-05, |
| "loss": 0.5467, |
| "step": 2227 |
| }, |
| { |
| "epoch": 3.16, |
| "grad_norm": 0.9860724806785583, |
| "learning_rate": 4.201561391057488e-05, |
| "loss": 0.5399, |
| "step": 2228 |
| }, |
| { |
| "epoch": 3.16, |
| "grad_norm": 0.9857328534126282, |
| "learning_rate": 4.1944641589779985e-05, |
| "loss": 0.5208, |
| "step": 2229 |
| }, |
| { |
| "epoch": 3.16, |
| "grad_norm": 1.07256281375885, |
| "learning_rate": 4.18736692689851e-05, |
| "loss": 0.5466, |
| "step": 2230 |
| }, |
| { |
| "epoch": 3.16, |
| "grad_norm": 1.0171128511428833, |
| "learning_rate": 4.1802696948190204e-05, |
| "loss": 0.5309, |
| "step": 2231 |
| }, |
| { |
| "epoch": 3.16, |
| "grad_norm": 1.0043413639068604, |
| "learning_rate": 4.1731724627395316e-05, |
| "loss": 0.518, |
| "step": 2232 |
| }, |
| { |
| "epoch": 3.16, |
| "grad_norm": 1.0043971538543701, |
| "learning_rate": 4.166075230660042e-05, |
| "loss": 0.5515, |
| "step": 2233 |
| }, |
| { |
| "epoch": 3.16, |
| "grad_norm": 1.00337553024292, |
| "learning_rate": 4.1589779985805535e-05, |
| "loss": 0.5425, |
| "step": 2234 |
| }, |
| { |
| "epoch": 3.17, |
| "grad_norm": 1.0087389945983887, |
| "learning_rate": 4.151880766501065e-05, |
| "loss": 0.5237, |
| "step": 2235 |
| }, |
| { |
| "epoch": 3.17, |
| "grad_norm": 1.0248241424560547, |
| "learning_rate": 4.144783534421576e-05, |
| "loss": 0.5685, |
| "step": 2236 |
| }, |
| { |
| "epoch": 3.17, |
| "grad_norm": 1.0191607475280762, |
| "learning_rate": 4.1376863023420867e-05, |
| "loss": 0.5436, |
| "step": 2237 |
| }, |
| { |
| "epoch": 3.17, |
| "grad_norm": 0.9885305762290955, |
| "learning_rate": 4.130589070262598e-05, |
| "loss": 0.5082, |
| "step": 2238 |
| }, |
| { |
| "epoch": 3.17, |
| "grad_norm": 0.9865046739578247, |
| "learning_rate": 4.123491838183109e-05, |
| "loss": 0.5384, |
| "step": 2239 |
| }, |
| { |
| "epoch": 3.17, |
| "grad_norm": 0.984719455242157, |
| "learning_rate": 4.11639460610362e-05, |
| "loss": 0.5017, |
| "step": 2240 |
| }, |
| { |
| "epoch": 3.17, |
| "grad_norm": 0.9513193964958191, |
| "learning_rate": 4.109297374024131e-05, |
| "loss": 0.497, |
| "step": 2241 |
| }, |
| { |
| "epoch": 3.18, |
| "grad_norm": 1.0050876140594482, |
| "learning_rate": 4.102200141944642e-05, |
| "loss": 0.598, |
| "step": 2242 |
| }, |
| { |
| "epoch": 3.18, |
| "grad_norm": 0.9792728424072266, |
| "learning_rate": 4.095102909865153e-05, |
| "loss": 0.4996, |
| "step": 2243 |
| }, |
| { |
| "epoch": 3.18, |
| "grad_norm": 1.059917688369751, |
| "learning_rate": 4.088005677785664e-05, |
| "loss": 0.5584, |
| "step": 2244 |
| }, |
| { |
| "epoch": 3.18, |
| "grad_norm": 0.962172269821167, |
| "learning_rate": 4.080908445706175e-05, |
| "loss": 0.485, |
| "step": 2245 |
| }, |
| { |
| "epoch": 3.18, |
| "grad_norm": 0.9715234637260437, |
| "learning_rate": 4.073811213626686e-05, |
| "loss": 0.4752, |
| "step": 2246 |
| }, |
| { |
| "epoch": 3.18, |
| "grad_norm": 1.0108722448349, |
| "learning_rate": 4.066713981547197e-05, |
| "loss": 0.5053, |
| "step": 2247 |
| }, |
| { |
| "epoch": 3.18, |
| "grad_norm": 0.9890355467796326, |
| "learning_rate": 4.059616749467708e-05, |
| "loss": 0.514, |
| "step": 2248 |
| }, |
| { |
| "epoch": 3.19, |
| "grad_norm": 0.966625988483429, |
| "learning_rate": 4.0525195173882185e-05, |
| "loss": 0.4707, |
| "step": 2249 |
| }, |
| { |
| "epoch": 3.19, |
| "grad_norm": 1.0721231698989868, |
| "learning_rate": 4.04542228530873e-05, |
| "loss": 0.5867, |
| "step": 2250 |
| }, |
| { |
| "epoch": 3.19, |
| "grad_norm": 0.9937222003936768, |
| "learning_rate": 4.038325053229241e-05, |
| "loss": 0.4945, |
| "step": 2251 |
| }, |
| { |
| "epoch": 3.19, |
| "grad_norm": 1.0190373659133911, |
| "learning_rate": 4.031227821149752e-05, |
| "loss": 0.5281, |
| "step": 2252 |
| }, |
| { |
| "epoch": 3.19, |
| "grad_norm": 1.028527855873108, |
| "learning_rate": 4.024130589070263e-05, |
| "loss": 0.5356, |
| "step": 2253 |
| }, |
| { |
| "epoch": 3.19, |
| "grad_norm": 0.9732937216758728, |
| "learning_rate": 4.0170333569907735e-05, |
| "loss": 0.5344, |
| "step": 2254 |
| }, |
| { |
| "epoch": 3.19, |
| "grad_norm": 0.9506965279579163, |
| "learning_rate": 4.009936124911285e-05, |
| "loss": 0.498, |
| "step": 2255 |
| }, |
| { |
| "epoch": 3.2, |
| "grad_norm": 1.0024874210357666, |
| "learning_rate": 4.0028388928317954e-05, |
| "loss": 0.5234, |
| "step": 2256 |
| }, |
| { |
| "epoch": 3.2, |
| "grad_norm": 0.9432622790336609, |
| "learning_rate": 3.995741660752307e-05, |
| "loss": 0.5332, |
| "step": 2257 |
| }, |
| { |
| "epoch": 3.2, |
| "grad_norm": 0.9941532015800476, |
| "learning_rate": 3.988644428672818e-05, |
| "loss": 0.5044, |
| "step": 2258 |
| }, |
| { |
| "epoch": 3.2, |
| "grad_norm": 1.0081837177276611, |
| "learning_rate": 3.9815471965933286e-05, |
| "loss": 0.5043, |
| "step": 2259 |
| }, |
| { |
| "epoch": 3.2, |
| "grad_norm": 1.062863826751709, |
| "learning_rate": 3.97444996451384e-05, |
| "loss": 0.5296, |
| "step": 2260 |
| }, |
| { |
| "epoch": 3.2, |
| "grad_norm": 0.9733378291130066, |
| "learning_rate": 3.9673527324343504e-05, |
| "loss": 0.5047, |
| "step": 2261 |
| }, |
| { |
| "epoch": 3.2, |
| "grad_norm": 0.983255922794342, |
| "learning_rate": 3.960255500354862e-05, |
| "loss": 0.4895, |
| "step": 2262 |
| }, |
| { |
| "epoch": 3.21, |
| "grad_norm": 0.9561431407928467, |
| "learning_rate": 3.953158268275372e-05, |
| "loss": 0.4546, |
| "step": 2263 |
| }, |
| { |
| "epoch": 3.21, |
| "grad_norm": 1.0252975225448608, |
| "learning_rate": 3.9460610361958836e-05, |
| "loss": 0.5129, |
| "step": 2264 |
| }, |
| { |
| "epoch": 3.21, |
| "grad_norm": 0.9920056462287903, |
| "learning_rate": 3.938963804116395e-05, |
| "loss": 0.5163, |
| "step": 2265 |
| }, |
| { |
| "epoch": 3.21, |
| "grad_norm": 1.0454301834106445, |
| "learning_rate": 3.931866572036906e-05, |
| "loss": 0.5528, |
| "step": 2266 |
| }, |
| { |
| "epoch": 3.21, |
| "grad_norm": 0.9582340717315674, |
| "learning_rate": 3.924769339957417e-05, |
| "loss": 0.5108, |
| "step": 2267 |
| }, |
| { |
| "epoch": 3.21, |
| "grad_norm": 1.0326735973358154, |
| "learning_rate": 3.917672107877928e-05, |
| "loss": 0.5287, |
| "step": 2268 |
| }, |
| { |
| "epoch": 3.21, |
| "grad_norm": 0.9848905801773071, |
| "learning_rate": 3.910574875798439e-05, |
| "loss": 0.5271, |
| "step": 2269 |
| }, |
| { |
| "epoch": 3.22, |
| "grad_norm": 0.949379563331604, |
| "learning_rate": 3.90347764371895e-05, |
| "loss": 0.4887, |
| "step": 2270 |
| }, |
| { |
| "epoch": 3.22, |
| "grad_norm": 0.9528018236160278, |
| "learning_rate": 3.896380411639461e-05, |
| "loss": 0.4874, |
| "step": 2271 |
| }, |
| { |
| "epoch": 3.22, |
| "grad_norm": 0.96091228723526, |
| "learning_rate": 3.889283179559972e-05, |
| "loss": 0.5189, |
| "step": 2272 |
| }, |
| { |
| "epoch": 3.22, |
| "grad_norm": 1.0250592231750488, |
| "learning_rate": 3.882185947480483e-05, |
| "loss": 0.5368, |
| "step": 2273 |
| }, |
| { |
| "epoch": 3.22, |
| "grad_norm": 1.0342363119125366, |
| "learning_rate": 3.875088715400994e-05, |
| "loss": 0.5076, |
| "step": 2274 |
| }, |
| { |
| "epoch": 3.22, |
| "grad_norm": 0.9662345051765442, |
| "learning_rate": 3.867991483321505e-05, |
| "loss": 0.5021, |
| "step": 2275 |
| }, |
| { |
| "epoch": 3.22, |
| "grad_norm": 1.0056486129760742, |
| "learning_rate": 3.860894251242016e-05, |
| "loss": 0.5173, |
| "step": 2276 |
| }, |
| { |
| "epoch": 3.23, |
| "grad_norm": 0.9594873785972595, |
| "learning_rate": 3.853797019162527e-05, |
| "loss": 0.498, |
| "step": 2277 |
| }, |
| { |
| "epoch": 3.23, |
| "grad_norm": 1.001089096069336, |
| "learning_rate": 3.846699787083038e-05, |
| "loss": 0.5125, |
| "step": 2278 |
| }, |
| { |
| "epoch": 3.23, |
| "grad_norm": 1.0407207012176514, |
| "learning_rate": 3.8396025550035486e-05, |
| "loss": 0.5567, |
| "step": 2279 |
| }, |
| { |
| "epoch": 3.23, |
| "grad_norm": 0.9708674550056458, |
| "learning_rate": 3.83250532292406e-05, |
| "loss": 0.4893, |
| "step": 2280 |
| }, |
| { |
| "epoch": 3.23, |
| "grad_norm": 0.9955769181251526, |
| "learning_rate": 3.8254080908445704e-05, |
| "loss": 0.4873, |
| "step": 2281 |
| }, |
| { |
| "epoch": 3.23, |
| "grad_norm": 0.9997064471244812, |
| "learning_rate": 3.818310858765082e-05, |
| "loss": 0.5166, |
| "step": 2282 |
| }, |
| { |
| "epoch": 3.23, |
| "grad_norm": 0.9996336698532104, |
| "learning_rate": 3.811213626685593e-05, |
| "loss": 0.5054, |
| "step": 2283 |
| }, |
| { |
| "epoch": 3.24, |
| "grad_norm": 0.9845335483551025, |
| "learning_rate": 3.8041163946061036e-05, |
| "loss": 0.4903, |
| "step": 2284 |
| }, |
| { |
| "epoch": 3.24, |
| "grad_norm": 1.0074700117111206, |
| "learning_rate": 3.797019162526615e-05, |
| "loss": 0.5121, |
| "step": 2285 |
| }, |
| { |
| "epoch": 3.24, |
| "grad_norm": 0.9627556800842285, |
| "learning_rate": 3.7899219304471255e-05, |
| "loss": 0.5175, |
| "step": 2286 |
| }, |
| { |
| "epoch": 3.24, |
| "grad_norm": 1.0189731121063232, |
| "learning_rate": 3.782824698367637e-05, |
| "loss": 0.5354, |
| "step": 2287 |
| }, |
| { |
| "epoch": 3.24, |
| "grad_norm": 1.014862060546875, |
| "learning_rate": 3.775727466288147e-05, |
| "loss": 0.5459, |
| "step": 2288 |
| }, |
| { |
| "epoch": 3.24, |
| "grad_norm": 0.9731964468955994, |
| "learning_rate": 3.7686302342086586e-05, |
| "loss": 0.5227, |
| "step": 2289 |
| }, |
| { |
| "epoch": 3.24, |
| "grad_norm": 0.9890885353088379, |
| "learning_rate": 3.76153300212917e-05, |
| "loss": 0.5083, |
| "step": 2290 |
| }, |
| { |
| "epoch": 3.25, |
| "grad_norm": 0.9976668357849121, |
| "learning_rate": 3.7544357700496805e-05, |
| "loss": 0.5373, |
| "step": 2291 |
| }, |
| { |
| "epoch": 3.25, |
| "grad_norm": 1.04860258102417, |
| "learning_rate": 3.747338537970192e-05, |
| "loss": 0.511, |
| "step": 2292 |
| }, |
| { |
| "epoch": 3.25, |
| "grad_norm": 1.0599074363708496, |
| "learning_rate": 3.740241305890702e-05, |
| "loss": 0.5566, |
| "step": 2293 |
| }, |
| { |
| "epoch": 3.25, |
| "grad_norm": 0.9964849352836609, |
| "learning_rate": 3.7331440738112136e-05, |
| "loss": 0.5177, |
| "step": 2294 |
| }, |
| { |
| "epoch": 3.25, |
| "grad_norm": 0.9439302086830139, |
| "learning_rate": 3.726046841731725e-05, |
| "loss": 0.5044, |
| "step": 2295 |
| }, |
| { |
| "epoch": 3.25, |
| "grad_norm": 1.0093363523483276, |
| "learning_rate": 3.718949609652236e-05, |
| "loss": 0.5031, |
| "step": 2296 |
| }, |
| { |
| "epoch": 3.25, |
| "grad_norm": 0.9915724396705627, |
| "learning_rate": 3.711852377572747e-05, |
| "loss": 0.5038, |
| "step": 2297 |
| }, |
| { |
| "epoch": 3.26, |
| "grad_norm": 0.9472718834877014, |
| "learning_rate": 3.704755145493258e-05, |
| "loss": 0.4533, |
| "step": 2298 |
| }, |
| { |
| "epoch": 3.26, |
| "grad_norm": 0.9775221347808838, |
| "learning_rate": 3.697657913413769e-05, |
| "loss": 0.5127, |
| "step": 2299 |
| }, |
| { |
| "epoch": 3.26, |
| "grad_norm": 1.0333958864212036, |
| "learning_rate": 3.69056068133428e-05, |
| "loss": 0.5467, |
| "step": 2300 |
| }, |
| { |
| "epoch": 3.26, |
| "grad_norm": 1.0134251117706299, |
| "learning_rate": 3.683463449254791e-05, |
| "loss": 0.4907, |
| "step": 2301 |
| }, |
| { |
| "epoch": 3.26, |
| "grad_norm": 0.979189932346344, |
| "learning_rate": 3.676366217175302e-05, |
| "loss": 0.4739, |
| "step": 2302 |
| }, |
| { |
| "epoch": 3.26, |
| "grad_norm": 0.9691347479820251, |
| "learning_rate": 3.669268985095813e-05, |
| "loss": 0.5066, |
| "step": 2303 |
| }, |
| { |
| "epoch": 3.26, |
| "grad_norm": 1.0259438753128052, |
| "learning_rate": 3.6621717530163236e-05, |
| "loss": 0.5329, |
| "step": 2304 |
| }, |
| { |
| "epoch": 3.27, |
| "grad_norm": 1.004991054534912, |
| "learning_rate": 3.655074520936835e-05, |
| "loss": 0.5315, |
| "step": 2305 |
| }, |
| { |
| "epoch": 3.27, |
| "grad_norm": 1.0028557777404785, |
| "learning_rate": 3.647977288857346e-05, |
| "loss": 0.505, |
| "step": 2306 |
| }, |
| { |
| "epoch": 3.27, |
| "grad_norm": 1.0041579008102417, |
| "learning_rate": 3.640880056777857e-05, |
| "loss": 0.5333, |
| "step": 2307 |
| }, |
| { |
| "epoch": 3.27, |
| "grad_norm": 0.9609581828117371, |
| "learning_rate": 3.633782824698368e-05, |
| "loss": 0.5016, |
| "step": 2308 |
| }, |
| { |
| "epoch": 3.27, |
| "grad_norm": 0.9761263728141785, |
| "learning_rate": 3.6266855926188786e-05, |
| "loss": 0.5014, |
| "step": 2309 |
| }, |
| { |
| "epoch": 3.27, |
| "grad_norm": 0.9792344570159912, |
| "learning_rate": 3.61958836053939e-05, |
| "loss": 0.4831, |
| "step": 2310 |
| }, |
| { |
| "epoch": 3.27, |
| "grad_norm": 1.0584090948104858, |
| "learning_rate": 3.6124911284599005e-05, |
| "loss": 0.5732, |
| "step": 2311 |
| }, |
| { |
| "epoch": 3.28, |
| "grad_norm": 0.9740521907806396, |
| "learning_rate": 3.605393896380412e-05, |
| "loss": 0.4929, |
| "step": 2312 |
| }, |
| { |
| "epoch": 3.28, |
| "grad_norm": 0.9656211137771606, |
| "learning_rate": 3.598296664300923e-05, |
| "loss": 0.5054, |
| "step": 2313 |
| }, |
| { |
| "epoch": 3.28, |
| "grad_norm": 0.9914239048957825, |
| "learning_rate": 3.5911994322214336e-05, |
| "loss": 0.5089, |
| "step": 2314 |
| }, |
| { |
| "epoch": 3.28, |
| "grad_norm": 1.0298843383789062, |
| "learning_rate": 3.584102200141945e-05, |
| "loss": 0.544, |
| "step": 2315 |
| }, |
| { |
| "epoch": 3.28, |
| "grad_norm": 0.9930390119552612, |
| "learning_rate": 3.5770049680624555e-05, |
| "loss": 0.5594, |
| "step": 2316 |
| }, |
| { |
| "epoch": 3.28, |
| "grad_norm": 1.0042483806610107, |
| "learning_rate": 3.569907735982967e-05, |
| "loss": 0.4972, |
| "step": 2317 |
| }, |
| { |
| "epoch": 3.28, |
| "grad_norm": 0.9482210874557495, |
| "learning_rate": 3.5628105039034774e-05, |
| "loss": 0.4537, |
| "step": 2318 |
| }, |
| { |
| "epoch": 3.29, |
| "grad_norm": 1.0086231231689453, |
| "learning_rate": 3.5557132718239886e-05, |
| "loss": 0.5497, |
| "step": 2319 |
| }, |
| { |
| "epoch": 3.29, |
| "grad_norm": 1.0223073959350586, |
| "learning_rate": 3.5486160397445e-05, |
| "loss": 0.4925, |
| "step": 2320 |
| }, |
| { |
| "epoch": 3.29, |
| "grad_norm": 1.0105258226394653, |
| "learning_rate": 3.5415188076650105e-05, |
| "loss": 0.513, |
| "step": 2321 |
| }, |
| { |
| "epoch": 3.29, |
| "grad_norm": 1.0318858623504639, |
| "learning_rate": 3.534421575585522e-05, |
| "loss": 0.5573, |
| "step": 2322 |
| }, |
| { |
| "epoch": 3.29, |
| "grad_norm": 0.9838947653770447, |
| "learning_rate": 3.5273243435060324e-05, |
| "loss": 0.5053, |
| "step": 2323 |
| }, |
| { |
| "epoch": 3.29, |
| "grad_norm": 0.9841655492782593, |
| "learning_rate": 3.5202271114265437e-05, |
| "loss": 0.4869, |
| "step": 2324 |
| }, |
| { |
| "epoch": 3.29, |
| "grad_norm": 1.0124472379684448, |
| "learning_rate": 3.513129879347055e-05, |
| "loss": 0.529, |
| "step": 2325 |
| }, |
| { |
| "epoch": 3.3, |
| "grad_norm": 1.0195896625518799, |
| "learning_rate": 3.506032647267566e-05, |
| "loss": 0.5419, |
| "step": 2326 |
| }, |
| { |
| "epoch": 3.3, |
| "grad_norm": 0.9941637516021729, |
| "learning_rate": 3.498935415188077e-05, |
| "loss": 0.5149, |
| "step": 2327 |
| }, |
| { |
| "epoch": 3.3, |
| "grad_norm": 1.036328911781311, |
| "learning_rate": 3.491838183108588e-05, |
| "loss": 0.5412, |
| "step": 2328 |
| }, |
| { |
| "epoch": 3.3, |
| "grad_norm": 0.9751386046409607, |
| "learning_rate": 3.484740951029099e-05, |
| "loss": 0.5206, |
| "step": 2329 |
| }, |
| { |
| "epoch": 3.3, |
| "grad_norm": 1.0403040647506714, |
| "learning_rate": 3.47764371894961e-05, |
| "loss": 0.5609, |
| "step": 2330 |
| }, |
| { |
| "epoch": 3.3, |
| "grad_norm": 0.9578166604042053, |
| "learning_rate": 3.470546486870121e-05, |
| "loss": 0.4783, |
| "step": 2331 |
| }, |
| { |
| "epoch": 3.3, |
| "grad_norm": 1.0178450345993042, |
| "learning_rate": 3.463449254790632e-05, |
| "loss": 0.531, |
| "step": 2332 |
| }, |
| { |
| "epoch": 3.31, |
| "grad_norm": 1.0271254777908325, |
| "learning_rate": 3.456352022711143e-05, |
| "loss": 0.5501, |
| "step": 2333 |
| }, |
| { |
| "epoch": 3.31, |
| "grad_norm": 0.9989926815032959, |
| "learning_rate": 3.449254790631654e-05, |
| "loss": 0.5176, |
| "step": 2334 |
| }, |
| { |
| "epoch": 3.31, |
| "grad_norm": 1.00890052318573, |
| "learning_rate": 3.442157558552165e-05, |
| "loss": 0.5352, |
| "step": 2335 |
| }, |
| { |
| "epoch": 3.31, |
| "grad_norm": 1.0696969032287598, |
| "learning_rate": 3.435060326472676e-05, |
| "loss": 0.539, |
| "step": 2336 |
| }, |
| { |
| "epoch": 3.31, |
| "grad_norm": 1.0061776638031006, |
| "learning_rate": 3.427963094393187e-05, |
| "loss": 0.52, |
| "step": 2337 |
| }, |
| { |
| "epoch": 3.31, |
| "grad_norm": 1.0269782543182373, |
| "learning_rate": 3.420865862313698e-05, |
| "loss": 0.5162, |
| "step": 2338 |
| }, |
| { |
| "epoch": 3.31, |
| "grad_norm": 0.9664796590805054, |
| "learning_rate": 3.413768630234209e-05, |
| "loss": 0.4884, |
| "step": 2339 |
| }, |
| { |
| "epoch": 3.32, |
| "grad_norm": 0.9960459470748901, |
| "learning_rate": 3.40667139815472e-05, |
| "loss": 0.5062, |
| "step": 2340 |
| }, |
| { |
| "epoch": 3.32, |
| "grad_norm": 1.0411757230758667, |
| "learning_rate": 3.3995741660752305e-05, |
| "loss": 0.5715, |
| "step": 2341 |
| }, |
| { |
| "epoch": 3.32, |
| "grad_norm": 0.9852669835090637, |
| "learning_rate": 3.392476933995742e-05, |
| "loss": 0.4707, |
| "step": 2342 |
| }, |
| { |
| "epoch": 3.32, |
| "grad_norm": 1.0226832628250122, |
| "learning_rate": 3.385379701916253e-05, |
| "loss": 0.5318, |
| "step": 2343 |
| }, |
| { |
| "epoch": 3.32, |
| "grad_norm": 1.077976107597351, |
| "learning_rate": 3.378282469836764e-05, |
| "loss": 0.5557, |
| "step": 2344 |
| }, |
| { |
| "epoch": 3.32, |
| "grad_norm": 1.0263750553131104, |
| "learning_rate": 3.371185237757275e-05, |
| "loss": 0.5563, |
| "step": 2345 |
| }, |
| { |
| "epoch": 3.32, |
| "grad_norm": 1.007447600364685, |
| "learning_rate": 3.3640880056777855e-05, |
| "loss": 0.5036, |
| "step": 2346 |
| }, |
| { |
| "epoch": 3.32, |
| "grad_norm": 0.9899296760559082, |
| "learning_rate": 3.356990773598297e-05, |
| "loss": 0.5017, |
| "step": 2347 |
| }, |
| { |
| "epoch": 3.33, |
| "grad_norm": 0.9798328876495361, |
| "learning_rate": 3.3498935415188074e-05, |
| "loss": 0.5436, |
| "step": 2348 |
| }, |
| { |
| "epoch": 3.33, |
| "grad_norm": 0.9951174259185791, |
| "learning_rate": 3.342796309439319e-05, |
| "loss": 0.5456, |
| "step": 2349 |
| }, |
| { |
| "epoch": 3.33, |
| "grad_norm": 1.0041948556900024, |
| "learning_rate": 3.335699077359829e-05, |
| "loss": 0.5188, |
| "step": 2350 |
| }, |
| { |
| "epoch": 3.33, |
| "grad_norm": 1.0109312534332275, |
| "learning_rate": 3.3286018452803406e-05, |
| "loss": 0.5191, |
| "step": 2351 |
| }, |
| { |
| "epoch": 3.33, |
| "grad_norm": 0.9915183186531067, |
| "learning_rate": 3.321504613200852e-05, |
| "loss": 0.532, |
| "step": 2352 |
| }, |
| { |
| "epoch": 3.33, |
| "grad_norm": 1.0262149572372437, |
| "learning_rate": 3.3144073811213624e-05, |
| "loss": 0.5295, |
| "step": 2353 |
| }, |
| { |
| "epoch": 3.33, |
| "grad_norm": 1.0315930843353271, |
| "learning_rate": 3.307310149041874e-05, |
| "loss": 0.5774, |
| "step": 2354 |
| }, |
| { |
| "epoch": 3.34, |
| "grad_norm": 1.0199474096298218, |
| "learning_rate": 3.300212916962385e-05, |
| "loss": 0.5729, |
| "step": 2355 |
| }, |
| { |
| "epoch": 3.34, |
| "grad_norm": 0.9959515333175659, |
| "learning_rate": 3.2931156848828956e-05, |
| "loss": 0.5117, |
| "step": 2356 |
| }, |
| { |
| "epoch": 3.34, |
| "grad_norm": 0.9782159328460693, |
| "learning_rate": 3.286018452803407e-05, |
| "loss": 0.4816, |
| "step": 2357 |
| }, |
| { |
| "epoch": 3.34, |
| "grad_norm": 1.0041710138320923, |
| "learning_rate": 3.278921220723918e-05, |
| "loss": 0.5353, |
| "step": 2358 |
| }, |
| { |
| "epoch": 3.34, |
| "grad_norm": 1.0350843667984009, |
| "learning_rate": 3.2718239886444294e-05, |
| "loss": 0.5499, |
| "step": 2359 |
| }, |
| { |
| "epoch": 3.34, |
| "grad_norm": 1.0022941827774048, |
| "learning_rate": 3.26472675656494e-05, |
| "loss": 0.5387, |
| "step": 2360 |
| }, |
| { |
| "epoch": 3.34, |
| "grad_norm": 0.978432834148407, |
| "learning_rate": 3.257629524485451e-05, |
| "loss": 0.5186, |
| "step": 2361 |
| }, |
| { |
| "epoch": 3.35, |
| "grad_norm": 0.9733918309211731, |
| "learning_rate": 3.250532292405962e-05, |
| "loss": 0.5027, |
| "step": 2362 |
| }, |
| { |
| "epoch": 3.35, |
| "grad_norm": 0.9844115376472473, |
| "learning_rate": 3.243435060326473e-05, |
| "loss": 0.5145, |
| "step": 2363 |
| }, |
| { |
| "epoch": 3.35, |
| "grad_norm": 0.9787808656692505, |
| "learning_rate": 3.236337828246984e-05, |
| "loss": 0.4654, |
| "step": 2364 |
| }, |
| { |
| "epoch": 3.35, |
| "grad_norm": 1.0499181747436523, |
| "learning_rate": 3.229240596167495e-05, |
| "loss": 0.5723, |
| "step": 2365 |
| }, |
| { |
| "epoch": 3.35, |
| "grad_norm": 0.9862911701202393, |
| "learning_rate": 3.222143364088006e-05, |
| "loss": 0.4858, |
| "step": 2366 |
| }, |
| { |
| "epoch": 3.35, |
| "grad_norm": 1.0450081825256348, |
| "learning_rate": 3.215046132008517e-05, |
| "loss": 0.5636, |
| "step": 2367 |
| }, |
| { |
| "epoch": 3.35, |
| "grad_norm": 1.0021899938583374, |
| "learning_rate": 3.207948899929028e-05, |
| "loss": 0.5229, |
| "step": 2368 |
| }, |
| { |
| "epoch": 3.36, |
| "grad_norm": 0.9917883276939392, |
| "learning_rate": 3.200851667849539e-05, |
| "loss": 0.5194, |
| "step": 2369 |
| }, |
| { |
| "epoch": 3.36, |
| "grad_norm": 1.058908462524414, |
| "learning_rate": 3.19375443577005e-05, |
| "loss": 0.5492, |
| "step": 2370 |
| }, |
| { |
| "epoch": 3.36, |
| "grad_norm": 1.0056339502334595, |
| "learning_rate": 3.1866572036905606e-05, |
| "loss": 0.5112, |
| "step": 2371 |
| }, |
| { |
| "epoch": 3.36, |
| "grad_norm": 1.0141656398773193, |
| "learning_rate": 3.179559971611072e-05, |
| "loss": 0.5233, |
| "step": 2372 |
| }, |
| { |
| "epoch": 3.36, |
| "grad_norm": 0.9797987937927246, |
| "learning_rate": 3.1724627395315825e-05, |
| "loss": 0.4747, |
| "step": 2373 |
| }, |
| { |
| "epoch": 3.36, |
| "grad_norm": 1.0097317695617676, |
| "learning_rate": 3.165365507452094e-05, |
| "loss": 0.5095, |
| "step": 2374 |
| }, |
| { |
| "epoch": 3.36, |
| "grad_norm": 1.0642932653427124, |
| "learning_rate": 3.158268275372605e-05, |
| "loss": 0.5635, |
| "step": 2375 |
| }, |
| { |
| "epoch": 3.37, |
| "grad_norm": 0.9883024096488953, |
| "learning_rate": 3.1511710432931156e-05, |
| "loss": 0.5196, |
| "step": 2376 |
| }, |
| { |
| "epoch": 3.37, |
| "grad_norm": 1.0617568492889404, |
| "learning_rate": 3.144073811213627e-05, |
| "loss": 0.5688, |
| "step": 2377 |
| }, |
| { |
| "epoch": 3.37, |
| "grad_norm": 0.9752853512763977, |
| "learning_rate": 3.1369765791341375e-05, |
| "loss": 0.5124, |
| "step": 2378 |
| }, |
| { |
| "epoch": 3.37, |
| "grad_norm": 1.0084364414215088, |
| "learning_rate": 3.129879347054649e-05, |
| "loss": 0.5194, |
| "step": 2379 |
| }, |
| { |
| "epoch": 3.37, |
| "grad_norm": 1.0282278060913086, |
| "learning_rate": 3.122782114975159e-05, |
| "loss": 0.5088, |
| "step": 2380 |
| }, |
| { |
| "epoch": 3.37, |
| "grad_norm": 1.0256778001785278, |
| "learning_rate": 3.1156848828956706e-05, |
| "loss": 0.5544, |
| "step": 2381 |
| }, |
| { |
| "epoch": 3.37, |
| "grad_norm": 1.0465688705444336, |
| "learning_rate": 3.108587650816182e-05, |
| "loss": 0.529, |
| "step": 2382 |
| }, |
| { |
| "epoch": 3.38, |
| "grad_norm": 1.0042644739151, |
| "learning_rate": 3.1014904187366925e-05, |
| "loss": 0.5066, |
| "step": 2383 |
| }, |
| { |
| "epoch": 3.38, |
| "grad_norm": 1.0506339073181152, |
| "learning_rate": 3.094393186657204e-05, |
| "loss": 0.5395, |
| "step": 2384 |
| }, |
| { |
| "epoch": 3.38, |
| "grad_norm": 0.9986273050308228, |
| "learning_rate": 3.087295954577715e-05, |
| "loss": 0.4854, |
| "step": 2385 |
| }, |
| { |
| "epoch": 3.38, |
| "grad_norm": 0.9996753334999084, |
| "learning_rate": 3.0801987224982256e-05, |
| "loss": 0.5069, |
| "step": 2386 |
| }, |
| { |
| "epoch": 3.38, |
| "grad_norm": 1.0293785333633423, |
| "learning_rate": 3.073101490418737e-05, |
| "loss": 0.5439, |
| "step": 2387 |
| }, |
| { |
| "epoch": 3.38, |
| "grad_norm": 1.012476921081543, |
| "learning_rate": 3.066004258339248e-05, |
| "loss": 0.5267, |
| "step": 2388 |
| }, |
| { |
| "epoch": 3.38, |
| "grad_norm": 1.0480303764343262, |
| "learning_rate": 3.0589070262597594e-05, |
| "loss": 0.5875, |
| "step": 2389 |
| }, |
| { |
| "epoch": 3.39, |
| "grad_norm": 1.0067976713180542, |
| "learning_rate": 3.05180979418027e-05, |
| "loss": 0.5189, |
| "step": 2390 |
| }, |
| { |
| "epoch": 3.39, |
| "grad_norm": 0.967479407787323, |
| "learning_rate": 3.044712562100781e-05, |
| "loss": 0.4723, |
| "step": 2391 |
| }, |
| { |
| "epoch": 3.39, |
| "grad_norm": 0.9812818169593811, |
| "learning_rate": 3.037615330021292e-05, |
| "loss": 0.5152, |
| "step": 2392 |
| }, |
| { |
| "epoch": 3.39, |
| "grad_norm": 1.037245512008667, |
| "learning_rate": 3.0305180979418028e-05, |
| "loss": 0.5402, |
| "step": 2393 |
| }, |
| { |
| "epoch": 3.39, |
| "grad_norm": 1.0045421123504639, |
| "learning_rate": 3.0234208658623138e-05, |
| "loss": 0.536, |
| "step": 2394 |
| }, |
| { |
| "epoch": 3.39, |
| "grad_norm": 0.9535611867904663, |
| "learning_rate": 3.016323633782825e-05, |
| "loss": 0.4775, |
| "step": 2395 |
| }, |
| { |
| "epoch": 3.39, |
| "grad_norm": 1.007875919342041, |
| "learning_rate": 3.0092264017033356e-05, |
| "loss": 0.4937, |
| "step": 2396 |
| }, |
| { |
| "epoch": 3.4, |
| "grad_norm": 1.0232481956481934, |
| "learning_rate": 3.002129169623847e-05, |
| "loss": 0.5054, |
| "step": 2397 |
| }, |
| { |
| "epoch": 3.4, |
| "grad_norm": 1.022507905960083, |
| "learning_rate": 2.9950319375443582e-05, |
| "loss": 0.5438, |
| "step": 2398 |
| }, |
| { |
| "epoch": 3.4, |
| "grad_norm": 1.0051137208938599, |
| "learning_rate": 2.9879347054648688e-05, |
| "loss": 0.5085, |
| "step": 2399 |
| }, |
| { |
| "epoch": 3.4, |
| "grad_norm": 0.9932114481925964, |
| "learning_rate": 2.98083747338538e-05, |
| "loss": 0.5117, |
| "step": 2400 |
| }, |
| { |
| "epoch": 3.4, |
| "grad_norm": 1.0523576736450195, |
| "learning_rate": 2.9737402413058906e-05, |
| "loss": 0.5187, |
| "step": 2401 |
| }, |
| { |
| "epoch": 3.4, |
| "grad_norm": 1.0372650623321533, |
| "learning_rate": 2.966643009226402e-05, |
| "loss": 0.5101, |
| "step": 2402 |
| }, |
| { |
| "epoch": 3.4, |
| "grad_norm": 0.9998854398727417, |
| "learning_rate": 2.9595457771469125e-05, |
| "loss": 0.5143, |
| "step": 2403 |
| }, |
| { |
| "epoch": 3.41, |
| "grad_norm": 1.0421397686004639, |
| "learning_rate": 2.9524485450674238e-05, |
| "loss": 0.5271, |
| "step": 2404 |
| }, |
| { |
| "epoch": 3.41, |
| "grad_norm": 1.0461199283599854, |
| "learning_rate": 2.945351312987935e-05, |
| "loss": 0.5193, |
| "step": 2405 |
| }, |
| { |
| "epoch": 3.41, |
| "grad_norm": 1.0345731973648071, |
| "learning_rate": 2.9382540809084456e-05, |
| "loss": 0.5227, |
| "step": 2406 |
| }, |
| { |
| "epoch": 3.41, |
| "grad_norm": 1.0057165622711182, |
| "learning_rate": 2.931156848828957e-05, |
| "loss": 0.507, |
| "step": 2407 |
| }, |
| { |
| "epoch": 3.41, |
| "grad_norm": 1.0056483745574951, |
| "learning_rate": 2.924059616749468e-05, |
| "loss": 0.5158, |
| "step": 2408 |
| }, |
| { |
| "epoch": 3.41, |
| "grad_norm": 1.0037814378738403, |
| "learning_rate": 2.916962384669979e-05, |
| "loss": 0.5409, |
| "step": 2409 |
| }, |
| { |
| "epoch": 3.41, |
| "grad_norm": 0.9976779818534851, |
| "learning_rate": 2.9098651525904897e-05, |
| "loss": 0.5262, |
| "step": 2410 |
| }, |
| { |
| "epoch": 3.42, |
| "grad_norm": 0.9764635562896729, |
| "learning_rate": 2.902767920511001e-05, |
| "loss": 0.5068, |
| "step": 2411 |
| }, |
| { |
| "epoch": 3.42, |
| "grad_norm": 1.001948595046997, |
| "learning_rate": 2.8956706884315123e-05, |
| "loss": 0.5317, |
| "step": 2412 |
| }, |
| { |
| "epoch": 3.42, |
| "grad_norm": 0.9676224589347839, |
| "learning_rate": 2.888573456352023e-05, |
| "loss": 0.499, |
| "step": 2413 |
| }, |
| { |
| "epoch": 3.42, |
| "grad_norm": 1.027353286743164, |
| "learning_rate": 2.881476224272534e-05, |
| "loss": 0.5309, |
| "step": 2414 |
| }, |
| { |
| "epoch": 3.42, |
| "grad_norm": 0.9962853193283081, |
| "learning_rate": 2.8743789921930447e-05, |
| "loss": 0.5111, |
| "step": 2415 |
| }, |
| { |
| "epoch": 3.42, |
| "grad_norm": 0.9848844408988953, |
| "learning_rate": 2.867281760113556e-05, |
| "loss": 0.5067, |
| "step": 2416 |
| }, |
| { |
| "epoch": 3.42, |
| "grad_norm": 0.9875693321228027, |
| "learning_rate": 2.8601845280340666e-05, |
| "loss": 0.5029, |
| "step": 2417 |
| }, |
| { |
| "epoch": 3.43, |
| "grad_norm": 0.9913346171379089, |
| "learning_rate": 2.853087295954578e-05, |
| "loss": 0.5344, |
| "step": 2418 |
| }, |
| { |
| "epoch": 3.43, |
| "grad_norm": 0.9555786848068237, |
| "learning_rate": 2.8459900638750885e-05, |
| "loss": 0.4685, |
| "step": 2419 |
| }, |
| { |
| "epoch": 3.43, |
| "grad_norm": 1.03782057762146, |
| "learning_rate": 2.8388928317955997e-05, |
| "loss": 0.5803, |
| "step": 2420 |
| }, |
| { |
| "epoch": 3.43, |
| "grad_norm": 1.0548306703567505, |
| "learning_rate": 2.831795599716111e-05, |
| "loss": 0.5648, |
| "step": 2421 |
| }, |
| { |
| "epoch": 3.43, |
| "grad_norm": 1.0190414190292358, |
| "learning_rate": 2.824698367636622e-05, |
| "loss": 0.5182, |
| "step": 2422 |
| }, |
| { |
| "epoch": 3.43, |
| "grad_norm": 1.0644923448562622, |
| "learning_rate": 2.817601135557133e-05, |
| "loss": 0.5258, |
| "step": 2423 |
| }, |
| { |
| "epoch": 3.43, |
| "grad_norm": 1.0226292610168457, |
| "learning_rate": 2.8105039034776438e-05, |
| "loss": 0.5299, |
| "step": 2424 |
| }, |
| { |
| "epoch": 3.44, |
| "grad_norm": 1.0438193082809448, |
| "learning_rate": 2.803406671398155e-05, |
| "loss": 0.5171, |
| "step": 2425 |
| }, |
| { |
| "epoch": 3.44, |
| "grad_norm": 1.064306378364563, |
| "learning_rate": 2.7963094393186657e-05, |
| "loss": 0.5664, |
| "step": 2426 |
| }, |
| { |
| "epoch": 3.44, |
| "grad_norm": 0.9808685183525085, |
| "learning_rate": 2.789212207239177e-05, |
| "loss": 0.4799, |
| "step": 2427 |
| }, |
| { |
| "epoch": 3.44, |
| "grad_norm": 1.047814965248108, |
| "learning_rate": 2.7821149751596882e-05, |
| "loss": 0.5405, |
| "step": 2428 |
| }, |
| { |
| "epoch": 3.44, |
| "grad_norm": 0.992023229598999, |
| "learning_rate": 2.7750177430801988e-05, |
| "loss": 0.5017, |
| "step": 2429 |
| }, |
| { |
| "epoch": 3.44, |
| "grad_norm": 1.0322489738464355, |
| "learning_rate": 2.76792051100071e-05, |
| "loss": 0.5467, |
| "step": 2430 |
| }, |
| { |
| "epoch": 3.44, |
| "grad_norm": 0.9788199663162231, |
| "learning_rate": 2.7608232789212207e-05, |
| "loss": 0.4835, |
| "step": 2431 |
| }, |
| { |
| "epoch": 3.45, |
| "grad_norm": 1.0321805477142334, |
| "learning_rate": 2.753726046841732e-05, |
| "loss": 0.4991, |
| "step": 2432 |
| }, |
| { |
| "epoch": 3.45, |
| "grad_norm": 1.0222502946853638, |
| "learning_rate": 2.7466288147622425e-05, |
| "loss": 0.5144, |
| "step": 2433 |
| }, |
| { |
| "epoch": 3.45, |
| "grad_norm": 1.042504072189331, |
| "learning_rate": 2.7395315826827538e-05, |
| "loss": 0.5507, |
| "step": 2434 |
| }, |
| { |
| "epoch": 3.45, |
| "grad_norm": 1.035918116569519, |
| "learning_rate": 2.7324343506032644e-05, |
| "loss": 0.5504, |
| "step": 2435 |
| }, |
| { |
| "epoch": 3.45, |
| "grad_norm": 1.0212652683258057, |
| "learning_rate": 2.7253371185237757e-05, |
| "loss": 0.5284, |
| "step": 2436 |
| }, |
| { |
| "epoch": 3.45, |
| "grad_norm": 1.0973209142684937, |
| "learning_rate": 2.718239886444287e-05, |
| "loss": 0.5614, |
| "step": 2437 |
| }, |
| { |
| "epoch": 3.45, |
| "grad_norm": 0.9826099276542664, |
| "learning_rate": 2.711142654364798e-05, |
| "loss": 0.4764, |
| "step": 2438 |
| }, |
| { |
| "epoch": 3.46, |
| "grad_norm": 1.0115211009979248, |
| "learning_rate": 2.7040454222853088e-05, |
| "loss": 0.5199, |
| "step": 2439 |
| }, |
| { |
| "epoch": 3.46, |
| "grad_norm": 1.0676625967025757, |
| "learning_rate": 2.6969481902058198e-05, |
| "loss": 0.5502, |
| "step": 2440 |
| }, |
| { |
| "epoch": 3.46, |
| "grad_norm": 1.050166130065918, |
| "learning_rate": 2.689850958126331e-05, |
| "loss": 0.5669, |
| "step": 2441 |
| }, |
| { |
| "epoch": 3.46, |
| "grad_norm": 1.061370849609375, |
| "learning_rate": 2.6827537260468416e-05, |
| "loss": 0.5565, |
| "step": 2442 |
| }, |
| { |
| "epoch": 3.46, |
| "grad_norm": 1.0100698471069336, |
| "learning_rate": 2.675656493967353e-05, |
| "loss": 0.5359, |
| "step": 2443 |
| }, |
| { |
| "epoch": 3.46, |
| "grad_norm": 0.9985615015029907, |
| "learning_rate": 2.6685592618878642e-05, |
| "loss": 0.5417, |
| "step": 2444 |
| }, |
| { |
| "epoch": 3.46, |
| "grad_norm": 1.0049725770950317, |
| "learning_rate": 2.6614620298083748e-05, |
| "loss": 0.4995, |
| "step": 2445 |
| }, |
| { |
| "epoch": 3.47, |
| "grad_norm": 0.973824143409729, |
| "learning_rate": 2.654364797728886e-05, |
| "loss": 0.4978, |
| "step": 2446 |
| }, |
| { |
| "epoch": 3.47, |
| "grad_norm": 1.0502867698669434, |
| "learning_rate": 2.6472675656493966e-05, |
| "loss": 0.5296, |
| "step": 2447 |
| }, |
| { |
| "epoch": 3.47, |
| "grad_norm": 0.9943152070045471, |
| "learning_rate": 2.640170333569908e-05, |
| "loss": 0.5324, |
| "step": 2448 |
| }, |
| { |
| "epoch": 3.47, |
| "grad_norm": 1.0061664581298828, |
| "learning_rate": 2.6330731014904185e-05, |
| "loss": 0.4898, |
| "step": 2449 |
| }, |
| { |
| "epoch": 3.47, |
| "grad_norm": 0.9694445133209229, |
| "learning_rate": 2.6259758694109298e-05, |
| "loss": 0.4816, |
| "step": 2450 |
| }, |
| { |
| "epoch": 3.47, |
| "grad_norm": 1.0240046977996826, |
| "learning_rate": 2.618878637331441e-05, |
| "loss": 0.5127, |
| "step": 2451 |
| }, |
| { |
| "epoch": 3.47, |
| "grad_norm": 1.0239776372909546, |
| "learning_rate": 2.6117814052519516e-05, |
| "loss": 0.5422, |
| "step": 2452 |
| }, |
| { |
| "epoch": 3.48, |
| "grad_norm": 1.0031204223632812, |
| "learning_rate": 2.604684173172463e-05, |
| "loss": 0.5236, |
| "step": 2453 |
| }, |
| { |
| "epoch": 3.48, |
| "grad_norm": 0.9446493983268738, |
| "learning_rate": 2.597586941092974e-05, |
| "loss": 0.4619, |
| "step": 2454 |
| }, |
| { |
| "epoch": 3.48, |
| "grad_norm": 1.073948621749878, |
| "learning_rate": 2.590489709013485e-05, |
| "loss": 0.5362, |
| "step": 2455 |
| }, |
| { |
| "epoch": 3.48, |
| "grad_norm": 0.9938516616821289, |
| "learning_rate": 2.5833924769339957e-05, |
| "loss": 0.5062, |
| "step": 2456 |
| }, |
| { |
| "epoch": 3.48, |
| "grad_norm": 1.0002453327178955, |
| "learning_rate": 2.576295244854507e-05, |
| "loss": 0.5083, |
| "step": 2457 |
| }, |
| { |
| "epoch": 3.48, |
| "grad_norm": 1.0422013998031616, |
| "learning_rate": 2.5691980127750176e-05, |
| "loss": 0.5477, |
| "step": 2458 |
| }, |
| { |
| "epoch": 3.48, |
| "grad_norm": 0.9741196632385254, |
| "learning_rate": 2.562100780695529e-05, |
| "loss": 0.4608, |
| "step": 2459 |
| }, |
| { |
| "epoch": 3.49, |
| "grad_norm": 1.0457109212875366, |
| "learning_rate": 2.55500354861604e-05, |
| "loss": 0.5356, |
| "step": 2460 |
| }, |
| { |
| "epoch": 3.49, |
| "grad_norm": 1.0058279037475586, |
| "learning_rate": 2.5479063165365507e-05, |
| "loss": 0.5145, |
| "step": 2461 |
| }, |
| { |
| "epoch": 3.49, |
| "grad_norm": 0.999752402305603, |
| "learning_rate": 2.540809084457062e-05, |
| "loss": 0.5211, |
| "step": 2462 |
| }, |
| { |
| "epoch": 3.49, |
| "grad_norm": 1.0829323530197144, |
| "learning_rate": 2.5337118523775726e-05, |
| "loss": 0.5486, |
| "step": 2463 |
| }, |
| { |
| "epoch": 3.49, |
| "grad_norm": 1.0869967937469482, |
| "learning_rate": 2.526614620298084e-05, |
| "loss": 0.5673, |
| "step": 2464 |
| }, |
| { |
| "epoch": 3.49, |
| "grad_norm": 1.0342005491256714, |
| "learning_rate": 2.5195173882185945e-05, |
| "loss": 0.5199, |
| "step": 2465 |
| }, |
| { |
| "epoch": 3.49, |
| "grad_norm": 0.9886661767959595, |
| "learning_rate": 2.5124201561391057e-05, |
| "loss": 0.4892, |
| "step": 2466 |
| }, |
| { |
| "epoch": 3.49, |
| "grad_norm": 0.9562023878097534, |
| "learning_rate": 2.505322924059617e-05, |
| "loss": 0.4768, |
| "step": 2467 |
| }, |
| { |
| "epoch": 3.5, |
| "grad_norm": 1.0431307554244995, |
| "learning_rate": 2.498225691980128e-05, |
| "loss": 0.5173, |
| "step": 2468 |
| }, |
| { |
| "epoch": 3.5, |
| "grad_norm": 0.9886699914932251, |
| "learning_rate": 2.491128459900639e-05, |
| "loss": 0.5218, |
| "step": 2469 |
| }, |
| { |
| "epoch": 3.5, |
| "grad_norm": 1.0256837606430054, |
| "learning_rate": 2.48403122782115e-05, |
| "loss": 0.5232, |
| "step": 2470 |
| }, |
| { |
| "epoch": 3.5, |
| "grad_norm": 1.0431032180786133, |
| "learning_rate": 2.476933995741661e-05, |
| "loss": 0.5279, |
| "step": 2471 |
| }, |
| { |
| "epoch": 3.5, |
| "grad_norm": 0.9854641556739807, |
| "learning_rate": 2.469836763662172e-05, |
| "loss": 0.498, |
| "step": 2472 |
| }, |
| { |
| "epoch": 3.5, |
| "grad_norm": 1.0274698734283447, |
| "learning_rate": 2.462739531582683e-05, |
| "loss": 0.5412, |
| "step": 2473 |
| }, |
| { |
| "epoch": 3.5, |
| "grad_norm": 1.0287938117980957, |
| "learning_rate": 2.455642299503194e-05, |
| "loss": 0.5629, |
| "step": 2474 |
| }, |
| { |
| "epoch": 3.51, |
| "grad_norm": 0.9638839364051819, |
| "learning_rate": 2.4485450674237048e-05, |
| "loss": 0.5243, |
| "step": 2475 |
| }, |
| { |
| "epoch": 3.51, |
| "grad_norm": 1.0268467664718628, |
| "learning_rate": 2.4414478353442157e-05, |
| "loss": 0.5069, |
| "step": 2476 |
| }, |
| { |
| "epoch": 3.51, |
| "grad_norm": 1.0005607604980469, |
| "learning_rate": 2.4343506032647267e-05, |
| "loss": 0.5016, |
| "step": 2477 |
| }, |
| { |
| "epoch": 3.51, |
| "grad_norm": 1.0163365602493286, |
| "learning_rate": 2.427253371185238e-05, |
| "loss": 0.5177, |
| "step": 2478 |
| }, |
| { |
| "epoch": 3.51, |
| "grad_norm": 1.0533087253570557, |
| "learning_rate": 2.420156139105749e-05, |
| "loss": 0.551, |
| "step": 2479 |
| }, |
| { |
| "epoch": 3.51, |
| "grad_norm": 1.117048740386963, |
| "learning_rate": 2.4130589070262598e-05, |
| "loss": 0.6015, |
| "step": 2480 |
| }, |
| { |
| "epoch": 3.51, |
| "grad_norm": 1.0275087356567383, |
| "learning_rate": 2.4059616749467708e-05, |
| "loss": 0.5362, |
| "step": 2481 |
| }, |
| { |
| "epoch": 3.52, |
| "grad_norm": 1.0228657722473145, |
| "learning_rate": 2.3988644428672817e-05, |
| "loss": 0.5356, |
| "step": 2482 |
| }, |
| { |
| "epoch": 3.52, |
| "grad_norm": 1.01448655128479, |
| "learning_rate": 2.391767210787793e-05, |
| "loss": 0.5046, |
| "step": 2483 |
| }, |
| { |
| "epoch": 3.52, |
| "grad_norm": 0.9837957620620728, |
| "learning_rate": 2.384669978708304e-05, |
| "loss": 0.4721, |
| "step": 2484 |
| }, |
| { |
| "epoch": 3.52, |
| "grad_norm": 1.0101453065872192, |
| "learning_rate": 2.3775727466288148e-05, |
| "loss": 0.5056, |
| "step": 2485 |
| }, |
| { |
| "epoch": 3.52, |
| "grad_norm": 1.0054988861083984, |
| "learning_rate": 2.370475514549326e-05, |
| "loss": 0.5063, |
| "step": 2486 |
| }, |
| { |
| "epoch": 3.52, |
| "grad_norm": 1.026909351348877, |
| "learning_rate": 2.363378282469837e-05, |
| "loss": 0.5346, |
| "step": 2487 |
| }, |
| { |
| "epoch": 3.52, |
| "grad_norm": 1.0565681457519531, |
| "learning_rate": 2.356281050390348e-05, |
| "loss": 0.5563, |
| "step": 2488 |
| }, |
| { |
| "epoch": 3.53, |
| "grad_norm": 0.9695550203323364, |
| "learning_rate": 2.349183818310859e-05, |
| "loss": 0.4942, |
| "step": 2489 |
| }, |
| { |
| "epoch": 3.53, |
| "grad_norm": 1.0256010293960571, |
| "learning_rate": 2.34208658623137e-05, |
| "loss": 0.5254, |
| "step": 2490 |
| }, |
| { |
| "epoch": 3.53, |
| "grad_norm": 0.9935940504074097, |
| "learning_rate": 2.3349893541518808e-05, |
| "loss": 0.5267, |
| "step": 2491 |
| }, |
| { |
| "epoch": 3.53, |
| "grad_norm": 1.0092248916625977, |
| "learning_rate": 2.3278921220723917e-05, |
| "loss": 0.5296, |
| "step": 2492 |
| }, |
| { |
| "epoch": 3.53, |
| "grad_norm": 1.0321085453033447, |
| "learning_rate": 2.320794889992903e-05, |
| "loss": 0.525, |
| "step": 2493 |
| }, |
| { |
| "epoch": 3.53, |
| "grad_norm": 1.0207672119140625, |
| "learning_rate": 2.313697657913414e-05, |
| "loss": 0.5035, |
| "step": 2494 |
| }, |
| { |
| "epoch": 3.53, |
| "grad_norm": 1.0441876649856567, |
| "learning_rate": 2.306600425833925e-05, |
| "loss": 0.541, |
| "step": 2495 |
| }, |
| { |
| "epoch": 3.54, |
| "grad_norm": 1.0213267803192139, |
| "learning_rate": 2.2995031937544358e-05, |
| "loss": 0.5041, |
| "step": 2496 |
| }, |
| { |
| "epoch": 3.54, |
| "grad_norm": 1.019863247871399, |
| "learning_rate": 2.2924059616749467e-05, |
| "loss": 0.5105, |
| "step": 2497 |
| }, |
| { |
| "epoch": 3.54, |
| "grad_norm": 0.9917488098144531, |
| "learning_rate": 2.285308729595458e-05, |
| "loss": 0.5081, |
| "step": 2498 |
| }, |
| { |
| "epoch": 3.54, |
| "grad_norm": 0.9887694716453552, |
| "learning_rate": 2.278211497515969e-05, |
| "loss": 0.4958, |
| "step": 2499 |
| }, |
| { |
| "epoch": 3.54, |
| "grad_norm": 1.0222865343093872, |
| "learning_rate": 2.27111426543648e-05, |
| "loss": 0.5191, |
| "step": 2500 |
| } |
| ], |
| "logging_steps": 1, |
| "max_steps": 2820, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 4, |
| "save_steps": 500, |
| "total_flos": 5.0634694852608e+17, |
| "train_batch_size": 8, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|