diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,12747 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 4.989690721649485, + "eval_steps": 500, + "global_step": 1815, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0027491408934707906, + "grad_norm": 6.006607135034925, + "learning_rate": 4.395604395604396e-07, + "loss": 1.1009, + "step": 1 + }, + { + "epoch": 0.005498281786941581, + "grad_norm": 6.018597667128207, + "learning_rate": 8.791208791208792e-07, + "loss": 1.1005, + "step": 2 + }, + { + "epoch": 0.008247422680412371, + "grad_norm": 6.073361462181881, + "learning_rate": 1.3186813186813187e-06, + "loss": 1.1055, + "step": 3 + }, + { + "epoch": 0.010996563573883162, + "grad_norm": 5.961879860389406, + "learning_rate": 1.7582417582417585e-06, + "loss": 1.1099, + "step": 4 + }, + { + "epoch": 0.013745704467353952, + "grad_norm": 5.617774190916827, + "learning_rate": 2.197802197802198e-06, + "loss": 1.081, + "step": 5 + }, + { + "epoch": 0.016494845360824743, + "grad_norm": 4.531649230749474, + "learning_rate": 2.6373626373626375e-06, + "loss": 1.0487, + "step": 6 + }, + { + "epoch": 0.019243986254295534, + "grad_norm": 4.217358609239622, + "learning_rate": 3.0769230769230774e-06, + "loss": 1.0556, + "step": 7 + }, + { + "epoch": 0.021993127147766325, + "grad_norm": 2.3655606096533908, + "learning_rate": 3.516483516483517e-06, + "loss": 1.0049, + "step": 8 + }, + { + "epoch": 0.024742268041237112, + "grad_norm": 2.0449739951596695, + "learning_rate": 3.9560439560439565e-06, + "loss": 0.9803, + "step": 9 + }, + { + "epoch": 0.027491408934707903, + "grad_norm": 2.8045358100308406, + "learning_rate": 4.395604395604396e-06, + "loss": 0.9589, + "step": 10 + }, + { + "epoch": 0.030240549828178694, + "grad_norm": 3.662777544001548, + "learning_rate": 4.8351648351648355e-06, + "loss": 0.9677, + "step": 11 + }, + { + "epoch": 0.032989690721649485, + "grad_norm": 3.5905401528260374, + "learning_rate": 5.274725274725275e-06, + "loss": 0.96, + "step": 12 + }, + { + "epoch": 0.03573883161512027, + "grad_norm": 3.3209223305008964, + "learning_rate": 5.7142857142857145e-06, + "loss": 0.9481, + "step": 13 + }, + { + "epoch": 0.03848797250859107, + "grad_norm": 2.3880367341332343, + "learning_rate": 6.153846153846155e-06, + "loss": 0.9057, + "step": 14 + }, + { + "epoch": 0.041237113402061855, + "grad_norm": 2.269345734711135, + "learning_rate": 6.5934065934065935e-06, + "loss": 0.9048, + "step": 15 + }, + { + "epoch": 0.04398625429553265, + "grad_norm": 1.7351295951376182, + "learning_rate": 7.032967032967034e-06, + "loss": 0.8869, + "step": 16 + }, + { + "epoch": 0.04673539518900344, + "grad_norm": 1.5080912194950085, + "learning_rate": 7.472527472527473e-06, + "loss": 0.8503, + "step": 17 + }, + { + "epoch": 0.049484536082474224, + "grad_norm": 1.4431136736456789, + "learning_rate": 7.912087912087913e-06, + "loss": 0.8405, + "step": 18 + }, + { + "epoch": 0.05223367697594502, + "grad_norm": 1.4712422848105347, + "learning_rate": 8.351648351648353e-06, + "loss": 0.8274, + "step": 19 + }, + { + "epoch": 0.054982817869415807, + "grad_norm": 1.3178930536171367, + "learning_rate": 8.791208791208792e-06, + "loss": 0.8428, + "step": 20 + }, + { + "epoch": 0.0577319587628866, + "grad_norm": 1.2703340647552794, + "learning_rate": 9.230769230769232e-06, + "loss": 0.8211, + "step": 21 + }, + { + "epoch": 0.06048109965635739, + "grad_norm": 1.2871317510857054, + "learning_rate": 9.670329670329671e-06, + "loss": 0.8053, + "step": 22 + }, + { + "epoch": 0.06323024054982818, + "grad_norm": 1.1237165728026322, + "learning_rate": 1.010989010989011e-05, + "loss": 0.8101, + "step": 23 + }, + { + "epoch": 0.06597938144329897, + "grad_norm": 0.9168689840127058, + "learning_rate": 1.054945054945055e-05, + "loss": 0.7837, + "step": 24 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 0.9355697137818322, + "learning_rate": 1.098901098901099e-05, + "loss": 0.79, + "step": 25 + }, + { + "epoch": 0.07147766323024055, + "grad_norm": 0.8919194529855764, + "learning_rate": 1.1428571428571429e-05, + "loss": 0.7927, + "step": 26 + }, + { + "epoch": 0.07422680412371134, + "grad_norm": 0.8013095068191387, + "learning_rate": 1.186813186813187e-05, + "loss": 0.7826, + "step": 27 + }, + { + "epoch": 0.07697594501718213, + "grad_norm": 0.7074089625670721, + "learning_rate": 1.230769230769231e-05, + "loss": 0.77, + "step": 28 + }, + { + "epoch": 0.07972508591065292, + "grad_norm": 0.5939851673949501, + "learning_rate": 1.2747252747252747e-05, + "loss": 0.764, + "step": 29 + }, + { + "epoch": 0.08247422680412371, + "grad_norm": 0.7291769706418423, + "learning_rate": 1.3186813186813187e-05, + "loss": 0.7667, + "step": 30 + }, + { + "epoch": 0.0852233676975945, + "grad_norm": 0.5476356706171729, + "learning_rate": 1.3626373626373627e-05, + "loss": 0.7669, + "step": 31 + }, + { + "epoch": 0.0879725085910653, + "grad_norm": 0.496258207605204, + "learning_rate": 1.4065934065934068e-05, + "loss": 0.7617, + "step": 32 + }, + { + "epoch": 0.09072164948453608, + "grad_norm": 0.5382769338196303, + "learning_rate": 1.4505494505494506e-05, + "loss": 0.7554, + "step": 33 + }, + { + "epoch": 0.09347079037800687, + "grad_norm": 0.4580628373124881, + "learning_rate": 1.4945054945054947e-05, + "loss": 0.75, + "step": 34 + }, + { + "epoch": 0.09621993127147767, + "grad_norm": 0.40847997114284523, + "learning_rate": 1.5384615384615387e-05, + "loss": 0.7467, + "step": 35 + }, + { + "epoch": 0.09896907216494845, + "grad_norm": 0.5031600722774845, + "learning_rate": 1.5824175824175826e-05, + "loss": 0.7383, + "step": 36 + }, + { + "epoch": 0.10171821305841924, + "grad_norm": 0.3529000118564782, + "learning_rate": 1.6263736263736265e-05, + "loss": 0.7524, + "step": 37 + }, + { + "epoch": 0.10446735395189004, + "grad_norm": 0.40270165373486655, + "learning_rate": 1.6703296703296707e-05, + "loss": 0.7371, + "step": 38 + }, + { + "epoch": 0.10721649484536082, + "grad_norm": 0.3743730526645748, + "learning_rate": 1.7142857142857142e-05, + "loss": 0.7332, + "step": 39 + }, + { + "epoch": 0.10996563573883161, + "grad_norm": 0.3802324817797124, + "learning_rate": 1.7582417582417584e-05, + "loss": 0.7301, + "step": 40 + }, + { + "epoch": 0.11271477663230241, + "grad_norm": 0.38122013544181327, + "learning_rate": 1.8021978021978023e-05, + "loss": 0.7481, + "step": 41 + }, + { + "epoch": 0.1154639175257732, + "grad_norm": 0.3807629715206352, + "learning_rate": 1.8461538461538465e-05, + "loss": 0.7234, + "step": 42 + }, + { + "epoch": 0.11821305841924398, + "grad_norm": 0.3690276896533298, + "learning_rate": 1.8901098901098903e-05, + "loss": 0.7382, + "step": 43 + }, + { + "epoch": 0.12096219931271478, + "grad_norm": 0.39354181368093777, + "learning_rate": 1.9340659340659342e-05, + "loss": 0.7274, + "step": 44 + }, + { + "epoch": 0.12371134020618557, + "grad_norm": 0.45930210506482494, + "learning_rate": 1.9780219780219784e-05, + "loss": 0.7222, + "step": 45 + }, + { + "epoch": 0.12646048109965635, + "grad_norm": 0.3257242749282543, + "learning_rate": 2.021978021978022e-05, + "loss": 0.7062, + "step": 46 + }, + { + "epoch": 0.12920962199312716, + "grad_norm": 0.4894322683246899, + "learning_rate": 2.0659340659340665e-05, + "loss": 0.7273, + "step": 47 + }, + { + "epoch": 0.13195876288659794, + "grad_norm": 0.4217505425727167, + "learning_rate": 2.10989010989011e-05, + "loss": 0.7144, + "step": 48 + }, + { + "epoch": 0.13470790378006872, + "grad_norm": 0.34926375879792276, + "learning_rate": 2.153846153846154e-05, + "loss": 0.7284, + "step": 49 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 0.37706775972489026, + "learning_rate": 2.197802197802198e-05, + "loss": 0.7197, + "step": 50 + }, + { + "epoch": 0.1402061855670103, + "grad_norm": 0.4461248020784725, + "learning_rate": 2.241758241758242e-05, + "loss": 0.7168, + "step": 51 + }, + { + "epoch": 0.1429553264604811, + "grad_norm": 0.45683903751388155, + "learning_rate": 2.2857142857142858e-05, + "loss": 0.7047, + "step": 52 + }, + { + "epoch": 0.1457044673539519, + "grad_norm": 0.4941785029111688, + "learning_rate": 2.32967032967033e-05, + "loss": 0.7203, + "step": 53 + }, + { + "epoch": 0.14845360824742268, + "grad_norm": 0.7101247658482969, + "learning_rate": 2.373626373626374e-05, + "loss": 0.7132, + "step": 54 + }, + { + "epoch": 0.15120274914089346, + "grad_norm": 1.3845226821923962, + "learning_rate": 2.4175824175824177e-05, + "loss": 0.7139, + "step": 55 + }, + { + "epoch": 0.15395189003436427, + "grad_norm": 0.9145522126443197, + "learning_rate": 2.461538461538462e-05, + "loss": 0.7201, + "step": 56 + }, + { + "epoch": 0.15670103092783505, + "grad_norm": 0.7316554124403013, + "learning_rate": 2.5054945054945058e-05, + "loss": 0.6997, + "step": 57 + }, + { + "epoch": 0.15945017182130583, + "grad_norm": 1.525604886016204, + "learning_rate": 2.5494505494505493e-05, + "loss": 0.708, + "step": 58 + }, + { + "epoch": 0.16219931271477664, + "grad_norm": 0.9569983399077112, + "learning_rate": 2.593406593406594e-05, + "loss": 0.7023, + "step": 59 + }, + { + "epoch": 0.16494845360824742, + "grad_norm": 1.222421966279637, + "learning_rate": 2.6373626373626374e-05, + "loss": 0.699, + "step": 60 + }, + { + "epoch": 0.1676975945017182, + "grad_norm": 1.1284003579139785, + "learning_rate": 2.6813186813186813e-05, + "loss": 0.6944, + "step": 61 + }, + { + "epoch": 0.170446735395189, + "grad_norm": 1.323080634931126, + "learning_rate": 2.7252747252747255e-05, + "loss": 0.6897, + "step": 62 + }, + { + "epoch": 0.1731958762886598, + "grad_norm": 0.9918797937768813, + "learning_rate": 2.7692307692307694e-05, + "loss": 0.6937, + "step": 63 + }, + { + "epoch": 0.1759450171821306, + "grad_norm": 1.4485583433380003, + "learning_rate": 2.8131868131868136e-05, + "loss": 0.6924, + "step": 64 + }, + { + "epoch": 0.17869415807560138, + "grad_norm": 0.9586990948210379, + "learning_rate": 2.8571428571428574e-05, + "loss": 0.7036, + "step": 65 + }, + { + "epoch": 0.18144329896907216, + "grad_norm": 1.5401125351907554, + "learning_rate": 2.9010989010989013e-05, + "loss": 0.7067, + "step": 66 + }, + { + "epoch": 0.18419243986254297, + "grad_norm": 1.149941673251316, + "learning_rate": 2.9450549450549455e-05, + "loss": 0.6846, + "step": 67 + }, + { + "epoch": 0.18694158075601375, + "grad_norm": 1.398728105276103, + "learning_rate": 2.9890109890109894e-05, + "loss": 0.7001, + "step": 68 + }, + { + "epoch": 0.18969072164948453, + "grad_norm": 1.3380685614311483, + "learning_rate": 3.0329670329670332e-05, + "loss": 0.7017, + "step": 69 + }, + { + "epoch": 0.19243986254295534, + "grad_norm": 0.966967948159332, + "learning_rate": 3.0769230769230774e-05, + "loss": 0.6945, + "step": 70 + }, + { + "epoch": 0.19518900343642612, + "grad_norm": 1.2949156130535089, + "learning_rate": 3.120879120879121e-05, + "loss": 0.6983, + "step": 71 + }, + { + "epoch": 0.1979381443298969, + "grad_norm": 0.9121501279786434, + "learning_rate": 3.164835164835165e-05, + "loss": 0.6944, + "step": 72 + }, + { + "epoch": 0.2006872852233677, + "grad_norm": 0.7744380611271706, + "learning_rate": 3.2087912087912094e-05, + "loss": 0.6797, + "step": 73 + }, + { + "epoch": 0.2034364261168385, + "grad_norm": 1.171893254170078, + "learning_rate": 3.252747252747253e-05, + "loss": 0.6916, + "step": 74 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 1.432067598703686, + "learning_rate": 3.296703296703297e-05, + "loss": 0.696, + "step": 75 + }, + { + "epoch": 0.20893470790378008, + "grad_norm": 2.015800500571526, + "learning_rate": 3.340659340659341e-05, + "loss": 0.6946, + "step": 76 + }, + { + "epoch": 0.21168384879725086, + "grad_norm": 1.1450267893677495, + "learning_rate": 3.384615384615385e-05, + "loss": 0.6938, + "step": 77 + }, + { + "epoch": 0.21443298969072164, + "grad_norm": 3.012444154989314, + "learning_rate": 3.4285714285714284e-05, + "loss": 0.708, + "step": 78 + }, + { + "epoch": 0.21718213058419245, + "grad_norm": 2.938842593069413, + "learning_rate": 3.4725274725274726e-05, + "loss": 0.7043, + "step": 79 + }, + { + "epoch": 0.21993127147766323, + "grad_norm": 1.2620210326118846, + "learning_rate": 3.516483516483517e-05, + "loss": 0.7015, + "step": 80 + }, + { + "epoch": 0.22268041237113403, + "grad_norm": 1.8136600496263213, + "learning_rate": 3.56043956043956e-05, + "loss": 0.6934, + "step": 81 + }, + { + "epoch": 0.22542955326460482, + "grad_norm": 1.4037155277378568, + "learning_rate": 3.6043956043956045e-05, + "loss": 0.6923, + "step": 82 + }, + { + "epoch": 0.2281786941580756, + "grad_norm": 1.3955889629454836, + "learning_rate": 3.648351648351649e-05, + "loss": 0.6932, + "step": 83 + }, + { + "epoch": 0.2309278350515464, + "grad_norm": 1.4401203218805492, + "learning_rate": 3.692307692307693e-05, + "loss": 0.6837, + "step": 84 + }, + { + "epoch": 0.23367697594501718, + "grad_norm": 1.1506286717804464, + "learning_rate": 3.7362637362637365e-05, + "loss": 0.6824, + "step": 85 + }, + { + "epoch": 0.23642611683848797, + "grad_norm": 1.4015178509079977, + "learning_rate": 3.7802197802197807e-05, + "loss": 0.6818, + "step": 86 + }, + { + "epoch": 0.23917525773195877, + "grad_norm": 0.8632967007668437, + "learning_rate": 3.824175824175825e-05, + "loss": 0.6881, + "step": 87 + }, + { + "epoch": 0.24192439862542955, + "grad_norm": 1.7108832289817475, + "learning_rate": 3.8681318681318684e-05, + "loss": 0.694, + "step": 88 + }, + { + "epoch": 0.24467353951890033, + "grad_norm": 1.2330487896649007, + "learning_rate": 3.9120879120879126e-05, + "loss": 0.682, + "step": 89 + }, + { + "epoch": 0.24742268041237114, + "grad_norm": 1.2807204902880718, + "learning_rate": 3.956043956043957e-05, + "loss": 0.6851, + "step": 90 + }, + { + "epoch": 0.2501718213058419, + "grad_norm": 1.327068957190879, + "learning_rate": 4e-05, + "loss": 0.6807, + "step": 91 + }, + { + "epoch": 0.2529209621993127, + "grad_norm": 1.0022125576592875, + "learning_rate": 4.043956043956044e-05, + "loss": 0.6782, + "step": 92 + }, + { + "epoch": 0.2556701030927835, + "grad_norm": 1.3221553663136452, + "learning_rate": 4.087912087912088e-05, + "loss": 0.6759, + "step": 93 + }, + { + "epoch": 0.2584192439862543, + "grad_norm": 0.9568766990099716, + "learning_rate": 4.131868131868133e-05, + "loss": 0.6873, + "step": 94 + }, + { + "epoch": 0.2611683848797251, + "grad_norm": 1.7172395359463215, + "learning_rate": 4.1758241758241765e-05, + "loss": 0.6814, + "step": 95 + }, + { + "epoch": 0.2639175257731959, + "grad_norm": 1.5142722720152748, + "learning_rate": 4.21978021978022e-05, + "loss": 0.682, + "step": 96 + }, + { + "epoch": 0.26666666666666666, + "grad_norm": 0.8979115368433407, + "learning_rate": 4.263736263736264e-05, + "loss": 0.6828, + "step": 97 + }, + { + "epoch": 0.26941580756013744, + "grad_norm": 1.3321313501935168, + "learning_rate": 4.307692307692308e-05, + "loss": 0.687, + "step": 98 + }, + { + "epoch": 0.2721649484536082, + "grad_norm": 1.6997841277714232, + "learning_rate": 4.351648351648352e-05, + "loss": 0.6759, + "step": 99 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.0580779198597339, + "learning_rate": 4.395604395604396e-05, + "loss": 0.6743, + "step": 100 + }, + { + "epoch": 0.27766323024054984, + "grad_norm": 2.3143610998320896, + "learning_rate": 4.4395604395604403e-05, + "loss": 0.6684, + "step": 101 + }, + { + "epoch": 0.2804123711340206, + "grad_norm": 1.370198490883026, + "learning_rate": 4.483516483516484e-05, + "loss": 0.6643, + "step": 102 + }, + { + "epoch": 0.2831615120274914, + "grad_norm": 2.7976008957802088, + "learning_rate": 4.527472527472528e-05, + "loss": 0.6878, + "step": 103 + }, + { + "epoch": 0.2859106529209622, + "grad_norm": 2.539570133020213, + "learning_rate": 4.5714285714285716e-05, + "loss": 0.6889, + "step": 104 + }, + { + "epoch": 0.28865979381443296, + "grad_norm": 1.7701837403812288, + "learning_rate": 4.615384615384615e-05, + "loss": 0.6794, + "step": 105 + }, + { + "epoch": 0.2914089347079038, + "grad_norm": 1.6244664002638955, + "learning_rate": 4.65934065934066e-05, + "loss": 0.6798, + "step": 106 + }, + { + "epoch": 0.2941580756013746, + "grad_norm": 1.658296656226508, + "learning_rate": 4.7032967032967035e-05, + "loss": 0.6852, + "step": 107 + }, + { + "epoch": 0.29690721649484536, + "grad_norm": 1.3133889484198134, + "learning_rate": 4.747252747252748e-05, + "loss": 0.6695, + "step": 108 + }, + { + "epoch": 0.29965635738831614, + "grad_norm": 1.5315701499343, + "learning_rate": 4.791208791208791e-05, + "loss": 0.6842, + "step": 109 + }, + { + "epoch": 0.3024054982817869, + "grad_norm": 1.5106029998832964, + "learning_rate": 4.8351648351648355e-05, + "loss": 0.6728, + "step": 110 + }, + { + "epoch": 0.30515463917525776, + "grad_norm": 1.2198017486153612, + "learning_rate": 4.87912087912088e-05, + "loss": 0.6804, + "step": 111 + }, + { + "epoch": 0.30790378006872854, + "grad_norm": 0.9408294254788567, + "learning_rate": 4.923076923076924e-05, + "loss": 0.6673, + "step": 112 + }, + { + "epoch": 0.3106529209621993, + "grad_norm": 1.4299436206183789, + "learning_rate": 4.9670329670329674e-05, + "loss": 0.6782, + "step": 113 + }, + { + "epoch": 0.3134020618556701, + "grad_norm": 1.3010524254683744, + "learning_rate": 5.0109890109890116e-05, + "loss": 0.6719, + "step": 114 + }, + { + "epoch": 0.3161512027491409, + "grad_norm": 1.0294736794964319, + "learning_rate": 5.054945054945055e-05, + "loss": 0.6734, + "step": 115 + }, + { + "epoch": 0.31890034364261166, + "grad_norm": 1.9844841183716888, + "learning_rate": 5.098901098901099e-05, + "loss": 0.6689, + "step": 116 + }, + { + "epoch": 0.3216494845360825, + "grad_norm": 1.0769609751269622, + "learning_rate": 5.1428571428571436e-05, + "loss": 0.6725, + "step": 117 + }, + { + "epoch": 0.3243986254295533, + "grad_norm": 1.9412353253336612, + "learning_rate": 5.186813186813188e-05, + "loss": 0.6788, + "step": 118 + }, + { + "epoch": 0.32714776632302406, + "grad_norm": 1.2983006823972247, + "learning_rate": 5.230769230769231e-05, + "loss": 0.6717, + "step": 119 + }, + { + "epoch": 0.32989690721649484, + "grad_norm": 1.78349719868111, + "learning_rate": 5.274725274725275e-05, + "loss": 0.6754, + "step": 120 + }, + { + "epoch": 0.3326460481099656, + "grad_norm": 1.2685409855209113, + "learning_rate": 5.318681318681319e-05, + "loss": 0.6689, + "step": 121 + }, + { + "epoch": 0.3353951890034364, + "grad_norm": 1.601626046067794, + "learning_rate": 5.3626373626373626e-05, + "loss": 0.6822, + "step": 122 + }, + { + "epoch": 0.33814432989690724, + "grad_norm": 1.6399160217312283, + "learning_rate": 5.4065934065934074e-05, + "loss": 0.6939, + "step": 123 + }, + { + "epoch": 0.340893470790378, + "grad_norm": 1.0496938208520434, + "learning_rate": 5.450549450549451e-05, + "loss": 0.6685, + "step": 124 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 1.8210641674833647, + "learning_rate": 5.494505494505495e-05, + "loss": 0.6726, + "step": 125 + }, + { + "epoch": 0.3463917525773196, + "grad_norm": 1.0994677896279192, + "learning_rate": 5.538461538461539e-05, + "loss": 0.6609, + "step": 126 + }, + { + "epoch": 0.34914089347079036, + "grad_norm": 1.7375889267050222, + "learning_rate": 5.582417582417583e-05, + "loss": 0.68, + "step": 127 + }, + { + "epoch": 0.3518900343642612, + "grad_norm": 1.1687845090860516, + "learning_rate": 5.626373626373627e-05, + "loss": 0.6715, + "step": 128 + }, + { + "epoch": 0.354639175257732, + "grad_norm": 1.395445256796641, + "learning_rate": 5.670329670329671e-05, + "loss": 0.6681, + "step": 129 + }, + { + "epoch": 0.35738831615120276, + "grad_norm": 1.355862476688157, + "learning_rate": 5.714285714285715e-05, + "loss": 0.6818, + "step": 130 + }, + { + "epoch": 0.36013745704467354, + "grad_norm": 1.0924489864173972, + "learning_rate": 5.7582417582417584e-05, + "loss": 0.6605, + "step": 131 + }, + { + "epoch": 0.3628865979381443, + "grad_norm": 1.9384854103627827, + "learning_rate": 5.8021978021978026e-05, + "loss": 0.6731, + "step": 132 + }, + { + "epoch": 0.3656357388316151, + "grad_norm": 1.5529834412680505, + "learning_rate": 5.846153846153846e-05, + "loss": 0.675, + "step": 133 + }, + { + "epoch": 0.36838487972508593, + "grad_norm": 0.9671293876862966, + "learning_rate": 5.890109890109891e-05, + "loss": 0.661, + "step": 134 + }, + { + "epoch": 0.3711340206185567, + "grad_norm": 1.5434547051402783, + "learning_rate": 5.9340659340659345e-05, + "loss": 0.673, + "step": 135 + }, + { + "epoch": 0.3738831615120275, + "grad_norm": 1.466260971358487, + "learning_rate": 5.978021978021979e-05, + "loss": 0.6799, + "step": 136 + }, + { + "epoch": 0.3766323024054983, + "grad_norm": 1.3025182971790092, + "learning_rate": 6.021978021978022e-05, + "loss": 0.656, + "step": 137 + }, + { + "epoch": 0.37938144329896906, + "grad_norm": 1.5407519214603846, + "learning_rate": 6.0659340659340665e-05, + "loss": 0.6703, + "step": 138 + }, + { + "epoch": 0.38213058419243984, + "grad_norm": 1.2435864649597408, + "learning_rate": 6.10989010989011e-05, + "loss": 0.6699, + "step": 139 + }, + { + "epoch": 0.3848797250859107, + "grad_norm": 1.32415071563785, + "learning_rate": 6.153846153846155e-05, + "loss": 0.6557, + "step": 140 + }, + { + "epoch": 0.38762886597938145, + "grad_norm": 1.5121087379761295, + "learning_rate": 6.197802197802199e-05, + "loss": 0.6659, + "step": 141 + }, + { + "epoch": 0.39037800687285223, + "grad_norm": 1.3570253258903069, + "learning_rate": 6.241758241758242e-05, + "loss": 0.6568, + "step": 142 + }, + { + "epoch": 0.393127147766323, + "grad_norm": 1.2775583571637077, + "learning_rate": 6.285714285714286e-05, + "loss": 0.6558, + "step": 143 + }, + { + "epoch": 0.3958762886597938, + "grad_norm": 1.029079383256909, + "learning_rate": 6.32967032967033e-05, + "loss": 0.6569, + "step": 144 + }, + { + "epoch": 0.39862542955326463, + "grad_norm": 1.379942185445747, + "learning_rate": 6.373626373626373e-05, + "loss": 0.6664, + "step": 145 + }, + { + "epoch": 0.4013745704467354, + "grad_norm": 1.7317621636752105, + "learning_rate": 6.417582417582419e-05, + "loss": 0.6675, + "step": 146 + }, + { + "epoch": 0.4041237113402062, + "grad_norm": 1.0553856530228192, + "learning_rate": 6.461538461538462e-05, + "loss": 0.6623, + "step": 147 + }, + { + "epoch": 0.406872852233677, + "grad_norm": 1.310533202428338, + "learning_rate": 6.505494505494506e-05, + "loss": 0.6658, + "step": 148 + }, + { + "epoch": 0.40962199312714775, + "grad_norm": 1.0365623836624396, + "learning_rate": 6.54945054945055e-05, + "loss": 0.6539, + "step": 149 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 1.5824321098093788, + "learning_rate": 6.593406593406594e-05, + "loss": 0.672, + "step": 150 + }, + { + "epoch": 0.41512027491408937, + "grad_norm": 1.2342253922087731, + "learning_rate": 6.637362637362638e-05, + "loss": 0.6609, + "step": 151 + }, + { + "epoch": 0.41786941580756015, + "grad_norm": 2.2157482513280797, + "learning_rate": 6.681318681318683e-05, + "loss": 0.6614, + "step": 152 + }, + { + "epoch": 0.42061855670103093, + "grad_norm": 0.8519496697821355, + "learning_rate": 6.725274725274725e-05, + "loss": 0.6558, + "step": 153 + }, + { + "epoch": 0.4233676975945017, + "grad_norm": 2.585667143290517, + "learning_rate": 6.76923076923077e-05, + "loss": 0.6712, + "step": 154 + }, + { + "epoch": 0.4261168384879725, + "grad_norm": 1.561949044082342, + "learning_rate": 6.813186813186814e-05, + "loss": 0.6606, + "step": 155 + }, + { + "epoch": 0.4288659793814433, + "grad_norm": 1.6541712978836747, + "learning_rate": 6.857142857142857e-05, + "loss": 0.6662, + "step": 156 + }, + { + "epoch": 0.4316151202749141, + "grad_norm": 1.471980620525164, + "learning_rate": 6.901098901098902e-05, + "loss": 0.6688, + "step": 157 + }, + { + "epoch": 0.4343642611683849, + "grad_norm": 1.4565326176621909, + "learning_rate": 6.945054945054945e-05, + "loss": 0.668, + "step": 158 + }, + { + "epoch": 0.43711340206185567, + "grad_norm": 1.3548150534245453, + "learning_rate": 6.98901098901099e-05, + "loss": 0.6628, + "step": 159 + }, + { + "epoch": 0.43986254295532645, + "grad_norm": 1.2045551451888672, + "learning_rate": 7.032967032967034e-05, + "loss": 0.6616, + "step": 160 + }, + { + "epoch": 0.44261168384879723, + "grad_norm": 1.7494121190993142, + "learning_rate": 7.076923076923078e-05, + "loss": 0.6581, + "step": 161 + }, + { + "epoch": 0.44536082474226807, + "grad_norm": 1.0147994550817765, + "learning_rate": 7.12087912087912e-05, + "loss": 0.6608, + "step": 162 + }, + { + "epoch": 0.44810996563573885, + "grad_norm": 1.0859192635780746, + "learning_rate": 7.164835164835166e-05, + "loss": 0.6451, + "step": 163 + }, + { + "epoch": 0.45085910652920963, + "grad_norm": 1.0492942035079653, + "learning_rate": 7.208791208791209e-05, + "loss": 0.6591, + "step": 164 + }, + { + "epoch": 0.4536082474226804, + "grad_norm": 2.379607115922538, + "learning_rate": 7.252747252747253e-05, + "loss": 0.6634, + "step": 165 + }, + { + "epoch": 0.4563573883161512, + "grad_norm": 1.112381416151378, + "learning_rate": 7.296703296703297e-05, + "loss": 0.6611, + "step": 166 + }, + { + "epoch": 0.45910652920962197, + "grad_norm": 2.0850617091598886, + "learning_rate": 7.34065934065934e-05, + "loss": 0.6569, + "step": 167 + }, + { + "epoch": 0.4618556701030928, + "grad_norm": 1.5623269496352417, + "learning_rate": 7.384615384615386e-05, + "loss": 0.672, + "step": 168 + }, + { + "epoch": 0.4646048109965636, + "grad_norm": 1.6783665454838335, + "learning_rate": 7.42857142857143e-05, + "loss": 0.6567, + "step": 169 + }, + { + "epoch": 0.46735395189003437, + "grad_norm": 1.0376996339879248, + "learning_rate": 7.472527472527473e-05, + "loss": 0.6606, + "step": 170 + }, + { + "epoch": 0.47010309278350515, + "grad_norm": 2.2216262071159356, + "learning_rate": 7.516483516483517e-05, + "loss": 0.6636, + "step": 171 + }, + { + "epoch": 0.47285223367697593, + "grad_norm": 1.3858824784808275, + "learning_rate": 7.560439560439561e-05, + "loss": 0.6673, + "step": 172 + }, + { + "epoch": 0.4756013745704467, + "grad_norm": 1.9605283100011852, + "learning_rate": 7.604395604395604e-05, + "loss": 0.6572, + "step": 173 + }, + { + "epoch": 0.47835051546391755, + "grad_norm": 1.4278083885748436, + "learning_rate": 7.64835164835165e-05, + "loss": 0.6645, + "step": 174 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.191754742216612, + "learning_rate": 7.692307692307693e-05, + "loss": 0.6631, + "step": 175 + }, + { + "epoch": 0.4838487972508591, + "grad_norm": 0.9597081578943443, + "learning_rate": 7.736263736263737e-05, + "loss": 0.655, + "step": 176 + }, + { + "epoch": 0.4865979381443299, + "grad_norm": 1.9487418698749246, + "learning_rate": 7.780219780219781e-05, + "loss": 0.6617, + "step": 177 + }, + { + "epoch": 0.48934707903780067, + "grad_norm": 1.4578651378155227, + "learning_rate": 7.824175824175825e-05, + "loss": 0.665, + "step": 178 + }, + { + "epoch": 0.49209621993127145, + "grad_norm": 0.9204633011470749, + "learning_rate": 7.868131868131868e-05, + "loss": 0.6559, + "step": 179 + }, + { + "epoch": 0.4948453608247423, + "grad_norm": 2.3519272780997893, + "learning_rate": 7.912087912087914e-05, + "loss": 0.661, + "step": 180 + }, + { + "epoch": 0.49759450171821307, + "grad_norm": 1.4425041211579737, + "learning_rate": 7.956043956043956e-05, + "loss": 0.6555, + "step": 181 + }, + { + "epoch": 0.5003436426116838, + "grad_norm": 2.2406732058079917, + "learning_rate": 8e-05, + "loss": 0.6672, + "step": 182 + }, + { + "epoch": 0.5030927835051546, + "grad_norm": 1.717445277482838, + "learning_rate": 7.999992597860977e-05, + "loss": 0.669, + "step": 183 + }, + { + "epoch": 0.5058419243986254, + "grad_norm": 1.9554853415214135, + "learning_rate": 7.999970391471297e-05, + "loss": 0.6484, + "step": 184 + }, + { + "epoch": 0.5085910652920962, + "grad_norm": 1.3337471142504986, + "learning_rate": 7.99993338091315e-05, + "loss": 0.6542, + "step": 185 + }, + { + "epoch": 0.511340206185567, + "grad_norm": 2.3628021636514913, + "learning_rate": 7.999881566323518e-05, + "loss": 0.6666, + "step": 186 + }, + { + "epoch": 0.5140893470790378, + "grad_norm": 1.6284990281451381, + "learning_rate": 7.999814947894166e-05, + "loss": 0.6713, + "step": 187 + }, + { + "epoch": 0.5168384879725086, + "grad_norm": 2.208780281202899, + "learning_rate": 7.999733525871655e-05, + "loss": 0.6564, + "step": 188 + }, + { + "epoch": 0.5195876288659794, + "grad_norm": 2.0324845611311653, + "learning_rate": 7.999637300557334e-05, + "loss": 0.6643, + "step": 189 + }, + { + "epoch": 0.5223367697594502, + "grad_norm": 1.411207407439057, + "learning_rate": 7.999526272307338e-05, + "loss": 0.6466, + "step": 190 + }, + { + "epoch": 0.525085910652921, + "grad_norm": 1.67795112571074, + "learning_rate": 7.999400441532593e-05, + "loss": 0.6567, + "step": 191 + }, + { + "epoch": 0.5278350515463918, + "grad_norm": 1.0171507403303963, + "learning_rate": 7.999259808698805e-05, + "loss": 0.6535, + "step": 192 + }, + { + "epoch": 0.5305841924398625, + "grad_norm": 1.8103499540778905, + "learning_rate": 7.999104374326465e-05, + "loss": 0.6546, + "step": 193 + }, + { + "epoch": 0.5333333333333333, + "grad_norm": 1.3619232333701012, + "learning_rate": 7.99893413899085e-05, + "loss": 0.6495, + "step": 194 + }, + { + "epoch": 0.5360824742268041, + "grad_norm": 1.7537074766579162, + "learning_rate": 7.99874910332201e-05, + "loss": 0.6513, + "step": 195 + }, + { + "epoch": 0.5388316151202749, + "grad_norm": 1.5125359865682328, + "learning_rate": 7.998549268004776e-05, + "loss": 0.6627, + "step": 196 + }, + { + "epoch": 0.5415807560137457, + "grad_norm": 1.8694904140870736, + "learning_rate": 7.998334633778752e-05, + "loss": 0.6552, + "step": 197 + }, + { + "epoch": 0.5443298969072164, + "grad_norm": 1.0313289192419033, + "learning_rate": 7.998105201438315e-05, + "loss": 0.6572, + "step": 198 + }, + { + "epoch": 0.5470790378006873, + "grad_norm": 1.7524234342414688, + "learning_rate": 7.997860971832609e-05, + "loss": 0.6485, + "step": 199 + }, + { + "epoch": 0.5498281786941581, + "grad_norm": 1.674466921847352, + "learning_rate": 7.997601945865545e-05, + "loss": 0.6567, + "step": 200 + }, + { + "epoch": 0.5525773195876289, + "grad_norm": 0.9044207656617541, + "learning_rate": 7.997328124495797e-05, + "loss": 0.6571, + "step": 201 + }, + { + "epoch": 0.5553264604810997, + "grad_norm": 1.5698427976284253, + "learning_rate": 7.997039508736794e-05, + "loss": 0.6552, + "step": 202 + }, + { + "epoch": 0.5580756013745705, + "grad_norm": 1.2149289312154958, + "learning_rate": 7.996736099656728e-05, + "loss": 0.6504, + "step": 203 + }, + { + "epoch": 0.5608247422680412, + "grad_norm": 1.5016916008686747, + "learning_rate": 7.996417898378532e-05, + "loss": 0.6501, + "step": 204 + }, + { + "epoch": 0.563573883161512, + "grad_norm": 1.3674415360691992, + "learning_rate": 7.996084906079895e-05, + "loss": 0.6521, + "step": 205 + }, + { + "epoch": 0.5663230240549828, + "grad_norm": 1.012486835623592, + "learning_rate": 7.995737123993242e-05, + "loss": 0.6533, + "step": 206 + }, + { + "epoch": 0.5690721649484536, + "grad_norm": 1.2333701554709846, + "learning_rate": 7.99537455340574e-05, + "loss": 0.6423, + "step": 207 + }, + { + "epoch": 0.5718213058419244, + "grad_norm": 1.432895391786843, + "learning_rate": 7.994997195659288e-05, + "loss": 0.6588, + "step": 208 + }, + { + "epoch": 0.5745704467353951, + "grad_norm": 1.3560618026372162, + "learning_rate": 7.994605052150512e-05, + "loss": 0.6419, + "step": 209 + }, + { + "epoch": 0.5773195876288659, + "grad_norm": 1.1169005294274361, + "learning_rate": 7.994198124330764e-05, + "loss": 0.643, + "step": 210 + }, + { + "epoch": 0.5800687285223368, + "grad_norm": 1.1551183527776216, + "learning_rate": 7.99377641370611e-05, + "loss": 0.6658, + "step": 211 + }, + { + "epoch": 0.5828178694158076, + "grad_norm": 1.494094722393518, + "learning_rate": 7.993339921837333e-05, + "loss": 0.6566, + "step": 212 + }, + { + "epoch": 0.5855670103092784, + "grad_norm": 1.1087501045845025, + "learning_rate": 7.992888650339918e-05, + "loss": 0.6389, + "step": 213 + }, + { + "epoch": 0.5883161512027492, + "grad_norm": 1.9002883231044971, + "learning_rate": 7.992422600884052e-05, + "loss": 0.6539, + "step": 214 + }, + { + "epoch": 0.5910652920962199, + "grad_norm": 0.7299685549898423, + "learning_rate": 7.991941775194619e-05, + "loss": 0.6465, + "step": 215 + }, + { + "epoch": 0.5938144329896907, + "grad_norm": 1.712564742362983, + "learning_rate": 7.991446175051184e-05, + "loss": 0.6474, + "step": 216 + }, + { + "epoch": 0.5965635738831615, + "grad_norm": 1.4830236426687247, + "learning_rate": 7.990935802288002e-05, + "loss": 0.6512, + "step": 217 + }, + { + "epoch": 0.5993127147766323, + "grad_norm": 1.1265182423363587, + "learning_rate": 7.990410658793994e-05, + "loss": 0.6541, + "step": 218 + }, + { + "epoch": 0.6020618556701031, + "grad_norm": 1.2971103607944785, + "learning_rate": 7.989870746512756e-05, + "loss": 0.6479, + "step": 219 + }, + { + "epoch": 0.6048109965635738, + "grad_norm": 1.2538177052036752, + "learning_rate": 7.989316067442539e-05, + "loss": 0.6497, + "step": 220 + }, + { + "epoch": 0.6075601374570446, + "grad_norm": 0.7957257777767685, + "learning_rate": 7.98874662363625e-05, + "loss": 0.6352, + "step": 221 + }, + { + "epoch": 0.6103092783505155, + "grad_norm": 0.9768965818664068, + "learning_rate": 7.988162417201437e-05, + "loss": 0.6443, + "step": 222 + }, + { + "epoch": 0.6130584192439863, + "grad_norm": 1.9580245398506602, + "learning_rate": 7.987563450300293e-05, + "loss": 0.6652, + "step": 223 + }, + { + "epoch": 0.6158075601374571, + "grad_norm": 1.065435200900762, + "learning_rate": 7.986949725149634e-05, + "loss": 0.6456, + "step": 224 + }, + { + "epoch": 0.6185567010309279, + "grad_norm": 1.9649018261160003, + "learning_rate": 7.986321244020901e-05, + "loss": 0.6479, + "step": 225 + }, + { + "epoch": 0.6213058419243986, + "grad_norm": 1.6213807682895127, + "learning_rate": 7.985678009240142e-05, + "loss": 0.6595, + "step": 226 + }, + { + "epoch": 0.6240549828178694, + "grad_norm": 1.4900550716535215, + "learning_rate": 7.985020023188018e-05, + "loss": 0.6359, + "step": 227 + }, + { + "epoch": 0.6268041237113402, + "grad_norm": 1.3214648767288044, + "learning_rate": 7.98434728829978e-05, + "loss": 0.647, + "step": 228 + }, + { + "epoch": 0.629553264604811, + "grad_norm": 1.4301295600290387, + "learning_rate": 7.983659807065267e-05, + "loss": 0.6475, + "step": 229 + }, + { + "epoch": 0.6323024054982818, + "grad_norm": 1.0375560871972622, + "learning_rate": 7.982957582028892e-05, + "loss": 0.6364, + "step": 230 + }, + { + "epoch": 0.6350515463917525, + "grad_norm": 1.3894087032049172, + "learning_rate": 7.982240615789641e-05, + "loss": 0.6426, + "step": 231 + }, + { + "epoch": 0.6378006872852233, + "grad_norm": 0.9850693356026884, + "learning_rate": 7.981508911001057e-05, + "loss": 0.6459, + "step": 232 + }, + { + "epoch": 0.6405498281786941, + "grad_norm": 1.4426682351463136, + "learning_rate": 7.980762470371228e-05, + "loss": 0.6474, + "step": 233 + }, + { + "epoch": 0.643298969072165, + "grad_norm": 1.0547468177759594, + "learning_rate": 7.980001296662784e-05, + "loss": 0.6469, + "step": 234 + }, + { + "epoch": 0.6460481099656358, + "grad_norm": 1.2056014455360582, + "learning_rate": 7.979225392692882e-05, + "loss": 0.6354, + "step": 235 + }, + { + "epoch": 0.6487972508591066, + "grad_norm": 1.535046209332006, + "learning_rate": 7.978434761333195e-05, + "loss": 0.6432, + "step": 236 + }, + { + "epoch": 0.6515463917525773, + "grad_norm": 0.7757101221034997, + "learning_rate": 7.977629405509905e-05, + "loss": 0.633, + "step": 237 + }, + { + "epoch": 0.6542955326460481, + "grad_norm": 0.992595763686358, + "learning_rate": 7.976809328203693e-05, + "loss": 0.64, + "step": 238 + }, + { + "epoch": 0.6570446735395189, + "grad_norm": 1.3919630066257997, + "learning_rate": 7.975974532449718e-05, + "loss": 0.6495, + "step": 239 + }, + { + "epoch": 0.6597938144329897, + "grad_norm": 1.130072368133697, + "learning_rate": 7.975125021337618e-05, + "loss": 0.6337, + "step": 240 + }, + { + "epoch": 0.6625429553264605, + "grad_norm": 1.0428416388066186, + "learning_rate": 7.974260798011494e-05, + "loss": 0.6515, + "step": 241 + }, + { + "epoch": 0.6652920962199312, + "grad_norm": 1.4678385771525282, + "learning_rate": 7.973381865669897e-05, + "loss": 0.6446, + "step": 242 + }, + { + "epoch": 0.668041237113402, + "grad_norm": 1.1731453018255897, + "learning_rate": 7.972488227565814e-05, + "loss": 0.6424, + "step": 243 + }, + { + "epoch": 0.6707903780068728, + "grad_norm": 1.1037532875258624, + "learning_rate": 7.971579887006663e-05, + "loss": 0.6356, + "step": 244 + }, + { + "epoch": 0.6735395189003437, + "grad_norm": 1.3601555121892395, + "learning_rate": 7.970656847354277e-05, + "loss": 0.6332, + "step": 245 + }, + { + "epoch": 0.6762886597938145, + "grad_norm": 0.6674453682989576, + "learning_rate": 7.969719112024889e-05, + "loss": 0.6341, + "step": 246 + }, + { + "epoch": 0.6790378006872853, + "grad_norm": 0.8185867421910552, + "learning_rate": 7.968766684489122e-05, + "loss": 0.6351, + "step": 247 + }, + { + "epoch": 0.681786941580756, + "grad_norm": 1.2859620347833656, + "learning_rate": 7.967799568271978e-05, + "loss": 0.636, + "step": 248 + }, + { + "epoch": 0.6845360824742268, + "grad_norm": 1.0894539357769442, + "learning_rate": 7.96681776695282e-05, + "loss": 0.6483, + "step": 249 + }, + { + "epoch": 0.6872852233676976, + "grad_norm": 1.536917534466485, + "learning_rate": 7.965821284165362e-05, + "loss": 0.6358, + "step": 250 + }, + { + "epoch": 0.6900343642611684, + "grad_norm": 0.8985662521364918, + "learning_rate": 7.964810123597659e-05, + "loss": 0.6345, + "step": 251 + }, + { + "epoch": 0.6927835051546392, + "grad_norm": 1.274319886309132, + "learning_rate": 7.963784288992085e-05, + "loss": 0.6416, + "step": 252 + }, + { + "epoch": 0.6955326460481099, + "grad_norm": 0.8642801778285559, + "learning_rate": 7.962743784145323e-05, + "loss": 0.6386, + "step": 253 + }, + { + "epoch": 0.6982817869415807, + "grad_norm": 0.9360706797234668, + "learning_rate": 7.961688612908358e-05, + "loss": 0.6311, + "step": 254 + }, + { + "epoch": 0.7010309278350515, + "grad_norm": 1.2347692000194708, + "learning_rate": 7.96061877918645e-05, + "loss": 0.6342, + "step": 255 + }, + { + "epoch": 0.7037800687285224, + "grad_norm": 1.3067449184972966, + "learning_rate": 7.959534286939126e-05, + "loss": 0.6348, + "step": 256 + }, + { + "epoch": 0.7065292096219932, + "grad_norm": 1.0853797825255465, + "learning_rate": 7.95843514018017e-05, + "loss": 0.6387, + "step": 257 + }, + { + "epoch": 0.709278350515464, + "grad_norm": 1.3876365868084906, + "learning_rate": 7.9573213429776e-05, + "loss": 0.6408, + "step": 258 + }, + { + "epoch": 0.7120274914089347, + "grad_norm": 1.2720820097889098, + "learning_rate": 7.956192899453656e-05, + "loss": 0.6403, + "step": 259 + }, + { + "epoch": 0.7147766323024055, + "grad_norm": 1.083330147390025, + "learning_rate": 7.955049813784787e-05, + "loss": 0.6326, + "step": 260 + }, + { + "epoch": 0.7175257731958763, + "grad_norm": 0.8536354360684208, + "learning_rate": 7.953892090201633e-05, + "loss": 0.6289, + "step": 261 + }, + { + "epoch": 0.7202749140893471, + "grad_norm": 0.6669283801432064, + "learning_rate": 7.952719732989007e-05, + "loss": 0.6311, + "step": 262 + }, + { + "epoch": 0.7230240549828179, + "grad_norm": 0.6028131310980143, + "learning_rate": 7.951532746485886e-05, + "loss": 0.6383, + "step": 263 + }, + { + "epoch": 0.7257731958762886, + "grad_norm": 1.0830419136461447, + "learning_rate": 7.950331135085389e-05, + "loss": 0.6257, + "step": 264 + }, + { + "epoch": 0.7285223367697594, + "grad_norm": 1.7725196403596395, + "learning_rate": 7.949114903234766e-05, + "loss": 0.6431, + "step": 265 + }, + { + "epoch": 0.7312714776632302, + "grad_norm": 0.7205576501203266, + "learning_rate": 7.947884055435371e-05, + "loss": 0.636, + "step": 266 + }, + { + "epoch": 0.734020618556701, + "grad_norm": 2.0339470836036386, + "learning_rate": 7.946638596242661e-05, + "loss": 0.6435, + "step": 267 + }, + { + "epoch": 0.7367697594501719, + "grad_norm": 1.0158450970399455, + "learning_rate": 7.945378530266166e-05, + "loss": 0.6342, + "step": 268 + }, + { + "epoch": 0.7395189003436426, + "grad_norm": 2.328741839753608, + "learning_rate": 7.944103862169478e-05, + "loss": 0.6496, + "step": 269 + }, + { + "epoch": 0.7422680412371134, + "grad_norm": 1.5377779904132143, + "learning_rate": 7.94281459667023e-05, + "loss": 0.6536, + "step": 270 + }, + { + "epoch": 0.7450171821305842, + "grad_norm": 2.1537049583171064, + "learning_rate": 7.941510738540086e-05, + "loss": 0.6411, + "step": 271 + }, + { + "epoch": 0.747766323024055, + "grad_norm": 1.7517265079351534, + "learning_rate": 7.940192292604714e-05, + "loss": 0.6489, + "step": 272 + }, + { + "epoch": 0.7505154639175258, + "grad_norm": 2.1942121182211527, + "learning_rate": 7.938859263743776e-05, + "loss": 0.6345, + "step": 273 + }, + { + "epoch": 0.7532646048109966, + "grad_norm": 1.863240282026407, + "learning_rate": 7.937511656890903e-05, + "loss": 0.65, + "step": 274 + }, + { + "epoch": 0.7560137457044673, + "grad_norm": 1.4298523411886686, + "learning_rate": 7.936149477033682e-05, + "loss": 0.6391, + "step": 275 + }, + { + "epoch": 0.7587628865979381, + "grad_norm": 1.2516811961606267, + "learning_rate": 7.934772729213634e-05, + "loss": 0.638, + "step": 276 + }, + { + "epoch": 0.7615120274914089, + "grad_norm": 1.5514968401316607, + "learning_rate": 7.9333814185262e-05, + "loss": 0.6464, + "step": 277 + }, + { + "epoch": 0.7642611683848797, + "grad_norm": 1.1588148872170223, + "learning_rate": 7.931975550120716e-05, + "loss": 0.642, + "step": 278 + }, + { + "epoch": 0.7670103092783506, + "grad_norm": 2.0239434861105403, + "learning_rate": 7.930555129200402e-05, + "loss": 0.6428, + "step": 279 + }, + { + "epoch": 0.7697594501718213, + "grad_norm": 1.6949238476000754, + "learning_rate": 7.929120161022329e-05, + "loss": 0.6333, + "step": 280 + }, + { + "epoch": 0.7725085910652921, + "grad_norm": 1.2740472233054914, + "learning_rate": 7.927670650897421e-05, + "loss": 0.6392, + "step": 281 + }, + { + "epoch": 0.7752577319587629, + "grad_norm": 1.2888896268808947, + "learning_rate": 7.92620660419041e-05, + "loss": 0.6324, + "step": 282 + }, + { + "epoch": 0.7780068728522337, + "grad_norm": 1.2148413669943177, + "learning_rate": 7.924728026319837e-05, + "loss": 0.646, + "step": 283 + }, + { + "epoch": 0.7807560137457045, + "grad_norm": 0.7141756639924294, + "learning_rate": 7.923234922758021e-05, + "loss": 0.6403, + "step": 284 + }, + { + "epoch": 0.7835051546391752, + "grad_norm": 1.0058322496482315, + "learning_rate": 7.921727299031042e-05, + "loss": 0.6418, + "step": 285 + }, + { + "epoch": 0.786254295532646, + "grad_norm": 0.794851634713387, + "learning_rate": 7.920205160718721e-05, + "loss": 0.6383, + "step": 286 + }, + { + "epoch": 0.7890034364261168, + "grad_norm": 0.7763515931120876, + "learning_rate": 7.918668513454598e-05, + "loss": 0.629, + "step": 287 + }, + { + "epoch": 0.7917525773195876, + "grad_norm": 1.0206658407382474, + "learning_rate": 7.917117362925907e-05, + "loss": 0.6369, + "step": 288 + }, + { + "epoch": 0.7945017182130584, + "grad_norm": 0.9044945170274734, + "learning_rate": 7.915551714873571e-05, + "loss": 0.6265, + "step": 289 + }, + { + "epoch": 0.7972508591065293, + "grad_norm": 1.7861242632179295, + "learning_rate": 7.913971575092157e-05, + "loss": 0.6354, + "step": 290 + }, + { + "epoch": 0.8, + "grad_norm": 0.8352377362963846, + "learning_rate": 7.912376949429875e-05, + "loss": 0.6258, + "step": 291 + }, + { + "epoch": 0.8027491408934708, + "grad_norm": 1.1285143530097073, + "learning_rate": 7.910767843788543e-05, + "loss": 0.6411, + "step": 292 + }, + { + "epoch": 0.8054982817869416, + "grad_norm": 2.0986758482850774, + "learning_rate": 7.909144264123575e-05, + "loss": 0.6432, + "step": 293 + }, + { + "epoch": 0.8082474226804124, + "grad_norm": 1.0319530318383152, + "learning_rate": 7.90750621644395e-05, + "loss": 0.6489, + "step": 294 + }, + { + "epoch": 0.8109965635738832, + "grad_norm": 2.628617525239813, + "learning_rate": 7.905853706812199e-05, + "loss": 0.6439, + "step": 295 + }, + { + "epoch": 0.813745704467354, + "grad_norm": 1.977433582449252, + "learning_rate": 7.904186741344373e-05, + "loss": 0.6482, + "step": 296 + }, + { + "epoch": 0.8164948453608247, + "grad_norm": 1.8360229102761563, + "learning_rate": 7.902505326210028e-05, + "loss": 0.65, + "step": 297 + }, + { + "epoch": 0.8192439862542955, + "grad_norm": 1.5433944283254346, + "learning_rate": 7.900809467632197e-05, + "loss": 0.6434, + "step": 298 + }, + { + "epoch": 0.8219931271477663, + "grad_norm": 1.6379518485922502, + "learning_rate": 7.899099171887373e-05, + "loss": 0.6553, + "step": 299 + }, + { + "epoch": 0.8247422680412371, + "grad_norm": 1.477099597856914, + "learning_rate": 7.897374445305478e-05, + "loss": 0.6392, + "step": 300 + }, + { + "epoch": 0.8274914089347079, + "grad_norm": 1.1857730742135408, + "learning_rate": 7.895635294269843e-05, + "loss": 0.6316, + "step": 301 + }, + { + "epoch": 0.8302405498281787, + "grad_norm": 1.361216152623904, + "learning_rate": 7.893881725217191e-05, + "loss": 0.6352, + "step": 302 + }, + { + "epoch": 0.8329896907216495, + "grad_norm": 0.8010015883602253, + "learning_rate": 7.892113744637599e-05, + "loss": 0.6382, + "step": 303 + }, + { + "epoch": 0.8357388316151203, + "grad_norm": 1.102241513039724, + "learning_rate": 7.890331359074488e-05, + "loss": 0.6447, + "step": 304 + }, + { + "epoch": 0.8384879725085911, + "grad_norm": 1.1821390144057153, + "learning_rate": 7.888534575124591e-05, + "loss": 0.6292, + "step": 305 + }, + { + "epoch": 0.8412371134020619, + "grad_norm": 0.7269296274858387, + "learning_rate": 7.886723399437931e-05, + "loss": 0.6319, + "step": 306 + }, + { + "epoch": 0.8439862542955326, + "grad_norm": 1.630523895781267, + "learning_rate": 7.884897838717792e-05, + "loss": 0.6169, + "step": 307 + }, + { + "epoch": 0.8467353951890034, + "grad_norm": 0.740638554856235, + "learning_rate": 7.883057899720703e-05, + "loss": 0.6355, + "step": 308 + }, + { + "epoch": 0.8494845360824742, + "grad_norm": 1.8497443172662162, + "learning_rate": 7.881203589256408e-05, + "loss": 0.6274, + "step": 309 + }, + { + "epoch": 0.852233676975945, + "grad_norm": 1.1910232798072693, + "learning_rate": 7.879334914187836e-05, + "loss": 0.6392, + "step": 310 + }, + { + "epoch": 0.8549828178694158, + "grad_norm": 2.1003839885669815, + "learning_rate": 7.877451881431086e-05, + "loss": 0.6296, + "step": 311 + }, + { + "epoch": 0.8577319587628865, + "grad_norm": 2.041585074702372, + "learning_rate": 7.87555449795539e-05, + "loss": 0.6352, + "step": 312 + }, + { + "epoch": 0.8604810996563574, + "grad_norm": 1.0287133005035243, + "learning_rate": 7.873642770783098e-05, + "loss": 0.6361, + "step": 313 + }, + { + "epoch": 0.8632302405498282, + "grad_norm": 1.4803932962834592, + "learning_rate": 7.871716706989645e-05, + "loss": 0.6433, + "step": 314 + }, + { + "epoch": 0.865979381443299, + "grad_norm": 0.9388113471557215, + "learning_rate": 7.869776313703528e-05, + "loss": 0.6346, + "step": 315 + }, + { + "epoch": 0.8687285223367698, + "grad_norm": 0.8575055727089447, + "learning_rate": 7.867821598106275e-05, + "loss": 0.6287, + "step": 316 + }, + { + "epoch": 0.8714776632302406, + "grad_norm": 0.9753760114729692, + "learning_rate": 7.865852567432428e-05, + "loss": 0.6386, + "step": 317 + }, + { + "epoch": 0.8742268041237113, + "grad_norm": 0.9688066608476287, + "learning_rate": 7.863869228969501e-05, + "loss": 0.631, + "step": 318 + }, + { + "epoch": 0.8769759450171821, + "grad_norm": 0.9593995798512949, + "learning_rate": 7.861871590057971e-05, + "loss": 0.6347, + "step": 319 + }, + { + "epoch": 0.8797250859106529, + "grad_norm": 1.2389383575654551, + "learning_rate": 7.85985965809124e-05, + "loss": 0.6346, + "step": 320 + }, + { + "epoch": 0.8824742268041237, + "grad_norm": 0.8827816805243792, + "learning_rate": 7.857833440515605e-05, + "loss": 0.6265, + "step": 321 + }, + { + "epoch": 0.8852233676975945, + "grad_norm": 0.6061777594451044, + "learning_rate": 7.85579294483024e-05, + "loss": 0.6286, + "step": 322 + }, + { + "epoch": 0.8879725085910652, + "grad_norm": 0.6113638623646376, + "learning_rate": 7.85373817858716e-05, + "loss": 0.6256, + "step": 323 + }, + { + "epoch": 0.8907216494845361, + "grad_norm": 0.8948079812828788, + "learning_rate": 7.851669149391198e-05, + "loss": 0.6238, + "step": 324 + }, + { + "epoch": 0.8934707903780069, + "grad_norm": 1.4121615992032686, + "learning_rate": 7.849585864899976e-05, + "loss": 0.633, + "step": 325 + }, + { + "epoch": 0.8962199312714777, + "grad_norm": 0.7940592789567055, + "learning_rate": 7.847488332823873e-05, + "loss": 0.6359, + "step": 326 + }, + { + "epoch": 0.8989690721649485, + "grad_norm": 0.5764914369379771, + "learning_rate": 7.845376560926002e-05, + "loss": 0.6375, + "step": 327 + }, + { + "epoch": 0.9017182130584193, + "grad_norm": 1.071092163645109, + "learning_rate": 7.843250557022177e-05, + "loss": 0.6196, + "step": 328 + }, + { + "epoch": 0.90446735395189, + "grad_norm": 1.6020470680476175, + "learning_rate": 7.841110328980887e-05, + "loss": 0.6293, + "step": 329 + }, + { + "epoch": 0.9072164948453608, + "grad_norm": 0.4752698709356062, + "learning_rate": 7.838955884723265e-05, + "loss": 0.626, + "step": 330 + }, + { + "epoch": 0.9099656357388316, + "grad_norm": 1.5726092706691321, + "learning_rate": 7.836787232223058e-05, + "loss": 0.6294, + "step": 331 + }, + { + "epoch": 0.9127147766323024, + "grad_norm": 1.01011634702882, + "learning_rate": 7.8346043795066e-05, + "loss": 0.6431, + "step": 332 + }, + { + "epoch": 0.9154639175257732, + "grad_norm": 0.8236758132081481, + "learning_rate": 7.83240733465278e-05, + "loss": 0.6237, + "step": 333 + }, + { + "epoch": 0.9182130584192439, + "grad_norm": 0.816087984512738, + "learning_rate": 7.830196105793017e-05, + "loss": 0.6304, + "step": 334 + }, + { + "epoch": 0.9209621993127147, + "grad_norm": 0.7514093388017495, + "learning_rate": 7.827970701111219e-05, + "loss": 0.6311, + "step": 335 + }, + { + "epoch": 0.9237113402061856, + "grad_norm": 0.9502215222777328, + "learning_rate": 7.825731128843762e-05, + "loss": 0.6317, + "step": 336 + }, + { + "epoch": 0.9264604810996564, + "grad_norm": 1.2376836455505118, + "learning_rate": 7.823477397279464e-05, + "loss": 0.6199, + "step": 337 + }, + { + "epoch": 0.9292096219931272, + "grad_norm": 0.7535199306698559, + "learning_rate": 7.821209514759539e-05, + "loss": 0.6248, + "step": 338 + }, + { + "epoch": 0.931958762886598, + "grad_norm": 0.8490031147278277, + "learning_rate": 7.818927489677577e-05, + "loss": 0.6267, + "step": 339 + }, + { + "epoch": 0.9347079037800687, + "grad_norm": 1.3764847868272945, + "learning_rate": 7.816631330479514e-05, + "loss": 0.6307, + "step": 340 + }, + { + "epoch": 0.9374570446735395, + "grad_norm": 0.6440823293174261, + "learning_rate": 7.814321045663594e-05, + "loss": 0.6366, + "step": 341 + }, + { + "epoch": 0.9402061855670103, + "grad_norm": 1.117461654132214, + "learning_rate": 7.811996643780339e-05, + "loss": 0.6284, + "step": 342 + }, + { + "epoch": 0.9429553264604811, + "grad_norm": 1.198712384928446, + "learning_rate": 7.809658133432526e-05, + "loss": 0.6179, + "step": 343 + }, + { + "epoch": 0.9457044673539519, + "grad_norm": 0.7800718683120617, + "learning_rate": 7.807305523275142e-05, + "loss": 0.6246, + "step": 344 + }, + { + "epoch": 0.9484536082474226, + "grad_norm": 1.106890600114741, + "learning_rate": 7.804938822015361e-05, + "loss": 0.6186, + "step": 345 + }, + { + "epoch": 0.9512027491408934, + "grad_norm": 1.1783968723372122, + "learning_rate": 7.802558038412509e-05, + "loss": 0.6249, + "step": 346 + }, + { + "epoch": 0.9539518900343643, + "grad_norm": 0.909115741821025, + "learning_rate": 7.800163181278033e-05, + "loss": 0.6244, + "step": 347 + }, + { + "epoch": 0.9567010309278351, + "grad_norm": 1.0133199245495836, + "learning_rate": 7.797754259475464e-05, + "loss": 0.6264, + "step": 348 + }, + { + "epoch": 0.9594501718213059, + "grad_norm": 1.195705135014382, + "learning_rate": 7.795331281920387e-05, + "loss": 0.6278, + "step": 349 + }, + { + "epoch": 0.9621993127147767, + "grad_norm": 0.7727285070098746, + "learning_rate": 7.792894257580415e-05, + "loss": 0.6322, + "step": 350 + }, + { + "epoch": 0.9649484536082474, + "grad_norm": 0.9523131612793756, + "learning_rate": 7.790443195475142e-05, + "loss": 0.6251, + "step": 351 + }, + { + "epoch": 0.9676975945017182, + "grad_norm": 1.3036288679788564, + "learning_rate": 7.78797810467612e-05, + "loss": 0.6218, + "step": 352 + }, + { + "epoch": 0.970446735395189, + "grad_norm": 0.5634490374165979, + "learning_rate": 7.785498994306821e-05, + "loss": 0.6221, + "step": 353 + }, + { + "epoch": 0.9731958762886598, + "grad_norm": 1.5260597480048004, + "learning_rate": 7.783005873542605e-05, + "loss": 0.619, + "step": 354 + }, + { + "epoch": 0.9759450171821306, + "grad_norm": 0.6409092943149695, + "learning_rate": 7.780498751610684e-05, + "loss": 0.6251, + "step": 355 + }, + { + "epoch": 0.9786941580756013, + "grad_norm": 1.6664465662420271, + "learning_rate": 7.777977637790092e-05, + "loss": 0.6357, + "step": 356 + }, + { + "epoch": 0.9814432989690721, + "grad_norm": 0.928708660543268, + "learning_rate": 7.775442541411647e-05, + "loss": 0.6218, + "step": 357 + }, + { + "epoch": 0.9841924398625429, + "grad_norm": 1.8449391562901856, + "learning_rate": 7.772893471857915e-05, + "loss": 0.6396, + "step": 358 + }, + { + "epoch": 0.9869415807560138, + "grad_norm": 1.6259312779533432, + "learning_rate": 7.77033043856318e-05, + "loss": 0.6368, + "step": 359 + }, + { + "epoch": 0.9896907216494846, + "grad_norm": 0.9340296294033434, + "learning_rate": 7.767753451013408e-05, + "loss": 0.6204, + "step": 360 + }, + { + "epoch": 0.9924398625429554, + "grad_norm": 1.2307139404659375, + "learning_rate": 7.765162518746207e-05, + "loss": 0.6177, + "step": 361 + }, + { + "epoch": 0.9951890034364261, + "grad_norm": 0.7744777080187526, + "learning_rate": 7.762557651350798e-05, + "loss": 0.6288, + "step": 362 + }, + { + "epoch": 0.9979381443298969, + "grad_norm": 1.0676481742466821, + "learning_rate": 7.759938858467979e-05, + "loss": 0.6272, + "step": 363 + }, + { + "epoch": 1.0006872852233677, + "grad_norm": 0.9232629543637857, + "learning_rate": 7.757306149790082e-05, + "loss": 0.7685, + "step": 364 + }, + { + "epoch": 1.0034364261168385, + "grad_norm": 1.0678574538408068, + "learning_rate": 7.754659535060943e-05, + "loss": 0.5916, + "step": 365 + }, + { + "epoch": 1.0061855670103093, + "grad_norm": 1.2551799774905172, + "learning_rate": 7.751999024075871e-05, + "loss": 0.5988, + "step": 366 + }, + { + "epoch": 1.00893470790378, + "grad_norm": 0.9263667600041875, + "learning_rate": 7.749324626681599e-05, + "loss": 0.5966, + "step": 367 + }, + { + "epoch": 1.0116838487972508, + "grad_norm": 0.731842781771422, + "learning_rate": 7.746636352776259e-05, + "loss": 0.5928, + "step": 368 + }, + { + "epoch": 1.0144329896907216, + "grad_norm": 0.8372632610382791, + "learning_rate": 7.74393421230934e-05, + "loss": 0.5919, + "step": 369 + }, + { + "epoch": 1.0171821305841924, + "grad_norm": 1.0188957663159135, + "learning_rate": 7.741218215281652e-05, + "loss": 0.5932, + "step": 370 + }, + { + "epoch": 1.0199312714776632, + "grad_norm": 1.3394999167844803, + "learning_rate": 7.738488371745287e-05, + "loss": 0.5879, + "step": 371 + }, + { + "epoch": 1.022680412371134, + "grad_norm": 0.8980665393748412, + "learning_rate": 7.735744691803588e-05, + "loss": 0.5942, + "step": 372 + }, + { + "epoch": 1.0254295532646047, + "grad_norm": 0.9856701598024391, + "learning_rate": 7.732987185611102e-05, + "loss": 0.5793, + "step": 373 + }, + { + "epoch": 1.0281786941580755, + "grad_norm": 0.8156719105288647, + "learning_rate": 7.730215863373554e-05, + "loss": 0.5953, + "step": 374 + }, + { + "epoch": 1.0309278350515463, + "grad_norm": 0.5095314934411239, + "learning_rate": 7.727430735347799e-05, + "loss": 0.5848, + "step": 375 + }, + { + "epoch": 1.0336769759450173, + "grad_norm": 0.5914859323410439, + "learning_rate": 7.724631811841789e-05, + "loss": 0.5877, + "step": 376 + }, + { + "epoch": 1.036426116838488, + "grad_norm": 0.6410263828936001, + "learning_rate": 7.721819103214536e-05, + "loss": 0.5852, + "step": 377 + }, + { + "epoch": 1.0391752577319588, + "grad_norm": 0.9724802661391164, + "learning_rate": 7.71899261987607e-05, + "loss": 0.5949, + "step": 378 + }, + { + "epoch": 1.0419243986254296, + "grad_norm": 1.3805105132332318, + "learning_rate": 7.716152372287399e-05, + "loss": 0.5912, + "step": 379 + }, + { + "epoch": 1.0446735395189004, + "grad_norm": 0.5961514316349491, + "learning_rate": 7.713298370960481e-05, + "loss": 0.5989, + "step": 380 + }, + { + "epoch": 1.0474226804123712, + "grad_norm": 1.0381265315374555, + "learning_rate": 7.710430626458171e-05, + "loss": 0.5853, + "step": 381 + }, + { + "epoch": 1.050171821305842, + "grad_norm": 0.88611297798641, + "learning_rate": 7.707549149394192e-05, + "loss": 0.5971, + "step": 382 + }, + { + "epoch": 1.0529209621993127, + "grad_norm": 0.9894621130663098, + "learning_rate": 7.704653950433092e-05, + "loss": 0.5839, + "step": 383 + }, + { + "epoch": 1.0556701030927835, + "grad_norm": 1.4525179754862534, + "learning_rate": 7.7017450402902e-05, + "loss": 0.5841, + "step": 384 + }, + { + "epoch": 1.0584192439862543, + "grad_norm": 0.6031744444046739, + "learning_rate": 7.698822429731595e-05, + "loss": 0.577, + "step": 385 + }, + { + "epoch": 1.061168384879725, + "grad_norm": 1.0814979553674235, + "learning_rate": 7.695886129574065e-05, + "loss": 0.5963, + "step": 386 + }, + { + "epoch": 1.0639175257731959, + "grad_norm": 1.2035521726093799, + "learning_rate": 7.692936150685059e-05, + "loss": 0.59, + "step": 387 + }, + { + "epoch": 1.0666666666666667, + "grad_norm": 1.065535271974634, + "learning_rate": 7.689972503982654e-05, + "loss": 0.59, + "step": 388 + }, + { + "epoch": 1.0694158075601374, + "grad_norm": 0.9885511702158032, + "learning_rate": 7.686995200435513e-05, + "loss": 0.5908, + "step": 389 + }, + { + "epoch": 1.0721649484536082, + "grad_norm": 1.0056851158221545, + "learning_rate": 7.684004251062844e-05, + "loss": 0.592, + "step": 390 + }, + { + "epoch": 1.074914089347079, + "grad_norm": 1.0202284999862656, + "learning_rate": 7.680999666934358e-05, + "loss": 0.5809, + "step": 391 + }, + { + "epoch": 1.0776632302405498, + "grad_norm": 1.1513878865642915, + "learning_rate": 7.677981459170228e-05, + "loss": 0.5878, + "step": 392 + }, + { + "epoch": 1.0804123711340206, + "grad_norm": 0.8699810682515629, + "learning_rate": 7.674949638941053e-05, + "loss": 0.5915, + "step": 393 + }, + { + "epoch": 1.0831615120274913, + "grad_norm": 0.4496053235798414, + "learning_rate": 7.671904217467808e-05, + "loss": 0.5801, + "step": 394 + }, + { + "epoch": 1.0859106529209621, + "grad_norm": 0.6745209087686392, + "learning_rate": 7.668845206021812e-05, + "loss": 0.5979, + "step": 395 + }, + { + "epoch": 1.088659793814433, + "grad_norm": 1.2754326592180227, + "learning_rate": 7.66577261592468e-05, + "loss": 0.5807, + "step": 396 + }, + { + "epoch": 1.0914089347079037, + "grad_norm": 0.7069250995821895, + "learning_rate": 7.662686458548276e-05, + "loss": 0.582, + "step": 397 + }, + { + "epoch": 1.0941580756013747, + "grad_norm": 0.4973387506195001, + "learning_rate": 7.659586745314689e-05, + "loss": 0.5855, + "step": 398 + }, + { + "epoch": 1.0969072164948455, + "grad_norm": 0.5388764587620251, + "learning_rate": 7.65647348769617e-05, + "loss": 0.5901, + "step": 399 + }, + { + "epoch": 1.0996563573883162, + "grad_norm": 0.5432430222716064, + "learning_rate": 7.653346697215103e-05, + "loss": 0.5852, + "step": 400 + }, + { + "epoch": 1.102405498281787, + "grad_norm": 0.7845045531047907, + "learning_rate": 7.650206385443958e-05, + "loss": 0.5815, + "step": 401 + }, + { + "epoch": 1.1051546391752578, + "grad_norm": 1.1593889092948502, + "learning_rate": 7.647052564005243e-05, + "loss": 0.5884, + "step": 402 + }, + { + "epoch": 1.1079037800687286, + "grad_norm": 0.9849246083500859, + "learning_rate": 7.643885244571474e-05, + "loss": 0.5884, + "step": 403 + }, + { + "epoch": 1.1106529209621994, + "grad_norm": 1.2327221523818575, + "learning_rate": 7.64070443886512e-05, + "loss": 0.5808, + "step": 404 + }, + { + "epoch": 1.1134020618556701, + "grad_norm": 0.8894760613915432, + "learning_rate": 7.637510158658563e-05, + "loss": 0.5808, + "step": 405 + }, + { + "epoch": 1.116151202749141, + "grad_norm": 0.8458010101721215, + "learning_rate": 7.634302415774057e-05, + "loss": 0.5921, + "step": 406 + }, + { + "epoch": 1.1189003436426117, + "grad_norm": 1.0269578881644645, + "learning_rate": 7.63108122208368e-05, + "loss": 0.5876, + "step": 407 + }, + { + "epoch": 1.1216494845360825, + "grad_norm": 0.9883024415526525, + "learning_rate": 7.627846589509295e-05, + "loss": 0.5938, + "step": 408 + }, + { + "epoch": 1.1243986254295533, + "grad_norm": 1.0374738987854506, + "learning_rate": 7.624598530022502e-05, + "loss": 0.5759, + "step": 409 + }, + { + "epoch": 1.127147766323024, + "grad_norm": 0.8968993730271081, + "learning_rate": 7.621337055644596e-05, + "loss": 0.5971, + "step": 410 + }, + { + "epoch": 1.1298969072164948, + "grad_norm": 0.645336647760953, + "learning_rate": 7.618062178446518e-05, + "loss": 0.5843, + "step": 411 + }, + { + "epoch": 1.1326460481099656, + "grad_norm": 0.8156198060430901, + "learning_rate": 7.614773910548816e-05, + "loss": 0.5834, + "step": 412 + }, + { + "epoch": 1.1353951890034364, + "grad_norm": 0.9969205762211508, + "learning_rate": 7.6114722641216e-05, + "loss": 0.5874, + "step": 413 + }, + { + "epoch": 1.1381443298969072, + "grad_norm": 0.9512507581396507, + "learning_rate": 7.608157251384493e-05, + "loss": 0.5907, + "step": 414 + }, + { + "epoch": 1.140893470790378, + "grad_norm": 0.46469337738492333, + "learning_rate": 7.604828884606587e-05, + "loss": 0.5854, + "step": 415 + }, + { + "epoch": 1.1436426116838487, + "grad_norm": 0.6875741181821637, + "learning_rate": 7.601487176106397e-05, + "loss": 0.5907, + "step": 416 + }, + { + "epoch": 1.1463917525773195, + "grad_norm": 0.8229065410063605, + "learning_rate": 7.59813213825182e-05, + "loss": 0.5835, + "step": 417 + }, + { + "epoch": 1.1491408934707903, + "grad_norm": 0.5535331052196206, + "learning_rate": 7.594763783460086e-05, + "loss": 0.5913, + "step": 418 + }, + { + "epoch": 1.151890034364261, + "grad_norm": 0.609930416365321, + "learning_rate": 7.591382124197708e-05, + "loss": 0.5861, + "step": 419 + }, + { + "epoch": 1.1546391752577319, + "grad_norm": 0.549391385583707, + "learning_rate": 7.587987172980443e-05, + "loss": 0.5774, + "step": 420 + }, + { + "epoch": 1.1573883161512026, + "grad_norm": 0.5299626474050247, + "learning_rate": 7.58457894237324e-05, + "loss": 0.5915, + "step": 421 + }, + { + "epoch": 1.1601374570446734, + "grad_norm": 0.5702316031822813, + "learning_rate": 7.581157444990199e-05, + "loss": 0.5771, + "step": 422 + }, + { + "epoch": 1.1628865979381444, + "grad_norm": 0.8908922680435231, + "learning_rate": 7.577722693494519e-05, + "loss": 0.5932, + "step": 423 + }, + { + "epoch": 1.1656357388316152, + "grad_norm": 1.4431986503442844, + "learning_rate": 7.574274700598453e-05, + "loss": 0.5813, + "step": 424 + }, + { + "epoch": 1.168384879725086, + "grad_norm": 0.6369720186684446, + "learning_rate": 7.570813479063265e-05, + "loss": 0.5845, + "step": 425 + }, + { + "epoch": 1.1711340206185568, + "grad_norm": 0.5256518804042303, + "learning_rate": 7.567339041699175e-05, + "loss": 0.5924, + "step": 426 + }, + { + "epoch": 1.1738831615120275, + "grad_norm": 0.8165966324851138, + "learning_rate": 7.563851401365316e-05, + "loss": 0.588, + "step": 427 + }, + { + "epoch": 1.1766323024054983, + "grad_norm": 1.1011809927414538, + "learning_rate": 7.56035057096969e-05, + "loss": 0.5878, + "step": 428 + }, + { + "epoch": 1.179381443298969, + "grad_norm": 1.2202082201319733, + "learning_rate": 7.556836563469111e-05, + "loss": 0.589, + "step": 429 + }, + { + "epoch": 1.1821305841924399, + "grad_norm": 0.8269982439978933, + "learning_rate": 7.553309391869167e-05, + "loss": 0.5846, + "step": 430 + }, + { + "epoch": 1.1848797250859107, + "grad_norm": 1.0285455684872933, + "learning_rate": 7.549769069224164e-05, + "loss": 0.5883, + "step": 431 + }, + { + "epoch": 1.1876288659793814, + "grad_norm": 1.4122193045161404, + "learning_rate": 7.546215608637083e-05, + "loss": 0.5832, + "step": 432 + }, + { + "epoch": 1.1903780068728522, + "grad_norm": 0.3041741502072138, + "learning_rate": 7.542649023259527e-05, + "loss": 0.5813, + "step": 433 + }, + { + "epoch": 1.193127147766323, + "grad_norm": 1.5989692268879376, + "learning_rate": 7.53906932629168e-05, + "loss": 0.5896, + "step": 434 + }, + { + "epoch": 1.1958762886597938, + "grad_norm": 0.5519101760067153, + "learning_rate": 7.535476530982244e-05, + "loss": 0.5719, + "step": 435 + }, + { + "epoch": 1.1986254295532646, + "grad_norm": 0.770417484741293, + "learning_rate": 7.53187065062841e-05, + "loss": 0.5881, + "step": 436 + }, + { + "epoch": 1.2013745704467353, + "grad_norm": 1.3484068230851232, + "learning_rate": 7.528251698575788e-05, + "loss": 0.5868, + "step": 437 + }, + { + "epoch": 1.2041237113402061, + "grad_norm": 0.5995693974246964, + "learning_rate": 7.524619688218372e-05, + "loss": 0.5812, + "step": 438 + }, + { + "epoch": 1.206872852233677, + "grad_norm": 0.9168760247232807, + "learning_rate": 7.520974632998485e-05, + "loss": 0.5877, + "step": 439 + }, + { + "epoch": 1.2096219931271477, + "grad_norm": 1.1013744178468492, + "learning_rate": 7.51731654640673e-05, + "loss": 0.592, + "step": 440 + }, + { + "epoch": 1.2123711340206185, + "grad_norm": 0.5790863138582544, + "learning_rate": 7.51364544198194e-05, + "loss": 0.5749, + "step": 441 + }, + { + "epoch": 1.2151202749140895, + "grad_norm": 0.5723305467246104, + "learning_rate": 7.509961333311126e-05, + "loss": 0.5905, + "step": 442 + }, + { + "epoch": 1.2178694158075603, + "grad_norm": 0.6893985276435765, + "learning_rate": 7.506264234029432e-05, + "loss": 0.5799, + "step": 443 + }, + { + "epoch": 1.220618556701031, + "grad_norm": 0.6306086582019484, + "learning_rate": 7.502554157820079e-05, + "loss": 0.583, + "step": 444 + }, + { + "epoch": 1.2233676975945018, + "grad_norm": 0.5828509036657884, + "learning_rate": 7.498831118414316e-05, + "loss": 0.5791, + "step": 445 + }, + { + "epoch": 1.2261168384879726, + "grad_norm": 0.9766394221436852, + "learning_rate": 7.495095129591373e-05, + "loss": 0.5876, + "step": 446 + }, + { + "epoch": 1.2288659793814434, + "grad_norm": 1.416296834720271, + "learning_rate": 7.4913462051784e-05, + "loss": 0.5876, + "step": 447 + }, + { + "epoch": 1.2316151202749142, + "grad_norm": 0.6702851111393332, + "learning_rate": 7.487584359050431e-05, + "loss": 0.5881, + "step": 448 + }, + { + "epoch": 1.234364261168385, + "grad_norm": 0.9384647234340839, + "learning_rate": 7.483809605130319e-05, + "loss": 0.5959, + "step": 449 + }, + { + "epoch": 1.2371134020618557, + "grad_norm": 1.2907110973467573, + "learning_rate": 7.480021957388691e-05, + "loss": 0.5785, + "step": 450 + }, + { + "epoch": 1.2398625429553265, + "grad_norm": 0.8298126531860124, + "learning_rate": 7.476221429843894e-05, + "loss": 0.5934, + "step": 451 + }, + { + "epoch": 1.2426116838487973, + "grad_norm": 1.1127124873762135, + "learning_rate": 7.472408036561942e-05, + "loss": 0.5833, + "step": 452 + }, + { + "epoch": 1.245360824742268, + "grad_norm": 0.7148549185439407, + "learning_rate": 7.468581791656473e-05, + "loss": 0.5927, + "step": 453 + }, + { + "epoch": 1.2481099656357388, + "grad_norm": 0.863808291677187, + "learning_rate": 7.464742709288683e-05, + "loss": 0.5955, + "step": 454 + }, + { + "epoch": 1.2508591065292096, + "grad_norm": 0.8446688366988994, + "learning_rate": 7.460890803667283e-05, + "loss": 0.5939, + "step": 455 + }, + { + "epoch": 1.2536082474226804, + "grad_norm": 1.0639580002350018, + "learning_rate": 7.457026089048445e-05, + "loss": 0.5839, + "step": 456 + }, + { + "epoch": 1.2563573883161512, + "grad_norm": 1.1091140834360853, + "learning_rate": 7.453148579735743e-05, + "loss": 0.5829, + "step": 457 + }, + { + "epoch": 1.259106529209622, + "grad_norm": 0.80462944337638, + "learning_rate": 7.449258290080112e-05, + "loss": 0.5845, + "step": 458 + }, + { + "epoch": 1.2618556701030927, + "grad_norm": 0.8412084155727457, + "learning_rate": 7.445355234479783e-05, + "loss": 0.5767, + "step": 459 + }, + { + "epoch": 1.2646048109965635, + "grad_norm": 0.6330472076770527, + "learning_rate": 7.441439427380235e-05, + "loss": 0.5816, + "step": 460 + }, + { + "epoch": 1.2673539518900343, + "grad_norm": 0.597903788175847, + "learning_rate": 7.437510883274144e-05, + "loss": 0.5729, + "step": 461 + }, + { + "epoch": 1.270103092783505, + "grad_norm": 0.4287714351702764, + "learning_rate": 7.433569616701324e-05, + "loss": 0.5932, + "step": 462 + }, + { + "epoch": 1.2728522336769759, + "grad_norm": 0.47550714789032805, + "learning_rate": 7.429615642248677e-05, + "loss": 0.5929, + "step": 463 + }, + { + "epoch": 1.2756013745704466, + "grad_norm": 0.5414468519874974, + "learning_rate": 7.425648974550138e-05, + "loss": 0.5774, + "step": 464 + }, + { + "epoch": 1.2783505154639174, + "grad_norm": 0.5640877600721679, + "learning_rate": 7.421669628286617e-05, + "loss": 0.586, + "step": 465 + }, + { + "epoch": 1.2810996563573882, + "grad_norm": 0.6048691932188638, + "learning_rate": 7.417677618185955e-05, + "loss": 0.597, + "step": 466 + }, + { + "epoch": 1.283848797250859, + "grad_norm": 0.7508265029471074, + "learning_rate": 7.413672959022856e-05, + "loss": 0.5961, + "step": 467 + }, + { + "epoch": 1.2865979381443298, + "grad_norm": 1.066042006928496, + "learning_rate": 7.409655665618843e-05, + "loss": 0.5837, + "step": 468 + }, + { + "epoch": 1.2893470790378008, + "grad_norm": 1.4460442111626333, + "learning_rate": 7.405625752842198e-05, + "loss": 0.5771, + "step": 469 + }, + { + "epoch": 1.2920962199312716, + "grad_norm": 0.40527253568707877, + "learning_rate": 7.401583235607908e-05, + "loss": 0.5769, + "step": 470 + }, + { + "epoch": 1.2948453608247423, + "grad_norm": 0.8934936601794983, + "learning_rate": 7.397528128877611e-05, + "loss": 0.585, + "step": 471 + }, + { + "epoch": 1.2975945017182131, + "grad_norm": 1.6705979909682174, + "learning_rate": 7.393460447659539e-05, + "loss": 0.5932, + "step": 472 + }, + { + "epoch": 1.300343642611684, + "grad_norm": 0.4422750143504963, + "learning_rate": 7.389380207008462e-05, + "loss": 0.5878, + "step": 473 + }, + { + "epoch": 1.3030927835051547, + "grad_norm": 1.5622517777634235, + "learning_rate": 7.385287422025635e-05, + "loss": 0.5895, + "step": 474 + }, + { + "epoch": 1.3058419243986255, + "grad_norm": 0.6900852059622544, + "learning_rate": 7.381182107858738e-05, + "loss": 0.5883, + "step": 475 + }, + { + "epoch": 1.3085910652920962, + "grad_norm": 0.9882097928385363, + "learning_rate": 7.377064279701827e-05, + "loss": 0.5894, + "step": 476 + }, + { + "epoch": 1.311340206185567, + "grad_norm": 0.9430491893537721, + "learning_rate": 7.37293395279527e-05, + "loss": 0.5921, + "step": 477 + }, + { + "epoch": 1.3140893470790378, + "grad_norm": 0.7682283056356898, + "learning_rate": 7.368791142425691e-05, + "loss": 0.5977, + "step": 478 + }, + { + "epoch": 1.3168384879725086, + "grad_norm": 1.0703629836228026, + "learning_rate": 7.364635863925922e-05, + "loss": 0.5829, + "step": 479 + }, + { + "epoch": 1.3195876288659794, + "grad_norm": 0.8976152595403926, + "learning_rate": 7.360468132674935e-05, + "loss": 0.5902, + "step": 480 + }, + { + "epoch": 1.3223367697594501, + "grad_norm": 0.7540663375222781, + "learning_rate": 7.356287964097795e-05, + "loss": 0.588, + "step": 481 + }, + { + "epoch": 1.325085910652921, + "grad_norm": 1.1037972218662753, + "learning_rate": 7.352095373665598e-05, + "loss": 0.5814, + "step": 482 + }, + { + "epoch": 1.3278350515463917, + "grad_norm": 0.6721571418198156, + "learning_rate": 7.347890376895407e-05, + "loss": 0.586, + "step": 483 + }, + { + "epoch": 1.3305841924398625, + "grad_norm": 0.6580601865881139, + "learning_rate": 7.343672989350214e-05, + "loss": 0.5726, + "step": 484 + }, + { + "epoch": 1.3333333333333333, + "grad_norm": 0.9232629737168565, + "learning_rate": 7.33944322663886e-05, + "loss": 0.5841, + "step": 485 + }, + { + "epoch": 1.3360824742268043, + "grad_norm": 0.921611093133066, + "learning_rate": 7.335201104415992e-05, + "loss": 0.5961, + "step": 486 + }, + { + "epoch": 1.338831615120275, + "grad_norm": 0.7555439012244609, + "learning_rate": 7.330946638381998e-05, + "loss": 0.5877, + "step": 487 + }, + { + "epoch": 1.3415807560137458, + "grad_norm": 0.8027350119891178, + "learning_rate": 7.326679844282953e-05, + "loss": 0.5917, + "step": 488 + }, + { + "epoch": 1.3443298969072166, + "grad_norm": 1.1401383602299868, + "learning_rate": 7.322400737910558e-05, + "loss": 0.5852, + "step": 489 + }, + { + "epoch": 1.3470790378006874, + "grad_norm": 0.805628145096471, + "learning_rate": 7.318109335102083e-05, + "loss": 0.5916, + "step": 490 + }, + { + "epoch": 1.3498281786941582, + "grad_norm": 0.5776321913613638, + "learning_rate": 7.31380565174031e-05, + "loss": 0.5846, + "step": 491 + }, + { + "epoch": 1.352577319587629, + "grad_norm": 0.730623963485602, + "learning_rate": 7.309489703753471e-05, + "loss": 0.5859, + "step": 492 + }, + { + "epoch": 1.3553264604810997, + "grad_norm": 0.6579197440168587, + "learning_rate": 7.305161507115185e-05, + "loss": 0.5796, + "step": 493 + }, + { + "epoch": 1.3580756013745705, + "grad_norm": 0.5630346265554919, + "learning_rate": 7.300821077844413e-05, + "loss": 0.5822, + "step": 494 + }, + { + "epoch": 1.3608247422680413, + "grad_norm": 0.7908819829643783, + "learning_rate": 7.296468432005382e-05, + "loss": 0.5709, + "step": 495 + }, + { + "epoch": 1.363573883161512, + "grad_norm": 1.1004873982746464, + "learning_rate": 7.292103585707541e-05, + "loss": 0.5923, + "step": 496 + }, + { + "epoch": 1.3663230240549828, + "grad_norm": 1.0680631626760493, + "learning_rate": 7.287726555105485e-05, + "loss": 0.5932, + "step": 497 + }, + { + "epoch": 1.3690721649484536, + "grad_norm": 0.9112494358591099, + "learning_rate": 7.283337356398911e-05, + "loss": 0.5785, + "step": 498 + }, + { + "epoch": 1.3718213058419244, + "grad_norm": 1.0337260282053968, + "learning_rate": 7.278936005832549e-05, + "loss": 0.5882, + "step": 499 + }, + { + "epoch": 1.3745704467353952, + "grad_norm": 1.3007806381700902, + "learning_rate": 7.274522519696102e-05, + "loss": 0.5798, + "step": 500 + }, + { + "epoch": 1.377319587628866, + "grad_norm": 0.43519904375561846, + "learning_rate": 7.270096914324189e-05, + "loss": 0.5791, + "step": 501 + }, + { + "epoch": 1.3800687285223368, + "grad_norm": 0.8544940867252295, + "learning_rate": 7.265659206096285e-05, + "loss": 0.5901, + "step": 502 + }, + { + "epoch": 1.3828178694158075, + "grad_norm": 1.2429689674231412, + "learning_rate": 7.261209411436654e-05, + "loss": 0.5809, + "step": 503 + }, + { + "epoch": 1.3855670103092783, + "grad_norm": 0.6308748621084995, + "learning_rate": 7.256747546814298e-05, + "loss": 0.5791, + "step": 504 + }, + { + "epoch": 1.388316151202749, + "grad_norm": 0.7148204479546915, + "learning_rate": 7.252273628742885e-05, + "loss": 0.5796, + "step": 505 + }, + { + "epoch": 1.3910652920962199, + "grad_norm": 0.9925558904292439, + "learning_rate": 7.2477876737807e-05, + "loss": 0.5858, + "step": 506 + }, + { + "epoch": 1.3938144329896907, + "grad_norm": 0.8578115038769086, + "learning_rate": 7.243289698530572e-05, + "loss": 0.5845, + "step": 507 + }, + { + "epoch": 1.3965635738831614, + "grad_norm": 0.7229862425355326, + "learning_rate": 7.23877971963982e-05, + "loss": 0.5755, + "step": 508 + }, + { + "epoch": 1.3993127147766322, + "grad_norm": 0.8939365007653768, + "learning_rate": 7.234257753800191e-05, + "loss": 0.5875, + "step": 509 + }, + { + "epoch": 1.402061855670103, + "grad_norm": 0.9280862590900643, + "learning_rate": 7.229723817747793e-05, + "loss": 0.5916, + "step": 510 + }, + { + "epoch": 1.4048109965635738, + "grad_norm": 0.7722106893797273, + "learning_rate": 7.225177928263042e-05, + "loss": 0.5817, + "step": 511 + }, + { + "epoch": 1.4075601374570446, + "grad_norm": 0.5590378429645152, + "learning_rate": 7.220620102170585e-05, + "loss": 0.5804, + "step": 512 + }, + { + "epoch": 1.4103092783505153, + "grad_norm": 0.5415497371424565, + "learning_rate": 7.21605035633926e-05, + "loss": 0.5829, + "step": 513 + }, + { + "epoch": 1.4130584192439861, + "grad_norm": 0.6720235507203421, + "learning_rate": 7.21146870768201e-05, + "loss": 0.5818, + "step": 514 + }, + { + "epoch": 1.4158075601374571, + "grad_norm": 0.5829950782167737, + "learning_rate": 7.206875173155834e-05, + "loss": 0.5922, + "step": 515 + }, + { + "epoch": 1.418556701030928, + "grad_norm": 0.3621229457523989, + "learning_rate": 7.202269769761726e-05, + "loss": 0.5842, + "step": 516 + }, + { + "epoch": 1.4213058419243987, + "grad_norm": 0.5023031289705603, + "learning_rate": 7.197652514544601e-05, + "loss": 0.576, + "step": 517 + }, + { + "epoch": 1.4240549828178695, + "grad_norm": 0.6099312033435136, + "learning_rate": 7.193023424593245e-05, + "loss": 0.5795, + "step": 518 + }, + { + "epoch": 1.4268041237113402, + "grad_norm": 0.5933868127441155, + "learning_rate": 7.188382517040237e-05, + "loss": 0.5736, + "step": 519 + }, + { + "epoch": 1.429553264604811, + "grad_norm": 0.582367928707779, + "learning_rate": 7.183729809061904e-05, + "loss": 0.5794, + "step": 520 + }, + { + "epoch": 1.4323024054982818, + "grad_norm": 0.7208839572689155, + "learning_rate": 7.179065317878237e-05, + "loss": 0.587, + "step": 521 + }, + { + "epoch": 1.4350515463917526, + "grad_norm": 1.0049567135040567, + "learning_rate": 7.174389060752845e-05, + "loss": 0.5863, + "step": 522 + }, + { + "epoch": 1.4378006872852234, + "grad_norm": 1.3116471278229, + "learning_rate": 7.16970105499288e-05, + "loss": 0.5873, + "step": 523 + }, + { + "epoch": 1.4405498281786941, + "grad_norm": 0.7052150815292837, + "learning_rate": 7.165001317948976e-05, + "loss": 0.5854, + "step": 524 + }, + { + "epoch": 1.443298969072165, + "grad_norm": 0.6674498164029434, + "learning_rate": 7.160289867015187e-05, + "loss": 0.58, + "step": 525 + }, + { + "epoch": 1.4460481099656357, + "grad_norm": 0.6479049706905784, + "learning_rate": 7.155566719628921e-05, + "loss": 0.583, + "step": 526 + }, + { + "epoch": 1.4487972508591065, + "grad_norm": 0.6003355447355602, + "learning_rate": 7.150831893270874e-05, + "loss": 0.5792, + "step": 527 + }, + { + "epoch": 1.4515463917525773, + "grad_norm": 0.7215636928866146, + "learning_rate": 7.146085405464968e-05, + "loss": 0.5746, + "step": 528 + }, + { + "epoch": 1.454295532646048, + "grad_norm": 0.6022527234482411, + "learning_rate": 7.141327273778284e-05, + "loss": 0.5808, + "step": 529 + }, + { + "epoch": 1.4570446735395188, + "grad_norm": 0.44820302089966574, + "learning_rate": 7.136557515820999e-05, + "loss": 0.5745, + "step": 530 + }, + { + "epoch": 1.4597938144329896, + "grad_norm": 0.42351586228242205, + "learning_rate": 7.13177614924632e-05, + "loss": 0.5858, + "step": 531 + }, + { + "epoch": 1.4625429553264606, + "grad_norm": 0.4466996033955547, + "learning_rate": 7.126983191750412e-05, + "loss": 0.5746, + "step": 532 + }, + { + "epoch": 1.4652920962199314, + "grad_norm": 0.40737004869624294, + "learning_rate": 7.12217866107235e-05, + "loss": 0.5818, + "step": 533 + }, + { + "epoch": 1.4680412371134022, + "grad_norm": 0.47191169515150533, + "learning_rate": 7.11736257499403e-05, + "loss": 0.5809, + "step": 534 + }, + { + "epoch": 1.470790378006873, + "grad_norm": 0.4813204578717719, + "learning_rate": 7.112534951340126e-05, + "loss": 0.5844, + "step": 535 + }, + { + "epoch": 1.4735395189003437, + "grad_norm": 0.5549078938525966, + "learning_rate": 7.107695807978007e-05, + "loss": 0.5812, + "step": 536 + }, + { + "epoch": 1.4762886597938145, + "grad_norm": 0.6761669063164707, + "learning_rate": 7.10284516281768e-05, + "loss": 0.5766, + "step": 537 + }, + { + "epoch": 1.4790378006872853, + "grad_norm": 0.6480701339305898, + "learning_rate": 7.097983033811718e-05, + "loss": 0.5805, + "step": 538 + }, + { + "epoch": 1.481786941580756, + "grad_norm": 0.5593193077179679, + "learning_rate": 7.0931094389552e-05, + "loss": 0.575, + "step": 539 + }, + { + "epoch": 1.4845360824742269, + "grad_norm": 0.4942110243815134, + "learning_rate": 7.088224396285638e-05, + "loss": 0.5833, + "step": 540 + }, + { + "epoch": 1.4872852233676976, + "grad_norm": 0.5797979347815979, + "learning_rate": 7.083327923882917e-05, + "loss": 0.5747, + "step": 541 + }, + { + "epoch": 1.4900343642611684, + "grad_norm": 0.9315007373980465, + "learning_rate": 7.078420039869218e-05, + "loss": 0.592, + "step": 542 + }, + { + "epoch": 1.4927835051546392, + "grad_norm": 1.616429951355777, + "learning_rate": 7.073500762408964e-05, + "loss": 0.5944, + "step": 543 + }, + { + "epoch": 1.49553264604811, + "grad_norm": 0.5055431608140836, + "learning_rate": 7.068570109708741e-05, + "loss": 0.5901, + "step": 544 + }, + { + "epoch": 1.4982817869415808, + "grad_norm": 1.0997422339745508, + "learning_rate": 7.06362810001724e-05, + "loss": 0.5902, + "step": 545 + }, + { + "epoch": 1.5010309278350515, + "grad_norm": 1.7657567775705625, + "learning_rate": 7.05867475162518e-05, + "loss": 0.5885, + "step": 546 + }, + { + "epoch": 1.5037800687285223, + "grad_norm": 0.7356302747726631, + "learning_rate": 7.053710082865247e-05, + "loss": 0.5861, + "step": 547 + }, + { + "epoch": 1.506529209621993, + "grad_norm": 2.1759109547514943, + "learning_rate": 7.048734112112026e-05, + "loss": 0.5997, + "step": 548 + }, + { + "epoch": 1.5092783505154639, + "grad_norm": 1.2978110027459113, + "learning_rate": 7.043746857781933e-05, + "loss": 0.6157, + "step": 549 + }, + { + "epoch": 1.5120274914089347, + "grad_norm": 2.219306449832288, + "learning_rate": 7.03874833833314e-05, + "loss": 0.607, + "step": 550 + }, + { + "epoch": 1.5147766323024054, + "grad_norm": 1.9591339401200079, + "learning_rate": 7.033738572265517e-05, + "loss": 0.6369, + "step": 551 + }, + { + "epoch": 1.5175257731958762, + "grad_norm": 1.478246394702369, + "learning_rate": 7.028717578120555e-05, + "loss": 0.6063, + "step": 552 + }, + { + "epoch": 1.520274914089347, + "grad_norm": 1.2579901505180804, + "learning_rate": 7.023685374481301e-05, + "loss": 0.6019, + "step": 553 + }, + { + "epoch": 1.5230240549828178, + "grad_norm": 1.46105914035749, + "learning_rate": 7.018641979972295e-05, + "loss": 0.6094, + "step": 554 + }, + { + "epoch": 1.5257731958762886, + "grad_norm": 0.8090152261985227, + "learning_rate": 7.013587413259486e-05, + "loss": 0.5913, + "step": 555 + }, + { + "epoch": 1.5285223367697593, + "grad_norm": 1.2609062811324916, + "learning_rate": 7.008521693050179e-05, + "loss": 0.5856, + "step": 556 + }, + { + "epoch": 1.5312714776632301, + "grad_norm": 1.0043747397077383, + "learning_rate": 7.003444838092957e-05, + "loss": 0.5989, + "step": 557 + }, + { + "epoch": 1.534020618556701, + "grad_norm": 0.7445889690645772, + "learning_rate": 6.998356867177613e-05, + "loss": 0.5865, + "step": 558 + }, + { + "epoch": 1.5367697594501717, + "grad_norm": 1.2566147555053582, + "learning_rate": 6.993257799135078e-05, + "loss": 0.5855, + "step": 559 + }, + { + "epoch": 1.5395189003436425, + "grad_norm": 0.623650378420435, + "learning_rate": 6.988147652837362e-05, + "loss": 0.5931, + "step": 560 + }, + { + "epoch": 1.5422680412371133, + "grad_norm": 1.0346995401552566, + "learning_rate": 6.983026447197469e-05, + "loss": 0.5914, + "step": 561 + }, + { + "epoch": 1.545017182130584, + "grad_norm": 0.9620047133783202, + "learning_rate": 6.977894201169336e-05, + "loss": 0.5915, + "step": 562 + }, + { + "epoch": 1.547766323024055, + "grad_norm": 0.6073721542597159, + "learning_rate": 6.972750933747765e-05, + "loss": 0.5834, + "step": 563 + }, + { + "epoch": 1.5505154639175258, + "grad_norm": 0.9240637148715215, + "learning_rate": 6.967596663968346e-05, + "loss": 0.5891, + "step": 564 + }, + { + "epoch": 1.5532646048109966, + "grad_norm": 0.7729700501754245, + "learning_rate": 6.962431410907387e-05, + "loss": 0.5864, + "step": 565 + }, + { + "epoch": 1.5560137457044674, + "grad_norm": 0.7023186108460778, + "learning_rate": 6.957255193681852e-05, + "loss": 0.5792, + "step": 566 + }, + { + "epoch": 1.5587628865979382, + "grad_norm": 0.8156415963129797, + "learning_rate": 6.952068031449278e-05, + "loss": 0.5804, + "step": 567 + }, + { + "epoch": 1.561512027491409, + "grad_norm": 0.4418458963414597, + "learning_rate": 6.946869943407714e-05, + "loss": 0.5796, + "step": 568 + }, + { + "epoch": 1.5642611683848797, + "grad_norm": 0.7517403935755494, + "learning_rate": 6.941660948795646e-05, + "loss": 0.5757, + "step": 569 + }, + { + "epoch": 1.5670103092783505, + "grad_norm": 0.6880319452283528, + "learning_rate": 6.936441066891924e-05, + "loss": 0.5866, + "step": 570 + }, + { + "epoch": 1.5697594501718213, + "grad_norm": 0.44431908466865444, + "learning_rate": 6.931210317015694e-05, + "loss": 0.5831, + "step": 571 + }, + { + "epoch": 1.572508591065292, + "grad_norm": 0.5206213485516508, + "learning_rate": 6.925968718526325e-05, + "loss": 0.5835, + "step": 572 + }, + { + "epoch": 1.5752577319587628, + "grad_norm": 0.4865149316480254, + "learning_rate": 6.920716290823337e-05, + "loss": 0.5872, + "step": 573 + }, + { + "epoch": 1.5780068728522336, + "grad_norm": 0.45702506236242896, + "learning_rate": 6.915453053346332e-05, + "loss": 0.5918, + "step": 574 + }, + { + "epoch": 1.5807560137457046, + "grad_norm": 0.44494079428514854, + "learning_rate": 6.910179025574916e-05, + "loss": 0.5889, + "step": 575 + }, + { + "epoch": 1.5835051546391754, + "grad_norm": 0.40161179223147797, + "learning_rate": 6.904894227028631e-05, + "loss": 0.5793, + "step": 576 + }, + { + "epoch": 1.5862542955326462, + "grad_norm": 0.40020351103635327, + "learning_rate": 6.899598677266888e-05, + "loss": 0.5791, + "step": 577 + }, + { + "epoch": 1.589003436426117, + "grad_norm": 0.5533864087207034, + "learning_rate": 6.89429239588888e-05, + "loss": 0.5686, + "step": 578 + }, + { + "epoch": 1.5917525773195877, + "grad_norm": 0.4609556358868059, + "learning_rate": 6.888975402533528e-05, + "loss": 0.5696, + "step": 579 + }, + { + "epoch": 1.5945017182130585, + "grad_norm": 0.376485016123922, + "learning_rate": 6.883647716879389e-05, + "loss": 0.5787, + "step": 580 + }, + { + "epoch": 1.5972508591065293, + "grad_norm": 0.3703464382281221, + "learning_rate": 6.878309358644602e-05, + "loss": 0.5786, + "step": 581 + }, + { + "epoch": 1.6, + "grad_norm": 0.36242738510972944, + "learning_rate": 6.8729603475868e-05, + "loss": 0.5838, + "step": 582 + }, + { + "epoch": 1.6027491408934709, + "grad_norm": 0.4005050809946878, + "learning_rate": 6.867600703503044e-05, + "loss": 0.5828, + "step": 583 + }, + { + "epoch": 1.6054982817869417, + "grad_norm": 0.42329454686286444, + "learning_rate": 6.86223044622975e-05, + "loss": 0.5803, + "step": 584 + }, + { + "epoch": 1.6082474226804124, + "grad_norm": 0.3329212652227201, + "learning_rate": 6.856849595642617e-05, + "loss": 0.5822, + "step": 585 + }, + { + "epoch": 1.6109965635738832, + "grad_norm": 0.46886870029058014, + "learning_rate": 6.851458171656541e-05, + "loss": 0.58, + "step": 586 + }, + { + "epoch": 1.613745704467354, + "grad_norm": 0.5556659975442035, + "learning_rate": 6.846056194225562e-05, + "loss": 0.5702, + "step": 587 + }, + { + "epoch": 1.6164948453608248, + "grad_norm": 0.5260721900968394, + "learning_rate": 6.84064368334277e-05, + "loss": 0.5867, + "step": 588 + }, + { + "epoch": 1.6192439862542956, + "grad_norm": 0.40786759602060935, + "learning_rate": 6.835220659040246e-05, + "loss": 0.5834, + "step": 589 + }, + { + "epoch": 1.6219931271477663, + "grad_norm": 0.2996276604956069, + "learning_rate": 6.82978714138898e-05, + "loss": 0.5776, + "step": 590 + }, + { + "epoch": 1.6247422680412371, + "grad_norm": 0.35270615435100505, + "learning_rate": 6.824343150498798e-05, + "loss": 0.5783, + "step": 591 + }, + { + "epoch": 1.627491408934708, + "grad_norm": 0.28376863619687237, + "learning_rate": 6.81888870651829e-05, + "loss": 0.5767, + "step": 592 + }, + { + "epoch": 1.6302405498281787, + "grad_norm": 0.36443318149486764, + "learning_rate": 6.813423829634732e-05, + "loss": 0.579, + "step": 593 + }, + { + "epoch": 1.6329896907216495, + "grad_norm": 0.481183239710988, + "learning_rate": 6.80794854007401e-05, + "loss": 0.5815, + "step": 594 + }, + { + "epoch": 1.6357388316151202, + "grad_norm": 0.4948913070938149, + "learning_rate": 6.802462858100556e-05, + "loss": 0.5847, + "step": 595 + }, + { + "epoch": 1.638487972508591, + "grad_norm": 0.3815071178544923, + "learning_rate": 6.796966804017258e-05, + "loss": 0.5724, + "step": 596 + }, + { + "epoch": 1.6412371134020618, + "grad_norm": 0.4407973877140981, + "learning_rate": 6.791460398165393e-05, + "loss": 0.5811, + "step": 597 + }, + { + "epoch": 1.6439862542955326, + "grad_norm": 0.43391575615156724, + "learning_rate": 6.785943660924553e-05, + "loss": 0.5827, + "step": 598 + }, + { + "epoch": 1.6467353951890034, + "grad_norm": 0.4035737413547384, + "learning_rate": 6.780416612712568e-05, + "loss": 0.5819, + "step": 599 + }, + { + "epoch": 1.6494845360824741, + "grad_norm": 0.5281104884380362, + "learning_rate": 6.774879273985423e-05, + "loss": 0.5775, + "step": 600 + }, + { + "epoch": 1.652233676975945, + "grad_norm": 0.7760361977252223, + "learning_rate": 6.769331665237198e-05, + "loss": 0.5899, + "step": 601 + }, + { + "epoch": 1.6549828178694157, + "grad_norm": 1.09403271674242, + "learning_rate": 6.763773806999977e-05, + "loss": 0.5859, + "step": 602 + }, + { + "epoch": 1.6577319587628865, + "grad_norm": 0.9348914419022682, + "learning_rate": 6.75820571984378e-05, + "loss": 0.5918, + "step": 603 + }, + { + "epoch": 1.6604810996563573, + "grad_norm": 0.7435836588260047, + "learning_rate": 6.752627424376484e-05, + "loss": 0.582, + "step": 604 + }, + { + "epoch": 1.663230240549828, + "grad_norm": 0.5416202424064285, + "learning_rate": 6.747038941243748e-05, + "loss": 0.5845, + "step": 605 + }, + { + "epoch": 1.6659793814432988, + "grad_norm": 0.4289104135369005, + "learning_rate": 6.741440291128938e-05, + "loss": 0.5841, + "step": 606 + }, + { + "epoch": 1.6687285223367696, + "grad_norm": 0.3842777168497921, + "learning_rate": 6.735831494753046e-05, + "loss": 0.5783, + "step": 607 + }, + { + "epoch": 1.6714776632302404, + "grad_norm": 0.3676465027204242, + "learning_rate": 6.730212572874618e-05, + "loss": 0.5708, + "step": 608 + }, + { + "epoch": 1.6742268041237114, + "grad_norm": 0.3960487648888393, + "learning_rate": 6.724583546289672e-05, + "loss": 0.5724, + "step": 609 + }, + { + "epoch": 1.6769759450171822, + "grad_norm": 0.4083333422479102, + "learning_rate": 6.71894443583163e-05, + "loss": 0.5925, + "step": 610 + }, + { + "epoch": 1.679725085910653, + "grad_norm": 0.4578552218287685, + "learning_rate": 6.713295262371232e-05, + "loss": 0.5715, + "step": 611 + }, + { + "epoch": 1.6824742268041237, + "grad_norm": 0.5656117537132986, + "learning_rate": 6.707636046816457e-05, + "loss": 0.5766, + "step": 612 + }, + { + "epoch": 1.6852233676975945, + "grad_norm": 0.6850771753028508, + "learning_rate": 6.70196681011246e-05, + "loss": 0.5928, + "step": 613 + }, + { + "epoch": 1.6879725085910653, + "grad_norm": 0.698725362743093, + "learning_rate": 6.696287573241478e-05, + "loss": 0.5897, + "step": 614 + }, + { + "epoch": 1.690721649484536, + "grad_norm": 0.7382677867943621, + "learning_rate": 6.690598357222764e-05, + "loss": 0.5668, + "step": 615 + }, + { + "epoch": 1.6934707903780069, + "grad_norm": 0.7105270810835767, + "learning_rate": 6.684899183112498e-05, + "loss": 0.5736, + "step": 616 + }, + { + "epoch": 1.6962199312714776, + "grad_norm": 0.6862323082435818, + "learning_rate": 6.679190072003722e-05, + "loss": 0.5765, + "step": 617 + }, + { + "epoch": 1.6989690721649484, + "grad_norm": 0.6299104085981699, + "learning_rate": 6.67347104502625e-05, + "loss": 0.5642, + "step": 618 + }, + { + "epoch": 1.7017182130584192, + "grad_norm": 0.6026002947844302, + "learning_rate": 6.667742123346603e-05, + "loss": 0.5804, + "step": 619 + }, + { + "epoch": 1.7044673539518902, + "grad_norm": 0.4462813007489562, + "learning_rate": 6.662003328167917e-05, + "loss": 0.5807, + "step": 620 + }, + { + "epoch": 1.707216494845361, + "grad_norm": 0.29472463703813456, + "learning_rate": 6.65625468072987e-05, + "loss": 0.5702, + "step": 621 + }, + { + "epoch": 1.7099656357388318, + "grad_norm": 0.320893033856143, + "learning_rate": 6.650496202308607e-05, + "loss": 0.5835, + "step": 622 + }, + { + "epoch": 1.7127147766323025, + "grad_norm": 0.39010773443389357, + "learning_rate": 6.644727914216656e-05, + "loss": 0.576, + "step": 623 + }, + { + "epoch": 1.7154639175257733, + "grad_norm": 0.5319160826521161, + "learning_rate": 6.638949837802855e-05, + "loss": 0.5798, + "step": 624 + }, + { + "epoch": 1.718213058419244, + "grad_norm": 0.6498383386047935, + "learning_rate": 6.633161994452262e-05, + "loss": 0.5795, + "step": 625 + }, + { + "epoch": 1.7209621993127149, + "grad_norm": 0.7259267311718851, + "learning_rate": 6.627364405586091e-05, + "loss": 0.5919, + "step": 626 + }, + { + "epoch": 1.7237113402061857, + "grad_norm": 0.8162977870752441, + "learning_rate": 6.621557092661621e-05, + "loss": 0.5753, + "step": 627 + }, + { + "epoch": 1.7264604810996564, + "grad_norm": 0.7728017259091454, + "learning_rate": 6.61574007717212e-05, + "loss": 0.5843, + "step": 628 + }, + { + "epoch": 1.7292096219931272, + "grad_norm": 0.7443225127869316, + "learning_rate": 6.609913380646767e-05, + "loss": 0.5802, + "step": 629 + }, + { + "epoch": 1.731958762886598, + "grad_norm": 0.7541481509168282, + "learning_rate": 6.604077024650571e-05, + "loss": 0.5841, + "step": 630 + }, + { + "epoch": 1.7347079037800688, + "grad_norm": 0.6874877805251565, + "learning_rate": 6.598231030784289e-05, + "loss": 0.5861, + "step": 631 + }, + { + "epoch": 1.7374570446735396, + "grad_norm": 0.6302784745574853, + "learning_rate": 6.592375420684354e-05, + "loss": 0.579, + "step": 632 + }, + { + "epoch": 1.7402061855670103, + "grad_norm": 0.3605005517279551, + "learning_rate": 6.586510216022785e-05, + "loss": 0.5761, + "step": 633 + }, + { + "epoch": 1.7429553264604811, + "grad_norm": 0.40851626295602705, + "learning_rate": 6.58063543850711e-05, + "loss": 0.5809, + "step": 634 + }, + { + "epoch": 1.745704467353952, + "grad_norm": 0.5889121925134353, + "learning_rate": 6.574751109880295e-05, + "loss": 0.589, + "step": 635 + }, + { + "epoch": 1.7484536082474227, + "grad_norm": 0.6855740977784578, + "learning_rate": 6.568857251920641e-05, + "loss": 0.5791, + "step": 636 + }, + { + "epoch": 1.7512027491408935, + "grad_norm": 0.7763253623405283, + "learning_rate": 6.562953886441731e-05, + "loss": 0.5698, + "step": 637 + }, + { + "epoch": 1.7539518900343642, + "grad_norm": 0.8027574724353924, + "learning_rate": 6.557041035292331e-05, + "loss": 0.5695, + "step": 638 + }, + { + "epoch": 1.756701030927835, + "grad_norm": 0.8124682042384597, + "learning_rate": 6.551118720356313e-05, + "loss": 0.5793, + "step": 639 + }, + { + "epoch": 1.7594501718213058, + "grad_norm": 0.7285676570172558, + "learning_rate": 6.545186963552578e-05, + "loss": 0.5809, + "step": 640 + }, + { + "epoch": 1.7621993127147766, + "grad_norm": 0.5493034420656492, + "learning_rate": 6.539245786834965e-05, + "loss": 0.5779, + "step": 641 + }, + { + "epoch": 1.7649484536082474, + "grad_norm": 0.4100237013175215, + "learning_rate": 6.533295212192189e-05, + "loss": 0.5634, + "step": 642 + }, + { + "epoch": 1.7676975945017182, + "grad_norm": 0.44386016701061204, + "learning_rate": 6.527335261647735e-05, + "loss": 0.5865, + "step": 643 + }, + { + "epoch": 1.770446735395189, + "grad_norm": 0.7605693123034108, + "learning_rate": 6.521365957259798e-05, + "loss": 0.5758, + "step": 644 + }, + { + "epoch": 1.7731958762886597, + "grad_norm": 0.9531489941794372, + "learning_rate": 6.515387321121185e-05, + "loss": 0.5845, + "step": 645 + }, + { + "epoch": 1.7759450171821305, + "grad_norm": 0.9748291614843003, + "learning_rate": 6.509399375359247e-05, + "loss": 0.5719, + "step": 646 + }, + { + "epoch": 1.7786941580756013, + "grad_norm": 0.7628578645914652, + "learning_rate": 6.503402142135787e-05, + "loss": 0.5768, + "step": 647 + }, + { + "epoch": 1.781443298969072, + "grad_norm": 0.5799935111185708, + "learning_rate": 6.497395643646981e-05, + "loss": 0.5701, + "step": 648 + }, + { + "epoch": 1.7841924398625428, + "grad_norm": 0.5277145245316653, + "learning_rate": 6.491379902123297e-05, + "loss": 0.5683, + "step": 649 + }, + { + "epoch": 1.7869415807560136, + "grad_norm": 0.4736528081812741, + "learning_rate": 6.485354939829414e-05, + "loss": 0.5779, + "step": 650 + }, + { + "epoch": 1.7896907216494844, + "grad_norm": 0.5669836094412534, + "learning_rate": 6.479320779064134e-05, + "loss": 0.5764, + "step": 651 + }, + { + "epoch": 1.7924398625429552, + "grad_norm": 0.559780538593531, + "learning_rate": 6.473277442160309e-05, + "loss": 0.5832, + "step": 652 + }, + { + "epoch": 1.795189003436426, + "grad_norm": 0.5065885370576576, + "learning_rate": 6.467224951484746e-05, + "loss": 0.5783, + "step": 653 + }, + { + "epoch": 1.797938144329897, + "grad_norm": 0.4559459768927469, + "learning_rate": 6.461163329438135e-05, + "loss": 0.5715, + "step": 654 + }, + { + "epoch": 1.8006872852233677, + "grad_norm": 0.42548224770937315, + "learning_rate": 6.45509259845496e-05, + "loss": 0.5751, + "step": 655 + }, + { + "epoch": 1.8034364261168385, + "grad_norm": 0.3455007524085001, + "learning_rate": 6.449012781003419e-05, + "loss": 0.5754, + "step": 656 + }, + { + "epoch": 1.8061855670103093, + "grad_norm": 0.4400225976066889, + "learning_rate": 6.442923899585338e-05, + "loss": 0.5941, + "step": 657 + }, + { + "epoch": 1.80893470790378, + "grad_norm": 0.5084943205879741, + "learning_rate": 6.436825976736092e-05, + "loss": 0.579, + "step": 658 + }, + { + "epoch": 1.8116838487972509, + "grad_norm": 0.5358918155905726, + "learning_rate": 6.430719035024515e-05, + "loss": 0.5748, + "step": 659 + }, + { + "epoch": 1.8144329896907216, + "grad_norm": 0.503047756943303, + "learning_rate": 6.424603097052826e-05, + "loss": 0.5794, + "step": 660 + }, + { + "epoch": 1.8171821305841924, + "grad_norm": 0.44321729048134617, + "learning_rate": 6.418478185456534e-05, + "loss": 0.5865, + "step": 661 + }, + { + "epoch": 1.8199312714776632, + "grad_norm": 0.3737976145350619, + "learning_rate": 6.412344322904362e-05, + "loss": 0.5762, + "step": 662 + }, + { + "epoch": 1.822680412371134, + "grad_norm": 0.2731612660975183, + "learning_rate": 6.406201532098165e-05, + "loss": 0.5691, + "step": 663 + }, + { + "epoch": 1.8254295532646048, + "grad_norm": 0.3311008105208644, + "learning_rate": 6.400049835772836e-05, + "loss": 0.5729, + "step": 664 + }, + { + "epoch": 1.8281786941580758, + "grad_norm": 0.3627533269491876, + "learning_rate": 6.393889256696231e-05, + "loss": 0.5729, + "step": 665 + }, + { + "epoch": 1.8309278350515465, + "grad_norm": 0.3316310187075983, + "learning_rate": 6.387719817669081e-05, + "loss": 0.582, + "step": 666 + }, + { + "epoch": 1.8336769759450173, + "grad_norm": 0.3431221379909935, + "learning_rate": 6.381541541524911e-05, + "loss": 0.5775, + "step": 667 + }, + { + "epoch": 1.836426116838488, + "grad_norm": 0.308724522020036, + "learning_rate": 6.375354451129949e-05, + "loss": 0.5759, + "step": 668 + }, + { + "epoch": 1.839175257731959, + "grad_norm": 0.2799060976007718, + "learning_rate": 6.369158569383045e-05, + "loss": 0.5826, + "step": 669 + }, + { + "epoch": 1.8419243986254297, + "grad_norm": 0.3820386705421886, + "learning_rate": 6.362953919215591e-05, + "loss": 0.5778, + "step": 670 + }, + { + "epoch": 1.8446735395189005, + "grad_norm": 0.38246444479750036, + "learning_rate": 6.356740523591427e-05, + "loss": 0.5749, + "step": 671 + }, + { + "epoch": 1.8474226804123712, + "grad_norm": 0.3271098200229515, + "learning_rate": 6.350518405506764e-05, + "loss": 0.5817, + "step": 672 + }, + { + "epoch": 1.850171821305842, + "grad_norm": 0.34228419827242806, + "learning_rate": 6.344287587990091e-05, + "loss": 0.5819, + "step": 673 + }, + { + "epoch": 1.8529209621993128, + "grad_norm": 0.32655548951800334, + "learning_rate": 6.338048094102096e-05, + "loss": 0.5749, + "step": 674 + }, + { + "epoch": 1.8556701030927836, + "grad_norm": 0.32871268032009765, + "learning_rate": 6.331799946935583e-05, + "loss": 0.5685, + "step": 675 + }, + { + "epoch": 1.8584192439862544, + "grad_norm": 0.4199684721453658, + "learning_rate": 6.325543169615378e-05, + "loss": 0.5725, + "step": 676 + }, + { + "epoch": 1.8611683848797251, + "grad_norm": 0.5756048364369786, + "learning_rate": 6.319277785298247e-05, + "loss": 0.5805, + "step": 677 + }, + { + "epoch": 1.863917525773196, + "grad_norm": 0.6547196962591697, + "learning_rate": 6.313003817172812e-05, + "loss": 0.5662, + "step": 678 + }, + { + "epoch": 1.8666666666666667, + "grad_norm": 0.7656825534443144, + "learning_rate": 6.30672128845947e-05, + "loss": 0.5864, + "step": 679 + }, + { + "epoch": 1.8694158075601375, + "grad_norm": 0.9358535366007116, + "learning_rate": 6.300430222410292e-05, + "loss": 0.5798, + "step": 680 + }, + { + "epoch": 1.8721649484536083, + "grad_norm": 0.9703236167568001, + "learning_rate": 6.294130642308952e-05, + "loss": 0.5752, + "step": 681 + }, + { + "epoch": 1.874914089347079, + "grad_norm": 0.8892433252848201, + "learning_rate": 6.287822571470636e-05, + "loss": 0.5709, + "step": 682 + }, + { + "epoch": 1.8776632302405498, + "grad_norm": 0.8810114036796293, + "learning_rate": 6.281506033241948e-05, + "loss": 0.5733, + "step": 683 + }, + { + "epoch": 1.8804123711340206, + "grad_norm": 0.853242199524319, + "learning_rate": 6.275181051000842e-05, + "loss": 0.5758, + "step": 684 + }, + { + "epoch": 1.8831615120274914, + "grad_norm": 0.6715868178268443, + "learning_rate": 6.26884764815651e-05, + "loss": 0.5681, + "step": 685 + }, + { + "epoch": 1.8859106529209622, + "grad_norm": 0.41927983624654164, + "learning_rate": 6.26250584814932e-05, + "loss": 0.5706, + "step": 686 + }, + { + "epoch": 1.888659793814433, + "grad_norm": 0.35123714928190247, + "learning_rate": 6.256155674450712e-05, + "loss": 0.5742, + "step": 687 + }, + { + "epoch": 1.8914089347079037, + "grad_norm": 0.5653996664319825, + "learning_rate": 6.249797150563124e-05, + "loss": 0.5731, + "step": 688 + }, + { + "epoch": 1.8941580756013745, + "grad_norm": 0.6649259671430193, + "learning_rate": 6.243430300019891e-05, + "loss": 0.5797, + "step": 689 + }, + { + "epoch": 1.8969072164948453, + "grad_norm": 0.60181480524739, + "learning_rate": 6.237055146385173e-05, + "loss": 0.5758, + "step": 690 + }, + { + "epoch": 1.899656357388316, + "grad_norm": 0.6235277209689051, + "learning_rate": 6.230671713253854e-05, + "loss": 0.5778, + "step": 691 + }, + { + "epoch": 1.9024054982817868, + "grad_norm": 0.6196187849528243, + "learning_rate": 6.224280024251466e-05, + "loss": 0.5698, + "step": 692 + }, + { + "epoch": 1.9051546391752576, + "grad_norm": 0.42709196066527666, + "learning_rate": 6.217880103034091e-05, + "loss": 0.5736, + "step": 693 + }, + { + "epoch": 1.9079037800687284, + "grad_norm": 0.34692776304267553, + "learning_rate": 6.211471973288285e-05, + "loss": 0.5712, + "step": 694 + }, + { + "epoch": 1.9106529209621992, + "grad_norm": 0.3110542487628964, + "learning_rate": 6.205055658730983e-05, + "loss": 0.5653, + "step": 695 + }, + { + "epoch": 1.91340206185567, + "grad_norm": 0.34540933819310177, + "learning_rate": 6.198631183109408e-05, + "loss": 0.576, + "step": 696 + }, + { + "epoch": 1.9161512027491407, + "grad_norm": 0.3126265820113025, + "learning_rate": 6.192198570200992e-05, + "loss": 0.5906, + "step": 697 + }, + { + "epoch": 1.9189003436426115, + "grad_norm": 0.29077210119155705, + "learning_rate": 6.185757843813282e-05, + "loss": 0.5799, + "step": 698 + }, + { + "epoch": 1.9216494845360823, + "grad_norm": 0.43001232514386034, + "learning_rate": 6.179309027783856e-05, + "loss": 0.5784, + "step": 699 + }, + { + "epoch": 1.9243986254295533, + "grad_norm": 0.4962679375465972, + "learning_rate": 6.172852145980228e-05, + "loss": 0.5731, + "step": 700 + }, + { + "epoch": 1.927147766323024, + "grad_norm": 0.5311714605519653, + "learning_rate": 6.166387222299767e-05, + "loss": 0.5812, + "step": 701 + }, + { + "epoch": 1.9298969072164949, + "grad_norm": 0.586368926903193, + "learning_rate": 6.159914280669607e-05, + "loss": 0.5638, + "step": 702 + }, + { + "epoch": 1.9326460481099657, + "grad_norm": 0.6549081294272183, + "learning_rate": 6.153433345046552e-05, + "loss": 0.5785, + "step": 703 + }, + { + "epoch": 1.9353951890034364, + "grad_norm": 0.7252681437363886, + "learning_rate": 6.146944439416997e-05, + "loss": 0.5751, + "step": 704 + }, + { + "epoch": 1.9381443298969072, + "grad_norm": 0.7845460102188683, + "learning_rate": 6.140447587796832e-05, + "loss": 0.5646, + "step": 705 + }, + { + "epoch": 1.940893470790378, + "grad_norm": 0.8048678290944055, + "learning_rate": 6.133942814231357e-05, + "loss": 0.5742, + "step": 706 + }, + { + "epoch": 1.9436426116838488, + "grad_norm": 0.7734169093084938, + "learning_rate": 6.127430142795192e-05, + "loss": 0.5876, + "step": 707 + }, + { + "epoch": 1.9463917525773196, + "grad_norm": 0.5484007054312127, + "learning_rate": 6.120909597592185e-05, + "loss": 0.5843, + "step": 708 + }, + { + "epoch": 1.9491408934707903, + "grad_norm": 0.3242608423209931, + "learning_rate": 6.114381202755328e-05, + "loss": 0.5654, + "step": 709 + }, + { + "epoch": 1.9518900343642611, + "grad_norm": 0.3539205469330434, + "learning_rate": 6.107844982446662e-05, + "loss": 0.5833, + "step": 710 + }, + { + "epoch": 1.9546391752577321, + "grad_norm": 0.5156015064072683, + "learning_rate": 6.1013009608571954e-05, + "loss": 0.583, + "step": 711 + }, + { + "epoch": 1.957388316151203, + "grad_norm": 0.6131503596319059, + "learning_rate": 6.094749162206807e-05, + "loss": 0.5742, + "step": 712 + }, + { + "epoch": 1.9601374570446737, + "grad_norm": 0.6333732757531064, + "learning_rate": 6.0881896107441575e-05, + "loss": 0.5751, + "step": 713 + }, + { + "epoch": 1.9628865979381445, + "grad_norm": 0.456390717989884, + "learning_rate": 6.081622330746603e-05, + "loss": 0.5844, + "step": 714 + }, + { + "epoch": 1.9656357388316152, + "grad_norm": 0.41591397010146025, + "learning_rate": 6.075047346520105e-05, + "loss": 0.5756, + "step": 715 + }, + { + "epoch": 1.968384879725086, + "grad_norm": 0.5054704227601636, + "learning_rate": 6.0684646823991345e-05, + "loss": 0.5688, + "step": 716 + }, + { + "epoch": 1.9711340206185568, + "grad_norm": 0.6526494286710874, + "learning_rate": 6.061874362746592e-05, + "loss": 0.568, + "step": 717 + }, + { + "epoch": 1.9738831615120276, + "grad_norm": 0.7155431076783062, + "learning_rate": 6.055276411953705e-05, + "loss": 0.5746, + "step": 718 + }, + { + "epoch": 1.9766323024054984, + "grad_norm": 0.5467249394073967, + "learning_rate": 6.048670854439952e-05, + "loss": 0.5793, + "step": 719 + }, + { + "epoch": 1.9793814432989691, + "grad_norm": 0.44545561246216775, + "learning_rate": 6.042057714652958e-05, + "loss": 0.5793, + "step": 720 + }, + { + "epoch": 1.98213058419244, + "grad_norm": 0.43668565117704844, + "learning_rate": 6.035437017068413e-05, + "loss": 0.5711, + "step": 721 + }, + { + "epoch": 1.9848797250859107, + "grad_norm": 0.4794911786968242, + "learning_rate": 6.02880878618998e-05, + "loss": 0.5764, + "step": 722 + }, + { + "epoch": 1.9876288659793815, + "grad_norm": 0.45936527802619287, + "learning_rate": 6.022173046549201e-05, + "loss": 0.5728, + "step": 723 + }, + { + "epoch": 1.9903780068728523, + "grad_norm": 0.3710590792543285, + "learning_rate": 6.0155298227054125e-05, + "loss": 0.5764, + "step": 724 + }, + { + "epoch": 1.993127147766323, + "grad_norm": 0.3684226778884953, + "learning_rate": 6.008879139245645e-05, + "loss": 0.5711, + "step": 725 + }, + { + "epoch": 1.9958762886597938, + "grad_norm": 0.37506727135987533, + "learning_rate": 6.00222102078454e-05, + "loss": 0.5655, + "step": 726 + }, + { + "epoch": 1.9986254295532646, + "grad_norm": 0.42203860157012285, + "learning_rate": 5.995555491964259e-05, + "loss": 0.6314, + "step": 727 + }, + { + "epoch": 2.0013745704467354, + "grad_norm": 0.6121901786626351, + "learning_rate": 5.988882577454386e-05, + "loss": 0.6437, + "step": 728 + }, + { + "epoch": 2.004123711340206, + "grad_norm": 0.838029390831891, + "learning_rate": 5.982202301951841e-05, + "loss": 0.5246, + "step": 729 + }, + { + "epoch": 2.006872852233677, + "grad_norm": 1.0035111815398425, + "learning_rate": 5.97551469018079e-05, + "loss": 0.532, + "step": 730 + }, + { + "epoch": 2.0096219931271477, + "grad_norm": 1.0550172252909944, + "learning_rate": 5.968819766892546e-05, + "loss": 0.5356, + "step": 731 + }, + { + "epoch": 2.0123711340206185, + "grad_norm": 0.7212990022902317, + "learning_rate": 5.962117556865489e-05, + "loss": 0.5149, + "step": 732 + }, + { + "epoch": 2.0151202749140893, + "grad_norm": 0.5084983736290086, + "learning_rate": 5.955408084904962e-05, + "loss": 0.5248, + "step": 733 + }, + { + "epoch": 2.01786941580756, + "grad_norm": 0.508857242926003, + "learning_rate": 5.948691375843187e-05, + "loss": 0.5309, + "step": 734 + }, + { + "epoch": 2.020618556701031, + "grad_norm": 0.5715174569632097, + "learning_rate": 5.941967454539171e-05, + "loss": 0.5247, + "step": 735 + }, + { + "epoch": 2.0233676975945016, + "grad_norm": 0.5565322020849539, + "learning_rate": 5.935236345878616e-05, + "loss": 0.5175, + "step": 736 + }, + { + "epoch": 2.0261168384879724, + "grad_norm": 0.5745604042771598, + "learning_rate": 5.928498074773823e-05, + "loss": 0.517, + "step": 737 + }, + { + "epoch": 2.028865979381443, + "grad_norm": 0.500350720005458, + "learning_rate": 5.921752666163598e-05, + "loss": 0.5215, + "step": 738 + }, + { + "epoch": 2.031615120274914, + "grad_norm": 0.4067063681679997, + "learning_rate": 5.9150001450131716e-05, + "loss": 0.5212, + "step": 739 + }, + { + "epoch": 2.0343642611683848, + "grad_norm": 0.4770810449857095, + "learning_rate": 5.9082405363140916e-05, + "loss": 0.521, + "step": 740 + }, + { + "epoch": 2.0371134020618555, + "grad_norm": 0.5487285643494075, + "learning_rate": 5.901473865084141e-05, + "loss": 0.5268, + "step": 741 + }, + { + "epoch": 2.0398625429553263, + "grad_norm": 0.4395320063115494, + "learning_rate": 5.894700156367238e-05, + "loss": 0.5243, + "step": 742 + }, + { + "epoch": 2.042611683848797, + "grad_norm": 0.29982074556969307, + "learning_rate": 5.887919435233352e-05, + "loss": 0.5298, + "step": 743 + }, + { + "epoch": 2.045360824742268, + "grad_norm": 0.41119709116602937, + "learning_rate": 5.881131726778402e-05, + "loss": 0.5145, + "step": 744 + }, + { + "epoch": 2.0481099656357387, + "grad_norm": 0.384320461752947, + "learning_rate": 5.8743370561241715e-05, + "loss": 0.5253, + "step": 745 + }, + { + "epoch": 2.0508591065292094, + "grad_norm": 0.3420792544611106, + "learning_rate": 5.867535448418203e-05, + "loss": 0.5183, + "step": 746 + }, + { + "epoch": 2.05360824742268, + "grad_norm": 0.36905823023095596, + "learning_rate": 5.8607269288337244e-05, + "loss": 0.5208, + "step": 747 + }, + { + "epoch": 2.056357388316151, + "grad_norm": 0.32511574569133067, + "learning_rate": 5.853911522569538e-05, + "loss": 0.5252, + "step": 748 + }, + { + "epoch": 2.059106529209622, + "grad_norm": 0.42000247500887966, + "learning_rate": 5.8470892548499384e-05, + "loss": 0.5182, + "step": 749 + }, + { + "epoch": 2.0618556701030926, + "grad_norm": 0.5749785701633202, + "learning_rate": 5.840260150924609e-05, + "loss": 0.5172, + "step": 750 + }, + { + "epoch": 2.064604810996564, + "grad_norm": 0.6483835292108968, + "learning_rate": 5.833424236068541e-05, + "loss": 0.5336, + "step": 751 + }, + { + "epoch": 2.0673539518900346, + "grad_norm": 0.6231601301759575, + "learning_rate": 5.8265815355819284e-05, + "loss": 0.527, + "step": 752 + }, + { + "epoch": 2.0701030927835053, + "grad_norm": 0.6400482537396563, + "learning_rate": 5.819732074790084e-05, + "loss": 0.5187, + "step": 753 + }, + { + "epoch": 2.072852233676976, + "grad_norm": 0.6112696586750945, + "learning_rate": 5.812875879043336e-05, + "loss": 0.5198, + "step": 754 + }, + { + "epoch": 2.075601374570447, + "grad_norm": 0.5115706970358354, + "learning_rate": 5.8060129737169416e-05, + "loss": 0.5279, + "step": 755 + }, + { + "epoch": 2.0783505154639177, + "grad_norm": 0.4058486284297744, + "learning_rate": 5.799143384210992e-05, + "loss": 0.5268, + "step": 756 + }, + { + "epoch": 2.0810996563573885, + "grad_norm": 0.32047749300200185, + "learning_rate": 5.7922671359503145e-05, + "loss": 0.5252, + "step": 757 + }, + { + "epoch": 2.0838487972508593, + "grad_norm": 0.37004296113308727, + "learning_rate": 5.785384254384382e-05, + "loss": 0.5239, + "step": 758 + }, + { + "epoch": 2.08659793814433, + "grad_norm": 0.4458786395281163, + "learning_rate": 5.7784947649872176e-05, + "loss": 0.5294, + "step": 759 + }, + { + "epoch": 2.089347079037801, + "grad_norm": 0.4272640389491108, + "learning_rate": 5.7715986932573e-05, + "loss": 0.5249, + "step": 760 + }, + { + "epoch": 2.0920962199312716, + "grad_norm": 0.3526739356483497, + "learning_rate": 5.7646960647174706e-05, + "loss": 0.5184, + "step": 761 + }, + { + "epoch": 2.0948453608247424, + "grad_norm": 0.27188265621194263, + "learning_rate": 5.757786904914838e-05, + "loss": 0.5223, + "step": 762 + }, + { + "epoch": 2.097594501718213, + "grad_norm": 0.3457734998315455, + "learning_rate": 5.750871239420681e-05, + "loss": 0.5227, + "step": 763 + }, + { + "epoch": 2.100343642611684, + "grad_norm": 0.4222137704724379, + "learning_rate": 5.7439490938303604e-05, + "loss": 0.5273, + "step": 764 + }, + { + "epoch": 2.1030927835051547, + "grad_norm": 0.45756362105195786, + "learning_rate": 5.7370204937632167e-05, + "loss": 0.5222, + "step": 765 + }, + { + "epoch": 2.1058419243986255, + "grad_norm": 0.3435489204211452, + "learning_rate": 5.7300854648624824e-05, + "loss": 0.5102, + "step": 766 + }, + { + "epoch": 2.1085910652920963, + "grad_norm": 0.24611635871938473, + "learning_rate": 5.723144032795179e-05, + "loss": 0.5221, + "step": 767 + }, + { + "epoch": 2.111340206185567, + "grad_norm": 0.3013065227913591, + "learning_rate": 5.7161962232520304e-05, + "loss": 0.532, + "step": 768 + }, + { + "epoch": 2.114089347079038, + "grad_norm": 0.3470862495086026, + "learning_rate": 5.7092420619473626e-05, + "loss": 0.5258, + "step": 769 + }, + { + "epoch": 2.1168384879725086, + "grad_norm": 0.38943011157017127, + "learning_rate": 5.70228157461901e-05, + "loss": 0.5264, + "step": 770 + }, + { + "epoch": 2.1195876288659794, + "grad_norm": 0.39287644862912424, + "learning_rate": 5.69531478702822e-05, + "loss": 0.5114, + "step": 771 + }, + { + "epoch": 2.12233676975945, + "grad_norm": 0.2916898291723961, + "learning_rate": 5.688341724959557e-05, + "loss": 0.5275, + "step": 772 + }, + { + "epoch": 2.125085910652921, + "grad_norm": 0.22962256743492102, + "learning_rate": 5.681362414220811e-05, + "loss": 0.5214, + "step": 773 + }, + { + "epoch": 2.1278350515463917, + "grad_norm": 0.27944270631912393, + "learning_rate": 5.674376880642893e-05, + "loss": 0.5264, + "step": 774 + }, + { + "epoch": 2.1305841924398625, + "grad_norm": 0.29044286341885533, + "learning_rate": 5.66738515007975e-05, + "loss": 0.5148, + "step": 775 + }, + { + "epoch": 2.1333333333333333, + "grad_norm": 0.28701568805373207, + "learning_rate": 5.6603872484082614e-05, + "loss": 0.5218, + "step": 776 + }, + { + "epoch": 2.136082474226804, + "grad_norm": 0.33569332451853506, + "learning_rate": 5.653383201528151e-05, + "loss": 0.5244, + "step": 777 + }, + { + "epoch": 2.138831615120275, + "grad_norm": 0.2997007948052843, + "learning_rate": 5.6463730353618795e-05, + "loss": 0.5207, + "step": 778 + }, + { + "epoch": 2.1415807560137456, + "grad_norm": 0.21237174047717156, + "learning_rate": 5.6393567758545616e-05, + "loss": 0.5103, + "step": 779 + }, + { + "epoch": 2.1443298969072164, + "grad_norm": 0.2360600600158792, + "learning_rate": 5.63233444897386e-05, + "loss": 0.5293, + "step": 780 + }, + { + "epoch": 2.147079037800687, + "grad_norm": 0.23546258348715238, + "learning_rate": 5.625306080709895e-05, + "loss": 0.526, + "step": 781 + }, + { + "epoch": 2.149828178694158, + "grad_norm": 0.23742640059602604, + "learning_rate": 5.618271697075147e-05, + "loss": 0.5207, + "step": 782 + }, + { + "epoch": 2.1525773195876288, + "grad_norm": 0.2697802862240549, + "learning_rate": 5.611231324104358e-05, + "loss": 0.5185, + "step": 783 + }, + { + "epoch": 2.1553264604810995, + "grad_norm": 0.33962286687598364, + "learning_rate": 5.604184987854437e-05, + "loss": 0.5324, + "step": 784 + }, + { + "epoch": 2.1580756013745703, + "grad_norm": 0.3496574863191211, + "learning_rate": 5.597132714404366e-05, + "loss": 0.5149, + "step": 785 + }, + { + "epoch": 2.160824742268041, + "grad_norm": 0.3002103817256998, + "learning_rate": 5.590074529855099e-05, + "loss": 0.5271, + "step": 786 + }, + { + "epoch": 2.163573883161512, + "grad_norm": 0.33431676858651316, + "learning_rate": 5.583010460329465e-05, + "loss": 0.5304, + "step": 787 + }, + { + "epoch": 2.1663230240549827, + "grad_norm": 0.3298687279509311, + "learning_rate": 5.5759405319720806e-05, + "loss": 0.5167, + "step": 788 + }, + { + "epoch": 2.1690721649484535, + "grad_norm": 0.21499388088920707, + "learning_rate": 5.568864770949237e-05, + "loss": 0.536, + "step": 789 + }, + { + "epoch": 2.1718213058419242, + "grad_norm": 0.26397548444272334, + "learning_rate": 5.5617832034488236e-05, + "loss": 0.5251, + "step": 790 + }, + { + "epoch": 2.174570446735395, + "grad_norm": 0.40340132985540694, + "learning_rate": 5.554695855680209e-05, + "loss": 0.5283, + "step": 791 + }, + { + "epoch": 2.177319587628866, + "grad_norm": 0.4098335019644846, + "learning_rate": 5.547602753874163e-05, + "loss": 0.5238, + "step": 792 + }, + { + "epoch": 2.1800687285223366, + "grad_norm": 0.3531497244525965, + "learning_rate": 5.540503924282746e-05, + "loss": 0.5231, + "step": 793 + }, + { + "epoch": 2.1828178694158074, + "grad_norm": 0.38987792040546504, + "learning_rate": 5.5333993931792224e-05, + "loss": 0.5366, + "step": 794 + }, + { + "epoch": 2.1855670103092786, + "grad_norm": 0.2647915747419579, + "learning_rate": 5.526289186857953e-05, + "loss": 0.5348, + "step": 795 + }, + { + "epoch": 2.1883161512027494, + "grad_norm": 0.26788341550791417, + "learning_rate": 5.5191733316343074e-05, + "loss": 0.5253, + "step": 796 + }, + { + "epoch": 2.19106529209622, + "grad_norm": 0.31976809906665954, + "learning_rate": 5.5120518538445595e-05, + "loss": 0.5225, + "step": 797 + }, + { + "epoch": 2.193814432989691, + "grad_norm": 0.27229292264708094, + "learning_rate": 5.504924779845794e-05, + "loss": 0.5169, + "step": 798 + }, + { + "epoch": 2.1965635738831617, + "grad_norm": 0.3204542340097227, + "learning_rate": 5.4977921360158096e-05, + "loss": 0.5256, + "step": 799 + }, + { + "epoch": 2.1993127147766325, + "grad_norm": 0.33648692429827426, + "learning_rate": 5.490653948753013e-05, + "loss": 0.523, + "step": 800 + }, + { + "epoch": 2.2020618556701033, + "grad_norm": 0.25849474311106246, + "learning_rate": 5.483510244476332e-05, + "loss": 0.5345, + "step": 801 + }, + { + "epoch": 2.204810996563574, + "grad_norm": 0.2777163411097152, + "learning_rate": 5.476361049625115e-05, + "loss": 0.5217, + "step": 802 + }, + { + "epoch": 2.207560137457045, + "grad_norm": 0.259120862015438, + "learning_rate": 5.469206390659028e-05, + "loss": 0.5293, + "step": 803 + }, + { + "epoch": 2.2103092783505156, + "grad_norm": 0.25014178976390844, + "learning_rate": 5.46204629405796e-05, + "loss": 0.5269, + "step": 804 + }, + { + "epoch": 2.2130584192439864, + "grad_norm": 0.223711086431831, + "learning_rate": 5.454880786321928e-05, + "loss": 0.5187, + "step": 805 + }, + { + "epoch": 2.215807560137457, + "grad_norm": 0.2438903572639829, + "learning_rate": 5.447709893970974e-05, + "loss": 0.5209, + "step": 806 + }, + { + "epoch": 2.218556701030928, + "grad_norm": 0.2545794798823825, + "learning_rate": 5.4405336435450684e-05, + "loss": 0.527, + "step": 807 + }, + { + "epoch": 2.2213058419243987, + "grad_norm": 0.18607426816746073, + "learning_rate": 5.4333520616040116e-05, + "loss": 0.5184, + "step": 808 + }, + { + "epoch": 2.2240549828178695, + "grad_norm": 0.29001196046138, + "learning_rate": 5.42616517472734e-05, + "loss": 0.5195, + "step": 809 + }, + { + "epoch": 2.2268041237113403, + "grad_norm": 0.2664754443847999, + "learning_rate": 5.418973009514221e-05, + "loss": 0.5254, + "step": 810 + }, + { + "epoch": 2.229553264604811, + "grad_norm": 0.2763244689974836, + "learning_rate": 5.411775592583358e-05, + "loss": 0.5225, + "step": 811 + }, + { + "epoch": 2.232302405498282, + "grad_norm": 0.2952802637754747, + "learning_rate": 5.4045729505728884e-05, + "loss": 0.5227, + "step": 812 + }, + { + "epoch": 2.2350515463917526, + "grad_norm": 0.23994298354796337, + "learning_rate": 5.397365110140295e-05, + "loss": 0.5305, + "step": 813 + }, + { + "epoch": 2.2378006872852234, + "grad_norm": 0.22190622512563482, + "learning_rate": 5.390152097962295e-05, + "loss": 0.532, + "step": 814 + }, + { + "epoch": 2.240549828178694, + "grad_norm": 0.2958910982821988, + "learning_rate": 5.382933940734747e-05, + "loss": 0.5215, + "step": 815 + }, + { + "epoch": 2.243298969072165, + "grad_norm": 0.3582968358435383, + "learning_rate": 5.375710665172554e-05, + "loss": 0.5206, + "step": 816 + }, + { + "epoch": 2.2460481099656358, + "grad_norm": 0.4048685164576461, + "learning_rate": 5.368482298009559e-05, + "loss": 0.5322, + "step": 817 + }, + { + "epoch": 2.2487972508591065, + "grad_norm": 0.35928125952347756, + "learning_rate": 5.3612488659984534e-05, + "loss": 0.5256, + "step": 818 + }, + { + "epoch": 2.2515463917525773, + "grad_norm": 0.27863313688997904, + "learning_rate": 5.3540103959106696e-05, + "loss": 0.5243, + "step": 819 + }, + { + "epoch": 2.254295532646048, + "grad_norm": 0.3504201948483768, + "learning_rate": 5.3467669145362914e-05, + "loss": 0.5369, + "step": 820 + }, + { + "epoch": 2.257044673539519, + "grad_norm": 0.4374644885930723, + "learning_rate": 5.339518448683945e-05, + "loss": 0.5206, + "step": 821 + }, + { + "epoch": 2.2597938144329897, + "grad_norm": 0.2820104015984913, + "learning_rate": 5.332265025180707e-05, + "loss": 0.5201, + "step": 822 + }, + { + "epoch": 2.2625429553264604, + "grad_norm": 0.3051225307345294, + "learning_rate": 5.325006670872002e-05, + "loss": 0.5248, + "step": 823 + }, + { + "epoch": 2.265292096219931, + "grad_norm": 0.41761093729690574, + "learning_rate": 5.3177434126215034e-05, + "loss": 0.5378, + "step": 824 + }, + { + "epoch": 2.268041237113402, + "grad_norm": 0.3487478544658144, + "learning_rate": 5.3104752773110345e-05, + "loss": 0.5245, + "step": 825 + }, + { + "epoch": 2.270790378006873, + "grad_norm": 0.32194879135312154, + "learning_rate": 5.303202291840471e-05, + "loss": 0.526, + "step": 826 + }, + { + "epoch": 2.2735395189003436, + "grad_norm": 0.27245901521747345, + "learning_rate": 5.295924483127635e-05, + "loss": 0.5184, + "step": 827 + }, + { + "epoch": 2.2762886597938143, + "grad_norm": 0.26804164140224673, + "learning_rate": 5.2886418781082066e-05, + "loss": 0.5201, + "step": 828 + }, + { + "epoch": 2.279037800687285, + "grad_norm": 0.23960902170409884, + "learning_rate": 5.2813545037356086e-05, + "loss": 0.5256, + "step": 829 + }, + { + "epoch": 2.281786941580756, + "grad_norm": 0.23457140639322555, + "learning_rate": 5.2740623869809234e-05, + "loss": 0.5209, + "step": 830 + }, + { + "epoch": 2.2845360824742267, + "grad_norm": 0.27297058783320943, + "learning_rate": 5.2667655548327796e-05, + "loss": 0.5315, + "step": 831 + }, + { + "epoch": 2.2872852233676975, + "grad_norm": 0.28200029368869456, + "learning_rate": 5.259464034297262e-05, + "loss": 0.527, + "step": 832 + }, + { + "epoch": 2.2900343642611682, + "grad_norm": 0.34366587435096785, + "learning_rate": 5.2521578523978046e-05, + "loss": 0.5166, + "step": 833 + }, + { + "epoch": 2.292783505154639, + "grad_norm": 0.33716783962672386, + "learning_rate": 5.2448470361750955e-05, + "loss": 0.5271, + "step": 834 + }, + { + "epoch": 2.29553264604811, + "grad_norm": 0.2883538098562466, + "learning_rate": 5.237531612686973e-05, + "loss": 0.5326, + "step": 835 + }, + { + "epoch": 2.2982817869415806, + "grad_norm": 0.23816842715139308, + "learning_rate": 5.2302116090083274e-05, + "loss": 0.521, + "step": 836 + }, + { + "epoch": 2.3010309278350514, + "grad_norm": 0.2676295491191559, + "learning_rate": 5.222887052231003e-05, + "loss": 0.5307, + "step": 837 + }, + { + "epoch": 2.303780068728522, + "grad_norm": 0.21899952831754288, + "learning_rate": 5.2155579694636916e-05, + "loss": 0.5254, + "step": 838 + }, + { + "epoch": 2.306529209621993, + "grad_norm": 0.2725930609511096, + "learning_rate": 5.208224387831839e-05, + "loss": 0.5243, + "step": 839 + }, + { + "epoch": 2.3092783505154637, + "grad_norm": 0.2670269822849005, + "learning_rate": 5.200886334477541e-05, + "loss": 0.5263, + "step": 840 + }, + { + "epoch": 2.3120274914089345, + "grad_norm": 0.2896259602041522, + "learning_rate": 5.1935438365594424e-05, + "loss": 0.5212, + "step": 841 + }, + { + "epoch": 2.3147766323024053, + "grad_norm": 0.3040525797226736, + "learning_rate": 5.186196921252638e-05, + "loss": 0.5272, + "step": 842 + }, + { + "epoch": 2.317525773195876, + "grad_norm": 0.402527650035371, + "learning_rate": 5.178845615748573e-05, + "loss": 0.5317, + "step": 843 + }, + { + "epoch": 2.320274914089347, + "grad_norm": 0.28589840098643327, + "learning_rate": 5.1714899472549394e-05, + "loss": 0.5302, + "step": 844 + }, + { + "epoch": 2.323024054982818, + "grad_norm": 0.2746598297288049, + "learning_rate": 5.164129942995578e-05, + "loss": 0.5237, + "step": 845 + }, + { + "epoch": 2.325773195876289, + "grad_norm": 0.35895254123689857, + "learning_rate": 5.156765630210375e-05, + "loss": 0.5321, + "step": 846 + }, + { + "epoch": 2.3285223367697596, + "grad_norm": 0.3075206472125044, + "learning_rate": 5.1493970361551666e-05, + "loss": 0.5261, + "step": 847 + }, + { + "epoch": 2.3312714776632304, + "grad_norm": 0.2631726158525417, + "learning_rate": 5.14202418810163e-05, + "loss": 0.5295, + "step": 848 + }, + { + "epoch": 2.334020618556701, + "grad_norm": 0.26253313412596246, + "learning_rate": 5.134647113337187e-05, + "loss": 0.5278, + "step": 849 + }, + { + "epoch": 2.336769759450172, + "grad_norm": 0.27292981975275177, + "learning_rate": 5.127265839164906e-05, + "loss": 0.5174, + "step": 850 + }, + { + "epoch": 2.3395189003436427, + "grad_norm": 0.29830820148659387, + "learning_rate": 5.119880392903396e-05, + "loss": 0.5235, + "step": 851 + }, + { + "epoch": 2.3422680412371135, + "grad_norm": 0.319178636597951, + "learning_rate": 5.112490801886706e-05, + "loss": 0.5249, + "step": 852 + }, + { + "epoch": 2.3450171821305843, + "grad_norm": 0.39509509521849606, + "learning_rate": 5.105097093464225e-05, + "loss": 0.53, + "step": 853 + }, + { + "epoch": 2.347766323024055, + "grad_norm": 0.43481008500350815, + "learning_rate": 5.0976992950005836e-05, + "loss": 0.5222, + "step": 854 + }, + { + "epoch": 2.350515463917526, + "grad_norm": 0.35580076547200273, + "learning_rate": 5.090297433875549e-05, + "loss": 0.5341, + "step": 855 + }, + { + "epoch": 2.3532646048109966, + "grad_norm": 0.27387444348548706, + "learning_rate": 5.082891537483921e-05, + "loss": 0.527, + "step": 856 + }, + { + "epoch": 2.3560137457044674, + "grad_norm": 0.28242295705533316, + "learning_rate": 5.0754816332354384e-05, + "loss": 0.5313, + "step": 857 + }, + { + "epoch": 2.358762886597938, + "grad_norm": 0.23633783144067932, + "learning_rate": 5.0680677485546724e-05, + "loss": 0.5256, + "step": 858 + }, + { + "epoch": 2.361512027491409, + "grad_norm": 0.20611827366274735, + "learning_rate": 5.060649910880926e-05, + "loss": 0.5246, + "step": 859 + }, + { + "epoch": 2.3642611683848798, + "grad_norm": 0.24527029533975683, + "learning_rate": 5.0532281476681295e-05, + "loss": 0.526, + "step": 860 + }, + { + "epoch": 2.3670103092783505, + "grad_norm": 0.2988866489058128, + "learning_rate": 5.0458024863847455e-05, + "loss": 0.5318, + "step": 861 + }, + { + "epoch": 2.3697594501718213, + "grad_norm": 0.3278941979649635, + "learning_rate": 5.038372954513664e-05, + "loss": 0.5208, + "step": 862 + }, + { + "epoch": 2.372508591065292, + "grad_norm": 0.3590976948020781, + "learning_rate": 5.030939579552098e-05, + "loss": 0.5288, + "step": 863 + }, + { + "epoch": 2.375257731958763, + "grad_norm": 0.3257482610923462, + "learning_rate": 5.0235023890114844e-05, + "loss": 0.5342, + "step": 864 + }, + { + "epoch": 2.3780068728522337, + "grad_norm": 0.26418250148170586, + "learning_rate": 5.016061410417384e-05, + "loss": 0.534, + "step": 865 + }, + { + "epoch": 2.3807560137457044, + "grad_norm": 0.25433731862506237, + "learning_rate": 5.008616671309374e-05, + "loss": 0.5349, + "step": 866 + }, + { + "epoch": 2.3835051546391752, + "grad_norm": 0.3098360496587967, + "learning_rate": 5.0011681992409526e-05, + "loss": 0.536, + "step": 867 + }, + { + "epoch": 2.386254295532646, + "grad_norm": 0.48700691274762764, + "learning_rate": 4.993716021779431e-05, + "loss": 0.5281, + "step": 868 + }, + { + "epoch": 2.389003436426117, + "grad_norm": 0.4325007530550783, + "learning_rate": 4.986260166505838e-05, + "loss": 0.5287, + "step": 869 + }, + { + "epoch": 2.3917525773195876, + "grad_norm": 0.40355654494603005, + "learning_rate": 4.978800661014811e-05, + "loss": 0.5211, + "step": 870 + }, + { + "epoch": 2.3945017182130583, + "grad_norm": 0.26544768260770013, + "learning_rate": 4.9713375329145e-05, + "loss": 0.5195, + "step": 871 + }, + { + "epoch": 2.397250859106529, + "grad_norm": 0.2791642651045501, + "learning_rate": 4.963870809826458e-05, + "loss": 0.5265, + "step": 872 + }, + { + "epoch": 2.4, + "grad_norm": 0.3408862865130159, + "learning_rate": 4.9564005193855486e-05, + "loss": 0.5391, + "step": 873 + }, + { + "epoch": 2.4027491408934707, + "grad_norm": 0.3331718549433266, + "learning_rate": 4.9489266892398346e-05, + "loss": 0.5216, + "step": 874 + }, + { + "epoch": 2.4054982817869415, + "grad_norm": 0.33450428391908577, + "learning_rate": 4.9414493470504834e-05, + "loss": 0.5274, + "step": 875 + }, + { + "epoch": 2.4082474226804123, + "grad_norm": 0.324321171570019, + "learning_rate": 4.933968520491654e-05, + "loss": 0.5373, + "step": 876 + }, + { + "epoch": 2.410996563573883, + "grad_norm": 0.28829591882553246, + "learning_rate": 4.9264842372504095e-05, + "loss": 0.5286, + "step": 877 + }, + { + "epoch": 2.413745704467354, + "grad_norm": 0.28746004519536006, + "learning_rate": 4.9189965250265994e-05, + "loss": 0.5199, + "step": 878 + }, + { + "epoch": 2.4164948453608246, + "grad_norm": 0.31445005880135923, + "learning_rate": 4.911505411532769e-05, + "loss": 0.5278, + "step": 879 + }, + { + "epoch": 2.4192439862542954, + "grad_norm": 0.47892737901970533, + "learning_rate": 4.9040109244940505e-05, + "loss": 0.5234, + "step": 880 + }, + { + "epoch": 2.421993127147766, + "grad_norm": 0.45215275967709107, + "learning_rate": 4.896513091648058e-05, + "loss": 0.535, + "step": 881 + }, + { + "epoch": 2.424742268041237, + "grad_norm": 0.34138582396119665, + "learning_rate": 4.889011940744796e-05, + "loss": 0.5164, + "step": 882 + }, + { + "epoch": 2.4274914089347077, + "grad_norm": 0.2682120490670279, + "learning_rate": 4.881507499546545e-05, + "loss": 0.5342, + "step": 883 + }, + { + "epoch": 2.430240549828179, + "grad_norm": 0.2692048784370494, + "learning_rate": 4.873999795827761e-05, + "loss": 0.5344, + "step": 884 + }, + { + "epoch": 2.4329896907216497, + "grad_norm": 0.3048032911190028, + "learning_rate": 4.866488857374979e-05, + "loss": 0.5131, + "step": 885 + }, + { + "epoch": 2.4357388316151205, + "grad_norm": 0.43798907642210033, + "learning_rate": 4.858974711986704e-05, + "loss": 0.5277, + "step": 886 + }, + { + "epoch": 2.4384879725085913, + "grad_norm": 0.4537544665998809, + "learning_rate": 4.851457387473312e-05, + "loss": 0.5272, + "step": 887 + }, + { + "epoch": 2.441237113402062, + "grad_norm": 0.275619974234131, + "learning_rate": 4.843936911656941e-05, + "loss": 0.5229, + "step": 888 + }, + { + "epoch": 2.443986254295533, + "grad_norm": 0.24697116802006067, + "learning_rate": 4.836413312371394e-05, + "loss": 0.5237, + "step": 889 + }, + { + "epoch": 2.4467353951890036, + "grad_norm": 0.2889992360936922, + "learning_rate": 4.828886617462039e-05, + "loss": 0.5214, + "step": 890 + }, + { + "epoch": 2.4494845360824744, + "grad_norm": 0.34236892659433343, + "learning_rate": 4.821356854785695e-05, + "loss": 0.5204, + "step": 891 + }, + { + "epoch": 2.452233676975945, + "grad_norm": 0.26986931392665453, + "learning_rate": 4.8138240522105365e-05, + "loss": 0.5285, + "step": 892 + }, + { + "epoch": 2.454982817869416, + "grad_norm": 0.2386727777106889, + "learning_rate": 4.806288237615989e-05, + "loss": 0.5258, + "step": 893 + }, + { + "epoch": 2.4577319587628867, + "grad_norm": 0.2400107390785469, + "learning_rate": 4.7987494388926275e-05, + "loss": 0.523, + "step": 894 + }, + { + "epoch": 2.4604810996563575, + "grad_norm": 0.30720815238105814, + "learning_rate": 4.7912076839420695e-05, + "loss": 0.5232, + "step": 895 + }, + { + "epoch": 2.4632302405498283, + "grad_norm": 0.25563044453886097, + "learning_rate": 4.7836630006768746e-05, + "loss": 0.5289, + "step": 896 + }, + { + "epoch": 2.465979381443299, + "grad_norm": 0.2618995944120738, + "learning_rate": 4.77611541702044e-05, + "loss": 0.5256, + "step": 897 + }, + { + "epoch": 2.46872852233677, + "grad_norm": 0.2872261399026541, + "learning_rate": 4.768564960906897e-05, + "loss": 0.527, + "step": 898 + }, + { + "epoch": 2.4714776632302407, + "grad_norm": 0.22134537204990035, + "learning_rate": 4.7610116602810096e-05, + "loss": 0.5343, + "step": 899 + }, + { + "epoch": 2.4742268041237114, + "grad_norm": 0.2565389994931295, + "learning_rate": 4.753455543098067e-05, + "loss": 0.5375, + "step": 900 + }, + { + "epoch": 2.476975945017182, + "grad_norm": 0.3740627121098048, + "learning_rate": 4.745896637323785e-05, + "loss": 0.5227, + "step": 901 + }, + { + "epoch": 2.479725085910653, + "grad_norm": 0.3714016967888814, + "learning_rate": 4.7383349709341994e-05, + "loss": 0.5216, + "step": 902 + }, + { + "epoch": 2.4824742268041238, + "grad_norm": 0.2863979507280106, + "learning_rate": 4.730770571915562e-05, + "loss": 0.5289, + "step": 903 + }, + { + "epoch": 2.4852233676975946, + "grad_norm": 0.2639569172110602, + "learning_rate": 4.7232034682642417e-05, + "loss": 0.5207, + "step": 904 + }, + { + "epoch": 2.4879725085910653, + "grad_norm": 0.26506986793397436, + "learning_rate": 4.715633687986613e-05, + "loss": 0.5276, + "step": 905 + }, + { + "epoch": 2.490721649484536, + "grad_norm": 0.3034302978268054, + "learning_rate": 4.7080612590989596e-05, + "loss": 0.5185, + "step": 906 + }, + { + "epoch": 2.493470790378007, + "grad_norm": 0.3242741496552742, + "learning_rate": 4.7004862096273674e-05, + "loss": 0.5292, + "step": 907 + }, + { + "epoch": 2.4962199312714777, + "grad_norm": 0.25198934191589384, + "learning_rate": 4.692908567607621e-05, + "loss": 0.5209, + "step": 908 + }, + { + "epoch": 2.4989690721649485, + "grad_norm": 0.2851823027369076, + "learning_rate": 4.6853283610851004e-05, + "loss": 0.5332, + "step": 909 + }, + { + "epoch": 2.5017182130584192, + "grad_norm": 0.2250586051327087, + "learning_rate": 4.677745618114674e-05, + "loss": 0.5326, + "step": 910 + }, + { + "epoch": 2.50446735395189, + "grad_norm": 0.2137916864449506, + "learning_rate": 4.670160366760606e-05, + "loss": 0.5214, + "step": 911 + }, + { + "epoch": 2.507216494845361, + "grad_norm": 0.23692573516235416, + "learning_rate": 4.6625726350964355e-05, + "loss": 0.5209, + "step": 912 + }, + { + "epoch": 2.5099656357388316, + "grad_norm": 0.2503089565651227, + "learning_rate": 4.654982451204885e-05, + "loss": 0.5228, + "step": 913 + }, + { + "epoch": 2.5127147766323024, + "grad_norm": 0.187246659026728, + "learning_rate": 4.6473898431777535e-05, + "loss": 0.5279, + "step": 914 + }, + { + "epoch": 2.515463917525773, + "grad_norm": 0.22968241737897607, + "learning_rate": 4.6397948391158104e-05, + "loss": 0.5229, + "step": 915 + }, + { + "epoch": 2.518213058419244, + "grad_norm": 0.34560044485812286, + "learning_rate": 4.632197467128695e-05, + "loss": 0.521, + "step": 916 + }, + { + "epoch": 2.5209621993127147, + "grad_norm": 0.36369842042382355, + "learning_rate": 4.624597755334807e-05, + "loss": 0.5222, + "step": 917 + }, + { + "epoch": 2.5237113402061855, + "grad_norm": 0.3273459802424622, + "learning_rate": 4.6169957318612096e-05, + "loss": 0.5269, + "step": 918 + }, + { + "epoch": 2.5264604810996563, + "grad_norm": 0.27046737413984473, + "learning_rate": 4.609391424843519e-05, + "loss": 0.5207, + "step": 919 + }, + { + "epoch": 2.529209621993127, + "grad_norm": 0.23973509685168032, + "learning_rate": 4.601784862425807e-05, + "loss": 0.5305, + "step": 920 + }, + { + "epoch": 2.531958762886598, + "grad_norm": 0.207584971564237, + "learning_rate": 4.594176072760485e-05, + "loss": 0.5391, + "step": 921 + }, + { + "epoch": 2.5347079037800686, + "grad_norm": 0.26389513176009577, + "learning_rate": 4.586565084008217e-05, + "loss": 0.5257, + "step": 922 + }, + { + "epoch": 2.5374570446735394, + "grad_norm": 0.27492964134545905, + "learning_rate": 4.5789519243377975e-05, + "loss": 0.5198, + "step": 923 + }, + { + "epoch": 2.54020618556701, + "grad_norm": 0.2550669423258001, + "learning_rate": 4.571336621926065e-05, + "loss": 0.5161, + "step": 924 + }, + { + "epoch": 2.542955326460481, + "grad_norm": 0.2434323842617188, + "learning_rate": 4.563719204957776e-05, + "loss": 0.5266, + "step": 925 + }, + { + "epoch": 2.5457044673539517, + "grad_norm": 0.2381537603353079, + "learning_rate": 4.5560997016255265e-05, + "loss": 0.5306, + "step": 926 + }, + { + "epoch": 2.5484536082474225, + "grad_norm": 0.25877756360414456, + "learning_rate": 4.548478140129624e-05, + "loss": 0.5279, + "step": 927 + }, + { + "epoch": 2.5512027491408933, + "grad_norm": 0.23404565240810715, + "learning_rate": 4.5408545486779996e-05, + "loss": 0.5325, + "step": 928 + }, + { + "epoch": 2.553951890034364, + "grad_norm": 0.20244068111276767, + "learning_rate": 4.533228955486094e-05, + "loss": 0.5245, + "step": 929 + }, + { + "epoch": 2.556701030927835, + "grad_norm": 0.20414770829038065, + "learning_rate": 4.525601388776758e-05, + "loss": 0.5193, + "step": 930 + }, + { + "epoch": 2.5594501718213056, + "grad_norm": 0.22610091916102834, + "learning_rate": 4.517971876780147e-05, + "loss": 0.5335, + "step": 931 + }, + { + "epoch": 2.5621993127147764, + "grad_norm": 0.2030707662947604, + "learning_rate": 4.5103404477336144e-05, + "loss": 0.5247, + "step": 932 + }, + { + "epoch": 2.564948453608247, + "grad_norm": 0.19565686931462759, + "learning_rate": 4.502707129881609e-05, + "loss": 0.5223, + "step": 933 + }, + { + "epoch": 2.567697594501718, + "grad_norm": 0.22241975043349305, + "learning_rate": 4.495071951475572e-05, + "loss": 0.5289, + "step": 934 + }, + { + "epoch": 2.5704467353951888, + "grad_norm": 0.25405146475942325, + "learning_rate": 4.487434940773828e-05, + "loss": 0.5216, + "step": 935 + }, + { + "epoch": 2.5731958762886595, + "grad_norm": 0.24090685844792697, + "learning_rate": 4.479796126041487e-05, + "loss": 0.5347, + "step": 936 + }, + { + "epoch": 2.5759450171821303, + "grad_norm": 0.20268315657201996, + "learning_rate": 4.472155535550331e-05, + "loss": 0.5174, + "step": 937 + }, + { + "epoch": 2.5786941580756015, + "grad_norm": 0.2164104030727069, + "learning_rate": 4.464513197578717e-05, + "loss": 0.5306, + "step": 938 + }, + { + "epoch": 2.5814432989690723, + "grad_norm": 0.20675687136906504, + "learning_rate": 4.45686914041147e-05, + "loss": 0.5217, + "step": 939 + }, + { + "epoch": 2.584192439862543, + "grad_norm": 0.2495476073768104, + "learning_rate": 4.449223392339776e-05, + "loss": 0.5233, + "step": 940 + }, + { + "epoch": 2.586941580756014, + "grad_norm": 0.2188055318879783, + "learning_rate": 4.44157598166108e-05, + "loss": 0.536, + "step": 941 + }, + { + "epoch": 2.5896907216494847, + "grad_norm": 0.19661959776334886, + "learning_rate": 4.43392693667898e-05, + "loss": 0.5167, + "step": 942 + }, + { + "epoch": 2.5924398625429554, + "grad_norm": 0.21281895098022574, + "learning_rate": 4.426276285703125e-05, + "loss": 0.5241, + "step": 943 + }, + { + "epoch": 2.5951890034364262, + "grad_norm": 0.24612261354981668, + "learning_rate": 4.418624057049106e-05, + "loss": 0.5288, + "step": 944 + }, + { + "epoch": 2.597938144329897, + "grad_norm": 0.26256604053792143, + "learning_rate": 4.410970279038351e-05, + "loss": 0.528, + "step": 945 + }, + { + "epoch": 2.600687285223368, + "grad_norm": 0.2146570261022671, + "learning_rate": 4.4033149799980265e-05, + "loss": 0.5336, + "step": 946 + }, + { + "epoch": 2.6034364261168386, + "grad_norm": 0.20864992098912383, + "learning_rate": 4.395658188260924e-05, + "loss": 0.5289, + "step": 947 + }, + { + "epoch": 2.6061855670103093, + "grad_norm": 0.23475812934644366, + "learning_rate": 4.3879999321653664e-05, + "loss": 0.5294, + "step": 948 + }, + { + "epoch": 2.60893470790378, + "grad_norm": 0.21552160156786482, + "learning_rate": 4.380340240055087e-05, + "loss": 0.5225, + "step": 949 + }, + { + "epoch": 2.611683848797251, + "grad_norm": 0.21451249484405446, + "learning_rate": 4.3726791402791405e-05, + "loss": 0.5309, + "step": 950 + }, + { + "epoch": 2.6144329896907217, + "grad_norm": 0.2228499608780446, + "learning_rate": 4.3650166611917904e-05, + "loss": 0.5218, + "step": 951 + }, + { + "epoch": 2.6171821305841925, + "grad_norm": 0.19761673711898892, + "learning_rate": 4.3573528311524056e-05, + "loss": 0.5263, + "step": 952 + }, + { + "epoch": 2.6199312714776632, + "grad_norm": 0.2717760446302159, + "learning_rate": 4.34968767852535e-05, + "loss": 0.5192, + "step": 953 + }, + { + "epoch": 2.622680412371134, + "grad_norm": 0.2307774736780584, + "learning_rate": 4.3420212316798895e-05, + "loss": 0.5224, + "step": 954 + }, + { + "epoch": 2.625429553264605, + "grad_norm": 0.26499404963083745, + "learning_rate": 4.334353518990075e-05, + "loss": 0.5284, + "step": 955 + }, + { + "epoch": 2.6281786941580756, + "grad_norm": 0.33949301525073333, + "learning_rate": 4.326684568834647e-05, + "loss": 0.5215, + "step": 956 + }, + { + "epoch": 2.6309278350515464, + "grad_norm": 0.33940964847094485, + "learning_rate": 4.31901440959692e-05, + "loss": 0.527, + "step": 957 + }, + { + "epoch": 2.633676975945017, + "grad_norm": 0.3034703314779094, + "learning_rate": 4.311343069664688e-05, + "loss": 0.5315, + "step": 958 + }, + { + "epoch": 2.636426116838488, + "grad_norm": 0.3285064736514596, + "learning_rate": 4.303670577430111e-05, + "loss": 0.5328, + "step": 959 + }, + { + "epoch": 2.6391752577319587, + "grad_norm": 0.3453164976173602, + "learning_rate": 4.295996961289619e-05, + "loss": 0.5323, + "step": 960 + }, + { + "epoch": 2.6419243986254295, + "grad_norm": 0.38444181745132805, + "learning_rate": 4.288322249643799e-05, + "loss": 0.5288, + "step": 961 + }, + { + "epoch": 2.6446735395189003, + "grad_norm": 0.2686311128572819, + "learning_rate": 4.2806464708972905e-05, + "loss": 0.525, + "step": 962 + }, + { + "epoch": 2.647422680412371, + "grad_norm": 0.3232991261053769, + "learning_rate": 4.272969653458685e-05, + "loss": 0.5278, + "step": 963 + }, + { + "epoch": 2.650171821305842, + "grad_norm": 0.32877598170142874, + "learning_rate": 4.265291825740416e-05, + "loss": 0.5176, + "step": 964 + }, + { + "epoch": 2.6529209621993126, + "grad_norm": 0.30844002727607134, + "learning_rate": 4.257613016158661e-05, + "loss": 0.5326, + "step": 965 + }, + { + "epoch": 2.6556701030927834, + "grad_norm": 0.2255639863284666, + "learning_rate": 4.249933253133224e-05, + "loss": 0.5266, + "step": 966 + }, + { + "epoch": 2.658419243986254, + "grad_norm": 0.25200678515207303, + "learning_rate": 4.2422525650874446e-05, + "loss": 0.5292, + "step": 967 + }, + { + "epoch": 2.661168384879725, + "grad_norm": 0.22993658765349623, + "learning_rate": 4.234570980448081e-05, + "loss": 0.5251, + "step": 968 + }, + { + "epoch": 2.6639175257731957, + "grad_norm": 0.2642010038722434, + "learning_rate": 4.226888527645215e-05, + "loss": 0.529, + "step": 969 + }, + { + "epoch": 2.6666666666666665, + "grad_norm": 0.28025388603567647, + "learning_rate": 4.219205235112135e-05, + "loss": 0.523, + "step": 970 + }, + { + "epoch": 2.6694158075601373, + "grad_norm": 0.24286835186859662, + "learning_rate": 4.2115211312852435e-05, + "loss": 0.5241, + "step": 971 + }, + { + "epoch": 2.6721649484536085, + "grad_norm": 0.260911275304071, + "learning_rate": 4.203836244603941e-05, + "loss": 0.5251, + "step": 972 + }, + { + "epoch": 2.6749140893470793, + "grad_norm": 0.23336974876081842, + "learning_rate": 4.1961506035105285e-05, + "loss": 0.5337, + "step": 973 + }, + { + "epoch": 2.67766323024055, + "grad_norm": 0.19717415494962492, + "learning_rate": 4.188464236450098e-05, + "loss": 0.518, + "step": 974 + }, + { + "epoch": 2.680412371134021, + "grad_norm": 0.2635705526316712, + "learning_rate": 4.180777171870427e-05, + "loss": 0.519, + "step": 975 + }, + { + "epoch": 2.6831615120274916, + "grad_norm": 0.2420795496999561, + "learning_rate": 4.173089438221876e-05, + "loss": 0.5329, + "step": 976 + }, + { + "epoch": 2.6859106529209624, + "grad_norm": 0.24604439574458017, + "learning_rate": 4.165401063957283e-05, + "loss": 0.5331, + "step": 977 + }, + { + "epoch": 2.688659793814433, + "grad_norm": 0.1948605941664403, + "learning_rate": 4.157712077531856e-05, + "loss": 0.5221, + "step": 978 + }, + { + "epoch": 2.691408934707904, + "grad_norm": 0.21270792355056445, + "learning_rate": 4.1500225074030654e-05, + "loss": 0.5355, + "step": 979 + }, + { + "epoch": 2.6941580756013748, + "grad_norm": 0.251861479262255, + "learning_rate": 4.142332382030547e-05, + "loss": 0.5336, + "step": 980 + }, + { + "epoch": 2.6969072164948455, + "grad_norm": 0.2112157139724383, + "learning_rate": 4.13464172987599e-05, + "loss": 0.5243, + "step": 981 + }, + { + "epoch": 2.6996563573883163, + "grad_norm": 0.20387549860274823, + "learning_rate": 4.1269505794030296e-05, + "loss": 0.5237, + "step": 982 + }, + { + "epoch": 2.702405498281787, + "grad_norm": 0.2515717159268574, + "learning_rate": 4.119258959077151e-05, + "loss": 0.5319, + "step": 983 + }, + { + "epoch": 2.705154639175258, + "grad_norm": 0.20847504323318589, + "learning_rate": 4.111566897365575e-05, + "loss": 0.52, + "step": 984 + }, + { + "epoch": 2.7079037800687287, + "grad_norm": 0.21911840052600298, + "learning_rate": 4.103874422737157e-05, + "loss": 0.5283, + "step": 985 + }, + { + "epoch": 2.7106529209621995, + "grad_norm": 0.2049343201857368, + "learning_rate": 4.096181563662279e-05, + "loss": 0.5263, + "step": 986 + }, + { + "epoch": 2.7134020618556702, + "grad_norm": 0.1947687795273586, + "learning_rate": 4.088488348612748e-05, + "loss": 0.5239, + "step": 987 + }, + { + "epoch": 2.716151202749141, + "grad_norm": 0.23438600392965853, + "learning_rate": 4.080794806061688e-05, + "loss": 0.5235, + "step": 988 + }, + { + "epoch": 2.718900343642612, + "grad_norm": 0.19851824128921414, + "learning_rate": 4.073100964483435e-05, + "loss": 0.5236, + "step": 989 + }, + { + "epoch": 2.7216494845360826, + "grad_norm": 0.23677872375792366, + "learning_rate": 4.06540685235343e-05, + "loss": 0.5307, + "step": 990 + }, + { + "epoch": 2.7243986254295534, + "grad_norm": 0.19484452355877876, + "learning_rate": 4.057712498148119e-05, + "loss": 0.533, + "step": 991 + }, + { + "epoch": 2.727147766323024, + "grad_norm": 0.24691106342879912, + "learning_rate": 4.05001793034484e-05, + "loss": 0.5282, + "step": 992 + }, + { + "epoch": 2.729896907216495, + "grad_norm": 0.23911573502762684, + "learning_rate": 4.0423231774217246e-05, + "loss": 0.5291, + "step": 993 + }, + { + "epoch": 2.7326460481099657, + "grad_norm": 0.24784394785326042, + "learning_rate": 4.034628267857587e-05, + "loss": 0.5299, + "step": 994 + }, + { + "epoch": 2.7353951890034365, + "grad_norm": 0.2530406886888911, + "learning_rate": 4.026933230131823e-05, + "loss": 0.5155, + "step": 995 + }, + { + "epoch": 2.7381443298969073, + "grad_norm": 0.24815572290601853, + "learning_rate": 4.0192380927243014e-05, + "loss": 0.5331, + "step": 996 + }, + { + "epoch": 2.740893470790378, + "grad_norm": 0.2756774245673258, + "learning_rate": 4.0115428841152637e-05, + "loss": 0.5334, + "step": 997 + }, + { + "epoch": 2.743642611683849, + "grad_norm": 0.24077726898420207, + "learning_rate": 4.0038476327852065e-05, + "loss": 0.5337, + "step": 998 + }, + { + "epoch": 2.7463917525773196, + "grad_norm": 0.20128889819397291, + "learning_rate": 3.9961523672147955e-05, + "loss": 0.5302, + "step": 999 + }, + { + "epoch": 2.7491408934707904, + "grad_norm": 0.20443940819422327, + "learning_rate": 3.988457115884739e-05, + "loss": 0.5284, + "step": 1000 + }, + { + "epoch": 2.751890034364261, + "grad_norm": 0.2208826727313727, + "learning_rate": 3.9807619072757e-05, + "loss": 0.5271, + "step": 1001 + }, + { + "epoch": 2.754639175257732, + "grad_norm": 0.24447959308366699, + "learning_rate": 3.973066769868178e-05, + "loss": 0.5283, + "step": 1002 + }, + { + "epoch": 2.7573883161512027, + "grad_norm": 0.1966957237573264, + "learning_rate": 3.965371732142415e-05, + "loss": 0.5292, + "step": 1003 + }, + { + "epoch": 2.7601374570446735, + "grad_norm": 0.21603427189344282, + "learning_rate": 3.957676822578276e-05, + "loss": 0.5237, + "step": 1004 + }, + { + "epoch": 2.7628865979381443, + "grad_norm": 0.18664057521742963, + "learning_rate": 3.949982069655161e-05, + "loss": 0.5233, + "step": 1005 + }, + { + "epoch": 2.765635738831615, + "grad_norm": 0.2234326703320951, + "learning_rate": 3.942287501851881e-05, + "loss": 0.5256, + "step": 1006 + }, + { + "epoch": 2.768384879725086, + "grad_norm": 0.2278011506232875, + "learning_rate": 3.9345931476465706e-05, + "loss": 0.5279, + "step": 1007 + }, + { + "epoch": 2.7711340206185566, + "grad_norm": 0.19798832981596853, + "learning_rate": 3.9268990355165664e-05, + "loss": 0.5211, + "step": 1008 + }, + { + "epoch": 2.7738831615120274, + "grad_norm": 0.2183804005479682, + "learning_rate": 3.9192051939383126e-05, + "loss": 0.5239, + "step": 1009 + }, + { + "epoch": 2.776632302405498, + "grad_norm": 0.22246588980190424, + "learning_rate": 3.911511651387253e-05, + "loss": 0.527, + "step": 1010 + }, + { + "epoch": 2.779381443298969, + "grad_norm": 0.1780086146278406, + "learning_rate": 3.903818436337722e-05, + "loss": 0.5284, + "step": 1011 + }, + { + "epoch": 2.7821305841924397, + "grad_norm": 0.19604513054901485, + "learning_rate": 3.896125577262845e-05, + "loss": 0.5162, + "step": 1012 + }, + { + "epoch": 2.7848797250859105, + "grad_norm": 0.18794858233130177, + "learning_rate": 3.888433102634425e-05, + "loss": 0.5216, + "step": 1013 + }, + { + "epoch": 2.7876288659793813, + "grad_norm": 0.21435085046499355, + "learning_rate": 3.8807410409228496e-05, + "loss": 0.5305, + "step": 1014 + }, + { + "epoch": 2.790378006872852, + "grad_norm": 0.17410246023922102, + "learning_rate": 3.8730494205969724e-05, + "loss": 0.5312, + "step": 1015 + }, + { + "epoch": 2.793127147766323, + "grad_norm": 0.1822314934502394, + "learning_rate": 3.865358270124013e-05, + "loss": 0.5189, + "step": 1016 + }, + { + "epoch": 2.7958762886597937, + "grad_norm": 0.206717747798629, + "learning_rate": 3.857667617969454e-05, + "loss": 0.5269, + "step": 1017 + }, + { + "epoch": 2.7986254295532644, + "grad_norm": 0.16863448749904908, + "learning_rate": 3.849977492596936e-05, + "loss": 0.5303, + "step": 1018 + }, + { + "epoch": 2.801374570446735, + "grad_norm": 0.23449573918173383, + "learning_rate": 3.8422879224681456e-05, + "loss": 0.5322, + "step": 1019 + }, + { + "epoch": 2.804123711340206, + "grad_norm": 0.17234697630039422, + "learning_rate": 3.8345989360427174e-05, + "loss": 0.5229, + "step": 1020 + }, + { + "epoch": 2.8068728522336768, + "grad_norm": 0.21339720138725765, + "learning_rate": 3.826910561778124e-05, + "loss": 0.5271, + "step": 1021 + }, + { + "epoch": 2.8096219931271476, + "grad_norm": 0.22865589786306695, + "learning_rate": 3.819222828129574e-05, + "loss": 0.5261, + "step": 1022 + }, + { + "epoch": 2.8123711340206183, + "grad_norm": 0.19395333537771012, + "learning_rate": 3.8115357635499045e-05, + "loss": 0.5307, + "step": 1023 + }, + { + "epoch": 2.815120274914089, + "grad_norm": 0.26952248831929015, + "learning_rate": 3.803849396489473e-05, + "loss": 0.5304, + "step": 1024 + }, + { + "epoch": 2.81786941580756, + "grad_norm": 0.2078676429133822, + "learning_rate": 3.7961637553960605e-05, + "loss": 0.521, + "step": 1025 + }, + { + "epoch": 2.8206185567010307, + "grad_norm": 0.2437865253535147, + "learning_rate": 3.788478868714758e-05, + "loss": 0.5284, + "step": 1026 + }, + { + "epoch": 2.8233676975945015, + "grad_norm": 0.23244204464146223, + "learning_rate": 3.780794764887866e-05, + "loss": 0.5347, + "step": 1027 + }, + { + "epoch": 2.8261168384879722, + "grad_norm": 0.1883460034103066, + "learning_rate": 3.7731114723547856e-05, + "loss": 0.5332, + "step": 1028 + }, + { + "epoch": 2.8288659793814435, + "grad_norm": 0.18889412109996112, + "learning_rate": 3.7654290195519195e-05, + "loss": 0.527, + "step": 1029 + }, + { + "epoch": 2.8316151202749142, + "grad_norm": 0.23037525353226912, + "learning_rate": 3.757747434912556e-05, + "loss": 0.5293, + "step": 1030 + }, + { + "epoch": 2.834364261168385, + "grad_norm": 0.21033380977426863, + "learning_rate": 3.750066746866778e-05, + "loss": 0.5226, + "step": 1031 + }, + { + "epoch": 2.837113402061856, + "grad_norm": 0.17225724075961382, + "learning_rate": 3.742386983841341e-05, + "loss": 0.5244, + "step": 1032 + }, + { + "epoch": 2.8398625429553266, + "grad_norm": 0.22879502842160185, + "learning_rate": 3.734708174259585e-05, + "loss": 0.5291, + "step": 1033 + }, + { + "epoch": 2.8426116838487974, + "grad_norm": 0.1859425423671784, + "learning_rate": 3.727030346541317e-05, + "loss": 0.5207, + "step": 1034 + }, + { + "epoch": 2.845360824742268, + "grad_norm": 0.21138789242685446, + "learning_rate": 3.7193535291027115e-05, + "loss": 0.5283, + "step": 1035 + }, + { + "epoch": 2.848109965635739, + "grad_norm": 0.22384899016314974, + "learning_rate": 3.7116777503562016e-05, + "loss": 0.528, + "step": 1036 + }, + { + "epoch": 2.8508591065292097, + "grad_norm": 0.2431535434052746, + "learning_rate": 3.7040030387103815e-05, + "loss": 0.5187, + "step": 1037 + }, + { + "epoch": 2.8536082474226805, + "grad_norm": 0.24399503353207871, + "learning_rate": 3.696329422569889e-05, + "loss": 0.5179, + "step": 1038 + }, + { + "epoch": 2.8563573883161513, + "grad_norm": 0.16921161494449857, + "learning_rate": 3.6886569303353136e-05, + "loss": 0.5292, + "step": 1039 + }, + { + "epoch": 2.859106529209622, + "grad_norm": 0.26728522647583636, + "learning_rate": 3.680985590403082e-05, + "loss": 0.5262, + "step": 1040 + }, + { + "epoch": 2.861855670103093, + "grad_norm": 0.27421229932307095, + "learning_rate": 3.673315431165355e-05, + "loss": 0.5252, + "step": 1041 + }, + { + "epoch": 2.8646048109965636, + "grad_norm": 0.2601105298366, + "learning_rate": 3.665646481009926e-05, + "loss": 0.5203, + "step": 1042 + }, + { + "epoch": 2.8673539518900344, + "grad_norm": 0.2997837262368804, + "learning_rate": 3.657978768320111e-05, + "loss": 0.5266, + "step": 1043 + }, + { + "epoch": 2.870103092783505, + "grad_norm": 0.1739319697716107, + "learning_rate": 3.650312321474651e-05, + "loss": 0.5285, + "step": 1044 + }, + { + "epoch": 2.872852233676976, + "grad_norm": 0.2868154357660894, + "learning_rate": 3.642647168847596e-05, + "loss": 0.5349, + "step": 1045 + }, + { + "epoch": 2.8756013745704467, + "grad_norm": 0.3150816781576196, + "learning_rate": 3.6349833388082096e-05, + "loss": 0.5213, + "step": 1046 + }, + { + "epoch": 2.8783505154639175, + "grad_norm": 0.21703226563550534, + "learning_rate": 3.6273208597208595e-05, + "loss": 0.5302, + "step": 1047 + }, + { + "epoch": 2.8810996563573883, + "grad_norm": 0.24416409278757065, + "learning_rate": 3.619659759944916e-05, + "loss": 0.5249, + "step": 1048 + }, + { + "epoch": 2.883848797250859, + "grad_norm": 0.2807527048682551, + "learning_rate": 3.612000067834636e-05, + "loss": 0.5249, + "step": 1049 + }, + { + "epoch": 2.88659793814433, + "grad_norm": 0.2744380467215706, + "learning_rate": 3.604341811739077e-05, + "loss": 0.5305, + "step": 1050 + }, + { + "epoch": 2.8893470790378006, + "grad_norm": 0.2271292993546667, + "learning_rate": 3.596685020001975e-05, + "loss": 0.5194, + "step": 1051 + }, + { + "epoch": 2.8920962199312714, + "grad_norm": 0.21515311729133313, + "learning_rate": 3.5890297209616507e-05, + "loss": 0.5242, + "step": 1052 + }, + { + "epoch": 2.894845360824742, + "grad_norm": 0.2904407669489142, + "learning_rate": 3.581375942950895e-05, + "loss": 0.5163, + "step": 1053 + }, + { + "epoch": 2.897594501718213, + "grad_norm": 0.23987566400445187, + "learning_rate": 3.5737237142968755e-05, + "loss": 0.5321, + "step": 1054 + }, + { + "epoch": 2.9003436426116838, + "grad_norm": 0.203928269390139, + "learning_rate": 3.56607306332102e-05, + "loss": 0.5203, + "step": 1055 + }, + { + "epoch": 2.9030927835051545, + "grad_norm": 0.26314705579076864, + "learning_rate": 3.558424018338922e-05, + "loss": 0.5267, + "step": 1056 + }, + { + "epoch": 2.9058419243986253, + "grad_norm": 0.22978092693646193, + "learning_rate": 3.5507766076602264e-05, + "loss": 0.5211, + "step": 1057 + }, + { + "epoch": 2.908591065292096, + "grad_norm": 0.1831775606016915, + "learning_rate": 3.5431308595885316e-05, + "loss": 0.5186, + "step": 1058 + }, + { + "epoch": 2.911340206185567, + "grad_norm": 0.2557072250330763, + "learning_rate": 3.535486802421284e-05, + "loss": 0.5204, + "step": 1059 + }, + { + "epoch": 2.9140893470790377, + "grad_norm": 0.2285012181274199, + "learning_rate": 3.5278444644496695e-05, + "loss": 0.5273, + "step": 1060 + }, + { + "epoch": 2.9168384879725084, + "grad_norm": 0.18355449593774995, + "learning_rate": 3.520203873958514e-05, + "loss": 0.531, + "step": 1061 + }, + { + "epoch": 2.9195876288659792, + "grad_norm": 0.21328292549083275, + "learning_rate": 3.512565059226172e-05, + "loss": 0.5229, + "step": 1062 + }, + { + "epoch": 2.9223367697594504, + "grad_norm": 0.21044538382380515, + "learning_rate": 3.5049280485244286e-05, + "loss": 0.5231, + "step": 1063 + }, + { + "epoch": 2.9250859106529212, + "grad_norm": 0.17119562274819797, + "learning_rate": 3.4972928701183925e-05, + "loss": 0.522, + "step": 1064 + }, + { + "epoch": 2.927835051546392, + "grad_norm": 0.22581900383360864, + "learning_rate": 3.489659552266388e-05, + "loss": 0.5246, + "step": 1065 + }, + { + "epoch": 2.930584192439863, + "grad_norm": 0.301379881236332, + "learning_rate": 3.482028123219855e-05, + "loss": 0.529, + "step": 1066 + }, + { + "epoch": 2.9333333333333336, + "grad_norm": 0.295319936733804, + "learning_rate": 3.4743986112232434e-05, + "loss": 0.5289, + "step": 1067 + }, + { + "epoch": 2.9360824742268044, + "grad_norm": 0.2388909601572767, + "learning_rate": 3.466771044513907e-05, + "loss": 0.5252, + "step": 1068 + }, + { + "epoch": 2.938831615120275, + "grad_norm": 0.2678249461923104, + "learning_rate": 3.459145451322002e-05, + "loss": 0.5262, + "step": 1069 + }, + { + "epoch": 2.941580756013746, + "grad_norm": 0.22322825864211238, + "learning_rate": 3.4515218598703765e-05, + "loss": 0.5297, + "step": 1070 + }, + { + "epoch": 2.9443298969072167, + "grad_norm": 0.179417097733591, + "learning_rate": 3.443900298374475e-05, + "loss": 0.5249, + "step": 1071 + }, + { + "epoch": 2.9470790378006875, + "grad_norm": 0.2551012950816274, + "learning_rate": 3.436280795042225e-05, + "loss": 0.5212, + "step": 1072 + }, + { + "epoch": 2.9498281786941583, + "grad_norm": 0.25848561223138594, + "learning_rate": 3.428663378073937e-05, + "loss": 0.528, + "step": 1073 + }, + { + "epoch": 2.952577319587629, + "grad_norm": 0.17137059703558605, + "learning_rate": 3.421048075662203e-05, + "loss": 0.524, + "step": 1074 + }, + { + "epoch": 2.9553264604811, + "grad_norm": 0.21809992041709833, + "learning_rate": 3.413434915991784e-05, + "loss": 0.5225, + "step": 1075 + }, + { + "epoch": 2.9580756013745706, + "grad_norm": 0.19281232602243703, + "learning_rate": 3.4058239272395156e-05, + "loss": 0.5198, + "step": 1076 + }, + { + "epoch": 2.9608247422680414, + "grad_norm": 0.18758042777019787, + "learning_rate": 3.398215137574194e-05, + "loss": 0.5145, + "step": 1077 + }, + { + "epoch": 2.963573883161512, + "grad_norm": 0.24212160366698007, + "learning_rate": 3.390608575156481e-05, + "loss": 0.529, + "step": 1078 + }, + { + "epoch": 2.966323024054983, + "grad_norm": 0.19965723896338422, + "learning_rate": 3.3830042681387904e-05, + "loss": 0.5345, + "step": 1079 + }, + { + "epoch": 2.9690721649484537, + "grad_norm": 0.16936178177221572, + "learning_rate": 3.375402244665194e-05, + "loss": 0.5208, + "step": 1080 + }, + { + "epoch": 2.9718213058419245, + "grad_norm": 0.20843282321759193, + "learning_rate": 3.367802532871306e-05, + "loss": 0.5254, + "step": 1081 + }, + { + "epoch": 2.9745704467353953, + "grad_norm": 0.1837127741621114, + "learning_rate": 3.360205160884191e-05, + "loss": 0.5313, + "step": 1082 + }, + { + "epoch": 2.977319587628866, + "grad_norm": 0.199460376426822, + "learning_rate": 3.352610156822248e-05, + "loss": 0.5292, + "step": 1083 + }, + { + "epoch": 2.980068728522337, + "grad_norm": 0.24820207311148457, + "learning_rate": 3.345017548795116e-05, + "loss": 0.5257, + "step": 1084 + }, + { + "epoch": 2.9828178694158076, + "grad_norm": 0.24493105512545987, + "learning_rate": 3.337427364903565e-05, + "loss": 0.5224, + "step": 1085 + }, + { + "epoch": 2.9855670103092784, + "grad_norm": 0.23786424160796374, + "learning_rate": 3.329839633239395e-05, + "loss": 0.5304, + "step": 1086 + }, + { + "epoch": 2.988316151202749, + "grad_norm": 0.27502699218694276, + "learning_rate": 3.322254381885325e-05, + "loss": 0.5281, + "step": 1087 + }, + { + "epoch": 2.99106529209622, + "grad_norm": 0.27049323273784326, + "learning_rate": 3.314671638914902e-05, + "loss": 0.5301, + "step": 1088 + }, + { + "epoch": 2.9938144329896907, + "grad_norm": 0.19737402579388286, + "learning_rate": 3.307091432392382e-05, + "loss": 0.5288, + "step": 1089 + }, + { + "epoch": 2.9965635738831615, + "grad_norm": 0.26232022578272246, + "learning_rate": 3.299513790372634e-05, + "loss": 0.521, + "step": 1090 + }, + { + "epoch": 2.9993127147766323, + "grad_norm": 0.23892201559966744, + "learning_rate": 3.2919387409010424e-05, + "loss": 0.6184, + "step": 1091 + }, + { + "epoch": 3.002061855670103, + "grad_norm": 0.3594917888207816, + "learning_rate": 3.284366312013388e-05, + "loss": 0.5355, + "step": 1092 + }, + { + "epoch": 3.004810996563574, + "grad_norm": 0.25599834111682307, + "learning_rate": 3.27679653173576e-05, + "loss": 0.4746, + "step": 1093 + }, + { + "epoch": 3.0075601374570446, + "grad_norm": 0.2254148080474178, + "learning_rate": 3.2692294280844374e-05, + "loss": 0.4669, + "step": 1094 + }, + { + "epoch": 3.0103092783505154, + "grad_norm": 0.2506660939835137, + "learning_rate": 3.261665029065801e-05, + "loss": 0.4682, + "step": 1095 + }, + { + "epoch": 3.013058419243986, + "grad_norm": 0.267195831442756, + "learning_rate": 3.254103362676217e-05, + "loss": 0.4715, + "step": 1096 + }, + { + "epoch": 3.015807560137457, + "grad_norm": 0.24179014490822495, + "learning_rate": 3.246544456901934e-05, + "loss": 0.4778, + "step": 1097 + }, + { + "epoch": 3.0185567010309278, + "grad_norm": 0.2329847121267379, + "learning_rate": 3.238988339718992e-05, + "loss": 0.4747, + "step": 1098 + }, + { + "epoch": 3.0213058419243985, + "grad_norm": 0.2043691678048163, + "learning_rate": 3.231435039093104e-05, + "loss": 0.4786, + "step": 1099 + }, + { + "epoch": 3.0240549828178693, + "grad_norm": 0.22745344698265332, + "learning_rate": 3.223884582979561e-05, + "loss": 0.4657, + "step": 1100 + }, + { + "epoch": 3.02680412371134, + "grad_norm": 0.22934476133037818, + "learning_rate": 3.216336999323127e-05, + "loss": 0.4889, + "step": 1101 + }, + { + "epoch": 3.029553264604811, + "grad_norm": 0.21801654584396532, + "learning_rate": 3.208792316057931e-05, + "loss": 0.4784, + "step": 1102 + }, + { + "epoch": 3.0323024054982817, + "grad_norm": 0.2431718990003574, + "learning_rate": 3.201250561107374e-05, + "loss": 0.4797, + "step": 1103 + }, + { + "epoch": 3.0350515463917525, + "grad_norm": 0.26915847727829645, + "learning_rate": 3.1937117623840114e-05, + "loss": 0.4826, + "step": 1104 + }, + { + "epoch": 3.0378006872852232, + "grad_norm": 0.2351108919147438, + "learning_rate": 3.1861759477894656e-05, + "loss": 0.4772, + "step": 1105 + }, + { + "epoch": 3.040549828178694, + "grad_norm": 0.2904532318483722, + "learning_rate": 3.178643145214307e-05, + "loss": 0.4803, + "step": 1106 + }, + { + "epoch": 3.043298969072165, + "grad_norm": 0.19635036055007402, + "learning_rate": 3.171113382537962e-05, + "loss": 0.4701, + "step": 1107 + }, + { + "epoch": 3.0460481099656356, + "grad_norm": 0.22223894507419642, + "learning_rate": 3.1635866876286064e-05, + "loss": 0.4741, + "step": 1108 + }, + { + "epoch": 3.0487972508591064, + "grad_norm": 0.2152706605758009, + "learning_rate": 3.156063088343061e-05, + "loss": 0.4746, + "step": 1109 + }, + { + "epoch": 3.051546391752577, + "grad_norm": 0.17409361205810703, + "learning_rate": 3.1485426125266896e-05, + "loss": 0.478, + "step": 1110 + }, + { + "epoch": 3.054295532646048, + "grad_norm": 0.21539232573809372, + "learning_rate": 3.141025288013296e-05, + "loss": 0.4706, + "step": 1111 + }, + { + "epoch": 3.0570446735395187, + "grad_norm": 0.19308259974547204, + "learning_rate": 3.1335111426250216e-05, + "loss": 0.4777, + "step": 1112 + }, + { + "epoch": 3.0597938144329895, + "grad_norm": 0.2276406440928922, + "learning_rate": 3.1260002041722405e-05, + "loss": 0.4823, + "step": 1113 + }, + { + "epoch": 3.0625429553264603, + "grad_norm": 0.1977395826548686, + "learning_rate": 3.118492500453456e-05, + "loss": 0.4781, + "step": 1114 + }, + { + "epoch": 3.0652920962199315, + "grad_norm": 0.17401988491729847, + "learning_rate": 3.1109880592552046e-05, + "loss": 0.474, + "step": 1115 + }, + { + "epoch": 3.0680412371134023, + "grad_norm": 0.1912931420538338, + "learning_rate": 3.103486908351943e-05, + "loss": 0.4763, + "step": 1116 + }, + { + "epoch": 3.070790378006873, + "grad_norm": 0.16713338636093636, + "learning_rate": 3.0959890755059515e-05, + "loss": 0.4796, + "step": 1117 + }, + { + "epoch": 3.073539518900344, + "grad_norm": 0.20702313839329717, + "learning_rate": 3.088494588467232e-05, + "loss": 0.4778, + "step": 1118 + }, + { + "epoch": 3.0762886597938146, + "grad_norm": 0.1616831225744788, + "learning_rate": 3.081003474973401e-05, + "loss": 0.4756, + "step": 1119 + }, + { + "epoch": 3.0790378006872854, + "grad_norm": 0.24558159371403487, + "learning_rate": 3.0735157627495925e-05, + "loss": 0.4789, + "step": 1120 + }, + { + "epoch": 3.081786941580756, + "grad_norm": 0.17163392192032517, + "learning_rate": 3.0660314795083475e-05, + "loss": 0.4768, + "step": 1121 + }, + { + "epoch": 3.084536082474227, + "grad_norm": 0.21495669147653343, + "learning_rate": 3.0585506529495186e-05, + "loss": 0.4757, + "step": 1122 + }, + { + "epoch": 3.0872852233676977, + "grad_norm": 0.15045704205058852, + "learning_rate": 3.0510733107601664e-05, + "loss": 0.473, + "step": 1123 + }, + { + "epoch": 3.0900343642611685, + "grad_norm": 0.2020087192391845, + "learning_rate": 3.0435994806144524e-05, + "loss": 0.4735, + "step": 1124 + }, + { + "epoch": 3.0927835051546393, + "grad_norm": 0.15940495290357568, + "learning_rate": 3.0361291901735428e-05, + "loss": 0.4709, + "step": 1125 + }, + { + "epoch": 3.09553264604811, + "grad_norm": 0.18904118815363216, + "learning_rate": 3.028662467085501e-05, + "loss": 0.4838, + "step": 1126 + }, + { + "epoch": 3.098281786941581, + "grad_norm": 0.17129808608598124, + "learning_rate": 3.0211993389851893e-05, + "loss": 0.4768, + "step": 1127 + }, + { + "epoch": 3.1010309278350516, + "grad_norm": 0.18416866902814874, + "learning_rate": 3.013739833494162e-05, + "loss": 0.4718, + "step": 1128 + }, + { + "epoch": 3.1037800687285224, + "grad_norm": 0.1705293761247234, + "learning_rate": 3.00628397822057e-05, + "loss": 0.4791, + "step": 1129 + }, + { + "epoch": 3.106529209621993, + "grad_norm": 0.20209221200923272, + "learning_rate": 2.9988318007590494e-05, + "loss": 0.4823, + "step": 1130 + }, + { + "epoch": 3.109278350515464, + "grad_norm": 0.2118645657673755, + "learning_rate": 2.991383328690627e-05, + "loss": 0.4782, + "step": 1131 + }, + { + "epoch": 3.1120274914089348, + "grad_norm": 0.17556823451574127, + "learning_rate": 2.9839385895826175e-05, + "loss": 0.4783, + "step": 1132 + }, + { + "epoch": 3.1147766323024055, + "grad_norm": 0.19795461740770434, + "learning_rate": 2.9764976109885166e-05, + "loss": 0.4818, + "step": 1133 + }, + { + "epoch": 3.1175257731958763, + "grad_norm": 0.16701933368048089, + "learning_rate": 2.9690604204479026e-05, + "loss": 0.476, + "step": 1134 + }, + { + "epoch": 3.120274914089347, + "grad_norm": 0.20618027859459478, + "learning_rate": 2.9616270454863368e-05, + "loss": 0.4798, + "step": 1135 + }, + { + "epoch": 3.123024054982818, + "grad_norm": 0.1431586178012281, + "learning_rate": 2.9541975136152548e-05, + "loss": 0.478, + "step": 1136 + }, + { + "epoch": 3.1257731958762887, + "grad_norm": 0.17793879100992058, + "learning_rate": 2.946771852331873e-05, + "loss": 0.4756, + "step": 1137 + }, + { + "epoch": 3.1285223367697594, + "grad_norm": 0.16681103898774272, + "learning_rate": 2.939350089119077e-05, + "loss": 0.4715, + "step": 1138 + }, + { + "epoch": 3.13127147766323, + "grad_norm": 0.17261193531928995, + "learning_rate": 2.9319322514453286e-05, + "loss": 0.4759, + "step": 1139 + }, + { + "epoch": 3.134020618556701, + "grad_norm": 0.1582640372762462, + "learning_rate": 2.9245183667645626e-05, + "loss": 0.4749, + "step": 1140 + }, + { + "epoch": 3.136769759450172, + "grad_norm": 0.17322436049835935, + "learning_rate": 2.9171084625160797e-05, + "loss": 0.4703, + "step": 1141 + }, + { + "epoch": 3.1395189003436426, + "grad_norm": 0.17367955729692558, + "learning_rate": 2.909702566124452e-05, + "loss": 0.4792, + "step": 1142 + }, + { + "epoch": 3.1422680412371133, + "grad_norm": 0.16722751357988883, + "learning_rate": 2.9023007049994157e-05, + "loss": 0.4717, + "step": 1143 + }, + { + "epoch": 3.145017182130584, + "grad_norm": 0.1728042548273763, + "learning_rate": 2.8949029065357753e-05, + "loss": 0.4818, + "step": 1144 + }, + { + "epoch": 3.147766323024055, + "grad_norm": 0.18069700472194267, + "learning_rate": 2.8875091981132963e-05, + "loss": 0.476, + "step": 1145 + }, + { + "epoch": 3.1505154639175257, + "grad_norm": 0.15820067646687608, + "learning_rate": 2.8801196070966054e-05, + "loss": 0.4666, + "step": 1146 + }, + { + "epoch": 3.1532646048109965, + "grad_norm": 0.15258574864023677, + "learning_rate": 2.8727341608350952e-05, + "loss": 0.4759, + "step": 1147 + }, + { + "epoch": 3.1560137457044672, + "grad_norm": 0.1960418259666068, + "learning_rate": 2.8653528866628132e-05, + "loss": 0.4769, + "step": 1148 + }, + { + "epoch": 3.158762886597938, + "grad_norm": 0.15371731621163243, + "learning_rate": 2.8579758118983716e-05, + "loss": 0.4842, + "step": 1149 + }, + { + "epoch": 3.161512027491409, + "grad_norm": 0.19068409005542658, + "learning_rate": 2.8506029638448347e-05, + "loss": 0.4776, + "step": 1150 + }, + { + "epoch": 3.1642611683848796, + "grad_norm": 0.17886606119360748, + "learning_rate": 2.843234369789625e-05, + "loss": 0.4744, + "step": 1151 + }, + { + "epoch": 3.1670103092783504, + "grad_norm": 0.19710185432909114, + "learning_rate": 2.8358700570044232e-05, + "loss": 0.4813, + "step": 1152 + }, + { + "epoch": 3.169759450171821, + "grad_norm": 0.18041901946396155, + "learning_rate": 2.8285100527450623e-05, + "loss": 0.4833, + "step": 1153 + }, + { + "epoch": 3.172508591065292, + "grad_norm": 0.16581599590174512, + "learning_rate": 2.8211543842514288e-05, + "loss": 0.4739, + "step": 1154 + }, + { + "epoch": 3.1752577319587627, + "grad_norm": 0.15872506391538035, + "learning_rate": 2.8138030787473635e-05, + "loss": 0.4854, + "step": 1155 + }, + { + "epoch": 3.1780068728522335, + "grad_norm": 0.18595712745784665, + "learning_rate": 2.806456163440559e-05, + "loss": 0.4791, + "step": 1156 + }, + { + "epoch": 3.1807560137457043, + "grad_norm": 0.1585604017012495, + "learning_rate": 2.7991136655224602e-05, + "loss": 0.4794, + "step": 1157 + }, + { + "epoch": 3.183505154639175, + "grad_norm": 0.16724675223224328, + "learning_rate": 2.791775612168161e-05, + "loss": 0.484, + "step": 1158 + }, + { + "epoch": 3.1862542955326463, + "grad_norm": 0.1598241225286718, + "learning_rate": 2.784442030536309e-05, + "loss": 0.4816, + "step": 1159 + }, + { + "epoch": 3.189003436426117, + "grad_norm": 0.16901071883801533, + "learning_rate": 2.7771129477689972e-05, + "loss": 0.473, + "step": 1160 + }, + { + "epoch": 3.191752577319588, + "grad_norm": 0.13853638340372265, + "learning_rate": 2.7697883909916732e-05, + "loss": 0.4736, + "step": 1161 + }, + { + "epoch": 3.1945017182130586, + "grad_norm": 0.16091561527631862, + "learning_rate": 2.7624683873130287e-05, + "loss": 0.4805, + "step": 1162 + }, + { + "epoch": 3.1972508591065294, + "grad_norm": 0.14598087437104398, + "learning_rate": 2.755152963824905e-05, + "loss": 0.4835, + "step": 1163 + }, + { + "epoch": 3.2, + "grad_norm": 0.175532838361529, + "learning_rate": 2.7478421476021968e-05, + "loss": 0.4743, + "step": 1164 + }, + { + "epoch": 3.202749140893471, + "grad_norm": 0.1337685529203635, + "learning_rate": 2.740535965702739e-05, + "loss": 0.4734, + "step": 1165 + }, + { + "epoch": 3.2054982817869417, + "grad_norm": 0.1735634483137809, + "learning_rate": 2.7332344451672214e-05, + "loss": 0.4709, + "step": 1166 + }, + { + "epoch": 3.2082474226804125, + "grad_norm": 0.15693421713978023, + "learning_rate": 2.7259376130190783e-05, + "loss": 0.4741, + "step": 1167 + }, + { + "epoch": 3.2109965635738833, + "grad_norm": 0.1330436144853641, + "learning_rate": 2.718645496264392e-05, + "loss": 0.4739, + "step": 1168 + }, + { + "epoch": 3.213745704467354, + "grad_norm": 0.17309959852487516, + "learning_rate": 2.711358121891795e-05, + "loss": 0.4705, + "step": 1169 + }, + { + "epoch": 3.216494845360825, + "grad_norm": 0.1642984807751716, + "learning_rate": 2.704075516872366e-05, + "loss": 0.4726, + "step": 1170 + }, + { + "epoch": 3.2192439862542956, + "grad_norm": 0.14722892838781462, + "learning_rate": 2.6967977081595304e-05, + "loss": 0.4775, + "step": 1171 + }, + { + "epoch": 3.2219931271477664, + "grad_norm": 0.16887717979369643, + "learning_rate": 2.6895247226889672e-05, + "loss": 0.4738, + "step": 1172 + }, + { + "epoch": 3.224742268041237, + "grad_norm": 0.17049853048281954, + "learning_rate": 2.682256587378498e-05, + "loss": 0.4719, + "step": 1173 + }, + { + "epoch": 3.227491408934708, + "grad_norm": 0.15138446686573354, + "learning_rate": 2.6749933291279994e-05, + "loss": 0.4856, + "step": 1174 + }, + { + "epoch": 3.2302405498281788, + "grad_norm": 0.16148723512701157, + "learning_rate": 2.6677349748192934e-05, + "loss": 0.4762, + "step": 1175 + }, + { + "epoch": 3.2329896907216495, + "grad_norm": 0.14912677660800724, + "learning_rate": 2.6604815513160556e-05, + "loss": 0.4653, + "step": 1176 + }, + { + "epoch": 3.2357388316151203, + "grad_norm": 0.13862596201937702, + "learning_rate": 2.6532330854637086e-05, + "loss": 0.4743, + "step": 1177 + }, + { + "epoch": 3.238487972508591, + "grad_norm": 0.15927565474045197, + "learning_rate": 2.645989604089331e-05, + "loss": 0.4746, + "step": 1178 + }, + { + "epoch": 3.241237113402062, + "grad_norm": 0.17434623830674337, + "learning_rate": 2.638751134001549e-05, + "loss": 0.48, + "step": 1179 + }, + { + "epoch": 3.2439862542955327, + "grad_norm": 0.1447492759368682, + "learning_rate": 2.6315177019904423e-05, + "loss": 0.4781, + "step": 1180 + }, + { + "epoch": 3.2467353951890034, + "grad_norm": 0.16927963210867344, + "learning_rate": 2.624289334827448e-05, + "loss": 0.4704, + "step": 1181 + }, + { + "epoch": 3.2494845360824742, + "grad_norm": 0.1665258090219738, + "learning_rate": 2.6170660592652545e-05, + "loss": 0.4781, + "step": 1182 + }, + { + "epoch": 3.252233676975945, + "grad_norm": 0.15855482153444905, + "learning_rate": 2.609847902037706e-05, + "loss": 0.4743, + "step": 1183 + }, + { + "epoch": 3.254982817869416, + "grad_norm": 0.17971555626146168, + "learning_rate": 2.6026348898597057e-05, + "loss": 0.4803, + "step": 1184 + }, + { + "epoch": 3.2577319587628866, + "grad_norm": 0.2816535496494584, + "learning_rate": 2.5954270494271116e-05, + "loss": 0.4761, + "step": 1185 + }, + { + "epoch": 3.2604810996563574, + "grad_norm": 0.23781070038647936, + "learning_rate": 2.588224407416645e-05, + "loss": 0.4729, + "step": 1186 + }, + { + "epoch": 3.263230240549828, + "grad_norm": 0.23075459880557803, + "learning_rate": 2.581026990485781e-05, + "loss": 0.4764, + "step": 1187 + }, + { + "epoch": 3.265979381443299, + "grad_norm": 0.14086275865969622, + "learning_rate": 2.5738348252726607e-05, + "loss": 0.4809, + "step": 1188 + }, + { + "epoch": 3.2687285223367697, + "grad_norm": 0.18978376603435007, + "learning_rate": 2.566647938395989e-05, + "loss": 0.4817, + "step": 1189 + }, + { + "epoch": 3.2714776632302405, + "grad_norm": 0.16648197235336828, + "learning_rate": 2.559466356454933e-05, + "loss": 0.4703, + "step": 1190 + }, + { + "epoch": 3.2742268041237113, + "grad_norm": 0.17731919695127468, + "learning_rate": 2.5522901060290272e-05, + "loss": 0.4757, + "step": 1191 + }, + { + "epoch": 3.276975945017182, + "grad_norm": 0.22533634711616554, + "learning_rate": 2.545119213678072e-05, + "loss": 0.4779, + "step": 1192 + }, + { + "epoch": 3.279725085910653, + "grad_norm": 0.29341047419363175, + "learning_rate": 2.53795370594204e-05, + "loss": 0.4809, + "step": 1193 + }, + { + "epoch": 3.2824742268041236, + "grad_norm": 0.17396435349550476, + "learning_rate": 2.530793609340974e-05, + "loss": 0.4796, + "step": 1194 + }, + { + "epoch": 3.2852233676975944, + "grad_norm": 0.20383849713641408, + "learning_rate": 2.523638950374886e-05, + "loss": 0.4721, + "step": 1195 + }, + { + "epoch": 3.287972508591065, + "grad_norm": 0.16601271855097022, + "learning_rate": 2.5164897555236686e-05, + "loss": 0.4804, + "step": 1196 + }, + { + "epoch": 3.290721649484536, + "grad_norm": 0.30470051683065796, + "learning_rate": 2.509346051246988e-05, + "loss": 0.4792, + "step": 1197 + }, + { + "epoch": 3.2934707903780067, + "grad_norm": 0.2812298427453436, + "learning_rate": 2.5022078639841918e-05, + "loss": 0.4782, + "step": 1198 + }, + { + "epoch": 3.2962199312714775, + "grad_norm": 0.1597994542210946, + "learning_rate": 2.495075220154206e-05, + "loss": 0.4823, + "step": 1199 + }, + { + "epoch": 3.2989690721649483, + "grad_norm": 0.3095999568945906, + "learning_rate": 2.4879481461554405e-05, + "loss": 0.4762, + "step": 1200 + }, + { + "epoch": 3.301718213058419, + "grad_norm": 0.24822013052998354, + "learning_rate": 2.4808266683656932e-05, + "loss": 0.4805, + "step": 1201 + }, + { + "epoch": 3.30446735395189, + "grad_norm": 0.19304267454670382, + "learning_rate": 2.473710813142049e-05, + "loss": 0.4709, + "step": 1202 + }, + { + "epoch": 3.3072164948453606, + "grad_norm": 0.22511180187157134, + "learning_rate": 2.4666006068207793e-05, + "loss": 0.4791, + "step": 1203 + }, + { + "epoch": 3.3099656357388314, + "grad_norm": 0.17597702188477718, + "learning_rate": 2.4594960757172555e-05, + "loss": 0.4825, + "step": 1204 + }, + { + "epoch": 3.312714776632302, + "grad_norm": 0.18231300214622423, + "learning_rate": 2.4523972461258386e-05, + "loss": 0.4794, + "step": 1205 + }, + { + "epoch": 3.315463917525773, + "grad_norm": 0.25489199823717157, + "learning_rate": 2.445304144319792e-05, + "loss": 0.4795, + "step": 1206 + }, + { + "epoch": 3.3182130584192437, + "grad_norm": 0.15105215480831594, + "learning_rate": 2.4382167965511774e-05, + "loss": 0.4902, + "step": 1207 + }, + { + "epoch": 3.320962199312715, + "grad_norm": 0.2799009342897516, + "learning_rate": 2.431135229050763e-05, + "loss": 0.4818, + "step": 1208 + }, + { + "epoch": 3.3237113402061857, + "grad_norm": 0.21951158609310864, + "learning_rate": 2.4240594680279204e-05, + "loss": 0.479, + "step": 1209 + }, + { + "epoch": 3.3264604810996565, + "grad_norm": 0.19395691144843719, + "learning_rate": 2.416989539670536e-05, + "loss": 0.479, + "step": 1210 + }, + { + "epoch": 3.3292096219931273, + "grad_norm": 0.17094551246350398, + "learning_rate": 2.409925470144903e-05, + "loss": 0.4808, + "step": 1211 + }, + { + "epoch": 3.331958762886598, + "grad_norm": 0.23964152078624051, + "learning_rate": 2.402867285595635e-05, + "loss": 0.4764, + "step": 1212 + }, + { + "epoch": 3.334707903780069, + "grad_norm": 0.2782973993431327, + "learning_rate": 2.3958150121455638e-05, + "loss": 0.4811, + "step": 1213 + }, + { + "epoch": 3.3374570446735397, + "grad_norm": 0.1640795235166955, + "learning_rate": 2.3887686758956425e-05, + "loss": 0.4793, + "step": 1214 + }, + { + "epoch": 3.3402061855670104, + "grad_norm": 0.3070072205373487, + "learning_rate": 2.3817283029248536e-05, + "loss": 0.475, + "step": 1215 + }, + { + "epoch": 3.342955326460481, + "grad_norm": 0.23275102863915093, + "learning_rate": 2.3746939192901052e-05, + "loss": 0.479, + "step": 1216 + }, + { + "epoch": 3.345704467353952, + "grad_norm": 0.23028817088895479, + "learning_rate": 2.3676655510261402e-05, + "loss": 0.4735, + "step": 1217 + }, + { + "epoch": 3.3484536082474228, + "grad_norm": 0.320766094957754, + "learning_rate": 2.3606432241454398e-05, + "loss": 0.4877, + "step": 1218 + }, + { + "epoch": 3.3512027491408936, + "grad_norm": 0.2518366256607952, + "learning_rate": 2.353626964638122e-05, + "loss": 0.4733, + "step": 1219 + }, + { + "epoch": 3.3539518900343643, + "grad_norm": 0.1904395352124866, + "learning_rate": 2.34661679847185e-05, + "loss": 0.4895, + "step": 1220 + }, + { + "epoch": 3.356701030927835, + "grad_norm": 0.3279385370633572, + "learning_rate": 2.3396127515917392e-05, + "loss": 0.4777, + "step": 1221 + }, + { + "epoch": 3.359450171821306, + "grad_norm": 0.20836286167731985, + "learning_rate": 2.3326148499202518e-05, + "loss": 0.4765, + "step": 1222 + }, + { + "epoch": 3.3621993127147767, + "grad_norm": 0.22895782643501986, + "learning_rate": 2.325623119357107e-05, + "loss": 0.4883, + "step": 1223 + }, + { + "epoch": 3.3649484536082475, + "grad_norm": 0.29237657945287854, + "learning_rate": 2.3186375857791896e-05, + "loss": 0.483, + "step": 1224 + }, + { + "epoch": 3.3676975945017182, + "grad_norm": 0.18465319808299466, + "learning_rate": 2.3116582750404427e-05, + "loss": 0.472, + "step": 1225 + }, + { + "epoch": 3.370446735395189, + "grad_norm": 0.20550527495867124, + "learning_rate": 2.304685212971781e-05, + "loss": 0.477, + "step": 1226 + }, + { + "epoch": 3.37319587628866, + "grad_norm": 0.2692822007314835, + "learning_rate": 2.2977184253809913e-05, + "loss": 0.4763, + "step": 1227 + }, + { + "epoch": 3.3759450171821306, + "grad_norm": 0.1708056032464783, + "learning_rate": 2.2907579380526387e-05, + "loss": 0.48, + "step": 1228 + }, + { + "epoch": 3.3786941580756014, + "grad_norm": 0.368849523723143, + "learning_rate": 2.2838037767479713e-05, + "loss": 0.4797, + "step": 1229 + }, + { + "epoch": 3.381443298969072, + "grad_norm": 0.21646292002270615, + "learning_rate": 2.2768559672048218e-05, + "loss": 0.474, + "step": 1230 + }, + { + "epoch": 3.384192439862543, + "grad_norm": 0.26373109002232037, + "learning_rate": 2.2699145351375186e-05, + "loss": 0.4803, + "step": 1231 + }, + { + "epoch": 3.3869415807560137, + "grad_norm": 0.2843624854578763, + "learning_rate": 2.262979506236784e-05, + "loss": 0.4826, + "step": 1232 + }, + { + "epoch": 3.3896907216494845, + "grad_norm": 0.19833041217898748, + "learning_rate": 2.256050906169641e-05, + "loss": 0.491, + "step": 1233 + }, + { + "epoch": 3.3924398625429553, + "grad_norm": 0.2800244352558394, + "learning_rate": 2.2491287605793204e-05, + "loss": 0.4707, + "step": 1234 + }, + { + "epoch": 3.395189003436426, + "grad_norm": 0.18325778358692565, + "learning_rate": 2.2422130950851643e-05, + "loss": 0.4744, + "step": 1235 + }, + { + "epoch": 3.397938144329897, + "grad_norm": 0.3008410533335858, + "learning_rate": 2.235303935282531e-05, + "loss": 0.4846, + "step": 1236 + }, + { + "epoch": 3.4006872852233676, + "grad_norm": 0.19194203147930453, + "learning_rate": 2.2284013067427025e-05, + "loss": 0.4718, + "step": 1237 + }, + { + "epoch": 3.4034364261168384, + "grad_norm": 0.34898181416099955, + "learning_rate": 2.2215052350127834e-05, + "loss": 0.484, + "step": 1238 + }, + { + "epoch": 3.406185567010309, + "grad_norm": 0.2356425803363679, + "learning_rate": 2.214615745615619e-05, + "loss": 0.4774, + "step": 1239 + }, + { + "epoch": 3.40893470790378, + "grad_norm": 0.21466246650328097, + "learning_rate": 2.207732864049686e-05, + "loss": 0.4812, + "step": 1240 + }, + { + "epoch": 3.4116838487972507, + "grad_norm": 0.26377375192350677, + "learning_rate": 2.200856615789009e-05, + "loss": 0.4912, + "step": 1241 + }, + { + "epoch": 3.4144329896907215, + "grad_norm": 0.15695872082264978, + "learning_rate": 2.1939870262830577e-05, + "loss": 0.4715, + "step": 1242 + }, + { + "epoch": 3.4171821305841923, + "grad_norm": 0.28736994518322967, + "learning_rate": 2.187124120956666e-05, + "loss": 0.4811, + "step": 1243 + }, + { + "epoch": 3.419931271477663, + "grad_norm": 0.24847231178584678, + "learning_rate": 2.1802679252099184e-05, + "loss": 0.4784, + "step": 1244 + }, + { + "epoch": 3.422680412371134, + "grad_norm": 0.17957114598770363, + "learning_rate": 2.1734184644180715e-05, + "loss": 0.4706, + "step": 1245 + }, + { + "epoch": 3.4254295532646046, + "grad_norm": 0.3386000360336158, + "learning_rate": 2.16657576393146e-05, + "loss": 0.4764, + "step": 1246 + }, + { + "epoch": 3.4281786941580754, + "grad_norm": 0.14262617741372224, + "learning_rate": 2.1597398490753917e-05, + "loss": 0.4645, + "step": 1247 + }, + { + "epoch": 3.4309278350515466, + "grad_norm": 0.2371298714043362, + "learning_rate": 2.152910745150063e-05, + "loss": 0.4778, + "step": 1248 + }, + { + "epoch": 3.4336769759450174, + "grad_norm": 0.15007549750939145, + "learning_rate": 2.1460884774304614e-05, + "loss": 0.481, + "step": 1249 + }, + { + "epoch": 3.436426116838488, + "grad_norm": 0.2666157528473445, + "learning_rate": 2.1392730711662755e-05, + "loss": 0.473, + "step": 1250 + }, + { + "epoch": 3.439175257731959, + "grad_norm": 0.19370621015791017, + "learning_rate": 2.1324645515817988e-05, + "loss": 0.4916, + "step": 1251 + }, + { + "epoch": 3.4419243986254298, + "grad_norm": 0.1363626153767201, + "learning_rate": 2.125662943875832e-05, + "loss": 0.4783, + "step": 1252 + }, + { + "epoch": 3.4446735395189005, + "grad_norm": 0.1558927651589561, + "learning_rate": 2.1188682732215978e-05, + "loss": 0.4863, + "step": 1253 + }, + { + "epoch": 3.4474226804123713, + "grad_norm": 0.1878171719282548, + "learning_rate": 2.1120805647666484e-05, + "loss": 0.483, + "step": 1254 + }, + { + "epoch": 3.450171821305842, + "grad_norm": 0.1706433093556562, + "learning_rate": 2.1052998436327624e-05, + "loss": 0.4766, + "step": 1255 + }, + { + "epoch": 3.452920962199313, + "grad_norm": 0.14594078534672697, + "learning_rate": 2.0985261349158606e-05, + "loss": 0.4701, + "step": 1256 + }, + { + "epoch": 3.4556701030927837, + "grad_norm": 0.2713964023439289, + "learning_rate": 2.0917594636859084e-05, + "loss": 0.4734, + "step": 1257 + }, + { + "epoch": 3.4584192439862544, + "grad_norm": 0.20301631548067403, + "learning_rate": 2.084999854986829e-05, + "loss": 0.4664, + "step": 1258 + }, + { + "epoch": 3.4611683848797252, + "grad_norm": 0.21017329606796895, + "learning_rate": 2.0782473338364034e-05, + "loss": 0.4801, + "step": 1259 + }, + { + "epoch": 3.463917525773196, + "grad_norm": 0.2024174569645721, + "learning_rate": 2.0715019252261786e-05, + "loss": 0.473, + "step": 1260 + }, + { + "epoch": 3.466666666666667, + "grad_norm": 0.13513087820592787, + "learning_rate": 2.0647636541213843e-05, + "loss": 0.479, + "step": 1261 + }, + { + "epoch": 3.4694158075601376, + "grad_norm": 0.15323818419427804, + "learning_rate": 2.0580325454608294e-05, + "loss": 0.479, + "step": 1262 + }, + { + "epoch": 3.4721649484536083, + "grad_norm": 0.1883180981275938, + "learning_rate": 2.051308624156815e-05, + "loss": 0.4797, + "step": 1263 + }, + { + "epoch": 3.474914089347079, + "grad_norm": 0.18535688822536886, + "learning_rate": 2.0445919150950387e-05, + "loss": 0.4828, + "step": 1264 + }, + { + "epoch": 3.47766323024055, + "grad_norm": 0.1735738790195007, + "learning_rate": 2.0378824431345116e-05, + "loss": 0.4752, + "step": 1265 + }, + { + "epoch": 3.4804123711340207, + "grad_norm": 0.25380262888772, + "learning_rate": 2.0311802331074543e-05, + "loss": 0.479, + "step": 1266 + }, + { + "epoch": 3.4831615120274915, + "grad_norm": 0.17442121730064505, + "learning_rate": 2.024485309819213e-05, + "loss": 0.4739, + "step": 1267 + }, + { + "epoch": 3.4859106529209622, + "grad_norm": 0.284320944684907, + "learning_rate": 2.01779769804816e-05, + "loss": 0.4843, + "step": 1268 + }, + { + "epoch": 3.488659793814433, + "grad_norm": 0.1625892440951168, + "learning_rate": 2.011117422545616e-05, + "loss": 0.4763, + "step": 1269 + }, + { + "epoch": 3.491408934707904, + "grad_norm": 0.28275001646900066, + "learning_rate": 2.004444508035743e-05, + "loss": 0.4699, + "step": 1270 + }, + { + "epoch": 3.4941580756013746, + "grad_norm": 0.17127917891669825, + "learning_rate": 1.9977789792154615e-05, + "loss": 0.4882, + "step": 1271 + }, + { + "epoch": 3.4969072164948454, + "grad_norm": 0.155941741872769, + "learning_rate": 1.991120860754356e-05, + "loss": 0.4692, + "step": 1272 + }, + { + "epoch": 3.499656357388316, + "grad_norm": 0.15459720154148457, + "learning_rate": 1.984470177294588e-05, + "loss": 0.4841, + "step": 1273 + }, + { + "epoch": 3.502405498281787, + "grad_norm": 0.2196341305641022, + "learning_rate": 1.9778269534507987e-05, + "loss": 0.4713, + "step": 1274 + }, + { + "epoch": 3.5051546391752577, + "grad_norm": 0.22478563840945578, + "learning_rate": 1.971191213810021e-05, + "loss": 0.4892, + "step": 1275 + }, + { + "epoch": 3.5079037800687285, + "grad_norm": 0.1739449107806518, + "learning_rate": 1.964562982931588e-05, + "loss": 0.4765, + "step": 1276 + }, + { + "epoch": 3.5106529209621993, + "grad_norm": 0.27214067372219086, + "learning_rate": 1.9579422853470436e-05, + "loss": 0.4815, + "step": 1277 + }, + { + "epoch": 3.51340206185567, + "grad_norm": 0.17400788077272825, + "learning_rate": 1.95132914556005e-05, + "loss": 0.4727, + "step": 1278 + }, + { + "epoch": 3.516151202749141, + "grad_norm": 0.19337030209527964, + "learning_rate": 1.9447235880462947e-05, + "loss": 0.4804, + "step": 1279 + }, + { + "epoch": 3.5189003436426116, + "grad_norm": 0.1514665359437226, + "learning_rate": 1.938125637253409e-05, + "loss": 0.4729, + "step": 1280 + }, + { + "epoch": 3.5216494845360824, + "grad_norm": 0.27810047698132523, + "learning_rate": 1.9315353176008655e-05, + "loss": 0.478, + "step": 1281 + }, + { + "epoch": 3.524398625429553, + "grad_norm": 0.2052552676995858, + "learning_rate": 1.9249526534798965e-05, + "loss": 0.4825, + "step": 1282 + }, + { + "epoch": 3.527147766323024, + "grad_norm": 0.1638869338895521, + "learning_rate": 1.9183776692533977e-05, + "loss": 0.4754, + "step": 1283 + }, + { + "epoch": 3.5298969072164947, + "grad_norm": 0.1623552411012285, + "learning_rate": 1.9118103892558438e-05, + "loss": 0.4721, + "step": 1284 + }, + { + "epoch": 3.5326460481099655, + "grad_norm": 0.2765823979801074, + "learning_rate": 1.9052508377931945e-05, + "loss": 0.4721, + "step": 1285 + }, + { + "epoch": 3.5353951890034363, + "grad_norm": 0.20787425872060086, + "learning_rate": 1.8986990391428056e-05, + "loss": 0.4721, + "step": 1286 + }, + { + "epoch": 3.538144329896907, + "grad_norm": 0.23103945726058686, + "learning_rate": 1.8921550175533384e-05, + "loss": 0.4804, + "step": 1287 + }, + { + "epoch": 3.540893470790378, + "grad_norm": 0.14819667238986017, + "learning_rate": 1.885618797244674e-05, + "loss": 0.4781, + "step": 1288 + }, + { + "epoch": 3.5436426116838486, + "grad_norm": 0.34832469601339244, + "learning_rate": 1.8790904024078166e-05, + "loss": 0.4853, + "step": 1289 + }, + { + "epoch": 3.5463917525773194, + "grad_norm": 0.17762443199441277, + "learning_rate": 1.8725698572048096e-05, + "loss": 0.4687, + "step": 1290 + }, + { + "epoch": 3.54914089347079, + "grad_norm": 0.220110309410904, + "learning_rate": 1.866057185768644e-05, + "loss": 0.4722, + "step": 1291 + }, + { + "epoch": 3.551890034364261, + "grad_norm": 0.1568411202207363, + "learning_rate": 1.8595524122031695e-05, + "loss": 0.477, + "step": 1292 + }, + { + "epoch": 3.5546391752577318, + "grad_norm": 0.1487231862533628, + "learning_rate": 1.8530555605830048e-05, + "loss": 0.4812, + "step": 1293 + }, + { + "epoch": 3.5573883161512025, + "grad_norm": 0.27212850028975205, + "learning_rate": 1.8465666549534487e-05, + "loss": 0.4716, + "step": 1294 + }, + { + "epoch": 3.5601374570446733, + "grad_norm": 0.26360242336986534, + "learning_rate": 1.840085719330394e-05, + "loss": 0.4794, + "step": 1295 + }, + { + "epoch": 3.562886597938144, + "grad_norm": 0.20568964864290215, + "learning_rate": 1.8336127777002336e-05, + "loss": 0.4808, + "step": 1296 + }, + { + "epoch": 3.565635738831615, + "grad_norm": 0.282051987566169, + "learning_rate": 1.8271478540197735e-05, + "loss": 0.4761, + "step": 1297 + }, + { + "epoch": 3.5683848797250857, + "grad_norm": 0.16289176663769198, + "learning_rate": 1.8206909722161444e-05, + "loss": 0.4774, + "step": 1298 + }, + { + "epoch": 3.5711340206185564, + "grad_norm": 0.2575145343236498, + "learning_rate": 1.814242156186718e-05, + "loss": 0.4795, + "step": 1299 + }, + { + "epoch": 3.5738831615120272, + "grad_norm": 0.1486906599375562, + "learning_rate": 1.8078014297990104e-05, + "loss": 0.4824, + "step": 1300 + }, + { + "epoch": 3.576632302405498, + "grad_norm": 0.1501909622248065, + "learning_rate": 1.8013688168905946e-05, + "loss": 0.4804, + "step": 1301 + }, + { + "epoch": 3.5793814432989692, + "grad_norm": 0.15230793704069007, + "learning_rate": 1.7949443412690182e-05, + "loss": 0.4751, + "step": 1302 + }, + { + "epoch": 3.58213058419244, + "grad_norm": 0.16635368754820404, + "learning_rate": 1.788528026711715e-05, + "loss": 0.4806, + "step": 1303 + }, + { + "epoch": 3.584879725085911, + "grad_norm": 0.1347568536197528, + "learning_rate": 1.78211989696591e-05, + "loss": 0.4723, + "step": 1304 + }, + { + "epoch": 3.5876288659793816, + "grad_norm": 0.19251735027744715, + "learning_rate": 1.7757199757485363e-05, + "loss": 0.4796, + "step": 1305 + }, + { + "epoch": 3.5903780068728524, + "grad_norm": 0.14511887693354616, + "learning_rate": 1.7693282867461464e-05, + "loss": 0.4875, + "step": 1306 + }, + { + "epoch": 3.593127147766323, + "grad_norm": 0.21823490371477205, + "learning_rate": 1.762944853614828e-05, + "loss": 0.4831, + "step": 1307 + }, + { + "epoch": 3.595876288659794, + "grad_norm": 0.14789140619202265, + "learning_rate": 1.756569699980111e-05, + "loss": 0.4719, + "step": 1308 + }, + { + "epoch": 3.5986254295532647, + "grad_norm": 0.1537689543052464, + "learning_rate": 1.7502028494368777e-05, + "loss": 0.4775, + "step": 1309 + }, + { + "epoch": 3.6013745704467355, + "grad_norm": 0.16438166123180578, + "learning_rate": 1.743844325549289e-05, + "loss": 0.4778, + "step": 1310 + }, + { + "epoch": 3.6041237113402063, + "grad_norm": 0.1309598142497098, + "learning_rate": 1.737494151850682e-05, + "loss": 0.4799, + "step": 1311 + }, + { + "epoch": 3.606872852233677, + "grad_norm": 0.13501349366551338, + "learning_rate": 1.731152351843492e-05, + "loss": 0.4778, + "step": 1312 + }, + { + "epoch": 3.609621993127148, + "grad_norm": 0.12271151015542832, + "learning_rate": 1.7248189489991594e-05, + "loss": 0.4804, + "step": 1313 + }, + { + "epoch": 3.6123711340206186, + "grad_norm": 0.1579684841055174, + "learning_rate": 1.7184939667580512e-05, + "loss": 0.4756, + "step": 1314 + }, + { + "epoch": 3.6151202749140894, + "grad_norm": 0.11550758392122117, + "learning_rate": 1.712177428529365e-05, + "loss": 0.4813, + "step": 1315 + }, + { + "epoch": 3.61786941580756, + "grad_norm": 0.1446519336760662, + "learning_rate": 1.7058693576910495e-05, + "loss": 0.4761, + "step": 1316 + }, + { + "epoch": 3.620618556701031, + "grad_norm": 0.1387283733233433, + "learning_rate": 1.6995697775897097e-05, + "loss": 0.4855, + "step": 1317 + }, + { + "epoch": 3.6233676975945017, + "grad_norm": 0.12450900277305127, + "learning_rate": 1.6932787115405318e-05, + "loss": 0.4768, + "step": 1318 + }, + { + "epoch": 3.6261168384879725, + "grad_norm": 0.15118692423046173, + "learning_rate": 1.6869961828271892e-05, + "loss": 0.4795, + "step": 1319 + }, + { + "epoch": 3.6288659793814433, + "grad_norm": 0.1275953061475887, + "learning_rate": 1.6807222147017558e-05, + "loss": 0.4844, + "step": 1320 + }, + { + "epoch": 3.631615120274914, + "grad_norm": 0.12357741430968726, + "learning_rate": 1.6744568303846237e-05, + "loss": 0.4718, + "step": 1321 + }, + { + "epoch": 3.634364261168385, + "grad_norm": 0.13324292495631146, + "learning_rate": 1.6682000530644177e-05, + "loss": 0.4828, + "step": 1322 + }, + { + "epoch": 3.6371134020618556, + "grad_norm": 0.12727908902135235, + "learning_rate": 1.6619519058979044e-05, + "loss": 0.4731, + "step": 1323 + }, + { + "epoch": 3.6398625429553264, + "grad_norm": 0.14031695435202224, + "learning_rate": 1.6557124120099108e-05, + "loss": 0.4723, + "step": 1324 + }, + { + "epoch": 3.642611683848797, + "grad_norm": 0.1238822517218841, + "learning_rate": 1.6494815944932376e-05, + "loss": 0.4767, + "step": 1325 + }, + { + "epoch": 3.645360824742268, + "grad_norm": 0.14383655055562664, + "learning_rate": 1.6432594764085738e-05, + "loss": 0.4744, + "step": 1326 + }, + { + "epoch": 3.6481099656357387, + "grad_norm": 0.13074212391556805, + "learning_rate": 1.6370460807844096e-05, + "loss": 0.4719, + "step": 1327 + }, + { + "epoch": 3.6508591065292095, + "grad_norm": 0.1408486433749572, + "learning_rate": 1.6308414306169545e-05, + "loss": 0.4792, + "step": 1328 + }, + { + "epoch": 3.6536082474226803, + "grad_norm": 0.1425866228634816, + "learning_rate": 1.6246455488700522e-05, + "loss": 0.4759, + "step": 1329 + }, + { + "epoch": 3.656357388316151, + "grad_norm": 0.13452543614110554, + "learning_rate": 1.61845845847509e-05, + "loss": 0.4825, + "step": 1330 + }, + { + "epoch": 3.659106529209622, + "grad_norm": 0.123447733182089, + "learning_rate": 1.6122801823309195e-05, + "loss": 0.4839, + "step": 1331 + }, + { + "epoch": 3.6618556701030927, + "grad_norm": 0.141262084894595, + "learning_rate": 1.6061107433037707e-05, + "loss": 0.4803, + "step": 1332 + }, + { + "epoch": 3.6646048109965634, + "grad_norm": 0.16335769694535568, + "learning_rate": 1.5999501642271664e-05, + "loss": 0.491, + "step": 1333 + }, + { + "epoch": 3.667353951890034, + "grad_norm": 0.13731203188970648, + "learning_rate": 1.593798467901837e-05, + "loss": 0.4764, + "step": 1334 + }, + { + "epoch": 3.670103092783505, + "grad_norm": 0.14344242124627954, + "learning_rate": 1.5876556770956394e-05, + "loss": 0.4776, + "step": 1335 + }, + { + "epoch": 3.672852233676976, + "grad_norm": 0.14287518229033996, + "learning_rate": 1.581521814543467e-05, + "loss": 0.4729, + "step": 1336 + }, + { + "epoch": 3.675601374570447, + "grad_norm": 0.13919137334446374, + "learning_rate": 1.575396902947175e-05, + "loss": 0.4803, + "step": 1337 + }, + { + "epoch": 3.678350515463918, + "grad_norm": 0.1332949080422285, + "learning_rate": 1.5692809649754855e-05, + "loss": 0.4732, + "step": 1338 + }, + { + "epoch": 3.6810996563573886, + "grad_norm": 0.15251079258406947, + "learning_rate": 1.563174023263909e-05, + "loss": 0.4721, + "step": 1339 + }, + { + "epoch": 3.6838487972508593, + "grad_norm": 0.12884291771360323, + "learning_rate": 1.557076100414663e-05, + "loss": 0.4771, + "step": 1340 + }, + { + "epoch": 3.68659793814433, + "grad_norm": 0.14077765554999186, + "learning_rate": 1.5509872189965826e-05, + "loss": 0.4765, + "step": 1341 + }, + { + "epoch": 3.689347079037801, + "grad_norm": 0.12097139782888651, + "learning_rate": 1.5449074015450416e-05, + "loss": 0.4816, + "step": 1342 + }, + { + "epoch": 3.6920962199312717, + "grad_norm": 0.12394128179144817, + "learning_rate": 1.5388366705618656e-05, + "loss": 0.4719, + "step": 1343 + }, + { + "epoch": 3.6948453608247425, + "grad_norm": 0.12482467627547454, + "learning_rate": 1.5327750485152546e-05, + "loss": 0.4763, + "step": 1344 + }, + { + "epoch": 3.6975945017182132, + "grad_norm": 0.12282644196525853, + "learning_rate": 1.5267225578396923e-05, + "loss": 0.4773, + "step": 1345 + }, + { + "epoch": 3.700343642611684, + "grad_norm": 0.12030648328844293, + "learning_rate": 1.5206792209358665e-05, + "loss": 0.486, + "step": 1346 + }, + { + "epoch": 3.703092783505155, + "grad_norm": 0.1242904271715769, + "learning_rate": 1.5146450601705866e-05, + "loss": 0.4788, + "step": 1347 + }, + { + "epoch": 3.7058419243986256, + "grad_norm": 0.12938487720592814, + "learning_rate": 1.5086200978767051e-05, + "loss": 0.482, + "step": 1348 + }, + { + "epoch": 3.7085910652920964, + "grad_norm": 0.1181010757249468, + "learning_rate": 1.5026043563530213e-05, + "loss": 0.4817, + "step": 1349 + }, + { + "epoch": 3.711340206185567, + "grad_norm": 0.12542244186954696, + "learning_rate": 1.4965978578642152e-05, + "loss": 0.4829, + "step": 1350 + }, + { + "epoch": 3.714089347079038, + "grad_norm": 0.1301001364291388, + "learning_rate": 1.4906006246407531e-05, + "loss": 0.4801, + "step": 1351 + }, + { + "epoch": 3.7168384879725087, + "grad_norm": 0.11752586529006614, + "learning_rate": 1.4846126788788153e-05, + "loss": 0.4824, + "step": 1352 + }, + { + "epoch": 3.7195876288659795, + "grad_norm": 0.1250323838865372, + "learning_rate": 1.4786340427402034e-05, + "loss": 0.4735, + "step": 1353 + }, + { + "epoch": 3.7223367697594503, + "grad_norm": 0.11762297628012565, + "learning_rate": 1.472664738352266e-05, + "loss": 0.4738, + "step": 1354 + }, + { + "epoch": 3.725085910652921, + "grad_norm": 0.1295998839343961, + "learning_rate": 1.4667047878078115e-05, + "loss": 0.4797, + "step": 1355 + }, + { + "epoch": 3.727835051546392, + "grad_norm": 0.12318111875882015, + "learning_rate": 1.460754213165036e-05, + "loss": 0.4749, + "step": 1356 + }, + { + "epoch": 3.7305841924398626, + "grad_norm": 0.1372696026783743, + "learning_rate": 1.4548130364474253e-05, + "loss": 0.4828, + "step": 1357 + }, + { + "epoch": 3.7333333333333334, + "grad_norm": 0.11717512263035454, + "learning_rate": 1.4488812796436871e-05, + "loss": 0.4672, + "step": 1358 + }, + { + "epoch": 3.736082474226804, + "grad_norm": 0.13561308791703292, + "learning_rate": 1.4429589647076693e-05, + "loss": 0.4786, + "step": 1359 + }, + { + "epoch": 3.738831615120275, + "grad_norm": 0.12422009286706612, + "learning_rate": 1.4370461135582691e-05, + "loss": 0.4851, + "step": 1360 + }, + { + "epoch": 3.7415807560137457, + "grad_norm": 0.11832155403134417, + "learning_rate": 1.4311427480793597e-05, + "loss": 0.4766, + "step": 1361 + }, + { + "epoch": 3.7443298969072165, + "grad_norm": 0.12867243461981434, + "learning_rate": 1.4252488901197064e-05, + "loss": 0.4724, + "step": 1362 + }, + { + "epoch": 3.7470790378006873, + "grad_norm": 0.1313834080728087, + "learning_rate": 1.4193645614928886e-05, + "loss": 0.4714, + "step": 1363 + }, + { + "epoch": 3.749828178694158, + "grad_norm": 0.129183260330846, + "learning_rate": 1.4134897839772155e-05, + "loss": 0.4809, + "step": 1364 + }, + { + "epoch": 3.752577319587629, + "grad_norm": 0.12556934848044315, + "learning_rate": 1.4076245793156477e-05, + "loss": 0.4769, + "step": 1365 + }, + { + "epoch": 3.7553264604810996, + "grad_norm": 0.12147447097394133, + "learning_rate": 1.4017689692157114e-05, + "loss": 0.4821, + "step": 1366 + }, + { + "epoch": 3.7580756013745704, + "grad_norm": 0.13051373867173985, + "learning_rate": 1.3959229753494312e-05, + "loss": 0.472, + "step": 1367 + }, + { + "epoch": 3.760824742268041, + "grad_norm": 0.1294424563340561, + "learning_rate": 1.3900866193532347e-05, + "loss": 0.4822, + "step": 1368 + }, + { + "epoch": 3.763573883161512, + "grad_norm": 0.1319274091460674, + "learning_rate": 1.3842599228278819e-05, + "loss": 0.4808, + "step": 1369 + }, + { + "epoch": 3.7663230240549828, + "grad_norm": 0.14039131963053228, + "learning_rate": 1.3784429073383794e-05, + "loss": 0.4878, + "step": 1370 + }, + { + "epoch": 3.7690721649484535, + "grad_norm": 0.15726224881254797, + "learning_rate": 1.3726355944139091e-05, + "loss": 0.4858, + "step": 1371 + }, + { + "epoch": 3.7718213058419243, + "grad_norm": 0.1414080124007642, + "learning_rate": 1.3668380055477383e-05, + "loss": 0.4843, + "step": 1372 + }, + { + "epoch": 3.774570446735395, + "grad_norm": 0.14834309789500114, + "learning_rate": 1.3610501621971466e-05, + "loss": 0.4817, + "step": 1373 + }, + { + "epoch": 3.777319587628866, + "grad_norm": 0.17635910401341698, + "learning_rate": 1.3552720857833449e-05, + "loss": 0.4791, + "step": 1374 + }, + { + "epoch": 3.7800687285223367, + "grad_norm": 0.12924512390652784, + "learning_rate": 1.3495037976913947e-05, + "loss": 0.4786, + "step": 1375 + }, + { + "epoch": 3.7828178694158074, + "grad_norm": 0.15614756263557175, + "learning_rate": 1.3437453192701315e-05, + "loss": 0.4805, + "step": 1376 + }, + { + "epoch": 3.7855670103092782, + "grad_norm": 0.13352957029862794, + "learning_rate": 1.3379966718320839e-05, + "loss": 0.476, + "step": 1377 + }, + { + "epoch": 3.788316151202749, + "grad_norm": 0.14136665810120994, + "learning_rate": 1.3322578766533968e-05, + "loss": 0.473, + "step": 1378 + }, + { + "epoch": 3.79106529209622, + "grad_norm": 0.1527290850046059, + "learning_rate": 1.3265289549737501e-05, + "loss": 0.4826, + "step": 1379 + }, + { + "epoch": 3.7938144329896906, + "grad_norm": 0.13142189843945631, + "learning_rate": 1.32080992799628e-05, + "loss": 0.4807, + "step": 1380 + }, + { + "epoch": 3.7965635738831613, + "grad_norm": 0.13227897526677368, + "learning_rate": 1.3151008168875037e-05, + "loss": 0.47, + "step": 1381 + }, + { + "epoch": 3.799312714776632, + "grad_norm": 0.1292721449879669, + "learning_rate": 1.309401642777238e-05, + "loss": 0.4747, + "step": 1382 + }, + { + "epoch": 3.802061855670103, + "grad_norm": 0.11592891333671732, + "learning_rate": 1.3037124267585228e-05, + "loss": 0.4795, + "step": 1383 + }, + { + "epoch": 3.8048109965635737, + "grad_norm": 0.12251003727405267, + "learning_rate": 1.298033189887541e-05, + "loss": 0.4841, + "step": 1384 + }, + { + "epoch": 3.8075601374570445, + "grad_norm": 0.11989246216482814, + "learning_rate": 1.2923639531835433e-05, + "loss": 0.4688, + "step": 1385 + }, + { + "epoch": 3.8103092783505152, + "grad_norm": 0.12756974223833703, + "learning_rate": 1.28670473762877e-05, + "loss": 0.4746, + "step": 1386 + }, + { + "epoch": 3.813058419243986, + "grad_norm": 0.1180466527779192, + "learning_rate": 1.2810555641683706e-05, + "loss": 0.4779, + "step": 1387 + }, + { + "epoch": 3.815807560137457, + "grad_norm": 0.11529174071752626, + "learning_rate": 1.2754164537103289e-05, + "loss": 0.4721, + "step": 1388 + }, + { + "epoch": 3.8185567010309276, + "grad_norm": 0.11876628327578556, + "learning_rate": 1.2697874271253844e-05, + "loss": 0.4719, + "step": 1389 + }, + { + "epoch": 3.8213058419243984, + "grad_norm": 0.1212315987997364, + "learning_rate": 1.2641685052469556e-05, + "loss": 0.4901, + "step": 1390 + }, + { + "epoch": 3.824054982817869, + "grad_norm": 0.11790905831570075, + "learning_rate": 1.2585597088710637e-05, + "loss": 0.4755, + "step": 1391 + }, + { + "epoch": 3.82680412371134, + "grad_norm": 0.11385644404016117, + "learning_rate": 1.252961058756252e-05, + "loss": 0.4844, + "step": 1392 + }, + { + "epoch": 3.829553264604811, + "grad_norm": 0.11530386738039233, + "learning_rate": 1.2473725756235165e-05, + "loss": 0.4756, + "step": 1393 + }, + { + "epoch": 3.832302405498282, + "grad_norm": 0.12379904902641005, + "learning_rate": 1.2417942801562201e-05, + "loss": 0.4702, + "step": 1394 + }, + { + "epoch": 3.8350515463917527, + "grad_norm": 0.12225777044844523, + "learning_rate": 1.236226193000023e-05, + "loss": 0.4734, + "step": 1395 + }, + { + "epoch": 3.8378006872852235, + "grad_norm": 0.12161313208268175, + "learning_rate": 1.2306683347628012e-05, + "loss": 0.4809, + "step": 1396 + }, + { + "epoch": 3.8405498281786943, + "grad_norm": 0.13058946315743358, + "learning_rate": 1.2251207260145775e-05, + "loss": 0.4787, + "step": 1397 + }, + { + "epoch": 3.843298969072165, + "grad_norm": 0.11523497518057338, + "learning_rate": 1.2195833872874343e-05, + "loss": 0.4723, + "step": 1398 + }, + { + "epoch": 3.846048109965636, + "grad_norm": 0.13253286053473112, + "learning_rate": 1.2140563390754485e-05, + "loss": 0.4765, + "step": 1399 + }, + { + "epoch": 3.8487972508591066, + "grad_norm": 0.12147651056509255, + "learning_rate": 1.208539601834608e-05, + "loss": 0.4715, + "step": 1400 + }, + { + "epoch": 3.8515463917525774, + "grad_norm": 0.12552379756828472, + "learning_rate": 1.203033195982744e-05, + "loss": 0.4785, + "step": 1401 + }, + { + "epoch": 3.854295532646048, + "grad_norm": 0.12156559670865116, + "learning_rate": 1.1975371418994457e-05, + "loss": 0.4865, + "step": 1402 + }, + { + "epoch": 3.857044673539519, + "grad_norm": 0.13699296116050447, + "learning_rate": 1.1920514599259908e-05, + "loss": 0.4754, + "step": 1403 + }, + { + "epoch": 3.8597938144329897, + "grad_norm": 0.14619958749768325, + "learning_rate": 1.1865761703652691e-05, + "loss": 0.4896, + "step": 1404 + }, + { + "epoch": 3.8625429553264605, + "grad_norm": 0.11322022044471265, + "learning_rate": 1.1811112934817115e-05, + "loss": 0.4809, + "step": 1405 + }, + { + "epoch": 3.8652920962199313, + "grad_norm": 0.12792071165635296, + "learning_rate": 1.1756568495012032e-05, + "loss": 0.4777, + "step": 1406 + }, + { + "epoch": 3.868041237113402, + "grad_norm": 0.12840707488542488, + "learning_rate": 1.1702128586110204e-05, + "loss": 0.4704, + "step": 1407 + }, + { + "epoch": 3.870790378006873, + "grad_norm": 0.12236846928122713, + "learning_rate": 1.164779340959755e-05, + "loss": 0.4734, + "step": 1408 + }, + { + "epoch": 3.8735395189003436, + "grad_norm": 0.12828972122388524, + "learning_rate": 1.1593563166572315e-05, + "loss": 0.4728, + "step": 1409 + }, + { + "epoch": 3.8762886597938144, + "grad_norm": 0.11963041349113263, + "learning_rate": 1.1539438057744405e-05, + "loss": 0.4798, + "step": 1410 + }, + { + "epoch": 3.879037800687285, + "grad_norm": 0.11889755394099005, + "learning_rate": 1.1485418283434591e-05, + "loss": 0.4765, + "step": 1411 + }, + { + "epoch": 3.881786941580756, + "grad_norm": 0.1220924766559957, + "learning_rate": 1.1431504043573845e-05, + "loss": 0.4792, + "step": 1412 + }, + { + "epoch": 3.8845360824742268, + "grad_norm": 0.10991608717200368, + "learning_rate": 1.1377695537702506e-05, + "loss": 0.4758, + "step": 1413 + }, + { + "epoch": 3.8872852233676976, + "grad_norm": 0.1185758525402704, + "learning_rate": 1.1323992964969568e-05, + "loss": 0.4764, + "step": 1414 + }, + { + "epoch": 3.8900343642611683, + "grad_norm": 0.1285483122063387, + "learning_rate": 1.1270396524132016e-05, + "loss": 0.4726, + "step": 1415 + }, + { + "epoch": 3.892783505154639, + "grad_norm": 0.11257438797691342, + "learning_rate": 1.1216906413553996e-05, + "loss": 0.4853, + "step": 1416 + }, + { + "epoch": 3.89553264604811, + "grad_norm": 0.1265940578758485, + "learning_rate": 1.1163522831206119e-05, + "loss": 0.4853, + "step": 1417 + }, + { + "epoch": 3.8982817869415807, + "grad_norm": 0.12742498500479152, + "learning_rate": 1.1110245974664741e-05, + "loss": 0.4848, + "step": 1418 + }, + { + "epoch": 3.9010309278350515, + "grad_norm": 0.11699707249742979, + "learning_rate": 1.1057076041111197e-05, + "loss": 0.4728, + "step": 1419 + }, + { + "epoch": 3.9037800687285222, + "grad_norm": 0.12027686818075006, + "learning_rate": 1.100401322733113e-05, + "loss": 0.4697, + "step": 1420 + }, + { + "epoch": 3.906529209621993, + "grad_norm": 0.12028837431785216, + "learning_rate": 1.0951057729713698e-05, + "loss": 0.4735, + "step": 1421 + }, + { + "epoch": 3.909278350515464, + "grad_norm": 0.11366367375993695, + "learning_rate": 1.0898209744250852e-05, + "loss": 0.4798, + "step": 1422 + }, + { + "epoch": 3.9120274914089346, + "grad_norm": 0.12271108404053097, + "learning_rate": 1.0845469466536689e-05, + "loss": 0.4767, + "step": 1423 + }, + { + "epoch": 3.9147766323024054, + "grad_norm": 0.11169948199394023, + "learning_rate": 1.0792837091766635e-05, + "loss": 0.4773, + "step": 1424 + }, + { + "epoch": 3.917525773195876, + "grad_norm": 0.12185308524300838, + "learning_rate": 1.0740312814736766e-05, + "loss": 0.481, + "step": 1425 + }, + { + "epoch": 3.920274914089347, + "grad_norm": 0.11498767957292035, + "learning_rate": 1.0687896829843068e-05, + "loss": 0.4815, + "step": 1426 + }, + { + "epoch": 3.923024054982818, + "grad_norm": 0.10760167455092437, + "learning_rate": 1.0635589331080771e-05, + "loss": 0.4716, + "step": 1427 + }, + { + "epoch": 3.925773195876289, + "grad_norm": 0.11652282613458785, + "learning_rate": 1.0583390512043557e-05, + "loss": 0.4716, + "step": 1428 + }, + { + "epoch": 3.9285223367697597, + "grad_norm": 0.11095078761860352, + "learning_rate": 1.0531300565922873e-05, + "loss": 0.4773, + "step": 1429 + }, + { + "epoch": 3.9312714776632305, + "grad_norm": 0.10567263314079245, + "learning_rate": 1.0479319685507234e-05, + "loss": 0.4818, + "step": 1430 + }, + { + "epoch": 3.9340206185567013, + "grad_norm": 0.11290338751098791, + "learning_rate": 1.04274480631815e-05, + "loss": 0.4736, + "step": 1431 + }, + { + "epoch": 3.936769759450172, + "grad_norm": 0.11142975595031625, + "learning_rate": 1.0375685890926142e-05, + "loss": 0.4759, + "step": 1432 + }, + { + "epoch": 3.939518900343643, + "grad_norm": 0.12580032061681007, + "learning_rate": 1.0324033360316559e-05, + "loss": 0.4787, + "step": 1433 + }, + { + "epoch": 3.9422680412371136, + "grad_norm": 0.10935541303336552, + "learning_rate": 1.0272490662522347e-05, + "loss": 0.4771, + "step": 1434 + }, + { + "epoch": 3.9450171821305844, + "grad_norm": 0.11261196339303021, + "learning_rate": 1.0221057988306637e-05, + "loss": 0.4711, + "step": 1435 + }, + { + "epoch": 3.947766323024055, + "grad_norm": 0.11321244000910012, + "learning_rate": 1.0169735528025319e-05, + "loss": 0.4851, + "step": 1436 + }, + { + "epoch": 3.950515463917526, + "grad_norm": 0.11089992172259956, + "learning_rate": 1.011852347162639e-05, + "loss": 0.4751, + "step": 1437 + }, + { + "epoch": 3.9532646048109967, + "grad_norm": 0.11766021872483373, + "learning_rate": 1.0067422008649221e-05, + "loss": 0.4797, + "step": 1438 + }, + { + "epoch": 3.9560137457044675, + "grad_norm": 0.10628228690278965, + "learning_rate": 1.0016431328223887e-05, + "loss": 0.4782, + "step": 1439 + }, + { + "epoch": 3.9587628865979383, + "grad_norm": 0.11378648141949942, + "learning_rate": 9.96555161907044e-06, + "loss": 0.4766, + "step": 1440 + }, + { + "epoch": 3.961512027491409, + "grad_norm": 0.11864702105913649, + "learning_rate": 9.91478306949821e-06, + "loss": 0.4786, + "step": 1441 + }, + { + "epoch": 3.96426116838488, + "grad_norm": 0.12126170054240307, + "learning_rate": 9.864125867405146e-06, + "loss": 0.4797, + "step": 1442 + }, + { + "epoch": 3.9670103092783506, + "grad_norm": 0.11758636840862939, + "learning_rate": 9.813580200277063e-06, + "loss": 0.4768, + "step": 1443 + }, + { + "epoch": 3.9697594501718214, + "grad_norm": 0.11275441649847655, + "learning_rate": 9.76314625518699e-06, + "loss": 0.4668, + "step": 1444 + }, + { + "epoch": 3.972508591065292, + "grad_norm": 0.1131262643697182, + "learning_rate": 9.712824218794457e-06, + "loss": 0.4733, + "step": 1445 + }, + { + "epoch": 3.975257731958763, + "grad_norm": 0.10279589610315747, + "learning_rate": 9.662614277344846e-06, + "loss": 0.479, + "step": 1446 + }, + { + "epoch": 3.9780068728522338, + "grad_norm": 0.11106926427912463, + "learning_rate": 9.612516616668612e-06, + "loss": 0.4822, + "step": 1447 + }, + { + "epoch": 3.9807560137457045, + "grad_norm": 0.10564345176951702, + "learning_rate": 9.562531422180674e-06, + "loss": 0.4745, + "step": 1448 + }, + { + "epoch": 3.9835051546391753, + "grad_norm": 0.10829788647274623, + "learning_rate": 9.512658878879737e-06, + "loss": 0.4846, + "step": 1449 + }, + { + "epoch": 3.986254295532646, + "grad_norm": 0.11018495327771764, + "learning_rate": 9.462899171347538e-06, + "loss": 0.4802, + "step": 1450 + }, + { + "epoch": 3.989003436426117, + "grad_norm": 0.12608551718647237, + "learning_rate": 9.413252483748216e-06, + "loss": 0.4766, + "step": 1451 + }, + { + "epoch": 3.9917525773195877, + "grad_norm": 0.11337813442101644, + "learning_rate": 9.363718999827612e-06, + "loss": 0.479, + "step": 1452 + }, + { + "epoch": 3.9945017182130584, + "grad_norm": 0.12372671414303754, + "learning_rate": 9.314298902912583e-06, + "loss": 0.479, + "step": 1453 + }, + { + "epoch": 3.997250859106529, + "grad_norm": 0.11011001077642653, + "learning_rate": 9.26499237591037e-06, + "loss": 0.4695, + "step": 1454 + }, + { + "epoch": 4.0, + "grad_norm": 0.14840181176358444, + "learning_rate": 9.215799601307828e-06, + "loss": 0.6096, + "step": 1455 + }, + { + "epoch": 4.002749140893471, + "grad_norm": 0.19421057731547003, + "learning_rate": 9.16672076117084e-06, + "loss": 0.4433, + "step": 1456 + }, + { + "epoch": 4.005498281786942, + "grad_norm": 0.15015912458886393, + "learning_rate": 9.117756037143622e-06, + "loss": 0.4425, + "step": 1457 + }, + { + "epoch": 4.008247422680412, + "grad_norm": 0.1383285525044493, + "learning_rate": 9.068905610448011e-06, + "loss": 0.4562, + "step": 1458 + }, + { + "epoch": 4.010996563573883, + "grad_norm": 0.15430721574490833, + "learning_rate": 9.02016966188283e-06, + "loss": 0.4484, + "step": 1459 + }, + { + "epoch": 4.013745704467354, + "grad_norm": 0.14905855161830828, + "learning_rate": 8.971548371823205e-06, + "loss": 0.446, + "step": 1460 + }, + { + "epoch": 4.016494845360825, + "grad_norm": 0.15470678416362293, + "learning_rate": 8.923041920219927e-06, + "loss": 0.4412, + "step": 1461 + }, + { + "epoch": 4.0192439862542955, + "grad_norm": 0.14564788987737176, + "learning_rate": 8.87465048659875e-06, + "loss": 0.4393, + "step": 1462 + }, + { + "epoch": 4.021993127147766, + "grad_norm": 0.13342656861373697, + "learning_rate": 8.826374250059704e-06, + "loss": 0.4498, + "step": 1463 + }, + { + "epoch": 4.024742268041237, + "grad_norm": 0.1302414374832906, + "learning_rate": 8.778213389276523e-06, + "loss": 0.4383, + "step": 1464 + }, + { + "epoch": 4.027491408934708, + "grad_norm": 0.14007280746183526, + "learning_rate": 8.730168082495885e-06, + "loss": 0.4474, + "step": 1465 + }, + { + "epoch": 4.030240549828179, + "grad_norm": 0.13171480431395038, + "learning_rate": 8.682238507536823e-06, + "loss": 0.4388, + "step": 1466 + }, + { + "epoch": 4.032989690721649, + "grad_norm": 0.12287866481381172, + "learning_rate": 8.634424841790014e-06, + "loss": 0.4384, + "step": 1467 + }, + { + "epoch": 4.03573883161512, + "grad_norm": 0.12293190753633977, + "learning_rate": 8.586727262217156e-06, + "loss": 0.4444, + "step": 1468 + }, + { + "epoch": 4.038487972508591, + "grad_norm": 0.12156568952497382, + "learning_rate": 8.539145945350324e-06, + "loss": 0.4393, + "step": 1469 + }, + { + "epoch": 4.041237113402062, + "grad_norm": 0.13520454469985085, + "learning_rate": 8.491681067291279e-06, + "loss": 0.4438, + "step": 1470 + }, + { + "epoch": 4.0439862542955325, + "grad_norm": 0.12818857279590956, + "learning_rate": 8.444332803710806e-06, + "loss": 0.4443, + "step": 1471 + }, + { + "epoch": 4.046735395189003, + "grad_norm": 0.11732641042539134, + "learning_rate": 8.397101329848146e-06, + "loss": 0.4443, + "step": 1472 + }, + { + "epoch": 4.049484536082474, + "grad_norm": 0.12736631288786282, + "learning_rate": 8.349986820510257e-06, + "loss": 0.4427, + "step": 1473 + }, + { + "epoch": 4.052233676975945, + "grad_norm": 0.1304818919917951, + "learning_rate": 8.302989450071219e-06, + "loss": 0.4478, + "step": 1474 + }, + { + "epoch": 4.054982817869416, + "grad_norm": 0.13641319860640636, + "learning_rate": 8.256109392471549e-06, + "loss": 0.4595, + "step": 1475 + }, + { + "epoch": 4.057731958762886, + "grad_norm": 0.11825978977034925, + "learning_rate": 8.20934682121763e-06, + "loss": 0.4461, + "step": 1476 + }, + { + "epoch": 4.060481099656357, + "grad_norm": 0.13030109067712098, + "learning_rate": 8.162701909380967e-06, + "loss": 0.4473, + "step": 1477 + }, + { + "epoch": 4.063230240549828, + "grad_norm": 0.12977389788164925, + "learning_rate": 8.116174829597625e-06, + "loss": 0.4506, + "step": 1478 + }, + { + "epoch": 4.065979381443299, + "grad_norm": 0.12148360180077224, + "learning_rate": 8.069765754067562e-06, + "loss": 0.4454, + "step": 1479 + }, + { + "epoch": 4.0687285223367695, + "grad_norm": 0.13076375704534893, + "learning_rate": 8.023474854553996e-06, + "loss": 0.4459, + "step": 1480 + }, + { + "epoch": 4.07147766323024, + "grad_norm": 0.13403343398060516, + "learning_rate": 7.977302302382757e-06, + "loss": 0.4392, + "step": 1481 + }, + { + "epoch": 4.074226804123711, + "grad_norm": 0.11166777845488875, + "learning_rate": 7.93124826844167e-06, + "loss": 0.4411, + "step": 1482 + }, + { + "epoch": 4.076975945017182, + "grad_norm": 0.13721899419089614, + "learning_rate": 7.885312923179912e-06, + "loss": 0.444, + "step": 1483 + }, + { + "epoch": 4.079725085910653, + "grad_norm": 0.13263433098857536, + "learning_rate": 7.839496436607406e-06, + "loss": 0.4459, + "step": 1484 + }, + { + "epoch": 4.082474226804123, + "grad_norm": 0.11590942502418243, + "learning_rate": 7.793798978294144e-06, + "loss": 0.4453, + "step": 1485 + }, + { + "epoch": 4.085223367697594, + "grad_norm": 0.13293144505263688, + "learning_rate": 7.748220717369594e-06, + "loss": 0.4413, + "step": 1486 + }, + { + "epoch": 4.087972508591065, + "grad_norm": 0.13483699801148805, + "learning_rate": 7.702761822522072e-06, + "loss": 0.4366, + "step": 1487 + }, + { + "epoch": 4.090721649484536, + "grad_norm": 0.11307879927149218, + "learning_rate": 7.6574224619981e-06, + "loss": 0.4352, + "step": 1488 + }, + { + "epoch": 4.0934707903780065, + "grad_norm": 0.11327258461306543, + "learning_rate": 7.612202803601812e-06, + "loss": 0.4513, + "step": 1489 + }, + { + "epoch": 4.096219931271477, + "grad_norm": 0.11998728903473971, + "learning_rate": 7.56710301469429e-06, + "loss": 0.4462, + "step": 1490 + }, + { + "epoch": 4.098969072164948, + "grad_norm": 0.11148518956657245, + "learning_rate": 7.522123262193011e-06, + "loss": 0.4414, + "step": 1491 + }, + { + "epoch": 4.101718213058419, + "grad_norm": 0.10980782127466164, + "learning_rate": 7.477263712571154e-06, + "loss": 0.4505, + "step": 1492 + }, + { + "epoch": 4.10446735395189, + "grad_norm": 0.13007209291051053, + "learning_rate": 7.432524531857037e-06, + "loss": 0.4404, + "step": 1493 + }, + { + "epoch": 4.10721649484536, + "grad_norm": 0.10633901487852752, + "learning_rate": 7.38790588563346e-06, + "loss": 0.4432, + "step": 1494 + }, + { + "epoch": 4.109965635738831, + "grad_norm": 0.10337495591575992, + "learning_rate": 7.343407939037166e-06, + "loss": 0.4442, + "step": 1495 + }, + { + "epoch": 4.112714776632302, + "grad_norm": 0.11699522345473407, + "learning_rate": 7.29903085675812e-06, + "loss": 0.4404, + "step": 1496 + }, + { + "epoch": 4.115463917525773, + "grad_norm": 0.1071569023006855, + "learning_rate": 7.254774803038986e-06, + "loss": 0.4448, + "step": 1497 + }, + { + "epoch": 4.118213058419244, + "grad_norm": 0.10590236218912877, + "learning_rate": 7.210639941674515e-06, + "loss": 0.4429, + "step": 1498 + }, + { + "epoch": 4.120962199312714, + "grad_norm": 0.10210210146806667, + "learning_rate": 7.166626436010893e-06, + "loss": 0.4415, + "step": 1499 + }, + { + "epoch": 4.123711340206185, + "grad_norm": 0.11125955970384037, + "learning_rate": 7.122734448945157e-06, + "loss": 0.4526, + "step": 1500 + }, + { + "epoch": 4.126460481099656, + "grad_norm": 0.12013542728499116, + "learning_rate": 7.078964142924607e-06, + "loss": 0.4419, + "step": 1501 + }, + { + "epoch": 4.129209621993128, + "grad_norm": 0.10484380190061886, + "learning_rate": 7.035315679946176e-06, + "loss": 0.4489, + "step": 1502 + }, + { + "epoch": 4.131958762886598, + "grad_norm": 0.09567503999966823, + "learning_rate": 6.991789221555887e-06, + "loss": 0.4372, + "step": 1503 + }, + { + "epoch": 4.134707903780069, + "grad_norm": 0.10540468880005585, + "learning_rate": 6.94838492884816e-06, + "loss": 0.4464, + "step": 1504 + }, + { + "epoch": 4.13745704467354, + "grad_norm": 0.11225166249471208, + "learning_rate": 6.905102962465302e-06, + "loss": 0.4439, + "step": 1505 + }, + { + "epoch": 4.140206185567011, + "grad_norm": 0.10759681388030853, + "learning_rate": 6.861943482596896e-06, + "loss": 0.4436, + "step": 1506 + }, + { + "epoch": 4.1429553264604815, + "grad_norm": 0.10130359403512029, + "learning_rate": 6.818906648979169e-06, + "loss": 0.4388, + "step": 1507 + }, + { + "epoch": 4.145704467353952, + "grad_norm": 0.1121174956770293, + "learning_rate": 6.775992620894434e-06, + "loss": 0.4507, + "step": 1508 + }, + { + "epoch": 4.148453608247423, + "grad_norm": 0.1014146829605852, + "learning_rate": 6.733201557170481e-06, + "loss": 0.4432, + "step": 1509 + }, + { + "epoch": 4.151202749140894, + "grad_norm": 0.09837158537375995, + "learning_rate": 6.690533616180031e-06, + "loss": 0.4416, + "step": 1510 + }, + { + "epoch": 4.153951890034365, + "grad_norm": 0.10857273055540849, + "learning_rate": 6.647988955840099e-06, + "loss": 0.4283, + "step": 1511 + }, + { + "epoch": 4.156701030927835, + "grad_norm": 0.11827148688247635, + "learning_rate": 6.6055677336114024e-06, + "loss": 0.4435, + "step": 1512 + }, + { + "epoch": 4.159450171821306, + "grad_norm": 0.10661184402237826, + "learning_rate": 6.563270106497866e-06, + "loss": 0.4493, + "step": 1513 + }, + { + "epoch": 4.162199312714777, + "grad_norm": 0.10027884158801532, + "learning_rate": 6.521096231045927e-06, + "loss": 0.4424, + "step": 1514 + }, + { + "epoch": 4.164948453608248, + "grad_norm": 0.10677714613702075, + "learning_rate": 6.479046263344044e-06, + "loss": 0.4407, + "step": 1515 + }, + { + "epoch": 4.1676975945017185, + "grad_norm": 0.10120025403165747, + "learning_rate": 6.43712035902206e-06, + "loss": 0.4482, + "step": 1516 + }, + { + "epoch": 4.170446735395189, + "grad_norm": 0.09576847158285014, + "learning_rate": 6.395318673250655e-06, + "loss": 0.4455, + "step": 1517 + }, + { + "epoch": 4.17319587628866, + "grad_norm": 0.10456502041919041, + "learning_rate": 6.353641360740796e-06, + "loss": 0.4442, + "step": 1518 + }, + { + "epoch": 4.175945017182131, + "grad_norm": 0.10975215070737157, + "learning_rate": 6.312088575743112e-06, + "loss": 0.4421, + "step": 1519 + }, + { + "epoch": 4.178694158075602, + "grad_norm": 0.09670376981272215, + "learning_rate": 6.270660472047318e-06, + "loss": 0.4357, + "step": 1520 + }, + { + "epoch": 4.181443298969072, + "grad_norm": 0.0989487056180283, + "learning_rate": 6.229357202981736e-06, + "loss": 0.4446, + "step": 1521 + }, + { + "epoch": 4.184192439862543, + "grad_norm": 0.11575820428367671, + "learning_rate": 6.188178921412622e-06, + "loss": 0.4376, + "step": 1522 + }, + { + "epoch": 4.186941580756014, + "grad_norm": 0.09506731515085455, + "learning_rate": 6.147125779743666e-06, + "loss": 0.4399, + "step": 1523 + }, + { + "epoch": 4.189690721649485, + "grad_norm": 0.10319725228868595, + "learning_rate": 6.106197929915385e-06, + "loss": 0.4494, + "step": 1524 + }, + { + "epoch": 4.1924398625429555, + "grad_norm": 0.10930094347383024, + "learning_rate": 6.065395523404616e-06, + "loss": 0.4416, + "step": 1525 + }, + { + "epoch": 4.195189003436426, + "grad_norm": 0.10333832350838476, + "learning_rate": 6.0247187112238936e-06, + "loss": 0.4419, + "step": 1526 + }, + { + "epoch": 4.197938144329897, + "grad_norm": 0.09863833187141094, + "learning_rate": 5.984167643920926e-06, + "loss": 0.45, + "step": 1527 + }, + { + "epoch": 4.200687285223368, + "grad_norm": 0.10338899938814769, + "learning_rate": 5.943742471578029e-06, + "loss": 0.4432, + "step": 1528 + }, + { + "epoch": 4.203436426116839, + "grad_norm": 0.10080408781457796, + "learning_rate": 5.903443343811583e-06, + "loss": 0.4498, + "step": 1529 + }, + { + "epoch": 4.206185567010309, + "grad_norm": 0.0995238925477759, + "learning_rate": 5.863270409771451e-06, + "loss": 0.4441, + "step": 1530 + }, + { + "epoch": 4.20893470790378, + "grad_norm": 0.09871347524798951, + "learning_rate": 5.823223818140458e-06, + "loss": 0.4465, + "step": 1531 + }, + { + "epoch": 4.211683848797251, + "grad_norm": 0.09741808192782717, + "learning_rate": 5.78330371713383e-06, + "loss": 0.4521, + "step": 1532 + }, + { + "epoch": 4.214432989690722, + "grad_norm": 0.10108809287384478, + "learning_rate": 5.7435102544986325e-06, + "loss": 0.4476, + "step": 1533 + }, + { + "epoch": 4.217182130584193, + "grad_norm": 0.09586236341789252, + "learning_rate": 5.70384357751323e-06, + "loss": 0.4417, + "step": 1534 + }, + { + "epoch": 4.219931271477663, + "grad_norm": 0.0982846297983466, + "learning_rate": 5.664303832986764e-06, + "loss": 0.4496, + "step": 1535 + }, + { + "epoch": 4.222680412371134, + "grad_norm": 0.09404435584314635, + "learning_rate": 5.624891167258569e-06, + "loss": 0.4415, + "step": 1536 + }, + { + "epoch": 4.225429553264605, + "grad_norm": 0.09792884175585087, + "learning_rate": 5.585605726197663e-06, + "loss": 0.4364, + "step": 1537 + }, + { + "epoch": 4.228178694158076, + "grad_norm": 0.0966404647926693, + "learning_rate": 5.54644765520219e-06, + "loss": 0.4464, + "step": 1538 + }, + { + "epoch": 4.2309278350515465, + "grad_norm": 0.09803372963199505, + "learning_rate": 5.507417099198886e-06, + "loss": 0.443, + "step": 1539 + }, + { + "epoch": 4.233676975945017, + "grad_norm": 0.10220414170891072, + "learning_rate": 5.468514202642574e-06, + "loss": 0.4416, + "step": 1540 + }, + { + "epoch": 4.236426116838488, + "grad_norm": 0.09940583554566296, + "learning_rate": 5.429739109515564e-06, + "loss": 0.4483, + "step": 1541 + }, + { + "epoch": 4.239175257731959, + "grad_norm": 0.09639208839569098, + "learning_rate": 5.3910919633271755e-06, + "loss": 0.442, + "step": 1542 + }, + { + "epoch": 4.24192439862543, + "grad_norm": 0.1117369320333374, + "learning_rate": 5.352572907113178e-06, + "loss": 0.4461, + "step": 1543 + }, + { + "epoch": 4.2446735395189, + "grad_norm": 0.09611791874864287, + "learning_rate": 5.314182083435282e-06, + "loss": 0.4346, + "step": 1544 + }, + { + "epoch": 4.247422680412371, + "grad_norm": 0.10794580998258632, + "learning_rate": 5.2759196343805885e-06, + "loss": 0.4415, + "step": 1545 + }, + { + "epoch": 4.250171821305842, + "grad_norm": 0.0952864724999734, + "learning_rate": 5.237785701561078e-06, + "loss": 0.4443, + "step": 1546 + }, + { + "epoch": 4.252920962199313, + "grad_norm": 0.10156924345931388, + "learning_rate": 5.1997804261131015e-06, + "loss": 0.449, + "step": 1547 + }, + { + "epoch": 4.2556701030927835, + "grad_norm": 0.09725478784660825, + "learning_rate": 5.161903948696813e-06, + "loss": 0.439, + "step": 1548 + }, + { + "epoch": 4.258419243986254, + "grad_norm": 0.10252085604277364, + "learning_rate": 5.124156409495693e-06, + "loss": 0.4385, + "step": 1549 + }, + { + "epoch": 4.261168384879725, + "grad_norm": 0.10131443009760514, + "learning_rate": 5.086537948216008e-06, + "loss": 0.4407, + "step": 1550 + }, + { + "epoch": 4.263917525773196, + "grad_norm": 0.09875784055342578, + "learning_rate": 5.049048704086295e-06, + "loss": 0.4469, + "step": 1551 + }, + { + "epoch": 4.266666666666667, + "grad_norm": 0.10157130046545569, + "learning_rate": 5.011688815856856e-06, + "loss": 0.4502, + "step": 1552 + }, + { + "epoch": 4.269415807560137, + "grad_norm": 0.10941012381717943, + "learning_rate": 4.974458421799231e-06, + "loss": 0.4411, + "step": 1553 + }, + { + "epoch": 4.272164948453608, + "grad_norm": 0.11064889415177875, + "learning_rate": 4.937357659705688e-06, + "loss": 0.4419, + "step": 1554 + }, + { + "epoch": 4.274914089347079, + "grad_norm": 0.10002641057536234, + "learning_rate": 4.90038666688875e-06, + "loss": 0.4558, + "step": 1555 + }, + { + "epoch": 4.27766323024055, + "grad_norm": 0.10016807122767976, + "learning_rate": 4.863545580180615e-06, + "loss": 0.4435, + "step": 1556 + }, + { + "epoch": 4.2804123711340205, + "grad_norm": 0.10851136506221658, + "learning_rate": 4.8268345359327075e-06, + "loss": 0.4398, + "step": 1557 + }, + { + "epoch": 4.283161512027491, + "grad_norm": 0.09488321817571037, + "learning_rate": 4.7902536700151504e-06, + "loss": 0.4429, + "step": 1558 + }, + { + "epoch": 4.285910652920962, + "grad_norm": 0.09962394867509372, + "learning_rate": 4.75380311781628e-06, + "loss": 0.4435, + "step": 1559 + }, + { + "epoch": 4.288659793814433, + "grad_norm": 0.10525730743423926, + "learning_rate": 4.717483014242134e-06, + "loss": 0.4488, + "step": 1560 + }, + { + "epoch": 4.291408934707904, + "grad_norm": 0.09444016052953329, + "learning_rate": 4.681293493715906e-06, + "loss": 0.4383, + "step": 1561 + }, + { + "epoch": 4.294158075601374, + "grad_norm": 0.0935555125757194, + "learning_rate": 4.645234690177556e-06, + "loss": 0.441, + "step": 1562 + }, + { + "epoch": 4.296907216494845, + "grad_norm": 0.09875362647621048, + "learning_rate": 4.6093067370832145e-06, + "loss": 0.4467, + "step": 1563 + }, + { + "epoch": 4.299656357388316, + "grad_norm": 0.10106687338028049, + "learning_rate": 4.573509767404733e-06, + "loss": 0.4398, + "step": 1564 + }, + { + "epoch": 4.302405498281787, + "grad_norm": 0.09430960821120914, + "learning_rate": 4.537843913629178e-06, + "loss": 0.4498, + "step": 1565 + }, + { + "epoch": 4.3051546391752575, + "grad_norm": 0.09341919086027475, + "learning_rate": 4.502309307758368e-06, + "loss": 0.4427, + "step": 1566 + }, + { + "epoch": 4.307903780068728, + "grad_norm": 0.09579646490730007, + "learning_rate": 4.46690608130834e-06, + "loss": 0.4504, + "step": 1567 + }, + { + "epoch": 4.310652920962199, + "grad_norm": 0.10140215061324537, + "learning_rate": 4.431634365308904e-06, + "loss": 0.4445, + "step": 1568 + }, + { + "epoch": 4.31340206185567, + "grad_norm": 0.09687704144452364, + "learning_rate": 4.39649429030311e-06, + "loss": 0.4478, + "step": 1569 + }, + { + "epoch": 4.316151202749141, + "grad_norm": 0.09438778342768989, + "learning_rate": 4.3614859863468425e-06, + "loss": 0.4386, + "step": 1570 + }, + { + "epoch": 4.318900343642611, + "grad_norm": 0.09156131739398227, + "learning_rate": 4.326609583008261e-06, + "loss": 0.4338, + "step": 1571 + }, + { + "epoch": 4.321649484536082, + "grad_norm": 0.09078660520618338, + "learning_rate": 4.2918652093673606e-06, + "loss": 0.442, + "step": 1572 + }, + { + "epoch": 4.324398625429553, + "grad_norm": 0.10260056112083875, + "learning_rate": 4.257252994015466e-06, + "loss": 0.4441, + "step": 1573 + }, + { + "epoch": 4.327147766323024, + "grad_norm": 0.10384708279251556, + "learning_rate": 4.222773065054817e-06, + "loss": 0.4466, + "step": 1574 + }, + { + "epoch": 4.329896907216495, + "grad_norm": 0.09444726168415894, + "learning_rate": 4.188425550098019e-06, + "loss": 0.444, + "step": 1575 + }, + { + "epoch": 4.332646048109965, + "grad_norm": 0.1139460077036934, + "learning_rate": 4.15421057626761e-06, + "loss": 0.4446, + "step": 1576 + }, + { + "epoch": 4.335395189003436, + "grad_norm": 0.09822740636361162, + "learning_rate": 4.120128270195585e-06, + "loss": 0.4424, + "step": 1577 + }, + { + "epoch": 4.338144329896907, + "grad_norm": 0.09913730861031846, + "learning_rate": 4.086178758022929e-06, + "loss": 0.4477, + "step": 1578 + }, + { + "epoch": 4.340893470790378, + "grad_norm": 0.09892692023825955, + "learning_rate": 4.052362165399148e-06, + "loss": 0.4462, + "step": 1579 + }, + { + "epoch": 4.3436426116838485, + "grad_norm": 0.09559985988178789, + "learning_rate": 4.018678617481797e-06, + "loss": 0.4518, + "step": 1580 + }, + { + "epoch": 4.346391752577319, + "grad_norm": 0.09645896728376689, + "learning_rate": 3.9851282389360336e-06, + "loss": 0.4411, + "step": 1581 + }, + { + "epoch": 4.34914089347079, + "grad_norm": 0.09621093774200687, + "learning_rate": 3.951711153934143e-06, + "loss": 0.4455, + "step": 1582 + }, + { + "epoch": 4.351890034364261, + "grad_norm": 0.09860601420290797, + "learning_rate": 3.918427486155079e-06, + "loss": 0.4452, + "step": 1583 + }, + { + "epoch": 4.354639175257732, + "grad_norm": 0.09460049776906895, + "learning_rate": 3.885277358784003e-06, + "loss": 0.4476, + "step": 1584 + }, + { + "epoch": 4.357388316151202, + "grad_norm": 0.0970914145620547, + "learning_rate": 3.85226089451185e-06, + "loss": 0.4486, + "step": 1585 + }, + { + "epoch": 4.360137457044673, + "grad_norm": 0.10970056178061441, + "learning_rate": 3.819378215534842e-06, + "loss": 0.4404, + "step": 1586 + }, + { + "epoch": 4.362886597938144, + "grad_norm": 0.09671374353988552, + "learning_rate": 3.7866294435540574e-06, + "loss": 0.4534, + "step": 1587 + }, + { + "epoch": 4.365635738831615, + "grad_norm": 0.09276644771447388, + "learning_rate": 3.7540146997749793e-06, + "loss": 0.4489, + "step": 1588 + }, + { + "epoch": 4.368384879725086, + "grad_norm": 0.09400993795205304, + "learning_rate": 3.72153410490705e-06, + "loss": 0.4467, + "step": 1589 + }, + { + "epoch": 4.371134020618557, + "grad_norm": 0.10312570233843393, + "learning_rate": 3.689187779163206e-06, + "loss": 0.4491, + "step": 1590 + }, + { + "epoch": 4.373883161512028, + "grad_norm": 0.10124141768595926, + "learning_rate": 3.6569758422594446e-06, + "loss": 0.4528, + "step": 1591 + }, + { + "epoch": 4.376632302405499, + "grad_norm": 0.0947775184692594, + "learning_rate": 3.6248984134143794e-06, + "loss": 0.4493, + "step": 1592 + }, + { + "epoch": 4.3793814432989695, + "grad_norm": 0.0949068368311883, + "learning_rate": 3.5929556113488117e-06, + "loss": 0.4502, + "step": 1593 + }, + { + "epoch": 4.38213058419244, + "grad_norm": 0.09225617183677981, + "learning_rate": 3.5611475542852714e-06, + "loss": 0.4378, + "step": 1594 + }, + { + "epoch": 4.384879725085911, + "grad_norm": 0.09712650449251893, + "learning_rate": 3.5294743599475754e-06, + "loss": 0.446, + "step": 1595 + }, + { + "epoch": 4.387628865979382, + "grad_norm": 0.10510272619911021, + "learning_rate": 3.4979361455604386e-06, + "loss": 0.4438, + "step": 1596 + }, + { + "epoch": 4.390378006872853, + "grad_norm": 0.10034299553088664, + "learning_rate": 3.4665330278489708e-06, + "loss": 0.4487, + "step": 1597 + }, + { + "epoch": 4.393127147766323, + "grad_norm": 0.09360138161898929, + "learning_rate": 3.4352651230383026e-06, + "loss": 0.4409, + "step": 1598 + }, + { + "epoch": 4.395876288659794, + "grad_norm": 0.09887187647716278, + "learning_rate": 3.4041325468531094e-06, + "loss": 0.4409, + "step": 1599 + }, + { + "epoch": 4.398625429553265, + "grad_norm": 0.10473071373698492, + "learning_rate": 3.3731354145172435e-06, + "loss": 0.4516, + "step": 1600 + }, + { + "epoch": 4.401374570446736, + "grad_norm": 0.1007375745415408, + "learning_rate": 3.342273840753221e-06, + "loss": 0.4383, + "step": 1601 + }, + { + "epoch": 4.4041237113402065, + "grad_norm": 0.09735792228082213, + "learning_rate": 3.311547939781887e-06, + "loss": 0.4409, + "step": 1602 + }, + { + "epoch": 4.406872852233677, + "grad_norm": 0.09459778267684918, + "learning_rate": 3.280957825321922e-06, + "loss": 0.4442, + "step": 1603 + }, + { + "epoch": 4.409621993127148, + "grad_norm": 0.09281792628016262, + "learning_rate": 3.250503610589482e-06, + "loss": 0.4491, + "step": 1604 + }, + { + "epoch": 4.412371134020619, + "grad_norm": 0.1007767987975053, + "learning_rate": 3.2201854082977292e-06, + "loss": 0.4517, + "step": 1605 + }, + { + "epoch": 4.41512027491409, + "grad_norm": 0.09789821294262532, + "learning_rate": 3.190003330656435e-06, + "loss": 0.4451, + "step": 1606 + }, + { + "epoch": 4.41786941580756, + "grad_norm": 0.10028034923814605, + "learning_rate": 3.159957489371559e-06, + "loss": 0.4367, + "step": 1607 + }, + { + "epoch": 4.420618556701031, + "grad_norm": 0.09081132192790976, + "learning_rate": 3.1300479956448693e-06, + "loss": 0.4387, + "step": 1608 + }, + { + "epoch": 4.423367697594502, + "grad_norm": 0.09711016308139389, + "learning_rate": 3.1002749601734618e-06, + "loss": 0.4428, + "step": 1609 + }, + { + "epoch": 4.426116838487973, + "grad_norm": 0.09435493012986142, + "learning_rate": 3.0706384931494137e-06, + "loss": 0.4328, + "step": 1610 + }, + { + "epoch": 4.4288659793814436, + "grad_norm": 0.09951785033015637, + "learning_rate": 3.0411387042593545e-06, + "loss": 0.4512, + "step": 1611 + }, + { + "epoch": 4.431615120274914, + "grad_norm": 0.0960124562000595, + "learning_rate": 3.0117757026840543e-06, + "loss": 0.4485, + "step": 1612 + }, + { + "epoch": 4.434364261168385, + "grad_norm": 0.09374569219824876, + "learning_rate": 2.9825495970980234e-06, + "loss": 0.4448, + "step": 1613 + }, + { + "epoch": 4.437113402061856, + "grad_norm": 0.09287833110723459, + "learning_rate": 2.953460495669096e-06, + "loss": 0.4473, + "step": 1614 + }, + { + "epoch": 4.439862542955327, + "grad_norm": 0.09330518761309428, + "learning_rate": 2.924508506058077e-06, + "loss": 0.4368, + "step": 1615 + }, + { + "epoch": 4.4426116838487975, + "grad_norm": 0.09851414675120082, + "learning_rate": 2.8956937354182923e-06, + "loss": 0.4438, + "step": 1616 + }, + { + "epoch": 4.445360824742268, + "grad_norm": 0.09503089697922397, + "learning_rate": 2.867016290395199e-06, + "loss": 0.4557, + "step": 1617 + }, + { + "epoch": 4.448109965635739, + "grad_norm": 0.09322925477662775, + "learning_rate": 2.838476277126012e-06, + "loss": 0.4428, + "step": 1618 + }, + { + "epoch": 4.45085910652921, + "grad_norm": 0.09242195549495436, + "learning_rate": 2.810073801239317e-06, + "loss": 0.446, + "step": 1619 + }, + { + "epoch": 4.453608247422681, + "grad_norm": 0.08945993365343624, + "learning_rate": 2.781808967854649e-06, + "loss": 0.4412, + "step": 1620 + }, + { + "epoch": 4.456357388316151, + "grad_norm": 0.09428643807864204, + "learning_rate": 2.753681881582115e-06, + "loss": 0.4416, + "step": 1621 + }, + { + "epoch": 4.459106529209622, + "grad_norm": 0.0931736838114746, + "learning_rate": 2.7256926465220177e-06, + "loss": 0.4472, + "step": 1622 + }, + { + "epoch": 4.461855670103093, + "grad_norm": 0.09181260953087753, + "learning_rate": 2.697841366264471e-06, + "loss": 0.4481, + "step": 1623 + }, + { + "epoch": 4.464604810996564, + "grad_norm": 0.09223991013911755, + "learning_rate": 2.670128143888988e-06, + "loss": 0.4417, + "step": 1624 + }, + { + "epoch": 4.4673539518900345, + "grad_norm": 0.09242485950713865, + "learning_rate": 2.6425530819641364e-06, + "loss": 0.4501, + "step": 1625 + }, + { + "epoch": 4.470103092783505, + "grad_norm": 0.09113105402491876, + "learning_rate": 2.6151162825471364e-06, + "loss": 0.4442, + "step": 1626 + }, + { + "epoch": 4.472852233676976, + "grad_norm": 0.09212808247265639, + "learning_rate": 2.587817847183489e-06, + "loss": 0.4429, + "step": 1627 + }, + { + "epoch": 4.475601374570447, + "grad_norm": 0.09767261371119439, + "learning_rate": 2.5606578769066026e-06, + "loss": 0.4432, + "step": 1628 + }, + { + "epoch": 4.478350515463918, + "grad_norm": 0.09743572389157769, + "learning_rate": 2.5336364722374106e-06, + "loss": 0.4487, + "step": 1629 + }, + { + "epoch": 4.481099656357388, + "grad_norm": 0.09102922314529087, + "learning_rate": 2.506753733184013e-06, + "loss": 0.4503, + "step": 1630 + }, + { + "epoch": 4.483848797250859, + "grad_norm": 0.09266255759834628, + "learning_rate": 2.480009759241302e-06, + "loss": 0.4405, + "step": 1631 + }, + { + "epoch": 4.48659793814433, + "grad_norm": 0.0925039627946117, + "learning_rate": 2.453404649390576e-06, + "loss": 0.4398, + "step": 1632 + }, + { + "epoch": 4.489347079037801, + "grad_norm": 0.09521064172876387, + "learning_rate": 2.4269385020991987e-06, + "loss": 0.4486, + "step": 1633 + }, + { + "epoch": 4.4920962199312715, + "grad_norm": 0.09113818118580695, + "learning_rate": 2.40061141532022e-06, + "loss": 0.4431, + "step": 1634 + }, + { + "epoch": 4.494845360824742, + "grad_norm": 0.10000887227672747, + "learning_rate": 2.374423486492021e-06, + "loss": 0.4388, + "step": 1635 + }, + { + "epoch": 4.497594501718213, + "grad_norm": 0.09939158503866538, + "learning_rate": 2.3483748125379434e-06, + "loss": 0.4526, + "step": 1636 + }, + { + "epoch": 4.500343642611684, + "grad_norm": 0.08865852194600819, + "learning_rate": 2.3224654898659348e-06, + "loss": 0.4458, + "step": 1637 + }, + { + "epoch": 4.503092783505155, + "grad_norm": 0.09118943093226906, + "learning_rate": 2.2966956143682094e-06, + "loss": 0.444, + "step": 1638 + }, + { + "epoch": 4.505841924398625, + "grad_norm": 0.09085329322674768, + "learning_rate": 2.2710652814208656e-06, + "loss": 0.4516, + "step": 1639 + }, + { + "epoch": 4.508591065292096, + "grad_norm": 0.09474948162033754, + "learning_rate": 2.2455745858835474e-06, + "loss": 0.4439, + "step": 1640 + }, + { + "epoch": 4.511340206185567, + "grad_norm": 0.09100873289075298, + "learning_rate": 2.2202236220990913e-06, + "loss": 0.4441, + "step": 1641 + }, + { + "epoch": 4.514089347079038, + "grad_norm": 0.09556416343413056, + "learning_rate": 2.1950124838931685e-06, + "loss": 0.4426, + "step": 1642 + }, + { + "epoch": 4.5168384879725085, + "grad_norm": 0.09124176627878805, + "learning_rate": 2.169941264573967e-06, + "loss": 0.4515, + "step": 1643 + }, + { + "epoch": 4.519587628865979, + "grad_norm": 0.08967523905753207, + "learning_rate": 2.1450100569317954e-06, + "loss": 0.4427, + "step": 1644 + }, + { + "epoch": 4.52233676975945, + "grad_norm": 0.08908910909605862, + "learning_rate": 2.1202189532387994e-06, + "loss": 0.4421, + "step": 1645 + }, + { + "epoch": 4.525085910652921, + "grad_norm": 0.08930544495492708, + "learning_rate": 2.0955680452485795e-06, + "loss": 0.4386, + "step": 1646 + }, + { + "epoch": 4.527835051546392, + "grad_norm": 0.09705957099996088, + "learning_rate": 2.0710574241958527e-06, + "loss": 0.4499, + "step": 1647 + }, + { + "epoch": 4.530584192439862, + "grad_norm": 0.09629213253097813, + "learning_rate": 2.0466871807961253e-06, + "loss": 0.4483, + "step": 1648 + }, + { + "epoch": 4.533333333333333, + "grad_norm": 0.08724520563465142, + "learning_rate": 2.0224574052453817e-06, + "loss": 0.4449, + "step": 1649 + }, + { + "epoch": 4.536082474226804, + "grad_norm": 0.09135773175337733, + "learning_rate": 1.9983681872196836e-06, + "loss": 0.4399, + "step": 1650 + }, + { + "epoch": 4.538831615120275, + "grad_norm": 0.09009008354611814, + "learning_rate": 1.9744196158749183e-06, + "loss": 0.4418, + "step": 1651 + }, + { + "epoch": 4.541580756013746, + "grad_norm": 0.09186018004376852, + "learning_rate": 1.9506117798463944e-06, + "loss": 0.4478, + "step": 1652 + }, + { + "epoch": 4.544329896907216, + "grad_norm": 0.08971975311980089, + "learning_rate": 1.9269447672485864e-06, + "loss": 0.4414, + "step": 1653 + }, + { + "epoch": 4.547079037800687, + "grad_norm": 0.08979949876575676, + "learning_rate": 1.9034186656747478e-06, + "loss": 0.4452, + "step": 1654 + }, + { + "epoch": 4.549828178694158, + "grad_norm": 0.0936250091293909, + "learning_rate": 1.8800335621966148e-06, + "loss": 0.4482, + "step": 1655 + }, + { + "epoch": 4.552577319587629, + "grad_norm": 0.08981143391720423, + "learning_rate": 1.8567895433640705e-06, + "loss": 0.4448, + "step": 1656 + }, + { + "epoch": 4.5553264604810995, + "grad_norm": 0.08587438163737723, + "learning_rate": 1.8336866952048683e-06, + "loss": 0.435, + "step": 1657 + }, + { + "epoch": 4.55807560137457, + "grad_norm": 0.09326486613993211, + "learning_rate": 1.8107251032242335e-06, + "loss": 0.4508, + "step": 1658 + }, + { + "epoch": 4.560824742268041, + "grad_norm": 0.09331518445515743, + "learning_rate": 1.7879048524046182e-06, + "loss": 0.4499, + "step": 1659 + }, + { + "epoch": 4.563573883161512, + "grad_norm": 0.08976792064689015, + "learning_rate": 1.765226027205369e-06, + "loss": 0.4466, + "step": 1660 + }, + { + "epoch": 4.566323024054983, + "grad_norm": 0.0875193893760576, + "learning_rate": 1.7426887115623791e-06, + "loss": 0.4452, + "step": 1661 + }, + { + "epoch": 4.569072164948453, + "grad_norm": 0.08920117183529058, + "learning_rate": 1.7202929888878329e-06, + "loss": 0.4401, + "step": 1662 + }, + { + "epoch": 4.571821305841924, + "grad_norm": 0.08832595498688149, + "learning_rate": 1.6980389420698395e-06, + "loss": 0.4451, + "step": 1663 + }, + { + "epoch": 4.574570446735395, + "grad_norm": 0.08859855283203406, + "learning_rate": 1.6759266534721952e-06, + "loss": 0.4431, + "step": 1664 + }, + { + "epoch": 4.577319587628866, + "grad_norm": 0.09152185332529074, + "learning_rate": 1.6539562049340085e-06, + "loss": 0.4402, + "step": 1665 + }, + { + "epoch": 4.5800687285223365, + "grad_norm": 0.08871925909133102, + "learning_rate": 1.6321276777694307e-06, + "loss": 0.4487, + "step": 1666 + }, + { + "epoch": 4.582817869415807, + "grad_norm": 0.09057802642441674, + "learning_rate": 1.6104411527673613e-06, + "loss": 0.4435, + "step": 1667 + }, + { + "epoch": 4.585567010309278, + "grad_norm": 0.09402629602885212, + "learning_rate": 1.588896710191139e-06, + "loss": 0.4489, + "step": 1668 + }, + { + "epoch": 4.588316151202749, + "grad_norm": 0.08990522146704653, + "learning_rate": 1.567494429778238e-06, + "loss": 0.4427, + "step": 1669 + }, + { + "epoch": 4.59106529209622, + "grad_norm": 0.09102458376196323, + "learning_rate": 1.546234390739998e-06, + "loss": 0.4498, + "step": 1670 + }, + { + "epoch": 4.59381443298969, + "grad_norm": 0.09383015002767303, + "learning_rate": 1.525116671761282e-06, + "loss": 0.4456, + "step": 1671 + }, + { + "epoch": 4.596563573883161, + "grad_norm": 0.087955107242829, + "learning_rate": 1.5041413510002544e-06, + "loss": 0.4495, + "step": 1672 + }, + { + "epoch": 4.599312714776632, + "grad_norm": 0.08874300385624094, + "learning_rate": 1.4833085060880349e-06, + "loss": 0.4412, + "step": 1673 + }, + { + "epoch": 4.602061855670103, + "grad_norm": 0.08466557475239722, + "learning_rate": 1.4626182141284085e-06, + "loss": 0.4455, + "step": 1674 + }, + { + "epoch": 4.6048109965635735, + "grad_norm": 0.08895580241312337, + "learning_rate": 1.4420705516976097e-06, + "loss": 0.4406, + "step": 1675 + }, + { + "epoch": 4.607560137457044, + "grad_norm": 0.08951149546757349, + "learning_rate": 1.421665594843953e-06, + "loss": 0.4485, + "step": 1676 + }, + { + "epoch": 4.610309278350515, + "grad_norm": 0.09239895115644584, + "learning_rate": 1.4014034190876057e-06, + "loss": 0.448, + "step": 1677 + }, + { + "epoch": 4.613058419243986, + "grad_norm": 0.08778403074207193, + "learning_rate": 1.3812840994202792e-06, + "loss": 0.44, + "step": 1678 + }, + { + "epoch": 4.615807560137457, + "grad_norm": 0.08817137101596591, + "learning_rate": 1.361307710304991e-06, + "loss": 0.4394, + "step": 1679 + }, + { + "epoch": 4.618556701030927, + "grad_norm": 0.09227710988150696, + "learning_rate": 1.3414743256757334e-06, + "loss": 0.4501, + "step": 1680 + }, + { + "epoch": 4.621305841924398, + "grad_norm": 0.08760382434561047, + "learning_rate": 1.3217840189372555e-06, + "loss": 0.4461, + "step": 1681 + }, + { + "epoch": 4.624054982817869, + "grad_norm": 0.09031853519071162, + "learning_rate": 1.3022368629647253e-06, + "loss": 0.4496, + "step": 1682 + }, + { + "epoch": 4.62680412371134, + "grad_norm": 0.09206239051131887, + "learning_rate": 1.2828329301035481e-06, + "loss": 0.4432, + "step": 1683 + }, + { + "epoch": 4.6295532646048105, + "grad_norm": 0.08843350392995919, + "learning_rate": 1.263572292169024e-06, + "loss": 0.4501, + "step": 1684 + }, + { + "epoch": 4.632302405498281, + "grad_norm": 0.09481473881768035, + "learning_rate": 1.2444550204461092e-06, + "loss": 0.4403, + "step": 1685 + }, + { + "epoch": 4.635051546391752, + "grad_norm": 0.08689601784166971, + "learning_rate": 1.2254811856891524e-06, + "loss": 0.4435, + "step": 1686 + }, + { + "epoch": 4.637800687285223, + "grad_norm": 0.08893870660419893, + "learning_rate": 1.2066508581216429e-06, + "loss": 0.4503, + "step": 1687 + }, + { + "epoch": 4.640549828178694, + "grad_norm": 0.08937431163006333, + "learning_rate": 1.187964107435926e-06, + "loss": 0.4456, + "step": 1688 + }, + { + "epoch": 4.643298969072165, + "grad_norm": 0.08865313646975571, + "learning_rate": 1.169421002792972e-06, + "loss": 0.4433, + "step": 1689 + }, + { + "epoch": 4.646048109965636, + "grad_norm": 0.08865228962259511, + "learning_rate": 1.151021612822092e-06, + "loss": 0.4491, + "step": 1690 + }, + { + "epoch": 4.648797250859107, + "grad_norm": 0.08885659723955662, + "learning_rate": 1.1327660056207113e-06, + "loss": 0.4368, + "step": 1691 + }, + { + "epoch": 4.651546391752578, + "grad_norm": 0.08962612308335349, + "learning_rate": 1.114654248754099e-06, + "loss": 0.4495, + "step": 1692 + }, + { + "epoch": 4.6542955326460484, + "grad_norm": 0.08766658442624285, + "learning_rate": 1.0966864092551233e-06, + "loss": 0.445, + "step": 1693 + }, + { + "epoch": 4.657044673539519, + "grad_norm": 0.08709960621618447, + "learning_rate": 1.0788625536240206e-06, + "loss": 0.4471, + "step": 1694 + }, + { + "epoch": 4.65979381443299, + "grad_norm": 0.09261985426164675, + "learning_rate": 1.0611827478281067e-06, + "loss": 0.4456, + "step": 1695 + }, + { + "epoch": 4.662542955326461, + "grad_norm": 0.08650196395240495, + "learning_rate": 1.0436470573015733e-06, + "loss": 0.4403, + "step": 1696 + }, + { + "epoch": 4.665292096219932, + "grad_norm": 0.08590676857219628, + "learning_rate": 1.026255546945234e-06, + "loss": 0.4449, + "step": 1697 + }, + { + "epoch": 4.668041237113402, + "grad_norm": 0.08741332778858373, + "learning_rate": 1.0090082811262802e-06, + "loss": 0.4492, + "step": 1698 + }, + { + "epoch": 4.670790378006873, + "grad_norm": 0.08738071184604224, + "learning_rate": 9.919053236780328e-07, + "loss": 0.444, + "step": 1699 + }, + { + "epoch": 4.673539518900344, + "grad_norm": 0.08949580602300157, + "learning_rate": 9.74946737899729e-07, + "loss": 0.4376, + "step": 1700 + }, + { + "epoch": 4.676288659793815, + "grad_norm": 0.08859670553407732, + "learning_rate": 9.581325865562775e-07, + "loss": 0.4486, + "step": 1701 + }, + { + "epoch": 4.6790378006872855, + "grad_norm": 0.08694870133225886, + "learning_rate": 9.414629318780189e-07, + "loss": 0.4483, + "step": 1702 + }, + { + "epoch": 4.681786941580756, + "grad_norm": 0.08702985035490321, + "learning_rate": 9.249378355605043e-07, + "loss": 0.4419, + "step": 1703 + }, + { + "epoch": 4.684536082474227, + "grad_norm": 0.08584597159667766, + "learning_rate": 9.085573587642637e-07, + "loss": 0.4517, + "step": 1704 + }, + { + "epoch": 4.687285223367698, + "grad_norm": 0.08578695182725778, + "learning_rate": 8.923215621145753e-07, + "loss": 0.4399, + "step": 1705 + }, + { + "epoch": 4.690034364261169, + "grad_norm": 0.08592571403893629, + "learning_rate": 8.762305057012654e-07, + "loss": 0.4515, + "step": 1706 + }, + { + "epoch": 4.692783505154639, + "grad_norm": 0.08693493065817048, + "learning_rate": 8.60284249078438e-07, + "loss": 0.4461, + "step": 1707 + }, + { + "epoch": 4.69553264604811, + "grad_norm": 0.08872474686819293, + "learning_rate": 8.444828512642966e-07, + "loss": 0.4545, + "step": 1708 + }, + { + "epoch": 4.698281786941581, + "grad_norm": 0.09336681238214921, + "learning_rate": 8.288263707409272e-07, + "loss": 0.4543, + "step": 1709 + }, + { + "epoch": 4.701030927835052, + "grad_norm": 0.0862849071211661, + "learning_rate": 8.133148654540402e-07, + "loss": 0.4409, + "step": 1710 + }, + { + "epoch": 4.7037800687285225, + "grad_norm": 0.08478357720964781, + "learning_rate": 7.979483928127974e-07, + "loss": 0.4422, + "step": 1711 + }, + { + "epoch": 4.706529209621993, + "grad_norm": 0.08654065268244254, + "learning_rate": 7.827270096895811e-07, + "loss": 0.4479, + "step": 1712 + }, + { + "epoch": 4.709278350515464, + "grad_norm": 0.08647792501003562, + "learning_rate": 7.676507724197946e-07, + "loss": 0.4484, + "step": 1713 + }, + { + "epoch": 4.712027491408935, + "grad_norm": 0.08597489991064104, + "learning_rate": 7.527197368016437e-07, + "loss": 0.4417, + "step": 1714 + }, + { + "epoch": 4.714776632302406, + "grad_norm": 0.0898277616665661, + "learning_rate": 7.37933958095911e-07, + "loss": 0.4514, + "step": 1715 + }, + { + "epoch": 4.717525773195876, + "grad_norm": 0.08697107341130339, + "learning_rate": 7.232934910258004e-07, + "loss": 0.4456, + "step": 1716 + }, + { + "epoch": 4.720274914089347, + "grad_norm": 0.08637312554758676, + "learning_rate": 7.087983897767059e-07, + "loss": 0.4477, + "step": 1717 + }, + { + "epoch": 4.723024054982818, + "grad_norm": 0.0882085833054912, + "learning_rate": 6.944487079959982e-07, + "loss": 0.4395, + "step": 1718 + }, + { + "epoch": 4.725773195876289, + "grad_norm": 0.085015391595313, + "learning_rate": 6.802444987928436e-07, + "loss": 0.4395, + "step": 1719 + }, + { + "epoch": 4.7285223367697595, + "grad_norm": 0.08451996713089384, + "learning_rate": 6.661858147380118e-07, + "loss": 0.437, + "step": 1720 + }, + { + "epoch": 4.73127147766323, + "grad_norm": 0.08738636017583253, + "learning_rate": 6.522727078636681e-07, + "loss": 0.4508, + "step": 1721 + }, + { + "epoch": 4.734020618556701, + "grad_norm": 0.08558858353740155, + "learning_rate": 6.385052296631955e-07, + "loss": 0.4464, + "step": 1722 + }, + { + "epoch": 4.736769759450172, + "grad_norm": 0.08661882774459452, + "learning_rate": 6.248834310909768e-07, + "loss": 0.4464, + "step": 1723 + }, + { + "epoch": 4.739518900343643, + "grad_norm": 0.08559450787680475, + "learning_rate": 6.114073625622396e-07, + "loss": 0.4372, + "step": 1724 + }, + { + "epoch": 4.742268041237113, + "grad_norm": 0.0894381151555916, + "learning_rate": 5.980770739528563e-07, + "loss": 0.4394, + "step": 1725 + }, + { + "epoch": 4.745017182130584, + "grad_norm": 0.0897214937051112, + "learning_rate": 5.84892614599144e-07, + "loss": 0.4511, + "step": 1726 + }, + { + "epoch": 4.747766323024055, + "grad_norm": 0.08890953858034158, + "learning_rate": 5.718540332977007e-07, + "loss": 0.4504, + "step": 1727 + }, + { + "epoch": 4.750515463917526, + "grad_norm": 0.08567497991113034, + "learning_rate": 5.589613783052317e-07, + "loss": 0.4422, + "step": 1728 + }, + { + "epoch": 4.7532646048109966, + "grad_norm": 0.08728822000081427, + "learning_rate": 5.462146973383453e-07, + "loss": 0.4589, + "step": 1729 + }, + { + "epoch": 4.756013745704467, + "grad_norm": 0.08723350175277615, + "learning_rate": 5.336140375733934e-07, + "loss": 0.4444, + "step": 1730 + }, + { + "epoch": 4.758762886597938, + "grad_norm": 0.0895784856053207, + "learning_rate": 5.211594456462932e-07, + "loss": 0.4502, + "step": 1731 + }, + { + "epoch": 4.761512027491409, + "grad_norm": 0.08733169216974808, + "learning_rate": 5.088509676523545e-07, + "loss": 0.4482, + "step": 1732 + }, + { + "epoch": 4.76426116838488, + "grad_norm": 0.08877952523971627, + "learning_rate": 4.966886491461109e-07, + "loss": 0.4478, + "step": 1733 + }, + { + "epoch": 4.7670103092783505, + "grad_norm": 0.08664355725321998, + "learning_rate": 4.846725351411507e-07, + "loss": 0.4463, + "step": 1734 + }, + { + "epoch": 4.769759450171821, + "grad_norm": 0.08722460501048601, + "learning_rate": 4.7280267010993974e-07, + "loss": 0.4484, + "step": 1735 + }, + { + "epoch": 4.772508591065292, + "grad_norm": 0.0869689702035606, + "learning_rate": 4.6107909798368324e-07, + "loss": 0.4449, + "step": 1736 + }, + { + "epoch": 4.775257731958763, + "grad_norm": 0.09408455047074887, + "learning_rate": 4.495018621521352e-07, + "loss": 0.4436, + "step": 1737 + }, + { + "epoch": 4.778006872852234, + "grad_norm": 0.09000664206402534, + "learning_rate": 4.3807100546344296e-07, + "loss": 0.44, + "step": 1738 + }, + { + "epoch": 4.780756013745704, + "grad_norm": 0.08893238330077334, + "learning_rate": 4.267865702240048e-07, + "loss": 0.4418, + "step": 1739 + }, + { + "epoch": 4.783505154639175, + "grad_norm": 0.08753276685918031, + "learning_rate": 4.1564859819830607e-07, + "loss": 0.4407, + "step": 1740 + }, + { + "epoch": 4.786254295532646, + "grad_norm": 0.08418852751539188, + "learning_rate": 4.046571306087499e-07, + "loss": 0.4429, + "step": 1741 + }, + { + "epoch": 4.789003436426117, + "grad_norm": 0.0880689357278319, + "learning_rate": 3.9381220813551555e-07, + "loss": 0.4517, + "step": 1742 + }, + { + "epoch": 4.7917525773195875, + "grad_norm": 0.08935991058559813, + "learning_rate": 3.831138709164295e-07, + "loss": 0.4464, + "step": 1743 + }, + { + "epoch": 4.794501718213058, + "grad_norm": 0.086751085582121, + "learning_rate": 3.725621585467698e-07, + "loss": 0.4504, + "step": 1744 + }, + { + "epoch": 4.797250859106529, + "grad_norm": 0.08660059471433466, + "learning_rate": 3.6215711007916434e-07, + "loss": 0.4457, + "step": 1745 + }, + { + "epoch": 4.8, + "grad_norm": 0.08795409500802658, + "learning_rate": 3.5189876402341727e-07, + "loss": 0.4487, + "step": 1746 + }, + { + "epoch": 4.802749140893471, + "grad_norm": 0.08517686102366333, + "learning_rate": 3.417871583463805e-07, + "loss": 0.442, + "step": 1747 + }, + { + "epoch": 4.805498281786941, + "grad_norm": 0.08656026625325718, + "learning_rate": 3.3182233047181154e-07, + "loss": 0.4482, + "step": 1748 + }, + { + "epoch": 4.808247422680412, + "grad_norm": 0.08742296055893738, + "learning_rate": 3.2200431728022676e-07, + "loss": 0.4497, + "step": 1749 + }, + { + "epoch": 4.810996563573883, + "grad_norm": 0.08706269427544457, + "learning_rate": 3.1233315510877714e-07, + "loss": 0.4399, + "step": 1750 + }, + { + "epoch": 4.813745704467354, + "grad_norm": 0.08541652658402427, + "learning_rate": 3.0280887975111087e-07, + "loss": 0.4463, + "step": 1751 + }, + { + "epoch": 4.8164948453608245, + "grad_norm": 0.08561518369119724, + "learning_rate": 2.9343152645723075e-07, + "loss": 0.4496, + "step": 1752 + }, + { + "epoch": 4.819243986254295, + "grad_norm": 0.08671209749993644, + "learning_rate": 2.8420112993337026e-07, + "loss": 0.4441, + "step": 1753 + }, + { + "epoch": 4.821993127147766, + "grad_norm": 0.0868113057408611, + "learning_rate": 2.75117724341869e-07, + "loss": 0.4539, + "step": 1754 + }, + { + "epoch": 4.824742268041237, + "grad_norm": 0.08787299513468588, + "learning_rate": 2.661813433010485e-07, + "loss": 0.4469, + "step": 1755 + }, + { + "epoch": 4.827491408934708, + "grad_norm": 0.08426954798208526, + "learning_rate": 2.5739201988506544e-07, + "loss": 0.4452, + "step": 1756 + }, + { + "epoch": 4.830240549828178, + "grad_norm": 0.08925999331930048, + "learning_rate": 2.487497866238231e-07, + "loss": 0.4431, + "step": 1757 + }, + { + "epoch": 4.832989690721649, + "grad_norm": 0.09009930022107024, + "learning_rate": 2.4025467550283345e-07, + "loss": 0.4428, + "step": 1758 + }, + { + "epoch": 4.83573883161512, + "grad_norm": 0.08903112575869974, + "learning_rate": 2.3190671796307963e-07, + "loss": 0.4376, + "step": 1759 + }, + { + "epoch": 4.838487972508591, + "grad_norm": 0.087097271598177, + "learning_rate": 2.237059449009449e-07, + "loss": 0.4514, + "step": 1760 + }, + { + "epoch": 4.8412371134020615, + "grad_norm": 0.08665721242897642, + "learning_rate": 2.1565238666805266e-07, + "loss": 0.4507, + "step": 1761 + }, + { + "epoch": 4.843986254295532, + "grad_norm": 0.08570901423189047, + "learning_rate": 2.0774607307118665e-07, + "loss": 0.444, + "step": 1762 + }, + { + "epoch": 4.846735395189003, + "grad_norm": 0.08855478539796897, + "learning_rate": 1.9998703337216207e-07, + "loss": 0.4459, + "step": 1763 + }, + { + "epoch": 4.849484536082474, + "grad_norm": 0.08830013933034192, + "learning_rate": 1.9237529628772345e-07, + "loss": 0.4471, + "step": 1764 + }, + { + "epoch": 4.852233676975945, + "grad_norm": 0.08632909822266122, + "learning_rate": 1.8491088998943806e-07, + "loss": 0.4443, + "step": 1765 + }, + { + "epoch": 4.854982817869415, + "grad_norm": 0.08788988835658808, + "learning_rate": 1.7759384210358943e-07, + "loss": 0.4495, + "step": 1766 + }, + { + "epoch": 4.857731958762886, + "grad_norm": 0.08641182871335162, + "learning_rate": 1.7042417971108837e-07, + "loss": 0.448, + "step": 1767 + }, + { + "epoch": 4.860481099656358, + "grad_norm": 0.08684528157216327, + "learning_rate": 1.6340192934734879e-07, + "loss": 0.4503, + "step": 1768 + }, + { + "epoch": 4.863230240549829, + "grad_norm": 0.08740681191652898, + "learning_rate": 1.5652711700220756e-07, + "loss": 0.4476, + "step": 1769 + }, + { + "epoch": 4.8659793814432994, + "grad_norm": 0.08455281840311087, + "learning_rate": 1.4979976811982267e-07, + "loss": 0.4501, + "step": 1770 + }, + { + "epoch": 4.86872852233677, + "grad_norm": 0.08621142637068548, + "learning_rate": 1.4321990759858406e-07, + "loss": 0.4427, + "step": 1771 + }, + { + "epoch": 4.871477663230241, + "grad_norm": 0.09048418177504676, + "learning_rate": 1.3678755979100734e-07, + "loss": 0.4414, + "step": 1772 + }, + { + "epoch": 4.874226804123712, + "grad_norm": 0.08552997152050767, + "learning_rate": 1.3050274850366252e-07, + "loss": 0.4452, + "step": 1773 + }, + { + "epoch": 4.876975945017183, + "grad_norm": 0.09163336347075388, + "learning_rate": 1.2436549699706756e-07, + "loss": 0.4421, + "step": 1774 + }, + { + "epoch": 4.879725085910653, + "grad_norm": 0.08581324177498632, + "learning_rate": 1.183758279856262e-07, + "loss": 0.445, + "step": 1775 + }, + { + "epoch": 4.882474226804124, + "grad_norm": 0.08520078540460027, + "learning_rate": 1.1253376363751234e-07, + "loss": 0.4398, + "step": 1776 + }, + { + "epoch": 4.885223367697595, + "grad_norm": 0.08389381645053694, + "learning_rate": 1.0683932557461696e-07, + "loss": 0.4509, + "step": 1777 + }, + { + "epoch": 4.887972508591066, + "grad_norm": 0.08463720514626481, + "learning_rate": 1.0129253487244584e-07, + "loss": 0.444, + "step": 1778 + }, + { + "epoch": 4.8907216494845365, + "grad_norm": 0.08376091905777829, + "learning_rate": 9.589341206006186e-08, + "loss": 0.4424, + "step": 1779 + }, + { + "epoch": 4.893470790378007, + "grad_norm": 0.0864473265773848, + "learning_rate": 9.064197711999178e-08, + "loss": 0.436, + "step": 1780 + }, + { + "epoch": 4.896219931271478, + "grad_norm": 0.08780409654999521, + "learning_rate": 8.553824948816402e-08, + "loss": 0.4439, + "step": 1781 + }, + { + "epoch": 4.898969072164949, + "grad_norm": 0.08643060928003499, + "learning_rate": 8.058224805382431e-08, + "loss": 0.457, + "step": 1782 + }, + { + "epoch": 4.90171821305842, + "grad_norm": 0.08999003837315114, + "learning_rate": 7.577399115948236e-08, + "loss": 0.4536, + "step": 1783 + }, + { + "epoch": 4.90446735395189, + "grad_norm": 0.08500920989818493, + "learning_rate": 7.111349660082756e-08, + "loss": 0.4524, + "step": 1784 + }, + { + "epoch": 4.907216494845361, + "grad_norm": 0.08713394706241256, + "learning_rate": 6.660078162667561e-08, + "loss": 0.4474, + "step": 1785 + }, + { + "epoch": 4.909965635738832, + "grad_norm": 0.08415301103290775, + "learning_rate": 6.223586293890193e-08, + "loss": 0.4388, + "step": 1786 + }, + { + "epoch": 4.912714776632303, + "grad_norm": 0.08641909751480724, + "learning_rate": 5.8018756692370673e-08, + "loss": 0.4452, + "step": 1787 + }, + { + "epoch": 4.9154639175257735, + "grad_norm": 0.085407296137395, + "learning_rate": 5.3949478494885745e-08, + "loss": 0.4442, + "step": 1788 + }, + { + "epoch": 4.918213058419244, + "grad_norm": 0.08620621228120745, + "learning_rate": 5.002804340712875e-08, + "loss": 0.4395, + "step": 1789 + }, + { + "epoch": 4.920962199312715, + "grad_norm": 0.08440435770555102, + "learning_rate": 4.625446594260563e-08, + "loss": 0.4475, + "step": 1790 + }, + { + "epoch": 4.923711340206186, + "grad_norm": 0.08498966775862096, + "learning_rate": 4.262876006758454e-08, + "loss": 0.4461, + "step": 1791 + }, + { + "epoch": 4.926460481099657, + "grad_norm": 0.0852940765401518, + "learning_rate": 3.915093920105584e-08, + "loss": 0.4508, + "step": 1792 + }, + { + "epoch": 4.929209621993127, + "grad_norm": 0.08708725826340913, + "learning_rate": 3.582101621467882e-08, + "loss": 0.4478, + "step": 1793 + }, + { + "epoch": 4.931958762886598, + "grad_norm": 0.08563954373853248, + "learning_rate": 3.263900343272841e-08, + "loss": 0.4506, + "step": 1794 + }, + { + "epoch": 4.934707903780069, + "grad_norm": 0.08351208793444787, + "learning_rate": 2.960491263205523e-08, + "loss": 0.4358, + "step": 1795 + }, + { + "epoch": 4.93745704467354, + "grad_norm": 0.08968304296903096, + "learning_rate": 2.6718755042041134e-08, + "loss": 0.4408, + "step": 1796 + }, + { + "epoch": 4.9402061855670105, + "grad_norm": 0.08687120751973657, + "learning_rate": 2.3980541344554852e-08, + "loss": 0.4472, + "step": 1797 + }, + { + "epoch": 4.942955326460481, + "grad_norm": 0.0843041818695698, + "learning_rate": 2.1390281673916435e-08, + "loss": 0.4433, + "step": 1798 + }, + { + "epoch": 4.945704467353952, + "grad_norm": 0.08678186311343745, + "learning_rate": 1.894798561685729e-08, + "loss": 0.4401, + "step": 1799 + }, + { + "epoch": 4.948453608247423, + "grad_norm": 0.08552567460187925, + "learning_rate": 1.6653662212484656e-08, + "loss": 0.4463, + "step": 1800 + }, + { + "epoch": 4.951202749140894, + "grad_norm": 0.08327662566119383, + "learning_rate": 1.4507319952246079e-08, + "loss": 0.4428, + "step": 1801 + }, + { + "epoch": 4.953951890034364, + "grad_norm": 0.08558816066436331, + "learning_rate": 1.2508966779907206e-08, + "loss": 0.4439, + "step": 1802 + }, + { + "epoch": 4.956701030927835, + "grad_norm": 0.08685318316335923, + "learning_rate": 1.0658610091507371e-08, + "loss": 0.4433, + "step": 1803 + }, + { + "epoch": 4.959450171821306, + "grad_norm": 0.08552301347515416, + "learning_rate": 8.956256735355162e-09, + "loss": 0.4462, + "step": 1804 + }, + { + "epoch": 4.962199312714777, + "grad_norm": 0.08918690199355128, + "learning_rate": 7.401913011966244e-09, + "loss": 0.4516, + "step": 1805 + }, + { + "epoch": 4.9649484536082475, + "grad_norm": 0.08439607746487938, + "learning_rate": 5.995584674085564e-09, + "loss": 0.4443, + "step": 1806 + }, + { + "epoch": 4.967697594501718, + "grad_norm": 0.0859086823143405, + "learning_rate": 4.737276926620738e-09, + "loss": 0.4477, + "step": 1807 + }, + { + "epoch": 4.970446735395189, + "grad_norm": 0.08603802799919767, + "learning_rate": 3.6269944266686953e-09, + "loss": 0.4398, + "step": 1808 + }, + { + "epoch": 4.97319587628866, + "grad_norm": 0.08797273612741212, + "learning_rate": 2.664741283453509e-09, + "loss": 0.4455, + "step": 1809 + }, + { + "epoch": 4.975945017182131, + "grad_norm": 0.08498147628736628, + "learning_rate": 1.8505210583441568e-09, + "loss": 0.4417, + "step": 1810 + }, + { + "epoch": 4.9786941580756015, + "grad_norm": 0.08573341738089828, + "learning_rate": 1.184336764823435e-09, + "loss": 0.4457, + "step": 1811 + }, + { + "epoch": 4.981443298969072, + "grad_norm": 0.08322231115466387, + "learning_rate": 6.661908684924002e-10, + "loss": 0.454, + "step": 1812 + }, + { + "epoch": 4.984192439862543, + "grad_norm": 0.08656315765889926, + "learning_rate": 2.9608528703928274e-10, + "loss": 0.4449, + "step": 1813 + }, + { + "epoch": 4.986941580756014, + "grad_norm": 0.08364091527237393, + "learning_rate": 7.402139024836886e-11, + "loss": 0.455, + "step": 1814 + }, + { + "epoch": 4.989690721649485, + "grad_norm": 0.08519652883753787, + "learning_rate": 0.0, + "loss": 0.4445, + "step": 1815 + }, + { + "epoch": 4.989690721649485, + "step": 1815, + "total_flos": 3.73703706563429e+19, + "train_loss": 0.5429748183767986, + "train_runtime": 94067.7478, + "train_samples_per_second": 9.898, + "train_steps_per_second": 0.019 + } + ], + "logging_steps": 1.0, + "max_steps": 1815, + "num_input_tokens_seen": 0, + "num_train_epochs": 5, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 3.73703706563429e+19, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +}