| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 1.0, |
| "eval_steps": 500, |
| "global_step": 518, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.0019305019305019305, |
| "grad_norm": 9.087823867797852, |
| "learning_rate": 9.980694980694981e-06, |
| "loss": 763.5779, |
| "step": 1 |
| }, |
| { |
| "epoch": 0.003861003861003861, |
| "grad_norm": 8.02014446258545, |
| "learning_rate": 9.961389961389962e-06, |
| "loss": 742.0983, |
| "step": 2 |
| }, |
| { |
| "epoch": 0.005791505791505791, |
| "grad_norm": 6.846795558929443, |
| "learning_rate": 9.942084942084944e-06, |
| "loss": 754.556, |
| "step": 3 |
| }, |
| { |
| "epoch": 0.007722007722007722, |
| "grad_norm": 5.642313003540039, |
| "learning_rate": 9.922779922779924e-06, |
| "loss": 760.8958, |
| "step": 4 |
| }, |
| { |
| "epoch": 0.009652509652509652, |
| "grad_norm": 4.954174995422363, |
| "learning_rate": 9.903474903474905e-06, |
| "loss": 773.4886, |
| "step": 5 |
| }, |
| { |
| "epoch": 0.011583011583011582, |
| "grad_norm": 4.209826946258545, |
| "learning_rate": 9.884169884169885e-06, |
| "loss": 752.0172, |
| "step": 6 |
| }, |
| { |
| "epoch": 0.013513513513513514, |
| "grad_norm": 3.8238916397094727, |
| "learning_rate": 9.864864864864865e-06, |
| "loss": 765.1586, |
| "step": 7 |
| }, |
| { |
| "epoch": 0.015444015444015444, |
| "grad_norm": 3.472975730895996, |
| "learning_rate": 9.845559845559847e-06, |
| "loss": 758.0544, |
| "step": 8 |
| }, |
| { |
| "epoch": 0.017374517374517374, |
| "grad_norm": 3.3365626335144043, |
| "learning_rate": 9.826254826254828e-06, |
| "loss": 760.6899, |
| "step": 9 |
| }, |
| { |
| "epoch": 0.019305019305019305, |
| "grad_norm": 3.205853223800659, |
| "learning_rate": 9.806949806949808e-06, |
| "loss": 763.1088, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.021235521235521235, |
| "grad_norm": 2.9912545680999756, |
| "learning_rate": 9.787644787644789e-06, |
| "loss": 751.3445, |
| "step": 11 |
| }, |
| { |
| "epoch": 0.023166023166023165, |
| "grad_norm": 2.9894907474517822, |
| "learning_rate": 9.768339768339769e-06, |
| "loss": 762.5296, |
| "step": 12 |
| }, |
| { |
| "epoch": 0.025096525096525095, |
| "grad_norm": 2.973778009414673, |
| "learning_rate": 9.749034749034751e-06, |
| "loss": 766.5692, |
| "step": 13 |
| }, |
| { |
| "epoch": 0.02702702702702703, |
| "grad_norm": 2.866206645965576, |
| "learning_rate": 9.729729729729732e-06, |
| "loss": 759.659, |
| "step": 14 |
| }, |
| { |
| "epoch": 0.02895752895752896, |
| "grad_norm": 2.7454729080200195, |
| "learning_rate": 9.71042471042471e-06, |
| "loss": 749.6011, |
| "step": 15 |
| }, |
| { |
| "epoch": 0.03088803088803089, |
| "grad_norm": 2.7870497703552246, |
| "learning_rate": 9.69111969111969e-06, |
| "loss": 757.2128, |
| "step": 16 |
| }, |
| { |
| "epoch": 0.032818532818532815, |
| "grad_norm": 2.727757453918457, |
| "learning_rate": 9.671814671814673e-06, |
| "loss": 752.8752, |
| "step": 17 |
| }, |
| { |
| "epoch": 0.03474903474903475, |
| "grad_norm": 2.761842966079712, |
| "learning_rate": 9.652509652509653e-06, |
| "loss": 763.5405, |
| "step": 18 |
| }, |
| { |
| "epoch": 0.03667953667953668, |
| "grad_norm": 2.64555025100708, |
| "learning_rate": 9.633204633204634e-06, |
| "loss": 752.1604, |
| "step": 19 |
| }, |
| { |
| "epoch": 0.03861003861003861, |
| "grad_norm": 2.604011297225952, |
| "learning_rate": 9.613899613899614e-06, |
| "loss": 745.1989, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.04054054054054054, |
| "grad_norm": 2.5989909172058105, |
| "learning_rate": 9.594594594594594e-06, |
| "loss": 744.575, |
| "step": 21 |
| }, |
| { |
| "epoch": 0.04247104247104247, |
| "grad_norm": 2.646167278289795, |
| "learning_rate": 9.575289575289576e-06, |
| "loss": 753.7839, |
| "step": 22 |
| }, |
| { |
| "epoch": 0.0444015444015444, |
| "grad_norm": 2.6674346923828125, |
| "learning_rate": 9.555984555984557e-06, |
| "loss": 764.2125, |
| "step": 23 |
| }, |
| { |
| "epoch": 0.04633204633204633, |
| "grad_norm": 2.6234052181243896, |
| "learning_rate": 9.536679536679537e-06, |
| "loss": 753.5481, |
| "step": 24 |
| }, |
| { |
| "epoch": 0.04826254826254826, |
| "grad_norm": 2.550206422805786, |
| "learning_rate": 9.517374517374518e-06, |
| "loss": 744.3314, |
| "step": 25 |
| }, |
| { |
| "epoch": 0.05019305019305019, |
| "grad_norm": 2.634002208709717, |
| "learning_rate": 9.498069498069498e-06, |
| "loss": 756.7723, |
| "step": 26 |
| }, |
| { |
| "epoch": 0.052123552123552123, |
| "grad_norm": 2.632336378097534, |
| "learning_rate": 9.47876447876448e-06, |
| "loss": 756.9872, |
| "step": 27 |
| }, |
| { |
| "epoch": 0.05405405405405406, |
| "grad_norm": 2.628897190093994, |
| "learning_rate": 9.45945945945946e-06, |
| "loss": 756.0316, |
| "step": 28 |
| }, |
| { |
| "epoch": 0.055984555984555984, |
| "grad_norm": 2.6032018661499023, |
| "learning_rate": 9.440154440154441e-06, |
| "loss": 754.0662, |
| "step": 29 |
| }, |
| { |
| "epoch": 0.05791505791505792, |
| "grad_norm": 2.5998055934906006, |
| "learning_rate": 9.420849420849421e-06, |
| "loss": 749.3468, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.059845559845559844, |
| "grad_norm": 2.623134136199951, |
| "learning_rate": 9.401544401544402e-06, |
| "loss": 753.1194, |
| "step": 31 |
| }, |
| { |
| "epoch": 0.06177606177606178, |
| "grad_norm": 2.636357069015503, |
| "learning_rate": 9.382239382239384e-06, |
| "loss": 755.1291, |
| "step": 32 |
| }, |
| { |
| "epoch": 0.0637065637065637, |
| "grad_norm": 2.709622621536255, |
| "learning_rate": 9.362934362934364e-06, |
| "loss": 774.9333, |
| "step": 33 |
| }, |
| { |
| "epoch": 0.06563706563706563, |
| "grad_norm": 2.6019175052642822, |
| "learning_rate": 9.343629343629345e-06, |
| "loss": 758.8455, |
| "step": 34 |
| }, |
| { |
| "epoch": 0.06756756756756757, |
| "grad_norm": 2.620014190673828, |
| "learning_rate": 9.324324324324325e-06, |
| "loss": 751.6978, |
| "step": 35 |
| }, |
| { |
| "epoch": 0.0694980694980695, |
| "grad_norm": 2.6363117694854736, |
| "learning_rate": 9.305019305019305e-06, |
| "loss": 755.4987, |
| "step": 36 |
| }, |
| { |
| "epoch": 0.07142857142857142, |
| "grad_norm": 2.697289228439331, |
| "learning_rate": 9.285714285714288e-06, |
| "loss": 769.0797, |
| "step": 37 |
| }, |
| { |
| "epoch": 0.07335907335907337, |
| "grad_norm": 2.684602975845337, |
| "learning_rate": 9.266409266409268e-06, |
| "loss": 771.785, |
| "step": 38 |
| }, |
| { |
| "epoch": 0.07528957528957529, |
| "grad_norm": 2.6869068145751953, |
| "learning_rate": 9.247104247104248e-06, |
| "loss": 771.0751, |
| "step": 39 |
| }, |
| { |
| "epoch": 0.07722007722007722, |
| "grad_norm": 2.65386962890625, |
| "learning_rate": 9.227799227799229e-06, |
| "loss": 766.6141, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.07915057915057915, |
| "grad_norm": 2.5492303371429443, |
| "learning_rate": 9.20849420849421e-06, |
| "loss": 741.012, |
| "step": 41 |
| }, |
| { |
| "epoch": 0.08108108108108109, |
| "grad_norm": 2.659393548965454, |
| "learning_rate": 9.189189189189191e-06, |
| "loss": 761.9696, |
| "step": 42 |
| }, |
| { |
| "epoch": 0.08301158301158301, |
| "grad_norm": 2.6672518253326416, |
| "learning_rate": 9.169884169884172e-06, |
| "loss": 762.8466, |
| "step": 43 |
| }, |
| { |
| "epoch": 0.08494208494208494, |
| "grad_norm": 2.60377836227417, |
| "learning_rate": 9.15057915057915e-06, |
| "loss": 758.5039, |
| "step": 44 |
| }, |
| { |
| "epoch": 0.08687258687258688, |
| "grad_norm": 2.622608184814453, |
| "learning_rate": 9.13127413127413e-06, |
| "loss": 750.5551, |
| "step": 45 |
| }, |
| { |
| "epoch": 0.0888030888030888, |
| "grad_norm": 2.6671438217163086, |
| "learning_rate": 9.111969111969113e-06, |
| "loss": 766.5024, |
| "step": 46 |
| }, |
| { |
| "epoch": 0.09073359073359073, |
| "grad_norm": 2.5831432342529297, |
| "learning_rate": 9.092664092664093e-06, |
| "loss": 747.624, |
| "step": 47 |
| }, |
| { |
| "epoch": 0.09266409266409266, |
| "grad_norm": 2.61625075340271, |
| "learning_rate": 9.073359073359074e-06, |
| "loss": 759.9783, |
| "step": 48 |
| }, |
| { |
| "epoch": 0.0945945945945946, |
| "grad_norm": 2.6503899097442627, |
| "learning_rate": 9.054054054054054e-06, |
| "loss": 766.5831, |
| "step": 49 |
| }, |
| { |
| "epoch": 0.09652509652509653, |
| "grad_norm": 2.713566303253174, |
| "learning_rate": 9.034749034749034e-06, |
| "loss": 764.6612, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.09845559845559845, |
| "grad_norm": 2.6607675552368164, |
| "learning_rate": 9.015444015444017e-06, |
| "loss": 759.2146, |
| "step": 51 |
| }, |
| { |
| "epoch": 0.10038610038610038, |
| "grad_norm": 2.703293800354004, |
| "learning_rate": 8.996138996138997e-06, |
| "loss": 771.3636, |
| "step": 52 |
| }, |
| { |
| "epoch": 0.10231660231660232, |
| "grad_norm": 2.6380341053009033, |
| "learning_rate": 8.976833976833977e-06, |
| "loss": 762.6519, |
| "step": 53 |
| }, |
| { |
| "epoch": 0.10424710424710425, |
| "grad_norm": 2.6804656982421875, |
| "learning_rate": 8.957528957528958e-06, |
| "loss": 764.4567, |
| "step": 54 |
| }, |
| { |
| "epoch": 0.10617760617760617, |
| "grad_norm": 2.6722824573516846, |
| "learning_rate": 8.938223938223938e-06, |
| "loss": 762.7928, |
| "step": 55 |
| }, |
| { |
| "epoch": 0.10810810810810811, |
| "grad_norm": 2.5956473350524902, |
| "learning_rate": 8.91891891891892e-06, |
| "loss": 757.1573, |
| "step": 56 |
| }, |
| { |
| "epoch": 0.11003861003861004, |
| "grad_norm": 2.67158579826355, |
| "learning_rate": 8.8996138996139e-06, |
| "loss": 762.9989, |
| "step": 57 |
| }, |
| { |
| "epoch": 0.11196911196911197, |
| "grad_norm": 2.6719319820404053, |
| "learning_rate": 8.880308880308881e-06, |
| "loss": 765.219, |
| "step": 58 |
| }, |
| { |
| "epoch": 0.1138996138996139, |
| "grad_norm": 2.690995931625366, |
| "learning_rate": 8.861003861003861e-06, |
| "loss": 755.7704, |
| "step": 59 |
| }, |
| { |
| "epoch": 0.11583011583011583, |
| "grad_norm": 2.672006845474243, |
| "learning_rate": 8.841698841698842e-06, |
| "loss": 755.6411, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.11776061776061776, |
| "grad_norm": 2.6625969409942627, |
| "learning_rate": 8.822393822393824e-06, |
| "loss": 755.6705, |
| "step": 61 |
| }, |
| { |
| "epoch": 0.11969111969111969, |
| "grad_norm": 2.6644089221954346, |
| "learning_rate": 8.803088803088804e-06, |
| "loss": 755.7343, |
| "step": 62 |
| }, |
| { |
| "epoch": 0.12162162162162163, |
| "grad_norm": 2.6913063526153564, |
| "learning_rate": 8.783783783783785e-06, |
| "loss": 760.8112, |
| "step": 63 |
| }, |
| { |
| "epoch": 0.12355212355212356, |
| "grad_norm": 2.640939474105835, |
| "learning_rate": 8.764478764478765e-06, |
| "loss": 752.9957, |
| "step": 64 |
| }, |
| { |
| "epoch": 0.12548262548262548, |
| "grad_norm": 2.706070899963379, |
| "learning_rate": 8.745173745173746e-06, |
| "loss": 751.4651, |
| "step": 65 |
| }, |
| { |
| "epoch": 0.1274131274131274, |
| "grad_norm": 2.6751809120178223, |
| "learning_rate": 8.725868725868728e-06, |
| "loss": 760.4512, |
| "step": 66 |
| }, |
| { |
| "epoch": 0.12934362934362933, |
| "grad_norm": 2.731750965118408, |
| "learning_rate": 8.706563706563708e-06, |
| "loss": 766.3732, |
| "step": 67 |
| }, |
| { |
| "epoch": 0.13127413127413126, |
| "grad_norm": 2.688737392425537, |
| "learning_rate": 8.687258687258689e-06, |
| "loss": 749.8923, |
| "step": 68 |
| }, |
| { |
| "epoch": 0.13320463320463322, |
| "grad_norm": 2.67354679107666, |
| "learning_rate": 8.667953667953669e-06, |
| "loss": 751.1083, |
| "step": 69 |
| }, |
| { |
| "epoch": 0.13513513513513514, |
| "grad_norm": 2.6414129734039307, |
| "learning_rate": 8.64864864864865e-06, |
| "loss": 733.0116, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.13706563706563707, |
| "grad_norm": 3.0136189460754395, |
| "learning_rate": 8.629343629343631e-06, |
| "loss": 764.5856, |
| "step": 71 |
| }, |
| { |
| "epoch": 0.138996138996139, |
| "grad_norm": 2.7174079418182373, |
| "learning_rate": 8.61003861003861e-06, |
| "loss": 753.4861, |
| "step": 72 |
| }, |
| { |
| "epoch": 0.14092664092664092, |
| "grad_norm": 2.714317560195923, |
| "learning_rate": 8.59073359073359e-06, |
| "loss": 754.7225, |
| "step": 73 |
| }, |
| { |
| "epoch": 0.14285714285714285, |
| "grad_norm": 2.7697298526763916, |
| "learning_rate": 8.571428571428571e-06, |
| "loss": 765.4945, |
| "step": 74 |
| }, |
| { |
| "epoch": 0.14478764478764478, |
| "grad_norm": 2.7908132076263428, |
| "learning_rate": 8.552123552123553e-06, |
| "loss": 779.0895, |
| "step": 75 |
| }, |
| { |
| "epoch": 0.14671814671814673, |
| "grad_norm": 2.7896623611450195, |
| "learning_rate": 8.532818532818533e-06, |
| "loss": 767.8585, |
| "step": 76 |
| }, |
| { |
| "epoch": 0.14864864864864866, |
| "grad_norm": 2.718611240386963, |
| "learning_rate": 8.513513513513514e-06, |
| "loss": 760.3116, |
| "step": 77 |
| }, |
| { |
| "epoch": 0.15057915057915058, |
| "grad_norm": 2.78110671043396, |
| "learning_rate": 8.494208494208494e-06, |
| "loss": 761.649, |
| "step": 78 |
| }, |
| { |
| "epoch": 0.1525096525096525, |
| "grad_norm": 2.822859048843384, |
| "learning_rate": 8.474903474903475e-06, |
| "loss": 766.2858, |
| "step": 79 |
| }, |
| { |
| "epoch": 0.15444015444015444, |
| "grad_norm": 2.719345808029175, |
| "learning_rate": 8.455598455598457e-06, |
| "loss": 755.3741, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.15637065637065636, |
| "grad_norm": 2.710378408432007, |
| "learning_rate": 8.436293436293437e-06, |
| "loss": 758.8147, |
| "step": 81 |
| }, |
| { |
| "epoch": 0.1583011583011583, |
| "grad_norm": 2.7957656383514404, |
| "learning_rate": 8.416988416988418e-06, |
| "loss": 761.3373, |
| "step": 82 |
| }, |
| { |
| "epoch": 0.16023166023166024, |
| "grad_norm": 2.8298914432525635, |
| "learning_rate": 8.397683397683398e-06, |
| "loss": 769.8754, |
| "step": 83 |
| }, |
| { |
| "epoch": 0.16216216216216217, |
| "grad_norm": 2.762465715408325, |
| "learning_rate": 8.378378378378378e-06, |
| "loss": 769.7825, |
| "step": 84 |
| }, |
| { |
| "epoch": 0.1640926640926641, |
| "grad_norm": 2.773038625717163, |
| "learning_rate": 8.35907335907336e-06, |
| "loss": 749.3358, |
| "step": 85 |
| }, |
| { |
| "epoch": 0.16602316602316602, |
| "grad_norm": 2.7703516483306885, |
| "learning_rate": 8.33976833976834e-06, |
| "loss": 757.9563, |
| "step": 86 |
| }, |
| { |
| "epoch": 0.16795366795366795, |
| "grad_norm": 2.930047035217285, |
| "learning_rate": 8.320463320463321e-06, |
| "loss": 764.3655, |
| "step": 87 |
| }, |
| { |
| "epoch": 0.16988416988416988, |
| "grad_norm": 2.7261102199554443, |
| "learning_rate": 8.301158301158302e-06, |
| "loss": 748.2144, |
| "step": 88 |
| }, |
| { |
| "epoch": 0.1718146718146718, |
| "grad_norm": 2.793009042739868, |
| "learning_rate": 8.281853281853282e-06, |
| "loss": 754.8672, |
| "step": 89 |
| }, |
| { |
| "epoch": 0.17374517374517376, |
| "grad_norm": 2.8894758224487305, |
| "learning_rate": 8.262548262548264e-06, |
| "loss": 764.7795, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.17567567567567569, |
| "grad_norm": 2.8510689735412598, |
| "learning_rate": 8.243243243243245e-06, |
| "loss": 753.4908, |
| "step": 91 |
| }, |
| { |
| "epoch": 0.1776061776061776, |
| "grad_norm": 2.757417917251587, |
| "learning_rate": 8.223938223938225e-06, |
| "loss": 760.0729, |
| "step": 92 |
| }, |
| { |
| "epoch": 0.17953667953667954, |
| "grad_norm": 2.8439671993255615, |
| "learning_rate": 8.204633204633205e-06, |
| "loss": 771.863, |
| "step": 93 |
| }, |
| { |
| "epoch": 0.18146718146718147, |
| "grad_norm": 2.8313705921173096, |
| "learning_rate": 8.185328185328186e-06, |
| "loss": 765.174, |
| "step": 94 |
| }, |
| { |
| "epoch": 0.1833976833976834, |
| "grad_norm": 2.7856626510620117, |
| "learning_rate": 8.166023166023168e-06, |
| "loss": 758.1021, |
| "step": 95 |
| }, |
| { |
| "epoch": 0.18532818532818532, |
| "grad_norm": 2.8734123706817627, |
| "learning_rate": 8.146718146718148e-06, |
| "loss": 768.5622, |
| "step": 96 |
| }, |
| { |
| "epoch": 0.18725868725868725, |
| "grad_norm": 2.833714723587036, |
| "learning_rate": 8.127413127413129e-06, |
| "loss": 753.1698, |
| "step": 97 |
| }, |
| { |
| "epoch": 0.1891891891891892, |
| "grad_norm": 2.8014986515045166, |
| "learning_rate": 8.108108108108109e-06, |
| "loss": 752.7419, |
| "step": 98 |
| }, |
| { |
| "epoch": 0.19111969111969113, |
| "grad_norm": 2.7595949172973633, |
| "learning_rate": 8.08880308880309e-06, |
| "loss": 755.9596, |
| "step": 99 |
| }, |
| { |
| "epoch": 0.19305019305019305, |
| "grad_norm": 2.7833123207092285, |
| "learning_rate": 8.06949806949807e-06, |
| "loss": 765.9607, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.19498069498069498, |
| "grad_norm": 2.777548313140869, |
| "learning_rate": 8.05019305019305e-06, |
| "loss": 766.981, |
| "step": 101 |
| }, |
| { |
| "epoch": 0.1969111969111969, |
| "grad_norm": 2.7974212169647217, |
| "learning_rate": 8.03088803088803e-06, |
| "loss": 753.3157, |
| "step": 102 |
| }, |
| { |
| "epoch": 0.19884169884169883, |
| "grad_norm": 3.194911003112793, |
| "learning_rate": 8.011583011583011e-06, |
| "loss": 773.3227, |
| "step": 103 |
| }, |
| { |
| "epoch": 0.20077220077220076, |
| "grad_norm": 2.7325661182403564, |
| "learning_rate": 7.992277992277993e-06, |
| "loss": 746.1143, |
| "step": 104 |
| }, |
| { |
| "epoch": 0.20270270270270271, |
| "grad_norm": 2.740203619003296, |
| "learning_rate": 7.972972972972974e-06, |
| "loss": 750.4532, |
| "step": 105 |
| }, |
| { |
| "epoch": 0.20463320463320464, |
| "grad_norm": 2.764615297317505, |
| "learning_rate": 7.953667953667954e-06, |
| "loss": 757.8738, |
| "step": 106 |
| }, |
| { |
| "epoch": 0.20656370656370657, |
| "grad_norm": 2.783639430999756, |
| "learning_rate": 7.934362934362934e-06, |
| "loss": 759.7949, |
| "step": 107 |
| }, |
| { |
| "epoch": 0.2084942084942085, |
| "grad_norm": 2.933793783187866, |
| "learning_rate": 7.915057915057915e-06, |
| "loss": 771.2281, |
| "step": 108 |
| }, |
| { |
| "epoch": 0.21042471042471042, |
| "grad_norm": 2.795241117477417, |
| "learning_rate": 7.895752895752897e-06, |
| "loss": 747.9264, |
| "step": 109 |
| }, |
| { |
| "epoch": 0.21235521235521235, |
| "grad_norm": 2.755584239959717, |
| "learning_rate": 7.876447876447877e-06, |
| "loss": 746.8754, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.21428571428571427, |
| "grad_norm": 2.802136182785034, |
| "learning_rate": 7.857142857142858e-06, |
| "loss": 745.1733, |
| "step": 111 |
| }, |
| { |
| "epoch": 0.21621621621621623, |
| "grad_norm": 3.5105533599853516, |
| "learning_rate": 7.837837837837838e-06, |
| "loss": 745.6346, |
| "step": 112 |
| }, |
| { |
| "epoch": 0.21814671814671815, |
| "grad_norm": 2.8462111949920654, |
| "learning_rate": 7.818532818532818e-06, |
| "loss": 757.0023, |
| "step": 113 |
| }, |
| { |
| "epoch": 0.22007722007722008, |
| "grad_norm": 3.079543113708496, |
| "learning_rate": 7.7992277992278e-06, |
| "loss": 757.2842, |
| "step": 114 |
| }, |
| { |
| "epoch": 0.222007722007722, |
| "grad_norm": 2.8306119441986084, |
| "learning_rate": 7.779922779922781e-06, |
| "loss": 752.3796, |
| "step": 115 |
| }, |
| { |
| "epoch": 0.22393822393822393, |
| "grad_norm": 3.1957056522369385, |
| "learning_rate": 7.760617760617761e-06, |
| "loss": 753.2894, |
| "step": 116 |
| }, |
| { |
| "epoch": 0.22586872586872586, |
| "grad_norm": 2.90783953666687, |
| "learning_rate": 7.741312741312742e-06, |
| "loss": 753.8042, |
| "step": 117 |
| }, |
| { |
| "epoch": 0.2277992277992278, |
| "grad_norm": 2.8529882431030273, |
| "learning_rate": 7.722007722007722e-06, |
| "loss": 758.0192, |
| "step": 118 |
| }, |
| { |
| "epoch": 0.22972972972972974, |
| "grad_norm": 2.8292617797851562, |
| "learning_rate": 7.702702702702704e-06, |
| "loss": 756.563, |
| "step": 119 |
| }, |
| { |
| "epoch": 0.23166023166023167, |
| "grad_norm": 2.8797755241394043, |
| "learning_rate": 7.683397683397685e-06, |
| "loss": 761.1346, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.2335907335907336, |
| "grad_norm": 2.7390382289886475, |
| "learning_rate": 7.664092664092665e-06, |
| "loss": 750.3627, |
| "step": 121 |
| }, |
| { |
| "epoch": 0.23552123552123552, |
| "grad_norm": 2.7686545848846436, |
| "learning_rate": 7.644787644787645e-06, |
| "loss": 750.1412, |
| "step": 122 |
| }, |
| { |
| "epoch": 0.23745173745173745, |
| "grad_norm": 2.798008680343628, |
| "learning_rate": 7.625482625482627e-06, |
| "loss": 742.588, |
| "step": 123 |
| }, |
| { |
| "epoch": 0.23938223938223938, |
| "grad_norm": 4.088344097137451, |
| "learning_rate": 7.606177606177607e-06, |
| "loss": 758.9719, |
| "step": 124 |
| }, |
| { |
| "epoch": 0.2413127413127413, |
| "grad_norm": 2.9229252338409424, |
| "learning_rate": 7.5868725868725875e-06, |
| "loss": 756.5762, |
| "step": 125 |
| }, |
| { |
| "epoch": 0.24324324324324326, |
| "grad_norm": 3.1966676712036133, |
| "learning_rate": 7.567567567567569e-06, |
| "loss": 768.2112, |
| "step": 126 |
| }, |
| { |
| "epoch": 0.24517374517374518, |
| "grad_norm": 3.255308151245117, |
| "learning_rate": 7.548262548262549e-06, |
| "loss": 768.8184, |
| "step": 127 |
| }, |
| { |
| "epoch": 0.2471042471042471, |
| "grad_norm": 2.803837537765503, |
| "learning_rate": 7.5289575289575304e-06, |
| "loss": 758.1424, |
| "step": 128 |
| }, |
| { |
| "epoch": 0.24903474903474904, |
| "grad_norm": 2.8075811862945557, |
| "learning_rate": 7.509652509652511e-06, |
| "loss": 748.3246, |
| "step": 129 |
| }, |
| { |
| "epoch": 0.25096525096525096, |
| "grad_norm": 3.1468892097473145, |
| "learning_rate": 7.49034749034749e-06, |
| "loss": 759.9876, |
| "step": 130 |
| }, |
| { |
| "epoch": 0.2528957528957529, |
| "grad_norm": 3.324525833129883, |
| "learning_rate": 7.471042471042471e-06, |
| "loss": 756.2051, |
| "step": 131 |
| }, |
| { |
| "epoch": 0.2548262548262548, |
| "grad_norm": 2.8439035415649414, |
| "learning_rate": 7.451737451737452e-06, |
| "loss": 755.392, |
| "step": 132 |
| }, |
| { |
| "epoch": 0.25675675675675674, |
| "grad_norm": 2.882779598236084, |
| "learning_rate": 7.4324324324324324e-06, |
| "loss": 771.8354, |
| "step": 133 |
| }, |
| { |
| "epoch": 0.25868725868725867, |
| "grad_norm": 3.0841407775878906, |
| "learning_rate": 7.413127413127414e-06, |
| "loss": 754.2871, |
| "step": 134 |
| }, |
| { |
| "epoch": 0.2606177606177606, |
| "grad_norm": 3.2702057361602783, |
| "learning_rate": 7.393822393822394e-06, |
| "loss": 742.1606, |
| "step": 135 |
| }, |
| { |
| "epoch": 0.2625482625482625, |
| "grad_norm": 2.7956244945526123, |
| "learning_rate": 7.3745173745173745e-06, |
| "loss": 750.0869, |
| "step": 136 |
| }, |
| { |
| "epoch": 0.2644787644787645, |
| "grad_norm": 3.004718542098999, |
| "learning_rate": 7.355212355212356e-06, |
| "loss": 781.2869, |
| "step": 137 |
| }, |
| { |
| "epoch": 0.26640926640926643, |
| "grad_norm": 3.8438968658447266, |
| "learning_rate": 7.335907335907336e-06, |
| "loss": 755.309, |
| "step": 138 |
| }, |
| { |
| "epoch": 0.26833976833976836, |
| "grad_norm": 3.0307114124298096, |
| "learning_rate": 7.316602316602317e-06, |
| "loss": 757.5778, |
| "step": 139 |
| }, |
| { |
| "epoch": 0.2702702702702703, |
| "grad_norm": 2.9648211002349854, |
| "learning_rate": 7.297297297297298e-06, |
| "loss": 753.4247, |
| "step": 140 |
| }, |
| { |
| "epoch": 0.2722007722007722, |
| "grad_norm": 2.9691784381866455, |
| "learning_rate": 7.277992277992278e-06, |
| "loss": 757.2734, |
| "step": 141 |
| }, |
| { |
| "epoch": 0.27413127413127414, |
| "grad_norm": 2.8389272689819336, |
| "learning_rate": 7.2586872586872595e-06, |
| "loss": 758.1973, |
| "step": 142 |
| }, |
| { |
| "epoch": 0.27606177606177607, |
| "grad_norm": 2.8457741737365723, |
| "learning_rate": 7.23938223938224e-06, |
| "loss": 759.3056, |
| "step": 143 |
| }, |
| { |
| "epoch": 0.277992277992278, |
| "grad_norm": 2.9947755336761475, |
| "learning_rate": 7.22007722007722e-06, |
| "loss": 756.6718, |
| "step": 144 |
| }, |
| { |
| "epoch": 0.2799227799227799, |
| "grad_norm": 3.285741090774536, |
| "learning_rate": 7.2007722007722015e-06, |
| "loss": 741.2257, |
| "step": 145 |
| }, |
| { |
| "epoch": 0.28185328185328185, |
| "grad_norm": 3.118593692779541, |
| "learning_rate": 7.181467181467182e-06, |
| "loss": 757.1205, |
| "step": 146 |
| }, |
| { |
| "epoch": 0.28378378378378377, |
| "grad_norm": 3.1158804893493652, |
| "learning_rate": 7.162162162162163e-06, |
| "loss": 747.1119, |
| "step": 147 |
| }, |
| { |
| "epoch": 0.2857142857142857, |
| "grad_norm": 2.9939019680023193, |
| "learning_rate": 7.1428571428571436e-06, |
| "loss": 738.4834, |
| "step": 148 |
| }, |
| { |
| "epoch": 0.2876447876447876, |
| "grad_norm": 3.792679786682129, |
| "learning_rate": 7.123552123552124e-06, |
| "loss": 758.1683, |
| "step": 149 |
| }, |
| { |
| "epoch": 0.28957528957528955, |
| "grad_norm": 3.3988595008850098, |
| "learning_rate": 7.104247104247105e-06, |
| "loss": 757.7793, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.2915057915057915, |
| "grad_norm": 3.4164488315582275, |
| "learning_rate": 7.084942084942086e-06, |
| "loss": 753.686, |
| "step": 151 |
| }, |
| { |
| "epoch": 0.29343629343629346, |
| "grad_norm": 4.295523166656494, |
| "learning_rate": 7.065637065637067e-06, |
| "loss": 749.4221, |
| "step": 152 |
| }, |
| { |
| "epoch": 0.2953667953667954, |
| "grad_norm": 3.1985435485839844, |
| "learning_rate": 7.046332046332047e-06, |
| "loss": 766.7484, |
| "step": 153 |
| }, |
| { |
| "epoch": 0.2972972972972973, |
| "grad_norm": 2.811443567276001, |
| "learning_rate": 7.027027027027028e-06, |
| "loss": 744.4705, |
| "step": 154 |
| }, |
| { |
| "epoch": 0.29922779922779924, |
| "grad_norm": 2.8985869884490967, |
| "learning_rate": 7.007722007722009e-06, |
| "loss": 771.2823, |
| "step": 155 |
| }, |
| { |
| "epoch": 0.30115830115830117, |
| "grad_norm": 2.913670301437378, |
| "learning_rate": 6.988416988416989e-06, |
| "loss": 767.1945, |
| "step": 156 |
| }, |
| { |
| "epoch": 0.3030888030888031, |
| "grad_norm": 3.091149091720581, |
| "learning_rate": 6.9691119691119706e-06, |
| "loss": 770.168, |
| "step": 157 |
| }, |
| { |
| "epoch": 0.305019305019305, |
| "grad_norm": 2.89846134185791, |
| "learning_rate": 6.949806949806951e-06, |
| "loss": 746.1279, |
| "step": 158 |
| }, |
| { |
| "epoch": 0.30694980694980695, |
| "grad_norm": 2.981823444366455, |
| "learning_rate": 6.9305019305019305e-06, |
| "loss": 749.2059, |
| "step": 159 |
| }, |
| { |
| "epoch": 0.3088803088803089, |
| "grad_norm": 3.6476705074310303, |
| "learning_rate": 6.911196911196911e-06, |
| "loss": 757.4434, |
| "step": 160 |
| }, |
| { |
| "epoch": 0.3108108108108108, |
| "grad_norm": 2.946174383163452, |
| "learning_rate": 6.891891891891892e-06, |
| "loss": 759.8816, |
| "step": 161 |
| }, |
| { |
| "epoch": 0.3127413127413127, |
| "grad_norm": 3.030103921890259, |
| "learning_rate": 6.872586872586873e-06, |
| "loss": 757.3468, |
| "step": 162 |
| }, |
| { |
| "epoch": 0.31467181467181465, |
| "grad_norm": 3.3172245025634766, |
| "learning_rate": 6.853281853281854e-06, |
| "loss": 753.3054, |
| "step": 163 |
| }, |
| { |
| "epoch": 0.3166023166023166, |
| "grad_norm": 3.128265857696533, |
| "learning_rate": 6.833976833976834e-06, |
| "loss": 764.0436, |
| "step": 164 |
| }, |
| { |
| "epoch": 0.3185328185328185, |
| "grad_norm": 2.8332438468933105, |
| "learning_rate": 6.814671814671815e-06, |
| "loss": 747.4676, |
| "step": 165 |
| }, |
| { |
| "epoch": 0.3204633204633205, |
| "grad_norm": 2.9192118644714355, |
| "learning_rate": 6.795366795366796e-06, |
| "loss": 759.479, |
| "step": 166 |
| }, |
| { |
| "epoch": 0.3223938223938224, |
| "grad_norm": 2.926225423812866, |
| "learning_rate": 6.776061776061776e-06, |
| "loss": 749.5502, |
| "step": 167 |
| }, |
| { |
| "epoch": 0.32432432432432434, |
| "grad_norm": 3.739107608795166, |
| "learning_rate": 6.7567567567567575e-06, |
| "loss": 761.9246, |
| "step": 168 |
| }, |
| { |
| "epoch": 0.32625482625482627, |
| "grad_norm": 3.1689376831054688, |
| "learning_rate": 6.737451737451738e-06, |
| "loss": 759.9313, |
| "step": 169 |
| }, |
| { |
| "epoch": 0.3281853281853282, |
| "grad_norm": 4.102996349334717, |
| "learning_rate": 6.718146718146718e-06, |
| "loss": 755.9019, |
| "step": 170 |
| }, |
| { |
| "epoch": 0.3301158301158301, |
| "grad_norm": 4.32927131652832, |
| "learning_rate": 6.6988416988417e-06, |
| "loss": 748.2408, |
| "step": 171 |
| }, |
| { |
| "epoch": 0.33204633204633205, |
| "grad_norm": 2.9341273307800293, |
| "learning_rate": 6.67953667953668e-06, |
| "loss": 767.1997, |
| "step": 172 |
| }, |
| { |
| "epoch": 0.333976833976834, |
| "grad_norm": 2.9747681617736816, |
| "learning_rate": 6.66023166023166e-06, |
| "loss": 749.2424, |
| "step": 173 |
| }, |
| { |
| "epoch": 0.3359073359073359, |
| "grad_norm": 3.03954815864563, |
| "learning_rate": 6.640926640926642e-06, |
| "loss": 753.1185, |
| "step": 174 |
| }, |
| { |
| "epoch": 0.33783783783783783, |
| "grad_norm": 2.896865129470825, |
| "learning_rate": 6.621621621621622e-06, |
| "loss": 760.0892, |
| "step": 175 |
| }, |
| { |
| "epoch": 0.33976833976833976, |
| "grad_norm": 4.517345905303955, |
| "learning_rate": 6.602316602316603e-06, |
| "loss": 761.2517, |
| "step": 176 |
| }, |
| { |
| "epoch": 0.3416988416988417, |
| "grad_norm": 3.4095852375030518, |
| "learning_rate": 6.583011583011584e-06, |
| "loss": 760.7, |
| "step": 177 |
| }, |
| { |
| "epoch": 0.3436293436293436, |
| "grad_norm": 3.5481455326080322, |
| "learning_rate": 6.563706563706564e-06, |
| "loss": 761.6207, |
| "step": 178 |
| }, |
| { |
| "epoch": 0.34555984555984554, |
| "grad_norm": 4.07427978515625, |
| "learning_rate": 6.544401544401545e-06, |
| "loss": 762.6121, |
| "step": 179 |
| }, |
| { |
| "epoch": 0.3474903474903475, |
| "grad_norm": 3.08604097366333, |
| "learning_rate": 6.525096525096526e-06, |
| "loss": 743.1675, |
| "step": 180 |
| }, |
| { |
| "epoch": 0.34942084942084944, |
| "grad_norm": 4.791824817657471, |
| "learning_rate": 6.505791505791507e-06, |
| "loss": 773.7855, |
| "step": 181 |
| }, |
| { |
| "epoch": 0.35135135135135137, |
| "grad_norm": 5.513062953948975, |
| "learning_rate": 6.486486486486487e-06, |
| "loss": 755.7749, |
| "step": 182 |
| }, |
| { |
| "epoch": 0.3532818532818533, |
| "grad_norm": 4.703281879425049, |
| "learning_rate": 6.467181467181468e-06, |
| "loss": 762.9774, |
| "step": 183 |
| }, |
| { |
| "epoch": 0.3552123552123552, |
| "grad_norm": 4.766940593719482, |
| "learning_rate": 6.447876447876449e-06, |
| "loss": 757.7225, |
| "step": 184 |
| }, |
| { |
| "epoch": 0.35714285714285715, |
| "grad_norm": 2.995068073272705, |
| "learning_rate": 6.4285714285714295e-06, |
| "loss": 757.5352, |
| "step": 185 |
| }, |
| { |
| "epoch": 0.3590733590733591, |
| "grad_norm": 4.172158241271973, |
| "learning_rate": 6.409266409266411e-06, |
| "loss": 758.1395, |
| "step": 186 |
| }, |
| { |
| "epoch": 0.361003861003861, |
| "grad_norm": 3.057084798812866, |
| "learning_rate": 6.389961389961391e-06, |
| "loss": 763.8903, |
| "step": 187 |
| }, |
| { |
| "epoch": 0.36293436293436293, |
| "grad_norm": 3.4208884239196777, |
| "learning_rate": 6.370656370656371e-06, |
| "loss": 755.9663, |
| "step": 188 |
| }, |
| { |
| "epoch": 0.36486486486486486, |
| "grad_norm": 3.5881717205047607, |
| "learning_rate": 6.351351351351351e-06, |
| "loss": 764.6107, |
| "step": 189 |
| }, |
| { |
| "epoch": 0.3667953667953668, |
| "grad_norm": 4.713850498199463, |
| "learning_rate": 6.332046332046332e-06, |
| "loss": 748.8102, |
| "step": 190 |
| }, |
| { |
| "epoch": 0.3687258687258687, |
| "grad_norm": 3.244720935821533, |
| "learning_rate": 6.312741312741313e-06, |
| "loss": 764.7598, |
| "step": 191 |
| }, |
| { |
| "epoch": 0.37065637065637064, |
| "grad_norm": 4.447683334350586, |
| "learning_rate": 6.293436293436294e-06, |
| "loss": 752.6754, |
| "step": 192 |
| }, |
| { |
| "epoch": 0.37258687258687256, |
| "grad_norm": 3.9772603511810303, |
| "learning_rate": 6.274131274131274e-06, |
| "loss": 756.0471, |
| "step": 193 |
| }, |
| { |
| "epoch": 0.3745173745173745, |
| "grad_norm": 4.803740978240967, |
| "learning_rate": 6.254826254826255e-06, |
| "loss": 769.636, |
| "step": 194 |
| }, |
| { |
| "epoch": 0.3764478764478765, |
| "grad_norm": 4.907277584075928, |
| "learning_rate": 6.235521235521236e-06, |
| "loss": 765.9659, |
| "step": 195 |
| }, |
| { |
| "epoch": 0.3783783783783784, |
| "grad_norm": 3.6414902210235596, |
| "learning_rate": 6.2162162162162164e-06, |
| "loss": 760.1011, |
| "step": 196 |
| }, |
| { |
| "epoch": 0.3803088803088803, |
| "grad_norm": 3.6037609577178955, |
| "learning_rate": 6.196911196911197e-06, |
| "loss": 750.7067, |
| "step": 197 |
| }, |
| { |
| "epoch": 0.38223938223938225, |
| "grad_norm": 2.8772242069244385, |
| "learning_rate": 6.177606177606178e-06, |
| "loss": 756.6414, |
| "step": 198 |
| }, |
| { |
| "epoch": 0.3841698841698842, |
| "grad_norm": 5.2583184242248535, |
| "learning_rate": 6.1583011583011585e-06, |
| "loss": 751.6912, |
| "step": 199 |
| }, |
| { |
| "epoch": 0.3861003861003861, |
| "grad_norm": 2.8050296306610107, |
| "learning_rate": 6.13899613899614e-06, |
| "loss": 743.5164, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.38803088803088803, |
| "grad_norm": 3.5677103996276855, |
| "learning_rate": 6.11969111969112e-06, |
| "loss": 764.8164, |
| "step": 201 |
| }, |
| { |
| "epoch": 0.38996138996138996, |
| "grad_norm": 3.1550533771514893, |
| "learning_rate": 6.1003861003861005e-06, |
| "loss": 762.9319, |
| "step": 202 |
| }, |
| { |
| "epoch": 0.3918918918918919, |
| "grad_norm": 2.991600751876831, |
| "learning_rate": 6.081081081081082e-06, |
| "loss": 770.1368, |
| "step": 203 |
| }, |
| { |
| "epoch": 0.3938223938223938, |
| "grad_norm": 2.9870667457580566, |
| "learning_rate": 6.061776061776062e-06, |
| "loss": 751.1348, |
| "step": 204 |
| }, |
| { |
| "epoch": 0.39575289575289574, |
| "grad_norm": 2.980593204498291, |
| "learning_rate": 6.0424710424710434e-06, |
| "loss": 769.3608, |
| "step": 205 |
| }, |
| { |
| "epoch": 0.39768339768339767, |
| "grad_norm": 3.3218741416931152, |
| "learning_rate": 6.023166023166024e-06, |
| "loss": 763.5135, |
| "step": 206 |
| }, |
| { |
| "epoch": 0.3996138996138996, |
| "grad_norm": 3.1396753787994385, |
| "learning_rate": 6.003861003861004e-06, |
| "loss": 757.1105, |
| "step": 207 |
| }, |
| { |
| "epoch": 0.4015444015444015, |
| "grad_norm": 3.2232918739318848, |
| "learning_rate": 5.9845559845559855e-06, |
| "loss": 743.5112, |
| "step": 208 |
| }, |
| { |
| "epoch": 0.4034749034749035, |
| "grad_norm": 3.3327605724334717, |
| "learning_rate": 5.965250965250966e-06, |
| "loss": 766.4379, |
| "step": 209 |
| }, |
| { |
| "epoch": 0.40540540540540543, |
| "grad_norm": 3.4697072505950928, |
| "learning_rate": 5.945945945945947e-06, |
| "loss": 767.7292, |
| "step": 210 |
| }, |
| { |
| "epoch": 0.40733590733590735, |
| "grad_norm": 4.161364555358887, |
| "learning_rate": 5.9266409266409275e-06, |
| "loss": 746.888, |
| "step": 211 |
| }, |
| { |
| "epoch": 0.4092664092664093, |
| "grad_norm": 3.0299108028411865, |
| "learning_rate": 5.907335907335908e-06, |
| "loss": 773.0205, |
| "step": 212 |
| }, |
| { |
| "epoch": 0.4111969111969112, |
| "grad_norm": 3.702420949935913, |
| "learning_rate": 5.888030888030889e-06, |
| "loss": 762.8275, |
| "step": 213 |
| }, |
| { |
| "epoch": 0.41312741312741313, |
| "grad_norm": 4.044224262237549, |
| "learning_rate": 5.86872586872587e-06, |
| "loss": 762.0264, |
| "step": 214 |
| }, |
| { |
| "epoch": 0.41505791505791506, |
| "grad_norm": 3.09975266456604, |
| "learning_rate": 5.84942084942085e-06, |
| "loss": 738.4187, |
| "step": 215 |
| }, |
| { |
| "epoch": 0.416988416988417, |
| "grad_norm": 3.779367208480835, |
| "learning_rate": 5.83011583011583e-06, |
| "loss": 764.1588, |
| "step": 216 |
| }, |
| { |
| "epoch": 0.4189189189189189, |
| "grad_norm": 3.9344935417175293, |
| "learning_rate": 5.810810810810811e-06, |
| "loss": 751.987, |
| "step": 217 |
| }, |
| { |
| "epoch": 0.42084942084942084, |
| "grad_norm": 3.073119878768921, |
| "learning_rate": 5.791505791505791e-06, |
| "loss": 763.5579, |
| "step": 218 |
| }, |
| { |
| "epoch": 0.42277992277992277, |
| "grad_norm": 5.221838474273682, |
| "learning_rate": 5.7722007722007725e-06, |
| "loss": 736.679, |
| "step": 219 |
| }, |
| { |
| "epoch": 0.4247104247104247, |
| "grad_norm": 3.210583209991455, |
| "learning_rate": 5.752895752895753e-06, |
| "loss": 743.9155, |
| "step": 220 |
| }, |
| { |
| "epoch": 0.4266409266409266, |
| "grad_norm": 3.032778739929199, |
| "learning_rate": 5.733590733590734e-06, |
| "loss": 756.9208, |
| "step": 221 |
| }, |
| { |
| "epoch": 0.42857142857142855, |
| "grad_norm": 3.052091598510742, |
| "learning_rate": 5.7142857142857145e-06, |
| "loss": 753.4189, |
| "step": 222 |
| }, |
| { |
| "epoch": 0.4305019305019305, |
| "grad_norm": 2.934615135192871, |
| "learning_rate": 5.694980694980695e-06, |
| "loss": 744.2302, |
| "step": 223 |
| }, |
| { |
| "epoch": 0.43243243243243246, |
| "grad_norm": 3.453640937805176, |
| "learning_rate": 5.675675675675676e-06, |
| "loss": 753.5289, |
| "step": 224 |
| }, |
| { |
| "epoch": 0.4343629343629344, |
| "grad_norm": 4.493204116821289, |
| "learning_rate": 5.6563706563706566e-06, |
| "loss": 773.2119, |
| "step": 225 |
| }, |
| { |
| "epoch": 0.4362934362934363, |
| "grad_norm": 3.1990344524383545, |
| "learning_rate": 5.637065637065637e-06, |
| "loss": 771.2296, |
| "step": 226 |
| }, |
| { |
| "epoch": 0.43822393822393824, |
| "grad_norm": 4.990328788757324, |
| "learning_rate": 5.617760617760618e-06, |
| "loss": 766.2954, |
| "step": 227 |
| }, |
| { |
| "epoch": 0.44015444015444016, |
| "grad_norm": 6.218008518218994, |
| "learning_rate": 5.598455598455599e-06, |
| "loss": 754.2188, |
| "step": 228 |
| }, |
| { |
| "epoch": 0.4420849420849421, |
| "grad_norm": 4.668118476867676, |
| "learning_rate": 5.57915057915058e-06, |
| "loss": 764.908, |
| "step": 229 |
| }, |
| { |
| "epoch": 0.444015444015444, |
| "grad_norm": 3.144949436187744, |
| "learning_rate": 5.55984555984556e-06, |
| "loss": 769.907, |
| "step": 230 |
| }, |
| { |
| "epoch": 0.44594594594594594, |
| "grad_norm": 3.761617422103882, |
| "learning_rate": 5.540540540540541e-06, |
| "loss": 762.5698, |
| "step": 231 |
| }, |
| { |
| "epoch": 0.44787644787644787, |
| "grad_norm": 3.3352720737457275, |
| "learning_rate": 5.521235521235522e-06, |
| "loss": 760.8379, |
| "step": 232 |
| }, |
| { |
| "epoch": 0.4498069498069498, |
| "grad_norm": 3.646754741668701, |
| "learning_rate": 5.501930501930502e-06, |
| "loss": 762.5335, |
| "step": 233 |
| }, |
| { |
| "epoch": 0.4517374517374517, |
| "grad_norm": 5.615658283233643, |
| "learning_rate": 5.4826254826254836e-06, |
| "loss": 763.7914, |
| "step": 234 |
| }, |
| { |
| "epoch": 0.45366795366795365, |
| "grad_norm": 3.1293540000915527, |
| "learning_rate": 5.463320463320464e-06, |
| "loss": 765.2655, |
| "step": 235 |
| }, |
| { |
| "epoch": 0.4555984555984556, |
| "grad_norm": 3.9679007530212402, |
| "learning_rate": 5.444015444015444e-06, |
| "loss": 776.4496, |
| "step": 236 |
| }, |
| { |
| "epoch": 0.4575289575289575, |
| "grad_norm": 4.110316753387451, |
| "learning_rate": 5.424710424710426e-06, |
| "loss": 769.2201, |
| "step": 237 |
| }, |
| { |
| "epoch": 0.4594594594594595, |
| "grad_norm": 3.5175106525421143, |
| "learning_rate": 5.405405405405406e-06, |
| "loss": 759.53, |
| "step": 238 |
| }, |
| { |
| "epoch": 0.4613899613899614, |
| "grad_norm": 3.206360101699829, |
| "learning_rate": 5.386100386100387e-06, |
| "loss": 753.7003, |
| "step": 239 |
| }, |
| { |
| "epoch": 0.46332046332046334, |
| "grad_norm": 2.994739294052124, |
| "learning_rate": 5.366795366795368e-06, |
| "loss": 750.8613, |
| "step": 240 |
| }, |
| { |
| "epoch": 0.46525096525096526, |
| "grad_norm": 2.966420888900757, |
| "learning_rate": 5.347490347490348e-06, |
| "loss": 755.6821, |
| "step": 241 |
| }, |
| { |
| "epoch": 0.4671814671814672, |
| "grad_norm": 5.035081386566162, |
| "learning_rate": 5.328185328185329e-06, |
| "loss": 753.4547, |
| "step": 242 |
| }, |
| { |
| "epoch": 0.4691119691119691, |
| "grad_norm": 3.399747371673584, |
| "learning_rate": 5.30888030888031e-06, |
| "loss": 752.4637, |
| "step": 243 |
| }, |
| { |
| "epoch": 0.47104247104247104, |
| "grad_norm": 2.9818832874298096, |
| "learning_rate": 5.28957528957529e-06, |
| "loss": 761.8242, |
| "step": 244 |
| }, |
| { |
| "epoch": 0.47297297297297297, |
| "grad_norm": 3.317462921142578, |
| "learning_rate": 5.2702702702702705e-06, |
| "loss": 741.8611, |
| "step": 245 |
| }, |
| { |
| "epoch": 0.4749034749034749, |
| "grad_norm": 3.9936470985412598, |
| "learning_rate": 5.250965250965251e-06, |
| "loss": 755.0592, |
| "step": 246 |
| }, |
| { |
| "epoch": 0.4768339768339768, |
| "grad_norm": 3.263794422149658, |
| "learning_rate": 5.231660231660231e-06, |
| "loss": 748.0864, |
| "step": 247 |
| }, |
| { |
| "epoch": 0.47876447876447875, |
| "grad_norm": 3.1874217987060547, |
| "learning_rate": 5.212355212355213e-06, |
| "loss": 751.8619, |
| "step": 248 |
| }, |
| { |
| "epoch": 0.4806949806949807, |
| "grad_norm": 4.754833698272705, |
| "learning_rate": 5.193050193050193e-06, |
| "loss": 763.7399, |
| "step": 249 |
| }, |
| { |
| "epoch": 0.4826254826254826, |
| "grad_norm": 3.21230149269104, |
| "learning_rate": 5.173745173745173e-06, |
| "loss": 757.1412, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.48455598455598453, |
| "grad_norm": 3.2287521362304688, |
| "learning_rate": 5.154440154440155e-06, |
| "loss": 749.6309, |
| "step": 251 |
| }, |
| { |
| "epoch": 0.4864864864864865, |
| "grad_norm": 3.137535572052002, |
| "learning_rate": 5.135135135135135e-06, |
| "loss": 759.4156, |
| "step": 252 |
| }, |
| { |
| "epoch": 0.48841698841698844, |
| "grad_norm": 3.6146438121795654, |
| "learning_rate": 5.115830115830116e-06, |
| "loss": 741.8262, |
| "step": 253 |
| }, |
| { |
| "epoch": 0.49034749034749037, |
| "grad_norm": 2.9531517028808594, |
| "learning_rate": 5.096525096525097e-06, |
| "loss": 763.7745, |
| "step": 254 |
| }, |
| { |
| "epoch": 0.4922779922779923, |
| "grad_norm": 3.0513949394226074, |
| "learning_rate": 5.077220077220077e-06, |
| "loss": 755.271, |
| "step": 255 |
| }, |
| { |
| "epoch": 0.4942084942084942, |
| "grad_norm": 3.2384207248687744, |
| "learning_rate": 5.057915057915058e-06, |
| "loss": 765.3377, |
| "step": 256 |
| }, |
| { |
| "epoch": 0.49613899613899615, |
| "grad_norm": 3.8821310997009277, |
| "learning_rate": 5.038610038610039e-06, |
| "loss": 763.2088, |
| "step": 257 |
| }, |
| { |
| "epoch": 0.4980694980694981, |
| "grad_norm": 2.926225423812866, |
| "learning_rate": 5.01930501930502e-06, |
| "loss": 758.5596, |
| "step": 258 |
| }, |
| { |
| "epoch": 0.5, |
| "grad_norm": 2.9751245975494385, |
| "learning_rate": 5e-06, |
| "loss": 757.9719, |
| "step": 259 |
| }, |
| { |
| "epoch": 0.5019305019305019, |
| "grad_norm": 4.772404193878174, |
| "learning_rate": 4.980694980694981e-06, |
| "loss": 757.3872, |
| "step": 260 |
| }, |
| { |
| "epoch": 0.5038610038610039, |
| "grad_norm": 3.096656322479248, |
| "learning_rate": 4.961389961389962e-06, |
| "loss": 760.4723, |
| "step": 261 |
| }, |
| { |
| "epoch": 0.5057915057915058, |
| "grad_norm": 3.504852056503296, |
| "learning_rate": 4.9420849420849425e-06, |
| "loss": 747.486, |
| "step": 262 |
| }, |
| { |
| "epoch": 0.5077220077220077, |
| "grad_norm": 3.33747935295105, |
| "learning_rate": 4.922779922779924e-06, |
| "loss": 758.9821, |
| "step": 263 |
| }, |
| { |
| "epoch": 0.5096525096525096, |
| "grad_norm": 3.4812850952148438, |
| "learning_rate": 4.903474903474904e-06, |
| "loss": 765.7698, |
| "step": 264 |
| }, |
| { |
| "epoch": 0.5115830115830116, |
| "grad_norm": 3.174971580505371, |
| "learning_rate": 4.8841698841698845e-06, |
| "loss": 756.7213, |
| "step": 265 |
| }, |
| { |
| "epoch": 0.5135135135135135, |
| "grad_norm": 3.601083755493164, |
| "learning_rate": 4.864864864864866e-06, |
| "loss": 762.7404, |
| "step": 266 |
| }, |
| { |
| "epoch": 0.5154440154440154, |
| "grad_norm": 3.3433072566986084, |
| "learning_rate": 4.845559845559845e-06, |
| "loss": 747.8812, |
| "step": 267 |
| }, |
| { |
| "epoch": 0.5173745173745173, |
| "grad_norm": 3.7494962215423584, |
| "learning_rate": 4.8262548262548266e-06, |
| "loss": 756.0776, |
| "step": 268 |
| }, |
| { |
| "epoch": 0.5193050193050193, |
| "grad_norm": 4.609169960021973, |
| "learning_rate": 4.806949806949807e-06, |
| "loss": 750.7801, |
| "step": 269 |
| }, |
| { |
| "epoch": 0.5212355212355212, |
| "grad_norm": 5.737427711486816, |
| "learning_rate": 4.787644787644788e-06, |
| "loss": 751.1343, |
| "step": 270 |
| }, |
| { |
| "epoch": 0.5231660231660231, |
| "grad_norm": 3.3391857147216797, |
| "learning_rate": 4.768339768339769e-06, |
| "loss": 760.2618, |
| "step": 271 |
| }, |
| { |
| "epoch": 0.525096525096525, |
| "grad_norm": 3.1758346557617188, |
| "learning_rate": 4.749034749034749e-06, |
| "loss": 756.4235, |
| "step": 272 |
| }, |
| { |
| "epoch": 0.527027027027027, |
| "grad_norm": 4.515329360961914, |
| "learning_rate": 4.72972972972973e-06, |
| "loss": 770.3212, |
| "step": 273 |
| }, |
| { |
| "epoch": 0.528957528957529, |
| "grad_norm": 3.2932803630828857, |
| "learning_rate": 4.710424710424711e-06, |
| "loss": 751.3934, |
| "step": 274 |
| }, |
| { |
| "epoch": 0.5308880308880309, |
| "grad_norm": 3.047992467880249, |
| "learning_rate": 4.691119691119692e-06, |
| "loss": 749.6434, |
| "step": 275 |
| }, |
| { |
| "epoch": 0.5328185328185329, |
| "grad_norm": 3.2720017433166504, |
| "learning_rate": 4.671814671814672e-06, |
| "loss": 771.3862, |
| "step": 276 |
| }, |
| { |
| "epoch": 0.5347490347490348, |
| "grad_norm": 3.6635942459106445, |
| "learning_rate": 4.652509652509653e-06, |
| "loss": 758.3706, |
| "step": 277 |
| }, |
| { |
| "epoch": 0.5366795366795367, |
| "grad_norm": 3.4579808712005615, |
| "learning_rate": 4.633204633204634e-06, |
| "loss": 747.9475, |
| "step": 278 |
| }, |
| { |
| "epoch": 0.5386100386100386, |
| "grad_norm": 3.525111675262451, |
| "learning_rate": 4.613899613899614e-06, |
| "loss": 756.5895, |
| "step": 279 |
| }, |
| { |
| "epoch": 0.5405405405405406, |
| "grad_norm": 3.336205005645752, |
| "learning_rate": 4.594594594594596e-06, |
| "loss": 772.9059, |
| "step": 280 |
| }, |
| { |
| "epoch": 0.5424710424710425, |
| "grad_norm": 3.758382558822632, |
| "learning_rate": 4.575289575289575e-06, |
| "loss": 769.8763, |
| "step": 281 |
| }, |
| { |
| "epoch": 0.5444015444015444, |
| "grad_norm": 3.054797649383545, |
| "learning_rate": 4.5559845559845564e-06, |
| "loss": 768.0063, |
| "step": 282 |
| }, |
| { |
| "epoch": 0.5463320463320464, |
| "grad_norm": 4.434202671051025, |
| "learning_rate": 4.536679536679537e-06, |
| "loss": 760.1692, |
| "step": 283 |
| }, |
| { |
| "epoch": 0.5482625482625483, |
| "grad_norm": 4.3827667236328125, |
| "learning_rate": 4.517374517374517e-06, |
| "loss": 744.6565, |
| "step": 284 |
| }, |
| { |
| "epoch": 0.5501930501930502, |
| "grad_norm": 6.237070560455322, |
| "learning_rate": 4.4980694980694985e-06, |
| "loss": 767.5054, |
| "step": 285 |
| }, |
| { |
| "epoch": 0.5521235521235521, |
| "grad_norm": 3.162343740463257, |
| "learning_rate": 4.478764478764479e-06, |
| "loss": 762.6232, |
| "step": 286 |
| }, |
| { |
| "epoch": 0.5540540540540541, |
| "grad_norm": 3.9073193073272705, |
| "learning_rate": 4.45945945945946e-06, |
| "loss": 778.0848, |
| "step": 287 |
| }, |
| { |
| "epoch": 0.555984555984556, |
| "grad_norm": 6.03090763092041, |
| "learning_rate": 4.4401544401544405e-06, |
| "loss": 759.6763, |
| "step": 288 |
| }, |
| { |
| "epoch": 0.5579150579150579, |
| "grad_norm": 4.3667311668396, |
| "learning_rate": 4.420849420849421e-06, |
| "loss": 762.0499, |
| "step": 289 |
| }, |
| { |
| "epoch": 0.5598455598455598, |
| "grad_norm": 3.5320377349853516, |
| "learning_rate": 4.401544401544402e-06, |
| "loss": 769.951, |
| "step": 290 |
| }, |
| { |
| "epoch": 0.5617760617760618, |
| "grad_norm": 3.7424259185791016, |
| "learning_rate": 4.382239382239383e-06, |
| "loss": 767.4916, |
| "step": 291 |
| }, |
| { |
| "epoch": 0.5637065637065637, |
| "grad_norm": 3.3271844387054443, |
| "learning_rate": 4.362934362934364e-06, |
| "loss": 761.4987, |
| "step": 292 |
| }, |
| { |
| "epoch": 0.5656370656370656, |
| "grad_norm": 2.957545757293701, |
| "learning_rate": 4.343629343629344e-06, |
| "loss": 762.5405, |
| "step": 293 |
| }, |
| { |
| "epoch": 0.5675675675675675, |
| "grad_norm": 3.087904453277588, |
| "learning_rate": 4.324324324324325e-06, |
| "loss": 752.0869, |
| "step": 294 |
| }, |
| { |
| "epoch": 0.5694980694980695, |
| "grad_norm": 3.4047775268554688, |
| "learning_rate": 4.305019305019305e-06, |
| "loss": 764.2693, |
| "step": 295 |
| }, |
| { |
| "epoch": 0.5714285714285714, |
| "grad_norm": 3.244701385498047, |
| "learning_rate": 4.2857142857142855e-06, |
| "loss": 754.4578, |
| "step": 296 |
| }, |
| { |
| "epoch": 0.5733590733590733, |
| "grad_norm": 3.3166253566741943, |
| "learning_rate": 4.266409266409267e-06, |
| "loss": 773.7292, |
| "step": 297 |
| }, |
| { |
| "epoch": 0.5752895752895753, |
| "grad_norm": 3.0471668243408203, |
| "learning_rate": 4.247104247104247e-06, |
| "loss": 778.6743, |
| "step": 298 |
| }, |
| { |
| "epoch": 0.5772200772200772, |
| "grad_norm": 3.6165595054626465, |
| "learning_rate": 4.227799227799228e-06, |
| "loss": 755.6964, |
| "step": 299 |
| }, |
| { |
| "epoch": 0.5791505791505791, |
| "grad_norm": 3.426254987716675, |
| "learning_rate": 4.208494208494209e-06, |
| "loss": 768.4984, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.581081081081081, |
| "grad_norm": 4.2183518409729, |
| "learning_rate": 4.189189189189189e-06, |
| "loss": 756.4444, |
| "step": 301 |
| }, |
| { |
| "epoch": 0.583011583011583, |
| "grad_norm": 4.252375602722168, |
| "learning_rate": 4.16988416988417e-06, |
| "loss": 760.7186, |
| "step": 302 |
| }, |
| { |
| "epoch": 0.584942084942085, |
| "grad_norm": 3.898979425430298, |
| "learning_rate": 4.150579150579151e-06, |
| "loss": 755.8693, |
| "step": 303 |
| }, |
| { |
| "epoch": 0.5868725868725869, |
| "grad_norm": 4.586599349975586, |
| "learning_rate": 4.131274131274132e-06, |
| "loss": 767.7788, |
| "step": 304 |
| }, |
| { |
| "epoch": 0.5888030888030888, |
| "grad_norm": 3.4545576572418213, |
| "learning_rate": 4.1119691119691125e-06, |
| "loss": 755.6498, |
| "step": 305 |
| }, |
| { |
| "epoch": 0.5907335907335908, |
| "grad_norm": 3.056624174118042, |
| "learning_rate": 4.092664092664093e-06, |
| "loss": 779.4406, |
| "step": 306 |
| }, |
| { |
| "epoch": 0.5926640926640927, |
| "grad_norm": 3.1570804119110107, |
| "learning_rate": 4.073359073359074e-06, |
| "loss": 756.1407, |
| "step": 307 |
| }, |
| { |
| "epoch": 0.5945945945945946, |
| "grad_norm": 5.404672622680664, |
| "learning_rate": 4.0540540540540545e-06, |
| "loss": 770.681, |
| "step": 308 |
| }, |
| { |
| "epoch": 0.5965250965250966, |
| "grad_norm": 3.0374257564544678, |
| "learning_rate": 4.034749034749035e-06, |
| "loss": 759.7466, |
| "step": 309 |
| }, |
| { |
| "epoch": 0.5984555984555985, |
| "grad_norm": 3.2685699462890625, |
| "learning_rate": 4.015444015444015e-06, |
| "loss": 762.7603, |
| "step": 310 |
| }, |
| { |
| "epoch": 0.6003861003861004, |
| "grad_norm": 7.29334831237793, |
| "learning_rate": 3.996138996138997e-06, |
| "loss": 742.7489, |
| "step": 311 |
| }, |
| { |
| "epoch": 0.6023166023166023, |
| "grad_norm": 3.852389335632324, |
| "learning_rate": 3.976833976833977e-06, |
| "loss": 767.3596, |
| "step": 312 |
| }, |
| { |
| "epoch": 0.6042471042471043, |
| "grad_norm": 3.200193166732788, |
| "learning_rate": 3.957528957528957e-06, |
| "loss": 766.8765, |
| "step": 313 |
| }, |
| { |
| "epoch": 0.6061776061776062, |
| "grad_norm": 3.5071451663970947, |
| "learning_rate": 3.938223938223939e-06, |
| "loss": 776.7089, |
| "step": 314 |
| }, |
| { |
| "epoch": 0.6081081081081081, |
| "grad_norm": 4.16587495803833, |
| "learning_rate": 3.918918918918919e-06, |
| "loss": 741.2016, |
| "step": 315 |
| }, |
| { |
| "epoch": 0.61003861003861, |
| "grad_norm": 5.703463554382324, |
| "learning_rate": 3.8996138996139e-06, |
| "loss": 756.0547, |
| "step": 316 |
| }, |
| { |
| "epoch": 0.611969111969112, |
| "grad_norm": 5.430912017822266, |
| "learning_rate": 3.880308880308881e-06, |
| "loss": 756.722, |
| "step": 317 |
| }, |
| { |
| "epoch": 0.6138996138996139, |
| "grad_norm": 8.650154113769531, |
| "learning_rate": 3.861003861003861e-06, |
| "loss": 745.7667, |
| "step": 318 |
| }, |
| { |
| "epoch": 0.6158301158301158, |
| "grad_norm": 13.80824089050293, |
| "learning_rate": 3.841698841698842e-06, |
| "loss": 755.2433, |
| "step": 319 |
| }, |
| { |
| "epoch": 0.6177606177606177, |
| "grad_norm": 7.08932638168335, |
| "learning_rate": 3.822393822393823e-06, |
| "loss": 753.7036, |
| "step": 320 |
| }, |
| { |
| "epoch": 0.6196911196911197, |
| "grad_norm": 3.6433534622192383, |
| "learning_rate": 3.8030888030888036e-06, |
| "loss": 773.3873, |
| "step": 321 |
| }, |
| { |
| "epoch": 0.6216216216216216, |
| "grad_norm": 3.2302098274230957, |
| "learning_rate": 3.7837837837837844e-06, |
| "loss": 758.14, |
| "step": 322 |
| }, |
| { |
| "epoch": 0.6235521235521235, |
| "grad_norm": 3.731865644454956, |
| "learning_rate": 3.7644787644787652e-06, |
| "loss": 766.5978, |
| "step": 323 |
| }, |
| { |
| "epoch": 0.6254826254826255, |
| "grad_norm": 3.244114875793457, |
| "learning_rate": 3.745173745173745e-06, |
| "loss": 748.5593, |
| "step": 324 |
| }, |
| { |
| "epoch": 0.6274131274131274, |
| "grad_norm": 6.71386194229126, |
| "learning_rate": 3.725868725868726e-06, |
| "loss": 750.8797, |
| "step": 325 |
| }, |
| { |
| "epoch": 0.6293436293436293, |
| "grad_norm": 6.857434272766113, |
| "learning_rate": 3.706563706563707e-06, |
| "loss": 751.0732, |
| "step": 326 |
| }, |
| { |
| "epoch": 0.6312741312741312, |
| "grad_norm": 3.1057608127593994, |
| "learning_rate": 3.6872586872586872e-06, |
| "loss": 751.9279, |
| "step": 327 |
| }, |
| { |
| "epoch": 0.6332046332046332, |
| "grad_norm": 4.462062835693359, |
| "learning_rate": 3.667953667953668e-06, |
| "loss": 761.2999, |
| "step": 328 |
| }, |
| { |
| "epoch": 0.6351351351351351, |
| "grad_norm": 5.394342422485352, |
| "learning_rate": 3.648648648648649e-06, |
| "loss": 765.5782, |
| "step": 329 |
| }, |
| { |
| "epoch": 0.637065637065637, |
| "grad_norm": 4.586513996124268, |
| "learning_rate": 3.6293436293436297e-06, |
| "loss": 759.7944, |
| "step": 330 |
| }, |
| { |
| "epoch": 0.638996138996139, |
| "grad_norm": 6.204144477844238, |
| "learning_rate": 3.61003861003861e-06, |
| "loss": 749.0181, |
| "step": 331 |
| }, |
| { |
| "epoch": 0.640926640926641, |
| "grad_norm": 3.065887451171875, |
| "learning_rate": 3.590733590733591e-06, |
| "loss": 764.4531, |
| "step": 332 |
| }, |
| { |
| "epoch": 0.6428571428571429, |
| "grad_norm": 3.5210766792297363, |
| "learning_rate": 3.5714285714285718e-06, |
| "loss": 757.1459, |
| "step": 333 |
| }, |
| { |
| "epoch": 0.6447876447876448, |
| "grad_norm": 4.145437240600586, |
| "learning_rate": 3.5521235521235526e-06, |
| "loss": 772.7172, |
| "step": 334 |
| }, |
| { |
| "epoch": 0.6467181467181468, |
| "grad_norm": 3.4393627643585205, |
| "learning_rate": 3.5328185328185334e-06, |
| "loss": 748.3174, |
| "step": 335 |
| }, |
| { |
| "epoch": 0.6486486486486487, |
| "grad_norm": 6.470264911651611, |
| "learning_rate": 3.513513513513514e-06, |
| "loss": 761.2096, |
| "step": 336 |
| }, |
| { |
| "epoch": 0.6505791505791506, |
| "grad_norm": 3.4703409671783447, |
| "learning_rate": 3.4942084942084947e-06, |
| "loss": 752.6991, |
| "step": 337 |
| }, |
| { |
| "epoch": 0.6525096525096525, |
| "grad_norm": 9.870447158813477, |
| "learning_rate": 3.4749034749034755e-06, |
| "loss": 771.034, |
| "step": 338 |
| }, |
| { |
| "epoch": 0.6544401544401545, |
| "grad_norm": 9.164502143859863, |
| "learning_rate": 3.4555984555984555e-06, |
| "loss": 759.8434, |
| "step": 339 |
| }, |
| { |
| "epoch": 0.6563706563706564, |
| "grad_norm": 8.678630828857422, |
| "learning_rate": 3.4362934362934363e-06, |
| "loss": 760.7397, |
| "step": 340 |
| }, |
| { |
| "epoch": 0.6583011583011583, |
| "grad_norm": 4.10188102722168, |
| "learning_rate": 3.416988416988417e-06, |
| "loss": 752.9614, |
| "step": 341 |
| }, |
| { |
| "epoch": 0.6602316602316602, |
| "grad_norm": 5.000187873840332, |
| "learning_rate": 3.397683397683398e-06, |
| "loss": 741.5472, |
| "step": 342 |
| }, |
| { |
| "epoch": 0.6621621621621622, |
| "grad_norm": 5.491066932678223, |
| "learning_rate": 3.3783783783783788e-06, |
| "loss": 752.3487, |
| "step": 343 |
| }, |
| { |
| "epoch": 0.6640926640926641, |
| "grad_norm": 4.743298530578613, |
| "learning_rate": 3.359073359073359e-06, |
| "loss": 740.1313, |
| "step": 344 |
| }, |
| { |
| "epoch": 0.666023166023166, |
| "grad_norm": 3.8254621028900146, |
| "learning_rate": 3.33976833976834e-06, |
| "loss": 751.6027, |
| "step": 345 |
| }, |
| { |
| "epoch": 0.667953667953668, |
| "grad_norm": 3.4094951152801514, |
| "learning_rate": 3.320463320463321e-06, |
| "loss": 747.6785, |
| "step": 346 |
| }, |
| { |
| "epoch": 0.6698841698841699, |
| "grad_norm": 4.3971757888793945, |
| "learning_rate": 3.3011583011583016e-06, |
| "loss": 765.18, |
| "step": 347 |
| }, |
| { |
| "epoch": 0.6718146718146718, |
| "grad_norm": 5.712260723114014, |
| "learning_rate": 3.281853281853282e-06, |
| "loss": 754.3199, |
| "step": 348 |
| }, |
| { |
| "epoch": 0.6737451737451737, |
| "grad_norm": 2.944014310836792, |
| "learning_rate": 3.262548262548263e-06, |
| "loss": 759.125, |
| "step": 349 |
| }, |
| { |
| "epoch": 0.6756756756756757, |
| "grad_norm": 5.154186725616455, |
| "learning_rate": 3.2432432432432437e-06, |
| "loss": 773.4568, |
| "step": 350 |
| }, |
| { |
| "epoch": 0.6776061776061776, |
| "grad_norm": 4.158810138702393, |
| "learning_rate": 3.2239382239382245e-06, |
| "loss": 773.0182, |
| "step": 351 |
| }, |
| { |
| "epoch": 0.6795366795366795, |
| "grad_norm": 2.9692142009735107, |
| "learning_rate": 3.2046332046332054e-06, |
| "loss": 759.3475, |
| "step": 352 |
| }, |
| { |
| "epoch": 0.6814671814671814, |
| "grad_norm": 5.837861061096191, |
| "learning_rate": 3.1853281853281853e-06, |
| "loss": 758.8844, |
| "step": 353 |
| }, |
| { |
| "epoch": 0.6833976833976834, |
| "grad_norm": 4.23380708694458, |
| "learning_rate": 3.166023166023166e-06, |
| "loss": 762.8468, |
| "step": 354 |
| }, |
| { |
| "epoch": 0.6853281853281853, |
| "grad_norm": 5.026414394378662, |
| "learning_rate": 3.146718146718147e-06, |
| "loss": 738.6548, |
| "step": 355 |
| }, |
| { |
| "epoch": 0.6872586872586872, |
| "grad_norm": 4.041588306427002, |
| "learning_rate": 3.1274131274131274e-06, |
| "loss": 766.1196, |
| "step": 356 |
| }, |
| { |
| "epoch": 0.6891891891891891, |
| "grad_norm": 4.39390230178833, |
| "learning_rate": 3.1081081081081082e-06, |
| "loss": 749.96, |
| "step": 357 |
| }, |
| { |
| "epoch": 0.6911196911196911, |
| "grad_norm": 4.870336055755615, |
| "learning_rate": 3.088803088803089e-06, |
| "loss": 750.3604, |
| "step": 358 |
| }, |
| { |
| "epoch": 0.693050193050193, |
| "grad_norm": 6.5269036293029785, |
| "learning_rate": 3.06949806949807e-06, |
| "loss": 765.9295, |
| "step": 359 |
| }, |
| { |
| "epoch": 0.694980694980695, |
| "grad_norm": 3.1647534370422363, |
| "learning_rate": 3.0501930501930503e-06, |
| "loss": 747.9913, |
| "step": 360 |
| }, |
| { |
| "epoch": 0.696911196911197, |
| "grad_norm": 3.435527801513672, |
| "learning_rate": 3.030888030888031e-06, |
| "loss": 770.7856, |
| "step": 361 |
| }, |
| { |
| "epoch": 0.6988416988416989, |
| "grad_norm": 5.05522346496582, |
| "learning_rate": 3.011583011583012e-06, |
| "loss": 773.0812, |
| "step": 362 |
| }, |
| { |
| "epoch": 0.7007722007722008, |
| "grad_norm": 4.161330223083496, |
| "learning_rate": 2.9922779922779927e-06, |
| "loss": 771.5272, |
| "step": 363 |
| }, |
| { |
| "epoch": 0.7027027027027027, |
| "grad_norm": 4.4345526695251465, |
| "learning_rate": 2.9729729729729736e-06, |
| "loss": 754.7959, |
| "step": 364 |
| }, |
| { |
| "epoch": 0.7046332046332047, |
| "grad_norm": 2.9777281284332275, |
| "learning_rate": 2.953667953667954e-06, |
| "loss": 749.3711, |
| "step": 365 |
| }, |
| { |
| "epoch": 0.7065637065637066, |
| "grad_norm": 2.855987310409546, |
| "learning_rate": 2.934362934362935e-06, |
| "loss": 749.5988, |
| "step": 366 |
| }, |
| { |
| "epoch": 0.7084942084942085, |
| "grad_norm": 3.307720184326172, |
| "learning_rate": 2.915057915057915e-06, |
| "loss": 761.0093, |
| "step": 367 |
| }, |
| { |
| "epoch": 0.7104247104247104, |
| "grad_norm": 3.787365198135376, |
| "learning_rate": 2.8957528957528956e-06, |
| "loss": 751.6588, |
| "step": 368 |
| }, |
| { |
| "epoch": 0.7123552123552124, |
| "grad_norm": 3.029486656188965, |
| "learning_rate": 2.8764478764478764e-06, |
| "loss": 755.2015, |
| "step": 369 |
| }, |
| { |
| "epoch": 0.7142857142857143, |
| "grad_norm": 3.0125091075897217, |
| "learning_rate": 2.8571428571428573e-06, |
| "loss": 766.543, |
| "step": 370 |
| }, |
| { |
| "epoch": 0.7162162162162162, |
| "grad_norm": 4.560346603393555, |
| "learning_rate": 2.837837837837838e-06, |
| "loss": 768.441, |
| "step": 371 |
| }, |
| { |
| "epoch": 0.7181467181467182, |
| "grad_norm": 3.190322160720825, |
| "learning_rate": 2.8185328185328185e-06, |
| "loss": 753.3053, |
| "step": 372 |
| }, |
| { |
| "epoch": 0.7200772200772201, |
| "grad_norm": 3.4040372371673584, |
| "learning_rate": 2.7992277992277993e-06, |
| "loss": 769.1228, |
| "step": 373 |
| }, |
| { |
| "epoch": 0.722007722007722, |
| "grad_norm": 5.434439182281494, |
| "learning_rate": 2.77992277992278e-06, |
| "loss": 754.5753, |
| "step": 374 |
| }, |
| { |
| "epoch": 0.7239382239382239, |
| "grad_norm": 3.3807380199432373, |
| "learning_rate": 2.760617760617761e-06, |
| "loss": 768.5261, |
| "step": 375 |
| }, |
| { |
| "epoch": 0.7258687258687259, |
| "grad_norm": 3.2653305530548096, |
| "learning_rate": 2.7413127413127418e-06, |
| "loss": 768.1964, |
| "step": 376 |
| }, |
| { |
| "epoch": 0.7277992277992278, |
| "grad_norm": 3.041628360748291, |
| "learning_rate": 2.722007722007722e-06, |
| "loss": 771.9323, |
| "step": 377 |
| }, |
| { |
| "epoch": 0.7297297297297297, |
| "grad_norm": 3.092245101928711, |
| "learning_rate": 2.702702702702703e-06, |
| "loss": 753.942, |
| "step": 378 |
| }, |
| { |
| "epoch": 0.7316602316602316, |
| "grad_norm": 4.467896938323975, |
| "learning_rate": 2.683397683397684e-06, |
| "loss": 749.8477, |
| "step": 379 |
| }, |
| { |
| "epoch": 0.7335907335907336, |
| "grad_norm": 2.989520788192749, |
| "learning_rate": 2.6640926640926647e-06, |
| "loss": 759.7245, |
| "step": 380 |
| }, |
| { |
| "epoch": 0.7355212355212355, |
| "grad_norm": 2.9288792610168457, |
| "learning_rate": 2.644787644787645e-06, |
| "loss": 761.0784, |
| "step": 381 |
| }, |
| { |
| "epoch": 0.7374517374517374, |
| "grad_norm": 3.411576986312866, |
| "learning_rate": 2.6254826254826255e-06, |
| "loss": 769.0735, |
| "step": 382 |
| }, |
| { |
| "epoch": 0.7393822393822393, |
| "grad_norm": 7.500954627990723, |
| "learning_rate": 2.6061776061776063e-06, |
| "loss": 752.552, |
| "step": 383 |
| }, |
| { |
| "epoch": 0.7413127413127413, |
| "grad_norm": 4.436896324157715, |
| "learning_rate": 2.5868725868725867e-06, |
| "loss": 772.7829, |
| "step": 384 |
| }, |
| { |
| "epoch": 0.7432432432432432, |
| "grad_norm": 3.8016252517700195, |
| "learning_rate": 2.5675675675675675e-06, |
| "loss": 756.5428, |
| "step": 385 |
| }, |
| { |
| "epoch": 0.7451737451737451, |
| "grad_norm": 3.1768035888671875, |
| "learning_rate": 2.5482625482625484e-06, |
| "loss": 761.3637, |
| "step": 386 |
| }, |
| { |
| "epoch": 0.747104247104247, |
| "grad_norm": 3.24792218208313, |
| "learning_rate": 2.528957528957529e-06, |
| "loss": 764.2406, |
| "step": 387 |
| }, |
| { |
| "epoch": 0.749034749034749, |
| "grad_norm": 3.6789443492889404, |
| "learning_rate": 2.50965250965251e-06, |
| "loss": 764.6515, |
| "step": 388 |
| }, |
| { |
| "epoch": 0.750965250965251, |
| "grad_norm": 2.998337507247925, |
| "learning_rate": 2.4903474903474904e-06, |
| "loss": 745.9545, |
| "step": 389 |
| }, |
| { |
| "epoch": 0.752895752895753, |
| "grad_norm": 5.5688796043396, |
| "learning_rate": 2.4710424710424712e-06, |
| "loss": 757.1937, |
| "step": 390 |
| }, |
| { |
| "epoch": 0.7548262548262549, |
| "grad_norm": 3.491425037384033, |
| "learning_rate": 2.451737451737452e-06, |
| "loss": 747.8929, |
| "step": 391 |
| }, |
| { |
| "epoch": 0.7567567567567568, |
| "grad_norm": 5.370115280151367, |
| "learning_rate": 2.432432432432433e-06, |
| "loss": 760.0563, |
| "step": 392 |
| }, |
| { |
| "epoch": 0.7586872586872587, |
| "grad_norm": 3.907552480697632, |
| "learning_rate": 2.4131274131274133e-06, |
| "loss": 752.7516, |
| "step": 393 |
| }, |
| { |
| "epoch": 0.7606177606177607, |
| "grad_norm": 3.052731513977051, |
| "learning_rate": 2.393822393822394e-06, |
| "loss": 752.1655, |
| "step": 394 |
| }, |
| { |
| "epoch": 0.7625482625482626, |
| "grad_norm": 5.308294296264648, |
| "learning_rate": 2.3745173745173745e-06, |
| "loss": 767.8257, |
| "step": 395 |
| }, |
| { |
| "epoch": 0.7644787644787645, |
| "grad_norm": 3.72119140625, |
| "learning_rate": 2.3552123552123553e-06, |
| "loss": 764.8139, |
| "step": 396 |
| }, |
| { |
| "epoch": 0.7664092664092664, |
| "grad_norm": 3.117110252380371, |
| "learning_rate": 2.335907335907336e-06, |
| "loss": 766.5566, |
| "step": 397 |
| }, |
| { |
| "epoch": 0.7683397683397684, |
| "grad_norm": 3.8100202083587646, |
| "learning_rate": 2.316602316602317e-06, |
| "loss": 767.1085, |
| "step": 398 |
| }, |
| { |
| "epoch": 0.7702702702702703, |
| "grad_norm": 3.1219418048858643, |
| "learning_rate": 2.297297297297298e-06, |
| "loss": 750.5438, |
| "step": 399 |
| }, |
| { |
| "epoch": 0.7722007722007722, |
| "grad_norm": 3.212282180786133, |
| "learning_rate": 2.2779922779922782e-06, |
| "loss": 748.5366, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.7741312741312741, |
| "grad_norm": 3.0505459308624268, |
| "learning_rate": 2.2586872586872586e-06, |
| "loss": 743.7197, |
| "step": 401 |
| }, |
| { |
| "epoch": 0.7760617760617761, |
| "grad_norm": 3.131042718887329, |
| "learning_rate": 2.2393822393822394e-06, |
| "loss": 756.7755, |
| "step": 402 |
| }, |
| { |
| "epoch": 0.777992277992278, |
| "grad_norm": 4.418597221374512, |
| "learning_rate": 2.2200772200772203e-06, |
| "loss": 754.0249, |
| "step": 403 |
| }, |
| { |
| "epoch": 0.7799227799227799, |
| "grad_norm": 4.271742820739746, |
| "learning_rate": 2.200772200772201e-06, |
| "loss": 768.3597, |
| "step": 404 |
| }, |
| { |
| "epoch": 0.7818532818532818, |
| "grad_norm": 3.788788080215454, |
| "learning_rate": 2.181467181467182e-06, |
| "loss": 758.9531, |
| "step": 405 |
| }, |
| { |
| "epoch": 0.7837837837837838, |
| "grad_norm": 3.5197062492370605, |
| "learning_rate": 2.1621621621621623e-06, |
| "loss": 754.1716, |
| "step": 406 |
| }, |
| { |
| "epoch": 0.7857142857142857, |
| "grad_norm": 3.414795398712158, |
| "learning_rate": 2.1428571428571427e-06, |
| "loss": 751.0524, |
| "step": 407 |
| }, |
| { |
| "epoch": 0.7876447876447876, |
| "grad_norm": 3.1258320808410645, |
| "learning_rate": 2.1235521235521236e-06, |
| "loss": 761.6645, |
| "step": 408 |
| }, |
| { |
| "epoch": 0.7895752895752896, |
| "grad_norm": 6.547064781188965, |
| "learning_rate": 2.1042471042471044e-06, |
| "loss": 765.9225, |
| "step": 409 |
| }, |
| { |
| "epoch": 0.7915057915057915, |
| "grad_norm": 4.772284030914307, |
| "learning_rate": 2.084942084942085e-06, |
| "loss": 764.9897, |
| "step": 410 |
| }, |
| { |
| "epoch": 0.7934362934362934, |
| "grad_norm": 4.287495136260986, |
| "learning_rate": 2.065637065637066e-06, |
| "loss": 756.7947, |
| "step": 411 |
| }, |
| { |
| "epoch": 0.7953667953667953, |
| "grad_norm": 5.286579132080078, |
| "learning_rate": 2.0463320463320464e-06, |
| "loss": 756.2988, |
| "step": 412 |
| }, |
| { |
| "epoch": 0.7972972972972973, |
| "grad_norm": 3.132582902908325, |
| "learning_rate": 2.0270270270270273e-06, |
| "loss": 760.6235, |
| "step": 413 |
| }, |
| { |
| "epoch": 0.7992277992277992, |
| "grad_norm": 4.088491439819336, |
| "learning_rate": 2.0077220077220077e-06, |
| "loss": 751.6579, |
| "step": 414 |
| }, |
| { |
| "epoch": 0.8011583011583011, |
| "grad_norm": 3.853825807571411, |
| "learning_rate": 1.9884169884169885e-06, |
| "loss": 751.7198, |
| "step": 415 |
| }, |
| { |
| "epoch": 0.803088803088803, |
| "grad_norm": 3.8805510997772217, |
| "learning_rate": 1.9691119691119693e-06, |
| "loss": 754.175, |
| "step": 416 |
| }, |
| { |
| "epoch": 0.805019305019305, |
| "grad_norm": 3.465831756591797, |
| "learning_rate": 1.94980694980695e-06, |
| "loss": 763.5966, |
| "step": 417 |
| }, |
| { |
| "epoch": 0.806949806949807, |
| "grad_norm": 3.4799764156341553, |
| "learning_rate": 1.9305019305019305e-06, |
| "loss": 757.6487, |
| "step": 418 |
| }, |
| { |
| "epoch": 0.8088803088803089, |
| "grad_norm": 4.306357383728027, |
| "learning_rate": 1.9111969111969114e-06, |
| "loss": 770.5115, |
| "step": 419 |
| }, |
| { |
| "epoch": 0.8108108108108109, |
| "grad_norm": 5.999144554138184, |
| "learning_rate": 1.8918918918918922e-06, |
| "loss": 750.0549, |
| "step": 420 |
| }, |
| { |
| "epoch": 0.8127413127413128, |
| "grad_norm": 3.353970527648926, |
| "learning_rate": 1.8725868725868726e-06, |
| "loss": 770.1407, |
| "step": 421 |
| }, |
| { |
| "epoch": 0.8146718146718147, |
| "grad_norm": 3.1123435497283936, |
| "learning_rate": 1.8532818532818534e-06, |
| "loss": 755.7714, |
| "step": 422 |
| }, |
| { |
| "epoch": 0.8166023166023166, |
| "grad_norm": 3.5919506549835205, |
| "learning_rate": 1.833976833976834e-06, |
| "loss": 768.7866, |
| "step": 423 |
| }, |
| { |
| "epoch": 0.8185328185328186, |
| "grad_norm": 5.815296649932861, |
| "learning_rate": 1.8146718146718149e-06, |
| "loss": 769.049, |
| "step": 424 |
| }, |
| { |
| "epoch": 0.8204633204633205, |
| "grad_norm": 6.508057117462158, |
| "learning_rate": 1.7953667953667955e-06, |
| "loss": 744.9335, |
| "step": 425 |
| }, |
| { |
| "epoch": 0.8223938223938224, |
| "grad_norm": 3.8051586151123047, |
| "learning_rate": 1.7760617760617763e-06, |
| "loss": 755.9141, |
| "step": 426 |
| }, |
| { |
| "epoch": 0.8243243243243243, |
| "grad_norm": 3.628269910812378, |
| "learning_rate": 1.756756756756757e-06, |
| "loss": 758.3513, |
| "step": 427 |
| }, |
| { |
| "epoch": 0.8262548262548263, |
| "grad_norm": 5.430158615112305, |
| "learning_rate": 1.7374517374517377e-06, |
| "loss": 754.9781, |
| "step": 428 |
| }, |
| { |
| "epoch": 0.8281853281853282, |
| "grad_norm": 3.8903748989105225, |
| "learning_rate": 1.7181467181467181e-06, |
| "loss": 762.0396, |
| "step": 429 |
| }, |
| { |
| "epoch": 0.8301158301158301, |
| "grad_norm": 3.2768187522888184, |
| "learning_rate": 1.698841698841699e-06, |
| "loss": 748.2692, |
| "step": 430 |
| }, |
| { |
| "epoch": 0.832046332046332, |
| "grad_norm": 4.247284889221191, |
| "learning_rate": 1.6795366795366796e-06, |
| "loss": 759.6968, |
| "step": 431 |
| }, |
| { |
| "epoch": 0.833976833976834, |
| "grad_norm": 3.205296039581299, |
| "learning_rate": 1.6602316602316604e-06, |
| "loss": 745.4493, |
| "step": 432 |
| }, |
| { |
| "epoch": 0.8359073359073359, |
| "grad_norm": 3.3943800926208496, |
| "learning_rate": 1.640926640926641e-06, |
| "loss": 759.4459, |
| "step": 433 |
| }, |
| { |
| "epoch": 0.8378378378378378, |
| "grad_norm": 3.8933515548706055, |
| "learning_rate": 1.6216216216216219e-06, |
| "loss": 768.5825, |
| "step": 434 |
| }, |
| { |
| "epoch": 0.8397683397683398, |
| "grad_norm": 3.123276710510254, |
| "learning_rate": 1.6023166023166027e-06, |
| "loss": 749.004, |
| "step": 435 |
| }, |
| { |
| "epoch": 0.8416988416988417, |
| "grad_norm": 3.028196334838867, |
| "learning_rate": 1.583011583011583e-06, |
| "loss": 754.6147, |
| "step": 436 |
| }, |
| { |
| "epoch": 0.8436293436293436, |
| "grad_norm": 5.627590179443359, |
| "learning_rate": 1.5637065637065637e-06, |
| "loss": 764.9931, |
| "step": 437 |
| }, |
| { |
| "epoch": 0.8455598455598455, |
| "grad_norm": 4.444796562194824, |
| "learning_rate": 1.5444015444015445e-06, |
| "loss": 756.2324, |
| "step": 438 |
| }, |
| { |
| "epoch": 0.8474903474903475, |
| "grad_norm": 4.994812965393066, |
| "learning_rate": 1.5250965250965251e-06, |
| "loss": 764.1788, |
| "step": 439 |
| }, |
| { |
| "epoch": 0.8494208494208494, |
| "grad_norm": 3.8100221157073975, |
| "learning_rate": 1.505791505791506e-06, |
| "loss": 748.0216, |
| "step": 440 |
| }, |
| { |
| "epoch": 0.8513513513513513, |
| "grad_norm": 3.284438371658325, |
| "learning_rate": 1.4864864864864868e-06, |
| "loss": 744.6481, |
| "step": 441 |
| }, |
| { |
| "epoch": 0.8532818532818532, |
| "grad_norm": 3.754772901535034, |
| "learning_rate": 1.4671814671814674e-06, |
| "loss": 759.6836, |
| "step": 442 |
| }, |
| { |
| "epoch": 0.8552123552123552, |
| "grad_norm": 3.4503631591796875, |
| "learning_rate": 1.4478764478764478e-06, |
| "loss": 757.1558, |
| "step": 443 |
| }, |
| { |
| "epoch": 0.8571428571428571, |
| "grad_norm": 3.9186127185821533, |
| "learning_rate": 1.4285714285714286e-06, |
| "loss": 749.7752, |
| "step": 444 |
| }, |
| { |
| "epoch": 0.859073359073359, |
| "grad_norm": 5.067336082458496, |
| "learning_rate": 1.4092664092664092e-06, |
| "loss": 768.12, |
| "step": 445 |
| }, |
| { |
| "epoch": 0.861003861003861, |
| "grad_norm": 3.5622620582580566, |
| "learning_rate": 1.38996138996139e-06, |
| "loss": 749.7698, |
| "step": 446 |
| }, |
| { |
| "epoch": 0.862934362934363, |
| "grad_norm": 3.2070393562316895, |
| "learning_rate": 1.3706563706563709e-06, |
| "loss": 765.8137, |
| "step": 447 |
| }, |
| { |
| "epoch": 0.8648648648648649, |
| "grad_norm": 3.920966386795044, |
| "learning_rate": 1.3513513513513515e-06, |
| "loss": 763.2885, |
| "step": 448 |
| }, |
| { |
| "epoch": 0.8667953667953668, |
| "grad_norm": 4.828410625457764, |
| "learning_rate": 1.3320463320463323e-06, |
| "loss": 744.1356, |
| "step": 449 |
| }, |
| { |
| "epoch": 0.8687258687258688, |
| "grad_norm": 2.9684271812438965, |
| "learning_rate": 1.3127413127413127e-06, |
| "loss": 748.5911, |
| "step": 450 |
| }, |
| { |
| "epoch": 0.8706563706563707, |
| "grad_norm": 3.4309277534484863, |
| "learning_rate": 1.2934362934362933e-06, |
| "loss": 754.4804, |
| "step": 451 |
| }, |
| { |
| "epoch": 0.8725868725868726, |
| "grad_norm": 3.2176759243011475, |
| "learning_rate": 1.2741312741312742e-06, |
| "loss": 763.7363, |
| "step": 452 |
| }, |
| { |
| "epoch": 0.8745173745173745, |
| "grad_norm": 4.412746429443359, |
| "learning_rate": 1.254826254826255e-06, |
| "loss": 762.2845, |
| "step": 453 |
| }, |
| { |
| "epoch": 0.8764478764478765, |
| "grad_norm": 3.3320093154907227, |
| "learning_rate": 1.2355212355212356e-06, |
| "loss": 737.4437, |
| "step": 454 |
| }, |
| { |
| "epoch": 0.8783783783783784, |
| "grad_norm": 3.1412718296051025, |
| "learning_rate": 1.2162162162162164e-06, |
| "loss": 752.1414, |
| "step": 455 |
| }, |
| { |
| "epoch": 0.8803088803088803, |
| "grad_norm": 2.9298524856567383, |
| "learning_rate": 1.196911196911197e-06, |
| "loss": 749.2881, |
| "step": 456 |
| }, |
| { |
| "epoch": 0.8822393822393823, |
| "grad_norm": 3.4683337211608887, |
| "learning_rate": 1.1776061776061777e-06, |
| "loss": 750.4893, |
| "step": 457 |
| }, |
| { |
| "epoch": 0.8841698841698842, |
| "grad_norm": 8.293436050415039, |
| "learning_rate": 1.1583011583011585e-06, |
| "loss": 759.465, |
| "step": 458 |
| }, |
| { |
| "epoch": 0.8861003861003861, |
| "grad_norm": 3.0078611373901367, |
| "learning_rate": 1.1389961389961391e-06, |
| "loss": 765.3209, |
| "step": 459 |
| }, |
| { |
| "epoch": 0.888030888030888, |
| "grad_norm": 3.5911474227905273, |
| "learning_rate": 1.1196911196911197e-06, |
| "loss": 748.1154, |
| "step": 460 |
| }, |
| { |
| "epoch": 0.88996138996139, |
| "grad_norm": 3.626723527908325, |
| "learning_rate": 1.1003861003861005e-06, |
| "loss": 763.6529, |
| "step": 461 |
| }, |
| { |
| "epoch": 0.8918918918918919, |
| "grad_norm": 3.322850465774536, |
| "learning_rate": 1.0810810810810812e-06, |
| "loss": 757.5947, |
| "step": 462 |
| }, |
| { |
| "epoch": 0.8938223938223938, |
| "grad_norm": 3.866856575012207, |
| "learning_rate": 1.0617760617760618e-06, |
| "loss": 761.439, |
| "step": 463 |
| }, |
| { |
| "epoch": 0.8957528957528957, |
| "grad_norm": 3.293731927871704, |
| "learning_rate": 1.0424710424710426e-06, |
| "loss": 763.1173, |
| "step": 464 |
| }, |
| { |
| "epoch": 0.8976833976833977, |
| "grad_norm": 3.4967000484466553, |
| "learning_rate": 1.0231660231660232e-06, |
| "loss": 771.8659, |
| "step": 465 |
| }, |
| { |
| "epoch": 0.8996138996138996, |
| "grad_norm": 2.9513790607452393, |
| "learning_rate": 1.0038610038610038e-06, |
| "loss": 759.2766, |
| "step": 466 |
| }, |
| { |
| "epoch": 0.9015444015444015, |
| "grad_norm": 3.051725387573242, |
| "learning_rate": 9.845559845559847e-07, |
| "loss": 761.7484, |
| "step": 467 |
| }, |
| { |
| "epoch": 0.9034749034749034, |
| "grad_norm": 3.222289562225342, |
| "learning_rate": 9.652509652509653e-07, |
| "loss": 753.144, |
| "step": 468 |
| }, |
| { |
| "epoch": 0.9054054054054054, |
| "grad_norm": 4.781485557556152, |
| "learning_rate": 9.459459459459461e-07, |
| "loss": 766.5778, |
| "step": 469 |
| }, |
| { |
| "epoch": 0.9073359073359073, |
| "grad_norm": 9.611944198608398, |
| "learning_rate": 9.266409266409267e-07, |
| "loss": 757.17, |
| "step": 470 |
| }, |
| { |
| "epoch": 0.9092664092664092, |
| "grad_norm": 5.166945457458496, |
| "learning_rate": 9.073359073359074e-07, |
| "loss": 758.192, |
| "step": 471 |
| }, |
| { |
| "epoch": 0.9111969111969112, |
| "grad_norm": 3.2668616771698, |
| "learning_rate": 8.880308880308882e-07, |
| "loss": 772.9589, |
| "step": 472 |
| }, |
| { |
| "epoch": 0.9131274131274131, |
| "grad_norm": 3.808668851852417, |
| "learning_rate": 8.687258687258689e-07, |
| "loss": 743.6304, |
| "step": 473 |
| }, |
| { |
| "epoch": 0.915057915057915, |
| "grad_norm": 3.6966071128845215, |
| "learning_rate": 8.494208494208495e-07, |
| "loss": 739.0352, |
| "step": 474 |
| }, |
| { |
| "epoch": 0.916988416988417, |
| "grad_norm": 5.23868465423584, |
| "learning_rate": 8.301158301158302e-07, |
| "loss": 758.2206, |
| "step": 475 |
| }, |
| { |
| "epoch": 0.918918918918919, |
| "grad_norm": 5.4658732414245605, |
| "learning_rate": 8.108108108108109e-07, |
| "loss": 745.1991, |
| "step": 476 |
| }, |
| { |
| "epoch": 0.9208494208494209, |
| "grad_norm": 4.291003227233887, |
| "learning_rate": 7.915057915057915e-07, |
| "loss": 765.403, |
| "step": 477 |
| }, |
| { |
| "epoch": 0.9227799227799228, |
| "grad_norm": 3.528947353363037, |
| "learning_rate": 7.722007722007723e-07, |
| "loss": 764.4042, |
| "step": 478 |
| }, |
| { |
| "epoch": 0.9247104247104247, |
| "grad_norm": 3.179927349090576, |
| "learning_rate": 7.52895752895753e-07, |
| "loss": 757.1093, |
| "step": 479 |
| }, |
| { |
| "epoch": 0.9266409266409267, |
| "grad_norm": 4.567697048187256, |
| "learning_rate": 7.335907335907337e-07, |
| "loss": 759.746, |
| "step": 480 |
| }, |
| { |
| "epoch": 0.9285714285714286, |
| "grad_norm": 3.929335355758667, |
| "learning_rate": 7.142857142857143e-07, |
| "loss": 758.0447, |
| "step": 481 |
| }, |
| { |
| "epoch": 0.9305019305019305, |
| "grad_norm": 5.094730377197266, |
| "learning_rate": 6.94980694980695e-07, |
| "loss": 757.1946, |
| "step": 482 |
| }, |
| { |
| "epoch": 0.9324324324324325, |
| "grad_norm": 3.7125465869903564, |
| "learning_rate": 6.756756756756758e-07, |
| "loss": 770.7656, |
| "step": 483 |
| }, |
| { |
| "epoch": 0.9343629343629344, |
| "grad_norm": 3.3825149536132812, |
| "learning_rate": 6.563706563706564e-07, |
| "loss": 747.5059, |
| "step": 484 |
| }, |
| { |
| "epoch": 0.9362934362934363, |
| "grad_norm": 3.3807835578918457, |
| "learning_rate": 6.370656370656371e-07, |
| "loss": 767.5899, |
| "step": 485 |
| }, |
| { |
| "epoch": 0.9382239382239382, |
| "grad_norm": 2.997955322265625, |
| "learning_rate": 6.177606177606178e-07, |
| "loss": 758.338, |
| "step": 486 |
| }, |
| { |
| "epoch": 0.9401544401544402, |
| "grad_norm": 3.478146553039551, |
| "learning_rate": 5.984555984555985e-07, |
| "loss": 758.3116, |
| "step": 487 |
| }, |
| { |
| "epoch": 0.9420849420849421, |
| "grad_norm": 3.0303502082824707, |
| "learning_rate": 5.791505791505792e-07, |
| "loss": 752.7023, |
| "step": 488 |
| }, |
| { |
| "epoch": 0.944015444015444, |
| "grad_norm": 4.182223796844482, |
| "learning_rate": 5.598455598455599e-07, |
| "loss": 767.7703, |
| "step": 489 |
| }, |
| { |
| "epoch": 0.9459459459459459, |
| "grad_norm": 3.0645620822906494, |
| "learning_rate": 5.405405405405406e-07, |
| "loss": 756.7206, |
| "step": 490 |
| }, |
| { |
| "epoch": 0.9478764478764479, |
| "grad_norm": 3.5093557834625244, |
| "learning_rate": 5.212355212355213e-07, |
| "loss": 759.0178, |
| "step": 491 |
| }, |
| { |
| "epoch": 0.9498069498069498, |
| "grad_norm": 5.231660842895508, |
| "learning_rate": 5.019305019305019e-07, |
| "loss": 758.1909, |
| "step": 492 |
| }, |
| { |
| "epoch": 0.9517374517374517, |
| "grad_norm": 4.316704273223877, |
| "learning_rate": 4.826254826254826e-07, |
| "loss": 756.6906, |
| "step": 493 |
| }, |
| { |
| "epoch": 0.9536679536679536, |
| "grad_norm": 3.842200994491577, |
| "learning_rate": 4.6332046332046336e-07, |
| "loss": 757.8163, |
| "step": 494 |
| }, |
| { |
| "epoch": 0.9555984555984556, |
| "grad_norm": 3.351325511932373, |
| "learning_rate": 4.440154440154441e-07, |
| "loss": 762.0304, |
| "step": 495 |
| }, |
| { |
| "epoch": 0.9575289575289575, |
| "grad_norm": 3.0159342288970947, |
| "learning_rate": 4.2471042471042474e-07, |
| "loss": 761.4176, |
| "step": 496 |
| }, |
| { |
| "epoch": 0.9594594594594594, |
| "grad_norm": 3.074551582336426, |
| "learning_rate": 4.0540540540540546e-07, |
| "loss": 751.5521, |
| "step": 497 |
| }, |
| { |
| "epoch": 0.9613899613899614, |
| "grad_norm": 3.896904945373535, |
| "learning_rate": 3.8610038610038613e-07, |
| "loss": 763.3478, |
| "step": 498 |
| }, |
| { |
| "epoch": 0.9633204633204633, |
| "grad_norm": 3.919020175933838, |
| "learning_rate": 3.6679536679536685e-07, |
| "loss": 768.0314, |
| "step": 499 |
| }, |
| { |
| "epoch": 0.9652509652509652, |
| "grad_norm": 2.949843645095825, |
| "learning_rate": 3.474903474903475e-07, |
| "loss": 755.5469, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.9671814671814671, |
| "grad_norm": 3.6389143466949463, |
| "learning_rate": 3.281853281853282e-07, |
| "loss": 744.184, |
| "step": 501 |
| }, |
| { |
| "epoch": 0.9691119691119691, |
| "grad_norm": 4.229357719421387, |
| "learning_rate": 3.088803088803089e-07, |
| "loss": 761.7263, |
| "step": 502 |
| }, |
| { |
| "epoch": 0.971042471042471, |
| "grad_norm": 3.1734883785247803, |
| "learning_rate": 2.895752895752896e-07, |
| "loss": 758.8358, |
| "step": 503 |
| }, |
| { |
| "epoch": 0.972972972972973, |
| "grad_norm": 3.089228868484497, |
| "learning_rate": 2.702702702702703e-07, |
| "loss": 760.397, |
| "step": 504 |
| }, |
| { |
| "epoch": 0.974903474903475, |
| "grad_norm": 3.926948070526123, |
| "learning_rate": 2.5096525096525096e-07, |
| "loss": 755.6573, |
| "step": 505 |
| }, |
| { |
| "epoch": 0.9768339768339769, |
| "grad_norm": 2.9857828617095947, |
| "learning_rate": 2.3166023166023168e-07, |
| "loss": 767.6666, |
| "step": 506 |
| }, |
| { |
| "epoch": 0.9787644787644788, |
| "grad_norm": 4.68814754486084, |
| "learning_rate": 2.1235521235521237e-07, |
| "loss": 767.6501, |
| "step": 507 |
| }, |
| { |
| "epoch": 0.9806949806949807, |
| "grad_norm": 3.457016706466675, |
| "learning_rate": 1.9305019305019306e-07, |
| "loss": 760.5441, |
| "step": 508 |
| }, |
| { |
| "epoch": 0.9826254826254827, |
| "grad_norm": 4.365758419036865, |
| "learning_rate": 1.7374517374517376e-07, |
| "loss": 760.8904, |
| "step": 509 |
| }, |
| { |
| "epoch": 0.9845559845559846, |
| "grad_norm": 3.6141843795776367, |
| "learning_rate": 1.5444015444015445e-07, |
| "loss": 754.2842, |
| "step": 510 |
| }, |
| { |
| "epoch": 0.9864864864864865, |
| "grad_norm": 4.281587600708008, |
| "learning_rate": 1.3513513513513515e-07, |
| "loss": 754.181, |
| "step": 511 |
| }, |
| { |
| "epoch": 0.9884169884169884, |
| "grad_norm": 3.3735556602478027, |
| "learning_rate": 1.1583011583011584e-07, |
| "loss": 755.9177, |
| "step": 512 |
| }, |
| { |
| "epoch": 0.9903474903474904, |
| "grad_norm": 3.126232624053955, |
| "learning_rate": 9.652509652509653e-08, |
| "loss": 750.3485, |
| "step": 513 |
| }, |
| { |
| "epoch": 0.9922779922779923, |
| "grad_norm": 5.9563727378845215, |
| "learning_rate": 7.722007722007723e-08, |
| "loss": 766.379, |
| "step": 514 |
| }, |
| { |
| "epoch": 0.9942084942084942, |
| "grad_norm": 3.0199837684631348, |
| "learning_rate": 5.791505791505792e-08, |
| "loss": 753.6818, |
| "step": 515 |
| }, |
| { |
| "epoch": 0.9961389961389961, |
| "grad_norm": 3.0909945964813232, |
| "learning_rate": 3.861003861003861e-08, |
| "loss": 749.71, |
| "step": 516 |
| }, |
| { |
| "epoch": 0.9980694980694981, |
| "grad_norm": 3.416806221008301, |
| "learning_rate": 1.9305019305019306e-08, |
| "loss": 772.7269, |
| "step": 517 |
| }, |
| { |
| "epoch": 1.0, |
| "grad_norm": 5.65108585357666, |
| "learning_rate": 0.0, |
| "loss": 744.4813, |
| "step": 518 |
| }, |
| { |
| "epoch": 1.0, |
| "step": 518, |
| "total_flos": 1.1089078720895386e+18, |
| "train_loss": 758.2892525573495, |
| "train_runtime": 2215.2409, |
| "train_samples_per_second": 239.051, |
| "train_steps_per_second": 0.234 |
| } |
| ], |
| "logging_steps": 1, |
| "max_steps": 518, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 1, |
| "save_steps": 500, |
| "total_flos": 1.1089078720895386e+18, |
| "train_batch_size": 1024, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|