| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 1.0, |
| "eval_steps": 500, |
| "global_step": 335, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.0029850746268656717, |
| "grad_norm": 8.253611034452266, |
| "learning_rate": 9.090909090909091e-07, |
| "loss": 2.2247, |
| "step": 1 |
| }, |
| { |
| "epoch": 0.005970149253731343, |
| "grad_norm": 7.395306126962346, |
| "learning_rate": 1.8181818181818183e-06, |
| "loss": 2.1255, |
| "step": 2 |
| }, |
| { |
| "epoch": 0.008955223880597015, |
| "grad_norm": 8.26748388260573, |
| "learning_rate": 2.7272727272727272e-06, |
| "loss": 2.1023, |
| "step": 3 |
| }, |
| { |
| "epoch": 0.011940298507462687, |
| "grad_norm": 7.1200500158632, |
| "learning_rate": 3.6363636363636366e-06, |
| "loss": 2.0025, |
| "step": 4 |
| }, |
| { |
| "epoch": 0.014925373134328358, |
| "grad_norm": 5.9633773670024315, |
| "learning_rate": 4.5454545454545455e-06, |
| "loss": 1.9076, |
| "step": 5 |
| }, |
| { |
| "epoch": 0.01791044776119403, |
| "grad_norm": 5.546547205735884, |
| "learning_rate": 5.4545454545454545e-06, |
| "loss": 1.8587, |
| "step": 6 |
| }, |
| { |
| "epoch": 0.020895522388059702, |
| "grad_norm": 5.042219157986929, |
| "learning_rate": 6.363636363636364e-06, |
| "loss": 1.8198, |
| "step": 7 |
| }, |
| { |
| "epoch": 0.023880597014925373, |
| "grad_norm": 3.5185430665767887, |
| "learning_rate": 7.272727272727273e-06, |
| "loss": 1.5117, |
| "step": 8 |
| }, |
| { |
| "epoch": 0.026865671641791045, |
| "grad_norm": 3.1354746744528867, |
| "learning_rate": 8.181818181818183e-06, |
| "loss": 1.4414, |
| "step": 9 |
| }, |
| { |
| "epoch": 0.029850746268656716, |
| "grad_norm": 2.757132384211869, |
| "learning_rate": 9.090909090909091e-06, |
| "loss": 1.4059, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.03283582089552239, |
| "grad_norm": 5.307800727766051, |
| "learning_rate": 1e-05, |
| "loss": 1.4367, |
| "step": 11 |
| }, |
| { |
| "epoch": 0.03582089552238806, |
| "grad_norm": 4.364662300190612, |
| "learning_rate": 9.99976495753613e-06, |
| "loss": 1.4175, |
| "step": 12 |
| }, |
| { |
| "epoch": 0.03880597014925373, |
| "grad_norm": 3.136010913151125, |
| "learning_rate": 9.999059852242508e-06, |
| "loss": 1.297, |
| "step": 13 |
| }, |
| { |
| "epoch": 0.041791044776119404, |
| "grad_norm": 2.6314270494566716, |
| "learning_rate": 9.997884750411004e-06, |
| "loss": 1.178, |
| "step": 14 |
| }, |
| { |
| "epoch": 0.04477611940298507, |
| "grad_norm": 2.8108610676101105, |
| "learning_rate": 9.996239762521152e-06, |
| "loss": 1.311, |
| "step": 15 |
| }, |
| { |
| "epoch": 0.04776119402985075, |
| "grad_norm": 2.646130122162001, |
| "learning_rate": 9.994125043229753e-06, |
| "loss": 1.1026, |
| "step": 16 |
| }, |
| { |
| "epoch": 0.050746268656716415, |
| "grad_norm": 2.3026683820565506, |
| "learning_rate": 9.991540791356342e-06, |
| "loss": 1.0708, |
| "step": 17 |
| }, |
| { |
| "epoch": 0.05373134328358209, |
| "grad_norm": 2.2060475680712157, |
| "learning_rate": 9.98848724986449e-06, |
| "loss": 1.1484, |
| "step": 18 |
| }, |
| { |
| "epoch": 0.056716417910447764, |
| "grad_norm": 2.22608658907851, |
| "learning_rate": 9.98496470583896e-06, |
| "loss": 1.1819, |
| "step": 19 |
| }, |
| { |
| "epoch": 0.05970149253731343, |
| "grad_norm": 2.012346588234655, |
| "learning_rate": 9.980973490458728e-06, |
| "loss": 1.1226, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.0626865671641791, |
| "grad_norm": 2.0805737688444155, |
| "learning_rate": 9.976513978965829e-06, |
| "loss": 1.0251, |
| "step": 21 |
| }, |
| { |
| "epoch": 0.06567164179104477, |
| "grad_norm": 2.384555566832582, |
| "learning_rate": 9.971586590630094e-06, |
| "loss": 1.0278, |
| "step": 22 |
| }, |
| { |
| "epoch": 0.06865671641791045, |
| "grad_norm": 2.1582387653704886, |
| "learning_rate": 9.966191788709716e-06, |
| "loss": 1.014, |
| "step": 23 |
| }, |
| { |
| "epoch": 0.07164179104477612, |
| "grad_norm": 2.0561968840026026, |
| "learning_rate": 9.960330080407712e-06, |
| "loss": 1.0074, |
| "step": 24 |
| }, |
| { |
| "epoch": 0.07462686567164178, |
| "grad_norm": 2.1634547485342828, |
| "learning_rate": 9.954002016824226e-06, |
| "loss": 1.0355, |
| "step": 25 |
| }, |
| { |
| "epoch": 0.07761194029850746, |
| "grad_norm": 2.000183275885172, |
| "learning_rate": 9.947208192904722e-06, |
| "loss": 1.0241, |
| "step": 26 |
| }, |
| { |
| "epoch": 0.08059701492537313, |
| "grad_norm": 1.9928462865297418, |
| "learning_rate": 9.939949247384046e-06, |
| "loss": 0.9917, |
| "step": 27 |
| }, |
| { |
| "epoch": 0.08358208955223881, |
| "grad_norm": 2.1100381727065094, |
| "learning_rate": 9.93222586272637e-06, |
| "loss": 1.0393, |
| "step": 28 |
| }, |
| { |
| "epoch": 0.08656716417910448, |
| "grad_norm": 1.9979189088773692, |
| "learning_rate": 9.924038765061042e-06, |
| "loss": 1.0134, |
| "step": 29 |
| }, |
| { |
| "epoch": 0.08955223880597014, |
| "grad_norm": 2.1566867938814784, |
| "learning_rate": 9.915388724114301e-06, |
| "loss": 0.937, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.09253731343283582, |
| "grad_norm": 2.0548646219588886, |
| "learning_rate": 9.906276553136924e-06, |
| "loss": 1.0227, |
| "step": 31 |
| }, |
| { |
| "epoch": 0.0955223880597015, |
| "grad_norm": 2.1094100666663493, |
| "learning_rate": 9.896703108827758e-06, |
| "loss": 0.9474, |
| "step": 32 |
| }, |
| { |
| "epoch": 0.09850746268656717, |
| "grad_norm": 2.196878885778515, |
| "learning_rate": 9.886669291253178e-06, |
| "loss": 0.8954, |
| "step": 33 |
| }, |
| { |
| "epoch": 0.10149253731343283, |
| "grad_norm": 1.9709535105103204, |
| "learning_rate": 9.876176043762467e-06, |
| "loss": 0.8858, |
| "step": 34 |
| }, |
| { |
| "epoch": 0.1044776119402985, |
| "grad_norm": 2.4157359301325028, |
| "learning_rate": 9.86522435289912e-06, |
| "loss": 0.9492, |
| "step": 35 |
| }, |
| { |
| "epoch": 0.10746268656716418, |
| "grad_norm": 2.0251547989723453, |
| "learning_rate": 9.853815248308101e-06, |
| "loss": 0.9819, |
| "step": 36 |
| }, |
| { |
| "epoch": 0.11044776119402985, |
| "grad_norm": 2.043073108115636, |
| "learning_rate": 9.841949802639031e-06, |
| "loss": 0.9856, |
| "step": 37 |
| }, |
| { |
| "epoch": 0.11343283582089553, |
| "grad_norm": 1.9051508783430386, |
| "learning_rate": 9.829629131445342e-06, |
| "loss": 0.8811, |
| "step": 38 |
| }, |
| { |
| "epoch": 0.11641791044776119, |
| "grad_norm": 1.9720465747167384, |
| "learning_rate": 9.816854393079402e-06, |
| "loss": 0.8809, |
| "step": 39 |
| }, |
| { |
| "epoch": 0.11940298507462686, |
| "grad_norm": 2.115788113435234, |
| "learning_rate": 9.803626788583603e-06, |
| "loss": 0.8754, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.12238805970149254, |
| "grad_norm": 2.0483615805890003, |
| "learning_rate": 9.789947561577445e-06, |
| "loss": 0.9111, |
| "step": 41 |
| }, |
| { |
| "epoch": 0.1253731343283582, |
| "grad_norm": 2.0931565027152925, |
| "learning_rate": 9.775817998140615e-06, |
| "loss": 0.883, |
| "step": 42 |
| }, |
| { |
| "epoch": 0.12835820895522387, |
| "grad_norm": 2.1619767559285896, |
| "learning_rate": 9.761239426692077e-06, |
| "loss": 0.8859, |
| "step": 43 |
| }, |
| { |
| "epoch": 0.13134328358208955, |
| "grad_norm": 2.1803394559484515, |
| "learning_rate": 9.74621321786517e-06, |
| "loss": 0.9756, |
| "step": 44 |
| }, |
| { |
| "epoch": 0.13432835820895522, |
| "grad_norm": 1.9473692197801435, |
| "learning_rate": 9.730740784378755e-06, |
| "loss": 0.8856, |
| "step": 45 |
| }, |
| { |
| "epoch": 0.1373134328358209, |
| "grad_norm": 1.9440539475047278, |
| "learning_rate": 9.71482358090438e-06, |
| "loss": 0.8619, |
| "step": 46 |
| }, |
| { |
| "epoch": 0.14029850746268657, |
| "grad_norm": 1.9268937369891654, |
| "learning_rate": 9.698463103929542e-06, |
| "loss": 0.8958, |
| "step": 47 |
| }, |
| { |
| "epoch": 0.14328358208955225, |
| "grad_norm": 1.9576735974697994, |
| "learning_rate": 9.681660891616967e-06, |
| "loss": 0.9014, |
| "step": 48 |
| }, |
| { |
| "epoch": 0.14626865671641792, |
| "grad_norm": 2.2502624703010787, |
| "learning_rate": 9.664418523660004e-06, |
| "loss": 0.9054, |
| "step": 49 |
| }, |
| { |
| "epoch": 0.14925373134328357, |
| "grad_norm": 1.9625578640953845, |
| "learning_rate": 9.646737621134112e-06, |
| "loss": 0.9438, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.15223880597014924, |
| "grad_norm": 2.11803005668703, |
| "learning_rate": 9.628619846344453e-06, |
| "loss": 0.933, |
| "step": 51 |
| }, |
| { |
| "epoch": 0.15522388059701492, |
| "grad_norm": 1.9405175828921, |
| "learning_rate": 9.610066902669593e-06, |
| "loss": 0.892, |
| "step": 52 |
| }, |
| { |
| "epoch": 0.1582089552238806, |
| "grad_norm": 1.940879796450495, |
| "learning_rate": 9.591080534401371e-06, |
| "loss": 0.8912, |
| "step": 53 |
| }, |
| { |
| "epoch": 0.16119402985074627, |
| "grad_norm": 2.1315443999032726, |
| "learning_rate": 9.571662526580898e-06, |
| "loss": 0.8837, |
| "step": 54 |
| }, |
| { |
| "epoch": 0.16417910447761194, |
| "grad_norm": 1.9936535994324662, |
| "learning_rate": 9.551814704830734e-06, |
| "loss": 0.936, |
| "step": 55 |
| }, |
| { |
| "epoch": 0.16716417910447762, |
| "grad_norm": 2.119381573598528, |
| "learning_rate": 9.531538935183252e-06, |
| "loss": 0.8668, |
| "step": 56 |
| }, |
| { |
| "epoch": 0.1701492537313433, |
| "grad_norm": 2.0767671264680665, |
| "learning_rate": 9.51083712390519e-06, |
| "loss": 0.9144, |
| "step": 57 |
| }, |
| { |
| "epoch": 0.17313432835820897, |
| "grad_norm": 1.7259669938296525, |
| "learning_rate": 9.48971121731844e-06, |
| "loss": 0.7978, |
| "step": 58 |
| }, |
| { |
| "epoch": 0.1761194029850746, |
| "grad_norm": 2.2712488027683895, |
| "learning_rate": 9.468163201617063e-06, |
| "loss": 0.8566, |
| "step": 59 |
| }, |
| { |
| "epoch": 0.1791044776119403, |
| "grad_norm": 2.1653559706182692, |
| "learning_rate": 9.446195102680531e-06, |
| "loss": 0.8772, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.18208955223880596, |
| "grad_norm": 2.019437924263334, |
| "learning_rate": 9.423808985883289e-06, |
| "loss": 0.8102, |
| "step": 61 |
| }, |
| { |
| "epoch": 0.18507462686567164, |
| "grad_norm": 2.1016655793843895, |
| "learning_rate": 9.401006955900555e-06, |
| "loss": 0.8829, |
| "step": 62 |
| }, |
| { |
| "epoch": 0.1880597014925373, |
| "grad_norm": 1.7789596254753794, |
| "learning_rate": 9.377791156510456e-06, |
| "loss": 0.8882, |
| "step": 63 |
| }, |
| { |
| "epoch": 0.191044776119403, |
| "grad_norm": 3.956571817793095, |
| "learning_rate": 9.35416377039246e-06, |
| "loss": 0.9216, |
| "step": 64 |
| }, |
| { |
| "epoch": 0.19402985074626866, |
| "grad_norm": 2.0814813313495764, |
| "learning_rate": 9.330127018922195e-06, |
| "loss": 0.8714, |
| "step": 65 |
| }, |
| { |
| "epoch": 0.19701492537313434, |
| "grad_norm": 2.16440601583317, |
| "learning_rate": 9.305683161962569e-06, |
| "loss": 0.9033, |
| "step": 66 |
| }, |
| { |
| "epoch": 0.2, |
| "grad_norm": 2.121322858320113, |
| "learning_rate": 9.280834497651334e-06, |
| "loss": 0.8866, |
| "step": 67 |
| }, |
| { |
| "epoch": 0.20298507462686566, |
| "grad_norm": 2.125810496652349, |
| "learning_rate": 9.255583362184998e-06, |
| "loss": 0.9084, |
| "step": 68 |
| }, |
| { |
| "epoch": 0.20597014925373133, |
| "grad_norm": 2.2332151510392926, |
| "learning_rate": 9.229932129599206e-06, |
| "loss": 0.8995, |
| "step": 69 |
| }, |
| { |
| "epoch": 0.208955223880597, |
| "grad_norm": 1.8844452106307579, |
| "learning_rate": 9.203883211545517e-06, |
| "loss": 0.8985, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.21194029850746268, |
| "grad_norm": 2.1170173520344253, |
| "learning_rate": 9.177439057064684e-06, |
| "loss": 0.8688, |
| "step": 71 |
| }, |
| { |
| "epoch": 0.21492537313432836, |
| "grad_norm": 1.9805965194267483, |
| "learning_rate": 9.150602152356394e-06, |
| "loss": 0.8909, |
| "step": 72 |
| }, |
| { |
| "epoch": 0.21791044776119403, |
| "grad_norm": 2.1059921689798116, |
| "learning_rate": 9.123375020545534e-06, |
| "loss": 0.9781, |
| "step": 73 |
| }, |
| { |
| "epoch": 0.2208955223880597, |
| "grad_norm": 1.9571325614730988, |
| "learning_rate": 9.09576022144496e-06, |
| "loss": 0.8519, |
| "step": 74 |
| }, |
| { |
| "epoch": 0.22388059701492538, |
| "grad_norm": 2.1730772437003867, |
| "learning_rate": 9.067760351314838e-06, |
| "loss": 0.8709, |
| "step": 75 |
| }, |
| { |
| "epoch": 0.22686567164179106, |
| "grad_norm": 2.236623739502403, |
| "learning_rate": 9.039378042618556e-06, |
| "loss": 1.0019, |
| "step": 76 |
| }, |
| { |
| "epoch": 0.2298507462686567, |
| "grad_norm": 2.049051045380969, |
| "learning_rate": 9.01061596377522e-06, |
| "loss": 0.8222, |
| "step": 77 |
| }, |
| { |
| "epoch": 0.23283582089552238, |
| "grad_norm": 2.0708856760876833, |
| "learning_rate": 8.981476818908778e-06, |
| "loss": 0.9427, |
| "step": 78 |
| }, |
| { |
| "epoch": 0.23582089552238805, |
| "grad_norm": 1.796129619411125, |
| "learning_rate": 8.951963347593797e-06, |
| "loss": 0.8325, |
| "step": 79 |
| }, |
| { |
| "epoch": 0.23880597014925373, |
| "grad_norm": 2.1961650618323687, |
| "learning_rate": 8.92207832459788e-06, |
| "loss": 0.9022, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.2417910447761194, |
| "grad_norm": 2.1631152148881023, |
| "learning_rate": 8.891824559620801e-06, |
| "loss": 0.8278, |
| "step": 81 |
| }, |
| { |
| "epoch": 0.24477611940298508, |
| "grad_norm": 2.025147970703757, |
| "learning_rate": 8.861204897030346e-06, |
| "loss": 0.7237, |
| "step": 82 |
| }, |
| { |
| "epoch": 0.24776119402985075, |
| "grad_norm": 1.9375472063035821, |
| "learning_rate": 8.83022221559489e-06, |
| "loss": 0.7376, |
| "step": 83 |
| }, |
| { |
| "epoch": 0.2507462686567164, |
| "grad_norm": 1.9368693802128247, |
| "learning_rate": 8.798879428212748e-06, |
| "loss": 0.8809, |
| "step": 84 |
| }, |
| { |
| "epoch": 0.2537313432835821, |
| "grad_norm": 1.7919474666326674, |
| "learning_rate": 8.767179481638303e-06, |
| "loss": 0.827, |
| "step": 85 |
| }, |
| { |
| "epoch": 0.25671641791044775, |
| "grad_norm": 1.8293764770314178, |
| "learning_rate": 8.735125356204982e-06, |
| "loss": 0.8348, |
| "step": 86 |
| }, |
| { |
| "epoch": 0.25970149253731345, |
| "grad_norm": 2.2466601113818983, |
| "learning_rate": 8.702720065545024e-06, |
| "loss": 0.7715, |
| "step": 87 |
| }, |
| { |
| "epoch": 0.2626865671641791, |
| "grad_norm": 2.007946173229526, |
| "learning_rate": 8.669966656306176e-06, |
| "loss": 0.7911, |
| "step": 88 |
| }, |
| { |
| "epoch": 0.2656716417910448, |
| "grad_norm": 1.8905119743926408, |
| "learning_rate": 8.636868207865244e-06, |
| "loss": 0.8525, |
| "step": 89 |
| }, |
| { |
| "epoch": 0.26865671641791045, |
| "grad_norm": 2.216132567769806, |
| "learning_rate": 8.603427832038574e-06, |
| "loss": 0.8743, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.2716417910447761, |
| "grad_norm": 2.0337720619135142, |
| "learning_rate": 8.569648672789496e-06, |
| "loss": 0.8652, |
| "step": 91 |
| }, |
| { |
| "epoch": 0.2746268656716418, |
| "grad_norm": 2.117953862074952, |
| "learning_rate": 8.535533905932739e-06, |
| "loss": 0.8499, |
| "step": 92 |
| }, |
| { |
| "epoch": 0.27761194029850744, |
| "grad_norm": 2.013179581976092, |
| "learning_rate": 8.501086738835843e-06, |
| "loss": 0.8635, |
| "step": 93 |
| }, |
| { |
| "epoch": 0.28059701492537314, |
| "grad_norm": 2.148194857680393, |
| "learning_rate": 8.466310410117622e-06, |
| "loss": 0.8781, |
| "step": 94 |
| }, |
| { |
| "epoch": 0.2835820895522388, |
| "grad_norm": 2.1704401998940024, |
| "learning_rate": 8.43120818934367e-06, |
| "loss": 0.8618, |
| "step": 95 |
| }, |
| { |
| "epoch": 0.2865671641791045, |
| "grad_norm": 2.065149455076995, |
| "learning_rate": 8.395783376718967e-06, |
| "loss": 0.8133, |
| "step": 96 |
| }, |
| { |
| "epoch": 0.28955223880597014, |
| "grad_norm": 2.0953263142220693, |
| "learning_rate": 8.360039302777614e-06, |
| "loss": 0.8189, |
| "step": 97 |
| }, |
| { |
| "epoch": 0.29253731343283584, |
| "grad_norm": 1.8700626853391247, |
| "learning_rate": 8.323979328069689e-06, |
| "loss": 0.8338, |
| "step": 98 |
| }, |
| { |
| "epoch": 0.2955223880597015, |
| "grad_norm": 1.886379993426298, |
| "learning_rate": 8.28760684284532e-06, |
| "loss": 0.8397, |
| "step": 99 |
| }, |
| { |
| "epoch": 0.29850746268656714, |
| "grad_norm": 4.887769993107663, |
| "learning_rate": 8.25092526673592e-06, |
| "loss": 0.9114, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.30149253731343284, |
| "grad_norm": 2.1831478421086237, |
| "learning_rate": 8.213938048432697e-06, |
| "loss": 0.838, |
| "step": 101 |
| }, |
| { |
| "epoch": 0.3044776119402985, |
| "grad_norm": 2.048417012849108, |
| "learning_rate": 8.176648665362426e-06, |
| "loss": 0.7323, |
| "step": 102 |
| }, |
| { |
| "epoch": 0.3074626865671642, |
| "grad_norm": 2.245486182262689, |
| "learning_rate": 8.139060623360494e-06, |
| "loss": 0.8361, |
| "step": 103 |
| }, |
| { |
| "epoch": 0.31044776119402984, |
| "grad_norm": 2.1131445651881937, |
| "learning_rate": 8.101177456341301e-06, |
| "loss": 0.8112, |
| "step": 104 |
| }, |
| { |
| "epoch": 0.31343283582089554, |
| "grad_norm": 1.9438690743970823, |
| "learning_rate": 8.063002725966014e-06, |
| "loss": 0.8172, |
| "step": 105 |
| }, |
| { |
| "epoch": 0.3164179104477612, |
| "grad_norm": 1.8330162019695133, |
| "learning_rate": 8.024540021307709e-06, |
| "loss": 0.7779, |
| "step": 106 |
| }, |
| { |
| "epoch": 0.3194029850746269, |
| "grad_norm": 1.9694508107519921, |
| "learning_rate": 7.985792958513932e-06, |
| "loss": 0.8582, |
| "step": 107 |
| }, |
| { |
| "epoch": 0.32238805970149254, |
| "grad_norm": 1.9336855933985957, |
| "learning_rate": 7.946765180466725e-06, |
| "loss": 0.7725, |
| "step": 108 |
| }, |
| { |
| "epoch": 0.3253731343283582, |
| "grad_norm": 2.0505954912615127, |
| "learning_rate": 7.907460356440133e-06, |
| "loss": 0.8778, |
| "step": 109 |
| }, |
| { |
| "epoch": 0.3283582089552239, |
| "grad_norm": 2.2263425751875183, |
| "learning_rate": 7.86788218175523e-06, |
| "loss": 0.8253, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.33134328358208953, |
| "grad_norm": 1.8159778540989733, |
| "learning_rate": 7.828034377432694e-06, |
| "loss": 0.8297, |
| "step": 111 |
| }, |
| { |
| "epoch": 0.33432835820895523, |
| "grad_norm": 1.9962851923382634, |
| "learning_rate": 7.787920689842965e-06, |
| "loss": 0.8761, |
| "step": 112 |
| }, |
| { |
| "epoch": 0.3373134328358209, |
| "grad_norm": 1.978659019282353, |
| "learning_rate": 7.747544890354031e-06, |
| "loss": 0.8364, |
| "step": 113 |
| }, |
| { |
| "epoch": 0.3402985074626866, |
| "grad_norm": 1.8570435757482742, |
| "learning_rate": 7.706910774976849e-06, |
| "loss": 0.7402, |
| "step": 114 |
| }, |
| { |
| "epoch": 0.34328358208955223, |
| "grad_norm": 1.9873552845505367, |
| "learning_rate": 7.666022164008458e-06, |
| "loss": 0.8159, |
| "step": 115 |
| }, |
| { |
| "epoch": 0.34626865671641793, |
| "grad_norm": 1.8774777253882997, |
| "learning_rate": 7.624882901672801e-06, |
| "loss": 0.8035, |
| "step": 116 |
| }, |
| { |
| "epoch": 0.3492537313432836, |
| "grad_norm": 2.042368536282115, |
| "learning_rate": 7.5834968557593155e-06, |
| "loss": 0.8852, |
| "step": 117 |
| }, |
| { |
| "epoch": 0.3522388059701492, |
| "grad_norm": 1.9585431439709604, |
| "learning_rate": 7.541867917259278e-06, |
| "loss": 0.9395, |
| "step": 118 |
| }, |
| { |
| "epoch": 0.35522388059701493, |
| "grad_norm": 1.9690363236772885, |
| "learning_rate": 7.500000000000001e-06, |
| "loss": 0.7769, |
| "step": 119 |
| }, |
| { |
| "epoch": 0.3582089552238806, |
| "grad_norm": 1.9997251291932896, |
| "learning_rate": 7.457897040276853e-06, |
| "loss": 0.8867, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.3611940298507463, |
| "grad_norm": 1.9091614861364634, |
| "learning_rate": 7.415562996483193e-06, |
| "loss": 0.8203, |
| "step": 121 |
| }, |
| { |
| "epoch": 0.3641791044776119, |
| "grad_norm": 2.0154748461484964, |
| "learning_rate": 7.373001848738203e-06, |
| "loss": 0.907, |
| "step": 122 |
| }, |
| { |
| "epoch": 0.36716417910447763, |
| "grad_norm": 1.8850762145928235, |
| "learning_rate": 7.330217598512696e-06, |
| "loss": 0.7268, |
| "step": 123 |
| }, |
| { |
| "epoch": 0.3701492537313433, |
| "grad_norm": 2.002378822284434, |
| "learning_rate": 7.2872142682529045e-06, |
| "loss": 0.75, |
| "step": 124 |
| }, |
| { |
| "epoch": 0.373134328358209, |
| "grad_norm": 1.8683346372861278, |
| "learning_rate": 7.243995901002312e-06, |
| "loss": 0.7998, |
| "step": 125 |
| }, |
| { |
| "epoch": 0.3761194029850746, |
| "grad_norm": 1.8603961929214259, |
| "learning_rate": 7.200566560021525e-06, |
| "loss": 0.7736, |
| "step": 126 |
| }, |
| { |
| "epoch": 0.37910447761194027, |
| "grad_norm": 2.086337061985904, |
| "learning_rate": 7.156930328406268e-06, |
| "loss": 0.8291, |
| "step": 127 |
| }, |
| { |
| "epoch": 0.382089552238806, |
| "grad_norm": 2.0488227138684203, |
| "learning_rate": 7.113091308703498e-06, |
| "loss": 0.8032, |
| "step": 128 |
| }, |
| { |
| "epoch": 0.3850746268656716, |
| "grad_norm": 2.2156317114896504, |
| "learning_rate": 7.069053622525697e-06, |
| "loss": 0.9299, |
| "step": 129 |
| }, |
| { |
| "epoch": 0.3880597014925373, |
| "grad_norm": 1.959090575118802, |
| "learning_rate": 7.0248214101633685e-06, |
| "loss": 0.7743, |
| "step": 130 |
| }, |
| { |
| "epoch": 0.39104477611940297, |
| "grad_norm": 2.040790454179876, |
| "learning_rate": 6.980398830195785e-06, |
| "loss": 0.9517, |
| "step": 131 |
| }, |
| { |
| "epoch": 0.3940298507462687, |
| "grad_norm": 2.0095877017198394, |
| "learning_rate": 6.9357900591000034e-06, |
| "loss": 0.8367, |
| "step": 132 |
| }, |
| { |
| "epoch": 0.3970149253731343, |
| "grad_norm": 2.053685963143701, |
| "learning_rate": 6.890999290858213e-06, |
| "loss": 0.908, |
| "step": 133 |
| }, |
| { |
| "epoch": 0.4, |
| "grad_norm": 1.9552949066858203, |
| "learning_rate": 6.8460307365634225e-06, |
| "loss": 0.7769, |
| "step": 134 |
| }, |
| { |
| "epoch": 0.40298507462686567, |
| "grad_norm": 1.9422682262777784, |
| "learning_rate": 6.800888624023552e-06, |
| "loss": 0.8613, |
| "step": 135 |
| }, |
| { |
| "epoch": 0.4059701492537313, |
| "grad_norm": 2.0018278352562784, |
| "learning_rate": 6.755577197363945e-06, |
| "loss": 0.7314, |
| "step": 136 |
| }, |
| { |
| "epoch": 0.408955223880597, |
| "grad_norm": 2.0748265555866023, |
| "learning_rate": 6.710100716628345e-06, |
| "loss": 0.7806, |
| "step": 137 |
| }, |
| { |
| "epoch": 0.41194029850746267, |
| "grad_norm": 1.9674200834017632, |
| "learning_rate": 6.6644634573783825e-06, |
| "loss": 0.8074, |
| "step": 138 |
| }, |
| { |
| "epoch": 0.41492537313432837, |
| "grad_norm": 1.9207512896621706, |
| "learning_rate": 6.618669710291607e-06, |
| "loss": 0.7148, |
| "step": 139 |
| }, |
| { |
| "epoch": 0.417910447761194, |
| "grad_norm": 1.905193379967269, |
| "learning_rate": 6.572723780758069e-06, |
| "loss": 0.8142, |
| "step": 140 |
| }, |
| { |
| "epoch": 0.4208955223880597, |
| "grad_norm": 2.1965158090202697, |
| "learning_rate": 6.526629988475567e-06, |
| "loss": 0.8123, |
| "step": 141 |
| }, |
| { |
| "epoch": 0.42388059701492536, |
| "grad_norm": 1.8753650627785756, |
| "learning_rate": 6.4803926670435e-06, |
| "loss": 0.8388, |
| "step": 142 |
| }, |
| { |
| "epoch": 0.42686567164179107, |
| "grad_norm": 2.0267686855160156, |
| "learning_rate": 6.434016163555452e-06, |
| "loss": 0.8065, |
| "step": 143 |
| }, |
| { |
| "epoch": 0.4298507462686567, |
| "grad_norm": 1.9454127213119508, |
| "learning_rate": 6.387504838190479e-06, |
| "loss": 0.7268, |
| "step": 144 |
| }, |
| { |
| "epoch": 0.43283582089552236, |
| "grad_norm": 1.9502519457009935, |
| "learning_rate": 6.340863063803187e-06, |
| "loss": 0.744, |
| "step": 145 |
| }, |
| { |
| "epoch": 0.43582089552238806, |
| "grad_norm": 1.9053847115703428, |
| "learning_rate": 6.294095225512604e-06, |
| "loss": 0.8448, |
| "step": 146 |
| }, |
| { |
| "epoch": 0.4388059701492537, |
| "grad_norm": 2.099027132432326, |
| "learning_rate": 6.247205720289907e-06, |
| "loss": 0.9298, |
| "step": 147 |
| }, |
| { |
| "epoch": 0.4417910447761194, |
| "grad_norm": 2.0781658570138153, |
| "learning_rate": 6.2001989565450305e-06, |
| "loss": 0.7678, |
| "step": 148 |
| }, |
| { |
| "epoch": 0.44477611940298506, |
| "grad_norm": 2.017544918765642, |
| "learning_rate": 6.153079353712201e-06, |
| "loss": 0.7815, |
| "step": 149 |
| }, |
| { |
| "epoch": 0.44776119402985076, |
| "grad_norm": 2.0613800967577203, |
| "learning_rate": 6.105851341834439e-06, |
| "loss": 0.7604, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.4507462686567164, |
| "grad_norm": 1.8534108525240118, |
| "learning_rate": 6.058519361147055e-06, |
| "loss": 0.7744, |
| "step": 151 |
| }, |
| { |
| "epoch": 0.4537313432835821, |
| "grad_norm": 2.0589449906784525, |
| "learning_rate": 6.011087861660191e-06, |
| "loss": 0.7458, |
| "step": 152 |
| }, |
| { |
| "epoch": 0.45671641791044776, |
| "grad_norm": 1.7521074071886629, |
| "learning_rate": 5.9635613027404495e-06, |
| "loss": 0.7427, |
| "step": 153 |
| }, |
| { |
| "epoch": 0.4597014925373134, |
| "grad_norm": 1.952094332549687, |
| "learning_rate": 5.915944152691634e-06, |
| "loss": 0.7228, |
| "step": 154 |
| }, |
| { |
| "epoch": 0.4626865671641791, |
| "grad_norm": 2.0059139166291162, |
| "learning_rate": 5.8682408883346535e-06, |
| "loss": 0.7888, |
| "step": 155 |
| }, |
| { |
| "epoch": 0.46567164179104475, |
| "grad_norm": 1.9863550415015798, |
| "learning_rate": 5.820455994586621e-06, |
| "loss": 0.7431, |
| "step": 156 |
| }, |
| { |
| "epoch": 0.46865671641791046, |
| "grad_norm": 2.186012143543713, |
| "learning_rate": 5.772593964039203e-06, |
| "loss": 0.9548, |
| "step": 157 |
| }, |
| { |
| "epoch": 0.4716417910447761, |
| "grad_norm": 2.0569327383361524, |
| "learning_rate": 5.724659296536234e-06, |
| "loss": 0.7608, |
| "step": 158 |
| }, |
| { |
| "epoch": 0.4746268656716418, |
| "grad_norm": 1.8427207627927604, |
| "learning_rate": 5.6766564987506564e-06, |
| "loss": 0.728, |
| "step": 159 |
| }, |
| { |
| "epoch": 0.47761194029850745, |
| "grad_norm": 2.0454749932374874, |
| "learning_rate": 5.628590083760815e-06, |
| "loss": 0.7987, |
| "step": 160 |
| }, |
| { |
| "epoch": 0.48059701492537316, |
| "grad_norm": 1.997088234967905, |
| "learning_rate": 5.5804645706261515e-06, |
| "loss": 0.87, |
| "step": 161 |
| }, |
| { |
| "epoch": 0.4835820895522388, |
| "grad_norm": 2.0854705687983897, |
| "learning_rate": 5.532284483962341e-06, |
| "loss": 0.7816, |
| "step": 162 |
| }, |
| { |
| "epoch": 0.48656716417910445, |
| "grad_norm": 2.063156543618824, |
| "learning_rate": 5.484054353515896e-06, |
| "loss": 0.7953, |
| "step": 163 |
| }, |
| { |
| "epoch": 0.48955223880597015, |
| "grad_norm": 2.0601249949949025, |
| "learning_rate": 5.435778713738292e-06, |
| "loss": 0.8433, |
| "step": 164 |
| }, |
| { |
| "epoch": 0.4925373134328358, |
| "grad_norm": 1.9082330488593775, |
| "learning_rate": 5.387462103359655e-06, |
| "loss": 0.78, |
| "step": 165 |
| }, |
| { |
| "epoch": 0.4955223880597015, |
| "grad_norm": 1.861403808976405, |
| "learning_rate": 5.339109064962047e-06, |
| "loss": 0.6939, |
| "step": 166 |
| }, |
| { |
| "epoch": 0.49850746268656715, |
| "grad_norm": 2.1165013908329393, |
| "learning_rate": 5.290724144552379e-06, |
| "loss": 0.6902, |
| "step": 167 |
| }, |
| { |
| "epoch": 0.5014925373134328, |
| "grad_norm": 2.4674070182101544, |
| "learning_rate": 5.242311891135016e-06, |
| "loss": 0.8382, |
| "step": 168 |
| }, |
| { |
| "epoch": 0.5044776119402985, |
| "grad_norm": 1.8675037579070943, |
| "learning_rate": 5.193876856284085e-06, |
| "loss": 0.7193, |
| "step": 169 |
| }, |
| { |
| "epoch": 0.5074626865671642, |
| "grad_norm": 2.0252014042338415, |
| "learning_rate": 5.145423593715558e-06, |
| "loss": 0.8031, |
| "step": 170 |
| }, |
| { |
| "epoch": 0.5104477611940299, |
| "grad_norm": 1.9108898103865828, |
| "learning_rate": 5.096956658859122e-06, |
| "loss": 0.7634, |
| "step": 171 |
| }, |
| { |
| "epoch": 0.5134328358208955, |
| "grad_norm": 2.139726380565598, |
| "learning_rate": 5.048480608429893e-06, |
| "loss": 0.7139, |
| "step": 172 |
| }, |
| { |
| "epoch": 0.5164179104477612, |
| "grad_norm": 1.9120901416594451, |
| "learning_rate": 5e-06, |
| "loss": 0.7729, |
| "step": 173 |
| }, |
| { |
| "epoch": 0.5194029850746269, |
| "grad_norm": 1.766403288334722, |
| "learning_rate": 4.951519391570108e-06, |
| "loss": 0.7792, |
| "step": 174 |
| }, |
| { |
| "epoch": 0.5223880597014925, |
| "grad_norm": 1.875054628847942, |
| "learning_rate": 4.903043341140879e-06, |
| "loss": 0.7026, |
| "step": 175 |
| }, |
| { |
| "epoch": 0.5253731343283582, |
| "grad_norm": 2.0649412907897817, |
| "learning_rate": 4.854576406284443e-06, |
| "loss": 0.6678, |
| "step": 176 |
| }, |
| { |
| "epoch": 0.5283582089552239, |
| "grad_norm": 1.9766676200083877, |
| "learning_rate": 4.806123143715916e-06, |
| "loss": 0.794, |
| "step": 177 |
| }, |
| { |
| "epoch": 0.5313432835820896, |
| "grad_norm": 1.9628257930544168, |
| "learning_rate": 4.7576881088649865e-06, |
| "loss": 0.8011, |
| "step": 178 |
| }, |
| { |
| "epoch": 0.5343283582089552, |
| "grad_norm": 1.993236983375051, |
| "learning_rate": 4.7092758554476215e-06, |
| "loss": 0.7871, |
| "step": 179 |
| }, |
| { |
| "epoch": 0.5373134328358209, |
| "grad_norm": 1.8504281412049604, |
| "learning_rate": 4.660890935037954e-06, |
| "loss": 0.7736, |
| "step": 180 |
| }, |
| { |
| "epoch": 0.5402985074626866, |
| "grad_norm": 1.9028018031804006, |
| "learning_rate": 4.6125378966403465e-06, |
| "loss": 0.7578, |
| "step": 181 |
| }, |
| { |
| "epoch": 0.5432835820895522, |
| "grad_norm": 1.8659479024734795, |
| "learning_rate": 4.564221286261709e-06, |
| "loss": 0.7557, |
| "step": 182 |
| }, |
| { |
| "epoch": 0.5462686567164179, |
| "grad_norm": 2.11712649211751, |
| "learning_rate": 4.515945646484105e-06, |
| "loss": 0.8088, |
| "step": 183 |
| }, |
| { |
| "epoch": 0.5492537313432836, |
| "grad_norm": 1.789112706708441, |
| "learning_rate": 4.467715516037659e-06, |
| "loss": 0.8087, |
| "step": 184 |
| }, |
| { |
| "epoch": 0.5522388059701493, |
| "grad_norm": 1.9892189797387887, |
| "learning_rate": 4.4195354293738484e-06, |
| "loss": 0.8395, |
| "step": 185 |
| }, |
| { |
| "epoch": 0.5552238805970149, |
| "grad_norm": 2.1883581521829316, |
| "learning_rate": 4.371409916239188e-06, |
| "loss": 0.7684, |
| "step": 186 |
| }, |
| { |
| "epoch": 0.5582089552238806, |
| "grad_norm": 2.1631251071113176, |
| "learning_rate": 4.323343501249346e-06, |
| "loss": 0.736, |
| "step": 187 |
| }, |
| { |
| "epoch": 0.5611940298507463, |
| "grad_norm": 1.9289233899504126, |
| "learning_rate": 4.275340703463767e-06, |
| "loss": 0.6405, |
| "step": 188 |
| }, |
| { |
| "epoch": 0.564179104477612, |
| "grad_norm": 1.9857433587359077, |
| "learning_rate": 4.227406035960798e-06, |
| "loss": 0.7959, |
| "step": 189 |
| }, |
| { |
| "epoch": 0.5671641791044776, |
| "grad_norm": 1.8258458953238406, |
| "learning_rate": 4.17954400541338e-06, |
| "loss": 0.7396, |
| "step": 190 |
| }, |
| { |
| "epoch": 0.5701492537313433, |
| "grad_norm": 1.9073211443331697, |
| "learning_rate": 4.131759111665349e-06, |
| "loss": 0.7894, |
| "step": 191 |
| }, |
| { |
| "epoch": 0.573134328358209, |
| "grad_norm": 2.1144318330609955, |
| "learning_rate": 4.084055847308367e-06, |
| "loss": 0.7564, |
| "step": 192 |
| }, |
| { |
| "epoch": 0.5761194029850746, |
| "grad_norm": 1.9704301567554263, |
| "learning_rate": 4.036438697259551e-06, |
| "loss": 0.768, |
| "step": 193 |
| }, |
| { |
| "epoch": 0.5791044776119403, |
| "grad_norm": 2.641711220430738, |
| "learning_rate": 3.988912138339812e-06, |
| "loss": 0.7862, |
| "step": 194 |
| }, |
| { |
| "epoch": 0.582089552238806, |
| "grad_norm": 2.1626033614930247, |
| "learning_rate": 3.941480638852948e-06, |
| "loss": 0.7856, |
| "step": 195 |
| }, |
| { |
| "epoch": 0.5850746268656717, |
| "grad_norm": 1.9284923265396965, |
| "learning_rate": 3.894148658165562e-06, |
| "loss": 0.7674, |
| "step": 196 |
| }, |
| { |
| "epoch": 0.5880597014925373, |
| "grad_norm": 2.058823111277626, |
| "learning_rate": 3.8469206462878e-06, |
| "loss": 0.7663, |
| "step": 197 |
| }, |
| { |
| "epoch": 0.591044776119403, |
| "grad_norm": 2.1351322093158744, |
| "learning_rate": 3.7998010434549716e-06, |
| "loss": 0.8399, |
| "step": 198 |
| }, |
| { |
| "epoch": 0.5940298507462687, |
| "grad_norm": 1.9380000299821818, |
| "learning_rate": 3.752794279710094e-06, |
| "loss": 0.7321, |
| "step": 199 |
| }, |
| { |
| "epoch": 0.5970149253731343, |
| "grad_norm": 1.8441505158750409, |
| "learning_rate": 3.705904774487396e-06, |
| "loss": 0.774, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.6, |
| "grad_norm": 2.0907330198564313, |
| "learning_rate": 3.6591369361968127e-06, |
| "loss": 0.7847, |
| "step": 201 |
| }, |
| { |
| "epoch": 0.6029850746268657, |
| "grad_norm": 1.9371066770940577, |
| "learning_rate": 3.6124951618095224e-06, |
| "loss": 0.748, |
| "step": 202 |
| }, |
| { |
| "epoch": 0.6059701492537314, |
| "grad_norm": 1.8374945672921557, |
| "learning_rate": 3.5659838364445505e-06, |
| "loss": 0.7537, |
| "step": 203 |
| }, |
| { |
| "epoch": 0.608955223880597, |
| "grad_norm": 2.081262815281359, |
| "learning_rate": 3.519607332956502e-06, |
| "loss": 0.815, |
| "step": 204 |
| }, |
| { |
| "epoch": 0.6119402985074627, |
| "grad_norm": 2.074788649262812, |
| "learning_rate": 3.473370011524435e-06, |
| "loss": 0.8674, |
| "step": 205 |
| }, |
| { |
| "epoch": 0.6149253731343284, |
| "grad_norm": 1.8698000805877122, |
| "learning_rate": 3.427276219241933e-06, |
| "loss": 0.7202, |
| "step": 206 |
| }, |
| { |
| "epoch": 0.6179104477611941, |
| "grad_norm": 1.9048492100127103, |
| "learning_rate": 3.3813302897083955e-06, |
| "loss": 0.8188, |
| "step": 207 |
| }, |
| { |
| "epoch": 0.6208955223880597, |
| "grad_norm": 1.979184398237071, |
| "learning_rate": 3.335536542621617e-06, |
| "loss": 0.7503, |
| "step": 208 |
| }, |
| { |
| "epoch": 0.6238805970149254, |
| "grad_norm": 1.9760275557890097, |
| "learning_rate": 3.289899283371657e-06, |
| "loss": 0.7997, |
| "step": 209 |
| }, |
| { |
| "epoch": 0.6268656716417911, |
| "grad_norm": 2.1124769220840944, |
| "learning_rate": 3.244422802636057e-06, |
| "loss": 0.831, |
| "step": 210 |
| }, |
| { |
| "epoch": 0.6298507462686567, |
| "grad_norm": 2.034333485591399, |
| "learning_rate": 3.1991113759764493e-06, |
| "loss": 0.7732, |
| "step": 211 |
| }, |
| { |
| "epoch": 0.6328358208955224, |
| "grad_norm": 1.897048944733691, |
| "learning_rate": 3.1539692634365788e-06, |
| "loss": 0.754, |
| "step": 212 |
| }, |
| { |
| "epoch": 0.6358208955223881, |
| "grad_norm": 1.9618035338024724, |
| "learning_rate": 3.1090007091417884e-06, |
| "loss": 0.6767, |
| "step": 213 |
| }, |
| { |
| "epoch": 0.6388059701492538, |
| "grad_norm": 2.0148083710716764, |
| "learning_rate": 3.0642099408999982e-06, |
| "loss": 0.8271, |
| "step": 214 |
| }, |
| { |
| "epoch": 0.6417910447761194, |
| "grad_norm": 2.283043013050116, |
| "learning_rate": 3.019601169804216e-06, |
| "loss": 0.7971, |
| "step": 215 |
| }, |
| { |
| "epoch": 0.6447761194029851, |
| "grad_norm": 2.11854512509762, |
| "learning_rate": 2.975178589836632e-06, |
| "loss": 0.7609, |
| "step": 216 |
| }, |
| { |
| "epoch": 0.6477611940298508, |
| "grad_norm": 1.8298105473763824, |
| "learning_rate": 2.9309463774743047e-06, |
| "loss": 0.8506, |
| "step": 217 |
| }, |
| { |
| "epoch": 0.6507462686567164, |
| "grad_norm": 1.7780813510143145, |
| "learning_rate": 2.886908691296504e-06, |
| "loss": 0.7632, |
| "step": 218 |
| }, |
| { |
| "epoch": 0.6537313432835821, |
| "grad_norm": 1.9126798345749187, |
| "learning_rate": 2.843069671593734e-06, |
| "loss": 0.7782, |
| "step": 219 |
| }, |
| { |
| "epoch": 0.6567164179104478, |
| "grad_norm": 1.9350084312265137, |
| "learning_rate": 2.7994334399784773e-06, |
| "loss": 0.7372, |
| "step": 220 |
| }, |
| { |
| "epoch": 0.6597014925373135, |
| "grad_norm": 1.9251939137126606, |
| "learning_rate": 2.7560040989976894e-06, |
| "loss": 0.7357, |
| "step": 221 |
| }, |
| { |
| "epoch": 0.6626865671641791, |
| "grad_norm": 1.9256071124163803, |
| "learning_rate": 2.7127857317470967e-06, |
| "loss": 0.7855, |
| "step": 222 |
| }, |
| { |
| "epoch": 0.6656716417910448, |
| "grad_norm": 2.0598971965285338, |
| "learning_rate": 2.6697824014873076e-06, |
| "loss": 0.7324, |
| "step": 223 |
| }, |
| { |
| "epoch": 0.6686567164179105, |
| "grad_norm": 1.823285258131258, |
| "learning_rate": 2.626998151261798e-06, |
| "loss": 0.6943, |
| "step": 224 |
| }, |
| { |
| "epoch": 0.6716417910447762, |
| "grad_norm": 2.7227864446392678, |
| "learning_rate": 2.5844370035168077e-06, |
| "loss": 0.7098, |
| "step": 225 |
| }, |
| { |
| "epoch": 0.6746268656716418, |
| "grad_norm": 2.1471535886447035, |
| "learning_rate": 2.5421029597231476e-06, |
| "loss": 0.7776, |
| "step": 226 |
| }, |
| { |
| "epoch": 0.6776119402985075, |
| "grad_norm": 2.429389297486461, |
| "learning_rate": 2.5000000000000015e-06, |
| "loss": 0.7514, |
| "step": 227 |
| }, |
| { |
| "epoch": 0.6805970149253732, |
| "grad_norm": 2.0726794837198725, |
| "learning_rate": 2.458132082740724e-06, |
| "loss": 0.7712, |
| "step": 228 |
| }, |
| { |
| "epoch": 0.6835820895522388, |
| "grad_norm": 2.1998151305942395, |
| "learning_rate": 2.4165031442406857e-06, |
| "loss": 0.7435, |
| "step": 229 |
| }, |
| { |
| "epoch": 0.6865671641791045, |
| "grad_norm": 2.02867494365551, |
| "learning_rate": 2.3751170983272e-06, |
| "loss": 0.8134, |
| "step": 230 |
| }, |
| { |
| "epoch": 0.6895522388059702, |
| "grad_norm": 1.9535756316081252, |
| "learning_rate": 2.333977835991545e-06, |
| "loss": 0.7191, |
| "step": 231 |
| }, |
| { |
| "epoch": 0.6925373134328359, |
| "grad_norm": 2.0484292921284477, |
| "learning_rate": 2.293089225023152e-06, |
| "loss": 0.8623, |
| "step": 232 |
| }, |
| { |
| "epoch": 0.6955223880597015, |
| "grad_norm": 2.0146733780323505, |
| "learning_rate": 2.2524551096459703e-06, |
| "loss": 0.7654, |
| "step": 233 |
| }, |
| { |
| "epoch": 0.6985074626865672, |
| "grad_norm": 2.007942017411499, |
| "learning_rate": 2.2120793101570366e-06, |
| "loss": 0.7197, |
| "step": 234 |
| }, |
| { |
| "epoch": 0.7014925373134329, |
| "grad_norm": 1.9512262040861592, |
| "learning_rate": 2.171965622567308e-06, |
| "loss": 0.795, |
| "step": 235 |
| }, |
| { |
| "epoch": 0.7044776119402985, |
| "grad_norm": 2.0223484864555608, |
| "learning_rate": 2.132117818244771e-06, |
| "loss": 0.73, |
| "step": 236 |
| }, |
| { |
| "epoch": 0.7074626865671642, |
| "grad_norm": 1.9076970970748008, |
| "learning_rate": 2.0925396435598665e-06, |
| "loss": 0.7341, |
| "step": 237 |
| }, |
| { |
| "epoch": 0.7104477611940299, |
| "grad_norm": 2.0099705479490084, |
| "learning_rate": 2.053234819533276e-06, |
| "loss": 0.706, |
| "step": 238 |
| }, |
| { |
| "epoch": 0.7134328358208956, |
| "grad_norm": 2.088602599462484, |
| "learning_rate": 2.0142070414860704e-06, |
| "loss": 0.8061, |
| "step": 239 |
| }, |
| { |
| "epoch": 0.7164179104477612, |
| "grad_norm": 2.1414530746534437, |
| "learning_rate": 1.9754599786922913e-06, |
| "loss": 0.7735, |
| "step": 240 |
| }, |
| { |
| "epoch": 0.7194029850746269, |
| "grad_norm": 1.756460059490162, |
| "learning_rate": 1.936997274033986e-06, |
| "loss": 0.652, |
| "step": 241 |
| }, |
| { |
| "epoch": 0.7223880597014926, |
| "grad_norm": 1.6835259072072402, |
| "learning_rate": 1.8988225436587005e-06, |
| "loss": 0.6575, |
| "step": 242 |
| }, |
| { |
| "epoch": 0.7253731343283583, |
| "grad_norm": 1.9338570463688378, |
| "learning_rate": 1.8609393766395083e-06, |
| "loss": 0.8504, |
| "step": 243 |
| }, |
| { |
| "epoch": 0.7283582089552239, |
| "grad_norm": 2.0279105685865884, |
| "learning_rate": 1.823351334637576e-06, |
| "loss": 0.8141, |
| "step": 244 |
| }, |
| { |
| "epoch": 0.7313432835820896, |
| "grad_norm": 1.9475282359983863, |
| "learning_rate": 1.7860619515673034e-06, |
| "loss": 0.7989, |
| "step": 245 |
| }, |
| { |
| "epoch": 0.7343283582089553, |
| "grad_norm": 2.0357960960016093, |
| "learning_rate": 1.7490747332640833e-06, |
| "loss": 0.8093, |
| "step": 246 |
| }, |
| { |
| "epoch": 0.7373134328358208, |
| "grad_norm": 2.0420189686519072, |
| "learning_rate": 1.7123931571546826e-06, |
| "loss": 0.8056, |
| "step": 247 |
| }, |
| { |
| "epoch": 0.7402985074626866, |
| "grad_norm": 1.7925260574915485, |
| "learning_rate": 1.6760206719303107e-06, |
| "loss": 0.7222, |
| "step": 248 |
| }, |
| { |
| "epoch": 0.7432835820895523, |
| "grad_norm": 1.9144780544797566, |
| "learning_rate": 1.639960697222388e-06, |
| "loss": 0.8697, |
| "step": 249 |
| }, |
| { |
| "epoch": 0.746268656716418, |
| "grad_norm": 1.9085137418012865, |
| "learning_rate": 1.6042166232810346e-06, |
| "loss": 0.7573, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.7492537313432835, |
| "grad_norm": 1.7594094715163107, |
| "learning_rate": 1.5687918106563326e-06, |
| "loss": 0.719, |
| "step": 251 |
| }, |
| { |
| "epoch": 0.7522388059701492, |
| "grad_norm": 1.9721722301749733, |
| "learning_rate": 1.5336895898823801e-06, |
| "loss": 0.805, |
| "step": 252 |
| }, |
| { |
| "epoch": 0.755223880597015, |
| "grad_norm": 2.2257145387956148, |
| "learning_rate": 1.4989132611641576e-06, |
| "loss": 0.7464, |
| "step": 253 |
| }, |
| { |
| "epoch": 0.7582089552238805, |
| "grad_norm": 2.2389879292796944, |
| "learning_rate": 1.4644660940672628e-06, |
| "loss": 0.7688, |
| "step": 254 |
| }, |
| { |
| "epoch": 0.7611940298507462, |
| "grad_norm": 2.2600400159677445, |
| "learning_rate": 1.4303513272105057e-06, |
| "loss": 0.6961, |
| "step": 255 |
| }, |
| { |
| "epoch": 0.764179104477612, |
| "grad_norm": 1.9828908431355627, |
| "learning_rate": 1.396572167961427e-06, |
| "loss": 0.8139, |
| "step": 256 |
| }, |
| { |
| "epoch": 0.7671641791044777, |
| "grad_norm": 2.123631920520936, |
| "learning_rate": 1.3631317921347564e-06, |
| "loss": 0.7817, |
| "step": 257 |
| }, |
| { |
| "epoch": 0.7701492537313432, |
| "grad_norm": 1.91427428961712, |
| "learning_rate": 1.330033343693824e-06, |
| "loss": 0.7411, |
| "step": 258 |
| }, |
| { |
| "epoch": 0.7731343283582089, |
| "grad_norm": 1.923869516924642, |
| "learning_rate": 1.297279934454978e-06, |
| "loss": 0.7464, |
| "step": 259 |
| }, |
| { |
| "epoch": 0.7761194029850746, |
| "grad_norm": 2.097695064363166, |
| "learning_rate": 1.264874643795021e-06, |
| "loss": 0.7835, |
| "step": 260 |
| }, |
| { |
| "epoch": 0.7791044776119403, |
| "grad_norm": 1.866847274620371, |
| "learning_rate": 1.2328205183616964e-06, |
| "loss": 0.7576, |
| "step": 261 |
| }, |
| { |
| "epoch": 0.7820895522388059, |
| "grad_norm": 1.9518448868919547, |
| "learning_rate": 1.2011205717872538e-06, |
| "loss": 0.8182, |
| "step": 262 |
| }, |
| { |
| "epoch": 0.7850746268656716, |
| "grad_norm": 2.0133337408527483, |
| "learning_rate": 1.1697777844051105e-06, |
| "loss": 0.7536, |
| "step": 263 |
| }, |
| { |
| "epoch": 0.7880597014925373, |
| "grad_norm": 1.9817193200479692, |
| "learning_rate": 1.1387951029696543e-06, |
| "loss": 0.7337, |
| "step": 264 |
| }, |
| { |
| "epoch": 0.7910447761194029, |
| "grad_norm": 1.9081666878508032, |
| "learning_rate": 1.1081754403792e-06, |
| "loss": 0.7247, |
| "step": 265 |
| }, |
| { |
| "epoch": 0.7940298507462686, |
| "grad_norm": 1.7424632653916463, |
| "learning_rate": 1.0779216754021215e-06, |
| "loss": 0.7659, |
| "step": 266 |
| }, |
| { |
| "epoch": 0.7970149253731343, |
| "grad_norm": 2.1707469563008694, |
| "learning_rate": 1.0480366524062041e-06, |
| "loss": 0.7172, |
| "step": 267 |
| }, |
| { |
| "epoch": 0.8, |
| "grad_norm": 1.9358665287041594, |
| "learning_rate": 1.0185231810912223e-06, |
| "loss": 0.7672, |
| "step": 268 |
| }, |
| { |
| "epoch": 0.8029850746268656, |
| "grad_norm": 1.9633165798810948, |
| "learning_rate": 9.893840362247809e-07, |
| "loss": 0.7251, |
| "step": 269 |
| }, |
| { |
| "epoch": 0.8059701492537313, |
| "grad_norm": 1.811709330190077, |
| "learning_rate": 9.606219573814447e-07, |
| "loss": 0.7109, |
| "step": 270 |
| }, |
| { |
| "epoch": 0.808955223880597, |
| "grad_norm": 2.113644270821526, |
| "learning_rate": 9.322396486851626e-07, |
| "loss": 0.7905, |
| "step": 271 |
| }, |
| { |
| "epoch": 0.8119402985074626, |
| "grad_norm": 1.827142174127426, |
| "learning_rate": 9.042397785550405e-07, |
| "loss": 0.7433, |
| "step": 272 |
| }, |
| { |
| "epoch": 0.8149253731343283, |
| "grad_norm": 2.014601055555163, |
| "learning_rate": 8.766249794544662e-07, |
| "loss": 0.7755, |
| "step": 273 |
| }, |
| { |
| "epoch": 0.817910447761194, |
| "grad_norm": 2.091449213566336, |
| "learning_rate": 8.49397847643606e-07, |
| "loss": 0.8266, |
| "step": 274 |
| }, |
| { |
| "epoch": 0.8208955223880597, |
| "grad_norm": 2.3792124584042407, |
| "learning_rate": 8.225609429353187e-07, |
| "loss": 0.8152, |
| "step": 275 |
| }, |
| { |
| "epoch": 0.8238805970149253, |
| "grad_norm": 2.236371795361171, |
| "learning_rate": 7.961167884544852e-07, |
| "loss": 0.8004, |
| "step": 276 |
| }, |
| { |
| "epoch": 0.826865671641791, |
| "grad_norm": 1.8675998248443555, |
| "learning_rate": 7.700678704007947e-07, |
| "loss": 0.7616, |
| "step": 277 |
| }, |
| { |
| "epoch": 0.8298507462686567, |
| "grad_norm": 1.8934578283049441, |
| "learning_rate": 7.444166378150014e-07, |
| "loss": 0.7384, |
| "step": 278 |
| }, |
| { |
| "epoch": 0.8328358208955224, |
| "grad_norm": 1.806959761192126, |
| "learning_rate": 7.191655023486682e-07, |
| "loss": 0.7859, |
| "step": 279 |
| }, |
| { |
| "epoch": 0.835820895522388, |
| "grad_norm": 2.0920960597937883, |
| "learning_rate": 6.94316838037431e-07, |
| "loss": 0.6582, |
| "step": 280 |
| }, |
| { |
| "epoch": 0.8388059701492537, |
| "grad_norm": 1.9049893572652696, |
| "learning_rate": 6.698729810778065e-07, |
| "loss": 0.7133, |
| "step": 281 |
| }, |
| { |
| "epoch": 0.8417910447761194, |
| "grad_norm": 2.021598834791704, |
| "learning_rate": 6.458362296075399e-07, |
| "loss": 0.6986, |
| "step": 282 |
| }, |
| { |
| "epoch": 0.844776119402985, |
| "grad_norm": 1.8785681628556798, |
| "learning_rate": 6.222088434895462e-07, |
| "loss": 0.7561, |
| "step": 283 |
| }, |
| { |
| "epoch": 0.8477611940298507, |
| "grad_norm": 1.7981294150989415, |
| "learning_rate": 5.989930440994451e-07, |
| "loss": 0.6988, |
| "step": 284 |
| }, |
| { |
| "epoch": 0.8507462686567164, |
| "grad_norm": 1.9804075333202549, |
| "learning_rate": 5.76191014116711e-07, |
| "loss": 0.7793, |
| "step": 285 |
| }, |
| { |
| "epoch": 0.8537313432835821, |
| "grad_norm": 1.8946196230697367, |
| "learning_rate": 5.538048973194699e-07, |
| "loss": 0.7378, |
| "step": 286 |
| }, |
| { |
| "epoch": 0.8567164179104477, |
| "grad_norm": 4.308097949167914, |
| "learning_rate": 5.318367983829393e-07, |
| "loss": 0.7545, |
| "step": 287 |
| }, |
| { |
| "epoch": 0.8597014925373134, |
| "grad_norm": 1.7900383798917394, |
| "learning_rate": 5.102887826815589e-07, |
| "loss": 0.6725, |
| "step": 288 |
| }, |
| { |
| "epoch": 0.8626865671641791, |
| "grad_norm": 1.8973145322914773, |
| "learning_rate": 4.891628760948114e-07, |
| "loss": 0.6943, |
| "step": 289 |
| }, |
| { |
| "epoch": 0.8656716417910447, |
| "grad_norm": 2.0037685592372516, |
| "learning_rate": 4.6846106481675035e-07, |
| "loss": 0.7904, |
| "step": 290 |
| }, |
| { |
| "epoch": 0.8686567164179104, |
| "grad_norm": 1.833347945444079, |
| "learning_rate": 4.481852951692672e-07, |
| "loss": 0.6899, |
| "step": 291 |
| }, |
| { |
| "epoch": 0.8716417910447761, |
| "grad_norm": 2.016728151056703, |
| "learning_rate": 4.283374734191037e-07, |
| "loss": 0.8379, |
| "step": 292 |
| }, |
| { |
| "epoch": 0.8746268656716418, |
| "grad_norm": 2.034748614996797, |
| "learning_rate": 4.089194655986306e-07, |
| "loss": 0.6884, |
| "step": 293 |
| }, |
| { |
| "epoch": 0.8776119402985074, |
| "grad_norm": 2.103995265841144, |
| "learning_rate": 3.899330973304083e-07, |
| "loss": 0.8119, |
| "step": 294 |
| }, |
| { |
| "epoch": 0.8805970149253731, |
| "grad_norm": 2.085775862013919, |
| "learning_rate": 3.7138015365554834e-07, |
| "loss": 0.7218, |
| "step": 295 |
| }, |
| { |
| "epoch": 0.8835820895522388, |
| "grad_norm": 1.859225098794404, |
| "learning_rate": 3.5326237886588734e-07, |
| "loss": 0.761, |
| "step": 296 |
| }, |
| { |
| "epoch": 0.8865671641791045, |
| "grad_norm": 1.9937848736813224, |
| "learning_rate": 3.355814763399973e-07, |
| "loss": 0.7734, |
| "step": 297 |
| }, |
| { |
| "epoch": 0.8895522388059701, |
| "grad_norm": 1.956060370495559, |
| "learning_rate": 3.183391083830345e-07, |
| "loss": 0.7522, |
| "step": 298 |
| }, |
| { |
| "epoch": 0.8925373134328358, |
| "grad_norm": 2.1663087142629953, |
| "learning_rate": 3.015368960704584e-07, |
| "loss": 0.7592, |
| "step": 299 |
| }, |
| { |
| "epoch": 0.8955223880597015, |
| "grad_norm": 1.8473510328793137, |
| "learning_rate": 2.8517641909562075e-07, |
| "loss": 0.7569, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.8985074626865671, |
| "grad_norm": 1.9818158957078058, |
| "learning_rate": 2.6925921562124867e-07, |
| "loss": 0.637, |
| "step": 301 |
| }, |
| { |
| "epoch": 0.9014925373134328, |
| "grad_norm": 1.9920228777916036, |
| "learning_rate": 2.5378678213483057e-07, |
| "loss": 0.766, |
| "step": 302 |
| }, |
| { |
| "epoch": 0.9044776119402985, |
| "grad_norm": 1.960285579935006, |
| "learning_rate": 2.3876057330792344e-07, |
| "loss": 0.7499, |
| "step": 303 |
| }, |
| { |
| "epoch": 0.9074626865671642, |
| "grad_norm": 1.94458918165242, |
| "learning_rate": 2.2418200185938488e-07, |
| "loss": 0.7358, |
| "step": 304 |
| }, |
| { |
| "epoch": 0.9104477611940298, |
| "grad_norm": 2.081159928237178, |
| "learning_rate": 2.1005243842255552e-07, |
| "loss": 0.882, |
| "step": 305 |
| }, |
| { |
| "epoch": 0.9134328358208955, |
| "grad_norm": 1.9550064790232975, |
| "learning_rate": 1.9637321141639743e-07, |
| "loss": 0.6894, |
| "step": 306 |
| }, |
| { |
| "epoch": 0.9164179104477612, |
| "grad_norm": 2.1049771274583486, |
| "learning_rate": 1.8314560692059836e-07, |
| "loss": 0.7991, |
| "step": 307 |
| }, |
| { |
| "epoch": 0.9194029850746268, |
| "grad_norm": 2.0857228086478448, |
| "learning_rate": 1.7037086855465902e-07, |
| "loss": 0.7768, |
| "step": 308 |
| }, |
| { |
| "epoch": 0.9223880597014925, |
| "grad_norm": 1.935207733713808, |
| "learning_rate": 1.5805019736097105e-07, |
| "loss": 0.7564, |
| "step": 309 |
| }, |
| { |
| "epoch": 0.9253731343283582, |
| "grad_norm": 2.072194118865995, |
| "learning_rate": 1.4618475169190017e-07, |
| "loss": 0.7604, |
| "step": 310 |
| }, |
| { |
| "epoch": 0.9283582089552239, |
| "grad_norm": 2.1552417858238493, |
| "learning_rate": 1.3477564710088097e-07, |
| "loss": 0.793, |
| "step": 311 |
| }, |
| { |
| "epoch": 0.9313432835820895, |
| "grad_norm": 1.951097534718573, |
| "learning_rate": 1.2382395623753484e-07, |
| "loss": 0.7764, |
| "step": 312 |
| }, |
| { |
| "epoch": 0.9343283582089552, |
| "grad_norm": 1.9504459470286735, |
| "learning_rate": 1.1333070874682217e-07, |
| "loss": 0.7777, |
| "step": 313 |
| }, |
| { |
| "epoch": 0.9373134328358209, |
| "grad_norm": 2.175522821443901, |
| "learning_rate": 1.0329689117224262e-07, |
| "loss": 0.8003, |
| "step": 314 |
| }, |
| { |
| "epoch": 0.9402985074626866, |
| "grad_norm": 1.896018319659114, |
| "learning_rate": 9.372344686307655e-08, |
| "loss": 0.6819, |
| "step": 315 |
| }, |
| { |
| "epoch": 0.9432835820895522, |
| "grad_norm": 1.8626576145737677, |
| "learning_rate": 8.461127588570039e-08, |
| "loss": 0.7407, |
| "step": 316 |
| }, |
| { |
| "epoch": 0.9462686567164179, |
| "grad_norm": 2.255720786589443, |
| "learning_rate": 7.59612349389599e-08, |
| "loss": 0.8032, |
| "step": 317 |
| }, |
| { |
| "epoch": 0.9492537313432836, |
| "grad_norm": 1.988888031528982, |
| "learning_rate": 6.777413727363069e-08, |
| "loss": 0.6934, |
| "step": 318 |
| }, |
| { |
| "epoch": 0.9522388059701492, |
| "grad_norm": 1.936424618352781, |
| "learning_rate": 6.005075261595495e-08, |
| "loss": 0.7447, |
| "step": 319 |
| }, |
| { |
| "epoch": 0.9552238805970149, |
| "grad_norm": 2.166079975131979, |
| "learning_rate": 5.279180709527765e-08, |
| "loss": 0.6995, |
| "step": 320 |
| }, |
| { |
| "epoch": 0.9582089552238806, |
| "grad_norm": 1.910155391887948, |
| "learning_rate": 4.599798317577342e-08, |
| "loss": 0.7723, |
| "step": 321 |
| }, |
| { |
| "epoch": 0.9611940298507463, |
| "grad_norm": 2.015438544350679, |
| "learning_rate": 3.9669919592288385e-08, |
| "loss": 0.7478, |
| "step": 322 |
| }, |
| { |
| "epoch": 0.9641791044776119, |
| "grad_norm": 1.7894327871602922, |
| "learning_rate": 3.3808211290284886e-08, |
| "loss": 0.6971, |
| "step": 323 |
| }, |
| { |
| "epoch": 0.9671641791044776, |
| "grad_norm": 2.243535756266807, |
| "learning_rate": 2.8413409369907887e-08, |
| "loss": 0.8014, |
| "step": 324 |
| }, |
| { |
| "epoch": 0.9701492537313433, |
| "grad_norm": 2.104094911680065, |
| "learning_rate": 2.3486021034170857e-08, |
| "loss": 0.7373, |
| "step": 325 |
| }, |
| { |
| "epoch": 0.9731343283582089, |
| "grad_norm": 1.9746672549193203, |
| "learning_rate": 1.9026509541272276e-08, |
| "loss": 0.7462, |
| "step": 326 |
| }, |
| { |
| "epoch": 0.9761194029850746, |
| "grad_norm": 1.8638633602470398, |
| "learning_rate": 1.5035294161039882e-08, |
| "loss": 0.7276, |
| "step": 327 |
| }, |
| { |
| "epoch": 0.9791044776119403, |
| "grad_norm": 1.9221365215727475, |
| "learning_rate": 1.1512750135511674e-08, |
| "loss": 0.6948, |
| "step": 328 |
| }, |
| { |
| "epoch": 0.982089552238806, |
| "grad_norm": 1.9696506075193372, |
| "learning_rate": 8.459208643659122e-09, |
| "loss": 0.8016, |
| "step": 329 |
| }, |
| { |
| "epoch": 0.9850746268656716, |
| "grad_norm": 1.9260689325158074, |
| "learning_rate": 5.874956770248186e-09, |
| "loss": 0.772, |
| "step": 330 |
| }, |
| { |
| "epoch": 0.9880597014925373, |
| "grad_norm": 1.9022494846934377, |
| "learning_rate": 3.760237478849793e-09, |
| "loss": 0.639, |
| "step": 331 |
| }, |
| { |
| "epoch": 0.991044776119403, |
| "grad_norm": 2.1198036949572523, |
| "learning_rate": 2.1152495889970035e-09, |
| "loss": 0.7524, |
| "step": 332 |
| }, |
| { |
| "epoch": 0.9940298507462687, |
| "grad_norm": 2.520497927670345, |
| "learning_rate": 9.401477574932927e-10, |
| "loss": 0.7804, |
| "step": 333 |
| }, |
| { |
| "epoch": 0.9970149253731343, |
| "grad_norm": 1.9463118318794423, |
| "learning_rate": 2.3504246386918394e-10, |
| "loss": 0.6975, |
| "step": 334 |
| }, |
| { |
| "epoch": 1.0, |
| "grad_norm": 2.0407770920954924, |
| "learning_rate": 0.0, |
| "loss": 0.7463, |
| "step": 335 |
| }, |
| { |
| "epoch": 1.0, |
| "step": 335, |
| "total_flos": 27958090317824.0, |
| "train_loss": 0.8481638431549072, |
| "train_runtime": 15355.884, |
| "train_samples_per_second": 0.175, |
| "train_steps_per_second": 0.022 |
| } |
| ], |
| "logging_steps": 1.0, |
| "max_steps": 335, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 1, |
| "save_steps": 2060, |
| "total_flos": 27958090317824.0, |
| "train_batch_size": 1, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|