| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 0.9986976744186047, |
| "eval_steps": 500, |
| "global_step": 671, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.0, |
| "learning_rate": 0.0, |
| "loss": 1.4084, |
| "step": 1 |
| }, |
| { |
| "epoch": 0.0, |
| "learning_rate": 9.523809523809525e-07, |
| "loss": 1.3285, |
| "step": 2 |
| }, |
| { |
| "epoch": 0.0, |
| "learning_rate": 1.904761904761905e-06, |
| "loss": 1.3301, |
| "step": 3 |
| }, |
| { |
| "epoch": 0.01, |
| "learning_rate": 2.8571428571428573e-06, |
| "loss": 1.3048, |
| "step": 4 |
| }, |
| { |
| "epoch": 0.01, |
| "learning_rate": 3.80952380952381e-06, |
| "loss": 1.2844, |
| "step": 5 |
| }, |
| { |
| "epoch": 0.01, |
| "learning_rate": 4.761904761904762e-06, |
| "loss": 1.2776, |
| "step": 6 |
| }, |
| { |
| "epoch": 0.01, |
| "learning_rate": 5.7142857142857145e-06, |
| "loss": 1.2673, |
| "step": 7 |
| }, |
| { |
| "epoch": 0.01, |
| "learning_rate": 6.666666666666667e-06, |
| "loss": 1.2424, |
| "step": 8 |
| }, |
| { |
| "epoch": 0.01, |
| "learning_rate": 7.61904761904762e-06, |
| "loss": 1.2499, |
| "step": 9 |
| }, |
| { |
| "epoch": 0.01, |
| "learning_rate": 8.571428571428571e-06, |
| "loss": 1.2529, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.02, |
| "learning_rate": 9.523809523809525e-06, |
| "loss": 1.279, |
| "step": 11 |
| }, |
| { |
| "epoch": 0.02, |
| "learning_rate": 1.0476190476190477e-05, |
| "loss": 1.271, |
| "step": 12 |
| }, |
| { |
| "epoch": 0.02, |
| "learning_rate": 1.1428571428571429e-05, |
| "loss": 1.1913, |
| "step": 13 |
| }, |
| { |
| "epoch": 0.02, |
| "learning_rate": 1.2380952380952383e-05, |
| "loss": 1.2517, |
| "step": 14 |
| }, |
| { |
| "epoch": 0.02, |
| "learning_rate": 1.3333333333333333e-05, |
| "loss": 1.2676, |
| "step": 15 |
| }, |
| { |
| "epoch": 0.02, |
| "learning_rate": 1.4285714285714287e-05, |
| "loss": 1.2522, |
| "step": 16 |
| }, |
| { |
| "epoch": 0.03, |
| "learning_rate": 1.523809523809524e-05, |
| "loss": 1.2597, |
| "step": 17 |
| }, |
| { |
| "epoch": 0.03, |
| "learning_rate": 1.6190476190476193e-05, |
| "loss": 1.2833, |
| "step": 18 |
| }, |
| { |
| "epoch": 0.03, |
| "learning_rate": 1.7142857142857142e-05, |
| "loss": 1.2838, |
| "step": 19 |
| }, |
| { |
| "epoch": 0.03, |
| "learning_rate": 1.8095238095238097e-05, |
| "loss": 1.2724, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.03, |
| "learning_rate": 1.904761904761905e-05, |
| "loss": 1.2284, |
| "step": 21 |
| }, |
| { |
| "epoch": 0.03, |
| "learning_rate": 1.9999883200175286e-05, |
| "loss": 1.2327, |
| "step": 22 |
| }, |
| { |
| "epoch": 0.03, |
| "learning_rate": 1.999953280342959e-05, |
| "loss": 1.2544, |
| "step": 23 |
| }, |
| { |
| "epoch": 0.04, |
| "learning_rate": 1.9998948817948157e-05, |
| "loss": 1.2326, |
| "step": 24 |
| }, |
| { |
| "epoch": 0.04, |
| "learning_rate": 1.9998131257372878e-05, |
| "loss": 1.2025, |
| "step": 25 |
| }, |
| { |
| "epoch": 0.04, |
| "learning_rate": 1.9997080140801932e-05, |
| "loss": 1.259, |
| "step": 26 |
| }, |
| { |
| "epoch": 0.04, |
| "learning_rate": 1.9995795492789368e-05, |
| "loss": 1.2632, |
| "step": 27 |
| }, |
| { |
| "epoch": 0.04, |
| "learning_rate": 1.999427734334452e-05, |
| "loss": 1.2132, |
| "step": 28 |
| }, |
| { |
| "epoch": 0.04, |
| "learning_rate": 1.9992525727931303e-05, |
| "loss": 1.2555, |
| "step": 29 |
| }, |
| { |
| "epoch": 0.04, |
| "learning_rate": 1.9990540687467394e-05, |
| "loss": 1.2657, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.05, |
| "learning_rate": 1.998832226832327e-05, |
| "loss": 1.2168, |
| "step": 31 |
| }, |
| { |
| "epoch": 0.05, |
| "learning_rate": 1.9985870522321118e-05, |
| "loss": 1.2495, |
| "step": 32 |
| }, |
| { |
| "epoch": 0.05, |
| "learning_rate": 1.9983185506733643e-05, |
| "loss": 1.2284, |
| "step": 33 |
| }, |
| { |
| "epoch": 0.05, |
| "learning_rate": 1.9980267284282718e-05, |
| "loss": 1.2396, |
| "step": 34 |
| }, |
| { |
| "epoch": 0.05, |
| "learning_rate": 1.9977115923137912e-05, |
| "loss": 1.212, |
| "step": 35 |
| }, |
| { |
| "epoch": 0.05, |
| "learning_rate": 1.9973731496914914e-05, |
| "loss": 1.2334, |
| "step": 36 |
| }, |
| { |
| "epoch": 0.06, |
| "learning_rate": 1.9970114084673796e-05, |
| "loss": 1.2277, |
| "step": 37 |
| }, |
| { |
| "epoch": 0.06, |
| "learning_rate": 1.9966263770917192e-05, |
| "loss": 1.2427, |
| "step": 38 |
| }, |
| { |
| "epoch": 0.06, |
| "learning_rate": 1.996218064558829e-05, |
| "loss": 1.27, |
| "step": 39 |
| }, |
| { |
| "epoch": 0.06, |
| "learning_rate": 1.9957864804068752e-05, |
| "loss": 1.2774, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.06, |
| "learning_rate": 1.995331634717649e-05, |
| "loss": 1.2491, |
| "step": 41 |
| }, |
| { |
| "epoch": 0.06, |
| "learning_rate": 1.994853538116329e-05, |
| "loss": 1.2824, |
| "step": 42 |
| }, |
| { |
| "epoch": 0.06, |
| "learning_rate": 1.994352201771236e-05, |
| "loss": 1.2848, |
| "step": 43 |
| }, |
| { |
| "epoch": 0.07, |
| "learning_rate": 1.9938276373935688e-05, |
| "loss": 1.2225, |
| "step": 44 |
| }, |
| { |
| "epoch": 0.07, |
| "learning_rate": 1.993279857237133e-05, |
| "loss": 1.2455, |
| "step": 45 |
| }, |
| { |
| "epoch": 0.07, |
| "learning_rate": 1.992708874098054e-05, |
| "loss": 1.2611, |
| "step": 46 |
| }, |
| { |
| "epoch": 0.07, |
| "learning_rate": 1.9921147013144782e-05, |
| "loss": 1.2313, |
| "step": 47 |
| }, |
| { |
| "epoch": 0.07, |
| "learning_rate": 1.99149735276626e-05, |
| "loss": 1.2363, |
| "step": 48 |
| }, |
| { |
| "epoch": 0.07, |
| "learning_rate": 1.9908568428746408e-05, |
| "loss": 1.2269, |
| "step": 49 |
| }, |
| { |
| "epoch": 0.07, |
| "learning_rate": 1.9901931866019087e-05, |
| "loss": 1.2967, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.08, |
| "learning_rate": 1.9895063994510512e-05, |
| "loss": 1.2253, |
| "step": 51 |
| }, |
| { |
| "epoch": 0.08, |
| "learning_rate": 1.988796497465392e-05, |
| "loss": 1.2445, |
| "step": 52 |
| }, |
| { |
| "epoch": 0.08, |
| "learning_rate": 1.9880634972282168e-05, |
| "loss": 1.2534, |
| "step": 53 |
| }, |
| { |
| "epoch": 0.08, |
| "learning_rate": 1.987307415862385e-05, |
| "loss": 1.2541, |
| "step": 54 |
| }, |
| { |
| "epoch": 0.08, |
| "learning_rate": 1.986528271029931e-05, |
| "loss": 1.1781, |
| "step": 55 |
| }, |
| { |
| "epoch": 0.08, |
| "learning_rate": 1.985726080931651e-05, |
| "loss": 1.2588, |
| "step": 56 |
| }, |
| { |
| "epoch": 0.08, |
| "learning_rate": 1.9849008643066774e-05, |
| "loss": 1.2065, |
| "step": 57 |
| }, |
| { |
| "epoch": 0.09, |
| "learning_rate": 1.9840526404320415e-05, |
| "loss": 1.1835, |
| "step": 58 |
| }, |
| { |
| "epoch": 0.09, |
| "learning_rate": 1.9831814291222233e-05, |
| "loss": 1.2175, |
| "step": 59 |
| }, |
| { |
| "epoch": 0.09, |
| "learning_rate": 1.982287250728689e-05, |
| "loss": 1.284, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.09, |
| "learning_rate": 1.9813701261394136e-05, |
| "loss": 1.2142, |
| "step": 61 |
| }, |
| { |
| "epoch": 0.09, |
| "learning_rate": 1.9804300767783958e-05, |
| "loss": 1.1987, |
| "step": 62 |
| }, |
| { |
| "epoch": 0.09, |
| "learning_rate": 1.979467124605156e-05, |
| "loss": 1.2183, |
| "step": 63 |
| }, |
| { |
| "epoch": 0.1, |
| "learning_rate": 1.9784812921142232e-05, |
| "loss": 1.2228, |
| "step": 64 |
| }, |
| { |
| "epoch": 0.1, |
| "learning_rate": 1.977472602334609e-05, |
| "loss": 1.2348, |
| "step": 65 |
| }, |
| { |
| "epoch": 0.1, |
| "learning_rate": 1.9764410788292724e-05, |
| "loss": 1.2709, |
| "step": 66 |
| }, |
| { |
| "epoch": 0.1, |
| "learning_rate": 1.9753867456945653e-05, |
| "loss": 1.2204, |
| "step": 67 |
| }, |
| { |
| "epoch": 0.1, |
| "learning_rate": 1.9743096275596735e-05, |
| "loss": 1.2384, |
| "step": 68 |
| }, |
| { |
| "epoch": 0.1, |
| "learning_rate": 1.9732097495860388e-05, |
| "loss": 1.2821, |
| "step": 69 |
| }, |
| { |
| "epoch": 0.1, |
| "learning_rate": 1.9720871374667714e-05, |
| "loss": 1.2486, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.11, |
| "learning_rate": 1.9709418174260523e-05, |
| "loss": 1.2789, |
| "step": 71 |
| }, |
| { |
| "epoch": 0.11, |
| "learning_rate": 1.9697738162185163e-05, |
| "loss": 1.2097, |
| "step": 72 |
| }, |
| { |
| "epoch": 0.11, |
| "learning_rate": 1.9685831611286312e-05, |
| "loss": 1.2084, |
| "step": 73 |
| }, |
| { |
| "epoch": 0.11, |
| "learning_rate": 1.9673698799700582e-05, |
| "loss": 1.2385, |
| "step": 74 |
| }, |
| { |
| "epoch": 0.11, |
| "learning_rate": 1.9661340010850025e-05, |
| "loss": 1.255, |
| "step": 75 |
| }, |
| { |
| "epoch": 0.11, |
| "learning_rate": 1.9648755533435517e-05, |
| "loss": 1.1625, |
| "step": 76 |
| }, |
| { |
| "epoch": 0.11, |
| "learning_rate": 1.9635945661430006e-05, |
| "loss": 1.2474, |
| "step": 77 |
| }, |
| { |
| "epoch": 0.12, |
| "learning_rate": 1.9622910694071654e-05, |
| "loss": 1.2155, |
| "step": 78 |
| }, |
| { |
| "epoch": 0.12, |
| "learning_rate": 1.9609650935856847e-05, |
| "loss": 1.2247, |
| "step": 79 |
| }, |
| { |
| "epoch": 0.12, |
| "learning_rate": 1.9596166696533062e-05, |
| "loss": 1.2019, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.12, |
| "learning_rate": 1.9582458291091664e-05, |
| "loss": 1.201, |
| "step": 81 |
| }, |
| { |
| "epoch": 0.12, |
| "learning_rate": 1.956852603976052e-05, |
| "loss": 1.1818, |
| "step": 82 |
| }, |
| { |
| "epoch": 0.12, |
| "learning_rate": 1.9554370267996537e-05, |
| "loss": 1.2636, |
| "step": 83 |
| }, |
| { |
| "epoch": 0.13, |
| "learning_rate": 1.9539991306478046e-05, |
| "loss": 1.2019, |
| "step": 84 |
| }, |
| { |
| "epoch": 0.13, |
| "learning_rate": 1.952538949109708e-05, |
| "loss": 1.2112, |
| "step": 85 |
| }, |
| { |
| "epoch": 0.13, |
| "learning_rate": 1.9510565162951538e-05, |
| "loss": 1.2328, |
| "step": 86 |
| }, |
| { |
| "epoch": 0.13, |
| "learning_rate": 1.9495518668337204e-05, |
| "loss": 1.187, |
| "step": 87 |
| }, |
| { |
| "epoch": 0.13, |
| "learning_rate": 1.9480250358739667e-05, |
| "loss": 1.2198, |
| "step": 88 |
| }, |
| { |
| "epoch": 0.13, |
| "learning_rate": 1.94647605908261e-05, |
| "loss": 1.2043, |
| "step": 89 |
| }, |
| { |
| "epoch": 0.13, |
| "learning_rate": 1.944904972643694e-05, |
| "loss": 1.2699, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.14, |
| "learning_rate": 1.9433118132577432e-05, |
| "loss": 1.2724, |
| "step": 91 |
| }, |
| { |
| "epoch": 0.14, |
| "learning_rate": 1.9416966181409047e-05, |
| "loss": 1.29, |
| "step": 92 |
| }, |
| { |
| "epoch": 0.14, |
| "learning_rate": 1.94005942502408e-05, |
| "loss": 1.2716, |
| "step": 93 |
| }, |
| { |
| "epoch": 0.14, |
| "learning_rate": 1.9384002721520423e-05, |
| "loss": 1.2757, |
| "step": 94 |
| }, |
| { |
| "epoch": 0.14, |
| "learning_rate": 1.936719198282545e-05, |
| "loss": 1.208, |
| "step": 95 |
| }, |
| { |
| "epoch": 0.14, |
| "learning_rate": 1.9350162426854152e-05, |
| "loss": 1.2125, |
| "step": 96 |
| }, |
| { |
| "epoch": 0.14, |
| "learning_rate": 1.933291445141635e-05, |
| "loss": 1.2842, |
| "step": 97 |
| }, |
| { |
| "epoch": 0.15, |
| "learning_rate": 1.931544845942415e-05, |
| "loss": 1.1984, |
| "step": 98 |
| }, |
| { |
| "epoch": 0.15, |
| "learning_rate": 1.9297764858882516e-05, |
| "loss": 1.2639, |
| "step": 99 |
| }, |
| { |
| "epoch": 0.15, |
| "learning_rate": 1.927986406287973e-05, |
| "loss": 1.232, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.15, |
| "learning_rate": 1.9261746489577767e-05, |
| "loss": 1.2204, |
| "step": 101 |
| }, |
| { |
| "epoch": 0.15, |
| "learning_rate": 1.92434125622025e-05, |
| "loss": 1.1784, |
| "step": 102 |
| }, |
| { |
| "epoch": 0.15, |
| "learning_rate": 1.9224862709033823e-05, |
| "loss": 1.2618, |
| "step": 103 |
| }, |
| { |
| "epoch": 0.15, |
| "learning_rate": 1.9206097363395668e-05, |
| "loss": 1.1905, |
| "step": 104 |
| }, |
| { |
| "epoch": 0.16, |
| "learning_rate": 1.9187116963645845e-05, |
| "loss": 1.2221, |
| "step": 105 |
| }, |
| { |
| "epoch": 0.16, |
| "learning_rate": 1.9167921953165827e-05, |
| "loss": 1.123, |
| "step": 106 |
| }, |
| { |
| "epoch": 0.16, |
| "learning_rate": 1.9148512780350384e-05, |
| "loss": 1.2856, |
| "step": 107 |
| }, |
| { |
| "epoch": 0.16, |
| "learning_rate": 1.9128889898597117e-05, |
| "loss": 1.2297, |
| "step": 108 |
| }, |
| { |
| "epoch": 0.16, |
| "learning_rate": 1.910905376629585e-05, |
| "loss": 1.2103, |
| "step": 109 |
| }, |
| { |
| "epoch": 0.16, |
| "learning_rate": 1.9089004846817947e-05, |
| "loss": 1.2346, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.17, |
| "learning_rate": 1.9068743608505454e-05, |
| "loss": 1.2455, |
| "step": 111 |
| }, |
| { |
| "epoch": 0.17, |
| "learning_rate": 1.9048270524660197e-05, |
| "loss": 1.2047, |
| "step": 112 |
| }, |
| { |
| "epoch": 0.17, |
| "learning_rate": 1.902758607353269e-05, |
| "loss": 1.2628, |
| "step": 113 |
| }, |
| { |
| "epoch": 0.17, |
| "learning_rate": 1.9006690738310988e-05, |
| "loss": 1.2368, |
| "step": 114 |
| }, |
| { |
| "epoch": 0.17, |
| "learning_rate": 1.898558500710939e-05, |
| "loss": 1.2111, |
| "step": 115 |
| }, |
| { |
| "epoch": 0.17, |
| "learning_rate": 1.896426937295704e-05, |
| "loss": 1.217, |
| "step": 116 |
| }, |
| { |
| "epoch": 0.17, |
| "learning_rate": 1.89427443337864e-05, |
| "loss": 1.2152, |
| "step": 117 |
| }, |
| { |
| "epoch": 0.18, |
| "learning_rate": 1.8921010392421628e-05, |
| "loss": 1.2946, |
| "step": 118 |
| }, |
| { |
| "epoch": 0.18, |
| "learning_rate": 1.889906805656684e-05, |
| "loss": 1.1985, |
| "step": 119 |
| }, |
| { |
| "epoch": 0.18, |
| "learning_rate": 1.8876917838794226e-05, |
| "loss": 1.2813, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.18, |
| "learning_rate": 1.8854560256532098e-05, |
| "loss": 1.2304, |
| "step": 121 |
| }, |
| { |
| "epoch": 0.18, |
| "learning_rate": 1.8831995832052802e-05, |
| "loss": 1.2149, |
| "step": 122 |
| }, |
| { |
| "epoch": 0.18, |
| "learning_rate": 1.8809225092460488e-05, |
| "loss": 1.2348, |
| "step": 123 |
| }, |
| { |
| "epoch": 0.18, |
| "learning_rate": 1.8786248569678847e-05, |
| "loss": 1.2094, |
| "step": 124 |
| }, |
| { |
| "epoch": 0.19, |
| "learning_rate": 1.8763066800438638e-05, |
| "loss": 1.2816, |
| "step": 125 |
| }, |
| { |
| "epoch": 0.19, |
| "learning_rate": 1.873968032626518e-05, |
| "loss": 1.2087, |
| "step": 126 |
| }, |
| { |
| "epoch": 0.19, |
| "learning_rate": 1.8716089693465693e-05, |
| "loss": 1.2106, |
| "step": 127 |
| }, |
| { |
| "epoch": 0.19, |
| "learning_rate": 1.869229545311653e-05, |
| "loss": 1.2248, |
| "step": 128 |
| }, |
| { |
| "epoch": 0.19, |
| "learning_rate": 1.8668298161050308e-05, |
| "loss": 1.2067, |
| "step": 129 |
| }, |
| { |
| "epoch": 0.19, |
| "learning_rate": 1.8644098377842934e-05, |
| "loss": 1.2321, |
| "step": 130 |
| }, |
| { |
| "epoch": 0.19, |
| "learning_rate": 1.8619696668800494e-05, |
| "loss": 1.2214, |
| "step": 131 |
| }, |
| { |
| "epoch": 0.2, |
| "learning_rate": 1.8595093603946053e-05, |
| "loss": 1.2007, |
| "step": 132 |
| }, |
| { |
| "epoch": 0.2, |
| "learning_rate": 1.8570289758006346e-05, |
| "loss": 1.2108, |
| "step": 133 |
| }, |
| { |
| "epoch": 0.2, |
| "learning_rate": 1.8545285710398343e-05, |
| "loss": 1.255, |
| "step": 134 |
| }, |
| { |
| "epoch": 0.2, |
| "learning_rate": 1.852008204521572e-05, |
| "loss": 1.1655, |
| "step": 135 |
| }, |
| { |
| "epoch": 0.2, |
| "learning_rate": 1.8494679351215212e-05, |
| "loss": 1.1728, |
| "step": 136 |
| }, |
| { |
| "epoch": 0.2, |
| "learning_rate": 1.846907822180286e-05, |
| "loss": 1.2236, |
| "step": 137 |
| }, |
| { |
| "epoch": 0.21, |
| "learning_rate": 1.8443279255020153e-05, |
| "loss": 1.1678, |
| "step": 138 |
| }, |
| { |
| "epoch": 0.21, |
| "learning_rate": 1.8417283053530047e-05, |
| "loss": 1.2252, |
| "step": 139 |
| }, |
| { |
| "epoch": 0.21, |
| "learning_rate": 1.8391090224602895e-05, |
| "loss": 1.2186, |
| "step": 140 |
| }, |
| { |
| "epoch": 0.21, |
| "learning_rate": 1.8364701380102267e-05, |
| "loss": 1.1997, |
| "step": 141 |
| }, |
| { |
| "epoch": 0.21, |
| "learning_rate": 1.8338117136470645e-05, |
| "loss": 1.2197, |
| "step": 142 |
| }, |
| { |
| "epoch": 0.21, |
| "learning_rate": 1.831133811471503e-05, |
| "loss": 1.2612, |
| "step": 143 |
| }, |
| { |
| "epoch": 0.21, |
| "learning_rate": 1.8284364940392426e-05, |
| "loss": 1.1799, |
| "step": 144 |
| }, |
| { |
| "epoch": 0.22, |
| "learning_rate": 1.825719824359524e-05, |
| "loss": 1.2105, |
| "step": 145 |
| }, |
| { |
| "epoch": 0.22, |
| "learning_rate": 1.8229838658936566e-05, |
| "loss": 1.2328, |
| "step": 146 |
| }, |
| { |
| "epoch": 0.22, |
| "learning_rate": 1.820228682553533e-05, |
| "loss": 1.2605, |
| "step": 147 |
| }, |
| { |
| "epoch": 0.22, |
| "learning_rate": 1.8174543387001403e-05, |
| "loss": 1.1525, |
| "step": 148 |
| }, |
| { |
| "epoch": 0.22, |
| "learning_rate": 1.8146608991420533e-05, |
| "loss": 1.1731, |
| "step": 149 |
| }, |
| { |
| "epoch": 0.22, |
| "learning_rate": 1.811848429133922e-05, |
| "loss": 1.2118, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.22, |
| "learning_rate": 1.8090169943749477e-05, |
| "loss": 1.1936, |
| "step": 151 |
| }, |
| { |
| "epoch": 0.23, |
| "learning_rate": 1.8061666610073465e-05, |
| "loss": 1.13, |
| "step": 152 |
| }, |
| { |
| "epoch": 0.23, |
| "learning_rate": 1.8032974956148064e-05, |
| "loss": 1.1965, |
| "step": 153 |
| }, |
| { |
| "epoch": 0.23, |
| "learning_rate": 1.8004095652209304e-05, |
| "loss": 1.221, |
| "step": 154 |
| }, |
| { |
| "epoch": 0.23, |
| "learning_rate": 1.7975029372876706e-05, |
| "loss": 1.1929, |
| "step": 155 |
| }, |
| { |
| "epoch": 0.23, |
| "learning_rate": 1.7945776797137544e-05, |
| "loss": 1.1709, |
| "step": 156 |
| }, |
| { |
| "epoch": 0.23, |
| "learning_rate": 1.791633860833096e-05, |
| "loss": 1.2288, |
| "step": 157 |
| }, |
| { |
| "epoch": 0.24, |
| "learning_rate": 1.7886715494132008e-05, |
| "loss": 1.266, |
| "step": 158 |
| }, |
| { |
| "epoch": 0.24, |
| "learning_rate": 1.7856908146535602e-05, |
| "loss": 1.1853, |
| "step": 159 |
| }, |
| { |
| "epoch": 0.24, |
| "learning_rate": 1.7826917261840337e-05, |
| "loss": 1.1828, |
| "step": 160 |
| }, |
| { |
| "epoch": 0.24, |
| "learning_rate": 1.7796743540632226e-05, |
| "loss": 1.1766, |
| "step": 161 |
| }, |
| { |
| "epoch": 0.24, |
| "learning_rate": 1.7766387687768338e-05, |
| "loss": 1.2138, |
| "step": 162 |
| }, |
| { |
| "epoch": 0.24, |
| "learning_rate": 1.7735850412360332e-05, |
| "loss": 1.1841, |
| "step": 163 |
| }, |
| { |
| "epoch": 0.24, |
| "learning_rate": 1.7705132427757895e-05, |
| "loss": 1.2339, |
| "step": 164 |
| }, |
| { |
| "epoch": 0.25, |
| "learning_rate": 1.7674234451532065e-05, |
| "loss": 1.2144, |
| "step": 165 |
| }, |
| { |
| "epoch": 0.25, |
| "learning_rate": 1.7643157205458483e-05, |
| "loss": 1.2213, |
| "step": 166 |
| }, |
| { |
| "epoch": 0.25, |
| "learning_rate": 1.7611901415500536e-05, |
| "loss": 1.218, |
| "step": 167 |
| }, |
| { |
| "epoch": 0.25, |
| "learning_rate": 1.7580467811792374e-05, |
| "loss": 1.205, |
| "step": 168 |
| }, |
| { |
| "epoch": 0.25, |
| "learning_rate": 1.7548857128621878e-05, |
| "loss": 1.191, |
| "step": 169 |
| }, |
| { |
| "epoch": 0.25, |
| "learning_rate": 1.7517070104413497e-05, |
| "loss": 1.2311, |
| "step": 170 |
| }, |
| { |
| "epoch": 0.25, |
| "learning_rate": 1.7485107481711014e-05, |
| "loss": 1.2155, |
| "step": 171 |
| }, |
| { |
| "epoch": 0.26, |
| "learning_rate": 1.745297000716016e-05, |
| "loss": 1.1845, |
| "step": 172 |
| }, |
| { |
| "epoch": 0.26, |
| "learning_rate": 1.7420658431491224e-05, |
| "loss": 1.1699, |
| "step": 173 |
| }, |
| { |
| "epoch": 0.26, |
| "learning_rate": 1.7388173509501475e-05, |
| "loss": 1.2239, |
| "step": 174 |
| }, |
| { |
| "epoch": 0.26, |
| "learning_rate": 1.7355516000037555e-05, |
| "loss": 1.179, |
| "step": 175 |
| }, |
| { |
| "epoch": 0.26, |
| "learning_rate": 1.7322686665977738e-05, |
| "loss": 1.2051, |
| "step": 176 |
| }, |
| { |
| "epoch": 0.26, |
| "learning_rate": 1.7289686274214116e-05, |
| "loss": 1.1782, |
| "step": 177 |
| }, |
| { |
| "epoch": 0.26, |
| "learning_rate": 1.7256515595634688e-05, |
| "loss": 1.2368, |
| "step": 178 |
| }, |
| { |
| "epoch": 0.27, |
| "learning_rate": 1.722317540510534e-05, |
| "loss": 1.2095, |
| "step": 179 |
| }, |
| { |
| "epoch": 0.27, |
| "learning_rate": 1.7189666481451755e-05, |
| "loss": 1.2139, |
| "step": 180 |
| }, |
| { |
| "epoch": 0.27, |
| "learning_rate": 1.715598960744121e-05, |
| "loss": 1.1589, |
| "step": 181 |
| }, |
| { |
| "epoch": 0.27, |
| "learning_rate": 1.712214556976431e-05, |
| "loss": 1.1475, |
| "step": 182 |
| }, |
| { |
| "epoch": 0.27, |
| "learning_rate": 1.7088135159016584e-05, |
| "loss": 1.1527, |
| "step": 183 |
| }, |
| { |
| "epoch": 0.27, |
| "learning_rate": 1.7053959169680033e-05, |
| "loss": 1.1701, |
| "step": 184 |
| }, |
| { |
| "epoch": 0.28, |
| "learning_rate": 1.7019618400104572e-05, |
| "loss": 1.2017, |
| "step": 185 |
| }, |
| { |
| "epoch": 0.28, |
| "learning_rate": 1.6985113652489374e-05, |
| "loss": 1.2087, |
| "step": 186 |
| }, |
| { |
| "epoch": 0.28, |
| "learning_rate": 1.695044573286413e-05, |
| "loss": 1.249, |
| "step": 187 |
| }, |
| { |
| "epoch": 0.28, |
| "learning_rate": 1.6915615451070234e-05, |
| "loss": 1.1857, |
| "step": 188 |
| }, |
| { |
| "epoch": 0.28, |
| "learning_rate": 1.688062362074184e-05, |
| "loss": 1.2133, |
| "step": 189 |
| }, |
| { |
| "epoch": 0.28, |
| "learning_rate": 1.684547105928689e-05, |
| "loss": 1.2234, |
| "step": 190 |
| }, |
| { |
| "epoch": 0.28, |
| "learning_rate": 1.6810158587867973e-05, |
| "loss": 1.1919, |
| "step": 191 |
| }, |
| { |
| "epoch": 0.29, |
| "learning_rate": 1.677468703138319e-05, |
| "loss": 1.2703, |
| "step": 192 |
| }, |
| { |
| "epoch": 0.29, |
| "learning_rate": 1.673905721844686e-05, |
| "loss": 1.1511, |
| "step": 193 |
| }, |
| { |
| "epoch": 0.29, |
| "learning_rate": 1.670326998137016e-05, |
| "loss": 1.1969, |
| "step": 194 |
| }, |
| { |
| "epoch": 0.29, |
| "learning_rate": 1.666732615614169e-05, |
| "loss": 1.1847, |
| "step": 195 |
| }, |
| { |
| "epoch": 0.29, |
| "learning_rate": 1.6631226582407954e-05, |
| "loss": 1.2302, |
| "step": 196 |
| }, |
| { |
| "epoch": 0.29, |
| "learning_rate": 1.6594972103453727e-05, |
| "loss": 1.2383, |
| "step": 197 |
| }, |
| { |
| "epoch": 0.29, |
| "learning_rate": 1.6558563566182365e-05, |
| "loss": 1.2046, |
| "step": 198 |
| }, |
| { |
| "epoch": 0.3, |
| "learning_rate": 1.652200182109602e-05, |
| "loss": 1.173, |
| "step": 199 |
| }, |
| { |
| "epoch": 0.3, |
| "learning_rate": 1.6485287722275783e-05, |
| "loss": 1.1651, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.3, |
| "learning_rate": 1.6448422127361707e-05, |
| "loss": 1.1685, |
| "step": 201 |
| }, |
| { |
| "epoch": 0.3, |
| "learning_rate": 1.64114058975328e-05, |
| "loss": 1.2085, |
| "step": 202 |
| }, |
| { |
| "epoch": 0.3, |
| "learning_rate": 1.63742398974869e-05, |
| "loss": 1.2296, |
| "step": 203 |
| }, |
| { |
| "epoch": 0.3, |
| "learning_rate": 1.6336924995420453e-05, |
| "loss": 1.1602, |
| "step": 204 |
| }, |
| { |
| "epoch": 0.31, |
| "learning_rate": 1.6299462063008272e-05, |
| "loss": 1.26, |
| "step": 205 |
| }, |
| { |
| "epoch": 0.31, |
| "learning_rate": 1.626185197538314e-05, |
| "loss": 1.1697, |
| "step": 206 |
| }, |
| { |
| "epoch": 0.31, |
| "learning_rate": 1.6224095611115385e-05, |
| "loss": 1.2056, |
| "step": 207 |
| }, |
| { |
| "epoch": 0.31, |
| "learning_rate": 1.6186193852192356e-05, |
| "loss": 1.205, |
| "step": 208 |
| }, |
| { |
| "epoch": 0.31, |
| "learning_rate": 1.6148147583997813e-05, |
| "loss": 1.1565, |
| "step": 209 |
| }, |
| { |
| "epoch": 0.31, |
| "learning_rate": 1.6109957695291246e-05, |
| "loss": 1.1724, |
| "step": 210 |
| }, |
| { |
| "epoch": 0.31, |
| "learning_rate": 1.6071625078187113e-05, |
| "loss": 1.1426, |
| "step": 211 |
| }, |
| { |
| "epoch": 0.32, |
| "learning_rate": 1.603315062813401e-05, |
| "loss": 1.2175, |
| "step": 212 |
| }, |
| { |
| "epoch": 0.32, |
| "learning_rate": 1.5994535243893742e-05, |
| "loss": 1.205, |
| "step": 213 |
| }, |
| { |
| "epoch": 0.32, |
| "learning_rate": 1.5955779827520327e-05, |
| "loss": 1.2054, |
| "step": 214 |
| }, |
| { |
| "epoch": 0.32, |
| "learning_rate": 1.5916885284338937e-05, |
| "loss": 1.2056, |
| "step": 215 |
| }, |
| { |
| "epoch": 0.32, |
| "learning_rate": 1.5877852522924733e-05, |
| "loss": 1.1235, |
| "step": 216 |
| }, |
| { |
| "epoch": 0.32, |
| "learning_rate": 1.5838682455081657e-05, |
| "loss": 1.189, |
| "step": 217 |
| }, |
| { |
| "epoch": 0.32, |
| "learning_rate": 1.5799375995821116e-05, |
| "loss": 1.2081, |
| "step": 218 |
| }, |
| { |
| "epoch": 0.33, |
| "learning_rate": 1.5759934063340627e-05, |
| "loss": 1.2584, |
| "step": 219 |
| }, |
| { |
| "epoch": 0.33, |
| "learning_rate": 1.5720357579002346e-05, |
| "loss": 1.2219, |
| "step": 220 |
| }, |
| { |
| "epoch": 0.33, |
| "learning_rate": 1.568064746731156e-05, |
| "loss": 1.1252, |
| "step": 221 |
| }, |
| { |
| "epoch": 0.33, |
| "learning_rate": 1.5640804655895086e-05, |
| "loss": 1.2149, |
| "step": 222 |
| }, |
| { |
| "epoch": 0.33, |
| "learning_rate": 1.5600830075479604e-05, |
| "loss": 1.144, |
| "step": 223 |
| }, |
| { |
| "epoch": 0.33, |
| "learning_rate": 1.5560724659869905e-05, |
| "loss": 1.1205, |
| "step": 224 |
| }, |
| { |
| "epoch": 0.33, |
| "learning_rate": 1.5520489345927095e-05, |
| "loss": 1.2103, |
| "step": 225 |
| }, |
| { |
| "epoch": 0.34, |
| "learning_rate": 1.5480125073546705e-05, |
| "loss": 1.1597, |
| "step": 226 |
| }, |
| { |
| "epoch": 0.34, |
| "learning_rate": 1.5439632785636707e-05, |
| "loss": 1.217, |
| "step": 227 |
| }, |
| { |
| "epoch": 0.34, |
| "learning_rate": 1.539901342809554e-05, |
| "loss": 1.148, |
| "step": 228 |
| }, |
| { |
| "epoch": 0.34, |
| "learning_rate": 1.5358267949789968e-05, |
| "loss": 1.2158, |
| "step": 229 |
| }, |
| { |
| "epoch": 0.34, |
| "learning_rate": 1.5317397302532933e-05, |
| "loss": 1.2159, |
| "step": 230 |
| }, |
| { |
| "epoch": 0.34, |
| "learning_rate": 1.527640244106133e-05, |
| "loss": 1.1752, |
| "step": 231 |
| }, |
| { |
| "epoch": 0.35, |
| "learning_rate": 1.5235284323013674e-05, |
| "loss": 1.2458, |
| "step": 232 |
| }, |
| { |
| "epoch": 0.35, |
| "learning_rate": 1.5194043908907774e-05, |
| "loss": 1.1763, |
| "step": 233 |
| }, |
| { |
| "epoch": 0.35, |
| "learning_rate": 1.515268216211825e-05, |
| "loss": 1.1833, |
| "step": 234 |
| }, |
| { |
| "epoch": 0.35, |
| "learning_rate": 1.5111200048854055e-05, |
| "loss": 1.1247, |
| "step": 235 |
| }, |
| { |
| "epoch": 0.35, |
| "learning_rate": 1.5069598538135905e-05, |
| "loss": 1.1844, |
| "step": 236 |
| }, |
| { |
| "epoch": 0.35, |
| "learning_rate": 1.5027878601773633e-05, |
| "loss": 1.234, |
| "step": 237 |
| }, |
| { |
| "epoch": 0.35, |
| "learning_rate": 1.4986041214343487e-05, |
| "loss": 1.1977, |
| "step": 238 |
| }, |
| { |
| "epoch": 0.36, |
| "learning_rate": 1.494408735316537e-05, |
| "loss": 1.1907, |
| "step": 239 |
| }, |
| { |
| "epoch": 0.36, |
| "learning_rate": 1.490201799828001e-05, |
| "loss": 1.1919, |
| "step": 240 |
| }, |
| { |
| "epoch": 0.36, |
| "learning_rate": 1.485983413242606e-05, |
| "loss": 1.2155, |
| "step": 241 |
| }, |
| { |
| "epoch": 0.36, |
| "learning_rate": 1.4817536741017153e-05, |
| "loss": 1.1924, |
| "step": 242 |
| }, |
| { |
| "epoch": 0.36, |
| "learning_rate": 1.4775126812118865e-05, |
| "loss": 1.1404, |
| "step": 243 |
| }, |
| { |
| "epoch": 0.36, |
| "learning_rate": 1.473260533642565e-05, |
| "loss": 1.1864, |
| "step": 244 |
| }, |
| { |
| "epoch": 0.36, |
| "learning_rate": 1.4689973307237687e-05, |
| "loss": 1.1742, |
| "step": 245 |
| }, |
| { |
| "epoch": 0.37, |
| "learning_rate": 1.4647231720437687e-05, |
| "loss": 1.1345, |
| "step": 246 |
| }, |
| { |
| "epoch": 0.37, |
| "learning_rate": 1.4604381574467616e-05, |
| "loss": 1.1472, |
| "step": 247 |
| }, |
| { |
| "epoch": 0.37, |
| "learning_rate": 1.4561423870305385e-05, |
| "loss": 1.2398, |
| "step": 248 |
| }, |
| { |
| "epoch": 0.37, |
| "learning_rate": 1.4518359611441452e-05, |
| "loss": 1.1712, |
| "step": 249 |
| }, |
| { |
| "epoch": 0.37, |
| "learning_rate": 1.4475189803855399e-05, |
| "loss": 1.1982, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.37, |
| "learning_rate": 1.4431915455992416e-05, |
| "loss": 1.1724, |
| "step": 251 |
| }, |
| { |
| "epoch": 0.38, |
| "learning_rate": 1.438853757873975e-05, |
| "loss": 1.1742, |
| "step": 252 |
| }, |
| { |
| "epoch": 0.38, |
| "learning_rate": 1.4345057185403098e-05, |
| "loss": 1.199, |
| "step": 253 |
| }, |
| { |
| "epoch": 0.38, |
| "learning_rate": 1.430147529168292e-05, |
| "loss": 1.1732, |
| "step": 254 |
| }, |
| { |
| "epoch": 0.38, |
| "learning_rate": 1.4257792915650728e-05, |
| "loss": 1.2167, |
| "step": 255 |
| }, |
| { |
| "epoch": 0.38, |
| "learning_rate": 1.4214011077725291e-05, |
| "loss": 1.1388, |
| "step": 256 |
| }, |
| { |
| "epoch": 0.38, |
| "learning_rate": 1.4170130800648814e-05, |
| "loss": 1.2166, |
| "step": 257 |
| }, |
| { |
| "epoch": 0.38, |
| "learning_rate": 1.4126153109463025e-05, |
| "loss": 1.1451, |
| "step": 258 |
| }, |
| { |
| "epoch": 0.39, |
| "learning_rate": 1.4082079031485253e-05, |
| "loss": 1.1513, |
| "step": 259 |
| }, |
| { |
| "epoch": 0.39, |
| "learning_rate": 1.4037909596284411e-05, |
| "loss": 1.1704, |
| "step": 260 |
| }, |
| { |
| "epoch": 0.39, |
| "learning_rate": 1.3993645835656957e-05, |
| "loss": 1.1971, |
| "step": 261 |
| }, |
| { |
| "epoch": 0.39, |
| "learning_rate": 1.394928878360279e-05, |
| "loss": 1.2009, |
| "step": 262 |
| }, |
| { |
| "epoch": 0.39, |
| "learning_rate": 1.3904839476301088e-05, |
| "loss": 1.2543, |
| "step": 263 |
| }, |
| { |
| "epoch": 0.39, |
| "learning_rate": 1.3860298952086115e-05, |
| "loss": 1.1814, |
| "step": 264 |
| }, |
| { |
| "epoch": 0.39, |
| "learning_rate": 1.3815668251422953e-05, |
| "loss": 1.1364, |
| "step": 265 |
| }, |
| { |
| "epoch": 0.4, |
| "learning_rate": 1.3770948416883205e-05, |
| "loss": 1.194, |
| "step": 266 |
| }, |
| { |
| "epoch": 0.4, |
| "learning_rate": 1.3726140493120639e-05, |
| "loss": 1.1886, |
| "step": 267 |
| }, |
| { |
| "epoch": 0.4, |
| "learning_rate": 1.3681245526846782e-05, |
| "loss": 1.1643, |
| "step": 268 |
| }, |
| { |
| "epoch": 0.4, |
| "learning_rate": 1.3636264566806473e-05, |
| "loss": 1.1556, |
| "step": 269 |
| }, |
| { |
| "epoch": 0.4, |
| "learning_rate": 1.3591198663753358e-05, |
| "loss": 1.1988, |
| "step": 270 |
| }, |
| { |
| "epoch": 0.4, |
| "learning_rate": 1.354604887042536e-05, |
| "loss": 1.1777, |
| "step": 271 |
| }, |
| { |
| "epoch": 0.4, |
| "learning_rate": 1.3500816241520059e-05, |
| "loss": 1.1831, |
| "step": 272 |
| }, |
| { |
| "epoch": 0.41, |
| "learning_rate": 1.3455501833670089e-05, |
| "loss": 1.1756, |
| "step": 273 |
| }, |
| { |
| "epoch": 0.41, |
| "learning_rate": 1.3410106705418424e-05, |
| "loss": 1.1909, |
| "step": 274 |
| }, |
| { |
| "epoch": 0.41, |
| "learning_rate": 1.336463191719367e-05, |
| "loss": 1.1974, |
| "step": 275 |
| }, |
| { |
| "epoch": 0.41, |
| "learning_rate": 1.3319078531285286e-05, |
| "loss": 1.1798, |
| "step": 276 |
| }, |
| { |
| "epoch": 0.41, |
| "learning_rate": 1.3273447611818768e-05, |
| "loss": 1.1379, |
| "step": 277 |
| }, |
| { |
| "epoch": 0.41, |
| "learning_rate": 1.3227740224730799e-05, |
| "loss": 1.1749, |
| "step": 278 |
| }, |
| { |
| "epoch": 0.42, |
| "learning_rate": 1.3181957437744334e-05, |
| "loss": 1.1885, |
| "step": 279 |
| }, |
| { |
| "epoch": 0.42, |
| "learning_rate": 1.3136100320343674e-05, |
| "loss": 1.1213, |
| "step": 280 |
| }, |
| { |
| "epoch": 0.42, |
| "learning_rate": 1.3090169943749475e-05, |
| "loss": 1.1782, |
| "step": 281 |
| }, |
| { |
| "epoch": 0.42, |
| "learning_rate": 1.3044167380893726e-05, |
| "loss": 1.2075, |
| "step": 282 |
| }, |
| { |
| "epoch": 0.42, |
| "learning_rate": 1.2998093706394674e-05, |
| "loss": 1.1648, |
| "step": 283 |
| }, |
| { |
| "epoch": 0.42, |
| "learning_rate": 1.295194999653175e-05, |
| "loss": 1.1872, |
| "step": 284 |
| }, |
| { |
| "epoch": 0.42, |
| "learning_rate": 1.2905737329220394e-05, |
| "loss": 1.2484, |
| "step": 285 |
| }, |
| { |
| "epoch": 0.43, |
| "learning_rate": 1.2859456783986892e-05, |
| "loss": 1.1544, |
| "step": 286 |
| }, |
| { |
| "epoch": 0.43, |
| "learning_rate": 1.2813109441943166e-05, |
| "loss": 1.2163, |
| "step": 287 |
| }, |
| { |
| "epoch": 0.43, |
| "learning_rate": 1.2766696385761494e-05, |
| "loss": 1.1837, |
| "step": 288 |
| }, |
| { |
| "epoch": 0.43, |
| "learning_rate": 1.2720218699649243e-05, |
| "loss": 1.153, |
| "step": 289 |
| }, |
| { |
| "epoch": 0.43, |
| "learning_rate": 1.2673677469323535e-05, |
| "loss": 1.2074, |
| "step": 290 |
| }, |
| { |
| "epoch": 0.43, |
| "learning_rate": 1.2627073781985873e-05, |
| "loss": 1.1652, |
| "step": 291 |
| }, |
| { |
| "epoch": 0.43, |
| "learning_rate": 1.258040872629676e-05, |
| "loss": 1.2019, |
| "step": 292 |
| }, |
| { |
| "epoch": 0.44, |
| "learning_rate": 1.2533683392350264e-05, |
| "loss": 1.1428, |
| "step": 293 |
| }, |
| { |
| "epoch": 0.44, |
| "learning_rate": 1.2486898871648547e-05, |
| "loss": 1.1073, |
| "step": 294 |
| }, |
| { |
| "epoch": 0.44, |
| "learning_rate": 1.2440056257076374e-05, |
| "loss": 1.2043, |
| "step": 295 |
| }, |
| { |
| "epoch": 0.44, |
| "learning_rate": 1.2393156642875579e-05, |
| "loss": 1.1542, |
| "step": 296 |
| }, |
| { |
| "epoch": 0.44, |
| "learning_rate": 1.2346201124619502e-05, |
| "loss": 1.1572, |
| "step": 297 |
| }, |
| { |
| "epoch": 0.44, |
| "learning_rate": 1.2299190799187405e-05, |
| "loss": 1.1742, |
| "step": 298 |
| }, |
| { |
| "epoch": 0.45, |
| "learning_rate": 1.2252126764738845e-05, |
| "loss": 1.2058, |
| "step": 299 |
| }, |
| { |
| "epoch": 0.45, |
| "learning_rate": 1.2205010120688012e-05, |
| "loss": 1.1983, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.45, |
| "learning_rate": 1.2157841967678064e-05, |
| "loss": 1.171, |
| "step": 301 |
| }, |
| { |
| "epoch": 0.45, |
| "learning_rate": 1.2110623407555398e-05, |
| "loss": 1.1433, |
| "step": 302 |
| }, |
| { |
| "epoch": 0.45, |
| "learning_rate": 1.2063355543343925e-05, |
| "loss": 1.2067, |
| "step": 303 |
| }, |
| { |
| "epoch": 0.45, |
| "learning_rate": 1.2016039479219293e-05, |
| "loss": 1.18, |
| "step": 304 |
| }, |
| { |
| "epoch": 0.45, |
| "learning_rate": 1.1968676320483103e-05, |
| "loss": 1.145, |
| "step": 305 |
| }, |
| { |
| "epoch": 0.46, |
| "learning_rate": 1.1921267173537083e-05, |
| "loss": 1.157, |
| "step": 306 |
| }, |
| { |
| "epoch": 0.46, |
| "learning_rate": 1.187381314585725e-05, |
| "loss": 1.2327, |
| "step": 307 |
| }, |
| { |
| "epoch": 0.46, |
| "learning_rate": 1.1826315345968014e-05, |
| "loss": 1.2182, |
| "step": 308 |
| }, |
| { |
| "epoch": 0.46, |
| "learning_rate": 1.1778774883416325e-05, |
| "loss": 1.2291, |
| "step": 309 |
| }, |
| { |
| "epoch": 0.46, |
| "learning_rate": 1.1731192868745717e-05, |
| "loss": 1.2624, |
| "step": 310 |
| }, |
| { |
| "epoch": 0.46, |
| "learning_rate": 1.1683570413470386e-05, |
| "loss": 1.1245, |
| "step": 311 |
| }, |
| { |
| "epoch": 0.46, |
| "learning_rate": 1.163590863004922e-05, |
| "loss": 1.2198, |
| "step": 312 |
| }, |
| { |
| "epoch": 0.47, |
| "learning_rate": 1.1588208631859808e-05, |
| "loss": 1.121, |
| "step": 313 |
| }, |
| { |
| "epoch": 0.47, |
| "learning_rate": 1.154047153317243e-05, |
| "loss": 1.1509, |
| "step": 314 |
| }, |
| { |
| "epoch": 0.47, |
| "learning_rate": 1.1492698449124042e-05, |
| "loss": 1.0908, |
| "step": 315 |
| }, |
| { |
| "epoch": 0.47, |
| "learning_rate": 1.1444890495692214e-05, |
| "loss": 1.2005, |
| "step": 316 |
| }, |
| { |
| "epoch": 0.47, |
| "learning_rate": 1.1397048789669061e-05, |
| "loss": 1.1822, |
| "step": 317 |
| }, |
| { |
| "epoch": 0.47, |
| "learning_rate": 1.1349174448635158e-05, |
| "loss": 1.1689, |
| "step": 318 |
| }, |
| { |
| "epoch": 0.47, |
| "learning_rate": 1.1301268590933434e-05, |
| "loss": 1.1987, |
| "step": 319 |
| }, |
| { |
| "epoch": 0.48, |
| "learning_rate": 1.1253332335643043e-05, |
| "loss": 1.1795, |
| "step": 320 |
| }, |
| { |
| "epoch": 0.48, |
| "learning_rate": 1.1205366802553233e-05, |
| "loss": 1.1771, |
| "step": 321 |
| }, |
| { |
| "epoch": 0.48, |
| "learning_rate": 1.1157373112137171e-05, |
| "loss": 1.1479, |
| "step": 322 |
| }, |
| { |
| "epoch": 0.48, |
| "learning_rate": 1.1109352385525782e-05, |
| "loss": 1.1888, |
| "step": 323 |
| }, |
| { |
| "epoch": 0.48, |
| "learning_rate": 1.1061305744481561e-05, |
| "loss": 1.1953, |
| "step": 324 |
| }, |
| { |
| "epoch": 0.48, |
| "learning_rate": 1.1013234311372353e-05, |
| "loss": 1.2657, |
| "step": 325 |
| }, |
| { |
| "epoch": 0.49, |
| "learning_rate": 1.096513920914515e-05, |
| "loss": 1.2082, |
| "step": 326 |
| }, |
| { |
| "epoch": 0.49, |
| "learning_rate": 1.0917021561299864e-05, |
| "loss": 1.212, |
| "step": 327 |
| }, |
| { |
| "epoch": 0.49, |
| "learning_rate": 1.0868882491863048e-05, |
| "loss": 1.1922, |
| "step": 328 |
| }, |
| { |
| "epoch": 0.49, |
| "learning_rate": 1.0820723125361685e-05, |
| "loss": 1.1676, |
| "step": 329 |
| }, |
| { |
| "epoch": 0.49, |
| "learning_rate": 1.077254458679689e-05, |
| "loss": 1.1622, |
| "step": 330 |
| }, |
| { |
| "epoch": 0.49, |
| "learning_rate": 1.0724348001617626e-05, |
| "loss": 1.1919, |
| "step": 331 |
| }, |
| { |
| "epoch": 0.49, |
| "learning_rate": 1.0676134495694437e-05, |
| "loss": 1.1744, |
| "step": 332 |
| }, |
| { |
| "epoch": 0.5, |
| "learning_rate": 1.0627905195293135e-05, |
| "loss": 1.1615, |
| "step": 333 |
| }, |
| { |
| "epoch": 0.5, |
| "learning_rate": 1.0579661227048484e-05, |
| "loss": 1.1561, |
| "step": 334 |
| }, |
| { |
| "epoch": 0.5, |
| "learning_rate": 1.0531403717937888e-05, |
| "loss": 1.194, |
| "step": 335 |
| }, |
| { |
| "epoch": 0.5, |
| "learning_rate": 1.0483133795255072e-05, |
| "loss": 1.1245, |
| "step": 336 |
| }, |
| { |
| "epoch": 0.5, |
| "learning_rate": 1.0434852586583734e-05, |
| "loss": 1.1741, |
| "step": 337 |
| }, |
| { |
| "epoch": 0.5, |
| "learning_rate": 1.0386561219771222e-05, |
| "loss": 1.1384, |
| "step": 338 |
| }, |
| { |
| "epoch": 0.5, |
| "learning_rate": 1.0338260822902166e-05, |
| "loss": 1.1889, |
| "step": 339 |
| }, |
| { |
| "epoch": 0.51, |
| "learning_rate": 1.0289952524272147e-05, |
| "loss": 1.2371, |
| "step": 340 |
| }, |
| { |
| "epoch": 0.51, |
| "learning_rate": 1.0241637452361327e-05, |
| "loss": 1.1536, |
| "step": 341 |
| }, |
| { |
| "epoch": 0.51, |
| "learning_rate": 1.0193316735808085e-05, |
| "loss": 1.1573, |
| "step": 342 |
| }, |
| { |
| "epoch": 0.51, |
| "learning_rate": 1.0144991503382676e-05, |
| "loss": 1.1573, |
| "step": 343 |
| }, |
| { |
| "epoch": 0.51, |
| "learning_rate": 1.0096662883960833e-05, |
| "loss": 1.1788, |
| "step": 344 |
| }, |
| { |
| "epoch": 0.51, |
| "learning_rate": 1.0048332006497406e-05, |
| "loss": 1.1736, |
| "step": 345 |
| }, |
| { |
| "epoch": 0.51, |
| "learning_rate": 1e-05, |
| "loss": 1.224, |
| "step": 346 |
| }, |
| { |
| "epoch": 0.52, |
| "learning_rate": 9.951667993502599e-06, |
| "loss": 1.1847, |
| "step": 347 |
| }, |
| { |
| "epoch": 0.52, |
| "learning_rate": 9.903337116039172e-06, |
| "loss": 1.2004, |
| "step": 348 |
| }, |
| { |
| "epoch": 0.52, |
| "learning_rate": 9.855008496617326e-06, |
| "loss": 1.1378, |
| "step": 349 |
| }, |
| { |
| "epoch": 0.52, |
| "learning_rate": 9.806683264191916e-06, |
| "loss": 1.171, |
| "step": 350 |
| }, |
| { |
| "epoch": 0.52, |
| "learning_rate": 9.75836254763868e-06, |
| "loss": 1.162, |
| "step": 351 |
| }, |
| { |
| "epoch": 0.52, |
| "learning_rate": 9.710047475727858e-06, |
| "loss": 1.1802, |
| "step": 352 |
| }, |
| { |
| "epoch": 0.53, |
| "learning_rate": 9.661739177097834e-06, |
| "loss": 1.1686, |
| "step": 353 |
| }, |
| { |
| "epoch": 0.53, |
| "learning_rate": 9.61343878022878e-06, |
| "loss": 1.1529, |
| "step": 354 |
| }, |
| { |
| "epoch": 0.53, |
| "learning_rate": 9.565147413416266e-06, |
| "loss": 1.1879, |
| "step": 355 |
| }, |
| { |
| "epoch": 0.53, |
| "learning_rate": 9.516866204744932e-06, |
| "loss": 1.1072, |
| "step": 356 |
| }, |
| { |
| "epoch": 0.53, |
| "learning_rate": 9.468596282062112e-06, |
| "loss": 1.1716, |
| "step": 357 |
| }, |
| { |
| "epoch": 0.53, |
| "learning_rate": 9.420338772951521e-06, |
| "loss": 1.2, |
| "step": 358 |
| }, |
| { |
| "epoch": 0.53, |
| "learning_rate": 9.372094804706867e-06, |
| "loss": 1.1524, |
| "step": 359 |
| }, |
| { |
| "epoch": 0.54, |
| "learning_rate": 9.323865504305566e-06, |
| "loss": 1.1684, |
| "step": 360 |
| }, |
| { |
| "epoch": 0.54, |
| "learning_rate": 9.275651998382377e-06, |
| "loss": 1.163, |
| "step": 361 |
| }, |
| { |
| "epoch": 0.54, |
| "learning_rate": 9.227455413203117e-06, |
| "loss": 1.1506, |
| "step": 362 |
| }, |
| { |
| "epoch": 0.54, |
| "learning_rate": 9.179276874638315e-06, |
| "loss": 1.1785, |
| "step": 363 |
| }, |
| { |
| "epoch": 0.54, |
| "learning_rate": 9.131117508136952e-06, |
| "loss": 1.1574, |
| "step": 364 |
| }, |
| { |
| "epoch": 0.54, |
| "learning_rate": 9.082978438700141e-06, |
| "loss": 1.1907, |
| "step": 365 |
| }, |
| { |
| "epoch": 0.54, |
| "learning_rate": 9.034860790854848e-06, |
| "loss": 1.1482, |
| "step": 366 |
| }, |
| { |
| "epoch": 0.55, |
| "learning_rate": 8.986765688627652e-06, |
| "loss": 1.1086, |
| "step": 367 |
| }, |
| { |
| "epoch": 0.55, |
| "learning_rate": 8.938694255518442e-06, |
| "loss": 1.1637, |
| "step": 368 |
| }, |
| { |
| "epoch": 0.55, |
| "learning_rate": 8.890647614474223e-06, |
| "loss": 1.1991, |
| "step": 369 |
| }, |
| { |
| "epoch": 0.55, |
| "learning_rate": 8.842626887862832e-06, |
| "loss": 1.1559, |
| "step": 370 |
| }, |
| { |
| "epoch": 0.55, |
| "learning_rate": 8.79463319744677e-06, |
| "loss": 1.2513, |
| "step": 371 |
| }, |
| { |
| "epoch": 0.55, |
| "learning_rate": 8.74666766435696e-06, |
| "loss": 1.1638, |
| "step": 372 |
| }, |
| { |
| "epoch": 0.56, |
| "learning_rate": 8.698731409066571e-06, |
| "loss": 1.1504, |
| "step": 373 |
| }, |
| { |
| "epoch": 0.56, |
| "learning_rate": 8.650825551364844e-06, |
| "loss": 1.1642, |
| "step": 374 |
| }, |
| { |
| "epoch": 0.56, |
| "learning_rate": 8.60295121033094e-06, |
| "loss": 1.2143, |
| "step": 375 |
| }, |
| { |
| "epoch": 0.56, |
| "learning_rate": 8.555109504307787e-06, |
| "loss": 1.167, |
| "step": 376 |
| }, |
| { |
| "epoch": 0.56, |
| "learning_rate": 8.50730155087596e-06, |
| "loss": 1.0684, |
| "step": 377 |
| }, |
| { |
| "epoch": 0.56, |
| "learning_rate": 8.459528466827576e-06, |
| "loss": 1.1521, |
| "step": 378 |
| }, |
| { |
| "epoch": 0.56, |
| "learning_rate": 8.411791368140197e-06, |
| "loss": 1.1821, |
| "step": 379 |
| }, |
| { |
| "epoch": 0.57, |
| "learning_rate": 8.364091369950783e-06, |
| "loss": 1.1589, |
| "step": 380 |
| }, |
| { |
| "epoch": 0.57, |
| "learning_rate": 8.316429586529616e-06, |
| "loss": 1.1123, |
| "step": 381 |
| }, |
| { |
| "epoch": 0.57, |
| "learning_rate": 8.268807131254288e-06, |
| "loss": 1.1787, |
| "step": 382 |
| }, |
| { |
| "epoch": 0.57, |
| "learning_rate": 8.22122511658368e-06, |
| "loss": 1.1102, |
| "step": 383 |
| }, |
| { |
| "epoch": 0.57, |
| "learning_rate": 8.173684654031986e-06, |
| "loss": 1.1509, |
| "step": 384 |
| }, |
| { |
| "epoch": 0.57, |
| "learning_rate": 8.126186854142754e-06, |
| "loss": 1.1295, |
| "step": 385 |
| }, |
| { |
| "epoch": 0.57, |
| "learning_rate": 8.078732826462917e-06, |
| "loss": 1.1561, |
| "step": 386 |
| }, |
| { |
| "epoch": 0.58, |
| "learning_rate": 8.0313236795169e-06, |
| "loss": 1.1393, |
| "step": 387 |
| }, |
| { |
| "epoch": 0.58, |
| "learning_rate": 7.983960520780712e-06, |
| "loss": 1.1318, |
| "step": 388 |
| }, |
| { |
| "epoch": 0.58, |
| "learning_rate": 7.936644456656082e-06, |
| "loss": 1.1277, |
| "step": 389 |
| }, |
| { |
| "epoch": 0.58, |
| "learning_rate": 7.889376592444605e-06, |
| "loss": 1.2151, |
| "step": 390 |
| }, |
| { |
| "epoch": 0.58, |
| "learning_rate": 7.84215803232194e-06, |
| "loss": 1.1762, |
| "step": 391 |
| }, |
| { |
| "epoch": 0.58, |
| "learning_rate": 7.794989879311991e-06, |
| "loss": 1.1384, |
| "step": 392 |
| }, |
| { |
| "epoch": 0.58, |
| "learning_rate": 7.74787323526116e-06, |
| "loss": 1.0996, |
| "step": 393 |
| }, |
| { |
| "epoch": 0.59, |
| "learning_rate": 7.700809200812598e-06, |
| "loss": 1.1601, |
| "step": 394 |
| }, |
| { |
| "epoch": 0.59, |
| "learning_rate": 7.653798875380498e-06, |
| "loss": 1.1501, |
| "step": 395 |
| }, |
| { |
| "epoch": 0.59, |
| "learning_rate": 7.6068433571244234e-06, |
| "loss": 1.1391, |
| "step": 396 |
| }, |
| { |
| "epoch": 0.59, |
| "learning_rate": 7.559943742923626e-06, |
| "loss": 1.1429, |
| "step": 397 |
| }, |
| { |
| "epoch": 0.59, |
| "learning_rate": 7.513101128351454e-06, |
| "loss": 1.1895, |
| "step": 398 |
| }, |
| { |
| "epoch": 0.59, |
| "learning_rate": 7.466316607649735e-06, |
| "loss": 1.171, |
| "step": 399 |
| }, |
| { |
| "epoch": 0.6, |
| "learning_rate": 7.419591273703245e-06, |
| "loss": 1.1202, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.6, |
| "learning_rate": 7.372926218014131e-06, |
| "loss": 1.1702, |
| "step": 401 |
| }, |
| { |
| "epoch": 0.6, |
| "learning_rate": 7.326322530676471e-06, |
| "loss": 1.1175, |
| "step": 402 |
| }, |
| { |
| "epoch": 0.6, |
| "learning_rate": 7.27978130035076e-06, |
| "loss": 1.1549, |
| "step": 403 |
| }, |
| { |
| "epoch": 0.6, |
| "learning_rate": 7.233303614238511e-06, |
| "loss": 1.167, |
| "step": 404 |
| }, |
| { |
| "epoch": 0.6, |
| "learning_rate": 7.186890558056836e-06, |
| "loss": 1.1407, |
| "step": 405 |
| }, |
| { |
| "epoch": 0.6, |
| "learning_rate": 7.1405432160131076e-06, |
| "loss": 1.1076, |
| "step": 406 |
| }, |
| { |
| "epoch": 0.61, |
| "learning_rate": 7.0942626707796094e-06, |
| "loss": 1.1575, |
| "step": 407 |
| }, |
| { |
| "epoch": 0.61, |
| "learning_rate": 7.048050003468252e-06, |
| "loss": 1.1732, |
| "step": 408 |
| }, |
| { |
| "epoch": 0.61, |
| "learning_rate": 7.001906293605329e-06, |
| "loss": 1.1869, |
| "step": 409 |
| }, |
| { |
| "epoch": 0.61, |
| "learning_rate": 6.9558326191062775e-06, |
| "loss": 1.1875, |
| "step": 410 |
| }, |
| { |
| "epoch": 0.61, |
| "learning_rate": 6.909830056250527e-06, |
| "loss": 1.1501, |
| "step": 411 |
| }, |
| { |
| "epoch": 0.61, |
| "learning_rate": 6.8638996796563275e-06, |
| "loss": 1.1729, |
| "step": 412 |
| }, |
| { |
| "epoch": 0.61, |
| "learning_rate": 6.81804256225567e-06, |
| "loss": 1.1432, |
| "step": 413 |
| }, |
| { |
| "epoch": 0.62, |
| "learning_rate": 6.7722597752692055e-06, |
| "loss": 1.1358, |
| "step": 414 |
| }, |
| { |
| "epoch": 0.62, |
| "learning_rate": 6.726552388181235e-06, |
| "loss": 1.143, |
| "step": 415 |
| }, |
| { |
| "epoch": 0.62, |
| "learning_rate": 6.6809214687147165e-06, |
| "loss": 1.1349, |
| "step": 416 |
| }, |
| { |
| "epoch": 0.62, |
| "learning_rate": 6.6353680828063306e-06, |
| "loss": 1.1159, |
| "step": 417 |
| }, |
| { |
| "epoch": 0.62, |
| "learning_rate": 6.589893294581579e-06, |
| "loss": 1.1903, |
| "step": 418 |
| }, |
| { |
| "epoch": 0.62, |
| "learning_rate": 6.5444981663299135e-06, |
| "loss": 1.1597, |
| "step": 419 |
| }, |
| { |
| "epoch": 0.63, |
| "learning_rate": 6.499183758479944e-06, |
| "loss": 1.1564, |
| "step": 420 |
| }, |
| { |
| "epoch": 0.63, |
| "learning_rate": 6.453951129574644e-06, |
| "loss": 1.1286, |
| "step": 421 |
| }, |
| { |
| "epoch": 0.63, |
| "learning_rate": 6.408801336246645e-06, |
| "loss": 1.1151, |
| "step": 422 |
| }, |
| { |
| "epoch": 0.63, |
| "learning_rate": 6.363735433193532e-06, |
| "loss": 1.1358, |
| "step": 423 |
| }, |
| { |
| "epoch": 0.63, |
| "learning_rate": 6.318754473153224e-06, |
| "loss": 1.14, |
| "step": 424 |
| }, |
| { |
| "epoch": 0.63, |
| "learning_rate": 6.273859506879365e-06, |
| "loss": 1.1771, |
| "step": 425 |
| }, |
| { |
| "epoch": 0.63, |
| "learning_rate": 6.229051583116799e-06, |
| "loss": 1.1251, |
| "step": 426 |
| }, |
| { |
| "epoch": 0.64, |
| "learning_rate": 6.184331748577049e-06, |
| "loss": 1.0971, |
| "step": 427 |
| }, |
| { |
| "epoch": 0.64, |
| "learning_rate": 6.139701047913885e-06, |
| "loss": 1.0982, |
| "step": 428 |
| }, |
| { |
| "epoch": 0.64, |
| "learning_rate": 6.095160523698913e-06, |
| "loss": 1.1474, |
| "step": 429 |
| }, |
| { |
| "epoch": 0.64, |
| "learning_rate": 6.0507112163972106e-06, |
| "loss": 1.1362, |
| "step": 430 |
| }, |
| { |
| "epoch": 0.64, |
| "learning_rate": 6.006354164343047e-06, |
| "loss": 1.1188, |
| "step": 431 |
| }, |
| { |
| "epoch": 0.64, |
| "learning_rate": 5.962090403715592e-06, |
| "loss": 1.1482, |
| "step": 432 |
| }, |
| { |
| "epoch": 0.64, |
| "learning_rate": 5.9179209685147525e-06, |
| "loss": 1.1055, |
| "step": 433 |
| }, |
| { |
| "epoch": 0.65, |
| "learning_rate": 5.873846890536977e-06, |
| "loss": 1.1933, |
| "step": 434 |
| }, |
| { |
| "epoch": 0.65, |
| "learning_rate": 5.829869199351188e-06, |
| "loss": 1.1561, |
| "step": 435 |
| }, |
| { |
| "epoch": 0.65, |
| "learning_rate": 5.785988922274711e-06, |
| "loss": 1.1167, |
| "step": 436 |
| }, |
| { |
| "epoch": 0.65, |
| "learning_rate": 5.742207084349274e-06, |
| "loss": 1.1961, |
| "step": 437 |
| }, |
| { |
| "epoch": 0.65, |
| "learning_rate": 5.698524708317082e-06, |
| "loss": 1.1883, |
| "step": 438 |
| }, |
| { |
| "epoch": 0.65, |
| "learning_rate": 5.654942814596902e-06, |
| "loss": 1.1935, |
| "step": 439 |
| }, |
| { |
| "epoch": 0.65, |
| "learning_rate": 5.611462421260251e-06, |
| "loss": 1.0944, |
| "step": 440 |
| }, |
| { |
| "epoch": 0.66, |
| "learning_rate": 5.5680845440075885e-06, |
| "loss": 1.1319, |
| "step": 441 |
| }, |
| { |
| "epoch": 0.66, |
| "learning_rate": 5.5248101961446065e-06, |
| "loss": 1.2082, |
| "step": 442 |
| }, |
| { |
| "epoch": 0.66, |
| "learning_rate": 5.481640388558551e-06, |
| "loss": 1.1499, |
| "step": 443 |
| }, |
| { |
| "epoch": 0.66, |
| "learning_rate": 5.43857612969462e-06, |
| "loss": 1.0991, |
| "step": 444 |
| }, |
| { |
| "epoch": 0.66, |
| "learning_rate": 5.3956184255323855e-06, |
| "loss": 1.1476, |
| "step": 445 |
| }, |
| { |
| "epoch": 0.66, |
| "learning_rate": 5.352768279562315e-06, |
| "loss": 1.1776, |
| "step": 446 |
| }, |
| { |
| "epoch": 0.67, |
| "learning_rate": 5.310026692762316e-06, |
| "loss": 1.1296, |
| "step": 447 |
| }, |
| { |
| "epoch": 0.67, |
| "learning_rate": 5.267394663574351e-06, |
| "loss": 1.087, |
| "step": 448 |
| }, |
| { |
| "epoch": 0.67, |
| "learning_rate": 5.224873187881136e-06, |
| "loss": 1.1738, |
| "step": 449 |
| }, |
| { |
| "epoch": 0.67, |
| "learning_rate": 5.1824632589828465e-06, |
| "loss": 1.1627, |
| "step": 450 |
| }, |
| { |
| "epoch": 0.67, |
| "learning_rate": 5.14016586757394e-06, |
| "loss": 1.1736, |
| "step": 451 |
| }, |
| { |
| "epoch": 0.67, |
| "learning_rate": 5.097982001719994e-06, |
| "loss": 1.1486, |
| "step": 452 |
| }, |
| { |
| "epoch": 0.67, |
| "learning_rate": 5.0559126468346354e-06, |
| "loss": 1.2196, |
| "step": 453 |
| }, |
| { |
| "epoch": 0.68, |
| "learning_rate": 5.013958785656516e-06, |
| "loss": 1.2005, |
| "step": 454 |
| }, |
| { |
| "epoch": 0.68, |
| "learning_rate": 4.972121398226371e-06, |
| "loss": 1.1786, |
| "step": 455 |
| }, |
| { |
| "epoch": 0.68, |
| "learning_rate": 4.930401461864096e-06, |
| "loss": 1.1405, |
| "step": 456 |
| }, |
| { |
| "epoch": 0.68, |
| "learning_rate": 4.888799951145948e-06, |
| "loss": 1.15, |
| "step": 457 |
| }, |
| { |
| "epoch": 0.68, |
| "learning_rate": 4.847317837881757e-06, |
| "loss": 1.1965, |
| "step": 458 |
| }, |
| { |
| "epoch": 0.68, |
| "learning_rate": 4.805956091092228e-06, |
| "loss": 1.1939, |
| "step": 459 |
| }, |
| { |
| "epoch": 0.68, |
| "learning_rate": 4.764715676986327e-06, |
| "loss": 1.0932, |
| "step": 460 |
| }, |
| { |
| "epoch": 0.69, |
| "learning_rate": 4.7235975589386715e-06, |
| "loss": 1.1657, |
| "step": 461 |
| }, |
| { |
| "epoch": 0.69, |
| "learning_rate": 4.6826026974670665e-06, |
| "loss": 1.1878, |
| "step": 462 |
| }, |
| { |
| "epoch": 0.69, |
| "learning_rate": 4.641732050210036e-06, |
| "loss": 1.1552, |
| "step": 463 |
| }, |
| { |
| "epoch": 0.69, |
| "learning_rate": 4.6009865719044645e-06, |
| "loss": 1.0927, |
| "step": 464 |
| }, |
| { |
| "epoch": 0.69, |
| "learning_rate": 4.560367214363295e-06, |
| "loss": 1.0978, |
| "step": 465 |
| }, |
| { |
| "epoch": 0.69, |
| "learning_rate": 4.519874926453303e-06, |
| "loss": 1.1518, |
| "step": 466 |
| }, |
| { |
| "epoch": 0.7, |
| "learning_rate": 4.479510654072905e-06, |
| "loss": 1.0652, |
| "step": 467 |
| }, |
| { |
| "epoch": 0.7, |
| "learning_rate": 4.439275340130099e-06, |
| "loss": 1.1441, |
| "step": 468 |
| }, |
| { |
| "epoch": 0.7, |
| "learning_rate": 4.399169924520403e-06, |
| "loss": 1.0733, |
| "step": 469 |
| }, |
| { |
| "epoch": 0.7, |
| "learning_rate": 4.359195344104916e-06, |
| "loss": 1.1216, |
| "step": 470 |
| }, |
| { |
| "epoch": 0.7, |
| "learning_rate": 4.319352532688444e-06, |
| "loss": 1.1248, |
| "step": 471 |
| }, |
| { |
| "epoch": 0.7, |
| "learning_rate": 4.279642420997655e-06, |
| "loss": 1.1487, |
| "step": 472 |
| }, |
| { |
| "epoch": 0.7, |
| "learning_rate": 4.240065936659374e-06, |
| "loss": 1.1352, |
| "step": 473 |
| }, |
| { |
| "epoch": 0.71, |
| "learning_rate": 4.200624004178886e-06, |
| "loss": 1.1473, |
| "step": 474 |
| }, |
| { |
| "epoch": 0.71, |
| "learning_rate": 4.1613175449183484e-06, |
| "loss": 1.1667, |
| "step": 475 |
| }, |
| { |
| "epoch": 0.71, |
| "learning_rate": 4.12214747707527e-06, |
| "loss": 1.1294, |
| "step": 476 |
| }, |
| { |
| "epoch": 0.71, |
| "learning_rate": 4.083114715661069e-06, |
| "loss": 1.1839, |
| "step": 477 |
| }, |
| { |
| "epoch": 0.71, |
| "learning_rate": 4.044220172479675e-06, |
| "loss": 1.0997, |
| "step": 478 |
| }, |
| { |
| "epoch": 0.71, |
| "learning_rate": 4.0054647561062625e-06, |
| "loss": 1.0926, |
| "step": 479 |
| }, |
| { |
| "epoch": 0.71, |
| "learning_rate": 3.9668493718659924e-06, |
| "loss": 1.1799, |
| "step": 480 |
| }, |
| { |
| "epoch": 0.72, |
| "learning_rate": 3.9283749218128885e-06, |
| "loss": 1.1347, |
| "step": 481 |
| }, |
| { |
| "epoch": 0.72, |
| "learning_rate": 3.890042304708758e-06, |
| "loss": 1.1829, |
| "step": 482 |
| }, |
| { |
| "epoch": 0.72, |
| "learning_rate": 3.8518524160021876e-06, |
| "loss": 1.1717, |
| "step": 483 |
| }, |
| { |
| "epoch": 0.72, |
| "learning_rate": 3.813806147807645e-06, |
| "loss": 1.1431, |
| "step": 484 |
| }, |
| { |
| "epoch": 0.72, |
| "learning_rate": 3.775904388884615e-06, |
| "loss": 1.1427, |
| "step": 485 |
| }, |
| { |
| "epoch": 0.72, |
| "learning_rate": 3.7381480246168665e-06, |
| "loss": 1.1423, |
| "step": 486 |
| }, |
| { |
| "epoch": 0.72, |
| "learning_rate": 3.700537936991733e-06, |
| "loss": 1.1204, |
| "step": 487 |
| }, |
| { |
| "epoch": 0.73, |
| "learning_rate": 3.6630750045795506e-06, |
| "loss": 1.1612, |
| "step": 488 |
| }, |
| { |
| "epoch": 0.73, |
| "learning_rate": 3.625760102513103e-06, |
| "loss": 1.1158, |
| "step": 489 |
| }, |
| { |
| "epoch": 0.73, |
| "learning_rate": 3.5885941024672e-06, |
| "loss": 1.1642, |
| "step": 490 |
| }, |
| { |
| "epoch": 0.73, |
| "learning_rate": 3.5515778726382933e-06, |
| "loss": 1.146, |
| "step": 491 |
| }, |
| { |
| "epoch": 0.73, |
| "learning_rate": 3.5147122777242203e-06, |
| "loss": 1.1832, |
| "step": 492 |
| }, |
| { |
| "epoch": 0.73, |
| "learning_rate": 3.477998178903982e-06, |
| "loss": 1.1724, |
| "step": 493 |
| }, |
| { |
| "epoch": 0.74, |
| "learning_rate": 3.4414364338176376e-06, |
| "loss": 1.1447, |
| "step": 494 |
| }, |
| { |
| "epoch": 0.74, |
| "learning_rate": 3.405027896546277e-06, |
| "loss": 1.1156, |
| "step": 495 |
| }, |
| { |
| "epoch": 0.74, |
| "learning_rate": 3.368773417592047e-06, |
| "loss": 1.1168, |
| "step": 496 |
| }, |
| { |
| "epoch": 0.74, |
| "learning_rate": 3.3326738438583116e-06, |
| "loss": 1.1344, |
| "step": 497 |
| }, |
| { |
| "epoch": 0.74, |
| "learning_rate": 3.2967300186298456e-06, |
| "loss": 1.1161, |
| "step": 498 |
| }, |
| { |
| "epoch": 0.74, |
| "learning_rate": 3.2609427815531448e-06, |
| "loss": 1.1473, |
| "step": 499 |
| }, |
| { |
| "epoch": 0.74, |
| "learning_rate": 3.2253129686168105e-06, |
| "loss": 1.0854, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.75, |
| "learning_rate": 3.18984141213203e-06, |
| "loss": 1.1898, |
| "step": 501 |
| }, |
| { |
| "epoch": 0.75, |
| "learning_rate": 3.1545289407131128e-06, |
| "loss": 1.1258, |
| "step": 502 |
| }, |
| { |
| "epoch": 0.75, |
| "learning_rate": 3.11937637925816e-06, |
| "loss": 1.1519, |
| "step": 503 |
| }, |
| { |
| "epoch": 0.75, |
| "learning_rate": 3.0843845489297698e-06, |
| "loss": 1.1353, |
| "step": 504 |
| }, |
| { |
| "epoch": 0.75, |
| "learning_rate": 3.0495542671358715e-06, |
| "loss": 1.1341, |
| "step": 505 |
| }, |
| { |
| "epoch": 0.75, |
| "learning_rate": 3.0148863475106315e-06, |
| "loss": 1.1278, |
| "step": 506 |
| }, |
| { |
| "epoch": 0.75, |
| "learning_rate": 2.98038159989543e-06, |
| "loss": 1.1183, |
| "step": 507 |
| }, |
| { |
| "epoch": 0.76, |
| "learning_rate": 2.9460408303199696e-06, |
| "loss": 1.1502, |
| "step": 508 |
| }, |
| { |
| "epoch": 0.76, |
| "learning_rate": 2.9118648409834205e-06, |
| "loss": 1.1151, |
| "step": 509 |
| }, |
| { |
| "epoch": 0.76, |
| "learning_rate": 2.8778544302356938e-06, |
| "loss": 1.0964, |
| "step": 510 |
| }, |
| { |
| "epoch": 0.76, |
| "learning_rate": 2.8440103925587904e-06, |
| "loss": 1.1387, |
| "step": 511 |
| }, |
| { |
| "epoch": 0.76, |
| "learning_rate": 2.810333518548246e-06, |
| "loss": 1.0998, |
| "step": 512 |
| }, |
| { |
| "epoch": 0.76, |
| "learning_rate": 2.7768245948946615e-06, |
| "loss": 1.1495, |
| "step": 513 |
| }, |
| { |
| "epoch": 0.77, |
| "learning_rate": 2.743484404365314e-06, |
| "loss": 1.153, |
| "step": 514 |
| }, |
| { |
| "epoch": 0.77, |
| "learning_rate": 2.7103137257858867e-06, |
| "loss": 1.1609, |
| "step": 515 |
| }, |
| { |
| "epoch": 0.77, |
| "learning_rate": 2.6773133340222647e-06, |
| "loss": 1.1579, |
| "step": 516 |
| }, |
| { |
| "epoch": 0.77, |
| "learning_rate": 2.6444839999624496e-06, |
| "loss": 1.1728, |
| "step": 517 |
| }, |
| { |
| "epoch": 0.77, |
| "learning_rate": 2.611826490498527e-06, |
| "loss": 1.1371, |
| "step": 518 |
| }, |
| { |
| "epoch": 0.77, |
| "learning_rate": 2.5793415685087797e-06, |
| "loss": 1.1787, |
| "step": 519 |
| }, |
| { |
| "epoch": 0.77, |
| "learning_rate": 2.5470299928398424e-06, |
| "loss": 1.1468, |
| "step": 520 |
| }, |
| { |
| "epoch": 0.78, |
| "learning_rate": 2.5148925182889916e-06, |
| "loss": 1.1859, |
| "step": 521 |
| }, |
| { |
| "epoch": 0.78, |
| "learning_rate": 2.4829298955865022e-06, |
| "loss": 1.088, |
| "step": 522 |
| }, |
| { |
| "epoch": 0.78, |
| "learning_rate": 2.451142871378124e-06, |
| "loss": 1.1379, |
| "step": 523 |
| }, |
| { |
| "epoch": 0.78, |
| "learning_rate": 2.4195321882076295e-06, |
| "loss": 1.1373, |
| "step": 524 |
| }, |
| { |
| "epoch": 0.78, |
| "learning_rate": 2.3880985844994674e-06, |
| "loss": 1.2, |
| "step": 525 |
| }, |
| { |
| "epoch": 0.78, |
| "learning_rate": 2.3568427945415196e-06, |
| "loss": 1.1295, |
| "step": 526 |
| }, |
| { |
| "epoch": 0.78, |
| "learning_rate": 2.3257655484679376e-06, |
| "loss": 1.1727, |
| "step": 527 |
| }, |
| { |
| "epoch": 0.79, |
| "learning_rate": 2.2948675722421086e-06, |
| "loss": 1.1532, |
| "step": 528 |
| }, |
| { |
| "epoch": 0.79, |
| "learning_rate": 2.264149587639668e-06, |
| "loss": 1.1731, |
| "step": 529 |
| }, |
| { |
| "epoch": 0.79, |
| "learning_rate": 2.2336123122316642e-06, |
| "loss": 1.1585, |
| "step": 530 |
| }, |
| { |
| "epoch": 0.79, |
| "learning_rate": 2.2032564593677773e-06, |
| "loss": 1.0844, |
| "step": 531 |
| }, |
| { |
| "epoch": 0.79, |
| "learning_rate": 2.1730827381596677e-06, |
| "loss": 1.1697, |
| "step": 532 |
| }, |
| { |
| "epoch": 0.79, |
| "learning_rate": 2.1430918534643996e-06, |
| "loss": 1.107, |
| "step": 533 |
| }, |
| { |
| "epoch": 0.79, |
| "learning_rate": 2.1132845058679917e-06, |
| "loss": 1.1906, |
| "step": 534 |
| }, |
| { |
| "epoch": 0.8, |
| "learning_rate": 2.083661391669043e-06, |
| "loss": 1.1275, |
| "step": 535 |
| }, |
| { |
| "epoch": 0.8, |
| "learning_rate": 2.0542232028624585e-06, |
| "loss": 1.0969, |
| "step": 536 |
| }, |
| { |
| "epoch": 0.8, |
| "learning_rate": 2.024970627123297e-06, |
| "loss": 1.1445, |
| "step": 537 |
| }, |
| { |
| "epoch": 0.8, |
| "learning_rate": 1.9959043477907e-06, |
| "loss": 1.0651, |
| "step": 538 |
| }, |
| { |
| "epoch": 0.8, |
| "learning_rate": 1.967025043851939e-06, |
| "loss": 1.0988, |
| "step": 539 |
| }, |
| { |
| "epoch": 0.8, |
| "learning_rate": 1.9383333899265368e-06, |
| "loss": 1.0866, |
| "step": 540 |
| }, |
| { |
| "epoch": 0.81, |
| "learning_rate": 1.9098300562505266e-06, |
| "loss": 1.1574, |
| "step": 541 |
| }, |
| { |
| "epoch": 0.81, |
| "learning_rate": 1.8815157086607826e-06, |
| "loss": 1.1884, |
| "step": 542 |
| }, |
| { |
| "epoch": 0.81, |
| "learning_rate": 1.8533910085794714e-06, |
| "loss": 1.1106, |
| "step": 543 |
| }, |
| { |
| "epoch": 0.81, |
| "learning_rate": 1.8254566129985996e-06, |
| "loss": 1.138, |
| "step": 544 |
| }, |
| { |
| "epoch": 0.81, |
| "learning_rate": 1.7977131744646692e-06, |
| "loss": 1.1997, |
| "step": 545 |
| }, |
| { |
| "epoch": 0.81, |
| "learning_rate": 1.7701613410634367e-06, |
| "loss": 1.1939, |
| "step": 546 |
| }, |
| { |
| "epoch": 0.81, |
| "learning_rate": 1.7428017564047594e-06, |
| "loss": 1.1176, |
| "step": 547 |
| }, |
| { |
| "epoch": 0.82, |
| "learning_rate": 1.7156350596075777e-06, |
| "loss": 1.1404, |
| "step": 548 |
| }, |
| { |
| "epoch": 0.82, |
| "learning_rate": 1.6886618852849723e-06, |
| "loss": 1.1449, |
| "step": 549 |
| }, |
| { |
| "epoch": 0.82, |
| "learning_rate": 1.6618828635293561e-06, |
| "loss": 1.1488, |
| "step": 550 |
| }, |
| { |
| "epoch": 0.82, |
| "learning_rate": 1.6352986198977327e-06, |
| "loss": 1.1442, |
| "step": 551 |
| }, |
| { |
| "epoch": 0.82, |
| "learning_rate": 1.6089097753971061e-06, |
| "loss": 1.0947, |
| "step": 552 |
| }, |
| { |
| "epoch": 0.82, |
| "learning_rate": 1.5827169464699576e-06, |
| "loss": 1.1533, |
| "step": 553 |
| }, |
| { |
| "epoch": 0.82, |
| "learning_rate": 1.5567207449798517e-06, |
| "loss": 1.1549, |
| "step": 554 |
| }, |
| { |
| "epoch": 0.83, |
| "learning_rate": 1.5309217781971419e-06, |
| "loss": 1.1368, |
| "step": 555 |
| }, |
| { |
| "epoch": 0.83, |
| "learning_rate": 1.5053206487847893e-06, |
| "loss": 1.1504, |
| "step": 556 |
| }, |
| { |
| "epoch": 0.83, |
| "learning_rate": 1.4799179547842823e-06, |
| "loss": 1.1365, |
| "step": 557 |
| }, |
| { |
| "epoch": 0.83, |
| "learning_rate": 1.4547142896016586e-06, |
| "loss": 1.1375, |
| "step": 558 |
| }, |
| { |
| "epoch": 0.83, |
| "learning_rate": 1.4297102419936582e-06, |
| "loss": 1.1443, |
| "step": 559 |
| }, |
| { |
| "epoch": 0.83, |
| "learning_rate": 1.4049063960539488e-06, |
| "loss": 1.1405, |
| "step": 560 |
| }, |
| { |
| "epoch": 0.83, |
| "learning_rate": 1.3803033311995096e-06, |
| "loss": 1.1526, |
| "step": 561 |
| }, |
| { |
| "epoch": 0.84, |
| "learning_rate": 1.3559016221570663e-06, |
| "loss": 1.0707, |
| "step": 562 |
| }, |
| { |
| "epoch": 0.84, |
| "learning_rate": 1.3317018389496927e-06, |
| "loss": 1.1649, |
| "step": 563 |
| }, |
| { |
| "epoch": 0.84, |
| "learning_rate": 1.3077045468834714e-06, |
| "loss": 1.1294, |
| "step": 564 |
| }, |
| { |
| "epoch": 0.84, |
| "learning_rate": 1.2839103065343084e-06, |
| "loss": 1.0976, |
| "step": 565 |
| }, |
| { |
| "epoch": 0.84, |
| "learning_rate": 1.2603196737348211e-06, |
| "loss": 1.1443, |
| "step": 566 |
| }, |
| { |
| "epoch": 0.84, |
| "learning_rate": 1.2369331995613643e-06, |
| "loss": 1.1315, |
| "step": 567 |
| }, |
| { |
| "epoch": 0.85, |
| "learning_rate": 1.213751430321156e-06, |
| "loss": 1.1398, |
| "step": 568 |
| }, |
| { |
| "epoch": 0.85, |
| "learning_rate": 1.1907749075395126e-06, |
| "loss": 1.1239, |
| "step": 569 |
| }, |
| { |
| "epoch": 0.85, |
| "learning_rate": 1.168004167947202e-06, |
| "loss": 1.1014, |
| "step": 570 |
| }, |
| { |
| "epoch": 0.85, |
| "learning_rate": 1.1454397434679022e-06, |
| "loss": 1.1451, |
| "step": 571 |
| }, |
| { |
| "epoch": 0.85, |
| "learning_rate": 1.1230821612057764e-06, |
| "loss": 1.1545, |
| "step": 572 |
| }, |
| { |
| "epoch": 0.85, |
| "learning_rate": 1.1009319434331623e-06, |
| "loss": 1.0954, |
| "step": 573 |
| }, |
| { |
| "epoch": 0.85, |
| "learning_rate": 1.0789896075783734e-06, |
| "loss": 1.1322, |
| "step": 574 |
| }, |
| { |
| "epoch": 0.86, |
| "learning_rate": 1.0572556662136036e-06, |
| "loss": 1.1098, |
| "step": 575 |
| }, |
| { |
| "epoch": 0.86, |
| "learning_rate": 1.0357306270429623e-06, |
| "loss": 1.1495, |
| "step": 576 |
| }, |
| { |
| "epoch": 0.86, |
| "learning_rate": 1.014414992890611e-06, |
| "loss": 1.1342, |
| "step": 577 |
| }, |
| { |
| "epoch": 0.86, |
| "learning_rate": 9.933092616890127e-07, |
| "loss": 1.1954, |
| "step": 578 |
| }, |
| { |
| "epoch": 0.86, |
| "learning_rate": 9.724139264673116e-07, |
| "loss": 1.1296, |
| "step": 579 |
| }, |
| { |
| "epoch": 0.86, |
| "learning_rate": 9.517294753398043e-07, |
| "loss": 1.1447, |
| "step": 580 |
| }, |
| { |
| "epoch": 0.86, |
| "learning_rate": 9.312563914945461e-07, |
| "loss": 1.082, |
| "step": 581 |
| }, |
| { |
| "epoch": 0.87, |
| "learning_rate": 9.10995153182056e-07, |
| "loss": 1.1625, |
| "step": 582 |
| }, |
| { |
| "epoch": 0.87, |
| "learning_rate": 8.909462337041508e-07, |
| "loss": 1.1774, |
| "step": 583 |
| }, |
| { |
| "epoch": 0.87, |
| "learning_rate": 8.711101014028855e-07, |
| "loss": 1.1183, |
| "step": 584 |
| }, |
| { |
| "epoch": 0.87, |
| "learning_rate": 8.514872196496182e-07, |
| "loss": 1.1269, |
| "step": 585 |
| }, |
| { |
| "epoch": 0.87, |
| "learning_rate": 8.320780468341761e-07, |
| "loss": 1.0736, |
| "step": 586 |
| }, |
| { |
| "epoch": 0.87, |
| "learning_rate": 8.128830363541574e-07, |
| "loss": 1.1115, |
| "step": 587 |
| }, |
| { |
| "epoch": 0.88, |
| "learning_rate": 7.939026366043346e-07, |
| "loss": 1.1085, |
| "step": 588 |
| }, |
| { |
| "epoch": 0.88, |
| "learning_rate": 7.75137290966177e-07, |
| "loss": 1.1869, |
| "step": 589 |
| }, |
| { |
| "epoch": 0.88, |
| "learning_rate": 7.565874377975046e-07, |
| "loss": 1.1166, |
| "step": 590 |
| }, |
| { |
| "epoch": 0.88, |
| "learning_rate": 7.382535104222344e-07, |
| "loss": 1.124, |
| "step": 591 |
| }, |
| { |
| "epoch": 0.88, |
| "learning_rate": 7.201359371202698e-07, |
| "loss": 1.1771, |
| "step": 592 |
| }, |
| { |
| "epoch": 0.88, |
| "learning_rate": 7.022351411174866e-07, |
| "loss": 1.1086, |
| "step": 593 |
| }, |
| { |
| "epoch": 0.88, |
| "learning_rate": 6.845515405758518e-07, |
| "loss": 1.1588, |
| "step": 594 |
| }, |
| { |
| "epoch": 0.89, |
| "learning_rate": 6.670855485836525e-07, |
| "loss": 1.1115, |
| "step": 595 |
| }, |
| { |
| "epoch": 0.89, |
| "learning_rate": 6.498375731458529e-07, |
| "loss": 1.1282, |
| "step": 596 |
| }, |
| { |
| "epoch": 0.89, |
| "learning_rate": 6.32808017174551e-07, |
| "loss": 1.16, |
| "step": 597 |
| }, |
| { |
| "epoch": 0.89, |
| "learning_rate": 6.159972784795798e-07, |
| "loss": 1.1252, |
| "step": 598 |
| }, |
| { |
| "epoch": 0.89, |
| "learning_rate": 5.994057497592054e-07, |
| "loss": 1.1086, |
| "step": 599 |
| }, |
| { |
| "epoch": 0.89, |
| "learning_rate": 5.830338185909545e-07, |
| "loss": 1.1208, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.89, |
| "learning_rate": 5.668818674225696e-07, |
| "loss": 1.1315, |
| "step": 601 |
| }, |
| { |
| "epoch": 0.9, |
| "learning_rate": 5.509502735630601e-07, |
| "loss": 1.0897, |
| "step": 602 |
| }, |
| { |
| "epoch": 0.9, |
| "learning_rate": 5.352394091739022e-07, |
| "loss": 1.1225, |
| "step": 603 |
| }, |
| { |
| "epoch": 0.9, |
| "learning_rate": 5.197496412603365e-07, |
| "loss": 1.1742, |
| "step": 604 |
| }, |
| { |
| "epoch": 0.9, |
| "learning_rate": 5.044813316627994e-07, |
| "loss": 1.1343, |
| "step": 605 |
| }, |
| { |
| "epoch": 0.9, |
| "learning_rate": 4.894348370484648e-07, |
| "loss": 1.1333, |
| "step": 606 |
| }, |
| { |
| "epoch": 0.9, |
| "learning_rate": 4.746105089029229e-07, |
| "loss": 1.154, |
| "step": 607 |
| }, |
| { |
| "epoch": 0.9, |
| "learning_rate": 4.6000869352195607e-07, |
| "loss": 1.0746, |
| "step": 608 |
| }, |
| { |
| "epoch": 0.91, |
| "learning_rate": 4.4562973200346413e-07, |
| "loss": 1.1104, |
| "step": 609 |
| }, |
| { |
| "epoch": 0.91, |
| "learning_rate": 4.314739602394813e-07, |
| "loss": 1.1653, |
| "step": 610 |
| }, |
| { |
| "epoch": 0.91, |
| "learning_rate": 4.1754170890833777e-07, |
| "loss": 1.1645, |
| "step": 611 |
| }, |
| { |
| "epoch": 0.91, |
| "learning_rate": 4.038333034669406e-07, |
| "loss": 1.1725, |
| "step": 612 |
| }, |
| { |
| "epoch": 0.91, |
| "learning_rate": 3.903490641431573e-07, |
| "loss": 1.1453, |
| "step": 613 |
| }, |
| { |
| "epoch": 0.91, |
| "learning_rate": 3.770893059283465e-07, |
| "loss": 1.1292, |
| "step": 614 |
| }, |
| { |
| "epoch": 0.92, |
| "learning_rate": 3.6405433856999684e-07, |
| "loss": 1.0625, |
| "step": 615 |
| }, |
| { |
| "epoch": 0.92, |
| "learning_rate": 3.5124446656448654e-07, |
| "loss": 1.1049, |
| "step": 616 |
| }, |
| { |
| "epoch": 0.92, |
| "learning_rate": 3.3865998914997645e-07, |
| "loss": 1.0797, |
| "step": 617 |
| }, |
| { |
| "epoch": 0.92, |
| "learning_rate": 3.2630120029942034e-07, |
| "loss": 1.0995, |
| "step": 618 |
| }, |
| { |
| "epoch": 0.92, |
| "learning_rate": 3.1416838871368925e-07, |
| "loss": 1.1127, |
| "step": 619 |
| }, |
| { |
| "epoch": 0.92, |
| "learning_rate": 3.0226183781483786e-07, |
| "loss": 1.1122, |
| "step": 620 |
| }, |
| { |
| "epoch": 0.92, |
| "learning_rate": 2.90581825739481e-07, |
| "loss": 1.1484, |
| "step": 621 |
| }, |
| { |
| "epoch": 0.93, |
| "learning_rate": 2.791286253322856e-07, |
| "loss": 1.0956, |
| "step": 622 |
| }, |
| { |
| "epoch": 0.93, |
| "learning_rate": 2.679025041396155e-07, |
| "loss": 1.0632, |
| "step": 623 |
| }, |
| { |
| "epoch": 0.93, |
| "learning_rate": 2.569037244032657e-07, |
| "loss": 1.1184, |
| "step": 624 |
| }, |
| { |
| "epoch": 0.93, |
| "learning_rate": 2.461325430543482e-07, |
| "loss": 1.1178, |
| "step": 625 |
| }, |
| { |
| "epoch": 0.93, |
| "learning_rate": 2.3558921170728e-07, |
| "loss": 1.1309, |
| "step": 626 |
| }, |
| { |
| "epoch": 0.93, |
| "learning_rate": 2.2527397665391137e-07, |
| "loss": 1.1154, |
| "step": 627 |
| }, |
| { |
| "epoch": 0.93, |
| "learning_rate": 2.1518707885777147e-07, |
| "loss": 1.1322, |
| "step": 628 |
| }, |
| { |
| "epoch": 0.94, |
| "learning_rate": 2.0532875394844053e-07, |
| "loss": 1.1287, |
| "step": 629 |
| }, |
| { |
| "epoch": 0.94, |
| "learning_rate": 1.9569923221604224e-07, |
| "loss": 1.1287, |
| "step": 630 |
| }, |
| { |
| "epoch": 0.94, |
| "learning_rate": 1.8629873860586567e-07, |
| "loss": 1.0926, |
| "step": 631 |
| }, |
| { |
| "epoch": 0.94, |
| "learning_rate": 1.7712749271311392e-07, |
| "loss": 1.1313, |
| "step": 632 |
| }, |
| { |
| "epoch": 0.94, |
| "learning_rate": 1.681857087777672e-07, |
| "loss": 1.1336, |
| "step": 633 |
| }, |
| { |
| "epoch": 0.94, |
| "learning_rate": 1.5947359567958677e-07, |
| "loss": 1.1612, |
| "step": 634 |
| }, |
| { |
| "epoch": 0.95, |
| "learning_rate": 1.5099135693322776e-07, |
| "loss": 1.1071, |
| "step": 635 |
| }, |
| { |
| "epoch": 0.95, |
| "learning_rate": 1.4273919068349184e-07, |
| "loss": 1.102, |
| "step": 636 |
| }, |
| { |
| "epoch": 0.95, |
| "learning_rate": 1.3471728970068986e-07, |
| "loss": 1.1045, |
| "step": 637 |
| }, |
| { |
| "epoch": 0.95, |
| "learning_rate": 1.2692584137615205e-07, |
| "loss": 1.125, |
| "step": 638 |
| }, |
| { |
| "epoch": 0.95, |
| "learning_rate": 1.1936502771783488e-07, |
| "loss": 1.1887, |
| "step": 639 |
| }, |
| { |
| "epoch": 0.95, |
| "learning_rate": 1.1203502534608113e-07, |
| "loss": 1.0997, |
| "step": 640 |
| }, |
| { |
| "epoch": 0.95, |
| "learning_rate": 1.0493600548948879e-07, |
| "loss": 1.0907, |
| "step": 641 |
| }, |
| { |
| "epoch": 0.96, |
| "learning_rate": 9.806813398091419e-08, |
| "loss": 1.1638, |
| "step": 642 |
| }, |
| { |
| "epoch": 0.96, |
| "learning_rate": 9.143157125359403e-08, |
| "loss": 1.1088, |
| "step": 643 |
| }, |
| { |
| "epoch": 0.96, |
| "learning_rate": 8.502647233740169e-08, |
| "loss": 1.069, |
| "step": 644 |
| }, |
| { |
| "epoch": 0.96, |
| "learning_rate": 7.885298685522235e-08, |
| "loss": 1.1003, |
| "step": 645 |
| }, |
| { |
| "epoch": 0.96, |
| "learning_rate": 7.291125901946027e-08, |
| "loss": 1.1347, |
| "step": 646 |
| }, |
| { |
| "epoch": 0.96, |
| "learning_rate": 6.720142762867032e-08, |
| "loss": 1.1776, |
| "step": 647 |
| }, |
| { |
| "epoch": 0.96, |
| "learning_rate": 6.172362606431281e-08, |
| "loss": 1.1104, |
| "step": 648 |
| }, |
| { |
| "epoch": 0.97, |
| "learning_rate": 5.647798228764156e-08, |
| "loss": 1.1623, |
| "step": 649 |
| }, |
| { |
| "epoch": 0.97, |
| "learning_rate": 5.146461883671072e-08, |
| "loss": 1.0729, |
| "step": 650 |
| }, |
| { |
| "epoch": 0.97, |
| "learning_rate": 4.6683652823513725e-08, |
| "loss": 1.0913, |
| "step": 651 |
| }, |
| { |
| "epoch": 0.97, |
| "learning_rate": 4.2135195931249925e-08, |
| "loss": 1.149, |
| "step": 652 |
| }, |
| { |
| "epoch": 0.97, |
| "learning_rate": 3.781935441171225e-08, |
| "loss": 1.0957, |
| "step": 653 |
| }, |
| { |
| "epoch": 0.97, |
| "learning_rate": 3.373622908280916e-08, |
| "loss": 1.0875, |
| "step": 654 |
| }, |
| { |
| "epoch": 0.97, |
| "learning_rate": 2.988591532620322e-08, |
| "loss": 1.1031, |
| "step": 655 |
| }, |
| { |
| "epoch": 0.98, |
| "learning_rate": 2.6268503085089547e-08, |
| "loss": 1.1561, |
| "step": 656 |
| }, |
| { |
| "epoch": 0.98, |
| "learning_rate": 2.2884076862089712e-08, |
| "loss": 1.2111, |
| "step": 657 |
| }, |
| { |
| "epoch": 0.98, |
| "learning_rate": 1.973271571728441e-08, |
| "loss": 1.1655, |
| "step": 658 |
| }, |
| { |
| "epoch": 0.98, |
| "learning_rate": 1.6814493266357202e-08, |
| "loss": 1.1432, |
| "step": 659 |
| }, |
| { |
| "epoch": 0.98, |
| "learning_rate": 1.4129477678884728e-08, |
| "loss": 1.1066, |
| "step": 660 |
| }, |
| { |
| "epoch": 0.98, |
| "learning_rate": 1.1677731676734694e-08, |
| "loss": 1.1371, |
| "step": 661 |
| }, |
| { |
| "epoch": 0.99, |
| "learning_rate": 9.459312532608122e-09, |
| "loss": 1.1792, |
| "step": 662 |
| }, |
| { |
| "epoch": 0.99, |
| "learning_rate": 7.474272068698219e-09, |
| "loss": 1.0863, |
| "step": 663 |
| }, |
| { |
| "epoch": 0.99, |
| "learning_rate": 5.722656655482439e-09, |
| "loss": 1.1954, |
| "step": 664 |
| }, |
| { |
| "epoch": 0.99, |
| "learning_rate": 4.204507210633368e-09, |
| "loss": 1.1196, |
| "step": 665 |
| }, |
| { |
| "epoch": 0.99, |
| "learning_rate": 2.9198591980705847e-09, |
| "loss": 1.1339, |
| "step": 666 |
| }, |
| { |
| "epoch": 0.99, |
| "learning_rate": 1.8687426271246646e-09, |
| "loss": 1.1454, |
| "step": 667 |
| }, |
| { |
| "epoch": 0.99, |
| "learning_rate": 1.0511820518432915e-09, |
| "loss": 1.1141, |
| "step": 668 |
| }, |
| { |
| "epoch": 1.0, |
| "learning_rate": 4.671965704128312e-10, |
| "loss": 1.2043, |
| "step": 669 |
| }, |
| { |
| "epoch": 1.0, |
| "learning_rate": 1.167998247131319e-10, |
| "loss": 1.1129, |
| "step": 670 |
| }, |
| { |
| "epoch": 1.0, |
| "learning_rate": 0.0, |
| "loss": 1.0999, |
| "step": 671 |
| }, |
| { |
| "epoch": 1.0, |
| "step": 671, |
| "total_flos": 2.178765965849998e+19, |
| "train_loss": 1.1755684873563876, |
| "train_runtime": 47161.9361, |
| "train_samples_per_second": 1.823, |
| "train_steps_per_second": 0.014 |
| } |
| ], |
| "logging_steps": 1.0, |
| "max_steps": 671, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 1, |
| "save_steps": 200, |
| "total_flos": 2.178765965849998e+19, |
| "train_batch_size": 2, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|