| { |
| "best_global_step": null, |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 0.2, |
| "eval_steps": 500, |
| "global_step": 2000, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "grad_norm": 1.6891276836395264, |
| "learning_rate": 1.8e-06, |
| "loss": 1.2585, |
| "step": 10 |
| }, |
| { |
| "grad_norm": 1.2421960830688477, |
| "learning_rate": 3.8e-06, |
| "loss": 1.2386, |
| "step": 20 |
| }, |
| { |
| "grad_norm": 1.0935646295547485, |
| "learning_rate": 5.8e-06, |
| "loss": 1.1761, |
| "step": 30 |
| }, |
| { |
| "grad_norm": 0.7210215330123901, |
| "learning_rate": 7.8e-06, |
| "loss": 1.0877, |
| "step": 40 |
| }, |
| { |
| "grad_norm": 0.34306979179382324, |
| "learning_rate": 9.800000000000001e-06, |
| "loss": 1.0339, |
| "step": 50 |
| }, |
| { |
| "grad_norm": 0.18134735524654388, |
| "learning_rate": 1.18e-05, |
| "loss": 1.0275, |
| "step": 60 |
| }, |
| { |
| "grad_norm": 0.15449777245521545, |
| "learning_rate": 1.3800000000000002e-05, |
| "loss": 1.0225, |
| "step": 70 |
| }, |
| { |
| "grad_norm": 0.15307500958442688, |
| "learning_rate": 1.58e-05, |
| "loss": 1.0188, |
| "step": 80 |
| }, |
| { |
| "grad_norm": 0.16959135234355927, |
| "learning_rate": 1.78e-05, |
| "loss": 1.0164, |
| "step": 90 |
| }, |
| { |
| "grad_norm": 0.1843302696943283, |
| "learning_rate": 1.9800000000000004e-05, |
| "loss": 1.019, |
| "step": 100 |
| }, |
| { |
| "grad_norm": 0.5004823803901672, |
| "learning_rate": 2.18e-05, |
| "loss": 1.007, |
| "step": 110 |
| }, |
| { |
| "grad_norm": 0.3533937931060791, |
| "learning_rate": 2.38e-05, |
| "loss": 0.9837, |
| "step": 120 |
| }, |
| { |
| "grad_norm": 0.492814302444458, |
| "learning_rate": 2.58e-05, |
| "loss": 0.95, |
| "step": 130 |
| }, |
| { |
| "grad_norm": 0.6054441332817078, |
| "learning_rate": 2.7800000000000005e-05, |
| "loss": 0.9005, |
| "step": 140 |
| }, |
| { |
| "grad_norm": 0.6737467646598816, |
| "learning_rate": 2.98e-05, |
| "loss": 0.8446, |
| "step": 150 |
| }, |
| { |
| "grad_norm": 1.3609671592712402, |
| "learning_rate": 3.18e-05, |
| "loss": 0.7917, |
| "step": 160 |
| }, |
| { |
| "grad_norm": 0.7027902603149414, |
| "learning_rate": 3.38e-05, |
| "loss": 0.7423, |
| "step": 170 |
| }, |
| { |
| "grad_norm": 0.6924164891242981, |
| "learning_rate": 3.58e-05, |
| "loss": 0.6849, |
| "step": 180 |
| }, |
| { |
| "grad_norm": 0.7687128186225891, |
| "learning_rate": 3.7800000000000004e-05, |
| "loss": 0.6385, |
| "step": 190 |
| }, |
| { |
| "grad_norm": 1.3768059015274048, |
| "learning_rate": 3.9800000000000005e-05, |
| "loss": 0.5934, |
| "step": 200 |
| }, |
| { |
| "grad_norm": 0.7467365860939026, |
| "learning_rate": 4.18e-05, |
| "loss": 0.5587, |
| "step": 210 |
| }, |
| { |
| "grad_norm": 0.9139049649238586, |
| "learning_rate": 4.38e-05, |
| "loss": 0.5173, |
| "step": 220 |
| }, |
| { |
| "grad_norm": 1.048671841621399, |
| "learning_rate": 4.58e-05, |
| "loss": 0.4742, |
| "step": 230 |
| }, |
| { |
| "grad_norm": 1.2771224975585938, |
| "learning_rate": 4.78e-05, |
| "loss": 0.4459, |
| "step": 240 |
| }, |
| { |
| "grad_norm": 1.6271154880523682, |
| "learning_rate": 4.9800000000000004e-05, |
| "loss": 0.4083, |
| "step": 250 |
| }, |
| { |
| "grad_norm": 1.4045976400375366, |
| "learning_rate": 5.1800000000000005e-05, |
| "loss": 0.3832, |
| "step": 260 |
| }, |
| { |
| "grad_norm": 1.031724214553833, |
| "learning_rate": 5.380000000000001e-05, |
| "loss": 0.3571, |
| "step": 270 |
| }, |
| { |
| "grad_norm": 0.9413377046585083, |
| "learning_rate": 5.580000000000001e-05, |
| "loss": 0.3294, |
| "step": 280 |
| }, |
| { |
| "grad_norm": 0.990053653717041, |
| "learning_rate": 5.7799999999999995e-05, |
| "loss": 0.3072, |
| "step": 290 |
| }, |
| { |
| "grad_norm": 0.9019013047218323, |
| "learning_rate": 5.9800000000000003e-05, |
| "loss": 0.2993, |
| "step": 300 |
| }, |
| { |
| "grad_norm": 0.9619300365447998, |
| "learning_rate": 6.18e-05, |
| "loss": 0.2908, |
| "step": 310 |
| }, |
| { |
| "grad_norm": 1.0667186975479126, |
| "learning_rate": 6.38e-05, |
| "loss": 0.2681, |
| "step": 320 |
| }, |
| { |
| "grad_norm": 1.119122862815857, |
| "learning_rate": 6.58e-05, |
| "loss": 0.2608, |
| "step": 330 |
| }, |
| { |
| "grad_norm": 1.1136410236358643, |
| "learning_rate": 6.780000000000001e-05, |
| "loss": 0.2568, |
| "step": 340 |
| }, |
| { |
| "grad_norm": 1.5547524690628052, |
| "learning_rate": 6.98e-05, |
| "loss": 0.2301, |
| "step": 350 |
| }, |
| { |
| "grad_norm": 1.0856785774230957, |
| "learning_rate": 7.18e-05, |
| "loss": 0.2166, |
| "step": 360 |
| }, |
| { |
| "grad_norm": 0.9325472712516785, |
| "learning_rate": 7.38e-05, |
| "loss": 0.1965, |
| "step": 370 |
| }, |
| { |
| "grad_norm": 0.9351710081100464, |
| "learning_rate": 7.58e-05, |
| "loss": 0.1845, |
| "step": 380 |
| }, |
| { |
| "grad_norm": 1.274675726890564, |
| "learning_rate": 7.780000000000001e-05, |
| "loss": 0.1701, |
| "step": 390 |
| }, |
| { |
| "grad_norm": 1.1908860206604004, |
| "learning_rate": 7.98e-05, |
| "loss": 0.1534, |
| "step": 400 |
| }, |
| { |
| "grad_norm": 1.1736817359924316, |
| "learning_rate": 8.18e-05, |
| "loss": 0.1455, |
| "step": 410 |
| }, |
| { |
| "grad_norm": 1.7282778024673462, |
| "learning_rate": 8.38e-05, |
| "loss": 0.1299, |
| "step": 420 |
| }, |
| { |
| "grad_norm": 1.0969237089157104, |
| "learning_rate": 8.58e-05, |
| "loss": 0.1252, |
| "step": 430 |
| }, |
| { |
| "grad_norm": 1.1998040676116943, |
| "learning_rate": 8.78e-05, |
| "loss": 0.1132, |
| "step": 440 |
| }, |
| { |
| "grad_norm": 1.1225069761276245, |
| "learning_rate": 8.98e-05, |
| "loss": 0.1111, |
| "step": 450 |
| }, |
| { |
| "grad_norm": 1.034788727760315, |
| "learning_rate": 9.180000000000001e-05, |
| "loss": 0.104, |
| "step": 460 |
| }, |
| { |
| "grad_norm": 1.0320745706558228, |
| "learning_rate": 9.38e-05, |
| "loss": 0.0981, |
| "step": 470 |
| }, |
| { |
| "grad_norm": 0.8197027444839478, |
| "learning_rate": 9.58e-05, |
| "loss": 0.0934, |
| "step": 480 |
| }, |
| { |
| "grad_norm": 1.0704108476638794, |
| "learning_rate": 9.78e-05, |
| "loss": 0.0844, |
| "step": 490 |
| }, |
| { |
| "grad_norm": 1.1035730838775635, |
| "learning_rate": 9.98e-05, |
| "loss": 0.0815, |
| "step": 500 |
| }, |
| { |
| "grad_norm": 0.8952515721321106, |
| "learning_rate": 9.9999778549206e-05, |
| "loss": 0.0764, |
| "step": 510 |
| }, |
| { |
| "grad_norm": 1.274247407913208, |
| "learning_rate": 9.999901304280685e-05, |
| "loss": 0.0774, |
| "step": 520 |
| }, |
| { |
| "grad_norm": 0.9264320135116577, |
| "learning_rate": 9.999770075521164e-05, |
| "loss": 0.0783, |
| "step": 530 |
| }, |
| { |
| "grad_norm": 0.8261769413948059, |
| "learning_rate": 9.99958417007713e-05, |
| "loss": 0.0771, |
| "step": 540 |
| }, |
| { |
| "grad_norm": 0.9270550608634949, |
| "learning_rate": 9.999343589981615e-05, |
| "loss": 0.0661, |
| "step": 550 |
| }, |
| { |
| "grad_norm": 1.1616880893707275, |
| "learning_rate": 9.999048337865568e-05, |
| "loss": 0.0755, |
| "step": 560 |
| }, |
| { |
| "grad_norm": 0.9049257636070251, |
| "learning_rate": 9.998698416957815e-05, |
| "loss": 0.0765, |
| "step": 570 |
| }, |
| { |
| "grad_norm": 1.1724106073379517, |
| "learning_rate": 9.998293831085037e-05, |
| "loss": 0.0712, |
| "step": 580 |
| }, |
| { |
| "grad_norm": 0.8528197407722473, |
| "learning_rate": 9.997834584671719e-05, |
| "loss": 0.0684, |
| "step": 590 |
| }, |
| { |
| "grad_norm": 0.9542710781097412, |
| "learning_rate": 9.997320682740107e-05, |
| "loss": 0.0716, |
| "step": 600 |
| }, |
| { |
| "grad_norm": 0.7609046101570129, |
| "learning_rate": 9.996752130910149e-05, |
| "loss": 0.0672, |
| "step": 610 |
| }, |
| { |
| "grad_norm": 0.770867109298706, |
| "learning_rate": 9.99612893539944e-05, |
| "loss": 0.0657, |
| "step": 620 |
| }, |
| { |
| "grad_norm": 0.5650806427001953, |
| "learning_rate": 9.995451103023144e-05, |
| "loss": 0.064, |
| "step": 630 |
| }, |
| { |
| "grad_norm": 0.9630634784698486, |
| "learning_rate": 9.994718641193928e-05, |
| "loss": 0.0668, |
| "step": 640 |
| }, |
| { |
| "grad_norm": 0.7161832451820374, |
| "learning_rate": 9.993931557921874e-05, |
| "loss": 0.0659, |
| "step": 650 |
| }, |
| { |
| "grad_norm": 0.9892486333847046, |
| "learning_rate": 9.993089861814402e-05, |
| "loss": 0.0627, |
| "step": 660 |
| }, |
| { |
| "grad_norm": 0.7624261379241943, |
| "learning_rate": 9.992193562076166e-05, |
| "loss": 0.0597, |
| "step": 670 |
| }, |
| { |
| "grad_norm": 0.8515845537185669, |
| "learning_rate": 9.991242668508954e-05, |
| "loss": 0.0556, |
| "step": 680 |
| }, |
| { |
| "grad_norm": 0.9307328462600708, |
| "learning_rate": 9.990237191511587e-05, |
| "loss": 0.053, |
| "step": 690 |
| }, |
| { |
| "grad_norm": 1.0245169401168823, |
| "learning_rate": 9.989177142079802e-05, |
| "loss": 0.0567, |
| "step": 700 |
| }, |
| { |
| "grad_norm": 0.8914056420326233, |
| "learning_rate": 9.988062531806126e-05, |
| "loss": 0.0565, |
| "step": 710 |
| }, |
| { |
| "grad_norm": 0.9028451442718506, |
| "learning_rate": 9.986893372879762e-05, |
| "loss": 0.0575, |
| "step": 720 |
| }, |
| { |
| "grad_norm": 0.8873146772384644, |
| "learning_rate": 9.985669678086443e-05, |
| "loss": 0.0602, |
| "step": 730 |
| }, |
| { |
| "grad_norm": 0.7956727147102356, |
| "learning_rate": 9.984391460808298e-05, |
| "loss": 0.0602, |
| "step": 740 |
| }, |
| { |
| "grad_norm": 0.7061982154846191, |
| "learning_rate": 9.983058735023709e-05, |
| "loss": 0.0587, |
| "step": 750 |
| }, |
| { |
| "grad_norm": 0.8428799510002136, |
| "learning_rate": 9.98167151530715e-05, |
| "loss": 0.0552, |
| "step": 760 |
| }, |
| { |
| "grad_norm": 0.9017972946166992, |
| "learning_rate": 9.980229816829034e-05, |
| "loss": 0.0595, |
| "step": 770 |
| }, |
| { |
| "grad_norm": 0.8163244128227234, |
| "learning_rate": 9.978733655355544e-05, |
| "loss": 0.057, |
| "step": 780 |
| }, |
| { |
| "grad_norm": 0.7705178260803223, |
| "learning_rate": 9.977183047248464e-05, |
| "loss": 0.0512, |
| "step": 790 |
| }, |
| { |
| "grad_norm": 1.0152888298034668, |
| "learning_rate": 9.975578009464992e-05, |
| "loss": 0.0532, |
| "step": 800 |
| }, |
| { |
| "grad_norm": 0.7115610241889954, |
| "learning_rate": 9.97391855955757e-05, |
| "loss": 0.0475, |
| "step": 810 |
| }, |
| { |
| "grad_norm": 0.8815870881080627, |
| "learning_rate": 9.972204715673669e-05, |
| "loss": 0.0495, |
| "step": 820 |
| }, |
| { |
| "grad_norm": 0.6733144521713257, |
| "learning_rate": 9.970436496555617e-05, |
| "loss": 0.05, |
| "step": 830 |
| }, |
| { |
| "grad_norm": 0.8259405493736267, |
| "learning_rate": 9.968613921540373e-05, |
| "loss": 0.0519, |
| "step": 840 |
| }, |
| { |
| "grad_norm": 0.7688018679618835, |
| "learning_rate": 9.966737010559326e-05, |
| "loss": 0.0525, |
| "step": 850 |
| }, |
| { |
| "grad_norm": 0.6979959607124329, |
| "learning_rate": 9.964805784138072e-05, |
| "loss": 0.0511, |
| "step": 860 |
| }, |
| { |
| "grad_norm": 0.7118434309959412, |
| "learning_rate": 9.962820263396195e-05, |
| "loss": 0.0482, |
| "step": 870 |
| }, |
| { |
| "grad_norm": 0.8046584129333496, |
| "learning_rate": 9.960780470047033e-05, |
| "loss": 0.0522, |
| "step": 880 |
| }, |
| { |
| "grad_norm": 0.6603678464889526, |
| "learning_rate": 9.958686426397437e-05, |
| "loss": 0.0499, |
| "step": 890 |
| }, |
| { |
| "grad_norm": 0.7854413986206055, |
| "learning_rate": 9.956538155347534e-05, |
| "loss": 0.052, |
| "step": 900 |
| }, |
| { |
| "grad_norm": 0.5703814029693604, |
| "learning_rate": 9.95433568039047e-05, |
| "loss": 0.0455, |
| "step": 910 |
| }, |
| { |
| "grad_norm": 0.8428476452827454, |
| "learning_rate": 9.952079025612162e-05, |
| "loss": 0.0443, |
| "step": 920 |
| }, |
| { |
| "grad_norm": 0.7631692886352539, |
| "learning_rate": 9.949768215691022e-05, |
| "loss": 0.0468, |
| "step": 930 |
| }, |
| { |
| "grad_norm": 0.7209548950195312, |
| "learning_rate": 9.9474032758977e-05, |
| "loss": 0.0474, |
| "step": 940 |
| }, |
| { |
| "grad_norm": 0.6969509124755859, |
| "learning_rate": 9.944984232094794e-05, |
| "loss": 0.0501, |
| "step": 950 |
| }, |
| { |
| "grad_norm": 0.7608206272125244, |
| "learning_rate": 9.942511110736584e-05, |
| "loss": 0.0471, |
| "step": 960 |
| }, |
| { |
| "grad_norm": 0.7608780860900879, |
| "learning_rate": 9.939983938868726e-05, |
| "loss": 0.0476, |
| "step": 970 |
| }, |
| { |
| "grad_norm": 0.6730809807777405, |
| "learning_rate": 9.93740274412797e-05, |
| "loss": 0.0431, |
| "step": 980 |
| }, |
| { |
| "grad_norm": 0.8944201469421387, |
| "learning_rate": 9.934767554741846e-05, |
| "loss": 0.0511, |
| "step": 990 |
| }, |
| { |
| "grad_norm": 0.8130436539649963, |
| "learning_rate": 9.932078399528361e-05, |
| "loss": 0.0424, |
| "step": 1000 |
| }, |
| { |
| "grad_norm": 0.7339826822280884, |
| "learning_rate": 9.929335307895689e-05, |
| "loss": 0.043, |
| "step": 1010 |
| }, |
| { |
| "grad_norm": 0.7107013463973999, |
| "learning_rate": 9.926538309841839e-05, |
| "loss": 0.0498, |
| "step": 1020 |
| }, |
| { |
| "grad_norm": 0.7686504125595093, |
| "learning_rate": 9.923687435954334e-05, |
| "loss": 0.0417, |
| "step": 1030 |
| }, |
| { |
| "grad_norm": 0.5518864989280701, |
| "learning_rate": 9.920782717409873e-05, |
| "loss": 0.0397, |
| "step": 1040 |
| }, |
| { |
| "grad_norm": 0.4921113848686218, |
| "learning_rate": 9.917824185973994e-05, |
| "loss": 0.0386, |
| "step": 1050 |
| }, |
| { |
| "grad_norm": 0.5414538979530334, |
| "learning_rate": 9.914811874000723e-05, |
| "loss": 0.0383, |
| "step": 1060 |
| }, |
| { |
| "grad_norm": 0.7523898482322693, |
| "learning_rate": 9.911745814432218e-05, |
| "loss": 0.0422, |
| "step": 1070 |
| }, |
| { |
| "grad_norm": 0.5954861640930176, |
| "learning_rate": 9.90862604079842e-05, |
| "loss": 0.0442, |
| "step": 1080 |
| }, |
| { |
| "grad_norm": 0.7658764719963074, |
| "learning_rate": 9.90545258721667e-05, |
| "loss": 0.0417, |
| "step": 1090 |
| }, |
| { |
| "grad_norm": 0.6343804001808167, |
| "learning_rate": 9.90222548839135e-05, |
| "loss": 0.0401, |
| "step": 1100 |
| }, |
| { |
| "grad_norm": 0.4654010534286499, |
| "learning_rate": 9.898944779613495e-05, |
| "loss": 0.0372, |
| "step": 1110 |
| }, |
| { |
| "grad_norm": 0.7204039096832275, |
| "learning_rate": 9.89561049676041e-05, |
| "loss": 0.0399, |
| "step": 1120 |
| }, |
| { |
| "grad_norm": 0.6943351626396179, |
| "learning_rate": 9.89222267629528e-05, |
| "loss": 0.0425, |
| "step": 1130 |
| }, |
| { |
| "grad_norm": 0.7369424104690552, |
| "learning_rate": 9.888781355266763e-05, |
| "loss": 0.0418, |
| "step": 1140 |
| }, |
| { |
| "grad_norm": 0.6650173664093018, |
| "learning_rate": 9.885286571308598e-05, |
| "loss": 0.0382, |
| "step": 1150 |
| }, |
| { |
| "grad_norm": 0.5390568375587463, |
| "learning_rate": 9.881738362639182e-05, |
| "loss": 0.0415, |
| "step": 1160 |
| }, |
| { |
| "grad_norm": 0.8118464946746826, |
| "learning_rate": 9.878136768061154e-05, |
| "loss": 0.0432, |
| "step": 1170 |
| }, |
| { |
| "grad_norm": 0.7557201981544495, |
| "learning_rate": 9.874481826960979e-05, |
| "loss": 0.0424, |
| "step": 1180 |
| }, |
| { |
| "grad_norm": 0.6853317022323608, |
| "learning_rate": 9.870773579308503e-05, |
| "loss": 0.0398, |
| "step": 1190 |
| }, |
| { |
| "grad_norm": 0.6386889815330505, |
| "learning_rate": 9.867012065656533e-05, |
| "loss": 0.0427, |
| "step": 1200 |
| }, |
| { |
| "grad_norm": 0.706305980682373, |
| "learning_rate": 9.863197327140376e-05, |
| "loss": 0.0375, |
| "step": 1210 |
| }, |
| { |
| "grad_norm": 0.7585138082504272, |
| "learning_rate": 9.859329405477403e-05, |
| "loss": 0.038, |
| "step": 1220 |
| }, |
| { |
| "grad_norm": 0.6497421264648438, |
| "learning_rate": 9.855408342966585e-05, |
| "loss": 0.0397, |
| "step": 1230 |
| }, |
| { |
| "grad_norm": 0.5648456811904907, |
| "learning_rate": 9.851434182488033e-05, |
| "loss": 0.038, |
| "step": 1240 |
| }, |
| { |
| "grad_norm": 0.7392773628234863, |
| "learning_rate": 9.84740696750253e-05, |
| "loss": 0.0375, |
| "step": 1250 |
| }, |
| { |
| "grad_norm": 0.7451732754707336, |
| "learning_rate": 9.843326742051055e-05, |
| "loss": 0.0384, |
| "step": 1260 |
| }, |
| { |
| "grad_norm": 0.5808190703392029, |
| "learning_rate": 9.839193550754297e-05, |
| "loss": 0.0374, |
| "step": 1270 |
| }, |
| { |
| "grad_norm": 0.7281758785247803, |
| "learning_rate": 9.835007438812177e-05, |
| "loss": 0.0418, |
| "step": 1280 |
| }, |
| { |
| "grad_norm": 0.6007897257804871, |
| "learning_rate": 9.830768452003341e-05, |
| "loss": 0.0375, |
| "step": 1290 |
| }, |
| { |
| "grad_norm": 0.6821262836456299, |
| "learning_rate": 9.826476636684671e-05, |
| "loss": 0.0369, |
| "step": 1300 |
| }, |
| { |
| "grad_norm": 0.6722646951675415, |
| "learning_rate": 9.822132039790773e-05, |
| "loss": 0.0368, |
| "step": 1310 |
| }, |
| { |
| "grad_norm": 0.5557746887207031, |
| "learning_rate": 9.817734708833461e-05, |
| "loss": 0.0337, |
| "step": 1320 |
| }, |
| { |
| "grad_norm": 0.6717876195907593, |
| "learning_rate": 9.813284691901243e-05, |
| "loss": 0.0423, |
| "step": 1330 |
| }, |
| { |
| "grad_norm": 0.6293877363204956, |
| "learning_rate": 9.808782037658792e-05, |
| "loss": 0.0408, |
| "step": 1340 |
| }, |
| { |
| "grad_norm": 0.6186095476150513, |
| "learning_rate": 9.804226795346411e-05, |
| "loss": 0.0382, |
| "step": 1350 |
| }, |
| { |
| "grad_norm": 0.8207311630249023, |
| "learning_rate": 9.799619014779503e-05, |
| "loss": 0.0371, |
| "step": 1360 |
| }, |
| { |
| "grad_norm": 0.5756433010101318, |
| "learning_rate": 9.794958746348013e-05, |
| "loss": 0.0384, |
| "step": 1370 |
| }, |
| { |
| "grad_norm": 0.653166651725769, |
| "learning_rate": 9.790246041015896e-05, |
| "loss": 0.034, |
| "step": 1380 |
| }, |
| { |
| "grad_norm": 0.6604922413825989, |
| "learning_rate": 9.785480950320538e-05, |
| "loss": 0.0347, |
| "step": 1390 |
| }, |
| { |
| "grad_norm": 0.7043635249137878, |
| "learning_rate": 9.78066352637221e-05, |
| "loss": 0.0338, |
| "step": 1400 |
| }, |
| { |
| "grad_norm": 0.6509787440299988, |
| "learning_rate": 9.775793821853488e-05, |
| "loss": 0.0355, |
| "step": 1410 |
| }, |
| { |
| "grad_norm": 0.6967816948890686, |
| "learning_rate": 9.77087189001868e-05, |
| "loss": 0.0353, |
| "step": 1420 |
| }, |
| { |
| "grad_norm": 0.8638598322868347, |
| "learning_rate": 9.765897784693243e-05, |
| "loss": 0.0361, |
| "step": 1430 |
| }, |
| { |
| "grad_norm": 0.6340733170509338, |
| "learning_rate": 9.760871560273197e-05, |
| "loss": 0.0351, |
| "step": 1440 |
| }, |
| { |
| "grad_norm": 0.6075161695480347, |
| "learning_rate": 9.755793271724526e-05, |
| "loss": 0.0342, |
| "step": 1450 |
| }, |
| { |
| "grad_norm": 0.6330815553665161, |
| "learning_rate": 9.750662974582584e-05, |
| "loss": 0.0406, |
| "step": 1460 |
| }, |
| { |
| "grad_norm": 0.5282254815101624, |
| "learning_rate": 9.745480724951473e-05, |
| "loss": 0.0351, |
| "step": 1470 |
| }, |
| { |
| "grad_norm": 0.7070049047470093, |
| "learning_rate": 9.740246579503447e-05, |
| "loss": 0.0366, |
| "step": 1480 |
| }, |
| { |
| "grad_norm": 0.7055975198745728, |
| "learning_rate": 9.734960595478284e-05, |
| "loss": 0.0357, |
| "step": 1490 |
| }, |
| { |
| "grad_norm": 0.6384962797164917, |
| "learning_rate": 9.729622830682657e-05, |
| "loss": 0.0343, |
| "step": 1500 |
| }, |
| { |
| "grad_norm": 0.5874096751213074, |
| "learning_rate": 9.724233343489504e-05, |
| "loss": 0.0372, |
| "step": 1510 |
| }, |
| { |
| "grad_norm": 0.5690801739692688, |
| "learning_rate": 9.718792192837396e-05, |
| "loss": 0.0319, |
| "step": 1520 |
| }, |
| { |
| "grad_norm": 0.6367931962013245, |
| "learning_rate": 9.713299438229886e-05, |
| "loss": 0.037, |
| "step": 1530 |
| }, |
| { |
| "grad_norm": 0.5469662547111511, |
| "learning_rate": 9.707755139734855e-05, |
| "loss": 0.0365, |
| "step": 1540 |
| }, |
| { |
| "grad_norm": 0.556199848651886, |
| "learning_rate": 9.702159357983866e-05, |
| "loss": 0.0342, |
| "step": 1550 |
| }, |
| { |
| "grad_norm": 0.6330277919769287, |
| "learning_rate": 9.696512154171492e-05, |
| "loss": 0.0363, |
| "step": 1560 |
| }, |
| { |
| "grad_norm": 0.6156578660011292, |
| "learning_rate": 9.690813590054645e-05, |
| "loss": 0.036, |
| "step": 1570 |
| }, |
| { |
| "grad_norm": 0.633952260017395, |
| "learning_rate": 9.685063727951914e-05, |
| "loss": 0.0342, |
| "step": 1580 |
| }, |
| { |
| "grad_norm": 0.4720134735107422, |
| "learning_rate": 9.679262630742865e-05, |
| "loss": 0.0352, |
| "step": 1590 |
| }, |
| { |
| "grad_norm": 0.6747561097145081, |
| "learning_rate": 9.673410361867373e-05, |
| "loss": 0.0361, |
| "step": 1600 |
| }, |
| { |
| "grad_norm": 0.5455501675605774, |
| "learning_rate": 9.667506985324909e-05, |
| "loss": 0.0332, |
| "step": 1610 |
| }, |
| { |
| "grad_norm": 0.698010265827179, |
| "learning_rate": 9.661552565673855e-05, |
| "loss": 0.0309, |
| "step": 1620 |
| }, |
| { |
| "grad_norm": 0.6324073672294617, |
| "learning_rate": 9.655547168030789e-05, |
| "loss": 0.035, |
| "step": 1630 |
| }, |
| { |
| "grad_norm": 0.6724660992622375, |
| "learning_rate": 9.649490858069777e-05, |
| "loss": 0.0327, |
| "step": 1640 |
| }, |
| { |
| "grad_norm": 0.4986797273159027, |
| "learning_rate": 9.643383702021658e-05, |
| "loss": 0.032, |
| "step": 1650 |
| }, |
| { |
| "grad_norm": 0.6412114500999451, |
| "learning_rate": 9.637225766673307e-05, |
| "loss": 0.0309, |
| "step": 1660 |
| }, |
| { |
| "grad_norm": 0.5809900164604187, |
| "learning_rate": 9.631017119366922e-05, |
| "loss": 0.0319, |
| "step": 1670 |
| }, |
| { |
| "grad_norm": 0.5835633277893066, |
| "learning_rate": 9.624757827999273e-05, |
| "loss": 0.0339, |
| "step": 1680 |
| }, |
| { |
| "grad_norm": 0.5545461773872375, |
| "learning_rate": 9.618447961020971e-05, |
| "loss": 0.039, |
| "step": 1690 |
| }, |
| { |
| "grad_norm": 0.6023017168045044, |
| "learning_rate": 9.612087587435707e-05, |
| "loss": 0.038, |
| "step": 1700 |
| }, |
| { |
| "grad_norm": 0.6422150135040283, |
| "learning_rate": 9.605676776799508e-05, |
| "loss": 0.0342, |
| "step": 1710 |
| }, |
| { |
| "grad_norm": 0.6100661158561707, |
| "learning_rate": 9.599215599219973e-05, |
| "loss": 0.0332, |
| "step": 1720 |
| }, |
| { |
| "grad_norm": 0.6025664806365967, |
| "learning_rate": 9.592704125355505e-05, |
| "loss": 0.0332, |
| "step": 1730 |
| }, |
| { |
| "grad_norm": 0.5365665555000305, |
| "learning_rate": 9.586142426414538e-05, |
| "loss": 0.0322, |
| "step": 1740 |
| }, |
| { |
| "grad_norm": 0.6416350603103638, |
| "learning_rate": 9.57953057415476e-05, |
| "loss": 0.0356, |
| "step": 1750 |
| }, |
| { |
| "grad_norm": 0.5091562867164612, |
| "learning_rate": 9.572868640882328e-05, |
| "loss": 0.0327, |
| "step": 1760 |
| }, |
| { |
| "grad_norm": 0.506897509098053, |
| "learning_rate": 9.56615669945108e-05, |
| "loss": 0.0347, |
| "step": 1770 |
| }, |
| { |
| "grad_norm": 0.5713956356048584, |
| "learning_rate": 9.55939482326173e-05, |
| "loss": 0.0324, |
| "step": 1780 |
| }, |
| { |
| "grad_norm": 0.47305065393447876, |
| "learning_rate": 9.552583086261069e-05, |
| "loss": 0.034, |
| "step": 1790 |
| }, |
| { |
| "grad_norm": 0.5677897334098816, |
| "learning_rate": 9.545721562941168e-05, |
| "loss": 0.0319, |
| "step": 1800 |
| }, |
| { |
| "grad_norm": 0.5385820865631104, |
| "learning_rate": 9.538810328338543e-05, |
| "loss": 0.0306, |
| "step": 1810 |
| }, |
| { |
| "grad_norm": 0.67743319272995, |
| "learning_rate": 9.531849458033349e-05, |
| "loss": 0.0336, |
| "step": 1820 |
| }, |
| { |
| "grad_norm": 0.6273674368858337, |
| "learning_rate": 9.524839028148547e-05, |
| "loss": 0.0325, |
| "step": 1830 |
| }, |
| { |
| "grad_norm": 0.5729127526283264, |
| "learning_rate": 9.517779115349077e-05, |
| "loss": 0.033, |
| "step": 1840 |
| }, |
| { |
| "grad_norm": 0.5738729238510132, |
| "learning_rate": 9.510669796841014e-05, |
| "loss": 0.0327, |
| "step": 1850 |
| }, |
| { |
| "grad_norm": 0.46439313888549805, |
| "learning_rate": 9.503511150370727e-05, |
| "loss": 0.0349, |
| "step": 1860 |
| }, |
| { |
| "grad_norm": 0.49889275431632996, |
| "learning_rate": 9.496303254224024e-05, |
| "loss": 0.0329, |
| "step": 1870 |
| }, |
| { |
| "grad_norm": 0.466863751411438, |
| "learning_rate": 9.489046187225306e-05, |
| "loss": 0.032, |
| "step": 1880 |
| }, |
| { |
| "grad_norm": 0.49700266122817993, |
| "learning_rate": 9.481740028736692e-05, |
| "loss": 0.0335, |
| "step": 1890 |
| }, |
| { |
| "grad_norm": 0.6205595135688782, |
| "learning_rate": 9.474384858657164e-05, |
| "loss": 0.0315, |
| "step": 1900 |
| }, |
| { |
| "grad_norm": 0.5160883665084839, |
| "learning_rate": 9.466980757421679e-05, |
| "loss": 0.0313, |
| "step": 1910 |
| }, |
| { |
| "grad_norm": 0.5717476010322571, |
| "learning_rate": 9.459527806000305e-05, |
| "loss": 0.0333, |
| "step": 1920 |
| }, |
| { |
| "grad_norm": 0.5541268587112427, |
| "learning_rate": 9.452026085897325e-05, |
| "loss": 0.0327, |
| "step": 1930 |
| }, |
| { |
| "grad_norm": 0.5027980208396912, |
| "learning_rate": 9.444475679150348e-05, |
| "loss": 0.0314, |
| "step": 1940 |
| }, |
| { |
| "grad_norm": 0.5508278608322144, |
| "learning_rate": 9.436876668329411e-05, |
| "loss": 0.0303, |
| "step": 1950 |
| }, |
| { |
| "grad_norm": 0.5171688199043274, |
| "learning_rate": 9.429229136536079e-05, |
| "loss": 0.0298, |
| "step": 1960 |
| }, |
| { |
| "grad_norm": 0.6098824143409729, |
| "learning_rate": 9.421533167402534e-05, |
| "loss": 0.0314, |
| "step": 1970 |
| }, |
| { |
| "grad_norm": 0.5086256265640259, |
| "learning_rate": 9.413788845090666e-05, |
| "loss": 0.0317, |
| "step": 1980 |
| }, |
| { |
| "grad_norm": 0.6548255681991577, |
| "learning_rate": 9.405996254291136e-05, |
| "loss": 0.0311, |
| "step": 1990 |
| }, |
| { |
| "grad_norm": 0.6564981341362, |
| "learning_rate": 9.398155480222474e-05, |
| "loss": 0.0309, |
| "step": 2000 |
| } |
| ], |
| "logging_steps": 10, |
| "max_steps": 10000, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 9223372036854775807, |
| "save_steps": 1000, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": false |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 0.0, |
| "train_batch_size": 64, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|