| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 3.0, |
| "eval_steps": 500, |
| "global_step": 3189, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.01, |
| "grad_norm": 20.48359489440918, |
| "learning_rate": 1.8796992481203008e-07, |
| "loss": 0.6209, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.02, |
| "grad_norm": 5.289457321166992, |
| "learning_rate": 3.7593984962406015e-07, |
| "loss": 0.2829, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.03, |
| "grad_norm": 0.8980396389961243, |
| "learning_rate": 5.639097744360903e-07, |
| "loss": 0.0863, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.04, |
| "grad_norm": 0.25141441822052, |
| "learning_rate": 7.518796992481203e-07, |
| "loss": 0.0941, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.05, |
| "grad_norm": 0.20232008397579193, |
| "learning_rate": 9.398496240601504e-07, |
| "loss": 0.0735, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.06, |
| "grad_norm": 1.138559341430664, |
| "learning_rate": 1.1278195488721805e-06, |
| "loss": 0.0802, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.07, |
| "grad_norm": 0.6884934306144714, |
| "learning_rate": 1.3157894736842106e-06, |
| "loss": 0.0741, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.08, |
| "grad_norm": 1.3850356340408325, |
| "learning_rate": 1.5037593984962406e-06, |
| "loss": 0.0631, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.08, |
| "grad_norm": 0.5666308403015137, |
| "learning_rate": 1.6917293233082707e-06, |
| "loss": 0.06, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.09, |
| "grad_norm": 0.5473531484603882, |
| "learning_rate": 1.8796992481203007e-06, |
| "loss": 0.0475, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.1, |
| "grad_norm": 0.7816307544708252, |
| "learning_rate": 2.067669172932331e-06, |
| "loss": 0.0582, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.11, |
| "grad_norm": 0.9976327419281006, |
| "learning_rate": 2.255639097744361e-06, |
| "loss": 0.0439, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.12, |
| "grad_norm": 1.050405740737915, |
| "learning_rate": 2.443609022556391e-06, |
| "loss": 0.0439, |
| "step": 130 |
| }, |
| { |
| "epoch": 0.13, |
| "grad_norm": 0.8269789814949036, |
| "learning_rate": 2.631578947368421e-06, |
| "loss": 0.059, |
| "step": 140 |
| }, |
| { |
| "epoch": 0.14, |
| "grad_norm": 0.14458781480789185, |
| "learning_rate": 2.8195488721804516e-06, |
| "loss": 0.0366, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.15, |
| "grad_norm": 1.4295405149459839, |
| "learning_rate": 3.007518796992481e-06, |
| "loss": 0.0602, |
| "step": 160 |
| }, |
| { |
| "epoch": 0.16, |
| "grad_norm": 1.3261005878448486, |
| "learning_rate": 3.1954887218045117e-06, |
| "loss": 0.0449, |
| "step": 170 |
| }, |
| { |
| "epoch": 0.17, |
| "grad_norm": 2.0573761463165283, |
| "learning_rate": 3.3834586466165413e-06, |
| "loss": 0.0564, |
| "step": 180 |
| }, |
| { |
| "epoch": 0.18, |
| "grad_norm": 0.9643514156341553, |
| "learning_rate": 3.5714285714285718e-06, |
| "loss": 0.0559, |
| "step": 190 |
| }, |
| { |
| "epoch": 0.19, |
| "grad_norm": 0.8114656805992126, |
| "learning_rate": 3.7593984962406014e-06, |
| "loss": 0.0402, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.2, |
| "grad_norm": 0.7962890267372131, |
| "learning_rate": 3.947368421052632e-06, |
| "loss": 0.0486, |
| "step": 210 |
| }, |
| { |
| "epoch": 0.21, |
| "grad_norm": 1.7510570287704468, |
| "learning_rate": 4.135338345864662e-06, |
| "loss": 0.0444, |
| "step": 220 |
| }, |
| { |
| "epoch": 0.22, |
| "grad_norm": 0.9593592882156372, |
| "learning_rate": 4.323308270676692e-06, |
| "loss": 0.0374, |
| "step": 230 |
| }, |
| { |
| "epoch": 0.23, |
| "grad_norm": 0.5025224089622498, |
| "learning_rate": 4.511278195488722e-06, |
| "loss": 0.0337, |
| "step": 240 |
| }, |
| { |
| "epoch": 0.24, |
| "grad_norm": 1.6147772073745728, |
| "learning_rate": 4.6992481203007525e-06, |
| "loss": 0.035, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.24, |
| "grad_norm": 1.0601171255111694, |
| "learning_rate": 4.887218045112782e-06, |
| "loss": 0.0445, |
| "step": 260 |
| }, |
| { |
| "epoch": 0.25, |
| "grad_norm": 1.0513237714767456, |
| "learning_rate": 5.075187969924813e-06, |
| "loss": 0.0495, |
| "step": 270 |
| }, |
| { |
| "epoch": 0.26, |
| "grad_norm": 1.3772203922271729, |
| "learning_rate": 5.263157894736842e-06, |
| "loss": 0.061, |
| "step": 280 |
| }, |
| { |
| "epoch": 0.27, |
| "grad_norm": 0.48920339345932007, |
| "learning_rate": 5.451127819548873e-06, |
| "loss": 0.0492, |
| "step": 290 |
| }, |
| { |
| "epoch": 0.28, |
| "grad_norm": 1.1981807947158813, |
| "learning_rate": 5.639097744360903e-06, |
| "loss": 0.0432, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.29, |
| "grad_norm": 0.28824716806411743, |
| "learning_rate": 5.827067669172934e-06, |
| "loss": 0.0412, |
| "step": 310 |
| }, |
| { |
| "epoch": 0.3, |
| "grad_norm": 0.5675559639930725, |
| "learning_rate": 6.015037593984962e-06, |
| "loss": 0.0419, |
| "step": 320 |
| }, |
| { |
| "epoch": 0.31, |
| "grad_norm": 0.18200090527534485, |
| "learning_rate": 6.203007518796993e-06, |
| "loss": 0.0278, |
| "step": 330 |
| }, |
| { |
| "epoch": 0.32, |
| "grad_norm": 1.1105334758758545, |
| "learning_rate": 6.390977443609023e-06, |
| "loss": 0.0454, |
| "step": 340 |
| }, |
| { |
| "epoch": 0.33, |
| "grad_norm": 0.6391922235488892, |
| "learning_rate": 6.578947368421054e-06, |
| "loss": 0.0343, |
| "step": 350 |
| }, |
| { |
| "epoch": 0.34, |
| "grad_norm": 0.5360444784164429, |
| "learning_rate": 6.766917293233083e-06, |
| "loss": 0.036, |
| "step": 360 |
| }, |
| { |
| "epoch": 0.35, |
| "grad_norm": 0.8997634053230286, |
| "learning_rate": 6.954887218045113e-06, |
| "loss": 0.033, |
| "step": 370 |
| }, |
| { |
| "epoch": 0.36, |
| "grad_norm": 0.6569644808769226, |
| "learning_rate": 7.1428571428571436e-06, |
| "loss": 0.0349, |
| "step": 380 |
| }, |
| { |
| "epoch": 0.37, |
| "grad_norm": 1.5074034929275513, |
| "learning_rate": 7.330827067669174e-06, |
| "loss": 0.0358, |
| "step": 390 |
| }, |
| { |
| "epoch": 0.38, |
| "grad_norm": 0.5094283819198608, |
| "learning_rate": 7.518796992481203e-06, |
| "loss": 0.0331, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.39, |
| "grad_norm": 0.9734311103820801, |
| "learning_rate": 7.706766917293233e-06, |
| "loss": 0.0507, |
| "step": 410 |
| }, |
| { |
| "epoch": 0.4, |
| "grad_norm": 0.4237803518772125, |
| "learning_rate": 7.894736842105265e-06, |
| "loss": 0.041, |
| "step": 420 |
| }, |
| { |
| "epoch": 0.4, |
| "grad_norm": 0.514431893825531, |
| "learning_rate": 8.082706766917294e-06, |
| "loss": 0.0476, |
| "step": 430 |
| }, |
| { |
| "epoch": 0.41, |
| "grad_norm": 1.05610990524292, |
| "learning_rate": 8.270676691729324e-06, |
| "loss": 0.0445, |
| "step": 440 |
| }, |
| { |
| "epoch": 0.42, |
| "grad_norm": 0.6562148332595825, |
| "learning_rate": 8.458646616541353e-06, |
| "loss": 0.0771, |
| "step": 450 |
| }, |
| { |
| "epoch": 0.43, |
| "grad_norm": 0.2789516746997833, |
| "learning_rate": 8.646616541353385e-06, |
| "loss": 0.0426, |
| "step": 460 |
| }, |
| { |
| "epoch": 0.44, |
| "grad_norm": 0.25616365671157837, |
| "learning_rate": 8.834586466165414e-06, |
| "loss": 0.0177, |
| "step": 470 |
| }, |
| { |
| "epoch": 0.45, |
| "grad_norm": 1.472976565361023, |
| "learning_rate": 9.022556390977444e-06, |
| "loss": 0.0344, |
| "step": 480 |
| }, |
| { |
| "epoch": 0.46, |
| "grad_norm": 0.6178507804870605, |
| "learning_rate": 9.210526315789474e-06, |
| "loss": 0.0534, |
| "step": 490 |
| }, |
| { |
| "epoch": 0.47, |
| "grad_norm": 1.1991667747497559, |
| "learning_rate": 9.398496240601505e-06, |
| "loss": 0.0338, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.48, |
| "grad_norm": 0.4456295073032379, |
| "learning_rate": 9.586466165413535e-06, |
| "loss": 0.0322, |
| "step": 510 |
| }, |
| { |
| "epoch": 0.49, |
| "grad_norm": 0.6090606451034546, |
| "learning_rate": 9.774436090225564e-06, |
| "loss": 0.047, |
| "step": 520 |
| }, |
| { |
| "epoch": 0.5, |
| "grad_norm": 0.4637684226036072, |
| "learning_rate": 9.962406015037594e-06, |
| "loss": 0.0479, |
| "step": 530 |
| }, |
| { |
| "epoch": 0.51, |
| "grad_norm": 0.7359185218811035, |
| "learning_rate": 9.983274095755803e-06, |
| "loss": 0.0365, |
| "step": 540 |
| }, |
| { |
| "epoch": 0.52, |
| "grad_norm": 0.5810590386390686, |
| "learning_rate": 9.962366715450555e-06, |
| "loss": 0.0276, |
| "step": 550 |
| }, |
| { |
| "epoch": 0.53, |
| "grad_norm": 0.31649720668792725, |
| "learning_rate": 9.941459335145307e-06, |
| "loss": 0.0327, |
| "step": 560 |
| }, |
| { |
| "epoch": 0.54, |
| "grad_norm": 1.11226224899292, |
| "learning_rate": 9.92055195484006e-06, |
| "loss": 0.0361, |
| "step": 570 |
| }, |
| { |
| "epoch": 0.55, |
| "grad_norm": 0.257317453622818, |
| "learning_rate": 9.89964457453481e-06, |
| "loss": 0.035, |
| "step": 580 |
| }, |
| { |
| "epoch": 0.56, |
| "grad_norm": 0.9321485757827759, |
| "learning_rate": 9.878737194229565e-06, |
| "loss": 0.0491, |
| "step": 590 |
| }, |
| { |
| "epoch": 0.56, |
| "grad_norm": 0.5542380809783936, |
| "learning_rate": 9.857829813924315e-06, |
| "loss": 0.0242, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.57, |
| "grad_norm": 0.8897249698638916, |
| "learning_rate": 9.836922433619069e-06, |
| "loss": 0.042, |
| "step": 610 |
| }, |
| { |
| "epoch": 0.58, |
| "grad_norm": 0.3682314157485962, |
| "learning_rate": 9.81601505331382e-06, |
| "loss": 0.024, |
| "step": 620 |
| }, |
| { |
| "epoch": 0.59, |
| "grad_norm": 1.0642801523208618, |
| "learning_rate": 9.795107673008573e-06, |
| "loss": 0.0402, |
| "step": 630 |
| }, |
| { |
| "epoch": 0.6, |
| "grad_norm": 0.27436748147010803, |
| "learning_rate": 9.774200292703325e-06, |
| "loss": 0.0318, |
| "step": 640 |
| }, |
| { |
| "epoch": 0.61, |
| "grad_norm": 0.13485893607139587, |
| "learning_rate": 9.753292912398077e-06, |
| "loss": 0.0265, |
| "step": 650 |
| }, |
| { |
| "epoch": 0.62, |
| "grad_norm": 1.0160752534866333, |
| "learning_rate": 9.73238553209283e-06, |
| "loss": 0.0257, |
| "step": 660 |
| }, |
| { |
| "epoch": 0.63, |
| "grad_norm": 0.2634833753108978, |
| "learning_rate": 9.711478151787582e-06, |
| "loss": 0.0358, |
| "step": 670 |
| }, |
| { |
| "epoch": 0.64, |
| "grad_norm": 0.27712008357048035, |
| "learning_rate": 9.690570771482334e-06, |
| "loss": 0.035, |
| "step": 680 |
| }, |
| { |
| "epoch": 0.65, |
| "grad_norm": 0.4950433075428009, |
| "learning_rate": 9.669663391177086e-06, |
| "loss": 0.0333, |
| "step": 690 |
| }, |
| { |
| "epoch": 0.66, |
| "grad_norm": 0.3358776867389679, |
| "learning_rate": 9.648756010871838e-06, |
| "loss": 0.0395, |
| "step": 700 |
| }, |
| { |
| "epoch": 0.67, |
| "grad_norm": 0.2755337655544281, |
| "learning_rate": 9.62784863056659e-06, |
| "loss": 0.0332, |
| "step": 710 |
| }, |
| { |
| "epoch": 0.68, |
| "grad_norm": 0.378814697265625, |
| "learning_rate": 9.606941250261344e-06, |
| "loss": 0.0386, |
| "step": 720 |
| }, |
| { |
| "epoch": 0.69, |
| "grad_norm": 0.40000224113464355, |
| "learning_rate": 9.586033869956095e-06, |
| "loss": 0.0336, |
| "step": 730 |
| }, |
| { |
| "epoch": 0.7, |
| "grad_norm": 0.19049006700515747, |
| "learning_rate": 9.565126489650848e-06, |
| "loss": 0.0313, |
| "step": 740 |
| }, |
| { |
| "epoch": 0.71, |
| "grad_norm": 0.11645769327878952, |
| "learning_rate": 9.544219109345599e-06, |
| "loss": 0.0278, |
| "step": 750 |
| }, |
| { |
| "epoch": 0.71, |
| "grad_norm": 0.1895197182893753, |
| "learning_rate": 9.523311729040352e-06, |
| "loss": 0.0272, |
| "step": 760 |
| }, |
| { |
| "epoch": 0.72, |
| "grad_norm": 0.22027094662189484, |
| "learning_rate": 9.502404348735104e-06, |
| "loss": 0.0337, |
| "step": 770 |
| }, |
| { |
| "epoch": 0.73, |
| "grad_norm": 0.25732624530792236, |
| "learning_rate": 9.481496968429856e-06, |
| "loss": 0.0358, |
| "step": 780 |
| }, |
| { |
| "epoch": 0.74, |
| "grad_norm": 0.4771791696548462, |
| "learning_rate": 9.460589588124608e-06, |
| "loss": 0.0324, |
| "step": 790 |
| }, |
| { |
| "epoch": 0.75, |
| "grad_norm": 0.18475191295146942, |
| "learning_rate": 9.43968220781936e-06, |
| "loss": 0.0348, |
| "step": 800 |
| }, |
| { |
| "epoch": 0.76, |
| "grad_norm": 0.36971214413642883, |
| "learning_rate": 9.418774827514114e-06, |
| "loss": 0.0366, |
| "step": 810 |
| }, |
| { |
| "epoch": 0.77, |
| "grad_norm": 0.6465145945549011, |
| "learning_rate": 9.397867447208866e-06, |
| "loss": 0.0304, |
| "step": 820 |
| }, |
| { |
| "epoch": 0.78, |
| "grad_norm": 1.547255516052246, |
| "learning_rate": 9.376960066903618e-06, |
| "loss": 0.0221, |
| "step": 830 |
| }, |
| { |
| "epoch": 0.79, |
| "grad_norm": 0.44453418254852295, |
| "learning_rate": 9.35605268659837e-06, |
| "loss": 0.0399, |
| "step": 840 |
| }, |
| { |
| "epoch": 0.8, |
| "grad_norm": 0.1706964075565338, |
| "learning_rate": 9.335145306293122e-06, |
| "loss": 0.0337, |
| "step": 850 |
| }, |
| { |
| "epoch": 0.81, |
| "grad_norm": 0.5651724934577942, |
| "learning_rate": 9.314237925987874e-06, |
| "loss": 0.0412, |
| "step": 860 |
| }, |
| { |
| "epoch": 0.82, |
| "grad_norm": 0.3307182490825653, |
| "learning_rate": 9.293330545682628e-06, |
| "loss": 0.0276, |
| "step": 870 |
| }, |
| { |
| "epoch": 0.83, |
| "grad_norm": 0.6166930794715881, |
| "learning_rate": 9.272423165377378e-06, |
| "loss": 0.0257, |
| "step": 880 |
| }, |
| { |
| "epoch": 0.84, |
| "grad_norm": 0.5756942629814148, |
| "learning_rate": 9.251515785072132e-06, |
| "loss": 0.036, |
| "step": 890 |
| }, |
| { |
| "epoch": 0.85, |
| "grad_norm": 0.1616719663143158, |
| "learning_rate": 9.230608404766884e-06, |
| "loss": 0.0157, |
| "step": 900 |
| }, |
| { |
| "epoch": 0.86, |
| "grad_norm": 0.6081432104110718, |
| "learning_rate": 9.209701024461636e-06, |
| "loss": 0.0452, |
| "step": 910 |
| }, |
| { |
| "epoch": 0.87, |
| "grad_norm": 0.9769769310951233, |
| "learning_rate": 9.188793644156388e-06, |
| "loss": 0.0305, |
| "step": 920 |
| }, |
| { |
| "epoch": 0.87, |
| "grad_norm": 0.5156053900718689, |
| "learning_rate": 9.16788626385114e-06, |
| "loss": 0.0344, |
| "step": 930 |
| }, |
| { |
| "epoch": 0.88, |
| "grad_norm": 0.30783510208129883, |
| "learning_rate": 9.146978883545892e-06, |
| "loss": 0.0193, |
| "step": 940 |
| }, |
| { |
| "epoch": 0.89, |
| "grad_norm": 0.44905951619148254, |
| "learning_rate": 9.126071503240644e-06, |
| "loss": 0.034, |
| "step": 950 |
| }, |
| { |
| "epoch": 0.9, |
| "grad_norm": 0.3008134067058563, |
| "learning_rate": 9.105164122935398e-06, |
| "loss": 0.0272, |
| "step": 960 |
| }, |
| { |
| "epoch": 0.91, |
| "grad_norm": 0.44791507720947266, |
| "learning_rate": 9.084256742630148e-06, |
| "loss": 0.0376, |
| "step": 970 |
| }, |
| { |
| "epoch": 0.92, |
| "grad_norm": 0.7329548001289368, |
| "learning_rate": 9.063349362324902e-06, |
| "loss": 0.0424, |
| "step": 980 |
| }, |
| { |
| "epoch": 0.93, |
| "grad_norm": 0.16488119959831238, |
| "learning_rate": 9.042441982019654e-06, |
| "loss": 0.0301, |
| "step": 990 |
| }, |
| { |
| "epoch": 0.94, |
| "grad_norm": 0.4265969395637512, |
| "learning_rate": 9.021534601714406e-06, |
| "loss": 0.022, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.95, |
| "grad_norm": 0.2555549740791321, |
| "learning_rate": 9.000627221409158e-06, |
| "loss": 0.0319, |
| "step": 1010 |
| }, |
| { |
| "epoch": 0.96, |
| "grad_norm": 0.7876251339912415, |
| "learning_rate": 8.97971984110391e-06, |
| "loss": 0.0409, |
| "step": 1020 |
| }, |
| { |
| "epoch": 0.97, |
| "grad_norm": 0.32860153913497925, |
| "learning_rate": 8.958812460798662e-06, |
| "loss": 0.0316, |
| "step": 1030 |
| }, |
| { |
| "epoch": 0.98, |
| "grad_norm": 0.36898553371429443, |
| "learning_rate": 8.937905080493416e-06, |
| "loss": 0.0325, |
| "step": 1040 |
| }, |
| { |
| "epoch": 0.99, |
| "grad_norm": 0.13984709978103638, |
| "learning_rate": 8.916997700188168e-06, |
| "loss": 0.0225, |
| "step": 1050 |
| }, |
| { |
| "epoch": 1.0, |
| "grad_norm": 0.09932324290275574, |
| "learning_rate": 8.89609031988292e-06, |
| "loss": 0.0175, |
| "step": 1060 |
| }, |
| { |
| "epoch": 1.01, |
| "grad_norm": 0.16241900622844696, |
| "learning_rate": 8.875182939577672e-06, |
| "loss": 0.0222, |
| "step": 1070 |
| }, |
| { |
| "epoch": 1.02, |
| "grad_norm": 0.14714168012142181, |
| "learning_rate": 8.854275559272424e-06, |
| "loss": 0.0207, |
| "step": 1080 |
| }, |
| { |
| "epoch": 1.03, |
| "grad_norm": 0.24781319499015808, |
| "learning_rate": 8.833368178967177e-06, |
| "loss": 0.0172, |
| "step": 1090 |
| }, |
| { |
| "epoch": 1.03, |
| "grad_norm": 0.3309984803199768, |
| "learning_rate": 8.812460798661928e-06, |
| "loss": 0.0377, |
| "step": 1100 |
| }, |
| { |
| "epoch": 1.04, |
| "grad_norm": 0.35171088576316833, |
| "learning_rate": 8.791553418356681e-06, |
| "loss": 0.0406, |
| "step": 1110 |
| }, |
| { |
| "epoch": 1.05, |
| "grad_norm": 0.3367606997489929, |
| "learning_rate": 8.770646038051432e-06, |
| "loss": 0.032, |
| "step": 1120 |
| }, |
| { |
| "epoch": 1.06, |
| "grad_norm": 0.13503430783748627, |
| "learning_rate": 8.749738657746185e-06, |
| "loss": 0.0299, |
| "step": 1130 |
| }, |
| { |
| "epoch": 1.07, |
| "grad_norm": 0.3619076907634735, |
| "learning_rate": 8.728831277440937e-06, |
| "loss": 0.0319, |
| "step": 1140 |
| }, |
| { |
| "epoch": 1.08, |
| "grad_norm": 0.40823495388031006, |
| "learning_rate": 8.70792389713569e-06, |
| "loss": 0.0249, |
| "step": 1150 |
| }, |
| { |
| "epoch": 1.09, |
| "grad_norm": 0.7125621438026428, |
| "learning_rate": 8.687016516830441e-06, |
| "loss": 0.0235, |
| "step": 1160 |
| }, |
| { |
| "epoch": 1.1, |
| "grad_norm": 0.12342959642410278, |
| "learning_rate": 8.666109136525193e-06, |
| "loss": 0.0288, |
| "step": 1170 |
| }, |
| { |
| "epoch": 1.11, |
| "grad_norm": 0.232752725481987, |
| "learning_rate": 8.645201756219947e-06, |
| "loss": 0.0354, |
| "step": 1180 |
| }, |
| { |
| "epoch": 1.12, |
| "grad_norm": 0.1255234330892563, |
| "learning_rate": 8.6242943759147e-06, |
| "loss": 0.0115, |
| "step": 1190 |
| }, |
| { |
| "epoch": 1.13, |
| "grad_norm": 0.26488494873046875, |
| "learning_rate": 8.603386995609451e-06, |
| "loss": 0.0247, |
| "step": 1200 |
| }, |
| { |
| "epoch": 1.14, |
| "grad_norm": 0.22945912182331085, |
| "learning_rate": 8.582479615304203e-06, |
| "loss": 0.0349, |
| "step": 1210 |
| }, |
| { |
| "epoch": 1.15, |
| "grad_norm": 0.3374781906604767, |
| "learning_rate": 8.561572234998955e-06, |
| "loss": 0.0303, |
| "step": 1220 |
| }, |
| { |
| "epoch": 1.16, |
| "grad_norm": 0.38748908042907715, |
| "learning_rate": 8.540664854693707e-06, |
| "loss": 0.0252, |
| "step": 1230 |
| }, |
| { |
| "epoch": 1.17, |
| "grad_norm": 0.2272762805223465, |
| "learning_rate": 8.519757474388461e-06, |
| "loss": 0.0325, |
| "step": 1240 |
| }, |
| { |
| "epoch": 1.18, |
| "grad_norm": 0.4768443703651428, |
| "learning_rate": 8.498850094083211e-06, |
| "loss": 0.0269, |
| "step": 1250 |
| }, |
| { |
| "epoch": 1.19, |
| "grad_norm": 0.32005059719085693, |
| "learning_rate": 8.477942713777965e-06, |
| "loss": 0.024, |
| "step": 1260 |
| }, |
| { |
| "epoch": 1.19, |
| "grad_norm": 0.2935084104537964, |
| "learning_rate": 8.457035333472717e-06, |
| "loss": 0.028, |
| "step": 1270 |
| }, |
| { |
| "epoch": 1.2, |
| "grad_norm": 0.17606884241104126, |
| "learning_rate": 8.436127953167469e-06, |
| "loss": 0.0314, |
| "step": 1280 |
| }, |
| { |
| "epoch": 1.21, |
| "grad_norm": 0.11081908643245697, |
| "learning_rate": 8.415220572862221e-06, |
| "loss": 0.0142, |
| "step": 1290 |
| }, |
| { |
| "epoch": 1.22, |
| "grad_norm": 0.06663521379232407, |
| "learning_rate": 8.394313192556973e-06, |
| "loss": 0.0217, |
| "step": 1300 |
| }, |
| { |
| "epoch": 1.23, |
| "grad_norm": 0.6221704483032227, |
| "learning_rate": 8.373405812251725e-06, |
| "loss": 0.0308, |
| "step": 1310 |
| }, |
| { |
| "epoch": 1.24, |
| "grad_norm": 0.13281384110450745, |
| "learning_rate": 8.352498431946477e-06, |
| "loss": 0.0295, |
| "step": 1320 |
| }, |
| { |
| "epoch": 1.25, |
| "grad_norm": 0.21760965883731842, |
| "learning_rate": 8.331591051641231e-06, |
| "loss": 0.0243, |
| "step": 1330 |
| }, |
| { |
| "epoch": 1.26, |
| "grad_norm": 0.29121580719947815, |
| "learning_rate": 8.310683671335981e-06, |
| "loss": 0.0207, |
| "step": 1340 |
| }, |
| { |
| "epoch": 1.27, |
| "grad_norm": 0.04138851910829544, |
| "learning_rate": 8.289776291030735e-06, |
| "loss": 0.03, |
| "step": 1350 |
| }, |
| { |
| "epoch": 1.28, |
| "grad_norm": 0.06128918379545212, |
| "learning_rate": 8.268868910725487e-06, |
| "loss": 0.0264, |
| "step": 1360 |
| }, |
| { |
| "epoch": 1.29, |
| "grad_norm": 0.3501634895801544, |
| "learning_rate": 8.247961530420239e-06, |
| "loss": 0.0193, |
| "step": 1370 |
| }, |
| { |
| "epoch": 1.3, |
| "grad_norm": 0.17542318999767303, |
| "learning_rate": 8.227054150114991e-06, |
| "loss": 0.0144, |
| "step": 1380 |
| }, |
| { |
| "epoch": 1.31, |
| "grad_norm": 0.10308554768562317, |
| "learning_rate": 8.206146769809743e-06, |
| "loss": 0.0543, |
| "step": 1390 |
| }, |
| { |
| "epoch": 1.32, |
| "grad_norm": 0.3240692615509033, |
| "learning_rate": 8.185239389504495e-06, |
| "loss": 0.0297, |
| "step": 1400 |
| }, |
| { |
| "epoch": 1.33, |
| "grad_norm": 0.09771730750799179, |
| "learning_rate": 8.164332009199249e-06, |
| "loss": 0.0334, |
| "step": 1410 |
| }, |
| { |
| "epoch": 1.34, |
| "grad_norm": 0.5148730278015137, |
| "learning_rate": 8.143424628894e-06, |
| "loss": 0.0209, |
| "step": 1420 |
| }, |
| { |
| "epoch": 1.35, |
| "grad_norm": 0.546535313129425, |
| "learning_rate": 8.122517248588753e-06, |
| "loss": 0.021, |
| "step": 1430 |
| }, |
| { |
| "epoch": 1.35, |
| "grad_norm": 0.7162196636199951, |
| "learning_rate": 8.101609868283505e-06, |
| "loss": 0.0218, |
| "step": 1440 |
| }, |
| { |
| "epoch": 1.36, |
| "grad_norm": 0.11191996932029724, |
| "learning_rate": 8.080702487978257e-06, |
| "loss": 0.0291, |
| "step": 1450 |
| }, |
| { |
| "epoch": 1.37, |
| "grad_norm": 0.04737411066889763, |
| "learning_rate": 8.05979510767301e-06, |
| "loss": 0.0254, |
| "step": 1460 |
| }, |
| { |
| "epoch": 1.38, |
| "grad_norm": 0.1907588094472885, |
| "learning_rate": 8.03888772736776e-06, |
| "loss": 0.0308, |
| "step": 1470 |
| }, |
| { |
| "epoch": 1.39, |
| "grad_norm": 0.18960894644260406, |
| "learning_rate": 8.017980347062515e-06, |
| "loss": 0.0157, |
| "step": 1480 |
| }, |
| { |
| "epoch": 1.4, |
| "grad_norm": 1.294573187828064, |
| "learning_rate": 7.997072966757265e-06, |
| "loss": 0.0162, |
| "step": 1490 |
| }, |
| { |
| "epoch": 1.41, |
| "grad_norm": 0.4640454947948456, |
| "learning_rate": 7.976165586452019e-06, |
| "loss": 0.0186, |
| "step": 1500 |
| }, |
| { |
| "epoch": 1.42, |
| "grad_norm": 0.6731734871864319, |
| "learning_rate": 7.95525820614677e-06, |
| "loss": 0.0144, |
| "step": 1510 |
| }, |
| { |
| "epoch": 1.43, |
| "grad_norm": 0.13755524158477783, |
| "learning_rate": 7.934350825841523e-06, |
| "loss": 0.0231, |
| "step": 1520 |
| }, |
| { |
| "epoch": 1.44, |
| "grad_norm": 0.16188235580921173, |
| "learning_rate": 7.913443445536275e-06, |
| "loss": 0.0276, |
| "step": 1530 |
| }, |
| { |
| "epoch": 1.45, |
| "grad_norm": 0.10896871984004974, |
| "learning_rate": 7.892536065231027e-06, |
| "loss": 0.0202, |
| "step": 1540 |
| }, |
| { |
| "epoch": 1.46, |
| "grad_norm": 0.08035355061292648, |
| "learning_rate": 7.871628684925779e-06, |
| "loss": 0.0143, |
| "step": 1550 |
| }, |
| { |
| "epoch": 1.47, |
| "grad_norm": 0.3662511706352234, |
| "learning_rate": 7.850721304620532e-06, |
| "loss": 0.0277, |
| "step": 1560 |
| }, |
| { |
| "epoch": 1.48, |
| "grad_norm": 0.5264787077903748, |
| "learning_rate": 7.829813924315284e-06, |
| "loss": 0.02, |
| "step": 1570 |
| }, |
| { |
| "epoch": 1.49, |
| "grad_norm": 0.9667835831642151, |
| "learning_rate": 7.808906544010036e-06, |
| "loss": 0.0236, |
| "step": 1580 |
| }, |
| { |
| "epoch": 1.5, |
| "grad_norm": 0.024623023346066475, |
| "learning_rate": 7.787999163704788e-06, |
| "loss": 0.0152, |
| "step": 1590 |
| }, |
| { |
| "epoch": 1.51, |
| "grad_norm": 0.01869453489780426, |
| "learning_rate": 7.76709178339954e-06, |
| "loss": 0.0165, |
| "step": 1600 |
| }, |
| { |
| "epoch": 1.51, |
| "grad_norm": 0.2934032678604126, |
| "learning_rate": 7.746184403094294e-06, |
| "loss": 0.0178, |
| "step": 1610 |
| }, |
| { |
| "epoch": 1.52, |
| "grad_norm": 0.1588401347398758, |
| "learning_rate": 7.725277022789044e-06, |
| "loss": 0.0174, |
| "step": 1620 |
| }, |
| { |
| "epoch": 1.53, |
| "grad_norm": 0.08647409826517105, |
| "learning_rate": 7.704369642483798e-06, |
| "loss": 0.0243, |
| "step": 1630 |
| }, |
| { |
| "epoch": 1.54, |
| "grad_norm": 0.16604632139205933, |
| "learning_rate": 7.683462262178549e-06, |
| "loss": 0.0323, |
| "step": 1640 |
| }, |
| { |
| "epoch": 1.55, |
| "grad_norm": 0.2372957319021225, |
| "learning_rate": 7.662554881873302e-06, |
| "loss": 0.0148, |
| "step": 1650 |
| }, |
| { |
| "epoch": 1.56, |
| "grad_norm": 0.1503574103116989, |
| "learning_rate": 7.641647501568054e-06, |
| "loss": 0.03, |
| "step": 1660 |
| }, |
| { |
| "epoch": 1.57, |
| "grad_norm": 0.029307017102837563, |
| "learning_rate": 7.620740121262806e-06, |
| "loss": 0.0221, |
| "step": 1670 |
| }, |
| { |
| "epoch": 1.58, |
| "grad_norm": 0.18366031348705292, |
| "learning_rate": 7.599832740957559e-06, |
| "loss": 0.015, |
| "step": 1680 |
| }, |
| { |
| "epoch": 1.59, |
| "grad_norm": 0.049557920545339584, |
| "learning_rate": 7.57892536065231e-06, |
| "loss": 0.0203, |
| "step": 1690 |
| }, |
| { |
| "epoch": 1.6, |
| "grad_norm": 0.18749569356441498, |
| "learning_rate": 7.558017980347063e-06, |
| "loss": 0.0221, |
| "step": 1700 |
| }, |
| { |
| "epoch": 1.61, |
| "grad_norm": 0.06821219623088837, |
| "learning_rate": 7.537110600041815e-06, |
| "loss": 0.0174, |
| "step": 1710 |
| }, |
| { |
| "epoch": 1.62, |
| "grad_norm": 0.27043628692626953, |
| "learning_rate": 7.516203219736567e-06, |
| "loss": 0.0214, |
| "step": 1720 |
| }, |
| { |
| "epoch": 1.63, |
| "grad_norm": 0.25328314304351807, |
| "learning_rate": 7.49529583943132e-06, |
| "loss": 0.0245, |
| "step": 1730 |
| }, |
| { |
| "epoch": 1.64, |
| "grad_norm": 0.11644481867551804, |
| "learning_rate": 7.474388459126072e-06, |
| "loss": 0.0218, |
| "step": 1740 |
| }, |
| { |
| "epoch": 1.65, |
| "grad_norm": 0.4062769114971161, |
| "learning_rate": 7.453481078820824e-06, |
| "loss": 0.0299, |
| "step": 1750 |
| }, |
| { |
| "epoch": 1.66, |
| "grad_norm": 0.30739670991897583, |
| "learning_rate": 7.432573698515576e-06, |
| "loss": 0.0351, |
| "step": 1760 |
| }, |
| { |
| "epoch": 1.67, |
| "grad_norm": 0.7801105380058289, |
| "learning_rate": 7.411666318210329e-06, |
| "loss": 0.0305, |
| "step": 1770 |
| }, |
| { |
| "epoch": 1.67, |
| "grad_norm": 0.29210716485977173, |
| "learning_rate": 7.390758937905082e-06, |
| "loss": 0.0274, |
| "step": 1780 |
| }, |
| { |
| "epoch": 1.68, |
| "grad_norm": 0.09251756221055984, |
| "learning_rate": 7.369851557599833e-06, |
| "loss": 0.0189, |
| "step": 1790 |
| }, |
| { |
| "epoch": 1.69, |
| "grad_norm": 0.1489870399236679, |
| "learning_rate": 7.348944177294586e-06, |
| "loss": 0.0256, |
| "step": 1800 |
| }, |
| { |
| "epoch": 1.7, |
| "grad_norm": 0.39114630222320557, |
| "learning_rate": 7.328036796989337e-06, |
| "loss": 0.0226, |
| "step": 1810 |
| }, |
| { |
| "epoch": 1.71, |
| "grad_norm": 0.19388756155967712, |
| "learning_rate": 7.30712941668409e-06, |
| "loss": 0.0287, |
| "step": 1820 |
| }, |
| { |
| "epoch": 1.72, |
| "grad_norm": 0.40138277411460876, |
| "learning_rate": 7.286222036378843e-06, |
| "loss": 0.0359, |
| "step": 1830 |
| }, |
| { |
| "epoch": 1.73, |
| "grad_norm": 0.34222084283828735, |
| "learning_rate": 7.265314656073594e-06, |
| "loss": 0.0298, |
| "step": 1840 |
| }, |
| { |
| "epoch": 1.74, |
| "grad_norm": 0.42924413084983826, |
| "learning_rate": 7.244407275768347e-06, |
| "loss": 0.0234, |
| "step": 1850 |
| }, |
| { |
| "epoch": 1.75, |
| "grad_norm": 0.26645031571388245, |
| "learning_rate": 7.223499895463099e-06, |
| "loss": 0.0202, |
| "step": 1860 |
| }, |
| { |
| "epoch": 1.76, |
| "grad_norm": 0.18649956583976746, |
| "learning_rate": 7.202592515157852e-06, |
| "loss": 0.0283, |
| "step": 1870 |
| }, |
| { |
| "epoch": 1.77, |
| "grad_norm": 0.23801521956920624, |
| "learning_rate": 7.181685134852604e-06, |
| "loss": 0.0187, |
| "step": 1880 |
| }, |
| { |
| "epoch": 1.78, |
| "grad_norm": 0.4391060769557953, |
| "learning_rate": 7.160777754547356e-06, |
| "loss": 0.0321, |
| "step": 1890 |
| }, |
| { |
| "epoch": 1.79, |
| "grad_norm": 0.8792116641998291, |
| "learning_rate": 7.139870374242109e-06, |
| "loss": 0.0272, |
| "step": 1900 |
| }, |
| { |
| "epoch": 1.8, |
| "grad_norm": 0.269652396440506, |
| "learning_rate": 7.11896299393686e-06, |
| "loss": 0.0304, |
| "step": 1910 |
| }, |
| { |
| "epoch": 1.81, |
| "grad_norm": 0.33860161900520325, |
| "learning_rate": 7.098055613631613e-06, |
| "loss": 0.0305, |
| "step": 1920 |
| }, |
| { |
| "epoch": 1.82, |
| "grad_norm": 0.4650591015815735, |
| "learning_rate": 7.0771482333263655e-06, |
| "loss": 0.0316, |
| "step": 1930 |
| }, |
| { |
| "epoch": 1.83, |
| "grad_norm": 0.1440097540616989, |
| "learning_rate": 7.056240853021117e-06, |
| "loss": 0.0223, |
| "step": 1940 |
| }, |
| { |
| "epoch": 1.83, |
| "grad_norm": 0.196376234292984, |
| "learning_rate": 7.0353334727158696e-06, |
| "loss": 0.0298, |
| "step": 1950 |
| }, |
| { |
| "epoch": 1.84, |
| "grad_norm": 0.617382824420929, |
| "learning_rate": 7.014426092410621e-06, |
| "loss": 0.0203, |
| "step": 1960 |
| }, |
| { |
| "epoch": 1.85, |
| "grad_norm": 0.14364871382713318, |
| "learning_rate": 6.993518712105374e-06, |
| "loss": 0.0217, |
| "step": 1970 |
| }, |
| { |
| "epoch": 1.86, |
| "grad_norm": 0.2357790470123291, |
| "learning_rate": 6.9726113318001265e-06, |
| "loss": 0.019, |
| "step": 1980 |
| }, |
| { |
| "epoch": 1.87, |
| "grad_norm": 0.13550205528736115, |
| "learning_rate": 6.9517039514948785e-06, |
| "loss": 0.0276, |
| "step": 1990 |
| }, |
| { |
| "epoch": 1.88, |
| "grad_norm": 0.20195119082927704, |
| "learning_rate": 6.9307965711896305e-06, |
| "loss": 0.0268, |
| "step": 2000 |
| }, |
| { |
| "epoch": 1.89, |
| "grad_norm": 0.19456946849822998, |
| "learning_rate": 6.9098891908843825e-06, |
| "loss": 0.0243, |
| "step": 2010 |
| }, |
| { |
| "epoch": 1.9, |
| "grad_norm": 0.5265508890151978, |
| "learning_rate": 6.888981810579135e-06, |
| "loss": 0.0342, |
| "step": 2020 |
| }, |
| { |
| "epoch": 1.91, |
| "grad_norm": 0.38090288639068604, |
| "learning_rate": 6.868074430273887e-06, |
| "loss": 0.0219, |
| "step": 2030 |
| }, |
| { |
| "epoch": 1.92, |
| "grad_norm": 0.30138471722602844, |
| "learning_rate": 6.847167049968639e-06, |
| "loss": 0.0193, |
| "step": 2040 |
| }, |
| { |
| "epoch": 1.93, |
| "grad_norm": 0.36553192138671875, |
| "learning_rate": 6.826259669663392e-06, |
| "loss": 0.0379, |
| "step": 2050 |
| }, |
| { |
| "epoch": 1.94, |
| "grad_norm": 0.27630284428596497, |
| "learning_rate": 6.8053522893581435e-06, |
| "loss": 0.016, |
| "step": 2060 |
| }, |
| { |
| "epoch": 1.95, |
| "grad_norm": 0.06610503047704697, |
| "learning_rate": 6.784444909052896e-06, |
| "loss": 0.0397, |
| "step": 2070 |
| }, |
| { |
| "epoch": 1.96, |
| "grad_norm": 0.257328599691391, |
| "learning_rate": 6.763537528747649e-06, |
| "loss": 0.0238, |
| "step": 2080 |
| }, |
| { |
| "epoch": 1.97, |
| "grad_norm": 0.1945256143808365, |
| "learning_rate": 6.7426301484424e-06, |
| "loss": 0.0212, |
| "step": 2090 |
| }, |
| { |
| "epoch": 1.98, |
| "grad_norm": 0.09793351590633392, |
| "learning_rate": 6.721722768137153e-06, |
| "loss": 0.0264, |
| "step": 2100 |
| }, |
| { |
| "epoch": 1.98, |
| "grad_norm": 0.3976469933986664, |
| "learning_rate": 6.700815387831905e-06, |
| "loss": 0.0294, |
| "step": 2110 |
| }, |
| { |
| "epoch": 1.99, |
| "grad_norm": 0.43862876296043396, |
| "learning_rate": 6.679908007526657e-06, |
| "loss": 0.0299, |
| "step": 2120 |
| }, |
| { |
| "epoch": 2.0, |
| "grad_norm": 0.3431900143623352, |
| "learning_rate": 6.659000627221409e-06, |
| "loss": 0.0196, |
| "step": 2130 |
| }, |
| { |
| "epoch": 2.01, |
| "grad_norm": 0.26596391201019287, |
| "learning_rate": 6.638093246916162e-06, |
| "loss": 0.0221, |
| "step": 2140 |
| }, |
| { |
| "epoch": 2.02, |
| "grad_norm": 0.5497652888298035, |
| "learning_rate": 6.617185866610914e-06, |
| "loss": 0.0144, |
| "step": 2150 |
| }, |
| { |
| "epoch": 2.03, |
| "grad_norm": 0.12056022882461548, |
| "learning_rate": 6.596278486305666e-06, |
| "loss": 0.0219, |
| "step": 2160 |
| }, |
| { |
| "epoch": 2.04, |
| "grad_norm": 0.2761884927749634, |
| "learning_rate": 6.575371106000419e-06, |
| "loss": 0.0152, |
| "step": 2170 |
| }, |
| { |
| "epoch": 2.05, |
| "grad_norm": 0.4963738024234772, |
| "learning_rate": 6.55446372569517e-06, |
| "loss": 0.012, |
| "step": 2180 |
| }, |
| { |
| "epoch": 2.06, |
| "grad_norm": 0.40073689818382263, |
| "learning_rate": 6.533556345389923e-06, |
| "loss": 0.027, |
| "step": 2190 |
| }, |
| { |
| "epoch": 2.07, |
| "grad_norm": 0.09300912916660309, |
| "learning_rate": 6.512648965084676e-06, |
| "loss": 0.0124, |
| "step": 2200 |
| }, |
| { |
| "epoch": 2.08, |
| "grad_norm": 0.49626946449279785, |
| "learning_rate": 6.491741584779427e-06, |
| "loss": 0.0122, |
| "step": 2210 |
| }, |
| { |
| "epoch": 2.09, |
| "grad_norm": 0.023487213999032974, |
| "learning_rate": 6.47083420447418e-06, |
| "loss": 0.0161, |
| "step": 2220 |
| }, |
| { |
| "epoch": 2.1, |
| "grad_norm": 0.0902651771903038, |
| "learning_rate": 6.449926824168932e-06, |
| "loss": 0.0269, |
| "step": 2230 |
| }, |
| { |
| "epoch": 2.11, |
| "grad_norm": 0.17983660101890564, |
| "learning_rate": 6.429019443863684e-06, |
| "loss": 0.0119, |
| "step": 2240 |
| }, |
| { |
| "epoch": 2.12, |
| "grad_norm": 0.1324489861726761, |
| "learning_rate": 6.408112063558437e-06, |
| "loss": 0.0278, |
| "step": 2250 |
| }, |
| { |
| "epoch": 2.13, |
| "grad_norm": 0.17838072776794434, |
| "learning_rate": 6.387204683253189e-06, |
| "loss": 0.0198, |
| "step": 2260 |
| }, |
| { |
| "epoch": 2.14, |
| "grad_norm": 0.2527213394641876, |
| "learning_rate": 6.366297302947941e-06, |
| "loss": 0.0154, |
| "step": 2270 |
| }, |
| { |
| "epoch": 2.14, |
| "grad_norm": 0.06320203095674515, |
| "learning_rate": 6.345389922642693e-06, |
| "loss": 0.0105, |
| "step": 2280 |
| }, |
| { |
| "epoch": 2.15, |
| "grad_norm": 0.17222613096237183, |
| "learning_rate": 6.324482542337446e-06, |
| "loss": 0.0131, |
| "step": 2290 |
| }, |
| { |
| "epoch": 2.16, |
| "grad_norm": 0.019979368895292282, |
| "learning_rate": 6.303575162032199e-06, |
| "loss": 0.0177, |
| "step": 2300 |
| }, |
| { |
| "epoch": 2.17, |
| "grad_norm": 0.6333135366439819, |
| "learning_rate": 6.28266778172695e-06, |
| "loss": 0.0205, |
| "step": 2310 |
| }, |
| { |
| "epoch": 2.18, |
| "grad_norm": 0.3780456483364105, |
| "learning_rate": 6.261760401421703e-06, |
| "loss": 0.0156, |
| "step": 2320 |
| }, |
| { |
| "epoch": 2.19, |
| "grad_norm": 0.3019196093082428, |
| "learning_rate": 6.240853021116454e-06, |
| "loss": 0.0139, |
| "step": 2330 |
| }, |
| { |
| "epoch": 2.2, |
| "grad_norm": 0.3984769582748413, |
| "learning_rate": 6.219945640811207e-06, |
| "loss": 0.0276, |
| "step": 2340 |
| }, |
| { |
| "epoch": 2.21, |
| "grad_norm": 0.22097636759281158, |
| "learning_rate": 6.19903826050596e-06, |
| "loss": 0.0198, |
| "step": 2350 |
| }, |
| { |
| "epoch": 2.22, |
| "grad_norm": 0.24413403868675232, |
| "learning_rate": 6.178130880200711e-06, |
| "loss": 0.018, |
| "step": 2360 |
| }, |
| { |
| "epoch": 2.23, |
| "grad_norm": 0.13032527267932892, |
| "learning_rate": 6.157223499895464e-06, |
| "loss": 0.0116, |
| "step": 2370 |
| }, |
| { |
| "epoch": 2.24, |
| "grad_norm": 0.2249244898557663, |
| "learning_rate": 6.136316119590216e-06, |
| "loss": 0.0118, |
| "step": 2380 |
| }, |
| { |
| "epoch": 2.25, |
| "grad_norm": 0.20351147651672363, |
| "learning_rate": 6.1154087392849685e-06, |
| "loss": 0.0257, |
| "step": 2390 |
| }, |
| { |
| "epoch": 2.26, |
| "grad_norm": 0.49516594409942627, |
| "learning_rate": 6.0945013589797206e-06, |
| "loss": 0.0154, |
| "step": 2400 |
| }, |
| { |
| "epoch": 2.27, |
| "grad_norm": 0.21651627123355865, |
| "learning_rate": 6.0735939786744726e-06, |
| "loss": 0.0135, |
| "step": 2410 |
| }, |
| { |
| "epoch": 2.28, |
| "grad_norm": 0.2528943121433258, |
| "learning_rate": 6.0526865983692254e-06, |
| "loss": 0.0298, |
| "step": 2420 |
| }, |
| { |
| "epoch": 2.29, |
| "grad_norm": 0.2160516232252121, |
| "learning_rate": 6.031779218063977e-06, |
| "loss": 0.0184, |
| "step": 2430 |
| }, |
| { |
| "epoch": 2.3, |
| "grad_norm": 0.14053639769554138, |
| "learning_rate": 6.0108718377587295e-06, |
| "loss": 0.0224, |
| "step": 2440 |
| }, |
| { |
| "epoch": 2.3, |
| "grad_norm": 0.23231235146522522, |
| "learning_rate": 5.989964457453482e-06, |
| "loss": 0.0152, |
| "step": 2450 |
| }, |
| { |
| "epoch": 2.31, |
| "grad_norm": 0.1623637080192566, |
| "learning_rate": 5.9690570771482335e-06, |
| "loss": 0.0089, |
| "step": 2460 |
| }, |
| { |
| "epoch": 2.32, |
| "grad_norm": 0.06655038148164749, |
| "learning_rate": 5.948149696842986e-06, |
| "loss": 0.0246, |
| "step": 2470 |
| }, |
| { |
| "epoch": 2.33, |
| "grad_norm": 0.508985698223114, |
| "learning_rate": 5.9272423165377375e-06, |
| "loss": 0.0144, |
| "step": 2480 |
| }, |
| { |
| "epoch": 2.34, |
| "grad_norm": 0.1423477828502655, |
| "learning_rate": 5.90633493623249e-06, |
| "loss": 0.0133, |
| "step": 2490 |
| }, |
| { |
| "epoch": 2.35, |
| "grad_norm": 0.29974058270454407, |
| "learning_rate": 5.885427555927243e-06, |
| "loss": 0.0133, |
| "step": 2500 |
| }, |
| { |
| "epoch": 2.36, |
| "grad_norm": 0.1467662900686264, |
| "learning_rate": 5.864520175621995e-06, |
| "loss": 0.0151, |
| "step": 2510 |
| }, |
| { |
| "epoch": 2.37, |
| "grad_norm": 0.09916812926530838, |
| "learning_rate": 5.843612795316747e-06, |
| "loss": 0.0173, |
| "step": 2520 |
| }, |
| { |
| "epoch": 2.38, |
| "grad_norm": 0.10352525860071182, |
| "learning_rate": 5.822705415011499e-06, |
| "loss": 0.0221, |
| "step": 2530 |
| }, |
| { |
| "epoch": 2.39, |
| "grad_norm": 0.5194036960601807, |
| "learning_rate": 5.801798034706252e-06, |
| "loss": 0.0258, |
| "step": 2540 |
| }, |
| { |
| "epoch": 2.4, |
| "grad_norm": 0.4094790816307068, |
| "learning_rate": 5.780890654401003e-06, |
| "loss": 0.018, |
| "step": 2550 |
| }, |
| { |
| "epoch": 2.41, |
| "grad_norm": 0.21202996373176575, |
| "learning_rate": 5.759983274095756e-06, |
| "loss": 0.0281, |
| "step": 2560 |
| }, |
| { |
| "epoch": 2.42, |
| "grad_norm": 0.11880209296941757, |
| "learning_rate": 5.739075893790509e-06, |
| "loss": 0.019, |
| "step": 2570 |
| }, |
| { |
| "epoch": 2.43, |
| "grad_norm": 0.15664508938789368, |
| "learning_rate": 5.71816851348526e-06, |
| "loss": 0.0162, |
| "step": 2580 |
| }, |
| { |
| "epoch": 2.44, |
| "grad_norm": 0.08462068438529968, |
| "learning_rate": 5.697261133180013e-06, |
| "loss": 0.0157, |
| "step": 2590 |
| }, |
| { |
| "epoch": 2.45, |
| "grad_norm": 0.5396077036857605, |
| "learning_rate": 5.676353752874765e-06, |
| "loss": 0.0184, |
| "step": 2600 |
| }, |
| { |
| "epoch": 2.46, |
| "grad_norm": 0.1727200448513031, |
| "learning_rate": 5.655446372569517e-06, |
| "loss": 0.0146, |
| "step": 2610 |
| }, |
| { |
| "epoch": 2.46, |
| "grad_norm": 0.27750787138938904, |
| "learning_rate": 5.63453899226427e-06, |
| "loss": 0.0137, |
| "step": 2620 |
| }, |
| { |
| "epoch": 2.47, |
| "grad_norm": 0.26835641264915466, |
| "learning_rate": 5.613631611959022e-06, |
| "loss": 0.0192, |
| "step": 2630 |
| }, |
| { |
| "epoch": 2.48, |
| "grad_norm": 0.8210089206695557, |
| "learning_rate": 5.592724231653774e-06, |
| "loss": 0.0195, |
| "step": 2640 |
| }, |
| { |
| "epoch": 2.49, |
| "grad_norm": 0.2197115570306778, |
| "learning_rate": 5.571816851348526e-06, |
| "loss": 0.021, |
| "step": 2650 |
| }, |
| { |
| "epoch": 2.5, |
| "grad_norm": 0.0288139246404171, |
| "learning_rate": 5.550909471043279e-06, |
| "loss": 0.0179, |
| "step": 2660 |
| }, |
| { |
| "epoch": 2.51, |
| "grad_norm": 0.4010894000530243, |
| "learning_rate": 5.530002090738031e-06, |
| "loss": 0.0163, |
| "step": 2670 |
| }, |
| { |
| "epoch": 2.52, |
| "grad_norm": 0.19797544181346893, |
| "learning_rate": 5.509094710432783e-06, |
| "loss": 0.0185, |
| "step": 2680 |
| }, |
| { |
| "epoch": 2.53, |
| "grad_norm": 0.06082729250192642, |
| "learning_rate": 5.488187330127536e-06, |
| "loss": 0.0245, |
| "step": 2690 |
| }, |
| { |
| "epoch": 2.54, |
| "grad_norm": 0.31951603293418884, |
| "learning_rate": 5.467279949822287e-06, |
| "loss": 0.0259, |
| "step": 2700 |
| }, |
| { |
| "epoch": 2.55, |
| "grad_norm": 0.11278670281171799, |
| "learning_rate": 5.44637256951704e-06, |
| "loss": 0.0131, |
| "step": 2710 |
| }, |
| { |
| "epoch": 2.56, |
| "grad_norm": 0.6295509934425354, |
| "learning_rate": 5.425465189211793e-06, |
| "loss": 0.0225, |
| "step": 2720 |
| }, |
| { |
| "epoch": 2.57, |
| "grad_norm": 0.12108682841062546, |
| "learning_rate": 5.404557808906544e-06, |
| "loss": 0.0139, |
| "step": 2730 |
| }, |
| { |
| "epoch": 2.58, |
| "grad_norm": 0.10168636590242386, |
| "learning_rate": 5.383650428601297e-06, |
| "loss": 0.0287, |
| "step": 2740 |
| }, |
| { |
| "epoch": 2.59, |
| "grad_norm": 0.06240009516477585, |
| "learning_rate": 5.362743048296049e-06, |
| "loss": 0.0155, |
| "step": 2750 |
| }, |
| { |
| "epoch": 2.6, |
| "grad_norm": 0.39674845337867737, |
| "learning_rate": 5.341835667990801e-06, |
| "loss": 0.019, |
| "step": 2760 |
| }, |
| { |
| "epoch": 2.61, |
| "grad_norm": 0.13627833127975464, |
| "learning_rate": 5.320928287685554e-06, |
| "loss": 0.0158, |
| "step": 2770 |
| }, |
| { |
| "epoch": 2.62, |
| "grad_norm": 0.15835390985012054, |
| "learning_rate": 5.300020907380306e-06, |
| "loss": 0.0145, |
| "step": 2780 |
| }, |
| { |
| "epoch": 2.62, |
| "grad_norm": 0.14348815381526947, |
| "learning_rate": 5.279113527075059e-06, |
| "loss": 0.0145, |
| "step": 2790 |
| }, |
| { |
| "epoch": 2.63, |
| "grad_norm": 0.08389858156442642, |
| "learning_rate": 5.25820614676981e-06, |
| "loss": 0.0085, |
| "step": 2800 |
| }, |
| { |
| "epoch": 2.64, |
| "grad_norm": 0.033944834023714066, |
| "learning_rate": 5.237298766464563e-06, |
| "loss": 0.0233, |
| "step": 2810 |
| }, |
| { |
| "epoch": 2.65, |
| "grad_norm": 0.031127430498600006, |
| "learning_rate": 5.2163913861593155e-06, |
| "loss": 0.0114, |
| "step": 2820 |
| }, |
| { |
| "epoch": 2.66, |
| "grad_norm": 0.1990656703710556, |
| "learning_rate": 5.195484005854067e-06, |
| "loss": 0.0162, |
| "step": 2830 |
| }, |
| { |
| "epoch": 2.67, |
| "grad_norm": 0.04372655227780342, |
| "learning_rate": 5.1745766255488195e-06, |
| "loss": 0.01, |
| "step": 2840 |
| }, |
| { |
| "epoch": 2.68, |
| "grad_norm": 0.34108391404151917, |
| "learning_rate": 5.153669245243571e-06, |
| "loss": 0.0166, |
| "step": 2850 |
| }, |
| { |
| "epoch": 2.69, |
| "grad_norm": 0.45837971568107605, |
| "learning_rate": 5.1327618649383236e-06, |
| "loss": 0.0192, |
| "step": 2860 |
| }, |
| { |
| "epoch": 2.7, |
| "grad_norm": 0.13491950929164886, |
| "learning_rate": 5.1118544846330764e-06, |
| "loss": 0.0162, |
| "step": 2870 |
| }, |
| { |
| "epoch": 2.71, |
| "grad_norm": 0.03369107097387314, |
| "learning_rate": 5.090947104327828e-06, |
| "loss": 0.0269, |
| "step": 2880 |
| }, |
| { |
| "epoch": 2.72, |
| "grad_norm": 0.28704699873924255, |
| "learning_rate": 5.0700397240225805e-06, |
| "loss": 0.0191, |
| "step": 2890 |
| }, |
| { |
| "epoch": 2.73, |
| "grad_norm": 0.6022350788116455, |
| "learning_rate": 5.0491323437173325e-06, |
| "loss": 0.0248, |
| "step": 2900 |
| }, |
| { |
| "epoch": 2.74, |
| "grad_norm": 0.3311866819858551, |
| "learning_rate": 5.028224963412085e-06, |
| "loss": 0.0337, |
| "step": 2910 |
| }, |
| { |
| "epoch": 2.75, |
| "grad_norm": 0.08945538103580475, |
| "learning_rate": 5.0073175831068365e-06, |
| "loss": 0.0166, |
| "step": 2920 |
| }, |
| { |
| "epoch": 2.76, |
| "grad_norm": 0.47838959097862244, |
| "learning_rate": 4.986410202801589e-06, |
| "loss": 0.0146, |
| "step": 2930 |
| }, |
| { |
| "epoch": 2.77, |
| "grad_norm": 0.0572611466050148, |
| "learning_rate": 4.965502822496341e-06, |
| "loss": 0.0086, |
| "step": 2940 |
| }, |
| { |
| "epoch": 2.78, |
| "grad_norm": 0.1355314403772354, |
| "learning_rate": 4.944595442191094e-06, |
| "loss": 0.0141, |
| "step": 2950 |
| }, |
| { |
| "epoch": 2.78, |
| "grad_norm": 0.10962551087141037, |
| "learning_rate": 4.923688061885846e-06, |
| "loss": 0.025, |
| "step": 2960 |
| }, |
| { |
| "epoch": 2.79, |
| "grad_norm": 0.3396119177341461, |
| "learning_rate": 4.902780681580598e-06, |
| "loss": 0.0171, |
| "step": 2970 |
| }, |
| { |
| "epoch": 2.8, |
| "grad_norm": 0.13699981570243835, |
| "learning_rate": 4.88187330127535e-06, |
| "loss": 0.012, |
| "step": 2980 |
| }, |
| { |
| "epoch": 2.81, |
| "grad_norm": 0.298997700214386, |
| "learning_rate": 4.860965920970102e-06, |
| "loss": 0.0218, |
| "step": 2990 |
| }, |
| { |
| "epoch": 2.82, |
| "grad_norm": 0.4527345895767212, |
| "learning_rate": 4.840058540664855e-06, |
| "loss": 0.0212, |
| "step": 3000 |
| }, |
| { |
| "epoch": 2.83, |
| "grad_norm": 0.15285594761371613, |
| "learning_rate": 4.819151160359607e-06, |
| "loss": 0.0162, |
| "step": 3010 |
| }, |
| { |
| "epoch": 2.84, |
| "grad_norm": 0.048979099839925766, |
| "learning_rate": 4.798243780054359e-06, |
| "loss": 0.0174, |
| "step": 3020 |
| }, |
| { |
| "epoch": 2.85, |
| "grad_norm": 0.05153890699148178, |
| "learning_rate": 4.777336399749112e-06, |
| "loss": 0.0149, |
| "step": 3030 |
| }, |
| { |
| "epoch": 2.86, |
| "grad_norm": 0.21725508570671082, |
| "learning_rate": 4.756429019443864e-06, |
| "loss": 0.0145, |
| "step": 3040 |
| }, |
| { |
| "epoch": 2.87, |
| "grad_norm": 0.16912169754505157, |
| "learning_rate": 4.735521639138616e-06, |
| "loss": 0.0179, |
| "step": 3050 |
| }, |
| { |
| "epoch": 2.88, |
| "grad_norm": 0.03088819980621338, |
| "learning_rate": 4.714614258833369e-06, |
| "loss": 0.0203, |
| "step": 3060 |
| }, |
| { |
| "epoch": 2.89, |
| "grad_norm": 0.17241007089614868, |
| "learning_rate": 4.693706878528121e-06, |
| "loss": 0.0151, |
| "step": 3070 |
| }, |
| { |
| "epoch": 2.9, |
| "grad_norm": 0.1420195996761322, |
| "learning_rate": 4.672799498222873e-06, |
| "loss": 0.0215, |
| "step": 3080 |
| }, |
| { |
| "epoch": 2.91, |
| "grad_norm": 0.23884275555610657, |
| "learning_rate": 4.651892117917625e-06, |
| "loss": 0.0135, |
| "step": 3090 |
| }, |
| { |
| "epoch": 2.92, |
| "grad_norm": 0.15318578481674194, |
| "learning_rate": 4.630984737612377e-06, |
| "loss": 0.0147, |
| "step": 3100 |
| }, |
| { |
| "epoch": 2.93, |
| "grad_norm": 0.20665739476680756, |
| "learning_rate": 4.61007735730713e-06, |
| "loss": 0.0143, |
| "step": 3110 |
| }, |
| { |
| "epoch": 2.94, |
| "grad_norm": 0.3664402663707733, |
| "learning_rate": 4.589169977001882e-06, |
| "loss": 0.0182, |
| "step": 3120 |
| }, |
| { |
| "epoch": 2.94, |
| "grad_norm": 0.3214040994644165, |
| "learning_rate": 4.568262596696634e-06, |
| "loss": 0.0199, |
| "step": 3130 |
| }, |
| { |
| "epoch": 2.95, |
| "grad_norm": 0.07582589983940125, |
| "learning_rate": 4.547355216391387e-06, |
| "loss": 0.012, |
| "step": 3140 |
| }, |
| { |
| "epoch": 2.96, |
| "grad_norm": 0.18084678053855896, |
| "learning_rate": 4.526447836086139e-06, |
| "loss": 0.0275, |
| "step": 3150 |
| }, |
| { |
| "epoch": 2.97, |
| "grad_norm": 0.9020859003067017, |
| "learning_rate": 4.505540455780891e-06, |
| "loss": 0.0208, |
| "step": 3160 |
| }, |
| { |
| "epoch": 2.98, |
| "grad_norm": 0.25799816846847534, |
| "learning_rate": 4.484633075475644e-06, |
| "loss": 0.0135, |
| "step": 3170 |
| }, |
| { |
| "epoch": 2.99, |
| "grad_norm": 0.10012619942426682, |
| "learning_rate": 4.463725695170396e-06, |
| "loss": 0.0202, |
| "step": 3180 |
| } |
| ], |
| "logging_steps": 10, |
| "max_steps": 5315, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 5, |
| "save_steps": 500, |
| "total_flos": 1.380935358480384e+18, |
| "train_batch_size": 224, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|