| { |
| "best_global_step": null, |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 0.9998474446987032, |
| "eval_steps": 500, |
| "global_step": 3277, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.0030511060259344014, |
| "grad_norm": 4.5989179611206055, |
| "learning_rate": 9e-06, |
| "loss": 1.1046, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.006102212051868803, |
| "grad_norm": 4.813294887542725, |
| "learning_rate": 1.9e-05, |
| "loss": 0.8003, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.009153318077803204, |
| "grad_norm": 4.563686847686768, |
| "learning_rate": 2.9e-05, |
| "loss": 0.7886, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.012204424103737605, |
| "grad_norm": 5.083153247833252, |
| "learning_rate": 3.9000000000000006e-05, |
| "loss": 0.7638, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.015255530129672006, |
| "grad_norm": 4.403526306152344, |
| "learning_rate": 4.9e-05, |
| "loss": 0.7635, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.018306636155606407, |
| "grad_norm": 3.6779983043670654, |
| "learning_rate": 4.986055159590952e-05, |
| "loss": 0.841, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.02135774218154081, |
| "grad_norm": 3.7986652851104736, |
| "learning_rate": 4.970560892469787e-05, |
| "loss": 0.9439, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.02440884820747521, |
| "grad_norm": 3.810528516769409, |
| "learning_rate": 4.955066625348621e-05, |
| "loss": 0.843, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.02745995423340961, |
| "grad_norm": 3.7793450355529785, |
| "learning_rate": 4.939572358227456e-05, |
| "loss": 0.9459, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.03051106025934401, |
| "grad_norm": 4.002682685852051, |
| "learning_rate": 4.924078091106291e-05, |
| "loss": 0.9242, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.033562166285278416, |
| "grad_norm": 3.5819356441497803, |
| "learning_rate": 4.908583823985126e-05, |
| "loss": 0.8683, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.036613272311212815, |
| "grad_norm": 3.728360891342163, |
| "learning_rate": 4.893089556863961e-05, |
| "loss": 0.9133, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.03966437833714721, |
| "grad_norm": 4.5038981437683105, |
| "learning_rate": 4.8775952897427956e-05, |
| "loss": 0.8466, |
| "step": 130 |
| }, |
| { |
| "epoch": 0.04271548436308162, |
| "grad_norm": 3.0992960929870605, |
| "learning_rate": 4.8621010226216305e-05, |
| "loss": 0.8424, |
| "step": 140 |
| }, |
| { |
| "epoch": 0.04576659038901602, |
| "grad_norm": 3.688044786453247, |
| "learning_rate": 4.8466067555004654e-05, |
| "loss": 0.8283, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.04881769641495042, |
| "grad_norm": 4.070577621459961, |
| "learning_rate": 4.8311124883792996e-05, |
| "loss": 0.9207, |
| "step": 160 |
| }, |
| { |
| "epoch": 0.05186880244088482, |
| "grad_norm": 3.9466311931610107, |
| "learning_rate": 4.8156182212581345e-05, |
| "loss": 0.8095, |
| "step": 170 |
| }, |
| { |
| "epoch": 0.05491990846681922, |
| "grad_norm": 3.0422556400299072, |
| "learning_rate": 4.8001239541369694e-05, |
| "loss": 0.8649, |
| "step": 180 |
| }, |
| { |
| "epoch": 0.057971014492753624, |
| "grad_norm": 3.414745330810547, |
| "learning_rate": 4.784629687015804e-05, |
| "loss": 0.8963, |
| "step": 190 |
| }, |
| { |
| "epoch": 0.06102212051868802, |
| "grad_norm": 3.7459030151367188, |
| "learning_rate": 4.769135419894639e-05, |
| "loss": 0.8352, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.06407322654462243, |
| "grad_norm": 2.6838042736053467, |
| "learning_rate": 4.753641152773474e-05, |
| "loss": 0.8089, |
| "step": 210 |
| }, |
| { |
| "epoch": 0.06712433257055683, |
| "grad_norm": 5.595691680908203, |
| "learning_rate": 4.738146885652309e-05, |
| "loss": 0.8701, |
| "step": 220 |
| }, |
| { |
| "epoch": 0.07017543859649122, |
| "grad_norm": 3.1771981716156006, |
| "learning_rate": 4.722652618531144e-05, |
| "loss": 0.8714, |
| "step": 230 |
| }, |
| { |
| "epoch": 0.07322654462242563, |
| "grad_norm": 2.728522539138794, |
| "learning_rate": 4.707158351409979e-05, |
| "loss": 0.8655, |
| "step": 240 |
| }, |
| { |
| "epoch": 0.07627765064836003, |
| "grad_norm": 3.781033515930176, |
| "learning_rate": 4.691664084288813e-05, |
| "loss": 0.7953, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.07932875667429443, |
| "grad_norm": 2.9846785068511963, |
| "learning_rate": 4.676169817167648e-05, |
| "loss": 0.8079, |
| "step": 260 |
| }, |
| { |
| "epoch": 0.08237986270022883, |
| "grad_norm": 3.5596115589141846, |
| "learning_rate": 4.6606755500464835e-05, |
| "loss": 0.8413, |
| "step": 270 |
| }, |
| { |
| "epoch": 0.08543096872616324, |
| "grad_norm": 3.188471555709839, |
| "learning_rate": 4.645181282925318e-05, |
| "loss": 0.8411, |
| "step": 280 |
| }, |
| { |
| "epoch": 0.08848207475209764, |
| "grad_norm": 4.387840270996094, |
| "learning_rate": 4.629687015804153e-05, |
| "loss": 0.9073, |
| "step": 290 |
| }, |
| { |
| "epoch": 0.09153318077803203, |
| "grad_norm": 3.842625379562378, |
| "learning_rate": 4.6141927486829876e-05, |
| "loss": 0.8269, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.09458428680396644, |
| "grad_norm": 4.832952499389648, |
| "learning_rate": 4.5986984815618225e-05, |
| "loss": 0.9064, |
| "step": 310 |
| }, |
| { |
| "epoch": 0.09763539282990084, |
| "grad_norm": 2.8753535747528076, |
| "learning_rate": 4.5832042144406574e-05, |
| "loss": 0.8873, |
| "step": 320 |
| }, |
| { |
| "epoch": 0.10068649885583524, |
| "grad_norm": 2.5598573684692383, |
| "learning_rate": 4.5677099473194916e-05, |
| "loss": 0.8913, |
| "step": 330 |
| }, |
| { |
| "epoch": 0.10373760488176964, |
| "grad_norm": 2.9012229442596436, |
| "learning_rate": 4.5522156801983265e-05, |
| "loss": 0.7988, |
| "step": 340 |
| }, |
| { |
| "epoch": 0.10678871090770405, |
| "grad_norm": 2.888648748397827, |
| "learning_rate": 4.536721413077162e-05, |
| "loss": 0.8404, |
| "step": 350 |
| }, |
| { |
| "epoch": 0.10983981693363844, |
| "grad_norm": 3.794020891189575, |
| "learning_rate": 4.521227145955996e-05, |
| "loss": 0.8429, |
| "step": 360 |
| }, |
| { |
| "epoch": 0.11289092295957284, |
| "grad_norm": 3.925490617752075, |
| "learning_rate": 4.505732878834831e-05, |
| "loss": 0.8768, |
| "step": 370 |
| }, |
| { |
| "epoch": 0.11594202898550725, |
| "grad_norm": 3.4535255432128906, |
| "learning_rate": 4.490238611713666e-05, |
| "loss": 0.7951, |
| "step": 380 |
| }, |
| { |
| "epoch": 0.11899313501144165, |
| "grad_norm": 3.9904849529266357, |
| "learning_rate": 4.474744344592501e-05, |
| "loss": 0.8524, |
| "step": 390 |
| }, |
| { |
| "epoch": 0.12204424103737604, |
| "grad_norm": 2.6464405059814453, |
| "learning_rate": 4.459250077471336e-05, |
| "loss": 0.8194, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.12509534706331046, |
| "grad_norm": 3.6185855865478516, |
| "learning_rate": 4.443755810350171e-05, |
| "loss": 0.7706, |
| "step": 410 |
| }, |
| { |
| "epoch": 0.12814645308924486, |
| "grad_norm": 4.222907543182373, |
| "learning_rate": 4.428261543229005e-05, |
| "loss": 0.7858, |
| "step": 420 |
| }, |
| { |
| "epoch": 0.13119755911517925, |
| "grad_norm": 2.8912553787231445, |
| "learning_rate": 4.4127672761078406e-05, |
| "loss": 0.7658, |
| "step": 430 |
| }, |
| { |
| "epoch": 0.13424866514111367, |
| "grad_norm": 2.9529941082000732, |
| "learning_rate": 4.3972730089866755e-05, |
| "loss": 0.8415, |
| "step": 440 |
| }, |
| { |
| "epoch": 0.13729977116704806, |
| "grad_norm": 3.296875238418579, |
| "learning_rate": 4.38177874186551e-05, |
| "loss": 0.818, |
| "step": 450 |
| }, |
| { |
| "epoch": 0.14035087719298245, |
| "grad_norm": 2.736250400543213, |
| "learning_rate": 4.366284474744345e-05, |
| "loss": 0.819, |
| "step": 460 |
| }, |
| { |
| "epoch": 0.14340198321891687, |
| "grad_norm": 3.2874772548675537, |
| "learning_rate": 4.3507902076231796e-05, |
| "loss": 0.8381, |
| "step": 470 |
| }, |
| { |
| "epoch": 0.14645308924485126, |
| "grad_norm": 2.9855446815490723, |
| "learning_rate": 4.3352959405020145e-05, |
| "loss": 0.7591, |
| "step": 480 |
| }, |
| { |
| "epoch": 0.14950419527078565, |
| "grad_norm": 3.37127947807312, |
| "learning_rate": 4.3198016733808494e-05, |
| "loss": 0.7815, |
| "step": 490 |
| }, |
| { |
| "epoch": 0.15255530129672007, |
| "grad_norm": 3.162848472595215, |
| "learning_rate": 4.3043074062596836e-05, |
| "loss": 0.8115, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.15560640732265446, |
| "grad_norm": 3.0527734756469727, |
| "learning_rate": 4.288813139138519e-05, |
| "loss": 0.8871, |
| "step": 510 |
| }, |
| { |
| "epoch": 0.15865751334858885, |
| "grad_norm": 3.1467580795288086, |
| "learning_rate": 4.273318872017354e-05, |
| "loss": 0.9112, |
| "step": 520 |
| }, |
| { |
| "epoch": 0.16170861937452327, |
| "grad_norm": 3.414907455444336, |
| "learning_rate": 4.257824604896188e-05, |
| "loss": 0.8089, |
| "step": 530 |
| }, |
| { |
| "epoch": 0.16475972540045766, |
| "grad_norm": 3.9576640129089355, |
| "learning_rate": 4.242330337775023e-05, |
| "loss": 0.7847, |
| "step": 540 |
| }, |
| { |
| "epoch": 0.16781083142639205, |
| "grad_norm": 3.8652496337890625, |
| "learning_rate": 4.226836070653858e-05, |
| "loss": 0.801, |
| "step": 550 |
| }, |
| { |
| "epoch": 0.17086193745232647, |
| "grad_norm": 3.028764247894287, |
| "learning_rate": 4.211341803532693e-05, |
| "loss": 0.8323, |
| "step": 560 |
| }, |
| { |
| "epoch": 0.17391304347826086, |
| "grad_norm": 3.1638927459716797, |
| "learning_rate": 4.195847536411528e-05, |
| "loss": 0.8575, |
| "step": 570 |
| }, |
| { |
| "epoch": 0.17696414950419528, |
| "grad_norm": 3.3378586769104004, |
| "learning_rate": 4.180353269290363e-05, |
| "loss": 0.8048, |
| "step": 580 |
| }, |
| { |
| "epoch": 0.18001525553012968, |
| "grad_norm": 2.5703723430633545, |
| "learning_rate": 4.164859002169198e-05, |
| "loss": 0.8557, |
| "step": 590 |
| }, |
| { |
| "epoch": 0.18306636155606407, |
| "grad_norm": 4.101804733276367, |
| "learning_rate": 4.1493647350480326e-05, |
| "loss": 0.8358, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.18611746758199849, |
| "grad_norm": 3.845407009124756, |
| "learning_rate": 4.1338704679268675e-05, |
| "loss": 0.8777, |
| "step": 610 |
| }, |
| { |
| "epoch": 0.18916857360793288, |
| "grad_norm": 3.370732545852661, |
| "learning_rate": 4.118376200805702e-05, |
| "loss": 0.8485, |
| "step": 620 |
| }, |
| { |
| "epoch": 0.19221967963386727, |
| "grad_norm": 2.168769121170044, |
| "learning_rate": 4.102881933684537e-05, |
| "loss": 0.7416, |
| "step": 630 |
| }, |
| { |
| "epoch": 0.1952707856598017, |
| "grad_norm": 3.819798707962036, |
| "learning_rate": 4.087387666563372e-05, |
| "loss": 0.9199, |
| "step": 640 |
| }, |
| { |
| "epoch": 0.19832189168573608, |
| "grad_norm": 3.19429612159729, |
| "learning_rate": 4.0718933994422065e-05, |
| "loss": 0.7747, |
| "step": 650 |
| }, |
| { |
| "epoch": 0.20137299771167047, |
| "grad_norm": 3.17229962348938, |
| "learning_rate": 4.0563991323210414e-05, |
| "loss": 0.7394, |
| "step": 660 |
| }, |
| { |
| "epoch": 0.2044241037376049, |
| "grad_norm": 2.316675901412964, |
| "learning_rate": 4.040904865199876e-05, |
| "loss": 0.7368, |
| "step": 670 |
| }, |
| { |
| "epoch": 0.20747520976353928, |
| "grad_norm": 3.350698471069336, |
| "learning_rate": 4.025410598078711e-05, |
| "loss": 0.8839, |
| "step": 680 |
| }, |
| { |
| "epoch": 0.21052631578947367, |
| "grad_norm": 2.5842721462249756, |
| "learning_rate": 4.009916330957546e-05, |
| "loss": 0.7665, |
| "step": 690 |
| }, |
| { |
| "epoch": 0.2135774218154081, |
| "grad_norm": 2.9108049869537354, |
| "learning_rate": 3.99442206383638e-05, |
| "loss": 0.8145, |
| "step": 700 |
| }, |
| { |
| "epoch": 0.21662852784134248, |
| "grad_norm": 2.9678006172180176, |
| "learning_rate": 3.978927796715215e-05, |
| "loss": 0.7719, |
| "step": 710 |
| }, |
| { |
| "epoch": 0.21967963386727687, |
| "grad_norm": 2.911322832107544, |
| "learning_rate": 3.963433529594051e-05, |
| "loss": 0.7354, |
| "step": 720 |
| }, |
| { |
| "epoch": 0.2227307398932113, |
| "grad_norm": 3.1599249839782715, |
| "learning_rate": 3.947939262472885e-05, |
| "loss": 0.7873, |
| "step": 730 |
| }, |
| { |
| "epoch": 0.22578184591914569, |
| "grad_norm": 2.6644296646118164, |
| "learning_rate": 3.93244499535172e-05, |
| "loss": 0.8311, |
| "step": 740 |
| }, |
| { |
| "epoch": 0.2288329519450801, |
| "grad_norm": 3.263995885848999, |
| "learning_rate": 3.916950728230555e-05, |
| "loss": 0.8185, |
| "step": 750 |
| }, |
| { |
| "epoch": 0.2318840579710145, |
| "grad_norm": 2.542343854904175, |
| "learning_rate": 3.90145646110939e-05, |
| "loss": 0.7769, |
| "step": 760 |
| }, |
| { |
| "epoch": 0.2349351639969489, |
| "grad_norm": 2.4859440326690674, |
| "learning_rate": 3.8859621939882246e-05, |
| "loss": 0.8328, |
| "step": 770 |
| }, |
| { |
| "epoch": 0.2379862700228833, |
| "grad_norm": 2.632722854614258, |
| "learning_rate": 3.8704679268670596e-05, |
| "loss": 0.7859, |
| "step": 780 |
| }, |
| { |
| "epoch": 0.2410373760488177, |
| "grad_norm": 3.209907054901123, |
| "learning_rate": 3.854973659745894e-05, |
| "loss": 0.8394, |
| "step": 790 |
| }, |
| { |
| "epoch": 0.2440884820747521, |
| "grad_norm": 2.4981114864349365, |
| "learning_rate": 3.8394793926247294e-05, |
| "loss": 0.8241, |
| "step": 800 |
| }, |
| { |
| "epoch": 0.2471395881006865, |
| "grad_norm": 2.5217268466949463, |
| "learning_rate": 3.823985125503564e-05, |
| "loss": 0.7501, |
| "step": 810 |
| }, |
| { |
| "epoch": 0.2501906941266209, |
| "grad_norm": 3.676504611968994, |
| "learning_rate": 3.8084908583823985e-05, |
| "loss": 0.8782, |
| "step": 820 |
| }, |
| { |
| "epoch": 0.2532418001525553, |
| "grad_norm": 2.651146173477173, |
| "learning_rate": 3.7929965912612334e-05, |
| "loss": 0.848, |
| "step": 830 |
| }, |
| { |
| "epoch": 0.2562929061784897, |
| "grad_norm": 3.1401736736297607, |
| "learning_rate": 3.777502324140068e-05, |
| "loss": 0.7991, |
| "step": 840 |
| }, |
| { |
| "epoch": 0.2593440122044241, |
| "grad_norm": 3.0251362323760986, |
| "learning_rate": 3.762008057018903e-05, |
| "loss": 0.8723, |
| "step": 850 |
| }, |
| { |
| "epoch": 0.2623951182303585, |
| "grad_norm": 2.1540679931640625, |
| "learning_rate": 3.746513789897738e-05, |
| "loss": 0.7478, |
| "step": 860 |
| }, |
| { |
| "epoch": 0.2654462242562929, |
| "grad_norm": 2.6065683364868164, |
| "learning_rate": 3.731019522776572e-05, |
| "loss": 0.7488, |
| "step": 870 |
| }, |
| { |
| "epoch": 0.26849733028222733, |
| "grad_norm": 2.697063684463501, |
| "learning_rate": 3.715525255655408e-05, |
| "loss": 0.7778, |
| "step": 880 |
| }, |
| { |
| "epoch": 0.2715484363081617, |
| "grad_norm": 3.4640157222747803, |
| "learning_rate": 3.700030988534243e-05, |
| "loss": 1.0205, |
| "step": 890 |
| }, |
| { |
| "epoch": 0.2745995423340961, |
| "grad_norm": 3.5426061153411865, |
| "learning_rate": 3.684536721413077e-05, |
| "loss": 0.762, |
| "step": 900 |
| }, |
| { |
| "epoch": 0.2776506483600305, |
| "grad_norm": 2.219358444213867, |
| "learning_rate": 3.669042454291912e-05, |
| "loss": 0.7996, |
| "step": 910 |
| }, |
| { |
| "epoch": 0.2807017543859649, |
| "grad_norm": 2.7404627799987793, |
| "learning_rate": 3.6535481871707475e-05, |
| "loss": 0.7611, |
| "step": 920 |
| }, |
| { |
| "epoch": 0.2837528604118993, |
| "grad_norm": 2.950205087661743, |
| "learning_rate": 3.638053920049582e-05, |
| "loss": 0.7685, |
| "step": 930 |
| }, |
| { |
| "epoch": 0.28680396643783374, |
| "grad_norm": 2.676258087158203, |
| "learning_rate": 3.6225596529284167e-05, |
| "loss": 0.7864, |
| "step": 940 |
| }, |
| { |
| "epoch": 0.2898550724637681, |
| "grad_norm": 2.5284500122070312, |
| "learning_rate": 3.6070653858072516e-05, |
| "loss": 0.7856, |
| "step": 950 |
| }, |
| { |
| "epoch": 0.2929061784897025, |
| "grad_norm": 2.2837953567504883, |
| "learning_rate": 3.5915711186860865e-05, |
| "loss": 0.7965, |
| "step": 960 |
| }, |
| { |
| "epoch": 0.2959572845156369, |
| "grad_norm": 3.223839044570923, |
| "learning_rate": 3.5760768515649214e-05, |
| "loss": 0.8123, |
| "step": 970 |
| }, |
| { |
| "epoch": 0.2990083905415713, |
| "grad_norm": 2.8053982257843018, |
| "learning_rate": 3.560582584443756e-05, |
| "loss": 0.7813, |
| "step": 980 |
| }, |
| { |
| "epoch": 0.30205949656750575, |
| "grad_norm": 2.799560785293579, |
| "learning_rate": 3.5450883173225905e-05, |
| "loss": 0.8418, |
| "step": 990 |
| }, |
| { |
| "epoch": 0.30511060259344014, |
| "grad_norm": 3.838677406311035, |
| "learning_rate": 3.529594050201426e-05, |
| "loss": 0.7948, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.30816170861937453, |
| "grad_norm": 2.587244749069214, |
| "learning_rate": 3.51409978308026e-05, |
| "loss": 0.784, |
| "step": 1010 |
| }, |
| { |
| "epoch": 0.3112128146453089, |
| "grad_norm": 3.555494785308838, |
| "learning_rate": 3.498605515959095e-05, |
| "loss": 0.7854, |
| "step": 1020 |
| }, |
| { |
| "epoch": 0.3142639206712433, |
| "grad_norm": 2.900275468826294, |
| "learning_rate": 3.48311124883793e-05, |
| "loss": 0.8536, |
| "step": 1030 |
| }, |
| { |
| "epoch": 0.3173150266971777, |
| "grad_norm": 3.058786153793335, |
| "learning_rate": 3.467616981716765e-05, |
| "loss": 0.8163, |
| "step": 1040 |
| }, |
| { |
| "epoch": 0.32036613272311215, |
| "grad_norm": 3.0138072967529297, |
| "learning_rate": 3.4521227145956e-05, |
| "loss": 0.8534, |
| "step": 1050 |
| }, |
| { |
| "epoch": 0.32341723874904654, |
| "grad_norm": 2.71114182472229, |
| "learning_rate": 3.436628447474435e-05, |
| "loss": 0.8401, |
| "step": 1060 |
| }, |
| { |
| "epoch": 0.32646834477498093, |
| "grad_norm": 2.574453592300415, |
| "learning_rate": 3.421134180353269e-05, |
| "loss": 0.7596, |
| "step": 1070 |
| }, |
| { |
| "epoch": 0.3295194508009153, |
| "grad_norm": 2.7125132083892822, |
| "learning_rate": 3.4056399132321046e-05, |
| "loss": 0.8042, |
| "step": 1080 |
| }, |
| { |
| "epoch": 0.3325705568268497, |
| "grad_norm": 3.0520405769348145, |
| "learning_rate": 3.3901456461109395e-05, |
| "loss": 0.7713, |
| "step": 1090 |
| }, |
| { |
| "epoch": 0.3356216628527841, |
| "grad_norm": 2.5527734756469727, |
| "learning_rate": 3.374651378989774e-05, |
| "loss": 0.7929, |
| "step": 1100 |
| }, |
| { |
| "epoch": 0.33867276887871856, |
| "grad_norm": 2.1792235374450684, |
| "learning_rate": 3.3591571118686087e-05, |
| "loss": 0.7102, |
| "step": 1110 |
| }, |
| { |
| "epoch": 0.34172387490465295, |
| "grad_norm": 3.192811965942383, |
| "learning_rate": 3.3436628447474436e-05, |
| "loss": 0.7956, |
| "step": 1120 |
| }, |
| { |
| "epoch": 0.34477498093058734, |
| "grad_norm": 3.0279996395111084, |
| "learning_rate": 3.3281685776262785e-05, |
| "loss": 0.7883, |
| "step": 1130 |
| }, |
| { |
| "epoch": 0.34782608695652173, |
| "grad_norm": 3.0758237838745117, |
| "learning_rate": 3.3126743105051134e-05, |
| "loss": 0.7457, |
| "step": 1140 |
| }, |
| { |
| "epoch": 0.3508771929824561, |
| "grad_norm": 2.2491872310638428, |
| "learning_rate": 3.297180043383948e-05, |
| "loss": 0.7712, |
| "step": 1150 |
| }, |
| { |
| "epoch": 0.35392829900839057, |
| "grad_norm": 2.543558120727539, |
| "learning_rate": 3.281685776262783e-05, |
| "loss": 0.8013, |
| "step": 1160 |
| }, |
| { |
| "epoch": 0.35697940503432496, |
| "grad_norm": 2.996823310852051, |
| "learning_rate": 3.266191509141618e-05, |
| "loss": 0.7769, |
| "step": 1170 |
| }, |
| { |
| "epoch": 0.36003051106025935, |
| "grad_norm": 2.1455206871032715, |
| "learning_rate": 3.250697242020452e-05, |
| "loss": 0.7778, |
| "step": 1180 |
| }, |
| { |
| "epoch": 0.36308161708619374, |
| "grad_norm": 3.1174466609954834, |
| "learning_rate": 3.235202974899287e-05, |
| "loss": 0.7519, |
| "step": 1190 |
| }, |
| { |
| "epoch": 0.36613272311212813, |
| "grad_norm": 2.6810905933380127, |
| "learning_rate": 3.219708707778122e-05, |
| "loss": 0.8112, |
| "step": 1200 |
| }, |
| { |
| "epoch": 0.3691838291380625, |
| "grad_norm": 2.6105597019195557, |
| "learning_rate": 3.204214440656957e-05, |
| "loss": 0.7344, |
| "step": 1210 |
| }, |
| { |
| "epoch": 0.37223493516399697, |
| "grad_norm": 3.3781626224517822, |
| "learning_rate": 3.188720173535792e-05, |
| "loss": 0.8401, |
| "step": 1220 |
| }, |
| { |
| "epoch": 0.37528604118993136, |
| "grad_norm": 3.3649299144744873, |
| "learning_rate": 3.173225906414627e-05, |
| "loss": 0.7849, |
| "step": 1230 |
| }, |
| { |
| "epoch": 0.37833714721586575, |
| "grad_norm": 2.6910207271575928, |
| "learning_rate": 3.157731639293462e-05, |
| "loss": 0.8165, |
| "step": 1240 |
| }, |
| { |
| "epoch": 0.38138825324180015, |
| "grad_norm": 2.5760819911956787, |
| "learning_rate": 3.1422373721722966e-05, |
| "loss": 0.8043, |
| "step": 1250 |
| }, |
| { |
| "epoch": 0.38443935926773454, |
| "grad_norm": 2.5498452186584473, |
| "learning_rate": 3.1267431050511315e-05, |
| "loss": 0.7566, |
| "step": 1260 |
| }, |
| { |
| "epoch": 0.38749046529366893, |
| "grad_norm": 1.8285728693008423, |
| "learning_rate": 3.111248837929966e-05, |
| "loss": 0.7152, |
| "step": 1270 |
| }, |
| { |
| "epoch": 0.3905415713196034, |
| "grad_norm": 2.509260654449463, |
| "learning_rate": 3.0957545708088007e-05, |
| "loss": 0.8489, |
| "step": 1280 |
| }, |
| { |
| "epoch": 0.39359267734553777, |
| "grad_norm": 2.905968427658081, |
| "learning_rate": 3.080260303687636e-05, |
| "loss": 0.7369, |
| "step": 1290 |
| }, |
| { |
| "epoch": 0.39664378337147216, |
| "grad_norm": 2.5083484649658203, |
| "learning_rate": 3.0647660365664705e-05, |
| "loss": 0.7331, |
| "step": 1300 |
| }, |
| { |
| "epoch": 0.39969488939740655, |
| "grad_norm": 2.2442944049835205, |
| "learning_rate": 3.0492717694453054e-05, |
| "loss": 0.7282, |
| "step": 1310 |
| }, |
| { |
| "epoch": 0.40274599542334094, |
| "grad_norm": 2.856735944747925, |
| "learning_rate": 3.0337775023241406e-05, |
| "loss": 0.7009, |
| "step": 1320 |
| }, |
| { |
| "epoch": 0.4057971014492754, |
| "grad_norm": 2.9385952949523926, |
| "learning_rate": 3.018283235202975e-05, |
| "loss": 0.7225, |
| "step": 1330 |
| }, |
| { |
| "epoch": 0.4088482074752098, |
| "grad_norm": 3.0979793071746826, |
| "learning_rate": 3.00278896808181e-05, |
| "loss": 0.764, |
| "step": 1340 |
| }, |
| { |
| "epoch": 0.41189931350114417, |
| "grad_norm": 2.6863820552825928, |
| "learning_rate": 2.9872947009606443e-05, |
| "loss": 0.7475, |
| "step": 1350 |
| }, |
| { |
| "epoch": 0.41495041952707856, |
| "grad_norm": 2.9140465259552, |
| "learning_rate": 2.9718004338394795e-05, |
| "loss": 0.7468, |
| "step": 1360 |
| }, |
| { |
| "epoch": 0.41800152555301295, |
| "grad_norm": 2.702855348587036, |
| "learning_rate": 2.9563061667183144e-05, |
| "loss": 0.7444, |
| "step": 1370 |
| }, |
| { |
| "epoch": 0.42105263157894735, |
| "grad_norm": 2.646045207977295, |
| "learning_rate": 2.940811899597149e-05, |
| "loss": 0.7911, |
| "step": 1380 |
| }, |
| { |
| "epoch": 0.4241037376048818, |
| "grad_norm": 2.568760395050049, |
| "learning_rate": 2.925317632475984e-05, |
| "loss": 0.7815, |
| "step": 1390 |
| }, |
| { |
| "epoch": 0.4271548436308162, |
| "grad_norm": 2.381138801574707, |
| "learning_rate": 2.909823365354819e-05, |
| "loss": 0.7411, |
| "step": 1400 |
| }, |
| { |
| "epoch": 0.4302059496567506, |
| "grad_norm": 2.6686244010925293, |
| "learning_rate": 2.8943290982336534e-05, |
| "loss": 0.7593, |
| "step": 1410 |
| }, |
| { |
| "epoch": 0.43325705568268497, |
| "grad_norm": 3.8091323375701904, |
| "learning_rate": 2.8788348311124886e-05, |
| "loss": 0.7731, |
| "step": 1420 |
| }, |
| { |
| "epoch": 0.43630816170861936, |
| "grad_norm": 3.0860376358032227, |
| "learning_rate": 2.8633405639913235e-05, |
| "loss": 0.7617, |
| "step": 1430 |
| }, |
| { |
| "epoch": 0.43935926773455375, |
| "grad_norm": 3.304323196411133, |
| "learning_rate": 2.847846296870158e-05, |
| "loss": 0.8343, |
| "step": 1440 |
| }, |
| { |
| "epoch": 0.4424103737604882, |
| "grad_norm": 2.6195857524871826, |
| "learning_rate": 2.832352029748993e-05, |
| "loss": 0.688, |
| "step": 1450 |
| }, |
| { |
| "epoch": 0.4454614797864226, |
| "grad_norm": 2.3896477222442627, |
| "learning_rate": 2.8168577626278282e-05, |
| "loss": 0.8032, |
| "step": 1460 |
| }, |
| { |
| "epoch": 0.448512585812357, |
| "grad_norm": 2.9045519828796387, |
| "learning_rate": 2.8013634955066625e-05, |
| "loss": 0.8131, |
| "step": 1470 |
| }, |
| { |
| "epoch": 0.45156369183829137, |
| "grad_norm": 2.5903823375701904, |
| "learning_rate": 2.7858692283854977e-05, |
| "loss": 0.7219, |
| "step": 1480 |
| }, |
| { |
| "epoch": 0.45461479786422576, |
| "grad_norm": 2.8301548957824707, |
| "learning_rate": 2.7703749612643326e-05, |
| "loss": 0.7661, |
| "step": 1490 |
| }, |
| { |
| "epoch": 0.4576659038901602, |
| "grad_norm": 2.5708298683166504, |
| "learning_rate": 2.7548806941431672e-05, |
| "loss": 0.7673, |
| "step": 1500 |
| }, |
| { |
| "epoch": 0.4607170099160946, |
| "grad_norm": 2.4464125633239746, |
| "learning_rate": 2.739386427022002e-05, |
| "loss": 0.7328, |
| "step": 1510 |
| }, |
| { |
| "epoch": 0.463768115942029, |
| "grad_norm": 3.007063627243042, |
| "learning_rate": 2.7238921599008366e-05, |
| "loss": 0.7449, |
| "step": 1520 |
| }, |
| { |
| "epoch": 0.4668192219679634, |
| "grad_norm": 2.8109517097473145, |
| "learning_rate": 2.7083978927796715e-05, |
| "loss": 0.8403, |
| "step": 1530 |
| }, |
| { |
| "epoch": 0.4698703279938978, |
| "grad_norm": 2.756274461746216, |
| "learning_rate": 2.6929036256585068e-05, |
| "loss": 0.786, |
| "step": 1540 |
| }, |
| { |
| "epoch": 0.47292143401983217, |
| "grad_norm": 3.142169237136841, |
| "learning_rate": 2.677409358537341e-05, |
| "loss": 0.7738, |
| "step": 1550 |
| }, |
| { |
| "epoch": 0.4759725400457666, |
| "grad_norm": 2.546050548553467, |
| "learning_rate": 2.6619150914161763e-05, |
| "loss": 0.7806, |
| "step": 1560 |
| }, |
| { |
| "epoch": 0.479023646071701, |
| "grad_norm": 2.3914120197296143, |
| "learning_rate": 2.646420824295011e-05, |
| "loss": 0.7641, |
| "step": 1570 |
| }, |
| { |
| "epoch": 0.4820747520976354, |
| "grad_norm": 2.156118869781494, |
| "learning_rate": 2.6309265571738457e-05, |
| "loss": 0.8034, |
| "step": 1580 |
| }, |
| { |
| "epoch": 0.4851258581235698, |
| "grad_norm": 2.6585230827331543, |
| "learning_rate": 2.6154322900526806e-05, |
| "loss": 0.7609, |
| "step": 1590 |
| }, |
| { |
| "epoch": 0.4881769641495042, |
| "grad_norm": 2.175424814224243, |
| "learning_rate": 2.5999380229315155e-05, |
| "loss": 0.6381, |
| "step": 1600 |
| }, |
| { |
| "epoch": 0.49122807017543857, |
| "grad_norm": 3.3766794204711914, |
| "learning_rate": 2.58444375581035e-05, |
| "loss": 0.7187, |
| "step": 1610 |
| }, |
| { |
| "epoch": 0.494279176201373, |
| "grad_norm": 2.665618419647217, |
| "learning_rate": 2.568949488689185e-05, |
| "loss": 0.7492, |
| "step": 1620 |
| }, |
| { |
| "epoch": 0.4973302822273074, |
| "grad_norm": 2.5544931888580322, |
| "learning_rate": 2.5534552215680202e-05, |
| "loss": 0.7473, |
| "step": 1630 |
| }, |
| { |
| "epoch": 0.5003813882532419, |
| "grad_norm": 2.091601848602295, |
| "learning_rate": 2.5379609544468548e-05, |
| "loss": 0.7364, |
| "step": 1640 |
| }, |
| { |
| "epoch": 0.5034324942791762, |
| "grad_norm": 2.7752933502197266, |
| "learning_rate": 2.5224666873256897e-05, |
| "loss": 0.8124, |
| "step": 1650 |
| }, |
| { |
| "epoch": 0.5064836003051106, |
| "grad_norm": 2.3963379859924316, |
| "learning_rate": 2.5069724202045246e-05, |
| "loss": 0.7213, |
| "step": 1660 |
| }, |
| { |
| "epoch": 0.509534706331045, |
| "grad_norm": 2.0520660877227783, |
| "learning_rate": 2.4914781530833592e-05, |
| "loss": 0.7954, |
| "step": 1670 |
| }, |
| { |
| "epoch": 0.5125858123569794, |
| "grad_norm": 2.9152419567108154, |
| "learning_rate": 2.475983885962194e-05, |
| "loss": 0.7114, |
| "step": 1680 |
| }, |
| { |
| "epoch": 0.5156369183829138, |
| "grad_norm": 3.4632716178894043, |
| "learning_rate": 2.460489618841029e-05, |
| "loss": 0.7557, |
| "step": 1690 |
| }, |
| { |
| "epoch": 0.5186880244088482, |
| "grad_norm": 3.0776329040527344, |
| "learning_rate": 2.4449953517198636e-05, |
| "loss": 0.7867, |
| "step": 1700 |
| }, |
| { |
| "epoch": 0.5217391304347826, |
| "grad_norm": 2.640956163406372, |
| "learning_rate": 2.4295010845986985e-05, |
| "loss": 0.717, |
| "step": 1710 |
| }, |
| { |
| "epoch": 0.524790236460717, |
| "grad_norm": 3.0681939125061035, |
| "learning_rate": 2.4140068174775334e-05, |
| "loss": 0.779, |
| "step": 1720 |
| }, |
| { |
| "epoch": 0.5278413424866514, |
| "grad_norm": 3.3912854194641113, |
| "learning_rate": 2.3985125503563683e-05, |
| "loss": 0.7602, |
| "step": 1730 |
| }, |
| { |
| "epoch": 0.5308924485125858, |
| "grad_norm": 3.4418396949768066, |
| "learning_rate": 2.3830182832352028e-05, |
| "loss": 0.7719, |
| "step": 1740 |
| }, |
| { |
| "epoch": 0.5339435545385202, |
| "grad_norm": 2.5723013877868652, |
| "learning_rate": 2.367524016114038e-05, |
| "loss": 0.6693, |
| "step": 1750 |
| }, |
| { |
| "epoch": 0.5369946605644547, |
| "grad_norm": 3.5987260341644287, |
| "learning_rate": 2.3520297489928726e-05, |
| "loss": 0.7202, |
| "step": 1760 |
| }, |
| { |
| "epoch": 0.540045766590389, |
| "grad_norm": 3.0256638526916504, |
| "learning_rate": 2.3365354818717075e-05, |
| "loss": 0.769, |
| "step": 1770 |
| }, |
| { |
| "epoch": 0.5430968726163234, |
| "grad_norm": 2.5137994289398193, |
| "learning_rate": 2.3210412147505424e-05, |
| "loss": 0.7133, |
| "step": 1780 |
| }, |
| { |
| "epoch": 0.5461479786422578, |
| "grad_norm": 2.394585132598877, |
| "learning_rate": 2.3055469476293773e-05, |
| "loss": 0.7089, |
| "step": 1790 |
| }, |
| { |
| "epoch": 0.5491990846681922, |
| "grad_norm": 2.6688249111175537, |
| "learning_rate": 2.290052680508212e-05, |
| "loss": 0.7819, |
| "step": 1800 |
| }, |
| { |
| "epoch": 0.5522501906941266, |
| "grad_norm": 3.030775785446167, |
| "learning_rate": 2.2745584133870468e-05, |
| "loss": 0.7545, |
| "step": 1810 |
| }, |
| { |
| "epoch": 0.555301296720061, |
| "grad_norm": 3.2252886295318604, |
| "learning_rate": 2.2590641462658817e-05, |
| "loss": 0.7854, |
| "step": 1820 |
| }, |
| { |
| "epoch": 0.5583524027459954, |
| "grad_norm": 1.8850343227386475, |
| "learning_rate": 2.2435698791447166e-05, |
| "loss": 0.708, |
| "step": 1830 |
| }, |
| { |
| "epoch": 0.5614035087719298, |
| "grad_norm": 3.511936902999878, |
| "learning_rate": 2.2280756120235512e-05, |
| "loss": 0.7384, |
| "step": 1840 |
| }, |
| { |
| "epoch": 0.5644546147978642, |
| "grad_norm": 3.171138286590576, |
| "learning_rate": 2.2125813449023864e-05, |
| "loss": 0.7831, |
| "step": 1850 |
| }, |
| { |
| "epoch": 0.5675057208237986, |
| "grad_norm": 2.5248336791992188, |
| "learning_rate": 2.197087077781221e-05, |
| "loss": 0.7011, |
| "step": 1860 |
| }, |
| { |
| "epoch": 0.5705568268497331, |
| "grad_norm": 3.0410759449005127, |
| "learning_rate": 2.181592810660056e-05, |
| "loss": 0.7252, |
| "step": 1870 |
| }, |
| { |
| "epoch": 0.5736079328756675, |
| "grad_norm": 3.691617012023926, |
| "learning_rate": 2.1660985435388905e-05, |
| "loss": 0.7585, |
| "step": 1880 |
| }, |
| { |
| "epoch": 0.5766590389016019, |
| "grad_norm": 2.653752565383911, |
| "learning_rate": 2.1506042764177257e-05, |
| "loss": 0.7948, |
| "step": 1890 |
| }, |
| { |
| "epoch": 0.5797101449275363, |
| "grad_norm": 2.0451462268829346, |
| "learning_rate": 2.1351100092965603e-05, |
| "loss": 0.6916, |
| "step": 1900 |
| }, |
| { |
| "epoch": 0.5827612509534706, |
| "grad_norm": 2.8858132362365723, |
| "learning_rate": 2.1196157421753952e-05, |
| "loss": 0.682, |
| "step": 1910 |
| }, |
| { |
| "epoch": 0.585812356979405, |
| "grad_norm": 2.553213596343994, |
| "learning_rate": 2.10412147505423e-05, |
| "loss": 0.7501, |
| "step": 1920 |
| }, |
| { |
| "epoch": 0.5888634630053394, |
| "grad_norm": 2.1616005897521973, |
| "learning_rate": 2.088627207933065e-05, |
| "loss": 0.7536, |
| "step": 1930 |
| }, |
| { |
| "epoch": 0.5919145690312738, |
| "grad_norm": 3.390302896499634, |
| "learning_rate": 2.0731329408118995e-05, |
| "loss": 0.787, |
| "step": 1940 |
| }, |
| { |
| "epoch": 0.5949656750572082, |
| "grad_norm": 2.695875883102417, |
| "learning_rate": 2.0576386736907348e-05, |
| "loss": 0.7717, |
| "step": 1950 |
| }, |
| { |
| "epoch": 0.5980167810831426, |
| "grad_norm": 2.517509698867798, |
| "learning_rate": 2.0421444065695693e-05, |
| "loss": 0.7527, |
| "step": 1960 |
| }, |
| { |
| "epoch": 0.601067887109077, |
| "grad_norm": 2.498511552810669, |
| "learning_rate": 2.0266501394484043e-05, |
| "loss": 0.6925, |
| "step": 1970 |
| }, |
| { |
| "epoch": 0.6041189931350115, |
| "grad_norm": 2.409759044647217, |
| "learning_rate": 2.0111558723272388e-05, |
| "loss": 0.7283, |
| "step": 1980 |
| }, |
| { |
| "epoch": 0.6071700991609459, |
| "grad_norm": 3.212092876434326, |
| "learning_rate": 1.995661605206074e-05, |
| "loss": 0.6655, |
| "step": 1990 |
| }, |
| { |
| "epoch": 0.6102212051868803, |
| "grad_norm": 2.577396869659424, |
| "learning_rate": 1.9801673380849086e-05, |
| "loss": 0.6565, |
| "step": 2000 |
| }, |
| { |
| "epoch": 0.6132723112128147, |
| "grad_norm": 2.715590715408325, |
| "learning_rate": 1.9646730709637435e-05, |
| "loss": 0.6705, |
| "step": 2010 |
| }, |
| { |
| "epoch": 0.6163234172387491, |
| "grad_norm": 2.6704838275909424, |
| "learning_rate": 1.9491788038425784e-05, |
| "loss": 0.7789, |
| "step": 2020 |
| }, |
| { |
| "epoch": 0.6193745232646835, |
| "grad_norm": 2.4763495922088623, |
| "learning_rate": 1.9336845367214133e-05, |
| "loss": 0.8014, |
| "step": 2030 |
| }, |
| { |
| "epoch": 0.6224256292906178, |
| "grad_norm": 2.754671573638916, |
| "learning_rate": 1.918190269600248e-05, |
| "loss": 0.7272, |
| "step": 2040 |
| }, |
| { |
| "epoch": 0.6254767353165522, |
| "grad_norm": 3.223806858062744, |
| "learning_rate": 1.9026960024790828e-05, |
| "loss": 0.7103, |
| "step": 2050 |
| }, |
| { |
| "epoch": 0.6285278413424866, |
| "grad_norm": 2.489527702331543, |
| "learning_rate": 1.8872017353579177e-05, |
| "loss": 0.6492, |
| "step": 2060 |
| }, |
| { |
| "epoch": 0.631578947368421, |
| "grad_norm": 3.088130235671997, |
| "learning_rate": 1.8717074682367526e-05, |
| "loss": 0.6808, |
| "step": 2070 |
| }, |
| { |
| "epoch": 0.6346300533943554, |
| "grad_norm": 3.130558967590332, |
| "learning_rate": 1.8562132011155872e-05, |
| "loss": 0.7399, |
| "step": 2080 |
| }, |
| { |
| "epoch": 0.6376811594202898, |
| "grad_norm": 2.672267436981201, |
| "learning_rate": 1.840718933994422e-05, |
| "loss": 0.7012, |
| "step": 2090 |
| }, |
| { |
| "epoch": 0.6407322654462243, |
| "grad_norm": 3.445889472961426, |
| "learning_rate": 1.825224666873257e-05, |
| "loss": 0.7329, |
| "step": 2100 |
| }, |
| { |
| "epoch": 0.6437833714721587, |
| "grad_norm": 2.1221354007720947, |
| "learning_rate": 1.809730399752092e-05, |
| "loss": 0.7105, |
| "step": 2110 |
| }, |
| { |
| "epoch": 0.6468344774980931, |
| "grad_norm": 2.5181918144226074, |
| "learning_rate": 1.7942361326309268e-05, |
| "loss": 0.7719, |
| "step": 2120 |
| }, |
| { |
| "epoch": 0.6498855835240275, |
| "grad_norm": 3.100332021713257, |
| "learning_rate": 1.7787418655097614e-05, |
| "loss": 0.6837, |
| "step": 2130 |
| }, |
| { |
| "epoch": 0.6529366895499619, |
| "grad_norm": 3.2809691429138184, |
| "learning_rate": 1.7632475983885963e-05, |
| "loss": 0.7173, |
| "step": 2140 |
| }, |
| { |
| "epoch": 0.6559877955758963, |
| "grad_norm": 2.191570520401001, |
| "learning_rate": 1.747753331267431e-05, |
| "loss": 0.6494, |
| "step": 2150 |
| }, |
| { |
| "epoch": 0.6590389016018307, |
| "grad_norm": 2.8507699966430664, |
| "learning_rate": 1.732259064146266e-05, |
| "loss": 0.7094, |
| "step": 2160 |
| }, |
| { |
| "epoch": 0.662090007627765, |
| "grad_norm": 2.7249112129211426, |
| "learning_rate": 1.7167647970251006e-05, |
| "loss": 0.6549, |
| "step": 2170 |
| }, |
| { |
| "epoch": 0.6651411136536994, |
| "grad_norm": 1.7811733484268188, |
| "learning_rate": 1.7012705299039355e-05, |
| "loss": 0.6353, |
| "step": 2180 |
| }, |
| { |
| "epoch": 0.6681922196796338, |
| "grad_norm": 2.451258659362793, |
| "learning_rate": 1.6857762627827704e-05, |
| "loss": 0.7104, |
| "step": 2190 |
| }, |
| { |
| "epoch": 0.6712433257055682, |
| "grad_norm": 2.2157397270202637, |
| "learning_rate": 1.6702819956616053e-05, |
| "loss": 0.7135, |
| "step": 2200 |
| }, |
| { |
| "epoch": 0.6742944317315027, |
| "grad_norm": 2.1415586471557617, |
| "learning_rate": 1.65478772854044e-05, |
| "loss": 0.6959, |
| "step": 2210 |
| }, |
| { |
| "epoch": 0.6773455377574371, |
| "grad_norm": 2.9360268115997314, |
| "learning_rate": 1.6392934614192748e-05, |
| "loss": 0.7511, |
| "step": 2220 |
| }, |
| { |
| "epoch": 0.6803966437833715, |
| "grad_norm": 2.7337160110473633, |
| "learning_rate": 1.6237991942981097e-05, |
| "loss": 0.7181, |
| "step": 2230 |
| }, |
| { |
| "epoch": 0.6834477498093059, |
| "grad_norm": 3.827877998352051, |
| "learning_rate": 1.6083049271769446e-05, |
| "loss": 0.6735, |
| "step": 2240 |
| }, |
| { |
| "epoch": 0.6864988558352403, |
| "grad_norm": 2.284428834915161, |
| "learning_rate": 1.5928106600557792e-05, |
| "loss": 0.6822, |
| "step": 2250 |
| }, |
| { |
| "epoch": 0.6895499618611747, |
| "grad_norm": 2.9292097091674805, |
| "learning_rate": 1.5773163929346144e-05, |
| "loss": 0.7522, |
| "step": 2260 |
| }, |
| { |
| "epoch": 0.6926010678871091, |
| "grad_norm": 2.607977867126465, |
| "learning_rate": 1.561822125813449e-05, |
| "loss": 0.656, |
| "step": 2270 |
| }, |
| { |
| "epoch": 0.6956521739130435, |
| "grad_norm": 3.1531264781951904, |
| "learning_rate": 1.546327858692284e-05, |
| "loss": 0.7427, |
| "step": 2280 |
| }, |
| { |
| "epoch": 0.6987032799389779, |
| "grad_norm": 2.421393871307373, |
| "learning_rate": 1.5308335915711188e-05, |
| "loss": 0.6408, |
| "step": 2290 |
| }, |
| { |
| "epoch": 0.7017543859649122, |
| "grad_norm": 3.7150917053222656, |
| "learning_rate": 1.5153393244499537e-05, |
| "loss": 0.7536, |
| "step": 2300 |
| }, |
| { |
| "epoch": 0.7048054919908466, |
| "grad_norm": 2.622455358505249, |
| "learning_rate": 1.4998450573287884e-05, |
| "loss": 0.7359, |
| "step": 2310 |
| }, |
| { |
| "epoch": 0.7078565980167811, |
| "grad_norm": 2.524982452392578, |
| "learning_rate": 1.4843507902076232e-05, |
| "loss": 0.689, |
| "step": 2320 |
| }, |
| { |
| "epoch": 0.7109077040427155, |
| "grad_norm": 3.156085729598999, |
| "learning_rate": 1.4688565230864582e-05, |
| "loss": 0.7111, |
| "step": 2330 |
| }, |
| { |
| "epoch": 0.7139588100686499, |
| "grad_norm": 2.8505616188049316, |
| "learning_rate": 1.453362255965293e-05, |
| "loss": 0.6698, |
| "step": 2340 |
| }, |
| { |
| "epoch": 0.7170099160945843, |
| "grad_norm": 2.8530240058898926, |
| "learning_rate": 1.4378679888441277e-05, |
| "loss": 0.7196, |
| "step": 2350 |
| }, |
| { |
| "epoch": 0.7200610221205187, |
| "grad_norm": 2.8734195232391357, |
| "learning_rate": 1.4223737217229626e-05, |
| "loss": 0.7656, |
| "step": 2360 |
| }, |
| { |
| "epoch": 0.7231121281464531, |
| "grad_norm": 2.84820556640625, |
| "learning_rate": 1.4068794546017975e-05, |
| "loss": 0.6777, |
| "step": 2370 |
| }, |
| { |
| "epoch": 0.7261632341723875, |
| "grad_norm": 2.4297478199005127, |
| "learning_rate": 1.3913851874806322e-05, |
| "loss": 0.7145, |
| "step": 2380 |
| }, |
| { |
| "epoch": 0.7292143401983219, |
| "grad_norm": 2.734090566635132, |
| "learning_rate": 1.375890920359467e-05, |
| "loss": 0.6577, |
| "step": 2390 |
| }, |
| { |
| "epoch": 0.7322654462242563, |
| "grad_norm": 2.2257931232452393, |
| "learning_rate": 1.3603966532383019e-05, |
| "loss": 0.6857, |
| "step": 2400 |
| }, |
| { |
| "epoch": 0.7353165522501907, |
| "grad_norm": 2.3681576251983643, |
| "learning_rate": 1.3449023861171368e-05, |
| "loss": 0.7619, |
| "step": 2410 |
| }, |
| { |
| "epoch": 0.738367658276125, |
| "grad_norm": 2.033923864364624, |
| "learning_rate": 1.3294081189959715e-05, |
| "loss": 0.691, |
| "step": 2420 |
| }, |
| { |
| "epoch": 0.7414187643020596, |
| "grad_norm": 2.131638526916504, |
| "learning_rate": 1.3139138518748064e-05, |
| "loss": 0.7191, |
| "step": 2430 |
| }, |
| { |
| "epoch": 0.7444698703279939, |
| "grad_norm": 2.609546661376953, |
| "learning_rate": 1.2984195847536412e-05, |
| "loss": 0.6737, |
| "step": 2440 |
| }, |
| { |
| "epoch": 0.7475209763539283, |
| "grad_norm": 2.4548799991607666, |
| "learning_rate": 1.282925317632476e-05, |
| "loss": 0.6301, |
| "step": 2450 |
| }, |
| { |
| "epoch": 0.7505720823798627, |
| "grad_norm": 2.644568920135498, |
| "learning_rate": 1.267431050511311e-05, |
| "loss": 0.6824, |
| "step": 2460 |
| }, |
| { |
| "epoch": 0.7536231884057971, |
| "grad_norm": 2.595654249191284, |
| "learning_rate": 1.2519367833901457e-05, |
| "loss": 0.654, |
| "step": 2470 |
| }, |
| { |
| "epoch": 0.7566742944317315, |
| "grad_norm": 2.7976434230804443, |
| "learning_rate": 1.2364425162689804e-05, |
| "loss": 0.7883, |
| "step": 2480 |
| }, |
| { |
| "epoch": 0.7597254004576659, |
| "grad_norm": 2.874528646469116, |
| "learning_rate": 1.2209482491478153e-05, |
| "loss": 0.7202, |
| "step": 2490 |
| }, |
| { |
| "epoch": 0.7627765064836003, |
| "grad_norm": 2.6641483306884766, |
| "learning_rate": 1.20545398202665e-05, |
| "loss": 0.6692, |
| "step": 2500 |
| }, |
| { |
| "epoch": 0.7658276125095347, |
| "grad_norm": 2.610469102859497, |
| "learning_rate": 1.189959714905485e-05, |
| "loss": 0.7201, |
| "step": 2510 |
| }, |
| { |
| "epoch": 0.7688787185354691, |
| "grad_norm": 2.2959210872650146, |
| "learning_rate": 1.1744654477843199e-05, |
| "loss": 0.6302, |
| "step": 2520 |
| }, |
| { |
| "epoch": 0.7719298245614035, |
| "grad_norm": 2.454118013381958, |
| "learning_rate": 1.1589711806631546e-05, |
| "loss": 0.7102, |
| "step": 2530 |
| }, |
| { |
| "epoch": 0.7749809305873379, |
| "grad_norm": 2.6348319053649902, |
| "learning_rate": 1.1434769135419895e-05, |
| "loss": 0.7214, |
| "step": 2540 |
| }, |
| { |
| "epoch": 0.7780320366132724, |
| "grad_norm": 2.254340887069702, |
| "learning_rate": 1.1279826464208242e-05, |
| "loss": 0.6973, |
| "step": 2550 |
| }, |
| { |
| "epoch": 0.7810831426392068, |
| "grad_norm": 2.1454594135284424, |
| "learning_rate": 1.1124883792996592e-05, |
| "loss": 0.7037, |
| "step": 2560 |
| }, |
| { |
| "epoch": 0.7841342486651411, |
| "grad_norm": 2.7335195541381836, |
| "learning_rate": 1.096994112178494e-05, |
| "loss": 0.7059, |
| "step": 2570 |
| }, |
| { |
| "epoch": 0.7871853546910755, |
| "grad_norm": 3.499448776245117, |
| "learning_rate": 1.0814998450573288e-05, |
| "loss": 0.6518, |
| "step": 2580 |
| }, |
| { |
| "epoch": 0.7902364607170099, |
| "grad_norm": 2.517699956893921, |
| "learning_rate": 1.0660055779361637e-05, |
| "loss": 0.6884, |
| "step": 2590 |
| }, |
| { |
| "epoch": 0.7932875667429443, |
| "grad_norm": 2.7374229431152344, |
| "learning_rate": 1.0505113108149984e-05, |
| "loss": 0.6146, |
| "step": 2600 |
| }, |
| { |
| "epoch": 0.7963386727688787, |
| "grad_norm": 3.5942227840423584, |
| "learning_rate": 1.0350170436938333e-05, |
| "loss": 0.7788, |
| "step": 2610 |
| }, |
| { |
| "epoch": 0.7993897787948131, |
| "grad_norm": 2.709524393081665, |
| "learning_rate": 1.019522776572668e-05, |
| "loss": 0.6843, |
| "step": 2620 |
| }, |
| { |
| "epoch": 0.8024408848207475, |
| "grad_norm": 2.7754933834075928, |
| "learning_rate": 1.004028509451503e-05, |
| "loss": 0.7694, |
| "step": 2630 |
| }, |
| { |
| "epoch": 0.8054919908466819, |
| "grad_norm": 2.025585412979126, |
| "learning_rate": 9.885342423303379e-06, |
| "loss": 0.6237, |
| "step": 2640 |
| }, |
| { |
| "epoch": 0.8085430968726163, |
| "grad_norm": 2.5337514877319336, |
| "learning_rate": 9.730399752091726e-06, |
| "loss": 0.7284, |
| "step": 2650 |
| }, |
| { |
| "epoch": 0.8115942028985508, |
| "grad_norm": 2.5185277462005615, |
| "learning_rate": 9.575457080880075e-06, |
| "loss": 0.695, |
| "step": 2660 |
| }, |
| { |
| "epoch": 0.8146453089244852, |
| "grad_norm": 3.0368494987487793, |
| "learning_rate": 9.420514409668422e-06, |
| "loss": 0.7291, |
| "step": 2670 |
| }, |
| { |
| "epoch": 0.8176964149504196, |
| "grad_norm": 2.6141366958618164, |
| "learning_rate": 9.265571738456771e-06, |
| "loss": 0.6768, |
| "step": 2680 |
| }, |
| { |
| "epoch": 0.820747520976354, |
| "grad_norm": 2.2215969562530518, |
| "learning_rate": 9.11062906724512e-06, |
| "loss": 0.6832, |
| "step": 2690 |
| }, |
| { |
| "epoch": 0.8237986270022883, |
| "grad_norm": 3.0464236736297607, |
| "learning_rate": 8.955686396033468e-06, |
| "loss": 0.7362, |
| "step": 2700 |
| }, |
| { |
| "epoch": 0.8268497330282227, |
| "grad_norm": 2.384838104248047, |
| "learning_rate": 8.800743724821817e-06, |
| "loss": 0.6501, |
| "step": 2710 |
| }, |
| { |
| "epoch": 0.8299008390541571, |
| "grad_norm": 2.2411282062530518, |
| "learning_rate": 8.645801053610164e-06, |
| "loss": 0.6274, |
| "step": 2720 |
| }, |
| { |
| "epoch": 0.8329519450800915, |
| "grad_norm": 2.8559658527374268, |
| "learning_rate": 8.490858382398513e-06, |
| "loss": 0.6282, |
| "step": 2730 |
| }, |
| { |
| "epoch": 0.8360030511060259, |
| "grad_norm": 2.729321002960205, |
| "learning_rate": 8.335915711186862e-06, |
| "loss": 0.6966, |
| "step": 2740 |
| }, |
| { |
| "epoch": 0.8390541571319603, |
| "grad_norm": 2.846073865890503, |
| "learning_rate": 8.18097303997521e-06, |
| "loss": 0.6925, |
| "step": 2750 |
| }, |
| { |
| "epoch": 0.8421052631578947, |
| "grad_norm": 2.1397573947906494, |
| "learning_rate": 8.026030368763559e-06, |
| "loss": 0.6435, |
| "step": 2760 |
| }, |
| { |
| "epoch": 0.8451563691838292, |
| "grad_norm": 3.3590445518493652, |
| "learning_rate": 7.871087697551906e-06, |
| "loss": 0.6592, |
| "step": 2770 |
| }, |
| { |
| "epoch": 0.8482074752097636, |
| "grad_norm": 2.7317698001861572, |
| "learning_rate": 7.716145026340255e-06, |
| "loss": 0.6915, |
| "step": 2780 |
| }, |
| { |
| "epoch": 0.851258581235698, |
| "grad_norm": 2.3746912479400635, |
| "learning_rate": 7.561202355128602e-06, |
| "loss": 0.7039, |
| "step": 2790 |
| }, |
| { |
| "epoch": 0.8543096872616324, |
| "grad_norm": 3.0164999961853027, |
| "learning_rate": 7.4062596839169505e-06, |
| "loss": 0.7024, |
| "step": 2800 |
| }, |
| { |
| "epoch": 0.8573607932875668, |
| "grad_norm": 3.2100493907928467, |
| "learning_rate": 7.2513170127053e-06, |
| "loss": 0.7137, |
| "step": 2810 |
| }, |
| { |
| "epoch": 0.8604118993135011, |
| "grad_norm": 2.1263020038604736, |
| "learning_rate": 7.096374341493647e-06, |
| "loss": 0.6334, |
| "step": 2820 |
| }, |
| { |
| "epoch": 0.8634630053394355, |
| "grad_norm": 2.614156484603882, |
| "learning_rate": 6.941431670281996e-06, |
| "loss": 0.7299, |
| "step": 2830 |
| }, |
| { |
| "epoch": 0.8665141113653699, |
| "grad_norm": 4.311347961425781, |
| "learning_rate": 6.786488999070343e-06, |
| "loss": 0.6464, |
| "step": 2840 |
| }, |
| { |
| "epoch": 0.8695652173913043, |
| "grad_norm": 2.2194840908050537, |
| "learning_rate": 6.631546327858692e-06, |
| "loss": 0.7055, |
| "step": 2850 |
| }, |
| { |
| "epoch": 0.8726163234172387, |
| "grad_norm": 2.8160476684570312, |
| "learning_rate": 6.476603656647041e-06, |
| "loss": 0.7233, |
| "step": 2860 |
| }, |
| { |
| "epoch": 0.8756674294431731, |
| "grad_norm": 1.9882564544677734, |
| "learning_rate": 6.321660985435389e-06, |
| "loss": 0.7041, |
| "step": 2870 |
| }, |
| { |
| "epoch": 0.8787185354691075, |
| "grad_norm": 2.381114959716797, |
| "learning_rate": 6.166718314223738e-06, |
| "loss": 0.6916, |
| "step": 2880 |
| }, |
| { |
| "epoch": 0.881769641495042, |
| "grad_norm": 2.911763906478882, |
| "learning_rate": 6.011775643012086e-06, |
| "loss": 0.6493, |
| "step": 2890 |
| }, |
| { |
| "epoch": 0.8848207475209764, |
| "grad_norm": 2.830087900161743, |
| "learning_rate": 5.856832971800434e-06, |
| "loss": 0.6497, |
| "step": 2900 |
| }, |
| { |
| "epoch": 0.8878718535469108, |
| "grad_norm": 2.609445333480835, |
| "learning_rate": 5.701890300588782e-06, |
| "loss": 0.6923, |
| "step": 2910 |
| }, |
| { |
| "epoch": 0.8909229595728452, |
| "grad_norm": 2.1896908283233643, |
| "learning_rate": 5.5469476293771305e-06, |
| "loss": 0.6325, |
| "step": 2920 |
| }, |
| { |
| "epoch": 0.8939740655987796, |
| "grad_norm": 3.2182352542877197, |
| "learning_rate": 5.392004958165479e-06, |
| "loss": 0.6613, |
| "step": 2930 |
| }, |
| { |
| "epoch": 0.897025171624714, |
| "grad_norm": 2.7768335342407227, |
| "learning_rate": 5.237062286953828e-06, |
| "loss": 0.6944, |
| "step": 2940 |
| }, |
| { |
| "epoch": 0.9000762776506483, |
| "grad_norm": 2.458332061767578, |
| "learning_rate": 5.082119615742176e-06, |
| "loss": 0.6689, |
| "step": 2950 |
| }, |
| { |
| "epoch": 0.9031273836765827, |
| "grad_norm": 2.2689878940582275, |
| "learning_rate": 4.927176944530524e-06, |
| "loss": 0.6729, |
| "step": 2960 |
| }, |
| { |
| "epoch": 0.9061784897025171, |
| "grad_norm": 3.256464719772339, |
| "learning_rate": 4.772234273318872e-06, |
| "loss": 0.6158, |
| "step": 2970 |
| }, |
| { |
| "epoch": 0.9092295957284515, |
| "grad_norm": 3.0589096546173096, |
| "learning_rate": 4.6172916021072205e-06, |
| "loss": 0.7551, |
| "step": 2980 |
| }, |
| { |
| "epoch": 0.9122807017543859, |
| "grad_norm": 2.630323648452759, |
| "learning_rate": 4.4623489308955695e-06, |
| "loss": 0.6235, |
| "step": 2990 |
| }, |
| { |
| "epoch": 0.9153318077803204, |
| "grad_norm": 2.285128593444824, |
| "learning_rate": 4.307406259683918e-06, |
| "loss": 0.6948, |
| "step": 3000 |
| }, |
| { |
| "epoch": 0.9183829138062548, |
| "grad_norm": 2.4126858711242676, |
| "learning_rate": 4.152463588472266e-06, |
| "loss": 0.6249, |
| "step": 3010 |
| }, |
| { |
| "epoch": 0.9214340198321892, |
| "grad_norm": 2.4043056964874268, |
| "learning_rate": 3.997520917260614e-06, |
| "loss": 0.6822, |
| "step": 3020 |
| }, |
| { |
| "epoch": 0.9244851258581236, |
| "grad_norm": 2.084949254989624, |
| "learning_rate": 3.842578246048962e-06, |
| "loss": 0.7345, |
| "step": 3030 |
| }, |
| { |
| "epoch": 0.927536231884058, |
| "grad_norm": 2.251657247543335, |
| "learning_rate": 3.68763557483731e-06, |
| "loss": 0.6048, |
| "step": 3040 |
| }, |
| { |
| "epoch": 0.9305873379099924, |
| "grad_norm": 2.20843505859375, |
| "learning_rate": 3.532692903625659e-06, |
| "loss": 0.6687, |
| "step": 3050 |
| }, |
| { |
| "epoch": 0.9336384439359268, |
| "grad_norm": 2.692772150039673, |
| "learning_rate": 3.3777502324140072e-06, |
| "loss": 0.651, |
| "step": 3060 |
| }, |
| { |
| "epoch": 0.9366895499618612, |
| "grad_norm": 2.778233766555786, |
| "learning_rate": 3.2228075612023554e-06, |
| "loss": 0.7709, |
| "step": 3070 |
| }, |
| { |
| "epoch": 0.9397406559877955, |
| "grad_norm": 2.2121570110321045, |
| "learning_rate": 3.0678648899907036e-06, |
| "loss": 0.6741, |
| "step": 3080 |
| }, |
| { |
| "epoch": 0.9427917620137299, |
| "grad_norm": 2.814641237258911, |
| "learning_rate": 2.9129222187790522e-06, |
| "loss": 0.6409, |
| "step": 3090 |
| }, |
| { |
| "epoch": 0.9458428680396643, |
| "grad_norm": 2.183281660079956, |
| "learning_rate": 2.7579795475674004e-06, |
| "loss": 0.6192, |
| "step": 3100 |
| }, |
| { |
| "epoch": 0.9488939740655988, |
| "grad_norm": 1.998976230621338, |
| "learning_rate": 2.6030368763557486e-06, |
| "loss": 0.5924, |
| "step": 3110 |
| }, |
| { |
| "epoch": 0.9519450800915332, |
| "grad_norm": 2.7774434089660645, |
| "learning_rate": 2.4480942051440968e-06, |
| "loss": 0.7145, |
| "step": 3120 |
| }, |
| { |
| "epoch": 0.9549961861174676, |
| "grad_norm": 2.510991334915161, |
| "learning_rate": 2.293151533932445e-06, |
| "loss": 0.6985, |
| "step": 3130 |
| }, |
| { |
| "epoch": 0.958047292143402, |
| "grad_norm": 2.820533037185669, |
| "learning_rate": 2.138208862720793e-06, |
| "loss": 0.673, |
| "step": 3140 |
| }, |
| { |
| "epoch": 0.9610983981693364, |
| "grad_norm": 3.7463274002075195, |
| "learning_rate": 1.9832661915091418e-06, |
| "loss": 0.6753, |
| "step": 3150 |
| }, |
| { |
| "epoch": 0.9641495041952708, |
| "grad_norm": 2.6098906993865967, |
| "learning_rate": 1.82832352029749e-06, |
| "loss": 0.6077, |
| "step": 3160 |
| }, |
| { |
| "epoch": 0.9672006102212052, |
| "grad_norm": 3.7235538959503174, |
| "learning_rate": 1.6733808490858381e-06, |
| "loss": 0.6528, |
| "step": 3170 |
| }, |
| { |
| "epoch": 0.9702517162471396, |
| "grad_norm": 2.551973342895508, |
| "learning_rate": 1.5184381778741865e-06, |
| "loss": 0.7123, |
| "step": 3180 |
| }, |
| { |
| "epoch": 0.973302822273074, |
| "grad_norm": 2.60445237159729, |
| "learning_rate": 1.363495506662535e-06, |
| "loss": 0.6368, |
| "step": 3190 |
| }, |
| { |
| "epoch": 0.9763539282990084, |
| "grad_norm": 2.7576358318328857, |
| "learning_rate": 1.2085528354508833e-06, |
| "loss": 0.6172, |
| "step": 3200 |
| }, |
| { |
| "epoch": 0.9794050343249427, |
| "grad_norm": 2.23207426071167, |
| "learning_rate": 1.0536101642392315e-06, |
| "loss": 0.6227, |
| "step": 3210 |
| }, |
| { |
| "epoch": 0.9824561403508771, |
| "grad_norm": 3.0700480937957764, |
| "learning_rate": 8.986674930275798e-07, |
| "loss": 0.7168, |
| "step": 3220 |
| }, |
| { |
| "epoch": 0.9855072463768116, |
| "grad_norm": 2.6852805614471436, |
| "learning_rate": 7.437248218159281e-07, |
| "loss": 0.6322, |
| "step": 3230 |
| }, |
| { |
| "epoch": 0.988558352402746, |
| "grad_norm": 2.6967718601226807, |
| "learning_rate": 5.887821506042764e-07, |
| "loss": 0.6357, |
| "step": 3240 |
| }, |
| { |
| "epoch": 0.9916094584286804, |
| "grad_norm": 3.2857532501220703, |
| "learning_rate": 4.3383947939262475e-07, |
| "loss": 0.7225, |
| "step": 3250 |
| }, |
| { |
| "epoch": 0.9946605644546148, |
| "grad_norm": 2.565335750579834, |
| "learning_rate": 2.788968081809731e-07, |
| "loss": 0.6709, |
| "step": 3260 |
| }, |
| { |
| "epoch": 0.9977116704805492, |
| "grad_norm": 3.099731922149658, |
| "learning_rate": 1.2395413696932136e-07, |
| "loss": 0.7549, |
| "step": 3270 |
| } |
| ], |
| "logging_steps": 10, |
| "max_steps": 3277, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 1, |
| "save_steps": 500, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 1.2252825778126848e+16, |
| "train_batch_size": 2, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|