| { |
| "best_global_step": null, |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 8.0, |
| "eval_steps": 500, |
| "global_step": 4576, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.017497812773403325, |
| "grad_norm": 15.0, |
| "learning_rate": 1.0465116279069768e-06, |
| "loss": 1.1604, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.03499562554680665, |
| "grad_norm": 9.6875, |
| "learning_rate": 2.2093023255813954e-06, |
| "loss": 1.1391, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.05249343832020997, |
| "grad_norm": 4.09375, |
| "learning_rate": 3.372093023255814e-06, |
| "loss": 1.0272, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.0699912510936133, |
| "grad_norm": 2.3125, |
| "learning_rate": 4.5348837209302326e-06, |
| "loss": 0.8968, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.08748906386701662, |
| "grad_norm": 1.515625, |
| "learning_rate": 5.697674418604652e-06, |
| "loss": 0.8108, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.10498687664041995, |
| "grad_norm": 1.421875, |
| "learning_rate": 6.86046511627907e-06, |
| "loss": 0.8071, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.12248468941382328, |
| "grad_norm": 1.1953125, |
| "learning_rate": 8.023255813953488e-06, |
| "loss": 0.759, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.1399825021872266, |
| "grad_norm": 1.25, |
| "learning_rate": 9.186046511627908e-06, |
| "loss": 0.767, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.15748031496062992, |
| "grad_norm": 1.3046875, |
| "learning_rate": 1.0348837209302327e-05, |
| "loss": 0.7569, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.17497812773403323, |
| "grad_norm": 1.1171875, |
| "learning_rate": 1.1511627906976746e-05, |
| "loss": 0.7603, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.19247594050743658, |
| "grad_norm": 1.046875, |
| "learning_rate": 1.2674418604651164e-05, |
| "loss": 0.7424, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.2099737532808399, |
| "grad_norm": 1.3125, |
| "learning_rate": 1.3837209302325583e-05, |
| "loss": 0.7358, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.2274715660542432, |
| "grad_norm": 1.1015625, |
| "learning_rate": 1.5000000000000002e-05, |
| "loss": 0.7119, |
| "step": 130 |
| }, |
| { |
| "epoch": 0.24496937882764655, |
| "grad_norm": 1.140625, |
| "learning_rate": 1.616279069767442e-05, |
| "loss": 0.7219, |
| "step": 140 |
| }, |
| { |
| "epoch": 0.26246719160104987, |
| "grad_norm": 1.0859375, |
| "learning_rate": 1.7325581395348837e-05, |
| "loss": 0.7073, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.2799650043744532, |
| "grad_norm": 1.078125, |
| "learning_rate": 1.8488372093023256e-05, |
| "loss": 0.7135, |
| "step": 160 |
| }, |
| { |
| "epoch": 0.2974628171478565, |
| "grad_norm": 1.046875, |
| "learning_rate": 1.9651162790697676e-05, |
| "loss": 0.72, |
| "step": 170 |
| }, |
| { |
| "epoch": 0.31496062992125984, |
| "grad_norm": 0.96484375, |
| "learning_rate": 1.9999929041918377e-05, |
| "loss": 0.7111, |
| "step": 180 |
| }, |
| { |
| "epoch": 0.3324584426946632, |
| "grad_norm": 1.0859375, |
| "learning_rate": 1.999958149482438e-05, |
| "loss": 0.7308, |
| "step": 190 |
| }, |
| { |
| "epoch": 0.34995625546806647, |
| "grad_norm": 1.09375, |
| "learning_rate": 1.9998944336771236e-05, |
| "loss": 0.711, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.3674540682414698, |
| "grad_norm": 1.03125, |
| "learning_rate": 1.9998017588263007e-05, |
| "loss": 0.7096, |
| "step": 210 |
| }, |
| { |
| "epoch": 0.38495188101487315, |
| "grad_norm": 1.15625, |
| "learning_rate": 1.9996801279122917e-05, |
| "loss": 0.7193, |
| "step": 220 |
| }, |
| { |
| "epoch": 0.40244969378827644, |
| "grad_norm": 0.97265625, |
| "learning_rate": 1.9995295448492383e-05, |
| "loss": 0.7187, |
| "step": 230 |
| }, |
| { |
| "epoch": 0.4199475065616798, |
| "grad_norm": 1.0546875, |
| "learning_rate": 1.9993500144829784e-05, |
| "loss": 0.7197, |
| "step": 240 |
| }, |
| { |
| "epoch": 0.4374453193350831, |
| "grad_norm": 1.046875, |
| "learning_rate": 1.9991415425908868e-05, |
| "loss": 0.7117, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.4549431321084864, |
| "grad_norm": 1.1328125, |
| "learning_rate": 1.9989041358816926e-05, |
| "loss": 0.6958, |
| "step": 260 |
| }, |
| { |
| "epoch": 0.47244094488188976, |
| "grad_norm": 1.0625, |
| "learning_rate": 1.9986378019952595e-05, |
| "loss": 0.7127, |
| "step": 270 |
| }, |
| { |
| "epoch": 0.4899387576552931, |
| "grad_norm": 1.1640625, |
| "learning_rate": 1.998342549502343e-05, |
| "loss": 0.7068, |
| "step": 280 |
| }, |
| { |
| "epoch": 0.5074365704286964, |
| "grad_norm": 1.0546875, |
| "learning_rate": 1.998018387904314e-05, |
| "loss": 0.6994, |
| "step": 290 |
| }, |
| { |
| "epoch": 0.5249343832020997, |
| "grad_norm": 1.078125, |
| "learning_rate": 1.997665327632852e-05, |
| "loss": 0.713, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.5424321959755031, |
| "grad_norm": 1.2109375, |
| "learning_rate": 1.9972833800496107e-05, |
| "loss": 0.7119, |
| "step": 310 |
| }, |
| { |
| "epoch": 0.5599300087489064, |
| "grad_norm": 1.0703125, |
| "learning_rate": 1.99687255744585e-05, |
| "loss": 0.6993, |
| "step": 320 |
| }, |
| { |
| "epoch": 0.5774278215223098, |
| "grad_norm": 0.97265625, |
| "learning_rate": 1.9964328730420445e-05, |
| "loss": 0.706, |
| "step": 330 |
| }, |
| { |
| "epoch": 0.594925634295713, |
| "grad_norm": 1.1484375, |
| "learning_rate": 1.995964340987454e-05, |
| "loss": 0.7167, |
| "step": 340 |
| }, |
| { |
| "epoch": 0.6124234470691163, |
| "grad_norm": 1.109375, |
| "learning_rate": 1.9954669763596708e-05, |
| "loss": 0.7206, |
| "step": 350 |
| }, |
| { |
| "epoch": 0.6299212598425197, |
| "grad_norm": 0.97265625, |
| "learning_rate": 1.9949407951641325e-05, |
| "loss": 0.6875, |
| "step": 360 |
| }, |
| { |
| "epoch": 0.647419072615923, |
| "grad_norm": 1.0625, |
| "learning_rate": 1.99438581433361e-05, |
| "loss": 0.7256, |
| "step": 370 |
| }, |
| { |
| "epoch": 0.6649168853893264, |
| "grad_norm": 1.1015625, |
| "learning_rate": 1.9938020517276583e-05, |
| "loss": 0.689, |
| "step": 380 |
| }, |
| { |
| "epoch": 0.6824146981627297, |
| "grad_norm": 0.98828125, |
| "learning_rate": 1.9931895261320463e-05, |
| "loss": 0.7029, |
| "step": 390 |
| }, |
| { |
| "epoch": 0.6999125109361329, |
| "grad_norm": 1.0078125, |
| "learning_rate": 1.9925482572581477e-05, |
| "loss": 0.7189, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.7174103237095363, |
| "grad_norm": 1.015625, |
| "learning_rate": 1.991878265742311e-05, |
| "loss": 0.7103, |
| "step": 410 |
| }, |
| { |
| "epoch": 0.7349081364829396, |
| "grad_norm": 0.98046875, |
| "learning_rate": 1.9911795731451928e-05, |
| "loss": 0.7045, |
| "step": 420 |
| }, |
| { |
| "epoch": 0.752405949256343, |
| "grad_norm": 1.1171875, |
| "learning_rate": 1.9904522019510647e-05, |
| "loss": 0.7112, |
| "step": 430 |
| }, |
| { |
| "epoch": 0.7699037620297463, |
| "grad_norm": 1.03125, |
| "learning_rate": 1.989696175567089e-05, |
| "loss": 0.711, |
| "step": 440 |
| }, |
| { |
| "epoch": 0.7874015748031497, |
| "grad_norm": 1.0546875, |
| "learning_rate": 1.988911518322566e-05, |
| "loss": 0.7174, |
| "step": 450 |
| }, |
| { |
| "epoch": 0.8048993875765529, |
| "grad_norm": 1.078125, |
| "learning_rate": 1.9880982554681534e-05, |
| "loss": 0.7023, |
| "step": 460 |
| }, |
| { |
| "epoch": 0.8223972003499562, |
| "grad_norm": 1.09375, |
| "learning_rate": 1.9872564131750488e-05, |
| "loss": 0.6956, |
| "step": 470 |
| }, |
| { |
| "epoch": 0.8398950131233596, |
| "grad_norm": 1.0859375, |
| "learning_rate": 1.9863860185341514e-05, |
| "loss": 0.7003, |
| "step": 480 |
| }, |
| { |
| "epoch": 0.8573928258967629, |
| "grad_norm": 1.03125, |
| "learning_rate": 1.985487099555189e-05, |
| "loss": 0.6849, |
| "step": 490 |
| }, |
| { |
| "epoch": 0.8748906386701663, |
| "grad_norm": 0.9140625, |
| "learning_rate": 1.9845596851658168e-05, |
| "loss": 0.7017, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.8923884514435696, |
| "grad_norm": 0.9609375, |
| "learning_rate": 1.9836038052106853e-05, |
| "loss": 0.6809, |
| "step": 510 |
| }, |
| { |
| "epoch": 0.9098862642169728, |
| "grad_norm": 0.96484375, |
| "learning_rate": 1.9826194904504824e-05, |
| "loss": 0.6961, |
| "step": 520 |
| }, |
| { |
| "epoch": 0.9273840769903762, |
| "grad_norm": 0.97265625, |
| "learning_rate": 1.9816067725609403e-05, |
| "loss": 0.698, |
| "step": 530 |
| }, |
| { |
| "epoch": 0.9448818897637795, |
| "grad_norm": 1.125, |
| "learning_rate": 1.9805656841318198e-05, |
| "loss": 0.7022, |
| "step": 540 |
| }, |
| { |
| "epoch": 0.9623797025371829, |
| "grad_norm": 0.97265625, |
| "learning_rate": 1.9794962586658582e-05, |
| "loss": 0.6896, |
| "step": 550 |
| }, |
| { |
| "epoch": 0.9798775153105862, |
| "grad_norm": 1.0390625, |
| "learning_rate": 1.978398530577693e-05, |
| "loss": 0.6915, |
| "step": 560 |
| }, |
| { |
| "epoch": 0.9973753280839895, |
| "grad_norm": 0.99609375, |
| "learning_rate": 1.977272535192754e-05, |
| "loss": 0.6996, |
| "step": 570 |
| }, |
| { |
| "epoch": 1.0139982502187226, |
| "grad_norm": 1.0390625, |
| "learning_rate": 1.9761183087461264e-05, |
| "loss": 0.6237, |
| "step": 580 |
| }, |
| { |
| "epoch": 1.031496062992126, |
| "grad_norm": 1.1328125, |
| "learning_rate": 1.974935888381385e-05, |
| "loss": 0.5954, |
| "step": 590 |
| }, |
| { |
| "epoch": 1.0489938757655293, |
| "grad_norm": 1.1484375, |
| "learning_rate": 1.9737253121493994e-05, |
| "loss": 0.6124, |
| "step": 600 |
| }, |
| { |
| "epoch": 1.0664916885389326, |
| "grad_norm": 1.1640625, |
| "learning_rate": 1.9724866190071075e-05, |
| "loss": 0.6023, |
| "step": 610 |
| }, |
| { |
| "epoch": 1.083989501312336, |
| "grad_norm": 1.1875, |
| "learning_rate": 1.971219848816264e-05, |
| "loss": 0.6128, |
| "step": 620 |
| }, |
| { |
| "epoch": 1.1014873140857393, |
| "grad_norm": 1.2109375, |
| "learning_rate": 1.9699250423421568e-05, |
| "loss": 0.6167, |
| "step": 630 |
| }, |
| { |
| "epoch": 1.1189851268591426, |
| "grad_norm": 1.203125, |
| "learning_rate": 1.9686022412522953e-05, |
| "loss": 0.6322, |
| "step": 640 |
| }, |
| { |
| "epoch": 1.136482939632546, |
| "grad_norm": 0.94921875, |
| "learning_rate": 1.967251488115069e-05, |
| "loss": 0.6, |
| "step": 650 |
| }, |
| { |
| "epoch": 1.1539807524059493, |
| "grad_norm": 1.0546875, |
| "learning_rate": 1.9658728263983783e-05, |
| "loss": 0.6204, |
| "step": 660 |
| }, |
| { |
| "epoch": 1.1714785651793527, |
| "grad_norm": 1.140625, |
| "learning_rate": 1.964466300468235e-05, |
| "loss": 0.627, |
| "step": 670 |
| }, |
| { |
| "epoch": 1.188976377952756, |
| "grad_norm": 1.0, |
| "learning_rate": 1.9630319555873364e-05, |
| "loss": 0.5988, |
| "step": 680 |
| }, |
| { |
| "epoch": 1.2064741907261591, |
| "grad_norm": 1.1015625, |
| "learning_rate": 1.961569837913605e-05, |
| "loss": 0.6141, |
| "step": 690 |
| }, |
| { |
| "epoch": 1.2239720034995625, |
| "grad_norm": 1.03125, |
| "learning_rate": 1.9600799944987077e-05, |
| "loss": 0.6171, |
| "step": 700 |
| }, |
| { |
| "epoch": 1.2414698162729658, |
| "grad_norm": 1.1328125, |
| "learning_rate": 1.958562473286538e-05, |
| "loss": 0.613, |
| "step": 710 |
| }, |
| { |
| "epoch": 1.2589676290463692, |
| "grad_norm": 1.109375, |
| "learning_rate": 1.957017323111675e-05, |
| "loss": 0.6169, |
| "step": 720 |
| }, |
| { |
| "epoch": 1.2764654418197725, |
| "grad_norm": 1.1015625, |
| "learning_rate": 1.955444593697811e-05, |
| "loss": 0.6312, |
| "step": 730 |
| }, |
| { |
| "epoch": 1.2939632545931758, |
| "grad_norm": 1.2109375, |
| "learning_rate": 1.9538443356561528e-05, |
| "loss": 0.6144, |
| "step": 740 |
| }, |
| { |
| "epoch": 1.3114610673665792, |
| "grad_norm": 1.140625, |
| "learning_rate": 1.9522166004837908e-05, |
| "loss": 0.6096, |
| "step": 750 |
| }, |
| { |
| "epoch": 1.3289588801399825, |
| "grad_norm": 1.1953125, |
| "learning_rate": 1.9505614405620436e-05, |
| "loss": 0.6225, |
| "step": 760 |
| }, |
| { |
| "epoch": 1.3464566929133859, |
| "grad_norm": 1.1953125, |
| "learning_rate": 1.9488789091547716e-05, |
| "loss": 0.6156, |
| "step": 770 |
| }, |
| { |
| "epoch": 1.3639545056867892, |
| "grad_norm": 1.078125, |
| "learning_rate": 1.9471690604066633e-05, |
| "loss": 0.644, |
| "step": 780 |
| }, |
| { |
| "epoch": 1.3814523184601923, |
| "grad_norm": 1.140625, |
| "learning_rate": 1.945431949341492e-05, |
| "loss": 0.6221, |
| "step": 790 |
| }, |
| { |
| "epoch": 1.3989501312335957, |
| "grad_norm": 1.046875, |
| "learning_rate": 1.9436676318603465e-05, |
| "loss": 0.6164, |
| "step": 800 |
| }, |
| { |
| "epoch": 1.416447944006999, |
| "grad_norm": 1.0546875, |
| "learning_rate": 1.941876164739831e-05, |
| "loss": 0.6187, |
| "step": 810 |
| }, |
| { |
| "epoch": 1.4339457567804024, |
| "grad_norm": 1.1953125, |
| "learning_rate": 1.940057605630239e-05, |
| "loss": 0.6041, |
| "step": 820 |
| }, |
| { |
| "epoch": 1.4514435695538057, |
| "grad_norm": 1.0078125, |
| "learning_rate": 1.938212013053697e-05, |
| "loss": 0.6172, |
| "step": 830 |
| }, |
| { |
| "epoch": 1.468941382327209, |
| "grad_norm": 1.125, |
| "learning_rate": 1.9363394464022814e-05, |
| "loss": 0.6107, |
| "step": 840 |
| }, |
| { |
| "epoch": 1.4864391951006124, |
| "grad_norm": 1.1484375, |
| "learning_rate": 1.934439965936109e-05, |
| "loss": 0.6109, |
| "step": 850 |
| }, |
| { |
| "epoch": 1.5039370078740157, |
| "grad_norm": 1.15625, |
| "learning_rate": 1.9325136327813945e-05, |
| "loss": 0.606, |
| "step": 860 |
| }, |
| { |
| "epoch": 1.521434820647419, |
| "grad_norm": 1.1171875, |
| "learning_rate": 1.9305605089284873e-05, |
| "loss": 0.6264, |
| "step": 870 |
| }, |
| { |
| "epoch": 1.5389326334208224, |
| "grad_norm": 1.140625, |
| "learning_rate": 1.9285806572298726e-05, |
| "loss": 0.6185, |
| "step": 880 |
| }, |
| { |
| "epoch": 1.5564304461942258, |
| "grad_norm": 1.1953125, |
| "learning_rate": 1.926574141398153e-05, |
| "loss": 0.6243, |
| "step": 890 |
| }, |
| { |
| "epoch": 1.5739282589676291, |
| "grad_norm": 1.0703125, |
| "learning_rate": 1.924541026003994e-05, |
| "loss": 0.6035, |
| "step": 900 |
| }, |
| { |
| "epoch": 1.5914260717410325, |
| "grad_norm": 1.203125, |
| "learning_rate": 1.9224813764740496e-05, |
| "loss": 0.6297, |
| "step": 910 |
| }, |
| { |
| "epoch": 1.6089238845144358, |
| "grad_norm": 1.078125, |
| "learning_rate": 1.9203952590888547e-05, |
| "loss": 0.6138, |
| "step": 920 |
| }, |
| { |
| "epoch": 1.6264216972878391, |
| "grad_norm": 1.1640625, |
| "learning_rate": 1.918282740980693e-05, |
| "loss": 0.6211, |
| "step": 930 |
| }, |
| { |
| "epoch": 1.6439195100612425, |
| "grad_norm": 1.140625, |
| "learning_rate": 1.9161438901314352e-05, |
| "loss": 0.6082, |
| "step": 940 |
| }, |
| { |
| "epoch": 1.6614173228346458, |
| "grad_norm": 1.09375, |
| "learning_rate": 1.9139787753703554e-05, |
| "loss": 0.6023, |
| "step": 950 |
| }, |
| { |
| "epoch": 1.678915135608049, |
| "grad_norm": 1.0859375, |
| "learning_rate": 1.91178746637191e-05, |
| "loss": 0.6152, |
| "step": 960 |
| }, |
| { |
| "epoch": 1.6964129483814523, |
| "grad_norm": 1.296875, |
| "learning_rate": 1.9095700336535017e-05, |
| "loss": 0.6076, |
| "step": 970 |
| }, |
| { |
| "epoch": 1.7139107611548556, |
| "grad_norm": 1.078125, |
| "learning_rate": 1.907326548573205e-05, |
| "loss": 0.6257, |
| "step": 980 |
| }, |
| { |
| "epoch": 1.731408573928259, |
| "grad_norm": 1.21875, |
| "learning_rate": 1.905057083327474e-05, |
| "loss": 0.616, |
| "step": 990 |
| }, |
| { |
| "epoch": 1.7489063867016623, |
| "grad_norm": 1.3203125, |
| "learning_rate": 1.9027617109488164e-05, |
| "loss": 0.6239, |
| "step": 1000 |
| }, |
| { |
| "epoch": 1.7664041994750657, |
| "grad_norm": 1.125, |
| "learning_rate": 1.9004405053034445e-05, |
| "loss": 0.5979, |
| "step": 1010 |
| }, |
| { |
| "epoch": 1.7839020122484688, |
| "grad_norm": 1.03125, |
| "learning_rate": 1.8980935410888973e-05, |
| "loss": 0.6177, |
| "step": 1020 |
| }, |
| { |
| "epoch": 1.8013998250218721, |
| "grad_norm": 1.0859375, |
| "learning_rate": 1.895720893831638e-05, |
| "loss": 0.5923, |
| "step": 1030 |
| }, |
| { |
| "epoch": 1.8188976377952755, |
| "grad_norm": 1.078125, |
| "learning_rate": 1.893322639884622e-05, |
| "loss": 0.6029, |
| "step": 1040 |
| }, |
| { |
| "epoch": 1.8363954505686788, |
| "grad_norm": 1.1328125, |
| "learning_rate": 1.8908988564248415e-05, |
| "loss": 0.6273, |
| "step": 1050 |
| }, |
| { |
| "epoch": 1.8538932633420822, |
| "grad_norm": 1.1953125, |
| "learning_rate": 1.8884496214508398e-05, |
| "loss": 0.6358, |
| "step": 1060 |
| }, |
| { |
| "epoch": 1.8713910761154855, |
| "grad_norm": 1.09375, |
| "learning_rate": 1.8859750137802037e-05, |
| "loss": 0.6321, |
| "step": 1070 |
| }, |
| { |
| "epoch": 1.8888888888888888, |
| "grad_norm": 1.1953125, |
| "learning_rate": 1.8834751130470247e-05, |
| "loss": 0.6118, |
| "step": 1080 |
| }, |
| { |
| "epoch": 1.9063867016622922, |
| "grad_norm": 1.140625, |
| "learning_rate": 1.8809499996993385e-05, |
| "loss": 0.6095, |
| "step": 1090 |
| }, |
| { |
| "epoch": 1.9238845144356955, |
| "grad_norm": 1.09375, |
| "learning_rate": 1.8783997549965344e-05, |
| "loss": 0.622, |
| "step": 1100 |
| }, |
| { |
| "epoch": 1.9413823272090989, |
| "grad_norm": 1.2421875, |
| "learning_rate": 1.875824461006741e-05, |
| "loss": 0.636, |
| "step": 1110 |
| }, |
| { |
| "epoch": 1.9588801399825022, |
| "grad_norm": 1.125, |
| "learning_rate": 1.8732242006041864e-05, |
| "loss": 0.6089, |
| "step": 1120 |
| }, |
| { |
| "epoch": 1.9763779527559056, |
| "grad_norm": 1.1328125, |
| "learning_rate": 1.8705990574665288e-05, |
| "loss": 0.6169, |
| "step": 1130 |
| }, |
| { |
| "epoch": 1.993875765529309, |
| "grad_norm": 1.1640625, |
| "learning_rate": 1.8679491160721654e-05, |
| "loss": 0.6174, |
| "step": 1140 |
| }, |
| { |
| "epoch": 2.010498687664042, |
| "grad_norm": 1.640625, |
| "learning_rate": 1.8652744616975142e-05, |
| "loss": 0.566, |
| "step": 1150 |
| }, |
| { |
| "epoch": 2.027996500437445, |
| "grad_norm": 1.6328125, |
| "learning_rate": 1.8625751804142685e-05, |
| "loss": 0.5181, |
| "step": 1160 |
| }, |
| { |
| "epoch": 2.0454943132108485, |
| "grad_norm": 1.2890625, |
| "learning_rate": 1.8598513590866278e-05, |
| "loss": 0.5257, |
| "step": 1170 |
| }, |
| { |
| "epoch": 2.062992125984252, |
| "grad_norm": 1.2890625, |
| "learning_rate": 1.8571030853685016e-05, |
| "loss": 0.5193, |
| "step": 1180 |
| }, |
| { |
| "epoch": 2.080489938757655, |
| "grad_norm": 1.3125, |
| "learning_rate": 1.8543304477006906e-05, |
| "loss": 0.5252, |
| "step": 1190 |
| }, |
| { |
| "epoch": 2.0979877515310585, |
| "grad_norm": 1.390625, |
| "learning_rate": 1.851533535308038e-05, |
| "loss": 0.5194, |
| "step": 1200 |
| }, |
| { |
| "epoch": 2.115485564304462, |
| "grad_norm": 1.1875, |
| "learning_rate": 1.848712438196561e-05, |
| "loss": 0.5126, |
| "step": 1210 |
| }, |
| { |
| "epoch": 2.1329833770778652, |
| "grad_norm": 1.2890625, |
| "learning_rate": 1.8458672471505522e-05, |
| "loss": 0.5007, |
| "step": 1220 |
| }, |
| { |
| "epoch": 2.1504811898512686, |
| "grad_norm": 1.2421875, |
| "learning_rate": 1.8429980537296587e-05, |
| "loss": 0.5205, |
| "step": 1230 |
| }, |
| { |
| "epoch": 2.167979002624672, |
| "grad_norm": 1.25, |
| "learning_rate": 1.840104950265936e-05, |
| "loss": 0.5018, |
| "step": 1240 |
| }, |
| { |
| "epoch": 2.1854768153980753, |
| "grad_norm": 1.359375, |
| "learning_rate": 1.8371880298608768e-05, |
| "loss": 0.5216, |
| "step": 1250 |
| }, |
| { |
| "epoch": 2.2029746281714786, |
| "grad_norm": 1.4921875, |
| "learning_rate": 1.834247386382414e-05, |
| "loss": 0.5109, |
| "step": 1260 |
| }, |
| { |
| "epoch": 2.220472440944882, |
| "grad_norm": 1.359375, |
| "learning_rate": 1.8312831144619024e-05, |
| "loss": 0.512, |
| "step": 1270 |
| }, |
| { |
| "epoch": 2.2379702537182853, |
| "grad_norm": 1.390625, |
| "learning_rate": 1.828295309491069e-05, |
| "loss": 0.5139, |
| "step": 1280 |
| }, |
| { |
| "epoch": 2.2554680664916886, |
| "grad_norm": 1.3203125, |
| "learning_rate": 1.8252840676189484e-05, |
| "loss": 0.5277, |
| "step": 1290 |
| }, |
| { |
| "epoch": 2.272965879265092, |
| "grad_norm": 1.4375, |
| "learning_rate": 1.8222494857487845e-05, |
| "loss": 0.5175, |
| "step": 1300 |
| }, |
| { |
| "epoch": 2.2904636920384953, |
| "grad_norm": 1.4609375, |
| "learning_rate": 1.819191661534915e-05, |
| "loss": 0.5247, |
| "step": 1310 |
| }, |
| { |
| "epoch": 2.3079615048118987, |
| "grad_norm": 1.3046875, |
| "learning_rate": 1.8161106933796268e-05, |
| "loss": 0.5173, |
| "step": 1320 |
| }, |
| { |
| "epoch": 2.325459317585302, |
| "grad_norm": 2.0, |
| "learning_rate": 1.813006680429991e-05, |
| "loss": 0.5262, |
| "step": 1330 |
| }, |
| { |
| "epoch": 2.3429571303587053, |
| "grad_norm": 1.328125, |
| "learning_rate": 1.8098797225746703e-05, |
| "loss": 0.5167, |
| "step": 1340 |
| }, |
| { |
| "epoch": 2.3604549431321082, |
| "grad_norm": 1.2890625, |
| "learning_rate": 1.8067299204407075e-05, |
| "loss": 0.5254, |
| "step": 1350 |
| }, |
| { |
| "epoch": 2.377952755905512, |
| "grad_norm": 1.40625, |
| "learning_rate": 1.803557375390284e-05, |
| "loss": 0.5234, |
| "step": 1360 |
| }, |
| { |
| "epoch": 2.395450568678915, |
| "grad_norm": 1.4375, |
| "learning_rate": 1.8003621895174603e-05, |
| "loss": 0.5113, |
| "step": 1370 |
| }, |
| { |
| "epoch": 2.4129483814523183, |
| "grad_norm": 1.296875, |
| "learning_rate": 1.797144465644889e-05, |
| "loss": 0.5177, |
| "step": 1380 |
| }, |
| { |
| "epoch": 2.4304461942257216, |
| "grad_norm": 1.34375, |
| "learning_rate": 1.7939043073205076e-05, |
| "loss": 0.5289, |
| "step": 1390 |
| }, |
| { |
| "epoch": 2.447944006999125, |
| "grad_norm": 1.3671875, |
| "learning_rate": 1.7906418188142045e-05, |
| "loss": 0.524, |
| "step": 1400 |
| }, |
| { |
| "epoch": 2.4654418197725283, |
| "grad_norm": 1.34375, |
| "learning_rate": 1.7873571051144644e-05, |
| "loss": 0.5193, |
| "step": 1410 |
| }, |
| { |
| "epoch": 2.4829396325459316, |
| "grad_norm": 1.421875, |
| "learning_rate": 1.7840502719249898e-05, |
| "loss": 0.5117, |
| "step": 1420 |
| }, |
| { |
| "epoch": 2.500437445319335, |
| "grad_norm": 1.375, |
| "learning_rate": 1.7807214256612995e-05, |
| "loss": 0.5436, |
| "step": 1430 |
| }, |
| { |
| "epoch": 2.5179352580927383, |
| "grad_norm": 1.3203125, |
| "learning_rate": 1.777370673447303e-05, |
| "loss": 0.5229, |
| "step": 1440 |
| }, |
| { |
| "epoch": 2.5354330708661417, |
| "grad_norm": 1.375, |
| "learning_rate": 1.7739981231118558e-05, |
| "loss": 0.5243, |
| "step": 1450 |
| }, |
| { |
| "epoch": 2.552930883639545, |
| "grad_norm": 1.3046875, |
| "learning_rate": 1.770603883185286e-05, |
| "loss": 0.5244, |
| "step": 1460 |
| }, |
| { |
| "epoch": 2.5704286964129484, |
| "grad_norm": 1.375, |
| "learning_rate": 1.7671880628959034e-05, |
| "loss": 0.5284, |
| "step": 1470 |
| }, |
| { |
| "epoch": 2.5879265091863517, |
| "grad_norm": 1.234375, |
| "learning_rate": 1.763750772166486e-05, |
| "loss": 0.5243, |
| "step": 1480 |
| }, |
| { |
| "epoch": 2.605424321959755, |
| "grad_norm": 1.4375, |
| "learning_rate": 1.7602921216107397e-05, |
| "loss": 0.5299, |
| "step": 1490 |
| }, |
| { |
| "epoch": 2.6229221347331584, |
| "grad_norm": 1.3046875, |
| "learning_rate": 1.7568122225297407e-05, |
| "loss": 0.5129, |
| "step": 1500 |
| }, |
| { |
| "epoch": 2.6404199475065617, |
| "grad_norm": 1.40625, |
| "learning_rate": 1.753311186908353e-05, |
| "loss": 0.5222, |
| "step": 1510 |
| }, |
| { |
| "epoch": 2.657917760279965, |
| "grad_norm": 1.3515625, |
| "learning_rate": 1.7497891274116246e-05, |
| "loss": 0.5314, |
| "step": 1520 |
| }, |
| { |
| "epoch": 2.6754155730533684, |
| "grad_norm": 1.3203125, |
| "learning_rate": 1.7462461573811632e-05, |
| "loss": 0.5145, |
| "step": 1530 |
| }, |
| { |
| "epoch": 2.6929133858267718, |
| "grad_norm": 1.3203125, |
| "learning_rate": 1.742682390831487e-05, |
| "loss": 0.5252, |
| "step": 1540 |
| }, |
| { |
| "epoch": 2.710411198600175, |
| "grad_norm": 1.515625, |
| "learning_rate": 1.739097942446356e-05, |
| "loss": 0.5354, |
| "step": 1550 |
| }, |
| { |
| "epoch": 2.7279090113735784, |
| "grad_norm": 1.328125, |
| "learning_rate": 1.7354929275750832e-05, |
| "loss": 0.5128, |
| "step": 1560 |
| }, |
| { |
| "epoch": 2.745406824146982, |
| "grad_norm": 1.2890625, |
| "learning_rate": 1.731867462228819e-05, |
| "loss": 0.5337, |
| "step": 1570 |
| }, |
| { |
| "epoch": 2.7629046369203847, |
| "grad_norm": 1.2890625, |
| "learning_rate": 1.7282216630768223e-05, |
| "loss": 0.51, |
| "step": 1580 |
| }, |
| { |
| "epoch": 2.7804024496937885, |
| "grad_norm": 1.296875, |
| "learning_rate": 1.7245556474427036e-05, |
| "loss": 0.5161, |
| "step": 1590 |
| }, |
| { |
| "epoch": 2.7979002624671914, |
| "grad_norm": 1.4765625, |
| "learning_rate": 1.7208695333006483e-05, |
| "loss": 0.5197, |
| "step": 1600 |
| }, |
| { |
| "epoch": 2.815398075240595, |
| "grad_norm": 1.3046875, |
| "learning_rate": 1.717163439271623e-05, |
| "loss": 0.5267, |
| "step": 1610 |
| }, |
| { |
| "epoch": 2.832895888013998, |
| "grad_norm": 1.3828125, |
| "learning_rate": 1.7134374846195566e-05, |
| "loss": 0.5144, |
| "step": 1620 |
| }, |
| { |
| "epoch": 2.850393700787402, |
| "grad_norm": 1.3046875, |
| "learning_rate": 1.7096917892475024e-05, |
| "loss": 0.5255, |
| "step": 1630 |
| }, |
| { |
| "epoch": 2.8678915135608047, |
| "grad_norm": 1.546875, |
| "learning_rate": 1.7059264736937796e-05, |
| "loss": 0.5197, |
| "step": 1640 |
| }, |
| { |
| "epoch": 2.885389326334208, |
| "grad_norm": 1.3828125, |
| "learning_rate": 1.702141659128095e-05, |
| "loss": 0.5165, |
| "step": 1650 |
| }, |
| { |
| "epoch": 2.9028871391076114, |
| "grad_norm": 1.296875, |
| "learning_rate": 1.6983374673476428e-05, |
| "loss": 0.5202, |
| "step": 1660 |
| }, |
| { |
| "epoch": 2.9203849518810148, |
| "grad_norm": 1.328125, |
| "learning_rate": 1.694514020773186e-05, |
| "loss": 0.5227, |
| "step": 1670 |
| }, |
| { |
| "epoch": 2.937882764654418, |
| "grad_norm": 1.328125, |
| "learning_rate": 1.6906714424451154e-05, |
| "loss": 0.5223, |
| "step": 1680 |
| }, |
| { |
| "epoch": 2.9553805774278215, |
| "grad_norm": 1.3359375, |
| "learning_rate": 1.6868098560194926e-05, |
| "loss": 0.528, |
| "step": 1690 |
| }, |
| { |
| "epoch": 2.972878390201225, |
| "grad_norm": 1.3046875, |
| "learning_rate": 1.6829293857640675e-05, |
| "loss": 0.518, |
| "step": 1700 |
| }, |
| { |
| "epoch": 2.990376202974628, |
| "grad_norm": 1.3671875, |
| "learning_rate": 1.679030156554282e-05, |
| "loss": 0.5345, |
| "step": 1710 |
| }, |
| { |
| "epoch": 3.0069991251093615, |
| "grad_norm": 1.6015625, |
| "learning_rate": 1.6751122938692493e-05, |
| "loss": 0.4821, |
| "step": 1720 |
| }, |
| { |
| "epoch": 3.024496937882765, |
| "grad_norm": 1.6796875, |
| "learning_rate": 1.6711759237877188e-05, |
| "loss": 0.4391, |
| "step": 1730 |
| }, |
| { |
| "epoch": 3.041994750656168, |
| "grad_norm": 1.890625, |
| "learning_rate": 1.667221172984015e-05, |
| "loss": 0.4258, |
| "step": 1740 |
| }, |
| { |
| "epoch": 3.059492563429571, |
| "grad_norm": 1.6796875, |
| "learning_rate": 1.663248168723964e-05, |
| "loss": 0.4333, |
| "step": 1750 |
| }, |
| { |
| "epoch": 3.0769903762029744, |
| "grad_norm": 1.7890625, |
| "learning_rate": 1.6592570388607978e-05, |
| "loss": 0.4318, |
| "step": 1760 |
| }, |
| { |
| "epoch": 3.094488188976378, |
| "grad_norm": 1.703125, |
| "learning_rate": 1.6552479118310378e-05, |
| "loss": 0.4146, |
| "step": 1770 |
| }, |
| { |
| "epoch": 3.111986001749781, |
| "grad_norm": 1.5078125, |
| "learning_rate": 1.651220916650364e-05, |
| "loss": 0.4209, |
| "step": 1780 |
| }, |
| { |
| "epoch": 3.1294838145231845, |
| "grad_norm": 1.6171875, |
| "learning_rate": 1.6471761829094623e-05, |
| "loss": 0.4331, |
| "step": 1790 |
| }, |
| { |
| "epoch": 3.146981627296588, |
| "grad_norm": 1.578125, |
| "learning_rate": 1.643113840769854e-05, |
| "loss": 0.428, |
| "step": 1800 |
| }, |
| { |
| "epoch": 3.164479440069991, |
| "grad_norm": 1.9609375, |
| "learning_rate": 1.6390340209597074e-05, |
| "loss": 0.4195, |
| "step": 1810 |
| }, |
| { |
| "epoch": 3.1819772528433945, |
| "grad_norm": 1.7109375, |
| "learning_rate": 1.6349368547696297e-05, |
| "loss": 0.4232, |
| "step": 1820 |
| }, |
| { |
| "epoch": 3.199475065616798, |
| "grad_norm": 1.6171875, |
| "learning_rate": 1.6308224740484456e-05, |
| "loss": 0.4258, |
| "step": 1830 |
| }, |
| { |
| "epoch": 3.216972878390201, |
| "grad_norm": 1.640625, |
| "learning_rate": 1.6266910111989498e-05, |
| "loss": 0.424, |
| "step": 1840 |
| }, |
| { |
| "epoch": 3.2344706911636045, |
| "grad_norm": 1.5703125, |
| "learning_rate": 1.622542599173649e-05, |
| "loss": 0.4326, |
| "step": 1850 |
| }, |
| { |
| "epoch": 3.251968503937008, |
| "grad_norm": 1.4921875, |
| "learning_rate": 1.6183773714704824e-05, |
| "loss": 0.436, |
| "step": 1860 |
| }, |
| { |
| "epoch": 3.269466316710411, |
| "grad_norm": 1.671875, |
| "learning_rate": 1.6141954621285267e-05, |
| "loss": 0.4423, |
| "step": 1870 |
| }, |
| { |
| "epoch": 3.2869641294838146, |
| "grad_norm": 1.71875, |
| "learning_rate": 1.6099970057236808e-05, |
| "loss": 0.4348, |
| "step": 1880 |
| }, |
| { |
| "epoch": 3.304461942257218, |
| "grad_norm": 1.8125, |
| "learning_rate": 1.605782137364338e-05, |
| "loss": 0.4306, |
| "step": 1890 |
| }, |
| { |
| "epoch": 3.3219597550306212, |
| "grad_norm": 1.609375, |
| "learning_rate": 1.6015509926870343e-05, |
| "loss": 0.4321, |
| "step": 1900 |
| }, |
| { |
| "epoch": 3.3394575678040246, |
| "grad_norm": 1.8203125, |
| "learning_rate": 1.597303707852087e-05, |
| "loss": 0.446, |
| "step": 1910 |
| }, |
| { |
| "epoch": 3.356955380577428, |
| "grad_norm": 1.6484375, |
| "learning_rate": 1.5930404195392114e-05, |
| "loss": 0.4246, |
| "step": 1920 |
| }, |
| { |
| "epoch": 3.3744531933508313, |
| "grad_norm": 1.7734375, |
| "learning_rate": 1.588761264943122e-05, |
| "loss": 0.4184, |
| "step": 1930 |
| }, |
| { |
| "epoch": 3.3919510061242346, |
| "grad_norm": 1.8828125, |
| "learning_rate": 1.5844663817691187e-05, |
| "loss": 0.4267, |
| "step": 1940 |
| }, |
| { |
| "epoch": 3.409448818897638, |
| "grad_norm": 2.0, |
| "learning_rate": 1.5801559082286546e-05, |
| "loss": 0.4358, |
| "step": 1950 |
| }, |
| { |
| "epoch": 3.4269466316710413, |
| "grad_norm": 1.671875, |
| "learning_rate": 1.5758299830348884e-05, |
| "loss": 0.4438, |
| "step": 1960 |
| }, |
| { |
| "epoch": 3.4444444444444446, |
| "grad_norm": 1.578125, |
| "learning_rate": 1.5714887453982204e-05, |
| "loss": 0.4413, |
| "step": 1970 |
| }, |
| { |
| "epoch": 3.4619422572178475, |
| "grad_norm": 1.7109375, |
| "learning_rate": 1.5671323350218135e-05, |
| "loss": 0.4425, |
| "step": 1980 |
| }, |
| { |
| "epoch": 3.4794400699912513, |
| "grad_norm": 1.640625, |
| "learning_rate": 1.562760892097096e-05, |
| "loss": 0.4326, |
| "step": 1990 |
| }, |
| { |
| "epoch": 3.4969378827646542, |
| "grad_norm": 1.59375, |
| "learning_rate": 1.5583745572992518e-05, |
| "loss": 0.4335, |
| "step": 2000 |
| }, |
| { |
| "epoch": 3.514435695538058, |
| "grad_norm": 1.734375, |
| "learning_rate": 1.553973471782692e-05, |
| "loss": 0.4405, |
| "step": 2010 |
| }, |
| { |
| "epoch": 3.531933508311461, |
| "grad_norm": 1.7265625, |
| "learning_rate": 1.5495577771765134e-05, |
| "loss": 0.4259, |
| "step": 2020 |
| }, |
| { |
| "epoch": 3.5494313210848643, |
| "grad_norm": 1.5546875, |
| "learning_rate": 1.5451276155799405e-05, |
| "loss": 0.4218, |
| "step": 2030 |
| }, |
| { |
| "epoch": 3.5669291338582676, |
| "grad_norm": 1.65625, |
| "learning_rate": 1.540683129557752e-05, |
| "loss": 0.4248, |
| "step": 2040 |
| }, |
| { |
| "epoch": 3.584426946631671, |
| "grad_norm": 1.71875, |
| "learning_rate": 1.5362244621356946e-05, |
| "loss": 0.4276, |
| "step": 2050 |
| }, |
| { |
| "epoch": 3.6019247594050743, |
| "grad_norm": 1.6640625, |
| "learning_rate": 1.531751756795879e-05, |
| "loss": 0.4408, |
| "step": 2060 |
| }, |
| { |
| "epoch": 3.6194225721784776, |
| "grad_norm": 1.59375, |
| "learning_rate": 1.5272651574721632e-05, |
| "loss": 0.4264, |
| "step": 2070 |
| }, |
| { |
| "epoch": 3.636920384951881, |
| "grad_norm": 1.7109375, |
| "learning_rate": 1.5227648085455202e-05, |
| "loss": 0.4295, |
| "step": 2080 |
| }, |
| { |
| "epoch": 3.6544181977252843, |
| "grad_norm": 1.7734375, |
| "learning_rate": 1.5182508548393917e-05, |
| "loss": 0.427, |
| "step": 2090 |
| }, |
| { |
| "epoch": 3.6719160104986877, |
| "grad_norm": 1.6953125, |
| "learning_rate": 1.5137234416150288e-05, |
| "loss": 0.4398, |
| "step": 2100 |
| }, |
| { |
| "epoch": 3.689413823272091, |
| "grad_norm": 1.546875, |
| "learning_rate": 1.5091827145668156e-05, |
| "loss": 0.4411, |
| "step": 2110 |
| }, |
| { |
| "epoch": 3.7069116360454943, |
| "grad_norm": 1.734375, |
| "learning_rate": 1.504628819817582e-05, |
| "loss": 0.4181, |
| "step": 2120 |
| }, |
| { |
| "epoch": 3.7244094488188977, |
| "grad_norm": 1.6484375, |
| "learning_rate": 1.5000619039139011e-05, |
| "loss": 0.4384, |
| "step": 2130 |
| }, |
| { |
| "epoch": 3.741907261592301, |
| "grad_norm": 1.7109375, |
| "learning_rate": 1.495482113821373e-05, |
| "loss": 0.4293, |
| "step": 2140 |
| }, |
| { |
| "epoch": 3.7594050743657044, |
| "grad_norm": 1.7734375, |
| "learning_rate": 1.4908895969198951e-05, |
| "loss": 0.4298, |
| "step": 2150 |
| }, |
| { |
| "epoch": 3.7769028871391077, |
| "grad_norm": 1.625, |
| "learning_rate": 1.4862845009989211e-05, |
| "loss": 0.4403, |
| "step": 2160 |
| }, |
| { |
| "epoch": 3.794400699912511, |
| "grad_norm": 1.6640625, |
| "learning_rate": 1.4816669742527018e-05, |
| "loss": 0.4373, |
| "step": 2170 |
| }, |
| { |
| "epoch": 3.8118985126859144, |
| "grad_norm": 1.546875, |
| "learning_rate": 1.4770371652755203e-05, |
| "loss": 0.4268, |
| "step": 2180 |
| }, |
| { |
| "epoch": 3.8293963254593177, |
| "grad_norm": 1.8203125, |
| "learning_rate": 1.4723952230569057e-05, |
| "loss": 0.4274, |
| "step": 2190 |
| }, |
| { |
| "epoch": 3.846894138232721, |
| "grad_norm": 1.6953125, |
| "learning_rate": 1.4677412969768427e-05, |
| "loss": 0.4313, |
| "step": 2200 |
| }, |
| { |
| "epoch": 3.864391951006124, |
| "grad_norm": 1.6484375, |
| "learning_rate": 1.4630755368009611e-05, |
| "loss": 0.4465, |
| "step": 2210 |
| }, |
| { |
| "epoch": 3.8818897637795278, |
| "grad_norm": 1.546875, |
| "learning_rate": 1.4583980926757184e-05, |
| "loss": 0.4454, |
| "step": 2220 |
| }, |
| { |
| "epoch": 3.8993875765529307, |
| "grad_norm": 1.921875, |
| "learning_rate": 1.4537091151235673e-05, |
| "loss": 0.4412, |
| "step": 2230 |
| }, |
| { |
| "epoch": 3.9168853893263345, |
| "grad_norm": 1.796875, |
| "learning_rate": 1.449008755038112e-05, |
| "loss": 0.4289, |
| "step": 2240 |
| }, |
| { |
| "epoch": 3.9343832020997374, |
| "grad_norm": 1.8515625, |
| "learning_rate": 1.4442971636792518e-05, |
| "loss": 0.4311, |
| "step": 2250 |
| }, |
| { |
| "epoch": 3.9518810148731407, |
| "grad_norm": 1.6640625, |
| "learning_rate": 1.4395744926683142e-05, |
| "loss": 0.4263, |
| "step": 2260 |
| }, |
| { |
| "epoch": 3.969378827646544, |
| "grad_norm": 1.875, |
| "learning_rate": 1.4348408939831758e-05, |
| "loss": 0.4435, |
| "step": 2270 |
| }, |
| { |
| "epoch": 3.9868766404199474, |
| "grad_norm": 1.65625, |
| "learning_rate": 1.4300965199533696e-05, |
| "loss": 0.433, |
| "step": 2280 |
| }, |
| { |
| "epoch": 4.003499562554681, |
| "grad_norm": 2.015625, |
| "learning_rate": 1.4253415232551861e-05, |
| "loss": 0.4271, |
| "step": 2290 |
| }, |
| { |
| "epoch": 4.020997375328084, |
| "grad_norm": 2.28125, |
| "learning_rate": 1.4205760569067577e-05, |
| "loss": 0.3527, |
| "step": 2300 |
| }, |
| { |
| "epoch": 4.038495188101487, |
| "grad_norm": 1.7265625, |
| "learning_rate": 1.4158002742631359e-05, |
| "loss": 0.3734, |
| "step": 2310 |
| }, |
| { |
| "epoch": 4.05599300087489, |
| "grad_norm": 2.296875, |
| "learning_rate": 1.4110143290113546e-05, |
| "loss": 0.3634, |
| "step": 2320 |
| }, |
| { |
| "epoch": 4.073490813648294, |
| "grad_norm": 1.9921875, |
| "learning_rate": 1.4062183751654868e-05, |
| "loss": 0.3462, |
| "step": 2330 |
| }, |
| { |
| "epoch": 4.090988626421697, |
| "grad_norm": 1.96875, |
| "learning_rate": 1.4014125670616856e-05, |
| "loss": 0.3487, |
| "step": 2340 |
| }, |
| { |
| "epoch": 4.108486439195101, |
| "grad_norm": 1.9921875, |
| "learning_rate": 1.3965970593532201e-05, |
| "loss": 0.3794, |
| "step": 2350 |
| }, |
| { |
| "epoch": 4.125984251968504, |
| "grad_norm": 1.890625, |
| "learning_rate": 1.3917720070054965e-05, |
| "loss": 0.3634, |
| "step": 2360 |
| }, |
| { |
| "epoch": 4.1434820647419075, |
| "grad_norm": 1.9609375, |
| "learning_rate": 1.386937565291073e-05, |
| "loss": 0.3601, |
| "step": 2370 |
| }, |
| { |
| "epoch": 4.16097987751531, |
| "grad_norm": 1.9453125, |
| "learning_rate": 1.382093889784662e-05, |
| "loss": 0.3654, |
| "step": 2380 |
| }, |
| { |
| "epoch": 4.178477690288714, |
| "grad_norm": 2.078125, |
| "learning_rate": 1.3772411363581238e-05, |
| "loss": 0.3492, |
| "step": 2390 |
| }, |
| { |
| "epoch": 4.195975503062117, |
| "grad_norm": 1.96875, |
| "learning_rate": 1.3723794611754502e-05, |
| "loss": 0.3683, |
| "step": 2400 |
| }, |
| { |
| "epoch": 4.213473315835521, |
| "grad_norm": 1.859375, |
| "learning_rate": 1.3675090206877399e-05, |
| "loss": 0.3634, |
| "step": 2410 |
| }, |
| { |
| "epoch": 4.230971128608924, |
| "grad_norm": 2.1875, |
| "learning_rate": 1.3626299716281639e-05, |
| "loss": 0.3623, |
| "step": 2420 |
| }, |
| { |
| "epoch": 4.248468941382328, |
| "grad_norm": 1.890625, |
| "learning_rate": 1.3577424710069202e-05, |
| "loss": 0.3596, |
| "step": 2430 |
| }, |
| { |
| "epoch": 4.2659667541557305, |
| "grad_norm": 2.25, |
| "learning_rate": 1.3528466761061839e-05, |
| "loss": 0.3623, |
| "step": 2440 |
| }, |
| { |
| "epoch": 4.283464566929134, |
| "grad_norm": 1.9453125, |
| "learning_rate": 1.3479427444750415e-05, |
| "loss": 0.3535, |
| "step": 2450 |
| }, |
| { |
| "epoch": 4.300962379702537, |
| "grad_norm": 2.09375, |
| "learning_rate": 1.343030833924426e-05, |
| "loss": 0.3592, |
| "step": 2460 |
| }, |
| { |
| "epoch": 4.318460192475941, |
| "grad_norm": 1.9765625, |
| "learning_rate": 1.338111102522035e-05, |
| "loss": 0.3644, |
| "step": 2470 |
| }, |
| { |
| "epoch": 4.335958005249344, |
| "grad_norm": 2.25, |
| "learning_rate": 1.3331837085872444e-05, |
| "loss": 0.3565, |
| "step": 2480 |
| }, |
| { |
| "epoch": 4.353455818022747, |
| "grad_norm": 2.09375, |
| "learning_rate": 1.3282488106860146e-05, |
| "loss": 0.3613, |
| "step": 2490 |
| }, |
| { |
| "epoch": 4.3709536307961505, |
| "grad_norm": 1.984375, |
| "learning_rate": 1.323306567625788e-05, |
| "loss": 0.3574, |
| "step": 2500 |
| }, |
| { |
| "epoch": 4.388451443569553, |
| "grad_norm": 2.484375, |
| "learning_rate": 1.3183571384503767e-05, |
| "loss": 0.3507, |
| "step": 2510 |
| }, |
| { |
| "epoch": 4.405949256342957, |
| "grad_norm": 1.8671875, |
| "learning_rate": 1.3134006824348464e-05, |
| "loss": 0.3583, |
| "step": 2520 |
| }, |
| { |
| "epoch": 4.42344706911636, |
| "grad_norm": 2.0, |
| "learning_rate": 1.3084373590803898e-05, |
| "loss": 0.3567, |
| "step": 2530 |
| }, |
| { |
| "epoch": 4.440944881889764, |
| "grad_norm": 1.96875, |
| "learning_rate": 1.303467328109193e-05, |
| "loss": 0.3538, |
| "step": 2540 |
| }, |
| { |
| "epoch": 4.458442694663167, |
| "grad_norm": 2.046875, |
| "learning_rate": 1.2984907494592983e-05, |
| "loss": 0.3623, |
| "step": 2550 |
| }, |
| { |
| "epoch": 4.475940507436571, |
| "grad_norm": 1.9375, |
| "learning_rate": 1.2935077832794533e-05, |
| "loss": 0.3649, |
| "step": 2560 |
| }, |
| { |
| "epoch": 4.4934383202099735, |
| "grad_norm": 1.984375, |
| "learning_rate": 1.2885185899239617e-05, |
| "loss": 0.3551, |
| "step": 2570 |
| }, |
| { |
| "epoch": 4.510936132983377, |
| "grad_norm": 2.140625, |
| "learning_rate": 1.2835233299475192e-05, |
| "loss": 0.354, |
| "step": 2580 |
| }, |
| { |
| "epoch": 4.52843394575678, |
| "grad_norm": 2.109375, |
| "learning_rate": 1.2785221641000487e-05, |
| "loss": 0.3652, |
| "step": 2590 |
| }, |
| { |
| "epoch": 4.545931758530184, |
| "grad_norm": 2.0625, |
| "learning_rate": 1.2735152533215275e-05, |
| "loss": 0.3597, |
| "step": 2600 |
| }, |
| { |
| "epoch": 4.563429571303587, |
| "grad_norm": 1.9375, |
| "learning_rate": 1.2685027587368072e-05, |
| "loss": 0.3567, |
| "step": 2610 |
| }, |
| { |
| "epoch": 4.580927384076991, |
| "grad_norm": 1.75, |
| "learning_rate": 1.2634848416504289e-05, |
| "loss": 0.3577, |
| "step": 2620 |
| }, |
| { |
| "epoch": 4.5984251968503935, |
| "grad_norm": 1.9453125, |
| "learning_rate": 1.2584616635414325e-05, |
| "loss": 0.3576, |
| "step": 2630 |
| }, |
| { |
| "epoch": 4.615923009623797, |
| "grad_norm": 1.9609375, |
| "learning_rate": 1.2534333860581607e-05, |
| "loss": 0.3609, |
| "step": 2640 |
| }, |
| { |
| "epoch": 4.6334208223972, |
| "grad_norm": 2.109375, |
| "learning_rate": 1.248400171013056e-05, |
| "loss": 0.3525, |
| "step": 2650 |
| }, |
| { |
| "epoch": 4.650918635170604, |
| "grad_norm": 1.7734375, |
| "learning_rate": 1.2433621803774542e-05, |
| "loss": 0.3567, |
| "step": 2660 |
| }, |
| { |
| "epoch": 4.668416447944007, |
| "grad_norm": 1.921875, |
| "learning_rate": 1.2383195762763718e-05, |
| "loss": 0.3623, |
| "step": 2670 |
| }, |
| { |
| "epoch": 4.685914260717411, |
| "grad_norm": 1.8046875, |
| "learning_rate": 1.2332725209832892e-05, |
| "loss": 0.3643, |
| "step": 2680 |
| }, |
| { |
| "epoch": 4.703412073490814, |
| "grad_norm": 1.8828125, |
| "learning_rate": 1.2282211769149283e-05, |
| "loss": 0.3642, |
| "step": 2690 |
| }, |
| { |
| "epoch": 4.7209098862642165, |
| "grad_norm": 1.96875, |
| "learning_rate": 1.2231657066260261e-05, |
| "loss": 0.348, |
| "step": 2700 |
| }, |
| { |
| "epoch": 4.73840769903762, |
| "grad_norm": 2.03125, |
| "learning_rate": 1.2181062728041029e-05, |
| "loss": 0.3545, |
| "step": 2710 |
| }, |
| { |
| "epoch": 4.755905511811024, |
| "grad_norm": 2.5625, |
| "learning_rate": 1.2130430382642276e-05, |
| "loss": 0.3624, |
| "step": 2720 |
| }, |
| { |
| "epoch": 4.773403324584427, |
| "grad_norm": 1.984375, |
| "learning_rate": 1.2079761659437781e-05, |
| "loss": 0.3719, |
| "step": 2730 |
| }, |
| { |
| "epoch": 4.79090113735783, |
| "grad_norm": 2.140625, |
| "learning_rate": 1.202905818897198e-05, |
| "loss": 0.362, |
| "step": 2740 |
| }, |
| { |
| "epoch": 4.808398950131234, |
| "grad_norm": 2.15625, |
| "learning_rate": 1.1978321602907497e-05, |
| "loss": 0.3691, |
| "step": 2750 |
| }, |
| { |
| "epoch": 4.8258967629046365, |
| "grad_norm": 1.953125, |
| "learning_rate": 1.192755353397262e-05, |
| "loss": 0.3611, |
| "step": 2760 |
| }, |
| { |
| "epoch": 4.84339457567804, |
| "grad_norm": 2.21875, |
| "learning_rate": 1.187675561590878e-05, |
| "loss": 0.356, |
| "step": 2770 |
| }, |
| { |
| "epoch": 4.860892388451443, |
| "grad_norm": 2.0, |
| "learning_rate": 1.1825929483417975e-05, |
| "loss": 0.349, |
| "step": 2780 |
| }, |
| { |
| "epoch": 4.878390201224847, |
| "grad_norm": 2.265625, |
| "learning_rate": 1.1775076772110136e-05, |
| "loss": 0.3643, |
| "step": 2790 |
| }, |
| { |
| "epoch": 4.89588801399825, |
| "grad_norm": 1.953125, |
| "learning_rate": 1.1724199118450537e-05, |
| "loss": 0.358, |
| "step": 2800 |
| }, |
| { |
| "epoch": 4.913385826771654, |
| "grad_norm": 2.359375, |
| "learning_rate": 1.1673298159707086e-05, |
| "loss": 0.358, |
| "step": 2810 |
| }, |
| { |
| "epoch": 4.930883639545057, |
| "grad_norm": 1.875, |
| "learning_rate": 1.1622375533897683e-05, |
| "loss": 0.3705, |
| "step": 2820 |
| }, |
| { |
| "epoch": 4.94838145231846, |
| "grad_norm": 2.171875, |
| "learning_rate": 1.157143287973747e-05, |
| "loss": 0.3638, |
| "step": 2830 |
| }, |
| { |
| "epoch": 4.965879265091863, |
| "grad_norm": 2.46875, |
| "learning_rate": 1.1520471836586115e-05, |
| "loss": 0.3417, |
| "step": 2840 |
| }, |
| { |
| "epoch": 4.983377077865267, |
| "grad_norm": 2.21875, |
| "learning_rate": 1.1469494044395055e-05, |
| "loss": 0.3732, |
| "step": 2850 |
| }, |
| { |
| "epoch": 5.0, |
| "grad_norm": 2.796875, |
| "learning_rate": 1.1418501143654713e-05, |
| "loss": 0.376, |
| "step": 2860 |
| }, |
| { |
| "epoch": 5.017497812773403, |
| "grad_norm": 2.6875, |
| "learning_rate": 1.1367494775341716e-05, |
| "loss": 0.3109, |
| "step": 2870 |
| }, |
| { |
| "epoch": 5.034995625546807, |
| "grad_norm": 2.265625, |
| "learning_rate": 1.1316476580866091e-05, |
| "loss": 0.3136, |
| "step": 2880 |
| }, |
| { |
| "epoch": 5.05249343832021, |
| "grad_norm": 2.78125, |
| "learning_rate": 1.1265448202018432e-05, |
| "loss": 0.3137, |
| "step": 2890 |
| }, |
| { |
| "epoch": 5.069991251093613, |
| "grad_norm": 2.71875, |
| "learning_rate": 1.121441128091706e-05, |
| "loss": 0.3082, |
| "step": 2900 |
| }, |
| { |
| "epoch": 5.087489063867016, |
| "grad_norm": 2.109375, |
| "learning_rate": 1.1163367459955208e-05, |
| "loss": 0.3102, |
| "step": 2910 |
| }, |
| { |
| "epoch": 5.10498687664042, |
| "grad_norm": 2.4375, |
| "learning_rate": 1.1112318381748135e-05, |
| "loss": 0.3092, |
| "step": 2920 |
| }, |
| { |
| "epoch": 5.122484689413823, |
| "grad_norm": 2.953125, |
| "learning_rate": 1.1061265689080296e-05, |
| "loss": 0.3054, |
| "step": 2930 |
| }, |
| { |
| "epoch": 5.139982502187227, |
| "grad_norm": 2.0625, |
| "learning_rate": 1.1010211024852445e-05, |
| "loss": 0.3165, |
| "step": 2940 |
| }, |
| { |
| "epoch": 5.15748031496063, |
| "grad_norm": 2.078125, |
| "learning_rate": 1.0959156032028789e-05, |
| "loss": 0.3081, |
| "step": 2950 |
| }, |
| { |
| "epoch": 5.174978127734033, |
| "grad_norm": 2.015625, |
| "learning_rate": 1.0908102353584117e-05, |
| "loss": 0.3047, |
| "step": 2960 |
| }, |
| { |
| "epoch": 5.192475940507436, |
| "grad_norm": 2.0, |
| "learning_rate": 1.0857051632450905e-05, |
| "loss": 0.3016, |
| "step": 2970 |
| }, |
| { |
| "epoch": 5.20997375328084, |
| "grad_norm": 2.1875, |
| "learning_rate": 1.0806005511466478e-05, |
| "loss": 0.306, |
| "step": 2980 |
| }, |
| { |
| "epoch": 5.227471566054243, |
| "grad_norm": 2.5, |
| "learning_rate": 1.075496563332011e-05, |
| "loss": 0.3036, |
| "step": 2990 |
| }, |
| { |
| "epoch": 5.244969378827647, |
| "grad_norm": 2.203125, |
| "learning_rate": 1.0703933640500183e-05, |
| "loss": 0.3041, |
| "step": 3000 |
| }, |
| { |
| "epoch": 5.26246719160105, |
| "grad_norm": 2.515625, |
| "learning_rate": 1.0652911175241336e-05, |
| "loss": 0.3183, |
| "step": 3010 |
| }, |
| { |
| "epoch": 5.2799650043744535, |
| "grad_norm": 2.515625, |
| "learning_rate": 1.0601899879471593e-05, |
| "loss": 0.2964, |
| "step": 3020 |
| }, |
| { |
| "epoch": 5.297462817147856, |
| "grad_norm": 2.453125, |
| "learning_rate": 1.0550901394759543e-05, |
| "loss": 0.3006, |
| "step": 3030 |
| }, |
| { |
| "epoch": 5.31496062992126, |
| "grad_norm": 2.25, |
| "learning_rate": 1.0499917362261505e-05, |
| "loss": 0.3146, |
| "step": 3040 |
| }, |
| { |
| "epoch": 5.332458442694663, |
| "grad_norm": 2.421875, |
| "learning_rate": 1.0448949422668729e-05, |
| "loss": 0.3153, |
| "step": 3050 |
| }, |
| { |
| "epoch": 5.349956255468067, |
| "grad_norm": 2.234375, |
| "learning_rate": 1.039799921615458e-05, |
| "loss": 0.3166, |
| "step": 3060 |
| }, |
| { |
| "epoch": 5.36745406824147, |
| "grad_norm": 2.234375, |
| "learning_rate": 1.0347068382321758e-05, |
| "loss": 0.3069, |
| "step": 3070 |
| }, |
| { |
| "epoch": 5.3849518810148735, |
| "grad_norm": 2.1875, |
| "learning_rate": 1.0296158560149554e-05, |
| "loss": 0.3031, |
| "step": 3080 |
| }, |
| { |
| "epoch": 5.402449693788276, |
| "grad_norm": 2.75, |
| "learning_rate": 1.0245271387941085e-05, |
| "loss": 0.308, |
| "step": 3090 |
| }, |
| { |
| "epoch": 5.41994750656168, |
| "grad_norm": 2.09375, |
| "learning_rate": 1.0194408503270582e-05, |
| "loss": 0.3081, |
| "step": 3100 |
| }, |
| { |
| "epoch": 5.437445319335083, |
| "grad_norm": 2.25, |
| "learning_rate": 1.0143571542930695e-05, |
| "loss": 0.3052, |
| "step": 3110 |
| }, |
| { |
| "epoch": 5.454943132108486, |
| "grad_norm": 2.140625, |
| "learning_rate": 1.0092762142879803e-05, |
| "loss": 0.3156, |
| "step": 3120 |
| }, |
| { |
| "epoch": 5.47244094488189, |
| "grad_norm": 2.109375, |
| "learning_rate": 1.0041981938189401e-05, |
| "loss": 0.3076, |
| "step": 3130 |
| }, |
| { |
| "epoch": 5.489938757655293, |
| "grad_norm": 2.171875, |
| "learning_rate": 9.99123256299145e-06, |
| "loss": 0.3164, |
| "step": 3140 |
| }, |
| { |
| "epoch": 5.5074365704286965, |
| "grad_norm": 2.234375, |
| "learning_rate": 9.94051565042581e-06, |
| "loss": 0.3215, |
| "step": 3150 |
| }, |
| { |
| "epoch": 5.524934383202099, |
| "grad_norm": 2.3125, |
| "learning_rate": 9.889832832587667e-06, |
| "loss": 0.3076, |
| "step": 3160 |
| }, |
| { |
| "epoch": 5.542432195975503, |
| "grad_norm": 2.296875, |
| "learning_rate": 9.83918574047503e-06, |
| "loss": 0.288, |
| "step": 3170 |
| }, |
| { |
| "epoch": 5.559930008748906, |
| "grad_norm": 2.25, |
| "learning_rate": 9.788576003936239e-06, |
| "loss": 0.3116, |
| "step": 3180 |
| }, |
| { |
| "epoch": 5.57742782152231, |
| "grad_norm": 2.203125, |
| "learning_rate": 9.738005251617508e-06, |
| "loss": 0.3083, |
| "step": 3190 |
| }, |
| { |
| "epoch": 5.594925634295713, |
| "grad_norm": 2.234375, |
| "learning_rate": 9.68747511091052e-06, |
| "loss": 0.3078, |
| "step": 3200 |
| }, |
| { |
| "epoch": 5.6124234470691166, |
| "grad_norm": 2.421875, |
| "learning_rate": 9.636987207900056e-06, |
| "loss": 0.3131, |
| "step": 3210 |
| }, |
| { |
| "epoch": 5.6299212598425195, |
| "grad_norm": 2.203125, |
| "learning_rate": 9.58654316731167e-06, |
| "loss": 0.3147, |
| "step": 3220 |
| }, |
| { |
| "epoch": 5.647419072615923, |
| "grad_norm": 2.109375, |
| "learning_rate": 9.536144612459399e-06, |
| "loss": 0.3131, |
| "step": 3230 |
| }, |
| { |
| "epoch": 5.664916885389326, |
| "grad_norm": 1.953125, |
| "learning_rate": 9.485793165193531e-06, |
| "loss": 0.3147, |
| "step": 3240 |
| }, |
| { |
| "epoch": 5.68241469816273, |
| "grad_norm": 2.234375, |
| "learning_rate": 9.435490445848403e-06, |
| "loss": 0.3124, |
| "step": 3250 |
| }, |
| { |
| "epoch": 5.699912510936133, |
| "grad_norm": 2.125, |
| "learning_rate": 9.385238073190264e-06, |
| "loss": 0.3091, |
| "step": 3260 |
| }, |
| { |
| "epoch": 5.717410323709537, |
| "grad_norm": 2.59375, |
| "learning_rate": 9.335037664365185e-06, |
| "loss": 0.2988, |
| "step": 3270 |
| }, |
| { |
| "epoch": 5.7349081364829395, |
| "grad_norm": 2.3125, |
| "learning_rate": 9.284890834847022e-06, |
| "loss": 0.306, |
| "step": 3280 |
| }, |
| { |
| "epoch": 5.752405949256343, |
| "grad_norm": 2.515625, |
| "learning_rate": 9.234799198385402e-06, |
| "loss": 0.3214, |
| "step": 3290 |
| }, |
| { |
| "epoch": 5.769903762029746, |
| "grad_norm": 2.046875, |
| "learning_rate": 9.184764366953831e-06, |
| "loss": 0.3178, |
| "step": 3300 |
| }, |
| { |
| "epoch": 5.78740157480315, |
| "grad_norm": 2.046875, |
| "learning_rate": 9.134787950697798e-06, |
| "loss": 0.3061, |
| "step": 3310 |
| }, |
| { |
| "epoch": 5.804899387576553, |
| "grad_norm": 2.25, |
| "learning_rate": 9.084871557882957e-06, |
| "loss": 0.3096, |
| "step": 3320 |
| }, |
| { |
| "epoch": 5.822397200349956, |
| "grad_norm": 2.28125, |
| "learning_rate": 9.035016794843383e-06, |
| "loss": 0.3131, |
| "step": 3330 |
| }, |
| { |
| "epoch": 5.83989501312336, |
| "grad_norm": 2.09375, |
| "learning_rate": 8.985225265929872e-06, |
| "loss": 0.307, |
| "step": 3340 |
| }, |
| { |
| "epoch": 5.857392825896763, |
| "grad_norm": 2.390625, |
| "learning_rate": 8.93549857345831e-06, |
| "loss": 0.3107, |
| "step": 3350 |
| }, |
| { |
| "epoch": 5.874890638670166, |
| "grad_norm": 2.09375, |
| "learning_rate": 8.885838317658126e-06, |
| "loss": 0.3041, |
| "step": 3360 |
| }, |
| { |
| "epoch": 5.892388451443569, |
| "grad_norm": 1.9765625, |
| "learning_rate": 8.83624609662078e-06, |
| "loss": 0.3123, |
| "step": 3370 |
| }, |
| { |
| "epoch": 5.909886264216973, |
| "grad_norm": 2.234375, |
| "learning_rate": 8.786723506248334e-06, |
| "loss": 0.316, |
| "step": 3380 |
| }, |
| { |
| "epoch": 5.927384076990376, |
| "grad_norm": 2.625, |
| "learning_rate": 8.737272140202105e-06, |
| "loss": 0.3015, |
| "step": 3390 |
| }, |
| { |
| "epoch": 5.94488188976378, |
| "grad_norm": 2.1875, |
| "learning_rate": 8.687893589851375e-06, |
| "loss": 0.3216, |
| "step": 3400 |
| }, |
| { |
| "epoch": 5.9623797025371825, |
| "grad_norm": 2.21875, |
| "learning_rate": 8.638589444222191e-06, |
| "loss": 0.3143, |
| "step": 3410 |
| }, |
| { |
| "epoch": 5.979877515310586, |
| "grad_norm": 2.40625, |
| "learning_rate": 8.5893612899462e-06, |
| "loss": 0.3089, |
| "step": 3420 |
| }, |
| { |
| "epoch": 5.997375328083989, |
| "grad_norm": 2.171875, |
| "learning_rate": 8.540210711209626e-06, |
| "loss": 0.3137, |
| "step": 3430 |
| }, |
| { |
| "epoch": 6.013998250218723, |
| "grad_norm": 2.40625, |
| "learning_rate": 8.491139289702262e-06, |
| "loss": 0.2933, |
| "step": 3440 |
| }, |
| { |
| "epoch": 6.031496062992126, |
| "grad_norm": 2.28125, |
| "learning_rate": 8.442148604566598e-06, |
| "loss": 0.2787, |
| "step": 3450 |
| }, |
| { |
| "epoch": 6.04899387576553, |
| "grad_norm": 2.46875, |
| "learning_rate": 8.393240232346976e-06, |
| "loss": 0.2763, |
| "step": 3460 |
| }, |
| { |
| "epoch": 6.066491688538933, |
| "grad_norm": 2.328125, |
| "learning_rate": 8.34441574693887e-06, |
| "loss": 0.2706, |
| "step": 3470 |
| }, |
| { |
| "epoch": 6.083989501312336, |
| "grad_norm": 2.21875, |
| "learning_rate": 8.295676719538231e-06, |
| "loss": 0.2732, |
| "step": 3480 |
| }, |
| { |
| "epoch": 6.101487314085739, |
| "grad_norm": 2.25, |
| "learning_rate": 8.247024718590941e-06, |
| "loss": 0.2772, |
| "step": 3490 |
| }, |
| { |
| "epoch": 6.118985126859142, |
| "grad_norm": 2.484375, |
| "learning_rate": 8.19846130974232e-06, |
| "loss": 0.2791, |
| "step": 3500 |
| }, |
| { |
| "epoch": 6.136482939632546, |
| "grad_norm": 2.484375, |
| "learning_rate": 8.14998805578675e-06, |
| "loss": 0.2808, |
| "step": 3510 |
| }, |
| { |
| "epoch": 6.153980752405949, |
| "grad_norm": 2.46875, |
| "learning_rate": 8.101606516617384e-06, |
| "loss": 0.2908, |
| "step": 3520 |
| }, |
| { |
| "epoch": 6.171478565179353, |
| "grad_norm": 2.359375, |
| "learning_rate": 8.053318249175955e-06, |
| "loss": 0.2809, |
| "step": 3530 |
| }, |
| { |
| "epoch": 6.188976377952756, |
| "grad_norm": 2.296875, |
| "learning_rate": 8.005124807402657e-06, |
| "loss": 0.2708, |
| "step": 3540 |
| }, |
| { |
| "epoch": 6.206474190726159, |
| "grad_norm": 2.265625, |
| "learning_rate": 7.957027742186153e-06, |
| "loss": 0.2789, |
| "step": 3550 |
| }, |
| { |
| "epoch": 6.223972003499562, |
| "grad_norm": 2.234375, |
| "learning_rate": 7.909028601313658e-06, |
| "loss": 0.2752, |
| "step": 3560 |
| }, |
| { |
| "epoch": 6.241469816272966, |
| "grad_norm": 2.234375, |
| "learning_rate": 7.861128929421133e-06, |
| "loss": 0.2778, |
| "step": 3570 |
| }, |
| { |
| "epoch": 6.258967629046369, |
| "grad_norm": 2.640625, |
| "learning_rate": 7.813330267943586e-06, |
| "loss": 0.2779, |
| "step": 3580 |
| }, |
| { |
| "epoch": 6.276465441819773, |
| "grad_norm": 2.015625, |
| "learning_rate": 7.765634155065451e-06, |
| "loss": 0.2696, |
| "step": 3590 |
| }, |
| { |
| "epoch": 6.293963254593176, |
| "grad_norm": 2.171875, |
| "learning_rate": 7.718042125671102e-06, |
| "loss": 0.2813, |
| "step": 3600 |
| }, |
| { |
| "epoch": 6.311461067366579, |
| "grad_norm": 2.21875, |
| "learning_rate": 7.670555711295446e-06, |
| "loss": 0.2847, |
| "step": 3610 |
| }, |
| { |
| "epoch": 6.328958880139982, |
| "grad_norm": 2.515625, |
| "learning_rate": 7.623176440074667e-06, |
| "loss": 0.2761, |
| "step": 3620 |
| }, |
| { |
| "epoch": 6.346456692913386, |
| "grad_norm": 2.546875, |
| "learning_rate": 7.575905836697008e-06, |
| "loss": 0.2801, |
| "step": 3630 |
| }, |
| { |
| "epoch": 6.363954505686789, |
| "grad_norm": 2.109375, |
| "learning_rate": 7.528745422353745e-06, |
| "loss": 0.2848, |
| "step": 3640 |
| }, |
| { |
| "epoch": 6.381452318460193, |
| "grad_norm": 2.75, |
| "learning_rate": 7.481696714690203e-06, |
| "loss": 0.2834, |
| "step": 3650 |
| }, |
| { |
| "epoch": 6.398950131233596, |
| "grad_norm": 2.140625, |
| "learning_rate": 7.4347612277569345e-06, |
| "loss": 0.2743, |
| "step": 3660 |
| }, |
| { |
| "epoch": 6.4164479440069995, |
| "grad_norm": 2.015625, |
| "learning_rate": 7.387940471961001e-06, |
| "loss": 0.2794, |
| "step": 3670 |
| }, |
| { |
| "epoch": 6.433945756780402, |
| "grad_norm": 2.234375, |
| "learning_rate": 7.341235954017351e-06, |
| "loss": 0.277, |
| "step": 3680 |
| }, |
| { |
| "epoch": 6.451443569553806, |
| "grad_norm": 2.15625, |
| "learning_rate": 7.294649176900344e-06, |
| "loss": 0.2817, |
| "step": 3690 |
| }, |
| { |
| "epoch": 6.468941382327209, |
| "grad_norm": 2.265625, |
| "learning_rate": 7.248181639795384e-06, |
| "loss": 0.2816, |
| "step": 3700 |
| }, |
| { |
| "epoch": 6.486439195100612, |
| "grad_norm": 2.328125, |
| "learning_rate": 7.201834838050668e-06, |
| "loss": 0.2765, |
| "step": 3710 |
| }, |
| { |
| "epoch": 6.503937007874016, |
| "grad_norm": 2.34375, |
| "learning_rate": 7.155610263129082e-06, |
| "loss": 0.2778, |
| "step": 3720 |
| }, |
| { |
| "epoch": 6.5214348206474195, |
| "grad_norm": 2.328125, |
| "learning_rate": 7.109509402560171e-06, |
| "loss": 0.27, |
| "step": 3730 |
| }, |
| { |
| "epoch": 6.538932633420822, |
| "grad_norm": 2.359375, |
| "learning_rate": 7.063533739892312e-06, |
| "loss": 0.2713, |
| "step": 3740 |
| }, |
| { |
| "epoch": 6.556430446194225, |
| "grad_norm": 2.375, |
| "learning_rate": 7.017684754644938e-06, |
| "loss": 0.2801, |
| "step": 3750 |
| }, |
| { |
| "epoch": 6.573928258967629, |
| "grad_norm": 2.234375, |
| "learning_rate": 6.971963922260954e-06, |
| "loss": 0.2893, |
| "step": 3760 |
| }, |
| { |
| "epoch": 6.591426071741032, |
| "grad_norm": 2.71875, |
| "learning_rate": 6.926372714059227e-06, |
| "loss": 0.2785, |
| "step": 3770 |
| }, |
| { |
| "epoch": 6.608923884514436, |
| "grad_norm": 2.28125, |
| "learning_rate": 6.880912597187266e-06, |
| "loss": 0.2762, |
| "step": 3780 |
| }, |
| { |
| "epoch": 6.626421697287839, |
| "grad_norm": 2.421875, |
| "learning_rate": 6.835585034573993e-06, |
| "loss": 0.2791, |
| "step": 3790 |
| }, |
| { |
| "epoch": 6.6439195100612425, |
| "grad_norm": 2.796875, |
| "learning_rate": 6.790391484882665e-06, |
| "loss": 0.2795, |
| "step": 3800 |
| }, |
| { |
| "epoch": 6.661417322834645, |
| "grad_norm": 2.078125, |
| "learning_rate": 6.745333402463944e-06, |
| "loss": 0.2845, |
| "step": 3810 |
| }, |
| { |
| "epoch": 6.678915135608049, |
| "grad_norm": 2.46875, |
| "learning_rate": 6.700412237309084e-06, |
| "loss": 0.2829, |
| "step": 3820 |
| }, |
| { |
| "epoch": 6.696412948381452, |
| "grad_norm": 2.421875, |
| "learning_rate": 6.655629435003274e-06, |
| "loss": 0.2814, |
| "step": 3830 |
| }, |
| { |
| "epoch": 6.713910761154856, |
| "grad_norm": 2.515625, |
| "learning_rate": 6.6109864366791225e-06, |
| "loss": 0.2703, |
| "step": 3840 |
| }, |
| { |
| "epoch": 6.731408573928259, |
| "grad_norm": 2.1875, |
| "learning_rate": 6.56648467897027e-06, |
| "loss": 0.2856, |
| "step": 3850 |
| }, |
| { |
| "epoch": 6.7489063867016625, |
| "grad_norm": 2.109375, |
| "learning_rate": 6.522125593965171e-06, |
| "loss": 0.2766, |
| "step": 3860 |
| }, |
| { |
| "epoch": 6.766404199475065, |
| "grad_norm": 2.546875, |
| "learning_rate": 6.477910609161004e-06, |
| "loss": 0.2792, |
| "step": 3870 |
| }, |
| { |
| "epoch": 6.783902012248469, |
| "grad_norm": 2.5, |
| "learning_rate": 6.433841147417717e-06, |
| "loss": 0.28, |
| "step": 3880 |
| }, |
| { |
| "epoch": 6.801399825021872, |
| "grad_norm": 2.125, |
| "learning_rate": 6.389918626912277e-06, |
| "loss": 0.2773, |
| "step": 3890 |
| }, |
| { |
| "epoch": 6.818897637795276, |
| "grad_norm": 2.3125, |
| "learning_rate": 6.346144461092991e-06, |
| "loss": 0.2761, |
| "step": 3900 |
| }, |
| { |
| "epoch": 6.836395450568679, |
| "grad_norm": 2.484375, |
| "learning_rate": 6.302520058634057e-06, |
| "loss": 0.2765, |
| "step": 3910 |
| }, |
| { |
| "epoch": 6.853893263342083, |
| "grad_norm": 2.125, |
| "learning_rate": 6.259046823390184e-06, |
| "loss": 0.2746, |
| "step": 3920 |
| }, |
| { |
| "epoch": 6.8713910761154855, |
| "grad_norm": 2.140625, |
| "learning_rate": 6.2157261543514825e-06, |
| "loss": 0.2832, |
| "step": 3930 |
| }, |
| { |
| "epoch": 6.888888888888889, |
| "grad_norm": 2.171875, |
| "learning_rate": 6.172559445598385e-06, |
| "loss": 0.2831, |
| "step": 3940 |
| }, |
| { |
| "epoch": 6.906386701662292, |
| "grad_norm": 2.09375, |
| "learning_rate": 6.129548086256816e-06, |
| "loss": 0.2821, |
| "step": 3950 |
| }, |
| { |
| "epoch": 6.923884514435695, |
| "grad_norm": 2.421875, |
| "learning_rate": 6.086693460453466e-06, |
| "loss": 0.2905, |
| "step": 3960 |
| }, |
| { |
| "epoch": 6.941382327209099, |
| "grad_norm": 2.046875, |
| "learning_rate": 6.0439969472712734e-06, |
| "loss": 0.2772, |
| "step": 3970 |
| }, |
| { |
| "epoch": 6.958880139982503, |
| "grad_norm": 2.28125, |
| "learning_rate": 6.001459920705046e-06, |
| "loss": 0.2814, |
| "step": 3980 |
| }, |
| { |
| "epoch": 6.9763779527559056, |
| "grad_norm": 2.59375, |
| "learning_rate": 5.95908374961721e-06, |
| "loss": 0.2674, |
| "step": 3990 |
| }, |
| { |
| "epoch": 6.9938757655293085, |
| "grad_norm": 2.234375, |
| "learning_rate": 5.916869797693794e-06, |
| "loss": 0.2746, |
| "step": 4000 |
| }, |
| { |
| "epoch": 7.010498687664042, |
| "grad_norm": 2.4375, |
| "learning_rate": 5.874819423400538e-06, |
| "loss": 0.2766, |
| "step": 4010 |
| }, |
| { |
| "epoch": 7.027996500437445, |
| "grad_norm": 2.1875, |
| "learning_rate": 5.832933979939177e-06, |
| "loss": 0.257, |
| "step": 4020 |
| }, |
| { |
| "epoch": 7.045494313210849, |
| "grad_norm": 2.328125, |
| "learning_rate": 5.791214815203874e-06, |
| "loss": 0.2575, |
| "step": 4030 |
| }, |
| { |
| "epoch": 7.062992125984252, |
| "grad_norm": 2.03125, |
| "learning_rate": 5.749663271737873e-06, |
| "loss": 0.2589, |
| "step": 4040 |
| }, |
| { |
| "epoch": 7.080489938757656, |
| "grad_norm": 2.046875, |
| "learning_rate": 5.708280686690284e-06, |
| "loss": 0.2585, |
| "step": 4050 |
| }, |
| { |
| "epoch": 7.0979877515310585, |
| "grad_norm": 2.109375, |
| "learning_rate": 5.667068391773045e-06, |
| "loss": 0.2696, |
| "step": 4060 |
| }, |
| { |
| "epoch": 7.115485564304462, |
| "grad_norm": 2.703125, |
| "learning_rate": 5.62602771321808e-06, |
| "loss": 0.263, |
| "step": 4070 |
| }, |
| { |
| "epoch": 7.132983377077865, |
| "grad_norm": 2.453125, |
| "learning_rate": 5.5851599717346106e-06, |
| "loss": 0.2598, |
| "step": 4080 |
| }, |
| { |
| "epoch": 7.150481189851269, |
| "grad_norm": 2.421875, |
| "learning_rate": 5.5444664824666594e-06, |
| "loss": 0.2683, |
| "step": 4090 |
| }, |
| { |
| "epoch": 7.167979002624672, |
| "grad_norm": 2.3125, |
| "learning_rate": 5.503948554950727e-06, |
| "loss": 0.2669, |
| "step": 4100 |
| }, |
| { |
| "epoch": 7.185476815398075, |
| "grad_norm": 2.328125, |
| "learning_rate": 5.4636074930736525e-06, |
| "loss": 0.2546, |
| "step": 4110 |
| }, |
| { |
| "epoch": 7.202974628171479, |
| "grad_norm": 2.328125, |
| "learning_rate": 5.423444595030648e-06, |
| "loss": 0.262, |
| "step": 4120 |
| }, |
| { |
| "epoch": 7.2204724409448815, |
| "grad_norm": 2.21875, |
| "learning_rate": 5.383461153283529e-06, |
| "loss": 0.2772, |
| "step": 4130 |
| }, |
| { |
| "epoch": 7.237970253718285, |
| "grad_norm": 2.203125, |
| "learning_rate": 5.343658454519113e-06, |
| "loss": 0.2584, |
| "step": 4140 |
| }, |
| { |
| "epoch": 7.255468066491688, |
| "grad_norm": 2.3125, |
| "learning_rate": 5.3040377796078295e-06, |
| "loss": 0.2626, |
| "step": 4150 |
| }, |
| { |
| "epoch": 7.272965879265092, |
| "grad_norm": 2.09375, |
| "learning_rate": 5.264600403562482e-06, |
| "loss": 0.2643, |
| "step": 4160 |
| }, |
| { |
| "epoch": 7.290463692038495, |
| "grad_norm": 2.1875, |
| "learning_rate": 5.225347595497234e-06, |
| "loss": 0.2598, |
| "step": 4170 |
| }, |
| { |
| "epoch": 7.307961504811899, |
| "grad_norm": 2.09375, |
| "learning_rate": 5.186280618586752e-06, |
| "loss": 0.2584, |
| "step": 4180 |
| }, |
| { |
| "epoch": 7.3254593175853016, |
| "grad_norm": 2.421875, |
| "learning_rate": 5.147400730025567e-06, |
| "loss": 0.2551, |
| "step": 4190 |
| }, |
| { |
| "epoch": 7.342957130358705, |
| "grad_norm": 2.03125, |
| "learning_rate": 5.108709180987623e-06, |
| "loss": 0.2573, |
| "step": 4200 |
| }, |
| { |
| "epoch": 7.360454943132108, |
| "grad_norm": 2.09375, |
| "learning_rate": 5.070207216586e-06, |
| "loss": 0.2546, |
| "step": 4210 |
| }, |
| { |
| "epoch": 7.377952755905512, |
| "grad_norm": 2.203125, |
| "learning_rate": 5.031896075832846e-06, |
| "loss": 0.2503, |
| "step": 4220 |
| }, |
| { |
| "epoch": 7.395450568678915, |
| "grad_norm": 2.171875, |
| "learning_rate": 4.993776991599511e-06, |
| "loss": 0.2612, |
| "step": 4230 |
| }, |
| { |
| "epoch": 7.412948381452319, |
| "grad_norm": 2.40625, |
| "learning_rate": 4.955851190576886e-06, |
| "loss": 0.2624, |
| "step": 4240 |
| }, |
| { |
| "epoch": 7.430446194225722, |
| "grad_norm": 2.328125, |
| "learning_rate": 4.918119893235894e-06, |
| "loss": 0.2627, |
| "step": 4250 |
| }, |
| { |
| "epoch": 7.447944006999125, |
| "grad_norm": 2.390625, |
| "learning_rate": 4.880584313788245e-06, |
| "loss": 0.2505, |
| "step": 4260 |
| }, |
| { |
| "epoch": 7.465441819772528, |
| "grad_norm": 2.546875, |
| "learning_rate": 4.843245660147346e-06, |
| "loss": 0.2545, |
| "step": 4270 |
| }, |
| { |
| "epoch": 7.482939632545932, |
| "grad_norm": 2.3125, |
| "learning_rate": 4.806105133889444e-06, |
| "loss": 0.265, |
| "step": 4280 |
| }, |
| { |
| "epoch": 7.500437445319335, |
| "grad_norm": 2.046875, |
| "learning_rate": 4.7691639302149365e-06, |
| "loss": 0.2661, |
| "step": 4290 |
| }, |
| { |
| "epoch": 7.517935258092739, |
| "grad_norm": 2.203125, |
| "learning_rate": 4.732423237909929e-06, |
| "loss": 0.2628, |
| "step": 4300 |
| }, |
| { |
| "epoch": 7.535433070866142, |
| "grad_norm": 2.140625, |
| "learning_rate": 4.695884239307972e-06, |
| "loss": 0.2691, |
| "step": 4310 |
| }, |
| { |
| "epoch": 7.5529308836395455, |
| "grad_norm": 2.296875, |
| "learning_rate": 4.659548110252012e-06, |
| "loss": 0.2632, |
| "step": 4320 |
| }, |
| { |
| "epoch": 7.570428696412948, |
| "grad_norm": 2.71875, |
| "learning_rate": 4.623416020056556e-06, |
| "loss": 0.2625, |
| "step": 4330 |
| }, |
| { |
| "epoch": 7.587926509186351, |
| "grad_norm": 2.296875, |
| "learning_rate": 4.587489131470039e-06, |
| "loss": 0.2514, |
| "step": 4340 |
| }, |
| { |
| "epoch": 7.605424321959755, |
| "grad_norm": 2.21875, |
| "learning_rate": 4.551768600637407e-06, |
| "loss": 0.2686, |
| "step": 4350 |
| }, |
| { |
| "epoch": 7.622922134733159, |
| "grad_norm": 2.515625, |
| "learning_rate": 4.516255577062913e-06, |
| "loss": 0.2701, |
| "step": 4360 |
| }, |
| { |
| "epoch": 7.640419947506562, |
| "grad_norm": 2.265625, |
| "learning_rate": 4.48095120357312e-06, |
| "loss": 0.2617, |
| "step": 4370 |
| }, |
| { |
| "epoch": 7.657917760279965, |
| "grad_norm": 2.09375, |
| "learning_rate": 4.445856616280136e-06, |
| "loss": 0.2597, |
| "step": 4380 |
| }, |
| { |
| "epoch": 7.675415573053368, |
| "grad_norm": 2.140625, |
| "learning_rate": 4.410972944545041e-06, |
| "loss": 0.2695, |
| "step": 4390 |
| }, |
| { |
| "epoch": 7.692913385826771, |
| "grad_norm": 2.21875, |
| "learning_rate": 4.376301310941552e-06, |
| "loss": 0.2632, |
| "step": 4400 |
| }, |
| { |
| "epoch": 7.710411198600175, |
| "grad_norm": 2.484375, |
| "learning_rate": 4.3418428312198835e-06, |
| "loss": 0.2629, |
| "step": 4410 |
| }, |
| { |
| "epoch": 7.727909011373578, |
| "grad_norm": 2.203125, |
| "learning_rate": 4.307598614270871e-06, |
| "loss": 0.2505, |
| "step": 4420 |
| }, |
| { |
| "epoch": 7.745406824146982, |
| "grad_norm": 2.171875, |
| "learning_rate": 4.273569762090261e-06, |
| "loss": 0.2695, |
| "step": 4430 |
| }, |
| { |
| "epoch": 7.762904636920385, |
| "grad_norm": 2.296875, |
| "learning_rate": 4.239757369743248e-06, |
| "loss": 0.2636, |
| "step": 4440 |
| }, |
| { |
| "epoch": 7.7804024496937885, |
| "grad_norm": 2.5, |
| "learning_rate": 4.2061625253292484e-06, |
| "loss": 0.257, |
| "step": 4450 |
| }, |
| { |
| "epoch": 7.797900262467191, |
| "grad_norm": 2.265625, |
| "learning_rate": 4.172786309946885e-06, |
| "loss": 0.2672, |
| "step": 4460 |
| }, |
| { |
| "epoch": 7.815398075240595, |
| "grad_norm": 2.40625, |
| "learning_rate": 4.139629797659188e-06, |
| "loss": 0.2641, |
| "step": 4470 |
| }, |
| { |
| "epoch": 7.832895888013998, |
| "grad_norm": 2.1875, |
| "learning_rate": 4.106694055459023e-06, |
| "loss": 0.2654, |
| "step": 4480 |
| }, |
| { |
| "epoch": 7.850393700787402, |
| "grad_norm": 2.671875, |
| "learning_rate": 4.073980143234777e-06, |
| "loss": 0.2606, |
| "step": 4490 |
| }, |
| { |
| "epoch": 7.867891513560805, |
| "grad_norm": 2.421875, |
| "learning_rate": 4.041489113736244e-06, |
| "loss": 0.2669, |
| "step": 4500 |
| }, |
| { |
| "epoch": 7.8853893263342085, |
| "grad_norm": 2.125, |
| "learning_rate": 4.009222012540725e-06, |
| "loss": 0.274, |
| "step": 4510 |
| }, |
| { |
| "epoch": 7.902887139107611, |
| "grad_norm": 2.125, |
| "learning_rate": 3.977179878019412e-06, |
| "loss": 0.259, |
| "step": 4520 |
| }, |
| { |
| "epoch": 7.920384951881015, |
| "grad_norm": 2.53125, |
| "learning_rate": 3.9453637413039536e-06, |
| "loss": 0.2611, |
| "step": 4530 |
| }, |
| { |
| "epoch": 7.937882764654418, |
| "grad_norm": 2.40625, |
| "learning_rate": 3.913774626253279e-06, |
| "loss": 0.2541, |
| "step": 4540 |
| }, |
| { |
| "epoch": 7.955380577427822, |
| "grad_norm": 2.171875, |
| "learning_rate": 3.882413549420649e-06, |
| "loss": 0.2657, |
| "step": 4550 |
| }, |
| { |
| "epoch": 7.972878390201225, |
| "grad_norm": 2.3125, |
| "learning_rate": 3.851281520020941e-06, |
| "loss": 0.2671, |
| "step": 4560 |
| }, |
| { |
| "epoch": 7.990376202974629, |
| "grad_norm": 1.9921875, |
| "learning_rate": 3.820379539898173e-06, |
| "loss": 0.271, |
| "step": 4570 |
| } |
| ], |
| "logging_steps": 10, |
| "max_steps": 5710, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 10, |
| "save_steps": 500, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": false |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 1.2912807922064032e+19, |
| "train_batch_size": 1, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|