NoLBERT / trainer_state.json
ojhfklsjhl's picture
Upload 18 files
717dccf verified
{
"best_global_step": 88000,
"best_metric": 1.4495242834091187,
"best_model_checkpoint": "debertav3-ddp-8gpu-continue/checkpoint-88000",
"epoch": 15.0,
"eval_steps": 1000,
"global_step": 88200,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.00017006802721088434,
"grad_norm": 3.9429101943969727,
"learning_rate": 0.0,
"loss": 10.4075,
"step": 1
},
{
"epoch": 0.008503401360544218,
"grad_norm": 1.971747636795044,
"learning_rate": 1.2250000000000001e-05,
"loss": 9.8035,
"step": 50
},
{
"epoch": 0.017006802721088437,
"grad_norm": 1.3740853071212769,
"learning_rate": 2.4750000000000002e-05,
"loss": 8.7689,
"step": 100
},
{
"epoch": 0.025510204081632654,
"grad_norm": 0.9082542061805725,
"learning_rate": 3.725e-05,
"loss": 7.832,
"step": 150
},
{
"epoch": 0.034013605442176874,
"grad_norm": 1.015873670578003,
"learning_rate": 4.975e-05,
"loss": 7.1484,
"step": 200
},
{
"epoch": 0.04251700680272109,
"grad_norm": 1.0678784847259521,
"learning_rate": 6.225e-05,
"loss": 6.6754,
"step": 250
},
{
"epoch": 0.05102040816326531,
"grad_norm": 1.0898432731628418,
"learning_rate": 7.475e-05,
"loss": 6.2519,
"step": 300
},
{
"epoch": 0.05952380952380952,
"grad_norm": 0.9323605298995972,
"learning_rate": 8.724999999999999e-05,
"loss": 5.8711,
"step": 350
},
{
"epoch": 0.06802721088435375,
"grad_norm": 0.7210457921028137,
"learning_rate": 9.975000000000001e-05,
"loss": 5.5604,
"step": 400
},
{
"epoch": 0.07653061224489796,
"grad_norm": 0.7947279214859009,
"learning_rate": 0.00011225,
"loss": 5.3079,
"step": 450
},
{
"epoch": 0.08503401360544217,
"grad_norm": 0.8039622902870178,
"learning_rate": 0.00012475,
"loss": 5.1044,
"step": 500
},
{
"epoch": 0.0935374149659864,
"grad_norm": 0.7226396799087524,
"learning_rate": 0.00013725,
"loss": 4.9271,
"step": 550
},
{
"epoch": 0.10204081632653061,
"grad_norm": 0.7884791493415833,
"learning_rate": 0.00014975,
"loss": 4.7654,
"step": 600
},
{
"epoch": 0.11054421768707483,
"grad_norm": 0.7507880330085754,
"learning_rate": 0.00016225000000000001,
"loss": 4.61,
"step": 650
},
{
"epoch": 0.11904761904761904,
"grad_norm": 0.6950189471244812,
"learning_rate": 0.00017475,
"loss": 4.474,
"step": 700
},
{
"epoch": 0.12755102040816327,
"grad_norm": 0.7212897539138794,
"learning_rate": 0.00018725,
"loss": 4.33,
"step": 750
},
{
"epoch": 0.1360544217687075,
"grad_norm": 0.8659721612930298,
"learning_rate": 0.00019975,
"loss": 4.177,
"step": 800
},
{
"epoch": 0.1445578231292517,
"grad_norm": 0.7730752229690552,
"learning_rate": 0.00021225,
"loss": 3.9974,
"step": 850
},
{
"epoch": 0.15306122448979592,
"grad_norm": 0.8463068604469299,
"learning_rate": 0.00022475000000000001,
"loss": 3.814,
"step": 900
},
{
"epoch": 0.16156462585034015,
"grad_norm": 0.8561194539070129,
"learning_rate": 0.00023725,
"loss": 3.6612,
"step": 950
},
{
"epoch": 0.17006802721088435,
"grad_norm": 0.7385916709899902,
"learning_rate": 0.00024975,
"loss": 3.5319,
"step": 1000
},
{
"epoch": 0.17006802721088435,
"eval_loss": 3.6021459102630615,
"eval_runtime": 75.3501,
"eval_samples_per_second": 1245.374,
"eval_steps_per_second": 4.871,
"step": 1000
},
{
"epoch": 0.17857142857142858,
"grad_norm": 0.7785052061080933,
"learning_rate": 0.00026225,
"loss": 3.4196,
"step": 1050
},
{
"epoch": 0.1870748299319728,
"grad_norm": 0.7699462175369263,
"learning_rate": 0.00027475,
"loss": 3.3248,
"step": 1100
},
{
"epoch": 0.195578231292517,
"grad_norm": 0.6943045258522034,
"learning_rate": 0.00028725,
"loss": 3.2327,
"step": 1150
},
{
"epoch": 0.20408163265306123,
"grad_norm": 0.7425329685211182,
"learning_rate": 0.00029975000000000005,
"loss": 3.1558,
"step": 1200
},
{
"epoch": 0.21258503401360543,
"grad_norm": 0.6596449613571167,
"learning_rate": 0.00031225000000000003,
"loss": 3.0896,
"step": 1250
},
{
"epoch": 0.22108843537414966,
"grad_norm": 0.6805656552314758,
"learning_rate": 0.00032475,
"loss": 3.0295,
"step": 1300
},
{
"epoch": 0.22959183673469388,
"grad_norm": 0.6616429686546326,
"learning_rate": 0.00033725,
"loss": 2.9716,
"step": 1350
},
{
"epoch": 0.23809523809523808,
"grad_norm": 0.699910581111908,
"learning_rate": 0.00034975,
"loss": 2.9154,
"step": 1400
},
{
"epoch": 0.2465986394557823,
"grad_norm": 0.6367819905281067,
"learning_rate": 0.00036225000000000005,
"loss": 2.88,
"step": 1450
},
{
"epoch": 0.25510204081632654,
"grad_norm": 0.633556067943573,
"learning_rate": 0.00037475000000000003,
"loss": 2.8288,
"step": 1500
},
{
"epoch": 0.26360544217687076,
"grad_norm": 0.6006012558937073,
"learning_rate": 0.00038725,
"loss": 2.8017,
"step": 1550
},
{
"epoch": 0.272108843537415,
"grad_norm": 0.6720077991485596,
"learning_rate": 0.00039975,
"loss": 2.762,
"step": 1600
},
{
"epoch": 0.28061224489795916,
"grad_norm": 0.6513054966926575,
"learning_rate": 0.00041225,
"loss": 2.7228,
"step": 1650
},
{
"epoch": 0.2891156462585034,
"grad_norm": 0.6120091676712036,
"learning_rate": 0.00042475000000000005,
"loss": 2.6978,
"step": 1700
},
{
"epoch": 0.2976190476190476,
"grad_norm": 0.6610727906227112,
"learning_rate": 0.00043725000000000003,
"loss": 2.6641,
"step": 1750
},
{
"epoch": 0.30612244897959184,
"grad_norm": 0.6299336552619934,
"learning_rate": 0.00044975,
"loss": 2.6343,
"step": 1800
},
{
"epoch": 0.31462585034013607,
"grad_norm": 0.6401428580284119,
"learning_rate": 0.00046225,
"loss": 2.6212,
"step": 1850
},
{
"epoch": 0.3231292517006803,
"grad_norm": 0.6018512845039368,
"learning_rate": 0.00047475,
"loss": 2.5919,
"step": 1900
},
{
"epoch": 0.33163265306122447,
"grad_norm": 0.63248610496521,
"learning_rate": 0.00048725000000000005,
"loss": 2.5739,
"step": 1950
},
{
"epoch": 0.3401360544217687,
"grad_norm": 0.646512508392334,
"learning_rate": 0.0004997500000000001,
"loss": 2.5493,
"step": 2000
},
{
"epoch": 0.3401360544217687,
"eval_loss": 2.755571126937866,
"eval_runtime": 75.4052,
"eval_samples_per_second": 1244.463,
"eval_steps_per_second": 4.867,
"step": 2000
},
{
"epoch": 0.3486394557823129,
"grad_norm": 0.6081250309944153,
"learning_rate": 0.000499568661971831,
"loss": 2.538,
"step": 2050
},
{
"epoch": 0.35714285714285715,
"grad_norm": 0.584749698638916,
"learning_rate": 0.0004991285211267605,
"loss": 2.5185,
"step": 2100
},
{
"epoch": 0.3656462585034014,
"grad_norm": 0.6018593907356262,
"learning_rate": 0.0004986883802816902,
"loss": 2.4893,
"step": 2150
},
{
"epoch": 0.3741496598639456,
"grad_norm": 0.5787435173988342,
"learning_rate": 0.0004982482394366197,
"loss": 2.4763,
"step": 2200
},
{
"epoch": 0.3826530612244898,
"grad_norm": 0.5765504240989685,
"learning_rate": 0.0004978080985915493,
"loss": 2.4471,
"step": 2250
},
{
"epoch": 0.391156462585034,
"grad_norm": 0.539054811000824,
"learning_rate": 0.0004973679577464789,
"loss": 2.4272,
"step": 2300
},
{
"epoch": 0.39965986394557823,
"grad_norm": 0.5542412400245667,
"learning_rate": 0.0004969278169014085,
"loss": 2.4178,
"step": 2350
},
{
"epoch": 0.40816326530612246,
"grad_norm": 0.5709657073020935,
"learning_rate": 0.0004964876760563381,
"loss": 2.4105,
"step": 2400
},
{
"epoch": 0.4166666666666667,
"grad_norm": 0.5530266165733337,
"learning_rate": 0.0004960475352112676,
"loss": 2.3812,
"step": 2450
},
{
"epoch": 0.42517006802721086,
"grad_norm": 0.5575669407844543,
"learning_rate": 0.0004956073943661972,
"loss": 2.3696,
"step": 2500
},
{
"epoch": 0.4336734693877551,
"grad_norm": 0.5466365814208984,
"learning_rate": 0.0004951672535211268,
"loss": 2.36,
"step": 2550
},
{
"epoch": 0.4421768707482993,
"grad_norm": 0.5788463354110718,
"learning_rate": 0.0004947271126760563,
"loss": 2.3409,
"step": 2600
},
{
"epoch": 0.45068027210884354,
"grad_norm": 0.5266286730766296,
"learning_rate": 0.0004942869718309859,
"loss": 2.3284,
"step": 2650
},
{
"epoch": 0.45918367346938777,
"grad_norm": 0.5419022440910339,
"learning_rate": 0.0004938468309859155,
"loss": 2.3175,
"step": 2700
},
{
"epoch": 0.467687074829932,
"grad_norm": 0.5173165202140808,
"learning_rate": 0.0004934066901408451,
"loss": 2.3021,
"step": 2750
},
{
"epoch": 0.47619047619047616,
"grad_norm": 0.5213769674301147,
"learning_rate": 0.0004929665492957746,
"loss": 2.2974,
"step": 2800
},
{
"epoch": 0.4846938775510204,
"grad_norm": 0.5393424034118652,
"learning_rate": 0.0004925264084507042,
"loss": 2.2834,
"step": 2850
},
{
"epoch": 0.4931972789115646,
"grad_norm": 0.5283843874931335,
"learning_rate": 0.0004920862676056338,
"loss": 2.2894,
"step": 2900
},
{
"epoch": 0.5017006802721088,
"grad_norm": 0.5366395711898804,
"learning_rate": 0.0004916461267605634,
"loss": 2.2643,
"step": 2950
},
{
"epoch": 0.5102040816326531,
"grad_norm": 0.5475978851318359,
"learning_rate": 0.000491205985915493,
"loss": 2.2525,
"step": 3000
},
{
"epoch": 0.5102040816326531,
"eval_loss": 2.454470634460449,
"eval_runtime": 75.3482,
"eval_samples_per_second": 1245.405,
"eval_steps_per_second": 4.871,
"step": 3000
},
{
"epoch": 0.5187074829931972,
"grad_norm": 0.5384633541107178,
"learning_rate": 0.0004907658450704226,
"loss": 2.2469,
"step": 3050
},
{
"epoch": 0.5272108843537415,
"grad_norm": 0.538545548915863,
"learning_rate": 0.0004903257042253521,
"loss": 2.2378,
"step": 3100
},
{
"epoch": 0.5357142857142857,
"grad_norm": 0.5223479270935059,
"learning_rate": 0.0004898855633802817,
"loss": 2.2343,
"step": 3150
},
{
"epoch": 0.54421768707483,
"grad_norm": 0.531447172164917,
"learning_rate": 0.0004894454225352113,
"loss": 2.2201,
"step": 3200
},
{
"epoch": 0.5527210884353742,
"grad_norm": 0.5139949321746826,
"learning_rate": 0.0004890052816901409,
"loss": 2.2098,
"step": 3250
},
{
"epoch": 0.5612244897959183,
"grad_norm": 0.5428385138511658,
"learning_rate": 0.0004885651408450705,
"loss": 2.2115,
"step": 3300
},
{
"epoch": 0.5697278911564626,
"grad_norm": 0.5274862051010132,
"learning_rate": 0.000488125,
"loss": 2.1987,
"step": 3350
},
{
"epoch": 0.5782312925170068,
"grad_norm": 0.5110312104225159,
"learning_rate": 0.0004876848591549296,
"loss": 2.1843,
"step": 3400
},
{
"epoch": 0.5867346938775511,
"grad_norm": 0.49753338098526,
"learning_rate": 0.00048724471830985914,
"loss": 2.1755,
"step": 3450
},
{
"epoch": 0.5952380952380952,
"grad_norm": 0.5169926881790161,
"learning_rate": 0.00048680457746478874,
"loss": 2.1779,
"step": 3500
},
{
"epoch": 0.6037414965986394,
"grad_norm": 0.5181500911712646,
"learning_rate": 0.0004863644366197183,
"loss": 2.1658,
"step": 3550
},
{
"epoch": 0.6122448979591837,
"grad_norm": 0.5171290636062622,
"learning_rate": 0.00048592429577464793,
"loss": 2.1668,
"step": 3600
},
{
"epoch": 0.6207482993197279,
"grad_norm": 0.4980362355709076,
"learning_rate": 0.00048548415492957747,
"loss": 2.1545,
"step": 3650
},
{
"epoch": 0.6292517006802721,
"grad_norm": 0.5099490284919739,
"learning_rate": 0.00048504401408450707,
"loss": 2.1415,
"step": 3700
},
{
"epoch": 0.6377551020408163,
"grad_norm": 0.5084192752838135,
"learning_rate": 0.0004846038732394366,
"loss": 2.1417,
"step": 3750
},
{
"epoch": 0.6462585034013606,
"grad_norm": 0.48829177021980286,
"learning_rate": 0.00048416373239436626,
"loss": 2.1359,
"step": 3800
},
{
"epoch": 0.6547619047619048,
"grad_norm": 0.5111795663833618,
"learning_rate": 0.0004837235915492958,
"loss": 2.1348,
"step": 3850
},
{
"epoch": 0.6632653061224489,
"grad_norm": 0.49867865443229675,
"learning_rate": 0.00048328345070422534,
"loss": 2.1199,
"step": 3900
},
{
"epoch": 0.6717687074829932,
"grad_norm": 0.4912058711051941,
"learning_rate": 0.00048284330985915493,
"loss": 2.1169,
"step": 3950
},
{
"epoch": 0.6802721088435374,
"grad_norm": 0.5115543603897095,
"learning_rate": 0.00048240316901408453,
"loss": 2.1093,
"step": 4000
},
{
"epoch": 0.6802721088435374,
"eval_loss": 2.320960521697998,
"eval_runtime": 75.9327,
"eval_samples_per_second": 1235.818,
"eval_steps_per_second": 4.833,
"step": 4000
},
{
"epoch": 0.6887755102040817,
"grad_norm": 0.5165016651153564,
"learning_rate": 0.0004819630281690141,
"loss": 2.1103,
"step": 4050
},
{
"epoch": 0.6972789115646258,
"grad_norm": 0.49575796723365784,
"learning_rate": 0.00048152288732394367,
"loss": 2.1027,
"step": 4100
},
{
"epoch": 0.70578231292517,
"grad_norm": 0.5010895133018494,
"learning_rate": 0.0004810827464788732,
"loss": 2.0939,
"step": 4150
},
{
"epoch": 0.7142857142857143,
"grad_norm": 0.49885183572769165,
"learning_rate": 0.00048064260563380286,
"loss": 2.0951,
"step": 4200
},
{
"epoch": 0.7227891156462585,
"grad_norm": 0.47758999466896057,
"learning_rate": 0.0004802024647887324,
"loss": 2.0834,
"step": 4250
},
{
"epoch": 0.7312925170068028,
"grad_norm": 0.48634713888168335,
"learning_rate": 0.000479762323943662,
"loss": 2.0833,
"step": 4300
},
{
"epoch": 0.7397959183673469,
"grad_norm": 0.5028336644172668,
"learning_rate": 0.00047932218309859153,
"loss": 2.0804,
"step": 4350
},
{
"epoch": 0.7482993197278912,
"grad_norm": 0.4951624870300293,
"learning_rate": 0.00047888204225352113,
"loss": 2.0697,
"step": 4400
},
{
"epoch": 0.7568027210884354,
"grad_norm": 0.5072620511054993,
"learning_rate": 0.0004784419014084507,
"loss": 2.0745,
"step": 4450
},
{
"epoch": 0.7653061224489796,
"grad_norm": 0.4807872474193573,
"learning_rate": 0.0004780017605633803,
"loss": 2.0618,
"step": 4500
},
{
"epoch": 0.7738095238095238,
"grad_norm": 0.49827975034713745,
"learning_rate": 0.00047756161971830986,
"loss": 2.0648,
"step": 4550
},
{
"epoch": 0.782312925170068,
"grad_norm": 0.48257699608802795,
"learning_rate": 0.0004771214788732394,
"loss": 2.0542,
"step": 4600
},
{
"epoch": 0.7908163265306123,
"grad_norm": 0.4961060583591461,
"learning_rate": 0.00047668133802816905,
"loss": 2.0531,
"step": 4650
},
{
"epoch": 0.7993197278911565,
"grad_norm": 0.48454368114471436,
"learning_rate": 0.0004762411971830986,
"loss": 2.0497,
"step": 4700
},
{
"epoch": 0.8078231292517006,
"grad_norm": 0.48959559202194214,
"learning_rate": 0.0004758010563380282,
"loss": 2.0539,
"step": 4750
},
{
"epoch": 0.8163265306122449,
"grad_norm": 0.4989263415336609,
"learning_rate": 0.00047536091549295773,
"loss": 2.0436,
"step": 4800
},
{
"epoch": 0.8248299319727891,
"grad_norm": 0.491557776927948,
"learning_rate": 0.0004749207746478874,
"loss": 2.0347,
"step": 4850
},
{
"epoch": 0.8333333333333334,
"grad_norm": 0.4861533045768738,
"learning_rate": 0.0004744806338028169,
"loss": 2.0269,
"step": 4900
},
{
"epoch": 0.8418367346938775,
"grad_norm": 0.5007792711257935,
"learning_rate": 0.00047404049295774646,
"loss": 2.0277,
"step": 4950
},
{
"epoch": 0.8503401360544217,
"grad_norm": 0.48347318172454834,
"learning_rate": 0.00047360035211267606,
"loss": 2.0278,
"step": 5000
},
{
"epoch": 0.8503401360544217,
"eval_loss": 2.221730947494507,
"eval_runtime": 75.3978,
"eval_samples_per_second": 1244.585,
"eval_steps_per_second": 4.868,
"step": 5000
},
{
"epoch": 0.858843537414966,
"grad_norm": 0.4758061468601227,
"learning_rate": 0.00047316021126760565,
"loss": 2.0207,
"step": 5050
},
{
"epoch": 0.8673469387755102,
"grad_norm": 0.5015333890914917,
"learning_rate": 0.00047272007042253525,
"loss": 2.0213,
"step": 5100
},
{
"epoch": 0.8758503401360545,
"grad_norm": 0.46741312742233276,
"learning_rate": 0.0004722799295774648,
"loss": 2.0115,
"step": 5150
},
{
"epoch": 0.8843537414965986,
"grad_norm": 0.4716244339942932,
"learning_rate": 0.0004718397887323944,
"loss": 2.0041,
"step": 5200
},
{
"epoch": 0.8928571428571429,
"grad_norm": 0.470093309879303,
"learning_rate": 0.000471399647887324,
"loss": 2.0101,
"step": 5250
},
{
"epoch": 0.9013605442176871,
"grad_norm": 0.46333491802215576,
"learning_rate": 0.0004709595070422535,
"loss": 1.9985,
"step": 5300
},
{
"epoch": 0.9098639455782312,
"grad_norm": 0.4772166609764099,
"learning_rate": 0.0004705193661971831,
"loss": 1.9956,
"step": 5350
},
{
"epoch": 0.9183673469387755,
"grad_norm": 0.49403470754623413,
"learning_rate": 0.00047007922535211266,
"loss": 1.991,
"step": 5400
},
{
"epoch": 0.9268707482993197,
"grad_norm": 0.4855027198791504,
"learning_rate": 0.00046963908450704225,
"loss": 1.9972,
"step": 5450
},
{
"epoch": 0.935374149659864,
"grad_norm": 0.4754594564437866,
"learning_rate": 0.00046919894366197185,
"loss": 1.9902,
"step": 5500
},
{
"epoch": 0.9438775510204082,
"grad_norm": 0.4977014660835266,
"learning_rate": 0.00046875880281690144,
"loss": 1.9904,
"step": 5550
},
{
"epoch": 0.9523809523809523,
"grad_norm": 0.48129725456237793,
"learning_rate": 0.000468318661971831,
"loss": 1.9855,
"step": 5600
},
{
"epoch": 0.9608843537414966,
"grad_norm": 0.484291672706604,
"learning_rate": 0.00046787852112676053,
"loss": 1.9781,
"step": 5650
},
{
"epoch": 0.9693877551020408,
"grad_norm": 0.4759155809879303,
"learning_rate": 0.0004674383802816902,
"loss": 1.9766,
"step": 5700
},
{
"epoch": 0.9778911564625851,
"grad_norm": 0.48138466477394104,
"learning_rate": 0.0004669982394366197,
"loss": 1.9678,
"step": 5750
},
{
"epoch": 0.9863945578231292,
"grad_norm": 0.48765894770622253,
"learning_rate": 0.0004665580985915493,
"loss": 1.9746,
"step": 5800
},
{
"epoch": 0.9948979591836735,
"grad_norm": 0.47016099095344543,
"learning_rate": 0.00046611795774647885,
"loss": 1.9745,
"step": 5850
},
{
"epoch": 1.0034013605442176,
"grad_norm": 0.4722880721092224,
"learning_rate": 0.0004656778169014085,
"loss": 1.961,
"step": 5900
},
{
"epoch": 1.0119047619047619,
"grad_norm": 0.47606250643730164,
"learning_rate": 0.00046523767605633805,
"loss": 1.9548,
"step": 5950
},
{
"epoch": 1.0204081632653061,
"grad_norm": 0.4867871403694153,
"learning_rate": 0.0004647975352112676,
"loss": 1.9531,
"step": 6000
},
{
"epoch": 1.0204081632653061,
"eval_loss": 2.1581504344940186,
"eval_runtime": 75.3496,
"eval_samples_per_second": 1245.381,
"eval_steps_per_second": 4.871,
"step": 6000
},
{
"epoch": 1.0289115646258504,
"grad_norm": 0.4676497280597687,
"learning_rate": 0.0004643573943661972,
"loss": 1.9505,
"step": 6050
},
{
"epoch": 1.0374149659863945,
"grad_norm": 0.47663334012031555,
"learning_rate": 0.0004639172535211268,
"loss": 1.9509,
"step": 6100
},
{
"epoch": 1.0459183673469388,
"grad_norm": 0.46838682889938354,
"learning_rate": 0.0004634771126760564,
"loss": 1.9522,
"step": 6150
},
{
"epoch": 1.054421768707483,
"grad_norm": 0.47476112842559814,
"learning_rate": 0.0004630369718309859,
"loss": 1.9575,
"step": 6200
},
{
"epoch": 1.0629251700680271,
"grad_norm": 0.4738709330558777,
"learning_rate": 0.0004625968309859155,
"loss": 1.9535,
"step": 6250
},
{
"epoch": 1.0714285714285714,
"grad_norm": 0.46784505248069763,
"learning_rate": 0.0004621566901408451,
"loss": 1.95,
"step": 6300
},
{
"epoch": 1.0799319727891157,
"grad_norm": 0.4729381501674652,
"learning_rate": 0.00046171654929577465,
"loss": 1.9382,
"step": 6350
},
{
"epoch": 1.08843537414966,
"grad_norm": 0.48842012882232666,
"learning_rate": 0.00046127640845070424,
"loss": 1.9383,
"step": 6400
},
{
"epoch": 1.096938775510204,
"grad_norm": 0.47705018520355225,
"learning_rate": 0.0004608362676056338,
"loss": 1.9298,
"step": 6450
},
{
"epoch": 1.1054421768707483,
"grad_norm": 0.47480863332748413,
"learning_rate": 0.00046039612676056343,
"loss": 1.9356,
"step": 6500
},
{
"epoch": 1.1139455782312926,
"grad_norm": 0.4836580455303192,
"learning_rate": 0.000459955985915493,
"loss": 1.936,
"step": 6550
},
{
"epoch": 1.1224489795918366,
"grad_norm": 0.47330114245414734,
"learning_rate": 0.00045951584507042257,
"loss": 1.9217,
"step": 6600
},
{
"epoch": 1.130952380952381,
"grad_norm": 0.45560622215270996,
"learning_rate": 0.0004590757042253521,
"loss": 1.9278,
"step": 6650
},
{
"epoch": 1.1394557823129252,
"grad_norm": 0.48356184363365173,
"learning_rate": 0.00045863556338028165,
"loss": 1.9204,
"step": 6700
},
{
"epoch": 1.1479591836734695,
"grad_norm": 0.48046737909317017,
"learning_rate": 0.0004581954225352113,
"loss": 1.9185,
"step": 6750
},
{
"epoch": 1.1564625850340136,
"grad_norm": 0.47801148891448975,
"learning_rate": 0.00045775528169014084,
"loss": 1.9084,
"step": 6800
},
{
"epoch": 1.1649659863945578,
"grad_norm": 0.46705126762390137,
"learning_rate": 0.00045731514084507044,
"loss": 1.9138,
"step": 6850
},
{
"epoch": 1.1734693877551021,
"grad_norm": 0.47566279768943787,
"learning_rate": 0.000456875,
"loss": 1.9139,
"step": 6900
},
{
"epoch": 1.1819727891156462,
"grad_norm": 0.4746539294719696,
"learning_rate": 0.00045643485915492963,
"loss": 1.916,
"step": 6950
},
{
"epoch": 1.1904761904761905,
"grad_norm": 0.49360892176628113,
"learning_rate": 0.00045599471830985917,
"loss": 1.9086,
"step": 7000
},
{
"epoch": 1.1904761904761905,
"eval_loss": 2.1105546951293945,
"eval_runtime": 75.3645,
"eval_samples_per_second": 1245.136,
"eval_steps_per_second": 4.87,
"step": 7000
},
{
"epoch": 1.1989795918367347,
"grad_norm": 0.49067452549934387,
"learning_rate": 0.0004555545774647887,
"loss": 1.9044,
"step": 7050
},
{
"epoch": 1.2074829931972788,
"grad_norm": 0.48024600744247437,
"learning_rate": 0.0004551144366197183,
"loss": 1.9143,
"step": 7100
},
{
"epoch": 1.215986394557823,
"grad_norm": 0.4815312922000885,
"learning_rate": 0.0004546742957746479,
"loss": 1.9038,
"step": 7150
},
{
"epoch": 1.2244897959183674,
"grad_norm": 0.47568100690841675,
"learning_rate": 0.0004542341549295775,
"loss": 1.9079,
"step": 7200
},
{
"epoch": 1.2329931972789117,
"grad_norm": 0.47568491101264954,
"learning_rate": 0.00045379401408450704,
"loss": 1.9024,
"step": 7250
},
{
"epoch": 1.2414965986394557,
"grad_norm": 0.46364548802375793,
"learning_rate": 0.00045335387323943663,
"loss": 1.9008,
"step": 7300
},
{
"epoch": 1.25,
"grad_norm": 0.4672023355960846,
"learning_rate": 0.00045291373239436623,
"loss": 1.9022,
"step": 7350
},
{
"epoch": 1.2585034013605443,
"grad_norm": 0.4671889841556549,
"learning_rate": 0.00045247359154929577,
"loss": 1.894,
"step": 7400
},
{
"epoch": 1.2670068027210886,
"grad_norm": 0.4657513201236725,
"learning_rate": 0.00045203345070422537,
"loss": 1.8978,
"step": 7450
},
{
"epoch": 1.2755102040816326,
"grad_norm": 0.4698520004749298,
"learning_rate": 0.0004515933098591549,
"loss": 1.8945,
"step": 7500
},
{
"epoch": 1.284013605442177,
"grad_norm": 0.4804818332195282,
"learning_rate": 0.00045115316901408456,
"loss": 1.8921,
"step": 7550
},
{
"epoch": 1.2925170068027212,
"grad_norm": 0.4709530174732208,
"learning_rate": 0.0004507130281690141,
"loss": 1.889,
"step": 7600
},
{
"epoch": 1.3010204081632653,
"grad_norm": 0.47395414113998413,
"learning_rate": 0.0004502728873239437,
"loss": 1.8837,
"step": 7650
},
{
"epoch": 1.3095238095238095,
"grad_norm": 0.4659099876880646,
"learning_rate": 0.00044983274647887323,
"loss": 1.8836,
"step": 7700
},
{
"epoch": 1.3180272108843538,
"grad_norm": 0.4623669683933258,
"learning_rate": 0.00044939260563380283,
"loss": 1.8823,
"step": 7750
},
{
"epoch": 1.3265306122448979,
"grad_norm": 0.4751649498939514,
"learning_rate": 0.0004489524647887324,
"loss": 1.8703,
"step": 7800
},
{
"epoch": 1.3350340136054422,
"grad_norm": 0.4874952435493469,
"learning_rate": 0.00044851232394366197,
"loss": 1.8765,
"step": 7850
},
{
"epoch": 1.3435374149659864,
"grad_norm": 0.48088109493255615,
"learning_rate": 0.00044807218309859156,
"loss": 1.8787,
"step": 7900
},
{
"epoch": 1.3520408163265305,
"grad_norm": 0.48666709661483765,
"learning_rate": 0.0004476320422535211,
"loss": 1.8723,
"step": 7950
},
{
"epoch": 1.3605442176870748,
"grad_norm": 0.5058099031448364,
"learning_rate": 0.00044719190140845075,
"loss": 1.8682,
"step": 8000
},
{
"epoch": 1.3605442176870748,
"eval_loss": 2.066969156265259,
"eval_runtime": 75.2958,
"eval_samples_per_second": 1246.271,
"eval_steps_per_second": 4.874,
"step": 8000
},
{
"epoch": 1.369047619047619,
"grad_norm": 0.47028714418411255,
"learning_rate": 0.0004467517605633803,
"loss": 1.8619,
"step": 8050
},
{
"epoch": 1.3775510204081631,
"grad_norm": 0.468432754278183,
"learning_rate": 0.00044631161971830983,
"loss": 1.8669,
"step": 8100
},
{
"epoch": 1.3860544217687074,
"grad_norm": 0.46689048409461975,
"learning_rate": 0.00044587147887323943,
"loss": 1.8657,
"step": 8150
},
{
"epoch": 1.3945578231292517,
"grad_norm": 0.464141309261322,
"learning_rate": 0.000445431338028169,
"loss": 1.863,
"step": 8200
},
{
"epoch": 1.403061224489796,
"grad_norm": 0.46267494559288025,
"learning_rate": 0.0004449911971830986,
"loss": 1.8696,
"step": 8250
},
{
"epoch": 1.4115646258503403,
"grad_norm": 0.4653218984603882,
"learning_rate": 0.00044455105633802816,
"loss": 1.8554,
"step": 8300
},
{
"epoch": 1.4200680272108843,
"grad_norm": 0.4844844937324524,
"learning_rate": 0.00044411091549295776,
"loss": 1.8525,
"step": 8350
},
{
"epoch": 1.4285714285714286,
"grad_norm": 0.47869589924812317,
"learning_rate": 0.00044367077464788735,
"loss": 1.8584,
"step": 8400
},
{
"epoch": 1.4370748299319729,
"grad_norm": 0.4788654148578644,
"learning_rate": 0.0004432306338028169,
"loss": 1.8562,
"step": 8450
},
{
"epoch": 1.445578231292517,
"grad_norm": 0.4887782335281372,
"learning_rate": 0.0004427904929577465,
"loss": 1.8583,
"step": 8500
},
{
"epoch": 1.4540816326530612,
"grad_norm": 0.48299601674079895,
"learning_rate": 0.00044235035211267603,
"loss": 1.8557,
"step": 8550
},
{
"epoch": 1.4625850340136055,
"grad_norm": 0.4597221910953522,
"learning_rate": 0.0004419102112676057,
"loss": 1.847,
"step": 8600
},
{
"epoch": 1.4710884353741496,
"grad_norm": 0.4786689877510071,
"learning_rate": 0.0004414700704225352,
"loss": 1.8393,
"step": 8650
},
{
"epoch": 1.4795918367346939,
"grad_norm": 0.46319809556007385,
"learning_rate": 0.0004410299295774648,
"loss": 1.8468,
"step": 8700
},
{
"epoch": 1.4880952380952381,
"grad_norm": 0.46916770935058594,
"learning_rate": 0.00044058978873239436,
"loss": 1.8414,
"step": 8750
},
{
"epoch": 1.4965986394557822,
"grad_norm": 0.4812500774860382,
"learning_rate": 0.00044014964788732395,
"loss": 1.8496,
"step": 8800
},
{
"epoch": 1.5051020408163265,
"grad_norm": 0.47240397334098816,
"learning_rate": 0.00043970950704225355,
"loss": 1.8415,
"step": 8850
},
{
"epoch": 1.5136054421768708,
"grad_norm": 0.4675607979297638,
"learning_rate": 0.0004392693661971831,
"loss": 1.8395,
"step": 8900
},
{
"epoch": 1.5221088435374148,
"grad_norm": 0.4568563401699066,
"learning_rate": 0.0004388380281690141,
"loss": 1.8349,
"step": 8950
},
{
"epoch": 1.5306122448979593,
"grad_norm": 0.4698484539985657,
"learning_rate": 0.00043839788732394363,
"loss": 1.8388,
"step": 9000
},
{
"epoch": 1.5306122448979593,
"eval_loss": 2.031731128692627,
"eval_runtime": 75.3373,
"eval_samples_per_second": 1245.584,
"eval_steps_per_second": 4.871,
"step": 9000
},
{
"epoch": 1.5391156462585034,
"grad_norm": 0.4721793532371521,
"learning_rate": 0.0004379577464788733,
"loss": 1.8368,
"step": 9050
},
{
"epoch": 1.5476190476190477,
"grad_norm": 0.48082154989242554,
"learning_rate": 0.0004375176056338028,
"loss": 1.8367,
"step": 9100
},
{
"epoch": 1.556122448979592,
"grad_norm": 0.47875961661338806,
"learning_rate": 0.00043707746478873237,
"loss": 1.834,
"step": 9150
},
{
"epoch": 1.564625850340136,
"grad_norm": 0.4505126178264618,
"learning_rate": 0.00043663732394366196,
"loss": 1.8293,
"step": 9200
},
{
"epoch": 1.5731292517006803,
"grad_norm": 0.4831013083457947,
"learning_rate": 0.00043619718309859156,
"loss": 1.8325,
"step": 9250
},
{
"epoch": 1.5816326530612246,
"grad_norm": 0.46912670135498047,
"learning_rate": 0.00043575704225352115,
"loss": 1.8201,
"step": 9300
},
{
"epoch": 1.5901360544217686,
"grad_norm": 0.4775020480155945,
"learning_rate": 0.0004353169014084507,
"loss": 1.8324,
"step": 9350
},
{
"epoch": 1.598639455782313,
"grad_norm": 0.4624606668949127,
"learning_rate": 0.0004348767605633803,
"loss": 1.8328,
"step": 9400
},
{
"epoch": 1.6071428571428572,
"grad_norm": 0.47817811369895935,
"learning_rate": 0.0004344366197183099,
"loss": 1.8265,
"step": 9450
},
{
"epoch": 1.6156462585034013,
"grad_norm": 0.4682134985923767,
"learning_rate": 0.0004339964788732395,
"loss": 1.8193,
"step": 9500
},
{
"epoch": 1.6241496598639455,
"grad_norm": 0.4736348092556,
"learning_rate": 0.000433556338028169,
"loss": 1.8144,
"step": 9550
},
{
"epoch": 1.6326530612244898,
"grad_norm": 0.47171303629875183,
"learning_rate": 0.00043311619718309856,
"loss": 1.8156,
"step": 9600
},
{
"epoch": 1.641156462585034,
"grad_norm": 0.4707530736923218,
"learning_rate": 0.0004326760563380282,
"loss": 1.8146,
"step": 9650
},
{
"epoch": 1.6496598639455784,
"grad_norm": 0.47999903559684753,
"learning_rate": 0.00043223591549295775,
"loss": 1.822,
"step": 9700
},
{
"epoch": 1.6581632653061225,
"grad_norm": 0.4823855757713318,
"learning_rate": 0.00043179577464788735,
"loss": 1.8161,
"step": 9750
},
{
"epoch": 1.6666666666666665,
"grad_norm": 0.47148072719573975,
"learning_rate": 0.0004313556338028169,
"loss": 1.8117,
"step": 9800
},
{
"epoch": 1.675170068027211,
"grad_norm": 0.46029239892959595,
"learning_rate": 0.00043091549295774654,
"loss": 1.8132,
"step": 9850
},
{
"epoch": 1.683673469387755,
"grad_norm": 0.47228437662124634,
"learning_rate": 0.0004304753521126761,
"loss": 1.816,
"step": 9900
},
{
"epoch": 1.6921768707482994,
"grad_norm": 0.4653479754924774,
"learning_rate": 0.0004300352112676056,
"loss": 1.8084,
"step": 9950
},
{
"epoch": 1.7006802721088436,
"grad_norm": 0.4696131646633148,
"learning_rate": 0.0004295950704225352,
"loss": 1.8072,
"step": 10000
},
{
"epoch": 1.7006802721088436,
"eval_loss": 1.9987208843231201,
"eval_runtime": 75.2953,
"eval_samples_per_second": 1246.28,
"eval_steps_per_second": 4.874,
"step": 10000
},
{
"epoch": 1.7091836734693877,
"grad_norm": 0.4656902253627777,
"learning_rate": 0.00042915492957746476,
"loss": 1.8049,
"step": 10050
},
{
"epoch": 1.717687074829932,
"grad_norm": 0.47671329975128174,
"learning_rate": 0.0004287147887323944,
"loss": 1.7934,
"step": 10100
},
{
"epoch": 1.7261904761904763,
"grad_norm": 0.46021008491516113,
"learning_rate": 0.00042827464788732395,
"loss": 1.8059,
"step": 10150
},
{
"epoch": 1.7346938775510203,
"grad_norm": 0.46965909004211426,
"learning_rate": 0.00042783450704225354,
"loss": 1.7958,
"step": 10200
},
{
"epoch": 1.7431972789115646,
"grad_norm": 0.4655004143714905,
"learning_rate": 0.0004273943661971831,
"loss": 1.7989,
"step": 10250
},
{
"epoch": 1.751700680272109,
"grad_norm": 0.48593828082084656,
"learning_rate": 0.0004269542253521127,
"loss": 1.7904,
"step": 10300
},
{
"epoch": 1.760204081632653,
"grad_norm": 0.4698497951030731,
"learning_rate": 0.0004265140845070423,
"loss": 1.8005,
"step": 10350
},
{
"epoch": 1.7687074829931972,
"grad_norm": 0.4775938391685486,
"learning_rate": 0.0004260739436619718,
"loss": 1.7992,
"step": 10400
},
{
"epoch": 1.7772108843537415,
"grad_norm": 0.45465579628944397,
"learning_rate": 0.0004256338028169014,
"loss": 1.7973,
"step": 10450
},
{
"epoch": 1.7857142857142856,
"grad_norm": 0.4619791507720947,
"learning_rate": 0.000425193661971831,
"loss": 1.7913,
"step": 10500
},
{
"epoch": 1.79421768707483,
"grad_norm": 0.4738638699054718,
"learning_rate": 0.0004247535211267606,
"loss": 1.7922,
"step": 10550
},
{
"epoch": 1.8027210884353742,
"grad_norm": 0.46855342388153076,
"learning_rate": 0.00042431338028169015,
"loss": 1.7894,
"step": 10600
},
{
"epoch": 1.8112244897959182,
"grad_norm": 0.48508965969085693,
"learning_rate": 0.0004238732394366197,
"loss": 1.7815,
"step": 10650
},
{
"epoch": 1.8197278911564627,
"grad_norm": 0.47747498750686646,
"learning_rate": 0.00042343309859154934,
"loss": 1.7907,
"step": 10700
},
{
"epoch": 1.8282312925170068,
"grad_norm": 0.47017601132392883,
"learning_rate": 0.0004229929577464789,
"loss": 1.7855,
"step": 10750
},
{
"epoch": 1.836734693877551,
"grad_norm": 0.4648725092411041,
"learning_rate": 0.00042255281690140847,
"loss": 1.7878,
"step": 10800
},
{
"epoch": 1.8452380952380953,
"grad_norm": 0.4595974385738373,
"learning_rate": 0.000422112676056338,
"loss": 1.7808,
"step": 10850
},
{
"epoch": 1.8537414965986394,
"grad_norm": 0.5028242468833923,
"learning_rate": 0.00042167253521126766,
"loss": 1.7838,
"step": 10900
},
{
"epoch": 1.8622448979591837,
"grad_norm": 0.4698718190193176,
"learning_rate": 0.0004212323943661972,
"loss": 1.7838,
"step": 10950
},
{
"epoch": 1.870748299319728,
"grad_norm": 0.46658000349998474,
"learning_rate": 0.00042079225352112675,
"loss": 1.7884,
"step": 11000
},
{
"epoch": 1.870748299319728,
"eval_loss": 1.9652538299560547,
"eval_runtime": 75.3728,
"eval_samples_per_second": 1244.998,
"eval_steps_per_second": 4.869,
"step": 11000
},
{
"epoch": 1.879251700680272,
"grad_norm": 0.4587540924549103,
"learning_rate": 0.00042035211267605634,
"loss": 1.7795,
"step": 11050
},
{
"epoch": 1.8877551020408163,
"grad_norm": 0.48107513785362244,
"learning_rate": 0.00041991197183098594,
"loss": 1.7843,
"step": 11100
},
{
"epoch": 1.8962585034013606,
"grad_norm": 0.468262255191803,
"learning_rate": 0.00041947183098591553,
"loss": 1.7831,
"step": 11150
},
{
"epoch": 1.9047619047619047,
"grad_norm": 0.4687144458293915,
"learning_rate": 0.0004190316901408451,
"loss": 1.7761,
"step": 11200
},
{
"epoch": 1.913265306122449,
"grad_norm": 0.48110878467559814,
"learning_rate": 0.00041859154929577467,
"loss": 1.7658,
"step": 11250
},
{
"epoch": 1.9217687074829932,
"grad_norm": 0.47303488850593567,
"learning_rate": 0.0004181514084507042,
"loss": 1.7733,
"step": 11300
},
{
"epoch": 1.9302721088435373,
"grad_norm": 0.48464885354042053,
"learning_rate": 0.0004177112676056338,
"loss": 1.765,
"step": 11350
},
{
"epoch": 1.9387755102040818,
"grad_norm": 0.4908495247364044,
"learning_rate": 0.0004172799295774648,
"loss": 1.7662,
"step": 11400
},
{
"epoch": 1.9472789115646258,
"grad_norm": 0.47335630655288696,
"learning_rate": 0.00041683978873239435,
"loss": 1.7689,
"step": 11450
},
{
"epoch": 1.95578231292517,
"grad_norm": 0.4794527292251587,
"learning_rate": 0.00041639964788732395,
"loss": 1.7659,
"step": 11500
},
{
"epoch": 1.9642857142857144,
"grad_norm": 0.47013363242149353,
"learning_rate": 0.00041595950704225354,
"loss": 1.7683,
"step": 11550
},
{
"epoch": 1.9727891156462585,
"grad_norm": 0.4743632674217224,
"learning_rate": 0.00041551936619718314,
"loss": 1.7674,
"step": 11600
},
{
"epoch": 1.9812925170068028,
"grad_norm": 0.46521708369255066,
"learning_rate": 0.0004150792253521127,
"loss": 1.7686,
"step": 11650
},
{
"epoch": 1.989795918367347,
"grad_norm": 0.48240792751312256,
"learning_rate": 0.0004146390845070422,
"loss": 1.7688,
"step": 11700
},
{
"epoch": 1.998299319727891,
"grad_norm": 0.453975111246109,
"learning_rate": 0.00041419894366197187,
"loss": 1.7658,
"step": 11750
},
{
"epoch": 2.006802721088435,
"grad_norm": 0.45248886942863464,
"learning_rate": 0.0004137588028169014,
"loss": 1.7543,
"step": 11800
},
{
"epoch": 2.0153061224489797,
"grad_norm": 0.47196707129478455,
"learning_rate": 0.0004133274647887324,
"loss": 1.751,
"step": 11850
},
{
"epoch": 2.0238095238095237,
"grad_norm": 0.46333810687065125,
"learning_rate": 0.00041288732394366195,
"loss": 1.7544,
"step": 11900
},
{
"epoch": 2.0323129251700682,
"grad_norm": 0.46540361642837524,
"learning_rate": 0.0004124471830985916,
"loss": 1.7586,
"step": 11950
},
{
"epoch": 2.0408163265306123,
"grad_norm": 0.45927169919013977,
"learning_rate": 0.00041200704225352114,
"loss": 1.753,
"step": 12000
},
{
"epoch": 2.0408163265306123,
"eval_loss": 1.948580265045166,
"eval_runtime": 75.3406,
"eval_samples_per_second": 1245.53,
"eval_steps_per_second": 4.871,
"step": 12000
},
{
"epoch": 2.0493197278911564,
"grad_norm": 0.4919711947441101,
"learning_rate": 0.0004115669014084507,
"loss": 1.7501,
"step": 12050
},
{
"epoch": 2.057823129251701,
"grad_norm": 0.47275787591934204,
"learning_rate": 0.0004111267605633803,
"loss": 1.7579,
"step": 12100
},
{
"epoch": 2.066326530612245,
"grad_norm": 0.4734228551387787,
"learning_rate": 0.0004106866197183099,
"loss": 1.7445,
"step": 12150
},
{
"epoch": 2.074829931972789,
"grad_norm": 0.4777526259422302,
"learning_rate": 0.00041024647887323947,
"loss": 1.7558,
"step": 12200
},
{
"epoch": 2.0833333333333335,
"grad_norm": 0.4509522020816803,
"learning_rate": 0.000409806338028169,
"loss": 1.754,
"step": 12250
},
{
"epoch": 2.0918367346938775,
"grad_norm": 0.4808884859085083,
"learning_rate": 0.0004093661971830986,
"loss": 1.7487,
"step": 12300
},
{
"epoch": 2.1003401360544216,
"grad_norm": 0.46674636006355286,
"learning_rate": 0.00040892605633802815,
"loss": 1.7428,
"step": 12350
},
{
"epoch": 2.108843537414966,
"grad_norm": 0.4768598973751068,
"learning_rate": 0.00040848591549295775,
"loss": 1.7372,
"step": 12400
},
{
"epoch": 2.11734693877551,
"grad_norm": 0.4708414673805237,
"learning_rate": 0.00040804577464788734,
"loss": 1.7458,
"step": 12450
},
{
"epoch": 2.1258503401360542,
"grad_norm": 0.47017523646354675,
"learning_rate": 0.0004076056338028169,
"loss": 1.7497,
"step": 12500
},
{
"epoch": 2.1343537414965987,
"grad_norm": 0.4945257306098938,
"learning_rate": 0.0004071654929577465,
"loss": 1.7451,
"step": 12550
},
{
"epoch": 2.142857142857143,
"grad_norm": 0.4588756859302521,
"learning_rate": 0.00040672535211267607,
"loss": 1.7393,
"step": 12600
},
{
"epoch": 2.1513605442176873,
"grad_norm": 0.49352967739105225,
"learning_rate": 0.00040628521126760567,
"loss": 1.7366,
"step": 12650
},
{
"epoch": 2.1598639455782314,
"grad_norm": 0.47481125593185425,
"learning_rate": 0.0004058450704225352,
"loss": 1.7373,
"step": 12700
},
{
"epoch": 2.1683673469387754,
"grad_norm": 0.48250919580459595,
"learning_rate": 0.00040540492957746475,
"loss": 1.7409,
"step": 12750
},
{
"epoch": 2.17687074829932,
"grad_norm": 0.4601617157459259,
"learning_rate": 0.0004049647887323944,
"loss": 1.7413,
"step": 12800
},
{
"epoch": 2.185374149659864,
"grad_norm": 0.47254276275634766,
"learning_rate": 0.00040452464788732394,
"loss": 1.7495,
"step": 12850
},
{
"epoch": 2.193877551020408,
"grad_norm": 0.48082679510116577,
"learning_rate": 0.00040408450704225354,
"loss": 1.7429,
"step": 12900
},
{
"epoch": 2.2023809523809526,
"grad_norm": 0.46352913975715637,
"learning_rate": 0.0004036443661971831,
"loss": 1.7401,
"step": 12950
},
{
"epoch": 2.2108843537414966,
"grad_norm": 0.4566120207309723,
"learning_rate": 0.00040321302816901414,
"loss": 1.7337,
"step": 13000
},
{
"epoch": 2.2108843537414966,
"eval_loss": 1.921891689300537,
"eval_runtime": 75.3847,
"eval_samples_per_second": 1244.802,
"eval_steps_per_second": 4.868,
"step": 13000
},
{
"epoch": 2.2193877551020407,
"grad_norm": 0.5124084949493408,
"learning_rate": 0.0004027728873239437,
"loss": 1.741,
"step": 13050
},
{
"epoch": 2.227891156462585,
"grad_norm": 0.4604070782661438,
"learning_rate": 0.00040233274647887327,
"loss": 1.7372,
"step": 13100
},
{
"epoch": 2.2363945578231292,
"grad_norm": 0.49075254797935486,
"learning_rate": 0.0004019014084507042,
"loss": 1.7411,
"step": 13150
},
{
"epoch": 2.2448979591836733,
"grad_norm": 0.4861091077327728,
"learning_rate": 0.00040146126760563376,
"loss": 1.7418,
"step": 13200
},
{
"epoch": 2.253401360544218,
"grad_norm": 0.47723260521888733,
"learning_rate": 0.0004010211267605634,
"loss": 1.7375,
"step": 13250
},
{
"epoch": 2.261904761904762,
"grad_norm": 0.4777429401874542,
"learning_rate": 0.00040058098591549295,
"loss": 1.7305,
"step": 13300
},
{
"epoch": 2.270408163265306,
"grad_norm": 0.4744473993778229,
"learning_rate": 0.00040014084507042255,
"loss": 1.7291,
"step": 13350
},
{
"epoch": 2.2789115646258504,
"grad_norm": 0.46367913484573364,
"learning_rate": 0.0003997007042253521,
"loss": 1.7305,
"step": 13400
},
{
"epoch": 2.2874149659863945,
"grad_norm": 0.4830113649368286,
"learning_rate": 0.00039926056338028174,
"loss": 1.7298,
"step": 13450
},
{
"epoch": 2.295918367346939,
"grad_norm": 0.48178356885910034,
"learning_rate": 0.0003988204225352113,
"loss": 1.715,
"step": 13500
},
{
"epoch": 2.304421768707483,
"grad_norm": 0.46279290318489075,
"learning_rate": 0.0003983802816901408,
"loss": 1.7265,
"step": 13550
},
{
"epoch": 2.312925170068027,
"grad_norm": 0.4898684620857239,
"learning_rate": 0.0003979401408450704,
"loss": 1.7257,
"step": 13600
},
{
"epoch": 2.3214285714285716,
"grad_norm": 0.47215622663497925,
"learning_rate": 0.0003975,
"loss": 1.7291,
"step": 13650
},
{
"epoch": 2.3299319727891157,
"grad_norm": 0.4575498104095459,
"learning_rate": 0.0003970598591549296,
"loss": 1.7165,
"step": 13700
},
{
"epoch": 2.3384353741496597,
"grad_norm": 0.4656940996646881,
"learning_rate": 0.00039661971830985915,
"loss": 1.7183,
"step": 13750
},
{
"epoch": 2.3469387755102042,
"grad_norm": 0.4670725464820862,
"learning_rate": 0.00039617957746478874,
"loss": 1.7217,
"step": 13800
},
{
"epoch": 2.3554421768707483,
"grad_norm": 0.47170403599739075,
"learning_rate": 0.00039573943661971834,
"loss": 1.7255,
"step": 13850
},
{
"epoch": 2.3639455782312924,
"grad_norm": 0.49639126658439636,
"learning_rate": 0.0003952992957746479,
"loss": 1.7246,
"step": 13900
},
{
"epoch": 2.372448979591837,
"grad_norm": 0.4941042959690094,
"learning_rate": 0.0003948591549295775,
"loss": 1.7183,
"step": 13950
},
{
"epoch": 2.380952380952381,
"grad_norm": 0.4859507381916046,
"learning_rate": 0.000394419014084507,
"loss": 1.7239,
"step": 14000
},
{
"epoch": 2.380952380952381,
"eval_loss": 1.9032775163650513,
"eval_runtime": 75.3175,
"eval_samples_per_second": 1245.912,
"eval_steps_per_second": 4.873,
"step": 14000
},
{
"epoch": 2.389455782312925,
"grad_norm": 0.48601582646369934,
"learning_rate": 0.00039397887323943667,
"loss": 1.721,
"step": 14050
},
{
"epoch": 2.3979591836734695,
"grad_norm": 0.4685114920139313,
"learning_rate": 0.0003935387323943662,
"loss": 1.7155,
"step": 14100
},
{
"epoch": 2.4064625850340136,
"grad_norm": 0.44943785667419434,
"learning_rate": 0.0003930985915492958,
"loss": 1.7154,
"step": 14150
},
{
"epoch": 2.4149659863945576,
"grad_norm": 0.48964107036590576,
"learning_rate": 0.00039265845070422535,
"loss": 1.7123,
"step": 14200
},
{
"epoch": 2.423469387755102,
"grad_norm": 0.457769513130188,
"learning_rate": 0.00039221830985915494,
"loss": 1.7084,
"step": 14250
},
{
"epoch": 2.431972789115646,
"grad_norm": 0.5480780005455017,
"learning_rate": 0.00039177816901408454,
"loss": 1.7129,
"step": 14300
},
{
"epoch": 2.4404761904761907,
"grad_norm": 0.48546379804611206,
"learning_rate": 0.0003913380281690141,
"loss": 1.7078,
"step": 14350
},
{
"epoch": 2.4489795918367347,
"grad_norm": 0.46493494510650635,
"learning_rate": 0.00039089788732394367,
"loss": 1.7104,
"step": 14400
},
{
"epoch": 2.457482993197279,
"grad_norm": 0.4606137275695801,
"learning_rate": 0.0003904577464788732,
"loss": 1.7045,
"step": 14450
},
{
"epoch": 2.4659863945578233,
"grad_norm": 0.46123144030570984,
"learning_rate": 0.00039001760563380286,
"loss": 1.7047,
"step": 14500
},
{
"epoch": 2.4744897959183674,
"grad_norm": 0.4724046289920807,
"learning_rate": 0.0003895774647887324,
"loss": 1.711,
"step": 14550
},
{
"epoch": 2.4829931972789114,
"grad_norm": 0.47036615014076233,
"learning_rate": 0.00038913732394366195,
"loss": 1.7077,
"step": 14600
},
{
"epoch": 2.491496598639456,
"grad_norm": 0.465732216835022,
"learning_rate": 0.00038869718309859154,
"loss": 1.7041,
"step": 14650
},
{
"epoch": 2.5,
"grad_norm": 0.4794946610927582,
"learning_rate": 0.00038825704225352114,
"loss": 1.7037,
"step": 14700
},
{
"epoch": 2.508503401360544,
"grad_norm": 0.47816696763038635,
"learning_rate": 0.00038781690140845073,
"loss": 1.7056,
"step": 14750
},
{
"epoch": 2.5170068027210886,
"grad_norm": 0.47812485694885254,
"learning_rate": 0.0003873767605633803,
"loss": 1.7063,
"step": 14800
},
{
"epoch": 2.5255102040816326,
"grad_norm": 0.4651651084423065,
"learning_rate": 0.00038693661971830987,
"loss": 1.693,
"step": 14850
},
{
"epoch": 2.534013605442177,
"grad_norm": 0.48417726159095764,
"learning_rate": 0.00038649647887323946,
"loss": 1.697,
"step": 14900
},
{
"epoch": 2.542517006802721,
"grad_norm": 0.4502563178539276,
"learning_rate": 0.000386056338028169,
"loss": 1.703,
"step": 14950
},
{
"epoch": 2.5510204081632653,
"grad_norm": 0.45644158124923706,
"learning_rate": 0.0003856161971830986,
"loss": 1.6959,
"step": 15000
},
{
"epoch": 2.5510204081632653,
"eval_loss": 1.8780722618103027,
"eval_runtime": 75.398,
"eval_samples_per_second": 1244.582,
"eval_steps_per_second": 4.868,
"step": 15000
},
{
"epoch": 2.5595238095238093,
"grad_norm": 0.4715486764907837,
"learning_rate": 0.00038517605633802814,
"loss": 1.7015,
"step": 15050
},
{
"epoch": 2.568027210884354,
"grad_norm": 0.44938182830810547,
"learning_rate": 0.0003847359154929578,
"loss": 1.6954,
"step": 15100
},
{
"epoch": 2.576530612244898,
"grad_norm": 0.4778023660182953,
"learning_rate": 0.00038429577464788733,
"loss": 1.6973,
"step": 15150
},
{
"epoch": 2.5850340136054424,
"grad_norm": 0.4602490961551666,
"learning_rate": 0.00038385563380281693,
"loss": 1.7041,
"step": 15200
},
{
"epoch": 2.5935374149659864,
"grad_norm": 0.4905562400817871,
"learning_rate": 0.0003834242957746479,
"loss": 1.6929,
"step": 15250
},
{
"epoch": 2.6020408163265305,
"grad_norm": 0.47374477982521057,
"learning_rate": 0.00038298415492957747,
"loss": 1.6963,
"step": 15300
},
{
"epoch": 2.6105442176870746,
"grad_norm": 0.45640283823013306,
"learning_rate": 0.00038254401408450707,
"loss": 1.6941,
"step": 15350
},
{
"epoch": 2.619047619047619,
"grad_norm": 0.46338459849357605,
"learning_rate": 0.0003821038732394366,
"loss": 1.6898,
"step": 15400
},
{
"epoch": 2.627551020408163,
"grad_norm": 0.458769291639328,
"learning_rate": 0.0003816637323943662,
"loss": 1.6897,
"step": 15450
},
{
"epoch": 2.6360544217687076,
"grad_norm": 0.44556647539138794,
"learning_rate": 0.00038122359154929575,
"loss": 1.6918,
"step": 15500
},
{
"epoch": 2.6445578231292517,
"grad_norm": 0.48705628514289856,
"learning_rate": 0.0003807834507042254,
"loss": 1.6948,
"step": 15550
},
{
"epoch": 2.6530612244897958,
"grad_norm": 0.46732062101364136,
"learning_rate": 0.00038034330985915494,
"loss": 1.6884,
"step": 15600
},
{
"epoch": 2.6615646258503403,
"grad_norm": 0.44862720370292664,
"learning_rate": 0.0003799031690140845,
"loss": 1.6921,
"step": 15650
},
{
"epoch": 2.6700680272108843,
"grad_norm": 0.4507545828819275,
"learning_rate": 0.0003794630281690141,
"loss": 1.6841,
"step": 15700
},
{
"epoch": 2.678571428571429,
"grad_norm": 0.4554910659790039,
"learning_rate": 0.00037902288732394367,
"loss": 1.6898,
"step": 15750
},
{
"epoch": 2.687074829931973,
"grad_norm": 0.47400516271591187,
"learning_rate": 0.00037858274647887326,
"loss": 1.6875,
"step": 15800
},
{
"epoch": 2.695578231292517,
"grad_norm": 0.46053117513656616,
"learning_rate": 0.0003781426056338028,
"loss": 1.6791,
"step": 15850
},
{
"epoch": 2.704081632653061,
"grad_norm": 0.4834235608577728,
"learning_rate": 0.0003777024647887324,
"loss": 1.6883,
"step": 15900
},
{
"epoch": 2.7125850340136055,
"grad_norm": 0.648454487323761,
"learning_rate": 0.000377262323943662,
"loss": 1.686,
"step": 15950
},
{
"epoch": 2.7210884353741496,
"grad_norm": 0.46427688002586365,
"learning_rate": 0.0003768221830985916,
"loss": 1.6771,
"step": 16000
},
{
"epoch": 2.7210884353741496,
"eval_loss": 1.8668512105941772,
"eval_runtime": 75.2617,
"eval_samples_per_second": 1246.836,
"eval_steps_per_second": 4.876,
"step": 16000
},
{
"epoch": 2.729591836734694,
"grad_norm": 0.4748222827911377,
"learning_rate": 0.00037638204225352113,
"loss": 1.6826,
"step": 16050
},
{
"epoch": 2.738095238095238,
"grad_norm": 0.4745228588581085,
"learning_rate": 0.0003759419014084507,
"loss": 1.6837,
"step": 16100
},
{
"epoch": 2.746598639455782,
"grad_norm": 0.4836510419845581,
"learning_rate": 0.0003755017605633803,
"loss": 1.6803,
"step": 16150
},
{
"epoch": 2.7551020408163263,
"grad_norm": 0.48129406571388245,
"learning_rate": 0.00037506161971830986,
"loss": 1.6804,
"step": 16200
},
{
"epoch": 2.7636054421768708,
"grad_norm": 0.45295995473861694,
"learning_rate": 0.00037462147887323946,
"loss": 1.6763,
"step": 16250
},
{
"epoch": 2.772108843537415,
"grad_norm": 0.45548656582832336,
"learning_rate": 0.000374181338028169,
"loss": 1.6793,
"step": 16300
},
{
"epoch": 2.7806122448979593,
"grad_norm": 0.46522030234336853,
"learning_rate": 0.00037374119718309865,
"loss": 1.6792,
"step": 16350
},
{
"epoch": 2.7891156462585034,
"grad_norm": 0.46659553050994873,
"learning_rate": 0.0003733010563380282,
"loss": 1.6705,
"step": 16400
},
{
"epoch": 2.7976190476190474,
"grad_norm": 0.4808714985847473,
"learning_rate": 0.00037286091549295773,
"loss": 1.6751,
"step": 16450
},
{
"epoch": 2.806122448979592,
"grad_norm": 0.46006572246551514,
"learning_rate": 0.00037242077464788733,
"loss": 1.6708,
"step": 16500
},
{
"epoch": 2.814625850340136,
"grad_norm": 0.4552723467350006,
"learning_rate": 0.0003719806338028169,
"loss": 1.6765,
"step": 16550
},
{
"epoch": 2.8231292517006805,
"grad_norm": 0.4788212180137634,
"learning_rate": 0.0003715404929577465,
"loss": 1.6774,
"step": 16600
},
{
"epoch": 2.8316326530612246,
"grad_norm": 0.4770471453666687,
"learning_rate": 0.00037110035211267606,
"loss": 1.6703,
"step": 16650
},
{
"epoch": 2.8401360544217686,
"grad_norm": 0.4665678143501282,
"learning_rate": 0.00037066021126760566,
"loss": 1.6754,
"step": 16700
},
{
"epoch": 2.8486394557823127,
"grad_norm": 0.46680372953414917,
"learning_rate": 0.0003702200704225352,
"loss": 1.67,
"step": 16750
},
{
"epoch": 2.857142857142857,
"grad_norm": 0.4522228240966797,
"learning_rate": 0.0003697799295774648,
"loss": 1.6718,
"step": 16800
},
{
"epoch": 2.8656462585034013,
"grad_norm": 0.4507662057876587,
"learning_rate": 0.0003693397887323944,
"loss": 1.6713,
"step": 16850
},
{
"epoch": 2.8741496598639458,
"grad_norm": 0.45078015327453613,
"learning_rate": 0.00036889964788732393,
"loss": 1.6641,
"step": 16900
},
{
"epoch": 2.88265306122449,
"grad_norm": 0.46813851594924927,
"learning_rate": 0.0003684595070422535,
"loss": 1.6684,
"step": 16950
},
{
"epoch": 2.891156462585034,
"grad_norm": 0.4676934480667114,
"learning_rate": 0.0003680193661971831,
"loss": 1.6678,
"step": 17000
},
{
"epoch": 2.891156462585034,
"eval_loss": 1.8514577150344849,
"eval_runtime": 75.2869,
"eval_samples_per_second": 1246.419,
"eval_steps_per_second": 4.875,
"step": 17000
},
{
"epoch": 2.8996598639455784,
"grad_norm": 0.4356513023376465,
"learning_rate": 0.0003675792253521127,
"loss": 1.6586,
"step": 17050
},
{
"epoch": 2.9081632653061225,
"grad_norm": 0.44688156247138977,
"learning_rate": 0.00036713908450704226,
"loss": 1.6596,
"step": 17100
},
{
"epoch": 2.9166666666666665,
"grad_norm": 0.46431487798690796,
"learning_rate": 0.0003666989436619718,
"loss": 1.6682,
"step": 17150
},
{
"epoch": 2.925170068027211,
"grad_norm": 0.44890233874320984,
"learning_rate": 0.00036625880281690145,
"loss": 1.6594,
"step": 17200
},
{
"epoch": 2.933673469387755,
"grad_norm": 0.45430788397789,
"learning_rate": 0.000365818661971831,
"loss": 1.6617,
"step": 17250
},
{
"epoch": 2.942176870748299,
"grad_norm": 0.4674071967601776,
"learning_rate": 0.0003653785211267606,
"loss": 1.6583,
"step": 17300
},
{
"epoch": 2.9506802721088436,
"grad_norm": 1.025103211402893,
"learning_rate": 0.0003649383802816901,
"loss": 1.6637,
"step": 17350
},
{
"epoch": 2.9591836734693877,
"grad_norm": 0.5043914318084717,
"learning_rate": 0.0003645070422535212,
"loss": 1.6673,
"step": 17400
},
{
"epoch": 2.967687074829932,
"grad_norm": 0.4789212942123413,
"learning_rate": 0.0003640669014084507,
"loss": 1.665,
"step": 17450
},
{
"epoch": 2.9761904761904763,
"grad_norm": 0.458146870136261,
"learning_rate": 0.00036362676056338027,
"loss": 1.6659,
"step": 17500
},
{
"epoch": 2.9846938775510203,
"grad_norm": 0.45948123931884766,
"learning_rate": 0.00036319542253521127,
"loss": 1.6559,
"step": 17550
},
{
"epoch": 2.9931972789115644,
"grad_norm": 0.44673070311546326,
"learning_rate": 0.0003627552816901408,
"loss": 1.6596,
"step": 17600
},
{
"epoch": 3.001700680272109,
"grad_norm": 0.46580860018730164,
"learning_rate": 0.00036231514084507046,
"loss": 1.6646,
"step": 17650
},
{
"epoch": 3.010204081632653,
"grad_norm": 0.48107218742370605,
"learning_rate": 0.000361875,
"loss": 1.6516,
"step": 17700
},
{
"epoch": 3.0187074829931975,
"grad_norm": 0.4705383777618408,
"learning_rate": 0.0003614348591549296,
"loss": 1.6506,
"step": 17750
},
{
"epoch": 3.0272108843537415,
"grad_norm": 0.4649271070957184,
"learning_rate": 0.00036099471830985914,
"loss": 1.6523,
"step": 17800
},
{
"epoch": 3.0357142857142856,
"grad_norm": 0.4697044789791107,
"learning_rate": 0.00036055457746478873,
"loss": 1.6556,
"step": 17850
},
{
"epoch": 3.04421768707483,
"grad_norm": 0.441441148519516,
"learning_rate": 0.00036011443661971833,
"loss": 1.6483,
"step": 17900
},
{
"epoch": 3.052721088435374,
"grad_norm": 0.4512791037559509,
"learning_rate": 0.00035967429577464787,
"loss": 1.655,
"step": 17950
},
{
"epoch": 3.061224489795918,
"grad_norm": 0.47064149379730225,
"learning_rate": 0.00035923415492957746,
"loss": 1.6493,
"step": 18000
},
{
"epoch": 3.061224489795918,
"eval_loss": 1.8354525566101074,
"eval_runtime": 75.3083,
"eval_samples_per_second": 1246.064,
"eval_steps_per_second": 4.873,
"step": 18000
},
{
"epoch": 3.0697278911564627,
"grad_norm": 0.4441589415073395,
"learning_rate": 0.00035879401408450706,
"loss": 1.6469,
"step": 18050
},
{
"epoch": 3.078231292517007,
"grad_norm": 0.46877768635749817,
"learning_rate": 0.00035835387323943666,
"loss": 1.6449,
"step": 18100
},
{
"epoch": 3.086734693877551,
"grad_norm": 0.9494704008102417,
"learning_rate": 0.0003579137323943662,
"loss": 1.6426,
"step": 18150
},
{
"epoch": 3.0952380952380953,
"grad_norm": 0.4520435631275177,
"learning_rate": 0.00035747359154929574,
"loss": 1.6521,
"step": 18200
},
{
"epoch": 3.1037414965986394,
"grad_norm": 0.46680590510368347,
"learning_rate": 0.0003570334507042254,
"loss": 1.655,
"step": 18250
},
{
"epoch": 3.1122448979591835,
"grad_norm": 0.47364234924316406,
"learning_rate": 0.00035659330985915493,
"loss": 1.6404,
"step": 18300
},
{
"epoch": 3.120748299319728,
"grad_norm": 0.47683462500572205,
"learning_rate": 0.0003561531690140845,
"loss": 1.6432,
"step": 18350
},
{
"epoch": 3.129251700680272,
"grad_norm": 0.4633096158504486,
"learning_rate": 0.00035571302816901407,
"loss": 1.6459,
"step": 18400
},
{
"epoch": 3.137755102040816,
"grad_norm": 0.4648244380950928,
"learning_rate": 0.0003552728873239437,
"loss": 1.6392,
"step": 18450
},
{
"epoch": 3.1462585034013606,
"grad_norm": 0.46435174345970154,
"learning_rate": 0.00035483274647887326,
"loss": 1.6396,
"step": 18500
},
{
"epoch": 3.1547619047619047,
"grad_norm": 0.45293596386909485,
"learning_rate": 0.0003543926056338028,
"loss": 1.6413,
"step": 18550
},
{
"epoch": 3.163265306122449,
"grad_norm": 0.46072807908058167,
"learning_rate": 0.0003539524647887324,
"loss": 1.6379,
"step": 18600
},
{
"epoch": 3.171768707482993,
"grad_norm": 0.45269516110420227,
"learning_rate": 0.000353512323943662,
"loss": 1.6409,
"step": 18650
},
{
"epoch": 3.1802721088435373,
"grad_norm": 0.4461529850959778,
"learning_rate": 0.0003530721830985916,
"loss": 1.6373,
"step": 18700
},
{
"epoch": 3.188775510204082,
"grad_norm": 0.45222413539886475,
"learning_rate": 0.0003526320422535211,
"loss": 1.6317,
"step": 18750
},
{
"epoch": 3.197278911564626,
"grad_norm": 0.44513842463493347,
"learning_rate": 0.0003521919014084507,
"loss": 1.6365,
"step": 18800
},
{
"epoch": 3.20578231292517,
"grad_norm": 0.4705728590488434,
"learning_rate": 0.00035175176056338026,
"loss": 1.6434,
"step": 18850
},
{
"epoch": 3.2142857142857144,
"grad_norm": 0.47750329971313477,
"learning_rate": 0.00035131161971830986,
"loss": 1.6337,
"step": 18900
},
{
"epoch": 3.2227891156462585,
"grad_norm": 0.4608128070831299,
"learning_rate": 0.00035087147887323945,
"loss": 1.642,
"step": 18950
},
{
"epoch": 3.2312925170068025,
"grad_norm": 0.46269285678863525,
"learning_rate": 0.000350431338028169,
"loss": 1.6326,
"step": 19000
},
{
"epoch": 3.2312925170068025,
"eval_loss": 1.8231135606765747,
"eval_runtime": 75.3579,
"eval_samples_per_second": 1245.245,
"eval_steps_per_second": 4.87,
"step": 19000
},
{
"epoch": 3.239795918367347,
"grad_norm": 0.44600149989128113,
"learning_rate": 0.0003499911971830986,
"loss": 1.6372,
"step": 19050
},
{
"epoch": 3.248299319727891,
"grad_norm": 0.45678427815437317,
"learning_rate": 0.0003495510563380282,
"loss": 1.6348,
"step": 19100
},
{
"epoch": 3.2568027210884356,
"grad_norm": 0.45453667640686035,
"learning_rate": 0.0003491109154929578,
"loss": 1.6422,
"step": 19150
},
{
"epoch": 3.2653061224489797,
"grad_norm": 0.4784705936908722,
"learning_rate": 0.0003486707746478873,
"loss": 1.6382,
"step": 19200
},
{
"epoch": 3.2738095238095237,
"grad_norm": 0.4635883569717407,
"learning_rate": 0.00034823063380281686,
"loss": 1.6388,
"step": 19250
},
{
"epoch": 3.282312925170068,
"grad_norm": 0.45983946323394775,
"learning_rate": 0.0003477904929577465,
"loss": 1.6408,
"step": 19300
},
{
"epoch": 3.2908163265306123,
"grad_norm": 0.4456377923488617,
"learning_rate": 0.00034735035211267605,
"loss": 1.6388,
"step": 19350
},
{
"epoch": 3.2993197278911564,
"grad_norm": 0.4579452872276306,
"learning_rate": 0.00034691021126760565,
"loss": 1.6283,
"step": 19400
},
{
"epoch": 3.307823129251701,
"grad_norm": 0.43990224599838257,
"learning_rate": 0.0003464700704225352,
"loss": 1.6327,
"step": 19450
},
{
"epoch": 3.316326530612245,
"grad_norm": 0.4531456530094147,
"learning_rate": 0.00034602992957746484,
"loss": 1.6302,
"step": 19500
},
{
"epoch": 3.324829931972789,
"grad_norm": 0.4910786747932434,
"learning_rate": 0.0003455897887323944,
"loss": 1.6275,
"step": 19550
},
{
"epoch": 3.3333333333333335,
"grad_norm": 0.46175333857536316,
"learning_rate": 0.0003451496478873239,
"loss": 1.6349,
"step": 19600
},
{
"epoch": 3.3418367346938775,
"grad_norm": 0.4583819508552551,
"learning_rate": 0.0003447095070422535,
"loss": 1.6279,
"step": 19650
},
{
"epoch": 3.3503401360544216,
"grad_norm": 0.4596096873283386,
"learning_rate": 0.0003442693661971831,
"loss": 1.6203,
"step": 19700
},
{
"epoch": 3.358843537414966,
"grad_norm": 0.476151704788208,
"learning_rate": 0.0003438292253521127,
"loss": 1.63,
"step": 19750
},
{
"epoch": 3.36734693877551,
"grad_norm": 0.45185205340385437,
"learning_rate": 0.00034338908450704225,
"loss": 1.6228,
"step": 19800
},
{
"epoch": 3.3758503401360542,
"grad_norm": 0.4486692249774933,
"learning_rate": 0.00034294894366197184,
"loss": 1.623,
"step": 19850
},
{
"epoch": 3.3843537414965987,
"grad_norm": 0.4323517084121704,
"learning_rate": 0.0003425088028169014,
"loss": 1.6203,
"step": 19900
},
{
"epoch": 3.392857142857143,
"grad_norm": 0.4673045575618744,
"learning_rate": 0.00034206866197183103,
"loss": 1.6229,
"step": 19950
},
{
"epoch": 3.4013605442176873,
"grad_norm": 0.5606664419174194,
"learning_rate": 0.0003416285211267606,
"loss": 1.6248,
"step": 20000
},
{
"epoch": 3.4013605442176873,
"eval_loss": 1.8098527193069458,
"eval_runtime": 75.3357,
"eval_samples_per_second": 1245.611,
"eval_steps_per_second": 4.872,
"step": 20000
},
{
"epoch": 3.4098639455782314,
"grad_norm": 0.4416612386703491,
"learning_rate": 0.0003411883802816901,
"loss": 1.6244,
"step": 20050
},
{
"epoch": 3.4183673469387754,
"grad_norm": 0.46314534544944763,
"learning_rate": 0.0003407482394366197,
"loss": 1.619,
"step": 20100
},
{
"epoch": 3.4268707482993195,
"grad_norm": 0.46182000637054443,
"learning_rate": 0.0003403080985915493,
"loss": 1.6198,
"step": 20150
},
{
"epoch": 3.435374149659864,
"grad_norm": 0.46425795555114746,
"learning_rate": 0.0003398679577464789,
"loss": 1.6198,
"step": 20200
},
{
"epoch": 3.443877551020408,
"grad_norm": 0.4472617208957672,
"learning_rate": 0.00033942781690140844,
"loss": 1.6208,
"step": 20250
},
{
"epoch": 3.4523809523809526,
"grad_norm": 0.4484054446220398,
"learning_rate": 0.000338987676056338,
"loss": 1.617,
"step": 20300
},
{
"epoch": 3.4608843537414966,
"grad_norm": 0.46471381187438965,
"learning_rate": 0.00033855633802816904,
"loss": 1.6235,
"step": 20350
},
{
"epoch": 3.4693877551020407,
"grad_norm": 0.471675306558609,
"learning_rate": 0.0003381161971830986,
"loss": 1.6211,
"step": 20400
},
{
"epoch": 3.477891156462585,
"grad_norm": 0.4538419246673584,
"learning_rate": 0.0003376760563380282,
"loss": 1.6159,
"step": 20450
},
{
"epoch": 3.4863945578231292,
"grad_norm": 0.4715994894504547,
"learning_rate": 0.0003372359154929577,
"loss": 1.6155,
"step": 20500
},
{
"epoch": 3.4948979591836733,
"grad_norm": 0.44585472345352173,
"learning_rate": 0.00033679577464788737,
"loss": 1.6169,
"step": 20550
},
{
"epoch": 3.503401360544218,
"grad_norm": 0.4732266962528229,
"learning_rate": 0.0003363556338028169,
"loss": 1.6175,
"step": 20600
},
{
"epoch": 3.511904761904762,
"grad_norm": 0.4438916742801666,
"learning_rate": 0.0003359154929577465,
"loss": 1.6179,
"step": 20650
},
{
"epoch": 3.520408163265306,
"grad_norm": 0.46406659483909607,
"learning_rate": 0.00033547535211267605,
"loss": 1.6169,
"step": 20700
},
{
"epoch": 3.5289115646258504,
"grad_norm": 0.438435435295105,
"learning_rate": 0.00033503521126760564,
"loss": 1.6155,
"step": 20750
},
{
"epoch": 3.5374149659863945,
"grad_norm": 0.4469165503978729,
"learning_rate": 0.00033459507042253524,
"loss": 1.6124,
"step": 20800
},
{
"epoch": 3.545918367346939,
"grad_norm": 0.4506942927837372,
"learning_rate": 0.0003341549295774648,
"loss": 1.6128,
"step": 20850
},
{
"epoch": 3.554421768707483,
"grad_norm": 0.4475231468677521,
"learning_rate": 0.0003337147887323944,
"loss": 1.6118,
"step": 20900
},
{
"epoch": 3.562925170068027,
"grad_norm": 0.45778709650039673,
"learning_rate": 0.0003332746478873239,
"loss": 1.6098,
"step": 20950
},
{
"epoch": 3.571428571428571,
"grad_norm": 0.44058629870414734,
"learning_rate": 0.00033283450704225357,
"loss": 1.6153,
"step": 21000
},
{
"epoch": 3.571428571428571,
"eval_loss": 1.791085124015808,
"eval_runtime": 75.3212,
"eval_samples_per_second": 1245.85,
"eval_steps_per_second": 4.872,
"step": 21000
},
{
"epoch": 3.5799319727891157,
"grad_norm": 0.44585561752319336,
"learning_rate": 0.0003323943661971831,
"loss": 1.6093,
"step": 21050
},
{
"epoch": 3.5884353741496597,
"grad_norm": 0.4633313715457916,
"learning_rate": 0.00033195422535211265,
"loss": 1.608,
"step": 21100
},
{
"epoch": 3.5969387755102042,
"grad_norm": 0.4513951539993286,
"learning_rate": 0.00033151408450704224,
"loss": 1.6066,
"step": 21150
},
{
"epoch": 3.6054421768707483,
"grad_norm": 0.44720685482025146,
"learning_rate": 0.00033107394366197184,
"loss": 1.6104,
"step": 21200
},
{
"epoch": 3.6139455782312924,
"grad_norm": 0.4755708873271942,
"learning_rate": 0.00033063380281690144,
"loss": 1.6138,
"step": 21250
},
{
"epoch": 3.622448979591837,
"grad_norm": 0.46023187041282654,
"learning_rate": 0.000330193661971831,
"loss": 1.6087,
"step": 21300
},
{
"epoch": 3.630952380952381,
"grad_norm": 0.45014435052871704,
"learning_rate": 0.00032975352112676057,
"loss": 1.606,
"step": 21350
},
{
"epoch": 3.6394557823129254,
"grad_norm": 0.440222829580307,
"learning_rate": 0.00032931338028169017,
"loss": 1.606,
"step": 21400
},
{
"epoch": 3.6479591836734695,
"grad_norm": 0.46397948265075684,
"learning_rate": 0.0003288732394366197,
"loss": 1.6085,
"step": 21450
},
{
"epoch": 3.6564625850340136,
"grad_norm": 0.5181043744087219,
"learning_rate": 0.0003284330985915493,
"loss": 1.6098,
"step": 21500
},
{
"epoch": 3.6649659863945576,
"grad_norm": 0.4351687431335449,
"learning_rate": 0.00032799295774647884,
"loss": 1.6087,
"step": 21550
},
{
"epoch": 3.673469387755102,
"grad_norm": 0.4317842721939087,
"learning_rate": 0.0003275528169014085,
"loss": 1.6068,
"step": 21600
},
{
"epoch": 3.681972789115646,
"grad_norm": 0.5115000009536743,
"learning_rate": 0.00032711267605633804,
"loss": 1.5992,
"step": 21650
},
{
"epoch": 3.6904761904761907,
"grad_norm": 0.4375057816505432,
"learning_rate": 0.00032667253521126763,
"loss": 1.6026,
"step": 21700
},
{
"epoch": 3.6989795918367347,
"grad_norm": 0.4488801956176758,
"learning_rate": 0.00032623239436619717,
"loss": 1.6057,
"step": 21750
},
{
"epoch": 3.707482993197279,
"grad_norm": 0.46254992485046387,
"learning_rate": 0.00032579225352112677,
"loss": 1.6031,
"step": 21800
},
{
"epoch": 3.715986394557823,
"grad_norm": 0.43931296467781067,
"learning_rate": 0.00032535211267605636,
"loss": 1.6035,
"step": 21850
},
{
"epoch": 3.7244897959183674,
"grad_norm": 0.4594497084617615,
"learning_rate": 0.0003249119718309859,
"loss": 1.6051,
"step": 21900
},
{
"epoch": 3.7329931972789114,
"grad_norm": 0.45316246151924133,
"learning_rate": 0.0003244718309859155,
"loss": 1.5937,
"step": 21950
},
{
"epoch": 3.741496598639456,
"grad_norm": 0.4681023359298706,
"learning_rate": 0.0003240316901408451,
"loss": 1.6015,
"step": 22000
},
{
"epoch": 3.741496598639456,
"eval_loss": 1.7835361957550049,
"eval_runtime": 75.3121,
"eval_samples_per_second": 1246.002,
"eval_steps_per_second": 4.873,
"step": 22000
},
{
"epoch": 3.75,
"grad_norm": 0.5095068216323853,
"learning_rate": 0.0003235915492957747,
"loss": 1.6031,
"step": 22050
},
{
"epoch": 3.758503401360544,
"grad_norm": 0.43521663546562195,
"learning_rate": 0.00032315140845070423,
"loss": 1.6038,
"step": 22100
},
{
"epoch": 3.7670068027210886,
"grad_norm": 0.4796842932701111,
"learning_rate": 0.00032271126760563377,
"loss": 1.6061,
"step": 22150
},
{
"epoch": 3.7755102040816326,
"grad_norm": 0.45405638217926025,
"learning_rate": 0.00032227112676056337,
"loss": 1.5938,
"step": 22200
},
{
"epoch": 3.784013605442177,
"grad_norm": 0.46786293387413025,
"learning_rate": 0.00032183098591549296,
"loss": 1.6031,
"step": 22250
},
{
"epoch": 3.792517006802721,
"grad_norm": 0.43882298469543457,
"learning_rate": 0.00032139084507042256,
"loss": 1.6018,
"step": 22300
},
{
"epoch": 3.8010204081632653,
"grad_norm": 0.45325416326522827,
"learning_rate": 0.0003209507042253521,
"loss": 1.5905,
"step": 22350
},
{
"epoch": 3.8095238095238093,
"grad_norm": 0.4605984687805176,
"learning_rate": 0.0003205105633802817,
"loss": 1.596,
"step": 22400
},
{
"epoch": 3.818027210884354,
"grad_norm": 0.5207810401916504,
"learning_rate": 0.0003200704225352113,
"loss": 1.6048,
"step": 22450
},
{
"epoch": 3.826530612244898,
"grad_norm": 0.45419174432754517,
"learning_rate": 0.00031963028169014083,
"loss": 1.5972,
"step": 22500
},
{
"epoch": 3.8350340136054424,
"grad_norm": 0.46196499466896057,
"learning_rate": 0.00031919014084507043,
"loss": 1.5952,
"step": 22550
},
{
"epoch": 3.8435374149659864,
"grad_norm": 0.48807084560394287,
"learning_rate": 0.00031874999999999997,
"loss": 1.6006,
"step": 22600
},
{
"epoch": 3.8520408163265305,
"grad_norm": 0.44178763031959534,
"learning_rate": 0.000318318661971831,
"loss": 1.6006,
"step": 22650
},
{
"epoch": 3.8605442176870746,
"grad_norm": 0.4443242847919464,
"learning_rate": 0.00031787852112676057,
"loss": 1.5875,
"step": 22700
},
{
"epoch": 3.869047619047619,
"grad_norm": 0.45137760043144226,
"learning_rate": 0.00031743838028169016,
"loss": 1.5874,
"step": 22750
},
{
"epoch": 3.877551020408163,
"grad_norm": 0.44187167286872864,
"learning_rate": 0.0003169982394366197,
"loss": 1.5935,
"step": 22800
},
{
"epoch": 3.8860544217687076,
"grad_norm": 0.46749913692474365,
"learning_rate": 0.0003165580985915493,
"loss": 1.5918,
"step": 22850
},
{
"epoch": 3.8945578231292517,
"grad_norm": 0.45157769322395325,
"learning_rate": 0.0003161179577464789,
"loss": 1.5965,
"step": 22900
},
{
"epoch": 3.9030612244897958,
"grad_norm": 0.6856850981712341,
"learning_rate": 0.00031567781690140844,
"loss": 1.5929,
"step": 22950
},
{
"epoch": 3.9115646258503403,
"grad_norm": 0.45894768834114075,
"learning_rate": 0.00031523767605633803,
"loss": 1.5915,
"step": 23000
},
{
"epoch": 3.9115646258503403,
"eval_loss": 1.7726092338562012,
"eval_runtime": 75.3291,
"eval_samples_per_second": 1245.72,
"eval_steps_per_second": 4.872,
"step": 23000
},
{
"epoch": 3.9200680272108843,
"grad_norm": 0.44938963651657104,
"learning_rate": 0.0003147975352112676,
"loss": 1.5913,
"step": 23050
},
{
"epoch": 3.928571428571429,
"grad_norm": 0.4493829309940338,
"learning_rate": 0.0003143573943661972,
"loss": 1.5999,
"step": 23100
},
{
"epoch": 3.937074829931973,
"grad_norm": 0.44146212935447693,
"learning_rate": 0.00031391725352112676,
"loss": 1.5937,
"step": 23150
},
{
"epoch": 3.945578231292517,
"grad_norm": 0.4436129927635193,
"learning_rate": 0.0003134771126760563,
"loss": 1.5937,
"step": 23200
},
{
"epoch": 3.954081632653061,
"grad_norm": 0.4842548668384552,
"learning_rate": 0.0003130369718309859,
"loss": 1.5921,
"step": 23250
},
{
"epoch": 3.9625850340136055,
"grad_norm": 0.4605764150619507,
"learning_rate": 0.0003125968309859155,
"loss": 1.585,
"step": 23300
},
{
"epoch": 3.9710884353741496,
"grad_norm": 0.446140319108963,
"learning_rate": 0.0003121566901408451,
"loss": 1.5773,
"step": 23350
},
{
"epoch": 3.979591836734694,
"grad_norm": 0.4490399658679962,
"learning_rate": 0.00031171654929577463,
"loss": 1.5862,
"step": 23400
},
{
"epoch": 3.988095238095238,
"grad_norm": 0.43997710943222046,
"learning_rate": 0.00031128521126760564,
"loss": 1.5873,
"step": 23450
},
{
"epoch": 3.996598639455782,
"grad_norm": 0.4559916853904724,
"learning_rate": 0.00031084507042253523,
"loss": 1.5936,
"step": 23500
},
{
"epoch": 4.005102040816326,
"grad_norm": 0.46398648619651794,
"learning_rate": 0.0003104049295774648,
"loss": 1.5861,
"step": 23550
},
{
"epoch": 4.01360544217687,
"grad_norm": 0.4367043972015381,
"learning_rate": 0.00030996478873239437,
"loss": 1.5814,
"step": 23600
},
{
"epoch": 4.022108843537415,
"grad_norm": 0.4262286126613617,
"learning_rate": 0.0003095246478873239,
"loss": 1.5831,
"step": 23650
},
{
"epoch": 4.030612244897959,
"grad_norm": 1.598874568939209,
"learning_rate": 0.00030908450704225356,
"loss": 1.5863,
"step": 23700
},
{
"epoch": 4.039115646258503,
"grad_norm": 0.44894129037857056,
"learning_rate": 0.0003086443661971831,
"loss": 1.5852,
"step": 23750
},
{
"epoch": 4.0476190476190474,
"grad_norm": 0.4517632722854614,
"learning_rate": 0.0003082042253521127,
"loss": 1.581,
"step": 23800
},
{
"epoch": 4.0561224489795915,
"grad_norm": 0.4445061683654785,
"learning_rate": 0.00030776408450704224,
"loss": 1.5777,
"step": 23850
},
{
"epoch": 4.0646258503401365,
"grad_norm": 0.4282425343990326,
"learning_rate": 0.0003073239436619719,
"loss": 1.579,
"step": 23900
},
{
"epoch": 4.0731292517006805,
"grad_norm": 0.48686137795448303,
"learning_rate": 0.0003068838028169014,
"loss": 1.5766,
"step": 23950
},
{
"epoch": 4.081632653061225,
"grad_norm": 0.4572753310203552,
"learning_rate": 0.00030644366197183097,
"loss": 1.5751,
"step": 24000
},
{
"epoch": 4.081632653061225,
"eval_loss": 1.7599811553955078,
"eval_runtime": 75.3428,
"eval_samples_per_second": 1245.494,
"eval_steps_per_second": 4.871,
"step": 24000
},
{
"epoch": 4.090136054421769,
"grad_norm": 0.44744428992271423,
"learning_rate": 0.00030600352112676056,
"loss": 1.573,
"step": 24050
},
{
"epoch": 4.098639455782313,
"grad_norm": 0.45439615845680237,
"learning_rate": 0.00030556338028169016,
"loss": 1.5717,
"step": 24100
},
{
"epoch": 4.107142857142857,
"grad_norm": 0.46508198976516724,
"learning_rate": 0.00030512323943661975,
"loss": 1.5716,
"step": 24150
},
{
"epoch": 4.115646258503402,
"grad_norm": 0.45242032408714294,
"learning_rate": 0.0003046830985915493,
"loss": 1.5692,
"step": 24200
},
{
"epoch": 4.124149659863946,
"grad_norm": 0.441568523645401,
"learning_rate": 0.0003042429577464789,
"loss": 1.569,
"step": 24250
},
{
"epoch": 4.13265306122449,
"grad_norm": 0.43780016899108887,
"learning_rate": 0.00030380281690140843,
"loss": 1.5751,
"step": 24300
},
{
"epoch": 4.141156462585034,
"grad_norm": 0.47256138920783997,
"learning_rate": 0.00030336267605633803,
"loss": 1.5711,
"step": 24350
},
{
"epoch": 4.149659863945578,
"grad_norm": 0.4236377477645874,
"learning_rate": 0.0003029225352112676,
"loss": 1.5737,
"step": 24400
},
{
"epoch": 4.158163265306122,
"grad_norm": 0.4356014132499695,
"learning_rate": 0.00030248239436619716,
"loss": 1.5683,
"step": 24450
},
{
"epoch": 4.166666666666667,
"grad_norm": 0.44677263498306274,
"learning_rate": 0.00030204225352112676,
"loss": 1.5661,
"step": 24500
},
{
"epoch": 4.175170068027211,
"grad_norm": 0.45358529686927795,
"learning_rate": 0.00030160211267605635,
"loss": 1.5718,
"step": 24550
},
{
"epoch": 4.183673469387755,
"grad_norm": 0.46313029527664185,
"learning_rate": 0.00030116197183098595,
"loss": 1.5726,
"step": 24600
},
{
"epoch": 4.192176870748299,
"grad_norm": 0.4524216055870056,
"learning_rate": 0.0003007218309859155,
"loss": 1.5659,
"step": 24650
},
{
"epoch": 4.200680272108843,
"grad_norm": 0.46442392468452454,
"learning_rate": 0.00030028169014084503,
"loss": 1.57,
"step": 24700
},
{
"epoch": 4.209183673469388,
"grad_norm": 0.43893301486968994,
"learning_rate": 0.0002998415492957747,
"loss": 1.5646,
"step": 24750
},
{
"epoch": 4.217687074829932,
"grad_norm": 0.4545600116252899,
"learning_rate": 0.0002994014084507042,
"loss": 1.5656,
"step": 24800
},
{
"epoch": 4.226190476190476,
"grad_norm": 0.43572577834129333,
"learning_rate": 0.0002989612676056338,
"loss": 1.5619,
"step": 24850
},
{
"epoch": 4.23469387755102,
"grad_norm": 0.4401288330554962,
"learning_rate": 0.00029852112676056336,
"loss": 1.5626,
"step": 24900
},
{
"epoch": 4.243197278911564,
"grad_norm": 0.4619799554347992,
"learning_rate": 0.000298080985915493,
"loss": 1.5679,
"step": 24950
},
{
"epoch": 4.2517006802721085,
"grad_norm": 0.43690764904022217,
"learning_rate": 0.00029764084507042255,
"loss": 1.5602,
"step": 25000
},
{
"epoch": 4.2517006802721085,
"eval_loss": 1.7416114807128906,
"eval_runtime": 75.3172,
"eval_samples_per_second": 1245.918,
"eval_steps_per_second": 4.873,
"step": 25000
},
{
"epoch": 4.260204081632653,
"grad_norm": 0.48442286252975464,
"learning_rate": 0.0002972007042253521,
"loss": 1.5592,
"step": 25050
},
{
"epoch": 4.2687074829931975,
"grad_norm": 0.44272610545158386,
"learning_rate": 0.0002967605633802817,
"loss": 1.5612,
"step": 25100
},
{
"epoch": 4.2772108843537415,
"grad_norm": 0.44438502192497253,
"learning_rate": 0.0002963204225352113,
"loss": 1.5622,
"step": 25150
},
{
"epoch": 4.285714285714286,
"grad_norm": 0.701813280582428,
"learning_rate": 0.0002958802816901409,
"loss": 1.5639,
"step": 25200
},
{
"epoch": 4.29421768707483,
"grad_norm": 0.4430041015148163,
"learning_rate": 0.0002954401408450704,
"loss": 1.5655,
"step": 25250
},
{
"epoch": 4.302721088435375,
"grad_norm": 0.4703254699707031,
"learning_rate": 0.000295,
"loss": 1.5585,
"step": 25300
},
{
"epoch": 4.311224489795919,
"grad_norm": 0.4456980228424072,
"learning_rate": 0.0002945598591549296,
"loss": 1.5602,
"step": 25350
},
{
"epoch": 4.319727891156463,
"grad_norm": 0.44341742992401123,
"learning_rate": 0.00029411971830985915,
"loss": 1.565,
"step": 25400
},
{
"epoch": 4.328231292517007,
"grad_norm": 0.6620036959648132,
"learning_rate": 0.00029367957746478875,
"loss": 1.563,
"step": 25450
},
{
"epoch": 4.336734693877551,
"grad_norm": 0.4276537597179413,
"learning_rate": 0.0002932482394366197,
"loss": 1.556,
"step": 25500
},
{
"epoch": 4.345238095238095,
"grad_norm": 0.5484892129898071,
"learning_rate": 0.0002928080985915493,
"loss": 1.5683,
"step": 25550
},
{
"epoch": 4.35374149659864,
"grad_norm": 0.4457222819328308,
"learning_rate": 0.0002923679577464789,
"loss": 1.5626,
"step": 25600
},
{
"epoch": 4.362244897959184,
"grad_norm": 0.4299587905406952,
"learning_rate": 0.0002919278169014085,
"loss": 1.5617,
"step": 25650
},
{
"epoch": 4.370748299319728,
"grad_norm": 0.6894689202308655,
"learning_rate": 0.000291487676056338,
"loss": 1.556,
"step": 25700
},
{
"epoch": 4.379251700680272,
"grad_norm": 0.4507542550563812,
"learning_rate": 0.00029104753521126756,
"loss": 1.5605,
"step": 25750
},
{
"epoch": 4.387755102040816,
"grad_norm": 0.45560920238494873,
"learning_rate": 0.0002906073943661972,
"loss": 1.5631,
"step": 25800
},
{
"epoch": 4.39625850340136,
"grad_norm": 0.4419881999492645,
"learning_rate": 0.00029016725352112676,
"loss": 1.558,
"step": 25850
},
{
"epoch": 4.404761904761905,
"grad_norm": 0.42712947726249695,
"learning_rate": 0.00028972711267605635,
"loss": 1.5571,
"step": 25900
},
{
"epoch": 4.413265306122449,
"grad_norm": 0.4276752769947052,
"learning_rate": 0.0002892869718309859,
"loss": 1.5517,
"step": 25950
},
{
"epoch": 4.421768707482993,
"grad_norm": 0.4421616494655609,
"learning_rate": 0.00028884683098591554,
"loss": 1.5549,
"step": 26000
},
{
"epoch": 4.421768707482993,
"eval_loss": 1.7275468111038208,
"eval_runtime": 75.3579,
"eval_samples_per_second": 1245.244,
"eval_steps_per_second": 4.87,
"step": 26000
},
{
"epoch": 4.430272108843537,
"grad_norm": 0.43331626057624817,
"learning_rate": 0.0002884066901408451,
"loss": 1.553,
"step": 26050
},
{
"epoch": 4.438775510204081,
"grad_norm": 0.42342087626457214,
"learning_rate": 0.0002879665492957746,
"loss": 1.5508,
"step": 26100
},
{
"epoch": 4.447278911564625,
"grad_norm": 0.4396851062774658,
"learning_rate": 0.0002875264084507042,
"loss": 1.5496,
"step": 26150
},
{
"epoch": 4.45578231292517,
"grad_norm": 0.4484754204750061,
"learning_rate": 0.0002870862676056338,
"loss": 1.559,
"step": 26200
},
{
"epoch": 4.464285714285714,
"grad_norm": 0.4413243532180786,
"learning_rate": 0.0002866461267605634,
"loss": 1.5497,
"step": 26250
},
{
"epoch": 4.4727891156462585,
"grad_norm": 0.509044885635376,
"learning_rate": 0.00028620598591549295,
"loss": 1.5507,
"step": 26300
},
{
"epoch": 4.4812925170068025,
"grad_norm": 0.4461870491504669,
"learning_rate": 0.00028576584507042255,
"loss": 1.5516,
"step": 26350
},
{
"epoch": 4.489795918367347,
"grad_norm": 0.48340821266174316,
"learning_rate": 0.00028532570422535214,
"loss": 1.5542,
"step": 26400
},
{
"epoch": 4.4982993197278915,
"grad_norm": 0.4417915344238281,
"learning_rate": 0.0002848855633802817,
"loss": 1.5539,
"step": 26450
},
{
"epoch": 4.506802721088436,
"grad_norm": 0.47127747535705566,
"learning_rate": 0.0002844454225352113,
"loss": 1.547,
"step": 26500
},
{
"epoch": 4.51530612244898,
"grad_norm": 0.7486183047294617,
"learning_rate": 0.0002840052816901408,
"loss": 1.5534,
"step": 26550
},
{
"epoch": 4.523809523809524,
"grad_norm": 0.4753170907497406,
"learning_rate": 0.0002835651408450704,
"loss": 1.5568,
"step": 26600
},
{
"epoch": 4.532312925170068,
"grad_norm": 0.43808746337890625,
"learning_rate": 0.000283125,
"loss": 1.5518,
"step": 26650
},
{
"epoch": 4.540816326530612,
"grad_norm": 0.4334624707698822,
"learning_rate": 0.0002826848591549296,
"loss": 1.5507,
"step": 26700
},
{
"epoch": 4.549319727891157,
"grad_norm": 0.4278333783149719,
"learning_rate": 0.00028224471830985915,
"loss": 1.5543,
"step": 26750
},
{
"epoch": 4.557823129251701,
"grad_norm": 0.44058939814567566,
"learning_rate": 0.0002818045774647887,
"loss": 1.5462,
"step": 26800
},
{
"epoch": 4.566326530612245,
"grad_norm": 0.42488718032836914,
"learning_rate": 0.00028136443661971834,
"loss": 1.5555,
"step": 26850
},
{
"epoch": 4.574829931972789,
"grad_norm": 0.4430455267429352,
"learning_rate": 0.0002809242957746479,
"loss": 1.5481,
"step": 26900
},
{
"epoch": 4.583333333333333,
"grad_norm": 1.010523796081543,
"learning_rate": 0.0002804841549295775,
"loss": 1.5446,
"step": 26950
},
{
"epoch": 4.591836734693878,
"grad_norm": 0.42729562520980835,
"learning_rate": 0.000280044014084507,
"loss": 1.5464,
"step": 27000
},
{
"epoch": 4.591836734693878,
"eval_loss": 1.7172296047210693,
"eval_runtime": 75.3539,
"eval_samples_per_second": 1245.31,
"eval_steps_per_second": 4.87,
"step": 27000
},
{
"epoch": 4.600340136054422,
"grad_norm": 0.4399702847003937,
"learning_rate": 0.00027960387323943667,
"loss": 1.5452,
"step": 27050
},
{
"epoch": 4.608843537414966,
"grad_norm": 0.4524286687374115,
"learning_rate": 0.0002791637323943662,
"loss": 1.5534,
"step": 27100
},
{
"epoch": 4.61734693877551,
"grad_norm": 0.43100252747535706,
"learning_rate": 0.00027872359154929575,
"loss": 1.5479,
"step": 27150
},
{
"epoch": 4.625850340136054,
"grad_norm": 0.43541115522384644,
"learning_rate": 0.00027828345070422534,
"loss": 1.542,
"step": 27200
},
{
"epoch": 4.634353741496598,
"grad_norm": 0.454149454832077,
"learning_rate": 0.00027784330985915494,
"loss": 1.5442,
"step": 27250
},
{
"epoch": 4.642857142857143,
"grad_norm": 0.4360577464103699,
"learning_rate": 0.00027740316901408453,
"loss": 1.541,
"step": 27300
},
{
"epoch": 4.651360544217687,
"grad_norm": 0.45018696784973145,
"learning_rate": 0.0002769630281690141,
"loss": 1.5522,
"step": 27350
},
{
"epoch": 4.659863945578231,
"grad_norm": 0.4641735851764679,
"learning_rate": 0.00027652288732394367,
"loss": 1.5409,
"step": 27400
},
{
"epoch": 4.668367346938775,
"grad_norm": 0.4704575836658478,
"learning_rate": 0.00027608274647887327,
"loss": 1.5386,
"step": 27450
},
{
"epoch": 4.6768707482993195,
"grad_norm": 0.4325176775455475,
"learning_rate": 0.0002756426056338028,
"loss": 1.5432,
"step": 27500
},
{
"epoch": 4.685374149659864,
"grad_norm": 0.5194480419158936,
"learning_rate": 0.0002752024647887324,
"loss": 1.5409,
"step": 27550
},
{
"epoch": 4.6938775510204085,
"grad_norm": 0.6043953895568848,
"learning_rate": 0.00027476232394366194,
"loss": 1.5412,
"step": 27600
},
{
"epoch": 4.7023809523809526,
"grad_norm": 0.49982741475105286,
"learning_rate": 0.0002743221830985916,
"loss": 1.544,
"step": 27650
},
{
"epoch": 4.710884353741497,
"grad_norm": 0.45368969440460205,
"learning_rate": 0.00027388204225352113,
"loss": 1.5452,
"step": 27700
},
{
"epoch": 4.719387755102041,
"grad_norm": 0.42949676513671875,
"learning_rate": 0.00027344190140845073,
"loss": 1.5431,
"step": 27750
},
{
"epoch": 4.727891156462585,
"grad_norm": 0.4460367262363434,
"learning_rate": 0.00027300176056338027,
"loss": 1.5379,
"step": 27800
},
{
"epoch": 4.736394557823129,
"grad_norm": 0.43941059708595276,
"learning_rate": 0.0002725616197183098,
"loss": 1.5438,
"step": 27850
},
{
"epoch": 4.744897959183674,
"grad_norm": 0.44306865334510803,
"learning_rate": 0.00027212147887323946,
"loss": 1.5403,
"step": 27900
},
{
"epoch": 4.753401360544218,
"grad_norm": 0.4779192805290222,
"learning_rate": 0.000271681338028169,
"loss": 1.5374,
"step": 27950
},
{
"epoch": 4.761904761904762,
"grad_norm": 0.44125354290008545,
"learning_rate": 0.0002712411971830986,
"loss": 1.5386,
"step": 28000
},
{
"epoch": 4.761904761904762,
"eval_loss": 1.7089167833328247,
"eval_runtime": 75.3103,
"eval_samples_per_second": 1246.032,
"eval_steps_per_second": 4.873,
"step": 28000
},
{
"epoch": 4.770408163265306,
"grad_norm": 0.47051918506622314,
"learning_rate": 0.00027080105633802814,
"loss": 1.5367,
"step": 28050
},
{
"epoch": 4.77891156462585,
"grad_norm": 0.44151777029037476,
"learning_rate": 0.0002703609154929578,
"loss": 1.5363,
"step": 28100
},
{
"epoch": 4.787414965986395,
"grad_norm": 0.43706709146499634,
"learning_rate": 0.00026992077464788733,
"loss": 1.5396,
"step": 28150
},
{
"epoch": 4.795918367346939,
"grad_norm": 0.4319562017917633,
"learning_rate": 0.0002694806338028169,
"loss": 1.5339,
"step": 28200
},
{
"epoch": 4.804421768707483,
"grad_norm": 0.4437416195869446,
"learning_rate": 0.00026904049295774647,
"loss": 1.5259,
"step": 28250
},
{
"epoch": 4.812925170068027,
"grad_norm": 0.45064008235931396,
"learning_rate": 0.00026860035211267606,
"loss": 1.5305,
"step": 28300
},
{
"epoch": 4.821428571428571,
"grad_norm": 0.4223947823047638,
"learning_rate": 0.00026816021126760566,
"loss": 1.5367,
"step": 28350
},
{
"epoch": 4.829931972789115,
"grad_norm": 0.41959166526794434,
"learning_rate": 0.0002677200704225352,
"loss": 1.5321,
"step": 28400
},
{
"epoch": 4.83843537414966,
"grad_norm": 0.4325518310070038,
"learning_rate": 0.0002672799295774648,
"loss": 1.5326,
"step": 28450
},
{
"epoch": 4.846938775510204,
"grad_norm": 0.43113574385643005,
"learning_rate": 0.0002668397887323944,
"loss": 1.5364,
"step": 28500
},
{
"epoch": 4.855442176870748,
"grad_norm": 0.4148525297641754,
"learning_rate": 0.000266399647887324,
"loss": 1.535,
"step": 28550
},
{
"epoch": 4.863945578231292,
"grad_norm": 0.4255242943763733,
"learning_rate": 0.0002659595070422535,
"loss": 1.5307,
"step": 28600
},
{
"epoch": 4.872448979591836,
"grad_norm": 0.4426940679550171,
"learning_rate": 0.00026551936619718307,
"loss": 1.5298,
"step": 28650
},
{
"epoch": 4.880952380952381,
"grad_norm": 0.4304163157939911,
"learning_rate": 0.0002650792253521127,
"loss": 1.5296,
"step": 28700
},
{
"epoch": 4.889455782312925,
"grad_norm": 0.43492448329925537,
"learning_rate": 0.00026463908450704226,
"loss": 1.5295,
"step": 28750
},
{
"epoch": 4.8979591836734695,
"grad_norm": 0.4191015660762787,
"learning_rate": 0.00026420774647887326,
"loss": 1.5275,
"step": 28800
},
{
"epoch": 4.906462585034014,
"grad_norm": 0.4442043602466583,
"learning_rate": 0.0002637676056338028,
"loss": 1.5241,
"step": 28850
},
{
"epoch": 4.914965986394558,
"grad_norm": 0.41631436347961426,
"learning_rate": 0.0002633274647887324,
"loss": 1.5296,
"step": 28900
},
{
"epoch": 4.923469387755102,
"grad_norm": 0.4311072528362274,
"learning_rate": 0.000262887323943662,
"loss": 1.5343,
"step": 28950
},
{
"epoch": 4.931972789115647,
"grad_norm": 0.4260921776294708,
"learning_rate": 0.00026244718309859154,
"loss": 1.528,
"step": 29000
},
{
"epoch": 4.931972789115647,
"eval_loss": 1.695530652999878,
"eval_runtime": 75.7527,
"eval_samples_per_second": 1238.755,
"eval_steps_per_second": 4.845,
"step": 29000
},
{
"epoch": 4.940476190476191,
"grad_norm": 0.43122732639312744,
"learning_rate": 0.00026200704225352113,
"loss": 1.5345,
"step": 29050
},
{
"epoch": 4.948979591836735,
"grad_norm": 0.44380810856819153,
"learning_rate": 0.00026156690140845067,
"loss": 1.5284,
"step": 29100
},
{
"epoch": 4.957482993197279,
"grad_norm": 0.43009868264198303,
"learning_rate": 0.0002611267605633803,
"loss": 1.5315,
"step": 29150
},
{
"epoch": 4.965986394557823,
"grad_norm": 0.4425157904624939,
"learning_rate": 0.00026068661971830986,
"loss": 1.5301,
"step": 29200
},
{
"epoch": 4.974489795918368,
"grad_norm": 0.41929659247398376,
"learning_rate": 0.00026024647887323946,
"loss": 1.533,
"step": 29250
},
{
"epoch": 4.982993197278912,
"grad_norm": 0.43218111991882324,
"learning_rate": 0.000259806338028169,
"loss": 1.5215,
"step": 29300
},
{
"epoch": 4.991496598639456,
"grad_norm": 0.4425402283668518,
"learning_rate": 0.0002593661971830986,
"loss": 1.5312,
"step": 29350
},
{
"epoch": 5.0,
"grad_norm": 0.44686195254325867,
"learning_rate": 0.0002589260563380282,
"loss": 1.5262,
"step": 29400
},
{
"epoch": 5.008503401360544,
"grad_norm": 0.42469632625579834,
"learning_rate": 0.00025848591549295773,
"loss": 1.5216,
"step": 29450
},
{
"epoch": 5.017006802721088,
"grad_norm": 0.44917765259742737,
"learning_rate": 0.0002580457746478873,
"loss": 1.5148,
"step": 29500
},
{
"epoch": 5.025510204081633,
"grad_norm": 0.42792636156082153,
"learning_rate": 0.0002576056338028169,
"loss": 1.5236,
"step": 29550
},
{
"epoch": 5.034013605442177,
"grad_norm": 0.42785459756851196,
"learning_rate": 0.0002571654929577465,
"loss": 1.5185,
"step": 29600
},
{
"epoch": 5.042517006802721,
"grad_norm": 0.4389972388744354,
"learning_rate": 0.00025672535211267606,
"loss": 1.5152,
"step": 29650
},
{
"epoch": 5.051020408163265,
"grad_norm": 0.4294348657131195,
"learning_rate": 0.0002562852112676056,
"loss": 1.5164,
"step": 29700
},
{
"epoch": 5.059523809523809,
"grad_norm": 0.4204474985599518,
"learning_rate": 0.00025584507042253525,
"loss": 1.5126,
"step": 29750
},
{
"epoch": 5.068027210884353,
"grad_norm": 0.4308267831802368,
"learning_rate": 0.0002554049295774648,
"loss": 1.5138,
"step": 29800
},
{
"epoch": 5.076530612244898,
"grad_norm": 0.43260395526885986,
"learning_rate": 0.0002549647887323944,
"loss": 1.5211,
"step": 29850
},
{
"epoch": 5.085034013605442,
"grad_norm": 0.4192730486392975,
"learning_rate": 0.00025452464788732393,
"loss": 1.5165,
"step": 29900
},
{
"epoch": 5.093537414965986,
"grad_norm": 0.418124794960022,
"learning_rate": 0.0002540845070422535,
"loss": 1.5152,
"step": 29950
},
{
"epoch": 5.1020408163265305,
"grad_norm": 0.4320596158504486,
"learning_rate": 0.0002536443661971831,
"loss": 1.5071,
"step": 30000
},
{
"epoch": 5.1020408163265305,
"eval_loss": 1.6892961263656616,
"eval_runtime": 75.4228,
"eval_samples_per_second": 1244.173,
"eval_steps_per_second": 4.866,
"step": 30000
},
{
"epoch": 5.110544217687075,
"grad_norm": 0.43125247955322266,
"learning_rate": 0.00025320422535211266,
"loss": 1.5116,
"step": 30050
},
{
"epoch": 5.119047619047619,
"grad_norm": 0.4130070209503174,
"learning_rate": 0.00025276408450704225,
"loss": 1.5125,
"step": 30100
},
{
"epoch": 5.127551020408164,
"grad_norm": 0.42717909812927246,
"learning_rate": 0.0002523239436619718,
"loss": 1.5122,
"step": 30150
},
{
"epoch": 5.136054421768708,
"grad_norm": 0.4337432086467743,
"learning_rate": 0.00025188380281690145,
"loss": 1.5103,
"step": 30200
},
{
"epoch": 5.144557823129252,
"grad_norm": 0.4125542938709259,
"learning_rate": 0.000251443661971831,
"loss": 1.5218,
"step": 30250
},
{
"epoch": 5.153061224489796,
"grad_norm": 0.428830623626709,
"learning_rate": 0.0002510035211267606,
"loss": 1.5068,
"step": 30300
},
{
"epoch": 5.16156462585034,
"grad_norm": 0.41956421732902527,
"learning_rate": 0.0002505633802816901,
"loss": 1.5148,
"step": 30350
},
{
"epoch": 5.170068027210885,
"grad_norm": 0.43243858218193054,
"learning_rate": 0.0002501232394366197,
"loss": 1.5153,
"step": 30400
},
{
"epoch": 5.178571428571429,
"grad_norm": 0.41967883706092834,
"learning_rate": 0.0002496830985915493,
"loss": 1.517,
"step": 30450
},
{
"epoch": 5.187074829931973,
"grad_norm": 0.43599751591682434,
"learning_rate": 0.00024924295774647886,
"loss": 1.5097,
"step": 30500
},
{
"epoch": 5.195578231292517,
"grad_norm": 0.4145863950252533,
"learning_rate": 0.00024880281690140845,
"loss": 1.5054,
"step": 30550
},
{
"epoch": 5.204081632653061,
"grad_norm": 0.44736286997795105,
"learning_rate": 0.00024836267605633805,
"loss": 1.512,
"step": 30600
},
{
"epoch": 5.212585034013605,
"grad_norm": 0.4339098036289215,
"learning_rate": 0.00024792253521126764,
"loss": 1.5119,
"step": 30650
},
{
"epoch": 5.22108843537415,
"grad_norm": 0.4354366064071655,
"learning_rate": 0.0002474823943661972,
"loss": 1.5134,
"step": 30700
},
{
"epoch": 5.229591836734694,
"grad_norm": 0.4405953288078308,
"learning_rate": 0.0002470422535211268,
"loss": 1.5117,
"step": 30750
},
{
"epoch": 5.238095238095238,
"grad_norm": 0.4392901659011841,
"learning_rate": 0.0002466021126760563,
"loss": 1.5097,
"step": 30800
},
{
"epoch": 5.246598639455782,
"grad_norm": 0.4320152699947357,
"learning_rate": 0.0002461619718309859,
"loss": 1.5185,
"step": 30850
},
{
"epoch": 5.255102040816326,
"grad_norm": 0.4348084628582001,
"learning_rate": 0.0002457218309859155,
"loss": 1.5073,
"step": 30900
},
{
"epoch": 5.263605442176871,
"grad_norm": 0.4216504693031311,
"learning_rate": 0.0002452816901408451,
"loss": 1.5055,
"step": 30950
},
{
"epoch": 5.272108843537415,
"grad_norm": 0.417338103055954,
"learning_rate": 0.00024484154929577465,
"loss": 1.5102,
"step": 31000
},
{
"epoch": 5.272108843537415,
"eval_loss": 1.6826726198196411,
"eval_runtime": 75.3283,
"eval_samples_per_second": 1245.734,
"eval_steps_per_second": 4.872,
"step": 31000
},
{
"epoch": 5.280612244897959,
"grad_norm": 0.425647109746933,
"learning_rate": 0.00024440140845070424,
"loss": 1.51,
"step": 31050
},
{
"epoch": 5.289115646258503,
"grad_norm": 0.4205061197280884,
"learning_rate": 0.0002439612676056338,
"loss": 1.5071,
"step": 31100
},
{
"epoch": 5.2976190476190474,
"grad_norm": 0.4191541075706482,
"learning_rate": 0.00024352112676056338,
"loss": 1.5072,
"step": 31150
},
{
"epoch": 5.3061224489795915,
"grad_norm": 0.4266960024833679,
"learning_rate": 0.00024308098591549297,
"loss": 1.5024,
"step": 31200
},
{
"epoch": 5.3146258503401365,
"grad_norm": 0.43113160133361816,
"learning_rate": 0.00024264084507042252,
"loss": 1.5045,
"step": 31250
},
{
"epoch": 5.3231292517006805,
"grad_norm": 0.46872562170028687,
"learning_rate": 0.0002422007042253521,
"loss": 1.5043,
"step": 31300
},
{
"epoch": 5.331632653061225,
"grad_norm": 0.444063663482666,
"learning_rate": 0.00024176056338028168,
"loss": 1.503,
"step": 31350
},
{
"epoch": 5.340136054421769,
"grad_norm": 0.46224623918533325,
"learning_rate": 0.00024132042253521127,
"loss": 1.5001,
"step": 31400
},
{
"epoch": 5.348639455782313,
"grad_norm": 0.4229288697242737,
"learning_rate": 0.00024088028169014084,
"loss": 1.5051,
"step": 31450
},
{
"epoch": 5.357142857142857,
"grad_norm": 0.4342740476131439,
"learning_rate": 0.00024044014084507044,
"loss": 1.5043,
"step": 31500
},
{
"epoch": 5.365646258503402,
"grad_norm": 0.4987669885158539,
"learning_rate": 0.00024,
"loss": 1.5067,
"step": 31550
},
{
"epoch": 5.374149659863946,
"grad_norm": 0.4277132451534271,
"learning_rate": 0.0002395598591549296,
"loss": 1.5005,
"step": 31600
},
{
"epoch": 5.38265306122449,
"grad_norm": 0.4543018639087677,
"learning_rate": 0.00023911971830985914,
"loss": 1.5081,
"step": 31650
},
{
"epoch": 5.391156462585034,
"grad_norm": 0.4164925217628479,
"learning_rate": 0.00023867957746478874,
"loss": 1.4998,
"step": 31700
},
{
"epoch": 5.399659863945578,
"grad_norm": 0.4165375828742981,
"learning_rate": 0.0002382394366197183,
"loss": 1.5047,
"step": 31750
},
{
"epoch": 5.408163265306122,
"grad_norm": 0.43289613723754883,
"learning_rate": 0.0002377992957746479,
"loss": 1.499,
"step": 31800
},
{
"epoch": 5.416666666666667,
"grad_norm": 0.4113534688949585,
"learning_rate": 0.00023735915492957747,
"loss": 1.4992,
"step": 31850
},
{
"epoch": 5.425170068027211,
"grad_norm": 0.4257815480232239,
"learning_rate": 0.00023691901408450707,
"loss": 1.5033,
"step": 31900
},
{
"epoch": 5.433673469387755,
"grad_norm": 0.423606812953949,
"learning_rate": 0.00023647887323943663,
"loss": 1.4985,
"step": 31950
},
{
"epoch": 5.442176870748299,
"grad_norm": 0.5088992118835449,
"learning_rate": 0.0002360387323943662,
"loss": 1.5025,
"step": 32000
},
{
"epoch": 5.442176870748299,
"eval_loss": 1.6751487255096436,
"eval_runtime": 75.6944,
"eval_samples_per_second": 1239.708,
"eval_steps_per_second": 4.848,
"step": 32000
},
{
"epoch": 5.450680272108843,
"grad_norm": 0.42682725191116333,
"learning_rate": 0.00023559859154929577,
"loss": 1.5012,
"step": 32050
},
{
"epoch": 5.459183673469388,
"grad_norm": 0.4484409689903259,
"learning_rate": 0.00023515845070422537,
"loss": 1.5056,
"step": 32100
},
{
"epoch": 5.467687074829932,
"grad_norm": 0.4330119788646698,
"learning_rate": 0.00023471830985915493,
"loss": 1.4997,
"step": 32150
},
{
"epoch": 5.476190476190476,
"grad_norm": 0.4262082874774933,
"learning_rate": 0.00023427816901408453,
"loss": 1.4992,
"step": 32200
},
{
"epoch": 5.48469387755102,
"grad_norm": 0.4434804320335388,
"learning_rate": 0.0002338380281690141,
"loss": 1.4975,
"step": 32250
},
{
"epoch": 5.493197278911564,
"grad_norm": 0.42507055401802063,
"learning_rate": 0.00023340669014084507,
"loss": 1.4923,
"step": 32300
},
{
"epoch": 5.5017006802721085,
"grad_norm": 0.415923535823822,
"learning_rate": 0.00023296654929577464,
"loss": 1.4899,
"step": 32350
},
{
"epoch": 5.510204081632653,
"grad_norm": 0.43220096826553345,
"learning_rate": 0.0002325264084507042,
"loss": 1.4947,
"step": 32400
},
{
"epoch": 5.5187074829931975,
"grad_norm": 0.4283994436264038,
"learning_rate": 0.0002320862676056338,
"loss": 1.4921,
"step": 32450
},
{
"epoch": 5.5272108843537415,
"grad_norm": 0.439207524061203,
"learning_rate": 0.00023164612676056337,
"loss": 1.4859,
"step": 32500
},
{
"epoch": 5.535714285714286,
"grad_norm": 0.4383983612060547,
"learning_rate": 0.00023120598591549297,
"loss": 1.4933,
"step": 32550
},
{
"epoch": 5.54421768707483,
"grad_norm": 0.44482484459877014,
"learning_rate": 0.00023076584507042254,
"loss": 1.4957,
"step": 32600
},
{
"epoch": 5.552721088435375,
"grad_norm": 0.4300253987312317,
"learning_rate": 0.00023032570422535213,
"loss": 1.49,
"step": 32650
},
{
"epoch": 5.561224489795919,
"grad_norm": 0.46058133244514465,
"learning_rate": 0.00022988556338028167,
"loss": 1.4872,
"step": 32700
},
{
"epoch": 5.569727891156463,
"grad_norm": 0.4194552004337311,
"learning_rate": 0.00022944542253521127,
"loss": 1.4911,
"step": 32750
},
{
"epoch": 5.578231292517007,
"grad_norm": 0.42726796865463257,
"learning_rate": 0.00022900528169014084,
"loss": 1.4926,
"step": 32800
},
{
"epoch": 5.586734693877551,
"grad_norm": 0.42779192328453064,
"learning_rate": 0.00022856514084507043,
"loss": 1.4917,
"step": 32850
},
{
"epoch": 5.595238095238095,
"grad_norm": 0.42153412103652954,
"learning_rate": 0.000228125,
"loss": 1.4824,
"step": 32900
},
{
"epoch": 5.603741496598639,
"grad_norm": 0.4386555254459381,
"learning_rate": 0.0002276848591549296,
"loss": 1.4916,
"step": 32950
},
{
"epoch": 5.612244897959184,
"grad_norm": 0.4412122964859009,
"learning_rate": 0.00022724471830985917,
"loss": 1.4908,
"step": 33000
},
{
"epoch": 5.612244897959184,
"eval_loss": 1.6691830158233643,
"eval_runtime": 75.3357,
"eval_samples_per_second": 1245.611,
"eval_steps_per_second": 4.872,
"step": 33000
},
{
"epoch": 5.620748299319728,
"grad_norm": 0.42562374472618103,
"learning_rate": 0.00022680457746478873,
"loss": 1.4867,
"step": 33050
},
{
"epoch": 5.629251700680272,
"grad_norm": 0.4305470585823059,
"learning_rate": 0.0002263644366197183,
"loss": 1.4874,
"step": 33100
},
{
"epoch": 5.637755102040816,
"grad_norm": 0.42887523770332336,
"learning_rate": 0.0002259242957746479,
"loss": 1.4993,
"step": 33150
},
{
"epoch": 5.646258503401361,
"grad_norm": 0.5242183804512024,
"learning_rate": 0.00022548415492957747,
"loss": 1.4934,
"step": 33200
},
{
"epoch": 5.654761904761905,
"grad_norm": 0.4118496775627136,
"learning_rate": 0.00022504401408450706,
"loss": 1.4872,
"step": 33250
},
{
"epoch": 5.663265306122449,
"grad_norm": 0.4159320890903473,
"learning_rate": 0.00022460387323943663,
"loss": 1.4782,
"step": 33300
},
{
"epoch": 5.671768707482993,
"grad_norm": 0.4275069534778595,
"learning_rate": 0.00022416373239436623,
"loss": 1.4857,
"step": 33350
},
{
"epoch": 5.680272108843537,
"grad_norm": 0.41769281029701233,
"learning_rate": 0.00022372359154929577,
"loss": 1.479,
"step": 33400
},
{
"epoch": 5.688775510204081,
"grad_norm": 0.41575032472610474,
"learning_rate": 0.00022328345070422533,
"loss": 1.4849,
"step": 33450
},
{
"epoch": 5.697278911564625,
"grad_norm": 0.4337887763977051,
"learning_rate": 0.00022284330985915493,
"loss": 1.4877,
"step": 33500
},
{
"epoch": 5.70578231292517,
"grad_norm": 0.41992053389549255,
"learning_rate": 0.0002224031690140845,
"loss": 1.4809,
"step": 33550
},
{
"epoch": 5.714285714285714,
"grad_norm": 0.42422786355018616,
"learning_rate": 0.0002219630281690141,
"loss": 1.4877,
"step": 33600
},
{
"epoch": 5.7227891156462585,
"grad_norm": 0.415855348110199,
"learning_rate": 0.00022152288732394366,
"loss": 1.4863,
"step": 33650
},
{
"epoch": 5.7312925170068025,
"grad_norm": 0.4512421488761902,
"learning_rate": 0.00022108274647887326,
"loss": 1.4894,
"step": 33700
},
{
"epoch": 5.739795918367347,
"grad_norm": 0.42408186197280884,
"learning_rate": 0.0002206426056338028,
"loss": 1.4767,
"step": 33750
},
{
"epoch": 5.7482993197278915,
"grad_norm": 0.42864853143692017,
"learning_rate": 0.0002202024647887324,
"loss": 1.4825,
"step": 33800
},
{
"epoch": 5.756802721088436,
"grad_norm": 0.4149647355079651,
"learning_rate": 0.00021976232394366196,
"loss": 1.4819,
"step": 33850
},
{
"epoch": 5.76530612244898,
"grad_norm": 0.45217257738113403,
"learning_rate": 0.00021932218309859156,
"loss": 1.487,
"step": 33900
},
{
"epoch": 5.773809523809524,
"grad_norm": 0.4308101236820221,
"learning_rate": 0.00021888204225352113,
"loss": 1.4842,
"step": 33950
},
{
"epoch": 5.782312925170068,
"grad_norm": 0.41859814524650574,
"learning_rate": 0.00021844190140845072,
"loss": 1.4856,
"step": 34000
},
{
"epoch": 5.782312925170068,
"eval_loss": 1.6499308347702026,
"eval_runtime": 75.302,
"eval_samples_per_second": 1246.169,
"eval_steps_per_second": 4.874,
"step": 34000
},
{
"epoch": 5.790816326530612,
"grad_norm": 0.4088199734687805,
"learning_rate": 0.0002180017605633803,
"loss": 1.4838,
"step": 34050
},
{
"epoch": 5.799319727891157,
"grad_norm": 0.8137006163597107,
"learning_rate": 0.00021756161971830989,
"loss": 1.4826,
"step": 34100
},
{
"epoch": 5.807823129251701,
"grad_norm": 0.4124939441680908,
"learning_rate": 0.00021712147887323943,
"loss": 1.4916,
"step": 34150
},
{
"epoch": 5.816326530612245,
"grad_norm": 0.4147464334964752,
"learning_rate": 0.00021668133802816902,
"loss": 1.485,
"step": 34200
},
{
"epoch": 5.824829931972789,
"grad_norm": 0.42168349027633667,
"learning_rate": 0.0002162411971830986,
"loss": 1.4787,
"step": 34250
},
{
"epoch": 5.833333333333333,
"grad_norm": 0.43687647581100464,
"learning_rate": 0.00021580105633802819,
"loss": 1.4859,
"step": 34300
},
{
"epoch": 5.841836734693878,
"grad_norm": 0.4150829017162323,
"learning_rate": 0.00021536091549295775,
"loss": 1.4806,
"step": 34350
},
{
"epoch": 5.850340136054422,
"grad_norm": 0.42533501982688904,
"learning_rate": 0.00021492077464788735,
"loss": 1.4774,
"step": 34400
},
{
"epoch": 5.858843537414966,
"grad_norm": 0.42881202697753906,
"learning_rate": 0.00021448063380281692,
"loss": 1.4793,
"step": 34450
},
{
"epoch": 5.86734693877551,
"grad_norm": 0.41647574305534363,
"learning_rate": 0.00021404049295774649,
"loss": 1.4844,
"step": 34500
},
{
"epoch": 5.875850340136054,
"grad_norm": 0.45438751578330994,
"learning_rate": 0.00021360035211267605,
"loss": 1.476,
"step": 34550
},
{
"epoch": 5.884353741496598,
"grad_norm": 0.4157373309135437,
"learning_rate": 0.00021316021126760562,
"loss": 1.4783,
"step": 34600
},
{
"epoch": 5.892857142857143,
"grad_norm": 0.4178527593612671,
"learning_rate": 0.00021272007042253522,
"loss": 1.477,
"step": 34650
},
{
"epoch": 5.901360544217687,
"grad_norm": 0.4168199896812439,
"learning_rate": 0.00021227992957746479,
"loss": 1.4685,
"step": 34700
},
{
"epoch": 5.909863945578231,
"grad_norm": 0.4487851858139038,
"learning_rate": 0.00021183978873239438,
"loss": 1.4782,
"step": 34750
},
{
"epoch": 5.918367346938775,
"grad_norm": 0.43783414363861084,
"learning_rate": 0.00021139964788732395,
"loss": 1.4727,
"step": 34800
},
{
"epoch": 5.9268707482993195,
"grad_norm": 0.4123859107494354,
"learning_rate": 0.00021095950704225352,
"loss": 1.4711,
"step": 34850
},
{
"epoch": 5.935374149659864,
"grad_norm": 0.41514667868614197,
"learning_rate": 0.00021051936619718309,
"loss": 1.4776,
"step": 34900
},
{
"epoch": 5.9438775510204085,
"grad_norm": 0.41285404562950134,
"learning_rate": 0.00021007922535211268,
"loss": 1.4796,
"step": 34950
},
{
"epoch": 5.9523809523809526,
"grad_norm": 0.407366544008255,
"learning_rate": 0.00020963908450704225,
"loss": 1.4687,
"step": 35000
},
{
"epoch": 5.9523809523809526,
"eval_loss": 1.6382627487182617,
"eval_runtime": 75.3966,
"eval_samples_per_second": 1244.606,
"eval_steps_per_second": 4.868,
"step": 35000
},
{
"epoch": 5.960884353741497,
"grad_norm": 0.4154813885688782,
"learning_rate": 0.00020919894366197185,
"loss": 1.4731,
"step": 35050
},
{
"epoch": 5.969387755102041,
"grad_norm": 0.40901270508766174,
"learning_rate": 0.0002087588028169014,
"loss": 1.4767,
"step": 35100
},
{
"epoch": 5.977891156462585,
"grad_norm": 0.41301828622817993,
"learning_rate": 0.000208318661971831,
"loss": 1.4688,
"step": 35150
},
{
"epoch": 5.986394557823129,
"grad_norm": 0.43283048272132874,
"learning_rate": 0.00020787852112676055,
"loss": 1.4673,
"step": 35200
},
{
"epoch": 5.994897959183674,
"grad_norm": 0.4242665469646454,
"learning_rate": 0.00020743838028169015,
"loss": 1.4749,
"step": 35250
},
{
"epoch": 6.003401360544218,
"grad_norm": 0.4176988899707794,
"learning_rate": 0.00020699823943661971,
"loss": 1.4713,
"step": 35300
},
{
"epoch": 6.011904761904762,
"grad_norm": 0.4250063896179199,
"learning_rate": 0.0002065580985915493,
"loss": 1.4688,
"step": 35350
},
{
"epoch": 6.020408163265306,
"grad_norm": 0.4247675836086273,
"learning_rate": 0.00020611795774647888,
"loss": 1.4661,
"step": 35400
},
{
"epoch": 6.02891156462585,
"grad_norm": 0.47026437520980835,
"learning_rate": 0.00020567781690140847,
"loss": 1.4641,
"step": 35450
},
{
"epoch": 6.037414965986395,
"grad_norm": 0.41622287034988403,
"learning_rate": 0.00020523767605633804,
"loss": 1.4676,
"step": 35500
},
{
"epoch": 6.045918367346939,
"grad_norm": 0.4315873980522156,
"learning_rate": 0.0002047975352112676,
"loss": 1.4701,
"step": 35550
},
{
"epoch": 6.054421768707483,
"grad_norm": 0.4140373170375824,
"learning_rate": 0.00020435739436619718,
"loss": 1.4625,
"step": 35600
},
{
"epoch": 6.062925170068027,
"grad_norm": 0.41319021582603455,
"learning_rate": 0.00020391725352112677,
"loss": 1.4674,
"step": 35650
},
{
"epoch": 6.071428571428571,
"grad_norm": 0.4266904294490814,
"learning_rate": 0.00020347711267605634,
"loss": 1.4651,
"step": 35700
},
{
"epoch": 6.079931972789115,
"grad_norm": 0.422049880027771,
"learning_rate": 0.00020304577464788732,
"loss": 1.4602,
"step": 35750
},
{
"epoch": 6.08843537414966,
"grad_norm": 0.4264945089817047,
"learning_rate": 0.0002026056338028169,
"loss": 1.4629,
"step": 35800
},
{
"epoch": 6.096938775510204,
"grad_norm": 0.47240814566612244,
"learning_rate": 0.00020216549295774648,
"loss": 1.4675,
"step": 35850
},
{
"epoch": 6.105442176870748,
"grad_norm": 0.4204316735267639,
"learning_rate": 0.00020172535211267608,
"loss": 1.4627,
"step": 35900
},
{
"epoch": 6.113945578231292,
"grad_norm": 0.41577088832855225,
"learning_rate": 0.00020128521126760562,
"loss": 1.4605,
"step": 35950
},
{
"epoch": 6.122448979591836,
"grad_norm": 0.4297136962413788,
"learning_rate": 0.0002008450704225352,
"loss": 1.4685,
"step": 36000
},
{
"epoch": 6.122448979591836,
"eval_loss": 1.6419459581375122,
"eval_runtime": 75.3644,
"eval_samples_per_second": 1245.137,
"eval_steps_per_second": 4.87,
"step": 36000
},
{
"epoch": 6.130952380952381,
"grad_norm": 0.408674418926239,
"learning_rate": 0.00020040492957746478,
"loss": 1.4666,
"step": 36050
},
{
"epoch": 6.139455782312925,
"grad_norm": 0.41451412439346313,
"learning_rate": 0.00019996478873239438,
"loss": 1.4611,
"step": 36100
},
{
"epoch": 6.1479591836734695,
"grad_norm": 0.4128754734992981,
"learning_rate": 0.00019952464788732395,
"loss": 1.4589,
"step": 36150
},
{
"epoch": 6.156462585034014,
"grad_norm": 0.4213225543498993,
"learning_rate": 0.00019908450704225354,
"loss": 1.462,
"step": 36200
},
{
"epoch": 6.164965986394558,
"grad_norm": 0.4226750433444977,
"learning_rate": 0.0001986443661971831,
"loss": 1.4605,
"step": 36250
},
{
"epoch": 6.173469387755102,
"grad_norm": 0.4176504611968994,
"learning_rate": 0.00019820422535211268,
"loss": 1.459,
"step": 36300
},
{
"epoch": 6.181972789115647,
"grad_norm": 0.42486241459846497,
"learning_rate": 0.00019776408450704225,
"loss": 1.4656,
"step": 36350
},
{
"epoch": 6.190476190476191,
"grad_norm": 0.40891167521476746,
"learning_rate": 0.00019732394366197184,
"loss": 1.4601,
"step": 36400
},
{
"epoch": 6.198979591836735,
"grad_norm": 0.4230293035507202,
"learning_rate": 0.0001968838028169014,
"loss": 1.4537,
"step": 36450
},
{
"epoch": 6.207482993197279,
"grad_norm": 0.4342026710510254,
"learning_rate": 0.000196443661971831,
"loss": 1.4575,
"step": 36500
},
{
"epoch": 6.215986394557823,
"grad_norm": 0.417348176240921,
"learning_rate": 0.00019600352112676057,
"loss": 1.4618,
"step": 36550
},
{
"epoch": 6.224489795918367,
"grad_norm": 0.4206349551677704,
"learning_rate": 0.00019556338028169017,
"loss": 1.4596,
"step": 36600
},
{
"epoch": 6.232993197278912,
"grad_norm": 0.42500296235084534,
"learning_rate": 0.0001951232394366197,
"loss": 1.4534,
"step": 36650
},
{
"epoch": 6.241496598639456,
"grad_norm": 0.42607611417770386,
"learning_rate": 0.0001946830985915493,
"loss": 1.4621,
"step": 36700
},
{
"epoch": 6.25,
"grad_norm": 0.4912338852882385,
"learning_rate": 0.00019424295774647887,
"loss": 1.4629,
"step": 36750
},
{
"epoch": 6.258503401360544,
"grad_norm": 0.40748193860054016,
"learning_rate": 0.00019380281690140847,
"loss": 1.4639,
"step": 36800
},
{
"epoch": 6.267006802721088,
"grad_norm": 0.41081297397613525,
"learning_rate": 0.00019336267605633804,
"loss": 1.4505,
"step": 36850
},
{
"epoch": 6.275510204081632,
"grad_norm": 0.41694825887680054,
"learning_rate": 0.0001929225352112676,
"loss": 1.4573,
"step": 36900
},
{
"epoch": 6.284013605442177,
"grad_norm": 0.4185777008533478,
"learning_rate": 0.0001924823943661972,
"loss": 1.4621,
"step": 36950
},
{
"epoch": 6.292517006802721,
"grad_norm": 0.6816411018371582,
"learning_rate": 0.00019204225352112674,
"loss": 1.4526,
"step": 37000
},
{
"epoch": 6.292517006802721,
"eval_loss": 1.6315878629684448,
"eval_runtime": 75.3702,
"eval_samples_per_second": 1245.041,
"eval_steps_per_second": 4.869,
"step": 37000
},
{
"epoch": 6.301020408163265,
"grad_norm": 0.4173048436641693,
"learning_rate": 0.00019160211267605634,
"loss": 1.4542,
"step": 37050
},
{
"epoch": 6.309523809523809,
"grad_norm": 0.4043162167072296,
"learning_rate": 0.0001911619718309859,
"loss": 1.4541,
"step": 37100
},
{
"epoch": 6.318027210884353,
"grad_norm": 0.4198301136493683,
"learning_rate": 0.0001907218309859155,
"loss": 1.4526,
"step": 37150
},
{
"epoch": 6.326530612244898,
"grad_norm": 0.4946387708187103,
"learning_rate": 0.00019028169014084507,
"loss": 1.4631,
"step": 37200
},
{
"epoch": 6.335034013605442,
"grad_norm": 0.431738018989563,
"learning_rate": 0.00018984154929577466,
"loss": 1.449,
"step": 37250
},
{
"epoch": 6.343537414965986,
"grad_norm": 0.41441062092781067,
"learning_rate": 0.00018940140845070423,
"loss": 1.4501,
"step": 37300
},
{
"epoch": 6.3520408163265305,
"grad_norm": 0.41452255845069885,
"learning_rate": 0.0001889612676056338,
"loss": 1.445,
"step": 37350
},
{
"epoch": 6.360544217687075,
"grad_norm": 0.4315445125102997,
"learning_rate": 0.00018852112676056337,
"loss": 1.4556,
"step": 37400
},
{
"epoch": 6.369047619047619,
"grad_norm": 0.41262757778167725,
"learning_rate": 0.00018808098591549297,
"loss": 1.4531,
"step": 37450
},
{
"epoch": 6.377551020408164,
"grad_norm": 0.42860937118530273,
"learning_rate": 0.00018764084507042253,
"loss": 1.4612,
"step": 37500
},
{
"epoch": 6.386054421768708,
"grad_norm": 0.4128934442996979,
"learning_rate": 0.00018720070422535213,
"loss": 1.4467,
"step": 37550
},
{
"epoch": 6.394557823129252,
"grad_norm": 0.4122573137283325,
"learning_rate": 0.0001867605633802817,
"loss": 1.446,
"step": 37600
},
{
"epoch": 6.403061224489796,
"grad_norm": 0.41068488359451294,
"learning_rate": 0.0001863204225352113,
"loss": 1.4515,
"step": 37650
},
{
"epoch": 6.41156462585034,
"grad_norm": 0.4696698784828186,
"learning_rate": 0.00018588028169014083,
"loss": 1.4508,
"step": 37700
},
{
"epoch": 6.420068027210885,
"grad_norm": 0.41070887446403503,
"learning_rate": 0.00018544014084507043,
"loss": 1.4496,
"step": 37750
},
{
"epoch": 6.428571428571429,
"grad_norm": 0.4017719626426697,
"learning_rate": 0.000185,
"loss": 1.4457,
"step": 37800
},
{
"epoch": 6.437074829931973,
"grad_norm": 0.420674592256546,
"learning_rate": 0.0001845598591549296,
"loss": 1.4524,
"step": 37850
},
{
"epoch": 6.445578231292517,
"grad_norm": 0.4091247618198395,
"learning_rate": 0.00018411971830985916,
"loss": 1.4454,
"step": 37900
},
{
"epoch": 6.454081632653061,
"grad_norm": 0.45321521162986755,
"learning_rate": 0.00018367957746478876,
"loss": 1.4511,
"step": 37950
},
{
"epoch": 6.462585034013605,
"grad_norm": 0.4094868004322052,
"learning_rate": 0.00018323943661971832,
"loss": 1.4523,
"step": 38000
},
{
"epoch": 6.462585034013605,
"eval_loss": 1.6195621490478516,
"eval_runtime": 75.373,
"eval_samples_per_second": 1244.995,
"eval_steps_per_second": 4.869,
"step": 38000
},
{
"epoch": 6.47108843537415,
"grad_norm": 0.41573402285575867,
"learning_rate": 0.00018279929577464787,
"loss": 1.4493,
"step": 38050
},
{
"epoch": 6.479591836734694,
"grad_norm": 0.4076358675956726,
"learning_rate": 0.00018235915492957746,
"loss": 1.4479,
"step": 38100
},
{
"epoch": 6.488095238095238,
"grad_norm": 0.41350802779197693,
"learning_rate": 0.00018192781690140844,
"loss": 1.4474,
"step": 38150
},
{
"epoch": 6.496598639455782,
"grad_norm": 0.42011725902557373,
"learning_rate": 0.00018149647887323944,
"loss": 1.4579,
"step": 38200
},
{
"epoch": 6.505102040816326,
"grad_norm": 0.4202333688735962,
"learning_rate": 0.000181056338028169,
"loss": 1.448,
"step": 38250
},
{
"epoch": 6.513605442176871,
"grad_norm": 0.41736331582069397,
"learning_rate": 0.0001806161971830986,
"loss": 1.4421,
"step": 38300
},
{
"epoch": 6.522108843537415,
"grad_norm": 0.4162420332431793,
"learning_rate": 0.00018017605633802817,
"loss": 1.4401,
"step": 38350
},
{
"epoch": 6.530612244897959,
"grad_norm": 0.40687423944473267,
"learning_rate": 0.00017973591549295777,
"loss": 1.4439,
"step": 38400
},
{
"epoch": 6.539115646258503,
"grad_norm": 0.42472130060195923,
"learning_rate": 0.0001792957746478873,
"loss": 1.4468,
"step": 38450
},
{
"epoch": 6.5476190476190474,
"grad_norm": 0.4216206669807434,
"learning_rate": 0.0001788556338028169,
"loss": 1.4454,
"step": 38500
},
{
"epoch": 6.5561224489795915,
"grad_norm": 0.4373023211956024,
"learning_rate": 0.00017841549295774647,
"loss": 1.4441,
"step": 38550
},
{
"epoch": 6.564625850340136,
"grad_norm": 0.4680120050907135,
"learning_rate": 0.00017797535211267607,
"loss": 1.4383,
"step": 38600
},
{
"epoch": 6.5731292517006805,
"grad_norm": 0.4241814911365509,
"learning_rate": 0.00017753521126760564,
"loss": 1.4413,
"step": 38650
},
{
"epoch": 6.581632653061225,
"grad_norm": 0.449913889169693,
"learning_rate": 0.00017709507042253523,
"loss": 1.4427,
"step": 38700
},
{
"epoch": 6.590136054421769,
"grad_norm": 0.4140756130218506,
"learning_rate": 0.0001766549295774648,
"loss": 1.4429,
"step": 38750
},
{
"epoch": 6.598639455782313,
"grad_norm": 0.4235837161540985,
"learning_rate": 0.00017621478873239437,
"loss": 1.4425,
"step": 38800
},
{
"epoch": 6.607142857142857,
"grad_norm": 0.4102327227592468,
"learning_rate": 0.00017577464788732394,
"loss": 1.4428,
"step": 38850
},
{
"epoch": 6.615646258503402,
"grad_norm": 0.41676509380340576,
"learning_rate": 0.00017533450704225353,
"loss": 1.4515,
"step": 38900
},
{
"epoch": 6.624149659863946,
"grad_norm": 0.41074737906455994,
"learning_rate": 0.0001748943661971831,
"loss": 1.4376,
"step": 38950
},
{
"epoch": 6.63265306122449,
"grad_norm": 0.4003889262676239,
"learning_rate": 0.00017445422535211267,
"loss": 1.4425,
"step": 39000
},
{
"epoch": 6.63265306122449,
"eval_loss": 1.6100125312805176,
"eval_runtime": 75.3679,
"eval_samples_per_second": 1245.078,
"eval_steps_per_second": 4.869,
"step": 39000
},
{
"epoch": 6.641156462585034,
"grad_norm": 0.4165302813053131,
"learning_rate": 0.00017401408450704226,
"loss": 1.4366,
"step": 39050
},
{
"epoch": 6.649659863945578,
"grad_norm": 0.4187438488006592,
"learning_rate": 0.00017357394366197183,
"loss": 1.4403,
"step": 39100
},
{
"epoch": 6.658163265306122,
"grad_norm": 0.41520532965660095,
"learning_rate": 0.0001731338028169014,
"loss": 1.4465,
"step": 39150
},
{
"epoch": 6.666666666666667,
"grad_norm": 0.47344011068344116,
"learning_rate": 0.00017269366197183097,
"loss": 1.4451,
"step": 39200
},
{
"epoch": 6.675170068027211,
"grad_norm": 0.4409021735191345,
"learning_rate": 0.00017225352112676057,
"loss": 1.4367,
"step": 39250
},
{
"epoch": 6.683673469387755,
"grad_norm": 0.4071614742279053,
"learning_rate": 0.00017181338028169013,
"loss": 1.4343,
"step": 39300
},
{
"epoch": 6.692176870748299,
"grad_norm": 0.43377685546875,
"learning_rate": 0.00017137323943661973,
"loss": 1.4387,
"step": 39350
},
{
"epoch": 6.700680272108843,
"grad_norm": 0.41328275203704834,
"learning_rate": 0.0001709330985915493,
"loss": 1.4308,
"step": 39400
},
{
"epoch": 6.709183673469388,
"grad_norm": 0.4269335865974426,
"learning_rate": 0.0001704929577464789,
"loss": 1.4329,
"step": 39450
},
{
"epoch": 6.717687074829932,
"grad_norm": 0.4136207401752472,
"learning_rate": 0.00017005281690140843,
"loss": 1.4386,
"step": 39500
},
{
"epoch": 6.726190476190476,
"grad_norm": 0.41490909457206726,
"learning_rate": 0.00016961267605633803,
"loss": 1.4348,
"step": 39550
},
{
"epoch": 6.73469387755102,
"grad_norm": 0.41925111413002014,
"learning_rate": 0.0001691725352112676,
"loss": 1.4329,
"step": 39600
},
{
"epoch": 6.743197278911564,
"grad_norm": 0.4161663353443146,
"learning_rate": 0.0001687323943661972,
"loss": 1.4325,
"step": 39650
},
{
"epoch": 6.7517006802721085,
"grad_norm": 0.4175470769405365,
"learning_rate": 0.00016829225352112676,
"loss": 1.4331,
"step": 39700
},
{
"epoch": 6.760204081632653,
"grad_norm": 0.40980637073516846,
"learning_rate": 0.00016785211267605636,
"loss": 1.436,
"step": 39750
},
{
"epoch": 6.7687074829931975,
"grad_norm": 0.4164717197418213,
"learning_rate": 0.00016741197183098592,
"loss": 1.4354,
"step": 39800
},
{
"epoch": 6.7772108843537415,
"grad_norm": 0.41800665855407715,
"learning_rate": 0.00016697183098591552,
"loss": 1.4321,
"step": 39850
},
{
"epoch": 6.785714285714286,
"grad_norm": 0.4796483814716339,
"learning_rate": 0.00016653169014084506,
"loss": 1.428,
"step": 39900
},
{
"epoch": 6.79421768707483,
"grad_norm": 0.40372607111930847,
"learning_rate": 0.00016609154929577466,
"loss": 1.4335,
"step": 39950
},
{
"epoch": 6.802721088435375,
"grad_norm": 0.42244842648506165,
"learning_rate": 0.00016565140845070423,
"loss": 1.4391,
"step": 40000
},
{
"epoch": 6.802721088435375,
"eval_loss": 1.6078286170959473,
"eval_runtime": 75.3844,
"eval_samples_per_second": 1244.806,
"eval_steps_per_second": 4.868,
"step": 40000
},
{
"epoch": 6.811224489795919,
"grad_norm": 0.41039007902145386,
"learning_rate": 0.00016521126760563382,
"loss": 1.4388,
"step": 40050
},
{
"epoch": 6.819727891156463,
"grad_norm": 0.4182026982307434,
"learning_rate": 0.0001647711267605634,
"loss": 1.4335,
"step": 40100
},
{
"epoch": 6.828231292517007,
"grad_norm": 0.4472804665565491,
"learning_rate": 0.00016433098591549296,
"loss": 1.4357,
"step": 40150
},
{
"epoch": 6.836734693877551,
"grad_norm": 0.4294123649597168,
"learning_rate": 0.00016389084507042255,
"loss": 1.4355,
"step": 40200
},
{
"epoch": 6.845238095238095,
"grad_norm": 0.41815003752708435,
"learning_rate": 0.0001634507042253521,
"loss": 1.4324,
"step": 40250
},
{
"epoch": 6.853741496598639,
"grad_norm": 0.4064941704273224,
"learning_rate": 0.0001630105633802817,
"loss": 1.4366,
"step": 40300
},
{
"epoch": 6.862244897959184,
"grad_norm": 0.42009902000427246,
"learning_rate": 0.00016257922535211267,
"loss": 1.4325,
"step": 40350
},
{
"epoch": 6.870748299319728,
"grad_norm": 0.3989739716053009,
"learning_rate": 0.00016213908450704226,
"loss": 1.4238,
"step": 40400
},
{
"epoch": 6.879251700680272,
"grad_norm": 0.4133879840373993,
"learning_rate": 0.00016169894366197183,
"loss": 1.4281,
"step": 40450
},
{
"epoch": 6.887755102040816,
"grad_norm": 0.42547741532325745,
"learning_rate": 0.00016125880281690142,
"loss": 1.4279,
"step": 40500
},
{
"epoch": 6.896258503401361,
"grad_norm": 0.3915853202342987,
"learning_rate": 0.000160818661971831,
"loss": 1.4351,
"step": 40550
},
{
"epoch": 6.904761904761905,
"grad_norm": 0.4148283302783966,
"learning_rate": 0.00016037852112676056,
"loss": 1.4332,
"step": 40600
},
{
"epoch": 6.913265306122449,
"grad_norm": 0.4044345021247864,
"learning_rate": 0.00015993838028169013,
"loss": 1.4304,
"step": 40650
},
{
"epoch": 6.921768707482993,
"grad_norm": 0.4137207269668579,
"learning_rate": 0.00015949823943661972,
"loss": 1.4342,
"step": 40700
},
{
"epoch": 6.930272108843537,
"grad_norm": 0.40803512930870056,
"learning_rate": 0.0001590580985915493,
"loss": 1.4287,
"step": 40750
},
{
"epoch": 6.938775510204081,
"grad_norm": 0.4186759889125824,
"learning_rate": 0.0001586179577464789,
"loss": 1.4269,
"step": 40800
},
{
"epoch": 6.947278911564625,
"grad_norm": 0.5999934077262878,
"learning_rate": 0.00015817781690140846,
"loss": 1.4267,
"step": 40850
},
{
"epoch": 6.95578231292517,
"grad_norm": 3.3650548458099365,
"learning_rate": 0.00015773767605633805,
"loss": 1.4273,
"step": 40900
},
{
"epoch": 6.964285714285714,
"grad_norm": 0.4189580976963043,
"learning_rate": 0.0001572975352112676,
"loss": 1.4246,
"step": 40950
},
{
"epoch": 6.9727891156462585,
"grad_norm": 0.4199207127094269,
"learning_rate": 0.0001568573943661972,
"loss": 1.4264,
"step": 41000
},
{
"epoch": 6.9727891156462585,
"eval_loss": 1.6002153158187866,
"eval_runtime": 75.3647,
"eval_samples_per_second": 1245.132,
"eval_steps_per_second": 4.87,
"step": 41000
},
{
"epoch": 6.9812925170068025,
"grad_norm": 0.4282417595386505,
"learning_rate": 0.00015641725352112676,
"loss": 1.4302,
"step": 41050
},
{
"epoch": 6.989795918367347,
"grad_norm": 0.4040624797344208,
"learning_rate": 0.00015597711267605635,
"loss": 1.4289,
"step": 41100
},
{
"epoch": 6.9982993197278915,
"grad_norm": 0.4131171703338623,
"learning_rate": 0.00015553697183098592,
"loss": 1.4265,
"step": 41150
},
{
"epoch": 7.006802721088436,
"grad_norm": 0.41220322251319885,
"learning_rate": 0.0001550968309859155,
"loss": 1.421,
"step": 41200
},
{
"epoch": 7.01530612244898,
"grad_norm": 0.41320499777793884,
"learning_rate": 0.00015465669014084508,
"loss": 1.4126,
"step": 41250
},
{
"epoch": 7.023809523809524,
"grad_norm": 0.43131372332572937,
"learning_rate": 0.00015421654929577463,
"loss": 1.4185,
"step": 41300
},
{
"epoch": 7.032312925170068,
"grad_norm": 0.41249164938926697,
"learning_rate": 0.00015377640845070422,
"loss": 1.4225,
"step": 41350
},
{
"epoch": 7.040816326530612,
"grad_norm": 0.41785308718681335,
"learning_rate": 0.0001533362676056338,
"loss": 1.4193,
"step": 41400
},
{
"epoch": 7.049319727891157,
"grad_norm": 0.4036597013473511,
"learning_rate": 0.00015289612676056338,
"loss": 1.417,
"step": 41450
},
{
"epoch": 7.057823129251701,
"grad_norm": 1.470516562461853,
"learning_rate": 0.00015245598591549295,
"loss": 1.42,
"step": 41500
},
{
"epoch": 7.066326530612245,
"grad_norm": 0.4129942059516907,
"learning_rate": 0.00015201584507042255,
"loss": 1.4191,
"step": 41550
},
{
"epoch": 7.074829931972789,
"grad_norm": 0.41847845911979675,
"learning_rate": 0.00015157570422535212,
"loss": 1.4146,
"step": 41600
},
{
"epoch": 7.083333333333333,
"grad_norm": 0.4097050428390503,
"learning_rate": 0.0001511355633802817,
"loss": 1.4202,
"step": 41650
},
{
"epoch": 7.091836734693878,
"grad_norm": 0.42190149426460266,
"learning_rate": 0.00015069542253521125,
"loss": 1.4191,
"step": 41700
},
{
"epoch": 7.100340136054422,
"grad_norm": 0.4126221835613251,
"learning_rate": 0.00015025528169014085,
"loss": 1.4157,
"step": 41750
},
{
"epoch": 7.108843537414966,
"grad_norm": 0.42145752906799316,
"learning_rate": 0.00014981514084507042,
"loss": 1.4152,
"step": 41800
},
{
"epoch": 7.11734693877551,
"grad_norm": 0.41415074467658997,
"learning_rate": 0.000149375,
"loss": 1.4233,
"step": 41850
},
{
"epoch": 7.125850340136054,
"grad_norm": 0.4122840166091919,
"learning_rate": 0.00014893485915492958,
"loss": 1.4194,
"step": 41900
},
{
"epoch": 7.134353741496598,
"grad_norm": 0.4208846986293793,
"learning_rate": 0.00014849471830985918,
"loss": 1.417,
"step": 41950
},
{
"epoch": 7.142857142857143,
"grad_norm": 0.41025856137275696,
"learning_rate": 0.00014805457746478874,
"loss": 1.4138,
"step": 42000
},
{
"epoch": 7.142857142857143,
"eval_loss": 1.5856564044952393,
"eval_runtime": 75.3402,
"eval_samples_per_second": 1245.536,
"eval_steps_per_second": 4.871,
"step": 42000
},
{
"epoch": 7.151360544217687,
"grad_norm": 0.4345760941505432,
"learning_rate": 0.0001476144366197183,
"loss": 1.4162,
"step": 42050
},
{
"epoch": 7.159863945578231,
"grad_norm": 0.42329126596450806,
"learning_rate": 0.00014717429577464788,
"loss": 1.4183,
"step": 42100
},
{
"epoch": 7.168367346938775,
"grad_norm": 0.45433273911476135,
"learning_rate": 0.00014673415492957748,
"loss": 1.4209,
"step": 42150
},
{
"epoch": 7.1768707482993195,
"grad_norm": 0.401620477437973,
"learning_rate": 0.00014629401408450704,
"loss": 1.4157,
"step": 42200
},
{
"epoch": 7.1853741496598635,
"grad_norm": 0.4111977219581604,
"learning_rate": 0.00014585387323943664,
"loss": 1.4176,
"step": 42250
},
{
"epoch": 7.1938775510204085,
"grad_norm": 0.4081551134586334,
"learning_rate": 0.0001454137323943662,
"loss": 1.4175,
"step": 42300
},
{
"epoch": 7.2023809523809526,
"grad_norm": 0.41629016399383545,
"learning_rate": 0.0001449735915492958,
"loss": 1.4156,
"step": 42350
},
{
"epoch": 7.210884353741497,
"grad_norm": 0.4202435314655304,
"learning_rate": 0.00014453345070422534,
"loss": 1.4158,
"step": 42400
},
{
"epoch": 7.219387755102041,
"grad_norm": 0.4117106795310974,
"learning_rate": 0.0001440933098591549,
"loss": 1.4128,
"step": 42450
},
{
"epoch": 7.227891156462585,
"grad_norm": 0.40955278277397156,
"learning_rate": 0.0001436531690140845,
"loss": 1.4101,
"step": 42500
},
{
"epoch": 7.23639455782313,
"grad_norm": 0.4062648415565491,
"learning_rate": 0.00014321302816901408,
"loss": 1.4123,
"step": 42550
},
{
"epoch": 7.244897959183674,
"grad_norm": 0.39645469188690186,
"learning_rate": 0.00014277288732394367,
"loss": 1.4122,
"step": 42600
},
{
"epoch": 7.253401360544218,
"grad_norm": 0.4102307856082916,
"learning_rate": 0.00014233274647887324,
"loss": 1.4142,
"step": 42650
},
{
"epoch": 7.261904761904762,
"grad_norm": 0.39169833064079285,
"learning_rate": 0.00014189260563380284,
"loss": 1.4123,
"step": 42700
},
{
"epoch": 7.270408163265306,
"grad_norm": 0.40885522961616516,
"learning_rate": 0.00014145246478873238,
"loss": 1.4099,
"step": 42750
},
{
"epoch": 7.27891156462585,
"grad_norm": 0.40849608182907104,
"learning_rate": 0.00014101232394366197,
"loss": 1.4112,
"step": 42800
},
{
"epoch": 7.287414965986395,
"grad_norm": 0.416453093290329,
"learning_rate": 0.00014057218309859154,
"loss": 1.4139,
"step": 42850
},
{
"epoch": 7.295918367346939,
"grad_norm": 0.40005016326904297,
"learning_rate": 0.00014013204225352114,
"loss": 1.4122,
"step": 42900
},
{
"epoch": 7.304421768707483,
"grad_norm": 0.40970712900161743,
"learning_rate": 0.0001396919014084507,
"loss": 1.4071,
"step": 42950
},
{
"epoch": 7.312925170068027,
"grad_norm": 0.3983543813228607,
"learning_rate": 0.0001392517605633803,
"loss": 1.4095,
"step": 43000
},
{
"epoch": 7.312925170068027,
"eval_loss": 1.581950068473816,
"eval_runtime": 75.8081,
"eval_samples_per_second": 1237.849,
"eval_steps_per_second": 4.841,
"step": 43000
},
{
"epoch": 7.321428571428571,
"grad_norm": 0.42998814582824707,
"learning_rate": 0.00013881161971830987,
"loss": 1.4165,
"step": 43050
},
{
"epoch": 7.329931972789115,
"grad_norm": 0.40602609515190125,
"learning_rate": 0.00013837147887323944,
"loss": 1.4079,
"step": 43100
},
{
"epoch": 7.33843537414966,
"grad_norm": 0.4095648229122162,
"learning_rate": 0.000137931338028169,
"loss": 1.4155,
"step": 43150
},
{
"epoch": 7.346938775510204,
"grad_norm": 0.5845966339111328,
"learning_rate": 0.0001374911971830986,
"loss": 1.3978,
"step": 43200
},
{
"epoch": 7.355442176870748,
"grad_norm": 0.39307621121406555,
"learning_rate": 0.00013705105633802817,
"loss": 1.4063,
"step": 43250
},
{
"epoch": 7.363945578231292,
"grad_norm": 0.41059941053390503,
"learning_rate": 0.00013661091549295776,
"loss": 1.4111,
"step": 43300
},
{
"epoch": 7.372448979591836,
"grad_norm": 0.4473901689052582,
"learning_rate": 0.00013617077464788733,
"loss": 1.4107,
"step": 43350
},
{
"epoch": 7.380952380952381,
"grad_norm": 0.4176762104034424,
"learning_rate": 0.00013573063380281693,
"loss": 1.4074,
"step": 43400
},
{
"epoch": 7.389455782312925,
"grad_norm": 0.40306830406188965,
"learning_rate": 0.00013529049295774647,
"loss": 1.4018,
"step": 43450
},
{
"epoch": 7.3979591836734695,
"grad_norm": 0.40745970606803894,
"learning_rate": 0.00013485035211267606,
"loss": 1.4094,
"step": 43500
},
{
"epoch": 7.406462585034014,
"grad_norm": 0.4084283411502838,
"learning_rate": 0.00013441021126760563,
"loss": 1.4065,
"step": 43550
},
{
"epoch": 7.414965986394558,
"grad_norm": 0.3982578217983246,
"learning_rate": 0.0001339700704225352,
"loss": 1.4058,
"step": 43600
},
{
"epoch": 7.423469387755102,
"grad_norm": 0.40555986762046814,
"learning_rate": 0.0001335299295774648,
"loss": 1.4071,
"step": 43650
},
{
"epoch": 7.431972789115647,
"grad_norm": 0.41015100479125977,
"learning_rate": 0.00013308978873239436,
"loss": 1.4055,
"step": 43700
},
{
"epoch": 7.440476190476191,
"grad_norm": 0.39785054326057434,
"learning_rate": 0.00013264964788732396,
"loss": 1.4052,
"step": 43750
},
{
"epoch": 7.448979591836735,
"grad_norm": 0.4046226441860199,
"learning_rate": 0.0001322095070422535,
"loss": 1.4099,
"step": 43800
},
{
"epoch": 7.457482993197279,
"grad_norm": 0.4013233482837677,
"learning_rate": 0.0001317693661971831,
"loss": 1.4105,
"step": 43850
},
{
"epoch": 7.465986394557823,
"grad_norm": 0.42137226462364197,
"learning_rate": 0.00013132922535211266,
"loss": 1.4074,
"step": 43900
},
{
"epoch": 7.474489795918368,
"grad_norm": 0.4147741496562958,
"learning_rate": 0.00013088908450704226,
"loss": 1.4055,
"step": 43950
},
{
"epoch": 7.482993197278912,
"grad_norm": 0.43600887060165405,
"learning_rate": 0.00013044894366197183,
"loss": 1.4049,
"step": 44000
},
{
"epoch": 7.482993197278912,
"eval_loss": 1.575681209564209,
"eval_runtime": 75.3426,
"eval_samples_per_second": 1245.498,
"eval_steps_per_second": 4.871,
"step": 44000
},
{
"epoch": 7.491496598639456,
"grad_norm": 0.40329524874687195,
"learning_rate": 0.00013000880281690142,
"loss": 1.406,
"step": 44050
},
{
"epoch": 7.5,
"grad_norm": 0.39978840947151184,
"learning_rate": 0.000129568661971831,
"loss": 1.4051,
"step": 44100
},
{
"epoch": 7.508503401360544,
"grad_norm": 0.4256359934806824,
"learning_rate": 0.00012912852112676056,
"loss": 1.4096,
"step": 44150
},
{
"epoch": 7.517006802721088,
"grad_norm": 0.41504326462745667,
"learning_rate": 0.00012868838028169013,
"loss": 1.411,
"step": 44200
},
{
"epoch": 7.525510204081632,
"grad_norm": 0.41557520627975464,
"learning_rate": 0.00012824823943661972,
"loss": 1.3988,
"step": 44250
},
{
"epoch": 7.534013605442177,
"grad_norm": 0.654144287109375,
"learning_rate": 0.0001278080985915493,
"loss": 1.4002,
"step": 44300
},
{
"epoch": 7.542517006802721,
"grad_norm": 0.40340688824653625,
"learning_rate": 0.0001273679577464789,
"loss": 1.407,
"step": 44350
},
{
"epoch": 7.551020408163265,
"grad_norm": 0.4111216068267822,
"learning_rate": 0.00012692781690140846,
"loss": 1.4051,
"step": 44400
},
{
"epoch": 7.559523809523809,
"grad_norm": 0.40301740169525146,
"learning_rate": 0.00012648767605633805,
"loss": 1.4063,
"step": 44450
},
{
"epoch": 7.568027210884353,
"grad_norm": 0.40716397762298584,
"learning_rate": 0.0001260475352112676,
"loss": 1.4026,
"step": 44500
},
{
"epoch": 7.576530612244898,
"grad_norm": 0.40946149826049805,
"learning_rate": 0.0001256073943661972,
"loss": 1.3999,
"step": 44550
},
{
"epoch": 7.585034013605442,
"grad_norm": 0.4165254533290863,
"learning_rate": 0.00012516725352112676,
"loss": 1.4061,
"step": 44600
},
{
"epoch": 7.593537414965986,
"grad_norm": 0.4061264097690582,
"learning_rate": 0.00012472711267605635,
"loss": 1.3952,
"step": 44650
},
{
"epoch": 7.6020408163265305,
"grad_norm": 0.4414292573928833,
"learning_rate": 0.00012428697183098592,
"loss": 1.4061,
"step": 44700
},
{
"epoch": 7.610544217687075,
"grad_norm": 0.39658012986183167,
"learning_rate": 0.0001238468309859155,
"loss": 1.3985,
"step": 44750
},
{
"epoch": 7.619047619047619,
"grad_norm": 0.7822299599647522,
"learning_rate": 0.00012340669014084508,
"loss": 1.3993,
"step": 44800
},
{
"epoch": 7.627551020408164,
"grad_norm": 0.3982444703578949,
"learning_rate": 0.00012296654929577465,
"loss": 1.4011,
"step": 44850
},
{
"epoch": 7.636054421768708,
"grad_norm": 0.40807053446769714,
"learning_rate": 0.00012252640845070422,
"loss": 1.3984,
"step": 44900
},
{
"epoch": 7.644557823129252,
"grad_norm": 0.4086779057979584,
"learning_rate": 0.00012208626760563382,
"loss": 1.403,
"step": 44950
},
{
"epoch": 7.653061224489796,
"grad_norm": 0.4078817367553711,
"learning_rate": 0.00012164612676056338,
"loss": 1.3974,
"step": 45000
},
{
"epoch": 7.653061224489796,
"eval_loss": 1.564176082611084,
"eval_runtime": 75.8799,
"eval_samples_per_second": 1236.678,
"eval_steps_per_second": 4.837,
"step": 45000
},
{
"epoch": 7.66156462585034,
"grad_norm": 0.4033880829811096,
"learning_rate": 0.00012120598591549297,
"loss": 1.396,
"step": 45050
},
{
"epoch": 7.670068027210885,
"grad_norm": 0.41422680020332336,
"learning_rate": 0.00012076584507042255,
"loss": 1.3999,
"step": 45100
},
{
"epoch": 7.678571428571429,
"grad_norm": 0.41265836358070374,
"learning_rate": 0.00012032570422535212,
"loss": 1.3974,
"step": 45150
},
{
"epoch": 7.687074829931973,
"grad_norm": 0.4050711989402771,
"learning_rate": 0.0001198855633802817,
"loss": 1.3992,
"step": 45200
},
{
"epoch": 7.695578231292517,
"grad_norm": 0.4541753828525543,
"learning_rate": 0.00011944542253521128,
"loss": 1.3956,
"step": 45250
},
{
"epoch": 7.704081632653061,
"grad_norm": 0.469691663980484,
"learning_rate": 0.00011900528169014085,
"loss": 1.3941,
"step": 45300
},
{
"epoch": 7.712585034013605,
"grad_norm": 0.4149336516857147,
"learning_rate": 0.00011856514084507042,
"loss": 1.3926,
"step": 45350
},
{
"epoch": 7.72108843537415,
"grad_norm": 0.3968436121940613,
"learning_rate": 0.000118125,
"loss": 1.3945,
"step": 45400
},
{
"epoch": 7.729591836734694,
"grad_norm": 0.4118001461029053,
"learning_rate": 0.00011768485915492958,
"loss": 1.3945,
"step": 45450
},
{
"epoch": 7.738095238095238,
"grad_norm": 0.4041975438594818,
"learning_rate": 0.00011724471830985915,
"loss": 1.3958,
"step": 45500
},
{
"epoch": 7.746598639455782,
"grad_norm": 0.4147978723049164,
"learning_rate": 0.00011680457746478873,
"loss": 1.3885,
"step": 45550
},
{
"epoch": 7.755102040816326,
"grad_norm": 0.4063569903373718,
"learning_rate": 0.00011636443661971831,
"loss": 1.3955,
"step": 45600
},
{
"epoch": 7.763605442176871,
"grad_norm": 0.4077317416667938,
"learning_rate": 0.00011592429577464788,
"loss": 1.3969,
"step": 45650
},
{
"epoch": 7.772108843537415,
"grad_norm": 0.4094686508178711,
"learning_rate": 0.00011548415492957746,
"loss": 1.3956,
"step": 45700
},
{
"epoch": 7.780612244897959,
"grad_norm": 0.46225419640541077,
"learning_rate": 0.00011504401408450704,
"loss": 1.3964,
"step": 45750
},
{
"epoch": 7.789115646258503,
"grad_norm": 0.40823882818222046,
"learning_rate": 0.00011460387323943663,
"loss": 1.3865,
"step": 45800
},
{
"epoch": 7.7976190476190474,
"grad_norm": 0.4115474224090576,
"learning_rate": 0.0001141637323943662,
"loss": 1.3946,
"step": 45850
},
{
"epoch": 7.8061224489795915,
"grad_norm": 0.4270274043083191,
"learning_rate": 0.00011372359154929578,
"loss": 1.3961,
"step": 45900
},
{
"epoch": 7.814625850340136,
"grad_norm": 0.4096102714538574,
"learning_rate": 0.00011328345070422536,
"loss": 1.3909,
"step": 45950
},
{
"epoch": 7.8231292517006805,
"grad_norm": 0.41915324330329895,
"learning_rate": 0.00011284330985915494,
"loss": 1.3942,
"step": 46000
},
{
"epoch": 7.8231292517006805,
"eval_loss": 1.5574978590011597,
"eval_runtime": 75.3679,
"eval_samples_per_second": 1245.08,
"eval_steps_per_second": 4.869,
"step": 46000
},
{
"epoch": 7.831632653061225,
"grad_norm": 0.42200446128845215,
"learning_rate": 0.00011241197183098592,
"loss": 1.3898,
"step": 46050
},
{
"epoch": 7.840136054421769,
"grad_norm": 0.40634581446647644,
"learning_rate": 0.0001119718309859155,
"loss": 1.3872,
"step": 46100
},
{
"epoch": 7.848639455782313,
"grad_norm": 0.40481993556022644,
"learning_rate": 0.00011153169014084508,
"loss": 1.394,
"step": 46150
},
{
"epoch": 7.857142857142857,
"grad_norm": 0.8925232887268066,
"learning_rate": 0.00011109154929577466,
"loss": 1.3921,
"step": 46200
},
{
"epoch": 7.865646258503402,
"grad_norm": 0.5260804295539856,
"learning_rate": 0.00011065140845070423,
"loss": 1.3892,
"step": 46250
},
{
"epoch": 7.874149659863946,
"grad_norm": 0.4104487895965576,
"learning_rate": 0.00011021126760563381,
"loss": 1.3966,
"step": 46300
},
{
"epoch": 7.88265306122449,
"grad_norm": 0.4222128391265869,
"learning_rate": 0.0001097711267605634,
"loss": 1.3911,
"step": 46350
},
{
"epoch": 7.891156462585034,
"grad_norm": 0.4017203748226166,
"learning_rate": 0.00010933098591549295,
"loss": 1.3858,
"step": 46400
},
{
"epoch": 7.899659863945578,
"grad_norm": 0.41633373498916626,
"learning_rate": 0.00010889084507042253,
"loss": 1.3888,
"step": 46450
},
{
"epoch": 7.908163265306122,
"grad_norm": 0.41837745904922485,
"learning_rate": 0.00010845070422535211,
"loss": 1.3827,
"step": 46500
},
{
"epoch": 7.916666666666667,
"grad_norm": 0.4108099341392517,
"learning_rate": 0.0001080105633802817,
"loss": 1.3867,
"step": 46550
},
{
"epoch": 7.925170068027211,
"grad_norm": 0.4089981019496918,
"learning_rate": 0.00010757042253521126,
"loss": 1.3883,
"step": 46600
},
{
"epoch": 7.933673469387755,
"grad_norm": 0.40665385127067566,
"learning_rate": 0.00010713028169014084,
"loss": 1.3867,
"step": 46650
},
{
"epoch": 7.942176870748299,
"grad_norm": 0.4587544798851013,
"learning_rate": 0.00010669014084507043,
"loss": 1.3887,
"step": 46700
},
{
"epoch": 7.950680272108843,
"grad_norm": 0.41734760999679565,
"learning_rate": 0.00010625,
"loss": 1.3865,
"step": 46750
},
{
"epoch": 7.959183673469388,
"grad_norm": 0.3909563422203064,
"learning_rate": 0.00010580985915492958,
"loss": 1.3826,
"step": 46800
},
{
"epoch": 7.967687074829932,
"grad_norm": 0.4237678647041321,
"learning_rate": 0.00010536971830985916,
"loss": 1.3844,
"step": 46850
},
{
"epoch": 7.976190476190476,
"grad_norm": 0.3990273177623749,
"learning_rate": 0.00010492957746478874,
"loss": 1.3834,
"step": 46900
},
{
"epoch": 7.98469387755102,
"grad_norm": 0.4431270360946655,
"learning_rate": 0.00010448943661971831,
"loss": 1.3829,
"step": 46950
},
{
"epoch": 7.993197278911564,
"grad_norm": 0.423221230506897,
"learning_rate": 0.00010404929577464789,
"loss": 1.3861,
"step": 47000
},
{
"epoch": 7.993197278911564,
"eval_loss": 1.5540677309036255,
"eval_runtime": 75.3265,
"eval_samples_per_second": 1245.764,
"eval_steps_per_second": 4.872,
"step": 47000
},
{
"epoch": 8.001700680272108,
"grad_norm": 0.4330544173717499,
"learning_rate": 0.00010360915492957747,
"loss": 1.3834,
"step": 47050
},
{
"epoch": 8.010204081632653,
"grad_norm": 0.4099094867706299,
"learning_rate": 0.00010316901408450704,
"loss": 1.3826,
"step": 47100
},
{
"epoch": 8.018707482993197,
"grad_norm": 0.41134965419769287,
"learning_rate": 0.00010272887323943662,
"loss": 1.3846,
"step": 47150
},
{
"epoch": 8.02721088435374,
"grad_norm": 0.4183215796947479,
"learning_rate": 0.0001022887323943662,
"loss": 1.379,
"step": 47200
},
{
"epoch": 8.035714285714286,
"grad_norm": 0.39894890785217285,
"learning_rate": 0.00010184859154929579,
"loss": 1.3839,
"step": 47250
},
{
"epoch": 8.04421768707483,
"grad_norm": 0.4777671992778778,
"learning_rate": 0.00010140845070422535,
"loss": 1.3789,
"step": 47300
},
{
"epoch": 8.052721088435375,
"grad_norm": 0.41188567876815796,
"learning_rate": 0.00010096830985915494,
"loss": 1.3848,
"step": 47350
},
{
"epoch": 8.061224489795919,
"grad_norm": 0.4015198349952698,
"learning_rate": 0.00010052816901408452,
"loss": 1.3883,
"step": 47400
},
{
"epoch": 8.069727891156463,
"grad_norm": 0.401507169008255,
"learning_rate": 0.00010008802816901409,
"loss": 1.381,
"step": 47450
},
{
"epoch": 8.078231292517007,
"grad_norm": 0.40761667490005493,
"learning_rate": 9.964788732394367e-05,
"loss": 1.3714,
"step": 47500
},
{
"epoch": 8.08673469387755,
"grad_norm": 0.427236407995224,
"learning_rate": 9.920774647887325e-05,
"loss": 1.3799,
"step": 47550
},
{
"epoch": 8.095238095238095,
"grad_norm": 0.40780577063560486,
"learning_rate": 9.876760563380282e-05,
"loss": 1.3762,
"step": 47600
},
{
"epoch": 8.103741496598639,
"grad_norm": 0.4071747064590454,
"learning_rate": 9.832746478873239e-05,
"loss": 1.3815,
"step": 47650
},
{
"epoch": 8.112244897959183,
"grad_norm": 0.40746966004371643,
"learning_rate": 9.788732394366197e-05,
"loss": 1.3794,
"step": 47700
},
{
"epoch": 8.120748299319727,
"grad_norm": 0.39780792593955994,
"learning_rate": 9.744718309859155e-05,
"loss": 1.3804,
"step": 47750
},
{
"epoch": 8.129251700680273,
"grad_norm": 0.4328888952732086,
"learning_rate": 9.700704225352112e-05,
"loss": 1.3792,
"step": 47800
},
{
"epoch": 8.137755102040817,
"grad_norm": 0.40212351083755493,
"learning_rate": 9.65669014084507e-05,
"loss": 1.3702,
"step": 47850
},
{
"epoch": 8.146258503401361,
"grad_norm": 0.3966144919395447,
"learning_rate": 9.612676056338028e-05,
"loss": 1.3761,
"step": 47900
},
{
"epoch": 8.154761904761905,
"grad_norm": 0.41284072399139404,
"learning_rate": 9.568661971830986e-05,
"loss": 1.3774,
"step": 47950
},
{
"epoch": 8.16326530612245,
"grad_norm": 0.41598987579345703,
"learning_rate": 9.524647887323943e-05,
"loss": 1.3746,
"step": 48000
},
{
"epoch": 8.16326530612245,
"eval_loss": 1.544783115386963,
"eval_runtime": 75.374,
"eval_samples_per_second": 1244.979,
"eval_steps_per_second": 4.869,
"step": 48000
},
{
"epoch": 8.171768707482993,
"grad_norm": 0.7141363024711609,
"learning_rate": 9.480633802816901e-05,
"loss": 1.378,
"step": 48050
},
{
"epoch": 8.180272108843537,
"grad_norm": 0.400291383266449,
"learning_rate": 9.43661971830986e-05,
"loss": 1.3805,
"step": 48100
},
{
"epoch": 8.188775510204081,
"grad_norm": 0.40866729617118835,
"learning_rate": 9.392605633802818e-05,
"loss": 1.3765,
"step": 48150
},
{
"epoch": 8.197278911564625,
"grad_norm": 0.39690279960632324,
"learning_rate": 9.348591549295775e-05,
"loss": 1.3747,
"step": 48200
},
{
"epoch": 8.20578231292517,
"grad_norm": 0.41751721501350403,
"learning_rate": 9.304577464788733e-05,
"loss": 1.376,
"step": 48250
},
{
"epoch": 8.214285714285714,
"grad_norm": 0.4105045795440674,
"learning_rate": 9.260563380281691e-05,
"loss": 1.383,
"step": 48300
},
{
"epoch": 8.22278911564626,
"grad_norm": 0.40793538093566895,
"learning_rate": 9.216549295774648e-05,
"loss": 1.3716,
"step": 48350
},
{
"epoch": 8.231292517006803,
"grad_norm": 0.746793270111084,
"learning_rate": 9.172535211267606e-05,
"loss": 1.3736,
"step": 48400
},
{
"epoch": 8.239795918367347,
"grad_norm": 0.41186049580574036,
"learning_rate": 9.128521126760564e-05,
"loss": 1.3751,
"step": 48450
},
{
"epoch": 8.248299319727892,
"grad_norm": 0.4152040481567383,
"learning_rate": 9.084507042253522e-05,
"loss": 1.3716,
"step": 48500
},
{
"epoch": 8.256802721088436,
"grad_norm": 0.4200298488140106,
"learning_rate": 9.040492957746479e-05,
"loss": 1.3762,
"step": 48550
},
{
"epoch": 8.26530612244898,
"grad_norm": 0.40081730484962463,
"learning_rate": 8.996478873239437e-05,
"loss": 1.371,
"step": 48600
},
{
"epoch": 8.273809523809524,
"grad_norm": 0.4015451967716217,
"learning_rate": 8.952464788732396e-05,
"loss": 1.3717,
"step": 48650
},
{
"epoch": 8.282312925170068,
"grad_norm": 0.41303500533103943,
"learning_rate": 8.908450704225352e-05,
"loss": 1.3793,
"step": 48700
},
{
"epoch": 8.290816326530612,
"grad_norm": 0.41865503787994385,
"learning_rate": 8.86443661971831e-05,
"loss": 1.3733,
"step": 48750
},
{
"epoch": 8.299319727891156,
"grad_norm": 0.39655590057373047,
"learning_rate": 8.820422535211267e-05,
"loss": 1.371,
"step": 48800
},
{
"epoch": 8.3078231292517,
"grad_norm": 0.4100252389907837,
"learning_rate": 8.776408450704226e-05,
"loss": 1.3778,
"step": 48850
},
{
"epoch": 8.316326530612244,
"grad_norm": 0.4039982259273529,
"learning_rate": 8.732394366197182e-05,
"loss": 1.3737,
"step": 48900
},
{
"epoch": 8.32482993197279,
"grad_norm": 0.40856197476387024,
"learning_rate": 8.68838028169014e-05,
"loss": 1.3687,
"step": 48950
},
{
"epoch": 8.333333333333334,
"grad_norm": 0.40824389457702637,
"learning_rate": 8.644366197183099e-05,
"loss": 1.3726,
"step": 49000
},
{
"epoch": 8.333333333333334,
"eval_loss": 1.5451842546463013,
"eval_runtime": 75.3658,
"eval_samples_per_second": 1245.114,
"eval_steps_per_second": 4.87,
"step": 49000
},
{
"epoch": 8.341836734693878,
"grad_norm": 0.40559035539627075,
"learning_rate": 8.600352112676056e-05,
"loss": 1.3714,
"step": 49050
},
{
"epoch": 8.350340136054422,
"grad_norm": 0.40669360756874084,
"learning_rate": 8.556338028169014e-05,
"loss": 1.3716,
"step": 49100
},
{
"epoch": 8.358843537414966,
"grad_norm": 0.40370744466781616,
"learning_rate": 8.513204225352113e-05,
"loss": 1.37,
"step": 49150
},
{
"epoch": 8.36734693877551,
"grad_norm": 0.5396425127983093,
"learning_rate": 8.469190140845071e-05,
"loss": 1.3769,
"step": 49200
},
{
"epoch": 8.375850340136054,
"grad_norm": 0.401902973651886,
"learning_rate": 8.425176056338028e-05,
"loss": 1.364,
"step": 49250
},
{
"epoch": 8.384353741496598,
"grad_norm": 0.7124184966087341,
"learning_rate": 8.381161971830986e-05,
"loss": 1.3658,
"step": 49300
},
{
"epoch": 8.392857142857142,
"grad_norm": 0.4373975396156311,
"learning_rate": 8.337147887323944e-05,
"loss": 1.3658,
"step": 49350
},
{
"epoch": 8.401360544217686,
"grad_norm": 0.40361514687538147,
"learning_rate": 8.293133802816902e-05,
"loss": 1.366,
"step": 49400
},
{
"epoch": 8.40986394557823,
"grad_norm": 0.4027824103832245,
"learning_rate": 8.249119718309859e-05,
"loss": 1.3726,
"step": 49450
},
{
"epoch": 8.418367346938776,
"grad_norm": 0.41340282559394836,
"learning_rate": 8.205105633802817e-05,
"loss": 1.3691,
"step": 49500
},
{
"epoch": 8.42687074829932,
"grad_norm": 0.41044774651527405,
"learning_rate": 8.161091549295776e-05,
"loss": 1.3715,
"step": 49550
},
{
"epoch": 8.435374149659864,
"grad_norm": 0.7769300937652588,
"learning_rate": 8.117077464788732e-05,
"loss": 1.3604,
"step": 49600
},
{
"epoch": 8.443877551020408,
"grad_norm": 0.3973638713359833,
"learning_rate": 8.07306338028169e-05,
"loss": 1.3692,
"step": 49650
},
{
"epoch": 8.452380952380953,
"grad_norm": 0.3998255133628845,
"learning_rate": 8.029049295774649e-05,
"loss": 1.3758,
"step": 49700
},
{
"epoch": 8.460884353741497,
"grad_norm": 0.4052984416484833,
"learning_rate": 7.985035211267607e-05,
"loss": 1.3695,
"step": 49750
},
{
"epoch": 8.46938775510204,
"grad_norm": 0.40513521432876587,
"learning_rate": 7.941021126760564e-05,
"loss": 1.365,
"step": 49800
},
{
"epoch": 8.477891156462585,
"grad_norm": 0.41850411891937256,
"learning_rate": 7.89700704225352e-05,
"loss": 1.371,
"step": 49850
},
{
"epoch": 8.486394557823129,
"grad_norm": 0.40076538920402527,
"learning_rate": 7.852992957746479e-05,
"loss": 1.3583,
"step": 49900
},
{
"epoch": 8.494897959183673,
"grad_norm": 0.40688443183898926,
"learning_rate": 7.808978873239436e-05,
"loss": 1.368,
"step": 49950
},
{
"epoch": 8.503401360544217,
"grad_norm": 0.399404913187027,
"learning_rate": 7.764964788732394e-05,
"loss": 1.3683,
"step": 50000
},
{
"epoch": 8.503401360544217,
"eval_loss": 1.5378692150115967,
"eval_runtime": 75.368,
"eval_samples_per_second": 1245.078,
"eval_steps_per_second": 4.869,
"step": 50000
},
{
"epoch": 8.511904761904763,
"grad_norm": 0.40362656116485596,
"learning_rate": 7.720950704225352e-05,
"loss": 1.3642,
"step": 50050
},
{
"epoch": 8.520408163265307,
"grad_norm": 0.41159501671791077,
"learning_rate": 7.67693661971831e-05,
"loss": 1.3559,
"step": 50100
},
{
"epoch": 8.52891156462585,
"grad_norm": 0.4064503014087677,
"learning_rate": 7.632922535211267e-05,
"loss": 1.3689,
"step": 50150
},
{
"epoch": 8.537414965986395,
"grad_norm": 0.3967163562774658,
"learning_rate": 7.588908450704225e-05,
"loss": 1.3645,
"step": 50200
},
{
"epoch": 8.545918367346939,
"grad_norm": 0.40805837512016296,
"learning_rate": 7.544894366197183e-05,
"loss": 1.3617,
"step": 50250
},
{
"epoch": 8.554421768707483,
"grad_norm": 0.39720582962036133,
"learning_rate": 7.500880281690142e-05,
"loss": 1.3606,
"step": 50300
},
{
"epoch": 8.562925170068027,
"grad_norm": 0.3969172537326813,
"learning_rate": 7.456866197183098e-05,
"loss": 1.3661,
"step": 50350
},
{
"epoch": 8.571428571428571,
"grad_norm": 0.40754255652427673,
"learning_rate": 7.412852112676057e-05,
"loss": 1.3674,
"step": 50400
},
{
"epoch": 8.579931972789115,
"grad_norm": 0.40162739157676697,
"learning_rate": 7.368838028169015e-05,
"loss": 1.3654,
"step": 50450
},
{
"epoch": 8.58843537414966,
"grad_norm": 0.4004862606525421,
"learning_rate": 7.324823943661972e-05,
"loss": 1.3567,
"step": 50500
},
{
"epoch": 8.596938775510203,
"grad_norm": 0.4237268269062042,
"learning_rate": 7.28080985915493e-05,
"loss": 1.3617,
"step": 50550
},
{
"epoch": 8.60544217687075,
"grad_norm": 0.4031754434108734,
"learning_rate": 7.236795774647888e-05,
"loss": 1.3597,
"step": 50600
},
{
"epoch": 8.613945578231293,
"grad_norm": 0.39399534463882446,
"learning_rate": 7.192781690140846e-05,
"loss": 1.3633,
"step": 50650
},
{
"epoch": 8.622448979591837,
"grad_norm": 0.4071154296398163,
"learning_rate": 7.148767605633803e-05,
"loss": 1.3634,
"step": 50700
},
{
"epoch": 8.630952380952381,
"grad_norm": 0.40059253573417664,
"learning_rate": 7.104753521126761e-05,
"loss": 1.3674,
"step": 50750
},
{
"epoch": 8.639455782312925,
"grad_norm": 0.4087289571762085,
"learning_rate": 7.060739436619719e-05,
"loss": 1.3646,
"step": 50800
},
{
"epoch": 8.64795918367347,
"grad_norm": 0.42607581615448,
"learning_rate": 7.016725352112676e-05,
"loss": 1.3645,
"step": 50850
},
{
"epoch": 8.656462585034014,
"grad_norm": 0.40754997730255127,
"learning_rate": 6.972711267605634e-05,
"loss": 1.3582,
"step": 50900
},
{
"epoch": 8.664965986394558,
"grad_norm": 0.4103423058986664,
"learning_rate": 6.928697183098592e-05,
"loss": 1.3635,
"step": 50950
},
{
"epoch": 8.673469387755102,
"grad_norm": 0.4093490540981293,
"learning_rate": 6.88468309859155e-05,
"loss": 1.3593,
"step": 51000
},
{
"epoch": 8.673469387755102,
"eval_loss": 1.5301626920700073,
"eval_runtime": 75.3714,
"eval_samples_per_second": 1245.021,
"eval_steps_per_second": 4.869,
"step": 51000
},
{
"epoch": 8.681972789115646,
"grad_norm": 0.403463751077652,
"learning_rate": 6.840669014084506e-05,
"loss": 1.3596,
"step": 51050
},
{
"epoch": 8.69047619047619,
"grad_norm": 0.406222403049469,
"learning_rate": 6.796654929577464e-05,
"loss": 1.3592,
"step": 51100
},
{
"epoch": 8.698979591836736,
"grad_norm": 0.41207653284072876,
"learning_rate": 6.752640845070423e-05,
"loss": 1.3574,
"step": 51150
},
{
"epoch": 8.70748299319728,
"grad_norm": 0.4042452573776245,
"learning_rate": 6.70862676056338e-05,
"loss": 1.3714,
"step": 51200
},
{
"epoch": 8.715986394557824,
"grad_norm": 0.3971654176712036,
"learning_rate": 6.664612676056338e-05,
"loss": 1.3557,
"step": 51250
},
{
"epoch": 8.724489795918368,
"grad_norm": 0.40285438299179077,
"learning_rate": 6.620598591549296e-05,
"loss": 1.3587,
"step": 51300
},
{
"epoch": 8.732993197278912,
"grad_norm": 0.4234530031681061,
"learning_rate": 6.576584507042254e-05,
"loss": 1.3682,
"step": 51350
},
{
"epoch": 8.741496598639456,
"grad_norm": 0.3918672502040863,
"learning_rate": 6.532570422535211e-05,
"loss": 1.3545,
"step": 51400
},
{
"epoch": 8.75,
"grad_norm": 0.4000791013240814,
"learning_rate": 6.488556338028169e-05,
"loss": 1.361,
"step": 51450
},
{
"epoch": 8.758503401360544,
"grad_norm": 0.41339510679244995,
"learning_rate": 6.444542253521127e-05,
"loss": 1.3595,
"step": 51500
},
{
"epoch": 8.767006802721088,
"grad_norm": 0.4353317618370056,
"learning_rate": 6.400528169014084e-05,
"loss": 1.3601,
"step": 51550
},
{
"epoch": 8.775510204081632,
"grad_norm": 0.39730769395828247,
"learning_rate": 6.356514084507042e-05,
"loss": 1.3547,
"step": 51600
},
{
"epoch": 8.784013605442176,
"grad_norm": 0.41856640577316284,
"learning_rate": 6.3125e-05,
"loss": 1.354,
"step": 51650
},
{
"epoch": 8.79251700680272,
"grad_norm": 0.40725892782211304,
"learning_rate": 6.268485915492958e-05,
"loss": 1.3609,
"step": 51700
},
{
"epoch": 8.801020408163264,
"grad_norm": 0.4020816385746002,
"learning_rate": 6.224471830985915e-05,
"loss": 1.3507,
"step": 51750
},
{
"epoch": 8.80952380952381,
"grad_norm": 0.4009547233581543,
"learning_rate": 6.180457746478873e-05,
"loss": 1.3588,
"step": 51800
},
{
"epoch": 8.818027210884354,
"grad_norm": 0.42511284351348877,
"learning_rate": 6.136443661971832e-05,
"loss": 1.3567,
"step": 51850
},
{
"epoch": 8.826530612244898,
"grad_norm": 0.4135643541812897,
"learning_rate": 6.0924295774647885e-05,
"loss": 1.3536,
"step": 51900
},
{
"epoch": 8.835034013605442,
"grad_norm": 0.39955195784568787,
"learning_rate": 6.048415492957746e-05,
"loss": 1.3564,
"step": 51950
},
{
"epoch": 8.843537414965986,
"grad_norm": 0.4043436348438263,
"learning_rate": 6.004401408450704e-05,
"loss": 1.3527,
"step": 52000
},
{
"epoch": 8.843537414965986,
"eval_loss": 1.521606206893921,
"eval_runtime": 75.3981,
"eval_samples_per_second": 1244.58,
"eval_steps_per_second": 4.867,
"step": 52000
},
{
"epoch": 8.85204081632653,
"grad_norm": 0.40607118606567383,
"learning_rate": 5.960387323943662e-05,
"loss": 1.3567,
"step": 52050
},
{
"epoch": 8.860544217687075,
"grad_norm": 0.4122581481933594,
"learning_rate": 5.91637323943662e-05,
"loss": 1.355,
"step": 52100
},
{
"epoch": 8.869047619047619,
"grad_norm": 0.4980255961418152,
"learning_rate": 5.8723591549295774e-05,
"loss": 1.3523,
"step": 52150
},
{
"epoch": 8.877551020408163,
"grad_norm": 0.4053143262863159,
"learning_rate": 5.8283450704225356e-05,
"loss": 1.3556,
"step": 52200
},
{
"epoch": 8.886054421768707,
"grad_norm": 0.40937647223472595,
"learning_rate": 5.785211267605634e-05,
"loss": 1.3597,
"step": 52250
},
{
"epoch": 8.89455782312925,
"grad_norm": 0.4028383195400238,
"learning_rate": 5.741197183098592e-05,
"loss": 1.3548,
"step": 52300
},
{
"epoch": 8.903061224489797,
"grad_norm": 0.40696024894714355,
"learning_rate": 5.6971830985915496e-05,
"loss": 1.3553,
"step": 52350
},
{
"epoch": 8.91156462585034,
"grad_norm": 0.4004703164100647,
"learning_rate": 5.653169014084507e-05,
"loss": 1.3515,
"step": 52400
},
{
"epoch": 8.920068027210885,
"grad_norm": 0.4049835503101349,
"learning_rate": 5.6091549295774646e-05,
"loss": 1.3529,
"step": 52450
},
{
"epoch": 8.928571428571429,
"grad_norm": 0.4118352234363556,
"learning_rate": 5.565140845070423e-05,
"loss": 1.3606,
"step": 52500
},
{
"epoch": 8.937074829931973,
"grad_norm": 0.40466126799583435,
"learning_rate": 5.52112676056338e-05,
"loss": 1.3509,
"step": 52550
},
{
"epoch": 8.945578231292517,
"grad_norm": 0.39501574635505676,
"learning_rate": 5.477112676056338e-05,
"loss": 1.3515,
"step": 52600
},
{
"epoch": 8.954081632653061,
"grad_norm": 0.4046393036842346,
"learning_rate": 5.433098591549296e-05,
"loss": 1.3584,
"step": 52650
},
{
"epoch": 8.962585034013605,
"grad_norm": 0.4039493203163147,
"learning_rate": 5.3890845070422535e-05,
"loss": 1.3513,
"step": 52700
},
{
"epoch": 8.97108843537415,
"grad_norm": 0.4195407032966614,
"learning_rate": 5.345070422535212e-05,
"loss": 1.3536,
"step": 52750
},
{
"epoch": 8.979591836734693,
"grad_norm": 0.39963847398757935,
"learning_rate": 5.301056338028169e-05,
"loss": 1.3511,
"step": 52800
},
{
"epoch": 8.988095238095237,
"grad_norm": 0.41947850584983826,
"learning_rate": 5.2570422535211274e-05,
"loss": 1.3558,
"step": 52850
},
{
"epoch": 8.996598639455783,
"grad_norm": 0.4046323597431183,
"learning_rate": 5.213028169014085e-05,
"loss": 1.3511,
"step": 52900
},
{
"epoch": 9.005102040816327,
"grad_norm": 0.4054872393608093,
"learning_rate": 5.1690140845070424e-05,
"loss": 1.3428,
"step": 52950
},
{
"epoch": 9.013605442176871,
"grad_norm": 0.4009184241294861,
"learning_rate": 5.125e-05,
"loss": 1.3434,
"step": 53000
},
{
"epoch": 9.013605442176871,
"eval_loss": 1.5165001153945923,
"eval_runtime": 75.3865,
"eval_samples_per_second": 1244.772,
"eval_steps_per_second": 4.868,
"step": 53000
},
{
"epoch": 9.022108843537415,
"grad_norm": 0.4195316433906555,
"learning_rate": 5.0809859154929574e-05,
"loss": 1.358,
"step": 53050
},
{
"epoch": 9.03061224489796,
"grad_norm": 0.40940186381340027,
"learning_rate": 5.0369718309859156e-05,
"loss": 1.3453,
"step": 53100
},
{
"epoch": 9.039115646258503,
"grad_norm": 0.4012869596481323,
"learning_rate": 4.992957746478873e-05,
"loss": 1.3531,
"step": 53150
},
{
"epoch": 9.047619047619047,
"grad_norm": 0.4076648950576782,
"learning_rate": 4.948943661971831e-05,
"loss": 1.3433,
"step": 53200
},
{
"epoch": 9.056122448979592,
"grad_norm": 0.40203115344047546,
"learning_rate": 4.904929577464789e-05,
"loss": 1.3491,
"step": 53250
},
{
"epoch": 9.064625850340136,
"grad_norm": 0.40481826663017273,
"learning_rate": 4.860915492957747e-05,
"loss": 1.3509,
"step": 53300
},
{
"epoch": 9.07312925170068,
"grad_norm": 0.4018038213253021,
"learning_rate": 4.8169014084507045e-05,
"loss": 1.3485,
"step": 53350
},
{
"epoch": 9.081632653061224,
"grad_norm": 0.4100002646446228,
"learning_rate": 4.772887323943662e-05,
"loss": 1.3428,
"step": 53400
},
{
"epoch": 9.09013605442177,
"grad_norm": 1.7801686525344849,
"learning_rate": 4.72887323943662e-05,
"loss": 1.3445,
"step": 53450
},
{
"epoch": 9.098639455782314,
"grad_norm": 0.39769893884658813,
"learning_rate": 4.684859154929578e-05,
"loss": 1.3467,
"step": 53500
},
{
"epoch": 9.107142857142858,
"grad_norm": 0.4195043444633484,
"learning_rate": 4.640845070422536e-05,
"loss": 1.348,
"step": 53550
},
{
"epoch": 9.115646258503402,
"grad_norm": 0.4053761959075928,
"learning_rate": 4.596830985915493e-05,
"loss": 1.3465,
"step": 53600
},
{
"epoch": 9.124149659863946,
"grad_norm": 0.4017139673233032,
"learning_rate": 4.552816901408451e-05,
"loss": 1.3448,
"step": 53650
},
{
"epoch": 9.13265306122449,
"grad_norm": 0.4139016270637512,
"learning_rate": 4.5088028169014084e-05,
"loss": 1.349,
"step": 53700
},
{
"epoch": 9.141156462585034,
"grad_norm": 0.39910098910331726,
"learning_rate": 4.464788732394366e-05,
"loss": 1.3379,
"step": 53750
},
{
"epoch": 9.149659863945578,
"grad_norm": 0.39038217067718506,
"learning_rate": 4.420774647887324e-05,
"loss": 1.3422,
"step": 53800
},
{
"epoch": 9.158163265306122,
"grad_norm": 0.4174290895462036,
"learning_rate": 4.3767605633802816e-05,
"loss": 1.3417,
"step": 53850
},
{
"epoch": 9.166666666666666,
"grad_norm": 0.399420827627182,
"learning_rate": 4.33274647887324e-05,
"loss": 1.3414,
"step": 53900
},
{
"epoch": 9.17517006802721,
"grad_norm": 0.3991691470146179,
"learning_rate": 4.288732394366197e-05,
"loss": 1.3457,
"step": 53950
},
{
"epoch": 9.183673469387756,
"grad_norm": 0.4257224500179291,
"learning_rate": 4.2447183098591555e-05,
"loss": 1.3389,
"step": 54000
},
{
"epoch": 9.183673469387756,
"eval_loss": 1.514726996421814,
"eval_runtime": 75.3819,
"eval_samples_per_second": 1244.848,
"eval_steps_per_second": 4.869,
"step": 54000
},
{
"epoch": 9.1921768707483,
"grad_norm": 0.40039774775505066,
"learning_rate": 4.200704225352113e-05,
"loss": 1.3415,
"step": 54050
},
{
"epoch": 9.200680272108844,
"grad_norm": 0.3933996558189392,
"learning_rate": 4.1566901408450705e-05,
"loss": 1.3426,
"step": 54100
},
{
"epoch": 9.209183673469388,
"grad_norm": 0.4128095805644989,
"learning_rate": 4.112676056338028e-05,
"loss": 1.3471,
"step": 54150
},
{
"epoch": 9.217687074829932,
"grad_norm": 0.4024725556373596,
"learning_rate": 4.0686619718309855e-05,
"loss": 1.3454,
"step": 54200
},
{
"epoch": 9.226190476190476,
"grad_norm": 0.4105578064918518,
"learning_rate": 4.024647887323944e-05,
"loss": 1.3399,
"step": 54250
},
{
"epoch": 9.23469387755102,
"grad_norm": 0.4047715663909912,
"learning_rate": 3.980633802816901e-05,
"loss": 1.3442,
"step": 54300
},
{
"epoch": 9.243197278911564,
"grad_norm": 0.3958626687526703,
"learning_rate": 3.9366197183098594e-05,
"loss": 1.3441,
"step": 54350
},
{
"epoch": 9.251700680272108,
"grad_norm": 0.40880346298217773,
"learning_rate": 3.892605633802817e-05,
"loss": 1.3416,
"step": 54400
},
{
"epoch": 9.260204081632653,
"grad_norm": 0.41129016876220703,
"learning_rate": 3.848591549295775e-05,
"loss": 1.3428,
"step": 54450
},
{
"epoch": 9.268707482993197,
"grad_norm": 0.39472246170043945,
"learning_rate": 3.8045774647887326e-05,
"loss": 1.3396,
"step": 54500
},
{
"epoch": 9.27721088435374,
"grad_norm": 0.40167415142059326,
"learning_rate": 3.76056338028169e-05,
"loss": 1.3425,
"step": 54550
},
{
"epoch": 9.285714285714286,
"grad_norm": 0.40154701471328735,
"learning_rate": 3.716549295774648e-05,
"loss": 1.3423,
"step": 54600
},
{
"epoch": 9.29421768707483,
"grad_norm": 0.40191853046417236,
"learning_rate": 3.672535211267606e-05,
"loss": 1.3397,
"step": 54650
},
{
"epoch": 9.302721088435375,
"grad_norm": 0.4014011025428772,
"learning_rate": 3.628521126760564e-05,
"loss": 1.3431,
"step": 54700
},
{
"epoch": 9.311224489795919,
"grad_norm": 0.4313969910144806,
"learning_rate": 3.584507042253521e-05,
"loss": 1.3413,
"step": 54750
},
{
"epoch": 9.319727891156463,
"grad_norm": 0.4143257141113281,
"learning_rate": 3.540492957746479e-05,
"loss": 1.3424,
"step": 54800
},
{
"epoch": 9.328231292517007,
"grad_norm": 0.40343379974365234,
"learning_rate": 3.4964788732394365e-05,
"loss": 1.3377,
"step": 54850
},
{
"epoch": 9.33673469387755,
"grad_norm": 0.4093693792819977,
"learning_rate": 3.452464788732394e-05,
"loss": 1.3426,
"step": 54900
},
{
"epoch": 9.345238095238095,
"grad_norm": 0.40565499663352966,
"learning_rate": 3.408450704225352e-05,
"loss": 1.3347,
"step": 54950
},
{
"epoch": 9.353741496598639,
"grad_norm": 0.40517303347587585,
"learning_rate": 3.3644366197183097e-05,
"loss": 1.3411,
"step": 55000
},
{
"epoch": 9.353741496598639,
"eval_loss": 1.5079046487808228,
"eval_runtime": 75.3606,
"eval_samples_per_second": 1245.199,
"eval_steps_per_second": 4.87,
"step": 55000
},
{
"epoch": 9.362244897959183,
"grad_norm": 0.4189765155315399,
"learning_rate": 3.320422535211268e-05,
"loss": 1.3391,
"step": 55050
},
{
"epoch": 9.370748299319727,
"grad_norm": 0.41205212473869324,
"learning_rate": 3.2764084507042253e-05,
"loss": 1.3335,
"step": 55100
},
{
"epoch": 9.379251700680273,
"grad_norm": 0.4138495624065399,
"learning_rate": 3.2323943661971835e-05,
"loss": 1.3385,
"step": 55150
},
{
"epoch": 9.387755102040817,
"grad_norm": 0.40465956926345825,
"learning_rate": 3.188380281690141e-05,
"loss": 1.3366,
"step": 55200
},
{
"epoch": 9.396258503401361,
"grad_norm": 0.41961055994033813,
"learning_rate": 3.144366197183099e-05,
"loss": 1.3386,
"step": 55250
},
{
"epoch": 9.404761904761905,
"grad_norm": 0.4017610251903534,
"learning_rate": 3.100352112676056e-05,
"loss": 1.3389,
"step": 55300
},
{
"epoch": 9.41326530612245,
"grad_norm": 0.3947182893753052,
"learning_rate": 3.056338028169014e-05,
"loss": 1.3366,
"step": 55350
},
{
"epoch": 9.421768707482993,
"grad_norm": 0.41199058294296265,
"learning_rate": 3.012323943661972e-05,
"loss": 1.3319,
"step": 55400
},
{
"epoch": 9.430272108843537,
"grad_norm": 0.44556739926338196,
"learning_rate": 2.96830985915493e-05,
"loss": 1.3386,
"step": 55450
},
{
"epoch": 9.438775510204081,
"grad_norm": 0.4042491316795349,
"learning_rate": 2.924295774647887e-05,
"loss": 1.3352,
"step": 55500
},
{
"epoch": 9.447278911564625,
"grad_norm": 0.4033842384815216,
"learning_rate": 2.880281690140845e-05,
"loss": 1.3333,
"step": 55550
},
{
"epoch": 9.45578231292517,
"grad_norm": 0.409463495016098,
"learning_rate": 2.8362676056338028e-05,
"loss": 1.3359,
"step": 55600
},
{
"epoch": 9.464285714285714,
"grad_norm": 0.4049001634120941,
"learning_rate": 2.7922535211267606e-05,
"loss": 1.3362,
"step": 55650
},
{
"epoch": 9.47278911564626,
"grad_norm": 0.4021693766117096,
"learning_rate": 2.7482394366197185e-05,
"loss": 1.3334,
"step": 55700
},
{
"epoch": 9.481292517006803,
"grad_norm": 0.3943755030632019,
"learning_rate": 2.7042253521126763e-05,
"loss": 1.3356,
"step": 55750
},
{
"epoch": 9.489795918367347,
"grad_norm": 0.4090036153793335,
"learning_rate": 2.660211267605634e-05,
"loss": 1.3359,
"step": 55800
},
{
"epoch": 9.498299319727892,
"grad_norm": 0.3975585103034973,
"learning_rate": 2.6161971830985917e-05,
"loss": 1.3356,
"step": 55850
},
{
"epoch": 9.506802721088436,
"grad_norm": 0.3984331488609314,
"learning_rate": 2.5721830985915492e-05,
"loss": 1.3334,
"step": 55900
},
{
"epoch": 9.51530612244898,
"grad_norm": 0.3891252279281616,
"learning_rate": 2.528169014084507e-05,
"loss": 1.3297,
"step": 55950
},
{
"epoch": 9.523809523809524,
"grad_norm": 0.4065133333206177,
"learning_rate": 2.484154929577465e-05,
"loss": 1.3357,
"step": 56000
},
{
"epoch": 9.523809523809524,
"eval_loss": 1.4985688924789429,
"eval_runtime": 75.3846,
"eval_samples_per_second": 1244.803,
"eval_steps_per_second": 4.868,
"step": 56000
},
{
"epoch": 9.532312925170068,
"grad_norm": 0.403832346200943,
"learning_rate": 2.4401408450704227e-05,
"loss": 1.3389,
"step": 56050
},
{
"epoch": 9.540816326530612,
"grad_norm": 0.39873868227005005,
"learning_rate": 2.3961267605633802e-05,
"loss": 1.3344,
"step": 56100
},
{
"epoch": 9.549319727891156,
"grad_norm": 0.40277689695358276,
"learning_rate": 2.352112676056338e-05,
"loss": 1.3335,
"step": 56150
},
{
"epoch": 9.5578231292517,
"grad_norm": 0.39874860644340515,
"learning_rate": 2.308098591549296e-05,
"loss": 1.3377,
"step": 56200
},
{
"epoch": 9.566326530612244,
"grad_norm": 0.4036062955856323,
"learning_rate": 2.2640845070422538e-05,
"loss": 1.335,
"step": 56250
},
{
"epoch": 9.57482993197279,
"grad_norm": 0.41648924350738525,
"learning_rate": 2.2200704225352113e-05,
"loss": 1.3341,
"step": 56300
},
{
"epoch": 9.583333333333334,
"grad_norm": 0.39397549629211426,
"learning_rate": 2.176056338028169e-05,
"loss": 1.3366,
"step": 56350
},
{
"epoch": 9.591836734693878,
"grad_norm": 0.39437320828437805,
"learning_rate": 2.1320422535211266e-05,
"loss": 1.335,
"step": 56400
},
{
"epoch": 9.600340136054422,
"grad_norm": 0.39650458097457886,
"learning_rate": 2.0880281690140845e-05,
"loss": 1.3385,
"step": 56450
},
{
"epoch": 9.608843537414966,
"grad_norm": 0.4140496850013733,
"learning_rate": 2.0440140845070423e-05,
"loss": 1.3284,
"step": 56500
},
{
"epoch": 9.61734693877551,
"grad_norm": 0.39376911520957947,
"learning_rate": 2e-05,
"loss": 1.3307,
"step": 56550
},
{
"epoch": 9.625850340136054,
"grad_norm": 0.4171138405799866,
"learning_rate": 1.955985915492958e-05,
"loss": 1.333,
"step": 56600
},
{
"epoch": 9.634353741496598,
"grad_norm": 0.41457417607307434,
"learning_rate": 1.911971830985916e-05,
"loss": 1.3366,
"step": 56650
},
{
"epoch": 9.642857142857142,
"grad_norm": 0.3986314535140991,
"learning_rate": 1.867957746478873e-05,
"loss": 1.3382,
"step": 56700
},
{
"epoch": 9.651360544217686,
"grad_norm": 0.4179774820804596,
"learning_rate": 1.823943661971831e-05,
"loss": 1.3332,
"step": 56750
},
{
"epoch": 9.65986394557823,
"grad_norm": 0.4044124186038971,
"learning_rate": 1.7799295774647887e-05,
"loss": 1.3273,
"step": 56800
},
{
"epoch": 9.668367346938776,
"grad_norm": 0.4135982096195221,
"learning_rate": 1.7359154929577466e-05,
"loss": 1.3311,
"step": 56850
},
{
"epoch": 9.67687074829932,
"grad_norm": 0.39998510479927063,
"learning_rate": 1.6919014084507044e-05,
"loss": 1.3366,
"step": 56900
},
{
"epoch": 9.685374149659864,
"grad_norm": 0.40354451537132263,
"learning_rate": 1.6478873239436623e-05,
"loss": 1.3326,
"step": 56950
},
{
"epoch": 9.693877551020408,
"grad_norm": 0.4005592167377472,
"learning_rate": 1.6038732394366198e-05,
"loss": 1.3319,
"step": 57000
},
{
"epoch": 9.693877551020408,
"eval_loss": 1.5001171827316284,
"eval_runtime": 75.6646,
"eval_samples_per_second": 1240.196,
"eval_steps_per_second": 4.85,
"step": 57000
},
{
"epoch": 9.702380952380953,
"grad_norm": 0.3942902386188507,
"learning_rate": 1.5598591549295773e-05,
"loss": 1.3331,
"step": 57050
},
{
"epoch": 9.710884353741497,
"grad_norm": 0.4057478606700897,
"learning_rate": 1.5158450704225353e-05,
"loss": 1.3295,
"step": 57100
},
{
"epoch": 9.71938775510204,
"grad_norm": 0.40495890378952026,
"learning_rate": 1.471830985915493e-05,
"loss": 1.3419,
"step": 57150
},
{
"epoch": 9.727891156462585,
"grad_norm": 0.3950183093547821,
"learning_rate": 1.4278169014084506e-05,
"loss": 1.3316,
"step": 57200
},
{
"epoch": 9.736394557823129,
"grad_norm": 0.4021071195602417,
"learning_rate": 1.3838028169014085e-05,
"loss": 1.3297,
"step": 57250
},
{
"epoch": 9.744897959183673,
"grad_norm": 0.40653911232948303,
"learning_rate": 1.3397887323943663e-05,
"loss": 1.3329,
"step": 57300
},
{
"epoch": 9.753401360544217,
"grad_norm": 0.40104779601097107,
"learning_rate": 1.2957746478873238e-05,
"loss": 1.3234,
"step": 57350
},
{
"epoch": 9.761904761904763,
"grad_norm": 0.408331036567688,
"learning_rate": 1.2517605633802817e-05,
"loss": 1.3301,
"step": 57400
},
{
"epoch": 9.770408163265307,
"grad_norm": 0.4001865088939667,
"learning_rate": 1.2077464788732395e-05,
"loss": 1.3282,
"step": 57450
},
{
"epoch": 9.77891156462585,
"grad_norm": 0.4088568389415741,
"learning_rate": 1.1637323943661972e-05,
"loss": 1.3292,
"step": 57500
},
{
"epoch": 9.787414965986395,
"grad_norm": 0.42663103342056274,
"learning_rate": 1.1197183098591549e-05,
"loss": 1.3327,
"step": 57550
},
{
"epoch": 9.795918367346939,
"grad_norm": 0.3960348963737488,
"learning_rate": 1.0757042253521127e-05,
"loss": 1.3326,
"step": 57600
},
{
"epoch": 9.804421768707483,
"grad_norm": 0.39671510457992554,
"learning_rate": 1.0316901408450704e-05,
"loss": 1.3306,
"step": 57650
},
{
"epoch": 9.812925170068027,
"grad_norm": 0.39632678031921387,
"learning_rate": 9.876760563380282e-06,
"loss": 1.3311,
"step": 57700
},
{
"epoch": 9.821428571428571,
"grad_norm": 0.4076337516307831,
"learning_rate": 9.43661971830986e-06,
"loss": 1.3348,
"step": 57750
},
{
"epoch": 9.829931972789115,
"grad_norm": 0.40061214566230774,
"learning_rate": 8.996478873239436e-06,
"loss": 1.3323,
"step": 57800
},
{
"epoch": 9.83843537414966,
"grad_norm": 0.3868488371372223,
"learning_rate": 8.556338028169014e-06,
"loss": 1.3274,
"step": 57850
},
{
"epoch": 9.846938775510203,
"grad_norm": 0.43311014771461487,
"learning_rate": 8.116197183098593e-06,
"loss": 1.3316,
"step": 57900
},
{
"epoch": 9.85544217687075,
"grad_norm": 0.39683276414871216,
"learning_rate": 7.67605633802817e-06,
"loss": 1.3302,
"step": 57950
},
{
"epoch": 9.863945578231293,
"grad_norm": 0.3994409143924713,
"learning_rate": 7.235915492957746e-06,
"loss": 1.3311,
"step": 58000
},
{
"epoch": 9.863945578231293,
"eval_loss": 1.499453067779541,
"eval_runtime": 75.4095,
"eval_samples_per_second": 1244.392,
"eval_steps_per_second": 4.867,
"step": 58000
},
{
"epoch": 9.872448979591837,
"grad_norm": 0.40057429671287537,
"learning_rate": 6.795774647887324e-06,
"loss": 1.331,
"step": 58050
},
{
"epoch": 9.880952380952381,
"grad_norm": 0.39245638251304626,
"learning_rate": 6.355633802816902e-06,
"loss": 1.334,
"step": 58100
},
{
"epoch": 9.889455782312925,
"grad_norm": 0.3993644714355469,
"learning_rate": 5.915492957746479e-06,
"loss": 1.3267,
"step": 58150
},
{
"epoch": 9.89795918367347,
"grad_norm": 0.39710524678230286,
"learning_rate": 5.475352112676056e-06,
"loss": 1.3201,
"step": 58200
},
{
"epoch": 9.906462585034014,
"grad_norm": 0.3986225426197052,
"learning_rate": 5.0352112676056345e-06,
"loss": 1.3275,
"step": 58250
},
{
"epoch": 9.914965986394558,
"grad_norm": 0.4060732126235962,
"learning_rate": 4.595070422535211e-06,
"loss": 1.3274,
"step": 58300
},
{
"epoch": 9.923469387755102,
"grad_norm": 0.3890203833580017,
"learning_rate": 4.154929577464789e-06,
"loss": 1.3259,
"step": 58350
},
{
"epoch": 9.931972789115646,
"grad_norm": 0.39963120222091675,
"learning_rate": 3.7147887323943665e-06,
"loss": 1.3343,
"step": 58400
},
{
"epoch": 9.94047619047619,
"grad_norm": 0.4122091233730316,
"learning_rate": 3.274647887323944e-06,
"loss": 1.3313,
"step": 58450
},
{
"epoch": 9.948979591836736,
"grad_norm": 0.40625423192977905,
"learning_rate": 2.8433098591549298e-06,
"loss": 1.329,
"step": 58500
},
{
"epoch": 9.95748299319728,
"grad_norm": 0.39917826652526855,
"learning_rate": 2.4031690140845074e-06,
"loss": 1.329,
"step": 58550
},
{
"epoch": 9.965986394557824,
"grad_norm": 0.4005703330039978,
"learning_rate": 1.9630281690140846e-06,
"loss": 1.3286,
"step": 58600
},
{
"epoch": 9.974489795918368,
"grad_norm": 0.4052295386791229,
"learning_rate": 1.522887323943662e-06,
"loss": 1.3334,
"step": 58650
},
{
"epoch": 9.982993197278912,
"grad_norm": 0.3987277150154114,
"learning_rate": 1.0827464788732394e-06,
"loss": 1.3279,
"step": 58700
},
{
"epoch": 9.991496598639456,
"grad_norm": 0.39191877841949463,
"learning_rate": 6.426056338028169e-07,
"loss": 1.322,
"step": 58750
},
{
"epoch": 10.0,
"grad_norm": 0.39657750725746155,
"learning_rate": 2.0246478873239435e-07,
"loss": 1.3269,
"step": 58800
},
{
"epoch": 10.008503401360544,
"grad_norm": 0.4391802251338959,
"learning_rate": 0.000170377030162413,
"loss": 1.3375,
"step": 58850
},
{
"epoch": 10.017006802721088,
"grad_norm": 0.4108564257621765,
"learning_rate": 0.00017008700696055683,
"loss": 1.3625,
"step": 58900
},
{
"epoch": 10.025510204081632,
"grad_norm": 0.4225695729255676,
"learning_rate": 0.0001697969837587007,
"loss": 1.3704,
"step": 58950
},
{
"epoch": 10.034013605442176,
"grad_norm": 0.41939666867256165,
"learning_rate": 0.00016950696055684455,
"loss": 1.3693,
"step": 59000
},
{
"epoch": 10.034013605442176,
"eval_loss": 1.5424796342849731,
"eval_runtime": 75.2733,
"eval_samples_per_second": 1246.644,
"eval_steps_per_second": 4.876,
"step": 59000
},
{
"epoch": 10.04251700680272,
"grad_norm": 0.925662100315094,
"learning_rate": 0.00016921693735498842,
"loss": 1.3652,
"step": 59050
},
{
"epoch": 10.051020408163266,
"grad_norm": 0.42733487486839294,
"learning_rate": 0.00016892691415313225,
"loss": 1.3702,
"step": 59100
},
{
"epoch": 10.05952380952381,
"grad_norm": 0.43742504715919495,
"learning_rate": 0.0001686368909512761,
"loss": 1.3748,
"step": 59150
},
{
"epoch": 10.068027210884354,
"grad_norm": 0.5494563579559326,
"learning_rate": 0.00016834686774941994,
"loss": 1.3775,
"step": 59200
},
{
"epoch": 10.076530612244898,
"grad_norm": 0.43776175379753113,
"learning_rate": 0.0001680568445475638,
"loss": 1.3784,
"step": 59250
},
{
"epoch": 10.085034013605442,
"grad_norm": 0.4281368851661682,
"learning_rate": 0.00016776682134570767,
"loss": 1.3695,
"step": 59300
},
{
"epoch": 10.093537414965986,
"grad_norm": 0.5244280695915222,
"learning_rate": 0.0001674767981438515,
"loss": 1.3764,
"step": 59350
},
{
"epoch": 10.10204081632653,
"grad_norm": 0.40682801604270935,
"learning_rate": 0.00016718677494199536,
"loss": 1.3829,
"step": 59400
},
{
"epoch": 10.110544217687075,
"grad_norm": 0.42796218395233154,
"learning_rate": 0.00016690255220417631,
"loss": 1.376,
"step": 59450
},
{
"epoch": 10.119047619047619,
"grad_norm": 0.43184956908226013,
"learning_rate": 0.00016661252900232018,
"loss": 1.3794,
"step": 59500
},
{
"epoch": 10.127551020408163,
"grad_norm": 0.4699791371822357,
"learning_rate": 0.00016632250580046404,
"loss": 1.3849,
"step": 59550
},
{
"epoch": 10.136054421768707,
"grad_norm": 0.4018096923828125,
"learning_rate": 0.0001660324825986079,
"loss": 1.3765,
"step": 59600
},
{
"epoch": 10.14455782312925,
"grad_norm": 0.42693692445755005,
"learning_rate": 0.00016574245939675176,
"loss": 1.3726,
"step": 59650
},
{
"epoch": 10.153061224489797,
"grad_norm": 0.42560431361198425,
"learning_rate": 0.0001654524361948956,
"loss": 1.376,
"step": 59700
},
{
"epoch": 10.16156462585034,
"grad_norm": 0.42239609360694885,
"learning_rate": 0.00016516241299303943,
"loss": 1.3863,
"step": 59750
},
{
"epoch": 10.170068027210885,
"grad_norm": 0.41921770572662354,
"learning_rate": 0.0001648723897911833,
"loss": 1.3819,
"step": 59800
},
{
"epoch": 10.178571428571429,
"grad_norm": 0.41046616435050964,
"learning_rate": 0.00016458236658932715,
"loss": 1.3748,
"step": 59850
},
{
"epoch": 10.187074829931973,
"grad_norm": 0.41564711928367615,
"learning_rate": 0.000164292343387471,
"loss": 1.3802,
"step": 59900
},
{
"epoch": 10.195578231292517,
"grad_norm": 0.4275333881378174,
"learning_rate": 0.00016400232018561484,
"loss": 1.3757,
"step": 59950
},
{
"epoch": 10.204081632653061,
"grad_norm": 0.47480013966560364,
"learning_rate": 0.0001637122969837587,
"loss": 1.3858,
"step": 60000
},
{
"epoch": 10.204081632653061,
"eval_loss": 1.5496501922607422,
"eval_runtime": 75.245,
"eval_samples_per_second": 1247.113,
"eval_steps_per_second": 4.877,
"step": 60000
},
{
"epoch": 10.212585034013605,
"grad_norm": 2.0938565731048584,
"learning_rate": 0.00016342227378190257,
"loss": 1.3816,
"step": 60050
},
{
"epoch": 10.22108843537415,
"grad_norm": 0.4088470935821533,
"learning_rate": 0.0001631322505800464,
"loss": 1.3806,
"step": 60100
},
{
"epoch": 10.229591836734693,
"grad_norm": 0.43842536211013794,
"learning_rate": 0.00016284222737819026,
"loss": 1.3796,
"step": 60150
},
{
"epoch": 10.238095238095237,
"grad_norm": 0.41963255405426025,
"learning_rate": 0.00016255220417633412,
"loss": 1.3798,
"step": 60200
},
{
"epoch": 10.246598639455783,
"grad_norm": 0.4422262907028198,
"learning_rate": 0.00016226218097447796,
"loss": 1.3778,
"step": 60250
},
{
"epoch": 10.255102040816327,
"grad_norm": 0.43372678756713867,
"learning_rate": 0.00016197215777262182,
"loss": 1.3779,
"step": 60300
},
{
"epoch": 10.263605442176871,
"grad_norm": 0.42680031061172485,
"learning_rate": 0.00016168213457076568,
"loss": 1.3771,
"step": 60350
},
{
"epoch": 10.272108843537415,
"grad_norm": 0.4302733540534973,
"learning_rate": 0.0001613921113689095,
"loss": 1.3848,
"step": 60400
},
{
"epoch": 10.28061224489796,
"grad_norm": 0.4111240804195404,
"learning_rate": 0.00016110208816705337,
"loss": 1.3784,
"step": 60450
},
{
"epoch": 10.289115646258503,
"grad_norm": 0.43267524242401123,
"learning_rate": 0.0001608120649651972,
"loss": 1.3828,
"step": 60500
},
{
"epoch": 10.297619047619047,
"grad_norm": 0.4088628888130188,
"learning_rate": 0.00016052204176334107,
"loss": 1.3809,
"step": 60550
},
{
"epoch": 10.306122448979592,
"grad_norm": 0.4188777208328247,
"learning_rate": 0.00016023201856148493,
"loss": 1.3772,
"step": 60600
},
{
"epoch": 10.314625850340136,
"grad_norm": 0.4147690534591675,
"learning_rate": 0.0001599419953596288,
"loss": 1.3788,
"step": 60650
},
{
"epoch": 10.32312925170068,
"grad_norm": 0.48966458439826965,
"learning_rate": 0.00015965777262180974,
"loss": 1.378,
"step": 60700
},
{
"epoch": 10.331632653061224,
"grad_norm": 0.40553244948387146,
"learning_rate": 0.0001593677494199536,
"loss": 1.3879,
"step": 60750
},
{
"epoch": 10.34013605442177,
"grad_norm": 0.41595011949539185,
"learning_rate": 0.00015907772621809747,
"loss": 1.3801,
"step": 60800
},
{
"epoch": 10.348639455782314,
"grad_norm": 0.41190576553344727,
"learning_rate": 0.0001587877030162413,
"loss": 1.3838,
"step": 60850
},
{
"epoch": 10.357142857142858,
"grad_norm": 0.4079788625240326,
"learning_rate": 0.00015849767981438516,
"loss": 1.3885,
"step": 60900
},
{
"epoch": 10.365646258503402,
"grad_norm": 0.41605037450790405,
"learning_rate": 0.000158207656612529,
"loss": 1.3869,
"step": 60950
},
{
"epoch": 10.374149659863946,
"grad_norm": 0.41318464279174805,
"learning_rate": 0.00015791763341067286,
"loss": 1.3789,
"step": 61000
},
{
"epoch": 10.374149659863946,
"eval_loss": 1.549613356590271,
"eval_runtime": 75.299,
"eval_samples_per_second": 1246.219,
"eval_steps_per_second": 4.874,
"step": 61000
},
{
"epoch": 10.38265306122449,
"grad_norm": 0.42197248339653015,
"learning_rate": 0.00015762761020881672,
"loss": 1.3862,
"step": 61050
},
{
"epoch": 10.391156462585034,
"grad_norm": 0.423498272895813,
"learning_rate": 0.00015733758700696055,
"loss": 1.3831,
"step": 61100
},
{
"epoch": 10.399659863945578,
"grad_norm": 0.41805723309516907,
"learning_rate": 0.0001570475638051044,
"loss": 1.3807,
"step": 61150
},
{
"epoch": 10.408163265306122,
"grad_norm": 0.40790635347366333,
"learning_rate": 0.00015675754060324827,
"loss": 1.3818,
"step": 61200
},
{
"epoch": 10.416666666666666,
"grad_norm": 0.42006179690361023,
"learning_rate": 0.0001564675174013921,
"loss": 1.3775,
"step": 61250
},
{
"epoch": 10.42517006802721,
"grad_norm": 0.41415610909461975,
"learning_rate": 0.00015617749419953597,
"loss": 1.391,
"step": 61300
},
{
"epoch": 10.433673469387756,
"grad_norm": 0.44713294506073,
"learning_rate": 0.00015588747099767983,
"loss": 1.383,
"step": 61350
},
{
"epoch": 10.4421768707483,
"grad_norm": 0.44019782543182373,
"learning_rate": 0.00015559744779582366,
"loss": 1.3904,
"step": 61400
},
{
"epoch": 10.450680272108844,
"grad_norm": 0.4164921045303345,
"learning_rate": 0.00015530742459396752,
"loss": 1.3821,
"step": 61450
},
{
"epoch": 10.459183673469388,
"grad_norm": 0.4174480438232422,
"learning_rate": 0.00015501740139211139,
"loss": 1.3791,
"step": 61500
},
{
"epoch": 10.467687074829932,
"grad_norm": 0.49047085642814636,
"learning_rate": 0.00015472737819025522,
"loss": 1.3818,
"step": 61550
},
{
"epoch": 10.476190476190476,
"grad_norm": 0.405862957239151,
"learning_rate": 0.00015443735498839908,
"loss": 1.3817,
"step": 61600
},
{
"epoch": 10.48469387755102,
"grad_norm": 0.4351731538772583,
"learning_rate": 0.00015414733178654291,
"loss": 1.3806,
"step": 61650
},
{
"epoch": 10.493197278911564,
"grad_norm": 0.4245615303516388,
"learning_rate": 0.00015385730858468678,
"loss": 1.383,
"step": 61700
},
{
"epoch": 10.501700680272108,
"grad_norm": 0.39436227083206177,
"learning_rate": 0.00015356728538283064,
"loss": 1.3805,
"step": 61750
},
{
"epoch": 10.510204081632653,
"grad_norm": 0.41128745675086975,
"learning_rate": 0.0001532772621809745,
"loss": 1.386,
"step": 61800
},
{
"epoch": 10.518707482993197,
"grad_norm": 0.42046472430229187,
"learning_rate": 0.00015298723897911833,
"loss": 1.3897,
"step": 61850
},
{
"epoch": 10.527210884353742,
"grad_norm": 0.4486936926841736,
"learning_rate": 0.00015269721577726217,
"loss": 1.383,
"step": 61900
},
{
"epoch": 10.535714285714286,
"grad_norm": 0.4166390001773834,
"learning_rate": 0.00015240719257540603,
"loss": 1.3793,
"step": 61950
},
{
"epoch": 10.54421768707483,
"grad_norm": 0.42586439847946167,
"learning_rate": 0.0001521171693735499,
"loss": 1.3778,
"step": 62000
},
{
"epoch": 10.54421768707483,
"eval_loss": 1.5420804023742676,
"eval_runtime": 75.2999,
"eval_samples_per_second": 1246.204,
"eval_steps_per_second": 4.874,
"step": 62000
},
{
"epoch": 10.552721088435375,
"grad_norm": 0.39667069911956787,
"learning_rate": 0.00015182714617169375,
"loss": 1.3749,
"step": 62050
},
{
"epoch": 10.561224489795919,
"grad_norm": 0.40961819887161255,
"learning_rate": 0.0001515371229698376,
"loss": 1.3797,
"step": 62100
},
{
"epoch": 10.569727891156463,
"grad_norm": 0.4282205104827881,
"learning_rate": 0.00015124709976798144,
"loss": 1.3803,
"step": 62150
},
{
"epoch": 10.578231292517007,
"grad_norm": 0.42050084471702576,
"learning_rate": 0.00015095707656612528,
"loss": 1.3812,
"step": 62200
},
{
"epoch": 10.58673469387755,
"grad_norm": 0.4289666712284088,
"learning_rate": 0.00015066705336426914,
"loss": 1.3824,
"step": 62250
},
{
"epoch": 10.595238095238095,
"grad_norm": 0.4385221600532532,
"learning_rate": 0.000150377030162413,
"loss": 1.3733,
"step": 62300
},
{
"epoch": 10.603741496598639,
"grad_norm": 2.321608781814575,
"learning_rate": 0.00015008700696055686,
"loss": 1.3793,
"step": 62350
},
{
"epoch": 10.612244897959183,
"grad_norm": 0.4276087284088135,
"learning_rate": 0.00014979698375870072,
"loss": 1.382,
"step": 62400
},
{
"epoch": 10.620748299319727,
"grad_norm": 0.41472572088241577,
"learning_rate": 0.00014950696055684453,
"loss": 1.3772,
"step": 62450
},
{
"epoch": 10.629251700680273,
"grad_norm": 0.40703973174095154,
"learning_rate": 0.0001492169373549884,
"loss": 1.381,
"step": 62500
},
{
"epoch": 10.637755102040817,
"grad_norm": 0.41853588819503784,
"learning_rate": 0.00014892691415313225,
"loss": 1.377,
"step": 62550
},
{
"epoch": 10.646258503401361,
"grad_norm": 0.4194204807281494,
"learning_rate": 0.0001486368909512761,
"loss": 1.3803,
"step": 62600
},
{
"epoch": 10.654761904761905,
"grad_norm": 0.41111990809440613,
"learning_rate": 0.00014834686774941997,
"loss": 1.3778,
"step": 62650
},
{
"epoch": 10.66326530612245,
"grad_norm": 0.43283921480178833,
"learning_rate": 0.0001480568445475638,
"loss": 1.3759,
"step": 62700
},
{
"epoch": 10.671768707482993,
"grad_norm": 0.4069305956363678,
"learning_rate": 0.00014776682134570764,
"loss": 1.3779,
"step": 62750
},
{
"epoch": 10.680272108843537,
"grad_norm": 0.4265955984592438,
"learning_rate": 0.0001474767981438515,
"loss": 1.3818,
"step": 62800
},
{
"epoch": 10.688775510204081,
"grad_norm": 0.43886256217956543,
"learning_rate": 0.00014718677494199536,
"loss": 1.3739,
"step": 62850
},
{
"epoch": 10.697278911564625,
"grad_norm": 0.4144219756126404,
"learning_rate": 0.00014689675174013922,
"loss": 1.3748,
"step": 62900
},
{
"epoch": 10.70578231292517,
"grad_norm": 0.41580018401145935,
"learning_rate": 0.00014660672853828309,
"loss": 1.379,
"step": 62950
},
{
"epoch": 10.714285714285714,
"grad_norm": 0.41576939821243286,
"learning_rate": 0.00014631670533642692,
"loss": 1.3789,
"step": 63000
},
{
"epoch": 10.714285714285714,
"eval_loss": 1.5434892177581787,
"eval_runtime": 75.2563,
"eval_samples_per_second": 1246.925,
"eval_steps_per_second": 4.877,
"step": 63000
},
{
"epoch": 10.722789115646258,
"grad_norm": 0.44234171509742737,
"learning_rate": 0.00014602668213457075,
"loss": 1.3822,
"step": 63050
},
{
"epoch": 10.731292517006803,
"grad_norm": 0.4127776026725769,
"learning_rate": 0.00014573665893271461,
"loss": 1.3789,
"step": 63100
},
{
"epoch": 10.739795918367347,
"grad_norm": 0.4461323320865631,
"learning_rate": 0.00014544663573085848,
"loss": 1.3744,
"step": 63150
},
{
"epoch": 10.748299319727892,
"grad_norm": 0.5485215187072754,
"learning_rate": 0.00014515661252900234,
"loss": 1.3762,
"step": 63200
},
{
"epoch": 10.756802721088436,
"grad_norm": 0.4205337464809418,
"learning_rate": 0.00014486658932714617,
"loss": 1.3753,
"step": 63250
},
{
"epoch": 10.76530612244898,
"grad_norm": 0.4096565246582031,
"learning_rate": 0.00014457656612529003,
"loss": 1.3752,
"step": 63300
},
{
"epoch": 10.773809523809524,
"grad_norm": 0.44153106212615967,
"learning_rate": 0.00014428654292343387,
"loss": 1.3784,
"step": 63350
},
{
"epoch": 10.782312925170068,
"grad_norm": 0.41935479640960693,
"learning_rate": 0.00014399651972157773,
"loss": 1.3803,
"step": 63400
},
{
"epoch": 10.790816326530612,
"grad_norm": 0.40989023447036743,
"learning_rate": 0.0001437064965197216,
"loss": 1.3779,
"step": 63450
},
{
"epoch": 10.799319727891156,
"grad_norm": 0.41036033630371094,
"learning_rate": 0.00014341647331786542,
"loss": 1.38,
"step": 63500
},
{
"epoch": 10.8078231292517,
"grad_norm": 0.4159608781337738,
"learning_rate": 0.00014312645011600928,
"loss": 1.3775,
"step": 63550
},
{
"epoch": 10.816326530612244,
"grad_norm": 0.4197967052459717,
"learning_rate": 0.00014283642691415314,
"loss": 1.3763,
"step": 63600
},
{
"epoch": 10.82482993197279,
"grad_norm": 0.9038332104682922,
"learning_rate": 0.00014254640371229698,
"loss": 1.3774,
"step": 63650
},
{
"epoch": 10.833333333333334,
"grad_norm": 0.41402915120124817,
"learning_rate": 0.00014225638051044084,
"loss": 1.3699,
"step": 63700
},
{
"epoch": 10.841836734693878,
"grad_norm": 0.4222056567668915,
"learning_rate": 0.00014197215777262182,
"loss": 1.3716,
"step": 63750
},
{
"epoch": 10.850340136054422,
"grad_norm": 0.4178094267845154,
"learning_rate": 0.00014168213457076568,
"loss": 1.3776,
"step": 63800
},
{
"epoch": 10.858843537414966,
"grad_norm": 0.42892855405807495,
"learning_rate": 0.00014139211136890951,
"loss": 1.3747,
"step": 63850
},
{
"epoch": 10.86734693877551,
"grad_norm": 0.4110173285007477,
"learning_rate": 0.00014110208816705335,
"loss": 1.3807,
"step": 63900
},
{
"epoch": 10.875850340136054,
"grad_norm": 0.4244990050792694,
"learning_rate": 0.0001408120649651972,
"loss": 1.375,
"step": 63950
},
{
"epoch": 10.884353741496598,
"grad_norm": 0.5068066716194153,
"learning_rate": 0.00014052204176334107,
"loss": 1.3742,
"step": 64000
},
{
"epoch": 10.884353741496598,
"eval_loss": 1.5442039966583252,
"eval_runtime": 75.3144,
"eval_samples_per_second": 1245.964,
"eval_steps_per_second": 4.873,
"step": 64000
},
{
"epoch": 10.892857142857142,
"grad_norm": 0.4364668130874634,
"learning_rate": 0.00014023201856148493,
"loss": 1.3739,
"step": 64050
},
{
"epoch": 10.901360544217686,
"grad_norm": 0.4433183968067169,
"learning_rate": 0.00013994199535962877,
"loss": 1.3756,
"step": 64100
},
{
"epoch": 10.90986394557823,
"grad_norm": 0.41508781909942627,
"learning_rate": 0.00013965197215777263,
"loss": 1.3792,
"step": 64150
},
{
"epoch": 10.918367346938776,
"grad_norm": 0.4308226406574249,
"learning_rate": 0.00013936194895591646,
"loss": 1.3818,
"step": 64200
},
{
"epoch": 10.92687074829932,
"grad_norm": 0.43690988421440125,
"learning_rate": 0.00013907192575406032,
"loss": 1.3723,
"step": 64250
},
{
"epoch": 10.935374149659864,
"grad_norm": 0.9234839081764221,
"learning_rate": 0.00013878190255220418,
"loss": 1.3769,
"step": 64300
},
{
"epoch": 10.943877551020408,
"grad_norm": 0.4184640049934387,
"learning_rate": 0.00013849187935034804,
"loss": 1.3783,
"step": 64350
},
{
"epoch": 10.952380952380953,
"grad_norm": 0.418977826833725,
"learning_rate": 0.00013820185614849188,
"loss": 1.3708,
"step": 64400
},
{
"epoch": 10.960884353741497,
"grad_norm": 0.42863503098487854,
"learning_rate": 0.00013791183294663574,
"loss": 1.3643,
"step": 64450
},
{
"epoch": 10.96938775510204,
"grad_norm": 0.41124334931373596,
"learning_rate": 0.00013762180974477957,
"loss": 1.3788,
"step": 64500
},
{
"epoch": 10.977891156462585,
"grad_norm": 0.46177518367767334,
"learning_rate": 0.00013733178654292343,
"loss": 1.376,
"step": 64550
},
{
"epoch": 10.986394557823129,
"grad_norm": 0.4240172207355499,
"learning_rate": 0.0001370417633410673,
"loss": 1.3674,
"step": 64600
},
{
"epoch": 10.994897959183673,
"grad_norm": 0.4008079767227173,
"learning_rate": 0.00013675174013921113,
"loss": 1.3676,
"step": 64650
},
{
"epoch": 11.003401360544217,
"grad_norm": 0.40654611587524414,
"learning_rate": 0.000136461716937355,
"loss": 1.3716,
"step": 64700
},
{
"epoch": 11.011904761904763,
"grad_norm": 0.4086385667324066,
"learning_rate": 0.00013617169373549885,
"loss": 1.3692,
"step": 64750
},
{
"epoch": 11.020408163265307,
"grad_norm": 0.43329235911369324,
"learning_rate": 0.0001358816705336427,
"loss": 1.3694,
"step": 64800
},
{
"epoch": 11.02891156462585,
"grad_norm": 0.43725255131721497,
"learning_rate": 0.00013559164733178655,
"loss": 1.3648,
"step": 64850
},
{
"epoch": 11.037414965986395,
"grad_norm": 0.4894959628582001,
"learning_rate": 0.00013530162412993038,
"loss": 1.3681,
"step": 64900
},
{
"epoch": 11.045918367346939,
"grad_norm": 0.4180058538913727,
"learning_rate": 0.00013501160092807424,
"loss": 1.3637,
"step": 64950
},
{
"epoch": 11.054421768707483,
"grad_norm": 0.42746037244796753,
"learning_rate": 0.0001347215777262181,
"loss": 1.3674,
"step": 65000
},
{
"epoch": 11.054421768707483,
"eval_loss": 1.5379548072814941,
"eval_runtime": 75.1076,
"eval_samples_per_second": 1249.394,
"eval_steps_per_second": 4.886,
"step": 65000
},
{
"epoch": 11.062925170068027,
"grad_norm": 0.4063352644443512,
"learning_rate": 0.00013443155452436196,
"loss": 1.3644,
"step": 65050
},
{
"epoch": 11.071428571428571,
"grad_norm": 0.41466256976127625,
"learning_rate": 0.00013414153132250582,
"loss": 1.367,
"step": 65100
},
{
"epoch": 11.079931972789115,
"grad_norm": 0.40846332907676697,
"learning_rate": 0.00013385150812064966,
"loss": 1.3586,
"step": 65150
},
{
"epoch": 11.08843537414966,
"grad_norm": 0.4066697955131531,
"learning_rate": 0.0001335614849187935,
"loss": 1.3681,
"step": 65200
},
{
"epoch": 11.096938775510203,
"grad_norm": 0.4276025891304016,
"learning_rate": 0.00013327146171693735,
"loss": 1.3638,
"step": 65250
},
{
"epoch": 11.10544217687075,
"grad_norm": 0.4069554805755615,
"learning_rate": 0.00013298143851508121,
"loss": 1.3661,
"step": 65300
},
{
"epoch": 11.113945578231293,
"grad_norm": 0.5378205180168152,
"learning_rate": 0.00013269141531322508,
"loss": 1.365,
"step": 65350
},
{
"epoch": 11.122448979591837,
"grad_norm": 0.4328392744064331,
"learning_rate": 0.00013240139211136894,
"loss": 1.3681,
"step": 65400
},
{
"epoch": 11.130952380952381,
"grad_norm": 1.8447155952453613,
"learning_rate": 0.00013211136890951274,
"loss": 1.3643,
"step": 65450
},
{
"epoch": 11.139455782312925,
"grad_norm": 0.4180423319339752,
"learning_rate": 0.0001318213457076566,
"loss": 1.3644,
"step": 65500
},
{
"epoch": 11.14795918367347,
"grad_norm": 0.4122379720211029,
"learning_rate": 0.00013153132250580047,
"loss": 1.3626,
"step": 65550
},
{
"epoch": 11.156462585034014,
"grad_norm": 0.431257963180542,
"learning_rate": 0.00013124129930394433,
"loss": 1.3653,
"step": 65600
},
{
"epoch": 11.164965986394558,
"grad_norm": 0.39916712045669556,
"learning_rate": 0.0001309512761020882,
"loss": 1.3615,
"step": 65650
},
{
"epoch": 11.173469387755102,
"grad_norm": 0.4114948809146881,
"learning_rate": 0.00013066125290023202,
"loss": 1.37,
"step": 65700
},
{
"epoch": 11.181972789115646,
"grad_norm": 0.42501741647720337,
"learning_rate": 0.00013037122969837586,
"loss": 1.369,
"step": 65750
},
{
"epoch": 11.19047619047619,
"grad_norm": 0.4058555066585541,
"learning_rate": 0.00013008120649651972,
"loss": 1.3609,
"step": 65800
},
{
"epoch": 11.198979591836734,
"grad_norm": 0.40838614106178284,
"learning_rate": 0.00012979118329466358,
"loss": 1.3605,
"step": 65850
},
{
"epoch": 11.20748299319728,
"grad_norm": 0.4708266854286194,
"learning_rate": 0.00012950116009280744,
"loss": 1.3606,
"step": 65900
},
{
"epoch": 11.215986394557824,
"grad_norm": 0.40549665689468384,
"learning_rate": 0.0001292111368909513,
"loss": 1.3652,
"step": 65950
},
{
"epoch": 11.224489795918368,
"grad_norm": 0.4081163704395294,
"learning_rate": 0.00012892111368909513,
"loss": 1.3609,
"step": 66000
},
{
"epoch": 11.224489795918368,
"eval_loss": 1.5325791835784912,
"eval_runtime": 75.302,
"eval_samples_per_second": 1246.169,
"eval_steps_per_second": 4.874,
"step": 66000
},
{
"epoch": 11.232993197278912,
"grad_norm": 0.42392808198928833,
"learning_rate": 0.00012863109048723897,
"loss": 1.3617,
"step": 66050
},
{
"epoch": 11.241496598639456,
"grad_norm": 0.4850771725177765,
"learning_rate": 0.00012834106728538283,
"loss": 1.3733,
"step": 66100
},
{
"epoch": 11.25,
"grad_norm": 0.42814481258392334,
"learning_rate": 0.0001280510440835267,
"loss": 1.3652,
"step": 66150
},
{
"epoch": 11.258503401360544,
"grad_norm": 0.40000179409980774,
"learning_rate": 0.00012776102088167055,
"loss": 1.3717,
"step": 66200
},
{
"epoch": 11.267006802721088,
"grad_norm": 0.4144078493118286,
"learning_rate": 0.00012747099767981438,
"loss": 1.3645,
"step": 66250
},
{
"epoch": 11.275510204081632,
"grad_norm": 0.49152231216430664,
"learning_rate": 0.00012718097447795825,
"loss": 1.3661,
"step": 66300
},
{
"epoch": 11.284013605442176,
"grad_norm": 0.4072152078151703,
"learning_rate": 0.00012689095127610208,
"loss": 1.3671,
"step": 66350
},
{
"epoch": 11.29251700680272,
"grad_norm": 0.40859153866767883,
"learning_rate": 0.00012660092807424594,
"loss": 1.3667,
"step": 66400
},
{
"epoch": 11.301020408163266,
"grad_norm": 0.4389115869998932,
"learning_rate": 0.0001263109048723898,
"loss": 1.3648,
"step": 66450
},
{
"epoch": 11.30952380952381,
"grad_norm": 0.42225921154022217,
"learning_rate": 0.00012602088167053364,
"loss": 1.3584,
"step": 66500
},
{
"epoch": 11.318027210884354,
"grad_norm": 0.40919360518455505,
"learning_rate": 0.0001257308584686775,
"loss": 1.3613,
"step": 66550
},
{
"epoch": 11.326530612244898,
"grad_norm": 0.41595298051834106,
"learning_rate": 0.00012544083526682136,
"loss": 1.359,
"step": 66600
},
{
"epoch": 11.335034013605442,
"grad_norm": 0.405838280916214,
"learning_rate": 0.0001251508120649652,
"loss": 1.3581,
"step": 66650
},
{
"epoch": 11.343537414965986,
"grad_norm": 0.4283597767353058,
"learning_rate": 0.00012486078886310905,
"loss": 1.3653,
"step": 66700
},
{
"epoch": 11.35204081632653,
"grad_norm": 0.42228442430496216,
"learning_rate": 0.00012457076566125291,
"loss": 1.3666,
"step": 66750
},
{
"epoch": 11.360544217687075,
"grad_norm": 0.4238821566104889,
"learning_rate": 0.00012428074245939675,
"loss": 1.3621,
"step": 66800
},
{
"epoch": 11.369047619047619,
"grad_norm": 0.40990692377090454,
"learning_rate": 0.0001239907192575406,
"loss": 1.3663,
"step": 66850
},
{
"epoch": 11.377551020408163,
"grad_norm": 0.43895843625068665,
"learning_rate": 0.00012370069605568447,
"loss": 1.3581,
"step": 66900
},
{
"epoch": 11.386054421768707,
"grad_norm": 0.42183375358581543,
"learning_rate": 0.0001234106728538283,
"loss": 1.3655,
"step": 66950
},
{
"epoch": 11.39455782312925,
"grad_norm": 0.41603147983551025,
"learning_rate": 0.00012312064965197216,
"loss": 1.353,
"step": 67000
},
{
"epoch": 11.39455782312925,
"eval_loss": 1.5269132852554321,
"eval_runtime": 75.2762,
"eval_samples_per_second": 1246.596,
"eval_steps_per_second": 4.875,
"step": 67000
},
{
"epoch": 11.403061224489797,
"grad_norm": 0.42194870114326477,
"learning_rate": 0.00012283062645011603,
"loss": 1.3639,
"step": 67050
},
{
"epoch": 11.41156462585034,
"grad_norm": 0.42385604977607727,
"learning_rate": 0.00012254060324825986,
"loss": 1.3605,
"step": 67100
},
{
"epoch": 11.420068027210885,
"grad_norm": 0.4089812636375427,
"learning_rate": 0.00012225058004640372,
"loss": 1.3553,
"step": 67150
},
{
"epoch": 11.428571428571429,
"grad_norm": 0.402220219373703,
"learning_rate": 0.00012196055684454757,
"loss": 1.3643,
"step": 67200
},
{
"epoch": 11.437074829931973,
"grad_norm": 0.4062860310077667,
"learning_rate": 0.00012167053364269142,
"loss": 1.3615,
"step": 67250
},
{
"epoch": 11.445578231292517,
"grad_norm": 0.4270987808704376,
"learning_rate": 0.00012138051044083528,
"loss": 1.3586,
"step": 67300
},
{
"epoch": 11.454081632653061,
"grad_norm": 0.39737579226493835,
"learning_rate": 0.00012109048723897912,
"loss": 1.3566,
"step": 67350
},
{
"epoch": 11.462585034013605,
"grad_norm": 0.4278363287448883,
"learning_rate": 0.00012080046403712297,
"loss": 1.3597,
"step": 67400
},
{
"epoch": 11.47108843537415,
"grad_norm": 0.4197821319103241,
"learning_rate": 0.00012051044083526683,
"loss": 1.3627,
"step": 67450
},
{
"epoch": 11.479591836734693,
"grad_norm": 0.4141803979873657,
"learning_rate": 0.00012022041763341067,
"loss": 1.3576,
"step": 67500
},
{
"epoch": 11.488095238095237,
"grad_norm": 0.41713324189186096,
"learning_rate": 0.00011993039443155453,
"loss": 1.3544,
"step": 67550
},
{
"epoch": 11.496598639455783,
"grad_norm": 0.4070891737937927,
"learning_rate": 0.00011964037122969839,
"loss": 1.364,
"step": 67600
},
{
"epoch": 11.505102040816327,
"grad_norm": 0.4234033524990082,
"learning_rate": 0.00011935034802784222,
"loss": 1.3549,
"step": 67650
},
{
"epoch": 11.513605442176871,
"grad_norm": 0.4253610372543335,
"learning_rate": 0.00011906032482598608,
"loss": 1.36,
"step": 67700
},
{
"epoch": 11.522108843537415,
"grad_norm": 0.4099927246570587,
"learning_rate": 0.00011877030162412995,
"loss": 1.3639,
"step": 67750
},
{
"epoch": 11.53061224489796,
"grad_norm": 0.4126720726490021,
"learning_rate": 0.00011848607888631091,
"loss": 1.3629,
"step": 67800
},
{
"epoch": 11.539115646258503,
"grad_norm": 0.42233848571777344,
"learning_rate": 0.00011819605568445476,
"loss": 1.358,
"step": 67850
},
{
"epoch": 11.547619047619047,
"grad_norm": 0.42125800251960754,
"learning_rate": 0.00011790603248259861,
"loss": 1.3552,
"step": 67900
},
{
"epoch": 11.556122448979592,
"grad_norm": 0.4108964204788208,
"learning_rate": 0.00011761600928074247,
"loss": 1.3603,
"step": 67950
},
{
"epoch": 11.564625850340136,
"grad_norm": 0.4113200902938843,
"learning_rate": 0.00011732598607888632,
"loss": 1.3627,
"step": 68000
},
{
"epoch": 11.564625850340136,
"eval_loss": 1.5228930711746216,
"eval_runtime": 75.2543,
"eval_samples_per_second": 1246.958,
"eval_steps_per_second": 4.877,
"step": 68000
},
{
"epoch": 11.57312925170068,
"grad_norm": 0.43190208077430725,
"learning_rate": 0.00011703596287703016,
"loss": 1.354,
"step": 68050
},
{
"epoch": 11.581632653061224,
"grad_norm": 0.41988831758499146,
"learning_rate": 0.00011674593967517402,
"loss": 1.3598,
"step": 68100
},
{
"epoch": 11.59013605442177,
"grad_norm": 0.41497015953063965,
"learning_rate": 0.00011645591647331787,
"loss": 1.3432,
"step": 68150
},
{
"epoch": 11.598639455782314,
"grad_norm": 0.42652803659439087,
"learning_rate": 0.00011616589327146172,
"loss": 1.3565,
"step": 68200
},
{
"epoch": 11.607142857142858,
"grad_norm": 0.43378782272338867,
"learning_rate": 0.00011587587006960557,
"loss": 1.3538,
"step": 68250
},
{
"epoch": 11.615646258503402,
"grad_norm": 0.4106690287590027,
"learning_rate": 0.00011558584686774943,
"loss": 1.3514,
"step": 68300
},
{
"epoch": 11.624149659863946,
"grad_norm": 0.4098680913448334,
"learning_rate": 0.00011529582366589328,
"loss": 1.3628,
"step": 68350
},
{
"epoch": 11.63265306122449,
"grad_norm": 0.4068293869495392,
"learning_rate": 0.00011500580046403712,
"loss": 1.3561,
"step": 68400
},
{
"epoch": 11.641156462585034,
"grad_norm": 0.4472546875476837,
"learning_rate": 0.00011471577726218098,
"loss": 1.3565,
"step": 68450
},
{
"epoch": 11.649659863945578,
"grad_norm": 0.4033961594104767,
"learning_rate": 0.00011442575406032482,
"loss": 1.3545,
"step": 68500
},
{
"epoch": 11.658163265306122,
"grad_norm": 0.4254186749458313,
"learning_rate": 0.00011413573085846868,
"loss": 1.3571,
"step": 68550
},
{
"epoch": 11.666666666666666,
"grad_norm": 0.412675142288208,
"learning_rate": 0.00011384570765661254,
"loss": 1.3513,
"step": 68600
},
{
"epoch": 11.67517006802721,
"grad_norm": 0.4190445840358734,
"learning_rate": 0.00011355568445475637,
"loss": 1.3549,
"step": 68650
},
{
"epoch": 11.683673469387756,
"grad_norm": 0.4222116470336914,
"learning_rate": 0.00011326566125290024,
"loss": 1.3561,
"step": 68700
},
{
"epoch": 11.6921768707483,
"grad_norm": 0.41736775636672974,
"learning_rate": 0.0001129756380510441,
"loss": 1.3541,
"step": 68750
},
{
"epoch": 11.700680272108844,
"grad_norm": 0.40362343192100525,
"learning_rate": 0.00011268561484918793,
"loss": 1.3489,
"step": 68800
},
{
"epoch": 11.709183673469388,
"grad_norm": 0.40062522888183594,
"learning_rate": 0.00011239559164733179,
"loss": 1.355,
"step": 68850
},
{
"epoch": 11.717687074829932,
"grad_norm": 0.40170249342918396,
"learning_rate": 0.00011210556844547564,
"loss": 1.3549,
"step": 68900
},
{
"epoch": 11.726190476190476,
"grad_norm": 0.4284593164920807,
"learning_rate": 0.00011181554524361949,
"loss": 1.3502,
"step": 68950
},
{
"epoch": 11.73469387755102,
"grad_norm": 0.4220130145549774,
"learning_rate": 0.00011152552204176335,
"loss": 1.3508,
"step": 69000
},
{
"epoch": 11.73469387755102,
"eval_loss": 1.517913818359375,
"eval_runtime": 75.269,
"eval_samples_per_second": 1246.716,
"eval_steps_per_second": 4.876,
"step": 69000
},
{
"epoch": 11.743197278911564,
"grad_norm": 0.41656509041786194,
"learning_rate": 0.0001112354988399072,
"loss": 1.3573,
"step": 69050
},
{
"epoch": 11.751700680272108,
"grad_norm": 0.4296277165412903,
"learning_rate": 0.00011094547563805104,
"loss": 1.3586,
"step": 69100
},
{
"epoch": 11.760204081632653,
"grad_norm": 0.5169918537139893,
"learning_rate": 0.0001106554524361949,
"loss": 1.3502,
"step": 69150
},
{
"epoch": 11.768707482993197,
"grad_norm": 0.40869855880737305,
"learning_rate": 0.00011036542923433875,
"loss": 1.3586,
"step": 69200
},
{
"epoch": 11.777210884353742,
"grad_norm": 0.41961830854415894,
"learning_rate": 0.0001100754060324826,
"loss": 1.3563,
"step": 69250
},
{
"epoch": 11.785714285714286,
"grad_norm": 0.40917539596557617,
"learning_rate": 0.00010978538283062645,
"loss": 1.3554,
"step": 69300
},
{
"epoch": 11.79421768707483,
"grad_norm": 0.4206612706184387,
"learning_rate": 0.00010949535962877031,
"loss": 1.3568,
"step": 69350
},
{
"epoch": 11.802721088435375,
"grad_norm": 0.41846826672554016,
"learning_rate": 0.00010920533642691415,
"loss": 1.3518,
"step": 69400
},
{
"epoch": 11.811224489795919,
"grad_norm": 0.4007967412471771,
"learning_rate": 0.000108915313225058,
"loss": 1.3488,
"step": 69450
},
{
"epoch": 11.819727891156463,
"grad_norm": 0.45398712158203125,
"learning_rate": 0.00010862529002320186,
"loss": 1.3522,
"step": 69500
},
{
"epoch": 11.828231292517007,
"grad_norm": 0.4125469923019409,
"learning_rate": 0.00010833526682134571,
"loss": 1.3475,
"step": 69550
},
{
"epoch": 11.83673469387755,
"grad_norm": 0.4218948781490326,
"learning_rate": 0.00010804524361948956,
"loss": 1.3476,
"step": 69600
},
{
"epoch": 11.845238095238095,
"grad_norm": 0.4130956828594208,
"learning_rate": 0.00010775522041763342,
"loss": 1.3406,
"step": 69650
},
{
"epoch": 11.853741496598639,
"grad_norm": 0.4252954125404358,
"learning_rate": 0.00010746519721577725,
"loss": 1.3497,
"step": 69700
},
{
"epoch": 11.862244897959183,
"grad_norm": 0.4810144603252411,
"learning_rate": 0.00010717517401392111,
"loss": 1.3513,
"step": 69750
},
{
"epoch": 11.870748299319727,
"grad_norm": 0.4183041453361511,
"learning_rate": 0.00010688515081206498,
"loss": 1.3464,
"step": 69800
},
{
"epoch": 11.879251700680273,
"grad_norm": 0.4231345057487488,
"learning_rate": 0.00010659512761020881,
"loss": 1.348,
"step": 69850
},
{
"epoch": 11.887755102040817,
"grad_norm": 0.4193786084651947,
"learning_rate": 0.00010630510440835267,
"loss": 1.3494,
"step": 69900
},
{
"epoch": 11.896258503401361,
"grad_norm": 0.41639646887779236,
"learning_rate": 0.00010601508120649653,
"loss": 1.3489,
"step": 69950
},
{
"epoch": 11.904761904761905,
"grad_norm": 0.4062506854534149,
"learning_rate": 0.00010572505800464037,
"loss": 1.3478,
"step": 70000
},
{
"epoch": 11.904761904761905,
"eval_loss": 1.5139620304107666,
"eval_runtime": 75.2979,
"eval_samples_per_second": 1246.237,
"eval_steps_per_second": 4.874,
"step": 70000
},
{
"epoch": 11.91326530612245,
"grad_norm": 0.49928614497184753,
"learning_rate": 0.00010543503480278423,
"loss": 1.3525,
"step": 70050
},
{
"epoch": 11.921768707482993,
"grad_norm": 0.4152166247367859,
"learning_rate": 0.0001051508120649652,
"loss": 1.3526,
"step": 70100
},
{
"epoch": 11.930272108843537,
"grad_norm": 0.42939451336860657,
"learning_rate": 0.00010486078886310905,
"loss": 1.352,
"step": 70150
},
{
"epoch": 11.938775510204081,
"grad_norm": 0.40568697452545166,
"learning_rate": 0.0001045707656612529,
"loss": 1.3493,
"step": 70200
},
{
"epoch": 11.947278911564625,
"grad_norm": 0.43555957078933716,
"learning_rate": 0.00010428074245939675,
"loss": 1.3425,
"step": 70250
},
{
"epoch": 11.95578231292517,
"grad_norm": 0.4200745224952698,
"learning_rate": 0.00010399071925754061,
"loss": 1.3477,
"step": 70300
},
{
"epoch": 11.964285714285714,
"grad_norm": 0.43165090680122375,
"learning_rate": 0.00010370069605568446,
"loss": 1.3522,
"step": 70350
},
{
"epoch": 11.972789115646258,
"grad_norm": 0.41488534212112427,
"learning_rate": 0.0001034106728538283,
"loss": 1.3509,
"step": 70400
},
{
"epoch": 11.981292517006803,
"grad_norm": 0.4268278479576111,
"learning_rate": 0.00010312064965197215,
"loss": 1.3567,
"step": 70450
},
{
"epoch": 11.989795918367347,
"grad_norm": 0.4317072927951813,
"learning_rate": 0.00010283062645011601,
"loss": 1.3441,
"step": 70500
},
{
"epoch": 11.998299319727892,
"grad_norm": 0.39999693632125854,
"learning_rate": 0.00010254060324825986,
"loss": 1.3486,
"step": 70550
},
{
"epoch": 12.006802721088436,
"grad_norm": 0.39868417382240295,
"learning_rate": 0.00010225058004640371,
"loss": 1.3422,
"step": 70600
},
{
"epoch": 12.01530612244898,
"grad_norm": 0.4184889793395996,
"learning_rate": 0.00010196055684454757,
"loss": 1.342,
"step": 70650
},
{
"epoch": 12.023809523809524,
"grad_norm": 0.4223806858062744,
"learning_rate": 0.00010167053364269142,
"loss": 1.339,
"step": 70700
},
{
"epoch": 12.032312925170068,
"grad_norm": 0.413501113653183,
"learning_rate": 0.00010138051044083527,
"loss": 1.3458,
"step": 70750
},
{
"epoch": 12.040816326530612,
"grad_norm": 0.41243672370910645,
"learning_rate": 0.00010109048723897913,
"loss": 1.3497,
"step": 70800
},
{
"epoch": 12.049319727891156,
"grad_norm": 0.4220627546310425,
"learning_rate": 0.00010080046403712296,
"loss": 1.3465,
"step": 70850
},
{
"epoch": 12.0578231292517,
"grad_norm": 0.4198216199874878,
"learning_rate": 0.00010051044083526682,
"loss": 1.3477,
"step": 70900
},
{
"epoch": 12.066326530612244,
"grad_norm": 0.4057486951351166,
"learning_rate": 0.00010022041763341068,
"loss": 1.3406,
"step": 70950
},
{
"epoch": 12.07482993197279,
"grad_norm": 0.4336649179458618,
"learning_rate": 9.993039443155452e-05,
"loss": 1.342,
"step": 71000
},
{
"epoch": 12.07482993197279,
"eval_loss": 1.5144281387329102,
"eval_runtime": 75.2698,
"eval_samples_per_second": 1246.702,
"eval_steps_per_second": 4.876,
"step": 71000
},
{
"epoch": 12.083333333333334,
"grad_norm": 0.4182901680469513,
"learning_rate": 9.964037122969838e-05,
"loss": 1.3432,
"step": 71050
},
{
"epoch": 12.091836734693878,
"grad_norm": 0.39864903688430786,
"learning_rate": 9.935034802784224e-05,
"loss": 1.3367,
"step": 71100
},
{
"epoch": 12.100340136054422,
"grad_norm": 0.40879112482070923,
"learning_rate": 9.906032482598607e-05,
"loss": 1.3431,
"step": 71150
},
{
"epoch": 12.108843537414966,
"grad_norm": 0.412240594625473,
"learning_rate": 9.877030162412993e-05,
"loss": 1.3406,
"step": 71200
},
{
"epoch": 12.11734693877551,
"grad_norm": 0.4245721995830536,
"learning_rate": 9.848027842227378e-05,
"loss": 1.3453,
"step": 71250
},
{
"epoch": 12.125850340136054,
"grad_norm": 0.40284958481788635,
"learning_rate": 9.819025522041763e-05,
"loss": 1.3383,
"step": 71300
},
{
"epoch": 12.134353741496598,
"grad_norm": 0.57562255859375,
"learning_rate": 9.790023201856149e-05,
"loss": 1.3417,
"step": 71350
},
{
"epoch": 12.142857142857142,
"grad_norm": 0.4117988348007202,
"learning_rate": 9.761020881670534e-05,
"loss": 1.3371,
"step": 71400
},
{
"epoch": 12.151360544217686,
"grad_norm": 0.4112182557582855,
"learning_rate": 9.732018561484918e-05,
"loss": 1.3447,
"step": 71450
},
{
"epoch": 12.15986394557823,
"grad_norm": 0.41666939854621887,
"learning_rate": 9.703016241299305e-05,
"loss": 1.3408,
"step": 71500
},
{
"epoch": 12.168367346938776,
"grad_norm": 0.4010598063468933,
"learning_rate": 9.674013921113689e-05,
"loss": 1.3443,
"step": 71550
},
{
"epoch": 12.17687074829932,
"grad_norm": 0.41796889901161194,
"learning_rate": 9.645011600928074e-05,
"loss": 1.3343,
"step": 71600
},
{
"epoch": 12.185374149659864,
"grad_norm": 0.42123886942863464,
"learning_rate": 9.616009280742459e-05,
"loss": 1.3401,
"step": 71650
},
{
"epoch": 12.193877551020408,
"grad_norm": 0.40508341789245605,
"learning_rate": 9.587006960556845e-05,
"loss": 1.3437,
"step": 71700
},
{
"epoch": 12.202380952380953,
"grad_norm": 0.44286808371543884,
"learning_rate": 9.558004640371231e-05,
"loss": 1.3419,
"step": 71750
},
{
"epoch": 12.210884353741497,
"grad_norm": 0.40285545587539673,
"learning_rate": 9.529002320185614e-05,
"loss": 1.333,
"step": 71800
},
{
"epoch": 12.21938775510204,
"grad_norm": 0.39993083477020264,
"learning_rate": 9.5e-05,
"loss": 1.3402,
"step": 71850
},
{
"epoch": 12.227891156462585,
"grad_norm": 0.4155488610267639,
"learning_rate": 9.470997679814387e-05,
"loss": 1.3378,
"step": 71900
},
{
"epoch": 12.236394557823129,
"grad_norm": 0.5050736665725708,
"learning_rate": 9.44199535962877e-05,
"loss": 1.3354,
"step": 71950
},
{
"epoch": 12.244897959183673,
"grad_norm": 0.4185927212238312,
"learning_rate": 9.412993039443156e-05,
"loss": 1.3357,
"step": 72000
},
{
"epoch": 12.244897959183673,
"eval_loss": 1.5064618587493896,
"eval_runtime": 75.3162,
"eval_samples_per_second": 1245.933,
"eval_steps_per_second": 4.873,
"step": 72000
},
{
"epoch": 12.253401360544217,
"grad_norm": 0.41753119230270386,
"learning_rate": 9.383990719257541e-05,
"loss": 1.3352,
"step": 72050
},
{
"epoch": 12.261904761904763,
"grad_norm": 0.4204673171043396,
"learning_rate": 9.354988399071926e-05,
"loss": 1.3436,
"step": 72100
},
{
"epoch": 12.270408163265307,
"grad_norm": 0.4101797044277191,
"learning_rate": 9.325986078886312e-05,
"loss": 1.3385,
"step": 72150
},
{
"epoch": 12.27891156462585,
"grad_norm": 0.4880974292755127,
"learning_rate": 9.296983758700696e-05,
"loss": 1.3322,
"step": 72200
},
{
"epoch": 12.287414965986395,
"grad_norm": 0.415211945772171,
"learning_rate": 9.267981438515081e-05,
"loss": 1.3409,
"step": 72250
},
{
"epoch": 12.295918367346939,
"grad_norm": 0.41376402974128723,
"learning_rate": 9.238979118329467e-05,
"loss": 1.3392,
"step": 72300
},
{
"epoch": 12.304421768707483,
"grad_norm": 0.43394094705581665,
"learning_rate": 9.209976798143852e-05,
"loss": 1.3345,
"step": 72350
},
{
"epoch": 12.312925170068027,
"grad_norm": 0.43497174978256226,
"learning_rate": 9.180974477958237e-05,
"loss": 1.3381,
"step": 72400
},
{
"epoch": 12.321428571428571,
"grad_norm": 0.4110974371433258,
"learning_rate": 9.151972157772622e-05,
"loss": 1.3375,
"step": 72450
},
{
"epoch": 12.329931972789115,
"grad_norm": 0.4111982583999634,
"learning_rate": 9.12354988399072e-05,
"loss": 1.3352,
"step": 72500
},
{
"epoch": 12.33843537414966,
"grad_norm": 0.4248039126396179,
"learning_rate": 9.094547563805104e-05,
"loss": 1.3353,
"step": 72550
},
{
"epoch": 12.346938775510203,
"grad_norm": 0.4378061294555664,
"learning_rate": 9.06554524361949e-05,
"loss": 1.3397,
"step": 72600
},
{
"epoch": 12.35544217687075,
"grad_norm": 0.41765543818473816,
"learning_rate": 9.036542923433875e-05,
"loss": 1.3372,
"step": 72650
},
{
"epoch": 12.363945578231293,
"grad_norm": 0.4655171036720276,
"learning_rate": 9.00754060324826e-05,
"loss": 1.337,
"step": 72700
},
{
"epoch": 12.372448979591837,
"grad_norm": 0.4178127348423004,
"learning_rate": 8.978538283062646e-05,
"loss": 1.3362,
"step": 72750
},
{
"epoch": 12.380952380952381,
"grad_norm": 0.4332452118396759,
"learning_rate": 8.94953596287703e-05,
"loss": 1.3379,
"step": 72800
},
{
"epoch": 12.389455782312925,
"grad_norm": 0.44955676794052124,
"learning_rate": 8.920533642691416e-05,
"loss": 1.3384,
"step": 72850
},
{
"epoch": 12.39795918367347,
"grad_norm": 0.42252740263938904,
"learning_rate": 8.891531322505802e-05,
"loss": 1.3398,
"step": 72900
},
{
"epoch": 12.406462585034014,
"grad_norm": 0.3996742069721222,
"learning_rate": 8.862529002320185e-05,
"loss": 1.3294,
"step": 72950
},
{
"epoch": 12.414965986394558,
"grad_norm": 0.5868241786956787,
"learning_rate": 8.833526682134571e-05,
"loss": 1.3347,
"step": 73000
},
{
"epoch": 12.414965986394558,
"eval_loss": 1.4998681545257568,
"eval_runtime": 75.3074,
"eval_samples_per_second": 1246.079,
"eval_steps_per_second": 4.873,
"step": 73000
},
{
"epoch": 12.423469387755102,
"grad_norm": 0.4118203818798065,
"learning_rate": 8.804524361948957e-05,
"loss": 1.3422,
"step": 73050
},
{
"epoch": 12.431972789115646,
"grad_norm": 0.5292280316352844,
"learning_rate": 8.775522041763341e-05,
"loss": 1.3353,
"step": 73100
},
{
"epoch": 12.44047619047619,
"grad_norm": 0.42021283507347107,
"learning_rate": 8.747099767981439e-05,
"loss": 1.3379,
"step": 73150
},
{
"epoch": 12.448979591836734,
"grad_norm": 0.42471328377723694,
"learning_rate": 8.718097447795824e-05,
"loss": 1.3377,
"step": 73200
},
{
"epoch": 12.45748299319728,
"grad_norm": 0.4189905524253845,
"learning_rate": 8.68909512761021e-05,
"loss": 1.3338,
"step": 73250
},
{
"epoch": 12.465986394557824,
"grad_norm": 0.4137709438800812,
"learning_rate": 8.660092807424594e-05,
"loss": 1.3341,
"step": 73300
},
{
"epoch": 12.474489795918368,
"grad_norm": 0.41046592593193054,
"learning_rate": 8.631090487238979e-05,
"loss": 1.3325,
"step": 73350
},
{
"epoch": 12.482993197278912,
"grad_norm": 0.40613409876823425,
"learning_rate": 8.602088167053364e-05,
"loss": 1.334,
"step": 73400
},
{
"epoch": 12.491496598639456,
"grad_norm": 0.4053729474544525,
"learning_rate": 8.57308584686775e-05,
"loss": 1.3297,
"step": 73450
},
{
"epoch": 12.5,
"grad_norm": 0.4099692404270172,
"learning_rate": 8.544083526682135e-05,
"loss": 1.3303,
"step": 73500
},
{
"epoch": 12.508503401360544,
"grad_norm": 0.40552228689193726,
"learning_rate": 8.51508120649652e-05,
"loss": 1.3286,
"step": 73550
},
{
"epoch": 12.517006802721088,
"grad_norm": 0.41349565982818604,
"learning_rate": 8.486078886310906e-05,
"loss": 1.327,
"step": 73600
},
{
"epoch": 12.525510204081632,
"grad_norm": 0.40340515971183777,
"learning_rate": 8.45707656612529e-05,
"loss": 1.3282,
"step": 73650
},
{
"epoch": 12.534013605442176,
"grad_norm": 0.4043871760368347,
"learning_rate": 8.428074245939675e-05,
"loss": 1.3316,
"step": 73700
},
{
"epoch": 12.54251700680272,
"grad_norm": 0.41814517974853516,
"learning_rate": 8.399071925754061e-05,
"loss": 1.3278,
"step": 73750
},
{
"epoch": 12.551020408163264,
"grad_norm": 0.41396060585975647,
"learning_rate": 8.370069605568445e-05,
"loss": 1.3283,
"step": 73800
},
{
"epoch": 12.55952380952381,
"grad_norm": 0.4073675572872162,
"learning_rate": 8.341067285382831e-05,
"loss": 1.3274,
"step": 73850
},
{
"epoch": 12.568027210884354,
"grad_norm": 0.40986111760139465,
"learning_rate": 8.312064965197217e-05,
"loss": 1.3252,
"step": 73900
},
{
"epoch": 12.576530612244898,
"grad_norm": 0.3996681869029999,
"learning_rate": 8.2830626450116e-05,
"loss": 1.3308,
"step": 73950
},
{
"epoch": 12.585034013605442,
"grad_norm": 0.4171459376811981,
"learning_rate": 8.254060324825986e-05,
"loss": 1.3263,
"step": 74000
},
{
"epoch": 12.585034013605442,
"eval_loss": 1.4951106309890747,
"eval_runtime": 75.2691,
"eval_samples_per_second": 1246.714,
"eval_steps_per_second": 4.876,
"step": 74000
},
{
"epoch": 12.593537414965986,
"grad_norm": 0.4198111593723297,
"learning_rate": 8.225058004640372e-05,
"loss": 1.3306,
"step": 74050
},
{
"epoch": 12.60204081632653,
"grad_norm": 0.4122917056083679,
"learning_rate": 8.196055684454756e-05,
"loss": 1.3325,
"step": 74100
},
{
"epoch": 12.610544217687075,
"grad_norm": 0.43273040652275085,
"learning_rate": 8.167053364269142e-05,
"loss": 1.3385,
"step": 74150
},
{
"epoch": 12.619047619047619,
"grad_norm": 0.4447864294052124,
"learning_rate": 8.138051044083527e-05,
"loss": 1.3386,
"step": 74200
},
{
"epoch": 12.627551020408163,
"grad_norm": 0.41947802901268005,
"learning_rate": 8.109048723897911e-05,
"loss": 1.3313,
"step": 74250
},
{
"epoch": 12.636054421768707,
"grad_norm": 0.41405126452445984,
"learning_rate": 8.080046403712298e-05,
"loss": 1.3331,
"step": 74300
},
{
"epoch": 12.64455782312925,
"grad_norm": 0.407975971698761,
"learning_rate": 8.051044083526682e-05,
"loss": 1.3243,
"step": 74350
},
{
"epoch": 12.653061224489797,
"grad_norm": 0.3966875374317169,
"learning_rate": 8.022041763341067e-05,
"loss": 1.3276,
"step": 74400
},
{
"epoch": 12.66156462585034,
"grad_norm": 0.43828269839286804,
"learning_rate": 7.993619489559165e-05,
"loss": 1.335,
"step": 74450
},
{
"epoch": 12.670068027210885,
"grad_norm": 0.40009811520576477,
"learning_rate": 7.96461716937355e-05,
"loss": 1.3217,
"step": 74500
},
{
"epoch": 12.678571428571429,
"grad_norm": 0.40851446986198425,
"learning_rate": 7.935614849187935e-05,
"loss": 1.3338,
"step": 74550
},
{
"epoch": 12.687074829931973,
"grad_norm": 0.4100090265274048,
"learning_rate": 7.906612529002321e-05,
"loss": 1.334,
"step": 74600
},
{
"epoch": 12.695578231292517,
"grad_norm": 0.41323432326316833,
"learning_rate": 7.877610208816705e-05,
"loss": 1.3279,
"step": 74650
},
{
"epoch": 12.704081632653061,
"grad_norm": 0.4219471514225006,
"learning_rate": 7.84860788863109e-05,
"loss": 1.3265,
"step": 74700
},
{
"epoch": 12.712585034013605,
"grad_norm": 0.4054030478000641,
"learning_rate": 7.819605568445476e-05,
"loss": 1.3275,
"step": 74750
},
{
"epoch": 12.72108843537415,
"grad_norm": 0.4156668782234192,
"learning_rate": 7.790603248259861e-05,
"loss": 1.3246,
"step": 74800
},
{
"epoch": 12.729591836734693,
"grad_norm": 0.4128476083278656,
"learning_rate": 7.761600928074246e-05,
"loss": 1.325,
"step": 74850
},
{
"epoch": 12.738095238095237,
"grad_norm": 0.4204627275466919,
"learning_rate": 7.732598607888632e-05,
"loss": 1.3303,
"step": 74900
},
{
"epoch": 12.746598639455783,
"grad_norm": 0.4113794267177582,
"learning_rate": 7.703596287703015e-05,
"loss": 1.3297,
"step": 74950
},
{
"epoch": 12.755102040816327,
"grad_norm": 0.4143586754798889,
"learning_rate": 7.674593967517401e-05,
"loss": 1.323,
"step": 75000
},
{
"epoch": 12.755102040816327,
"eval_loss": 1.4961031675338745,
"eval_runtime": 75.2707,
"eval_samples_per_second": 1246.687,
"eval_steps_per_second": 4.876,
"step": 75000
},
{
"epoch": 12.763605442176871,
"grad_norm": 0.4443499743938446,
"learning_rate": 7.645591647331788e-05,
"loss": 1.3285,
"step": 75050
},
{
"epoch": 12.772108843537415,
"grad_norm": 0.4180935025215149,
"learning_rate": 7.616589327146171e-05,
"loss": 1.3299,
"step": 75100
},
{
"epoch": 12.78061224489796,
"grad_norm": 0.41669797897338867,
"learning_rate": 7.587587006960557e-05,
"loss": 1.3314,
"step": 75150
},
{
"epoch": 12.789115646258503,
"grad_norm": 0.40955501794815063,
"learning_rate": 7.558584686774943e-05,
"loss": 1.3285,
"step": 75200
},
{
"epoch": 12.797619047619047,
"grad_norm": 0.4140739142894745,
"learning_rate": 7.529582366589327e-05,
"loss": 1.331,
"step": 75250
},
{
"epoch": 12.806122448979592,
"grad_norm": 0.4092477858066559,
"learning_rate": 7.500580046403713e-05,
"loss": 1.3268,
"step": 75300
},
{
"epoch": 12.814625850340136,
"grad_norm": 0.412955105304718,
"learning_rate": 7.471577726218097e-05,
"loss": 1.324,
"step": 75350
},
{
"epoch": 12.82312925170068,
"grad_norm": 0.4366496205329895,
"learning_rate": 7.442575406032482e-05,
"loss": 1.3257,
"step": 75400
},
{
"epoch": 12.831632653061224,
"grad_norm": 0.4090063273906708,
"learning_rate": 7.413573085846868e-05,
"loss": 1.3261,
"step": 75450
},
{
"epoch": 12.84013605442177,
"grad_norm": 0.41486817598342896,
"learning_rate": 7.384570765661253e-05,
"loss": 1.3224,
"step": 75500
},
{
"epoch": 12.848639455782314,
"grad_norm": 0.40485501289367676,
"learning_rate": 7.355568445475638e-05,
"loss": 1.3285,
"step": 75550
},
{
"epoch": 12.857142857142858,
"grad_norm": 0.43988388776779175,
"learning_rate": 7.326566125290024e-05,
"loss": 1.3245,
"step": 75600
},
{
"epoch": 12.865646258503402,
"grad_norm": 0.43231526017189026,
"learning_rate": 7.297563805104409e-05,
"loss": 1.3266,
"step": 75650
},
{
"epoch": 12.874149659863946,
"grad_norm": 0.40446582436561584,
"learning_rate": 7.268561484918793e-05,
"loss": 1.3248,
"step": 75700
},
{
"epoch": 12.88265306122449,
"grad_norm": 0.40424802899360657,
"learning_rate": 7.239559164733178e-05,
"loss": 1.3187,
"step": 75750
},
{
"epoch": 12.891156462585034,
"grad_norm": 0.410151869058609,
"learning_rate": 7.210556844547564e-05,
"loss": 1.3212,
"step": 75800
},
{
"epoch": 12.899659863945578,
"grad_norm": 0.41777828335762024,
"learning_rate": 7.181554524361949e-05,
"loss": 1.3204,
"step": 75850
},
{
"epoch": 12.908163265306122,
"grad_norm": 0.4275606870651245,
"learning_rate": 7.152552204176334e-05,
"loss": 1.3239,
"step": 75900
},
{
"epoch": 12.916666666666666,
"grad_norm": 0.4166664481163025,
"learning_rate": 7.12354988399072e-05,
"loss": 1.3174,
"step": 75950
},
{
"epoch": 12.92517006802721,
"grad_norm": 0.3995276391506195,
"learning_rate": 7.094547563805105e-05,
"loss": 1.3231,
"step": 76000
},
{
"epoch": 12.92517006802721,
"eval_loss": 1.490061640739441,
"eval_runtime": 75.2857,
"eval_samples_per_second": 1246.439,
"eval_steps_per_second": 4.875,
"step": 76000
},
{
"epoch": 12.933673469387756,
"grad_norm": 0.4179072082042694,
"learning_rate": 7.06554524361949e-05,
"loss": 1.3217,
"step": 76050
},
{
"epoch": 12.9421768707483,
"grad_norm": 0.43183988332748413,
"learning_rate": 7.036542923433875e-05,
"loss": 1.3218,
"step": 76100
},
{
"epoch": 12.950680272108844,
"grad_norm": 0.4066503345966339,
"learning_rate": 7.007540603248259e-05,
"loss": 1.326,
"step": 76150
},
{
"epoch": 12.959183673469388,
"grad_norm": 0.4215805232524872,
"learning_rate": 6.978538283062645e-05,
"loss": 1.325,
"step": 76200
},
{
"epoch": 12.967687074829932,
"grad_norm": 0.4248071312904358,
"learning_rate": 6.949535962877031e-05,
"loss": 1.3179,
"step": 76250
},
{
"epoch": 12.976190476190476,
"grad_norm": 0.42110559344291687,
"learning_rate": 6.920533642691414e-05,
"loss": 1.3292,
"step": 76300
},
{
"epoch": 12.98469387755102,
"grad_norm": 0.4082868695259094,
"learning_rate": 6.8915313225058e-05,
"loss": 1.3254,
"step": 76350
},
{
"epoch": 12.993197278911564,
"grad_norm": 0.4124908447265625,
"learning_rate": 6.862529002320187e-05,
"loss": 1.3227,
"step": 76400
},
{
"epoch": 13.001700680272108,
"grad_norm": 0.5426220297813416,
"learning_rate": 6.83352668213457e-05,
"loss": 1.3198,
"step": 76450
},
{
"epoch": 13.010204081632653,
"grad_norm": 0.4062071442604065,
"learning_rate": 6.804524361948956e-05,
"loss": 1.3183,
"step": 76500
},
{
"epoch": 13.018707482993197,
"grad_norm": 0.40301281213760376,
"learning_rate": 6.775522041763341e-05,
"loss": 1.3232,
"step": 76550
},
{
"epoch": 13.02721088435374,
"grad_norm": 0.44012153148651123,
"learning_rate": 6.746519721577726e-05,
"loss": 1.318,
"step": 76600
},
{
"epoch": 13.035714285714286,
"grad_norm": 0.41154050827026367,
"learning_rate": 6.717517401392112e-05,
"loss": 1.3176,
"step": 76650
},
{
"epoch": 13.04421768707483,
"grad_norm": 0.42961668968200684,
"learning_rate": 6.688515081206497e-05,
"loss": 1.3188,
"step": 76700
},
{
"epoch": 13.052721088435375,
"grad_norm": 0.41780275106430054,
"learning_rate": 6.659512761020881e-05,
"loss": 1.3191,
"step": 76750
},
{
"epoch": 13.061224489795919,
"grad_norm": 0.4239339828491211,
"learning_rate": 6.630510440835267e-05,
"loss": 1.3225,
"step": 76800
},
{
"epoch": 13.069727891156463,
"grad_norm": 0.40867921710014343,
"learning_rate": 6.601508120649652e-05,
"loss": 1.3254,
"step": 76850
},
{
"epoch": 13.078231292517007,
"grad_norm": 0.4113767743110657,
"learning_rate": 6.572505800464038e-05,
"loss": 1.3151,
"step": 76900
},
{
"epoch": 13.08673469387755,
"grad_norm": 0.4053804874420166,
"learning_rate": 6.543503480278422e-05,
"loss": 1.3196,
"step": 76950
},
{
"epoch": 13.095238095238095,
"grad_norm": 0.40238648653030396,
"learning_rate": 6.514501160092808e-05,
"loss": 1.3119,
"step": 77000
},
{
"epoch": 13.095238095238095,
"eval_loss": 1.486020803451538,
"eval_runtime": 75.2657,
"eval_samples_per_second": 1246.769,
"eval_steps_per_second": 4.876,
"step": 77000
},
{
"epoch": 13.103741496598639,
"grad_norm": 0.4077170193195343,
"learning_rate": 6.485498839907194e-05,
"loss": 1.3185,
"step": 77050
},
{
"epoch": 13.112244897959183,
"grad_norm": 0.40546542406082153,
"learning_rate": 6.456496519721577e-05,
"loss": 1.3221,
"step": 77100
},
{
"epoch": 13.120748299319727,
"grad_norm": 0.4073767364025116,
"learning_rate": 6.427494199535963e-05,
"loss": 1.3168,
"step": 77150
},
{
"epoch": 13.129251700680273,
"grad_norm": 0.4032026529312134,
"learning_rate": 6.39849187935035e-05,
"loss": 1.3162,
"step": 77200
},
{
"epoch": 13.137755102040817,
"grad_norm": 0.4208144545555115,
"learning_rate": 6.369489559164733e-05,
"loss": 1.3218,
"step": 77250
},
{
"epoch": 13.146258503401361,
"grad_norm": 0.4274178743362427,
"learning_rate": 6.340487238979119e-05,
"loss": 1.3165,
"step": 77300
},
{
"epoch": 13.154761904761905,
"grad_norm": 0.41824012994766235,
"learning_rate": 6.311484918793504e-05,
"loss": 1.3185,
"step": 77350
},
{
"epoch": 13.16326530612245,
"grad_norm": 0.395487517118454,
"learning_rate": 6.282482598607888e-05,
"loss": 1.3115,
"step": 77400
},
{
"epoch": 13.171768707482993,
"grad_norm": 0.4073810577392578,
"learning_rate": 6.253480278422275e-05,
"loss": 1.3156,
"step": 77450
},
{
"epoch": 13.180272108843537,
"grad_norm": 0.4188859760761261,
"learning_rate": 6.224477958236659e-05,
"loss": 1.3164,
"step": 77500
},
{
"epoch": 13.188775510204081,
"grad_norm": 0.4151723086833954,
"learning_rate": 6.195475638051044e-05,
"loss": 1.316,
"step": 77550
},
{
"epoch": 13.197278911564625,
"grad_norm": 0.42075613141059875,
"learning_rate": 6.166473317865429e-05,
"loss": 1.3136,
"step": 77600
},
{
"epoch": 13.20578231292517,
"grad_norm": 0.41158735752105713,
"learning_rate": 6.137470997679815e-05,
"loss": 1.3157,
"step": 77650
},
{
"epoch": 13.214285714285714,
"grad_norm": 0.4071219563484192,
"learning_rate": 6.1084686774942e-05,
"loss": 1.3167,
"step": 77700
},
{
"epoch": 13.22278911564626,
"grad_norm": 0.430393248796463,
"learning_rate": 6.079466357308585e-05,
"loss": 1.3142,
"step": 77750
},
{
"epoch": 13.231292517006803,
"grad_norm": 0.402025043964386,
"learning_rate": 6.05046403712297e-05,
"loss": 1.3149,
"step": 77800
},
{
"epoch": 13.239795918367347,
"grad_norm": 0.39697858691215515,
"learning_rate": 6.0214617169373546e-05,
"loss": 1.3163,
"step": 77850
},
{
"epoch": 13.248299319727892,
"grad_norm": 0.45332589745521545,
"learning_rate": 5.992459396751741e-05,
"loss": 1.31,
"step": 77900
},
{
"epoch": 13.256802721088436,
"grad_norm": 0.42384395003318787,
"learning_rate": 5.9634570765661255e-05,
"loss": 1.3102,
"step": 77950
},
{
"epoch": 13.26530612244898,
"grad_norm": 0.42941874265670776,
"learning_rate": 5.934454756380511e-05,
"loss": 1.3184,
"step": 78000
},
{
"epoch": 13.26530612244898,
"eval_loss": 1.4811251163482666,
"eval_runtime": 75.2963,
"eval_samples_per_second": 1246.264,
"eval_steps_per_second": 4.874,
"step": 78000
},
{
"epoch": 13.273809523809524,
"grad_norm": 0.41413602232933044,
"learning_rate": 5.9054524361948956e-05,
"loss": 1.3109,
"step": 78050
},
{
"epoch": 13.282312925170068,
"grad_norm": 0.4183179438114166,
"learning_rate": 5.876450116009281e-05,
"loss": 1.3129,
"step": 78100
},
{
"epoch": 13.290816326530612,
"grad_norm": 0.41364216804504395,
"learning_rate": 5.8474477958236665e-05,
"loss": 1.3132,
"step": 78150
},
{
"epoch": 13.299319727891156,
"grad_norm": 0.41207000613212585,
"learning_rate": 5.818445475638051e-05,
"loss": 1.3238,
"step": 78200
},
{
"epoch": 13.3078231292517,
"grad_norm": 0.42450812458992004,
"learning_rate": 5.789443155452436e-05,
"loss": 1.3113,
"step": 78250
},
{
"epoch": 13.316326530612244,
"grad_norm": 0.4294716715812683,
"learning_rate": 5.760440835266822e-05,
"loss": 1.3195,
"step": 78300
},
{
"epoch": 13.32482993197279,
"grad_norm": 0.4036255478858948,
"learning_rate": 5.731438515081207e-05,
"loss": 1.3148,
"step": 78350
},
{
"epoch": 13.333333333333334,
"grad_norm": 0.3960818946361542,
"learning_rate": 5.7024361948955916e-05,
"loss": 1.3098,
"step": 78400
},
{
"epoch": 13.341836734693878,
"grad_norm": 0.40689873695373535,
"learning_rate": 5.673433874709977e-05,
"loss": 1.3136,
"step": 78450
},
{
"epoch": 13.350340136054422,
"grad_norm": 0.4078476130962372,
"learning_rate": 5.6444315545243625e-05,
"loss": 1.3111,
"step": 78500
},
{
"epoch": 13.358843537414966,
"grad_norm": 0.4066517651081085,
"learning_rate": 5.615429234338747e-05,
"loss": 1.3117,
"step": 78550
},
{
"epoch": 13.36734693877551,
"grad_norm": 0.4128108024597168,
"learning_rate": 5.5864269141531326e-05,
"loss": 1.3108,
"step": 78600
},
{
"epoch": 13.375850340136054,
"grad_norm": 0.40743228793144226,
"learning_rate": 5.5574245939675174e-05,
"loss": 1.3099,
"step": 78650
},
{
"epoch": 13.384353741496598,
"grad_norm": 0.43066033720970154,
"learning_rate": 5.528422273781902e-05,
"loss": 1.312,
"step": 78700
},
{
"epoch": 13.392857142857142,
"grad_norm": 0.3970576226711273,
"learning_rate": 5.499419953596288e-05,
"loss": 1.3128,
"step": 78750
},
{
"epoch": 13.401360544217686,
"grad_norm": 0.42503124475479126,
"learning_rate": 5.470417633410673e-05,
"loss": 1.3057,
"step": 78800
},
{
"epoch": 13.40986394557823,
"grad_norm": 0.4038301408290863,
"learning_rate": 5.441415313225058e-05,
"loss": 1.3093,
"step": 78850
},
{
"epoch": 13.418367346938776,
"grad_norm": 0.4250807762145996,
"learning_rate": 5.412412993039443e-05,
"loss": 1.3086,
"step": 78900
},
{
"epoch": 13.42687074829932,
"grad_norm": 0.4134768843650818,
"learning_rate": 5.3834106728538286e-05,
"loss": 1.3135,
"step": 78950
},
{
"epoch": 13.435374149659864,
"grad_norm": 0.39782798290252686,
"learning_rate": 5.3544083526682134e-05,
"loss": 1.305,
"step": 79000
},
{
"epoch": 13.435374149659864,
"eval_loss": 1.4755269289016724,
"eval_runtime": 75.3173,
"eval_samples_per_second": 1245.915,
"eval_steps_per_second": 4.873,
"step": 79000
},
{
"epoch": 13.443877551020408,
"grad_norm": 0.42068231105804443,
"learning_rate": 5.325406032482599e-05,
"loss": 1.3107,
"step": 79050
},
{
"epoch": 13.452380952380953,
"grad_norm": 0.40822988748550415,
"learning_rate": 5.2964037122969835e-05,
"loss": 1.3076,
"step": 79100
},
{
"epoch": 13.460884353741497,
"grad_norm": 0.42288413643836975,
"learning_rate": 5.267401392111369e-05,
"loss": 1.3128,
"step": 79150
},
{
"epoch": 13.46938775510204,
"grad_norm": 0.41481590270996094,
"learning_rate": 5.2383990719257544e-05,
"loss": 1.3083,
"step": 79200
},
{
"epoch": 13.477891156462585,
"grad_norm": 0.4217277467250824,
"learning_rate": 5.209396751740139e-05,
"loss": 1.2985,
"step": 79250
},
{
"epoch": 13.486394557823129,
"grad_norm": 0.4183300733566284,
"learning_rate": 5.180394431554524e-05,
"loss": 1.3067,
"step": 79300
},
{
"epoch": 13.494897959183673,
"grad_norm": 0.44343501329421997,
"learning_rate": 5.15139211136891e-05,
"loss": 1.3101,
"step": 79350
},
{
"epoch": 13.503401360544217,
"grad_norm": 0.47200411558151245,
"learning_rate": 5.122389791183295e-05,
"loss": 1.3113,
"step": 79400
},
{
"epoch": 13.511904761904763,
"grad_norm": 0.40812692046165466,
"learning_rate": 5.0933874709976795e-05,
"loss": 1.3085,
"step": 79450
},
{
"epoch": 13.520408163265307,
"grad_norm": 0.4125157594680786,
"learning_rate": 5.064385150812065e-05,
"loss": 1.3117,
"step": 79500
},
{
"epoch": 13.52891156462585,
"grad_norm": 0.42148134112358093,
"learning_rate": 5.0353828306264504e-05,
"loss": 1.3096,
"step": 79550
},
{
"epoch": 13.537414965986395,
"grad_norm": 0.4138471782207489,
"learning_rate": 5.006380510440835e-05,
"loss": 1.3099,
"step": 79600
},
{
"epoch": 13.545918367346939,
"grad_norm": 0.7425023317337036,
"learning_rate": 4.9773781902552205e-05,
"loss": 1.3098,
"step": 79650
},
{
"epoch": 13.554421768707483,
"grad_norm": 0.4228558838367462,
"learning_rate": 4.948375870069605e-05,
"loss": 1.3055,
"step": 79700
},
{
"epoch": 13.562925170068027,
"grad_norm": 0.39013656973838806,
"learning_rate": 4.919373549883991e-05,
"loss": 1.3109,
"step": 79750
},
{
"epoch": 13.571428571428571,
"grad_norm": 0.4059963822364807,
"learning_rate": 4.890371229698376e-05,
"loss": 1.3062,
"step": 79800
},
{
"epoch": 13.579931972789115,
"grad_norm": 0.41088271141052246,
"learning_rate": 4.861368909512761e-05,
"loss": 1.3104,
"step": 79850
},
{
"epoch": 13.58843537414966,
"grad_norm": 0.4267512857913971,
"learning_rate": 4.8323665893271456e-05,
"loss": 1.3094,
"step": 79900
},
{
"epoch": 13.596938775510203,
"grad_norm": 0.4166390001773834,
"learning_rate": 4.803364269141532e-05,
"loss": 1.3063,
"step": 79950
},
{
"epoch": 13.60544217687075,
"grad_norm": 0.43058088421821594,
"learning_rate": 4.7743619489559165e-05,
"loss": 1.3092,
"step": 80000
},
{
"epoch": 13.60544217687075,
"eval_loss": 1.4739412069320679,
"eval_runtime": 75.2973,
"eval_samples_per_second": 1246.247,
"eval_steps_per_second": 4.874,
"step": 80000
},
{
"epoch": 13.613945578231293,
"grad_norm": 0.4472084939479828,
"learning_rate": 4.745359628770301e-05,
"loss": 1.3092,
"step": 80050
},
{
"epoch": 13.622448979591837,
"grad_norm": 0.40930888056755066,
"learning_rate": 4.716357308584687e-05,
"loss": 1.2986,
"step": 80100
},
{
"epoch": 13.630952380952381,
"grad_norm": 0.4269891083240509,
"learning_rate": 4.687354988399072e-05,
"loss": 1.3085,
"step": 80150
},
{
"epoch": 13.639455782312925,
"grad_norm": 0.4400569498538971,
"learning_rate": 4.6583526682134575e-05,
"loss": 1.3091,
"step": 80200
},
{
"epoch": 13.64795918367347,
"grad_norm": 0.40618574619293213,
"learning_rate": 4.629350348027842e-05,
"loss": 1.3053,
"step": 80250
},
{
"epoch": 13.656462585034014,
"grad_norm": 0.40548551082611084,
"learning_rate": 4.600348027842227e-05,
"loss": 1.3097,
"step": 80300
},
{
"epoch": 13.664965986394558,
"grad_norm": 0.4029221534729004,
"learning_rate": 4.571345707656613e-05,
"loss": 1.3051,
"step": 80350
},
{
"epoch": 13.673469387755102,
"grad_norm": 0.4272826015949249,
"learning_rate": 4.542343387470998e-05,
"loss": 1.3078,
"step": 80400
},
{
"epoch": 13.681972789115646,
"grad_norm": 0.41242125630378723,
"learning_rate": 4.5133410672853826e-05,
"loss": 1.3068,
"step": 80450
},
{
"epoch": 13.69047619047619,
"grad_norm": 0.405517578125,
"learning_rate": 4.484338747099768e-05,
"loss": 1.3061,
"step": 80500
},
{
"epoch": 13.698979591836736,
"grad_norm": 0.4161379635334015,
"learning_rate": 4.4553364269141535e-05,
"loss": 1.3053,
"step": 80550
},
{
"epoch": 13.70748299319728,
"grad_norm": 0.4203460216522217,
"learning_rate": 4.426334106728538e-05,
"loss": 1.3043,
"step": 80600
},
{
"epoch": 13.715986394557824,
"grad_norm": 0.4214410185813904,
"learning_rate": 4.397331786542924e-05,
"loss": 1.3066,
"step": 80650
},
{
"epoch": 13.724489795918368,
"grad_norm": 0.42133694887161255,
"learning_rate": 4.3683294663573084e-05,
"loss": 1.3143,
"step": 80700
},
{
"epoch": 13.732993197278912,
"grad_norm": 0.39356616139411926,
"learning_rate": 4.339327146171694e-05,
"loss": 1.3078,
"step": 80750
},
{
"epoch": 13.741496598639456,
"grad_norm": 0.41399866342544556,
"learning_rate": 4.310324825986079e-05,
"loss": 1.3112,
"step": 80800
},
{
"epoch": 13.75,
"grad_norm": 0.41746941208839417,
"learning_rate": 4.281322505800464e-05,
"loss": 1.3019,
"step": 80850
},
{
"epoch": 13.758503401360544,
"grad_norm": 0.4164547324180603,
"learning_rate": 4.252320185614849e-05,
"loss": 1.309,
"step": 80900
},
{
"epoch": 13.767006802721088,
"grad_norm": 0.4242299795150757,
"learning_rate": 4.223317865429235e-05,
"loss": 1.3023,
"step": 80950
},
{
"epoch": 13.775510204081632,
"grad_norm": 0.4084872603416443,
"learning_rate": 4.1943155452436197e-05,
"loss": 1.3039,
"step": 81000
},
{
"epoch": 13.775510204081632,
"eval_loss": 1.4660990238189697,
"eval_runtime": 75.2655,
"eval_samples_per_second": 1246.773,
"eval_steps_per_second": 4.876,
"step": 81000
},
{
"epoch": 13.784013605442176,
"grad_norm": 0.4138684868812561,
"learning_rate": 4.1653132250580044e-05,
"loss": 1.3075,
"step": 81050
},
{
"epoch": 13.79251700680272,
"grad_norm": 0.40258848667144775,
"learning_rate": 4.13631090487239e-05,
"loss": 1.3068,
"step": 81100
},
{
"epoch": 13.801020408163264,
"grad_norm": 0.4167514145374298,
"learning_rate": 4.107308584686775e-05,
"loss": 1.305,
"step": 81150
},
{
"epoch": 13.80952380952381,
"grad_norm": 0.41513994336128235,
"learning_rate": 4.07830626450116e-05,
"loss": 1.3087,
"step": 81200
},
{
"epoch": 13.818027210884354,
"grad_norm": 0.42223405838012695,
"learning_rate": 4.0493039443155454e-05,
"loss": 1.3048,
"step": 81250
},
{
"epoch": 13.826530612244898,
"grad_norm": 0.3972738981246948,
"learning_rate": 4.02030162412993e-05,
"loss": 1.2986,
"step": 81300
},
{
"epoch": 13.835034013605442,
"grad_norm": 0.40687882900238037,
"learning_rate": 3.9912993039443156e-05,
"loss": 1.3106,
"step": 81350
},
{
"epoch": 13.843537414965986,
"grad_norm": 0.40394240617752075,
"learning_rate": 3.962296983758701e-05,
"loss": 1.309,
"step": 81400
},
{
"epoch": 13.85204081632653,
"grad_norm": 0.40160682797431946,
"learning_rate": 3.933294663573086e-05,
"loss": 1.3063,
"step": 81450
},
{
"epoch": 13.860544217687075,
"grad_norm": 0.401882529258728,
"learning_rate": 3.9042923433874705e-05,
"loss": 1.3085,
"step": 81500
},
{
"epoch": 13.869047619047619,
"grad_norm": 0.40280336141586304,
"learning_rate": 3.8758700696055686e-05,
"loss": 1.2966,
"step": 81550
},
{
"epoch": 13.877551020408163,
"grad_norm": 0.43757128715515137,
"learning_rate": 3.8468677494199534e-05,
"loss": 1.301,
"step": 81600
},
{
"epoch": 13.886054421768707,
"grad_norm": 0.40991008281707764,
"learning_rate": 3.817865429234339e-05,
"loss": 1.3037,
"step": 81650
},
{
"epoch": 13.89455782312925,
"grad_norm": 0.413152277469635,
"learning_rate": 3.788863109048724e-05,
"loss": 1.303,
"step": 81700
},
{
"epoch": 13.903061224489797,
"grad_norm": 0.40487709641456604,
"learning_rate": 3.759860788863109e-05,
"loss": 1.3071,
"step": 81750
},
{
"epoch": 13.91156462585034,
"grad_norm": 0.4063573181629181,
"learning_rate": 3.7308584686774944e-05,
"loss": 1.3037,
"step": 81800
},
{
"epoch": 13.920068027210885,
"grad_norm": 0.4249866306781769,
"learning_rate": 3.701856148491879e-05,
"loss": 1.3005,
"step": 81850
},
{
"epoch": 13.928571428571429,
"grad_norm": 0.41051676869392395,
"learning_rate": 3.6728538283062646e-05,
"loss": 1.3071,
"step": 81900
},
{
"epoch": 13.937074829931973,
"grad_norm": 0.39749205112457275,
"learning_rate": 3.64385150812065e-05,
"loss": 1.3049,
"step": 81950
},
{
"epoch": 13.945578231292517,
"grad_norm": 0.4107550382614136,
"learning_rate": 3.614849187935035e-05,
"loss": 1.3005,
"step": 82000
},
{
"epoch": 13.945578231292517,
"eval_loss": 1.4647165536880493,
"eval_runtime": 75.3191,
"eval_samples_per_second": 1245.886,
"eval_steps_per_second": 4.873,
"step": 82000
},
{
"epoch": 13.954081632653061,
"grad_norm": 0.4149150848388672,
"learning_rate": 3.5858468677494195e-05,
"loss": 1.3065,
"step": 82050
},
{
"epoch": 13.962585034013605,
"grad_norm": 0.4155055284500122,
"learning_rate": 3.5568445475638056e-05,
"loss": 1.2891,
"step": 82100
},
{
"epoch": 13.97108843537415,
"grad_norm": 0.40817248821258545,
"learning_rate": 3.5278422273781904e-05,
"loss": 1.303,
"step": 82150
},
{
"epoch": 13.979591836734693,
"grad_norm": 0.41482388973236084,
"learning_rate": 3.498839907192575e-05,
"loss": 1.3021,
"step": 82200
},
{
"epoch": 13.988095238095237,
"grad_norm": 0.41232186555862427,
"learning_rate": 3.4698375870069605e-05,
"loss": 1.2998,
"step": 82250
},
{
"epoch": 13.996598639455783,
"grad_norm": 0.40600040555000305,
"learning_rate": 3.440835266821346e-05,
"loss": 1.2983,
"step": 82300
},
{
"epoch": 14.005102040816327,
"grad_norm": 0.40963810682296753,
"learning_rate": 3.411832946635731e-05,
"loss": 1.2898,
"step": 82350
},
{
"epoch": 14.013605442176871,
"grad_norm": 0.41621676087379456,
"learning_rate": 3.382830626450116e-05,
"loss": 1.2984,
"step": 82400
},
{
"epoch": 14.022108843537415,
"grad_norm": 0.3934304118156433,
"learning_rate": 3.353828306264501e-05,
"loss": 1.2949,
"step": 82450
},
{
"epoch": 14.03061224489796,
"grad_norm": 0.4079570770263672,
"learning_rate": 3.324825986078887e-05,
"loss": 1.2966,
"step": 82500
},
{
"epoch": 14.039115646258503,
"grad_norm": 0.41960811614990234,
"learning_rate": 3.295823665893272e-05,
"loss": 1.3005,
"step": 82550
},
{
"epoch": 14.047619047619047,
"grad_norm": 0.40414050221443176,
"learning_rate": 3.2668213457076565e-05,
"loss": 1.299,
"step": 82600
},
{
"epoch": 14.056122448979592,
"grad_norm": 0.4004068970680237,
"learning_rate": 3.237819025522042e-05,
"loss": 1.2898,
"step": 82650
},
{
"epoch": 14.064625850340136,
"grad_norm": 0.41004568338394165,
"learning_rate": 3.2088167053364274e-05,
"loss": 1.2961,
"step": 82700
},
{
"epoch": 14.07312925170068,
"grad_norm": 0.4094617962837219,
"learning_rate": 3.179814385150812e-05,
"loss": 1.2972,
"step": 82750
},
{
"epoch": 14.081632653061224,
"grad_norm": 0.422753244638443,
"learning_rate": 3.1508120649651975e-05,
"loss": 1.3021,
"step": 82800
},
{
"epoch": 14.09013605442177,
"grad_norm": 0.41975682973861694,
"learning_rate": 3.121809744779582e-05,
"loss": 1.2996,
"step": 82850
},
{
"epoch": 14.098639455782314,
"grad_norm": 0.3985610008239746,
"learning_rate": 3.092807424593968e-05,
"loss": 1.3014,
"step": 82900
},
{
"epoch": 14.107142857142858,
"grad_norm": 0.4374600350856781,
"learning_rate": 3.063805104408353e-05,
"loss": 1.2963,
"step": 82950
},
{
"epoch": 14.115646258503402,
"grad_norm": 0.4096509516239166,
"learning_rate": 3.034802784222738e-05,
"loss": 1.2925,
"step": 83000
},
{
"epoch": 14.115646258503402,
"eval_loss": 1.4645365476608276,
"eval_runtime": 75.2852,
"eval_samples_per_second": 1246.447,
"eval_steps_per_second": 4.875,
"step": 83000
},
{
"epoch": 14.124149659863946,
"grad_norm": 0.6984773874282837,
"learning_rate": 3.0058004640371233e-05,
"loss": 1.2946,
"step": 83050
},
{
"epoch": 14.13265306122449,
"grad_norm": 0.4192226827144623,
"learning_rate": 2.976798143851508e-05,
"loss": 1.2906,
"step": 83100
},
{
"epoch": 14.141156462585034,
"grad_norm": 0.4226321280002594,
"learning_rate": 2.9477958236658935e-05,
"loss": 1.304,
"step": 83150
},
{
"epoch": 14.149659863945578,
"grad_norm": 0.4252030849456787,
"learning_rate": 2.9187935034802786e-05,
"loss": 1.2966,
"step": 83200
},
{
"epoch": 14.158163265306122,
"grad_norm": 0.3973104953765869,
"learning_rate": 2.8897911832946637e-05,
"loss": 1.2916,
"step": 83250
},
{
"epoch": 14.166666666666666,
"grad_norm": 0.42370930314064026,
"learning_rate": 2.8607888631090488e-05,
"loss": 1.2924,
"step": 83300
},
{
"epoch": 14.17517006802721,
"grad_norm": 1.85109543800354,
"learning_rate": 2.8317865429234342e-05,
"loss": 1.2927,
"step": 83350
},
{
"epoch": 14.183673469387756,
"grad_norm": 0.41108080744743347,
"learning_rate": 2.802784222737819e-05,
"loss": 1.2924,
"step": 83400
},
{
"epoch": 14.1921768707483,
"grad_norm": 0.45318904519081116,
"learning_rate": 2.7737819025522044e-05,
"loss": 1.2936,
"step": 83450
},
{
"epoch": 14.200680272108844,
"grad_norm": 0.402097225189209,
"learning_rate": 2.7447795823665895e-05,
"loss": 1.2966,
"step": 83500
},
{
"epoch": 14.209183673469388,
"grad_norm": 0.40355297923088074,
"learning_rate": 2.7157772621809746e-05,
"loss": 1.2926,
"step": 83550
},
{
"epoch": 14.217687074829932,
"grad_norm": 0.5472226738929749,
"learning_rate": 2.6873549883990723e-05,
"loss": 1.2938,
"step": 83600
},
{
"epoch": 14.226190476190476,
"grad_norm": 0.40354666113853455,
"learning_rate": 2.658352668213457e-05,
"loss": 1.2997,
"step": 83650
},
{
"epoch": 14.23469387755102,
"grad_norm": 0.399110347032547,
"learning_rate": 2.6293503480278425e-05,
"loss": 1.2916,
"step": 83700
},
{
"epoch": 14.243197278911564,
"grad_norm": 0.42456942796707153,
"learning_rate": 2.6003480278422276e-05,
"loss": 1.2987,
"step": 83750
},
{
"epoch": 14.251700680272108,
"grad_norm": 0.4085513949394226,
"learning_rate": 2.5713457076566123e-05,
"loss": 1.2922,
"step": 83800
},
{
"epoch": 14.260204081632653,
"grad_norm": 0.41955679655075073,
"learning_rate": 2.5423433874709977e-05,
"loss": 1.2942,
"step": 83850
},
{
"epoch": 14.268707482993197,
"grad_norm": 0.42755693197250366,
"learning_rate": 2.5133410672853828e-05,
"loss": 1.2927,
"step": 83900
},
{
"epoch": 14.27721088435374,
"grad_norm": 0.7680487036705017,
"learning_rate": 2.484338747099768e-05,
"loss": 1.2896,
"step": 83950
},
{
"epoch": 14.285714285714286,
"grad_norm": 0.41442444920539856,
"learning_rate": 2.455336426914153e-05,
"loss": 1.2993,
"step": 84000
},
{
"epoch": 14.285714285714286,
"eval_loss": 1.4609616994857788,
"eval_runtime": 75.2756,
"eval_samples_per_second": 1246.605,
"eval_steps_per_second": 4.875,
"step": 84000
},
{
"epoch": 14.29421768707483,
"grad_norm": 0.40402066707611084,
"learning_rate": 2.4263341067285384e-05,
"loss": 1.2965,
"step": 84050
},
{
"epoch": 14.302721088435375,
"grad_norm": 0.4172896146774292,
"learning_rate": 2.3973317865429232e-05,
"loss": 1.2912,
"step": 84100
},
{
"epoch": 14.311224489795919,
"grad_norm": 0.4188857078552246,
"learning_rate": 2.3683294663573086e-05,
"loss": 1.2941,
"step": 84150
},
{
"epoch": 14.319727891156463,
"grad_norm": 0.4177609384059906,
"learning_rate": 2.3393271461716937e-05,
"loss": 1.2956,
"step": 84200
},
{
"epoch": 14.328231292517007,
"grad_norm": 0.4155607223510742,
"learning_rate": 2.3103248259860788e-05,
"loss": 1.2909,
"step": 84250
},
{
"epoch": 14.33673469387755,
"grad_norm": 0.39458438754081726,
"learning_rate": 2.281322505800464e-05,
"loss": 1.2991,
"step": 84300
},
{
"epoch": 14.345238095238095,
"grad_norm": 0.3875535726547241,
"learning_rate": 2.2523201856148493e-05,
"loss": 1.2914,
"step": 84350
},
{
"epoch": 14.353741496598639,
"grad_norm": 0.39373454451560974,
"learning_rate": 2.223317865429234e-05,
"loss": 1.2921,
"step": 84400
},
{
"epoch": 14.362244897959183,
"grad_norm": 0.4142104685306549,
"learning_rate": 2.1943155452436195e-05,
"loss": 1.2924,
"step": 84450
},
{
"epoch": 14.370748299319727,
"grad_norm": 0.4059418737888336,
"learning_rate": 2.1653132250580046e-05,
"loss": 1.2927,
"step": 84500
},
{
"epoch": 14.379251700680273,
"grad_norm": 0.42010223865509033,
"learning_rate": 2.13631090487239e-05,
"loss": 1.2946,
"step": 84550
},
{
"epoch": 14.387755102040817,
"grad_norm": 0.4005703330039978,
"learning_rate": 2.1073085846867748e-05,
"loss": 1.2889,
"step": 84600
},
{
"epoch": 14.396258503401361,
"grad_norm": 0.434023916721344,
"learning_rate": 2.0783062645011602e-05,
"loss": 1.2925,
"step": 84650
},
{
"epoch": 14.404761904761905,
"grad_norm": 0.41903188824653625,
"learning_rate": 2.0493039443155453e-05,
"loss": 1.2982,
"step": 84700
},
{
"epoch": 14.41326530612245,
"grad_norm": 0.41402989625930786,
"learning_rate": 2.0203016241299304e-05,
"loss": 1.2972,
"step": 84750
},
{
"epoch": 14.421768707482993,
"grad_norm": 0.4167090952396393,
"learning_rate": 1.9912993039443155e-05,
"loss": 1.2975,
"step": 84800
},
{
"epoch": 14.430272108843537,
"grad_norm": 0.40198904275894165,
"learning_rate": 1.962296983758701e-05,
"loss": 1.2898,
"step": 84850
},
{
"epoch": 14.438775510204081,
"grad_norm": 0.412724107503891,
"learning_rate": 1.9332946635730856e-05,
"loss": 1.2955,
"step": 84900
},
{
"epoch": 14.447278911564625,
"grad_norm": 0.4073164165019989,
"learning_rate": 1.904292343387471e-05,
"loss": 1.2907,
"step": 84950
},
{
"epoch": 14.45578231292517,
"grad_norm": 0.4109826982021332,
"learning_rate": 1.875290023201856e-05,
"loss": 1.2935,
"step": 85000
},
{
"epoch": 14.45578231292517,
"eval_loss": 1.457992672920227,
"eval_runtime": 75.28,
"eval_samples_per_second": 1246.534,
"eval_steps_per_second": 4.875,
"step": 85000
},
{
"epoch": 14.464285714285714,
"grad_norm": 0.40559759736061096,
"learning_rate": 1.8462877030162412e-05,
"loss": 1.2888,
"step": 85050
},
{
"epoch": 14.47278911564626,
"grad_norm": 0.4154794216156006,
"learning_rate": 1.8172853828306263e-05,
"loss": 1.289,
"step": 85100
},
{
"epoch": 14.481292517006803,
"grad_norm": 0.4155610501766205,
"learning_rate": 1.7882830626450118e-05,
"loss": 1.2922,
"step": 85150
},
{
"epoch": 14.489795918367347,
"grad_norm": 0.4222477674484253,
"learning_rate": 1.7592807424593965e-05,
"loss": 1.2867,
"step": 85200
},
{
"epoch": 14.498299319727892,
"grad_norm": 0.4063816964626312,
"learning_rate": 1.730278422273782e-05,
"loss": 1.2888,
"step": 85250
},
{
"epoch": 14.506802721088436,
"grad_norm": 0.3986765742301941,
"learning_rate": 1.701276102088167e-05,
"loss": 1.285,
"step": 85300
},
{
"epoch": 14.51530612244898,
"grad_norm": 0.4038733243942261,
"learning_rate": 1.672273781902552e-05,
"loss": 1.2913,
"step": 85350
},
{
"epoch": 14.523809523809524,
"grad_norm": 0.40419521927833557,
"learning_rate": 1.6432714617169372e-05,
"loss": 1.2872,
"step": 85400
},
{
"epoch": 14.532312925170068,
"grad_norm": 0.41449958086013794,
"learning_rate": 1.6142691415313226e-05,
"loss": 1.2853,
"step": 85450
},
{
"epoch": 14.540816326530612,
"grad_norm": 0.4067400097846985,
"learning_rate": 1.5852668213457077e-05,
"loss": 1.2919,
"step": 85500
},
{
"epoch": 14.549319727891156,
"grad_norm": 0.6428622007369995,
"learning_rate": 1.5562645011600928e-05,
"loss": 1.2935,
"step": 85550
},
{
"epoch": 14.5578231292517,
"grad_norm": 0.4078386723995209,
"learning_rate": 1.527262180974478e-05,
"loss": 1.2892,
"step": 85600
},
{
"epoch": 14.566326530612244,
"grad_norm": 0.4054696261882782,
"learning_rate": 1.4982598607888632e-05,
"loss": 1.2871,
"step": 85650
},
{
"epoch": 14.57482993197279,
"grad_norm": 0.4058522880077362,
"learning_rate": 1.4692575406032483e-05,
"loss": 1.286,
"step": 85700
},
{
"epoch": 14.583333333333334,
"grad_norm": 0.40443381667137146,
"learning_rate": 1.4402552204176335e-05,
"loss": 1.2888,
"step": 85750
},
{
"epoch": 14.591836734693878,
"grad_norm": 0.4228183925151825,
"learning_rate": 1.4112529002320186e-05,
"loss": 1.2875,
"step": 85800
},
{
"epoch": 14.600340136054422,
"grad_norm": 0.4046574831008911,
"learning_rate": 1.3822505800464039e-05,
"loss": 1.2883,
"step": 85850
},
{
"epoch": 14.608843537414966,
"grad_norm": 0.4183114767074585,
"learning_rate": 1.353248259860789e-05,
"loss": 1.2929,
"step": 85900
},
{
"epoch": 14.61734693877551,
"grad_norm": 0.40642350912094116,
"learning_rate": 1.324245939675174e-05,
"loss": 1.2892,
"step": 85950
},
{
"epoch": 14.625850340136054,
"grad_norm": 0.40780848264694214,
"learning_rate": 1.2952436194895593e-05,
"loss": 1.2888,
"step": 86000
},
{
"epoch": 14.625850340136054,
"eval_loss": 1.4564687013626099,
"eval_runtime": 75.2583,
"eval_samples_per_second": 1246.893,
"eval_steps_per_second": 4.877,
"step": 86000
},
{
"epoch": 14.634353741496598,
"grad_norm": 0.399654746055603,
"learning_rate": 1.2662412993039444e-05,
"loss": 1.292,
"step": 86050
},
{
"epoch": 14.642857142857142,
"grad_norm": 0.4076179563999176,
"learning_rate": 1.2372389791183295e-05,
"loss": 1.2894,
"step": 86100
},
{
"epoch": 14.651360544217686,
"grad_norm": 0.4201255738735199,
"learning_rate": 1.2082366589327147e-05,
"loss": 1.289,
"step": 86150
},
{
"epoch": 14.65986394557823,
"grad_norm": 0.39747655391693115,
"learning_rate": 1.1792343387470998e-05,
"loss": 1.2915,
"step": 86200
},
{
"epoch": 14.668367346938776,
"grad_norm": 0.4104771316051483,
"learning_rate": 1.1508120649651974e-05,
"loss": 1.2864,
"step": 86250
},
{
"epoch": 14.67687074829932,
"grad_norm": 0.41255298256874084,
"learning_rate": 1.1218097447795825e-05,
"loss": 1.2916,
"step": 86300
},
{
"epoch": 14.685374149659864,
"grad_norm": 0.4036099910736084,
"learning_rate": 1.0928074245939676e-05,
"loss": 1.2921,
"step": 86350
},
{
"epoch": 14.693877551020408,
"grad_norm": 0.40650486946105957,
"learning_rate": 1.0638051044083526e-05,
"loss": 1.2896,
"step": 86400
},
{
"epoch": 14.702380952380953,
"grad_norm": 0.47484561800956726,
"learning_rate": 1.0348027842227377e-05,
"loss": 1.2917,
"step": 86450
},
{
"epoch": 14.710884353741497,
"grad_norm": 0.4110647737979889,
"learning_rate": 1.0058004640371228e-05,
"loss": 1.2878,
"step": 86500
},
{
"epoch": 14.71938775510204,
"grad_norm": 0.4205218553543091,
"learning_rate": 9.767981438515081e-06,
"loss": 1.2879,
"step": 86550
},
{
"epoch": 14.727891156462585,
"grad_norm": 0.4059794843196869,
"learning_rate": 9.477958236658932e-06,
"loss": 1.2871,
"step": 86600
},
{
"epoch": 14.736394557823129,
"grad_norm": 0.41226640343666077,
"learning_rate": 9.187935034802784e-06,
"loss": 1.2938,
"step": 86650
},
{
"epoch": 14.744897959183673,
"grad_norm": 0.41290587186813354,
"learning_rate": 8.897911832946635e-06,
"loss": 1.286,
"step": 86700
},
{
"epoch": 14.753401360544217,
"grad_norm": 0.4120457172393799,
"learning_rate": 8.607888631090486e-06,
"loss": 1.289,
"step": 86750
},
{
"epoch": 14.761904761904763,
"grad_norm": 0.4931107461452484,
"learning_rate": 8.317865429234339e-06,
"loss": 1.2896,
"step": 86800
},
{
"epoch": 14.770408163265307,
"grad_norm": 0.4076496660709381,
"learning_rate": 8.02784222737819e-06,
"loss": 1.2877,
"step": 86850
},
{
"epoch": 14.77891156462585,
"grad_norm": 0.41810768842697144,
"learning_rate": 7.737819025522042e-06,
"loss": 1.2873,
"step": 86900
},
{
"epoch": 14.787414965986395,
"grad_norm": 0.41356149315834045,
"learning_rate": 7.447795823665894e-06,
"loss": 1.2886,
"step": 86950
},
{
"epoch": 14.795918367346939,
"grad_norm": 0.43218472599983215,
"learning_rate": 7.157772621809745e-06,
"loss": 1.291,
"step": 87000
},
{
"epoch": 14.795918367346939,
"eval_loss": 1.4570873975753784,
"eval_runtime": 75.2933,
"eval_samples_per_second": 1246.313,
"eval_steps_per_second": 4.874,
"step": 87000
},
{
"epoch": 14.804421768707483,
"grad_norm": 0.4102339446544647,
"learning_rate": 6.867749419953597e-06,
"loss": 1.2848,
"step": 87050
},
{
"epoch": 14.812925170068027,
"grad_norm": 0.41269227862358093,
"learning_rate": 6.577726218097448e-06,
"loss": 1.2888,
"step": 87100
},
{
"epoch": 14.821428571428571,
"grad_norm": 0.4198719561100006,
"learning_rate": 6.287703016241299e-06,
"loss": 1.2904,
"step": 87150
},
{
"epoch": 14.829931972789115,
"grad_norm": 0.40524616837501526,
"learning_rate": 5.997679814385151e-06,
"loss": 1.2916,
"step": 87200
},
{
"epoch": 14.83843537414966,
"grad_norm": 0.40271249413490295,
"learning_rate": 5.707656612529003e-06,
"loss": 1.2821,
"step": 87250
},
{
"epoch": 14.846938775510203,
"grad_norm": 0.404710590839386,
"learning_rate": 5.4176334106728545e-06,
"loss": 1.2848,
"step": 87300
},
{
"epoch": 14.85544217687075,
"grad_norm": 0.40877318382263184,
"learning_rate": 5.127610208816705e-06,
"loss": 1.2824,
"step": 87350
},
{
"epoch": 14.863945578231293,
"grad_norm": 0.40743860602378845,
"learning_rate": 4.837587006960557e-06,
"loss": 1.2875,
"step": 87400
},
{
"epoch": 14.872448979591837,
"grad_norm": 0.40123429894447327,
"learning_rate": 4.547563805104409e-06,
"loss": 1.2879,
"step": 87450
},
{
"epoch": 14.880952380952381,
"grad_norm": 0.4120638072490692,
"learning_rate": 4.257540603248261e-06,
"loss": 1.2886,
"step": 87500
},
{
"epoch": 14.889455782312925,
"grad_norm": 0.4441770017147064,
"learning_rate": 3.9675174013921115e-06,
"loss": 1.2903,
"step": 87550
},
{
"epoch": 14.89795918367347,
"grad_norm": 0.39684155583381653,
"learning_rate": 3.677494199535963e-06,
"loss": 1.2894,
"step": 87600
},
{
"epoch": 14.906462585034014,
"grad_norm": 0.40048104524612427,
"learning_rate": 3.387470997679814e-06,
"loss": 1.2872,
"step": 87650
},
{
"epoch": 14.914965986394558,
"grad_norm": 0.40761691331863403,
"learning_rate": 3.097447795823666e-06,
"loss": 1.2856,
"step": 87700
},
{
"epoch": 14.923469387755102,
"grad_norm": 0.41914671659469604,
"learning_rate": 2.807424593967517e-06,
"loss": 1.2864,
"step": 87750
},
{
"epoch": 14.931972789115646,
"grad_norm": 0.41371914744377136,
"learning_rate": 2.517401392111369e-06,
"loss": 1.2807,
"step": 87800
},
{
"epoch": 14.94047619047619,
"grad_norm": 0.4007938802242279,
"learning_rate": 2.2273781902552203e-06,
"loss": 1.2869,
"step": 87850
},
{
"epoch": 14.948979591836736,
"grad_norm": 0.4132489264011383,
"learning_rate": 1.937354988399072e-06,
"loss": 1.2822,
"step": 87900
},
{
"epoch": 14.95748299319728,
"grad_norm": 0.4170679748058319,
"learning_rate": 1.6473317865429233e-06,
"loss": 1.2867,
"step": 87950
},
{
"epoch": 14.965986394557824,
"grad_norm": 0.4113543629646301,
"learning_rate": 1.3573085846867749e-06,
"loss": 1.2852,
"step": 88000
},
{
"epoch": 14.965986394557824,
"eval_loss": 1.4495242834091187,
"eval_runtime": 75.6807,
"eval_samples_per_second": 1239.933,
"eval_steps_per_second": 4.849,
"step": 88000
},
{
"epoch": 14.974489795918368,
"grad_norm": 0.4077957272529602,
"learning_rate": 1.0672853828306264e-06,
"loss": 1.2852,
"step": 88050
},
{
"epoch": 14.982993197278912,
"grad_norm": 0.403070330619812,
"learning_rate": 7.772621809744779e-07,
"loss": 1.2854,
"step": 88100
},
{
"epoch": 14.991496598639456,
"grad_norm": 0.41275766491889954,
"learning_rate": 4.872389791183295e-07,
"loss": 1.288,
"step": 88150
},
{
"epoch": 15.0,
"grad_norm": 0.4174801707267761,
"learning_rate": 1.97215777262181e-07,
"loss": 1.2938,
"step": 88200
}
],
"logging_steps": 50,
"max_steps": 88200,
"num_input_tokens_seen": 0,
"num_train_epochs": 15,
"save_steps": 1000,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 2.377209771467368e+19,
"train_batch_size": 32,
"trial_name": null,
"trial_params": null
}