9b-31 / trainer_state.json
furproxy's picture
Upload folder using huggingface_hub
da081e5 verified
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 3.0,
"eval_steps": 500,
"global_step": 1029,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.00583941605839416,
"grad_norm": 1.9921875,
"learning_rate": 1.9230769230769234e-07,
"loss": 2.052279233932495,
"step": 2
},
{
"epoch": 0.01167883211678832,
"grad_norm": 0.71875,
"learning_rate": 5.76923076923077e-07,
"loss": 1.9450315237045288,
"step": 4
},
{
"epoch": 0.017518248175182483,
"grad_norm": 0.97265625,
"learning_rate": 9.615384615384617e-07,
"loss": 2.0441408157348633,
"step": 6
},
{
"epoch": 0.02335766423357664,
"grad_norm": 1.390625,
"learning_rate": 1.3461538461538462e-06,
"loss": 1.9473555088043213,
"step": 8
},
{
"epoch": 0.029197080291970802,
"grad_norm": 0.6875,
"learning_rate": 1.7307692307692308e-06,
"loss": 1.9234904050827026,
"step": 10
},
{
"epoch": 0.035036496350364967,
"grad_norm": 1.234375,
"learning_rate": 2.1153846153846155e-06,
"loss": 2.0447089672088623,
"step": 12
},
{
"epoch": 0.040875912408759124,
"grad_norm": 1.171875,
"learning_rate": 2.5e-06,
"loss": 1.8234096765518188,
"step": 14
},
{
"epoch": 0.04671532846715328,
"grad_norm": 0.51953125,
"learning_rate": 2.8846153846153845e-06,
"loss": 1.8641804456710815,
"step": 16
},
{
"epoch": 0.052554744525547446,
"grad_norm": 0.6640625,
"learning_rate": 3.2692307692307696e-06,
"loss": 1.9037718772888184,
"step": 18
},
{
"epoch": 0.058394160583941604,
"grad_norm": 0.7734375,
"learning_rate": 3.653846153846154e-06,
"loss": 1.9216861724853516,
"step": 20
},
{
"epoch": 0.06423357664233577,
"grad_norm": 1.3125,
"learning_rate": 4.0384615384615385e-06,
"loss": 1.9886118173599243,
"step": 22
},
{
"epoch": 0.07007299270072993,
"grad_norm": 1.015625,
"learning_rate": 4.423076923076924e-06,
"loss": 1.7973345518112183,
"step": 24
},
{
"epoch": 0.07591240875912408,
"grad_norm": 0.53515625,
"learning_rate": 4.807692307692308e-06,
"loss": 1.8787648677825928,
"step": 26
},
{
"epoch": 0.08175182481751825,
"grad_norm": 0.765625,
"learning_rate": 5.192307692307693e-06,
"loss": 1.8880407810211182,
"step": 28
},
{
"epoch": 0.08759124087591241,
"grad_norm": 0.5,
"learning_rate": 5.576923076923077e-06,
"loss": 1.8372446298599243,
"step": 30
},
{
"epoch": 0.09343065693430656,
"grad_norm": 0.65625,
"learning_rate": 5.961538461538462e-06,
"loss": 1.8118259906768799,
"step": 32
},
{
"epoch": 0.09927007299270073,
"grad_norm": 1.09375,
"learning_rate": 6.3461538461538466e-06,
"loss": 1.872291922569275,
"step": 34
},
{
"epoch": 0.10510948905109489,
"grad_norm": 1.125,
"learning_rate": 6.730769230769232e-06,
"loss": 1.9201974868774414,
"step": 36
},
{
"epoch": 0.11094890510948906,
"grad_norm": 0.6875,
"learning_rate": 7.115384615384616e-06,
"loss": 1.821286678314209,
"step": 38
},
{
"epoch": 0.11678832116788321,
"grad_norm": 0.361328125,
"learning_rate": 7.500000000000001e-06,
"loss": 1.6810513734817505,
"step": 40
},
{
"epoch": 0.12262773722627737,
"grad_norm": 0.50390625,
"learning_rate": 7.884615384615384e-06,
"loss": 1.7490516901016235,
"step": 42
},
{
"epoch": 0.12846715328467154,
"grad_norm": 0.400390625,
"learning_rate": 8.26923076923077e-06,
"loss": 1.6871154308319092,
"step": 44
},
{
"epoch": 0.1343065693430657,
"grad_norm": 0.9765625,
"learning_rate": 8.653846153846155e-06,
"loss": 1.8664501905441284,
"step": 46
},
{
"epoch": 0.14014598540145987,
"grad_norm": 0.57421875,
"learning_rate": 9.03846153846154e-06,
"loss": 1.76151442527771,
"step": 48
},
{
"epoch": 0.145985401459854,
"grad_norm": 0.6953125,
"learning_rate": 9.423076923076923e-06,
"loss": 1.7157841920852661,
"step": 50
},
{
"epoch": 0.15182481751824817,
"grad_norm": 0.474609375,
"learning_rate": 9.807692307692308e-06,
"loss": 1.7865409851074219,
"step": 52
},
{
"epoch": 0.15766423357664233,
"grad_norm": 0.43359375,
"learning_rate": 9.999976735551495e-06,
"loss": 1.7557110786437988,
"step": 54
},
{
"epoch": 0.1635036496350365,
"grad_norm": 0.359375,
"learning_rate": 9.999790621406734e-06,
"loss": 1.6408637762069702,
"step": 56
},
{
"epoch": 0.16934306569343066,
"grad_norm": 0.62890625,
"learning_rate": 9.999418400814693e-06,
"loss": 1.7107805013656616,
"step": 58
},
{
"epoch": 0.17518248175182483,
"grad_norm": 11.9375,
"learning_rate": 9.998860089170008e-06,
"loss": 1.6930475234985352,
"step": 60
},
{
"epoch": 0.181021897810219,
"grad_norm": 0.90625,
"learning_rate": 9.998115709563843e-06,
"loss": 1.8085753917694092,
"step": 62
},
{
"epoch": 0.18686131386861313,
"grad_norm": 0.54296875,
"learning_rate": 9.997185292782932e-06,
"loss": 1.6834930181503296,
"step": 64
},
{
"epoch": 0.1927007299270073,
"grad_norm": 0.80859375,
"learning_rate": 9.996068877308297e-06,
"loss": 1.5944346189498901,
"step": 66
},
{
"epoch": 0.19854014598540146,
"grad_norm": 0.80859375,
"learning_rate": 9.994766509313667e-06,
"loss": 1.7082901000976562,
"step": 68
},
{
"epoch": 0.20437956204379562,
"grad_norm": 0.3671875,
"learning_rate": 9.993278242663572e-06,
"loss": 1.6360015869140625,
"step": 70
},
{
"epoch": 0.21021897810218979,
"grad_norm": 0.419921875,
"learning_rate": 9.9916041389111e-06,
"loss": 1.7114381790161133,
"step": 72
},
{
"epoch": 0.21605839416058395,
"grad_norm": 0.4296875,
"learning_rate": 9.989744267295359e-06,
"loss": 1.5103245973587036,
"step": 74
},
{
"epoch": 0.22189781021897811,
"grad_norm": 0.46875,
"learning_rate": 9.987698704738617e-06,
"loss": 1.4727965593338013,
"step": 76
},
{
"epoch": 0.22773722627737225,
"grad_norm": 0.71875,
"learning_rate": 9.985467535843121e-06,
"loss": 1.6391849517822266,
"step": 78
},
{
"epoch": 0.23357664233576642,
"grad_norm": 3.625,
"learning_rate": 9.983050852887587e-06,
"loss": 1.7767753601074219,
"step": 80
},
{
"epoch": 0.23941605839416058,
"grad_norm": 0.57421875,
"learning_rate": 9.9804487558234e-06,
"loss": 1.6842796802520752,
"step": 82
},
{
"epoch": 0.24525547445255474,
"grad_norm": 0.8515625,
"learning_rate": 9.977661352270457e-06,
"loss": 1.6517852544784546,
"step": 84
},
{
"epoch": 0.2510948905109489,
"grad_norm": 0.72265625,
"learning_rate": 9.974688757512742e-06,
"loss": 1.501247763633728,
"step": 86
},
{
"epoch": 0.2569343065693431,
"grad_norm": 0.462890625,
"learning_rate": 9.971531094493541e-06,
"loss": 1.5306718349456787,
"step": 88
},
{
"epoch": 0.26277372262773724,
"grad_norm": 0.3671875,
"learning_rate": 9.968188493810365e-06,
"loss": 1.6086313724517822,
"step": 90
},
{
"epoch": 0.2686131386861314,
"grad_norm": 0.59375,
"learning_rate": 9.964661093709545e-06,
"loss": 1.5198537111282349,
"step": 92
},
{
"epoch": 0.27445255474452557,
"grad_norm": 0.9375,
"learning_rate": 9.960949040080507e-06,
"loss": 1.5648375749588013,
"step": 94
},
{
"epoch": 0.28029197080291973,
"grad_norm": 0.5390625,
"learning_rate": 9.95705248644976e-06,
"loss": 1.6112207174301147,
"step": 96
},
{
"epoch": 0.28613138686131384,
"grad_norm": 0.38671875,
"learning_rate": 9.95297159397452e-06,
"loss": 1.575645923614502,
"step": 98
},
{
"epoch": 0.291970802919708,
"grad_norm": 0.455078125,
"learning_rate": 9.948706531436067e-06,
"loss": 1.4316264390945435,
"step": 100
},
{
"epoch": 0.29781021897810217,
"grad_norm": 0.365234375,
"learning_rate": 9.944257475232746e-06,
"loss": 1.4882372617721558,
"step": 102
},
{
"epoch": 0.30364963503649633,
"grad_norm": 0.392578125,
"learning_rate": 9.939624609372686e-06,
"loss": 1.4284008741378784,
"step": 104
},
{
"epoch": 0.3094890510948905,
"grad_norm": 0.43359375,
"learning_rate": 9.934808125466181e-06,
"loss": 1.468398928642273,
"step": 106
},
{
"epoch": 0.31532846715328466,
"grad_norm": 0.37109375,
"learning_rate": 9.929808222717769e-06,
"loss": 1.465457797050476,
"step": 108
},
{
"epoch": 0.32116788321167883,
"grad_norm": 0.359375,
"learning_rate": 9.924625107917991e-06,
"loss": 1.425096035003662,
"step": 110
},
{
"epoch": 0.327007299270073,
"grad_norm": 0.337890625,
"learning_rate": 9.919258995434838e-06,
"loss": 1.4254180192947388,
"step": 112
},
{
"epoch": 0.33284671532846716,
"grad_norm": 0.349609375,
"learning_rate": 9.913710107204888e-06,
"loss": 1.4670501947402954,
"step": 114
},
{
"epoch": 0.3386861313868613,
"grad_norm": 0.41796875,
"learning_rate": 9.907978672724125e-06,
"loss": 1.549682378768921,
"step": 116
},
{
"epoch": 0.3445255474452555,
"grad_norm": 0.322265625,
"learning_rate": 9.902064929038447e-06,
"loss": 1.448873519897461,
"step": 118
},
{
"epoch": 0.35036496350364965,
"grad_norm": 0.40625,
"learning_rate": 9.89596912073386e-06,
"loss": 1.4342598915100098,
"step": 120
},
{
"epoch": 0.3562043795620438,
"grad_norm": 0.396484375,
"learning_rate": 9.889691499926372e-06,
"loss": 1.424146294593811,
"step": 122
},
{
"epoch": 0.362043795620438,
"grad_norm": 0.337890625,
"learning_rate": 9.883232326251553e-06,
"loss": 1.3757034540176392,
"step": 124
},
{
"epoch": 0.3678832116788321,
"grad_norm": 0.7421875,
"learning_rate": 9.876591866853798e-06,
"loss": 1.5297839641571045,
"step": 126
},
{
"epoch": 0.37372262773722625,
"grad_norm": 0.443359375,
"learning_rate": 9.869770396375291e-06,
"loss": 1.4363529682159424,
"step": 128
},
{
"epoch": 0.3795620437956204,
"grad_norm": 0.515625,
"learning_rate": 9.862768196944636e-06,
"loss": 1.5066875219345093,
"step": 130
},
{
"epoch": 0.3854014598540146,
"grad_norm": 0.609375,
"learning_rate": 9.855585558165181e-06,
"loss": 1.4090046882629395,
"step": 132
},
{
"epoch": 0.39124087591240875,
"grad_norm": 0.46875,
"learning_rate": 9.848222777103063e-06,
"loss": 1.5839899778366089,
"step": 134
},
{
"epoch": 0.3970802919708029,
"grad_norm": 0.7421875,
"learning_rate": 9.840680158274895e-06,
"loss": 1.425782561302185,
"step": 136
},
{
"epoch": 0.4029197080291971,
"grad_norm": 0.376953125,
"learning_rate": 9.832958013635195e-06,
"loss": 1.3894191980361938,
"step": 138
},
{
"epoch": 0.40875912408759124,
"grad_norm": 0.322265625,
"learning_rate": 9.825056662563464e-06,
"loss": 1.5015205144882202,
"step": 140
},
{
"epoch": 0.4145985401459854,
"grad_norm": 1.234375,
"learning_rate": 9.816976431850994e-06,
"loss": 1.3953148126602173,
"step": 142
},
{
"epoch": 0.42043795620437957,
"grad_norm": 0.66015625,
"learning_rate": 9.808717655687344e-06,
"loss": 1.3676221370697021,
"step": 144
},
{
"epoch": 0.42627737226277373,
"grad_norm": 1.046875,
"learning_rate": 9.800280675646508e-06,
"loss": 1.4441938400268555,
"step": 146
},
{
"epoch": 0.4321167883211679,
"grad_norm": 0.609375,
"learning_rate": 9.791665840672813e-06,
"loss": 1.3372529745101929,
"step": 148
},
{
"epoch": 0.43795620437956206,
"grad_norm": 0.4765625,
"learning_rate": 9.782873507066466e-06,
"loss": 1.3514975309371948,
"step": 150
},
{
"epoch": 0.44379562043795623,
"grad_norm": 0.515625,
"learning_rate": 9.773904038468822e-06,
"loss": 1.3599399328231812,
"step": 152
},
{
"epoch": 0.44963503649635034,
"grad_norm": 0.5546875,
"learning_rate": 9.764757805847352e-06,
"loss": 1.4376461505889893,
"step": 154
},
{
"epoch": 0.4554744525547445,
"grad_norm": 0.443359375,
"learning_rate": 9.755435187480288e-06,
"loss": 1.4733713865280151,
"step": 156
},
{
"epoch": 0.46131386861313867,
"grad_norm": 0.51171875,
"learning_rate": 9.745936568940992e-06,
"loss": 1.4277862310409546,
"step": 158
},
{
"epoch": 0.46715328467153283,
"grad_norm": 0.8046875,
"learning_rate": 9.736262343081998e-06,
"loss": 1.4021371603012085,
"step": 160
},
{
"epoch": 0.472992700729927,
"grad_norm": 0.63671875,
"learning_rate": 9.726412910018768e-06,
"loss": 1.4141305685043335,
"step": 162
},
{
"epoch": 0.47883211678832116,
"grad_norm": 0.53125,
"learning_rate": 9.716388677113145e-06,
"loss": 1.3931472301483154,
"step": 164
},
{
"epoch": 0.4846715328467153,
"grad_norm": 0.38671875,
"learning_rate": 9.7061900589565e-06,
"loss": 1.3798588514328003,
"step": 166
},
{
"epoch": 0.4905109489051095,
"grad_norm": 0.34375,
"learning_rate": 9.695817477352593e-06,
"loss": 1.3861608505249023,
"step": 168
},
{
"epoch": 0.49635036496350365,
"grad_norm": 0.4453125,
"learning_rate": 9.685271361300117e-06,
"loss": 1.4044109582901,
"step": 170
},
{
"epoch": 0.5021897810218978,
"grad_norm": 0.345703125,
"learning_rate": 9.67455214697497e-06,
"loss": 1.392524003982544,
"step": 172
},
{
"epoch": 0.5080291970802919,
"grad_norm": 0.73046875,
"learning_rate": 9.663660277712194e-06,
"loss": 1.4445301294326782,
"step": 174
},
{
"epoch": 0.5138686131386861,
"grad_norm": 0.4453125,
"learning_rate": 9.652596203987667e-06,
"loss": 1.47966468334198,
"step": 176
},
{
"epoch": 0.5197080291970803,
"grad_norm": 0.431640625,
"learning_rate": 9.64136038339944e-06,
"loss": 1.392290711402893,
"step": 178
},
{
"epoch": 0.5255474452554745,
"grad_norm": 0.390625,
"learning_rate": 9.62995328064884e-06,
"loss": 1.5381602048873901,
"step": 180
},
{
"epoch": 0.5313868613138686,
"grad_norm": 0.36328125,
"learning_rate": 9.618375367521232e-06,
"loss": 1.3802764415740967,
"step": 182
},
{
"epoch": 0.5372262773722628,
"grad_norm": 0.453125,
"learning_rate": 9.606627122866513e-06,
"loss": 1.4806022644042969,
"step": 184
},
{
"epoch": 0.5430656934306569,
"grad_norm": 0.546875,
"learning_rate": 9.594709032579307e-06,
"loss": 1.4255928993225098,
"step": 186
},
{
"epoch": 0.5489051094890511,
"grad_norm": 0.73828125,
"learning_rate": 9.582621589578873e-06,
"loss": 1.336686372756958,
"step": 188
},
{
"epoch": 0.5547445255474452,
"grad_norm": 0.51171875,
"learning_rate": 9.570365293788707e-06,
"loss": 1.4040147066116333,
"step": 190
},
{
"epoch": 0.5605839416058395,
"grad_norm": 0.337890625,
"learning_rate": 9.557940652115875e-06,
"loss": 1.416895866394043,
"step": 192
},
{
"epoch": 0.5664233576642336,
"grad_norm": 0.455078125,
"learning_rate": 9.545348178430051e-06,
"loss": 1.3130196332931519,
"step": 194
},
{
"epoch": 0.5722627737226277,
"grad_norm": 0.3671875,
"learning_rate": 9.53258839354225e-06,
"loss": 1.3405402898788452,
"step": 196
},
{
"epoch": 0.5781021897810219,
"grad_norm": 0.337890625,
"learning_rate": 9.519661825183303e-06,
"loss": 1.4502712488174438,
"step": 198
},
{
"epoch": 0.583941605839416,
"grad_norm": 0.9765625,
"learning_rate": 9.506569007982023e-06,
"loss": 1.4316283464431763,
"step": 200
},
{
"epoch": 0.5897810218978102,
"grad_norm": 1.0859375,
"learning_rate": 9.493310483443089e-06,
"loss": 1.4556365013122559,
"step": 202
},
{
"epoch": 0.5956204379562043,
"grad_norm": 0.376953125,
"learning_rate": 9.479886799924663e-06,
"loss": 1.413832426071167,
"step": 204
},
{
"epoch": 0.6014598540145986,
"grad_norm": 0.56640625,
"learning_rate": 9.466298512615697e-06,
"loss": 1.252786636352539,
"step": 206
},
{
"epoch": 0.6072992700729927,
"grad_norm": 0.4765625,
"learning_rate": 9.45254618351298e-06,
"loss": 1.369625210762024,
"step": 208
},
{
"epoch": 0.6131386861313869,
"grad_norm": 1.40625,
"learning_rate": 9.43863038139789e-06,
"loss": 1.3296895027160645,
"step": 210
},
{
"epoch": 0.618978102189781,
"grad_norm": 0.388671875,
"learning_rate": 9.424551681812871e-06,
"loss": 1.4407665729522705,
"step": 212
},
{
"epoch": 0.6248175182481752,
"grad_norm": 0.400390625,
"learning_rate": 9.41031066703763e-06,
"loss": 1.363607406616211,
"step": 214
},
{
"epoch": 0.6306569343065693,
"grad_norm": 0.6953125,
"learning_rate": 9.395907926065052e-06,
"loss": 1.4068963527679443,
"step": 216
},
{
"epoch": 0.6364963503649635,
"grad_norm": 0.484375,
"learning_rate": 9.381344054576845e-06,
"loss": 1.4279429912567139,
"step": 218
},
{
"epoch": 0.6423357664233577,
"grad_norm": 0.4375,
"learning_rate": 9.366619654918895e-06,
"loss": 1.4228907823562622,
"step": 220
},
{
"epoch": 0.6481751824817519,
"grad_norm": 0.330078125,
"learning_rate": 9.351735336076363e-06,
"loss": 1.4166944026947021,
"step": 222
},
{
"epoch": 0.654014598540146,
"grad_norm": 0.6171875,
"learning_rate": 9.33669171364849e-06,
"loss": 1.438708782196045,
"step": 224
},
{
"epoch": 0.6598540145985401,
"grad_norm": 0.59375,
"learning_rate": 9.32148940982315e-06,
"loss": 1.3473215103149414,
"step": 226
},
{
"epoch": 0.6656934306569343,
"grad_norm": 0.326171875,
"learning_rate": 9.306129053351087e-06,
"loss": 1.3401542901992798,
"step": 228
},
{
"epoch": 0.6715328467153284,
"grad_norm": 0.59765625,
"learning_rate": 9.290611279519952e-06,
"loss": 1.4278738498687744,
"step": 230
},
{
"epoch": 0.6773722627737226,
"grad_norm": 0.47265625,
"learning_rate": 9.274936730127991e-06,
"loss": 1.4194695949554443,
"step": 232
},
{
"epoch": 0.6832116788321168,
"grad_norm": 0.59375,
"learning_rate": 9.259106053457529e-06,
"loss": 1.4373456239700317,
"step": 234
},
{
"epoch": 0.689051094890511,
"grad_norm": 0.51171875,
"learning_rate": 9.243119904248136e-06,
"loss": 1.3994792699813843,
"step": 236
},
{
"epoch": 0.6948905109489051,
"grad_norm": 0.36328125,
"learning_rate": 9.226978943669562e-06,
"loss": 1.3777506351470947,
"step": 238
},
{
"epoch": 0.7007299270072993,
"grad_norm": 0.37890625,
"learning_rate": 9.210683839294386e-06,
"loss": 1.3514622449874878,
"step": 240
},
{
"epoch": 0.7065693430656934,
"grad_norm": 0.5390625,
"learning_rate": 9.194235265070406e-06,
"loss": 1.4253668785095215,
"step": 242
},
{
"epoch": 0.7124087591240876,
"grad_norm": 0.470703125,
"learning_rate": 9.177633901292767e-06,
"loss": 1.3638951778411865,
"step": 244
},
{
"epoch": 0.7182481751824817,
"grad_norm": 0.310546875,
"learning_rate": 9.160880434575823e-06,
"loss": 1.3718801736831665,
"step": 246
},
{
"epoch": 0.724087591240876,
"grad_norm": 0.46484375,
"learning_rate": 9.143975557824743e-06,
"loss": 1.3095505237579346,
"step": 248
},
{
"epoch": 0.7299270072992701,
"grad_norm": 0.56640625,
"learning_rate": 9.126919970206843e-06,
"loss": 1.3069571256637573,
"step": 250
},
{
"epoch": 0.7357664233576642,
"grad_norm": 0.27734375,
"learning_rate": 9.109714377122685e-06,
"loss": 1.341015100479126,
"step": 252
},
{
"epoch": 0.7416058394160584,
"grad_norm": 0.5859375,
"learning_rate": 9.092359490176892e-06,
"loss": 1.3410066366195679,
"step": 254
},
{
"epoch": 0.7474452554744525,
"grad_norm": 0.396484375,
"learning_rate": 9.074856027148715e-06,
"loss": 1.265282154083252,
"step": 256
},
{
"epoch": 0.7532846715328467,
"grad_norm": 0.349609375,
"learning_rate": 9.057204711962352e-06,
"loss": 1.2000367641448975,
"step": 258
},
{
"epoch": 0.7591240875912408,
"grad_norm": 0.50390625,
"learning_rate": 9.03940627465701e-06,
"loss": 1.4280476570129395,
"step": 260
},
{
"epoch": 0.7649635036496351,
"grad_norm": 0.51171875,
"learning_rate": 9.021461451356695e-06,
"loss": 1.3344112634658813,
"step": 262
},
{
"epoch": 0.7708029197080292,
"grad_norm": 0.859375,
"learning_rate": 9.003370984239795e-06,
"loss": 1.4776599407196045,
"step": 264
},
{
"epoch": 0.7766423357664234,
"grad_norm": 0.6328125,
"learning_rate": 8.985135621508355e-06,
"loss": 1.4331424236297607,
"step": 266
},
{
"epoch": 0.7824817518248175,
"grad_norm": 0.314453125,
"learning_rate": 8.966756117357152e-06,
"loss": 1.3084166049957275,
"step": 268
},
{
"epoch": 0.7883211678832117,
"grad_norm": 2.640625,
"learning_rate": 8.948233231942489e-06,
"loss": 1.3645600080490112,
"step": 270
},
{
"epoch": 0.7941605839416058,
"grad_norm": 0.703125,
"learning_rate": 8.929567731350767e-06,
"loss": 1.3116084337234497,
"step": 272
},
{
"epoch": 0.8,
"grad_norm": 0.65625,
"learning_rate": 8.9107603875668e-06,
"loss": 1.363153338432312,
"step": 274
},
{
"epoch": 0.8058394160583942,
"grad_norm": 0.4140625,
"learning_rate": 8.891811978441871e-06,
"loss": 1.3285009860992432,
"step": 276
},
{
"epoch": 0.8116788321167884,
"grad_norm": 0.40234375,
"learning_rate": 8.87272328766158e-06,
"loss": 1.4433071613311768,
"step": 278
},
{
"epoch": 0.8175182481751825,
"grad_norm": 1.09375,
"learning_rate": 8.85349510471342e-06,
"loss": 1.4583128690719604,
"step": 280
},
{
"epoch": 0.8233576642335766,
"grad_norm": 0.298828125,
"learning_rate": 8.834128224854133e-06,
"loss": 1.3218973875045776,
"step": 282
},
{
"epoch": 0.8291970802919708,
"grad_norm": 0.97265625,
"learning_rate": 8.814623449076809e-06,
"loss": 1.291555404663086,
"step": 284
},
{
"epoch": 0.8350364963503649,
"grad_norm": 0.337890625,
"learning_rate": 8.794981584077767e-06,
"loss": 1.3497284650802612,
"step": 286
},
{
"epoch": 0.8408759124087591,
"grad_norm": 0.322265625,
"learning_rate": 8.775203442223186e-06,
"loss": 1.2834782600402832,
"step": 288
},
{
"epoch": 0.8467153284671532,
"grad_norm": 0.35546875,
"learning_rate": 8.755289841515507e-06,
"loss": 1.3291265964508057,
"step": 290
},
{
"epoch": 0.8525547445255475,
"grad_norm": 0.373046875,
"learning_rate": 8.735241605559603e-06,
"loss": 1.3475388288497925,
"step": 292
},
{
"epoch": 0.8583941605839416,
"grad_norm": 0.349609375,
"learning_rate": 8.715059563528713e-06,
"loss": 1.37130606174469,
"step": 294
},
{
"epoch": 0.8642335766423358,
"grad_norm": 0.443359375,
"learning_rate": 8.694744550130153e-06,
"loss": 1.4104875326156616,
"step": 296
},
{
"epoch": 0.8700729927007299,
"grad_norm": 1.1875,
"learning_rate": 8.674297405570782e-06,
"loss": 1.3223466873168945,
"step": 298
},
{
"epoch": 0.8759124087591241,
"grad_norm": 0.52734375,
"learning_rate": 8.653718975522271e-06,
"loss": 1.3034943342208862,
"step": 300
},
{
"epoch": 0.8817518248175182,
"grad_norm": 0.46875,
"learning_rate": 8.63301011108611e-06,
"loss": 1.4011861085891724,
"step": 302
},
{
"epoch": 0.8875912408759125,
"grad_norm": 0.7578125,
"learning_rate": 8.612171668758413e-06,
"loss": 1.4873169660568237,
"step": 304
},
{
"epoch": 0.8934306569343066,
"grad_norm": 0.4296875,
"learning_rate": 8.591204510394498e-06,
"loss": 1.2987056970596313,
"step": 306
},
{
"epoch": 0.8992700729927007,
"grad_norm": 0.61328125,
"learning_rate": 8.570109503173233e-06,
"loss": 1.3402009010314941,
"step": 308
},
{
"epoch": 0.9051094890510949,
"grad_norm": 0.5859375,
"learning_rate": 8.548887519561184e-06,
"loss": 1.2947343587875366,
"step": 310
},
{
"epoch": 0.910948905109489,
"grad_norm": 0.89453125,
"learning_rate": 8.527539437276515e-06,
"loss": 1.3383480310440063,
"step": 312
},
{
"epoch": 0.9167883211678832,
"grad_norm": 0.72265625,
"learning_rate": 8.506066139252696e-06,
"loss": 1.3967725038528442,
"step": 314
},
{
"epoch": 0.9226277372262773,
"grad_norm": 0.3828125,
"learning_rate": 8.484468513601987e-06,
"loss": 1.2097219228744507,
"step": 316
},
{
"epoch": 0.9284671532846716,
"grad_norm": 1.546875,
"learning_rate": 8.462747453578698e-06,
"loss": 1.3612643480300903,
"step": 318
},
{
"epoch": 0.9343065693430657,
"grad_norm": 0.578125,
"learning_rate": 8.440903857542253e-06,
"loss": 1.3607382774353027,
"step": 320
},
{
"epoch": 0.9401459854014599,
"grad_norm": 0.384765625,
"learning_rate": 8.418938628920038e-06,
"loss": 1.276464819908142,
"step": 322
},
{
"epoch": 0.945985401459854,
"grad_norm": 0.5390625,
"learning_rate": 8.396852676170018e-06,
"loss": 1.3733280897140503,
"step": 324
},
{
"epoch": 0.9518248175182482,
"grad_norm": 0.6328125,
"learning_rate": 8.374646912743187e-06,
"loss": 1.365743637084961,
"step": 326
},
{
"epoch": 0.9576642335766423,
"grad_norm": 0.373046875,
"learning_rate": 8.352322257045775e-06,
"loss": 1.3169546127319336,
"step": 328
},
{
"epoch": 0.9635036496350365,
"grad_norm": 0.49609375,
"learning_rate": 8.329879632401265e-06,
"loss": 1.3615431785583496,
"step": 330
},
{
"epoch": 0.9693430656934306,
"grad_norm": 0.341796875,
"learning_rate": 8.307319967012211e-06,
"loss": 1.3420912027359009,
"step": 332
},
{
"epoch": 0.9751824817518249,
"grad_norm": 0.8203125,
"learning_rate": 8.284644193921848e-06,
"loss": 1.28525710105896,
"step": 334
},
{
"epoch": 0.981021897810219,
"grad_norm": 0.48046875,
"learning_rate": 8.261853250975487e-06,
"loss": 1.3461776971817017,
"step": 336
},
{
"epoch": 0.9868613138686131,
"grad_norm": 0.546875,
"learning_rate": 8.238948080781752e-06,
"loss": 1.3771872520446777,
"step": 338
},
{
"epoch": 0.9927007299270073,
"grad_norm": 0.71484375,
"learning_rate": 8.215929630673573e-06,
"loss": 1.1941386461257935,
"step": 340
},
{
"epoch": 0.9985401459854014,
"grad_norm": 0.3359375,
"learning_rate": 8.192798852669016e-06,
"loss": 1.3167439699172974,
"step": 342
},
{
"epoch": 1.0029197080291972,
"grad_norm": 0.291015625,
"learning_rate": 8.169556703431916e-06,
"loss": 1.196553111076355,
"step": 344
},
{
"epoch": 1.0087591240875913,
"grad_norm": 0.376953125,
"learning_rate": 8.146204144232285e-06,
"loss": 1.3300058841705322,
"step": 346
},
{
"epoch": 1.0145985401459854,
"grad_norm": 0.388671875,
"learning_rate": 8.122742140906585e-06,
"loss": 1.3176811933517456,
"step": 348
},
{
"epoch": 1.0204379562043795,
"grad_norm": 0.486328125,
"learning_rate": 8.099171663817758e-06,
"loss": 1.2936145067214966,
"step": 350
},
{
"epoch": 1.0262773722627738,
"grad_norm": 0.375,
"learning_rate": 8.07549368781511e-06,
"loss": 1.2210314273834229,
"step": 352
},
{
"epoch": 1.032116788321168,
"grad_norm": 0.625,
"learning_rate": 8.05170919219398e-06,
"loss": 1.2909876108169556,
"step": 354
},
{
"epoch": 1.037956204379562,
"grad_norm": 0.58203125,
"learning_rate": 8.027819160655248e-06,
"loss": 1.2588341236114502,
"step": 356
},
{
"epoch": 1.0437956204379562,
"grad_norm": 0.267578125,
"learning_rate": 8.003824581264639e-06,
"loss": 1.2091121673583984,
"step": 358
},
{
"epoch": 1.0496350364963503,
"grad_norm": 0.259765625,
"learning_rate": 7.979726446411868e-06,
"loss": 1.2468533515930176,
"step": 360
},
{
"epoch": 1.0554744525547446,
"grad_norm": 0.703125,
"learning_rate": 7.955525752769588e-06,
"loss": 1.2384498119354248,
"step": 362
},
{
"epoch": 1.0613138686131387,
"grad_norm": 0.447265625,
"learning_rate": 7.931223501252176e-06,
"loss": 1.2103887796401978,
"step": 364
},
{
"epoch": 1.0671532846715328,
"grad_norm": 0.375,
"learning_rate": 7.906820696974325e-06,
"loss": 1.2475626468658447,
"step": 366
},
{
"epoch": 1.072992700729927,
"grad_norm": 0.474609375,
"learning_rate": 7.882318349209492e-06,
"loss": 1.3049901723861694,
"step": 368
},
{
"epoch": 1.0788321167883212,
"grad_norm": 0.478515625,
"learning_rate": 7.857717471348135e-06,
"loss": 1.3034197092056274,
"step": 370
},
{
"epoch": 1.0846715328467154,
"grad_norm": 0.5078125,
"learning_rate": 7.833019080855812e-06,
"loss": 1.177399754524231,
"step": 372
},
{
"epoch": 1.0905109489051095,
"grad_norm": 0.69921875,
"learning_rate": 7.808224199231099e-06,
"loss": 1.2648054361343384,
"step": 374
},
{
"epoch": 1.0963503649635036,
"grad_norm": 0.349609375,
"learning_rate": 7.783333851963337e-06,
"loss": 1.1547051668167114,
"step": 376
},
{
"epoch": 1.102189781021898,
"grad_norm": 0.337890625,
"learning_rate": 7.75834906849023e-06,
"loss": 1.2063677310943604,
"step": 378
},
{
"epoch": 1.108029197080292,
"grad_norm": 0.431640625,
"learning_rate": 7.733270882155252e-06,
"loss": 1.294480323791504,
"step": 380
},
{
"epoch": 1.1138686131386861,
"grad_norm": 0.5078125,
"learning_rate": 7.708100330164922e-06,
"loss": 1.1963160037994385,
"step": 382
},
{
"epoch": 1.1197080291970802,
"grad_norm": 0.82421875,
"learning_rate": 7.682838453545898e-06,
"loss": 1.2486681938171387,
"step": 384
},
{
"epoch": 1.1255474452554743,
"grad_norm": 0.294921875,
"learning_rate": 7.657486297101933e-06,
"loss": 1.2629107236862183,
"step": 386
},
{
"epoch": 1.1313868613138687,
"grad_norm": 0.486328125,
"learning_rate": 7.63204490937065e-06,
"loss": 1.200867772102356,
"step": 388
},
{
"epoch": 1.1372262773722628,
"grad_norm": 0.53125,
"learning_rate": 7.606515342580181e-06,
"loss": 1.2637931108474731,
"step": 390
},
{
"epoch": 1.143065693430657,
"grad_norm": 0.369140625,
"learning_rate": 7.580898652605649e-06,
"loss": 1.2616581916809082,
"step": 392
},
{
"epoch": 1.148905109489051,
"grad_norm": 0.48828125,
"learning_rate": 7.555195898925497e-06,
"loss": 1.3153358697891235,
"step": 394
},
{
"epoch": 1.1547445255474453,
"grad_norm": 0.58203125,
"learning_rate": 7.529408144577673e-06,
"loss": 1.3324670791625977,
"step": 396
},
{
"epoch": 1.1605839416058394,
"grad_norm": 0.283203125,
"learning_rate": 7.503536456115653e-06,
"loss": 1.226647138595581,
"step": 398
},
{
"epoch": 1.1664233576642336,
"grad_norm": 1.125,
"learning_rate": 7.47758190356434e-06,
"loss": 1.2243517637252808,
"step": 400
},
{
"epoch": 1.1722627737226277,
"grad_norm": 0.75,
"learning_rate": 7.451545560375808e-06,
"loss": 1.2697927951812744,
"step": 402
},
{
"epoch": 1.178102189781022,
"grad_norm": 1.0703125,
"learning_rate": 7.425428503384898e-06,
"loss": 1.2327685356140137,
"step": 404
},
{
"epoch": 1.183941605839416,
"grad_norm": 1.390625,
"learning_rate": 7.399231812764695e-06,
"loss": 1.3016244173049927,
"step": 406
},
{
"epoch": 1.1897810218978102,
"grad_norm": 0.3046875,
"learning_rate": 7.372956571981831e-06,
"loss": 1.189962387084961,
"step": 408
},
{
"epoch": 1.1956204379562043,
"grad_norm": 2.15625,
"learning_rate": 7.346603867751698e-06,
"loss": 1.235486626625061,
"step": 410
},
{
"epoch": 1.2014598540145984,
"grad_norm": 1.0703125,
"learning_rate": 7.32017478999349e-06,
"loss": 1.2581639289855957,
"step": 412
},
{
"epoch": 1.2072992700729928,
"grad_norm": 0.44140625,
"learning_rate": 7.293670431785123e-06,
"loss": 1.2778874635696411,
"step": 414
},
{
"epoch": 1.2131386861313869,
"grad_norm": 0.66015625,
"learning_rate": 7.267091889318033e-06,
"loss": 1.3071149587631226,
"step": 416
},
{
"epoch": 1.218978102189781,
"grad_norm": 0.404296875,
"learning_rate": 7.240440261851839e-06,
"loss": 1.2892022132873535,
"step": 418
},
{
"epoch": 1.224817518248175,
"grad_norm": 0.828125,
"learning_rate": 7.21371665166887e-06,
"loss": 1.1825231313705444,
"step": 420
},
{
"epoch": 1.2306569343065694,
"grad_norm": 0.703125,
"learning_rate": 7.186922164028589e-06,
"loss": 1.2527143955230713,
"step": 422
},
{
"epoch": 1.2364963503649635,
"grad_norm": 0.71875,
"learning_rate": 7.160057907121868e-06,
"loss": 1.1351896524429321,
"step": 424
},
{
"epoch": 1.2423357664233576,
"grad_norm": 0.47265625,
"learning_rate": 7.133124992025161e-06,
"loss": 1.2882559299468994,
"step": 426
},
{
"epoch": 1.2481751824817517,
"grad_norm": 2.796875,
"learning_rate": 7.106124532654553e-06,
"loss": 1.3231326341629028,
"step": 428
},
{
"epoch": 1.254014598540146,
"grad_norm": 0.328125,
"learning_rate": 7.079057645719686e-06,
"loss": 1.138244867324829,
"step": 430
},
{
"epoch": 1.2598540145985402,
"grad_norm": 0.39453125,
"learning_rate": 7.051925450677568e-06,
"loss": 1.3060168027877808,
"step": 432
},
{
"epoch": 1.2656934306569343,
"grad_norm": 0.484375,
"learning_rate": 7.024729069686288e-06,
"loss": 1.2885905504226685,
"step": 434
},
{
"epoch": 1.2715328467153284,
"grad_norm": 0.2890625,
"learning_rate": 6.997469627558591e-06,
"loss": 1.2515441179275513,
"step": 436
},
{
"epoch": 1.2773722627737225,
"grad_norm": 0.41796875,
"learning_rate": 6.970148251715363e-06,
"loss": 1.2661933898925781,
"step": 438
},
{
"epoch": 1.2832116788321168,
"grad_norm": 0.90625,
"learning_rate": 6.942766072139e-06,
"loss": 1.2927354574203491,
"step": 440
},
{
"epoch": 1.289051094890511,
"grad_norm": 0.62890625,
"learning_rate": 6.9153242213266745e-06,
"loss": 1.2908434867858887,
"step": 442
},
{
"epoch": 1.294890510948905,
"grad_norm": 0.482421875,
"learning_rate": 6.887823834243501e-06,
"loss": 1.2376095056533813,
"step": 444
},
{
"epoch": 1.3007299270072994,
"grad_norm": 0.36328125,
"learning_rate": 6.860266048275585e-06,
"loss": 1.2220968008041382,
"step": 446
},
{
"epoch": 1.3065693430656935,
"grad_norm": 0.52734375,
"learning_rate": 6.832652003182989e-06,
"loss": 1.2590991258621216,
"step": 448
},
{
"epoch": 1.3124087591240876,
"grad_norm": 0.5234375,
"learning_rate": 6.804982841052593e-06,
"loss": 1.2670376300811768,
"step": 450
},
{
"epoch": 1.3182481751824817,
"grad_norm": 1.3984375,
"learning_rate": 6.777259706250856e-06,
"loss": 1.3193187713623047,
"step": 452
},
{
"epoch": 1.3240875912408758,
"grad_norm": 1.0234375,
"learning_rate": 6.749483745376489e-06,
"loss": 1.2359987497329712,
"step": 454
},
{
"epoch": 1.3299270072992702,
"grad_norm": 0.546875,
"learning_rate": 6.721656107213032e-06,
"loss": 1.2815067768096924,
"step": 456
},
{
"epoch": 1.3357664233576643,
"grad_norm": 0.486328125,
"learning_rate": 6.693777942681339e-06,
"loss": 1.2145709991455078,
"step": 458
},
{
"epoch": 1.3416058394160584,
"grad_norm": 0.435546875,
"learning_rate": 6.665850404791986e-06,
"loss": 1.2475862503051758,
"step": 460
},
{
"epoch": 1.3474452554744525,
"grad_norm": 0.4453125,
"learning_rate": 6.637874648597571e-06,
"loss": 1.2392385005950928,
"step": 462
},
{
"epoch": 1.3532846715328466,
"grad_norm": 1.1875,
"learning_rate": 6.60985183114495e-06,
"loss": 1.2735258340835571,
"step": 464
},
{
"epoch": 1.359124087591241,
"grad_norm": 0.34765625,
"learning_rate": 6.5817831114273845e-06,
"loss": 1.26645827293396,
"step": 466
},
{
"epoch": 1.364963503649635,
"grad_norm": 0.396484375,
"learning_rate": 6.553669650336599e-06,
"loss": 1.2172024250030518,
"step": 468
},
{
"epoch": 1.3708029197080291,
"grad_norm": 0.37890625,
"learning_rate": 6.5255126106147785e-06,
"loss": 1.3016058206558228,
"step": 470
},
{
"epoch": 1.3766423357664235,
"grad_norm": 0.400390625,
"learning_rate": 6.497313156806467e-06,
"loss": 1.3059654235839844,
"step": 472
},
{
"epoch": 1.3824817518248176,
"grad_norm": 0.7421875,
"learning_rate": 6.469072455210414e-06,
"loss": 1.2682950496673584,
"step": 474
},
{
"epoch": 1.3883211678832117,
"grad_norm": 0.41015625,
"learning_rate": 6.4407916738313256e-06,
"loss": 1.190187931060791,
"step": 476
},
{
"epoch": 1.3941605839416058,
"grad_norm": 0.51953125,
"learning_rate": 6.412471982331574e-06,
"loss": 1.2461662292480469,
"step": 478
},
{
"epoch": 1.4,
"grad_norm": 0.5,
"learning_rate": 6.384114551982804e-06,
"loss": 1.2167097330093384,
"step": 480
},
{
"epoch": 1.4058394160583942,
"grad_norm": 0.390625,
"learning_rate": 6.3557205556175e-06,
"loss": 1.1439507007598877,
"step": 482
},
{
"epoch": 1.4116788321167884,
"grad_norm": 1.1328125,
"learning_rate": 6.327291167580478e-06,
"loss": 1.207546353340149,
"step": 484
},
{
"epoch": 1.4175182481751825,
"grad_norm": 0.486328125,
"learning_rate": 6.298827563680314e-06,
"loss": 1.292868971824646,
"step": 486
},
{
"epoch": 1.4233576642335766,
"grad_norm": 0.42578125,
"learning_rate": 6.270330921140718e-06,
"loss": 1.3323169946670532,
"step": 488
},
{
"epoch": 1.4291970802919707,
"grad_norm": 0.458984375,
"learning_rate": 6.241802418551838e-06,
"loss": 1.19215726852417,
"step": 490
},
{
"epoch": 1.435036496350365,
"grad_norm": 0.5703125,
"learning_rate": 6.21324323582152e-06,
"loss": 1.2297883033752441,
"step": 492
},
{
"epoch": 1.4408759124087591,
"grad_norm": 0.33984375,
"learning_rate": 6.1846545541265115e-06,
"loss": 1.3055942058563232,
"step": 494
},
{
"epoch": 1.4467153284671532,
"grad_norm": 0.4140625,
"learning_rate": 6.156037555863597e-06,
"loss": 1.2372398376464844,
"step": 496
},
{
"epoch": 1.4525547445255476,
"grad_norm": 0.7265625,
"learning_rate": 6.1273934246007105e-06,
"loss": 1.3352288007736206,
"step": 498
},
{
"epoch": 1.4583941605839417,
"grad_norm": 0.5390625,
"learning_rate": 6.098723345027977e-06,
"loss": 1.2846771478652954,
"step": 500
},
{
"epoch": 1.4642335766423358,
"grad_norm": 0.5703125,
"learning_rate": 6.070028502908709e-06,
"loss": 1.2137339115142822,
"step": 502
},
{
"epoch": 1.4700729927007299,
"grad_norm": 0.470703125,
"learning_rate": 6.041310085030378e-06,
"loss": 1.2333766222000122,
"step": 504
},
{
"epoch": 1.475912408759124,
"grad_norm": 0.400390625,
"learning_rate": 6.012569279155521e-06,
"loss": 1.217051386833191,
"step": 506
},
{
"epoch": 1.4817518248175183,
"grad_norm": 1.078125,
"learning_rate": 5.983807273972617e-06,
"loss": 1.1864327192306519,
"step": 508
},
{
"epoch": 1.4875912408759124,
"grad_norm": 0.46484375,
"learning_rate": 5.955025259046927e-06,
"loss": 1.2389144897460938,
"step": 510
},
{
"epoch": 1.4934306569343065,
"grad_norm": 0.71875,
"learning_rate": 5.926224424771291e-06,
"loss": 1.2382243871688843,
"step": 512
},
{
"epoch": 1.4992700729927007,
"grad_norm": 0.4921875,
"learning_rate": 5.897405962316899e-06,
"loss": 1.208208680152893,
"step": 514
},
{
"epoch": 1.5051094890510948,
"grad_norm": 0.44921875,
"learning_rate": 5.868571063584023e-06,
"loss": 1.286102533340454,
"step": 516
},
{
"epoch": 1.510948905109489,
"grad_norm": 0.494140625,
"learning_rate": 5.83972092115272e-06,
"loss": 1.2459356784820557,
"step": 518
},
{
"epoch": 1.5167883211678832,
"grad_norm": 0.53515625,
"learning_rate": 5.810856728233513e-06,
"loss": 1.276041865348816,
"step": 520
},
{
"epoch": 1.5226277372262773,
"grad_norm": 0.734375,
"learning_rate": 5.781979678618033e-06,
"loss": 1.27890944480896,
"step": 522
},
{
"epoch": 1.5284671532846716,
"grad_norm": 1.7265625,
"learning_rate": 5.753090966629654e-06,
"loss": 1.1399537324905396,
"step": 524
},
{
"epoch": 1.5343065693430655,
"grad_norm": 0.388671875,
"learning_rate": 5.7241917870740935e-06,
"loss": 1.2000788450241089,
"step": 526
},
{
"epoch": 1.5401459854014599,
"grad_norm": 0.451171875,
"learning_rate": 5.695283335189992e-06,
"loss": 1.3109184503555298,
"step": 528
},
{
"epoch": 1.545985401459854,
"grad_norm": 0.5859375,
"learning_rate": 5.666366806599488e-06,
"loss": 1.2306164503097534,
"step": 530
},
{
"epoch": 1.551824817518248,
"grad_norm": 0.3359375,
"learning_rate": 5.63744339725876e-06,
"loss": 1.2434693574905396,
"step": 532
},
{
"epoch": 1.5576642335766424,
"grad_norm": 0.8125,
"learning_rate": 5.608514303408572e-06,
"loss": 1.1840685606002808,
"step": 534
},
{
"epoch": 1.5635036496350365,
"grad_norm": 0.90234375,
"learning_rate": 5.579580721524786e-06,
"loss": 1.2241274118423462,
"step": 536
},
{
"epoch": 1.5693430656934306,
"grad_norm": 0.478515625,
"learning_rate": 5.550643848268889e-06,
"loss": 1.206296682357788,
"step": 538
},
{
"epoch": 1.575182481751825,
"grad_norm": 1.6484375,
"learning_rate": 5.5217048804384945e-06,
"loss": 1.2080531120300293,
"step": 540
},
{
"epoch": 1.5810218978102188,
"grad_norm": 0.498046875,
"learning_rate": 5.492765014917845e-06,
"loss": 1.207125186920166,
"step": 542
},
{
"epoch": 1.5868613138686132,
"grad_norm": 1.0546875,
"learning_rate": 5.463825448628314e-06,
"loss": 1.218297004699707,
"step": 544
},
{
"epoch": 1.5927007299270073,
"grad_norm": 0.431640625,
"learning_rate": 5.434887378478892e-06,
"loss": 1.1724162101745605,
"step": 546
},
{
"epoch": 1.5985401459854014,
"grad_norm": 0.4765625,
"learning_rate": 5.405952001316697e-06,
"loss": 1.215348720550537,
"step": 548
},
{
"epoch": 1.6043795620437957,
"grad_norm": 0.3359375,
"learning_rate": 5.377020513877463e-06,
"loss": 1.2883554697036743,
"step": 550
},
{
"epoch": 1.6102189781021898,
"grad_norm": 0.490234375,
"learning_rate": 5.348094112736057e-06,
"loss": 1.253550410270691,
"step": 552
},
{
"epoch": 1.616058394160584,
"grad_norm": 0.4921875,
"learning_rate": 5.319173994256973e-06,
"loss": 1.29800546169281,
"step": 554
},
{
"epoch": 1.621897810218978,
"grad_norm": 0.390625,
"learning_rate": 5.290261354544867e-06,
"loss": 1.1871838569641113,
"step": 556
},
{
"epoch": 1.6277372262773722,
"grad_norm": 0.65625,
"learning_rate": 5.261357389395078e-06,
"loss": 1.1837831735610962,
"step": 558
},
{
"epoch": 1.6335766423357665,
"grad_norm": 1.125,
"learning_rate": 5.232463294244178e-06,
"loss": 1.2513107061386108,
"step": 560
},
{
"epoch": 1.6394160583941606,
"grad_norm": 0.52734375,
"learning_rate": 5.203580264120521e-06,
"loss": 1.3240464925765991,
"step": 562
},
{
"epoch": 1.6452554744525547,
"grad_norm": 0.52734375,
"learning_rate": 5.1747094935948325e-06,
"loss": 1.2752066850662231,
"step": 564
},
{
"epoch": 1.651094890510949,
"grad_norm": 0.546875,
"learning_rate": 5.145852176730786e-06,
"loss": 1.1872999668121338,
"step": 566
},
{
"epoch": 1.656934306569343,
"grad_norm": 0.69140625,
"learning_rate": 5.117009507035628e-06,
"loss": 1.2202606201171875,
"step": 568
},
{
"epoch": 1.6627737226277373,
"grad_norm": 0.625,
"learning_rate": 5.088182677410819e-06,
"loss": 1.2224456071853638,
"step": 570
},
{
"epoch": 1.6686131386861314,
"grad_norm": 0.6328125,
"learning_rate": 5.059372880102683e-06,
"loss": 1.2351001501083374,
"step": 572
},
{
"epoch": 1.6744525547445255,
"grad_norm": 0.609375,
"learning_rate": 5.0305813066531136e-06,
"loss": 1.2855942249298096,
"step": 574
},
{
"epoch": 1.6802919708029198,
"grad_norm": 0.98828125,
"learning_rate": 5.001809147850282e-06,
"loss": 1.2410361766815186,
"step": 576
},
{
"epoch": 1.686131386861314,
"grad_norm": 0.431640625,
"learning_rate": 4.97305759367939e-06,
"loss": 1.2131555080413818,
"step": 578
},
{
"epoch": 1.691970802919708,
"grad_norm": 0.4375,
"learning_rate": 4.944327833273456e-06,
"loss": 1.2389878034591675,
"step": 580
},
{
"epoch": 1.6978102189781021,
"grad_norm": 0.46484375,
"learning_rate": 4.9156210548641315e-06,
"loss": 1.2704704999923706,
"step": 582
},
{
"epoch": 1.7036496350364962,
"grad_norm": 0.50390625,
"learning_rate": 4.886938445732557e-06,
"loss": 1.2079095840454102,
"step": 584
},
{
"epoch": 1.7094890510948906,
"grad_norm": 0.5703125,
"learning_rate": 4.858281192160258e-06,
"loss": 1.3086167573928833,
"step": 586
},
{
"epoch": 1.7153284671532847,
"grad_norm": 0.37109375,
"learning_rate": 4.82965047938008e-06,
"loss": 1.1482274532318115,
"step": 588
},
{
"epoch": 1.7211678832116788,
"grad_norm": 0.349609375,
"learning_rate": 4.801047491527176e-06,
"loss": 1.316062331199646,
"step": 590
},
{
"epoch": 1.7270072992700731,
"grad_norm": 0.8359375,
"learning_rate": 4.772473411590022e-06,
"loss": 1.271776556968689,
"step": 592
},
{
"epoch": 1.732846715328467,
"grad_norm": 0.427734375,
"learning_rate": 4.743929421361492e-06,
"loss": 1.1389895677566528,
"step": 594
},
{
"epoch": 1.7386861313868613,
"grad_norm": 7.53125,
"learning_rate": 4.715416701389985e-06,
"loss": 1.196537733078003,
"step": 596
},
{
"epoch": 1.7445255474452555,
"grad_norm": 0.333984375,
"learning_rate": 4.686936430930597e-06,
"loss": 1.2567116022109985,
"step": 598
},
{
"epoch": 1.7503649635036496,
"grad_norm": 0.427734375,
"learning_rate": 4.658489787896346e-06,
"loss": 1.3270188570022583,
"step": 600
},
{
"epoch": 1.756204379562044,
"grad_norm": 0.458984375,
"learning_rate": 4.630077948809457e-06,
"loss": 1.2750407457351685,
"step": 602
},
{
"epoch": 1.762043795620438,
"grad_norm": 0.447265625,
"learning_rate": 4.601702088752702e-06,
"loss": 1.2777055501937866,
"step": 604
},
{
"epoch": 1.7678832116788321,
"grad_norm": 0.361328125,
"learning_rate": 4.573363381320795e-06,
"loss": 1.1461223363876343,
"step": 606
},
{
"epoch": 1.7737226277372264,
"grad_norm": 0.498046875,
"learning_rate": 4.5450629985718655e-06,
"loss": 1.2125712633132935,
"step": 608
},
{
"epoch": 1.7795620437956203,
"grad_norm": 0.453125,
"learning_rate": 4.516802110978968e-06,
"loss": 1.259176254272461,
"step": 610
},
{
"epoch": 1.7854014598540147,
"grad_norm": 1.3203125,
"learning_rate": 4.488581887381689e-06,
"loss": 1.1314493417739868,
"step": 612
},
{
"epoch": 1.7912408759124088,
"grad_norm": 0.39453125,
"learning_rate": 4.460403494937787e-06,
"loss": 1.2192261219024658,
"step": 614
},
{
"epoch": 1.7970802919708029,
"grad_norm": 0.5859375,
"learning_rate": 4.4322680990749365e-06,
"loss": 1.1592247486114502,
"step": 616
},
{
"epoch": 1.8029197080291972,
"grad_norm": 0.404296875,
"learning_rate": 4.404176863442515e-06,
"loss": 1.300370454788208,
"step": 618
},
{
"epoch": 1.808759124087591,
"grad_norm": 1.65625,
"learning_rate": 4.376130949863487e-06,
"loss": 1.2521164417266846,
"step": 620
},
{
"epoch": 1.8145985401459854,
"grad_norm": 0.5703125,
"learning_rate": 4.3481315182863405e-06,
"loss": 1.1387872695922852,
"step": 622
},
{
"epoch": 1.8204379562043795,
"grad_norm": 0.58984375,
"learning_rate": 4.320179726737118e-06,
"loss": 1.2439383268356323,
"step": 624
},
{
"epoch": 1.8262773722627736,
"grad_norm": 0.859375,
"learning_rate": 4.292276731271528e-06,
"loss": 1.1349395513534546,
"step": 626
},
{
"epoch": 1.832116788321168,
"grad_norm": 0.51953125,
"learning_rate": 4.264423685927123e-06,
"loss": 1.2615232467651367,
"step": 628
},
{
"epoch": 1.837956204379562,
"grad_norm": 1.1328125,
"learning_rate": 4.236621742675572e-06,
"loss": 1.297728419303894,
"step": 630
},
{
"epoch": 1.8437956204379562,
"grad_norm": 0.89453125,
"learning_rate": 4.208872051375016e-06,
"loss": 1.3184691667556763,
"step": 632
},
{
"epoch": 1.8496350364963505,
"grad_norm": 0.4140625,
"learning_rate": 4.1811757597225154e-06,
"loss": 1.2532655000686646,
"step": 634
},
{
"epoch": 1.8554744525547444,
"grad_norm": 0.7421875,
"learning_rate": 4.1535340132065775e-06,
"loss": 1.2408058643341064,
"step": 636
},
{
"epoch": 1.8613138686131387,
"grad_norm": 0.52734375,
"learning_rate": 4.125947955059781e-06,
"loss": 1.298876404762268,
"step": 638
},
{
"epoch": 1.8671532846715329,
"grad_norm": 0.58203125,
"learning_rate": 4.098418726211497e-06,
"loss": 1.1866939067840576,
"step": 640
},
{
"epoch": 1.872992700729927,
"grad_norm": 0.515625,
"learning_rate": 4.070947465240695e-06,
"loss": 1.2406346797943115,
"step": 642
},
{
"epoch": 1.8788321167883213,
"grad_norm": 0.45703125,
"learning_rate": 4.043535308328861e-06,
"loss": 1.2553025484085083,
"step": 644
},
{
"epoch": 1.8846715328467152,
"grad_norm": 0.50390625,
"learning_rate": 4.0161833892129944e-06,
"loss": 1.2756036520004272,
"step": 646
},
{
"epoch": 1.8905109489051095,
"grad_norm": 0.546875,
"learning_rate": 3.988892839138731e-06,
"loss": 1.185132622718811,
"step": 648
},
{
"epoch": 1.8963503649635036,
"grad_norm": 0.83203125,
"learning_rate": 3.96166478681355e-06,
"loss": 1.221949815750122,
"step": 650
},
{
"epoch": 1.9021897810218977,
"grad_norm": 0.63671875,
"learning_rate": 3.934500358360085e-06,
"loss": 1.2973777055740356,
"step": 652
},
{
"epoch": 1.908029197080292,
"grad_norm": 0.373046875,
"learning_rate": 3.907400677269563e-06,
"loss": 1.2753914594650269,
"step": 654
},
{
"epoch": 1.9138686131386862,
"grad_norm": 0.4609375,
"learning_rate": 3.880366864355325e-06,
"loss": 1.2708187103271484,
"step": 656
},
{
"epoch": 1.9197080291970803,
"grad_norm": 0.380859375,
"learning_rate": 3.853400037706478e-06,
"loss": 1.2944742441177368,
"step": 658
},
{
"epoch": 1.9255474452554746,
"grad_norm": 0.41015625,
"learning_rate": 3.826501312641655e-06,
"loss": 1.2255587577819824,
"step": 660
},
{
"epoch": 1.9313868613138685,
"grad_norm": 1.4609375,
"learning_rate": 3.7996718016628697e-06,
"loss": 1.168819546699524,
"step": 662
},
{
"epoch": 1.9372262773722628,
"grad_norm": 0.62109375,
"learning_rate": 3.772912614409526e-06,
"loss": 1.141089916229248,
"step": 664
},
{
"epoch": 1.943065693430657,
"grad_norm": 0.388671875,
"learning_rate": 3.7462248576125125e-06,
"loss": 1.2469316720962524,
"step": 666
},
{
"epoch": 1.948905109489051,
"grad_norm": 0.53125,
"learning_rate": 3.719609635048431e-06,
"loss": 1.222388505935669,
"step": 668
},
{
"epoch": 1.9547445255474454,
"grad_norm": 0.47265625,
"learning_rate": 3.6930680474939486e-06,
"loss": 1.1344715356826782,
"step": 670
},
{
"epoch": 1.9605839416058393,
"grad_norm": 0.5625,
"learning_rate": 3.6666011926802647e-06,
"loss": 1.2021689414978027,
"step": 672
},
{
"epoch": 1.9664233576642336,
"grad_norm": 0.51171875,
"learning_rate": 3.6402101652477163e-06,
"loss": 1.251044750213623,
"step": 674
},
{
"epoch": 1.9722627737226277,
"grad_norm": 0.578125,
"learning_rate": 3.613896056700502e-06,
"loss": 1.258431315422058,
"step": 676
},
{
"epoch": 1.9781021897810218,
"grad_norm": 0.51171875,
"learning_rate": 3.58765995536154e-06,
"loss": 1.2117798328399658,
"step": 678
},
{
"epoch": 1.9839416058394161,
"grad_norm": 1.390625,
"learning_rate": 3.561502946327452e-06,
"loss": 1.2486507892608643,
"step": 680
},
{
"epoch": 1.9897810218978103,
"grad_norm": 0.73046875,
"learning_rate": 3.535426111423691e-06,
"loss": 1.1538894176483154,
"step": 682
},
{
"epoch": 1.9956204379562044,
"grad_norm": 0.5703125,
"learning_rate": 3.509430529159794e-06,
"loss": 1.2577342987060547,
"step": 684
},
{
"epoch": 2.0,
"grad_norm": 1.2421875,
"learning_rate": 3.483517274684778e-06,
"loss": 1.3043779134750366,
"step": 686
},
{
"epoch": 2.0058394160583943,
"grad_norm": 0.8359375,
"learning_rate": 3.457687419742668e-06,
"loss": 1.1992483139038086,
"step": 688
},
{
"epoch": 2.011678832116788,
"grad_norm": 0.42578125,
"learning_rate": 3.431942032628184e-06,
"loss": 1.2118330001831055,
"step": 690
},
{
"epoch": 2.0175182481751825,
"grad_norm": 0.34375,
"learning_rate": 3.4062821781425402e-06,
"loss": 1.1599669456481934,
"step": 692
},
{
"epoch": 2.0233576642335764,
"grad_norm": 0.30078125,
"learning_rate": 3.3807089175494175e-06,
"loss": 1.1471362113952637,
"step": 694
},
{
"epoch": 2.0291970802919708,
"grad_norm": 0.421875,
"learning_rate": 3.355223308531066e-06,
"loss": 1.3080707788467407,
"step": 696
},
{
"epoch": 2.035036496350365,
"grad_norm": 0.55859375,
"learning_rate": 3.3298264051445655e-06,
"loss": 1.2860289812088013,
"step": 698
},
{
"epoch": 2.040875912408759,
"grad_norm": 0.4921875,
"learning_rate": 3.3045192577782214e-06,
"loss": 1.1995092630386353,
"step": 700
},
{
"epoch": 2.0467153284671533,
"grad_norm": 0.5078125,
"learning_rate": 3.2793029131081335e-06,
"loss": 1.1927409172058105,
"step": 702
},
{
"epoch": 2.0525547445255476,
"grad_norm": 0.38671875,
"learning_rate": 3.254178414054896e-06,
"loss": 1.1319977045059204,
"step": 704
},
{
"epoch": 2.0583941605839415,
"grad_norm": 0.41015625,
"learning_rate": 3.2291467997404747e-06,
"loss": 1.238900899887085,
"step": 706
},
{
"epoch": 2.064233576642336,
"grad_norm": 0.439453125,
"learning_rate": 3.2042091054452175e-06,
"loss": 1.1581428050994873,
"step": 708
},
{
"epoch": 2.0700729927007298,
"grad_norm": 0.31640625,
"learning_rate": 3.1793663625650444e-06,
"loss": 1.2304531335830688,
"step": 710
},
{
"epoch": 2.075912408759124,
"grad_norm": 0.37109375,
"learning_rate": 3.154619598568789e-06,
"loss": 1.2318830490112305,
"step": 712
},
{
"epoch": 2.0817518248175184,
"grad_norm": 0.52734375,
"learning_rate": 3.1299698369557026e-06,
"loss": 1.2299096584320068,
"step": 714
},
{
"epoch": 2.0875912408759123,
"grad_norm": 0.52734375,
"learning_rate": 3.105418097213121e-06,
"loss": 1.154068112373352,
"step": 716
},
{
"epoch": 2.0934306569343066,
"grad_norm": 0.41796875,
"learning_rate": 3.0809653947743044e-06,
"loss": 1.0612831115722656,
"step": 718
},
{
"epoch": 2.0992700729927005,
"grad_norm": 0.298828125,
"learning_rate": 3.0566127409764377e-06,
"loss": 1.154129147529602,
"step": 720
},
{
"epoch": 2.105109489051095,
"grad_norm": 0.7890625,
"learning_rate": 3.0323611430188026e-06,
"loss": 1.1408308744430542,
"step": 722
},
{
"epoch": 2.110948905109489,
"grad_norm": 0.734375,
"learning_rate": 3.008211603921118e-06,
"loss": 1.1843266487121582,
"step": 724
},
{
"epoch": 2.116788321167883,
"grad_norm": 0.41015625,
"learning_rate": 2.9841651224820656e-06,
"loss": 1.0983469486236572,
"step": 726
},
{
"epoch": 2.1226277372262774,
"grad_norm": 0.361328125,
"learning_rate": 2.96022269323797e-06,
"loss": 1.2294151782989502,
"step": 728
},
{
"epoch": 2.1284671532846717,
"grad_norm": 0.400390625,
"learning_rate": 2.9363853064216706e-06,
"loss": 1.3076847791671753,
"step": 730
},
{
"epoch": 2.1343065693430656,
"grad_norm": 0.61328125,
"learning_rate": 2.912653947921567e-06,
"loss": 1.185577392578125,
"step": 732
},
{
"epoch": 2.14014598540146,
"grad_norm": 0.408203125,
"learning_rate": 2.8890295992408425e-06,
"loss": 1.1940393447875977,
"step": 734
},
{
"epoch": 2.145985401459854,
"grad_norm": 0.4296875,
"learning_rate": 2.86551323745687e-06,
"loss": 1.1575770378112793,
"step": 736
},
{
"epoch": 2.151824817518248,
"grad_norm": 0.35546875,
"learning_rate": 2.8421058351808055e-06,
"loss": 1.1326178312301636,
"step": 738
},
{
"epoch": 2.1576642335766425,
"grad_norm": 0.6171875,
"learning_rate": 2.818808360517349e-06,
"loss": 1.167670726776123,
"step": 740
},
{
"epoch": 2.1635036496350364,
"grad_norm": 0.482421875,
"learning_rate": 2.7956217770247262e-06,
"loss": 1.3139435052871704,
"step": 742
},
{
"epoch": 2.1693430656934307,
"grad_norm": 0.46875,
"learning_rate": 2.7725470436748165e-06,
"loss": 1.12082839012146,
"step": 744
},
{
"epoch": 2.1751824817518246,
"grad_norm": 0.53125,
"learning_rate": 2.7495851148135005e-06,
"loss": 1.1051329374313354,
"step": 746
},
{
"epoch": 2.181021897810219,
"grad_norm": 0.93359375,
"learning_rate": 2.7267369401211895e-06,
"loss": 1.1831530332565308,
"step": 748
},
{
"epoch": 2.1868613138686133,
"grad_norm": 0.36328125,
"learning_rate": 2.704003464573544e-06,
"loss": 1.1686221361160278,
"step": 750
},
{
"epoch": 2.192700729927007,
"grad_norm": 0.326171875,
"learning_rate": 2.6813856284023943e-06,
"loss": 1.1672006845474243,
"step": 752
},
{
"epoch": 2.1985401459854015,
"grad_norm": 0.404296875,
"learning_rate": 2.6588843670568505e-06,
"loss": 1.254637360572815,
"step": 754
},
{
"epoch": 2.204379562043796,
"grad_norm": 0.52734375,
"learning_rate": 2.6365006111646152e-06,
"loss": 1.1789019107818604,
"step": 756
},
{
"epoch": 2.2102189781021897,
"grad_norm": 0.474609375,
"learning_rate": 2.614235286493494e-06,
"loss": 1.1955108642578125,
"step": 758
},
{
"epoch": 2.216058394160584,
"grad_norm": 0.87109375,
"learning_rate": 2.5920893139131043e-06,
"loss": 1.2021061182022095,
"step": 760
},
{
"epoch": 2.221897810218978,
"grad_norm": 0.451171875,
"learning_rate": 2.570063609356791e-06,
"loss": 1.2603737115859985,
"step": 762
},
{
"epoch": 2.2277372262773723,
"grad_norm": 0.431640625,
"learning_rate": 2.5481590837837477e-06,
"loss": 1.1586928367614746,
"step": 764
},
{
"epoch": 2.2335766423357666,
"grad_norm": 0.443359375,
"learning_rate": 2.5263766431413327e-06,
"loss": 1.3295210599899292,
"step": 766
},
{
"epoch": 2.2394160583941605,
"grad_norm": 0.69140625,
"learning_rate": 2.5047171883276065e-06,
"loss": 1.116708517074585,
"step": 768
},
{
"epoch": 2.245255474452555,
"grad_norm": 0.37890625,
"learning_rate": 2.4831816151540682e-06,
"loss": 1.0877715349197388,
"step": 770
},
{
"epoch": 2.2510948905109487,
"grad_norm": 0.49609375,
"learning_rate": 2.4617708143086085e-06,
"loss": 1.3073028326034546,
"step": 772
},
{
"epoch": 2.256934306569343,
"grad_norm": 0.7578125,
"learning_rate": 2.440485671318668e-06,
"loss": 1.2231556177139282,
"step": 774
},
{
"epoch": 2.2627737226277373,
"grad_norm": 0.796875,
"learning_rate": 2.41932706651462e-06,
"loss": 1.1234259605407715,
"step": 776
},
{
"epoch": 2.2686131386861312,
"grad_norm": 0.443359375,
"learning_rate": 2.398295874993347e-06,
"loss": 1.1716923713684082,
"step": 778
},
{
"epoch": 2.2744525547445256,
"grad_norm": 0.65625,
"learning_rate": 2.3773929665820662e-06,
"loss": 1.2562777996063232,
"step": 780
},
{
"epoch": 2.28029197080292,
"grad_norm": 0.458984375,
"learning_rate": 2.35661920580234e-06,
"loss": 1.2565009593963623,
"step": 782
},
{
"epoch": 2.286131386861314,
"grad_norm": 1.140625,
"learning_rate": 2.3359754518343255e-06,
"loss": 1.2188069820404053,
"step": 784
},
{
"epoch": 2.291970802919708,
"grad_norm": 1.3125,
"learning_rate": 2.315462558481241e-06,
"loss": 1.1537082195281982,
"step": 786
},
{
"epoch": 2.297810218978102,
"grad_norm": 0.357421875,
"learning_rate": 2.29508137413405e-06,
"loss": 1.0930520296096802,
"step": 788
},
{
"epoch": 2.3036496350364963,
"grad_norm": 0.5078125,
"learning_rate": 2.274832741736376e-06,
"loss": 1.107776165008545,
"step": 790
},
{
"epoch": 2.3094890510948907,
"grad_norm": 0.40234375,
"learning_rate": 2.254717498749638e-06,
"loss": 1.291731357574463,
"step": 792
},
{
"epoch": 2.3153284671532846,
"grad_norm": 0.396484375,
"learning_rate": 2.234736477118413e-06,
"loss": 1.1373404264450073,
"step": 794
},
{
"epoch": 2.321167883211679,
"grad_norm": 0.380859375,
"learning_rate": 2.2148905032360305e-06,
"loss": 1.1539984941482544,
"step": 796
},
{
"epoch": 2.3270072992700728,
"grad_norm": 0.46875,
"learning_rate": 2.195180397910389e-06,
"loss": 1.1758666038513184,
"step": 798
},
{
"epoch": 2.332846715328467,
"grad_norm": 1.8984375,
"learning_rate": 2.1756069763300158e-06,
"loss": 1.2303135395050049,
"step": 800
},
{
"epoch": 2.3386861313868614,
"grad_norm": 0.546875,
"learning_rate": 2.1561710480303435e-06,
"loss": 1.1783814430236816,
"step": 802
},
{
"epoch": 2.3445255474452553,
"grad_norm": 0.345703125,
"learning_rate": 2.1368734168602318e-06,
"loss": 1.2231682538986206,
"step": 804
},
{
"epoch": 2.3503649635036497,
"grad_norm": 0.4453125,
"learning_rate": 2.1177148809487237e-06,
"loss": 1.1812344789505005,
"step": 806
},
{
"epoch": 2.356204379562044,
"grad_norm": 1.5546875,
"learning_rate": 2.0986962326720328e-06,
"loss": 1.1925325393676758,
"step": 808
},
{
"epoch": 2.362043795620438,
"grad_norm": 0.4140625,
"learning_rate": 2.079818258620771e-06,
"loss": 1.1922979354858398,
"step": 810
},
{
"epoch": 2.367883211678832,
"grad_norm": 1.7578125,
"learning_rate": 2.0610817395674197e-06,
"loss": 1.2363067865371704,
"step": 812
},
{
"epoch": 2.373722627737226,
"grad_norm": 0.41015625,
"learning_rate": 2.042487450434033e-06,
"loss": 1.1505279541015625,
"step": 814
},
{
"epoch": 2.3795620437956204,
"grad_norm": 0.765625,
"learning_rate": 2.0240361602601906e-06,
"loss": 1.1627535820007324,
"step": 816
},
{
"epoch": 2.3854014598540147,
"grad_norm": 0.33203125,
"learning_rate": 2.0057286321711924e-06,
"loss": 1.155818223953247,
"step": 818
},
{
"epoch": 2.3912408759124086,
"grad_norm": 0.5078125,
"learning_rate": 1.987565623346492e-06,
"loss": 1.2554137706756592,
"step": 820
},
{
"epoch": 2.397080291970803,
"grad_norm": 0.546875,
"learning_rate": 1.9695478849883867e-06,
"loss": 1.1215136051177979,
"step": 822
},
{
"epoch": 2.402919708029197,
"grad_norm": 0.3515625,
"learning_rate": 1.9516761622909423e-06,
"loss": 1.1090680360794067,
"step": 824
},
{
"epoch": 2.408759124087591,
"grad_norm": 0.416015625,
"learning_rate": 1.9339511944091773e-06,
"loss": 1.1436755657196045,
"step": 826
},
{
"epoch": 2.4145985401459855,
"grad_norm": 0.48046875,
"learning_rate": 1.91637371442849e-06,
"loss": 1.140608549118042,
"step": 828
},
{
"epoch": 2.4204379562043794,
"grad_norm": 0.54296875,
"learning_rate": 1.8989444493343401e-06,
"loss": 1.2412205934524536,
"step": 830
},
{
"epoch": 2.4262773722627737,
"grad_norm": 0.71484375,
"learning_rate": 1.8816641199821797e-06,
"loss": 1.2433428764343262,
"step": 832
},
{
"epoch": 2.432116788321168,
"grad_norm": 0.4140625,
"learning_rate": 1.8645334410676413e-06,
"loss": 1.2072176933288574,
"step": 834
},
{
"epoch": 2.437956204379562,
"grad_norm": 0.5703125,
"learning_rate": 1.8475531210969766e-06,
"loss": 1.174179196357727,
"step": 836
},
{
"epoch": 2.4437956204379563,
"grad_norm": 0.87890625,
"learning_rate": 1.8307238623577588e-06,
"loss": 1.2336622476577759,
"step": 838
},
{
"epoch": 2.44963503649635,
"grad_norm": 0.40625,
"learning_rate": 1.814046360889829e-06,
"loss": 1.2075152397155762,
"step": 840
},
{
"epoch": 2.4554744525547445,
"grad_norm": 0.3515625,
"learning_rate": 1.7975213064565136e-06,
"loss": 1.1945725679397583,
"step": 842
},
{
"epoch": 2.461313868613139,
"grad_norm": 0.5703125,
"learning_rate": 1.7811493825160952e-06,
"loss": 1.255374789237976,
"step": 844
},
{
"epoch": 2.4671532846715327,
"grad_norm": 0.326171875,
"learning_rate": 1.7649312661935463e-06,
"loss": 1.0935304164886475,
"step": 846
},
{
"epoch": 2.472992700729927,
"grad_norm": 1.2109375,
"learning_rate": 1.7488676282525236e-06,
"loss": 1.1871508359909058,
"step": 848
},
{
"epoch": 2.478832116788321,
"grad_norm": 0.515625,
"learning_rate": 1.7329591330676255e-06,
"loss": 1.236677646636963,
"step": 850
},
{
"epoch": 2.4846715328467153,
"grad_norm": 0.59375,
"learning_rate": 1.7172064385969144e-06,
"loss": 1.1770851612091064,
"step": 852
},
{
"epoch": 2.4905109489051096,
"grad_norm": 0.76953125,
"learning_rate": 1.7016101963547063e-06,
"loss": 1.1361277103424072,
"step": 854
},
{
"epoch": 2.4963503649635035,
"grad_norm": 0.5078125,
"learning_rate": 1.68617105138462e-06,
"loss": 1.1203017234802246,
"step": 856
},
{
"epoch": 2.502189781021898,
"grad_norm": 0.453125,
"learning_rate": 1.6708896422329056e-06,
"loss": 1.2030003070831299,
"step": 858
},
{
"epoch": 2.508029197080292,
"grad_norm": 0.5234375,
"learning_rate": 1.6557666009220274e-06,
"loss": 1.194663405418396,
"step": 860
},
{
"epoch": 2.513868613138686,
"grad_norm": 0.74609375,
"learning_rate": 1.6408025529245324e-06,
"loss": 1.3168246746063232,
"step": 862
},
{
"epoch": 2.5197080291970804,
"grad_norm": 0.43359375,
"learning_rate": 1.6259981171371736e-06,
"loss": 1.1622366905212402,
"step": 864
},
{
"epoch": 2.5255474452554747,
"grad_norm": 0.333984375,
"learning_rate": 1.611353905855317e-06,
"loss": 1.1341776847839355,
"step": 866
},
{
"epoch": 2.5313868613138686,
"grad_norm": 0.47265625,
"learning_rate": 1.5968705247476192e-06,
"loss": 1.1897858381271362,
"step": 868
},
{
"epoch": 2.537226277372263,
"grad_norm": 0.486328125,
"learning_rate": 1.582548572830973e-06,
"loss": 1.1871055364608765,
"step": 870
},
{
"epoch": 2.543065693430657,
"grad_norm": 0.55859375,
"learning_rate": 1.5683886424457373e-06,
"loss": 1.1016008853912354,
"step": 872
},
{
"epoch": 2.548905109489051,
"grad_norm": 0.7734375,
"learning_rate": 1.5543913192312373e-06,
"loss": 1.1918286085128784,
"step": 874
},
{
"epoch": 2.554744525547445,
"grad_norm": 0.5078125,
"learning_rate": 1.5405571821015402e-06,
"loss": 1.2634363174438477,
"step": 876
},
{
"epoch": 2.5605839416058394,
"grad_norm": 0.376953125,
"learning_rate": 1.5268868032215152e-06,
"loss": 1.1621180772781372,
"step": 878
},
{
"epoch": 2.5664233576642337,
"grad_norm": 0.4765625,
"learning_rate": 1.513380747983169e-06,
"loss": 1.1664128303527832,
"step": 880
},
{
"epoch": 2.5722627737226276,
"grad_norm": 0.609375,
"learning_rate": 1.5000395749822597e-06,
"loss": 1.1880872249603271,
"step": 882
},
{
"epoch": 2.578102189781022,
"grad_norm": 0.4921875,
"learning_rate": 1.4868638359951963e-06,
"loss": 1.2146211862564087,
"step": 884
},
{
"epoch": 2.5839416058394162,
"grad_norm": 0.55859375,
"learning_rate": 1.4738540759562175e-06,
"loss": 1.1511411666870117,
"step": 886
},
{
"epoch": 2.58978102189781,
"grad_norm": 0.515625,
"learning_rate": 1.4610108329348514e-06,
"loss": 1.2108086347579956,
"step": 888
},
{
"epoch": 2.5956204379562045,
"grad_norm": 0.400390625,
"learning_rate": 1.4483346381136653e-06,
"loss": 1.0855056047439575,
"step": 890
},
{
"epoch": 2.601459854014599,
"grad_norm": 0.3671875,
"learning_rate": 1.4358260157662928e-06,
"loss": 1.1662517786026,
"step": 892
},
{
"epoch": 2.6072992700729927,
"grad_norm": 0.79296875,
"learning_rate": 1.4234854832357536e-06,
"loss": 1.195744514465332,
"step": 894
},
{
"epoch": 2.613138686131387,
"grad_norm": 0.53515625,
"learning_rate": 1.4113135509130547e-06,
"loss": 1.2275428771972656,
"step": 896
},
{
"epoch": 2.618978102189781,
"grad_norm": 0.50390625,
"learning_rate": 1.3993107222160817e-06,
"loss": 1.2926044464111328,
"step": 898
},
{
"epoch": 2.624817518248175,
"grad_norm": 0.60546875,
"learning_rate": 1.3874774935687782e-06,
"loss": 1.2193105220794678,
"step": 900
},
{
"epoch": 2.630656934306569,
"grad_norm": 0.5859375,
"learning_rate": 1.3758143543806143e-06,
"loss": 1.227401852607727,
"step": 902
},
{
"epoch": 2.6364963503649634,
"grad_norm": 0.86328125,
"learning_rate": 1.3643217870263448e-06,
"loss": 1.2654823064804077,
"step": 904
},
{
"epoch": 2.6423357664233578,
"grad_norm": 0.373046875,
"learning_rate": 1.35300026682606e-06,
"loss": 1.1640398502349854,
"step": 906
},
{
"epoch": 2.6481751824817517,
"grad_norm": 0.625,
"learning_rate": 1.341850262025524e-06,
"loss": 1.1758366823196411,
"step": 908
},
{
"epoch": 2.654014598540146,
"grad_norm": 0.462890625,
"learning_rate": 1.330872233776811e-06,
"loss": 1.1808586120605469,
"step": 910
},
{
"epoch": 2.6598540145985403,
"grad_norm": 0.458984375,
"learning_rate": 1.3200666361192349e-06,
"loss": 1.1935012340545654,
"step": 912
},
{
"epoch": 2.665693430656934,
"grad_norm": 0.396484375,
"learning_rate": 1.3094339159605627e-06,
"loss": 1.1293021440505981,
"step": 914
},
{
"epoch": 2.6715328467153285,
"grad_norm": 0.875,
"learning_rate": 1.2989745130585407e-06,
"loss": 1.2715429067611694,
"step": 916
},
{
"epoch": 2.677372262773723,
"grad_norm": 0.462890625,
"learning_rate": 1.2886888600026983e-06,
"loss": 1.1346051692962646,
"step": 918
},
{
"epoch": 2.6832116788321168,
"grad_norm": 0.412109375,
"learning_rate": 1.2785773821964636e-06,
"loss": 1.2460830211639404,
"step": 920
},
{
"epoch": 2.689051094890511,
"grad_norm": 1.6171875,
"learning_rate": 1.2686404978395626e-06,
"loss": 1.1654274463653564,
"step": 922
},
{
"epoch": 2.694890510948905,
"grad_norm": 1.65625,
"learning_rate": 1.2588786179107279e-06,
"loss": 1.253066897392273,
"step": 924
},
{
"epoch": 2.7007299270072993,
"grad_norm": 0.42578125,
"learning_rate": 1.2492921461506994e-06,
"loss": 1.188564658164978,
"step": 926
},
{
"epoch": 2.706569343065693,
"grad_norm": 0.466796875,
"learning_rate": 1.2398814790455244e-06,
"loss": 1.244341254234314,
"step": 928
},
{
"epoch": 2.7124087591240875,
"grad_norm": 0.443359375,
"learning_rate": 1.2306470058101612e-06,
"loss": 1.1333363056182861,
"step": 930
},
{
"epoch": 2.718248175182482,
"grad_norm": 0.66015625,
"learning_rate": 1.2215891083723842e-06,
"loss": 1.263460397720337,
"step": 932
},
{
"epoch": 2.7240875912408757,
"grad_norm": 0.359375,
"learning_rate": 1.2127081613569809e-06,
"loss": 1.2082524299621582,
"step": 934
},
{
"epoch": 2.72992700729927,
"grad_norm": 0.408203125,
"learning_rate": 1.2040045320702626e-06,
"loss": 1.1242833137512207,
"step": 936
},
{
"epoch": 2.7357664233576644,
"grad_norm": 0.63671875,
"learning_rate": 1.1954785804848739e-06,
"loss": 1.14191734790802,
"step": 938
},
{
"epoch": 2.7416058394160583,
"grad_norm": 1.65625,
"learning_rate": 1.1871306592249008e-06,
"loss": 1.1483235359191895,
"step": 940
},
{
"epoch": 2.7474452554744526,
"grad_norm": 0.5546875,
"learning_rate": 1.1789611135512903e-06,
"loss": 1.1981879472732544,
"step": 942
},
{
"epoch": 2.753284671532847,
"grad_norm": 1.171875,
"learning_rate": 1.1709702813475674e-06,
"loss": 1.1760236024856567,
"step": 944
},
{
"epoch": 2.759124087591241,
"grad_norm": 0.380859375,
"learning_rate": 1.163158493105864e-06,
"loss": 1.2009891271591187,
"step": 946
},
{
"epoch": 2.764963503649635,
"grad_norm": 0.65234375,
"learning_rate": 1.1555260719132474e-06,
"loss": 1.1685845851898193,
"step": 948
},
{
"epoch": 2.770802919708029,
"grad_norm": 0.423828125,
"learning_rate": 1.1480733334383599e-06,
"loss": 1.2095947265625,
"step": 950
},
{
"epoch": 2.7766423357664234,
"grad_norm": 0.357421875,
"learning_rate": 1.1408005859183595e-06,
"loss": 1.1101715564727783,
"step": 952
},
{
"epoch": 2.7824817518248173,
"grad_norm": 0.69921875,
"learning_rate": 1.1337081301461775e-06,
"loss": 1.2515991926193237,
"step": 954
},
{
"epoch": 2.7883211678832116,
"grad_norm": 1.9609375,
"learning_rate": 1.1267962594580712e-06,
"loss": 1.2288193702697754,
"step": 956
},
{
"epoch": 2.794160583941606,
"grad_norm": 0.90234375,
"learning_rate": 1.1200652597214982e-06,
"loss": 1.1902008056640625,
"step": 958
},
{
"epoch": 2.8,
"grad_norm": 0.458984375,
"learning_rate": 1.1135154093232868e-06,
"loss": 1.1392861604690552,
"step": 960
},
{
"epoch": 2.805839416058394,
"grad_norm": 0.625,
"learning_rate": 1.107146979158129e-06,
"loss": 1.1379271745681763,
"step": 962
},
{
"epoch": 2.8116788321167885,
"grad_norm": 0.57421875,
"learning_rate": 1.100960232617371e-06,
"loss": 1.274213433265686,
"step": 964
},
{
"epoch": 2.8175182481751824,
"grad_norm": 0.984375,
"learning_rate": 1.094955425578123e-06,
"loss": 1.2406063079833984,
"step": 966
},
{
"epoch": 2.8233576642335767,
"grad_norm": 1.0859375,
"learning_rate": 1.0891328063926742e-06,
"loss": 1.1469717025756836,
"step": 968
},
{
"epoch": 2.829197080291971,
"grad_norm": 0.3046875,
"learning_rate": 1.0834926158782235e-06,
"loss": 1.2098904848098755,
"step": 970
},
{
"epoch": 2.835036496350365,
"grad_norm": 0.640625,
"learning_rate": 1.0780350873069184e-06,
"loss": 1.1746724843978882,
"step": 972
},
{
"epoch": 2.8408759124087593,
"grad_norm": 0.37109375,
"learning_rate": 1.0727604463962058e-06,
"loss": 1.187951683998108,
"step": 974
},
{
"epoch": 2.846715328467153,
"grad_norm": 0.546875,
"learning_rate": 1.0676689112994977e-06,
"loss": 1.1875951290130615,
"step": 976
},
{
"epoch": 2.8525547445255475,
"grad_norm": 0.50390625,
"learning_rate": 1.06276069259715e-06,
"loss": 1.1863131523132324,
"step": 978
},
{
"epoch": 2.8583941605839414,
"grad_norm": 0.3984375,
"learning_rate": 1.0580359932877516e-06,
"loss": 1.1864418983459473,
"step": 980
},
{
"epoch": 2.8642335766423357,
"grad_norm": 0.66796875,
"learning_rate": 1.0534950087797282e-06,
"loss": 1.255091667175293,
"step": 982
},
{
"epoch": 2.87007299270073,
"grad_norm": 0.357421875,
"learning_rate": 1.049137926883261e-06,
"loss": 1.1269922256469727,
"step": 984
},
{
"epoch": 2.875912408759124,
"grad_norm": 0.64453125,
"learning_rate": 1.0449649278025208e-06,
"loss": 1.1925872564315796,
"step": 986
},
{
"epoch": 2.8817518248175182,
"grad_norm": 0.3828125,
"learning_rate": 1.0409761841282112e-06,
"loss": 1.1946667432785034,
"step": 988
},
{
"epoch": 2.8875912408759126,
"grad_norm": 0.408203125,
"learning_rate": 1.037171860830434e-06,
"loss": 1.141740083694458,
"step": 990
},
{
"epoch": 2.8934306569343065,
"grad_norm": 0.4921875,
"learning_rate": 1.0335521152518637e-06,
"loss": 1.0821702480316162,
"step": 992
},
{
"epoch": 2.899270072992701,
"grad_norm": 0.453125,
"learning_rate": 1.0301170971012418e-06,
"loss": 1.1304783821105957,
"step": 994
},
{
"epoch": 2.905109489051095,
"grad_norm": 0.84765625,
"learning_rate": 1.0268669484471839e-06,
"loss": 1.1460580825805664,
"step": 996
},
{
"epoch": 2.910948905109489,
"grad_norm": 0.58984375,
"learning_rate": 1.0238018037123042e-06,
"loss": 1.216837763786316,
"step": 998
},
{
"epoch": 2.9167883211678833,
"grad_norm": 0.46484375,
"learning_rate": 1.020921789667656e-06,
"loss": 1.217172622680664,
"step": 1000
},
{
"epoch": 2.9226277372262772,
"grad_norm": 0.400390625,
"learning_rate": 1.0182270254274888e-06,
"loss": 1.2125324010849,
"step": 1002
},
{
"epoch": 2.9284671532846716,
"grad_norm": 0.83984375,
"learning_rate": 1.015717622444321e-06,
"loss": 1.1410317420959473,
"step": 1004
},
{
"epoch": 2.9343065693430654,
"grad_norm": 0.578125,
"learning_rate": 1.0133936845043322e-06,
"loss": 1.2082126140594482,
"step": 1006
},
{
"epoch": 2.9401459854014598,
"grad_norm": 0.6484375,
"learning_rate": 1.011255307723068e-06,
"loss": 1.2457588911056519,
"step": 1008
},
{
"epoch": 2.945985401459854,
"grad_norm": 1.140625,
"learning_rate": 1.0093025805414676e-06,
"loss": 1.1455327272415161,
"step": 1010
},
{
"epoch": 2.951824817518248,
"grad_norm": 0.5546875,
"learning_rate": 1.007535583722203e-06,
"loss": 1.1338199377059937,
"step": 1012
},
{
"epoch": 2.9576642335766423,
"grad_norm": 0.419921875,
"learning_rate": 1.0059543903463426e-06,
"loss": 1.2006813287734985,
"step": 1014
},
{
"epoch": 2.9635036496350367,
"grad_norm": 0.392578125,
"learning_rate": 1.004559065810324e-06,
"loss": 1.2222992181777954,
"step": 1016
},
{
"epoch": 2.9693430656934305,
"grad_norm": 0.349609375,
"learning_rate": 1.003349667823253e-06,
"loss": 1.1396572589874268,
"step": 1018
},
{
"epoch": 2.975182481751825,
"grad_norm": 0.4453125,
"learning_rate": 1.0023262464045155e-06,
"loss": 1.2639334201812744,
"step": 1020
},
{
"epoch": 2.981021897810219,
"grad_norm": 0.640625,
"learning_rate": 1.0014888438817083e-06,
"loss": 1.206993579864502,
"step": 1022
},
{
"epoch": 2.986861313868613,
"grad_norm": 0.58203125,
"learning_rate": 1.0008374948888896e-06,
"loss": 1.1673427820205688,
"step": 1024
},
{
"epoch": 2.9927007299270074,
"grad_norm": 0.39453125,
"learning_rate": 1.0003722263651458e-06,
"loss": 1.2105653285980225,
"step": 1026
},
{
"epoch": 2.9985401459854013,
"grad_norm": 0.78125,
"learning_rate": 1.0000930575534762e-06,
"loss": 1.1724438667297363,
"step": 1028
},
{
"epoch": 3.0,
"step": 1029,
"total_flos": 2.909158325117518e+18,
"train_loss": 1.307906815331462,
"train_runtime": 13215.9511,
"train_samples_per_second": 2.488,
"train_steps_per_second": 0.078
}
],
"logging_steps": 2,
"max_steps": 1029,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 9999999,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 2.909158325117518e+18,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}