Qwen3_tiny_ep3 / trainer_state.json
DecoderImmortal's picture
Upload folder using huggingface_hub
e16424d verified
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 2.0,
"eval_steps": 200.0,
"global_step": 10410,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.00019212295869356388,
"grad_norm": 0.08390819281339645,
"learning_rate": 5.758157389635316e-07,
"loss": 3.274601936340332,
"step": 1
},
{
"epoch": 0.0009606147934678194,
"grad_norm": 0.08173709362745285,
"learning_rate": 2.879078694817658e-06,
"loss": 3.2788190841674805,
"step": 5
},
{
"epoch": 0.0019212295869356388,
"grad_norm": 0.09356535971164703,
"learning_rate": 5.758157389635316e-06,
"loss": 3.2803062438964843,
"step": 10
},
{
"epoch": 0.002881844380403458,
"grad_norm": 0.11097322404384613,
"learning_rate": 8.637236084452974e-06,
"loss": 3.277487564086914,
"step": 15
},
{
"epoch": 0.0038424591738712775,
"grad_norm": 0.15006589889526367,
"learning_rate": 1.1516314779270632e-05,
"loss": 3.281480407714844,
"step": 20
},
{
"epoch": 0.004803073967339097,
"grad_norm": 0.14452151954174042,
"learning_rate": 1.439539347408829e-05,
"loss": 3.2803359985351563,
"step": 25
},
{
"epoch": 0.005763688760806916,
"grad_norm": 0.16946451365947723,
"learning_rate": 1.7274472168905948e-05,
"loss": 3.287946319580078,
"step": 30
},
{
"epoch": 0.0067243035542747355,
"grad_norm": 0.14511597156524658,
"learning_rate": 2.015355086372361e-05,
"loss": 3.278676986694336,
"step": 35
},
{
"epoch": 0.007684918347742555,
"grad_norm": 0.21068508923053741,
"learning_rate": 2.3032629558541264e-05,
"loss": 3.2816986083984374,
"step": 40
},
{
"epoch": 0.008645533141210375,
"grad_norm": 0.2152123898267746,
"learning_rate": 2.591170825335892e-05,
"loss": 3.2830833435058593,
"step": 45
},
{
"epoch": 0.009606147934678195,
"grad_norm": 0.2621569037437439,
"learning_rate": 2.879078694817658e-05,
"loss": 3.279366302490234,
"step": 50
},
{
"epoch": 0.010566762728146013,
"grad_norm": 0.29736942052841187,
"learning_rate": 3.166986564299424e-05,
"loss": 3.2836788177490233,
"step": 55
},
{
"epoch": 0.011527377521613832,
"grad_norm": 0.31358596682548523,
"learning_rate": 3.4548944337811895e-05,
"loss": 3.2797149658203124,
"step": 60
},
{
"epoch": 0.012487992315081652,
"grad_norm": 0.42882370948791504,
"learning_rate": 3.742802303262955e-05,
"loss": 3.2872390747070312,
"step": 65
},
{
"epoch": 0.013448607108549471,
"grad_norm": 0.3247455060482025,
"learning_rate": 4.030710172744722e-05,
"loss": 3.2825721740722655,
"step": 70
},
{
"epoch": 0.01440922190201729,
"grad_norm": 0.40256041288375854,
"learning_rate": 4.318618042226487e-05,
"loss": 3.279859924316406,
"step": 75
},
{
"epoch": 0.01536983669548511,
"grad_norm": 0.5199999809265137,
"learning_rate": 4.606525911708253e-05,
"loss": 3.284808349609375,
"step": 80
},
{
"epoch": 0.01633045148895293,
"grad_norm": 0.6197025179862976,
"learning_rate": 4.894433781190019e-05,
"loss": 3.2813262939453125,
"step": 85
},
{
"epoch": 0.01729106628242075,
"grad_norm": 0.4358411133289337,
"learning_rate": 5.182341650671784e-05,
"loss": 3.2836116790771483,
"step": 90
},
{
"epoch": 0.01825168107588857,
"grad_norm": 0.5619639158248901,
"learning_rate": 5.4702495201535504e-05,
"loss": 3.281964874267578,
"step": 95
},
{
"epoch": 0.01921229586935639,
"grad_norm": 0.42559099197387695,
"learning_rate": 5.758157389635316e-05,
"loss": 3.284071350097656,
"step": 100
},
{
"epoch": 0.020172910662824207,
"grad_norm": 0.5353518128395081,
"learning_rate": 6.0460652591170814e-05,
"loss": 3.286463165283203,
"step": 105
},
{
"epoch": 0.021133525456292025,
"grad_norm": 0.8242971897125244,
"learning_rate": 6.333973128598848e-05,
"loss": 3.28314094543457,
"step": 110
},
{
"epoch": 0.022094140249759846,
"grad_norm": 0.5128299593925476,
"learning_rate": 6.621880998080614e-05,
"loss": 3.2824722290039063,
"step": 115
},
{
"epoch": 0.023054755043227664,
"grad_norm": 0.7521133422851562,
"learning_rate": 6.909788867562379e-05,
"loss": 3.2825164794921875,
"step": 120
},
{
"epoch": 0.024015369836695485,
"grad_norm": 0.7099725008010864,
"learning_rate": 7.197696737044145e-05,
"loss": 3.2856216430664062,
"step": 125
},
{
"epoch": 0.024975984630163303,
"grad_norm": 0.673387885093689,
"learning_rate": 7.48560460652591e-05,
"loss": 3.2845367431640624,
"step": 130
},
{
"epoch": 0.025936599423631124,
"grad_norm": 0.47837865352630615,
"learning_rate": 7.773512476007677e-05,
"loss": 3.2803184509277346,
"step": 135
},
{
"epoch": 0.026897214217098942,
"grad_norm": 0.5746335983276367,
"learning_rate": 8.061420345489444e-05,
"loss": 3.2853370666503907,
"step": 140
},
{
"epoch": 0.027857829010566763,
"grad_norm": 0.9516947269439697,
"learning_rate": 8.349328214971209e-05,
"loss": 3.2811763763427733,
"step": 145
},
{
"epoch": 0.02881844380403458,
"grad_norm": 0.7206563949584961,
"learning_rate": 8.637236084452975e-05,
"loss": 3.2858177185058595,
"step": 150
},
{
"epoch": 0.029779058597502402,
"grad_norm": 0.9468950033187866,
"learning_rate": 8.92514395393474e-05,
"loss": 3.2817359924316407,
"step": 155
},
{
"epoch": 0.03073967339097022,
"grad_norm": 0.9247293472290039,
"learning_rate": 9.213051823416505e-05,
"loss": 3.283827209472656,
"step": 160
},
{
"epoch": 0.03170028818443804,
"grad_norm": 0.7639815807342529,
"learning_rate": 9.500959692898272e-05,
"loss": 3.2805789947509765,
"step": 165
},
{
"epoch": 0.03266090297790586,
"grad_norm": 1.0864217281341553,
"learning_rate": 9.788867562380038e-05,
"loss": 3.282984161376953,
"step": 170
},
{
"epoch": 0.03362151777137368,
"grad_norm": 1.0888683795928955,
"learning_rate": 0.00010076775431861803,
"loss": 3.2868572235107423,
"step": 175
},
{
"epoch": 0.0345821325648415,
"grad_norm": 0.7901690602302551,
"learning_rate": 0.00010364683301343569,
"loss": 3.2854034423828127,
"step": 180
},
{
"epoch": 0.03554274735830932,
"grad_norm": 1.132866621017456,
"learning_rate": 0.00010652591170825334,
"loss": 3.283061218261719,
"step": 185
},
{
"epoch": 0.03650336215177714,
"grad_norm": 1.1301084756851196,
"learning_rate": 0.00010940499040307101,
"loss": 3.2865623474121093,
"step": 190
},
{
"epoch": 0.037463976945244955,
"grad_norm": 0.9956747889518738,
"learning_rate": 0.00011228406909788866,
"loss": 3.2860641479492188,
"step": 195
},
{
"epoch": 0.03842459173871278,
"grad_norm": 1.1139202117919922,
"learning_rate": 0.00011516314779270632,
"loss": 3.2860763549804686,
"step": 200
},
{
"epoch": 0.0393852065321806,
"grad_norm": 1.1494112014770508,
"learning_rate": 0.00011804222648752397,
"loss": 3.2874046325683595,
"step": 205
},
{
"epoch": 0.040345821325648415,
"grad_norm": 1.1791850328445435,
"learning_rate": 0.00012092130518234163,
"loss": 3.28511962890625,
"step": 210
},
{
"epoch": 0.04130643611911623,
"grad_norm": 0.8412664532661438,
"learning_rate": 0.0001238003838771593,
"loss": 3.2826396942138674,
"step": 215
},
{
"epoch": 0.04226705091258405,
"grad_norm": 1.0537320375442505,
"learning_rate": 0.00012667946257197696,
"loss": 3.2859474182128907,
"step": 220
},
{
"epoch": 0.043227665706051875,
"grad_norm": 0.9535348415374756,
"learning_rate": 0.00012955854126679462,
"loss": 3.2872699737548827,
"step": 225
},
{
"epoch": 0.04418828049951969,
"grad_norm": 1.2287579774856567,
"learning_rate": 0.00013243761996161227,
"loss": 3.2854686737060548,
"step": 230
},
{
"epoch": 0.04514889529298751,
"grad_norm": 0.8720296621322632,
"learning_rate": 0.00013531669865642993,
"loss": 3.2873931884765626,
"step": 235
},
{
"epoch": 0.04610951008645533,
"grad_norm": 0.9984288811683655,
"learning_rate": 0.00013819577735124758,
"loss": 3.28588752746582,
"step": 240
},
{
"epoch": 0.04707012487992315,
"grad_norm": 1.3232122659683228,
"learning_rate": 0.00014107485604606524,
"loss": 3.2899200439453127,
"step": 245
},
{
"epoch": 0.04803073967339097,
"grad_norm": 1.0983723402023315,
"learning_rate": 0.0001439539347408829,
"loss": 3.287739562988281,
"step": 250
},
{
"epoch": 0.04899135446685879,
"grad_norm": 0.9880014657974243,
"learning_rate": 0.00014683301343570055,
"loss": 3.2914413452148437,
"step": 255
},
{
"epoch": 0.049951969260326606,
"grad_norm": 1.0988938808441162,
"learning_rate": 0.0001497120921305182,
"loss": 3.2892181396484377,
"step": 260
},
{
"epoch": 0.05091258405379443,
"grad_norm": 0.9497994780540466,
"learning_rate": 0.00015259117082533588,
"loss": 3.289907455444336,
"step": 265
},
{
"epoch": 0.05187319884726225,
"grad_norm": 1.4786131381988525,
"learning_rate": 0.00015547024952015354,
"loss": 3.2885669708251952,
"step": 270
},
{
"epoch": 0.052833813640730067,
"grad_norm": 1.1913130283355713,
"learning_rate": 0.0001583493282149712,
"loss": 3.2930023193359377,
"step": 275
},
{
"epoch": 0.053794428434197884,
"grad_norm": 1.0576707124710083,
"learning_rate": 0.00016122840690978887,
"loss": 3.285129165649414,
"step": 280
},
{
"epoch": 0.05475504322766571,
"grad_norm": 1.3840175867080688,
"learning_rate": 0.0001641074856046065,
"loss": 3.2881790161132813,
"step": 285
},
{
"epoch": 0.05571565802113353,
"grad_norm": 1.1912461519241333,
"learning_rate": 0.00016698656429942418,
"loss": 3.2907310485839845,
"step": 290
},
{
"epoch": 0.056676272814601344,
"grad_norm": 0.8021464347839355,
"learning_rate": 0.0001698656429942418,
"loss": 3.2897079467773436,
"step": 295
},
{
"epoch": 0.05763688760806916,
"grad_norm": 1.2219209671020508,
"learning_rate": 0.0001727447216890595,
"loss": 3.288962554931641,
"step": 300
},
{
"epoch": 0.05859750240153699,
"grad_norm": 1.13152015209198,
"learning_rate": 0.00017562380038387714,
"loss": 3.2893699645996093,
"step": 305
},
{
"epoch": 0.059558117195004805,
"grad_norm": 1.5439852476119995,
"learning_rate": 0.0001785028790786948,
"loss": 3.293457794189453,
"step": 310
},
{
"epoch": 0.06051873198847262,
"grad_norm": 1.4060869216918945,
"learning_rate": 0.00018138195777351245,
"loss": 3.2903827667236327,
"step": 315
},
{
"epoch": 0.06147934678194044,
"grad_norm": 1.0909712314605713,
"learning_rate": 0.0001842610364683301,
"loss": 3.293254852294922,
"step": 320
},
{
"epoch": 0.06243996157540826,
"grad_norm": 1.4758312702178955,
"learning_rate": 0.00018714011516314776,
"loss": 3.291452407836914,
"step": 325
},
{
"epoch": 0.06340057636887608,
"grad_norm": 1.3612134456634521,
"learning_rate": 0.00019001919385796544,
"loss": 3.2947055816650392,
"step": 330
},
{
"epoch": 0.0643611911623439,
"grad_norm": 1.017069935798645,
"learning_rate": 0.00019289827255278307,
"loss": 3.2914665222167967,
"step": 335
},
{
"epoch": 0.06532180595581172,
"grad_norm": 1.114537239074707,
"learning_rate": 0.00019577735124760075,
"loss": 3.2918880462646483,
"step": 340
},
{
"epoch": 0.06628242074927954,
"grad_norm": 1.359885334968567,
"learning_rate": 0.00019865642994241838,
"loss": 3.291158676147461,
"step": 345
},
{
"epoch": 0.06724303554274735,
"grad_norm": 1.5216479301452637,
"learning_rate": 0.00020153550863723606,
"loss": 3.2951950073242187,
"step": 350
},
{
"epoch": 0.06820365033621517,
"grad_norm": 1.0338709354400635,
"learning_rate": 0.00020441458733205374,
"loss": 3.2910877227783204,
"step": 355
},
{
"epoch": 0.069164265129683,
"grad_norm": 1.3676224946975708,
"learning_rate": 0.00020729366602687137,
"loss": 3.2979248046875,
"step": 360
},
{
"epoch": 0.07012487992315082,
"grad_norm": 1.508606195449829,
"learning_rate": 0.00021017274472168905,
"loss": 3.2947929382324217,
"step": 365
},
{
"epoch": 0.07108549471661864,
"grad_norm": 1.4694204330444336,
"learning_rate": 0.00021305182341650668,
"loss": 3.290261077880859,
"step": 370
},
{
"epoch": 0.07204610951008646,
"grad_norm": 1.629552960395813,
"learning_rate": 0.00021593090211132436,
"loss": 3.292864990234375,
"step": 375
},
{
"epoch": 0.07300672430355427,
"grad_norm": 1.3059035539627075,
"learning_rate": 0.00021880998080614202,
"loss": 3.297530746459961,
"step": 380
},
{
"epoch": 0.07396733909702209,
"grad_norm": 1.6734189987182617,
"learning_rate": 0.00022168905950095967,
"loss": 3.2937080383300783,
"step": 385
},
{
"epoch": 0.07492795389048991,
"grad_norm": 1.9466580152511597,
"learning_rate": 0.00022456813819577733,
"loss": 3.2989707946777345,
"step": 390
},
{
"epoch": 0.07588856868395773,
"grad_norm": 1.4589389562606812,
"learning_rate": 0.00022744721689059498,
"loss": 3.2969600677490236,
"step": 395
},
{
"epoch": 0.07684918347742556,
"grad_norm": 1.273814082145691,
"learning_rate": 0.00023032629558541264,
"loss": 3.2937450408935547,
"step": 400
},
{
"epoch": 0.07780979827089338,
"grad_norm": 1.6920170783996582,
"learning_rate": 0.00023320537428023032,
"loss": 3.2936920166015624,
"step": 405
},
{
"epoch": 0.0787704130643612,
"grad_norm": 1.7424761056900024,
"learning_rate": 0.00023608445297504794,
"loss": 3.29752197265625,
"step": 410
},
{
"epoch": 0.07973102785782901,
"grad_norm": 1.4278359413146973,
"learning_rate": 0.00023896353166986563,
"loss": 3.2931358337402346,
"step": 415
},
{
"epoch": 0.08069164265129683,
"grad_norm": 1.3596338033676147,
"learning_rate": 0.00024184261036468325,
"loss": 3.2951515197753904,
"step": 420
},
{
"epoch": 0.08165225744476465,
"grad_norm": 1.4509109258651733,
"learning_rate": 0.00024472168905950096,
"loss": 3.2955284118652344,
"step": 425
},
{
"epoch": 0.08261287223823247,
"grad_norm": 2.0564208030700684,
"learning_rate": 0.0002476007677543186,
"loss": 3.2958652496337892,
"step": 430
},
{
"epoch": 0.08357348703170028,
"grad_norm": 1.80084228515625,
"learning_rate": 0.00025047984644913627,
"loss": 3.299582672119141,
"step": 435
},
{
"epoch": 0.0845341018251681,
"grad_norm": 1.2458295822143555,
"learning_rate": 0.0002533589251439539,
"loss": 3.2957221984863283,
"step": 440
},
{
"epoch": 0.08549471661863593,
"grad_norm": 1.5943655967712402,
"learning_rate": 0.0002562380038387716,
"loss": 3.296315002441406,
"step": 445
},
{
"epoch": 0.08645533141210375,
"grad_norm": 1.4959216117858887,
"learning_rate": 0.00025911708253358924,
"loss": 3.2936065673828123,
"step": 450
},
{
"epoch": 0.08741594620557157,
"grad_norm": 0.9593575596809387,
"learning_rate": 0.0002619961612284069,
"loss": 3.2923255920410157,
"step": 455
},
{
"epoch": 0.08837656099903939,
"grad_norm": 1.7525019645690918,
"learning_rate": 0.00026487523992322454,
"loss": 3.297338104248047,
"step": 460
},
{
"epoch": 0.0893371757925072,
"grad_norm": 1.5253691673278809,
"learning_rate": 0.0002677543186180422,
"loss": 3.29698486328125,
"step": 465
},
{
"epoch": 0.09029779058597502,
"grad_norm": 1.7394542694091797,
"learning_rate": 0.00027063339731285985,
"loss": 3.2963401794433596,
"step": 470
},
{
"epoch": 0.09125840537944284,
"grad_norm": 1.3029968738555908,
"learning_rate": 0.0002735124760076775,
"loss": 3.2992328643798827,
"step": 475
},
{
"epoch": 0.09221902017291066,
"grad_norm": 1.3961071968078613,
"learning_rate": 0.00027639155470249516,
"loss": 3.3040702819824217,
"step": 480
},
{
"epoch": 0.09317963496637849,
"grad_norm": 0.9128854274749756,
"learning_rate": 0.0002792706333973128,
"loss": 3.2994720458984377,
"step": 485
},
{
"epoch": 0.0941402497598463,
"grad_norm": 1.39821195602417,
"learning_rate": 0.00028214971209213047,
"loss": 3.3009670257568358,
"step": 490
},
{
"epoch": 0.09510086455331412,
"grad_norm": 1.3767975568771362,
"learning_rate": 0.0002850287907869481,
"loss": 3.298174285888672,
"step": 495
},
{
"epoch": 0.09606147934678194,
"grad_norm": 1.5919218063354492,
"learning_rate": 0.0002879078694817658,
"loss": 3.2930580139160157,
"step": 500
},
{
"epoch": 0.09702209414024976,
"grad_norm": 2.060764789581299,
"learning_rate": 0.0002907869481765835,
"loss": 3.3007606506347655,
"step": 505
},
{
"epoch": 0.09798270893371758,
"grad_norm": 1.864414095878601,
"learning_rate": 0.0002936660268714011,
"loss": 3.302916717529297,
"step": 510
},
{
"epoch": 0.0989433237271854,
"grad_norm": 1.410025954246521,
"learning_rate": 0.0002965451055662188,
"loss": 3.297088623046875,
"step": 515
},
{
"epoch": 0.09990393852065321,
"grad_norm": 1.4724565744400024,
"learning_rate": 0.0002994241842610364,
"loss": 3.2969940185546873,
"step": 520
},
{
"epoch": 0.10086455331412104,
"grad_norm": 1.1308706998825073,
"learning_rate": 0.0002999998788910666,
"loss": 3.295259475708008,
"step": 525
},
{
"epoch": 0.10182516810758886,
"grad_norm": 2.0092132091522217,
"learning_rate": 0.00029999938688636,
"loss": 3.3064605712890627,
"step": 530
},
{
"epoch": 0.10278578290105668,
"grad_norm": 1.5756242275238037,
"learning_rate": 0.0002999985164178122,
"loss": 3.302138900756836,
"step": 535
},
{
"epoch": 0.1037463976945245,
"grad_norm": 1.6854816675186157,
"learning_rate": 0.00029999726748761955,
"loss": 3.29990234375,
"step": 540
},
{
"epoch": 0.10470701248799232,
"grad_norm": 1.5925079584121704,
"learning_rate": 0.0002999956400989331,
"loss": 3.3017444610595703,
"step": 545
},
{
"epoch": 0.10566762728146013,
"grad_norm": 0.9805465340614319,
"learning_rate": 0.00029999363425585907,
"loss": 3.2947647094726564,
"step": 550
},
{
"epoch": 0.10662824207492795,
"grad_norm": 1.143500804901123,
"learning_rate": 0.0002999912499634584,
"loss": 3.3032691955566404,
"step": 555
},
{
"epoch": 0.10758885686839577,
"grad_norm": 1.57884681224823,
"learning_rate": 0.00029998848722774676,
"loss": 3.3033897399902346,
"step": 560
},
{
"epoch": 0.10854947166186359,
"grad_norm": 1.9447771310806274,
"learning_rate": 0.00029998534605569496,
"loss": 3.2998809814453125,
"step": 565
},
{
"epoch": 0.10951008645533142,
"grad_norm": 1.4967366456985474,
"learning_rate": 0.0002999818264552284,
"loss": 3.2998199462890625,
"step": 570
},
{
"epoch": 0.11047070124879924,
"grad_norm": 1.5965176820755005,
"learning_rate": 0.0002999779284352275,
"loss": 3.2984855651855467,
"step": 575
},
{
"epoch": 0.11143131604226705,
"grad_norm": 1.3109067678451538,
"learning_rate": 0.0002999736520055273,
"loss": 3.293305206298828,
"step": 580
},
{
"epoch": 0.11239193083573487,
"grad_norm": 1.374250054359436,
"learning_rate": 0.00029996899717691766,
"loss": 3.2990135192871093,
"step": 585
},
{
"epoch": 0.11335254562920269,
"grad_norm": 1.8241665363311768,
"learning_rate": 0.0002999639639611432,
"loss": 3.2949737548828124,
"step": 590
},
{
"epoch": 0.1143131604226705,
"grad_norm": 1.1245280504226685,
"learning_rate": 0.00029995855237090327,
"loss": 3.3007225036621093,
"step": 595
},
{
"epoch": 0.11527377521613832,
"grad_norm": 1.290839433670044,
"learning_rate": 0.0002999527624198518,
"loss": 3.2996570587158205,
"step": 600
},
{
"epoch": 0.11623439000960614,
"grad_norm": 1.100143551826477,
"learning_rate": 0.00029994659412259746,
"loss": 3.2947521209716797,
"step": 605
},
{
"epoch": 0.11719500480307397,
"grad_norm": 1.6502522230148315,
"learning_rate": 0.0002999400474947036,
"loss": 3.295918273925781,
"step": 610
},
{
"epoch": 0.11815561959654179,
"grad_norm": 1.4078381061553955,
"learning_rate": 0.0002999331225526879,
"loss": 3.299440383911133,
"step": 615
},
{
"epoch": 0.11911623439000961,
"grad_norm": 1.4280340671539307,
"learning_rate": 0.0002999258193140227,
"loss": 3.296113967895508,
"step": 620
},
{
"epoch": 0.12007684918347743,
"grad_norm": 1.210864543914795,
"learning_rate": 0.00029991813779713485,
"loss": 3.292288970947266,
"step": 625
},
{
"epoch": 0.12103746397694524,
"grad_norm": 1.7756609916687012,
"learning_rate": 0.0002999100780214057,
"loss": 3.291936492919922,
"step": 630
},
{
"epoch": 0.12199807877041306,
"grad_norm": 2.3280141353607178,
"learning_rate": 0.0002999016400071708,
"loss": 3.2992881774902343,
"step": 635
},
{
"epoch": 0.12295869356388088,
"grad_norm": 1.2652477025985718,
"learning_rate": 0.00029989282377572006,
"loss": 3.302785110473633,
"step": 640
},
{
"epoch": 0.1239193083573487,
"grad_norm": 1.1368358135223389,
"learning_rate": 0.00029988362934929786,
"loss": 3.298853302001953,
"step": 645
},
{
"epoch": 0.12487992315081652,
"grad_norm": 1.5556303262710571,
"learning_rate": 0.0002998740567511026,
"loss": 3.2962833404541017,
"step": 650
},
{
"epoch": 0.12584053794428435,
"grad_norm": 1.2470626831054688,
"learning_rate": 0.000299864106005287,
"loss": 3.296977996826172,
"step": 655
},
{
"epoch": 0.12680115273775217,
"grad_norm": 1.8531115055084229,
"learning_rate": 0.00029985377713695775,
"loss": 3.294302749633789,
"step": 660
},
{
"epoch": 0.12776176753121998,
"grad_norm": 1.524565577507019,
"learning_rate": 0.0002998430701721757,
"loss": 3.2964263916015626,
"step": 665
},
{
"epoch": 0.1287223823246878,
"grad_norm": 1.3688217401504517,
"learning_rate": 0.00029983198513795556,
"loss": 3.2987823486328125,
"step": 670
},
{
"epoch": 0.12968299711815562,
"grad_norm": 1.3249455690383911,
"learning_rate": 0.000299820522062266,
"loss": 3.293718719482422,
"step": 675
},
{
"epoch": 0.13064361191162344,
"grad_norm": 1.0875604152679443,
"learning_rate": 0.0002998086809740296,
"loss": 3.294192886352539,
"step": 680
},
{
"epoch": 0.13160422670509125,
"grad_norm": 1.4566972255706787,
"learning_rate": 0.00029979646190312264,
"loss": 3.2981857299804687,
"step": 685
},
{
"epoch": 0.13256484149855907,
"grad_norm": 1.506569743156433,
"learning_rate": 0.0002997838648803751,
"loss": 3.2970245361328123,
"step": 690
},
{
"epoch": 0.1335254562920269,
"grad_norm": 1.2062523365020752,
"learning_rate": 0.00029977088993757045,
"loss": 3.294635772705078,
"step": 695
},
{
"epoch": 0.1344860710854947,
"grad_norm": 1.4316554069519043,
"learning_rate": 0.0002997575371074458,
"loss": 3.294428253173828,
"step": 700
},
{
"epoch": 0.13544668587896252,
"grad_norm": 1.0954806804656982,
"learning_rate": 0.0002997438064236919,
"loss": 3.2926597595214844,
"step": 705
},
{
"epoch": 0.13640730067243034,
"grad_norm": 1.2398245334625244,
"learning_rate": 0.0002997296979209526,
"loss": 3.293048858642578,
"step": 710
},
{
"epoch": 0.1373679154658982,
"grad_norm": 1.7894244194030762,
"learning_rate": 0.00029971521163482494,
"loss": 3.2915740966796876,
"step": 715
},
{
"epoch": 0.138328530259366,
"grad_norm": 1.1798651218414307,
"learning_rate": 0.0002997003476018595,
"loss": 3.2928890228271483,
"step": 720
},
{
"epoch": 0.13928914505283382,
"grad_norm": 1.4055883884429932,
"learning_rate": 0.0002996851058595596,
"loss": 3.295462417602539,
"step": 725
},
{
"epoch": 0.14024975984630164,
"grad_norm": 1.3189594745635986,
"learning_rate": 0.00029966948644638184,
"loss": 3.2915382385253906,
"step": 730
},
{
"epoch": 0.14121037463976946,
"grad_norm": 2.0678322315216064,
"learning_rate": 0.0002996534894017356,
"loss": 3.2929306030273438,
"step": 735
},
{
"epoch": 0.14217098943323728,
"grad_norm": 1.0185853242874146,
"learning_rate": 0.000299637114765983,
"loss": 3.2973453521728517,
"step": 740
},
{
"epoch": 0.1431316042267051,
"grad_norm": 1.1556357145309448,
"learning_rate": 0.00029962036258043893,
"loss": 3.2935462951660157,
"step": 745
},
{
"epoch": 0.1440922190201729,
"grad_norm": 1.392240285873413,
"learning_rate": 0.00029960323288737097,
"loss": 3.2977447509765625,
"step": 750
},
{
"epoch": 0.14505283381364073,
"grad_norm": 1.6778427362442017,
"learning_rate": 0.000299585725729999,
"loss": 3.2953746795654295,
"step": 755
},
{
"epoch": 0.14601344860710855,
"grad_norm": 0.8692209720611572,
"learning_rate": 0.00029956784115249535,
"loss": 3.2932708740234373,
"step": 760
},
{
"epoch": 0.14697406340057637,
"grad_norm": 1.0066865682601929,
"learning_rate": 0.0002995495791999847,
"loss": 3.291333770751953,
"step": 765
},
{
"epoch": 0.14793467819404418,
"grad_norm": 1.1920307874679565,
"learning_rate": 0.0002995309399185439,
"loss": 3.2948654174804686,
"step": 770
},
{
"epoch": 0.148895292987512,
"grad_norm": 0.9732314348220825,
"learning_rate": 0.0002995119233552016,
"loss": 3.2924430847167967,
"step": 775
},
{
"epoch": 0.14985590778097982,
"grad_norm": 1.306164264678955,
"learning_rate": 0.0002994925295579386,
"loss": 3.2954177856445312,
"step": 780
},
{
"epoch": 0.15081652257444764,
"grad_norm": 1.2705974578857422,
"learning_rate": 0.0002994727585756875,
"loss": 3.291736602783203,
"step": 785
},
{
"epoch": 0.15177713736791545,
"grad_norm": 1.4319415092468262,
"learning_rate": 0.00029945261045833256,
"loss": 3.2872940063476563,
"step": 790
},
{
"epoch": 0.15273775216138327,
"grad_norm": 1.5740435123443604,
"learning_rate": 0.0002994320852567094,
"loss": 3.2921607971191404,
"step": 795
},
{
"epoch": 0.15369836695485112,
"grad_norm": 1.2682417631149292,
"learning_rate": 0.0002994111830226053,
"loss": 3.2847610473632813,
"step": 800
},
{
"epoch": 0.15465898174831894,
"grad_norm": 1.3516907691955566,
"learning_rate": 0.0002993899038087587,
"loss": 3.2877117156982423,
"step": 805
},
{
"epoch": 0.15561959654178675,
"grad_norm": 1.1541163921356201,
"learning_rate": 0.00029936824766885927,
"loss": 3.28739013671875,
"step": 810
},
{
"epoch": 0.15658021133525457,
"grad_norm": 1.4818735122680664,
"learning_rate": 0.0002993462146575477,
"loss": 3.2846580505371095,
"step": 815
},
{
"epoch": 0.1575408261287224,
"grad_norm": 1.0636875629425049,
"learning_rate": 0.00029932380483041547,
"loss": 3.2877876281738283,
"step": 820
},
{
"epoch": 0.1585014409221902,
"grad_norm": 1.1833419799804688,
"learning_rate": 0.00029930101824400486,
"loss": 3.289693069458008,
"step": 825
},
{
"epoch": 0.15946205571565802,
"grad_norm": 1.19170343875885,
"learning_rate": 0.00029927785495580887,
"loss": 3.283536911010742,
"step": 830
},
{
"epoch": 0.16042267050912584,
"grad_norm": 1.074803352355957,
"learning_rate": 0.00029925431502427077,
"loss": 3.2888442993164064,
"step": 835
},
{
"epoch": 0.16138328530259366,
"grad_norm": 1.205644130706787,
"learning_rate": 0.00029923039850878423,
"loss": 3.2893348693847657,
"step": 840
},
{
"epoch": 0.16234390009606148,
"grad_norm": 1.6655492782592773,
"learning_rate": 0.00029920610546969314,
"loss": 3.2857383728027343,
"step": 845
},
{
"epoch": 0.1633045148895293,
"grad_norm": 1.3908863067626953,
"learning_rate": 0.0002991814359682912,
"loss": 3.2931854248046877,
"step": 850
},
{
"epoch": 0.1642651296829971,
"grad_norm": 1.3854031562805176,
"learning_rate": 0.00029915639006682223,
"loss": 3.291329193115234,
"step": 855
},
{
"epoch": 0.16522574447646493,
"grad_norm": 1.3702648878097534,
"learning_rate": 0.0002991309678284795,
"loss": 3.283129119873047,
"step": 860
},
{
"epoch": 0.16618635926993275,
"grad_norm": 1.3857735395431519,
"learning_rate": 0.000299105169317406,
"loss": 3.285200500488281,
"step": 865
},
{
"epoch": 0.16714697406340057,
"grad_norm": 1.0670974254608154,
"learning_rate": 0.000299078994598694,
"loss": 3.287001037597656,
"step": 870
},
{
"epoch": 0.16810758885686838,
"grad_norm": 1.2670496702194214,
"learning_rate": 0.000299052443738385,
"loss": 3.28797607421875,
"step": 875
},
{
"epoch": 0.1690682036503362,
"grad_norm": 1.4439146518707275,
"learning_rate": 0.00029902551680346956,
"loss": 3.2858448028564453,
"step": 880
},
{
"epoch": 0.17002881844380405,
"grad_norm": 1.4315037727355957,
"learning_rate": 0.000298998213861887,
"loss": 3.28741340637207,
"step": 885
},
{
"epoch": 0.17098943323727187,
"grad_norm": 1.227054238319397,
"learning_rate": 0.0002989705349825256,
"loss": 3.285395050048828,
"step": 890
},
{
"epoch": 0.17195004803073968,
"grad_norm": 1.9081028699874878,
"learning_rate": 0.0002989424802352218,
"loss": 3.2881591796875,
"step": 895
},
{
"epoch": 0.1729106628242075,
"grad_norm": 1.5766661167144775,
"learning_rate": 0.0002989140496907609,
"loss": 3.282776641845703,
"step": 900
},
{
"epoch": 0.17387127761767532,
"grad_norm": 1.3651994466781616,
"learning_rate": 0.00029888524342087583,
"loss": 3.288460540771484,
"step": 905
},
{
"epoch": 0.17483189241114314,
"grad_norm": 1.1733394861221313,
"learning_rate": 0.00029885606149824786,
"loss": 3.2853572845458983,
"step": 910
},
{
"epoch": 0.17579250720461095,
"grad_norm": 1.798338770866394,
"learning_rate": 0.00029882650399650596,
"loss": 3.2855796813964844,
"step": 915
},
{
"epoch": 0.17675312199807877,
"grad_norm": 1.5101557970046997,
"learning_rate": 0.00029879657099022677,
"loss": 3.2913845062255858,
"step": 920
},
{
"epoch": 0.1777137367915466,
"grad_norm": 1.1141709089279175,
"learning_rate": 0.0002987662625549343,
"loss": 3.2880714416503904,
"step": 925
},
{
"epoch": 0.1786743515850144,
"grad_norm": 1.4672577381134033,
"learning_rate": 0.00029873557876709984,
"loss": 3.2870162963867187,
"step": 930
},
{
"epoch": 0.17963496637848222,
"grad_norm": 1.2008938789367676,
"learning_rate": 0.00029870451970414176,
"loss": 3.281757354736328,
"step": 935
},
{
"epoch": 0.18059558117195004,
"grad_norm": 1.4659417867660522,
"learning_rate": 0.00029867308544442525,
"loss": 3.2835357666015623,
"step": 940
},
{
"epoch": 0.18155619596541786,
"grad_norm": 1.200221300125122,
"learning_rate": 0.0002986412760672622,
"loss": 3.2824310302734374,
"step": 945
},
{
"epoch": 0.18251681075888568,
"grad_norm": 1.1019620895385742,
"learning_rate": 0.0002986090916529109,
"loss": 3.2842033386230467,
"step": 950
},
{
"epoch": 0.1834774255523535,
"grad_norm": 1.3469750881195068,
"learning_rate": 0.00029857653228257587,
"loss": 3.285472106933594,
"step": 955
},
{
"epoch": 0.1844380403458213,
"grad_norm": 1.3108643293380737,
"learning_rate": 0.0002985435980384078,
"loss": 3.285649871826172,
"step": 960
},
{
"epoch": 0.18539865513928913,
"grad_norm": 1.452244758605957,
"learning_rate": 0.0002985102890035032,
"loss": 3.2803878784179688,
"step": 965
},
{
"epoch": 0.18635926993275698,
"grad_norm": 1.229927897453308,
"learning_rate": 0.0002984766052619041,
"loss": 3.283854675292969,
"step": 970
},
{
"epoch": 0.1873198847262248,
"grad_norm": 1.1009323596954346,
"learning_rate": 0.0002984425468985982,
"loss": 3.2806129455566406,
"step": 975
},
{
"epoch": 0.1882804995196926,
"grad_norm": 1.5917284488677979,
"learning_rate": 0.00029840811399951806,
"loss": 3.282550048828125,
"step": 980
},
{
"epoch": 0.18924111431316043,
"grad_norm": 1.1452780961990356,
"learning_rate": 0.00029837330665154156,
"loss": 3.2815521240234373,
"step": 985
},
{
"epoch": 0.19020172910662825,
"grad_norm": 0.9764254093170166,
"learning_rate": 0.000298338124942491,
"loss": 3.284914016723633,
"step": 990
},
{
"epoch": 0.19116234390009607,
"grad_norm": 1.045482873916626,
"learning_rate": 0.0002983025689611337,
"loss": 3.2865413665771483,
"step": 995
},
{
"epoch": 0.19212295869356388,
"grad_norm": 1.3126083612442017,
"learning_rate": 0.0002982666387971809,
"loss": 3.281140899658203,
"step": 1000
},
{
"epoch": 0.1930835734870317,
"grad_norm": 1.3957022428512573,
"learning_rate": 0.0002982303345412881,
"loss": 3.2785770416259767,
"step": 1005
},
{
"epoch": 0.19404418828049952,
"grad_norm": 1.2035447359085083,
"learning_rate": 0.0002981936562850547,
"loss": 3.279229736328125,
"step": 1010
},
{
"epoch": 0.19500480307396734,
"grad_norm": 0.8585781455039978,
"learning_rate": 0.00029815660412102366,
"loss": 3.2802391052246094,
"step": 1015
},
{
"epoch": 0.19596541786743515,
"grad_norm": 1.203058123588562,
"learning_rate": 0.00029811917814268144,
"loss": 3.2790596008300783,
"step": 1020
},
{
"epoch": 0.19692603266090297,
"grad_norm": 1.4647698402404785,
"learning_rate": 0.0002980813784444576,
"loss": 3.2819541931152343,
"step": 1025
},
{
"epoch": 0.1978866474543708,
"grad_norm": 1.4942032098770142,
"learning_rate": 0.0002980432051217246,
"loss": 3.2802181243896484,
"step": 1030
},
{
"epoch": 0.1988472622478386,
"grad_norm": 1.1856545209884644,
"learning_rate": 0.0002980046582707978,
"loss": 3.290009307861328,
"step": 1035
},
{
"epoch": 0.19980787704130643,
"grad_norm": 1.166818380355835,
"learning_rate": 0.00029796573798893464,
"loss": 3.28170166015625,
"step": 1040
},
{
"epoch": 0.20076849183477424,
"grad_norm": 1.310017466545105,
"learning_rate": 0.0002979264443743352,
"loss": 3.282215118408203,
"step": 1045
},
{
"epoch": 0.2017291066282421,
"grad_norm": 1.7260593175888062,
"learning_rate": 0.00029788677752614116,
"loss": 3.2847213745117188,
"step": 1050
},
{
"epoch": 0.2026897214217099,
"grad_norm": 1.083449125289917,
"learning_rate": 0.0002978467375444361,
"loss": 3.278899002075195,
"step": 1055
},
{
"epoch": 0.20365033621517772,
"grad_norm": 1.2556031942367554,
"learning_rate": 0.000297806324530245,
"loss": 3.280004119873047,
"step": 1060
},
{
"epoch": 0.20461095100864554,
"grad_norm": 0.8801208138465881,
"learning_rate": 0.00029776553858553405,
"loss": 3.2757003784179686,
"step": 1065
},
{
"epoch": 0.20557156580211336,
"grad_norm": 1.419739007949829,
"learning_rate": 0.0002977243798132103,
"loss": 3.2807285308837892,
"step": 1070
},
{
"epoch": 0.20653218059558118,
"grad_norm": 1.3385144472122192,
"learning_rate": 0.00029768284831712175,
"loss": 3.280821990966797,
"step": 1075
},
{
"epoch": 0.207492795389049,
"grad_norm": 1.0410194396972656,
"learning_rate": 0.0002976409442020564,
"loss": 3.279698944091797,
"step": 1080
},
{
"epoch": 0.2084534101825168,
"grad_norm": 1.2109322547912598,
"learning_rate": 0.0002975986675737427,
"loss": 3.2776737213134766,
"step": 1085
},
{
"epoch": 0.20941402497598463,
"grad_norm": 1.598090648651123,
"learning_rate": 0.00029755601853884893,
"loss": 3.277222442626953,
"step": 1090
},
{
"epoch": 0.21037463976945245,
"grad_norm": 1.0197858810424805,
"learning_rate": 0.000297512997204983,
"loss": 3.2803329467773437,
"step": 1095
},
{
"epoch": 0.21133525456292027,
"grad_norm": 0.9741032123565674,
"learning_rate": 0.0002974696036806922,
"loss": 3.2853607177734374,
"step": 1100
},
{
"epoch": 0.21229586935638808,
"grad_norm": 1.3451876640319824,
"learning_rate": 0.00029742583807546274,
"loss": 3.276333236694336,
"step": 1105
},
{
"epoch": 0.2132564841498559,
"grad_norm": 1.2260080575942993,
"learning_rate": 0.00029738170049971974,
"loss": 3.2778556823730467,
"step": 1110
},
{
"epoch": 0.21421709894332372,
"grad_norm": 1.4238629341125488,
"learning_rate": 0.00029733719106482694,
"loss": 3.27607421875,
"step": 1115
},
{
"epoch": 0.21517771373679154,
"grad_norm": 1.0617904663085938,
"learning_rate": 0.0002972923098830861,
"loss": 3.278319549560547,
"step": 1120
},
{
"epoch": 0.21613832853025935,
"grad_norm": 1.3632889986038208,
"learning_rate": 0.0002972470570677371,
"loss": 3.2795860290527346,
"step": 1125
},
{
"epoch": 0.21709894332372717,
"grad_norm": 1.2691421508789062,
"learning_rate": 0.00029720143273295735,
"loss": 3.277021026611328,
"step": 1130
},
{
"epoch": 0.21805955811719502,
"grad_norm": 1.2249642610549927,
"learning_rate": 0.0002971554369938619,
"loss": 3.274774169921875,
"step": 1135
},
{
"epoch": 0.21902017291066284,
"grad_norm": 1.6009410619735718,
"learning_rate": 0.0002971090699665025,
"loss": 3.2758060455322267,
"step": 1140
},
{
"epoch": 0.21998078770413065,
"grad_norm": 1.1862974166870117,
"learning_rate": 0.0002970623317678681,
"loss": 3.278969955444336,
"step": 1145
},
{
"epoch": 0.22094140249759847,
"grad_norm": 1.1896966695785522,
"learning_rate": 0.00029701522251588395,
"loss": 3.277644729614258,
"step": 1150
},
{
"epoch": 0.2219020172910663,
"grad_norm": 1.2047041654586792,
"learning_rate": 0.00029696774232941135,
"loss": 3.276856231689453,
"step": 1155
},
{
"epoch": 0.2228626320845341,
"grad_norm": 1.4511812925338745,
"learning_rate": 0.0002969198913282479,
"loss": 3.2762985229492188,
"step": 1160
},
{
"epoch": 0.22382324687800192,
"grad_norm": 1.0419248342514038,
"learning_rate": 0.0002968716696331265,
"loss": 3.2781463623046876,
"step": 1165
},
{
"epoch": 0.22478386167146974,
"grad_norm": 1.0537165403366089,
"learning_rate": 0.00029682307736571533,
"loss": 3.280584716796875,
"step": 1170
},
{
"epoch": 0.22574447646493756,
"grad_norm": 0.9234645366668701,
"learning_rate": 0.00029677411464861783,
"loss": 3.2776565551757812,
"step": 1175
},
{
"epoch": 0.22670509125840538,
"grad_norm": 1.0644463300704956,
"learning_rate": 0.00029672478160537184,
"loss": 3.2758762359619142,
"step": 1180
},
{
"epoch": 0.2276657060518732,
"grad_norm": 0.9237310290336609,
"learning_rate": 0.00029667507836044984,
"loss": 3.2722846984863283,
"step": 1185
},
{
"epoch": 0.228626320845341,
"grad_norm": 1.5153244733810425,
"learning_rate": 0.00029662500503925803,
"loss": 3.2741416931152343,
"step": 1190
},
{
"epoch": 0.22958693563880883,
"grad_norm": 1.3842270374298096,
"learning_rate": 0.00029657456176813664,
"loss": 3.2764862060546873,
"step": 1195
},
{
"epoch": 0.23054755043227665,
"grad_norm": 1.1134364604949951,
"learning_rate": 0.0002965237486743592,
"loss": 3.275590515136719,
"step": 1200
},
{
"epoch": 0.23150816522574447,
"grad_norm": 1.4811969995498657,
"learning_rate": 0.0002964725658861323,
"loss": 3.2718353271484375,
"step": 1205
},
{
"epoch": 0.23246878001921228,
"grad_norm": 1.2888314723968506,
"learning_rate": 0.0002964210135325955,
"loss": 3.2731605529785157,
"step": 1210
},
{
"epoch": 0.2334293948126801,
"grad_norm": 1.530621886253357,
"learning_rate": 0.0002963690917438206,
"loss": 3.2751121520996094,
"step": 1215
},
{
"epoch": 0.23439000960614795,
"grad_norm": 1.0951229333877563,
"learning_rate": 0.0002963168006508116,
"loss": 3.2797927856445312,
"step": 1220
},
{
"epoch": 0.23535062439961577,
"grad_norm": 1.0309784412384033,
"learning_rate": 0.00029626414038550424,
"loss": 3.279781723022461,
"step": 1225
},
{
"epoch": 0.23631123919308358,
"grad_norm": 1.211702823638916,
"learning_rate": 0.00029621111108076584,
"loss": 3.276560592651367,
"step": 1230
},
{
"epoch": 0.2372718539865514,
"grad_norm": 1.0704128742218018,
"learning_rate": 0.0002961577128703948,
"loss": 3.2736587524414062,
"step": 1235
},
{
"epoch": 0.23823246878001922,
"grad_norm": 1.4519094228744507,
"learning_rate": 0.0002961039458891202,
"loss": 3.274006652832031,
"step": 1240
},
{
"epoch": 0.23919308357348704,
"grad_norm": 1.2808977365493774,
"learning_rate": 0.00029604981027260167,
"loss": 3.27347412109375,
"step": 1245
},
{
"epoch": 0.24015369836695485,
"grad_norm": 0.8584306836128235,
"learning_rate": 0.00029599530615742896,
"loss": 3.2714912414550783,
"step": 1250
},
{
"epoch": 0.24111431316042267,
"grad_norm": 1.1275231838226318,
"learning_rate": 0.0002959404336811215,
"loss": 3.2730926513671874,
"step": 1255
},
{
"epoch": 0.2420749279538905,
"grad_norm": 1.422745943069458,
"learning_rate": 0.0002958851929821283,
"loss": 3.27635498046875,
"step": 1260
},
{
"epoch": 0.2430355427473583,
"grad_norm": 1.0575655698776245,
"learning_rate": 0.00029582958419982713,
"loss": 3.272149658203125,
"step": 1265
},
{
"epoch": 0.24399615754082613,
"grad_norm": 1.2576707601547241,
"learning_rate": 0.0002957736074745248,
"loss": 3.275560760498047,
"step": 1270
},
{
"epoch": 0.24495677233429394,
"grad_norm": 1.328376054763794,
"learning_rate": 0.0002957172629474562,
"loss": 3.272699737548828,
"step": 1275
},
{
"epoch": 0.24591738712776176,
"grad_norm": 1.0527936220169067,
"learning_rate": 0.00029566055076078447,
"loss": 3.269004058837891,
"step": 1280
},
{
"epoch": 0.24687800192122958,
"grad_norm": 1.1557375192642212,
"learning_rate": 0.00029560347105760023,
"loss": 3.2709724426269533,
"step": 1285
},
{
"epoch": 0.2478386167146974,
"grad_norm": 1.1958216428756714,
"learning_rate": 0.00029554602398192137,
"loss": 3.276508331298828,
"step": 1290
},
{
"epoch": 0.24879923150816521,
"grad_norm": 1.0567305088043213,
"learning_rate": 0.00029548820967869287,
"loss": 3.270316314697266,
"step": 1295
},
{
"epoch": 0.24975984630163303,
"grad_norm": 1.1860175132751465,
"learning_rate": 0.00029543002829378605,
"loss": 3.274994659423828,
"step": 1300
},
{
"epoch": 0.2507204610951009,
"grad_norm": 1.2902470827102661,
"learning_rate": 0.0002953714799739986,
"loss": 3.2740650177001953,
"step": 1305
},
{
"epoch": 0.2516810758885687,
"grad_norm": 1.0249444246292114,
"learning_rate": 0.0002953125648670538,
"loss": 3.273143768310547,
"step": 1310
},
{
"epoch": 0.2526416906820365,
"grad_norm": 1.051010251045227,
"learning_rate": 0.00029525328312160064,
"loss": 3.270255279541016,
"step": 1315
},
{
"epoch": 0.25360230547550433,
"grad_norm": 1.5545085668563843,
"learning_rate": 0.00029519363488721297,
"loss": 3.270277404785156,
"step": 1320
},
{
"epoch": 0.25456292026897215,
"grad_norm": 1.096474289894104,
"learning_rate": 0.0002951336203143895,
"loss": 3.2695423126220704,
"step": 1325
},
{
"epoch": 0.25552353506243997,
"grad_norm": 1.3266576528549194,
"learning_rate": 0.0002950732395545531,
"loss": 3.268444061279297,
"step": 1330
},
{
"epoch": 0.2564841498559078,
"grad_norm": 1.3308000564575195,
"learning_rate": 0.0002950124927600505,
"loss": 3.270121765136719,
"step": 1335
},
{
"epoch": 0.2574447646493756,
"grad_norm": 0.9661091566085815,
"learning_rate": 0.0002949513800841523,
"loss": 3.2689208984375,
"step": 1340
},
{
"epoch": 0.2584053794428434,
"grad_norm": 1.0783677101135254,
"learning_rate": 0.0002948899016810519,
"loss": 3.2726470947265627,
"step": 1345
},
{
"epoch": 0.25936599423631124,
"grad_norm": 1.44166898727417,
"learning_rate": 0.00029482805770586566,
"loss": 3.2676158905029298,
"step": 1350
},
{
"epoch": 0.26032660902977905,
"grad_norm": 0.9048498868942261,
"learning_rate": 0.0002947658483146323,
"loss": 3.268976593017578,
"step": 1355
},
{
"epoch": 0.2612872238232469,
"grad_norm": 0.9976827502250671,
"learning_rate": 0.00029470327366431257,
"loss": 3.2651641845703123,
"step": 1360
},
{
"epoch": 0.2622478386167147,
"grad_norm": 1.4668549299240112,
"learning_rate": 0.00029464033391278857,
"loss": 3.267919158935547,
"step": 1365
},
{
"epoch": 0.2632084534101825,
"grad_norm": 1.3420720100402832,
"learning_rate": 0.0002945770292188638,
"loss": 3.2697959899902345,
"step": 1370
},
{
"epoch": 0.2641690682036503,
"grad_norm": 1.3224592208862305,
"learning_rate": 0.00029451335974226255,
"loss": 3.2723613739013673,
"step": 1375
},
{
"epoch": 0.26512968299711814,
"grad_norm": 1.1924461126327515,
"learning_rate": 0.0002944493256436294,
"loss": 3.270499038696289,
"step": 1380
},
{
"epoch": 0.26609029779058596,
"grad_norm": 1.2803032398223877,
"learning_rate": 0.000294384927084529,
"loss": 3.269664001464844,
"step": 1385
},
{
"epoch": 0.2670509125840538,
"grad_norm": 1.0187225341796875,
"learning_rate": 0.0002943201642274455,
"loss": 3.2647491455078126,
"step": 1390
},
{
"epoch": 0.2680115273775216,
"grad_norm": 1.3384113311767578,
"learning_rate": 0.00029425503723578216,
"loss": 3.2721004486083984,
"step": 1395
},
{
"epoch": 0.2689721421709894,
"grad_norm": 1.2812883853912354,
"learning_rate": 0.00029418954627386115,
"loss": 3.269911193847656,
"step": 1400
},
{
"epoch": 0.26993275696445723,
"grad_norm": 1.1734299659729004,
"learning_rate": 0.0002941236915069228,
"loss": 3.2672080993652344,
"step": 1405
},
{
"epoch": 0.27089337175792505,
"grad_norm": 1.069024682044983,
"learning_rate": 0.00029405747310112557,
"loss": 3.2702381134033205,
"step": 1410
},
{
"epoch": 0.27185398655139287,
"grad_norm": 1.286638617515564,
"learning_rate": 0.0002939908912235452,
"loss": 3.2697906494140625,
"step": 1415
},
{
"epoch": 0.2728146013448607,
"grad_norm": 1.1734787225723267,
"learning_rate": 0.0002939239460421746,
"loss": 3.2664981842041017,
"step": 1420
},
{
"epoch": 0.2737752161383285,
"grad_norm": 1.150191307067871,
"learning_rate": 0.00029385663772592334,
"loss": 3.2677589416503907,
"step": 1425
},
{
"epoch": 0.2747358309317964,
"grad_norm": 1.3920680284500122,
"learning_rate": 0.00029378896644461716,
"loss": 3.2696159362792967,
"step": 1430
},
{
"epoch": 0.2756964457252642,
"grad_norm": 1.1501007080078125,
"learning_rate": 0.0002937209323689978,
"loss": 3.268462371826172,
"step": 1435
},
{
"epoch": 0.276657060518732,
"grad_norm": 1.2256152629852295,
"learning_rate": 0.00029365253567072207,
"loss": 3.269298553466797,
"step": 1440
},
{
"epoch": 0.27761767531219983,
"grad_norm": 1.2467095851898193,
"learning_rate": 0.00029358377652236186,
"loss": 3.2696155548095702,
"step": 1445
},
{
"epoch": 0.27857829010566765,
"grad_norm": 0.7786490321159363,
"learning_rate": 0.00029351465509740366,
"loss": 3.265594482421875,
"step": 1450
},
{
"epoch": 0.27953890489913547,
"grad_norm": 0.9412602186203003,
"learning_rate": 0.00029344517157024773,
"loss": 3.2679397583007814,
"step": 1455
},
{
"epoch": 0.2804995196926033,
"grad_norm": 1.1962471008300781,
"learning_rate": 0.0002933753261162084,
"loss": 3.276123046875,
"step": 1460
},
{
"epoch": 0.2814601344860711,
"grad_norm": 1.1592360734939575,
"learning_rate": 0.00029330511891151263,
"loss": 3.2704566955566405,
"step": 1465
},
{
"epoch": 0.2824207492795389,
"grad_norm": 0.8931114673614502,
"learning_rate": 0.00029323455013330064,
"loss": 3.2649574279785156,
"step": 1470
},
{
"epoch": 0.28338136407300674,
"grad_norm": 0.8980585932731628,
"learning_rate": 0.00029316361995962453,
"loss": 3.2683528900146483,
"step": 1475
},
{
"epoch": 0.28434197886647455,
"grad_norm": 1.330986499786377,
"learning_rate": 0.0002930923285694485,
"loss": 3.2716033935546873,
"step": 1480
},
{
"epoch": 0.28530259365994237,
"grad_norm": 1.0797970294952393,
"learning_rate": 0.000293020676142648,
"loss": 3.268834686279297,
"step": 1485
},
{
"epoch": 0.2862632084534102,
"grad_norm": 1.2156567573547363,
"learning_rate": 0.00029294866286000946,
"loss": 3.2620380401611326,
"step": 1490
},
{
"epoch": 0.287223823246878,
"grad_norm": 1.3845075368881226,
"learning_rate": 0.00029287628890322986,
"loss": 3.2627891540527343,
"step": 1495
},
{
"epoch": 0.2881844380403458,
"grad_norm": 1.2039449214935303,
"learning_rate": 0.00029280355445491596,
"loss": 3.2676162719726562,
"step": 1500
},
{
"epoch": 0.28914505283381364,
"grad_norm": 1.2778565883636475,
"learning_rate": 0.00029273045969858437,
"loss": 3.267996978759766,
"step": 1505
},
{
"epoch": 0.29010566762728146,
"grad_norm": 0.9892801642417908,
"learning_rate": 0.0002926570048186606,
"loss": 3.2660484313964844,
"step": 1510
},
{
"epoch": 0.2910662824207493,
"grad_norm": 1.0218427181243896,
"learning_rate": 0.00029258319000047885,
"loss": 3.269457244873047,
"step": 1515
},
{
"epoch": 0.2920268972142171,
"grad_norm": 1.0072425603866577,
"learning_rate": 0.00029250901543028144,
"loss": 3.267464828491211,
"step": 1520
},
{
"epoch": 0.2929875120076849,
"grad_norm": 1.4437980651855469,
"learning_rate": 0.00029243448129521847,
"loss": 3.2681037902832033,
"step": 1525
},
{
"epoch": 0.29394812680115273,
"grad_norm": 0.8555430769920349,
"learning_rate": 0.0002923595877833472,
"loss": 3.2668888092041017,
"step": 1530
},
{
"epoch": 0.29490874159462055,
"grad_norm": 1.2450693845748901,
"learning_rate": 0.0002922843350836317,
"loss": 3.2664356231689453,
"step": 1535
},
{
"epoch": 0.29586935638808837,
"grad_norm": 0.9553253650665283,
"learning_rate": 0.00029220872338594215,
"loss": 3.263041687011719,
"step": 1540
},
{
"epoch": 0.2968299711815562,
"grad_norm": 1.2076423168182373,
"learning_rate": 0.0002921327528810547,
"loss": 3.266950988769531,
"step": 1545
},
{
"epoch": 0.297790585975024,
"grad_norm": 2.1191205978393555,
"learning_rate": 0.00029205642376065066,
"loss": 3.2656558990478515,
"step": 1550
},
{
"epoch": 0.2987512007684918,
"grad_norm": 1.4229607582092285,
"learning_rate": 0.0002919797362173163,
"loss": 3.2715381622314452,
"step": 1555
},
{
"epoch": 0.29971181556195964,
"grad_norm": 1.2368606328964233,
"learning_rate": 0.00029190269044454206,
"loss": 3.2648460388183596,
"step": 1560
},
{
"epoch": 0.30067243035542746,
"grad_norm": 1.266961693763733,
"learning_rate": 0.00029182528663672245,
"loss": 3.2667465209960938,
"step": 1565
},
{
"epoch": 0.3016330451488953,
"grad_norm": 0.9238030314445496,
"learning_rate": 0.00029174752498915515,
"loss": 3.262348175048828,
"step": 1570
},
{
"epoch": 0.3025936599423631,
"grad_norm": 1.0845513343811035,
"learning_rate": 0.0002916694056980408,
"loss": 3.266324996948242,
"step": 1575
},
{
"epoch": 0.3035542747358309,
"grad_norm": 1.050512433052063,
"learning_rate": 0.0002915909289604823,
"loss": 3.2667007446289062,
"step": 1580
},
{
"epoch": 0.3045148895292987,
"grad_norm": 1.25586998462677,
"learning_rate": 0.00029151209497448467,
"loss": 3.262495422363281,
"step": 1585
},
{
"epoch": 0.30547550432276654,
"grad_norm": 1.1228755712509155,
"learning_rate": 0.000291432903938954,
"loss": 3.26649169921875,
"step": 1590
},
{
"epoch": 0.30643611911623436,
"grad_norm": 1.2967936992645264,
"learning_rate": 0.0002913533560536975,
"loss": 3.264292526245117,
"step": 1595
},
{
"epoch": 0.30739673390970224,
"grad_norm": 1.1702841520309448,
"learning_rate": 0.00029127345151942253,
"loss": 3.2614276885986326,
"step": 1600
},
{
"epoch": 0.30835734870317005,
"grad_norm": 1.438833236694336,
"learning_rate": 0.00029119319053773655,
"loss": 3.2649909973144533,
"step": 1605
},
{
"epoch": 0.30931796349663787,
"grad_norm": 1.1444823741912842,
"learning_rate": 0.00029111257331114617,
"loss": 3.2618698120117187,
"step": 1610
},
{
"epoch": 0.3102785782901057,
"grad_norm": 1.2421047687530518,
"learning_rate": 0.00029103160004305693,
"loss": 3.260860061645508,
"step": 1615
},
{
"epoch": 0.3112391930835735,
"grad_norm": 1.3885422945022583,
"learning_rate": 0.00029095027093777276,
"loss": 3.2636131286621093,
"step": 1620
},
{
"epoch": 0.3121998078770413,
"grad_norm": 1.403899073600769,
"learning_rate": 0.00029086858620049527,
"loss": 3.261060333251953,
"step": 1625
},
{
"epoch": 0.31316042267050914,
"grad_norm": 1.4193172454833984,
"learning_rate": 0.0002907865460373234,
"loss": 3.2587600708007813,
"step": 1630
},
{
"epoch": 0.31412103746397696,
"grad_norm": 0.9989815354347229,
"learning_rate": 0.00029070415065525295,
"loss": 3.2618736267089843,
"step": 1635
},
{
"epoch": 0.3150816522574448,
"grad_norm": 1.0307117700576782,
"learning_rate": 0.00029062140026217596,
"loss": 3.25679931640625,
"step": 1640
},
{
"epoch": 0.3160422670509126,
"grad_norm": 1.3937115669250488,
"learning_rate": 0.0002905382950668801,
"loss": 3.2610740661621094,
"step": 1645
},
{
"epoch": 0.3170028818443804,
"grad_norm": 1.0371273756027222,
"learning_rate": 0.0002904548352790483,
"loss": 3.264708709716797,
"step": 1650
},
{
"epoch": 0.31796349663784823,
"grad_norm": 1.202774167060852,
"learning_rate": 0.0002903710211092582,
"loss": 3.261525344848633,
"step": 1655
},
{
"epoch": 0.31892411143131605,
"grad_norm": 1.2215559482574463,
"learning_rate": 0.0002902868527689816,
"loss": 3.264019012451172,
"step": 1660
},
{
"epoch": 0.31988472622478387,
"grad_norm": 0.8910490274429321,
"learning_rate": 0.0002902023304705837,
"loss": 3.262919616699219,
"step": 1665
},
{
"epoch": 0.3208453410182517,
"grad_norm": 1.35845148563385,
"learning_rate": 0.00029011745442732303,
"loss": 3.2655372619628906,
"step": 1670
},
{
"epoch": 0.3218059558117195,
"grad_norm": 1.354686975479126,
"learning_rate": 0.0002900322248533506,
"loss": 3.26285400390625,
"step": 1675
},
{
"epoch": 0.3227665706051873,
"grad_norm": 1.0969462394714355,
"learning_rate": 0.0002899466419637093,
"loss": 3.257940673828125,
"step": 1680
},
{
"epoch": 0.32372718539865514,
"grad_norm": 1.291715145111084,
"learning_rate": 0.00028986070597433354,
"loss": 3.2633174896240233,
"step": 1685
},
{
"epoch": 0.32468780019212296,
"grad_norm": 1.4123157262802124,
"learning_rate": 0.00028977441710204867,
"loss": 3.2593170166015626,
"step": 1690
},
{
"epoch": 0.3256484149855908,
"grad_norm": 1.1881605386734009,
"learning_rate": 0.0002896877755645703,
"loss": 3.2590873718261717,
"step": 1695
},
{
"epoch": 0.3266090297790586,
"grad_norm": 1.043522596359253,
"learning_rate": 0.00028960078158050403,
"loss": 3.258655548095703,
"step": 1700
},
{
"epoch": 0.3275696445725264,
"grad_norm": 1.0990368127822876,
"learning_rate": 0.0002895134353693445,
"loss": 3.258600616455078,
"step": 1705
},
{
"epoch": 0.3285302593659942,
"grad_norm": 1.4628812074661255,
"learning_rate": 0.0002894257371514752,
"loss": 3.262346649169922,
"step": 1710
},
{
"epoch": 0.32949087415946204,
"grad_norm": 0.8175418376922607,
"learning_rate": 0.0002893376871481678,
"loss": 3.255454254150391,
"step": 1715
},
{
"epoch": 0.33045148895292986,
"grad_norm": 1.0393476486206055,
"learning_rate": 0.0002892492855815814,
"loss": 3.2613197326660157,
"step": 1720
},
{
"epoch": 0.3314121037463977,
"grad_norm": 1.250258445739746,
"learning_rate": 0.0002891605326747622,
"loss": 3.260877227783203,
"step": 1725
},
{
"epoch": 0.3323727185398655,
"grad_norm": 0.8695818781852722,
"learning_rate": 0.00028907142865164305,
"loss": 3.259082794189453,
"step": 1730
},
{
"epoch": 0.3333333333333333,
"grad_norm": 1.0892332792282104,
"learning_rate": 0.00028898197373704234,
"loss": 3.261730194091797,
"step": 1735
},
{
"epoch": 0.33429394812680113,
"grad_norm": 1.2874714136123657,
"learning_rate": 0.00028889216815666406,
"loss": 3.2563186645507813,
"step": 1740
},
{
"epoch": 0.33525456292026895,
"grad_norm": 1.2988804578781128,
"learning_rate": 0.000288802012137097,
"loss": 3.2569385528564454,
"step": 1745
},
{
"epoch": 0.33621517771373677,
"grad_norm": 1.2552379369735718,
"learning_rate": 0.0002887115059058139,
"loss": 3.2592254638671876,
"step": 1750
},
{
"epoch": 0.3371757925072046,
"grad_norm": 0.9785555601119995,
"learning_rate": 0.0002886206496911714,
"loss": 3.257631301879883,
"step": 1755
},
{
"epoch": 0.3381364073006724,
"grad_norm": 1.3783308267593384,
"learning_rate": 0.00028852944372240896,
"loss": 3.263458251953125,
"step": 1760
},
{
"epoch": 0.3390970220941403,
"grad_norm": 1.2814546823501587,
"learning_rate": 0.00028843788822964864,
"loss": 3.2578887939453125,
"step": 1765
},
{
"epoch": 0.3400576368876081,
"grad_norm": 1.2692376375198364,
"learning_rate": 0.00028834598344389437,
"loss": 3.2599555969238283,
"step": 1770
},
{
"epoch": 0.3410182516810759,
"grad_norm": 1.3274823427200317,
"learning_rate": 0.0002882537295970312,
"loss": 3.261073684692383,
"step": 1775
},
{
"epoch": 0.34197886647454373,
"grad_norm": 1.1028993129730225,
"learning_rate": 0.00028816112692182526,
"loss": 3.2600685119628907,
"step": 1780
},
{
"epoch": 0.34293948126801155,
"grad_norm": 1.0940430164337158,
"learning_rate": 0.0002880681756519224,
"loss": 3.2631156921386717,
"step": 1785
},
{
"epoch": 0.34390009606147937,
"grad_norm": 1.1741334199905396,
"learning_rate": 0.0002879748760218484,
"loss": 3.2562606811523436,
"step": 1790
},
{
"epoch": 0.3448607108549472,
"grad_norm": 1.4189056158065796,
"learning_rate": 0.0002878812282670077,
"loss": 3.2629486083984376,
"step": 1795
},
{
"epoch": 0.345821325648415,
"grad_norm": 0.9202475547790527,
"learning_rate": 0.00028778723262368313,
"loss": 3.258461761474609,
"step": 1800
},
{
"epoch": 0.3467819404418828,
"grad_norm": 1.0866518020629883,
"learning_rate": 0.00028769288932903547,
"loss": 3.2561798095703125,
"step": 1805
},
{
"epoch": 0.34774255523535064,
"grad_norm": 1.1435341835021973,
"learning_rate": 0.00028759819862110244,
"loss": 3.253205108642578,
"step": 1810
},
{
"epoch": 0.34870317002881845,
"grad_norm": 0.9997341632843018,
"learning_rate": 0.00028750316073879845,
"loss": 3.256957244873047,
"step": 1815
},
{
"epoch": 0.34966378482228627,
"grad_norm": 1.5304245948791504,
"learning_rate": 0.0002874077759219138,
"loss": 3.253832244873047,
"step": 1820
},
{
"epoch": 0.3506243996157541,
"grad_norm": 1.3517911434173584,
"learning_rate": 0.0002873120444111142,
"loss": 3.2599609375,
"step": 1825
},
{
"epoch": 0.3515850144092219,
"grad_norm": 1.1517263650894165,
"learning_rate": 0.0002872159664479401,
"loss": 3.253156280517578,
"step": 1830
},
{
"epoch": 0.3525456292026897,
"grad_norm": 1.2207057476043701,
"learning_rate": 0.00028711954227480596,
"loss": 3.255885696411133,
"step": 1835
},
{
"epoch": 0.35350624399615754,
"grad_norm": 1.4356719255447388,
"learning_rate": 0.00028702277213499993,
"loss": 3.2568824768066404,
"step": 1840
},
{
"epoch": 0.35446685878962536,
"grad_norm": 0.9663187265396118,
"learning_rate": 0.000286925656272683,
"loss": 3.2607173919677734,
"step": 1845
},
{
"epoch": 0.3554274735830932,
"grad_norm": 1.0548568964004517,
"learning_rate": 0.0002868281949328884,
"loss": 3.254198455810547,
"step": 1850
},
{
"epoch": 0.356388088376561,
"grad_norm": 1.216261386871338,
"learning_rate": 0.0002867303883615212,
"loss": 3.25797119140625,
"step": 1855
},
{
"epoch": 0.3573487031700288,
"grad_norm": 0.9959939122200012,
"learning_rate": 0.00028663223680535727,
"loss": 3.2549095153808594,
"step": 1860
},
{
"epoch": 0.35830931796349663,
"grad_norm": 1.0812616348266602,
"learning_rate": 0.0002865337405120432,
"loss": 3.2560733795166015,
"step": 1865
},
{
"epoch": 0.35926993275696445,
"grad_norm": 1.2685303688049316,
"learning_rate": 0.0002864348997300951,
"loss": 3.2555168151855467,
"step": 1870
},
{
"epoch": 0.36023054755043227,
"grad_norm": 1.0879360437393188,
"learning_rate": 0.0002863357147088985,
"loss": 3.255245590209961,
"step": 1875
},
{
"epoch": 0.3611911623439001,
"grad_norm": 1.0502004623413086,
"learning_rate": 0.0002862361856987073,
"loss": 3.260142517089844,
"step": 1880
},
{
"epoch": 0.3621517771373679,
"grad_norm": 1.2763992547988892,
"learning_rate": 0.0002861363129506435,
"loss": 3.2576175689697267,
"step": 1885
},
{
"epoch": 0.3631123919308357,
"grad_norm": 1.2673193216323853,
"learning_rate": 0.0002860360967166963,
"loss": 3.259668731689453,
"step": 1890
},
{
"epoch": 0.36407300672430354,
"grad_norm": 1.0005592107772827,
"learning_rate": 0.00028593553724972146,
"loss": 3.2603923797607424,
"step": 1895
},
{
"epoch": 0.36503362151777136,
"grad_norm": 1.2387863397598267,
"learning_rate": 0.0002858346348034408,
"loss": 3.2571121215820313,
"step": 1900
},
{
"epoch": 0.3659942363112392,
"grad_norm": 1.0775574445724487,
"learning_rate": 0.0002857333896324417,
"loss": 3.2519359588623047,
"step": 1905
},
{
"epoch": 0.366954851104707,
"grad_norm": 1.1767851114273071,
"learning_rate": 0.00028563180199217596,
"loss": 3.256037139892578,
"step": 1910
},
{
"epoch": 0.3679154658981748,
"grad_norm": 1.2488412857055664,
"learning_rate": 0.00028552987213895965,
"loss": 3.2543380737304686,
"step": 1915
},
{
"epoch": 0.3688760806916426,
"grad_norm": 1.2731399536132812,
"learning_rate": 0.0002854276003299723,
"loss": 3.251921844482422,
"step": 1920
},
{
"epoch": 0.36983669548511044,
"grad_norm": 1.2652751207351685,
"learning_rate": 0.00028532498682325613,
"loss": 3.254206085205078,
"step": 1925
},
{
"epoch": 0.37079731027857826,
"grad_norm": 1.3380461931228638,
"learning_rate": 0.0002852220318777156,
"loss": 3.2542865753173826,
"step": 1930
},
{
"epoch": 0.37175792507204614,
"grad_norm": 1.0234968662261963,
"learning_rate": 0.0002851187357531165,
"loss": 3.2551605224609377,
"step": 1935
},
{
"epoch": 0.37271853986551395,
"grad_norm": 0.7750507593154907,
"learning_rate": 0.00028501509871008563,
"loss": 3.258992004394531,
"step": 1940
},
{
"epoch": 0.37367915465898177,
"grad_norm": 0.9654103517532349,
"learning_rate": 0.00028491112101010986,
"loss": 3.2557037353515623,
"step": 1945
},
{
"epoch": 0.3746397694524496,
"grad_norm": 1.1453462839126587,
"learning_rate": 0.0002848068029155356,
"loss": 3.249644470214844,
"step": 1950
},
{
"epoch": 0.3756003842459174,
"grad_norm": 1.0247918367385864,
"learning_rate": 0.0002847021446895681,
"loss": 3.253766632080078,
"step": 1955
},
{
"epoch": 0.3765609990393852,
"grad_norm": 1.4280420541763306,
"learning_rate": 0.0002845971465962708,
"loss": 3.2554515838623046,
"step": 1960
},
{
"epoch": 0.37752161383285304,
"grad_norm": 1.534204125404358,
"learning_rate": 0.00028449180890056464,
"loss": 3.2532962799072265,
"step": 1965
},
{
"epoch": 0.37848222862632086,
"grad_norm": 1.2856396436691284,
"learning_rate": 0.00028438613186822746,
"loss": 3.251531219482422,
"step": 1970
},
{
"epoch": 0.3794428434197887,
"grad_norm": 1.0911295413970947,
"learning_rate": 0.00028428011576589327,
"loss": 3.250580978393555,
"step": 1975
},
{
"epoch": 0.3804034582132565,
"grad_norm": 1.142147421836853,
"learning_rate": 0.00028417376086105155,
"loss": 3.254068374633789,
"step": 1980
},
{
"epoch": 0.3813640730067243,
"grad_norm": 1.335033655166626,
"learning_rate": 0.0002840670674220466,
"loss": 3.2494667053222654,
"step": 1985
},
{
"epoch": 0.38232468780019213,
"grad_norm": 1.410131812095642,
"learning_rate": 0.00028396003571807697,
"loss": 3.2515232086181642,
"step": 1990
},
{
"epoch": 0.38328530259365995,
"grad_norm": 1.3424606323242188,
"learning_rate": 0.0002838526660191946,
"loss": 3.250862884521484,
"step": 1995
},
{
"epoch": 0.38424591738712777,
"grad_norm": 0.8202113509178162,
"learning_rate": 0.0002837449585963043,
"loss": 3.249342346191406,
"step": 2000
},
{
"epoch": 0.3852065321805956,
"grad_norm": 0.9713074564933777,
"learning_rate": 0.00028363691372116284,
"loss": 3.2535247802734375,
"step": 2005
},
{
"epoch": 0.3861671469740634,
"grad_norm": 0.7178729176521301,
"learning_rate": 0.0002835285316663786,
"loss": 3.2526882171630858,
"step": 2010
},
{
"epoch": 0.3871277617675312,
"grad_norm": 0.9730674624443054,
"learning_rate": 0.0002834198127054106,
"loss": 3.2526809692382814,
"step": 2015
},
{
"epoch": 0.38808837656099904,
"grad_norm": 1.3725463151931763,
"learning_rate": 0.000283310757112568,
"loss": 3.2512813568115235,
"step": 2020
},
{
"epoch": 0.38904899135446686,
"grad_norm": 1.5489643812179565,
"learning_rate": 0.0002832013651630091,
"loss": 3.2502918243408203,
"step": 2025
},
{
"epoch": 0.3900096061479347,
"grad_norm": 0.8844899535179138,
"learning_rate": 0.0002830916371327412,
"loss": 3.250738525390625,
"step": 2030
},
{
"epoch": 0.3909702209414025,
"grad_norm": 1.0500115156173706,
"learning_rate": 0.0002829815732986192,
"loss": 3.2540382385253905,
"step": 2035
},
{
"epoch": 0.3919308357348703,
"grad_norm": 1.0166023969650269,
"learning_rate": 0.00028287117393834563,
"loss": 3.2548370361328125,
"step": 2040
},
{
"epoch": 0.3928914505283381,
"grad_norm": 1.216377854347229,
"learning_rate": 0.00028276043933046926,
"loss": 3.248933792114258,
"step": 2045
},
{
"epoch": 0.39385206532180594,
"grad_norm": 0.9477090239524841,
"learning_rate": 0.0002826493697543849,
"loss": 3.2496387481689455,
"step": 2050
},
{
"epoch": 0.39481268011527376,
"grad_norm": 1.4546550512313843,
"learning_rate": 0.00028253796549033245,
"loss": 3.257366943359375,
"step": 2055
},
{
"epoch": 0.3957732949087416,
"grad_norm": 1.3477628231048584,
"learning_rate": 0.0002824262268193964,
"loss": 3.2546245574951174,
"step": 2060
},
{
"epoch": 0.3967339097022094,
"grad_norm": 1.1829005479812622,
"learning_rate": 0.00028231415402350476,
"loss": 3.2527755737304687,
"step": 2065
},
{
"epoch": 0.3976945244956772,
"grad_norm": 1.00674569606781,
"learning_rate": 0.00028220174738542876,
"loss": 3.249166488647461,
"step": 2070
},
{
"epoch": 0.39865513928914503,
"grad_norm": 1.143075942993164,
"learning_rate": 0.00028208900718878187,
"loss": 3.2508075714111326,
"step": 2075
},
{
"epoch": 0.39961575408261285,
"grad_norm": 0.9893442392349243,
"learning_rate": 0.00028197593371801915,
"loss": 3.256524658203125,
"step": 2080
},
{
"epoch": 0.40057636887608067,
"grad_norm": 0.9596649408340454,
"learning_rate": 0.00028186252725843664,
"loss": 3.2511486053466796,
"step": 2085
},
{
"epoch": 0.4015369836695485,
"grad_norm": 1.4209188222885132,
"learning_rate": 0.00028174878809617037,
"loss": 3.2548614501953126,
"step": 2090
},
{
"epoch": 0.4024975984630163,
"grad_norm": 1.0415889024734497,
"learning_rate": 0.000281634716518196,
"loss": 3.2491207122802734,
"step": 2095
},
{
"epoch": 0.4034582132564842,
"grad_norm": 1.0843944549560547,
"learning_rate": 0.00028152031281232775,
"loss": 3.2516613006591797,
"step": 2100
},
{
"epoch": 0.404418828049952,
"grad_norm": 1.5569794178009033,
"learning_rate": 0.00028140557726721795,
"loss": 3.2495628356933595,
"step": 2105
},
{
"epoch": 0.4053794428434198,
"grad_norm": 0.8837640881538391,
"learning_rate": 0.00028129051017235614,
"loss": 3.249878692626953,
"step": 2110
},
{
"epoch": 0.40634005763688763,
"grad_norm": 1.1749989986419678,
"learning_rate": 0.0002811751118180684,
"loss": 3.2475624084472656,
"step": 2115
},
{
"epoch": 0.40730067243035545,
"grad_norm": 1.2945916652679443,
"learning_rate": 0.00028105938249551666,
"loss": 3.2487873077392577,
"step": 2120
},
{
"epoch": 0.40826128722382327,
"grad_norm": 0.9443672299385071,
"learning_rate": 0.0002809433224966978,
"loss": 3.247914123535156,
"step": 2125
},
{
"epoch": 0.4092219020172911,
"grad_norm": 1.1064391136169434,
"learning_rate": 0.0002808269321144431,
"loss": 3.250864028930664,
"step": 2130
},
{
"epoch": 0.4101825168107589,
"grad_norm": 1.2150980234146118,
"learning_rate": 0.00028071021164241755,
"loss": 3.248881530761719,
"step": 2135
},
{
"epoch": 0.4111431316042267,
"grad_norm": 0.8171827793121338,
"learning_rate": 0.00028059316137511877,
"loss": 3.246689224243164,
"step": 2140
},
{
"epoch": 0.41210374639769454,
"grad_norm": 1.093009114265442,
"learning_rate": 0.0002804757816078766,
"loss": 3.2493213653564452,
"step": 2145
},
{
"epoch": 0.41306436119116235,
"grad_norm": 1.021937370300293,
"learning_rate": 0.00028035807263685224,
"loss": 3.2454933166503905,
"step": 2150
},
{
"epoch": 0.4140249759846302,
"grad_norm": 1.1384049654006958,
"learning_rate": 0.0002802400347590376,
"loss": 3.2516769409179687,
"step": 2155
},
{
"epoch": 0.414985590778098,
"grad_norm": 1.0551131963729858,
"learning_rate": 0.00028012166827225425,
"loss": 3.2476879119873048,
"step": 2160
},
{
"epoch": 0.4159462055715658,
"grad_norm": 0.9579499363899231,
"learning_rate": 0.0002800029734751529,
"loss": 3.247665023803711,
"step": 2165
},
{
"epoch": 0.4169068203650336,
"grad_norm": 1.290895938873291,
"learning_rate": 0.0002798839506672129,
"loss": 3.248125457763672,
"step": 2170
},
{
"epoch": 0.41786743515850144,
"grad_norm": 1.1702419519424438,
"learning_rate": 0.00027976460014874087,
"loss": 3.2500679016113283,
"step": 2175
},
{
"epoch": 0.41882804995196926,
"grad_norm": 1.2696362733840942,
"learning_rate": 0.0002796449222208704,
"loss": 3.2472507476806642,
"step": 2180
},
{
"epoch": 0.4197886647454371,
"grad_norm": 1.1748182773590088,
"learning_rate": 0.0002795249171855613,
"loss": 3.245960998535156,
"step": 2185
},
{
"epoch": 0.4207492795389049,
"grad_norm": 1.2495172023773193,
"learning_rate": 0.00027940458534559846,
"loss": 3.2419815063476562,
"step": 2190
},
{
"epoch": 0.4217098943323727,
"grad_norm": 1.2086557149887085,
"learning_rate": 0.0002792839270045916,
"loss": 3.2496551513671874,
"step": 2195
},
{
"epoch": 0.42267050912584053,
"grad_norm": 0.9928983449935913,
"learning_rate": 0.0002791629424669739,
"loss": 3.2451755523681642,
"step": 2200
},
{
"epoch": 0.42363112391930835,
"grad_norm": 1.264974594116211,
"learning_rate": 0.00027904163203800185,
"loss": 3.2482452392578125,
"step": 2205
},
{
"epoch": 0.42459173871277617,
"grad_norm": 1.2751965522766113,
"learning_rate": 0.0002789199960237542,
"loss": 3.251111602783203,
"step": 2210
},
{
"epoch": 0.425552353506244,
"grad_norm": 1.115212082862854,
"learning_rate": 0.0002787980347311309,
"loss": 3.2442550659179688,
"step": 2215
},
{
"epoch": 0.4265129682997118,
"grad_norm": 1.3036905527114868,
"learning_rate": 0.00027867574846785295,
"loss": 3.250843048095703,
"step": 2220
},
{
"epoch": 0.4274735830931796,
"grad_norm": 1.3169866800308228,
"learning_rate": 0.0002785531375424611,
"loss": 3.2423397064208985,
"step": 2225
},
{
"epoch": 0.42843419788664744,
"grad_norm": 1.1663439273834229,
"learning_rate": 0.00027843020226431535,
"loss": 3.2481651306152344,
"step": 2230
},
{
"epoch": 0.42939481268011526,
"grad_norm": 1.014664649963379,
"learning_rate": 0.0002783069429435939,
"loss": 3.245722198486328,
"step": 2235
},
{
"epoch": 0.4303554274735831,
"grad_norm": 1.1078412532806396,
"learning_rate": 0.00027818335989129275,
"loss": 3.2450366973876954,
"step": 2240
},
{
"epoch": 0.4313160422670509,
"grad_norm": 1.0631850957870483,
"learning_rate": 0.00027805945341922457,
"loss": 3.2482345581054686,
"step": 2245
},
{
"epoch": 0.4322766570605187,
"grad_norm": 1.182057499885559,
"learning_rate": 0.0002779352238400181,
"loss": 3.2490921020507812,
"step": 2250
},
{
"epoch": 0.4332372718539865,
"grad_norm": 1.021649718284607,
"learning_rate": 0.0002778106714671174,
"loss": 3.2460681915283205,
"step": 2255
},
{
"epoch": 0.43419788664745435,
"grad_norm": 1.0771024227142334,
"learning_rate": 0.00027768579661478077,
"loss": 3.2438682556152343,
"step": 2260
},
{
"epoch": 0.43515850144092216,
"grad_norm": 1.1866998672485352,
"learning_rate": 0.0002775605995980803,
"loss": 3.2406051635742186,
"step": 2265
},
{
"epoch": 0.43611911623439004,
"grad_norm": 1.2904285192489624,
"learning_rate": 0.00027743508073290097,
"loss": 3.247574234008789,
"step": 2270
},
{
"epoch": 0.43707973102785785,
"grad_norm": 0.990243136882782,
"learning_rate": 0.0002773092403359397,
"loss": 3.2439697265625,
"step": 2275
},
{
"epoch": 0.43804034582132567,
"grad_norm": 0.976154625415802,
"learning_rate": 0.0002771830787247047,
"loss": 3.251276397705078,
"step": 2280
},
{
"epoch": 0.4390009606147935,
"grad_norm": 1.2034424543380737,
"learning_rate": 0.0002770565962175148,
"loss": 3.2444232940673827,
"step": 2285
},
{
"epoch": 0.4399615754082613,
"grad_norm": 1.4228894710540771,
"learning_rate": 0.00027692979313349815,
"loss": 3.2472274780273436,
"step": 2290
},
{
"epoch": 0.4409221902017291,
"grad_norm": 1.3287744522094727,
"learning_rate": 0.000276802669792592,
"loss": 3.247750091552734,
"step": 2295
},
{
"epoch": 0.44188280499519694,
"grad_norm": 1.1008442640304565,
"learning_rate": 0.0002766752265155416,
"loss": 3.2413604736328123,
"step": 2300
},
{
"epoch": 0.44284341978866476,
"grad_norm": 1.3397494554519653,
"learning_rate": 0.0002765474636238994,
"loss": 3.243372344970703,
"step": 2305
},
{
"epoch": 0.4438040345821326,
"grad_norm": 1.2448689937591553,
"learning_rate": 0.00027641938144002434,
"loss": 3.24469108581543,
"step": 2310
},
{
"epoch": 0.4447646493756004,
"grad_norm": 1.2328872680664062,
"learning_rate": 0.0002762909802870809,
"loss": 3.2434837341308596,
"step": 2315
},
{
"epoch": 0.4457252641690682,
"grad_norm": 1.3019057512283325,
"learning_rate": 0.0002761622604890382,
"loss": 3.2452354431152344,
"step": 2320
},
{
"epoch": 0.44668587896253603,
"grad_norm": 1.0617128610610962,
"learning_rate": 0.0002760332223706697,
"loss": 3.243347930908203,
"step": 2325
},
{
"epoch": 0.44764649375600385,
"grad_norm": 1.2202798128128052,
"learning_rate": 0.0002759038662575518,
"loss": 3.2428775787353517,
"step": 2330
},
{
"epoch": 0.44860710854947167,
"grad_norm": 0.8098715543746948,
"learning_rate": 0.0002757741924760631,
"loss": 3.2433307647705076,
"step": 2335
},
{
"epoch": 0.4495677233429395,
"grad_norm": 1.2258954048156738,
"learning_rate": 0.0002756442013533839,
"loss": 3.2418006896972655,
"step": 2340
},
{
"epoch": 0.4505283381364073,
"grad_norm": 1.155310869216919,
"learning_rate": 0.0002755138932174952,
"loss": 3.2446517944335938,
"step": 2345
},
{
"epoch": 0.4514889529298751,
"grad_norm": 0.9408679604530334,
"learning_rate": 0.0002753832683971778,
"loss": 3.2486572265625,
"step": 2350
},
{
"epoch": 0.45244956772334294,
"grad_norm": 1.1427699327468872,
"learning_rate": 0.0002752523272220114,
"loss": 3.2479103088378904,
"step": 2355
},
{
"epoch": 0.45341018251681076,
"grad_norm": 0.9570598602294922,
"learning_rate": 0.00027512107002237415,
"loss": 3.247629165649414,
"step": 2360
},
{
"epoch": 0.4543707973102786,
"grad_norm": 0.8978891968727112,
"learning_rate": 0.0002749894971294414,
"loss": 3.242329406738281,
"step": 2365
},
{
"epoch": 0.4553314121037464,
"grad_norm": 1.0361249446868896,
"learning_rate": 0.000274857608875185,
"loss": 3.243325042724609,
"step": 2370
},
{
"epoch": 0.4562920268972142,
"grad_norm": 0.9221097230911255,
"learning_rate": 0.0002747254055923726,
"loss": 3.2444053649902345,
"step": 2375
},
{
"epoch": 0.457252641690682,
"grad_norm": 1.2986423969268799,
"learning_rate": 0.0002745928876145667,
"loss": 3.2442516326904296,
"step": 2380
},
{
"epoch": 0.45821325648414984,
"grad_norm": 0.9883876442909241,
"learning_rate": 0.0002744600552761236,
"loss": 3.2429885864257812,
"step": 2385
},
{
"epoch": 0.45917387127761766,
"grad_norm": 1.4296432733535767,
"learning_rate": 0.000274326908912193,
"loss": 3.2417251586914064,
"step": 2390
},
{
"epoch": 0.4601344860710855,
"grad_norm": 1.216537594795227,
"learning_rate": 0.00027419344885871685,
"loss": 3.2435916900634765,
"step": 2395
},
{
"epoch": 0.4610951008645533,
"grad_norm": 1.0442991256713867,
"learning_rate": 0.0002740596754524285,
"loss": 3.2400821685791015,
"step": 2400
},
{
"epoch": 0.4620557156580211,
"grad_norm": 1.4031963348388672,
"learning_rate": 0.0002739255890308521,
"loss": 3.2466384887695314,
"step": 2405
},
{
"epoch": 0.46301633045148893,
"grad_norm": 1.301132082939148,
"learning_rate": 0.0002737911899323013,
"loss": 3.241625213623047,
"step": 2410
},
{
"epoch": 0.46397694524495675,
"grad_norm": 1.3836039304733276,
"learning_rate": 0.00027365647849587893,
"loss": 3.2370471954345703,
"step": 2415
},
{
"epoch": 0.46493756003842457,
"grad_norm": 0.9351392984390259,
"learning_rate": 0.0002735214550614757,
"loss": 3.2448040008544923,
"step": 2420
},
{
"epoch": 0.4658981748318924,
"grad_norm": 0.9278304576873779,
"learning_rate": 0.00027338611996976955,
"loss": 3.239888000488281,
"step": 2425
},
{
"epoch": 0.4668587896253602,
"grad_norm": 0.8525119423866272,
"learning_rate": 0.0002732504735622248,
"loss": 3.2448410034179687,
"step": 2430
},
{
"epoch": 0.4678194044188281,
"grad_norm": 1.2899324893951416,
"learning_rate": 0.00027311451618109144,
"loss": 3.2423675537109373,
"step": 2435
},
{
"epoch": 0.4687800192122959,
"grad_norm": 0.8825019598007202,
"learning_rate": 0.00027297824816940365,
"loss": 3.240159606933594,
"step": 2440
},
{
"epoch": 0.4697406340057637,
"grad_norm": 1.301020860671997,
"learning_rate": 0.00027284166987097977,
"loss": 3.24395751953125,
"step": 2445
},
{
"epoch": 0.47070124879923153,
"grad_norm": 0.9350459575653076,
"learning_rate": 0.0002727047816304208,
"loss": 3.2402999877929686,
"step": 2450
},
{
"epoch": 0.47166186359269935,
"grad_norm": 1.1744569540023804,
"learning_rate": 0.00027256758379310975,
"loss": 3.239767837524414,
"step": 2455
},
{
"epoch": 0.47262247838616717,
"grad_norm": 0.9879583716392517,
"learning_rate": 0.00027243007670521086,
"loss": 3.242339324951172,
"step": 2460
},
{
"epoch": 0.473583093179635,
"grad_norm": 0.8559085726737976,
"learning_rate": 0.00027229226071366874,
"loss": 3.241751861572266,
"step": 2465
},
{
"epoch": 0.4745437079731028,
"grad_norm": 1.3251603841781616,
"learning_rate": 0.00027215413616620714,
"loss": 3.2432788848876952,
"step": 2470
},
{
"epoch": 0.4755043227665706,
"grad_norm": 1.1395015716552734,
"learning_rate": 0.00027201570341132855,
"loss": 3.239895248413086,
"step": 2475
},
{
"epoch": 0.47646493756003844,
"grad_norm": 1.1372860670089722,
"learning_rate": 0.0002718769627983129,
"loss": 3.2404502868652343,
"step": 2480
},
{
"epoch": 0.47742555235350626,
"grad_norm": 1.014133334159851,
"learning_rate": 0.00027173791467721714,
"loss": 3.238873291015625,
"step": 2485
},
{
"epoch": 0.4783861671469741,
"grad_norm": 1.2768558263778687,
"learning_rate": 0.00027159855939887386,
"loss": 3.2407760620117188,
"step": 2490
},
{
"epoch": 0.4793467819404419,
"grad_norm": 1.0559134483337402,
"learning_rate": 0.0002714588973148907,
"loss": 3.2391632080078123,
"step": 2495
},
{
"epoch": 0.4803073967339097,
"grad_norm": 1.212760090827942,
"learning_rate": 0.00027131892877764945,
"loss": 3.242381286621094,
"step": 2500
},
{
"epoch": 0.4812680115273775,
"grad_norm": 1.3705852031707764,
"learning_rate": 0.0002711786541403051,
"loss": 3.2453689575195312,
"step": 2505
},
{
"epoch": 0.48222862632084534,
"grad_norm": 1.1016364097595215,
"learning_rate": 0.00027103807375678494,
"loss": 3.240717315673828,
"step": 2510
},
{
"epoch": 0.48318924111431316,
"grad_norm": 1.0440564155578613,
"learning_rate": 0.0002708971879817877,
"loss": 3.2409019470214844,
"step": 2515
},
{
"epoch": 0.484149855907781,
"grad_norm": 1.175846815109253,
"learning_rate": 0.0002707559971707827,
"loss": 3.2366703033447264,
"step": 2520
},
{
"epoch": 0.4851104707012488,
"grad_norm": 1.2363582849502563,
"learning_rate": 0.00027061450168000875,
"loss": 3.235219192504883,
"step": 2525
},
{
"epoch": 0.4860710854947166,
"grad_norm": 1.2251267433166504,
"learning_rate": 0.00027047270186647353,
"loss": 3.2362335205078123,
"step": 2530
},
{
"epoch": 0.48703170028818443,
"grad_norm": 0.8210157752037048,
"learning_rate": 0.00027033059808795253,
"loss": 3.2397132873535157,
"step": 2535
},
{
"epoch": 0.48799231508165225,
"grad_norm": 1.229904055595398,
"learning_rate": 0.0002701881907029881,
"loss": 3.2385082244873047,
"step": 2540
},
{
"epoch": 0.48895292987512007,
"grad_norm": 1.1676335334777832,
"learning_rate": 0.00027004548007088876,
"loss": 3.2385066986083983,
"step": 2545
},
{
"epoch": 0.4899135446685879,
"grad_norm": 1.271478533744812,
"learning_rate": 0.000269902466551728,
"loss": 3.2415950775146483,
"step": 2550
},
{
"epoch": 0.4908741594620557,
"grad_norm": 1.395169973373413,
"learning_rate": 0.00026975915050634367,
"loss": 3.2399360656738283,
"step": 2555
},
{
"epoch": 0.4918347742555235,
"grad_norm": 1.3647792339324951,
"learning_rate": 0.00026961553229633685,
"loss": 3.243121337890625,
"step": 2560
},
{
"epoch": 0.49279538904899134,
"grad_norm": 1.0769857168197632,
"learning_rate": 0.000269471612284071,
"loss": 3.2393962860107424,
"step": 2565
},
{
"epoch": 0.49375600384245916,
"grad_norm": 0.9694387316703796,
"learning_rate": 0.00026932739083267105,
"loss": 3.2386848449707033,
"step": 2570
},
{
"epoch": 0.494716618635927,
"grad_norm": 0.9222965240478516,
"learning_rate": 0.00026918286830602254,
"loss": 3.238786315917969,
"step": 2575
},
{
"epoch": 0.4956772334293948,
"grad_norm": 1.1549302339553833,
"learning_rate": 0.00026903804506877064,
"loss": 3.239440155029297,
"step": 2580
},
{
"epoch": 0.4966378482228626,
"grad_norm": 1.0019875764846802,
"learning_rate": 0.0002688929214863192,
"loss": 3.237133026123047,
"step": 2585
},
{
"epoch": 0.49759846301633043,
"grad_norm": 1.1600396633148193,
"learning_rate": 0.0002687474979248299,
"loss": 3.2444324493408203,
"step": 2590
},
{
"epoch": 0.49855907780979825,
"grad_norm": 1.2407069206237793,
"learning_rate": 0.00026860177475122133,
"loss": 3.2405441284179686,
"step": 2595
},
{
"epoch": 0.49951969260326606,
"grad_norm": 1.313706874847412,
"learning_rate": 0.00026845575233316806,
"loss": 3.2400081634521483,
"step": 2600
},
{
"epoch": 0.5004803073967339,
"grad_norm": 1.102402925491333,
"learning_rate": 0.00026830943103909954,
"loss": 3.236441421508789,
"step": 2605
},
{
"epoch": 0.5014409221902018,
"grad_norm": 1.4062278270721436,
"learning_rate": 0.00026816281123819946,
"loss": 3.24122314453125,
"step": 2610
},
{
"epoch": 0.5024015369836695,
"grad_norm": 1.1859440803527832,
"learning_rate": 0.00026801589330040465,
"loss": 3.238700103759766,
"step": 2615
},
{
"epoch": 0.5033621517771374,
"grad_norm": 0.8828426599502563,
"learning_rate": 0.000267868677596404,
"loss": 3.2366287231445314,
"step": 2620
},
{
"epoch": 0.5043227665706052,
"grad_norm": 1.1908822059631348,
"learning_rate": 0.000267721164497638,
"loss": 3.2335018157958983,
"step": 2625
},
{
"epoch": 0.505283381364073,
"grad_norm": 1.3185487985610962,
"learning_rate": 0.00026757335437629725,
"loss": 3.2354408264160157,
"step": 2630
},
{
"epoch": 0.5062439961575408,
"grad_norm": 1.1835713386535645,
"learning_rate": 0.00026742524760532183,
"loss": 3.235929489135742,
"step": 2635
},
{
"epoch": 0.5072046109510087,
"grad_norm": 1.6619583368301392,
"learning_rate": 0.00026727684455840037,
"loss": 3.236317443847656,
"step": 2640
},
{
"epoch": 0.5081652257444764,
"grad_norm": 1.0703696012496948,
"learning_rate": 0.00026712814560996896,
"loss": 3.2386764526367187,
"step": 2645
},
{
"epoch": 0.5091258405379443,
"grad_norm": 1.149133324623108,
"learning_rate": 0.0002669791511352103,
"loss": 3.23681526184082,
"step": 2650
},
{
"epoch": 0.5100864553314121,
"grad_norm": 1.1290984153747559,
"learning_rate": 0.0002668298615100527,
"loss": 3.2398101806640627,
"step": 2655
},
{
"epoch": 0.5110470701248799,
"grad_norm": 1.0490314960479736,
"learning_rate": 0.0002666802771111693,
"loss": 3.238353729248047,
"step": 2660
},
{
"epoch": 0.5120076849183477,
"grad_norm": 0.849686861038208,
"learning_rate": 0.0002665303983159767,
"loss": 3.2377159118652346,
"step": 2665
},
{
"epoch": 0.5129682997118156,
"grad_norm": 1.13809072971344,
"learning_rate": 0.00026638022550263467,
"loss": 3.2373073577880858,
"step": 2670
},
{
"epoch": 0.5139289145052833,
"grad_norm": 1.1344239711761475,
"learning_rate": 0.0002662297590500445,
"loss": 3.233965301513672,
"step": 2675
},
{
"epoch": 0.5148895292987512,
"grad_norm": 1.3232028484344482,
"learning_rate": 0.0002660789993378486,
"loss": 3.236408233642578,
"step": 2680
},
{
"epoch": 0.515850144092219,
"grad_norm": 1.2224704027175903,
"learning_rate": 0.0002659279467464291,
"loss": 3.238209915161133,
"step": 2685
},
{
"epoch": 0.5168107588856868,
"grad_norm": 1.106103777885437,
"learning_rate": 0.0002657766016569072,
"loss": 3.2342296600341798,
"step": 2690
},
{
"epoch": 0.5177713736791547,
"grad_norm": 0.8532865643501282,
"learning_rate": 0.00026562496445114205,
"loss": 3.238364410400391,
"step": 2695
},
{
"epoch": 0.5187319884726225,
"grad_norm": 0.8444566130638123,
"learning_rate": 0.00026547303551172996,
"loss": 3.236472320556641,
"step": 2700
},
{
"epoch": 0.5196926032660903,
"grad_norm": 0.8221146464347839,
"learning_rate": 0.0002653208152220032,
"loss": 3.2367286682128906,
"step": 2705
},
{
"epoch": 0.5206532180595581,
"grad_norm": 1.1430490016937256,
"learning_rate": 0.00026516830396602914,
"loss": 3.2309627532958984,
"step": 2710
},
{
"epoch": 0.521613832853026,
"grad_norm": 1.3973969221115112,
"learning_rate": 0.00026501550212860937,
"loss": 3.2348331451416015,
"step": 2715
},
{
"epoch": 0.5225744476464937,
"grad_norm": 1.2749481201171875,
"learning_rate": 0.0002648624100952786,
"loss": 3.2363750457763674,
"step": 2720
},
{
"epoch": 0.5235350624399616,
"grad_norm": 1.2370373010635376,
"learning_rate": 0.00026470902825230367,
"loss": 3.2376075744628907,
"step": 2725
},
{
"epoch": 0.5244956772334294,
"grad_norm": 1.0846977233886719,
"learning_rate": 0.0002645553569866828,
"loss": 3.2337451934814454,
"step": 2730
},
{
"epoch": 0.5254562920268973,
"grad_norm": 1.026764154434204,
"learning_rate": 0.00026440139668614427,
"loss": 3.237920379638672,
"step": 2735
},
{
"epoch": 0.526416906820365,
"grad_norm": 1.045480489730835,
"learning_rate": 0.00026424714773914574,
"loss": 3.231882858276367,
"step": 2740
},
{
"epoch": 0.5273775216138329,
"grad_norm": 0.9559823274612427,
"learning_rate": 0.0002640926105348732,
"loss": 3.23189697265625,
"step": 2745
},
{
"epoch": 0.5283381364073007,
"grad_norm": 1.3470317125320435,
"learning_rate": 0.00026393778546323976,
"loss": 3.2387470245361327,
"step": 2750
},
{
"epoch": 0.5292987512007685,
"grad_norm": 1.1389282941818237,
"learning_rate": 0.00026378267291488506,
"loss": 3.2349002838134764,
"step": 2755
},
{
"epoch": 0.5302593659942363,
"grad_norm": 1.0445148944854736,
"learning_rate": 0.00026362727328117384,
"loss": 3.2324363708496096,
"step": 2760
},
{
"epoch": 0.5312199807877042,
"grad_norm": 1.0291074514389038,
"learning_rate": 0.00026347158695419546,
"loss": 3.2345924377441406,
"step": 2765
},
{
"epoch": 0.5321805955811719,
"grad_norm": 1.23056161403656,
"learning_rate": 0.00026331561432676244,
"loss": 3.2309837341308594,
"step": 2770
},
{
"epoch": 0.5331412103746398,
"grad_norm": 0.658456563949585,
"learning_rate": 0.0002631593557924097,
"loss": 3.232289123535156,
"step": 2775
},
{
"epoch": 0.5341018251681076,
"grad_norm": 1.1292603015899658,
"learning_rate": 0.0002630028117453936,
"loss": 3.2309852600097657,
"step": 2780
},
{
"epoch": 0.5350624399615754,
"grad_norm": 0.8553731441497803,
"learning_rate": 0.00026284598258069074,
"loss": 3.235504150390625,
"step": 2785
},
{
"epoch": 0.5360230547550432,
"grad_norm": 0.8991851210594177,
"learning_rate": 0.00026268886869399723,
"loss": 3.233042907714844,
"step": 2790
},
{
"epoch": 0.5369836695485111,
"grad_norm": 0.916896402835846,
"learning_rate": 0.0002625314704817276,
"loss": 3.2310935974121096,
"step": 2795
},
{
"epoch": 0.5379442843419788,
"grad_norm": 1.1668161153793335,
"learning_rate": 0.0002623737883410136,
"loss": 3.232282257080078,
"step": 2800
},
{
"epoch": 0.5389048991354467,
"grad_norm": 1.223141074180603,
"learning_rate": 0.00026221582266970346,
"loss": 3.2328109741210938,
"step": 2805
},
{
"epoch": 0.5398655139289145,
"grad_norm": 1.0517135858535767,
"learning_rate": 0.00026205757386636085,
"loss": 3.2323383331298827,
"step": 2810
},
{
"epoch": 0.5408261287223823,
"grad_norm": 0.7666857242584229,
"learning_rate": 0.0002618990423302636,
"loss": 3.234081268310547,
"step": 2815
},
{
"epoch": 0.5417867435158501,
"grad_norm": 1.253327488899231,
"learning_rate": 0.0002617402284614031,
"loss": 3.233143997192383,
"step": 2820
},
{
"epoch": 0.542747358309318,
"grad_norm": 0.7495427131652832,
"learning_rate": 0.00026158113266048317,
"loss": 3.236143112182617,
"step": 2825
},
{
"epoch": 0.5437079731027857,
"grad_norm": 1.1047379970550537,
"learning_rate": 0.0002614217553289186,
"loss": 3.234874725341797,
"step": 2830
},
{
"epoch": 0.5446685878962536,
"grad_norm": 1.2154449224472046,
"learning_rate": 0.0002612620968688349,
"loss": 3.2348342895507813,
"step": 2835
},
{
"epoch": 0.5456292026897214,
"grad_norm": 1.1259676218032837,
"learning_rate": 0.0002611021576830667,
"loss": 3.232221221923828,
"step": 2840
},
{
"epoch": 0.5465898174831892,
"grad_norm": 1.2192593812942505,
"learning_rate": 0.00026094193817515697,
"loss": 3.2312171936035154,
"step": 2845
},
{
"epoch": 0.547550432276657,
"grad_norm": 1.3673197031021118,
"learning_rate": 0.00026078143874935604,
"loss": 3.229291534423828,
"step": 2850
},
{
"epoch": 0.5485110470701249,
"grad_norm": 1.3008506298065186,
"learning_rate": 0.00026062065981062024,
"loss": 3.2330032348632813,
"step": 2855
},
{
"epoch": 0.5494716618635928,
"grad_norm": 1.1764352321624756,
"learning_rate": 0.00026045960176461155,
"loss": 3.232608413696289,
"step": 2860
},
{
"epoch": 0.5504322766570605,
"grad_norm": 1.0877325534820557,
"learning_rate": 0.0002602982650176958,
"loss": 3.232305908203125,
"step": 2865
},
{
"epoch": 0.5513928914505284,
"grad_norm": 1.1681047677993774,
"learning_rate": 0.0002601366499769422,
"loss": 3.233287811279297,
"step": 2870
},
{
"epoch": 0.5523535062439962,
"grad_norm": 1.0753793716430664,
"learning_rate": 0.00025997475705012203,
"loss": 3.2348102569580077,
"step": 2875
},
{
"epoch": 0.553314121037464,
"grad_norm": 0.9505957365036011,
"learning_rate": 0.0002598125866457078,
"loss": 3.2324668884277346,
"step": 2880
},
{
"epoch": 0.5542747358309318,
"grad_norm": 1.2333307266235352,
"learning_rate": 0.0002596501391728721,
"loss": 3.235234832763672,
"step": 2885
},
{
"epoch": 0.5552353506243997,
"grad_norm": 1.0360201597213745,
"learning_rate": 0.00025948741504148656,
"loss": 3.234451675415039,
"step": 2890
},
{
"epoch": 0.5561959654178674,
"grad_norm": 0.9275029301643372,
"learning_rate": 0.0002593244146621208,
"loss": 3.2318180084228514,
"step": 2895
},
{
"epoch": 0.5571565802113353,
"grad_norm": 0.8183401823043823,
"learning_rate": 0.00025916113844604157,
"loss": 3.2336944580078124,
"step": 2900
},
{
"epoch": 0.5581171950048031,
"grad_norm": 0.7255450487136841,
"learning_rate": 0.0002589975868052114,
"loss": 3.2311965942382814,
"step": 2905
},
{
"epoch": 0.5590778097982709,
"grad_norm": 1.1232422590255737,
"learning_rate": 0.000258833760152288,
"loss": 3.2349414825439453,
"step": 2910
},
{
"epoch": 0.5600384245917387,
"grad_norm": 1.311683177947998,
"learning_rate": 0.00025866965890062276,
"loss": 3.233409881591797,
"step": 2915
},
{
"epoch": 0.5609990393852066,
"grad_norm": 1.104836106300354,
"learning_rate": 0.0002585052834642599,
"loss": 3.228401947021484,
"step": 2920
},
{
"epoch": 0.5619596541786743,
"grad_norm": 1.170188069343567,
"learning_rate": 0.0002583406342579357,
"loss": 3.2292003631591797,
"step": 2925
},
{
"epoch": 0.5629202689721422,
"grad_norm": 1.1829124689102173,
"learning_rate": 0.0002581757116970768,
"loss": 3.2359298706054687,
"step": 2930
},
{
"epoch": 0.56388088376561,
"grad_norm": 0.885960042476654,
"learning_rate": 0.00025801051619779987,
"loss": 3.2320457458496095,
"step": 2935
},
{
"epoch": 0.5648414985590778,
"grad_norm": 1.0792263746261597,
"learning_rate": 0.0002578450481769099,
"loss": 3.2303844451904298,
"step": 2940
},
{
"epoch": 0.5658021133525456,
"grad_norm": 1.2442518472671509,
"learning_rate": 0.0002576793080519,
"loss": 3.2286468505859376,
"step": 2945
},
{
"epoch": 0.5667627281460135,
"grad_norm": 0.976241409778595,
"learning_rate": 0.00025751329624094926,
"loss": 3.226253128051758,
"step": 2950
},
{
"epoch": 0.5677233429394812,
"grad_norm": 1.2932735681533813,
"learning_rate": 0.00025734701316292263,
"loss": 3.2291339874267577,
"step": 2955
},
{
"epoch": 0.5686839577329491,
"grad_norm": 1.17381751537323,
"learning_rate": 0.0002571804592373693,
"loss": 3.2318046569824217,
"step": 2960
},
{
"epoch": 0.5696445725264169,
"grad_norm": 1.1502513885498047,
"learning_rate": 0.00025701363488452196,
"loss": 3.230632781982422,
"step": 2965
},
{
"epoch": 0.5706051873198847,
"grad_norm": 1.2376641035079956,
"learning_rate": 0.00025684654052529556,
"loss": 3.2285648345947267,
"step": 2970
},
{
"epoch": 0.5715658021133525,
"grad_norm": 1.2463771104812622,
"learning_rate": 0.0002566791765812862,
"loss": 3.232989501953125,
"step": 2975
},
{
"epoch": 0.5725264169068204,
"grad_norm": 1.148836612701416,
"learning_rate": 0.0002565115434747705,
"loss": 3.2307807922363283,
"step": 2980
},
{
"epoch": 0.5734870317002881,
"grad_norm": 1.313881516456604,
"learning_rate": 0.00025634364162870386,
"loss": 3.2307106018066407,
"step": 2985
},
{
"epoch": 0.574447646493756,
"grad_norm": 1.0571279525756836,
"learning_rate": 0.0002561754714667198,
"loss": 3.230280303955078,
"step": 2990
},
{
"epoch": 0.5754082612872238,
"grad_norm": 0.9680122137069702,
"learning_rate": 0.000256007033413129,
"loss": 3.2283321380615235,
"step": 2995
},
{
"epoch": 0.5763688760806917,
"grad_norm": 1.11472749710083,
"learning_rate": 0.00025583832789291794,
"loss": 3.229805755615234,
"step": 3000
},
{
"epoch": 0.5773294908741594,
"grad_norm": 1.0987354516983032,
"learning_rate": 0.0002556693553317479,
"loss": 3.2307418823242187,
"step": 3005
},
{
"epoch": 0.5782901056676273,
"grad_norm": 0.8845497369766235,
"learning_rate": 0.0002555001161559539,
"loss": 3.226715850830078,
"step": 3010
},
{
"epoch": 0.579250720461095,
"grad_norm": 1.37449049949646,
"learning_rate": 0.000255330610792544,
"loss": 3.230916976928711,
"step": 3015
},
{
"epoch": 0.5802113352545629,
"grad_norm": 1.2311089038848877,
"learning_rate": 0.00025516083966919744,
"loss": 3.228615570068359,
"step": 3020
},
{
"epoch": 0.5811719500480308,
"grad_norm": 1.0169094800949097,
"learning_rate": 0.00025499080321426417,
"loss": 3.2289390563964844,
"step": 3025
},
{
"epoch": 0.5821325648414986,
"grad_norm": 1.280392050743103,
"learning_rate": 0.00025482050185676367,
"loss": 3.227009963989258,
"step": 3030
},
{
"epoch": 0.5830931796349664,
"grad_norm": 0.9485029578208923,
"learning_rate": 0.0002546499360263837,
"loss": 3.231037139892578,
"step": 3035
},
{
"epoch": 0.5840537944284342,
"grad_norm": 1.0025532245635986,
"learning_rate": 0.0002544791061534794,
"loss": 3.231890869140625,
"step": 3040
},
{
"epoch": 0.5850144092219021,
"grad_norm": 0.9485952854156494,
"learning_rate": 0.000254308012669072,
"loss": 3.2329727172851563,
"step": 3045
},
{
"epoch": 0.5859750240153698,
"grad_norm": 0.9976146817207336,
"learning_rate": 0.00025413665600484796,
"loss": 3.228594207763672,
"step": 3050
},
{
"epoch": 0.5869356388088377,
"grad_norm": 0.9106408953666687,
"learning_rate": 0.0002539650365931577,
"loss": 3.225025177001953,
"step": 3055
},
{
"epoch": 0.5878962536023055,
"grad_norm": 0.823363721370697,
"learning_rate": 0.00025379315486701454,
"loss": 3.227878189086914,
"step": 3060
},
{
"epoch": 0.5888568683957733,
"grad_norm": 0.9073953032493591,
"learning_rate": 0.00025362101126009387,
"loss": 3.2288444519042967,
"step": 3065
},
{
"epoch": 0.5898174831892411,
"grad_norm": 1.0130420923233032,
"learning_rate": 0.0002534486062067316,
"loss": 3.228134536743164,
"step": 3070
},
{
"epoch": 0.590778097982709,
"grad_norm": 0.8417114615440369,
"learning_rate": 0.00025327594014192326,
"loss": 3.2314247131347655,
"step": 3075
},
{
"epoch": 0.5917387127761767,
"grad_norm": 0.8221418261528015,
"learning_rate": 0.0002531030135013232,
"loss": 3.226838302612305,
"step": 3080
},
{
"epoch": 0.5926993275696446,
"grad_norm": 1.0958610773086548,
"learning_rate": 0.0002529298267212429,
"loss": 3.2264358520507814,
"step": 3085
},
{
"epoch": 0.5936599423631124,
"grad_norm": 1.3347340822219849,
"learning_rate": 0.00025275638023865055,
"loss": 3.2262367248535155,
"step": 3090
},
{
"epoch": 0.5946205571565802,
"grad_norm": 0.8695631623268127,
"learning_rate": 0.0002525826744911693,
"loss": 3.228030776977539,
"step": 3095
},
{
"epoch": 0.595581171950048,
"grad_norm": 0.8885836601257324,
"learning_rate": 0.00025240870991707665,
"loss": 3.229400634765625,
"step": 3100
},
{
"epoch": 0.5965417867435159,
"grad_norm": 0.8682327270507812,
"learning_rate": 0.000252234486955303,
"loss": 3.2293159484863283,
"step": 3105
},
{
"epoch": 0.5975024015369836,
"grad_norm": 1.130051851272583,
"learning_rate": 0.00025206000604543083,
"loss": 3.2304660797119142,
"step": 3110
},
{
"epoch": 0.5984630163304515,
"grad_norm": 1.0737762451171875,
"learning_rate": 0.00025188526762769326,
"loss": 3.2278984069824217,
"step": 3115
},
{
"epoch": 0.5994236311239193,
"grad_norm": 1.096584677696228,
"learning_rate": 0.0002517102721429734,
"loss": 3.2275615692138673,
"step": 3120
},
{
"epoch": 0.6003842459173871,
"grad_norm": 1.1471421718597412,
"learning_rate": 0.00025153502003280267,
"loss": 3.225168991088867,
"step": 3125
},
{
"epoch": 0.6013448607108549,
"grad_norm": 0.8774202466011047,
"learning_rate": 0.0002513595117393602,
"loss": 3.225673294067383,
"step": 3130
},
{
"epoch": 0.6023054755043228,
"grad_norm": 1.1272785663604736,
"learning_rate": 0.00025118374770547136,
"loss": 3.228855514526367,
"step": 3135
},
{
"epoch": 0.6032660902977905,
"grad_norm": 1.0108392238616943,
"learning_rate": 0.0002510077283746069,
"loss": 3.226605224609375,
"step": 3140
},
{
"epoch": 0.6042267050912584,
"grad_norm": 0.9391249418258667,
"learning_rate": 0.00025083145419088165,
"loss": 3.2272396087646484,
"step": 3145
},
{
"epoch": 0.6051873198847262,
"grad_norm": 1.0647284984588623,
"learning_rate": 0.0002506549255990534,
"loss": 3.2265804290771483,
"step": 3150
},
{
"epoch": 0.6061479346781941,
"grad_norm": 1.143602967262268,
"learning_rate": 0.00025047814304452197,
"loss": 3.2242530822753905,
"step": 3155
},
{
"epoch": 0.6071085494716618,
"grad_norm": 1.098870873451233,
"learning_rate": 0.0002503011069733278,
"loss": 3.225702667236328,
"step": 3160
},
{
"epoch": 0.6080691642651297,
"grad_norm": 1.1901029348373413,
"learning_rate": 0.00025012381783215117,
"loss": 3.2249305725097654,
"step": 3165
},
{
"epoch": 0.6090297790585975,
"grad_norm": 1.254163384437561,
"learning_rate": 0.0002499462760683106,
"loss": 3.2239353179931642,
"step": 3170
},
{
"epoch": 0.6099903938520653,
"grad_norm": 1.3494505882263184,
"learning_rate": 0.00024976848212976237,
"loss": 3.225982666015625,
"step": 3175
},
{
"epoch": 0.6109510086455331,
"grad_norm": 0.9600538015365601,
"learning_rate": 0.00024959043646509866,
"loss": 3.2219696044921875,
"step": 3180
},
{
"epoch": 0.611911623439001,
"grad_norm": 1.1406611204147339,
"learning_rate": 0.00024941213952354703,
"loss": 3.225663757324219,
"step": 3185
},
{
"epoch": 0.6128722382324687,
"grad_norm": 1.276875376701355,
"learning_rate": 0.000249233591754969,
"loss": 3.226406478881836,
"step": 3190
},
{
"epoch": 0.6138328530259366,
"grad_norm": 1.0921087265014648,
"learning_rate": 0.00024905479360985884,
"loss": 3.227063751220703,
"step": 3195
},
{
"epoch": 0.6147934678194045,
"grad_norm": 1.0305696725845337,
"learning_rate": 0.0002488757455393426,
"loss": 3.2252223968505858,
"step": 3200
},
{
"epoch": 0.6157540826128722,
"grad_norm": 1.0612218379974365,
"learning_rate": 0.0002486964479951769,
"loss": 3.2239383697509765,
"step": 3205
},
{
"epoch": 0.6167146974063401,
"grad_norm": 1.011340856552124,
"learning_rate": 0.00024851690142974796,
"loss": 3.2294452667236326,
"step": 3210
},
{
"epoch": 0.6176753121998079,
"grad_norm": 1.0465643405914307,
"learning_rate": 0.0002483371062960701,
"loss": 3.2301746368408204,
"step": 3215
},
{
"epoch": 0.6186359269932757,
"grad_norm": 0.8956519961357117,
"learning_rate": 0.0002481570630477849,
"loss": 3.226789093017578,
"step": 3220
},
{
"epoch": 0.6195965417867435,
"grad_norm": 0.8143766522407532,
"learning_rate": 0.00024797677213915997,
"loss": 3.2227046966552733,
"step": 3225
},
{
"epoch": 0.6205571565802114,
"grad_norm": 1.0714408159255981,
"learning_rate": 0.00024779623402508766,
"loss": 3.225188064575195,
"step": 3230
},
{
"epoch": 0.6215177713736791,
"grad_norm": 1.0253653526306152,
"learning_rate": 0.0002476154491610843,
"loss": 3.225335693359375,
"step": 3235
},
{
"epoch": 0.622478386167147,
"grad_norm": 1.0260114669799805,
"learning_rate": 0.00024743441800328866,
"loss": 3.2250225067138674,
"step": 3240
},
{
"epoch": 0.6234390009606148,
"grad_norm": 1.0167043209075928,
"learning_rate": 0.0002472531410084607,
"loss": 3.222640609741211,
"step": 3245
},
{
"epoch": 0.6243996157540826,
"grad_norm": 1.0281862020492554,
"learning_rate": 0.00024707161863398114,
"loss": 3.223117446899414,
"step": 3250
},
{
"epoch": 0.6253602305475504,
"grad_norm": 1.1766680479049683,
"learning_rate": 0.0002468898513378494,
"loss": 3.2250850677490233,
"step": 3255
},
{
"epoch": 0.6263208453410183,
"grad_norm": 0.9479981064796448,
"learning_rate": 0.0002467078395786831,
"loss": 3.2229190826416017,
"step": 3260
},
{
"epoch": 0.627281460134486,
"grad_norm": 0.9690991640090942,
"learning_rate": 0.00024652558381571646,
"loss": 3.2238792419433593,
"step": 3265
},
{
"epoch": 0.6282420749279539,
"grad_norm": 1.127388596534729,
"learning_rate": 0.00024634308450879953,
"loss": 3.22578125,
"step": 3270
},
{
"epoch": 0.6292026897214217,
"grad_norm": 1.0700501203536987,
"learning_rate": 0.0002461603421183968,
"loss": 3.222941589355469,
"step": 3275
},
{
"epoch": 0.6301633045148896,
"grad_norm": 1.1317108869552612,
"learning_rate": 0.00024597735710558613,
"loss": 3.228685760498047,
"step": 3280
},
{
"epoch": 0.6311239193083573,
"grad_norm": 1.0256825685501099,
"learning_rate": 0.0002457941299320574,
"loss": 3.2236000061035157,
"step": 3285
},
{
"epoch": 0.6320845341018252,
"grad_norm": 1.0949205160140991,
"learning_rate": 0.0002456106610601115,
"loss": 3.223644256591797,
"step": 3290
},
{
"epoch": 0.633045148895293,
"grad_norm": 1.147100567817688,
"learning_rate": 0.0002454269509526593,
"loss": 3.222336196899414,
"step": 3295
},
{
"epoch": 0.6340057636887608,
"grad_norm": 1.0118300914764404,
"learning_rate": 0.0002452430000732203,
"loss": 3.2236583709716795,
"step": 3300
},
{
"epoch": 0.6349663784822286,
"grad_norm": 1.4460432529449463,
"learning_rate": 0.00024505880888592134,
"loss": 3.225933074951172,
"step": 3305
},
{
"epoch": 0.6359269932756965,
"grad_norm": 1.0727514028549194,
"learning_rate": 0.0002448743778554957,
"loss": 3.222777557373047,
"step": 3310
},
{
"epoch": 0.6368876080691642,
"grad_norm": 0.9514535069465637,
"learning_rate": 0.0002446897074472819,
"loss": 3.2218833923339845,
"step": 3315
},
{
"epoch": 0.6378482228626321,
"grad_norm": 0.8115454912185669,
"learning_rate": 0.0002445047981272221,
"loss": 3.2228424072265627,
"step": 3320
},
{
"epoch": 0.6388088376560999,
"grad_norm": 1.2015719413757324,
"learning_rate": 0.00024431965036186175,
"loss": 3.2233604431152343,
"step": 3325
},
{
"epoch": 0.6397694524495677,
"grad_norm": 1.0239508152008057,
"learning_rate": 0.00024413426461834746,
"loss": 3.223210906982422,
"step": 3330
},
{
"epoch": 0.6407300672430355,
"grad_norm": 0.8938039541244507,
"learning_rate": 0.0002439486413644265,
"loss": 3.2242591857910154,
"step": 3335
},
{
"epoch": 0.6416906820365034,
"grad_norm": 1.0734831094741821,
"learning_rate": 0.00024376278106844548,
"loss": 3.221482849121094,
"step": 3340
},
{
"epoch": 0.6426512968299711,
"grad_norm": 0.9349012970924377,
"learning_rate": 0.00024357668419934894,
"loss": 3.2233955383300783,
"step": 3345
},
{
"epoch": 0.643611911623439,
"grad_norm": 1.0495328903198242,
"learning_rate": 0.00024339035122667837,
"loss": 3.2246463775634764,
"step": 3350
},
{
"epoch": 0.6445725264169068,
"grad_norm": 1.152726173400879,
"learning_rate": 0.00024320378262057102,
"loss": 3.221533203125,
"step": 3355
},
{
"epoch": 0.6455331412103746,
"grad_norm": 1.1071127653121948,
"learning_rate": 0.00024301697885175847,
"loss": 3.2202079772949217,
"step": 3360
},
{
"epoch": 0.6464937560038425,
"grad_norm": 1.143314003944397,
"learning_rate": 0.00024282994039156598,
"loss": 3.228096771240234,
"step": 3365
},
{
"epoch": 0.6474543707973103,
"grad_norm": 1.0042517185211182,
"learning_rate": 0.00024264266771191066,
"loss": 3.2260887145996096,
"step": 3370
},
{
"epoch": 0.6484149855907781,
"grad_norm": 0.9083545804023743,
"learning_rate": 0.00024245516128530077,
"loss": 3.2202564239501954,
"step": 3375
},
{
"epoch": 0.6493756003842459,
"grad_norm": 0.8550533056259155,
"learning_rate": 0.0002422674215848342,
"loss": 3.22034912109375,
"step": 3380
},
{
"epoch": 0.6503362151777138,
"grad_norm": 0.9592113494873047,
"learning_rate": 0.0002420794490841975,
"loss": 3.22401123046875,
"step": 3385
},
{
"epoch": 0.6512968299711815,
"grad_norm": 0.8267397880554199,
"learning_rate": 0.00024189124425766454,
"loss": 3.2178253173828124,
"step": 3390
},
{
"epoch": 0.6522574447646494,
"grad_norm": 0.9734513759613037,
"learning_rate": 0.00024170280758009546,
"loss": 3.2190723419189453,
"step": 3395
},
{
"epoch": 0.6532180595581172,
"grad_norm": 1.0415617227554321,
"learning_rate": 0.00024151413952693524,
"loss": 3.2226608276367186,
"step": 3400
},
{
"epoch": 0.654178674351585,
"grad_norm": 0.9090771675109863,
"learning_rate": 0.00024132524057421287,
"loss": 3.221820831298828,
"step": 3405
},
{
"epoch": 0.6551392891450528,
"grad_norm": 0.8799076676368713,
"learning_rate": 0.0002411361111985396,
"loss": 3.2193443298339846,
"step": 3410
},
{
"epoch": 0.6560999039385207,
"grad_norm": 1.0817033052444458,
"learning_rate": 0.00024094675187710847,
"loss": 3.2185131072998048,
"step": 3415
},
{
"epoch": 0.6570605187319885,
"grad_norm": 0.9831625819206238,
"learning_rate": 0.0002407571630876923,
"loss": 3.2214286804199217,
"step": 3420
},
{
"epoch": 0.6580211335254563,
"grad_norm": 0.987235963344574,
"learning_rate": 0.00024056734530864313,
"loss": 3.219521713256836,
"step": 3425
},
{
"epoch": 0.6589817483189241,
"grad_norm": 1.047991394996643,
"learning_rate": 0.00024037729901889068,
"loss": 3.2219139099121095,
"step": 3430
},
{
"epoch": 0.659942363112392,
"grad_norm": 0.8672451376914978,
"learning_rate": 0.00024018702469794125,
"loss": 3.2201393127441404,
"step": 3435
},
{
"epoch": 0.6609029779058597,
"grad_norm": 1.1878256797790527,
"learning_rate": 0.0002399965228258764,
"loss": 3.2203353881835937,
"step": 3440
},
{
"epoch": 0.6618635926993276,
"grad_norm": 1.1251111030578613,
"learning_rate": 0.00023980579388335193,
"loss": 3.21944580078125,
"step": 3445
},
{
"epoch": 0.6628242074927954,
"grad_norm": 0.9487177729606628,
"learning_rate": 0.0002396148383515966,
"loss": 3.2208953857421876,
"step": 3450
},
{
"epoch": 0.6637848222862632,
"grad_norm": 0.9404230713844299,
"learning_rate": 0.00023942365671241072,
"loss": 3.22216796875,
"step": 3455
},
{
"epoch": 0.664745437079731,
"grad_norm": 1.1221790313720703,
"learning_rate": 0.00023923224944816522,
"loss": 3.220103454589844,
"step": 3460
},
{
"epoch": 0.6657060518731989,
"grad_norm": 1.1422487497329712,
"learning_rate": 0.00023904061704180027,
"loss": 3.2190887451171877,
"step": 3465
},
{
"epoch": 0.6666666666666666,
"grad_norm": 0.9574413299560547,
"learning_rate": 0.00023884875997682404,
"loss": 3.215955352783203,
"step": 3470
},
{
"epoch": 0.6676272814601345,
"grad_norm": 1.1348673105239868,
"learning_rate": 0.00023865667873731168,
"loss": 3.2160018920898437,
"step": 3475
},
{
"epoch": 0.6685878962536023,
"grad_norm": 1.0261040925979614,
"learning_rate": 0.00023846437380790368,
"loss": 3.221507263183594,
"step": 3480
},
{
"epoch": 0.6695485110470701,
"grad_norm": 1.009157657623291,
"learning_rate": 0.0002382718456738053,
"loss": 3.2182361602783205,
"step": 3485
},
{
"epoch": 0.6705091258405379,
"grad_norm": 0.9334181547164917,
"learning_rate": 0.00023807909482078475,
"loss": 3.221875762939453,
"step": 3490
},
{
"epoch": 0.6714697406340058,
"grad_norm": 1.0333514213562012,
"learning_rate": 0.0002378861217351721,
"loss": 3.2202327728271483,
"step": 3495
},
{
"epoch": 0.6724303554274735,
"grad_norm": 0.8510705828666687,
"learning_rate": 0.00023769292690385831,
"loss": 3.2205406188964845,
"step": 3500
},
{
"epoch": 0.6733909702209414,
"grad_norm": 1.0097713470458984,
"learning_rate": 0.0002374995108142938,
"loss": 3.2197288513183593,
"step": 3505
},
{
"epoch": 0.6743515850144092,
"grad_norm": 1.3777557611465454,
"learning_rate": 0.0002373058739544871,
"loss": 3.2197029113769533,
"step": 3510
},
{
"epoch": 0.675312199807877,
"grad_norm": 1.1440997123718262,
"learning_rate": 0.0002371120168130039,
"loss": 3.2237571716308593,
"step": 3515
},
{
"epoch": 0.6762728146013448,
"grad_norm": 1.1652058362960815,
"learning_rate": 0.0002369179398789657,
"loss": 3.216419982910156,
"step": 3520
},
{
"epoch": 0.6772334293948127,
"grad_norm": 0.7069166302680969,
"learning_rate": 0.00023672364364204853,
"loss": 3.2144538879394533,
"step": 3525
},
{
"epoch": 0.6781940441882806,
"grad_norm": 1.2215242385864258,
"learning_rate": 0.00023652912859248166,
"loss": 3.221971130371094,
"step": 3530
},
{
"epoch": 0.6791546589817483,
"grad_norm": 0.978817880153656,
"learning_rate": 0.00023633439522104658,
"loss": 3.21834716796875,
"step": 3535
},
{
"epoch": 0.6801152737752162,
"grad_norm": 1.0850077867507935,
"learning_rate": 0.0002361394440190755,
"loss": 3.216111755371094,
"step": 3540
},
{
"epoch": 0.681075888568684,
"grad_norm": 1.0549561977386475,
"learning_rate": 0.00023594427547845033,
"loss": 3.2226284027099608,
"step": 3545
},
{
"epoch": 0.6820365033621518,
"grad_norm": 1.0915031433105469,
"learning_rate": 0.00023574889009160134,
"loss": 3.220775604248047,
"step": 3550
},
{
"epoch": 0.6829971181556196,
"grad_norm": 1.2648931741714478,
"learning_rate": 0.00023555328835150587,
"loss": 3.219307708740234,
"step": 3555
},
{
"epoch": 0.6839577329490875,
"grad_norm": 0.8500558733940125,
"learning_rate": 0.00023535747075168726,
"loss": 3.2166213989257812,
"step": 3560
},
{
"epoch": 0.6849183477425552,
"grad_norm": 0.9885119795799255,
"learning_rate": 0.00023516143778621327,
"loss": 3.219788360595703,
"step": 3565
},
{
"epoch": 0.6858789625360231,
"grad_norm": 1.1124404668807983,
"learning_rate": 0.00023496518994969528,
"loss": 3.218476104736328,
"step": 3570
},
{
"epoch": 0.6868395773294909,
"grad_norm": 1.0537521839141846,
"learning_rate": 0.00023476872773728666,
"loss": 3.217354583740234,
"step": 3575
},
{
"epoch": 0.6878001921229587,
"grad_norm": 1.0711948871612549,
"learning_rate": 0.00023457205164468173,
"loss": 3.2162689208984374,
"step": 3580
},
{
"epoch": 0.6887608069164265,
"grad_norm": 1.0931755304336548,
"learning_rate": 0.0002343751621681145,
"loss": 3.217786407470703,
"step": 3585
},
{
"epoch": 0.6897214217098944,
"grad_norm": 0.9976537227630615,
"learning_rate": 0.00023417805980435736,
"loss": 3.216781997680664,
"step": 3590
},
{
"epoch": 0.6906820365033621,
"grad_norm": 0.8797841668128967,
"learning_rate": 0.00023398074505071964,
"loss": 3.21632080078125,
"step": 3595
},
{
"epoch": 0.69164265129683,
"grad_norm": 1.0468313694000244,
"learning_rate": 0.0002337832184050468,
"loss": 3.2173301696777346,
"step": 3600
},
{
"epoch": 0.6926032660902978,
"grad_norm": 1.2071698904037476,
"learning_rate": 0.0002335854803657188,
"loss": 3.2191944122314453,
"step": 3605
},
{
"epoch": 0.6935638808837656,
"grad_norm": 1.0584746599197388,
"learning_rate": 0.00023338753143164906,
"loss": 3.2196414947509764,
"step": 3610
},
{
"epoch": 0.6945244956772334,
"grad_norm": 0.9132346510887146,
"learning_rate": 0.00023318937210228295,
"loss": 3.215652847290039,
"step": 3615
},
{
"epoch": 0.6954851104707013,
"grad_norm": 0.9715785384178162,
"learning_rate": 0.00023299100287759686,
"loss": 3.215423583984375,
"step": 3620
},
{
"epoch": 0.696445725264169,
"grad_norm": 1.1859434843063354,
"learning_rate": 0.00023279242425809667,
"loss": 3.211594009399414,
"step": 3625
},
{
"epoch": 0.6974063400576369,
"grad_norm": 1.0725688934326172,
"learning_rate": 0.00023259363674481666,
"loss": 3.2200057983398436,
"step": 3630
},
{
"epoch": 0.6983669548511047,
"grad_norm": 1.0033189058303833,
"learning_rate": 0.00023239464083931802,
"loss": 3.217108154296875,
"step": 3635
},
{
"epoch": 0.6993275696445725,
"grad_norm": 1.0625147819519043,
"learning_rate": 0.00023219543704368792,
"loss": 3.2169322967529297,
"step": 3640
},
{
"epoch": 0.7002881844380403,
"grad_norm": 0.9788894653320312,
"learning_rate": 0.00023199602586053793,
"loss": 3.2101593017578125,
"step": 3645
},
{
"epoch": 0.7012487992315082,
"grad_norm": 0.9568902254104614,
"learning_rate": 0.0002317964077930029,
"loss": 3.215932846069336,
"step": 3650
},
{
"epoch": 0.7022094140249759,
"grad_norm": 1.3549705743789673,
"learning_rate": 0.00023159658334473974,
"loss": 3.2176162719726564,
"step": 3655
},
{
"epoch": 0.7031700288184438,
"grad_norm": 0.9975321292877197,
"learning_rate": 0.00023139655301992603,
"loss": 3.218878173828125,
"step": 3660
},
{
"epoch": 0.7041306436119116,
"grad_norm": 0.8793084621429443,
"learning_rate": 0.00023119631732325882,
"loss": 3.2155433654785157,
"step": 3665
},
{
"epoch": 0.7050912584053795,
"grad_norm": 1.2263166904449463,
"learning_rate": 0.0002309958767599532,
"loss": 3.2179286956787108,
"step": 3670
},
{
"epoch": 0.7060518731988472,
"grad_norm": 1.0385061502456665,
"learning_rate": 0.0002307952318357414,
"loss": 3.2134330749511717,
"step": 3675
},
{
"epoch": 0.7070124879923151,
"grad_norm": 1.019902229309082,
"learning_rate": 0.0002305943830568711,
"loss": 3.2176845550537108,
"step": 3680
},
{
"epoch": 0.7079731027857828,
"grad_norm": 0.9984716773033142,
"learning_rate": 0.00023039333093010433,
"loss": 3.218429946899414,
"step": 3685
},
{
"epoch": 0.7089337175792507,
"grad_norm": 1.0424875020980835,
"learning_rate": 0.00023019207596271634,
"loss": 3.2165111541748046,
"step": 3690
},
{
"epoch": 0.7098943323727186,
"grad_norm": 1.0895493030548096,
"learning_rate": 0.00022999061866249397,
"loss": 3.2162498474121093,
"step": 3695
},
{
"epoch": 0.7108549471661864,
"grad_norm": 1.0710501670837402,
"learning_rate": 0.00022978895953773464,
"loss": 3.2143184661865236,
"step": 3700
},
{
"epoch": 0.7118155619596542,
"grad_norm": 1.1927725076675415,
"learning_rate": 0.00022958709909724512,
"loss": 3.2180755615234373,
"step": 3705
},
{
"epoch": 0.712776176753122,
"grad_norm": 0.9517913460731506,
"learning_rate": 0.00022938503785033991,
"loss": 3.2159660339355467,
"step": 3710
},
{
"epoch": 0.7137367915465899,
"grad_norm": 1.0126185417175293,
"learning_rate": 0.00022918277630684032,
"loss": 3.2123146057128906,
"step": 3715
},
{
"epoch": 0.7146974063400576,
"grad_norm": 1.0625087022781372,
"learning_rate": 0.0002289803149770729,
"loss": 3.215711212158203,
"step": 3720
},
{
"epoch": 0.7156580211335255,
"grad_norm": 1.2499691247940063,
"learning_rate": 0.00022877765437186843,
"loss": 3.215605545043945,
"step": 3725
},
{
"epoch": 0.7166186359269933,
"grad_norm": 1.0240933895111084,
"learning_rate": 0.0002285747950025604,
"loss": 3.2169639587402346,
"step": 3730
},
{
"epoch": 0.7175792507204611,
"grad_norm": 1.0153264999389648,
"learning_rate": 0.00022837173738098375,
"loss": 3.2149864196777345,
"step": 3735
},
{
"epoch": 0.7185398655139289,
"grad_norm": 0.7960460782051086,
"learning_rate": 0.00022816848201947376,
"loss": 3.217276382446289,
"step": 3740
},
{
"epoch": 0.7195004803073968,
"grad_norm": 0.886752724647522,
"learning_rate": 0.0002279650294308645,
"loss": 3.2126708984375,
"step": 3745
},
{
"epoch": 0.7204610951008645,
"grad_norm": 0.9281694889068604,
"learning_rate": 0.00022776138012848776,
"loss": 3.2142318725585937,
"step": 3750
},
{
"epoch": 0.7214217098943324,
"grad_norm": 0.9548509120941162,
"learning_rate": 0.00022755753462617162,
"loss": 3.2146896362304687,
"step": 3755
},
{
"epoch": 0.7223823246878002,
"grad_norm": 1.099318265914917,
"learning_rate": 0.0002273534934382392,
"loss": 3.2121913909912108,
"step": 3760
},
{
"epoch": 0.723342939481268,
"grad_norm": 1.0480355024337769,
"learning_rate": 0.00022714925707950734,
"loss": 3.2157440185546875,
"step": 3765
},
{
"epoch": 0.7243035542747358,
"grad_norm": 0.8996202349662781,
"learning_rate": 0.0002269448260652853,
"loss": 3.2162132263183594,
"step": 3770
},
{
"epoch": 0.7252641690682037,
"grad_norm": 1.0007187128067017,
"learning_rate": 0.00022674020091137357,
"loss": 3.216075897216797,
"step": 3775
},
{
"epoch": 0.7262247838616714,
"grad_norm": 1.0066853761672974,
"learning_rate": 0.0002265353821340623,
"loss": 3.2114944458007812,
"step": 3780
},
{
"epoch": 0.7271853986551393,
"grad_norm": 0.8113794922828674,
"learning_rate": 0.00022633037025013034,
"loss": 3.212854766845703,
"step": 3785
},
{
"epoch": 0.7281460134486071,
"grad_norm": 0.8382619619369507,
"learning_rate": 0.00022612516577684373,
"loss": 3.2154186248779295,
"step": 3790
},
{
"epoch": 0.729106628242075,
"grad_norm": 0.7889550924301147,
"learning_rate": 0.0002259197692319544,
"loss": 3.212788391113281,
"step": 3795
},
{
"epoch": 0.7300672430355427,
"grad_norm": 0.988488495349884,
"learning_rate": 0.00022571418113369885,
"loss": 3.2164848327636717,
"step": 3800
},
{
"epoch": 0.7310278578290106,
"grad_norm": 1.2173198461532593,
"learning_rate": 0.00022550840200079696,
"loss": 3.216341018676758,
"step": 3805
},
{
"epoch": 0.7319884726224783,
"grad_norm": 0.9247362017631531,
"learning_rate": 0.00022530243235245067,
"loss": 3.214260482788086,
"step": 3810
},
{
"epoch": 0.7329490874159462,
"grad_norm": 1.0379600524902344,
"learning_rate": 0.0002250962727083424,
"loss": 3.211573028564453,
"step": 3815
},
{
"epoch": 0.733909702209414,
"grad_norm": 1.0794739723205566,
"learning_rate": 0.00022488992358863416,
"loss": 3.2143383026123047,
"step": 3820
},
{
"epoch": 0.7348703170028819,
"grad_norm": 0.7588670253753662,
"learning_rate": 0.00022468338551396598,
"loss": 3.2139129638671875,
"step": 3825
},
{
"epoch": 0.7358309317963496,
"grad_norm": 0.9358298182487488,
"learning_rate": 0.00022447665900545452,
"loss": 3.2157051086425783,
"step": 3830
},
{
"epoch": 0.7367915465898175,
"grad_norm": 1.0112359523773193,
"learning_rate": 0.00022426974458469206,
"loss": 3.2114688873291017,
"step": 3835
},
{
"epoch": 0.7377521613832853,
"grad_norm": 1.0056138038635254,
"learning_rate": 0.00022406264277374487,
"loss": 3.211741638183594,
"step": 3840
},
{
"epoch": 0.7387127761767531,
"grad_norm": 1.0002721548080444,
"learning_rate": 0.00022385535409515203,
"loss": 3.214643096923828,
"step": 3845
},
{
"epoch": 0.7396733909702209,
"grad_norm": 1.0511516332626343,
"learning_rate": 0.0002236478790719242,
"loss": 3.2131919860839844,
"step": 3850
},
{
"epoch": 0.7406340057636888,
"grad_norm": 1.0013073682785034,
"learning_rate": 0.00022344021822754212,
"loss": 3.2114883422851563,
"step": 3855
},
{
"epoch": 0.7415946205571565,
"grad_norm": 0.9666048288345337,
"learning_rate": 0.0002232323720859554,
"loss": 3.2139522552490236,
"step": 3860
},
{
"epoch": 0.7425552353506244,
"grad_norm": 0.9765871167182922,
"learning_rate": 0.0002230243411715812,
"loss": 3.216571044921875,
"step": 3865
},
{
"epoch": 0.7435158501440923,
"grad_norm": 0.8558294773101807,
"learning_rate": 0.00022281612600930282,
"loss": 3.211492156982422,
"step": 3870
},
{
"epoch": 0.74447646493756,
"grad_norm": 0.9790803790092468,
"learning_rate": 0.00022260772712446848,
"loss": 3.212940979003906,
"step": 3875
},
{
"epoch": 0.7454370797310279,
"grad_norm": 1.115885853767395,
"learning_rate": 0.00022239914504289002,
"loss": 3.215241241455078,
"step": 3880
},
{
"epoch": 0.7463976945244957,
"grad_norm": 0.8012208342552185,
"learning_rate": 0.00022219038029084134,
"loss": 3.2129531860351563,
"step": 3885
},
{
"epoch": 0.7473583093179635,
"grad_norm": 0.9198108315467834,
"learning_rate": 0.00022198143339505736,
"loss": 3.2131282806396486,
"step": 3890
},
{
"epoch": 0.7483189241114313,
"grad_norm": 0.9384759068489075,
"learning_rate": 0.00022177230488273255,
"loss": 3.2148651123046874,
"step": 3895
},
{
"epoch": 0.7492795389048992,
"grad_norm": 0.9330629706382751,
"learning_rate": 0.00022156299528151957,
"loss": 3.2130111694335937,
"step": 3900
},
{
"epoch": 0.7502401536983669,
"grad_norm": 1.0961406230926514,
"learning_rate": 0.00022135350511952805,
"loss": 3.209431457519531,
"step": 3905
},
{
"epoch": 0.7512007684918348,
"grad_norm": 1.0598336458206177,
"learning_rate": 0.0002211438349253231,
"loss": 3.2150421142578125,
"step": 3910
},
{
"epoch": 0.7521613832853026,
"grad_norm": 0.8614126443862915,
"learning_rate": 0.00022093398522792418,
"loss": 3.2113304138183594,
"step": 3915
},
{
"epoch": 0.7531219980787704,
"grad_norm": 1.2163376808166504,
"learning_rate": 0.0002207239565568036,
"loss": 3.213220977783203,
"step": 3920
},
{
"epoch": 0.7540826128722382,
"grad_norm": 1.068865180015564,
"learning_rate": 0.00022051374944188524,
"loss": 3.2152915954589845,
"step": 3925
},
{
"epoch": 0.7550432276657061,
"grad_norm": 0.8628993034362793,
"learning_rate": 0.0002203033644135432,
"loss": 3.2100101470947267,
"step": 3930
},
{
"epoch": 0.7560038424591738,
"grad_norm": 1.0070104598999023,
"learning_rate": 0.00022009280200260053,
"loss": 3.214314651489258,
"step": 3935
},
{
"epoch": 0.7569644572526417,
"grad_norm": 1.1911405324935913,
"learning_rate": 0.00021988206274032776,
"loss": 3.2116607666015624,
"step": 3940
},
{
"epoch": 0.7579250720461095,
"grad_norm": 1.0605741739273071,
"learning_rate": 0.00021967114715844163,
"loss": 3.210451126098633,
"step": 3945
},
{
"epoch": 0.7588856868395774,
"grad_norm": 1.010385274887085,
"learning_rate": 0.0002194600557891039,
"loss": 3.2117488861083983,
"step": 3950
},
{
"epoch": 0.7598463016330451,
"grad_norm": 0.9285851716995239,
"learning_rate": 0.00021924878916491962,
"loss": 3.2095130920410155,
"step": 3955
},
{
"epoch": 0.760806916426513,
"grad_norm": 0.9447981715202332,
"learning_rate": 0.00021903734781893625,
"loss": 3.214548873901367,
"step": 3960
},
{
"epoch": 0.7617675312199808,
"grad_norm": 1.0844841003417969,
"learning_rate": 0.000218825732284642,
"loss": 3.2142662048339843,
"step": 3965
},
{
"epoch": 0.7627281460134486,
"grad_norm": 0.9654716849327087,
"learning_rate": 0.00021861394309596446,
"loss": 3.214899444580078,
"step": 3970
},
{
"epoch": 0.7636887608069164,
"grad_norm": 0.8219507336616516,
"learning_rate": 0.00021840198078726964,
"loss": 3.2075408935546874,
"step": 3975
},
{
"epoch": 0.7646493756003843,
"grad_norm": 1.0514135360717773,
"learning_rate": 0.00021818984589336006,
"loss": 3.2124794006347654,
"step": 3980
},
{
"epoch": 0.765609990393852,
"grad_norm": 0.764786958694458,
"learning_rate": 0.0002179775389494739,
"loss": 3.210370635986328,
"step": 3985
},
{
"epoch": 0.7665706051873199,
"grad_norm": 0.9463275671005249,
"learning_rate": 0.0002177650604912833,
"loss": 3.214555358886719,
"step": 3990
},
{
"epoch": 0.7675312199807877,
"grad_norm": 0.9087357521057129,
"learning_rate": 0.0002175524110548932,
"loss": 3.211328125,
"step": 3995
},
{
"epoch": 0.7684918347742555,
"grad_norm": 0.8839055299758911,
"learning_rate": 0.00021733959117684008,
"loss": 3.208209991455078,
"step": 4000
},
{
"epoch": 0.7694524495677233,
"grad_norm": 1.0021069049835205,
"learning_rate": 0.00021712660139409015,
"loss": 3.213267517089844,
"step": 4005
},
{
"epoch": 0.7704130643611912,
"grad_norm": 1.089827060699463,
"learning_rate": 0.0002169134422440386,
"loss": 3.209787368774414,
"step": 4010
},
{
"epoch": 0.7713736791546589,
"grad_norm": 0.9058536887168884,
"learning_rate": 0.00021670011426450772,
"loss": 3.2052375793457033,
"step": 4015
},
{
"epoch": 0.7723342939481268,
"grad_norm": 0.8882672190666199,
"learning_rate": 0.000216486617993746,
"loss": 3.2103240966796873,
"step": 4020
},
{
"epoch": 0.7732949087415946,
"grad_norm": 0.7981977462768555,
"learning_rate": 0.00021627295397042635,
"loss": 3.2096282958984377,
"step": 4025
},
{
"epoch": 0.7742555235350624,
"grad_norm": 0.9157530665397644,
"learning_rate": 0.00021605912273364513,
"loss": 3.213287353515625,
"step": 4030
},
{
"epoch": 0.7752161383285303,
"grad_norm": 0.7963248491287231,
"learning_rate": 0.00021584512482292038,
"loss": 3.2098575592041017,
"step": 4035
},
{
"epoch": 0.7761767531219981,
"grad_norm": 1.0376338958740234,
"learning_rate": 0.00021563096077819083,
"loss": 3.2072002410888674,
"step": 4040
},
{
"epoch": 0.777137367915466,
"grad_norm": 1.159191608428955,
"learning_rate": 0.00021541663113981433,
"loss": 3.2127113342285156,
"step": 4045
},
{
"epoch": 0.7780979827089337,
"grad_norm": 1.213563323020935,
"learning_rate": 0.0002152021364485665,
"loss": 3.2094375610351564,
"step": 4050
},
{
"epoch": 0.7790585975024016,
"grad_norm": 1.1174225807189941,
"learning_rate": 0.00021498747724563953,
"loss": 3.2087074279785157,
"step": 4055
},
{
"epoch": 0.7800192122958693,
"grad_norm": 1.0365922451019287,
"learning_rate": 0.00021477265407264051,
"loss": 3.2079147338867187,
"step": 4060
},
{
"epoch": 0.7809798270893372,
"grad_norm": 1.0354453325271606,
"learning_rate": 0.00021455766747159044,
"loss": 3.2089813232421873,
"step": 4065
},
{
"epoch": 0.781940441882805,
"grad_norm": 1.0720642805099487,
"learning_rate": 0.0002143425179849226,
"loss": 3.207681655883789,
"step": 4070
},
{
"epoch": 0.7829010566762729,
"grad_norm": 0.9447735548019409,
"learning_rate": 0.00021412720615548105,
"loss": 3.2113746643066405,
"step": 4075
},
{
"epoch": 0.7838616714697406,
"grad_norm": 0.9376322627067566,
"learning_rate": 0.00021391173252651978,
"loss": 3.2086830139160156,
"step": 4080
},
{
"epoch": 0.7848222862632085,
"grad_norm": 1.0265626907348633,
"learning_rate": 0.00021369609764170075,
"loss": 3.2098388671875,
"step": 4085
},
{
"epoch": 0.7857829010566763,
"grad_norm": 1.1230889558792114,
"learning_rate": 0.00021348030204509303,
"loss": 3.208446502685547,
"step": 4090
},
{
"epoch": 0.7867435158501441,
"grad_norm": 0.8800603747367859,
"learning_rate": 0.00021326434628117088,
"loss": 3.208513641357422,
"step": 4095
},
{
"epoch": 0.7877041306436119,
"grad_norm": 1.0674163103103638,
"learning_rate": 0.00021304823089481298,
"loss": 3.2110557556152344,
"step": 4100
},
{
"epoch": 0.7886647454370798,
"grad_norm": 0.8815765976905823,
"learning_rate": 0.00021283195643130058,
"loss": 3.210008239746094,
"step": 4105
},
{
"epoch": 0.7896253602305475,
"grad_norm": 1.0902719497680664,
"learning_rate": 0.00021261552343631633,
"loss": 3.2107887268066406,
"step": 4110
},
{
"epoch": 0.7905859750240154,
"grad_norm": 0.859643816947937,
"learning_rate": 0.00021239893245594287,
"loss": 3.210626220703125,
"step": 4115
},
{
"epoch": 0.7915465898174832,
"grad_norm": 0.9734321236610413,
"learning_rate": 0.00021218218403666148,
"loss": 3.207859420776367,
"step": 4120
},
{
"epoch": 0.792507204610951,
"grad_norm": 0.8096666932106018,
"learning_rate": 0.00021196527872535068,
"loss": 3.2104820251464843,
"step": 4125
},
{
"epoch": 0.7934678194044188,
"grad_norm": 1.0135759115219116,
"learning_rate": 0.0002117482170692847,
"loss": 3.2095088958740234,
"step": 4130
},
{
"epoch": 0.7944284341978867,
"grad_norm": 0.9503042101860046,
"learning_rate": 0.00021153099961613257,
"loss": 3.2078937530517577,
"step": 4135
},
{
"epoch": 0.7953890489913544,
"grad_norm": 0.8183991312980652,
"learning_rate": 0.00021131362691395608,
"loss": 3.2082897186279298,
"step": 4140
},
{
"epoch": 0.7963496637848223,
"grad_norm": 0.9838517308235168,
"learning_rate": 0.00021109609951120887,
"loss": 3.2070068359375,
"step": 4145
},
{
"epoch": 0.7973102785782901,
"grad_norm": 1.0721310377120972,
"learning_rate": 0.00021087841795673488,
"loss": 3.206983947753906,
"step": 4150
},
{
"epoch": 0.7982708933717579,
"grad_norm": 1.0762015581130981,
"learning_rate": 0.00021066058279976704,
"loss": 3.2114967346191405,
"step": 4155
},
{
"epoch": 0.7992315081652257,
"grad_norm": 0.7675787806510925,
"learning_rate": 0.0002104425945899258,
"loss": 3.2034461975097654,
"step": 4160
},
{
"epoch": 0.8001921229586936,
"grad_norm": 1.0134726762771606,
"learning_rate": 0.00021022445387721767,
"loss": 3.207468032836914,
"step": 4165
},
{
"epoch": 0.8011527377521613,
"grad_norm": 0.9967617988586426,
"learning_rate": 0.00021000616121203422,
"loss": 3.2145057678222657,
"step": 4170
},
{
"epoch": 0.8021133525456292,
"grad_norm": 1.0330275297164917,
"learning_rate": 0.00020978771714515016,
"loss": 3.205029296875,
"step": 4175
},
{
"epoch": 0.803073967339097,
"grad_norm": 0.8626118898391724,
"learning_rate": 0.00020956912222772222,
"loss": 3.2084095001220705,
"step": 4180
},
{
"epoch": 0.8040345821325648,
"grad_norm": 1.186023235321045,
"learning_rate": 0.0002093503770112879,
"loss": 3.209725189208984,
"step": 4185
},
{
"epoch": 0.8049951969260326,
"grad_norm": 0.943437397480011,
"learning_rate": 0.00020913148204776378,
"loss": 3.209857177734375,
"step": 4190
},
{
"epoch": 0.8059558117195005,
"grad_norm": 1.167366623878479,
"learning_rate": 0.00020891243788944428,
"loss": 3.2103233337402344,
"step": 4195
},
{
"epoch": 0.8069164265129684,
"grad_norm": 0.9998597502708435,
"learning_rate": 0.00020869324508900026,
"loss": 3.211322784423828,
"step": 4200
},
{
"epoch": 0.8078770413064361,
"grad_norm": 0.9104480147361755,
"learning_rate": 0.00020847390419947778,
"loss": 3.2085220336914064,
"step": 4205
},
{
"epoch": 0.808837656099904,
"grad_norm": 1.0278260707855225,
"learning_rate": 0.0002082544157742963,
"loss": 3.2045223236083986,
"step": 4210
},
{
"epoch": 0.8097982708933718,
"grad_norm": 0.9279727339744568,
"learning_rate": 0.0002080347803672476,
"loss": 3.2089134216308595,
"step": 4215
},
{
"epoch": 0.8107588856868396,
"grad_norm": 0.9136155247688293,
"learning_rate": 0.00020781499853249441,
"loss": 3.2083240509033204,
"step": 4220
},
{
"epoch": 0.8117195004803074,
"grad_norm": 0.8539382219314575,
"learning_rate": 0.00020759507082456885,
"loss": 3.204279327392578,
"step": 4225
},
{
"epoch": 0.8126801152737753,
"grad_norm": 1.1511064767837524,
"learning_rate": 0.00020737499779837105,
"loss": 3.2057140350341795,
"step": 4230
},
{
"epoch": 0.813640730067243,
"grad_norm": 0.8153745532035828,
"learning_rate": 0.00020715478000916783,
"loss": 3.2046875,
"step": 4235
},
{
"epoch": 0.8146013448607109,
"grad_norm": 0.8916031122207642,
"learning_rate": 0.00020693441801259135,
"loss": 3.210984802246094,
"step": 4240
},
{
"epoch": 0.8155619596541787,
"grad_norm": 0.9601288437843323,
"learning_rate": 0.00020671391236463742,
"loss": 3.2044132232666014,
"step": 4245
},
{
"epoch": 0.8165225744476465,
"grad_norm": 0.9480450749397278,
"learning_rate": 0.00020649326362166449,
"loss": 3.2047096252441407,
"step": 4250
},
{
"epoch": 0.8174831892411143,
"grad_norm": 1.3009711503982544,
"learning_rate": 0.0002062724723403919,
"loss": 3.210066223144531,
"step": 4255
},
{
"epoch": 0.8184438040345822,
"grad_norm": 0.8482257127761841,
"learning_rate": 0.00020605153907789876,
"loss": 3.2040420532226563,
"step": 4260
},
{
"epoch": 0.8194044188280499,
"grad_norm": 1.1215349435806274,
"learning_rate": 0.00020583046439162232,
"loss": 3.209978485107422,
"step": 4265
},
{
"epoch": 0.8203650336215178,
"grad_norm": 1.0416927337646484,
"learning_rate": 0.00020560924883935674,
"loss": 3.20634765625,
"step": 4270
},
{
"epoch": 0.8213256484149856,
"grad_norm": 0.9002500176429749,
"learning_rate": 0.00020538789297925154,
"loss": 3.210291290283203,
"step": 4275
},
{
"epoch": 0.8222862632084534,
"grad_norm": 0.7959718108177185,
"learning_rate": 0.00020516639736981027,
"loss": 3.205914306640625,
"step": 4280
},
{
"epoch": 0.8232468780019212,
"grad_norm": 0.9458216428756714,
"learning_rate": 0.000204944762569889,
"loss": 3.2048385620117186,
"step": 4285
},
{
"epoch": 0.8242074927953891,
"grad_norm": 1.0240002870559692,
"learning_rate": 0.00020472298913869514,
"loss": 3.2106887817382814,
"step": 4290
},
{
"epoch": 0.8251681075888568,
"grad_norm": 0.9878635406494141,
"learning_rate": 0.00020450107763578582,
"loss": 3.206460952758789,
"step": 4295
},
{
"epoch": 0.8261287223823247,
"grad_norm": 0.961821973323822,
"learning_rate": 0.00020427902862106645,
"loss": 3.208537292480469,
"step": 4300
},
{
"epoch": 0.8270893371757925,
"grad_norm": 1.0129098892211914,
"learning_rate": 0.00020405684265478955,
"loss": 3.2066085815429686,
"step": 4305
},
{
"epoch": 0.8280499519692603,
"grad_norm": 1.2535499334335327,
"learning_rate": 0.00020383452029755308,
"loss": 3.206789016723633,
"step": 4310
},
{
"epoch": 0.8290105667627281,
"grad_norm": 1.0553781986236572,
"learning_rate": 0.00020361206211029915,
"loss": 3.208069610595703,
"step": 4315
},
{
"epoch": 0.829971181556196,
"grad_norm": 1.033327341079712,
"learning_rate": 0.00020338946865431257,
"loss": 3.208452606201172,
"step": 4320
},
{
"epoch": 0.8309317963496637,
"grad_norm": 1.1007755994796753,
"learning_rate": 0.00020316674049121954,
"loss": 3.205517578125,
"step": 4325
},
{
"epoch": 0.8318924111431316,
"grad_norm": 1.0913561582565308,
"learning_rate": 0.00020294387818298596,
"loss": 3.205488586425781,
"step": 4330
},
{
"epoch": 0.8328530259365994,
"grad_norm": 1.073475956916809,
"learning_rate": 0.00020272088229191638,
"loss": 3.2070037841796877,
"step": 4335
},
{
"epoch": 0.8338136407300673,
"grad_norm": 0.8359492421150208,
"learning_rate": 0.00020249775338065224,
"loss": 3.2025604248046875,
"step": 4340
},
{
"epoch": 0.834774255523535,
"grad_norm": 1.1163265705108643,
"learning_rate": 0.0002022744920121707,
"loss": 3.205812454223633,
"step": 4345
},
{
"epoch": 0.8357348703170029,
"grad_norm": 1.0195002555847168,
"learning_rate": 0.0002020510987497832,
"loss": 3.2072071075439452,
"step": 4350
},
{
"epoch": 0.8366954851104706,
"grad_norm": 1.06333327293396,
"learning_rate": 0.0002018275741571337,
"loss": 3.2045135498046875,
"step": 4355
},
{
"epoch": 0.8376560999039385,
"grad_norm": 1.07416570186615,
"learning_rate": 0.00020160391879819775,
"loss": 3.2041862487792967,
"step": 4360
},
{
"epoch": 0.8386167146974063,
"grad_norm": 0.8247685432434082,
"learning_rate": 0.00020138013323728072,
"loss": 3.204395294189453,
"step": 4365
},
{
"epoch": 0.8395773294908742,
"grad_norm": 0.7512625455856323,
"learning_rate": 0.00020115621803901658,
"loss": 3.2026763916015626,
"step": 4370
},
{
"epoch": 0.840537944284342,
"grad_norm": 0.6788026094436646,
"learning_rate": 0.0002009321737683664,
"loss": 3.202665328979492,
"step": 4375
},
{
"epoch": 0.8414985590778098,
"grad_norm": 1.178909182548523,
"learning_rate": 0.00020070800099061676,
"loss": 3.203816604614258,
"step": 4380
},
{
"epoch": 0.8424591738712777,
"grad_norm": 1.056028127670288,
"learning_rate": 0.00020048370027137864,
"loss": 3.203491973876953,
"step": 4385
},
{
"epoch": 0.8434197886647454,
"grad_norm": 0.9850196838378906,
"learning_rate": 0.0002002592721765857,
"loss": 3.203214645385742,
"step": 4390
},
{
"epoch": 0.8443804034582133,
"grad_norm": 0.812002956867218,
"learning_rate": 0.00020003471727249317,
"loss": 3.2064640045166017,
"step": 4395
},
{
"epoch": 0.8453410182516811,
"grad_norm": 0.7780818343162537,
"learning_rate": 0.00019981003612567594,
"loss": 3.2039905548095704,
"step": 4400
},
{
"epoch": 0.8463016330451489,
"grad_norm": 0.7282761931419373,
"learning_rate": 0.0001995852293030277,
"loss": 3.2027183532714845,
"step": 4405
},
{
"epoch": 0.8472622478386167,
"grad_norm": 1.597205400466919,
"learning_rate": 0.0001993602973717591,
"loss": 3.2074440002441404,
"step": 4410
},
{
"epoch": 0.8482228626320846,
"grad_norm": 0.7743385434150696,
"learning_rate": 0.0001991352408993965,
"loss": 3.20322265625,
"step": 4415
},
{
"epoch": 0.8491834774255523,
"grad_norm": 0.8802571296691895,
"learning_rate": 0.00019891006045378047,
"loss": 3.2041389465332033,
"step": 4420
},
{
"epoch": 0.8501440922190202,
"grad_norm": 0.9674323797225952,
"learning_rate": 0.00019868475660306435,
"loss": 3.2038337707519533,
"step": 4425
},
{
"epoch": 0.851104707012488,
"grad_norm": 0.8163370490074158,
"learning_rate": 0.00019845932991571284,
"loss": 3.204595947265625,
"step": 4430
},
{
"epoch": 0.8520653218059558,
"grad_norm": 0.8586170077323914,
"learning_rate": 0.00019823378096050067,
"loss": 3.204967498779297,
"step": 4435
},
{
"epoch": 0.8530259365994236,
"grad_norm": 0.9709478616714478,
"learning_rate": 0.00019800811030651095,
"loss": 3.200875091552734,
"step": 4440
},
{
"epoch": 0.8539865513928915,
"grad_norm": 0.8611642718315125,
"learning_rate": 0.00019778231852313386,
"loss": 3.2068603515625,
"step": 4445
},
{
"epoch": 0.8549471661863592,
"grad_norm": 0.8717654347419739,
"learning_rate": 0.00019755640618006532,
"loss": 3.2012969970703127,
"step": 4450
},
{
"epoch": 0.8559077809798271,
"grad_norm": 0.9735947251319885,
"learning_rate": 0.0001973303738473053,
"loss": 3.2008651733398437,
"step": 4455
},
{
"epoch": 0.8568683957732949,
"grad_norm": 0.9653201699256897,
"learning_rate": 0.0001971042220951565,
"loss": 3.2013267517089843,
"step": 4460
},
{
"epoch": 0.8578290105667628,
"grad_norm": 0.8183594346046448,
"learning_rate": 0.00019687795149422315,
"loss": 3.204022979736328,
"step": 4465
},
{
"epoch": 0.8587896253602305,
"grad_norm": 0.8697395920753479,
"learning_rate": 0.00019665156261540898,
"loss": 3.202830505371094,
"step": 4470
},
{
"epoch": 0.8597502401536984,
"grad_norm": 0.829225480556488,
"learning_rate": 0.00019642505602991654,
"loss": 3.2021392822265624,
"step": 4475
},
{
"epoch": 0.8607108549471661,
"grad_norm": 1.1149920225143433,
"learning_rate": 0.0001961984323092451,
"loss": 3.205643081665039,
"step": 4480
},
{
"epoch": 0.861671469740634,
"grad_norm": 0.8502190113067627,
"learning_rate": 0.00019597169202518954,
"loss": 3.2061065673828124,
"step": 4485
},
{
"epoch": 0.8626320845341018,
"grad_norm": 0.7430519461631775,
"learning_rate": 0.0001957448357498389,
"loss": 3.2033714294433593,
"step": 4490
},
{
"epoch": 0.8635926993275697,
"grad_norm": 0.8692427277565002,
"learning_rate": 0.0001955178640555748,
"loss": 3.2003097534179688,
"step": 4495
},
{
"epoch": 0.8645533141210374,
"grad_norm": 0.809934139251709,
"learning_rate": 0.00019529077751507016,
"loss": 3.2053035736083983,
"step": 4500
},
{
"epoch": 0.8655139289145053,
"grad_norm": 0.6470674276351929,
"learning_rate": 0.00019506357670128749,
"loss": 3.2052642822265627,
"step": 4505
},
{
"epoch": 0.866474543707973,
"grad_norm": 0.9643762111663818,
"learning_rate": 0.00019483626218747794,
"loss": 3.203343963623047,
"step": 4510
},
{
"epoch": 0.8674351585014409,
"grad_norm": 1.0639313459396362,
"learning_rate": 0.00019460883454717922,
"loss": 3.2024681091308596,
"step": 4515
},
{
"epoch": 0.8683957732949087,
"grad_norm": 0.8861836194992065,
"learning_rate": 0.00019438129435421462,
"loss": 3.201047897338867,
"step": 4520
},
{
"epoch": 0.8693563880883766,
"grad_norm": 1.0320836305618286,
"learning_rate": 0.00019415364218269146,
"loss": 3.2003639221191404,
"step": 4525
},
{
"epoch": 0.8703170028818443,
"grad_norm": 0.9835689663887024,
"learning_rate": 0.00019392587860699942,
"loss": 3.2048595428466795,
"step": 4530
},
{
"epoch": 0.8712776176753122,
"grad_norm": 0.7003117799758911,
"learning_rate": 0.00019369800420180943,
"loss": 3.204084014892578,
"step": 4535
},
{
"epoch": 0.8722382324687801,
"grad_norm": 0.9389500021934509,
"learning_rate": 0.00019347001954207193,
"loss": 3.2038955688476562,
"step": 4540
},
{
"epoch": 0.8731988472622478,
"grad_norm": 1.1108181476593018,
"learning_rate": 0.00019324192520301566,
"loss": 3.200461196899414,
"step": 4545
},
{
"epoch": 0.8741594620557157,
"grad_norm": 0.84727942943573,
"learning_rate": 0.00019301372176014605,
"loss": 3.2037403106689455,
"step": 4550
},
{
"epoch": 0.8751200768491835,
"grad_norm": 0.776715874671936,
"learning_rate": 0.00019278540978924378,
"loss": 3.1999557495117186,
"step": 4555
},
{
"epoch": 0.8760806916426513,
"grad_norm": 0.873356282711029,
"learning_rate": 0.0001925569898663633,
"loss": 3.201426696777344,
"step": 4560
},
{
"epoch": 0.8770413064361191,
"grad_norm": 1.0719802379608154,
"learning_rate": 0.00019232846256783163,
"loss": 3.2020469665527345,
"step": 4565
},
{
"epoch": 0.878001921229587,
"grad_norm": 0.8912343978881836,
"learning_rate": 0.00019209982847024655,
"loss": 3.200605010986328,
"step": 4570
},
{
"epoch": 0.8789625360230547,
"grad_norm": 0.8289459347724915,
"learning_rate": 0.00019187108815047523,
"loss": 3.2027835845947266,
"step": 4575
},
{
"epoch": 0.8799231508165226,
"grad_norm": 0.7044758200645447,
"learning_rate": 0.00019164224218565313,
"loss": 3.2026336669921873,
"step": 4580
},
{
"epoch": 0.8808837656099904,
"grad_norm": 0.8316428661346436,
"learning_rate": 0.00019141329115318203,
"loss": 3.204214096069336,
"step": 4585
},
{
"epoch": 0.8818443804034583,
"grad_norm": 0.9814732074737549,
"learning_rate": 0.00019118423563072885,
"loss": 3.2025177001953127,
"step": 4590
},
{
"epoch": 0.882804995196926,
"grad_norm": 0.6341441869735718,
"learning_rate": 0.0001909550761962242,
"loss": 3.2031497955322266,
"step": 4595
},
{
"epoch": 0.8837656099903939,
"grad_norm": 0.977326512336731,
"learning_rate": 0.00019072581342786084,
"loss": 3.199177551269531,
"step": 4600
},
{
"epoch": 0.8847262247838616,
"grad_norm": 0.7784291505813599,
"learning_rate": 0.00019049644790409225,
"loss": 3.198267936706543,
"step": 4605
},
{
"epoch": 0.8856868395773295,
"grad_norm": 1.1206945180892944,
"learning_rate": 0.00019026698020363107,
"loss": 3.205250549316406,
"step": 4610
},
{
"epoch": 0.8866474543707973,
"grad_norm": 0.8689852356910706,
"learning_rate": 0.00019003741090544804,
"loss": 3.2022071838378907,
"step": 4615
},
{
"epoch": 0.8876080691642652,
"grad_norm": 1.1133830547332764,
"learning_rate": 0.00018980774058876995,
"loss": 3.2035804748535157,
"step": 4620
},
{
"epoch": 0.8885686839577329,
"grad_norm": 0.8632625937461853,
"learning_rate": 0.00018957796983307858,
"loss": 3.2034885406494142,
"step": 4625
},
{
"epoch": 0.8895292987512008,
"grad_norm": 0.8747223615646362,
"learning_rate": 0.0001893480992181091,
"loss": 3.2007545471191405,
"step": 4630
},
{
"epoch": 0.8904899135446686,
"grad_norm": 1.075469732284546,
"learning_rate": 0.00018911812932384872,
"loss": 3.2042964935302733,
"step": 4635
},
{
"epoch": 0.8914505283381364,
"grad_norm": 1.0559624433517456,
"learning_rate": 0.000188888060730535,
"loss": 3.199691963195801,
"step": 4640
},
{
"epoch": 0.8924111431316042,
"grad_norm": 1.0888924598693848,
"learning_rate": 0.00018865789401865458,
"loss": 3.202299118041992,
"step": 4645
},
{
"epoch": 0.8933717579250721,
"grad_norm": 0.7396308779716492,
"learning_rate": 0.0001884276297689418,
"loss": 3.19757080078125,
"step": 4650
},
{
"epoch": 0.8943323727185398,
"grad_norm": 0.8118330240249634,
"learning_rate": 0.00018819726856237685,
"loss": 3.2041015625,
"step": 4655
},
{
"epoch": 0.8952929875120077,
"grad_norm": 0.9625673890113831,
"learning_rate": 0.00018796681098018486,
"loss": 3.2014694213867188,
"step": 4660
},
{
"epoch": 0.8962536023054755,
"grad_norm": 1.0025596618652344,
"learning_rate": 0.00018773625760383375,
"loss": 3.2010726928710938,
"step": 4665
},
{
"epoch": 0.8972142170989433,
"grad_norm": 1.0931344032287598,
"learning_rate": 0.00018750560901503353,
"loss": 3.2006298065185548,
"step": 4670
},
{
"epoch": 0.8981748318924111,
"grad_norm": 1.0470645427703857,
"learning_rate": 0.00018727486579573407,
"loss": 3.1973636627197264,
"step": 4675
},
{
"epoch": 0.899135446685879,
"grad_norm": 0.9046833515167236,
"learning_rate": 0.00018704402852812431,
"loss": 3.2017478942871094,
"step": 4680
},
{
"epoch": 0.9000960614793467,
"grad_norm": 0.9769369959831238,
"learning_rate": 0.00018681309779463033,
"loss": 3.2017196655273437,
"step": 4685
},
{
"epoch": 0.9010566762728146,
"grad_norm": 0.8879526853561401,
"learning_rate": 0.00018658207417791405,
"loss": 3.2009010314941406,
"step": 4690
},
{
"epoch": 0.9020172910662824,
"grad_norm": 0.7592387795448303,
"learning_rate": 0.00018635095826087175,
"loss": 3.2003753662109373,
"step": 4695
},
{
"epoch": 0.9029779058597502,
"grad_norm": 0.9972590804100037,
"learning_rate": 0.00018611975062663263,
"loss": 3.1992095947265624,
"step": 4700
},
{
"epoch": 0.9039385206532181,
"grad_norm": 0.9059445261955261,
"learning_rate": 0.0001858884518585572,
"loss": 3.195656585693359,
"step": 4705
},
{
"epoch": 0.9048991354466859,
"grad_norm": 0.8296246528625488,
"learning_rate": 0.00018565706254023605,
"loss": 3.196135711669922,
"step": 4710
},
{
"epoch": 0.9058597502401537,
"grad_norm": 0.9912506937980652,
"learning_rate": 0.00018542558325548814,
"loss": 3.1977779388427736,
"step": 4715
},
{
"epoch": 0.9068203650336215,
"grad_norm": 0.7257289290428162,
"learning_rate": 0.00018519401458835948,
"loss": 3.197937774658203,
"step": 4720
},
{
"epoch": 0.9077809798270894,
"grad_norm": 1.0594534873962402,
"learning_rate": 0.00018496235712312154,
"loss": 3.193804168701172,
"step": 4725
},
{
"epoch": 0.9087415946205571,
"grad_norm": 1.0351169109344482,
"learning_rate": 0.00018473061144426986,
"loss": 3.196752166748047,
"step": 4730
},
{
"epoch": 0.909702209414025,
"grad_norm": 0.8392277359962463,
"learning_rate": 0.0001844987781365226,
"loss": 3.2042442321777345,
"step": 4735
},
{
"epoch": 0.9106628242074928,
"grad_norm": 0.6720691323280334,
"learning_rate": 0.00018426685778481897,
"loss": 3.2033058166503907,
"step": 4740
},
{
"epoch": 0.9116234390009607,
"grad_norm": 1.009544014930725,
"learning_rate": 0.00018403485097431778,
"loss": 3.200358200073242,
"step": 4745
},
{
"epoch": 0.9125840537944284,
"grad_norm": 0.8520932793617249,
"learning_rate": 0.00018380275829039602,
"loss": 3.201191711425781,
"step": 4750
},
{
"epoch": 0.9135446685878963,
"grad_norm": 0.9514408111572266,
"learning_rate": 0.0001835705803186474,
"loss": 3.194804000854492,
"step": 4755
},
{
"epoch": 0.914505283381364,
"grad_norm": 0.8069214224815369,
"learning_rate": 0.00018333831764488065,
"loss": 3.1965736389160155,
"step": 4760
},
{
"epoch": 0.9154658981748319,
"grad_norm": 0.8542043566703796,
"learning_rate": 0.00018310597085511844,
"loss": 3.1960010528564453,
"step": 4765
},
{
"epoch": 0.9164265129682997,
"grad_norm": 0.7166752815246582,
"learning_rate": 0.00018287354053559546,
"loss": 3.1996959686279296,
"step": 4770
},
{
"epoch": 0.9173871277617676,
"grad_norm": 1.2815790176391602,
"learning_rate": 0.0001826410272727574,
"loss": 3.200403594970703,
"step": 4775
},
{
"epoch": 0.9183477425552353,
"grad_norm": 0.9036345481872559,
"learning_rate": 0.00018240843165325882,
"loss": 3.203938674926758,
"step": 4780
},
{
"epoch": 0.9193083573487032,
"grad_norm": 0.9565702080726624,
"learning_rate": 0.00018217575426396256,
"loss": 3.197901153564453,
"step": 4785
},
{
"epoch": 0.920268972142171,
"grad_norm": 0.8386558890342712,
"learning_rate": 0.00018194299569193747,
"loss": 3.194584274291992,
"step": 4790
},
{
"epoch": 0.9212295869356388,
"grad_norm": 1.0965979099273682,
"learning_rate": 0.0001817101565244573,
"loss": 3.20220947265625,
"step": 4795
},
{
"epoch": 0.9221902017291066,
"grad_norm": 1.1644456386566162,
"learning_rate": 0.00018147723734899916,
"loss": 3.1982275009155274,
"step": 4800
},
{
"epoch": 0.9231508165225745,
"grad_norm": 1.0380982160568237,
"learning_rate": 0.00018124423875324198,
"loss": 3.1980897903442385,
"step": 4805
},
{
"epoch": 0.9241114313160422,
"grad_norm": 0.823890209197998,
"learning_rate": 0.00018101116132506522,
"loss": 3.1994667053222656,
"step": 4810
},
{
"epoch": 0.9250720461095101,
"grad_norm": 0.5867325663566589,
"learning_rate": 0.00018077800565254702,
"loss": 3.1975902557373046,
"step": 4815
},
{
"epoch": 0.9260326609029779,
"grad_norm": 1.1908833980560303,
"learning_rate": 0.00018054477232396312,
"loss": 3.198208808898926,
"step": 4820
},
{
"epoch": 0.9269932756964457,
"grad_norm": 1.0711902379989624,
"learning_rate": 0.00018031146192778517,
"loss": 3.1953447341918944,
"step": 4825
},
{
"epoch": 0.9279538904899135,
"grad_norm": 0.9716858863830566,
"learning_rate": 0.0001800780750526792,
"loss": 3.1982431411743164,
"step": 4830
},
{
"epoch": 0.9289145052833814,
"grad_norm": 0.9620060920715332,
"learning_rate": 0.00017984461228750422,
"loss": 3.196470260620117,
"step": 4835
},
{
"epoch": 0.9298751200768491,
"grad_norm": 1.015932321548462,
"learning_rate": 0.00017961107422131075,
"loss": 3.199555206298828,
"step": 4840
},
{
"epoch": 0.930835734870317,
"grad_norm": 0.9855188727378845,
"learning_rate": 0.00017937746144333934,
"loss": 3.2012577056884766,
"step": 4845
},
{
"epoch": 0.9317963496637848,
"grad_norm": 0.8330841660499573,
"learning_rate": 0.0001791437745430189,
"loss": 3.1955251693725586,
"step": 4850
},
{
"epoch": 0.9327569644572526,
"grad_norm": 0.8580245971679688,
"learning_rate": 0.00017891001410996554,
"loss": 3.1982753753662108,
"step": 4855
},
{
"epoch": 0.9337175792507204,
"grad_norm": 1.1361298561096191,
"learning_rate": 0.00017867618073398074,
"loss": 3.202219009399414,
"step": 4860
},
{
"epoch": 0.9346781940441883,
"grad_norm": 0.9495155215263367,
"learning_rate": 0.00017844227500505016,
"loss": 3.1957366943359373,
"step": 4865
},
{
"epoch": 0.9356388088376562,
"grad_norm": 0.8877089619636536,
"learning_rate": 0.00017820829751334194,
"loss": 3.198602867126465,
"step": 4870
},
{
"epoch": 0.9365994236311239,
"grad_norm": 0.7768687605857849,
"learning_rate": 0.0001779742488492052,
"loss": 3.1972635269165037,
"step": 4875
},
{
"epoch": 0.9375600384245918,
"grad_norm": 1.0583457946777344,
"learning_rate": 0.00017774012960316887,
"loss": 3.1948143005371095,
"step": 4880
},
{
"epoch": 0.9385206532180596,
"grad_norm": 0.9406375288963318,
"learning_rate": 0.00017750594036593974,
"loss": 3.1969860076904295,
"step": 4885
},
{
"epoch": 0.9394812680115274,
"grad_norm": 1.0013177394866943,
"learning_rate": 0.00017727168172840133,
"loss": 3.1973243713378907,
"step": 4890
},
{
"epoch": 0.9404418828049952,
"grad_norm": 0.7509773373603821,
"learning_rate": 0.00017703735428161216,
"loss": 3.198769760131836,
"step": 4895
},
{
"epoch": 0.9414024975984631,
"grad_norm": 0.8109461069107056,
"learning_rate": 0.00017680295861680445,
"loss": 3.1947399139404298,
"step": 4900
},
{
"epoch": 0.9423631123919308,
"grad_norm": 0.8327275514602661,
"learning_rate": 0.0001765684953253825,
"loss": 3.196604919433594,
"step": 4905
},
{
"epoch": 0.9433237271853987,
"grad_norm": 0.8777264356613159,
"learning_rate": 0.00017633396499892125,
"loss": 3.197701644897461,
"step": 4910
},
{
"epoch": 0.9442843419788665,
"grad_norm": 1.045627474784851,
"learning_rate": 0.00017609936822916475,
"loss": 3.204631805419922,
"step": 4915
},
{
"epoch": 0.9452449567723343,
"grad_norm": 1.0442372560501099,
"learning_rate": 0.0001758647056080248,
"loss": 3.2004959106445314,
"step": 4920
},
{
"epoch": 0.9462055715658021,
"grad_norm": 0.7651126384735107,
"learning_rate": 0.00017562997772757916,
"loss": 3.1948537826538086,
"step": 4925
},
{
"epoch": 0.94716618635927,
"grad_norm": 0.790465235710144,
"learning_rate": 0.00017539518518007043,
"loss": 3.196913719177246,
"step": 4930
},
{
"epoch": 0.9481268011527377,
"grad_norm": 0.9765505194664001,
"learning_rate": 0.00017516032855790424,
"loss": 3.1958955764770507,
"step": 4935
},
{
"epoch": 0.9490874159462056,
"grad_norm": 1.0204505920410156,
"learning_rate": 0.00017492540845364798,
"loss": 3.197770690917969,
"step": 4940
},
{
"epoch": 0.9500480307396734,
"grad_norm": 0.7875639796257019,
"learning_rate": 0.00017469042546002913,
"loss": 3.1972192764282226,
"step": 4945
},
{
"epoch": 0.9510086455331412,
"grad_norm": 0.8441899418830872,
"learning_rate": 0.00017445538016993393,
"loss": 3.1970443725585938,
"step": 4950
},
{
"epoch": 0.951969260326609,
"grad_norm": 0.9154574275016785,
"learning_rate": 0.00017422027317640575,
"loss": 3.1967355728149416,
"step": 4955
},
{
"epoch": 0.9529298751200769,
"grad_norm": 0.6774746775627136,
"learning_rate": 0.00017398510507264363,
"loss": 3.1978105545043944,
"step": 4960
},
{
"epoch": 0.9538904899135446,
"grad_norm": 0.9463170170783997,
"learning_rate": 0.0001737498764520009,
"loss": 3.195131301879883,
"step": 4965
},
{
"epoch": 0.9548511047070125,
"grad_norm": 1.0156409740447998,
"learning_rate": 0.00017351458790798338,
"loss": 3.197483253479004,
"step": 4970
},
{
"epoch": 0.9558117195004803,
"grad_norm": 0.8678054213523865,
"learning_rate": 0.0001732792400342483,
"loss": 3.1987661361694335,
"step": 4975
},
{
"epoch": 0.9567723342939481,
"grad_norm": 0.8004742860794067,
"learning_rate": 0.00017304383342460242,
"loss": 3.1965717315673827,
"step": 4980
},
{
"epoch": 0.9577329490874159,
"grad_norm": 1.0178972482681274,
"learning_rate": 0.0001728083686730008,
"loss": 3.197255325317383,
"step": 4985
},
{
"epoch": 0.9586935638808838,
"grad_norm": 0.7633374333381653,
"learning_rate": 0.00017257284637354524,
"loss": 3.1934268951416014,
"step": 4990
},
{
"epoch": 0.9596541786743515,
"grad_norm": 0.735908031463623,
"learning_rate": 0.00017233726712048253,
"loss": 3.194230079650879,
"step": 4995
},
{
"epoch": 0.9606147934678194,
"grad_norm": 0.936837911605835,
"learning_rate": 0.0001721016315082034,
"loss": 3.1968971252441407,
"step": 5000
},
{
"epoch": 0.9615754082612872,
"grad_norm": 0.8966473937034607,
"learning_rate": 0.00017186594013124057,
"loss": 3.2022254943847654,
"step": 5005
},
{
"epoch": 0.962536023054755,
"grad_norm": 0.8859227299690247,
"learning_rate": 0.00017163019358426763,
"loss": 3.19702205657959,
"step": 5010
},
{
"epoch": 0.9634966378482228,
"grad_norm": 0.6885091662406921,
"learning_rate": 0.00017139439246209728,
"loss": 3.1955480575561523,
"step": 5015
},
{
"epoch": 0.9644572526416907,
"grad_norm": 0.6677486896514893,
"learning_rate": 0.00017115853735967995,
"loss": 3.1977615356445312,
"step": 5020
},
{
"epoch": 0.9654178674351584,
"grad_norm": 1.0436347723007202,
"learning_rate": 0.00017092262887210232,
"loss": 3.195005416870117,
"step": 5025
},
{
"epoch": 0.9663784822286263,
"grad_norm": 0.9044774174690247,
"learning_rate": 0.0001706866675945856,
"loss": 3.1969493865966796,
"step": 5030
},
{
"epoch": 0.9673390970220941,
"grad_norm": 0.9574539661407471,
"learning_rate": 0.00017045065412248434,
"loss": 3.194741058349609,
"step": 5035
},
{
"epoch": 0.968299711815562,
"grad_norm": 0.9477970004081726,
"learning_rate": 0.00017021458905128477,
"loss": 3.1963451385498045,
"step": 5040
},
{
"epoch": 0.9692603266090298,
"grad_norm": 0.8571361899375916,
"learning_rate": 0.00016997847297660324,
"loss": 3.194466972351074,
"step": 5045
},
{
"epoch": 0.9702209414024976,
"grad_norm": 0.8901039361953735,
"learning_rate": 0.00016974230649418487,
"loss": 3.1929143905639648,
"step": 5050
},
{
"epoch": 0.9711815561959655,
"grad_norm": 1.027086615562439,
"learning_rate": 0.00016950609019990187,
"loss": 3.191559982299805,
"step": 5055
},
{
"epoch": 0.9721421709894332,
"grad_norm": 0.8591023683547974,
"learning_rate": 0.00016926982468975225,
"loss": 3.199761962890625,
"step": 5060
},
{
"epoch": 0.9731027857829011,
"grad_norm": 1.0826267004013062,
"learning_rate": 0.00016903351055985806,
"loss": 3.1961299896240236,
"step": 5065
},
{
"epoch": 0.9740634005763689,
"grad_norm": 1.0031397342681885,
"learning_rate": 0.0001687971484064642,
"loss": 3.1959823608398437,
"step": 5070
},
{
"epoch": 0.9750240153698367,
"grad_norm": 0.7752434015274048,
"learning_rate": 0.00016856073882593646,
"loss": 3.1921688079833985,
"step": 5075
},
{
"epoch": 0.9759846301633045,
"grad_norm": 0.911412239074707,
"learning_rate": 0.00016832428241476063,
"loss": 3.191498947143555,
"step": 5080
},
{
"epoch": 0.9769452449567724,
"grad_norm": 0.9647287726402283,
"learning_rate": 0.00016808777976954042,
"loss": 3.1929500579833983,
"step": 5085
},
{
"epoch": 0.9779058597502401,
"grad_norm": 0.7081168293952942,
"learning_rate": 0.00016785123148699624,
"loss": 3.1945589065551756,
"step": 5090
},
{
"epoch": 0.978866474543708,
"grad_norm": 0.8361444473266602,
"learning_rate": 0.00016761463816396374,
"loss": 3.1967111587524415,
"step": 5095
},
{
"epoch": 0.9798270893371758,
"grad_norm": 0.8908597230911255,
"learning_rate": 0.00016737800039739212,
"loss": 3.1946334838867188,
"step": 5100
},
{
"epoch": 0.9807877041306436,
"grad_norm": 0.8121721744537354,
"learning_rate": 0.0001671413187843427,
"loss": 3.195412826538086,
"step": 5105
},
{
"epoch": 0.9817483189241114,
"grad_norm": 0.6848476529121399,
"learning_rate": 0.00016690459392198752,
"loss": 3.191677284240723,
"step": 5110
},
{
"epoch": 0.9827089337175793,
"grad_norm": 0.6763195991516113,
"learning_rate": 0.00016666782640760766,
"loss": 3.1958782196044924,
"step": 5115
},
{
"epoch": 0.983669548511047,
"grad_norm": 0.8343232870101929,
"learning_rate": 0.00016643101683859186,
"loss": 3.197767639160156,
"step": 5120
},
{
"epoch": 0.9846301633045149,
"grad_norm": 0.9125416874885559,
"learning_rate": 0.00016619416581243498,
"loss": 3.1921886444091796,
"step": 5125
},
{
"epoch": 0.9855907780979827,
"grad_norm": 0.8966071009635925,
"learning_rate": 0.00016595727392673643,
"loss": 3.1952947616577148,
"step": 5130
},
{
"epoch": 0.9865513928914506,
"grad_norm": 0.6765633225440979,
"learning_rate": 0.00016572034177919875,
"loss": 3.1933788299560546,
"step": 5135
},
{
"epoch": 0.9875120076849183,
"grad_norm": 0.814222514629364,
"learning_rate": 0.00016548336996762608,
"loss": 3.1904666900634764,
"step": 5140
},
{
"epoch": 0.9884726224783862,
"grad_norm": 0.9140084981918335,
"learning_rate": 0.00016524635908992253,
"loss": 3.1964300155639647,
"step": 5145
},
{
"epoch": 0.989433237271854,
"grad_norm": 0.9220796227455139,
"learning_rate": 0.00016500930974409092,
"loss": 3.191013526916504,
"step": 5150
},
{
"epoch": 0.9903938520653218,
"grad_norm": 0.7357509732246399,
"learning_rate": 0.00016477222252823107,
"loss": 3.192600059509277,
"step": 5155
},
{
"epoch": 0.9913544668587896,
"grad_norm": 0.9331154227256775,
"learning_rate": 0.00016453509804053833,
"loss": 3.195221519470215,
"step": 5160
},
{
"epoch": 0.9923150816522575,
"grad_norm": 0.8373104929924011,
"learning_rate": 0.00016429793687930215,
"loss": 3.195628356933594,
"step": 5165
},
{
"epoch": 0.9932756964457252,
"grad_norm": 0.8984804153442383,
"learning_rate": 0.0001640607396429044,
"loss": 3.192228889465332,
"step": 5170
},
{
"epoch": 0.9942363112391931,
"grad_norm": 0.8978838324546814,
"learning_rate": 0.00016382350692981812,
"loss": 3.1939905166625975,
"step": 5175
},
{
"epoch": 0.9951969260326609,
"grad_norm": 0.6365435719490051,
"learning_rate": 0.00016358623933860567,
"loss": 3.1925621032714844,
"step": 5180
},
{
"epoch": 0.9961575408261287,
"grad_norm": 0.7753834128379822,
"learning_rate": 0.00016334893746791762,
"loss": 3.192973327636719,
"step": 5185
},
{
"epoch": 0.9971181556195965,
"grad_norm": 0.9735800623893738,
"learning_rate": 0.0001631116019164909,
"loss": 3.1959808349609373,
"step": 5190
},
{
"epoch": 0.9980787704130644,
"grad_norm": 0.6787250638008118,
"learning_rate": 0.00016287423328314746,
"loss": 3.193808746337891,
"step": 5195
},
{
"epoch": 0.9990393852065321,
"grad_norm": 0.7312700152397156,
"learning_rate": 0.00016263683216679274,
"loss": 3.1927206039428713,
"step": 5200
},
{
"epoch": 1.0,
"grad_norm": 0.7704550623893738,
"learning_rate": 0.00016239939916641404,
"loss": 3.1963579177856447,
"step": 5205
},
{
"epoch": 1.0009606147934678,
"grad_norm": 0.7807424664497375,
"learning_rate": 0.00016216193488107926,
"loss": 3.1920791625976563,
"step": 5210
},
{
"epoch": 1.0019212295869357,
"grad_norm": 0.9543339014053345,
"learning_rate": 0.0001619244399099351,
"loss": 3.1910972595214844,
"step": 5215
},
{
"epoch": 1.0028818443804035,
"grad_norm": 0.7766204476356506,
"learning_rate": 0.00016168691485220573,
"loss": 3.194061851501465,
"step": 5220
},
{
"epoch": 1.0038424591738713,
"grad_norm": 0.8086463809013367,
"learning_rate": 0.00016144936030719126,
"loss": 3.1898380279541017,
"step": 5225
},
{
"epoch": 1.004803073967339,
"grad_norm": 1.1027806997299194,
"learning_rate": 0.00016121177687426617,
"loss": 3.193951416015625,
"step": 5230
},
{
"epoch": 1.005763688760807,
"grad_norm": 0.996457576751709,
"learning_rate": 0.00016097416515287787,
"loss": 3.188690757751465,
"step": 5235
},
{
"epoch": 1.0067243035542748,
"grad_norm": 0.7975702881813049,
"learning_rate": 0.00016073652574254504,
"loss": 3.1924358367919923,
"step": 5240
},
{
"epoch": 1.0076849183477425,
"grad_norm": 0.7981067299842834,
"learning_rate": 0.00016049885924285638,
"loss": 3.1881412506103515,
"step": 5245
},
{
"epoch": 1.0086455331412103,
"grad_norm": 0.8105031251907349,
"learning_rate": 0.00016026116625346876,
"loss": 3.192718505859375,
"step": 5250
},
{
"epoch": 1.0096061479346783,
"grad_norm": 0.7112568616867065,
"learning_rate": 0.00016002344737410602,
"loss": 3.187815856933594,
"step": 5255
},
{
"epoch": 1.010566762728146,
"grad_norm": 0.8126310706138611,
"learning_rate": 0.0001597857032045573,
"loss": 3.191788673400879,
"step": 5260
},
{
"epoch": 1.0115273775216138,
"grad_norm": 0.733974814414978,
"learning_rate": 0.00015954793434467545,
"loss": 3.1942237854003905,
"step": 5265
},
{
"epoch": 1.0124879923150816,
"grad_norm": 0.6596633791923523,
"learning_rate": 0.0001593101413943758,
"loss": 3.189275932312012,
"step": 5270
},
{
"epoch": 1.0134486071085496,
"grad_norm": 0.8669559955596924,
"learning_rate": 0.0001590723249536343,
"loss": 3.1921119689941406,
"step": 5275
},
{
"epoch": 1.0144092219020173,
"grad_norm": 1.2266993522644043,
"learning_rate": 0.00015883448562248624,
"loss": 3.191496467590332,
"step": 5280
},
{
"epoch": 1.015369836695485,
"grad_norm": 0.8078704476356506,
"learning_rate": 0.00015859662400102462,
"loss": 3.1911800384521483,
"step": 5285
},
{
"epoch": 1.0163304514889528,
"grad_norm": 0.6551411151885986,
"learning_rate": 0.00015835874068939878,
"loss": 3.188446044921875,
"step": 5290
},
{
"epoch": 1.0172910662824208,
"grad_norm": 0.8124401569366455,
"learning_rate": 0.0001581208362878126,
"loss": 3.1922882080078123,
"step": 5295
},
{
"epoch": 1.0182516810758886,
"grad_norm": 0.8051562309265137,
"learning_rate": 0.00015788291139652344,
"loss": 3.190486717224121,
"step": 5300
},
{
"epoch": 1.0192122958693564,
"grad_norm": 0.9743736386299133,
"learning_rate": 0.00015764496661584013,
"loss": 3.1894975662231446,
"step": 5305
},
{
"epoch": 1.0201729106628241,
"grad_norm": 0.7459861040115356,
"learning_rate": 0.00015740700254612177,
"loss": 3.193043518066406,
"step": 5310
},
{
"epoch": 1.021133525456292,
"grad_norm": 1.2092807292938232,
"learning_rate": 0.0001571690197877761,
"loss": 3.1933567047119142,
"step": 5315
},
{
"epoch": 1.0220941402497599,
"grad_norm": 0.7785642147064209,
"learning_rate": 0.00015693101894125806,
"loss": 3.1891590118408204,
"step": 5320
},
{
"epoch": 1.0230547550432276,
"grad_norm": 0.7812630534172058,
"learning_rate": 0.00015669300060706823,
"loss": 3.190948486328125,
"step": 5325
},
{
"epoch": 1.0240153698366954,
"grad_norm": 0.9370684027671814,
"learning_rate": 0.0001564549653857512,
"loss": 3.1883705139160154,
"step": 5330
},
{
"epoch": 1.0249759846301634,
"grad_norm": 0.8281055092811584,
"learning_rate": 0.00015621691387789432,
"loss": 3.1863418579101563,
"step": 5335
},
{
"epoch": 1.0259365994236311,
"grad_norm": 1.0316886901855469,
"learning_rate": 0.000155978846684126,
"loss": 3.19388370513916,
"step": 5340
},
{
"epoch": 1.026897214217099,
"grad_norm": 0.758702278137207,
"learning_rate": 0.00015574076440511407,
"loss": 3.1906862258911133,
"step": 5345
},
{
"epoch": 1.0278578290105667,
"grad_norm": 0.5779961347579956,
"learning_rate": 0.00015550266764156466,
"loss": 3.190965461730957,
"step": 5350
},
{
"epoch": 1.0288184438040346,
"grad_norm": 0.9982998371124268,
"learning_rate": 0.00015526455699422023,
"loss": 3.190749740600586,
"step": 5355
},
{
"epoch": 1.0297790585975024,
"grad_norm": 1.105873465538025,
"learning_rate": 0.00015502643306385846,
"loss": 3.1928503036499025,
"step": 5360
},
{
"epoch": 1.0307396733909702,
"grad_norm": 0.9287274479866028,
"learning_rate": 0.00015478829645129046,
"loss": 3.1860063552856444,
"step": 5365
},
{
"epoch": 1.031700288184438,
"grad_norm": 0.7572025656700134,
"learning_rate": 0.00015455014775735923,
"loss": 3.1904781341552733,
"step": 5370
},
{
"epoch": 1.032660902977906,
"grad_norm": 0.8866919279098511,
"learning_rate": 0.0001543119875829385,
"loss": 3.1870819091796876,
"step": 5375
},
{
"epoch": 1.0336215177713737,
"grad_norm": 0.7839595675468445,
"learning_rate": 0.00015407381652893066,
"loss": 3.1915761947631838,
"step": 5380
},
{
"epoch": 1.0345821325648414,
"grad_norm": 0.7243456244468689,
"learning_rate": 0.00015383563519626582,
"loss": 3.1893665313720705,
"step": 5385
},
{
"epoch": 1.0355427473583094,
"grad_norm": 1.0272892713546753,
"learning_rate": 0.0001535974441858999,
"loss": 3.187778854370117,
"step": 5390
},
{
"epoch": 1.0365033621517772,
"grad_norm": 1.069861650466919,
"learning_rate": 0.00015335924409881323,
"loss": 3.1904586791992187,
"step": 5395
},
{
"epoch": 1.037463976945245,
"grad_norm": 0.8298363089561462,
"learning_rate": 0.00015312103553600913,
"loss": 3.1916831970214843,
"step": 5400
},
{
"epoch": 1.0384245917387127,
"grad_norm": 0.7426701784133911,
"learning_rate": 0.00015288281909851213,
"loss": 3.1897172927856445,
"step": 5405
},
{
"epoch": 1.0393852065321807,
"grad_norm": 0.8290877938270569,
"learning_rate": 0.00015264459538736686,
"loss": 3.1861778259277345,
"step": 5410
},
{
"epoch": 1.0403458213256485,
"grad_norm": 0.5183548927307129,
"learning_rate": 0.00015240636500363607,
"loss": 3.185122489929199,
"step": 5415
},
{
"epoch": 1.0413064361191162,
"grad_norm": 0.9525085687637329,
"learning_rate": 0.00015216812854839957,
"loss": 3.187310791015625,
"step": 5420
},
{
"epoch": 1.042267050912584,
"grad_norm": 0.9072858691215515,
"learning_rate": 0.0001519298866227523,
"loss": 3.1853969573974608,
"step": 5425
},
{
"epoch": 1.043227665706052,
"grad_norm": 0.8549609780311584,
"learning_rate": 0.00015169163982780304,
"loss": 3.1908397674560547,
"step": 5430
},
{
"epoch": 1.0441882804995197,
"grad_norm": 0.5929837822914124,
"learning_rate": 0.00015145338876467303,
"loss": 3.1879791259765624,
"step": 5435
},
{
"epoch": 1.0451488952929875,
"grad_norm": 0.8749310970306396,
"learning_rate": 0.00015121513403449402,
"loss": 3.1909906387329103,
"step": 5440
},
{
"epoch": 1.0461095100864553,
"grad_norm": 0.6831730604171753,
"learning_rate": 0.00015097687623840726,
"loss": 3.188145637512207,
"step": 5445
},
{
"epoch": 1.0470701248799232,
"grad_norm": 0.7838398814201355,
"learning_rate": 0.00015073861597756144,
"loss": 3.1877973556518553,
"step": 5450
},
{
"epoch": 1.048030739673391,
"grad_norm": 0.6628211140632629,
"learning_rate": 0.00015050035385311182,
"loss": 3.1896636962890623,
"step": 5455
},
{
"epoch": 1.0489913544668588,
"grad_norm": 0.737322211265564,
"learning_rate": 0.00015026209046621803,
"loss": 3.1869300842285155,
"step": 5460
},
{
"epoch": 1.0499519692603265,
"grad_norm": 1.0209139585494995,
"learning_rate": 0.0001500238264180431,
"loss": 3.187608528137207,
"step": 5465
},
{
"epoch": 1.0509125840537945,
"grad_norm": 0.7242445945739746,
"learning_rate": 0.00014978556230975173,
"loss": 3.188776969909668,
"step": 5470
},
{
"epoch": 1.0518731988472623,
"grad_norm": 0.968614399433136,
"learning_rate": 0.0001495472987425086,
"loss": 3.1897449493408203,
"step": 5475
},
{
"epoch": 1.05283381364073,
"grad_norm": 1.0760611295700073,
"learning_rate": 0.00014930903631747724,
"loss": 3.1843488693237303,
"step": 5480
},
{
"epoch": 1.0537944284341978,
"grad_norm": 0.8381465673446655,
"learning_rate": 0.00014907077563581809,
"loss": 3.189168930053711,
"step": 5485
},
{
"epoch": 1.0547550432276658,
"grad_norm": 0.5522369742393494,
"learning_rate": 0.00014883251729868737,
"loss": 3.1856075286865235,
"step": 5490
},
{
"epoch": 1.0557156580211335,
"grad_norm": 0.8688974380493164,
"learning_rate": 0.00014859426190723523,
"loss": 3.1877534866333006,
"step": 5495
},
{
"epoch": 1.0566762728146013,
"grad_norm": 1.0312248468399048,
"learning_rate": 0.00014835601006260455,
"loss": 3.187721824645996,
"step": 5500
},
{
"epoch": 1.057636887608069,
"grad_norm": 1.0165605545043945,
"learning_rate": 0.00014811776236592914,
"loss": 3.1891069412231445,
"step": 5505
},
{
"epoch": 1.058597502401537,
"grad_norm": 0.6776289939880371,
"learning_rate": 0.00014787951941833236,
"loss": 3.188289451599121,
"step": 5510
},
{
"epoch": 1.0595581171950048,
"grad_norm": 0.6294692158699036,
"learning_rate": 0.00014764128182092557,
"loss": 3.191664123535156,
"step": 5515
},
{
"epoch": 1.0605187319884726,
"grad_norm": 0.8317521810531616,
"learning_rate": 0.00014740305017480674,
"loss": 3.1893966674804686,
"step": 5520
},
{
"epoch": 1.0614793467819403,
"grad_norm": 0.9894767999649048,
"learning_rate": 0.0001471648250810588,
"loss": 3.1885601043701173,
"step": 5525
},
{
"epoch": 1.0624399615754083,
"grad_norm": 0.7484564185142517,
"learning_rate": 0.00014692660714074796,
"loss": 3.18869571685791,
"step": 5530
},
{
"epoch": 1.063400576368876,
"grad_norm": 0.8479475975036621,
"learning_rate": 0.00014668839695492264,
"loss": 3.184678649902344,
"step": 5535
},
{
"epoch": 1.0643611911623438,
"grad_norm": 0.6609562039375305,
"learning_rate": 0.00014645019512461157,
"loss": 3.1846710205078126,
"step": 5540
},
{
"epoch": 1.0653218059558118,
"grad_norm": 0.955204963684082,
"learning_rate": 0.0001462120022508223,
"loss": 3.19210205078125,
"step": 5545
},
{
"epoch": 1.0662824207492796,
"grad_norm": 0.7516992092132568,
"learning_rate": 0.00014597381893453998,
"loss": 3.188296318054199,
"step": 5550
},
{
"epoch": 1.0672430355427474,
"grad_norm": 0.7409364581108093,
"learning_rate": 0.0001457356457767255,
"loss": 3.1862789154052735,
"step": 5555
},
{
"epoch": 1.0682036503362151,
"grad_norm": 0.6613131165504456,
"learning_rate": 0.00014549748337831423,
"loss": 3.185582733154297,
"step": 5560
},
{
"epoch": 1.069164265129683,
"grad_norm": 0.6710831522941589,
"learning_rate": 0.00014525933234021428,
"loss": 3.1856903076171874,
"step": 5565
},
{
"epoch": 1.0701248799231509,
"grad_norm": 0.6263184547424316,
"learning_rate": 0.00014502119326330516,
"loss": 3.186991310119629,
"step": 5570
},
{
"epoch": 1.0710854947166186,
"grad_norm": 0.838955283164978,
"learning_rate": 0.0001447830667484362,
"loss": 3.1853305816650392,
"step": 5575
},
{
"epoch": 1.0720461095100864,
"grad_norm": 0.6884387135505676,
"learning_rate": 0.00014454495339642504,
"loss": 3.189616584777832,
"step": 5580
},
{
"epoch": 1.0730067243035544,
"grad_norm": 0.8635831475257874,
"learning_rate": 0.00014430685380805605,
"loss": 3.185613822937012,
"step": 5585
},
{
"epoch": 1.0739673390970221,
"grad_norm": 0.842369794845581,
"learning_rate": 0.00014406876858407894,
"loss": 3.1873950958251953,
"step": 5590
},
{
"epoch": 1.07492795389049,
"grad_norm": 0.6380950212478638,
"learning_rate": 0.00014383069832520721,
"loss": 3.1872686386108398,
"step": 5595
},
{
"epoch": 1.0758885686839577,
"grad_norm": 0.6167466044425964,
"learning_rate": 0.00014359264363211647,
"loss": 3.1889183044433596,
"step": 5600
},
{
"epoch": 1.0768491834774256,
"grad_norm": 0.8871692419052124,
"learning_rate": 0.00014335460510544318,
"loss": 3.190326118469238,
"step": 5605
},
{
"epoch": 1.0778097982708934,
"grad_norm": 0.8428598642349243,
"learning_rate": 0.00014311658334578296,
"loss": 3.190188407897949,
"step": 5610
},
{
"epoch": 1.0787704130643612,
"grad_norm": 0.7600328922271729,
"learning_rate": 0.00014287857895368906,
"loss": 3.1856243133544924,
"step": 5615
},
{
"epoch": 1.079731027857829,
"grad_norm": 0.7061536908149719,
"learning_rate": 0.00014264059252967108,
"loss": 3.1858592987060548,
"step": 5620
},
{
"epoch": 1.080691642651297,
"grad_norm": 0.6776458024978638,
"learning_rate": 0.0001424026246741931,
"loss": 3.188393402099609,
"step": 5625
},
{
"epoch": 1.0816522574447647,
"grad_norm": 0.9490513205528259,
"learning_rate": 0.00014216467598767248,
"loss": 3.1876445770263673,
"step": 5630
},
{
"epoch": 1.0826128722382324,
"grad_norm": 0.8668431043624878,
"learning_rate": 0.0001419267470704781,
"loss": 3.1881153106689455,
"step": 5635
},
{
"epoch": 1.0835734870317002,
"grad_norm": 0.7393624782562256,
"learning_rate": 0.00014168883852292915,
"loss": 3.186983108520508,
"step": 5640
},
{
"epoch": 1.0845341018251682,
"grad_norm": 0.7788834571838379,
"learning_rate": 0.00014145095094529318,
"loss": 3.187802314758301,
"step": 5645
},
{
"epoch": 1.085494716618636,
"grad_norm": 0.6699324250221252,
"learning_rate": 0.00014121308493778503,
"loss": 3.1878355026245115,
"step": 5650
},
{
"epoch": 1.0864553314121037,
"grad_norm": 0.7733721733093262,
"learning_rate": 0.00014097524110056494,
"loss": 3.1843936920166014,
"step": 5655
},
{
"epoch": 1.0874159462055715,
"grad_norm": 0.711729884147644,
"learning_rate": 0.0001407374200337374,
"loss": 3.1883363723754883,
"step": 5660
},
{
"epoch": 1.0883765609990395,
"grad_norm": 0.6013932824134827,
"learning_rate": 0.0001404996223373494,
"loss": 3.1905609130859376,
"step": 5665
},
{
"epoch": 1.0893371757925072,
"grad_norm": 0.7252037525177002,
"learning_rate": 0.0001402618486113888,
"loss": 3.1879077911376954,
"step": 5670
},
{
"epoch": 1.090297790585975,
"grad_norm": 1.1090692281723022,
"learning_rate": 0.0001400240994557832,
"loss": 3.1889415740966798,
"step": 5675
},
{
"epoch": 1.0912584053794427,
"grad_norm": 1.0110430717468262,
"learning_rate": 0.0001397863754703981,
"loss": 3.185939407348633,
"step": 5680
},
{
"epoch": 1.0922190201729107,
"grad_norm": 0.8803393840789795,
"learning_rate": 0.00013954867725503542,
"loss": 3.1873733520507814,
"step": 5685
},
{
"epoch": 1.0931796349663785,
"grad_norm": 0.6381970047950745,
"learning_rate": 0.00013931100540943227,
"loss": 3.1853010177612306,
"step": 5690
},
{
"epoch": 1.0941402497598463,
"grad_norm": 0.5555366277694702,
"learning_rate": 0.00013907336053325896,
"loss": 3.184972381591797,
"step": 5695
},
{
"epoch": 1.0951008645533142,
"grad_norm": 1.0896075963974,
"learning_rate": 0.00013883574322611801,
"loss": 3.186721992492676,
"step": 5700
},
{
"epoch": 1.096061479346782,
"grad_norm": 0.8639760613441467,
"learning_rate": 0.00013859815408754218,
"loss": 3.1882835388183595,
"step": 5705
},
{
"epoch": 1.0970220941402498,
"grad_norm": 0.7950151562690735,
"learning_rate": 0.00013836059371699332,
"loss": 3.187421417236328,
"step": 5710
},
{
"epoch": 1.0979827089337175,
"grad_norm": 0.6552711725234985,
"learning_rate": 0.00013812306271386048,
"loss": 3.184224319458008,
"step": 5715
},
{
"epoch": 1.0989433237271853,
"grad_norm": 0.7803718447685242,
"learning_rate": 0.00013788556167745886,
"loss": 3.1838443756103514,
"step": 5720
},
{
"epoch": 1.0999039385206533,
"grad_norm": 0.8713034987449646,
"learning_rate": 0.00013764809120702797,
"loss": 3.188005828857422,
"step": 5725
},
{
"epoch": 1.100864553314121,
"grad_norm": 0.638888955116272,
"learning_rate": 0.00013741065190173008,
"loss": 3.183907890319824,
"step": 5730
},
{
"epoch": 1.1018251681075888,
"grad_norm": 0.8170698285102844,
"learning_rate": 0.00013717324436064897,
"loss": 3.184910011291504,
"step": 5735
},
{
"epoch": 1.1027857829010568,
"grad_norm": 0.8812254667282104,
"learning_rate": 0.0001369358691827882,
"loss": 3.1871992111206056,
"step": 5740
},
{
"epoch": 1.1037463976945245,
"grad_norm": 0.6346907615661621,
"learning_rate": 0.0001366985269670697,
"loss": 3.1861663818359376,
"step": 5745
},
{
"epoch": 1.1047070124879923,
"grad_norm": 0.930957019329071,
"learning_rate": 0.00013646121831233226,
"loss": 3.1821567535400392,
"step": 5750
},
{
"epoch": 1.10566762728146,
"grad_norm": 0.7975543141365051,
"learning_rate": 0.00013622394381732986,
"loss": 3.1874359130859373,
"step": 5755
},
{
"epoch": 1.106628242074928,
"grad_norm": 0.6229592561721802,
"learning_rate": 0.0001359867040807305,
"loss": 3.1849109649658205,
"step": 5760
},
{
"epoch": 1.1075888568683958,
"grad_norm": 0.7393524050712585,
"learning_rate": 0.0001357494997011143,
"loss": 3.1865949630737305,
"step": 5765
},
{
"epoch": 1.1085494716618636,
"grad_norm": 0.7825970649719238,
"learning_rate": 0.0001355123312769723,
"loss": 3.179779815673828,
"step": 5770
},
{
"epoch": 1.1095100864553313,
"grad_norm": 1.018021821975708,
"learning_rate": 0.00013527519940670474,
"loss": 3.1880813598632813,
"step": 5775
},
{
"epoch": 1.1104707012487993,
"grad_norm": 0.7721784114837646,
"learning_rate": 0.00013503810468861967,
"loss": 3.1860368728637694,
"step": 5780
},
{
"epoch": 1.111431316042267,
"grad_norm": 0.7420386075973511,
"learning_rate": 0.00013480104772093136,
"loss": 3.181845855712891,
"step": 5785
},
{
"epoch": 1.1123919308357348,
"grad_norm": 0.7086904644966125,
"learning_rate": 0.0001345640291017588,
"loss": 3.181869125366211,
"step": 5790
},
{
"epoch": 1.1133525456292026,
"grad_norm": 0.5291205048561096,
"learning_rate": 0.00013432704942912445,
"loss": 3.1842918395996094,
"step": 5795
},
{
"epoch": 1.1143131604226706,
"grad_norm": 0.6153315901756287,
"learning_rate": 0.0001340901093009522,
"loss": 3.180573654174805,
"step": 5800
},
{
"epoch": 1.1152737752161384,
"grad_norm": 0.8189062476158142,
"learning_rate": 0.00013385320931506636,
"loss": 3.181578826904297,
"step": 5805
},
{
"epoch": 1.1162343900096061,
"grad_norm": 0.8425748348236084,
"learning_rate": 0.00013361635006918987,
"loss": 3.1840728759765624,
"step": 5810
},
{
"epoch": 1.1171950048030739,
"grad_norm": 0.6964288353919983,
"learning_rate": 0.00013337953216094293,
"loss": 3.183258056640625,
"step": 5815
},
{
"epoch": 1.1181556195965419,
"grad_norm": 0.8112155795097351,
"learning_rate": 0.0001331427561878414,
"loss": 3.1822595596313477,
"step": 5820
},
{
"epoch": 1.1191162343900096,
"grad_norm": 0.5789711475372314,
"learning_rate": 0.0001329060227472953,
"loss": 3.1870792388916014,
"step": 5825
},
{
"epoch": 1.1200768491834774,
"grad_norm": 0.7857502102851868,
"learning_rate": 0.00013266933243660748,
"loss": 3.1816539764404297,
"step": 5830
},
{
"epoch": 1.1210374639769451,
"grad_norm": 0.8584936857223511,
"learning_rate": 0.0001324326858529718,
"loss": 3.1840972900390625,
"step": 5835
},
{
"epoch": 1.1219980787704131,
"grad_norm": 0.7861515879631042,
"learning_rate": 0.00013219608359347194,
"loss": 3.1865585327148436,
"step": 5840
},
{
"epoch": 1.122958693563881,
"grad_norm": 0.6675422787666321,
"learning_rate": 0.0001319595262550796,
"loss": 3.1822277069091798,
"step": 5845
},
{
"epoch": 1.1239193083573487,
"grad_norm": 0.7415298223495483,
"learning_rate": 0.00013172301443465327,
"loss": 3.18204345703125,
"step": 5850
},
{
"epoch": 1.1248799231508164,
"grad_norm": 0.7668789029121399,
"learning_rate": 0.00013148654872893647,
"loss": 3.1823986053466795,
"step": 5855
},
{
"epoch": 1.1258405379442844,
"grad_norm": 1.0284775495529175,
"learning_rate": 0.00013125012973455645,
"loss": 3.186284637451172,
"step": 5860
},
{
"epoch": 1.1268011527377522,
"grad_norm": 0.9501076936721802,
"learning_rate": 0.00013101375804802268,
"loss": 3.180257034301758,
"step": 5865
},
{
"epoch": 1.12776176753122,
"grad_norm": 0.6620021462440491,
"learning_rate": 0.00013077743426572508,
"loss": 3.1850921630859377,
"step": 5870
},
{
"epoch": 1.1287223823246877,
"grad_norm": 0.6377175450325012,
"learning_rate": 0.0001305411589839328,
"loss": 3.1847787857055665,
"step": 5875
},
{
"epoch": 1.1296829971181557,
"grad_norm": 0.8956018686294556,
"learning_rate": 0.00013030493279879267,
"loss": 3.1820547103881838,
"step": 5880
},
{
"epoch": 1.1306436119116234,
"grad_norm": 0.9648638367652893,
"learning_rate": 0.00013006875630632757,
"loss": 3.1855083465576173,
"step": 5885
},
{
"epoch": 1.1316042267050912,
"grad_norm": 0.7749320864677429,
"learning_rate": 0.000129832630102435,
"loss": 3.18371639251709,
"step": 5890
},
{
"epoch": 1.1325648414985592,
"grad_norm": 0.7162895202636719,
"learning_rate": 0.00012959655478288556,
"loss": 3.1814332962036134,
"step": 5895
},
{
"epoch": 1.133525456292027,
"grad_norm": 0.7946479320526123,
"learning_rate": 0.00012936053094332158,
"loss": 3.183310699462891,
"step": 5900
},
{
"epoch": 1.1344860710854947,
"grad_norm": 0.9230923652648926,
"learning_rate": 0.00012912455917925535,
"loss": 3.1882291793823243,
"step": 5905
},
{
"epoch": 1.1354466858789625,
"grad_norm": 0.6934447288513184,
"learning_rate": 0.0001288886400860679,
"loss": 3.1866472244262694,
"step": 5910
},
{
"epoch": 1.1364073006724302,
"grad_norm": 0.5946722626686096,
"learning_rate": 0.00012865277425900724,
"loss": 3.1836269378662108,
"step": 5915
},
{
"epoch": 1.1373679154658982,
"grad_norm": 0.6142674088478088,
"learning_rate": 0.00012841696229318712,
"loss": 3.1850711822509767,
"step": 5920
},
{
"epoch": 1.138328530259366,
"grad_norm": 0.5882613062858582,
"learning_rate": 0.00012818120478358522,
"loss": 3.1808708190917967,
"step": 5925
},
{
"epoch": 1.1392891450528337,
"grad_norm": 1.1265987157821655,
"learning_rate": 0.000127945502325042,
"loss": 3.1862953186035154,
"step": 5930
},
{
"epoch": 1.1402497598463017,
"grad_norm": 0.8293631076812744,
"learning_rate": 0.00012770985551225898,
"loss": 3.1817962646484377,
"step": 5935
},
{
"epoch": 1.1412103746397695,
"grad_norm": 0.6100813150405884,
"learning_rate": 0.00012747426493979713,
"loss": 3.182262420654297,
"step": 5940
},
{
"epoch": 1.1421709894332372,
"grad_norm": 0.6388895511627197,
"learning_rate": 0.00012723873120207575,
"loss": 3.183791732788086,
"step": 5945
},
{
"epoch": 1.143131604226705,
"grad_norm": 1.1153887510299683,
"learning_rate": 0.00012700325489337056,
"loss": 3.1867645263671873,
"step": 5950
},
{
"epoch": 1.144092219020173,
"grad_norm": 0.6275951862335205,
"learning_rate": 0.00012676783660781245,
"loss": 3.183899688720703,
"step": 5955
},
{
"epoch": 1.1450528338136408,
"grad_norm": 0.6649777293205261,
"learning_rate": 0.00012653247693938588,
"loss": 3.1856327056884766,
"step": 5960
},
{
"epoch": 1.1460134486071085,
"grad_norm": 0.7300867438316345,
"learning_rate": 0.00012629717648192747,
"loss": 3.1813177108764648,
"step": 5965
},
{
"epoch": 1.1469740634005763,
"grad_norm": 0.8946056365966797,
"learning_rate": 0.00012606193582912446,
"loss": 3.182071304321289,
"step": 5970
},
{
"epoch": 1.1479346781940443,
"grad_norm": 0.8307191729545593,
"learning_rate": 0.00012582675557451306,
"loss": 3.1808849334716798,
"step": 5975
},
{
"epoch": 1.148895292987512,
"grad_norm": 0.8783102631568909,
"learning_rate": 0.00012559163631147723,
"loss": 3.184329032897949,
"step": 5980
},
{
"epoch": 1.1498559077809798,
"grad_norm": 0.7081605792045593,
"learning_rate": 0.00012535657863324695,
"loss": 3.1825115203857424,
"step": 5985
},
{
"epoch": 1.1508165225744476,
"grad_norm": 0.5934545993804932,
"learning_rate": 0.0001251215831328969,
"loss": 3.1839290618896485,
"step": 5990
},
{
"epoch": 1.1517771373679155,
"grad_norm": 0.8041390180587769,
"learning_rate": 0.00012488665040334479,
"loss": 3.183102607727051,
"step": 5995
},
{
"epoch": 1.1527377521613833,
"grad_norm": 0.6526947617530823,
"learning_rate": 0.00012465178103735,
"loss": 3.1817670822143556,
"step": 6000
},
{
"epoch": 1.153698366954851,
"grad_norm": 0.6892653703689575,
"learning_rate": 0.00012441697562751215,
"loss": 3.180023193359375,
"step": 6005
},
{
"epoch": 1.154658981748319,
"grad_norm": 0.6279175877571106,
"learning_rate": 0.00012418223476626925,
"loss": 3.180769157409668,
"step": 6010
},
{
"epoch": 1.1556195965417868,
"grad_norm": 0.8991310596466064,
"learning_rate": 0.00012394755904589664,
"loss": 3.185033416748047,
"step": 6015
},
{
"epoch": 1.1565802113352546,
"grad_norm": 0.6720035672187805,
"learning_rate": 0.00012371294905850518,
"loss": 3.1769649505615236,
"step": 6020
},
{
"epoch": 1.1575408261287223,
"grad_norm": 0.8395649790763855,
"learning_rate": 0.00012347840539603998,
"loss": 3.1827545166015625,
"step": 6025
},
{
"epoch": 1.15850144092219,
"grad_norm": 0.8189507722854614,
"learning_rate": 0.00012324392865027873,
"loss": 3.1804759979248045,
"step": 6030
},
{
"epoch": 1.159462055715658,
"grad_norm": 0.9116944074630737,
"learning_rate": 0.00012300951941283036,
"loss": 3.185091018676758,
"step": 6035
},
{
"epoch": 1.1604226705091258,
"grad_norm": 0.6200742125511169,
"learning_rate": 0.00012277517827513341,
"loss": 3.1775371551513674,
"step": 6040
},
{
"epoch": 1.1613832853025936,
"grad_norm": 0.6865299344062805,
"learning_rate": 0.00012254090582845457,
"loss": 3.1813201904296875,
"step": 6045
},
{
"epoch": 1.1623439000960616,
"grad_norm": 0.7450284957885742,
"learning_rate": 0.00012230670266388728,
"loss": 3.18134765625,
"step": 6050
},
{
"epoch": 1.1633045148895294,
"grad_norm": 0.814132571220398,
"learning_rate": 0.00012207256937235014,
"loss": 3.182879638671875,
"step": 6055
},
{
"epoch": 1.1642651296829971,
"grad_norm": 0.8822999000549316,
"learning_rate": 0.00012183850654458553,
"loss": 3.1848114013671873,
"step": 6060
},
{
"epoch": 1.1652257444764649,
"grad_norm": 0.5908654928207397,
"learning_rate": 0.00012160451477115783,
"loss": 3.1844045639038088,
"step": 6065
},
{
"epoch": 1.1661863592699326,
"grad_norm": 0.5196489095687866,
"learning_rate": 0.00012137059464245242,
"loss": 3.181051254272461,
"step": 6070
},
{
"epoch": 1.1671469740634006,
"grad_norm": 0.7493326663970947,
"learning_rate": 0.00012113674674867379,
"loss": 3.180977630615234,
"step": 6075
},
{
"epoch": 1.1681075888568684,
"grad_norm": 0.8588430881500244,
"learning_rate": 0.00012090297167984409,
"loss": 3.183562088012695,
"step": 6080
},
{
"epoch": 1.1690682036503361,
"grad_norm": 0.707949697971344,
"learning_rate": 0.00012066927002580185,
"loss": 3.186185836791992,
"step": 6085
},
{
"epoch": 1.1700288184438041,
"grad_norm": 0.7008398771286011,
"learning_rate": 0.00012043564237620028,
"loss": 3.183705520629883,
"step": 6090
},
{
"epoch": 1.170989433237272,
"grad_norm": 0.5148990750312805,
"learning_rate": 0.00012020208932050595,
"loss": 3.180450439453125,
"step": 6095
},
{
"epoch": 1.1719500480307397,
"grad_norm": 1.0000126361846924,
"learning_rate": 0.00011996861144799712,
"loss": 3.1828609466552735,
"step": 6100
},
{
"epoch": 1.1729106628242074,
"grad_norm": 0.9941655993461609,
"learning_rate": 0.0001197352093477625,
"loss": 3.1851844787597656,
"step": 6105
},
{
"epoch": 1.1738712776176754,
"grad_norm": 0.6262122988700867,
"learning_rate": 0.00011950188360869947,
"loss": 3.186477279663086,
"step": 6110
},
{
"epoch": 1.1748318924111432,
"grad_norm": 0.7170460224151611,
"learning_rate": 0.00011926863481951279,
"loss": 3.181361770629883,
"step": 6115
},
{
"epoch": 1.175792507204611,
"grad_norm": 0.5651366114616394,
"learning_rate": 0.00011903546356871315,
"loss": 3.180009460449219,
"step": 6120
},
{
"epoch": 1.1767531219980787,
"grad_norm": 0.7348089814186096,
"learning_rate": 0.00011880237044461546,
"loss": 3.1843181610107423,
"step": 6125
},
{
"epoch": 1.1777137367915467,
"grad_norm": 0.8490554690361023,
"learning_rate": 0.00011856935603533759,
"loss": 3.1855024337768554,
"step": 6130
},
{
"epoch": 1.1786743515850144,
"grad_norm": 0.688126802444458,
"learning_rate": 0.0001183364209287989,
"loss": 3.1825450897216796,
"step": 6135
},
{
"epoch": 1.1796349663784822,
"grad_norm": 0.6043757796287537,
"learning_rate": 0.00011810356571271847,
"loss": 3.1808750152587892,
"step": 6140
},
{
"epoch": 1.18059558117195,
"grad_norm": 0.6653470396995544,
"learning_rate": 0.00011787079097461394,
"loss": 3.185806655883789,
"step": 6145
},
{
"epoch": 1.181556195965418,
"grad_norm": 0.5727336406707764,
"learning_rate": 0.0001176380973017998,
"loss": 3.182196617126465,
"step": 6150
},
{
"epoch": 1.1825168107588857,
"grad_norm": 0.4174487292766571,
"learning_rate": 0.00011740548528138613,
"loss": 3.1803747177124024,
"step": 6155
},
{
"epoch": 1.1834774255523535,
"grad_norm": 0.8878217339515686,
"learning_rate": 0.00011717295550027685,
"loss": 3.1840145111083986,
"step": 6160
},
{
"epoch": 1.1844380403458212,
"grad_norm": 0.8910025954246521,
"learning_rate": 0.00011694050854516853,
"loss": 3.181249237060547,
"step": 6165
},
{
"epoch": 1.1853986551392892,
"grad_norm": 0.7205166220664978,
"learning_rate": 0.00011670814500254858,
"loss": 3.1778148651123046,
"step": 6170
},
{
"epoch": 1.186359269932757,
"grad_norm": 0.6521825790405273,
"learning_rate": 0.00011647586545869418,
"loss": 3.17816276550293,
"step": 6175
},
{
"epoch": 1.1873198847262247,
"grad_norm": 0.6897614002227783,
"learning_rate": 0.00011624367049967037,
"loss": 3.1780323028564452,
"step": 6180
},
{
"epoch": 1.1882804995196925,
"grad_norm": 0.5899947881698608,
"learning_rate": 0.00011601156071132883,
"loss": 3.181020164489746,
"step": 6185
},
{
"epoch": 1.1892411143131605,
"grad_norm": 0.6569263935089111,
"learning_rate": 0.00011577953667930641,
"loss": 3.1823570251464846,
"step": 6190
},
{
"epoch": 1.1902017291066282,
"grad_norm": 0.6824354529380798,
"learning_rate": 0.00011554759898902355,
"loss": 3.1782264709472656,
"step": 6195
},
{
"epoch": 1.191162343900096,
"grad_norm": 0.5225342512130737,
"learning_rate": 0.00011531574822568278,
"loss": 3.1797101974487303,
"step": 6200
},
{
"epoch": 1.192122958693564,
"grad_norm": 0.5605508685112,
"learning_rate": 0.00011508398497426745,
"loss": 3.1792762756347654,
"step": 6205
},
{
"epoch": 1.1930835734870318,
"grad_norm": 0.8098820447921753,
"learning_rate": 0.00011485230981953997,
"loss": 3.183140182495117,
"step": 6210
},
{
"epoch": 1.1940441882804995,
"grad_norm": 0.6241822838783264,
"learning_rate": 0.00011462072334604053,
"loss": 3.1846290588378907,
"step": 6215
},
{
"epoch": 1.1950048030739673,
"grad_norm": 0.6163104772567749,
"learning_rate": 0.00011438922613808554,
"loss": 3.184892272949219,
"step": 6220
},
{
"epoch": 1.195965417867435,
"grad_norm": 0.6423901915550232,
"learning_rate": 0.00011415781877976628,
"loss": 3.180209922790527,
"step": 6225
},
{
"epoch": 1.196926032660903,
"grad_norm": 0.743615448474884,
"learning_rate": 0.00011392650185494712,
"loss": 3.181998634338379,
"step": 6230
},
{
"epoch": 1.1978866474543708,
"grad_norm": 0.7482926845550537,
"learning_rate": 0.00011369527594726452,
"loss": 3.178305244445801,
"step": 6235
},
{
"epoch": 1.1988472622478386,
"grad_norm": 0.6290075778961182,
"learning_rate": 0.0001134641416401251,
"loss": 3.182933807373047,
"step": 6240
},
{
"epoch": 1.1998078770413065,
"grad_norm": 0.5498515963554382,
"learning_rate": 0.00011323309951670444,
"loss": 3.1798477172851562,
"step": 6245
},
{
"epoch": 1.2007684918347743,
"grad_norm": 0.6985566020011902,
"learning_rate": 0.00011300215015994554,
"loss": 3.1814619064331056,
"step": 6250
},
{
"epoch": 1.201729106628242,
"grad_norm": 0.8303685784339905,
"learning_rate": 0.00011277129415255727,
"loss": 3.1828535079956053,
"step": 6255
},
{
"epoch": 1.2026897214217098,
"grad_norm": 0.7280902862548828,
"learning_rate": 0.00011254053207701308,
"loss": 3.17972354888916,
"step": 6260
},
{
"epoch": 1.2036503362151778,
"grad_norm": 0.6531400084495544,
"learning_rate": 0.00011230986451554932,
"loss": 3.1785831451416016,
"step": 6265
},
{
"epoch": 1.2046109510086456,
"grad_norm": 0.6152121424674988,
"learning_rate": 0.00011207929205016386,
"loss": 3.177712631225586,
"step": 6270
},
{
"epoch": 1.2055715658021133,
"grad_norm": 0.5569039583206177,
"learning_rate": 0.00011184881526261483,
"loss": 3.1826234817504884,
"step": 6275
},
{
"epoch": 1.206532180595581,
"grad_norm": 1.1272635459899902,
"learning_rate": 0.00011161843473441867,
"loss": 3.179414749145508,
"step": 6280
},
{
"epoch": 1.207492795389049,
"grad_norm": 0.6206833720207214,
"learning_rate": 0.0001113881510468492,
"loss": 3.184890365600586,
"step": 6285
},
{
"epoch": 1.2084534101825168,
"grad_norm": 0.6447250247001648,
"learning_rate": 0.00011115796478093569,
"loss": 3.1766143798828126,
"step": 6290
},
{
"epoch": 1.2094140249759846,
"grad_norm": 0.7120223641395569,
"learning_rate": 0.00011092787651746178,
"loss": 3.1796751022338867,
"step": 6295
},
{
"epoch": 1.2103746397694524,
"grad_norm": 0.620508074760437,
"learning_rate": 0.00011069788683696366,
"loss": 3.1746284484863283,
"step": 6300
},
{
"epoch": 1.2113352545629203,
"grad_norm": 0.6917296051979065,
"learning_rate": 0.00011046799631972903,
"loss": 3.1808353424072267,
"step": 6305
},
{
"epoch": 1.2122958693563881,
"grad_norm": 0.7654911875724792,
"learning_rate": 0.00011023820554579512,
"loss": 3.1830776214599608,
"step": 6310
},
{
"epoch": 1.2132564841498559,
"grad_norm": 0.7507694959640503,
"learning_rate": 0.00011000851509494771,
"loss": 3.181525230407715,
"step": 6315
},
{
"epoch": 1.2142170989433236,
"grad_norm": 0.6086501479148865,
"learning_rate": 0.00010977892554671935,
"loss": 3.1778955459594727,
"step": 6320
},
{
"epoch": 1.2151777137367916,
"grad_norm": 0.5555998682975769,
"learning_rate": 0.00010954943748038798,
"loss": 3.1771930694580077,
"step": 6325
},
{
"epoch": 1.2161383285302594,
"grad_norm": 0.6753772497177124,
"learning_rate": 0.00010932005147497559,
"loss": 3.178631591796875,
"step": 6330
},
{
"epoch": 1.2170989433237271,
"grad_norm": 0.5674150586128235,
"learning_rate": 0.00010909076810924654,
"loss": 3.183444786071777,
"step": 6335
},
{
"epoch": 1.218059558117195,
"grad_norm": 0.8342090249061584,
"learning_rate": 0.00010886158796170633,
"loss": 3.181511116027832,
"step": 6340
},
{
"epoch": 1.219020172910663,
"grad_norm": 0.9122212529182434,
"learning_rate": 0.00010863251161060002,
"loss": 3.1852695465087892,
"step": 6345
},
{
"epoch": 1.2199807877041307,
"grad_norm": 0.44202694296836853,
"learning_rate": 0.00010840353963391065,
"loss": 3.1827775955200197,
"step": 6350
},
{
"epoch": 1.2209414024975984,
"grad_norm": 0.5556421279907227,
"learning_rate": 0.0001081746726093581,
"loss": 3.176057815551758,
"step": 6355
},
{
"epoch": 1.2219020172910664,
"grad_norm": 0.5319374203681946,
"learning_rate": 0.00010794591111439729,
"loss": 3.177410125732422,
"step": 6360
},
{
"epoch": 1.2228626320845342,
"grad_norm": 0.7626725435256958,
"learning_rate": 0.000107717255726217,
"loss": 3.182592582702637,
"step": 6365
},
{
"epoch": 1.223823246878002,
"grad_norm": 0.7391705513000488,
"learning_rate": 0.00010748870702173815,
"loss": 3.178932952880859,
"step": 6370
},
{
"epoch": 1.2247838616714697,
"grad_norm": 0.4964456260204315,
"learning_rate": 0.00010726026557761264,
"loss": 3.1790546417236327,
"step": 6375
},
{
"epoch": 1.2257444764649374,
"grad_norm": 0.5090710520744324,
"learning_rate": 0.00010703193197022159,
"loss": 3.1741233825683595,
"step": 6380
},
{
"epoch": 1.2267050912584054,
"grad_norm": 0.8102352023124695,
"learning_rate": 0.00010680370677567423,
"loss": 3.184102249145508,
"step": 6385
},
{
"epoch": 1.2276657060518732,
"grad_norm": 0.7811996340751648,
"learning_rate": 0.00010657559056980603,
"loss": 3.1772319793701174,
"step": 6390
},
{
"epoch": 1.228626320845341,
"grad_norm": 0.7701186537742615,
"learning_rate": 0.00010634758392817762,
"loss": 3.1799732208251954,
"step": 6395
},
{
"epoch": 1.229586935638809,
"grad_norm": 0.685007631778717,
"learning_rate": 0.00010611968742607318,
"loss": 3.176517105102539,
"step": 6400
},
{
"epoch": 1.2305475504322767,
"grad_norm": 0.5316605567932129,
"learning_rate": 0.00010589190163849885,
"loss": 3.1800348281860353,
"step": 6405
},
{
"epoch": 1.2315081652257445,
"grad_norm": 0.6539739370346069,
"learning_rate": 0.00010566422714018167,
"loss": 3.1817600250244142,
"step": 6410
},
{
"epoch": 1.2324687800192122,
"grad_norm": 0.5445916056632996,
"learning_rate": 0.00010543666450556774,
"loss": 3.1751848220825196,
"step": 6415
},
{
"epoch": 1.23342939481268,
"grad_norm": 0.6994644403457642,
"learning_rate": 0.00010520921430882088,
"loss": 3.182989311218262,
"step": 6420
},
{
"epoch": 1.234390009606148,
"grad_norm": 0.6963391304016113,
"learning_rate": 0.00010498187712382132,
"loss": 3.1813974380493164,
"step": 6425
},
{
"epoch": 1.2353506243996157,
"grad_norm": 0.8707190752029419,
"learning_rate": 0.00010475465352416402,
"loss": 3.1847021102905275,
"step": 6430
},
{
"epoch": 1.2363112391930835,
"grad_norm": 0.628430962562561,
"learning_rate": 0.00010452754408315754,
"loss": 3.180718994140625,
"step": 6435
},
{
"epoch": 1.2372718539865515,
"grad_norm": 0.4727374017238617,
"learning_rate": 0.00010430054937382219,
"loss": 3.179873466491699,
"step": 6440
},
{
"epoch": 1.2382324687800192,
"grad_norm": 0.6456632614135742,
"learning_rate": 0.000104073669968889,
"loss": 3.1764720916748046,
"step": 6445
},
{
"epoch": 1.239193083573487,
"grad_norm": 0.9211506843566895,
"learning_rate": 0.0001038469064407979,
"loss": 3.175563430786133,
"step": 6450
},
{
"epoch": 1.2401536983669548,
"grad_norm": 0.7014700770378113,
"learning_rate": 0.00010362025936169664,
"loss": 3.1777694702148436,
"step": 6455
},
{
"epoch": 1.2411143131604228,
"grad_norm": 0.612933874130249,
"learning_rate": 0.00010339372930343896,
"loss": 3.177769088745117,
"step": 6460
},
{
"epoch": 1.2420749279538905,
"grad_norm": 0.47281619906425476,
"learning_rate": 0.00010316731683758346,
"loss": 3.177565574645996,
"step": 6465
},
{
"epoch": 1.2430355427473583,
"grad_norm": 0.6775256395339966,
"learning_rate": 0.00010294102253539204,
"loss": 3.1776294708251953,
"step": 6470
},
{
"epoch": 1.243996157540826,
"grad_norm": 0.7892163395881653,
"learning_rate": 0.00010271484696782837,
"loss": 3.175709342956543,
"step": 6475
},
{
"epoch": 1.244956772334294,
"grad_norm": 0.6570128202438354,
"learning_rate": 0.00010248879070555668,
"loss": 3.1776430130004885,
"step": 6480
},
{
"epoch": 1.2459173871277618,
"grad_norm": 0.5200539231300354,
"learning_rate": 0.00010226285431894013,
"loss": 3.176724433898926,
"step": 6485
},
{
"epoch": 1.2468780019212296,
"grad_norm": 0.7223069667816162,
"learning_rate": 0.00010203703837803934,
"loss": 3.176532745361328,
"step": 6490
},
{
"epoch": 1.2478386167146973,
"grad_norm": 0.8052407503128052,
"learning_rate": 0.00010181134345261115,
"loss": 3.176080322265625,
"step": 6495
},
{
"epoch": 1.2487992315081653,
"grad_norm": 0.48731186985969543,
"learning_rate": 0.00010158577011210695,
"loss": 3.173239517211914,
"step": 6500
},
{
"epoch": 1.249759846301633,
"grad_norm": 0.48055845499038696,
"learning_rate": 0.00010136031892567145,
"loss": 3.1757953643798826,
"step": 6505
},
{
"epoch": 1.2507204610951008,
"grad_norm": 0.9334460496902466,
"learning_rate": 0.00010113499046214113,
"loss": 3.173488998413086,
"step": 6510
},
{
"epoch": 1.2516810758885688,
"grad_norm": 0.747382402420044,
"learning_rate": 0.00010090978529004284,
"loss": 3.1791568756103517,
"step": 6515
},
{
"epoch": 1.2526416906820366,
"grad_norm": 0.5684623718261719,
"learning_rate": 0.00010068470397759226,
"loss": 3.1786075592041017,
"step": 6520
},
{
"epoch": 1.2536023054755043,
"grad_norm": 0.7101691365242004,
"learning_rate": 0.0001004597470926927,
"loss": 3.17834587097168,
"step": 6525
},
{
"epoch": 1.254562920268972,
"grad_norm": 0.6113804578781128,
"learning_rate": 0.0001002349152029334,
"loss": 3.1760730743408203,
"step": 6530
},
{
"epoch": 1.2555235350624399,
"grad_norm": 0.634519100189209,
"learning_rate": 0.00010001020887558837,
"loss": 3.1759071350097656,
"step": 6535
},
{
"epoch": 1.2564841498559078,
"grad_norm": 0.6819270253181458,
"learning_rate": 9.978562867761465e-05,
"loss": 3.1800254821777343,
"step": 6540
},
{
"epoch": 1.2574447646493756,
"grad_norm": 0.6920965313911438,
"learning_rate": 9.95611751756511e-05,
"loss": 3.179019546508789,
"step": 6545
},
{
"epoch": 1.2584053794428434,
"grad_norm": 0.7164304256439209,
"learning_rate": 9.9336848936017e-05,
"loss": 3.174900436401367,
"step": 6550
},
{
"epoch": 1.2593659942363113,
"grad_norm": 0.5404880046844482,
"learning_rate": 9.911265052471046e-05,
"loss": 3.1754596710205076,
"step": 6555
},
{
"epoch": 1.260326609029779,
"grad_norm": 0.6277545690536499,
"learning_rate": 9.888858050740703e-05,
"loss": 3.1767545700073243,
"step": 6560
},
{
"epoch": 1.2612872238232469,
"grad_norm": 0.4282405972480774,
"learning_rate": 9.866463944945839e-05,
"loss": 3.175259590148926,
"step": 6565
},
{
"epoch": 1.2622478386167146,
"grad_norm": 0.740606963634491,
"learning_rate": 9.844082791589075e-05,
"loss": 3.1768466949462892,
"step": 6570
},
{
"epoch": 1.2632084534101824,
"grad_norm": 0.6880332231521606,
"learning_rate": 9.821714647140363e-05,
"loss": 3.1770793914794924,
"step": 6575
},
{
"epoch": 1.2641690682036504,
"grad_norm": 0.5514328479766846,
"learning_rate": 9.799359568036821e-05,
"loss": 3.180289649963379,
"step": 6580
},
{
"epoch": 1.2651296829971181,
"grad_norm": 0.569648027420044,
"learning_rate": 9.777017610682612e-05,
"loss": 3.1760627746582033,
"step": 6585
},
{
"epoch": 1.266090297790586,
"grad_norm": 0.6740429401397705,
"learning_rate": 9.754688831448781e-05,
"loss": 3.173795700073242,
"step": 6590
},
{
"epoch": 1.267050912584054,
"grad_norm": 0.6410177946090698,
"learning_rate": 9.732373286673135e-05,
"loss": 3.178542709350586,
"step": 6595
},
{
"epoch": 1.2680115273775217,
"grad_norm": 0.5730255246162415,
"learning_rate": 9.710071032660076e-05,
"loss": 3.1776718139648437,
"step": 6600
},
{
"epoch": 1.2689721421709894,
"grad_norm": 0.6753981113433838,
"learning_rate": 9.687782125680484e-05,
"loss": 3.1791826248168946,
"step": 6605
},
{
"epoch": 1.2699327569644572,
"grad_norm": 0.5752228498458862,
"learning_rate": 9.66550662197155e-05,
"loss": 3.173518753051758,
"step": 6610
},
{
"epoch": 1.270893371757925,
"grad_norm": 0.5868836045265198,
"learning_rate": 9.643244577736667e-05,
"loss": 3.173557090759277,
"step": 6615
},
{
"epoch": 1.271853986551393,
"grad_norm": 0.5383107662200928,
"learning_rate": 9.620996049145247e-05,
"loss": 3.1773754119873048,
"step": 6620
},
{
"epoch": 1.2728146013448607,
"grad_norm": 0.5838056802749634,
"learning_rate": 9.598761092332616e-05,
"loss": 3.1778125762939453,
"step": 6625
},
{
"epoch": 1.2737752161383284,
"grad_norm": 0.5867598056793213,
"learning_rate": 9.576539763399847e-05,
"loss": 3.183711814880371,
"step": 6630
},
{
"epoch": 1.2747358309317964,
"grad_norm": 0.6111621856689453,
"learning_rate": 9.554332118413635e-05,
"loss": 3.1793869018554686,
"step": 6635
},
{
"epoch": 1.2756964457252642,
"grad_norm": 0.8043511509895325,
"learning_rate": 9.532138213406143e-05,
"loss": 3.175552749633789,
"step": 6640
},
{
"epoch": 1.276657060518732,
"grad_norm": 0.8225224018096924,
"learning_rate": 9.509958104374877e-05,
"loss": 3.1786468505859373,
"step": 6645
},
{
"epoch": 1.2776176753121997,
"grad_norm": 0.6611747741699219,
"learning_rate": 9.487791847282517e-05,
"loss": 3.1788875579833986,
"step": 6650
},
{
"epoch": 1.2785782901056677,
"grad_norm": 0.49061068892478943,
"learning_rate": 9.465639498056815e-05,
"loss": 3.1779201507568358,
"step": 6655
},
{
"epoch": 1.2795389048991355,
"grad_norm": 0.6303700804710388,
"learning_rate": 9.443501112590413e-05,
"loss": 3.176443862915039,
"step": 6660
},
{
"epoch": 1.2804995196926032,
"grad_norm": 0.7213342189788818,
"learning_rate": 9.421376746740733e-05,
"loss": 3.1703159332275392,
"step": 6665
},
{
"epoch": 1.2814601344860712,
"grad_norm": 0.6647571921348572,
"learning_rate": 9.399266456329815e-05,
"loss": 3.1765552520751954,
"step": 6670
},
{
"epoch": 1.282420749279539,
"grad_norm": 0.6062517762184143,
"learning_rate": 9.377170297144196e-05,
"loss": 3.1765026092529296,
"step": 6675
},
{
"epoch": 1.2833813640730067,
"grad_norm": 0.4922938048839569,
"learning_rate": 9.355088324934741e-05,
"loss": 3.176412582397461,
"step": 6680
},
{
"epoch": 1.2843419788664745,
"grad_norm": 0.5771006941795349,
"learning_rate": 9.333020595416548e-05,
"loss": 3.1751476287841798,
"step": 6685
},
{
"epoch": 1.2853025936599423,
"grad_norm": 0.6197171211242676,
"learning_rate": 9.310967164268749e-05,
"loss": 3.174680709838867,
"step": 6690
},
{
"epoch": 1.2862632084534102,
"grad_norm": 0.5435570478439331,
"learning_rate": 9.28892808713442e-05,
"loss": 3.1765819549560548,
"step": 6695
},
{
"epoch": 1.287223823246878,
"grad_norm": 0.5521527528762817,
"learning_rate": 9.266903419620411e-05,
"loss": 3.173064041137695,
"step": 6700
},
{
"epoch": 1.2881844380403458,
"grad_norm": 0.5107563138008118,
"learning_rate": 9.24489321729722e-05,
"loss": 3.1766895294189452,
"step": 6705
},
{
"epoch": 1.2891450528338138,
"grad_norm": 0.66566401720047,
"learning_rate": 9.222897535698841e-05,
"loss": 3.175777053833008,
"step": 6710
},
{
"epoch": 1.2901056676272815,
"grad_norm": 0.501441478729248,
"learning_rate": 9.20091643032264e-05,
"loss": 3.1773338317871094,
"step": 6715
},
{
"epoch": 1.2910662824207493,
"grad_norm": 0.7562515735626221,
"learning_rate": 9.178949956629195e-05,
"loss": 3.1776708602905273,
"step": 6720
},
{
"epoch": 1.292026897214217,
"grad_norm": 0.7048882246017456,
"learning_rate": 9.156998170042182e-05,
"loss": 3.1779850006103514,
"step": 6725
},
{
"epoch": 1.2929875120076848,
"grad_norm": 0.5082115530967712,
"learning_rate": 9.1350611259482e-05,
"loss": 3.174315071105957,
"step": 6730
},
{
"epoch": 1.2939481268011528,
"grad_norm": 0.5481780171394348,
"learning_rate": 9.113138879696667e-05,
"loss": 3.174916648864746,
"step": 6735
},
{
"epoch": 1.2949087415946205,
"grad_norm": 0.5947979688644409,
"learning_rate": 9.091231486599656e-05,
"loss": 3.177842903137207,
"step": 6740
},
{
"epoch": 1.2958693563880883,
"grad_norm": 0.5829041600227356,
"learning_rate": 9.069339001931775e-05,
"loss": 3.1755516052246096,
"step": 6745
},
{
"epoch": 1.2968299711815563,
"grad_norm": 0.7779812812805176,
"learning_rate": 9.047461480929996e-05,
"loss": 3.1758398056030273,
"step": 6750
},
{
"epoch": 1.297790585975024,
"grad_norm": 0.5685781836509705,
"learning_rate": 9.025598978793564e-05,
"loss": 3.1764198303222657,
"step": 6755
},
{
"epoch": 1.2987512007684918,
"grad_norm": 0.5285525918006897,
"learning_rate": 9.003751550683803e-05,
"loss": 3.1752038955688477,
"step": 6760
},
{
"epoch": 1.2997118155619596,
"grad_norm": 0.6162580251693726,
"learning_rate": 8.981919251724023e-05,
"loss": 3.172860336303711,
"step": 6765
},
{
"epoch": 1.3006724303554273,
"grad_norm": 0.6464502215385437,
"learning_rate": 8.960102136999346e-05,
"loss": 3.174723815917969,
"step": 6770
},
{
"epoch": 1.3016330451488953,
"grad_norm": 0.472971111536026,
"learning_rate": 8.938300261556599e-05,
"loss": 3.176887321472168,
"step": 6775
},
{
"epoch": 1.302593659942363,
"grad_norm": 0.46003440022468567,
"learning_rate": 8.916513680404145e-05,
"loss": 3.176510047912598,
"step": 6780
},
{
"epoch": 1.3035542747358309,
"grad_norm": 0.7004385590553284,
"learning_rate": 8.894742448511766e-05,
"loss": 3.1788423538208006,
"step": 6785
},
{
"epoch": 1.3045148895292988,
"grad_norm": 0.6769760847091675,
"learning_rate": 8.872986620810508e-05,
"loss": 3.175991439819336,
"step": 6790
},
{
"epoch": 1.3054755043227666,
"grad_norm": 0.593636691570282,
"learning_rate": 8.851246252192566e-05,
"loss": 3.1727279663085937,
"step": 6795
},
{
"epoch": 1.3064361191162344,
"grad_norm": 0.580614447593689,
"learning_rate": 8.82952139751111e-05,
"loss": 3.178467559814453,
"step": 6800
},
{
"epoch": 1.3073967339097021,
"grad_norm": 0.519120991230011,
"learning_rate": 8.807812111580186e-05,
"loss": 3.1754281997680662,
"step": 6805
},
{
"epoch": 1.30835734870317,
"grad_norm": 0.5653340816497803,
"learning_rate": 8.786118449174541e-05,
"loss": 3.1773509979248047,
"step": 6810
},
{
"epoch": 1.3093179634966379,
"grad_norm": 0.7600631713867188,
"learning_rate": 8.764440465029513e-05,
"loss": 3.1753978729248047,
"step": 6815
},
{
"epoch": 1.3102785782901056,
"grad_norm": 0.5709202885627747,
"learning_rate": 8.742778213840888e-05,
"loss": 3.176059341430664,
"step": 6820
},
{
"epoch": 1.3112391930835736,
"grad_norm": 0.5290617346763611,
"learning_rate": 8.721131750264737e-05,
"loss": 3.176310348510742,
"step": 6825
},
{
"epoch": 1.3121998078770414,
"grad_norm": 0.5382722020149231,
"learning_rate": 8.699501128917321e-05,
"loss": 3.1754171371459963,
"step": 6830
},
{
"epoch": 1.3131604226705091,
"grad_norm": 0.5834631323814392,
"learning_rate": 8.677886404374903e-05,
"loss": 3.176200103759766,
"step": 6835
},
{
"epoch": 1.314121037463977,
"grad_norm": 0.44905054569244385,
"learning_rate": 8.656287631173656e-05,
"loss": 3.178561210632324,
"step": 6840
},
{
"epoch": 1.3150816522574447,
"grad_norm": 0.4696488380432129,
"learning_rate": 8.634704863809501e-05,
"loss": 3.1775485992431642,
"step": 6845
},
{
"epoch": 1.3160422670509127,
"grad_norm": 0.5930847525596619,
"learning_rate": 8.61313815673798e-05,
"loss": 3.176204299926758,
"step": 6850
},
{
"epoch": 1.3170028818443804,
"grad_norm": 0.6078386902809143,
"learning_rate": 8.591587564374094e-05,
"loss": 3.175630569458008,
"step": 6855
},
{
"epoch": 1.3179634966378482,
"grad_norm": 0.6279536485671997,
"learning_rate": 8.570053141092209e-05,
"loss": 3.1791570663452147,
"step": 6860
},
{
"epoch": 1.3189241114313162,
"grad_norm": 0.4797205328941345,
"learning_rate": 8.548534941225877e-05,
"loss": 3.1787494659423827,
"step": 6865
},
{
"epoch": 1.319884726224784,
"grad_norm": 0.692002534866333,
"learning_rate": 8.527033019067736e-05,
"loss": 3.1755470275878905,
"step": 6870
},
{
"epoch": 1.3208453410182517,
"grad_norm": 0.609710693359375,
"learning_rate": 8.505547428869326e-05,
"loss": 3.1747051239013673,
"step": 6875
},
{
"epoch": 1.3218059558117194,
"grad_norm": 0.511199414730072,
"learning_rate": 8.484078224841002e-05,
"loss": 3.176365280151367,
"step": 6880
},
{
"epoch": 1.3227665706051872,
"grad_norm": 0.4836551547050476,
"learning_rate": 8.462625461151769e-05,
"loss": 3.1734235763549803,
"step": 6885
},
{
"epoch": 1.3237271853986552,
"grad_norm": 0.6896669268608093,
"learning_rate": 8.441189191929152e-05,
"loss": 3.1753162384033202,
"step": 6890
},
{
"epoch": 1.324687800192123,
"grad_norm": 0.5728012323379517,
"learning_rate": 8.419769471259053e-05,
"loss": 3.1740385055541993,
"step": 6895
},
{
"epoch": 1.3256484149855907,
"grad_norm": 0.6955984234809875,
"learning_rate": 8.398366353185633e-05,
"loss": 3.175432586669922,
"step": 6900
},
{
"epoch": 1.3266090297790587,
"grad_norm": 0.48423975706100464,
"learning_rate": 8.376979891711145e-05,
"loss": 3.1750982284545897,
"step": 6905
},
{
"epoch": 1.3275696445725265,
"grad_norm": 0.4743121862411499,
"learning_rate": 8.355610140795827e-05,
"loss": 3.1710472106933594,
"step": 6910
},
{
"epoch": 1.3285302593659942,
"grad_norm": 0.6560218334197998,
"learning_rate": 8.334257154357755e-05,
"loss": 3.1815349578857424,
"step": 6915
},
{
"epoch": 1.329490874159462,
"grad_norm": 0.536353588104248,
"learning_rate": 8.31292098627271e-05,
"loss": 3.1730472564697267,
"step": 6920
},
{
"epoch": 1.3304514889529298,
"grad_norm": 0.49340781569480896,
"learning_rate": 8.291601690374021e-05,
"loss": 3.1722347259521486,
"step": 6925
},
{
"epoch": 1.3314121037463977,
"grad_norm": 0.5234084129333496,
"learning_rate": 8.270299320452467e-05,
"loss": 3.172749710083008,
"step": 6930
},
{
"epoch": 1.3323727185398655,
"grad_norm": 0.6601616740226746,
"learning_rate": 8.249013930256108e-05,
"loss": 3.1766475677490233,
"step": 6935
},
{
"epoch": 1.3333333333333333,
"grad_norm": 0.7026968002319336,
"learning_rate": 8.227745573490176e-05,
"loss": 3.173655319213867,
"step": 6940
},
{
"epoch": 1.3342939481268012,
"grad_norm": 0.5302594304084778,
"learning_rate": 8.206494303816904e-05,
"loss": 3.178236389160156,
"step": 6945
},
{
"epoch": 1.335254562920269,
"grad_norm": 0.46280986070632935,
"learning_rate": 8.185260174855433e-05,
"loss": 3.173746871948242,
"step": 6950
},
{
"epoch": 1.3362151777137368,
"grad_norm": 0.5142591595649719,
"learning_rate": 8.164043240181646e-05,
"loss": 3.173639106750488,
"step": 6955
},
{
"epoch": 1.3371757925072045,
"grad_norm": 0.682403028011322,
"learning_rate": 8.142843553328048e-05,
"loss": 3.175322151184082,
"step": 6960
},
{
"epoch": 1.3381364073006723,
"grad_norm": 0.6334633231163025,
"learning_rate": 8.121661167783625e-05,
"loss": 3.17366943359375,
"step": 6965
},
{
"epoch": 1.3390970220941403,
"grad_norm": 0.606118381023407,
"learning_rate": 8.100496136993712e-05,
"loss": 3.1736032485961916,
"step": 6970
},
{
"epoch": 1.340057636887608,
"grad_norm": 0.5890917778015137,
"learning_rate": 8.07934851435984e-05,
"loss": 3.172843360900879,
"step": 6975
},
{
"epoch": 1.341018251681076,
"grad_norm": 0.47088661789894104,
"learning_rate": 8.05821835323964e-05,
"loss": 3.1722528457641603,
"step": 6980
},
{
"epoch": 1.3419788664745438,
"grad_norm": 0.48279380798339844,
"learning_rate": 8.037105706946673e-05,
"loss": 3.177801513671875,
"step": 6985
},
{
"epoch": 1.3429394812680115,
"grad_norm": 0.5353664755821228,
"learning_rate": 8.01601062875032e-05,
"loss": 3.1740434646606444,
"step": 6990
},
{
"epoch": 1.3439000960614793,
"grad_norm": 0.5050175189971924,
"learning_rate": 7.994933171875613e-05,
"loss": 3.1727882385253907,
"step": 6995
},
{
"epoch": 1.344860710854947,
"grad_norm": 0.44954636693000793,
"learning_rate": 7.973873389503149e-05,
"loss": 3.1749853134155273,
"step": 7000
},
{
"epoch": 1.345821325648415,
"grad_norm": 0.6471356749534607,
"learning_rate": 7.952831334768913e-05,
"loss": 3.1730751037597655,
"step": 7005
},
{
"epoch": 1.3467819404418828,
"grad_norm": 0.5839700102806091,
"learning_rate": 7.931807060764176e-05,
"loss": 3.172525405883789,
"step": 7010
},
{
"epoch": 1.3477425552353506,
"grad_norm": 0.5643643140792847,
"learning_rate": 7.910800620535329e-05,
"loss": 3.172414016723633,
"step": 7015
},
{
"epoch": 1.3487031700288186,
"grad_norm": 0.5194916129112244,
"learning_rate": 7.889812067083777e-05,
"loss": 3.1767629623413085,
"step": 7020
},
{
"epoch": 1.3496637848222863,
"grad_norm": 0.4932091236114502,
"learning_rate": 7.868841453365797e-05,
"loss": 3.1722953796386717,
"step": 7025
},
{
"epoch": 1.350624399615754,
"grad_norm": 0.5307772159576416,
"learning_rate": 7.847888832292397e-05,
"loss": 3.1748939514160157,
"step": 7030
},
{
"epoch": 1.3515850144092219,
"grad_norm": 0.5100224614143372,
"learning_rate": 7.826954256729188e-05,
"loss": 3.1737871170043945,
"step": 7035
},
{
"epoch": 1.3525456292026896,
"grad_norm": 0.5833878517150879,
"learning_rate": 7.806037779496264e-05,
"loss": 3.171864128112793,
"step": 7040
},
{
"epoch": 1.3535062439961576,
"grad_norm": 0.4274962842464447,
"learning_rate": 7.785139453368025e-05,
"loss": 3.174275207519531,
"step": 7045
},
{
"epoch": 1.3544668587896254,
"grad_norm": 0.43694546818733215,
"learning_rate": 7.764259331073096e-05,
"loss": 3.1718505859375,
"step": 7050
},
{
"epoch": 1.3554274735830931,
"grad_norm": 0.5787099003791809,
"learning_rate": 7.743397465294174e-05,
"loss": 3.171575164794922,
"step": 7055
},
{
"epoch": 1.356388088376561,
"grad_norm": 0.45274344086647034,
"learning_rate": 7.72255390866789e-05,
"loss": 3.169318199157715,
"step": 7060
},
{
"epoch": 1.3573487031700289,
"grad_norm": 0.490500807762146,
"learning_rate": 7.701728713784662e-05,
"loss": 3.172596740722656,
"step": 7065
},
{
"epoch": 1.3583093179634966,
"grad_norm": 0.8986951112747192,
"learning_rate": 7.6809219331886e-05,
"loss": 3.177069664001465,
"step": 7070
},
{
"epoch": 1.3592699327569644,
"grad_norm": 0.6794254183769226,
"learning_rate": 7.660133619377346e-05,
"loss": 3.1766523361206054,
"step": 7075
},
{
"epoch": 1.3602305475504322,
"grad_norm": 0.4294492304325104,
"learning_rate": 7.639363824801957e-05,
"loss": 3.171103668212891,
"step": 7080
},
{
"epoch": 1.3611911623439001,
"grad_norm": 0.4653557240962982,
"learning_rate": 7.618612601866738e-05,
"loss": 3.1739612579345704,
"step": 7085
},
{
"epoch": 1.362151777137368,
"grad_norm": 0.48339158296585083,
"learning_rate": 7.597880002929155e-05,
"loss": 3.174298095703125,
"step": 7090
},
{
"epoch": 1.3631123919308357,
"grad_norm": 0.4108154773712158,
"learning_rate": 7.577166080299704e-05,
"loss": 3.1711090087890623,
"step": 7095
},
{
"epoch": 1.3640730067243036,
"grad_norm": 0.5386460423469543,
"learning_rate": 7.556470886241715e-05,
"loss": 3.172881317138672,
"step": 7100
},
{
"epoch": 1.3650336215177714,
"grad_norm": 0.521826982498169,
"learning_rate": 7.535794472971292e-05,
"loss": 3.1787548065185547,
"step": 7105
},
{
"epoch": 1.3659942363112392,
"grad_norm": 0.5726050138473511,
"learning_rate": 7.515136892657152e-05,
"loss": 3.1746589660644533,
"step": 7110
},
{
"epoch": 1.366954851104707,
"grad_norm": 0.49367332458496094,
"learning_rate": 7.494498197420486e-05,
"loss": 3.171868896484375,
"step": 7115
},
{
"epoch": 1.3679154658981747,
"grad_norm": 0.6521613597869873,
"learning_rate": 7.473878439334833e-05,
"loss": 3.1738697052001954,
"step": 7120
},
{
"epoch": 1.3688760806916427,
"grad_norm": 0.4904996454715729,
"learning_rate": 7.453277670425964e-05,
"loss": 3.1751945495605467,
"step": 7125
},
{
"epoch": 1.3698366954851104,
"grad_norm": 0.5470160841941833,
"learning_rate": 7.432695942671736e-05,
"loss": 3.1728076934814453,
"step": 7130
},
{
"epoch": 1.3707973102785782,
"grad_norm": 0.6032967567443848,
"learning_rate": 7.412133308001952e-05,
"loss": 3.166953468322754,
"step": 7135
},
{
"epoch": 1.3717579250720462,
"grad_norm": 0.4871714413166046,
"learning_rate": 7.39158981829825e-05,
"loss": 3.1728256225585936,
"step": 7140
},
{
"epoch": 1.372718539865514,
"grad_norm": 0.42759451270103455,
"learning_rate": 7.371065525393965e-05,
"loss": 3.1744380950927735,
"step": 7145
},
{
"epoch": 1.3736791546589817,
"grad_norm": 0.5663604736328125,
"learning_rate": 7.350560481073996e-05,
"loss": 3.1777429580688477,
"step": 7150
},
{
"epoch": 1.3746397694524495,
"grad_norm": 0.5459421873092651,
"learning_rate": 7.330074737074665e-05,
"loss": 3.170821762084961,
"step": 7155
},
{
"epoch": 1.3756003842459175,
"grad_norm": 0.46091675758361816,
"learning_rate": 7.309608345083605e-05,
"loss": 3.1742374420166017,
"step": 7160
},
{
"epoch": 1.3765609990393852,
"grad_norm": 0.38046908378601074,
"learning_rate": 7.289161356739638e-05,
"loss": 3.1713001251220705,
"step": 7165
},
{
"epoch": 1.377521613832853,
"grad_norm": 0.468205064535141,
"learning_rate": 7.268733823632601e-05,
"loss": 3.1741426467895506,
"step": 7170
},
{
"epoch": 1.378482228626321,
"grad_norm": 0.6682205200195312,
"learning_rate": 7.248325797303256e-05,
"loss": 3.1712989807128906,
"step": 7175
},
{
"epoch": 1.3794428434197887,
"grad_norm": 0.5630512833595276,
"learning_rate": 7.227937329243149e-05,
"loss": 3.171531867980957,
"step": 7180
},
{
"epoch": 1.3804034582132565,
"grad_norm": 0.54999178647995,
"learning_rate": 7.20756847089448e-05,
"loss": 3.173867416381836,
"step": 7185
},
{
"epoch": 1.3813640730067243,
"grad_norm": 0.43921148777008057,
"learning_rate": 7.187219273649962e-05,
"loss": 3.170378494262695,
"step": 7190
},
{
"epoch": 1.382324687800192,
"grad_norm": 0.4150092601776123,
"learning_rate": 7.166889788852707e-05,
"loss": 3.170474624633789,
"step": 7195
},
{
"epoch": 1.38328530259366,
"grad_norm": 0.38194531202316284,
"learning_rate": 7.146580067796102e-05,
"loss": 3.1690601348876952,
"step": 7200
},
{
"epoch": 1.3842459173871278,
"grad_norm": 0.7853124141693115,
"learning_rate": 7.126290161723642e-05,
"loss": 3.1698192596435546,
"step": 7205
},
{
"epoch": 1.3852065321805955,
"grad_norm": 0.4423576593399048,
"learning_rate": 7.106020121828848e-05,
"loss": 3.1732460021972657,
"step": 7210
},
{
"epoch": 1.3861671469740635,
"grad_norm": 0.4560396671295166,
"learning_rate": 7.085769999255108e-05,
"loss": 3.1727630615234377,
"step": 7215
},
{
"epoch": 1.3871277617675313,
"grad_norm": 0.41349244117736816,
"learning_rate": 7.065539845095567e-05,
"loss": 3.1731325149536134,
"step": 7220
},
{
"epoch": 1.388088376560999,
"grad_norm": 0.47393569350242615,
"learning_rate": 7.045329710392967e-05,
"loss": 3.1727859497070314,
"step": 7225
},
{
"epoch": 1.3890489913544668,
"grad_norm": 0.7396698594093323,
"learning_rate": 7.025139646139553e-05,
"loss": 3.1715877532958983,
"step": 7230
},
{
"epoch": 1.3900096061479346,
"grad_norm": 0.7079951763153076,
"learning_rate": 7.004969703276941e-05,
"loss": 3.169532585144043,
"step": 7235
},
{
"epoch": 1.3909702209414025,
"grad_norm": 0.46009525656700134,
"learning_rate": 6.984819932695956e-05,
"loss": 3.169749450683594,
"step": 7240
},
{
"epoch": 1.3919308357348703,
"grad_norm": 0.4487673044204712,
"learning_rate": 6.964690385236534e-05,
"loss": 3.1695636749267577,
"step": 7245
},
{
"epoch": 1.392891450528338,
"grad_norm": 0.4718167781829834,
"learning_rate": 6.944581111687593e-05,
"loss": 3.168662452697754,
"step": 7250
},
{
"epoch": 1.393852065321806,
"grad_norm": 0.46017783880233765,
"learning_rate": 6.924492162786898e-05,
"loss": 3.169230842590332,
"step": 7255
},
{
"epoch": 1.3948126801152738,
"grad_norm": 0.45555198192596436,
"learning_rate": 6.904423589220917e-05,
"loss": 3.169890594482422,
"step": 7260
},
{
"epoch": 1.3957732949087416,
"grad_norm": 0.4948507845401764,
"learning_rate": 6.884375441624724e-05,
"loss": 3.173093795776367,
"step": 7265
},
{
"epoch": 1.3967339097022093,
"grad_norm": 0.43780985474586487,
"learning_rate": 6.864347770581859e-05,
"loss": 3.168751335144043,
"step": 7270
},
{
"epoch": 1.397694524495677,
"grad_norm": 0.4438318908214569,
"learning_rate": 6.844340626624178e-05,
"loss": 3.1701900482177736,
"step": 7275
},
{
"epoch": 1.398655139289145,
"grad_norm": 0.5062385201454163,
"learning_rate": 6.824354060231765e-05,
"loss": 3.1665103912353514,
"step": 7280
},
{
"epoch": 1.3996157540826129,
"grad_norm": 0.4086364507675171,
"learning_rate": 6.804388121832777e-05,
"loss": 3.1716842651367188,
"step": 7285
},
{
"epoch": 1.4005763688760806,
"grad_norm": 0.5124282836914062,
"learning_rate": 6.784442861803331e-05,
"loss": 3.1696632385253904,
"step": 7290
},
{
"epoch": 1.4015369836695486,
"grad_norm": 0.5204159617424011,
"learning_rate": 6.764518330467346e-05,
"loss": 3.1707353591918945,
"step": 7295
},
{
"epoch": 1.4024975984630164,
"grad_norm": 0.6575141549110413,
"learning_rate": 6.744614578096475e-05,
"loss": 3.171962547302246,
"step": 7300
},
{
"epoch": 1.4034582132564841,
"grad_norm": 0.5063356757164001,
"learning_rate": 6.72473165490993e-05,
"loss": 3.1709272384643556,
"step": 7305
},
{
"epoch": 1.4044188280499519,
"grad_norm": 0.4604819416999817,
"learning_rate": 6.704869611074351e-05,
"loss": 3.173125076293945,
"step": 7310
},
{
"epoch": 1.4053794428434199,
"grad_norm": 0.4766218960285187,
"learning_rate": 6.685028496703719e-05,
"loss": 3.1713457107543945,
"step": 7315
},
{
"epoch": 1.4063400576368876,
"grad_norm": 0.48007673025131226,
"learning_rate": 6.665208361859203e-05,
"loss": 3.1723331451416015,
"step": 7320
},
{
"epoch": 1.4073006724303554,
"grad_norm": 0.3352242410182953,
"learning_rate": 6.64540925654904e-05,
"loss": 3.171708679199219,
"step": 7325
},
{
"epoch": 1.4082612872238234,
"grad_norm": 0.6097206473350525,
"learning_rate": 6.625631230728393e-05,
"loss": 3.169724464416504,
"step": 7330
},
{
"epoch": 1.4092219020172911,
"grad_norm": 0.45327913761138916,
"learning_rate": 6.605874334299257e-05,
"loss": 3.172129821777344,
"step": 7335
},
{
"epoch": 1.410182516810759,
"grad_norm": 0.3918154537677765,
"learning_rate": 6.586138617110312e-05,
"loss": 3.1696929931640625,
"step": 7340
},
{
"epoch": 1.4111431316042267,
"grad_norm": 0.45734044909477234,
"learning_rate": 6.566424128956788e-05,
"loss": 3.169087600708008,
"step": 7345
},
{
"epoch": 1.4121037463976944,
"grad_norm": 0.41583287715911865,
"learning_rate": 6.546730919580365e-05,
"loss": 3.169902801513672,
"step": 7350
},
{
"epoch": 1.4130643611911624,
"grad_norm": 0.3783586919307709,
"learning_rate": 6.527059038669032e-05,
"loss": 3.173455810546875,
"step": 7355
},
{
"epoch": 1.4140249759846302,
"grad_norm": 0.5969278216362,
"learning_rate": 6.507408535856968e-05,
"loss": 3.171731185913086,
"step": 7360
},
{
"epoch": 1.414985590778098,
"grad_norm": 0.5529734492301941,
"learning_rate": 6.487779460724388e-05,
"loss": 3.169993591308594,
"step": 7365
},
{
"epoch": 1.415946205571566,
"grad_norm": 0.460746705532074,
"learning_rate": 6.468171862797484e-05,
"loss": 3.1684539794921873,
"step": 7370
},
{
"epoch": 1.4169068203650337,
"grad_norm": 0.40012434124946594,
"learning_rate": 6.448585791548231e-05,
"loss": 3.170420837402344,
"step": 7375
},
{
"epoch": 1.4178674351585014,
"grad_norm": 0.4528000056743622,
"learning_rate": 6.42902129639429e-05,
"loss": 3.1724185943603516,
"step": 7380
},
{
"epoch": 1.4188280499519692,
"grad_norm": 0.5503178834915161,
"learning_rate": 6.409478426698893e-05,
"loss": 3.170566749572754,
"step": 7385
},
{
"epoch": 1.419788664745437,
"grad_norm": 0.4782489240169525,
"learning_rate": 6.389957231770705e-05,
"loss": 3.1753549575805664,
"step": 7390
},
{
"epoch": 1.420749279538905,
"grad_norm": 0.5017015933990479,
"learning_rate": 6.370457760863708e-05,
"loss": 3.1712177276611326,
"step": 7395
},
{
"epoch": 1.4217098943323727,
"grad_norm": 0.4609530568122864,
"learning_rate": 6.35098006317706e-05,
"loss": 3.1715621948242188,
"step": 7400
},
{
"epoch": 1.4226705091258405,
"grad_norm": 0.43000540137290955,
"learning_rate": 6.331524187854992e-05,
"loss": 3.1696197509765627,
"step": 7405
},
{
"epoch": 1.4236311239193085,
"grad_norm": 0.6466848850250244,
"learning_rate": 6.31209018398668e-05,
"loss": 3.171028900146484,
"step": 7410
},
{
"epoch": 1.4245917387127762,
"grad_norm": 0.4881599247455597,
"learning_rate": 6.292678100606101e-05,
"loss": 3.1710824966430664,
"step": 7415
},
{
"epoch": 1.425552353506244,
"grad_norm": 0.42242470383644104,
"learning_rate": 6.273287986691934e-05,
"loss": 3.1717771530151366,
"step": 7420
},
{
"epoch": 1.4265129682997117,
"grad_norm": 0.40897586941719055,
"learning_rate": 6.253919891167427e-05,
"loss": 3.168376159667969,
"step": 7425
},
{
"epoch": 1.4274735830931795,
"grad_norm": 0.31631234288215637,
"learning_rate": 6.23457386290028e-05,
"loss": 3.167464828491211,
"step": 7430
},
{
"epoch": 1.4284341978866475,
"grad_norm": 0.423149973154068,
"learning_rate": 6.215249950702488e-05,
"loss": 3.1704282760620117,
"step": 7435
},
{
"epoch": 1.4293948126801153,
"grad_norm": 0.663411557674408,
"learning_rate": 6.195948203330282e-05,
"loss": 3.170808219909668,
"step": 7440
},
{
"epoch": 1.430355427473583,
"grad_norm": 0.35881131887435913,
"learning_rate": 6.176668669483948e-05,
"loss": 3.1693355560302736,
"step": 7445
},
{
"epoch": 1.431316042267051,
"grad_norm": 0.3775772452354431,
"learning_rate": 6.15741139780772e-05,
"loss": 3.1732666015625,
"step": 7450
},
{
"epoch": 1.4322766570605188,
"grad_norm": 0.40818941593170166,
"learning_rate": 6.138176436889672e-05,
"loss": 3.1699262619018556,
"step": 7455
},
{
"epoch": 1.4332372718539865,
"grad_norm": 0.31840816140174866,
"learning_rate": 6.118963835261588e-05,
"loss": 3.171256256103516,
"step": 7460
},
{
"epoch": 1.4341978866474543,
"grad_norm": 0.45461708307266235,
"learning_rate": 6.099773641398834e-05,
"loss": 3.1705402374267577,
"step": 7465
},
{
"epoch": 1.435158501440922,
"grad_norm": 0.49488258361816406,
"learning_rate": 6.080605903720229e-05,
"loss": 3.1715473175048827,
"step": 7470
},
{
"epoch": 1.43611911623439,
"grad_norm": 0.4354032278060913,
"learning_rate": 6.061460670587948e-05,
"loss": 3.170640563964844,
"step": 7475
},
{
"epoch": 1.4370797310278578,
"grad_norm": 0.39679399132728577,
"learning_rate": 6.0423379903073824e-05,
"loss": 3.170828437805176,
"step": 7480
},
{
"epoch": 1.4380403458213258,
"grad_norm": 0.37695419788360596,
"learning_rate": 6.023237911127005e-05,
"loss": 3.167501449584961,
"step": 7485
},
{
"epoch": 1.4390009606147935,
"grad_norm": 0.37574145197868347,
"learning_rate": 6.004160481238281e-05,
"loss": 3.172044372558594,
"step": 7490
},
{
"epoch": 1.4399615754082613,
"grad_norm": 0.5872917771339417,
"learning_rate": 5.9851057487755225e-05,
"loss": 3.167539596557617,
"step": 7495
},
{
"epoch": 1.440922190201729,
"grad_norm": 0.43774786591529846,
"learning_rate": 5.966073761815776e-05,
"loss": 3.170111656188965,
"step": 7500
},
{
"epoch": 1.4418828049951968,
"grad_norm": 0.319992333650589,
"learning_rate": 5.947064568378696e-05,
"loss": 3.1693355560302736,
"step": 7505
},
{
"epoch": 1.4428434197886648,
"grad_norm": 0.4077332019805908,
"learning_rate": 5.928078216426427e-05,
"loss": 3.1672155380249025,
"step": 7510
},
{
"epoch": 1.4438040345821326,
"grad_norm": 0.355669766664505,
"learning_rate": 5.909114753863488e-05,
"loss": 3.1689855575561525,
"step": 7515
},
{
"epoch": 1.4447646493756003,
"grad_norm": 0.5554761290550232,
"learning_rate": 5.8901742285366315e-05,
"loss": 3.171320343017578,
"step": 7520
},
{
"epoch": 1.4457252641690683,
"grad_norm": 0.4790358543395996,
"learning_rate": 5.8712566882347504e-05,
"loss": 3.1696239471435548,
"step": 7525
},
{
"epoch": 1.446685878962536,
"grad_norm": 0.4286898076534271,
"learning_rate": 5.8523621806887374e-05,
"loss": 3.169162368774414,
"step": 7530
},
{
"epoch": 1.4476464937560038,
"grad_norm": 0.42978528141975403,
"learning_rate": 5.8334907535713836e-05,
"loss": 3.172647476196289,
"step": 7535
},
{
"epoch": 1.4486071085494716,
"grad_norm": 0.4055570363998413,
"learning_rate": 5.8146424544972205e-05,
"loss": 3.1680522918701173,
"step": 7540
},
{
"epoch": 1.4495677233429394,
"grad_norm": 0.3212641775608063,
"learning_rate": 5.7958173310224465e-05,
"loss": 3.168552017211914,
"step": 7545
},
{
"epoch": 1.4505283381364074,
"grad_norm": 0.40570881962776184,
"learning_rate": 5.7770154306447866e-05,
"loss": 3.172994041442871,
"step": 7550
},
{
"epoch": 1.4514889529298751,
"grad_norm": 0.4512649476528168,
"learning_rate": 5.758236800803351e-05,
"loss": 3.171183395385742,
"step": 7555
},
{
"epoch": 1.4524495677233429,
"grad_norm": 0.5307871103286743,
"learning_rate": 5.739481488878558e-05,
"loss": 3.172830581665039,
"step": 7560
},
{
"epoch": 1.4534101825168109,
"grad_norm": 0.38034340739250183,
"learning_rate": 5.720749542191979e-05,
"loss": 3.1710861206054686,
"step": 7565
},
{
"epoch": 1.4543707973102786,
"grad_norm": 0.46515336632728577,
"learning_rate": 5.7020410080062416e-05,
"loss": 3.1718137741088865,
"step": 7570
},
{
"epoch": 1.4553314121037464,
"grad_norm": 0.4164998531341553,
"learning_rate": 5.683355933524894e-05,
"loss": 3.1689435958862306,
"step": 7575
},
{
"epoch": 1.4562920268972142,
"grad_norm": 0.48741379380226135,
"learning_rate": 5.6646943658922975e-05,
"loss": 3.167072296142578,
"step": 7580
},
{
"epoch": 1.457252641690682,
"grad_norm": 0.39116013050079346,
"learning_rate": 5.646056352193505e-05,
"loss": 3.171469497680664,
"step": 7585
},
{
"epoch": 1.45821325648415,
"grad_norm": 0.4745250642299652,
"learning_rate": 5.6274419394541296e-05,
"loss": 3.168760871887207,
"step": 7590
},
{
"epoch": 1.4591738712776177,
"grad_norm": 0.5160530805587769,
"learning_rate": 5.608851174640247e-05,
"loss": 3.168941307067871,
"step": 7595
},
{
"epoch": 1.4601344860710854,
"grad_norm": 0.3649737238883972,
"learning_rate": 5.590284104658264e-05,
"loss": 3.1685501098632813,
"step": 7600
},
{
"epoch": 1.4610951008645534,
"grad_norm": 0.37525928020477295,
"learning_rate": 5.571740776354811e-05,
"loss": 3.169935607910156,
"step": 7605
},
{
"epoch": 1.4620557156580212,
"grad_norm": 0.35180380940437317,
"learning_rate": 5.553221236516594e-05,
"loss": 3.1680299758911135,
"step": 7610
},
{
"epoch": 1.463016330451489,
"grad_norm": 0.5641520023345947,
"learning_rate": 5.534725531870317e-05,
"loss": 3.171617126464844,
"step": 7615
},
{
"epoch": 1.4639769452449567,
"grad_norm": 0.5281386375427246,
"learning_rate": 5.516253709082547e-05,
"loss": 3.1680675506591798,
"step": 7620
},
{
"epoch": 1.4649375600384245,
"grad_norm": 0.37240076065063477,
"learning_rate": 5.4978058147595796e-05,
"loss": 3.164999008178711,
"step": 7625
},
{
"epoch": 1.4658981748318924,
"grad_norm": 0.37205061316490173,
"learning_rate": 5.479381895447346e-05,
"loss": 3.1697301864624023,
"step": 7630
},
{
"epoch": 1.4668587896253602,
"grad_norm": 0.35913726687431335,
"learning_rate": 5.4609819976312854e-05,
"loss": 3.1634567260742186,
"step": 7635
},
{
"epoch": 1.4678194044188282,
"grad_norm": 0.42885205149650574,
"learning_rate": 5.4426061677362284e-05,
"loss": 3.173727798461914,
"step": 7640
},
{
"epoch": 1.468780019212296,
"grad_norm": 0.3557131588459015,
"learning_rate": 5.424254452126279e-05,
"loss": 3.1679935455322266,
"step": 7645
},
{
"epoch": 1.4697406340057637,
"grad_norm": 0.4141709804534912,
"learning_rate": 5.4059268971047e-05,
"loss": 3.1674577713012697,
"step": 7650
},
{
"epoch": 1.4707012487992315,
"grad_norm": 0.40807923674583435,
"learning_rate": 5.387623548913795e-05,
"loss": 3.171774673461914,
"step": 7655
},
{
"epoch": 1.4716618635926992,
"grad_norm": 0.3850151598453522,
"learning_rate": 5.36934445373478e-05,
"loss": 3.164442443847656,
"step": 7660
},
{
"epoch": 1.4726224783861672,
"grad_norm": 0.43482470512390137,
"learning_rate": 5.3510896576876924e-05,
"loss": 3.1658775329589846,
"step": 7665
},
{
"epoch": 1.473583093179635,
"grad_norm": 0.40651193261146545,
"learning_rate": 5.3328592068312565e-05,
"loss": 3.164666748046875,
"step": 7670
},
{
"epoch": 1.4745437079731027,
"grad_norm": 0.33272117376327515,
"learning_rate": 5.3146531471627737e-05,
"loss": 3.169040298461914,
"step": 7675
},
{
"epoch": 1.4755043227665707,
"grad_norm": 0.3396502435207367,
"learning_rate": 5.29647152461799e-05,
"loss": 3.1710128784179688,
"step": 7680
},
{
"epoch": 1.4764649375600385,
"grad_norm": 0.27867111563682556,
"learning_rate": 5.278314385071011e-05,
"loss": 3.169612693786621,
"step": 7685
},
{
"epoch": 1.4774255523535063,
"grad_norm": 0.5024138689041138,
"learning_rate": 5.260181774334165e-05,
"loss": 3.1680776596069338,
"step": 7690
},
{
"epoch": 1.478386167146974,
"grad_norm": 0.5927129983901978,
"learning_rate": 5.2420737381578814e-05,
"loss": 3.1725555419921876,
"step": 7695
},
{
"epoch": 1.4793467819404418,
"grad_norm": 0.2990652322769165,
"learning_rate": 5.223990322230596e-05,
"loss": 3.1685482025146485,
"step": 7700
},
{
"epoch": 1.4803073967339098,
"grad_norm": 0.41576090455055237,
"learning_rate": 5.205931572178625e-05,
"loss": 3.1704280853271483,
"step": 7705
},
{
"epoch": 1.4812680115273775,
"grad_norm": 0.33696505427360535,
"learning_rate": 5.187897533566047e-05,
"loss": 3.1678043365478517,
"step": 7710
},
{
"epoch": 1.4822286263208453,
"grad_norm": 0.33144330978393555,
"learning_rate": 5.169888251894587e-05,
"loss": 3.1680809020996095,
"step": 7715
},
{
"epoch": 1.4831892411143133,
"grad_norm": 0.35232898592948914,
"learning_rate": 5.151903772603517e-05,
"loss": 3.1694143295288084,
"step": 7720
},
{
"epoch": 1.484149855907781,
"grad_norm": 0.482440710067749,
"learning_rate": 5.1339441410695225e-05,
"loss": 3.167070007324219,
"step": 7725
},
{
"epoch": 1.4851104707012488,
"grad_norm": 0.44148990511894226,
"learning_rate": 5.116009402606591e-05,
"loss": 3.1692365646362304,
"step": 7730
},
{
"epoch": 1.4860710854947166,
"grad_norm": 0.4089405834674835,
"learning_rate": 5.0980996024659075e-05,
"loss": 3.165630912780762,
"step": 7735
},
{
"epoch": 1.4870317002881843,
"grad_norm": 0.3344769775867462,
"learning_rate": 5.0802147858357386e-05,
"loss": 3.1695476531982423,
"step": 7740
},
{
"epoch": 1.4879923150816523,
"grad_norm": 0.389249324798584,
"learning_rate": 5.0623549978413166e-05,
"loss": 3.171805000305176,
"step": 7745
},
{
"epoch": 1.48895292987512,
"grad_norm": 0.3440057635307312,
"learning_rate": 5.0445202835447056e-05,
"loss": 3.1670265197753906,
"step": 7750
},
{
"epoch": 1.4899135446685878,
"grad_norm": 0.3163948059082031,
"learning_rate": 5.026710687944728e-05,
"loss": 3.1670303344726562,
"step": 7755
},
{
"epoch": 1.4908741594620558,
"grad_norm": 0.3952416777610779,
"learning_rate": 5.0089262559768246e-05,
"loss": 3.171236038208008,
"step": 7760
},
{
"epoch": 1.4918347742555236,
"grad_norm": 0.3361209034919739,
"learning_rate": 4.9911670325129304e-05,
"loss": 3.1686822891235353,
"step": 7765
},
{
"epoch": 1.4927953890489913,
"grad_norm": 0.35905590653419495,
"learning_rate": 4.9734330623613924e-05,
"loss": 3.1694427490234376,
"step": 7770
},
{
"epoch": 1.493756003842459,
"grad_norm": 0.4923272132873535,
"learning_rate": 4.955724390266841e-05,
"loss": 3.171158218383789,
"step": 7775
},
{
"epoch": 1.4947166186359269,
"grad_norm": 0.3285467326641083,
"learning_rate": 4.9380410609100674e-05,
"loss": 3.1719661712646485,
"step": 7780
},
{
"epoch": 1.4956772334293948,
"grad_norm": 0.35682541131973267,
"learning_rate": 4.920383118907929e-05,
"loss": 3.1671146392822265,
"step": 7785
},
{
"epoch": 1.4966378482228626,
"grad_norm": 0.5638254880905151,
"learning_rate": 4.902750608813222e-05,
"loss": 3.1699798583984373,
"step": 7790
},
{
"epoch": 1.4975984630163304,
"grad_norm": 0.2942904531955719,
"learning_rate": 4.885143575114587e-05,
"loss": 3.1689422607421873,
"step": 7795
},
{
"epoch": 1.4985590778097984,
"grad_norm": 0.3493422269821167,
"learning_rate": 4.8675620622363645e-05,
"loss": 3.168619918823242,
"step": 7800
},
{
"epoch": 1.4995196926032661,
"grad_norm": 0.40712451934814453,
"learning_rate": 4.850006114538519e-05,
"loss": 3.169388771057129,
"step": 7805
},
{
"epoch": 1.5004803073967339,
"grad_norm": 0.31108608841896057,
"learning_rate": 4.8324757763165075e-05,
"loss": 3.1686771392822264,
"step": 7810
},
{
"epoch": 1.5014409221902016,
"grad_norm": 0.5020172595977783,
"learning_rate": 4.814971091801179e-05,
"loss": 3.1644512176513673,
"step": 7815
},
{
"epoch": 1.5024015369836694,
"grad_norm": 0.4801461696624756,
"learning_rate": 4.7974921051586385e-05,
"loss": 3.1699331283569334,
"step": 7820
},
{
"epoch": 1.5033621517771374,
"grad_norm": 0.31796136498451233,
"learning_rate": 4.780038860490164e-05,
"loss": 3.1699861526489257,
"step": 7825
},
{
"epoch": 1.5043227665706052,
"grad_norm": 0.3169126510620117,
"learning_rate": 4.762611401832089e-05,
"loss": 3.1705883026123045,
"step": 7830
},
{
"epoch": 1.5052833813640731,
"grad_norm": 0.3533933758735657,
"learning_rate": 4.745209773155671e-05,
"loss": 3.165675926208496,
"step": 7835
},
{
"epoch": 1.506243996157541,
"grad_norm": 0.4423210024833679,
"learning_rate": 4.727834018367007e-05,
"loss": 3.1672183990478517,
"step": 7840
},
{
"epoch": 1.5072046109510087,
"grad_norm": 0.33951136469841003,
"learning_rate": 4.710484181306912e-05,
"loss": 3.169887733459473,
"step": 7845
},
{
"epoch": 1.5081652257444764,
"grad_norm": 0.298172265291214,
"learning_rate": 4.693160305750801e-05,
"loss": 3.1656867980957033,
"step": 7850
},
{
"epoch": 1.5091258405379442,
"grad_norm": 0.3188192546367645,
"learning_rate": 4.675862435408591e-05,
"loss": 3.1662307739257813,
"step": 7855
},
{
"epoch": 1.510086455331412,
"grad_norm": 0.3231668174266815,
"learning_rate": 4.6585906139245834e-05,
"loss": 3.1684810638427736,
"step": 7860
},
{
"epoch": 1.51104707012488,
"grad_norm": 0.46347707509994507,
"learning_rate": 4.641344884877362e-05,
"loss": 3.1662145614624024,
"step": 7865
},
{
"epoch": 1.5120076849183477,
"grad_norm": 0.4303688704967499,
"learning_rate": 4.6241252917796576e-05,
"loss": 3.166950798034668,
"step": 7870
},
{
"epoch": 1.5129682997118157,
"grad_norm": 0.3407163918018341,
"learning_rate": 4.6069318780782765e-05,
"loss": 3.1677284240722656,
"step": 7875
},
{
"epoch": 1.5139289145052834,
"grad_norm": 0.2888704240322113,
"learning_rate": 4.589764687153967e-05,
"loss": 3.162090301513672,
"step": 7880
},
{
"epoch": 1.5148895292987512,
"grad_norm": 0.3005317747592926,
"learning_rate": 4.5726237623213155e-05,
"loss": 3.1687442779541017,
"step": 7885
},
{
"epoch": 1.515850144092219,
"grad_norm": 0.2978266477584839,
"learning_rate": 4.555509146828624e-05,
"loss": 3.1686517715454103,
"step": 7890
},
{
"epoch": 1.5168107588856867,
"grad_norm": 0.2926294505596161,
"learning_rate": 4.53842088385783e-05,
"loss": 3.170303726196289,
"step": 7895
},
{
"epoch": 1.5177713736791547,
"grad_norm": 0.47925063967704773,
"learning_rate": 4.521359016524376e-05,
"loss": 3.169039726257324,
"step": 7900
},
{
"epoch": 1.5187319884726225,
"grad_norm": 0.4029441475868225,
"learning_rate": 4.5043235878770965e-05,
"loss": 3.166218376159668,
"step": 7905
},
{
"epoch": 1.5196926032660905,
"grad_norm": 0.31125885248184204,
"learning_rate": 4.4873146408981295e-05,
"loss": 3.1654186248779297,
"step": 7910
},
{
"epoch": 1.5206532180595582,
"grad_norm": 0.3251868784427643,
"learning_rate": 4.4703322185027926e-05,
"loss": 3.1693817138671876,
"step": 7915
},
{
"epoch": 1.521613832853026,
"grad_norm": 0.31480100750923157,
"learning_rate": 4.453376363539481e-05,
"loss": 3.1616186141967773,
"step": 7920
},
{
"epoch": 1.5225744476464937,
"grad_norm": 0.3208870589733124,
"learning_rate": 4.436447118789555e-05,
"loss": 3.168798637390137,
"step": 7925
},
{
"epoch": 1.5235350624399615,
"grad_norm": 0.3242574632167816,
"learning_rate": 4.419544526967238e-05,
"loss": 3.171648406982422,
"step": 7930
},
{
"epoch": 1.5244956772334293,
"grad_norm": 0.3922894597053528,
"learning_rate": 4.402668630719504e-05,
"loss": 3.1695817947387694,
"step": 7935
},
{
"epoch": 1.5254562920268973,
"grad_norm": 0.3969297409057617,
"learning_rate": 4.385819472625963e-05,
"loss": 3.1659820556640623,
"step": 7940
},
{
"epoch": 1.526416906820365,
"grad_norm": 0.3534747362136841,
"learning_rate": 4.368997095198775e-05,
"loss": 3.164035415649414,
"step": 7945
},
{
"epoch": 1.527377521613833,
"grad_norm": 0.387599915266037,
"learning_rate": 4.352201540882523e-05,
"loss": 3.166557502746582,
"step": 7950
},
{
"epoch": 1.5283381364073008,
"grad_norm": 0.3567873537540436,
"learning_rate": 4.33543285205412e-05,
"loss": 3.1643226623535154,
"step": 7955
},
{
"epoch": 1.5292987512007685,
"grad_norm": 0.3381181061267853,
"learning_rate": 4.318691071022676e-05,
"loss": 3.1638862609863283,
"step": 7960
},
{
"epoch": 1.5302593659942363,
"grad_norm": 0.2984130084514618,
"learning_rate": 4.301976240029428e-05,
"loss": 3.1657569885253904,
"step": 7965
},
{
"epoch": 1.531219980787704,
"grad_norm": 0.2934906780719757,
"learning_rate": 4.285288401247614e-05,
"loss": 3.165040969848633,
"step": 7970
},
{
"epoch": 1.5321805955811718,
"grad_norm": 0.2868797779083252,
"learning_rate": 4.268627596782354e-05,
"loss": 3.166617202758789,
"step": 7975
},
{
"epoch": 1.5331412103746398,
"grad_norm": 0.40537068247795105,
"learning_rate": 4.251993868670569e-05,
"loss": 3.1657033920288087,
"step": 7980
},
{
"epoch": 1.5341018251681076,
"grad_norm": 0.3606870770454407,
"learning_rate": 4.235387258880871e-05,
"loss": 3.1681283950805663,
"step": 7985
},
{
"epoch": 1.5350624399615755,
"grad_norm": 0.3049268126487732,
"learning_rate": 4.218807809313428e-05,
"loss": 3.165813446044922,
"step": 7990
},
{
"epoch": 1.5360230547550433,
"grad_norm": 0.3047685921192169,
"learning_rate": 4.202255561799897e-05,
"loss": 3.1700771331787108,
"step": 7995
},
{
"epoch": 1.536983669548511,
"grad_norm": 0.32123667001724243,
"learning_rate": 4.1857305581032974e-05,
"loss": 3.16713924407959,
"step": 8000
},
{
"epoch": 1.5379442843419788,
"grad_norm": 0.31456029415130615,
"learning_rate": 4.1692328399179134e-05,
"loss": 3.1665233612060546,
"step": 8005
},
{
"epoch": 1.5389048991354466,
"grad_norm": 0.2823289632797241,
"learning_rate": 4.1527624488691706e-05,
"loss": 3.165792465209961,
"step": 8010
},
{
"epoch": 1.5398655139289144,
"grad_norm": 0.41505372524261475,
"learning_rate": 4.1363194265135584e-05,
"loss": 3.165696907043457,
"step": 8015
},
{
"epoch": 1.5408261287223823,
"grad_norm": 0.3873119056224823,
"learning_rate": 4.1199038143385114e-05,
"loss": 3.1616024017333983,
"step": 8020
},
{
"epoch": 1.54178674351585,
"grad_norm": 0.29045772552490234,
"learning_rate": 4.1035156537623056e-05,
"loss": 3.163679504394531,
"step": 8025
},
{
"epoch": 1.542747358309318,
"grad_norm": 0.31360870599746704,
"learning_rate": 4.087154986133944e-05,
"loss": 3.1689876556396483,
"step": 8030
},
{
"epoch": 1.5437079731027858,
"grad_norm": 0.34343641996383667,
"learning_rate": 4.070821852733074e-05,
"loss": 3.1644660949707033,
"step": 8035
},
{
"epoch": 1.5446685878962536,
"grad_norm": 0.3275468647480011,
"learning_rate": 4.054516294769871e-05,
"loss": 3.161859321594238,
"step": 8040
},
{
"epoch": 1.5456292026897214,
"grad_norm": 0.3492441773414612,
"learning_rate": 4.038238353384919e-05,
"loss": 3.166157531738281,
"step": 8045
},
{
"epoch": 1.5465898174831891,
"grad_norm": 0.3080005645751953,
"learning_rate": 4.021988069649138e-05,
"loss": 3.169676399230957,
"step": 8050
},
{
"epoch": 1.547550432276657,
"grad_norm": 0.2594451904296875,
"learning_rate": 4.0057654845636714e-05,
"loss": 3.170206642150879,
"step": 8055
},
{
"epoch": 1.5485110470701249,
"grad_norm": 0.2814268469810486,
"learning_rate": 3.989570639059753e-05,
"loss": 3.164537811279297,
"step": 8060
},
{
"epoch": 1.5494716618635929,
"grad_norm": 0.3419535756111145,
"learning_rate": 3.973403573998647e-05,
"loss": 3.169506645202637,
"step": 8065
},
{
"epoch": 1.5504322766570606,
"grad_norm": 0.3430168032646179,
"learning_rate": 3.957264330171511e-05,
"loss": 3.1708099365234377,
"step": 8070
},
{
"epoch": 1.5513928914505284,
"grad_norm": 0.3074200749397278,
"learning_rate": 3.941152948299321e-05,
"loss": 3.1663097381591796,
"step": 8075
},
{
"epoch": 1.5523535062439962,
"grad_norm": 0.29986321926116943,
"learning_rate": 3.9250694690327383e-05,
"loss": 3.1678709030151366,
"step": 8080
},
{
"epoch": 1.553314121037464,
"grad_norm": 0.305226594209671,
"learning_rate": 3.9090139329520327e-05,
"loss": 3.163295364379883,
"step": 8085
},
{
"epoch": 1.5542747358309317,
"grad_norm": 0.27455848455429077,
"learning_rate": 3.8929863805669706e-05,
"loss": 3.163608741760254,
"step": 8090
},
{
"epoch": 1.5552353506243997,
"grad_norm": 0.30126896500587463,
"learning_rate": 3.876986852316715e-05,
"loss": 3.1656238555908205,
"step": 8095
},
{
"epoch": 1.5561959654178674,
"grad_norm": 0.3914654552936554,
"learning_rate": 3.861015388569709e-05,
"loss": 3.1641334533691405,
"step": 8100
},
{
"epoch": 1.5571565802113354,
"grad_norm": 0.3612956404685974,
"learning_rate": 3.845072029623598e-05,
"loss": 3.1632358551025392,
"step": 8105
},
{
"epoch": 1.5581171950048032,
"grad_norm": 0.2885013222694397,
"learning_rate": 3.8291568157051154e-05,
"loss": 3.1646907806396483,
"step": 8110
},
{
"epoch": 1.559077809798271,
"grad_norm": 0.29719772934913635,
"learning_rate": 3.8132697869699705e-05,
"loss": 3.1659112930297852,
"step": 8115
},
{
"epoch": 1.5600384245917387,
"grad_norm": 0.308212012052536,
"learning_rate": 3.797410983502766e-05,
"loss": 3.1631364822387695,
"step": 8120
},
{
"epoch": 1.5609990393852065,
"grad_norm": 0.30908966064453125,
"learning_rate": 3.781580445316906e-05,
"loss": 3.1642078399658202,
"step": 8125
},
{
"epoch": 1.5619596541786742,
"grad_norm": 0.3072703182697296,
"learning_rate": 3.765778212354445e-05,
"loss": 3.1633129119873047,
"step": 8130
},
{
"epoch": 1.5629202689721422,
"grad_norm": 0.3363180458545685,
"learning_rate": 3.7500043244860475e-05,
"loss": 3.16600341796875,
"step": 8135
},
{
"epoch": 1.56388088376561,
"grad_norm": 0.41240906715393066,
"learning_rate": 3.7342588215108446e-05,
"loss": 3.168244743347168,
"step": 8140
},
{
"epoch": 1.564841498559078,
"grad_norm": 0.3254190683364868,
"learning_rate": 3.7185417431563644e-05,
"loss": 3.1647628784179687,
"step": 8145
},
{
"epoch": 1.5658021133525457,
"grad_norm": 0.3488558232784271,
"learning_rate": 3.702853129078398e-05,
"loss": 3.1683582305908202,
"step": 8150
},
{
"epoch": 1.5667627281460135,
"grad_norm": 0.34358084201812744,
"learning_rate": 3.6871930188609325e-05,
"loss": 3.1676990509033205,
"step": 8155
},
{
"epoch": 1.5677233429394812,
"grad_norm": 0.3370071053504944,
"learning_rate": 3.671561452016033e-05,
"loss": 3.1652057647705076,
"step": 8160
},
{
"epoch": 1.568683957732949,
"grad_norm": 0.23271813988685608,
"learning_rate": 3.655958467983749e-05,
"loss": 3.164310073852539,
"step": 8165
},
{
"epoch": 1.5696445725264168,
"grad_norm": 0.33882594108581543,
"learning_rate": 3.6403841061320026e-05,
"loss": 3.1641658782958983,
"step": 8170
},
{
"epoch": 1.5706051873198847,
"grad_norm": 0.25975945591926575,
"learning_rate": 3.6248384057565104e-05,
"loss": 3.1639142990112306,
"step": 8175
},
{
"epoch": 1.5715658021133525,
"grad_norm": 0.3010103404521942,
"learning_rate": 3.6093214060806686e-05,
"loss": 3.166419792175293,
"step": 8180
},
{
"epoch": 1.5725264169068205,
"grad_norm": 0.26971593499183655,
"learning_rate": 3.593833146255461e-05,
"loss": 3.1657012939453124,
"step": 8185
},
{
"epoch": 1.5734870317002883,
"grad_norm": 0.2859204411506653,
"learning_rate": 3.5783736653593546e-05,
"loss": 3.163465118408203,
"step": 8190
},
{
"epoch": 1.574447646493756,
"grad_norm": 0.2561207115650177,
"learning_rate": 3.56294300239821e-05,
"loss": 3.164503288269043,
"step": 8195
},
{
"epoch": 1.5754082612872238,
"grad_norm": 0.22760829329490662,
"learning_rate": 3.547541196305166e-05,
"loss": 3.1677932739257812,
"step": 8200
},
{
"epoch": 1.5763688760806915,
"grad_norm": 0.3261941969394684,
"learning_rate": 3.53216828594056e-05,
"loss": 3.164171600341797,
"step": 8205
},
{
"epoch": 1.5773294908741593,
"grad_norm": 0.3237496018409729,
"learning_rate": 3.5168243100918254e-05,
"loss": 3.1644121170043946,
"step": 8210
},
{
"epoch": 1.5782901056676273,
"grad_norm": 0.308450311422348,
"learning_rate": 3.501509307473391e-05,
"loss": 3.167286682128906,
"step": 8215
},
{
"epoch": 1.579250720461095,
"grad_norm": 0.2737598717212677,
"learning_rate": 3.486223316726569e-05,
"loss": 3.159839630126953,
"step": 8220
},
{
"epoch": 1.580211335254563,
"grad_norm": 0.2531875967979431,
"learning_rate": 3.470966376419489e-05,
"loss": 3.165401268005371,
"step": 8225
},
{
"epoch": 1.5811719500480308,
"grad_norm": 0.3886658847332001,
"learning_rate": 3.455738525046976e-05,
"loss": 3.164730453491211,
"step": 8230
},
{
"epoch": 1.5821325648414986,
"grad_norm": 0.32052338123321533,
"learning_rate": 3.440539801030463e-05,
"loss": 3.1636079788208007,
"step": 8235
},
{
"epoch": 1.5830931796349663,
"grad_norm": 0.3415239751338959,
"learning_rate": 3.425370242717887e-05,
"loss": 3.1651962280273436,
"step": 8240
},
{
"epoch": 1.584053794428434,
"grad_norm": 0.27457112073898315,
"learning_rate": 3.410229888383597e-05,
"loss": 3.169073486328125,
"step": 8245
},
{
"epoch": 1.585014409221902,
"grad_norm": 0.3607783615589142,
"learning_rate": 3.395118776228265e-05,
"loss": 3.1643707275390627,
"step": 8250
},
{
"epoch": 1.5859750240153698,
"grad_norm": 0.44193530082702637,
"learning_rate": 3.380036944378775e-05,
"loss": 3.165296936035156,
"step": 8255
},
{
"epoch": 1.5869356388088378,
"grad_norm": 0.27032527327537537,
"learning_rate": 3.364984430888138e-05,
"loss": 3.1653570175170898,
"step": 8260
},
{
"epoch": 1.5878962536023056,
"grad_norm": 0.3761611580848694,
"learning_rate": 3.34996127373539e-05,
"loss": 3.1637210845947266,
"step": 8265
},
{
"epoch": 1.5888568683957733,
"grad_norm": 0.2229749858379364,
"learning_rate": 3.3349675108254946e-05,
"loss": 3.165303421020508,
"step": 8270
},
{
"epoch": 1.589817483189241,
"grad_norm": 0.23744769394397736,
"learning_rate": 3.320003179989254e-05,
"loss": 3.16616268157959,
"step": 8275
},
{
"epoch": 1.5907780979827089,
"grad_norm": 0.28989988565444946,
"learning_rate": 3.305068318983211e-05,
"loss": 3.1658939361572265,
"step": 8280
},
{
"epoch": 1.5917387127761766,
"grad_norm": 0.2983841300010681,
"learning_rate": 3.290162965489558e-05,
"loss": 3.1631855010986327,
"step": 8285
},
{
"epoch": 1.5926993275696446,
"grad_norm": 0.2570306658744812,
"learning_rate": 3.275287157116021e-05,
"loss": 3.1675746917724608,
"step": 8290
},
{
"epoch": 1.5936599423631124,
"grad_norm": 0.3952568471431732,
"learning_rate": 3.260440931395794e-05,
"loss": 3.1651771545410154,
"step": 8295
},
{
"epoch": 1.5946205571565804,
"grad_norm": 0.3825446367263794,
"learning_rate": 3.24562432578743e-05,
"loss": 3.1654695510864257,
"step": 8300
},
{
"epoch": 1.5955811719500481,
"grad_norm": 0.2898649275302887,
"learning_rate": 3.230837377674746e-05,
"loss": 3.1621742248535156,
"step": 8305
},
{
"epoch": 1.5965417867435159,
"grad_norm": 0.24470216035842896,
"learning_rate": 3.216080124366724e-05,
"loss": 3.1656517028808593,
"step": 8310
},
{
"epoch": 1.5975024015369836,
"grad_norm": 0.23829184472560883,
"learning_rate": 3.20135260309743e-05,
"loss": 3.166254425048828,
"step": 8315
},
{
"epoch": 1.5984630163304514,
"grad_norm": 0.2703918218612671,
"learning_rate": 3.186654851025911e-05,
"loss": 3.1674873352050783,
"step": 8320
},
{
"epoch": 1.5994236311239192,
"grad_norm": 0.2798570394515991,
"learning_rate": 3.171986905236104e-05,
"loss": 3.16519718170166,
"step": 8325
},
{
"epoch": 1.6003842459173871,
"grad_norm": 0.29663723707199097,
"learning_rate": 3.15734880273674e-05,
"loss": 3.1651836395263673,
"step": 8330
},
{
"epoch": 1.601344860710855,
"grad_norm": 0.2850891351699829,
"learning_rate": 3.142740580461261e-05,
"loss": 3.1634387969970703,
"step": 8335
},
{
"epoch": 1.602305475504323,
"grad_norm": 0.3124440610408783,
"learning_rate": 3.128162275267697e-05,
"loss": 3.1671958923339845,
"step": 8340
},
{
"epoch": 1.6032660902977907,
"grad_norm": 0.3228466808795929,
"learning_rate": 3.113613923938614e-05,
"loss": 3.1625482559204103,
"step": 8345
},
{
"epoch": 1.6042267050912584,
"grad_norm": 0.23934929072856903,
"learning_rate": 3.0990955631809965e-05,
"loss": 3.1659454345703124,
"step": 8350
},
{
"epoch": 1.6051873198847262,
"grad_norm": 0.2973260283470154,
"learning_rate": 3.08460722962616e-05,
"loss": 3.167291259765625,
"step": 8355
},
{
"epoch": 1.606147934678194,
"grad_norm": 0.26656073331832886,
"learning_rate": 3.070148959829649e-05,
"loss": 3.165090560913086,
"step": 8360
},
{
"epoch": 1.6071085494716617,
"grad_norm": 0.25156885385513306,
"learning_rate": 3.055720790271164e-05,
"loss": 3.1673526763916016,
"step": 8365
},
{
"epoch": 1.6080691642651297,
"grad_norm": 0.2427086979150772,
"learning_rate": 3.0413227573544592e-05,
"loss": 3.167298698425293,
"step": 8370
},
{
"epoch": 1.6090297790585975,
"grad_norm": 0.33012548089027405,
"learning_rate": 3.026954897407252e-05,
"loss": 3.16772403717041,
"step": 8375
},
{
"epoch": 1.6099903938520654,
"grad_norm": 0.19765010476112366,
"learning_rate": 3.01261724668112e-05,
"loss": 3.1641408920288088,
"step": 8380
},
{
"epoch": 1.6109510086455332,
"grad_norm": 0.25813835859298706,
"learning_rate": 2.9983098413514284e-05,
"loss": 3.168695068359375,
"step": 8385
},
{
"epoch": 1.611911623439001,
"grad_norm": 0.2662206292152405,
"learning_rate": 2.9840327175172295e-05,
"loss": 3.1604537963867188,
"step": 8390
},
{
"epoch": 1.6128722382324687,
"grad_norm": 0.2579902410507202,
"learning_rate": 2.969785911201172e-05,
"loss": 3.1639808654785155,
"step": 8395
},
{
"epoch": 1.6138328530259365,
"grad_norm": 0.27471011877059937,
"learning_rate": 2.9555694583494095e-05,
"loss": 3.1644439697265625,
"step": 8400
},
{
"epoch": 1.6147934678194045,
"grad_norm": 0.2601883113384247,
"learning_rate": 2.9413833948315163e-05,
"loss": 3.1667598724365233,
"step": 8405
},
{
"epoch": 1.6157540826128722,
"grad_norm": 0.27512237429618835,
"learning_rate": 2.9272277564403746e-05,
"loss": 3.1641109466552733,
"step": 8410
},
{
"epoch": 1.6167146974063402,
"grad_norm": 0.25258708000183105,
"learning_rate": 2.9131025788921193e-05,
"loss": 3.1663232803344727,
"step": 8415
},
{
"epoch": 1.617675312199808,
"grad_norm": 0.22575309872627258,
"learning_rate": 2.8990078978260216e-05,
"loss": 3.166961669921875,
"step": 8420
},
{
"epoch": 1.6186359269932757,
"grad_norm": 0.24490588903427124,
"learning_rate": 2.8849437488044118e-05,
"loss": 3.1663583755493163,
"step": 8425
},
{
"epoch": 1.6195965417867435,
"grad_norm": 0.2155570238828659,
"learning_rate": 2.8709101673125728e-05,
"loss": 3.1656875610351562,
"step": 8430
},
{
"epoch": 1.6205571565802113,
"grad_norm": 0.2795376479625702,
"learning_rate": 2.856907188758674e-05,
"loss": 3.162942314147949,
"step": 8435
},
{
"epoch": 1.621517771373679,
"grad_norm": 0.3358766734600067,
"learning_rate": 2.8429348484736658e-05,
"loss": 3.1658599853515623,
"step": 8440
},
{
"epoch": 1.622478386167147,
"grad_norm": 0.2666330933570862,
"learning_rate": 2.8289931817112027e-05,
"loss": 3.1691190719604494,
"step": 8445
},
{
"epoch": 1.6234390009606148,
"grad_norm": 0.2384926676750183,
"learning_rate": 2.8150822236475283e-05,
"loss": 3.1659183502197266,
"step": 8450
},
{
"epoch": 1.6243996157540828,
"grad_norm": 0.25563499331474304,
"learning_rate": 2.8012020093814252e-05,
"loss": 3.1584651947021483,
"step": 8455
},
{
"epoch": 1.6253602305475505,
"grad_norm": 0.2299119532108307,
"learning_rate": 2.7873525739340973e-05,
"loss": 3.1645713806152345,
"step": 8460
},
{
"epoch": 1.6263208453410183,
"grad_norm": 0.26834285259246826,
"learning_rate": 2.773533952249088e-05,
"loss": 3.166657257080078,
"step": 8465
},
{
"epoch": 1.627281460134486,
"grad_norm": 0.2765657901763916,
"learning_rate": 2.7597461791921987e-05,
"loss": 3.165719223022461,
"step": 8470
},
{
"epoch": 1.6282420749279538,
"grad_norm": 0.26235702633857727,
"learning_rate": 2.745989289551401e-05,
"loss": 3.163796615600586,
"step": 8475
},
{
"epoch": 1.6292026897214216,
"grad_norm": 0.2585254907608032,
"learning_rate": 2.7322633180367286e-05,
"loss": 3.1652366638183596,
"step": 8480
},
{
"epoch": 1.6301633045148896,
"grad_norm": 0.24528291821479797,
"learning_rate": 2.7185682992802215e-05,
"loss": 3.165533447265625,
"step": 8485
},
{
"epoch": 1.6311239193083573,
"grad_norm": 0.24676673114299774,
"learning_rate": 2.7049042678358157e-05,
"loss": 3.162363624572754,
"step": 8490
},
{
"epoch": 1.6320845341018253,
"grad_norm": 0.28905969858169556,
"learning_rate": 2.6912712581792684e-05,
"loss": 3.165178680419922,
"step": 8495
},
{
"epoch": 1.633045148895293,
"grad_norm": 0.25018957257270813,
"learning_rate": 2.6776693047080546e-05,
"loss": 3.163674736022949,
"step": 8500
},
{
"epoch": 1.6340057636887608,
"grad_norm": 0.24032709002494812,
"learning_rate": 2.6640984417412996e-05,
"loss": 3.1674957275390625,
"step": 8505
},
{
"epoch": 1.6349663784822286,
"grad_norm": 0.2825995981693268,
"learning_rate": 2.6505587035196862e-05,
"loss": 3.164845085144043,
"step": 8510
},
{
"epoch": 1.6359269932756964,
"grad_norm": 0.3082216680049896,
"learning_rate": 2.6370501242053655e-05,
"loss": 3.164422607421875,
"step": 8515
},
{
"epoch": 1.6368876080691641,
"grad_norm": 0.2581336796283722,
"learning_rate": 2.6235727378818617e-05,
"loss": 3.1643039703369142,
"step": 8520
},
{
"epoch": 1.637848222862632,
"grad_norm": 0.2364456057548523,
"learning_rate": 2.6101265785540054e-05,
"loss": 3.1647056579589843,
"step": 8525
},
{
"epoch": 1.6388088376560999,
"grad_norm": 0.2648424804210663,
"learning_rate": 2.596711680147837e-05,
"loss": 3.1666595458984377,
"step": 8530
},
{
"epoch": 1.6397694524495678,
"grad_norm": 0.21114224195480347,
"learning_rate": 2.5833280765105218e-05,
"loss": 3.1652767181396486,
"step": 8535
},
{
"epoch": 1.6407300672430356,
"grad_norm": 0.29811277985572815,
"learning_rate": 2.5699758014102627e-05,
"loss": 3.1638210296630858,
"step": 8540
},
{
"epoch": 1.6416906820365034,
"grad_norm": 0.22697743773460388,
"learning_rate": 2.5566548885362233e-05,
"loss": 3.1660308837890625,
"step": 8545
},
{
"epoch": 1.6426512968299711,
"grad_norm": 0.2113056480884552,
"learning_rate": 2.543365371498434e-05,
"loss": 3.1632022857666016,
"step": 8550
},
{
"epoch": 1.643611911623439,
"grad_norm": 0.24464410543441772,
"learning_rate": 2.5301072838277026e-05,
"loss": 3.1658231735229494,
"step": 8555
},
{
"epoch": 1.6445725264169067,
"grad_norm": 0.23001989722251892,
"learning_rate": 2.5168806589755497e-05,
"loss": 3.163945960998535,
"step": 8560
},
{
"epoch": 1.6455331412103746,
"grad_norm": 0.20726712048053741,
"learning_rate": 2.503685530314109e-05,
"loss": 3.162770080566406,
"step": 8565
},
{
"epoch": 1.6464937560038426,
"grad_norm": 0.23915457725524902,
"learning_rate": 2.490521931136036e-05,
"loss": 3.1642734527587892,
"step": 8570
},
{
"epoch": 1.6474543707973104,
"grad_norm": 0.2551625370979309,
"learning_rate": 2.4773898946544473e-05,
"loss": 3.1642024993896483,
"step": 8575
},
{
"epoch": 1.6484149855907781,
"grad_norm": 0.2408856302499771,
"learning_rate": 2.4642894540028164e-05,
"loss": 3.1599807739257812,
"step": 8580
},
{
"epoch": 1.649375600384246,
"grad_norm": 0.2513860762119293,
"learning_rate": 2.4512206422349024e-05,
"loss": 3.163848876953125,
"step": 8585
},
{
"epoch": 1.6503362151777137,
"grad_norm": 0.21447288990020752,
"learning_rate": 2.438183492324654e-05,
"loss": 3.1619583129882813,
"step": 8590
},
{
"epoch": 1.6512968299711814,
"grad_norm": 0.1881914734840393,
"learning_rate": 2.4251780371661373e-05,
"loss": 3.1636112213134764,
"step": 8595
},
{
"epoch": 1.6522574447646494,
"grad_norm": 0.19982366263866425,
"learning_rate": 2.4122043095734518e-05,
"loss": 3.1659828186035157,
"step": 8600
},
{
"epoch": 1.6532180595581172,
"grad_norm": 0.2119777649641037,
"learning_rate": 2.3992623422806444e-05,
"loss": 3.161235809326172,
"step": 8605
},
{
"epoch": 1.6541786743515852,
"grad_norm": 0.2393619865179062,
"learning_rate": 2.3863521679416237e-05,
"loss": 3.1633071899414062,
"step": 8610
},
{
"epoch": 1.655139289145053,
"grad_norm": 0.24367739260196686,
"learning_rate": 2.3734738191300862e-05,
"loss": 3.16229133605957,
"step": 8615
},
{
"epoch": 1.6560999039385207,
"grad_norm": 0.29672330617904663,
"learning_rate": 2.36062732833943e-05,
"loss": 3.1618356704711914,
"step": 8620
},
{
"epoch": 1.6570605187319885,
"grad_norm": 0.24469807744026184,
"learning_rate": 2.347812727982661e-05,
"loss": 3.161136245727539,
"step": 8625
},
{
"epoch": 1.6580211335254562,
"grad_norm": 0.21228572726249695,
"learning_rate": 2.3350300503923352e-05,
"loss": 3.1614845275878904,
"step": 8630
},
{
"epoch": 1.658981748318924,
"grad_norm": 0.20836694538593292,
"learning_rate": 2.3222793278204626e-05,
"loss": 3.1641204833984373,
"step": 8635
},
{
"epoch": 1.659942363112392,
"grad_norm": 0.2582319974899292,
"learning_rate": 2.309560592438417e-05,
"loss": 3.1627979278564453,
"step": 8640
},
{
"epoch": 1.6609029779058597,
"grad_norm": 0.23718759417533875,
"learning_rate": 2.2968738763368765e-05,
"loss": 3.1604846954345702,
"step": 8645
},
{
"epoch": 1.6618635926993277,
"grad_norm": 0.19198426604270935,
"learning_rate": 2.2842192115257295e-05,
"loss": 3.1641172409057616,
"step": 8650
},
{
"epoch": 1.6628242074927955,
"grad_norm": 0.258806049823761,
"learning_rate": 2.271596629933992e-05,
"loss": 3.16113224029541,
"step": 8655
},
{
"epoch": 1.6637848222862632,
"grad_norm": 0.21029417216777802,
"learning_rate": 2.25900616340973e-05,
"loss": 3.1632869720458983,
"step": 8660
},
{
"epoch": 1.664745437079731,
"grad_norm": 0.21884505450725555,
"learning_rate": 2.2464478437199808e-05,
"loss": 3.163255310058594,
"step": 8665
},
{
"epoch": 1.6657060518731988,
"grad_norm": 0.22930586338043213,
"learning_rate": 2.2339217025506813e-05,
"loss": 3.163426399230957,
"step": 8670
},
{
"epoch": 1.6666666666666665,
"grad_norm": 0.2358139008283615,
"learning_rate": 2.2214277715065636e-05,
"loss": 3.1608362197875977,
"step": 8675
},
{
"epoch": 1.6676272814601345,
"grad_norm": 0.22611092031002045,
"learning_rate": 2.2089660821110988e-05,
"loss": 3.1634220123291015,
"step": 8680
},
{
"epoch": 1.6685878962536023,
"grad_norm": 0.19509707391262054,
"learning_rate": 2.1965366658064086e-05,
"loss": 3.158903884887695,
"step": 8685
},
{
"epoch": 1.6695485110470702,
"grad_norm": 0.19261038303375244,
"learning_rate": 2.1841395539531893e-05,
"loss": 3.161344528198242,
"step": 8690
},
{
"epoch": 1.670509125840538,
"grad_norm": 0.1918260157108307,
"learning_rate": 2.1717747778306177e-05,
"loss": 3.1661489486694334,
"step": 8695
},
{
"epoch": 1.6714697406340058,
"grad_norm": 0.20543381571769714,
"learning_rate": 2.159442368636297e-05,
"loss": 3.1635005950927733,
"step": 8700
},
{
"epoch": 1.6724303554274735,
"grad_norm": 0.2433311641216278,
"learning_rate": 2.147142357486164e-05,
"loss": 3.163587951660156,
"step": 8705
},
{
"epoch": 1.6733909702209413,
"grad_norm": 0.21309548616409302,
"learning_rate": 2.1348747754144004e-05,
"loss": 3.1650224685668946,
"step": 8710
},
{
"epoch": 1.674351585014409,
"grad_norm": 0.2069951891899109,
"learning_rate": 2.1226396533733796e-05,
"loss": 3.164658546447754,
"step": 8715
},
{
"epoch": 1.675312199807877,
"grad_norm": 0.24692699313163757,
"learning_rate": 2.1104370222335688e-05,
"loss": 3.1658843994140624,
"step": 8720
},
{
"epoch": 1.6762728146013448,
"grad_norm": 0.23039954900741577,
"learning_rate": 2.0982669127834622e-05,
"loss": 3.1638277053833006,
"step": 8725
},
{
"epoch": 1.6772334293948128,
"grad_norm": 0.2396543025970459,
"learning_rate": 2.0861293557294862e-05,
"loss": 3.162588119506836,
"step": 8730
},
{
"epoch": 1.6781940441882806,
"grad_norm": 0.2332344651222229,
"learning_rate": 2.0740243816959452e-05,
"loss": 3.163272476196289,
"step": 8735
},
{
"epoch": 1.6791546589817483,
"grad_norm": 0.21478639543056488,
"learning_rate": 2.061952021224938e-05,
"loss": 3.162495803833008,
"step": 8740
},
{
"epoch": 1.680115273775216,
"grad_norm": 0.19731956720352173,
"learning_rate": 2.0499123047762576e-05,
"loss": 3.162389945983887,
"step": 8745
},
{
"epoch": 1.6810758885686838,
"grad_norm": 0.19572487473487854,
"learning_rate": 2.03790526272735e-05,
"loss": 3.1616092681884767,
"step": 8750
},
{
"epoch": 1.6820365033621518,
"grad_norm": 0.184941828250885,
"learning_rate": 2.025930925373213e-05,
"loss": 3.1640493392944338,
"step": 8755
},
{
"epoch": 1.6829971181556196,
"grad_norm": 0.1868741363286972,
"learning_rate": 2.013989322926331e-05,
"loss": 3.164847183227539,
"step": 8760
},
{
"epoch": 1.6839577329490876,
"grad_norm": 0.19907650351524353,
"learning_rate": 2.0020804855165857e-05,
"loss": 3.164986801147461,
"step": 8765
},
{
"epoch": 1.6849183477425553,
"grad_norm": 0.19891948997974396,
"learning_rate": 1.9902044431912e-05,
"loss": 3.16265869140625,
"step": 8770
},
{
"epoch": 1.685878962536023,
"grad_norm": 0.23641318082809448,
"learning_rate": 1.9783612259146485e-05,
"loss": 3.164572525024414,
"step": 8775
},
{
"epoch": 1.6868395773294909,
"grad_norm": 0.2543870210647583,
"learning_rate": 1.9665508635685767e-05,
"loss": 3.1619945526123048,
"step": 8780
},
{
"epoch": 1.6878001921229586,
"grad_norm": 0.1898890733718872,
"learning_rate": 1.9547733859517443e-05,
"loss": 3.161852264404297,
"step": 8785
},
{
"epoch": 1.6887608069164264,
"grad_norm": 0.18471160531044006,
"learning_rate": 1.9430288227799367e-05,
"loss": 3.167014312744141,
"step": 8790
},
{
"epoch": 1.6897214217098944,
"grad_norm": 0.1837644726037979,
"learning_rate": 1.931317203685893e-05,
"loss": 3.1675960540771486,
"step": 8795
},
{
"epoch": 1.6906820365033621,
"grad_norm": 0.16458193957805634,
"learning_rate": 1.9196385582192247e-05,
"loss": 3.161328125,
"step": 8800
},
{
"epoch": 1.6916426512968301,
"grad_norm": 0.18657121062278748,
"learning_rate": 1.9079929158463526e-05,
"loss": 3.163207244873047,
"step": 8805
},
{
"epoch": 1.6926032660902979,
"grad_norm": 0.2095474749803543,
"learning_rate": 1.896380305950434e-05,
"loss": 3.1646053314208986,
"step": 8810
},
{
"epoch": 1.6935638808837656,
"grad_norm": 0.1963958740234375,
"learning_rate": 1.8848007578312686e-05,
"loss": 3.164917755126953,
"step": 8815
},
{
"epoch": 1.6945244956772334,
"grad_norm": 0.18170610070228577,
"learning_rate": 1.8732543007052452e-05,
"loss": 3.1576461791992188,
"step": 8820
},
{
"epoch": 1.6954851104707012,
"grad_norm": 0.2065751552581787,
"learning_rate": 1.8617409637052606e-05,
"loss": 3.166869354248047,
"step": 8825
},
{
"epoch": 1.696445725264169,
"grad_norm": 0.18570922315120697,
"learning_rate": 1.8502607758806487e-05,
"loss": 3.162985992431641,
"step": 8830
},
{
"epoch": 1.697406340057637,
"grad_norm": 0.22870124876499176,
"learning_rate": 1.8388137661970948e-05,
"loss": 3.165153694152832,
"step": 8835
},
{
"epoch": 1.6983669548511047,
"grad_norm": 0.18405385315418243,
"learning_rate": 1.8273999635365806e-05,
"loss": 3.162175750732422,
"step": 8840
},
{
"epoch": 1.6993275696445727,
"grad_norm": 0.1959793120622635,
"learning_rate": 1.8160193966973047e-05,
"loss": 3.161460113525391,
"step": 8845
},
{
"epoch": 1.7002881844380404,
"grad_norm": 0.20207500457763672,
"learning_rate": 1.8046720943936e-05,
"loss": 3.1632549285888674,
"step": 8850
},
{
"epoch": 1.7012487992315082,
"grad_norm": 0.20162688195705414,
"learning_rate": 1.7933580852558742e-05,
"loss": 3.1618804931640625,
"step": 8855
},
{
"epoch": 1.702209414024976,
"grad_norm": 0.18540622293949127,
"learning_rate": 1.7820773978305365e-05,
"loss": 3.161595916748047,
"step": 8860
},
{
"epoch": 1.7031700288184437,
"grad_norm": 0.19820909202098846,
"learning_rate": 1.7708300605799202e-05,
"loss": 3.1606048583984374,
"step": 8865
},
{
"epoch": 1.7041306436119115,
"grad_norm": 0.2218032330274582,
"learning_rate": 1.7596161018822007e-05,
"loss": 3.1620101928710938,
"step": 8870
},
{
"epoch": 1.7050912584053795,
"grad_norm": 0.1894785612821579,
"learning_rate": 1.7484355500313568e-05,
"loss": 3.161252975463867,
"step": 8875
},
{
"epoch": 1.7060518731988472,
"grad_norm": 0.20882797241210938,
"learning_rate": 1.7372884332370677e-05,
"loss": 3.162582015991211,
"step": 8880
},
{
"epoch": 1.7070124879923152,
"grad_norm": 0.2097301483154297,
"learning_rate": 1.7261747796246478e-05,
"loss": 3.162227249145508,
"step": 8885
},
{
"epoch": 1.707973102785783,
"grad_norm": 0.18620242178440094,
"learning_rate": 1.7150946172349844e-05,
"loss": 3.163176345825195,
"step": 8890
},
{
"epoch": 1.7089337175792507,
"grad_norm": 0.1744890809059143,
"learning_rate": 1.7040479740244684e-05,
"loss": 3.1600067138671877,
"step": 8895
},
{
"epoch": 1.7098943323727185,
"grad_norm": 0.20441320538520813,
"learning_rate": 1.693034877864913e-05,
"loss": 3.161859130859375,
"step": 8900
},
{
"epoch": 1.7108549471661862,
"grad_norm": 0.18480157852172852,
"learning_rate": 1.682055356543487e-05,
"loss": 3.1619277954101563,
"step": 8905
},
{
"epoch": 1.7118155619596542,
"grad_norm": 0.19677984714508057,
"learning_rate": 1.6711094377626495e-05,
"loss": 3.1647516250610352,
"step": 8910
},
{
"epoch": 1.712776176753122,
"grad_norm": 0.19380095601081848,
"learning_rate": 1.66019714914008e-05,
"loss": 3.1602848052978514,
"step": 8915
},
{
"epoch": 1.71373679154659,
"grad_norm": 0.17474035918712616,
"learning_rate": 1.6493185182085967e-05,
"loss": 3.164753532409668,
"step": 8920
},
{
"epoch": 1.7146974063400577,
"grad_norm": 0.18277983367443085,
"learning_rate": 1.6384735724161045e-05,
"loss": 3.162521743774414,
"step": 8925
},
{
"epoch": 1.7156580211335255,
"grad_norm": 0.17841650545597076,
"learning_rate": 1.6276623391255146e-05,
"loss": 3.1669170379638674,
"step": 8930
},
{
"epoch": 1.7166186359269933,
"grad_norm": 0.18692415952682495,
"learning_rate": 1.6168848456146793e-05,
"loss": 3.1644006729125977,
"step": 8935
},
{
"epoch": 1.717579250720461,
"grad_norm": 0.22790934145450592,
"learning_rate": 1.606141119076314e-05,
"loss": 3.164959716796875,
"step": 8940
},
{
"epoch": 1.7185398655139288,
"grad_norm": 0.19892485439777374,
"learning_rate": 1.595431186617948e-05,
"loss": 3.1620708465576173,
"step": 8945
},
{
"epoch": 1.7195004803073968,
"grad_norm": 0.19590741395950317,
"learning_rate": 1.5847550752618427e-05,
"loss": 3.158066177368164,
"step": 8950
},
{
"epoch": 1.7204610951008645,
"grad_norm": 0.2052772343158722,
"learning_rate": 1.5741128119449153e-05,
"loss": 3.161151885986328,
"step": 8955
},
{
"epoch": 1.7214217098943325,
"grad_norm": 0.18683621287345886,
"learning_rate": 1.5635044235186906e-05,
"loss": 3.1654678344726563,
"step": 8960
},
{
"epoch": 1.7223823246878003,
"grad_norm": 0.19518622756004333,
"learning_rate": 1.5529299367492208e-05,
"loss": 3.160481262207031,
"step": 8965
},
{
"epoch": 1.723342939481268,
"grad_norm": 0.1885916143655777,
"learning_rate": 1.5423893783170262e-05,
"loss": 3.162572479248047,
"step": 8970
},
{
"epoch": 1.7243035542747358,
"grad_norm": 0.18152864277362823,
"learning_rate": 1.531882774817007e-05,
"loss": 3.1627574920654298,
"step": 8975
},
{
"epoch": 1.7252641690682036,
"grad_norm": 0.17760129272937775,
"learning_rate": 1.5214101527584071e-05,
"loss": 3.1614215850830076,
"step": 8980
},
{
"epoch": 1.7262247838616713,
"grad_norm": 0.20811888575553894,
"learning_rate": 1.5109715385647297e-05,
"loss": 3.163728141784668,
"step": 8985
},
{
"epoch": 1.7271853986551393,
"grad_norm": 0.18515148758888245,
"learning_rate": 1.5005669585736618e-05,
"loss": 3.161965179443359,
"step": 8990
},
{
"epoch": 1.728146013448607,
"grad_norm": 0.1933458298444748,
"learning_rate": 1.4901964390370308e-05,
"loss": 3.163407325744629,
"step": 8995
},
{
"epoch": 1.729106628242075,
"grad_norm": 0.20849090814590454,
"learning_rate": 1.4798600061207195e-05,
"loss": 3.163631057739258,
"step": 9000
},
{
"epoch": 1.7300672430355428,
"grad_norm": 0.18622250854969025,
"learning_rate": 1.4695576859046127e-05,
"loss": 3.159561538696289,
"step": 9005
},
{
"epoch": 1.7310278578290106,
"grad_norm": 0.1835029423236847,
"learning_rate": 1.4592895043825126e-05,
"loss": 3.163999557495117,
"step": 9010
},
{
"epoch": 1.7319884726224783,
"grad_norm": 0.1808691918849945,
"learning_rate": 1.449055487462102e-05,
"loss": 3.1594869613647463,
"step": 9015
},
{
"epoch": 1.732949087415946,
"grad_norm": 0.16524961590766907,
"learning_rate": 1.4388556609648572e-05,
"loss": 3.163185691833496,
"step": 9020
},
{
"epoch": 1.7339097022094139,
"grad_norm": 0.16896623373031616,
"learning_rate": 1.428690050625979e-05,
"loss": 3.1632381439208985,
"step": 9025
},
{
"epoch": 1.7348703170028819,
"grad_norm": 0.1609111875295639,
"learning_rate": 1.4185586820943506e-05,
"loss": 3.1643299102783202,
"step": 9030
},
{
"epoch": 1.7358309317963496,
"grad_norm": 0.16877403855323792,
"learning_rate": 1.4084615809324523e-05,
"loss": 3.164197540283203,
"step": 9035
},
{
"epoch": 1.7367915465898176,
"grad_norm": 0.17374147474765778,
"learning_rate": 1.3983987726163087e-05,
"loss": 3.1597652435302734,
"step": 9040
},
{
"epoch": 1.7377521613832854,
"grad_norm": 0.1869489699602127,
"learning_rate": 1.3883702825354138e-05,
"loss": 3.160333251953125,
"step": 9045
},
{
"epoch": 1.7387127761767531,
"grad_norm": 0.19941848516464233,
"learning_rate": 1.3783761359926771e-05,
"loss": 3.1596229553222654,
"step": 9050
},
{
"epoch": 1.739673390970221,
"grad_norm": 0.16694855690002441,
"learning_rate": 1.3684163582043595e-05,
"loss": 3.159475898742676,
"step": 9055
},
{
"epoch": 1.7406340057636887,
"grad_norm": 0.1921820491552353,
"learning_rate": 1.3584909742999978e-05,
"loss": 3.162578010559082,
"step": 9060
},
{
"epoch": 1.7415946205571564,
"grad_norm": 0.17421427369117737,
"learning_rate": 1.3486000093223565e-05,
"loss": 3.1633106231689454,
"step": 9065
},
{
"epoch": 1.7425552353506244,
"grad_norm": 0.16822272539138794,
"learning_rate": 1.3387434882273529e-05,
"loss": 3.1628774642944335,
"step": 9070
},
{
"epoch": 1.7435158501440924,
"grad_norm": 0.2147689312696457,
"learning_rate": 1.3289214358840022e-05,
"loss": 3.164061737060547,
"step": 9075
},
{
"epoch": 1.7444764649375601,
"grad_norm": 0.19877059757709503,
"learning_rate": 1.3191338770743493e-05,
"loss": 3.164179039001465,
"step": 9080
},
{
"epoch": 1.745437079731028,
"grad_norm": 0.17470191419124603,
"learning_rate": 1.30938083649341e-05,
"loss": 3.162880706787109,
"step": 9085
},
{
"epoch": 1.7463976945244957,
"grad_norm": 0.17637498676776886,
"learning_rate": 1.2996623387491085e-05,
"loss": 3.163307952880859,
"step": 9090
},
{
"epoch": 1.7473583093179634,
"grad_norm": 0.17012238502502441,
"learning_rate": 1.2899784083622067e-05,
"loss": 3.1613311767578125,
"step": 9095
},
{
"epoch": 1.7483189241114312,
"grad_norm": 0.15726859867572784,
"learning_rate": 1.2803290697662566e-05,
"loss": 3.161721038818359,
"step": 9100
},
{
"epoch": 1.7492795389048992,
"grad_norm": 0.1671910136938095,
"learning_rate": 1.2707143473075299e-05,
"loss": 3.1644134521484375,
"step": 9105
},
{
"epoch": 1.750240153698367,
"grad_norm": 0.17212195694446564,
"learning_rate": 1.2611342652449597e-05,
"loss": 3.1598962783813476,
"step": 9110
},
{
"epoch": 1.751200768491835,
"grad_norm": 0.17955084145069122,
"learning_rate": 1.2515888477500708e-05,
"loss": 3.161183166503906,
"step": 9115
},
{
"epoch": 1.7521613832853027,
"grad_norm": 0.15037564933300018,
"learning_rate": 1.2420781189069346e-05,
"loss": 3.164091873168945,
"step": 9120
},
{
"epoch": 1.7531219980787704,
"grad_norm": 0.16244594752788544,
"learning_rate": 1.2326021027120958e-05,
"loss": 3.162504196166992,
"step": 9125
},
{
"epoch": 1.7540826128722382,
"grad_norm": 0.1728479117155075,
"learning_rate": 1.2231608230745128e-05,
"loss": 3.159666633605957,
"step": 9130
},
{
"epoch": 1.755043227665706,
"grad_norm": 0.17677444219589233,
"learning_rate": 1.2137543038155034e-05,
"loss": 3.16094970703125,
"step": 9135
},
{
"epoch": 1.7560038424591737,
"grad_norm": 0.16001811623573303,
"learning_rate": 1.2043825686686798e-05,
"loss": 3.1602264404296876,
"step": 9140
},
{
"epoch": 1.7569644572526417,
"grad_norm": 0.1979171186685562,
"learning_rate": 1.1950456412798954e-05,
"loss": 3.160991096496582,
"step": 9145
},
{
"epoch": 1.7579250720461095,
"grad_norm": 0.18909943103790283,
"learning_rate": 1.1857435452071706e-05,
"loss": 3.164293098449707,
"step": 9150
},
{
"epoch": 1.7588856868395775,
"grad_norm": 0.1578613966703415,
"learning_rate": 1.1764763039206516e-05,
"loss": 3.1651378631591798,
"step": 9155
},
{
"epoch": 1.7598463016330452,
"grad_norm": 0.17254026234149933,
"learning_rate": 1.1672439408025409e-05,
"loss": 3.1591054916381838,
"step": 9160
},
{
"epoch": 1.760806916426513,
"grad_norm": 0.1548478901386261,
"learning_rate": 1.1580464791470317e-05,
"loss": 3.162530517578125,
"step": 9165
},
{
"epoch": 1.7617675312199808,
"grad_norm": 0.16993309557437897,
"learning_rate": 1.1488839421602663e-05,
"loss": 3.1617578506469726,
"step": 9170
},
{
"epoch": 1.7627281460134485,
"grad_norm": 0.18223033845424652,
"learning_rate": 1.1397563529602654e-05,
"loss": 3.1608226776123045,
"step": 9175
},
{
"epoch": 1.7636887608069163,
"grad_norm": 0.15401305258274078,
"learning_rate": 1.130663734576877e-05,
"loss": 3.1587234497070313,
"step": 9180
},
{
"epoch": 1.7646493756003843,
"grad_norm": 0.19235272705554962,
"learning_rate": 1.1216061099517016e-05,
"loss": 3.158829116821289,
"step": 9185
},
{
"epoch": 1.765609990393852,
"grad_norm": 0.1655486822128296,
"learning_rate": 1.1125835019380614e-05,
"loss": 3.160569190979004,
"step": 9190
},
{
"epoch": 1.76657060518732,
"grad_norm": 0.17558416724205017,
"learning_rate": 1.1035959333009231e-05,
"loss": 3.1613176345825194,
"step": 9195
},
{
"epoch": 1.7675312199807878,
"grad_norm": 0.16277319192886353,
"learning_rate": 1.0946434267168375e-05,
"loss": 3.1619541168212892,
"step": 9200
},
{
"epoch": 1.7684918347742555,
"grad_norm": 0.15174464881420135,
"learning_rate": 1.0857260047739025e-05,
"loss": 3.1618457794189454,
"step": 9205
},
{
"epoch": 1.7694524495677233,
"grad_norm": 0.14863047003746033,
"learning_rate": 1.076843689971687e-05,
"loss": 3.1616443634033202,
"step": 9210
},
{
"epoch": 1.770413064361191,
"grad_norm": 0.16424240171909332,
"learning_rate": 1.0679965047211841e-05,
"loss": 3.1608753204345703,
"step": 9215
},
{
"epoch": 1.7713736791546588,
"grad_norm": 0.15550029277801514,
"learning_rate": 1.0591844713447479e-05,
"loss": 3.15771427154541,
"step": 9220
},
{
"epoch": 1.7723342939481268,
"grad_norm": 0.15241944789886475,
"learning_rate": 1.0504076120760413e-05,
"loss": 3.1600357055664063,
"step": 9225
},
{
"epoch": 1.7732949087415946,
"grad_norm": 0.15490137040615082,
"learning_rate": 1.0416659490599871e-05,
"loss": 3.163216400146484,
"step": 9230
},
{
"epoch": 1.7742555235350626,
"grad_norm": 0.17097872495651245,
"learning_rate": 1.0329595043526905e-05,
"loss": 3.164168930053711,
"step": 9235
},
{
"epoch": 1.7752161383285303,
"grad_norm": 0.19601082801818848,
"learning_rate": 1.024288299921408e-05,
"loss": 3.1567230224609375,
"step": 9240
},
{
"epoch": 1.776176753121998,
"grad_norm": 0.1483290195465088,
"learning_rate": 1.015652357644477e-05,
"loss": 3.1643510818481446,
"step": 9245
},
{
"epoch": 1.7771373679154658,
"grad_norm": 0.1415608525276184,
"learning_rate": 1.0070516993112676e-05,
"loss": 3.157514953613281,
"step": 9250
},
{
"epoch": 1.7780979827089336,
"grad_norm": 0.14610423147678375,
"learning_rate": 9.984863466221199e-06,
"loss": 3.1627222061157227,
"step": 9255
},
{
"epoch": 1.7790585975024016,
"grad_norm": 0.17516079545021057,
"learning_rate": 9.899563211883e-06,
"loss": 3.1590530395507814,
"step": 9260
},
{
"epoch": 1.7800192122958693,
"grad_norm": 0.1540340632200241,
"learning_rate": 9.814616445319384e-06,
"loss": 3.1593671798706056,
"step": 9265
},
{
"epoch": 1.7809798270893373,
"grad_norm": 0.18057873845100403,
"learning_rate": 9.730023380859725e-06,
"loss": 3.162105751037598,
"step": 9270
},
{
"epoch": 1.781940441882805,
"grad_norm": 0.17139260470867157,
"learning_rate": 9.645784231941005e-06,
"loss": 3.1630107879638674,
"step": 9275
},
{
"epoch": 1.7829010566762729,
"grad_norm": 0.16848276555538177,
"learning_rate": 9.561899211107244e-06,
"loss": 3.1598880767822264,
"step": 9280
},
{
"epoch": 1.7838616714697406,
"grad_norm": 0.16549763083457947,
"learning_rate": 9.478368530008967e-06,
"loss": 3.160906219482422,
"step": 9285
},
{
"epoch": 1.7848222862632084,
"grad_norm": 0.15802277624607086,
"learning_rate": 9.39519239940264e-06,
"loss": 3.161147689819336,
"step": 9290
},
{
"epoch": 1.7857829010566761,
"grad_norm": 0.1761743128299713,
"learning_rate": 9.31237102915019e-06,
"loss": 3.158751678466797,
"step": 9295
},
{
"epoch": 1.7867435158501441,
"grad_norm": 0.1671990603208542,
"learning_rate": 9.229904628218427e-06,
"loss": 3.1624494552612306,
"step": 9300
},
{
"epoch": 1.7877041306436119,
"grad_norm": 0.19958704710006714,
"learning_rate": 9.14779340467851e-06,
"loss": 3.1618110656738283,
"step": 9305
},
{
"epoch": 1.7886647454370799,
"grad_norm": 0.16844980418682098,
"learning_rate": 9.06603756570552e-06,
"loss": 3.156951141357422,
"step": 9310
},
{
"epoch": 1.7896253602305476,
"grad_norm": 0.14970090985298157,
"learning_rate": 8.984637317577782e-06,
"loss": 3.1624423980712892,
"step": 9315
},
{
"epoch": 1.7905859750240154,
"grad_norm": 0.15226224064826965,
"learning_rate": 8.90359286567654e-06,
"loss": 3.1624835968017577,
"step": 9320
},
{
"epoch": 1.7915465898174832,
"grad_norm": 0.1699807345867157,
"learning_rate": 8.822904414485194e-06,
"loss": 3.1621606826782225,
"step": 9325
},
{
"epoch": 1.792507204610951,
"grad_norm": 0.15113097429275513,
"learning_rate": 8.742572167589008e-06,
"loss": 3.16312255859375,
"step": 9330
},
{
"epoch": 1.7934678194044187,
"grad_norm": 0.16513291001319885,
"learning_rate": 8.662596327674499e-06,
"loss": 3.1624629974365233,
"step": 9335
},
{
"epoch": 1.7944284341978867,
"grad_norm": 0.154737189412117,
"learning_rate": 8.582977096528887e-06,
"loss": 3.1638771057128907,
"step": 9340
},
{
"epoch": 1.7953890489913544,
"grad_norm": 0.18483127653598785,
"learning_rate": 8.503714675039663e-06,
"loss": 3.16489315032959,
"step": 9345
},
{
"epoch": 1.7963496637848224,
"grad_norm": 0.16609624028205872,
"learning_rate": 8.424809263194054e-06,
"loss": 3.1592342376708986,
"step": 9350
},
{
"epoch": 1.7973102785782902,
"grad_norm": 0.13648296892642975,
"learning_rate": 8.346261060078524e-06,
"loss": 3.163716506958008,
"step": 9355
},
{
"epoch": 1.798270893371758,
"grad_norm": 0.16231390833854675,
"learning_rate": 8.268070263878223e-06,
"loss": 3.1625925064086915,
"step": 9360
},
{
"epoch": 1.7992315081652257,
"grad_norm": 0.15581952035427094,
"learning_rate": 8.190237071876576e-06,
"loss": 3.161655235290527,
"step": 9365
},
{
"epoch": 1.8001921229586935,
"grad_norm": 0.15568803250789642,
"learning_rate": 8.11276168045471e-06,
"loss": 3.1620370864868166,
"step": 9370
},
{
"epoch": 1.8011527377521612,
"grad_norm": 0.1662697196006775,
"learning_rate": 8.035644285090958e-06,
"loss": 3.1623306274414062,
"step": 9375
},
{
"epoch": 1.8021133525456292,
"grad_norm": 0.16496489942073822,
"learning_rate": 7.958885080360445e-06,
"loss": 3.1613933563232424,
"step": 9380
},
{
"epoch": 1.803073967339097,
"grad_norm": 0.15053115785121918,
"learning_rate": 7.882484259934497e-06,
"loss": 3.1627952575683596,
"step": 9385
},
{
"epoch": 1.804034582132565,
"grad_norm": 0.14695799350738525,
"learning_rate": 7.806442016580267e-06,
"loss": 3.160305404663086,
"step": 9390
},
{
"epoch": 1.8049951969260327,
"grad_norm": 0.15005071461200714,
"learning_rate": 7.730758542160064e-06,
"loss": 3.1611255645751952,
"step": 9395
},
{
"epoch": 1.8059558117195005,
"grad_norm": 0.15226683020591736,
"learning_rate": 7.655434027631124e-06,
"loss": 3.1629497528076174,
"step": 9400
},
{
"epoch": 1.8069164265129682,
"grad_norm": 0.15546758472919464,
"learning_rate": 7.580468663044903e-06,
"loss": 3.1586400985717775,
"step": 9405
},
{
"epoch": 1.807877041306436,
"grad_norm": 0.1452452838420868,
"learning_rate": 7.505862637546684e-06,
"loss": 3.160877227783203,
"step": 9410
},
{
"epoch": 1.808837656099904,
"grad_norm": 0.13006171584129333,
"learning_rate": 7.43161613937514e-06,
"loss": 3.1579397201538084,
"step": 9415
},
{
"epoch": 1.8097982708933718,
"grad_norm": 0.15625065565109253,
"learning_rate": 7.3577293558618075e-06,
"loss": 3.1598045349121096,
"step": 9420
},
{
"epoch": 1.8107588856868397,
"grad_norm": 0.16364313662052155,
"learning_rate": 7.284202473430612e-06,
"loss": 3.1616336822509767,
"step": 9425
},
{
"epoch": 1.8117195004803075,
"grad_norm": 0.15080556273460388,
"learning_rate": 7.211035677597421e-06,
"loss": 3.1628395080566407,
"step": 9430
},
{
"epoch": 1.8126801152737753,
"grad_norm": 0.1453072428703308,
"learning_rate": 7.138229152969582e-06,
"loss": 3.161026382446289,
"step": 9435
},
{
"epoch": 1.813640730067243,
"grad_norm": 0.1453263759613037,
"learning_rate": 7.065783083245435e-06,
"loss": 3.16085262298584,
"step": 9440
},
{
"epoch": 1.8146013448607108,
"grad_norm": 0.13254213333129883,
"learning_rate": 6.993697651213798e-06,
"loss": 3.1604400634765626,
"step": 9445
},
{
"epoch": 1.8155619596541785,
"grad_norm": 0.1310938000679016,
"learning_rate": 6.921973038753647e-06,
"loss": 3.158857727050781,
"step": 9450
},
{
"epoch": 1.8165225744476465,
"grad_norm": 0.13702644407749176,
"learning_rate": 6.8506094268335255e-06,
"loss": 3.16072998046875,
"step": 9455
},
{
"epoch": 1.8174831892411143,
"grad_norm": 0.15652552247047424,
"learning_rate": 6.779606995511183e-06,
"loss": 3.161703109741211,
"step": 9460
},
{
"epoch": 1.8184438040345823,
"grad_norm": 0.1748208999633789,
"learning_rate": 6.7089659239329685e-06,
"loss": 3.162480926513672,
"step": 9465
},
{
"epoch": 1.81940441882805,
"grad_norm": 0.15371812880039215,
"learning_rate": 6.638686390333592e-06,
"loss": 3.1619606018066406,
"step": 9470
},
{
"epoch": 1.8203650336215178,
"grad_norm": 0.15862195193767548,
"learning_rate": 6.56876857203556e-06,
"loss": 3.1612228393554687,
"step": 9475
},
{
"epoch": 1.8213256484149856,
"grad_norm": 0.14958879351615906,
"learning_rate": 6.499212645448626e-06,
"loss": 3.161460113525391,
"step": 9480
},
{
"epoch": 1.8222862632084533,
"grad_norm": 0.14320014417171478,
"learning_rate": 6.430018786069574e-06,
"loss": 3.1611602783203123,
"step": 9485
},
{
"epoch": 1.823246878001921,
"grad_norm": 0.14477957785129547,
"learning_rate": 6.361187168481618e-06,
"loss": 3.162501907348633,
"step": 9490
},
{
"epoch": 1.824207492795389,
"grad_norm": 0.14906539022922516,
"learning_rate": 6.292717966353955e-06,
"loss": 3.1564708709716798,
"step": 9495
},
{
"epoch": 1.8251681075888568,
"grad_norm": 0.1561676263809204,
"learning_rate": 6.224611352441444e-06,
"loss": 3.164084053039551,
"step": 9500
},
{
"epoch": 1.8261287223823248,
"grad_norm": 0.15524768829345703,
"learning_rate": 6.156867498584028e-06,
"loss": 3.164693832397461,
"step": 9505
},
{
"epoch": 1.8270893371757926,
"grad_norm": 0.13482025265693665,
"learning_rate": 6.089486575706431e-06,
"loss": 3.161751556396484,
"step": 9510
},
{
"epoch": 1.8280499519692603,
"grad_norm": 0.1538052260875702,
"learning_rate": 6.022468753817611e-06,
"loss": 3.162434196472168,
"step": 9515
},
{
"epoch": 1.829010566762728,
"grad_norm": 0.13271762430667877,
"learning_rate": 5.955814202010406e-06,
"loss": 3.1618423461914062,
"step": 9520
},
{
"epoch": 1.8299711815561959,
"grad_norm": 0.14370056986808777,
"learning_rate": 5.88952308846109e-06,
"loss": 3.1616491317749023,
"step": 9525
},
{
"epoch": 1.8309317963496636,
"grad_norm": 0.13553740084171295,
"learning_rate": 5.823595580428969e-06,
"loss": 3.1615522384643553,
"step": 9530
},
{
"epoch": 1.8318924111431316,
"grad_norm": 0.14558495581150055,
"learning_rate": 5.758031844255884e-06,
"loss": 3.163438415527344,
"step": 9535
},
{
"epoch": 1.8328530259365994,
"grad_norm": 0.1292106956243515,
"learning_rate": 5.692832045365858e-06,
"loss": 3.1632017135620116,
"step": 9540
},
{
"epoch": 1.8338136407300674,
"grad_norm": 0.13839663565158844,
"learning_rate": 5.627996348264685e-06,
"loss": 3.1595331192016602,
"step": 9545
},
{
"epoch": 1.8347742555235351,
"grad_norm": 0.13135622441768646,
"learning_rate": 5.563524916539508e-06,
"loss": 3.161990165710449,
"step": 9550
},
{
"epoch": 1.8357348703170029,
"grad_norm": 0.13140968978405,
"learning_rate": 5.499417912858289e-06,
"loss": 3.159449005126953,
"step": 9555
},
{
"epoch": 1.8366954851104706,
"grad_norm": 0.14151158928871155,
"learning_rate": 5.435675498969661e-06,
"loss": 3.1604162216186524,
"step": 9560
},
{
"epoch": 1.8376560999039384,
"grad_norm": 0.1417761594057083,
"learning_rate": 5.372297835702222e-06,
"loss": 3.1592178344726562,
"step": 9565
},
{
"epoch": 1.8386167146974062,
"grad_norm": 0.13075558841228485,
"learning_rate": 5.309285082964343e-06,
"loss": 3.159617233276367,
"step": 9570
},
{
"epoch": 1.8395773294908742,
"grad_norm": 0.13188965618610382,
"learning_rate": 5.24663739974363e-06,
"loss": 3.1601367950439454,
"step": 9575
},
{
"epoch": 1.8405379442843421,
"grad_norm": 0.13191638886928558,
"learning_rate": 5.184354944106661e-06,
"loss": 3.16494197845459,
"step": 9580
},
{
"epoch": 1.84149855907781,
"grad_norm": 0.16160184144973755,
"learning_rate": 5.122437873198415e-06,
"loss": 3.159535217285156,
"step": 9585
},
{
"epoch": 1.8424591738712777,
"grad_norm": 0.13549911975860596,
"learning_rate": 5.060886343242043e-06,
"loss": 3.1587257385253906,
"step": 9590
},
{
"epoch": 1.8434197886647454,
"grad_norm": 0.13970741629600525,
"learning_rate": 4.999700509538368e-06,
"loss": 3.1649681091308595,
"step": 9595
},
{
"epoch": 1.8443804034582132,
"grad_norm": 0.134404718875885,
"learning_rate": 4.938880526465516e-06,
"loss": 3.1567209243774412,
"step": 9600
},
{
"epoch": 1.845341018251681,
"grad_norm": 0.14060495793819427,
"learning_rate": 4.878426547478537e-06,
"loss": 3.161220741271973,
"step": 9605
},
{
"epoch": 1.846301633045149,
"grad_norm": 0.14048533141613007,
"learning_rate": 4.818338725109033e-06,
"loss": 3.1618812561035154,
"step": 9610
},
{
"epoch": 1.8472622478386167,
"grad_norm": 0.13674895465373993,
"learning_rate": 4.758617210964749e-06,
"loss": 3.1565229415893556,
"step": 9615
},
{
"epoch": 1.8482228626320847,
"grad_norm": 0.13577382266521454,
"learning_rate": 4.6992621557292e-06,
"loss": 3.158913230895996,
"step": 9620
},
{
"epoch": 1.8491834774255524,
"grad_norm": 0.13337448239326477,
"learning_rate": 4.640273709161224e-06,
"loss": 3.161669158935547,
"step": 9625
},
{
"epoch": 1.8501440922190202,
"grad_norm": 0.134219229221344,
"learning_rate": 4.5816520200948005e-06,
"loss": 3.160819435119629,
"step": 9630
},
{
"epoch": 1.851104707012488,
"grad_norm": 0.13151714205741882,
"learning_rate": 4.5233972364383975e-06,
"loss": 3.1615375518798827,
"step": 9635
},
{
"epoch": 1.8520653218059557,
"grad_norm": 0.11864073574542999,
"learning_rate": 4.465509505174858e-06,
"loss": 3.1609018325805662,
"step": 9640
},
{
"epoch": 1.8530259365994235,
"grad_norm": 0.14346528053283691,
"learning_rate": 4.407988972360849e-06,
"loss": 3.161561393737793,
"step": 9645
},
{
"epoch": 1.8539865513928915,
"grad_norm": 0.13390202820301056,
"learning_rate": 4.350835783126594e-06,
"loss": 3.1598655700683596,
"step": 9650
},
{
"epoch": 1.8549471661863592,
"grad_norm": 0.12858052551746368,
"learning_rate": 4.2940500816754094e-06,
"loss": 3.160798263549805,
"step": 9655
},
{
"epoch": 1.8559077809798272,
"grad_norm": 0.13176590204238892,
"learning_rate": 4.237632011283471e-06,
"loss": 3.158690643310547,
"step": 9660
},
{
"epoch": 1.856868395773295,
"grad_norm": 0.12532593309879303,
"learning_rate": 4.181581714299359e-06,
"loss": 3.159891128540039,
"step": 9665
},
{
"epoch": 1.8578290105667628,
"grad_norm": 0.11781672388315201,
"learning_rate": 4.12589933214375e-06,
"loss": 3.161539077758789,
"step": 9670
},
{
"epoch": 1.8587896253602305,
"grad_norm": 0.1212988868355751,
"learning_rate": 4.070585005308946e-06,
"loss": 3.16253662109375,
"step": 9675
},
{
"epoch": 1.8597502401536983,
"grad_norm": 0.11885207146406174,
"learning_rate": 4.015638873358707e-06,
"loss": 3.162197303771973,
"step": 9680
},
{
"epoch": 1.860710854947166,
"grad_norm": 0.12261178344488144,
"learning_rate": 3.961061074927752e-06,
"loss": 3.1611923217773437,
"step": 9685
},
{
"epoch": 1.861671469740634,
"grad_norm": 0.13136336207389832,
"learning_rate": 3.906851747721495e-06,
"loss": 3.160177993774414,
"step": 9690
},
{
"epoch": 1.8626320845341018,
"grad_norm": 0.13114266097545624,
"learning_rate": 3.853011028515563e-06,
"loss": 3.15836124420166,
"step": 9695
},
{
"epoch": 1.8635926993275698,
"grad_norm": 0.13357418775558472,
"learning_rate": 3.79953905315567e-06,
"loss": 3.161663627624512,
"step": 9700
},
{
"epoch": 1.8645533141210375,
"grad_norm": 0.11809264123439789,
"learning_rate": 3.746435956557081e-06,
"loss": 3.1624309539794924,
"step": 9705
},
{
"epoch": 1.8655139289145053,
"grad_norm": 0.1218031793832779,
"learning_rate": 3.6937018727043523e-06,
"loss": 3.1634284973144533,
"step": 9710
},
{
"epoch": 1.866474543707973,
"grad_norm": 0.12914207577705383,
"learning_rate": 3.6413369346509848e-06,
"loss": 3.1623727798461916,
"step": 9715
},
{
"epoch": 1.8674351585014408,
"grad_norm": 0.12263166159391403,
"learning_rate": 3.5893412745191085e-06,
"loss": 3.1578353881835937,
"step": 9720
},
{
"epoch": 1.8683957732949086,
"grad_norm": 0.11959100514650345,
"learning_rate": 3.5377150234990824e-06,
"loss": 3.1602123260498045,
"step": 9725
},
{
"epoch": 1.8693563880883766,
"grad_norm": 0.132080078125,
"learning_rate": 3.4864583118492438e-06,
"loss": 3.1630222320556642,
"step": 9730
},
{
"epoch": 1.8703170028818443,
"grad_norm": 0.13228577375411987,
"learning_rate": 3.435571268895526e-06,
"loss": 3.155730438232422,
"step": 9735
},
{
"epoch": 1.8712776176753123,
"grad_norm": 0.12898947298526764,
"learning_rate": 3.3850540230311918e-06,
"loss": 3.1614667892456056,
"step": 9740
},
{
"epoch": 1.87223823246878,
"grad_norm": 0.1316165179014206,
"learning_rate": 3.3349067017163833e-06,
"loss": 3.162592315673828,
"step": 9745
},
{
"epoch": 1.8731988472622478,
"grad_norm": 0.12914599478244781,
"learning_rate": 3.28512943147799e-06,
"loss": 3.160664749145508,
"step": 9750
},
{
"epoch": 1.8741594620557156,
"grad_norm": 0.11813253164291382,
"learning_rate": 3.23572233790913e-06,
"loss": 3.1575859069824217,
"step": 9755
},
{
"epoch": 1.8751200768491834,
"grad_norm": 0.1261204481124878,
"learning_rate": 3.1866855456689878e-06,
"loss": 3.1581016540527345,
"step": 9760
},
{
"epoch": 1.8760806916426513,
"grad_norm": 0.120733842253685,
"learning_rate": 3.138019178482409e-06,
"loss": 3.1588390350341795,
"step": 9765
},
{
"epoch": 1.877041306436119,
"grad_norm": 0.13560213148593903,
"learning_rate": 3.0897233591396562e-06,
"loss": 3.1578319549560545,
"step": 9770
},
{
"epoch": 1.878001921229587,
"grad_norm": 0.11367765814065933,
"learning_rate": 3.0417982094960212e-06,
"loss": 3.1566787719726563,
"step": 9775
},
{
"epoch": 1.8789625360230549,
"grad_norm": 0.11928807944059372,
"learning_rate": 2.994243850471545e-06,
"loss": 3.162850570678711,
"step": 9780
},
{
"epoch": 1.8799231508165226,
"grad_norm": 0.11847756057977676,
"learning_rate": 2.9470604020507994e-06,
"loss": 3.1612308502197264,
"step": 9785
},
{
"epoch": 1.8808837656099904,
"grad_norm": 0.1250849962234497,
"learning_rate": 2.9002479832824553e-06,
"loss": 3.1637001037597656,
"step": 9790
},
{
"epoch": 1.8818443804034581,
"grad_norm": 0.12311087548732758,
"learning_rate": 2.8538067122790164e-06,
"loss": 3.1589778900146483,
"step": 9795
},
{
"epoch": 1.882804995196926,
"grad_norm": 0.11985550820827484,
"learning_rate": 2.8077367062166013e-06,
"loss": 3.1609506607055664,
"step": 9800
},
{
"epoch": 1.8837656099903939,
"grad_norm": 0.12335570156574249,
"learning_rate": 2.762038081334561e-06,
"loss": 3.1584909439086912,
"step": 9805
},
{
"epoch": 1.8847262247838616,
"grad_norm": 0.12690205872058868,
"learning_rate": 2.716710952935214e-06,
"loss": 3.160411071777344,
"step": 9810
},
{
"epoch": 1.8856868395773296,
"grad_norm": 0.12958942353725433,
"learning_rate": 2.6717554353835435e-06,
"loss": 3.16156005859375,
"step": 9815
},
{
"epoch": 1.8866474543707974,
"grad_norm": 0.1530323475599289,
"learning_rate": 2.6271716421069344e-06,
"loss": 3.1630611419677734,
"step": 9820
},
{
"epoch": 1.8876080691642652,
"grad_norm": 0.11445850133895874,
"learning_rate": 2.5829596855948876e-06,
"loss": 3.160757064819336,
"step": 9825
},
{
"epoch": 1.888568683957733,
"grad_norm": 0.12335135042667389,
"learning_rate": 2.539119677398671e-06,
"loss": 3.159043884277344,
"step": 9830
},
{
"epoch": 1.8895292987512007,
"grad_norm": 0.12193870544433594,
"learning_rate": 2.495651728131154e-06,
"loss": 3.1591602325439454,
"step": 9835
},
{
"epoch": 1.8904899135446684,
"grad_norm": 0.11826537549495697,
"learning_rate": 2.452555947466439e-06,
"loss": 3.161988067626953,
"step": 9840
},
{
"epoch": 1.8914505283381364,
"grad_norm": 0.12225425243377686,
"learning_rate": 2.4098324441395645e-06,
"loss": 3.164390754699707,
"step": 9845
},
{
"epoch": 1.8924111431316042,
"grad_norm": 0.11919926851987839,
"learning_rate": 2.367481325946352e-06,
"loss": 3.160402297973633,
"step": 9850
},
{
"epoch": 1.8933717579250722,
"grad_norm": 0.11494944989681244,
"learning_rate": 2.325502699742976e-06,
"loss": 3.159711456298828,
"step": 9855
},
{
"epoch": 1.89433237271854,
"grad_norm": 0.11843105405569077,
"learning_rate": 2.283896671445862e-06,
"loss": 3.159711456298828,
"step": 9860
},
{
"epoch": 1.8952929875120077,
"grad_norm": 0.12161135673522949,
"learning_rate": 2.2426633460312373e-06,
"loss": 3.1607955932617187,
"step": 9865
},
{
"epoch": 1.8962536023054755,
"grad_norm": 0.114869125187397,
"learning_rate": 2.2018028275350652e-06,
"loss": 3.161262321472168,
"step": 9870
},
{
"epoch": 1.8972142170989432,
"grad_norm": 0.129729226231575,
"learning_rate": 2.1613152190525785e-06,
"loss": 3.161954879760742,
"step": 9875
},
{
"epoch": 1.898174831892411,
"grad_norm": 0.12527115643024445,
"learning_rate": 2.1212006227382117e-06,
"loss": 3.159313201904297,
"step": 9880
},
{
"epoch": 1.899135446685879,
"grad_norm": 0.1106216311454773,
"learning_rate": 2.0814591398051527e-06,
"loss": 3.160965347290039,
"step": 9885
},
{
"epoch": 1.9000960614793467,
"grad_norm": 0.11833595484495163,
"learning_rate": 2.0420908705252926e-06,
"loss": 3.160472106933594,
"step": 9890
},
{
"epoch": 1.9010566762728147,
"grad_norm": 0.10949314385652542,
"learning_rate": 2.003095914228775e-06,
"loss": 3.163528060913086,
"step": 9895
},
{
"epoch": 1.9020172910662825,
"grad_norm": 0.1126236766576767,
"learning_rate": 1.9644743693038977e-06,
"loss": 3.1650867462158203,
"step": 9900
},
{
"epoch": 1.9029779058597502,
"grad_norm": 0.1123654693365097,
"learning_rate": 1.9262263331967785e-06,
"loss": 3.1654218673706054,
"step": 9905
},
{
"epoch": 1.903938520653218,
"grad_norm": 0.12411278486251831,
"learning_rate": 1.8883519024111227e-06,
"loss": 3.156511688232422,
"step": 9910
},
{
"epoch": 1.9048991354466858,
"grad_norm": 0.1198534443974495,
"learning_rate": 1.8508511725080388e-06,
"loss": 3.160737991333008,
"step": 9915
},
{
"epoch": 1.9058597502401537,
"grad_norm": 0.11347641795873642,
"learning_rate": 1.8137242381056571e-06,
"loss": 3.1631191253662108,
"step": 9920
},
{
"epoch": 1.9068203650336215,
"grad_norm": 0.11204273998737335,
"learning_rate": 1.7769711928790953e-06,
"loss": 3.158831787109375,
"step": 9925
},
{
"epoch": 1.9077809798270895,
"grad_norm": 0.12282492965459824,
"learning_rate": 1.7405921295600421e-06,
"loss": 3.1593994140625,
"step": 9930
},
{
"epoch": 1.9087415946205573,
"grad_norm": 0.11055224388837814,
"learning_rate": 1.7045871399365918e-06,
"loss": 3.159510040283203,
"step": 9935
},
{
"epoch": 1.909702209414025,
"grad_norm": 0.11742359399795532,
"learning_rate": 1.668956314853026e-06,
"loss": 3.161764144897461,
"step": 9940
},
{
"epoch": 1.9106628242074928,
"grad_norm": 0.10971947759389877,
"learning_rate": 1.6336997442095823e-06,
"loss": 3.1589191436767576,
"step": 9945
},
{
"epoch": 1.9116234390009605,
"grad_norm": 0.11144320666790009,
"learning_rate": 1.5988175169622197e-06,
"loss": 3.1609643936157226,
"step": 9950
},
{
"epoch": 1.9125840537944283,
"grad_norm": 0.11729393154382706,
"learning_rate": 1.5643097211223533e-06,
"loss": 3.1646373748779295,
"step": 9955
},
{
"epoch": 1.9135446685878963,
"grad_norm": 0.11545123904943466,
"learning_rate": 1.5301764437567032e-06,
"loss": 3.1612579345703127,
"step": 9960
},
{
"epoch": 1.914505283381364,
"grad_norm": 0.1158137395977974,
"learning_rate": 1.4964177709870128e-06,
"loss": 3.1609870910644533,
"step": 9965
},
{
"epoch": 1.915465898174832,
"grad_norm": 0.1098775789141655,
"learning_rate": 1.4630337879899146e-06,
"loss": 3.1606449127197265,
"step": 9970
},
{
"epoch": 1.9164265129682998,
"grad_norm": 0.11874253302812576,
"learning_rate": 1.4300245789965803e-06,
"loss": 3.1594825744628907,
"step": 9975
},
{
"epoch": 1.9173871277617676,
"grad_norm": 0.11549612134695053,
"learning_rate": 1.3973902272926384e-06,
"loss": 3.160371780395508,
"step": 9980
},
{
"epoch": 1.9183477425552353,
"grad_norm": 0.10717111080884933,
"learning_rate": 1.3651308152179407e-06,
"loss": 3.1619895935058593,
"step": 9985
},
{
"epoch": 1.919308357348703,
"grad_norm": 0.11006828397512436,
"learning_rate": 1.3332464241662622e-06,
"loss": 3.1590812683105467,
"step": 9990
},
{
"epoch": 1.9202689721421708,
"grad_norm": 0.11138733476400375,
"learning_rate": 1.3017371345851846e-06,
"loss": 3.158272933959961,
"step": 9995
},
{
"epoch": 1.9212295869356388,
"grad_norm": 0.1066170483827591,
"learning_rate": 1.2706030259759137e-06,
"loss": 3.158770942687988,
"step": 10000
},
{
"epoch": 1.9221902017291066,
"grad_norm": 0.11191676557064056,
"learning_rate": 1.2398441768929624e-06,
"loss": 3.1578929901123045,
"step": 10005
},
{
"epoch": 1.9231508165225746,
"grad_norm": 0.11436916887760162,
"learning_rate": 1.2094606649440841e-06,
"loss": 3.1585205078125,
"step": 10010
},
{
"epoch": 1.9241114313160423,
"grad_norm": 0.11085876077413559,
"learning_rate": 1.1794525667899901e-06,
"loss": 3.1594661712646483,
"step": 10015
},
{
"epoch": 1.92507204610951,
"grad_norm": 0.11293391138315201,
"learning_rate": 1.149819958144199e-06,
"loss": 3.1591583251953126,
"step": 10020
},
{
"epoch": 1.9260326609029779,
"grad_norm": 0.10681528598070145,
"learning_rate": 1.1205629137728044e-06,
"loss": 3.160277557373047,
"step": 10025
},
{
"epoch": 1.9269932756964456,
"grad_norm": 0.1138753667473793,
"learning_rate": 1.0916815074943242e-06,
"loss": 3.15789909362793,
"step": 10030
},
{
"epoch": 1.9279538904899134,
"grad_norm": 0.10544762760400772,
"learning_rate": 1.0631758121795508e-06,
"loss": 3.1582752227783204,
"step": 10035
},
{
"epoch": 1.9289145052833814,
"grad_norm": 0.11651872098445892,
"learning_rate": 1.0350458997512356e-06,
"loss": 3.1609378814697267,
"step": 10040
},
{
"epoch": 1.9298751200768491,
"grad_norm": 0.11160679161548615,
"learning_rate": 1.0072918411840548e-06,
"loss": 3.161101722717285,
"step": 10045
},
{
"epoch": 1.9308357348703171,
"grad_norm": 0.11564800143241882,
"learning_rate": 9.799137065043428e-07,
"loss": 3.1580835342407227,
"step": 10050
},
{
"epoch": 1.9317963496637849,
"grad_norm": 0.12115105986595154,
"learning_rate": 9.529115647899599e-07,
"loss": 3.160675811767578,
"step": 10055
},
{
"epoch": 1.9327569644572526,
"grad_norm": 0.11305255442857742,
"learning_rate": 9.262854841700418e-07,
"loss": 3.1640155792236326,
"step": 10060
},
{
"epoch": 1.9337175792507204,
"grad_norm": 0.11431337147951126,
"learning_rate": 9.00035531824933e-07,
"loss": 3.162336730957031,
"step": 10065
},
{
"epoch": 1.9346781940441882,
"grad_norm": 0.10593244433403015,
"learning_rate": 8.741617739859708e-07,
"loss": 3.1616371154785154,
"step": 10070
},
{
"epoch": 1.9356388088376562,
"grad_norm": 0.10147248208522797,
"learning_rate": 8.486642759353013e-07,
"loss": 3.1602996826171874,
"step": 10075
},
{
"epoch": 1.936599423631124,
"grad_norm": 0.11662466824054718,
"learning_rate": 8.23543102005697e-07,
"loss": 3.1570274353027346,
"step": 10080
},
{
"epoch": 1.937560038424592,
"grad_norm": 0.10574644058942795,
"learning_rate": 7.9879831558049e-07,
"loss": 3.156863975524902,
"step": 10085
},
{
"epoch": 1.9385206532180597,
"grad_norm": 0.10977693647146225,
"learning_rate": 7.744299790933217e-07,
"loss": 3.1590835571289064,
"step": 10090
},
{
"epoch": 1.9394812680115274,
"grad_norm": 0.10733279585838318,
"learning_rate": 7.504381540279603e-07,
"loss": 3.1599714279174806,
"step": 10095
},
{
"epoch": 1.9404418828049952,
"grad_norm": 0.1001046821475029,
"learning_rate": 7.268229009182836e-07,
"loss": 3.158583068847656,
"step": 10100
},
{
"epoch": 1.941402497598463,
"grad_norm": 0.10647624731063843,
"learning_rate": 7.035842793479795e-07,
"loss": 3.158983612060547,
"step": 10105
},
{
"epoch": 1.9423631123919307,
"grad_norm": 0.10598830133676529,
"learning_rate": 6.807223479504631e-07,
"loss": 3.1631649017333983,
"step": 10110
},
{
"epoch": 1.9433237271853987,
"grad_norm": 0.10469937324523926,
"learning_rate": 6.582371644087592e-07,
"loss": 3.1588485717773436,
"step": 10115
},
{
"epoch": 1.9442843419788665,
"grad_norm": 0.10287413746118546,
"learning_rate": 6.361287854552865e-07,
"loss": 3.161878967285156,
"step": 10120
},
{
"epoch": 1.9452449567723344,
"grad_norm": 0.10505318641662598,
"learning_rate": 6.143972668717411e-07,
"loss": 3.1627128601074217,
"step": 10125
},
{
"epoch": 1.9462055715658022,
"grad_norm": 0.1040089949965477,
"learning_rate": 5.930426634889796e-07,
"loss": 3.1591156005859373,
"step": 10130
},
{
"epoch": 1.94716618635927,
"grad_norm": 0.10172892361879349,
"learning_rate": 5.720650291868523e-07,
"loss": 3.1622541427612303,
"step": 10135
},
{
"epoch": 1.9481268011527377,
"grad_norm": 0.10863189399242401,
"learning_rate": 5.514644168941041e-07,
"loss": 3.161451721191406,
"step": 10140
},
{
"epoch": 1.9490874159462055,
"grad_norm": 0.10447549819946289,
"learning_rate": 5.312408785881573e-07,
"loss": 3.1583093643188476,
"step": 10145
},
{
"epoch": 1.9500480307396733,
"grad_norm": 0.10685736685991287,
"learning_rate": 5.113944652951118e-07,
"loss": 3.1573591232299805,
"step": 10150
},
{
"epoch": 1.9510086455331412,
"grad_norm": 0.10184746235609055,
"learning_rate": 4.919252270894625e-07,
"loss": 3.155690383911133,
"step": 10155
},
{
"epoch": 1.951969260326609,
"grad_norm": 0.10571934282779694,
"learning_rate": 4.7283321309413145e-07,
"loss": 3.159967613220215,
"step": 10160
},
{
"epoch": 1.952929875120077,
"grad_norm": 0.1015496551990509,
"learning_rate": 4.541184714802193e-07,
"loss": 3.1597368240356447,
"step": 10165
},
{
"epoch": 1.9538904899135447,
"grad_norm": 0.1008259505033493,
"learning_rate": 4.3578104946692114e-07,
"loss": 3.156237030029297,
"step": 10170
},
{
"epoch": 1.9548511047070125,
"grad_norm": 0.10270283371210098,
"learning_rate": 4.1782099332146046e-07,
"loss": 3.162929916381836,
"step": 10175
},
{
"epoch": 1.9558117195004803,
"grad_norm": 0.10830524563789368,
"learning_rate": 4.002383483588723e-07,
"loss": 3.157078170776367,
"step": 10180
},
{
"epoch": 1.956772334293948,
"grad_norm": 0.09829416126012802,
"learning_rate": 3.8303315894200326e-07,
"loss": 3.1587419509887695,
"step": 10185
},
{
"epoch": 1.9577329490874158,
"grad_norm": 0.10378840565681458,
"learning_rate": 3.662054684812954e-07,
"loss": 3.1602014541625976,
"step": 10190
},
{
"epoch": 1.9586935638808838,
"grad_norm": 0.10607574880123138,
"learning_rate": 3.497553194347358e-07,
"loss": 3.1587203979492187,
"step": 10195
},
{
"epoch": 1.9596541786743515,
"grad_norm": 0.10446880012750626,
"learning_rate": 3.3368275330775686e-07,
"loss": 3.160129928588867,
"step": 10200
},
{
"epoch": 1.9606147934678195,
"grad_norm": 0.10566939413547516,
"learning_rate": 3.1798781065305314e-07,
"loss": 3.162166213989258,
"step": 10205
},
{
"epoch": 1.9615754082612873,
"grad_norm": 0.10455495119094849,
"learning_rate": 3.0267053107061456e-07,
"loss": 3.1622539520263673,
"step": 10210
},
{
"epoch": 1.962536023054755,
"grad_norm": 0.10329185426235199,
"learning_rate": 2.877309532074934e-07,
"loss": 3.155780029296875,
"step": 10215
},
{
"epoch": 1.9634966378482228,
"grad_norm": 0.09786950796842575,
"learning_rate": 2.7316911475777083e-07,
"loss": 3.160215377807617,
"step": 10220
},
{
"epoch": 1.9644572526416906,
"grad_norm": 0.10430426150560379,
"learning_rate": 2.589850524624737e-07,
"loss": 3.158669662475586,
"step": 10225
},
{
"epoch": 1.9654178674351583,
"grad_norm": 0.10500769317150116,
"learning_rate": 2.451788021094414e-07,
"loss": 3.1637641906738283,
"step": 10230
},
{
"epoch": 1.9663784822286263,
"grad_norm": 0.09861162304878235,
"learning_rate": 2.317503985332425e-07,
"loss": 3.1611236572265624,
"step": 10235
},
{
"epoch": 1.967339097022094,
"grad_norm": 0.09918565303087234,
"learning_rate": 2.1869987561514147e-07,
"loss": 3.15871639251709,
"step": 10240
},
{
"epoch": 1.968299711815562,
"grad_norm": 0.1011766865849495,
"learning_rate": 2.060272662829321e-07,
"loss": 3.16119499206543,
"step": 10245
},
{
"epoch": 1.9692603266090298,
"grad_norm": 0.11001937091350555,
"learning_rate": 1.9373260251092092e-07,
"loss": 3.1602630615234375,
"step": 10250
},
{
"epoch": 1.9702209414024976,
"grad_norm": 0.10449342429637909,
"learning_rate": 1.8181591531977736e-07,
"loss": 3.160749816894531,
"step": 10255
},
{
"epoch": 1.9711815561959654,
"grad_norm": 0.09981942176818848,
"learning_rate": 1.7027723477656686e-07,
"loss": 3.159575653076172,
"step": 10260
},
{
"epoch": 1.9721421709894331,
"grad_norm": 0.10080744326114655,
"learning_rate": 1.5911658999453458e-07,
"loss": 3.16530704498291,
"step": 10265
},
{
"epoch": 1.973102785782901,
"grad_norm": 0.09806544333696365,
"learning_rate": 1.4833400913313864e-07,
"loss": 3.1608566284179687,
"step": 10270
},
{
"epoch": 1.9740634005763689,
"grad_norm": 0.10161542147397995,
"learning_rate": 1.379295193979335e-07,
"loss": 3.1610599517822267,
"step": 10275
},
{
"epoch": 1.9750240153698368,
"grad_norm": 0.10026436299085617,
"learning_rate": 1.2790314704052008e-07,
"loss": 3.1599550247192383,
"step": 10280
},
{
"epoch": 1.9759846301633046,
"grad_norm": 0.09924504905939102,
"learning_rate": 1.1825491735846237e-07,
"loss": 3.1612056732177733,
"step": 10285
},
{
"epoch": 1.9769452449567724,
"grad_norm": 0.10245411098003387,
"learning_rate": 1.0898485469523766e-07,
"loss": 3.1604587554931642,
"step": 10290
},
{
"epoch": 1.9779058597502401,
"grad_norm": 0.09821955114603043,
"learning_rate": 1.0009298244016972e-07,
"loss": 3.1620269775390626,
"step": 10295
},
{
"epoch": 1.978866474543708,
"grad_norm": 0.09943889081478119,
"learning_rate": 9.15793230283457e-08,
"loss": 3.160098648071289,
"step": 10300
},
{
"epoch": 1.9798270893371757,
"grad_norm": 0.09800975024700165,
"learning_rate": 8.3443897940616e-08,
"loss": 3.1579994201660155,
"step": 10305
},
{
"epoch": 1.9807877041306436,
"grad_norm": 0.0975160002708435,
"learning_rate": 7.568672770349449e-08,
"loss": 3.160677909851074,
"step": 10310
},
{
"epoch": 1.9817483189241114,
"grad_norm": 0.09866613149642944,
"learning_rate": 6.830783188910838e-08,
"loss": 3.15645809173584,
"step": 10315
},
{
"epoch": 1.9827089337175794,
"grad_norm": 0.09947178512811661,
"learning_rate": 6.130722911516505e-08,
"loss": 3.1583110809326174,
"step": 10320
},
{
"epoch": 1.9836695485110472,
"grad_norm": 0.09944622218608856,
"learning_rate": 5.468493704491872e-08,
"loss": 3.1639808654785155,
"step": 10325
},
{
"epoch": 1.984630163304515,
"grad_norm": 0.09704981744289398,
"learning_rate": 4.844097238708711e-08,
"loss": 3.1581130981445313,
"step": 10330
},
{
"epoch": 1.9855907780979827,
"grad_norm": 0.09650828689336777,
"learning_rate": 4.257535089581821e-08,
"loss": 3.156038284301758,
"step": 10335
},
{
"epoch": 1.9865513928914504,
"grad_norm": 0.09726572781801224,
"learning_rate": 3.70880873707069e-08,
"loss": 3.1570310592651367,
"step": 10340
},
{
"epoch": 1.9875120076849182,
"grad_norm": 0.09816308319568634,
"learning_rate": 3.1979195656645084e-08,
"loss": 3.160786819458008,
"step": 10345
},
{
"epoch": 1.9884726224783862,
"grad_norm": 0.09846552461385727,
"learning_rate": 2.72486886439216e-08,
"loss": 3.1611724853515626,
"step": 10350
},
{
"epoch": 1.989433237271854,
"grad_norm": 0.09757398068904877,
"learning_rate": 2.289657826807234e-08,
"loss": 3.1581047058105467,
"step": 10355
},
{
"epoch": 1.990393852065322,
"grad_norm": 0.10126172751188278,
"learning_rate": 1.8922875509930212e-08,
"loss": 3.160317611694336,
"step": 10360
},
{
"epoch": 1.9913544668587897,
"grad_norm": 0.09841560572385788,
"learning_rate": 1.532759039554188e-08,
"loss": 3.158650207519531,
"step": 10365
},
{
"epoch": 1.9923150816522575,
"grad_norm": 0.09537240862846375,
"learning_rate": 1.2110731996201051e-08,
"loss": 3.1590465545654296,
"step": 10370
},
{
"epoch": 1.9932756964457252,
"grad_norm": 0.09531023353338242,
"learning_rate": 9.272308428348586e-09,
"loss": 3.159567642211914,
"step": 10375
},
{
"epoch": 1.994236311239193,
"grad_norm": 0.10250640660524368,
"learning_rate": 6.812326853639083e-09,
"loss": 3.1602352142333983,
"step": 10380
},
{
"epoch": 1.9951969260326607,
"grad_norm": 0.09563779830932617,
"learning_rate": 4.730793478840977e-09,
"loss": 3.157382583618164,
"step": 10385
},
{
"epoch": 1.9961575408261287,
"grad_norm": 0.09357914328575134,
"learning_rate": 3.0277135558864906e-09,
"loss": 3.1614322662353516,
"step": 10390
},
{
"epoch": 1.9971181556195965,
"grad_norm": 0.09247033298015594,
"learning_rate": 1.7030913818050217e-09,
"loss": 3.161212921142578,
"step": 10395
},
{
"epoch": 1.9980787704130645,
"grad_norm": 0.09770449995994568,
"learning_rate": 7.569302987897597e-10,
"loss": 3.1566755294799806,
"step": 10400
},
{
"epoch": 1.9990393852065322,
"grad_norm": 0.0940788984298706,
"learning_rate": 1.8923269406445618e-10,
"loss": 3.157200050354004,
"step": 10405
},
{
"epoch": 2.0,
"grad_norm": 0.10163447260856628,
"learning_rate": 0.0,
"loss": 3.162412643432617,
"step": 10410
}
],
"logging_steps": 5,
"max_steps": 10410,
"num_input_tokens_seen": 0,
"num_train_epochs": 2,
"save_steps": 1000,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 1.9541356371946714e+20,
"train_batch_size": 64,
"trial_name": null,
"trial_params": null
}