ckpt1.6 / trainer_state.json
semran1's picture
Upload folder using huggingface_hub
ff3bcfd verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.01,
"eval_steps": 1000,
"global_step": 1000,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 1e-05,
"grad_norm": 0.37634041905403137,
"learning_rate": 5e-06,
"loss": 0.169,
"loss/crossentropy": 2.8720462918281555,
"loss/hidden": 0.0,
"loss/logits": 0.16897856071591377,
"loss/reg": 4.4040703773498535,
"step": 1
},
{
"epoch": 2e-05,
"grad_norm": 0.35649582743644714,
"learning_rate": 1e-05,
"loss": 0.1696,
"loss/crossentropy": 2.715533673763275,
"loss/hidden": 0.0,
"loss/logits": 0.1695844642817974,
"loss/reg": 4.399058818817139,
"step": 2
},
{
"epoch": 3e-05,
"grad_norm": 0.3591013252735138,
"learning_rate": 1.5e-05,
"loss": 0.1782,
"loss/crossentropy": 2.6291310787200928,
"loss/hidden": 0.0,
"loss/logits": 0.1782267540693283,
"loss/reg": 4.394084930419922,
"step": 3
},
{
"epoch": 4e-05,
"grad_norm": 0.36401960253715515,
"learning_rate": 2e-05,
"loss": 0.1843,
"loss/crossentropy": 2.7142109274864197,
"loss/hidden": 0.0,
"loss/logits": 0.1843317598104477,
"loss/reg": 4.389005661010742,
"step": 4
},
{
"epoch": 5e-05,
"grad_norm": 0.3119131922721863,
"learning_rate": 2.5e-05,
"loss": 0.1625,
"loss/crossentropy": 2.7586326003074646,
"loss/hidden": 0.0,
"loss/logits": 0.1625315584242344,
"loss/reg": 4.3841166496276855,
"step": 5
},
{
"epoch": 6e-05,
"grad_norm": 0.3388400673866272,
"learning_rate": 3e-05,
"loss": 0.1844,
"loss/crossentropy": 2.8104345202445984,
"loss/hidden": 0.0,
"loss/logits": 0.1844346523284912,
"loss/reg": 4.3792877197265625,
"step": 6
},
{
"epoch": 7e-05,
"grad_norm": 0.4783320426940918,
"learning_rate": 3.5e-05,
"loss": 0.1843,
"loss/crossentropy": 2.8321655988693237,
"loss/hidden": 0.0,
"loss/logits": 0.18431555479764938,
"loss/reg": 4.37478494644165,
"step": 7
},
{
"epoch": 8e-05,
"grad_norm": 0.29636114835739136,
"learning_rate": 4e-05,
"loss": 0.1589,
"loss/crossentropy": 2.6809526681900024,
"loss/hidden": 0.0,
"loss/logits": 0.15894119441509247,
"loss/reg": 4.370139122009277,
"step": 8
},
{
"epoch": 9e-05,
"grad_norm": 0.30071625113487244,
"learning_rate": 4.5e-05,
"loss": 0.1657,
"loss/crossentropy": 2.6759764552116394,
"loss/hidden": 0.0,
"loss/logits": 0.16574353352189064,
"loss/reg": 4.365106105804443,
"step": 9
},
{
"epoch": 0.0001,
"grad_norm": 0.28883349895477295,
"learning_rate": 5e-05,
"loss": 0.1572,
"loss/crossentropy": 2.808637499809265,
"loss/hidden": 0.0,
"loss/logits": 0.15719739720225334,
"loss/reg": 4.360220909118652,
"step": 10
},
{
"epoch": 0.00011,
"grad_norm": 0.28243017196655273,
"learning_rate": 5e-05,
"loss": 0.1426,
"loss/crossentropy": 2.72423392534256,
"loss/hidden": 0.0,
"loss/logits": 0.14257685840129852,
"loss/reg": 4.355813503265381,
"step": 11
},
{
"epoch": 0.00012,
"grad_norm": 0.31152331829071045,
"learning_rate": 5e-05,
"loss": 0.147,
"loss/crossentropy": 2.710044264793396,
"loss/hidden": 0.0,
"loss/logits": 0.14701137319207191,
"loss/reg": 4.351265907287598,
"step": 12
},
{
"epoch": 0.00013,
"grad_norm": 0.2739678919315338,
"learning_rate": 5e-05,
"loss": 0.1499,
"loss/crossentropy": 2.7644649744033813,
"loss/hidden": 0.0,
"loss/logits": 0.149860430508852,
"loss/reg": 4.346287727355957,
"step": 13
},
{
"epoch": 0.00014,
"grad_norm": 0.2712353467941284,
"learning_rate": 5e-05,
"loss": 0.1454,
"loss/crossentropy": 2.7370432019233704,
"loss/hidden": 0.0,
"loss/logits": 0.14539287611842155,
"loss/reg": 4.340969085693359,
"step": 14
},
{
"epoch": 0.00015,
"grad_norm": 0.2667863667011261,
"learning_rate": 5e-05,
"loss": 0.1403,
"loss/crossentropy": 2.5638718008995056,
"loss/hidden": 0.0,
"loss/logits": 0.14029696956276894,
"loss/reg": 4.336019515991211,
"step": 15
},
{
"epoch": 0.00016,
"grad_norm": 0.30467212200164795,
"grad_norm_var": 0.0029449483710212204,
"learning_rate": 5e-05,
"loss": 0.1361,
"loss/crossentropy": 2.797445595264435,
"loss/hidden": 0.0,
"loss/logits": 0.13607431203126907,
"loss/reg": 4.330692291259766,
"step": 16
},
{
"epoch": 0.00017,
"grad_norm": 0.2617621421813965,
"grad_norm_var": 0.0029635281595075556,
"learning_rate": 5e-05,
"loss": 0.1443,
"loss/crossentropy": 2.7542406916618347,
"loss/hidden": 0.0,
"loss/logits": 0.14427556470036507,
"loss/reg": 4.325323581695557,
"step": 17
},
{
"epoch": 0.00018,
"grad_norm": 0.28648674488067627,
"grad_norm_var": 0.0028982593896559215,
"learning_rate": 5e-05,
"loss": 0.1396,
"loss/crossentropy": 2.674492835998535,
"loss/hidden": 0.0,
"loss/logits": 0.13961521908640862,
"loss/reg": 4.31995153427124,
"step": 18
},
{
"epoch": 0.00019,
"grad_norm": 0.26269060373306274,
"grad_norm_var": 0.002877724259904054,
"learning_rate": 5e-05,
"loss": 0.141,
"loss/crossentropy": 2.8323662281036377,
"loss/hidden": 0.0,
"loss/logits": 0.14103225618600845,
"loss/reg": 4.315446853637695,
"step": 19
},
{
"epoch": 0.0002,
"grad_norm": 0.2718074321746826,
"grad_norm_var": 0.0026993307095730186,
"learning_rate": 5e-05,
"loss": 0.1314,
"loss/crossentropy": 2.63212913274765,
"loss/hidden": 0.0,
"loss/logits": 0.1313977725803852,
"loss/reg": 4.310704708099365,
"step": 20
},
{
"epoch": 0.00021,
"grad_norm": 0.2430431842803955,
"grad_norm_var": 0.0028911751903802204,
"learning_rate": 5e-05,
"loss": 0.1324,
"loss/crossentropy": 2.664808928966522,
"loss/hidden": 0.0,
"loss/logits": 0.1324238833039999,
"loss/reg": 4.305792808532715,
"step": 21
},
{
"epoch": 0.00022,
"grad_norm": 0.24898661673069,
"grad_norm_var": 0.00288514612507397,
"learning_rate": 5e-05,
"loss": 0.1242,
"loss/crossentropy": 2.7142711877822876,
"loss/hidden": 0.0,
"loss/logits": 0.12423932552337646,
"loss/reg": 4.300712585449219,
"step": 22
},
{
"epoch": 0.00023,
"grad_norm": 0.3123313784599304,
"grad_norm_var": 0.0004523056580034851,
"learning_rate": 5e-05,
"loss": 0.1321,
"loss/crossentropy": 2.7829225063323975,
"loss/hidden": 0.0,
"loss/logits": 0.13212688639760017,
"loss/reg": 4.295501232147217,
"step": 23
},
{
"epoch": 0.00024,
"grad_norm": 0.25187963247299194,
"grad_norm_var": 0.00048027979198491945,
"learning_rate": 5e-05,
"loss": 0.1248,
"loss/crossentropy": 2.692659854888916,
"loss/hidden": 0.0,
"loss/logits": 0.12482420355081558,
"loss/reg": 4.2908830642700195,
"step": 24
},
{
"epoch": 0.00025,
"grad_norm": 0.2151177078485489,
"grad_norm_var": 0.0006726495064564575,
"learning_rate": 5e-05,
"loss": 0.1232,
"loss/crossentropy": 2.738182246685028,
"loss/hidden": 0.0,
"loss/logits": 0.1231868714094162,
"loss/reg": 4.285846710205078,
"step": 25
},
{
"epoch": 0.00026,
"grad_norm": 0.23308518528938293,
"grad_norm_var": 0.0007424884519799501,
"learning_rate": 5e-05,
"loss": 0.1174,
"loss/crossentropy": 2.555102586746216,
"loss/hidden": 0.0,
"loss/logits": 0.11737299524247646,
"loss/reg": 4.281113147735596,
"step": 26
},
{
"epoch": 0.00027,
"grad_norm": 0.24523235857486725,
"grad_norm_var": 0.0007604384721796281,
"learning_rate": 5e-05,
"loss": 0.1201,
"loss/crossentropy": 2.6816893815994263,
"loss/hidden": 0.0,
"loss/logits": 0.12014555744826794,
"loss/reg": 4.2765069007873535,
"step": 27
},
{
"epoch": 0.00028,
"grad_norm": 0.25897473096847534,
"grad_norm_var": 0.0006160828367585275,
"learning_rate": 5e-05,
"loss": 0.1227,
"loss/crossentropy": 2.7505548000335693,
"loss/hidden": 0.0,
"loss/logits": 0.12271320074796677,
"loss/reg": 4.27158260345459,
"step": 28
},
{
"epoch": 0.00029,
"grad_norm": 0.23087331652641296,
"grad_norm_var": 0.0006691547004593392,
"learning_rate": 5e-05,
"loss": 0.1181,
"loss/crossentropy": 2.8483291268348694,
"loss/hidden": 0.0,
"loss/logits": 0.11810225620865822,
"loss/reg": 4.267061233520508,
"step": 29
},
{
"epoch": 0.0003,
"grad_norm": 1.2210192680358887,
"grad_norm_var": 0.05843327221954173,
"learning_rate": 5e-05,
"loss": 0.1723,
"loss/crossentropy": 2.8535077571868896,
"loss/hidden": 0.0,
"loss/logits": 0.17234252952039242,
"loss/reg": 4.262645244598389,
"step": 30
},
{
"epoch": 0.00031,
"grad_norm": 0.2712586224079132,
"grad_norm_var": 0.058402986662709634,
"learning_rate": 5e-05,
"loss": 0.1156,
"loss/crossentropy": 2.6525614261627197,
"loss/hidden": 0.0,
"loss/logits": 0.11560441367328167,
"loss/reg": 4.258092403411865,
"step": 31
},
{
"epoch": 0.00032,
"grad_norm": 0.5226843953132629,
"grad_norm_var": 0.06092943089461011,
"learning_rate": 5e-05,
"loss": 0.1537,
"loss/crossentropy": 2.6228127479553223,
"loss/hidden": 0.0,
"loss/logits": 0.15369537472724915,
"loss/reg": 4.253781318664551,
"step": 32
},
{
"epoch": 0.00033,
"grad_norm": 0.35246461629867554,
"grad_norm_var": 0.06057510886832484,
"learning_rate": 5e-05,
"loss": 0.1216,
"loss/crossentropy": 2.6986429691314697,
"loss/hidden": 0.0,
"loss/logits": 0.12163393199443817,
"loss/reg": 4.249208450317383,
"step": 33
},
{
"epoch": 0.00034,
"grad_norm": 0.2868311405181885,
"grad_norm_var": 0.060572693607631393,
"learning_rate": 5e-05,
"loss": 0.1215,
"loss/crossentropy": 2.7423174381256104,
"loss/hidden": 0.0,
"loss/logits": 0.12151895463466644,
"loss/reg": 4.244677543640137,
"step": 34
},
{
"epoch": 0.00035,
"grad_norm": 0.2556142210960388,
"grad_norm_var": 0.06064807497415105,
"learning_rate": 5e-05,
"loss": 0.1137,
"loss/crossentropy": 2.7171207070350647,
"loss/hidden": 0.0,
"loss/logits": 0.1137176975607872,
"loss/reg": 4.2399797439575195,
"step": 35
},
{
"epoch": 0.00036,
"grad_norm": 0.2783287763595581,
"grad_norm_var": 0.060592460146055585,
"learning_rate": 5e-05,
"loss": 0.1138,
"loss/crossentropy": 2.7394094467163086,
"loss/hidden": 0.0,
"loss/logits": 0.11381806619465351,
"loss/reg": 4.235424041748047,
"step": 36
},
{
"epoch": 0.00037,
"grad_norm": 0.3065175712108612,
"grad_norm_var": 0.06003019540430902,
"learning_rate": 5e-05,
"loss": 0.1235,
"loss/crossentropy": 2.755502223968506,
"loss/hidden": 0.0,
"loss/logits": 0.12348765879869461,
"loss/reg": 4.2310051918029785,
"step": 37
},
{
"epoch": 0.00038,
"grad_norm": 0.26492562890052795,
"grad_norm_var": 0.059845851287469956,
"learning_rate": 5e-05,
"loss": 0.1119,
"loss/crossentropy": 2.8106552362442017,
"loss/hidden": 0.0,
"loss/logits": 0.11191634088754654,
"loss/reg": 4.226707935333252,
"step": 38
},
{
"epoch": 0.00039,
"grad_norm": 0.24673967063426971,
"grad_norm_var": 0.06039341868271975,
"learning_rate": 5e-05,
"loss": 0.1161,
"loss/crossentropy": 2.7490118741989136,
"loss/hidden": 0.0,
"loss/logits": 0.11609707958996296,
"loss/reg": 4.222842216491699,
"step": 39
},
{
"epoch": 0.0004,
"grad_norm": 0.2973298132419586,
"grad_norm_var": 0.05998792869591778,
"learning_rate": 5e-05,
"loss": 0.1124,
"loss/crossentropy": 2.7798808813095093,
"loss/hidden": 0.0,
"loss/logits": 0.11244922317564487,
"loss/reg": 4.218531131744385,
"step": 40
},
{
"epoch": 0.00041,
"grad_norm": 0.7517657279968262,
"grad_norm_var": 0.06884148715130983,
"learning_rate": 5e-05,
"loss": 0.1545,
"loss/crossentropy": 2.749855697154999,
"loss/hidden": 0.0,
"loss/logits": 0.15445118583738804,
"loss/reg": 4.214253902435303,
"step": 41
},
{
"epoch": 0.00042,
"grad_norm": 0.2417730987071991,
"grad_norm_var": 0.06868010027414732,
"learning_rate": 5e-05,
"loss": 0.1099,
"loss/crossentropy": 2.751042366027832,
"loss/hidden": 0.0,
"loss/logits": 0.1099155992269516,
"loss/reg": 4.2101359367370605,
"step": 42
},
{
"epoch": 0.00043,
"grad_norm": 0.2631951570510864,
"grad_norm_var": 0.06838462807177058,
"learning_rate": 5e-05,
"loss": 0.1165,
"loss/crossentropy": 2.7250843048095703,
"loss/hidden": 0.0,
"loss/logits": 0.11648696288466454,
"loss/reg": 4.206397533416748,
"step": 43
},
{
"epoch": 0.00044,
"grad_norm": 0.2518296241760254,
"grad_norm_var": 0.06850134865244813,
"learning_rate": 5e-05,
"loss": 0.1111,
"loss/crossentropy": 2.7153283953666687,
"loss/hidden": 0.0,
"loss/logits": 0.11108221486210823,
"loss/reg": 4.201878547668457,
"step": 44
},
{
"epoch": 0.00045,
"grad_norm": 0.24082158505916595,
"grad_norm_var": 0.06831278207672915,
"learning_rate": 5e-05,
"loss": 0.1177,
"loss/crossentropy": 2.6632660627365112,
"loss/hidden": 0.0,
"loss/logits": 0.11769118346273899,
"loss/reg": 4.19778299331665,
"step": 45
},
{
"epoch": 0.00046,
"grad_norm": 0.260890394449234,
"grad_norm_var": 0.018048092726357542,
"learning_rate": 5e-05,
"loss": 0.1227,
"loss/crossentropy": 2.7315176129341125,
"loss/hidden": 0.0,
"loss/logits": 0.12269957736134529,
"loss/reg": 4.193592071533203,
"step": 46
},
{
"epoch": 0.00047,
"grad_norm": 0.25268790125846863,
"grad_norm_var": 0.018186152495949234,
"learning_rate": 5e-05,
"loss": 0.1178,
"loss/crossentropy": 2.774504065513611,
"loss/hidden": 0.0,
"loss/logits": 0.11776839196681976,
"loss/reg": 4.189169406890869,
"step": 47
},
{
"epoch": 0.00048,
"grad_norm": 0.2759403884410858,
"grad_norm_var": 0.015229396543742831,
"learning_rate": 5e-05,
"loss": 0.1289,
"loss/crossentropy": 2.8515073657035828,
"loss/hidden": 0.0,
"loss/logits": 0.12885254248976707,
"loss/reg": 4.185054779052734,
"step": 48
},
{
"epoch": 0.00049,
"grad_norm": 0.24765782058238983,
"grad_norm_var": 0.015206926335741973,
"learning_rate": 5e-05,
"loss": 0.1256,
"loss/crossentropy": 2.7131593823432922,
"loss/hidden": 0.0,
"loss/logits": 0.1256290916353464,
"loss/reg": 4.1810526847839355,
"step": 49
},
{
"epoch": 0.0005,
"grad_norm": 0.3096969425678253,
"grad_norm_var": 0.015214156358291781,
"learning_rate": 5e-05,
"loss": 0.1401,
"loss/crossentropy": 2.7528311014175415,
"loss/hidden": 0.0,
"loss/logits": 0.14005928859114647,
"loss/reg": 4.176880359649658,
"step": 50
},
{
"epoch": 0.00051,
"grad_norm": 0.33225017786026,
"grad_norm_var": 0.015162352298149247,
"learning_rate": 5e-05,
"loss": 0.1618,
"loss/crossentropy": 2.73341304063797,
"loss/hidden": 0.0,
"loss/logits": 0.1618291698396206,
"loss/reg": 4.173260688781738,
"step": 51
},
{
"epoch": 0.00052,
"grad_norm": 0.33166685700416565,
"grad_norm_var": 0.015176107188209845,
"learning_rate": 5e-05,
"loss": 0.1704,
"loss/crossentropy": 2.824883460998535,
"loss/hidden": 0.0,
"loss/logits": 0.1703827939927578,
"loss/reg": 4.168625354766846,
"step": 52
},
{
"epoch": 0.00053,
"grad_norm": 0.4255874752998352,
"grad_norm_var": 0.01609058098027729,
"learning_rate": 5e-05,
"loss": 0.1856,
"loss/crossentropy": 2.8565452694892883,
"loss/hidden": 0.0,
"loss/logits": 0.18561138212680817,
"loss/reg": 4.164296627044678,
"step": 53
},
{
"epoch": 0.00054,
"grad_norm": 0.33207008242607117,
"grad_norm_var": 0.015949373509081675,
"learning_rate": 5e-05,
"loss": 0.1762,
"loss/crossentropy": 2.7211243510246277,
"loss/hidden": 0.0,
"loss/logits": 0.1762254200875759,
"loss/reg": 4.16010856628418,
"step": 54
},
{
"epoch": 0.00055,
"grad_norm": 0.3105420470237732,
"grad_norm_var": 0.01561146008609899,
"learning_rate": 5e-05,
"loss": 0.172,
"loss/crossentropy": 2.7821205854415894,
"loss/hidden": 0.0,
"loss/logits": 0.17203472182154655,
"loss/reg": 4.155950546264648,
"step": 55
},
{
"epoch": 0.00056,
"grad_norm": 0.3342844247817993,
"grad_norm_var": 0.015583353488029018,
"learning_rate": 5e-05,
"loss": 0.1675,
"loss/crossentropy": 2.783965766429901,
"loss/hidden": 0.0,
"loss/logits": 0.1675088219344616,
"loss/reg": 4.151437759399414,
"step": 56
},
{
"epoch": 0.00057,
"grad_norm": 0.3392151892185211,
"grad_norm_var": 0.0026173613848745727,
"learning_rate": 5e-05,
"loss": 0.1675,
"loss/crossentropy": 2.782883048057556,
"loss/hidden": 0.0,
"loss/logits": 0.16754426062107086,
"loss/reg": 4.1469950675964355,
"step": 57
},
{
"epoch": 0.00058,
"grad_norm": 0.46169230341911316,
"grad_norm_var": 0.004024211017059094,
"learning_rate": 5e-05,
"loss": 0.1828,
"loss/crossentropy": 2.6869139075279236,
"loss/hidden": 0.0,
"loss/logits": 0.18278859555721283,
"loss/reg": 4.142712116241455,
"step": 58
},
{
"epoch": 0.00059,
"grad_norm": 0.35874953866004944,
"grad_norm_var": 0.00399056950783742,
"learning_rate": 5e-05,
"loss": 0.179,
"loss/crossentropy": 2.683705747127533,
"loss/hidden": 0.0,
"loss/logits": 0.17896704375743866,
"loss/reg": 4.138728141784668,
"step": 59
},
{
"epoch": 0.0006,
"grad_norm": 0.3390788435935974,
"grad_norm_var": 0.0037128700604173097,
"learning_rate": 5e-05,
"loss": 0.1824,
"loss/crossentropy": 2.6724974513053894,
"loss/hidden": 0.0,
"loss/logits": 0.18236950412392616,
"loss/reg": 4.1345534324646,
"step": 60
},
{
"epoch": 0.00061,
"grad_norm": 0.3341596722602844,
"grad_norm_var": 0.003246451116369023,
"learning_rate": 5e-05,
"loss": 0.1694,
"loss/crossentropy": 2.956072986125946,
"loss/hidden": 0.0,
"loss/logits": 0.16935936734080315,
"loss/reg": 4.130521774291992,
"step": 61
},
{
"epoch": 0.00062,
"grad_norm": 0.33658263087272644,
"grad_norm_var": 0.0029283974011622186,
"learning_rate": 5e-05,
"loss": 0.1668,
"loss/crossentropy": 2.8409587144851685,
"loss/hidden": 0.0,
"loss/logits": 0.16678539663553238,
"loss/reg": 4.126163005828857,
"step": 62
},
{
"epoch": 0.00063,
"grad_norm": 0.33723217248916626,
"grad_norm_var": 0.0024741312804299983,
"learning_rate": 5e-05,
"loss": 0.1856,
"loss/crossentropy": 2.7388935685157776,
"loss/hidden": 0.0,
"loss/logits": 0.18555545806884766,
"loss/reg": 4.121931552886963,
"step": 63
},
{
"epoch": 0.00064,
"grad_norm": 0.34580445289611816,
"grad_norm_var": 0.0022020224702210757,
"learning_rate": 5e-05,
"loss": 0.1658,
"loss/crossentropy": 2.6729788780212402,
"loss/hidden": 0.0,
"loss/logits": 0.16578427329659462,
"loss/reg": 4.117753982543945,
"step": 64
},
{
"epoch": 0.00065,
"grad_norm": 0.33867374062538147,
"grad_norm_var": 0.0015716415803633144,
"learning_rate": 5e-05,
"loss": 0.1643,
"loss/crossentropy": 2.8432253003120422,
"loss/hidden": 0.0,
"loss/logits": 0.16425132378935814,
"loss/reg": 4.113894939422607,
"step": 65
},
{
"epoch": 0.00066,
"grad_norm": 0.42098623514175415,
"grad_norm_var": 0.001778022217079652,
"learning_rate": 5e-05,
"loss": 0.2155,
"loss/crossentropy": 2.6712504625320435,
"loss/hidden": 0.0,
"loss/logits": 0.21550852805376053,
"loss/reg": 4.10945463180542,
"step": 66
},
{
"epoch": 0.00067,
"grad_norm": 0.35403043031692505,
"grad_norm_var": 0.0017418631675115888,
"learning_rate": 5e-05,
"loss": 0.1798,
"loss/crossentropy": 2.7415149211883545,
"loss/hidden": 0.0,
"loss/logits": 0.1797672137618065,
"loss/reg": 4.1049418449401855,
"step": 67
},
{
"epoch": 0.00068,
"grad_norm": 0.34834232926368713,
"grad_norm_var": 0.0017045350753313,
"learning_rate": 5e-05,
"loss": 0.1783,
"loss/crossentropy": 2.6858341097831726,
"loss/hidden": 0.0,
"loss/logits": 0.17833665013313293,
"loss/reg": 4.100775718688965,
"step": 68
},
{
"epoch": 0.00069,
"grad_norm": 0.3541049063205719,
"grad_norm_var": 0.0013731843169029498,
"learning_rate": 5e-05,
"loss": 0.1744,
"loss/crossentropy": 2.8710713982582092,
"loss/hidden": 0.0,
"loss/logits": 0.1744227409362793,
"loss/reg": 4.096506595611572,
"step": 69
},
{
"epoch": 0.0007,
"grad_norm": 0.3736323118209839,
"grad_norm_var": 0.0013660110363047928,
"learning_rate": 5e-05,
"loss": 0.1994,
"loss/crossentropy": 2.858128011226654,
"loss/hidden": 0.0,
"loss/logits": 0.19940509647130966,
"loss/reg": 4.091678142547607,
"step": 70
},
{
"epoch": 0.00071,
"grad_norm": 0.33025625348091125,
"grad_norm_var": 0.001272272953577754,
"learning_rate": 5e-05,
"loss": 0.1646,
"loss/crossentropy": 2.692229390144348,
"loss/hidden": 0.0,
"loss/logits": 0.16458340734243393,
"loss/reg": 4.087361812591553,
"step": 71
},
{
"epoch": 0.00072,
"grad_norm": 0.6907688975334167,
"grad_norm_var": 0.00815051878013667,
"learning_rate": 5e-05,
"loss": 0.1757,
"loss/crossentropy": 2.886055052280426,
"loss/hidden": 0.0,
"loss/logits": 0.1757429726421833,
"loss/reg": 4.08318567276001,
"step": 72
},
{
"epoch": 0.00073,
"grad_norm": 0.3311053514480591,
"grad_norm_var": 0.008197602515626375,
"learning_rate": 5e-05,
"loss": 0.1682,
"loss/crossentropy": 2.704796850681305,
"loss/hidden": 0.0,
"loss/logits": 0.1681583784520626,
"loss/reg": 4.079033374786377,
"step": 73
},
{
"epoch": 0.00074,
"grad_norm": 0.3336332142353058,
"grad_norm_var": 0.0078012237613196535,
"learning_rate": 5e-05,
"loss": 0.1689,
"loss/crossentropy": 2.6181225776672363,
"loss/hidden": 0.0,
"loss/logits": 0.16892167925834656,
"loss/reg": 4.074740409851074,
"step": 74
},
{
"epoch": 0.00075,
"grad_norm": 0.33766406774520874,
"grad_norm_var": 0.007861895340318493,
"learning_rate": 5e-05,
"loss": 0.1712,
"loss/crossentropy": 2.756729245185852,
"loss/hidden": 0.0,
"loss/logits": 0.17122048512101173,
"loss/reg": 4.070303916931152,
"step": 75
},
{
"epoch": 0.00076,
"grad_norm": 0.34048837423324585,
"grad_norm_var": 0.007856372064757134,
"learning_rate": 5e-05,
"loss": 0.1763,
"loss/crossentropy": 2.62674218416214,
"loss/hidden": 0.0,
"loss/logits": 0.17628077790141106,
"loss/reg": 4.065893650054932,
"step": 76
},
{
"epoch": 0.00077,
"grad_norm": 0.3368911147117615,
"grad_norm_var": 0.007844070912018693,
"learning_rate": 5e-05,
"loss": 0.1789,
"loss/crossentropy": 2.838981509208679,
"loss/hidden": 0.0,
"loss/logits": 0.17892110347747803,
"loss/reg": 4.061193943023682,
"step": 77
},
{
"epoch": 0.00078,
"grad_norm": 0.2983826696872711,
"grad_norm_var": 0.008102358070792626,
"learning_rate": 5e-05,
"loss": 0.151,
"loss/crossentropy": 2.8157095909118652,
"loss/hidden": 0.0,
"loss/logits": 0.15098581835627556,
"loss/reg": 4.05631685256958,
"step": 78
},
{
"epoch": 0.00079,
"grad_norm": 0.34036847949028015,
"grad_norm_var": 0.008090524798600873,
"learning_rate": 5e-05,
"loss": 0.1772,
"loss/crossentropy": 2.742383122444153,
"loss/hidden": 0.0,
"loss/logits": 0.17718595638871193,
"loss/reg": 4.051788330078125,
"step": 79
},
{
"epoch": 0.0008,
"grad_norm": 0.3196929097175598,
"grad_norm_var": 0.008207612908988405,
"learning_rate": 5e-05,
"loss": 0.1574,
"loss/crossentropy": 2.64748877286911,
"loss/hidden": 0.0,
"loss/logits": 0.15740340948104858,
"loss/reg": 4.046438694000244,
"step": 80
},
{
"epoch": 0.00081,
"grad_norm": 0.3145473897457123,
"grad_norm_var": 0.008330494258097032,
"learning_rate": 5e-05,
"loss": 0.1591,
"loss/crossentropy": 2.7640033960342407,
"loss/hidden": 0.0,
"loss/logits": 0.15912048518657684,
"loss/reg": 4.041863441467285,
"step": 81
},
{
"epoch": 0.00082,
"grad_norm": 0.37658828496932983,
"grad_norm_var": 0.008116681055328008,
"learning_rate": 5e-05,
"loss": 0.1783,
"loss/crossentropy": 2.8226330876350403,
"loss/hidden": 0.0,
"loss/logits": 0.17833809927105904,
"loss/reg": 4.0372796058654785,
"step": 82
},
{
"epoch": 0.00083,
"grad_norm": 0.36421865224838257,
"grad_norm_var": 0.00811331907494814,
"learning_rate": 5e-05,
"loss": 0.1636,
"loss/crossentropy": 2.762717604637146,
"loss/hidden": 0.0,
"loss/logits": 0.16359057649970055,
"loss/reg": 4.032177925109863,
"step": 83
},
{
"epoch": 0.00084,
"grad_norm": 0.3138120174407959,
"grad_norm_var": 0.00825034262581384,
"learning_rate": 5e-05,
"loss": 0.1606,
"loss/crossentropy": 2.625426709651947,
"loss/hidden": 0.0,
"loss/logits": 0.16061001271009445,
"loss/reg": 4.027446269989014,
"step": 84
},
{
"epoch": 0.00085,
"grad_norm": 0.34441590309143066,
"grad_norm_var": 0.00826351514204321,
"learning_rate": 5e-05,
"loss": 0.1667,
"loss/crossentropy": 2.8294222950935364,
"loss/hidden": 0.0,
"loss/logits": 0.16673466563224792,
"loss/reg": 4.022748947143555,
"step": 85
},
{
"epoch": 0.00086,
"grad_norm": 0.316683828830719,
"grad_norm_var": 0.00835627592765974,
"learning_rate": 5e-05,
"loss": 0.1564,
"loss/crossentropy": 2.8250383734703064,
"loss/hidden": 0.0,
"loss/logits": 0.1564498096704483,
"loss/reg": 4.017378330230713,
"step": 86
},
{
"epoch": 0.00087,
"grad_norm": 0.3178180456161499,
"grad_norm_var": 0.008407967451986308,
"learning_rate": 5e-05,
"loss": 0.1589,
"loss/crossentropy": 2.831330358982086,
"loss/hidden": 0.0,
"loss/logits": 0.15890633687376976,
"loss/reg": 4.012408256530762,
"step": 87
},
{
"epoch": 0.00088,
"grad_norm": 0.33865824341773987,
"grad_norm_var": 0.00038455914158520567,
"learning_rate": 5e-05,
"loss": 0.1665,
"loss/crossentropy": 2.8202422857284546,
"loss/hidden": 0.0,
"loss/logits": 0.16647625714540482,
"loss/reg": 4.00655460357666,
"step": 88
},
{
"epoch": 0.00089,
"grad_norm": 0.33375900983810425,
"grad_norm_var": 0.00038439593085719167,
"learning_rate": 5e-05,
"loss": 0.1655,
"loss/crossentropy": 2.748092472553253,
"loss/hidden": 0.0,
"loss/logits": 0.1655096672475338,
"loss/reg": 4.000852584838867,
"step": 89
},
{
"epoch": 0.0009,
"grad_norm": 0.41060250997543335,
"grad_norm_var": 0.000761403690223957,
"learning_rate": 5e-05,
"loss": 0.1679,
"loss/crossentropy": 2.8519994616508484,
"loss/hidden": 0.0,
"loss/logits": 0.1679377369582653,
"loss/reg": 3.9966533184051514,
"step": 90
},
{
"epoch": 0.00091,
"grad_norm": 0.3349744379520416,
"grad_norm_var": 0.0007618998964447029,
"learning_rate": 5e-05,
"loss": 0.1663,
"loss/crossentropy": 2.8302014470100403,
"loss/hidden": 0.0,
"loss/logits": 0.16629018262028694,
"loss/reg": 3.9916272163391113,
"step": 91
},
{
"epoch": 0.00092,
"grad_norm": 0.40859073400497437,
"grad_norm_var": 0.0010778266384652254,
"learning_rate": 5e-05,
"loss": 0.1631,
"loss/crossentropy": 2.831357002258301,
"loss/hidden": 0.0,
"loss/logits": 0.16314184293150902,
"loss/reg": 3.9862587451934814,
"step": 92
},
{
"epoch": 0.00093,
"grad_norm": 0.3679395616054535,
"grad_norm_var": 0.0011174436691973562,
"learning_rate": 5e-05,
"loss": 0.1749,
"loss/crossentropy": 2.653463125228882,
"loss/hidden": 0.0,
"loss/logits": 0.17491210997104645,
"loss/reg": 3.9809703826904297,
"step": 93
},
{
"epoch": 0.00094,
"grad_norm": 0.33192068338394165,
"grad_norm_var": 0.000984578674839117,
"learning_rate": 5e-05,
"loss": 0.1689,
"loss/crossentropy": 2.9128816723823547,
"loss/hidden": 0.0,
"loss/logits": 0.16890091821551323,
"loss/reg": 3.9768238067626953,
"step": 94
},
{
"epoch": 0.00095,
"grad_norm": 0.33981162309646606,
"grad_norm_var": 0.000985009641976816,
"learning_rate": 5e-05,
"loss": 0.1651,
"loss/crossentropy": 2.8998738527297974,
"loss/hidden": 0.0,
"loss/logits": 0.1651129573583603,
"loss/reg": 3.9723405838012695,
"step": 95
},
{
"epoch": 0.00096,
"grad_norm": 0.31845277547836304,
"grad_norm_var": 0.0009894353533322537,
"learning_rate": 5e-05,
"loss": 0.1566,
"loss/crossentropy": 2.738618314266205,
"loss/hidden": 0.0,
"loss/logits": 0.15662826597690582,
"loss/reg": 3.9680373668670654,
"step": 96
},
{
"epoch": 0.00097,
"grad_norm": 0.3521839678287506,
"grad_norm_var": 0.0009211371554959176,
"learning_rate": 5e-05,
"loss": 0.1571,
"loss/crossentropy": 2.896687388420105,
"loss/hidden": 0.0,
"loss/logits": 0.15710216015577316,
"loss/reg": 3.964097499847412,
"step": 97
},
{
"epoch": 0.00098,
"grad_norm": 0.41529935598373413,
"grad_norm_var": 0.0011615701056859014,
"learning_rate": 5e-05,
"loss": 0.1761,
"loss/crossentropy": 2.6711183190345764,
"loss/hidden": 0.0,
"loss/logits": 0.176058791577816,
"loss/reg": 3.959585428237915,
"step": 98
},
{
"epoch": 0.00099,
"grad_norm": 0.3406970202922821,
"grad_norm_var": 0.0011533483453351997,
"learning_rate": 5e-05,
"loss": 0.1755,
"loss/crossentropy": 2.762200713157654,
"loss/hidden": 0.0,
"loss/logits": 0.17553818225860596,
"loss/reg": 3.9551267623901367,
"step": 99
},
{
"epoch": 0.001,
"grad_norm": 0.3295409083366394,
"grad_norm_var": 0.0010948026927074712,
"learning_rate": 5e-05,
"loss": 0.1791,
"loss/crossentropy": 2.666721522808075,
"loss/hidden": 0.0,
"loss/logits": 0.17914289608597755,
"loss/reg": 3.9509167671203613,
"step": 100
},
{
"epoch": 0.00101,
"grad_norm": 0.3429720401763916,
"grad_norm_var": 0.001096024238407974,
"learning_rate": 5e-05,
"loss": 0.1793,
"loss/crossentropy": 2.82060843706131,
"loss/hidden": 0.0,
"loss/logits": 0.1792576014995575,
"loss/reg": 3.9469358921051025,
"step": 101
},
{
"epoch": 0.00102,
"grad_norm": 0.3215195834636688,
"grad_norm_var": 0.0010760084324249537,
"learning_rate": 5e-05,
"loss": 0.1632,
"loss/crossentropy": 2.808405876159668,
"loss/hidden": 0.0,
"loss/logits": 0.16316882148385048,
"loss/reg": 3.943436622619629,
"step": 102
},
{
"epoch": 0.00103,
"grad_norm": 0.33158427476882935,
"grad_norm_var": 0.0010282390377130302,
"learning_rate": 5e-05,
"loss": 0.1783,
"loss/crossentropy": 2.8497248888015747,
"loss/hidden": 0.0,
"loss/logits": 0.1783306896686554,
"loss/reg": 3.9394803047180176,
"step": 103
},
{
"epoch": 0.00104,
"grad_norm": 0.3384368121623993,
"grad_norm_var": 0.001028611107856688,
"learning_rate": 5e-05,
"loss": 0.1773,
"loss/crossentropy": 2.8479551672935486,
"loss/hidden": 0.0,
"loss/logits": 0.17731818184256554,
"loss/reg": 3.935678243637085,
"step": 104
},
{
"epoch": 0.00105,
"grad_norm": 0.3275454342365265,
"grad_norm_var": 0.0010454262321925218,
"learning_rate": 5e-05,
"loss": 0.172,
"loss/crossentropy": 2.7240310311317444,
"loss/hidden": 0.0,
"loss/logits": 0.17204875499010086,
"loss/reg": 3.932224750518799,
"step": 105
},
{
"epoch": 0.00106,
"grad_norm": 0.3352244198322296,
"grad_norm_var": 0.0007990449288615142,
"learning_rate": 5e-05,
"loss": 0.1687,
"loss/crossentropy": 2.657980978488922,
"loss/hidden": 0.0,
"loss/logits": 0.16869833320379257,
"loss/reg": 3.92889142036438,
"step": 106
},
{
"epoch": 0.00107,
"grad_norm": 0.3195781409740448,
"grad_norm_var": 0.00083658300653268,
"learning_rate": 5e-05,
"loss": 0.1642,
"loss/crossentropy": 2.7351735830307007,
"loss/hidden": 0.0,
"loss/logits": 0.16421591117978096,
"loss/reg": 3.9260904788970947,
"step": 107
},
{
"epoch": 0.00108,
"grad_norm": 0.3216703534126282,
"grad_norm_var": 0.0005727423089818255,
"learning_rate": 5e-05,
"loss": 0.1611,
"loss/crossentropy": 2.835266649723053,
"loss/hidden": 0.0,
"loss/logits": 0.1611352562904358,
"loss/reg": 3.923356533050537,
"step": 108
},
{
"epoch": 0.00109,
"grad_norm": 0.3534785807132721,
"grad_norm_var": 0.0005312635552543169,
"learning_rate": 5e-05,
"loss": 0.1689,
"loss/crossentropy": 2.8821677565574646,
"loss/hidden": 0.0,
"loss/logits": 0.1689467802643776,
"loss/reg": 3.9208316802978516,
"step": 109
},
{
"epoch": 0.0011,
"grad_norm": 0.33851271867752075,
"grad_norm_var": 0.0005279815580263729,
"learning_rate": 5e-05,
"loss": 0.171,
"loss/crossentropy": 2.7201637029647827,
"loss/hidden": 0.0,
"loss/logits": 0.17095838487148285,
"loss/reg": 3.918743133544922,
"step": 110
},
{
"epoch": 0.00111,
"grad_norm": 0.32998839020729065,
"grad_norm_var": 0.0005331548233647158,
"learning_rate": 5e-05,
"loss": 0.166,
"loss/crossentropy": 2.6836928725242615,
"loss/hidden": 0.0,
"loss/logits": 0.16604754701256752,
"loss/reg": 3.914886951446533,
"step": 111
},
{
"epoch": 0.00112,
"grad_norm": 0.420744925737381,
"grad_norm_var": 0.0009131281860373264,
"learning_rate": 5e-05,
"loss": 0.1738,
"loss/crossentropy": 2.568650722503662,
"loss/hidden": 0.0,
"loss/logits": 0.1737859919667244,
"loss/reg": 3.9106812477111816,
"step": 112
},
{
"epoch": 0.00113,
"grad_norm": 0.3349835276603699,
"grad_norm_var": 0.000914996833659265,
"learning_rate": 5e-05,
"loss": 0.1522,
"loss/crossentropy": 2.7411792278289795,
"loss/hidden": 0.0,
"loss/logits": 0.15223057195544243,
"loss/reg": 3.9069032669067383,
"step": 113
},
{
"epoch": 0.00114,
"grad_norm": 0.34276068210601807,
"grad_norm_var": 0.0005529241807124034,
"learning_rate": 5e-05,
"loss": 0.1567,
"loss/crossentropy": 2.80877947807312,
"loss/hidden": 0.0,
"loss/logits": 0.1567244492471218,
"loss/reg": 3.90332293510437,
"step": 114
},
{
"epoch": 0.00115,
"grad_norm": 0.35375383496284485,
"grad_norm_var": 0.0005659636539689298,
"learning_rate": 5e-05,
"loss": 0.1657,
"loss/crossentropy": 2.698065936565399,
"loss/hidden": 0.0,
"loss/logits": 0.16574329882860184,
"loss/reg": 3.8998756408691406,
"step": 115
},
{
"epoch": 0.00116,
"grad_norm": 0.33278602361679077,
"grad_norm_var": 0.0005620343134485931,
"learning_rate": 5e-05,
"loss": 0.1739,
"loss/crossentropy": 2.7814364433288574,
"loss/hidden": 0.0,
"loss/logits": 0.17385346069931984,
"loss/reg": 3.8964290618896484,
"step": 116
},
{
"epoch": 0.00117,
"grad_norm": 0.35139891505241394,
"grad_norm_var": 0.0005694228893132684,
"learning_rate": 5e-05,
"loss": 0.1701,
"loss/crossentropy": 2.7721198201179504,
"loss/hidden": 0.0,
"loss/logits": 0.1701316274702549,
"loss/reg": 3.8925936222076416,
"step": 117
},
{
"epoch": 0.00118,
"grad_norm": 0.3708522915840149,
"grad_norm_var": 0.0005942298534055627,
"learning_rate": 5e-05,
"loss": 0.1723,
"loss/crossentropy": 2.8753750920295715,
"loss/hidden": 0.0,
"loss/logits": 0.17226434499025345,
"loss/reg": 3.888739824295044,
"step": 118
},
{
"epoch": 0.00119,
"grad_norm": 0.32619452476501465,
"grad_norm_var": 0.0006049363247454272,
"learning_rate": 5e-05,
"loss": 0.1559,
"loss/crossentropy": 2.792622923851013,
"loss/hidden": 0.0,
"loss/logits": 0.15585486218333244,
"loss/reg": 3.8849120140075684,
"step": 119
},
{
"epoch": 0.0012,
"grad_norm": 0.3160404562950134,
"grad_norm_var": 0.0006517621123632485,
"learning_rate": 5e-05,
"loss": 0.1657,
"loss/crossentropy": 2.833389937877655,
"loss/hidden": 0.0,
"loss/logits": 0.16574294120073318,
"loss/reg": 3.8814921379089355,
"step": 120
},
{
"epoch": 0.00121,
"grad_norm": 2.6332755088806152,
"grad_norm_var": 0.328414929277446,
"learning_rate": 5e-05,
"loss": 0.2807,
"loss/crossentropy": 2.960978329181671,
"loss/hidden": 0.0,
"loss/logits": 0.280683059245348,
"loss/reg": 3.8778162002563477,
"step": 121
},
{
"epoch": 0.00122,
"grad_norm": 0.39280807971954346,
"grad_norm_var": 0.32746202761424736,
"learning_rate": 5e-05,
"loss": 0.1791,
"loss/crossentropy": 2.8656354546546936,
"loss/hidden": 0.0,
"loss/logits": 0.17905254289507866,
"loss/reg": 3.8742706775665283,
"step": 122
},
{
"epoch": 0.00123,
"grad_norm": 0.36644095182418823,
"grad_norm_var": 0.3265348837601918,
"learning_rate": 5e-05,
"loss": 0.1765,
"loss/crossentropy": 2.776346266269684,
"loss/hidden": 0.0,
"loss/logits": 0.1764557734131813,
"loss/reg": 3.8701822757720947,
"step": 123
},
{
"epoch": 0.00124,
"grad_norm": 0.39717525243759155,
"grad_norm_var": 0.3251678188828664,
"learning_rate": 5e-05,
"loss": 0.1796,
"loss/crossentropy": 2.9204375743865967,
"loss/hidden": 0.0,
"loss/logits": 0.1796155981719494,
"loss/reg": 3.866316556930542,
"step": 124
},
{
"epoch": 0.00125,
"grad_norm": 0.366623193025589,
"grad_norm_var": 0.3249260727271075,
"learning_rate": 5e-05,
"loss": 0.1654,
"loss/crossentropy": 2.42034849524498,
"loss/hidden": 0.0,
"loss/logits": 0.165392205119133,
"loss/reg": 3.8625807762145996,
"step": 125
},
{
"epoch": 0.00126,
"grad_norm": 0.3638598918914795,
"grad_norm_var": 0.32442588175429127,
"learning_rate": 5e-05,
"loss": 0.1601,
"loss/crossentropy": 2.936553716659546,
"loss/hidden": 0.0,
"loss/logits": 0.16014225035905838,
"loss/reg": 3.8585283756256104,
"step": 126
},
{
"epoch": 0.00127,
"grad_norm": 0.3437521159648895,
"grad_norm_var": 0.3241257586372512,
"learning_rate": 5e-05,
"loss": 0.1603,
"loss/crossentropy": 2.8428520560264587,
"loss/hidden": 0.0,
"loss/logits": 0.16030794754624367,
"loss/reg": 3.854602813720703,
"step": 127
},
{
"epoch": 0.00128,
"grad_norm": 0.3604683578014374,
"grad_norm_var": 0.3249965569466151,
"learning_rate": 5e-05,
"loss": 0.1688,
"loss/crossentropy": 2.717309355735779,
"loss/hidden": 0.0,
"loss/logits": 0.1687549129128456,
"loss/reg": 3.85067081451416,
"step": 128
},
{
"epoch": 0.00129,
"grad_norm": 0.3499651849269867,
"grad_norm_var": 0.32468680185211135,
"learning_rate": 5e-05,
"loss": 0.1748,
"loss/crossentropy": 2.819560468196869,
"loss/hidden": 0.0,
"loss/logits": 0.17475899681448936,
"loss/reg": 3.8467037677764893,
"step": 129
},
{
"epoch": 0.0013,
"grad_norm": 0.3231496512889862,
"grad_norm_var": 0.32511678466571453,
"learning_rate": 5e-05,
"loss": 0.1695,
"loss/crossentropy": 2.5843223929405212,
"loss/hidden": 0.0,
"loss/logits": 0.16951489821076393,
"loss/reg": 3.843282699584961,
"step": 130
},
{
"epoch": 0.00131,
"grad_norm": 0.3588982820510864,
"grad_norm_var": 0.325020330590364,
"learning_rate": 5e-05,
"loss": 0.169,
"loss/crossentropy": 2.725651264190674,
"loss/hidden": 0.0,
"loss/logits": 0.16896183416247368,
"loss/reg": 3.839895725250244,
"step": 131
},
{
"epoch": 0.00132,
"grad_norm": 0.37743306159973145,
"grad_norm_var": 0.32416673149153025,
"learning_rate": 5e-05,
"loss": 0.1833,
"loss/crossentropy": 3.0410608053207397,
"loss/hidden": 0.0,
"loss/logits": 0.1833292953670025,
"loss/reg": 3.8355963230133057,
"step": 132
},
{
"epoch": 0.00133,
"grad_norm": 0.32988330721855164,
"grad_norm_var": 0.32462166470000664,
"learning_rate": 5e-05,
"loss": 0.1654,
"loss/crossentropy": 2.7005507349967957,
"loss/hidden": 0.0,
"loss/logits": 0.1653790920972824,
"loss/reg": 3.831345558166504,
"step": 133
},
{
"epoch": 0.00134,
"grad_norm": 0.35988613963127136,
"grad_norm_var": 0.32481589623167567,
"learning_rate": 5e-05,
"loss": 0.1792,
"loss/crossentropy": 2.7048683762550354,
"loss/hidden": 0.0,
"loss/logits": 0.17917973920702934,
"loss/reg": 3.8267781734466553,
"step": 134
},
{
"epoch": 0.00135,
"grad_norm": 0.32649827003479004,
"grad_norm_var": 0.324808949416691,
"learning_rate": 5e-05,
"loss": 0.1642,
"loss/crossentropy": 2.791461765766144,
"loss/hidden": 0.0,
"loss/logits": 0.16420895606279373,
"loss/reg": 3.8223133087158203,
"step": 135
},
{
"epoch": 0.00136,
"grad_norm": 0.6779212355613708,
"grad_norm_var": 0.32421967313153754,
"learning_rate": 5e-05,
"loss": 0.2361,
"loss/crossentropy": 3.063343107700348,
"loss/hidden": 0.0,
"loss/logits": 0.2360655590891838,
"loss/reg": 3.818582057952881,
"step": 136
},
{
"epoch": 0.00137,
"grad_norm": 0.4217770993709564,
"grad_norm_var": 0.0069040846383882,
"learning_rate": 5e-05,
"loss": 0.1936,
"loss/crossentropy": 2.8291149735450745,
"loss/hidden": 0.0,
"loss/logits": 0.19361505657434464,
"loss/reg": 3.814713716506958,
"step": 137
},
{
"epoch": 0.00138,
"grad_norm": 0.3183574378490448,
"grad_norm_var": 0.0071460434004817905,
"learning_rate": 5e-05,
"loss": 0.1596,
"loss/crossentropy": 2.733646512031555,
"loss/hidden": 0.0,
"loss/logits": 0.15959006920456886,
"loss/reg": 3.8112361431121826,
"step": 138
},
{
"epoch": 0.00139,
"grad_norm": 0.35119444131851196,
"grad_norm_var": 0.007183318962822194,
"learning_rate": 5e-05,
"loss": 0.1706,
"loss/crossentropy": 2.777931809425354,
"loss/hidden": 0.0,
"loss/logits": 0.17056189104914665,
"loss/reg": 3.807130813598633,
"step": 139
},
{
"epoch": 0.0014,
"grad_norm": 0.3381962478160858,
"grad_norm_var": 0.007239536480815012,
"learning_rate": 5e-05,
"loss": 0.1651,
"loss/crossentropy": 2.865752935409546,
"loss/hidden": 0.0,
"loss/logits": 0.16511252894997597,
"loss/reg": 3.8030734062194824,
"step": 140
},
{
"epoch": 0.00141,
"grad_norm": 0.35082533955574036,
"grad_norm_var": 0.007268548808216302,
"learning_rate": 5e-05,
"loss": 0.1608,
"loss/crossentropy": 2.734546184539795,
"loss/hidden": 0.0,
"loss/logits": 0.16080284118652344,
"loss/reg": 3.7996251583099365,
"step": 141
},
{
"epoch": 0.00142,
"grad_norm": 0.4269000291824341,
"grad_norm_var": 0.007448472313405929,
"learning_rate": 5e-05,
"loss": 0.1806,
"loss/crossentropy": 2.9227113127708435,
"loss/hidden": 0.0,
"loss/logits": 0.1805506870150566,
"loss/reg": 3.7955057621002197,
"step": 142
},
{
"epoch": 0.00143,
"grad_norm": 0.3532395660877228,
"grad_norm_var": 0.0074133753520221855,
"learning_rate": 5e-05,
"loss": 0.1588,
"loss/crossentropy": 2.9407125115394592,
"loss/hidden": 0.0,
"loss/logits": 0.15880529955029488,
"loss/reg": 3.791508197784424,
"step": 143
},
{
"epoch": 0.00144,
"grad_norm": 0.3449239134788513,
"grad_norm_var": 0.007461781173789813,
"learning_rate": 5e-05,
"loss": 0.1652,
"loss/crossentropy": 2.8305121660232544,
"loss/hidden": 0.0,
"loss/logits": 0.1652398444712162,
"loss/reg": 3.7871744632720947,
"step": 144
},
{
"epoch": 0.00145,
"grad_norm": 0.3272966742515564,
"grad_norm_var": 0.007571273873210712,
"learning_rate": 5e-05,
"loss": 0.1714,
"loss/crossentropy": 2.876939594745636,
"loss/hidden": 0.0,
"loss/logits": 0.17143940553069115,
"loss/reg": 3.7832887172698975,
"step": 145
},
{
"epoch": 0.00146,
"grad_norm": 0.31960922479629517,
"grad_norm_var": 0.007596131782178968,
"learning_rate": 5e-05,
"loss": 0.1558,
"loss/crossentropy": 2.7597694993019104,
"loss/hidden": 0.0,
"loss/logits": 0.15579523891210556,
"loss/reg": 3.77976393699646,
"step": 146
},
{
"epoch": 0.00147,
"grad_norm": 0.3329758048057556,
"grad_norm_var": 0.007690076208493398,
"learning_rate": 5e-05,
"loss": 0.1602,
"loss/crossentropy": 2.823091506958008,
"loss/hidden": 0.0,
"loss/logits": 0.16016652062535286,
"loss/reg": 3.776364326477051,
"step": 147
},
{
"epoch": 0.00148,
"grad_norm": 0.3245135545730591,
"grad_norm_var": 0.007828939248271782,
"learning_rate": 5e-05,
"loss": 0.1608,
"loss/crossentropy": 2.622242748737335,
"loss/hidden": 0.0,
"loss/logits": 0.16081608831882477,
"loss/reg": 3.7724997997283936,
"step": 148
},
{
"epoch": 0.00149,
"grad_norm": 0.3239537179470062,
"grad_norm_var": 0.007862062788276463,
"learning_rate": 5e-05,
"loss": 0.1559,
"loss/crossentropy": 2.826173484325409,
"loss/hidden": 0.0,
"loss/logits": 0.15591008588671684,
"loss/reg": 3.7680001258850098,
"step": 149
},
{
"epoch": 0.0015,
"grad_norm": 0.3199516534805298,
"grad_norm_var": 0.00800828926831548,
"learning_rate": 5e-05,
"loss": 0.1705,
"loss/crossentropy": 2.73406845331192,
"loss/hidden": 0.0,
"loss/logits": 0.17048393934965134,
"loss/reg": 3.7640268802642822,
"step": 150
},
{
"epoch": 0.00151,
"grad_norm": 0.3810157775878906,
"grad_norm_var": 0.00790594146931481,
"learning_rate": 5e-05,
"loss": 0.1772,
"loss/crossentropy": 2.746786952018738,
"loss/hidden": 0.0,
"loss/logits": 0.17715823650360107,
"loss/reg": 3.760627508163452,
"step": 151
},
{
"epoch": 0.00152,
"grad_norm": 0.33840498328208923,
"grad_norm_var": 0.0011503711202599262,
"learning_rate": 5e-05,
"loss": 0.168,
"loss/crossentropy": 2.7671576738357544,
"loss/hidden": 0.0,
"loss/logits": 0.1679898537695408,
"loss/reg": 3.7571685314178467,
"step": 152
},
{
"epoch": 0.00153,
"grad_norm": 0.35103219747543335,
"grad_norm_var": 0.0007702874374444798,
"learning_rate": 5e-05,
"loss": 0.1683,
"loss/crossentropy": 2.8394588828086853,
"loss/hidden": 0.0,
"loss/logits": 0.1682782731950283,
"loss/reg": 3.7533957958221436,
"step": 153
},
{
"epoch": 0.00154,
"grad_norm": 0.34948527812957764,
"grad_norm_var": 0.000724837481797543,
"learning_rate": 5e-05,
"loss": 0.1551,
"loss/crossentropy": 2.637475073337555,
"loss/hidden": 0.0,
"loss/logits": 0.1551469974219799,
"loss/reg": 3.7496984004974365,
"step": 154
},
{
"epoch": 0.00155,
"grad_norm": 0.32411250472068787,
"grad_norm_var": 0.000751360146424022,
"learning_rate": 5e-05,
"loss": 0.1655,
"loss/crossentropy": 2.65782767534256,
"loss/hidden": 0.0,
"loss/logits": 0.16551653295755386,
"loss/reg": 3.7462174892425537,
"step": 155
},
{
"epoch": 0.00156,
"grad_norm": 0.3659244775772095,
"grad_norm_var": 0.0007773935392291246,
"learning_rate": 5e-05,
"loss": 0.1618,
"loss/crossentropy": 2.8054139614105225,
"loss/hidden": 0.0,
"loss/logits": 0.16182733327150345,
"loss/reg": 3.7422730922698975,
"step": 156
},
{
"epoch": 0.00157,
"grad_norm": 0.3639696538448334,
"grad_norm_var": 0.0007968496539047743,
"learning_rate": 5e-05,
"loss": 0.172,
"loss/crossentropy": 2.643721103668213,
"loss/hidden": 0.0,
"loss/logits": 0.17196981981396675,
"loss/reg": 3.7390189170837402,
"step": 157
},
{
"epoch": 0.00158,
"grad_norm": 0.372111439704895,
"grad_norm_var": 0.0003986384080602812,
"learning_rate": 5e-05,
"loss": 0.1752,
"loss/crossentropy": 2.6860750317573547,
"loss/hidden": 0.0,
"loss/logits": 0.17522458359599113,
"loss/reg": 3.7351748943328857,
"step": 158
},
{
"epoch": 0.00159,
"grad_norm": 0.3412966728210449,
"grad_norm_var": 0.0003916975034196302,
"learning_rate": 5e-05,
"loss": 0.1732,
"loss/crossentropy": 2.7506829500198364,
"loss/hidden": 0.0,
"loss/logits": 0.17320549115538597,
"loss/reg": 3.731645345687866,
"step": 159
},
{
"epoch": 0.0016,
"grad_norm": 0.31508323550224304,
"grad_norm_var": 0.0004378510847698321,
"learning_rate": 5e-05,
"loss": 0.1676,
"loss/crossentropy": 2.672293782234192,
"loss/hidden": 0.0,
"loss/logits": 0.16758090257644653,
"loss/reg": 3.727598190307617,
"step": 160
},
{
"epoch": 0.00161,
"grad_norm": 0.39773106575012207,
"grad_norm_var": 0.0006223116385708161,
"learning_rate": 5e-05,
"loss": 0.1867,
"loss/crossentropy": 2.975751519203186,
"loss/hidden": 0.0,
"loss/logits": 0.18666821345686913,
"loss/reg": 3.7237842082977295,
"step": 161
},
{
"epoch": 0.00162,
"grad_norm": 0.3057797849178314,
"grad_norm_var": 0.0006812186499233134,
"learning_rate": 5e-05,
"loss": 0.1511,
"loss/crossentropy": 2.768982172012329,
"loss/hidden": 0.0,
"loss/logits": 0.15112394466996193,
"loss/reg": 3.7201473712921143,
"step": 162
},
{
"epoch": 0.00163,
"grad_norm": 0.39109617471694946,
"grad_norm_var": 0.0008052929738533592,
"learning_rate": 5e-05,
"loss": 0.1692,
"loss/crossentropy": 2.7556854486465454,
"loss/hidden": 0.0,
"loss/logits": 0.1692204400897026,
"loss/reg": 3.715847969055176,
"step": 163
},
{
"epoch": 0.00164,
"grad_norm": 0.3230038285255432,
"grad_norm_var": 0.0008101312463145642,
"learning_rate": 5e-05,
"loss": 0.158,
"loss/crossentropy": 2.663906216621399,
"loss/hidden": 0.0,
"loss/logits": 0.1579984687268734,
"loss/reg": 3.712200403213501,
"step": 164
},
{
"epoch": 0.00165,
"grad_norm": 0.32820436358451843,
"grad_norm_var": 0.0007977755717131292,
"learning_rate": 5e-05,
"loss": 0.1535,
"loss/crossentropy": 2.7556238174438477,
"loss/hidden": 0.0,
"loss/logits": 0.15348907560110092,
"loss/reg": 3.7093729972839355,
"step": 165
},
{
"epoch": 0.00166,
"grad_norm": 0.37247714400291443,
"grad_norm_var": 0.0007736858685811421,
"learning_rate": 5e-05,
"loss": 0.168,
"loss/crossentropy": 2.623964309692383,
"loss/hidden": 0.0,
"loss/logits": 0.16797634214162827,
"loss/reg": 3.7055835723876953,
"step": 166
},
{
"epoch": 0.00167,
"grad_norm": 0.31921809911727905,
"grad_norm_var": 0.0007674848471050747,
"learning_rate": 5e-05,
"loss": 0.1618,
"loss/crossentropy": 2.6233983039855957,
"loss/hidden": 0.0,
"loss/logits": 0.16180693730711937,
"loss/reg": 3.7018704414367676,
"step": 167
},
{
"epoch": 0.00168,
"grad_norm": 0.41518375277519226,
"grad_norm_var": 0.0010434978692974088,
"learning_rate": 5e-05,
"loss": 0.1842,
"loss/crossentropy": 2.794585347175598,
"loss/hidden": 0.0,
"loss/logits": 0.18423354998230934,
"loss/reg": 3.6984987258911133,
"step": 168
},
{
"epoch": 0.00169,
"grad_norm": 0.3530808985233307,
"grad_norm_var": 0.0010434324942960296,
"learning_rate": 5e-05,
"loss": 0.1818,
"loss/crossentropy": 2.725895941257477,
"loss/hidden": 0.0,
"loss/logits": 0.18175217881798744,
"loss/reg": 3.6949737071990967,
"step": 169
},
{
"epoch": 0.0017,
"grad_norm": 0.35729339718818665,
"grad_norm_var": 0.001044250197534243,
"learning_rate": 5e-05,
"loss": 0.1758,
"loss/crossentropy": 2.8144423365592957,
"loss/hidden": 0.0,
"loss/logits": 0.1758369542658329,
"loss/reg": 3.6909079551696777,
"step": 170
},
{
"epoch": 0.00171,
"grad_norm": 0.3258056044578552,
"grad_norm_var": 0.0010379424391956011,
"learning_rate": 5e-05,
"loss": 0.1615,
"loss/crossentropy": 2.6860609650611877,
"loss/hidden": 0.0,
"loss/logits": 0.1615053378045559,
"loss/reg": 3.6872336864471436,
"step": 171
},
{
"epoch": 0.00172,
"grad_norm": 0.3320024907588959,
"grad_norm_var": 0.0010511954351829684,
"learning_rate": 5e-05,
"loss": 0.1669,
"loss/crossentropy": 2.7618680596351624,
"loss/hidden": 0.0,
"loss/logits": 0.16686224937438965,
"loss/reg": 3.684033155441284,
"step": 172
},
{
"epoch": 0.00173,
"grad_norm": 0.32370057702064514,
"grad_norm_var": 0.001082015111668518,
"learning_rate": 5e-05,
"loss": 0.1568,
"loss/crossentropy": 2.8911356329917908,
"loss/hidden": 0.0,
"loss/logits": 0.15675026923418045,
"loss/reg": 3.6797101497650146,
"step": 173
},
{
"epoch": 0.00174,
"grad_norm": 0.3590388298034668,
"grad_norm_var": 0.0010512214971074684,
"learning_rate": 5e-05,
"loss": 0.1608,
"loss/crossentropy": 2.894763946533203,
"loss/hidden": 0.0,
"loss/logits": 0.16077794507145882,
"loss/reg": 3.676694393157959,
"step": 174
},
{
"epoch": 0.00175,
"grad_norm": 0.362693190574646,
"grad_norm_var": 0.0010621381304175893,
"learning_rate": 5e-05,
"loss": 0.181,
"loss/crossentropy": 2.9355967044830322,
"loss/hidden": 0.0,
"loss/logits": 0.18095535412430763,
"loss/reg": 3.6728715896606445,
"step": 175
},
{
"epoch": 0.00176,
"grad_norm": 0.3421201705932617,
"grad_norm_var": 0.0009861454944628978,
"learning_rate": 5e-05,
"loss": 0.1752,
"loss/crossentropy": 2.771928548812866,
"loss/hidden": 0.0,
"loss/logits": 0.17522436380386353,
"loss/reg": 3.6698157787323,
"step": 176
},
{
"epoch": 0.00177,
"grad_norm": 0.3921768069267273,
"grad_norm_var": 0.0009531156716223066,
"learning_rate": 5e-05,
"loss": 0.1682,
"loss/crossentropy": 2.9020140171051025,
"loss/hidden": 0.0,
"loss/logits": 0.16816864535212517,
"loss/reg": 3.6669130325317383,
"step": 177
},
{
"epoch": 0.00178,
"grad_norm": 0.414460688829422,
"grad_norm_var": 0.0010479472090343092,
"learning_rate": 5e-05,
"loss": 0.1652,
"loss/crossentropy": 2.871070384979248,
"loss/hidden": 0.0,
"loss/logits": 0.1651761755347252,
"loss/reg": 3.6637353897094727,
"step": 178
},
{
"epoch": 0.00179,
"grad_norm": 0.37821123003959656,
"grad_norm_var": 0.0009996989224075473,
"learning_rate": 5e-05,
"loss": 0.1618,
"loss/crossentropy": 2.8318552374839783,
"loss/hidden": 0.0,
"loss/logits": 0.16177014261484146,
"loss/reg": 3.6601526737213135,
"step": 179
},
{
"epoch": 0.0018,
"grad_norm": 0.33756861090660095,
"grad_norm_var": 0.0009485554235717804,
"learning_rate": 5e-05,
"loss": 0.164,
"loss/crossentropy": 2.7179840803146362,
"loss/hidden": 0.0,
"loss/logits": 0.16402991488575935,
"loss/reg": 3.655977725982666,
"step": 180
},
{
"epoch": 0.00181,
"grad_norm": 0.3508152663707733,
"grad_norm_var": 0.0008934631549546879,
"learning_rate": 5e-05,
"loss": 0.1824,
"loss/crossentropy": 2.655538856983185,
"loss/hidden": 0.0,
"loss/logits": 0.18240001425147057,
"loss/reg": 3.6522390842437744,
"step": 181
},
{
"epoch": 0.00182,
"grad_norm": 0.4800889194011688,
"grad_norm_var": 0.0018179163356779901,
"learning_rate": 5e-05,
"loss": 0.1773,
"loss/crossentropy": 2.9170504808425903,
"loss/hidden": 0.0,
"loss/logits": 0.17730093747377396,
"loss/reg": 3.648420810699463,
"step": 182
},
{
"epoch": 0.00183,
"grad_norm": 0.32715606689453125,
"grad_norm_var": 0.0017731703957083382,
"learning_rate": 5e-05,
"loss": 0.1599,
"loss/crossentropy": 2.6978230476379395,
"loss/hidden": 0.0,
"loss/logits": 0.15988203510642052,
"loss/reg": 3.644439458847046,
"step": 183
},
{
"epoch": 0.00184,
"grad_norm": 0.3219493329524994,
"grad_norm_var": 0.0017014689354580615,
"learning_rate": 5e-05,
"loss": 0.1588,
"loss/crossentropy": 2.772395610809326,
"loss/hidden": 0.0,
"loss/logits": 0.158803328871727,
"loss/reg": 3.6410605907440186,
"step": 184
},
{
"epoch": 0.00185,
"grad_norm": 0.3204100728034973,
"grad_norm_var": 0.0017978203455529696,
"learning_rate": 5e-05,
"loss": 0.1595,
"loss/crossentropy": 2.7290788292884827,
"loss/hidden": 0.0,
"loss/logits": 0.15948805212974548,
"loss/reg": 3.637272596359253,
"step": 185
},
{
"epoch": 0.00186,
"grad_norm": 0.34646865725517273,
"grad_norm_var": 0.0018059373173852718,
"learning_rate": 5e-05,
"loss": 0.1723,
"loss/crossentropy": 2.68435937166214,
"loss/hidden": 0.0,
"loss/logits": 0.17226089164614677,
"loss/reg": 3.6334545612335205,
"step": 186
},
{
"epoch": 0.00187,
"grad_norm": 0.35515356063842773,
"grad_norm_var": 0.001737051018651666,
"learning_rate": 5e-05,
"loss": 0.1656,
"loss/crossentropy": 2.8159299492836,
"loss/hidden": 0.0,
"loss/logits": 0.16557194665074348,
"loss/reg": 3.629215717315674,
"step": 187
},
{
"epoch": 0.00188,
"grad_norm": 0.31605055928230286,
"grad_norm_var": 0.0018103786054489293,
"learning_rate": 5e-05,
"loss": 0.1587,
"loss/crossentropy": 2.737620174884796,
"loss/hidden": 0.0,
"loss/logits": 0.15867746248841286,
"loss/reg": 3.6252663135528564,
"step": 188
},
{
"epoch": 0.00189,
"grad_norm": 0.3383916914463043,
"grad_norm_var": 0.0017566740185558556,
"learning_rate": 5e-05,
"loss": 0.1621,
"loss/crossentropy": 2.7829577326774597,
"loss/hidden": 0.0,
"loss/logits": 0.162098228931427,
"loss/reg": 3.621067523956299,
"step": 189
},
{
"epoch": 0.0019,
"grad_norm": 0.4556836783885956,
"grad_norm_var": 0.0023419423247556044,
"learning_rate": 5e-05,
"loss": 0.1687,
"loss/crossentropy": 2.9624626636505127,
"loss/hidden": 0.0,
"loss/logits": 0.1686898171901703,
"loss/reg": 3.6163265705108643,
"step": 190
},
{
"epoch": 0.00191,
"grad_norm": 0.3975931406021118,
"grad_norm_var": 0.0024075083289669527,
"learning_rate": 5e-05,
"loss": 0.155,
"loss/crossentropy": 2.731001079082489,
"loss/hidden": 0.0,
"loss/logits": 0.15501929074525833,
"loss/reg": 3.6121585369110107,
"step": 191
},
{
"epoch": 0.00192,
"grad_norm": 0.37328633666038513,
"grad_norm_var": 0.002364231645687964,
"learning_rate": 5e-05,
"loss": 0.1683,
"loss/crossentropy": 2.754942238330841,
"loss/hidden": 0.0,
"loss/logits": 0.16829831898212433,
"loss/reg": 3.6073873043060303,
"step": 192
},
{
"epoch": 0.00193,
"grad_norm": 0.3342723250389099,
"grad_norm_var": 0.0023955576435807737,
"learning_rate": 5e-05,
"loss": 0.1663,
"loss/crossentropy": 2.7424720525741577,
"loss/hidden": 0.0,
"loss/logits": 0.16633369401097298,
"loss/reg": 3.6036906242370605,
"step": 193
},
{
"epoch": 0.00194,
"grad_norm": 0.38286155462265015,
"grad_norm_var": 0.002251566346172081,
"learning_rate": 5e-05,
"loss": 0.1652,
"loss/crossentropy": 2.9778133630752563,
"loss/hidden": 0.0,
"loss/logits": 0.16522743180394173,
"loss/reg": 3.600316286087036,
"step": 194
},
{
"epoch": 0.00195,
"grad_norm": 0.36051952838897705,
"grad_norm_var": 0.0022364206403587715,
"learning_rate": 5e-05,
"loss": 0.1772,
"loss/crossentropy": 2.6842609643936157,
"loss/hidden": 0.0,
"loss/logits": 0.1771794743835926,
"loss/reg": 3.596491813659668,
"step": 195
},
{
"epoch": 0.00196,
"grad_norm": 0.3526027202606201,
"grad_norm_var": 0.0022007878333510996,
"learning_rate": 5e-05,
"loss": 0.1561,
"loss/crossentropy": 2.7837477922439575,
"loss/hidden": 0.0,
"loss/logits": 0.15605639293789864,
"loss/reg": 3.5926032066345215,
"step": 196
},
{
"epoch": 0.00197,
"grad_norm": 0.35895583033561707,
"grad_norm_var": 0.002191344445433652,
"learning_rate": 5e-05,
"loss": 0.1801,
"loss/crossentropy": 2.85478812456131,
"loss/hidden": 0.0,
"loss/logits": 0.1800978109240532,
"loss/reg": 3.589280843734741,
"step": 197
},
{
"epoch": 0.00198,
"grad_norm": 0.3372839689254761,
"grad_norm_var": 0.0012524713862786308,
"learning_rate": 5e-05,
"loss": 0.1571,
"loss/crossentropy": 2.805725872516632,
"loss/hidden": 0.0,
"loss/logits": 0.15710647776722908,
"loss/reg": 3.5851662158966064,
"step": 198
},
{
"epoch": 0.00199,
"grad_norm": 0.33652499318122864,
"grad_norm_var": 0.0012232813247675149,
"learning_rate": 5e-05,
"loss": 0.1652,
"loss/crossentropy": 2.657254457473755,
"loss/hidden": 0.0,
"loss/logits": 0.1651643067598343,
"loss/reg": 3.581798553466797,
"step": 199
},
{
"epoch": 0.002,
"grad_norm": 0.36757001280784607,
"grad_norm_var": 0.001149275638629573,
"learning_rate": 5e-05,
"loss": 0.1756,
"loss/crossentropy": 2.7496553659439087,
"loss/hidden": 0.0,
"loss/logits": 0.17555997148156166,
"loss/reg": 3.577878475189209,
"step": 200
},
{
"epoch": 0.00201,
"grad_norm": 0.4317435324192047,
"grad_norm_var": 0.0013607474972908151,
"learning_rate": 5e-05,
"loss": 0.1643,
"loss/crossentropy": 3.168861448764801,
"loss/hidden": 0.0,
"loss/logits": 0.164311021566391,
"loss/reg": 3.5741024017333984,
"step": 201
},
{
"epoch": 0.00202,
"grad_norm": 0.3569833040237427,
"grad_norm_var": 0.0013412425012825579,
"learning_rate": 5e-05,
"loss": 0.1778,
"loss/crossentropy": 2.7941558957099915,
"loss/hidden": 0.0,
"loss/logits": 0.17778108268976212,
"loss/reg": 3.5706787109375,
"step": 202
},
{
"epoch": 0.00203,
"grad_norm": 0.31648150086402893,
"grad_norm_var": 0.0014904716039333447,
"learning_rate": 5e-05,
"loss": 0.156,
"loss/crossentropy": 2.872058689594269,
"loss/hidden": 0.0,
"loss/logits": 0.1559964008629322,
"loss/reg": 3.5671305656433105,
"step": 203
},
{
"epoch": 0.00204,
"grad_norm": 0.32686129212379456,
"grad_norm_var": 0.0014293085106024154,
"learning_rate": 5e-05,
"loss": 0.1593,
"loss/crossentropy": 2.7316592931747437,
"loss/hidden": 0.0,
"loss/logits": 0.15925980731844902,
"loss/reg": 3.5632758140563965,
"step": 204
},
{
"epoch": 0.00205,
"grad_norm": 0.3191937506198883,
"grad_norm_var": 0.001518472211395964,
"learning_rate": 5e-05,
"loss": 0.1527,
"loss/crossentropy": 2.7802085876464844,
"loss/hidden": 0.0,
"loss/logits": 0.15268265083432198,
"loss/reg": 3.559633493423462,
"step": 205
},
{
"epoch": 0.00206,
"grad_norm": 0.34924882650375366,
"grad_norm_var": 0.0009115629505157467,
"learning_rate": 5e-05,
"loss": 0.1773,
"loss/crossentropy": 2.792604923248291,
"loss/hidden": 0.0,
"loss/logits": 0.17729893326759338,
"loss/reg": 3.555882453918457,
"step": 206
},
{
"epoch": 0.00207,
"grad_norm": 0.38204553723335266,
"grad_norm_var": 0.0008412229229646054,
"learning_rate": 5e-05,
"loss": 0.1735,
"loss/crossentropy": 2.729912281036377,
"loss/hidden": 0.0,
"loss/logits": 0.17347190529108047,
"loss/reg": 3.551867723464966,
"step": 207
},
{
"epoch": 0.00208,
"grad_norm": 0.316631555557251,
"grad_norm_var": 0.0009067368526577339,
"learning_rate": 5e-05,
"loss": 0.1521,
"loss/crossentropy": 2.6910020112991333,
"loss/hidden": 0.0,
"loss/logits": 0.15211007744073868,
"loss/reg": 3.547140598297119,
"step": 208
},
{
"epoch": 0.00209,
"grad_norm": 0.3024788200855255,
"grad_norm_var": 0.0010444754089082963,
"learning_rate": 5e-05,
"loss": 0.1534,
"loss/crossentropy": 2.6174367666244507,
"loss/hidden": 0.0,
"loss/logits": 0.15340904891490936,
"loss/reg": 3.5430798530578613,
"step": 209
},
{
"epoch": 0.0021,
"grad_norm": 0.31879743933677673,
"grad_norm_var": 0.0010192142441715734,
"learning_rate": 5e-05,
"loss": 0.1644,
"loss/crossentropy": 2.6434658765792847,
"loss/hidden": 0.0,
"loss/logits": 0.164449330419302,
"loss/reg": 3.539293050765991,
"step": 210
},
{
"epoch": 0.00211,
"grad_norm": 0.37038934230804443,
"grad_norm_var": 0.0010445807718520773,
"learning_rate": 5e-05,
"loss": 0.1618,
"loss/crossentropy": 2.7187950015068054,
"loss/hidden": 0.0,
"loss/logits": 0.16179471090435982,
"loss/reg": 3.5359723567962646,
"step": 211
},
{
"epoch": 0.00212,
"grad_norm": 0.3256055414676666,
"grad_norm_var": 0.0010681195543044476,
"learning_rate": 5e-05,
"loss": 0.1634,
"loss/crossentropy": 2.6802476048469543,
"loss/hidden": 0.0,
"loss/logits": 0.16339639574289322,
"loss/reg": 3.5320651531219482,
"step": 212
},
{
"epoch": 0.00213,
"grad_norm": 0.363210529088974,
"grad_norm_var": 0.0010772816324646883,
"learning_rate": 5e-05,
"loss": 0.1682,
"loss/crossentropy": 2.925456941127777,
"loss/hidden": 0.0,
"loss/logits": 0.16816257312893867,
"loss/reg": 3.527592420578003,
"step": 213
},
{
"epoch": 0.00214,
"grad_norm": 0.3341169059276581,
"grad_norm_var": 0.0010811945233913268,
"learning_rate": 5e-05,
"loss": 0.169,
"loss/crossentropy": 2.8775156140327454,
"loss/hidden": 0.0,
"loss/logits": 0.1689641959965229,
"loss/reg": 3.5241305828094482,
"step": 214
},
{
"epoch": 0.00215,
"grad_norm": 0.7971848249435425,
"grad_norm_var": 0.013831743286372744,
"learning_rate": 5e-05,
"loss": 0.1898,
"loss/crossentropy": 2.769020676612854,
"loss/hidden": 0.0,
"loss/logits": 0.18977811932563782,
"loss/reg": 3.5197558403015137,
"step": 215
},
{
"epoch": 0.00216,
"grad_norm": 0.3044687807559967,
"grad_norm_var": 0.014131832632900828,
"learning_rate": 5e-05,
"loss": 0.1467,
"loss/crossentropy": 2.792181670665741,
"loss/hidden": 0.0,
"loss/logits": 0.1466773971915245,
"loss/reg": 3.516072988510132,
"step": 216
},
{
"epoch": 0.00217,
"grad_norm": 0.3434732258319855,
"grad_norm_var": 0.013888774653188173,
"learning_rate": 5e-05,
"loss": 0.1698,
"loss/crossentropy": 2.577077627182007,
"loss/hidden": 0.0,
"loss/logits": 0.16977669671177864,
"loss/reg": 3.512517213821411,
"step": 217
},
{
"epoch": 0.00218,
"grad_norm": 0.37019920349121094,
"grad_norm_var": 0.013886977393692842,
"learning_rate": 5e-05,
"loss": 0.1943,
"loss/crossentropy": 2.722847878932953,
"loss/hidden": 0.0,
"loss/logits": 0.19428952783346176,
"loss/reg": 3.5091969966888428,
"step": 218
},
{
"epoch": 0.00219,
"grad_norm": 0.31637635827064514,
"grad_norm_var": 0.013887658605223226,
"learning_rate": 5e-05,
"loss": 0.1547,
"loss/crossentropy": 2.787532150745392,
"loss/hidden": 0.0,
"loss/logits": 0.1546883024275303,
"loss/reg": 3.5059781074523926,
"step": 219
},
{
"epoch": 0.0022,
"grad_norm": 0.368344783782959,
"grad_norm_var": 0.013784165910995568,
"learning_rate": 5e-05,
"loss": 0.1773,
"loss/crossentropy": 2.7095659971237183,
"loss/hidden": 0.0,
"loss/logits": 0.1773015893995762,
"loss/reg": 3.502683162689209,
"step": 220
},
{
"epoch": 0.00221,
"grad_norm": 0.3447912037372589,
"grad_norm_var": 0.013659872247631084,
"learning_rate": 5e-05,
"loss": 0.1688,
"loss/crossentropy": 2.7072474360466003,
"loss/hidden": 0.0,
"loss/logits": 0.1687602400779724,
"loss/reg": 3.4986109733581543,
"step": 221
},
{
"epoch": 0.00222,
"grad_norm": 0.3812227249145508,
"grad_norm_var": 0.013638668912457892,
"learning_rate": 5e-05,
"loss": 0.1811,
"loss/crossentropy": 2.8128660917282104,
"loss/hidden": 0.0,
"loss/logits": 0.18113631010055542,
"loss/reg": 3.4947147369384766,
"step": 222
},
{
"epoch": 0.00223,
"grad_norm": 0.339374303817749,
"grad_norm_var": 0.013690814024359154,
"learning_rate": 5e-05,
"loss": 0.1716,
"loss/crossentropy": 2.885101020336151,
"loss/hidden": 0.0,
"loss/logits": 0.17156245186924934,
"loss/reg": 3.4905753135681152,
"step": 223
},
{
"epoch": 0.00224,
"grad_norm": 0.3169143497943878,
"grad_norm_var": 0.013688861707923295,
"learning_rate": 5e-05,
"loss": 0.1589,
"loss/crossentropy": 2.6434147357940674,
"loss/hidden": 0.0,
"loss/logits": 0.15887855738401413,
"loss/reg": 3.486919641494751,
"step": 224
},
{
"epoch": 0.00225,
"grad_norm": 0.4436502456665039,
"grad_norm_var": 0.013690624557478688,
"learning_rate": 5e-05,
"loss": 0.2037,
"loss/crossentropy": 2.9042821526527405,
"loss/hidden": 0.0,
"loss/logits": 0.20374128222465515,
"loss/reg": 3.483499765396118,
"step": 225
},
{
"epoch": 0.00226,
"grad_norm": 0.44937804341316223,
"grad_norm_var": 0.01373632101878638,
"learning_rate": 5e-05,
"loss": 0.1588,
"loss/crossentropy": 2.79194039106369,
"loss/hidden": 0.0,
"loss/logits": 0.1587841510772705,
"loss/reg": 3.480142593383789,
"step": 226
},
{
"epoch": 0.00227,
"grad_norm": 0.3453376889228821,
"grad_norm_var": 0.013826164241530992,
"learning_rate": 5e-05,
"loss": 0.1659,
"loss/crossentropy": 2.7488330006599426,
"loss/hidden": 0.0,
"loss/logits": 0.16589100658893585,
"loss/reg": 3.476062297821045,
"step": 227
},
{
"epoch": 0.00228,
"grad_norm": 0.3845584988594055,
"grad_norm_var": 0.013584549048495138,
"learning_rate": 5e-05,
"loss": 0.1842,
"loss/crossentropy": 2.6935607194900513,
"loss/hidden": 0.0,
"loss/logits": 0.18416164070367813,
"loss/reg": 3.4726946353912354,
"step": 228
},
{
"epoch": 0.00229,
"grad_norm": 0.3347846567630768,
"grad_norm_var": 0.013727727146734722,
"learning_rate": 5e-05,
"loss": 0.1767,
"loss/crossentropy": 2.6182947754859924,
"loss/hidden": 0.0,
"loss/logits": 0.1767422929406166,
"loss/reg": 3.469238519668579,
"step": 229
},
{
"epoch": 0.0023,
"grad_norm": 0.35126739740371704,
"grad_norm_var": 0.01362772883112919,
"learning_rate": 5e-05,
"loss": 0.1694,
"loss/crossentropy": 2.8005401492118835,
"loss/hidden": 0.0,
"loss/logits": 0.1694028675556183,
"loss/reg": 3.4662599563598633,
"step": 230
},
{
"epoch": 0.00231,
"grad_norm": 0.37644773721694946,
"grad_norm_var": 0.0016784352808341082,
"learning_rate": 5e-05,
"loss": 0.1677,
"loss/crossentropy": 2.7537949085235596,
"loss/hidden": 0.0,
"loss/logits": 0.16772692278027534,
"loss/reg": 3.4623701572418213,
"step": 231
},
{
"epoch": 0.00232,
"grad_norm": 0.33086928725242615,
"grad_norm_var": 0.0015241936410912834,
"learning_rate": 5e-05,
"loss": 0.1624,
"loss/crossentropy": 2.7844293117523193,
"loss/hidden": 0.0,
"loss/logits": 0.1624348722398281,
"loss/reg": 3.459073066711426,
"step": 232
},
{
"epoch": 0.00233,
"grad_norm": 0.3152429461479187,
"grad_norm_var": 0.0016449122438399724,
"learning_rate": 5e-05,
"loss": 0.1607,
"loss/crossentropy": 2.5863555669784546,
"loss/hidden": 0.0,
"loss/logits": 0.16065017879009247,
"loss/reg": 3.456038475036621,
"step": 233
},
{
"epoch": 0.00234,
"grad_norm": 0.34679386019706726,
"grad_norm_var": 0.001649030072333372,
"learning_rate": 5e-05,
"loss": 0.1656,
"loss/crossentropy": 2.9068891406059265,
"loss/hidden": 0.0,
"loss/logits": 0.16555847227573395,
"loss/reg": 3.452618360519409,
"step": 234
},
{
"epoch": 0.00235,
"grad_norm": 0.36684513092041016,
"grad_norm_var": 0.001520832425550959,
"learning_rate": 5e-05,
"loss": 0.1878,
"loss/crossentropy": 2.6781840920448303,
"loss/hidden": 0.0,
"loss/logits": 0.18775511160492897,
"loss/reg": 3.4493637084960938,
"step": 235
},
{
"epoch": 0.00236,
"grad_norm": 0.39043235778808594,
"grad_norm_var": 0.0015693055369300879,
"learning_rate": 5e-05,
"loss": 0.1559,
"loss/crossentropy": 2.9237093925476074,
"loss/hidden": 0.0,
"loss/logits": 0.15592358261346817,
"loss/reg": 3.446392059326172,
"step": 236
},
{
"epoch": 0.00237,
"grad_norm": 0.3486286997795105,
"grad_norm_var": 0.0015605921838873513,
"learning_rate": 5e-05,
"loss": 0.1524,
"loss/crossentropy": 2.8276549577713013,
"loss/hidden": 0.0,
"loss/logits": 0.1523873247206211,
"loss/reg": 3.443490505218506,
"step": 237
},
{
"epoch": 0.00238,
"grad_norm": 0.4030380845069885,
"grad_norm_var": 0.0016408419596595262,
"learning_rate": 5e-05,
"loss": 0.1839,
"loss/crossentropy": 2.7374503016471863,
"loss/hidden": 0.0,
"loss/logits": 0.1839219257235527,
"loss/reg": 3.440230131149292,
"step": 238
},
{
"epoch": 0.00239,
"grad_norm": 0.3677695095539093,
"grad_norm_var": 0.0015933721835237928,
"learning_rate": 5e-05,
"loss": 0.1725,
"loss/crossentropy": 2.637487053871155,
"loss/hidden": 0.0,
"loss/logits": 0.1725292131304741,
"loss/reg": 3.4369444847106934,
"step": 239
},
{
"epoch": 0.0024,
"grad_norm": 0.3092736303806305,
"grad_norm_var": 0.001648043714460871,
"learning_rate": 5e-05,
"loss": 0.1608,
"loss/crossentropy": 2.785566747188568,
"loss/hidden": 0.0,
"loss/logits": 0.16075557842850685,
"loss/reg": 3.4329705238342285,
"step": 240
},
{
"epoch": 0.00241,
"grad_norm": 0.3242727518081665,
"grad_norm_var": 0.001311046071157899,
"learning_rate": 5e-05,
"loss": 0.1641,
"loss/crossentropy": 2.7823829650878906,
"loss/hidden": 0.0,
"loss/logits": 0.16410458087921143,
"loss/reg": 3.429222345352173,
"step": 241
},
{
"epoch": 0.00242,
"grad_norm": 0.3544396758079529,
"grad_norm_var": 0.0007310749754719385,
"learning_rate": 5e-05,
"loss": 0.1742,
"loss/crossentropy": 2.7899482250213623,
"loss/hidden": 0.0,
"loss/logits": 0.1741911694407463,
"loss/reg": 3.4251747131347656,
"step": 242
},
{
"epoch": 0.00243,
"grad_norm": 0.3156209886074066,
"grad_norm_var": 0.0008171231835736463,
"learning_rate": 5e-05,
"loss": 0.159,
"loss/crossentropy": 2.7414376735687256,
"loss/hidden": 0.0,
"loss/logits": 0.15898872911930084,
"loss/reg": 3.421576976776123,
"step": 243
},
{
"epoch": 0.00244,
"grad_norm": 0.3353999853134155,
"grad_norm_var": 0.000749955482525048,
"learning_rate": 5e-05,
"loss": 0.1669,
"loss/crossentropy": 2.707472503185272,
"loss/hidden": 0.0,
"loss/logits": 0.16693224385380745,
"loss/reg": 3.417820930480957,
"step": 244
},
{
"epoch": 0.00245,
"grad_norm": 0.32766133546829224,
"grad_norm_var": 0.0007658640613261528,
"learning_rate": 5e-05,
"loss": 0.1761,
"loss/crossentropy": 2.6950490474700928,
"loss/hidden": 0.0,
"loss/logits": 0.17608999833464622,
"loss/reg": 3.414095640182495,
"step": 245
},
{
"epoch": 0.00246,
"grad_norm": 0.31360548734664917,
"grad_norm_var": 0.0008368534177580581,
"learning_rate": 5e-05,
"loss": 0.1578,
"loss/crossentropy": 2.6977627873420715,
"loss/hidden": 0.0,
"loss/logits": 0.15783175826072693,
"loss/reg": 3.409533739089966,
"step": 246
},
{
"epoch": 0.00247,
"grad_norm": 0.35324403643608093,
"grad_norm_var": 0.0007744365123312817,
"learning_rate": 5e-05,
"loss": 0.1688,
"loss/crossentropy": 2.8509859442710876,
"loss/hidden": 0.0,
"loss/logits": 0.16875524446368217,
"loss/reg": 3.4052798748016357,
"step": 247
},
{
"epoch": 0.00248,
"grad_norm": 0.41796907782554626,
"grad_norm_var": 0.0010967197155327421,
"learning_rate": 5e-05,
"loss": 0.18,
"loss/crossentropy": 2.701251804828644,
"loss/hidden": 0.0,
"loss/logits": 0.1800428181886673,
"loss/reg": 3.4006803035736084,
"step": 248
},
{
"epoch": 0.00249,
"grad_norm": 0.33844876289367676,
"grad_norm_var": 0.0010247223552569313,
"learning_rate": 5e-05,
"loss": 0.1737,
"loss/crossentropy": 2.7646324038505554,
"loss/hidden": 0.0,
"loss/logits": 0.17367269843816757,
"loss/reg": 3.3968873023986816,
"step": 249
},
{
"epoch": 0.0025,
"grad_norm": 0.31011876463890076,
"grad_norm_var": 0.0011285754764581786,
"learning_rate": 5e-05,
"loss": 0.1591,
"loss/crossentropy": 2.7303661704063416,
"loss/hidden": 0.0,
"loss/logits": 0.15912269055843353,
"loss/reg": 3.3925790786743164,
"step": 250
},
{
"epoch": 0.00251,
"grad_norm": 0.4837491512298584,
"grad_norm_var": 0.0022679356659945546,
"learning_rate": 5e-05,
"loss": 0.1845,
"loss/crossentropy": 2.718783438205719,
"loss/hidden": 0.0,
"loss/logits": 0.18454211205244064,
"loss/reg": 3.389193296432495,
"step": 251
},
{
"epoch": 0.00252,
"grad_norm": 0.30302709341049194,
"grad_norm_var": 0.002342444325573334,
"learning_rate": 5e-05,
"loss": 0.1527,
"loss/crossentropy": 2.7513213753700256,
"loss/hidden": 0.0,
"loss/logits": 0.15272299572825432,
"loss/reg": 3.384976863861084,
"step": 252
},
{
"epoch": 0.00253,
"grad_norm": 0.3376671075820923,
"grad_norm_var": 0.002352530797232196,
"learning_rate": 5e-05,
"loss": 0.1717,
"loss/crossentropy": 2.7082377672195435,
"loss/hidden": 0.0,
"loss/logits": 0.1717442087829113,
"loss/reg": 3.381958246231079,
"step": 253
},
{
"epoch": 0.00254,
"grad_norm": 0.3470434546470642,
"grad_norm_var": 0.00215032290339258,
"learning_rate": 5e-05,
"loss": 0.1751,
"loss/crossentropy": 2.7747623324394226,
"loss/hidden": 0.0,
"loss/logits": 0.17506984621286392,
"loss/reg": 3.3780975341796875,
"step": 254
},
{
"epoch": 0.00255,
"grad_norm": 0.35893791913986206,
"grad_norm_var": 0.002129806794166807,
"learning_rate": 5e-05,
"loss": 0.176,
"loss/crossentropy": 2.6670790910720825,
"loss/hidden": 0.0,
"loss/logits": 0.17602670192718506,
"loss/reg": 3.374431848526001,
"step": 255
},
{
"epoch": 0.00256,
"grad_norm": 0.33274415135383606,
"grad_norm_var": 0.002050384071076557,
"learning_rate": 5e-05,
"loss": 0.1683,
"loss/crossentropy": 2.930284321308136,
"loss/hidden": 0.0,
"loss/logits": 0.16833152994513512,
"loss/reg": 3.3709969520568848,
"step": 256
},
{
"epoch": 0.00257,
"grad_norm": 0.3107251822948456,
"grad_norm_var": 0.0021031284267367073,
"learning_rate": 5e-05,
"loss": 0.154,
"loss/crossentropy": 2.7738651037216187,
"loss/hidden": 0.0,
"loss/logits": 0.15401111543178558,
"loss/reg": 3.36681866645813,
"step": 257
},
{
"epoch": 0.00258,
"grad_norm": 0.3238702118396759,
"grad_norm_var": 0.00212825610345269,
"learning_rate": 5e-05,
"loss": 0.1485,
"loss/crossentropy": 2.7926384806632996,
"loss/hidden": 0.0,
"loss/logits": 0.14850713685154915,
"loss/reg": 3.363298177719116,
"step": 258
},
{
"epoch": 0.00259,
"grad_norm": 0.3937188982963562,
"grad_norm_var": 0.0022101531057158843,
"learning_rate": 5e-05,
"loss": 0.1796,
"loss/crossentropy": 2.732594311237335,
"loss/hidden": 0.0,
"loss/logits": 0.17957409471273422,
"loss/reg": 3.3592464923858643,
"step": 259
},
{
"epoch": 0.0026,
"grad_norm": 0.35869738459587097,
"grad_norm_var": 0.002201067050087302,
"learning_rate": 5e-05,
"loss": 0.1608,
"loss/crossentropy": 2.6412158608436584,
"loss/hidden": 0.0,
"loss/logits": 0.1607726439833641,
"loss/reg": 3.355531692504883,
"step": 260
},
{
"epoch": 0.00261,
"grad_norm": 0.342753529548645,
"grad_norm_var": 0.002168938888458849,
"learning_rate": 5e-05,
"loss": 0.1679,
"loss/crossentropy": 2.8253002762794495,
"loss/hidden": 0.0,
"loss/logits": 0.16785955801606178,
"loss/reg": 3.3510327339172363,
"step": 261
},
{
"epoch": 0.00262,
"grad_norm": 0.3396557867527008,
"grad_norm_var": 0.0020792270475482005,
"learning_rate": 5e-05,
"loss": 0.1719,
"loss/crossentropy": 2.5446697473526,
"loss/hidden": 0.0,
"loss/logits": 0.1719457022845745,
"loss/reg": 3.3462460041046143,
"step": 262
},
{
"epoch": 0.00263,
"grad_norm": 0.326615571975708,
"grad_norm_var": 0.002123647634079288,
"learning_rate": 5e-05,
"loss": 0.1662,
"loss/crossentropy": 2.7185133695602417,
"loss/hidden": 0.0,
"loss/logits": 0.16621045768260956,
"loss/reg": 3.342698097229004,
"step": 263
},
{
"epoch": 0.00264,
"grad_norm": 0.372024804353714,
"grad_norm_var": 0.0018490612448516057,
"learning_rate": 5e-05,
"loss": 0.1785,
"loss/crossentropy": 2.90339195728302,
"loss/hidden": 0.0,
"loss/logits": 0.17848360165953636,
"loss/reg": 3.3390297889709473,
"step": 264
},
{
"epoch": 0.00265,
"grad_norm": 0.336412638425827,
"grad_norm_var": 0.0018521135396843155,
"learning_rate": 5e-05,
"loss": 0.1685,
"loss/crossentropy": 2.6998194456100464,
"loss/hidden": 0.0,
"loss/logits": 0.1684669330716133,
"loss/reg": 3.3360044956207275,
"step": 265
},
{
"epoch": 0.00266,
"grad_norm": 0.3179170787334442,
"grad_norm_var": 0.0018158920564407192,
"learning_rate": 5e-05,
"loss": 0.164,
"loss/crossentropy": 2.713620126247406,
"loss/hidden": 0.0,
"loss/logits": 0.16402245312929153,
"loss/reg": 3.33297061920166,
"step": 266
},
{
"epoch": 0.00267,
"grad_norm": 0.32180216908454895,
"grad_norm_var": 0.0005475447645484169,
"learning_rate": 5e-05,
"loss": 0.1561,
"loss/crossentropy": 2.8285736441612244,
"loss/hidden": 0.0,
"loss/logits": 0.15614933148026466,
"loss/reg": 3.330070972442627,
"step": 267
},
{
"epoch": 0.00268,
"grad_norm": 0.34155359864234924,
"grad_norm_var": 0.00045564919377512797,
"learning_rate": 5e-05,
"loss": 0.1666,
"loss/crossentropy": 2.698326587677002,
"loss/hidden": 0.0,
"loss/logits": 0.16663997247815132,
"loss/reg": 3.326782464981079,
"step": 268
},
{
"epoch": 0.00269,
"grad_norm": 0.3281239867210388,
"grad_norm_var": 0.0004660702159405468,
"learning_rate": 5e-05,
"loss": 0.1547,
"loss/crossentropy": 2.7132135033607483,
"loss/hidden": 0.0,
"loss/logits": 0.15473050251603127,
"loss/reg": 3.3232901096343994,
"step": 269
},
{
"epoch": 0.0027,
"grad_norm": 0.3694444000720978,
"grad_norm_var": 0.0005161187813034911,
"learning_rate": 5e-05,
"loss": 0.1658,
"loss/crossentropy": 2.943029820919037,
"loss/hidden": 0.0,
"loss/logits": 0.16579603031277657,
"loss/reg": 3.319425344467163,
"step": 270
},
{
"epoch": 0.00271,
"grad_norm": 0.3521305024623871,
"grad_norm_var": 0.0005038113254072218,
"learning_rate": 5e-05,
"loss": 0.1762,
"loss/crossentropy": 2.813421130180359,
"loss/hidden": 0.0,
"loss/logits": 0.17624986171722412,
"loss/reg": 3.31587553024292,
"step": 271
},
{
"epoch": 0.00272,
"grad_norm": 0.3419167995452881,
"grad_norm_var": 0.0004980410714001496,
"learning_rate": 5e-05,
"loss": 0.1579,
"loss/crossentropy": 2.6725985407829285,
"loss/hidden": 0.0,
"loss/logits": 0.1579086296260357,
"loss/reg": 3.313774347305298,
"step": 272
},
{
"epoch": 0.00273,
"grad_norm": 0.45973771810531616,
"grad_norm_var": 0.001257799356739812,
"learning_rate": 5e-05,
"loss": 0.1806,
"loss/crossentropy": 2.7593576908111572,
"loss/hidden": 0.0,
"loss/logits": 0.1806396320462227,
"loss/reg": 3.311671257019043,
"step": 273
},
{
"epoch": 0.00274,
"grad_norm": 0.327812522649765,
"grad_norm_var": 0.00124416933097297,
"learning_rate": 5e-05,
"loss": 0.1544,
"loss/crossentropy": 2.7368595004081726,
"loss/hidden": 0.0,
"loss/logits": 0.15438436716794968,
"loss/reg": 3.308312177658081,
"step": 274
},
{
"epoch": 0.00275,
"grad_norm": 0.43593037128448486,
"grad_norm_var": 0.001590926391083336,
"learning_rate": 5e-05,
"loss": 0.1721,
"loss/crossentropy": 2.8178694248199463,
"loss/hidden": 0.0,
"loss/logits": 0.17214355245232582,
"loss/reg": 3.30526065826416,
"step": 275
},
{
"epoch": 0.00276,
"grad_norm": 0.361247181892395,
"grad_norm_var": 0.0015927484925991053,
"learning_rate": 5e-05,
"loss": 0.1735,
"loss/crossentropy": 2.734869599342346,
"loss/hidden": 0.0,
"loss/logits": 0.17345493659377098,
"loss/reg": 3.3015174865722656,
"step": 276
},
{
"epoch": 0.00277,
"grad_norm": 0.3708873689174652,
"grad_norm_var": 0.0015974331537993436,
"learning_rate": 5e-05,
"loss": 0.1628,
"loss/crossentropy": 2.67022705078125,
"loss/hidden": 0.0,
"loss/logits": 0.16277828440070152,
"loss/reg": 3.2979342937469482,
"step": 277
},
{
"epoch": 0.00278,
"grad_norm": 0.3481086790561676,
"grad_norm_var": 0.0015829700282981919,
"learning_rate": 5e-05,
"loss": 0.1577,
"loss/crossentropy": 2.700168251991272,
"loss/hidden": 0.0,
"loss/logits": 0.15773406997323036,
"loss/reg": 3.294177532196045,
"step": 278
},
{
"epoch": 0.00279,
"grad_norm": 0.5134589076042175,
"grad_norm_var": 0.003008442642246208,
"learning_rate": 5e-05,
"loss": 0.1766,
"loss/crossentropy": 2.9033528566360474,
"loss/hidden": 0.0,
"loss/logits": 0.17655130848288536,
"loss/reg": 3.2909915447235107,
"step": 279
},
{
"epoch": 0.0028,
"grad_norm": 0.38205716013908386,
"grad_norm_var": 0.0030192383608610503,
"learning_rate": 5e-05,
"loss": 0.1934,
"loss/crossentropy": 2.7160211205482483,
"loss/hidden": 0.0,
"loss/logits": 0.19341961666941643,
"loss/reg": 3.2878634929656982,
"step": 280
},
{
"epoch": 0.00281,
"grad_norm": 0.3628558814525604,
"grad_norm_var": 0.002947045102075485,
"learning_rate": 5e-05,
"loss": 0.1709,
"loss/crossentropy": 2.6912715435028076,
"loss/hidden": 0.0,
"loss/logits": 0.17094064503908157,
"loss/reg": 3.284353256225586,
"step": 281
},
{
"epoch": 0.00282,
"grad_norm": 0.32757696509361267,
"grad_norm_var": 0.002884588952342071,
"learning_rate": 5e-05,
"loss": 0.1702,
"loss/crossentropy": 2.7818912267684937,
"loss/hidden": 0.0,
"loss/logits": 0.17016061395406723,
"loss/reg": 3.281097412109375,
"step": 282
},
{
"epoch": 0.00283,
"grad_norm": 0.5035110712051392,
"grad_norm_var": 0.003743174506031214,
"learning_rate": 5e-05,
"loss": 0.1891,
"loss/crossentropy": 2.8356027603149414,
"loss/hidden": 0.0,
"loss/logits": 0.1890563629567623,
"loss/reg": 3.277557611465454,
"step": 283
},
{
"epoch": 0.00284,
"grad_norm": 0.4021988809108734,
"grad_norm_var": 0.003638735284583853,
"learning_rate": 5e-05,
"loss": 0.1574,
"loss/crossentropy": 2.768595337867737,
"loss/hidden": 0.0,
"loss/logits": 0.15735788643360138,
"loss/reg": 3.274083375930786,
"step": 284
},
{
"epoch": 0.00285,
"grad_norm": 0.3557356297969818,
"grad_norm_var": 0.0034707811870306284,
"learning_rate": 5e-05,
"loss": 0.1618,
"loss/crossentropy": 2.7620763182640076,
"loss/hidden": 0.0,
"loss/logits": 0.16183337569236755,
"loss/reg": 3.2709619998931885,
"step": 285
},
{
"epoch": 0.00286,
"grad_norm": 0.35383257269859314,
"grad_norm_var": 0.0035254991255895857,
"learning_rate": 5e-05,
"loss": 0.164,
"loss/crossentropy": 2.7777530550956726,
"loss/hidden": 0.0,
"loss/logits": 0.16404272243380547,
"loss/reg": 3.2678396701812744,
"step": 286
},
{
"epoch": 0.00287,
"grad_norm": 0.3291935324668884,
"grad_norm_var": 0.0036663583934404683,
"learning_rate": 5e-05,
"loss": 0.1545,
"loss/crossentropy": 2.7828534841537476,
"loss/hidden": 0.0,
"loss/logits": 0.15450828149914742,
"loss/reg": 3.2639055252075195,
"step": 287
},
{
"epoch": 0.00288,
"grad_norm": 0.3174595534801483,
"grad_norm_var": 0.003847509504795695,
"learning_rate": 5e-05,
"loss": 0.163,
"loss/crossentropy": 2.7973127365112305,
"loss/hidden": 0.0,
"loss/logits": 0.16296324506402016,
"loss/reg": 3.2608911991119385,
"step": 288
},
{
"epoch": 0.00289,
"grad_norm": 0.3723791539669037,
"grad_norm_var": 0.0034478366033269445,
"learning_rate": 5e-05,
"loss": 0.1757,
"loss/crossentropy": 2.630415976047516,
"loss/hidden": 0.0,
"loss/logits": 0.17573364078998566,
"loss/reg": 3.257523775100708,
"step": 289
},
{
"epoch": 0.0029,
"grad_norm": 0.38034215569496155,
"grad_norm_var": 0.003261674725795945,
"learning_rate": 5e-05,
"loss": 0.1689,
"loss/crossentropy": 2.704579532146454,
"loss/hidden": 0.0,
"loss/logits": 0.16892588511109352,
"loss/reg": 3.254065752029419,
"step": 290
},
{
"epoch": 0.00291,
"grad_norm": 1.2464065551757812,
"grad_norm_var": 0.05011180607704591,
"learning_rate": 5e-05,
"loss": 0.1921,
"loss/crossentropy": 2.8595897555351257,
"loss/hidden": 0.0,
"loss/logits": 0.19211571291089058,
"loss/reg": 3.2510294914245605,
"step": 291
},
{
"epoch": 0.00292,
"grad_norm": 0.3307066559791565,
"grad_norm_var": 0.05046209325623486,
"learning_rate": 5e-05,
"loss": 0.1681,
"loss/crossentropy": 2.8195151686668396,
"loss/hidden": 0.0,
"loss/logits": 0.16810721158981323,
"loss/reg": 3.2478625774383545,
"step": 292
},
{
"epoch": 0.00293,
"grad_norm": 0.33664193749427795,
"grad_norm_var": 0.05081007066127065,
"learning_rate": 5e-05,
"loss": 0.1602,
"loss/crossentropy": 2.912789523601532,
"loss/hidden": 0.0,
"loss/logits": 0.16015203669667244,
"loss/reg": 3.2451751232147217,
"step": 293
},
{
"epoch": 0.00294,
"grad_norm": 0.42365437746047974,
"grad_norm_var": 0.05035293502217161,
"learning_rate": 5e-05,
"loss": 0.1833,
"loss/crossentropy": 2.7459517121315002,
"loss/hidden": 0.0,
"loss/logits": 0.18333038315176964,
"loss/reg": 3.242023229598999,
"step": 294
},
{
"epoch": 0.00295,
"grad_norm": 0.40840578079223633,
"grad_norm_var": 0.049924464393714924,
"learning_rate": 5e-05,
"loss": 0.176,
"loss/crossentropy": 2.778249144554138,
"loss/hidden": 0.0,
"loss/logits": 0.17595936357975006,
"loss/reg": 3.2388389110565186,
"step": 295
},
{
"epoch": 0.00296,
"grad_norm": 0.3591618835926056,
"grad_norm_var": 0.05009460642018338,
"learning_rate": 5e-05,
"loss": 0.1722,
"loss/crossentropy": 2.7121748328208923,
"loss/hidden": 0.0,
"loss/logits": 0.17217102646827698,
"loss/reg": 3.2364017963409424,
"step": 296
},
{
"epoch": 0.00297,
"grad_norm": 0.31615006923675537,
"grad_norm_var": 0.05062186135785553,
"learning_rate": 5e-05,
"loss": 0.1488,
"loss/crossentropy": 2.7933038473129272,
"loss/hidden": 0.0,
"loss/logits": 0.14879318699240685,
"loss/reg": 3.2344272136688232,
"step": 297
},
{
"epoch": 0.00298,
"grad_norm": 0.3586377799510956,
"grad_norm_var": 0.050288172636787816,
"learning_rate": 5e-05,
"loss": 0.1492,
"loss/crossentropy": 2.869333803653717,
"loss/hidden": 0.0,
"loss/logits": 0.14924299344420433,
"loss/reg": 3.232086181640625,
"step": 298
},
{
"epoch": 0.00299,
"grad_norm": 0.37798771262168884,
"grad_norm_var": 0.04995309393064352,
"learning_rate": 5e-05,
"loss": 0.1652,
"loss/crossentropy": 2.913083255290985,
"loss/hidden": 0.0,
"loss/logits": 0.1651969812810421,
"loss/reg": 3.2296016216278076,
"step": 299
},
{
"epoch": 0.003,
"grad_norm": 0.5914519429206848,
"grad_norm_var": 0.05182304954391634,
"learning_rate": 5e-05,
"loss": 0.1893,
"loss/crossentropy": 2.8007007241249084,
"loss/hidden": 0.0,
"loss/logits": 0.18929021432995796,
"loss/reg": 3.2266576290130615,
"step": 300
},
{
"epoch": 0.00301,
"grad_norm": 0.3292617201805115,
"grad_norm_var": 0.05212417516215169,
"learning_rate": 5e-05,
"loss": 0.16,
"loss/crossentropy": 2.7326099276542664,
"loss/hidden": 0.0,
"loss/logits": 0.15998771041631699,
"loss/reg": 3.2236239910125732,
"step": 301
},
{
"epoch": 0.00302,
"grad_norm": 0.3807355761528015,
"grad_norm_var": 0.05190702763823275,
"learning_rate": 5e-05,
"loss": 0.1831,
"loss/crossentropy": 2.7199636101722717,
"loss/hidden": 0.0,
"loss/logits": 0.1831417679786682,
"loss/reg": 3.220425844192505,
"step": 302
},
{
"epoch": 0.00303,
"grad_norm": 0.4008902907371521,
"grad_norm_var": 0.05127743798183474,
"learning_rate": 5e-05,
"loss": 0.1777,
"loss/crossentropy": 2.7570589184761047,
"loss/hidden": 0.0,
"loss/logits": 0.177694384008646,
"loss/reg": 3.2171859741210938,
"step": 303
},
{
"epoch": 0.00304,
"grad_norm": 0.35962697863578796,
"grad_norm_var": 0.05073816419262332,
"learning_rate": 5e-05,
"loss": 0.1574,
"loss/crossentropy": 2.809377670288086,
"loss/hidden": 0.0,
"loss/logits": 0.15741629898548126,
"loss/reg": 3.2137656211853027,
"step": 304
},
{
"epoch": 0.00305,
"grad_norm": 0.404453843832016,
"grad_norm_var": 0.05053133217663517,
"learning_rate": 5e-05,
"loss": 0.196,
"loss/crossentropy": 2.6374824047088623,
"loss/hidden": 0.0,
"loss/logits": 0.19603004679083824,
"loss/reg": 3.2106897830963135,
"step": 305
},
{
"epoch": 0.00306,
"grad_norm": 0.3411775231361389,
"grad_norm_var": 0.05092714807133293,
"learning_rate": 5e-05,
"loss": 0.1701,
"loss/crossentropy": 2.7098072171211243,
"loss/hidden": 0.0,
"loss/logits": 0.17005891352891922,
"loss/reg": 3.2072975635528564,
"step": 306
},
{
"epoch": 0.00307,
"grad_norm": 0.48913073539733887,
"grad_norm_var": 0.004874772049479148,
"learning_rate": 5e-05,
"loss": 0.2396,
"loss/crossentropy": 2.8529589772224426,
"loss/hidden": 0.0,
"loss/logits": 0.23958228901028633,
"loss/reg": 3.2032482624053955,
"step": 307
},
{
"epoch": 0.00308,
"grad_norm": 0.3359135389328003,
"grad_norm_var": 0.004836687315538634,
"learning_rate": 5e-05,
"loss": 0.154,
"loss/crossentropy": 2.825865149497986,
"loss/hidden": 0.0,
"loss/logits": 0.15401727706193924,
"loss/reg": 3.2001187801361084,
"step": 308
},
{
"epoch": 0.00309,
"grad_norm": 0.3673790693283081,
"grad_norm_var": 0.004683902714770716,
"learning_rate": 5e-05,
"loss": 0.1701,
"loss/crossentropy": 2.757752239704132,
"loss/hidden": 0.0,
"loss/logits": 0.1700810343027115,
"loss/reg": 3.197371244430542,
"step": 309
},
{
"epoch": 0.0031,
"grad_norm": 0.3675878643989563,
"grad_norm_var": 0.004630661781797581,
"learning_rate": 5e-05,
"loss": 0.1742,
"loss/crossentropy": 2.9055893421173096,
"loss/hidden": 0.0,
"loss/logits": 0.17424792051315308,
"loss/reg": 3.193922758102417,
"step": 310
},
{
"epoch": 0.00311,
"grad_norm": 0.3216918110847473,
"grad_norm_var": 0.004850203191397428,
"learning_rate": 5e-05,
"loss": 0.1529,
"loss/crossentropy": 2.7421942353248596,
"loss/hidden": 0.0,
"loss/logits": 0.15289029106497765,
"loss/reg": 3.1913256645202637,
"step": 311
},
{
"epoch": 0.00312,
"grad_norm": 0.30283358693122864,
"grad_norm_var": 0.005214980747276743,
"learning_rate": 5e-05,
"loss": 0.1501,
"loss/crossentropy": 2.68456107378006,
"loss/hidden": 0.0,
"loss/logits": 0.15012749284505844,
"loss/reg": 3.188302755355835,
"step": 312
},
{
"epoch": 0.00313,
"grad_norm": 0.3840731978416443,
"grad_norm_var": 0.004944937932171186,
"learning_rate": 5e-05,
"loss": 0.1565,
"loss/crossentropy": 2.7386457920074463,
"loss/hidden": 0.0,
"loss/logits": 0.15645165741443634,
"loss/reg": 3.1856327056884766,
"step": 313
},
{
"epoch": 0.00314,
"grad_norm": 0.3471456468105316,
"grad_norm_var": 0.004989069609234416,
"learning_rate": 5e-05,
"loss": 0.1661,
"loss/crossentropy": 2.8941837549209595,
"loss/hidden": 0.0,
"loss/logits": 0.16613885760307312,
"loss/reg": 3.182285785675049,
"step": 314
},
{
"epoch": 0.00315,
"grad_norm": 0.40083590149879456,
"grad_norm_var": 0.005011503442318803,
"learning_rate": 5e-05,
"loss": 0.1816,
"loss/crossentropy": 2.830922782421112,
"loss/hidden": 0.0,
"loss/logits": 0.18162691593170166,
"loss/reg": 3.1795294284820557,
"step": 315
},
{
"epoch": 0.00316,
"grad_norm": 0.40112313628196716,
"grad_norm_var": 0.001979603921073251,
"learning_rate": 5e-05,
"loss": 0.163,
"loss/crossentropy": 2.5529216527938843,
"loss/hidden": 0.0,
"loss/logits": 0.1629548817873001,
"loss/reg": 3.1763863563537598,
"step": 316
},
{
"epoch": 0.00317,
"grad_norm": 0.3936459720134735,
"grad_norm_var": 0.001881530067811854,
"learning_rate": 5e-05,
"loss": 0.1851,
"loss/crossentropy": 2.780943751335144,
"loss/hidden": 0.0,
"loss/logits": 0.18506472185254097,
"loss/reg": 3.173980474472046,
"step": 317
},
{
"epoch": 0.00318,
"grad_norm": 0.3727342486381531,
"grad_norm_var": 0.0018792953911145364,
"learning_rate": 5e-05,
"loss": 0.1827,
"loss/crossentropy": 2.76874041557312,
"loss/hidden": 0.0,
"loss/logits": 0.1826922371983528,
"loss/reg": 3.1704628467559814,
"step": 318
},
{
"epoch": 0.00319,
"grad_norm": 0.3470066785812378,
"grad_norm_var": 0.0018703712702832478,
"learning_rate": 5e-05,
"loss": 0.1598,
"loss/crossentropy": 2.740228831768036,
"loss/hidden": 0.0,
"loss/logits": 0.1597808077931404,
"loss/reg": 3.1674654483795166,
"step": 319
},
{
"epoch": 0.0032,
"grad_norm": 0.3653993010520935,
"grad_norm_var": 0.0018636832816178708,
"learning_rate": 5e-05,
"loss": 0.1549,
"loss/crossentropy": 2.883521616458893,
"loss/hidden": 0.0,
"loss/logits": 0.15490083023905754,
"loss/reg": 3.1641883850097656,
"step": 320
},
{
"epoch": 0.00321,
"grad_norm": 0.3510638475418091,
"grad_norm_var": 0.0018064205203171017,
"learning_rate": 5e-05,
"loss": 0.1577,
"loss/crossentropy": 2.9007150530815125,
"loss/hidden": 0.0,
"loss/logits": 0.15773681923747063,
"loss/reg": 3.1610915660858154,
"step": 321
},
{
"epoch": 0.00322,
"grad_norm": 0.5068875551223755,
"grad_norm_var": 0.0029290004167608335,
"learning_rate": 5e-05,
"loss": 0.2013,
"loss/crossentropy": 2.716179847717285,
"loss/hidden": 0.0,
"loss/logits": 0.2013130635023117,
"loss/reg": 3.1571173667907715,
"step": 322
},
{
"epoch": 0.00323,
"grad_norm": 0.40178200602531433,
"grad_norm_var": 0.0021162756618779235,
"learning_rate": 5e-05,
"loss": 0.1809,
"loss/crossentropy": 2.9381837844848633,
"loss/hidden": 0.0,
"loss/logits": 0.1809130534529686,
"loss/reg": 3.1540093421936035,
"step": 323
},
{
"epoch": 0.00324,
"grad_norm": 0.35252845287323,
"grad_norm_var": 0.0020514948206895294,
"learning_rate": 5e-05,
"loss": 0.1617,
"loss/crossentropy": 2.743869721889496,
"loss/hidden": 0.0,
"loss/logits": 0.16171807795763016,
"loss/reg": 3.1503779888153076,
"step": 324
},
{
"epoch": 0.00325,
"grad_norm": 0.36802011728286743,
"grad_norm_var": 0.0020509560983741053,
"learning_rate": 5e-05,
"loss": 0.1722,
"loss/crossentropy": 2.934039294719696,
"loss/hidden": 0.0,
"loss/logits": 0.17217914387583733,
"loss/reg": 3.1467440128326416,
"step": 325
},
{
"epoch": 0.00326,
"grad_norm": 0.4267319142818451,
"grad_norm_var": 0.0022188398751517274,
"learning_rate": 5e-05,
"loss": 0.1924,
"loss/crossentropy": 2.802468180656433,
"loss/hidden": 0.0,
"loss/logits": 0.19241869449615479,
"loss/reg": 3.1427693367004395,
"step": 326
},
{
"epoch": 0.00327,
"grad_norm": 0.34044548869132996,
"grad_norm_var": 0.0021007258044081806,
"learning_rate": 5e-05,
"loss": 0.1522,
"loss/crossentropy": 2.8443135619163513,
"loss/hidden": 0.0,
"loss/logits": 0.1522187888622284,
"loss/reg": 3.1388471126556396,
"step": 327
},
{
"epoch": 0.00328,
"grad_norm": 0.4276120066642761,
"grad_norm_var": 0.001808451579785623,
"learning_rate": 5e-05,
"loss": 0.1708,
"loss/crossentropy": 2.8915366530418396,
"loss/hidden": 0.0,
"loss/logits": 0.17084889113903046,
"loss/reg": 3.134800434112549,
"step": 328
},
{
"epoch": 0.00329,
"grad_norm": 0.3486219346523285,
"grad_norm_var": 0.0018993689379833108,
"learning_rate": 5e-05,
"loss": 0.1687,
"loss/crossentropy": 2.575106978416443,
"loss/hidden": 0.0,
"loss/logits": 0.16874929517507553,
"loss/reg": 3.1303272247314453,
"step": 329
},
{
"epoch": 0.0033,
"grad_norm": 0.365105539560318,
"grad_norm_var": 0.0018301403367672127,
"learning_rate": 5e-05,
"loss": 0.1842,
"loss/crossentropy": 2.6813217401504517,
"loss/hidden": 0.0,
"loss/logits": 0.18416454643011093,
"loss/reg": 3.1269419193267822,
"step": 330
},
{
"epoch": 0.00331,
"grad_norm": 0.5757820010185242,
"grad_norm_var": 0.004098500311938921,
"learning_rate": 5e-05,
"loss": 0.1935,
"loss/crossentropy": 2.9679067730903625,
"loss/hidden": 0.0,
"loss/logits": 0.19354857876896858,
"loss/reg": 3.1228113174438477,
"step": 331
},
{
"epoch": 0.00332,
"grad_norm": 0.405617356300354,
"grad_norm_var": 0.004102514647771457,
"learning_rate": 5e-05,
"loss": 0.1716,
"loss/crossentropy": 2.843691408634186,
"loss/hidden": 0.0,
"loss/logits": 0.1716487891972065,
"loss/reg": 3.120131015777588,
"step": 332
},
{
"epoch": 0.00333,
"grad_norm": 0.3825243413448334,
"grad_norm_var": 0.004114939464605513,
"learning_rate": 5e-05,
"loss": 0.1692,
"loss/crossentropy": 2.722069561481476,
"loss/hidden": 0.0,
"loss/logits": 0.16915880143642426,
"loss/reg": 3.117812395095825,
"step": 333
},
{
"epoch": 0.00334,
"grad_norm": 0.38414397835731506,
"grad_norm_var": 0.004087504594686679,
"learning_rate": 5e-05,
"loss": 0.1678,
"loss/crossentropy": 2.727014124393463,
"loss/hidden": 0.0,
"loss/logits": 0.1677936352789402,
"loss/reg": 3.115847587585449,
"step": 334
},
{
"epoch": 0.00335,
"grad_norm": 0.4531976580619812,
"grad_norm_var": 0.0040868556651997325,
"learning_rate": 5e-05,
"loss": 0.1726,
"loss/crossentropy": 2.5817691683769226,
"loss/hidden": 0.0,
"loss/logits": 0.17255331575870514,
"loss/reg": 3.1140594482421875,
"step": 335
},
{
"epoch": 0.00336,
"grad_norm": 0.33963721990585327,
"grad_norm_var": 0.004259094561609115,
"learning_rate": 5e-05,
"loss": 0.1605,
"loss/crossentropy": 2.7009602189064026,
"loss/hidden": 0.0,
"loss/logits": 0.16050074249505997,
"loss/reg": 3.1128299236297607,
"step": 336
},
{
"epoch": 0.00337,
"grad_norm": 0.36085280776023865,
"grad_norm_var": 0.00419878945557195,
"learning_rate": 5e-05,
"loss": 0.1641,
"loss/crossentropy": 2.7012510299682617,
"loss/hidden": 0.0,
"loss/logits": 0.16410352289676666,
"loss/reg": 3.109898805618286,
"step": 337
},
{
"epoch": 0.00338,
"grad_norm": 0.3331363797187805,
"grad_norm_var": 0.003666565441549352,
"learning_rate": 5e-05,
"loss": 0.1619,
"loss/crossentropy": 2.802642047405243,
"loss/hidden": 0.0,
"loss/logits": 0.1618719846010208,
"loss/reg": 3.1067681312561035,
"step": 338
},
{
"epoch": 0.00339,
"grad_norm": 0.41531723737716675,
"grad_norm_var": 0.003696375336840474,
"learning_rate": 5e-05,
"loss": 0.1877,
"loss/crossentropy": 2.622368335723877,
"loss/hidden": 0.0,
"loss/logits": 0.18770882859826088,
"loss/reg": 3.103231191635132,
"step": 339
},
{
"epoch": 0.0034,
"grad_norm": 0.3483443260192871,
"grad_norm_var": 0.0037197436901762657,
"learning_rate": 5e-05,
"loss": 0.1603,
"loss/crossentropy": 2.7523834109306335,
"loss/hidden": 0.0,
"loss/logits": 0.16034872457385063,
"loss/reg": 3.1008808612823486,
"step": 340
},
{
"epoch": 0.00341,
"grad_norm": 0.46117284893989563,
"grad_norm_var": 0.003961845355148208,
"learning_rate": 5e-05,
"loss": 0.1801,
"loss/crossentropy": 2.820544958114624,
"loss/hidden": 0.0,
"loss/logits": 0.18007208034396172,
"loss/reg": 3.098954439163208,
"step": 341
},
{
"epoch": 0.00342,
"grad_norm": 0.38020917773246765,
"grad_norm_var": 0.003918987421685794,
"learning_rate": 5e-05,
"loss": 0.1711,
"loss/crossentropy": 2.864526093006134,
"loss/hidden": 0.0,
"loss/logits": 0.1710633859038353,
"loss/reg": 3.0957016944885254,
"step": 342
},
{
"epoch": 0.00343,
"grad_norm": 0.3978392779827118,
"grad_norm_var": 0.0037065638898653073,
"learning_rate": 5e-05,
"loss": 0.1782,
"loss/crossentropy": 2.859494388103485,
"loss/hidden": 0.0,
"loss/logits": 0.1782137230038643,
"loss/reg": 3.09333872795105,
"step": 343
},
{
"epoch": 0.00344,
"grad_norm": 0.36975786089897156,
"grad_norm_var": 0.0036926924317912187,
"learning_rate": 5e-05,
"loss": 0.1658,
"loss/crossentropy": 2.7827839255332947,
"loss/hidden": 0.0,
"loss/logits": 0.1657763496041298,
"loss/reg": 3.090590715408325,
"step": 344
},
{
"epoch": 0.00345,
"grad_norm": 0.333897203207016,
"grad_norm_var": 0.0037974520830182084,
"learning_rate": 5e-05,
"loss": 0.1612,
"loss/crossentropy": 2.7804853320121765,
"loss/hidden": 0.0,
"loss/logits": 0.1611798331141472,
"loss/reg": 3.0877137184143066,
"step": 345
},
{
"epoch": 0.00346,
"grad_norm": 0.43794891238212585,
"grad_norm_var": 0.0038469119530984567,
"learning_rate": 5e-05,
"loss": 0.1678,
"loss/crossentropy": 2.7229984402656555,
"loss/hidden": 0.0,
"loss/logits": 0.1677897423505783,
"loss/reg": 3.085155725479126,
"step": 346
},
{
"epoch": 0.00347,
"grad_norm": 0.33257824182510376,
"grad_norm_var": 0.0018017603976316725,
"learning_rate": 5e-05,
"loss": 0.1745,
"loss/crossentropy": 2.7936434745788574,
"loss/hidden": 0.0,
"loss/logits": 0.17448442056775093,
"loss/reg": 3.0829925537109375,
"step": 347
},
{
"epoch": 0.00348,
"grad_norm": 0.393646240234375,
"grad_norm_var": 0.001775431972661142,
"learning_rate": 5e-05,
"loss": 0.1647,
"loss/crossentropy": 2.8590177297592163,
"loss/hidden": 0.0,
"loss/logits": 0.16474304348230362,
"loss/reg": 3.080383062362671,
"step": 348
},
{
"epoch": 0.00349,
"grad_norm": 0.34549105167388916,
"grad_norm_var": 0.0018623256252658482,
"learning_rate": 5e-05,
"loss": 0.1678,
"loss/crossentropy": 2.7182729840278625,
"loss/hidden": 0.0,
"loss/logits": 0.16776488721370697,
"loss/reg": 3.0792622566223145,
"step": 349
},
{
"epoch": 0.0035,
"grad_norm": 0.9833559393882751,
"grad_norm_var": 0.024598539346201563,
"learning_rate": 5e-05,
"loss": 0.1911,
"loss/crossentropy": 2.720784068107605,
"loss/hidden": 0.0,
"loss/logits": 0.19106518849730492,
"loss/reg": 3.0780370235443115,
"step": 350
},
{
"epoch": 0.00351,
"grad_norm": 0.3550430238246918,
"grad_norm_var": 0.02473872020472854,
"learning_rate": 5e-05,
"loss": 0.1683,
"loss/crossentropy": 2.7388776540756226,
"loss/hidden": 0.0,
"loss/logits": 0.16827991232275963,
"loss/reg": 3.07700777053833,
"step": 351
},
{
"epoch": 0.00352,
"grad_norm": 0.32236865162849426,
"grad_norm_var": 0.024923428623413246,
"learning_rate": 5e-05,
"loss": 0.1555,
"loss/crossentropy": 2.787019371986389,
"loss/hidden": 0.0,
"loss/logits": 0.1554781049489975,
"loss/reg": 3.075169324874878,
"step": 352
},
{
"epoch": 0.00353,
"grad_norm": 0.34089383482933044,
"grad_norm_var": 0.025080939274787682,
"learning_rate": 5e-05,
"loss": 0.1625,
"loss/crossentropy": 2.763973832130432,
"loss/hidden": 0.0,
"loss/logits": 0.1625489443540573,
"loss/reg": 3.0732734203338623,
"step": 353
},
{
"epoch": 0.00354,
"grad_norm": 0.35467466711997986,
"grad_norm_var": 0.02489081345717287,
"learning_rate": 5e-05,
"loss": 0.161,
"loss/crossentropy": 2.6696255207061768,
"loss/hidden": 0.0,
"loss/logits": 0.1610397771000862,
"loss/reg": 3.071441411972046,
"step": 354
},
{
"epoch": 0.00355,
"grad_norm": 0.3465348184108734,
"grad_norm_var": 0.02514492485323772,
"learning_rate": 5e-05,
"loss": 0.1692,
"loss/crossentropy": 2.8481903076171875,
"loss/hidden": 0.0,
"loss/logits": 0.16923030093312263,
"loss/reg": 3.068796157836914,
"step": 355
},
{
"epoch": 0.00356,
"grad_norm": 0.3337138891220093,
"grad_norm_var": 0.025271718941650815,
"learning_rate": 5e-05,
"loss": 0.1675,
"loss/crossentropy": 2.734935760498047,
"loss/hidden": 0.0,
"loss/logits": 0.1674855425953865,
"loss/reg": 3.066350221633911,
"step": 356
},
{
"epoch": 0.00357,
"grad_norm": 0.3486330509185791,
"grad_norm_var": 0.02522896182872459,
"learning_rate": 5e-05,
"loss": 0.1751,
"loss/crossentropy": 2.8282878398895264,
"loss/hidden": 0.0,
"loss/logits": 0.17508375644683838,
"loss/reg": 3.0633251667022705,
"step": 357
},
{
"epoch": 0.00358,
"grad_norm": 0.3714129626750946,
"grad_norm_var": 0.025255292610223575,
"learning_rate": 5e-05,
"loss": 0.1798,
"loss/crossentropy": 2.723433256149292,
"loss/hidden": 0.0,
"loss/logits": 0.17982058972120285,
"loss/reg": 3.060797929763794,
"step": 358
},
{
"epoch": 0.00359,
"grad_norm": 0.38819992542266846,
"grad_norm_var": 0.025261289598676597,
"learning_rate": 5e-05,
"loss": 0.179,
"loss/crossentropy": 2.5971017479896545,
"loss/hidden": 0.0,
"loss/logits": 0.17904112860560417,
"loss/reg": 3.057884693145752,
"step": 359
},
{
"epoch": 0.0036,
"grad_norm": 0.3948271870613098,
"grad_norm_var": 0.02520822524046924,
"learning_rate": 5e-05,
"loss": 0.1826,
"loss/crossentropy": 2.631825864315033,
"loss/hidden": 0.0,
"loss/logits": 0.1826026625931263,
"loss/reg": 3.0550246238708496,
"step": 360
},
{
"epoch": 0.00361,
"grad_norm": 0.46747469902038574,
"grad_norm_var": 0.025164775676019657,
"learning_rate": 5e-05,
"loss": 0.1849,
"loss/crossentropy": 2.628718376159668,
"loss/hidden": 0.0,
"loss/logits": 0.1848563477396965,
"loss/reg": 3.052072525024414,
"step": 361
},
{
"epoch": 0.00362,
"grad_norm": 0.9672635197639465,
"grad_norm_var": 0.044838716189940266,
"learning_rate": 5e-05,
"loss": 0.2236,
"loss/crossentropy": 2.7798518538475037,
"loss/hidden": 0.0,
"loss/logits": 0.22359847277402878,
"loss/reg": 3.048891305923462,
"step": 362
},
{
"epoch": 0.00363,
"grad_norm": 0.3722783029079437,
"grad_norm_var": 0.04436658011174813,
"learning_rate": 5e-05,
"loss": 0.175,
"loss/crossentropy": 2.7808294892311096,
"loss/hidden": 0.0,
"loss/logits": 0.17498808726668358,
"loss/reg": 3.0457394123077393,
"step": 363
},
{
"epoch": 0.00364,
"grad_norm": 0.3547132611274719,
"grad_norm_var": 0.04471680473078544,
"learning_rate": 5e-05,
"loss": 0.1602,
"loss/crossentropy": 2.7656018137931824,
"loss/hidden": 0.0,
"loss/logits": 0.16020696610212326,
"loss/reg": 3.043097734451294,
"step": 364
},
{
"epoch": 0.00365,
"grad_norm": 0.4774816632270813,
"grad_norm_var": 0.04413484451680517,
"learning_rate": 5e-05,
"loss": 0.1831,
"loss/crossentropy": 2.9051772356033325,
"loss/hidden": 0.0,
"loss/logits": 0.18311960250139236,
"loss/reg": 3.0403990745544434,
"step": 365
},
{
"epoch": 0.00366,
"grad_norm": 0.41332709789276123,
"grad_norm_var": 0.0238056716485936,
"learning_rate": 5e-05,
"loss": 0.1705,
"loss/crossentropy": 2.8529672026634216,
"loss/hidden": 0.0,
"loss/logits": 0.17049206793308258,
"loss/reg": 3.037370204925537,
"step": 366
},
{
"epoch": 0.00367,
"grad_norm": 0.39109355211257935,
"grad_norm_var": 0.023608062717162412,
"learning_rate": 5e-05,
"loss": 0.1798,
"loss/crossentropy": 2.796768307685852,
"loss/hidden": 0.0,
"loss/logits": 0.17983367666602135,
"loss/reg": 3.0343563556671143,
"step": 367
},
{
"epoch": 0.00368,
"grad_norm": 0.36531057953834534,
"grad_norm_var": 0.023191193861390014,
"learning_rate": 5e-05,
"loss": 0.1715,
"loss/crossentropy": 2.8276050686836243,
"loss/hidden": 0.0,
"loss/logits": 0.1715383380651474,
"loss/reg": 3.031611919403076,
"step": 368
},
{
"epoch": 0.00369,
"grad_norm": 0.33283501863479614,
"grad_norm_var": 0.023278092934366657,
"learning_rate": 5e-05,
"loss": 0.1499,
"loss/crossentropy": 2.6641258597373962,
"loss/hidden": 0.0,
"loss/logits": 0.14990831911563873,
"loss/reg": 3.0287797451019287,
"step": 369
},
{
"epoch": 0.0037,
"grad_norm": 0.4542810618877411,
"grad_norm_var": 0.023063995994232415,
"learning_rate": 5e-05,
"loss": 0.1722,
"loss/crossentropy": 2.7453941702842712,
"loss/hidden": 0.0,
"loss/logits": 0.1721569411456585,
"loss/reg": 3.0258800983428955,
"step": 370
},
{
"epoch": 0.00371,
"grad_norm": 0.3705763816833496,
"grad_norm_var": 0.022852728398777448,
"learning_rate": 5e-05,
"loss": 0.1849,
"loss/crossentropy": 2.635721504688263,
"loss/hidden": 0.0,
"loss/logits": 0.18491211906075478,
"loss/reg": 3.0235960483551025,
"step": 371
},
{
"epoch": 0.00372,
"grad_norm": 0.4085729420185089,
"grad_norm_var": 0.022289690361487074,
"learning_rate": 5e-05,
"loss": 0.1871,
"loss/crossentropy": 2.732766628265381,
"loss/hidden": 0.0,
"loss/logits": 0.18710973486304283,
"loss/reg": 3.0203185081481934,
"step": 372
},
{
"epoch": 0.00373,
"grad_norm": 0.3334612250328064,
"grad_norm_var": 0.022468457594482887,
"learning_rate": 5e-05,
"loss": 0.1662,
"loss/crossentropy": 2.783429443836212,
"loss/hidden": 0.0,
"loss/logits": 0.1662071831524372,
"loss/reg": 3.0171730518341064,
"step": 373
},
{
"epoch": 0.00374,
"grad_norm": 0.35536903142929077,
"grad_norm_var": 0.022607616164545874,
"learning_rate": 5e-05,
"loss": 0.1654,
"loss/crossentropy": 2.818749785423279,
"loss/hidden": 0.0,
"loss/logits": 0.16541225090622902,
"loss/reg": 3.0140058994293213,
"step": 374
},
{
"epoch": 0.00375,
"grad_norm": 0.348376989364624,
"grad_norm_var": 0.022917750109528078,
"learning_rate": 5e-05,
"loss": 0.1628,
"loss/crossentropy": 2.9099320769309998,
"loss/hidden": 0.0,
"loss/logits": 0.1627991460263729,
"loss/reg": 3.010875701904297,
"step": 375
},
{
"epoch": 0.00376,
"grad_norm": 0.3394787311553955,
"grad_norm_var": 0.023335225496050292,
"learning_rate": 5e-05,
"loss": 0.1605,
"loss/crossentropy": 2.811407744884491,
"loss/hidden": 0.0,
"loss/logits": 0.16051743179559708,
"loss/reg": 3.0073623657226562,
"step": 376
},
{
"epoch": 0.00377,
"grad_norm": 0.42454567551612854,
"grad_norm_var": 0.02319007765550817,
"learning_rate": 5e-05,
"loss": 0.1645,
"loss/crossentropy": 2.773725748062134,
"loss/hidden": 0.0,
"loss/logits": 0.16450630128383636,
"loss/reg": 3.004185676574707,
"step": 377
},
{
"epoch": 0.00378,
"grad_norm": 0.3412385582923889,
"grad_norm_var": 0.001946629707866813,
"learning_rate": 5e-05,
"loss": 0.1742,
"loss/crossentropy": 2.785709500312805,
"loss/hidden": 0.0,
"loss/logits": 0.17420916631817818,
"loss/reg": 3.0007028579711914,
"step": 378
},
{
"epoch": 0.00379,
"grad_norm": 0.3544103503227234,
"grad_norm_var": 0.001985417588834304,
"learning_rate": 5e-05,
"loss": 0.1716,
"loss/crossentropy": 2.670408546924591,
"loss/hidden": 0.0,
"loss/logits": 0.17155754566192627,
"loss/reg": 2.9972825050354004,
"step": 379
},
{
"epoch": 0.0038,
"grad_norm": 0.36286091804504395,
"grad_norm_var": 0.001963109812450625,
"learning_rate": 5e-05,
"loss": 0.1786,
"loss/crossentropy": 2.822770357131958,
"loss/hidden": 0.0,
"loss/logits": 0.17864727228879929,
"loss/reg": 2.993536949157715,
"step": 380
},
{
"epoch": 0.00381,
"grad_norm": 0.5003440976142883,
"grad_norm_var": 0.002294225514870618,
"learning_rate": 5e-05,
"loss": 0.1896,
"loss/crossentropy": 2.790800392627716,
"loss/hidden": 0.0,
"loss/logits": 0.18961890786886215,
"loss/reg": 2.990283489227295,
"step": 381
},
{
"epoch": 0.00382,
"grad_norm": 0.3698294758796692,
"grad_norm_var": 0.0022250210916228584,
"learning_rate": 5e-05,
"loss": 0.1648,
"loss/crossentropy": 2.8308547139167786,
"loss/hidden": 0.0,
"loss/logits": 0.16477400809526443,
"loss/reg": 2.986691474914551,
"step": 382
},
{
"epoch": 0.00383,
"grad_norm": 0.36506953835487366,
"grad_norm_var": 0.0022229105132923347,
"learning_rate": 5e-05,
"loss": 0.1682,
"loss/crossentropy": 2.744426727294922,
"loss/hidden": 0.0,
"loss/logits": 0.16819821670651436,
"loss/reg": 2.983008861541748,
"step": 383
},
{
"epoch": 0.00384,
"grad_norm": 0.3243113160133362,
"grad_norm_var": 0.002390011819316588,
"learning_rate": 5e-05,
"loss": 0.1596,
"loss/crossentropy": 2.89188152551651,
"loss/hidden": 0.0,
"loss/logits": 0.1595698669552803,
"loss/reg": 2.979191541671753,
"step": 384
},
{
"epoch": 0.00385,
"grad_norm": 0.36836785078048706,
"grad_norm_var": 0.002273433106164295,
"learning_rate": 5e-05,
"loss": 0.1747,
"loss/crossentropy": 2.902570128440857,
"loss/hidden": 0.0,
"loss/logits": 0.1746898777782917,
"loss/reg": 2.975698947906494,
"step": 385
},
{
"epoch": 0.00386,
"grad_norm": 0.3365325927734375,
"grad_norm_var": 0.001915978849994604,
"learning_rate": 5e-05,
"loss": 0.1608,
"loss/crossentropy": 2.766001045703888,
"loss/hidden": 0.0,
"loss/logits": 0.16081618145108223,
"loss/reg": 2.972667694091797,
"step": 386
},
{
"epoch": 0.00387,
"grad_norm": 0.35417604446411133,
"grad_norm_var": 0.0019292530227877358,
"learning_rate": 5e-05,
"loss": 0.1605,
"loss/crossentropy": 2.7814798951148987,
"loss/hidden": 0.0,
"loss/logits": 0.16046970710158348,
"loss/reg": 2.969967842102051,
"step": 387
},
{
"epoch": 0.00388,
"grad_norm": 1.9537514448165894,
"grad_norm_var": 0.15952536292831518,
"learning_rate": 5e-05,
"loss": 0.1926,
"loss/crossentropy": 2.782427728176117,
"loss/hidden": 0.0,
"loss/logits": 0.1925731934607029,
"loss/reg": 2.9671437740325928,
"step": 388
},
{
"epoch": 0.00389,
"grad_norm": 0.3620118498802185,
"grad_norm_var": 0.15907744774636304,
"learning_rate": 5e-05,
"loss": 0.1714,
"loss/crossentropy": 2.716952919960022,
"loss/hidden": 0.0,
"loss/logits": 0.17136194929480553,
"loss/reg": 2.9639716148376465,
"step": 389
},
{
"epoch": 0.0039,
"grad_norm": 0.3765539526939392,
"grad_norm_var": 0.1587921781193889,
"learning_rate": 5e-05,
"loss": 0.1689,
"loss/crossentropy": 2.8395472168922424,
"loss/hidden": 0.0,
"loss/logits": 0.16893662884831429,
"loss/reg": 2.9617481231689453,
"step": 390
},
{
"epoch": 0.00391,
"grad_norm": 0.39779096841812134,
"grad_norm_var": 0.15815917569478716,
"learning_rate": 5e-05,
"loss": 0.1677,
"loss/crossentropy": 2.813215434551239,
"loss/hidden": 0.0,
"loss/logits": 0.1676802597939968,
"loss/reg": 2.958872079849243,
"step": 391
},
{
"epoch": 0.00392,
"grad_norm": 2.267273187637329,
"grad_norm_var": 0.35670344578828533,
"learning_rate": 5e-05,
"loss": 0.1898,
"loss/crossentropy": 2.710484206676483,
"loss/hidden": 0.0,
"loss/logits": 0.18981827050447464,
"loss/reg": 2.9562907218933105,
"step": 392
},
{
"epoch": 0.00393,
"grad_norm": 0.4732659161090851,
"grad_norm_var": 0.35576926148027355,
"learning_rate": 5e-05,
"loss": 0.1789,
"loss/crossentropy": 2.7463297247886658,
"loss/hidden": 0.0,
"loss/logits": 0.17890166491270065,
"loss/reg": 2.953505277633667,
"step": 393
},
{
"epoch": 0.00394,
"grad_norm": 0.46487849950790405,
"grad_norm_var": 0.3525539310677323,
"learning_rate": 5e-05,
"loss": 0.1737,
"loss/crossentropy": 2.781617820262909,
"loss/hidden": 0.0,
"loss/logits": 0.17365656793117523,
"loss/reg": 2.951185464859009,
"step": 394
},
{
"epoch": 0.00395,
"grad_norm": 0.36613309383392334,
"grad_norm_var": 0.352175585204308,
"learning_rate": 5e-05,
"loss": 0.1749,
"loss/crossentropy": 2.9521047472953796,
"loss/hidden": 0.0,
"loss/logits": 0.1749158501625061,
"loss/reg": 2.949521780014038,
"step": 395
},
{
"epoch": 0.00396,
"grad_norm": 0.33889761567115784,
"grad_norm_var": 0.35297777688562104,
"learning_rate": 5e-05,
"loss": 0.1611,
"loss/crossentropy": 2.709399461746216,
"loss/hidden": 0.0,
"loss/logits": 0.161102045327425,
"loss/reg": 2.94758677482605,
"step": 396
},
{
"epoch": 0.00397,
"grad_norm": 0.37460586428642273,
"grad_norm_var": 0.3556567542520952,
"learning_rate": 5e-05,
"loss": 0.1709,
"loss/crossentropy": 2.8318939208984375,
"loss/hidden": 0.0,
"loss/logits": 0.1709057316184044,
"loss/reg": 2.9449052810668945,
"step": 397
},
{
"epoch": 0.00398,
"grad_norm": 0.36912715435028076,
"grad_norm_var": 0.3556777153015602,
"learning_rate": 5e-05,
"loss": 0.1699,
"loss/crossentropy": 2.7699413895606995,
"loss/hidden": 0.0,
"loss/logits": 0.16991987824440002,
"loss/reg": 2.9416375160217285,
"step": 398
},
{
"epoch": 0.00399,
"grad_norm": 0.4335935711860657,
"grad_norm_var": 0.35388598085202433,
"learning_rate": 5e-05,
"loss": 0.1621,
"loss/crossentropy": 2.6929262280464172,
"loss/hidden": 0.0,
"loss/logits": 0.16208457946777344,
"loss/reg": 2.9385571479797363,
"step": 399
},
{
"epoch": 0.004,
"grad_norm": 0.36466526985168457,
"grad_norm_var": 0.35251743192284957,
"learning_rate": 5e-05,
"loss": 0.1603,
"loss/crossentropy": 2.9028329849243164,
"loss/hidden": 0.0,
"loss/logits": 0.16026458516716957,
"loss/reg": 2.935270071029663,
"step": 400
},
{
"epoch": 0.00401,
"grad_norm": 0.31859657168388367,
"grad_norm_var": 0.3542100800677372,
"learning_rate": 5e-05,
"loss": 0.155,
"loss/crossentropy": 2.7707905769348145,
"loss/hidden": 0.0,
"loss/logits": 0.15500668808817863,
"loss/reg": 2.9319961071014404,
"step": 401
},
{
"epoch": 0.00402,
"grad_norm": 0.39714375138282776,
"grad_norm_var": 0.35233479687143326,
"learning_rate": 5e-05,
"loss": 0.1722,
"loss/crossentropy": 2.5947351455688477,
"loss/hidden": 0.0,
"loss/logits": 0.17224042490124702,
"loss/reg": 2.929413318634033,
"step": 402
},
{
"epoch": 0.00403,
"grad_norm": 2.267681121826172,
"grad_norm_var": 0.518261838886723,
"learning_rate": 5e-05,
"loss": 0.2313,
"loss/crossentropy": 2.703626811504364,
"loss/hidden": 0.0,
"loss/logits": 0.23128759488463402,
"loss/reg": 2.925968647003174,
"step": 403
},
{
"epoch": 0.00404,
"grad_norm": 0.38179340958595276,
"grad_norm_var": 0.414193396025083,
"learning_rate": 5e-05,
"loss": 0.1852,
"loss/crossentropy": 2.820598065853119,
"loss/hidden": 0.0,
"loss/logits": 0.18524771928787231,
"loss/reg": 2.9223124980926514,
"step": 404
},
{
"epoch": 0.00405,
"grad_norm": 0.35225149989128113,
"grad_norm_var": 0.4145378570625937,
"learning_rate": 5e-05,
"loss": 0.1672,
"loss/crossentropy": 2.761472165584564,
"loss/hidden": 0.0,
"loss/logits": 0.1672290712594986,
"loss/reg": 2.9187073707580566,
"step": 405
},
{
"epoch": 0.00406,
"grad_norm": 0.35987603664398193,
"grad_norm_var": 0.4150999685132215,
"learning_rate": 5e-05,
"loss": 0.1731,
"loss/crossentropy": 2.7906153202056885,
"loss/hidden": 0.0,
"loss/logits": 0.17313192784786224,
"loss/reg": 2.915867805480957,
"step": 406
},
{
"epoch": 0.00407,
"grad_norm": 0.36714085936546326,
"grad_norm_var": 0.416068714723823,
"learning_rate": 5e-05,
"loss": 0.1729,
"loss/crossentropy": 2.7815486192703247,
"loss/hidden": 0.0,
"loss/logits": 0.17289156094193459,
"loss/reg": 2.9124350547790527,
"step": 407
},
{
"epoch": 0.00408,
"grad_norm": 0.43249061703681946,
"grad_norm_var": 0.22313248530400895,
"learning_rate": 5e-05,
"loss": 0.1886,
"loss/crossentropy": 2.7781424522399902,
"loss/hidden": 0.0,
"loss/logits": 0.18864024803042412,
"loss/reg": 2.9094510078430176,
"step": 408
},
{
"epoch": 0.00409,
"grad_norm": 0.34418484568595886,
"grad_norm_var": 0.22470081409508588,
"learning_rate": 5e-05,
"loss": 0.1713,
"loss/crossentropy": 2.7818892002105713,
"loss/hidden": 0.0,
"loss/logits": 0.17131192237138748,
"loss/reg": 2.9064505100250244,
"step": 409
},
{
"epoch": 0.0041,
"grad_norm": 0.5792982578277588,
"grad_norm_var": 0.2250470715469535,
"learning_rate": 5e-05,
"loss": 0.1925,
"loss/crossentropy": 2.8059155344963074,
"loss/hidden": 0.0,
"loss/logits": 0.1925133354961872,
"loss/reg": 2.9032788276672363,
"step": 410
},
{
"epoch": 0.00411,
"grad_norm": 0.35917767882347107,
"grad_norm_var": 0.22517699381034958,
"learning_rate": 5e-05,
"loss": 0.1597,
"loss/crossentropy": 2.9948400259017944,
"loss/hidden": 0.0,
"loss/logits": 0.1597156822681427,
"loss/reg": 2.900125741958618,
"step": 411
},
{
"epoch": 0.00412,
"grad_norm": 0.32936394214630127,
"grad_norm_var": 0.2253906803631866,
"learning_rate": 5e-05,
"loss": 0.16,
"loss/crossentropy": 2.8464353680610657,
"loss/hidden": 0.0,
"loss/logits": 0.15996254980564117,
"loss/reg": 2.897120952606201,
"step": 412
},
{
"epoch": 0.00413,
"grad_norm": 0.3636591136455536,
"grad_norm_var": 0.22558401797348096,
"learning_rate": 5e-05,
"loss": 0.1896,
"loss/crossentropy": 2.5940242409706116,
"loss/hidden": 0.0,
"loss/logits": 0.18961919099092484,
"loss/reg": 2.8938040733337402,
"step": 413
},
{
"epoch": 0.00414,
"grad_norm": 0.3614409565925598,
"grad_norm_var": 0.2257231161008428,
"learning_rate": 5e-05,
"loss": 0.1669,
"loss/crossentropy": 2.9938586950302124,
"loss/hidden": 0.0,
"loss/logits": 0.16688520461320877,
"loss/reg": 2.891237497329712,
"step": 414
},
{
"epoch": 0.00415,
"grad_norm": 0.33793795108795166,
"grad_norm_var": 0.2271517945765009,
"learning_rate": 5e-05,
"loss": 0.1662,
"loss/crossentropy": 2.7566803693771362,
"loss/hidden": 0.0,
"loss/logits": 0.16617370769381523,
"loss/reg": 2.888444185256958,
"step": 415
},
{
"epoch": 0.00416,
"grad_norm": 0.33697640895843506,
"grad_norm_var": 0.22768012665927795,
"learning_rate": 5e-05,
"loss": 0.1609,
"loss/crossentropy": 2.7538956999778748,
"loss/hidden": 0.0,
"loss/logits": 0.1609276346862316,
"loss/reg": 2.8854165077209473,
"step": 416
},
{
"epoch": 0.00417,
"grad_norm": 0.33163169026374817,
"grad_norm_var": 0.22738752034767185,
"learning_rate": 5e-05,
"loss": 0.1686,
"loss/crossentropy": 2.716135025024414,
"loss/hidden": 0.0,
"loss/logits": 0.16858552396297455,
"loss/reg": 2.883502244949341,
"step": 417
},
{
"epoch": 0.00418,
"grad_norm": 0.3197973072528839,
"grad_norm_var": 0.22875903165210631,
"learning_rate": 5e-05,
"loss": 0.1622,
"loss/crossentropy": 2.8385114669799805,
"loss/hidden": 0.0,
"loss/logits": 0.16216163337230682,
"loss/reg": 2.8822600841522217,
"step": 418
},
{
"epoch": 0.00419,
"grad_norm": 0.3929068446159363,
"grad_norm_var": 0.0038269076397950408,
"learning_rate": 5e-05,
"loss": 0.1747,
"loss/crossentropy": 2.7871418595314026,
"loss/hidden": 0.0,
"loss/logits": 0.17471741139888763,
"loss/reg": 2.8793692588806152,
"step": 419
},
{
"epoch": 0.0042,
"grad_norm": 0.3870161473751068,
"grad_norm_var": 0.0038355224442556198,
"learning_rate": 5e-05,
"loss": 0.1704,
"loss/crossentropy": 2.7993595600128174,
"loss/hidden": 0.0,
"loss/logits": 0.1703585907816887,
"loss/reg": 2.8776159286499023,
"step": 420
},
{
"epoch": 0.00421,
"grad_norm": 0.35682201385498047,
"grad_norm_var": 0.0038246732894099337,
"learning_rate": 5e-05,
"loss": 0.1886,
"loss/crossentropy": 2.655856966972351,
"loss/hidden": 0.0,
"loss/logits": 0.18859218060970306,
"loss/reg": 2.87612247467041,
"step": 421
},
{
"epoch": 0.00422,
"grad_norm": 0.33115604519844055,
"grad_norm_var": 0.003924500155300174,
"learning_rate": 5e-05,
"loss": 0.1611,
"loss/crossentropy": 2.8695462942123413,
"loss/hidden": 0.0,
"loss/logits": 0.1611364483833313,
"loss/reg": 2.873897075653076,
"step": 422
},
{
"epoch": 0.00423,
"grad_norm": 0.4912989139556885,
"grad_norm_var": 0.004829238325957341,
"learning_rate": 5e-05,
"loss": 0.1784,
"loss/crossentropy": 2.7167177200317383,
"loss/hidden": 0.0,
"loss/logits": 0.17839327454566956,
"loss/reg": 2.8718714714050293,
"step": 423
},
{
"epoch": 0.00424,
"grad_norm": 0.3349898159503937,
"grad_norm_var": 0.004720821391959795,
"learning_rate": 5e-05,
"loss": 0.1615,
"loss/crossentropy": 2.7473002076148987,
"loss/hidden": 0.0,
"loss/logits": 0.1615452691912651,
"loss/reg": 2.8705270290374756,
"step": 424
},
{
"epoch": 0.00425,
"grad_norm": 0.3930635154247284,
"grad_norm_var": 0.004686561363231038,
"learning_rate": 5e-05,
"loss": 0.1775,
"loss/crossentropy": 2.785146713256836,
"loss/hidden": 0.0,
"loss/logits": 0.17748162522912025,
"loss/reg": 2.868584394454956,
"step": 425
},
{
"epoch": 0.00426,
"grad_norm": 0.3448260426521301,
"grad_norm_var": 0.0017484410160554464,
"learning_rate": 5e-05,
"loss": 0.1758,
"loss/crossentropy": 2.726165235042572,
"loss/hidden": 0.0,
"loss/logits": 0.17576807364821434,
"loss/reg": 2.8673741817474365,
"step": 426
},
{
"epoch": 0.00427,
"grad_norm": 0.3846610188484192,
"grad_norm_var": 0.0017836724819081718,
"learning_rate": 5e-05,
"loss": 0.1762,
"loss/crossentropy": 2.7086002230644226,
"loss/hidden": 0.0,
"loss/logits": 0.1762431263923645,
"loss/reg": 2.8651485443115234,
"step": 427
},
{
"epoch": 0.00428,
"grad_norm": 0.3494791090488434,
"grad_norm_var": 0.0017205006490997802,
"learning_rate": 5e-05,
"loss": 0.1818,
"loss/crossentropy": 2.7305288314819336,
"loss/hidden": 0.0,
"loss/logits": 0.18181117624044418,
"loss/reg": 2.8629541397094727,
"step": 428
},
{
"epoch": 0.00429,
"grad_norm": 0.3337409794330597,
"grad_norm_var": 0.0017762239427149495,
"learning_rate": 5e-05,
"loss": 0.1661,
"loss/crossentropy": 2.840167284011841,
"loss/hidden": 0.0,
"loss/logits": 0.1660769209265709,
"loss/reg": 2.8615217208862305,
"step": 429
},
{
"epoch": 0.0043,
"grad_norm": 0.4685284495353699,
"grad_norm_var": 0.0024887722894077887,
"learning_rate": 5e-05,
"loss": 0.1741,
"loss/crossentropy": 2.7593295574188232,
"loss/hidden": 0.0,
"loss/logits": 0.1740834154188633,
"loss/reg": 2.860365629196167,
"step": 430
},
{
"epoch": 0.00431,
"grad_norm": 0.35651838779449463,
"grad_norm_var": 0.002434815976952542,
"learning_rate": 5e-05,
"loss": 0.1673,
"loss/crossentropy": 2.777701735496521,
"loss/hidden": 0.0,
"loss/logits": 0.16725125908851624,
"loss/reg": 2.858222007751465,
"step": 431
},
{
"epoch": 0.00432,
"grad_norm": 0.34670454263687134,
"grad_norm_var": 0.0023984303943363817,
"learning_rate": 5e-05,
"loss": 0.165,
"loss/crossentropy": 2.749099850654602,
"loss/hidden": 0.0,
"loss/logits": 0.16504037007689476,
"loss/reg": 2.855973958969116,
"step": 432
},
{
"epoch": 0.00433,
"grad_norm": 0.3284713923931122,
"grad_norm_var": 0.0024153046998328874,
"learning_rate": 5e-05,
"loss": 0.1521,
"loss/crossentropy": 2.7869237661361694,
"loss/hidden": 0.0,
"loss/logits": 0.15208067372441292,
"loss/reg": 2.853942632675171,
"step": 433
},
{
"epoch": 0.00434,
"grad_norm": 0.48883649706840515,
"grad_norm_var": 0.0030697262784900037,
"learning_rate": 5e-05,
"loss": 0.1759,
"loss/crossentropy": 2.9403671622276306,
"loss/hidden": 0.0,
"loss/logits": 0.17589127644896507,
"loss/reg": 2.851728916168213,
"step": 434
},
{
"epoch": 0.00435,
"grad_norm": 0.36951273679733276,
"grad_norm_var": 0.0030654307324534447,
"learning_rate": 5e-05,
"loss": 0.185,
"loss/crossentropy": 2.7797312140464783,
"loss/hidden": 0.0,
"loss/logits": 0.1850355602800846,
"loss/reg": 2.8488640785217285,
"step": 435
},
{
"epoch": 0.00436,
"grad_norm": 0.4184967577457428,
"grad_norm_var": 0.0031605906698184564,
"learning_rate": 5e-05,
"loss": 0.1855,
"loss/crossentropy": 2.812410533428192,
"loss/hidden": 0.0,
"loss/logits": 0.18553681299090385,
"loss/reg": 2.8465514183044434,
"step": 436
},
{
"epoch": 0.00437,
"grad_norm": 0.4329560101032257,
"grad_norm_var": 0.0032767273553133062,
"learning_rate": 5e-05,
"loss": 0.1784,
"loss/crossentropy": 2.840768814086914,
"loss/hidden": 0.0,
"loss/logits": 0.17843929678201675,
"loss/reg": 2.844315767288208,
"step": 437
},
{
"epoch": 0.00438,
"grad_norm": 0.6038658022880554,
"grad_norm_var": 0.005936964872234999,
"learning_rate": 5e-05,
"loss": 0.1949,
"loss/crossentropy": 2.7183879017829895,
"loss/hidden": 0.0,
"loss/logits": 0.19491342082619667,
"loss/reg": 2.842548131942749,
"step": 438
},
{
"epoch": 0.00439,
"grad_norm": 0.4069391191005707,
"grad_norm_var": 0.005387125873613382,
"learning_rate": 5e-05,
"loss": 0.1875,
"loss/crossentropy": 2.7780433297157288,
"loss/hidden": 0.0,
"loss/logits": 0.18749799579381943,
"loss/reg": 2.840106964111328,
"step": 439
},
{
"epoch": 0.0044,
"grad_norm": 0.35941290855407715,
"grad_norm_var": 0.005220523762257064,
"learning_rate": 5e-05,
"loss": 0.1639,
"loss/crossentropy": 2.7595601081848145,
"loss/hidden": 0.0,
"loss/logits": 0.1638675332069397,
"loss/reg": 2.8378958702087402,
"step": 440
},
{
"epoch": 0.00441,
"grad_norm": 0.3669149875640869,
"grad_norm_var": 0.005284393934492052,
"learning_rate": 5e-05,
"loss": 0.1714,
"loss/crossentropy": 2.673116147518158,
"loss/hidden": 0.0,
"loss/logits": 0.1713937260210514,
"loss/reg": 2.8351917266845703,
"step": 441
},
{
"epoch": 0.00442,
"grad_norm": 0.3643859922885895,
"grad_norm_var": 0.0051709546313713755,
"learning_rate": 5e-05,
"loss": 0.1795,
"loss/crossentropy": 2.8026832342147827,
"loss/hidden": 0.0,
"loss/logits": 0.17951133847236633,
"loss/reg": 2.8323209285736084,
"step": 442
},
{
"epoch": 0.00443,
"grad_norm": 0.34250232577323914,
"grad_norm_var": 0.005361033629373261,
"learning_rate": 5e-05,
"loss": 0.1729,
"loss/crossentropy": 2.7829501032829285,
"loss/hidden": 0.0,
"loss/logits": 0.17286691814661026,
"loss/reg": 2.8302509784698486,
"step": 443
},
{
"epoch": 0.00444,
"grad_norm": 0.3323063552379608,
"grad_norm_var": 0.005486165176397177,
"learning_rate": 5e-05,
"loss": 0.1685,
"loss/crossentropy": 2.7025471329689026,
"loss/hidden": 0.0,
"loss/logits": 0.1684984639286995,
"loss/reg": 2.827455997467041,
"step": 444
},
{
"epoch": 0.00445,
"grad_norm": 0.35889074206352234,
"grad_norm_var": 0.005320257567319386,
"learning_rate": 5e-05,
"loss": 0.1872,
"loss/crossentropy": 2.6426368355751038,
"loss/hidden": 0.0,
"loss/logits": 0.18716050684452057,
"loss/reg": 2.8253841400146484,
"step": 445
},
{
"epoch": 0.00446,
"grad_norm": 0.39696604013442993,
"grad_norm_var": 0.004953801905317123,
"learning_rate": 5e-05,
"loss": 0.1844,
"loss/crossentropy": 2.6918662786483765,
"loss/hidden": 0.0,
"loss/logits": 0.18442435935139656,
"loss/reg": 2.8228838443756104,
"step": 446
},
{
"epoch": 0.00447,
"grad_norm": 0.3320043981075287,
"grad_norm_var": 0.005107676487313561,
"learning_rate": 5e-05,
"loss": 0.1689,
"loss/crossentropy": 2.7309769988059998,
"loss/hidden": 0.0,
"loss/logits": 0.16887113079428673,
"loss/reg": 2.8211417198181152,
"step": 447
},
{
"epoch": 0.00448,
"grad_norm": 0.3350951373577118,
"grad_norm_var": 0.0051840048892141,
"learning_rate": 5e-05,
"loss": 0.1736,
"loss/crossentropy": 2.7696202397346497,
"loss/hidden": 0.0,
"loss/logits": 0.17358100041747093,
"loss/reg": 2.818211793899536,
"step": 448
},
{
"epoch": 0.00449,
"grad_norm": 0.35995370149612427,
"grad_norm_var": 0.004988316730949372,
"learning_rate": 5e-05,
"loss": 0.185,
"loss/crossentropy": 2.7628800868988037,
"loss/hidden": 0.0,
"loss/logits": 0.1849852055311203,
"loss/reg": 2.8163435459136963,
"step": 449
},
{
"epoch": 0.0045,
"grad_norm": 0.3433259427547455,
"grad_norm_var": 0.004429295151525636,
"learning_rate": 5e-05,
"loss": 0.1776,
"loss/crossentropy": 2.92121821641922,
"loss/hidden": 0.0,
"loss/logits": 0.1775917150080204,
"loss/reg": 2.8140530586242676,
"step": 450
},
{
"epoch": 0.00451,
"grad_norm": 0.3525676131248474,
"grad_norm_var": 0.004477082320186199,
"learning_rate": 5e-05,
"loss": 0.1765,
"loss/crossentropy": 2.8223352432250977,
"loss/hidden": 0.0,
"loss/logits": 0.1765166036784649,
"loss/reg": 2.8114964962005615,
"step": 451
},
{
"epoch": 0.00452,
"grad_norm": 0.3349536955356598,
"grad_norm_var": 0.004502986709870172,
"learning_rate": 5e-05,
"loss": 0.1617,
"loss/crossentropy": 2.58266818523407,
"loss/hidden": 0.0,
"loss/logits": 0.16168920323252678,
"loss/reg": 2.8082778453826904,
"step": 452
},
{
"epoch": 0.00453,
"grad_norm": 0.3272739350795746,
"grad_norm_var": 0.004404667304666754,
"learning_rate": 5e-05,
"loss": 0.1606,
"loss/crossentropy": 2.791221022605896,
"loss/hidden": 0.0,
"loss/logits": 0.16057174652814865,
"loss/reg": 2.8055834770202637,
"step": 453
},
{
"epoch": 0.00454,
"grad_norm": 0.35802412033081055,
"grad_norm_var": 0.0005107777789474354,
"learning_rate": 5e-05,
"loss": 0.1758,
"loss/crossentropy": 2.856186628341675,
"loss/hidden": 0.0,
"loss/logits": 0.17580854520201683,
"loss/reg": 2.8030307292938232,
"step": 454
},
{
"epoch": 0.00455,
"grad_norm": 0.34605538845062256,
"grad_norm_var": 0.0003165176266941239,
"learning_rate": 5e-05,
"loss": 0.1651,
"loss/crossentropy": 2.734806716442108,
"loss/hidden": 0.0,
"loss/logits": 0.16513444855809212,
"loss/reg": 2.7998156547546387,
"step": 455
},
{
"epoch": 0.00456,
"grad_norm": 0.35396888852119446,
"grad_norm_var": 0.0003120198180476634,
"learning_rate": 5e-05,
"loss": 0.1701,
"loss/crossentropy": 2.8904529213905334,
"loss/hidden": 0.0,
"loss/logits": 0.17014532163739204,
"loss/reg": 2.796231508255005,
"step": 456
},
{
"epoch": 0.00457,
"grad_norm": 0.3613145649433136,
"grad_norm_var": 0.00030159148728288546,
"learning_rate": 5e-05,
"loss": 0.1835,
"loss/crossentropy": 2.813112735748291,
"loss/hidden": 0.0,
"loss/logits": 0.1834680140018463,
"loss/reg": 2.7929091453552246,
"step": 457
},
{
"epoch": 0.00458,
"grad_norm": 0.36372610926628113,
"grad_norm_var": 0.00030035069871777733,
"learning_rate": 5e-05,
"loss": 0.167,
"loss/crossentropy": 2.795239508152008,
"loss/hidden": 0.0,
"loss/logits": 0.1669648103415966,
"loss/reg": 2.789707899093628,
"step": 458
},
{
"epoch": 0.00459,
"grad_norm": 0.3581913113594055,
"grad_norm_var": 0.00030019062479412403,
"learning_rate": 5e-05,
"loss": 0.165,
"loss/crossentropy": 2.797567903995514,
"loss/hidden": 0.0,
"loss/logits": 0.16498373076319695,
"loss/reg": 2.786154270172119,
"step": 459
},
{
"epoch": 0.0046,
"grad_norm": 0.3571149408817291,
"grad_norm_var": 0.00027710791712463786,
"learning_rate": 5e-05,
"loss": 0.161,
"loss/crossentropy": 2.743883192539215,
"loss/hidden": 0.0,
"loss/logits": 0.16101711615920067,
"loss/reg": 2.7824792861938477,
"step": 460
},
{
"epoch": 0.00461,
"grad_norm": 0.356715589761734,
"grad_norm_var": 0.0002755397827386948,
"learning_rate": 5e-05,
"loss": 0.1732,
"loss/crossentropy": 2.724743604660034,
"loss/hidden": 0.0,
"loss/logits": 0.1731928214430809,
"loss/reg": 2.778890371322632,
"step": 461
},
{
"epoch": 0.00462,
"grad_norm": 0.3243059813976288,
"grad_norm_var": 0.00017305590364662023,
"learning_rate": 5e-05,
"loss": 0.1592,
"loss/crossentropy": 2.7731017470359802,
"loss/hidden": 0.0,
"loss/logits": 0.15923070535063744,
"loss/reg": 2.775613784790039,
"step": 462
},
{
"epoch": 0.00463,
"grad_norm": 0.3843972980976105,
"grad_norm_var": 0.00023436686166613171,
"learning_rate": 5e-05,
"loss": 0.1776,
"loss/crossentropy": 2.6426811814308167,
"loss/hidden": 0.0,
"loss/logits": 0.17761223763227463,
"loss/reg": 2.772561550140381,
"step": 463
},
{
"epoch": 0.00464,
"grad_norm": 0.3468632102012634,
"grad_norm_var": 0.00021796986892265539,
"learning_rate": 5e-05,
"loss": 0.1742,
"loss/crossentropy": 2.882816791534424,
"loss/hidden": 0.0,
"loss/logits": 0.17419364303350449,
"loss/reg": 2.769129991531372,
"step": 464
},
{
"epoch": 0.00465,
"grad_norm": 0.3967267870903015,
"grad_norm_var": 0.0003424789629975648,
"learning_rate": 5e-05,
"loss": 0.1785,
"loss/crossentropy": 2.9460648894309998,
"loss/hidden": 0.0,
"loss/logits": 0.1784828118979931,
"loss/reg": 2.767021656036377,
"step": 465
},
{
"epoch": 0.00466,
"grad_norm": 0.36927327513694763,
"grad_norm_var": 0.0003472996962895862,
"learning_rate": 5e-05,
"loss": 0.1723,
"loss/crossentropy": 2.7243736386299133,
"loss/hidden": 0.0,
"loss/logits": 0.17225344851613045,
"loss/reg": 2.7637596130371094,
"step": 466
},
{
"epoch": 0.00467,
"grad_norm": 0.6855224967002869,
"grad_norm_var": 0.007136168552583877,
"learning_rate": 5e-05,
"loss": 0.1725,
"loss/crossentropy": 2.9400131702423096,
"loss/hidden": 0.0,
"loss/logits": 0.17254997044801712,
"loss/reg": 2.7604963779449463,
"step": 467
},
{
"epoch": 0.00468,
"grad_norm": 0.41388872265815735,
"grad_norm_var": 0.007088047286249225,
"learning_rate": 5e-05,
"loss": 0.1608,
"loss/crossentropy": 2.7375746369361877,
"loss/hidden": 0.0,
"loss/logits": 0.16082635894417763,
"loss/reg": 2.7572667598724365,
"step": 468
},
{
"epoch": 0.00469,
"grad_norm": 0.3997427225112915,
"grad_norm_var": 0.006892705403346755,
"learning_rate": 5e-05,
"loss": 0.1704,
"loss/crossentropy": 2.8761582374572754,
"loss/hidden": 0.0,
"loss/logits": 0.17043552175164223,
"loss/reg": 2.754149913787842,
"step": 469
},
{
"epoch": 0.0047,
"grad_norm": 0.33424052596092224,
"grad_norm_var": 0.007016741295476203,
"learning_rate": 5e-05,
"loss": 0.1652,
"loss/crossentropy": 2.8255309462547302,
"loss/hidden": 0.0,
"loss/logits": 0.16519855335354805,
"loss/reg": 2.751276731491089,
"step": 470
},
{
"epoch": 0.00471,
"grad_norm": 0.41415977478027344,
"grad_norm_var": 0.006957502567752493,
"learning_rate": 5e-05,
"loss": 0.1832,
"loss/crossentropy": 2.676911950111389,
"loss/hidden": 0.0,
"loss/logits": 0.18318749964237213,
"loss/reg": 2.7486090660095215,
"step": 471
},
{
"epoch": 0.00472,
"grad_norm": 0.3688299357891083,
"grad_norm_var": 0.006902369057221236,
"learning_rate": 5e-05,
"loss": 0.162,
"loss/crossentropy": 2.6736281514167786,
"loss/hidden": 0.0,
"loss/logits": 0.16196409612894058,
"loss/reg": 2.745225667953491,
"step": 472
},
{
"epoch": 0.00473,
"grad_norm": 0.42543119192123413,
"grad_norm_var": 0.006916738926360373,
"learning_rate": 5e-05,
"loss": 0.1685,
"loss/crossentropy": 2.7871673703193665,
"loss/hidden": 0.0,
"loss/logits": 0.16850638762116432,
"loss/reg": 2.7421023845672607,
"step": 473
},
{
"epoch": 0.00474,
"grad_norm": 0.3542870581150055,
"grad_norm_var": 0.006960025235757868,
"learning_rate": 5e-05,
"loss": 0.1695,
"loss/crossentropy": 2.727401077747345,
"loss/hidden": 0.0,
"loss/logits": 0.1695132479071617,
"loss/reg": 2.738083839416504,
"step": 474
},
{
"epoch": 0.00475,
"grad_norm": 0.42508408427238464,
"grad_norm_var": 0.006928287935252916,
"learning_rate": 5e-05,
"loss": 0.1967,
"loss/crossentropy": 2.756391167640686,
"loss/hidden": 0.0,
"loss/logits": 0.19667796790599823,
"loss/reg": 2.734565496444702,
"step": 475
},
{
"epoch": 0.00476,
"grad_norm": 0.3327108919620514,
"grad_norm_var": 0.007096223362361916,
"learning_rate": 5e-05,
"loss": 0.1698,
"loss/crossentropy": 2.662286937236786,
"loss/hidden": 0.0,
"loss/logits": 0.16976173967123032,
"loss/reg": 2.730898141860962,
"step": 476
},
{
"epoch": 0.00477,
"grad_norm": 0.3538263142108917,
"grad_norm_var": 0.007111786918880665,
"learning_rate": 5e-05,
"loss": 0.1728,
"loss/crossentropy": 2.6600981950759888,
"loss/hidden": 0.0,
"loss/logits": 0.17283405736088753,
"loss/reg": 2.7276227474212646,
"step": 477
},
{
"epoch": 0.00478,
"grad_norm": 0.6719810962677002,
"grad_norm_var": 0.011362604241119423,
"learning_rate": 5e-05,
"loss": 0.1956,
"loss/crossentropy": 2.9192944169044495,
"loss/hidden": 0.0,
"loss/logits": 0.1955549158155918,
"loss/reg": 2.7246785163879395,
"step": 478
},
{
"epoch": 0.00479,
"grad_norm": 0.40175002813339233,
"grad_norm_var": 0.011305273259017534,
"learning_rate": 5e-05,
"loss": 0.1707,
"loss/crossentropy": 2.8099315762519836,
"loss/hidden": 0.0,
"loss/logits": 0.1706707924604416,
"loss/reg": 2.721214771270752,
"step": 479
},
{
"epoch": 0.0048,
"grad_norm": 0.5700014233589172,
"grad_norm_var": 0.012288996380569357,
"learning_rate": 5e-05,
"loss": 0.175,
"loss/crossentropy": 2.764845371246338,
"loss/hidden": 0.0,
"loss/logits": 0.17503220587968826,
"loss/reg": 2.71852970123291,
"step": 480
},
{
"epoch": 0.00481,
"grad_norm": 0.3602856993675232,
"grad_norm_var": 0.012545036289332723,
"learning_rate": 5e-05,
"loss": 0.1675,
"loss/crossentropy": 2.817295730113983,
"loss/hidden": 0.0,
"loss/logits": 0.16745564714074135,
"loss/reg": 2.7155778408050537,
"step": 481
},
{
"epoch": 0.00482,
"grad_norm": 0.37470605969429016,
"grad_norm_var": 0.012502846327791594,
"learning_rate": 5e-05,
"loss": 0.1722,
"loss/crossentropy": 2.7710434794425964,
"loss/hidden": 0.0,
"loss/logits": 0.172159094363451,
"loss/reg": 2.7133727073669434,
"step": 482
},
{
"epoch": 0.00483,
"grad_norm": 0.319488525390625,
"grad_norm_var": 0.008425663660975724,
"learning_rate": 5e-05,
"loss": 0.1524,
"loss/crossentropy": 2.783412456512451,
"loss/hidden": 0.0,
"loss/logits": 0.15241163223981857,
"loss/reg": 2.710848331451416,
"step": 483
},
{
"epoch": 0.00484,
"grad_norm": 0.3474343419075012,
"grad_norm_var": 0.008645296689366684,
"learning_rate": 5e-05,
"loss": 0.1582,
"loss/crossentropy": 2.8712441325187683,
"loss/hidden": 0.0,
"loss/logits": 0.1582440249621868,
"loss/reg": 2.708759307861328,
"step": 484
},
{
"epoch": 0.00485,
"grad_norm": 0.3881974518299103,
"grad_norm_var": 0.008659215056144554,
"learning_rate": 5e-05,
"loss": 0.168,
"loss/crossentropy": 2.7801290154457092,
"loss/hidden": 0.0,
"loss/logits": 0.16804108768701553,
"loss/reg": 2.707090139389038,
"step": 485
},
{
"epoch": 0.00486,
"grad_norm": 0.3865320086479187,
"grad_norm_var": 0.008353144350497502,
"learning_rate": 5e-05,
"loss": 0.1622,
"loss/crossentropy": 2.762259840965271,
"loss/hidden": 0.0,
"loss/logits": 0.1622123382985592,
"loss/reg": 2.704341173171997,
"step": 486
},
{
"epoch": 0.00487,
"grad_norm": 0.3601287007331848,
"grad_norm_var": 0.008476237288047564,
"learning_rate": 5e-05,
"loss": 0.1717,
"loss/crossentropy": 2.7640222311019897,
"loss/hidden": 0.0,
"loss/logits": 0.1716899275779724,
"loss/reg": 2.702449321746826,
"step": 487
},
{
"epoch": 0.00488,
"grad_norm": 0.3476349711418152,
"grad_norm_var": 0.008599584577097493,
"learning_rate": 5e-05,
"loss": 0.1771,
"loss/crossentropy": 2.8057321906089783,
"loss/hidden": 0.0,
"loss/logits": 0.1770944595336914,
"loss/reg": 2.6996164321899414,
"step": 488
},
{
"epoch": 0.00489,
"grad_norm": 0.35880380868911743,
"grad_norm_var": 0.008661929013581293,
"learning_rate": 5e-05,
"loss": 0.1759,
"loss/crossentropy": 2.7546836137771606,
"loss/hidden": 0.0,
"loss/logits": 0.17589908093214035,
"loss/reg": 2.69681978225708,
"step": 489
},
{
"epoch": 0.0049,
"grad_norm": 0.3216891586780548,
"grad_norm_var": 0.008914221483014847,
"learning_rate": 5e-05,
"loss": 0.1628,
"loss/crossentropy": 2.755846858024597,
"loss/hidden": 0.0,
"loss/logits": 0.16277150437235832,
"loss/reg": 2.6948180198669434,
"step": 490
},
{
"epoch": 0.00491,
"grad_norm": 0.3739294409751892,
"grad_norm_var": 0.008872687766587003,
"learning_rate": 5e-05,
"loss": 0.1737,
"loss/crossentropy": 2.758453607559204,
"loss/hidden": 0.0,
"loss/logits": 0.173715490847826,
"loss/reg": 2.692713737487793,
"step": 491
},
{
"epoch": 0.00492,
"grad_norm": 0.3633546531200409,
"grad_norm_var": 0.008689872848313067,
"learning_rate": 5e-05,
"loss": 0.1755,
"loss/crossentropy": 2.756626844406128,
"loss/hidden": 0.0,
"loss/logits": 0.17549088224768639,
"loss/reg": 2.6911702156066895,
"step": 492
},
{
"epoch": 0.00493,
"grad_norm": 0.4165309965610504,
"grad_norm_var": 0.00860196217059566,
"learning_rate": 5e-05,
"loss": 0.1882,
"loss/crossentropy": 2.69485205411911,
"loss/hidden": 0.0,
"loss/logits": 0.18822569772601128,
"loss/reg": 2.6890034675598145,
"step": 493
},
{
"epoch": 0.00494,
"grad_norm": 0.34585461020469666,
"grad_norm_var": 0.0033206140596304815,
"learning_rate": 5e-05,
"loss": 0.1742,
"loss/crossentropy": 2.73829984664917,
"loss/hidden": 0.0,
"loss/logits": 0.17416464537382126,
"loss/reg": 2.687713623046875,
"step": 494
},
{
"epoch": 0.00495,
"grad_norm": 0.3443280756473541,
"grad_norm_var": 0.003339269529387142,
"learning_rate": 5e-05,
"loss": 0.1613,
"loss/crossentropy": 2.745963931083679,
"loss/hidden": 0.0,
"loss/logits": 0.161319550126791,
"loss/reg": 2.685638189315796,
"step": 495
},
{
"epoch": 0.00496,
"grad_norm": 0.3689098656177521,
"grad_norm_var": 0.000602855553897171,
"learning_rate": 5e-05,
"loss": 0.1735,
"loss/crossentropy": 2.6744813919067383,
"loss/hidden": 0.0,
"loss/logits": 0.17346932739019394,
"loss/reg": 2.6839358806610107,
"step": 496
},
{
"epoch": 0.00497,
"grad_norm": 0.36457374691963196,
"grad_norm_var": 0.0006035317496342747,
"learning_rate": 5e-05,
"loss": 0.1735,
"loss/crossentropy": 2.799642562866211,
"loss/hidden": 0.0,
"loss/logits": 0.17352834343910217,
"loss/reg": 2.683485507965088,
"step": 497
},
{
"epoch": 0.00498,
"grad_norm": 0.35222312808036804,
"grad_norm_var": 0.0005951796117876367,
"learning_rate": 5e-05,
"loss": 0.1742,
"loss/crossentropy": 2.739416480064392,
"loss/hidden": 0.0,
"loss/logits": 0.1742345169186592,
"loss/reg": 2.6830482482910156,
"step": 498
},
{
"epoch": 0.00499,
"grad_norm": 0.3427422344684601,
"grad_norm_var": 0.0005034448418147264,
"learning_rate": 5e-05,
"loss": 0.1673,
"loss/crossentropy": 2.772252082824707,
"loss/hidden": 0.0,
"loss/logits": 0.16728588938713074,
"loss/reg": 2.682189464569092,
"step": 499
},
{
"epoch": 0.005,
"grad_norm": 0.39299577474594116,
"grad_norm_var": 0.0005481683329227494,
"learning_rate": 5e-05,
"loss": 0.1926,
"loss/crossentropy": 2.7702752351760864,
"loss/hidden": 0.0,
"loss/logits": 0.1926349699497223,
"loss/reg": 2.6808254718780518,
"step": 500
},
{
"epoch": 0.00501,
"grad_norm": 0.3431949019432068,
"grad_norm_var": 0.0005312130675710792,
"learning_rate": 5e-05,
"loss": 0.1634,
"loss/crossentropy": 2.7881234288215637,
"loss/hidden": 0.0,
"loss/logits": 0.1633942425251007,
"loss/reg": 2.6798629760742188,
"step": 501
},
{
"epoch": 0.00502,
"grad_norm": 0.36641839146614075,
"grad_norm_var": 0.0004892704880637311,
"learning_rate": 5e-05,
"loss": 0.1787,
"loss/crossentropy": 2.848407030105591,
"loss/hidden": 0.0,
"loss/logits": 0.17872987315058708,
"loss/reg": 2.677311658859253,
"step": 502
},
{
"epoch": 0.00503,
"grad_norm": 0.3278079330921173,
"grad_norm_var": 0.000554897538649816,
"learning_rate": 5e-05,
"loss": 0.1587,
"loss/crossentropy": 2.75662362575531,
"loss/hidden": 0.0,
"loss/logits": 0.15868044644594193,
"loss/reg": 2.675185203552246,
"step": 503
},
{
"epoch": 0.00504,
"grad_norm": 0.3251039683818817,
"grad_norm_var": 0.0006183250665441046,
"learning_rate": 5e-05,
"loss": 0.1551,
"loss/crossentropy": 2.731416165828705,
"loss/hidden": 0.0,
"loss/logits": 0.15506618097424507,
"loss/reg": 2.673948287963867,
"step": 504
},
{
"epoch": 0.00505,
"grad_norm": 0.35344070196151733,
"grad_norm_var": 0.0006186746986458047,
"learning_rate": 5e-05,
"loss": 0.167,
"loss/crossentropy": 2.740668296813965,
"loss/hidden": 0.0,
"loss/logits": 0.16695522889494896,
"loss/reg": 2.6712498664855957,
"step": 505
},
{
"epoch": 0.00506,
"grad_norm": 0.36658284068107605,
"grad_norm_var": 0.0005366058949143918,
"learning_rate": 5e-05,
"loss": 0.1668,
"loss/crossentropy": 2.802608013153076,
"loss/hidden": 0.0,
"loss/logits": 0.16682763025164604,
"loss/reg": 2.669286012649536,
"step": 506
},
{
"epoch": 0.00507,
"grad_norm": 0.4423954486846924,
"grad_norm_var": 0.000963591213409624,
"learning_rate": 5e-05,
"loss": 0.1963,
"loss/crossentropy": 2.811932861804962,
"loss/hidden": 0.0,
"loss/logits": 0.19632378965616226,
"loss/reg": 2.666898012161255,
"step": 507
},
{
"epoch": 0.00508,
"grad_norm": 0.3770610988140106,
"grad_norm_var": 0.000975015024308116,
"learning_rate": 5e-05,
"loss": 0.1753,
"loss/crossentropy": 2.7279282808303833,
"loss/hidden": 0.0,
"loss/logits": 0.17527905479073524,
"loss/reg": 2.664764881134033,
"step": 508
},
{
"epoch": 0.00509,
"grad_norm": 0.35589146614074707,
"grad_norm_var": 0.0007832244440521922,
"learning_rate": 5e-05,
"loss": 0.168,
"loss/crossentropy": 2.8977367281913757,
"loss/hidden": 0.0,
"loss/logits": 0.16798892244696617,
"loss/reg": 2.6622438430786133,
"step": 509
},
{
"epoch": 0.0051,
"grad_norm": 0.3419097661972046,
"grad_norm_var": 0.0007919503322765919,
"learning_rate": 5e-05,
"loss": 0.1729,
"loss/crossentropy": 2.6906025409698486,
"loss/hidden": 0.0,
"loss/logits": 0.172856405377388,
"loss/reg": 2.6595211029052734,
"step": 510
},
{
"epoch": 0.00511,
"grad_norm": 0.3972381353378296,
"grad_norm_var": 0.0008538971282195384,
"learning_rate": 5e-05,
"loss": 0.1825,
"loss/crossentropy": 2.762513279914856,
"loss/hidden": 0.0,
"loss/logits": 0.18245646730065346,
"loss/reg": 2.6572320461273193,
"step": 511
},
{
"epoch": 0.00512,
"grad_norm": 0.3489353358745575,
"grad_norm_var": 0.0008648399289393501,
"learning_rate": 5e-05,
"loss": 0.179,
"loss/crossentropy": 2.7472071647644043,
"loss/hidden": 0.0,
"loss/logits": 0.1789936050772667,
"loss/reg": 2.6544342041015625,
"step": 512
},
{
"epoch": 0.00513,
"grad_norm": 0.3673308491706848,
"grad_norm_var": 0.0008661114894439326,
"learning_rate": 5e-05,
"loss": 0.1745,
"loss/crossentropy": 2.69700163602829,
"loss/hidden": 0.0,
"loss/logits": 0.1745261810719967,
"loss/reg": 2.6520497798919678,
"step": 513
},
{
"epoch": 0.00514,
"grad_norm": 0.33870744705200195,
"grad_norm_var": 0.0008961917113334199,
"learning_rate": 5e-05,
"loss": 0.1649,
"loss/crossentropy": 2.762860357761383,
"loss/hidden": 0.0,
"loss/logits": 0.16494135558605194,
"loss/reg": 2.6500847339630127,
"step": 514
},
{
"epoch": 0.00515,
"grad_norm": 0.40411266684532166,
"grad_norm_var": 0.0009761766654230563,
"learning_rate": 5e-05,
"loss": 0.1628,
"loss/crossentropy": 3.01085501909256,
"loss/hidden": 0.0,
"loss/logits": 0.16284478455781937,
"loss/reg": 2.648311138153076,
"step": 515
},
{
"epoch": 0.00516,
"grad_norm": 0.37194308638572693,
"grad_norm_var": 0.0009268939874421604,
"learning_rate": 5e-05,
"loss": 0.182,
"loss/crossentropy": 2.721080005168915,
"loss/hidden": 0.0,
"loss/logits": 0.18202906847000122,
"loss/reg": 2.6469640731811523,
"step": 516
},
{
"epoch": 0.00517,
"grad_norm": 0.3380615711212158,
"grad_norm_var": 0.0009429551352979477,
"learning_rate": 5e-05,
"loss": 0.1639,
"loss/crossentropy": 2.6788495779037476,
"loss/hidden": 0.0,
"loss/logits": 0.16385124996304512,
"loss/reg": 2.6445441246032715,
"step": 517
},
{
"epoch": 0.00518,
"grad_norm": 0.37696361541748047,
"grad_norm_var": 0.0009533986625055632,
"learning_rate": 5e-05,
"loss": 0.1587,
"loss/crossentropy": 2.6845511198043823,
"loss/hidden": 0.0,
"loss/logits": 0.1586880125105381,
"loss/reg": 2.6424736976623535,
"step": 518
},
{
"epoch": 0.00519,
"grad_norm": 0.32983675599098206,
"grad_norm_var": 0.0009437052369864992,
"learning_rate": 5e-05,
"loss": 0.1585,
"loss/crossentropy": 2.5984672904014587,
"loss/hidden": 0.0,
"loss/logits": 0.15849433466792107,
"loss/reg": 2.639796257019043,
"step": 519
},
{
"epoch": 0.0052,
"grad_norm": 0.3439983129501343,
"grad_norm_var": 0.000866215802107521,
"learning_rate": 5e-05,
"loss": 0.1578,
"loss/crossentropy": 2.7057528495788574,
"loss/hidden": 0.0,
"loss/logits": 0.15775253251194954,
"loss/reg": 2.636976480484009,
"step": 520
},
{
"epoch": 0.00521,
"grad_norm": 0.4739494323730469,
"grad_norm_var": 0.0015736599047053415,
"learning_rate": 5e-05,
"loss": 0.178,
"loss/crossentropy": 2.6239394545555115,
"loss/hidden": 0.0,
"loss/logits": 0.17801255360245705,
"loss/reg": 2.6342852115631104,
"step": 521
},
{
"epoch": 0.00522,
"grad_norm": 0.5270029306411743,
"grad_norm_var": 0.003035565907296726,
"learning_rate": 5e-05,
"loss": 0.1801,
"loss/crossentropy": 2.890467643737793,
"loss/hidden": 0.0,
"loss/logits": 0.18007512018084526,
"loss/reg": 2.631289005279541,
"step": 522
},
{
"epoch": 0.00523,
"grad_norm": 0.42719003558158875,
"grad_norm_var": 0.002930528350278516,
"learning_rate": 5e-05,
"loss": 0.1781,
"loss/crossentropy": 2.9749565720558167,
"loss/hidden": 0.0,
"loss/logits": 0.17812742665410042,
"loss/reg": 2.6284420490264893,
"step": 523
},
{
"epoch": 0.00524,
"grad_norm": 0.37133005261421204,
"grad_norm_var": 0.0029367435634455913,
"learning_rate": 5e-05,
"loss": 0.1597,
"loss/crossentropy": 2.692670702934265,
"loss/hidden": 0.0,
"loss/logits": 0.1597190946340561,
"loss/reg": 2.6251702308654785,
"step": 524
},
{
"epoch": 0.00525,
"grad_norm": 0.3646347224712372,
"grad_norm_var": 0.0029109098946428253,
"learning_rate": 5e-05,
"loss": 0.1676,
"loss/crossentropy": 2.8696910738945007,
"loss/hidden": 0.0,
"loss/logits": 0.16764900088310242,
"loss/reg": 2.621973991394043,
"step": 525
},
{
"epoch": 0.00526,
"grad_norm": 0.3347557485103607,
"grad_norm_var": 0.002953013887398237,
"learning_rate": 5e-05,
"loss": 0.1659,
"loss/crossentropy": 2.844240427017212,
"loss/hidden": 0.0,
"loss/logits": 0.16587505862116814,
"loss/reg": 2.6180617809295654,
"step": 526
},
{
"epoch": 0.00527,
"grad_norm": 0.3301764726638794,
"grad_norm_var": 0.003100070614909223,
"learning_rate": 5e-05,
"loss": 0.1554,
"loss/crossentropy": 2.822225272655487,
"loss/hidden": 0.0,
"loss/logits": 0.15541274286806583,
"loss/reg": 2.6156363487243652,
"step": 527
},
{
"epoch": 0.00528,
"grad_norm": 0.3668423593044281,
"grad_norm_var": 0.003050578439524883,
"learning_rate": 5e-05,
"loss": 0.1725,
"loss/crossentropy": 2.873881459236145,
"loss/hidden": 0.0,
"loss/logits": 0.17249644920229912,
"loss/reg": 2.6129000186920166,
"step": 528
},
{
"epoch": 0.00529,
"grad_norm": 0.33062636852264404,
"grad_norm_var": 0.0031927551041592986,
"learning_rate": 5e-05,
"loss": 0.169,
"loss/crossentropy": 2.7202290296554565,
"loss/hidden": 0.0,
"loss/logits": 0.16901781037449837,
"loss/reg": 2.6098320484161377,
"step": 529
},
{
"epoch": 0.0053,
"grad_norm": 0.33170488476753235,
"grad_norm_var": 0.003231463613689256,
"learning_rate": 5e-05,
"loss": 0.1708,
"loss/crossentropy": 2.7543463706970215,
"loss/hidden": 0.0,
"loss/logits": 0.17077547311782837,
"loss/reg": 2.606674909591675,
"step": 530
},
{
"epoch": 0.00531,
"grad_norm": 0.3436318337917328,
"grad_norm_var": 0.0032369737172315838,
"learning_rate": 5e-05,
"loss": 0.1822,
"loss/crossentropy": 2.6231788992881775,
"loss/hidden": 0.0,
"loss/logits": 0.1821577101945877,
"loss/reg": 2.603997230529785,
"step": 531
},
{
"epoch": 0.00532,
"grad_norm": 0.33105242252349854,
"grad_norm_var": 0.0033454153420392264,
"learning_rate": 5e-05,
"loss": 0.1661,
"loss/crossentropy": 2.819184124469757,
"loss/hidden": 0.0,
"loss/logits": 0.16607840731739998,
"loss/reg": 2.6012203693389893,
"step": 532
},
{
"epoch": 0.00533,
"grad_norm": 0.3485148847103119,
"grad_norm_var": 0.0033075767398377588,
"learning_rate": 5e-05,
"loss": 0.1676,
"loss/crossentropy": 2.8594303727149963,
"loss/hidden": 0.0,
"loss/logits": 0.1676221825182438,
"loss/reg": 2.5986573696136475,
"step": 533
},
{
"epoch": 0.00534,
"grad_norm": 0.3541623651981354,
"grad_norm_var": 0.003321219936842216,
"learning_rate": 5e-05,
"loss": 0.1742,
"loss/crossentropy": 2.7382256984710693,
"loss/hidden": 0.0,
"loss/logits": 0.17420669272542,
"loss/reg": 2.5956368446350098,
"step": 534
},
{
"epoch": 0.00535,
"grad_norm": 0.362183541059494,
"grad_norm_var": 0.003216249066768325,
"learning_rate": 5e-05,
"loss": 0.1732,
"loss/crossentropy": 2.820302128791809,
"loss/hidden": 0.0,
"loss/logits": 0.1731831431388855,
"loss/reg": 2.591860294342041,
"step": 535
},
{
"epoch": 0.00536,
"grad_norm": 0.340348482131958,
"grad_norm_var": 0.0032303969391706505,
"learning_rate": 5e-05,
"loss": 0.1677,
"loss/crossentropy": 2.978896915912628,
"loss/hidden": 0.0,
"loss/logits": 0.16773569583892822,
"loss/reg": 2.5879762172698975,
"step": 536
},
{
"epoch": 0.00537,
"grad_norm": 0.359326034784317,
"grad_norm_var": 0.002480178301492671,
"learning_rate": 5e-05,
"loss": 0.1767,
"loss/crossentropy": 2.7645240426063538,
"loss/hidden": 0.0,
"loss/logits": 0.17668773606419563,
"loss/reg": 2.5847809314727783,
"step": 537
},
{
"epoch": 0.00538,
"grad_norm": 0.3420480489730835,
"grad_norm_var": 0.0005976425682412671,
"learning_rate": 5e-05,
"loss": 0.1758,
"loss/crossentropy": 2.7278724908828735,
"loss/hidden": 0.0,
"loss/logits": 0.17576001212000847,
"loss/reg": 2.581143379211426,
"step": 538
},
{
"epoch": 0.00539,
"grad_norm": 0.33362701535224915,
"grad_norm_var": 0.00021185911019383125,
"learning_rate": 5e-05,
"loss": 0.1708,
"loss/crossentropy": 2.6828721165657043,
"loss/hidden": 0.0,
"loss/logits": 0.1707863062620163,
"loss/reg": 2.5769548416137695,
"step": 539
},
{
"epoch": 0.0054,
"grad_norm": 0.6795082092285156,
"grad_norm_var": 0.007165518560383773,
"learning_rate": 5e-05,
"loss": 0.2001,
"loss/crossentropy": 2.7977577447891235,
"loss/hidden": 0.0,
"loss/logits": 0.20013980567455292,
"loss/reg": 2.5742313861846924,
"step": 540
},
{
"epoch": 0.00541,
"grad_norm": 0.35366278886795044,
"grad_norm_var": 0.007174778628811747,
"learning_rate": 5e-05,
"loss": 0.1703,
"loss/crossentropy": 2.859143853187561,
"loss/hidden": 0.0,
"loss/logits": 0.17025134339928627,
"loss/reg": 2.57037353515625,
"step": 541
},
{
"epoch": 0.00542,
"grad_norm": 0.3655552566051483,
"grad_norm_var": 0.007109308326582827,
"learning_rate": 5e-05,
"loss": 0.1775,
"loss/crossentropy": 2.8502614498138428,
"loss/hidden": 0.0,
"loss/logits": 0.1775321438908577,
"loss/reg": 2.566716432571411,
"step": 542
},
{
"epoch": 0.00543,
"grad_norm": 0.3574732542037964,
"grad_norm_var": 0.007021635262450318,
"learning_rate": 5e-05,
"loss": 0.1735,
"loss/crossentropy": 2.892129361629486,
"loss/hidden": 0.0,
"loss/logits": 0.17345865443348885,
"loss/reg": 2.563842296600342,
"step": 543
},
{
"epoch": 0.00544,
"grad_norm": 0.36598220467567444,
"grad_norm_var": 0.007021902205424502,
"learning_rate": 5e-05,
"loss": 0.1721,
"loss/crossentropy": 2.7138225436210632,
"loss/hidden": 0.0,
"loss/logits": 0.17211398482322693,
"loss/reg": 2.5608513355255127,
"step": 544
},
{
"epoch": 0.00545,
"grad_norm": 0.35922348499298096,
"grad_norm_var": 0.0069277921155704155,
"learning_rate": 5e-05,
"loss": 0.1695,
"loss/crossentropy": 2.8138818740844727,
"loss/hidden": 0.0,
"loss/logits": 0.16949571669101715,
"loss/reg": 2.557931423187256,
"step": 545
},
{
"epoch": 0.00546,
"grad_norm": 0.3538724184036255,
"grad_norm_var": 0.006843838113958241,
"learning_rate": 5e-05,
"loss": 0.1722,
"loss/crossentropy": 2.7698569893836975,
"loss/hidden": 0.0,
"loss/logits": 0.17218982055783272,
"loss/reg": 2.5551669597625732,
"step": 546
},
{
"epoch": 0.00547,
"grad_norm": 0.38070008158683777,
"grad_norm_var": 0.006790073386325786,
"learning_rate": 5e-05,
"loss": 0.1843,
"loss/crossentropy": 2.631078600883484,
"loss/hidden": 0.0,
"loss/logits": 0.1842627413570881,
"loss/reg": 2.5517876148223877,
"step": 547
},
{
"epoch": 0.00548,
"grad_norm": 0.35319533944129944,
"grad_norm_var": 0.006693321782661003,
"learning_rate": 5e-05,
"loss": 0.16,
"loss/crossentropy": 2.850399076938629,
"loss/hidden": 0.0,
"loss/logits": 0.15995023399591446,
"loss/reg": 2.548754930496216,
"step": 548
},
{
"epoch": 0.00549,
"grad_norm": 0.4596186578273773,
"grad_norm_var": 0.0070637908733671,
"learning_rate": 5e-05,
"loss": 0.164,
"loss/crossentropy": 2.868459641933441,
"loss/hidden": 0.0,
"loss/logits": 0.16395244374871254,
"loss/reg": 2.5462560653686523,
"step": 549
},
{
"epoch": 0.0055,
"grad_norm": 0.3474785387516022,
"grad_norm_var": 0.007091863949161529,
"learning_rate": 5e-05,
"loss": 0.1641,
"loss/crossentropy": 2.6114882230758667,
"loss/hidden": 0.0,
"loss/logits": 0.1641043722629547,
"loss/reg": 2.5431270599365234,
"step": 550
},
{
"epoch": 0.00551,
"grad_norm": 0.3570033013820648,
"grad_norm_var": 0.007107306178779664,
"learning_rate": 5e-05,
"loss": 0.1579,
"loss/crossentropy": 2.8220438957214355,
"loss/hidden": 0.0,
"loss/logits": 0.15791887789964676,
"loss/reg": 2.539910078048706,
"step": 551
},
{
"epoch": 0.00552,
"grad_norm": 0.32915255427360535,
"grad_norm_var": 0.0071770024029156184,
"learning_rate": 5e-05,
"loss": 0.1612,
"loss/crossentropy": 2.878856658935547,
"loss/hidden": 0.0,
"loss/logits": 0.16122740507125854,
"loss/reg": 2.5368034839630127,
"step": 552
},
{
"epoch": 0.00553,
"grad_norm": 0.3565903604030609,
"grad_norm_var": 0.007185408405122592,
"learning_rate": 5e-05,
"loss": 0.1689,
"loss/crossentropy": 2.705552637577057,
"loss/hidden": 0.0,
"loss/logits": 0.16892266646027565,
"loss/reg": 2.5333409309387207,
"step": 553
},
{
"epoch": 0.00554,
"grad_norm": 0.31767770648002625,
"grad_norm_var": 0.0073488319211029345,
"learning_rate": 5e-05,
"loss": 0.1579,
"loss/crossentropy": 2.685009717941284,
"loss/hidden": 0.0,
"loss/logits": 0.15788856148719788,
"loss/reg": 2.5296521186828613,
"step": 554
},
{
"epoch": 0.00555,
"grad_norm": 0.35047340393066406,
"grad_norm_var": 0.0072637659398345844,
"learning_rate": 5e-05,
"loss": 0.1746,
"loss/crossentropy": 2.5745012760162354,
"loss/hidden": 0.0,
"loss/logits": 0.17458590865135193,
"loss/reg": 2.5270395278930664,
"step": 555
},
{
"epoch": 0.00556,
"grad_norm": 0.3832140266895294,
"grad_norm_var": 0.0009360149891549837,
"learning_rate": 5e-05,
"loss": 0.1676,
"loss/crossentropy": 2.8551809787750244,
"loss/hidden": 0.0,
"loss/logits": 0.16756092011928558,
"loss/reg": 2.523982048034668,
"step": 556
},
{
"epoch": 0.00557,
"grad_norm": 0.4020329713821411,
"grad_norm_var": 0.0010289291164416311,
"learning_rate": 5e-05,
"loss": 0.1755,
"loss/crossentropy": 2.787672698497772,
"loss/hidden": 0.0,
"loss/logits": 0.17547398060560226,
"loss/reg": 2.520615816116333,
"step": 557
},
{
"epoch": 0.00558,
"grad_norm": 0.38412049412727356,
"grad_norm_var": 0.0010519623608851428,
"learning_rate": 5e-05,
"loss": 0.1815,
"loss/crossentropy": 2.846573293209076,
"loss/hidden": 0.0,
"loss/logits": 0.1815263032913208,
"loss/reg": 2.518004894256592,
"step": 558
},
{
"epoch": 0.00559,
"grad_norm": 0.3456071615219116,
"grad_norm_var": 0.0010744320361522322,
"learning_rate": 5e-05,
"loss": 0.1746,
"loss/crossentropy": 2.922893524169922,
"loss/hidden": 0.0,
"loss/logits": 0.17459525167942047,
"loss/reg": 2.5153868198394775,
"step": 559
},
{
"epoch": 0.0056,
"grad_norm": 0.36563733220100403,
"grad_norm_var": 0.0010744113839659304,
"learning_rate": 5e-05,
"loss": 0.169,
"loss/crossentropy": 2.7154372334480286,
"loss/hidden": 0.0,
"loss/logits": 0.16903281211853027,
"loss/reg": 2.5126099586486816,
"step": 560
},
{
"epoch": 0.00561,
"grad_norm": 0.3387238085269928,
"grad_norm_var": 0.00111742135319511,
"learning_rate": 5e-05,
"loss": 0.1666,
"loss/crossentropy": 2.5947054624557495,
"loss/hidden": 0.0,
"loss/logits": 0.16662058234214783,
"loss/reg": 2.509439706802368,
"step": 561
},
{
"epoch": 0.00562,
"grad_norm": 0.45574790239334106,
"grad_norm_var": 0.0016275854789366514,
"learning_rate": 5e-05,
"loss": 0.1814,
"loss/crossentropy": 2.961915969848633,
"loss/hidden": 0.0,
"loss/logits": 0.1813669353723526,
"loss/reg": 2.5074241161346436,
"step": 562
},
{
"epoch": 0.00563,
"grad_norm": 0.39113175868988037,
"grad_norm_var": 0.0016486631382784092,
"learning_rate": 5e-05,
"loss": 0.1643,
"loss/crossentropy": 2.7337673902511597,
"loss/hidden": 0.0,
"loss/logits": 0.164301548153162,
"loss/reg": 2.5046684741973877,
"step": 563
},
{
"epoch": 0.00564,
"grad_norm": 0.36300358176231384,
"grad_norm_var": 0.0016312765518430934,
"learning_rate": 5e-05,
"loss": 0.1602,
"loss/crossentropy": 2.713749051094055,
"loss/hidden": 0.0,
"loss/logits": 0.16021040827035904,
"loss/reg": 2.5020864009857178,
"step": 564
},
{
"epoch": 0.00565,
"grad_norm": 0.3250221312046051,
"grad_norm_var": 0.001185749693630452,
"learning_rate": 5e-05,
"loss": 0.1661,
"loss/crossentropy": 2.739534556865692,
"loss/hidden": 0.0,
"loss/logits": 0.166114691644907,
"loss/reg": 2.500089645385742,
"step": 565
},
{
"epoch": 0.00566,
"grad_norm": 0.3059675395488739,
"grad_norm_var": 0.0013809527139825861,
"learning_rate": 5e-05,
"loss": 0.1528,
"loss/crossentropy": 2.7676697373390198,
"loss/hidden": 0.0,
"loss/logits": 0.15275665000081062,
"loss/reg": 2.497802257537842,
"step": 566
},
{
"epoch": 0.00567,
"grad_norm": 0.41637444496154785,
"grad_norm_var": 0.0015720438674995396,
"learning_rate": 5e-05,
"loss": 0.1899,
"loss/crossentropy": 2.7852693796157837,
"loss/hidden": 0.0,
"loss/logits": 0.18987080082297325,
"loss/reg": 2.4960269927978516,
"step": 567
},
{
"epoch": 0.00568,
"grad_norm": 0.48216167092323303,
"grad_norm_var": 0.002316091582714641,
"learning_rate": 5e-05,
"loss": 0.179,
"loss/crossentropy": 2.919625759124756,
"loss/hidden": 0.0,
"loss/logits": 0.17901213094592094,
"loss/reg": 2.4943020343780518,
"step": 568
},
{
"epoch": 0.00569,
"grad_norm": 0.34773337841033936,
"grad_norm_var": 0.0023415161321106623,
"learning_rate": 5e-05,
"loss": 0.1689,
"loss/crossentropy": 2.829575002193451,
"loss/hidden": 0.0,
"loss/logits": 0.1688704527914524,
"loss/reg": 2.4922549724578857,
"step": 569
},
{
"epoch": 0.0057,
"grad_norm": 0.42466020584106445,
"grad_norm_var": 0.0022617987789910494,
"learning_rate": 5e-05,
"loss": 0.2065,
"loss/crossentropy": 2.847673773765564,
"loss/hidden": 0.0,
"loss/logits": 0.20647098124027252,
"loss/reg": 2.4896240234375,
"step": 570
},
{
"epoch": 0.00571,
"grad_norm": 0.39025840163230896,
"grad_norm_var": 0.0022035635328787567,
"learning_rate": 5e-05,
"loss": 0.181,
"loss/crossentropy": 2.9154597520828247,
"loss/hidden": 0.0,
"loss/logits": 0.1810290329158306,
"loss/reg": 2.487513542175293,
"step": 571
},
{
"epoch": 0.00572,
"grad_norm": 0.3611275851726532,
"grad_norm_var": 0.002232206094215346,
"learning_rate": 5e-05,
"loss": 0.1687,
"loss/crossentropy": 2.813008964061737,
"loss/hidden": 0.0,
"loss/logits": 0.16871189698576927,
"loss/reg": 2.4849367141723633,
"step": 572
},
{
"epoch": 0.00573,
"grad_norm": 0.37163245677948,
"grad_norm_var": 0.002205551603401897,
"learning_rate": 5e-05,
"loss": 0.1736,
"loss/crossentropy": 2.829798102378845,
"loss/hidden": 0.0,
"loss/logits": 0.173641849309206,
"loss/reg": 2.481811046600342,
"step": 573
},
{
"epoch": 0.00574,
"grad_norm": 0.37662971019744873,
"grad_norm_var": 0.002204250880404842,
"learning_rate": 5e-05,
"loss": 0.1641,
"loss/crossentropy": 2.786403477191925,
"loss/hidden": 0.0,
"loss/logits": 0.16413037478923798,
"loss/reg": 2.479344606399536,
"step": 574
},
{
"epoch": 0.00575,
"grad_norm": 0.4090428948402405,
"grad_norm_var": 0.002174681113915019,
"learning_rate": 5e-05,
"loss": 0.1684,
"loss/crossentropy": 2.685749888420105,
"loss/hidden": 0.0,
"loss/logits": 0.16840650886297226,
"loss/reg": 2.476419448852539,
"step": 575
},
{
"epoch": 0.00576,
"grad_norm": 0.35688483715057373,
"grad_norm_var": 0.0021995018187083346,
"learning_rate": 5e-05,
"loss": 0.1611,
"loss/crossentropy": 2.809792697429657,
"loss/hidden": 0.0,
"loss/logits": 0.1611488163471222,
"loss/reg": 2.4737355709075928,
"step": 576
},
{
"epoch": 0.00577,
"grad_norm": 0.38194504380226135,
"grad_norm_var": 0.002065385566742454,
"learning_rate": 5e-05,
"loss": 0.1615,
"loss/crossentropy": 2.850769340991974,
"loss/hidden": 0.0,
"loss/logits": 0.1615295149385929,
"loss/reg": 2.4713802337646484,
"step": 577
},
{
"epoch": 0.00578,
"grad_norm": 0.3567502200603485,
"grad_norm_var": 0.001743510873329986,
"learning_rate": 5e-05,
"loss": 0.1689,
"loss/crossentropy": 2.7103776335716248,
"loss/hidden": 0.0,
"loss/logits": 0.16894375160336494,
"loss/reg": 2.4685709476470947,
"step": 578
},
{
"epoch": 0.00579,
"grad_norm": 0.3396901786327362,
"grad_norm_var": 0.001824115359163836,
"learning_rate": 5e-05,
"loss": 0.166,
"loss/crossentropy": 2.7079854607582092,
"loss/hidden": 0.0,
"loss/logits": 0.1659584417939186,
"loss/reg": 2.465658664703369,
"step": 579
},
{
"epoch": 0.0058,
"grad_norm": 0.358395516872406,
"grad_norm_var": 0.0018331543648902808,
"learning_rate": 5e-05,
"loss": 0.1813,
"loss/crossentropy": 2.8853692412376404,
"loss/hidden": 0.0,
"loss/logits": 0.18133477121591568,
"loss/reg": 2.4633235931396484,
"step": 580
},
{
"epoch": 0.00581,
"grad_norm": 0.3434228301048279,
"grad_norm_var": 0.0017310432323107805,
"learning_rate": 5e-05,
"loss": 0.1739,
"loss/crossentropy": 2.666011691093445,
"loss/hidden": 0.0,
"loss/logits": 0.1739240102469921,
"loss/reg": 2.460447072982788,
"step": 581
},
{
"epoch": 0.00582,
"grad_norm": 0.3482820689678192,
"grad_norm_var": 0.0014454775261250163,
"learning_rate": 5e-05,
"loss": 0.1846,
"loss/crossentropy": 2.6244596242904663,
"loss/hidden": 0.0,
"loss/logits": 0.1845875158905983,
"loss/reg": 2.457307815551758,
"step": 582
},
{
"epoch": 0.00583,
"grad_norm": 0.36450985074043274,
"grad_norm_var": 0.0013555723186838029,
"learning_rate": 5e-05,
"loss": 0.1741,
"loss/crossentropy": 2.6909091472625732,
"loss/hidden": 0.0,
"loss/logits": 0.174148079007864,
"loss/reg": 2.4545810222625732,
"step": 583
},
{
"epoch": 0.00584,
"grad_norm": 0.34841907024383545,
"grad_norm_var": 0.0005772011049318792,
"learning_rate": 5e-05,
"loss": 0.1621,
"loss/crossentropy": 2.7992460131645203,
"loss/hidden": 0.0,
"loss/logits": 0.162076648324728,
"loss/reg": 2.4516048431396484,
"step": 584
},
{
"epoch": 0.00585,
"grad_norm": 0.36560627818107605,
"grad_norm_var": 0.0005501529366056079,
"learning_rate": 5e-05,
"loss": 0.1612,
"loss/crossentropy": 2.7556354999542236,
"loss/hidden": 0.0,
"loss/logits": 0.16119826585054398,
"loss/reg": 2.4482715129852295,
"step": 585
},
{
"epoch": 0.00586,
"grad_norm": 0.37393423914909363,
"grad_norm_var": 0.00033166715444868193,
"learning_rate": 5e-05,
"loss": 0.1779,
"loss/crossentropy": 2.6222774982452393,
"loss/hidden": 0.0,
"loss/logits": 0.1779084950685501,
"loss/reg": 2.4452362060546875,
"step": 586
},
{
"epoch": 0.00587,
"grad_norm": 0.3511587679386139,
"grad_norm_var": 0.0002976648126369145,
"learning_rate": 5e-05,
"loss": 0.1761,
"loss/crossentropy": 2.7342361211776733,
"loss/hidden": 0.0,
"loss/logits": 0.17614838480949402,
"loss/reg": 2.441678524017334,
"step": 587
},
{
"epoch": 0.00588,
"grad_norm": 0.33847615122795105,
"grad_norm_var": 0.0003352805276915209,
"learning_rate": 5e-05,
"loss": 0.173,
"loss/crossentropy": 2.7935328483581543,
"loss/hidden": 0.0,
"loss/logits": 0.17295999452471733,
"loss/reg": 2.437959671020508,
"step": 588
},
{
"epoch": 0.00589,
"grad_norm": 0.351034015417099,
"grad_norm_var": 0.00033410454836428903,
"learning_rate": 5e-05,
"loss": 0.1775,
"loss/crossentropy": 2.7590489387512207,
"loss/hidden": 0.0,
"loss/logits": 0.1775294505059719,
"loss/reg": 2.4346938133239746,
"step": 589
},
{
"epoch": 0.0059,
"grad_norm": 0.37800535559654236,
"grad_norm_var": 0.0003372250971240794,
"learning_rate": 5e-05,
"loss": 0.1645,
"loss/crossentropy": 2.75826096534729,
"loss/hidden": 0.0,
"loss/logits": 0.16445999220013618,
"loss/reg": 2.4319543838500977,
"step": 590
},
{
"epoch": 0.00591,
"grad_norm": 0.3323316276073456,
"grad_norm_var": 0.0002069473145354402,
"learning_rate": 5e-05,
"loss": 0.1644,
"loss/crossentropy": 2.963920295238495,
"loss/hidden": 0.0,
"loss/logits": 0.164412472397089,
"loss/reg": 2.4295387268066406,
"step": 591
},
{
"epoch": 0.00592,
"grad_norm": 0.8281128406524658,
"grad_norm_var": 0.014169124282143371,
"learning_rate": 5e-05,
"loss": 0.2256,
"loss/crossentropy": 2.9319988489151,
"loss/hidden": 0.0,
"loss/logits": 0.22560855001211166,
"loss/reg": 2.427125930786133,
"step": 592
},
{
"epoch": 0.00593,
"grad_norm": 0.37988972663879395,
"grad_norm_var": 0.014170226758262046,
"learning_rate": 5e-05,
"loss": 0.1795,
"loss/crossentropy": 2.92197585105896,
"loss/hidden": 0.0,
"loss/logits": 0.1794501654803753,
"loss/reg": 2.4247610569000244,
"step": 593
},
{
"epoch": 0.00594,
"grad_norm": 0.37449878454208374,
"grad_norm_var": 0.014123355612102569,
"learning_rate": 5e-05,
"loss": 0.1756,
"loss/crossentropy": 2.734030842781067,
"loss/hidden": 0.0,
"loss/logits": 0.17557094618678093,
"loss/reg": 2.4224398136138916,
"step": 594
},
{
"epoch": 0.00595,
"grad_norm": 0.3890518248081207,
"grad_norm_var": 0.013970946553029018,
"learning_rate": 5e-05,
"loss": 0.1721,
"loss/crossentropy": 2.6910988688468933,
"loss/hidden": 0.0,
"loss/logits": 0.17206770926713943,
"loss/reg": 2.4206151962280273,
"step": 595
},
{
"epoch": 0.00596,
"grad_norm": 0.45764538645744324,
"grad_norm_var": 0.014180672563351104,
"learning_rate": 5e-05,
"loss": 0.1886,
"loss/crossentropy": 2.706140458583832,
"loss/hidden": 0.0,
"loss/logits": 0.1886041909456253,
"loss/reg": 2.418341636657715,
"step": 596
},
{
"epoch": 0.00597,
"grad_norm": 0.3294787108898163,
"grad_norm_var": 0.014289226884282809,
"learning_rate": 5e-05,
"loss": 0.1693,
"loss/crossentropy": 2.772903263568878,
"loss/hidden": 0.0,
"loss/logits": 0.16925981268286705,
"loss/reg": 2.415613889694214,
"step": 597
},
{
"epoch": 0.00598,
"grad_norm": 0.3425086438655853,
"grad_norm_var": 0.014326812953815705,
"learning_rate": 5e-05,
"loss": 0.17,
"loss/crossentropy": 2.7024609446525574,
"loss/hidden": 0.0,
"loss/logits": 0.16995511576533318,
"loss/reg": 2.4122676849365234,
"step": 598
},
{
"epoch": 0.00599,
"grad_norm": 0.37222734093666077,
"grad_norm_var": 0.014300147579082,
"learning_rate": 5e-05,
"loss": 0.1789,
"loss/crossentropy": 2.8698896765708923,
"loss/hidden": 0.0,
"loss/logits": 0.17886632308363914,
"loss/reg": 2.4099037647247314,
"step": 599
},
{
"epoch": 0.006,
"grad_norm": 0.39135316014289856,
"grad_norm_var": 0.014151428197242365,
"learning_rate": 5e-05,
"loss": 0.1747,
"loss/crossentropy": 2.700629711151123,
"loss/hidden": 0.0,
"loss/logits": 0.17468373104929924,
"loss/reg": 2.40794038772583,
"step": 600
},
{
"epoch": 0.00601,
"grad_norm": 0.3728218376636505,
"grad_norm_var": 0.014124279912823712,
"learning_rate": 5e-05,
"loss": 0.169,
"loss/crossentropy": 2.794102430343628,
"loss/hidden": 0.0,
"loss/logits": 0.16897983103990555,
"loss/reg": 2.405545711517334,
"step": 601
},
{
"epoch": 0.00602,
"grad_norm": 0.37317147850990295,
"grad_norm_var": 0.014126729018321404,
"learning_rate": 5e-05,
"loss": 0.1747,
"loss/crossentropy": 2.6252577900886536,
"loss/hidden": 0.0,
"loss/logits": 0.17474820092320442,
"loss/reg": 2.402970790863037,
"step": 602
},
{
"epoch": 0.00603,
"grad_norm": 0.35492607951164246,
"grad_norm_var": 0.01410428304541661,
"learning_rate": 5e-05,
"loss": 0.1809,
"loss/crossentropy": 2.6527358889579773,
"loss/hidden": 0.0,
"loss/logits": 0.18093448877334595,
"loss/reg": 2.4007880687713623,
"step": 603
},
{
"epoch": 0.00604,
"grad_norm": 0.408010870218277,
"grad_norm_var": 0.013856041692491945,
"learning_rate": 5e-05,
"loss": 0.2018,
"loss/crossentropy": 2.874286651611328,
"loss/hidden": 0.0,
"loss/logits": 0.20177744701504707,
"loss/reg": 2.3988449573516846,
"step": 604
},
{
"epoch": 0.00605,
"grad_norm": 0.3291812837123871,
"grad_norm_var": 0.014034946168994126,
"learning_rate": 5e-05,
"loss": 0.1614,
"loss/crossentropy": 2.7926167249679565,
"loss/hidden": 0.0,
"loss/logits": 0.16140995919704437,
"loss/reg": 2.3966500759124756,
"step": 605
},
{
"epoch": 0.00606,
"grad_norm": 0.34659212827682495,
"grad_norm_var": 0.014192203001449558,
"learning_rate": 5e-05,
"loss": 0.1709,
"loss/crossentropy": 2.8195464611053467,
"loss/hidden": 0.0,
"loss/logits": 0.1709096021950245,
"loss/reg": 2.394033670425415,
"step": 606
},
{
"epoch": 0.00607,
"grad_norm": 0.32253992557525635,
"grad_norm_var": 0.014285055545239086,
"learning_rate": 5e-05,
"loss": 0.1649,
"loss/crossentropy": 2.7236337065696716,
"loss/hidden": 0.0,
"loss/logits": 0.16493552178144455,
"loss/reg": 2.3918449878692627,
"step": 607
},
{
"epoch": 0.00608,
"grad_norm": 0.350931316614151,
"grad_norm_var": 0.0011668026056699994,
"learning_rate": 5e-05,
"loss": 0.1718,
"loss/crossentropy": 2.800759196281433,
"loss/hidden": 0.0,
"loss/logits": 0.17175282910466194,
"loss/reg": 2.38920521736145,
"step": 608
},
{
"epoch": 0.00609,
"grad_norm": 0.40333986282348633,
"grad_norm_var": 0.001237012928824995,
"learning_rate": 5e-05,
"loss": 0.2046,
"loss/crossentropy": 2.7574119567871094,
"loss/hidden": 0.0,
"loss/logits": 0.2046247273683548,
"loss/reg": 2.3865227699279785,
"step": 609
},
{
"epoch": 0.0061,
"grad_norm": 0.3773089349269867,
"grad_norm_var": 0.0012392324335123346,
"learning_rate": 5e-05,
"loss": 0.1641,
"loss/crossentropy": 2.6313101649284363,
"loss/hidden": 0.0,
"loss/logits": 0.16410250216722488,
"loss/reg": 2.3836612701416016,
"step": 610
},
{
"epoch": 0.00611,
"grad_norm": 0.438357949256897,
"grad_norm_var": 0.0015159779907225465,
"learning_rate": 5e-05,
"loss": 0.2098,
"loss/crossentropy": 2.780138611793518,
"loss/hidden": 0.0,
"loss/logits": 0.20979087427258492,
"loss/reg": 2.380405902862549,
"step": 611
},
{
"epoch": 0.00612,
"grad_norm": 0.34121251106262207,
"grad_norm_var": 0.0010515226822608785,
"learning_rate": 5e-05,
"loss": 0.161,
"loss/crossentropy": 2.669090151786804,
"loss/hidden": 0.0,
"loss/logits": 0.1609921157360077,
"loss/reg": 2.3776426315307617,
"step": 612
},
{
"epoch": 0.00613,
"grad_norm": 0.36169829964637756,
"grad_norm_var": 0.0009600577824135296,
"learning_rate": 5e-05,
"loss": 0.1734,
"loss/crossentropy": 2.7925440073013306,
"loss/hidden": 0.0,
"loss/logits": 0.17342102900147438,
"loss/reg": 2.3744184970855713,
"step": 613
},
{
"epoch": 0.00614,
"grad_norm": 0.522160530090332,
"grad_norm_var": 0.002369345725690275,
"learning_rate": 5e-05,
"loss": 0.1663,
"loss/crossentropy": 2.698939800262451,
"loss/hidden": 0.0,
"loss/logits": 0.16627153754234314,
"loss/reg": 2.370917320251465,
"step": 614
},
{
"epoch": 0.00615,
"grad_norm": 0.4562234580516815,
"grad_norm_var": 0.002733171284208069,
"learning_rate": 5e-05,
"loss": 0.1686,
"loss/crossentropy": 2.8971627950668335,
"loss/hidden": 0.0,
"loss/logits": 0.16856613755226135,
"loss/reg": 2.368067979812622,
"step": 615
},
{
"epoch": 0.00616,
"grad_norm": 0.5767056345939636,
"grad_norm_var": 0.0050531115809510415,
"learning_rate": 5e-05,
"loss": 0.171,
"loss/crossentropy": 2.822002112865448,
"loss/hidden": 0.0,
"loss/logits": 0.17102698609232903,
"loss/reg": 2.3653640747070312,
"step": 616
},
{
"epoch": 0.00617,
"grad_norm": 0.3703908324241638,
"grad_norm_var": 0.005060977204500819,
"learning_rate": 5e-05,
"loss": 0.1799,
"loss/crossentropy": 2.761395037174225,
"loss/hidden": 0.0,
"loss/logits": 0.17990661412477493,
"loss/reg": 2.362797975540161,
"step": 617
},
{
"epoch": 0.00618,
"grad_norm": 0.44375622272491455,
"grad_norm_var": 0.005159430065949637,
"learning_rate": 5e-05,
"loss": 0.1662,
"loss/crossentropy": 2.8019450306892395,
"loss/hidden": 0.0,
"loss/logits": 0.1661831997334957,
"loss/reg": 2.3597350120544434,
"step": 618
},
{
"epoch": 0.00619,
"grad_norm": 0.41226035356521606,
"grad_norm_var": 0.005018716701479123,
"learning_rate": 5e-05,
"loss": 0.1737,
"loss/crossentropy": 2.837542712688446,
"loss/hidden": 0.0,
"loss/logits": 0.1737065464258194,
"loss/reg": 2.356935977935791,
"step": 619
},
{
"epoch": 0.0062,
"grad_norm": 0.36850520968437195,
"grad_norm_var": 0.005094037089036248,
"learning_rate": 5e-05,
"loss": 0.1691,
"loss/crossentropy": 2.872538685798645,
"loss/hidden": 0.0,
"loss/logits": 0.16909406706690788,
"loss/reg": 2.354841709136963,
"step": 620
},
{
"epoch": 0.00621,
"grad_norm": 0.3547448217868805,
"grad_norm_var": 0.004888988248098869,
"learning_rate": 5e-05,
"loss": 0.1777,
"loss/crossentropy": 2.727312922477722,
"loss/hidden": 0.0,
"loss/logits": 0.17773358151316643,
"loss/reg": 2.3518083095550537,
"step": 621
},
{
"epoch": 0.00622,
"grad_norm": 0.3340252637863159,
"grad_norm_var": 0.0049932414292845895,
"learning_rate": 5e-05,
"loss": 0.1673,
"loss/crossentropy": 2.7399535179138184,
"loss/hidden": 0.0,
"loss/logits": 0.16725115478038788,
"loss/reg": 2.349299907684326,
"step": 622
},
{
"epoch": 0.00623,
"grad_norm": 0.328477680683136,
"grad_norm_var": 0.004932429457390519,
"learning_rate": 5e-05,
"loss": 0.1658,
"loss/crossentropy": 2.7973376512527466,
"loss/hidden": 0.0,
"loss/logits": 0.1658070906996727,
"loss/reg": 2.346407175064087,
"step": 623
},
{
"epoch": 0.00624,
"grad_norm": 0.3988572061061859,
"grad_norm_var": 0.004746415643168555,
"learning_rate": 5e-05,
"loss": 0.1808,
"loss/crossentropy": 2.886197090148926,
"loss/hidden": 0.0,
"loss/logits": 0.18076446652412415,
"loss/reg": 2.343637228012085,
"step": 624
},
{
"epoch": 0.00625,
"grad_norm": 0.3653312921524048,
"grad_norm_var": 0.0048476613679717525,
"learning_rate": 5e-05,
"loss": 0.1752,
"loss/crossentropy": 2.6095593571662903,
"loss/hidden": 0.0,
"loss/logits": 0.17522242665290833,
"loss/reg": 2.34155011177063,
"step": 625
},
{
"epoch": 0.00626,
"grad_norm": 0.3519672751426697,
"grad_norm_var": 0.004975031863489395,
"learning_rate": 5e-05,
"loss": 0.1685,
"loss/crossentropy": 2.724495232105255,
"loss/hidden": 0.0,
"loss/logits": 0.16851425543427467,
"loss/reg": 2.339081048965454,
"step": 626
},
{
"epoch": 0.00627,
"grad_norm": 0.3507337272167206,
"grad_norm_var": 0.005024779798457324,
"learning_rate": 5e-05,
"loss": 0.1661,
"loss/crossentropy": 2.775688886642456,
"loss/hidden": 0.0,
"loss/logits": 0.1661130003631115,
"loss/reg": 2.3362221717834473,
"step": 627
},
{
"epoch": 0.00628,
"grad_norm": 0.35331088304519653,
"grad_norm_var": 0.004945443478871292,
"learning_rate": 5e-05,
"loss": 0.1803,
"loss/crossentropy": 2.6876689195632935,
"loss/hidden": 0.0,
"loss/logits": 0.18032584339380264,
"loss/reg": 2.3339033126831055,
"step": 628
},
{
"epoch": 0.00629,
"grad_norm": 0.3569658696651459,
"grad_norm_var": 0.004969005818722216,
"learning_rate": 5e-05,
"loss": 0.1646,
"loss/crossentropy": 2.87895804643631,
"loss/hidden": 0.0,
"loss/logits": 0.16462786123156548,
"loss/reg": 2.3317511081695557,
"step": 629
},
{
"epoch": 0.0063,
"grad_norm": 0.37102508544921875,
"grad_norm_var": 0.0038649155689368443,
"learning_rate": 5e-05,
"loss": 0.1807,
"loss/crossentropy": 2.8995742201805115,
"loss/hidden": 0.0,
"loss/logits": 0.1807471290230751,
"loss/reg": 2.3286077976226807,
"step": 630
},
{
"epoch": 0.00631,
"grad_norm": 0.37091144919395447,
"grad_norm_var": 0.0035332975201383715,
"learning_rate": 5e-05,
"loss": 0.1679,
"loss/crossentropy": 2.755174398422241,
"loss/hidden": 0.0,
"loss/logits": 0.16785116121172905,
"loss/reg": 2.3263347148895264,
"step": 631
},
{
"epoch": 0.00632,
"grad_norm": 0.3764369487762451,
"grad_norm_var": 0.000834165955391919,
"learning_rate": 5e-05,
"loss": 0.1597,
"loss/crossentropy": 2.7826634645462036,
"loss/hidden": 0.0,
"loss/logits": 0.1596829891204834,
"loss/reg": 2.3239433765411377,
"step": 632
},
{
"epoch": 0.00633,
"grad_norm": 0.34151408076286316,
"grad_norm_var": 0.0008818179956038841,
"learning_rate": 5e-05,
"loss": 0.1628,
"loss/crossentropy": 2.805456221103668,
"loss/hidden": 0.0,
"loss/logits": 0.16277796775102615,
"loss/reg": 2.3215837478637695,
"step": 633
},
{
"epoch": 0.00634,
"grad_norm": 0.7558053731918335,
"grad_norm_var": 0.010143553337954326,
"learning_rate": 5e-05,
"loss": 0.1858,
"loss/crossentropy": 2.7678999304771423,
"loss/hidden": 0.0,
"loss/logits": 0.18575545772910118,
"loss/reg": 2.3192989826202393,
"step": 634
},
{
"epoch": 0.00635,
"grad_norm": 0.3809748589992523,
"grad_norm_var": 0.010099062255010161,
"learning_rate": 5e-05,
"loss": 0.1792,
"loss/crossentropy": 2.86500483751297,
"loss/hidden": 0.0,
"loss/logits": 0.1791505441069603,
"loss/reg": 2.317030906677246,
"step": 635
},
{
"epoch": 0.00636,
"grad_norm": 0.40578746795654297,
"grad_norm_var": 0.010104068412990375,
"learning_rate": 5e-05,
"loss": 0.1882,
"loss/crossentropy": 2.8707818388938904,
"loss/hidden": 0.0,
"loss/logits": 0.188164584338665,
"loss/reg": 2.3146989345550537,
"step": 636
},
{
"epoch": 0.00637,
"grad_norm": 0.415227472782135,
"grad_norm_var": 0.010070131470069877,
"learning_rate": 5e-05,
"loss": 0.1748,
"loss/crossentropy": 2.831197440624237,
"loss/hidden": 0.0,
"loss/logits": 0.17476488277316093,
"loss/reg": 2.311936616897583,
"step": 637
},
{
"epoch": 0.00638,
"grad_norm": 0.4119730293750763,
"grad_norm_var": 0.00985685373482662,
"learning_rate": 5e-05,
"loss": 0.1699,
"loss/crossentropy": 2.6559138894081116,
"loss/hidden": 0.0,
"loss/logits": 0.16989587992429733,
"loss/reg": 2.3090522289276123,
"step": 638
},
{
"epoch": 0.00639,
"grad_norm": 0.3662709593772888,
"grad_norm_var": 0.009606093056996168,
"learning_rate": 5e-05,
"loss": 0.1775,
"loss/crossentropy": 2.767539858818054,
"loss/hidden": 0.0,
"loss/logits": 0.17754964902997017,
"loss/reg": 2.3056743144989014,
"step": 639
},
{
"epoch": 0.0064,
"grad_norm": 0.38491374254226685,
"grad_norm_var": 0.009617242443139995,
"learning_rate": 5e-05,
"loss": 0.1827,
"loss/crossentropy": 2.6669586896896362,
"loss/hidden": 0.0,
"loss/logits": 0.18266603723168373,
"loss/reg": 2.303258180618286,
"step": 640
},
{
"epoch": 0.00641,
"grad_norm": 0.4197373390197754,
"grad_norm_var": 0.009569272862985524,
"learning_rate": 5e-05,
"loss": 0.1778,
"loss/crossentropy": 2.7964502573013306,
"loss/hidden": 0.0,
"loss/logits": 0.17782465368509293,
"loss/reg": 2.300361156463623,
"step": 641
},
{
"epoch": 0.00642,
"grad_norm": 0.4097757339477539,
"grad_norm_var": 0.00940137989136159,
"learning_rate": 5e-05,
"loss": 0.1856,
"loss/crossentropy": 2.735614001750946,
"loss/hidden": 0.0,
"loss/logits": 0.1855894774198532,
"loss/reg": 2.2972419261932373,
"step": 642
},
{
"epoch": 0.00643,
"grad_norm": 0.35904356837272644,
"grad_norm_var": 0.00934616788177974,
"learning_rate": 5e-05,
"loss": 0.1833,
"loss/crossentropy": 2.774403393268585,
"loss/hidden": 0.0,
"loss/logits": 0.1832551322877407,
"loss/reg": 2.293954610824585,
"step": 643
},
{
"epoch": 0.00644,
"grad_norm": 0.34157049655914307,
"grad_norm_var": 0.009435664127140328,
"learning_rate": 5e-05,
"loss": 0.1618,
"loss/crossentropy": 2.8616234064102173,
"loss/hidden": 0.0,
"loss/logits": 0.16181085631251335,
"loss/reg": 2.2906179428100586,
"step": 644
},
{
"epoch": 0.00645,
"grad_norm": 0.4255986213684082,
"grad_norm_var": 0.009297406924193945,
"learning_rate": 5e-05,
"loss": 0.1853,
"loss/crossentropy": 2.654071033000946,
"loss/hidden": 0.0,
"loss/logits": 0.18530349805951118,
"loss/reg": 2.287349224090576,
"step": 645
},
{
"epoch": 0.00646,
"grad_norm": 0.3393001854419708,
"grad_norm_var": 0.009518979339113423,
"learning_rate": 5e-05,
"loss": 0.1665,
"loss/crossentropy": 2.73319810628891,
"loss/hidden": 0.0,
"loss/logits": 0.16647282242774963,
"loss/reg": 2.2834725379943848,
"step": 646
},
{
"epoch": 0.00647,
"grad_norm": 0.34969252347946167,
"grad_norm_var": 0.00964795505733251,
"learning_rate": 5e-05,
"loss": 0.1833,
"loss/crossentropy": 2.7993005514144897,
"loss/hidden": 0.0,
"loss/logits": 0.1833389550447464,
"loss/reg": 2.2799694538116455,
"step": 647
},
{
"epoch": 0.00648,
"grad_norm": 0.35388484597206116,
"grad_norm_var": 0.009766310746661707,
"learning_rate": 5e-05,
"loss": 0.1749,
"loss/crossentropy": 2.766145169734955,
"loss/hidden": 0.0,
"loss/logits": 0.17488964274525642,
"loss/reg": 2.277585029602051,
"step": 648
},
{
"epoch": 0.00649,
"grad_norm": 0.5462765097618103,
"grad_norm_var": 0.010685818975949597,
"learning_rate": 5e-05,
"loss": 0.1882,
"loss/crossentropy": 2.7475533485412598,
"loss/hidden": 0.0,
"loss/logits": 0.18824508786201477,
"loss/reg": 2.2750473022460938,
"step": 649
},
{
"epoch": 0.0065,
"grad_norm": 0.3537692725658417,
"grad_norm_var": 0.002605622083243005,
"learning_rate": 5e-05,
"loss": 0.1761,
"loss/crossentropy": 2.756273865699768,
"loss/hidden": 0.0,
"loss/logits": 0.17612234875559807,
"loss/reg": 2.2722957134246826,
"step": 650
},
{
"epoch": 0.00651,
"grad_norm": 0.3770252466201782,
"grad_norm_var": 0.0026121330513858157,
"learning_rate": 5e-05,
"loss": 0.1897,
"loss/crossentropy": 2.7889973521232605,
"loss/hidden": 0.0,
"loss/logits": 0.1897362545132637,
"loss/reg": 2.2701008319854736,
"step": 651
},
{
"epoch": 0.00652,
"grad_norm": 0.4475138187408447,
"grad_norm_var": 0.0028018836674080227,
"learning_rate": 5e-05,
"loss": 0.1951,
"loss/crossentropy": 2.531024992465973,
"loss/hidden": 0.0,
"loss/logits": 0.1951226033270359,
"loss/reg": 2.267695665359497,
"step": 652
},
{
"epoch": 0.00653,
"grad_norm": 0.3947466313838959,
"grad_norm_var": 0.002769718525090366,
"learning_rate": 5e-05,
"loss": 0.1958,
"loss/crossentropy": 2.886034905910492,
"loss/hidden": 0.0,
"loss/logits": 0.19575949385762215,
"loss/reg": 2.2648732662200928,
"step": 653
},
{
"epoch": 0.00654,
"grad_norm": 0.3775857090950012,
"grad_norm_var": 0.0027546537142078996,
"learning_rate": 5e-05,
"loss": 0.1746,
"loss/crossentropy": 2.7190786600112915,
"loss/hidden": 0.0,
"loss/logits": 0.17461128905415535,
"loss/reg": 2.2620925903320312,
"step": 654
},
{
"epoch": 0.00655,
"grad_norm": 0.34534481167793274,
"grad_norm_var": 0.002849399631435645,
"learning_rate": 5e-05,
"loss": 0.1792,
"loss/crossentropy": 2.8847506046295166,
"loss/hidden": 0.0,
"loss/logits": 0.1791832633316517,
"loss/reg": 2.2593743801116943,
"step": 655
},
{
"epoch": 0.00656,
"grad_norm": 0.3607633411884308,
"grad_norm_var": 0.0028993682580486144,
"learning_rate": 5e-05,
"loss": 0.1792,
"loss/crossentropy": 2.815674066543579,
"loss/hidden": 0.0,
"loss/logits": 0.179163109511137,
"loss/reg": 2.2569730281829834,
"step": 656
},
{
"epoch": 0.00657,
"grad_norm": 0.38781270384788513,
"grad_norm_var": 0.002826278400635814,
"learning_rate": 5e-05,
"loss": 0.1618,
"loss/crossentropy": 2.6106160283088684,
"loss/hidden": 0.0,
"loss/logits": 0.1617795117199421,
"loss/reg": 2.254523277282715,
"step": 657
},
{
"epoch": 0.00658,
"grad_norm": 0.40386784076690674,
"grad_norm_var": 0.0028094212847462165,
"learning_rate": 5e-05,
"loss": 0.1747,
"loss/crossentropy": 2.8471227884292603,
"loss/hidden": 0.0,
"loss/logits": 0.17474086582660675,
"loss/reg": 2.252164125442505,
"step": 658
},
{
"epoch": 0.00659,
"grad_norm": 0.36319243907928467,
"grad_norm_var": 0.002796007207749466,
"learning_rate": 5e-05,
"loss": 0.163,
"loss/crossentropy": 2.7625906467437744,
"loss/hidden": 0.0,
"loss/logits": 0.16296877712011337,
"loss/reg": 2.249460458755493,
"step": 659
},
{
"epoch": 0.0066,
"grad_norm": 0.3657222092151642,
"grad_norm_var": 0.00269101182172804,
"learning_rate": 5e-05,
"loss": 0.1777,
"loss/crossentropy": 2.781547486782074,
"loss/hidden": 0.0,
"loss/logits": 0.1777149885892868,
"loss/reg": 2.246467113494873,
"step": 660
},
{
"epoch": 0.00661,
"grad_norm": 0.38363057374954224,
"grad_norm_var": 0.0025851401210759276,
"learning_rate": 5e-05,
"loss": 0.1844,
"loss/crossentropy": 2.82689893245697,
"loss/hidden": 0.0,
"loss/logits": 0.1844283789396286,
"loss/reg": 2.2436530590057373,
"step": 661
},
{
"epoch": 0.00662,
"grad_norm": 0.4096749424934387,
"grad_norm_var": 0.0024716520181473594,
"learning_rate": 5e-05,
"loss": 0.1745,
"loss/crossentropy": 2.8063756823539734,
"loss/hidden": 0.0,
"loss/logits": 0.17451731115579605,
"loss/reg": 2.241178035736084,
"step": 662
},
{
"epoch": 0.00663,
"grad_norm": 0.42931249737739563,
"grad_norm_var": 0.0024528927297352344,
"learning_rate": 5e-05,
"loss": 0.186,
"loss/crossentropy": 2.8724401593208313,
"loss/hidden": 0.0,
"loss/logits": 0.1859952136874199,
"loss/reg": 2.2383124828338623,
"step": 663
},
{
"epoch": 0.00664,
"grad_norm": 0.3530314862728119,
"grad_norm_var": 0.0024574750299312478,
"learning_rate": 5e-05,
"loss": 0.1696,
"loss/crossentropy": 2.9292226433753967,
"loss/hidden": 0.0,
"loss/logits": 0.16955319419503212,
"loss/reg": 2.2356791496276855,
"step": 664
},
{
"epoch": 0.00665,
"grad_norm": 0.4304611384868622,
"grad_norm_var": 0.0009397736187397402,
"learning_rate": 5e-05,
"loss": 0.1902,
"loss/crossentropy": 2.7114855647087097,
"loss/hidden": 0.0,
"loss/logits": 0.19023016840219498,
"loss/reg": 2.2330329418182373,
"step": 665
},
{
"epoch": 0.00666,
"grad_norm": 0.32996541261672974,
"grad_norm_var": 0.0010789617804694747,
"learning_rate": 5e-05,
"loss": 0.158,
"loss/crossentropy": 2.8920618891716003,
"loss/hidden": 0.0,
"loss/logits": 0.1580132134258747,
"loss/reg": 2.2296440601348877,
"step": 666
},
{
"epoch": 0.00667,
"grad_norm": 0.3874596953392029,
"grad_norm_var": 0.0010747020479673205,
"learning_rate": 5e-05,
"loss": 0.1819,
"loss/crossentropy": 2.7297377586364746,
"loss/hidden": 0.0,
"loss/logits": 0.18189727514982224,
"loss/reg": 2.226966619491577,
"step": 667
},
{
"epoch": 0.00668,
"grad_norm": 0.36097773909568787,
"grad_norm_var": 0.000828712243429038,
"learning_rate": 5e-05,
"loss": 0.1726,
"loss/crossentropy": 2.6433697938919067,
"loss/hidden": 0.0,
"loss/logits": 0.17256683483719826,
"loss/reg": 2.2244138717651367,
"step": 668
},
{
"epoch": 0.00669,
"grad_norm": 0.3509676158428192,
"grad_norm_var": 0.0008637156407869958,
"learning_rate": 5e-05,
"loss": 0.1723,
"loss/crossentropy": 2.8315157890319824,
"loss/hidden": 0.0,
"loss/logits": 0.17234884947538376,
"loss/reg": 2.220712661743164,
"step": 669
},
{
"epoch": 0.0067,
"grad_norm": 0.3578469157218933,
"grad_norm_var": 0.0008878035089742793,
"learning_rate": 5e-05,
"loss": 0.1657,
"loss/crossentropy": 2.788190722465515,
"loss/hidden": 0.0,
"loss/logits": 0.16572094336152077,
"loss/reg": 2.217878818511963,
"step": 670
},
{
"epoch": 0.00671,
"grad_norm": 0.4930081069469452,
"grad_norm_var": 0.0016420680378558003,
"learning_rate": 5e-05,
"loss": 0.1818,
"loss/crossentropy": 3.013857901096344,
"loss/hidden": 0.0,
"loss/logits": 0.18175816163420677,
"loss/reg": 2.2148284912109375,
"step": 671
},
{
"epoch": 0.00672,
"grad_norm": 0.36925604939460754,
"grad_norm_var": 0.0016185866984450236,
"learning_rate": 5e-05,
"loss": 0.1642,
"loss/crossentropy": 2.8940696716308594,
"loss/hidden": 0.0,
"loss/logits": 0.1642276532948017,
"loss/reg": 2.2119295597076416,
"step": 672
},
{
"epoch": 0.00673,
"grad_norm": 0.4327005445957184,
"grad_norm_var": 0.0017552981165500747,
"learning_rate": 5e-05,
"loss": 0.1742,
"loss/crossentropy": 2.87309467792511,
"loss/hidden": 0.0,
"loss/logits": 0.17423933744430542,
"loss/reg": 2.209021806716919,
"step": 673
},
{
"epoch": 0.00674,
"grad_norm": 0.738524854183197,
"grad_norm_var": 0.009426579051544037,
"learning_rate": 5e-05,
"loss": 0.1868,
"loss/crossentropy": 2.8040258288383484,
"loss/hidden": 0.0,
"loss/logits": 0.1867678351700306,
"loss/reg": 2.2066619396209717,
"step": 674
},
{
"epoch": 0.00675,
"grad_norm": 0.4364205002784729,
"grad_norm_var": 0.009307313279513674,
"learning_rate": 5e-05,
"loss": 0.1796,
"loss/crossentropy": 2.718536138534546,
"loss/hidden": 0.0,
"loss/logits": 0.17956989258527756,
"loss/reg": 2.203990936279297,
"step": 675
},
{
"epoch": 0.00676,
"grad_norm": 0.41067376732826233,
"grad_norm_var": 0.009142390414932911,
"learning_rate": 5e-05,
"loss": 0.1642,
"loss/crossentropy": 2.72940456867218,
"loss/hidden": 0.0,
"loss/logits": 0.16421709582209587,
"loss/reg": 2.2013046741485596,
"step": 676
},
{
"epoch": 0.00677,
"grad_norm": 0.4327182173728943,
"grad_norm_var": 0.009073804614162174,
"learning_rate": 5e-05,
"loss": 0.1905,
"loss/crossentropy": 2.7371246814727783,
"loss/hidden": 0.0,
"loss/logits": 0.19054419547319412,
"loss/reg": 2.1991653442382812,
"step": 677
},
{
"epoch": 0.00678,
"grad_norm": 0.3779783844947815,
"grad_norm_var": 0.009181024716334075,
"learning_rate": 5e-05,
"loss": 0.1589,
"loss/crossentropy": 2.7500953674316406,
"loss/hidden": 0.0,
"loss/logits": 0.15891055390238762,
"loss/reg": 2.197261333465576,
"step": 678
},
{
"epoch": 0.00679,
"grad_norm": 0.3585035502910614,
"grad_norm_var": 0.009389539404841711,
"learning_rate": 5e-05,
"loss": 0.1836,
"loss/crossentropy": 2.869826376438141,
"loss/hidden": 0.0,
"loss/logits": 0.18359991908073425,
"loss/reg": 2.195239782333374,
"step": 679
},
{
"epoch": 0.0068,
"grad_norm": 0.3534944951534271,
"grad_norm_var": 0.009385802469305704,
"learning_rate": 5e-05,
"loss": 0.1689,
"loss/crossentropy": 2.8476794362068176,
"loss/hidden": 0.0,
"loss/logits": 0.1689487136900425,
"loss/reg": 2.1929402351379395,
"step": 680
},
{
"epoch": 0.00681,
"grad_norm": 0.3718988001346588,
"grad_norm_var": 0.009470130435250168,
"learning_rate": 5e-05,
"loss": 0.1704,
"loss/crossentropy": 2.7930009365081787,
"loss/hidden": 0.0,
"loss/logits": 0.17038631066679955,
"loss/reg": 2.19075608253479,
"step": 681
},
{
"epoch": 0.00682,
"grad_norm": 0.4854961037635803,
"grad_norm_var": 0.009319177707927173,
"learning_rate": 5e-05,
"loss": 0.1705,
"loss/crossentropy": 2.8028470277786255,
"loss/hidden": 0.0,
"loss/logits": 0.17054682224988937,
"loss/reg": 2.1886627674102783,
"step": 682
},
{
"epoch": 0.00683,
"grad_norm": 0.3880312144756317,
"grad_norm_var": 0.00931672834921676,
"learning_rate": 5e-05,
"loss": 0.1766,
"loss/crossentropy": 2.963544547557831,
"loss/hidden": 0.0,
"loss/logits": 0.17662956938147545,
"loss/reg": 2.1860859394073486,
"step": 683
},
{
"epoch": 0.00684,
"grad_norm": 0.3488878309726715,
"grad_norm_var": 0.009420855437860176,
"learning_rate": 5e-05,
"loss": 0.1667,
"loss/crossentropy": 2.959736704826355,
"loss/hidden": 0.0,
"loss/logits": 0.16670886427164078,
"loss/reg": 2.183668375015259,
"step": 684
},
{
"epoch": 0.00685,
"grad_norm": 0.8154363632202148,
"grad_norm_var": 0.018681551405985854,
"learning_rate": 5e-05,
"loss": 0.2213,
"loss/crossentropy": 2.911233067512512,
"loss/hidden": 0.0,
"loss/logits": 0.2212524674832821,
"loss/reg": 2.1813154220581055,
"step": 685
},
{
"epoch": 0.00686,
"grad_norm": 0.4155946969985962,
"grad_norm_var": 0.018194440840509217,
"learning_rate": 5e-05,
"loss": 0.1964,
"loss/crossentropy": 2.7186298966407776,
"loss/hidden": 0.0,
"loss/logits": 0.19640850275754929,
"loss/reg": 2.1795501708984375,
"step": 686
},
{
"epoch": 0.00687,
"grad_norm": 0.38160914182662964,
"grad_norm_var": 0.01835781299917098,
"learning_rate": 5e-05,
"loss": 0.1708,
"loss/crossentropy": 2.7301290035247803,
"loss/hidden": 0.0,
"loss/logits": 0.17081937566399574,
"loss/reg": 2.177623748779297,
"step": 687
},
{
"epoch": 0.00688,
"grad_norm": 0.41628003120422363,
"grad_norm_var": 0.01802219976069038,
"learning_rate": 5e-05,
"loss": 0.1771,
"loss/crossentropy": 2.7876546382904053,
"loss/hidden": 0.0,
"loss/logits": 0.17706667259335518,
"loss/reg": 2.175504446029663,
"step": 688
},
{
"epoch": 0.00689,
"grad_norm": 0.4177417755126953,
"grad_norm_var": 0.018066232212721724,
"learning_rate": 5e-05,
"loss": 0.1786,
"loss/crossentropy": 2.763257145881653,
"loss/hidden": 0.0,
"loss/logits": 0.1785966381430626,
"loss/reg": 2.173213481903076,
"step": 689
},
{
"epoch": 0.0069,
"grad_norm": 0.3603265583515167,
"grad_norm_var": 0.012296751904473697,
"learning_rate": 5e-05,
"loss": 0.1678,
"loss/crossentropy": 2.774847447872162,
"loss/hidden": 0.0,
"loss/logits": 0.16783085092902184,
"loss/reg": 2.1712427139282227,
"step": 690
},
{
"epoch": 0.00691,
"grad_norm": 0.4307333827018738,
"grad_norm_var": 0.01228874334383105,
"learning_rate": 5e-05,
"loss": 0.2034,
"loss/crossentropy": 2.6488924622535706,
"loss/hidden": 0.0,
"loss/logits": 0.20339511707425117,
"loss/reg": 2.170015573501587,
"step": 691
},
{
"epoch": 0.00692,
"grad_norm": 0.3678703010082245,
"grad_norm_var": 0.012472673417673882,
"learning_rate": 5e-05,
"loss": 0.1793,
"loss/crossentropy": 2.8285900950431824,
"loss/hidden": 0.0,
"loss/logits": 0.179282795637846,
"loss/reg": 2.1680784225463867,
"step": 692
},
{
"epoch": 0.00693,
"grad_norm": 0.3516632914543152,
"grad_norm_var": 0.012747599104723136,
"learning_rate": 5e-05,
"loss": 0.1638,
"loss/crossentropy": 2.72187340259552,
"loss/hidden": 0.0,
"loss/logits": 0.16377655416727066,
"loss/reg": 2.166708469390869,
"step": 693
},
{
"epoch": 0.00694,
"grad_norm": 0.37773895263671875,
"grad_norm_var": 0.012748787659448176,
"learning_rate": 5e-05,
"loss": 0.2,
"loss/crossentropy": 2.5079989433288574,
"loss/hidden": 0.0,
"loss/logits": 0.19995050877332687,
"loss/reg": 2.1644845008850098,
"step": 694
},
{
"epoch": 0.00695,
"grad_norm": 0.33557403087615967,
"grad_norm_var": 0.012954622340141124,
"learning_rate": 5e-05,
"loss": 0.173,
"loss/crossentropy": 2.733457326889038,
"loss/hidden": 0.0,
"loss/logits": 0.1730196811258793,
"loss/reg": 2.162649631500244,
"step": 695
},
{
"epoch": 0.00696,
"grad_norm": 0.3414340615272522,
"grad_norm_var": 0.01306044443406886,
"learning_rate": 5e-05,
"loss": 0.1699,
"loss/crossentropy": 2.770694136619568,
"loss/hidden": 0.0,
"loss/logits": 0.1698729656636715,
"loss/reg": 2.1605873107910156,
"step": 696
},
{
"epoch": 0.00697,
"grad_norm": 0.39742037653923035,
"grad_norm_var": 0.012961649579914787,
"learning_rate": 5e-05,
"loss": 0.1753,
"loss/crossentropy": 2.747798502445221,
"loss/hidden": 0.0,
"loss/logits": 0.17528066039085388,
"loss/reg": 2.158661127090454,
"step": 697
},
{
"epoch": 0.00698,
"grad_norm": 0.4672209620475769,
"grad_norm_var": 0.012809503544980934,
"learning_rate": 5e-05,
"loss": 0.1961,
"loss/crossentropy": 2.764335811138153,
"loss/hidden": 0.0,
"loss/logits": 0.1961456499993801,
"loss/reg": 2.157139539718628,
"step": 698
},
{
"epoch": 0.00699,
"grad_norm": 0.40900057554244995,
"grad_norm_var": 0.012766202979620484,
"learning_rate": 5e-05,
"loss": 0.1826,
"loss/crossentropy": 2.9526583552360535,
"loss/hidden": 0.0,
"loss/logits": 0.1826096773147583,
"loss/reg": 2.1556172370910645,
"step": 699
},
{
"epoch": 0.007,
"grad_norm": 0.45763787627220154,
"grad_norm_var": 0.01255169197725956,
"learning_rate": 5e-05,
"loss": 0.1845,
"loss/crossentropy": 2.9059385657310486,
"loss/hidden": 0.0,
"loss/logits": 0.18454358726739883,
"loss/reg": 2.1542842388153076,
"step": 700
},
{
"epoch": 0.00701,
"grad_norm": 0.568651020526886,
"grad_norm_var": 0.0033942912710514268,
"learning_rate": 5e-05,
"loss": 0.1801,
"loss/crossentropy": 2.771495759487152,
"loss/hidden": 0.0,
"loss/logits": 0.18007055297493935,
"loss/reg": 2.1522164344787598,
"step": 701
},
{
"epoch": 0.00702,
"grad_norm": 0.3590672016143799,
"grad_norm_var": 0.00352192003862181,
"learning_rate": 5e-05,
"loss": 0.1651,
"loss/crossentropy": 2.750881016254425,
"loss/hidden": 0.0,
"loss/logits": 0.1650897115468979,
"loss/reg": 2.1509296894073486,
"step": 702
},
{
"epoch": 0.00703,
"grad_norm": 0.36948493123054504,
"grad_norm_var": 0.0035648755964216056,
"learning_rate": 5e-05,
"loss": 0.1785,
"loss/crossentropy": 2.8027891516685486,
"loss/hidden": 0.0,
"loss/logits": 0.17848948016762733,
"loss/reg": 2.149231195449829,
"step": 703
},
{
"epoch": 0.00704,
"grad_norm": 0.3613908588886261,
"grad_norm_var": 0.0036467673242235915,
"learning_rate": 5e-05,
"loss": 0.1682,
"loss/crossentropy": 2.763719141483307,
"loss/hidden": 0.0,
"loss/logits": 0.16823140904307365,
"loss/reg": 2.1478073596954346,
"step": 704
},
{
"epoch": 0.00705,
"grad_norm": 0.38240060210227966,
"grad_norm_var": 0.003633263034560896,
"learning_rate": 5e-05,
"loss": 0.178,
"loss/crossentropy": 2.705716133117676,
"loss/hidden": 0.0,
"loss/logits": 0.1780022643506527,
"loss/reg": 2.1462342739105225,
"step": 705
},
{
"epoch": 0.00706,
"grad_norm": 0.3587467074394226,
"grad_norm_var": 0.0036409547879681387,
"learning_rate": 5e-05,
"loss": 0.1658,
"loss/crossentropy": 2.7772558331489563,
"loss/hidden": 0.0,
"loss/logits": 0.16575098782777786,
"loss/reg": 2.144062042236328,
"step": 706
},
{
"epoch": 0.00707,
"grad_norm": 0.36025822162628174,
"grad_norm_var": 0.0036250184261099458,
"learning_rate": 5e-05,
"loss": 0.1724,
"loss/crossentropy": 2.634014904499054,
"loss/hidden": 0.0,
"loss/logits": 0.17235567048192024,
"loss/reg": 2.1429154872894287,
"step": 707
},
{
"epoch": 0.00708,
"grad_norm": 0.35575759410858154,
"grad_norm_var": 0.0036725083584184842,
"learning_rate": 5e-05,
"loss": 0.1794,
"loss/crossentropy": 2.6474004983901978,
"loss/hidden": 0.0,
"loss/logits": 0.17943605780601501,
"loss/reg": 2.1421751976013184,
"step": 708
},
{
"epoch": 0.00709,
"grad_norm": 0.3865105211734772,
"grad_norm_var": 0.003566375202589933,
"learning_rate": 5e-05,
"loss": 0.1817,
"loss/crossentropy": 2.9082140922546387,
"loss/hidden": 0.0,
"loss/logits": 0.18172414600849152,
"loss/reg": 2.1403071880340576,
"step": 709
},
{
"epoch": 0.0071,
"grad_norm": 0.368362694978714,
"grad_norm_var": 0.003590971719305887,
"learning_rate": 5e-05,
"loss": 0.1711,
"loss/crossentropy": 2.6766469478607178,
"loss/hidden": 0.0,
"loss/logits": 0.17112310975790024,
"loss/reg": 2.138167142868042,
"step": 710
},
{
"epoch": 0.00711,
"grad_norm": 0.34797176718711853,
"grad_norm_var": 0.003506589552138199,
"learning_rate": 5e-05,
"loss": 0.163,
"loss/crossentropy": 2.8545928597450256,
"loss/hidden": 0.0,
"loss/logits": 0.16299721226096153,
"loss/reg": 2.1356112957000732,
"step": 711
},
{
"epoch": 0.00712,
"grad_norm": 0.3511999547481537,
"grad_norm_var": 0.0034451354888741254,
"learning_rate": 5e-05,
"loss": 0.1706,
"loss/crossentropy": 2.7516467571258545,
"loss/hidden": 0.0,
"loss/logits": 0.1705768182873726,
"loss/reg": 2.13396954536438,
"step": 712
},
{
"epoch": 0.00713,
"grad_norm": 0.4692562520503998,
"grad_norm_var": 0.0038021677070381584,
"learning_rate": 5e-05,
"loss": 0.1772,
"loss/crossentropy": 2.7622230648994446,
"loss/hidden": 0.0,
"loss/logits": 0.1771526113152504,
"loss/reg": 2.1316707134246826,
"step": 713
},
{
"epoch": 0.00714,
"grad_norm": 0.3500974774360657,
"grad_norm_var": 0.003583350276630167,
"learning_rate": 5e-05,
"loss": 0.1651,
"loss/crossentropy": 2.7385149598121643,
"loss/hidden": 0.0,
"loss/logits": 0.16512250155210495,
"loss/reg": 2.1301488876342773,
"step": 714
},
{
"epoch": 0.00715,
"grad_norm": 0.33279696106910706,
"grad_norm_var": 0.0037632620297312364,
"learning_rate": 5e-05,
"loss": 0.1639,
"loss/crossentropy": 2.6760587096214294,
"loss/hidden": 0.0,
"loss/logits": 0.16387901455163956,
"loss/reg": 2.128563165664673,
"step": 715
},
{
"epoch": 0.00716,
"grad_norm": 0.36436334252357483,
"grad_norm_var": 0.003418879723208453,
"learning_rate": 5e-05,
"loss": 0.1675,
"loss/crossentropy": 2.7055559158325195,
"loss/hidden": 0.0,
"loss/logits": 0.16751762479543686,
"loss/reg": 2.1272294521331787,
"step": 716
},
{
"epoch": 0.00717,
"grad_norm": 0.35308849811553955,
"grad_norm_var": 0.0009122804473129371,
"learning_rate": 5e-05,
"loss": 0.1685,
"loss/crossentropy": 2.827264368534088,
"loss/hidden": 0.0,
"loss/logits": 0.16845671087503433,
"loss/reg": 2.125559091567993,
"step": 717
},
{
"epoch": 0.00718,
"grad_norm": 0.36609259247779846,
"grad_norm_var": 0.0009080073745675876,
"learning_rate": 5e-05,
"loss": 0.1878,
"loss/crossentropy": 2.7995529770851135,
"loss/hidden": 0.0,
"loss/logits": 0.18778567016124725,
"loss/reg": 2.12422513961792,
"step": 718
},
{
"epoch": 0.00719,
"grad_norm": 0.3564467430114746,
"grad_norm_var": 0.0009149400496893722,
"learning_rate": 5e-05,
"loss": 0.1665,
"loss/crossentropy": 2.848701000213623,
"loss/hidden": 0.0,
"loss/logits": 0.16652807220816612,
"loss/reg": 2.1227705478668213,
"step": 719
},
{
"epoch": 0.0072,
"grad_norm": 0.3523035943508148,
"grad_norm_var": 0.0009263477116920882,
"learning_rate": 5e-05,
"loss": 0.169,
"loss/crossentropy": 2.7714666724205017,
"loss/hidden": 0.0,
"loss/logits": 0.16896242648363113,
"loss/reg": 2.1214191913604736,
"step": 720
},
{
"epoch": 0.00721,
"grad_norm": 0.39885270595550537,
"grad_norm_var": 0.0009792887842439849,
"learning_rate": 5e-05,
"loss": 0.1701,
"loss/crossentropy": 2.835131287574768,
"loss/hidden": 0.0,
"loss/logits": 0.1700747236609459,
"loss/reg": 2.1200404167175293,
"step": 721
},
{
"epoch": 0.00722,
"grad_norm": 0.40293964743614197,
"grad_norm_var": 0.0010526817455953927,
"learning_rate": 5e-05,
"loss": 0.1819,
"loss/crossentropy": 2.744925618171692,
"loss/hidden": 0.0,
"loss/logits": 0.1818903423845768,
"loss/reg": 2.1181347370147705,
"step": 722
},
{
"epoch": 0.00723,
"grad_norm": 0.5598530769348145,
"grad_norm_var": 0.0032894654306610577,
"learning_rate": 5e-05,
"loss": 0.1662,
"loss/crossentropy": 2.983691990375519,
"loss/hidden": 0.0,
"loss/logits": 0.16615596786141396,
"loss/reg": 2.1172993183135986,
"step": 723
},
{
"epoch": 0.00724,
"grad_norm": 0.39669546484947205,
"grad_norm_var": 0.003249640426166478,
"learning_rate": 5e-05,
"loss": 0.1735,
"loss/crossentropy": 2.709549069404602,
"loss/hidden": 0.0,
"loss/logits": 0.17345992848277092,
"loss/reg": 2.1152801513671875,
"step": 724
},
{
"epoch": 0.00725,
"grad_norm": 0.35726040601730347,
"grad_norm_var": 0.0032964500726321067,
"learning_rate": 5e-05,
"loss": 0.1624,
"loss/crossentropy": 2.7148231267929077,
"loss/hidden": 0.0,
"loss/logits": 0.16241873800754547,
"loss/reg": 2.1130924224853516,
"step": 725
},
{
"epoch": 0.00726,
"grad_norm": 0.3927571177482605,
"grad_norm_var": 0.0032861190572127997,
"learning_rate": 5e-05,
"loss": 0.1795,
"loss/crossentropy": 2.7939482927322388,
"loss/hidden": 0.0,
"loss/logits": 0.1794990859925747,
"loss/reg": 2.111480712890625,
"step": 726
},
{
"epoch": 0.00727,
"grad_norm": 0.3941044807434082,
"grad_norm_var": 0.0031944564404073005,
"learning_rate": 5e-05,
"loss": 0.1712,
"loss/crossentropy": 2.834249794483185,
"loss/hidden": 0.0,
"loss/logits": 0.1711888276040554,
"loss/reg": 2.109204053878784,
"step": 727
},
{
"epoch": 0.00728,
"grad_norm": 0.4828793704509735,
"grad_norm_var": 0.0036429198556795937,
"learning_rate": 5e-05,
"loss": 0.2082,
"loss/crossentropy": 2.899094045162201,
"loss/hidden": 0.0,
"loss/logits": 0.20823358744382858,
"loss/reg": 2.1061551570892334,
"step": 728
},
{
"epoch": 0.00729,
"grad_norm": 0.3574215769767761,
"grad_norm_var": 0.003326472236741973,
"learning_rate": 5e-05,
"loss": 0.1596,
"loss/crossentropy": 2.7636680603027344,
"loss/hidden": 0.0,
"loss/logits": 0.15961402654647827,
"loss/reg": 2.104001760482788,
"step": 729
},
{
"epoch": 0.0073,
"grad_norm": 0.40163764357566833,
"grad_norm_var": 0.003227754706050412,
"learning_rate": 5e-05,
"loss": 0.1797,
"loss/crossentropy": 2.8588566184043884,
"loss/hidden": 0.0,
"loss/logits": 0.17974677309393883,
"loss/reg": 2.102442741394043,
"step": 730
},
{
"epoch": 0.00731,
"grad_norm": 0.37189754843711853,
"grad_norm_var": 0.003015475193035148,
"learning_rate": 5e-05,
"loss": 0.1673,
"loss/crossentropy": 2.826458215713501,
"loss/hidden": 0.0,
"loss/logits": 0.16728204488754272,
"loss/reg": 2.1002049446105957,
"step": 731
},
{
"epoch": 0.00732,
"grad_norm": 0.3587784171104431,
"grad_norm_var": 0.003039707591927121,
"learning_rate": 5e-05,
"loss": 0.1645,
"loss/crossentropy": 2.731923222541809,
"loss/hidden": 0.0,
"loss/logits": 0.16448039561510086,
"loss/reg": 2.098315954208374,
"step": 732
},
{
"epoch": 0.00733,
"grad_norm": 0.37631648778915405,
"grad_norm_var": 0.002946915065401934,
"learning_rate": 5e-05,
"loss": 0.1714,
"loss/crossentropy": 2.880859136581421,
"loss/hidden": 0.0,
"loss/logits": 0.17137856781482697,
"loss/reg": 2.0962650775909424,
"step": 733
},
{
"epoch": 0.00734,
"grad_norm": 0.3563605844974518,
"grad_norm_var": 0.002990850657754888,
"learning_rate": 5e-05,
"loss": 0.1628,
"loss/crossentropy": 2.6355279088020325,
"loss/hidden": 0.0,
"loss/logits": 0.16278789564967155,
"loss/reg": 2.094142436981201,
"step": 734
},
{
"epoch": 0.00735,
"grad_norm": 0.37199047207832336,
"grad_norm_var": 0.0029265023383142925,
"learning_rate": 5e-05,
"loss": 0.1802,
"loss/crossentropy": 2.769617021083832,
"loss/hidden": 0.0,
"loss/logits": 0.18019907549023628,
"loss/reg": 2.0921125411987305,
"step": 735
},
{
"epoch": 0.00736,
"grad_norm": 0.489103764295578,
"grad_norm_var": 0.0033036264225515164,
"learning_rate": 5e-05,
"loss": 0.1832,
"loss/crossentropy": 2.7491883039474487,
"loss/hidden": 0.0,
"loss/logits": 0.1831774264574051,
"loss/reg": 2.0904338359832764,
"step": 736
},
{
"epoch": 0.00737,
"grad_norm": 0.5059826970100403,
"grad_norm_var": 0.003943075932518525,
"learning_rate": 5e-05,
"loss": 0.1817,
"loss/crossentropy": 2.8231146931648254,
"loss/hidden": 0.0,
"loss/logits": 0.18174266442656517,
"loss/reg": 2.0879714488983154,
"step": 737
},
{
"epoch": 0.00738,
"grad_norm": 0.6662333011627197,
"grad_norm_var": 0.007992879009924207,
"learning_rate": 5e-05,
"loss": 0.1861,
"loss/crossentropy": 2.7952335476875305,
"loss/hidden": 0.0,
"loss/logits": 0.18612126260995865,
"loss/reg": 2.0855941772460938,
"step": 738
},
{
"epoch": 0.00739,
"grad_norm": 0.43555790185928345,
"grad_norm_var": 0.006764259520141217,
"learning_rate": 5e-05,
"loss": 0.1823,
"loss/crossentropy": 2.7390406727790833,
"loss/hidden": 0.0,
"loss/logits": 0.1822943352162838,
"loss/reg": 2.083002805709839,
"step": 739
},
{
"epoch": 0.0074,
"grad_norm": 0.36206063628196716,
"grad_norm_var": 0.0069454028516603905,
"learning_rate": 5e-05,
"loss": 0.1747,
"loss/crossentropy": 2.6245489716529846,
"loss/hidden": 0.0,
"loss/logits": 0.17474905773997307,
"loss/reg": 2.0807597637176514,
"step": 740
},
{
"epoch": 0.00741,
"grad_norm": 0.4077146649360657,
"grad_norm_var": 0.006699115025232619,
"learning_rate": 5e-05,
"loss": 0.21,
"loss/crossentropy": 2.734727144241333,
"loss/hidden": 0.0,
"loss/logits": 0.20996900647878647,
"loss/reg": 2.0777878761291504,
"step": 741
},
{
"epoch": 0.00742,
"grad_norm": 0.4748740792274475,
"grad_norm_var": 0.0068148961293998615,
"learning_rate": 5e-05,
"loss": 0.1926,
"loss/crossentropy": 2.757317006587982,
"loss/hidden": 0.0,
"loss/logits": 0.19259492680430412,
"loss/reg": 2.0748398303985596,
"step": 742
},
{
"epoch": 0.00743,
"grad_norm": 0.3738694190979004,
"grad_norm_var": 0.006926021168675212,
"learning_rate": 5e-05,
"loss": 0.1671,
"loss/crossentropy": 2.7804144620895386,
"loss/hidden": 0.0,
"loss/logits": 0.1671152375638485,
"loss/reg": 2.072981119155884,
"step": 743
},
{
"epoch": 0.00744,
"grad_norm": 0.4373812675476074,
"grad_norm_var": 0.006701504868688938,
"learning_rate": 5e-05,
"loss": 0.1836,
"loss/crossentropy": 2.8251866698265076,
"loss/hidden": 0.0,
"loss/logits": 0.183602724224329,
"loss/reg": 2.069178581237793,
"step": 744
},
{
"epoch": 0.00745,
"grad_norm": 0.41339626908302307,
"grad_norm_var": 0.006417608208757027,
"learning_rate": 5e-05,
"loss": 0.1665,
"loss/crossentropy": 2.7784698605537415,
"loss/hidden": 0.0,
"loss/logits": 0.16648468375205994,
"loss/reg": 2.066897392272949,
"step": 745
},
{
"epoch": 0.00746,
"grad_norm": 0.36906108260154724,
"grad_norm_var": 0.0065862671267569286,
"learning_rate": 5e-05,
"loss": 0.1752,
"loss/crossentropy": 2.7134994864463806,
"loss/hidden": 0.0,
"loss/logits": 0.17522963881492615,
"loss/reg": 2.063711404800415,
"step": 746
},
{
"epoch": 0.00747,
"grad_norm": 0.3699776232242584,
"grad_norm_var": 0.006599620482715507,
"learning_rate": 5e-05,
"loss": 0.1816,
"loss/crossentropy": 2.6448380947113037,
"loss/hidden": 0.0,
"loss/logits": 0.1815556287765503,
"loss/reg": 2.0618152618408203,
"step": 747
},
{
"epoch": 0.00748,
"grad_norm": 0.35848432779312134,
"grad_norm_var": 0.006602145753337363,
"learning_rate": 5e-05,
"loss": 0.1759,
"loss/crossentropy": 2.577029287815094,
"loss/hidden": 0.0,
"loss/logits": 0.1758727729320526,
"loss/reg": 2.0592784881591797,
"step": 748
},
{
"epoch": 0.00749,
"grad_norm": 0.40015411376953125,
"grad_norm_var": 0.006489211309593653,
"learning_rate": 5e-05,
"loss": 0.2007,
"loss/crossentropy": 2.7719894647598267,
"loss/hidden": 0.0,
"loss/logits": 0.20066174119710922,
"loss/reg": 2.056396484375,
"step": 749
},
{
"epoch": 0.0075,
"grad_norm": 0.34235846996307373,
"grad_norm_var": 0.006628701391081989,
"learning_rate": 5e-05,
"loss": 0.1661,
"loss/crossentropy": 2.8526532649993896,
"loss/hidden": 0.0,
"loss/logits": 0.1660567931830883,
"loss/reg": 2.053225040435791,
"step": 750
},
{
"epoch": 0.00751,
"grad_norm": 0.37578198313713074,
"grad_norm_var": 0.006603490490161393,
"learning_rate": 5e-05,
"loss": 0.192,
"loss/crossentropy": 2.803673267364502,
"loss/hidden": 0.0,
"loss/logits": 0.19199685007333755,
"loss/reg": 2.0500073432922363,
"step": 751
},
{
"epoch": 0.00752,
"grad_norm": 0.3724234700202942,
"grad_norm_var": 0.006439587327084632,
"learning_rate": 5e-05,
"loss": 0.1611,
"loss/crossentropy": 2.8497246503829956,
"loss/hidden": 0.0,
"loss/logits": 0.16111965849995613,
"loss/reg": 2.0481040477752686,
"step": 752
},
{
"epoch": 0.00753,
"grad_norm": 0.37283533811569214,
"grad_norm_var": 0.005960471364599432,
"learning_rate": 5e-05,
"loss": 0.1816,
"loss/crossentropy": 2.580562174320221,
"loss/hidden": 0.0,
"loss/logits": 0.18162427470088005,
"loss/reg": 2.045363187789917,
"step": 753
},
{
"epoch": 0.00754,
"grad_norm": 0.42849722504615784,
"grad_norm_var": 0.0013156070888824681,
"learning_rate": 5e-05,
"loss": 0.1885,
"loss/crossentropy": 2.7384997606277466,
"loss/hidden": 0.0,
"loss/logits": 0.1885378062725067,
"loss/reg": 2.0420894622802734,
"step": 754
},
{
"epoch": 0.00755,
"grad_norm": 0.3246319890022278,
"grad_norm_var": 0.0014611472372319412,
"learning_rate": 5e-05,
"loss": 0.152,
"loss/crossentropy": 2.827781558036804,
"loss/hidden": 0.0,
"loss/logits": 0.15203238278627396,
"loss/reg": 2.0399134159088135,
"step": 755
},
{
"epoch": 0.00756,
"grad_norm": 0.3523566722869873,
"grad_norm_var": 0.0014986135555234647,
"learning_rate": 5e-05,
"loss": 0.1799,
"loss/crossentropy": 2.7049853801727295,
"loss/hidden": 0.0,
"loss/logits": 0.17991740256547928,
"loss/reg": 2.036095142364502,
"step": 756
},
{
"epoch": 0.00757,
"grad_norm": 0.3352646827697754,
"grad_norm_var": 0.0016155829783374783,
"learning_rate": 5e-05,
"loss": 0.1612,
"loss/crossentropy": 2.715296685695648,
"loss/hidden": 0.0,
"loss/logits": 0.16122159361839294,
"loss/reg": 2.0335283279418945,
"step": 757
},
{
"epoch": 0.00758,
"grad_norm": 0.36173179745674133,
"grad_norm_var": 0.001004548523534495,
"learning_rate": 5e-05,
"loss": 0.1768,
"loss/crossentropy": 2.8797001242637634,
"loss/hidden": 0.0,
"loss/logits": 0.17676853761076927,
"loss/reg": 2.0315983295440674,
"step": 758
},
{
"epoch": 0.00759,
"grad_norm": 0.43379032611846924,
"grad_norm_var": 0.0012258123535982288,
"learning_rate": 5e-05,
"loss": 0.2008,
"loss/crossentropy": 2.7367305159568787,
"loss/hidden": 0.0,
"loss/logits": 0.20076703280210495,
"loss/reg": 2.028825521469116,
"step": 759
},
{
"epoch": 0.0076,
"grad_norm": 0.7135851979255676,
"grad_norm_var": 0.008180404333396396,
"learning_rate": 5e-05,
"loss": 0.1988,
"loss/crossentropy": 2.7777557373046875,
"loss/hidden": 0.0,
"loss/logits": 0.19884883239865303,
"loss/reg": 2.0258262157440186,
"step": 760
},
{
"epoch": 0.00761,
"grad_norm": 0.36141064763069153,
"grad_norm_var": 0.008223674749041798,
"learning_rate": 5e-05,
"loss": 0.1723,
"loss/crossentropy": 2.755949318408966,
"loss/hidden": 0.0,
"loss/logits": 0.17225057631731033,
"loss/reg": 2.022468090057373,
"step": 761
},
{
"epoch": 0.00762,
"grad_norm": 0.47610145807266235,
"grad_norm_var": 0.008612084301677063,
"learning_rate": 5e-05,
"loss": 0.1786,
"loss/crossentropy": 2.701655924320221,
"loss/hidden": 0.0,
"loss/logits": 0.17863870784640312,
"loss/reg": 2.0194478034973145,
"step": 762
},
{
"epoch": 0.00763,
"grad_norm": 0.35960420966148376,
"grad_norm_var": 0.0086585523494028,
"learning_rate": 5e-05,
"loss": 0.1719,
"loss/crossentropy": 2.6496411561965942,
"loss/hidden": 0.0,
"loss/logits": 0.17191722244024277,
"loss/reg": 2.0173516273498535,
"step": 763
},
{
"epoch": 0.00764,
"grad_norm": 0.3759422302246094,
"grad_norm_var": 0.008585472348376118,
"learning_rate": 5e-05,
"loss": 0.1665,
"loss/crossentropy": 2.7261382937431335,
"loss/hidden": 0.0,
"loss/logits": 0.1664697714149952,
"loss/reg": 2.014254331588745,
"step": 764
},
{
"epoch": 0.00765,
"grad_norm": 0.3791477680206299,
"grad_norm_var": 0.008610251361000461,
"learning_rate": 5e-05,
"loss": 0.1716,
"loss/crossentropy": 2.779210090637207,
"loss/hidden": 0.0,
"loss/logits": 0.1715676300227642,
"loss/reg": 2.0109994411468506,
"step": 765
},
{
"epoch": 0.00766,
"grad_norm": 0.37698858976364136,
"grad_norm_var": 0.00842901980982322,
"learning_rate": 5e-05,
"loss": 0.178,
"loss/crossentropy": 2.5693264603614807,
"loss/hidden": 0.0,
"loss/logits": 0.17796850576996803,
"loss/reg": 2.007894277572632,
"step": 766
},
{
"epoch": 0.00767,
"grad_norm": 0.324692040681839,
"grad_norm_var": 0.008757168987509056,
"learning_rate": 5e-05,
"loss": 0.163,
"loss/crossentropy": 2.790699005126953,
"loss/hidden": 0.0,
"loss/logits": 0.16297711431980133,
"loss/reg": 2.0054080486297607,
"step": 767
},
{
"epoch": 0.00768,
"grad_norm": 0.37725430727005005,
"grad_norm_var": 0.008742918144709544,
"learning_rate": 5e-05,
"loss": 0.1701,
"loss/crossentropy": 2.90560781955719,
"loss/hidden": 0.0,
"loss/logits": 0.17007537558674812,
"loss/reg": 2.0026705265045166,
"step": 768
},
{
"epoch": 0.00769,
"grad_norm": 0.3565872013568878,
"grad_norm_var": 0.008812017421293783,
"learning_rate": 5e-05,
"loss": 0.1748,
"loss/crossentropy": 2.8573551774024963,
"loss/hidden": 0.0,
"loss/logits": 0.1747995764017105,
"loss/reg": 1.9999767541885376,
"step": 769
},
{
"epoch": 0.0077,
"grad_norm": 0.32768821716308594,
"grad_norm_var": 0.009011701837686615,
"learning_rate": 5e-05,
"loss": 0.1647,
"loss/crossentropy": 2.7850446105003357,
"loss/hidden": 0.0,
"loss/logits": 0.16465429961681366,
"loss/reg": 1.9974277019500732,
"step": 770
},
{
"epoch": 0.00771,
"grad_norm": 0.34194430708885193,
"grad_norm_var": 0.008880009468442519,
"learning_rate": 5e-05,
"loss": 0.1681,
"loss/crossentropy": 2.89225697517395,
"loss/hidden": 0.0,
"loss/logits": 0.16806093603372574,
"loss/reg": 1.9941447973251343,
"step": 771
},
{
"epoch": 0.00772,
"grad_norm": 0.36788639426231384,
"grad_norm_var": 0.008815313943155234,
"learning_rate": 5e-05,
"loss": 0.1824,
"loss/crossentropy": 2.8710330724716187,
"loss/hidden": 0.0,
"loss/logits": 0.1823921650648117,
"loss/reg": 1.991845965385437,
"step": 772
},
{
"epoch": 0.00773,
"grad_norm": 0.33500465750694275,
"grad_norm_var": 0.008817280025891942,
"learning_rate": 5e-05,
"loss": 0.1669,
"loss/crossentropy": 2.7821491956710815,
"loss/hidden": 0.0,
"loss/logits": 0.16685106977820396,
"loss/reg": 1.9901355504989624,
"step": 773
},
{
"epoch": 0.00774,
"grad_norm": 0.33815550804138184,
"grad_norm_var": 0.008946649562538052,
"learning_rate": 5e-05,
"loss": 0.162,
"loss/crossentropy": 2.7051143050193787,
"loss/hidden": 0.0,
"loss/logits": 0.16199326515197754,
"loss/reg": 1.9880555868148804,
"step": 774
},
{
"epoch": 0.00775,
"grad_norm": 0.32524728775024414,
"grad_norm_var": 0.009054478596347363,
"learning_rate": 5e-05,
"loss": 0.169,
"loss/crossentropy": 2.72264701128006,
"loss/hidden": 0.0,
"loss/logits": 0.16902651265263557,
"loss/reg": 1.986093521118164,
"step": 775
},
{
"epoch": 0.00776,
"grad_norm": 0.34697458148002625,
"grad_norm_var": 0.0013234442614042051,
"learning_rate": 5e-05,
"loss": 0.1691,
"loss/crossentropy": 2.780848979949951,
"loss/hidden": 0.0,
"loss/logits": 0.1690516211092472,
"loss/reg": 1.9844779968261719,
"step": 776
},
{
"epoch": 0.00777,
"grad_norm": 0.33995282649993896,
"grad_norm_var": 0.0013500864177136548,
"learning_rate": 5e-05,
"loss": 0.1612,
"loss/crossentropy": 2.772739827632904,
"loss/hidden": 0.0,
"loss/logits": 0.16119593381881714,
"loss/reg": 1.9825077056884766,
"step": 777
},
{
"epoch": 0.00778,
"grad_norm": 0.35139432549476624,
"grad_norm_var": 0.0003803343966673219,
"learning_rate": 5e-05,
"loss": 0.1668,
"loss/crossentropy": 2.7008825540542603,
"loss/hidden": 0.0,
"loss/logits": 0.16681700944900513,
"loss/reg": 1.9806305170059204,
"step": 778
},
{
"epoch": 0.00779,
"grad_norm": 0.4588527977466583,
"grad_norm_var": 0.00110283708340256,
"learning_rate": 5e-05,
"loss": 0.1907,
"loss/crossentropy": 2.632855713367462,
"loss/hidden": 0.0,
"loss/logits": 0.19068260118365288,
"loss/reg": 1.9793086051940918,
"step": 779
},
{
"epoch": 0.0078,
"grad_norm": 0.3829444646835327,
"grad_norm_var": 0.0011229031183707624,
"learning_rate": 5e-05,
"loss": 0.1875,
"loss/crossentropy": 2.9350045323371887,
"loss/hidden": 0.0,
"loss/logits": 0.18754199519753456,
"loss/reg": 1.9774302244186401,
"step": 780
},
{
"epoch": 0.00781,
"grad_norm": 0.46253493428230286,
"grad_norm_var": 0.0017907320044085833,
"learning_rate": 5e-05,
"loss": 0.1957,
"loss/crossentropy": 2.7478776574134827,
"loss/hidden": 0.0,
"loss/logits": 0.19565920531749725,
"loss/reg": 1.9757329225540161,
"step": 781
},
{
"epoch": 0.00782,
"grad_norm": 0.35229969024658203,
"grad_norm_var": 0.0017840355007145352,
"learning_rate": 5e-05,
"loss": 0.1611,
"loss/crossentropy": 2.805756628513336,
"loss/hidden": 0.0,
"loss/logits": 0.1610955037176609,
"loss/reg": 1.973933458328247,
"step": 782
},
{
"epoch": 0.00783,
"grad_norm": 0.3324076533317566,
"grad_norm_var": 0.0017495419673394963,
"learning_rate": 5e-05,
"loss": 0.1706,
"loss/crossentropy": 2.8343148827552795,
"loss/hidden": 0.0,
"loss/logits": 0.1706329919397831,
"loss/reg": 1.9718844890594482,
"step": 783
},
{
"epoch": 0.00784,
"grad_norm": 0.3563413619995117,
"grad_norm_var": 0.0017352353042652258,
"learning_rate": 5e-05,
"loss": 0.1768,
"loss/crossentropy": 2.825278103351593,
"loss/hidden": 0.0,
"loss/logits": 0.17681827396154404,
"loss/reg": 1.969612956047058,
"step": 784
},
{
"epoch": 0.00785,
"grad_norm": 0.33560603857040405,
"grad_norm_var": 0.0017751309342711038,
"learning_rate": 5e-05,
"loss": 0.1558,
"loss/crossentropy": 2.8132280111312866,
"loss/hidden": 0.0,
"loss/logits": 0.1558120921254158,
"loss/reg": 1.9675753116607666,
"step": 785
},
{
"epoch": 0.00786,
"grad_norm": 0.39733415842056274,
"grad_norm_var": 0.0017810049820061401,
"learning_rate": 5e-05,
"loss": 0.1849,
"loss/crossentropy": 2.8960456252098083,
"loss/hidden": 0.0,
"loss/logits": 0.18493180349469185,
"loss/reg": 1.965217113494873,
"step": 786
},
{
"epoch": 0.00787,
"grad_norm": 0.561698317527771,
"grad_norm_var": 0.004151387117344507,
"learning_rate": 5e-05,
"loss": 0.1987,
"loss/crossentropy": 2.66669100522995,
"loss/hidden": 0.0,
"loss/logits": 0.19869648292660713,
"loss/reg": 1.9629205465316772,
"step": 787
},
{
"epoch": 0.00788,
"grad_norm": 0.35911333560943604,
"grad_norm_var": 0.004167781816727311,
"learning_rate": 5e-05,
"loss": 0.1742,
"loss/crossentropy": 2.723667323589325,
"loss/hidden": 0.0,
"loss/logits": 0.17424843832850456,
"loss/reg": 1.9608547687530518,
"step": 788
},
{
"epoch": 0.00789,
"grad_norm": 0.3422897160053253,
"grad_norm_var": 0.004130072564222831,
"learning_rate": 5e-05,
"loss": 0.1683,
"loss/crossentropy": 2.853653848171234,
"loss/hidden": 0.0,
"loss/logits": 0.16829833760857582,
"loss/reg": 1.959040880203247,
"step": 789
},
{
"epoch": 0.0079,
"grad_norm": 0.373519629240036,
"grad_norm_var": 0.0040217911733014585,
"learning_rate": 5e-05,
"loss": 0.1697,
"loss/crossentropy": 2.7469093799591064,
"loss/hidden": 0.0,
"loss/logits": 0.1697460599243641,
"loss/reg": 1.9578640460968018,
"step": 790
},
{
"epoch": 0.00791,
"grad_norm": 0.42586550116539,
"grad_norm_var": 0.003921241787547673,
"learning_rate": 5e-05,
"loss": 0.1896,
"loss/crossentropy": 2.9876235127449036,
"loss/hidden": 0.0,
"loss/logits": 0.18961479887366295,
"loss/reg": 1.9558684825897217,
"step": 791
},
{
"epoch": 0.00792,
"grad_norm": 0.34371063113212585,
"grad_norm_var": 0.00393897634361432,
"learning_rate": 5e-05,
"loss": 0.1624,
"loss/crossentropy": 2.922863006591797,
"loss/hidden": 0.0,
"loss/logits": 0.16239817067980766,
"loss/reg": 1.9541829824447632,
"step": 792
},
{
"epoch": 0.00793,
"grad_norm": 0.3611912727355957,
"grad_norm_var": 0.0038367960883469387,
"learning_rate": 5e-05,
"loss": 0.1767,
"loss/crossentropy": 2.751186192035675,
"loss/hidden": 0.0,
"loss/logits": 0.17671825364232063,
"loss/reg": 1.9524297714233398,
"step": 793
},
{
"epoch": 0.00794,
"grad_norm": 0.3787733018398285,
"grad_norm_var": 0.0037525025406884736,
"learning_rate": 5e-05,
"loss": 0.1695,
"loss/crossentropy": 2.6539193391799927,
"loss/hidden": 0.0,
"loss/logits": 0.16952653229236603,
"loss/reg": 1.951439619064331,
"step": 794
},
{
"epoch": 0.00795,
"grad_norm": 0.37621310353279114,
"grad_norm_var": 0.003409985625982037,
"learning_rate": 5e-05,
"loss": 0.1827,
"loss/crossentropy": 2.672878086566925,
"loss/hidden": 0.0,
"loss/logits": 0.1826501600444317,
"loss/reg": 1.9504698514938354,
"step": 795
},
{
"epoch": 0.00796,
"grad_norm": 0.3580264747142792,
"grad_norm_var": 0.0034518512961513536,
"learning_rate": 5e-05,
"loss": 0.1741,
"loss/crossentropy": 2.8564891815185547,
"loss/hidden": 0.0,
"loss/logits": 0.17409207299351692,
"loss/reg": 1.9492477178573608,
"step": 796
},
{
"epoch": 0.00797,
"grad_norm": 0.3552623689174652,
"grad_norm_var": 0.0030235748866805395,
"learning_rate": 5e-05,
"loss": 0.1679,
"loss/crossentropy": 2.9642611145973206,
"loss/hidden": 0.0,
"loss/logits": 0.16786304488778114,
"loss/reg": 1.9484763145446777,
"step": 797
},
{
"epoch": 0.00798,
"grad_norm": 0.37029561400413513,
"grad_norm_var": 0.0029878997549970957,
"learning_rate": 5e-05,
"loss": 0.1837,
"loss/crossentropy": 2.7581509947776794,
"loss/hidden": 0.0,
"loss/logits": 0.18365685641765594,
"loss/reg": 1.946696400642395,
"step": 798
},
{
"epoch": 0.00799,
"grad_norm": 0.37257152795791626,
"grad_norm_var": 0.00285137706672662,
"learning_rate": 5e-05,
"loss": 0.1768,
"loss/crossentropy": 2.766197443008423,
"loss/hidden": 0.0,
"loss/logits": 0.17676663026213646,
"loss/reg": 1.9451103210449219,
"step": 799
},
{
"epoch": 0.008,
"grad_norm": 0.3937225043773651,
"grad_norm_var": 0.0028245897421089812,
"learning_rate": 5e-05,
"loss": 0.1593,
"loss/crossentropy": 2.968823492527008,
"loss/hidden": 0.0,
"loss/logits": 0.15929469466209412,
"loss/reg": 1.943403959274292,
"step": 800
},
{
"epoch": 0.00801,
"grad_norm": 0.5229995846748352,
"grad_norm_var": 0.003870799660257873,
"learning_rate": 5e-05,
"loss": 0.1904,
"loss/crossentropy": 2.5733524560928345,
"loss/hidden": 0.0,
"loss/logits": 0.19040565192699432,
"loss/reg": 1.9423651695251465,
"step": 801
},
{
"epoch": 0.00802,
"grad_norm": 0.4087795913219452,
"grad_norm_var": 0.003885163701405114,
"learning_rate": 5e-05,
"loss": 0.2011,
"loss/crossentropy": 2.643693685531616,
"loss/hidden": 0.0,
"loss/logits": 0.20110392943024635,
"loss/reg": 1.941137433052063,
"step": 802
},
{
"epoch": 0.00803,
"grad_norm": 0.369555801153183,
"grad_norm_var": 0.0018963737991296507,
"learning_rate": 5e-05,
"loss": 0.1768,
"loss/crossentropy": 2.702915072441101,
"loss/hidden": 0.0,
"loss/logits": 0.17682579904794693,
"loss/reg": 1.9396127462387085,
"step": 803
},
{
"epoch": 0.00804,
"grad_norm": 0.3822772204875946,
"grad_norm_var": 0.001859244513831604,
"learning_rate": 5e-05,
"loss": 0.1674,
"loss/crossentropy": 2.8051819801330566,
"loss/hidden": 0.0,
"loss/logits": 0.1674252152442932,
"loss/reg": 1.9382424354553223,
"step": 804
},
{
"epoch": 0.00805,
"grad_norm": 0.42195388674736023,
"grad_norm_var": 0.0018187903132861672,
"learning_rate": 5e-05,
"loss": 0.1982,
"loss/crossentropy": 2.7962412238121033,
"loss/hidden": 0.0,
"loss/logits": 0.19819854572415352,
"loss/reg": 1.9367636442184448,
"step": 805
},
{
"epoch": 0.00806,
"grad_norm": 0.39215224981307983,
"grad_norm_var": 0.001803471303692028,
"learning_rate": 5e-05,
"loss": 0.1862,
"loss/crossentropy": 2.838093400001526,
"loss/hidden": 0.0,
"loss/logits": 0.1862441450357437,
"loss/reg": 1.9356895685195923,
"step": 806
},
{
"epoch": 0.00807,
"grad_norm": 0.36561474204063416,
"grad_norm_var": 0.0017388941933010808,
"learning_rate": 5e-05,
"loss": 0.1659,
"loss/crossentropy": 2.8486820459365845,
"loss/hidden": 0.0,
"loss/logits": 0.16586757823824883,
"loss/reg": 1.9336965084075928,
"step": 807
},
{
"epoch": 0.00808,
"grad_norm": 0.3940856456756592,
"grad_norm_var": 0.0016146705961780842,
"learning_rate": 5e-05,
"loss": 0.1695,
"loss/crossentropy": 2.7590547800064087,
"loss/hidden": 0.0,
"loss/logits": 0.16953302919864655,
"loss/reg": 1.9318287372589111,
"step": 808
},
{
"epoch": 0.00809,
"grad_norm": 0.37031009793281555,
"grad_norm_var": 0.0015860965038246484,
"learning_rate": 5e-05,
"loss": 0.1663,
"loss/crossentropy": 2.803991198539734,
"loss/hidden": 0.0,
"loss/logits": 0.16627426072955132,
"loss/reg": 1.9298937320709229,
"step": 809
},
{
"epoch": 0.0081,
"grad_norm": 0.36467787623405457,
"grad_norm_var": 0.001618743456786816,
"learning_rate": 5e-05,
"loss": 0.1798,
"loss/crossentropy": 2.7522680163383484,
"loss/hidden": 0.0,
"loss/logits": 0.17983945459127426,
"loss/reg": 1.928220272064209,
"step": 810
},
{
"epoch": 0.00811,
"grad_norm": 0.644191324710846,
"grad_norm_var": 0.005662418748027209,
"learning_rate": 5e-05,
"loss": 0.1823,
"loss/crossentropy": 2.9207261204719543,
"loss/hidden": 0.0,
"loss/logits": 0.18232716247439384,
"loss/reg": 1.9264674186706543,
"step": 811
},
{
"epoch": 0.00812,
"grad_norm": 0.4135313034057617,
"grad_norm_var": 0.00550433789682554,
"learning_rate": 5e-05,
"loss": 0.1754,
"loss/crossentropy": 2.78128319978714,
"loss/hidden": 0.0,
"loss/logits": 0.17538663744926453,
"loss/reg": 1.9246680736541748,
"step": 812
},
{
"epoch": 0.00813,
"grad_norm": 0.44594907760620117,
"grad_norm_var": 0.005370096537218135,
"learning_rate": 5e-05,
"loss": 0.1898,
"loss/crossentropy": 2.7069836258888245,
"loss/hidden": 0.0,
"loss/logits": 0.18982965499162674,
"loss/reg": 1.9229440689086914,
"step": 813
},
{
"epoch": 0.00814,
"grad_norm": 0.41460415720939636,
"grad_norm_var": 0.005231401879877403,
"learning_rate": 5e-05,
"loss": 0.1965,
"loss/crossentropy": 2.658607244491577,
"loss/hidden": 0.0,
"loss/logits": 0.1964995227754116,
"loss/reg": 1.9211541414260864,
"step": 814
},
{
"epoch": 0.00815,
"grad_norm": 0.40847840905189514,
"grad_norm_var": 0.0050977892227572616,
"learning_rate": 5e-05,
"loss": 0.1865,
"loss/crossentropy": 2.789508819580078,
"loss/hidden": 0.0,
"loss/logits": 0.18652214854955673,
"loss/reg": 1.91935396194458,
"step": 815
},
{
"epoch": 0.00816,
"grad_norm": 0.39475539326667786,
"grad_norm_var": 0.005094298258556392,
"learning_rate": 5e-05,
"loss": 0.1876,
"loss/crossentropy": 2.771743655204773,
"loss/hidden": 0.0,
"loss/logits": 0.18758049979805946,
"loss/reg": 1.9176177978515625,
"step": 816
},
{
"epoch": 0.00817,
"grad_norm": 0.3788897395133972,
"grad_norm_var": 0.004405869730473085,
"learning_rate": 5e-05,
"loss": 0.1765,
"loss/crossentropy": 2.807315766811371,
"loss/hidden": 0.0,
"loss/logits": 0.17645375058054924,
"loss/reg": 1.9157112836837769,
"step": 817
},
{
"epoch": 0.00818,
"grad_norm": 0.3527612090110779,
"grad_norm_var": 0.004615691680188668,
"learning_rate": 5e-05,
"loss": 0.1644,
"loss/crossentropy": 2.8354954719543457,
"loss/hidden": 0.0,
"loss/logits": 0.16440149024128914,
"loss/reg": 1.914186716079712,
"step": 818
},
{
"epoch": 0.00819,
"grad_norm": 0.45434367656707764,
"grad_norm_var": 0.004640431192599914,
"learning_rate": 5e-05,
"loss": 0.2004,
"loss/crossentropy": 2.7823927998542786,
"loss/hidden": 0.0,
"loss/logits": 0.20038331300020218,
"loss/reg": 1.9128350019454956,
"step": 819
},
{
"epoch": 0.0082,
"grad_norm": 0.4440554082393646,
"grad_norm_var": 0.004630750512825665,
"learning_rate": 5e-05,
"loss": 0.198,
"loss/crossentropy": 2.7826399207115173,
"loss/hidden": 0.0,
"loss/logits": 0.19803617522120476,
"loss/reg": 1.9112329483032227,
"step": 820
},
{
"epoch": 0.00821,
"grad_norm": 0.8357638120651245,
"grad_norm_var": 0.015646654980531947,
"learning_rate": 5e-05,
"loss": 0.2236,
"loss/crossentropy": 2.855618476867676,
"loss/hidden": 0.0,
"loss/logits": 0.22355607897043228,
"loss/reg": 1.9094618558883667,
"step": 821
},
{
"epoch": 0.00822,
"grad_norm": 0.369484543800354,
"grad_norm_var": 0.01582983572015086,
"learning_rate": 5e-05,
"loss": 0.1688,
"loss/crossentropy": 2.7591440081596375,
"loss/hidden": 0.0,
"loss/logits": 0.16877064853906631,
"loss/reg": 1.9078381061553955,
"step": 822
},
{
"epoch": 0.00823,
"grad_norm": 0.3682078421115875,
"grad_norm_var": 0.015804289096973827,
"learning_rate": 5e-05,
"loss": 0.1661,
"loss/crossentropy": 2.6714991331100464,
"loss/hidden": 0.0,
"loss/logits": 0.16609660163521767,
"loss/reg": 1.9066871404647827,
"step": 823
},
{
"epoch": 0.00824,
"grad_norm": 0.4925972521305084,
"grad_norm_var": 0.015796176553567597,
"learning_rate": 5e-05,
"loss": 0.1983,
"loss/crossentropy": 2.690047025680542,
"loss/hidden": 0.0,
"loss/logits": 0.19832593947649002,
"loss/reg": 1.9051308631896973,
"step": 824
},
{
"epoch": 0.00825,
"grad_norm": 0.4004671275615692,
"grad_norm_var": 0.01554450060197241,
"learning_rate": 5e-05,
"loss": 0.1867,
"loss/crossentropy": 3.0430662631988525,
"loss/hidden": 0.0,
"loss/logits": 0.18668686598539352,
"loss/reg": 1.9037506580352783,
"step": 825
},
{
"epoch": 0.00826,
"grad_norm": 0.40644222497940063,
"grad_norm_var": 0.01518439463368978,
"learning_rate": 5e-05,
"loss": 0.1992,
"loss/crossentropy": 2.856709599494934,
"loss/hidden": 0.0,
"loss/logits": 0.19924000278115273,
"loss/reg": 1.9035279750823975,
"step": 826
},
{
"epoch": 0.00827,
"grad_norm": 0.387662410736084,
"grad_norm_var": 0.012707668169475368,
"learning_rate": 5e-05,
"loss": 0.171,
"loss/crossentropy": 2.8205041885375977,
"loss/hidden": 0.0,
"loss/logits": 0.17101648449897766,
"loss/reg": 1.9019997119903564,
"step": 827
},
{
"epoch": 0.00828,
"grad_norm": 0.4229760766029358,
"grad_norm_var": 0.012685578660426963,
"learning_rate": 5e-05,
"loss": 0.1705,
"loss/crossentropy": 2.8298428058624268,
"loss/hidden": 0.0,
"loss/logits": 0.1704978421330452,
"loss/reg": 1.9013198614120483,
"step": 828
},
{
"epoch": 0.00829,
"grad_norm": 0.4192207455635071,
"grad_norm_var": 0.012695092968616347,
"learning_rate": 5e-05,
"loss": 0.1699,
"loss/crossentropy": 2.8572763800621033,
"loss/hidden": 0.0,
"loss/logits": 0.16986168175935745,
"loss/reg": 1.9007604122161865,
"step": 829
},
{
"epoch": 0.0083,
"grad_norm": 0.3887600898742676,
"grad_norm_var": 0.012805118489640084,
"learning_rate": 5e-05,
"loss": 0.202,
"loss/crossentropy": 2.786255419254303,
"loss/hidden": 0.0,
"loss/logits": 0.20195355266332626,
"loss/reg": 1.9000667333602905,
"step": 830
},
{
"epoch": 0.00831,
"grad_norm": 0.48604434728622437,
"grad_norm_var": 0.012929568590754843,
"learning_rate": 5e-05,
"loss": 0.1961,
"loss/crossentropy": 2.845152735710144,
"loss/hidden": 0.0,
"loss/logits": 0.19609695672988892,
"loss/reg": 1.899495005607605,
"step": 831
},
{
"epoch": 0.00832,
"grad_norm": 0.38712021708488464,
"grad_norm_var": 0.012976881832390848,
"learning_rate": 5e-05,
"loss": 0.1889,
"loss/crossentropy": 2.7007412910461426,
"loss/hidden": 0.0,
"loss/logits": 0.1888689175248146,
"loss/reg": 1.8983198404312134,
"step": 832
},
{
"epoch": 0.00833,
"grad_norm": 0.3749590814113617,
"grad_norm_var": 0.013008393945473115,
"learning_rate": 5e-05,
"loss": 0.1785,
"loss/crossentropy": 2.607687532901764,
"loss/hidden": 0.0,
"loss/logits": 0.1784559190273285,
"loss/reg": 1.897759199142456,
"step": 833
},
{
"epoch": 0.00834,
"grad_norm": 0.35202544927597046,
"grad_norm_var": 0.013016684761580717,
"learning_rate": 5e-05,
"loss": 0.1707,
"loss/crossentropy": 2.7777926325798035,
"loss/hidden": 0.0,
"loss/logits": 0.17071311548352242,
"loss/reg": 1.8961632251739502,
"step": 834
},
{
"epoch": 0.00835,
"grad_norm": 0.3441760540008545,
"grad_norm_var": 0.013518763280912912,
"learning_rate": 5e-05,
"loss": 0.1615,
"loss/crossentropy": 2.835566759109497,
"loss/hidden": 0.0,
"loss/logits": 0.1615000143647194,
"loss/reg": 1.8950566053390503,
"step": 835
},
{
"epoch": 0.00836,
"grad_norm": 0.362005352973938,
"grad_norm_var": 0.013785734718565416,
"learning_rate": 5e-05,
"loss": 0.1631,
"loss/crossentropy": 2.818268299102783,
"loss/hidden": 0.0,
"loss/logits": 0.16310900822281837,
"loss/reg": 1.8935774564743042,
"step": 836
},
{
"epoch": 0.00837,
"grad_norm": 0.3725143074989319,
"grad_norm_var": 0.0018186987426010584,
"learning_rate": 5e-05,
"loss": 0.1929,
"loss/crossentropy": 2.763745427131653,
"loss/hidden": 0.0,
"loss/logits": 0.192863829433918,
"loss/reg": 1.8919721841812134,
"step": 837
},
{
"epoch": 0.00838,
"grad_norm": 0.39604651927948,
"grad_norm_var": 0.0017691837659245461,
"learning_rate": 5e-05,
"loss": 0.1847,
"loss/crossentropy": 2.729005455970764,
"loss/hidden": 0.0,
"loss/logits": 0.18469301983714104,
"loss/reg": 1.8908017873764038,
"step": 838
},
{
"epoch": 0.00839,
"grad_norm": 0.39175912737846375,
"grad_norm_var": 0.001711627371570279,
"learning_rate": 5e-05,
"loss": 0.1841,
"loss/crossentropy": 2.8563897609710693,
"loss/hidden": 0.0,
"loss/logits": 0.18407713249325752,
"loss/reg": 1.8892946243286133,
"step": 839
},
{
"epoch": 0.0084,
"grad_norm": 0.3497207462787628,
"grad_norm_var": 0.0012053613127933737,
"learning_rate": 5e-05,
"loss": 0.1657,
"loss/crossentropy": 2.9635773301124573,
"loss/hidden": 0.0,
"loss/logits": 0.1656595915555954,
"loss/reg": 1.8886394500732422,
"step": 840
},
{
"epoch": 0.00841,
"grad_norm": 0.36070436239242554,
"grad_norm_var": 0.0012493146014174172,
"learning_rate": 5e-05,
"loss": 0.1659,
"loss/crossentropy": 2.905772030353546,
"loss/hidden": 0.0,
"loss/logits": 0.16585366800427437,
"loss/reg": 1.8883072137832642,
"step": 841
},
{
"epoch": 0.00842,
"grad_norm": 0.5194427371025085,
"grad_norm_var": 0.002330769361460483,
"learning_rate": 5e-05,
"loss": 0.1903,
"loss/crossentropy": 2.914414703845978,
"loss/hidden": 0.0,
"loss/logits": 0.1903173327445984,
"loss/reg": 1.8872804641723633,
"step": 842
},
{
"epoch": 0.00843,
"grad_norm": 0.3658028841018677,
"grad_norm_var": 0.0023811347132161485,
"learning_rate": 5e-05,
"loss": 0.1722,
"loss/crossentropy": 2.8683270812034607,
"loss/hidden": 0.0,
"loss/logits": 0.17216329649090767,
"loss/reg": 1.8860183954238892,
"step": 843
},
{
"epoch": 0.00844,
"grad_norm": 0.3355120122432709,
"grad_norm_var": 0.0025135271396979795,
"learning_rate": 5e-05,
"loss": 0.166,
"loss/crossentropy": 2.7672330141067505,
"loss/hidden": 0.0,
"loss/logits": 0.16596197709441185,
"loss/reg": 1.8848904371261597,
"step": 844
},
{
"epoch": 0.00845,
"grad_norm": 0.45907968282699585,
"grad_norm_var": 0.002779472587279837,
"learning_rate": 5e-05,
"loss": 0.1881,
"loss/crossentropy": 2.79125440120697,
"loss/hidden": 0.0,
"loss/logits": 0.1881270818412304,
"loss/reg": 1.8831804990768433,
"step": 845
},
{
"epoch": 0.00846,
"grad_norm": 0.3753393888473511,
"grad_norm_var": 0.0027935829770950843,
"learning_rate": 5e-05,
"loss": 0.1801,
"loss/crossentropy": 2.753562033176422,
"loss/hidden": 0.0,
"loss/logits": 0.18013736233115196,
"loss/reg": 1.8817567825317383,
"step": 846
},
{
"epoch": 0.00847,
"grad_norm": 0.41996800899505615,
"grad_norm_var": 0.002216029114339835,
"learning_rate": 5e-05,
"loss": 0.1871,
"loss/crossentropy": 2.8630106449127197,
"loss/hidden": 0.0,
"loss/logits": 0.18711163103580475,
"loss/reg": 1.8805315494537354,
"step": 847
},
{
"epoch": 0.00848,
"grad_norm": 0.40139615535736084,
"grad_norm_var": 0.0022320677834542836,
"learning_rate": 5e-05,
"loss": 0.182,
"loss/crossentropy": 2.8663435578346252,
"loss/hidden": 0.0,
"loss/logits": 0.18201814219355583,
"loss/reg": 1.8795160055160522,
"step": 848
},
{
"epoch": 0.00849,
"grad_norm": 0.44251078367233276,
"grad_norm_var": 0.0024153193390402117,
"learning_rate": 5e-05,
"loss": 0.1721,
"loss/crossentropy": 2.8493316173553467,
"loss/hidden": 0.0,
"loss/logits": 0.1721041165292263,
"loss/reg": 1.8777539730072021,
"step": 849
},
{
"epoch": 0.0085,
"grad_norm": 0.39363330602645874,
"grad_norm_var": 0.00231007314672006,
"learning_rate": 5e-05,
"loss": 0.1933,
"loss/crossentropy": 2.8225064277648926,
"loss/hidden": 0.0,
"loss/logits": 0.19327203929424286,
"loss/reg": 1.8762940168380737,
"step": 850
},
{
"epoch": 0.00851,
"grad_norm": 0.3834942579269409,
"grad_norm_var": 0.0021502092497398652,
"learning_rate": 5e-05,
"loss": 0.1919,
"loss/crossentropy": 2.7274433970451355,
"loss/hidden": 0.0,
"loss/logits": 0.19194044917821884,
"loss/reg": 1.8752055168151855,
"step": 851
},
{
"epoch": 0.00852,
"grad_norm": 0.36249086260795593,
"grad_norm_var": 0.0021480519578248518,
"learning_rate": 5e-05,
"loss": 0.1759,
"loss/crossentropy": 2.725217640399933,
"loss/hidden": 0.0,
"loss/logits": 0.17594841867685318,
"loss/reg": 1.8736830949783325,
"step": 852
},
{
"epoch": 0.00853,
"grad_norm": 0.3869001269340515,
"grad_norm_var": 0.002116727725912885,
"learning_rate": 5e-05,
"loss": 0.1823,
"loss/crossentropy": 2.738001227378845,
"loss/hidden": 0.0,
"loss/logits": 0.1822870336472988,
"loss/reg": 1.8728076219558716,
"step": 853
},
{
"epoch": 0.00854,
"grad_norm": 0.3708244860172272,
"grad_norm_var": 0.002157970353941248,
"learning_rate": 5e-05,
"loss": 0.1668,
"loss/crossentropy": 2.788993239402771,
"loss/hidden": 0.0,
"loss/logits": 0.16682880371809006,
"loss/reg": 1.8720353841781616,
"step": 854
},
{
"epoch": 0.00855,
"grad_norm": 0.372335821390152,
"grad_norm_var": 0.002189712517136307,
"learning_rate": 5e-05,
"loss": 0.1718,
"loss/crossentropy": 2.813368082046509,
"loss/hidden": 0.0,
"loss/logits": 0.17184938862919807,
"loss/reg": 1.8711519241333008,
"step": 855
},
{
"epoch": 0.00856,
"grad_norm": 0.35767483711242676,
"grad_norm_var": 0.0021470276840197164,
"learning_rate": 5e-05,
"loss": 0.175,
"loss/crossentropy": 2.779674708843231,
"loss/hidden": 0.0,
"loss/logits": 0.17495984584093094,
"loss/reg": 1.869391679763794,
"step": 856
},
{
"epoch": 0.00857,
"grad_norm": 0.3517981767654419,
"grad_norm_var": 0.002191754274186038,
"learning_rate": 5e-05,
"loss": 0.1688,
"loss/crossentropy": 2.7776423692703247,
"loss/hidden": 0.0,
"loss/logits": 0.16876182705163956,
"loss/reg": 1.868484616279602,
"step": 857
},
{
"epoch": 0.00858,
"grad_norm": 0.8127824664115906,
"grad_norm_var": 0.012490247842596114,
"learning_rate": 5e-05,
"loss": 0.2224,
"loss/crossentropy": 2.9876713156700134,
"loss/hidden": 0.0,
"loss/logits": 0.2223958522081375,
"loss/reg": 1.8681334257125854,
"step": 858
},
{
"epoch": 0.00859,
"grad_norm": 0.4339921474456787,
"grad_norm_var": 0.012361098720851383,
"learning_rate": 5e-05,
"loss": 0.1752,
"loss/crossentropy": 2.8306267857551575,
"loss/hidden": 0.0,
"loss/logits": 0.17517483979463577,
"loss/reg": 1.8669747114181519,
"step": 859
},
{
"epoch": 0.0086,
"grad_norm": 0.5807726383209229,
"grad_norm_var": 0.013480947234538975,
"learning_rate": 5e-05,
"loss": 0.1787,
"loss/crossentropy": 2.731403112411499,
"loss/hidden": 0.0,
"loss/logits": 0.17869474738836288,
"loss/reg": 1.865167498588562,
"step": 860
},
{
"epoch": 0.00861,
"grad_norm": 0.377247154712677,
"grad_norm_var": 0.013599237642109623,
"learning_rate": 5e-05,
"loss": 0.1714,
"loss/crossentropy": 2.8269473910331726,
"loss/hidden": 0.0,
"loss/logits": 0.1713988333940506,
"loss/reg": 1.8640793561935425,
"step": 861
},
{
"epoch": 0.00862,
"grad_norm": 0.37849879264831543,
"grad_norm_var": 0.013578332002263205,
"learning_rate": 5e-05,
"loss": 0.169,
"loss/crossentropy": 2.8433790802955627,
"loss/hidden": 0.0,
"loss/logits": 0.16895778477191925,
"loss/reg": 1.8629158735275269,
"step": 862
},
{
"epoch": 0.00863,
"grad_norm": 0.4124751091003418,
"grad_norm_var": 0.013588511645486826,
"learning_rate": 5e-05,
"loss": 0.1803,
"loss/crossentropy": 2.7584823966026306,
"loss/hidden": 0.0,
"loss/logits": 0.18026690557599068,
"loss/reg": 1.8609492778778076,
"step": 863
},
{
"epoch": 0.00864,
"grad_norm": 0.37336814403533936,
"grad_norm_var": 0.013730216300815038,
"learning_rate": 5e-05,
"loss": 0.1707,
"loss/crossentropy": 2.7589592933654785,
"loss/hidden": 0.0,
"loss/logits": 0.17070752009749413,
"loss/reg": 1.8597759008407593,
"step": 864
},
{
"epoch": 0.00865,
"grad_norm": 0.8337989449501038,
"grad_norm_var": 0.02424293784323857,
"learning_rate": 5e-05,
"loss": 0.1781,
"loss/crossentropy": 2.8036633133888245,
"loss/hidden": 0.0,
"loss/logits": 0.17813356593251228,
"loss/reg": 1.8581993579864502,
"step": 865
},
{
"epoch": 0.00866,
"grad_norm": 0.35601744055747986,
"grad_norm_var": 0.02460846166740538,
"learning_rate": 5e-05,
"loss": 0.1759,
"loss/crossentropy": 2.715932846069336,
"loss/hidden": 0.0,
"loss/logits": 0.17593519389629364,
"loss/reg": 1.8565593957901,
"step": 866
},
{
"epoch": 0.00867,
"grad_norm": 0.425502747297287,
"grad_norm_var": 0.02436568774250706,
"learning_rate": 5e-05,
"loss": 0.1821,
"loss/crossentropy": 2.7275202870368958,
"loss/hidden": 0.0,
"loss/logits": 0.18210354447364807,
"loss/reg": 1.8554532527923584,
"step": 867
},
{
"epoch": 0.00868,
"grad_norm": 0.3844553232192993,
"grad_norm_var": 0.024142035969486622,
"learning_rate": 5e-05,
"loss": 0.1796,
"loss/crossentropy": 2.8464134335517883,
"loss/hidden": 0.0,
"loss/logits": 0.1795903705060482,
"loss/reg": 1.855208396911621,
"step": 868
},
{
"epoch": 0.00869,
"grad_norm": 0.35618311166763306,
"grad_norm_var": 0.02446160042257303,
"learning_rate": 5e-05,
"loss": 0.1758,
"loss/crossentropy": 2.87707781791687,
"loss/hidden": 0.0,
"loss/logits": 0.17583919316530228,
"loss/reg": 1.8549201488494873,
"step": 869
},
{
"epoch": 0.0087,
"grad_norm": 0.41672077775001526,
"grad_norm_var": 0.024117258377413187,
"learning_rate": 5e-05,
"loss": 0.1759,
"loss/crossentropy": 2.71548855304718,
"loss/hidden": 0.0,
"loss/logits": 0.1758672632277012,
"loss/reg": 1.8541263341903687,
"step": 870
},
{
"epoch": 0.00871,
"grad_norm": 0.3646162450313568,
"grad_norm_var": 0.024202440513241764,
"learning_rate": 5e-05,
"loss": 0.186,
"loss/crossentropy": 2.786548674106598,
"loss/hidden": 0.0,
"loss/logits": 0.18596061319112778,
"loss/reg": 1.8524823188781738,
"step": 871
},
{
"epoch": 0.00872,
"grad_norm": 0.37939974665641785,
"grad_norm_var": 0.023961625350842352,
"learning_rate": 5e-05,
"loss": 0.1748,
"loss/crossentropy": 2.835165321826935,
"loss/hidden": 0.0,
"loss/logits": 0.17476912215352058,
"loss/reg": 1.8506335020065308,
"step": 872
},
{
"epoch": 0.00873,
"grad_norm": 0.3887036442756653,
"grad_norm_var": 0.023551954015331347,
"learning_rate": 5e-05,
"loss": 0.1789,
"loss/crossentropy": 2.732525408267975,
"loss/hidden": 0.0,
"loss/logits": 0.17892278358340263,
"loss/reg": 1.849176287651062,
"step": 873
},
{
"epoch": 0.00874,
"grad_norm": 0.387320876121521,
"grad_norm_var": 0.014549813961372993,
"learning_rate": 5e-05,
"loss": 0.1941,
"loss/crossentropy": 2.792769968509674,
"loss/hidden": 0.0,
"loss/logits": 0.19409611076116562,
"loss/reg": 1.848083734512329,
"step": 874
},
{
"epoch": 0.00875,
"grad_norm": 0.3818178176879883,
"grad_norm_var": 0.014678730624869341,
"learning_rate": 5e-05,
"loss": 0.1865,
"loss/crossentropy": 2.762765884399414,
"loss/hidden": 0.0,
"loss/logits": 0.18646146729588509,
"loss/reg": 1.8465189933776855,
"step": 875
},
{
"epoch": 0.00876,
"grad_norm": 0.36014363169670105,
"grad_norm_var": 0.013132955726787365,
"learning_rate": 5e-05,
"loss": 0.1847,
"loss/crossentropy": 2.8113619089126587,
"loss/hidden": 0.0,
"loss/logits": 0.1847200095653534,
"loss/reg": 1.8454153537750244,
"step": 876
},
{
"epoch": 0.00877,
"grad_norm": 0.3916279077529907,
"grad_norm_var": 0.013081129963419124,
"learning_rate": 5e-05,
"loss": 0.181,
"loss/crossentropy": 2.767706513404846,
"loss/hidden": 0.0,
"loss/logits": 0.18097594752907753,
"loss/reg": 1.8445773124694824,
"step": 877
},
{
"epoch": 0.00878,
"grad_norm": 0.38396528363227844,
"grad_norm_var": 0.013058641234249413,
"learning_rate": 5e-05,
"loss": 0.1799,
"loss/crossentropy": 2.750400483608246,
"loss/hidden": 0.0,
"loss/logits": 0.179927259683609,
"loss/reg": 1.8441030979156494,
"step": 878
},
{
"epoch": 0.00879,
"grad_norm": 0.3784838020801544,
"grad_norm_var": 0.013129867131250705,
"learning_rate": 5e-05,
"loss": 0.1783,
"loss/crossentropy": 2.6576608419418335,
"loss/hidden": 0.0,
"loss/logits": 0.1783200539648533,
"loss/reg": 1.842759370803833,
"step": 879
},
{
"epoch": 0.0088,
"grad_norm": 0.3373940885066986,
"grad_norm_var": 0.013387093786405535,
"learning_rate": 5e-05,
"loss": 0.1665,
"loss/crossentropy": 2.7494055032730103,
"loss/hidden": 0.0,
"loss/logits": 0.16646360978484154,
"loss/reg": 1.8418775796890259,
"step": 880
},
{
"epoch": 0.00881,
"grad_norm": 0.3743676543235779,
"grad_norm_var": 0.000488954453456858,
"learning_rate": 5e-05,
"loss": 0.1718,
"loss/crossentropy": 2.7373632192611694,
"loss/hidden": 0.0,
"loss/logits": 0.17177791520953178,
"loss/reg": 1.8405792713165283,
"step": 881
},
{
"epoch": 0.00882,
"grad_norm": 0.36398422718048096,
"grad_norm_var": 0.0004683277690547882,
"learning_rate": 5e-05,
"loss": 0.1713,
"loss/crossentropy": 2.689941644668579,
"loss/hidden": 0.0,
"loss/logits": 0.1712586209177971,
"loss/reg": 1.8396137952804565,
"step": 882
},
{
"epoch": 0.00883,
"grad_norm": 0.36932700872421265,
"grad_norm_var": 0.0003222525763987627,
"learning_rate": 5e-05,
"loss": 0.1738,
"loss/crossentropy": 3.0094715356826782,
"loss/hidden": 0.0,
"loss/logits": 0.17375321686267853,
"loss/reg": 1.8387012481689453,
"step": 883
},
{
"epoch": 0.00884,
"grad_norm": 0.37647050619125366,
"grad_norm_var": 0.0003174026053568609,
"learning_rate": 5e-05,
"loss": 0.1734,
"loss/crossentropy": 2.675420820713043,
"loss/hidden": 0.0,
"loss/logits": 0.17337032034993172,
"loss/reg": 1.8372994661331177,
"step": 884
},
{
"epoch": 0.00885,
"grad_norm": 0.3657122850418091,
"grad_norm_var": 0.0002983341146215642,
"learning_rate": 5e-05,
"loss": 0.175,
"loss/crossentropy": 2.785289704799652,
"loss/hidden": 0.0,
"loss/logits": 0.1750231385231018,
"loss/reg": 1.836666464805603,
"step": 885
},
{
"epoch": 0.00886,
"grad_norm": 0.3565351963043213,
"grad_norm_var": 0.00019998832643003023,
"learning_rate": 5e-05,
"loss": 0.1611,
"loss/crossentropy": 2.67407763004303,
"loss/hidden": 0.0,
"loss/logits": 0.16111686453223228,
"loss/reg": 1.8361395597457886,
"step": 886
},
{
"epoch": 0.00887,
"grad_norm": 0.38317063450813293,
"grad_norm_var": 0.00020202125672466782,
"learning_rate": 5e-05,
"loss": 0.1793,
"loss/crossentropy": 2.7103012204170227,
"loss/hidden": 0.0,
"loss/logits": 0.17928585410118103,
"loss/reg": 1.8344452381134033,
"step": 887
},
{
"epoch": 0.00888,
"grad_norm": 0.39307963848114014,
"grad_norm_var": 0.00022420215532017082,
"learning_rate": 5e-05,
"loss": 0.1899,
"loss/crossentropy": 2.7159000635147095,
"loss/hidden": 0.0,
"loss/logits": 0.18990719318389893,
"loss/reg": 1.8328790664672852,
"step": 888
},
{
"epoch": 0.00889,
"grad_norm": 0.35862287878990173,
"grad_norm_var": 0.00022381402201028245,
"learning_rate": 5e-05,
"loss": 0.1834,
"loss/crossentropy": 2.847196877002716,
"loss/hidden": 0.0,
"loss/logits": 0.18342823907732964,
"loss/reg": 1.8318103551864624,
"step": 889
},
{
"epoch": 0.0089,
"grad_norm": 0.3539126515388489,
"grad_norm_var": 0.0002281156381275314,
"learning_rate": 5e-05,
"loss": 0.1777,
"loss/crossentropy": 2.853213608264923,
"loss/hidden": 0.0,
"loss/logits": 0.1777319796383381,
"loss/reg": 1.829829454421997,
"step": 890
},
{
"epoch": 0.00891,
"grad_norm": 0.41561535000801086,
"grad_norm_var": 0.000350336348254295,
"learning_rate": 5e-05,
"loss": 0.1895,
"loss/crossentropy": 2.6837574243545532,
"loss/hidden": 0.0,
"loss/logits": 0.18954132869839668,
"loss/reg": 1.8281813859939575,
"step": 891
},
{
"epoch": 0.00892,
"grad_norm": 0.3593007028102875,
"grad_norm_var": 0.00035178644306217054,
"learning_rate": 5e-05,
"loss": 0.1639,
"loss/crossentropy": 2.7398064136505127,
"loss/hidden": 0.0,
"loss/logits": 0.163859985768795,
"loss/reg": 1.8272664546966553,
"step": 892
},
{
"epoch": 0.00893,
"grad_norm": 0.3928586542606354,
"grad_norm_var": 0.00035500389449958367,
"learning_rate": 5e-05,
"loss": 0.187,
"loss/crossentropy": 2.7214816212654114,
"loss/hidden": 0.0,
"loss/logits": 0.18696707114577293,
"loss/reg": 1.8260765075683594,
"step": 893
},
{
"epoch": 0.00894,
"grad_norm": 0.38060957193374634,
"grad_norm_var": 0.00035065611870696014,
"learning_rate": 5e-05,
"loss": 0.1816,
"loss/crossentropy": 2.8307188153266907,
"loss/hidden": 0.0,
"loss/logits": 0.18159236386418343,
"loss/reg": 1.8246755599975586,
"step": 894
},
{
"epoch": 0.00895,
"grad_norm": 0.34957438707351685,
"grad_norm_var": 0.0003796919232549693,
"learning_rate": 5e-05,
"loss": 0.1801,
"loss/crossentropy": 2.7387137413024902,
"loss/hidden": 0.0,
"loss/logits": 0.18012140691280365,
"loss/reg": 1.8235573768615723,
"step": 895
},
{
"epoch": 0.00896,
"grad_norm": 0.3666534721851349,
"grad_norm_var": 0.00030342620785123544,
"learning_rate": 5e-05,
"loss": 0.2036,
"loss/crossentropy": 2.758805215358734,
"loss/hidden": 0.0,
"loss/logits": 0.20357270538806915,
"loss/reg": 1.8219776153564453,
"step": 896
},
{
"epoch": 0.00897,
"grad_norm": 0.5404136180877686,
"grad_norm_var": 0.0020682628614343557,
"learning_rate": 5e-05,
"loss": 0.1933,
"loss/crossentropy": 2.970340132713318,
"loss/hidden": 0.0,
"loss/logits": 0.19332898035645485,
"loss/reg": 1.8206608295440674,
"step": 897
},
{
"epoch": 0.00898,
"grad_norm": 0.3982648551464081,
"grad_norm_var": 0.0020554109287465,
"learning_rate": 5e-05,
"loss": 0.1768,
"loss/crossentropy": 2.6868785619735718,
"loss/hidden": 0.0,
"loss/logits": 0.1768476814031601,
"loss/reg": 1.8192402124404907,
"step": 898
},
{
"epoch": 0.00899,
"grad_norm": 0.38741791248321533,
"grad_norm_var": 0.002038042531465663,
"learning_rate": 5e-05,
"loss": 0.1808,
"loss/crossentropy": 2.816374719142914,
"loss/hidden": 0.0,
"loss/logits": 0.18077386170625687,
"loss/reg": 1.8179749250411987,
"step": 899
},
{
"epoch": 0.009,
"grad_norm": 0.3847026526927948,
"grad_norm_var": 0.0020316665201686695,
"learning_rate": 5e-05,
"loss": 0.1982,
"loss/crossentropy": 2.8102923035621643,
"loss/hidden": 0.0,
"loss/logits": 0.19821672514081,
"loss/reg": 1.8163453340530396,
"step": 900
},
{
"epoch": 0.00901,
"grad_norm": 0.3916863203048706,
"grad_norm_var": 0.0020013109603508254,
"learning_rate": 5e-05,
"loss": 0.1778,
"loss/crossentropy": 2.883694589138031,
"loss/hidden": 0.0,
"loss/logits": 0.17775952070951462,
"loss/reg": 1.8147519826889038,
"step": 901
},
{
"epoch": 0.00902,
"grad_norm": 0.36217865347862244,
"grad_norm_var": 0.001979417665481912,
"learning_rate": 5e-05,
"loss": 0.1669,
"loss/crossentropy": 2.8100743293762207,
"loss/hidden": 0.0,
"loss/logits": 0.1668529324233532,
"loss/reg": 1.8133151531219482,
"step": 902
},
{
"epoch": 0.00903,
"grad_norm": 0.42102572321891785,
"grad_norm_var": 0.00204143104015622,
"learning_rate": 5e-05,
"loss": 0.181,
"loss/crossentropy": 2.8006924986839294,
"loss/hidden": 0.0,
"loss/logits": 0.18102795258164406,
"loss/reg": 1.8111281394958496,
"step": 903
},
{
"epoch": 0.00904,
"grad_norm": 0.388874351978302,
"grad_norm_var": 0.002041367346731493,
"learning_rate": 5e-05,
"loss": 0.1818,
"loss/crossentropy": 2.8004772067070007,
"loss/hidden": 0.0,
"loss/logits": 0.18177441507577896,
"loss/reg": 1.8097094297409058,
"step": 904
},
{
"epoch": 0.00905,
"grad_norm": 0.3800460696220398,
"grad_norm_var": 0.0019783346485648203,
"learning_rate": 5e-05,
"loss": 0.176,
"loss/crossentropy": 2.693452537059784,
"loss/hidden": 0.0,
"loss/logits": 0.1760004386305809,
"loss/reg": 1.8086622953414917,
"step": 905
},
{
"epoch": 0.00906,
"grad_norm": 0.32960033416748047,
"grad_norm_var": 0.0021389732007735863,
"learning_rate": 5e-05,
"loss": 0.1641,
"loss/crossentropy": 2.7406028509140015,
"loss/hidden": 0.0,
"loss/logits": 0.1641414910554886,
"loss/reg": 1.8070775270462036,
"step": 906
},
{
"epoch": 0.00907,
"grad_norm": 0.41262900829315186,
"grad_norm_var": 0.002129550660359725,
"learning_rate": 5e-05,
"loss": 0.1639,
"loss/crossentropy": 2.815558433532715,
"loss/hidden": 0.0,
"loss/logits": 0.1639426201581955,
"loss/reg": 1.80557119846344,
"step": 907
},
{
"epoch": 0.00908,
"grad_norm": 0.34168651700019836,
"grad_norm_var": 0.0022218976438497353,
"learning_rate": 5e-05,
"loss": 0.1689,
"loss/crossentropy": 2.7727773189544678,
"loss/hidden": 0.0,
"loss/logits": 0.16886158660054207,
"loss/reg": 1.8040578365325928,
"step": 908
},
{
"epoch": 0.00909,
"grad_norm": 0.3481311500072479,
"grad_norm_var": 0.0023254939668475396,
"learning_rate": 5e-05,
"loss": 0.1723,
"loss/crossentropy": 2.7662405967712402,
"loss/hidden": 0.0,
"loss/logits": 0.17232706770300865,
"loss/reg": 1.802317500114441,
"step": 909
},
{
"epoch": 0.0091,
"grad_norm": 0.34673434495925903,
"grad_norm_var": 0.0024236772610501315,
"learning_rate": 5e-05,
"loss": 0.1721,
"loss/crossentropy": 2.7547109723091125,
"loss/hidden": 0.0,
"loss/logits": 0.17211218550801277,
"loss/reg": 1.8002270460128784,
"step": 910
},
{
"epoch": 0.00911,
"grad_norm": 0.358116090297699,
"grad_norm_var": 0.002388630196925971,
"learning_rate": 5e-05,
"loss": 0.1778,
"loss/crossentropy": 2.8174885511398315,
"loss/hidden": 0.0,
"loss/logits": 0.17777465283870697,
"loss/reg": 1.7986141443252563,
"step": 911
},
{
"epoch": 0.00912,
"grad_norm": 0.37328216433525085,
"grad_norm_var": 0.0023752628687049343,
"learning_rate": 5e-05,
"loss": 0.1846,
"loss/crossentropy": 2.7423484921455383,
"loss/hidden": 0.0,
"loss/logits": 0.18462468683719635,
"loss/reg": 1.7969509363174438,
"step": 912
},
{
"epoch": 0.00913,
"grad_norm": 0.4073435366153717,
"grad_norm_var": 0.000729848525378903,
"learning_rate": 5e-05,
"loss": 0.1656,
"loss/crossentropy": 2.6962223649024963,
"loss/hidden": 0.0,
"loss/logits": 0.16563431546092033,
"loss/reg": 1.7949342727661133,
"step": 913
},
{
"epoch": 0.00914,
"grad_norm": 0.37720003724098206,
"grad_norm_var": 0.0006978068548327905,
"learning_rate": 5e-05,
"loss": 0.1853,
"loss/crossentropy": 2.710608184337616,
"loss/hidden": 0.0,
"loss/logits": 0.18529681861400604,
"loss/reg": 1.7920844554901123,
"step": 914
},
{
"epoch": 0.00915,
"grad_norm": 0.36010050773620605,
"grad_norm_var": 0.0007016424011595597,
"learning_rate": 5e-05,
"loss": 0.1833,
"loss/crossentropy": 2.8862733840942383,
"loss/hidden": 0.0,
"loss/logits": 0.1833462007343769,
"loss/reg": 1.7895822525024414,
"step": 915
},
{
"epoch": 0.00916,
"grad_norm": 0.35757148265838623,
"grad_norm_var": 0.0007087821480995478,
"learning_rate": 5e-05,
"loss": 0.1741,
"loss/crossentropy": 2.736583173274994,
"loss/hidden": 0.0,
"loss/logits": 0.1740923710167408,
"loss/reg": 1.7870407104492188,
"step": 916
},
{
"epoch": 0.00917,
"grad_norm": 0.38085147738456726,
"grad_norm_var": 0.0006880592910958772,
"learning_rate": 5e-05,
"loss": 0.1758,
"loss/crossentropy": 2.722678780555725,
"loss/hidden": 0.0,
"loss/logits": 0.17577889189124107,
"loss/reg": 1.78484308719635,
"step": 917
},
{
"epoch": 0.00918,
"grad_norm": 0.3618144989013672,
"grad_norm_var": 0.000688524329092799,
"learning_rate": 5e-05,
"loss": 0.1729,
"loss/crossentropy": 2.7107303738594055,
"loss/hidden": 0.0,
"loss/logits": 0.17288268730044365,
"loss/reg": 1.781936764717102,
"step": 918
},
{
"epoch": 0.00919,
"grad_norm": 0.4562152326107025,
"grad_norm_var": 0.000997994245971834,
"learning_rate": 5e-05,
"loss": 0.209,
"loss/crossentropy": 2.6911511421203613,
"loss/hidden": 0.0,
"loss/logits": 0.20904593169689178,
"loss/reg": 1.7793291807174683,
"step": 919
},
{
"epoch": 0.0092,
"grad_norm": 0.7892553806304932,
"grad_norm_var": 0.011823798595284762,
"learning_rate": 5e-05,
"loss": 0.2098,
"loss/crossentropy": 2.8024474382400513,
"loss/hidden": 0.0,
"loss/logits": 0.20984026044607162,
"loss/reg": 1.7767542600631714,
"step": 920
},
{
"epoch": 0.00921,
"grad_norm": 0.4952187240123749,
"grad_norm_var": 0.012365066103201613,
"learning_rate": 5e-05,
"loss": 0.2215,
"loss/crossentropy": 2.836692988872528,
"loss/hidden": 0.0,
"loss/logits": 0.22146976366639137,
"loss/reg": 1.77475905418396,
"step": 921
},
{
"epoch": 0.00922,
"grad_norm": 0.4500190317630768,
"grad_norm_var": 0.01204494814009713,
"learning_rate": 5e-05,
"loss": 0.1957,
"loss/crossentropy": 2.7740437984466553,
"loss/hidden": 0.0,
"loss/logits": 0.1957194283604622,
"loss/reg": 1.7725263833999634,
"step": 922
},
{
"epoch": 0.00923,
"grad_norm": 0.4018624424934387,
"grad_norm_var": 0.012053458598524087,
"learning_rate": 5e-05,
"loss": 0.1985,
"loss/crossentropy": 2.783832609653473,
"loss/hidden": 0.0,
"loss/logits": 0.19849245250225067,
"loss/reg": 1.7706068754196167,
"step": 923
},
{
"epoch": 0.00924,
"grad_norm": 0.4053579866886139,
"grad_norm_var": 0.01170279735803306,
"learning_rate": 5e-05,
"loss": 0.1847,
"loss/crossentropy": 2.8584959506988525,
"loss/hidden": 0.0,
"loss/logits": 0.1846720688045025,
"loss/reg": 1.7687995433807373,
"step": 924
},
{
"epoch": 0.00925,
"grad_norm": 0.4355542063713074,
"grad_norm_var": 0.011379840002585932,
"learning_rate": 5e-05,
"loss": 0.1911,
"loss/crossentropy": 2.8072018027305603,
"loss/hidden": 0.0,
"loss/logits": 0.19105902686715126,
"loss/reg": 1.766356348991394,
"step": 925
},
{
"epoch": 0.00926,
"grad_norm": 0.41985246539115906,
"grad_norm_var": 0.010977469936536095,
"learning_rate": 5e-05,
"loss": 0.1772,
"loss/crossentropy": 2.825522303581238,
"loss/hidden": 0.0,
"loss/logits": 0.17719140276312828,
"loss/reg": 1.764008641242981,
"step": 926
},
{
"epoch": 0.00927,
"grad_norm": 0.4020366370677948,
"grad_norm_var": 0.010695516965112247,
"learning_rate": 5e-05,
"loss": 0.168,
"loss/crossentropy": 2.946666181087494,
"loss/hidden": 0.0,
"loss/logits": 0.16798604279756546,
"loss/reg": 1.7619134187698364,
"step": 927
},
{
"epoch": 0.00928,
"grad_norm": 0.4333237111568451,
"grad_norm_var": 0.01047000612816995,
"learning_rate": 5e-05,
"loss": 0.1732,
"loss/crossentropy": 2.904057264328003,
"loss/hidden": 0.0,
"loss/logits": 0.1731964722275734,
"loss/reg": 1.7594822645187378,
"step": 928
},
{
"epoch": 0.00929,
"grad_norm": 0.44914835691452026,
"grad_norm_var": 0.010434282299518182,
"learning_rate": 5e-05,
"loss": 0.1822,
"loss/crossentropy": 2.8881112933158875,
"loss/hidden": 0.0,
"loss/logits": 0.1821521669626236,
"loss/reg": 1.757930874824524,
"step": 929
},
{
"epoch": 0.0093,
"grad_norm": 0.8063428401947021,
"grad_norm_var": 0.018582235883409348,
"learning_rate": 5e-05,
"loss": 0.2108,
"loss/crossentropy": 3.000428855419159,
"loss/hidden": 0.0,
"loss/logits": 0.21083774790167809,
"loss/reg": 1.756495714187622,
"step": 930
},
{
"epoch": 0.00931,
"grad_norm": 0.37262633442878723,
"grad_norm_var": 0.018420551139004416,
"learning_rate": 5e-05,
"loss": 0.1727,
"loss/crossentropy": 2.899652659893036,
"loss/hidden": 0.0,
"loss/logits": 0.1727372407913208,
"loss/reg": 1.755653977394104,
"step": 931
},
{
"epoch": 0.00932,
"grad_norm": 0.44574448466300964,
"grad_norm_var": 0.017660345874116586,
"learning_rate": 5e-05,
"loss": 0.1771,
"loss/crossentropy": 2.7930703163146973,
"loss/hidden": 0.0,
"loss/logits": 0.17705539613962173,
"loss/reg": 1.7541528940200806,
"step": 932
},
{
"epoch": 0.00933,
"grad_norm": 0.4381559193134308,
"grad_norm_var": 0.017191491981390843,
"learning_rate": 5e-05,
"loss": 0.1928,
"loss/crossentropy": 3.009516716003418,
"loss/hidden": 0.0,
"loss/logits": 0.19276633486151695,
"loss/reg": 1.7523828744888306,
"step": 933
},
{
"epoch": 0.00934,
"grad_norm": 0.37119948863983154,
"grad_norm_var": 0.01705829482260827,
"learning_rate": 5e-05,
"loss": 0.1784,
"loss/crossentropy": 2.7011741995811462,
"loss/hidden": 0.0,
"loss/logits": 0.17841476574540138,
"loss/reg": 1.750117540359497,
"step": 934
},
{
"epoch": 0.00935,
"grad_norm": 0.38776126503944397,
"grad_norm_var": 0.017506596591750172,
"learning_rate": 5e-05,
"loss": 0.1765,
"loss/crossentropy": 2.755903959274292,
"loss/hidden": 0.0,
"loss/logits": 0.17649077624082565,
"loss/reg": 1.7485483884811401,
"step": 935
},
{
"epoch": 0.00936,
"grad_norm": 0.3909890949726105,
"grad_norm_var": 0.010412048008983836,
"learning_rate": 5e-05,
"loss": 0.1718,
"loss/crossentropy": 2.7981409430503845,
"loss/hidden": 0.0,
"loss/logits": 0.1718210205435753,
"loss/reg": 1.7470717430114746,
"step": 936
},
{
"epoch": 0.00937,
"grad_norm": 0.4047463536262512,
"grad_norm_var": 0.010306674977021387,
"learning_rate": 5e-05,
"loss": 0.1944,
"loss/crossentropy": 2.653374135494232,
"loss/hidden": 0.0,
"loss/logits": 0.1944383941590786,
"loss/reg": 1.7457630634307861,
"step": 937
},
{
"epoch": 0.00938,
"grad_norm": 0.4641042947769165,
"grad_norm_var": 0.010340857957444612,
"learning_rate": 5e-05,
"loss": 0.1933,
"loss/crossentropy": 3.054188370704651,
"loss/hidden": 0.0,
"loss/logits": 0.19333792477846146,
"loss/reg": 1.7441859245300293,
"step": 938
},
{
"epoch": 0.00939,
"grad_norm": 0.463642418384552,
"grad_norm_var": 0.01027101724812616,
"learning_rate": 5e-05,
"loss": 0.1906,
"loss/crossentropy": 2.7980846166610718,
"loss/hidden": 0.0,
"loss/logits": 0.19064636901021004,
"loss/reg": 1.7424683570861816,
"step": 939
},
{
"epoch": 0.0094,
"grad_norm": 0.3763858675956726,
"grad_norm_var": 0.010469512228889604,
"learning_rate": 5e-05,
"loss": 0.1817,
"loss/crossentropy": 2.8554503321647644,
"loss/hidden": 0.0,
"loss/logits": 0.1817203275859356,
"loss/reg": 1.7405589818954468,
"step": 940
},
{
"epoch": 0.00941,
"grad_norm": 0.41792160272598267,
"grad_norm_var": 0.010502572032980106,
"learning_rate": 5e-05,
"loss": 0.2081,
"loss/crossentropy": 2.7464479207992554,
"loss/hidden": 0.0,
"loss/logits": 0.20806986466050148,
"loss/reg": 1.7390387058258057,
"step": 941
},
{
"epoch": 0.00942,
"grad_norm": 0.405609130859375,
"grad_norm_var": 0.010553986517832181,
"learning_rate": 5e-05,
"loss": 0.1923,
"loss/crossentropy": 2.9190812706947327,
"loss/hidden": 0.0,
"loss/logits": 0.1922554075717926,
"loss/reg": 1.737121343612671,
"step": 942
},
{
"epoch": 0.00943,
"grad_norm": 0.5186859369277954,
"grad_norm_var": 0.010823950074372254,
"learning_rate": 5e-05,
"loss": 0.1892,
"loss/crossentropy": 2.846211016178131,
"loss/hidden": 0.0,
"loss/logits": 0.1892014741897583,
"loss/reg": 1.7357187271118164,
"step": 943
},
{
"epoch": 0.00944,
"grad_norm": 0.3441806137561798,
"grad_norm_var": 0.011478989118617007,
"learning_rate": 5e-05,
"loss": 0.1701,
"loss/crossentropy": 2.8220054507255554,
"loss/hidden": 0.0,
"loss/logits": 0.1700589321553707,
"loss/reg": 1.733402967453003,
"step": 944
},
{
"epoch": 0.00945,
"grad_norm": 0.37400367856025696,
"grad_norm_var": 0.011751047533430632,
"learning_rate": 5e-05,
"loss": 0.179,
"loss/crossentropy": 2.7692030668258667,
"loss/hidden": 0.0,
"loss/logits": 0.17900892347097397,
"loss/reg": 1.7316250801086426,
"step": 945
},
{
"epoch": 0.00946,
"grad_norm": 0.4089336395263672,
"grad_norm_var": 0.0020184395330867097,
"learning_rate": 5e-05,
"loss": 0.1982,
"loss/crossentropy": 2.891884744167328,
"loss/hidden": 0.0,
"loss/logits": 0.19816706702113152,
"loss/reg": 1.7298880815505981,
"step": 946
},
{
"epoch": 0.00947,
"grad_norm": 0.36752966046333313,
"grad_norm_var": 0.002046509202798789,
"learning_rate": 5e-05,
"loss": 0.1843,
"loss/crossentropy": 2.858055591583252,
"loss/hidden": 0.0,
"loss/logits": 0.18428384885191917,
"loss/reg": 1.7284300327301025,
"step": 947
},
{
"epoch": 0.00948,
"grad_norm": 0.36644455790519714,
"grad_norm_var": 0.002074549092487384,
"learning_rate": 5e-05,
"loss": 0.1714,
"loss/crossentropy": 2.848255932331085,
"loss/hidden": 0.0,
"loss/logits": 0.17144014686346054,
"loss/reg": 1.7270002365112305,
"step": 948
},
{
"epoch": 0.00949,
"grad_norm": 0.5910805463790894,
"grad_norm_var": 0.004186356490923812,
"learning_rate": 5e-05,
"loss": 0.23,
"loss/crossentropy": 2.7590108513832092,
"loss/hidden": 0.0,
"loss/logits": 0.22996008396148682,
"loss/reg": 1.7256983518600464,
"step": 949
},
{
"epoch": 0.0095,
"grad_norm": 0.35803112387657166,
"grad_norm_var": 0.00427554901524122,
"learning_rate": 5e-05,
"loss": 0.1684,
"loss/crossentropy": 2.7760064005851746,
"loss/hidden": 0.0,
"loss/logits": 0.16840620338916779,
"loss/reg": 1.723679542541504,
"step": 950
},
{
"epoch": 0.00951,
"grad_norm": 0.412681519985199,
"grad_norm_var": 0.004223846207916952,
"learning_rate": 5e-05,
"loss": 0.2018,
"loss/crossentropy": 2.9045740365982056,
"loss/hidden": 0.0,
"loss/logits": 0.20182525366544724,
"loss/reg": 1.72231125831604,
"step": 951
},
{
"epoch": 0.00952,
"grad_norm": 0.4021626114845276,
"grad_norm_var": 0.004193552649382352,
"learning_rate": 5e-05,
"loss": 0.1848,
"loss/crossentropy": 2.6521793007850647,
"loss/hidden": 0.0,
"loss/logits": 0.1848319098353386,
"loss/reg": 1.720641851425171,
"step": 952
},
{
"epoch": 0.00953,
"grad_norm": 0.3750251233577728,
"grad_norm_var": 0.00429834750938114,
"learning_rate": 5e-05,
"loss": 0.1791,
"loss/crossentropy": 2.7560397386550903,
"loss/hidden": 0.0,
"loss/logits": 0.17908834293484688,
"loss/reg": 1.7182477712631226,
"step": 953
},
{
"epoch": 0.00954,
"grad_norm": 0.5893900990486145,
"grad_norm_var": 0.006092951728716223,
"learning_rate": 5e-05,
"loss": 0.2129,
"loss/crossentropy": 2.835801601409912,
"loss/hidden": 0.0,
"loss/logits": 0.21293479949235916,
"loss/reg": 1.716722011566162,
"step": 954
},
{
"epoch": 0.00955,
"grad_norm": 0.40877264738082886,
"grad_norm_var": 0.005985476808116985,
"learning_rate": 5e-05,
"loss": 0.1938,
"loss/crossentropy": 2.689119517803192,
"loss/hidden": 0.0,
"loss/logits": 0.19383220747113228,
"loss/reg": 1.714568853378296,
"step": 955
},
{
"epoch": 0.00956,
"grad_norm": 0.38810843229293823,
"grad_norm_var": 0.005926205055061354,
"learning_rate": 5e-05,
"loss": 0.1705,
"loss/crossentropy": 2.948507070541382,
"loss/hidden": 0.0,
"loss/logits": 0.1705201156437397,
"loss/reg": 1.7119203805923462,
"step": 956
},
{
"epoch": 0.00957,
"grad_norm": 0.4206679165363312,
"grad_norm_var": 0.005925719462670757,
"learning_rate": 5e-05,
"loss": 0.1807,
"loss/crossentropy": 2.78257417678833,
"loss/hidden": 0.0,
"loss/logits": 0.18074193224310875,
"loss/reg": 1.7095727920532227,
"step": 957
},
{
"epoch": 0.00958,
"grad_norm": 0.3933105766773224,
"grad_norm_var": 0.005959929971731507,
"learning_rate": 5e-05,
"loss": 0.2045,
"loss/crossentropy": 2.7225964665412903,
"loss/hidden": 0.0,
"loss/logits": 0.2044883407652378,
"loss/reg": 1.707101583480835,
"step": 958
},
{
"epoch": 0.00959,
"grad_norm": 0.3582659661769867,
"grad_norm_var": 0.005456189170996354,
"learning_rate": 5e-05,
"loss": 0.1603,
"loss/crossentropy": 2.7268422842025757,
"loss/hidden": 0.0,
"loss/logits": 0.1602596789598465,
"loss/reg": 1.7055177688598633,
"step": 959
},
{
"epoch": 0.0096,
"grad_norm": 0.397594153881073,
"grad_norm_var": 0.0051663773874797295,
"learning_rate": 5e-05,
"loss": 0.1733,
"loss/crossentropy": 2.877332389354706,
"loss/hidden": 0.0,
"loss/logits": 0.1733493208885193,
"loss/reg": 1.7030569314956665,
"step": 960
},
{
"epoch": 0.00961,
"grad_norm": 0.4981625974178314,
"grad_norm_var": 0.005480135764708397,
"learning_rate": 5e-05,
"loss": 0.1826,
"loss/crossentropy": 2.656112492084503,
"loss/hidden": 0.0,
"loss/logits": 0.18259770050644875,
"loss/reg": 1.7019816637039185,
"step": 961
},
{
"epoch": 0.00962,
"grad_norm": 0.937751054763794,
"grad_norm_var": 0.022106629800203694,
"learning_rate": 5e-05,
"loss": 0.2083,
"loss/crossentropy": 2.907736301422119,
"loss/hidden": 0.0,
"loss/logits": 0.20831404626369476,
"loss/reg": 1.700728416442871,
"step": 962
},
{
"epoch": 0.00963,
"grad_norm": 0.3895174264907837,
"grad_norm_var": 0.021883161578962466,
"learning_rate": 5e-05,
"loss": 0.1803,
"loss/crossentropy": 2.845858633518219,
"loss/hidden": 0.0,
"loss/logits": 0.18031802773475647,
"loss/reg": 1.6990742683410645,
"step": 963
},
{
"epoch": 0.00964,
"grad_norm": 0.34548690915107727,
"grad_norm_var": 0.02215928485241057,
"learning_rate": 5e-05,
"loss": 0.1645,
"loss/crossentropy": 2.8244311213493347,
"loss/hidden": 0.0,
"loss/logits": 0.1645219847559929,
"loss/reg": 1.697798252105713,
"step": 964
},
{
"epoch": 0.00965,
"grad_norm": 0.36824318766593933,
"grad_norm_var": 0.021193656582446117,
"learning_rate": 5e-05,
"loss": 0.1717,
"loss/crossentropy": 2.884181797504425,
"loss/hidden": 0.0,
"loss/logits": 0.17166699841618538,
"loss/reg": 1.6957753896713257,
"step": 965
},
{
"epoch": 0.00966,
"grad_norm": 0.37774839997291565,
"grad_norm_var": 0.021001939954339834,
"learning_rate": 5e-05,
"loss": 0.1824,
"loss/crossentropy": 3.025804340839386,
"loss/hidden": 0.0,
"loss/logits": 0.18241329863667488,
"loss/reg": 1.6944239139556885,
"step": 966
},
{
"epoch": 0.00967,
"grad_norm": 0.36408743262290955,
"grad_norm_var": 0.02133579751543382,
"learning_rate": 5e-05,
"loss": 0.1816,
"loss/crossentropy": 2.6597015261650085,
"loss/hidden": 0.0,
"loss/logits": 0.1815933845937252,
"loss/reg": 1.692252516746521,
"step": 967
},
{
"epoch": 0.00968,
"grad_norm": 0.34311729669570923,
"grad_norm_var": 0.02183892884845392,
"learning_rate": 5e-05,
"loss": 0.1662,
"loss/crossentropy": 2.845684826374054,
"loss/hidden": 0.0,
"loss/logits": 0.1661607250571251,
"loss/reg": 1.6907639503479004,
"step": 968
},
{
"epoch": 0.00969,
"grad_norm": 0.38303908705711365,
"grad_norm_var": 0.021779175231247044,
"learning_rate": 5e-05,
"loss": 0.1693,
"loss/crossentropy": 2.558404505252838,
"loss/hidden": 0.0,
"loss/logits": 0.16925161331892014,
"loss/reg": 1.6893569231033325,
"step": 969
},
{
"epoch": 0.0097,
"grad_norm": 0.3850949704647064,
"grad_norm_var": 0.02018777587365078,
"learning_rate": 5e-05,
"loss": 0.1685,
"loss/crossentropy": 2.890751600265503,
"loss/hidden": 0.0,
"loss/logits": 0.1684512346982956,
"loss/reg": 1.688266634941101,
"step": 970
},
{
"epoch": 0.00971,
"grad_norm": 0.4068422317504883,
"grad_norm_var": 0.020191525445637973,
"learning_rate": 5e-05,
"loss": 0.1857,
"loss/crossentropy": 2.707846701145172,
"loss/hidden": 0.0,
"loss/logits": 0.18569114059209824,
"loss/reg": 1.6867531538009644,
"step": 971
},
{
"epoch": 0.00972,
"grad_norm": 0.3924512565135956,
"grad_norm_var": 0.020172897207266394,
"learning_rate": 5e-05,
"loss": 0.1829,
"loss/crossentropy": 2.848098576068878,
"loss/hidden": 0.0,
"loss/logits": 0.1828712299466133,
"loss/reg": 1.6852294206619263,
"step": 972
},
{
"epoch": 0.00973,
"grad_norm": 0.3714575469493866,
"grad_norm_var": 0.020336838096992275,
"learning_rate": 5e-05,
"loss": 0.1773,
"loss/crossentropy": 2.8703532814979553,
"loss/hidden": 0.0,
"loss/logits": 0.17731666564941406,
"loss/reg": 1.6838881969451904,
"step": 973
},
{
"epoch": 0.00974,
"grad_norm": 0.35195186734199524,
"grad_norm_var": 0.020588227081264298,
"learning_rate": 5e-05,
"loss": 0.1862,
"loss/crossentropy": 2.863659620285034,
"loss/hidden": 0.0,
"loss/logits": 0.18621815741062164,
"loss/reg": 1.6821165084838867,
"step": 974
},
{
"epoch": 0.00975,
"grad_norm": 0.441755086183548,
"grad_norm_var": 0.02037088575085001,
"learning_rate": 5e-05,
"loss": 0.1933,
"loss/crossentropy": 2.810901939868927,
"loss/hidden": 0.0,
"loss/logits": 0.19334488362073898,
"loss/reg": 1.6802574396133423,
"step": 975
},
{
"epoch": 0.00976,
"grad_norm": 0.40233367681503296,
"grad_norm_var": 0.02035677589008348,
"learning_rate": 5e-05,
"loss": 0.1983,
"loss/crossentropy": 2.681654691696167,
"loss/hidden": 0.0,
"loss/logits": 0.19833911955356598,
"loss/reg": 1.6781818866729736,
"step": 976
},
{
"epoch": 0.00977,
"grad_norm": 0.6531580686569214,
"grad_norm_var": 0.023423138566671976,
"learning_rate": 5e-05,
"loss": 0.2212,
"loss/crossentropy": 2.82851505279541,
"loss/hidden": 0.0,
"loss/logits": 0.2212192267179489,
"loss/reg": 1.6763266324996948,
"step": 977
},
{
"epoch": 0.00978,
"grad_norm": 0.3646674156188965,
"grad_norm_var": 0.005314392422501734,
"learning_rate": 5e-05,
"loss": 0.1715,
"loss/crossentropy": 2.7788134813308716,
"loss/hidden": 0.0,
"loss/logits": 0.17147252708673477,
"loss/reg": 1.674770712852478,
"step": 978
},
{
"epoch": 0.00979,
"grad_norm": 0.40374529361724854,
"grad_norm_var": 0.005314159555871933,
"learning_rate": 5e-05,
"loss": 0.2092,
"loss/crossentropy": 2.7746172547340393,
"loss/hidden": 0.0,
"loss/logits": 0.2092289738357067,
"loss/reg": 1.6723932027816772,
"step": 979
},
{
"epoch": 0.0098,
"grad_norm": 0.3737621009349823,
"grad_norm_var": 0.005169172562247167,
"learning_rate": 5e-05,
"loss": 0.1847,
"loss/crossentropy": 2.871635138988495,
"loss/hidden": 0.0,
"loss/logits": 0.18473126366734505,
"loss/reg": 1.6707124710083008,
"step": 980
},
{
"epoch": 0.00981,
"grad_norm": 0.37383797764778137,
"grad_norm_var": 0.005148210609649088,
"learning_rate": 5e-05,
"loss": 0.1733,
"loss/crossentropy": 2.760922133922577,
"loss/hidden": 0.0,
"loss/logits": 0.1733292043209076,
"loss/reg": 1.6693061590194702,
"step": 981
},
{
"epoch": 0.00982,
"grad_norm": 0.38922467827796936,
"grad_norm_var": 0.00512344066750369,
"learning_rate": 5e-05,
"loss": 0.164,
"loss/crossentropy": 2.8191832304000854,
"loss/hidden": 0.0,
"loss/logits": 0.16395087912678719,
"loss/reg": 1.6671026945114136,
"step": 982
},
{
"epoch": 0.00983,
"grad_norm": 0.40670332312583923,
"grad_norm_var": 0.0050327015332547465,
"learning_rate": 5e-05,
"loss": 0.1892,
"loss/crossentropy": 2.847275197505951,
"loss/hidden": 0.0,
"loss/logits": 0.18915896490216255,
"loss/reg": 1.6655491590499878,
"step": 983
},
{
"epoch": 0.00984,
"grad_norm": 0.3739645183086395,
"grad_norm_var": 0.004847126969690196,
"learning_rate": 5e-05,
"loss": 0.194,
"loss/crossentropy": 2.817361056804657,
"loss/hidden": 0.0,
"loss/logits": 0.1940479725599289,
"loss/reg": 1.664434552192688,
"step": 984
},
{
"epoch": 0.00985,
"grad_norm": 0.36827903985977173,
"grad_norm_var": 0.00490322302848593,
"learning_rate": 5e-05,
"loss": 0.1793,
"loss/crossentropy": 2.9278652667999268,
"loss/hidden": 0.0,
"loss/logits": 0.17926159501075745,
"loss/reg": 1.6631444692611694,
"step": 985
},
{
"epoch": 0.00986,
"grad_norm": 0.36838048696517944,
"grad_norm_var": 0.0049621510753778035,
"learning_rate": 5e-05,
"loss": 0.182,
"loss/crossentropy": 2.625900149345398,
"loss/hidden": 0.0,
"loss/logits": 0.1820085123181343,
"loss/reg": 1.6615486145019531,
"step": 986
},
{
"epoch": 0.00987,
"grad_norm": 0.406107097864151,
"grad_norm_var": 0.00496177464005331,
"learning_rate": 5e-05,
"loss": 0.165,
"loss/crossentropy": 2.6364856362342834,
"loss/hidden": 0.0,
"loss/logits": 0.16496483236551285,
"loss/reg": 1.6604608297348022,
"step": 987
},
{
"epoch": 0.00988,
"grad_norm": 0.3886563777923584,
"grad_norm_var": 0.004967815483619401,
"learning_rate": 5e-05,
"loss": 0.1803,
"loss/crossentropy": 2.796413004398346,
"loss/hidden": 0.0,
"loss/logits": 0.1803182028234005,
"loss/reg": 1.6593427658081055,
"step": 988
},
{
"epoch": 0.00989,
"grad_norm": 0.35161423683166504,
"grad_norm_var": 0.005074223354079936,
"learning_rate": 5e-05,
"loss": 0.1644,
"loss/crossentropy": 2.8155667185783386,
"loss/hidden": 0.0,
"loss/logits": 0.16441339999437332,
"loss/reg": 1.6583104133605957,
"step": 989
},
{
"epoch": 0.0099,
"grad_norm": 0.39407941699028015,
"grad_norm_var": 0.0049088886087087355,
"learning_rate": 5e-05,
"loss": 0.1928,
"loss/crossentropy": 2.8455575704574585,
"loss/hidden": 0.0,
"loss/logits": 0.19284628704190254,
"loss/reg": 1.658367395401001,
"step": 990
},
{
"epoch": 0.00991,
"grad_norm": 0.3695094883441925,
"grad_norm_var": 0.004869171230278472,
"learning_rate": 5e-05,
"loss": 0.1814,
"loss/crossentropy": 2.80289888381958,
"loss/hidden": 0.0,
"loss/logits": 0.18142832443118095,
"loss/reg": 1.6575336456298828,
"step": 991
},
{
"epoch": 0.00992,
"grad_norm": 0.3505973815917969,
"grad_norm_var": 0.005015199761620141,
"learning_rate": 5e-05,
"loss": 0.1769,
"loss/crossentropy": 2.731611430644989,
"loss/hidden": 0.0,
"loss/logits": 0.17688723653554916,
"loss/reg": 1.6567095518112183,
"step": 992
},
{
"epoch": 0.00993,
"grad_norm": 0.38008397817611694,
"grad_norm_var": 0.0003133497280889556,
"learning_rate": 5e-05,
"loss": 0.1836,
"loss/crossentropy": 2.7869952917099,
"loss/hidden": 0.0,
"loss/logits": 0.18361864984035492,
"loss/reg": 1.6552574634552002,
"step": 993
},
{
"epoch": 0.00994,
"grad_norm": 0.38647469878196716,
"grad_norm_var": 0.00030154116815576856,
"learning_rate": 5e-05,
"loss": 0.1794,
"loss/crossentropy": 2.744201898574829,
"loss/hidden": 0.0,
"loss/logits": 0.17939525097608566,
"loss/reg": 1.6545374393463135,
"step": 994
},
{
"epoch": 0.00995,
"grad_norm": 0.3995093107223511,
"grad_norm_var": 0.0002894285610608412,
"learning_rate": 5e-05,
"loss": 0.1916,
"loss/crossentropy": 2.694726526737213,
"loss/hidden": 0.0,
"loss/logits": 0.1916041001677513,
"loss/reg": 1.6533586978912354,
"step": 995
},
{
"epoch": 0.00996,
"grad_norm": 0.34584900736808777,
"grad_norm_var": 0.0003615231269390488,
"learning_rate": 5e-05,
"loss": 0.1739,
"loss/crossentropy": 2.8016315698623657,
"loss/hidden": 0.0,
"loss/logits": 0.1739257462322712,
"loss/reg": 1.651719570159912,
"step": 996
},
{
"epoch": 0.00997,
"grad_norm": 0.3925288915634155,
"grad_norm_var": 0.0003722265532580001,
"learning_rate": 5e-05,
"loss": 0.177,
"loss/crossentropy": 2.897447168827057,
"loss/hidden": 0.0,
"loss/logits": 0.17699377238750458,
"loss/reg": 1.6504392623901367,
"step": 997
},
{
"epoch": 0.00998,
"grad_norm": 0.39880403876304626,
"grad_norm_var": 0.0003904176090236522,
"learning_rate": 5e-05,
"loss": 0.1919,
"loss/crossentropy": 2.8275578022003174,
"loss/hidden": 0.0,
"loss/logits": 0.19189641624689102,
"loss/reg": 1.6491367816925049,
"step": 998
},
{
"epoch": 0.00999,
"grad_norm": 0.4310808479785919,
"grad_norm_var": 0.0005141220319849537,
"learning_rate": 5e-05,
"loss": 0.2298,
"loss/crossentropy": 2.7063609957695007,
"loss/hidden": 0.0,
"loss/logits": 0.22976921498775482,
"loss/reg": 1.6471576690673828,
"step": 999
},
{
"epoch": 0.01,
"grad_norm": 0.3714313507080078,
"grad_norm_var": 0.0005171003041950173,
"learning_rate": 5e-05,
"loss": 0.1901,
"loss/crossentropy": 2.7168938517570496,
"loss/hidden": 0.0,
"loss/logits": 0.19012651592493057,
"loss/reg": 1.6457953453063965,
"step": 1000
}
],
"logging_steps": 1,
"max_steps": 100000,
"num_input_tokens_seen": 0,
"num_train_epochs": 9223372036854775807,
"save_steps": 1000,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": true,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 5.15246892515328e+17,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}