| { |
| "best_global_step": null, |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 0.33831406822667043, |
| "eval_steps": 500, |
| "global_step": 900, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.0037590452025185604, |
| "grad_norm": 2.616206169128418, |
| "learning_rate": 1.6000000000000003e-05, |
| "loss": 4.4449, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.007518090405037121, |
| "grad_norm": 1.0848251581192017, |
| "learning_rate": 3.6e-05, |
| "loss": 3.5961, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.01127713560755568, |
| "grad_norm": 1.1111085414886475, |
| "learning_rate": 5.6000000000000006e-05, |
| "loss": 2.9221, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.015036180810074242, |
| "grad_norm": 1.2459989786148071, |
| "learning_rate": 7.6e-05, |
| "loss": 1.9429, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.0187952260125928, |
| "grad_norm": 0.8213669657707214, |
| "learning_rate": 9.6e-05, |
| "loss": 1.9535, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.02255427121511136, |
| "grad_norm": 0.8069262504577637, |
| "learning_rate": 9.979695431472081e-05, |
| "loss": 1.7172, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.026313316417629923, |
| "grad_norm": 1.422139286994934, |
| "learning_rate": 9.954314720812184e-05, |
| "loss": 1.6907, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.030072361620148483, |
| "grad_norm": 1.3577821254730225, |
| "learning_rate": 9.928934010152285e-05, |
| "loss": 1.8705, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.03383140682266704, |
| "grad_norm": 1.192848801612854, |
| "learning_rate": 9.903553299492386e-05, |
| "loss": 1.5954, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.0375904520251856, |
| "grad_norm": 1.2164387702941895, |
| "learning_rate": 9.878172588832489e-05, |
| "loss": 1.5556, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.04134949722770416, |
| "grad_norm": 1.162097692489624, |
| "learning_rate": 9.852791878172589e-05, |
| "loss": 1.5447, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.04510854243022272, |
| "grad_norm": 0.9176503419876099, |
| "learning_rate": 9.827411167512691e-05, |
| "loss": 1.4747, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.04886758763274129, |
| "grad_norm": 1.1206398010253906, |
| "learning_rate": 9.802030456852792e-05, |
| "loss": 1.5876, |
| "step": 130 |
| }, |
| { |
| "epoch": 0.05262663283525985, |
| "grad_norm": 0.6738734841346741, |
| "learning_rate": 9.776649746192893e-05, |
| "loss": 1.5071, |
| "step": 140 |
| }, |
| { |
| "epoch": 0.05638567803777841, |
| "grad_norm": 1.0791672468185425, |
| "learning_rate": 9.751269035532995e-05, |
| "loss": 1.3875, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.06014472324029697, |
| "grad_norm": 0.8023204803466797, |
| "learning_rate": 9.725888324873097e-05, |
| "loss": 1.4511, |
| "step": 160 |
| }, |
| { |
| "epoch": 0.06390376844281552, |
| "grad_norm": 1.1953641176223755, |
| "learning_rate": 9.700507614213198e-05, |
| "loss": 1.551, |
| "step": 170 |
| }, |
| { |
| "epoch": 0.06766281364533408, |
| "grad_norm": 0.9535043835639954, |
| "learning_rate": 9.6751269035533e-05, |
| "loss": 1.6191, |
| "step": 180 |
| }, |
| { |
| "epoch": 0.07142185884785264, |
| "grad_norm": 1.397887945175171, |
| "learning_rate": 9.649746192893402e-05, |
| "loss": 1.5753, |
| "step": 190 |
| }, |
| { |
| "epoch": 0.0751809040503712, |
| "grad_norm": 0.7615554332733154, |
| "learning_rate": 9.624365482233503e-05, |
| "loss": 1.3002, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.07893994925288976, |
| "grad_norm": 1.2040668725967407, |
| "learning_rate": 9.598984771573605e-05, |
| "loss": 1.3894, |
| "step": 210 |
| }, |
| { |
| "epoch": 0.08269899445540832, |
| "grad_norm": 0.8942911028862, |
| "learning_rate": 9.573604060913706e-05, |
| "loss": 1.4224, |
| "step": 220 |
| }, |
| { |
| "epoch": 0.08645803965792688, |
| "grad_norm": 0.9624324440956116, |
| "learning_rate": 9.548223350253807e-05, |
| "loss": 1.4281, |
| "step": 230 |
| }, |
| { |
| "epoch": 0.09021708486044544, |
| "grad_norm": 0.9280639886856079, |
| "learning_rate": 9.522842639593908e-05, |
| "loss": 1.4149, |
| "step": 240 |
| }, |
| { |
| "epoch": 0.09397613006296401, |
| "grad_norm": 1.081101655960083, |
| "learning_rate": 9.497461928934011e-05, |
| "loss": 1.4395, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.09773517526548257, |
| "grad_norm": 0.6382322311401367, |
| "learning_rate": 9.472081218274112e-05, |
| "loss": 1.2719, |
| "step": 260 |
| }, |
| { |
| "epoch": 0.10149422046800113, |
| "grad_norm": 0.6878702044487, |
| "learning_rate": 9.446700507614213e-05, |
| "loss": 1.358, |
| "step": 270 |
| }, |
| { |
| "epoch": 0.1052532656705197, |
| "grad_norm": 0.6313768029212952, |
| "learning_rate": 9.421319796954316e-05, |
| "loss": 1.323, |
| "step": 280 |
| }, |
| { |
| "epoch": 0.10901231087303825, |
| "grad_norm": 0.7255018353462219, |
| "learning_rate": 9.395939086294417e-05, |
| "loss": 1.2451, |
| "step": 290 |
| }, |
| { |
| "epoch": 0.11277135607555681, |
| "grad_norm": 0.9848262667655945, |
| "learning_rate": 9.370558375634518e-05, |
| "loss": 1.3681, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.11653040127807537, |
| "grad_norm": 0.9399502277374268, |
| "learning_rate": 9.34517766497462e-05, |
| "loss": 1.3431, |
| "step": 310 |
| }, |
| { |
| "epoch": 0.12028944648059393, |
| "grad_norm": 0.5780265927314758, |
| "learning_rate": 9.31979695431472e-05, |
| "loss": 1.4748, |
| "step": 320 |
| }, |
| { |
| "epoch": 0.1240484916831125, |
| "grad_norm": 0.8427544832229614, |
| "learning_rate": 9.294416243654823e-05, |
| "loss": 1.3739, |
| "step": 330 |
| }, |
| { |
| "epoch": 0.12780753688563104, |
| "grad_norm": 0.9551161527633667, |
| "learning_rate": 9.269035532994924e-05, |
| "loss": 1.3434, |
| "step": 340 |
| }, |
| { |
| "epoch": 0.13156658208814961, |
| "grad_norm": 0.5216005444526672, |
| "learning_rate": 9.243654822335026e-05, |
| "loss": 1.2726, |
| "step": 350 |
| }, |
| { |
| "epoch": 0.13532562729066816, |
| "grad_norm": 1.0278292894363403, |
| "learning_rate": 9.218274111675127e-05, |
| "loss": 1.3311, |
| "step": 360 |
| }, |
| { |
| "epoch": 0.13908467249318673, |
| "grad_norm": 0.8886487483978271, |
| "learning_rate": 9.192893401015229e-05, |
| "loss": 1.5468, |
| "step": 370 |
| }, |
| { |
| "epoch": 0.14284371769570528, |
| "grad_norm": 0.7291779518127441, |
| "learning_rate": 9.16751269035533e-05, |
| "loss": 1.366, |
| "step": 380 |
| }, |
| { |
| "epoch": 0.14660276289822385, |
| "grad_norm": 0.8320682048797607, |
| "learning_rate": 9.142131979695432e-05, |
| "loss": 1.4284, |
| "step": 390 |
| }, |
| { |
| "epoch": 0.1503618081007424, |
| "grad_norm": 0.5894179940223694, |
| "learning_rate": 9.116751269035534e-05, |
| "loss": 1.451, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.15412085330326097, |
| "grad_norm": 0.5339919924736023, |
| "learning_rate": 9.091370558375635e-05, |
| "loss": 1.3939, |
| "step": 410 |
| }, |
| { |
| "epoch": 0.15787989850577952, |
| "grad_norm": 0.6264607310295105, |
| "learning_rate": 9.065989847715737e-05, |
| "loss": 1.4581, |
| "step": 420 |
| }, |
| { |
| "epoch": 0.1616389437082981, |
| "grad_norm": 1.09451162815094, |
| "learning_rate": 9.040609137055838e-05, |
| "loss": 1.3846, |
| "step": 430 |
| }, |
| { |
| "epoch": 0.16539798891081664, |
| "grad_norm": 0.7401901483535767, |
| "learning_rate": 9.015228426395939e-05, |
| "loss": 1.4449, |
| "step": 440 |
| }, |
| { |
| "epoch": 0.16915703411333521, |
| "grad_norm": 0.8674114346504211, |
| "learning_rate": 8.98984771573604e-05, |
| "loss": 1.5506, |
| "step": 450 |
| }, |
| { |
| "epoch": 0.17291607931585376, |
| "grad_norm": 0.6718773245811462, |
| "learning_rate": 8.964467005076143e-05, |
| "loss": 1.4187, |
| "step": 460 |
| }, |
| { |
| "epoch": 0.17667512451837233, |
| "grad_norm": 0.858181893825531, |
| "learning_rate": 8.939086294416244e-05, |
| "loss": 1.4919, |
| "step": 470 |
| }, |
| { |
| "epoch": 0.18043416972089088, |
| "grad_norm": 0.8369438052177429, |
| "learning_rate": 8.913705583756345e-05, |
| "loss": 1.3876, |
| "step": 480 |
| }, |
| { |
| "epoch": 0.18419321492340945, |
| "grad_norm": 0.5989262461662292, |
| "learning_rate": 8.888324873096448e-05, |
| "loss": 1.2856, |
| "step": 490 |
| }, |
| { |
| "epoch": 0.18795226012592803, |
| "grad_norm": 0.4753406345844269, |
| "learning_rate": 8.862944162436549e-05, |
| "loss": 1.1997, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.19171130532844657, |
| "grad_norm": 0.6064810752868652, |
| "learning_rate": 8.83756345177665e-05, |
| "loss": 1.4085, |
| "step": 510 |
| }, |
| { |
| "epoch": 0.19547035053096515, |
| "grad_norm": 0.7949738502502441, |
| "learning_rate": 8.812182741116751e-05, |
| "loss": 1.4494, |
| "step": 520 |
| }, |
| { |
| "epoch": 0.1992293957334837, |
| "grad_norm": 0.6465169191360474, |
| "learning_rate": 8.786802030456854e-05, |
| "loss": 1.3264, |
| "step": 530 |
| }, |
| { |
| "epoch": 0.20298844093600227, |
| "grad_norm": 0.7491289377212524, |
| "learning_rate": 8.761421319796955e-05, |
| "loss": 1.3157, |
| "step": 540 |
| }, |
| { |
| "epoch": 0.20674748613852081, |
| "grad_norm": 0.9102064967155457, |
| "learning_rate": 8.736040609137056e-05, |
| "loss": 1.4135, |
| "step": 550 |
| }, |
| { |
| "epoch": 0.2105065313410394, |
| "grad_norm": 0.9291871786117554, |
| "learning_rate": 8.710659898477158e-05, |
| "loss": 1.3683, |
| "step": 560 |
| }, |
| { |
| "epoch": 0.21426557654355793, |
| "grad_norm": 0.8890318274497986, |
| "learning_rate": 8.685279187817259e-05, |
| "loss": 1.3436, |
| "step": 570 |
| }, |
| { |
| "epoch": 0.2180246217460765, |
| "grad_norm": 0.6270791292190552, |
| "learning_rate": 8.659898477157361e-05, |
| "loss": 1.2768, |
| "step": 580 |
| }, |
| { |
| "epoch": 0.22178366694859505, |
| "grad_norm": 0.8151299953460693, |
| "learning_rate": 8.634517766497462e-05, |
| "loss": 1.2226, |
| "step": 590 |
| }, |
| { |
| "epoch": 0.22554271215111363, |
| "grad_norm": 0.5168191194534302, |
| "learning_rate": 8.609137055837564e-05, |
| "loss": 1.315, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.22930175735363217, |
| "grad_norm": 0.5865817666053772, |
| "learning_rate": 8.583756345177666e-05, |
| "loss": 1.2823, |
| "step": 610 |
| }, |
| { |
| "epoch": 0.23306080255615075, |
| "grad_norm": 0.9748789072036743, |
| "learning_rate": 8.558375634517767e-05, |
| "loss": 1.3953, |
| "step": 620 |
| }, |
| { |
| "epoch": 0.2368198477586693, |
| "grad_norm": 0.8284083604812622, |
| "learning_rate": 8.532994923857869e-05, |
| "loss": 1.3654, |
| "step": 630 |
| }, |
| { |
| "epoch": 0.24057889296118787, |
| "grad_norm": 0.9949294924736023, |
| "learning_rate": 8.50761421319797e-05, |
| "loss": 1.2715, |
| "step": 640 |
| }, |
| { |
| "epoch": 0.24433793816370641, |
| "grad_norm": 0.5778189897537231, |
| "learning_rate": 8.482233502538071e-05, |
| "loss": 1.2411, |
| "step": 650 |
| }, |
| { |
| "epoch": 0.248096983366225, |
| "grad_norm": 0.6740812659263611, |
| "learning_rate": 8.456852791878172e-05, |
| "loss": 1.2492, |
| "step": 660 |
| }, |
| { |
| "epoch": 0.25185602856874356, |
| "grad_norm": 0.9166800379753113, |
| "learning_rate": 8.431472081218275e-05, |
| "loss": 1.4728, |
| "step": 670 |
| }, |
| { |
| "epoch": 0.2556150737712621, |
| "grad_norm": 0.6058619618415833, |
| "learning_rate": 8.406091370558376e-05, |
| "loss": 1.4004, |
| "step": 680 |
| }, |
| { |
| "epoch": 0.25937411897378065, |
| "grad_norm": 1.2171576023101807, |
| "learning_rate": 8.380710659898477e-05, |
| "loss": 1.3007, |
| "step": 690 |
| }, |
| { |
| "epoch": 0.26313316417629923, |
| "grad_norm": 0.5887770056724548, |
| "learning_rate": 8.35532994923858e-05, |
| "loss": 1.374, |
| "step": 700 |
| }, |
| { |
| "epoch": 0.2668922093788178, |
| "grad_norm": 0.7483983635902405, |
| "learning_rate": 8.329949238578681e-05, |
| "loss": 1.3807, |
| "step": 710 |
| }, |
| { |
| "epoch": 0.2706512545813363, |
| "grad_norm": 0.7405831217765808, |
| "learning_rate": 8.304568527918782e-05, |
| "loss": 1.3439, |
| "step": 720 |
| }, |
| { |
| "epoch": 0.2744102997838549, |
| "grad_norm": 0.8032345771789551, |
| "learning_rate": 8.279187817258883e-05, |
| "loss": 1.2951, |
| "step": 730 |
| }, |
| { |
| "epoch": 0.27816934498637347, |
| "grad_norm": 0.7847606539726257, |
| "learning_rate": 8.253807106598986e-05, |
| "loss": 1.2761, |
| "step": 740 |
| }, |
| { |
| "epoch": 0.28192839018889204, |
| "grad_norm": 0.8183022737503052, |
| "learning_rate": 8.228426395939086e-05, |
| "loss": 1.1964, |
| "step": 750 |
| }, |
| { |
| "epoch": 0.28568743539141056, |
| "grad_norm": 0.6903379559516907, |
| "learning_rate": 8.203045685279188e-05, |
| "loss": 1.3832, |
| "step": 760 |
| }, |
| { |
| "epoch": 0.28944648059392913, |
| "grad_norm": 0.8767275214195251, |
| "learning_rate": 8.17766497461929e-05, |
| "loss": 1.4244, |
| "step": 770 |
| }, |
| { |
| "epoch": 0.2932055257964477, |
| "grad_norm": 0.830820620059967, |
| "learning_rate": 8.152284263959391e-05, |
| "loss": 1.2934, |
| "step": 780 |
| }, |
| { |
| "epoch": 0.2969645709989663, |
| "grad_norm": 0.8802516460418701, |
| "learning_rate": 8.126903553299493e-05, |
| "loss": 1.4081, |
| "step": 790 |
| }, |
| { |
| "epoch": 0.3007236162014848, |
| "grad_norm": 0.7873942255973816, |
| "learning_rate": 8.101522842639594e-05, |
| "loss": 1.3219, |
| "step": 800 |
| }, |
| { |
| "epoch": 0.3044826614040034, |
| "grad_norm": 1.127156376838684, |
| "learning_rate": 8.076142131979696e-05, |
| "loss": 1.2824, |
| "step": 810 |
| }, |
| { |
| "epoch": 0.30824170660652195, |
| "grad_norm": 0.9335768818855286, |
| "learning_rate": 8.050761421319797e-05, |
| "loss": 1.3021, |
| "step": 820 |
| }, |
| { |
| "epoch": 0.3120007518090405, |
| "grad_norm": 0.6120972633361816, |
| "learning_rate": 8.0253807106599e-05, |
| "loss": 1.3155, |
| "step": 830 |
| }, |
| { |
| "epoch": 0.31575979701155904, |
| "grad_norm": 0.5859377980232239, |
| "learning_rate": 8e-05, |
| "loss": 1.3117, |
| "step": 840 |
| }, |
| { |
| "epoch": 0.3195188422140776, |
| "grad_norm": 0.6444947719573975, |
| "learning_rate": 7.974619289340102e-05, |
| "loss": 1.3348, |
| "step": 850 |
| }, |
| { |
| "epoch": 0.3232778874165962, |
| "grad_norm": 0.907409131526947, |
| "learning_rate": 7.949238578680203e-05, |
| "loss": 1.2427, |
| "step": 860 |
| }, |
| { |
| "epoch": 0.32703693261911476, |
| "grad_norm": 0.5900988578796387, |
| "learning_rate": 7.923857868020304e-05, |
| "loss": 1.3427, |
| "step": 870 |
| }, |
| { |
| "epoch": 0.3307959778216333, |
| "grad_norm": 0.8546301126480103, |
| "learning_rate": 7.898477157360407e-05, |
| "loss": 1.3824, |
| "step": 880 |
| }, |
| { |
| "epoch": 0.33455502302415185, |
| "grad_norm": 0.9503180384635925, |
| "learning_rate": 7.873096446700508e-05, |
| "loss": 1.2671, |
| "step": 890 |
| }, |
| { |
| "epoch": 0.33831406822667043, |
| "grad_norm": 0.6486964821815491, |
| "learning_rate": 7.847715736040609e-05, |
| "loss": 1.3737, |
| "step": 900 |
| } |
| ], |
| "logging_steps": 10, |
| "max_steps": 3990, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 2, |
| "save_steps": 300, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": false |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 5.663021612477645e+16, |
| "train_batch_size": 2, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|