| { |
| "best_global_step": 1000, |
| "best_metric": 0.7969154715538025, |
| "best_model_checkpoint": "speecht5_finetuned_nepali/checkpoint-1000", |
| "epoch": 2000.0, |
| "eval_steps": 500, |
| "global_step": 2000, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 25.0, |
| "grad_norm": 10.529245376586914, |
| "learning_rate": 9.5e-06, |
| "loss": 1.0362, |
| "step": 25 |
| }, |
| { |
| "epoch": 50.0, |
| "grad_norm": 14.513840675354004, |
| "learning_rate": 2.2000000000000003e-05, |
| "loss": 0.4845, |
| "step": 50 |
| }, |
| { |
| "epoch": 75.0, |
| "grad_norm": 22.360095977783203, |
| "learning_rate": 3.45e-05, |
| "loss": 0.415, |
| "step": 75 |
| }, |
| { |
| "epoch": 100.0, |
| "grad_norm": 3.0262670516967773, |
| "learning_rate": 4.7e-05, |
| "loss": 0.2888, |
| "step": 100 |
| }, |
| { |
| "epoch": 125.0, |
| "grad_norm": 3.495602607727051, |
| "learning_rate": 5.95e-05, |
| "loss": 0.3118, |
| "step": 125 |
| }, |
| { |
| "epoch": 150.0, |
| "grad_norm": 12.870991706848145, |
| "learning_rate": 7.2e-05, |
| "loss": 0.2475, |
| "step": 150 |
| }, |
| { |
| "epoch": 175.0, |
| "grad_norm": 25.45779800415039, |
| "learning_rate": 8.450000000000001e-05, |
| "loss": 0.2191, |
| "step": 175 |
| }, |
| { |
| "epoch": 200.0, |
| "grad_norm": 16.621234893798828, |
| "learning_rate": 9.7e-05, |
| "loss": 0.2357, |
| "step": 200 |
| }, |
| { |
| "epoch": 225.0, |
| "grad_norm": 3.157877206802368, |
| "learning_rate": 9.894444444444446e-05, |
| "loss": 0.2045, |
| "step": 225 |
| }, |
| { |
| "epoch": 250.0, |
| "grad_norm": 3.6257405281066895, |
| "learning_rate": 9.755555555555555e-05, |
| "loss": 0.2164, |
| "step": 250 |
| }, |
| { |
| "epoch": 275.0, |
| "grad_norm": 3.596989870071411, |
| "learning_rate": 9.616666666666667e-05, |
| "loss": 0.2195, |
| "step": 275 |
| }, |
| { |
| "epoch": 300.0, |
| "grad_norm": 9.64238166809082, |
| "learning_rate": 9.477777777777779e-05, |
| "loss": 0.2001, |
| "step": 300 |
| }, |
| { |
| "epoch": 325.0, |
| "grad_norm": 2.2049977779388428, |
| "learning_rate": 9.33888888888889e-05, |
| "loss": 0.2109, |
| "step": 325 |
| }, |
| { |
| "epoch": 350.0, |
| "grad_norm": 13.762038230895996, |
| "learning_rate": 9.200000000000001e-05, |
| "loss": 0.1858, |
| "step": 350 |
| }, |
| { |
| "epoch": 375.0, |
| "grad_norm": 16.161832809448242, |
| "learning_rate": 9.061111111111112e-05, |
| "loss": 0.2076, |
| "step": 375 |
| }, |
| { |
| "epoch": 400.0, |
| "grad_norm": 4.481337070465088, |
| "learning_rate": 8.922222222222223e-05, |
| "loss": 0.1771, |
| "step": 400 |
| }, |
| { |
| "epoch": 425.0, |
| "grad_norm": 4.335459232330322, |
| "learning_rate": 8.783333333333333e-05, |
| "loss": 0.1854, |
| "step": 425 |
| }, |
| { |
| "epoch": 450.0, |
| "grad_norm": 4.443661212921143, |
| "learning_rate": 8.644444444444445e-05, |
| "loss": 0.1635, |
| "step": 450 |
| }, |
| { |
| "epoch": 475.0, |
| "grad_norm": 5.2760796546936035, |
| "learning_rate": 8.505555555555556e-05, |
| "loss": 0.1767, |
| "step": 475 |
| }, |
| { |
| "epoch": 500.0, |
| "grad_norm": 13.874415397644043, |
| "learning_rate": 8.366666666666668e-05, |
| "loss": 0.1732, |
| "step": 500 |
| }, |
| { |
| "epoch": 500.0, |
| "eval_loss": 0.8349313735961914, |
| "eval_runtime": 0.0239, |
| "eval_samples_per_second": 41.882, |
| "eval_steps_per_second": 41.882, |
| "step": 500 |
| }, |
| { |
| "epoch": 525.0, |
| "grad_norm": 7.144843101501465, |
| "learning_rate": 8.227777777777778e-05, |
| "loss": 0.1622, |
| "step": 525 |
| }, |
| { |
| "epoch": 550.0, |
| "grad_norm": 9.63769245147705, |
| "learning_rate": 8.088888888888889e-05, |
| "loss": 0.1594, |
| "step": 550 |
| }, |
| { |
| "epoch": 575.0, |
| "grad_norm": 18.466543197631836, |
| "learning_rate": 7.950000000000001e-05, |
| "loss": 0.1584, |
| "step": 575 |
| }, |
| { |
| "epoch": 600.0, |
| "grad_norm": 4.807222843170166, |
| "learning_rate": 7.811111111111111e-05, |
| "loss": 0.1928, |
| "step": 600 |
| }, |
| { |
| "epoch": 625.0, |
| "grad_norm": 7.026573657989502, |
| "learning_rate": 7.672222222222223e-05, |
| "loss": 0.1403, |
| "step": 625 |
| }, |
| { |
| "epoch": 650.0, |
| "grad_norm": 1.3481919765472412, |
| "learning_rate": 7.533333333333334e-05, |
| "loss": 0.1454, |
| "step": 650 |
| }, |
| { |
| "epoch": 675.0, |
| "grad_norm": 1.3561595678329468, |
| "learning_rate": 7.394444444444445e-05, |
| "loss": 0.128, |
| "step": 675 |
| }, |
| { |
| "epoch": 700.0, |
| "grad_norm": 8.730093955993652, |
| "learning_rate": 7.255555555555555e-05, |
| "loss": 0.1306, |
| "step": 700 |
| }, |
| { |
| "epoch": 725.0, |
| "grad_norm": 1.6442005634307861, |
| "learning_rate": 7.116666666666667e-05, |
| "loss": 0.1332, |
| "step": 725 |
| }, |
| { |
| "epoch": 750.0, |
| "grad_norm": 2.5606157779693604, |
| "learning_rate": 6.977777777777779e-05, |
| "loss": 0.1289, |
| "step": 750 |
| }, |
| { |
| "epoch": 775.0, |
| "grad_norm": 10.112384796142578, |
| "learning_rate": 6.83888888888889e-05, |
| "loss": 0.1354, |
| "step": 775 |
| }, |
| { |
| "epoch": 800.0, |
| "grad_norm": 1.0964512825012207, |
| "learning_rate": 6.7e-05, |
| "loss": 0.1147, |
| "step": 800 |
| }, |
| { |
| "epoch": 825.0, |
| "grad_norm": 1.3629651069641113, |
| "learning_rate": 6.561111111111111e-05, |
| "loss": 0.1129, |
| "step": 825 |
| }, |
| { |
| "epoch": 850.0, |
| "grad_norm": 12.741216659545898, |
| "learning_rate": 6.422222222222223e-05, |
| "loss": 0.1098, |
| "step": 850 |
| }, |
| { |
| "epoch": 875.0, |
| "grad_norm": 2.6955878734588623, |
| "learning_rate": 6.283333333333333e-05, |
| "loss": 0.1103, |
| "step": 875 |
| }, |
| { |
| "epoch": 900.0, |
| "grad_norm": 1.8039289712905884, |
| "learning_rate": 6.144444444444445e-05, |
| "loss": 0.114, |
| "step": 900 |
| }, |
| { |
| "epoch": 925.0, |
| "grad_norm": 2.290956735610962, |
| "learning_rate": 6.005555555555555e-05, |
| "loss": 0.1182, |
| "step": 925 |
| }, |
| { |
| "epoch": 950.0, |
| "grad_norm": 0.9769994616508484, |
| "learning_rate": 5.866666666666667e-05, |
| "loss": 0.1236, |
| "step": 950 |
| }, |
| { |
| "epoch": 975.0, |
| "grad_norm": 0.7922130227088928, |
| "learning_rate": 5.7277777777777785e-05, |
| "loss": 0.1029, |
| "step": 975 |
| }, |
| { |
| "epoch": 1000.0, |
| "grad_norm": 0.9130666255950928, |
| "learning_rate": 5.588888888888889e-05, |
| "loss": 0.1032, |
| "step": 1000 |
| }, |
| { |
| "epoch": 1000.0, |
| "eval_loss": 0.7969154715538025, |
| "eval_runtime": 0.0205, |
| "eval_samples_per_second": 48.798, |
| "eval_steps_per_second": 48.798, |
| "step": 1000 |
| }, |
| { |
| "epoch": 1025.0, |
| "grad_norm": 3.981950283050537, |
| "learning_rate": 5.45e-05, |
| "loss": 0.1481, |
| "step": 1025 |
| }, |
| { |
| "epoch": 1050.0, |
| "grad_norm": 1.1825660467147827, |
| "learning_rate": 5.311111111111111e-05, |
| "loss": 0.104, |
| "step": 1050 |
| }, |
| { |
| "epoch": 1075.0, |
| "grad_norm": 2.0291762351989746, |
| "learning_rate": 5.172222222222223e-05, |
| "loss": 0.1049, |
| "step": 1075 |
| }, |
| { |
| "epoch": 1100.0, |
| "grad_norm": 0.8333075046539307, |
| "learning_rate": 5.0333333333333335e-05, |
| "loss": 0.0917, |
| "step": 1100 |
| }, |
| { |
| "epoch": 1125.0, |
| "grad_norm": 19.428544998168945, |
| "learning_rate": 4.894444444444445e-05, |
| "loss": 0.1106, |
| "step": 1125 |
| }, |
| { |
| "epoch": 1150.0, |
| "grad_norm": 1.1518399715423584, |
| "learning_rate": 4.755555555555556e-05, |
| "loss": 0.0911, |
| "step": 1150 |
| }, |
| { |
| "epoch": 1175.0, |
| "grad_norm": 11.079072952270508, |
| "learning_rate": 4.6166666666666666e-05, |
| "loss": 0.0853, |
| "step": 1175 |
| }, |
| { |
| "epoch": 1200.0, |
| "grad_norm": 3.140981435775757, |
| "learning_rate": 4.477777777777778e-05, |
| "loss": 0.1094, |
| "step": 1200 |
| }, |
| { |
| "epoch": 1225.0, |
| "grad_norm": 1.3652321100234985, |
| "learning_rate": 4.338888888888889e-05, |
| "loss": 0.085, |
| "step": 1225 |
| }, |
| { |
| "epoch": 1250.0, |
| "grad_norm": 4.102111339569092, |
| "learning_rate": 4.2e-05, |
| "loss": 0.0715, |
| "step": 1250 |
| }, |
| { |
| "epoch": 1275.0, |
| "grad_norm": 0.7971464395523071, |
| "learning_rate": 4.061111111111111e-05, |
| "loss": 0.0995, |
| "step": 1275 |
| }, |
| { |
| "epoch": 1300.0, |
| "grad_norm": 2.4831583499908447, |
| "learning_rate": 3.922222222222223e-05, |
| "loss": 0.0911, |
| "step": 1300 |
| }, |
| { |
| "epoch": 1325.0, |
| "grad_norm": 0.8036356568336487, |
| "learning_rate": 3.7833333333333336e-05, |
| "loss": 0.0744, |
| "step": 1325 |
| }, |
| { |
| "epoch": 1350.0, |
| "grad_norm": 0.7885045409202576, |
| "learning_rate": 3.644444444444445e-05, |
| "loss": 0.0819, |
| "step": 1350 |
| }, |
| { |
| "epoch": 1375.0, |
| "grad_norm": 0.8334118127822876, |
| "learning_rate": 3.505555555555556e-05, |
| "loss": 0.0814, |
| "step": 1375 |
| }, |
| { |
| "epoch": 1400.0, |
| "grad_norm": 3.9760286808013916, |
| "learning_rate": 3.366666666666667e-05, |
| "loss": 0.0787, |
| "step": 1400 |
| }, |
| { |
| "epoch": 1425.0, |
| "grad_norm": 1.063104271888733, |
| "learning_rate": 3.227777777777778e-05, |
| "loss": 0.0922, |
| "step": 1425 |
| }, |
| { |
| "epoch": 1450.0, |
| "grad_norm": 1.2303999662399292, |
| "learning_rate": 3.088888888888889e-05, |
| "loss": 0.0834, |
| "step": 1450 |
| }, |
| { |
| "epoch": 1475.0, |
| "grad_norm": 3.905423164367676, |
| "learning_rate": 2.95e-05, |
| "loss": 0.0704, |
| "step": 1475 |
| }, |
| { |
| "epoch": 1500.0, |
| "grad_norm": 0.5851226449012756, |
| "learning_rate": 2.811111111111111e-05, |
| "loss": 0.0938, |
| "step": 1500 |
| }, |
| { |
| "epoch": 1500.0, |
| "eval_loss": 0.8552892804145813, |
| "eval_runtime": 0.0192, |
| "eval_samples_per_second": 52.216, |
| "eval_steps_per_second": 52.216, |
| "step": 1500 |
| }, |
| { |
| "epoch": 1525.0, |
| "grad_norm": 0.6819572448730469, |
| "learning_rate": 2.6722222222222228e-05, |
| "loss": 0.0679, |
| "step": 1525 |
| }, |
| { |
| "epoch": 1550.0, |
| "grad_norm": 0.800021767616272, |
| "learning_rate": 2.5333333333333337e-05, |
| "loss": 0.0695, |
| "step": 1550 |
| }, |
| { |
| "epoch": 1575.0, |
| "grad_norm": 0.6414686441421509, |
| "learning_rate": 2.3944444444444443e-05, |
| "loss": 0.069, |
| "step": 1575 |
| }, |
| { |
| "epoch": 1600.0, |
| "grad_norm": 0.7216551303863525, |
| "learning_rate": 2.255555555555556e-05, |
| "loss": 0.0688, |
| "step": 1600 |
| }, |
| { |
| "epoch": 1625.0, |
| "grad_norm": 2.296966075897217, |
| "learning_rate": 2.116666666666667e-05, |
| "loss": 0.0705, |
| "step": 1625 |
| }, |
| { |
| "epoch": 1650.0, |
| "grad_norm": 0.7505176067352295, |
| "learning_rate": 1.9777777777777778e-05, |
| "loss": 0.0788, |
| "step": 1650 |
| }, |
| { |
| "epoch": 1675.0, |
| "grad_norm": 0.9566255807876587, |
| "learning_rate": 1.838888888888889e-05, |
| "loss": 0.0799, |
| "step": 1675 |
| }, |
| { |
| "epoch": 1700.0, |
| "grad_norm": 3.685331106185913, |
| "learning_rate": 1.7000000000000003e-05, |
| "loss": 0.0805, |
| "step": 1700 |
| }, |
| { |
| "epoch": 1725.0, |
| "grad_norm": 0.3868086338043213, |
| "learning_rate": 1.5611111111111113e-05, |
| "loss": 0.075, |
| "step": 1725 |
| }, |
| { |
| "epoch": 1750.0, |
| "grad_norm": 0.690929114818573, |
| "learning_rate": 1.4222222222222224e-05, |
| "loss": 0.0637, |
| "step": 1750 |
| }, |
| { |
| "epoch": 1775.0, |
| "grad_norm": 6.572366714477539, |
| "learning_rate": 1.2833333333333333e-05, |
| "loss": 0.0709, |
| "step": 1775 |
| }, |
| { |
| "epoch": 1800.0, |
| "grad_norm": 0.6316695213317871, |
| "learning_rate": 1.1444444444444446e-05, |
| "loss": 0.0708, |
| "step": 1800 |
| }, |
| { |
| "epoch": 1825.0, |
| "grad_norm": 16.782623291015625, |
| "learning_rate": 1.0055555555555555e-05, |
| "loss": 0.0711, |
| "step": 1825 |
| }, |
| { |
| "epoch": 1850.0, |
| "grad_norm": 0.42628633975982666, |
| "learning_rate": 8.666666666666668e-06, |
| "loss": 0.0713, |
| "step": 1850 |
| }, |
| { |
| "epoch": 1875.0, |
| "grad_norm": 0.48583984375, |
| "learning_rate": 7.277777777777778e-06, |
| "loss": 0.0665, |
| "step": 1875 |
| }, |
| { |
| "epoch": 1900.0, |
| "grad_norm": 0.5549265146255493, |
| "learning_rate": 5.888888888888889e-06, |
| "loss": 0.0658, |
| "step": 1900 |
| }, |
| { |
| "epoch": 1925.0, |
| "grad_norm": 2.4642038345336914, |
| "learning_rate": 4.5e-06, |
| "loss": 0.0687, |
| "step": 1925 |
| }, |
| { |
| "epoch": 1950.0, |
| "grad_norm": 0.607486367225647, |
| "learning_rate": 3.111111111111111e-06, |
| "loss": 0.0733, |
| "step": 1950 |
| }, |
| { |
| "epoch": 1975.0, |
| "grad_norm": 0.6398457884788513, |
| "learning_rate": 1.7222222222222222e-06, |
| "loss": 0.066, |
| "step": 1975 |
| }, |
| { |
| "epoch": 2000.0, |
| "grad_norm": 0.43928271532058716, |
| "learning_rate": 3.3333333333333335e-07, |
| "loss": 0.1062, |
| "step": 2000 |
| }, |
| { |
| "epoch": 2000.0, |
| "eval_loss": 0.8336465954780579, |
| "eval_runtime": 0.0207, |
| "eval_samples_per_second": 48.371, |
| "eval_steps_per_second": 48.371, |
| "step": 2000 |
| } |
| ], |
| "logging_steps": 25, |
| "max_steps": 2000, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 2000, |
| "save_steps": 500, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 48501262656000.0, |
| "train_batch_size": 8, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|