| { |
| "best_global_step": 400, |
| "best_metric": 2.051694393157959, |
| "best_model_checkpoint": "models/MNLP_M3_rag_model_test/checkpoint-400", |
| "epoch": 2.08355091383812, |
| "eval_steps": 200, |
| "global_step": 400, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.05221932114882506, |
| "grad_norm": 17.204919815063477, |
| "learning_rate": 8.000000000000001e-07, |
| "loss": 3.6299, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.10443864229765012, |
| "grad_norm": 10.213739395141602, |
| "learning_rate": 1.8000000000000001e-06, |
| "loss": 3.263, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.1566579634464752, |
| "grad_norm": 10.35930347442627, |
| "learning_rate": 2.8000000000000003e-06, |
| "loss": 2.7275, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.20887728459530025, |
| "grad_norm": 13.71743106842041, |
| "learning_rate": 3.8000000000000005e-06, |
| "loss": 2.7291, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.26109660574412535, |
| "grad_norm": 9.006918907165527, |
| "learning_rate": 4.800000000000001e-06, |
| "loss": 2.551, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.3133159268929504, |
| "grad_norm": 7.384293556213379, |
| "learning_rate": 5.8e-06, |
| "loss": 2.4662, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.36553524804177545, |
| "grad_norm": 7.212233543395996, |
| "learning_rate": 6.800000000000001e-06, |
| "loss": 2.3987, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.4177545691906005, |
| "grad_norm": 8.014248847961426, |
| "learning_rate": 7.800000000000002e-06, |
| "loss": 2.3993, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.4699738903394256, |
| "grad_norm": 7.093222618103027, |
| "learning_rate": 8.8e-06, |
| "loss": 2.3603, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.5221932114882507, |
| "grad_norm": 9.149968147277832, |
| "learning_rate": 9.800000000000001e-06, |
| "loss": 2.3923, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.5744125326370757, |
| "grad_norm": 8.030098915100098, |
| "learning_rate": 9.943977591036416e-06, |
| "loss": 2.2983, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.6266318537859008, |
| "grad_norm": 7.210897445678711, |
| "learning_rate": 9.873949579831935e-06, |
| "loss": 2.3049, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.6788511749347258, |
| "grad_norm": 7.169530391693115, |
| "learning_rate": 9.803921568627451e-06, |
| "loss": 2.2872, |
| "step": 130 |
| }, |
| { |
| "epoch": 0.7310704960835509, |
| "grad_norm": 7.033287525177002, |
| "learning_rate": 9.73389355742297e-06, |
| "loss": 2.279, |
| "step": 140 |
| }, |
| { |
| "epoch": 0.783289817232376, |
| "grad_norm": 7.60181188583374, |
| "learning_rate": 9.663865546218488e-06, |
| "loss": 2.2453, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.835509138381201, |
| "grad_norm": 7.923684597015381, |
| "learning_rate": 9.593837535014006e-06, |
| "loss": 2.2335, |
| "step": 160 |
| }, |
| { |
| "epoch": 0.8877284595300261, |
| "grad_norm": 6.4125165939331055, |
| "learning_rate": 9.523809523809525e-06, |
| "loss": 2.2682, |
| "step": 170 |
| }, |
| { |
| "epoch": 0.9399477806788512, |
| "grad_norm": 9.670829772949219, |
| "learning_rate": 9.453781512605043e-06, |
| "loss": 2.1589, |
| "step": 180 |
| }, |
| { |
| "epoch": 0.9921671018276762, |
| "grad_norm": 6.57144832611084, |
| "learning_rate": 9.383753501400561e-06, |
| "loss": 2.04, |
| "step": 190 |
| }, |
| { |
| "epoch": 1.04177545691906, |
| "grad_norm": 6.862838268280029, |
| "learning_rate": 9.31372549019608e-06, |
| "loss": 1.867, |
| "step": 200 |
| }, |
| { |
| "epoch": 1.04177545691906, |
| "eval_loss": 2.155261278152466, |
| "eval_runtime": 4.7541, |
| "eval_samples_per_second": 35.969, |
| "eval_steps_per_second": 4.628, |
| "step": 200 |
| }, |
| { |
| "epoch": 1.0939947780678851, |
| "grad_norm": 12.507933616638184, |
| "learning_rate": 9.243697478991598e-06, |
| "loss": 1.7035, |
| "step": 210 |
| }, |
| { |
| "epoch": 1.1462140992167102, |
| "grad_norm": 8.804220199584961, |
| "learning_rate": 9.173669467787115e-06, |
| "loss": 1.6772, |
| "step": 220 |
| }, |
| { |
| "epoch": 1.1984334203655354, |
| "grad_norm": 8.77147102355957, |
| "learning_rate": 9.103641456582633e-06, |
| "loss": 1.6972, |
| "step": 230 |
| }, |
| { |
| "epoch": 1.2506527415143602, |
| "grad_norm": 8.48830795288086, |
| "learning_rate": 9.033613445378152e-06, |
| "loss": 1.5959, |
| "step": 240 |
| }, |
| { |
| "epoch": 1.3028720626631853, |
| "grad_norm": 7.4547648429870605, |
| "learning_rate": 8.96358543417367e-06, |
| "loss": 1.6931, |
| "step": 250 |
| }, |
| { |
| "epoch": 1.3550913838120104, |
| "grad_norm": 10.059548377990723, |
| "learning_rate": 8.893557422969188e-06, |
| "loss": 1.6475, |
| "step": 260 |
| }, |
| { |
| "epoch": 1.4073107049608355, |
| "grad_norm": 8.088046073913574, |
| "learning_rate": 8.823529411764707e-06, |
| "loss": 1.7921, |
| "step": 270 |
| }, |
| { |
| "epoch": 1.4595300261096606, |
| "grad_norm": 7.830897808074951, |
| "learning_rate": 8.753501400560225e-06, |
| "loss": 1.6743, |
| "step": 280 |
| }, |
| { |
| "epoch": 1.5117493472584855, |
| "grad_norm": 10.651778221130371, |
| "learning_rate": 8.683473389355744e-06, |
| "loss": 1.614, |
| "step": 290 |
| }, |
| { |
| "epoch": 1.5639686684073109, |
| "grad_norm": 8.227381706237793, |
| "learning_rate": 8.613445378151262e-06, |
| "loss": 1.612, |
| "step": 300 |
| }, |
| { |
| "epoch": 1.6161879895561357, |
| "grad_norm": 8.035416603088379, |
| "learning_rate": 8.543417366946779e-06, |
| "loss": 1.6633, |
| "step": 310 |
| }, |
| { |
| "epoch": 1.6684073107049608, |
| "grad_norm": 8.318634033203125, |
| "learning_rate": 8.473389355742297e-06, |
| "loss": 1.6989, |
| "step": 320 |
| }, |
| { |
| "epoch": 1.720626631853786, |
| "grad_norm": 6.585997581481934, |
| "learning_rate": 8.403361344537815e-06, |
| "loss": 1.6686, |
| "step": 330 |
| }, |
| { |
| "epoch": 1.7728459530026108, |
| "grad_norm": 6.752286911010742, |
| "learning_rate": 8.333333333333334e-06, |
| "loss": 1.7218, |
| "step": 340 |
| }, |
| { |
| "epoch": 1.8250652741514362, |
| "grad_norm": 8.7686128616333, |
| "learning_rate": 8.263305322128852e-06, |
| "loss": 1.588, |
| "step": 350 |
| }, |
| { |
| "epoch": 1.877284595300261, |
| "grad_norm": 7.1550703048706055, |
| "learning_rate": 8.19327731092437e-06, |
| "loss": 1.5188, |
| "step": 360 |
| }, |
| { |
| "epoch": 1.9295039164490861, |
| "grad_norm": 8.00949478149414, |
| "learning_rate": 8.123249299719889e-06, |
| "loss": 1.5708, |
| "step": 370 |
| }, |
| { |
| "epoch": 1.9817232375979112, |
| "grad_norm": 8.731331825256348, |
| "learning_rate": 8.053221288515407e-06, |
| "loss": 1.565, |
| "step": 380 |
| }, |
| { |
| "epoch": 2.031331592689295, |
| "grad_norm": 8.245712280273438, |
| "learning_rate": 7.983193277310926e-06, |
| "loss": 1.3907, |
| "step": 390 |
| }, |
| { |
| "epoch": 2.08355091383812, |
| "grad_norm": 10.174018859863281, |
| "learning_rate": 7.913165266106442e-06, |
| "loss": 1.1657, |
| "step": 400 |
| }, |
| { |
| "epoch": 2.08355091383812, |
| "eval_loss": 2.051694393157959, |
| "eval_runtime": 4.6523, |
| "eval_samples_per_second": 36.756, |
| "eval_steps_per_second": 4.729, |
| "step": 400 |
| } |
| ], |
| "logging_steps": 10, |
| "max_steps": 1528, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 8, |
| "save_steps": 200, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": false |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 4316440119214080.0, |
| "train_batch_size": 2, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|