| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 4.0, |
| "eval_steps": 500, |
| "global_step": 4120, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.24271844660194175, |
| "grad_norm": 0.29302138090133667, |
| "learning_rate": 9.393203883495146e-05, |
| "loss": 2.8318, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.4854368932038835, |
| "grad_norm": 0.24823348224163055, |
| "learning_rate": 8.786407766990292e-05, |
| "loss": 2.4178, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.7281553398058253, |
| "grad_norm": 0.2736216187477112, |
| "learning_rate": 8.179611650485438e-05, |
| "loss": 2.3486, |
| "step": 750 |
| }, |
| { |
| "epoch": 0.970873786407767, |
| "grad_norm": 0.23155324161052704, |
| "learning_rate": 7.572815533980583e-05, |
| "loss": 2.321, |
| "step": 1000 |
| }, |
| { |
| "epoch": 1.2135922330097086, |
| "grad_norm": 0.3087535500526428, |
| "learning_rate": 6.966019417475728e-05, |
| "loss": 2.3124, |
| "step": 1250 |
| }, |
| { |
| "epoch": 1.4563106796116505, |
| "grad_norm": 0.2022610604763031, |
| "learning_rate": 6.359223300970875e-05, |
| "loss": 2.307, |
| "step": 1500 |
| }, |
| { |
| "epoch": 1.6990291262135924, |
| "grad_norm": 0.1987154483795166, |
| "learning_rate": 5.752427184466019e-05, |
| "loss": 2.3019, |
| "step": 1750 |
| }, |
| { |
| "epoch": 1.941747572815534, |
| "grad_norm": 0.16502627730369568, |
| "learning_rate": 5.145631067961165e-05, |
| "loss": 2.3073, |
| "step": 2000 |
| }, |
| { |
| "epoch": 2.1844660194174756, |
| "grad_norm": 0.19471481442451477, |
| "learning_rate": 4.538834951456311e-05, |
| "loss": 2.2879, |
| "step": 2250 |
| }, |
| { |
| "epoch": 2.4271844660194173, |
| "grad_norm": 0.20921674370765686, |
| "learning_rate": 3.9320388349514564e-05, |
| "loss": 2.2888, |
| "step": 2500 |
| }, |
| { |
| "epoch": 2.6699029126213594, |
| "grad_norm": 0.2216738611459732, |
| "learning_rate": 3.325242718446602e-05, |
| "loss": 2.2997, |
| "step": 2750 |
| }, |
| { |
| "epoch": 2.912621359223301, |
| "grad_norm": 0.2476879507303238, |
| "learning_rate": 2.7184466019417475e-05, |
| "loss": 2.2748, |
| "step": 3000 |
| }, |
| { |
| "epoch": 3.1553398058252426, |
| "grad_norm": 0.2942667603492737, |
| "learning_rate": 2.111650485436893e-05, |
| "loss": 2.2749, |
| "step": 3250 |
| }, |
| { |
| "epoch": 3.3980582524271843, |
| "grad_norm": 0.2994736135005951, |
| "learning_rate": 1.5048543689320387e-05, |
| "loss": 2.2738, |
| "step": 3500 |
| }, |
| { |
| "epoch": 3.6407766990291264, |
| "grad_norm": 0.2753419876098633, |
| "learning_rate": 8.980582524271845e-06, |
| "loss": 2.2557, |
| "step": 3750 |
| }, |
| { |
| "epoch": 3.883495145631068, |
| "grad_norm": 0.277048796415329, |
| "learning_rate": 2.912621359223301e-06, |
| "loss": 2.2512, |
| "step": 4000 |
| }, |
| { |
| "epoch": 4.0, |
| "step": 4120, |
| "total_flos": 7.617568236584731e+18, |
| "train_loss": 2.3324776880949445, |
| "train_runtime": 61881.1241, |
| "train_samples_per_second": 8.522, |
| "train_steps_per_second": 0.067 |
| } |
| ], |
| "logging_steps": 250, |
| "max_steps": 4120, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 4, |
| "save_steps": 500, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": false, |
| "should_training_stop": false |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 7.617568236584731e+18, |
| "train_batch_size": 8, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|