| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 2.3823704586063132, |
| "eval_steps": 500, |
| "global_step": 2000, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.11911852293031566, |
| "grad_norm": 0.3590008318424225, |
| "learning_rate": 1.6666666666666667e-05, |
| "loss": 0.7627, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.23823704586063132, |
| "grad_norm": 0.5137978196144104, |
| "learning_rate": 3.3333333333333335e-05, |
| "loss": 0.6067, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.357355568790947, |
| "grad_norm": 0.46571800112724304, |
| "learning_rate": 5e-05, |
| "loss": 0.5131, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.47647409172126265, |
| "grad_norm": 0.48380178213119507, |
| "learning_rate": 4.957432749209755e-05, |
| "loss": 0.4791, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.5955926146515783, |
| "grad_norm": 0.5578784942626953, |
| "learning_rate": 4.8311805735108894e-05, |
| "loss": 0.4766, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.714711137581894, |
| "grad_norm": 0.4985765516757965, |
| "learning_rate": 4.625542839324036e-05, |
| "loss": 0.4803, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.8338296605122096, |
| "grad_norm": 0.5892139673233032, |
| "learning_rate": 4.347522293051648e-05, |
| "loss": 0.4718, |
| "step": 700 |
| }, |
| { |
| "epoch": 0.9529481834425253, |
| "grad_norm": 0.5380077362060547, |
| "learning_rate": 4.0065865909481417e-05, |
| "loss": 0.4626, |
| "step": 800 |
| }, |
| { |
| "epoch": 1.072066706372841, |
| "grad_norm": 0.576122522354126, |
| "learning_rate": 3.6143458894413465e-05, |
| "loss": 0.4564, |
| "step": 900 |
| }, |
| { |
| "epoch": 1.1911852293031566, |
| "grad_norm": 0.543662428855896, |
| "learning_rate": 3.1841574751802076e-05, |
| "loss": 0.4526, |
| "step": 1000 |
| }, |
| { |
| "epoch": 1.3103037522334722, |
| "grad_norm": 0.6222002506256104, |
| "learning_rate": 2.7306708986582553e-05, |
| "loss": 0.4626, |
| "step": 1100 |
| }, |
| { |
| "epoch": 1.429422275163788, |
| "grad_norm": 0.5597625374794006, |
| "learning_rate": 2.2693291013417453e-05, |
| "loss": 0.4581, |
| "step": 1200 |
| }, |
| { |
| "epoch": 1.5485407980941037, |
| "grad_norm": 0.8017860651016235, |
| "learning_rate": 1.815842524819793e-05, |
| "loss": 0.4646, |
| "step": 1300 |
| }, |
| { |
| "epoch": 1.6676593210244193, |
| "grad_norm": 0.696183443069458, |
| "learning_rate": 1.3856541105586545e-05, |
| "loss": 0.4548, |
| "step": 1400 |
| }, |
| { |
| "epoch": 1.7867778439547348, |
| "grad_norm": 0.6249508261680603, |
| "learning_rate": 9.934134090518593e-06, |
| "loss": 0.4668, |
| "step": 1500 |
| }, |
| { |
| "epoch": 1.9058963668850506, |
| "grad_norm": 0.679349422454834, |
| "learning_rate": 6.524777069483526e-06, |
| "loss": 0.4519, |
| "step": 1600 |
| }, |
| { |
| "epoch": 2.0250148898153664, |
| "grad_norm": 0.6183383464813232, |
| "learning_rate": 3.7445716067596503e-06, |
| "loss": 0.4554, |
| "step": 1700 |
| }, |
| { |
| "epoch": 2.144133412745682, |
| "grad_norm": 0.7539538145065308, |
| "learning_rate": 1.6881942648911076e-06, |
| "loss": 0.4568, |
| "step": 1800 |
| }, |
| { |
| "epoch": 2.2632519356759975, |
| "grad_norm": 0.7621870636940002, |
| "learning_rate": 4.256725079024554e-07, |
| "loss": 0.4577, |
| "step": 1900 |
| }, |
| { |
| "epoch": 2.3823704586063132, |
| "grad_norm": 0.45056673884391785, |
| "learning_rate": 0.0, |
| "loss": 0.4445, |
| "step": 2000 |
| } |
| ], |
| "logging_steps": 100, |
| "max_steps": 2000, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 3, |
| "save_steps": 200, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 6.15307218125783e+16, |
| "train_batch_size": 8, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|