| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 1.0, |
| "eval_steps": 500, |
| "global_step": 24, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.04, |
| "grad_norm": 0.049072265625, |
| "learning_rate": 0.0002, |
| "loss": 1.2558, |
| "step": 1 |
| }, |
| { |
| "epoch": 0.08, |
| "grad_norm": 0.046630859375, |
| "learning_rate": 0.0002, |
| "loss": 1.276, |
| "step": 2 |
| }, |
| { |
| "epoch": 0.12, |
| "grad_norm": 0.046875, |
| "learning_rate": 0.0002, |
| "loss": 1.2758, |
| "step": 3 |
| }, |
| { |
| "epoch": 0.17, |
| "grad_norm": 0.054443359375, |
| "learning_rate": 0.0002, |
| "loss": 1.2889, |
| "step": 4 |
| }, |
| { |
| "epoch": 0.21, |
| "grad_norm": 0.058349609375, |
| "learning_rate": 0.0002, |
| "loss": 1.3781, |
| "step": 5 |
| }, |
| { |
| "epoch": 0.25, |
| "grad_norm": 0.042724609375, |
| "learning_rate": 0.0002, |
| "loss": 1.2866, |
| "step": 6 |
| }, |
| { |
| "epoch": 0.29, |
| "grad_norm": 0.041748046875, |
| "learning_rate": 0.0002, |
| "loss": 1.1998, |
| "step": 7 |
| }, |
| { |
| "epoch": 0.33, |
| "grad_norm": 0.044677734375, |
| "learning_rate": 0.0002, |
| "loss": 1.2973, |
| "step": 8 |
| }, |
| { |
| "epoch": 0.38, |
| "grad_norm": 0.044921875, |
| "learning_rate": 0.0002, |
| "loss": 1.2811, |
| "step": 9 |
| }, |
| { |
| "epoch": 0.42, |
| "grad_norm": 0.039794921875, |
| "learning_rate": 0.0002, |
| "loss": 1.2072, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.46, |
| "grad_norm": 0.040283203125, |
| "learning_rate": 0.0002, |
| "loss": 1.2256, |
| "step": 11 |
| }, |
| { |
| "epoch": 0.5, |
| "grad_norm": 0.0380859375, |
| "learning_rate": 0.0002, |
| "loss": 1.3164, |
| "step": 12 |
| }, |
| { |
| "epoch": 0.54, |
| "grad_norm": 0.0380859375, |
| "learning_rate": 0.0002, |
| "loss": 1.2636, |
| "step": 13 |
| }, |
| { |
| "epoch": 0.58, |
| "grad_norm": 0.039794921875, |
| "learning_rate": 0.0002, |
| "loss": 1.2758, |
| "step": 14 |
| }, |
| { |
| "epoch": 0.62, |
| "grad_norm": 0.0400390625, |
| "learning_rate": 0.0002, |
| "loss": 1.097, |
| "step": 15 |
| }, |
| { |
| "epoch": 0.67, |
| "grad_norm": 0.04248046875, |
| "learning_rate": 0.0002, |
| "loss": 1.1723, |
| "step": 16 |
| }, |
| { |
| "epoch": 0.71, |
| "grad_norm": 0.0419921875, |
| "learning_rate": 0.0002, |
| "loss": 1.2034, |
| "step": 17 |
| }, |
| { |
| "epoch": 0.75, |
| "grad_norm": 0.043212890625, |
| "learning_rate": 0.0002, |
| "loss": 1.1135, |
| "step": 18 |
| }, |
| { |
| "epoch": 0.79, |
| "grad_norm": 0.044921875, |
| "learning_rate": 0.0002, |
| "loss": 1.2245, |
| "step": 19 |
| }, |
| { |
| "epoch": 0.83, |
| "grad_norm": 0.04541015625, |
| "learning_rate": 0.0002, |
| "loss": 1.1489, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.88, |
| "grad_norm": 0.042724609375, |
| "learning_rate": 0.0002, |
| "loss": 1.193, |
| "step": 21 |
| }, |
| { |
| "epoch": 0.92, |
| "grad_norm": 0.040771484375, |
| "learning_rate": 0.0002, |
| "loss": 1.2408, |
| "step": 22 |
| }, |
| { |
| "epoch": 0.96, |
| "grad_norm": 0.046142578125, |
| "learning_rate": 0.0002, |
| "loss": 1.1367, |
| "step": 23 |
| }, |
| { |
| "epoch": 1.0, |
| "grad_norm": 0.0419921875, |
| "learning_rate": 0.0002, |
| "loss": 1.0999, |
| "step": 24 |
| } |
| ], |
| "logging_steps": 1, |
| "max_steps": 24, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 1, |
| "save_steps": 8, |
| "total_flos": 4961451135467520.0, |
| "train_batch_size": 8, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|