| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 9.0, |
| "eval_steps": 500, |
| "global_step": 999, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.009009009009009009, |
| "grad_norm": 1.7692377372839305, |
| "learning_rate": 5e-06, |
| "loss": 0.3442, |
| "step": 1 |
| }, |
| { |
| "epoch": 0.018018018018018018, |
| "grad_norm": 1.655774168281545, |
| "learning_rate": 5e-06, |
| "loss": 0.374, |
| "step": 2 |
| }, |
| { |
| "epoch": 0.02702702702702703, |
| "grad_norm": 1.5838611317965265, |
| "learning_rate": 5e-06, |
| "loss": 0.3763, |
| "step": 3 |
| }, |
| { |
| "epoch": 0.036036036036036036, |
| "grad_norm": 1.4722490643600856, |
| "learning_rate": 5e-06, |
| "loss": 0.3667, |
| "step": 4 |
| }, |
| { |
| "epoch": 0.04504504504504504, |
| "grad_norm": 0.967982129724269, |
| "learning_rate": 5e-06, |
| "loss": 0.3378, |
| "step": 5 |
| }, |
| { |
| "epoch": 0.05405405405405406, |
| "grad_norm": 0.9655512366546067, |
| "learning_rate": 5e-06, |
| "loss": 0.325, |
| "step": 6 |
| }, |
| { |
| "epoch": 0.06306306306306306, |
| "grad_norm": 0.7980444967597017, |
| "learning_rate": 5e-06, |
| "loss": 0.3418, |
| "step": 7 |
| }, |
| { |
| "epoch": 0.07207207207207207, |
| "grad_norm": 0.5186921114651042, |
| "learning_rate": 5e-06, |
| "loss": 0.3508, |
| "step": 8 |
| }, |
| { |
| "epoch": 0.08108108108108109, |
| "grad_norm": 0.5518215648538942, |
| "learning_rate": 5e-06, |
| "loss": 0.348, |
| "step": 9 |
| }, |
| { |
| "epoch": 0.09009009009009009, |
| "grad_norm": 0.6622509236535837, |
| "learning_rate": 5e-06, |
| "loss": 0.3539, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.0990990990990991, |
| "grad_norm": 0.6537261351376887, |
| "learning_rate": 5e-06, |
| "loss": 0.3364, |
| "step": 11 |
| }, |
| { |
| "epoch": 0.10810810810810811, |
| "grad_norm": 0.6557224204301801, |
| "learning_rate": 5e-06, |
| "loss": 0.3529, |
| "step": 12 |
| }, |
| { |
| "epoch": 0.11711711711711711, |
| "grad_norm": 0.6677743317643713, |
| "learning_rate": 5e-06, |
| "loss": 0.3233, |
| "step": 13 |
| }, |
| { |
| "epoch": 0.12612612612612611, |
| "grad_norm": 0.5771734482767436, |
| "learning_rate": 5e-06, |
| "loss": 0.3433, |
| "step": 14 |
| }, |
| { |
| "epoch": 0.13513513513513514, |
| "grad_norm": 0.5194262746227281, |
| "learning_rate": 5e-06, |
| "loss": 0.2968, |
| "step": 15 |
| }, |
| { |
| "epoch": 0.14414414414414414, |
| "grad_norm": 0.5871866323370637, |
| "learning_rate": 5e-06, |
| "loss": 0.3177, |
| "step": 16 |
| }, |
| { |
| "epoch": 0.15315315315315314, |
| "grad_norm": 0.6823752349157315, |
| "learning_rate": 5e-06, |
| "loss": 0.3217, |
| "step": 17 |
| }, |
| { |
| "epoch": 0.16216216216216217, |
| "grad_norm": 0.5298328303770766, |
| "learning_rate": 5e-06, |
| "loss": 0.333, |
| "step": 18 |
| }, |
| { |
| "epoch": 0.17117117117117117, |
| "grad_norm": 0.45672917289622006, |
| "learning_rate": 5e-06, |
| "loss": 0.2998, |
| "step": 19 |
| }, |
| { |
| "epoch": 0.18018018018018017, |
| "grad_norm": 0.4892453526407057, |
| "learning_rate": 5e-06, |
| "loss": 0.3301, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.1891891891891892, |
| "grad_norm": 0.4157035225188495, |
| "learning_rate": 5e-06, |
| "loss": 0.3097, |
| "step": 21 |
| }, |
| { |
| "epoch": 0.1981981981981982, |
| "grad_norm": 0.42144355038756004, |
| "learning_rate": 5e-06, |
| "loss": 0.336, |
| "step": 22 |
| }, |
| { |
| "epoch": 0.2072072072072072, |
| "grad_norm": 0.40449172267977285, |
| "learning_rate": 5e-06, |
| "loss": 0.3203, |
| "step": 23 |
| }, |
| { |
| "epoch": 0.21621621621621623, |
| "grad_norm": 0.3817504264369776, |
| "learning_rate": 5e-06, |
| "loss": 0.3282, |
| "step": 24 |
| }, |
| { |
| "epoch": 0.22522522522522523, |
| "grad_norm": 0.37458931065383283, |
| "learning_rate": 5e-06, |
| "loss": 0.3427, |
| "step": 25 |
| }, |
| { |
| "epoch": 0.23423423423423423, |
| "grad_norm": 0.43415654347436194, |
| "learning_rate": 5e-06, |
| "loss": 0.3361, |
| "step": 26 |
| }, |
| { |
| "epoch": 0.24324324324324326, |
| "grad_norm": 0.34734907350951355, |
| "learning_rate": 5e-06, |
| "loss": 0.3081, |
| "step": 27 |
| }, |
| { |
| "epoch": 0.25225225225225223, |
| "grad_norm": 0.3446691978222806, |
| "learning_rate": 5e-06, |
| "loss": 0.3104, |
| "step": 28 |
| }, |
| { |
| "epoch": 0.26126126126126126, |
| "grad_norm": 0.3219457244434707, |
| "learning_rate": 5e-06, |
| "loss": 0.3154, |
| "step": 29 |
| }, |
| { |
| "epoch": 0.2702702702702703, |
| "grad_norm": 0.35333024684448033, |
| "learning_rate": 5e-06, |
| "loss": 0.3238, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.27927927927927926, |
| "grad_norm": 0.38018940900412435, |
| "learning_rate": 5e-06, |
| "loss": 0.3479, |
| "step": 31 |
| }, |
| { |
| "epoch": 0.2882882882882883, |
| "grad_norm": 0.4160537077429581, |
| "learning_rate": 5e-06, |
| "loss": 0.3225, |
| "step": 32 |
| }, |
| { |
| "epoch": 0.2972972972972973, |
| "grad_norm": 0.4001899610048794, |
| "learning_rate": 5e-06, |
| "loss": 0.3378, |
| "step": 33 |
| }, |
| { |
| "epoch": 0.3063063063063063, |
| "grad_norm": 0.3966450451230361, |
| "learning_rate": 5e-06, |
| "loss": 0.3136, |
| "step": 34 |
| }, |
| { |
| "epoch": 0.3153153153153153, |
| "grad_norm": 0.35442342787868963, |
| "learning_rate": 5e-06, |
| "loss": 0.3272, |
| "step": 35 |
| }, |
| { |
| "epoch": 0.32432432432432434, |
| "grad_norm": 0.31417075347024526, |
| "learning_rate": 5e-06, |
| "loss": 0.34, |
| "step": 36 |
| }, |
| { |
| "epoch": 0.3333333333333333, |
| "grad_norm": 0.31460209634883374, |
| "learning_rate": 5e-06, |
| "loss": 0.3068, |
| "step": 37 |
| }, |
| { |
| "epoch": 0.34234234234234234, |
| "grad_norm": 0.27638346613404846, |
| "learning_rate": 5e-06, |
| "loss": 0.3355, |
| "step": 38 |
| }, |
| { |
| "epoch": 0.35135135135135137, |
| "grad_norm": 0.31966876717000925, |
| "learning_rate": 5e-06, |
| "loss": 0.3176, |
| "step": 39 |
| }, |
| { |
| "epoch": 0.36036036036036034, |
| "grad_norm": 0.2841478766107157, |
| "learning_rate": 5e-06, |
| "loss": 0.3031, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.36936936936936937, |
| "grad_norm": 0.30103555060540843, |
| "learning_rate": 5e-06, |
| "loss": 0.3149, |
| "step": 41 |
| }, |
| { |
| "epoch": 0.3783783783783784, |
| "grad_norm": 0.3245469606074819, |
| "learning_rate": 5e-06, |
| "loss": 0.3147, |
| "step": 42 |
| }, |
| { |
| "epoch": 0.38738738738738737, |
| "grad_norm": 0.32434042643762057, |
| "learning_rate": 5e-06, |
| "loss": 0.3305, |
| "step": 43 |
| }, |
| { |
| "epoch": 0.3963963963963964, |
| "grad_norm": 0.27778799916309627, |
| "learning_rate": 5e-06, |
| "loss": 0.3384, |
| "step": 44 |
| }, |
| { |
| "epoch": 0.40540540540540543, |
| "grad_norm": 0.2745056010877783, |
| "learning_rate": 5e-06, |
| "loss": 0.3156, |
| "step": 45 |
| }, |
| { |
| "epoch": 0.4144144144144144, |
| "grad_norm": 0.29410832050755714, |
| "learning_rate": 5e-06, |
| "loss": 0.3112, |
| "step": 46 |
| }, |
| { |
| "epoch": 0.42342342342342343, |
| "grad_norm": 0.39421096404575884, |
| "learning_rate": 5e-06, |
| "loss": 0.3504, |
| "step": 47 |
| }, |
| { |
| "epoch": 0.43243243243243246, |
| "grad_norm": 0.30297987367745016, |
| "learning_rate": 5e-06, |
| "loss": 0.3133, |
| "step": 48 |
| }, |
| { |
| "epoch": 0.44144144144144143, |
| "grad_norm": 0.312599049596589, |
| "learning_rate": 5e-06, |
| "loss": 0.3298, |
| "step": 49 |
| }, |
| { |
| "epoch": 0.45045045045045046, |
| "grad_norm": 0.28890139188869196, |
| "learning_rate": 5e-06, |
| "loss": 0.3016, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.4594594594594595, |
| "grad_norm": 0.27234641580243496, |
| "learning_rate": 5e-06, |
| "loss": 0.3349, |
| "step": 51 |
| }, |
| { |
| "epoch": 0.46846846846846846, |
| "grad_norm": 0.30882782510454476, |
| "learning_rate": 5e-06, |
| "loss": 0.3307, |
| "step": 52 |
| }, |
| { |
| "epoch": 0.4774774774774775, |
| "grad_norm": 0.2657310651267706, |
| "learning_rate": 5e-06, |
| "loss": 0.3246, |
| "step": 53 |
| }, |
| { |
| "epoch": 0.4864864864864865, |
| "grad_norm": 0.2876695765716273, |
| "learning_rate": 5e-06, |
| "loss": 0.336, |
| "step": 54 |
| }, |
| { |
| "epoch": 0.4954954954954955, |
| "grad_norm": 0.29656571676225046, |
| "learning_rate": 5e-06, |
| "loss": 0.3428, |
| "step": 55 |
| }, |
| { |
| "epoch": 0.5045045045045045, |
| "grad_norm": 0.25789947550982967, |
| "learning_rate": 5e-06, |
| "loss": 0.3035, |
| "step": 56 |
| }, |
| { |
| "epoch": 0.5135135135135135, |
| "grad_norm": 0.3359664317488606, |
| "learning_rate": 5e-06, |
| "loss": 0.3221, |
| "step": 57 |
| }, |
| { |
| "epoch": 0.5225225225225225, |
| "grad_norm": 0.26901646941539337, |
| "learning_rate": 5e-06, |
| "loss": 0.3061, |
| "step": 58 |
| }, |
| { |
| "epoch": 0.5315315315315315, |
| "grad_norm": 0.26500112714488566, |
| "learning_rate": 5e-06, |
| "loss": 0.3203, |
| "step": 59 |
| }, |
| { |
| "epoch": 0.5405405405405406, |
| "grad_norm": 0.2614586643859284, |
| "learning_rate": 5e-06, |
| "loss": 0.3068, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.5495495495495496, |
| "grad_norm": 0.28994271054547277, |
| "learning_rate": 5e-06, |
| "loss": 0.3568, |
| "step": 61 |
| }, |
| { |
| "epoch": 0.5585585585585585, |
| "grad_norm": 0.3372155822417667, |
| "learning_rate": 5e-06, |
| "loss": 0.3559, |
| "step": 62 |
| }, |
| { |
| "epoch": 0.5675675675675675, |
| "grad_norm": 0.30224128387648297, |
| "learning_rate": 5e-06, |
| "loss": 0.3069, |
| "step": 63 |
| }, |
| { |
| "epoch": 0.5765765765765766, |
| "grad_norm": 0.3130672270163632, |
| "learning_rate": 5e-06, |
| "loss": 0.3513, |
| "step": 64 |
| }, |
| { |
| "epoch": 0.5855855855855856, |
| "grad_norm": 0.3065414445284105, |
| "learning_rate": 5e-06, |
| "loss": 0.3194, |
| "step": 65 |
| }, |
| { |
| "epoch": 0.5945945945945946, |
| "grad_norm": 0.29075353592758474, |
| "learning_rate": 5e-06, |
| "loss": 0.364, |
| "step": 66 |
| }, |
| { |
| "epoch": 0.6036036036036037, |
| "grad_norm": 0.28085597006714386, |
| "learning_rate": 5e-06, |
| "loss": 0.3626, |
| "step": 67 |
| }, |
| { |
| "epoch": 0.6126126126126126, |
| "grad_norm": 0.30828909246343983, |
| "learning_rate": 5e-06, |
| "loss": 0.3278, |
| "step": 68 |
| }, |
| { |
| "epoch": 0.6216216216216216, |
| "grad_norm": 0.30901462421223835, |
| "learning_rate": 5e-06, |
| "loss": 0.3338, |
| "step": 69 |
| }, |
| { |
| "epoch": 0.6306306306306306, |
| "grad_norm": 0.3316361212286006, |
| "learning_rate": 5e-06, |
| "loss": 0.3444, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.6396396396396397, |
| "grad_norm": 0.26217545165384337, |
| "learning_rate": 5e-06, |
| "loss": 0.3226, |
| "step": 71 |
| }, |
| { |
| "epoch": 0.6486486486486487, |
| "grad_norm": 0.2563886493400457, |
| "learning_rate": 5e-06, |
| "loss": 0.3409, |
| "step": 72 |
| }, |
| { |
| "epoch": 0.6576576576576577, |
| "grad_norm": 0.2962337946705661, |
| "learning_rate": 5e-06, |
| "loss": 0.3196, |
| "step": 73 |
| }, |
| { |
| "epoch": 0.6666666666666666, |
| "grad_norm": 0.2542256020612804, |
| "learning_rate": 5e-06, |
| "loss": 0.3281, |
| "step": 74 |
| }, |
| { |
| "epoch": 0.6756756756756757, |
| "grad_norm": 0.32938420923589096, |
| "learning_rate": 5e-06, |
| "loss": 0.3185, |
| "step": 75 |
| }, |
| { |
| "epoch": 0.6846846846846847, |
| "grad_norm": 0.33155654515742616, |
| "learning_rate": 5e-06, |
| "loss": 0.3153, |
| "step": 76 |
| }, |
| { |
| "epoch": 0.6936936936936937, |
| "grad_norm": 0.25618184255532905, |
| "learning_rate": 5e-06, |
| "loss": 0.3109, |
| "step": 77 |
| }, |
| { |
| "epoch": 0.7027027027027027, |
| "grad_norm": 0.27091352477500336, |
| "learning_rate": 5e-06, |
| "loss": 0.3371, |
| "step": 78 |
| }, |
| { |
| "epoch": 0.7117117117117117, |
| "grad_norm": 0.270383658268609, |
| "learning_rate": 5e-06, |
| "loss": 0.3325, |
| "step": 79 |
| }, |
| { |
| "epoch": 0.7207207207207207, |
| "grad_norm": 0.2525642887483178, |
| "learning_rate": 5e-06, |
| "loss": 0.3288, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.7297297297297297, |
| "grad_norm": 0.3027811916633369, |
| "learning_rate": 5e-06, |
| "loss": 0.3472, |
| "step": 81 |
| }, |
| { |
| "epoch": 0.7387387387387387, |
| "grad_norm": 0.3506129591935139, |
| "learning_rate": 5e-06, |
| "loss": 0.3383, |
| "step": 82 |
| }, |
| { |
| "epoch": 0.7477477477477478, |
| "grad_norm": 0.32710280320818547, |
| "learning_rate": 5e-06, |
| "loss": 0.3095, |
| "step": 83 |
| }, |
| { |
| "epoch": 0.7567567567567568, |
| "grad_norm": 0.28423266582526613, |
| "learning_rate": 5e-06, |
| "loss": 0.2909, |
| "step": 84 |
| }, |
| { |
| "epoch": 0.7657657657657657, |
| "grad_norm": 0.30514786428468144, |
| "learning_rate": 5e-06, |
| "loss": 0.3181, |
| "step": 85 |
| }, |
| { |
| "epoch": 0.7747747747747747, |
| "grad_norm": 0.3219013280475637, |
| "learning_rate": 5e-06, |
| "loss": 0.3421, |
| "step": 86 |
| }, |
| { |
| "epoch": 0.7837837837837838, |
| "grad_norm": 0.30765019613171724, |
| "learning_rate": 5e-06, |
| "loss": 0.339, |
| "step": 87 |
| }, |
| { |
| "epoch": 0.7927927927927928, |
| "grad_norm": 0.31363666903509363, |
| "learning_rate": 5e-06, |
| "loss": 0.32, |
| "step": 88 |
| }, |
| { |
| "epoch": 0.8018018018018018, |
| "grad_norm": 0.2802553985535834, |
| "learning_rate": 5e-06, |
| "loss": 0.3698, |
| "step": 89 |
| }, |
| { |
| "epoch": 0.8108108108108109, |
| "grad_norm": 0.29928509597489333, |
| "learning_rate": 5e-06, |
| "loss": 0.3465, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.8198198198198198, |
| "grad_norm": 0.30368274583450106, |
| "learning_rate": 5e-06, |
| "loss": 0.321, |
| "step": 91 |
| }, |
| { |
| "epoch": 0.8288288288288288, |
| "grad_norm": 0.28901190479096217, |
| "learning_rate": 5e-06, |
| "loss": 0.3034, |
| "step": 92 |
| }, |
| { |
| "epoch": 0.8378378378378378, |
| "grad_norm": 0.27035852334114224, |
| "learning_rate": 5e-06, |
| "loss": 0.3379, |
| "step": 93 |
| }, |
| { |
| "epoch": 0.8468468468468469, |
| "grad_norm": 0.2757989755002078, |
| "learning_rate": 5e-06, |
| "loss": 0.335, |
| "step": 94 |
| }, |
| { |
| "epoch": 0.8558558558558559, |
| "grad_norm": 0.30063030136785046, |
| "learning_rate": 5e-06, |
| "loss": 0.3299, |
| "step": 95 |
| }, |
| { |
| "epoch": 0.8648648648648649, |
| "grad_norm": 0.3436429105109027, |
| "learning_rate": 5e-06, |
| "loss": 0.2906, |
| "step": 96 |
| }, |
| { |
| "epoch": 0.8738738738738738, |
| "grad_norm": 0.2995609579715489, |
| "learning_rate": 5e-06, |
| "loss": 0.317, |
| "step": 97 |
| }, |
| { |
| "epoch": 0.8828828828828829, |
| "grad_norm": 0.2860865100969785, |
| "learning_rate": 5e-06, |
| "loss": 0.3281, |
| "step": 98 |
| }, |
| { |
| "epoch": 0.8918918918918919, |
| "grad_norm": 0.29202094172851817, |
| "learning_rate": 5e-06, |
| "loss": 0.3225, |
| "step": 99 |
| }, |
| { |
| "epoch": 0.9009009009009009, |
| "grad_norm": 0.2931365896073913, |
| "learning_rate": 5e-06, |
| "loss": 0.3022, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.9099099099099099, |
| "grad_norm": 0.30610410355543166, |
| "learning_rate": 5e-06, |
| "loss": 0.3287, |
| "step": 101 |
| }, |
| { |
| "epoch": 0.918918918918919, |
| "grad_norm": 0.27050744175601266, |
| "learning_rate": 5e-06, |
| "loss": 0.3004, |
| "step": 102 |
| }, |
| { |
| "epoch": 0.9279279279279279, |
| "grad_norm": 0.2530762314307683, |
| "learning_rate": 5e-06, |
| "loss": 0.3153, |
| "step": 103 |
| }, |
| { |
| "epoch": 0.9369369369369369, |
| "grad_norm": 0.2939696187606388, |
| "learning_rate": 5e-06, |
| "loss": 0.3039, |
| "step": 104 |
| }, |
| { |
| "epoch": 0.9459459459459459, |
| "grad_norm": 0.269725936200039, |
| "learning_rate": 5e-06, |
| "loss": 0.3028, |
| "step": 105 |
| }, |
| { |
| "epoch": 0.954954954954955, |
| "grad_norm": 0.32481323273559976, |
| "learning_rate": 5e-06, |
| "loss": 0.3328, |
| "step": 106 |
| }, |
| { |
| "epoch": 0.963963963963964, |
| "grad_norm": 0.3297388133110706, |
| "learning_rate": 5e-06, |
| "loss": 0.3303, |
| "step": 107 |
| }, |
| { |
| "epoch": 0.972972972972973, |
| "grad_norm": 0.3137683488542705, |
| "learning_rate": 5e-06, |
| "loss": 0.3317, |
| "step": 108 |
| }, |
| { |
| "epoch": 0.9819819819819819, |
| "grad_norm": 0.2724212797943338, |
| "learning_rate": 5e-06, |
| "loss": 0.3231, |
| "step": 109 |
| }, |
| { |
| "epoch": 0.990990990990991, |
| "grad_norm": 0.26974252035068974, |
| "learning_rate": 5e-06, |
| "loss": 0.3052, |
| "step": 110 |
| }, |
| { |
| "epoch": 1.0, |
| "grad_norm": 0.27546705234954955, |
| "learning_rate": 5e-06, |
| "loss": 0.2872, |
| "step": 111 |
| }, |
| { |
| "epoch": 1.009009009009009, |
| "grad_norm": 0.26532136740094475, |
| "learning_rate": 5e-06, |
| "loss": 0.2831, |
| "step": 112 |
| }, |
| { |
| "epoch": 1.018018018018018, |
| "grad_norm": 0.2847617719081207, |
| "learning_rate": 5e-06, |
| "loss": 0.3127, |
| "step": 113 |
| }, |
| { |
| "epoch": 1.027027027027027, |
| "grad_norm": 0.25187489870567525, |
| "learning_rate": 5e-06, |
| "loss": 0.3179, |
| "step": 114 |
| }, |
| { |
| "epoch": 1.0360360360360361, |
| "grad_norm": 0.2470210561590589, |
| "learning_rate": 5e-06, |
| "loss": 0.2888, |
| "step": 115 |
| }, |
| { |
| "epoch": 1.045045045045045, |
| "grad_norm": 0.2908873792372198, |
| "learning_rate": 5e-06, |
| "loss": 0.3172, |
| "step": 116 |
| }, |
| { |
| "epoch": 1.054054054054054, |
| "grad_norm": 0.2545755890884819, |
| "learning_rate": 5e-06, |
| "loss": 0.3044, |
| "step": 117 |
| }, |
| { |
| "epoch": 1.063063063063063, |
| "grad_norm": 0.2720375854552878, |
| "learning_rate": 5e-06, |
| "loss": 0.3171, |
| "step": 118 |
| }, |
| { |
| "epoch": 1.072072072072072, |
| "grad_norm": 0.2804009954248822, |
| "learning_rate": 5e-06, |
| "loss": 0.2903, |
| "step": 119 |
| }, |
| { |
| "epoch": 1.0810810810810811, |
| "grad_norm": 0.2584639986814767, |
| "learning_rate": 5e-06, |
| "loss": 0.2786, |
| "step": 120 |
| }, |
| { |
| "epoch": 1.09009009009009, |
| "grad_norm": 0.2523704924311713, |
| "learning_rate": 5e-06, |
| "loss": 0.3009, |
| "step": 121 |
| }, |
| { |
| "epoch": 1.0990990990990992, |
| "grad_norm": 0.27166739983138516, |
| "learning_rate": 5e-06, |
| "loss": 0.3144, |
| "step": 122 |
| }, |
| { |
| "epoch": 1.1081081081081081, |
| "grad_norm": 0.291934322919287, |
| "learning_rate": 5e-06, |
| "loss": 0.3199, |
| "step": 123 |
| }, |
| { |
| "epoch": 1.117117117117117, |
| "grad_norm": 0.2869424658137007, |
| "learning_rate": 5e-06, |
| "loss": 0.2768, |
| "step": 124 |
| }, |
| { |
| "epoch": 1.1261261261261262, |
| "grad_norm": 0.35542461802439873, |
| "learning_rate": 5e-06, |
| "loss": 0.2972, |
| "step": 125 |
| }, |
| { |
| "epoch": 1.135135135135135, |
| "grad_norm": 0.25765779721736737, |
| "learning_rate": 5e-06, |
| "loss": 0.2715, |
| "step": 126 |
| }, |
| { |
| "epoch": 1.1441441441441442, |
| "grad_norm": 0.2850720419420103, |
| "learning_rate": 5e-06, |
| "loss": 0.2861, |
| "step": 127 |
| }, |
| { |
| "epoch": 1.1531531531531531, |
| "grad_norm": 0.2869267701696132, |
| "learning_rate": 5e-06, |
| "loss": 0.2797, |
| "step": 128 |
| }, |
| { |
| "epoch": 1.1621621621621623, |
| "grad_norm": 0.27437916265446266, |
| "learning_rate": 5e-06, |
| "loss": 0.284, |
| "step": 129 |
| }, |
| { |
| "epoch": 1.1711711711711712, |
| "grad_norm": 0.26640743341471523, |
| "learning_rate": 5e-06, |
| "loss": 0.282, |
| "step": 130 |
| }, |
| { |
| "epoch": 1.1801801801801801, |
| "grad_norm": 0.2600732173679119, |
| "learning_rate": 5e-06, |
| "loss": 0.2919, |
| "step": 131 |
| }, |
| { |
| "epoch": 1.1891891891891893, |
| "grad_norm": 0.2665092682109021, |
| "learning_rate": 5e-06, |
| "loss": 0.3143, |
| "step": 132 |
| }, |
| { |
| "epoch": 1.1981981981981982, |
| "grad_norm": 0.24683974895824953, |
| "learning_rate": 5e-06, |
| "loss": 0.2987, |
| "step": 133 |
| }, |
| { |
| "epoch": 1.2072072072072073, |
| "grad_norm": 0.2908036694917544, |
| "learning_rate": 5e-06, |
| "loss": 0.3158, |
| "step": 134 |
| }, |
| { |
| "epoch": 1.2162162162162162, |
| "grad_norm": 0.2945953899064198, |
| "learning_rate": 5e-06, |
| "loss": 0.3152, |
| "step": 135 |
| }, |
| { |
| "epoch": 1.2252252252252251, |
| "grad_norm": 0.2616231868963709, |
| "learning_rate": 5e-06, |
| "loss": 0.3152, |
| "step": 136 |
| }, |
| { |
| "epoch": 1.2342342342342343, |
| "grad_norm": 0.27650089751312973, |
| "learning_rate": 5e-06, |
| "loss": 0.3029, |
| "step": 137 |
| }, |
| { |
| "epoch": 1.2432432432432432, |
| "grad_norm": 0.2631481660529609, |
| "learning_rate": 5e-06, |
| "loss": 0.3084, |
| "step": 138 |
| }, |
| { |
| "epoch": 1.2522522522522523, |
| "grad_norm": 0.28830473220819297, |
| "learning_rate": 5e-06, |
| "loss": 0.3251, |
| "step": 139 |
| }, |
| { |
| "epoch": 1.2612612612612613, |
| "grad_norm": 0.3062303048487267, |
| "learning_rate": 5e-06, |
| "loss": 0.3093, |
| "step": 140 |
| }, |
| { |
| "epoch": 1.2702702702702702, |
| "grad_norm": 0.3066815320224598, |
| "learning_rate": 5e-06, |
| "loss": 0.2881, |
| "step": 141 |
| }, |
| { |
| "epoch": 1.2792792792792793, |
| "grad_norm": 0.29129920876550947, |
| "learning_rate": 5e-06, |
| "loss": 0.282, |
| "step": 142 |
| }, |
| { |
| "epoch": 1.2882882882882882, |
| "grad_norm": 0.2895564905632834, |
| "learning_rate": 5e-06, |
| "loss": 0.3076, |
| "step": 143 |
| }, |
| { |
| "epoch": 1.2972972972972974, |
| "grad_norm": 0.25687914463290057, |
| "learning_rate": 5e-06, |
| "loss": 0.283, |
| "step": 144 |
| }, |
| { |
| "epoch": 1.3063063063063063, |
| "grad_norm": 0.2543976032045274, |
| "learning_rate": 5e-06, |
| "loss": 0.2987, |
| "step": 145 |
| }, |
| { |
| "epoch": 1.3153153153153152, |
| "grad_norm": 0.27423309545031693, |
| "learning_rate": 5e-06, |
| "loss": 0.2981, |
| "step": 146 |
| }, |
| { |
| "epoch": 1.3243243243243243, |
| "grad_norm": 0.3127504643012831, |
| "learning_rate": 5e-06, |
| "loss": 0.3091, |
| "step": 147 |
| }, |
| { |
| "epoch": 1.3333333333333333, |
| "grad_norm": 0.2738777266583336, |
| "learning_rate": 5e-06, |
| "loss": 0.2577, |
| "step": 148 |
| }, |
| { |
| "epoch": 1.3423423423423424, |
| "grad_norm": 0.2669333852747903, |
| "learning_rate": 5e-06, |
| "loss": 0.2855, |
| "step": 149 |
| }, |
| { |
| "epoch": 1.3513513513513513, |
| "grad_norm": 0.26761386479699967, |
| "learning_rate": 5e-06, |
| "loss": 0.3019, |
| "step": 150 |
| }, |
| { |
| "epoch": 1.3603603603603602, |
| "grad_norm": 0.25789802423884284, |
| "learning_rate": 5e-06, |
| "loss": 0.3107, |
| "step": 151 |
| }, |
| { |
| "epoch": 1.3693693693693694, |
| "grad_norm": 0.27940713368034126, |
| "learning_rate": 5e-06, |
| "loss": 0.2843, |
| "step": 152 |
| }, |
| { |
| "epoch": 1.3783783783783785, |
| "grad_norm": 0.277366156708692, |
| "learning_rate": 5e-06, |
| "loss": 0.2799, |
| "step": 153 |
| }, |
| { |
| "epoch": 1.3873873873873874, |
| "grad_norm": 0.2607843059312788, |
| "learning_rate": 5e-06, |
| "loss": 0.292, |
| "step": 154 |
| }, |
| { |
| "epoch": 1.3963963963963963, |
| "grad_norm": 0.2649281612489507, |
| "learning_rate": 5e-06, |
| "loss": 0.3134, |
| "step": 155 |
| }, |
| { |
| "epoch": 1.4054054054054055, |
| "grad_norm": 0.27271972468771527, |
| "learning_rate": 5e-06, |
| "loss": 0.2972, |
| "step": 156 |
| }, |
| { |
| "epoch": 1.4144144144144144, |
| "grad_norm": 0.26207901754212165, |
| "learning_rate": 5e-06, |
| "loss": 0.2822, |
| "step": 157 |
| }, |
| { |
| "epoch": 1.4234234234234235, |
| "grad_norm": 0.2641717963089793, |
| "learning_rate": 5e-06, |
| "loss": 0.2971, |
| "step": 158 |
| }, |
| { |
| "epoch": 1.4324324324324325, |
| "grad_norm": 0.2579842614638958, |
| "learning_rate": 5e-06, |
| "loss": 0.3024, |
| "step": 159 |
| }, |
| { |
| "epoch": 1.4414414414414414, |
| "grad_norm": 0.2870255938899811, |
| "learning_rate": 5e-06, |
| "loss": 0.2885, |
| "step": 160 |
| }, |
| { |
| "epoch": 1.4504504504504505, |
| "grad_norm": 0.2777224839264993, |
| "learning_rate": 5e-06, |
| "loss": 0.2892, |
| "step": 161 |
| }, |
| { |
| "epoch": 1.4594594594594594, |
| "grad_norm": 0.27625106290913043, |
| "learning_rate": 5e-06, |
| "loss": 0.2805, |
| "step": 162 |
| }, |
| { |
| "epoch": 1.4684684684684686, |
| "grad_norm": 0.2700016737510603, |
| "learning_rate": 5e-06, |
| "loss": 0.2992, |
| "step": 163 |
| }, |
| { |
| "epoch": 1.4774774774774775, |
| "grad_norm": 0.25372514988722056, |
| "learning_rate": 5e-06, |
| "loss": 0.2972, |
| "step": 164 |
| }, |
| { |
| "epoch": 1.4864864864864864, |
| "grad_norm": 0.28782834487825465, |
| "learning_rate": 5e-06, |
| "loss": 0.3018, |
| "step": 165 |
| }, |
| { |
| "epoch": 1.4954954954954955, |
| "grad_norm": 0.27036226357763354, |
| "learning_rate": 5e-06, |
| "loss": 0.2968, |
| "step": 166 |
| }, |
| { |
| "epoch": 1.5045045045045045, |
| "grad_norm": 0.24997568182394178, |
| "learning_rate": 5e-06, |
| "loss": 0.2482, |
| "step": 167 |
| }, |
| { |
| "epoch": 1.5135135135135136, |
| "grad_norm": 0.28025540658752757, |
| "learning_rate": 5e-06, |
| "loss": 0.3314, |
| "step": 168 |
| }, |
| { |
| "epoch": 1.5225225225225225, |
| "grad_norm": 0.25563343479526396, |
| "learning_rate": 5e-06, |
| "loss": 0.3163, |
| "step": 169 |
| }, |
| { |
| "epoch": 1.5315315315315314, |
| "grad_norm": 0.3556162754506623, |
| "learning_rate": 5e-06, |
| "loss": 0.2925, |
| "step": 170 |
| }, |
| { |
| "epoch": 1.5405405405405406, |
| "grad_norm": 0.27599016482238853, |
| "learning_rate": 5e-06, |
| "loss": 0.2838, |
| "step": 171 |
| }, |
| { |
| "epoch": 1.5495495495495497, |
| "grad_norm": 0.272343971725021, |
| "learning_rate": 5e-06, |
| "loss": 0.3088, |
| "step": 172 |
| }, |
| { |
| "epoch": 1.5585585585585586, |
| "grad_norm": 0.28693003610171597, |
| "learning_rate": 5e-06, |
| "loss": 0.2921, |
| "step": 173 |
| }, |
| { |
| "epoch": 1.5675675675675675, |
| "grad_norm": 0.2955327518594707, |
| "learning_rate": 5e-06, |
| "loss": 0.2777, |
| "step": 174 |
| }, |
| { |
| "epoch": 1.5765765765765765, |
| "grad_norm": 0.27961760151449894, |
| "learning_rate": 5e-06, |
| "loss": 0.2838, |
| "step": 175 |
| }, |
| { |
| "epoch": 1.5855855855855856, |
| "grad_norm": 0.24665431850909808, |
| "learning_rate": 5e-06, |
| "loss": 0.2781, |
| "step": 176 |
| }, |
| { |
| "epoch": 1.5945945945945947, |
| "grad_norm": 0.26426261640553667, |
| "learning_rate": 5e-06, |
| "loss": 0.2816, |
| "step": 177 |
| }, |
| { |
| "epoch": 1.6036036036036037, |
| "grad_norm": 0.2711333903704824, |
| "learning_rate": 5e-06, |
| "loss": 0.3142, |
| "step": 178 |
| }, |
| { |
| "epoch": 1.6126126126126126, |
| "grad_norm": 0.2722379287245898, |
| "learning_rate": 5e-06, |
| "loss": 0.2816, |
| "step": 179 |
| }, |
| { |
| "epoch": 1.6216216216216215, |
| "grad_norm": 0.3012330875667607, |
| "learning_rate": 5e-06, |
| "loss": 0.3263, |
| "step": 180 |
| }, |
| { |
| "epoch": 1.6306306306306306, |
| "grad_norm": 0.2669108739090265, |
| "learning_rate": 5e-06, |
| "loss": 0.2952, |
| "step": 181 |
| }, |
| { |
| "epoch": 1.6396396396396398, |
| "grad_norm": 0.2748579289599078, |
| "learning_rate": 5e-06, |
| "loss": 0.2823, |
| "step": 182 |
| }, |
| { |
| "epoch": 1.6486486486486487, |
| "grad_norm": 0.29837425745633833, |
| "learning_rate": 5e-06, |
| "loss": 0.3038, |
| "step": 183 |
| }, |
| { |
| "epoch": 1.6576576576576576, |
| "grad_norm": 0.3305979404285009, |
| "learning_rate": 5e-06, |
| "loss": 0.3017, |
| "step": 184 |
| }, |
| { |
| "epoch": 1.6666666666666665, |
| "grad_norm": 0.26365462645864157, |
| "learning_rate": 5e-06, |
| "loss": 0.2952, |
| "step": 185 |
| }, |
| { |
| "epoch": 1.6756756756756757, |
| "grad_norm": 0.27117354048602127, |
| "learning_rate": 5e-06, |
| "loss": 0.2713, |
| "step": 186 |
| }, |
| { |
| "epoch": 1.6846846846846848, |
| "grad_norm": 0.2618109082938301, |
| "learning_rate": 5e-06, |
| "loss": 0.3259, |
| "step": 187 |
| }, |
| { |
| "epoch": 1.6936936936936937, |
| "grad_norm": 0.24890174240606217, |
| "learning_rate": 5e-06, |
| "loss": 0.2864, |
| "step": 188 |
| }, |
| { |
| "epoch": 1.7027027027027026, |
| "grad_norm": 0.28948368345439884, |
| "learning_rate": 5e-06, |
| "loss": 0.3242, |
| "step": 189 |
| }, |
| { |
| "epoch": 1.7117117117117115, |
| "grad_norm": 0.2659473815766033, |
| "learning_rate": 5e-06, |
| "loss": 0.2928, |
| "step": 190 |
| }, |
| { |
| "epoch": 1.7207207207207207, |
| "grad_norm": 0.26435921312812555, |
| "learning_rate": 5e-06, |
| "loss": 0.2747, |
| "step": 191 |
| }, |
| { |
| "epoch": 1.7297297297297298, |
| "grad_norm": 0.2834566804404197, |
| "learning_rate": 5e-06, |
| "loss": 0.3373, |
| "step": 192 |
| }, |
| { |
| "epoch": 1.7387387387387387, |
| "grad_norm": 0.26226690378932954, |
| "learning_rate": 5e-06, |
| "loss": 0.2713, |
| "step": 193 |
| }, |
| { |
| "epoch": 1.7477477477477477, |
| "grad_norm": 0.2574908549961044, |
| "learning_rate": 5e-06, |
| "loss": 0.2949, |
| "step": 194 |
| }, |
| { |
| "epoch": 1.7567567567567568, |
| "grad_norm": 0.2670216430713444, |
| "learning_rate": 5e-06, |
| "loss": 0.3105, |
| "step": 195 |
| }, |
| { |
| "epoch": 1.7657657657657657, |
| "grad_norm": 0.2644549565961117, |
| "learning_rate": 5e-06, |
| "loss": 0.2974, |
| "step": 196 |
| }, |
| { |
| "epoch": 1.7747747747747749, |
| "grad_norm": 0.2754975911578592, |
| "learning_rate": 5e-06, |
| "loss": 0.3019, |
| "step": 197 |
| }, |
| { |
| "epoch": 1.7837837837837838, |
| "grad_norm": 0.28446592391114817, |
| "learning_rate": 5e-06, |
| "loss": 0.3148, |
| "step": 198 |
| }, |
| { |
| "epoch": 1.7927927927927927, |
| "grad_norm": 0.28893386362511947, |
| "learning_rate": 5e-06, |
| "loss": 0.3189, |
| "step": 199 |
| }, |
| { |
| "epoch": 1.8018018018018018, |
| "grad_norm": 0.2869246676669029, |
| "learning_rate": 5e-06, |
| "loss": 0.3015, |
| "step": 200 |
| }, |
| { |
| "epoch": 1.810810810810811, |
| "grad_norm": 0.2847178633594474, |
| "learning_rate": 5e-06, |
| "loss": 0.2628, |
| "step": 201 |
| }, |
| { |
| "epoch": 1.8198198198198199, |
| "grad_norm": 0.2946725850660284, |
| "learning_rate": 5e-06, |
| "loss": 0.2768, |
| "step": 202 |
| }, |
| { |
| "epoch": 1.8288288288288288, |
| "grad_norm": 0.29608299277278, |
| "learning_rate": 5e-06, |
| "loss": 0.3171, |
| "step": 203 |
| }, |
| { |
| "epoch": 1.8378378378378377, |
| "grad_norm": 0.28628382246998885, |
| "learning_rate": 5e-06, |
| "loss": 0.3096, |
| "step": 204 |
| }, |
| { |
| "epoch": 1.8468468468468469, |
| "grad_norm": 0.2660371973699119, |
| "learning_rate": 5e-06, |
| "loss": 0.2685, |
| "step": 205 |
| }, |
| { |
| "epoch": 1.855855855855856, |
| "grad_norm": 0.2514264016055165, |
| "learning_rate": 5e-06, |
| "loss": 0.2622, |
| "step": 206 |
| }, |
| { |
| "epoch": 1.864864864864865, |
| "grad_norm": 0.2675623714158383, |
| "learning_rate": 5e-06, |
| "loss": 0.3324, |
| "step": 207 |
| }, |
| { |
| "epoch": 1.8738738738738738, |
| "grad_norm": 0.2817065371989752, |
| "learning_rate": 5e-06, |
| "loss": 0.2926, |
| "step": 208 |
| }, |
| { |
| "epoch": 1.8828828828828827, |
| "grad_norm": 0.24376840027264843, |
| "learning_rate": 5e-06, |
| "loss": 0.2695, |
| "step": 209 |
| }, |
| { |
| "epoch": 1.8918918918918919, |
| "grad_norm": 0.2679237524654036, |
| "learning_rate": 5e-06, |
| "loss": 0.2606, |
| "step": 210 |
| }, |
| { |
| "epoch": 1.900900900900901, |
| "grad_norm": 0.2593077892544588, |
| "learning_rate": 5e-06, |
| "loss": 0.2781, |
| "step": 211 |
| }, |
| { |
| "epoch": 1.90990990990991, |
| "grad_norm": 0.2555343741606999, |
| "learning_rate": 5e-06, |
| "loss": 0.2616, |
| "step": 212 |
| }, |
| { |
| "epoch": 1.9189189189189189, |
| "grad_norm": 0.27065363914180135, |
| "learning_rate": 5e-06, |
| "loss": 0.309, |
| "step": 213 |
| }, |
| { |
| "epoch": 1.9279279279279278, |
| "grad_norm": 0.29950662348843465, |
| "learning_rate": 5e-06, |
| "loss": 0.2953, |
| "step": 214 |
| }, |
| { |
| "epoch": 1.936936936936937, |
| "grad_norm": 0.30392398016557, |
| "learning_rate": 5e-06, |
| "loss": 0.3302, |
| "step": 215 |
| }, |
| { |
| "epoch": 1.945945945945946, |
| "grad_norm": 0.2688781676455933, |
| "learning_rate": 5e-06, |
| "loss": 0.2946, |
| "step": 216 |
| }, |
| { |
| "epoch": 1.954954954954955, |
| "grad_norm": 0.27334249580678227, |
| "learning_rate": 5e-06, |
| "loss": 0.3169, |
| "step": 217 |
| }, |
| { |
| "epoch": 1.9639639639639639, |
| "grad_norm": 0.2637661232011851, |
| "learning_rate": 5e-06, |
| "loss": 0.2923, |
| "step": 218 |
| }, |
| { |
| "epoch": 1.972972972972973, |
| "grad_norm": 0.24845919128888916, |
| "learning_rate": 5e-06, |
| "loss": 0.2956, |
| "step": 219 |
| }, |
| { |
| "epoch": 1.981981981981982, |
| "grad_norm": 0.2677476120892863, |
| "learning_rate": 5e-06, |
| "loss": 0.2725, |
| "step": 220 |
| }, |
| { |
| "epoch": 1.990990990990991, |
| "grad_norm": 0.27245457118100547, |
| "learning_rate": 5e-06, |
| "loss": 0.304, |
| "step": 221 |
| }, |
| { |
| "epoch": 2.0, |
| "grad_norm": 0.2632364290696338, |
| "learning_rate": 5e-06, |
| "loss": 0.2759, |
| "step": 222 |
| }, |
| { |
| "epoch": 2.009009009009009, |
| "grad_norm": 0.29524131111947416, |
| "learning_rate": 5e-06, |
| "loss": 0.2467, |
| "step": 223 |
| }, |
| { |
| "epoch": 2.018018018018018, |
| "grad_norm": 0.26959444826517864, |
| "learning_rate": 5e-06, |
| "loss": 0.2509, |
| "step": 224 |
| }, |
| { |
| "epoch": 2.027027027027027, |
| "grad_norm": 0.24776989679141162, |
| "learning_rate": 5e-06, |
| "loss": 0.2647, |
| "step": 225 |
| }, |
| { |
| "epoch": 2.036036036036036, |
| "grad_norm": 0.24922491602278132, |
| "learning_rate": 5e-06, |
| "loss": 0.2734, |
| "step": 226 |
| }, |
| { |
| "epoch": 2.045045045045045, |
| "grad_norm": 0.2637011140567836, |
| "learning_rate": 5e-06, |
| "loss": 0.2538, |
| "step": 227 |
| }, |
| { |
| "epoch": 2.054054054054054, |
| "grad_norm": 0.24677968833597697, |
| "learning_rate": 5e-06, |
| "loss": 0.2569, |
| "step": 228 |
| }, |
| { |
| "epoch": 2.063063063063063, |
| "grad_norm": 0.25749179244984177, |
| "learning_rate": 5e-06, |
| "loss": 0.266, |
| "step": 229 |
| }, |
| { |
| "epoch": 2.0720720720720722, |
| "grad_norm": 0.2704364348984915, |
| "learning_rate": 5e-06, |
| "loss": 0.2645, |
| "step": 230 |
| }, |
| { |
| "epoch": 2.081081081081081, |
| "grad_norm": 0.2848341811917101, |
| "learning_rate": 5e-06, |
| "loss": 0.258, |
| "step": 231 |
| }, |
| { |
| "epoch": 2.09009009009009, |
| "grad_norm": 0.2539455237645273, |
| "learning_rate": 5e-06, |
| "loss": 0.2648, |
| "step": 232 |
| }, |
| { |
| "epoch": 2.099099099099099, |
| "grad_norm": 0.2534894136461773, |
| "learning_rate": 5e-06, |
| "loss": 0.2611, |
| "step": 233 |
| }, |
| { |
| "epoch": 2.108108108108108, |
| "grad_norm": 0.2666435185167066, |
| "learning_rate": 5e-06, |
| "loss": 0.2671, |
| "step": 234 |
| }, |
| { |
| "epoch": 2.1171171171171173, |
| "grad_norm": 0.275032039682747, |
| "learning_rate": 5e-06, |
| "loss": 0.2807, |
| "step": 235 |
| }, |
| { |
| "epoch": 2.126126126126126, |
| "grad_norm": 0.24537895004936466, |
| "learning_rate": 5e-06, |
| "loss": 0.2777, |
| "step": 236 |
| }, |
| { |
| "epoch": 2.135135135135135, |
| "grad_norm": 0.29459998669694115, |
| "learning_rate": 5e-06, |
| "loss": 0.2782, |
| "step": 237 |
| }, |
| { |
| "epoch": 2.144144144144144, |
| "grad_norm": 0.2727554788191977, |
| "learning_rate": 5e-06, |
| "loss": 0.2687, |
| "step": 238 |
| }, |
| { |
| "epoch": 2.153153153153153, |
| "grad_norm": 0.30880501847599995, |
| "learning_rate": 5e-06, |
| "loss": 0.2878, |
| "step": 239 |
| }, |
| { |
| "epoch": 2.1621621621621623, |
| "grad_norm": 0.2886633976684916, |
| "learning_rate": 5e-06, |
| "loss": 0.267, |
| "step": 240 |
| }, |
| { |
| "epoch": 2.171171171171171, |
| "grad_norm": 0.2597628174067978, |
| "learning_rate": 5e-06, |
| "loss": 0.2648, |
| "step": 241 |
| }, |
| { |
| "epoch": 2.18018018018018, |
| "grad_norm": 0.2534324931692372, |
| "learning_rate": 5e-06, |
| "loss": 0.2981, |
| "step": 242 |
| }, |
| { |
| "epoch": 2.189189189189189, |
| "grad_norm": 0.2563993838591747, |
| "learning_rate": 5e-06, |
| "loss": 0.2487, |
| "step": 243 |
| }, |
| { |
| "epoch": 2.1981981981981984, |
| "grad_norm": 0.2852726219398302, |
| "learning_rate": 5e-06, |
| "loss": 0.2543, |
| "step": 244 |
| }, |
| { |
| "epoch": 2.2072072072072073, |
| "grad_norm": 0.30478195170068134, |
| "learning_rate": 5e-06, |
| "loss": 0.2562, |
| "step": 245 |
| }, |
| { |
| "epoch": 2.2162162162162162, |
| "grad_norm": 0.24772685929517294, |
| "learning_rate": 5e-06, |
| "loss": 0.2869, |
| "step": 246 |
| }, |
| { |
| "epoch": 2.225225225225225, |
| "grad_norm": 0.26428977786941277, |
| "learning_rate": 5e-06, |
| "loss": 0.2709, |
| "step": 247 |
| }, |
| { |
| "epoch": 2.234234234234234, |
| "grad_norm": 0.2447098843485426, |
| "learning_rate": 5e-06, |
| "loss": 0.2241, |
| "step": 248 |
| }, |
| { |
| "epoch": 2.2432432432432434, |
| "grad_norm": 0.2841804786817898, |
| "learning_rate": 5e-06, |
| "loss": 0.2398, |
| "step": 249 |
| }, |
| { |
| "epoch": 2.2522522522522523, |
| "grad_norm": 0.2837413945636495, |
| "learning_rate": 5e-06, |
| "loss": 0.2755, |
| "step": 250 |
| }, |
| { |
| "epoch": 2.2612612612612613, |
| "grad_norm": 0.27688677145182117, |
| "learning_rate": 5e-06, |
| "loss": 0.2581, |
| "step": 251 |
| }, |
| { |
| "epoch": 2.27027027027027, |
| "grad_norm": 0.2524013812037196, |
| "learning_rate": 5e-06, |
| "loss": 0.2447, |
| "step": 252 |
| }, |
| { |
| "epoch": 2.279279279279279, |
| "grad_norm": 0.25708866849265744, |
| "learning_rate": 5e-06, |
| "loss": 0.2629, |
| "step": 253 |
| }, |
| { |
| "epoch": 2.2882882882882885, |
| "grad_norm": 0.31089756790372536, |
| "learning_rate": 5e-06, |
| "loss": 0.262, |
| "step": 254 |
| }, |
| { |
| "epoch": 2.2972972972972974, |
| "grad_norm": 0.2580437334513352, |
| "learning_rate": 5e-06, |
| "loss": 0.2672, |
| "step": 255 |
| }, |
| { |
| "epoch": 2.3063063063063063, |
| "grad_norm": 0.25589033140205797, |
| "learning_rate": 5e-06, |
| "loss": 0.2854, |
| "step": 256 |
| }, |
| { |
| "epoch": 2.315315315315315, |
| "grad_norm": 0.2851188761111017, |
| "learning_rate": 5e-06, |
| "loss": 0.2847, |
| "step": 257 |
| }, |
| { |
| "epoch": 2.3243243243243246, |
| "grad_norm": 0.2742352435214708, |
| "learning_rate": 5e-06, |
| "loss": 0.2863, |
| "step": 258 |
| }, |
| { |
| "epoch": 2.3333333333333335, |
| "grad_norm": 0.25574343614682743, |
| "learning_rate": 5e-06, |
| "loss": 0.2744, |
| "step": 259 |
| }, |
| { |
| "epoch": 2.3423423423423424, |
| "grad_norm": 0.2704501372387818, |
| "learning_rate": 5e-06, |
| "loss": 0.2356, |
| "step": 260 |
| }, |
| { |
| "epoch": 2.3513513513513513, |
| "grad_norm": 0.2694883625074875, |
| "learning_rate": 5e-06, |
| "loss": 0.2845, |
| "step": 261 |
| }, |
| { |
| "epoch": 2.3603603603603602, |
| "grad_norm": 0.2749897171746042, |
| "learning_rate": 5e-06, |
| "loss": 0.2745, |
| "step": 262 |
| }, |
| { |
| "epoch": 2.3693693693693696, |
| "grad_norm": 0.33678826387641014, |
| "learning_rate": 5e-06, |
| "loss": 0.3088, |
| "step": 263 |
| }, |
| { |
| "epoch": 2.3783783783783785, |
| "grad_norm": 0.2773165283746789, |
| "learning_rate": 5e-06, |
| "loss": 0.2946, |
| "step": 264 |
| }, |
| { |
| "epoch": 2.3873873873873874, |
| "grad_norm": 0.31677913584086903, |
| "learning_rate": 5e-06, |
| "loss": 0.3179, |
| "step": 265 |
| }, |
| { |
| "epoch": 2.3963963963963963, |
| "grad_norm": 0.2563051452749462, |
| "learning_rate": 5e-06, |
| "loss": 0.2625, |
| "step": 266 |
| }, |
| { |
| "epoch": 2.4054054054054053, |
| "grad_norm": 0.321688693489085, |
| "learning_rate": 5e-06, |
| "loss": 0.2671, |
| "step": 267 |
| }, |
| { |
| "epoch": 2.4144144144144146, |
| "grad_norm": 0.26634437339972133, |
| "learning_rate": 5e-06, |
| "loss": 0.2613, |
| "step": 268 |
| }, |
| { |
| "epoch": 2.4234234234234235, |
| "grad_norm": 0.27171211584580457, |
| "learning_rate": 5e-06, |
| "loss": 0.2628, |
| "step": 269 |
| }, |
| { |
| "epoch": 2.4324324324324325, |
| "grad_norm": 0.2555430715005437, |
| "learning_rate": 5e-06, |
| "loss": 0.2687, |
| "step": 270 |
| }, |
| { |
| "epoch": 2.4414414414414414, |
| "grad_norm": 0.24255848197003171, |
| "learning_rate": 5e-06, |
| "loss": 0.2941, |
| "step": 271 |
| }, |
| { |
| "epoch": 2.4504504504504503, |
| "grad_norm": 0.29538238957980967, |
| "learning_rate": 5e-06, |
| "loss": 0.2777, |
| "step": 272 |
| }, |
| { |
| "epoch": 2.4594594594594597, |
| "grad_norm": 0.2876545631402078, |
| "learning_rate": 5e-06, |
| "loss": 0.2764, |
| "step": 273 |
| }, |
| { |
| "epoch": 2.4684684684684686, |
| "grad_norm": 0.2773762933353327, |
| "learning_rate": 5e-06, |
| "loss": 0.2834, |
| "step": 274 |
| }, |
| { |
| "epoch": 2.4774774774774775, |
| "grad_norm": 0.25275190194965114, |
| "learning_rate": 5e-06, |
| "loss": 0.2625, |
| "step": 275 |
| }, |
| { |
| "epoch": 2.4864864864864864, |
| "grad_norm": 0.30548139692249815, |
| "learning_rate": 5e-06, |
| "loss": 0.264, |
| "step": 276 |
| }, |
| { |
| "epoch": 2.4954954954954953, |
| "grad_norm": 0.2857116539220258, |
| "learning_rate": 5e-06, |
| "loss": 0.2663, |
| "step": 277 |
| }, |
| { |
| "epoch": 2.5045045045045047, |
| "grad_norm": 0.27127459034653845, |
| "learning_rate": 5e-06, |
| "loss": 0.289, |
| "step": 278 |
| }, |
| { |
| "epoch": 2.5135135135135136, |
| "grad_norm": 0.29403524162565264, |
| "learning_rate": 5e-06, |
| "loss": 0.2665, |
| "step": 279 |
| }, |
| { |
| "epoch": 2.5225225225225225, |
| "grad_norm": 0.2982604039719257, |
| "learning_rate": 5e-06, |
| "loss": 0.2635, |
| "step": 280 |
| }, |
| { |
| "epoch": 2.5315315315315314, |
| "grad_norm": 0.25776587175299304, |
| "learning_rate": 5e-06, |
| "loss": 0.2592, |
| "step": 281 |
| }, |
| { |
| "epoch": 2.5405405405405403, |
| "grad_norm": 0.2646598986862087, |
| "learning_rate": 5e-06, |
| "loss": 0.2579, |
| "step": 282 |
| }, |
| { |
| "epoch": 2.5495495495495497, |
| "grad_norm": 0.24717949544087905, |
| "learning_rate": 5e-06, |
| "loss": 0.2544, |
| "step": 283 |
| }, |
| { |
| "epoch": 2.5585585585585586, |
| "grad_norm": 0.2657887766041429, |
| "learning_rate": 5e-06, |
| "loss": 0.2656, |
| "step": 284 |
| }, |
| { |
| "epoch": 2.5675675675675675, |
| "grad_norm": 0.27748457946008864, |
| "learning_rate": 5e-06, |
| "loss": 0.2675, |
| "step": 285 |
| }, |
| { |
| "epoch": 2.5765765765765765, |
| "grad_norm": 0.25089374600320746, |
| "learning_rate": 5e-06, |
| "loss": 0.2816, |
| "step": 286 |
| }, |
| { |
| "epoch": 2.5855855855855854, |
| "grad_norm": 0.28897866413916584, |
| "learning_rate": 5e-06, |
| "loss": 0.2589, |
| "step": 287 |
| }, |
| { |
| "epoch": 2.5945945945945947, |
| "grad_norm": 0.26235423487495346, |
| "learning_rate": 5e-06, |
| "loss": 0.271, |
| "step": 288 |
| }, |
| { |
| "epoch": 2.6036036036036037, |
| "grad_norm": 0.29773828111895406, |
| "learning_rate": 5e-06, |
| "loss": 0.2884, |
| "step": 289 |
| }, |
| { |
| "epoch": 2.6126126126126126, |
| "grad_norm": 0.2732062490555635, |
| "learning_rate": 5e-06, |
| "loss": 0.297, |
| "step": 290 |
| }, |
| { |
| "epoch": 2.6216216216216215, |
| "grad_norm": 0.28269145506341653, |
| "learning_rate": 5e-06, |
| "loss": 0.2794, |
| "step": 291 |
| }, |
| { |
| "epoch": 2.6306306306306304, |
| "grad_norm": 0.2592351362804753, |
| "learning_rate": 5e-06, |
| "loss": 0.2653, |
| "step": 292 |
| }, |
| { |
| "epoch": 2.6396396396396398, |
| "grad_norm": 0.27363184791488976, |
| "learning_rate": 5e-06, |
| "loss": 0.2659, |
| "step": 293 |
| }, |
| { |
| "epoch": 2.6486486486486487, |
| "grad_norm": 0.2687283362268144, |
| "learning_rate": 5e-06, |
| "loss": 0.2881, |
| "step": 294 |
| }, |
| { |
| "epoch": 2.6576576576576576, |
| "grad_norm": 0.2669999794761192, |
| "learning_rate": 5e-06, |
| "loss": 0.2658, |
| "step": 295 |
| }, |
| { |
| "epoch": 2.6666666666666665, |
| "grad_norm": 0.2584313873251436, |
| "learning_rate": 5e-06, |
| "loss": 0.2597, |
| "step": 296 |
| }, |
| { |
| "epoch": 2.6756756756756754, |
| "grad_norm": 0.2870412914872632, |
| "learning_rate": 5e-06, |
| "loss": 0.273, |
| "step": 297 |
| }, |
| { |
| "epoch": 2.684684684684685, |
| "grad_norm": 0.2565405158611234, |
| "learning_rate": 5e-06, |
| "loss": 0.2471, |
| "step": 298 |
| }, |
| { |
| "epoch": 2.6936936936936937, |
| "grad_norm": 0.2718920473228364, |
| "learning_rate": 5e-06, |
| "loss": 0.2556, |
| "step": 299 |
| }, |
| { |
| "epoch": 2.7027027027027026, |
| "grad_norm": 0.2732398668856954, |
| "learning_rate": 5e-06, |
| "loss": 0.2729, |
| "step": 300 |
| }, |
| { |
| "epoch": 2.7117117117117115, |
| "grad_norm": 0.25213076888264274, |
| "learning_rate": 5e-06, |
| "loss": 0.2659, |
| "step": 301 |
| }, |
| { |
| "epoch": 2.7207207207207205, |
| "grad_norm": 0.25342262780535696, |
| "learning_rate": 5e-06, |
| "loss": 0.2791, |
| "step": 302 |
| }, |
| { |
| "epoch": 2.72972972972973, |
| "grad_norm": 0.2929513672092119, |
| "learning_rate": 5e-06, |
| "loss": 0.2714, |
| "step": 303 |
| }, |
| { |
| "epoch": 2.7387387387387387, |
| "grad_norm": 0.27482309634629043, |
| "learning_rate": 5e-06, |
| "loss": 0.2646, |
| "step": 304 |
| }, |
| { |
| "epoch": 2.7477477477477477, |
| "grad_norm": 0.26495695016553, |
| "learning_rate": 5e-06, |
| "loss": 0.2655, |
| "step": 305 |
| }, |
| { |
| "epoch": 2.756756756756757, |
| "grad_norm": 0.2751450071843517, |
| "learning_rate": 5e-06, |
| "loss": 0.282, |
| "step": 306 |
| }, |
| { |
| "epoch": 2.7657657657657655, |
| "grad_norm": 0.2492074837362159, |
| "learning_rate": 5e-06, |
| "loss": 0.2728, |
| "step": 307 |
| }, |
| { |
| "epoch": 2.774774774774775, |
| "grad_norm": 0.24588259514355568, |
| "learning_rate": 5e-06, |
| "loss": 0.2506, |
| "step": 308 |
| }, |
| { |
| "epoch": 2.7837837837837838, |
| "grad_norm": 0.290865691950273, |
| "learning_rate": 5e-06, |
| "loss": 0.3019, |
| "step": 309 |
| }, |
| { |
| "epoch": 2.7927927927927927, |
| "grad_norm": 0.24649105252907824, |
| "learning_rate": 5e-06, |
| "loss": 0.3028, |
| "step": 310 |
| }, |
| { |
| "epoch": 2.801801801801802, |
| "grad_norm": 0.24865219694730992, |
| "learning_rate": 5e-06, |
| "loss": 0.2647, |
| "step": 311 |
| }, |
| { |
| "epoch": 2.810810810810811, |
| "grad_norm": 0.2641273618850612, |
| "learning_rate": 5e-06, |
| "loss": 0.2743, |
| "step": 312 |
| }, |
| { |
| "epoch": 2.81981981981982, |
| "grad_norm": 0.27036448999028867, |
| "learning_rate": 5e-06, |
| "loss": 0.2483, |
| "step": 313 |
| }, |
| { |
| "epoch": 2.828828828828829, |
| "grad_norm": 0.277820288498933, |
| "learning_rate": 5e-06, |
| "loss": 0.2478, |
| "step": 314 |
| }, |
| { |
| "epoch": 2.8378378378378377, |
| "grad_norm": 0.25834412495274456, |
| "learning_rate": 5e-06, |
| "loss": 0.2395, |
| "step": 315 |
| }, |
| { |
| "epoch": 2.846846846846847, |
| "grad_norm": 0.25827263911198917, |
| "learning_rate": 5e-06, |
| "loss": 0.222, |
| "step": 316 |
| }, |
| { |
| "epoch": 2.855855855855856, |
| "grad_norm": 0.28475747286608616, |
| "learning_rate": 5e-06, |
| "loss": 0.2747, |
| "step": 317 |
| }, |
| { |
| "epoch": 2.864864864864865, |
| "grad_norm": 0.25037323222188695, |
| "learning_rate": 5e-06, |
| "loss": 0.2689, |
| "step": 318 |
| }, |
| { |
| "epoch": 2.873873873873874, |
| "grad_norm": 0.2652972773806203, |
| "learning_rate": 5e-06, |
| "loss": 0.2477, |
| "step": 319 |
| }, |
| { |
| "epoch": 2.8828828828828827, |
| "grad_norm": 0.26279014403702605, |
| "learning_rate": 5e-06, |
| "loss": 0.2734, |
| "step": 320 |
| }, |
| { |
| "epoch": 2.891891891891892, |
| "grad_norm": 0.2854042712916503, |
| "learning_rate": 5e-06, |
| "loss": 0.2679, |
| "step": 321 |
| }, |
| { |
| "epoch": 2.900900900900901, |
| "grad_norm": 0.26077805779165003, |
| "learning_rate": 5e-06, |
| "loss": 0.2608, |
| "step": 322 |
| }, |
| { |
| "epoch": 2.90990990990991, |
| "grad_norm": 0.255112093170312, |
| "learning_rate": 5e-06, |
| "loss": 0.2695, |
| "step": 323 |
| }, |
| { |
| "epoch": 2.918918918918919, |
| "grad_norm": 0.26211588620202336, |
| "learning_rate": 5e-06, |
| "loss": 0.2424, |
| "step": 324 |
| }, |
| { |
| "epoch": 2.9279279279279278, |
| "grad_norm": 0.2685084403266774, |
| "learning_rate": 5e-06, |
| "loss": 0.235, |
| "step": 325 |
| }, |
| { |
| "epoch": 2.936936936936937, |
| "grad_norm": 0.27269803144536753, |
| "learning_rate": 5e-06, |
| "loss": 0.2759, |
| "step": 326 |
| }, |
| { |
| "epoch": 2.945945945945946, |
| "grad_norm": 0.26751393672770624, |
| "learning_rate": 5e-06, |
| "loss": 0.2564, |
| "step": 327 |
| }, |
| { |
| "epoch": 2.954954954954955, |
| "grad_norm": 0.2665543902683488, |
| "learning_rate": 5e-06, |
| "loss": 0.2763, |
| "step": 328 |
| }, |
| { |
| "epoch": 2.963963963963964, |
| "grad_norm": 0.28496550173938856, |
| "learning_rate": 5e-06, |
| "loss": 0.2762, |
| "step": 329 |
| }, |
| { |
| "epoch": 2.972972972972973, |
| "grad_norm": 0.2567341688659859, |
| "learning_rate": 5e-06, |
| "loss": 0.2756, |
| "step": 330 |
| }, |
| { |
| "epoch": 2.981981981981982, |
| "grad_norm": 0.2584671428651672, |
| "learning_rate": 5e-06, |
| "loss": 0.3001, |
| "step": 331 |
| }, |
| { |
| "epoch": 2.990990990990991, |
| "grad_norm": 0.2804525292556161, |
| "learning_rate": 5e-06, |
| "loss": 0.2785, |
| "step": 332 |
| }, |
| { |
| "epoch": 3.0, |
| "grad_norm": 0.24187503112431247, |
| "learning_rate": 5e-06, |
| "loss": 0.2454, |
| "step": 333 |
| }, |
| { |
| "epoch": 3.009009009009009, |
| "grad_norm": 0.26503328616806615, |
| "learning_rate": 5e-06, |
| "loss": 0.2332, |
| "step": 334 |
| }, |
| { |
| "epoch": 3.018018018018018, |
| "grad_norm": 0.2631846355641096, |
| "learning_rate": 5e-06, |
| "loss": 0.2658, |
| "step": 335 |
| }, |
| { |
| "epoch": 3.027027027027027, |
| "grad_norm": 0.2786137851508687, |
| "learning_rate": 5e-06, |
| "loss": 0.2519, |
| "step": 336 |
| }, |
| { |
| "epoch": 3.036036036036036, |
| "grad_norm": 0.2755722489630619, |
| "learning_rate": 5e-06, |
| "loss": 0.206, |
| "step": 337 |
| }, |
| { |
| "epoch": 3.045045045045045, |
| "grad_norm": 0.2530795628224832, |
| "learning_rate": 5e-06, |
| "loss": 0.2029, |
| "step": 338 |
| }, |
| { |
| "epoch": 3.054054054054054, |
| "grad_norm": 0.25959049991529565, |
| "learning_rate": 5e-06, |
| "loss": 0.2371, |
| "step": 339 |
| }, |
| { |
| "epoch": 3.063063063063063, |
| "grad_norm": 0.2916294807412774, |
| "learning_rate": 5e-06, |
| "loss": 0.2556, |
| "step": 340 |
| }, |
| { |
| "epoch": 3.0720720720720722, |
| "grad_norm": 0.2790615318391773, |
| "learning_rate": 5e-06, |
| "loss": 0.2198, |
| "step": 341 |
| }, |
| { |
| "epoch": 3.081081081081081, |
| "grad_norm": 0.3203392671142568, |
| "learning_rate": 5e-06, |
| "loss": 0.2693, |
| "step": 342 |
| }, |
| { |
| "epoch": 3.09009009009009, |
| "grad_norm": 0.2576637679316666, |
| "learning_rate": 5e-06, |
| "loss": 0.2304, |
| "step": 343 |
| }, |
| { |
| "epoch": 3.099099099099099, |
| "grad_norm": 0.24928248944605377, |
| "learning_rate": 5e-06, |
| "loss": 0.2291, |
| "step": 344 |
| }, |
| { |
| "epoch": 3.108108108108108, |
| "grad_norm": 0.26793696602953165, |
| "learning_rate": 5e-06, |
| "loss": 0.2445, |
| "step": 345 |
| }, |
| { |
| "epoch": 3.1171171171171173, |
| "grad_norm": 0.2971915014155351, |
| "learning_rate": 5e-06, |
| "loss": 0.2416, |
| "step": 346 |
| }, |
| { |
| "epoch": 3.126126126126126, |
| "grad_norm": 0.27752473751722373, |
| "learning_rate": 5e-06, |
| "loss": 0.219, |
| "step": 347 |
| }, |
| { |
| "epoch": 3.135135135135135, |
| "grad_norm": 0.2794534994824032, |
| "learning_rate": 5e-06, |
| "loss": 0.2297, |
| "step": 348 |
| }, |
| { |
| "epoch": 3.144144144144144, |
| "grad_norm": 0.25926948541703204, |
| "learning_rate": 5e-06, |
| "loss": 0.2411, |
| "step": 349 |
| }, |
| { |
| "epoch": 3.153153153153153, |
| "grad_norm": 0.24722918054941584, |
| "learning_rate": 5e-06, |
| "loss": 0.2419, |
| "step": 350 |
| }, |
| { |
| "epoch": 3.1621621621621623, |
| "grad_norm": 0.26203389917376085, |
| "learning_rate": 5e-06, |
| "loss": 0.234, |
| "step": 351 |
| }, |
| { |
| "epoch": 3.171171171171171, |
| "grad_norm": 0.2472074514309984, |
| "learning_rate": 5e-06, |
| "loss": 0.2363, |
| "step": 352 |
| }, |
| { |
| "epoch": 3.18018018018018, |
| "grad_norm": 0.2945063702553609, |
| "learning_rate": 5e-06, |
| "loss": 0.2435, |
| "step": 353 |
| }, |
| { |
| "epoch": 3.189189189189189, |
| "grad_norm": 0.27925373635853185, |
| "learning_rate": 5e-06, |
| "loss": 0.2746, |
| "step": 354 |
| }, |
| { |
| "epoch": 3.1981981981981984, |
| "grad_norm": 0.24996405356591392, |
| "learning_rate": 5e-06, |
| "loss": 0.2357, |
| "step": 355 |
| }, |
| { |
| "epoch": 3.2072072072072073, |
| "grad_norm": 0.2556106250304069, |
| "learning_rate": 5e-06, |
| "loss": 0.2499, |
| "step": 356 |
| }, |
| { |
| "epoch": 3.2162162162162162, |
| "grad_norm": 0.26114830248277804, |
| "learning_rate": 5e-06, |
| "loss": 0.2435, |
| "step": 357 |
| }, |
| { |
| "epoch": 3.225225225225225, |
| "grad_norm": 0.2469470177899144, |
| "learning_rate": 5e-06, |
| "loss": 0.2391, |
| "step": 358 |
| }, |
| { |
| "epoch": 3.234234234234234, |
| "grad_norm": 0.2641345310685226, |
| "learning_rate": 5e-06, |
| "loss": 0.1981, |
| "step": 359 |
| }, |
| { |
| "epoch": 3.2432432432432434, |
| "grad_norm": 0.2630942786949833, |
| "learning_rate": 5e-06, |
| "loss": 0.2098, |
| "step": 360 |
| }, |
| { |
| "epoch": 3.2522522522522523, |
| "grad_norm": 0.24708329710543495, |
| "learning_rate": 5e-06, |
| "loss": 0.2228, |
| "step": 361 |
| }, |
| { |
| "epoch": 3.2612612612612613, |
| "grad_norm": 0.25000693689900794, |
| "learning_rate": 5e-06, |
| "loss": 0.242, |
| "step": 362 |
| }, |
| { |
| "epoch": 3.27027027027027, |
| "grad_norm": 0.2554644897448756, |
| "learning_rate": 5e-06, |
| "loss": 0.2558, |
| "step": 363 |
| }, |
| { |
| "epoch": 3.279279279279279, |
| "grad_norm": 0.25264038317978293, |
| "learning_rate": 5e-06, |
| "loss": 0.2009, |
| "step": 364 |
| }, |
| { |
| "epoch": 3.2882882882882885, |
| "grad_norm": 0.2743512388274681, |
| "learning_rate": 5e-06, |
| "loss": 0.2331, |
| "step": 365 |
| }, |
| { |
| "epoch": 3.2972972972972974, |
| "grad_norm": 0.2728238972210015, |
| "learning_rate": 5e-06, |
| "loss": 0.238, |
| "step": 366 |
| }, |
| { |
| "epoch": 3.3063063063063063, |
| "grad_norm": 0.2602352997656632, |
| "learning_rate": 5e-06, |
| "loss": 0.2595, |
| "step": 367 |
| }, |
| { |
| "epoch": 3.315315315315315, |
| "grad_norm": 0.27036311534944873, |
| "learning_rate": 5e-06, |
| "loss": 0.2093, |
| "step": 368 |
| }, |
| { |
| "epoch": 3.3243243243243246, |
| "grad_norm": 0.264625202176752, |
| "learning_rate": 5e-06, |
| "loss": 0.2579, |
| "step": 369 |
| }, |
| { |
| "epoch": 3.3333333333333335, |
| "grad_norm": 0.259895631515348, |
| "learning_rate": 5e-06, |
| "loss": 0.2375, |
| "step": 370 |
| }, |
| { |
| "epoch": 3.3423423423423424, |
| "grad_norm": 0.2563353260712296, |
| "learning_rate": 5e-06, |
| "loss": 0.2364, |
| "step": 371 |
| }, |
| { |
| "epoch": 3.3513513513513513, |
| "grad_norm": 0.28822107627305354, |
| "learning_rate": 5e-06, |
| "loss": 0.2493, |
| "step": 372 |
| }, |
| { |
| "epoch": 3.3603603603603602, |
| "grad_norm": 0.25680447088580227, |
| "learning_rate": 5e-06, |
| "loss": 0.2501, |
| "step": 373 |
| }, |
| { |
| "epoch": 3.3693693693693696, |
| "grad_norm": 0.27784185650966475, |
| "learning_rate": 5e-06, |
| "loss": 0.245, |
| "step": 374 |
| }, |
| { |
| "epoch": 3.3783783783783785, |
| "grad_norm": 0.2627541742958857, |
| "learning_rate": 5e-06, |
| "loss": 0.2387, |
| "step": 375 |
| }, |
| { |
| "epoch": 3.3873873873873874, |
| "grad_norm": 0.24193274859474298, |
| "learning_rate": 5e-06, |
| "loss": 0.2375, |
| "step": 376 |
| }, |
| { |
| "epoch": 3.3963963963963963, |
| "grad_norm": 0.258378796876473, |
| "learning_rate": 5e-06, |
| "loss": 0.2281, |
| "step": 377 |
| }, |
| { |
| "epoch": 3.4054054054054053, |
| "grad_norm": 0.2749899330352957, |
| "learning_rate": 5e-06, |
| "loss": 0.24, |
| "step": 378 |
| }, |
| { |
| "epoch": 3.4144144144144146, |
| "grad_norm": 0.25777164751813997, |
| "learning_rate": 5e-06, |
| "loss": 0.2524, |
| "step": 379 |
| }, |
| { |
| "epoch": 3.4234234234234235, |
| "grad_norm": 0.2805168544005753, |
| "learning_rate": 5e-06, |
| "loss": 0.2415, |
| "step": 380 |
| }, |
| { |
| "epoch": 3.4324324324324325, |
| "grad_norm": 0.25842839628916536, |
| "learning_rate": 5e-06, |
| "loss": 0.2433, |
| "step": 381 |
| }, |
| { |
| "epoch": 3.4414414414414414, |
| "grad_norm": 0.26639980982056893, |
| "learning_rate": 5e-06, |
| "loss": 0.2403, |
| "step": 382 |
| }, |
| { |
| "epoch": 3.4504504504504503, |
| "grad_norm": 0.3060982219805088, |
| "learning_rate": 5e-06, |
| "loss": 0.2236, |
| "step": 383 |
| }, |
| { |
| "epoch": 3.4594594594594597, |
| "grad_norm": 0.26146902459280136, |
| "learning_rate": 5e-06, |
| "loss": 0.2504, |
| "step": 384 |
| }, |
| { |
| "epoch": 3.4684684684684686, |
| "grad_norm": 0.25380491317438975, |
| "learning_rate": 5e-06, |
| "loss": 0.2464, |
| "step": 385 |
| }, |
| { |
| "epoch": 3.4774774774774775, |
| "grad_norm": 0.27324232509875496, |
| "learning_rate": 5e-06, |
| "loss": 0.2404, |
| "step": 386 |
| }, |
| { |
| "epoch": 3.4864864864864864, |
| "grad_norm": 0.2651723560610241, |
| "learning_rate": 5e-06, |
| "loss": 0.226, |
| "step": 387 |
| }, |
| { |
| "epoch": 3.4954954954954953, |
| "grad_norm": 0.2689389917124243, |
| "learning_rate": 5e-06, |
| "loss": 0.2751, |
| "step": 388 |
| }, |
| { |
| "epoch": 3.5045045045045047, |
| "grad_norm": 0.2643418768447757, |
| "learning_rate": 5e-06, |
| "loss": 0.2298, |
| "step": 389 |
| }, |
| { |
| "epoch": 3.5135135135135136, |
| "grad_norm": 0.24935046689303417, |
| "learning_rate": 5e-06, |
| "loss": 0.2505, |
| "step": 390 |
| }, |
| { |
| "epoch": 3.5225225225225225, |
| "grad_norm": 0.2508765789856499, |
| "learning_rate": 5e-06, |
| "loss": 0.2478, |
| "step": 391 |
| }, |
| { |
| "epoch": 3.5315315315315314, |
| "grad_norm": 0.26705709850776205, |
| "learning_rate": 5e-06, |
| "loss": 0.2225, |
| "step": 392 |
| }, |
| { |
| "epoch": 3.5405405405405403, |
| "grad_norm": 0.2573422869010653, |
| "learning_rate": 5e-06, |
| "loss": 0.2122, |
| "step": 393 |
| }, |
| { |
| "epoch": 3.5495495495495497, |
| "grad_norm": 0.2770154762726231, |
| "learning_rate": 5e-06, |
| "loss": 0.2802, |
| "step": 394 |
| }, |
| { |
| "epoch": 3.5585585585585586, |
| "grad_norm": 0.26710684568846427, |
| "learning_rate": 5e-06, |
| "loss": 0.2275, |
| "step": 395 |
| }, |
| { |
| "epoch": 3.5675675675675675, |
| "grad_norm": 0.2527476600992376, |
| "learning_rate": 5e-06, |
| "loss": 0.2445, |
| "step": 396 |
| }, |
| { |
| "epoch": 3.5765765765765765, |
| "grad_norm": 0.2521141774058005, |
| "learning_rate": 5e-06, |
| "loss": 0.2342, |
| "step": 397 |
| }, |
| { |
| "epoch": 3.5855855855855854, |
| "grad_norm": 0.2689995200221707, |
| "learning_rate": 5e-06, |
| "loss": 0.2518, |
| "step": 398 |
| }, |
| { |
| "epoch": 3.5945945945945947, |
| "grad_norm": 0.25908754443823273, |
| "learning_rate": 5e-06, |
| "loss": 0.2386, |
| "step": 399 |
| }, |
| { |
| "epoch": 3.6036036036036037, |
| "grad_norm": 0.273518168337783, |
| "learning_rate": 5e-06, |
| "loss": 0.2641, |
| "step": 400 |
| }, |
| { |
| "epoch": 3.6126126126126126, |
| "grad_norm": 0.26669639385445737, |
| "learning_rate": 5e-06, |
| "loss": 0.2359, |
| "step": 401 |
| }, |
| { |
| "epoch": 3.6216216216216215, |
| "grad_norm": 0.2560702170541, |
| "learning_rate": 5e-06, |
| "loss": 0.2337, |
| "step": 402 |
| }, |
| { |
| "epoch": 3.6306306306306304, |
| "grad_norm": 0.2461177958525498, |
| "learning_rate": 5e-06, |
| "loss": 0.2401, |
| "step": 403 |
| }, |
| { |
| "epoch": 3.6396396396396398, |
| "grad_norm": 0.2648097200804019, |
| "learning_rate": 5e-06, |
| "loss": 0.2144, |
| "step": 404 |
| }, |
| { |
| "epoch": 3.6486486486486487, |
| "grad_norm": 0.2646834290329095, |
| "learning_rate": 5e-06, |
| "loss": 0.2493, |
| "step": 405 |
| }, |
| { |
| "epoch": 3.6576576576576576, |
| "grad_norm": 0.2796973639180676, |
| "learning_rate": 5e-06, |
| "loss": 0.2467, |
| "step": 406 |
| }, |
| { |
| "epoch": 3.6666666666666665, |
| "grad_norm": 0.25308085485220105, |
| "learning_rate": 5e-06, |
| "loss": 0.227, |
| "step": 407 |
| }, |
| { |
| "epoch": 3.6756756756756754, |
| "grad_norm": 0.2587012503429008, |
| "learning_rate": 5e-06, |
| "loss": 0.2285, |
| "step": 408 |
| }, |
| { |
| "epoch": 3.684684684684685, |
| "grad_norm": 0.2958300777778266, |
| "learning_rate": 5e-06, |
| "loss": 0.2547, |
| "step": 409 |
| }, |
| { |
| "epoch": 3.6936936936936937, |
| "grad_norm": 0.25334395158267925, |
| "learning_rate": 5e-06, |
| "loss": 0.2693, |
| "step": 410 |
| }, |
| { |
| "epoch": 3.7027027027027026, |
| "grad_norm": 0.29019457155713096, |
| "learning_rate": 5e-06, |
| "loss": 0.2518, |
| "step": 411 |
| }, |
| { |
| "epoch": 3.7117117117117115, |
| "grad_norm": 0.2473020184393372, |
| "learning_rate": 5e-06, |
| "loss": 0.2344, |
| "step": 412 |
| }, |
| { |
| "epoch": 3.7207207207207205, |
| "grad_norm": 0.270453761649425, |
| "learning_rate": 5e-06, |
| "loss": 0.2275, |
| "step": 413 |
| }, |
| { |
| "epoch": 3.72972972972973, |
| "grad_norm": 0.2602131546551776, |
| "learning_rate": 5e-06, |
| "loss": 0.2428, |
| "step": 414 |
| }, |
| { |
| "epoch": 3.7387387387387387, |
| "grad_norm": 0.29110180180417683, |
| "learning_rate": 5e-06, |
| "loss": 0.2301, |
| "step": 415 |
| }, |
| { |
| "epoch": 3.7477477477477477, |
| "grad_norm": 0.25367703106621997, |
| "learning_rate": 5e-06, |
| "loss": 0.2377, |
| "step": 416 |
| }, |
| { |
| "epoch": 3.756756756756757, |
| "grad_norm": 0.257299738969486, |
| "learning_rate": 5e-06, |
| "loss": 0.2137, |
| "step": 417 |
| }, |
| { |
| "epoch": 3.7657657657657655, |
| "grad_norm": 0.257656312443973, |
| "learning_rate": 5e-06, |
| "loss": 0.2567, |
| "step": 418 |
| }, |
| { |
| "epoch": 3.774774774774775, |
| "grad_norm": 0.2808325095308855, |
| "learning_rate": 5e-06, |
| "loss": 0.2571, |
| "step": 419 |
| }, |
| { |
| "epoch": 3.7837837837837838, |
| "grad_norm": 0.2657618644204265, |
| "learning_rate": 5e-06, |
| "loss": 0.2382, |
| "step": 420 |
| }, |
| { |
| "epoch": 3.7927927927927927, |
| "grad_norm": 0.27556658674977147, |
| "learning_rate": 5e-06, |
| "loss": 0.2748, |
| "step": 421 |
| }, |
| { |
| "epoch": 3.801801801801802, |
| "grad_norm": 0.2783118243091199, |
| "learning_rate": 5e-06, |
| "loss": 0.2349, |
| "step": 422 |
| }, |
| { |
| "epoch": 3.810810810810811, |
| "grad_norm": 0.27683880390148435, |
| "learning_rate": 5e-06, |
| "loss": 0.2545, |
| "step": 423 |
| }, |
| { |
| "epoch": 3.81981981981982, |
| "grad_norm": 0.24903071725050696, |
| "learning_rate": 5e-06, |
| "loss": 0.2436, |
| "step": 424 |
| }, |
| { |
| "epoch": 3.828828828828829, |
| "grad_norm": 0.27140890180707533, |
| "learning_rate": 5e-06, |
| "loss": 0.2103, |
| "step": 425 |
| }, |
| { |
| "epoch": 3.8378378378378377, |
| "grad_norm": 0.25999693913444694, |
| "learning_rate": 5e-06, |
| "loss": 0.2703, |
| "step": 426 |
| }, |
| { |
| "epoch": 3.846846846846847, |
| "grad_norm": 0.28165585165926776, |
| "learning_rate": 5e-06, |
| "loss": 0.2395, |
| "step": 427 |
| }, |
| { |
| "epoch": 3.855855855855856, |
| "grad_norm": 0.26800670806664434, |
| "learning_rate": 5e-06, |
| "loss": 0.2855, |
| "step": 428 |
| }, |
| { |
| "epoch": 3.864864864864865, |
| "grad_norm": 0.26752171553410126, |
| "learning_rate": 5e-06, |
| "loss": 0.2525, |
| "step": 429 |
| }, |
| { |
| "epoch": 3.873873873873874, |
| "grad_norm": 0.2550812423474624, |
| "learning_rate": 5e-06, |
| "loss": 0.2357, |
| "step": 430 |
| }, |
| { |
| "epoch": 3.8828828828828827, |
| "grad_norm": 0.25341757674985854, |
| "learning_rate": 5e-06, |
| "loss": 0.2752, |
| "step": 431 |
| }, |
| { |
| "epoch": 3.891891891891892, |
| "grad_norm": 0.2714456590973952, |
| "learning_rate": 5e-06, |
| "loss": 0.2399, |
| "step": 432 |
| }, |
| { |
| "epoch": 3.900900900900901, |
| "grad_norm": 0.2832850264958553, |
| "learning_rate": 5e-06, |
| "loss": 0.232, |
| "step": 433 |
| }, |
| { |
| "epoch": 3.90990990990991, |
| "grad_norm": 0.2560994537050628, |
| "learning_rate": 5e-06, |
| "loss": 0.2457, |
| "step": 434 |
| }, |
| { |
| "epoch": 3.918918918918919, |
| "grad_norm": 0.2624403245035626, |
| "learning_rate": 5e-06, |
| "loss": 0.2782, |
| "step": 435 |
| }, |
| { |
| "epoch": 3.9279279279279278, |
| "grad_norm": 0.2645012258501843, |
| "learning_rate": 5e-06, |
| "loss": 0.2432, |
| "step": 436 |
| }, |
| { |
| "epoch": 3.936936936936937, |
| "grad_norm": 0.26607477226554654, |
| "learning_rate": 5e-06, |
| "loss": 0.2202, |
| "step": 437 |
| }, |
| { |
| "epoch": 3.945945945945946, |
| "grad_norm": 0.2731452758231204, |
| "learning_rate": 5e-06, |
| "loss": 0.2689, |
| "step": 438 |
| }, |
| { |
| "epoch": 3.954954954954955, |
| "grad_norm": 0.2964590337329977, |
| "learning_rate": 5e-06, |
| "loss": 0.229, |
| "step": 439 |
| }, |
| { |
| "epoch": 3.963963963963964, |
| "grad_norm": 0.2787999534447745, |
| "learning_rate": 5e-06, |
| "loss": 0.2527, |
| "step": 440 |
| }, |
| { |
| "epoch": 3.972972972972973, |
| "grad_norm": 0.24055312465968123, |
| "learning_rate": 5e-06, |
| "loss": 0.2231, |
| "step": 441 |
| }, |
| { |
| "epoch": 3.981981981981982, |
| "grad_norm": 0.2757745274177008, |
| "learning_rate": 5e-06, |
| "loss": 0.2437, |
| "step": 442 |
| }, |
| { |
| "epoch": 3.990990990990991, |
| "grad_norm": 0.26536706718909975, |
| "learning_rate": 5e-06, |
| "loss": 0.2335, |
| "step": 443 |
| }, |
| { |
| "epoch": 4.0, |
| "grad_norm": 0.2390963333912312, |
| "learning_rate": 5e-06, |
| "loss": 0.2237, |
| "step": 444 |
| }, |
| { |
| "epoch": 4.009009009009009, |
| "grad_norm": 0.2720207934109716, |
| "learning_rate": 5e-06, |
| "loss": 0.2316, |
| "step": 445 |
| }, |
| { |
| "epoch": 4.018018018018018, |
| "grad_norm": 0.2673459557274162, |
| "learning_rate": 5e-06, |
| "loss": 0.2205, |
| "step": 446 |
| }, |
| { |
| "epoch": 4.027027027027027, |
| "grad_norm": 0.24447403903164172, |
| "learning_rate": 5e-06, |
| "loss": 0.2033, |
| "step": 447 |
| }, |
| { |
| "epoch": 4.036036036036036, |
| "grad_norm": 0.29354577394627634, |
| "learning_rate": 5e-06, |
| "loss": 0.2059, |
| "step": 448 |
| }, |
| { |
| "epoch": 4.045045045045045, |
| "grad_norm": 0.28252004790921936, |
| "learning_rate": 5e-06, |
| "loss": 0.2222, |
| "step": 449 |
| }, |
| { |
| "epoch": 4.054054054054054, |
| "grad_norm": 0.279624558559084, |
| "learning_rate": 5e-06, |
| "loss": 0.2485, |
| "step": 450 |
| }, |
| { |
| "epoch": 4.063063063063063, |
| "grad_norm": 0.2742544682456035, |
| "learning_rate": 5e-06, |
| "loss": 0.2153, |
| "step": 451 |
| }, |
| { |
| "epoch": 4.072072072072072, |
| "grad_norm": 0.26315979594288036, |
| "learning_rate": 5e-06, |
| "loss": 0.1967, |
| "step": 452 |
| }, |
| { |
| "epoch": 4.081081081081081, |
| "grad_norm": 0.25548950244986113, |
| "learning_rate": 5e-06, |
| "loss": 0.2108, |
| "step": 453 |
| }, |
| { |
| "epoch": 4.09009009009009, |
| "grad_norm": 0.260763131351132, |
| "learning_rate": 5e-06, |
| "loss": 0.21, |
| "step": 454 |
| }, |
| { |
| "epoch": 4.099099099099099, |
| "grad_norm": 0.2705243300559351, |
| "learning_rate": 5e-06, |
| "loss": 0.1926, |
| "step": 455 |
| }, |
| { |
| "epoch": 4.108108108108108, |
| "grad_norm": 0.2572296275587587, |
| "learning_rate": 5e-06, |
| "loss": 0.2182, |
| "step": 456 |
| }, |
| { |
| "epoch": 4.117117117117117, |
| "grad_norm": 0.32370825872086306, |
| "learning_rate": 5e-06, |
| "loss": 0.1912, |
| "step": 457 |
| }, |
| { |
| "epoch": 4.126126126126126, |
| "grad_norm": 0.24556795850306926, |
| "learning_rate": 5e-06, |
| "loss": 0.2071, |
| "step": 458 |
| }, |
| { |
| "epoch": 4.135135135135135, |
| "grad_norm": 0.23389148126428516, |
| "learning_rate": 5e-06, |
| "loss": 0.234, |
| "step": 459 |
| }, |
| { |
| "epoch": 4.1441441441441444, |
| "grad_norm": 0.2428236778448457, |
| "learning_rate": 5e-06, |
| "loss": 0.2119, |
| "step": 460 |
| }, |
| { |
| "epoch": 4.153153153153153, |
| "grad_norm": 0.31106881930176683, |
| "learning_rate": 5e-06, |
| "loss": 0.2124, |
| "step": 461 |
| }, |
| { |
| "epoch": 4.162162162162162, |
| "grad_norm": 0.27122185214756195, |
| "learning_rate": 5e-06, |
| "loss": 0.2226, |
| "step": 462 |
| }, |
| { |
| "epoch": 4.171171171171171, |
| "grad_norm": 0.2996732981773459, |
| "learning_rate": 5e-06, |
| "loss": 0.2299, |
| "step": 463 |
| }, |
| { |
| "epoch": 4.18018018018018, |
| "grad_norm": 0.27023462008753, |
| "learning_rate": 5e-06, |
| "loss": 0.2557, |
| "step": 464 |
| }, |
| { |
| "epoch": 4.1891891891891895, |
| "grad_norm": 0.25842796305339033, |
| "learning_rate": 5e-06, |
| "loss": 0.2325, |
| "step": 465 |
| }, |
| { |
| "epoch": 4.198198198198198, |
| "grad_norm": 0.2437169161762717, |
| "learning_rate": 5e-06, |
| "loss": 0.1797, |
| "step": 466 |
| }, |
| { |
| "epoch": 4.207207207207207, |
| "grad_norm": 0.26780073229070595, |
| "learning_rate": 5e-06, |
| "loss": 0.2067, |
| "step": 467 |
| }, |
| { |
| "epoch": 4.216216216216216, |
| "grad_norm": 0.2670888124294135, |
| "learning_rate": 5e-06, |
| "loss": 0.2205, |
| "step": 468 |
| }, |
| { |
| "epoch": 4.225225225225225, |
| "grad_norm": 0.25879921020859936, |
| "learning_rate": 5e-06, |
| "loss": 0.2085, |
| "step": 469 |
| }, |
| { |
| "epoch": 4.2342342342342345, |
| "grad_norm": 0.26317981293875226, |
| "learning_rate": 5e-06, |
| "loss": 0.2336, |
| "step": 470 |
| }, |
| { |
| "epoch": 4.243243243243243, |
| "grad_norm": 0.23931715866089387, |
| "learning_rate": 5e-06, |
| "loss": 0.2124, |
| "step": 471 |
| }, |
| { |
| "epoch": 4.252252252252252, |
| "grad_norm": 0.2691126922298142, |
| "learning_rate": 5e-06, |
| "loss": 0.2065, |
| "step": 472 |
| }, |
| { |
| "epoch": 4.261261261261261, |
| "grad_norm": 0.23991879914940956, |
| "learning_rate": 5e-06, |
| "loss": 0.2045, |
| "step": 473 |
| }, |
| { |
| "epoch": 4.27027027027027, |
| "grad_norm": 0.2548563923839949, |
| "learning_rate": 5e-06, |
| "loss": 0.1836, |
| "step": 474 |
| }, |
| { |
| "epoch": 4.2792792792792795, |
| "grad_norm": 0.24697361737276458, |
| "learning_rate": 5e-06, |
| "loss": 0.1792, |
| "step": 475 |
| }, |
| { |
| "epoch": 4.288288288288288, |
| "grad_norm": 0.2829022630675641, |
| "learning_rate": 5e-06, |
| "loss": 0.2098, |
| "step": 476 |
| }, |
| { |
| "epoch": 4.297297297297297, |
| "grad_norm": 0.2620700761228271, |
| "learning_rate": 5e-06, |
| "loss": 0.2102, |
| "step": 477 |
| }, |
| { |
| "epoch": 4.306306306306306, |
| "grad_norm": 0.2628063026021744, |
| "learning_rate": 5e-06, |
| "loss": 0.2025, |
| "step": 478 |
| }, |
| { |
| "epoch": 4.315315315315315, |
| "grad_norm": 0.2863724297024661, |
| "learning_rate": 5e-06, |
| "loss": 0.237, |
| "step": 479 |
| }, |
| { |
| "epoch": 4.324324324324325, |
| "grad_norm": 0.25990971129318524, |
| "learning_rate": 5e-06, |
| "loss": 0.2116, |
| "step": 480 |
| }, |
| { |
| "epoch": 4.333333333333333, |
| "grad_norm": 0.2606038664504591, |
| "learning_rate": 5e-06, |
| "loss": 0.2092, |
| "step": 481 |
| }, |
| { |
| "epoch": 4.342342342342342, |
| "grad_norm": 0.253863280864317, |
| "learning_rate": 5e-06, |
| "loss": 0.164, |
| "step": 482 |
| }, |
| { |
| "epoch": 4.351351351351352, |
| "grad_norm": 0.24650022322727727, |
| "learning_rate": 5e-06, |
| "loss": 0.1852, |
| "step": 483 |
| }, |
| { |
| "epoch": 4.36036036036036, |
| "grad_norm": 0.25369962373757826, |
| "learning_rate": 5e-06, |
| "loss": 0.2183, |
| "step": 484 |
| }, |
| { |
| "epoch": 4.36936936936937, |
| "grad_norm": 0.28375278856958064, |
| "learning_rate": 5e-06, |
| "loss": 0.2052, |
| "step": 485 |
| }, |
| { |
| "epoch": 4.378378378378378, |
| "grad_norm": 0.24267201305207473, |
| "learning_rate": 5e-06, |
| "loss": 0.2367, |
| "step": 486 |
| }, |
| { |
| "epoch": 4.387387387387387, |
| "grad_norm": 0.25205690964559024, |
| "learning_rate": 5e-06, |
| "loss": 0.2212, |
| "step": 487 |
| }, |
| { |
| "epoch": 4.396396396396397, |
| "grad_norm": 0.25716800876310375, |
| "learning_rate": 5e-06, |
| "loss": 0.1906, |
| "step": 488 |
| }, |
| { |
| "epoch": 4.405405405405405, |
| "grad_norm": 0.23704968081876604, |
| "learning_rate": 5e-06, |
| "loss": 0.2075, |
| "step": 489 |
| }, |
| { |
| "epoch": 4.414414414414415, |
| "grad_norm": 0.3201956523912786, |
| "learning_rate": 5e-06, |
| "loss": 0.198, |
| "step": 490 |
| }, |
| { |
| "epoch": 4.423423423423423, |
| "grad_norm": 0.26301398337918436, |
| "learning_rate": 5e-06, |
| "loss": 0.1878, |
| "step": 491 |
| }, |
| { |
| "epoch": 4.4324324324324325, |
| "grad_norm": 0.27402028797210554, |
| "learning_rate": 5e-06, |
| "loss": 0.2104, |
| "step": 492 |
| }, |
| { |
| "epoch": 4.441441441441442, |
| "grad_norm": 0.33955450203665727, |
| "learning_rate": 5e-06, |
| "loss": 0.2205, |
| "step": 493 |
| }, |
| { |
| "epoch": 4.45045045045045, |
| "grad_norm": 0.26220441610422024, |
| "learning_rate": 5e-06, |
| "loss": 0.218, |
| "step": 494 |
| }, |
| { |
| "epoch": 4.45945945945946, |
| "grad_norm": 0.281656218031479, |
| "learning_rate": 5e-06, |
| "loss": 0.2235, |
| "step": 495 |
| }, |
| { |
| "epoch": 4.468468468468468, |
| "grad_norm": 0.29159551817654267, |
| "learning_rate": 5e-06, |
| "loss": 0.2178, |
| "step": 496 |
| }, |
| { |
| "epoch": 4.4774774774774775, |
| "grad_norm": 0.2623117148967965, |
| "learning_rate": 5e-06, |
| "loss": 0.2178, |
| "step": 497 |
| }, |
| { |
| "epoch": 4.486486486486487, |
| "grad_norm": 0.2531840506893455, |
| "learning_rate": 5e-06, |
| "loss": 0.2086, |
| "step": 498 |
| }, |
| { |
| "epoch": 4.495495495495495, |
| "grad_norm": 0.25528977769788064, |
| "learning_rate": 5e-06, |
| "loss": 0.2186, |
| "step": 499 |
| }, |
| { |
| "epoch": 4.504504504504505, |
| "grad_norm": 0.2679628655435481, |
| "learning_rate": 5e-06, |
| "loss": 0.2237, |
| "step": 500 |
| }, |
| { |
| "epoch": 4.513513513513513, |
| "grad_norm": 0.263719988749634, |
| "learning_rate": 5e-06, |
| "loss": 0.194, |
| "step": 501 |
| }, |
| { |
| "epoch": 4.5225225225225225, |
| "grad_norm": 0.273138889734998, |
| "learning_rate": 5e-06, |
| "loss": 0.2086, |
| "step": 502 |
| }, |
| { |
| "epoch": 4.531531531531532, |
| "grad_norm": 0.292878429342998, |
| "learning_rate": 5e-06, |
| "loss": 0.2154, |
| "step": 503 |
| }, |
| { |
| "epoch": 4.54054054054054, |
| "grad_norm": 0.27619815070018144, |
| "learning_rate": 5e-06, |
| "loss": 0.2049, |
| "step": 504 |
| }, |
| { |
| "epoch": 4.54954954954955, |
| "grad_norm": 0.27527630799114594, |
| "learning_rate": 5e-06, |
| "loss": 0.2213, |
| "step": 505 |
| }, |
| { |
| "epoch": 4.558558558558558, |
| "grad_norm": 0.26879133234631997, |
| "learning_rate": 5e-06, |
| "loss": 0.2238, |
| "step": 506 |
| }, |
| { |
| "epoch": 4.5675675675675675, |
| "grad_norm": 0.272548643979066, |
| "learning_rate": 5e-06, |
| "loss": 0.2114, |
| "step": 507 |
| }, |
| { |
| "epoch": 4.576576576576577, |
| "grad_norm": 0.27819059711468064, |
| "learning_rate": 5e-06, |
| "loss": 0.2157, |
| "step": 508 |
| }, |
| { |
| "epoch": 4.585585585585585, |
| "grad_norm": 0.27618387944584083, |
| "learning_rate": 5e-06, |
| "loss": 0.2048, |
| "step": 509 |
| }, |
| { |
| "epoch": 4.594594594594595, |
| "grad_norm": 0.2549425189875316, |
| "learning_rate": 5e-06, |
| "loss": 0.2174, |
| "step": 510 |
| }, |
| { |
| "epoch": 4.603603603603604, |
| "grad_norm": 0.2645903835474375, |
| "learning_rate": 5e-06, |
| "loss": 0.2216, |
| "step": 511 |
| }, |
| { |
| "epoch": 4.612612612612613, |
| "grad_norm": 0.2640684028376376, |
| "learning_rate": 5e-06, |
| "loss": 0.2182, |
| "step": 512 |
| }, |
| { |
| "epoch": 4.621621621621622, |
| "grad_norm": 0.26051198776980117, |
| "learning_rate": 5e-06, |
| "loss": 0.1808, |
| "step": 513 |
| }, |
| { |
| "epoch": 4.63063063063063, |
| "grad_norm": 0.2931023356142575, |
| "learning_rate": 5e-06, |
| "loss": 0.1941, |
| "step": 514 |
| }, |
| { |
| "epoch": 4.63963963963964, |
| "grad_norm": 0.25284181276914397, |
| "learning_rate": 5e-06, |
| "loss": 0.2362, |
| "step": 515 |
| }, |
| { |
| "epoch": 4.648648648648649, |
| "grad_norm": 0.2590084071736973, |
| "learning_rate": 5e-06, |
| "loss": 0.2417, |
| "step": 516 |
| }, |
| { |
| "epoch": 4.657657657657658, |
| "grad_norm": 0.30404451969520124, |
| "learning_rate": 5e-06, |
| "loss": 0.2186, |
| "step": 517 |
| }, |
| { |
| "epoch": 4.666666666666667, |
| "grad_norm": 0.2673580882682002, |
| "learning_rate": 5e-06, |
| "loss": 0.2224, |
| "step": 518 |
| }, |
| { |
| "epoch": 4.675675675675675, |
| "grad_norm": 0.2636588657441614, |
| "learning_rate": 5e-06, |
| "loss": 0.2324, |
| "step": 519 |
| }, |
| { |
| "epoch": 4.684684684684685, |
| "grad_norm": 0.2876900527962799, |
| "learning_rate": 5e-06, |
| "loss": 0.2246, |
| "step": 520 |
| }, |
| { |
| "epoch": 4.693693693693694, |
| "grad_norm": 0.33566773437219993, |
| "learning_rate": 5e-06, |
| "loss": 0.2148, |
| "step": 521 |
| }, |
| { |
| "epoch": 4.702702702702703, |
| "grad_norm": 0.25837694824532986, |
| "learning_rate": 5e-06, |
| "loss": 0.2435, |
| "step": 522 |
| }, |
| { |
| "epoch": 4.711711711711712, |
| "grad_norm": 0.2618996341262811, |
| "learning_rate": 5e-06, |
| "loss": 0.2504, |
| "step": 523 |
| }, |
| { |
| "epoch": 4.7207207207207205, |
| "grad_norm": 0.2916721768764094, |
| "learning_rate": 5e-06, |
| "loss": 0.1969, |
| "step": 524 |
| }, |
| { |
| "epoch": 4.72972972972973, |
| "grad_norm": 0.2695124629228616, |
| "learning_rate": 5e-06, |
| "loss": 0.2242, |
| "step": 525 |
| }, |
| { |
| "epoch": 4.738738738738739, |
| "grad_norm": 0.25003767055634085, |
| "learning_rate": 5e-06, |
| "loss": 0.2396, |
| "step": 526 |
| }, |
| { |
| "epoch": 4.747747747747748, |
| "grad_norm": 0.26273587385726827, |
| "learning_rate": 5e-06, |
| "loss": 0.2389, |
| "step": 527 |
| }, |
| { |
| "epoch": 4.756756756756757, |
| "grad_norm": 0.2633999928270432, |
| "learning_rate": 5e-06, |
| "loss": 0.1999, |
| "step": 528 |
| }, |
| { |
| "epoch": 4.7657657657657655, |
| "grad_norm": 0.2657486138733691, |
| "learning_rate": 5e-06, |
| "loss": 0.1973, |
| "step": 529 |
| }, |
| { |
| "epoch": 4.774774774774775, |
| "grad_norm": 0.2615424263109113, |
| "learning_rate": 5e-06, |
| "loss": 0.2172, |
| "step": 530 |
| }, |
| { |
| "epoch": 4.783783783783784, |
| "grad_norm": 0.2725460425087256, |
| "learning_rate": 5e-06, |
| "loss": 0.2316, |
| "step": 531 |
| }, |
| { |
| "epoch": 4.792792792792793, |
| "grad_norm": 0.29663406158664646, |
| "learning_rate": 5e-06, |
| "loss": 0.215, |
| "step": 532 |
| }, |
| { |
| "epoch": 4.801801801801802, |
| "grad_norm": 0.2680114226198382, |
| "learning_rate": 5e-06, |
| "loss": 0.1913, |
| "step": 533 |
| }, |
| { |
| "epoch": 4.8108108108108105, |
| "grad_norm": 0.2717779322025023, |
| "learning_rate": 5e-06, |
| "loss": 0.2134, |
| "step": 534 |
| }, |
| { |
| "epoch": 4.81981981981982, |
| "grad_norm": 0.2461871817136421, |
| "learning_rate": 5e-06, |
| "loss": 0.2242, |
| "step": 535 |
| }, |
| { |
| "epoch": 4.828828828828829, |
| "grad_norm": 0.23898230675599963, |
| "learning_rate": 5e-06, |
| "loss": 0.1906, |
| "step": 536 |
| }, |
| { |
| "epoch": 4.837837837837838, |
| "grad_norm": 0.24493103786606743, |
| "learning_rate": 5e-06, |
| "loss": 0.2157, |
| "step": 537 |
| }, |
| { |
| "epoch": 4.846846846846847, |
| "grad_norm": 0.2513533399485069, |
| "learning_rate": 5e-06, |
| "loss": 0.2251, |
| "step": 538 |
| }, |
| { |
| "epoch": 4.8558558558558556, |
| "grad_norm": 0.25335345934289205, |
| "learning_rate": 5e-06, |
| "loss": 0.1981, |
| "step": 539 |
| }, |
| { |
| "epoch": 4.864864864864865, |
| "grad_norm": 0.24569861483369518, |
| "learning_rate": 5e-06, |
| "loss": 0.1923, |
| "step": 540 |
| }, |
| { |
| "epoch": 4.873873873873874, |
| "grad_norm": 0.3107988513160903, |
| "learning_rate": 5e-06, |
| "loss": 0.2022, |
| "step": 541 |
| }, |
| { |
| "epoch": 4.882882882882883, |
| "grad_norm": 0.2440474159047901, |
| "learning_rate": 5e-06, |
| "loss": 0.2102, |
| "step": 542 |
| }, |
| { |
| "epoch": 4.891891891891892, |
| "grad_norm": 0.269910179414699, |
| "learning_rate": 5e-06, |
| "loss": 0.2023, |
| "step": 543 |
| }, |
| { |
| "epoch": 4.900900900900901, |
| "grad_norm": 0.28697178165278897, |
| "learning_rate": 5e-06, |
| "loss": 0.2026, |
| "step": 544 |
| }, |
| { |
| "epoch": 4.90990990990991, |
| "grad_norm": 0.27623354559228885, |
| "learning_rate": 5e-06, |
| "loss": 0.2477, |
| "step": 545 |
| }, |
| { |
| "epoch": 4.918918918918919, |
| "grad_norm": 0.2598034982021407, |
| "learning_rate": 5e-06, |
| "loss": 0.202, |
| "step": 546 |
| }, |
| { |
| "epoch": 4.927927927927928, |
| "grad_norm": 0.2982050262473221, |
| "learning_rate": 5e-06, |
| "loss": 0.1987, |
| "step": 547 |
| }, |
| { |
| "epoch": 4.936936936936937, |
| "grad_norm": 0.2506438136769937, |
| "learning_rate": 5e-06, |
| "loss": 0.2332, |
| "step": 548 |
| }, |
| { |
| "epoch": 4.945945945945946, |
| "grad_norm": 0.2619541945846186, |
| "learning_rate": 5e-06, |
| "loss": 0.1929, |
| "step": 549 |
| }, |
| { |
| "epoch": 4.954954954954955, |
| "grad_norm": 0.263321542176826, |
| "learning_rate": 5e-06, |
| "loss": 0.199, |
| "step": 550 |
| }, |
| { |
| "epoch": 4.963963963963964, |
| "grad_norm": 0.2601674233941214, |
| "learning_rate": 5e-06, |
| "loss": 0.2515, |
| "step": 551 |
| }, |
| { |
| "epoch": 4.972972972972973, |
| "grad_norm": 0.29934162295077227, |
| "learning_rate": 5e-06, |
| "loss": 0.2177, |
| "step": 552 |
| }, |
| { |
| "epoch": 4.981981981981982, |
| "grad_norm": 0.24535555883333707, |
| "learning_rate": 5e-06, |
| "loss": 0.2038, |
| "step": 553 |
| }, |
| { |
| "epoch": 4.990990990990991, |
| "grad_norm": 0.2743717244598402, |
| "learning_rate": 5e-06, |
| "loss": 0.1904, |
| "step": 554 |
| }, |
| { |
| "epoch": 5.0, |
| "grad_norm": 0.2346279054988279, |
| "learning_rate": 5e-06, |
| "loss": 0.2318, |
| "step": 555 |
| }, |
| { |
| "epoch": 5.009009009009009, |
| "grad_norm": 0.3139762280953865, |
| "learning_rate": 5e-06, |
| "loss": 0.2065, |
| "step": 556 |
| }, |
| { |
| "epoch": 5.018018018018018, |
| "grad_norm": 0.2318927222535076, |
| "learning_rate": 5e-06, |
| "loss": 0.2063, |
| "step": 557 |
| }, |
| { |
| "epoch": 5.027027027027027, |
| "grad_norm": 0.25650614267529076, |
| "learning_rate": 5e-06, |
| "loss": 0.2042, |
| "step": 558 |
| }, |
| { |
| "epoch": 5.036036036036036, |
| "grad_norm": 0.25768317605269925, |
| "learning_rate": 5e-06, |
| "loss": 0.2057, |
| "step": 559 |
| }, |
| { |
| "epoch": 5.045045045045045, |
| "grad_norm": 0.29060238578973707, |
| "learning_rate": 5e-06, |
| "loss": 0.1944, |
| "step": 560 |
| }, |
| { |
| "epoch": 5.054054054054054, |
| "grad_norm": 0.28407299845741896, |
| "learning_rate": 5e-06, |
| "loss": 0.1718, |
| "step": 561 |
| }, |
| { |
| "epoch": 5.063063063063063, |
| "grad_norm": 0.29213793767158686, |
| "learning_rate": 5e-06, |
| "loss": 0.1878, |
| "step": 562 |
| }, |
| { |
| "epoch": 5.072072072072072, |
| "grad_norm": 0.26810675570875164, |
| "learning_rate": 5e-06, |
| "loss": 0.2125, |
| "step": 563 |
| }, |
| { |
| "epoch": 5.081081081081081, |
| "grad_norm": 0.2692377641775085, |
| "learning_rate": 5e-06, |
| "loss": 0.1846, |
| "step": 564 |
| }, |
| { |
| "epoch": 5.09009009009009, |
| "grad_norm": 0.405649877673358, |
| "learning_rate": 5e-06, |
| "loss": 0.1837, |
| "step": 565 |
| }, |
| { |
| "epoch": 5.099099099099099, |
| "grad_norm": 0.26726682072971775, |
| "learning_rate": 5e-06, |
| "loss": 0.2062, |
| "step": 566 |
| }, |
| { |
| "epoch": 5.108108108108108, |
| "grad_norm": 0.2940841675590565, |
| "learning_rate": 5e-06, |
| "loss": 0.2165, |
| "step": 567 |
| }, |
| { |
| "epoch": 5.117117117117117, |
| "grad_norm": 0.3398159316706572, |
| "learning_rate": 5e-06, |
| "loss": 0.1926, |
| "step": 568 |
| }, |
| { |
| "epoch": 5.126126126126126, |
| "grad_norm": 0.2826251512922728, |
| "learning_rate": 5e-06, |
| "loss": 0.1848, |
| "step": 569 |
| }, |
| { |
| "epoch": 5.135135135135135, |
| "grad_norm": 0.25092563468699364, |
| "learning_rate": 5e-06, |
| "loss": 0.193, |
| "step": 570 |
| }, |
| { |
| "epoch": 5.1441441441441444, |
| "grad_norm": 0.25159248777659954, |
| "learning_rate": 5e-06, |
| "loss": 0.1723, |
| "step": 571 |
| }, |
| { |
| "epoch": 5.153153153153153, |
| "grad_norm": 0.2681017671845892, |
| "learning_rate": 5e-06, |
| "loss": 0.1634, |
| "step": 572 |
| }, |
| { |
| "epoch": 5.162162162162162, |
| "grad_norm": 0.2733469299319058, |
| "learning_rate": 5e-06, |
| "loss": 0.1702, |
| "step": 573 |
| }, |
| { |
| "epoch": 5.171171171171171, |
| "grad_norm": 0.2643697126326926, |
| "learning_rate": 5e-06, |
| "loss": 0.2108, |
| "step": 574 |
| }, |
| { |
| "epoch": 5.18018018018018, |
| "grad_norm": 0.2929652382664824, |
| "learning_rate": 5e-06, |
| "loss": 0.2104, |
| "step": 575 |
| }, |
| { |
| "epoch": 5.1891891891891895, |
| "grad_norm": 0.30518478646977765, |
| "learning_rate": 5e-06, |
| "loss": 0.2049, |
| "step": 576 |
| }, |
| { |
| "epoch": 5.198198198198198, |
| "grad_norm": 0.29565787595285775, |
| "learning_rate": 5e-06, |
| "loss": 0.1701, |
| "step": 577 |
| }, |
| { |
| "epoch": 5.207207207207207, |
| "grad_norm": 0.24799846379048632, |
| "learning_rate": 5e-06, |
| "loss": 0.1849, |
| "step": 578 |
| }, |
| { |
| "epoch": 5.216216216216216, |
| "grad_norm": 0.26812878158143444, |
| "learning_rate": 5e-06, |
| "loss": 0.1939, |
| "step": 579 |
| }, |
| { |
| "epoch": 5.225225225225225, |
| "grad_norm": 0.2832327785366025, |
| "learning_rate": 5e-06, |
| "loss": 0.1748, |
| "step": 580 |
| }, |
| { |
| "epoch": 5.2342342342342345, |
| "grad_norm": 0.24530353488882148, |
| "learning_rate": 5e-06, |
| "loss": 0.1672, |
| "step": 581 |
| }, |
| { |
| "epoch": 5.243243243243243, |
| "grad_norm": 0.267893260322143, |
| "learning_rate": 5e-06, |
| "loss": 0.1982, |
| "step": 582 |
| }, |
| { |
| "epoch": 5.252252252252252, |
| "grad_norm": 0.28205728775241223, |
| "learning_rate": 5e-06, |
| "loss": 0.1402, |
| "step": 583 |
| }, |
| { |
| "epoch": 5.261261261261261, |
| "grad_norm": 0.2616195565718879, |
| "learning_rate": 5e-06, |
| "loss": 0.1669, |
| "step": 584 |
| }, |
| { |
| "epoch": 5.27027027027027, |
| "grad_norm": 0.2623448971573745, |
| "learning_rate": 5e-06, |
| "loss": 0.2006, |
| "step": 585 |
| }, |
| { |
| "epoch": 5.2792792792792795, |
| "grad_norm": 0.24193944254287217, |
| "learning_rate": 5e-06, |
| "loss": 0.1737, |
| "step": 586 |
| }, |
| { |
| "epoch": 5.288288288288288, |
| "grad_norm": 0.27208641316196014, |
| "learning_rate": 5e-06, |
| "loss": 0.1837, |
| "step": 587 |
| }, |
| { |
| "epoch": 5.297297297297297, |
| "grad_norm": 0.25067910651417047, |
| "learning_rate": 5e-06, |
| "loss": 0.1934, |
| "step": 588 |
| }, |
| { |
| "epoch": 5.306306306306306, |
| "grad_norm": 0.25385900871383876, |
| "learning_rate": 5e-06, |
| "loss": 0.1689, |
| "step": 589 |
| }, |
| { |
| "epoch": 5.315315315315315, |
| "grad_norm": 0.32902079040677734, |
| "learning_rate": 5e-06, |
| "loss": 0.1534, |
| "step": 590 |
| }, |
| { |
| "epoch": 5.324324324324325, |
| "grad_norm": 0.2529027343155485, |
| "learning_rate": 5e-06, |
| "loss": 0.2102, |
| "step": 591 |
| }, |
| { |
| "epoch": 5.333333333333333, |
| "grad_norm": 0.28906659508958055, |
| "learning_rate": 5e-06, |
| "loss": 0.181, |
| "step": 592 |
| }, |
| { |
| "epoch": 5.342342342342342, |
| "grad_norm": 0.282108480924088, |
| "learning_rate": 5e-06, |
| "loss": 0.2128, |
| "step": 593 |
| }, |
| { |
| "epoch": 5.351351351351352, |
| "grad_norm": 0.2604161116106256, |
| "learning_rate": 5e-06, |
| "loss": 0.1844, |
| "step": 594 |
| }, |
| { |
| "epoch": 5.36036036036036, |
| "grad_norm": 0.2789492989923241, |
| "learning_rate": 5e-06, |
| "loss": 0.1683, |
| "step": 595 |
| }, |
| { |
| "epoch": 5.36936936936937, |
| "grad_norm": 0.2559431308271593, |
| "learning_rate": 5e-06, |
| "loss": 0.2137, |
| "step": 596 |
| }, |
| { |
| "epoch": 5.378378378378378, |
| "grad_norm": 0.30088029917481107, |
| "learning_rate": 5e-06, |
| "loss": 0.1892, |
| "step": 597 |
| }, |
| { |
| "epoch": 5.387387387387387, |
| "grad_norm": 0.26253812275245714, |
| "learning_rate": 5e-06, |
| "loss": 0.1765, |
| "step": 598 |
| }, |
| { |
| "epoch": 5.396396396396397, |
| "grad_norm": 0.26495943964336816, |
| "learning_rate": 5e-06, |
| "loss": 0.1843, |
| "step": 599 |
| }, |
| { |
| "epoch": 5.405405405405405, |
| "grad_norm": 0.25894821975432253, |
| "learning_rate": 5e-06, |
| "loss": 0.1881, |
| "step": 600 |
| }, |
| { |
| "epoch": 5.414414414414415, |
| "grad_norm": 0.24931805970093998, |
| "learning_rate": 5e-06, |
| "loss": 0.1878, |
| "step": 601 |
| }, |
| { |
| "epoch": 5.423423423423423, |
| "grad_norm": 0.23455479372929255, |
| "learning_rate": 5e-06, |
| "loss": 0.1906, |
| "step": 602 |
| }, |
| { |
| "epoch": 5.4324324324324325, |
| "grad_norm": 0.25467507848802673, |
| "learning_rate": 5e-06, |
| "loss": 0.1717, |
| "step": 603 |
| }, |
| { |
| "epoch": 5.441441441441442, |
| "grad_norm": 0.33202611172740315, |
| "learning_rate": 5e-06, |
| "loss": 0.1688, |
| "step": 604 |
| }, |
| { |
| "epoch": 5.45045045045045, |
| "grad_norm": 0.29109320447844156, |
| "learning_rate": 5e-06, |
| "loss": 0.2029, |
| "step": 605 |
| }, |
| { |
| "epoch": 5.45945945945946, |
| "grad_norm": 0.24981105367499418, |
| "learning_rate": 5e-06, |
| "loss": 0.1941, |
| "step": 606 |
| }, |
| { |
| "epoch": 5.468468468468468, |
| "grad_norm": 0.24367601204155379, |
| "learning_rate": 5e-06, |
| "loss": 0.1671, |
| "step": 607 |
| }, |
| { |
| "epoch": 5.4774774774774775, |
| "grad_norm": 0.2932072155115799, |
| "learning_rate": 5e-06, |
| "loss": 0.1705, |
| "step": 608 |
| }, |
| { |
| "epoch": 5.486486486486487, |
| "grad_norm": 0.2882005482228378, |
| "learning_rate": 5e-06, |
| "loss": 0.2065, |
| "step": 609 |
| }, |
| { |
| "epoch": 5.495495495495495, |
| "grad_norm": 0.25719022643699463, |
| "learning_rate": 5e-06, |
| "loss": 0.2111, |
| "step": 610 |
| }, |
| { |
| "epoch": 5.504504504504505, |
| "grad_norm": 0.2611846325545377, |
| "learning_rate": 5e-06, |
| "loss": 0.2016, |
| "step": 611 |
| }, |
| { |
| "epoch": 5.513513513513513, |
| "grad_norm": 0.23251839540489064, |
| "learning_rate": 5e-06, |
| "loss": 0.1711, |
| "step": 612 |
| }, |
| { |
| "epoch": 5.5225225225225225, |
| "grad_norm": 0.3956644880260737, |
| "learning_rate": 5e-06, |
| "loss": 0.1794, |
| "step": 613 |
| }, |
| { |
| "epoch": 5.531531531531532, |
| "grad_norm": 0.27250839467887433, |
| "learning_rate": 5e-06, |
| "loss": 0.1676, |
| "step": 614 |
| }, |
| { |
| "epoch": 5.54054054054054, |
| "grad_norm": 0.2638663157341973, |
| "learning_rate": 5e-06, |
| "loss": 0.189, |
| "step": 615 |
| }, |
| { |
| "epoch": 5.54954954954955, |
| "grad_norm": 0.2635087015420886, |
| "learning_rate": 5e-06, |
| "loss": 0.1894, |
| "step": 616 |
| }, |
| { |
| "epoch": 5.558558558558558, |
| "grad_norm": 0.25884441144311887, |
| "learning_rate": 5e-06, |
| "loss": 0.1625, |
| "step": 617 |
| }, |
| { |
| "epoch": 5.5675675675675675, |
| "grad_norm": 0.33989481367732455, |
| "learning_rate": 5e-06, |
| "loss": 0.1632, |
| "step": 618 |
| }, |
| { |
| "epoch": 5.576576576576577, |
| "grad_norm": 0.33951958077722966, |
| "learning_rate": 5e-06, |
| "loss": 0.1956, |
| "step": 619 |
| }, |
| { |
| "epoch": 5.585585585585585, |
| "grad_norm": 0.2547652235180218, |
| "learning_rate": 5e-06, |
| "loss": 0.2291, |
| "step": 620 |
| }, |
| { |
| "epoch": 5.594594594594595, |
| "grad_norm": 0.24750295719042112, |
| "learning_rate": 5e-06, |
| "loss": 0.1726, |
| "step": 621 |
| }, |
| { |
| "epoch": 5.603603603603604, |
| "grad_norm": 0.2508541551180729, |
| "learning_rate": 5e-06, |
| "loss": 0.1776, |
| "step": 622 |
| }, |
| { |
| "epoch": 5.612612612612613, |
| "grad_norm": 0.2506039467248062, |
| "learning_rate": 5e-06, |
| "loss": 0.2185, |
| "step": 623 |
| }, |
| { |
| "epoch": 5.621621621621622, |
| "grad_norm": 0.26927980609500457, |
| "learning_rate": 5e-06, |
| "loss": 0.2051, |
| "step": 624 |
| }, |
| { |
| "epoch": 5.63063063063063, |
| "grad_norm": 0.2902598041361342, |
| "learning_rate": 5e-06, |
| "loss": 0.1836, |
| "step": 625 |
| }, |
| { |
| "epoch": 5.63963963963964, |
| "grad_norm": 0.27400647533943007, |
| "learning_rate": 5e-06, |
| "loss": 0.187, |
| "step": 626 |
| }, |
| { |
| "epoch": 5.648648648648649, |
| "grad_norm": 0.29199710457207273, |
| "learning_rate": 5e-06, |
| "loss": 0.1969, |
| "step": 627 |
| }, |
| { |
| "epoch": 5.657657657657658, |
| "grad_norm": 0.3025760209241755, |
| "learning_rate": 5e-06, |
| "loss": 0.1745, |
| "step": 628 |
| }, |
| { |
| "epoch": 5.666666666666667, |
| "grad_norm": 0.2863200552497931, |
| "learning_rate": 5e-06, |
| "loss": 0.1763, |
| "step": 629 |
| }, |
| { |
| "epoch": 5.675675675675675, |
| "grad_norm": 0.3046187504171871, |
| "learning_rate": 5e-06, |
| "loss": 0.1765, |
| "step": 630 |
| }, |
| { |
| "epoch": 5.684684684684685, |
| "grad_norm": 0.2594010152562734, |
| "learning_rate": 5e-06, |
| "loss": 0.1922, |
| "step": 631 |
| }, |
| { |
| "epoch": 5.693693693693694, |
| "grad_norm": 0.25276964471192975, |
| "learning_rate": 5e-06, |
| "loss": 0.2011, |
| "step": 632 |
| }, |
| { |
| "epoch": 5.702702702702703, |
| "grad_norm": 0.2457717647956263, |
| "learning_rate": 5e-06, |
| "loss": 0.1878, |
| "step": 633 |
| }, |
| { |
| "epoch": 5.711711711711712, |
| "grad_norm": 0.27348692878164155, |
| "learning_rate": 5e-06, |
| "loss": 0.1992, |
| "step": 634 |
| }, |
| { |
| "epoch": 5.7207207207207205, |
| "grad_norm": 0.2599835122727351, |
| "learning_rate": 5e-06, |
| "loss": 0.1649, |
| "step": 635 |
| }, |
| { |
| "epoch": 5.72972972972973, |
| "grad_norm": 0.2712466634459408, |
| "learning_rate": 5e-06, |
| "loss": 0.2089, |
| "step": 636 |
| }, |
| { |
| "epoch": 5.738738738738739, |
| "grad_norm": 0.2732874374632016, |
| "learning_rate": 5e-06, |
| "loss": 0.1613, |
| "step": 637 |
| }, |
| { |
| "epoch": 5.747747747747748, |
| "grad_norm": 0.3335330997010001, |
| "learning_rate": 5e-06, |
| "loss": 0.2057, |
| "step": 638 |
| }, |
| { |
| "epoch": 5.756756756756757, |
| "grad_norm": 0.29795878105581997, |
| "learning_rate": 5e-06, |
| "loss": 0.1619, |
| "step": 639 |
| }, |
| { |
| "epoch": 5.7657657657657655, |
| "grad_norm": 0.27969811406236256, |
| "learning_rate": 5e-06, |
| "loss": 0.212, |
| "step": 640 |
| }, |
| { |
| "epoch": 5.774774774774775, |
| "grad_norm": 0.26108347760571876, |
| "learning_rate": 5e-06, |
| "loss": 0.1898, |
| "step": 641 |
| }, |
| { |
| "epoch": 5.783783783783784, |
| "grad_norm": 0.2954357533804664, |
| "learning_rate": 5e-06, |
| "loss": 0.1845, |
| "step": 642 |
| }, |
| { |
| "epoch": 5.792792792792793, |
| "grad_norm": 0.2897237531310712, |
| "learning_rate": 5e-06, |
| "loss": 0.1778, |
| "step": 643 |
| }, |
| { |
| "epoch": 5.801801801801802, |
| "grad_norm": 0.2862916112801224, |
| "learning_rate": 5e-06, |
| "loss": 0.1616, |
| "step": 644 |
| }, |
| { |
| "epoch": 5.8108108108108105, |
| "grad_norm": 0.2655242503575118, |
| "learning_rate": 5e-06, |
| "loss": 0.2125, |
| "step": 645 |
| }, |
| { |
| "epoch": 5.81981981981982, |
| "grad_norm": 0.29888735697965757, |
| "learning_rate": 5e-06, |
| "loss": 0.17, |
| "step": 646 |
| }, |
| { |
| "epoch": 5.828828828828829, |
| "grad_norm": 0.3878759366680586, |
| "learning_rate": 5e-06, |
| "loss": 0.1747, |
| "step": 647 |
| }, |
| { |
| "epoch": 5.837837837837838, |
| "grad_norm": 0.2780142798396489, |
| "learning_rate": 5e-06, |
| "loss": 0.208, |
| "step": 648 |
| }, |
| { |
| "epoch": 5.846846846846847, |
| "grad_norm": 0.32189822402929597, |
| "learning_rate": 5e-06, |
| "loss": 0.1903, |
| "step": 649 |
| }, |
| { |
| "epoch": 5.8558558558558556, |
| "grad_norm": 0.26666263423680436, |
| "learning_rate": 5e-06, |
| "loss": 0.1753, |
| "step": 650 |
| }, |
| { |
| "epoch": 5.864864864864865, |
| "grad_norm": 0.26508147192359016, |
| "learning_rate": 5e-06, |
| "loss": 0.1956, |
| "step": 651 |
| }, |
| { |
| "epoch": 5.873873873873874, |
| "grad_norm": 0.2800648058751269, |
| "learning_rate": 5e-06, |
| "loss": 0.1796, |
| "step": 652 |
| }, |
| { |
| "epoch": 5.882882882882883, |
| "grad_norm": 0.25602425755319697, |
| "learning_rate": 5e-06, |
| "loss": 0.2028, |
| "step": 653 |
| }, |
| { |
| "epoch": 5.891891891891892, |
| "grad_norm": 0.296930917045288, |
| "learning_rate": 5e-06, |
| "loss": 0.1924, |
| "step": 654 |
| }, |
| { |
| "epoch": 5.900900900900901, |
| "grad_norm": 0.3452776489094155, |
| "learning_rate": 5e-06, |
| "loss": 0.1964, |
| "step": 655 |
| }, |
| { |
| "epoch": 5.90990990990991, |
| "grad_norm": 0.2910698717731606, |
| "learning_rate": 5e-06, |
| "loss": 0.1751, |
| "step": 656 |
| }, |
| { |
| "epoch": 5.918918918918919, |
| "grad_norm": 0.2591517828645954, |
| "learning_rate": 5e-06, |
| "loss": 0.187, |
| "step": 657 |
| }, |
| { |
| "epoch": 5.927927927927928, |
| "grad_norm": 0.32146446439072224, |
| "learning_rate": 5e-06, |
| "loss": 0.1945, |
| "step": 658 |
| }, |
| { |
| "epoch": 5.936936936936937, |
| "grad_norm": 0.29057113691944186, |
| "learning_rate": 5e-06, |
| "loss": 0.1796, |
| "step": 659 |
| }, |
| { |
| "epoch": 5.945945945945946, |
| "grad_norm": 0.2673772774533524, |
| "learning_rate": 5e-06, |
| "loss": 0.1871, |
| "step": 660 |
| }, |
| { |
| "epoch": 5.954954954954955, |
| "grad_norm": 0.25292557260096377, |
| "learning_rate": 5e-06, |
| "loss": 0.1748, |
| "step": 661 |
| }, |
| { |
| "epoch": 5.963963963963964, |
| "grad_norm": 0.24315435369686791, |
| "learning_rate": 5e-06, |
| "loss": 0.1838, |
| "step": 662 |
| }, |
| { |
| "epoch": 5.972972972972973, |
| "grad_norm": 0.30275438050027514, |
| "learning_rate": 5e-06, |
| "loss": 0.1633, |
| "step": 663 |
| }, |
| { |
| "epoch": 5.981981981981982, |
| "grad_norm": 0.28436057893273076, |
| "learning_rate": 5e-06, |
| "loss": 0.1876, |
| "step": 664 |
| }, |
| { |
| "epoch": 5.990990990990991, |
| "grad_norm": 0.28562922979220184, |
| "learning_rate": 5e-06, |
| "loss": 0.2009, |
| "step": 665 |
| }, |
| { |
| "epoch": 6.0, |
| "grad_norm": 0.2653424601600143, |
| "learning_rate": 5e-06, |
| "loss": 0.1905, |
| "step": 666 |
| }, |
| { |
| "epoch": 6.009009009009009, |
| "grad_norm": 0.2860376096075966, |
| "learning_rate": 5e-06, |
| "loss": 0.1957, |
| "step": 667 |
| }, |
| { |
| "epoch": 6.018018018018018, |
| "grad_norm": 0.25196665164096865, |
| "learning_rate": 5e-06, |
| "loss": 0.1563, |
| "step": 668 |
| }, |
| { |
| "epoch": 6.027027027027027, |
| "grad_norm": 0.24029344524647256, |
| "learning_rate": 5e-06, |
| "loss": 0.1879, |
| "step": 669 |
| }, |
| { |
| "epoch": 6.036036036036036, |
| "grad_norm": 0.2620085799429486, |
| "learning_rate": 5e-06, |
| "loss": 0.1811, |
| "step": 670 |
| }, |
| { |
| "epoch": 6.045045045045045, |
| "grad_norm": 0.27308115959180734, |
| "learning_rate": 5e-06, |
| "loss": 0.155, |
| "step": 671 |
| }, |
| { |
| "epoch": 6.054054054054054, |
| "grad_norm": 0.26803737868546207, |
| "learning_rate": 5e-06, |
| "loss": 0.1537, |
| "step": 672 |
| }, |
| { |
| "epoch": 6.063063063063063, |
| "grad_norm": 0.30441930072274076, |
| "learning_rate": 5e-06, |
| "loss": 0.1902, |
| "step": 673 |
| }, |
| { |
| "epoch": 6.072072072072072, |
| "grad_norm": 0.2465984202629159, |
| "learning_rate": 5e-06, |
| "loss": 0.1807, |
| "step": 674 |
| }, |
| { |
| "epoch": 6.081081081081081, |
| "grad_norm": 0.2674335217467193, |
| "learning_rate": 5e-06, |
| "loss": 0.1487, |
| "step": 675 |
| }, |
| { |
| "epoch": 6.09009009009009, |
| "grad_norm": 0.2905204800351543, |
| "learning_rate": 5e-06, |
| "loss": 0.1567, |
| "step": 676 |
| }, |
| { |
| "epoch": 6.099099099099099, |
| "grad_norm": 0.2954597077236978, |
| "learning_rate": 5e-06, |
| "loss": 0.1535, |
| "step": 677 |
| }, |
| { |
| "epoch": 6.108108108108108, |
| "grad_norm": 0.3045298345267351, |
| "learning_rate": 5e-06, |
| "loss": 0.1689, |
| "step": 678 |
| }, |
| { |
| "epoch": 6.117117117117117, |
| "grad_norm": 0.2740781768489349, |
| "learning_rate": 5e-06, |
| "loss": 0.1797, |
| "step": 679 |
| }, |
| { |
| "epoch": 6.126126126126126, |
| "grad_norm": 0.39475528450021763, |
| "learning_rate": 5e-06, |
| "loss": 0.1617, |
| "step": 680 |
| }, |
| { |
| "epoch": 6.135135135135135, |
| "grad_norm": 0.6235225287605396, |
| "learning_rate": 5e-06, |
| "loss": 0.1238, |
| "step": 681 |
| }, |
| { |
| "epoch": 6.1441441441441444, |
| "grad_norm": 0.26845753887421847, |
| "learning_rate": 5e-06, |
| "loss": 0.1511, |
| "step": 682 |
| }, |
| { |
| "epoch": 6.153153153153153, |
| "grad_norm": 0.3602960092750115, |
| "learning_rate": 5e-06, |
| "loss": 0.1678, |
| "step": 683 |
| }, |
| { |
| "epoch": 6.162162162162162, |
| "grad_norm": 0.33224893659794336, |
| "learning_rate": 5e-06, |
| "loss": 0.1412, |
| "step": 684 |
| }, |
| { |
| "epoch": 6.171171171171171, |
| "grad_norm": 0.24094527332728147, |
| "learning_rate": 5e-06, |
| "loss": 0.1695, |
| "step": 685 |
| }, |
| { |
| "epoch": 6.18018018018018, |
| "grad_norm": 0.341428905288911, |
| "learning_rate": 5e-06, |
| "loss": 0.1839, |
| "step": 686 |
| }, |
| { |
| "epoch": 6.1891891891891895, |
| "grad_norm": 0.2956801407312396, |
| "learning_rate": 5e-06, |
| "loss": 0.1692, |
| "step": 687 |
| }, |
| { |
| "epoch": 6.198198198198198, |
| "grad_norm": 0.3054866867274709, |
| "learning_rate": 5e-06, |
| "loss": 0.1546, |
| "step": 688 |
| }, |
| { |
| "epoch": 6.207207207207207, |
| "grad_norm": 0.23806841375933424, |
| "learning_rate": 5e-06, |
| "loss": 0.1554, |
| "step": 689 |
| }, |
| { |
| "epoch": 6.216216216216216, |
| "grad_norm": 0.24481339250784975, |
| "learning_rate": 5e-06, |
| "loss": 0.1953, |
| "step": 690 |
| }, |
| { |
| "epoch": 6.225225225225225, |
| "grad_norm": 0.3014128409778474, |
| "learning_rate": 5e-06, |
| "loss": 0.1674, |
| "step": 691 |
| }, |
| { |
| "epoch": 6.2342342342342345, |
| "grad_norm": 0.2708989701315342, |
| "learning_rate": 5e-06, |
| "loss": 0.1479, |
| "step": 692 |
| }, |
| { |
| "epoch": 6.243243243243243, |
| "grad_norm": 0.31024609108715306, |
| "learning_rate": 5e-06, |
| "loss": 0.1553, |
| "step": 693 |
| }, |
| { |
| "epoch": 6.252252252252252, |
| "grad_norm": 0.29134393470015496, |
| "learning_rate": 5e-06, |
| "loss": 0.1437, |
| "step": 694 |
| }, |
| { |
| "epoch": 6.261261261261261, |
| "grad_norm": 0.2793592485054197, |
| "learning_rate": 5e-06, |
| "loss": 0.1684, |
| "step": 695 |
| }, |
| { |
| "epoch": 6.27027027027027, |
| "grad_norm": 0.30498815020407055, |
| "learning_rate": 5e-06, |
| "loss": 0.1968, |
| "step": 696 |
| }, |
| { |
| "epoch": 6.2792792792792795, |
| "grad_norm": 0.2652672098205942, |
| "learning_rate": 5e-06, |
| "loss": 0.1565, |
| "step": 697 |
| }, |
| { |
| "epoch": 6.288288288288288, |
| "grad_norm": 0.3018458330908521, |
| "learning_rate": 5e-06, |
| "loss": 0.1726, |
| "step": 698 |
| }, |
| { |
| "epoch": 6.297297297297297, |
| "grad_norm": 0.2592172426217306, |
| "learning_rate": 5e-06, |
| "loss": 0.1656, |
| "step": 699 |
| }, |
| { |
| "epoch": 6.306306306306306, |
| "grad_norm": 0.2565291008895072, |
| "learning_rate": 5e-06, |
| "loss": 0.1981, |
| "step": 700 |
| }, |
| { |
| "epoch": 6.315315315315315, |
| "grad_norm": 0.2980025331247, |
| "learning_rate": 5e-06, |
| "loss": 0.1656, |
| "step": 701 |
| }, |
| { |
| "epoch": 6.324324324324325, |
| "grad_norm": 0.26039013888986284, |
| "learning_rate": 5e-06, |
| "loss": 0.1656, |
| "step": 702 |
| }, |
| { |
| "epoch": 6.333333333333333, |
| "grad_norm": 0.2568935923552546, |
| "learning_rate": 5e-06, |
| "loss": 0.1451, |
| "step": 703 |
| }, |
| { |
| "epoch": 6.342342342342342, |
| "grad_norm": 0.29222987592831656, |
| "learning_rate": 5e-06, |
| "loss": 0.161, |
| "step": 704 |
| }, |
| { |
| "epoch": 6.351351351351352, |
| "grad_norm": 0.2622511449178775, |
| "learning_rate": 5e-06, |
| "loss": 0.1553, |
| "step": 705 |
| }, |
| { |
| "epoch": 6.36036036036036, |
| "grad_norm": 0.2703894332854895, |
| "learning_rate": 5e-06, |
| "loss": 0.1803, |
| "step": 706 |
| }, |
| { |
| "epoch": 6.36936936936937, |
| "grad_norm": 0.2516505913848481, |
| "learning_rate": 5e-06, |
| "loss": 0.16, |
| "step": 707 |
| }, |
| { |
| "epoch": 6.378378378378378, |
| "grad_norm": 0.26750256687760715, |
| "learning_rate": 5e-06, |
| "loss": 0.1461, |
| "step": 708 |
| }, |
| { |
| "epoch": 6.387387387387387, |
| "grad_norm": 0.2539871109081379, |
| "learning_rate": 5e-06, |
| "loss": 0.1488, |
| "step": 709 |
| }, |
| { |
| "epoch": 6.396396396396397, |
| "grad_norm": 0.2769403607227516, |
| "learning_rate": 5e-06, |
| "loss": 0.1924, |
| "step": 710 |
| }, |
| { |
| "epoch": 6.405405405405405, |
| "grad_norm": 0.2946720991492928, |
| "learning_rate": 5e-06, |
| "loss": 0.1503, |
| "step": 711 |
| }, |
| { |
| "epoch": 6.414414414414415, |
| "grad_norm": 0.24458166835948247, |
| "learning_rate": 5e-06, |
| "loss": 0.1855, |
| "step": 712 |
| }, |
| { |
| "epoch": 6.423423423423423, |
| "grad_norm": 0.2840232732624716, |
| "learning_rate": 5e-06, |
| "loss": 0.1643, |
| "step": 713 |
| }, |
| { |
| "epoch": 6.4324324324324325, |
| "grad_norm": 0.31757015141649597, |
| "learning_rate": 5e-06, |
| "loss": 0.1622, |
| "step": 714 |
| }, |
| { |
| "epoch": 6.441441441441442, |
| "grad_norm": 0.28847324036631117, |
| "learning_rate": 5e-06, |
| "loss": 0.1365, |
| "step": 715 |
| }, |
| { |
| "epoch": 6.45045045045045, |
| "grad_norm": 0.24694988398848988, |
| "learning_rate": 5e-06, |
| "loss": 0.1594, |
| "step": 716 |
| }, |
| { |
| "epoch": 6.45945945945946, |
| "grad_norm": 0.29307213864672693, |
| "learning_rate": 5e-06, |
| "loss": 0.1411, |
| "step": 717 |
| }, |
| { |
| "epoch": 6.468468468468468, |
| "grad_norm": 0.30163977699200506, |
| "learning_rate": 5e-06, |
| "loss": 0.1649, |
| "step": 718 |
| }, |
| { |
| "epoch": 6.4774774774774775, |
| "grad_norm": 0.2854457863377953, |
| "learning_rate": 5e-06, |
| "loss": 0.1918, |
| "step": 719 |
| }, |
| { |
| "epoch": 6.486486486486487, |
| "grad_norm": 0.27342932900047295, |
| "learning_rate": 5e-06, |
| "loss": 0.176, |
| "step": 720 |
| }, |
| { |
| "epoch": 6.495495495495495, |
| "grad_norm": 0.28175783115173536, |
| "learning_rate": 5e-06, |
| "loss": 0.1422, |
| "step": 721 |
| }, |
| { |
| "epoch": 6.504504504504505, |
| "grad_norm": 0.2840989876099184, |
| "learning_rate": 5e-06, |
| "loss": 0.1633, |
| "step": 722 |
| }, |
| { |
| "epoch": 6.513513513513513, |
| "grad_norm": 0.2867793910350591, |
| "learning_rate": 5e-06, |
| "loss": 0.1796, |
| "step": 723 |
| }, |
| { |
| "epoch": 6.5225225225225225, |
| "grad_norm": 0.28428808863989385, |
| "learning_rate": 5e-06, |
| "loss": 0.1529, |
| "step": 724 |
| }, |
| { |
| "epoch": 6.531531531531532, |
| "grad_norm": 0.29279024558574074, |
| "learning_rate": 5e-06, |
| "loss": 0.1392, |
| "step": 725 |
| }, |
| { |
| "epoch": 6.54054054054054, |
| "grad_norm": 0.23321656651720726, |
| "learning_rate": 5e-06, |
| "loss": 0.1472, |
| "step": 726 |
| }, |
| { |
| "epoch": 6.54954954954955, |
| "grad_norm": 0.27521328727823563, |
| "learning_rate": 5e-06, |
| "loss": 0.1577, |
| "step": 727 |
| }, |
| { |
| "epoch": 6.558558558558558, |
| "grad_norm": 0.32541783429708115, |
| "learning_rate": 5e-06, |
| "loss": 0.1782, |
| "step": 728 |
| }, |
| { |
| "epoch": 6.5675675675675675, |
| "grad_norm": 0.4716962818206086, |
| "learning_rate": 5e-06, |
| "loss": 0.1247, |
| "step": 729 |
| }, |
| { |
| "epoch": 6.576576576576577, |
| "grad_norm": 0.26285903206886113, |
| "learning_rate": 5e-06, |
| "loss": 0.1744, |
| "step": 730 |
| }, |
| { |
| "epoch": 6.585585585585585, |
| "grad_norm": 0.30168633716148247, |
| "learning_rate": 5e-06, |
| "loss": 0.1422, |
| "step": 731 |
| }, |
| { |
| "epoch": 6.594594594594595, |
| "grad_norm": 0.2745027764141301, |
| "learning_rate": 5e-06, |
| "loss": 0.1672, |
| "step": 732 |
| }, |
| { |
| "epoch": 6.603603603603604, |
| "grad_norm": 0.35380479132918236, |
| "learning_rate": 5e-06, |
| "loss": 0.1663, |
| "step": 733 |
| }, |
| { |
| "epoch": 6.612612612612613, |
| "grad_norm": 0.27601944134435535, |
| "learning_rate": 5e-06, |
| "loss": 0.1695, |
| "step": 734 |
| }, |
| { |
| "epoch": 6.621621621621622, |
| "grad_norm": 0.2528980448808799, |
| "learning_rate": 5e-06, |
| "loss": 0.2017, |
| "step": 735 |
| }, |
| { |
| "epoch": 6.63063063063063, |
| "grad_norm": 0.30767286206094524, |
| "learning_rate": 5e-06, |
| "loss": 0.1261, |
| "step": 736 |
| }, |
| { |
| "epoch": 6.63963963963964, |
| "grad_norm": 0.2602137688236013, |
| "learning_rate": 5e-06, |
| "loss": 0.1636, |
| "step": 737 |
| }, |
| { |
| "epoch": 6.648648648648649, |
| "grad_norm": 0.23824598012350529, |
| "learning_rate": 5e-06, |
| "loss": 0.1584, |
| "step": 738 |
| }, |
| { |
| "epoch": 6.657657657657658, |
| "grad_norm": 0.29134756631872455, |
| "learning_rate": 5e-06, |
| "loss": 0.1835, |
| "step": 739 |
| }, |
| { |
| "epoch": 6.666666666666667, |
| "grad_norm": 0.25765774787058354, |
| "learning_rate": 5e-06, |
| "loss": 0.1603, |
| "step": 740 |
| }, |
| { |
| "epoch": 6.675675675675675, |
| "grad_norm": 0.2600078403016356, |
| "learning_rate": 5e-06, |
| "loss": 0.1826, |
| "step": 741 |
| }, |
| { |
| "epoch": 6.684684684684685, |
| "grad_norm": 0.2617835836231004, |
| "learning_rate": 5e-06, |
| "loss": 0.1464, |
| "step": 742 |
| }, |
| { |
| "epoch": 6.693693693693694, |
| "grad_norm": 0.32078684057749896, |
| "learning_rate": 5e-06, |
| "loss": 0.1252, |
| "step": 743 |
| }, |
| { |
| "epoch": 6.702702702702703, |
| "grad_norm": 0.29351670808174113, |
| "learning_rate": 5e-06, |
| "loss": 0.1548, |
| "step": 744 |
| }, |
| { |
| "epoch": 6.711711711711712, |
| "grad_norm": 0.30854251276850175, |
| "learning_rate": 5e-06, |
| "loss": 0.137, |
| "step": 745 |
| }, |
| { |
| "epoch": 6.7207207207207205, |
| "grad_norm": 0.26688862536435537, |
| "learning_rate": 5e-06, |
| "loss": 0.124, |
| "step": 746 |
| }, |
| { |
| "epoch": 6.72972972972973, |
| "grad_norm": 0.44923760580414157, |
| "learning_rate": 5e-06, |
| "loss": 0.1736, |
| "step": 747 |
| }, |
| { |
| "epoch": 6.738738738738739, |
| "grad_norm": 0.39218610199513526, |
| "learning_rate": 5e-06, |
| "loss": 0.1418, |
| "step": 748 |
| }, |
| { |
| "epoch": 6.747747747747748, |
| "grad_norm": 0.2664334967715308, |
| "learning_rate": 5e-06, |
| "loss": 0.1517, |
| "step": 749 |
| }, |
| { |
| "epoch": 6.756756756756757, |
| "grad_norm": 0.28834080697901254, |
| "learning_rate": 5e-06, |
| "loss": 0.1422, |
| "step": 750 |
| }, |
| { |
| "epoch": 6.7657657657657655, |
| "grad_norm": 0.29170783307220777, |
| "learning_rate": 5e-06, |
| "loss": 0.1466, |
| "step": 751 |
| }, |
| { |
| "epoch": 6.774774774774775, |
| "grad_norm": 0.2802071333171322, |
| "learning_rate": 5e-06, |
| "loss": 0.2085, |
| "step": 752 |
| }, |
| { |
| "epoch": 6.783783783783784, |
| "grad_norm": 0.27926590125257916, |
| "learning_rate": 5e-06, |
| "loss": 0.1923, |
| "step": 753 |
| }, |
| { |
| "epoch": 6.792792792792793, |
| "grad_norm": 0.26518681255237936, |
| "learning_rate": 5e-06, |
| "loss": 0.1766, |
| "step": 754 |
| }, |
| { |
| "epoch": 6.801801801801802, |
| "grad_norm": 0.347354240325402, |
| "learning_rate": 5e-06, |
| "loss": 0.1825, |
| "step": 755 |
| }, |
| { |
| "epoch": 6.8108108108108105, |
| "grad_norm": 0.2783190286987182, |
| "learning_rate": 5e-06, |
| "loss": 0.169, |
| "step": 756 |
| }, |
| { |
| "epoch": 6.81981981981982, |
| "grad_norm": 0.25398065438526435, |
| "learning_rate": 5e-06, |
| "loss": 0.1586, |
| "step": 757 |
| }, |
| { |
| "epoch": 6.828828828828829, |
| "grad_norm": 0.2677100544625917, |
| "learning_rate": 5e-06, |
| "loss": 0.1422, |
| "step": 758 |
| }, |
| { |
| "epoch": 6.837837837837838, |
| "grad_norm": 0.25767020061357093, |
| "learning_rate": 5e-06, |
| "loss": 0.1562, |
| "step": 759 |
| }, |
| { |
| "epoch": 6.846846846846847, |
| "grad_norm": 0.2741476785712207, |
| "learning_rate": 5e-06, |
| "loss": 0.1672, |
| "step": 760 |
| }, |
| { |
| "epoch": 6.8558558558558556, |
| "grad_norm": 0.3040893781704407, |
| "learning_rate": 5e-06, |
| "loss": 0.1571, |
| "step": 761 |
| }, |
| { |
| "epoch": 6.864864864864865, |
| "grad_norm": 0.3190900071250576, |
| "learning_rate": 5e-06, |
| "loss": 0.1424, |
| "step": 762 |
| }, |
| { |
| "epoch": 6.873873873873874, |
| "grad_norm": 0.29513021972878545, |
| "learning_rate": 5e-06, |
| "loss": 0.1553, |
| "step": 763 |
| }, |
| { |
| "epoch": 6.882882882882883, |
| "grad_norm": 0.24584895713037455, |
| "learning_rate": 5e-06, |
| "loss": 0.1648, |
| "step": 764 |
| }, |
| { |
| "epoch": 6.891891891891892, |
| "grad_norm": 0.29616482928194166, |
| "learning_rate": 5e-06, |
| "loss": 0.1628, |
| "step": 765 |
| }, |
| { |
| "epoch": 6.900900900900901, |
| "grad_norm": 0.2698717209619555, |
| "learning_rate": 5e-06, |
| "loss": 0.1486, |
| "step": 766 |
| }, |
| { |
| "epoch": 6.90990990990991, |
| "grad_norm": 0.2842949721560408, |
| "learning_rate": 5e-06, |
| "loss": 0.1533, |
| "step": 767 |
| }, |
| { |
| "epoch": 6.918918918918919, |
| "grad_norm": 0.2548658270502879, |
| "learning_rate": 5e-06, |
| "loss": 0.1828, |
| "step": 768 |
| }, |
| { |
| "epoch": 6.927927927927928, |
| "grad_norm": 0.29650593537088255, |
| "learning_rate": 5e-06, |
| "loss": 0.1609, |
| "step": 769 |
| }, |
| { |
| "epoch": 6.936936936936937, |
| "grad_norm": 0.28258003962728084, |
| "learning_rate": 5e-06, |
| "loss": 0.1593, |
| "step": 770 |
| }, |
| { |
| "epoch": 6.945945945945946, |
| "grad_norm": 0.2635611766361993, |
| "learning_rate": 5e-06, |
| "loss": 0.1854, |
| "step": 771 |
| }, |
| { |
| "epoch": 6.954954954954955, |
| "grad_norm": 0.30424832142174796, |
| "learning_rate": 5e-06, |
| "loss": 0.146, |
| "step": 772 |
| }, |
| { |
| "epoch": 6.963963963963964, |
| "grad_norm": 0.3060917294529799, |
| "learning_rate": 5e-06, |
| "loss": 0.1733, |
| "step": 773 |
| }, |
| { |
| "epoch": 6.972972972972973, |
| "grad_norm": 0.32441537080653826, |
| "learning_rate": 5e-06, |
| "loss": 0.1537, |
| "step": 774 |
| }, |
| { |
| "epoch": 6.981981981981982, |
| "grad_norm": 0.2765707057627649, |
| "learning_rate": 5e-06, |
| "loss": 0.1693, |
| "step": 775 |
| }, |
| { |
| "epoch": 6.990990990990991, |
| "grad_norm": 0.2603233373640257, |
| "learning_rate": 5e-06, |
| "loss": 0.1796, |
| "step": 776 |
| }, |
| { |
| "epoch": 7.0, |
| "grad_norm": 0.2781503743653767, |
| "learning_rate": 5e-06, |
| "loss": 0.184, |
| "step": 777 |
| }, |
| { |
| "epoch": 7.009009009009009, |
| "grad_norm": 0.28846916522682, |
| "learning_rate": 5e-06, |
| "loss": 0.1327, |
| "step": 778 |
| }, |
| { |
| "epoch": 7.018018018018018, |
| "grad_norm": 0.2909211896726909, |
| "learning_rate": 5e-06, |
| "loss": 0.1436, |
| "step": 779 |
| }, |
| { |
| "epoch": 7.027027027027027, |
| "grad_norm": 0.29059846677673873, |
| "learning_rate": 5e-06, |
| "loss": 0.1264, |
| "step": 780 |
| }, |
| { |
| "epoch": 7.036036036036036, |
| "grad_norm": 0.2589502571701869, |
| "learning_rate": 5e-06, |
| "loss": 0.1433, |
| "step": 781 |
| }, |
| { |
| "epoch": 7.045045045045045, |
| "grad_norm": 0.30299652950475636, |
| "learning_rate": 5e-06, |
| "loss": 0.1139, |
| "step": 782 |
| }, |
| { |
| "epoch": 7.054054054054054, |
| "grad_norm": 0.3100291804047275, |
| "learning_rate": 5e-06, |
| "loss": 0.1592, |
| "step": 783 |
| }, |
| { |
| "epoch": 7.063063063063063, |
| "grad_norm": 0.2975691545809747, |
| "learning_rate": 5e-06, |
| "loss": 0.157, |
| "step": 784 |
| }, |
| { |
| "epoch": 7.072072072072072, |
| "grad_norm": 0.3068091793086541, |
| "learning_rate": 5e-06, |
| "loss": 0.1232, |
| "step": 785 |
| }, |
| { |
| "epoch": 7.081081081081081, |
| "grad_norm": 0.3089806282089307, |
| "learning_rate": 5e-06, |
| "loss": 0.1221, |
| "step": 786 |
| }, |
| { |
| "epoch": 7.09009009009009, |
| "grad_norm": 0.2847038649852651, |
| "learning_rate": 5e-06, |
| "loss": 0.1281, |
| "step": 787 |
| }, |
| { |
| "epoch": 7.099099099099099, |
| "grad_norm": 0.2615997547703096, |
| "learning_rate": 5e-06, |
| "loss": 0.1434, |
| "step": 788 |
| }, |
| { |
| "epoch": 7.108108108108108, |
| "grad_norm": 0.2807779495824356, |
| "learning_rate": 5e-06, |
| "loss": 0.1245, |
| "step": 789 |
| }, |
| { |
| "epoch": 7.117117117117117, |
| "grad_norm": 0.32154910892821653, |
| "learning_rate": 5e-06, |
| "loss": 0.1398, |
| "step": 790 |
| }, |
| { |
| "epoch": 7.126126126126126, |
| "grad_norm": 0.2909515213375792, |
| "learning_rate": 5e-06, |
| "loss": 0.1617, |
| "step": 791 |
| }, |
| { |
| "epoch": 7.135135135135135, |
| "grad_norm": 0.3077346518188213, |
| "learning_rate": 5e-06, |
| "loss": 0.1798, |
| "step": 792 |
| }, |
| { |
| "epoch": 7.1441441441441444, |
| "grad_norm": 0.2950118643236569, |
| "learning_rate": 5e-06, |
| "loss": 0.1189, |
| "step": 793 |
| }, |
| { |
| "epoch": 7.153153153153153, |
| "grad_norm": 0.3358905154061822, |
| "learning_rate": 5e-06, |
| "loss": 0.1145, |
| "step": 794 |
| }, |
| { |
| "epoch": 7.162162162162162, |
| "grad_norm": 0.2784628731056912, |
| "learning_rate": 5e-06, |
| "loss": 0.1413, |
| "step": 795 |
| }, |
| { |
| "epoch": 7.171171171171171, |
| "grad_norm": 0.332241278578818, |
| "learning_rate": 5e-06, |
| "loss": 0.141, |
| "step": 796 |
| }, |
| { |
| "epoch": 7.18018018018018, |
| "grad_norm": 0.3319493746279513, |
| "learning_rate": 5e-06, |
| "loss": 0.1265, |
| "step": 797 |
| }, |
| { |
| "epoch": 7.1891891891891895, |
| "grad_norm": 0.2918711891065202, |
| "learning_rate": 5e-06, |
| "loss": 0.1848, |
| "step": 798 |
| }, |
| { |
| "epoch": 7.198198198198198, |
| "grad_norm": 0.294971929143932, |
| "learning_rate": 5e-06, |
| "loss": 0.1102, |
| "step": 799 |
| }, |
| { |
| "epoch": 7.207207207207207, |
| "grad_norm": 0.28582274936569596, |
| "learning_rate": 5e-06, |
| "loss": 0.1364, |
| "step": 800 |
| }, |
| { |
| "epoch": 7.216216216216216, |
| "grad_norm": 0.2659870012399625, |
| "learning_rate": 5e-06, |
| "loss": 0.1402, |
| "step": 801 |
| }, |
| { |
| "epoch": 7.225225225225225, |
| "grad_norm": 0.27334027202909716, |
| "learning_rate": 5e-06, |
| "loss": 0.16, |
| "step": 802 |
| }, |
| { |
| "epoch": 7.2342342342342345, |
| "grad_norm": 0.29814746528630565, |
| "learning_rate": 5e-06, |
| "loss": 0.1313, |
| "step": 803 |
| }, |
| { |
| "epoch": 7.243243243243243, |
| "grad_norm": 0.30947580830786586, |
| "learning_rate": 5e-06, |
| "loss": 0.16, |
| "step": 804 |
| }, |
| { |
| "epoch": 7.252252252252252, |
| "grad_norm": 0.38586634200713993, |
| "learning_rate": 5e-06, |
| "loss": 0.1162, |
| "step": 805 |
| }, |
| { |
| "epoch": 7.261261261261261, |
| "grad_norm": 0.24887533724067495, |
| "learning_rate": 5e-06, |
| "loss": 0.1632, |
| "step": 806 |
| }, |
| { |
| "epoch": 7.27027027027027, |
| "grad_norm": 0.26959801719398596, |
| "learning_rate": 5e-06, |
| "loss": 0.152, |
| "step": 807 |
| }, |
| { |
| "epoch": 7.2792792792792795, |
| "grad_norm": 0.29035577733855455, |
| "learning_rate": 5e-06, |
| "loss": 0.1283, |
| "step": 808 |
| }, |
| { |
| "epoch": 7.288288288288288, |
| "grad_norm": 0.28290473947861045, |
| "learning_rate": 5e-06, |
| "loss": 0.1411, |
| "step": 809 |
| }, |
| { |
| "epoch": 7.297297297297297, |
| "grad_norm": 0.32523778027288563, |
| "learning_rate": 5e-06, |
| "loss": 0.1198, |
| "step": 810 |
| }, |
| { |
| "epoch": 7.306306306306306, |
| "grad_norm": 0.27833761872914975, |
| "learning_rate": 5e-06, |
| "loss": 0.1555, |
| "step": 811 |
| }, |
| { |
| "epoch": 7.315315315315315, |
| "grad_norm": 0.3367892879965876, |
| "learning_rate": 5e-06, |
| "loss": 0.1314, |
| "step": 812 |
| }, |
| { |
| "epoch": 7.324324324324325, |
| "grad_norm": 0.277313010335673, |
| "learning_rate": 5e-06, |
| "loss": 0.1364, |
| "step": 813 |
| }, |
| { |
| "epoch": 7.333333333333333, |
| "grad_norm": 0.27375407683101277, |
| "learning_rate": 5e-06, |
| "loss": 0.1331, |
| "step": 814 |
| }, |
| { |
| "epoch": 7.342342342342342, |
| "grad_norm": 0.26697608186333877, |
| "learning_rate": 5e-06, |
| "loss": 0.1458, |
| "step": 815 |
| }, |
| { |
| "epoch": 7.351351351351352, |
| "grad_norm": 0.2974176115893814, |
| "learning_rate": 5e-06, |
| "loss": 0.1394, |
| "step": 816 |
| }, |
| { |
| "epoch": 7.36036036036036, |
| "grad_norm": 0.2764189750660692, |
| "learning_rate": 5e-06, |
| "loss": 0.1421, |
| "step": 817 |
| }, |
| { |
| "epoch": 7.36936936936937, |
| "grad_norm": 0.26103798171790754, |
| "learning_rate": 5e-06, |
| "loss": 0.1277, |
| "step": 818 |
| }, |
| { |
| "epoch": 7.378378378378378, |
| "grad_norm": 0.2702444951963583, |
| "learning_rate": 5e-06, |
| "loss": 0.1248, |
| "step": 819 |
| }, |
| { |
| "epoch": 7.387387387387387, |
| "grad_norm": 0.3091293494651806, |
| "learning_rate": 5e-06, |
| "loss": 0.107, |
| "step": 820 |
| }, |
| { |
| "epoch": 7.396396396396397, |
| "grad_norm": 0.27618979276963385, |
| "learning_rate": 5e-06, |
| "loss": 0.1286, |
| "step": 821 |
| }, |
| { |
| "epoch": 7.405405405405405, |
| "grad_norm": 0.2850325224623505, |
| "learning_rate": 5e-06, |
| "loss": 0.1524, |
| "step": 822 |
| }, |
| { |
| "epoch": 7.414414414414415, |
| "grad_norm": 0.2699678297509792, |
| "learning_rate": 5e-06, |
| "loss": 0.1504, |
| "step": 823 |
| }, |
| { |
| "epoch": 7.423423423423423, |
| "grad_norm": 0.2684653976073876, |
| "learning_rate": 5e-06, |
| "loss": 0.1472, |
| "step": 824 |
| }, |
| { |
| "epoch": 7.4324324324324325, |
| "grad_norm": 0.27562604313146905, |
| "learning_rate": 5e-06, |
| "loss": 0.1647, |
| "step": 825 |
| }, |
| { |
| "epoch": 7.441441441441442, |
| "grad_norm": 0.32271326740564915, |
| "learning_rate": 5e-06, |
| "loss": 0.1373, |
| "step": 826 |
| }, |
| { |
| "epoch": 7.45045045045045, |
| "grad_norm": 0.3017797868640937, |
| "learning_rate": 5e-06, |
| "loss": 0.1177, |
| "step": 827 |
| }, |
| { |
| "epoch": 7.45945945945946, |
| "grad_norm": 0.3746544705141892, |
| "learning_rate": 5e-06, |
| "loss": 0.1266, |
| "step": 828 |
| }, |
| { |
| "epoch": 7.468468468468468, |
| "grad_norm": 0.36162531031963435, |
| "learning_rate": 5e-06, |
| "loss": 0.1723, |
| "step": 829 |
| }, |
| { |
| "epoch": 7.4774774774774775, |
| "grad_norm": 0.30931420713180063, |
| "learning_rate": 5e-06, |
| "loss": 0.1497, |
| "step": 830 |
| }, |
| { |
| "epoch": 7.486486486486487, |
| "grad_norm": 0.27294359139491653, |
| "learning_rate": 5e-06, |
| "loss": 0.1409, |
| "step": 831 |
| }, |
| { |
| "epoch": 7.495495495495495, |
| "grad_norm": 0.2643297412103037, |
| "learning_rate": 5e-06, |
| "loss": 0.1339, |
| "step": 832 |
| }, |
| { |
| "epoch": 7.504504504504505, |
| "grad_norm": 0.27159435674152455, |
| "learning_rate": 5e-06, |
| "loss": 0.1432, |
| "step": 833 |
| }, |
| { |
| "epoch": 7.513513513513513, |
| "grad_norm": 0.29349995019825675, |
| "learning_rate": 5e-06, |
| "loss": 0.1431, |
| "step": 834 |
| }, |
| { |
| "epoch": 7.5225225225225225, |
| "grad_norm": 0.31813922083817525, |
| "learning_rate": 5e-06, |
| "loss": 0.1237, |
| "step": 835 |
| }, |
| { |
| "epoch": 7.531531531531532, |
| "grad_norm": 0.25431234598026253, |
| "learning_rate": 5e-06, |
| "loss": 0.1231, |
| "step": 836 |
| }, |
| { |
| "epoch": 7.54054054054054, |
| "grad_norm": 0.26549876685780915, |
| "learning_rate": 5e-06, |
| "loss": 0.1079, |
| "step": 837 |
| }, |
| { |
| "epoch": 7.54954954954955, |
| "grad_norm": 0.2843904679866454, |
| "learning_rate": 5e-06, |
| "loss": 0.1456, |
| "step": 838 |
| }, |
| { |
| "epoch": 7.558558558558558, |
| "grad_norm": 0.2764008229289936, |
| "learning_rate": 5e-06, |
| "loss": 0.1294, |
| "step": 839 |
| }, |
| { |
| "epoch": 7.5675675675675675, |
| "grad_norm": 0.27550196745644295, |
| "learning_rate": 5e-06, |
| "loss": 0.1169, |
| "step": 840 |
| }, |
| { |
| "epoch": 7.576576576576577, |
| "grad_norm": 0.2657966726978357, |
| "learning_rate": 5e-06, |
| "loss": 0.1283, |
| "step": 841 |
| }, |
| { |
| "epoch": 7.585585585585585, |
| "grad_norm": 0.279726882287188, |
| "learning_rate": 5e-06, |
| "loss": 0.1151, |
| "step": 842 |
| }, |
| { |
| "epoch": 7.594594594594595, |
| "grad_norm": 0.27489297214494474, |
| "learning_rate": 5e-06, |
| "loss": 0.1515, |
| "step": 843 |
| }, |
| { |
| "epoch": 7.603603603603604, |
| "grad_norm": 0.277774516155618, |
| "learning_rate": 5e-06, |
| "loss": 0.1503, |
| "step": 844 |
| }, |
| { |
| "epoch": 7.612612612612613, |
| "grad_norm": 0.27480641761427765, |
| "learning_rate": 5e-06, |
| "loss": 0.1075, |
| "step": 845 |
| }, |
| { |
| "epoch": 7.621621621621622, |
| "grad_norm": 0.2846350822817088, |
| "learning_rate": 5e-06, |
| "loss": 0.1514, |
| "step": 846 |
| }, |
| { |
| "epoch": 7.63063063063063, |
| "grad_norm": 0.28181647241406504, |
| "learning_rate": 5e-06, |
| "loss": 0.1314, |
| "step": 847 |
| }, |
| { |
| "epoch": 7.63963963963964, |
| "grad_norm": 0.3260999375199513, |
| "learning_rate": 5e-06, |
| "loss": 0.138, |
| "step": 848 |
| }, |
| { |
| "epoch": 7.648648648648649, |
| "grad_norm": 0.26191958014959504, |
| "learning_rate": 5e-06, |
| "loss": 0.1258, |
| "step": 849 |
| }, |
| { |
| "epoch": 7.657657657657658, |
| "grad_norm": 0.3229999390728544, |
| "learning_rate": 5e-06, |
| "loss": 0.15, |
| "step": 850 |
| }, |
| { |
| "epoch": 7.666666666666667, |
| "grad_norm": 0.26451999606394416, |
| "learning_rate": 5e-06, |
| "loss": 0.1495, |
| "step": 851 |
| }, |
| { |
| "epoch": 7.675675675675675, |
| "grad_norm": 0.26990766634628194, |
| "learning_rate": 5e-06, |
| "loss": 0.136, |
| "step": 852 |
| }, |
| { |
| "epoch": 7.684684684684685, |
| "grad_norm": 0.3024735158151713, |
| "learning_rate": 5e-06, |
| "loss": 0.1506, |
| "step": 853 |
| }, |
| { |
| "epoch": 7.693693693693694, |
| "grad_norm": 0.2924088710387093, |
| "learning_rate": 5e-06, |
| "loss": 0.1327, |
| "step": 854 |
| }, |
| { |
| "epoch": 7.702702702702703, |
| "grad_norm": 0.3127786707020021, |
| "learning_rate": 5e-06, |
| "loss": 0.139, |
| "step": 855 |
| }, |
| { |
| "epoch": 7.711711711711712, |
| "grad_norm": 0.284180680266691, |
| "learning_rate": 5e-06, |
| "loss": 0.144, |
| "step": 856 |
| }, |
| { |
| "epoch": 7.7207207207207205, |
| "grad_norm": 0.28516328758030296, |
| "learning_rate": 5e-06, |
| "loss": 0.1428, |
| "step": 857 |
| }, |
| { |
| "epoch": 7.72972972972973, |
| "grad_norm": 0.2862117655255194, |
| "learning_rate": 5e-06, |
| "loss": 0.1579, |
| "step": 858 |
| }, |
| { |
| "epoch": 7.738738738738739, |
| "grad_norm": 0.3006220321307271, |
| "learning_rate": 5e-06, |
| "loss": 0.1603, |
| "step": 859 |
| }, |
| { |
| "epoch": 7.747747747747748, |
| "grad_norm": 0.29806627084091325, |
| "learning_rate": 5e-06, |
| "loss": 0.138, |
| "step": 860 |
| }, |
| { |
| "epoch": 7.756756756756757, |
| "grad_norm": 0.2947193709313226, |
| "learning_rate": 5e-06, |
| "loss": 0.1523, |
| "step": 861 |
| }, |
| { |
| "epoch": 7.7657657657657655, |
| "grad_norm": 0.27653789730185635, |
| "learning_rate": 5e-06, |
| "loss": 0.0974, |
| "step": 862 |
| }, |
| { |
| "epoch": 7.774774774774775, |
| "grad_norm": 0.3359884991528938, |
| "learning_rate": 5e-06, |
| "loss": 0.1498, |
| "step": 863 |
| }, |
| { |
| "epoch": 7.783783783783784, |
| "grad_norm": 0.3548503598800355, |
| "learning_rate": 5e-06, |
| "loss": 0.1437, |
| "step": 864 |
| }, |
| { |
| "epoch": 7.792792792792793, |
| "grad_norm": 0.2847477764044701, |
| "learning_rate": 5e-06, |
| "loss": 0.1393, |
| "step": 865 |
| }, |
| { |
| "epoch": 7.801801801801802, |
| "grad_norm": 0.3065450470724209, |
| "learning_rate": 5e-06, |
| "loss": 0.1575, |
| "step": 866 |
| }, |
| { |
| "epoch": 7.8108108108108105, |
| "grad_norm": 0.3012473965068844, |
| "learning_rate": 5e-06, |
| "loss": 0.1032, |
| "step": 867 |
| }, |
| { |
| "epoch": 7.81981981981982, |
| "grad_norm": 0.3010781272384048, |
| "learning_rate": 5e-06, |
| "loss": 0.1475, |
| "step": 868 |
| }, |
| { |
| "epoch": 7.828828828828829, |
| "grad_norm": 0.2802542506009257, |
| "learning_rate": 5e-06, |
| "loss": 0.1493, |
| "step": 869 |
| }, |
| { |
| "epoch": 7.837837837837838, |
| "grad_norm": 0.257772116445583, |
| "learning_rate": 5e-06, |
| "loss": 0.1712, |
| "step": 870 |
| }, |
| { |
| "epoch": 7.846846846846847, |
| "grad_norm": 0.2619565806462764, |
| "learning_rate": 5e-06, |
| "loss": 0.1511, |
| "step": 871 |
| }, |
| { |
| "epoch": 7.8558558558558556, |
| "grad_norm": 0.24721830116730928, |
| "learning_rate": 5e-06, |
| "loss": 0.11, |
| "step": 872 |
| }, |
| { |
| "epoch": 7.864864864864865, |
| "grad_norm": 0.2608797590307874, |
| "learning_rate": 5e-06, |
| "loss": 0.1434, |
| "step": 873 |
| }, |
| { |
| "epoch": 7.873873873873874, |
| "grad_norm": 0.3098206461383644, |
| "learning_rate": 5e-06, |
| "loss": 0.1244, |
| "step": 874 |
| }, |
| { |
| "epoch": 7.882882882882883, |
| "grad_norm": 0.2543474293117471, |
| "learning_rate": 5e-06, |
| "loss": 0.1484, |
| "step": 875 |
| }, |
| { |
| "epoch": 7.891891891891892, |
| "grad_norm": 0.275008000690447, |
| "learning_rate": 5e-06, |
| "loss": 0.1866, |
| "step": 876 |
| }, |
| { |
| "epoch": 7.900900900900901, |
| "grad_norm": 0.2574884831625064, |
| "learning_rate": 5e-06, |
| "loss": 0.1163, |
| "step": 877 |
| }, |
| { |
| "epoch": 7.90990990990991, |
| "grad_norm": 0.2735186605433472, |
| "learning_rate": 5e-06, |
| "loss": 0.181, |
| "step": 878 |
| }, |
| { |
| "epoch": 7.918918918918919, |
| "grad_norm": 0.274445320782787, |
| "learning_rate": 5e-06, |
| "loss": 0.1501, |
| "step": 879 |
| }, |
| { |
| "epoch": 7.927927927927928, |
| "grad_norm": 0.32592488504503725, |
| "learning_rate": 5e-06, |
| "loss": 0.149, |
| "step": 880 |
| }, |
| { |
| "epoch": 7.936936936936937, |
| "grad_norm": 0.2733852933643102, |
| "learning_rate": 5e-06, |
| "loss": 0.1434, |
| "step": 881 |
| }, |
| { |
| "epoch": 7.945945945945946, |
| "grad_norm": 0.2801042118029067, |
| "learning_rate": 5e-06, |
| "loss": 0.1348, |
| "step": 882 |
| }, |
| { |
| "epoch": 7.954954954954955, |
| "grad_norm": 0.27111497067523815, |
| "learning_rate": 5e-06, |
| "loss": 0.1597, |
| "step": 883 |
| }, |
| { |
| "epoch": 7.963963963963964, |
| "grad_norm": 0.27209799660825884, |
| "learning_rate": 5e-06, |
| "loss": 0.1521, |
| "step": 884 |
| }, |
| { |
| "epoch": 7.972972972972973, |
| "grad_norm": 0.3930775839000158, |
| "learning_rate": 5e-06, |
| "loss": 0.1559, |
| "step": 885 |
| }, |
| { |
| "epoch": 7.981981981981982, |
| "grad_norm": 0.2678257713350625, |
| "learning_rate": 5e-06, |
| "loss": 0.1541, |
| "step": 886 |
| }, |
| { |
| "epoch": 7.990990990990991, |
| "grad_norm": 0.2557726554508295, |
| "learning_rate": 5e-06, |
| "loss": 0.1293, |
| "step": 887 |
| }, |
| { |
| "epoch": 8.0, |
| "grad_norm": 0.30361837110340506, |
| "learning_rate": 5e-06, |
| "loss": 0.118, |
| "step": 888 |
| }, |
| { |
| "epoch": 8.00900900900901, |
| "grad_norm": 0.3724300858125284, |
| "learning_rate": 5e-06, |
| "loss": 0.119, |
| "step": 889 |
| }, |
| { |
| "epoch": 8.018018018018019, |
| "grad_norm": 0.278602815740515, |
| "learning_rate": 5e-06, |
| "loss": 0.1553, |
| "step": 890 |
| }, |
| { |
| "epoch": 8.027027027027026, |
| "grad_norm": 0.2757271522278008, |
| "learning_rate": 5e-06, |
| "loss": 0.1182, |
| "step": 891 |
| }, |
| { |
| "epoch": 8.036036036036036, |
| "grad_norm": 0.2929517079994954, |
| "learning_rate": 5e-06, |
| "loss": 0.1158, |
| "step": 892 |
| }, |
| { |
| "epoch": 8.045045045045045, |
| "grad_norm": 0.36372793530477965, |
| "learning_rate": 5e-06, |
| "loss": 0.1134, |
| "step": 893 |
| }, |
| { |
| "epoch": 8.054054054054054, |
| "grad_norm": 0.2817436945034752, |
| "learning_rate": 5e-06, |
| "loss": 0.0886, |
| "step": 894 |
| }, |
| { |
| "epoch": 8.063063063063064, |
| "grad_norm": 0.26617606652724307, |
| "learning_rate": 5e-06, |
| "loss": 0.1196, |
| "step": 895 |
| }, |
| { |
| "epoch": 8.072072072072071, |
| "grad_norm": 0.28867983451275775, |
| "learning_rate": 5e-06, |
| "loss": 0.1194, |
| "step": 896 |
| }, |
| { |
| "epoch": 8.08108108108108, |
| "grad_norm": 0.32961696028266857, |
| "learning_rate": 5e-06, |
| "loss": 0.1137, |
| "step": 897 |
| }, |
| { |
| "epoch": 8.09009009009009, |
| "grad_norm": 0.3261806017299068, |
| "learning_rate": 5e-06, |
| "loss": 0.1164, |
| "step": 898 |
| }, |
| { |
| "epoch": 8.0990990990991, |
| "grad_norm": 0.3047004644766596, |
| "learning_rate": 5e-06, |
| "loss": 0.1126, |
| "step": 899 |
| }, |
| { |
| "epoch": 8.108108108108109, |
| "grad_norm": 0.33711404617474894, |
| "learning_rate": 5e-06, |
| "loss": 0.1321, |
| "step": 900 |
| }, |
| { |
| "epoch": 8.117117117117116, |
| "grad_norm": 0.28781948503164106, |
| "learning_rate": 5e-06, |
| "loss": 0.0933, |
| "step": 901 |
| }, |
| { |
| "epoch": 8.126126126126126, |
| "grad_norm": 0.28954810553500565, |
| "learning_rate": 5e-06, |
| "loss": 0.1232, |
| "step": 902 |
| }, |
| { |
| "epoch": 8.135135135135135, |
| "grad_norm": 0.2934734159929239, |
| "learning_rate": 5e-06, |
| "loss": 0.1197, |
| "step": 903 |
| }, |
| { |
| "epoch": 8.144144144144144, |
| "grad_norm": 0.2851801626338276, |
| "learning_rate": 5e-06, |
| "loss": 0.1371, |
| "step": 904 |
| }, |
| { |
| "epoch": 8.153153153153154, |
| "grad_norm": 0.27920553933946246, |
| "learning_rate": 5e-06, |
| "loss": 0.1166, |
| "step": 905 |
| }, |
| { |
| "epoch": 8.162162162162161, |
| "grad_norm": 0.31343774042661315, |
| "learning_rate": 5e-06, |
| "loss": 0.1351, |
| "step": 906 |
| }, |
| { |
| "epoch": 8.17117117117117, |
| "grad_norm": 0.28444182940880375, |
| "learning_rate": 5e-06, |
| "loss": 0.1102, |
| "step": 907 |
| }, |
| { |
| "epoch": 8.18018018018018, |
| "grad_norm": 0.27773525186278575, |
| "learning_rate": 5e-06, |
| "loss": 0.1051, |
| "step": 908 |
| }, |
| { |
| "epoch": 8.18918918918919, |
| "grad_norm": 0.2822074526087139, |
| "learning_rate": 5e-06, |
| "loss": 0.1147, |
| "step": 909 |
| }, |
| { |
| "epoch": 8.198198198198199, |
| "grad_norm": 0.3445008346134781, |
| "learning_rate": 5e-06, |
| "loss": 0.1627, |
| "step": 910 |
| }, |
| { |
| "epoch": 8.207207207207206, |
| "grad_norm": 0.2948133646560585, |
| "learning_rate": 5e-06, |
| "loss": 0.1243, |
| "step": 911 |
| }, |
| { |
| "epoch": 8.216216216216216, |
| "grad_norm": 0.3548137336190423, |
| "learning_rate": 5e-06, |
| "loss": 0.1034, |
| "step": 912 |
| }, |
| { |
| "epoch": 8.225225225225225, |
| "grad_norm": 0.24835233982448732, |
| "learning_rate": 5e-06, |
| "loss": 0.1083, |
| "step": 913 |
| }, |
| { |
| "epoch": 8.234234234234235, |
| "grad_norm": 0.28608021315126475, |
| "learning_rate": 5e-06, |
| "loss": 0.0924, |
| "step": 914 |
| }, |
| { |
| "epoch": 8.243243243243244, |
| "grad_norm": 0.2489920981375781, |
| "learning_rate": 5e-06, |
| "loss": 0.0732, |
| "step": 915 |
| }, |
| { |
| "epoch": 8.252252252252251, |
| "grad_norm": 0.32648292494227393, |
| "learning_rate": 5e-06, |
| "loss": 0.101, |
| "step": 916 |
| }, |
| { |
| "epoch": 8.26126126126126, |
| "grad_norm": 0.26393238433265215, |
| "learning_rate": 5e-06, |
| "loss": 0.129, |
| "step": 917 |
| }, |
| { |
| "epoch": 8.27027027027027, |
| "grad_norm": 0.33000321897677853, |
| "learning_rate": 5e-06, |
| "loss": 0.1092, |
| "step": 918 |
| }, |
| { |
| "epoch": 8.27927927927928, |
| "grad_norm": 0.36101227626943727, |
| "learning_rate": 5e-06, |
| "loss": 0.1164, |
| "step": 919 |
| }, |
| { |
| "epoch": 8.288288288288289, |
| "grad_norm": 0.31480014766124803, |
| "learning_rate": 5e-06, |
| "loss": 0.1042, |
| "step": 920 |
| }, |
| { |
| "epoch": 8.297297297297296, |
| "grad_norm": 0.3411446684131361, |
| "learning_rate": 5e-06, |
| "loss": 0.1165, |
| "step": 921 |
| }, |
| { |
| "epoch": 8.306306306306306, |
| "grad_norm": 0.332814645717202, |
| "learning_rate": 5e-06, |
| "loss": 0.1078, |
| "step": 922 |
| }, |
| { |
| "epoch": 8.315315315315315, |
| "grad_norm": 0.5298958645427678, |
| "learning_rate": 5e-06, |
| "loss": 0.129, |
| "step": 923 |
| }, |
| { |
| "epoch": 8.324324324324325, |
| "grad_norm": 0.359036677437228, |
| "learning_rate": 5e-06, |
| "loss": 0.1297, |
| "step": 924 |
| }, |
| { |
| "epoch": 8.333333333333334, |
| "grad_norm": 0.3038540419282529, |
| "learning_rate": 5e-06, |
| "loss": 0.0996, |
| "step": 925 |
| }, |
| { |
| "epoch": 8.342342342342342, |
| "grad_norm": 0.2836788711371018, |
| "learning_rate": 5e-06, |
| "loss": 0.1064, |
| "step": 926 |
| }, |
| { |
| "epoch": 8.35135135135135, |
| "grad_norm": 0.2840741884034298, |
| "learning_rate": 5e-06, |
| "loss": 0.1068, |
| "step": 927 |
| }, |
| { |
| "epoch": 8.36036036036036, |
| "grad_norm": 0.3606217340289843, |
| "learning_rate": 5e-06, |
| "loss": 0.1161, |
| "step": 928 |
| }, |
| { |
| "epoch": 8.36936936936937, |
| "grad_norm": 0.33639738597690616, |
| "learning_rate": 5e-06, |
| "loss": 0.1172, |
| "step": 929 |
| }, |
| { |
| "epoch": 8.378378378378379, |
| "grad_norm": 0.34030156462417316, |
| "learning_rate": 5e-06, |
| "loss": 0.0924, |
| "step": 930 |
| }, |
| { |
| "epoch": 8.387387387387387, |
| "grad_norm": 0.2833799915899549, |
| "learning_rate": 5e-06, |
| "loss": 0.1433, |
| "step": 931 |
| }, |
| { |
| "epoch": 8.396396396396396, |
| "grad_norm": 0.29693654573168704, |
| "learning_rate": 5e-06, |
| "loss": 0.1192, |
| "step": 932 |
| }, |
| { |
| "epoch": 8.405405405405405, |
| "grad_norm": 0.2834266663051093, |
| "learning_rate": 5e-06, |
| "loss": 0.1289, |
| "step": 933 |
| }, |
| { |
| "epoch": 8.414414414414415, |
| "grad_norm": 0.29926603005274255, |
| "learning_rate": 5e-06, |
| "loss": 0.1238, |
| "step": 934 |
| }, |
| { |
| "epoch": 8.423423423423424, |
| "grad_norm": 0.27131987409072494, |
| "learning_rate": 5e-06, |
| "loss": 0.1137, |
| "step": 935 |
| }, |
| { |
| "epoch": 8.432432432432432, |
| "grad_norm": 0.27245579434724637, |
| "learning_rate": 5e-06, |
| "loss": 0.0845, |
| "step": 936 |
| }, |
| { |
| "epoch": 8.441441441441441, |
| "grad_norm": 0.29790096192104826, |
| "learning_rate": 5e-06, |
| "loss": 0.1455, |
| "step": 937 |
| }, |
| { |
| "epoch": 8.45045045045045, |
| "grad_norm": 0.30112523923077406, |
| "learning_rate": 5e-06, |
| "loss": 0.1243, |
| "step": 938 |
| }, |
| { |
| "epoch": 8.45945945945946, |
| "grad_norm": 0.2975979835875494, |
| "learning_rate": 5e-06, |
| "loss": 0.1124, |
| "step": 939 |
| }, |
| { |
| "epoch": 8.468468468468469, |
| "grad_norm": 0.325062595930168, |
| "learning_rate": 5e-06, |
| "loss": 0.125, |
| "step": 940 |
| }, |
| { |
| "epoch": 8.477477477477478, |
| "grad_norm": 0.316039949527555, |
| "learning_rate": 5e-06, |
| "loss": 0.1423, |
| "step": 941 |
| }, |
| { |
| "epoch": 8.486486486486486, |
| "grad_norm": 0.38300001583678706, |
| "learning_rate": 5e-06, |
| "loss": 0.1402, |
| "step": 942 |
| }, |
| { |
| "epoch": 8.495495495495495, |
| "grad_norm": 0.368069280096857, |
| "learning_rate": 5e-06, |
| "loss": 0.1394, |
| "step": 943 |
| }, |
| { |
| "epoch": 8.504504504504505, |
| "grad_norm": 0.2836701503469006, |
| "learning_rate": 5e-06, |
| "loss": 0.1212, |
| "step": 944 |
| }, |
| { |
| "epoch": 8.513513513513514, |
| "grad_norm": 0.28208846723398523, |
| "learning_rate": 5e-06, |
| "loss": 0.1176, |
| "step": 945 |
| }, |
| { |
| "epoch": 8.522522522522522, |
| "grad_norm": 0.3223580674636738, |
| "learning_rate": 5e-06, |
| "loss": 0.1085, |
| "step": 946 |
| }, |
| { |
| "epoch": 8.531531531531531, |
| "grad_norm": 0.2975838495736659, |
| "learning_rate": 5e-06, |
| "loss": 0.1388, |
| "step": 947 |
| }, |
| { |
| "epoch": 8.54054054054054, |
| "grad_norm": 0.2926384544838666, |
| "learning_rate": 5e-06, |
| "loss": 0.1446, |
| "step": 948 |
| }, |
| { |
| "epoch": 8.54954954954955, |
| "grad_norm": 0.253328197509007, |
| "learning_rate": 5e-06, |
| "loss": 0.1106, |
| "step": 949 |
| }, |
| { |
| "epoch": 8.558558558558559, |
| "grad_norm": 0.29078679203113755, |
| "learning_rate": 5e-06, |
| "loss": 0.147, |
| "step": 950 |
| }, |
| { |
| "epoch": 8.567567567567568, |
| "grad_norm": 0.3242982227120515, |
| "learning_rate": 5e-06, |
| "loss": 0.114, |
| "step": 951 |
| }, |
| { |
| "epoch": 8.576576576576576, |
| "grad_norm": 0.31916745252108, |
| "learning_rate": 5e-06, |
| "loss": 0.1467, |
| "step": 952 |
| }, |
| { |
| "epoch": 8.585585585585585, |
| "grad_norm": 0.30713744803388243, |
| "learning_rate": 5e-06, |
| "loss": 0.1165, |
| "step": 953 |
| }, |
| { |
| "epoch": 8.594594594594595, |
| "grad_norm": 0.31006975320296604, |
| "learning_rate": 5e-06, |
| "loss": 0.1293, |
| "step": 954 |
| }, |
| { |
| "epoch": 8.603603603603604, |
| "grad_norm": 0.2929902011566849, |
| "learning_rate": 5e-06, |
| "loss": 0.1134, |
| "step": 955 |
| }, |
| { |
| "epoch": 8.612612612612612, |
| "grad_norm": 0.26235543384430693, |
| "learning_rate": 5e-06, |
| "loss": 0.1053, |
| "step": 956 |
| }, |
| { |
| "epoch": 8.621621621621621, |
| "grad_norm": 0.310009086101237, |
| "learning_rate": 5e-06, |
| "loss": 0.1353, |
| "step": 957 |
| }, |
| { |
| "epoch": 8.63063063063063, |
| "grad_norm": 0.4012614816736551, |
| "learning_rate": 5e-06, |
| "loss": 0.1248, |
| "step": 958 |
| }, |
| { |
| "epoch": 8.63963963963964, |
| "grad_norm": 0.2922213629865694, |
| "learning_rate": 5e-06, |
| "loss": 0.1229, |
| "step": 959 |
| }, |
| { |
| "epoch": 8.64864864864865, |
| "grad_norm": 0.28311443461185315, |
| "learning_rate": 5e-06, |
| "loss": 0.1309, |
| "step": 960 |
| }, |
| { |
| "epoch": 8.657657657657658, |
| "grad_norm": 0.3882325365638355, |
| "learning_rate": 5e-06, |
| "loss": 0.134, |
| "step": 961 |
| }, |
| { |
| "epoch": 8.666666666666666, |
| "grad_norm": 0.3162063136020091, |
| "learning_rate": 5e-06, |
| "loss": 0.1166, |
| "step": 962 |
| }, |
| { |
| "epoch": 8.675675675675675, |
| "grad_norm": 0.3408953528226783, |
| "learning_rate": 5e-06, |
| "loss": 0.1289, |
| "step": 963 |
| }, |
| { |
| "epoch": 8.684684684684685, |
| "grad_norm": 0.2558993641119161, |
| "learning_rate": 5e-06, |
| "loss": 0.0915, |
| "step": 964 |
| }, |
| { |
| "epoch": 8.693693693693694, |
| "grad_norm": 0.3240415851794722, |
| "learning_rate": 5e-06, |
| "loss": 0.1283, |
| "step": 965 |
| }, |
| { |
| "epoch": 8.702702702702704, |
| "grad_norm": 0.2802378425767391, |
| "learning_rate": 5e-06, |
| "loss": 0.1462, |
| "step": 966 |
| }, |
| { |
| "epoch": 8.711711711711711, |
| "grad_norm": 0.32417830620793386, |
| "learning_rate": 5e-06, |
| "loss": 0.1332, |
| "step": 967 |
| }, |
| { |
| "epoch": 8.72072072072072, |
| "grad_norm": 0.27199769909957633, |
| "learning_rate": 5e-06, |
| "loss": 0.111, |
| "step": 968 |
| }, |
| { |
| "epoch": 8.72972972972973, |
| "grad_norm": 0.3143617324505027, |
| "learning_rate": 5e-06, |
| "loss": 0.1215, |
| "step": 969 |
| }, |
| { |
| "epoch": 8.73873873873874, |
| "grad_norm": 0.3294346785496534, |
| "learning_rate": 5e-06, |
| "loss": 0.1054, |
| "step": 970 |
| }, |
| { |
| "epoch": 8.747747747747749, |
| "grad_norm": 0.30155526964094, |
| "learning_rate": 5e-06, |
| "loss": 0.1222, |
| "step": 971 |
| }, |
| { |
| "epoch": 8.756756756756756, |
| "grad_norm": 0.3117950940446377, |
| "learning_rate": 5e-06, |
| "loss": 0.114, |
| "step": 972 |
| }, |
| { |
| "epoch": 8.765765765765765, |
| "grad_norm": 0.3392300144846105, |
| "learning_rate": 5e-06, |
| "loss": 0.1163, |
| "step": 973 |
| }, |
| { |
| "epoch": 8.774774774774775, |
| "grad_norm": 0.2977388674828091, |
| "learning_rate": 5e-06, |
| "loss": 0.1386, |
| "step": 974 |
| }, |
| { |
| "epoch": 8.783783783783784, |
| "grad_norm": 0.3088070527312149, |
| "learning_rate": 5e-06, |
| "loss": 0.1622, |
| "step": 975 |
| }, |
| { |
| "epoch": 8.792792792792794, |
| "grad_norm": 0.28460989793331587, |
| "learning_rate": 5e-06, |
| "loss": 0.1321, |
| "step": 976 |
| }, |
| { |
| "epoch": 8.801801801801801, |
| "grad_norm": 0.301114864385651, |
| "learning_rate": 5e-06, |
| "loss": 0.1084, |
| "step": 977 |
| }, |
| { |
| "epoch": 8.81081081081081, |
| "grad_norm": 0.3080454311504172, |
| "learning_rate": 5e-06, |
| "loss": 0.0958, |
| "step": 978 |
| }, |
| { |
| "epoch": 8.81981981981982, |
| "grad_norm": 0.35110800631668737, |
| "learning_rate": 5e-06, |
| "loss": 0.1272, |
| "step": 979 |
| }, |
| { |
| "epoch": 8.82882882882883, |
| "grad_norm": 0.27956217271886274, |
| "learning_rate": 5e-06, |
| "loss": 0.1327, |
| "step": 980 |
| }, |
| { |
| "epoch": 8.837837837837839, |
| "grad_norm": 0.32361648737642695, |
| "learning_rate": 5e-06, |
| "loss": 0.1214, |
| "step": 981 |
| }, |
| { |
| "epoch": 8.846846846846846, |
| "grad_norm": 0.33228900753392643, |
| "learning_rate": 5e-06, |
| "loss": 0.0843, |
| "step": 982 |
| }, |
| { |
| "epoch": 8.855855855855856, |
| "grad_norm": 0.32323655451004957, |
| "learning_rate": 5e-06, |
| "loss": 0.117, |
| "step": 983 |
| }, |
| { |
| "epoch": 8.864864864864865, |
| "grad_norm": 0.3411596228576446, |
| "learning_rate": 5e-06, |
| "loss": 0.1137, |
| "step": 984 |
| }, |
| { |
| "epoch": 8.873873873873874, |
| "grad_norm": 0.33618399554078643, |
| "learning_rate": 5e-06, |
| "loss": 0.1232, |
| "step": 985 |
| }, |
| { |
| "epoch": 8.882882882882884, |
| "grad_norm": 0.31438715504842607, |
| "learning_rate": 5e-06, |
| "loss": 0.1188, |
| "step": 986 |
| }, |
| { |
| "epoch": 8.891891891891891, |
| "grad_norm": 0.29235691269480235, |
| "learning_rate": 5e-06, |
| "loss": 0.0907, |
| "step": 987 |
| }, |
| { |
| "epoch": 8.9009009009009, |
| "grad_norm": 0.31569994309412647, |
| "learning_rate": 5e-06, |
| "loss": 0.139, |
| "step": 988 |
| }, |
| { |
| "epoch": 8.90990990990991, |
| "grad_norm": 0.3488819032640533, |
| "learning_rate": 5e-06, |
| "loss": 0.0925, |
| "step": 989 |
| }, |
| { |
| "epoch": 8.91891891891892, |
| "grad_norm": 0.3287782461836467, |
| "learning_rate": 5e-06, |
| "loss": 0.1144, |
| "step": 990 |
| }, |
| { |
| "epoch": 8.927927927927929, |
| "grad_norm": 0.342018883415981, |
| "learning_rate": 5e-06, |
| "loss": 0.1554, |
| "step": 991 |
| }, |
| { |
| "epoch": 8.936936936936936, |
| "grad_norm": 0.30922097124521764, |
| "learning_rate": 5e-06, |
| "loss": 0.0773, |
| "step": 992 |
| }, |
| { |
| "epoch": 8.945945945945946, |
| "grad_norm": 0.3170605884389048, |
| "learning_rate": 5e-06, |
| "loss": 0.0961, |
| "step": 993 |
| }, |
| { |
| "epoch": 8.954954954954955, |
| "grad_norm": 0.2894394712507756, |
| "learning_rate": 5e-06, |
| "loss": 0.1175, |
| "step": 994 |
| }, |
| { |
| "epoch": 8.963963963963964, |
| "grad_norm": 0.3507612251253632, |
| "learning_rate": 5e-06, |
| "loss": 0.0898, |
| "step": 995 |
| }, |
| { |
| "epoch": 8.972972972972974, |
| "grad_norm": 0.2916461407037756, |
| "learning_rate": 5e-06, |
| "loss": 0.1162, |
| "step": 996 |
| }, |
| { |
| "epoch": 8.981981981981981, |
| "grad_norm": 0.3148022366204299, |
| "learning_rate": 5e-06, |
| "loss": 0.1263, |
| "step": 997 |
| }, |
| { |
| "epoch": 8.99099099099099, |
| "grad_norm": 0.2917019835417808, |
| "learning_rate": 5e-06, |
| "loss": 0.1088, |
| "step": 998 |
| }, |
| { |
| "epoch": 9.0, |
| "grad_norm": 0.27462543957892144, |
| "learning_rate": 5e-06, |
| "loss": 0.1158, |
| "step": 999 |
| } |
| ], |
| "logging_steps": 1.0, |
| "max_steps": 11100, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 100, |
| "save_steps": 500, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": false |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 1278638161920000.0, |
| "train_batch_size": 1, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|