| { |
| "best_global_step": null, |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 2.9890909090909092, |
| "eval_steps": 500, |
| "global_step": 153, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.019393939393939394, |
| "grad_norm": 87.14871215820312, |
| "learning_rate": 0.0, |
| "loss": 2.7491, |
| "step": 1 |
| }, |
| { |
| "epoch": 0.03878787878787879, |
| "grad_norm": 88.4898452758789, |
| "learning_rate": 4e-05, |
| "loss": 2.7719, |
| "step": 2 |
| }, |
| { |
| "epoch": 0.05818181818181818, |
| "grad_norm": 55.55705642700195, |
| "learning_rate": 8e-05, |
| "loss": 1.3445, |
| "step": 3 |
| }, |
| { |
| "epoch": 0.07757575757575758, |
| "grad_norm": 68.65280151367188, |
| "learning_rate": 0.00012, |
| "loss": 2.8264, |
| "step": 4 |
| }, |
| { |
| "epoch": 0.09696969696969697, |
| "grad_norm": 60.984527587890625, |
| "learning_rate": 0.00016, |
| "loss": 1.2853, |
| "step": 5 |
| }, |
| { |
| "epoch": 0.11636363636363636, |
| "grad_norm": 79.5303955078125, |
| "learning_rate": 0.0002, |
| "loss": 1.5542, |
| "step": 6 |
| }, |
| { |
| "epoch": 0.13575757575757577, |
| "grad_norm": 122.48136138916016, |
| "learning_rate": 0.00019997747161747695, |
| "loss": 4.8128, |
| "step": 7 |
| }, |
| { |
| "epoch": 0.15515151515151515, |
| "grad_norm": 115.57413482666016, |
| "learning_rate": 0.00019990989662046818, |
| "loss": 7.7786, |
| "step": 8 |
| }, |
| { |
| "epoch": 0.17454545454545456, |
| "grad_norm": 106.0721664428711, |
| "learning_rate": 0.00019979730545608126, |
| "loss": 14.3271, |
| "step": 9 |
| }, |
| { |
| "epoch": 0.19393939393939394, |
| "grad_norm": 187.45887756347656, |
| "learning_rate": 0.00019963974885425266, |
| "loss": 10.5438, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.21333333333333335, |
| "grad_norm": 67.8653564453125, |
| "learning_rate": 0.00019943729780489027, |
| "loss": 8.8936, |
| "step": 11 |
| }, |
| { |
| "epoch": 0.23272727272727273, |
| "grad_norm": 53.751426696777344, |
| "learning_rate": 0.00019919004352588767, |
| "loss": 6.9044, |
| "step": 12 |
| }, |
| { |
| "epoch": 0.25212121212121213, |
| "grad_norm": 68.24635314941406, |
| "learning_rate": 0.00019889809742202455, |
| "loss": 6.6015, |
| "step": 13 |
| }, |
| { |
| "epoch": 0.27151515151515154, |
| "grad_norm": 49.91704559326172, |
| "learning_rate": 0.00019856159103477086, |
| "loss": 4.7958, |
| "step": 14 |
| }, |
| { |
| "epoch": 0.2909090909090909, |
| "grad_norm": 84.79557800292969, |
| "learning_rate": 0.0001981806759830189, |
| "loss": 4.4889, |
| "step": 15 |
| }, |
| { |
| "epoch": 0.3103030303030303, |
| "grad_norm": 42.62947463989258, |
| "learning_rate": 0.00019775552389476864, |
| "loss": 3.852, |
| "step": 16 |
| }, |
| { |
| "epoch": 0.3296969696969697, |
| "grad_norm": 56.70987319946289, |
| "learning_rate": 0.00019728632632979746, |
| "loss": 3.7326, |
| "step": 17 |
| }, |
| { |
| "epoch": 0.3490909090909091, |
| "grad_norm": 48.838932037353516, |
| "learning_rate": 0.0001967732946933499, |
| "loss": 3.5072, |
| "step": 18 |
| }, |
| { |
| "epoch": 0.36848484848484847, |
| "grad_norm": 36.361454010009766, |
| "learning_rate": 0.00019621666014088494, |
| "loss": 3.3605, |
| "step": 19 |
| }, |
| { |
| "epoch": 0.3878787878787879, |
| "grad_norm": 42.21967697143555, |
| "learning_rate": 0.00019561667347392508, |
| "loss": 3.405, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.4072727272727273, |
| "grad_norm": 38.328067779541016, |
| "learning_rate": 0.0001949736050270532, |
| "loss": 3.2796, |
| "step": 21 |
| }, |
| { |
| "epoch": 0.4266666666666667, |
| "grad_norm": 35.21055221557617, |
| "learning_rate": 0.00019428774454610843, |
| "loss": 3.1108, |
| "step": 22 |
| }, |
| { |
| "epoch": 0.44606060606060605, |
| "grad_norm": 36.039955139160156, |
| "learning_rate": 0.0001935594010576362, |
| "loss": 2.989, |
| "step": 23 |
| }, |
| { |
| "epoch": 0.46545454545454545, |
| "grad_norm": 46.623226165771484, |
| "learning_rate": 0.00019278890272965096, |
| "loss": 2.9046, |
| "step": 24 |
| }, |
| { |
| "epoch": 0.48484848484848486, |
| "grad_norm": 42.583335876464844, |
| "learning_rate": 0.0001919765967237739, |
| "loss": 2.7366, |
| "step": 25 |
| }, |
| { |
| "epoch": 0.5042424242424243, |
| "grad_norm": 27.625574111938477, |
| "learning_rate": 0.0001911228490388136, |
| "loss": 2.538, |
| "step": 26 |
| }, |
| { |
| "epoch": 0.5236363636363637, |
| "grad_norm": 22.885229110717773, |
| "learning_rate": 0.00019022804434585852, |
| "loss": 2.4008, |
| "step": 27 |
| }, |
| { |
| "epoch": 0.5430303030303031, |
| "grad_norm": 25.352842330932617, |
| "learning_rate": 0.00018929258581495685, |
| "loss": 2.2888, |
| "step": 28 |
| }, |
| { |
| "epoch": 0.5624242424242424, |
| "grad_norm": 25.622007369995117, |
| "learning_rate": 0.00018831689493346095, |
| "loss": 2.1831, |
| "step": 29 |
| }, |
| { |
| "epoch": 0.5818181818181818, |
| "grad_norm": 24.7369327545166, |
| "learning_rate": 0.00018730141131611882, |
| "loss": 2.0357, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.6012121212121212, |
| "grad_norm": 24.530712127685547, |
| "learning_rate": 0.00018624659250699805, |
| "loss": 1.91, |
| "step": 31 |
| }, |
| { |
| "epoch": 0.6206060606060606, |
| "grad_norm": 21.335567474365234, |
| "learning_rate": 0.00018515291377333112, |
| "loss": 1.7888, |
| "step": 32 |
| }, |
| { |
| "epoch": 0.64, |
| "grad_norm": 21.168838500976562, |
| "learning_rate": 0.00018402086789137546, |
| "loss": 1.6856, |
| "step": 33 |
| }, |
| { |
| "epoch": 0.6593939393939394, |
| "grad_norm": 22.99504280090332, |
| "learning_rate": 0.00018285096492438424, |
| "loss": 1.5913, |
| "step": 34 |
| }, |
| { |
| "epoch": 0.6787878787878788, |
| "grad_norm": 19.430891036987305, |
| "learning_rate": 0.00018164373199278856, |
| "loss": 1.5296, |
| "step": 35 |
| }, |
| { |
| "epoch": 0.6981818181818182, |
| "grad_norm": 17.32403564453125, |
| "learning_rate": 0.00018039971303669407, |
| "loss": 1.4921, |
| "step": 36 |
| }, |
| { |
| "epoch": 0.7175757575757575, |
| "grad_norm": 13.494332313537598, |
| "learning_rate": 0.00017911946857079888, |
| "loss": 1.4273, |
| "step": 37 |
| }, |
| { |
| "epoch": 0.7369696969696969, |
| "grad_norm": 13.716304779052734, |
| "learning_rate": 0.00017780357543184397, |
| "loss": 1.3661, |
| "step": 38 |
| }, |
| { |
| "epoch": 0.7563636363636363, |
| "grad_norm": 18.049741744995117, |
| "learning_rate": 0.00017645262651870926, |
| "loss": 1.3225, |
| "step": 39 |
| }, |
| { |
| "epoch": 0.7757575757575758, |
| "grad_norm": 12.06884765625, |
| "learning_rate": 0.00017506723052527242, |
| "loss": 1.2614, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.7951515151515152, |
| "grad_norm": 10.950291633605957, |
| "learning_rate": 0.00017364801166615124, |
| "loss": 1.2235, |
| "step": 41 |
| }, |
| { |
| "epoch": 0.8145454545454546, |
| "grad_norm": 10.462716102600098, |
| "learning_rate": 0.00017219560939545246, |
| "loss": 1.1887, |
| "step": 42 |
| }, |
| { |
| "epoch": 0.833939393939394, |
| "grad_norm": 14.221378326416016, |
| "learning_rate": 0.00017071067811865476, |
| "loss": 1.1758, |
| "step": 43 |
| }, |
| { |
| "epoch": 0.8533333333333334, |
| "grad_norm": 9.887701034545898, |
| "learning_rate": 0.00016919388689775464, |
| "loss": 1.1214, |
| "step": 44 |
| }, |
| { |
| "epoch": 0.8727272727272727, |
| "grad_norm": 13.127241134643555, |
| "learning_rate": 0.0001676459191498087, |
| "loss": 1.0835, |
| "step": 45 |
| }, |
| { |
| "epoch": 0.8921212121212121, |
| "grad_norm": 16.6353816986084, |
| "learning_rate": 0.00016606747233900815, |
| "loss": 1.0416, |
| "step": 46 |
| }, |
| { |
| "epoch": 0.9115151515151515, |
| "grad_norm": 14.955666542053223, |
| "learning_rate": 0.00016445925766242391, |
| "loss": 1.0035, |
| "step": 47 |
| }, |
| { |
| "epoch": 0.9309090909090909, |
| "grad_norm": 9.893986701965332, |
| "learning_rate": 0.00016282199972956425, |
| "loss": 0.9478, |
| "step": 48 |
| }, |
| { |
| "epoch": 0.9503030303030303, |
| "grad_norm": 7.6303606033325195, |
| "learning_rate": 0.00016115643623588915, |
| "loss": 0.9117, |
| "step": 49 |
| }, |
| { |
| "epoch": 0.9696969696969697, |
| "grad_norm": 6.176935195922852, |
| "learning_rate": 0.00015946331763042867, |
| "loss": 0.8864, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.9890909090909091, |
| "grad_norm": 5.813427448272705, |
| "learning_rate": 0.0001577434067776548, |
| "loss": 0.8923, |
| "step": 51 |
| }, |
| { |
| "epoch": 1.0193939393939393, |
| "grad_norm": 5.501270294189453, |
| "learning_rate": 0.00015599747861375955, |
| "loss": 1.1926, |
| "step": 52 |
| }, |
| { |
| "epoch": 1.0387878787878788, |
| "grad_norm": 6.7219743728637695, |
| "learning_rate": 0.00015422631979749354, |
| "loss": 0.8721, |
| "step": 53 |
| }, |
| { |
| "epoch": 1.0581818181818181, |
| "grad_norm": 5.997743606567383, |
| "learning_rate": 0.00015243072835572318, |
| "loss": 0.8581, |
| "step": 54 |
| }, |
| { |
| "epoch": 1.0775757575757576, |
| "grad_norm": 6.01072359085083, |
| "learning_rate": 0.00015061151332386566, |
| "loss": 0.8375, |
| "step": 55 |
| }, |
| { |
| "epoch": 1.096969696969697, |
| "grad_norm": 7.074451923370361, |
| "learning_rate": 0.00014876949438136347, |
| "loss": 0.8521, |
| "step": 56 |
| }, |
| { |
| "epoch": 1.1163636363636364, |
| "grad_norm": 6.878782272338867, |
| "learning_rate": 0.0001469055014823637, |
| "loss": 0.8376, |
| "step": 57 |
| }, |
| { |
| "epoch": 1.1357575757575757, |
| "grad_norm": 7.401325225830078, |
| "learning_rate": 0.00014502037448176734, |
| "loss": 0.8258, |
| "step": 58 |
| }, |
| { |
| "epoch": 1.1551515151515153, |
| "grad_norm": 6.986009120941162, |
| "learning_rate": 0.00014311496275681783, |
| "loss": 0.8152, |
| "step": 59 |
| }, |
| { |
| "epoch": 1.1745454545454546, |
| "grad_norm": 3.373999834060669, |
| "learning_rate": 0.0001411901248243993, |
| "loss": 0.8212, |
| "step": 60 |
| }, |
| { |
| "epoch": 1.1939393939393939, |
| "grad_norm": 4.864870548248291, |
| "learning_rate": 0.00013924672795421637, |
| "loss": 0.8272, |
| "step": 61 |
| }, |
| { |
| "epoch": 1.2133333333333334, |
| "grad_norm": 5.978538513183594, |
| "learning_rate": 0.00013728564777803088, |
| "loss": 0.8253, |
| "step": 62 |
| }, |
| { |
| "epoch": 1.2327272727272727, |
| "grad_norm": 5.256051540374756, |
| "learning_rate": 0.0001353077678951301, |
| "loss": 0.8115, |
| "step": 63 |
| }, |
| { |
| "epoch": 1.2521212121212122, |
| "grad_norm": 4.685987949371338, |
| "learning_rate": 0.00013331397947420576, |
| "loss": 0.806, |
| "step": 64 |
| }, |
| { |
| "epoch": 1.2715151515151515, |
| "grad_norm": 5.210947513580322, |
| "learning_rate": 0.00013130518085182225, |
| "loss": 0.8114, |
| "step": 65 |
| }, |
| { |
| "epoch": 1.290909090909091, |
| "grad_norm": 7.189337730407715, |
| "learning_rate": 0.00012928227712765504, |
| "loss": 0.8136, |
| "step": 66 |
| }, |
| { |
| "epoch": 1.3103030303030303, |
| "grad_norm": 4.788082599639893, |
| "learning_rate": 0.0001272461797566823, |
| "loss": 0.798, |
| "step": 67 |
| }, |
| { |
| "epoch": 1.3296969696969696, |
| "grad_norm": 6.347202301025391, |
| "learning_rate": 0.00012519780613851254, |
| "loss": 0.8164, |
| "step": 68 |
| }, |
| { |
| "epoch": 1.3490909090909091, |
| "grad_norm": 4.3109307289123535, |
| "learning_rate": 0.00012313807920403419, |
| "loss": 0.7968, |
| "step": 69 |
| }, |
| { |
| "epoch": 1.3684848484848484, |
| "grad_norm": 4.202847480773926, |
| "learning_rate": 0.00012106792699957263, |
| "loss": 0.8084, |
| "step": 70 |
| }, |
| { |
| "epoch": 1.387878787878788, |
| "grad_norm": 5.670694351196289, |
| "learning_rate": 0.00011898828226874284, |
| "loss": 0.8105, |
| "step": 71 |
| }, |
| { |
| "epoch": 1.4072727272727272, |
| "grad_norm": 6.197960376739502, |
| "learning_rate": 0.00011690008203218493, |
| "loss": 0.8075, |
| "step": 72 |
| }, |
| { |
| "epoch": 1.4266666666666667, |
| "grad_norm": 5.756638526916504, |
| "learning_rate": 0.00011480426716537315, |
| "loss": 0.791, |
| "step": 73 |
| }, |
| { |
| "epoch": 1.446060606060606, |
| "grad_norm": 4.505667209625244, |
| "learning_rate": 0.00011270178197468789, |
| "loss": 0.8066, |
| "step": 74 |
| }, |
| { |
| "epoch": 1.4654545454545453, |
| "grad_norm": 4.306074142456055, |
| "learning_rate": 0.00011059357377194161, |
| "loss": 0.7946, |
| "step": 75 |
| }, |
| { |
| "epoch": 1.4848484848484849, |
| "grad_norm": 5.666556358337402, |
| "learning_rate": 0.00010848059244755093, |
| "loss": 0.8056, |
| "step": 76 |
| }, |
| { |
| "epoch": 1.5042424242424244, |
| "grad_norm": 4.2490315437316895, |
| "learning_rate": 0.00010636379004254664, |
| "loss": 0.8098, |
| "step": 77 |
| }, |
| { |
| "epoch": 1.5236363636363637, |
| "grad_norm": 4.633942604064941, |
| "learning_rate": 0.00010424412031961484, |
| "loss": 0.8032, |
| "step": 78 |
| }, |
| { |
| "epoch": 1.543030303030303, |
| "grad_norm": 4.449273109436035, |
| "learning_rate": 0.00010212253833336237, |
| "loss": 0.7973, |
| "step": 79 |
| }, |
| { |
| "epoch": 1.5624242424242425, |
| "grad_norm": 4.270414352416992, |
| "learning_rate": 0.0001, |
| "loss": 0.8055, |
| "step": 80 |
| }, |
| { |
| "epoch": 1.5818181818181818, |
| "grad_norm": 3.798518657684326, |
| "learning_rate": 9.787746166663764e-05, |
| "loss": 0.7913, |
| "step": 81 |
| }, |
| { |
| "epoch": 1.601212121212121, |
| "grad_norm": 4.9506425857543945, |
| "learning_rate": 9.57558796803852e-05, |
| "loss": 0.7878, |
| "step": 82 |
| }, |
| { |
| "epoch": 1.6206060606060606, |
| "grad_norm": 2.22216534614563, |
| "learning_rate": 9.363620995745337e-05, |
| "loss": 0.7962, |
| "step": 83 |
| }, |
| { |
| "epoch": 1.6400000000000001, |
| "grad_norm": 4.047158241271973, |
| "learning_rate": 9.151940755244912e-05, |
| "loss": 0.7892, |
| "step": 84 |
| }, |
| { |
| "epoch": 1.6593939393939394, |
| "grad_norm": 3.936147689819336, |
| "learning_rate": 8.94064262280584e-05, |
| "loss": 0.7937, |
| "step": 85 |
| }, |
| { |
| "epoch": 1.6787878787878787, |
| "grad_norm": 2.8039066791534424, |
| "learning_rate": 8.729821802531212e-05, |
| "loss": 0.7902, |
| "step": 86 |
| }, |
| { |
| "epoch": 1.6981818181818182, |
| "grad_norm": 3.999570608139038, |
| "learning_rate": 8.519573283462687e-05, |
| "loss": 0.7902, |
| "step": 87 |
| }, |
| { |
| "epoch": 1.7175757575757575, |
| "grad_norm": 6.398386478424072, |
| "learning_rate": 8.309991796781511e-05, |
| "loss": 0.7874, |
| "step": 88 |
| }, |
| { |
| "epoch": 1.7369696969696968, |
| "grad_norm": 3.3710646629333496, |
| "learning_rate": 8.101171773125716e-05, |
| "loss": 0.8017, |
| "step": 89 |
| }, |
| { |
| "epoch": 1.7563636363636363, |
| "grad_norm": 4.403554439544678, |
| "learning_rate": 7.89320730004274e-05, |
| "loss": 0.7909, |
| "step": 90 |
| }, |
| { |
| "epoch": 1.7757575757575759, |
| "grad_norm": 3.6332623958587646, |
| "learning_rate": 7.686192079596586e-05, |
| "loss": 0.7914, |
| "step": 91 |
| }, |
| { |
| "epoch": 1.7951515151515152, |
| "grad_norm": 2.992091655731201, |
| "learning_rate": 7.48021938614875e-05, |
| "loss": 0.8005, |
| "step": 92 |
| }, |
| { |
| "epoch": 1.8145454545454545, |
| "grad_norm": 3.749993324279785, |
| "learning_rate": 7.275382024331772e-05, |
| "loss": 0.7916, |
| "step": 93 |
| }, |
| { |
| "epoch": 1.833939393939394, |
| "grad_norm": 5.296722888946533, |
| "learning_rate": 7.071772287234497e-05, |
| "loss": 0.7968, |
| "step": 94 |
| }, |
| { |
| "epoch": 1.8533333333333335, |
| "grad_norm": 3.761016607284546, |
| "learning_rate": 6.869481914817779e-05, |
| "loss": 0.7856, |
| "step": 95 |
| }, |
| { |
| "epoch": 1.8727272727272726, |
| "grad_norm": 2.1738433837890625, |
| "learning_rate": 6.668602052579424e-05, |
| "loss": 0.7907, |
| "step": 96 |
| }, |
| { |
| "epoch": 1.892121212121212, |
| "grad_norm": 4.787725448608398, |
| "learning_rate": 6.469223210486992e-05, |
| "loss": 0.7969, |
| "step": 97 |
| }, |
| { |
| "epoch": 1.9115151515151516, |
| "grad_norm": 3.5583176612854004, |
| "learning_rate": 6.271435222196916e-05, |
| "loss": 0.789, |
| "step": 98 |
| }, |
| { |
| "epoch": 1.930909090909091, |
| "grad_norm": 3.7247867584228516, |
| "learning_rate": 6.0753272045783625e-05, |
| "loss": 0.7984, |
| "step": 99 |
| }, |
| { |
| "epoch": 1.9503030303030302, |
| "grad_norm": 3.7892508506774902, |
| "learning_rate": 5.880987517560075e-05, |
| "loss": 0.7951, |
| "step": 100 |
| }, |
| { |
| "epoch": 1.9696969696969697, |
| "grad_norm": 3.6910500526428223, |
| "learning_rate": 5.688503724318217e-05, |
| "loss": 0.7995, |
| "step": 101 |
| }, |
| { |
| "epoch": 1.9890909090909092, |
| "grad_norm": 3.318086862564087, |
| "learning_rate": 5.497962551823266e-05, |
| "loss": 0.7816, |
| "step": 102 |
| }, |
| { |
| "epoch": 2.0193939393939395, |
| "grad_norm": 4.396894454956055, |
| "learning_rate": 5.309449851763633e-05, |
| "loss": 1.092, |
| "step": 103 |
| }, |
| { |
| "epoch": 2.0387878787878786, |
| "grad_norm": 2.913602828979492, |
| "learning_rate": 5.123050561863657e-05, |
| "loss": 0.7868, |
| "step": 104 |
| }, |
| { |
| "epoch": 2.058181818181818, |
| "grad_norm": 5.820463180541992, |
| "learning_rate": 4.938848667613436e-05, |
| "loss": 0.792, |
| "step": 105 |
| }, |
| { |
| "epoch": 2.0775757575757576, |
| "grad_norm": 3.3575682640075684, |
| "learning_rate": 4.756927164427685e-05, |
| "loss": 0.7894, |
| "step": 106 |
| }, |
| { |
| "epoch": 2.096969696969697, |
| "grad_norm": 4.6300458908081055, |
| "learning_rate": 4.57736802025065e-05, |
| "loss": 0.7882, |
| "step": 107 |
| }, |
| { |
| "epoch": 2.1163636363636362, |
| "grad_norm": 3.036693572998047, |
| "learning_rate": 4.4002521386240466e-05, |
| "loss": 0.7934, |
| "step": 108 |
| }, |
| { |
| "epoch": 2.1357575757575757, |
| "grad_norm": 3.7198660373687744, |
| "learning_rate": 4.2256593222345185e-05, |
| "loss": 0.7864, |
| "step": 109 |
| }, |
| { |
| "epoch": 2.1551515151515153, |
| "grad_norm": 3.7882423400878906, |
| "learning_rate": 4.053668236957134e-05, |
| "loss": 0.785, |
| "step": 110 |
| }, |
| { |
| "epoch": 2.174545454545455, |
| "grad_norm": 2.843419075012207, |
| "learning_rate": 3.884356376411089e-05, |
| "loss": 0.7836, |
| "step": 111 |
| }, |
| { |
| "epoch": 2.193939393939394, |
| "grad_norm": 2.6506614685058594, |
| "learning_rate": 3.717800027043576e-05, |
| "loss": 0.7894, |
| "step": 112 |
| }, |
| { |
| "epoch": 2.2133333333333334, |
| "grad_norm": 3.587935447692871, |
| "learning_rate": 3.554074233757608e-05, |
| "loss": 0.7884, |
| "step": 113 |
| }, |
| { |
| "epoch": 2.232727272727273, |
| "grad_norm": 3.2883381843566895, |
| "learning_rate": 3.393252766099187e-05, |
| "loss": 0.7926, |
| "step": 114 |
| }, |
| { |
| "epoch": 2.252121212121212, |
| "grad_norm": 4.512854099273682, |
| "learning_rate": 3.2354080850191324e-05, |
| "loss": 0.7839, |
| "step": 115 |
| }, |
| { |
| "epoch": 2.2715151515151515, |
| "grad_norm": 4.273723125457764, |
| "learning_rate": 3.080611310224539e-05, |
| "loss": 0.7902, |
| "step": 116 |
| }, |
| { |
| "epoch": 2.290909090909091, |
| "grad_norm": 3.590773820877075, |
| "learning_rate": 2.9289321881345254e-05, |
| "loss": 0.7818, |
| "step": 117 |
| }, |
| { |
| "epoch": 2.3103030303030305, |
| "grad_norm": 3.1294124126434326, |
| "learning_rate": 2.7804390604547557e-05, |
| "loss": 0.786, |
| "step": 118 |
| }, |
| { |
| "epoch": 2.3296969696969696, |
| "grad_norm": 3.600764751434326, |
| "learning_rate": 2.6351988333848788e-05, |
| "loss": 0.7853, |
| "step": 119 |
| }, |
| { |
| "epoch": 2.349090909090909, |
| "grad_norm": 3.205591917037964, |
| "learning_rate": 2.493276947472756e-05, |
| "loss": 0.7835, |
| "step": 120 |
| }, |
| { |
| "epoch": 2.3684848484848486, |
| "grad_norm": 2.840078592300415, |
| "learning_rate": 2.354737348129077e-05, |
| "loss": 0.785, |
| "step": 121 |
| }, |
| { |
| "epoch": 2.3878787878787877, |
| "grad_norm": 3.47402286529541, |
| "learning_rate": 2.2196424568156073e-05, |
| "loss": 0.7822, |
| "step": 122 |
| }, |
| { |
| "epoch": 2.4072727272727272, |
| "grad_norm": 3.246713638305664, |
| "learning_rate": 2.0880531429201145e-05, |
| "loss": 0.779, |
| "step": 123 |
| }, |
| { |
| "epoch": 2.4266666666666667, |
| "grad_norm": 3.6749753952026367, |
| "learning_rate": 1.9600286963305957e-05, |
| "loss": 0.7974, |
| "step": 124 |
| }, |
| { |
| "epoch": 2.4460606060606063, |
| "grad_norm": 3.423552989959717, |
| "learning_rate": 1.835626800721144e-05, |
| "loss": 0.795, |
| "step": 125 |
| }, |
| { |
| "epoch": 2.4654545454545453, |
| "grad_norm": 5.248719215393066, |
| "learning_rate": 1.7149035075615794e-05, |
| "loss": 0.7816, |
| "step": 126 |
| }, |
| { |
| "epoch": 2.484848484848485, |
| "grad_norm": 2.561350107192993, |
| "learning_rate": 1.5979132108624574e-05, |
| "loss": 0.7886, |
| "step": 127 |
| }, |
| { |
| "epoch": 2.5042424242424244, |
| "grad_norm": 2.6997573375701904, |
| "learning_rate": 1.4847086226668872e-05, |
| "loss": 0.788, |
| "step": 128 |
| }, |
| { |
| "epoch": 2.5236363636363635, |
| "grad_norm": 2.400233268737793, |
| "learning_rate": 1.3753407493001968e-05, |
| "loss": 0.7785, |
| "step": 129 |
| }, |
| { |
| "epoch": 2.543030303030303, |
| "grad_norm": 4.139961242675781, |
| "learning_rate": 1.2698588683881186e-05, |
| "loss": 0.7846, |
| "step": 130 |
| }, |
| { |
| "epoch": 2.5624242424242425, |
| "grad_norm": 3.8779916763305664, |
| "learning_rate": 1.1683105066539068e-05, |
| "loss": 0.7877, |
| "step": 131 |
| }, |
| { |
| "epoch": 2.581818181818182, |
| "grad_norm": 3.5718817710876465, |
| "learning_rate": 1.0707414185043163e-05, |
| "loss": 0.7755, |
| "step": 132 |
| }, |
| { |
| "epoch": 2.601212121212121, |
| "grad_norm": 2.733354330062866, |
| "learning_rate": 9.771955654141496e-06, |
| "loss": 0.7923, |
| "step": 133 |
| }, |
| { |
| "epoch": 2.6206060606060606, |
| "grad_norm": 3.5972788333892822, |
| "learning_rate": 8.87715096118642e-06, |
| "loss": 0.7829, |
| "step": 134 |
| }, |
| { |
| "epoch": 2.64, |
| "grad_norm": 2.625112533569336, |
| "learning_rate": 8.023403276226126e-06, |
| "loss": 0.7758, |
| "step": 135 |
| }, |
| { |
| "epoch": 2.659393939393939, |
| "grad_norm": 1.7941921949386597, |
| "learning_rate": 7.211097270349066e-06, |
| "loss": 0.7896, |
| "step": 136 |
| }, |
| { |
| "epoch": 2.6787878787878787, |
| "grad_norm": 3.4884462356567383, |
| "learning_rate": 6.440598942363796e-06, |
| "loss": 0.787, |
| "step": 137 |
| }, |
| { |
| "epoch": 2.6981818181818182, |
| "grad_norm": 3.0694332122802734, |
| "learning_rate": 5.71225545389158e-06, |
| "loss": 0.7787, |
| "step": 138 |
| }, |
| { |
| "epoch": 2.7175757575757578, |
| "grad_norm": 3.9330761432647705, |
| "learning_rate": 5.026394972946813e-06, |
| "loss": 0.7724, |
| "step": 139 |
| }, |
| { |
| "epoch": 2.736969696969697, |
| "grad_norm": 4.874517917633057, |
| "learning_rate": 4.383326526074916e-06, |
| "loss": 0.7737, |
| "step": 140 |
| }, |
| { |
| "epoch": 2.7563636363636363, |
| "grad_norm": 3.3836331367492676, |
| "learning_rate": 3.783339859115065e-06, |
| "loss": 0.7892, |
| "step": 141 |
| }, |
| { |
| "epoch": 2.775757575757576, |
| "grad_norm": 3.0882411003112793, |
| "learning_rate": 3.226705306650113e-06, |
| "loss": 0.7905, |
| "step": 142 |
| }, |
| { |
| "epoch": 2.795151515151515, |
| "grad_norm": 3.052604913711548, |
| "learning_rate": 2.7136736702025433e-06, |
| "loss": 0.7853, |
| "step": 143 |
| }, |
| { |
| "epoch": 2.8145454545454545, |
| "grad_norm": 1.9272351264953613, |
| "learning_rate": 2.2444761052313856e-06, |
| "loss": 0.7732, |
| "step": 144 |
| }, |
| { |
| "epoch": 2.833939393939394, |
| "grad_norm": 2.5749495029449463, |
| "learning_rate": 1.8193240169810943e-06, |
| "loss": 0.7846, |
| "step": 145 |
| }, |
| { |
| "epoch": 2.8533333333333335, |
| "grad_norm": 1.8724554777145386, |
| "learning_rate": 1.4384089652291543e-06, |
| "loss": 0.7803, |
| "step": 146 |
| }, |
| { |
| "epoch": 2.8727272727272726, |
| "grad_norm": 2.784907102584839, |
| "learning_rate": 1.1019025779754666e-06, |
| "loss": 0.7838, |
| "step": 147 |
| }, |
| { |
| "epoch": 2.892121212121212, |
| "grad_norm": 4.25577449798584, |
| "learning_rate": 8.099564741123166e-07, |
| "loss": 0.7863, |
| "step": 148 |
| }, |
| { |
| "epoch": 2.9115151515151516, |
| "grad_norm": 2.0086898803710938, |
| "learning_rate": 5.627021951097545e-07, |
| "loss": 0.7808, |
| "step": 149 |
| }, |
| { |
| "epoch": 2.9309090909090907, |
| "grad_norm": 2.6451563835144043, |
| "learning_rate": 3.6025114574734785e-07, |
| "loss": 0.7877, |
| "step": 150 |
| }, |
| { |
| "epoch": 2.95030303030303, |
| "grad_norm": 3.0813286304473877, |
| "learning_rate": 2.0269454391874666e-07, |
| "loss": 0.7727, |
| "step": 151 |
| }, |
| { |
| "epoch": 2.9696969696969697, |
| "grad_norm": 2.1434407234191895, |
| "learning_rate": 9.010337953185843e-08, |
| "loss": 0.7861, |
| "step": 152 |
| }, |
| { |
| "epoch": 2.9890909090909092, |
| "grad_norm": 2.7659671306610107, |
| "learning_rate": 2.2528382523057113e-08, |
| "loss": 0.7782, |
| "step": 153 |
| }, |
| { |
| "epoch": 2.9890909090909092, |
| "step": 153, |
| "total_flos": 2.4132084696088576e+18, |
| "train_loss": 1.5338537124247333, |
| "train_runtime": 26135.9705, |
| "train_samples_per_second": 0.757, |
| "train_steps_per_second": 0.006 |
| } |
| ], |
| "logging_steps": 1.0, |
| "max_steps": 153, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 3, |
| "save_steps": 100, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 2.4132084696088576e+18, |
| "train_batch_size": 1, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|