diff --git "a/backup/output/backup/sft/full_data/lora_qwen2_5_vl/v0-20251210-172217/logging.jsonl" "b/backup/output/backup/sft/full_data/lora_qwen2_5_vl/v0-20251210-172217/logging.jsonl" new file mode 100644--- /dev/null +++ "b/backup/output/backup/sft/full_data/lora_qwen2_5_vl/v0-20251210-172217/logging.jsonl" @@ -0,0 +1,2515 @@ +{"loss": 0.99492186, "grad_norm": 0.2155518, "learning_rate": 1.32e-06, "epoch": 0.00039805, "global_step/max_steps": "1/2513", "percentage": "0.04%", "elapsed_time": "19s", "remaining_time": "13h 36m 25s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.051281} +{"loss": 0.94367695, "grad_norm": 0.2189005, "learning_rate": 2.63e-06, "epoch": 0.0007961, "global_step/max_steps": "2/2513", "percentage": "0.08%", "elapsed_time": "30s", "remaining_time": "10h 30m 7s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.066415} +{"loss": 1.06688213, "grad_norm": 0.21416435, "learning_rate": 3.95e-06, "epoch": 0.00119415, "global_step/max_steps": "3/2513", "percentage": "0.12%", "elapsed_time": "40s", "remaining_time": "9h 29m 26s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.073465} +{"loss": 0.98178893, "grad_norm": 0.21602339, "learning_rate": 5.26e-06, "epoch": 0.0015922, "global_step/max_steps": "4/2513", "percentage": "0.16%", "elapsed_time": "51s", "remaining_time": "8h 57m 23s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.077813} +{"loss": 1.06129098, "grad_norm": 0.21177924, "learning_rate": 6.58e-06, "epoch": 0.00199025, "global_step/max_steps": "5/2513", "percentage": "0.20%", "elapsed_time": "1m 0s", "remaining_time": "8h 29m 23s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.08206} +{"loss": 0.99530178, "grad_norm": 0.20776736, "learning_rate": 7.89e-06, "epoch": 0.0023883, "global_step/max_steps": "6/2513", "percentage": "0.24%", "elapsed_time": "1m 10s", "remaining_time": "8h 12m 43s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.0848} +{"loss": 0.93607241, "grad_norm": 0.19928025, "learning_rate": 9.21e-06, "epoch": 0.00278635, "global_step/max_steps": "7/2513", "percentage": "0.28%", "elapsed_time": "1m 20s", "remaining_time": "7h 59m 59s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.087016} +{"loss": 1.00627804, "grad_norm": 0.21057382, "learning_rate": 1.053e-05, "epoch": 0.0031844, "global_step/max_steps": "8/2513", "percentage": "0.32%", "elapsed_time": "1m 31s", "remaining_time": "7h 55m 11s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.087859} +{"loss": 1.00552511, "grad_norm": 0.23786748, "learning_rate": 1.184e-05, "epoch": 0.00358245, "global_step/max_steps": "9/2513", "percentage": "0.36%", "elapsed_time": "1m 40s", "remaining_time": "7h 45m 36s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.089631} +{"loss": 0.93177068, "grad_norm": 0.2672223, "learning_rate": 1.316e-05, "epoch": 0.0039805, "global_step/max_steps": "10/2513", "percentage": "0.40%", "elapsed_time": "1m 49s", "remaining_time": "7h 37m 54s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.091104} +{"loss": 0.9052282, "grad_norm": 0.21399761, "learning_rate": 1.447e-05, "epoch": 0.00437855, "global_step/max_steps": "11/2513", "percentage": "0.44%", "elapsed_time": "2m 0s", "remaining_time": "7h 35m 23s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.091569} +{"loss": 0.88549763, "grad_norm": 0.22456881, "learning_rate": 1.579e-05, "epoch": 0.00477659, "global_step/max_steps": "12/2513", "percentage": "0.48%", "elapsed_time": "2m 9s", "remaining_time": "7h 31m 2s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.092416} +{"loss": 0.82977152, "grad_norm": 0.18524821, "learning_rate": 1.711e-05, "epoch": 0.00517464, "global_step/max_steps": "13/2513", "percentage": "0.52%", "elapsed_time": "2m 22s", "remaining_time": "7h 37m 5s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.091154} +{"loss": 0.82335645, "grad_norm": 0.1307091, "learning_rate": 1.842e-05, "epoch": 0.00557269, "global_step/max_steps": "14/2513", "percentage": "0.56%", "elapsed_time": "2m 32s", "remaining_time": "7h 34m 26s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.09165} +{"loss": 0.88316786, "grad_norm": 0.13550087, "learning_rate": 1.974e-05, "epoch": 0.00597074, "global_step/max_steps": "15/2513", "percentage": "0.60%", "elapsed_time": "2m 42s", "remaining_time": "7h 30m 22s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.092441} +{"loss": 0.69601846, "grad_norm": 0.12782316, "learning_rate": 2.105e-05, "epoch": 0.00636879, "global_step/max_steps": "16/2513", "percentage": "0.64%", "elapsed_time": "2m 52s", "remaining_time": "7h 27m 25s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.093015} +{"loss": 0.71622086, "grad_norm": 0.13018635, "learning_rate": 2.237e-05, "epoch": 0.00676684, "global_step/max_steps": "17/2513", "percentage": "0.68%", "elapsed_time": "3m 1s", "remaining_time": "7h 24m 40s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.093553} +{"loss": 0.70752454, "grad_norm": 0.12816942, "learning_rate": 2.368e-05, "epoch": 0.00716489, "global_step/max_steps": "18/2513", "percentage": "0.72%", "elapsed_time": "3m 11s", "remaining_time": "7h 22m 7s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.094054} +{"loss": 0.69916421, "grad_norm": 0.13075613, "learning_rate": 2.5e-05, "epoch": 0.00756294, "global_step/max_steps": "19/2513", "percentage": "0.76%", "elapsed_time": "3m 20s", "remaining_time": "7h 19m 41s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.094537} +{"loss": 0.68453735, "grad_norm": 0.13933954, "learning_rate": 2.632e-05, "epoch": 0.00796099, "global_step/max_steps": "20/2513", "percentage": "0.80%", "elapsed_time": "3m 31s", "remaining_time": "7h 19m 4s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.094631} +{"loss": 0.68185133, "grad_norm": 0.09653552, "learning_rate": 2.763e-05, "epoch": 0.00835904, "global_step/max_steps": "21/2513", "percentage": "0.84%", "elapsed_time": "3m 41s", "remaining_time": "7h 18m 52s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.094637} +{"loss": 0.62866157, "grad_norm": 0.09195902, "learning_rate": 2.895e-05, "epoch": 0.00875709, "global_step/max_steps": "22/2513", "percentage": "0.88%", "elapsed_time": "3m 51s", "remaining_time": "7h 17m 16s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.094943} +{"loss": 0.61280859, "grad_norm": 0.10400124, "learning_rate": 3.026e-05, "epoch": 0.00915514, "global_step/max_steps": "23/2513", "percentage": "0.92%", "elapsed_time": "4m 1s", "remaining_time": "7h 15m 1s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.095396} +{"loss": 0.62549853, "grad_norm": 0.10874229, "learning_rate": 3.158e-05, "epoch": 0.00955319, "global_step/max_steps": "24/2513", "percentage": "0.96%", "elapsed_time": "4m 11s", "remaining_time": "7h 13m 53s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.095607} +{"loss": 0.58290291, "grad_norm": 0.10870522, "learning_rate": 3.289e-05, "epoch": 0.00995124, "global_step/max_steps": "25/2513", "percentage": "0.99%", "elapsed_time": "4m 21s", "remaining_time": "7h 14m 21s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.095466} +{"loss": 0.64286339, "grad_norm": 0.1047467, "learning_rate": 3.421e-05, "epoch": 0.01034929, "global_step/max_steps": "26/2513", "percentage": "1.03%", "elapsed_time": "4m 31s", "remaining_time": "7h 12m 45s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.095783} +{"loss": 0.61306274, "grad_norm": 0.10712946, "learning_rate": 3.553e-05, "epoch": 0.01074734, "global_step/max_steps": "27/2513", "percentage": "1.07%", "elapsed_time": "4m 41s", "remaining_time": "7h 11m 44s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.095967} +{"loss": 0.56914639, "grad_norm": 0.09571935, "learning_rate": 3.684e-05, "epoch": 0.01114539, "global_step/max_steps": "28/2513", "percentage": "1.11%", "elapsed_time": "4m 53s", "remaining_time": "7h 13m 32s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.095532} +{"loss": 0.58046091, "grad_norm": 0.06699314, "learning_rate": 3.816e-05, "epoch": 0.01154344, "global_step/max_steps": "29/2513", "percentage": "1.15%", "elapsed_time": "5m 2s", "remaining_time": "7h 12m 4s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.095815} +{"loss": 0.53397465, "grad_norm": 0.04525255, "learning_rate": 3.947e-05, "epoch": 0.01194149, "global_step/max_steps": "30/2513", "percentage": "1.19%", "elapsed_time": "5m 12s", "remaining_time": "7h 11m 1s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.096013} +{"loss": 0.52872682, "grad_norm": 0.04023329, "learning_rate": 4.079e-05, "epoch": 0.01233954, "global_step/max_steps": "31/2513", "percentage": "1.23%", "elapsed_time": "5m 22s", "remaining_time": "7h 10m 13s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.096152} +{"loss": 0.58563882, "grad_norm": 0.06605311, "learning_rate": 4.211e-05, "epoch": 0.01273759, "global_step/max_steps": "32/2513", "percentage": "1.27%", "elapsed_time": "5m 32s", "remaining_time": "7h 9m 2s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.096378} +{"loss": 0.54210782, "grad_norm": 0.04610144, "learning_rate": 4.342e-05, "epoch": 0.01313564, "global_step/max_steps": "33/2513", "percentage": "1.31%", "elapsed_time": "5m 41s", "remaining_time": "7h 7m 46s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.096625} +{"loss": 0.50995529, "grad_norm": 0.04927177, "learning_rate": 4.474e-05, "epoch": 0.01353368, "global_step/max_steps": "34/2513", "percentage": "1.35%", "elapsed_time": "5m 51s", "remaining_time": "7h 6m 54s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.09678} +{"loss": 0.52196729, "grad_norm": 0.04793802, "learning_rate": 4.605e-05, "epoch": 0.01393173, "global_step/max_steps": "35/2513", "percentage": "1.39%", "elapsed_time": "6m 0s", "remaining_time": "7h 5m 46s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.097} +{"loss": 0.46674502, "grad_norm": 0.03821297, "learning_rate": 4.737e-05, "epoch": 0.01432978, "global_step/max_steps": "36/2513", "percentage": "1.43%", "elapsed_time": "6m 10s", "remaining_time": "7h 5m 12s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.097088} +{"loss": 0.46464723, "grad_norm": 0.03418196, "learning_rate": 4.868e-05, "epoch": 0.01472783, "global_step/max_steps": "37/2513", "percentage": "1.47%", "elapsed_time": "6m 20s", "remaining_time": "7h 4m 35s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.097191} +{"loss": 0.54680771, "grad_norm": 0.04802583, "learning_rate": 5e-05, "epoch": 0.01512588, "global_step/max_steps": "38/2513", "percentage": "1.51%", "elapsed_time": "6m 30s", "remaining_time": "7h 4m 2s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.09728} +{"loss": 0.54605794, "grad_norm": 0.0487497, "learning_rate": 5.132e-05, "epoch": 0.01552393, "global_step/max_steps": "39/2513", "percentage": "1.55%", "elapsed_time": "6m 40s", "remaining_time": "7h 2m 56s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.097491} +{"loss": 0.53157222, "grad_norm": 0.04837357, "learning_rate": 5.263e-05, "epoch": 0.01592198, "global_step/max_steps": "40/2513", "percentage": "1.59%", "elapsed_time": "6m 50s", "remaining_time": "7h 2m 36s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.097529} +{"loss": 0.51745248, "grad_norm": 0.04149935, "learning_rate": 5.395e-05, "epoch": 0.01632003, "global_step/max_steps": "41/2513", "percentage": "1.63%", "elapsed_time": "6m 59s", "remaining_time": "7h 1m 46s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.097681} +{"loss": 0.54102689, "grad_norm": 0.04199228, "learning_rate": 5.526e-05, "epoch": 0.01671808, "global_step/max_steps": "42/2513", "percentage": "1.67%", "elapsed_time": "7m 10s", "remaining_time": "7h 1m 57s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.097599} +{"loss": 0.54316354, "grad_norm": 0.03990499, "learning_rate": 5.658e-05, "epoch": 0.01711613, "global_step/max_steps": "43/2513", "percentage": "1.71%", "elapsed_time": "7m 21s", "remaining_time": "7h 2m 12s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.097504} +{"loss": 0.52153653, "grad_norm": 0.03669535, "learning_rate": 5.789e-05, "epoch": 0.01751418, "global_step/max_steps": "44/2513", "percentage": "1.75%", "elapsed_time": "7m 30s", "remaining_time": "7h 1m 24s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.097647} +{"loss": 0.5254761, "grad_norm": 0.03882231, "learning_rate": 5.921e-05, "epoch": 0.01791223, "global_step/max_steps": "45/2513", "percentage": "1.79%", "elapsed_time": "7m 40s", "remaining_time": "7h 0m 41s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.097774} +{"loss": 0.52390671, "grad_norm": 0.04098727, "learning_rate": 6.053e-05, "epoch": 0.01831028, "global_step/max_steps": "46/2513", "percentage": "1.83%", "elapsed_time": "7m 49s", "remaining_time": "6h 59m 51s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.09793} +{"loss": 0.44943276, "grad_norm": 0.03812268, "learning_rate": 6.184e-05, "epoch": 0.01870833, "global_step/max_steps": "47/2513", "percentage": "1.87%", "elapsed_time": "7m 59s", "remaining_time": "6h 59m 9s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.098055} +{"loss": 0.52037275, "grad_norm": 0.03938031, "learning_rate": 6.316e-05, "epoch": 0.01910638, "global_step/max_steps": "48/2513", "percentage": "1.91%", "elapsed_time": "8m 9s", "remaining_time": "6h 58m 45s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.098106} +{"loss": 0.52805287, "grad_norm": 0.03527931, "learning_rate": 6.447e-05, "epoch": 0.01950443, "global_step/max_steps": "49/2513", "percentage": "1.95%", "elapsed_time": "8m 18s", "remaining_time": "6h 58m 0s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.098246} +{"loss": 0.50398469, "grad_norm": 0.03625924, "learning_rate": 6.579e-05, "epoch": 0.01990248, "global_step/max_steps": "50/2513", "percentage": "1.99%", "elapsed_time": "8m 28s", "remaining_time": "6h 57m 40s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.098284} +{"loss": 0.45674366, "grad_norm": 0.03593372, "learning_rate": 6.711e-05, "epoch": 0.02030053, "global_step/max_steps": "51/2513", "percentage": "2.03%", "elapsed_time": "8m 38s", "remaining_time": "6h 57m 13s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.098346} +{"loss": 0.54829997, "grad_norm": 0.04369781, "learning_rate": 6.842e-05, "epoch": 0.02069858, "global_step/max_steps": "52/2513", "percentage": "2.07%", "elapsed_time": "8m 48s", "remaining_time": "6h 56m 48s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.098407} +{"loss": 0.48109418, "grad_norm": 0.03834309, "learning_rate": 6.974e-05, "epoch": 0.02109663, "global_step/max_steps": "53/2513", "percentage": "2.11%", "elapsed_time": "8m 57s", "remaining_time": "6h 56m 8s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.098523} +{"loss": 0.47800878, "grad_norm": 0.035735, "learning_rate": 7.105e-05, "epoch": 0.02149468, "global_step/max_steps": "54/2513", "percentage": "2.15%", "elapsed_time": "9m 7s", "remaining_time": "6h 55m 42s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.098588} +{"loss": 0.50193167, "grad_norm": 0.03294038, "learning_rate": 7.237e-05, "epoch": 0.02189273, "global_step/max_steps": "55/2513", "percentage": "2.19%", "elapsed_time": "9m 17s", "remaining_time": "6h 55m 15s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.098655} +{"loss": 0.51210964, "grad_norm": 0.03878693, "learning_rate": 7.368e-05, "epoch": 0.02229078, "global_step/max_steps": "56/2513", "percentage": "2.23%", "elapsed_time": "9m 27s", "remaining_time": "6h 55m 10s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.098634} +{"loss": 0.5112983, "grad_norm": 0.03802566, "learning_rate": 7.5e-05, "epoch": 0.02268882, "global_step/max_steps": "57/2513", "percentage": "2.27%", "elapsed_time": "9m 39s", "remaining_time": "6h 55m 55s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.098413} +{"loss": 0.50983649, "grad_norm": 0.03701293, "learning_rate": 7.632e-05, "epoch": 0.02308687, "global_step/max_steps": "58/2513", "percentage": "2.31%", "elapsed_time": "9m 49s", "remaining_time": "6h 55m 32s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.098466} +{"loss": 0.52957296, "grad_norm": 0.0373461, "learning_rate": 7.763e-05, "epoch": 0.02348492, "global_step/max_steps": "59/2513", "percentage": "2.35%", "elapsed_time": "9m 58s", "remaining_time": "6h 55m 2s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.098545} +{"loss": 0.43766993, "grad_norm": 0.03432748, "learning_rate": 7.895e-05, "epoch": 0.02388297, "global_step/max_steps": "60/2513", "percentage": "2.39%", "elapsed_time": "10m 8s", "remaining_time": "6h 54m 43s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.098578} +{"loss": 0.46333522, "grad_norm": 0.03595589, "learning_rate": 8.026e-05, "epoch": 0.02428102, "global_step/max_steps": "61/2513", "percentage": "2.43%", "elapsed_time": "10m 18s", "remaining_time": "6h 54m 12s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.09866} +{"loss": 0.45054978, "grad_norm": 0.03505861, "learning_rate": 8.158e-05, "epoch": 0.02467907, "global_step/max_steps": "62/2513", "percentage": "2.47%", "elapsed_time": "10m 27s", "remaining_time": "6h 53m 40s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.09875} +{"loss": 0.47529572, "grad_norm": 0.04025973, "learning_rate": 8.289e-05, "epoch": 0.02507712, "global_step/max_steps": "63/2513", "percentage": "2.51%", "elapsed_time": "10m 37s", "remaining_time": "6h 53m 23s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.098776} +{"loss": 0.42414466, "grad_norm": 0.0355258, "learning_rate": 8.421e-05, "epoch": 0.02547517, "global_step/max_steps": "64/2513", "percentage": "2.55%", "elapsed_time": "10m 47s", "remaining_time": "6h 52m 55s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.098848} +{"loss": 0.48888195, "grad_norm": 0.03537808, "learning_rate": 8.553e-05, "epoch": 0.02587322, "global_step/max_steps": "65/2513", "percentage": "2.59%", "elapsed_time": "10m 57s", "remaining_time": "6h 52m 40s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.098866} +{"loss": 0.46754813, "grad_norm": 0.03906525, "learning_rate": 8.684e-05, "epoch": 0.02627127, "global_step/max_steps": "66/2513", "percentage": "2.63%", "elapsed_time": "11m 7s", "remaining_time": "6h 52m 14s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.098929} +{"loss": 0.44222131, "grad_norm": 0.03745347, "learning_rate": 8.816e-05, "epoch": 0.02666932, "global_step/max_steps": "67/2513", "percentage": "2.67%", "elapsed_time": "11m 17s", "remaining_time": "6h 52m 22s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.098858} +{"loss": 0.41981095, "grad_norm": 0.03608848, "learning_rate": 8.947e-05, "epoch": 0.02706737, "global_step/max_steps": "68/2513", "percentage": "2.71%", "elapsed_time": "11m 27s", "remaining_time": "6h 52m 1s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.0989} +{"loss": 0.44852066, "grad_norm": 0.03446097, "learning_rate": 9.079e-05, "epoch": 0.02746542, "global_step/max_steps": "69/2513", "percentage": "2.75%", "elapsed_time": "11m 37s", "remaining_time": "6h 51m 44s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.09893} +{"loss": 0.42335641, "grad_norm": 0.0367477, "learning_rate": 9.211e-05, "epoch": 0.02786347, "global_step/max_steps": "70/2513", "percentage": "2.79%", "elapsed_time": "11m 47s", "remaining_time": "6h 51m 20s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.098985} +{"loss": 0.42593247, "grad_norm": 0.03501367, "learning_rate": 9.342e-05, "epoch": 0.02826152, "global_step/max_steps": "71/2513", "percentage": "2.83%", "elapsed_time": "11m 58s", "remaining_time": "6h 51m 45s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.098846} +{"loss": 0.45563731, "grad_norm": 0.04298533, "learning_rate": 9.474e-05, "epoch": 0.02865957, "global_step/max_steps": "72/2513", "percentage": "2.87%", "elapsed_time": "12m 9s", "remaining_time": "6h 52m 1s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.098739} +{"loss": 0.46544915, "grad_norm": 0.03907185, "learning_rate": 9.605e-05, "epoch": 0.02905762, "global_step/max_steps": "73/2513", "percentage": "2.90%", "elapsed_time": "12m 18s", "remaining_time": "6h 51m 39s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.098787} +{"loss": 0.52784073, "grad_norm": 0.04066806, "learning_rate": 9.737e-05, "epoch": 0.02945567, "global_step/max_steps": "74/2513", "percentage": "2.94%", "elapsed_time": "12m 29s", "remaining_time": "6h 51m 27s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.098793} +{"loss": 0.46847758, "grad_norm": 0.03859466, "learning_rate": 9.868e-05, "epoch": 0.02985372, "global_step/max_steps": "75/2513", "percentage": "2.98%", "elapsed_time": "12m 39s", "remaining_time": "6h 51m 15s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.098803} +{"loss": 0.4965198, "grad_norm": 0.03836249, "learning_rate": 0.0001, "epoch": 0.03025177, "global_step/max_steps": "76/2513", "percentage": "3.02%", "elapsed_time": "12m 48s", "remaining_time": "6h 50m 51s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.09886} +{"loss": 0.43584037, "grad_norm": 0.04042613, "learning_rate": 0.0001, "epoch": 0.03064982, "global_step/max_steps": "77/2513", "percentage": "3.06%", "elapsed_time": "12m 58s", "remaining_time": "6h 50m 26s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.098918} +{"loss": 0.43018526, "grad_norm": 0.03756863, "learning_rate": 0.0001, "epoch": 0.03104787, "global_step/max_steps": "78/2513", "percentage": "3.10%", "elapsed_time": "13m 9s", "remaining_time": "6h 50m 33s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.098851} +{"loss": 0.44210824, "grad_norm": 0.03331263, "learning_rate": 0.0001, "epoch": 0.03144592, "global_step/max_steps": "79/2513", "percentage": "3.14%", "elapsed_time": "13m 18s", "remaining_time": "6h 50m 7s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.098913} +{"loss": 0.4545646, "grad_norm": 0.04031183, "learning_rate": 0.0001, "epoch": 0.03184396, "global_step/max_steps": "80/2513", "percentage": "3.18%", "elapsed_time": "13m 28s", "remaining_time": "6h 49m 41s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.098976} +{"loss": 0.51598227, "grad_norm": 0.04213561, "learning_rate": 0.0001, "epoch": 0.03224201, "global_step/max_steps": "81/2513", "percentage": "3.22%", "elapsed_time": "13m 38s", "remaining_time": "6h 49m 31s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.098975} +{"loss": 0.47242194, "grad_norm": 0.03805201, "learning_rate": 0.0001, "epoch": 0.03264006, "global_step/max_steps": "82/2513", "percentage": "3.26%", "elapsed_time": "13m 49s", "remaining_time": "6h 49m 38s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.098909} +{"loss": 0.41448423, "grad_norm": 0.03853696, "learning_rate": 0.0001, "epoch": 0.03303811, "global_step/max_steps": "83/2513", "percentage": "3.30%", "elapsed_time": "13m 58s", "remaining_time": "6h 49m 12s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.098972} +{"loss": 0.45854294, "grad_norm": 0.04499345, "learning_rate": 0.0001, "epoch": 0.03343616, "global_step/max_steps": "84/2513", "percentage": "3.34%", "elapsed_time": "14m 8s", "remaining_time": "6h 48m 54s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.099004} +{"loss": 0.46574044, "grad_norm": 0.04198587, "learning_rate": 0.0001, "epoch": 0.03383421, "global_step/max_steps": "85/2513", "percentage": "3.38%", "elapsed_time": "14m 18s", "remaining_time": "6h 48m 52s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.098969} +{"loss": 0.43082204, "grad_norm": 0.04056991, "learning_rate": 0.0001, "epoch": 0.03423226, "global_step/max_steps": "86/2513", "percentage": "3.42%", "elapsed_time": "14m 29s", "remaining_time": "6h 49m 7s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.098869} +{"loss": 0.49034557, "grad_norm": 0.03974547, "learning_rate": 9.999e-05, "epoch": 0.03463031, "global_step/max_steps": "87/2513", "percentage": "3.46%", "elapsed_time": "14m 40s", "remaining_time": "6h 49m 3s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.098847} +{"loss": 0.43225497, "grad_norm": 0.03939761, "learning_rate": 9.999e-05, "epoch": 0.03502836, "global_step/max_steps": "88/2513", "percentage": "3.50%", "elapsed_time": "14m 49s", "remaining_time": "6h 48m 37s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.098909} +{"loss": 0.46949512, "grad_norm": 0.03816381, "learning_rate": 9.999e-05, "epoch": 0.03542641, "global_step/max_steps": "89/2513", "percentage": "3.54%", "elapsed_time": "14m 59s", "remaining_time": "6h 48m 16s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.098954} +{"loss": 0.44680744, "grad_norm": 0.03871481, "learning_rate": 9.999e-05, "epoch": 0.03582446, "global_step/max_steps": "90/2513", "percentage": "3.58%", "elapsed_time": "15m 9s", "remaining_time": "6h 48m 3s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.098967} +{"loss": 0.3941884, "grad_norm": 0.03732983, "learning_rate": 9.999e-05, "epoch": 0.03622251, "global_step/max_steps": "91/2513", "percentage": "3.62%", "elapsed_time": "15m 19s", "remaining_time": "6h 47m 48s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.098984} +{"loss": 0.4268159, "grad_norm": 0.03932844, "learning_rate": 9.999e-05, "epoch": 0.03662056, "global_step/max_steps": "92/2513", "percentage": "3.66%", "elapsed_time": "15m 28s", "remaining_time": "6h 47m 21s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.099053} +{"loss": 0.39057562, "grad_norm": 0.03829151, "learning_rate": 9.999e-05, "epoch": 0.03701861, "global_step/max_steps": "93/2513", "percentage": "3.70%", "elapsed_time": "15m 38s", "remaining_time": "6h 46m 53s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.099126} +{"loss": 0.4239282, "grad_norm": 0.03774095, "learning_rate": 9.999e-05, "epoch": 0.03741666, "global_step/max_steps": "94/2513", "percentage": "3.74%", "elapsed_time": "15m 47s", "remaining_time": "6h 46m 32s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.099168} +{"loss": 0.47255814, "grad_norm": 0.042899, "learning_rate": 9.999e-05, "epoch": 0.03781471, "global_step/max_steps": "95/2513", "percentage": "3.78%", "elapsed_time": "15m 57s", "remaining_time": "6h 46m 12s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.099211} +{"loss": 0.42352629, "grad_norm": 0.04366745, "learning_rate": 9.998e-05, "epoch": 0.03821276, "global_step/max_steps": "96/2513", "percentage": "3.82%", "elapsed_time": "16m 6s", "remaining_time": "6h 45m 42s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.099293} +{"loss": 0.43142402, "grad_norm": 0.04007267, "learning_rate": 9.998e-05, "epoch": 0.03861081, "global_step/max_steps": "97/2513", "percentage": "3.86%", "elapsed_time": "16m 16s", "remaining_time": "6h 45m 19s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.099344} +{"loss": 0.45048943, "grad_norm": 0.04459287, "learning_rate": 9.998e-05, "epoch": 0.03900886, "global_step/max_steps": "98/2513", "percentage": "3.90%", "elapsed_time": "16m 26s", "remaining_time": "6h 45m 6s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.099355} +{"loss": 0.44291624, "grad_norm": 0.04025343, "learning_rate": 9.998e-05, "epoch": 0.03940691, "global_step/max_steps": "99/2513", "percentage": "3.94%", "elapsed_time": "16m 36s", "remaining_time": "6h 44m 46s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.099397} +{"loss": 0.44428122, "grad_norm": 0.04076025, "learning_rate": 9.998e-05, "epoch": 0.03980496, "global_step/max_steps": "100/2513", "percentage": "3.98%", "elapsed_time": "16m 45s", "remaining_time": "6h 44m 29s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.099425} +{"loss": 0.44870013, "grad_norm": 0.0394497, "learning_rate": 9.997e-05, "epoch": 0.04020301, "global_step/max_steps": "101/2513", "percentage": "4.02%", "elapsed_time": "16m 58s", "remaining_time": "6h 45m 22s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.099167} +{"loss": 0.48024178, "grad_norm": 0.04446409, "learning_rate": 9.997e-05, "epoch": 0.04060105, "global_step/max_steps": "102/2513", "percentage": "4.06%", "elapsed_time": "17m 8s", "remaining_time": "6h 45m 4s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.099201} +{"loss": 0.44365811, "grad_norm": 0.03753713, "learning_rate": 9.997e-05, "epoch": 0.0409991, "global_step/max_steps": "103/2513", "percentage": "4.10%", "elapsed_time": "17m 18s", "remaining_time": "6h 44m 49s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.099221} +{"loss": 0.45750669, "grad_norm": 0.0493593, "learning_rate": 9.997e-05, "epoch": 0.04139715, "global_step/max_steps": "104/2513", "percentage": "4.14%", "elapsed_time": "17m 27s", "remaining_time": "6h 44m 26s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.099271} +{"loss": 0.43207157, "grad_norm": 0.04282277, "learning_rate": 9.997e-05, "epoch": 0.0417952, "global_step/max_steps": "105/2513", "percentage": "4.18%", "elapsed_time": "17m 37s", "remaining_time": "6h 44m 8s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.099307} +{"loss": 0.42911652, "grad_norm": 0.03503739, "learning_rate": 9.996e-05, "epoch": 0.04219325, "global_step/max_steps": "106/2513", "percentage": "4.22%", "elapsed_time": "17m 46s", "remaining_time": "6h 43m 43s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.099368} +{"loss": 0.45157364, "grad_norm": 0.04159521, "learning_rate": 9.996e-05, "epoch": 0.0425913, "global_step/max_steps": "107/2513", "percentage": "4.26%", "elapsed_time": "17m 56s", "remaining_time": "6h 43m 21s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.099416} +{"loss": 0.4506799, "grad_norm": 0.04353627, "learning_rate": 9.996e-05, "epoch": 0.04298935, "global_step/max_steps": "108/2513", "percentage": "4.30%", "elapsed_time": "18m 5s", "remaining_time": "6h 42m 59s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.099464} +{"loss": 0.45106912, "grad_norm": 0.04386578, "learning_rate": 9.995e-05, "epoch": 0.0433874, "global_step/max_steps": "109/2513", "percentage": "4.34%", "elapsed_time": "18m 15s", "remaining_time": "6h 42m 44s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.099483} +{"loss": 0.5013119, "grad_norm": 0.04238444, "learning_rate": 9.995e-05, "epoch": 0.04378545, "global_step/max_steps": "110/2513", "percentage": "4.38%", "elapsed_time": "18m 25s", "remaining_time": "6h 42m 32s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.099492} +{"loss": 0.37856752, "grad_norm": 0.0386101, "learning_rate": 9.995e-05, "epoch": 0.0441835, "global_step/max_steps": "111/2513", "percentage": "4.42%", "elapsed_time": "18m 35s", "remaining_time": "6h 42m 17s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.099512} +{"loss": 0.40496111, "grad_norm": 0.04140529, "learning_rate": 9.995e-05, "epoch": 0.04458155, "global_step/max_steps": "112/2513", "percentage": "4.46%", "elapsed_time": "18m 44s", "remaining_time": "6h 41m 56s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.09956} +{"loss": 0.43783048, "grad_norm": 0.04486702, "learning_rate": 9.994e-05, "epoch": 0.0449796, "global_step/max_steps": "113/2513", "percentage": "4.50%", "elapsed_time": "18m 54s", "remaining_time": "6h 41m 44s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.099565} +{"loss": 0.42617247, "grad_norm": 0.03791196, "learning_rate": 9.994e-05, "epoch": 0.04537765, "global_step/max_steps": "114/2513", "percentage": "4.54%", "elapsed_time": "19m 4s", "remaining_time": "6h 41m 28s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.099589} +{"loss": 0.40904814, "grad_norm": 0.04239538, "learning_rate": 9.994e-05, "epoch": 0.0457757, "global_step/max_steps": "115/2513", "percentage": "4.58%", "elapsed_time": "19m 14s", "remaining_time": "6h 41m 10s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.099625} +{"loss": 0.4320094, "grad_norm": 0.03892144, "learning_rate": 9.993e-05, "epoch": 0.04617375, "global_step/max_steps": "116/2513", "percentage": "4.62%", "elapsed_time": "19m 25s", "remaining_time": "6h 41m 21s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.099539} +{"loss": 0.35490415, "grad_norm": 0.03638756, "learning_rate": 9.993e-05, "epoch": 0.0465718, "global_step/max_steps": "117/2513", "percentage": "4.66%", "elapsed_time": "19m 35s", "remaining_time": "6h 41m 6s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.099556} +{"loss": 0.43866396, "grad_norm": 0.03842415, "learning_rate": 9.993e-05, "epoch": 0.04696985, "global_step/max_steps": "118/2513", "percentage": "4.70%", "elapsed_time": "19m 44s", "remaining_time": "6h 40m 46s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.099598} +{"loss": 0.36374098, "grad_norm": 0.04200448, "learning_rate": 9.992e-05, "epoch": 0.0473679, "global_step/max_steps": "119/2513", "percentage": "4.74%", "elapsed_time": "19m 54s", "remaining_time": "6h 40m 27s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.099636} +{"loss": 0.41988105, "grad_norm": 0.03868981, "learning_rate": 9.992e-05, "epoch": 0.04776595, "global_step/max_steps": "120/2513", "percentage": "4.78%", "elapsed_time": "20m 3s", "remaining_time": "6h 40m 6s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.099682} +{"loss": 0.45327681, "grad_norm": 0.04202842, "learning_rate": 9.992e-05, "epoch": 0.048164, "global_step/max_steps": "121/2513", "percentage": "4.81%", "elapsed_time": "20m 13s", "remaining_time": "6h 39m 39s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.09975} +{"loss": 0.41151503, "grad_norm": 0.03830932, "learning_rate": 9.991e-05, "epoch": 0.04856205, "global_step/max_steps": "122/2513", "percentage": "4.85%", "elapsed_time": "20m 22s", "remaining_time": "6h 39m 21s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.099786} +{"loss": 0.43575126, "grad_norm": 0.04025004, "learning_rate": 9.991e-05, "epoch": 0.0489601, "global_step/max_steps": "123/2513", "percentage": "4.89%", "elapsed_time": "20m 32s", "remaining_time": "6h 39m 7s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.0998} +{"loss": 0.42165613, "grad_norm": 0.04027884, "learning_rate": 9.99e-05, "epoch": 0.04935815, "global_step/max_steps": "124/2513", "percentage": "4.93%", "elapsed_time": "20m 42s", "remaining_time": "6h 38m 54s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.099814} +{"loss": 0.42360121, "grad_norm": 0.04131161, "learning_rate": 9.99e-05, "epoch": 0.04975619, "global_step/max_steps": "125/2513", "percentage": "4.97%", "elapsed_time": "20m 51s", "remaining_time": "6h 38m 35s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.099851} +{"loss": 0.4609946, "grad_norm": 0.0400711, "learning_rate": 9.99e-05, "epoch": 0.05015424, "global_step/max_steps": "126/2513", "percentage": "5.01%", "elapsed_time": "21m 1s", "remaining_time": "6h 38m 17s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.099886} +{"loss": 0.48313099, "grad_norm": 0.03947784, "learning_rate": 9.989e-05, "epoch": 0.05055229, "global_step/max_steps": "127/2513", "percentage": "5.05%", "elapsed_time": "21m 11s", "remaining_time": "6h 38m 6s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.099891} +{"loss": 0.47838131, "grad_norm": 0.04620794, "learning_rate": 9.989e-05, "epoch": 0.05095034, "global_step/max_steps": "128/2513", "percentage": "5.09%", "elapsed_time": "21m 21s", "remaining_time": "6h 37m 56s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.09989} +{"loss": 0.43852988, "grad_norm": 0.04389549, "learning_rate": 9.988e-05, "epoch": 0.05134839, "global_step/max_steps": "129/2513", "percentage": "5.13%", "elapsed_time": "21m 31s", "remaining_time": "6h 37m 43s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.099902} +{"loss": 0.43673021, "grad_norm": 0.03762936, "learning_rate": 9.988e-05, "epoch": 0.05174644, "global_step/max_steps": "130/2513", "percentage": "5.17%", "elapsed_time": "21m 42s", "remaining_time": "6h 37m 48s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.09984} +{"loss": 0.42734182, "grad_norm": 0.0404209, "learning_rate": 9.987e-05, "epoch": 0.05214449, "global_step/max_steps": "131/2513", "percentage": "5.21%", "elapsed_time": "21m 52s", "remaining_time": "6h 37m 48s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.099798} +{"loss": 0.42770153, "grad_norm": 0.04365689, "learning_rate": 9.987e-05, "epoch": 0.05254254, "global_step/max_steps": "132/2513", "percentage": "5.25%", "elapsed_time": "22m 2s", "remaining_time": "6h 37m 30s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.099832} +{"loss": 0.43427616, "grad_norm": 0.04318428, "learning_rate": 9.987e-05, "epoch": 0.05294059, "global_step/max_steps": "133/2513", "percentage": "5.29%", "elapsed_time": "22m 12s", "remaining_time": "6h 37m 19s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.099836} +{"loss": 0.43279463, "grad_norm": 0.0389939, "learning_rate": 9.986e-05, "epoch": 0.05333864, "global_step/max_steps": "134/2513", "percentage": "5.33%", "elapsed_time": "22m 21s", "remaining_time": "6h 36m 59s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.099878} +{"loss": 0.47584808, "grad_norm": 0.04170423, "learning_rate": 9.986e-05, "epoch": 0.05373669, "global_step/max_steps": "135/2513", "percentage": "5.37%", "elapsed_time": "22m 31s", "remaining_time": "6h 36m 41s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.099908} +{"loss": 0.41349176, "grad_norm": 0.05131029, "learning_rate": 9.985e-05, "epoch": 0.05413474, "global_step/max_steps": "136/2513", "percentage": "5.41%", "elapsed_time": "22m 41s", "remaining_time": "6h 36m 27s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.099926} +{"loss": 0.47103179, "grad_norm": 0.04477016, "learning_rate": 9.985e-05, "epoch": 0.05453279, "global_step/max_steps": "137/2513", "percentage": "5.45%", "elapsed_time": "22m 50s", "remaining_time": "6h 36m 10s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.099957} +{"loss": 0.44553721, "grad_norm": 0.04221403, "learning_rate": 9.984e-05, "epoch": 0.05493084, "global_step/max_steps": "138/2513", "percentage": "5.49%", "elapsed_time": "23m 0s", "remaining_time": "6h 35m 56s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.099974} +{"loss": 0.44288799, "grad_norm": 0.04525968, "learning_rate": 9.984e-05, "epoch": 0.05532889, "global_step/max_steps": "139/2513", "percentage": "5.53%", "elapsed_time": "23m 9s", "remaining_time": "6h 35m 39s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.1} +{"loss": 0.38288018, "grad_norm": 0.04262387, "learning_rate": 9.983e-05, "epoch": 0.05572694, "global_step/max_steps": "140/2513", "percentage": "5.57%", "elapsed_time": "23m 19s", "remaining_time": "6h 35m 24s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100022} +{"loss": 0.40578493, "grad_norm": 0.03923057, "learning_rate": 9.982e-05, "epoch": 0.05612499, "global_step/max_steps": "141/2513", "percentage": "5.61%", "elapsed_time": "23m 29s", "remaining_time": "6h 35m 6s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100055} +{"loss": 0.41789925, "grad_norm": 0.03766845, "learning_rate": 9.982e-05, "epoch": 0.05652304, "global_step/max_steps": "142/2513", "percentage": "5.65%", "elapsed_time": "23m 38s", "remaining_time": "6h 34m 50s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100083} +{"loss": 0.41657943, "grad_norm": 0.04392187, "learning_rate": 9.981e-05, "epoch": 0.05692109, "global_step/max_steps": "143/2513", "percentage": "5.69%", "elapsed_time": "23m 48s", "remaining_time": "6h 34m 32s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100116} +{"loss": 0.42920128, "grad_norm": 0.0393295, "learning_rate": 9.981e-05, "epoch": 0.05731914, "global_step/max_steps": "144/2513", "percentage": "5.73%", "elapsed_time": "23m 59s", "remaining_time": "6h 34m 35s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100061} +{"loss": 0.42768866, "grad_norm": 0.0406664, "learning_rate": 9.98e-05, "epoch": 0.05771719, "global_step/max_steps": "145/2513", "percentage": "5.77%", "elapsed_time": "24m 9s", "remaining_time": "6h 34m 37s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10001} +{"loss": 0.46423417, "grad_norm": 0.04011628, "learning_rate": 9.98e-05, "epoch": 0.05811524, "global_step/max_steps": "146/2513", "percentage": "5.81%", "elapsed_time": "24m 20s", "remaining_time": "6h 34m 30s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.099998} +{"loss": 0.49012855, "grad_norm": 0.04006903, "learning_rate": 9.979e-05, "epoch": 0.05851328, "global_step/max_steps": "147/2513", "percentage": "5.85%", "elapsed_time": "24m 29s", "remaining_time": "6h 34m 16s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100017} +{"loss": 0.42608371, "grad_norm": 0.04280027, "learning_rate": 9.978e-05, "epoch": 0.05891133, "global_step/max_steps": "148/2513", "percentage": "5.89%", "elapsed_time": "24m 39s", "remaining_time": "6h 33m 58s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10005} +{"loss": 0.46243805, "grad_norm": 0.0396251, "learning_rate": 9.978e-05, "epoch": 0.05930938, "global_step/max_steps": "149/2513", "percentage": "5.93%", "elapsed_time": "24m 48s", "remaining_time": "6h 33m 38s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100091} +{"loss": 0.40444827, "grad_norm": 0.0368268, "learning_rate": 9.977e-05, "epoch": 0.05970743, "global_step/max_steps": "150/2513", "percentage": "5.97%", "elapsed_time": "24m 58s", "remaining_time": "6h 33m 22s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100118} +{"loss": 0.40355939, "grad_norm": 0.04181711, "learning_rate": 9.977e-05, "epoch": 0.06010548, "global_step/max_steps": "151/2513", "percentage": "6.01%", "elapsed_time": "25m 8s", "remaining_time": "6h 33m 10s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100123} +{"loss": 0.38007605, "grad_norm": 0.04006226, "learning_rate": 9.976e-05, "epoch": 0.06050353, "global_step/max_steps": "152/2513", "percentage": "6.05%", "elapsed_time": "25m 17s", "remaining_time": "6h 32m 58s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100134} +{"loss": 0.44188285, "grad_norm": 0.04224454, "learning_rate": 9.975e-05, "epoch": 0.06090158, "global_step/max_steps": "153/2513", "percentage": "6.09%", "elapsed_time": "25m 27s", "remaining_time": "6h 32m 42s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100159} +{"loss": 0.38021713, "grad_norm": 0.03948209, "learning_rate": 9.975e-05, "epoch": 0.06129963, "global_step/max_steps": "154/2513", "percentage": "6.13%", "elapsed_time": "25m 36s", "remaining_time": "6h 32m 21s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100204} +{"loss": 0.4046151, "grad_norm": 0.04034552, "learning_rate": 9.974e-05, "epoch": 0.06169768, "global_step/max_steps": "155/2513", "percentage": "6.17%", "elapsed_time": "25m 46s", "remaining_time": "6h 32m 8s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10022} +{"loss": 0.45719361, "grad_norm": 0.04262317, "learning_rate": 9.973e-05, "epoch": 0.06209573, "global_step/max_steps": "156/2513", "percentage": "6.21%", "elapsed_time": "25m 56s", "remaining_time": "6h 31m 51s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100249} +{"loss": 0.44340298, "grad_norm": 0.04202209, "learning_rate": 9.973e-05, "epoch": 0.06249378, "global_step/max_steps": "157/2513", "percentage": "6.25%", "elapsed_time": "26m 5s", "remaining_time": "6h 31m 37s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100265} +{"loss": 0.4208805, "grad_norm": 0.04386898, "learning_rate": 9.972e-05, "epoch": 0.06289183, "global_step/max_steps": "158/2513", "percentage": "6.29%", "elapsed_time": "26m 16s", "remaining_time": "6h 31m 34s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100237} +{"loss": 0.42155051, "grad_norm": 0.03714408, "learning_rate": 9.971e-05, "epoch": 0.06328988, "global_step/max_steps": "159/2513", "percentage": "6.33%", "elapsed_time": "26m 26s", "remaining_time": "6h 31m 29s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100215} +{"loss": 0.4108274, "grad_norm": 0.04400492, "learning_rate": 9.971e-05, "epoch": 0.06368793, "global_step/max_steps": "160/2513", "percentage": "6.37%", "elapsed_time": "26m 37s", "remaining_time": "6h 31m 33s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100156} +{"loss": 0.40703824, "grad_norm": 0.03960423, "learning_rate": 9.97e-05, "epoch": 0.06408598, "global_step/max_steps": "161/2513", "percentage": "6.41%", "elapsed_time": "26m 46s", "remaining_time": "6h 31m 14s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100195} +{"loss": 0.44042516, "grad_norm": 0.04117991, "learning_rate": 9.969e-05, "epoch": 0.06448403, "global_step/max_steps": "162/2513", "percentage": "6.45%", "elapsed_time": "26m 56s", "remaining_time": "6h 30m 56s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100226} +{"loss": 0.42304483, "grad_norm": 0.04577458, "learning_rate": 9.969e-05, "epoch": 0.06488208, "global_step/max_steps": "163/2513", "percentage": "6.49%", "elapsed_time": "27m 5s", "remaining_time": "6h 30m 40s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100253} +{"loss": 0.38179314, "grad_norm": 0.03980219, "learning_rate": 9.968e-05, "epoch": 0.06528013, "global_step/max_steps": "164/2513", "percentage": "6.53%", "elapsed_time": "27m 15s", "remaining_time": "6h 30m 25s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100276} +{"loss": 0.45506608, "grad_norm": 0.04363962, "learning_rate": 9.967e-05, "epoch": 0.06567818, "global_step/max_steps": "165/2513", "percentage": "6.57%", "elapsed_time": "27m 24s", "remaining_time": "6h 30m 6s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100314} +{"loss": 0.42806238, "grad_norm": 0.04431815, "learning_rate": 9.966e-05, "epoch": 0.06607623, "global_step/max_steps": "166/2513", "percentage": "6.61%", "elapsed_time": "27m 34s", "remaining_time": "6h 29m 46s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100357} +{"loss": 0.44230402, "grad_norm": 0.04031154, "learning_rate": 9.966e-05, "epoch": 0.06647428, "global_step/max_steps": "167/2513", "percentage": "6.65%", "elapsed_time": "27m 43s", "remaining_time": "6h 29m 31s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100378} +{"loss": 0.38662475, "grad_norm": 0.03662058, "learning_rate": 9.965e-05, "epoch": 0.06687233, "global_step/max_steps": "168/2513", "percentage": "6.69%", "elapsed_time": "27m 53s", "remaining_time": "6h 29m 21s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100377} +{"loss": 0.42420629, "grad_norm": 0.0408593, "learning_rate": 9.964e-05, "epoch": 0.06727038, "global_step/max_steps": "169/2513", "percentage": "6.73%", "elapsed_time": "28m 3s", "remaining_time": "6h 29m 11s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10038} +{"loss": 0.38610825, "grad_norm": 0.04358715, "learning_rate": 9.963e-05, "epoch": 0.06766842, "global_step/max_steps": "170/2513", "percentage": "6.76%", "elapsed_time": "28m 13s", "remaining_time": "6h 28m 55s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100406} +{"loss": 0.42256689, "grad_norm": 0.04487713, "learning_rate": 9.963e-05, "epoch": 0.06806647, "global_step/max_steps": "171/2513", "percentage": "6.80%", "elapsed_time": "28m 23s", "remaining_time": "6h 28m 52s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100374} +{"loss": 0.37971812, "grad_norm": 0.03857103, "learning_rate": 9.962e-05, "epoch": 0.06846452, "global_step/max_steps": "172/2513", "percentage": "6.84%", "elapsed_time": "28m 33s", "remaining_time": "6h 28m 37s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100396} +{"loss": 0.46156105, "grad_norm": 0.04515174, "learning_rate": 9.961e-05, "epoch": 0.06886257, "global_step/max_steps": "173/2513", "percentage": "6.88%", "elapsed_time": "28m 43s", "remaining_time": "6h 28m 28s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100392} +{"loss": 0.39867973, "grad_norm": 0.03983258, "learning_rate": 9.96e-05, "epoch": 0.06926062, "global_step/max_steps": "174/2513", "percentage": "6.92%", "elapsed_time": "28m 54s", "remaining_time": "6h 28m 31s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100338} +{"loss": 0.44861019, "grad_norm": 0.04497159, "learning_rate": 9.959e-05, "epoch": 0.06965867, "global_step/max_steps": "175/2513", "percentage": "6.96%", "elapsed_time": "29m 4s", "remaining_time": "6h 28m 23s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100328} +{"loss": 0.46851137, "grad_norm": 0.04425122, "learning_rate": 9.959e-05, "epoch": 0.07005672, "global_step/max_steps": "176/2513", "percentage": "7.00%", "elapsed_time": "29m 13s", "remaining_time": "6h 28m 9s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100346} +{"loss": 0.4479115, "grad_norm": 0.04343737, "learning_rate": 9.958e-05, "epoch": 0.07045477, "global_step/max_steps": "177/2513", "percentage": "7.04%", "elapsed_time": "29m 23s", "remaining_time": "6h 27m 56s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100359} +{"loss": 0.39941305, "grad_norm": 0.03739605, "learning_rate": 9.957e-05, "epoch": 0.07085282, "global_step/max_steps": "178/2513", "percentage": "7.08%", "elapsed_time": "29m 33s", "remaining_time": "6h 27m 42s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100374} +{"loss": 0.3496874, "grad_norm": 0.04127823, "learning_rate": 9.956e-05, "epoch": 0.07125087, "global_step/max_steps": "179/2513", "percentage": "7.12%", "elapsed_time": "29m 43s", "remaining_time": "6h 27m 31s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100381} +{"loss": 0.41580027, "grad_norm": 0.04055145, "learning_rate": 9.955e-05, "epoch": 0.07164892, "global_step/max_steps": "180/2513", "percentage": "7.16%", "elapsed_time": "29m 52s", "remaining_time": "6h 27m 19s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100391} +{"loss": 0.41803792, "grad_norm": 0.04232781, "learning_rate": 9.954e-05, "epoch": 0.07204697, "global_step/max_steps": "181/2513", "percentage": "7.20%", "elapsed_time": "30m 2s", "remaining_time": "6h 27m 2s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100419} +{"loss": 0.48920214, "grad_norm": 0.04501472, "learning_rate": 9.953e-05, "epoch": 0.07244502, "global_step/max_steps": "182/2513", "percentage": "7.24%", "elapsed_time": "30m 12s", "remaining_time": "6h 26m 54s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100414} +{"loss": 0.36994201, "grad_norm": 0.042242, "learning_rate": 9.953e-05, "epoch": 0.07284307, "global_step/max_steps": "183/2513", "percentage": "7.28%", "elapsed_time": "30m 21s", "remaining_time": "6h 26m 36s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100444} +{"loss": 0.41931775, "grad_norm": 0.04251162, "learning_rate": 9.952e-05, "epoch": 0.07324112, "global_step/max_steps": "184/2513", "percentage": "7.32%", "elapsed_time": "30m 31s", "remaining_time": "6h 26m 22s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100464} +{"loss": 0.45457688, "grad_norm": 0.04130198, "learning_rate": 9.951e-05, "epoch": 0.07363917, "global_step/max_steps": "185/2513", "percentage": "7.36%", "elapsed_time": "30m 41s", "remaining_time": "6h 26m 10s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100471} +{"loss": 0.41679853, "grad_norm": 0.039039, "learning_rate": 9.95e-05, "epoch": 0.07403722, "global_step/max_steps": "186/2513", "percentage": "7.40%", "elapsed_time": "30m 50s", "remaining_time": "6h 25m 54s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100499} +{"loss": 0.42772388, "grad_norm": 0.03830274, "learning_rate": 9.949e-05, "epoch": 0.07443527, "global_step/max_steps": "187/2513", "percentage": "7.44%", "elapsed_time": "31m 0s", "remaining_time": "6h 25m 37s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100531} +{"loss": 0.39904475, "grad_norm": 0.03690683, "learning_rate": 9.948e-05, "epoch": 0.07483332, "global_step/max_steps": "188/2513", "percentage": "7.48%", "elapsed_time": "31m 10s", "remaining_time": "6h 25m 33s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100503} +{"loss": 0.44359234, "grad_norm": 0.04514274, "learning_rate": 9.947e-05, "epoch": 0.07523137, "global_step/max_steps": "189/2513", "percentage": "7.52%", "elapsed_time": "31m 21s", "remaining_time": "6h 25m 34s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100454} +{"loss": 0.41584384, "grad_norm": 0.04113187, "learning_rate": 9.946e-05, "epoch": 0.07562942, "global_step/max_steps": "190/2513", "percentage": "7.56%", "elapsed_time": "31m 31s", "remaining_time": "6h 25m 22s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100465} +{"loss": 0.43835366, "grad_norm": 0.04583373, "learning_rate": 9.945e-05, "epoch": 0.07602747, "global_step/max_steps": "191/2513", "percentage": "7.60%", "elapsed_time": "31m 41s", "remaining_time": "6h 25m 11s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10047} +{"loss": 0.39380682, "grad_norm": 0.04142394, "learning_rate": 9.944e-05, "epoch": 0.07642551, "global_step/max_steps": "192/2513", "percentage": "7.64%", "elapsed_time": "31m 50s", "remaining_time": "6h 24m 56s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10049} +{"loss": 0.41637111, "grad_norm": 0.04151365, "learning_rate": 9.943e-05, "epoch": 0.07682356, "global_step/max_steps": "193/2513", "percentage": "7.68%", "elapsed_time": "32m 0s", "remaining_time": "6h 24m 42s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100508} +{"loss": 0.46362555, "grad_norm": 0.04510412, "learning_rate": 9.942e-05, "epoch": 0.07722161, "global_step/max_steps": "194/2513", "percentage": "7.72%", "elapsed_time": "32m 9s", "remaining_time": "6h 24m 26s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100535} +{"loss": 0.42084819, "grad_norm": 0.04305697, "learning_rate": 9.941e-05, "epoch": 0.07761966, "global_step/max_steps": "195/2513", "percentage": "7.76%", "elapsed_time": "32m 19s", "remaining_time": "6h 24m 13s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10055} +{"loss": 0.38593468, "grad_norm": 0.03656732, "learning_rate": 9.94e-05, "epoch": 0.07801771, "global_step/max_steps": "196/2513", "percentage": "7.80%", "elapsed_time": "32m 29s", "remaining_time": "6h 24m 0s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100564} +{"loss": 0.44058341, "grad_norm": 0.03997835, "learning_rate": 9.939e-05, "epoch": 0.07841576, "global_step/max_steps": "197/2513", "percentage": "7.84%", "elapsed_time": "32m 38s", "remaining_time": "6h 23m 47s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100577} +{"loss": 0.39587572, "grad_norm": 0.03827296, "learning_rate": 9.938e-05, "epoch": 0.07881381, "global_step/max_steps": "198/2513", "percentage": "7.88%", "elapsed_time": "32m 48s", "remaining_time": "6h 23m 35s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100585} +{"loss": 0.39354119, "grad_norm": 0.0425915, "learning_rate": 9.937e-05, "epoch": 0.07921186, "global_step/max_steps": "199/2513", "percentage": "7.92%", "elapsed_time": "32m 58s", "remaining_time": "6h 23m 21s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100601} +{"loss": 0.45415545, "grad_norm": 0.04769524, "learning_rate": 9.936e-05, "epoch": 0.07960991, "global_step/max_steps": "200/2513", "percentage": "7.96%", "elapsed_time": "33m 8s", "remaining_time": "6h 23m 12s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100596} +{"loss": 0.37071097, "grad_norm": 0.0353796, "learning_rate": 9.935e-05, "epoch": 0.08000796, "global_step/max_steps": "201/2513", "percentage": "8.00%", "elapsed_time": "33m 19s", "remaining_time": "6h 23m 18s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10053} +{"loss": 0.44706288, "grad_norm": 0.04283047, "learning_rate": 9.934e-05, "epoch": 0.08040601, "global_step/max_steps": "202/2513", "percentage": "8.04%", "elapsed_time": "33m 29s", "remaining_time": "6h 23m 12s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100512} +{"loss": 0.42771566, "grad_norm": 0.03941539, "learning_rate": 9.933e-05, "epoch": 0.08080406, "global_step/max_steps": "203/2513", "percentage": "8.08%", "elapsed_time": "33m 40s", "remaining_time": "6h 23m 9s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10048} +{"loss": 0.42674387, "grad_norm": 0.04517798, "learning_rate": 9.932e-05, "epoch": 0.08120211, "global_step/max_steps": "204/2513", "percentage": "8.12%", "elapsed_time": "33m 50s", "remaining_time": "6h 23m 5s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100455} +{"loss": 0.45254311, "grad_norm": 0.04286423, "learning_rate": 9.931e-05, "epoch": 0.08160016, "global_step/max_steps": "205/2513", "percentage": "8.16%", "elapsed_time": "34m 0s", "remaining_time": "6h 22m 53s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100464} +{"loss": 0.41461694, "grad_norm": 0.03787321, "learning_rate": 9.93e-05, "epoch": 0.08199821, "global_step/max_steps": "206/2513", "percentage": "8.20%", "elapsed_time": "34m 10s", "remaining_time": "6h 22m 39s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10048} +{"loss": 0.43528092, "grad_norm": 0.04302459, "learning_rate": 9.929e-05, "epoch": 0.08239626, "global_step/max_steps": "207/2513", "percentage": "8.24%", "elapsed_time": "34m 20s", "remaining_time": "6h 22m 37s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100445} +{"loss": 0.40190589, "grad_norm": 0.0387694, "learning_rate": 9.928e-05, "epoch": 0.08279431, "global_step/max_steps": "208/2513", "percentage": "8.28%", "elapsed_time": "34m 30s", "remaining_time": "6h 22m 25s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100455} +{"loss": 0.44457823, "grad_norm": 0.04057871, "learning_rate": 9.927e-05, "epoch": 0.08319236, "global_step/max_steps": "209/2513", "percentage": "8.32%", "elapsed_time": "34m 40s", "remaining_time": "6h 22m 12s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100469} +{"loss": 0.41339478, "grad_norm": 0.04051652, "learning_rate": 9.926e-05, "epoch": 0.08359041, "global_step/max_steps": "210/2513", "percentage": "8.36%", "elapsed_time": "34m 50s", "remaining_time": "6h 22m 2s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100471} +{"loss": 0.42213231, "grad_norm": 0.0409466, "learning_rate": 9.924e-05, "epoch": 0.08398846, "global_step/max_steps": "211/2513", "percentage": "8.40%", "elapsed_time": "35m 0s", "remaining_time": "6h 21m 51s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100476} +{"loss": 0.37747318, "grad_norm": 0.04093783, "learning_rate": 9.923e-05, "epoch": 0.08438651, "global_step/max_steps": "212/2513", "percentage": "8.44%", "elapsed_time": "35m 9s", "remaining_time": "6h 21m 36s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100496} +{"loss": 0.41167977, "grad_norm": 0.03921886, "learning_rate": 9.922e-05, "epoch": 0.08478456, "global_step/max_steps": "213/2513", "percentage": "8.48%", "elapsed_time": "35m 19s", "remaining_time": "6h 21m 27s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10049} +{"loss": 0.4256435, "grad_norm": 0.03942347, "learning_rate": 9.921e-05, "epoch": 0.08518261, "global_step/max_steps": "214/2513", "percentage": "8.52%", "elapsed_time": "35m 29s", "remaining_time": "6h 21m 14s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100503} +{"loss": 0.44722259, "grad_norm": 0.04276415, "learning_rate": 9.92e-05, "epoch": 0.08558065, "global_step/max_steps": "215/2513", "percentage": "8.56%", "elapsed_time": "35m 39s", "remaining_time": "6h 21m 3s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100508} +{"loss": 0.46179205, "grad_norm": 0.04206607, "learning_rate": 9.919e-05, "epoch": 0.0859787, "global_step/max_steps": "216/2513", "percentage": "8.60%", "elapsed_time": "35m 49s", "remaining_time": "6h 20m 54s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100506} +{"loss": 0.35863683, "grad_norm": 0.03802703, "learning_rate": 9.918e-05, "epoch": 0.08637675, "global_step/max_steps": "217/2513", "percentage": "8.64%", "elapsed_time": "35m 59s", "remaining_time": "6h 20m 52s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10047} +{"loss": 0.41153759, "grad_norm": 0.0407595, "learning_rate": 9.916e-05, "epoch": 0.0867748, "global_step/max_steps": "218/2513", "percentage": "8.67%", "elapsed_time": "36m 10s", "remaining_time": "6h 20m 50s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100436} +{"loss": 0.42727566, "grad_norm": 0.0435717, "learning_rate": 9.915e-05, "epoch": 0.08717285, "global_step/max_steps": "219/2513", "percentage": "8.71%", "elapsed_time": "36m 20s", "remaining_time": "6h 20m 44s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100419} +{"loss": 0.40917438, "grad_norm": 0.04182773, "learning_rate": 9.914e-05, "epoch": 0.0875709, "global_step/max_steps": "220/2513", "percentage": "8.75%", "elapsed_time": "36m 30s", "remaining_time": "6h 20m 32s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100428} +{"loss": 0.43094593, "grad_norm": 0.04157738, "learning_rate": 9.913e-05, "epoch": 0.08796895, "global_step/max_steps": "221/2513", "percentage": "8.79%", "elapsed_time": "36m 40s", "remaining_time": "6h 20m 20s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100437} +{"loss": 0.3783294, "grad_norm": 0.03694082, "learning_rate": 9.912e-05, "epoch": 0.088367, "global_step/max_steps": "222/2513", "percentage": "8.83%", "elapsed_time": "36m 49s", "remaining_time": "6h 20m 4s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10046} +{"loss": 0.43967968, "grad_norm": 0.0412313, "learning_rate": 9.91e-05, "epoch": 0.08876505, "global_step/max_steps": "223/2513", "percentage": "8.87%", "elapsed_time": "36m 59s", "remaining_time": "6h 19m 50s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10048} +{"loss": 0.38267374, "grad_norm": 0.03779399, "learning_rate": 9.909e-05, "epoch": 0.0891631, "global_step/max_steps": "224/2513", "percentage": "8.91%", "elapsed_time": "37m 9s", "remaining_time": "6h 19m 39s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100484} +{"loss": 0.42980745, "grad_norm": 0.04180725, "learning_rate": 9.908e-05, "epoch": 0.08956115, "global_step/max_steps": "225/2513", "percentage": "8.95%", "elapsed_time": "37m 19s", "remaining_time": "6h 19m 30s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100481} +{"loss": 0.41687331, "grad_norm": 0.0415656, "learning_rate": 9.907e-05, "epoch": 0.0899592, "global_step/max_steps": "226/2513", "percentage": "8.99%", "elapsed_time": "37m 28s", "remaining_time": "6h 19m 17s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100496} +{"loss": 0.39908093, "grad_norm": 0.04048435, "learning_rate": 9.906e-05, "epoch": 0.09035725, "global_step/max_steps": "227/2513", "percentage": "9.03%", "elapsed_time": "37m 38s", "remaining_time": "6h 19m 4s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100509} +{"loss": 0.41864547, "grad_norm": 0.03815248, "learning_rate": 9.904e-05, "epoch": 0.0907553, "global_step/max_steps": "228/2513", "percentage": "9.07%", "elapsed_time": "37m 48s", "remaining_time": "6h 18m 50s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100527} +{"loss": 0.39103919, "grad_norm": 0.03708798, "learning_rate": 9.903e-05, "epoch": 0.09115335, "global_step/max_steps": "229/2513", "percentage": "9.11%", "elapsed_time": "37m 57s", "remaining_time": "6h 18m 40s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100528} +{"loss": 0.40441704, "grad_norm": 0.03885075, "learning_rate": 9.902e-05, "epoch": 0.0915514, "global_step/max_steps": "230/2513", "percentage": "9.15%", "elapsed_time": "38m 7s", "remaining_time": "6h 18m 29s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100531} +{"loss": 0.40901124, "grad_norm": 0.04401333, "learning_rate": 9.901e-05, "epoch": 0.09194945, "global_step/max_steps": "231/2513", "percentage": "9.19%", "elapsed_time": "38m 17s", "remaining_time": "6h 18m 17s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100538} +{"loss": 0.38673049, "grad_norm": 0.03772669, "learning_rate": 9.899e-05, "epoch": 0.0923475, "global_step/max_steps": "232/2513", "percentage": "9.23%", "elapsed_time": "38m 28s", "remaining_time": "6h 18m 12s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100517} +{"loss": 0.41287696, "grad_norm": 0.03990328, "learning_rate": 9.898e-05, "epoch": 0.09274555, "global_step/max_steps": "233/2513", "percentage": "9.27%", "elapsed_time": "38m 39s", "remaining_time": "6h 18m 14s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100464} +{"loss": 0.45152929, "grad_norm": 0.0409604, "learning_rate": 9.897e-05, "epoch": 0.0931436, "global_step/max_steps": "234/2513", "percentage": "9.31%", "elapsed_time": "38m 49s", "remaining_time": "6h 18m 7s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100454} +{"loss": 0.42877167, "grad_norm": 0.03932695, "learning_rate": 9.895e-05, "epoch": 0.09354165, "global_step/max_steps": "235/2513", "percentage": "9.35%", "elapsed_time": "38m 59s", "remaining_time": "6h 17m 56s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100456} +{"loss": 0.43262047, "grad_norm": 0.03980766, "learning_rate": 9.894e-05, "epoch": 0.0939397, "global_step/max_steps": "236/2513", "percentage": "9.39%", "elapsed_time": "39m 9s", "remaining_time": "6h 17m 45s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100459} +{"loss": 0.44375235, "grad_norm": 0.03895334, "learning_rate": 9.893e-05, "epoch": 0.09433775, "global_step/max_steps": "237/2513", "percentage": "9.43%", "elapsed_time": "39m 19s", "remaining_time": "6h 17m 34s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100466} +{"loss": 0.3981334, "grad_norm": 0.04013609, "learning_rate": 9.891e-05, "epoch": 0.09473579, "global_step/max_steps": "238/2513", "percentage": "9.47%", "elapsed_time": "39m 28s", "remaining_time": "6h 17m 22s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100474} +{"loss": 0.4182632, "grad_norm": 0.03949301, "learning_rate": 9.89e-05, "epoch": 0.09513384, "global_step/max_steps": "239/2513", "percentage": "9.51%", "elapsed_time": "39m 38s", "remaining_time": "6h 17m 9s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100487} +{"loss": 0.39613017, "grad_norm": 0.03941763, "learning_rate": 9.889e-05, "epoch": 0.09553189, "global_step/max_steps": "240/2513", "percentage": "9.55%", "elapsed_time": "39m 48s", "remaining_time": "6h 16m 56s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100501} +{"loss": 0.36194277, "grad_norm": 0.04054167, "learning_rate": 9.887e-05, "epoch": 0.09592994, "global_step/max_steps": "241/2513", "percentage": "9.59%", "elapsed_time": "39m 57s", "remaining_time": "6h 16m 44s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10051} +{"loss": 0.45620698, "grad_norm": 0.04391407, "learning_rate": 9.886e-05, "epoch": 0.09632799, "global_step/max_steps": "242/2513", "percentage": "9.63%", "elapsed_time": "40m 7s", "remaining_time": "6h 16m 31s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100524} +{"loss": 0.31752869, "grad_norm": 0.03593024, "learning_rate": 9.885e-05, "epoch": 0.09672604, "global_step/max_steps": "243/2513", "percentage": "9.67%", "elapsed_time": "40m 16s", "remaining_time": "6h 16m 18s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100539} +{"loss": 0.3816753, "grad_norm": 0.0435882, "learning_rate": 9.883e-05, "epoch": 0.09712409, "global_step/max_steps": "244/2513", "percentage": "9.71%", "elapsed_time": "40m 26s", "remaining_time": "6h 16m 6s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100547} +{"loss": 0.41833979, "grad_norm": 0.03788602, "learning_rate": 9.882e-05, "epoch": 0.09752214, "global_step/max_steps": "245/2513", "percentage": "9.75%", "elapsed_time": "40m 36s", "remaining_time": "6h 15m 54s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100554} +{"loss": 0.4607124, "grad_norm": 0.04262907, "learning_rate": 9.88e-05, "epoch": 0.09792019, "global_step/max_steps": "246/2513", "percentage": "9.79%", "elapsed_time": "40m 46s", "remaining_time": "6h 15m 44s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100556} +{"loss": 0.39882523, "grad_norm": 0.04106831, "learning_rate": 9.879e-05, "epoch": 0.09831824, "global_step/max_steps": "247/2513", "percentage": "9.83%", "elapsed_time": "40m 57s", "remaining_time": "6h 15m 43s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100516} +{"loss": 0.37272185, "grad_norm": 0.03894518, "learning_rate": 9.878e-05, "epoch": 0.09871629, "global_step/max_steps": "248/2513", "percentage": "9.87%", "elapsed_time": "41m 7s", "remaining_time": "6h 15m 39s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100492} +{"loss": 0.41568145, "grad_norm": 0.03973812, "learning_rate": 9.876e-05, "epoch": 0.09911434, "global_step/max_steps": "249/2513", "percentage": "9.91%", "elapsed_time": "41m 17s", "remaining_time": "6h 15m 27s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100499} +{"loss": 0.43539131, "grad_norm": 0.04112942, "learning_rate": 9.875e-05, "epoch": 0.09951239, "global_step/max_steps": "250/2513", "percentage": "9.95%", "elapsed_time": "41m 27s", "remaining_time": "6h 15m 15s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100507} +{"loss": 0.40725282, "grad_norm": 0.03531125, "learning_rate": 9.873e-05, "epoch": 0.09991044, "global_step/max_steps": "251/2513", "percentage": "9.99%", "elapsed_time": "41m 38s", "remaining_time": "6h 15m 14s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10047} +{"loss": 0.39336836, "grad_norm": 0.03836723, "learning_rate": 9.872e-05, "epoch": 0.10030849, "global_step/max_steps": "252/2513", "percentage": "10.03%", "elapsed_time": "41m 48s", "remaining_time": "6h 15m 3s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100474} +{"loss": 0.41192147, "grad_norm": 0.03955725, "learning_rate": 9.87e-05, "epoch": 0.10070654, "global_step/max_steps": "253/2513", "percentage": "10.07%", "elapsed_time": "41m 57s", "remaining_time": "6h 14m 50s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100487} +{"loss": 0.4600248, "grad_norm": 0.04591149, "learning_rate": 9.869e-05, "epoch": 0.10110459, "global_step/max_steps": "254/2513", "percentage": "10.11%", "elapsed_time": "42m 7s", "remaining_time": "6h 14m 38s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100495} +{"loss": 0.4163883, "grad_norm": 0.04046923, "learning_rate": 9.867e-05, "epoch": 0.10150264, "global_step/max_steps": "255/2513", "percentage": "10.15%", "elapsed_time": "42m 17s", "remaining_time": "6h 14m 25s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100508} +{"loss": 0.37764031, "grad_norm": 0.04061568, "learning_rate": 9.866e-05, "epoch": 0.10190069, "global_step/max_steps": "256/2513", "percentage": "10.19%", "elapsed_time": "42m 26s", "remaining_time": "6h 14m 15s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100512} +{"loss": 0.43196949, "grad_norm": 0.04025549, "learning_rate": 9.865e-05, "epoch": 0.10229874, "global_step/max_steps": "257/2513", "percentage": "10.23%", "elapsed_time": "42m 36s", "remaining_time": "6h 14m 2s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100521} +{"loss": 0.43864051, "grad_norm": 0.0411945, "learning_rate": 9.863e-05, "epoch": 0.10269679, "global_step/max_steps": "258/2513", "percentage": "10.27%", "elapsed_time": "42m 46s", "remaining_time": "6h 13m 51s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100526} +{"loss": 0.38343686, "grad_norm": 0.03824353, "learning_rate": 9.862e-05, "epoch": 0.10309484, "global_step/max_steps": "259/2513", "percentage": "10.31%", "elapsed_time": "42m 56s", "remaining_time": "6h 13m 40s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100531} +{"loss": 0.42167279, "grad_norm": 0.04346565, "learning_rate": 9.86e-05, "epoch": 0.10349288, "global_step/max_steps": "260/2513", "percentage": "10.35%", "elapsed_time": "43m 5s", "remaining_time": "6h 13m 26s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10055} +{"loss": 0.38373193, "grad_norm": 0.04362537, "learning_rate": 9.858e-05, "epoch": 0.10389093, "global_step/max_steps": "261/2513", "percentage": "10.39%", "elapsed_time": "43m 16s", "remaining_time": "6h 13m 20s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100532} +{"loss": 0.41099548, "grad_norm": 0.04128423, "learning_rate": 9.857e-05, "epoch": 0.10428898, "global_step/max_steps": "262/2513", "percentage": "10.43%", "elapsed_time": "43m 26s", "remaining_time": "6h 13m 16s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100507} +{"loss": 0.40315944, "grad_norm": 0.04578854, "learning_rate": 9.855e-05, "epoch": 0.10468703, "global_step/max_steps": "263/2513", "percentage": "10.47%", "elapsed_time": "43m 37s", "remaining_time": "6h 13m 9s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100492} +{"loss": 0.38049459, "grad_norm": 0.03888287, "learning_rate": 9.854e-05, "epoch": 0.10508508, "global_step/max_steps": "264/2513", "percentage": "10.51%", "elapsed_time": "43m 46s", "remaining_time": "6h 12m 56s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100509} +{"loss": 0.36438257, "grad_norm": 0.03827494, "learning_rate": 9.852e-05, "epoch": 0.10548313, "global_step/max_steps": "265/2513", "percentage": "10.55%", "elapsed_time": "43m 56s", "remaining_time": "6h 12m 42s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100526} +{"loss": 0.45318666, "grad_norm": 0.04026962, "learning_rate": 9.851e-05, "epoch": 0.10588118, "global_step/max_steps": "266/2513", "percentage": "10.58%", "elapsed_time": "44m 5s", "remaining_time": "6h 12m 30s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100533} +{"loss": 0.43060514, "grad_norm": 0.03975171, "learning_rate": 9.849e-05, "epoch": 0.10627923, "global_step/max_steps": "267/2513", "percentage": "10.62%", "elapsed_time": "44m 15s", "remaining_time": "6h 12m 19s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100538} +{"loss": 0.39059567, "grad_norm": 0.04206689, "learning_rate": 9.848e-05, "epoch": 0.10667728, "global_step/max_steps": "268/2513", "percentage": "10.66%", "elapsed_time": "44m 25s", "remaining_time": "6h 12m 6s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100555} +{"loss": 0.42062604, "grad_norm": 0.04100228, "learning_rate": 9.846e-05, "epoch": 0.10707533, "global_step/max_steps": "269/2513", "percentage": "10.70%", "elapsed_time": "44m 34s", "remaining_time": "6h 11m 53s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100566} +{"loss": 0.42568034, "grad_norm": 0.04123776, "learning_rate": 9.844e-05, "epoch": 0.10747338, "global_step/max_steps": "270/2513", "percentage": "10.74%", "elapsed_time": "44m 44s", "remaining_time": "6h 11m 40s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10058} +{"loss": 0.41734195, "grad_norm": 0.04407817, "learning_rate": 9.843e-05, "epoch": 0.10787143, "global_step/max_steps": "271/2513", "percentage": "10.78%", "elapsed_time": "44m 54s", "remaining_time": "6h 11m 29s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100583} +{"loss": 0.36896312, "grad_norm": 0.0391029, "learning_rate": 9.841e-05, "epoch": 0.10826948, "global_step/max_steps": "272/2513", "percentage": "10.82%", "elapsed_time": "45m 4s", "remaining_time": "6h 11m 19s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100585} +{"loss": 0.42982653, "grad_norm": 0.042474, "learning_rate": 9.84e-05, "epoch": 0.10866753, "global_step/max_steps": "273/2513", "percentage": "10.86%", "elapsed_time": "45m 13s", "remaining_time": "6h 11m 7s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100593} +{"loss": 0.4388338, "grad_norm": 0.04123165, "learning_rate": 9.838e-05, "epoch": 0.10906558, "global_step/max_steps": "274/2513", "percentage": "10.90%", "elapsed_time": "45m 23s", "remaining_time": "6h 10m 52s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100617} +{"loss": 0.38788861, "grad_norm": 0.04026947, "learning_rate": 9.836e-05, "epoch": 0.10946363, "global_step/max_steps": "275/2513", "percentage": "10.94%", "elapsed_time": "45m 32s", "remaining_time": "6h 10m 41s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100623} +{"loss": 0.35903955, "grad_norm": 0.03945483, "learning_rate": 9.835e-05, "epoch": 0.10986168, "global_step/max_steps": "276/2513", "percentage": "10.98%", "elapsed_time": "45m 43s", "remaining_time": "6h 10m 39s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100586} +{"loss": 0.39957649, "grad_norm": 0.04015244, "learning_rate": 9.833e-05, "epoch": 0.11025973, "global_step/max_steps": "277/2513", "percentage": "11.02%", "elapsed_time": "45m 54s", "remaining_time": "6h 10m 33s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100571} +{"loss": 0.40955788, "grad_norm": 0.03867542, "learning_rate": 9.831e-05, "epoch": 0.11065778, "global_step/max_steps": "278/2513", "percentage": "11.06%", "elapsed_time": "46m 4s", "remaining_time": "6h 10m 24s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100566} +{"loss": 0.43313831, "grad_norm": 0.03991583, "learning_rate": 9.83e-05, "epoch": 0.11105583, "global_step/max_steps": "279/2513", "percentage": "11.10%", "elapsed_time": "46m 13s", "remaining_time": "6h 10m 9s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100585} +{"loss": 0.45130181, "grad_norm": 0.03932567, "learning_rate": 9.828e-05, "epoch": 0.11145388, "global_step/max_steps": "280/2513", "percentage": "11.14%", "elapsed_time": "46m 23s", "remaining_time": "6h 9m 57s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100598} +{"loss": 0.3920207, "grad_norm": 0.03882095, "learning_rate": 9.826e-05, "epoch": 0.11185193, "global_step/max_steps": "281/2513", "percentage": "11.18%", "elapsed_time": "46m 32s", "remaining_time": "6h 9m 44s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100609} +{"loss": 0.39698452, "grad_norm": 0.03905571, "learning_rate": 9.825e-05, "epoch": 0.11224998, "global_step/max_steps": "282/2513", "percentage": "11.22%", "elapsed_time": "46m 42s", "remaining_time": "6h 9m 34s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10061} +{"loss": 0.39969841, "grad_norm": 0.03772588, "learning_rate": 9.823e-05, "epoch": 0.11264802, "global_step/max_steps": "283/2513", "percentage": "11.26%", "elapsed_time": "46m 52s", "remaining_time": "6h 9m 24s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100612} +{"loss": 0.44417632, "grad_norm": 0.04025221, "learning_rate": 9.821e-05, "epoch": 0.11304607, "global_step/max_steps": "284/2513", "percentage": "11.30%", "elapsed_time": "47m 2s", "remaining_time": "6h 9m 10s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10063} +{"loss": 0.37657857, "grad_norm": 0.03929229, "learning_rate": 9.82e-05, "epoch": 0.11344412, "global_step/max_steps": "285/2513", "percentage": "11.34%", "elapsed_time": "47m 11s", "remaining_time": "6h 8m 58s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100638} +{"loss": 0.39685008, "grad_norm": 0.03794652, "learning_rate": 9.818e-05, "epoch": 0.11384217, "global_step/max_steps": "286/2513", "percentage": "11.38%", "elapsed_time": "47m 21s", "remaining_time": "6h 8m 46s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100649} +{"loss": 0.45299655, "grad_norm": 0.04021295, "learning_rate": 9.816e-05, "epoch": 0.11424022, "global_step/max_steps": "287/2513", "percentage": "11.42%", "elapsed_time": "47m 31s", "remaining_time": "6h 8m 34s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100656} +{"loss": 0.42825806, "grad_norm": 0.0441743, "learning_rate": 9.814e-05, "epoch": 0.11463827, "global_step/max_steps": "288/2513", "percentage": "11.46%", "elapsed_time": "47m 40s", "remaining_time": "6h 8m 22s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100667} +{"loss": 0.40483481, "grad_norm": 0.04330594, "learning_rate": 9.813e-05, "epoch": 0.11503632, "global_step/max_steps": "289/2513", "percentage": "11.50%", "elapsed_time": "47m 50s", "remaining_time": "6h 8m 8s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100684} +{"loss": 0.42288607, "grad_norm": 0.04215585, "learning_rate": 9.811e-05, "epoch": 0.11543437, "global_step/max_steps": "290/2513", "percentage": "11.54%", "elapsed_time": "48m 0s", "remaining_time": "6h 8m 1s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100672} +{"loss": 0.34067124, "grad_norm": 0.04109865, "learning_rate": 9.809e-05, "epoch": 0.11583242, "global_step/max_steps": "291/2513", "percentage": "11.58%", "elapsed_time": "48m 10s", "remaining_time": "6h 7m 53s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100662} +{"loss": 0.4766348, "grad_norm": 0.04569076, "learning_rate": 9.807e-05, "epoch": 0.11623047, "global_step/max_steps": "292/2513", "percentage": "11.62%", "elapsed_time": "48m 21s", "remaining_time": "6h 7m 45s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100655} +{"loss": 0.41257304, "grad_norm": 0.04060029, "learning_rate": 9.806e-05, "epoch": 0.11662852, "global_step/max_steps": "293/2513", "percentage": "11.66%", "elapsed_time": "48m 31s", "remaining_time": "6h 7m 39s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100637} +{"loss": 0.43231922, "grad_norm": 0.03825737, "learning_rate": 9.804e-05, "epoch": 0.11702657, "global_step/max_steps": "294/2513", "percentage": "11.70%", "elapsed_time": "48m 41s", "remaining_time": "6h 7m 28s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100641} +{"loss": 0.41773084, "grad_norm": 0.04086457, "learning_rate": 9.802e-05, "epoch": 0.11742462, "global_step/max_steps": "295/2513", "percentage": "11.74%", "elapsed_time": "48m 51s", "remaining_time": "6h 7m 17s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100647} +{"loss": 0.38483745, "grad_norm": 0.03856602, "learning_rate": 9.8e-05, "epoch": 0.11782267, "global_step/max_steps": "296/2513", "percentage": "11.78%", "elapsed_time": "49m 0s", "remaining_time": "6h 7m 6s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100653} +{"loss": 0.41235018, "grad_norm": 0.04212108, "learning_rate": 9.798e-05, "epoch": 0.11822072, "global_step/max_steps": "297/2513", "percentage": "11.82%", "elapsed_time": "49m 10s", "remaining_time": "6h 6m 55s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100656} +{"loss": 0.48127258, "grad_norm": 0.04580515, "learning_rate": 9.797e-05, "epoch": 0.11861877, "global_step/max_steps": "298/2513", "percentage": "11.86%", "elapsed_time": "49m 19s", "remaining_time": "6h 6m 41s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100676} +{"loss": 0.41397792, "grad_norm": 0.04175884, "learning_rate": 9.795e-05, "epoch": 0.11901682, "global_step/max_steps": "299/2513", "percentage": "11.90%", "elapsed_time": "49m 29s", "remaining_time": "6h 6m 30s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100679} +{"loss": 0.38807914, "grad_norm": 0.03863097, "learning_rate": 9.793e-05, "epoch": 0.11941487, "global_step/max_steps": "300/2513", "percentage": "11.94%", "elapsed_time": "49m 39s", "remaining_time": "6h 6m 19s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100687} +{"loss": 0.42934817, "grad_norm": 0.04147158, "learning_rate": 9.791e-05, "epoch": 0.11981292, "global_step/max_steps": "301/2513", "percentage": "11.98%", "elapsed_time": "49m 50s", "remaining_time": "6h 6m 18s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100642} +{"loss": 0.42295891, "grad_norm": 0.0425803, "learning_rate": 9.789e-05, "epoch": 0.12021097, "global_step/max_steps": "302/2513", "percentage": "12.02%", "elapsed_time": "50m 0s", "remaining_time": "6h 6m 8s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100646} +{"loss": 0.41160041, "grad_norm": 0.03980079, "learning_rate": 9.787e-05, "epoch": 0.12060902, "global_step/max_steps": "303/2513", "percentage": "12.06%", "elapsed_time": "50m 10s", "remaining_time": "6h 5m 57s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100647} +{"loss": 0.4014942, "grad_norm": 0.03907022, "learning_rate": 9.786e-05, "epoch": 0.12100707, "global_step/max_steps": "304/2513", "percentage": "12.10%", "elapsed_time": "50m 20s", "remaining_time": "6h 5m 51s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100629} +{"loss": 0.41204223, "grad_norm": 0.04301613, "learning_rate": 9.784e-05, "epoch": 0.12140511, "global_step/max_steps": "305/2513", "percentage": "12.14%", "elapsed_time": "50m 31s", "remaining_time": "6h 5m 44s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100619} +{"loss": 0.38755208, "grad_norm": 0.03971666, "learning_rate": 9.782e-05, "epoch": 0.12180316, "global_step/max_steps": "306/2513", "percentage": "12.18%", "elapsed_time": "50m 41s", "remaining_time": "6h 5m 36s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100608} +{"loss": 0.42322284, "grad_norm": 0.03904421, "learning_rate": 9.78e-05, "epoch": 0.12220121, "global_step/max_steps": "307/2513", "percentage": "12.22%", "elapsed_time": "50m 51s", "remaining_time": "6h 5m 29s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100597} +{"loss": 0.37198657, "grad_norm": 0.03534119, "learning_rate": 9.778e-05, "epoch": 0.12259926, "global_step/max_steps": "308/2513", "percentage": "12.26%", "elapsed_time": "51m 2s", "remaining_time": "6h 5m 21s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100585} +{"loss": 0.39048862, "grad_norm": 0.03688101, "learning_rate": 9.776e-05, "epoch": 0.12299731, "global_step/max_steps": "309/2513", "percentage": "12.30%", "elapsed_time": "51m 11s", "remaining_time": "6h 5m 10s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100593} +{"loss": 0.4022575, "grad_norm": 0.04171477, "learning_rate": 9.774e-05, "epoch": 0.12339536, "global_step/max_steps": "310/2513", "percentage": "12.34%", "elapsed_time": "51m 21s", "remaining_time": "6h 4m 58s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100601} +{"loss": 0.38312411, "grad_norm": 0.03963149, "learning_rate": 9.772e-05, "epoch": 0.12379341, "global_step/max_steps": "311/2513", "percentage": "12.38%", "elapsed_time": "51m 31s", "remaining_time": "6h 4m 46s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100609} +{"loss": 0.39071497, "grad_norm": 0.04415069, "learning_rate": 9.77e-05, "epoch": 0.12419146, "global_step/max_steps": "312/2513", "percentage": "12.42%", "elapsed_time": "51m 40s", "remaining_time": "6h 4m 35s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100616} +{"loss": 0.40064955, "grad_norm": 0.03856095, "learning_rate": 9.768e-05, "epoch": 0.12458951, "global_step/max_steps": "313/2513", "percentage": "12.46%", "elapsed_time": "51m 50s", "remaining_time": "6h 4m 23s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100624} +{"loss": 0.39380339, "grad_norm": 0.0413079, "learning_rate": 9.767e-05, "epoch": 0.12498756, "global_step/max_steps": "314/2513", "percentage": "12.50%", "elapsed_time": "52m 0s", "remaining_time": "6h 4m 11s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100633} +{"loss": 0.43045393, "grad_norm": 0.04210878, "learning_rate": 9.765e-05, "epoch": 0.12538561, "global_step/max_steps": "315/2513", "percentage": "12.53%", "elapsed_time": "52m 10s", "remaining_time": "6h 4m 1s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100636} +{"loss": 0.40216905, "grad_norm": 0.0388288, "learning_rate": 9.763e-05, "epoch": 0.12578366, "global_step/max_steps": "316/2513", "percentage": "12.57%", "elapsed_time": "52m 19s", "remaining_time": "6h 3m 49s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100644} +{"loss": 0.43507838, "grad_norm": 0.04354027, "learning_rate": 9.761e-05, "epoch": 0.12618171, "global_step/max_steps": "317/2513", "percentage": "12.61%", "elapsed_time": "52m 29s", "remaining_time": "6h 3m 38s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100648} +{"loss": 0.37054333, "grad_norm": 0.04167061, "learning_rate": 9.759e-05, "epoch": 0.12657976, "global_step/max_steps": "318/2513", "percentage": "12.65%", "elapsed_time": "52m 39s", "remaining_time": "6h 3m 25s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100661} +{"loss": 0.39669126, "grad_norm": 0.04182671, "learning_rate": 9.757e-05, "epoch": 0.12697781, "global_step/max_steps": "319/2513", "percentage": "12.69%", "elapsed_time": "52m 49s", "remaining_time": "6h 3m 16s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10066} +{"loss": 0.395558, "grad_norm": 0.04321408, "learning_rate": 9.755e-05, "epoch": 0.12737586, "global_step/max_steps": "320/2513", "percentage": "12.73%", "elapsed_time": "52m 59s", "remaining_time": "6h 3m 6s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100657} +{"loss": 0.37100756, "grad_norm": 0.04009045, "learning_rate": 9.753e-05, "epoch": 0.12777391, "global_step/max_steps": "321/2513", "percentage": "12.77%", "elapsed_time": "53m 9s", "remaining_time": "6h 2m 58s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100651} +{"loss": 0.45274442, "grad_norm": 0.0429568, "learning_rate": 9.751e-05, "epoch": 0.12817196, "global_step/max_steps": "322/2513", "percentage": "12.81%", "elapsed_time": "53m 19s", "remaining_time": "6h 2m 49s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100646} +{"loss": 0.39147022, "grad_norm": 0.03819731, "learning_rate": 9.749e-05, "epoch": 0.12857001, "global_step/max_steps": "323/2513", "percentage": "12.85%", "elapsed_time": "53m 29s", "remaining_time": "6h 2m 41s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100637} +{"loss": 0.41818932, "grad_norm": 0.03987284, "learning_rate": 9.747e-05, "epoch": 0.12896806, "global_step/max_steps": "324/2513", "percentage": "12.89%", "elapsed_time": "53m 39s", "remaining_time": "6h 2m 31s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100637} +{"loss": 0.40861034, "grad_norm": 0.03977091, "learning_rate": 9.745e-05, "epoch": 0.12936611, "global_step/max_steps": "325/2513", "percentage": "12.93%", "elapsed_time": "53m 48s", "remaining_time": "6h 2m 17s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100656} +{"loss": 0.44170418, "grad_norm": 0.04240593, "learning_rate": 9.743e-05, "epoch": 0.12976416, "global_step/max_steps": "326/2513", "percentage": "12.97%", "elapsed_time": "53m 58s", "remaining_time": "6h 2m 5s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100666} +{"loss": 0.46128315, "grad_norm": 0.04224736, "learning_rate": 9.741e-05, "epoch": 0.13016221, "global_step/max_steps": "327/2513", "percentage": "13.01%", "elapsed_time": "54m 8s", "remaining_time": "6h 1m 56s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100663} +{"loss": 0.41231966, "grad_norm": 0.03751595, "learning_rate": 9.738e-05, "epoch": 0.13056025, "global_step/max_steps": "328/2513", "percentage": "13.05%", "elapsed_time": "54m 18s", "remaining_time": "6h 1m 45s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100665} +{"loss": 0.41921073, "grad_norm": 0.04015876, "learning_rate": 9.736e-05, "epoch": 0.1309583, "global_step/max_steps": "329/2513", "percentage": "13.09%", "elapsed_time": "54m 27s", "remaining_time": "6h 1m 31s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100683} +{"loss": 0.4141019, "grad_norm": 0.03938042, "learning_rate": 9.734e-05, "epoch": 0.13135635, "global_step/max_steps": "330/2513", "percentage": "13.13%", "elapsed_time": "54m 37s", "remaining_time": "6h 1m 21s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100683} +{"loss": 0.44357675, "grad_norm": 0.04300262, "learning_rate": 9.732e-05, "epoch": 0.1317544, "global_step/max_steps": "331/2513", "percentage": "13.17%", "elapsed_time": "54m 47s", "remaining_time": "6h 1m 11s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100687} +{"loss": 0.47668165, "grad_norm": 0.04216665, "learning_rate": 9.73e-05, "epoch": 0.13215245, "global_step/max_steps": "332/2513", "percentage": "13.21%", "elapsed_time": "54m 57s", "remaining_time": "6h 1m 0s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100691} +{"loss": 0.39411467, "grad_norm": 0.03812058, "learning_rate": 9.728e-05, "epoch": 0.1325505, "global_step/max_steps": "333/2513", "percentage": "13.25%", "elapsed_time": "55m 7s", "remaining_time": "6h 0m 50s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100689} +{"loss": 0.46367088, "grad_norm": 0.04117224, "learning_rate": 9.726e-05, "epoch": 0.13294855, "global_step/max_steps": "334/2513", "percentage": "13.29%", "elapsed_time": "55m 17s", "remaining_time": "6h 0m 44s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100674} +{"loss": 0.38619298, "grad_norm": 0.03664182, "learning_rate": 9.724e-05, "epoch": 0.1333466, "global_step/max_steps": "335/2513", "percentage": "13.33%", "elapsed_time": "55m 28s", "remaining_time": "6h 0m 39s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10065} +{"loss": 0.34580255, "grad_norm": 0.03416269, "learning_rate": 9.722e-05, "epoch": 0.13374465, "global_step/max_steps": "336/2513", "percentage": "13.37%", "elapsed_time": "55m 38s", "remaining_time": "6h 0m 27s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100659} +{"loss": 0.38384396, "grad_norm": 0.03885616, "learning_rate": 9.72e-05, "epoch": 0.1341427, "global_step/max_steps": "337/2513", "percentage": "13.41%", "elapsed_time": "55m 48s", "remaining_time": "6h 0m 19s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100649} +{"loss": 0.41347879, "grad_norm": 0.04123837, "learning_rate": 9.718e-05, "epoch": 0.13454075, "global_step/max_steps": "338/2513", "percentage": "13.45%", "elapsed_time": "55m 57s", "remaining_time": "6h 0m 6s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100662} +{"loss": 0.41235802, "grad_norm": 0.03998745, "learning_rate": 9.715e-05, "epoch": 0.1349388, "global_step/max_steps": "339/2513", "percentage": "13.49%", "elapsed_time": "56m 7s", "remaining_time": "5h 59m 54s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100676} +{"loss": 0.36377186, "grad_norm": 0.03729174, "learning_rate": 9.713e-05, "epoch": 0.13533685, "global_step/max_steps": "340/2513", "percentage": "13.53%", "elapsed_time": "56m 17s", "remaining_time": "5h 59m 43s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100681} +{"loss": 0.39249617, "grad_norm": 0.03820098, "learning_rate": 9.711e-05, "epoch": 0.1357349, "global_step/max_steps": "341/2513", "percentage": "13.57%", "elapsed_time": "56m 26s", "remaining_time": "5h 59m 32s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100685} +{"loss": 0.37631404, "grad_norm": 0.04064572, "learning_rate": 9.709e-05, "epoch": 0.13613295, "global_step/max_steps": "342/2513", "percentage": "13.61%", "elapsed_time": "56m 36s", "remaining_time": "5h 59m 21s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100689} +{"loss": 0.40443963, "grad_norm": 0.03852529, "learning_rate": 9.707e-05, "epoch": 0.136531, "global_step/max_steps": "343/2513", "percentage": "13.65%", "elapsed_time": "56m 45s", "remaining_time": "5h 59m 8s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100705} +{"loss": 0.3854112, "grad_norm": 0.0386616, "learning_rate": 9.705e-05, "epoch": 0.13692905, "global_step/max_steps": "344/2513", "percentage": "13.69%", "elapsed_time": "56m 55s", "remaining_time": "5h 58m 56s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100712} +{"loss": 0.42870551, "grad_norm": 0.04411174, "learning_rate": 9.702e-05, "epoch": 0.1373271, "global_step/max_steps": "345/2513", "percentage": "13.73%", "elapsed_time": "57m 5s", "remaining_time": "5h 58m 44s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100721} +{"loss": 0.40259701, "grad_norm": 0.03737822, "learning_rate": 9.7e-05, "epoch": 0.13772515, "global_step/max_steps": "346/2513", "percentage": "13.77%", "elapsed_time": "57m 15s", "remaining_time": "5h 58m 34s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100722} +{"loss": 0.38169205, "grad_norm": 0.03968716, "learning_rate": 9.698e-05, "epoch": 0.1381232, "global_step/max_steps": "347/2513", "percentage": "13.81%", "elapsed_time": "57m 24s", "remaining_time": "5h 58m 22s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100733} +{"loss": 0.4149861, "grad_norm": 0.04068502, "learning_rate": 9.696e-05, "epoch": 0.13852125, "global_step/max_steps": "348/2513", "percentage": "13.85%", "elapsed_time": "57m 34s", "remaining_time": "5h 58m 9s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100747} +{"loss": 0.44331664, "grad_norm": 0.04177055, "learning_rate": 9.694e-05, "epoch": 0.1389193, "global_step/max_steps": "349/2513", "percentage": "13.89%", "elapsed_time": "57m 45s", "remaining_time": "5h 58m 5s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100719} +{"loss": 0.41006905, "grad_norm": 0.03967073, "learning_rate": 9.691e-05, "epoch": 0.13931735, "global_step/max_steps": "350/2513", "percentage": "13.93%", "elapsed_time": "57m 55s", "remaining_time": "5h 58m 1s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10069} +{"loss": 0.40971041, "grad_norm": 0.04059121, "learning_rate": 9.689e-05, "epoch": 0.13971539, "global_step/max_steps": "351/2513", "percentage": "13.97%", "elapsed_time": "58m 6s", "remaining_time": "5h 57m 55s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100673} +{"loss": 0.37801492, "grad_norm": 0.04236503, "learning_rate": 9.687e-05, "epoch": 0.14011344, "global_step/max_steps": "352/2513", "percentage": "14.01%", "elapsed_time": "58m 16s", "remaining_time": "5h 57m 45s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100672} +{"loss": 0.3946231, "grad_norm": 0.04267127, "learning_rate": 9.685e-05, "epoch": 0.14051149, "global_step/max_steps": "353/2513", "percentage": "14.05%", "elapsed_time": "58m 25s", "remaining_time": "5h 57m 30s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100696} +{"loss": 0.41969782, "grad_norm": 0.04377445, "learning_rate": 9.682e-05, "epoch": 0.14090954, "global_step/max_steps": "354/2513", "percentage": "14.09%", "elapsed_time": "58m 35s", "remaining_time": "5h 57m 17s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100709} +{"loss": 0.40527058, "grad_norm": 0.03753189, "learning_rate": 9.68e-05, "epoch": 0.14130759, "global_step/max_steps": "355/2513", "percentage": "14.13%", "elapsed_time": "58m 44s", "remaining_time": "5h 57m 4s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100726} +{"loss": 0.38835439, "grad_norm": 0.04048691, "learning_rate": 9.678e-05, "epoch": 0.14170564, "global_step/max_steps": "356/2513", "percentage": "14.17%", "elapsed_time": "58m 53s", "remaining_time": "5h 56m 52s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100737} +{"loss": 0.42066181, "grad_norm": 0.04416409, "learning_rate": 9.676e-05, "epoch": 0.14210369, "global_step/max_steps": "357/2513", "percentage": "14.21%", "elapsed_time": "59m 3s", "remaining_time": "5h 56m 41s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10074} +{"loss": 0.39708108, "grad_norm": 0.04227872, "learning_rate": 9.673e-05, "epoch": 0.14250174, "global_step/max_steps": "358/2513", "percentage": "14.25%", "elapsed_time": "59m 13s", "remaining_time": "5h 56m 29s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100752} +{"loss": 0.41697669, "grad_norm": 0.04237761, "learning_rate": 9.671e-05, "epoch": 0.14289979, "global_step/max_steps": "359/2513", "percentage": "14.29%", "elapsed_time": "59m 22s", "remaining_time": "5h 56m 16s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100765} +{"loss": 0.4037194, "grad_norm": 0.04684227, "learning_rate": 9.669e-05, "epoch": 0.14329784, "global_step/max_steps": "360/2513", "percentage": "14.33%", "elapsed_time": "59m 32s", "remaining_time": "5h 56m 2s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100782} +{"loss": 0.40961647, "grad_norm": 0.04208418, "learning_rate": 9.666e-05, "epoch": 0.14369589, "global_step/max_steps": "361/2513", "percentage": "14.37%", "elapsed_time": "59m 41s", "remaining_time": "5h 55m 52s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100784} +{"loss": 0.35609904, "grad_norm": 0.03825622, "learning_rate": 9.664e-05, "epoch": 0.14409394, "global_step/max_steps": "362/2513", "percentage": "14.41%", "elapsed_time": "59m 51s", "remaining_time": "5h 55m 40s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100795} +{"loss": 0.44964477, "grad_norm": 0.03913233, "learning_rate": 9.662e-05, "epoch": 0.14449199, "global_step/max_steps": "363/2513", "percentage": "14.44%", "elapsed_time": "1h 0m 1s", "remaining_time": "5h 55m 30s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100796} +{"loss": 0.42290819, "grad_norm": 0.04063351, "learning_rate": 9.659e-05, "epoch": 0.14489004, "global_step/max_steps": "364/2513", "percentage": "14.48%", "elapsed_time": "1h 0m 11s", "remaining_time": "5h 55m 19s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100801} +{"loss": 0.35195896, "grad_norm": 0.03388202, "learning_rate": 9.657e-05, "epoch": 0.14528809, "global_step/max_steps": "365/2513", "percentage": "14.52%", "elapsed_time": "1h 0m 21s", "remaining_time": "5h 55m 10s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100793} +{"loss": 0.4291034, "grad_norm": 0.04369249, "learning_rate": 9.655e-05, "epoch": 0.14568614, "global_step/max_steps": "366/2513", "percentage": "14.56%", "elapsed_time": "1h 0m 31s", "remaining_time": "5h 55m 3s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100783} +{"loss": 0.39379632, "grad_norm": 0.04124169, "learning_rate": 9.652e-05, "epoch": 0.14608419, "global_step/max_steps": "367/2513", "percentage": "14.60%", "elapsed_time": "1h 0m 41s", "remaining_time": "5h 54m 51s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100792} +{"loss": 0.3776283, "grad_norm": 0.03945122, "learning_rate": 9.65e-05, "epoch": 0.14648224, "global_step/max_steps": "368/2513", "percentage": "14.64%", "elapsed_time": "1h 0m 50s", "remaining_time": "5h 54m 39s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100801} +{"loss": 0.46613732, "grad_norm": 0.04100883, "learning_rate": 9.648e-05, "epoch": 0.14688029, "global_step/max_steps": "369/2513", "percentage": "14.68%", "elapsed_time": "1h 1m 0s", "remaining_time": "5h 54m 30s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100797} +{"loss": 0.44081387, "grad_norm": 0.04015291, "learning_rate": 9.645e-05, "epoch": 0.14727834, "global_step/max_steps": "370/2513", "percentage": "14.72%", "elapsed_time": "1h 1m 10s", "remaining_time": "5h 54m 20s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100798} +{"loss": 0.35208356, "grad_norm": 0.03851292, "learning_rate": 9.643e-05, "epoch": 0.14767639, "global_step/max_steps": "371/2513", "percentage": "14.76%", "elapsed_time": "1h 1m 20s", "remaining_time": "5h 54m 8s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100807} +{"loss": 0.39081532, "grad_norm": 0.03898424, "learning_rate": 9.64e-05, "epoch": 0.14807444, "global_step/max_steps": "372/2513", "percentage": "14.80%", "elapsed_time": "1h 1m 29s", "remaining_time": "5h 53m 54s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100826} +{"loss": 0.37188843, "grad_norm": 0.03719999, "learning_rate": 9.638e-05, "epoch": 0.14847248, "global_step/max_steps": "373/2513", "percentage": "14.84%", "elapsed_time": "1h 1m 38s", "remaining_time": "5h 53m 42s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100839} +{"loss": 0.38032365, "grad_norm": 0.03822905, "learning_rate": 9.636e-05, "epoch": 0.14887053, "global_step/max_steps": "374/2513", "percentage": "14.88%", "elapsed_time": "1h 1m 49s", "remaining_time": "5h 53m 33s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100833} +{"loss": 0.41489112, "grad_norm": 0.04025198, "learning_rate": 9.633e-05, "epoch": 0.14926858, "global_step/max_steps": "375/2513", "percentage": "14.92%", "elapsed_time": "1h 1m 58s", "remaining_time": "5h 53m 19s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100851} +{"loss": 0.39321521, "grad_norm": 0.04006611, "learning_rate": 9.631e-05, "epoch": 0.14966663, "global_step/max_steps": "376/2513", "percentage": "14.96%", "elapsed_time": "1h 2m 7s", "remaining_time": "5h 53m 7s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100863} +{"loss": 0.43051639, "grad_norm": 0.04543954, "learning_rate": 9.628e-05, "epoch": 0.15006468, "global_step/max_steps": "377/2513", "percentage": "15.00%", "elapsed_time": "1h 2m 17s", "remaining_time": "5h 52m 56s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100866} +{"loss": 0.37973845, "grad_norm": 0.04128501, "learning_rate": 9.626e-05, "epoch": 0.15046273, "global_step/max_steps": "378/2513", "percentage": "15.04%", "elapsed_time": "1h 2m 27s", "remaining_time": "5h 52m 48s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100859} +{"loss": 0.40130854, "grad_norm": 0.0403388, "learning_rate": 9.623e-05, "epoch": 0.15086078, "global_step/max_steps": "379/2513", "percentage": "15.08%", "elapsed_time": "1h 2m 38s", "remaining_time": "5h 52m 43s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100832} +{"loss": 0.45146304, "grad_norm": 0.04144462, "learning_rate": 9.621e-05, "epoch": 0.15125883, "global_step/max_steps": "380/2513", "percentage": "15.12%", "elapsed_time": "1h 2m 48s", "remaining_time": "5h 52m 35s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100823} +{"loss": 0.43836653, "grad_norm": 0.04119306, "learning_rate": 9.618e-05, "epoch": 0.15165688, "global_step/max_steps": "381/2513", "percentage": "15.16%", "elapsed_time": "1h 2m 59s", "remaining_time": "5h 52m 26s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10082} +{"loss": 0.42115957, "grad_norm": 0.03994948, "learning_rate": 9.616e-05, "epoch": 0.15205493, "global_step/max_steps": "382/2513", "percentage": "15.20%", "elapsed_time": "1h 3m 8s", "remaining_time": "5h 52m 15s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100824} +{"loss": 0.387079, "grad_norm": 0.03877375, "learning_rate": 9.614e-05, "epoch": 0.15245298, "global_step/max_steps": "383/2513", "percentage": "15.24%", "elapsed_time": "1h 3m 18s", "remaining_time": "5h 52m 3s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100837} +{"loss": 0.41162485, "grad_norm": 0.04017994, "learning_rate": 9.611e-05, "epoch": 0.15285103, "global_step/max_steps": "384/2513", "percentage": "15.28%", "elapsed_time": "1h 3m 28s", "remaining_time": "5h 51m 53s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100834} +{"loss": 0.41282216, "grad_norm": 0.03797473, "learning_rate": 9.609e-05, "epoch": 0.15324908, "global_step/max_steps": "385/2513", "percentage": "15.32%", "elapsed_time": "1h 3m 38s", "remaining_time": "5h 51m 48s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100814} +{"loss": 0.42371783, "grad_norm": 0.03872111, "learning_rate": 9.606e-05, "epoch": 0.15364713, "global_step/max_steps": "386/2513", "percentage": "15.36%", "elapsed_time": "1h 3m 48s", "remaining_time": "5h 51m 35s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100827} +{"loss": 0.36211634, "grad_norm": 0.0415573, "learning_rate": 9.604e-05, "epoch": 0.15404518, "global_step/max_steps": "387/2513", "percentage": "15.40%", "elapsed_time": "1h 3m 57s", "remaining_time": "5h 51m 23s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100839} +{"loss": 0.4007386, "grad_norm": 0.04230562, "learning_rate": 9.601e-05, "epoch": 0.15444323, "global_step/max_steps": "388/2513", "percentage": "15.44%", "elapsed_time": "1h 4m 7s", "remaining_time": "5h 51m 10s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100852} +{"loss": 0.40365285, "grad_norm": 0.03758115, "learning_rate": 9.598e-05, "epoch": 0.15484128, "global_step/max_steps": "389/2513", "percentage": "15.48%", "elapsed_time": "1h 4m 16s", "remaining_time": "5h 50m 58s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100861} +{"loss": 0.441055, "grad_norm": 0.03985709, "learning_rate": 9.596e-05, "epoch": 0.15523933, "global_step/max_steps": "390/2513", "percentage": "15.52%", "elapsed_time": "1h 4m 26s", "remaining_time": "5h 50m 48s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100861} +{"loss": 0.42898738, "grad_norm": 0.03956949, "learning_rate": 9.593e-05, "epoch": 0.15563738, "global_step/max_steps": "391/2513", "percentage": "15.56%", "elapsed_time": "1h 4m 36s", "remaining_time": "5h 50m 37s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100865} +{"loss": 0.38816804, "grad_norm": 0.03945186, "learning_rate": 9.591e-05, "epoch": 0.15603543, "global_step/max_steps": "392/2513", "percentage": "15.60%", "elapsed_time": "1h 4m 45s", "remaining_time": "5h 50m 25s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100877} +{"loss": 0.42077178, "grad_norm": 0.0397406, "learning_rate": 9.588e-05, "epoch": 0.15643348, "global_step/max_steps": "393/2513", "percentage": "15.64%", "elapsed_time": "1h 4m 56s", "remaining_time": "5h 50m 18s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100861} +{"loss": 0.44623476, "grad_norm": 0.04148711, "learning_rate": 9.586e-05, "epoch": 0.15683153, "global_step/max_steps": "394/2513", "percentage": "15.68%", "elapsed_time": "1h 5m 6s", "remaining_time": "5h 50m 11s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100848} +{"loss": 0.38908857, "grad_norm": 0.03649422, "learning_rate": 9.583e-05, "epoch": 0.15722958, "global_step/max_steps": "395/2513", "percentage": "15.72%", "elapsed_time": "1h 5m 17s", "remaining_time": "5h 50m 3s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100841} +{"loss": 0.40280598, "grad_norm": 0.04060103, "learning_rate": 9.581e-05, "epoch": 0.15762762, "global_step/max_steps": "396/2513", "percentage": "15.76%", "elapsed_time": "1h 5m 26s", "remaining_time": "5h 49m 52s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100845} +{"loss": 0.35843867, "grad_norm": 0.03932205, "learning_rate": 9.578e-05, "epoch": 0.15802567, "global_step/max_steps": "397/2513", "percentage": "15.80%", "elapsed_time": "1h 5m 36s", "remaining_time": "5h 49m 40s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100857} +{"loss": 0.40448287, "grad_norm": 0.0444594, "learning_rate": 9.575e-05, "epoch": 0.15842372, "global_step/max_steps": "398/2513", "percentage": "15.84%", "elapsed_time": "1h 5m 45s", "remaining_time": "5h 49m 28s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100866} +{"loss": 0.39401859, "grad_norm": 0.03553969, "learning_rate": 9.573e-05, "epoch": 0.15882177, "global_step/max_steps": "399/2513", "percentage": "15.88%", "elapsed_time": "1h 5m 55s", "remaining_time": "5h 49m 17s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100869} +{"loss": 0.44274786, "grad_norm": 0.04111161, "learning_rate": 9.57e-05, "epoch": 0.15921982, "global_step/max_steps": "400/2513", "percentage": "15.92%", "elapsed_time": "1h 6m 5s", "remaining_time": "5h 49m 8s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100868} +{"loss": 0.3624045, "grad_norm": 0.03763436, "learning_rate": 9.568e-05, "epoch": 0.15961787, "global_step/max_steps": "401/2513", "percentage": "15.96%", "elapsed_time": "1h 6m 16s", "remaining_time": "5h 49m 5s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100835} +{"loss": 0.42059588, "grad_norm": 0.03973778, "learning_rate": 9.565e-05, "epoch": 0.16001592, "global_step/max_steps": "402/2513", "percentage": "16.00%", "elapsed_time": "1h 6m 26s", "remaining_time": "5h 48m 54s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100837} +{"loss": 0.38174838, "grad_norm": 0.0370774, "learning_rate": 9.562e-05, "epoch": 0.16041397, "global_step/max_steps": "403/2513", "percentage": "16.04%", "elapsed_time": "1h 6m 36s", "remaining_time": "5h 48m 42s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100848} +{"loss": 0.4320589, "grad_norm": 0.04118223, "learning_rate": 9.56e-05, "epoch": 0.16081202, "global_step/max_steps": "404/2513", "percentage": "16.08%", "elapsed_time": "1h 6m 46s", "remaining_time": "5h 48m 33s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100846} +{"loss": 0.37695897, "grad_norm": 0.0381271, "learning_rate": 9.557e-05, "epoch": 0.16121007, "global_step/max_steps": "405/2513", "percentage": "16.12%", "elapsed_time": "1h 6m 55s", "remaining_time": "5h 48m 22s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100848} +{"loss": 0.35138699, "grad_norm": 0.03685558, "learning_rate": 9.554e-05, "epoch": 0.16160812, "global_step/max_steps": "406/2513", "percentage": "16.16%", "elapsed_time": "1h 7m 5s", "remaining_time": "5h 48m 10s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100859} +{"loss": 0.3933624, "grad_norm": 0.03753622, "learning_rate": 9.552e-05, "epoch": 0.16200617, "global_step/max_steps": "407/2513", "percentage": "16.20%", "elapsed_time": "1h 7m 15s", "remaining_time": "5h 48m 0s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100859} +{"loss": 0.39433825, "grad_norm": 0.03766764, "learning_rate": 9.549e-05, "epoch": 0.16240422, "global_step/max_steps": "408/2513", "percentage": "16.24%", "elapsed_time": "1h 7m 25s", "remaining_time": "5h 47m 53s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100846} +{"loss": 0.37000287, "grad_norm": 0.0346412, "learning_rate": 9.546e-05, "epoch": 0.16280227, "global_step/max_steps": "409/2513", "percentage": "16.28%", "elapsed_time": "1h 7m 36s", "remaining_time": "5h 47m 45s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100838} +{"loss": 0.42637259, "grad_norm": 0.04077097, "learning_rate": 9.544e-05, "epoch": 0.16320032, "global_step/max_steps": "410/2513", "percentage": "16.32%", "elapsed_time": "1h 7m 46s", "remaining_time": "5h 47m 36s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100832} +{"loss": 0.35502476, "grad_norm": 0.03991504, "learning_rate": 9.541e-05, "epoch": 0.16359837, "global_step/max_steps": "411/2513", "percentage": "16.35%", "elapsed_time": "1h 7m 56s", "remaining_time": "5h 47m 27s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100828} +{"loss": 0.40059108, "grad_norm": 0.03661646, "learning_rate": 9.538e-05, "epoch": 0.16399642, "global_step/max_steps": "412/2513", "percentage": "16.39%", "elapsed_time": "1h 8m 5s", "remaining_time": "5h 47m 15s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100837} +{"loss": 0.40296042, "grad_norm": 0.04154465, "learning_rate": 9.536e-05, "epoch": 0.16439447, "global_step/max_steps": "413/2513", "percentage": "16.43%", "elapsed_time": "1h 8m 15s", "remaining_time": "5h 47m 5s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10084} +{"loss": 0.42562741, "grad_norm": 0.03859769, "learning_rate": 9.533e-05, "epoch": 0.16479252, "global_step/max_steps": "414/2513", "percentage": "16.47%", "elapsed_time": "1h 8m 25s", "remaining_time": "5h 46m 54s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100842} +{"loss": 0.37159795, "grad_norm": 0.04059911, "learning_rate": 9.53e-05, "epoch": 0.16519057, "global_step/max_steps": "415/2513", "percentage": "16.51%", "elapsed_time": "1h 8m 35s", "remaining_time": "5h 46m 44s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100843} +{"loss": 0.39319608, "grad_norm": 0.03826781, "learning_rate": 9.527e-05, "epoch": 0.16558862, "global_step/max_steps": "416/2513", "percentage": "16.55%", "elapsed_time": "1h 8m 44s", "remaining_time": "5h 46m 32s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100852} +{"loss": 0.38508332, "grad_norm": 0.03537513, "learning_rate": 9.525e-05, "epoch": 0.16598667, "global_step/max_steps": "417/2513", "percentage": "16.59%", "elapsed_time": "1h 8m 54s", "remaining_time": "5h 46m 21s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100857} +{"loss": 0.44964257, "grad_norm": 0.04100261, "learning_rate": 9.522e-05, "epoch": 0.16638471, "global_step/max_steps": "418/2513", "percentage": "16.63%", "elapsed_time": "1h 9m 4s", "remaining_time": "5h 46m 10s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100865} +{"loss": 0.3994332, "grad_norm": 0.03698952, "learning_rate": 9.519e-05, "epoch": 0.16678276, "global_step/max_steps": "419/2513", "percentage": "16.67%", "elapsed_time": "1h 9m 14s", "remaining_time": "5h 46m 0s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100866} +{"loss": 0.4067435, "grad_norm": 0.03828394, "learning_rate": 9.516e-05, "epoch": 0.16718081, "global_step/max_steps": "420/2513", "percentage": "16.71%", "elapsed_time": "1h 9m 24s", "remaining_time": "5h 45m 51s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100861} +{"loss": 0.42447317, "grad_norm": 0.04202333, "learning_rate": 9.514e-05, "epoch": 0.16757886, "global_step/max_steps": "421/2513", "percentage": "16.75%", "elapsed_time": "1h 9m 33s", "remaining_time": "5h 45m 39s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100872} +{"loss": 0.36914086, "grad_norm": 0.04032471, "learning_rate": 9.511e-05, "epoch": 0.16797691, "global_step/max_steps": "422/2513", "percentage": "16.79%", "elapsed_time": "1h 9m 43s", "remaining_time": "5h 45m 28s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100874} +{"loss": 0.35767967, "grad_norm": 0.036829, "learning_rate": 9.508e-05, "epoch": 0.16837496, "global_step/max_steps": "423/2513", "percentage": "16.83%", "elapsed_time": "1h 9m 53s", "remaining_time": "5h 45m 20s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100865} +{"loss": 0.39381129, "grad_norm": 0.03793851, "learning_rate": 9.505e-05, "epoch": 0.16877301, "global_step/max_steps": "424/2513", "percentage": "16.87%", "elapsed_time": "1h 10m 3s", "remaining_time": "5h 45m 10s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100868} +{"loss": 0.42446527, "grad_norm": 0.03970136, "learning_rate": 9.502e-05, "epoch": 0.16917106, "global_step/max_steps": "425/2513", "percentage": "16.91%", "elapsed_time": "1h 10m 13s", "remaining_time": "5h 45m 1s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100862} +{"loss": 0.41325915, "grad_norm": 0.04201335, "learning_rate": 9.5e-05, "epoch": 0.16956911, "global_step/max_steps": "426/2513", "percentage": "16.95%", "elapsed_time": "1h 10m 23s", "remaining_time": "5h 44m 51s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100861} +{"loss": 0.40549806, "grad_norm": 0.04040714, "learning_rate": 9.497e-05, "epoch": 0.16996716, "global_step/max_steps": "427/2513", "percentage": "16.99%", "elapsed_time": "1h 10m 33s", "remaining_time": "5h 44m 40s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10087} +{"loss": 0.39220238, "grad_norm": 0.03759424, "learning_rate": 9.494e-05, "epoch": 0.17036521, "global_step/max_steps": "428/2513", "percentage": "17.03%", "elapsed_time": "1h 10m 43s", "remaining_time": "5h 44m 30s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100868} +{"loss": 0.4125315, "grad_norm": 0.03873876, "learning_rate": 9.491e-05, "epoch": 0.17076326, "global_step/max_steps": "429/2513", "percentage": "17.07%", "elapsed_time": "1h 10m 53s", "remaining_time": "5h 44m 21s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100865} +{"loss": 0.41127166, "grad_norm": 0.03731771, "learning_rate": 9.488e-05, "epoch": 0.17116131, "global_step/max_steps": "430/2513", "percentage": "17.11%", "elapsed_time": "1h 11m 2s", "remaining_time": "5h 44m 10s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100871} +{"loss": 0.40110093, "grad_norm": 0.04093385, "learning_rate": 9.485e-05, "epoch": 0.17155936, "global_step/max_steps": "431/2513", "percentage": "17.15%", "elapsed_time": "1h 11m 12s", "remaining_time": "5h 43m 57s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100882} +{"loss": 0.38587248, "grad_norm": 0.03581341, "learning_rate": 9.483e-05, "epoch": 0.17195741, "global_step/max_steps": "432/2513", "percentage": "17.19%", "elapsed_time": "1h 11m 21s", "remaining_time": "5h 43m 46s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100889} +{"loss": 0.41169593, "grad_norm": 0.03945048, "learning_rate": 9.48e-05, "epoch": 0.17235546, "global_step/max_steps": "433/2513", "percentage": "17.23%", "elapsed_time": "1h 11m 31s", "remaining_time": "5h 43m 36s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10089} +{"loss": 0.37987077, "grad_norm": 0.04387198, "learning_rate": 9.477e-05, "epoch": 0.17275351, "global_step/max_steps": "434/2513", "percentage": "17.27%", "elapsed_time": "1h 11m 41s", "remaining_time": "5h 43m 25s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100894} +{"loss": 0.3973982, "grad_norm": 0.04172361, "learning_rate": 9.474e-05, "epoch": 0.17315156, "global_step/max_steps": "435/2513", "percentage": "17.31%", "elapsed_time": "1h 11m 51s", "remaining_time": "5h 43m 15s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100895} +{"loss": 0.46508586, "grad_norm": 0.0393256, "learning_rate": 9.471e-05, "epoch": 0.17354961, "global_step/max_steps": "436/2513", "percentage": "17.35%", "elapsed_time": "1h 12m 1s", "remaining_time": "5h 43m 6s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100894} +{"loss": 0.41073555, "grad_norm": 0.04074104, "learning_rate": 9.468e-05, "epoch": 0.17394766, "global_step/max_steps": "437/2513", "percentage": "17.39%", "elapsed_time": "1h 12m 11s", "remaining_time": "5h 42m 56s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100891} +{"loss": 0.39878404, "grad_norm": 0.03794933, "learning_rate": 9.465e-05, "epoch": 0.17434571, "global_step/max_steps": "438/2513", "percentage": "17.43%", "elapsed_time": "1h 12m 22s", "remaining_time": "5h 42m 50s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100873} +{"loss": 0.39110249, "grad_norm": 0.03904588, "learning_rate": 9.462e-05, "epoch": 0.17474376, "global_step/max_steps": "439/2513", "percentage": "17.47%", "elapsed_time": "1h 12m 32s", "remaining_time": "5h 42m 42s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100865} +{"loss": 0.38752234, "grad_norm": 0.03783071, "learning_rate": 9.46e-05, "epoch": 0.17514181, "global_step/max_steps": "440/2513", "percentage": "17.51%", "elapsed_time": "1h 12m 42s", "remaining_time": "5h 42m 32s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100865} +{"loss": 0.41689059, "grad_norm": 0.04476051, "learning_rate": 9.457e-05, "epoch": 0.17553985, "global_step/max_steps": "441/2513", "percentage": "17.55%", "elapsed_time": "1h 12m 51s", "remaining_time": "5h 42m 20s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100873} +{"loss": 0.39316142, "grad_norm": 0.04061846, "learning_rate": 9.454e-05, "epoch": 0.1759379, "global_step/max_steps": "442/2513", "percentage": "17.59%", "elapsed_time": "1h 13m 1s", "remaining_time": "5h 42m 10s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100876} +{"loss": 0.41543996, "grad_norm": 0.0392601, "learning_rate": 9.451e-05, "epoch": 0.17633595, "global_step/max_steps": "443/2513", "percentage": "17.63%", "elapsed_time": "1h 13m 11s", "remaining_time": "5h 42m 1s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100871} +{"loss": 0.4274646, "grad_norm": 0.04104031, "learning_rate": 9.448e-05, "epoch": 0.176734, "global_step/max_steps": "444/2513", "percentage": "17.67%", "elapsed_time": "1h 13m 21s", "remaining_time": "5h 41m 49s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100881} +{"loss": 0.38442725, "grad_norm": 0.04008734, "learning_rate": 9.445e-05, "epoch": 0.17713205, "global_step/max_steps": "445/2513", "percentage": "17.71%", "elapsed_time": "1h 13m 30s", "remaining_time": "5h 41m 37s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10089} +{"loss": 0.41134369, "grad_norm": 0.04208915, "learning_rate": 9.442e-05, "epoch": 0.1775301, "global_step/max_steps": "446/2513", "percentage": "17.75%", "elapsed_time": "1h 13m 40s", "remaining_time": "5h 41m 26s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100896} +{"loss": 0.38136849, "grad_norm": 0.03787953, "learning_rate": 9.439e-05, "epoch": 0.17792815, "global_step/max_steps": "447/2513", "percentage": "17.79%", "elapsed_time": "1h 13m 50s", "remaining_time": "5h 41m 15s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100902} +{"loss": 0.36658823, "grad_norm": 0.03798918, "learning_rate": 9.436e-05, "epoch": 0.1783262, "global_step/max_steps": "448/2513", "percentage": "17.83%", "elapsed_time": "1h 14m 0s", "remaining_time": "5h 41m 6s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100897} +{"loss": 0.43690854, "grad_norm": 0.0401314, "learning_rate": 9.433e-05, "epoch": 0.17872425, "global_step/max_steps": "449/2513", "percentage": "17.87%", "elapsed_time": "1h 14m 9s", "remaining_time": "5h 40m 55s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100901} +{"loss": 0.40165126, "grad_norm": 0.03887462, "learning_rate": 9.43e-05, "epoch": 0.1791223, "global_step/max_steps": "450/2513", "percentage": "17.91%", "elapsed_time": "1h 14m 19s", "remaining_time": "5h 40m 44s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100907} +{"loss": 0.40920132, "grad_norm": 0.04034262, "learning_rate": 9.427e-05, "epoch": 0.17952035, "global_step/max_steps": "451/2513", "percentage": "17.95%", "elapsed_time": "1h 14m 29s", "remaining_time": "5h 40m 33s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100911} +{"loss": 0.41792637, "grad_norm": 0.04057187, "learning_rate": 9.424e-05, "epoch": 0.1799184, "global_step/max_steps": "452/2513", "percentage": "17.99%", "elapsed_time": "1h 14m 39s", "remaining_time": "5h 40m 24s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100907} +{"loss": 0.42316848, "grad_norm": 0.03979832, "learning_rate": 9.421e-05, "epoch": 0.18031645, "global_step/max_steps": "453/2513", "percentage": "18.03%", "elapsed_time": "1h 14m 49s", "remaining_time": "5h 40m 16s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100898} +{"loss": 0.41397715, "grad_norm": 0.0387321, "learning_rate": 9.418e-05, "epoch": 0.1807145, "global_step/max_steps": "454/2513", "percentage": "18.07%", "elapsed_time": "1h 14m 59s", "remaining_time": "5h 40m 8s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100891} +{"loss": 0.39180169, "grad_norm": 0.04097088, "learning_rate": 9.415e-05, "epoch": 0.18111255, "global_step/max_steps": "455/2513", "percentage": "18.11%", "elapsed_time": "1h 15m 9s", "remaining_time": "5h 39m 58s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100889} +{"loss": 0.43606865, "grad_norm": 0.04152085, "learning_rate": 9.412e-05, "epoch": 0.1815106, "global_step/max_steps": "456/2513", "percentage": "18.15%", "elapsed_time": "1h 15m 20s", "remaining_time": "5h 39m 53s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100867} +{"loss": 0.42962742, "grad_norm": 0.03906917, "learning_rate": 9.409e-05, "epoch": 0.18190865, "global_step/max_steps": "457/2513", "percentage": "18.19%", "elapsed_time": "1h 15m 30s", "remaining_time": "5h 39m 40s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100881} +{"loss": 0.41636765, "grad_norm": 0.03645085, "learning_rate": 9.406e-05, "epoch": 0.1823067, "global_step/max_steps": "458/2513", "percentage": "18.23%", "elapsed_time": "1h 15m 39s", "remaining_time": "5h 39m 29s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100887} +{"loss": 0.40902728, "grad_norm": 0.03743212, "learning_rate": 9.403e-05, "epoch": 0.18270475, "global_step/max_steps": "459/2513", "percentage": "18.27%", "elapsed_time": "1h 15m 49s", "remaining_time": "5h 39m 18s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100891} +{"loss": 0.41183734, "grad_norm": 0.04007659, "learning_rate": 9.4e-05, "epoch": 0.1831028, "global_step/max_steps": "460/2513", "percentage": "18.30%", "elapsed_time": "1h 15m 58s", "remaining_time": "5h 39m 6s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100902} +{"loss": 0.3716127, "grad_norm": 0.04007126, "learning_rate": 9.397e-05, "epoch": 0.18350085, "global_step/max_steps": "461/2513", "percentage": "18.34%", "elapsed_time": "1h 16m 8s", "remaining_time": "5h 38m 55s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100905} +{"loss": 0.40928483, "grad_norm": 0.04038055, "learning_rate": 9.394e-05, "epoch": 0.1838989, "global_step/max_steps": "462/2513", "percentage": "18.38%", "elapsed_time": "1h 16m 18s", "remaining_time": "5h 38m 44s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10091} +{"loss": 0.42964721, "grad_norm": 0.04065052, "learning_rate": 9.391e-05, "epoch": 0.18429694, "global_step/max_steps": "463/2513", "percentage": "18.42%", "elapsed_time": "1h 16m 27s", "remaining_time": "5h 38m 33s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100917} +{"loss": 0.41068912, "grad_norm": 0.03746908, "learning_rate": 9.387e-05, "epoch": 0.18469499, "global_step/max_steps": "464/2513", "percentage": "18.46%", "elapsed_time": "1h 16m 37s", "remaining_time": "5h 38m 23s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100917} +{"loss": 0.3829487, "grad_norm": 0.03968377, "learning_rate": 9.384e-05, "epoch": 0.18509304, "global_step/max_steps": "465/2513", "percentage": "18.50%", "elapsed_time": "1h 16m 47s", "remaining_time": "5h 38m 13s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100919} +{"loss": 0.42573422, "grad_norm": 0.0400569, "learning_rate": 9.381e-05, "epoch": 0.18549109, "global_step/max_steps": "466/2513", "percentage": "18.54%", "elapsed_time": "1h 16m 57s", "remaining_time": "5h 38m 1s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10093} +{"loss": 0.39892587, "grad_norm": 0.04000848, "learning_rate": 9.378e-05, "epoch": 0.18588914, "global_step/max_steps": "467/2513", "percentage": "18.58%", "elapsed_time": "1h 17m 7s", "remaining_time": "5h 37m 55s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10091} +{"loss": 0.43865126, "grad_norm": 0.04361434, "learning_rate": 9.375e-05, "epoch": 0.18628719, "global_step/max_steps": "468/2513", "percentage": "18.62%", "elapsed_time": "1h 17m 18s", "remaining_time": "5h 37m 47s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.1009} +{"loss": 0.39352435, "grad_norm": 0.03814779, "learning_rate": 9.372e-05, "epoch": 0.18668524, "global_step/max_steps": "469/2513", "percentage": "18.66%", "elapsed_time": "1h 17m 28s", "remaining_time": "5h 37m 38s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100895} +{"loss": 0.37169394, "grad_norm": 0.03982125, "learning_rate": 9.369e-05, "epoch": 0.18708329, "global_step/max_steps": "470/2513", "percentage": "18.70%", "elapsed_time": "1h 17m 38s", "remaining_time": "5h 37m 30s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100888} +{"loss": 0.35818386, "grad_norm": 0.0355771, "learning_rate": 9.366e-05, "epoch": 0.18748134, "global_step/max_steps": "471/2513", "percentage": "18.74%", "elapsed_time": "1h 17m 48s", "remaining_time": "5h 37m 20s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100885} +{"loss": 0.47496349, "grad_norm": 0.04149422, "learning_rate": 9.363e-05, "epoch": 0.18787939, "global_step/max_steps": "472/2513", "percentage": "18.78%", "elapsed_time": "1h 17m 58s", "remaining_time": "5h 37m 9s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100894} +{"loss": 0.44812238, "grad_norm": 0.04100587, "learning_rate": 9.359e-05, "epoch": 0.18827744, "global_step/max_steps": "473/2513", "percentage": "18.82%", "elapsed_time": "1h 18m 7s", "remaining_time": "5h 36m 58s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.1009} +{"loss": 0.41410154, "grad_norm": 0.03690934, "learning_rate": 9.356e-05, "epoch": 0.18867549, "global_step/max_steps": "474/2513", "percentage": "18.86%", "elapsed_time": "1h 18m 17s", "remaining_time": "5h 36m 46s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100907} +{"loss": 0.39779037, "grad_norm": 0.03822509, "learning_rate": 9.353e-05, "epoch": 0.18907354, "global_step/max_steps": "475/2513", "percentage": "18.90%", "elapsed_time": "1h 18m 26s", "remaining_time": "5h 36m 35s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100915} +{"loss": 0.39638102, "grad_norm": 0.03625893, "learning_rate": 9.35e-05, "epoch": 0.18947159, "global_step/max_steps": "476/2513", "percentage": "18.94%", "elapsed_time": "1h 18m 36s", "remaining_time": "5h 36m 24s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10092} +{"loss": 0.3970871, "grad_norm": 0.04061795, "learning_rate": 9.347e-05, "epoch": 0.18986964, "global_step/max_steps": "477/2513", "percentage": "18.98%", "elapsed_time": "1h 18m 46s", "remaining_time": "5h 36m 12s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100927} +{"loss": 0.3837882, "grad_norm": 0.03723517, "learning_rate": 9.343e-05, "epoch": 0.19026769, "global_step/max_steps": "478/2513", "percentage": "19.02%", "elapsed_time": "1h 18m 55s", "remaining_time": "5h 36m 2s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10093} +{"loss": 0.39741993, "grad_norm": 0.03875985, "learning_rate": 9.34e-05, "epoch": 0.19066574, "global_step/max_steps": "479/2513", "percentage": "19.06%", "elapsed_time": "1h 19m 5s", "remaining_time": "5h 35m 52s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10093} +{"loss": 0.45172897, "grad_norm": 0.04503752, "learning_rate": 9.337e-05, "epoch": 0.19106379, "global_step/max_steps": "480/2513", "percentage": "19.10%", "elapsed_time": "1h 19m 15s", "remaining_time": "5h 35m 42s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100932} +{"loss": 0.36869797, "grad_norm": 0.03737544, "learning_rate": 9.334e-05, "epoch": 0.19146184, "global_step/max_steps": "481/2513", "percentage": "19.14%", "elapsed_time": "1h 19m 25s", "remaining_time": "5h 35m 32s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10093} +{"loss": 0.39546326, "grad_norm": 0.04267163, "learning_rate": 9.331e-05, "epoch": 0.19185989, "global_step/max_steps": "482/2513", "percentage": "19.18%", "elapsed_time": "1h 19m 36s", "remaining_time": "5h 35m 24s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10092} +{"loss": 0.39608961, "grad_norm": 0.03936427, "learning_rate": 9.327e-05, "epoch": 0.19225794, "global_step/max_steps": "483/2513", "percentage": "19.22%", "elapsed_time": "1h 19m 46s", "remaining_time": "5h 35m 18s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100903} +{"loss": 0.41353303, "grad_norm": 0.03846092, "learning_rate": 9.324e-05, "epoch": 0.19265599, "global_step/max_steps": "484/2513", "percentage": "19.26%", "elapsed_time": "1h 19m 57s", "remaining_time": "5h 35m 10s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100894} +{"loss": 0.37784418, "grad_norm": 0.03765392, "learning_rate": 9.321e-05, "epoch": 0.19305404, "global_step/max_steps": "485/2513", "percentage": "19.30%", "elapsed_time": "1h 20m 7s", "remaining_time": "5h 35m 0s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100892} +{"loss": 0.3790015, "grad_norm": 0.0385027, "learning_rate": 9.318e-05, "epoch": 0.19345208, "global_step/max_steps": "486/2513", "percentage": "19.34%", "elapsed_time": "1h 20m 16s", "remaining_time": "5h 34m 49s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100896} +{"loss": 0.37882614, "grad_norm": 0.03707709, "learning_rate": 9.314e-05, "epoch": 0.19385013, "global_step/max_steps": "487/2513", "percentage": "19.38%", "elapsed_time": "1h 20m 26s", "remaining_time": "5h 34m 38s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100906} +{"loss": 0.32182536, "grad_norm": 0.03648027, "learning_rate": 9.311e-05, "epoch": 0.19424818, "global_step/max_steps": "488/2513", "percentage": "19.42%", "elapsed_time": "1h 20m 35s", "remaining_time": "5h 34m 27s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10091} +{"loss": 0.45690417, "grad_norm": 0.04030609, "learning_rate": 9.308e-05, "epoch": 0.19464623, "global_step/max_steps": "489/2513", "percentage": "19.46%", "elapsed_time": "1h 20m 45s", "remaining_time": "5h 34m 16s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100916} +{"loss": 0.41689241, "grad_norm": 0.03897972, "learning_rate": 9.305e-05, "epoch": 0.19504428, "global_step/max_steps": "490/2513", "percentage": "19.50%", "elapsed_time": "1h 20m 55s", "remaining_time": "5h 34m 5s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10092} +{"loss": 0.38784996, "grad_norm": 0.03785496, "learning_rate": 9.301e-05, "epoch": 0.19544233, "global_step/max_steps": "491/2513", "percentage": "19.54%", "elapsed_time": "1h 21m 4s", "remaining_time": "5h 33m 54s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100928} +{"loss": 0.42070574, "grad_norm": 0.03988143, "learning_rate": 9.298e-05, "epoch": 0.19584038, "global_step/max_steps": "492/2513", "percentage": "19.58%", "elapsed_time": "1h 21m 14s", "remaining_time": "5h 33m 43s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10093} +{"loss": 0.38060418, "grad_norm": 0.03444765, "learning_rate": 9.295e-05, "epoch": 0.19623843, "global_step/max_steps": "493/2513", "percentage": "19.62%", "elapsed_time": "1h 21m 24s", "remaining_time": "5h 33m 33s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100934} +{"loss": 0.44172695, "grad_norm": 0.04030796, "learning_rate": 9.291e-05, "epoch": 0.19663648, "global_step/max_steps": "494/2513", "percentage": "19.66%", "elapsed_time": "1h 21m 34s", "remaining_time": "5h 33m 23s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100934} +{"loss": 0.36312973, "grad_norm": 0.03473888, "learning_rate": 9.288e-05, "epoch": 0.19703453, "global_step/max_steps": "495/2513", "percentage": "19.70%", "elapsed_time": "1h 21m 43s", "remaining_time": "5h 33m 11s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100942} +{"loss": 0.38864464, "grad_norm": 0.03611589, "learning_rate": 9.285e-05, "epoch": 0.19743258, "global_step/max_steps": "496/2513", "percentage": "19.74%", "elapsed_time": "1h 21m 53s", "remaining_time": "5h 33m 1s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100942} +{"loss": 0.37512273, "grad_norm": 0.04147916, "learning_rate": 9.282e-05, "epoch": 0.19783063, "global_step/max_steps": "497/2513", "percentage": "19.78%", "elapsed_time": "1h 22m 4s", "remaining_time": "5h 32m 53s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100931} +{"loss": 0.35291207, "grad_norm": 0.03694937, "learning_rate": 9.278e-05, "epoch": 0.19822868, "global_step/max_steps": "498/2513", "percentage": "19.82%", "elapsed_time": "1h 22m 14s", "remaining_time": "5h 32m 44s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100928} +{"loss": 0.42491508, "grad_norm": 0.03978024, "learning_rate": 9.275e-05, "epoch": 0.19862673, "global_step/max_steps": "499/2513", "percentage": "19.86%", "elapsed_time": "1h 22m 24s", "remaining_time": "5h 32m 38s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100911} +{"loss": 0.3922123, "grad_norm": 0.04116228, "learning_rate": 9.272e-05, "epoch": 0.19902478, "global_step/max_steps": "500/2513", "percentage": "19.90%", "elapsed_time": "1h 22m 34s", "remaining_time": "5h 32m 28s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100912} +{"loss": 0.39566165, "grad_norm": 0.04098085, "learning_rate": 9.268e-05, "epoch": 0.19942283, "global_step/max_steps": "501/2513", "percentage": "19.94%", "elapsed_time": "1h 22m 46s", "remaining_time": "5h 32m 24s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100879} +{"loss": 0.39756042, "grad_norm": 0.03801301, "learning_rate": 9.265e-05, "epoch": 0.19982088, "global_step/max_steps": "502/2513", "percentage": "19.98%", "elapsed_time": "1h 22m 55s", "remaining_time": "5h 32m 12s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100888} +{"loss": 0.39644167, "grad_norm": 0.03925881, "learning_rate": 9.261e-05, "epoch": 0.20021893, "global_step/max_steps": "503/2513", "percentage": "20.02%", "elapsed_time": "1h 23m 5s", "remaining_time": "5h 32m 0s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100899} +{"loss": 0.44583815, "grad_norm": 0.04089521, "learning_rate": 9.258e-05, "epoch": 0.20061698, "global_step/max_steps": "504/2513", "percentage": "20.06%", "elapsed_time": "1h 23m 14s", "remaining_time": "5h 31m 50s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100901} +{"loss": 0.3832089, "grad_norm": 0.03932423, "learning_rate": 9.255e-05, "epoch": 0.20101503, "global_step/max_steps": "505/2513", "percentage": "20.10%", "elapsed_time": "1h 23m 24s", "remaining_time": "5h 31m 39s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100906} +{"loss": 0.39235264, "grad_norm": 0.03687468, "learning_rate": 9.251e-05, "epoch": 0.20141308, "global_step/max_steps": "506/2513", "percentage": "20.14%", "elapsed_time": "1h 23m 34s", "remaining_time": "5h 31m 28s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100912} +{"loss": 0.418194, "grad_norm": 0.04046179, "learning_rate": 9.248e-05, "epoch": 0.20181113, "global_step/max_steps": "507/2513", "percentage": "20.18%", "elapsed_time": "1h 23m 43s", "remaining_time": "5h 31m 17s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100919} +{"loss": 0.39845073, "grad_norm": 0.03969061, "learning_rate": 9.244e-05, "epoch": 0.20220918, "global_step/max_steps": "508/2513", "percentage": "20.21%", "elapsed_time": "1h 23m 53s", "remaining_time": "5h 31m 7s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10092} +{"loss": 0.37429163, "grad_norm": 0.03512306, "learning_rate": 9.241e-05, "epoch": 0.20260722, "global_step/max_steps": "509/2513", "percentage": "20.25%", "elapsed_time": "1h 24m 3s", "remaining_time": "5h 30m 56s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100925} +{"loss": 0.42061782, "grad_norm": 0.0444808, "learning_rate": 9.238e-05, "epoch": 0.20300527, "global_step/max_steps": "510/2513", "percentage": "20.29%", "elapsed_time": "1h 24m 12s", "remaining_time": "5h 30m 44s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100933} +{"loss": 0.40756789, "grad_norm": 0.03743252, "learning_rate": 9.234e-05, "epoch": 0.20340332, "global_step/max_steps": "511/2513", "percentage": "20.33%", "elapsed_time": "1h 24m 23s", "remaining_time": "5h 30m 36s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100924} +{"loss": 0.36916035, "grad_norm": 0.03811298, "learning_rate": 9.231e-05, "epoch": 0.20380137, "global_step/max_steps": "512/2513", "percentage": "20.37%", "elapsed_time": "1h 24m 33s", "remaining_time": "5h 30m 30s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100907} +{"loss": 0.38644856, "grad_norm": 0.0368386, "learning_rate": 9.227e-05, "epoch": 0.20419942, "global_step/max_steps": "513/2513", "percentage": "20.41%", "elapsed_time": "1h 24m 44s", "remaining_time": "5h 30m 21s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100901} +{"loss": 0.35839933, "grad_norm": 0.03550445, "learning_rate": 9.224e-05, "epoch": 0.20459747, "global_step/max_steps": "514/2513", "percentage": "20.45%", "elapsed_time": "1h 24m 54s", "remaining_time": "5h 30m 11s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100902} +{"loss": 0.41361269, "grad_norm": 0.03939651, "learning_rate": 9.22e-05, "epoch": 0.20499552, "global_step/max_steps": "515/2513", "percentage": "20.49%", "elapsed_time": "1h 25m 4s", "remaining_time": "5h 30m 1s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100901} +{"loss": 0.3962667, "grad_norm": 0.03843812, "learning_rate": 9.217e-05, "epoch": 0.20539357, "global_step/max_steps": "516/2513", "percentage": "20.53%", "elapsed_time": "1h 25m 13s", "remaining_time": "5h 29m 50s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100907} +{"loss": 0.35477391, "grad_norm": 0.03576682, "learning_rate": 9.214e-05, "epoch": 0.20579162, "global_step/max_steps": "517/2513", "percentage": "20.57%", "elapsed_time": "1h 25m 23s", "remaining_time": "5h 29m 39s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100913} +{"loss": 0.40599695, "grad_norm": 0.03784699, "learning_rate": 9.21e-05, "epoch": 0.20618967, "global_step/max_steps": "518/2513", "percentage": "20.61%", "elapsed_time": "1h 25m 33s", "remaining_time": "5h 29m 29s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100914} +{"loss": 0.4233681, "grad_norm": 0.03763746, "learning_rate": 9.207e-05, "epoch": 0.20658772, "global_step/max_steps": "519/2513", "percentage": "20.65%", "elapsed_time": "1h 25m 43s", "remaining_time": "5h 29m 19s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100913} +{"loss": 0.4129765, "grad_norm": 0.04273812, "learning_rate": 9.203e-05, "epoch": 0.20698577, "global_step/max_steps": "520/2513", "percentage": "20.69%", "elapsed_time": "1h 25m 52s", "remaining_time": "5h 29m 9s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100916} +{"loss": 0.4090094, "grad_norm": 0.03803043, "learning_rate": 9.2e-05, "epoch": 0.20738382, "global_step/max_steps": "521/2513", "percentage": "20.73%", "elapsed_time": "1h 26m 2s", "remaining_time": "5h 28m 58s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100921} +{"loss": 0.3777045, "grad_norm": 0.03523743, "learning_rate": 9.196e-05, "epoch": 0.20778187, "global_step/max_steps": "522/2513", "percentage": "20.77%", "elapsed_time": "1h 26m 12s", "remaining_time": "5h 28m 48s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100922} +{"loss": 0.42499089, "grad_norm": 0.03427975, "learning_rate": 9.193e-05, "epoch": 0.20817992, "global_step/max_steps": "523/2513", "percentage": "20.81%", "elapsed_time": "1h 26m 21s", "remaining_time": "5h 28m 36s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100928} +{"loss": 0.44479716, "grad_norm": 0.03947001, "learning_rate": 9.189e-05, "epoch": 0.20857797, "global_step/max_steps": "524/2513", "percentage": "20.85%", "elapsed_time": "1h 26m 31s", "remaining_time": "5h 28m 26s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10093} +{"loss": 0.40261805, "grad_norm": 0.03960082, "learning_rate": 9.186e-05, "epoch": 0.20897602, "global_step/max_steps": "525/2513", "percentage": "20.89%", "elapsed_time": "1h 26m 41s", "remaining_time": "5h 28m 16s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100933} +{"loss": 0.38961071, "grad_norm": 0.03807428, "learning_rate": 9.182e-05, "epoch": 0.20937407, "global_step/max_steps": "526/2513", "percentage": "20.93%", "elapsed_time": "1h 26m 52s", "remaining_time": "5h 28m 9s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100916} +{"loss": 0.36818451, "grad_norm": 0.03831663, "learning_rate": 9.178e-05, "epoch": 0.20977212, "global_step/max_steps": "527/2513", "percentage": "20.97%", "elapsed_time": "1h 27m 3s", "remaining_time": "5h 28m 3s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100899} +{"loss": 0.42621905, "grad_norm": 0.03953371, "learning_rate": 9.175e-05, "epoch": 0.21017017, "global_step/max_steps": "528/2513", "percentage": "21.01%", "elapsed_time": "1h 27m 13s", "remaining_time": "5h 27m 53s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100897} +{"loss": 0.41625607, "grad_norm": 0.0370616, "learning_rate": 9.171e-05, "epoch": 0.21056822, "global_step/max_steps": "529/2513", "percentage": "21.05%", "elapsed_time": "1h 27m 22s", "remaining_time": "5h 27m 42s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100902} +{"loss": 0.39323205, "grad_norm": 0.03508428, "learning_rate": 9.168e-05, "epoch": 0.21096627, "global_step/max_steps": "530/2513", "percentage": "21.09%", "elapsed_time": "1h 27m 31s", "remaining_time": "5h 27m 30s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100914} +{"loss": 0.43993542, "grad_norm": 0.040819, "learning_rate": 9.164e-05, "epoch": 0.21136431, "global_step/max_steps": "531/2513", "percentage": "21.13%", "elapsed_time": "1h 27m 41s", "remaining_time": "5h 27m 19s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100921} +{"loss": 0.38119888, "grad_norm": 0.04233171, "learning_rate": 9.161e-05, "epoch": 0.21176236, "global_step/max_steps": "532/2513", "percentage": "21.17%", "elapsed_time": "1h 27m 51s", "remaining_time": "5h 27m 8s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100924} +{"loss": 0.41985881, "grad_norm": 0.03821015, "learning_rate": 9.157e-05, "epoch": 0.21216041, "global_step/max_steps": "533/2513", "percentage": "21.21%", "elapsed_time": "1h 28m 0s", "remaining_time": "5h 26m 57s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100933} +{"loss": 0.3473092, "grad_norm": 0.03504725, "learning_rate": 9.154e-05, "epoch": 0.21255846, "global_step/max_steps": "534/2513", "percentage": "21.25%", "elapsed_time": "1h 28m 10s", "remaining_time": "5h 26m 47s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100932} +{"loss": 0.35103363, "grad_norm": 0.034596, "learning_rate": 9.15e-05, "epoch": 0.21295651, "global_step/max_steps": "535/2513", "percentage": "21.29%", "elapsed_time": "1h 28m 20s", "remaining_time": "5h 26m 37s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100933} +{"loss": 0.39056081, "grad_norm": 0.03639331, "learning_rate": 9.146e-05, "epoch": 0.21335456, "global_step/max_steps": "536/2513", "percentage": "21.33%", "elapsed_time": "1h 28m 29s", "remaining_time": "5h 26m 24s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100949} +{"loss": 0.39027387, "grad_norm": 0.03797147, "learning_rate": 9.143e-05, "epoch": 0.21375261, "global_step/max_steps": "537/2513", "percentage": "21.37%", "elapsed_time": "1h 28m 39s", "remaining_time": "5h 26m 13s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100952} +{"loss": 0.39158937, "grad_norm": 0.0381827, "learning_rate": 9.139e-05, "epoch": 0.21415066, "global_step/max_steps": "538/2513", "percentage": "21.41%", "elapsed_time": "1h 28m 49s", "remaining_time": "5h 26m 3s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100953} +{"loss": 0.38021201, "grad_norm": 0.03883036, "learning_rate": 9.136e-05, "epoch": 0.21454871, "global_step/max_steps": "539/2513", "percentage": "21.45%", "elapsed_time": "1h 28m 58s", "remaining_time": "5h 25m 52s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100961} +{"loss": 0.42220891, "grad_norm": 0.04075331, "learning_rate": 9.132e-05, "epoch": 0.21494676, "global_step/max_steps": "540/2513", "percentage": "21.49%", "elapsed_time": "1h 29m 8s", "remaining_time": "5h 25m 40s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10097} +{"loss": 0.40389717, "grad_norm": 0.03957849, "learning_rate": 9.128e-05, "epoch": 0.21534481, "global_step/max_steps": "541/2513", "percentage": "21.53%", "elapsed_time": "1h 29m 18s", "remaining_time": "5h 25m 33s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100953} +{"loss": 0.42625231, "grad_norm": 0.04314502, "learning_rate": 9.125e-05, "epoch": 0.21574286, "global_step/max_steps": "542/2513", "percentage": "21.57%", "elapsed_time": "1h 29m 29s", "remaining_time": "5h 25m 25s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100947} +{"loss": 0.42303833, "grad_norm": 0.0387302, "learning_rate": 9.121e-05, "epoch": 0.21614091, "global_step/max_steps": "543/2513", "percentage": "21.61%", "elapsed_time": "1h 29m 38s", "remaining_time": "5h 25m 14s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100949} +{"loss": 0.41239572, "grad_norm": 0.04393616, "learning_rate": 9.117e-05, "epoch": 0.21653896, "global_step/max_steps": "544/2513", "percentage": "21.65%", "elapsed_time": "1h 29m 48s", "remaining_time": "5h 25m 4s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100952} +{"loss": 0.36517668, "grad_norm": 0.03820521, "learning_rate": 9.114e-05, "epoch": 0.21693701, "global_step/max_steps": "545/2513", "percentage": "21.69%", "elapsed_time": "1h 29m 58s", "remaining_time": "5h 24m 53s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100957} +{"loss": 0.36060643, "grad_norm": 0.03855309, "learning_rate": 9.11e-05, "epoch": 0.21733506, "global_step/max_steps": "546/2513", "percentage": "21.73%", "elapsed_time": "1h 30m 8s", "remaining_time": "5h 24m 43s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100955} +{"loss": 0.43454906, "grad_norm": 0.04500091, "learning_rate": 9.106e-05, "epoch": 0.21773311, "global_step/max_steps": "547/2513", "percentage": "21.77%", "elapsed_time": "1h 30m 17s", "remaining_time": "5h 24m 32s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100962} +{"loss": 0.39555484, "grad_norm": 0.0362787, "learning_rate": 9.103e-05, "epoch": 0.21813116, "global_step/max_steps": "548/2513", "percentage": "21.81%", "elapsed_time": "1h 30m 27s", "remaining_time": "5h 24m 21s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100971} +{"loss": 0.40128547, "grad_norm": 0.03772022, "learning_rate": 9.099e-05, "epoch": 0.21852921, "global_step/max_steps": "549/2513", "percentage": "21.85%", "elapsed_time": "1h 30m 36s", "remaining_time": "5h 24m 9s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10098} +{"loss": 0.39255732, "grad_norm": 0.03752493, "learning_rate": 9.095e-05, "epoch": 0.21892726, "global_step/max_steps": "550/2513", "percentage": "21.89%", "elapsed_time": "1h 30m 46s", "remaining_time": "5h 23m 59s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100982} +{"loss": 0.36319792, "grad_norm": 0.03709894, "learning_rate": 9.092e-05, "epoch": 0.21932531, "global_step/max_steps": "551/2513", "percentage": "21.93%", "elapsed_time": "1h 30m 56s", "remaining_time": "5h 23m 48s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100988} +{"loss": 0.35345137, "grad_norm": 0.03547273, "learning_rate": 9.088e-05, "epoch": 0.21972336, "global_step/max_steps": "552/2513", "percentage": "21.97%", "elapsed_time": "1h 31m 5s", "remaining_time": "5h 23m 37s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100989} +{"loss": 0.3926402, "grad_norm": 0.03779873, "learning_rate": 9.084e-05, "epoch": 0.22012141, "global_step/max_steps": "553/2513", "percentage": "22.01%", "elapsed_time": "1h 31m 15s", "remaining_time": "5h 23m 27s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10099} +{"loss": 0.3902384, "grad_norm": 0.03866392, "learning_rate": 9.08e-05, "epoch": 0.22051945, "global_step/max_steps": "554/2513", "percentage": "22.05%", "elapsed_time": "1h 31m 25s", "remaining_time": "5h 23m 16s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100998} +{"loss": 0.37022391, "grad_norm": 0.0359829, "learning_rate": 9.077e-05, "epoch": 0.2209175, "global_step/max_steps": "555/2513", "percentage": "22.09%", "elapsed_time": "1h 31m 35s", "remaining_time": "5h 23m 6s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100997} +{"loss": 0.43384758, "grad_norm": 0.03781654, "learning_rate": 9.073e-05, "epoch": 0.22131555, "global_step/max_steps": "556/2513", "percentage": "22.12%", "elapsed_time": "1h 31m 45s", "remaining_time": "5h 22m 59s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100982} +{"loss": 0.35919768, "grad_norm": 0.0359275, "learning_rate": 9.069e-05, "epoch": 0.2217136, "global_step/max_steps": "557/2513", "percentage": "22.16%", "elapsed_time": "1h 31m 56s", "remaining_time": "5h 22m 50s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100978} +{"loss": 0.36309657, "grad_norm": 0.03726579, "learning_rate": 9.065e-05, "epoch": 0.22211165, "global_step/max_steps": "558/2513", "percentage": "22.20%", "elapsed_time": "1h 32m 5s", "remaining_time": "5h 22m 40s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100979} +{"loss": 0.38933527, "grad_norm": 0.03648731, "learning_rate": 9.062e-05, "epoch": 0.2225097, "global_step/max_steps": "559/2513", "percentage": "22.24%", "elapsed_time": "1h 32m 15s", "remaining_time": "5h 22m 28s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100987} +{"loss": 0.37831461, "grad_norm": 0.03580398, "learning_rate": 9.058e-05, "epoch": 0.22290775, "global_step/max_steps": "560/2513", "percentage": "22.28%", "elapsed_time": "1h 32m 25s", "remaining_time": "5h 22m 18s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10099} +{"loss": 0.38781294, "grad_norm": 0.04111247, "learning_rate": 9.054e-05, "epoch": 0.2233058, "global_step/max_steps": "561/2513", "percentage": "22.32%", "elapsed_time": "1h 32m 34s", "remaining_time": "5h 22m 7s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100997} +{"loss": 0.42957723, "grad_norm": 0.03731127, "learning_rate": 9.05e-05, "epoch": 0.22370385, "global_step/max_steps": "562/2513", "percentage": "22.36%", "elapsed_time": "1h 32m 44s", "remaining_time": "5h 21m 56s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101} +{"loss": 0.39483675, "grad_norm": 0.03602529, "learning_rate": 9.047e-05, "epoch": 0.2241019, "global_step/max_steps": "563/2513", "percentage": "22.40%", "elapsed_time": "1h 32m 54s", "remaining_time": "5h 21m 46s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101002} +{"loss": 0.38599169, "grad_norm": 0.04004959, "learning_rate": 9.043e-05, "epoch": 0.22449995, "global_step/max_steps": "564/2513", "percentage": "22.44%", "elapsed_time": "1h 33m 3s", "remaining_time": "5h 21m 35s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101007} +{"loss": 0.35610121, "grad_norm": 0.03546475, "learning_rate": 9.039e-05, "epoch": 0.224898, "global_step/max_steps": "565/2513", "percentage": "22.48%", "elapsed_time": "1h 33m 13s", "remaining_time": "5h 21m 24s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101011} +{"loss": 0.39040476, "grad_norm": 0.03779387, "learning_rate": 9.035e-05, "epoch": 0.22529605, "global_step/max_steps": "566/2513", "percentage": "22.52%", "elapsed_time": "1h 33m 23s", "remaining_time": "5h 21m 14s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101015} +{"loss": 0.45170942, "grad_norm": 0.04277943, "learning_rate": 9.031e-05, "epoch": 0.2256941, "global_step/max_steps": "567/2513", "percentage": "22.56%", "elapsed_time": "1h 33m 32s", "remaining_time": "5h 21m 2s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101025} +{"loss": 0.40703094, "grad_norm": 0.04190017, "learning_rate": 9.028e-05, "epoch": 0.22609215, "global_step/max_steps": "568/2513", "percentage": "22.60%", "elapsed_time": "1h 33m 42s", "remaining_time": "5h 20m 52s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101026} +{"loss": 0.44918862, "grad_norm": 0.04270778, "learning_rate": 9.024e-05, "epoch": 0.2264902, "global_step/max_steps": "569/2513", "percentage": "22.64%", "elapsed_time": "1h 33m 52s", "remaining_time": "5h 20m 41s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10103} +{"loss": 0.34330329, "grad_norm": 0.03679695, "learning_rate": 9.02e-05, "epoch": 0.22688825, "global_step/max_steps": "570/2513", "percentage": "22.68%", "elapsed_time": "1h 34m 2s", "remaining_time": "5h 20m 32s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101026} +{"loss": 0.42687726, "grad_norm": 0.03855987, "learning_rate": 9.016e-05, "epoch": 0.2272863, "global_step/max_steps": "571/2513", "percentage": "22.72%", "elapsed_time": "1h 34m 13s", "remaining_time": "5h 20m 26s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101007} +{"loss": 0.40833154, "grad_norm": 0.03842543, "learning_rate": 9.012e-05, "epoch": 0.22768435, "global_step/max_steps": "572/2513", "percentage": "22.76%", "elapsed_time": "1h 34m 23s", "remaining_time": "5h 20m 16s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101006} +{"loss": 0.39338559, "grad_norm": 0.03960088, "learning_rate": 9.008e-05, "epoch": 0.2280824, "global_step/max_steps": "573/2513", "percentage": "22.80%", "elapsed_time": "1h 34m 33s", "remaining_time": "5h 20m 7s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101004} +{"loss": 0.39086765, "grad_norm": 0.0362416, "learning_rate": 9.005e-05, "epoch": 0.22848045, "global_step/max_steps": "574/2513", "percentage": "22.84%", "elapsed_time": "1h 34m 42s", "remaining_time": "5h 19m 55s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101015} +{"loss": 0.39349979, "grad_norm": 0.03771998, "learning_rate": 9.001e-05, "epoch": 0.2288785, "global_step/max_steps": "575/2513", "percentage": "22.88%", "elapsed_time": "1h 34m 52s", "remaining_time": "5h 19m 44s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101017} +{"loss": 0.39537632, "grad_norm": 0.03509784, "learning_rate": 8.997e-05, "epoch": 0.22927654, "global_step/max_steps": "576/2513", "percentage": "22.92%", "elapsed_time": "1h 35m 1s", "remaining_time": "5h 19m 33s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101024} +{"loss": 0.40087405, "grad_norm": 0.03583907, "learning_rate": 8.993e-05, "epoch": 0.22967459, "global_step/max_steps": "577/2513", "percentage": "22.96%", "elapsed_time": "1h 35m 11s", "remaining_time": "5h 19m 23s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101026} +{"loss": 0.37616548, "grad_norm": 0.04187142, "learning_rate": 8.989e-05, "epoch": 0.23007264, "global_step/max_steps": "578/2513", "percentage": "23.00%", "elapsed_time": "1h 35m 21s", "remaining_time": "5h 19m 12s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10103} +{"loss": 0.42560536, "grad_norm": 0.04017587, "learning_rate": 8.985e-05, "epoch": 0.23047069, "global_step/max_steps": "579/2513", "percentage": "23.04%", "elapsed_time": "1h 35m 30s", "remaining_time": "5h 19m 2s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101032} +{"loss": 0.4153198, "grad_norm": 0.03964649, "learning_rate": 8.981e-05, "epoch": 0.23086874, "global_step/max_steps": "580/2513", "percentage": "23.08%", "elapsed_time": "1h 35m 40s", "remaining_time": "5h 18m 52s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101034} +{"loss": 0.37644234, "grad_norm": 0.03662604, "learning_rate": 8.977e-05, "epoch": 0.23126679, "global_step/max_steps": "581/2513", "percentage": "23.12%", "elapsed_time": "1h 35m 50s", "remaining_time": "5h 18m 42s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101034} +{"loss": 0.45945823, "grad_norm": 0.03921081, "learning_rate": 8.973e-05, "epoch": 0.23166484, "global_step/max_steps": "582/2513", "percentage": "23.16%", "elapsed_time": "1h 36m 0s", "remaining_time": "5h 18m 31s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101038} +{"loss": 0.36577487, "grad_norm": 0.03665758, "learning_rate": 8.97e-05, "epoch": 0.23206289, "global_step/max_steps": "583/2513", "percentage": "23.20%", "elapsed_time": "1h 36m 9s", "remaining_time": "5h 18m 20s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101042} +{"loss": 0.37444687, "grad_norm": 0.03549711, "learning_rate": 8.966e-05, "epoch": 0.23246094, "global_step/max_steps": "584/2513", "percentage": "23.24%", "elapsed_time": "1h 36m 20s", "remaining_time": "5h 18m 11s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101038} +{"loss": 0.36727947, "grad_norm": 0.03807237, "learning_rate": 8.962e-05, "epoch": 0.23285899, "global_step/max_steps": "585/2513", "percentage": "23.28%", "elapsed_time": "1h 36m 30s", "remaining_time": "5h 18m 2s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101036} +{"loss": 0.43532416, "grad_norm": 0.04306592, "learning_rate": 8.958e-05, "epoch": 0.23325704, "global_step/max_steps": "586/2513", "percentage": "23.32%", "elapsed_time": "1h 36m 41s", "remaining_time": "5h 17m 56s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101014} +{"loss": 0.40147498, "grad_norm": 0.03941208, "learning_rate": 8.954e-05, "epoch": 0.23365509, "global_step/max_steps": "587/2513", "percentage": "23.36%", "elapsed_time": "1h 36m 51s", "remaining_time": "5h 17m 46s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101013} +{"loss": 0.38080549, "grad_norm": 0.03781646, "learning_rate": 8.95e-05, "epoch": 0.23405314, "global_step/max_steps": "588/2513", "percentage": "23.40%", "elapsed_time": "1h 37m 0s", "remaining_time": "5h 17m 35s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101021} +{"loss": 0.36971763, "grad_norm": 0.03676564, "learning_rate": 8.946e-05, "epoch": 0.23445119, "global_step/max_steps": "589/2513", "percentage": "23.44%", "elapsed_time": "1h 37m 10s", "remaining_time": "5h 17m 24s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101027} +{"loss": 0.4530744, "grad_norm": 0.04133803, "learning_rate": 8.942e-05, "epoch": 0.23484924, "global_step/max_steps": "590/2513", "percentage": "23.48%", "elapsed_time": "1h 37m 19s", "remaining_time": "5h 17m 13s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101031} +{"loss": 0.35865319, "grad_norm": 0.03641838, "learning_rate": 8.938e-05, "epoch": 0.23524729, "global_step/max_steps": "591/2513", "percentage": "23.52%", "elapsed_time": "1h 37m 29s", "remaining_time": "5h 17m 2s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101037} +{"loss": 0.33918083, "grad_norm": 0.03123681, "learning_rate": 8.934e-05, "epoch": 0.23564534, "global_step/max_steps": "592/2513", "percentage": "23.56%", "elapsed_time": "1h 37m 39s", "remaining_time": "5h 16m 52s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101037} +{"loss": 0.41952905, "grad_norm": 0.03708918, "learning_rate": 8.93e-05, "epoch": 0.23604339, "global_step/max_steps": "593/2513", "percentage": "23.60%", "elapsed_time": "1h 37m 49s", "remaining_time": "5h 16m 42s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101037} +{"loss": 0.42962152, "grad_norm": 0.04066467, "learning_rate": 8.926e-05, "epoch": 0.23644144, "global_step/max_steps": "594/2513", "percentage": "23.64%", "elapsed_time": "1h 37m 58s", "remaining_time": "5h 16m 32s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101042} +{"loss": 0.37639561, "grad_norm": 0.0386877, "learning_rate": 8.922e-05, "epoch": 0.23683949, "global_step/max_steps": "595/2513", "percentage": "23.68%", "elapsed_time": "1h 38m 8s", "remaining_time": "5h 16m 21s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101046} +{"loss": 0.37441319, "grad_norm": 0.03781223, "learning_rate": 8.918e-05, "epoch": 0.23723754, "global_step/max_steps": "596/2513", "percentage": "23.72%", "elapsed_time": "1h 38m 18s", "remaining_time": "5h 16m 11s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101048} +{"loss": 0.38545045, "grad_norm": 0.03664033, "learning_rate": 8.914e-05, "epoch": 0.23763559, "global_step/max_steps": "597/2513", "percentage": "23.76%", "elapsed_time": "1h 38m 27s", "remaining_time": "5h 16m 0s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101052} +{"loss": 0.38350588, "grad_norm": 0.03559083, "learning_rate": 8.91e-05, "epoch": 0.23803364, "global_step/max_steps": "598/2513", "percentage": "23.80%", "elapsed_time": "1h 38m 37s", "remaining_time": "5h 15m 49s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101057} +{"loss": 0.38010955, "grad_norm": 0.03315098, "learning_rate": 8.906e-05, "epoch": 0.23843168, "global_step/max_steps": "599/2513", "percentage": "23.84%", "elapsed_time": "1h 38m 47s", "remaining_time": "5h 15m 40s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101056} +{"loss": 0.33961427, "grad_norm": 0.03633326, "learning_rate": 8.902e-05, "epoch": 0.23882973, "global_step/max_steps": "600/2513", "percentage": "23.88%", "elapsed_time": "1h 38m 57s", "remaining_time": "5h 15m 30s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101055} +{"loss": 0.40508196, "grad_norm": 0.03525442, "learning_rate": 8.898e-05, "epoch": 0.23922778, "global_step/max_steps": "601/2513", "percentage": "23.92%", "elapsed_time": "1h 39m 9s", "remaining_time": "5h 15m 28s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101012} +{"loss": 0.39710331, "grad_norm": 0.03686844, "learning_rate": 8.894e-05, "epoch": 0.23962583, "global_step/max_steps": "602/2513", "percentage": "23.96%", "elapsed_time": "1h 39m 19s", "remaining_time": "5h 15m 19s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101008} +{"loss": 0.37115729, "grad_norm": 0.03506747, "learning_rate": 8.89e-05, "epoch": 0.24002388, "global_step/max_steps": "603/2513", "percentage": "24.00%", "elapsed_time": "1h 39m 29s", "remaining_time": "5h 15m 8s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101013} +{"loss": 0.38503239, "grad_norm": 0.03669997, "learning_rate": 8.886e-05, "epoch": 0.24042193, "global_step/max_steps": "604/2513", "percentage": "24.04%", "elapsed_time": "1h 39m 39s", "remaining_time": "5h 14m 57s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101019} +{"loss": 0.4442997, "grad_norm": 0.04198588, "learning_rate": 8.882e-05, "epoch": 0.24081998, "global_step/max_steps": "605/2513", "percentage": "24.07%", "elapsed_time": "1h 39m 48s", "remaining_time": "5h 14m 45s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101028} +{"loss": 0.37965497, "grad_norm": 0.03790022, "learning_rate": 8.878e-05, "epoch": 0.24121803, "global_step/max_steps": "606/2513", "percentage": "24.11%", "elapsed_time": "1h 39m 58s", "remaining_time": "5h 14m 34s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101033} +{"loss": 0.33939552, "grad_norm": 0.03628371, "learning_rate": 8.874e-05, "epoch": 0.24161608, "global_step/max_steps": "607/2513", "percentage": "24.15%", "elapsed_time": "1h 40m 7s", "remaining_time": "5h 14m 24s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101037} +{"loss": 0.43349418, "grad_norm": 0.03983185, "learning_rate": 8.87e-05, "epoch": 0.24201413, "global_step/max_steps": "608/2513", "percentage": "24.19%", "elapsed_time": "1h 40m 17s", "remaining_time": "5h 14m 13s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10104} +{"loss": 0.37992203, "grad_norm": 0.03712036, "learning_rate": 8.865e-05, "epoch": 0.24241218, "global_step/max_steps": "609/2513", "percentage": "24.23%", "elapsed_time": "1h 40m 26s", "remaining_time": "5h 14m 2s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101046} +{"loss": 0.33312225, "grad_norm": 0.03577974, "learning_rate": 8.861e-05, "epoch": 0.24281023, "global_step/max_steps": "610/2513", "percentage": "24.27%", "elapsed_time": "1h 40m 36s", "remaining_time": "5h 13m 51s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101053} +{"loss": 0.39519745, "grad_norm": 0.03571225, "learning_rate": 8.857e-05, "epoch": 0.24320828, "global_step/max_steps": "611/2513", "percentage": "24.31%", "elapsed_time": "1h 40m 45s", "remaining_time": "5h 13m 40s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10106} +{"loss": 0.39843971, "grad_norm": 0.03864844, "learning_rate": 8.853e-05, "epoch": 0.24360633, "global_step/max_steps": "612/2513", "percentage": "24.35%", "elapsed_time": "1h 40m 55s", "remaining_time": "5h 13m 30s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101061} +{"loss": 0.40167737, "grad_norm": 0.04120313, "learning_rate": 8.849e-05, "epoch": 0.24400438, "global_step/max_steps": "613/2513", "percentage": "24.39%", "elapsed_time": "1h 41m 5s", "remaining_time": "5h 13m 21s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101056} +{"loss": 0.38644046, "grad_norm": 0.03540329, "learning_rate": 8.845e-05, "epoch": 0.24440243, "global_step/max_steps": "614/2513", "percentage": "24.43%", "elapsed_time": "1h 41m 15s", "remaining_time": "5h 13m 11s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101059} +{"loss": 0.37281376, "grad_norm": 0.03678903, "learning_rate": 8.841e-05, "epoch": 0.24480048, "global_step/max_steps": "615/2513", "percentage": "24.47%", "elapsed_time": "1h 41m 26s", "remaining_time": "5h 13m 3s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101047} +{"loss": 0.44499701, "grad_norm": 0.04377965, "learning_rate": 8.837e-05, "epoch": 0.24519853, "global_step/max_steps": "616/2513", "percentage": "24.51%", "elapsed_time": "1h 41m 36s", "remaining_time": "5h 12m 54s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101042} +{"loss": 0.41185725, "grad_norm": 0.03920707, "learning_rate": 8.833e-05, "epoch": 0.24559658, "global_step/max_steps": "617/2513", "percentage": "24.55%", "elapsed_time": "1h 41m 46s", "remaining_time": "5h 12m 44s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101043} +{"loss": 0.34447324, "grad_norm": 0.03882386, "learning_rate": 8.828e-05, "epoch": 0.24599463, "global_step/max_steps": "618/2513", "percentage": "24.59%", "elapsed_time": "1h 41m 55s", "remaining_time": "5h 12m 32s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101052} +{"loss": 0.40506345, "grad_norm": 0.03504897, "learning_rate": 8.824e-05, "epoch": 0.24639268, "global_step/max_steps": "619/2513", "percentage": "24.63%", "elapsed_time": "1h 42m 5s", "remaining_time": "5h 12m 21s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10106} +{"loss": 0.37286371, "grad_norm": 0.03910166, "learning_rate": 8.82e-05, "epoch": 0.24679073, "global_step/max_steps": "620/2513", "percentage": "24.67%", "elapsed_time": "1h 42m 14s", "remaining_time": "5h 12m 10s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101064} +{"loss": 0.41125309, "grad_norm": 0.04355651, "learning_rate": 8.816e-05, "epoch": 0.24718878, "global_step/max_steps": "621/2513", "percentage": "24.71%", "elapsed_time": "1h 42m 24s", "remaining_time": "5h 12m 0s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101067} +{"loss": 0.3808305, "grad_norm": 0.04083378, "learning_rate": 8.812e-05, "epoch": 0.24758682, "global_step/max_steps": "622/2513", "percentage": "24.75%", "elapsed_time": "1h 42m 33s", "remaining_time": "5h 11m 49s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101072} +{"loss": 0.37254614, "grad_norm": 0.03554908, "learning_rate": 8.808e-05, "epoch": 0.24798487, "global_step/max_steps": "623/2513", "percentage": "24.79%", "elapsed_time": "1h 42m 43s", "remaining_time": "5h 11m 38s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101077} +{"loss": 0.39604348, "grad_norm": 0.03806756, "learning_rate": 8.803e-05, "epoch": 0.24838292, "global_step/max_steps": "624/2513", "percentage": "24.83%", "elapsed_time": "1h 42m 53s", "remaining_time": "5h 11m 28s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10108} +{"loss": 0.40497732, "grad_norm": 0.04058397, "learning_rate": 8.799e-05, "epoch": 0.24878097, "global_step/max_steps": "625/2513", "percentage": "24.87%", "elapsed_time": "1h 43m 3s", "remaining_time": "5h 11m 18s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101081} +{"loss": 0.39086023, "grad_norm": 0.03966281, "learning_rate": 8.795e-05, "epoch": 0.24917902, "global_step/max_steps": "626/2513", "percentage": "24.91%", "elapsed_time": "1h 43m 12s", "remaining_time": "5h 11m 7s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101086} +{"loss": 0.39297181, "grad_norm": 0.03541655, "learning_rate": 8.791e-05, "epoch": 0.24957707, "global_step/max_steps": "627/2513", "percentage": "24.95%", "elapsed_time": "1h 43m 22s", "remaining_time": "5h 10m 56s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101091} +{"loss": 0.39785007, "grad_norm": 0.03735066, "learning_rate": 8.787e-05, "epoch": 0.24997512, "global_step/max_steps": "628/2513", "percentage": "24.99%", "elapsed_time": "1h 43m 32s", "remaining_time": "5h 10m 47s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101086} +{"loss": 0.36805356, "grad_norm": 0.03652359, "learning_rate": 8.782e-05, "epoch": 0.25037317, "global_step/max_steps": "629/2513", "percentage": "25.03%", "elapsed_time": "1h 43m 42s", "remaining_time": "5h 10m 37s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101086} +{"loss": 0.34212345, "grad_norm": 0.03670732, "learning_rate": 8.778e-05, "epoch": 0.25077122, "global_step/max_steps": "630/2513", "percentage": "25.07%", "elapsed_time": "1h 43m 53s", "remaining_time": "5h 10m 30s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101069} +{"loss": 0.3185553, "grad_norm": 0.03610923, "learning_rate": 8.774e-05, "epoch": 0.25116927, "global_step/max_steps": "631/2513", "percentage": "25.11%", "elapsed_time": "1h 44m 3s", "remaining_time": "5h 10m 21s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101068} +{"loss": 0.43920586, "grad_norm": 0.04579917, "learning_rate": 8.77e-05, "epoch": 0.25156732, "global_step/max_steps": "632/2513", "percentage": "25.15%", "elapsed_time": "1h 44m 13s", "remaining_time": "5h 10m 11s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101064} +{"loss": 0.37778306, "grad_norm": 0.04125069, "learning_rate": 8.765e-05, "epoch": 0.25196537, "global_step/max_steps": "633/2513", "percentage": "25.19%", "elapsed_time": "1h 44m 22s", "remaining_time": "5h 10m 0s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101071} +{"loss": 0.38616067, "grad_norm": 0.03686643, "learning_rate": 8.761e-05, "epoch": 0.25236342, "global_step/max_steps": "634/2513", "percentage": "25.23%", "elapsed_time": "1h 44m 32s", "remaining_time": "5h 9m 49s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101077} +{"loss": 0.39831376, "grad_norm": 0.03867249, "learning_rate": 8.757e-05, "epoch": 0.25276147, "global_step/max_steps": "635/2513", "percentage": "25.27%", "elapsed_time": "1h 44m 42s", "remaining_time": "5h 9m 40s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101076} +{"loss": 0.39499515, "grad_norm": 0.03965486, "learning_rate": 8.753e-05, "epoch": 0.25315952, "global_step/max_steps": "636/2513", "percentage": "25.31%", "elapsed_time": "1h 44m 52s", "remaining_time": "5h 9m 29s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101078} +{"loss": 0.35555133, "grad_norm": 0.03444407, "learning_rate": 8.748e-05, "epoch": 0.25355757, "global_step/max_steps": "637/2513", "percentage": "25.35%", "elapsed_time": "1h 45m 2s", "remaining_time": "5h 9m 19s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101079} +{"loss": 0.39645553, "grad_norm": 0.039656, "learning_rate": 8.744e-05, "epoch": 0.25395562, "global_step/max_steps": "638/2513", "percentage": "25.39%", "elapsed_time": "1h 45m 11s", "remaining_time": "5h 9m 9s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101079} +{"loss": 0.40436426, "grad_norm": 0.04249811, "learning_rate": 8.74e-05, "epoch": 0.25435367, "global_step/max_steps": "639/2513", "percentage": "25.43%", "elapsed_time": "1h 45m 21s", "remaining_time": "5h 8m 59s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101082} +{"loss": 0.43688819, "grad_norm": 0.04144375, "learning_rate": 8.736e-05, "epoch": 0.25475172, "global_step/max_steps": "640/2513", "percentage": "25.47%", "elapsed_time": "1h 45m 31s", "remaining_time": "5h 8m 49s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101082} +{"loss": 0.39327866, "grad_norm": 0.06059236, "learning_rate": 8.731e-05, "epoch": 0.25514977, "global_step/max_steps": "641/2513", "percentage": "25.51%", "elapsed_time": "1h 45m 40s", "remaining_time": "5h 8m 38s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101088} +{"loss": 0.37052983, "grad_norm": 0.03513333, "learning_rate": 8.727e-05, "epoch": 0.25554782, "global_step/max_steps": "642/2513", "percentage": "25.55%", "elapsed_time": "1h 45m 50s", "remaining_time": "5h 8m 28s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101089} +{"loss": 0.39603406, "grad_norm": 0.0384026, "learning_rate": 8.723e-05, "epoch": 0.25594587, "global_step/max_steps": "643/2513", "percentage": "25.59%", "elapsed_time": "1h 46m 1s", "remaining_time": "5h 8m 19s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101084} +{"loss": 0.4593173, "grad_norm": 0.03829339, "learning_rate": 8.718e-05, "epoch": 0.25634391, "global_step/max_steps": "644/2513", "percentage": "25.63%", "elapsed_time": "1h 46m 10s", "remaining_time": "5h 8m 8s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10109} +{"loss": 0.35913962, "grad_norm": 0.03794221, "learning_rate": 8.714e-05, "epoch": 0.25674196, "global_step/max_steps": "645/2513", "percentage": "25.67%", "elapsed_time": "1h 46m 21s", "remaining_time": "5h 8m 2s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101071} +{"loss": 0.38493675, "grad_norm": 0.03515271, "learning_rate": 8.71e-05, "epoch": 0.25714001, "global_step/max_steps": "646/2513", "percentage": "25.71%", "elapsed_time": "1h 46m 31s", "remaining_time": "5h 7m 53s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101064} +{"loss": 0.38373309, "grad_norm": 0.03647069, "learning_rate": 8.705e-05, "epoch": 0.25753806, "global_step/max_steps": "647/2513", "percentage": "25.75%", "elapsed_time": "1h 46m 41s", "remaining_time": "5h 7m 43s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101065} +{"loss": 0.37519491, "grad_norm": 0.0380325, "learning_rate": 8.701e-05, "epoch": 0.25793611, "global_step/max_steps": "648/2513", "percentage": "25.79%", "elapsed_time": "1h 46m 51s", "remaining_time": "5h 7m 32s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101068} +{"loss": 0.39679164, "grad_norm": 0.0399088, "learning_rate": 8.697e-05, "epoch": 0.25833416, "global_step/max_steps": "649/2513", "percentage": "25.83%", "elapsed_time": "1h 47m 1s", "remaining_time": "5h 7m 22s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10107} +{"loss": 0.41685632, "grad_norm": 0.03682487, "learning_rate": 8.692e-05, "epoch": 0.25873221, "global_step/max_steps": "650/2513", "percentage": "25.87%", "elapsed_time": "1h 47m 11s", "remaining_time": "5h 7m 12s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101072} +{"loss": 0.35115075, "grad_norm": 0.03625213, "learning_rate": 8.688e-05, "epoch": 0.25913026, "global_step/max_steps": "651/2513", "percentage": "25.91%", "elapsed_time": "1h 47m 20s", "remaining_time": "5h 7m 1s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101076} +{"loss": 0.39155069, "grad_norm": 0.03809184, "learning_rate": 8.684e-05, "epoch": 0.25952831, "global_step/max_steps": "652/2513", "percentage": "25.95%", "elapsed_time": "1h 47m 30s", "remaining_time": "5h 6m 52s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101072} +{"loss": 0.37945676, "grad_norm": 0.0374719, "learning_rate": 8.679e-05, "epoch": 0.25992636, "global_step/max_steps": "653/2513", "percentage": "25.98%", "elapsed_time": "1h 47m 40s", "remaining_time": "5h 6m 42s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101074} +{"loss": 0.39786661, "grad_norm": 0.04076172, "learning_rate": 8.675e-05, "epoch": 0.26032441, "global_step/max_steps": "654/2513", "percentage": "26.02%", "elapsed_time": "1h 47m 50s", "remaining_time": "5h 6m 32s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101074} +{"loss": 0.37394655, "grad_norm": 0.03720197, "learning_rate": 8.671e-05, "epoch": 0.26072246, "global_step/max_steps": "655/2513", "percentage": "26.06%", "elapsed_time": "1h 48m 0s", "remaining_time": "5h 6m 22s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101073} +{"loss": 0.3743127, "grad_norm": 0.03898171, "learning_rate": 8.666e-05, "epoch": 0.26112051, "global_step/max_steps": "656/2513", "percentage": "26.10%", "elapsed_time": "1h 48m 10s", "remaining_time": "5h 6m 12s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101076} +{"loss": 0.39312509, "grad_norm": 0.03732446, "learning_rate": 8.662e-05, "epoch": 0.26151856, "global_step/max_steps": "657/2513", "percentage": "26.14%", "elapsed_time": "1h 48m 20s", "remaining_time": "5h 6m 2s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101076} +{"loss": 0.42060244, "grad_norm": 0.0371281, "learning_rate": 8.658e-05, "epoch": 0.26191661, "global_step/max_steps": "658/2513", "percentage": "26.18%", "elapsed_time": "1h 48m 30s", "remaining_time": "5h 5m 53s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101073} +{"loss": 0.38030276, "grad_norm": 0.03870429, "learning_rate": 8.653e-05, "epoch": 0.26231466, "global_step/max_steps": "659/2513", "percentage": "26.22%", "elapsed_time": "1h 48m 41s", "remaining_time": "5h 5m 48s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101044} +{"loss": 0.37964958, "grad_norm": 0.03427542, "learning_rate": 8.649e-05, "epoch": 0.26271271, "global_step/max_steps": "660/2513", "percentage": "26.26%", "elapsed_time": "1h 48m 52s", "remaining_time": "5h 5m 39s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101037} +{"loss": 0.36614838, "grad_norm": 0.03532339, "learning_rate": 8.644e-05, "epoch": 0.26311076, "global_step/max_steps": "661/2513", "percentage": "26.30%", "elapsed_time": "1h 49m 2s", "remaining_time": "5h 5m 30s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101032} +{"loss": 0.39407814, "grad_norm": 0.0359318, "learning_rate": 8.64e-05, "epoch": 0.26350881, "global_step/max_steps": "662/2513", "percentage": "26.34%", "elapsed_time": "1h 49m 12s", "remaining_time": "5h 5m 19s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101037} +{"loss": 0.39178139, "grad_norm": 0.03841859, "learning_rate": 8.635e-05, "epoch": 0.26390686, "global_step/max_steps": "663/2513", "percentage": "26.38%", "elapsed_time": "1h 49m 21s", "remaining_time": "5h 5m 8s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101045} +{"loss": 0.41544059, "grad_norm": 0.03424106, "learning_rate": 8.631e-05, "epoch": 0.26430491, "global_step/max_steps": "664/2513", "percentage": "26.42%", "elapsed_time": "1h 49m 31s", "remaining_time": "5h 4m 58s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101048} +{"loss": 0.40125495, "grad_norm": 0.03461291, "learning_rate": 8.627e-05, "epoch": 0.26470296, "global_step/max_steps": "665/2513", "percentage": "26.46%", "elapsed_time": "1h 49m 40s", "remaining_time": "5h 4m 47s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101052} +{"loss": 0.38738841, "grad_norm": 0.03606515, "learning_rate": 8.622e-05, "epoch": 0.26510101, "global_step/max_steps": "666/2513", "percentage": "26.50%", "elapsed_time": "1h 49m 50s", "remaining_time": "5h 4m 37s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101056} +{"loss": 0.40063784, "grad_norm": 0.03849525, "learning_rate": 8.618e-05, "epoch": 0.26549905, "global_step/max_steps": "667/2513", "percentage": "26.54%", "elapsed_time": "1h 50m 0s", "remaining_time": "5h 4m 27s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101055} +{"loss": 0.40470701, "grad_norm": 0.03991099, "learning_rate": 8.613e-05, "epoch": 0.2658971, "global_step/max_steps": "668/2513", "percentage": "26.58%", "elapsed_time": "1h 50m 10s", "remaining_time": "5h 4m 17s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101057} +{"loss": 0.35702801, "grad_norm": 0.03426716, "learning_rate": 8.609e-05, "epoch": 0.26629515, "global_step/max_steps": "669/2513", "percentage": "26.62%", "elapsed_time": "1h 50m 20s", "remaining_time": "5h 4m 7s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101057} +{"loss": 0.37606984, "grad_norm": 0.03521996, "learning_rate": 8.604e-05, "epoch": 0.2666932, "global_step/max_steps": "670/2513", "percentage": "26.66%", "elapsed_time": "1h 50m 29s", "remaining_time": "5h 3m 56s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101061} +{"loss": 0.39551708, "grad_norm": 0.03771663, "learning_rate": 8.6e-05, "epoch": 0.26709125, "global_step/max_steps": "671/2513", "percentage": "26.70%", "elapsed_time": "1h 50m 39s", "remaining_time": "5h 3m 46s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101063} +{"loss": 0.36155632, "grad_norm": 0.03981155, "learning_rate": 8.595e-05, "epoch": 0.2674893, "global_step/max_steps": "672/2513", "percentage": "26.74%", "elapsed_time": "1h 50m 49s", "remaining_time": "5h 3m 36s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10106} +{"loss": 0.38449603, "grad_norm": 0.03759706, "learning_rate": 8.591e-05, "epoch": 0.26788735, "global_step/max_steps": "673/2513", "percentage": "26.78%", "elapsed_time": "1h 50m 59s", "remaining_time": "5h 3m 26s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101061} +{"loss": 0.40448612, "grad_norm": 0.03492925, "learning_rate": 8.586e-05, "epoch": 0.2682854, "global_step/max_steps": "674/2513", "percentage": "26.82%", "elapsed_time": "1h 51m 10s", "remaining_time": "5h 3m 19s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101046} +{"loss": 0.36171103, "grad_norm": 0.03672885, "learning_rate": 8.582e-05, "epoch": 0.26868345, "global_step/max_steps": "675/2513", "percentage": "26.86%", "elapsed_time": "1h 51m 20s", "remaining_time": "5h 3m 10s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101043} +{"loss": 0.39992759, "grad_norm": 0.034638, "learning_rate": 8.577e-05, "epoch": 0.2690815, "global_step/max_steps": "676/2513", "percentage": "26.90%", "elapsed_time": "1h 51m 30s", "remaining_time": "5h 3m 0s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101042} +{"loss": 0.40268034, "grad_norm": 0.03807736, "learning_rate": 8.573e-05, "epoch": 0.26947955, "global_step/max_steps": "677/2513", "percentage": "26.94%", "elapsed_time": "1h 51m 40s", "remaining_time": "5h 2m 51s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101036} +{"loss": 0.36609086, "grad_norm": 0.03440988, "learning_rate": 8.568e-05, "epoch": 0.2698776, "global_step/max_steps": "678/2513", "percentage": "26.98%", "elapsed_time": "1h 51m 50s", "remaining_time": "5h 2m 40s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101042} +{"loss": 0.40611994, "grad_norm": 0.03714269, "learning_rate": 8.564e-05, "epoch": 0.27027565, "global_step/max_steps": "679/2513", "percentage": "27.02%", "elapsed_time": "1h 52m 0s", "remaining_time": "5h 2m 31s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101037} +{"loss": 0.41878104, "grad_norm": 0.03894625, "learning_rate": 8.559e-05, "epoch": 0.2706737, "global_step/max_steps": "680/2513", "percentage": "27.06%", "elapsed_time": "1h 52m 10s", "remaining_time": "5h 2m 22s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101035} +{"loss": 0.39566824, "grad_norm": 0.03596859, "learning_rate": 8.555e-05, "epoch": 0.27107175, "global_step/max_steps": "681/2513", "percentage": "27.10%", "elapsed_time": "1h 52m 20s", "remaining_time": "5h 2m 12s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101036} +{"loss": 0.40405285, "grad_norm": 0.04209324, "learning_rate": 8.55e-05, "epoch": 0.2714698, "global_step/max_steps": "682/2513", "percentage": "27.14%", "elapsed_time": "1h 52m 29s", "remaining_time": "5h 2m 1s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101041} +{"loss": 0.38496608, "grad_norm": 0.03402385, "learning_rate": 8.546e-05, "epoch": 0.27186785, "global_step/max_steps": "683/2513", "percentage": "27.18%", "elapsed_time": "1h 52m 39s", "remaining_time": "5h 1m 50s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101044} +{"loss": 0.41825902, "grad_norm": 0.03753219, "learning_rate": 8.541e-05, "epoch": 0.2722659, "global_step/max_steps": "684/2513", "percentage": "27.22%", "elapsed_time": "1h 52m 49s", "remaining_time": "5h 1m 40s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101045} +{"loss": 0.39444262, "grad_norm": 0.03580362, "learning_rate": 8.537e-05, "epoch": 0.27266395, "global_step/max_steps": "685/2513", "percentage": "27.26%", "elapsed_time": "1h 52m 59s", "remaining_time": "5h 1m 31s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101042} +{"loss": 0.36176807, "grad_norm": 0.03716755, "learning_rate": 8.532e-05, "epoch": 0.273062, "global_step/max_steps": "686/2513", "percentage": "27.30%", "elapsed_time": "1h 53m 9s", "remaining_time": "5h 1m 21s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101043} +{"loss": 0.42424139, "grad_norm": 0.03883287, "learning_rate": 8.528e-05, "epoch": 0.27346005, "global_step/max_steps": "687/2513", "percentage": "27.34%", "elapsed_time": "1h 53m 19s", "remaining_time": "5h 1m 11s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101043} +{"loss": 0.42573267, "grad_norm": 0.03876291, "learning_rate": 8.523e-05, "epoch": 0.2738581, "global_step/max_steps": "688/2513", "percentage": "27.38%", "elapsed_time": "1h 53m 28s", "remaining_time": "5h 1m 0s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101047} +{"loss": 0.37972438, "grad_norm": 0.03618768, "learning_rate": 8.518e-05, "epoch": 0.27425614, "global_step/max_steps": "689/2513", "percentage": "27.42%", "elapsed_time": "1h 53m 39s", "remaining_time": "5h 0m 54s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10103} +{"loss": 0.41350919, "grad_norm": 0.03884387, "learning_rate": 8.514e-05, "epoch": 0.27465419, "global_step/max_steps": "690/2513", "percentage": "27.46%", "elapsed_time": "1h 53m 49s", "remaining_time": "5h 0m 44s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101025} +{"loss": 0.39474934, "grad_norm": 0.03837453, "learning_rate": 8.509e-05, "epoch": 0.27505224, "global_step/max_steps": "691/2513", "percentage": "27.50%", "elapsed_time": "1h 53m 59s", "remaining_time": "5h 0m 35s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101025} +{"loss": 0.35641754, "grad_norm": 0.03580492, "learning_rate": 8.505e-05, "epoch": 0.27545029, "global_step/max_steps": "692/2513", "percentage": "27.54%", "elapsed_time": "1h 54m 9s", "remaining_time": "5h 0m 25s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101026} +{"loss": 0.370547, "grad_norm": 0.03648506, "learning_rate": 8.5e-05, "epoch": 0.27584834, "global_step/max_steps": "693/2513", "percentage": "27.58%", "elapsed_time": "1h 54m 19s", "remaining_time": "5h 0m 15s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101024} +{"loss": 0.40173179, "grad_norm": 0.03981817, "learning_rate": 8.495e-05, "epoch": 0.27624639, "global_step/max_steps": "694/2513", "percentage": "27.62%", "elapsed_time": "1h 54m 29s", "remaining_time": "5h 0m 5s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101027} +{"loss": 0.38630959, "grad_norm": 0.03647744, "learning_rate": 8.491e-05, "epoch": 0.27664444, "global_step/max_steps": "695/2513", "percentage": "27.66%", "elapsed_time": "1h 54m 39s", "remaining_time": "4h 59m 54s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101029} +{"loss": 0.33930445, "grad_norm": 0.03579109, "learning_rate": 8.486e-05, "epoch": 0.27704249, "global_step/max_steps": "696/2513", "percentage": "27.70%", "elapsed_time": "1h 54m 49s", "remaining_time": "4h 59m 45s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101027} +{"loss": 0.41143751, "grad_norm": 0.04083904, "learning_rate": 8.482e-05, "epoch": 0.27744054, "global_step/max_steps": "697/2513", "percentage": "27.74%", "elapsed_time": "1h 54m 58s", "remaining_time": "4h 59m 34s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101032} +{"loss": 0.4096427, "grad_norm": 0.03970016, "learning_rate": 8.477e-05, "epoch": 0.27783859, "global_step/max_steps": "698/2513", "percentage": "27.78%", "elapsed_time": "1h 55m 8s", "remaining_time": "4h 59m 24s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101035} +{"loss": 0.39472672, "grad_norm": 0.0379714, "learning_rate": 8.472e-05, "epoch": 0.27823664, "global_step/max_steps": "699/2513", "percentage": "27.82%", "elapsed_time": "1h 55m 18s", "remaining_time": "4h 59m 14s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101035} +{"loss": 0.38011849, "grad_norm": 0.03774771, "learning_rate": 8.468e-05, "epoch": 0.27863469, "global_step/max_steps": "700/2513", "percentage": "27.86%", "elapsed_time": "1h 55m 28s", "remaining_time": "4h 59m 4s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101034} +{"loss": 0.39733642, "grad_norm": 0.0422117, "learning_rate": 8.463e-05, "epoch": 0.27903274, "global_step/max_steps": "701/2513", "percentage": "27.89%", "elapsed_time": "1h 55m 39s", "remaining_time": "4h 58m 58s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10101} +{"loss": 0.39891315, "grad_norm": 0.04200006, "learning_rate": 8.458e-05, "epoch": 0.27943079, "global_step/max_steps": "702/2513", "percentage": "27.93%", "elapsed_time": "1h 55m 49s", "remaining_time": "4h 58m 48s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101011} +{"loss": 0.42260283, "grad_norm": 0.04541105, "learning_rate": 8.454e-05, "epoch": 0.27982884, "global_step/max_steps": "703/2513", "percentage": "27.97%", "elapsed_time": "1h 56m 0s", "remaining_time": "4h 58m 41s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100997} +{"loss": 0.3719666, "grad_norm": 0.03682005, "learning_rate": 8.449e-05, "epoch": 0.28022689, "global_step/max_steps": "704/2513", "percentage": "28.01%", "elapsed_time": "1h 56m 10s", "remaining_time": "4h 58m 32s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100991} +{"loss": 0.41426429, "grad_norm": 0.04157876, "learning_rate": 8.444e-05, "epoch": 0.28062494, "global_step/max_steps": "705/2513", "percentage": "28.05%", "elapsed_time": "1h 56m 21s", "remaining_time": "4h 58m 23s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100986} +{"loss": 0.39552915, "grad_norm": 0.03769764, "learning_rate": 8.44e-05, "epoch": 0.28102299, "global_step/max_steps": "706/2513", "percentage": "28.09%", "elapsed_time": "1h 56m 31s", "remaining_time": "4h 58m 13s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100986} +{"loss": 0.36415681, "grad_norm": 0.03797006, "learning_rate": 8.435e-05, "epoch": 0.28142104, "global_step/max_steps": "707/2513", "percentage": "28.13%", "elapsed_time": "1h 56m 41s", "remaining_time": "4h 58m 5s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100978} +{"loss": 0.35721722, "grad_norm": 0.03883169, "learning_rate": 8.43e-05, "epoch": 0.28181909, "global_step/max_steps": "708/2513", "percentage": "28.17%", "elapsed_time": "1h 56m 51s", "remaining_time": "4h 57m 54s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100981} +{"loss": 0.34997267, "grad_norm": 0.03456209, "learning_rate": 8.426e-05, "epoch": 0.28221714, "global_step/max_steps": "709/2513", "percentage": "28.21%", "elapsed_time": "1h 57m 1s", "remaining_time": "4h 57m 45s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100979} +{"loss": 0.4360469, "grad_norm": 0.03532075, "learning_rate": 8.421e-05, "epoch": 0.28261519, "global_step/max_steps": "710/2513", "percentage": "28.25%", "elapsed_time": "1h 57m 10s", "remaining_time": "4h 57m 34s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100984} +{"loss": 0.41823775, "grad_norm": 0.03706688, "learning_rate": 8.416e-05, "epoch": 0.28301324, "global_step/max_steps": "711/2513", "percentage": "28.29%", "elapsed_time": "1h 57m 20s", "remaining_time": "4h 57m 24s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100985} +{"loss": 0.35273445, "grad_norm": 0.0366187, "learning_rate": 8.412e-05, "epoch": 0.28341128, "global_step/max_steps": "712/2513", "percentage": "28.33%", "elapsed_time": "1h 57m 30s", "remaining_time": "4h 57m 13s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100992} +{"loss": 0.40333992, "grad_norm": 0.03754918, "learning_rate": 8.407e-05, "epoch": 0.28380933, "global_step/max_steps": "713/2513", "percentage": "28.37%", "elapsed_time": "1h 57m 39s", "remaining_time": "4h 57m 3s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100992} +{"loss": 0.38567916, "grad_norm": 0.03333773, "learning_rate": 8.402e-05, "epoch": 0.28420738, "global_step/max_steps": "714/2513", "percentage": "28.41%", "elapsed_time": "1h 57m 49s", "remaining_time": "4h 56m 52s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100998} +{"loss": 0.42924565, "grad_norm": 0.03905861, "learning_rate": 8.397e-05, "epoch": 0.28460543, "global_step/max_steps": "715/2513", "percentage": "28.45%", "elapsed_time": "1h 57m 59s", "remaining_time": "4h 56m 42s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100997} +{"loss": 0.38623309, "grad_norm": 0.03676739, "learning_rate": 8.393e-05, "epoch": 0.28500348, "global_step/max_steps": "716/2513", "percentage": "28.49%", "elapsed_time": "1h 58m 9s", "remaining_time": "4h 56m 32s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100997} +{"loss": 0.4225421, "grad_norm": 0.03985614, "learning_rate": 8.388e-05, "epoch": 0.28540153, "global_step/max_steps": "717/2513", "percentage": "28.53%", "elapsed_time": "1h 58m 18s", "remaining_time": "4h 56m 22s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101001} +{"loss": 0.4101502, "grad_norm": 0.03909147, "learning_rate": 8.383e-05, "epoch": 0.28579958, "global_step/max_steps": "718/2513", "percentage": "28.57%", "elapsed_time": "1h 58m 29s", "remaining_time": "4h 56m 14s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10099} +{"loss": 0.41978562, "grad_norm": 0.03949572, "learning_rate": 8.378e-05, "epoch": 0.28619763, "global_step/max_steps": "719/2513", "percentage": "28.61%", "elapsed_time": "1h 58m 39s", "remaining_time": "4h 56m 4s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100986} +{"loss": 0.37958357, "grad_norm": 0.03499122, "learning_rate": 8.374e-05, "epoch": 0.28659568, "global_step/max_steps": "720/2513", "percentage": "28.65%", "elapsed_time": "1h 58m 49s", "remaining_time": "4h 55m 53s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100992} +{"loss": 0.31989092, "grad_norm": 0.03602173, "learning_rate": 8.369e-05, "epoch": 0.28699373, "global_step/max_steps": "721/2513", "percentage": "28.69%", "elapsed_time": "1h 58m 58s", "remaining_time": "4h 55m 43s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100997} +{"loss": 0.35879034, "grad_norm": 0.03494573, "learning_rate": 8.364e-05, "epoch": 0.28739178, "global_step/max_steps": "722/2513", "percentage": "28.73%", "elapsed_time": "1h 59m 8s", "remaining_time": "4h 55m 32s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101001} +{"loss": 0.40202129, "grad_norm": 0.04014835, "learning_rate": 8.359e-05, "epoch": 0.28778983, "global_step/max_steps": "723/2513", "percentage": "28.77%", "elapsed_time": "1h 59m 18s", "remaining_time": "4h 55m 22s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101004} +{"loss": 0.41356546, "grad_norm": 0.04008379, "learning_rate": 8.355e-05, "epoch": 0.28818788, "global_step/max_steps": "724/2513", "percentage": "28.81%", "elapsed_time": "1h 59m 27s", "remaining_time": "4h 55m 11s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101007} +{"loss": 0.41120791, "grad_norm": 0.03669295, "learning_rate": 8.35e-05, "epoch": 0.28858593, "global_step/max_steps": "725/2513", "percentage": "28.85%", "elapsed_time": "1h 59m 37s", "remaining_time": "4h 55m 1s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101007} +{"loss": 0.37856883, "grad_norm": 0.03877611, "learning_rate": 8.345e-05, "epoch": 0.28898398, "global_step/max_steps": "726/2513", "percentage": "28.89%", "elapsed_time": "1h 59m 47s", "remaining_time": "4h 54m 51s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101008} +{"loss": 0.40101224, "grad_norm": 0.03857209, "learning_rate": 8.34e-05, "epoch": 0.28938203, "global_step/max_steps": "727/2513", "percentage": "28.93%", "elapsed_time": "1h 59m 57s", "remaining_time": "4h 54m 41s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101011} +{"loss": 0.37130973, "grad_norm": 0.03994598, "learning_rate": 8.335e-05, "epoch": 0.28978008, "global_step/max_steps": "728/2513", "percentage": "28.97%", "elapsed_time": "2h 0m 6s", "remaining_time": "4h 54m 30s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101014} +{"loss": 0.40202034, "grad_norm": 0.04022467, "learning_rate": 8.331e-05, "epoch": 0.29017813, "global_step/max_steps": "729/2513", "percentage": "29.01%", "elapsed_time": "2h 0m 16s", "remaining_time": "4h 54m 20s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101018} +{"loss": 0.38623488, "grad_norm": 0.03825641, "learning_rate": 8.326e-05, "epoch": 0.29057618, "global_step/max_steps": "730/2513", "percentage": "29.05%", "elapsed_time": "2h 0m 26s", "remaining_time": "4h 54m 9s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10102} +{"loss": 0.42868155, "grad_norm": 0.03931245, "learning_rate": 8.321e-05, "epoch": 0.29097423, "global_step/max_steps": "731/2513", "percentage": "29.09%", "elapsed_time": "2h 0m 36s", "remaining_time": "4h 54m 0s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101019} +{"loss": 0.34358948, "grad_norm": 0.03545261, "learning_rate": 8.316e-05, "epoch": 0.29137228, "global_step/max_steps": "732/2513", "percentage": "29.13%", "elapsed_time": "2h 0m 45s", "remaining_time": "4h 53m 49s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101022} +{"loss": 0.40178448, "grad_norm": 0.03722585, "learning_rate": 8.311e-05, "epoch": 0.29177033, "global_step/max_steps": "733/2513", "percentage": "29.17%", "elapsed_time": "2h 0m 57s", "remaining_time": "4h 53m 43s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101004} +{"loss": 0.37925902, "grad_norm": 0.03626354, "learning_rate": 8.307e-05, "epoch": 0.29216837, "global_step/max_steps": "734/2513", "percentage": "29.21%", "elapsed_time": "2h 1m 7s", "remaining_time": "4h 53m 33s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101001} +{"loss": 0.38703293, "grad_norm": 0.04367719, "learning_rate": 8.302e-05, "epoch": 0.29256642, "global_step/max_steps": "735/2513", "percentage": "29.25%", "elapsed_time": "2h 1m 16s", "remaining_time": "4h 53m 23s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101004} +{"loss": 0.39513361, "grad_norm": 0.04119886, "learning_rate": 8.297e-05, "epoch": 0.29296447, "global_step/max_steps": "736/2513", "percentage": "29.29%", "elapsed_time": "2h 1m 26s", "remaining_time": "4h 53m 12s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101009} +{"loss": 0.40662491, "grad_norm": 0.03741539, "learning_rate": 8.292e-05, "epoch": 0.29336252, "global_step/max_steps": "737/2513", "percentage": "29.33%", "elapsed_time": "2h 1m 36s", "remaining_time": "4h 53m 3s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101006} +{"loss": 0.4023543, "grad_norm": 0.03964094, "learning_rate": 8.287e-05, "epoch": 0.29376057, "global_step/max_steps": "738/2513", "percentage": "29.37%", "elapsed_time": "2h 1m 46s", "remaining_time": "4h 52m 52s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101008} +{"loss": 0.36680454, "grad_norm": 0.03448436, "learning_rate": 8.282e-05, "epoch": 0.29415862, "global_step/max_steps": "739/2513", "percentage": "29.41%", "elapsed_time": "2h 1m 55s", "remaining_time": "4h 52m 42s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101012} +{"loss": 0.3486613, "grad_norm": 0.03742253, "learning_rate": 8.277e-05, "epoch": 0.29455667, "global_step/max_steps": "740/2513", "percentage": "29.45%", "elapsed_time": "2h 2m 5s", "remaining_time": "4h 52m 32s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101012} +{"loss": 0.37178475, "grad_norm": 0.03917089, "learning_rate": 8.273e-05, "epoch": 0.29495472, "global_step/max_steps": "741/2513", "percentage": "29.49%", "elapsed_time": "2h 2m 15s", "remaining_time": "4h 52m 22s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101012} +{"loss": 0.34186471, "grad_norm": 0.0395753, "learning_rate": 8.268e-05, "epoch": 0.29535277, "global_step/max_steps": "742/2513", "percentage": "29.53%", "elapsed_time": "2h 2m 25s", "remaining_time": "4h 52m 12s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101013} +{"loss": 0.4077208, "grad_norm": 0.03736283, "learning_rate": 8.263e-05, "epoch": 0.29575082, "global_step/max_steps": "743/2513", "percentage": "29.57%", "elapsed_time": "2h 2m 35s", "remaining_time": "4h 52m 2s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101013} +{"loss": 0.41754994, "grad_norm": 0.0389197, "learning_rate": 8.258e-05, "epoch": 0.29614887, "global_step/max_steps": "744/2513", "percentage": "29.61%", "elapsed_time": "2h 2m 45s", "remaining_time": "4h 51m 52s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101016} +{"loss": 0.40642977, "grad_norm": 0.03840404, "learning_rate": 8.253e-05, "epoch": 0.29654692, "global_step/max_steps": "745/2513", "percentage": "29.65%", "elapsed_time": "2h 2m 55s", "remaining_time": "4h 51m 42s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101012} +{"loss": 0.36818451, "grad_norm": 0.03695843, "learning_rate": 8.248e-05, "epoch": 0.29694497, "global_step/max_steps": "746/2513", "percentage": "29.69%", "elapsed_time": "2h 3m 5s", "remaining_time": "4h 51m 33s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101009} +{"loss": 0.38987881, "grad_norm": 0.03644519, "learning_rate": 8.243e-05, "epoch": 0.29734302, "global_step/max_steps": "747/2513", "percentage": "29.73%", "elapsed_time": "2h 3m 16s", "remaining_time": "4h 51m 25s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101} +{"loss": 0.3393364, "grad_norm": 0.03847351, "learning_rate": 8.238e-05, "epoch": 0.29774107, "global_step/max_steps": "748/2513", "percentage": "29.77%", "elapsed_time": "2h 3m 26s", "remaining_time": "4h 51m 16s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100991} +{"loss": 0.40684456, "grad_norm": 0.03883474, "learning_rate": 8.233e-05, "epoch": 0.29813912, "global_step/max_steps": "749/2513", "percentage": "29.81%", "elapsed_time": "2h 3m 36s", "remaining_time": "4h 51m 6s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100994} +{"loss": 0.37084144, "grad_norm": 0.0362354, "learning_rate": 8.228e-05, "epoch": 0.29853717, "global_step/max_steps": "750/2513", "percentage": "29.84%", "elapsed_time": "2h 3m 46s", "remaining_time": "4h 50m 56s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100994} +{"loss": 0.36967647, "grad_norm": 0.03688348, "learning_rate": 8.224e-05, "epoch": 0.29893522, "global_step/max_steps": "751/2513", "percentage": "29.88%", "elapsed_time": "2h 3m 55s", "remaining_time": "4h 50m 46s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100995} +{"loss": 0.39964575, "grad_norm": 0.03913074, "learning_rate": 8.219e-05, "epoch": 0.29933327, "global_step/max_steps": "752/2513", "percentage": "29.92%", "elapsed_time": "2h 4m 5s", "remaining_time": "4h 50m 36s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100997} +{"loss": 0.37190983, "grad_norm": 0.03456016, "learning_rate": 8.214e-05, "epoch": 0.29973132, "global_step/max_steps": "753/2513", "percentage": "29.96%", "elapsed_time": "2h 4m 15s", "remaining_time": "4h 50m 26s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100996} +{"loss": 0.35222709, "grad_norm": 0.03588318, "learning_rate": 8.209e-05, "epoch": 0.30012937, "global_step/max_steps": "754/2513", "percentage": "30.00%", "elapsed_time": "2h 4m 25s", "remaining_time": "4h 50m 16s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100999} +{"loss": 0.3662225, "grad_norm": 0.03820868, "learning_rate": 8.204e-05, "epoch": 0.30052742, "global_step/max_steps": "755/2513", "percentage": "30.04%", "elapsed_time": "2h 4m 35s", "remaining_time": "4h 50m 6s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100996} +{"loss": 0.39328417, "grad_norm": 0.03870692, "learning_rate": 8.199e-05, "epoch": 0.30092547, "global_step/max_steps": "756/2513", "percentage": "30.08%", "elapsed_time": "2h 4m 45s", "remaining_time": "4h 49m 55s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101} +{"loss": 0.33109969, "grad_norm": 0.03621302, "learning_rate": 8.194e-05, "epoch": 0.30132351, "global_step/max_steps": "757/2513", "percentage": "30.12%", "elapsed_time": "2h 4m 54s", "remaining_time": "4h 49m 45s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101002} +{"loss": 0.43397754, "grad_norm": 0.03801266, "learning_rate": 8.189e-05, "epoch": 0.30172156, "global_step/max_steps": "758/2513", "percentage": "30.16%", "elapsed_time": "2h 5m 4s", "remaining_time": "4h 49m 35s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101005} +{"loss": 0.40455064, "grad_norm": 0.0352982, "learning_rate": 8.184e-05, "epoch": 0.30211961, "global_step/max_steps": "759/2513", "percentage": "30.20%", "elapsed_time": "2h 5m 14s", "remaining_time": "4h 49m 24s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101008} +{"loss": 0.42623281, "grad_norm": 0.038002, "learning_rate": 8.179e-05, "epoch": 0.30251766, "global_step/max_steps": "760/2513", "percentage": "30.24%", "elapsed_time": "2h 5m 24s", "remaining_time": "4h 49m 15s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101007} +{"loss": 0.33651096, "grad_norm": 0.03281367, "learning_rate": 8.174e-05, "epoch": 0.30291571, "global_step/max_steps": "761/2513", "percentage": "30.28%", "elapsed_time": "2h 5m 34s", "remaining_time": "4h 49m 5s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101008} +{"loss": 0.44936183, "grad_norm": 0.03887233, "learning_rate": 8.169e-05, "epoch": 0.30331376, "global_step/max_steps": "762/2513", "percentage": "30.32%", "elapsed_time": "2h 5m 44s", "remaining_time": "4h 48m 57s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100998} +{"loss": 0.39224872, "grad_norm": 0.03915822, "learning_rate": 8.164e-05, "epoch": 0.30371181, "global_step/max_steps": "763/2513", "percentage": "30.36%", "elapsed_time": "2h 5m 55s", "remaining_time": "4h 48m 48s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10099} +{"loss": 0.36710587, "grad_norm": 0.03842421, "learning_rate": 8.159e-05, "epoch": 0.30410986, "global_step/max_steps": "764/2513", "percentage": "30.40%", "elapsed_time": "2h 6m 5s", "remaining_time": "4h 48m 38s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100991} +{"loss": 0.43730214, "grad_norm": 0.05072376, "learning_rate": 8.154e-05, "epoch": 0.30450791, "global_step/max_steps": "765/2513", "percentage": "30.44%", "elapsed_time": "2h 6m 14s", "remaining_time": "4h 48m 27s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100998} +{"loss": 0.40010905, "grad_norm": 0.0384521, "learning_rate": 8.149e-05, "epoch": 0.30490596, "global_step/max_steps": "766/2513", "percentage": "30.48%", "elapsed_time": "2h 6m 24s", "remaining_time": "4h 48m 17s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100999} +{"loss": 0.32358426, "grad_norm": 0.0363764, "learning_rate": 8.144e-05, "epoch": 0.30530401, "global_step/max_steps": "767/2513", "percentage": "30.52%", "elapsed_time": "2h 6m 33s", "remaining_time": "4h 48m 6s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101001} +{"loss": 0.44243458, "grad_norm": 0.03886676, "learning_rate": 8.139e-05, "epoch": 0.30570206, "global_step/max_steps": "768/2513", "percentage": "30.56%", "elapsed_time": "2h 6m 43s", "remaining_time": "4h 47m 56s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101006} +{"loss": 0.36869061, "grad_norm": 0.03560178, "learning_rate": 8.134e-05, "epoch": 0.30610011, "global_step/max_steps": "769/2513", "percentage": "30.60%", "elapsed_time": "2h 6m 53s", "remaining_time": "4h 47m 45s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101011} +{"loss": 0.3589423, "grad_norm": 0.03837978, "learning_rate": 8.129e-05, "epoch": 0.30649816, "global_step/max_steps": "770/2513", "percentage": "30.64%", "elapsed_time": "2h 7m 2s", "remaining_time": "4h 47m 35s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101013} +{"loss": 0.3978762, "grad_norm": 0.04281505, "learning_rate": 8.124e-05, "epoch": 0.30689621, "global_step/max_steps": "771/2513", "percentage": "30.68%", "elapsed_time": "2h 7m 12s", "remaining_time": "4h 47m 25s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101015} +{"loss": 0.38797921, "grad_norm": 0.03942709, "learning_rate": 8.119e-05, "epoch": 0.30729426, "global_step/max_steps": "772/2513", "percentage": "30.72%", "elapsed_time": "2h 7m 22s", "remaining_time": "4h 47m 14s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101017} +{"loss": 0.42902845, "grad_norm": 0.03898477, "learning_rate": 8.114e-05, "epoch": 0.30769231, "global_step/max_steps": "773/2513", "percentage": "30.76%", "elapsed_time": "2h 7m 32s", "remaining_time": "4h 47m 4s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101019} +{"loss": 0.35802001, "grad_norm": 0.03595875, "learning_rate": 8.109e-05, "epoch": 0.30809036, "global_step/max_steps": "774/2513", "percentage": "30.80%", "elapsed_time": "2h 7m 41s", "remaining_time": "4h 46m 54s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101021} +{"loss": 0.34291556, "grad_norm": 0.0376358, "learning_rate": 8.104e-05, "epoch": 0.30848841, "global_step/max_steps": "775/2513", "percentage": "30.84%", "elapsed_time": "2h 7m 51s", "remaining_time": "4h 46m 44s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101019} +{"loss": 0.35383379, "grad_norm": 0.03662343, "learning_rate": 8.099e-05, "epoch": 0.30888646, "global_step/max_steps": "776/2513", "percentage": "30.88%", "elapsed_time": "2h 8m 2s", "remaining_time": "4h 46m 36s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101008} +{"loss": 0.40335572, "grad_norm": 0.03932497, "learning_rate": 8.094e-05, "epoch": 0.30928451, "global_step/max_steps": "777/2513", "percentage": "30.92%", "elapsed_time": "2h 8m 13s", "remaining_time": "4h 46m 30s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100989} +{"loss": 0.40639478, "grad_norm": 0.04063162, "learning_rate": 8.089e-05, "epoch": 0.30968256, "global_step/max_steps": "778/2513", "percentage": "30.96%", "elapsed_time": "2h 8m 23s", "remaining_time": "4h 46m 19s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100991} +{"loss": 0.40820551, "grad_norm": 0.03950801, "learning_rate": 8.083e-05, "epoch": 0.31008061, "global_step/max_steps": "779/2513", "percentage": "31.00%", "elapsed_time": "2h 8m 33s", "remaining_time": "4h 46m 9s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100993} +{"loss": 0.37763506, "grad_norm": 0.03847155, "learning_rate": 8.078e-05, "epoch": 0.31047865, "global_step/max_steps": "780/2513", "percentage": "31.04%", "elapsed_time": "2h 8m 43s", "remaining_time": "4h 45m 59s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100991} +{"loss": 0.36905432, "grad_norm": 0.0384823, "learning_rate": 8.073e-05, "epoch": 0.3108767, "global_step/max_steps": "781/2513", "percentage": "31.08%", "elapsed_time": "2h 8m 53s", "remaining_time": "4h 45m 49s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100992} +{"loss": 0.38688648, "grad_norm": 0.03854913, "learning_rate": 8.068e-05, "epoch": 0.31127475, "global_step/max_steps": "782/2513", "percentage": "31.12%", "elapsed_time": "2h 9m 3s", "remaining_time": "4h 45m 40s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100992} +{"loss": 0.40389818, "grad_norm": 0.03648758, "learning_rate": 8.063e-05, "epoch": 0.3116728, "global_step/max_steps": "783/2513", "percentage": "31.16%", "elapsed_time": "2h 9m 12s", "remaining_time": "4h 45m 29s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100993} +{"loss": 0.39148971, "grad_norm": 0.03787342, "learning_rate": 8.058e-05, "epoch": 0.31207085, "global_step/max_steps": "784/2513", "percentage": "31.20%", "elapsed_time": "2h 9m 22s", "remaining_time": "4h 45m 20s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100992} +{"loss": 0.4006823, "grad_norm": 0.03845968, "learning_rate": 8.053e-05, "epoch": 0.3124689, "global_step/max_steps": "785/2513", "percentage": "31.24%", "elapsed_time": "2h 9m 32s", "remaining_time": "4h 45m 9s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100997} +{"loss": 0.38881841, "grad_norm": 0.04092117, "learning_rate": 8.048e-05, "epoch": 0.31286695, "global_step/max_steps": "786/2513", "percentage": "31.28%", "elapsed_time": "2h 9m 42s", "remaining_time": "4h 44m 59s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100999} +{"loss": 0.43073025, "grad_norm": 0.03700325, "learning_rate": 8.043e-05, "epoch": 0.313265, "global_step/max_steps": "787/2513", "percentage": "31.32%", "elapsed_time": "2h 9m 52s", "remaining_time": "4h 44m 49s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101001} +{"loss": 0.37892127, "grad_norm": 0.03972764, "learning_rate": 8.038e-05, "epoch": 0.31366305, "global_step/max_steps": "788/2513", "percentage": "31.36%", "elapsed_time": "2h 10m 1s", "remaining_time": "4h 44m 38s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101004} +{"loss": 0.41667813, "grad_norm": 0.03953572, "learning_rate": 8.033e-05, "epoch": 0.3140611, "global_step/max_steps": "789/2513", "percentage": "31.40%", "elapsed_time": "2h 10m 11s", "remaining_time": "4h 44m 28s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101006} +{"loss": 0.38850617, "grad_norm": 0.03531028, "learning_rate": 8.027e-05, "epoch": 0.31445915, "global_step/max_steps": "790/2513", "percentage": "31.44%", "elapsed_time": "2h 10m 21s", "remaining_time": "4h 44m 18s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101006} +{"loss": 0.39348304, "grad_norm": 0.03681746, "learning_rate": 8.022e-05, "epoch": 0.3148572, "global_step/max_steps": "791/2513", "percentage": "31.48%", "elapsed_time": "2h 10m 31s", "remaining_time": "4h 44m 9s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100999} +{"loss": 0.41172081, "grad_norm": 0.03976989, "learning_rate": 8.017e-05, "epoch": 0.31525525, "global_step/max_steps": "792/2513", "percentage": "31.52%", "elapsed_time": "2h 10m 43s", "remaining_time": "4h 44m 3s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100979} +{"loss": 0.4127056, "grad_norm": 0.03716265, "learning_rate": 8.012e-05, "epoch": 0.3156533, "global_step/max_steps": "793/2513", "percentage": "31.56%", "elapsed_time": "2h 10m 52s", "remaining_time": "4h 43m 52s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100983} +{"loss": 0.37577024, "grad_norm": 0.04042723, "learning_rate": 8.007e-05, "epoch": 0.31605135, "global_step/max_steps": "794/2513", "percentage": "31.60%", "elapsed_time": "2h 11m 2s", "remaining_time": "4h 43m 42s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100982} +{"loss": 0.37990063, "grad_norm": 0.038814, "learning_rate": 8.002e-05, "epoch": 0.3164494, "global_step/max_steps": "795/2513", "percentage": "31.64%", "elapsed_time": "2h 11m 12s", "remaining_time": "4h 43m 32s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100986} +{"loss": 0.40193933, "grad_norm": 0.0392374, "learning_rate": 7.997e-05, "epoch": 0.31684745, "global_step/max_steps": "796/2513", "percentage": "31.68%", "elapsed_time": "2h 11m 22s", "remaining_time": "4h 43m 22s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100987} +{"loss": 0.34289101, "grad_norm": 0.03778835, "learning_rate": 7.991e-05, "epoch": 0.3172455, "global_step/max_steps": "797/2513", "percentage": "31.72%", "elapsed_time": "2h 11m 31s", "remaining_time": "4h 43m 11s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100992} +{"loss": 0.36352071, "grad_norm": 0.03668873, "learning_rate": 7.986e-05, "epoch": 0.31764355, "global_step/max_steps": "798/2513", "percentage": "31.75%", "elapsed_time": "2h 11m 41s", "remaining_time": "4h 43m 0s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100996} +{"loss": 0.45587265, "grad_norm": 0.0413241, "learning_rate": 7.981e-05, "epoch": 0.3180416, "global_step/max_steps": "799/2513", "percentage": "31.79%", "elapsed_time": "2h 11m 50s", "remaining_time": "4h 42m 50s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101} +{"loss": 0.36033323, "grad_norm": 0.04178113, "learning_rate": 7.976e-05, "epoch": 0.31843965, "global_step/max_steps": "800/2513", "percentage": "31.83%", "elapsed_time": "2h 12m 0s", "remaining_time": "4h 42m 39s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101004} +{"loss": 0.40316498, "grad_norm": 0.04242905, "learning_rate": 7.971e-05, "epoch": 0.3188377, "global_step/max_steps": "801/2513", "percentage": "31.87%", "elapsed_time": "2h 12m 11s", "remaining_time": "4h 42m 32s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100987} +{"loss": 0.37317044, "grad_norm": 0.04079316, "learning_rate": 7.965e-05, "epoch": 0.31923574, "global_step/max_steps": "802/2513", "percentage": "31.91%", "elapsed_time": "2h 12m 21s", "remaining_time": "4h 42m 22s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10099} +{"loss": 0.35714197, "grad_norm": 0.03941842, "learning_rate": 7.96e-05, "epoch": 0.31963379, "global_step/max_steps": "803/2513", "percentage": "31.95%", "elapsed_time": "2h 12m 30s", "remaining_time": "4h 42m 11s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100995} +{"loss": 0.37631887, "grad_norm": 0.03901567, "learning_rate": 7.955e-05, "epoch": 0.32003184, "global_step/max_steps": "804/2513", "percentage": "31.99%", "elapsed_time": "2h 12m 40s", "remaining_time": "4h 42m 1s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100998} +{"loss": 0.3798742, "grad_norm": 0.03501267, "learning_rate": 7.95e-05, "epoch": 0.32042989, "global_step/max_steps": "805/2513", "percentage": "32.03%", "elapsed_time": "2h 12m 50s", "remaining_time": "4h 41m 50s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101} +{"loss": 0.42090937, "grad_norm": 0.03984685, "learning_rate": 7.945e-05, "epoch": 0.32082794, "global_step/max_steps": "806/2513", "percentage": "32.07%", "elapsed_time": "2h 13m 0s", "remaining_time": "4h 41m 41s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100994} +{"loss": 0.35065854, "grad_norm": 0.04165682, "learning_rate": 7.939e-05, "epoch": 0.32122599, "global_step/max_steps": "807/2513", "percentage": "32.11%", "elapsed_time": "2h 13m 12s", "remaining_time": "4h 41m 35s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100976} +{"loss": 0.36929151, "grad_norm": 0.0336529, "learning_rate": 7.934e-05, "epoch": 0.32162404, "global_step/max_steps": "808/2513", "percentage": "32.15%", "elapsed_time": "2h 13m 22s", "remaining_time": "4h 41m 25s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100974} +{"loss": 0.36210445, "grad_norm": 0.03756483, "learning_rate": 7.929e-05, "epoch": 0.32202209, "global_step/max_steps": "809/2513", "percentage": "32.19%", "elapsed_time": "2h 13m 31s", "remaining_time": "4h 41m 15s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100977} +{"loss": 0.36490309, "grad_norm": 0.03876079, "learning_rate": 7.924e-05, "epoch": 0.32242014, "global_step/max_steps": "810/2513", "percentage": "32.23%", "elapsed_time": "2h 13m 41s", "remaining_time": "4h 41m 5s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100975} +{"loss": 0.38519478, "grad_norm": 0.0368257, "learning_rate": 7.919e-05, "epoch": 0.32281819, "global_step/max_steps": "811/2513", "percentage": "32.27%", "elapsed_time": "2h 13m 51s", "remaining_time": "4h 40m 54s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100981} +{"loss": 0.3771666, "grad_norm": 0.03685035, "learning_rate": 7.913e-05, "epoch": 0.32321624, "global_step/max_steps": "812/2513", "percentage": "32.31%", "elapsed_time": "2h 14m 0s", "remaining_time": "4h 40m 43s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100986} +{"loss": 0.3616257, "grad_norm": 0.03450239, "learning_rate": 7.908e-05, "epoch": 0.32361429, "global_step/max_steps": "813/2513", "percentage": "32.35%", "elapsed_time": "2h 14m 10s", "remaining_time": "4h 40m 33s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100987} +{"loss": 0.3554754, "grad_norm": 0.0341185, "learning_rate": 7.903e-05, "epoch": 0.32401234, "global_step/max_steps": "814/2513", "percentage": "32.39%", "elapsed_time": "2h 14m 20s", "remaining_time": "4h 40m 23s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100989} +{"loss": 0.38767582, "grad_norm": 0.0387764, "learning_rate": 7.898e-05, "epoch": 0.32441039, "global_step/max_steps": "815/2513", "percentage": "32.43%", "elapsed_time": "2h 14m 30s", "remaining_time": "4h 40m 13s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100988} +{"loss": 0.3509081, "grad_norm": 0.03615564, "learning_rate": 7.892e-05, "epoch": 0.32480844, "global_step/max_steps": "816/2513", "percentage": "32.47%", "elapsed_time": "2h 14m 39s", "remaining_time": "4h 40m 3s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100991} +{"loss": 0.38403076, "grad_norm": 0.04262887, "learning_rate": 7.887e-05, "epoch": 0.32520649, "global_step/max_steps": "817/2513", "percentage": "32.51%", "elapsed_time": "2h 14m 49s", "remaining_time": "4h 39m 53s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100993} +{"loss": 0.38940755, "grad_norm": 0.03902119, "learning_rate": 7.882e-05, "epoch": 0.32560454, "global_step/max_steps": "818/2513", "percentage": "32.55%", "elapsed_time": "2h 14m 59s", "remaining_time": "4h 39m 42s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100998} +{"loss": 0.38104567, "grad_norm": 0.03814755, "learning_rate": 7.877e-05, "epoch": 0.32600259, "global_step/max_steps": "819/2513", "percentage": "32.59%", "elapsed_time": "2h 15m 9s", "remaining_time": "4h 39m 32s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100997} +{"loss": 0.45925599, "grad_norm": 0.04135591, "learning_rate": 7.871e-05, "epoch": 0.32640064, "global_step/max_steps": "820/2513", "percentage": "32.63%", "elapsed_time": "2h 15m 18s", "remaining_time": "4h 39m 22s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100999} +{"loss": 0.37731951, "grad_norm": 0.03688401, "learning_rate": 7.866e-05, "epoch": 0.32679869, "global_step/max_steps": "821/2513", "percentage": "32.67%", "elapsed_time": "2h 15m 29s", "remaining_time": "4h 39m 14s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100989} +{"loss": 0.3715212, "grad_norm": 0.03686271, "learning_rate": 7.861e-05, "epoch": 0.32719674, "global_step/max_steps": "822/2513", "percentage": "32.71%", "elapsed_time": "2h 15m 40s", "remaining_time": "4h 39m 5s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100982} +{"loss": 0.41119152, "grad_norm": 0.04200467, "learning_rate": 7.855e-05, "epoch": 0.32759479, "global_step/max_steps": "823/2513", "percentage": "32.75%", "elapsed_time": "2h 15m 49s", "remaining_time": "4h 38m 54s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100987} +{"loss": 0.37336019, "grad_norm": 0.03685286, "learning_rate": 7.85e-05, "epoch": 0.32799284, "global_step/max_steps": "824/2513", "percentage": "32.79%", "elapsed_time": "2h 15m 59s", "remaining_time": "4h 38m 44s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100988} +{"loss": 0.39030004, "grad_norm": 0.03846774, "learning_rate": 7.845e-05, "epoch": 0.32839088, "global_step/max_steps": "825/2513", "percentage": "32.83%", "elapsed_time": "2h 16m 9s", "remaining_time": "4h 38m 34s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100991} +{"loss": 0.37967491, "grad_norm": 0.0355303, "learning_rate": 7.84e-05, "epoch": 0.32878893, "global_step/max_steps": "826/2513", "percentage": "32.87%", "elapsed_time": "2h 16m 18s", "remaining_time": "4h 38m 23s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100995} +{"loss": 0.3854385, "grad_norm": 0.03697951, "learning_rate": 7.834e-05, "epoch": 0.32918698, "global_step/max_steps": "827/2513", "percentage": "32.91%", "elapsed_time": "2h 16m 28s", "remaining_time": "4h 38m 13s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100998} +{"loss": 0.39857015, "grad_norm": 0.03818137, "learning_rate": 7.829e-05, "epoch": 0.32958503, "global_step/max_steps": "828/2513", "percentage": "32.95%", "elapsed_time": "2h 16m 37s", "remaining_time": "4h 38m 3s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101001} +{"loss": 0.37179261, "grad_norm": 0.03797709, "learning_rate": 7.824e-05, "epoch": 0.32998308, "global_step/max_steps": "829/2513", "percentage": "32.99%", "elapsed_time": "2h 16m 48s", "remaining_time": "4h 37m 53s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100999} +{"loss": 0.37895507, "grad_norm": 0.03993439, "learning_rate": 7.818e-05, "epoch": 0.33038113, "global_step/max_steps": "830/2513", "percentage": "33.03%", "elapsed_time": "2h 16m 57s", "remaining_time": "4h 37m 42s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101005} +{"loss": 0.43075043, "grad_norm": 0.03772867, "learning_rate": 7.813e-05, "epoch": 0.33077918, "global_step/max_steps": "831/2513", "percentage": "33.07%", "elapsed_time": "2h 17m 6s", "remaining_time": "4h 37m 31s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10101} +{"loss": 0.41675153, "grad_norm": 0.04086562, "learning_rate": 7.808e-05, "epoch": 0.33117723, "global_step/max_steps": "832/2513", "percentage": "33.11%", "elapsed_time": "2h 17m 16s", "remaining_time": "4h 37m 21s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101012} +{"loss": 0.38082051, "grad_norm": 0.0369474, "learning_rate": 7.802e-05, "epoch": 0.33157528, "global_step/max_steps": "833/2513", "percentage": "33.15%", "elapsed_time": "2h 17m 26s", "remaining_time": "4h 37m 11s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101013} +{"loss": 0.35852081, "grad_norm": 0.0377942, "learning_rate": 7.797e-05, "epoch": 0.33197333, "global_step/max_steps": "834/2513", "percentage": "33.19%", "elapsed_time": "2h 17m 36s", "remaining_time": "4h 37m 2s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10101} +{"loss": 0.33808962, "grad_norm": 0.03649956, "learning_rate": 7.792e-05, "epoch": 0.33237138, "global_step/max_steps": "835/2513", "percentage": "33.23%", "elapsed_time": "2h 17m 46s", "remaining_time": "4h 36m 52s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101006} +{"loss": 0.41049653, "grad_norm": 0.04058657, "learning_rate": 7.786e-05, "epoch": 0.33276943, "global_step/max_steps": "836/2513", "percentage": "33.27%", "elapsed_time": "2h 17m 57s", "remaining_time": "4h 36m 44s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100996} +{"loss": 0.38160712, "grad_norm": 0.04103353, "learning_rate": 7.781e-05, "epoch": 0.33316748, "global_step/max_steps": "837/2513", "percentage": "33.31%", "elapsed_time": "2h 18m 7s", "remaining_time": "4h 36m 35s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100992} +{"loss": 0.44044662, "grad_norm": 0.03967549, "learning_rate": 7.776e-05, "epoch": 0.33356553, "global_step/max_steps": "838/2513", "percentage": "33.35%", "elapsed_time": "2h 18m 17s", "remaining_time": "4h 36m 25s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100989} +{"loss": 0.36082491, "grad_norm": 0.0382564, "learning_rate": 7.77e-05, "epoch": 0.33396358, "global_step/max_steps": "839/2513", "percentage": "33.39%", "elapsed_time": "2h 18m 27s", "remaining_time": "4h 36m 15s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100995} +{"loss": 0.32890093, "grad_norm": 0.0368955, "learning_rate": 7.765e-05, "epoch": 0.33436163, "global_step/max_steps": "840/2513", "percentage": "33.43%", "elapsed_time": "2h 18m 37s", "remaining_time": "4h 36m 5s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100993} +{"loss": 0.38580719, "grad_norm": 0.04153829, "learning_rate": 7.759e-05, "epoch": 0.33475968, "global_step/max_steps": "841/2513", "percentage": "33.47%", "elapsed_time": "2h 18m 47s", "remaining_time": "4h 35m 55s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100995} +{"loss": 0.38952315, "grad_norm": 0.03997328, "learning_rate": 7.754e-05, "epoch": 0.33515773, "global_step/max_steps": "842/2513", "percentage": "33.51%", "elapsed_time": "2h 18m 56s", "remaining_time": "4h 35m 45s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100997} +{"loss": 0.3573083, "grad_norm": 0.03689604, "learning_rate": 7.749e-05, "epoch": 0.33555578, "global_step/max_steps": "843/2513", "percentage": "33.55%", "elapsed_time": "2h 19m 6s", "remaining_time": "4h 35m 34s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101001} +{"loss": 0.37234917, "grad_norm": 0.03780849, "learning_rate": 7.743e-05, "epoch": 0.33595383, "global_step/max_steps": "844/2513", "percentage": "33.59%", "elapsed_time": "2h 19m 16s", "remaining_time": "4h 35m 24s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101004} +{"loss": 0.42667228, "grad_norm": 0.045156, "learning_rate": 7.738e-05, "epoch": 0.33635188, "global_step/max_steps": "845/2513", "percentage": "33.63%", "elapsed_time": "2h 19m 25s", "remaining_time": "4h 35m 13s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101005} +{"loss": 0.43250334, "grad_norm": 0.03775199, "learning_rate": 7.732e-05, "epoch": 0.33674993, "global_step/max_steps": "846/2513", "percentage": "33.66%", "elapsed_time": "2h 19m 35s", "remaining_time": "4h 35m 3s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10101} +{"loss": 0.37967938, "grad_norm": 0.03737088, "learning_rate": 7.727e-05, "epoch": 0.33714797, "global_step/max_steps": "847/2513", "percentage": "33.70%", "elapsed_time": "2h 19m 45s", "remaining_time": "4h 34m 52s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101013} +{"loss": 0.37069595, "grad_norm": 0.03673559, "learning_rate": 7.722e-05, "epoch": 0.33754602, "global_step/max_steps": "848/2513", "percentage": "33.74%", "elapsed_time": "2h 19m 54s", "remaining_time": "4h 34m 42s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101018} +{"loss": 0.37293565, "grad_norm": 0.0410213, "learning_rate": 7.716e-05, "epoch": 0.33794407, "global_step/max_steps": "849/2513", "percentage": "33.78%", "elapsed_time": "2h 20m 4s", "remaining_time": "4h 34m 32s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101016} +{"loss": 0.40777352, "grad_norm": 0.04338143, "learning_rate": 7.711e-05, "epoch": 0.33834212, "global_step/max_steps": "850/2513", "percentage": "33.82%", "elapsed_time": "2h 20m 15s", "remaining_time": "4h 34m 23s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10101} +{"loss": 0.4113583, "grad_norm": 0.04091765, "learning_rate": 7.705e-05, "epoch": 0.33874017, "global_step/max_steps": "851/2513", "percentage": "33.86%", "elapsed_time": "2h 20m 26s", "remaining_time": "4h 34m 16s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100995} +{"loss": 0.40371776, "grad_norm": 0.03801095, "learning_rate": 7.7e-05, "epoch": 0.33913822, "global_step/max_steps": "852/2513", "percentage": "33.90%", "elapsed_time": "2h 20m 36s", "remaining_time": "4h 34m 6s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100995} +{"loss": 0.40404898, "grad_norm": 0.03659429, "learning_rate": 7.695e-05, "epoch": 0.33953627, "global_step/max_steps": "853/2513", "percentage": "33.94%", "elapsed_time": "2h 20m 45s", "remaining_time": "4h 33m 56s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100996} +{"loss": 0.37208742, "grad_norm": 0.03978645, "learning_rate": 7.689e-05, "epoch": 0.33993432, "global_step/max_steps": "854/2513", "percentage": "33.98%", "elapsed_time": "2h 20m 55s", "remaining_time": "4h 33m 45s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101} +{"loss": 0.41594142, "grad_norm": 0.03977927, "learning_rate": 7.684e-05, "epoch": 0.34033237, "global_step/max_steps": "855/2513", "percentage": "34.02%", "elapsed_time": "2h 21m 5s", "remaining_time": "4h 33m 35s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101003} +{"loss": 0.39853957, "grad_norm": 0.03896909, "learning_rate": 7.678e-05, "epoch": 0.34073042, "global_step/max_steps": "856/2513", "percentage": "34.06%", "elapsed_time": "2h 21m 14s", "remaining_time": "4h 33m 25s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101003} +{"loss": 0.37848955, "grad_norm": 0.03740137, "learning_rate": 7.673e-05, "epoch": 0.34112847, "global_step/max_steps": "857/2513", "percentage": "34.10%", "elapsed_time": "2h 21m 24s", "remaining_time": "4h 33m 14s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101008} +{"loss": 0.38843659, "grad_norm": 0.03886664, "learning_rate": 7.667e-05, "epoch": 0.34152652, "global_step/max_steps": "858/2513", "percentage": "34.14%", "elapsed_time": "2h 21m 34s", "remaining_time": "4h 33m 4s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101009} +{"loss": 0.38097715, "grad_norm": 0.03455376, "learning_rate": 7.662e-05, "epoch": 0.34192457, "global_step/max_steps": "859/2513", "percentage": "34.18%", "elapsed_time": "2h 21m 44s", "remaining_time": "4h 32m 54s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101008} +{"loss": 0.35322568, "grad_norm": 0.03473708, "learning_rate": 7.656e-05, "epoch": 0.34232262, "global_step/max_steps": "860/2513", "percentage": "34.22%", "elapsed_time": "2h 21m 54s", "remaining_time": "4h 32m 44s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101008} +{"loss": 0.39355797, "grad_norm": 0.036222, "learning_rate": 7.651e-05, "epoch": 0.34272067, "global_step/max_steps": "861/2513", "percentage": "34.26%", "elapsed_time": "2h 22m 3s", "remaining_time": "4h 32m 34s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101009} +{"loss": 0.38389623, "grad_norm": 0.03840085, "learning_rate": 7.646e-05, "epoch": 0.34311872, "global_step/max_steps": "862/2513", "percentage": "34.30%", "elapsed_time": "2h 22m 13s", "remaining_time": "4h 32m 24s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101015} +{"loss": 0.31250823, "grad_norm": 0.03214924, "learning_rate": 7.64e-05, "epoch": 0.34351677, "global_step/max_steps": "863/2513", "percentage": "34.34%", "elapsed_time": "2h 22m 23s", "remaining_time": "4h 32m 13s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101016} +{"loss": 0.39971405, "grad_norm": 0.03537471, "learning_rate": 7.635e-05, "epoch": 0.34391482, "global_step/max_steps": "864/2513", "percentage": "34.38%", "elapsed_time": "2h 22m 33s", "remaining_time": "4h 32m 4s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101013} +{"loss": 0.38381764, "grad_norm": 0.03700933, "learning_rate": 7.629e-05, "epoch": 0.34431287, "global_step/max_steps": "865/2513", "percentage": "34.42%", "elapsed_time": "2h 22m 44s", "remaining_time": "4h 31m 56s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101001} +{"loss": 0.40065113, "grad_norm": 0.03646943, "learning_rate": 7.624e-05, "epoch": 0.34471092, "global_step/max_steps": "866/2513", "percentage": "34.46%", "elapsed_time": "2h 22m 54s", "remaining_time": "4h 31m 48s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100992} +{"loss": 0.37887979, "grad_norm": 0.03927763, "learning_rate": 7.618e-05, "epoch": 0.34510897, "global_step/max_steps": "867/2513", "percentage": "34.50%", "elapsed_time": "2h 23m 4s", "remaining_time": "4h 31m 37s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100996} +{"loss": 0.35138291, "grad_norm": 0.03903328, "learning_rate": 7.613e-05, "epoch": 0.34550702, "global_step/max_steps": "868/2513", "percentage": "34.54%", "elapsed_time": "2h 23m 14s", "remaining_time": "4h 31m 27s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100996} +{"loss": 0.40744275, "grad_norm": 0.03821612, "learning_rate": 7.607e-05, "epoch": 0.34590507, "global_step/max_steps": "869/2513", "percentage": "34.58%", "elapsed_time": "2h 23m 23s", "remaining_time": "4h 31m 17s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101} +{"loss": 0.38508648, "grad_norm": 0.0361767, "learning_rate": 7.602e-05, "epoch": 0.34630311, "global_step/max_steps": "870/2513", "percentage": "34.62%", "elapsed_time": "2h 23m 33s", "remaining_time": "4h 31m 7s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101002} +{"loss": 0.38719559, "grad_norm": 0.03674431, "learning_rate": 7.596e-05, "epoch": 0.34670116, "global_step/max_steps": "871/2513", "percentage": "34.66%", "elapsed_time": "2h 23m 43s", "remaining_time": "4h 30m 56s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101005} +{"loss": 0.36946744, "grad_norm": 0.03750655, "learning_rate": 7.591e-05, "epoch": 0.34709921, "global_step/max_steps": "872/2513", "percentage": "34.70%", "elapsed_time": "2h 23m 53s", "remaining_time": "4h 30m 46s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101006} +{"loss": 0.37711191, "grad_norm": 0.03507634, "learning_rate": 7.585e-05, "epoch": 0.34749726, "global_step/max_steps": "873/2513", "percentage": "34.74%", "elapsed_time": "2h 24m 2s", "remaining_time": "4h 30m 36s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101007} +{"loss": 0.37760052, "grad_norm": 0.03891025, "learning_rate": 7.58e-05, "epoch": 0.34789531, "global_step/max_steps": "874/2513", "percentage": "34.78%", "elapsed_time": "2h 24m 12s", "remaining_time": "4h 30m 26s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101008} +{"loss": 0.45565644, "grad_norm": 0.040917, "learning_rate": 7.574e-05, "epoch": 0.34829336, "global_step/max_steps": "875/2513", "percentage": "34.82%", "elapsed_time": "2h 24m 22s", "remaining_time": "4h 30m 16s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101009} +{"loss": 0.42572892, "grad_norm": 0.04082896, "learning_rate": 7.569e-05, "epoch": 0.34869141, "global_step/max_steps": "876/2513", "percentage": "34.86%", "elapsed_time": "2h 24m 32s", "remaining_time": "4h 30m 6s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101011} +{"loss": 0.40921286, "grad_norm": 0.04092672, "learning_rate": 7.563e-05, "epoch": 0.34908946, "global_step/max_steps": "877/2513", "percentage": "34.90%", "elapsed_time": "2h 24m 42s", "remaining_time": "4h 29m 56s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101011} +{"loss": 0.40336752, "grad_norm": 0.04617387, "learning_rate": 7.557e-05, "epoch": 0.34948751, "global_step/max_steps": "878/2513", "percentage": "34.94%", "elapsed_time": "2h 24m 52s", "remaining_time": "4h 29m 46s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10101} +{"loss": 0.38324648, "grad_norm": 0.03994398, "learning_rate": 7.552e-05, "epoch": 0.34988556, "global_step/max_steps": "879/2513", "percentage": "34.98%", "elapsed_time": "2h 25m 2s", "remaining_time": "4h 29m 37s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101006} +{"loss": 0.38350207, "grad_norm": 0.03736774, "learning_rate": 7.546e-05, "epoch": 0.35028361, "global_step/max_steps": "880/2513", "percentage": "35.02%", "elapsed_time": "2h 25m 13s", "remaining_time": "4h 29m 30s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100989} +{"loss": 0.38762969, "grad_norm": 0.03676005, "learning_rate": 7.541e-05, "epoch": 0.35068166, "global_step/max_steps": "881/2513", "percentage": "35.06%", "elapsed_time": "2h 25m 23s", "remaining_time": "4h 29m 20s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100988} +{"loss": 0.38616794, "grad_norm": 0.03804822, "learning_rate": 7.535e-05, "epoch": 0.35107971, "global_step/max_steps": "882/2513", "percentage": "35.10%", "elapsed_time": "2h 25m 33s", "remaining_time": "4h 29m 10s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10099} +{"loss": 0.38520694, "grad_norm": 0.03622513, "learning_rate": 7.53e-05, "epoch": 0.35147776, "global_step/max_steps": "883/2513", "percentage": "35.14%", "elapsed_time": "2h 25m 43s", "remaining_time": "4h 28m 59s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100991} +{"loss": 0.34872723, "grad_norm": 0.03466821, "learning_rate": 7.524e-05, "epoch": 0.35187581, "global_step/max_steps": "884/2513", "percentage": "35.18%", "elapsed_time": "2h 25m 52s", "remaining_time": "4h 28m 49s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100996} +{"loss": 0.40488541, "grad_norm": 0.0393163, "learning_rate": 7.519e-05, "epoch": 0.35227386, "global_step/max_steps": "885/2513", "percentage": "35.22%", "elapsed_time": "2h 26m 2s", "remaining_time": "4h 28m 39s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100997} +{"loss": 0.37199894, "grad_norm": 0.03879182, "learning_rate": 7.513e-05, "epoch": 0.35267191, "global_step/max_steps": "886/2513", "percentage": "35.26%", "elapsed_time": "2h 26m 12s", "remaining_time": "4h 28m 29s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100998} +{"loss": 0.39584643, "grad_norm": 0.03818459, "learning_rate": 7.507e-05, "epoch": 0.35306996, "global_step/max_steps": "887/2513", "percentage": "35.30%", "elapsed_time": "2h 26m 21s", "remaining_time": "4h 28m 18s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101003} +{"loss": 0.33084017, "grad_norm": 0.03564368, "learning_rate": 7.502e-05, "epoch": 0.35346801, "global_step/max_steps": "888/2513", "percentage": "35.34%", "elapsed_time": "2h 26m 31s", "remaining_time": "4h 28m 7s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101011} +{"loss": 0.44291592, "grad_norm": 0.04019806, "learning_rate": 7.496e-05, "epoch": 0.35386606, "global_step/max_steps": "889/2513", "percentage": "35.38%", "elapsed_time": "2h 26m 40s", "remaining_time": "4h 27m 57s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101012} +{"loss": 0.36531907, "grad_norm": 0.03671797, "learning_rate": 7.491e-05, "epoch": 0.35426411, "global_step/max_steps": "890/2513", "percentage": "35.42%", "elapsed_time": "2h 26m 50s", "remaining_time": "4h 27m 46s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101016} +{"loss": 0.36903489, "grad_norm": 0.03758556, "learning_rate": 7.485e-05, "epoch": 0.35466216, "global_step/max_steps": "891/2513", "percentage": "35.46%", "elapsed_time": "2h 27m 0s", "remaining_time": "4h 27m 36s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101018} +{"loss": 0.36947808, "grad_norm": 0.03336417, "learning_rate": 7.48e-05, "epoch": 0.3550602, "global_step/max_steps": "892/2513", "percentage": "35.50%", "elapsed_time": "2h 27m 9s", "remaining_time": "4h 27m 26s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101021} +{"loss": 0.3765167, "grad_norm": 0.04533408, "learning_rate": 7.474e-05, "epoch": 0.35545825, "global_step/max_steps": "893/2513", "percentage": "35.54%", "elapsed_time": "2h 27m 19s", "remaining_time": "4h 27m 15s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101023} +{"loss": 0.38363227, "grad_norm": 0.03382675, "learning_rate": 7.468e-05, "epoch": 0.3558563, "global_step/max_steps": "894/2513", "percentage": "35.58%", "elapsed_time": "2h 27m 29s", "remaining_time": "4h 27m 6s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101018} +{"loss": 0.37906677, "grad_norm": 0.03885265, "learning_rate": 7.463e-05, "epoch": 0.35625435, "global_step/max_steps": "895/2513", "percentage": "35.61%", "elapsed_time": "2h 27m 40s", "remaining_time": "4h 26m 58s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101008} +{"loss": 0.3474654, "grad_norm": 0.03593923, "learning_rate": 7.457e-05, "epoch": 0.3566524, "global_step/max_steps": "896/2513", "percentage": "35.65%", "elapsed_time": "2h 27m 50s", "remaining_time": "4h 26m 48s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10101} +{"loss": 0.39322144, "grad_norm": 0.03644192, "learning_rate": 7.451e-05, "epoch": 0.35705045, "global_step/max_steps": "897/2513", "percentage": "35.69%", "elapsed_time": "2h 28m 0s", "remaining_time": "4h 26m 38s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101011} +{"loss": 0.33685333, "grad_norm": 0.03276667, "learning_rate": 7.446e-05, "epoch": 0.3574485, "global_step/max_steps": "898/2513", "percentage": "35.73%", "elapsed_time": "2h 28m 9s", "remaining_time": "4h 26m 28s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101013} +{"loss": 0.37606442, "grad_norm": 0.03812505, "learning_rate": 7.44e-05, "epoch": 0.35784655, "global_step/max_steps": "899/2513", "percentage": "35.77%", "elapsed_time": "2h 28m 20s", "remaining_time": "4h 26m 18s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101011} +{"loss": 0.36392602, "grad_norm": 0.03463996, "learning_rate": 7.435e-05, "epoch": 0.3582446, "global_step/max_steps": "900/2513", "percentage": "35.81%", "elapsed_time": "2h 28m 29s", "remaining_time": "4h 26m 7s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101015} +{"loss": 0.36136517, "grad_norm": 0.03409098, "learning_rate": 7.429e-05, "epoch": 0.35864265, "global_step/max_steps": "901/2513", "percentage": "35.85%", "elapsed_time": "2h 28m 40s", "remaining_time": "4h 26m 0s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100999} +{"loss": 0.40745592, "grad_norm": 0.03762789, "learning_rate": 7.423e-05, "epoch": 0.3590407, "global_step/max_steps": "902/2513", "percentage": "35.89%", "elapsed_time": "2h 28m 50s", "remaining_time": "4h 25m 49s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101004} +{"loss": 0.38606927, "grad_norm": 0.03687134, "learning_rate": 7.418e-05, "epoch": 0.35943875, "global_step/max_steps": "903/2513", "percentage": "35.93%", "elapsed_time": "2h 28m 59s", "remaining_time": "4h 25m 39s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101009} +{"loss": 0.39273688, "grad_norm": 0.03592107, "learning_rate": 7.412e-05, "epoch": 0.3598368, "global_step/max_steps": "904/2513", "percentage": "35.97%", "elapsed_time": "2h 29m 9s", "remaining_time": "4h 25m 29s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101009} +{"loss": 0.3452507, "grad_norm": 0.03862759, "learning_rate": 7.406e-05, "epoch": 0.36023485, "global_step/max_steps": "905/2513", "percentage": "36.01%", "elapsed_time": "2h 29m 19s", "remaining_time": "4h 25m 18s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101013} +{"loss": 0.40722784, "grad_norm": 0.04001311, "learning_rate": 7.401e-05, "epoch": 0.3606329, "global_step/max_steps": "906/2513", "percentage": "36.05%", "elapsed_time": "2h 29m 29s", "remaining_time": "4h 25m 8s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101013} +{"loss": 0.4047904, "grad_norm": 0.04036228, "learning_rate": 7.395e-05, "epoch": 0.36103095, "global_step/max_steps": "907/2513", "percentage": "36.09%", "elapsed_time": "2h 29m 38s", "remaining_time": "4h 24m 58s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101015} +{"loss": 0.39961457, "grad_norm": 0.03661867, "learning_rate": 7.389e-05, "epoch": 0.361429, "global_step/max_steps": "908/2513", "percentage": "36.13%", "elapsed_time": "2h 29m 49s", "remaining_time": "4h 24m 49s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101012} +{"loss": 0.45012933, "grad_norm": 0.04301021, "learning_rate": 7.384e-05, "epoch": 0.36182705, "global_step/max_steps": "909/2513", "percentage": "36.17%", "elapsed_time": "2h 29m 59s", "remaining_time": "4h 24m 41s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101} +{"loss": 0.38506526, "grad_norm": 0.03582938, "learning_rate": 7.378e-05, "epoch": 0.3622251, "global_step/max_steps": "910/2513", "percentage": "36.21%", "elapsed_time": "2h 30m 9s", "remaining_time": "4h 24m 31s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101} +{"loss": 0.38369614, "grad_norm": 0.04281361, "learning_rate": 7.372e-05, "epoch": 0.36262315, "global_step/max_steps": "911/2513", "percentage": "36.25%", "elapsed_time": "2h 30m 19s", "remaining_time": "4h 24m 21s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.100999} +{"loss": 0.35413826, "grad_norm": 0.03640938, "learning_rate": 7.367e-05, "epoch": 0.3630212, "global_step/max_steps": "912/2513", "percentage": "36.29%", "elapsed_time": "2h 30m 29s", "remaining_time": "4h 24m 11s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101} +{"loss": 0.3729496, "grad_norm": 0.03458349, "learning_rate": 7.361e-05, "epoch": 0.36341925, "global_step/max_steps": "913/2513", "percentage": "36.33%", "elapsed_time": "2h 30m 39s", "remaining_time": "4h 24m 1s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101003} +{"loss": 0.43280846, "grad_norm": 0.03715081, "learning_rate": 7.355e-05, "epoch": 0.3638173, "global_step/max_steps": "914/2513", "percentage": "36.37%", "elapsed_time": "2h 30m 49s", "remaining_time": "4h 23m 51s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101001} +{"loss": 0.43558112, "grad_norm": 0.0418675, "learning_rate": 7.35e-05, "epoch": 0.36421534, "global_step/max_steps": "915/2513", "percentage": "36.41%", "elapsed_time": "2h 30m 58s", "remaining_time": "4h 23m 40s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101007} +{"loss": 0.39991227, "grad_norm": 0.04160349, "learning_rate": 7.344e-05, "epoch": 0.36461339, "global_step/max_steps": "916/2513", "percentage": "36.45%", "elapsed_time": "2h 31m 8s", "remaining_time": "4h 23m 29s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101012} +{"loss": 0.38336569, "grad_norm": 0.03914576, "learning_rate": 7.338e-05, "epoch": 0.36501144, "global_step/max_steps": "917/2513", "percentage": "36.49%", "elapsed_time": "2h 31m 17s", "remaining_time": "4h 23m 19s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101015} +{"loss": 0.42035204, "grad_norm": 0.03768742, "learning_rate": 7.333e-05, "epoch": 0.36540949, "global_step/max_steps": "918/2513", "percentage": "36.53%", "elapsed_time": "2h 31m 27s", "remaining_time": "4h 23m 9s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101019} +{"loss": 0.40368515, "grad_norm": 0.03736846, "learning_rate": 7.327e-05, "epoch": 0.36580754, "global_step/max_steps": "919/2513", "percentage": "36.57%", "elapsed_time": "2h 31m 36s", "remaining_time": "4h 22m 58s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101022} +{"loss": 0.38748667, "grad_norm": 0.03442677, "learning_rate": 7.321e-05, "epoch": 0.36620559, "global_step/max_steps": "920/2513", "percentage": "36.61%", "elapsed_time": "2h 31m 46s", "remaining_time": "4h 22m 48s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101022} +{"loss": 0.38036674, "grad_norm": 0.03756398, "learning_rate": 7.315e-05, "epoch": 0.36660364, "global_step/max_steps": "921/2513", "percentage": "36.65%", "elapsed_time": "2h 31m 56s", "remaining_time": "4h 22m 38s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101025} +{"loss": 0.41424477, "grad_norm": 0.03837287, "learning_rate": 7.31e-05, "epoch": 0.36700169, "global_step/max_steps": "922/2513", "percentage": "36.69%", "elapsed_time": "2h 32m 6s", "remaining_time": "4h 22m 28s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101026} +{"loss": 0.34526056, "grad_norm": 0.03994035, "learning_rate": 7.304e-05, "epoch": 0.36739974, "global_step/max_steps": "923/2513", "percentage": "36.73%", "elapsed_time": "2h 32m 15s", "remaining_time": "4h 22m 17s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101032} +{"loss": 0.34458128, "grad_norm": 0.03501834, "learning_rate": 7.298e-05, "epoch": 0.36779779, "global_step/max_steps": "924/2513", "percentage": "36.77%", "elapsed_time": "2h 32m 27s", "remaining_time": "4h 22m 10s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101016} +{"loss": 0.39862898, "grad_norm": 0.03901136, "learning_rate": 7.293e-05, "epoch": 0.36819584, "global_step/max_steps": "925/2513", "percentage": "36.81%", "elapsed_time": "2h 32m 37s", "remaining_time": "4h 22m 1s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101011} +{"loss": 0.37571847, "grad_norm": 0.03585529, "learning_rate": 7.287e-05, "epoch": 0.36859389, "global_step/max_steps": "926/2513", "percentage": "36.85%", "elapsed_time": "2h 32m 47s", "remaining_time": "4h 21m 51s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101012} +{"loss": 0.36580124, "grad_norm": 0.03419143, "learning_rate": 7.281e-05, "epoch": 0.36899194, "global_step/max_steps": "927/2513", "percentage": "36.89%", "elapsed_time": "2h 32m 57s", "remaining_time": "4h 21m 41s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101011} +{"loss": 0.32010889, "grad_norm": 0.03543742, "learning_rate": 7.275e-05, "epoch": 0.36938999, "global_step/max_steps": "928/2513", "percentage": "36.93%", "elapsed_time": "2h 33m 6s", "remaining_time": "4h 21m 30s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101016} +{"loss": 0.39869839, "grad_norm": 0.03974191, "learning_rate": 7.27e-05, "epoch": 0.36978804, "global_step/max_steps": "929/2513", "percentage": "36.97%", "elapsed_time": "2h 33m 16s", "remaining_time": "4h 21m 20s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101019} +{"loss": 0.3677727, "grad_norm": 0.03906206, "learning_rate": 7.264e-05, "epoch": 0.37018609, "global_step/max_steps": "930/2513", "percentage": "37.01%", "elapsed_time": "2h 33m 25s", "remaining_time": "4h 21m 9s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101023} +{"loss": 0.40108192, "grad_norm": 0.03816475, "learning_rate": 7.258e-05, "epoch": 0.37058414, "global_step/max_steps": "931/2513", "percentage": "37.05%", "elapsed_time": "2h 33m 35s", "remaining_time": "4h 20m 59s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101027} +{"loss": 0.42170107, "grad_norm": 0.04174533, "learning_rate": 7.252e-05, "epoch": 0.37098219, "global_step/max_steps": "932/2513", "percentage": "37.09%", "elapsed_time": "2h 33m 45s", "remaining_time": "4h 20m 49s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101028} +{"loss": 0.38418043, "grad_norm": 0.03630768, "learning_rate": 7.247e-05, "epoch": 0.37138024, "global_step/max_steps": "933/2513", "percentage": "37.13%", "elapsed_time": "2h 33m 55s", "remaining_time": "4h 20m 39s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101028} +{"loss": 0.37611216, "grad_norm": 0.03948016, "learning_rate": 7.241e-05, "epoch": 0.37177829, "global_step/max_steps": "934/2513", "percentage": "37.17%", "elapsed_time": "2h 34m 4s", "remaining_time": "4h 20m 28s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101034} +{"loss": 0.40509993, "grad_norm": 0.03466319, "learning_rate": 7.235e-05, "epoch": 0.37217634, "global_step/max_steps": "935/2513", "percentage": "37.21%", "elapsed_time": "2h 34m 15s", "remaining_time": "4h 20m 20s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101024} +{"loss": 0.38211221, "grad_norm": 0.03894168, "learning_rate": 7.229e-05, "epoch": 0.37257439, "global_step/max_steps": "936/2513", "percentage": "37.25%", "elapsed_time": "2h 34m 24s", "remaining_time": "4h 20m 9s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101028} +{"loss": 0.39059779, "grad_norm": 0.03636175, "learning_rate": 7.224e-05, "epoch": 0.37297244, "global_step/max_steps": "937/2513", "percentage": "37.29%", "elapsed_time": "2h 34m 34s", "remaining_time": "4h 19m 59s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101027} +{"loss": 0.39571631, "grad_norm": 0.03563861, "learning_rate": 7.218e-05, "epoch": 0.37337048, "global_step/max_steps": "938/2513", "percentage": "37.33%", "elapsed_time": "2h 34m 44s", "remaining_time": "4h 19m 50s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101026} +{"loss": 0.38543636, "grad_norm": 0.03826202, "learning_rate": 7.212e-05, "epoch": 0.37376853, "global_step/max_steps": "939/2513", "percentage": "37.37%", "elapsed_time": "2h 34m 55s", "remaining_time": "4h 19m 41s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101014} +{"loss": 0.37814438, "grad_norm": 0.03884284, "learning_rate": 7.206e-05, "epoch": 0.37416658, "global_step/max_steps": "940/2513", "percentage": "37.41%", "elapsed_time": "2h 35m 5s", "remaining_time": "4h 19m 32s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101011} +{"loss": 0.39927667, "grad_norm": 0.03769216, "learning_rate": 7.2e-05, "epoch": 0.37456463, "global_step/max_steps": "941/2513", "percentage": "37.45%", "elapsed_time": "2h 35m 15s", "remaining_time": "4h 19m 22s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101011} +{"loss": 0.39600465, "grad_norm": 0.03688247, "learning_rate": 7.195e-05, "epoch": 0.37496268, "global_step/max_steps": "942/2513", "percentage": "37.49%", "elapsed_time": "2h 35m 25s", "remaining_time": "4h 19m 12s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101013} +{"loss": 0.36443537, "grad_norm": 0.03425132, "learning_rate": 7.189e-05, "epoch": 0.37536073, "global_step/max_steps": "943/2513", "percentage": "37.52%", "elapsed_time": "2h 35m 35s", "remaining_time": "4h 19m 2s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101016} +{"loss": 0.36200106, "grad_norm": 0.03508241, "learning_rate": 7.183e-05, "epoch": 0.37575878, "global_step/max_steps": "944/2513", "percentage": "37.56%", "elapsed_time": "2h 35m 44s", "remaining_time": "4h 18m 51s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101019} +{"loss": 0.41134423, "grad_norm": 0.03681577, "learning_rate": 7.177e-05, "epoch": 0.37615683, "global_step/max_steps": "945/2513", "percentage": "37.60%", "elapsed_time": "2h 35m 54s", "remaining_time": "4h 18m 41s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101019} +{"loss": 0.34571284, "grad_norm": 0.03714686, "learning_rate": 7.171e-05, "epoch": 0.37655488, "global_step/max_steps": "946/2513", "percentage": "37.64%", "elapsed_time": "2h 36m 4s", "remaining_time": "4h 18m 31s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101019} +{"loss": 0.36968923, "grad_norm": 0.03629059, "learning_rate": 7.166e-05, "epoch": 0.37695293, "global_step/max_steps": "947/2513", "percentage": "37.68%", "elapsed_time": "2h 36m 14s", "remaining_time": "4h 18m 21s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101019} +{"loss": 0.36034191, "grad_norm": 0.03533786, "learning_rate": 7.16e-05, "epoch": 0.37735098, "global_step/max_steps": "948/2513", "percentage": "37.72%", "elapsed_time": "2h 36m 24s", "remaining_time": "4h 18m 12s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101019} +{"loss": 0.36499146, "grad_norm": 0.03734614, "learning_rate": 7.154e-05, "epoch": 0.37774903, "global_step/max_steps": "949/2513", "percentage": "37.76%", "elapsed_time": "2h 36m 34s", "remaining_time": "4h 18m 2s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101021} +{"loss": 0.38827372, "grad_norm": 0.03693075, "learning_rate": 7.148e-05, "epoch": 0.37814708, "global_step/max_steps": "950/2513", "percentage": "37.80%", "elapsed_time": "2h 36m 43s", "remaining_time": "4h 17m 51s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101024} +{"loss": 0.38592327, "grad_norm": 0.03621794, "learning_rate": 7.142e-05, "epoch": 0.37854513, "global_step/max_steps": "951/2513", "percentage": "37.84%", "elapsed_time": "2h 36m 53s", "remaining_time": "4h 17m 41s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101025} +{"loss": 0.38558152, "grad_norm": 0.03601804, "learning_rate": 7.137e-05, "epoch": 0.37894318, "global_step/max_steps": "952/2513", "percentage": "37.88%", "elapsed_time": "2h 37m 3s", "remaining_time": "4h 17m 31s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101025} +{"loss": 0.4045001, "grad_norm": 0.04196599, "learning_rate": 7.131e-05, "epoch": 0.37934123, "global_step/max_steps": "953/2513", "percentage": "37.92%", "elapsed_time": "2h 37m 13s", "remaining_time": "4h 17m 22s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101019} +{"loss": 0.38026965, "grad_norm": 0.03897747, "learning_rate": 7.125e-05, "epoch": 0.37973928, "global_step/max_steps": "954/2513", "percentage": "37.96%", "elapsed_time": "2h 37m 24s", "remaining_time": "4h 17m 13s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101014} +{"loss": 0.3695735, "grad_norm": 0.03477455, "learning_rate": 7.119e-05, "epoch": 0.38013733, "global_step/max_steps": "955/2513", "percentage": "38.00%", "elapsed_time": "2h 37m 34s", "remaining_time": "4h 17m 4s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101006} +{"loss": 0.37135607, "grad_norm": 0.03789524, "learning_rate": 7.113e-05, "epoch": 0.38053538, "global_step/max_steps": "956/2513", "percentage": "38.04%", "elapsed_time": "2h 37m 44s", "remaining_time": "4h 16m 54s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101009} +{"loss": 0.40044227, "grad_norm": 0.03737127, "learning_rate": 7.107e-05, "epoch": 0.38093343, "global_step/max_steps": "957/2513", "percentage": "38.08%", "elapsed_time": "2h 37m 54s", "remaining_time": "4h 16m 43s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101013} +{"loss": 0.41522104, "grad_norm": 0.03598981, "learning_rate": 7.102e-05, "epoch": 0.38133148, "global_step/max_steps": "958/2513", "percentage": "38.12%", "elapsed_time": "2h 38m 3s", "remaining_time": "4h 16m 33s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101019} +{"loss": 0.40096131, "grad_norm": 0.03774059, "learning_rate": 7.096e-05, "epoch": 0.38172953, "global_step/max_steps": "959/2513", "percentage": "38.16%", "elapsed_time": "2h 38m 12s", "remaining_time": "4h 16m 22s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101023} +{"loss": 0.3933754, "grad_norm": 0.03564509, "learning_rate": 7.09e-05, "epoch": 0.38212757, "global_step/max_steps": "960/2513", "percentage": "38.20%", "elapsed_time": "2h 38m 22s", "remaining_time": "4h 16m 12s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101026} +{"loss": 0.39521086, "grad_norm": 0.03513177, "learning_rate": 7.084e-05, "epoch": 0.38252562, "global_step/max_steps": "961/2513", "percentage": "38.24%", "elapsed_time": "2h 38m 32s", "remaining_time": "4h 16m 2s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101028} +{"loss": 0.36216098, "grad_norm": 0.0335679, "learning_rate": 7.078e-05, "epoch": 0.38292367, "global_step/max_steps": "962/2513", "percentage": "38.28%", "elapsed_time": "2h 38m 42s", "remaining_time": "4h 15m 52s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101028} +{"loss": 0.37822139, "grad_norm": 0.03735408, "learning_rate": 7.072e-05, "epoch": 0.38332172, "global_step/max_steps": "963/2513", "percentage": "38.32%", "elapsed_time": "2h 38m 52s", "remaining_time": "4h 15m 42s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101028} +{"loss": 0.35074729, "grad_norm": 0.03581619, "learning_rate": 7.066e-05, "epoch": 0.38371977, "global_step/max_steps": "964/2513", "percentage": "38.36%", "elapsed_time": "2h 39m 1s", "remaining_time": "4h 15m 31s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101033} +{"loss": 0.38919592, "grad_norm": 0.03781261, "learning_rate": 7.061e-05, "epoch": 0.38411782, "global_step/max_steps": "965/2513", "percentage": "38.40%", "elapsed_time": "2h 39m 10s", "remaining_time": "4h 15m 21s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101037} +{"loss": 0.38271934, "grad_norm": 0.03527826, "learning_rate": 7.055e-05, "epoch": 0.38451587, "global_step/max_steps": "966/2513", "percentage": "38.44%", "elapsed_time": "2h 39m 20s", "remaining_time": "4h 15m 11s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101037} +{"loss": 0.38837332, "grad_norm": 0.03620968, "learning_rate": 7.049e-05, "epoch": 0.38491392, "global_step/max_steps": "967/2513", "percentage": "38.48%", "elapsed_time": "2h 39m 30s", "remaining_time": "4h 15m 0s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101041} +{"loss": 0.39330301, "grad_norm": 0.03308778, "learning_rate": 7.043e-05, "epoch": 0.38531197, "global_step/max_steps": "968/2513", "percentage": "38.52%", "elapsed_time": "2h 39m 41s", "remaining_time": "4h 14m 52s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101032} +{"loss": 0.35627541, "grad_norm": 0.03545079, "learning_rate": 7.037e-05, "epoch": 0.38571002, "global_step/max_steps": "969/2513", "percentage": "38.56%", "elapsed_time": "2h 39m 51s", "remaining_time": "4h 14m 42s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101031} +{"loss": 0.38529137, "grad_norm": 0.0402377, "learning_rate": 7.031e-05, "epoch": 0.38610807, "global_step/max_steps": "970/2513", "percentage": "38.60%", "elapsed_time": "2h 40m 1s", "remaining_time": "4h 14m 32s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101029} +{"loss": 0.35601002, "grad_norm": 0.03400808, "learning_rate": 7.025e-05, "epoch": 0.38650612, "global_step/max_steps": "971/2513", "percentage": "38.64%", "elapsed_time": "2h 40m 11s", "remaining_time": "4h 14m 23s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101023} +{"loss": 0.39265147, "grad_norm": 0.0333173, "learning_rate": 7.019e-05, "epoch": 0.38690417, "global_step/max_steps": "972/2513", "percentage": "38.68%", "elapsed_time": "2h 40m 21s", "remaining_time": "4h 14m 13s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101025} +{"loss": 0.39435601, "grad_norm": 0.05299295, "learning_rate": 7.013e-05, "epoch": 0.38730222, "global_step/max_steps": "973/2513", "percentage": "38.72%", "elapsed_time": "2h 40m 30s", "remaining_time": "4h 14m 3s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101029} +{"loss": 0.39438865, "grad_norm": 0.03971574, "learning_rate": 7.008e-05, "epoch": 0.38770027, "global_step/max_steps": "974/2513", "percentage": "38.76%", "elapsed_time": "2h 40m 40s", "remaining_time": "4h 13m 52s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101035} +{"loss": 0.38243371, "grad_norm": 0.03557932, "learning_rate": 7.002e-05, "epoch": 0.38809832, "global_step/max_steps": "975/2513", "percentage": "38.80%", "elapsed_time": "2h 40m 50s", "remaining_time": "4h 13m 42s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101034} +{"loss": 0.37422031, "grad_norm": 0.0385891, "learning_rate": 6.996e-05, "epoch": 0.38849637, "global_step/max_steps": "976/2513", "percentage": "38.84%", "elapsed_time": "2h 40m 59s", "remaining_time": "4h 13m 32s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101037} +{"loss": 0.34791243, "grad_norm": 0.03510102, "learning_rate": 6.99e-05, "epoch": 0.38889442, "global_step/max_steps": "977/2513", "percentage": "38.88%", "elapsed_time": "2h 41m 9s", "remaining_time": "4h 13m 22s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101039} +{"loss": 0.35134307, "grad_norm": 0.03573894, "learning_rate": 6.984e-05, "epoch": 0.38929247, "global_step/max_steps": "978/2513", "percentage": "38.92%", "elapsed_time": "2h 41m 19s", "remaining_time": "4h 13m 12s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101039} +{"loss": 0.38994175, "grad_norm": 0.03673695, "learning_rate": 6.978e-05, "epoch": 0.38969052, "global_step/max_steps": "979/2513", "percentage": "38.96%", "elapsed_time": "2h 41m 29s", "remaining_time": "4h 13m 2s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101039} +{"loss": 0.36288595, "grad_norm": 0.03825318, "learning_rate": 6.972e-05, "epoch": 0.39008857, "global_step/max_steps": "980/2513", "percentage": "39.00%", "elapsed_time": "2h 41m 39s", "remaining_time": "4h 12m 53s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101034} +{"loss": 0.36086243, "grad_norm": 0.03661553, "learning_rate": 6.966e-05, "epoch": 0.39048662, "global_step/max_steps": "981/2513", "percentage": "39.04%", "elapsed_time": "2h 41m 49s", "remaining_time": "4h 12m 43s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101034} +{"loss": 0.37888929, "grad_norm": 0.03842705, "learning_rate": 6.96e-05, "epoch": 0.39088467, "global_step/max_steps": "982/2513", "percentage": "39.08%", "elapsed_time": "2h 41m 59s", "remaining_time": "4h 12m 33s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101033} +{"loss": 0.36904302, "grad_norm": 0.03584861, "learning_rate": 6.954e-05, "epoch": 0.39128271, "global_step/max_steps": "983/2513", "percentage": "39.12%", "elapsed_time": "2h 42m 9s", "remaining_time": "4h 12m 24s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101028} +{"loss": 0.37160385, "grad_norm": 0.03762048, "learning_rate": 6.948e-05, "epoch": 0.39168076, "global_step/max_steps": "984/2513", "percentage": "39.16%", "elapsed_time": "2h 42m 20s", "remaining_time": "4h 12m 15s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10102} +{"loss": 0.41812199, "grad_norm": 0.04049577, "learning_rate": 6.942e-05, "epoch": 0.39207881, "global_step/max_steps": "985/2513", "percentage": "39.20%", "elapsed_time": "2h 42m 29s", "remaining_time": "4h 12m 4s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101026} +{"loss": 0.32994658, "grad_norm": 0.03413456, "learning_rate": 6.936e-05, "epoch": 0.39247686, "global_step/max_steps": "986/2513", "percentage": "39.24%", "elapsed_time": "2h 42m 39s", "remaining_time": "4h 11m 54s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101028} +{"loss": 0.39913845, "grad_norm": 0.04013721, "learning_rate": 6.931e-05, "epoch": 0.39287491, "global_step/max_steps": "987/2513", "percentage": "39.28%", "elapsed_time": "2h 42m 49s", "remaining_time": "4h 11m 44s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10103} +{"loss": 0.38746616, "grad_norm": 0.0342662, "learning_rate": 6.925e-05, "epoch": 0.39327296, "global_step/max_steps": "988/2513", "percentage": "39.32%", "elapsed_time": "2h 43m 0s", "remaining_time": "4h 11m 35s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101021} +{"loss": 0.34796464, "grad_norm": 0.03895526, "learning_rate": 6.919e-05, "epoch": 0.39367101, "global_step/max_steps": "989/2513", "percentage": "39.36%", "elapsed_time": "2h 43m 9s", "remaining_time": "4h 11m 25s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101023} +{"loss": 0.38768473, "grad_norm": 0.03836879, "learning_rate": 6.913e-05, "epoch": 0.39406906, "global_step/max_steps": "990/2513", "percentage": "39.40%", "elapsed_time": "2h 43m 19s", "remaining_time": "4h 11m 15s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101026} +{"loss": 0.34458268, "grad_norm": 0.03478513, "learning_rate": 6.907e-05, "epoch": 0.39446711, "global_step/max_steps": "991/2513", "percentage": "39.43%", "elapsed_time": "2h 43m 29s", "remaining_time": "4h 11m 5s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101025} +{"loss": 0.40342209, "grad_norm": 0.03790583, "learning_rate": 6.901e-05, "epoch": 0.39486516, "global_step/max_steps": "992/2513", "percentage": "39.47%", "elapsed_time": "2h 43m 39s", "remaining_time": "4h 10m 55s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101024} +{"loss": 0.39587304, "grad_norm": 0.0409203, "learning_rate": 6.895e-05, "epoch": 0.39526321, "global_step/max_steps": "993/2513", "percentage": "39.51%", "elapsed_time": "2h 43m 48s", "remaining_time": "4h 10m 45s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101029} +{"loss": 0.38266701, "grad_norm": 0.03559684, "learning_rate": 6.889e-05, "epoch": 0.39566126, "global_step/max_steps": "994/2513", "percentage": "39.55%", "elapsed_time": "2h 43m 58s", "remaining_time": "4h 10m 34s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101035} +{"loss": 0.39486161, "grad_norm": 0.03560005, "learning_rate": 6.883e-05, "epoch": 0.39605931, "global_step/max_steps": "995/2513", "percentage": "39.59%", "elapsed_time": "2h 44m 8s", "remaining_time": "4h 10m 25s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101031} +{"loss": 0.35067332, "grad_norm": 0.03726057, "learning_rate": 6.877e-05, "epoch": 0.39645736, "global_step/max_steps": "996/2513", "percentage": "39.63%", "elapsed_time": "2h 44m 18s", "remaining_time": "4h 10m 15s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10103} +{"loss": 0.34614921, "grad_norm": 0.03447244, "learning_rate": 6.871e-05, "epoch": 0.39685541, "global_step/max_steps": "997/2513", "percentage": "39.67%", "elapsed_time": "2h 44m 28s", "remaining_time": "4h 10m 5s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10103} +{"loss": 0.41518676, "grad_norm": 0.03898026, "learning_rate": 6.865e-05, "epoch": 0.39725346, "global_step/max_steps": "998/2513", "percentage": "39.71%", "elapsed_time": "2h 44m 38s", "remaining_time": "4h 9m 55s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101027} +{"loss": 0.36745521, "grad_norm": 0.04149122, "learning_rate": 6.859e-05, "epoch": 0.39765151, "global_step/max_steps": "999/2513", "percentage": "39.75%", "elapsed_time": "2h 44m 48s", "remaining_time": "4h 9m 46s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101021} +{"loss": 0.36087105, "grad_norm": 0.03627334, "learning_rate": 6.853e-05, "epoch": 0.39804956, "global_step/max_steps": "1000/2513", "percentage": "39.79%", "elapsed_time": "2h 44m 58s", "remaining_time": "4h 9m 36s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101026} +{"loss": 0.39311105, "grad_norm": 0.03857073, "learning_rate": 6.847e-05, "epoch": 0.39844761, "global_step/max_steps": "1001/2513", "percentage": "39.83%", "elapsed_time": "2h 45m 9s", "remaining_time": "4h 9m 28s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101013} +{"loss": 0.38886791, "grad_norm": 0.03862334, "learning_rate": 6.841e-05, "epoch": 0.39884566, "global_step/max_steps": "1002/2513", "percentage": "39.87%", "elapsed_time": "2h 45m 19s", "remaining_time": "4h 9m 17s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101016} +{"loss": 0.33682191, "grad_norm": 0.03479142, "learning_rate": 6.835e-05, "epoch": 0.39924371, "global_step/max_steps": "1003/2513", "percentage": "39.91%", "elapsed_time": "2h 45m 28s", "remaining_time": "4h 9m 7s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101018} +{"loss": 0.33000332, "grad_norm": 0.03764594, "learning_rate": 6.829e-05, "epoch": 0.39964176, "global_step/max_steps": "1004/2513", "percentage": "39.95%", "elapsed_time": "2h 45m 38s", "remaining_time": "4h 8m 57s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101021} +{"loss": 0.38881731, "grad_norm": 0.07310547, "learning_rate": 6.823e-05, "epoch": 0.4000398, "global_step/max_steps": "1005/2513", "percentage": "39.99%", "elapsed_time": "2h 45m 48s", "remaining_time": "4h 8m 47s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101023} +{"loss": 0.39167526, "grad_norm": 0.03882829, "learning_rate": 6.817e-05, "epoch": 0.40043785, "global_step/max_steps": "1006/2513", "percentage": "40.03%", "elapsed_time": "2h 45m 58s", "remaining_time": "4h 8m 37s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101023} +{"loss": 0.36362571, "grad_norm": 0.03453396, "learning_rate": 6.811e-05, "epoch": 0.4008359, "global_step/max_steps": "1007/2513", "percentage": "40.07%", "elapsed_time": "2h 46m 7s", "remaining_time": "4h 8m 26s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101027} +{"loss": 0.37414801, "grad_norm": 0.03290429, "learning_rate": 6.805e-05, "epoch": 0.40123395, "global_step/max_steps": "1008/2513", "percentage": "40.11%", "elapsed_time": "2h 46m 17s", "remaining_time": "4h 8m 16s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101028} +{"loss": 0.3900916, "grad_norm": 0.03916964, "learning_rate": 6.799e-05, "epoch": 0.401632, "global_step/max_steps": "1009/2513", "percentage": "40.15%", "elapsed_time": "2h 46m 27s", "remaining_time": "4h 8m 7s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101026} +{"loss": 0.42974371, "grad_norm": 0.03819191, "learning_rate": 6.793e-05, "epoch": 0.40203005, "global_step/max_steps": "1010/2513", "percentage": "40.19%", "elapsed_time": "2h 46m 37s", "remaining_time": "4h 7m 57s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101026} +{"loss": 0.36876351, "grad_norm": 0.03859538, "learning_rate": 6.787e-05, "epoch": 0.4024281, "global_step/max_steps": "1011/2513", "percentage": "40.23%", "elapsed_time": "2h 46m 47s", "remaining_time": "4h 7m 47s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101029} +{"loss": 0.36860508, "grad_norm": 0.0385901, "learning_rate": 6.781e-05, "epoch": 0.40282615, "global_step/max_steps": "1012/2513", "percentage": "40.27%", "elapsed_time": "2h 46m 57s", "remaining_time": "4h 7m 38s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101018} +{"loss": 0.35071564, "grad_norm": 0.03621693, "learning_rate": 6.775e-05, "epoch": 0.4032242, "global_step/max_steps": "1013/2513", "percentage": "40.31%", "elapsed_time": "2h 47m 8s", "remaining_time": "4h 7m 29s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101016} +{"loss": 0.39238366, "grad_norm": 0.03686262, "learning_rate": 6.769e-05, "epoch": 0.40362225, "global_step/max_steps": "1014/2513", "percentage": "40.35%", "elapsed_time": "2h 47m 17s", "remaining_time": "4h 7m 18s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101018} +{"loss": 0.36293536, "grad_norm": 0.03702721, "learning_rate": 6.763e-05, "epoch": 0.4040203, "global_step/max_steps": "1015/2513", "percentage": "40.39%", "elapsed_time": "2h 47m 27s", "remaining_time": "4h 7m 8s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101023} +{"loss": 0.3900308, "grad_norm": 0.03907849, "learning_rate": 6.757e-05, "epoch": 0.40441835, "global_step/max_steps": "1016/2513", "percentage": "40.43%", "elapsed_time": "2h 47m 37s", "remaining_time": "4h 6m 58s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101024} +{"loss": 0.36020175, "grad_norm": 0.03628784, "learning_rate": 6.751e-05, "epoch": 0.4048164, "global_step/max_steps": "1017/2513", "percentage": "40.47%", "elapsed_time": "2h 47m 46s", "remaining_time": "4h 6m 47s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101027} +{"loss": 0.35455018, "grad_norm": 0.03467605, "learning_rate": 6.745e-05, "epoch": 0.40521445, "global_step/max_steps": "1018/2513", "percentage": "40.51%", "elapsed_time": "2h 47m 56s", "remaining_time": "4h 6m 37s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101031} +{"loss": 0.36921439, "grad_norm": 0.03868725, "learning_rate": 6.739e-05, "epoch": 0.4056125, "global_step/max_steps": "1019/2513", "percentage": "40.55%", "elapsed_time": "2h 48m 5s", "remaining_time": "4h 6m 27s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101034} +{"loss": 0.39923328, "grad_norm": 0.05340121, "learning_rate": 6.733e-05, "epoch": 0.40601055, "global_step/max_steps": "1020/2513", "percentage": "40.59%", "elapsed_time": "2h 48m 15s", "remaining_time": "4h 6m 17s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101032} +{"loss": 0.43419299, "grad_norm": 0.04043588, "learning_rate": 6.727e-05, "epoch": 0.4064086, "global_step/max_steps": "1021/2513", "percentage": "40.63%", "elapsed_time": "2h 48m 25s", "remaining_time": "4h 6m 6s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101037} +{"loss": 0.36826307, "grad_norm": 0.03483811, "learning_rate": 6.721e-05, "epoch": 0.40680665, "global_step/max_steps": "1022/2513", "percentage": "40.67%", "elapsed_time": "2h 48m 35s", "remaining_time": "4h 5m 56s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101038} +{"loss": 0.40029258, "grad_norm": 0.03897471, "learning_rate": 6.714e-05, "epoch": 0.4072047, "global_step/max_steps": "1023/2513", "percentage": "40.71%", "elapsed_time": "2h 48m 44s", "remaining_time": "4h 5m 46s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101039} +{"loss": 0.37274992, "grad_norm": 0.03863596, "learning_rate": 6.708e-05, "epoch": 0.40760275, "global_step/max_steps": "1024/2513", "percentage": "40.75%", "elapsed_time": "2h 48m 54s", "remaining_time": "4h 5m 36s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10104} +{"loss": 0.37175894, "grad_norm": 0.03718995, "learning_rate": 6.702e-05, "epoch": 0.4080008, "global_step/max_steps": "1025/2513", "percentage": "40.79%", "elapsed_time": "2h 49m 4s", "remaining_time": "4h 5m 27s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101039} +{"loss": 0.39320028, "grad_norm": 0.04263934, "learning_rate": 6.696e-05, "epoch": 0.40839885, "global_step/max_steps": "1026/2513", "percentage": "40.83%", "elapsed_time": "2h 49m 14s", "remaining_time": "4h 5m 17s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101034} +{"loss": 0.38057494, "grad_norm": 0.03721488, "learning_rate": 6.69e-05, "epoch": 0.4087969, "global_step/max_steps": "1027/2513", "percentage": "40.87%", "elapsed_time": "2h 49m 25s", "remaining_time": "4h 5m 9s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101025} +{"loss": 0.3735382, "grad_norm": 0.03557236, "learning_rate": 6.684e-05, "epoch": 0.40919494, "global_step/max_steps": "1028/2513", "percentage": "40.91%", "elapsed_time": "2h 49m 35s", "remaining_time": "4h 4m 59s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101024} +{"loss": 0.36271822, "grad_norm": 0.0343892, "learning_rate": 6.678e-05, "epoch": 0.40959299, "global_step/max_steps": "1029/2513", "percentage": "40.95%", "elapsed_time": "2h 49m 45s", "remaining_time": "4h 4m 49s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101025} +{"loss": 0.41627949, "grad_norm": 0.03857729, "learning_rate": 6.672e-05, "epoch": 0.40999104, "global_step/max_steps": "1030/2513", "percentage": "40.99%", "elapsed_time": "2h 49m 55s", "remaining_time": "4h 4m 39s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101025} +{"loss": 0.38395107, "grad_norm": 0.0374249, "learning_rate": 6.666e-05, "epoch": 0.41038909, "global_step/max_steps": "1031/2513", "percentage": "41.03%", "elapsed_time": "2h 50m 5s", "remaining_time": "4h 4m 29s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101027} +{"loss": 0.35306975, "grad_norm": 0.03432616, "learning_rate": 6.66e-05, "epoch": 0.41078714, "global_step/max_steps": "1032/2513", "percentage": "41.07%", "elapsed_time": "2h 50m 14s", "remaining_time": "4h 4m 18s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101032} +{"loss": 0.34373286, "grad_norm": 0.0364811, "learning_rate": 6.654e-05, "epoch": 0.41118519, "global_step/max_steps": "1033/2513", "percentage": "41.11%", "elapsed_time": "2h 50m 24s", "remaining_time": "4h 4m 8s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101031} +{"loss": 0.35526019, "grad_norm": 0.03453759, "learning_rate": 6.648e-05, "epoch": 0.41158324, "global_step/max_steps": "1034/2513", "percentage": "41.15%", "elapsed_time": "2h 50m 34s", "remaining_time": "4h 3m 59s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101031} +{"loss": 0.38312346, "grad_norm": 0.03792815, "learning_rate": 6.642e-05, "epoch": 0.41198129, "global_step/max_steps": "1035/2513", "percentage": "41.19%", "elapsed_time": "2h 50m 44s", "remaining_time": "4h 3m 49s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10103} +{"loss": 0.36205882, "grad_norm": 0.03888046, "learning_rate": 6.636e-05, "epoch": 0.41237934, "global_step/max_steps": "1036/2513", "percentage": "41.23%", "elapsed_time": "2h 50m 54s", "remaining_time": "4h 3m 39s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101032} +{"loss": 0.42208478, "grad_norm": 0.04188436, "learning_rate": 6.629e-05, "epoch": 0.41277739, "global_step/max_steps": "1037/2513", "percentage": "41.27%", "elapsed_time": "2h 51m 3s", "remaining_time": "4h 3m 28s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101035} +{"loss": 0.39329237, "grad_norm": 0.04107098, "learning_rate": 6.623e-05, "epoch": 0.41317544, "global_step/max_steps": "1038/2513", "percentage": "41.31%", "elapsed_time": "2h 51m 13s", "remaining_time": "4h 3m 18s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101039} +{"loss": 0.40946269, "grad_norm": 0.03816457, "learning_rate": 6.617e-05, "epoch": 0.41357349, "global_step/max_steps": "1039/2513", "percentage": "41.35%", "elapsed_time": "2h 51m 23s", "remaining_time": "4h 3m 8s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10104} +{"loss": 0.42808077, "grad_norm": 0.03973761, "learning_rate": 6.611e-05, "epoch": 0.41397154, "global_step/max_steps": "1040/2513", "percentage": "41.38%", "elapsed_time": "2h 51m 32s", "remaining_time": "4h 2m 58s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101041} +{"loss": 0.38644284, "grad_norm": 0.04066068, "learning_rate": 6.605e-05, "epoch": 0.41436959, "global_step/max_steps": "1041/2513", "percentage": "41.42%", "elapsed_time": "2h 51m 43s", "remaining_time": "4h 2m 49s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101035} +{"loss": 0.32538009, "grad_norm": 0.03521776, "learning_rate": 6.599e-05, "epoch": 0.41476764, "global_step/max_steps": "1042/2513", "percentage": "41.46%", "elapsed_time": "2h 51m 53s", "remaining_time": "4h 2m 40s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101029} +{"loss": 0.35929465, "grad_norm": 0.03586114, "learning_rate": 6.593e-05, "epoch": 0.41516569, "global_step/max_steps": "1043/2513", "percentage": "41.50%", "elapsed_time": "2h 52m 3s", "remaining_time": "4h 2m 30s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101027} +{"loss": 0.3878721, "grad_norm": 0.03556656, "learning_rate": 6.587e-05, "epoch": 0.41556374, "global_step/max_steps": "1044/2513", "percentage": "41.54%", "elapsed_time": "2h 52m 13s", "remaining_time": "4h 2m 20s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10103} +{"loss": 0.37210184, "grad_norm": 0.0354483, "learning_rate": 6.581e-05, "epoch": 0.41596179, "global_step/max_steps": "1045/2513", "percentage": "41.58%", "elapsed_time": "2h 52m 23s", "remaining_time": "4h 2m 10s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101032} +{"loss": 0.37046653, "grad_norm": 0.03880815, "learning_rate": 6.574e-05, "epoch": 0.41635984, "global_step/max_steps": "1046/2513", "percentage": "41.62%", "elapsed_time": "2h 52m 33s", "remaining_time": "4h 2m 0s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101031} +{"loss": 0.33823562, "grad_norm": 0.04151273, "learning_rate": 6.568e-05, "epoch": 0.41675789, "global_step/max_steps": "1047/2513", "percentage": "41.66%", "elapsed_time": "2h 52m 42s", "remaining_time": "4h 1m 49s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101034} +{"loss": 0.42243224, "grad_norm": 0.03997839, "learning_rate": 6.562e-05, "epoch": 0.41715594, "global_step/max_steps": "1048/2513", "percentage": "41.70%", "elapsed_time": "2h 52m 52s", "remaining_time": "4h 1m 39s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101039} +{"loss": 0.39314941, "grad_norm": 0.03510961, "learning_rate": 6.556e-05, "epoch": 0.41755399, "global_step/max_steps": "1049/2513", "percentage": "41.74%", "elapsed_time": "2h 53m 2s", "remaining_time": "4h 1m 29s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101039} +{"loss": 0.39251381, "grad_norm": 0.03867108, "learning_rate": 6.55e-05, "epoch": 0.41795204, "global_step/max_steps": "1050/2513", "percentage": "41.78%", "elapsed_time": "2h 53m 11s", "remaining_time": "4h 1m 19s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10104} +{"loss": 0.40018684, "grad_norm": 0.03682449, "learning_rate": 6.544e-05, "epoch": 0.41835008, "global_step/max_steps": "1051/2513", "percentage": "41.82%", "elapsed_time": "2h 53m 21s", "remaining_time": "4h 1m 8s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101045} +{"loss": 0.3927407, "grad_norm": 0.03811367, "learning_rate": 6.538e-05, "epoch": 0.41874813, "global_step/max_steps": "1052/2513", "percentage": "41.86%", "elapsed_time": "2h 53m 30s", "remaining_time": "4h 0m 58s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101047} +{"loss": 0.37567517, "grad_norm": 0.03864213, "learning_rate": 6.532e-05, "epoch": 0.41914618, "global_step/max_steps": "1053/2513", "percentage": "41.90%", "elapsed_time": "2h 53m 40s", "remaining_time": "4h 0m 48s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101049} +{"loss": 0.36238465, "grad_norm": 0.0347639, "learning_rate": 6.525e-05, "epoch": 0.41954423, "global_step/max_steps": "1054/2513", "percentage": "41.94%", "elapsed_time": "2h 53m 50s", "remaining_time": "4h 0m 38s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101051} +{"loss": 0.36084986, "grad_norm": 0.03614914, "learning_rate": 6.519e-05, "epoch": 0.41994228, "global_step/max_steps": "1055/2513", "percentage": "41.98%", "elapsed_time": "2h 54m 0s", "remaining_time": "4h 0m 28s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101052} +{"loss": 0.39993241, "grad_norm": 0.03659558, "learning_rate": 6.513e-05, "epoch": 0.42034033, "global_step/max_steps": "1056/2513", "percentage": "42.02%", "elapsed_time": "2h 54m 10s", "remaining_time": "4h 0m 19s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101045} +{"loss": 0.37827516, "grad_norm": 0.03375262, "learning_rate": 6.507e-05, "epoch": 0.42073838, "global_step/max_steps": "1057/2513", "percentage": "42.06%", "elapsed_time": "2h 54m 21s", "remaining_time": "4h 0m 10s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101037} +{"loss": 0.37344238, "grad_norm": 0.03791124, "learning_rate": 6.501e-05, "epoch": 0.42113643, "global_step/max_steps": "1058/2513", "percentage": "42.10%", "elapsed_time": "2h 54m 31s", "remaining_time": "4h 0m 1s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101034} +{"loss": 0.3893612, "grad_norm": 0.03696891, "learning_rate": 6.495e-05, "epoch": 0.42153448, "global_step/max_steps": "1059/2513", "percentage": "42.14%", "elapsed_time": "2h 54m 41s", "remaining_time": "3h 59m 51s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101034} +{"loss": 0.3908959, "grad_norm": 0.03595777, "learning_rate": 6.489e-05, "epoch": 0.42193253, "global_step/max_steps": "1060/2513", "percentage": "42.18%", "elapsed_time": "2h 54m 51s", "remaining_time": "3h 59m 41s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101036} +{"loss": 0.41778976, "grad_norm": 0.03978825, "learning_rate": 6.482e-05, "epoch": 0.42233058, "global_step/max_steps": "1061/2513", "percentage": "42.22%", "elapsed_time": "2h 55m 1s", "remaining_time": "3h 59m 30s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101037} +{"loss": 0.37240964, "grad_norm": 0.03804863, "learning_rate": 6.476e-05, "epoch": 0.42272863, "global_step/max_steps": "1062/2513", "percentage": "42.26%", "elapsed_time": "2h 55m 10s", "remaining_time": "3h 59m 20s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101041} +{"loss": 0.3569819, "grad_norm": 0.04022294, "learning_rate": 6.47e-05, "epoch": 0.42312668, "global_step/max_steps": "1063/2513", "percentage": "42.30%", "elapsed_time": "2h 55m 20s", "remaining_time": "3h 59m 10s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101045} +{"loss": 0.35362881, "grad_norm": 0.03587215, "learning_rate": 6.464e-05, "epoch": 0.42352473, "global_step/max_steps": "1064/2513", "percentage": "42.34%", "elapsed_time": "2h 55m 29s", "remaining_time": "3h 58m 59s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101047} +{"loss": 0.3394568, "grad_norm": 0.03579485, "learning_rate": 6.458e-05, "epoch": 0.42392278, "global_step/max_steps": "1065/2513", "percentage": "42.38%", "elapsed_time": "2h 55m 39s", "remaining_time": "3h 58m 49s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101049} +{"loss": 0.40600216, "grad_norm": 0.04022887, "learning_rate": 6.452e-05, "epoch": 0.42432083, "global_step/max_steps": "1066/2513", "percentage": "42.42%", "elapsed_time": "2h 55m 49s", "remaining_time": "3h 58m 39s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10105} +{"loss": 0.36610496, "grad_norm": 0.0360794, "learning_rate": 6.445e-05, "epoch": 0.42471888, "global_step/max_steps": "1067/2513", "percentage": "42.46%", "elapsed_time": "2h 55m 58s", "remaining_time": "3h 58m 29s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101054} +{"loss": 0.35441926, "grad_norm": 0.03759089, "learning_rate": 6.439e-05, "epoch": 0.42511693, "global_step/max_steps": "1068/2513", "percentage": "42.50%", "elapsed_time": "2h 56m 8s", "remaining_time": "3h 58m 19s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101056} +{"loss": 0.35290968, "grad_norm": 0.03809513, "learning_rate": 6.433e-05, "epoch": 0.42551498, "global_step/max_steps": "1069/2513", "percentage": "42.54%", "elapsed_time": "2h 56m 18s", "remaining_time": "3h 58m 9s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101054} +{"loss": 0.4438535, "grad_norm": 0.04101435, "learning_rate": 6.427e-05, "epoch": 0.42591303, "global_step/max_steps": "1070/2513", "percentage": "42.58%", "elapsed_time": "2h 56m 28s", "remaining_time": "3h 57m 59s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101051} +{"loss": 0.39407629, "grad_norm": 0.04115495, "learning_rate": 6.421e-05, "epoch": 0.42631108, "global_step/max_steps": "1071/2513", "percentage": "42.62%", "elapsed_time": "2h 56m 39s", "remaining_time": "3h 57m 51s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101043} +{"loss": 0.36172527, "grad_norm": 0.03800834, "learning_rate": 6.415e-05, "epoch": 0.42670913, "global_step/max_steps": "1072/2513", "percentage": "42.66%", "elapsed_time": "2h 56m 49s", "remaining_time": "3h 57m 41s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101041} +{"loss": 0.3715145, "grad_norm": 0.0371082, "learning_rate": 6.408e-05, "epoch": 0.42710717, "global_step/max_steps": "1073/2513", "percentage": "42.70%", "elapsed_time": "2h 56m 59s", "remaining_time": "3h 57m 31s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10104} +{"loss": 0.40530574, "grad_norm": 0.03905117, "learning_rate": 6.402e-05, "epoch": 0.42750522, "global_step/max_steps": "1074/2513", "percentage": "42.74%", "elapsed_time": "2h 57m 9s", "remaining_time": "3h 57m 21s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101042} +{"loss": 0.40361369, "grad_norm": 0.0420209, "learning_rate": 6.396e-05, "epoch": 0.42790327, "global_step/max_steps": "1075/2513", "percentage": "42.78%", "elapsed_time": "2h 57m 18s", "remaining_time": "3h 57m 11s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101044} +{"loss": 0.39499703, "grad_norm": 0.03719841, "learning_rate": 6.39e-05, "epoch": 0.42830132, "global_step/max_steps": "1076/2513", "percentage": "42.82%", "elapsed_time": "2h 57m 28s", "remaining_time": "3h 57m 1s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101045} +{"loss": 0.38722125, "grad_norm": 0.03627866, "learning_rate": 6.384e-05, "epoch": 0.42869937, "global_step/max_steps": "1077/2513", "percentage": "42.86%", "elapsed_time": "2h 57m 38s", "remaining_time": "3h 56m 50s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101049} +{"loss": 0.35349095, "grad_norm": 0.03528201, "learning_rate": 6.377e-05, "epoch": 0.42909742, "global_step/max_steps": "1078/2513", "percentage": "42.90%", "elapsed_time": "2h 57m 47s", "remaining_time": "3h 56m 40s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101052} +{"loss": 0.36078969, "grad_norm": 0.03721738, "learning_rate": 6.371e-05, "epoch": 0.42949547, "global_step/max_steps": "1079/2513", "percentage": "42.94%", "elapsed_time": "2h 57m 57s", "remaining_time": "3h 56m 30s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101052} +{"loss": 0.39640614, "grad_norm": 0.03676761, "learning_rate": 6.365e-05, "epoch": 0.42989352, "global_step/max_steps": "1080/2513", "percentage": "42.98%", "elapsed_time": "2h 58m 7s", "remaining_time": "3h 56m 20s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101053} +{"loss": 0.37652555, "grad_norm": 0.03617187, "learning_rate": 6.359e-05, "epoch": 0.43029157, "global_step/max_steps": "1081/2513", "percentage": "43.02%", "elapsed_time": "2h 58m 17s", "remaining_time": "3h 56m 10s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101055} +{"loss": 0.37932208, "grad_norm": 0.03778558, "learning_rate": 6.353e-05, "epoch": 0.43068962, "global_step/max_steps": "1082/2513", "percentage": "43.06%", "elapsed_time": "2h 58m 26s", "remaining_time": "3h 56m 0s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101056} +{"loss": 0.35589808, "grad_norm": 0.03550905, "learning_rate": 6.346e-05, "epoch": 0.43108767, "global_step/max_steps": "1083/2513", "percentage": "43.10%", "elapsed_time": "2h 58m 36s", "remaining_time": "3h 55m 50s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101057} +{"loss": 0.37179118, "grad_norm": 0.03969688, "learning_rate": 6.34e-05, "epoch": 0.43148572, "global_step/max_steps": "1084/2513", "percentage": "43.14%", "elapsed_time": "2h 58m 46s", "remaining_time": "3h 55m 40s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101061} +{"loss": 0.38533807, "grad_norm": 0.03770191, "learning_rate": 6.334e-05, "epoch": 0.43188377, "global_step/max_steps": "1085/2513", "percentage": "43.18%", "elapsed_time": "2h 58m 55s", "remaining_time": "3h 55m 29s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101062} +{"loss": 0.3488403, "grad_norm": 0.0360841, "learning_rate": 6.328e-05, "epoch": 0.43228182, "global_step/max_steps": "1086/2513", "percentage": "43.22%", "elapsed_time": "2h 59m 6s", "remaining_time": "3h 55m 20s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101056} +{"loss": 0.40375316, "grad_norm": 0.03967943, "learning_rate": 6.322e-05, "epoch": 0.43267987, "global_step/max_steps": "1087/2513", "percentage": "43.26%", "elapsed_time": "2h 59m 16s", "remaining_time": "3h 55m 11s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101052} +{"loss": 0.34396219, "grad_norm": 0.04057393, "learning_rate": 6.315e-05, "epoch": 0.43307792, "global_step/max_steps": "1088/2513", "percentage": "43.29%", "elapsed_time": "2h 59m 26s", "remaining_time": "3h 55m 1s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101052} +{"loss": 0.37530136, "grad_norm": 0.03957298, "learning_rate": 6.309e-05, "epoch": 0.43347597, "global_step/max_steps": "1089/2513", "percentage": "43.33%", "elapsed_time": "2h 59m 36s", "remaining_time": "3h 54m 51s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101053} +{"loss": 0.38569647, "grad_norm": 0.03696757, "learning_rate": 6.303e-05, "epoch": 0.43387402, "global_step/max_steps": "1090/2513", "percentage": "43.37%", "elapsed_time": "2h 59m 45s", "remaining_time": "3h 54m 41s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101057} +{"loss": 0.35508496, "grad_norm": 0.03715165, "learning_rate": 6.297e-05, "epoch": 0.43427207, "global_step/max_steps": "1091/2513", "percentage": "43.41%", "elapsed_time": "2h 59m 55s", "remaining_time": "3h 54m 30s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101061} +{"loss": 0.37459671, "grad_norm": 0.04010713, "learning_rate": 6.29e-05, "epoch": 0.43467012, "global_step/max_steps": "1092/2513", "percentage": "43.45%", "elapsed_time": "3h 0m 5s", "remaining_time": "3h 54m 20s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101063} +{"loss": 0.43121785, "grad_norm": 0.04025779, "learning_rate": 6.284e-05, "epoch": 0.43506817, "global_step/max_steps": "1093/2513", "percentage": "43.49%", "elapsed_time": "3h 0m 15s", "remaining_time": "3h 54m 10s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101062} +{"loss": 0.38961393, "grad_norm": 0.03546382, "learning_rate": 6.278e-05, "epoch": 0.43546622, "global_step/max_steps": "1094/2513", "percentage": "43.53%", "elapsed_time": "3h 0m 25s", "remaining_time": "3h 54m 0s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101062} +{"loss": 0.37125725, "grad_norm": 0.03597318, "learning_rate": 6.272e-05, "epoch": 0.43586427, "global_step/max_steps": "1095/2513", "percentage": "43.57%", "elapsed_time": "3h 0m 34s", "remaining_time": "3h 53m 51s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101062} +{"loss": 0.36575615, "grad_norm": 0.03615457, "learning_rate": 6.266e-05, "epoch": 0.43626231, "global_step/max_steps": "1096/2513", "percentage": "43.61%", "elapsed_time": "3h 0m 44s", "remaining_time": "3h 53m 41s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101062} +{"loss": 0.39060766, "grad_norm": 0.0423942, "learning_rate": 6.259e-05, "epoch": 0.43666036, "global_step/max_steps": "1097/2513", "percentage": "43.65%", "elapsed_time": "3h 0m 54s", "remaining_time": "3h 53m 30s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101065} +{"loss": 0.40133461, "grad_norm": 0.0387206, "learning_rate": 6.253e-05, "epoch": 0.43705841, "global_step/max_steps": "1098/2513", "percentage": "43.69%", "elapsed_time": "3h 1m 4s", "remaining_time": "3h 53m 20s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101065} +{"loss": 0.37237364, "grad_norm": 0.03516437, "learning_rate": 6.247e-05, "epoch": 0.43745646, "global_step/max_steps": "1099/2513", "percentage": "43.73%", "elapsed_time": "3h 1m 13s", "remaining_time": "3h 53m 10s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101067} +{"loss": 0.39604819, "grad_norm": 0.03850228, "learning_rate": 6.241e-05, "epoch": 0.43785451, "global_step/max_steps": "1100/2513", "percentage": "43.77%", "elapsed_time": "3h 1m 23s", "remaining_time": "3h 53m 0s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101067} +{"loss": 0.39908373, "grad_norm": 0.04322362, "learning_rate": 6.234e-05, "epoch": 0.43825256, "global_step/max_steps": "1101/2513", "percentage": "43.81%", "elapsed_time": "3h 1m 35s", "remaining_time": "3h 52m 53s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101046} +{"loss": 0.38882154, "grad_norm": 0.0346978, "learning_rate": 6.228e-05, "epoch": 0.43865061, "global_step/max_steps": "1102/2513", "percentage": "43.85%", "elapsed_time": "3h 1m 46s", "remaining_time": "3h 52m 44s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101043} +{"loss": 0.37507421, "grad_norm": 0.03493532, "learning_rate": 6.222e-05, "epoch": 0.43904866, "global_step/max_steps": "1103/2513", "percentage": "43.89%", "elapsed_time": "3h 1m 55s", "remaining_time": "3h 52m 34s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101046} +{"loss": 0.40341347, "grad_norm": 0.03607218, "learning_rate": 6.216e-05, "epoch": 0.43944671, "global_step/max_steps": "1104/2513", "percentage": "43.93%", "elapsed_time": "3h 2m 5s", "remaining_time": "3h 52m 23s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101049} +{"loss": 0.42253938, "grad_norm": 0.03797553, "learning_rate": 6.209e-05, "epoch": 0.43984476, "global_step/max_steps": "1105/2513", "percentage": "43.97%", "elapsed_time": "3h 2m 14s", "remaining_time": "3h 52m 13s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101053} +{"loss": 0.40432471, "grad_norm": 0.0374091, "learning_rate": 6.203e-05, "epoch": 0.44024281, "global_step/max_steps": "1106/2513", "percentage": "44.01%", "elapsed_time": "3h 2m 24s", "remaining_time": "3h 52m 3s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101052} +{"loss": 0.36049569, "grad_norm": 0.03630008, "learning_rate": 6.197e-05, "epoch": 0.44064086, "global_step/max_steps": "1107/2513", "percentage": "44.05%", "elapsed_time": "3h 2m 34s", "remaining_time": "3h 51m 53s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101054} +{"loss": 0.38416547, "grad_norm": 0.03547363, "learning_rate": 6.191e-05, "epoch": 0.44103891, "global_step/max_steps": "1108/2513", "percentage": "44.09%", "elapsed_time": "3h 2m 44s", "remaining_time": "3h 51m 43s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101055} +{"loss": 0.39441198, "grad_norm": 0.03759926, "learning_rate": 6.184e-05, "epoch": 0.44143696, "global_step/max_steps": "1109/2513", "percentage": "44.13%", "elapsed_time": "3h 2m 54s", "remaining_time": "3h 51m 33s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101057} +{"loss": 0.33682716, "grad_norm": 0.03514341, "learning_rate": 6.178e-05, "epoch": 0.44183501, "global_step/max_steps": "1110/2513", "percentage": "44.17%", "elapsed_time": "3h 3m 3s", "remaining_time": "3h 51m 22s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101059} +{"loss": 0.35018164, "grad_norm": 0.03445338, "learning_rate": 6.172e-05, "epoch": 0.44223306, "global_step/max_steps": "1111/2513", "percentage": "44.21%", "elapsed_time": "3h 3m 13s", "remaining_time": "3h 51m 12s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101061} +{"loss": 0.42314303, "grad_norm": 0.03970608, "learning_rate": 6.166e-05, "epoch": 0.44263111, "global_step/max_steps": "1112/2513", "percentage": "44.25%", "elapsed_time": "3h 3m 23s", "remaining_time": "3h 51m 2s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101063} +{"loss": 0.34312087, "grad_norm": 0.03598126, "learning_rate": 6.159e-05, "epoch": 0.44302916, "global_step/max_steps": "1113/2513", "percentage": "44.29%", "elapsed_time": "3h 3m 33s", "remaining_time": "3h 50m 53s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101061} +{"loss": 0.35786617, "grad_norm": 0.0358012, "learning_rate": 6.153e-05, "epoch": 0.44342721, "global_step/max_steps": "1114/2513", "percentage": "44.33%", "elapsed_time": "3h 3m 43s", "remaining_time": "3h 50m 43s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10106} +{"loss": 0.42129296, "grad_norm": 0.04205231, "learning_rate": 6.147e-05, "epoch": 0.44382526, "global_step/max_steps": "1115/2513", "percentage": "44.37%", "elapsed_time": "3h 3m 53s", "remaining_time": "3h 50m 34s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101052} +{"loss": 0.38793576, "grad_norm": 0.03444266, "learning_rate": 6.14e-05, "epoch": 0.44422331, "global_step/max_steps": "1116/2513", "percentage": "44.41%", "elapsed_time": "3h 4m 4s", "remaining_time": "3h 50m 24s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101049} +{"loss": 0.37043685, "grad_norm": 0.03584027, "learning_rate": 6.134e-05, "epoch": 0.44462136, "global_step/max_steps": "1117/2513", "percentage": "44.45%", "elapsed_time": "3h 4m 14s", "remaining_time": "3h 50m 15s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101049} +{"loss": 0.39155704, "grad_norm": 0.03663245, "learning_rate": 6.128e-05, "epoch": 0.4450194, "global_step/max_steps": "1118/2513", "percentage": "44.49%", "elapsed_time": "3h 4m 23s", "remaining_time": "3h 50m 4s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101051} +{"loss": 0.35198519, "grad_norm": 0.04009599, "learning_rate": 6.122e-05, "epoch": 0.44541745, "global_step/max_steps": "1119/2513", "percentage": "44.53%", "elapsed_time": "3h 4m 33s", "remaining_time": "3h 49m 55s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101051} +{"loss": 0.35481364, "grad_norm": 0.03719425, "learning_rate": 6.115e-05, "epoch": 0.4458155, "global_step/max_steps": "1120/2513", "percentage": "44.57%", "elapsed_time": "3h 4m 43s", "remaining_time": "3h 49m 45s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101052} +{"loss": 0.36680609, "grad_norm": 0.03580325, "learning_rate": 6.109e-05, "epoch": 0.44621355, "global_step/max_steps": "1121/2513", "percentage": "44.61%", "elapsed_time": "3h 4m 52s", "remaining_time": "3h 49m 34s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101056} +{"loss": 0.36021712, "grad_norm": 0.03613036, "learning_rate": 6.103e-05, "epoch": 0.4466116, "global_step/max_steps": "1122/2513", "percentage": "44.65%", "elapsed_time": "3h 5m 2s", "remaining_time": "3h 49m 24s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101058} +{"loss": 0.38199908, "grad_norm": 0.03346833, "learning_rate": 6.096e-05, "epoch": 0.44700965, "global_step/max_steps": "1123/2513", "percentage": "44.69%", "elapsed_time": "3h 5m 12s", "remaining_time": "3h 49m 14s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101057} +{"loss": 0.37550539, "grad_norm": 0.03640752, "learning_rate": 6.09e-05, "epoch": 0.4474077, "global_step/max_steps": "1124/2513", "percentage": "44.73%", "elapsed_time": "3h 5m 22s", "remaining_time": "3h 49m 4s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101056} +{"loss": 0.34824577, "grad_norm": 0.03671646, "learning_rate": 6.084e-05, "epoch": 0.44780575, "global_step/max_steps": "1125/2513", "percentage": "44.77%", "elapsed_time": "3h 5m 32s", "remaining_time": "3h 48m 54s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101058} +{"loss": 0.35808405, "grad_norm": 0.03705852, "learning_rate": 6.078e-05, "epoch": 0.4482038, "global_step/max_steps": "1126/2513", "percentage": "44.81%", "elapsed_time": "3h 5m 41s", "remaining_time": "3h 48m 44s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101061} +{"loss": 0.41514286, "grad_norm": 0.04111872, "learning_rate": 6.071e-05, "epoch": 0.44860185, "global_step/max_steps": "1127/2513", "percentage": "44.85%", "elapsed_time": "3h 5m 51s", "remaining_time": "3h 48m 34s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101062} +{"loss": 0.40184826, "grad_norm": 0.04047347, "learning_rate": 6.065e-05, "epoch": 0.4489999, "global_step/max_steps": "1128/2513", "percentage": "44.89%", "elapsed_time": "3h 6m 1s", "remaining_time": "3h 48m 24s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101062} +{"loss": 0.41583133, "grad_norm": 0.03781884, "learning_rate": 6.059e-05, "epoch": 0.44939795, "global_step/max_steps": "1129/2513", "percentage": "44.93%", "elapsed_time": "3h 6m 11s", "remaining_time": "3h 48m 14s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10106} +{"loss": 0.36352566, "grad_norm": 0.03499691, "learning_rate": 6.052e-05, "epoch": 0.449796, "global_step/max_steps": "1130/2513", "percentage": "44.97%", "elapsed_time": "3h 6m 21s", "remaining_time": "3h 48m 5s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101055} +{"loss": 0.37551302, "grad_norm": 0.03638792, "learning_rate": 6.046e-05, "epoch": 0.45019405, "global_step/max_steps": "1131/2513", "percentage": "45.01%", "elapsed_time": "3h 6m 31s", "remaining_time": "3h 47m 55s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101056} +{"loss": 0.3957026, "grad_norm": 0.03906949, "learning_rate": 6.04e-05, "epoch": 0.4505921, "global_step/max_steps": "1132/2513", "percentage": "45.05%", "elapsed_time": "3h 6m 41s", "remaining_time": "3h 47m 45s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101056} +{"loss": 0.39556828, "grad_norm": 0.03620204, "learning_rate": 6.033e-05, "epoch": 0.45099015, "global_step/max_steps": "1133/2513", "percentage": "45.09%", "elapsed_time": "3h 6m 51s", "remaining_time": "3h 47m 35s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101059} +{"loss": 0.41138157, "grad_norm": 0.03742876, "learning_rate": 6.027e-05, "epoch": 0.4513882, "global_step/max_steps": "1134/2513", "percentage": "45.13%", "elapsed_time": "3h 7m 1s", "remaining_time": "3h 47m 25s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10106} +{"loss": 0.39060813, "grad_norm": 0.0390697, "learning_rate": 6.021e-05, "epoch": 0.45178625, "global_step/max_steps": "1135/2513", "percentage": "45.17%", "elapsed_time": "3h 7m 10s", "remaining_time": "3h 47m 15s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101063} +{"loss": 0.4037551, "grad_norm": 0.04188641, "learning_rate": 6.015e-05, "epoch": 0.4521843, "global_step/max_steps": "1136/2513", "percentage": "45.20%", "elapsed_time": "3h 7m 20s", "remaining_time": "3h 47m 4s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101065} +{"loss": 0.38211691, "grad_norm": 0.04009415, "learning_rate": 6.008e-05, "epoch": 0.45258235, "global_step/max_steps": "1137/2513", "percentage": "45.24%", "elapsed_time": "3h 7m 29s", "remaining_time": "3h 46m 54s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101067} +{"loss": 0.38647017, "grad_norm": 0.03627751, "learning_rate": 6.002e-05, "epoch": 0.4529804, "global_step/max_steps": "1138/2513", "percentage": "45.28%", "elapsed_time": "3h 7m 39s", "remaining_time": "3h 46m 44s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10107} +{"loss": 0.40443397, "grad_norm": 0.03754392, "learning_rate": 5.996e-05, "epoch": 0.45337845, "global_step/max_steps": "1139/2513", "percentage": "45.32%", "elapsed_time": "3h 7m 49s", "remaining_time": "3h 46m 34s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101071} +{"loss": 0.39402047, "grad_norm": 0.03487633, "learning_rate": 5.989e-05, "epoch": 0.4537765, "global_step/max_steps": "1140/2513", "percentage": "45.36%", "elapsed_time": "3h 7m 58s", "remaining_time": "3h 46m 24s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101073} +{"loss": 0.354581, "grad_norm": 0.03659962, "learning_rate": 5.983e-05, "epoch": 0.45417454, "global_step/max_steps": "1141/2513", "percentage": "45.40%", "elapsed_time": "3h 8m 8s", "remaining_time": "3h 46m 14s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101074} +{"loss": 0.37728304, "grad_norm": 0.03315014, "learning_rate": 5.977e-05, "epoch": 0.45457259, "global_step/max_steps": "1142/2513", "percentage": "45.44%", "elapsed_time": "3h 8m 18s", "remaining_time": "3h 46m 4s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101074} +{"loss": 0.36136186, "grad_norm": 0.03818796, "learning_rate": 5.97e-05, "epoch": 0.45497064, "global_step/max_steps": "1143/2513", "percentage": "45.48%", "elapsed_time": "3h 8m 28s", "remaining_time": "3h 45m 54s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101075} +{"loss": 0.37427959, "grad_norm": 0.03588859, "learning_rate": 5.964e-05, "epoch": 0.45536869, "global_step/max_steps": "1144/2513", "percentage": "45.52%", "elapsed_time": "3h 8m 38s", "remaining_time": "3h 45m 44s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101075} +{"loss": 0.38884956, "grad_norm": 0.03632775, "learning_rate": 5.958e-05, "epoch": 0.45576674, "global_step/max_steps": "1145/2513", "percentage": "45.56%", "elapsed_time": "3h 8m 49s", "remaining_time": "3h 45m 35s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101068} +{"loss": 0.37677479, "grad_norm": 0.03488244, "learning_rate": 5.951e-05, "epoch": 0.45616479, "global_step/max_steps": "1146/2513", "percentage": "45.60%", "elapsed_time": "3h 8m 58s", "remaining_time": "3h 45m 25s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101071} +{"loss": 0.38016599, "grad_norm": 0.03772219, "learning_rate": 5.945e-05, "epoch": 0.45656284, "global_step/max_steps": "1147/2513", "percentage": "45.64%", "elapsed_time": "3h 9m 8s", "remaining_time": "3h 45m 15s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101069} +{"loss": 0.32646745, "grad_norm": 0.03444507, "learning_rate": 5.939e-05, "epoch": 0.45696089, "global_step/max_steps": "1148/2513", "percentage": "45.68%", "elapsed_time": "3h 9m 18s", "remaining_time": "3h 45m 4s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101074} +{"loss": 0.36697328, "grad_norm": 0.03609757, "learning_rate": 5.932e-05, "epoch": 0.45735894, "global_step/max_steps": "1149/2513", "percentage": "45.72%", "elapsed_time": "3h 9m 27s", "remaining_time": "3h 44m 54s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101076} +{"loss": 0.38726741, "grad_norm": 0.0396512, "learning_rate": 5.926e-05, "epoch": 0.45775699, "global_step/max_steps": "1150/2513", "percentage": "45.76%", "elapsed_time": "3h 9m 37s", "remaining_time": "3h 44m 44s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10108} +{"loss": 0.39077583, "grad_norm": 0.03950308, "learning_rate": 5.92e-05, "epoch": 0.45815504, "global_step/max_steps": "1151/2513", "percentage": "45.80%", "elapsed_time": "3h 9m 46s", "remaining_time": "3h 44m 34s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101083} +{"loss": 0.41659963, "grad_norm": 0.03989048, "learning_rate": 5.913e-05, "epoch": 0.45855309, "global_step/max_steps": "1152/2513", "percentage": "45.84%", "elapsed_time": "3h 9m 56s", "remaining_time": "3h 44m 24s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101084} +{"loss": 0.37665009, "grad_norm": 0.03775197, "learning_rate": 5.907e-05, "epoch": 0.45895114, "global_step/max_steps": "1153/2513", "percentage": "45.88%", "elapsed_time": "3h 10m 6s", "remaining_time": "3h 44m 14s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101085} +{"loss": 0.37049919, "grad_norm": 0.03410515, "learning_rate": 5.901e-05, "epoch": 0.45934919, "global_step/max_steps": "1154/2513", "percentage": "45.92%", "elapsed_time": "3h 10m 16s", "remaining_time": "3h 44m 4s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101086} +{"loss": 0.3637377, "grad_norm": 0.03806553, "learning_rate": 5.894e-05, "epoch": 0.45974724, "global_step/max_steps": "1155/2513", "percentage": "45.96%", "elapsed_time": "3h 10m 25s", "remaining_time": "3h 43m 53s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101089} +{"loss": 0.33799082, "grad_norm": 0.03919118, "learning_rate": 5.888e-05, "epoch": 0.46014529, "global_step/max_steps": "1156/2513", "percentage": "46.00%", "elapsed_time": "3h 10m 35s", "remaining_time": "3h 43m 43s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101091} +{"loss": 0.35624778, "grad_norm": 0.0342521, "learning_rate": 5.882e-05, "epoch": 0.46054334, "global_step/max_steps": "1157/2513", "percentage": "46.04%", "elapsed_time": "3h 10m 45s", "remaining_time": "3h 43m 33s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10109} +{"loss": 0.37567577, "grad_norm": 0.03558575, "learning_rate": 5.875e-05, "epoch": 0.46094139, "global_step/max_steps": "1158/2513", "percentage": "46.08%", "elapsed_time": "3h 10m 55s", "remaining_time": "3h 43m 23s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10109} +{"loss": 0.33755851, "grad_norm": 0.03679212, "learning_rate": 5.869e-05, "epoch": 0.46133944, "global_step/max_steps": "1159/2513", "percentage": "46.12%", "elapsed_time": "3h 11m 5s", "remaining_time": "3h 43m 14s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101086} +{"loss": 0.31002015, "grad_norm": 0.03512711, "learning_rate": 5.863e-05, "epoch": 0.46173749, "global_step/max_steps": "1160/2513", "percentage": "46.16%", "elapsed_time": "3h 11m 16s", "remaining_time": "3h 43m 5s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.10108} +{"loss": 0.38598204, "grad_norm": 0.03916606, "learning_rate": 5.856e-05, "epoch": 0.46213554, "global_step/max_steps": "1161/2513", "percentage": "46.20%", "elapsed_time": "3h 11m 26s", "remaining_time": "3h 42m 55s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101079} +{"loss": 0.36039001, "grad_norm": 0.03672372, "learning_rate": 5.85e-05, "epoch": 0.46253359, "global_step/max_steps": "1162/2513", "percentage": "46.24%", "elapsed_time": "3h 11m 35s", "remaining_time": "3h 42m 45s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101083} +{"loss": 0.36806807, "grad_norm": 0.03857403, "learning_rate": 5.844e-05, "epoch": 0.46293163, "global_step/max_steps": "1163/2513", "percentage": "46.28%", "elapsed_time": "3h 11m 45s", "remaining_time": "3h 42m 35s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101081} +{"loss": 0.33681911, "grad_norm": 0.03692723, "learning_rate": 5.837e-05, "epoch": 0.46332968, "global_step/max_steps": "1164/2513", "percentage": "46.32%", "elapsed_time": "3h 11m 56s", "remaining_time": "3h 42m 26s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101072} +{"loss": 0.39577252, "grad_norm": 0.03624738, "learning_rate": 5.831e-05, "epoch": 0.46372773, "global_step/max_steps": "1165/2513", "percentage": "46.36%", "elapsed_time": "3h 12m 6s", "remaining_time": "3h 42m 16s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101075} +{"loss": 0.35047776, "grad_norm": 0.03664861, "learning_rate": 5.824e-05, "epoch": 0.46412578, "global_step/max_steps": "1166/2513", "percentage": "46.40%", "elapsed_time": "3h 12m 16s", "remaining_time": "3h 42m 6s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101075} +{"loss": 0.3636395, "grad_norm": 0.03496392, "learning_rate": 5.818e-05, "epoch": 0.46452383, "global_step/max_steps": "1167/2513", "percentage": "46.44%", "elapsed_time": "3h 12m 25s", "remaining_time": "3h 41m 56s", "memory(GiB)": 77.61, "train_speed(iter/s)": 0.101078} +{"loss": 0.40208444, "grad_norm": 0.03934608, "learning_rate": 5.812e-05, "epoch": 0.46492188, "global_step/max_steps": "1168/2513", "percentage": "46.48%", "elapsed_time": "3h 12m 36s", "remaining_time": "3h 41m 47s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101071} +{"loss": 0.39269075, "grad_norm": 0.04071297, "learning_rate": 5.805e-05, "epoch": 0.46531993, "global_step/max_steps": "1169/2513", "percentage": "46.52%", "elapsed_time": "3h 12m 45s", "remaining_time": "3h 41m 37s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101074} +{"loss": 0.35051721, "grad_norm": 0.0360652, "learning_rate": 5.799e-05, "epoch": 0.46571798, "global_step/max_steps": "1170/2513", "percentage": "46.56%", "elapsed_time": "3h 12m 55s", "remaining_time": "3h 41m 26s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101077} +{"loss": 0.29757929, "grad_norm": 0.03601246, "learning_rate": 5.793e-05, "epoch": 0.46611603, "global_step/max_steps": "1171/2513", "percentage": "46.60%", "elapsed_time": "3h 13m 4s", "remaining_time": "3h 41m 16s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101081} +{"loss": 0.39720482, "grad_norm": 0.03646464, "learning_rate": 5.786e-05, "epoch": 0.46651408, "global_step/max_steps": "1172/2513", "percentage": "46.64%", "elapsed_time": "3h 13m 14s", "remaining_time": "3h 41m 6s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101081} +{"loss": 0.38668117, "grad_norm": 0.03678505, "learning_rate": 5.78e-05, "epoch": 0.46691213, "global_step/max_steps": "1173/2513", "percentage": "46.68%", "elapsed_time": "3h 13m 24s", "remaining_time": "3h 40m 56s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101083} +{"loss": 0.37609956, "grad_norm": 0.03761296, "learning_rate": 5.774e-05, "epoch": 0.46731018, "global_step/max_steps": "1174/2513", "percentage": "46.72%", "elapsed_time": "3h 13m 34s", "remaining_time": "3h 40m 46s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10108} +{"loss": 0.35901994, "grad_norm": 0.03865211, "learning_rate": 5.767e-05, "epoch": 0.46770823, "global_step/max_steps": "1175/2513", "percentage": "46.76%", "elapsed_time": "3h 13m 44s", "remaining_time": "3h 40m 37s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101078} +{"loss": 0.40018871, "grad_norm": 0.03790101, "learning_rate": 5.761e-05, "epoch": 0.46810628, "global_step/max_steps": "1176/2513", "percentage": "46.80%", "elapsed_time": "3h 13m 54s", "remaining_time": "3h 40m 27s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101077} +{"loss": 0.41470829, "grad_norm": 0.0378965, "learning_rate": 5.754e-05, "epoch": 0.46850433, "global_step/max_steps": "1177/2513", "percentage": "46.84%", "elapsed_time": "3h 14m 4s", "remaining_time": "3h 40m 17s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101077} +{"loss": 0.34036025, "grad_norm": 0.03446351, "learning_rate": 5.748e-05, "epoch": 0.46890238, "global_step/max_steps": "1178/2513", "percentage": "46.88%", "elapsed_time": "3h 14m 14s", "remaining_time": "3h 40m 7s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101078} +{"loss": 0.38730511, "grad_norm": 0.0393194, "learning_rate": 5.742e-05, "epoch": 0.46930043, "global_step/max_steps": "1179/2513", "percentage": "46.92%", "elapsed_time": "3h 14m 24s", "remaining_time": "3h 39m 58s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101075} +{"loss": 0.37844628, "grad_norm": 0.03770138, "learning_rate": 5.735e-05, "epoch": 0.46969848, "global_step/max_steps": "1180/2513", "percentage": "46.96%", "elapsed_time": "3h 14m 34s", "remaining_time": "3h 39m 48s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101076} +{"loss": 0.32191223, "grad_norm": 0.03997274, "learning_rate": 5.729e-05, "epoch": 0.47009653, "global_step/max_steps": "1181/2513", "percentage": "47.00%", "elapsed_time": "3h 14m 44s", "remaining_time": "3h 39m 38s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101076} +{"loss": 0.34468067, "grad_norm": 0.03860892, "learning_rate": 5.723e-05, "epoch": 0.47049458, "global_step/max_steps": "1182/2513", "percentage": "47.04%", "elapsed_time": "3h 14m 53s", "remaining_time": "3h 39m 28s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101078} +{"loss": 0.39943171, "grad_norm": 0.03755921, "learning_rate": 5.716e-05, "epoch": 0.47089263, "global_step/max_steps": "1183/2513", "percentage": "47.08%", "elapsed_time": "3h 15m 3s", "remaining_time": "3h 39m 18s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101079} +{"loss": 0.33117825, "grad_norm": 0.03419051, "learning_rate": 5.71e-05, "epoch": 0.47129068, "global_step/max_steps": "1184/2513", "percentage": "47.12%", "elapsed_time": "3h 15m 13s", "remaining_time": "3h 39m 7s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101081} +{"loss": 0.3891235, "grad_norm": 0.04114524, "learning_rate": 5.703e-05, "epoch": 0.47168873, "global_step/max_steps": "1185/2513", "percentage": "47.15%", "elapsed_time": "3h 15m 23s", "remaining_time": "3h 38m 57s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101082} +{"loss": 0.34660047, "grad_norm": 0.03748944, "learning_rate": 5.697e-05, "epoch": 0.47208677, "global_step/max_steps": "1186/2513", "percentage": "47.19%", "elapsed_time": "3h 15m 33s", "remaining_time": "3h 38m 48s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101082} +{"loss": 0.35886666, "grad_norm": 0.03697407, "learning_rate": 5.691e-05, "epoch": 0.47248482, "global_step/max_steps": "1187/2513", "percentage": "47.23%", "elapsed_time": "3h 15m 43s", "remaining_time": "3h 38m 38s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10108} +{"loss": 0.3751004, "grad_norm": 0.03679509, "learning_rate": 5.684e-05, "epoch": 0.47288287, "global_step/max_steps": "1188/2513", "percentage": "47.27%", "elapsed_time": "3h 15m 53s", "remaining_time": "3h 38m 28s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101077} +{"loss": 0.40294698, "grad_norm": 0.04071149, "learning_rate": 5.678e-05, "epoch": 0.47328092, "global_step/max_steps": "1189/2513", "percentage": "47.31%", "elapsed_time": "3h 16m 4s", "remaining_time": "3h 38m 20s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101068} +{"loss": 0.41659129, "grad_norm": 0.03904453, "learning_rate": 5.672e-05, "epoch": 0.47367897, "global_step/max_steps": "1190/2513", "percentage": "47.35%", "elapsed_time": "3h 16m 15s", "remaining_time": "3h 38m 11s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101058} +{"loss": 0.3819834, "grad_norm": 0.03918282, "learning_rate": 5.665e-05, "epoch": 0.47407702, "global_step/max_steps": "1191/2513", "percentage": "47.39%", "elapsed_time": "3h 16m 25s", "remaining_time": "3h 38m 1s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101057} +{"loss": 0.44376856, "grad_norm": 0.0367634, "learning_rate": 5.659e-05, "epoch": 0.47447507, "global_step/max_steps": "1192/2513", "percentage": "47.43%", "elapsed_time": "3h 16m 35s", "remaining_time": "3h 37m 51s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101057} +{"loss": 0.33498031, "grad_norm": 0.03511262, "learning_rate": 5.652e-05, "epoch": 0.47487312, "global_step/max_steps": "1193/2513", "percentage": "47.47%", "elapsed_time": "3h 16m 45s", "remaining_time": "3h 37m 41s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101057} +{"loss": 0.33646488, "grad_norm": 0.03619285, "learning_rate": 5.646e-05, "epoch": 0.47527117, "global_step/max_steps": "1194/2513", "percentage": "47.51%", "elapsed_time": "3h 16m 54s", "remaining_time": "3h 37m 31s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101059} +{"loss": 0.39649653, "grad_norm": 0.038946, "learning_rate": 5.64e-05, "epoch": 0.47566922, "global_step/max_steps": "1195/2513", "percentage": "47.55%", "elapsed_time": "3h 17m 4s", "remaining_time": "3h 37m 21s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101059} +{"loss": 0.39574903, "grad_norm": 0.0352308, "learning_rate": 5.633e-05, "epoch": 0.47606727, "global_step/max_steps": "1196/2513", "percentage": "47.59%", "elapsed_time": "3h 17m 14s", "remaining_time": "3h 37m 11s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10106} +{"loss": 0.32222331, "grad_norm": 0.03354615, "learning_rate": 5.627e-05, "epoch": 0.47646532, "global_step/max_steps": "1197/2513", "percentage": "47.63%", "elapsed_time": "3h 17m 24s", "remaining_time": "3h 37m 1s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101061} +{"loss": 0.39282608, "grad_norm": 0.03742928, "learning_rate": 5.62e-05, "epoch": 0.47686337, "global_step/max_steps": "1198/2513", "percentage": "47.67%", "elapsed_time": "3h 17m 34s", "remaining_time": "3h 36m 51s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101062} +{"loss": 0.3769995, "grad_norm": 0.03668065, "learning_rate": 5.614e-05, "epoch": 0.47726142, "global_step/max_steps": "1199/2513", "percentage": "47.71%", "elapsed_time": "3h 17m 44s", "remaining_time": "3h 36m 42s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101061} +{"loss": 0.37359971, "grad_norm": 0.0359235, "learning_rate": 5.608e-05, "epoch": 0.47765947, "global_step/max_steps": "1200/2513", "percentage": "47.75%", "elapsed_time": "3h 17m 54s", "remaining_time": "3h 36m 32s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101058} +{"loss": 0.35367182, "grad_norm": 0.03754924, "learning_rate": 5.601e-05, "epoch": 0.47805752, "global_step/max_steps": "1201/2513", "percentage": "47.79%", "elapsed_time": "3h 18m 5s", "remaining_time": "3h 36m 24s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101045} +{"loss": 0.3567031, "grad_norm": 0.03625955, "learning_rate": 5.595e-05, "epoch": 0.47845557, "global_step/max_steps": "1202/2513", "percentage": "47.83%", "elapsed_time": "3h 18m 15s", "remaining_time": "3h 36m 14s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101044} +{"loss": 0.36814481, "grad_norm": 0.03444956, "learning_rate": 5.588e-05, "epoch": 0.47885362, "global_step/max_steps": "1203/2513", "percentage": "47.87%", "elapsed_time": "3h 18m 26s", "remaining_time": "3h 36m 5s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101038} +{"loss": 0.36888465, "grad_norm": 0.03636791, "learning_rate": 5.582e-05, "epoch": 0.47925167, "global_step/max_steps": "1204/2513", "percentage": "47.91%", "elapsed_time": "3h 18m 36s", "remaining_time": "3h 35m 55s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101035} +{"loss": 0.38934633, "grad_norm": 0.03847006, "learning_rate": 5.576e-05, "epoch": 0.47964972, "global_step/max_steps": "1205/2513", "percentage": "47.95%", "elapsed_time": "3h 18m 47s", "remaining_time": "3h 35m 46s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101029} +{"loss": 0.3635394, "grad_norm": 0.03649595, "learning_rate": 5.569e-05, "epoch": 0.48004777, "global_step/max_steps": "1206/2513", "percentage": "47.99%", "elapsed_time": "3h 18m 57s", "remaining_time": "3h 35m 36s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101028} +{"loss": 0.37135464, "grad_norm": 0.03677517, "learning_rate": 5.563e-05, "epoch": 0.48044582, "global_step/max_steps": "1207/2513", "percentage": "48.03%", "elapsed_time": "3h 19m 7s", "remaining_time": "3h 35m 27s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101027} +{"loss": 0.38271117, "grad_norm": 0.03912143, "learning_rate": 5.556e-05, "epoch": 0.48084387, "global_step/max_steps": "1208/2513", "percentage": "48.07%", "elapsed_time": "3h 19m 16s", "remaining_time": "3h 35m 17s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101029} +{"loss": 0.3473011, "grad_norm": 0.03645533, "learning_rate": 5.55e-05, "epoch": 0.48124191, "global_step/max_steps": "1209/2513", "percentage": "48.11%", "elapsed_time": "3h 19m 26s", "remaining_time": "3h 35m 6s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101031} +{"loss": 0.38035986, "grad_norm": 0.04479135, "learning_rate": 5.544e-05, "epoch": 0.48163996, "global_step/max_steps": "1210/2513", "percentage": "48.15%", "elapsed_time": "3h 19m 36s", "remaining_time": "3h 34m 56s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101032} +{"loss": 0.34143922, "grad_norm": 0.03792173, "learning_rate": 5.537e-05, "epoch": 0.48203801, "global_step/max_steps": "1211/2513", "percentage": "48.19%", "elapsed_time": "3h 19m 46s", "remaining_time": "3h 34m 47s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101031} +{"loss": 0.37220907, "grad_norm": 0.03521024, "learning_rate": 5.531e-05, "epoch": 0.48243606, "global_step/max_steps": "1212/2513", "percentage": "48.23%", "elapsed_time": "3h 19m 56s", "remaining_time": "3h 34m 37s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101032} +{"loss": 0.3634944, "grad_norm": 0.03976798, "learning_rate": 5.524e-05, "epoch": 0.48283411, "global_step/max_steps": "1213/2513", "percentage": "48.27%", "elapsed_time": "3h 20m 6s", "remaining_time": "3h 34m 27s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101031} +{"loss": 0.3623094, "grad_norm": 0.03785743, "learning_rate": 5.518e-05, "epoch": 0.48323216, "global_step/max_steps": "1214/2513", "percentage": "48.31%", "elapsed_time": "3h 20m 15s", "remaining_time": "3h 34m 17s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101033} +{"loss": 0.36577442, "grad_norm": 0.03426451, "learning_rate": 5.512e-05, "epoch": 0.48363021, "global_step/max_steps": "1215/2513", "percentage": "48.35%", "elapsed_time": "3h 20m 25s", "remaining_time": "3h 34m 7s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101033} +{"loss": 0.35658857, "grad_norm": 0.03872727, "learning_rate": 5.505e-05, "epoch": 0.48402826, "global_step/max_steps": "1216/2513", "percentage": "48.39%", "elapsed_time": "3h 20m 36s", "remaining_time": "3h 33m 58s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101028} +{"loss": 0.3797532, "grad_norm": 0.0363913, "learning_rate": 5.499e-05, "epoch": 0.48442631, "global_step/max_steps": "1217/2513", "percentage": "48.43%", "elapsed_time": "3h 20m 46s", "remaining_time": "3h 33m 48s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101025} +{"loss": 0.37077758, "grad_norm": 0.04081394, "learning_rate": 5.492e-05, "epoch": 0.48482436, "global_step/max_steps": "1218/2513", "percentage": "48.47%", "elapsed_time": "3h 20m 57s", "remaining_time": "3h 33m 39s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10102} +{"loss": 0.32786605, "grad_norm": 0.03829842, "learning_rate": 5.486e-05, "epoch": 0.48522241, "global_step/max_steps": "1219/2513", "percentage": "48.51%", "elapsed_time": "3h 21m 7s", "remaining_time": "3h 33m 29s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101017} +{"loss": 0.37261289, "grad_norm": 0.0370207, "learning_rate": 5.479e-05, "epoch": 0.48562046, "global_step/max_steps": "1220/2513", "percentage": "48.55%", "elapsed_time": "3h 21m 17s", "remaining_time": "3h 33m 20s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101015} +{"loss": 0.34285596, "grad_norm": 0.03510558, "learning_rate": 5.473e-05, "epoch": 0.48601851, "global_step/max_steps": "1221/2513", "percentage": "48.59%", "elapsed_time": "3h 21m 27s", "remaining_time": "3h 33m 9s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101017} +{"loss": 0.36556143, "grad_norm": 0.03751659, "learning_rate": 5.467e-05, "epoch": 0.48641656, "global_step/max_steps": "1222/2513", "percentage": "48.63%", "elapsed_time": "3h 21m 36s", "remaining_time": "3h 32m 59s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101017} +{"loss": 0.40216714, "grad_norm": 0.04110838, "learning_rate": 5.46e-05, "epoch": 0.48681461, "global_step/max_steps": "1223/2513", "percentage": "48.67%", "elapsed_time": "3h 21m 46s", "remaining_time": "3h 32m 50s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101016} +{"loss": 0.35844371, "grad_norm": 0.03502567, "learning_rate": 5.454e-05, "epoch": 0.48721266, "global_step/max_steps": "1224/2513", "percentage": "48.71%", "elapsed_time": "3h 21m 57s", "remaining_time": "3h 32m 40s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101014} +{"loss": 0.4160603, "grad_norm": 0.03805001, "learning_rate": 5.447e-05, "epoch": 0.48761071, "global_step/max_steps": "1225/2513", "percentage": "48.75%", "elapsed_time": "3h 22m 6s", "remaining_time": "3h 32m 30s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101015} +{"loss": 0.38591427, "grad_norm": 0.03870676, "learning_rate": 5.441e-05, "epoch": 0.48800876, "global_step/max_steps": "1226/2513", "percentage": "48.79%", "elapsed_time": "3h 22m 16s", "remaining_time": "3h 32m 20s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101017} +{"loss": 0.42594597, "grad_norm": 0.03865747, "learning_rate": 5.435e-05, "epoch": 0.48840681, "global_step/max_steps": "1227/2513", "percentage": "48.83%", "elapsed_time": "3h 22m 26s", "remaining_time": "3h 32m 10s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101019} +{"loss": 0.32996798, "grad_norm": 0.03664633, "learning_rate": 5.428e-05, "epoch": 0.48880486, "global_step/max_steps": "1228/2513", "percentage": "48.87%", "elapsed_time": "3h 22m 36s", "remaining_time": "3h 32m 0s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101019} +{"loss": 0.40491444, "grad_norm": 0.03749396, "learning_rate": 5.422e-05, "epoch": 0.48920291, "global_step/max_steps": "1229/2513", "percentage": "48.91%", "elapsed_time": "3h 22m 45s", "remaining_time": "3h 31m 50s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101021} +{"loss": 0.36731169, "grad_norm": 0.03696006, "learning_rate": 5.415e-05, "epoch": 0.48960096, "global_step/max_steps": "1230/2513", "percentage": "48.95%", "elapsed_time": "3h 22m 55s", "remaining_time": "3h 31m 39s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101024} +{"loss": 0.40206733, "grad_norm": 0.03798194, "learning_rate": 5.409e-05, "epoch": 0.489999, "global_step/max_steps": "1231/2513", "percentage": "48.99%", "elapsed_time": "3h 23m 5s", "remaining_time": "3h 31m 30s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101022} +{"loss": 0.36567724, "grad_norm": 0.03723979, "learning_rate": 5.402e-05, "epoch": 0.49039705, "global_step/max_steps": "1232/2513", "percentage": "49.03%", "elapsed_time": "3h 23m 16s", "remaining_time": "3h 31m 21s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101015} +{"loss": 0.36912525, "grad_norm": 0.03981821, "learning_rate": 5.396e-05, "epoch": 0.4907951, "global_step/max_steps": "1233/2513", "percentage": "49.06%", "elapsed_time": "3h 23m 26s", "remaining_time": "3h 31m 12s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101009} +{"loss": 0.38233459, "grad_norm": 0.03772346, "learning_rate": 5.39e-05, "epoch": 0.49119315, "global_step/max_steps": "1234/2513", "percentage": "49.10%", "elapsed_time": "3h 23m 37s", "remaining_time": "3h 31m 2s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101003} +{"loss": 0.34432328, "grad_norm": 0.03540582, "learning_rate": 5.383e-05, "epoch": 0.4915912, "global_step/max_steps": "1235/2513", "percentage": "49.14%", "elapsed_time": "3h 23m 47s", "remaining_time": "3h 30m 53s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101001} +{"loss": 0.33088958, "grad_norm": 0.03398752, "learning_rate": 5.377e-05, "epoch": 0.49198925, "global_step/max_steps": "1236/2513", "percentage": "49.18%", "elapsed_time": "3h 23m 57s", "remaining_time": "3h 30m 43s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101001} +{"loss": 0.36710811, "grad_norm": 0.0346679, "learning_rate": 5.37e-05, "epoch": 0.4923873, "global_step/max_steps": "1237/2513", "percentage": "49.22%", "elapsed_time": "3h 24m 6s", "remaining_time": "3h 30m 33s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101005} +{"loss": 0.37891957, "grad_norm": 0.03652882, "learning_rate": 5.364e-05, "epoch": 0.49278535, "global_step/max_steps": "1238/2513", "percentage": "49.26%", "elapsed_time": "3h 24m 16s", "remaining_time": "3h 30m 23s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101004} +{"loss": 0.3727563, "grad_norm": 0.03489281, "learning_rate": 5.357e-05, "epoch": 0.4931834, "global_step/max_steps": "1239/2513", "percentage": "49.30%", "elapsed_time": "3h 24m 26s", "remaining_time": "3h 30m 13s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101004} +{"loss": 0.38735795, "grad_norm": 0.03834851, "learning_rate": 5.351e-05, "epoch": 0.49358145, "global_step/max_steps": "1240/2513", "percentage": "49.34%", "elapsed_time": "3h 24m 36s", "remaining_time": "3h 30m 3s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101005} +{"loss": 0.37230498, "grad_norm": 0.03734789, "learning_rate": 5.345e-05, "epoch": 0.4939795, "global_step/max_steps": "1241/2513", "percentage": "49.38%", "elapsed_time": "3h 24m 46s", "remaining_time": "3h 29m 53s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101007} +{"loss": 0.38357788, "grad_norm": 0.0383225, "learning_rate": 5.338e-05, "epoch": 0.49437755, "global_step/max_steps": "1242/2513", "percentage": "49.42%", "elapsed_time": "3h 24m 56s", "remaining_time": "3h 29m 43s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101008} +{"loss": 0.35744685, "grad_norm": 0.03831731, "learning_rate": 5.332e-05, "epoch": 0.4947756, "global_step/max_steps": "1243/2513", "percentage": "49.46%", "elapsed_time": "3h 25m 5s", "remaining_time": "3h 29m 33s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101009} +{"loss": 0.38007569, "grad_norm": 0.03610678, "learning_rate": 5.325e-05, "epoch": 0.49517365, "global_step/max_steps": "1244/2513", "percentage": "49.50%", "elapsed_time": "3h 25m 15s", "remaining_time": "3h 29m 23s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10101} +{"loss": 0.39526927, "grad_norm": 0.03919041, "learning_rate": 5.319e-05, "epoch": 0.4955717, "global_step/max_steps": "1245/2513", "percentage": "49.54%", "elapsed_time": "3h 25m 25s", "remaining_time": "3h 29m 13s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10101} +{"loss": 0.3889434, "grad_norm": 0.03830917, "learning_rate": 5.312e-05, "epoch": 0.49596975, "global_step/max_steps": "1246/2513", "percentage": "49.58%", "elapsed_time": "3h 25m 35s", "remaining_time": "3h 29m 3s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101011} +{"loss": 0.43630469, "grad_norm": 0.03773646, "learning_rate": 5.306e-05, "epoch": 0.4963678, "global_step/max_steps": "1247/2513", "percentage": "49.62%", "elapsed_time": "3h 25m 45s", "remaining_time": "3h 28m 53s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101007} +{"loss": 0.34351647, "grad_norm": 0.03209431, "learning_rate": 5.3e-05, "epoch": 0.49676585, "global_step/max_steps": "1248/2513", "percentage": "49.66%", "elapsed_time": "3h 25m 56s", "remaining_time": "3h 28m 44s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101003} +{"loss": 0.3725895, "grad_norm": 0.03631428, "learning_rate": 5.293e-05, "epoch": 0.4971639, "global_step/max_steps": "1249/2513", "percentage": "49.70%", "elapsed_time": "3h 26m 6s", "remaining_time": "3h 28m 35s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100997} +{"loss": 0.41162264, "grad_norm": 0.03778494, "learning_rate": 5.287e-05, "epoch": 0.49756195, "global_step/max_steps": "1250/2513", "percentage": "49.74%", "elapsed_time": "3h 26m 16s", "remaining_time": "3h 28m 24s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101002} +{"loss": 0.39077091, "grad_norm": 0.0368973, "learning_rate": 5.28e-05, "epoch": 0.49796, "global_step/max_steps": "1251/2513", "percentage": "49.78%", "elapsed_time": "3h 26m 26s", "remaining_time": "3h 28m 14s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101} +{"loss": 0.35946283, "grad_norm": 0.03895072, "learning_rate": 5.274e-05, "epoch": 0.49835805, "global_step/max_steps": "1252/2513", "percentage": "49.82%", "elapsed_time": "3h 26m 35s", "remaining_time": "3h 28m 4s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101002} +{"loss": 0.38457608, "grad_norm": 0.03820881, "learning_rate": 5.267e-05, "epoch": 0.4987561, "global_step/max_steps": "1253/2513", "percentage": "49.86%", "elapsed_time": "3h 26m 46s", "remaining_time": "3h 27m 55s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100999} +{"loss": 0.4018513, "grad_norm": 0.04070884, "learning_rate": 5.261e-05, "epoch": 0.49915414, "global_step/max_steps": "1254/2513", "percentage": "49.90%", "elapsed_time": "3h 26m 55s", "remaining_time": "3h 27m 45s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101001} +{"loss": 0.39805907, "grad_norm": 0.03874368, "learning_rate": 5.254e-05, "epoch": 0.49955219, "global_step/max_steps": "1255/2513", "percentage": "49.94%", "elapsed_time": "3h 27m 5s", "remaining_time": "3h 27m 35s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101001} +{"loss": 0.3691985, "grad_norm": 0.03875529, "learning_rate": 5.248e-05, "epoch": 0.49995024, "global_step/max_steps": "1256/2513", "percentage": "49.98%", "elapsed_time": "3h 27m 15s", "remaining_time": "3h 27m 25s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101} +{"loss": 0.39525294, "grad_norm": 0.03712848, "learning_rate": 5.242e-05, "epoch": 0.50034829, "global_step/max_steps": "1257/2513", "percentage": "50.02%", "elapsed_time": "3h 27m 25s", "remaining_time": "3h 27m 15s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100999} +{"loss": 0.39156938, "grad_norm": 0.03643491, "learning_rate": 5.235e-05, "epoch": 0.50074634, "global_step/max_steps": "1258/2513", "percentage": "50.06%", "elapsed_time": "3h 27m 35s", "remaining_time": "3h 27m 5s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101001} +{"loss": 0.33865926, "grad_norm": 0.0406871, "learning_rate": 5.229e-05, "epoch": 0.50114439, "global_step/max_steps": "1259/2513", "percentage": "50.10%", "elapsed_time": "3h 27m 45s", "remaining_time": "3h 26m 55s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101} +{"loss": 0.4151684, "grad_norm": 0.03763057, "learning_rate": 5.222e-05, "epoch": 0.50154244, "global_step/max_steps": "1260/2513", "percentage": "50.14%", "elapsed_time": "3h 27m 55s", "remaining_time": "3h 26m 46s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100995} +{"loss": 0.35214472, "grad_norm": 0.036032, "learning_rate": 5.216e-05, "epoch": 0.50194049, "global_step/max_steps": "1261/2513", "percentage": "50.18%", "elapsed_time": "3h 28m 5s", "remaining_time": "3h 26m 36s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100996} +{"loss": 0.40353113, "grad_norm": 0.0393556, "learning_rate": 5.209e-05, "epoch": 0.50233854, "global_step/max_steps": "1262/2513", "percentage": "50.22%", "elapsed_time": "3h 28m 16s", "remaining_time": "3h 26m 27s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100988} +{"loss": 0.43143892, "grad_norm": 0.04055529, "learning_rate": 5.203e-05, "epoch": 0.50273659, "global_step/max_steps": "1263/2513", "percentage": "50.26%", "elapsed_time": "3h 28m 27s", "remaining_time": "3h 26m 18s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100982} +{"loss": 0.34920207, "grad_norm": 0.03460906, "learning_rate": 5.197e-05, "epoch": 0.50313464, "global_step/max_steps": "1264/2513", "percentage": "50.30%", "elapsed_time": "3h 28m 37s", "remaining_time": "3h 26m 8s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100982} +{"loss": 0.38549444, "grad_norm": 0.03628419, "learning_rate": 5.19e-05, "epoch": 0.50353269, "global_step/max_steps": "1265/2513", "percentage": "50.34%", "elapsed_time": "3h 28m 47s", "remaining_time": "3h 25m 58s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100981} +{"loss": 0.35885334, "grad_norm": 0.03308801, "learning_rate": 5.184e-05, "epoch": 0.50393074, "global_step/max_steps": "1266/2513", "percentage": "50.38%", "elapsed_time": "3h 28m 57s", "remaining_time": "3h 25m 48s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10098} +{"loss": 0.41268259, "grad_norm": 0.03837495, "learning_rate": 5.177e-05, "epoch": 0.50432879, "global_step/max_steps": "1267/2513", "percentage": "50.42%", "elapsed_time": "3h 29m 6s", "remaining_time": "3h 25m 38s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100983} +{"loss": 0.39642352, "grad_norm": 0.04021688, "learning_rate": 5.171e-05, "epoch": 0.50472684, "global_step/max_steps": "1268/2513", "percentage": "50.46%", "elapsed_time": "3h 29m 16s", "remaining_time": "3h 25m 28s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100985} +{"loss": 0.3539179, "grad_norm": 0.03368013, "learning_rate": 5.164e-05, "epoch": 0.50512489, "global_step/max_steps": "1269/2513", "percentage": "50.50%", "elapsed_time": "3h 29m 26s", "remaining_time": "3h 25m 18s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100985} +{"loss": 0.35676029, "grad_norm": 0.03547221, "learning_rate": 5.158e-05, "epoch": 0.50552294, "global_step/max_steps": "1270/2513", "percentage": "50.54%", "elapsed_time": "3h 29m 36s", "remaining_time": "3h 25m 8s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100986} +{"loss": 0.37826374, "grad_norm": 0.03869432, "learning_rate": 5.151e-05, "epoch": 0.50592099, "global_step/max_steps": "1271/2513", "percentage": "50.58%", "elapsed_time": "3h 29m 45s", "remaining_time": "3h 24m 58s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100986} +{"loss": 0.36010313, "grad_norm": 0.03736697, "learning_rate": 5.145e-05, "epoch": 0.50631904, "global_step/max_steps": "1272/2513", "percentage": "50.62%", "elapsed_time": "3h 29m 55s", "remaining_time": "3h 24m 48s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100987} +{"loss": 0.39260602, "grad_norm": 0.03985176, "learning_rate": 5.139e-05, "epoch": 0.50671709, "global_step/max_steps": "1273/2513", "percentage": "50.66%", "elapsed_time": "3h 30m 5s", "remaining_time": "3h 24m 38s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100989} +{"loss": 0.36220038, "grad_norm": 0.03634655, "learning_rate": 5.132e-05, "epoch": 0.50711514, "global_step/max_steps": "1274/2513", "percentage": "50.70%", "elapsed_time": "3h 30m 15s", "remaining_time": "3h 24m 28s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100991} +{"loss": 0.34238517, "grad_norm": 0.03758319, "learning_rate": 5.126e-05, "epoch": 0.50751319, "global_step/max_steps": "1275/2513", "percentage": "50.74%", "elapsed_time": "3h 30m 25s", "remaining_time": "3h 24m 18s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100988} +{"loss": 0.39759135, "grad_norm": 0.03710954, "learning_rate": 5.119e-05, "epoch": 0.50791123, "global_step/max_steps": "1276/2513", "percentage": "50.78%", "elapsed_time": "3h 30m 35s", "remaining_time": "3h 24m 8s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100988} +{"loss": 0.35478517, "grad_norm": 0.0526224, "learning_rate": 5.113e-05, "epoch": 0.50830928, "global_step/max_steps": "1277/2513", "percentage": "50.82%", "elapsed_time": "3h 30m 46s", "remaining_time": "3h 24m 0s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10098} +{"loss": 0.34199661, "grad_norm": 0.03560114, "learning_rate": 5.106e-05, "epoch": 0.50870733, "global_step/max_steps": "1278/2513", "percentage": "50.86%", "elapsed_time": "3h 30m 56s", "remaining_time": "3h 23m 51s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100972} +{"loss": 0.38995785, "grad_norm": 0.03713617, "learning_rate": 5.1e-05, "epoch": 0.50910538, "global_step/max_steps": "1279/2513", "percentage": "50.90%", "elapsed_time": "3h 31m 6s", "remaining_time": "3h 23m 41s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100973} +{"loss": 0.36334723, "grad_norm": 0.03812359, "learning_rate": 5.093e-05, "epoch": 0.50950343, "global_step/max_steps": "1280/2513", "percentage": "50.94%", "elapsed_time": "3h 31m 16s", "remaining_time": "3h 23m 31s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100972} +{"loss": 0.36144537, "grad_norm": 0.0385777, "learning_rate": 5.087e-05, "epoch": 0.50990148, "global_step/max_steps": "1281/2513", "percentage": "50.97%", "elapsed_time": "3h 31m 26s", "remaining_time": "3h 23m 21s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100972} +{"loss": 0.33524022, "grad_norm": 0.03672687, "learning_rate": 5.081e-05, "epoch": 0.51029953, "global_step/max_steps": "1282/2513", "percentage": "51.01%", "elapsed_time": "3h 31m 36s", "remaining_time": "3h 23m 11s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100971} +{"loss": 0.34491435, "grad_norm": 0.03760311, "learning_rate": 5.074e-05, "epoch": 0.51069758, "global_step/max_steps": "1283/2513", "percentage": "51.05%", "elapsed_time": "3h 31m 46s", "remaining_time": "3h 23m 1s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100971} +{"loss": 0.31565681, "grad_norm": 0.03600815, "learning_rate": 5.068e-05, "epoch": 0.51109563, "global_step/max_steps": "1284/2513", "percentage": "51.09%", "elapsed_time": "3h 31m 56s", "remaining_time": "3h 22m 51s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100974} +{"loss": 0.38552779, "grad_norm": 0.03863073, "learning_rate": 5.061e-05, "epoch": 0.51149368, "global_step/max_steps": "1285/2513", "percentage": "51.13%", "elapsed_time": "3h 32m 6s", "remaining_time": "3h 22m 41s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100972} +{"loss": 0.36938938, "grad_norm": 0.0345849, "learning_rate": 5.055e-05, "epoch": 0.51189173, "global_step/max_steps": "1286/2513", "percentage": "51.17%", "elapsed_time": "3h 32m 16s", "remaining_time": "3h 22m 31s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100973} +{"loss": 0.36609429, "grad_norm": 0.03979463, "learning_rate": 5.048e-05, "epoch": 0.51228978, "global_step/max_steps": "1287/2513", "percentage": "51.21%", "elapsed_time": "3h 32m 26s", "remaining_time": "3h 22m 21s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100972} +{"loss": 0.3963834, "grad_norm": 0.03875119, "learning_rate": 5.042e-05, "epoch": 0.51268783, "global_step/max_steps": "1288/2513", "percentage": "51.25%", "elapsed_time": "3h 32m 35s", "remaining_time": "3h 22m 12s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100972} +{"loss": 0.39369199, "grad_norm": 0.03752268, "learning_rate": 5.035e-05, "epoch": 0.51308588, "global_step/max_steps": "1289/2513", "percentage": "51.29%", "elapsed_time": "3h 32m 45s", "remaining_time": "3h 22m 2s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100972} +{"loss": 0.41168717, "grad_norm": 0.03722548, "learning_rate": 5.029e-05, "epoch": 0.51348393, "global_step/max_steps": "1290/2513", "percentage": "51.33%", "elapsed_time": "3h 32m 56s", "remaining_time": "3h 21m 52s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100969} +{"loss": 0.35430419, "grad_norm": 0.03312449, "learning_rate": 5.023e-05, "epoch": 0.51388198, "global_step/max_steps": "1291/2513", "percentage": "51.37%", "elapsed_time": "3h 33m 6s", "remaining_time": "3h 21m 42s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100968} +{"loss": 0.39056104, "grad_norm": 0.04113855, "learning_rate": 5.016e-05, "epoch": 0.51428003, "global_step/max_steps": "1292/2513", "percentage": "51.41%", "elapsed_time": "3h 33m 16s", "remaining_time": "3h 21m 33s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100965} +{"loss": 0.38059914, "grad_norm": 0.03870867, "learning_rate": 5.01e-05, "epoch": 0.51467808, "global_step/max_steps": "1293/2513", "percentage": "51.45%", "elapsed_time": "3h 33m 26s", "remaining_time": "3h 21m 23s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100961} +{"loss": 0.346957, "grad_norm": 0.03360431, "learning_rate": 5.003e-05, "epoch": 0.51507613, "global_step/max_steps": "1294/2513", "percentage": "51.49%", "elapsed_time": "3h 33m 36s", "remaining_time": "3h 21m 13s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100962} +{"loss": 0.33001453, "grad_norm": 0.03544368, "learning_rate": 4.997e-05, "epoch": 0.51547418, "global_step/max_steps": "1295/2513", "percentage": "51.53%", "elapsed_time": "3h 33m 46s", "remaining_time": "3h 21m 3s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100965} +{"loss": 0.39894712, "grad_norm": 0.03892443, "learning_rate": 4.99e-05, "epoch": 0.51587223, "global_step/max_steps": "1296/2513", "percentage": "51.57%", "elapsed_time": "3h 33m 55s", "remaining_time": "3h 20m 53s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100967} +{"loss": 0.38934627, "grad_norm": 0.03642319, "learning_rate": 4.984e-05, "epoch": 0.51627028, "global_step/max_steps": "1297/2513", "percentage": "51.61%", "elapsed_time": "3h 34m 5s", "remaining_time": "3h 20m 43s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10097} +{"loss": 0.41923901, "grad_norm": 0.04307564, "learning_rate": 4.977e-05, "epoch": 0.51666833, "global_step/max_steps": "1298/2513", "percentage": "51.65%", "elapsed_time": "3h 34m 15s", "remaining_time": "3h 20m 33s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100971} +{"loss": 0.36835116, "grad_norm": 0.03850428, "learning_rate": 4.971e-05, "epoch": 0.51706637, "global_step/max_steps": "1299/2513", "percentage": "51.69%", "elapsed_time": "3h 34m 24s", "remaining_time": "3h 20m 23s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100972} +{"loss": 0.38497368, "grad_norm": 0.03706643, "learning_rate": 4.965e-05, "epoch": 0.51746442, "global_step/max_steps": "1300/2513", "percentage": "51.73%", "elapsed_time": "3h 34m 34s", "remaining_time": "3h 20m 13s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100972} +{"loss": 0.33923721, "grad_norm": 0.0401908, "learning_rate": 4.958e-05, "epoch": 0.51786247, "global_step/max_steps": "1301/2513", "percentage": "51.77%", "elapsed_time": "3h 34m 46s", "remaining_time": "3h 20m 4s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100962} +{"loss": 0.40083486, "grad_norm": 0.03945491, "learning_rate": 4.952e-05, "epoch": 0.51826052, "global_step/max_steps": "1302/2513", "percentage": "51.81%", "elapsed_time": "3h 34m 55s", "remaining_time": "3h 19m 54s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100964} +{"loss": 0.32278591, "grad_norm": 0.03711731, "learning_rate": 4.945e-05, "epoch": 0.51865857, "global_step/max_steps": "1303/2513", "percentage": "51.85%", "elapsed_time": "3h 35m 5s", "remaining_time": "3h 19m 44s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100967} +{"loss": 0.35633451, "grad_norm": 0.03702312, "learning_rate": 4.939e-05, "epoch": 0.51905662, "global_step/max_steps": "1304/2513", "percentage": "51.89%", "elapsed_time": "3h 35m 15s", "remaining_time": "3h 19m 34s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100968} +{"loss": 0.35817808, "grad_norm": 0.03903353, "learning_rate": 4.932e-05, "epoch": 0.51945467, "global_step/max_steps": "1305/2513", "percentage": "51.93%", "elapsed_time": "3h 35m 25s", "remaining_time": "3h 19m 24s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100967} +{"loss": 0.37760222, "grad_norm": 0.0370478, "learning_rate": 4.926e-05, "epoch": 0.51985272, "global_step/max_steps": "1306/2513", "percentage": "51.97%", "elapsed_time": "3h 35m 35s", "remaining_time": "3h 19m 15s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100961} +{"loss": 0.35247433, "grad_norm": 0.03590401, "learning_rate": 4.919e-05, "epoch": 0.52025077, "global_step/max_steps": "1307/2513", "percentage": "52.01%", "elapsed_time": "3h 35m 45s", "remaining_time": "3h 19m 5s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100961} +{"loss": 0.36863491, "grad_norm": 0.03783801, "learning_rate": 4.913e-05, "epoch": 0.52064882, "global_step/max_steps": "1308/2513", "percentage": "52.05%", "elapsed_time": "3h 35m 56s", "remaining_time": "3h 18m 55s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100955} +{"loss": 0.32801026, "grad_norm": 0.03124318, "learning_rate": 4.907e-05, "epoch": 0.52104687, "global_step/max_steps": "1309/2513", "percentage": "52.09%", "elapsed_time": "3h 36m 5s", "remaining_time": "3h 18m 45s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100957} +{"loss": 0.40247536, "grad_norm": 0.0393014, "learning_rate": 4.9e-05, "epoch": 0.52144492, "global_step/max_steps": "1310/2513", "percentage": "52.13%", "elapsed_time": "3h 36m 15s", "remaining_time": "3h 18m 35s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100958} +{"loss": 0.36921054, "grad_norm": 0.04139529, "learning_rate": 4.894e-05, "epoch": 0.52184297, "global_step/max_steps": "1311/2513", "percentage": "52.17%", "elapsed_time": "3h 36m 25s", "remaining_time": "3h 18m 26s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100957} +{"loss": 0.37401515, "grad_norm": 0.03612501, "learning_rate": 4.887e-05, "epoch": 0.52224102, "global_step/max_steps": "1312/2513", "percentage": "52.21%", "elapsed_time": "3h 36m 35s", "remaining_time": "3h 18m 15s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10096} +{"loss": 0.34609225, "grad_norm": 0.03690437, "learning_rate": 4.881e-05, "epoch": 0.52263907, "global_step/max_steps": "1313/2513", "percentage": "52.25%", "elapsed_time": "3h 36m 44s", "remaining_time": "3h 18m 5s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100962} +{"loss": 0.36105001, "grad_norm": 0.03833219, "learning_rate": 4.874e-05, "epoch": 0.52303712, "global_step/max_steps": "1314/2513", "percentage": "52.29%", "elapsed_time": "3h 36m 54s", "remaining_time": "3h 17m 55s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100966} +{"loss": 0.36963314, "grad_norm": 0.03985429, "learning_rate": 4.868e-05, "epoch": 0.52343517, "global_step/max_steps": "1315/2513", "percentage": "52.33%", "elapsed_time": "3h 37m 3s", "remaining_time": "3h 17m 45s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100968} +{"loss": 0.37718442, "grad_norm": 0.0353495, "learning_rate": 4.861e-05, "epoch": 0.52383322, "global_step/max_steps": "1316/2513", "percentage": "52.37%", "elapsed_time": "3h 37m 13s", "remaining_time": "3h 17m 34s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100971} +{"loss": 0.37829328, "grad_norm": 0.03632031, "learning_rate": 4.855e-05, "epoch": 0.52423127, "global_step/max_steps": "1317/2513", "percentage": "52.41%", "elapsed_time": "3h 37m 23s", "remaining_time": "3h 17m 24s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100971} +{"loss": 0.42233205, "grad_norm": 0.03871157, "learning_rate": 4.849e-05, "epoch": 0.52462932, "global_step/max_steps": "1318/2513", "percentage": "52.45%", "elapsed_time": "3h 37m 32s", "remaining_time": "3h 17m 14s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100974} +{"loss": 0.32020435, "grad_norm": 0.03422884, "learning_rate": 4.842e-05, "epoch": 0.52502737, "global_step/max_steps": "1319/2513", "percentage": "52.49%", "elapsed_time": "3h 37m 42s", "remaining_time": "3h 17m 4s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100975} +{"loss": 0.34961897, "grad_norm": 0.03998562, "learning_rate": 4.836e-05, "epoch": 0.52542542, "global_step/max_steps": "1320/2513", "percentage": "52.53%", "elapsed_time": "3h 37m 52s", "remaining_time": "3h 16m 54s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100977} +{"loss": 0.33273718, "grad_norm": 0.03615602, "learning_rate": 4.829e-05, "epoch": 0.52582347, "global_step/max_steps": "1321/2513", "percentage": "52.57%", "elapsed_time": "3h 38m 2s", "remaining_time": "3h 16m 45s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100972} +{"loss": 0.39682668, "grad_norm": 0.03670788, "learning_rate": 4.823e-05, "epoch": 0.52622151, "global_step/max_steps": "1322/2513", "percentage": "52.61%", "elapsed_time": "3h 38m 12s", "remaining_time": "3h 16m 35s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10097} +{"loss": 0.41369328, "grad_norm": 0.04222092, "learning_rate": 4.816e-05, "epoch": 0.52661956, "global_step/max_steps": "1323/2513", "percentage": "52.65%", "elapsed_time": "3h 38m 22s", "remaining_time": "3h 16m 25s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10097} +{"loss": 0.34909222, "grad_norm": 0.03463815, "learning_rate": 4.81e-05, "epoch": 0.52701761, "global_step/max_steps": "1324/2513", "percentage": "52.69%", "elapsed_time": "3h 38m 32s", "remaining_time": "3h 16m 15s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100969} +{"loss": 0.3742249, "grad_norm": 0.03418667, "learning_rate": 4.803e-05, "epoch": 0.52741566, "global_step/max_steps": "1325/2513", "percentage": "52.73%", "elapsed_time": "3h 38m 42s", "remaining_time": "3h 16m 5s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100972} +{"loss": 0.32572961, "grad_norm": 0.03641287, "learning_rate": 4.797e-05, "epoch": 0.52781371, "global_step/max_steps": "1326/2513", "percentage": "52.77%", "elapsed_time": "3h 38m 52s", "remaining_time": "3h 15m 55s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100974} +{"loss": 0.3707996, "grad_norm": 0.03642299, "learning_rate": 4.791e-05, "epoch": 0.52821176, "global_step/max_steps": "1327/2513", "percentage": "52.81%", "elapsed_time": "3h 39m 1s", "remaining_time": "3h 15m 45s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100975} +{"loss": 0.37160036, "grad_norm": 0.03557209, "learning_rate": 4.784e-05, "epoch": 0.52860981, "global_step/max_steps": "1328/2513", "percentage": "52.85%", "elapsed_time": "3h 39m 11s", "remaining_time": "3h 15m 35s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100975} +{"loss": 0.39504528, "grad_norm": 0.03689565, "learning_rate": 4.778e-05, "epoch": 0.52900786, "global_step/max_steps": "1329/2513", "percentage": "52.88%", "elapsed_time": "3h 39m 21s", "remaining_time": "3h 15m 25s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100978} +{"loss": 0.37509695, "grad_norm": 0.03804505, "learning_rate": 4.771e-05, "epoch": 0.52940591, "global_step/max_steps": "1330/2513", "percentage": "52.92%", "elapsed_time": "3h 39m 30s", "remaining_time": "3h 15m 14s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100982} +{"loss": 0.33185112, "grad_norm": 0.03767237, "learning_rate": 4.765e-05, "epoch": 0.52980396, "global_step/max_steps": "1331/2513", "percentage": "52.96%", "elapsed_time": "3h 39m 39s", "remaining_time": "3h 15m 4s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100987} +{"loss": 0.36523396, "grad_norm": 0.03635045, "learning_rate": 4.758e-05, "epoch": 0.53020201, "global_step/max_steps": "1332/2513", "percentage": "53.00%", "elapsed_time": "3h 39m 49s", "remaining_time": "3h 14m 54s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100987} +{"loss": 0.34177065, "grad_norm": 0.03767948, "learning_rate": 4.752e-05, "epoch": 0.53060006, "global_step/max_steps": "1333/2513", "percentage": "53.04%", "elapsed_time": "3h 39m 59s", "remaining_time": "3h 14m 44s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100991} +{"loss": 0.37291819, "grad_norm": 0.03922132, "learning_rate": 4.746e-05, "epoch": 0.53099811, "global_step/max_steps": "1334/2513", "percentage": "53.08%", "elapsed_time": "3h 40m 9s", "remaining_time": "3h 14m 34s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100989} +{"loss": 0.43030453, "grad_norm": 0.04264135, "learning_rate": 4.739e-05, "epoch": 0.53139616, "global_step/max_steps": "1335/2513", "percentage": "53.12%", "elapsed_time": "3h 40m 19s", "remaining_time": "3h 14m 25s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100985} +{"loss": 0.38451949, "grad_norm": 0.03807441, "learning_rate": 4.733e-05, "epoch": 0.53179421, "global_step/max_steps": "1336/2513", "percentage": "53.16%", "elapsed_time": "3h 40m 29s", "remaining_time": "3h 14m 15s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100983} +{"loss": 0.40532896, "grad_norm": 0.04007441, "learning_rate": 4.726e-05, "epoch": 0.53219226, "global_step/max_steps": "1337/2513", "percentage": "53.20%", "elapsed_time": "3h 40m 39s", "remaining_time": "3h 14m 5s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100982} +{"loss": 0.35036999, "grad_norm": 0.03393676, "learning_rate": 4.72e-05, "epoch": 0.53259031, "global_step/max_steps": "1338/2513", "percentage": "53.24%", "elapsed_time": "3h 40m 49s", "remaining_time": "3h 13m 55s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100982} +{"loss": 0.37265122, "grad_norm": 0.03660138, "learning_rate": 4.713e-05, "epoch": 0.53298836, "global_step/max_steps": "1339/2513", "percentage": "53.28%", "elapsed_time": "3h 40m 59s", "remaining_time": "3h 13m 45s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100984} +{"loss": 0.38289908, "grad_norm": 0.03836186, "learning_rate": 4.707e-05, "epoch": 0.53338641, "global_step/max_steps": "1340/2513", "percentage": "53.32%", "elapsed_time": "3h 41m 9s", "remaining_time": "3h 13m 35s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100984} +{"loss": 0.35110337, "grad_norm": 0.04074544, "learning_rate": 4.7e-05, "epoch": 0.53378446, "global_step/max_steps": "1341/2513", "percentage": "53.36%", "elapsed_time": "3h 41m 18s", "remaining_time": "3h 13m 25s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100987} +{"loss": 0.34207028, "grad_norm": 0.03557878, "learning_rate": 4.694e-05, "epoch": 0.53418251, "global_step/max_steps": "1342/2513", "percentage": "53.40%", "elapsed_time": "3h 41m 28s", "remaining_time": "3h 13m 15s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100988} +{"loss": 0.3989321, "grad_norm": 0.03753023, "learning_rate": 4.688e-05, "epoch": 0.53458056, "global_step/max_steps": "1343/2513", "percentage": "53.44%", "elapsed_time": "3h 41m 38s", "remaining_time": "3h 13m 5s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10099} +{"loss": 0.35332039, "grad_norm": 0.03479154, "learning_rate": 4.681e-05, "epoch": 0.5349786, "global_step/max_steps": "1344/2513", "percentage": "53.48%", "elapsed_time": "3h 41m 47s", "remaining_time": "3h 12m 55s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100992} +{"loss": 0.38157904, "grad_norm": 0.03921328, "learning_rate": 4.675e-05, "epoch": 0.53537665, "global_step/max_steps": "1345/2513", "percentage": "53.52%", "elapsed_time": "3h 41m 57s", "remaining_time": "3h 12m 44s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100996} +{"loss": 0.36454064, "grad_norm": 0.03795063, "learning_rate": 4.668e-05, "epoch": 0.5357747, "global_step/max_steps": "1346/2513", "percentage": "53.56%", "elapsed_time": "3h 42m 6s", "remaining_time": "3h 12m 34s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100998} +{"loss": 0.40122724, "grad_norm": 0.0425463, "learning_rate": 4.662e-05, "epoch": 0.53617275, "global_step/max_steps": "1347/2513", "percentage": "53.60%", "elapsed_time": "3h 42m 16s", "remaining_time": "3h 12m 24s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101} +{"loss": 0.34140268, "grad_norm": 0.03414402, "learning_rate": 4.655e-05, "epoch": 0.5365708, "global_step/max_steps": "1348/2513", "percentage": "53.64%", "elapsed_time": "3h 42m 26s", "remaining_time": "3h 12m 14s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101001} +{"loss": 0.34526968, "grad_norm": 0.03681366, "learning_rate": 4.649e-05, "epoch": 0.53696885, "global_step/max_steps": "1349/2513", "percentage": "53.68%", "elapsed_time": "3h 42m 36s", "remaining_time": "3h 12m 5s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100998} +{"loss": 0.43353897, "grad_norm": 0.04225372, "learning_rate": 4.643e-05, "epoch": 0.5373669, "global_step/max_steps": "1350/2513", "percentage": "53.72%", "elapsed_time": "3h 42m 46s", "remaining_time": "3h 11m 55s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100996} +{"loss": 0.39200604, "grad_norm": 0.03817278, "learning_rate": 4.636e-05, "epoch": 0.53776495, "global_step/max_steps": "1351/2513", "percentage": "53.76%", "elapsed_time": "3h 42m 56s", "remaining_time": "3h 11m 45s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100995} +{"loss": 0.39699453, "grad_norm": 0.03755343, "learning_rate": 4.63e-05, "epoch": 0.538163, "global_step/max_steps": "1352/2513", "percentage": "53.80%", "elapsed_time": "3h 43m 7s", "remaining_time": "3h 11m 35s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100993} +{"loss": 0.36047161, "grad_norm": 0.03700577, "learning_rate": 4.623e-05, "epoch": 0.53856105, "global_step/max_steps": "1353/2513", "percentage": "53.84%", "elapsed_time": "3h 43m 16s", "remaining_time": "3h 11m 25s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100995} +{"loss": 0.37551385, "grad_norm": 0.04312841, "learning_rate": 4.617e-05, "epoch": 0.5389591, "global_step/max_steps": "1354/2513", "percentage": "53.88%", "elapsed_time": "3h 43m 26s", "remaining_time": "3h 11m 15s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100997} +{"loss": 0.40843418, "grad_norm": 0.03973819, "learning_rate": 4.61e-05, "epoch": 0.53935715, "global_step/max_steps": "1355/2513", "percentage": "53.92%", "elapsed_time": "3h 43m 36s", "remaining_time": "3h 11m 5s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100996} +{"loss": 0.37786973, "grad_norm": 0.03981602, "learning_rate": 4.604e-05, "epoch": 0.5397552, "global_step/max_steps": "1356/2513", "percentage": "53.96%", "elapsed_time": "3h 43m 46s", "remaining_time": "3h 10m 55s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100997} +{"loss": 0.38227087, "grad_norm": 0.03928725, "learning_rate": 4.598e-05, "epoch": 0.54015325, "global_step/max_steps": "1357/2513", "percentage": "54.00%", "elapsed_time": "3h 43m 55s", "remaining_time": "3h 10m 45s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100999} +{"loss": 0.38476223, "grad_norm": 0.04069302, "learning_rate": 4.591e-05, "epoch": 0.5405513, "global_step/max_steps": "1358/2513", "percentage": "54.04%", "elapsed_time": "3h 44m 5s", "remaining_time": "3h 10m 35s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101001} +{"loss": 0.37604663, "grad_norm": 0.03420386, "learning_rate": 4.585e-05, "epoch": 0.54094935, "global_step/max_steps": "1359/2513", "percentage": "54.08%", "elapsed_time": "3h 44m 15s", "remaining_time": "3h 10m 25s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101001} +{"loss": 0.34540278, "grad_norm": 0.03461434, "learning_rate": 4.578e-05, "epoch": 0.5413474, "global_step/max_steps": "1360/2513", "percentage": "54.12%", "elapsed_time": "3h 44m 25s", "remaining_time": "3h 10m 15s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101002} +{"loss": 0.39446601, "grad_norm": 0.03784185, "learning_rate": 4.572e-05, "epoch": 0.54174545, "global_step/max_steps": "1361/2513", "percentage": "54.16%", "elapsed_time": "3h 44m 34s", "remaining_time": "3h 10m 5s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101002} +{"loss": 0.34893268, "grad_norm": 0.03920869, "learning_rate": 4.565e-05, "epoch": 0.5421435, "global_step/max_steps": "1362/2513", "percentage": "54.20%", "elapsed_time": "3h 44m 44s", "remaining_time": "3h 9m 55s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101004} +{"loss": 0.34133744, "grad_norm": 0.03496657, "learning_rate": 4.559e-05, "epoch": 0.54254155, "global_step/max_steps": "1363/2513", "percentage": "54.24%", "elapsed_time": "3h 44m 54s", "remaining_time": "3h 9m 45s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101003} +{"loss": 0.39633057, "grad_norm": 0.04084803, "learning_rate": 4.553e-05, "epoch": 0.5429396, "global_step/max_steps": "1364/2513", "percentage": "54.28%", "elapsed_time": "3h 45m 4s", "remaining_time": "3h 9m 35s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101004} +{"loss": 0.36317021, "grad_norm": 0.03600458, "learning_rate": 4.546e-05, "epoch": 0.54333765, "global_step/max_steps": "1365/2513", "percentage": "54.32%", "elapsed_time": "3h 45m 14s", "remaining_time": "3h 9m 26s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100999} +{"loss": 0.34968325, "grad_norm": 0.03446973, "learning_rate": 4.54e-05, "epoch": 0.5437357, "global_step/max_steps": "1366/2513", "percentage": "54.36%", "elapsed_time": "3h 45m 25s", "remaining_time": "3h 9m 16s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100996} +{"loss": 0.41146138, "grad_norm": 0.04281944, "learning_rate": 4.533e-05, "epoch": 0.54413374, "global_step/max_steps": "1367/2513", "percentage": "54.40%", "elapsed_time": "3h 45m 35s", "remaining_time": "3h 9m 7s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100995} +{"loss": 0.39069033, "grad_norm": 0.04281891, "learning_rate": 4.527e-05, "epoch": 0.54453179, "global_step/max_steps": "1368/2513", "percentage": "54.44%", "elapsed_time": "3h 45m 44s", "remaining_time": "3h 8m 56s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100997} +{"loss": 0.36535078, "grad_norm": 0.03582647, "learning_rate": 4.521e-05, "epoch": 0.54492984, "global_step/max_steps": "1369/2513", "percentage": "54.48%", "elapsed_time": "3h 45m 54s", "remaining_time": "3h 8m 46s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100999} +{"loss": 0.38735628, "grad_norm": 0.04045545, "learning_rate": 4.514e-05, "epoch": 0.54532789, "global_step/max_steps": "1370/2513", "percentage": "54.52%", "elapsed_time": "3h 46m 3s", "remaining_time": "3h 8m 36s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101003} +{"loss": 0.40221182, "grad_norm": 0.03865031, "learning_rate": 4.508e-05, "epoch": 0.54572594, "global_step/max_steps": "1371/2513", "percentage": "54.56%", "elapsed_time": "3h 46m 13s", "remaining_time": "3h 8m 26s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101004} +{"loss": 0.34497645, "grad_norm": 0.03547088, "learning_rate": 4.501e-05, "epoch": 0.54612399, "global_step/max_steps": "1372/2513", "percentage": "54.60%", "elapsed_time": "3h 46m 23s", "remaining_time": "3h 8m 16s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101006} +{"loss": 0.40504533, "grad_norm": 0.03862525, "learning_rate": 4.495e-05, "epoch": 0.54652204, "global_step/max_steps": "1373/2513", "percentage": "54.64%", "elapsed_time": "3h 46m 33s", "remaining_time": "3h 8m 6s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101004} +{"loss": 0.36528563, "grad_norm": 0.04012805, "learning_rate": 4.488e-05, "epoch": 0.54692009, "global_step/max_steps": "1374/2513", "percentage": "54.68%", "elapsed_time": "3h 46m 43s", "remaining_time": "3h 7m 56s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101004} +{"loss": 0.36388785, "grad_norm": 0.03789472, "learning_rate": 4.482e-05, "epoch": 0.54731814, "global_step/max_steps": "1375/2513", "percentage": "54.72%", "elapsed_time": "3h 46m 53s", "remaining_time": "3h 7m 46s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101006} +{"loss": 0.39473492, "grad_norm": 0.03766723, "learning_rate": 4.476e-05, "epoch": 0.54771619, "global_step/max_steps": "1376/2513", "percentage": "54.76%", "elapsed_time": "3h 47m 2s", "remaining_time": "3h 7m 36s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101008} +{"loss": 0.40239975, "grad_norm": 0.04033206, "learning_rate": 4.469e-05, "epoch": 0.54811424, "global_step/max_steps": "1377/2513", "percentage": "54.80%", "elapsed_time": "3h 47m 12s", "remaining_time": "3h 7m 26s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10101} +{"loss": 0.36376143, "grad_norm": 0.03660902, "learning_rate": 4.463e-05, "epoch": 0.54851229, "global_step/max_steps": "1378/2513", "percentage": "54.83%", "elapsed_time": "3h 47m 22s", "remaining_time": "3h 7m 16s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101007} +{"loss": 0.40608698, "grad_norm": 0.03860856, "learning_rate": 4.456e-05, "epoch": 0.54891034, "global_step/max_steps": "1379/2513", "percentage": "54.87%", "elapsed_time": "3h 47m 32s", "remaining_time": "3h 7m 6s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101007} +{"loss": 0.40045738, "grad_norm": 0.04286909, "learning_rate": 4.45e-05, "epoch": 0.54930839, "global_step/max_steps": "1380/2513", "percentage": "54.91%", "elapsed_time": "3h 47m 43s", "remaining_time": "3h 6m 57s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101001} +{"loss": 0.37275603, "grad_norm": 0.0382822, "learning_rate": 4.444e-05, "epoch": 0.54970644, "global_step/max_steps": "1381/2513", "percentage": "54.95%", "elapsed_time": "3h 47m 53s", "remaining_time": "3h 6m 48s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100996} +{"loss": 0.37239677, "grad_norm": 0.03808504, "learning_rate": 4.437e-05, "epoch": 0.55010449, "global_step/max_steps": "1382/2513", "percentage": "54.99%", "elapsed_time": "3h 48m 3s", "remaining_time": "3h 6m 38s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100997} +{"loss": 0.40236539, "grad_norm": 0.03928037, "learning_rate": 4.431e-05, "epoch": 0.55050254, "global_step/max_steps": "1383/2513", "percentage": "55.03%", "elapsed_time": "3h 48m 13s", "remaining_time": "3h 6m 28s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100997} +{"loss": 0.38583815, "grad_norm": 0.03829184, "learning_rate": 4.424e-05, "epoch": 0.55090059, "global_step/max_steps": "1384/2513", "percentage": "55.07%", "elapsed_time": "3h 48m 23s", "remaining_time": "3h 6m 18s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100994} +{"loss": 0.36716634, "grad_norm": 0.04245746, "learning_rate": 4.418e-05, "epoch": 0.55129864, "global_step/max_steps": "1385/2513", "percentage": "55.11%", "elapsed_time": "3h 48m 34s", "remaining_time": "3h 6m 9s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100991} +{"loss": 0.37201574, "grad_norm": 0.03953209, "learning_rate": 4.412e-05, "epoch": 0.55169669, "global_step/max_steps": "1386/2513", "percentage": "55.15%", "elapsed_time": "3h 48m 43s", "remaining_time": "3h 5m 59s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100992} +{"loss": 0.38283211, "grad_norm": 0.04004745, "learning_rate": 4.405e-05, "epoch": 0.55209474, "global_step/max_steps": "1387/2513", "percentage": "55.19%", "elapsed_time": "3h 48m 53s", "remaining_time": "3h 5m 49s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100994} +{"loss": 0.36801037, "grad_norm": 0.0384211, "learning_rate": 4.399e-05, "epoch": 0.55249279, "global_step/max_steps": "1388/2513", "percentage": "55.23%", "elapsed_time": "3h 49m 3s", "remaining_time": "3h 5m 39s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100994} +{"loss": 0.35916981, "grad_norm": 0.0346489, "learning_rate": 4.392e-05, "epoch": 0.55289083, "global_step/max_steps": "1389/2513", "percentage": "55.27%", "elapsed_time": "3h 49m 13s", "remaining_time": "3h 5m 29s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100995} +{"loss": 0.32543117, "grad_norm": 0.03799957, "learning_rate": 4.386e-05, "epoch": 0.55328888, "global_step/max_steps": "1390/2513", "percentage": "55.31%", "elapsed_time": "3h 49m 22s", "remaining_time": "3h 5m 19s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100997} +{"loss": 0.34680611, "grad_norm": 0.03797158, "learning_rate": 4.38e-05, "epoch": 0.55368693, "global_step/max_steps": "1391/2513", "percentage": "55.35%", "elapsed_time": "3h 49m 32s", "remaining_time": "3h 5m 9s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100998} +{"loss": 0.37590811, "grad_norm": 0.03794139, "learning_rate": 4.373e-05, "epoch": 0.55408498, "global_step/max_steps": "1392/2513", "percentage": "55.39%", "elapsed_time": "3h 49m 42s", "remaining_time": "3h 4m 59s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100997} +{"loss": 0.35631314, "grad_norm": 0.0385678, "learning_rate": 4.367e-05, "epoch": 0.55448303, "global_step/max_steps": "1393/2513", "percentage": "55.43%", "elapsed_time": "3h 49m 52s", "remaining_time": "3h 4m 49s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100996} +{"loss": 0.32522434, "grad_norm": 0.03295492, "learning_rate": 4.36e-05, "epoch": 0.55488108, "global_step/max_steps": "1394/2513", "percentage": "55.47%", "elapsed_time": "3h 50m 2s", "remaining_time": "3h 4m 39s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100993} +{"loss": 0.36037034, "grad_norm": 0.04268104, "learning_rate": 4.354e-05, "epoch": 0.55527913, "global_step/max_steps": "1395/2513", "percentage": "55.51%", "elapsed_time": "3h 50m 13s", "remaining_time": "3h 4m 30s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100989} +{"loss": 0.39499259, "grad_norm": 0.0396797, "learning_rate": 4.348e-05, "epoch": 0.55567718, "global_step/max_steps": "1396/2513", "percentage": "55.55%", "elapsed_time": "3h 50m 24s", "remaining_time": "3h 4m 21s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100983} +{"loss": 0.32952806, "grad_norm": 0.03355445, "learning_rate": 4.341e-05, "epoch": 0.55607523, "global_step/max_steps": "1397/2513", "percentage": "55.59%", "elapsed_time": "3h 50m 33s", "remaining_time": "3h 4m 11s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100985} +{"loss": 0.34427074, "grad_norm": 0.03746034, "learning_rate": 4.335e-05, "epoch": 0.55647328, "global_step/max_steps": "1398/2513", "percentage": "55.63%", "elapsed_time": "3h 50m 43s", "remaining_time": "3h 4m 1s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100984} +{"loss": 0.35273507, "grad_norm": 0.03802476, "learning_rate": 4.328e-05, "epoch": 0.55687133, "global_step/max_steps": "1399/2513", "percentage": "55.67%", "elapsed_time": "3h 50m 53s", "remaining_time": "3h 3m 51s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100985} +{"loss": 0.36229616, "grad_norm": 0.0414679, "learning_rate": 4.322e-05, "epoch": 0.55726938, "global_step/max_steps": "1400/2513", "percentage": "55.71%", "elapsed_time": "3h 51m 3s", "remaining_time": "3h 3m 41s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100987} +{"loss": 0.37150732, "grad_norm": 0.03952656, "learning_rate": 4.316e-05, "epoch": 0.55766743, "global_step/max_steps": "1401/2513", "percentage": "55.75%", "elapsed_time": "3h 51m 14s", "remaining_time": "3h 3m 32s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100977} +{"loss": 0.39060307, "grad_norm": 0.03938337, "learning_rate": 4.309e-05, "epoch": 0.55806548, "global_step/max_steps": "1402/2513", "percentage": "55.79%", "elapsed_time": "3h 51m 24s", "remaining_time": "3h 3m 22s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100977} +{"loss": 0.42115986, "grad_norm": 0.04020094, "learning_rate": 4.303e-05, "epoch": 0.55846353, "global_step/max_steps": "1403/2513", "percentage": "55.83%", "elapsed_time": "3h 51m 34s", "remaining_time": "3h 3m 12s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100977} +{"loss": 0.36175019, "grad_norm": 0.03955844, "learning_rate": 4.297e-05, "epoch": 0.55886158, "global_step/max_steps": "1404/2513", "percentage": "55.87%", "elapsed_time": "3h 51m 44s", "remaining_time": "3h 3m 2s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100976} +{"loss": 0.41246024, "grad_norm": 0.03886866, "learning_rate": 4.29e-05, "epoch": 0.55925963, "global_step/max_steps": "1405/2513", "percentage": "55.91%", "elapsed_time": "3h 51m 54s", "remaining_time": "3h 2m 52s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100977} +{"loss": 0.36490273, "grad_norm": 0.03607539, "learning_rate": 4.284e-05, "epoch": 0.55965768, "global_step/max_steps": "1406/2513", "percentage": "55.95%", "elapsed_time": "3h 52m 4s", "remaining_time": "3h 2m 42s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100976} +{"loss": 0.35380167, "grad_norm": 0.03726033, "learning_rate": 4.277e-05, "epoch": 0.56005573, "global_step/max_steps": "1407/2513", "percentage": "55.99%", "elapsed_time": "3h 52m 14s", "remaining_time": "3h 2m 33s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100975} +{"loss": 0.34340632, "grad_norm": 0.03425944, "learning_rate": 4.271e-05, "epoch": 0.56045378, "global_step/max_steps": "1408/2513", "percentage": "56.03%", "elapsed_time": "3h 52m 24s", "remaining_time": "3h 2m 23s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100972} +{"loss": 0.33478332, "grad_norm": 0.03358534, "learning_rate": 4.265e-05, "epoch": 0.56085183, "global_step/max_steps": "1409/2513", "percentage": "56.07%", "elapsed_time": "3h 52m 35s", "remaining_time": "3h 2m 14s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100963} +{"loss": 0.35495773, "grad_norm": 0.04075749, "learning_rate": 4.258e-05, "epoch": 0.56124988, "global_step/max_steps": "1410/2513", "percentage": "56.11%", "elapsed_time": "3h 52m 45s", "remaining_time": "3h 2m 4s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100961} +{"loss": 0.38263845, "grad_norm": 0.03837966, "learning_rate": 4.252e-05, "epoch": 0.56164793, "global_step/max_steps": "1411/2513", "percentage": "56.15%", "elapsed_time": "3h 52m 56s", "remaining_time": "3h 1m 55s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100959} +{"loss": 0.36798567, "grad_norm": 0.03515039, "learning_rate": 4.246e-05, "epoch": 0.56204597, "global_step/max_steps": "1412/2513", "percentage": "56.19%", "elapsed_time": "3h 53m 5s", "remaining_time": "3h 1m 45s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100961} +{"loss": 0.37892288, "grad_norm": 0.0372568, "learning_rate": 4.239e-05, "epoch": 0.56244402, "global_step/max_steps": "1413/2513", "percentage": "56.23%", "elapsed_time": "3h 53m 15s", "remaining_time": "3h 1m 35s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100962} +{"loss": 0.34580702, "grad_norm": 0.03869077, "learning_rate": 4.233e-05, "epoch": 0.56284207, "global_step/max_steps": "1414/2513", "percentage": "56.27%", "elapsed_time": "3h 53m 25s", "remaining_time": "3h 1m 25s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100962} +{"loss": 0.36654642, "grad_norm": 0.03757107, "learning_rate": 4.226e-05, "epoch": 0.56324012, "global_step/max_steps": "1415/2513", "percentage": "56.31%", "elapsed_time": "3h 53m 35s", "remaining_time": "3h 1m 15s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10096} +{"loss": 0.32155281, "grad_norm": 0.03641445, "learning_rate": 4.22e-05, "epoch": 0.56363817, "global_step/max_steps": "1416/2513", "percentage": "56.35%", "elapsed_time": "3h 53m 45s", "remaining_time": "3h 1m 5s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100961} +{"loss": 0.39481252, "grad_norm": 0.03945908, "learning_rate": 4.214e-05, "epoch": 0.56403622, "global_step/max_steps": "1417/2513", "percentage": "56.39%", "elapsed_time": "3h 53m 54s", "remaining_time": "3h 0m 55s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100962} +{"loss": 0.36973178, "grad_norm": 0.03679854, "learning_rate": 4.207e-05, "epoch": 0.56443427, "global_step/max_steps": "1418/2513", "percentage": "56.43%", "elapsed_time": "3h 54m 4s", "remaining_time": "3h 0m 45s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100961} +{"loss": 0.40154645, "grad_norm": 0.04083598, "learning_rate": 4.201e-05, "epoch": 0.56483232, "global_step/max_steps": "1419/2513", "percentage": "56.47%", "elapsed_time": "3h 54m 14s", "remaining_time": "3h 0m 35s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100963} +{"loss": 0.38732916, "grad_norm": 0.03627798, "learning_rate": 4.195e-05, "epoch": 0.56523037, "global_step/max_steps": "1420/2513", "percentage": "56.51%", "elapsed_time": "3h 54m 24s", "remaining_time": "3h 0m 25s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100963} +{"loss": 0.39931726, "grad_norm": 0.0378959, "learning_rate": 4.188e-05, "epoch": 0.56562842, "global_step/max_steps": "1421/2513", "percentage": "56.55%", "elapsed_time": "3h 54m 34s", "remaining_time": "3h 0m 16s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100961} +{"loss": 0.34670359, "grad_norm": 0.0374721, "learning_rate": 4.182e-05, "epoch": 0.56602647, "global_step/max_steps": "1422/2513", "percentage": "56.59%", "elapsed_time": "3h 54m 44s", "remaining_time": "3h 0m 6s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10096} +{"loss": 0.33692595, "grad_norm": 0.03523171, "learning_rate": 4.176e-05, "epoch": 0.56642452, "global_step/max_steps": "1423/2513", "percentage": "56.63%", "elapsed_time": "3h 54m 54s", "remaining_time": "2h 59m 56s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100959} +{"loss": 0.39996135, "grad_norm": 0.03869125, "learning_rate": 4.169e-05, "epoch": 0.56682257, "global_step/max_steps": "1424/2513", "percentage": "56.67%", "elapsed_time": "3h 55m 5s", "remaining_time": "2h 59m 47s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100954} +{"loss": 0.37516314, "grad_norm": 0.03596675, "learning_rate": 4.163e-05, "epoch": 0.56722062, "global_step/max_steps": "1425/2513", "percentage": "56.71%", "elapsed_time": "3h 55m 16s", "remaining_time": "2h 59m 37s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100949} +{"loss": 0.39666414, "grad_norm": 0.03910181, "learning_rate": 4.156e-05, "epoch": 0.56761867, "global_step/max_steps": "1426/2513", "percentage": "56.74%", "elapsed_time": "3h 55m 26s", "remaining_time": "2h 59m 27s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100949} +{"loss": 0.3407647, "grad_norm": 0.03667938, "learning_rate": 4.15e-05, "epoch": 0.56801672, "global_step/max_steps": "1427/2513", "percentage": "56.78%", "elapsed_time": "3h 55m 35s", "remaining_time": "2h 59m 17s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100949} +{"loss": 0.35089999, "grad_norm": 0.03746576, "learning_rate": 4.144e-05, "epoch": 0.56841477, "global_step/max_steps": "1428/2513", "percentage": "56.82%", "elapsed_time": "3h 55m 45s", "remaining_time": "2h 59m 8s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100949} +{"loss": 0.31115404, "grad_norm": 0.03700372, "learning_rate": 4.137e-05, "epoch": 0.56881282, "global_step/max_steps": "1429/2513", "percentage": "56.86%", "elapsed_time": "3h 55m 55s", "remaining_time": "2h 58m 58s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100949} +{"loss": 0.35011768, "grad_norm": 0.04035016, "learning_rate": 4.131e-05, "epoch": 0.56921087, "global_step/max_steps": "1430/2513", "percentage": "56.90%", "elapsed_time": "3h 56m 5s", "remaining_time": "2h 58m 47s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100952} +{"loss": 0.31710124, "grad_norm": 0.03718734, "learning_rate": 4.125e-05, "epoch": 0.56960892, "global_step/max_steps": "1431/2513", "percentage": "56.94%", "elapsed_time": "3h 56m 14s", "remaining_time": "2h 58m 37s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100953} +{"loss": 0.36210138, "grad_norm": 0.0400697, "learning_rate": 4.118e-05, "epoch": 0.57000697, "global_step/max_steps": "1432/2513", "percentage": "56.98%", "elapsed_time": "3h 56m 24s", "remaining_time": "2h 58m 28s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100952} +{"loss": 0.3863377, "grad_norm": 0.03969403, "learning_rate": 4.112e-05, "epoch": 0.57040502, "global_step/max_steps": "1433/2513", "percentage": "57.02%", "elapsed_time": "3h 56m 34s", "remaining_time": "2h 58m 18s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100952} +{"loss": 0.38386485, "grad_norm": 0.03931983, "learning_rate": 4.106e-05, "epoch": 0.57080306, "global_step/max_steps": "1434/2513", "percentage": "57.06%", "elapsed_time": "3h 56m 44s", "remaining_time": "2h 58m 8s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100953} +{"loss": 0.35639378, "grad_norm": 0.04093411, "learning_rate": 4.099e-05, "epoch": 0.57120111, "global_step/max_steps": "1435/2513", "percentage": "57.10%", "elapsed_time": "3h 56m 54s", "remaining_time": "2h 57m 58s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100952} +{"loss": 0.40131211, "grad_norm": 0.04046551, "learning_rate": 4.093e-05, "epoch": 0.57159916, "global_step/max_steps": "1436/2513", "percentage": "57.14%", "elapsed_time": "3h 57m 4s", "remaining_time": "2h 57m 48s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100954} +{"loss": 0.39017725, "grad_norm": 0.04278498, "learning_rate": 4.087e-05, "epoch": 0.57199721, "global_step/max_steps": "1437/2513", "percentage": "57.18%", "elapsed_time": "3h 57m 14s", "remaining_time": "2h 57m 38s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100953} +{"loss": 0.35565865, "grad_norm": 0.03625567, "learning_rate": 4.08e-05, "epoch": 0.57239526, "global_step/max_steps": "1438/2513", "percentage": "57.22%", "elapsed_time": "3h 57m 24s", "remaining_time": "2h 57m 28s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100953} +{"loss": 0.38117015, "grad_norm": 0.04104637, "learning_rate": 4.074e-05, "epoch": 0.57279331, "global_step/max_steps": "1439/2513", "percentage": "57.26%", "elapsed_time": "3h 57m 35s", "remaining_time": "2h 57m 19s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100945} +{"loss": 0.37277254, "grad_norm": 0.03350739, "learning_rate": 4.068e-05, "epoch": 0.57319136, "global_step/max_steps": "1440/2513", "percentage": "57.30%", "elapsed_time": "3h 57m 45s", "remaining_time": "2h 57m 10s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10094} +{"loss": 0.31599796, "grad_norm": 0.03697387, "learning_rate": 4.061e-05, "epoch": 0.57358941, "global_step/max_steps": "1441/2513", "percentage": "57.34%", "elapsed_time": "3h 57m 55s", "remaining_time": "2h 57m 0s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100939} +{"loss": 0.36663687, "grad_norm": 0.03787301, "learning_rate": 4.055e-05, "epoch": 0.57398746, "global_step/max_steps": "1442/2513", "percentage": "57.38%", "elapsed_time": "3h 58m 5s", "remaining_time": "2h 56m 50s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100941} +{"loss": 0.33801582, "grad_norm": 0.03486211, "learning_rate": 4.049e-05, "epoch": 0.57438551, "global_step/max_steps": "1443/2513", "percentage": "57.42%", "elapsed_time": "3h 58m 15s", "remaining_time": "2h 56m 39s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100943} +{"loss": 0.36514801, "grad_norm": 0.0364836, "learning_rate": 4.042e-05, "epoch": 0.57478356, "global_step/max_steps": "1444/2513", "percentage": "57.46%", "elapsed_time": "3h 58m 24s", "remaining_time": "2h 56m 29s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100945} +{"loss": 0.38395208, "grad_norm": 0.03847795, "learning_rate": 4.036e-05, "epoch": 0.57518161, "global_step/max_steps": "1445/2513", "percentage": "57.50%", "elapsed_time": "3h 58m 34s", "remaining_time": "2h 56m 19s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100947} +{"loss": 0.41461712, "grad_norm": 0.04433949, "learning_rate": 4.03e-05, "epoch": 0.57557966, "global_step/max_steps": "1446/2513", "percentage": "57.54%", "elapsed_time": "3h 58m 44s", "remaining_time": "2h 56m 9s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100947} +{"loss": 0.3579922, "grad_norm": 0.03997282, "learning_rate": 4.023e-05, "epoch": 0.57597771, "global_step/max_steps": "1447/2513", "percentage": "57.58%", "elapsed_time": "3h 58m 53s", "remaining_time": "2h 55m 59s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100949} +{"loss": 0.33641693, "grad_norm": 0.03520917, "learning_rate": 4.017e-05, "epoch": 0.57637576, "global_step/max_steps": "1448/2513", "percentage": "57.62%", "elapsed_time": "3h 59m 3s", "remaining_time": "2h 55m 49s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100952} +{"loss": 0.39083076, "grad_norm": 0.03962765, "learning_rate": 4.011e-05, "epoch": 0.57677381, "global_step/max_steps": "1449/2513", "percentage": "57.66%", "elapsed_time": "3h 59m 13s", "remaining_time": "2h 55m 39s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100952} +{"loss": 0.33700448, "grad_norm": 0.03714294, "learning_rate": 4.004e-05, "epoch": 0.57717186, "global_step/max_steps": "1450/2513", "percentage": "57.70%", "elapsed_time": "3h 59m 22s", "remaining_time": "2h 55m 29s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100955} +{"loss": 0.35564882, "grad_norm": 0.03937982, "learning_rate": 3.998e-05, "epoch": 0.57756991, "global_step/max_steps": "1451/2513", "percentage": "57.74%", "elapsed_time": "3h 59m 32s", "remaining_time": "2h 55m 19s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100956} +{"loss": 0.34496945, "grad_norm": 0.03872374, "learning_rate": 3.992e-05, "epoch": 0.57796796, "global_step/max_steps": "1452/2513", "percentage": "57.78%", "elapsed_time": "3h 59m 42s", "remaining_time": "2h 55m 9s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100956} +{"loss": 0.36534297, "grad_norm": 0.03746958, "learning_rate": 3.985e-05, "epoch": 0.57836601, "global_step/max_steps": "1453/2513", "percentage": "57.82%", "elapsed_time": "3h 59m 52s", "remaining_time": "2h 54m 59s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100958} +{"loss": 0.40701407, "grad_norm": 0.0444858, "learning_rate": 3.979e-05, "epoch": 0.57876406, "global_step/max_steps": "1454/2513", "percentage": "57.86%", "elapsed_time": "4h 0m 2s", "remaining_time": "2h 54m 49s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100956} +{"loss": 0.41107982, "grad_norm": 0.03810847, "learning_rate": 3.973e-05, "epoch": 0.57916211, "global_step/max_steps": "1455/2513", "percentage": "57.90%", "elapsed_time": "4h 0m 12s", "remaining_time": "2h 54m 39s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100956} +{"loss": 0.36823094, "grad_norm": 0.03773183, "learning_rate": 3.967e-05, "epoch": 0.57956016, "global_step/max_steps": "1456/2513", "percentage": "57.94%", "elapsed_time": "4h 0m 22s", "remaining_time": "2h 54m 30s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100955} +{"loss": 0.39830062, "grad_norm": 0.04293749, "learning_rate": 3.96e-05, "epoch": 0.5799582, "global_step/max_steps": "1457/2513", "percentage": "57.98%", "elapsed_time": "4h 0m 32s", "remaining_time": "2h 54m 20s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100955} +{"loss": 0.40512216, "grad_norm": 0.03908537, "learning_rate": 3.954e-05, "epoch": 0.58035625, "global_step/max_steps": "1458/2513", "percentage": "58.02%", "elapsed_time": "4h 0m 42s", "remaining_time": "2h 54m 10s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100955} +{"loss": 0.38582432, "grad_norm": 0.04024162, "learning_rate": 3.948e-05, "epoch": 0.5807543, "global_step/max_steps": "1459/2513", "percentage": "58.06%", "elapsed_time": "4h 0m 51s", "remaining_time": "2h 54m 0s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100956} +{"loss": 0.36988622, "grad_norm": 0.03934519, "learning_rate": 3.941e-05, "epoch": 0.58115235, "global_step/max_steps": "1460/2513", "percentage": "58.10%", "elapsed_time": "4h 1m 1s", "remaining_time": "2h 53m 50s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100957} +{"loss": 0.36239266, "grad_norm": 0.04040368, "learning_rate": 3.935e-05, "epoch": 0.5815504, "global_step/max_steps": "1461/2513", "percentage": "58.14%", "elapsed_time": "4h 1m 11s", "remaining_time": "2h 53m 40s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100958} +{"loss": 0.33075246, "grad_norm": 0.03826056, "learning_rate": 3.929e-05, "epoch": 0.58194845, "global_step/max_steps": "1462/2513", "percentage": "58.18%", "elapsed_time": "4h 1m 21s", "remaining_time": "2h 53m 30s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100959} +{"loss": 0.36559707, "grad_norm": 0.0357701, "learning_rate": 3.922e-05, "epoch": 0.5823465, "global_step/max_steps": "1463/2513", "percentage": "58.22%", "elapsed_time": "4h 1m 30s", "remaining_time": "2h 53m 20s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100961} +{"loss": 0.38902792, "grad_norm": 0.03555416, "learning_rate": 3.916e-05, "epoch": 0.58274455, "global_step/max_steps": "1464/2513", "percentage": "58.26%", "elapsed_time": "4h 1m 40s", "remaining_time": "2h 53m 10s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100962} +{"loss": 0.39375296, "grad_norm": 0.04126147, "learning_rate": 3.91e-05, "epoch": 0.5831426, "global_step/max_steps": "1465/2513", "percentage": "58.30%", "elapsed_time": "4h 1m 50s", "remaining_time": "2h 52m 59s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100965} +{"loss": 0.38048601, "grad_norm": 0.03551867, "learning_rate": 3.904e-05, "epoch": 0.58354065, "global_step/max_steps": "1466/2513", "percentage": "58.34%", "elapsed_time": "4h 1m 59s", "remaining_time": "2h 52m 49s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100966} +{"loss": 0.3808915, "grad_norm": 0.03520523, "learning_rate": 3.897e-05, "epoch": 0.5839387, "global_step/max_steps": "1467/2513", "percentage": "58.38%", "elapsed_time": "4h 2m 9s", "remaining_time": "2h 52m 39s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100966} +{"loss": 0.36888418, "grad_norm": 0.03472139, "learning_rate": 3.891e-05, "epoch": 0.58433675, "global_step/max_steps": "1468/2513", "percentage": "58.42%", "elapsed_time": "4h 2m 19s", "remaining_time": "2h 52m 30s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100963} +{"loss": 0.32074308, "grad_norm": 0.0327358, "learning_rate": 3.885e-05, "epoch": 0.5847348, "global_step/max_steps": "1469/2513", "percentage": "58.46%", "elapsed_time": "4h 2m 30s", "remaining_time": "2h 52m 20s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100961} +{"loss": 0.42769533, "grad_norm": 0.03967023, "learning_rate": 3.878e-05, "epoch": 0.58513285, "global_step/max_steps": "1470/2513", "percentage": "58.50%", "elapsed_time": "4h 2m 40s", "remaining_time": "2h 52m 10s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100959} +{"loss": 0.37810838, "grad_norm": 0.03917043, "learning_rate": 3.872e-05, "epoch": 0.5855309, "global_step/max_steps": "1471/2513", "percentage": "58.54%", "elapsed_time": "4h 2m 50s", "remaining_time": "2h 52m 1s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100958} +{"loss": 0.35458335, "grad_norm": 0.03990793, "learning_rate": 3.866e-05, "epoch": 0.58592895, "global_step/max_steps": "1472/2513", "percentage": "58.58%", "elapsed_time": "4h 3m 0s", "remaining_time": "2h 51m 51s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10096} +{"loss": 0.35634691, "grad_norm": 0.0385593, "learning_rate": 3.86e-05, "epoch": 0.586327, "global_step/max_steps": "1473/2513", "percentage": "58.62%", "elapsed_time": "4h 3m 9s", "remaining_time": "2h 51m 41s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10096} +{"loss": 0.37370116, "grad_norm": 0.03692088, "learning_rate": 3.853e-05, "epoch": 0.58672505, "global_step/max_steps": "1474/2513", "percentage": "58.65%", "elapsed_time": "4h 3m 19s", "remaining_time": "2h 51m 31s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100961} +{"loss": 0.41675586, "grad_norm": 0.039023, "learning_rate": 3.847e-05, "epoch": 0.5871231, "global_step/max_steps": "1475/2513", "percentage": "58.69%", "elapsed_time": "4h 3m 29s", "remaining_time": "2h 51m 21s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100961} +{"loss": 0.37387848, "grad_norm": 0.03789366, "learning_rate": 3.841e-05, "epoch": 0.58752115, "global_step/max_steps": "1476/2513", "percentage": "58.73%", "elapsed_time": "4h 3m 39s", "remaining_time": "2h 51m 11s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100963} +{"loss": 0.40386045, "grad_norm": 0.03820579, "learning_rate": 3.834e-05, "epoch": 0.5879192, "global_step/max_steps": "1477/2513", "percentage": "58.77%", "elapsed_time": "4h 3m 48s", "remaining_time": "2h 51m 0s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100966} +{"loss": 0.34913149, "grad_norm": 0.03643611, "learning_rate": 3.828e-05, "epoch": 0.58831725, "global_step/max_steps": "1478/2513", "percentage": "58.81%", "elapsed_time": "4h 3m 58s", "remaining_time": "2h 50m 50s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100967} +{"loss": 0.4069643, "grad_norm": 0.03817791, "learning_rate": 3.822e-05, "epoch": 0.5887153, "global_step/max_steps": "1479/2513", "percentage": "58.85%", "elapsed_time": "4h 4m 8s", "remaining_time": "2h 50m 40s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100968} +{"loss": 0.38055563, "grad_norm": 0.04099875, "learning_rate": 3.816e-05, "epoch": 0.58911334, "global_step/max_steps": "1480/2513", "percentage": "58.89%", "elapsed_time": "4h 4m 18s", "remaining_time": "2h 50m 31s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100967} +{"loss": 0.31980118, "grad_norm": 0.03444165, "learning_rate": 3.809e-05, "epoch": 0.58951139, "global_step/max_steps": "1481/2513", "percentage": "58.93%", "elapsed_time": "4h 4m 28s", "remaining_time": "2h 50m 21s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100967} +{"loss": 0.36182103, "grad_norm": 0.03552155, "learning_rate": 3.803e-05, "epoch": 0.58990944, "global_step/max_steps": "1482/2513", "percentage": "58.97%", "elapsed_time": "4h 4m 37s", "remaining_time": "2h 50m 11s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100968} +{"loss": 0.41183114, "grad_norm": 0.03896853, "learning_rate": 3.797e-05, "epoch": 0.59030749, "global_step/max_steps": "1483/2513", "percentage": "59.01%", "elapsed_time": "4h 4m 48s", "remaining_time": "2h 50m 1s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100966} +{"loss": 0.3632004, "grad_norm": 0.03353433, "learning_rate": 3.791e-05, "epoch": 0.59070554, "global_step/max_steps": "1484/2513", "percentage": "59.05%", "elapsed_time": "4h 4m 58s", "remaining_time": "2h 49m 52s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10096} +{"loss": 0.36162415, "grad_norm": 0.0390172, "learning_rate": 3.784e-05, "epoch": 0.59110359, "global_step/max_steps": "1485/2513", "percentage": "59.09%", "elapsed_time": "4h 5m 9s", "remaining_time": "2h 49m 42s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100956} +{"loss": 0.36183882, "grad_norm": 0.03537493, "learning_rate": 3.778e-05, "epoch": 0.59150164, "global_step/max_steps": "1486/2513", "percentage": "59.13%", "elapsed_time": "4h 5m 18s", "remaining_time": "2h 49m 32s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10096} +{"loss": 0.39573267, "grad_norm": 0.03544066, "learning_rate": 3.772e-05, "epoch": 0.59189969, "global_step/max_steps": "1487/2513", "percentage": "59.17%", "elapsed_time": "4h 5m 28s", "remaining_time": "2h 49m 22s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100961} +{"loss": 0.34043145, "grad_norm": 0.03726188, "learning_rate": 3.766e-05, "epoch": 0.59229774, "global_step/max_steps": "1488/2513", "percentage": "59.21%", "elapsed_time": "4h 5m 37s", "remaining_time": "2h 49m 12s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100964} +{"loss": 0.36909539, "grad_norm": 0.03257512, "learning_rate": 3.759e-05, "epoch": 0.59269579, "global_step/max_steps": "1489/2513", "percentage": "59.25%", "elapsed_time": "4h 5m 47s", "remaining_time": "2h 49m 2s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100964} +{"loss": 0.3592093, "grad_norm": 0.03711158, "learning_rate": 3.753e-05, "epoch": 0.59309384, "global_step/max_steps": "1490/2513", "percentage": "59.29%", "elapsed_time": "4h 5m 57s", "remaining_time": "2h 48m 52s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100964} +{"loss": 0.4075599, "grad_norm": 0.03792725, "learning_rate": 3.747e-05, "epoch": 0.59349189, "global_step/max_steps": "1491/2513", "percentage": "59.33%", "elapsed_time": "4h 6m 7s", "remaining_time": "2h 48m 42s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100967} +{"loss": 0.35083982, "grad_norm": 0.03495212, "learning_rate": 3.741e-05, "epoch": 0.59388994, "global_step/max_steps": "1492/2513", "percentage": "59.37%", "elapsed_time": "4h 6m 17s", "remaining_time": "2h 48m 32s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100967} +{"loss": 0.38064867, "grad_norm": 0.03959202, "learning_rate": 3.734e-05, "epoch": 0.59428799, "global_step/max_steps": "1493/2513", "percentage": "59.41%", "elapsed_time": "4h 6m 26s", "remaining_time": "2h 48m 22s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100967} +{"loss": 0.40008307, "grad_norm": 0.03804065, "learning_rate": 3.728e-05, "epoch": 0.59468604, "global_step/max_steps": "1494/2513", "percentage": "59.45%", "elapsed_time": "4h 6m 36s", "remaining_time": "2h 48m 12s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100967} +{"loss": 0.36677662, "grad_norm": 0.03791442, "learning_rate": 3.722e-05, "epoch": 0.59508409, "global_step/max_steps": "1495/2513", "percentage": "59.49%", "elapsed_time": "4h 6m 46s", "remaining_time": "2h 48m 2s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100968} +{"loss": 0.35296899, "grad_norm": 0.07573886, "learning_rate": 3.716e-05, "epoch": 0.59548214, "global_step/max_steps": "1496/2513", "percentage": "59.53%", "elapsed_time": "4h 6m 56s", "remaining_time": "2h 47m 52s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100969} +{"loss": 0.42288888, "grad_norm": 0.04235904, "learning_rate": 3.71e-05, "epoch": 0.59588019, "global_step/max_steps": "1497/2513", "percentage": "59.57%", "elapsed_time": "4h 7m 7s", "remaining_time": "2h 47m 42s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100964} +{"loss": 0.33718145, "grad_norm": 0.03611832, "learning_rate": 3.703e-05, "epoch": 0.59627824, "global_step/max_steps": "1498/2513", "percentage": "59.61%", "elapsed_time": "4h 7m 17s", "remaining_time": "2h 47m 33s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100959} +{"loss": 0.36648691, "grad_norm": 0.04010546, "learning_rate": 3.697e-05, "epoch": 0.59667629, "global_step/max_steps": "1499/2513", "percentage": "59.65%", "elapsed_time": "4h 7m 28s", "remaining_time": "2h 47m 24s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100954} +{"loss": 0.37895593, "grad_norm": 0.03652614, "learning_rate": 3.691e-05, "epoch": 0.59707434, "global_step/max_steps": "1500/2513", "percentage": "59.69%", "elapsed_time": "4h 7m 38s", "remaining_time": "2h 47m 14s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100955} +{"loss": 0.33035508, "grad_norm": 0.03679032, "learning_rate": 3.685e-05, "epoch": 0.59747239, "global_step/max_steps": "1501/2513", "percentage": "59.73%", "elapsed_time": "4h 7m 49s", "remaining_time": "2h 47m 5s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100946} +{"loss": 0.33475915, "grad_norm": 0.03513373, "learning_rate": 3.678e-05, "epoch": 0.59787043, "global_step/max_steps": "1502/2513", "percentage": "59.77%", "elapsed_time": "4h 7m 59s", "remaining_time": "2h 46m 55s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100948} +{"loss": 0.39417312, "grad_norm": 0.0356796, "learning_rate": 3.672e-05, "epoch": 0.59826848, "global_step/max_steps": "1503/2513", "percentage": "59.81%", "elapsed_time": "4h 8m 8s", "remaining_time": "2h 46m 45s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100948} +{"loss": 0.35183981, "grad_norm": 0.03936377, "learning_rate": 3.666e-05, "epoch": 0.59866653, "global_step/max_steps": "1504/2513", "percentage": "59.85%", "elapsed_time": "4h 8m 18s", "remaining_time": "2h 46m 35s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100949} +{"loss": 0.36732501, "grad_norm": 0.03648141, "learning_rate": 3.66e-05, "epoch": 0.59906458, "global_step/max_steps": "1505/2513", "percentage": "59.89%", "elapsed_time": "4h 8m 28s", "remaining_time": "2h 46m 25s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100951} +{"loss": 0.3964535, "grad_norm": 0.03890665, "learning_rate": 3.654e-05, "epoch": 0.59946263, "global_step/max_steps": "1506/2513", "percentage": "59.93%", "elapsed_time": "4h 8m 37s", "remaining_time": "2h 46m 14s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100953} +{"loss": 0.35273749, "grad_norm": 0.03708293, "learning_rate": 3.647e-05, "epoch": 0.59986068, "global_step/max_steps": "1507/2513", "percentage": "59.97%", "elapsed_time": "4h 8m 47s", "remaining_time": "2h 46m 4s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100955} +{"loss": 0.33175606, "grad_norm": 0.04028777, "learning_rate": 3.641e-05, "epoch": 0.60025873, "global_step/max_steps": "1508/2513", "percentage": "60.01%", "elapsed_time": "4h 8m 57s", "remaining_time": "2h 45m 54s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100955} +{"loss": 0.35214457, "grad_norm": 0.035231, "learning_rate": 3.635e-05, "epoch": 0.60065678, "global_step/max_steps": "1509/2513", "percentage": "60.05%", "elapsed_time": "4h 9m 7s", "remaining_time": "2h 45m 45s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100955} +{"loss": 0.3462131, "grad_norm": 0.038315, "learning_rate": 3.629e-05, "epoch": 0.60105483, "global_step/max_steps": "1510/2513", "percentage": "60.09%", "elapsed_time": "4h 9m 17s", "remaining_time": "2h 45m 35s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100956} +{"loss": 0.34019825, "grad_norm": 0.03755053, "learning_rate": 3.623e-05, "epoch": 0.60145288, "global_step/max_steps": "1511/2513", "percentage": "60.13%", "elapsed_time": "4h 9m 27s", "remaining_time": "2h 45m 25s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100954} +{"loss": 0.38927525, "grad_norm": 0.04055936, "learning_rate": 3.616e-05, "epoch": 0.60185093, "global_step/max_steps": "1512/2513", "percentage": "60.17%", "elapsed_time": "4h 9m 37s", "remaining_time": "2h 45m 15s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100953} +{"loss": 0.39452839, "grad_norm": 0.03775851, "learning_rate": 3.61e-05, "epoch": 0.60224898, "global_step/max_steps": "1513/2513", "percentage": "60.21%", "elapsed_time": "4h 9m 48s", "remaining_time": "2h 45m 6s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100947} +{"loss": 0.38824242, "grad_norm": 0.04177283, "learning_rate": 3.604e-05, "epoch": 0.60264703, "global_step/max_steps": "1514/2513", "percentage": "60.25%", "elapsed_time": "4h 9m 58s", "remaining_time": "2h 44m 56s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100943} +{"loss": 0.35585546, "grad_norm": 0.04168914, "learning_rate": 3.598e-05, "epoch": 0.60304508, "global_step/max_steps": "1515/2513", "percentage": "60.29%", "elapsed_time": "4h 10m 8s", "remaining_time": "2h 44m 46s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100946} +{"loss": 0.34383908, "grad_norm": 0.03643807, "learning_rate": 3.592e-05, "epoch": 0.60344313, "global_step/max_steps": "1516/2513", "percentage": "60.33%", "elapsed_time": "4h 10m 17s", "remaining_time": "2h 44m 36s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100949} +{"loss": 0.34994048, "grad_norm": 0.03631479, "learning_rate": 3.585e-05, "epoch": 0.60384118, "global_step/max_steps": "1517/2513", "percentage": "60.37%", "elapsed_time": "4h 10m 27s", "remaining_time": "2h 44m 26s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100951} +{"loss": 0.38504106, "grad_norm": 0.03567856, "learning_rate": 3.579e-05, "epoch": 0.60423923, "global_step/max_steps": "1518/2513", "percentage": "60.41%", "elapsed_time": "4h 10m 36s", "remaining_time": "2h 44m 16s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100951} +{"loss": 0.33750182, "grad_norm": 0.03555704, "learning_rate": 3.573e-05, "epoch": 0.60463728, "global_step/max_steps": "1519/2513", "percentage": "60.45%", "elapsed_time": "4h 10m 46s", "remaining_time": "2h 44m 5s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100955} +{"loss": 0.39423478, "grad_norm": 0.03753847, "learning_rate": 3.567e-05, "epoch": 0.60503533, "global_step/max_steps": "1520/2513", "percentage": "60.49%", "elapsed_time": "4h 10m 55s", "remaining_time": "2h 43m 55s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100957} +{"loss": 0.37369218, "grad_norm": 0.03871876, "learning_rate": 3.561e-05, "epoch": 0.60543338, "global_step/max_steps": "1521/2513", "percentage": "60.53%", "elapsed_time": "4h 11m 5s", "remaining_time": "2h 43m 45s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100959} +{"loss": 0.38148263, "grad_norm": 0.0402595, "learning_rate": 3.555e-05, "epoch": 0.60583143, "global_step/max_steps": "1522/2513", "percentage": "60.57%", "elapsed_time": "4h 11m 15s", "remaining_time": "2h 43m 35s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100961} +{"loss": 0.36272112, "grad_norm": 0.04355194, "learning_rate": 3.548e-05, "epoch": 0.60622948, "global_step/max_steps": "1523/2513", "percentage": "60.60%", "elapsed_time": "4h 11m 25s", "remaining_time": "2h 43m 25s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100961} +{"loss": 0.3532044, "grad_norm": 0.03628937, "learning_rate": 3.542e-05, "epoch": 0.60662753, "global_step/max_steps": "1524/2513", "percentage": "60.64%", "elapsed_time": "4h 11m 34s", "remaining_time": "2h 43m 15s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100963} +{"loss": 0.36282018, "grad_norm": 0.0366938, "learning_rate": 3.536e-05, "epoch": 0.60702557, "global_step/max_steps": "1525/2513", "percentage": "60.68%", "elapsed_time": "4h 11m 44s", "remaining_time": "2h 43m 5s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100965} +{"loss": 0.39458847, "grad_norm": 0.0390285, "learning_rate": 3.53e-05, "epoch": 0.60742362, "global_step/max_steps": "1526/2513", "percentage": "60.72%", "elapsed_time": "4h 11m 54s", "remaining_time": "2h 42m 55s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100963} +{"loss": 0.3780196, "grad_norm": 0.03788117, "learning_rate": 3.524e-05, "epoch": 0.60782167, "global_step/max_steps": "1527/2513", "percentage": "60.76%", "elapsed_time": "4h 12m 4s", "remaining_time": "2h 42m 46s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10096} +{"loss": 0.393053, "grad_norm": 0.03770903, "learning_rate": 3.518e-05, "epoch": 0.60821972, "global_step/max_steps": "1528/2513", "percentage": "60.80%", "elapsed_time": "4h 12m 15s", "remaining_time": "2h 42m 36s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100957} +{"loss": 0.35779375, "grad_norm": 0.03882351, "learning_rate": 3.511e-05, "epoch": 0.60861777, "global_step/max_steps": "1529/2513", "percentage": "60.84%", "elapsed_time": "4h 12m 25s", "remaining_time": "2h 42m 26s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100954} +{"loss": 0.39719585, "grad_norm": 0.03992942, "learning_rate": 3.505e-05, "epoch": 0.60901582, "global_step/max_steps": "1530/2513", "percentage": "60.88%", "elapsed_time": "4h 12m 35s", "remaining_time": "2h 42m 16s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100956} +{"loss": 0.35630018, "grad_norm": 0.03783708, "learning_rate": 3.499e-05, "epoch": 0.60941387, "global_step/max_steps": "1531/2513", "percentage": "60.92%", "elapsed_time": "4h 12m 45s", "remaining_time": "2h 42m 7s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100956} +{"loss": 0.33066615, "grad_norm": 0.03527415, "learning_rate": 3.493e-05, "epoch": 0.60981192, "global_step/max_steps": "1532/2513", "percentage": "60.96%", "elapsed_time": "4h 12m 54s", "remaining_time": "2h 41m 56s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100959} +{"loss": 0.37296903, "grad_norm": 0.03909087, "learning_rate": 3.487e-05, "epoch": 0.61020997, "global_step/max_steps": "1533/2513", "percentage": "61.00%", "elapsed_time": "4h 13m 4s", "remaining_time": "2h 41m 46s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100958} +{"loss": 0.36312398, "grad_norm": 0.04261815, "learning_rate": 3.481e-05, "epoch": 0.61060802, "global_step/max_steps": "1534/2513", "percentage": "61.04%", "elapsed_time": "4h 13m 14s", "remaining_time": "2h 41m 36s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100961} +{"loss": 0.36335972, "grad_norm": 0.04821886, "learning_rate": 3.475e-05, "epoch": 0.61100607, "global_step/max_steps": "1535/2513", "percentage": "61.08%", "elapsed_time": "4h 13m 23s", "remaining_time": "2h 41m 26s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100962} +{"loss": 0.36789638, "grad_norm": 0.04488537, "learning_rate": 3.468e-05, "epoch": 0.61140412, "global_step/max_steps": "1536/2513", "percentage": "61.12%", "elapsed_time": "4h 13m 33s", "remaining_time": "2h 41m 16s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100964} +{"loss": 0.36835268, "grad_norm": 0.03674157, "learning_rate": 3.462e-05, "epoch": 0.61180217, "global_step/max_steps": "1537/2513", "percentage": "61.16%", "elapsed_time": "4h 13m 43s", "remaining_time": "2h 41m 6s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100965} +{"loss": 0.36548218, "grad_norm": 0.0400808, "learning_rate": 3.456e-05, "epoch": 0.61220022, "global_step/max_steps": "1538/2513", "percentage": "61.20%", "elapsed_time": "4h 13m 52s", "remaining_time": "2h 40m 56s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100968} +{"loss": 0.4040086, "grad_norm": 0.03726365, "learning_rate": 3.45e-05, "epoch": 0.61259827, "global_step/max_steps": "1539/2513", "percentage": "61.24%", "elapsed_time": "4h 14m 2s", "remaining_time": "2h 40m 46s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100969} +{"loss": 0.39234787, "grad_norm": 0.03912639, "learning_rate": 3.444e-05, "epoch": 0.61299632, "global_step/max_steps": "1540/2513", "percentage": "61.28%", "elapsed_time": "4h 14m 12s", "remaining_time": "2h 40m 36s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100969} +{"loss": 0.38392362, "grad_norm": 0.03853862, "learning_rate": 3.438e-05, "epoch": 0.61339437, "global_step/max_steps": "1541/2513", "percentage": "61.32%", "elapsed_time": "4h 14m 22s", "remaining_time": "2h 40m 26s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100969} +{"loss": 0.39207363, "grad_norm": 0.0379183, "learning_rate": 3.432e-05, "epoch": 0.61379242, "global_step/max_steps": "1542/2513", "percentage": "61.36%", "elapsed_time": "4h 14m 32s", "remaining_time": "2h 40m 17s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100965} +{"loss": 0.3739506, "grad_norm": 0.03989874, "learning_rate": 3.426e-05, "epoch": 0.61419047, "global_step/max_steps": "1543/2513", "percentage": "61.40%", "elapsed_time": "4h 14m 42s", "remaining_time": "2h 40m 7s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100962} +{"loss": 0.33461636, "grad_norm": 0.03293255, "learning_rate": 3.419e-05, "epoch": 0.61458852, "global_step/max_steps": "1544/2513", "percentage": "61.44%", "elapsed_time": "4h 14m 53s", "remaining_time": "2h 39m 57s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100961} +{"loss": 0.36379361, "grad_norm": 0.03907281, "learning_rate": 3.413e-05, "epoch": 0.61498657, "global_step/max_steps": "1545/2513", "percentage": "61.48%", "elapsed_time": "4h 15m 2s", "remaining_time": "2h 39m 47s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100962} +{"loss": 0.34264791, "grad_norm": 0.03938064, "learning_rate": 3.407e-05, "epoch": 0.61538462, "global_step/max_steps": "1546/2513", "percentage": "61.52%", "elapsed_time": "4h 15m 12s", "remaining_time": "2h 39m 37s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100964} +{"loss": 0.35504484, "grad_norm": 0.04289872, "learning_rate": 3.401e-05, "epoch": 0.61578266, "global_step/max_steps": "1547/2513", "percentage": "61.56%", "elapsed_time": "4h 15m 21s", "remaining_time": "2h 39m 27s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100968} +{"loss": 0.36354321, "grad_norm": 0.0401614, "learning_rate": 3.395e-05, "epoch": 0.61618071, "global_step/max_steps": "1548/2513", "percentage": "61.60%", "elapsed_time": "4h 15m 31s", "remaining_time": "2h 39m 17s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10097} +{"loss": 0.36788747, "grad_norm": 0.03873255, "learning_rate": 3.389e-05, "epoch": 0.61657876, "global_step/max_steps": "1549/2513", "percentage": "61.64%", "elapsed_time": "4h 15m 40s", "remaining_time": "2h 39m 7s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100974} +{"loss": 0.37185338, "grad_norm": 0.04017128, "learning_rate": 3.383e-05, "epoch": 0.61697681, "global_step/max_steps": "1550/2513", "percentage": "61.68%", "elapsed_time": "4h 15m 50s", "remaining_time": "2h 38m 56s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100976} +{"loss": 0.37030068, "grad_norm": 0.03723158, "learning_rate": 3.377e-05, "epoch": 0.61737486, "global_step/max_steps": "1551/2513", "percentage": "61.72%", "elapsed_time": "4h 15m 59s", "remaining_time": "2h 38m 46s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100978} +{"loss": 0.39511293, "grad_norm": 0.03999367, "learning_rate": 3.371e-05, "epoch": 0.61777291, "global_step/max_steps": "1552/2513", "percentage": "61.76%", "elapsed_time": "4h 16m 9s", "remaining_time": "2h 38m 36s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100979} +{"loss": 0.3959403, "grad_norm": 0.0413309, "learning_rate": 3.364e-05, "epoch": 0.61817096, "global_step/max_steps": "1553/2513", "percentage": "61.80%", "elapsed_time": "4h 16m 19s", "remaining_time": "2h 38m 26s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100981} +{"loss": 0.40123606, "grad_norm": 0.04074311, "learning_rate": 3.358e-05, "epoch": 0.61856901, "global_step/max_steps": "1554/2513", "percentage": "61.84%", "elapsed_time": "4h 16m 28s", "remaining_time": "2h 38m 16s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100984} +{"loss": 0.35023776, "grad_norm": 0.03647155, "learning_rate": 3.352e-05, "epoch": 0.61896706, "global_step/max_steps": "1555/2513", "percentage": "61.88%", "elapsed_time": "4h 16m 38s", "remaining_time": "2h 38m 6s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100983} +{"loss": 0.39172196, "grad_norm": 0.04116356, "learning_rate": 3.346e-05, "epoch": 0.61936511, "global_step/max_steps": "1556/2513", "percentage": "61.92%", "elapsed_time": "4h 16m 48s", "remaining_time": "2h 37m 56s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100983} +{"loss": 0.39477575, "grad_norm": 0.04245973, "learning_rate": 3.34e-05, "epoch": 0.61976316, "global_step/max_steps": "1557/2513", "percentage": "61.96%", "elapsed_time": "4h 16m 58s", "remaining_time": "2h 37m 47s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10098} +{"loss": 0.36894682, "grad_norm": 0.04321522, "learning_rate": 3.334e-05, "epoch": 0.62016121, "global_step/max_steps": "1558/2513", "percentage": "62.00%", "elapsed_time": "4h 17m 9s", "remaining_time": "2h 37m 37s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100978} +{"loss": 0.3456648, "grad_norm": 0.03839972, "learning_rate": 3.328e-05, "epoch": 0.62055926, "global_step/max_steps": "1559/2513", "percentage": "62.04%", "elapsed_time": "4h 17m 19s", "remaining_time": "2h 37m 27s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100977} +{"loss": 0.36785117, "grad_norm": 0.04041547, "learning_rate": 3.322e-05, "epoch": 0.62095731, "global_step/max_steps": "1560/2513", "percentage": "62.08%", "elapsed_time": "4h 17m 28s", "remaining_time": "2h 37m 17s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100979} +{"loss": 0.37332669, "grad_norm": 0.0373837, "learning_rate": 3.316e-05, "epoch": 0.62135536, "global_step/max_steps": "1561/2513", "percentage": "62.12%", "elapsed_time": "4h 17m 38s", "remaining_time": "2h 37m 7s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100981} +{"loss": 0.39044139, "grad_norm": 0.03773076, "learning_rate": 3.31e-05, "epoch": 0.62175341, "global_step/max_steps": "1562/2513", "percentage": "62.16%", "elapsed_time": "4h 17m 48s", "remaining_time": "2h 36m 57s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100981} +{"loss": 0.39470798, "grad_norm": 0.03846467, "learning_rate": 3.304e-05, "epoch": 0.62215146, "global_step/max_steps": "1563/2513", "percentage": "62.20%", "elapsed_time": "4h 17m 57s", "remaining_time": "2h 36m 47s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100985} +{"loss": 0.39453796, "grad_norm": 0.03948133, "learning_rate": 3.298e-05, "epoch": 0.62254951, "global_step/max_steps": "1564/2513", "percentage": "62.24%", "elapsed_time": "4h 18m 7s", "remaining_time": "2h 36m 37s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100987} +{"loss": 0.35658464, "grad_norm": 0.0382171, "learning_rate": 3.292e-05, "epoch": 0.62294756, "global_step/max_steps": "1565/2513", "percentage": "62.28%", "elapsed_time": "4h 18m 16s", "remaining_time": "2h 36m 27s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100988} +{"loss": 0.36390218, "grad_norm": 0.03891461, "learning_rate": 3.286e-05, "epoch": 0.62334561, "global_step/max_steps": "1566/2513", "percentage": "62.32%", "elapsed_time": "4h 18m 26s", "remaining_time": "2h 36m 16s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100993} +{"loss": 0.37035024, "grad_norm": 0.03782075, "learning_rate": 3.279e-05, "epoch": 0.62374366, "global_step/max_steps": "1567/2513", "percentage": "62.36%", "elapsed_time": "4h 18m 36s", "remaining_time": "2h 36m 7s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100992} +{"loss": 0.3822028, "grad_norm": 0.04018079, "learning_rate": 3.273e-05, "epoch": 0.62414171, "global_step/max_steps": "1568/2513", "percentage": "62.40%", "elapsed_time": "4h 18m 45s", "remaining_time": "2h 35m 57s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100994} +{"loss": 0.39694425, "grad_norm": 0.03852001, "learning_rate": 3.267e-05, "epoch": 0.62453976, "global_step/max_steps": "1569/2513", "percentage": "62.44%", "elapsed_time": "4h 18m 55s", "remaining_time": "2h 35m 46s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100996} +{"loss": 0.38366309, "grad_norm": 0.0440445, "learning_rate": 3.261e-05, "epoch": 0.6249378, "global_step/max_steps": "1570/2513", "percentage": "62.48%", "elapsed_time": "4h 19m 5s", "remaining_time": "2h 35m 37s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100993} +{"loss": 0.39173558, "grad_norm": 0.04131713, "learning_rate": 3.255e-05, "epoch": 0.62533585, "global_step/max_steps": "1571/2513", "percentage": "62.51%", "elapsed_time": "4h 19m 15s", "remaining_time": "2h 35m 27s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100994} +{"loss": 0.37002724, "grad_norm": 0.03651191, "learning_rate": 3.249e-05, "epoch": 0.6257339, "global_step/max_steps": "1572/2513", "percentage": "62.55%", "elapsed_time": "4h 19m 25s", "remaining_time": "2h 35m 17s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10099} +{"loss": 0.36519819, "grad_norm": 0.03575857, "learning_rate": 3.243e-05, "epoch": 0.62613195, "global_step/max_steps": "1573/2513", "percentage": "62.59%", "elapsed_time": "4h 19m 35s", "remaining_time": "2h 35m 7s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100991} +{"loss": 0.33849251, "grad_norm": 0.03416847, "learning_rate": 3.237e-05, "epoch": 0.62653, "global_step/max_steps": "1574/2513", "percentage": "62.63%", "elapsed_time": "4h 19m 45s", "remaining_time": "2h 34m 57s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100991} +{"loss": 0.37432986, "grad_norm": 0.03898856, "learning_rate": 3.231e-05, "epoch": 0.62692805, "global_step/max_steps": "1575/2513", "percentage": "62.67%", "elapsed_time": "4h 19m 55s", "remaining_time": "2h 34m 47s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100992} +{"loss": 0.36584654, "grad_norm": 0.03862986, "learning_rate": 3.225e-05, "epoch": 0.6273261, "global_step/max_steps": "1576/2513", "percentage": "62.71%", "elapsed_time": "4h 20m 4s", "remaining_time": "2h 34m 37s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100996} +{"loss": 0.37974766, "grad_norm": 0.03882704, "learning_rate": 3.219e-05, "epoch": 0.62772415, "global_step/max_steps": "1577/2513", "percentage": "62.75%", "elapsed_time": "4h 20m 14s", "remaining_time": "2h 34m 27s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100997} +{"loss": 0.34756967, "grad_norm": 0.0379452, "learning_rate": 3.213e-05, "epoch": 0.6281222, "global_step/max_steps": "1578/2513", "percentage": "62.79%", "elapsed_time": "4h 20m 24s", "remaining_time": "2h 34m 17s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100998} +{"loss": 0.36661047, "grad_norm": 0.03903212, "learning_rate": 3.207e-05, "epoch": 0.62852025, "global_step/max_steps": "1579/2513", "percentage": "62.83%", "elapsed_time": "4h 20m 34s", "remaining_time": "2h 34m 7s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100997} +{"loss": 0.40112048, "grad_norm": 0.0384569, "learning_rate": 3.201e-05, "epoch": 0.6289183, "global_step/max_steps": "1580/2513", "percentage": "62.87%", "elapsed_time": "4h 20m 43s", "remaining_time": "2h 33m 57s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101} +{"loss": 0.34700954, "grad_norm": 0.03830873, "learning_rate": 3.195e-05, "epoch": 0.62931635, "global_step/max_steps": "1581/2513", "percentage": "62.91%", "elapsed_time": "4h 20m 52s", "remaining_time": "2h 33m 47s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101003} +{"loss": 0.37663931, "grad_norm": 0.0404099, "learning_rate": 3.189e-05, "epoch": 0.6297144, "global_step/max_steps": "1582/2513", "percentage": "62.95%", "elapsed_time": "4h 21m 2s", "remaining_time": "2h 33m 37s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101005} +{"loss": 0.37491572, "grad_norm": 0.04262237, "learning_rate": 3.183e-05, "epoch": 0.63011245, "global_step/max_steps": "1583/2513", "percentage": "62.99%", "elapsed_time": "4h 21m 12s", "remaining_time": "2h 33m 27s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101006} +{"loss": 0.35837135, "grad_norm": 0.04088216, "learning_rate": 3.177e-05, "epoch": 0.6305105, "global_step/max_steps": "1584/2513", "percentage": "63.03%", "elapsed_time": "4h 21m 22s", "remaining_time": "2h 33m 17s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101004} +{"loss": 0.30990639, "grad_norm": 0.0405705, "learning_rate": 3.171e-05, "epoch": 0.63090855, "global_step/max_steps": "1585/2513", "percentage": "63.07%", "elapsed_time": "4h 21m 32s", "remaining_time": "2h 33m 7s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101006} +{"loss": 0.34537557, "grad_norm": 0.03749521, "learning_rate": 3.165e-05, "epoch": 0.6313066, "global_step/max_steps": "1586/2513", "percentage": "63.11%", "elapsed_time": "4h 21m 42s", "remaining_time": "2h 32m 57s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101003} +{"loss": 0.34633276, "grad_norm": 0.03708453, "learning_rate": 3.159e-05, "epoch": 0.63170465, "global_step/max_steps": "1587/2513", "percentage": "63.15%", "elapsed_time": "4h 21m 53s", "remaining_time": "2h 32m 48s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100999} +{"loss": 0.34020251, "grad_norm": 0.03472205, "learning_rate": 3.153e-05, "epoch": 0.6321027, "global_step/max_steps": "1588/2513", "percentage": "63.19%", "elapsed_time": "4h 22m 2s", "remaining_time": "2h 32m 38s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100999} +{"loss": 0.37348887, "grad_norm": 0.03851137, "learning_rate": 3.147e-05, "epoch": 0.63250075, "global_step/max_steps": "1589/2513", "percentage": "63.23%", "elapsed_time": "4h 22m 12s", "remaining_time": "2h 32m 28s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101} +{"loss": 0.43036011, "grad_norm": 0.0460428, "learning_rate": 3.141e-05, "epoch": 0.6328988, "global_step/max_steps": "1590/2513", "percentage": "63.27%", "elapsed_time": "4h 22m 22s", "remaining_time": "2h 32m 18s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101001} +{"loss": 0.41827506, "grad_norm": 0.04373275, "learning_rate": 3.135e-05, "epoch": 0.63329685, "global_step/max_steps": "1591/2513", "percentage": "63.31%", "elapsed_time": "4h 22m 32s", "remaining_time": "2h 32m 8s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101002} +{"loss": 0.45960158, "grad_norm": 0.04118875, "learning_rate": 3.129e-05, "epoch": 0.6336949, "global_step/max_steps": "1592/2513", "percentage": "63.35%", "elapsed_time": "4h 22m 41s", "remaining_time": "2h 31m 58s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101004} +{"loss": 0.38923526, "grad_norm": 0.04090707, "learning_rate": 3.123e-05, "epoch": 0.63409294, "global_step/max_steps": "1593/2513", "percentage": "63.39%", "elapsed_time": "4h 22m 51s", "remaining_time": "2h 31m 48s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101005} +{"loss": 0.3489958, "grad_norm": 0.04013298, "learning_rate": 3.117e-05, "epoch": 0.63449099, "global_step/max_steps": "1594/2513", "percentage": "63.43%", "elapsed_time": "4h 23m 1s", "remaining_time": "2h 31m 38s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101005} +{"loss": 0.42754441, "grad_norm": 0.03971415, "learning_rate": 3.111e-05, "epoch": 0.63488904, "global_step/max_steps": "1595/2513", "percentage": "63.47%", "elapsed_time": "4h 23m 11s", "remaining_time": "2h 31m 28s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101006} +{"loss": 0.3744359, "grad_norm": 0.03831673, "learning_rate": 3.105e-05, "epoch": 0.63528709, "global_step/max_steps": "1596/2513", "percentage": "63.51%", "elapsed_time": "4h 23m 20s", "remaining_time": "2h 31m 18s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101008} +{"loss": 0.33047438, "grad_norm": 0.03367131, "learning_rate": 3.099e-05, "epoch": 0.63568514, "global_step/max_steps": "1597/2513", "percentage": "63.55%", "elapsed_time": "4h 23m 30s", "remaining_time": "2h 31m 8s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101009} +{"loss": 0.40924203, "grad_norm": 0.03982999, "learning_rate": 3.093e-05, "epoch": 0.63608319, "global_step/max_steps": "1598/2513", "percentage": "63.59%", "elapsed_time": "4h 23m 40s", "remaining_time": "2h 30m 58s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101009} +{"loss": 0.37411255, "grad_norm": 0.03791627, "learning_rate": 3.087e-05, "epoch": 0.63648124, "global_step/max_steps": "1599/2513", "percentage": "63.63%", "elapsed_time": "4h 23m 50s", "remaining_time": "2h 30m 48s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101006} +{"loss": 0.33181834, "grad_norm": 0.03746769, "learning_rate": 3.081e-05, "epoch": 0.63687929, "global_step/max_steps": "1600/2513", "percentage": "63.67%", "elapsed_time": "4h 24m 0s", "remaining_time": "2h 30m 38s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101008} +{"loss": 0.37335387, "grad_norm": 0.03943905, "learning_rate": 3.075e-05, "epoch": 0.63727734, "global_step/max_steps": "1601/2513", "percentage": "63.71%", "elapsed_time": "4h 24m 12s", "remaining_time": "2h 30m 30s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100994} +{"loss": 0.34863734, "grad_norm": 0.0405632, "learning_rate": 3.069e-05, "epoch": 0.63767539, "global_step/max_steps": "1602/2513", "percentage": "63.75%", "elapsed_time": "4h 24m 22s", "remaining_time": "2h 30m 20s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100992} +{"loss": 0.40071848, "grad_norm": 0.04345919, "learning_rate": 3.064e-05, "epoch": 0.63807344, "global_step/max_steps": "1603/2513", "percentage": "63.79%", "elapsed_time": "4h 24m 32s", "remaining_time": "2h 30m 10s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100992} +{"loss": 0.37800312, "grad_norm": 0.03718189, "learning_rate": 3.058e-05, "epoch": 0.63847149, "global_step/max_steps": "1604/2513", "percentage": "63.83%", "elapsed_time": "4h 24m 41s", "remaining_time": "2h 30m 0s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100995} +{"loss": 0.35330874, "grad_norm": 0.0354245, "learning_rate": 3.052e-05, "epoch": 0.63886954, "global_step/max_steps": "1605/2513", "percentage": "63.87%", "elapsed_time": "4h 24m 51s", "remaining_time": "2h 29m 50s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100995} +{"loss": 0.36983031, "grad_norm": 0.04170507, "learning_rate": 3.046e-05, "epoch": 0.63926759, "global_step/max_steps": "1606/2513", "percentage": "63.91%", "elapsed_time": "4h 25m 1s", "remaining_time": "2h 29m 40s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100994} +{"loss": 0.34446695, "grad_norm": 0.03670848, "learning_rate": 3.04e-05, "epoch": 0.63966564, "global_step/max_steps": "1607/2513", "percentage": "63.95%", "elapsed_time": "4h 25m 11s", "remaining_time": "2h 29m 30s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100996} +{"loss": 0.37884668, "grad_norm": 0.03988962, "learning_rate": 3.034e-05, "epoch": 0.64006369, "global_step/max_steps": "1608/2513", "percentage": "63.99%", "elapsed_time": "4h 25m 21s", "remaining_time": "2h 29m 20s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100998} +{"loss": 0.3601526, "grad_norm": 0.0367263, "learning_rate": 3.028e-05, "epoch": 0.64046174, "global_step/max_steps": "1609/2513", "percentage": "64.03%", "elapsed_time": "4h 25m 30s", "remaining_time": "2h 29m 10s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100999} +{"loss": 0.33960003, "grad_norm": 0.03737989, "learning_rate": 3.022e-05, "epoch": 0.64085979, "global_step/max_steps": "1610/2513", "percentage": "64.07%", "elapsed_time": "4h 25m 40s", "remaining_time": "2h 29m 0s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101001} +{"loss": 0.36789399, "grad_norm": 0.03858215, "learning_rate": 3.016e-05, "epoch": 0.64125784, "global_step/max_steps": "1611/2513", "percentage": "64.11%", "elapsed_time": "4h 25m 50s", "remaining_time": "2h 28m 50s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101} +{"loss": 0.36852688, "grad_norm": 0.03728789, "learning_rate": 3.01e-05, "epoch": 0.64165589, "global_step/max_steps": "1612/2513", "percentage": "64.15%", "elapsed_time": "4h 26m 0s", "remaining_time": "2h 28m 40s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100999} +{"loss": 0.34108204, "grad_norm": 0.03798959, "learning_rate": 3.004e-05, "epoch": 0.64205394, "global_step/max_steps": "1613/2513", "percentage": "64.19%", "elapsed_time": "4h 26m 10s", "remaining_time": "2h 28m 30s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101} +{"loss": 0.3347196, "grad_norm": 0.036655, "learning_rate": 2.998e-05, "epoch": 0.64245199, "global_step/max_steps": "1614/2513", "percentage": "64.23%", "elapsed_time": "4h 26m 20s", "remaining_time": "2h 28m 21s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101} +{"loss": 0.36279312, "grad_norm": 0.03473402, "learning_rate": 2.992e-05, "epoch": 0.64285003, "global_step/max_steps": "1615/2513", "percentage": "64.27%", "elapsed_time": "4h 26m 30s", "remaining_time": "2h 28m 11s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100999} +{"loss": 0.34431526, "grad_norm": 0.03585938, "learning_rate": 2.987e-05, "epoch": 0.64324808, "global_step/max_steps": "1616/2513", "percentage": "64.31%", "elapsed_time": "4h 26m 40s", "remaining_time": "2h 28m 1s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100995} +{"loss": 0.32344428, "grad_norm": 0.03710978, "learning_rate": 2.981e-05, "epoch": 0.64364613, "global_step/max_steps": "1617/2513", "percentage": "64.35%", "elapsed_time": "4h 26m 50s", "remaining_time": "2h 27m 51s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100996} +{"loss": 0.33458808, "grad_norm": 0.0381574, "learning_rate": 2.975e-05, "epoch": 0.64404418, "global_step/max_steps": "1618/2513", "percentage": "64.39%", "elapsed_time": "4h 27m 0s", "remaining_time": "2h 27m 41s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100994} +{"loss": 0.39929196, "grad_norm": 0.03921615, "learning_rate": 2.969e-05, "epoch": 0.64444223, "global_step/max_steps": "1619/2513", "percentage": "64.42%", "elapsed_time": "4h 27m 10s", "remaining_time": "2h 27m 31s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100995} +{"loss": 0.38812518, "grad_norm": 0.03822768, "learning_rate": 2.963e-05, "epoch": 0.64484028, "global_step/max_steps": "1620/2513", "percentage": "64.46%", "elapsed_time": "4h 27m 20s", "remaining_time": "2h 27m 22s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100994} +{"loss": 0.39688736, "grad_norm": 0.04180869, "learning_rate": 2.957e-05, "epoch": 0.64523833, "global_step/max_steps": "1621/2513", "percentage": "64.50%", "elapsed_time": "4h 27m 30s", "remaining_time": "2h 27m 12s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100994} +{"loss": 0.34579521, "grad_norm": 0.04116908, "learning_rate": 2.951e-05, "epoch": 0.64563638, "global_step/max_steps": "1622/2513", "percentage": "64.54%", "elapsed_time": "4h 27m 39s", "remaining_time": "2h 27m 2s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100997} +{"loss": 0.37222159, "grad_norm": 0.04168343, "learning_rate": 2.945e-05, "epoch": 0.64603443, "global_step/max_steps": "1623/2513", "percentage": "64.58%", "elapsed_time": "4h 27m 49s", "remaining_time": "2h 26m 52s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100997} +{"loss": 0.41384387, "grad_norm": 0.04367976, "learning_rate": 2.939e-05, "epoch": 0.64643248, "global_step/max_steps": "1624/2513", "percentage": "64.62%", "elapsed_time": "4h 27m 59s", "remaining_time": "2h 26m 42s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100999} +{"loss": 0.36632547, "grad_norm": 0.03929461, "learning_rate": 2.934e-05, "epoch": 0.64683053, "global_step/max_steps": "1625/2513", "percentage": "64.66%", "elapsed_time": "4h 28m 8s", "remaining_time": "2h 26m 31s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101002} +{"loss": 0.35642809, "grad_norm": 0.03611796, "learning_rate": 2.928e-05, "epoch": 0.64722858, "global_step/max_steps": "1626/2513", "percentage": "64.70%", "elapsed_time": "4h 28m 18s", "remaining_time": "2h 26m 21s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101004} +{"loss": 0.35962564, "grad_norm": 0.03878482, "learning_rate": 2.922e-05, "epoch": 0.64762663, "global_step/max_steps": "1627/2513", "percentage": "64.74%", "elapsed_time": "4h 28m 28s", "remaining_time": "2h 26m 11s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101005} +{"loss": 0.41412386, "grad_norm": 0.04113202, "learning_rate": 2.916e-05, "epoch": 0.64802468, "global_step/max_steps": "1628/2513", "percentage": "64.78%", "elapsed_time": "4h 28m 38s", "remaining_time": "2h 26m 1s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101004} +{"loss": 0.35877639, "grad_norm": 0.03652216, "learning_rate": 2.91e-05, "epoch": 0.64842273, "global_step/max_steps": "1629/2513", "percentage": "64.82%", "elapsed_time": "4h 28m 47s", "remaining_time": "2h 25m 51s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101006} +{"loss": 0.39685392, "grad_norm": 0.03887513, "learning_rate": 2.904e-05, "epoch": 0.64882078, "global_step/max_steps": "1630/2513", "percentage": "64.86%", "elapsed_time": "4h 28m 58s", "remaining_time": "2h 25m 42s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101004} +{"loss": 0.39990556, "grad_norm": 0.04162888, "learning_rate": 2.898e-05, "epoch": 0.64921883, "global_step/max_steps": "1631/2513", "percentage": "64.90%", "elapsed_time": "4h 29m 9s", "remaining_time": "2h 25m 32s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100997} +{"loss": 0.35807765, "grad_norm": 0.0346882, "learning_rate": 2.893e-05, "epoch": 0.64961688, "global_step/max_steps": "1632/2513", "percentage": "64.94%", "elapsed_time": "4h 29m 18s", "remaining_time": "2h 25m 22s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100998} +{"loss": 0.38522372, "grad_norm": 0.04036905, "learning_rate": 2.887e-05, "epoch": 0.65001493, "global_step/max_steps": "1633/2513", "percentage": "64.98%", "elapsed_time": "4h 29m 28s", "remaining_time": "2h 25m 13s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100997} +{"loss": 0.34281713, "grad_norm": 0.03903314, "learning_rate": 2.881e-05, "epoch": 0.65041298, "global_step/max_steps": "1634/2513", "percentage": "65.02%", "elapsed_time": "4h 29m 38s", "remaining_time": "2h 25m 3s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.100998} +{"loss": 0.38619012, "grad_norm": 0.03856281, "learning_rate": 2.875e-05, "epoch": 0.65081103, "global_step/max_steps": "1635/2513", "percentage": "65.06%", "elapsed_time": "4h 29m 48s", "remaining_time": "2h 24m 53s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101} +{"loss": 0.41080421, "grad_norm": 0.04331513, "learning_rate": 2.869e-05, "epoch": 0.65120908, "global_step/max_steps": "1636/2513", "percentage": "65.10%", "elapsed_time": "4h 29m 57s", "remaining_time": "2h 24m 43s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101001} +{"loss": 0.32044229, "grad_norm": 0.03729568, "learning_rate": 2.863e-05, "epoch": 0.65160713, "global_step/max_steps": "1637/2513", "percentage": "65.14%", "elapsed_time": "4h 30m 7s", "remaining_time": "2h 24m 33s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101002} +{"loss": 0.36403358, "grad_norm": 0.03599924, "learning_rate": 2.858e-05, "epoch": 0.65200517, "global_step/max_steps": "1638/2513", "percentage": "65.18%", "elapsed_time": "4h 30m 17s", "remaining_time": "2h 24m 23s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101003} +{"loss": 0.332268, "grad_norm": 0.03464879, "learning_rate": 2.852e-05, "epoch": 0.65240322, "global_step/max_steps": "1639/2513", "percentage": "65.22%", "elapsed_time": "4h 30m 27s", "remaining_time": "2h 24m 13s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101003} +{"loss": 0.41394335, "grad_norm": 0.03970818, "learning_rate": 2.846e-05, "epoch": 0.65280127, "global_step/max_steps": "1640/2513", "percentage": "65.26%", "elapsed_time": "4h 30m 36s", "remaining_time": "2h 24m 3s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101005} +{"loss": 0.37546599, "grad_norm": 0.03748483, "learning_rate": 2.84e-05, "epoch": 0.65319932, "global_step/max_steps": "1641/2513", "percentage": "65.30%", "elapsed_time": "4h 30m 46s", "remaining_time": "2h 23m 53s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101006} +{"loss": 0.34531429, "grad_norm": 0.0385652, "learning_rate": 2.834e-05, "epoch": 0.65359737, "global_step/max_steps": "1642/2513", "percentage": "65.34%", "elapsed_time": "4h 30m 56s", "remaining_time": "2h 23m 43s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101008} +{"loss": 0.40652072, "grad_norm": 0.04268194, "learning_rate": 2.829e-05, "epoch": 0.65399542, "global_step/max_steps": "1643/2513", "percentage": "65.38%", "elapsed_time": "4h 31m 6s", "remaining_time": "2h 23m 33s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101007} +{"loss": 0.37890726, "grad_norm": 0.03809291, "learning_rate": 2.823e-05, "epoch": 0.65439347, "global_step/max_steps": "1644/2513", "percentage": "65.42%", "elapsed_time": "4h 31m 15s", "remaining_time": "2h 23m 23s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101009} +{"loss": 0.34563494, "grad_norm": 0.03777844, "learning_rate": 2.817e-05, "epoch": 0.65479152, "global_step/max_steps": "1645/2513", "percentage": "65.46%", "elapsed_time": "4h 31m 26s", "remaining_time": "2h 23m 13s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101004} +{"loss": 0.39303678, "grad_norm": 0.04272969, "learning_rate": 2.811e-05, "epoch": 0.65518957, "global_step/max_steps": "1646/2513", "percentage": "65.50%", "elapsed_time": "4h 31m 36s", "remaining_time": "2h 23m 3s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101003} +{"loss": 0.35381448, "grad_norm": 0.03759883, "learning_rate": 2.805e-05, "epoch": 0.65558762, "global_step/max_steps": "1647/2513", "percentage": "65.54%", "elapsed_time": "4h 31m 46s", "remaining_time": "2h 22m 53s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101004} +{"loss": 0.37197113, "grad_norm": 0.0423496, "learning_rate": 2.8e-05, "epoch": 0.65598567, "global_step/max_steps": "1648/2513", "percentage": "65.58%", "elapsed_time": "4h 31m 56s", "remaining_time": "2h 22m 44s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101003} +{"loss": 0.33250934, "grad_norm": 0.03713213, "learning_rate": 2.794e-05, "epoch": 0.65638372, "global_step/max_steps": "1649/2513", "percentage": "65.62%", "elapsed_time": "4h 32m 5s", "remaining_time": "2h 22m 33s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101006} +{"loss": 0.39621997, "grad_norm": 0.04230755, "learning_rate": 2.788e-05, "epoch": 0.65678177, "global_step/max_steps": "1650/2513", "percentage": "65.66%", "elapsed_time": "4h 32m 15s", "remaining_time": "2h 22m 23s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101008} +{"loss": 0.42060459, "grad_norm": 0.04357388, "learning_rate": 2.782e-05, "epoch": 0.65717982, "global_step/max_steps": "1651/2513", "percentage": "65.70%", "elapsed_time": "4h 32m 25s", "remaining_time": "2h 22m 13s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101009} +{"loss": 0.37576786, "grad_norm": 0.03783972, "learning_rate": 2.776e-05, "epoch": 0.65757787, "global_step/max_steps": "1652/2513", "percentage": "65.74%", "elapsed_time": "4h 32m 34s", "remaining_time": "2h 22m 3s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101013} +{"loss": 0.3686347, "grad_norm": 0.0392834, "learning_rate": 2.771e-05, "epoch": 0.65797592, "global_step/max_steps": "1653/2513", "percentage": "65.78%", "elapsed_time": "4h 32m 43s", "remaining_time": "2h 21m 53s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101016} +{"loss": 0.30996573, "grad_norm": 0.03892113, "learning_rate": 2.765e-05, "epoch": 0.65837397, "global_step/max_steps": "1654/2513", "percentage": "65.82%", "elapsed_time": "4h 32m 53s", "remaining_time": "2h 21m 43s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101018} +{"loss": 0.35090464, "grad_norm": 0.03663349, "learning_rate": 2.759e-05, "epoch": 0.65877202, "global_step/max_steps": "1655/2513", "percentage": "65.86%", "elapsed_time": "4h 33m 2s", "remaining_time": "2h 21m 33s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10102} +{"loss": 0.38814625, "grad_norm": 0.03985255, "learning_rate": 2.753e-05, "epoch": 0.65917007, "global_step/max_steps": "1656/2513", "percentage": "65.90%", "elapsed_time": "4h 33m 12s", "remaining_time": "2h 21m 23s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101019} +{"loss": 0.37154725, "grad_norm": 0.0399961, "learning_rate": 2.748e-05, "epoch": 0.65956812, "global_step/max_steps": "1657/2513", "percentage": "65.94%", "elapsed_time": "4h 33m 23s", "remaining_time": "2h 21m 13s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101018} +{"loss": 0.35475475, "grad_norm": 0.03998553, "learning_rate": 2.742e-05, "epoch": 0.65996617, "global_step/max_steps": "1658/2513", "percentage": "65.98%", "elapsed_time": "4h 33m 33s", "remaining_time": "2h 21m 4s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101016} +{"loss": 0.38041323, "grad_norm": 0.03905061, "learning_rate": 2.736e-05, "epoch": 0.66036422, "global_step/max_steps": "1659/2513", "percentage": "66.02%", "elapsed_time": "4h 33m 42s", "remaining_time": "2h 20m 53s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101018} +{"loss": 0.38736016, "grad_norm": 0.03677526, "learning_rate": 2.73e-05, "epoch": 0.66076226, "global_step/max_steps": "1660/2513", "percentage": "66.06%", "elapsed_time": "4h 33m 53s", "remaining_time": "2h 20m 44s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10101} +{"loss": 0.337327, "grad_norm": 0.03675432, "learning_rate": 2.725e-05, "epoch": 0.66116031, "global_step/max_steps": "1661/2513", "percentage": "66.10%", "elapsed_time": "4h 34m 3s", "remaining_time": "2h 20m 34s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10101} +{"loss": 0.36478263, "grad_norm": 0.04081826, "learning_rate": 2.719e-05, "epoch": 0.66155836, "global_step/max_steps": "1662/2513", "percentage": "66.14%", "elapsed_time": "4h 34m 13s", "remaining_time": "2h 20m 24s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101012} +{"loss": 0.37575978, "grad_norm": 0.03990385, "learning_rate": 2.713e-05, "epoch": 0.66195641, "global_step/max_steps": "1663/2513", "percentage": "66.18%", "elapsed_time": "4h 34m 23s", "remaining_time": "2h 20m 14s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101011} +{"loss": 0.39164755, "grad_norm": 0.03746461, "learning_rate": 2.707e-05, "epoch": 0.66235446, "global_step/max_steps": "1664/2513", "percentage": "66.22%", "elapsed_time": "4h 34m 33s", "remaining_time": "2h 20m 4s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101012} +{"loss": 0.40180117, "grad_norm": 0.04268776, "learning_rate": 2.702e-05, "epoch": 0.66275251, "global_step/max_steps": "1665/2513", "percentage": "66.26%", "elapsed_time": "4h 34m 43s", "remaining_time": "2h 19m 54s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101013} +{"loss": 0.36567646, "grad_norm": 0.04058706, "learning_rate": 2.696e-05, "epoch": 0.66315056, "global_step/max_steps": "1666/2513", "percentage": "66.30%", "elapsed_time": "4h 34m 53s", "remaining_time": "2h 19m 45s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101012} +{"loss": 0.35151401, "grad_norm": 0.03574344, "learning_rate": 2.69e-05, "epoch": 0.66354861, "global_step/max_steps": "1667/2513", "percentage": "66.34%", "elapsed_time": "4h 35m 2s", "remaining_time": "2h 19m 35s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101013} +{"loss": 0.35783851, "grad_norm": 0.03956663, "learning_rate": 2.685e-05, "epoch": 0.66394666, "global_step/max_steps": "1668/2513", "percentage": "66.37%", "elapsed_time": "4h 35m 12s", "remaining_time": "2h 19m 25s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101015} +{"loss": 0.39122915, "grad_norm": 0.04120494, "learning_rate": 2.679e-05, "epoch": 0.66434471, "global_step/max_steps": "1669/2513", "percentage": "66.41%", "elapsed_time": "4h 35m 22s", "remaining_time": "2h 19m 15s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101015} +{"loss": 0.38768435, "grad_norm": 0.04040757, "learning_rate": 2.673e-05, "epoch": 0.66474276, "global_step/max_steps": "1670/2513", "percentage": "66.45%", "elapsed_time": "4h 35m 31s", "remaining_time": "2h 19m 5s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101016} +{"loss": 0.35812706, "grad_norm": 0.03793013, "learning_rate": 2.667e-05, "epoch": 0.66514081, "global_step/max_steps": "1671/2513", "percentage": "66.49%", "elapsed_time": "4h 35m 41s", "remaining_time": "2h 18m 55s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101016} +{"loss": 0.37794426, "grad_norm": 0.03875937, "learning_rate": 2.662e-05, "epoch": 0.66553886, "global_step/max_steps": "1672/2513", "percentage": "66.53%", "elapsed_time": "4h 35m 51s", "remaining_time": "2h 18m 45s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101015} +{"loss": 0.37504044, "grad_norm": 0.04057793, "learning_rate": 2.656e-05, "epoch": 0.66593691, "global_step/max_steps": "1673/2513", "percentage": "66.57%", "elapsed_time": "4h 36m 1s", "remaining_time": "2h 18m 35s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101015} +{"loss": 0.34376818, "grad_norm": 0.03890877, "learning_rate": 2.65e-05, "epoch": 0.66633496, "global_step/max_steps": "1674/2513", "percentage": "66.61%", "elapsed_time": "4h 36m 11s", "remaining_time": "2h 18m 25s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101015} +{"loss": 0.39865416, "grad_norm": 0.03767243, "learning_rate": 2.645e-05, "epoch": 0.66673301, "global_step/max_steps": "1675/2513", "percentage": "66.65%", "elapsed_time": "4h 36m 22s", "remaining_time": "2h 18m 16s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101009} +{"loss": 0.38955808, "grad_norm": 0.04067006, "learning_rate": 2.639e-05, "epoch": 0.66713106, "global_step/max_steps": "1676/2513", "percentage": "66.69%", "elapsed_time": "4h 36m 32s", "remaining_time": "2h 18m 6s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101009} +{"loss": 0.3503094, "grad_norm": 0.04008522, "learning_rate": 2.633e-05, "epoch": 0.66752911, "global_step/max_steps": "1677/2513", "percentage": "66.73%", "elapsed_time": "4h 36m 42s", "remaining_time": "2h 17m 56s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101008} +{"loss": 0.33471161, "grad_norm": 0.03913618, "learning_rate": 2.628e-05, "epoch": 0.66792716, "global_step/max_steps": "1678/2513", "percentage": "66.77%", "elapsed_time": "4h 36m 52s", "remaining_time": "2h 17m 46s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101009} +{"loss": 0.33756086, "grad_norm": 0.03750346, "learning_rate": 2.622e-05, "epoch": 0.66832521, "global_step/max_steps": "1679/2513", "percentage": "66.81%", "elapsed_time": "4h 37m 2s", "remaining_time": "2h 17m 36s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10101} +{"loss": 0.39282456, "grad_norm": 0.0398236, "learning_rate": 2.616e-05, "epoch": 0.66872326, "global_step/max_steps": "1680/2513", "percentage": "66.85%", "elapsed_time": "4h 37m 11s", "remaining_time": "2h 17m 26s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101013} +{"loss": 0.36724785, "grad_norm": 0.03555205, "learning_rate": 2.611e-05, "epoch": 0.66912131, "global_step/max_steps": "1681/2513", "percentage": "66.89%", "elapsed_time": "4h 37m 21s", "remaining_time": "2h 17m 16s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101014} +{"loss": 0.32578641, "grad_norm": 0.03672085, "learning_rate": 2.605e-05, "epoch": 0.66951936, "global_step/max_steps": "1682/2513", "percentage": "66.93%", "elapsed_time": "4h 37m 31s", "remaining_time": "2h 17m 6s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101014} +{"loss": 0.33327672, "grad_norm": 0.03772152, "learning_rate": 2.599e-05, "epoch": 0.6699174, "global_step/max_steps": "1683/2513", "percentage": "66.97%", "elapsed_time": "4h 37m 40s", "remaining_time": "2h 16m 56s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101014} +{"loss": 0.40556836, "grad_norm": 0.04259647, "learning_rate": 2.594e-05, "epoch": 0.67031545, "global_step/max_steps": "1684/2513", "percentage": "67.01%", "elapsed_time": "4h 37m 50s", "remaining_time": "2h 16m 46s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101016} +{"loss": 0.33025688, "grad_norm": 0.03611214, "learning_rate": 2.588e-05, "epoch": 0.6707135, "global_step/max_steps": "1685/2513", "percentage": "67.05%", "elapsed_time": "4h 38m 0s", "remaining_time": "2h 16m 36s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101017} +{"loss": 0.38295314, "grad_norm": 0.03850259, "learning_rate": 2.582e-05, "epoch": 0.67111155, "global_step/max_steps": "1686/2513", "percentage": "67.09%", "elapsed_time": "4h 38m 9s", "remaining_time": "2h 16m 26s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10102} +{"loss": 0.38121805, "grad_norm": 0.03909293, "learning_rate": 2.577e-05, "epoch": 0.6715096, "global_step/max_steps": "1687/2513", "percentage": "67.13%", "elapsed_time": "4h 38m 19s", "remaining_time": "2h 16m 16s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101019} +{"loss": 0.34039992, "grad_norm": 0.03808339, "learning_rate": 2.571e-05, "epoch": 0.67190765, "global_step/max_steps": "1688/2513", "percentage": "67.17%", "elapsed_time": "4h 38m 29s", "remaining_time": "2h 16m 6s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101018} +{"loss": 0.38779733, "grad_norm": 0.03730431, "learning_rate": 2.565e-05, "epoch": 0.6723057, "global_step/max_steps": "1689/2513", "percentage": "67.21%", "elapsed_time": "4h 38m 39s", "remaining_time": "2h 15m 56s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101021} +{"loss": 0.34276545, "grad_norm": 0.04059694, "learning_rate": 2.56e-05, "epoch": 0.67270375, "global_step/max_steps": "1690/2513", "percentage": "67.25%", "elapsed_time": "4h 38m 50s", "remaining_time": "2h 15m 47s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101015} +{"loss": 0.36955801, "grad_norm": 0.03816649, "learning_rate": 2.554e-05, "epoch": 0.6731018, "global_step/max_steps": "1691/2513", "percentage": "67.29%", "elapsed_time": "4h 38m 59s", "remaining_time": "2h 15m 37s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101018} +{"loss": 0.36844257, "grad_norm": 0.0423115, "learning_rate": 2.549e-05, "epoch": 0.67349985, "global_step/max_steps": "1692/2513", "percentage": "67.33%", "elapsed_time": "4h 39m 9s", "remaining_time": "2h 15m 27s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101019} +{"loss": 0.37610084, "grad_norm": 0.03740901, "learning_rate": 2.543e-05, "epoch": 0.6738979, "global_step/max_steps": "1693/2513", "percentage": "67.37%", "elapsed_time": "4h 39m 19s", "remaining_time": "2h 15m 17s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101018} +{"loss": 0.36372715, "grad_norm": 0.03731059, "learning_rate": 2.537e-05, "epoch": 0.67429595, "global_step/max_steps": "1694/2513", "percentage": "67.41%", "elapsed_time": "4h 39m 29s", "remaining_time": "2h 15m 7s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10102} +{"loss": 0.32387352, "grad_norm": 0.0348171, "learning_rate": 2.532e-05, "epoch": 0.674694, "global_step/max_steps": "1695/2513", "percentage": "67.45%", "elapsed_time": "4h 39m 38s", "remaining_time": "2h 14m 57s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101022} +{"loss": 0.37851012, "grad_norm": 0.03870973, "learning_rate": 2.526e-05, "epoch": 0.67509205, "global_step/max_steps": "1696/2513", "percentage": "67.49%", "elapsed_time": "4h 39m 47s", "remaining_time": "2h 14m 47s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101025} +{"loss": 0.35799712, "grad_norm": 0.03745406, "learning_rate": 2.52e-05, "epoch": 0.6754901, "global_step/max_steps": "1697/2513", "percentage": "67.53%", "elapsed_time": "4h 39m 57s", "remaining_time": "2h 14m 37s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101025} +{"loss": 0.3526001, "grad_norm": 0.03942648, "learning_rate": 2.515e-05, "epoch": 0.67588815, "global_step/max_steps": "1698/2513", "percentage": "67.57%", "elapsed_time": "4h 40m 7s", "remaining_time": "2h 14m 27s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101024} +{"loss": 0.40692759, "grad_norm": 0.04284486, "learning_rate": 2.509e-05, "epoch": 0.6762862, "global_step/max_steps": "1699/2513", "percentage": "67.61%", "elapsed_time": "4h 40m 17s", "remaining_time": "2h 14m 17s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101026} +{"loss": 0.36663929, "grad_norm": 0.04011741, "learning_rate": 2.504e-05, "epoch": 0.67668425, "global_step/max_steps": "1700/2513", "percentage": "67.65%", "elapsed_time": "4h 40m 27s", "remaining_time": "2h 14m 7s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101028} +{"loss": 0.41603261, "grad_norm": 0.04463996, "learning_rate": 2.498e-05, "epoch": 0.6770823, "global_step/max_steps": "1701/2513", "percentage": "67.69%", "elapsed_time": "4h 40m 38s", "remaining_time": "2h 13m 58s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101018} +{"loss": 0.33445114, "grad_norm": 0.03701621, "learning_rate": 2.493e-05, "epoch": 0.67748035, "global_step/max_steps": "1702/2513", "percentage": "67.73%", "elapsed_time": "4h 40m 48s", "remaining_time": "2h 13m 48s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101019} +{"loss": 0.39529878, "grad_norm": 0.03748686, "learning_rate": 2.487e-05, "epoch": 0.6778784, "global_step/max_steps": "1703/2513", "percentage": "67.77%", "elapsed_time": "4h 40m 57s", "remaining_time": "2h 13m 38s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10102} +{"loss": 0.42135257, "grad_norm": 0.04364387, "learning_rate": 2.481e-05, "epoch": 0.67827645, "global_step/max_steps": "1704/2513", "percentage": "67.81%", "elapsed_time": "4h 41m 8s", "remaining_time": "2h 13m 28s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101019} +{"loss": 0.37974128, "grad_norm": 0.04096739, "learning_rate": 2.476e-05, "epoch": 0.67867449, "global_step/max_steps": "1705/2513", "percentage": "67.85%", "elapsed_time": "4h 41m 19s", "remaining_time": "2h 13m 19s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101012} +{"loss": 0.36499029, "grad_norm": 0.03626869, "learning_rate": 2.47e-05, "epoch": 0.67907254, "global_step/max_steps": "1706/2513", "percentage": "67.89%", "elapsed_time": "4h 41m 28s", "remaining_time": "2h 13m 8s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101015} +{"loss": 0.33662844, "grad_norm": 0.03633578, "learning_rate": 2.465e-05, "epoch": 0.67947059, "global_step/max_steps": "1707/2513", "percentage": "67.93%", "elapsed_time": "4h 41m 38s", "remaining_time": "2h 12m 58s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101016} +{"loss": 0.3673788, "grad_norm": 0.03855757, "learning_rate": 2.459e-05, "epoch": 0.67986864, "global_step/max_steps": "1708/2513", "percentage": "67.97%", "elapsed_time": "4h 41m 47s", "remaining_time": "2h 12m 48s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101018} +{"loss": 0.37156409, "grad_norm": 0.03889152, "learning_rate": 2.454e-05, "epoch": 0.68026669, "global_step/max_steps": "1709/2513", "percentage": "68.01%", "elapsed_time": "4h 41m 57s", "remaining_time": "2h 12m 38s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101021} +{"loss": 0.36864406, "grad_norm": 0.03882059, "learning_rate": 2.448e-05, "epoch": 0.68066474, "global_step/max_steps": "1710/2513", "percentage": "68.05%", "elapsed_time": "4h 42m 6s", "remaining_time": "2h 12m 28s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101022} +{"loss": 0.33472297, "grad_norm": 0.03643123, "learning_rate": 2.443e-05, "epoch": 0.68106279, "global_step/max_steps": "1711/2513", "percentage": "68.09%", "elapsed_time": "4h 42m 16s", "remaining_time": "2h 12m 18s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101022} +{"loss": 0.37367082, "grad_norm": 0.04131015, "learning_rate": 2.437e-05, "epoch": 0.68146084, "global_step/max_steps": "1712/2513", "percentage": "68.13%", "elapsed_time": "4h 42m 26s", "remaining_time": "2h 12m 8s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101023} +{"loss": 0.41095272, "grad_norm": 0.03865982, "learning_rate": 2.431e-05, "epoch": 0.68185889, "global_step/max_steps": "1713/2513", "percentage": "68.17%", "elapsed_time": "4h 42m 35s", "remaining_time": "2h 11m 58s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101026} +{"loss": 0.4054299, "grad_norm": 0.04096902, "learning_rate": 2.426e-05, "epoch": 0.68225694, "global_step/max_steps": "1714/2513", "percentage": "68.21%", "elapsed_time": "4h 42m 45s", "remaining_time": "2h 11m 48s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101029} +{"loss": 0.342053, "grad_norm": 0.03989183, "learning_rate": 2.42e-05, "epoch": 0.68265499, "global_step/max_steps": "1715/2513", "percentage": "68.25%", "elapsed_time": "4h 42m 55s", "remaining_time": "2h 11m 38s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10103} +{"loss": 0.38889518, "grad_norm": 0.038575, "learning_rate": 2.415e-05, "epoch": 0.68305304, "global_step/max_steps": "1716/2513", "percentage": "68.28%", "elapsed_time": "4h 43m 5s", "remaining_time": "2h 11m 28s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101029} +{"loss": 0.37262499, "grad_norm": 0.03987692, "learning_rate": 2.409e-05, "epoch": 0.68345109, "global_step/max_steps": "1717/2513", "percentage": "68.32%", "elapsed_time": "4h 43m 15s", "remaining_time": "2h 11m 18s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101028} +{"loss": 0.39424062, "grad_norm": 0.03865157, "learning_rate": 2.404e-05, "epoch": 0.68384914, "global_step/max_steps": "1718/2513", "percentage": "68.36%", "elapsed_time": "4h 43m 25s", "remaining_time": "2h 11m 9s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101029} +{"loss": 0.3339681, "grad_norm": 0.03579866, "learning_rate": 2.398e-05, "epoch": 0.68424719, "global_step/max_steps": "1719/2513", "percentage": "68.40%", "elapsed_time": "4h 43m 36s", "remaining_time": "2h 10m 59s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101022} +{"loss": 0.36899966, "grad_norm": 0.03928339, "learning_rate": 2.393e-05, "epoch": 0.68464524, "global_step/max_steps": "1720/2513", "percentage": "68.44%", "elapsed_time": "4h 43m 46s", "remaining_time": "2h 10m 49s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101021} +{"loss": 0.37091765, "grad_norm": 0.04209876, "learning_rate": 2.387e-05, "epoch": 0.68504329, "global_step/max_steps": "1721/2513", "percentage": "68.48%", "elapsed_time": "4h 43m 56s", "remaining_time": "2h 10m 40s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10102} +{"loss": 0.35617763, "grad_norm": 0.03772846, "learning_rate": 2.382e-05, "epoch": 0.68544134, "global_step/max_steps": "1722/2513", "percentage": "68.52%", "elapsed_time": "4h 44m 5s", "remaining_time": "2h 10m 29s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101022} +{"loss": 0.34399015, "grad_norm": 0.03754446, "learning_rate": 2.376e-05, "epoch": 0.68583939, "global_step/max_steps": "1723/2513", "percentage": "68.56%", "elapsed_time": "4h 44m 15s", "remaining_time": "2h 10m 20s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101022} +{"loss": 0.38143772, "grad_norm": 0.03641704, "learning_rate": 2.371e-05, "epoch": 0.68623744, "global_step/max_steps": "1724/2513", "percentage": "68.60%", "elapsed_time": "4h 44m 25s", "remaining_time": "2h 10m 10s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101022} +{"loss": 0.35302213, "grad_norm": 0.03922261, "learning_rate": 2.365e-05, "epoch": 0.68663549, "global_step/max_steps": "1725/2513", "percentage": "68.64%", "elapsed_time": "4h 44m 35s", "remaining_time": "2h 10m 0s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101022} +{"loss": 0.34373856, "grad_norm": 0.0377318, "learning_rate": 2.36e-05, "epoch": 0.68703354, "global_step/max_steps": "1726/2513", "percentage": "68.68%", "elapsed_time": "4h 44m 45s", "remaining_time": "2h 9m 50s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101022} +{"loss": 0.34589103, "grad_norm": 0.04049447, "learning_rate": 2.354e-05, "epoch": 0.68743159, "global_step/max_steps": "1727/2513", "percentage": "68.72%", "elapsed_time": "4h 44m 55s", "remaining_time": "2h 9m 40s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101023} +{"loss": 0.40712196, "grad_norm": 0.04055978, "learning_rate": 2.349e-05, "epoch": 0.68782963, "global_step/max_steps": "1728/2513", "percentage": "68.76%", "elapsed_time": "4h 45m 4s", "remaining_time": "2h 9m 30s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101024} +{"loss": 0.38450766, "grad_norm": 0.03705246, "learning_rate": 2.344e-05, "epoch": 0.68822768, "global_step/max_steps": "1729/2513", "percentage": "68.80%", "elapsed_time": "4h 45m 14s", "remaining_time": "2h 9m 20s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101026} +{"loss": 0.36765724, "grad_norm": 0.03788842, "learning_rate": 2.338e-05, "epoch": 0.68862573, "global_step/max_steps": "1730/2513", "percentage": "68.84%", "elapsed_time": "4h 45m 24s", "remaining_time": "2h 9m 10s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101025} +{"loss": 0.33265209, "grad_norm": 0.03846188, "learning_rate": 2.333e-05, "epoch": 0.68902378, "global_step/max_steps": "1731/2513", "percentage": "68.88%", "elapsed_time": "4h 45m 34s", "remaining_time": "2h 9m 0s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101024} +{"loss": 0.33946782, "grad_norm": 0.03826766, "learning_rate": 2.327e-05, "epoch": 0.68942183, "global_step/max_steps": "1732/2513", "percentage": "68.92%", "elapsed_time": "4h 45m 44s", "remaining_time": "2h 8m 50s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101024} +{"loss": 0.35594052, "grad_norm": 0.04001828, "learning_rate": 2.322e-05, "epoch": 0.68981988, "global_step/max_steps": "1733/2513", "percentage": "68.96%", "elapsed_time": "4h 45m 54s", "remaining_time": "2h 8m 40s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101025} +{"loss": 0.39923346, "grad_norm": 0.04015152, "learning_rate": 2.316e-05, "epoch": 0.69021793, "global_step/max_steps": "1734/2513", "percentage": "69.00%", "elapsed_time": "4h 46m 4s", "remaining_time": "2h 8m 31s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101021} +{"loss": 0.36918914, "grad_norm": 0.04043986, "learning_rate": 2.311e-05, "epoch": 0.69061598, "global_step/max_steps": "1735/2513", "percentage": "69.04%", "elapsed_time": "4h 46m 14s", "remaining_time": "2h 8m 21s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101023} +{"loss": 0.35840827, "grad_norm": 0.03745337, "learning_rate": 2.305e-05, "epoch": 0.69101403, "global_step/max_steps": "1736/2513", "percentage": "69.08%", "elapsed_time": "4h 46m 24s", "remaining_time": "2h 8m 11s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101023} +{"loss": 0.40225488, "grad_norm": 0.03791191, "learning_rate": 2.3e-05, "epoch": 0.69141208, "global_step/max_steps": "1737/2513", "percentage": "69.12%", "elapsed_time": "4h 46m 34s", "remaining_time": "2h 8m 1s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101023} +{"loss": 0.3335543, "grad_norm": 0.0362825, "learning_rate": 2.295e-05, "epoch": 0.69181013, "global_step/max_steps": "1738/2513", "percentage": "69.16%", "elapsed_time": "4h 46m 43s", "remaining_time": "2h 7m 51s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101024} +{"loss": 0.38916606, "grad_norm": 0.04297907, "learning_rate": 2.289e-05, "epoch": 0.69220818, "global_step/max_steps": "1739/2513", "percentage": "69.20%", "elapsed_time": "4h 46m 53s", "remaining_time": "2h 7m 41s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101024} +{"loss": 0.34901023, "grad_norm": 0.04204021, "learning_rate": 2.284e-05, "epoch": 0.69260623, "global_step/max_steps": "1740/2513", "percentage": "69.24%", "elapsed_time": "4h 47m 3s", "remaining_time": "2h 7m 31s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101025} +{"loss": 0.38676226, "grad_norm": 0.04050739, "learning_rate": 2.278e-05, "epoch": 0.69300428, "global_step/max_steps": "1741/2513", "percentage": "69.28%", "elapsed_time": "4h 47m 13s", "remaining_time": "2h 7m 21s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101026} +{"loss": 0.3527596, "grad_norm": 0.03960122, "learning_rate": 2.273e-05, "epoch": 0.69340233, "global_step/max_steps": "1742/2513", "percentage": "69.32%", "elapsed_time": "4h 47m 22s", "remaining_time": "2h 7m 11s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101028} +{"loss": 0.37321407, "grad_norm": 0.03816335, "learning_rate": 2.268e-05, "epoch": 0.69380038, "global_step/max_steps": "1743/2513", "percentage": "69.36%", "elapsed_time": "4h 47m 32s", "remaining_time": "2h 7m 1s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10103} +{"loss": 0.35742301, "grad_norm": 0.0353569, "learning_rate": 2.262e-05, "epoch": 0.69419843, "global_step/max_steps": "1744/2513", "percentage": "69.40%", "elapsed_time": "4h 47m 41s", "remaining_time": "2h 6m 51s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101035} +{"loss": 0.36755186, "grad_norm": 0.03778437, "learning_rate": 2.257e-05, "epoch": 0.69459648, "global_step/max_steps": "1745/2513", "percentage": "69.44%", "elapsed_time": "4h 47m 51s", "remaining_time": "2h 6m 41s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101035} +{"loss": 0.40731531, "grad_norm": 0.04141267, "learning_rate": 2.251e-05, "epoch": 0.69499453, "global_step/max_steps": "1746/2513", "percentage": "69.48%", "elapsed_time": "4h 48m 1s", "remaining_time": "2h 6m 31s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101032} +{"loss": 0.40930805, "grad_norm": 0.04073383, "learning_rate": 2.246e-05, "epoch": 0.69539258, "global_step/max_steps": "1747/2513", "percentage": "69.52%", "elapsed_time": "4h 48m 11s", "remaining_time": "2h 6m 21s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101033} +{"loss": 0.36926103, "grad_norm": 0.03627899, "learning_rate": 2.241e-05, "epoch": 0.69579063, "global_step/max_steps": "1748/2513", "percentage": "69.56%", "elapsed_time": "4h 48m 21s", "remaining_time": "2h 6m 11s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10103} +{"loss": 0.37798733, "grad_norm": 0.03761037, "learning_rate": 2.235e-05, "epoch": 0.69618868, "global_step/max_steps": "1749/2513", "percentage": "69.60%", "elapsed_time": "4h 48m 32s", "remaining_time": "2h 6m 2s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101026} +{"loss": 0.38563278, "grad_norm": 0.03608972, "learning_rate": 2.23e-05, "epoch": 0.69658673, "global_step/max_steps": "1750/2513", "percentage": "69.64%", "elapsed_time": "4h 48m 42s", "remaining_time": "2h 5m 52s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101025} +{"loss": 0.37910593, "grad_norm": 0.03897486, "learning_rate": 2.224e-05, "epoch": 0.69698477, "global_step/max_steps": "1751/2513", "percentage": "69.68%", "elapsed_time": "4h 48m 52s", "remaining_time": "2h 5m 42s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101026} +{"loss": 0.3545416, "grad_norm": 0.03610627, "learning_rate": 2.219e-05, "epoch": 0.69738282, "global_step/max_steps": "1752/2513", "percentage": "69.72%", "elapsed_time": "4h 49m 1s", "remaining_time": "2h 5m 32s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101028} +{"loss": 0.35215077, "grad_norm": 0.03736172, "learning_rate": 2.214e-05, "epoch": 0.69778087, "global_step/max_steps": "1753/2513", "percentage": "69.76%", "elapsed_time": "4h 49m 11s", "remaining_time": "2h 5m 22s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101031} +{"loss": 0.34322196, "grad_norm": 0.03900706, "learning_rate": 2.208e-05, "epoch": 0.69817892, "global_step/max_steps": "1754/2513", "percentage": "69.80%", "elapsed_time": "4h 49m 21s", "remaining_time": "2h 5m 12s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10103} +{"loss": 0.40238816, "grad_norm": 0.03709559, "learning_rate": 2.203e-05, "epoch": 0.69857697, "global_step/max_steps": "1755/2513", "percentage": "69.84%", "elapsed_time": "4h 49m 30s", "remaining_time": "2h 5m 2s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101033} +{"loss": 0.32320744, "grad_norm": 0.03663081, "learning_rate": 2.198e-05, "epoch": 0.69897502, "global_step/max_steps": "1756/2513", "percentage": "69.88%", "elapsed_time": "4h 49m 40s", "remaining_time": "2h 4m 52s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101034} +{"loss": 0.36835384, "grad_norm": 0.04275472, "learning_rate": 2.192e-05, "epoch": 0.69937307, "global_step/max_steps": "1757/2513", "percentage": "69.92%", "elapsed_time": "4h 49m 49s", "remaining_time": "2h 4m 42s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101036} +{"loss": 0.38563326, "grad_norm": 0.03509533, "learning_rate": 2.187e-05, "epoch": 0.69977112, "global_step/max_steps": "1758/2513", "percentage": "69.96%", "elapsed_time": "4h 49m 59s", "remaining_time": "2h 4m 32s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101036} +{"loss": 0.38056272, "grad_norm": 0.03920912, "learning_rate": 2.182e-05, "epoch": 0.70016917, "global_step/max_steps": "1759/2513", "percentage": "70.00%", "elapsed_time": "4h 50m 8s", "remaining_time": "2h 4m 22s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10104} +{"loss": 0.35117185, "grad_norm": 0.03540643, "learning_rate": 2.176e-05, "epoch": 0.70056722, "global_step/max_steps": "1760/2513", "percentage": "70.04%", "elapsed_time": "4h 50m 19s", "remaining_time": "2h 4m 12s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101038} +{"loss": 0.38909143, "grad_norm": 0.04158143, "learning_rate": 2.171e-05, "epoch": 0.70096527, "global_step/max_steps": "1761/2513", "percentage": "70.08%", "elapsed_time": "4h 50m 28s", "remaining_time": "2h 4m 2s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101039} +{"loss": 0.42399383, "grad_norm": 0.04063628, "learning_rate": 2.166e-05, "epoch": 0.70136332, "global_step/max_steps": "1762/2513", "percentage": "70.12%", "elapsed_time": "4h 50m 38s", "remaining_time": "2h 3m 52s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101041} +{"loss": 0.3346025, "grad_norm": 0.03459314, "learning_rate": 2.16e-05, "epoch": 0.70176137, "global_step/max_steps": "1763/2513", "percentage": "70.16%", "elapsed_time": "4h 50m 49s", "remaining_time": "2h 3m 43s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101037} +{"loss": 0.36880231, "grad_norm": 0.03827954, "learning_rate": 2.155e-05, "epoch": 0.70215942, "global_step/max_steps": "1764/2513", "percentage": "70.19%", "elapsed_time": "4h 50m 59s", "remaining_time": "2h 3m 33s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101036} +{"loss": 0.3721042, "grad_norm": 0.0382896, "learning_rate": 2.15e-05, "epoch": 0.70255747, "global_step/max_steps": "1765/2513", "percentage": "70.23%", "elapsed_time": "4h 51m 9s", "remaining_time": "2h 3m 23s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101036} +{"loss": 0.39900473, "grad_norm": 0.03912022, "learning_rate": 2.145e-05, "epoch": 0.70295552, "global_step/max_steps": "1766/2513", "percentage": "70.27%", "elapsed_time": "4h 51m 18s", "remaining_time": "2h 3m 13s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101038} +{"loss": 0.34453005, "grad_norm": 0.04076134, "learning_rate": 2.139e-05, "epoch": 0.70335357, "global_step/max_steps": "1767/2513", "percentage": "70.31%", "elapsed_time": "4h 51m 28s", "remaining_time": "2h 3m 3s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101039} +{"loss": 0.38108423, "grad_norm": 0.03963088, "learning_rate": 2.134e-05, "epoch": 0.70375162, "global_step/max_steps": "1768/2513", "percentage": "70.35%", "elapsed_time": "4h 51m 37s", "remaining_time": "2h 2m 53s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10104} +{"loss": 0.41983587, "grad_norm": 0.04214012, "learning_rate": 2.129e-05, "epoch": 0.70414967, "global_step/max_steps": "1769/2513", "percentage": "70.39%", "elapsed_time": "4h 51m 47s", "remaining_time": "2h 2m 43s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101043} +{"loss": 0.35811085, "grad_norm": 0.03692876, "learning_rate": 2.123e-05, "epoch": 0.70454772, "global_step/max_steps": "1770/2513", "percentage": "70.43%", "elapsed_time": "4h 51m 56s", "remaining_time": "2h 2m 33s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101047} +{"loss": 0.34532464, "grad_norm": 0.03860535, "learning_rate": 2.118e-05, "epoch": 0.70494577, "global_step/max_steps": "1771/2513", "percentage": "70.47%", "elapsed_time": "4h 52m 6s", "remaining_time": "2h 2m 23s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101046} +{"loss": 0.36013481, "grad_norm": 0.03945192, "learning_rate": 2.113e-05, "epoch": 0.70534382, "global_step/max_steps": "1772/2513", "percentage": "70.51%", "elapsed_time": "4h 52m 16s", "remaining_time": "2h 2m 13s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101047} +{"loss": 0.37384105, "grad_norm": 0.04258168, "learning_rate": 2.108e-05, "epoch": 0.70574186, "global_step/max_steps": "1773/2513", "percentage": "70.55%", "elapsed_time": "4h 52m 26s", "remaining_time": "2h 2m 3s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101047} +{"loss": 0.39233699, "grad_norm": 0.04354592, "learning_rate": 2.102e-05, "epoch": 0.70613991, "global_step/max_steps": "1774/2513", "percentage": "70.59%", "elapsed_time": "4h 52m 36s", "remaining_time": "2h 1m 53s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101048} +{"loss": 0.34916127, "grad_norm": 0.03617363, "learning_rate": 2.097e-05, "epoch": 0.70653796, "global_step/max_steps": "1775/2513", "percentage": "70.63%", "elapsed_time": "4h 52m 45s", "remaining_time": "2h 1m 43s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101049} +{"loss": 0.3509599, "grad_norm": 0.03611675, "learning_rate": 2.092e-05, "epoch": 0.70693601, "global_step/max_steps": "1776/2513", "percentage": "70.67%", "elapsed_time": "4h 52m 55s", "remaining_time": "2h 1m 33s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101047} +{"loss": 0.37690383, "grad_norm": 0.03994339, "learning_rate": 2.087e-05, "epoch": 0.70733406, "global_step/max_steps": "1777/2513", "percentage": "70.71%", "elapsed_time": "4h 53m 5s", "remaining_time": "2h 1m 23s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101047} +{"loss": 0.39262581, "grad_norm": 0.04250378, "learning_rate": 2.081e-05, "epoch": 0.70773211, "global_step/max_steps": "1778/2513", "percentage": "70.75%", "elapsed_time": "4h 53m 16s", "remaining_time": "2h 1m 14s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101043} +{"loss": 0.35077327, "grad_norm": 0.04850718, "learning_rate": 2.076e-05, "epoch": 0.70813016, "global_step/max_steps": "1779/2513", "percentage": "70.79%", "elapsed_time": "4h 53m 26s", "remaining_time": "2h 1m 4s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101042} +{"loss": 0.37103581, "grad_norm": 0.03729432, "learning_rate": 2.071e-05, "epoch": 0.70852821, "global_step/max_steps": "1780/2513", "percentage": "70.83%", "elapsed_time": "4h 53m 36s", "remaining_time": "2h 0m 54s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101042} +{"loss": 0.34160873, "grad_norm": 0.03880646, "learning_rate": 2.066e-05, "epoch": 0.70892626, "global_step/max_steps": "1781/2513", "percentage": "70.87%", "elapsed_time": "4h 53m 46s", "remaining_time": "2h 0m 44s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101043} +{"loss": 0.40955335, "grad_norm": 0.04018359, "learning_rate": 2.061e-05, "epoch": 0.70932431, "global_step/max_steps": "1782/2513", "percentage": "70.91%", "elapsed_time": "4h 53m 55s", "remaining_time": "2h 0m 34s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101047} +{"loss": 0.41789895, "grad_norm": 0.04189721, "learning_rate": 2.055e-05, "epoch": 0.70972236, "global_step/max_steps": "1783/2513", "percentage": "70.95%", "elapsed_time": "4h 54m 4s", "remaining_time": "2h 0m 24s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101049} +{"loss": 0.35929736, "grad_norm": 0.03927404, "learning_rate": 2.05e-05, "epoch": 0.71012041, "global_step/max_steps": "1784/2513", "percentage": "70.99%", "elapsed_time": "4h 54m 14s", "remaining_time": "2h 0m 14s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101049} +{"loss": 0.35312706, "grad_norm": 0.03834073, "learning_rate": 2.045e-05, "epoch": 0.71051846, "global_step/max_steps": "1785/2513", "percentage": "71.03%", "elapsed_time": "4h 54m 24s", "remaining_time": "2h 0m 4s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101051} +{"loss": 0.36717188, "grad_norm": 0.04015695, "learning_rate": 2.04e-05, "epoch": 0.71091651, "global_step/max_steps": "1786/2513", "percentage": "71.07%", "elapsed_time": "4h 54m 34s", "remaining_time": "1h 59m 54s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101052} +{"loss": 0.35131428, "grad_norm": 0.03731744, "learning_rate": 2.035e-05, "epoch": 0.71131456, "global_step/max_steps": "1787/2513", "percentage": "71.11%", "elapsed_time": "4h 54m 43s", "remaining_time": "1h 59m 44s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101054} +{"loss": 0.32992893, "grad_norm": 0.03625372, "learning_rate": 2.029e-05, "epoch": 0.71171261, "global_step/max_steps": "1788/2513", "percentage": "71.15%", "elapsed_time": "4h 54m 53s", "remaining_time": "1h 59m 34s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101056} +{"loss": 0.38970241, "grad_norm": 0.03708371, "learning_rate": 2.024e-05, "epoch": 0.71211066, "global_step/max_steps": "1789/2513", "percentage": "71.19%", "elapsed_time": "4h 55m 3s", "remaining_time": "1h 59m 24s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101056} +{"loss": 0.40430444, "grad_norm": 0.0432088, "learning_rate": 2.019e-05, "epoch": 0.71250871, "global_step/max_steps": "1790/2513", "percentage": "71.23%", "elapsed_time": "4h 55m 12s", "remaining_time": "1h 59m 14s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101057} +{"loss": 0.38373882, "grad_norm": 0.04147463, "learning_rate": 2.014e-05, "epoch": 0.71290676, "global_step/max_steps": "1791/2513", "percentage": "71.27%", "elapsed_time": "4h 55m 22s", "remaining_time": "1h 59m 4s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101058} +{"loss": 0.39703143, "grad_norm": 0.0407105, "learning_rate": 2.009e-05, "epoch": 0.71330481, "global_step/max_steps": "1792/2513", "percentage": "71.31%", "elapsed_time": "4h 55m 32s", "remaining_time": "1h 58m 54s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101056} +{"loss": 0.33006203, "grad_norm": 0.03695577, "learning_rate": 2.003e-05, "epoch": 0.71370286, "global_step/max_steps": "1793/2513", "percentage": "71.35%", "elapsed_time": "4h 55m 42s", "remaining_time": "1h 58m 44s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101055} +{"loss": 0.38934562, "grad_norm": 0.03888919, "learning_rate": 1.998e-05, "epoch": 0.71410091, "global_step/max_steps": "1794/2513", "percentage": "71.39%", "elapsed_time": "4h 55m 53s", "remaining_time": "1h 58m 35s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101053} +{"loss": 0.35393673, "grad_norm": 0.03818859, "learning_rate": 1.993e-05, "epoch": 0.71449896, "global_step/max_steps": "1795/2513", "percentage": "71.43%", "elapsed_time": "4h 56m 3s", "remaining_time": "1h 58m 25s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101052} +{"loss": 0.37780899, "grad_norm": 0.04022615, "learning_rate": 1.988e-05, "epoch": 0.714897, "global_step/max_steps": "1796/2513", "percentage": "71.47%", "elapsed_time": "4h 56m 12s", "remaining_time": "1h 58m 15s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101054} +{"loss": 0.39242059, "grad_norm": 0.04117887, "learning_rate": 1.983e-05, "epoch": 0.71529505, "global_step/max_steps": "1797/2513", "percentage": "71.51%", "elapsed_time": "4h 56m 22s", "remaining_time": "1h 58m 5s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101057} +{"loss": 0.34969938, "grad_norm": 0.03661471, "learning_rate": 1.978e-05, "epoch": 0.7156931, "global_step/max_steps": "1798/2513", "percentage": "71.55%", "elapsed_time": "4h 56m 32s", "remaining_time": "1h 57m 55s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101056} +{"loss": 0.33093539, "grad_norm": 0.03934725, "learning_rate": 1.973e-05, "epoch": 0.71609115, "global_step/max_steps": "1799/2513", "percentage": "71.59%", "elapsed_time": "4h 56m 41s", "remaining_time": "1h 57m 45s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101059} +{"loss": 0.35199416, "grad_norm": 0.03857031, "learning_rate": 1.967e-05, "epoch": 0.7164892, "global_step/max_steps": "1800/2513", "percentage": "71.63%", "elapsed_time": "4h 56m 51s", "remaining_time": "1h 57m 35s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10106} +{"loss": 0.37082911, "grad_norm": 0.03787124, "learning_rate": 1.962e-05, "epoch": 0.71688725, "global_step/max_steps": "1801/2513", "percentage": "71.67%", "elapsed_time": "4h 57m 2s", "remaining_time": "1h 57m 25s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101053} +{"loss": 0.41362002, "grad_norm": 0.04250321, "learning_rate": 1.957e-05, "epoch": 0.7172853, "global_step/max_steps": "1802/2513", "percentage": "71.71%", "elapsed_time": "4h 57m 11s", "remaining_time": "1h 57m 15s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101055} +{"loss": 0.38556227, "grad_norm": 0.04424244, "learning_rate": 1.952e-05, "epoch": 0.71768335, "global_step/max_steps": "1803/2513", "percentage": "71.75%", "elapsed_time": "4h 57m 21s", "remaining_time": "1h 57m 5s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101056} +{"loss": 0.34056082, "grad_norm": 0.03953006, "learning_rate": 1.947e-05, "epoch": 0.7180814, "global_step/max_steps": "1804/2513", "percentage": "71.79%", "elapsed_time": "4h 57m 31s", "remaining_time": "1h 56m 55s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101055} +{"loss": 0.38728324, "grad_norm": 0.04074144, "learning_rate": 1.942e-05, "epoch": 0.71847945, "global_step/max_steps": "1805/2513", "percentage": "71.83%", "elapsed_time": "4h 57m 41s", "remaining_time": "1h 56m 45s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101057} +{"loss": 0.37379441, "grad_norm": 0.0390771, "learning_rate": 1.937e-05, "epoch": 0.7188775, "global_step/max_steps": "1806/2513", "percentage": "71.87%", "elapsed_time": "4h 57m 51s", "remaining_time": "1h 56m 36s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101055} +{"loss": 0.38787037, "grad_norm": 0.03868654, "learning_rate": 1.932e-05, "epoch": 0.71927555, "global_step/max_steps": "1807/2513", "percentage": "71.91%", "elapsed_time": "4h 58m 1s", "remaining_time": "1h 56m 26s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101054} +{"loss": 0.3896558, "grad_norm": 0.03914303, "learning_rate": 1.927e-05, "epoch": 0.7196736, "global_step/max_steps": "1808/2513", "percentage": "71.95%", "elapsed_time": "4h 58m 11s", "remaining_time": "1h 56m 16s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101053} +{"loss": 0.34955633, "grad_norm": 0.03914408, "learning_rate": 1.922e-05, "epoch": 0.72007165, "global_step/max_steps": "1809/2513", "percentage": "71.99%", "elapsed_time": "4h 58m 21s", "remaining_time": "1h 56m 6s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101053} +{"loss": 0.37548244, "grad_norm": 0.04025864, "learning_rate": 1.917e-05, "epoch": 0.7204697, "global_step/max_steps": "1810/2513", "percentage": "72.03%", "elapsed_time": "4h 58m 31s", "remaining_time": "1h 55m 56s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101053} +{"loss": 0.3833428, "grad_norm": 0.0369743, "learning_rate": 1.911e-05, "epoch": 0.72086775, "global_step/max_steps": "1811/2513", "percentage": "72.07%", "elapsed_time": "4h 58m 41s", "remaining_time": "1h 55m 46s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101052} +{"loss": 0.37075204, "grad_norm": 0.04008234, "learning_rate": 1.906e-05, "epoch": 0.7212658, "global_step/max_steps": "1812/2513", "percentage": "72.11%", "elapsed_time": "4h 58m 51s", "remaining_time": "1h 55m 36s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101053} +{"loss": 0.37993902, "grad_norm": 0.04029291, "learning_rate": 1.901e-05, "epoch": 0.72166385, "global_step/max_steps": "1813/2513", "percentage": "72.14%", "elapsed_time": "4h 59m 0s", "remaining_time": "1h 55m 26s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101057} +{"loss": 0.35922405, "grad_norm": 0.03980331, "learning_rate": 1.896e-05, "epoch": 0.7220619, "global_step/max_steps": "1814/2513", "percentage": "72.18%", "elapsed_time": "4h 59m 10s", "remaining_time": "1h 55m 16s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101058} +{"loss": 0.3176159, "grad_norm": 0.03837536, "learning_rate": 1.891e-05, "epoch": 0.72245995, "global_step/max_steps": "1815/2513", "percentage": "72.22%", "elapsed_time": "4h 59m 19s", "remaining_time": "1h 55m 6s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101059} +{"loss": 0.35177201, "grad_norm": 0.03776906, "learning_rate": 1.886e-05, "epoch": 0.722858, "global_step/max_steps": "1816/2513", "percentage": "72.26%", "elapsed_time": "4h 59m 29s", "remaining_time": "1h 54m 56s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10106} +{"loss": 0.35345674, "grad_norm": 0.03722289, "learning_rate": 1.881e-05, "epoch": 0.72325605, "global_step/max_steps": "1817/2513", "percentage": "72.30%", "elapsed_time": "4h 59m 39s", "remaining_time": "1h 54m 47s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10106} +{"loss": 0.36935577, "grad_norm": 0.03791772, "learning_rate": 1.876e-05, "epoch": 0.72365409, "global_step/max_steps": "1818/2513", "percentage": "72.34%", "elapsed_time": "4h 59m 49s", "remaining_time": "1h 54m 37s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101061} +{"loss": 0.38714874, "grad_norm": 0.0400701, "learning_rate": 1.871e-05, "epoch": 0.72405214, "global_step/max_steps": "1819/2513", "percentage": "72.38%", "elapsed_time": "4h 59m 59s", "remaining_time": "1h 54m 27s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10106} +{"loss": 0.34742707, "grad_norm": 0.03880936, "learning_rate": 1.866e-05, "epoch": 0.72445019, "global_step/max_steps": "1820/2513", "percentage": "72.42%", "elapsed_time": "5h 0m 8s", "remaining_time": "1h 54m 17s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101062} +{"loss": 0.35624254, "grad_norm": 0.03707651, "learning_rate": 1.861e-05, "epoch": 0.72484824, "global_step/max_steps": "1821/2513", "percentage": "72.46%", "elapsed_time": "5h 0m 18s", "remaining_time": "1h 54m 7s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10106} +{"loss": 0.38619131, "grad_norm": 0.03874494, "learning_rate": 1.856e-05, "epoch": 0.72524629, "global_step/max_steps": "1822/2513", "percentage": "72.50%", "elapsed_time": "5h 0m 29s", "remaining_time": "1h 53m 57s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101059} +{"loss": 0.35902989, "grad_norm": 0.03993434, "learning_rate": 1.851e-05, "epoch": 0.72564434, "global_step/max_steps": "1823/2513", "percentage": "72.54%", "elapsed_time": "5h 0m 39s", "remaining_time": "1h 53m 47s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101056} +{"loss": 0.37589458, "grad_norm": 0.04415081, "learning_rate": 1.846e-05, "epoch": 0.72604239, "global_step/max_steps": "1824/2513", "percentage": "72.58%", "elapsed_time": "5h 0m 49s", "remaining_time": "1h 53m 37s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101057} +{"loss": 0.36447322, "grad_norm": 0.03700173, "learning_rate": 1.841e-05, "epoch": 0.72644044, "global_step/max_steps": "1825/2513", "percentage": "72.62%", "elapsed_time": "5h 0m 58s", "remaining_time": "1h 53m 27s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101058} +{"loss": 0.36553437, "grad_norm": 0.03852186, "learning_rate": 1.836e-05, "epoch": 0.72683849, "global_step/max_steps": "1826/2513", "percentage": "72.66%", "elapsed_time": "5h 1m 8s", "remaining_time": "1h 53m 18s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101057} +{"loss": 0.39745116, "grad_norm": 0.03713828, "learning_rate": 1.831e-05, "epoch": 0.72723654, "global_step/max_steps": "1827/2513", "percentage": "72.70%", "elapsed_time": "5h 1m 18s", "remaining_time": "1h 53m 8s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10106} +{"loss": 0.37314034, "grad_norm": 0.03959654, "learning_rate": 1.826e-05, "epoch": 0.72763459, "global_step/max_steps": "1828/2513", "percentage": "72.74%", "elapsed_time": "5h 1m 28s", "remaining_time": "1h 52m 58s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101059} +{"loss": 0.41604489, "grad_norm": 0.04043249, "learning_rate": 1.821e-05, "epoch": 0.72803264, "global_step/max_steps": "1829/2513", "percentage": "72.78%", "elapsed_time": "5h 1m 38s", "remaining_time": "1h 52m 48s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101061} +{"loss": 0.37108803, "grad_norm": 0.03775241, "learning_rate": 1.816e-05, "epoch": 0.72843069, "global_step/max_steps": "1830/2513", "percentage": "72.82%", "elapsed_time": "5h 1m 47s", "remaining_time": "1h 52m 38s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101061} +{"loss": 0.32959175, "grad_norm": 0.03517871, "learning_rate": 1.811e-05, "epoch": 0.72882874, "global_step/max_steps": "1831/2513", "percentage": "72.86%", "elapsed_time": "5h 1m 57s", "remaining_time": "1h 52m 28s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101062} +{"loss": 0.40543675, "grad_norm": 0.04140364, "learning_rate": 1.806e-05, "epoch": 0.72922679, "global_step/max_steps": "1832/2513", "percentage": "72.90%", "elapsed_time": "5h 2m 7s", "remaining_time": "1h 52m 18s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101064} +{"loss": 0.31942064, "grad_norm": 0.03858238, "learning_rate": 1.801e-05, "epoch": 0.72962484, "global_step/max_steps": "1833/2513", "percentage": "72.94%", "elapsed_time": "5h 2m 17s", "remaining_time": "1h 52m 8s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101063} +{"loss": 0.36630699, "grad_norm": 0.03760737, "learning_rate": 1.796e-05, "epoch": 0.73002289, "global_step/max_steps": "1834/2513", "percentage": "72.98%", "elapsed_time": "5h 2m 27s", "remaining_time": "1h 51m 58s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101062} +{"loss": 0.34710792, "grad_norm": 0.04073143, "learning_rate": 1.791e-05, "epoch": 0.73042094, "global_step/max_steps": "1835/2513", "percentage": "73.02%", "elapsed_time": "5h 2m 36s", "remaining_time": "1h 51m 48s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101063} +{"loss": 0.36890906, "grad_norm": 0.04342727, "learning_rate": 1.786e-05, "epoch": 0.73081899, "global_step/max_steps": "1836/2513", "percentage": "73.06%", "elapsed_time": "5h 2m 46s", "remaining_time": "1h 51m 38s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101063} +{"loss": 0.36565033, "grad_norm": 0.04161928, "learning_rate": 1.781e-05, "epoch": 0.73121704, "global_step/max_steps": "1837/2513", "percentage": "73.10%", "elapsed_time": "5h 2m 57s", "remaining_time": "1h 51m 29s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10106} +{"loss": 0.37575477, "grad_norm": 0.03813613, "learning_rate": 1.776e-05, "epoch": 0.73161509, "global_step/max_steps": "1838/2513", "percentage": "73.14%", "elapsed_time": "5h 3m 7s", "remaining_time": "1h 51m 19s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101056} +{"loss": 0.3881017, "grad_norm": 0.04207354, "learning_rate": 1.772e-05, "epoch": 0.73201314, "global_step/max_steps": "1839/2513", "percentage": "73.18%", "elapsed_time": "5h 3m 17s", "remaining_time": "1h 51m 9s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101059} +{"loss": 0.34133539, "grad_norm": 0.03740626, "learning_rate": 1.767e-05, "epoch": 0.73241119, "global_step/max_steps": "1840/2513", "percentage": "73.22%", "elapsed_time": "5h 3m 27s", "remaining_time": "1h 50m 59s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101058} +{"loss": 0.33870956, "grad_norm": 0.0378768, "learning_rate": 1.762e-05, "epoch": 0.73280923, "global_step/max_steps": "1841/2513", "percentage": "73.26%", "elapsed_time": "5h 3m 37s", "remaining_time": "1h 50m 49s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101058} +{"loss": 0.36646724, "grad_norm": 0.0405163, "learning_rate": 1.757e-05, "epoch": 0.73320728, "global_step/max_steps": "1842/2513", "percentage": "73.30%", "elapsed_time": "5h 3m 46s", "remaining_time": "1h 50m 39s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10106} +{"loss": 0.39348623, "grad_norm": 0.04170314, "learning_rate": 1.752e-05, "epoch": 0.73360533, "global_step/max_steps": "1843/2513", "percentage": "73.34%", "elapsed_time": "5h 3m 56s", "remaining_time": "1h 50m 29s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101062} +{"loss": 0.3260591, "grad_norm": 0.0376968, "learning_rate": 1.747e-05, "epoch": 0.73400338, "global_step/max_steps": "1844/2513", "percentage": "73.38%", "elapsed_time": "5h 4m 6s", "remaining_time": "1h 50m 19s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101062} +{"loss": 0.36723554, "grad_norm": 0.03831903, "learning_rate": 1.742e-05, "epoch": 0.73440143, "global_step/max_steps": "1845/2513", "percentage": "73.42%", "elapsed_time": "5h 4m 16s", "remaining_time": "1h 50m 9s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101062} +{"loss": 0.35812247, "grad_norm": 0.03775706, "learning_rate": 1.737e-05, "epoch": 0.73479948, "global_step/max_steps": "1846/2513", "percentage": "73.46%", "elapsed_time": "5h 4m 25s", "remaining_time": "1h 49m 59s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101064} +{"loss": 0.37857318, "grad_norm": 0.03802053, "learning_rate": 1.732e-05, "epoch": 0.73519753, "global_step/max_steps": "1847/2513", "percentage": "73.50%", "elapsed_time": "5h 4m 35s", "remaining_time": "1h 49m 49s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101066} +{"loss": 0.33334559, "grad_norm": 0.03904004, "learning_rate": 1.727e-05, "epoch": 0.73559558, "global_step/max_steps": "1848/2513", "percentage": "73.54%", "elapsed_time": "5h 4m 45s", "remaining_time": "1h 49m 39s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101066} +{"loss": 0.36983332, "grad_norm": 0.04055828, "learning_rate": 1.723e-05, "epoch": 0.73599363, "global_step/max_steps": "1849/2513", "percentage": "73.58%", "elapsed_time": "5h 4m 54s", "remaining_time": "1h 49m 29s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101067} +{"loss": 0.39640674, "grad_norm": 0.04201368, "learning_rate": 1.718e-05, "epoch": 0.73639168, "global_step/max_steps": "1850/2513", "percentage": "73.62%", "elapsed_time": "5h 5m 4s", "remaining_time": "1h 49m 19s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101069} +{"loss": 0.33556151, "grad_norm": 0.03590813, "learning_rate": 1.713e-05, "epoch": 0.73678973, "global_step/max_steps": "1851/2513", "percentage": "73.66%", "elapsed_time": "5h 5m 14s", "remaining_time": "1h 49m 10s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101069} +{"loss": 0.31716907, "grad_norm": 0.03675658, "learning_rate": 1.708e-05, "epoch": 0.73718778, "global_step/max_steps": "1852/2513", "percentage": "73.70%", "elapsed_time": "5h 5m 24s", "remaining_time": "1h 49m 0s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101066} +{"loss": 0.3456237, "grad_norm": 0.03879976, "learning_rate": 1.703e-05, "epoch": 0.73758583, "global_step/max_steps": "1853/2513", "percentage": "73.74%", "elapsed_time": "5h 5m 34s", "remaining_time": "1h 48m 50s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101064} +{"loss": 0.38257179, "grad_norm": 0.04097363, "learning_rate": 1.698e-05, "epoch": 0.73798388, "global_step/max_steps": "1854/2513", "percentage": "73.78%", "elapsed_time": "5h 5m 45s", "remaining_time": "1h 48m 40s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101062} +{"loss": 0.35344136, "grad_norm": 0.03673316, "learning_rate": 1.693e-05, "epoch": 0.73838193, "global_step/max_steps": "1855/2513", "percentage": "73.82%", "elapsed_time": "5h 5m 54s", "remaining_time": "1h 48m 30s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101064} +{"loss": 0.33982864, "grad_norm": 0.0368421, "learning_rate": 1.689e-05, "epoch": 0.73877998, "global_step/max_steps": "1856/2513", "percentage": "73.86%", "elapsed_time": "5h 6m 4s", "remaining_time": "1h 48m 20s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101066} +{"loss": 0.33021301, "grad_norm": 0.03490047, "learning_rate": 1.684e-05, "epoch": 0.73917803, "global_step/max_steps": "1857/2513", "percentage": "73.90%", "elapsed_time": "5h 6m 13s", "remaining_time": "1h 48m 10s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101067} +{"loss": 0.40567502, "grad_norm": 0.04064079, "learning_rate": 1.679e-05, "epoch": 0.73957608, "global_step/max_steps": "1858/2513", "percentage": "73.94%", "elapsed_time": "5h 6m 23s", "remaining_time": "1h 48m 0s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101069} +{"loss": 0.37482098, "grad_norm": 0.04336725, "learning_rate": 1.674e-05, "epoch": 0.73997413, "global_step/max_steps": "1859/2513", "percentage": "73.98%", "elapsed_time": "5h 6m 33s", "remaining_time": "1h 47m 50s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101071} +{"loss": 0.37345767, "grad_norm": 0.04169437, "learning_rate": 1.669e-05, "epoch": 0.74037218, "global_step/max_steps": "1860/2513", "percentage": "74.02%", "elapsed_time": "5h 6m 42s", "remaining_time": "1h 47m 40s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101072} +{"loss": 0.37432873, "grad_norm": 0.03964141, "learning_rate": 1.665e-05, "epoch": 0.74077023, "global_step/max_steps": "1861/2513", "percentage": "74.05%", "elapsed_time": "5h 6m 52s", "remaining_time": "1h 47m 30s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101073} +{"loss": 0.3696022, "grad_norm": 0.03778457, "learning_rate": 1.66e-05, "epoch": 0.74116828, "global_step/max_steps": "1862/2513", "percentage": "74.09%", "elapsed_time": "5h 7m 1s", "remaining_time": "1h 47m 20s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101075} +{"loss": 0.3448371, "grad_norm": 0.03981556, "learning_rate": 1.655e-05, "epoch": 0.74156633, "global_step/max_steps": "1863/2513", "percentage": "74.13%", "elapsed_time": "5h 7m 11s", "remaining_time": "1h 47m 10s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101076} +{"loss": 0.37971669, "grad_norm": 0.03996774, "learning_rate": 1.65e-05, "epoch": 0.74196437, "global_step/max_steps": "1864/2513", "percentage": "74.17%", "elapsed_time": "5h 7m 21s", "remaining_time": "1h 47m 0s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101076} +{"loss": 0.32652441, "grad_norm": 0.03601867, "learning_rate": 1.645e-05, "epoch": 0.74236242, "global_step/max_steps": "1865/2513", "percentage": "74.21%", "elapsed_time": "5h 7m 32s", "remaining_time": "1h 46m 51s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101073} +{"loss": 0.3705422, "grad_norm": 0.04083525, "learning_rate": 1.641e-05, "epoch": 0.74276047, "global_step/max_steps": "1866/2513", "percentage": "74.25%", "elapsed_time": "5h 7m 41s", "remaining_time": "1h 46m 41s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101073} +{"loss": 0.35126948, "grad_norm": 0.03797588, "learning_rate": 1.636e-05, "epoch": 0.74315852, "global_step/max_steps": "1867/2513", "percentage": "74.29%", "elapsed_time": "5h 7m 52s", "remaining_time": "1h 46m 31s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10107} +{"loss": 0.39321914, "grad_norm": 0.03902425, "learning_rate": 1.631e-05, "epoch": 0.74355657, "global_step/max_steps": "1868/2513", "percentage": "74.33%", "elapsed_time": "5h 8m 2s", "remaining_time": "1h 46m 21s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101071} +{"loss": 0.36076933, "grad_norm": 0.04242645, "learning_rate": 1.626e-05, "epoch": 0.74395462, "global_step/max_steps": "1869/2513", "percentage": "74.37%", "elapsed_time": "5h 8m 12s", "remaining_time": "1h 46m 11s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101071} +{"loss": 0.38649249, "grad_norm": 0.03817252, "learning_rate": 1.622e-05, "epoch": 0.74435267, "global_step/max_steps": "1870/2513", "percentage": "74.41%", "elapsed_time": "5h 8m 21s", "remaining_time": "1h 46m 1s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101073} +{"loss": 0.38602272, "grad_norm": 0.03887994, "learning_rate": 1.617e-05, "epoch": 0.74475072, "global_step/max_steps": "1871/2513", "percentage": "74.45%", "elapsed_time": "5h 8m 31s", "remaining_time": "1h 45m 51s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101075} +{"loss": 0.3704794, "grad_norm": 0.04201199, "learning_rate": 1.612e-05, "epoch": 0.74514877, "global_step/max_steps": "1872/2513", "percentage": "74.49%", "elapsed_time": "5h 8m 40s", "remaining_time": "1h 45m 41s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101077} +{"loss": 0.39102566, "grad_norm": 0.0365371, "learning_rate": 1.607e-05, "epoch": 0.74554682, "global_step/max_steps": "1873/2513", "percentage": "74.53%", "elapsed_time": "5h 8m 50s", "remaining_time": "1h 45m 31s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101078} +{"loss": 0.38802272, "grad_norm": 0.04105331, "learning_rate": 1.603e-05, "epoch": 0.74594487, "global_step/max_steps": "1874/2513", "percentage": "74.57%", "elapsed_time": "5h 9m 0s", "remaining_time": "1h 45m 21s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101079} +{"loss": 0.3432194, "grad_norm": 0.0384548, "learning_rate": 1.598e-05, "epoch": 0.74634292, "global_step/max_steps": "1875/2513", "percentage": "74.61%", "elapsed_time": "5h 9m 9s", "remaining_time": "1h 45m 11s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101082} +{"loss": 0.34918952, "grad_norm": 0.04013681, "learning_rate": 1.593e-05, "epoch": 0.74674097, "global_step/max_steps": "1876/2513", "percentage": "74.65%", "elapsed_time": "5h 9m 19s", "remaining_time": "1h 45m 1s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101082} +{"loss": 0.36736614, "grad_norm": 0.03774971, "learning_rate": 1.588e-05, "epoch": 0.74713902, "global_step/max_steps": "1877/2513", "percentage": "74.69%", "elapsed_time": "5h 9m 29s", "remaining_time": "1h 44m 51s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101082} +{"loss": 0.31855345, "grad_norm": 0.03744502, "learning_rate": 1.584e-05, "epoch": 0.74753707, "global_step/max_steps": "1878/2513", "percentage": "74.73%", "elapsed_time": "5h 9m 38s", "remaining_time": "1h 44m 41s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101083} +{"loss": 0.36753911, "grad_norm": 0.04297927, "learning_rate": 1.579e-05, "epoch": 0.74793512, "global_step/max_steps": "1879/2513", "percentage": "74.77%", "elapsed_time": "5h 9m 48s", "remaining_time": "1h 44m 32s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101082} +{"loss": 0.36029488, "grad_norm": 0.03820128, "learning_rate": 1.574e-05, "epoch": 0.74833317, "global_step/max_steps": "1880/2513", "percentage": "74.81%", "elapsed_time": "5h 9m 59s", "remaining_time": "1h 44m 22s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101081} +{"loss": 0.38766336, "grad_norm": 0.03929848, "learning_rate": 1.57e-05, "epoch": 0.74873122, "global_step/max_steps": "1881/2513", "percentage": "74.85%", "elapsed_time": "5h 10m 9s", "remaining_time": "1h 44m 12s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101079} +{"loss": 0.35205883, "grad_norm": 0.0419052, "learning_rate": 1.565e-05, "epoch": 0.74912927, "global_step/max_steps": "1882/2513", "percentage": "74.89%", "elapsed_time": "5h 10m 19s", "remaining_time": "1h 44m 2s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101075} +{"loss": 0.35468313, "grad_norm": 0.0405512, "learning_rate": 1.56e-05, "epoch": 0.74952732, "global_step/max_steps": "1883/2513", "percentage": "74.93%", "elapsed_time": "5h 10m 29s", "remaining_time": "1h 43m 52s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101076} +{"loss": 0.38502449, "grad_norm": 0.04452367, "learning_rate": 1.556e-05, "epoch": 0.74992537, "global_step/max_steps": "1884/2513", "percentage": "74.97%", "elapsed_time": "5h 10m 39s", "remaining_time": "1h 43m 43s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101076} +{"loss": 0.3527185, "grad_norm": 0.03976179, "learning_rate": 1.551e-05, "epoch": 0.75032342, "global_step/max_steps": "1885/2513", "percentage": "75.01%", "elapsed_time": "5h 10m 49s", "remaining_time": "1h 43m 33s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101076} +{"loss": 0.37351748, "grad_norm": 0.03798196, "learning_rate": 1.546e-05, "epoch": 0.75072146, "global_step/max_steps": "1886/2513", "percentage": "75.05%", "elapsed_time": "5h 10m 59s", "remaining_time": "1h 43m 23s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101076} +{"loss": 0.34388891, "grad_norm": 0.03648869, "learning_rate": 1.542e-05, "epoch": 0.75111951, "global_step/max_steps": "1887/2513", "percentage": "75.09%", "elapsed_time": "5h 11m 9s", "remaining_time": "1h 43m 13s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101077} +{"loss": 0.33297259, "grad_norm": 0.03728346, "learning_rate": 1.537e-05, "epoch": 0.75151756, "global_step/max_steps": "1888/2513", "percentage": "75.13%", "elapsed_time": "5h 11m 18s", "remaining_time": "1h 43m 3s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101078} +{"loss": 0.35659751, "grad_norm": 0.03883943, "learning_rate": 1.532e-05, "epoch": 0.75191561, "global_step/max_steps": "1889/2513", "percentage": "75.17%", "elapsed_time": "5h 11m 28s", "remaining_time": "1h 42m 53s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101078} +{"loss": 0.40059513, "grad_norm": 0.03877813, "learning_rate": 1.528e-05, "epoch": 0.75231366, "global_step/max_steps": "1890/2513", "percentage": "75.21%", "elapsed_time": "5h 11m 38s", "remaining_time": "1h 42m 43s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101079} +{"loss": 0.39456999, "grad_norm": 0.04768571, "learning_rate": 1.523e-05, "epoch": 0.75271171, "global_step/max_steps": "1891/2513", "percentage": "75.25%", "elapsed_time": "5h 11m 47s", "remaining_time": "1h 42m 33s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10108} +{"loss": 0.35792139, "grad_norm": 0.03774198, "learning_rate": 1.518e-05, "epoch": 0.75310976, "global_step/max_steps": "1892/2513", "percentage": "75.29%", "elapsed_time": "5h 11m 57s", "remaining_time": "1h 42m 23s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101081} +{"loss": 0.36374241, "grad_norm": 0.0423401, "learning_rate": 1.514e-05, "epoch": 0.75350781, "global_step/max_steps": "1893/2513", "percentage": "75.33%", "elapsed_time": "5h 12m 7s", "remaining_time": "1h 42m 13s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101081} +{"loss": 0.3571651, "grad_norm": 0.03895366, "learning_rate": 1.509e-05, "epoch": 0.75390586, "global_step/max_steps": "1894/2513", "percentage": "75.37%", "elapsed_time": "5h 12m 17s", "remaining_time": "1h 42m 3s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101078} +{"loss": 0.39077002, "grad_norm": 0.039138, "learning_rate": 1.505e-05, "epoch": 0.75430391, "global_step/max_steps": "1895/2513", "percentage": "75.41%", "elapsed_time": "5h 12m 27s", "remaining_time": "1h 41m 54s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101079} +{"loss": 0.3750689, "grad_norm": 0.03887721, "learning_rate": 1.5e-05, "epoch": 0.75470196, "global_step/max_steps": "1896/2513", "percentage": "75.45%", "elapsed_time": "5h 12m 37s", "remaining_time": "1h 41m 44s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101077} +{"loss": 0.39710945, "grad_norm": 0.03909974, "learning_rate": 1.495e-05, "epoch": 0.75510001, "global_step/max_steps": "1897/2513", "percentage": "75.49%", "elapsed_time": "5h 12m 48s", "remaining_time": "1h 41m 34s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101073} +{"loss": 0.31810266, "grad_norm": 0.03838901, "learning_rate": 1.491e-05, "epoch": 0.75549806, "global_step/max_steps": "1898/2513", "percentage": "75.53%", "elapsed_time": "5h 12m 58s", "remaining_time": "1h 41m 24s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101073} +{"loss": 0.3467221, "grad_norm": 0.03803818, "learning_rate": 1.486e-05, "epoch": 0.75589611, "global_step/max_steps": "1899/2513", "percentage": "75.57%", "elapsed_time": "5h 13m 8s", "remaining_time": "1h 41m 14s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101071} +{"loss": 0.39533064, "grad_norm": 0.04070631, "learning_rate": 1.482e-05, "epoch": 0.75629416, "global_step/max_steps": "1900/2513", "percentage": "75.61%", "elapsed_time": "5h 13m 18s", "remaining_time": "1h 41m 4s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101073} +{"loss": 0.39789477, "grad_norm": 0.04040477, "learning_rate": 1.477e-05, "epoch": 0.75669221, "global_step/max_steps": "1901/2513", "percentage": "75.65%", "elapsed_time": "5h 13m 29s", "remaining_time": "1h 40m 55s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101065} +{"loss": 0.34767541, "grad_norm": 0.03855732, "learning_rate": 1.472e-05, "epoch": 0.75709026, "global_step/max_steps": "1902/2513", "percentage": "75.69%", "elapsed_time": "5h 13m 39s", "remaining_time": "1h 40m 45s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101065} +{"loss": 0.38886076, "grad_norm": 0.04176292, "learning_rate": 1.468e-05, "epoch": 0.75748831, "global_step/max_steps": "1903/2513", "percentage": "75.73%", "elapsed_time": "5h 13m 49s", "remaining_time": "1h 40m 35s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101064} +{"loss": 0.33032289, "grad_norm": 0.04052975, "learning_rate": 1.463e-05, "epoch": 0.75788636, "global_step/max_steps": "1904/2513", "percentage": "75.77%", "elapsed_time": "5h 13m 59s", "remaining_time": "1h 40m 25s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101064} +{"loss": 0.36374396, "grad_norm": 0.03945574, "learning_rate": 1.459e-05, "epoch": 0.75828441, "global_step/max_steps": "1905/2513", "percentage": "75.81%", "elapsed_time": "5h 14m 9s", "remaining_time": "1h 40m 15s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101065} +{"loss": 0.34634697, "grad_norm": 0.03944394, "learning_rate": 1.454e-05, "epoch": 0.75868246, "global_step/max_steps": "1906/2513", "percentage": "75.85%", "elapsed_time": "5h 14m 19s", "remaining_time": "1h 40m 6s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101063} +{"loss": 0.36094019, "grad_norm": 0.03798659, "learning_rate": 1.45e-05, "epoch": 0.75908051, "global_step/max_steps": "1907/2513", "percentage": "75.89%", "elapsed_time": "5h 14m 29s", "remaining_time": "1h 39m 56s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101064} +{"loss": 0.3177118, "grad_norm": 0.03582692, "learning_rate": 1.445e-05, "epoch": 0.75947856, "global_step/max_steps": "1908/2513", "percentage": "75.93%", "elapsed_time": "5h 14m 38s", "remaining_time": "1h 39m 46s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101065} +{"loss": 0.36276042, "grad_norm": 0.04204422, "learning_rate": 1.441e-05, "epoch": 0.7598766, "global_step/max_steps": "1909/2513", "percentage": "75.96%", "elapsed_time": "5h 14m 49s", "remaining_time": "1h 39m 36s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101062} +{"loss": 0.40873659, "grad_norm": 0.03915032, "learning_rate": 1.436e-05, "epoch": 0.76027465, "global_step/max_steps": "1910/2513", "percentage": "76.00%", "elapsed_time": "5h 14m 59s", "remaining_time": "1h 39m 26s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10106} +{"loss": 0.36074921, "grad_norm": 0.04263062, "learning_rate": 1.432e-05, "epoch": 0.7606727, "global_step/max_steps": "1911/2513", "percentage": "76.04%", "elapsed_time": "5h 15m 9s", "remaining_time": "1h 39m 16s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10106} +{"loss": 0.3942782, "grad_norm": 0.04006838, "learning_rate": 1.427e-05, "epoch": 0.76107075, "global_step/max_steps": "1912/2513", "percentage": "76.08%", "elapsed_time": "5h 15m 20s", "remaining_time": "1h 39m 7s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101056} +{"loss": 0.32699031, "grad_norm": 0.03621234, "learning_rate": 1.423e-05, "epoch": 0.7614688, "global_step/max_steps": "1913/2513", "percentage": "76.12%", "elapsed_time": "5h 15m 29s", "remaining_time": "1h 38m 57s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101058} +{"loss": 0.37883139, "grad_norm": 0.03786376, "learning_rate": 1.418e-05, "epoch": 0.76186685, "global_step/max_steps": "1914/2513", "percentage": "76.16%", "elapsed_time": "5h 15m 39s", "remaining_time": "1h 38m 47s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101057} +{"loss": 0.38859761, "grad_norm": 0.04029632, "learning_rate": 1.414e-05, "epoch": 0.7622649, "global_step/max_steps": "1915/2513", "percentage": "76.20%", "elapsed_time": "5h 15m 49s", "remaining_time": "1h 38m 37s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101058} +{"loss": 0.33786058, "grad_norm": 0.04466378, "learning_rate": 1.409e-05, "epoch": 0.76266295, "global_step/max_steps": "1916/2513", "percentage": "76.24%", "elapsed_time": "5h 15m 59s", "remaining_time": "1h 38m 27s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101059} +{"loss": 0.33965743, "grad_norm": 0.03759015, "learning_rate": 1.405e-05, "epoch": 0.763061, "global_step/max_steps": "1917/2513", "percentage": "76.28%", "elapsed_time": "5h 16m 8s", "remaining_time": "1h 38m 17s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10106} +{"loss": 0.33524436, "grad_norm": 0.04003431, "learning_rate": 1.4e-05, "epoch": 0.76345905, "global_step/max_steps": "1918/2513", "percentage": "76.32%", "elapsed_time": "5h 16m 18s", "remaining_time": "1h 38m 7s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10106} +{"loss": 0.41854227, "grad_norm": 0.04534606, "learning_rate": 1.396e-05, "epoch": 0.7638571, "global_step/max_steps": "1919/2513", "percentage": "76.36%", "elapsed_time": "5h 16m 28s", "remaining_time": "1h 37m 57s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10106} +{"loss": 0.32377911, "grad_norm": 0.03622863, "learning_rate": 1.391e-05, "epoch": 0.76425515, "global_step/max_steps": "1920/2513", "percentage": "76.40%", "elapsed_time": "5h 16m 38s", "remaining_time": "1h 37m 47s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101059} +{"loss": 0.4118349, "grad_norm": 0.04343867, "learning_rate": 1.387e-05, "epoch": 0.7646532, "global_step/max_steps": "1921/2513", "percentage": "76.44%", "elapsed_time": "5h 16m 48s", "remaining_time": "1h 37m 37s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101059} +{"loss": 0.3696124, "grad_norm": 0.04703899, "learning_rate": 1.382e-05, "epoch": 0.76505125, "global_step/max_steps": "1922/2513", "percentage": "76.48%", "elapsed_time": "5h 16m 58s", "remaining_time": "1h 37m 28s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10106} +{"loss": 0.36796573, "grad_norm": 0.04217617, "learning_rate": 1.378e-05, "epoch": 0.7654493, "global_step/max_steps": "1923/2513", "percentage": "76.52%", "elapsed_time": "5h 17m 8s", "remaining_time": "1h 37m 18s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10106} +{"loss": 0.3981359, "grad_norm": 0.04474946, "learning_rate": 1.373e-05, "epoch": 0.76584735, "global_step/max_steps": "1924/2513", "percentage": "76.56%", "elapsed_time": "5h 17m 18s", "remaining_time": "1h 37m 8s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101061} +{"loss": 0.37930667, "grad_norm": 0.04455562, "learning_rate": 1.369e-05, "epoch": 0.7662454, "global_step/max_steps": "1925/2513", "percentage": "76.60%", "elapsed_time": "5h 17m 28s", "remaining_time": "1h 36m 58s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101059} +{"loss": 0.35652533, "grad_norm": 0.04252191, "learning_rate": 1.365e-05, "epoch": 0.76664345, "global_step/max_steps": "1926/2513", "percentage": "76.64%", "elapsed_time": "5h 17m 37s", "remaining_time": "1h 36m 48s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101061} +{"loss": 0.42226034, "grad_norm": 0.04326122, "learning_rate": 1.36e-05, "epoch": 0.7670415, "global_step/max_steps": "1927/2513", "percentage": "76.68%", "elapsed_time": "5h 17m 48s", "remaining_time": "1h 36m 38s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101059} +{"loss": 0.35534239, "grad_norm": 0.04054495, "learning_rate": 1.356e-05, "epoch": 0.76743955, "global_step/max_steps": "1928/2513", "percentage": "76.72%", "elapsed_time": "5h 17m 57s", "remaining_time": "1h 36m 28s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10106} +{"loss": 0.31003755, "grad_norm": 0.03590789, "learning_rate": 1.351e-05, "epoch": 0.7678376, "global_step/max_steps": "1929/2513", "percentage": "76.76%", "elapsed_time": "5h 18m 7s", "remaining_time": "1h 36m 18s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10106} +{"loss": 0.36408636, "grad_norm": 0.04477759, "learning_rate": 1.347e-05, "epoch": 0.76823565, "global_step/max_steps": "1930/2513", "percentage": "76.80%", "elapsed_time": "5h 18m 17s", "remaining_time": "1h 36m 8s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101062} +{"loss": 0.37083399, "grad_norm": 0.04041375, "learning_rate": 1.342e-05, "epoch": 0.76863369, "global_step/max_steps": "1931/2513", "percentage": "76.84%", "elapsed_time": "5h 18m 26s", "remaining_time": "1h 35m 58s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101063} +{"loss": 0.39379069, "grad_norm": 0.0449026, "learning_rate": 1.338e-05, "epoch": 0.76903174, "global_step/max_steps": "1932/2513", "percentage": "76.88%", "elapsed_time": "5h 18m 36s", "remaining_time": "1h 35m 48s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101063} +{"loss": 0.37123626, "grad_norm": 0.03767939, "learning_rate": 1.334e-05, "epoch": 0.76942979, "global_step/max_steps": "1933/2513", "percentage": "76.92%", "elapsed_time": "5h 18m 46s", "remaining_time": "1h 35m 38s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101064} +{"loss": 0.36597198, "grad_norm": 0.03965018, "learning_rate": 1.329e-05, "epoch": 0.76982784, "global_step/max_steps": "1934/2513", "percentage": "76.96%", "elapsed_time": "5h 18m 56s", "remaining_time": "1h 35m 29s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101064} +{"loss": 0.40107292, "grad_norm": 0.04538547, "learning_rate": 1.325e-05, "epoch": 0.77022589, "global_step/max_steps": "1935/2513", "percentage": "77.00%", "elapsed_time": "5h 19m 5s", "remaining_time": "1h 35m 19s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101066} +{"loss": 0.32567978, "grad_norm": 0.03677708, "learning_rate": 1.321e-05, "epoch": 0.77062394, "global_step/max_steps": "1936/2513", "percentage": "77.04%", "elapsed_time": "5h 19m 15s", "remaining_time": "1h 35m 9s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101065} +{"loss": 0.35815305, "grad_norm": 0.04285465, "learning_rate": 1.316e-05, "epoch": 0.77102199, "global_step/max_steps": "1937/2513", "percentage": "77.08%", "elapsed_time": "5h 19m 25s", "remaining_time": "1h 34m 59s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101065} +{"loss": 0.31480998, "grad_norm": 0.03984445, "learning_rate": 1.312e-05, "epoch": 0.77142004, "global_step/max_steps": "1938/2513", "percentage": "77.12%", "elapsed_time": "5h 19m 36s", "remaining_time": "1h 34m 49s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101064} +{"loss": 0.37249231, "grad_norm": 0.03982322, "learning_rate": 1.308e-05, "epoch": 0.77181809, "global_step/max_steps": "1939/2513", "percentage": "77.16%", "elapsed_time": "5h 19m 46s", "remaining_time": "1h 34m 39s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101063} +{"loss": 0.36110464, "grad_norm": 0.04047265, "learning_rate": 1.303e-05, "epoch": 0.77221614, "global_step/max_steps": "1940/2513", "percentage": "77.20%", "elapsed_time": "5h 19m 55s", "remaining_time": "1h 34m 29s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101063} +{"loss": 0.3512505, "grad_norm": 0.03931322, "learning_rate": 1.299e-05, "epoch": 0.77261419, "global_step/max_steps": "1941/2513", "percentage": "77.24%", "elapsed_time": "5h 20m 6s", "remaining_time": "1h 34m 19s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101061} +{"loss": 0.40162927, "grad_norm": 0.04439027, "learning_rate": 1.295e-05, "epoch": 0.77301224, "global_step/max_steps": "1942/2513", "percentage": "77.28%", "elapsed_time": "5h 20m 16s", "remaining_time": "1h 34m 10s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101059} +{"loss": 0.33123946, "grad_norm": 0.03725913, "learning_rate": 1.29e-05, "epoch": 0.77341029, "global_step/max_steps": "1943/2513", "percentage": "77.32%", "elapsed_time": "5h 20m 26s", "remaining_time": "1h 34m 0s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101059} +{"loss": 0.32399261, "grad_norm": 0.03773821, "learning_rate": 1.286e-05, "epoch": 0.77380834, "global_step/max_steps": "1944/2513", "percentage": "77.36%", "elapsed_time": "5h 20m 36s", "remaining_time": "1h 33m 50s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10106} +{"loss": 0.37109438, "grad_norm": 0.04246744, "learning_rate": 1.282e-05, "epoch": 0.77420639, "global_step/max_steps": "1945/2513", "percentage": "77.40%", "elapsed_time": "5h 20m 45s", "remaining_time": "1h 33m 40s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101062} +{"loss": 0.39362007, "grad_norm": 0.03854192, "learning_rate": 1.277e-05, "epoch": 0.77460444, "global_step/max_steps": "1946/2513", "percentage": "77.44%", "elapsed_time": "5h 20m 55s", "remaining_time": "1h 33m 30s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101062} +{"loss": 0.37573633, "grad_norm": 0.03845974, "learning_rate": 1.273e-05, "epoch": 0.77500249, "global_step/max_steps": "1947/2513", "percentage": "77.48%", "elapsed_time": "5h 21m 5s", "remaining_time": "1h 33m 20s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101063} +{"loss": 0.38928396, "grad_norm": 0.0404848, "learning_rate": 1.269e-05, "epoch": 0.77540054, "global_step/max_steps": "1948/2513", "percentage": "77.52%", "elapsed_time": "5h 21m 14s", "remaining_time": "1h 33m 10s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101065} +{"loss": 0.33032751, "grad_norm": 0.03678518, "learning_rate": 1.264e-05, "epoch": 0.77579859, "global_step/max_steps": "1949/2513", "percentage": "77.56%", "elapsed_time": "5h 21m 24s", "remaining_time": "1h 33m 0s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101065} +{"loss": 0.36823171, "grad_norm": 0.04036023, "learning_rate": 1.26e-05, "epoch": 0.77619664, "global_step/max_steps": "1950/2513", "percentage": "77.60%", "elapsed_time": "5h 21m 34s", "remaining_time": "1h 32m 50s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101065} +{"loss": 0.39706525, "grad_norm": 0.04211203, "learning_rate": 1.256e-05, "epoch": 0.77659469, "global_step/max_steps": "1951/2513", "percentage": "77.64%", "elapsed_time": "5h 21m 44s", "remaining_time": "1h 32m 40s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101066} +{"loss": 0.34768623, "grad_norm": 0.04157782, "learning_rate": 1.252e-05, "epoch": 0.77699274, "global_step/max_steps": "1952/2513", "percentage": "77.68%", "elapsed_time": "5h 21m 53s", "remaining_time": "1h 32m 30s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101068} +{"loss": 0.37712672, "grad_norm": 0.03724625, "learning_rate": 1.247e-05, "epoch": 0.77739079, "global_step/max_steps": "1953/2513", "percentage": "77.72%", "elapsed_time": "5h 22m 3s", "remaining_time": "1h 32m 20s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101067} +{"loss": 0.37824041, "grad_norm": 0.03976941, "learning_rate": 1.243e-05, "epoch": 0.77778883, "global_step/max_steps": "1954/2513", "percentage": "77.76%", "elapsed_time": "5h 22m 14s", "remaining_time": "1h 32m 11s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101064} +{"loss": 0.36917591, "grad_norm": 0.03768756, "learning_rate": 1.239e-05, "epoch": 0.77818688, "global_step/max_steps": "1955/2513", "percentage": "77.80%", "elapsed_time": "5h 22m 23s", "remaining_time": "1h 32m 1s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101066} +{"loss": 0.35425287, "grad_norm": 0.03794918, "learning_rate": 1.235e-05, "epoch": 0.77858493, "global_step/max_steps": "1956/2513", "percentage": "77.84%", "elapsed_time": "5h 22m 34s", "remaining_time": "1h 31m 51s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101061} +{"loss": 0.36399645, "grad_norm": 0.03714208, "learning_rate": 1.23e-05, "epoch": 0.77898298, "global_step/max_steps": "1957/2513", "percentage": "77.88%", "elapsed_time": "5h 22m 44s", "remaining_time": "1h 31m 41s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101059} +{"loss": 0.37532696, "grad_norm": 0.04188837, "learning_rate": 1.226e-05, "epoch": 0.77938103, "global_step/max_steps": "1958/2513", "percentage": "77.91%", "elapsed_time": "5h 22m 54s", "remaining_time": "1h 31m 31s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101059} +{"loss": 0.32579798, "grad_norm": 0.0343931, "learning_rate": 1.222e-05, "epoch": 0.77977908, "global_step/max_steps": "1959/2513", "percentage": "77.95%", "elapsed_time": "5h 23m 4s", "remaining_time": "1h 31m 21s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101059} +{"loss": 0.41564846, "grad_norm": 0.04085084, "learning_rate": 1.218e-05, "epoch": 0.78017713, "global_step/max_steps": "1960/2513", "percentage": "77.99%", "elapsed_time": "5h 23m 14s", "remaining_time": "1h 31m 11s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101061} +{"loss": 0.357032, "grad_norm": 0.04029486, "learning_rate": 1.213e-05, "epoch": 0.78057518, "global_step/max_steps": "1961/2513", "percentage": "78.03%", "elapsed_time": "5h 23m 23s", "remaining_time": "1h 31m 1s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101063} +{"loss": 0.33224469, "grad_norm": 0.03934089, "learning_rate": 1.209e-05, "epoch": 0.78097323, "global_step/max_steps": "1962/2513", "percentage": "78.07%", "elapsed_time": "5h 23m 33s", "remaining_time": "1h 30m 52s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101063} +{"loss": 0.38714918, "grad_norm": 0.03974557, "learning_rate": 1.205e-05, "epoch": 0.78137128, "global_step/max_steps": "1963/2513", "percentage": "78.11%", "elapsed_time": "5h 23m 43s", "remaining_time": "1h 30m 42s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101063} +{"loss": 0.34509501, "grad_norm": 0.04202912, "learning_rate": 1.201e-05, "epoch": 0.78176933, "global_step/max_steps": "1964/2513", "percentage": "78.15%", "elapsed_time": "5h 23m 53s", "remaining_time": "1h 30m 32s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101063} +{"loss": 0.39171597, "grad_norm": 0.04688504, "learning_rate": 1.197e-05, "epoch": 0.78216738, "global_step/max_steps": "1965/2513", "percentage": "78.19%", "elapsed_time": "5h 24m 3s", "remaining_time": "1h 30m 22s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101063} +{"loss": 0.36350125, "grad_norm": 0.04203723, "learning_rate": 1.192e-05, "epoch": 0.78256543, "global_step/max_steps": "1966/2513", "percentage": "78.23%", "elapsed_time": "5h 24m 13s", "remaining_time": "1h 30m 12s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101063} +{"loss": 0.33169082, "grad_norm": 0.03746529, "learning_rate": 1.188e-05, "epoch": 0.78296348, "global_step/max_steps": "1967/2513", "percentage": "78.27%", "elapsed_time": "5h 24m 22s", "remaining_time": "1h 30m 2s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101064} +{"loss": 0.34454846, "grad_norm": 0.04029046, "learning_rate": 1.184e-05, "epoch": 0.78336153, "global_step/max_steps": "1968/2513", "percentage": "78.31%", "elapsed_time": "5h 24m 32s", "remaining_time": "1h 29m 52s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101063} +{"loss": 0.35759562, "grad_norm": 0.03768001, "learning_rate": 1.18e-05, "epoch": 0.78375958, "global_step/max_steps": "1969/2513", "percentage": "78.35%", "elapsed_time": "5h 24m 43s", "remaining_time": "1h 29m 42s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101061} +{"loss": 0.38424394, "grad_norm": 0.04024763, "learning_rate": 1.176e-05, "epoch": 0.78415763, "global_step/max_steps": "1970/2513", "percentage": "78.39%", "elapsed_time": "5h 24m 53s", "remaining_time": "1h 29m 33s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101059} +{"loss": 0.34937567, "grad_norm": 0.03979493, "learning_rate": 1.172e-05, "epoch": 0.78455568, "global_step/max_steps": "1971/2513", "percentage": "78.43%", "elapsed_time": "5h 25m 3s", "remaining_time": "1h 29m 23s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101057} +{"loss": 0.36846766, "grad_norm": 0.0392788, "learning_rate": 1.167e-05, "epoch": 0.78495373, "global_step/max_steps": "1972/2513", "percentage": "78.47%", "elapsed_time": "5h 25m 13s", "remaining_time": "1h 29m 13s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101057} +{"loss": 0.37376934, "grad_norm": 0.03666881, "learning_rate": 1.163e-05, "epoch": 0.78535178, "global_step/max_steps": "1973/2513", "percentage": "78.51%", "elapsed_time": "5h 25m 23s", "remaining_time": "1h 29m 3s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101058} +{"loss": 0.36314034, "grad_norm": 0.03694188, "learning_rate": 1.159e-05, "epoch": 0.78574983, "global_step/max_steps": "1974/2513", "percentage": "78.55%", "elapsed_time": "5h 25m 33s", "remaining_time": "1h 28m 53s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101058} +{"loss": 0.40818527, "grad_norm": 0.03887508, "learning_rate": 1.155e-05, "epoch": 0.78614788, "global_step/max_steps": "1975/2513", "percentage": "78.59%", "elapsed_time": "5h 25m 43s", "remaining_time": "1h 28m 43s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101058} +{"loss": 0.3414796, "grad_norm": 0.0382983, "learning_rate": 1.151e-05, "epoch": 0.78654592, "global_step/max_steps": "1976/2513", "percentage": "78.63%", "elapsed_time": "5h 25m 52s", "remaining_time": "1h 28m 33s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10106} +{"loss": 0.39339215, "grad_norm": 0.04072575, "learning_rate": 1.147e-05, "epoch": 0.78694397, "global_step/max_steps": "1977/2513", "percentage": "78.67%", "elapsed_time": "5h 26m 2s", "remaining_time": "1h 28m 23s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101061} +{"loss": 0.36391804, "grad_norm": 0.03780003, "learning_rate": 1.143e-05, "epoch": 0.78734202, "global_step/max_steps": "1978/2513", "percentage": "78.71%", "elapsed_time": "5h 26m 12s", "remaining_time": "1h 28m 13s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101061} +{"loss": 0.40788645, "grad_norm": 0.04190721, "learning_rate": 1.139e-05, "epoch": 0.78774007, "global_step/max_steps": "1979/2513", "percentage": "78.75%", "elapsed_time": "5h 26m 22s", "remaining_time": "1h 28m 4s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10106} +{"loss": 0.39039826, "grad_norm": 0.04344618, "learning_rate": 1.135e-05, "epoch": 0.78813812, "global_step/max_steps": "1980/2513", "percentage": "78.79%", "elapsed_time": "5h 26m 32s", "remaining_time": "1h 27m 54s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101061} +{"loss": 0.34266365, "grad_norm": 0.03742013, "learning_rate": 1.13e-05, "epoch": 0.78853617, "global_step/max_steps": "1981/2513", "percentage": "78.83%", "elapsed_time": "5h 26m 41s", "remaining_time": "1h 27m 44s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101062} +{"loss": 0.32671759, "grad_norm": 0.03599561, "learning_rate": 1.126e-05, "epoch": 0.78893422, "global_step/max_steps": "1982/2513", "percentage": "78.87%", "elapsed_time": "5h 26m 51s", "remaining_time": "1h 27m 34s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101061} +{"loss": 0.37228471, "grad_norm": 0.04263633, "learning_rate": 1.122e-05, "epoch": 0.78933227, "global_step/max_steps": "1983/2513", "percentage": "78.91%", "elapsed_time": "5h 27m 2s", "remaining_time": "1h 27m 24s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10106} +{"loss": 0.37408119, "grad_norm": 0.03770704, "learning_rate": 1.118e-05, "epoch": 0.78973032, "global_step/max_steps": "1984/2513", "percentage": "78.95%", "elapsed_time": "5h 27m 12s", "remaining_time": "1h 27m 14s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101058} +{"loss": 0.30473968, "grad_norm": 0.0371539, "learning_rate": 1.114e-05, "epoch": 0.79012837, "global_step/max_steps": "1985/2513", "percentage": "78.99%", "elapsed_time": "5h 27m 22s", "remaining_time": "1h 27m 4s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101054} +{"loss": 0.37497675, "grad_norm": 0.03867068, "learning_rate": 1.11e-05, "epoch": 0.79052642, "global_step/max_steps": "1986/2513", "percentage": "79.03%", "elapsed_time": "5h 27m 33s", "remaining_time": "1h 26m 55s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101052} +{"loss": 0.30487201, "grad_norm": 0.03535695, "learning_rate": 1.106e-05, "epoch": 0.79092447, "global_step/max_steps": "1987/2513", "percentage": "79.07%", "elapsed_time": "5h 27m 43s", "remaining_time": "1h 26m 45s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101052} +{"loss": 0.36047637, "grad_norm": 0.04010689, "learning_rate": 1.102e-05, "epoch": 0.79132252, "global_step/max_steps": "1988/2513", "percentage": "79.11%", "elapsed_time": "5h 27m 52s", "remaining_time": "1h 26m 35s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101054} +{"loss": 0.36772418, "grad_norm": 0.04356168, "learning_rate": 1.098e-05, "epoch": 0.79172057, "global_step/max_steps": "1989/2513", "percentage": "79.15%", "elapsed_time": "5h 28m 2s", "remaining_time": "1h 26m 25s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101056} +{"loss": 0.38908514, "grad_norm": 0.04441492, "learning_rate": 1.094e-05, "epoch": 0.79211862, "global_step/max_steps": "1990/2513", "percentage": "79.19%", "elapsed_time": "5h 28m 11s", "remaining_time": "1h 26m 15s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101058} +{"loss": 0.38077801, "grad_norm": 0.038967, "learning_rate": 1.09e-05, "epoch": 0.79251667, "global_step/max_steps": "1991/2513", "percentage": "79.23%", "elapsed_time": "5h 28m 21s", "remaining_time": "1h 26m 5s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101059} +{"loss": 0.34488118, "grad_norm": 0.04011983, "learning_rate": 1.086e-05, "epoch": 0.79291472, "global_step/max_steps": "1992/2513", "percentage": "79.27%", "elapsed_time": "5h 28m 30s", "remaining_time": "1h 25m 55s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101061} +{"loss": 0.34160879, "grad_norm": 0.03495017, "learning_rate": 1.082e-05, "epoch": 0.79331277, "global_step/max_steps": "1993/2513", "percentage": "79.31%", "elapsed_time": "5h 28m 40s", "remaining_time": "1h 25m 45s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101063} +{"loss": 0.35820547, "grad_norm": 0.03974383, "learning_rate": 1.078e-05, "epoch": 0.79371082, "global_step/max_steps": "1994/2513", "percentage": "79.35%", "elapsed_time": "5h 28m 50s", "remaining_time": "1h 25m 35s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101062} +{"loss": 0.35152584, "grad_norm": 0.03806406, "learning_rate": 1.074e-05, "epoch": 0.79410887, "global_step/max_steps": "1995/2513", "percentage": "79.39%", "elapsed_time": "5h 29m 0s", "remaining_time": "1h 25m 25s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101062} +{"loss": 0.34477448, "grad_norm": 0.03726882, "learning_rate": 1.07e-05, "epoch": 0.79450692, "global_step/max_steps": "1996/2513", "percentage": "79.43%", "elapsed_time": "5h 29m 10s", "remaining_time": "1h 25m 15s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101062} +{"loss": 0.39558387, "grad_norm": 0.04417658, "learning_rate": 1.066e-05, "epoch": 0.79490497, "global_step/max_steps": "1997/2513", "percentage": "79.47%", "elapsed_time": "5h 29m 19s", "remaining_time": "1h 25m 5s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101063} +{"loss": 0.37508118, "grad_norm": 0.03873208, "learning_rate": 1.062e-05, "epoch": 0.79530302, "global_step/max_steps": "1998/2513", "percentage": "79.51%", "elapsed_time": "5h 29m 29s", "remaining_time": "1h 24m 55s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101063} +{"loss": 0.38779992, "grad_norm": 0.04076428, "learning_rate": 1.058e-05, "epoch": 0.79570106, "global_step/max_steps": "1999/2513", "percentage": "79.55%", "elapsed_time": "5h 29m 40s", "remaining_time": "1h 24m 46s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101062} +{"loss": 0.37395871, "grad_norm": 0.04551465, "learning_rate": 1.054e-05, "epoch": 0.79609911, "global_step/max_steps": "2000/2513", "percentage": "79.59%", "elapsed_time": "5h 29m 50s", "remaining_time": "1h 24m 36s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101059} +{"loss": 0.39389849, "grad_norm": 0.04051378, "learning_rate": 1.05e-05, "epoch": 0.79649716, "global_step/max_steps": "2001/2513", "percentage": "79.63%", "elapsed_time": "5h 30m 1s", "remaining_time": "1h 24m 26s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101052} +{"loss": 0.34851813, "grad_norm": 0.04164845, "learning_rate": 1.046e-05, "epoch": 0.79689521, "global_step/max_steps": "2002/2513", "percentage": "79.67%", "elapsed_time": "5h 30m 12s", "remaining_time": "1h 24m 16s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101049} +{"loss": 0.37838423, "grad_norm": 0.03830553, "learning_rate": 1.042e-05, "epoch": 0.79729326, "global_step/max_steps": "2003/2513", "percentage": "79.71%", "elapsed_time": "5h 30m 22s", "remaining_time": "1h 24m 7s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101049} +{"loss": 0.31143978, "grad_norm": 0.03868942, "learning_rate": 1.038e-05, "epoch": 0.79769131, "global_step/max_steps": "2004/2513", "percentage": "79.75%", "elapsed_time": "5h 30m 31s", "remaining_time": "1h 23m 57s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10105} +{"loss": 0.31295842, "grad_norm": 0.04153764, "learning_rate": 1.034e-05, "epoch": 0.79808936, "global_step/max_steps": "2005/2513", "percentage": "79.79%", "elapsed_time": "5h 30m 41s", "remaining_time": "1h 23m 47s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101052} +{"loss": 0.3587119, "grad_norm": 0.04093274, "learning_rate": 1.03e-05, "epoch": 0.79848741, "global_step/max_steps": "2006/2513", "percentage": "79.82%", "elapsed_time": "5h 30m 50s", "remaining_time": "1h 23m 37s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101055} +{"loss": 0.36349571, "grad_norm": 0.04055065, "learning_rate": 1.027e-05, "epoch": 0.79888546, "global_step/max_steps": "2007/2513", "percentage": "79.86%", "elapsed_time": "5h 31m 0s", "remaining_time": "1h 23m 27s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101055} +{"loss": 0.38490453, "grad_norm": 0.04202674, "learning_rate": 1.023e-05, "epoch": 0.79928351, "global_step/max_steps": "2008/2513", "percentage": "79.90%", "elapsed_time": "5h 31m 9s", "remaining_time": "1h 23m 17s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101058} +{"loss": 0.35616803, "grad_norm": 0.03553919, "learning_rate": 1.019e-05, "epoch": 0.79968156, "global_step/max_steps": "2009/2513", "percentage": "79.94%", "elapsed_time": "5h 31m 19s", "remaining_time": "1h 23m 7s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101056} +{"loss": 0.39624143, "grad_norm": 0.04174896, "learning_rate": 1.015e-05, "epoch": 0.80007961, "global_step/max_steps": "2010/2513", "percentage": "79.98%", "elapsed_time": "5h 31m 29s", "remaining_time": "1h 22m 57s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101058} +{"loss": 0.3657546, "grad_norm": 0.04048875, "learning_rate": 1.011e-05, "epoch": 0.80047766, "global_step/max_steps": "2011/2513", "percentage": "80.02%", "elapsed_time": "5h 31m 39s", "remaining_time": "1h 22m 47s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101058} +{"loss": 0.38026732, "grad_norm": 0.04171957, "learning_rate": 1.007e-05, "epoch": 0.80087571, "global_step/max_steps": "2012/2513", "percentage": "80.06%", "elapsed_time": "5h 31m 49s", "remaining_time": "1h 22m 37s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101059} +{"loss": 0.3395586, "grad_norm": 0.03896374, "learning_rate": 1.003e-05, "epoch": 0.80127376, "global_step/max_steps": "2013/2513", "percentage": "80.10%", "elapsed_time": "5h 31m 59s", "remaining_time": "1h 22m 27s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101059} +{"loss": 0.32884824, "grad_norm": 0.03813218, "learning_rate": 9.99e-06, "epoch": 0.80167181, "global_step/max_steps": "2014/2513", "percentage": "80.14%", "elapsed_time": "5h 32m 8s", "remaining_time": "1h 22m 17s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101059} +{"loss": 0.33733934, "grad_norm": 0.03795874, "learning_rate": 9.95e-06, "epoch": 0.80206986, "global_step/max_steps": "2015/2513", "percentage": "80.18%", "elapsed_time": "5h 32m 19s", "remaining_time": "1h 22m 8s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101054} +{"loss": 0.37382942, "grad_norm": 0.03884255, "learning_rate": 9.92e-06, "epoch": 0.80246791, "global_step/max_steps": "2016/2513", "percentage": "80.22%", "elapsed_time": "5h 32m 29s", "remaining_time": "1h 21m 58s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101055} +{"loss": 0.36346269, "grad_norm": 0.04034569, "learning_rate": 9.88e-06, "epoch": 0.80286596, "global_step/max_steps": "2017/2513", "percentage": "80.26%", "elapsed_time": "5h 32m 39s", "remaining_time": "1h 21m 48s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101054} +{"loss": 0.30781314, "grad_norm": 0.0391828, "learning_rate": 9.84e-06, "epoch": 0.80326401, "global_step/max_steps": "2018/2513", "percentage": "80.30%", "elapsed_time": "5h 32m 49s", "remaining_time": "1h 21m 38s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101055} +{"loss": 0.38639274, "grad_norm": 0.04205259, "learning_rate": 9.8e-06, "epoch": 0.80366206, "global_step/max_steps": "2019/2513", "percentage": "80.34%", "elapsed_time": "5h 32m 58s", "remaining_time": "1h 21m 28s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101058} +{"loss": 0.38922006, "grad_norm": 0.04438174, "learning_rate": 9.76e-06, "epoch": 0.80406011, "global_step/max_steps": "2020/2513", "percentage": "80.38%", "elapsed_time": "5h 33m 8s", "remaining_time": "1h 21m 18s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10106} +{"loss": 0.34741277, "grad_norm": 0.03966521, "learning_rate": 9.72e-06, "epoch": 0.80445816, "global_step/max_steps": "2021/2513", "percentage": "80.42%", "elapsed_time": "5h 33m 17s", "remaining_time": "1h 21m 8s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101062} +{"loss": 0.3677187, "grad_norm": 0.0396163, "learning_rate": 9.69e-06, "epoch": 0.8048562, "global_step/max_steps": "2022/2513", "percentage": "80.46%", "elapsed_time": "5h 33m 27s", "remaining_time": "1h 20m 58s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101063} +{"loss": 0.37601122, "grad_norm": 0.04187283, "learning_rate": 9.65e-06, "epoch": 0.80525425, "global_step/max_steps": "2023/2513", "percentage": "80.50%", "elapsed_time": "5h 33m 37s", "remaining_time": "1h 20m 48s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101063} +{"loss": 0.3588649, "grad_norm": 0.04061914, "learning_rate": 9.61e-06, "epoch": 0.8056523, "global_step/max_steps": "2024/2513", "percentage": "80.54%", "elapsed_time": "5h 33m 47s", "remaining_time": "1h 20m 38s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101062} +{"loss": 0.38807896, "grad_norm": 0.03904059, "learning_rate": 9.57e-06, "epoch": 0.80605035, "global_step/max_steps": "2025/2513", "percentage": "80.58%", "elapsed_time": "5h 33m 56s", "remaining_time": "1h 20m 28s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101063} +{"loss": 0.37364161, "grad_norm": 0.04030292, "learning_rate": 9.53e-06, "epoch": 0.8064484, "global_step/max_steps": "2026/2513", "percentage": "80.62%", "elapsed_time": "5h 34m 6s", "remaining_time": "1h 20m 18s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101063} +{"loss": 0.37979415, "grad_norm": 0.04338404, "learning_rate": 9.5e-06, "epoch": 0.80684645, "global_step/max_steps": "2027/2513", "percentage": "80.66%", "elapsed_time": "5h 34m 16s", "remaining_time": "1h 20m 8s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101065} +{"loss": 0.3128593, "grad_norm": 0.03825925, "learning_rate": 9.46e-06, "epoch": 0.8072445, "global_step/max_steps": "2028/2513", "percentage": "80.70%", "elapsed_time": "5h 34m 26s", "remaining_time": "1h 19m 58s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101064} +{"loss": 0.36262634, "grad_norm": 0.04090073, "learning_rate": 9.42e-06, "epoch": 0.80764255, "global_step/max_steps": "2029/2513", "percentage": "80.74%", "elapsed_time": "5h 34m 36s", "remaining_time": "1h 19m 49s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101065} +{"loss": 0.35354108, "grad_norm": 0.04015323, "learning_rate": 9.38e-06, "epoch": 0.8080406, "global_step/max_steps": "2030/2513", "percentage": "80.78%", "elapsed_time": "5h 34m 46s", "remaining_time": "1h 19m 39s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101063} +{"loss": 0.35941499, "grad_norm": 0.04150788, "learning_rate": 9.35e-06, "epoch": 0.80843865, "global_step/max_steps": "2031/2513", "percentage": "80.82%", "elapsed_time": "5h 34m 56s", "remaining_time": "1h 19m 29s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101063} +{"loss": 0.34741953, "grad_norm": 0.0383022, "learning_rate": 9.31e-06, "epoch": 0.8088367, "global_step/max_steps": "2032/2513", "percentage": "80.86%", "elapsed_time": "5h 35m 6s", "remaining_time": "1h 19m 19s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10106} +{"loss": 0.35166016, "grad_norm": 0.0419706, "learning_rate": 9.27e-06, "epoch": 0.80923475, "global_step/max_steps": "2033/2513", "percentage": "80.90%", "elapsed_time": "5h 35m 16s", "remaining_time": "1h 19m 9s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101062} +{"loss": 0.36289847, "grad_norm": 0.04159596, "learning_rate": 9.23e-06, "epoch": 0.8096328, "global_step/max_steps": "2034/2513", "percentage": "80.94%", "elapsed_time": "5h 35m 26s", "remaining_time": "1h 18m 59s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101062} +{"loss": 0.39327312, "grad_norm": 0.04329556, "learning_rate": 9.2e-06, "epoch": 0.81003085, "global_step/max_steps": "2035/2513", "percentage": "80.98%", "elapsed_time": "5h 35m 35s", "remaining_time": "1h 18m 49s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101064} +{"loss": 0.36441407, "grad_norm": 0.04169905, "learning_rate": 9.16e-06, "epoch": 0.8104289, "global_step/max_steps": "2036/2513", "percentage": "81.02%", "elapsed_time": "5h 35m 45s", "remaining_time": "1h 18m 39s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101065} +{"loss": 0.32994089, "grad_norm": 0.03953179, "learning_rate": 9.12e-06, "epoch": 0.81082695, "global_step/max_steps": "2037/2513", "percentage": "81.06%", "elapsed_time": "5h 35m 55s", "remaining_time": "1h 18m 29s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101066} +{"loss": 0.36862844, "grad_norm": 0.04022645, "learning_rate": 9.08e-06, "epoch": 0.811225, "global_step/max_steps": "2038/2513", "percentage": "81.10%", "elapsed_time": "5h 36m 5s", "remaining_time": "1h 18m 19s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101066} +{"loss": 0.35520199, "grad_norm": 0.04052185, "learning_rate": 9.05e-06, "epoch": 0.81162305, "global_step/max_steps": "2039/2513", "percentage": "81.14%", "elapsed_time": "5h 36m 14s", "remaining_time": "1h 18m 9s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101066} +{"loss": 0.39048946, "grad_norm": 0.04071084, "learning_rate": 9.01e-06, "epoch": 0.8120211, "global_step/max_steps": "2040/2513", "percentage": "81.18%", "elapsed_time": "5h 36m 24s", "remaining_time": "1h 18m 0s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101066} +{"loss": 0.41372782, "grad_norm": 0.04209202, "learning_rate": 8.97e-06, "epoch": 0.81241915, "global_step/max_steps": "2041/2513", "percentage": "81.22%", "elapsed_time": "5h 36m 34s", "remaining_time": "1h 17m 50s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101065} +{"loss": 0.37003464, "grad_norm": 0.03981309, "learning_rate": 8.94e-06, "epoch": 0.8128172, "global_step/max_steps": "2042/2513", "percentage": "81.26%", "elapsed_time": "5h 36m 44s", "remaining_time": "1h 17m 40s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101066} +{"loss": 0.39241987, "grad_norm": 0.04703304, "learning_rate": 8.9e-06, "epoch": 0.81321525, "global_step/max_steps": "2043/2513", "percentage": "81.30%", "elapsed_time": "5h 36m 54s", "remaining_time": "1h 17m 30s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101066} +{"loss": 0.33717036, "grad_norm": 0.04471432, "learning_rate": 8.86e-06, "epoch": 0.81361329, "global_step/max_steps": "2044/2513", "percentage": "81.34%", "elapsed_time": "5h 37m 3s", "remaining_time": "1h 17m 20s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101068} +{"loss": 0.36130521, "grad_norm": 0.04061074, "learning_rate": 8.83e-06, "epoch": 0.81401134, "global_step/max_steps": "2045/2513", "percentage": "81.38%", "elapsed_time": "5h 37m 14s", "remaining_time": "1h 17m 10s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101066} +{"loss": 0.36163989, "grad_norm": 0.03905942, "learning_rate": 8.79e-06, "epoch": 0.81440939, "global_step/max_steps": "2046/2513", "percentage": "81.42%", "elapsed_time": "5h 37m 24s", "remaining_time": "1h 17m 0s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101064} +{"loss": 0.3682487, "grad_norm": 0.04171975, "learning_rate": 8.75e-06, "epoch": 0.81480744, "global_step/max_steps": "2047/2513", "percentage": "81.46%", "elapsed_time": "5h 37m 34s", "remaining_time": "1h 16m 50s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101065} +{"loss": 0.34108201, "grad_norm": 0.04030538, "learning_rate": 8.72e-06, "epoch": 0.81520549, "global_step/max_steps": "2048/2513", "percentage": "81.50%", "elapsed_time": "5h 37m 44s", "remaining_time": "1h 16m 40s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101065} +{"loss": 0.37490213, "grad_norm": 0.04193002, "learning_rate": 8.68e-06, "epoch": 0.81560354, "global_step/max_steps": "2049/2513", "percentage": "81.54%", "elapsed_time": "5h 37m 53s", "remaining_time": "1h 16m 31s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101067} +{"loss": 0.39479139, "grad_norm": 0.04209036, "learning_rate": 8.64e-06, "epoch": 0.81600159, "global_step/max_steps": "2050/2513", "percentage": "81.58%", "elapsed_time": "5h 38m 3s", "remaining_time": "1h 16m 21s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101068} +{"loss": 0.39804065, "grad_norm": 0.03987862, "learning_rate": 8.61e-06, "epoch": 0.81639964, "global_step/max_steps": "2051/2513", "percentage": "81.62%", "elapsed_time": "5h 38m 12s", "remaining_time": "1h 16m 11s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10107} +{"loss": 0.33522764, "grad_norm": 0.03938688, "learning_rate": 8.57e-06, "epoch": 0.81679769, "global_step/max_steps": "2052/2513", "percentage": "81.66%", "elapsed_time": "5h 38m 22s", "remaining_time": "1h 16m 1s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101072} +{"loss": 0.34963593, "grad_norm": 0.04109553, "learning_rate": 8.54e-06, "epoch": 0.81719574, "global_step/max_steps": "2053/2513", "percentage": "81.70%", "elapsed_time": "5h 38m 32s", "remaining_time": "1h 15m 51s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101071} +{"loss": 0.3409014, "grad_norm": 0.03813348, "learning_rate": 8.5e-06, "epoch": 0.81759379, "global_step/max_steps": "2054/2513", "percentage": "81.73%", "elapsed_time": "5h 38m 42s", "remaining_time": "1h 15m 41s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101072} +{"loss": 0.37786868, "grad_norm": 0.0409543, "learning_rate": 8.46e-06, "epoch": 0.81799184, "global_step/max_steps": "2055/2513", "percentage": "81.77%", "elapsed_time": "5h 38m 52s", "remaining_time": "1h 15m 31s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101071} +{"loss": 0.38916132, "grad_norm": 0.04163813, "learning_rate": 8.43e-06, "epoch": 0.81838989, "global_step/max_steps": "2056/2513", "percentage": "81.81%", "elapsed_time": "5h 39m 2s", "remaining_time": "1h 15m 21s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101071} +{"loss": 0.37232256, "grad_norm": 0.03710061, "learning_rate": 8.39e-06, "epoch": 0.81878794, "global_step/max_steps": "2057/2513", "percentage": "81.85%", "elapsed_time": "5h 39m 11s", "remaining_time": "1h 15m 11s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101072} +{"loss": 0.38773882, "grad_norm": 0.04355095, "learning_rate": 8.36e-06, "epoch": 0.81918599, "global_step/max_steps": "2058/2513", "percentage": "81.89%", "elapsed_time": "5h 39m 21s", "remaining_time": "1h 15m 1s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101072} +{"loss": 0.36635879, "grad_norm": 0.04233346, "learning_rate": 8.32e-06, "epoch": 0.81958404, "global_step/max_steps": "2059/2513", "percentage": "81.93%", "elapsed_time": "5h 39m 32s", "remaining_time": "1h 14m 51s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101069} +{"loss": 0.37373599, "grad_norm": 0.03896614, "learning_rate": 8.29e-06, "epoch": 0.81998209, "global_step/max_steps": "2060/2513", "percentage": "81.97%", "elapsed_time": "5h 39m 41s", "remaining_time": "1h 14m 42s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10107} +{"loss": 0.3864997, "grad_norm": 0.04091047, "learning_rate": 8.25e-06, "epoch": 0.82038014, "global_step/max_steps": "2061/2513", "percentage": "82.01%", "elapsed_time": "5h 39m 51s", "remaining_time": "1h 14m 32s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101069} +{"loss": 0.37433499, "grad_norm": 0.03987834, "learning_rate": 8.22e-06, "epoch": 0.82077819, "global_step/max_steps": "2062/2513", "percentage": "82.05%", "elapsed_time": "5h 40m 1s", "remaining_time": "1h 14m 22s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101071} +{"loss": 0.33426067, "grad_norm": 0.04473417, "learning_rate": 8.18e-06, "epoch": 0.82117624, "global_step/max_steps": "2063/2513", "percentage": "82.09%", "elapsed_time": "5h 40m 11s", "remaining_time": "1h 14m 12s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101072} +{"loss": 0.31508046, "grad_norm": 0.03875923, "learning_rate": 8.14e-06, "epoch": 0.82157429, "global_step/max_steps": "2064/2513", "percentage": "82.13%", "elapsed_time": "5h 40m 20s", "remaining_time": "1h 14m 2s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101073} +{"loss": 0.34756041, "grad_norm": 0.04024164, "learning_rate": 8.11e-06, "epoch": 0.82197234, "global_step/max_steps": "2065/2513", "percentage": "82.17%", "elapsed_time": "5h 40m 30s", "remaining_time": "1h 13m 52s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101074} +{"loss": 0.40478748, "grad_norm": 0.04208288, "learning_rate": 8.07e-06, "epoch": 0.82237039, "global_step/max_steps": "2066/2513", "percentage": "82.21%", "elapsed_time": "5h 40m 40s", "remaining_time": "1h 13m 42s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101076} +{"loss": 0.35319117, "grad_norm": 0.03808355, "learning_rate": 8.04e-06, "epoch": 0.82276843, "global_step/max_steps": "2067/2513", "percentage": "82.25%", "elapsed_time": "5h 40m 50s", "remaining_time": "1h 13m 32s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101075} +{"loss": 0.37212938, "grad_norm": 0.03654209, "learning_rate": 8e-06, "epoch": 0.82316648, "global_step/max_steps": "2068/2513", "percentage": "82.29%", "elapsed_time": "5h 40m 59s", "remaining_time": "1h 13m 22s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101076} +{"loss": 0.34939677, "grad_norm": 0.04084612, "learning_rate": 7.97e-06, "epoch": 0.82356453, "global_step/max_steps": "2069/2513", "percentage": "82.33%", "elapsed_time": "5h 41m 9s", "remaining_time": "1h 13m 12s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101077} +{"loss": 0.35277516, "grad_norm": 0.03631011, "learning_rate": 7.93e-06, "epoch": 0.82396258, "global_step/max_steps": "2070/2513", "percentage": "82.37%", "elapsed_time": "5h 41m 19s", "remaining_time": "1h 13m 2s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101075} +{"loss": 0.38636345, "grad_norm": 0.04066951, "learning_rate": 7.9e-06, "epoch": 0.82436063, "global_step/max_steps": "2071/2513", "percentage": "82.41%", "elapsed_time": "5h 41m 29s", "remaining_time": "1h 12m 52s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101075} +{"loss": 0.35045111, "grad_norm": 0.04034306, "learning_rate": 7.86e-06, "epoch": 0.82475868, "global_step/max_steps": "2072/2513", "percentage": "82.45%", "elapsed_time": "5h 41m 39s", "remaining_time": "1h 12m 43s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101075} +{"loss": 0.37488797, "grad_norm": 0.04083956, "learning_rate": 7.83e-06, "epoch": 0.82515673, "global_step/max_steps": "2073/2513", "percentage": "82.49%", "elapsed_time": "5h 41m 49s", "remaining_time": "1h 12m 33s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101074} +{"loss": 0.38263696, "grad_norm": 0.04377902, "learning_rate": 7.8e-06, "epoch": 0.82555478, "global_step/max_steps": "2074/2513", "percentage": "82.53%", "elapsed_time": "5h 42m 0s", "remaining_time": "1h 12m 23s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101071} +{"loss": 0.30791813, "grad_norm": 0.0356127, "learning_rate": 7.76e-06, "epoch": 0.82595283, "global_step/max_steps": "2075/2513", "percentage": "82.57%", "elapsed_time": "5h 42m 10s", "remaining_time": "1h 12m 13s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10107} +{"loss": 0.35002565, "grad_norm": 0.04035079, "learning_rate": 7.73e-06, "epoch": 0.82635088, "global_step/max_steps": "2076/2513", "percentage": "82.61%", "elapsed_time": "5h 42m 20s", "remaining_time": "1h 12m 3s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101069} +{"loss": 0.35692823, "grad_norm": 0.0381856, "learning_rate": 7.69e-06, "epoch": 0.82674893, "global_step/max_steps": "2077/2513", "percentage": "82.65%", "elapsed_time": "5h 42m 29s", "remaining_time": "1h 11m 53s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101071} +{"loss": 0.359496, "grad_norm": 0.03895749, "learning_rate": 7.66e-06, "epoch": 0.82714698, "global_step/max_steps": "2078/2513", "percentage": "82.69%", "elapsed_time": "5h 42m 39s", "remaining_time": "1h 11m 43s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101071} +{"loss": 0.35159439, "grad_norm": 0.03624024, "learning_rate": 7.62e-06, "epoch": 0.82754503, "global_step/max_steps": "2079/2513", "percentage": "82.73%", "elapsed_time": "5h 42m 49s", "remaining_time": "1h 11m 33s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101072} +{"loss": 0.34090334, "grad_norm": 0.04048537, "learning_rate": 7.59e-06, "epoch": 0.82794308, "global_step/max_steps": "2080/2513", "percentage": "82.77%", "elapsed_time": "5h 42m 59s", "remaining_time": "1h 11m 24s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101073} +{"loss": 0.37881315, "grad_norm": 0.04049302, "learning_rate": 7.56e-06, "epoch": 0.82834113, "global_step/max_steps": "2081/2513", "percentage": "82.81%", "elapsed_time": "5h 43m 8s", "remaining_time": "1h 11m 14s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101074} +{"loss": 0.3721543, "grad_norm": 0.0385901, "learning_rate": 7.52e-06, "epoch": 0.82873918, "global_step/max_steps": "2082/2513", "percentage": "82.85%", "elapsed_time": "5h 43m 18s", "remaining_time": "1h 11m 4s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101073} +{"loss": 0.37891251, "grad_norm": 0.03933053, "learning_rate": 7.49e-06, "epoch": 0.82913723, "global_step/max_steps": "2083/2513", "percentage": "82.89%", "elapsed_time": "5h 43m 28s", "remaining_time": "1h 10m 54s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101076} +{"loss": 0.39318901, "grad_norm": 0.04133896, "learning_rate": 7.45e-06, "epoch": 0.82953528, "global_step/max_steps": "2084/2513", "percentage": "82.93%", "elapsed_time": "5h 43m 37s", "remaining_time": "1h 10m 44s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101078} +{"loss": 0.35631859, "grad_norm": 0.03791454, "learning_rate": 7.42e-06, "epoch": 0.82993333, "global_step/max_steps": "2085/2513", "percentage": "82.97%", "elapsed_time": "5h 43m 48s", "remaining_time": "1h 10m 34s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101075} +{"loss": 0.42495322, "grad_norm": 0.0443142, "learning_rate": 7.39e-06, "epoch": 0.83033138, "global_step/max_steps": "2086/2513", "percentage": "83.01%", "elapsed_time": "5h 43m 57s", "remaining_time": "1h 10m 24s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101076} +{"loss": 0.38101256, "grad_norm": 0.04092602, "learning_rate": 7.35e-06, "epoch": 0.83072943, "global_step/max_steps": "2087/2513", "percentage": "83.05%", "elapsed_time": "5h 44m 8s", "remaining_time": "1h 10m 14s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101073} +{"loss": 0.34306684, "grad_norm": 0.03732677, "learning_rate": 7.32e-06, "epoch": 0.83112748, "global_step/max_steps": "2088/2513", "percentage": "83.09%", "elapsed_time": "5h 44m 18s", "remaining_time": "1h 10m 4s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101072} +{"loss": 0.35934502, "grad_norm": 0.04301722, "learning_rate": 7.28e-06, "epoch": 0.83152552, "global_step/max_steps": "2089/2513", "percentage": "83.13%", "elapsed_time": "5h 44m 28s", "remaining_time": "1h 9m 55s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101071} +{"loss": 0.36983132, "grad_norm": 0.03960834, "learning_rate": 7.25e-06, "epoch": 0.83192357, "global_step/max_steps": "2090/2513", "percentage": "83.17%", "elapsed_time": "5h 44m 38s", "remaining_time": "1h 9m 45s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101072} +{"loss": 0.38930321, "grad_norm": 0.04100133, "learning_rate": 7.22e-06, "epoch": 0.83232162, "global_step/max_steps": "2091/2513", "percentage": "83.21%", "elapsed_time": "5h 44m 48s", "remaining_time": "1h 9m 35s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101073} +{"loss": 0.33863983, "grad_norm": 0.03810031, "learning_rate": 7.18e-06, "epoch": 0.83271967, "global_step/max_steps": "2092/2513", "percentage": "83.25%", "elapsed_time": "5h 44m 57s", "remaining_time": "1h 9m 25s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101075} +{"loss": 0.37825865, "grad_norm": 0.04180864, "learning_rate": 7.15e-06, "epoch": 0.83311772, "global_step/max_steps": "2093/2513", "percentage": "83.29%", "elapsed_time": "5h 45m 7s", "remaining_time": "1h 9m 15s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101077} +{"loss": 0.36807168, "grad_norm": 0.03715655, "learning_rate": 7.12e-06, "epoch": 0.83351577, "global_step/max_steps": "2094/2513", "percentage": "83.33%", "elapsed_time": "5h 45m 16s", "remaining_time": "1h 9m 5s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101077} +{"loss": 0.34559321, "grad_norm": 0.04225916, "learning_rate": 7.09e-06, "epoch": 0.83391382, "global_step/max_steps": "2095/2513", "percentage": "83.37%", "elapsed_time": "5h 45m 26s", "remaining_time": "1h 8m 55s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101078} +{"loss": 0.34308964, "grad_norm": 0.03610347, "learning_rate": 7.05e-06, "epoch": 0.83431187, "global_step/max_steps": "2096/2513", "percentage": "83.41%", "elapsed_time": "5h 45m 36s", "remaining_time": "1h 8m 45s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101078} +{"loss": 0.37005204, "grad_norm": 0.04160344, "learning_rate": 7.02e-06, "epoch": 0.83470992, "global_step/max_steps": "2097/2513", "percentage": "83.45%", "elapsed_time": "5h 45m 46s", "remaining_time": "1h 8m 35s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101077} +{"loss": 0.3301225, "grad_norm": 0.03631332, "learning_rate": 6.99e-06, "epoch": 0.83510797, "global_step/max_steps": "2098/2513", "percentage": "83.49%", "elapsed_time": "5h 45m 56s", "remaining_time": "1h 8m 25s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101079} +{"loss": 0.3504681, "grad_norm": 0.0380315, "learning_rate": 6.95e-06, "epoch": 0.83550602, "global_step/max_steps": "2099/2513", "percentage": "83.53%", "elapsed_time": "5h 46m 5s", "remaining_time": "1h 8m 15s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101079} +{"loss": 0.33947125, "grad_norm": 0.03992773, "learning_rate": 6.92e-06, "epoch": 0.83590407, "global_step/max_steps": "2100/2513", "percentage": "83.57%", "elapsed_time": "5h 46m 15s", "remaining_time": "1h 8m 5s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101078} +{"loss": 0.31385031, "grad_norm": 0.03749635, "learning_rate": 6.89e-06, "epoch": 0.83630212, "global_step/max_steps": "2101/2513", "percentage": "83.61%", "elapsed_time": "5h 46m 27s", "remaining_time": "1h 7m 56s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101071} +{"loss": 0.36834651, "grad_norm": 0.03945381, "learning_rate": 6.86e-06, "epoch": 0.83670017, "global_step/max_steps": "2102/2513", "percentage": "83.65%", "elapsed_time": "5h 46m 37s", "remaining_time": "1h 7m 46s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101071} +{"loss": 0.36173022, "grad_norm": 0.0415095, "learning_rate": 6.82e-06, "epoch": 0.83709822, "global_step/max_steps": "2103/2513", "percentage": "83.68%", "elapsed_time": "5h 46m 47s", "remaining_time": "1h 7m 36s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101067} +{"loss": 0.33934292, "grad_norm": 0.03674014, "learning_rate": 6.79e-06, "epoch": 0.83749627, "global_step/max_steps": "2104/2513", "percentage": "83.72%", "elapsed_time": "5h 46m 57s", "remaining_time": "1h 7m 26s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101069} +{"loss": 0.38391331, "grad_norm": 0.040629, "learning_rate": 6.76e-06, "epoch": 0.83789432, "global_step/max_steps": "2105/2513", "percentage": "83.76%", "elapsed_time": "5h 47m 7s", "remaining_time": "1h 7m 16s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101069} +{"loss": 0.41013852, "grad_norm": 0.04311998, "learning_rate": 6.73e-06, "epoch": 0.83829237, "global_step/max_steps": "2106/2513", "percentage": "83.80%", "elapsed_time": "5h 47m 16s", "remaining_time": "1h 7m 6s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101071} +{"loss": 0.34708768, "grad_norm": 0.03877266, "learning_rate": 6.69e-06, "epoch": 0.83869042, "global_step/max_steps": "2107/2513", "percentage": "83.84%", "elapsed_time": "5h 47m 26s", "remaining_time": "1h 6m 56s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101072} +{"loss": 0.38662606, "grad_norm": 0.04261489, "learning_rate": 6.66e-06, "epoch": 0.83908847, "global_step/max_steps": "2108/2513", "percentage": "83.88%", "elapsed_time": "5h 47m 36s", "remaining_time": "1h 6m 47s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101071} +{"loss": 0.3468321, "grad_norm": 0.03690625, "learning_rate": 6.63e-06, "epoch": 0.83948652, "global_step/max_steps": "2109/2513", "percentage": "83.92%", "elapsed_time": "5h 47m 46s", "remaining_time": "1h 6m 37s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101071} +{"loss": 0.36573303, "grad_norm": 0.03959831, "learning_rate": 6.6e-06, "epoch": 0.83988457, "global_step/max_steps": "2110/2513", "percentage": "83.96%", "elapsed_time": "5h 47m 55s", "remaining_time": "1h 6m 27s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101073} +{"loss": 0.38700318, "grad_norm": 0.04075491, "learning_rate": 6.57e-06, "epoch": 0.84028262, "global_step/max_steps": "2111/2513", "percentage": "84.00%", "elapsed_time": "5h 48m 5s", "remaining_time": "1h 6m 17s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101075} +{"loss": 0.37660688, "grad_norm": 0.04194655, "learning_rate": 6.53e-06, "epoch": 0.84068066, "global_step/max_steps": "2112/2513", "percentage": "84.04%", "elapsed_time": "5h 48m 15s", "remaining_time": "1h 6m 7s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101075} +{"loss": 0.36976993, "grad_norm": 0.03826378, "learning_rate": 6.5e-06, "epoch": 0.84107871, "global_step/max_steps": "2113/2513", "percentage": "84.08%", "elapsed_time": "5h 48m 25s", "remaining_time": "1h 5m 57s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101075} +{"loss": 0.32539928, "grad_norm": 0.03939155, "learning_rate": 6.47e-06, "epoch": 0.84147676, "global_step/max_steps": "2114/2513", "percentage": "84.12%", "elapsed_time": "5h 48m 34s", "remaining_time": "1h 5m 47s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101076} +{"loss": 0.37705085, "grad_norm": 0.03993818, "learning_rate": 6.44e-06, "epoch": 0.84187481, "global_step/max_steps": "2115/2513", "percentage": "84.16%", "elapsed_time": "5h 48m 45s", "remaining_time": "1h 5m 37s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101075} +{"loss": 0.35076773, "grad_norm": 0.03737573, "learning_rate": 6.41e-06, "epoch": 0.84227286, "global_step/max_steps": "2116/2513", "percentage": "84.20%", "elapsed_time": "5h 48m 54s", "remaining_time": "1h 5m 27s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101076} +{"loss": 0.34597629, "grad_norm": 0.04266817, "learning_rate": 6.37e-06, "epoch": 0.84267091, "global_step/max_steps": "2117/2513", "percentage": "84.24%", "elapsed_time": "5h 49m 4s", "remaining_time": "1h 5m 17s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101077} +{"loss": 0.38571548, "grad_norm": 0.04609189, "learning_rate": 6.34e-06, "epoch": 0.84306896, "global_step/max_steps": "2118/2513", "percentage": "84.28%", "elapsed_time": "5h 49m 15s", "remaining_time": "1h 5m 8s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101073} +{"loss": 0.38780248, "grad_norm": 0.04044531, "learning_rate": 6.31e-06, "epoch": 0.84346701, "global_step/max_steps": "2119/2513", "percentage": "84.32%", "elapsed_time": "5h 49m 25s", "remaining_time": "1h 4m 58s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101072} +{"loss": 0.37337556, "grad_norm": 0.04197374, "learning_rate": 6.28e-06, "epoch": 0.84386506, "global_step/max_steps": "2120/2513", "percentage": "84.36%", "elapsed_time": "5h 49m 34s", "remaining_time": "1h 4m 48s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101074} +{"loss": 0.33493698, "grad_norm": 0.03953559, "learning_rate": 6.25e-06, "epoch": 0.84426311, "global_step/max_steps": "2121/2513", "percentage": "84.40%", "elapsed_time": "5h 49m 44s", "remaining_time": "1h 4m 38s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101075} +{"loss": 0.32592019, "grad_norm": 0.03652446, "learning_rate": 6.22e-06, "epoch": 0.84466116, "global_step/max_steps": "2122/2513", "percentage": "84.44%", "elapsed_time": "5h 49m 53s", "remaining_time": "1h 4m 28s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101077} +{"loss": 0.35057607, "grad_norm": 0.03878179, "learning_rate": 6.19e-06, "epoch": 0.84505921, "global_step/max_steps": "2123/2513", "percentage": "84.48%", "elapsed_time": "5h 50m 3s", "remaining_time": "1h 4m 18s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101078} +{"loss": 0.32772392, "grad_norm": 0.03973481, "learning_rate": 6.16e-06, "epoch": 0.84545726, "global_step/max_steps": "2124/2513", "percentage": "84.52%", "elapsed_time": "5h 50m 13s", "remaining_time": "1h 4m 8s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101079} +{"loss": 0.35657924, "grad_norm": 0.04207392, "learning_rate": 6.13e-06, "epoch": 0.84585531, "global_step/max_steps": "2125/2513", "percentage": "84.56%", "elapsed_time": "5h 50m 23s", "remaining_time": "1h 3m 58s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101079} +{"loss": 0.34233898, "grad_norm": 0.04444461, "learning_rate": 6.09e-06, "epoch": 0.84625336, "global_step/max_steps": "2126/2513", "percentage": "84.60%", "elapsed_time": "5h 50m 32s", "remaining_time": "1h 3m 48s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101081} +{"loss": 0.38287687, "grad_norm": 0.04232915, "learning_rate": 6.06e-06, "epoch": 0.84665141, "global_step/max_steps": "2127/2513", "percentage": "84.64%", "elapsed_time": "5h 50m 42s", "remaining_time": "1h 3m 38s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101084} +{"loss": 0.3680782, "grad_norm": 0.04071359, "learning_rate": 6.03e-06, "epoch": 0.84704946, "global_step/max_steps": "2128/2513", "percentage": "84.68%", "elapsed_time": "5h 50m 52s", "remaining_time": "1h 3m 28s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101082} +{"loss": 0.33345464, "grad_norm": 0.0368135, "learning_rate": 6e-06, "epoch": 0.84744751, "global_step/max_steps": "2129/2513", "percentage": "84.72%", "elapsed_time": "5h 51m 2s", "remaining_time": "1h 3m 18s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101082} +{"loss": 0.39715534, "grad_norm": 0.04663152, "learning_rate": 5.97e-06, "epoch": 0.84784556, "global_step/max_steps": "2130/2513", "percentage": "84.76%", "elapsed_time": "5h 51m 12s", "remaining_time": "1h 3m 9s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101081} +{"loss": 0.35019737, "grad_norm": 0.03796367, "learning_rate": 5.94e-06, "epoch": 0.84824361, "global_step/max_steps": "2131/2513", "percentage": "84.80%", "elapsed_time": "5h 51m 22s", "remaining_time": "1h 2m 59s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10108} +{"loss": 0.36915004, "grad_norm": 0.03951093, "learning_rate": 5.91e-06, "epoch": 0.84864166, "global_step/max_steps": "2132/2513", "percentage": "84.84%", "elapsed_time": "5h 51m 32s", "remaining_time": "1h 2m 49s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101079} +{"loss": 0.37086228, "grad_norm": 0.03840027, "learning_rate": 5.88e-06, "epoch": 0.84903971, "global_step/max_steps": "2133/2513", "percentage": "84.88%", "elapsed_time": "5h 51m 42s", "remaining_time": "1h 2m 39s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101078} +{"loss": 0.37275726, "grad_norm": 0.03941028, "learning_rate": 5.85e-06, "epoch": 0.84943776, "global_step/max_steps": "2134/2513", "percentage": "84.92%", "elapsed_time": "5h 51m 52s", "remaining_time": "1h 2m 29s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101077} +{"loss": 0.40081808, "grad_norm": 0.03895602, "learning_rate": 5.82e-06, "epoch": 0.8498358, "global_step/max_steps": "2135/2513", "percentage": "84.96%", "elapsed_time": "5h 52m 2s", "remaining_time": "1h 2m 19s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101078} +{"loss": 0.31907624, "grad_norm": 0.04176057, "learning_rate": 5.79e-06, "epoch": 0.85023385, "global_step/max_steps": "2136/2513", "percentage": "85.00%", "elapsed_time": "5h 52m 11s", "remaining_time": "1h 2m 9s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101081} +{"loss": 0.36144733, "grad_norm": 0.03888028, "learning_rate": 5.76e-06, "epoch": 0.8506319, "global_step/max_steps": "2137/2513", "percentage": "85.04%", "elapsed_time": "5h 52m 21s", "remaining_time": "1h 1m 59s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101082} +{"loss": 0.34855777, "grad_norm": 0.04201547, "learning_rate": 5.73e-06, "epoch": 0.85102995, "global_step/max_steps": "2138/2513", "percentage": "85.08%", "elapsed_time": "5h 52m 30s", "remaining_time": "1h 1m 49s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101083} +{"loss": 0.32882449, "grad_norm": 0.03725719, "learning_rate": 5.7e-06, "epoch": 0.851428, "global_step/max_steps": "2139/2513", "percentage": "85.12%", "elapsed_time": "5h 52m 40s", "remaining_time": "1h 1m 39s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101085} +{"loss": 0.3259902, "grad_norm": 0.0387679, "learning_rate": 5.67e-06, "epoch": 0.85182605, "global_step/max_steps": "2140/2513", "percentage": "85.16%", "elapsed_time": "5h 52m 49s", "remaining_time": "1h 1m 29s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101088} +{"loss": 0.3368575, "grad_norm": 0.0394495, "learning_rate": 5.64e-06, "epoch": 0.8522241, "global_step/max_steps": "2141/2513", "percentage": "85.20%", "elapsed_time": "5h 52m 59s", "remaining_time": "1h 1m 19s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10109} +{"loss": 0.36627787, "grad_norm": 0.03849701, "learning_rate": 5.61e-06, "epoch": 0.85262215, "global_step/max_steps": "2142/2513", "percentage": "85.24%", "elapsed_time": "5h 53m 9s", "remaining_time": "1h 1m 10s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101089} +{"loss": 0.35969591, "grad_norm": 0.03747489, "learning_rate": 5.58e-06, "epoch": 0.8530202, "global_step/max_steps": "2143/2513", "percentage": "85.28%", "elapsed_time": "5h 53m 18s", "remaining_time": "1h 1m 0s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101091} +{"loss": 0.32660896, "grad_norm": 0.03719087, "learning_rate": 5.55e-06, "epoch": 0.85341825, "global_step/max_steps": "2144/2513", "percentage": "85.32%", "elapsed_time": "5h 53m 28s", "remaining_time": "1h 0m 50s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101089} +{"loss": 0.35032052, "grad_norm": 0.03823757, "learning_rate": 5.52e-06, "epoch": 0.8538163, "global_step/max_steps": "2145/2513", "percentage": "85.36%", "elapsed_time": "5h 53m 39s", "remaining_time": "1h 0m 40s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101088} +{"loss": 0.34910637, "grad_norm": 0.03663702, "learning_rate": 5.49e-06, "epoch": 0.85421435, "global_step/max_steps": "2146/2513", "percentage": "85.40%", "elapsed_time": "5h 53m 49s", "remaining_time": "1h 0m 30s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101088} +{"loss": 0.38543648, "grad_norm": 0.04090904, "learning_rate": 5.46e-06, "epoch": 0.8546124, "global_step/max_steps": "2147/2513", "percentage": "85.44%", "elapsed_time": "5h 53m 59s", "remaining_time": "1h 0m 20s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101087} +{"loss": 0.36182731, "grad_norm": 0.03792441, "learning_rate": 5.43e-06, "epoch": 0.85501045, "global_step/max_steps": "2148/2513", "percentage": "85.48%", "elapsed_time": "5h 54m 9s", "remaining_time": "1h 0m 10s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101086} +{"loss": 0.3461861, "grad_norm": 0.04111122, "learning_rate": 5.4e-06, "epoch": 0.8554085, "global_step/max_steps": "2149/2513", "percentage": "85.52%", "elapsed_time": "5h 54m 19s", "remaining_time": "1h 0m 0s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101087} +{"loss": 0.34431762, "grad_norm": 0.03773953, "learning_rate": 5.38e-06, "epoch": 0.85580655, "global_step/max_steps": "2150/2513", "percentage": "85.56%", "elapsed_time": "5h 54m 28s", "remaining_time": "59m 50s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101088} +{"loss": 0.35677448, "grad_norm": 0.03656835, "learning_rate": 5.35e-06, "epoch": 0.8562046, "global_step/max_steps": "2151/2513", "percentage": "85.59%", "elapsed_time": "5h 54m 38s", "remaining_time": "59m 41s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101089} +{"loss": 0.36865777, "grad_norm": 0.0365462, "learning_rate": 5.32e-06, "epoch": 0.85660265, "global_step/max_steps": "2152/2513", "percentage": "85.63%", "elapsed_time": "5h 54m 47s", "remaining_time": "59m 31s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101091} +{"loss": 0.40909201, "grad_norm": 0.04435601, "learning_rate": 5.29e-06, "epoch": 0.8570007, "global_step/max_steps": "2153/2513", "percentage": "85.67%", "elapsed_time": "5h 54m 57s", "remaining_time": "59m 21s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101093} +{"loss": 0.33442852, "grad_norm": 0.0377262, "learning_rate": 5.26e-06, "epoch": 0.85739875, "global_step/max_steps": "2154/2513", "percentage": "85.71%", "elapsed_time": "5h 55m 6s", "remaining_time": "59m 11s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101094} +{"loss": 0.35277247, "grad_norm": 0.04202267, "learning_rate": 5.23e-06, "epoch": 0.8577968, "global_step/max_steps": "2155/2513", "percentage": "85.75%", "elapsed_time": "5h 55m 16s", "remaining_time": "59m 1s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101097} +{"loss": 0.38496518, "grad_norm": 0.04203579, "learning_rate": 5.2e-06, "epoch": 0.85819485, "global_step/max_steps": "2156/2513", "percentage": "85.79%", "elapsed_time": "5h 55m 25s", "remaining_time": "58m 51s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101098} +{"loss": 0.37442374, "grad_norm": 0.04404248, "learning_rate": 5.17e-06, "epoch": 0.85859289, "global_step/max_steps": "2157/2513", "percentage": "85.83%", "elapsed_time": "5h 55m 35s", "remaining_time": "58m 41s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101098} +{"loss": 0.38560098, "grad_norm": 0.04205486, "learning_rate": 5.15e-06, "epoch": 0.85899094, "global_step/max_steps": "2158/2513", "percentage": "85.87%", "elapsed_time": "5h 55m 45s", "remaining_time": "58m 31s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101099} +{"loss": 0.38596523, "grad_norm": 0.04096209, "learning_rate": 5.12e-06, "epoch": 0.85938899, "global_step/max_steps": "2159/2513", "percentage": "85.91%", "elapsed_time": "5h 55m 55s", "remaining_time": "58m 21s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101099} +{"loss": 0.37994742, "grad_norm": 0.04237923, "learning_rate": 5.09e-06, "epoch": 0.85978704, "global_step/max_steps": "2160/2513", "percentage": "85.95%", "elapsed_time": "5h 56m 5s", "remaining_time": "58m 11s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101098} +{"loss": 0.349208, "grad_norm": 0.03724098, "learning_rate": 5.06e-06, "epoch": 0.86018509, "global_step/max_steps": "2161/2513", "percentage": "85.99%", "elapsed_time": "5h 56m 14s", "remaining_time": "58m 1s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.1011} +{"loss": 0.3662765, "grad_norm": 0.04243735, "learning_rate": 5.03e-06, "epoch": 0.86058314, "global_step/max_steps": "2162/2513", "percentage": "86.03%", "elapsed_time": "5h 56m 25s", "remaining_time": "57m 51s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101096} +{"loss": 0.3834967, "grad_norm": 0.0376328, "learning_rate": 5e-06, "epoch": 0.86098119, "global_step/max_steps": "2163/2513", "percentage": "86.07%", "elapsed_time": "5h 56m 35s", "remaining_time": "57m 42s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101095} +{"loss": 0.33936056, "grad_norm": 0.03766158, "learning_rate": 4.98e-06, "epoch": 0.86137924, "global_step/max_steps": "2164/2513", "percentage": "86.11%", "elapsed_time": "5h 56m 45s", "remaining_time": "57m 32s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101094} +{"loss": 0.37360421, "grad_norm": 0.03763127, "learning_rate": 4.95e-06, "epoch": 0.86177729, "global_step/max_steps": "2165/2513", "percentage": "86.15%", "elapsed_time": "5h 56m 55s", "remaining_time": "57m 22s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101095} +{"loss": 0.38482469, "grad_norm": 0.04179162, "learning_rate": 4.92e-06, "epoch": 0.86217534, "global_step/max_steps": "2166/2513", "percentage": "86.19%", "elapsed_time": "5h 57m 5s", "remaining_time": "57m 12s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101095} +{"loss": 0.36701921, "grad_norm": 0.04267217, "learning_rate": 4.89e-06, "epoch": 0.86257339, "global_step/max_steps": "2167/2513", "percentage": "86.23%", "elapsed_time": "5h 57m 15s", "remaining_time": "57m 2s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101096} +{"loss": 0.37867695, "grad_norm": 0.04359603, "learning_rate": 4.86e-06, "epoch": 0.86297144, "global_step/max_steps": "2168/2513", "percentage": "86.27%", "elapsed_time": "5h 57m 24s", "remaining_time": "56m 52s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101098} +{"loss": 0.3414278, "grad_norm": 0.03888197, "learning_rate": 4.84e-06, "epoch": 0.86336949, "global_step/max_steps": "2169/2513", "percentage": "86.31%", "elapsed_time": "5h 57m 34s", "remaining_time": "56m 42s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101099} +{"loss": 0.3548418, "grad_norm": 0.0373586, "learning_rate": 4.81e-06, "epoch": 0.86376754, "global_step/max_steps": "2170/2513", "percentage": "86.35%", "elapsed_time": "5h 57m 43s", "remaining_time": "56m 32s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101101} +{"loss": 0.37828535, "grad_norm": 0.04071718, "learning_rate": 4.78e-06, "epoch": 0.86416559, "global_step/max_steps": "2171/2513", "percentage": "86.39%", "elapsed_time": "5h 57m 53s", "remaining_time": "56m 22s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101101} +{"loss": 0.3610025, "grad_norm": 0.03844276, "learning_rate": 4.75e-06, "epoch": 0.86456364, "global_step/max_steps": "2172/2513", "percentage": "86.43%", "elapsed_time": "5h 58m 3s", "remaining_time": "56m 12s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101102} +{"loss": 0.34575397, "grad_norm": 0.03758214, "learning_rate": 4.73e-06, "epoch": 0.86496169, "global_step/max_steps": "2173/2513", "percentage": "86.47%", "elapsed_time": "5h 58m 13s", "remaining_time": "56m 2s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101102} +{"loss": 0.37464029, "grad_norm": 0.03813827, "learning_rate": 4.7e-06, "epoch": 0.86535974, "global_step/max_steps": "2174/2513", "percentage": "86.51%", "elapsed_time": "5h 58m 23s", "remaining_time": "55m 53s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101102} +{"loss": 0.38723269, "grad_norm": 0.03950563, "learning_rate": 4.67e-06, "epoch": 0.86575779, "global_step/max_steps": "2175/2513", "percentage": "86.55%", "elapsed_time": "5h 58m 33s", "remaining_time": "55m 43s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101101} +{"loss": 0.34315604, "grad_norm": 0.03914243, "learning_rate": 4.64e-06, "epoch": 0.86615584, "global_step/max_steps": "2176/2513", "percentage": "86.59%", "elapsed_time": "5h 58m 43s", "remaining_time": "55m 33s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.1011} +{"loss": 0.37327057, "grad_norm": 0.04174009, "learning_rate": 4.62e-06, "epoch": 0.86655389, "global_step/max_steps": "2177/2513", "percentage": "86.63%", "elapsed_time": "5h 58m 53s", "remaining_time": "55m 23s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101098} +{"loss": 0.39304739, "grad_norm": 0.04313336, "learning_rate": 4.59e-06, "epoch": 0.86695194, "global_step/max_steps": "2178/2513", "percentage": "86.67%", "elapsed_time": "5h 59m 3s", "remaining_time": "55m 13s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101098} +{"loss": 0.36381969, "grad_norm": 0.0378747, "learning_rate": 4.56e-06, "epoch": 0.86734999, "global_step/max_steps": "2179/2513", "percentage": "86.71%", "elapsed_time": "5h 59m 13s", "remaining_time": "55m 3s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101097} +{"loss": 0.34005791, "grad_norm": 0.03794488, "learning_rate": 4.54e-06, "epoch": 0.86774803, "global_step/max_steps": "2180/2513", "percentage": "86.75%", "elapsed_time": "5h 59m 23s", "remaining_time": "54m 53s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101098} +{"loss": 0.31525534, "grad_norm": 0.03994507, "learning_rate": 4.51e-06, "epoch": 0.86814608, "global_step/max_steps": "2181/2513", "percentage": "86.79%", "elapsed_time": "5h 59m 33s", "remaining_time": "54m 43s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101098} +{"loss": 0.29810363, "grad_norm": 0.03695108, "learning_rate": 4.48e-06, "epoch": 0.86854413, "global_step/max_steps": "2182/2513", "percentage": "86.83%", "elapsed_time": "5h 59m 42s", "remaining_time": "54m 33s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.1011} +{"loss": 0.36575848, "grad_norm": 0.04142537, "learning_rate": 4.46e-06, "epoch": 0.86894218, "global_step/max_steps": "2183/2513", "percentage": "86.87%", "elapsed_time": "5h 59m 52s", "remaining_time": "54m 24s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101099} +{"loss": 0.40696484, "grad_norm": 0.04383482, "learning_rate": 4.43e-06, "epoch": 0.86934023, "global_step/max_steps": "2184/2513", "percentage": "86.91%", "elapsed_time": "6h 0m 2s", "remaining_time": "54m 14s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101099} +{"loss": 0.35977161, "grad_norm": 0.04039531, "learning_rate": 4.4e-06, "epoch": 0.86973828, "global_step/max_steps": "2185/2513", "percentage": "86.95%", "elapsed_time": "6h 0m 12s", "remaining_time": "54m 4s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.1011} +{"loss": 0.33971286, "grad_norm": 0.03787889, "learning_rate": 4.38e-06, "epoch": 0.87013633, "global_step/max_steps": "2186/2513", "percentage": "86.99%", "elapsed_time": "6h 0m 22s", "remaining_time": "53m 54s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.1011} +{"loss": 0.34984386, "grad_norm": 0.04213157, "learning_rate": 4.35e-06, "epoch": 0.87053438, "global_step/max_steps": "2187/2513", "percentage": "87.03%", "elapsed_time": "6h 0m 31s", "remaining_time": "53m 44s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.1011} +{"loss": 0.38575047, "grad_norm": 0.05366591, "learning_rate": 4.32e-06, "epoch": 0.87093243, "global_step/max_steps": "2188/2513", "percentage": "87.07%", "elapsed_time": "6h 0m 41s", "remaining_time": "53m 34s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101101} +{"loss": 0.37617257, "grad_norm": 0.03915643, "learning_rate": 4.3e-06, "epoch": 0.87133048, "global_step/max_steps": "2189/2513", "percentage": "87.11%", "elapsed_time": "6h 0m 51s", "remaining_time": "53m 24s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101101} +{"loss": 0.38529482, "grad_norm": 0.04501158, "learning_rate": 4.27e-06, "epoch": 0.87172853, "global_step/max_steps": "2190/2513", "percentage": "87.15%", "elapsed_time": "6h 1m 1s", "remaining_time": "53m 14s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101101} +{"loss": 0.3696782, "grad_norm": 0.0403552, "learning_rate": 4.25e-06, "epoch": 0.87212658, "global_step/max_steps": "2191/2513", "percentage": "87.19%", "elapsed_time": "6h 1m 11s", "remaining_time": "53m 5s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101098} +{"loss": 0.3674534, "grad_norm": 0.03437992, "learning_rate": 4.22e-06, "epoch": 0.87252463, "global_step/max_steps": "2192/2513", "percentage": "87.23%", "elapsed_time": "6h 1m 22s", "remaining_time": "52m 55s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101097} +{"loss": 0.36684996, "grad_norm": 0.0408709, "learning_rate": 4.19e-06, "epoch": 0.87292268, "global_step/max_steps": "2193/2513", "percentage": "87.27%", "elapsed_time": "6h 1m 32s", "remaining_time": "52m 45s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101096} +{"loss": 0.37281135, "grad_norm": 0.0403044, "learning_rate": 4.17e-06, "epoch": 0.87332073, "global_step/max_steps": "2194/2513", "percentage": "87.31%", "elapsed_time": "6h 1m 42s", "remaining_time": "52m 35s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101096} +{"loss": 0.38367888, "grad_norm": 0.04469379, "learning_rate": 4.14e-06, "epoch": 0.87371878, "global_step/max_steps": "2195/2513", "percentage": "87.35%", "elapsed_time": "6h 1m 51s", "remaining_time": "52m 25s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101097} +{"loss": 0.39335081, "grad_norm": 0.04024485, "learning_rate": 4.12e-06, "epoch": 0.87411683, "global_step/max_steps": "2196/2513", "percentage": "87.39%", "elapsed_time": "6h 2m 1s", "remaining_time": "52m 15s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101097} +{"loss": 0.32582563, "grad_norm": 0.03781102, "learning_rate": 4.09e-06, "epoch": 0.87451488, "global_step/max_steps": "2197/2513", "percentage": "87.43%", "elapsed_time": "6h 2m 11s", "remaining_time": "52m 5s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.1011} +{"loss": 0.40820944, "grad_norm": 0.04021759, "learning_rate": 4.07e-06, "epoch": 0.87491293, "global_step/max_steps": "2198/2513", "percentage": "87.47%", "elapsed_time": "6h 2m 20s", "remaining_time": "51m 55s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.1011} +{"loss": 0.41344512, "grad_norm": 0.04191153, "learning_rate": 4.04e-06, "epoch": 0.87531098, "global_step/max_steps": "2199/2513", "percentage": "87.50%", "elapsed_time": "6h 2m 30s", "remaining_time": "51m 45s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101101} +{"loss": 0.37733951, "grad_norm": 0.04079167, "learning_rate": 4.02e-06, "epoch": 0.87570903, "global_step/max_steps": "2200/2513", "percentage": "87.54%", "elapsed_time": "6h 2m 40s", "remaining_time": "51m 35s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101102} +{"loss": 0.31417567, "grad_norm": 0.03767161, "learning_rate": 3.99e-06, "epoch": 0.87610708, "global_step/max_steps": "2201/2513", "percentage": "87.58%", "elapsed_time": "6h 2m 51s", "remaining_time": "51m 26s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101093} +{"loss": 0.32329354, "grad_norm": 0.03865152, "learning_rate": 3.96e-06, "epoch": 0.87650512, "global_step/max_steps": "2202/2513", "percentage": "87.62%", "elapsed_time": "6h 3m 1s", "remaining_time": "51m 16s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101093} +{"loss": 0.37448198, "grad_norm": 0.03996215, "learning_rate": 3.94e-06, "epoch": 0.87690317, "global_step/max_steps": "2203/2513", "percentage": "87.66%", "elapsed_time": "6h 3m 11s", "remaining_time": "51m 6s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101093} +{"loss": 0.37661761, "grad_norm": 0.0462843, "learning_rate": 3.91e-06, "epoch": 0.87730122, "global_step/max_steps": "2204/2513", "percentage": "87.70%", "elapsed_time": "6h 3m 22s", "remaining_time": "50m 56s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10109} +{"loss": 0.40636426, "grad_norm": 0.04326352, "learning_rate": 3.89e-06, "epoch": 0.87769927, "global_step/max_steps": "2205/2513", "percentage": "87.74%", "elapsed_time": "6h 3m 32s", "remaining_time": "50m 46s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10109} +{"loss": 0.37760407, "grad_norm": 0.03971325, "learning_rate": 3.86e-06, "epoch": 0.87809732, "global_step/max_steps": "2206/2513", "percentage": "87.78%", "elapsed_time": "6h 3m 42s", "remaining_time": "50m 36s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101089} +{"loss": 0.3596822, "grad_norm": 0.04036427, "learning_rate": 3.84e-06, "epoch": 0.87849537, "global_step/max_steps": "2207/2513", "percentage": "87.82%", "elapsed_time": "6h 3m 53s", "remaining_time": "50m 27s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101085} +{"loss": 0.36632499, "grad_norm": 0.04038747, "learning_rate": 3.82e-06, "epoch": 0.87889342, "global_step/max_steps": "2208/2513", "percentage": "87.86%", "elapsed_time": "6h 4m 3s", "remaining_time": "50m 17s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101084} +{"loss": 0.35491669, "grad_norm": 0.03932468, "learning_rate": 3.79e-06, "epoch": 0.87929147, "global_step/max_steps": "2209/2513", "percentage": "87.90%", "elapsed_time": "6h 4m 13s", "remaining_time": "50m 7s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101084} +{"loss": 0.3656663, "grad_norm": 0.03828991, "learning_rate": 3.77e-06, "epoch": 0.87968952, "global_step/max_steps": "2210/2513", "percentage": "87.94%", "elapsed_time": "6h 4m 22s", "remaining_time": "49m 57s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101085} +{"loss": 0.40944248, "grad_norm": 0.04036062, "learning_rate": 3.74e-06, "epoch": 0.88008757, "global_step/max_steps": "2211/2513", "percentage": "87.98%", "elapsed_time": "6h 4m 32s", "remaining_time": "49m 47s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101084} +{"loss": 0.39650723, "grad_norm": 0.04189367, "learning_rate": 3.72e-06, "epoch": 0.88048562, "global_step/max_steps": "2212/2513", "percentage": "88.02%", "elapsed_time": "6h 4m 42s", "remaining_time": "49m 37s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101085} +{"loss": 0.38627499, "grad_norm": 0.04029617, "learning_rate": 3.69e-06, "epoch": 0.88088367, "global_step/max_steps": "2213/2513", "percentage": "88.06%", "elapsed_time": "6h 4m 52s", "remaining_time": "49m 27s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101084} +{"loss": 0.34809583, "grad_norm": 0.03619998, "learning_rate": 3.67e-06, "epoch": 0.88128172, "global_step/max_steps": "2214/2513", "percentage": "88.10%", "elapsed_time": "6h 5m 2s", "remaining_time": "49m 17s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101085} +{"loss": 0.33765477, "grad_norm": 0.03556319, "learning_rate": 3.64e-06, "epoch": 0.88167977, "global_step/max_steps": "2215/2513", "percentage": "88.14%", "elapsed_time": "6h 5m 12s", "remaining_time": "49m 7s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101086} +{"loss": 0.34325784, "grad_norm": 0.04236738, "learning_rate": 3.62e-06, "epoch": 0.88207782, "global_step/max_steps": "2216/2513", "percentage": "88.18%", "elapsed_time": "6h 5m 22s", "remaining_time": "48m 58s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101085} +{"loss": 0.32166404, "grad_norm": 0.03734674, "learning_rate": 3.6e-06, "epoch": 0.88247587, "global_step/max_steps": "2217/2513", "percentage": "88.22%", "elapsed_time": "6h 5m 31s", "remaining_time": "48m 48s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101086} +{"loss": 0.37296626, "grad_norm": 0.04065389, "learning_rate": 3.57e-06, "epoch": 0.88287392, "global_step/max_steps": "2218/2513", "percentage": "88.26%", "elapsed_time": "6h 5m 41s", "remaining_time": "48m 38s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101086} +{"loss": 0.36913946, "grad_norm": 0.03709412, "learning_rate": 3.55e-06, "epoch": 0.88327197, "global_step/max_steps": "2219/2513", "percentage": "88.30%", "elapsed_time": "6h 5m 51s", "remaining_time": "48m 28s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101085} +{"loss": 0.34303272, "grad_norm": 0.03888328, "learning_rate": 3.52e-06, "epoch": 0.88367002, "global_step/max_steps": "2220/2513", "percentage": "88.34%", "elapsed_time": "6h 6m 1s", "remaining_time": "48m 18s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101084} +{"loss": 0.3831625, "grad_norm": 0.04107862, "learning_rate": 3.5e-06, "epoch": 0.88406807, "global_step/max_steps": "2221/2513", "percentage": "88.38%", "elapsed_time": "6h 6m 12s", "remaining_time": "48m 8s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101083} +{"loss": 0.30752301, "grad_norm": 0.0370807, "learning_rate": 3.48e-06, "epoch": 0.88446612, "global_step/max_steps": "2222/2513", "percentage": "88.42%", "elapsed_time": "6h 6m 21s", "remaining_time": "47m 58s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101083} +{"loss": 0.37410748, "grad_norm": 0.04173565, "learning_rate": 3.45e-06, "epoch": 0.88486417, "global_step/max_steps": "2223/2513", "percentage": "88.46%", "elapsed_time": "6h 6m 32s", "remaining_time": "47m 48s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101082} +{"loss": 0.36984426, "grad_norm": 0.03966333, "learning_rate": 3.43e-06, "epoch": 0.88526222, "global_step/max_steps": "2224/2513", "percentage": "88.50%", "elapsed_time": "6h 6m 42s", "remaining_time": "47m 39s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10108} +{"loss": 0.35860941, "grad_norm": 0.04361383, "learning_rate": 3.41e-06, "epoch": 0.88566026, "global_step/max_steps": "2225/2513", "percentage": "88.54%", "elapsed_time": "6h 6m 52s", "remaining_time": "47m 29s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10108} +{"loss": 0.35638416, "grad_norm": 0.03765189, "learning_rate": 3.38e-06, "epoch": 0.88605831, "global_step/max_steps": "2226/2513", "percentage": "88.58%", "elapsed_time": "6h 7m 2s", "remaining_time": "47m 19s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101081} +{"loss": 0.33128649, "grad_norm": 0.03686494, "learning_rate": 3.36e-06, "epoch": 0.88645636, "global_step/max_steps": "2227/2513", "percentage": "88.62%", "elapsed_time": "6h 7m 12s", "remaining_time": "47m 9s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10108} +{"loss": 0.34662697, "grad_norm": 0.04405836, "learning_rate": 3.34e-06, "epoch": 0.88685441, "global_step/max_steps": "2228/2513", "percentage": "88.66%", "elapsed_time": "6h 7m 21s", "remaining_time": "46m 59s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10108} +{"loss": 0.31270927, "grad_norm": 0.03707828, "learning_rate": 3.31e-06, "epoch": 0.88725246, "global_step/max_steps": "2229/2513", "percentage": "88.70%", "elapsed_time": "6h 7m 31s", "remaining_time": "46m 49s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101082} +{"loss": 0.37119246, "grad_norm": 0.0384765, "learning_rate": 3.29e-06, "epoch": 0.88765051, "global_step/max_steps": "2230/2513", "percentage": "88.74%", "elapsed_time": "6h 7m 41s", "remaining_time": "46m 39s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101081} +{"loss": 0.35559061, "grad_norm": 0.03836597, "learning_rate": 3.27e-06, "epoch": 0.88804856, "global_step/max_steps": "2231/2513", "percentage": "88.78%", "elapsed_time": "6h 7m 51s", "remaining_time": "46m 29s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10108} +{"loss": 0.37231514, "grad_norm": 0.04008477, "learning_rate": 3.24e-06, "epoch": 0.88844661, "global_step/max_steps": "2232/2513", "percentage": "88.82%", "elapsed_time": "6h 8m 1s", "remaining_time": "46m 19s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101082} +{"loss": 0.32981849, "grad_norm": 0.04240018, "learning_rate": 3.22e-06, "epoch": 0.88884466, "global_step/max_steps": "2233/2513", "percentage": "88.86%", "elapsed_time": "6h 8m 11s", "remaining_time": "46m 10s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101082} +{"loss": 0.41944242, "grad_norm": 0.04900022, "learning_rate": 3.2e-06, "epoch": 0.88924271, "global_step/max_steps": "2234/2513", "percentage": "88.90%", "elapsed_time": "6h 8m 20s", "remaining_time": "46m 0s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101082} +{"loss": 0.30460715, "grad_norm": 0.03438408, "learning_rate": 3.18e-06, "epoch": 0.88964076, "global_step/max_steps": "2235/2513", "percentage": "88.94%", "elapsed_time": "6h 8m 31s", "remaining_time": "45m 50s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101079} +{"loss": 0.40063599, "grad_norm": 0.04068603, "learning_rate": 3.15e-06, "epoch": 0.89003881, "global_step/max_steps": "2236/2513", "percentage": "88.98%", "elapsed_time": "6h 8m 41s", "remaining_time": "45m 40s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101076} +{"loss": 0.35501951, "grad_norm": 0.0391092, "learning_rate": 3.13e-06, "epoch": 0.89043686, "global_step/max_steps": "2237/2513", "percentage": "89.02%", "elapsed_time": "6h 8m 52s", "remaining_time": "45m 30s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101075} +{"loss": 0.36831766, "grad_norm": 0.04266644, "learning_rate": 3.11e-06, "epoch": 0.89083491, "global_step/max_steps": "2238/2513", "percentage": "89.06%", "elapsed_time": "6h 9m 1s", "remaining_time": "45m 20s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101076} +{"loss": 0.37359357, "grad_norm": 0.03971936, "learning_rate": 3.09e-06, "epoch": 0.89123296, "global_step/max_steps": "2239/2513", "percentage": "89.10%", "elapsed_time": "6h 9m 11s", "remaining_time": "45m 10s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101077} +{"loss": 0.32507765, "grad_norm": 0.03996241, "learning_rate": 3.06e-06, "epoch": 0.89163101, "global_step/max_steps": "2240/2513", "percentage": "89.14%", "elapsed_time": "6h 9m 21s", "remaining_time": "45m 0s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101078} +{"loss": 0.33446622, "grad_norm": 0.03862912, "learning_rate": 3.04e-06, "epoch": 0.89202906, "global_step/max_steps": "2241/2513", "percentage": "89.18%", "elapsed_time": "6h 9m 30s", "remaining_time": "44m 50s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101078} +{"loss": 0.38389549, "grad_norm": 0.03831474, "learning_rate": 3.02e-06, "epoch": 0.89242711, "global_step/max_steps": "2242/2513", "percentage": "89.22%", "elapsed_time": "6h 9m 40s", "remaining_time": "44m 41s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101078} +{"loss": 0.31626701, "grad_norm": 0.04022759, "learning_rate": 3e-06, "epoch": 0.89282516, "global_step/max_steps": "2243/2513", "percentage": "89.26%", "elapsed_time": "6h 9m 50s", "remaining_time": "44m 31s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10108} +{"loss": 0.38242441, "grad_norm": 0.0417961, "learning_rate": 2.98e-06, "epoch": 0.89322321, "global_step/max_steps": "2244/2513", "percentage": "89.30%", "elapsed_time": "6h 9m 59s", "remaining_time": "44m 21s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101081} +{"loss": 0.34294111, "grad_norm": 0.0468599, "learning_rate": 2.95e-06, "epoch": 0.89362126, "global_step/max_steps": "2245/2513", "percentage": "89.34%", "elapsed_time": "6h 10m 9s", "remaining_time": "44m 11s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101084} +{"loss": 0.36994165, "grad_norm": 0.04209406, "learning_rate": 2.93e-06, "epoch": 0.89401931, "global_step/max_steps": "2246/2513", "percentage": "89.38%", "elapsed_time": "6h 10m 18s", "remaining_time": "44m 1s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101086} +{"loss": 0.38397443, "grad_norm": 0.03963103, "learning_rate": 2.91e-06, "epoch": 0.89441735, "global_step/max_steps": "2247/2513", "percentage": "89.42%", "elapsed_time": "6h 10m 28s", "remaining_time": "43m 51s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101086} +{"loss": 0.3652494, "grad_norm": 0.04243485, "learning_rate": 2.89e-06, "epoch": 0.8948154, "global_step/max_steps": "2248/2513", "percentage": "89.45%", "elapsed_time": "6h 10m 38s", "remaining_time": "43m 41s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101087} +{"loss": 0.3320421, "grad_norm": 0.03635827, "learning_rate": 2.87e-06, "epoch": 0.89521345, "global_step/max_steps": "2249/2513", "percentage": "89.49%", "elapsed_time": "6h 10m 48s", "remaining_time": "43m 31s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101087} +{"loss": 0.41458976, "grad_norm": 0.04151411, "learning_rate": 2.85e-06, "epoch": 0.8956115, "global_step/max_steps": "2250/2513", "percentage": "89.53%", "elapsed_time": "6h 10m 58s", "remaining_time": "43m 21s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101085} +{"loss": 0.38790372, "grad_norm": 0.04292895, "learning_rate": 2.82e-06, "epoch": 0.89600955, "global_step/max_steps": "2251/2513", "percentage": "89.57%", "elapsed_time": "6h 11m 9s", "remaining_time": "43m 11s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101082} +{"loss": 0.35432416, "grad_norm": 0.03812751, "learning_rate": 2.8e-06, "epoch": 0.8964076, "global_step/max_steps": "2252/2513", "percentage": "89.61%", "elapsed_time": "6h 11m 18s", "remaining_time": "43m 2s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101082} +{"loss": 0.42483199, "grad_norm": 0.04423217, "learning_rate": 2.78e-06, "epoch": 0.89680565, "global_step/max_steps": "2253/2513", "percentage": "89.65%", "elapsed_time": "6h 11m 28s", "remaining_time": "42m 52s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101082} +{"loss": 0.36042818, "grad_norm": 0.03988797, "learning_rate": 2.76e-06, "epoch": 0.8972037, "global_step/max_steps": "2254/2513", "percentage": "89.69%", "elapsed_time": "6h 11m 38s", "remaining_time": "42m 42s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101082} +{"loss": 0.33240834, "grad_norm": 0.03966881, "learning_rate": 2.74e-06, "epoch": 0.89760175, "global_step/max_steps": "2255/2513", "percentage": "89.73%", "elapsed_time": "6h 11m 48s", "remaining_time": "42m 32s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101083} +{"loss": 0.34562558, "grad_norm": 0.04072718, "learning_rate": 2.72e-06, "epoch": 0.8979998, "global_step/max_steps": "2256/2513", "percentage": "89.77%", "elapsed_time": "6h 11m 58s", "remaining_time": "42m 22s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101083} +{"loss": 0.37229759, "grad_norm": 0.03867706, "learning_rate": 2.7e-06, "epoch": 0.89839785, "global_step/max_steps": "2257/2513", "percentage": "89.81%", "elapsed_time": "6h 12m 8s", "remaining_time": "42m 12s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101082} +{"loss": 0.35123467, "grad_norm": 0.03819757, "learning_rate": 2.68e-06, "epoch": 0.8987959, "global_step/max_steps": "2258/2513", "percentage": "89.85%", "elapsed_time": "6h 12m 18s", "remaining_time": "42m 2s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101082} +{"loss": 0.3269639, "grad_norm": 0.03822682, "learning_rate": 2.66e-06, "epoch": 0.89919395, "global_step/max_steps": "2259/2513", "percentage": "89.89%", "elapsed_time": "6h 12m 28s", "remaining_time": "41m 52s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101082} +{"loss": 0.34399396, "grad_norm": 0.04077029, "learning_rate": 2.64e-06, "epoch": 0.899592, "global_step/max_steps": "2260/2513", "percentage": "89.93%", "elapsed_time": "6h 12m 38s", "remaining_time": "41m 42s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101081} +{"loss": 0.36668068, "grad_norm": 0.03866393, "learning_rate": 2.62e-06, "epoch": 0.89999005, "global_step/max_steps": "2261/2513", "percentage": "89.97%", "elapsed_time": "6h 12m 48s", "remaining_time": "41m 33s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101081} +{"loss": 0.41560078, "grad_norm": 0.04148083, "learning_rate": 2.59e-06, "epoch": 0.9003881, "global_step/max_steps": "2262/2513", "percentage": "90.01%", "elapsed_time": "6h 12m 58s", "remaining_time": "41m 23s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10108} +{"loss": 0.35127276, "grad_norm": 0.03838582, "learning_rate": 2.57e-06, "epoch": 0.90078615, "global_step/max_steps": "2263/2513", "percentage": "90.05%", "elapsed_time": "6h 13m 8s", "remaining_time": "41m 13s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101081} +{"loss": 0.35800761, "grad_norm": 0.03593715, "learning_rate": 2.55e-06, "epoch": 0.9011842, "global_step/max_steps": "2264/2513", "percentage": "90.09%", "elapsed_time": "6h 13m 18s", "remaining_time": "41m 3s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101079} +{"loss": 0.35945469, "grad_norm": 0.04078573, "learning_rate": 2.53e-06, "epoch": 0.90158225, "global_step/max_steps": "2265/2513", "percentage": "90.13%", "elapsed_time": "6h 13m 28s", "remaining_time": "40m 53s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101076} +{"loss": 0.34527776, "grad_norm": 0.04223143, "learning_rate": 2.51e-06, "epoch": 0.9019803, "global_step/max_steps": "2266/2513", "percentage": "90.17%", "elapsed_time": "6h 13m 39s", "remaining_time": "40m 43s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101072} +{"loss": 0.32971302, "grad_norm": 0.04139362, "learning_rate": 2.49e-06, "epoch": 0.90237835, "global_step/max_steps": "2267/2513", "percentage": "90.21%", "elapsed_time": "6h 13m 49s", "remaining_time": "40m 33s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101071} +{"loss": 0.32308063, "grad_norm": 0.03800265, "learning_rate": 2.47e-06, "epoch": 0.9027764, "global_step/max_steps": "2268/2513", "percentage": "90.25%", "elapsed_time": "6h 13m 59s", "remaining_time": "40m 24s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101072} +{"loss": 0.37794253, "grad_norm": 0.04097464, "learning_rate": 2.45e-06, "epoch": 0.90317445, "global_step/max_steps": "2269/2513", "percentage": "90.29%", "elapsed_time": "6h 14m 8s", "remaining_time": "40m 14s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101074} +{"loss": 0.36214212, "grad_norm": 0.04060321, "learning_rate": 2.43e-06, "epoch": 0.90357249, "global_step/max_steps": "2270/2513", "percentage": "90.33%", "elapsed_time": "6h 14m 18s", "remaining_time": "40m 4s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101074} +{"loss": 0.33523506, "grad_norm": 0.0418859, "learning_rate": 2.41e-06, "epoch": 0.90397054, "global_step/max_steps": "2271/2513", "percentage": "90.37%", "elapsed_time": "6h 14m 28s", "remaining_time": "39m 54s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101075} +{"loss": 0.32962832, "grad_norm": 0.03599798, "learning_rate": 2.39e-06, "epoch": 0.90436859, "global_step/max_steps": "2272/2513", "percentage": "90.41%", "elapsed_time": "6h 14m 38s", "remaining_time": "39m 44s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101076} +{"loss": 0.36390254, "grad_norm": 0.04024254, "learning_rate": 2.37e-06, "epoch": 0.90476664, "global_step/max_steps": "2273/2513", "percentage": "90.45%", "elapsed_time": "6h 14m 48s", "remaining_time": "39m 34s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101076} +{"loss": 0.40286541, "grad_norm": 0.0430071, "learning_rate": 2.35e-06, "epoch": 0.90516469, "global_step/max_steps": "2274/2513", "percentage": "90.49%", "elapsed_time": "6h 14m 57s", "remaining_time": "39m 24s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101078} +{"loss": 0.36374265, "grad_norm": 0.04241627, "learning_rate": 2.33e-06, "epoch": 0.90556274, "global_step/max_steps": "2275/2513", "percentage": "90.53%", "elapsed_time": "6h 15m 7s", "remaining_time": "39m 14s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101078} +{"loss": 0.38054872, "grad_norm": 0.04053234, "learning_rate": 2.32e-06, "epoch": 0.90596079, "global_step/max_steps": "2276/2513", "percentage": "90.57%", "elapsed_time": "6h 15m 17s", "remaining_time": "39m 4s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101079} +{"loss": 0.34706214, "grad_norm": 0.04053876, "learning_rate": 2.3e-06, "epoch": 0.90635884, "global_step/max_steps": "2277/2513", "percentage": "90.61%", "elapsed_time": "6h 15m 26s", "remaining_time": "38m 54s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101079} +{"loss": 0.39276174, "grad_norm": 0.03890637, "learning_rate": 2.28e-06, "epoch": 0.90675689, "global_step/max_steps": "2278/2513", "percentage": "90.65%", "elapsed_time": "6h 15m 36s", "remaining_time": "38m 44s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101078} +{"loss": 0.35797739, "grad_norm": 0.04116286, "learning_rate": 2.26e-06, "epoch": 0.90715494, "global_step/max_steps": "2279/2513", "percentage": "90.69%", "elapsed_time": "6h 15m 47s", "remaining_time": "38m 35s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101078} +{"loss": 0.35574582, "grad_norm": 0.04430363, "learning_rate": 2.24e-06, "epoch": 0.90755299, "global_step/max_steps": "2280/2513", "percentage": "90.73%", "elapsed_time": "6h 15m 57s", "remaining_time": "38m 25s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101077} +{"loss": 0.33935666, "grad_norm": 0.0405487, "learning_rate": 2.22e-06, "epoch": 0.90795104, "global_step/max_steps": "2281/2513", "percentage": "90.77%", "elapsed_time": "6h 16m 7s", "remaining_time": "38m 15s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101074} +{"loss": 0.37190425, "grad_norm": 0.04142748, "learning_rate": 2.2e-06, "epoch": 0.90834909, "global_step/max_steps": "2282/2513", "percentage": "90.81%", "elapsed_time": "6h 16m 17s", "remaining_time": "38m 5s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101075} +{"loss": 0.33104166, "grad_norm": 0.04036277, "learning_rate": 2.18e-06, "epoch": 0.90874714, "global_step/max_steps": "2283/2513", "percentage": "90.85%", "elapsed_time": "6h 16m 27s", "remaining_time": "37m 55s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101075} +{"loss": 0.36250615, "grad_norm": 0.03748367, "learning_rate": 2.16e-06, "epoch": 0.90914519, "global_step/max_steps": "2284/2513", "percentage": "90.89%", "elapsed_time": "6h 16m 36s", "remaining_time": "37m 45s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101075} +{"loss": 0.36618155, "grad_norm": 0.05130217, "learning_rate": 2.14e-06, "epoch": 0.90954324, "global_step/max_steps": "2285/2513", "percentage": "90.93%", "elapsed_time": "6h 16m 46s", "remaining_time": "37m 35s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101077} +{"loss": 0.37342227, "grad_norm": 0.03995247, "learning_rate": 2.13e-06, "epoch": 0.90994129, "global_step/max_steps": "2286/2513", "percentage": "90.97%", "elapsed_time": "6h 16m 56s", "remaining_time": "37m 25s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101078} +{"loss": 0.37165105, "grad_norm": 0.04011515, "learning_rate": 2.11e-06, "epoch": 0.91033934, "global_step/max_steps": "2287/2513", "percentage": "91.01%", "elapsed_time": "6h 17m 5s", "remaining_time": "37m 15s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101079} +{"loss": 0.37005189, "grad_norm": 0.0419786, "learning_rate": 2.09e-06, "epoch": 0.91073739, "global_step/max_steps": "2288/2513", "percentage": "91.05%", "elapsed_time": "6h 17m 15s", "remaining_time": "37m 5s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10108} +{"loss": 0.39621967, "grad_norm": 0.03906064, "learning_rate": 2.07e-06, "epoch": 0.91113544, "global_step/max_steps": "2289/2513", "percentage": "91.09%", "elapsed_time": "6h 17m 25s", "remaining_time": "36m 56s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101079} +{"loss": 0.36071128, "grad_norm": 0.04155518, "learning_rate": 2.05e-06, "epoch": 0.91153349, "global_step/max_steps": "2290/2513", "percentage": "91.13%", "elapsed_time": "6h 17m 35s", "remaining_time": "36m 46s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101079} +{"loss": 0.39086241, "grad_norm": 0.04061673, "learning_rate": 2.03e-06, "epoch": 0.91193154, "global_step/max_steps": "2291/2513", "percentage": "91.17%", "elapsed_time": "6h 17m 45s", "remaining_time": "36m 36s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10108} +{"loss": 0.33939528, "grad_norm": 0.03880685, "learning_rate": 2.02e-06, "epoch": 0.91232959, "global_step/max_steps": "2292/2513", "percentage": "91.21%", "elapsed_time": "6h 17m 54s", "remaining_time": "36m 26s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101082} +{"loss": 0.38026813, "grad_norm": 0.04236106, "learning_rate": 2e-06, "epoch": 0.91272763, "global_step/max_steps": "2293/2513", "percentage": "91.25%", "elapsed_time": "6h 18m 4s", "remaining_time": "36m 16s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101081} +{"loss": 0.3615261, "grad_norm": 0.04093775, "learning_rate": 1.98e-06, "epoch": 0.91312568, "global_step/max_steps": "2294/2513", "percentage": "91.29%", "elapsed_time": "6h 18m 14s", "remaining_time": "36m 6s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101081} +{"loss": 0.3776471, "grad_norm": 0.04144548, "learning_rate": 1.96e-06, "epoch": 0.91352373, "global_step/max_steps": "2295/2513", "percentage": "91.33%", "elapsed_time": "6h 18m 25s", "remaining_time": "35m 56s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101077} +{"loss": 0.40605325, "grad_norm": 0.04297891, "learning_rate": 1.94e-06, "epoch": 0.91392178, "global_step/max_steps": "2296/2513", "percentage": "91.36%", "elapsed_time": "6h 18m 35s", "remaining_time": "35m 46s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101076} +{"loss": 0.36857826, "grad_norm": 0.03936837, "learning_rate": 1.93e-06, "epoch": 0.91431983, "global_step/max_steps": "2297/2513", "percentage": "91.40%", "elapsed_time": "6h 18m 45s", "remaining_time": "35m 37s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101076} +{"loss": 0.36139214, "grad_norm": 0.04007238, "learning_rate": 1.91e-06, "epoch": 0.91471788, "global_step/max_steps": "2298/2513", "percentage": "91.44%", "elapsed_time": "6h 18m 55s", "remaining_time": "35m 27s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101076} +{"loss": 0.34474704, "grad_norm": 0.04181386, "learning_rate": 1.89e-06, "epoch": 0.91511593, "global_step/max_steps": "2299/2513", "percentage": "91.48%", "elapsed_time": "6h 19m 5s", "remaining_time": "35m 17s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101075} +{"loss": 0.39937466, "grad_norm": 0.0427062, "learning_rate": 1.87e-06, "epoch": 0.91551398, "global_step/max_steps": "2300/2513", "percentage": "91.52%", "elapsed_time": "6h 19m 15s", "remaining_time": "35m 7s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101076} +{"loss": 0.34876651, "grad_norm": 0.04071445, "learning_rate": 1.86e-06, "epoch": 0.91591203, "global_step/max_steps": "2301/2513", "percentage": "91.56%", "elapsed_time": "6h 19m 26s", "remaining_time": "34m 57s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10107} +{"loss": 0.33937454, "grad_norm": 0.03956643, "learning_rate": 1.84e-06, "epoch": 0.91631008, "global_step/max_steps": "2302/2513", "percentage": "91.60%", "elapsed_time": "6h 19m 36s", "remaining_time": "34m 47s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10107} +{"loss": 0.33821669, "grad_norm": 0.03763698, "learning_rate": 1.82e-06, "epoch": 0.91670813, "global_step/max_steps": "2303/2513", "percentage": "91.64%", "elapsed_time": "6h 19m 45s", "remaining_time": "34m 37s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101072} +{"loss": 0.3445918, "grad_norm": 0.0420421, "learning_rate": 1.8e-06, "epoch": 0.91710618, "global_step/max_steps": "2304/2513", "percentage": "91.68%", "elapsed_time": "6h 19m 56s", "remaining_time": "34m 27s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10107} +{"loss": 0.36559412, "grad_norm": 0.03885804, "learning_rate": 1.79e-06, "epoch": 0.91750423, "global_step/max_steps": "2305/2513", "percentage": "91.72%", "elapsed_time": "6h 20m 6s", "remaining_time": "34m 17s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10107} +{"loss": 0.31613851, "grad_norm": 0.0391092, "learning_rate": 1.77e-06, "epoch": 0.91790228, "global_step/max_steps": "2306/2513", "percentage": "91.76%", "elapsed_time": "6h 20m 15s", "remaining_time": "34m 8s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101071} +{"loss": 0.35485518, "grad_norm": 0.03877636, "learning_rate": 1.75e-06, "epoch": 0.91830033, "global_step/max_steps": "2307/2513", "percentage": "91.80%", "elapsed_time": "6h 20m 25s", "remaining_time": "33m 58s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101071} +{"loss": 0.3258276, "grad_norm": 0.03822896, "learning_rate": 1.74e-06, "epoch": 0.91869838, "global_step/max_steps": "2308/2513", "percentage": "91.84%", "elapsed_time": "6h 20m 35s", "remaining_time": "33m 48s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10107} +{"loss": 0.32304305, "grad_norm": 0.03978028, "learning_rate": 1.72e-06, "epoch": 0.91909643, "global_step/max_steps": "2309/2513", "percentage": "91.88%", "elapsed_time": "6h 20m 45s", "remaining_time": "33m 38s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10107} +{"loss": 0.34750715, "grad_norm": 0.04068442, "learning_rate": 1.7e-06, "epoch": 0.91949448, "global_step/max_steps": "2310/2513", "percentage": "91.92%", "elapsed_time": "6h 20m 56s", "remaining_time": "33m 28s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101066} +{"loss": 0.40368778, "grad_norm": 0.03886694, "learning_rate": 1.69e-06, "epoch": 0.91989253, "global_step/max_steps": "2311/2513", "percentage": "91.96%", "elapsed_time": "6h 21m 6s", "remaining_time": "33m 18s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101066} +{"loss": 0.36323765, "grad_norm": 0.03778246, "learning_rate": 1.67e-06, "epoch": 0.92029058, "global_step/max_steps": "2312/2513", "percentage": "92.00%", "elapsed_time": "6h 21m 16s", "remaining_time": "33m 8s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101066} +{"loss": 0.32095307, "grad_norm": 0.03727116, "learning_rate": 1.65e-06, "epoch": 0.92068863, "global_step/max_steps": "2313/2513", "percentage": "92.04%", "elapsed_time": "6h 21m 25s", "remaining_time": "32m 58s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101067} +{"loss": 0.36258116, "grad_norm": 0.04452861, "learning_rate": 1.64e-06, "epoch": 0.92108668, "global_step/max_steps": "2314/2513", "percentage": "92.08%", "elapsed_time": "6h 21m 35s", "remaining_time": "32m 48s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101069} +{"loss": 0.31803966, "grad_norm": 0.03509194, "learning_rate": 1.62e-06, "epoch": 0.92148472, "global_step/max_steps": "2315/2513", "percentage": "92.12%", "elapsed_time": "6h 21m 44s", "remaining_time": "32m 39s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10107} +{"loss": 0.372814, "grad_norm": 0.04277603, "learning_rate": 1.6e-06, "epoch": 0.92188277, "global_step/max_steps": "2316/2513", "percentage": "92.16%", "elapsed_time": "6h 21m 54s", "remaining_time": "32m 29s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101071} +{"loss": 0.36151916, "grad_norm": 0.03953929, "learning_rate": 1.59e-06, "epoch": 0.92228082, "global_step/max_steps": "2317/2513", "percentage": "92.20%", "elapsed_time": "6h 22m 4s", "remaining_time": "32m 19s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101071} +{"loss": 0.33866534, "grad_norm": 0.03882176, "learning_rate": 1.57e-06, "epoch": 0.92267887, "global_step/max_steps": "2318/2513", "percentage": "92.24%", "elapsed_time": "6h 22m 14s", "remaining_time": "32m 9s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101072} +{"loss": 0.35268423, "grad_norm": 0.03559231, "learning_rate": 1.56e-06, "epoch": 0.92307692, "global_step/max_steps": "2319/2513", "percentage": "92.28%", "elapsed_time": "6h 22m 24s", "remaining_time": "31m 59s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101071} +{"loss": 0.37416431, "grad_norm": 0.0406917, "learning_rate": 1.54e-06, "epoch": 0.92347497, "global_step/max_steps": "2320/2513", "percentage": "92.32%", "elapsed_time": "6h 22m 34s", "remaining_time": "31m 49s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101071} +{"loss": 0.30533743, "grad_norm": 0.03751959, "learning_rate": 1.52e-06, "epoch": 0.92387302, "global_step/max_steps": "2321/2513", "percentage": "92.36%", "elapsed_time": "6h 22m 43s", "remaining_time": "31m 39s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101073} +{"loss": 0.32351527, "grad_norm": 0.04068908, "learning_rate": 1.51e-06, "epoch": 0.92427107, "global_step/max_steps": "2322/2513", "percentage": "92.40%", "elapsed_time": "6h 22m 53s", "remaining_time": "31m 29s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101075} +{"loss": 0.35042393, "grad_norm": 0.03912885, "learning_rate": 1.49e-06, "epoch": 0.92466912, "global_step/max_steps": "2323/2513", "percentage": "92.44%", "elapsed_time": "6h 23m 3s", "remaining_time": "31m 19s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101074} +{"loss": 0.35271269, "grad_norm": 0.03770638, "learning_rate": 1.48e-06, "epoch": 0.92506717, "global_step/max_steps": "2324/2513", "percentage": "92.48%", "elapsed_time": "6h 23m 13s", "remaining_time": "31m 9s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101072} +{"loss": 0.38025656, "grad_norm": 0.04136386, "learning_rate": 1.46e-06, "epoch": 0.92546522, "global_step/max_steps": "2325/2513", "percentage": "92.52%", "elapsed_time": "6h 23m 24s", "remaining_time": "31m 0s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101069} +{"loss": 0.39575815, "grad_norm": 0.04151049, "learning_rate": 1.45e-06, "epoch": 0.92586327, "global_step/max_steps": "2326/2513", "percentage": "92.56%", "elapsed_time": "6h 23m 34s", "remaining_time": "30m 50s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101069} +{"loss": 0.35952765, "grad_norm": 0.04292822, "learning_rate": 1.43e-06, "epoch": 0.92626132, "global_step/max_steps": "2327/2513", "percentage": "92.60%", "elapsed_time": "6h 23m 43s", "remaining_time": "30m 40s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10107} +{"loss": 0.37900248, "grad_norm": 0.04298764, "learning_rate": 1.42e-06, "epoch": 0.92665937, "global_step/max_steps": "2328/2513", "percentage": "92.64%", "elapsed_time": "6h 23m 53s", "remaining_time": "30m 30s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101072} +{"loss": 0.37870771, "grad_norm": 0.04100342, "learning_rate": 1.4e-06, "epoch": 0.92705742, "global_step/max_steps": "2329/2513", "percentage": "92.68%", "elapsed_time": "6h 24m 2s", "remaining_time": "30m 20s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101072} +{"loss": 0.35161889, "grad_norm": 0.03834893, "learning_rate": 1.38e-06, "epoch": 0.92745547, "global_step/max_steps": "2330/2513", "percentage": "92.72%", "elapsed_time": "6h 24m 12s", "remaining_time": "30m 10s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101073} +{"loss": 0.33716995, "grad_norm": 0.0380656, "learning_rate": 1.37e-06, "epoch": 0.92785352, "global_step/max_steps": "2331/2513", "percentage": "92.76%", "elapsed_time": "6h 24m 22s", "remaining_time": "30m 0s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101074} +{"loss": 0.34911275, "grad_norm": 0.03845935, "learning_rate": 1.35e-06, "epoch": 0.92825157, "global_step/max_steps": "2332/2513", "percentage": "92.80%", "elapsed_time": "6h 24m 31s", "remaining_time": "29m 50s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101076} +{"loss": 0.39662045, "grad_norm": 0.04232449, "learning_rate": 1.34e-06, "epoch": 0.92864962, "global_step/max_steps": "2333/2513", "percentage": "92.84%", "elapsed_time": "6h 24m 42s", "remaining_time": "29m 40s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101074} +{"loss": 0.32521129, "grad_norm": 0.04058937, "learning_rate": 1.33e-06, "epoch": 0.92904767, "global_step/max_steps": "2334/2513", "percentage": "92.88%", "elapsed_time": "6h 24m 51s", "remaining_time": "29m 30s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101075} +{"loss": 0.45049649, "grad_norm": 0.04345571, "learning_rate": 1.31e-06, "epoch": 0.92944572, "global_step/max_steps": "2335/2513", "percentage": "92.92%", "elapsed_time": "6h 25m 1s", "remaining_time": "29m 21s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101077} +{"loss": 0.35118556, "grad_norm": 0.04004823, "learning_rate": 1.3e-06, "epoch": 0.92984377, "global_step/max_steps": "2336/2513", "percentage": "92.96%", "elapsed_time": "6h 25m 10s", "remaining_time": "29m 11s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101077} +{"loss": 0.32746637, "grad_norm": 0.04053207, "learning_rate": 1.28e-06, "epoch": 0.93024182, "global_step/max_steps": "2337/2513", "percentage": "93.00%", "elapsed_time": "6h 25m 20s", "remaining_time": "29m 1s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101078} +{"loss": 0.38243601, "grad_norm": 0.04118387, "learning_rate": 1.27e-06, "epoch": 0.93063986, "global_step/max_steps": "2338/2513", "percentage": "93.04%", "elapsed_time": "6h 25m 30s", "remaining_time": "28m 51s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101079} +{"loss": 0.31870872, "grad_norm": 0.03852138, "learning_rate": 1.25e-06, "epoch": 0.93103791, "global_step/max_steps": "2339/2513", "percentage": "93.08%", "elapsed_time": "6h 25m 41s", "remaining_time": "28m 41s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101075} +{"loss": 0.34626919, "grad_norm": 0.03714069, "learning_rate": 1.24e-06, "epoch": 0.93143596, "global_step/max_steps": "2340/2513", "percentage": "93.12%", "elapsed_time": "6h 25m 50s", "remaining_time": "28m 31s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101076} +{"loss": 0.38229185, "grad_norm": 0.04353078, "learning_rate": 1.22e-06, "epoch": 0.93183401, "global_step/max_steps": "2341/2513", "percentage": "93.16%", "elapsed_time": "6h 26m 0s", "remaining_time": "28m 21s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101077} +{"loss": 0.35440129, "grad_norm": 0.03950735, "learning_rate": 1.21e-06, "epoch": 0.93223206, "global_step/max_steps": "2342/2513", "percentage": "93.20%", "elapsed_time": "6h 26m 10s", "remaining_time": "28m 11s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101077} +{"loss": 0.3785947, "grad_norm": 0.04293374, "learning_rate": 1.2e-06, "epoch": 0.93263011, "global_step/max_steps": "2343/2513", "percentage": "93.24%", "elapsed_time": "6h 26m 20s", "remaining_time": "28m 1s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101077} +{"loss": 0.36053681, "grad_norm": 0.04237041, "learning_rate": 1.18e-06, "epoch": 0.93302816, "global_step/max_steps": "2344/2513", "percentage": "93.27%", "elapsed_time": "6h 26m 30s", "remaining_time": "27m 51s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101077} +{"loss": 0.38432181, "grad_norm": 0.04064271, "learning_rate": 1.17e-06, "epoch": 0.93342621, "global_step/max_steps": "2345/2513", "percentage": "93.31%", "elapsed_time": "6h 26m 39s", "remaining_time": "27m 42s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101079} +{"loss": 0.31761372, "grad_norm": 0.03972471, "learning_rate": 1.15e-06, "epoch": 0.93382426, "global_step/max_steps": "2346/2513", "percentage": "93.35%", "elapsed_time": "6h 26m 49s", "remaining_time": "27m 32s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10108} +{"loss": 0.35444987, "grad_norm": 0.03805418, "learning_rate": 1.14e-06, "epoch": 0.93422231, "global_step/max_steps": "2347/2513", "percentage": "93.39%", "elapsed_time": "6h 26m 59s", "remaining_time": "27m 22s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101081} +{"loss": 0.36155215, "grad_norm": 0.03880517, "learning_rate": 1.13e-06, "epoch": 0.93462036, "global_step/max_steps": "2348/2513", "percentage": "93.43%", "elapsed_time": "6h 27m 9s", "remaining_time": "27m 12s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10108} +{"loss": 0.35682619, "grad_norm": 0.03925406, "learning_rate": 1.11e-06, "epoch": 0.93501841, "global_step/max_steps": "2349/2513", "percentage": "93.47%", "elapsed_time": "6h 27m 18s", "remaining_time": "27m 2s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101081} +{"loss": 0.37006968, "grad_norm": 0.04190918, "learning_rate": 1.1e-06, "epoch": 0.93541646, "global_step/max_steps": "2350/2513", "percentage": "93.51%", "elapsed_time": "6h 27m 28s", "remaining_time": "26m 52s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101081} +{"loss": 0.36935604, "grad_norm": 0.04076891, "learning_rate": 1.09e-06, "epoch": 0.93581451, "global_step/max_steps": "2351/2513", "percentage": "93.55%", "elapsed_time": "6h 27m 38s", "remaining_time": "26m 42s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101081} +{"loss": 0.34833634, "grad_norm": 0.03901775, "learning_rate": 1.07e-06, "epoch": 0.93621256, "global_step/max_steps": "2352/2513", "percentage": "93.59%", "elapsed_time": "6h 27m 48s", "remaining_time": "26m 32s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10108} +{"loss": 0.39730331, "grad_norm": 0.042417, "learning_rate": 1.06e-06, "epoch": 0.93661061, "global_step/max_steps": "2353/2513", "percentage": "93.63%", "elapsed_time": "6h 27m 58s", "remaining_time": "26m 22s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10108} +{"loss": 0.39014766, "grad_norm": 0.04191247, "learning_rate": 1.05e-06, "epoch": 0.93700866, "global_step/max_steps": "2354/2513", "percentage": "93.67%", "elapsed_time": "6h 28m 8s", "remaining_time": "26m 13s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101078} +{"loss": 0.34607255, "grad_norm": 0.03851504, "learning_rate": 1.03e-06, "epoch": 0.93740671, "global_step/max_steps": "2355/2513", "percentage": "93.71%", "elapsed_time": "6h 28m 18s", "remaining_time": "26m 3s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101078} +{"loss": 0.33962965, "grad_norm": 0.03812294, "learning_rate": 1.02e-06, "epoch": 0.93780476, "global_step/max_steps": "2356/2513", "percentage": "93.75%", "elapsed_time": "6h 28m 28s", "remaining_time": "25m 53s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101077} +{"loss": 0.32129636, "grad_norm": 0.03800034, "learning_rate": 1.01e-06, "epoch": 0.93820281, "global_step/max_steps": "2357/2513", "percentage": "93.79%", "elapsed_time": "6h 28m 38s", "remaining_time": "25m 43s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101078} +{"loss": 0.41838747, "grad_norm": 0.04114883, "learning_rate": 9.9e-07, "epoch": 0.93860086, "global_step/max_steps": "2358/2513", "percentage": "93.83%", "elapsed_time": "6h 28m 48s", "remaining_time": "25m 33s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101077} +{"loss": 0.34027201, "grad_norm": 0.04147637, "learning_rate": 9.8e-07, "epoch": 0.93899891, "global_step/max_steps": "2359/2513", "percentage": "93.87%", "elapsed_time": "6h 28m 58s", "remaining_time": "25m 23s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101078} +{"loss": 0.41134495, "grad_norm": 0.04389985, "learning_rate": 9.7e-07, "epoch": 0.93939695, "global_step/max_steps": "2360/2513", "percentage": "93.91%", "elapsed_time": "6h 29m 8s", "remaining_time": "25m 13s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101079} +{"loss": 0.34925258, "grad_norm": 0.03774102, "learning_rate": 9.6e-07, "epoch": 0.939795, "global_step/max_steps": "2361/2513", "percentage": "93.95%", "elapsed_time": "6h 29m 17s", "remaining_time": "25m 3s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10108} +{"loss": 0.35928488, "grad_norm": 0.04152105, "learning_rate": 9.4e-07, "epoch": 0.94019305, "global_step/max_steps": "2362/2513", "percentage": "93.99%", "elapsed_time": "6h 29m 27s", "remaining_time": "24m 53s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101082} +{"loss": 0.37261111, "grad_norm": 0.04157441, "learning_rate": 9.3e-07, "epoch": 0.9405911, "global_step/max_steps": "2363/2513", "percentage": "94.03%", "elapsed_time": "6h 29m 37s", "remaining_time": "24m 43s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101082} +{"loss": 0.36916512, "grad_norm": 0.0382176, "learning_rate": 9.2e-07, "epoch": 0.94098915, "global_step/max_steps": "2364/2513", "percentage": "94.07%", "elapsed_time": "6h 29m 46s", "remaining_time": "24m 34s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101083} +{"loss": 0.36080468, "grad_norm": 0.04133954, "learning_rate": 9.1e-07, "epoch": 0.9413872, "global_step/max_steps": "2365/2513", "percentage": "94.11%", "elapsed_time": "6h 29m 56s", "remaining_time": "24m 24s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101082} +{"loss": 0.35129017, "grad_norm": 0.04018632, "learning_rate": 9e-07, "epoch": 0.94178525, "global_step/max_steps": "2366/2513", "percentage": "94.15%", "elapsed_time": "6h 30m 6s", "remaining_time": "24m 14s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101084} +{"loss": 0.3327139, "grad_norm": 0.0357193, "learning_rate": 8.8e-07, "epoch": 0.9421833, "global_step/max_steps": "2367/2513", "percentage": "94.19%", "elapsed_time": "6h 30m 15s", "remaining_time": "24m 4s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101086} +{"loss": 0.36158621, "grad_norm": 0.03714375, "learning_rate": 8.7e-07, "epoch": 0.94258135, "global_step/max_steps": "2368/2513", "percentage": "94.23%", "elapsed_time": "6h 30m 25s", "remaining_time": "23m 54s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101087} +{"loss": 0.37232116, "grad_norm": 0.03907057, "learning_rate": 8.6e-07, "epoch": 0.9429794, "global_step/max_steps": "2369/2513", "percentage": "94.27%", "elapsed_time": "6h 30m 35s", "remaining_time": "23m 44s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101084} +{"loss": 0.33634853, "grad_norm": 0.03942348, "learning_rate": 8.5e-07, "epoch": 0.94337745, "global_step/max_steps": "2370/2513", "percentage": "94.31%", "elapsed_time": "6h 30m 45s", "remaining_time": "23m 34s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101085} +{"loss": 0.37813228, "grad_norm": 0.04208194, "learning_rate": 8.4e-07, "epoch": 0.9437755, "global_step/max_steps": "2371/2513", "percentage": "94.35%", "elapsed_time": "6h 30m 55s", "remaining_time": "23m 24s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101085} +{"loss": 0.34923208, "grad_norm": 0.04161219, "learning_rate": 8.2e-07, "epoch": 0.94417355, "global_step/max_steps": "2372/2513", "percentage": "94.39%", "elapsed_time": "6h 31m 5s", "remaining_time": "23m 14s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101085} +{"loss": 0.36615092, "grad_norm": 0.04166213, "learning_rate": 8.1e-07, "epoch": 0.9445716, "global_step/max_steps": "2373/2513", "percentage": "94.43%", "elapsed_time": "6h 31m 14s", "remaining_time": "23m 4s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101086} +{"loss": 0.35554674, "grad_norm": 0.04110198, "learning_rate": 8e-07, "epoch": 0.94496965, "global_step/max_steps": "2374/2513", "percentage": "94.47%", "elapsed_time": "6h 31m 24s", "remaining_time": "22m 55s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101088} +{"loss": 0.36775833, "grad_norm": 0.03841995, "learning_rate": 7.9e-07, "epoch": 0.9453677, "global_step/max_steps": "2375/2513", "percentage": "94.51%", "elapsed_time": "6h 31m 33s", "remaining_time": "22m 45s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101092} +{"loss": 0.38564122, "grad_norm": 0.04472275, "learning_rate": 7.8e-07, "epoch": 0.94576575, "global_step/max_steps": "2376/2513", "percentage": "94.55%", "elapsed_time": "6h 31m 43s", "remaining_time": "22m 35s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101092} +{"loss": 0.35119003, "grad_norm": 0.04261597, "learning_rate": 7.7e-07, "epoch": 0.9461638, "global_step/max_steps": "2377/2513", "percentage": "94.59%", "elapsed_time": "6h 31m 52s", "remaining_time": "22m 25s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101093} +{"loss": 0.37496576, "grad_norm": 0.04053713, "learning_rate": 7.6e-07, "epoch": 0.94656185, "global_step/max_steps": "2378/2513", "percentage": "94.63%", "elapsed_time": "6h 32m 2s", "remaining_time": "22m 15s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101093} +{"loss": 0.39919543, "grad_norm": 0.03757986, "learning_rate": 7.4e-07, "epoch": 0.9469599, "global_step/max_steps": "2379/2513", "percentage": "94.67%", "elapsed_time": "6h 32m 12s", "remaining_time": "22m 5s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101095} +{"loss": 0.33090729, "grad_norm": 0.04106264, "learning_rate": 7.3e-07, "epoch": 0.94735795, "global_step/max_steps": "2380/2513", "percentage": "94.71%", "elapsed_time": "6h 32m 22s", "remaining_time": "21m 55s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101095} +{"loss": 0.39112014, "grad_norm": 0.04324697, "learning_rate": 7.2e-07, "epoch": 0.947756, "global_step/max_steps": "2381/2513", "percentage": "94.75%", "elapsed_time": "6h 32m 32s", "remaining_time": "21m 45s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101095} +{"loss": 0.39259177, "grad_norm": 0.03751466, "learning_rate": 7.1e-07, "epoch": 0.94815405, "global_step/max_steps": "2382/2513", "percentage": "94.79%", "elapsed_time": "6h 32m 42s", "remaining_time": "21m 35s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101094} +{"loss": 0.36136287, "grad_norm": 0.04028756, "learning_rate": 7e-07, "epoch": 0.94855209, "global_step/max_steps": "2383/2513", "percentage": "94.83%", "elapsed_time": "6h 32m 52s", "remaining_time": "21m 25s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101093} +{"loss": 0.3835378, "grad_norm": 0.03913161, "learning_rate": 6.9e-07, "epoch": 0.94895014, "global_step/max_steps": "2384/2513", "percentage": "94.87%", "elapsed_time": "6h 33m 3s", "remaining_time": "21m 16s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101089} +{"loss": 0.34877121, "grad_norm": 0.03939532, "learning_rate": 6.8e-07, "epoch": 0.94934819, "global_step/max_steps": "2385/2513", "percentage": "94.91%", "elapsed_time": "6h 33m 12s", "remaining_time": "21m 6s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10109} +{"loss": 0.34555095, "grad_norm": 0.03892179, "learning_rate": 6.7e-07, "epoch": 0.94974624, "global_step/max_steps": "2386/2513", "percentage": "94.95%", "elapsed_time": "6h 33m 22s", "remaining_time": "20m 56s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10109} +{"loss": 0.39180022, "grad_norm": 0.04131938, "learning_rate": 6.6e-07, "epoch": 0.95014429, "global_step/max_steps": "2387/2513", "percentage": "94.99%", "elapsed_time": "6h 33m 32s", "remaining_time": "20m 46s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101092} +{"loss": 0.37276644, "grad_norm": 0.03815892, "learning_rate": 6.5e-07, "epoch": 0.95054234, "global_step/max_steps": "2388/2513", "percentage": "95.03%", "elapsed_time": "6h 33m 41s", "remaining_time": "20m 36s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101094} +{"loss": 0.36109895, "grad_norm": 0.04265879, "learning_rate": 6.4e-07, "epoch": 0.95094039, "global_step/max_steps": "2389/2513", "percentage": "95.07%", "elapsed_time": "6h 33m 51s", "remaining_time": "20m 26s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101093} +{"loss": 0.38555294, "grad_norm": 0.04147407, "learning_rate": 6.3e-07, "epoch": 0.95133844, "global_step/max_steps": "2390/2513", "percentage": "95.11%", "elapsed_time": "6h 34m 1s", "remaining_time": "20m 16s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101094} +{"loss": 0.35941046, "grad_norm": 0.03926016, "learning_rate": 6.2e-07, "epoch": 0.95173649, "global_step/max_steps": "2391/2513", "percentage": "95.15%", "elapsed_time": "6h 34m 11s", "remaining_time": "20m 6s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101094} +{"loss": 0.34895092, "grad_norm": 0.04185753, "learning_rate": 6.1e-07, "epoch": 0.95213454, "global_step/max_steps": "2392/2513", "percentage": "95.19%", "elapsed_time": "6h 34m 20s", "remaining_time": "19m 56s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101095} +{"loss": 0.36087129, "grad_norm": 0.03904166, "learning_rate": 6e-07, "epoch": 0.95253259, "global_step/max_steps": "2393/2513", "percentage": "95.22%", "elapsed_time": "6h 34m 30s", "remaining_time": "19m 47s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101094} +{"loss": 0.32956189, "grad_norm": 0.03739024, "learning_rate": 5.9e-07, "epoch": 0.95293064, "global_step/max_steps": "2394/2513", "percentage": "95.26%", "elapsed_time": "6h 34m 40s", "remaining_time": "19m 37s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101096} +{"loss": 0.37316823, "grad_norm": 0.04031613, "learning_rate": 5.8e-07, "epoch": 0.95332869, "global_step/max_steps": "2395/2513", "percentage": "95.30%", "elapsed_time": "6h 34m 50s", "remaining_time": "19m 27s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101095} +{"loss": 0.38456589, "grad_norm": 0.04026843, "learning_rate": 5.7e-07, "epoch": 0.95372674, "global_step/max_steps": "2396/2513", "percentage": "95.34%", "elapsed_time": "6h 35m 0s", "remaining_time": "19m 17s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101095} +{"loss": 0.41073257, "grad_norm": 0.04164356, "learning_rate": 5.6e-07, "epoch": 0.95412479, "global_step/max_steps": "2397/2513", "percentage": "95.38%", "elapsed_time": "6h 35m 10s", "remaining_time": "19m 7s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101094} +{"loss": 0.38050663, "grad_norm": 0.04480749, "learning_rate": 5.5e-07, "epoch": 0.95452284, "global_step/max_steps": "2398/2513", "percentage": "95.42%", "elapsed_time": "6h 35m 20s", "remaining_time": "18m 57s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101094} +{"loss": 0.29678121, "grad_norm": 0.03849535, "learning_rate": 5.4e-07, "epoch": 0.95492089, "global_step/max_steps": "2399/2513", "percentage": "95.46%", "elapsed_time": "6h 35m 30s", "remaining_time": "18m 47s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101092} +{"loss": 0.35244429, "grad_norm": 0.03949043, "learning_rate": 5.3e-07, "epoch": 0.95531894, "global_step/max_steps": "2400/2513", "percentage": "95.50%", "elapsed_time": "6h 35m 40s", "remaining_time": "18m 37s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101093} +{"loss": 0.40524226, "grad_norm": 0.04149273, "learning_rate": 5.2e-07, "epoch": 0.95571699, "global_step/max_steps": "2401/2513", "percentage": "95.54%", "elapsed_time": "6h 35m 51s", "remaining_time": "18m 27s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101087} +{"loss": 0.36410797, "grad_norm": 0.03775362, "learning_rate": 5.1e-07, "epoch": 0.95611504, "global_step/max_steps": "2402/2513", "percentage": "95.58%", "elapsed_time": "6h 36m 1s", "remaining_time": "18m 18s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101088} +{"loss": 0.35580438, "grad_norm": 0.04125267, "learning_rate": 5e-07, "epoch": 0.95651309, "global_step/max_steps": "2403/2513", "percentage": "95.62%", "elapsed_time": "6h 36m 11s", "remaining_time": "18m 8s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101089} +{"loss": 0.32058328, "grad_norm": 0.03839222, "learning_rate": 4.9e-07, "epoch": 0.95691114, "global_step/max_steps": "2404/2513", "percentage": "95.66%", "elapsed_time": "6h 36m 20s", "remaining_time": "17m 58s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101089} +{"loss": 0.35519528, "grad_norm": 0.04378365, "learning_rate": 4.8e-07, "epoch": 0.95730918, "global_step/max_steps": "2405/2513", "percentage": "95.70%", "elapsed_time": "6h 36m 30s", "remaining_time": "17m 48s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101091} +{"loss": 0.35669029, "grad_norm": 0.04444516, "learning_rate": 4.7e-07, "epoch": 0.95770723, "global_step/max_steps": "2406/2513", "percentage": "95.74%", "elapsed_time": "6h 36m 39s", "remaining_time": "17m 38s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101093} +{"loss": 0.36643177, "grad_norm": 0.04185406, "learning_rate": 4.7e-07, "epoch": 0.95810528, "global_step/max_steps": "2407/2513", "percentage": "95.78%", "elapsed_time": "6h 36m 49s", "remaining_time": "17m 28s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101092} +{"loss": 0.36393964, "grad_norm": 0.04192755, "learning_rate": 4.6e-07, "epoch": 0.95850333, "global_step/max_steps": "2408/2513", "percentage": "95.82%", "elapsed_time": "6h 36m 59s", "remaining_time": "17m 18s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101095} +{"loss": 0.38664556, "grad_norm": 0.04152222, "learning_rate": 4.5e-07, "epoch": 0.95890138, "global_step/max_steps": "2409/2513", "percentage": "95.86%", "elapsed_time": "6h 37m 8s", "remaining_time": "17m 8s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101096} +{"loss": 0.34916133, "grad_norm": 0.03669818, "learning_rate": 4.4e-07, "epoch": 0.95929943, "global_step/max_steps": "2410/2513", "percentage": "95.90%", "elapsed_time": "6h 37m 19s", "remaining_time": "16m 58s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101094} +{"loss": 0.40478939, "grad_norm": 0.04390503, "learning_rate": 4.3e-07, "epoch": 0.95969748, "global_step/max_steps": "2411/2513", "percentage": "95.94%", "elapsed_time": "6h 37m 28s", "remaining_time": "16m 48s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101095} +{"loss": 0.40016967, "grad_norm": 0.04371899, "learning_rate": 4.2e-07, "epoch": 0.96009553, "global_step/max_steps": "2412/2513", "percentage": "95.98%", "elapsed_time": "6h 37m 38s", "remaining_time": "16m 39s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101095} +{"loss": 0.34790862, "grad_norm": 0.04183855, "learning_rate": 4.1e-07, "epoch": 0.96049358, "global_step/max_steps": "2413/2513", "percentage": "96.02%", "elapsed_time": "6h 37m 49s", "remaining_time": "16m 29s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101093} +{"loss": 0.35914248, "grad_norm": 0.04017278, "learning_rate": 4.1e-07, "epoch": 0.96089163, "global_step/max_steps": "2414/2513", "percentage": "96.06%", "elapsed_time": "6h 37m 59s", "remaining_time": "16m 19s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101092} +{"loss": 0.37020493, "grad_norm": 0.03896724, "learning_rate": 4e-07, "epoch": 0.96128968, "global_step/max_steps": "2415/2513", "percentage": "96.10%", "elapsed_time": "6h 38m 9s", "remaining_time": "16m 9s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101092} +{"loss": 0.36602402, "grad_norm": 0.04008624, "learning_rate": 3.9e-07, "epoch": 0.96168773, "global_step/max_steps": "2416/2513", "percentage": "96.14%", "elapsed_time": "6h 38m 18s", "remaining_time": "15m 59s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101094} +{"loss": 0.35130784, "grad_norm": 0.03829276, "learning_rate": 3.8e-07, "epoch": 0.96208578, "global_step/max_steps": "2417/2513", "percentage": "96.18%", "elapsed_time": "6h 38m 28s", "remaining_time": "15m 49s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101094} +{"loss": 0.40820184, "grad_norm": 0.0425351, "learning_rate": 3.7e-07, "epoch": 0.96248383, "global_step/max_steps": "2418/2513", "percentage": "96.22%", "elapsed_time": "6h 38m 38s", "remaining_time": "15m 39s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101095} +{"loss": 0.37923878, "grad_norm": 0.04272079, "learning_rate": 3.7e-07, "epoch": 0.96288188, "global_step/max_steps": "2419/2513", "percentage": "96.26%", "elapsed_time": "6h 38m 47s", "remaining_time": "15m 29s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101096} +{"loss": 0.37518451, "grad_norm": 0.03982388, "learning_rate": 3.6e-07, "epoch": 0.96327993, "global_step/max_steps": "2420/2513", "percentage": "96.30%", "elapsed_time": "6h 38m 57s", "remaining_time": "15m 19s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101097} +{"loss": 0.33693361, "grad_norm": 0.03714872, "learning_rate": 3.5e-07, "epoch": 0.96367798, "global_step/max_steps": "2421/2513", "percentage": "96.34%", "elapsed_time": "6h 39m 7s", "remaining_time": "15m 10s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101097} +{"loss": 0.36756214, "grad_norm": 0.04006861, "learning_rate": 3.4e-07, "epoch": 0.96407603, "global_step/max_steps": "2422/2513", "percentage": "96.38%", "elapsed_time": "6h 39m 17s", "remaining_time": "15m 0s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101097} +{"loss": 0.36346009, "grad_norm": 0.04211525, "learning_rate": 3.4e-07, "epoch": 0.96447408, "global_step/max_steps": "2423/2513", "percentage": "96.42%", "elapsed_time": "6h 39m 26s", "remaining_time": "14m 50s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101099} +{"loss": 0.36829334, "grad_norm": 0.04307553, "learning_rate": 3.3e-07, "epoch": 0.96487213, "global_step/max_steps": "2424/2513", "percentage": "96.46%", "elapsed_time": "6h 39m 36s", "remaining_time": "14m 40s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101099} +{"loss": 0.34866554, "grad_norm": 0.04074364, "learning_rate": 3.2e-07, "epoch": 0.96527018, "global_step/max_steps": "2425/2513", "percentage": "96.50%", "elapsed_time": "6h 39m 46s", "remaining_time": "14m 30s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101097} +{"loss": 0.3433246, "grad_norm": 0.03876328, "learning_rate": 3.1e-07, "epoch": 0.96566823, "global_step/max_steps": "2426/2513", "percentage": "96.54%", "elapsed_time": "6h 39m 56s", "remaining_time": "14m 20s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101097} +{"loss": 0.31501496, "grad_norm": 0.0387354, "learning_rate": 3.1e-07, "epoch": 0.96606628, "global_step/max_steps": "2427/2513", "percentage": "96.58%", "elapsed_time": "6h 40m 6s", "remaining_time": "14m 10s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101099} +{"loss": 0.36158657, "grad_norm": 0.03956446, "learning_rate": 3e-07, "epoch": 0.96646432, "global_step/max_steps": "2428/2513", "percentage": "96.62%", "elapsed_time": "6h 40m 16s", "remaining_time": "14m 0s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101096} +{"loss": 0.34443161, "grad_norm": 0.03613061, "learning_rate": 2.9e-07, "epoch": 0.96686237, "global_step/max_steps": "2429/2513", "percentage": "96.66%", "elapsed_time": "6h 40m 27s", "remaining_time": "13m 50s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101095} +{"loss": 0.35088283, "grad_norm": 0.04041115, "learning_rate": 2.9e-07, "epoch": 0.96726042, "global_step/max_steps": "2430/2513", "percentage": "96.70%", "elapsed_time": "6h 40m 36s", "remaining_time": "13m 41s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101096} +{"loss": 0.3514255, "grad_norm": 0.03796744, "learning_rate": 2.8e-07, "epoch": 0.96765847, "global_step/max_steps": "2431/2513", "percentage": "96.74%", "elapsed_time": "6h 40m 46s", "remaining_time": "13m 31s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101096} +{"loss": 0.37111676, "grad_norm": 0.03971968, "learning_rate": 2.7e-07, "epoch": 0.96805652, "global_step/max_steps": "2432/2513", "percentage": "96.78%", "elapsed_time": "6h 40m 56s", "remaining_time": "13m 21s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101097} +{"loss": 0.34118137, "grad_norm": 0.03600189, "learning_rate": 2.7e-07, "epoch": 0.96845457, "global_step/max_steps": "2433/2513", "percentage": "96.82%", "elapsed_time": "6h 41m 5s", "remaining_time": "13m 11s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101098} +{"loss": 0.34862149, "grad_norm": 0.0372078, "learning_rate": 2.6e-07, "epoch": 0.96885262, "global_step/max_steps": "2434/2513", "percentage": "96.86%", "elapsed_time": "6h 41m 15s", "remaining_time": "13m 1s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101098} +{"loss": 0.35761368, "grad_norm": 0.0428785, "learning_rate": 2.5e-07, "epoch": 0.96925067, "global_step/max_steps": "2435/2513", "percentage": "96.90%", "elapsed_time": "6h 41m 25s", "remaining_time": "12m 51s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101099} +{"loss": 0.38516122, "grad_norm": 0.04179702, "learning_rate": 2.5e-07, "epoch": 0.96964872, "global_step/max_steps": "2436/2513", "percentage": "96.94%", "elapsed_time": "6h 41m 34s", "remaining_time": "12m 41s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.1011} +{"loss": 0.3851552, "grad_norm": 0.03913363, "learning_rate": 2.4e-07, "epoch": 0.97004677, "global_step/max_steps": "2437/2513", "percentage": "96.98%", "elapsed_time": "6h 41m 45s", "remaining_time": "12m 31s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101099} +{"loss": 0.32142413, "grad_norm": 0.03878484, "learning_rate": 2.3e-07, "epoch": 0.97044482, "global_step/max_steps": "2438/2513", "percentage": "97.02%", "elapsed_time": "6h 41m 54s", "remaining_time": "12m 21s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.1011} +{"loss": 0.40969428, "grad_norm": 0.04041745, "learning_rate": 2.3e-07, "epoch": 0.97084287, "global_step/max_steps": "2439/2513", "percentage": "97.06%", "elapsed_time": "6h 42m 4s", "remaining_time": "12m 11s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101099} +{"loss": 0.34962061, "grad_norm": 0.03502026, "learning_rate": 2.2e-07, "epoch": 0.97124092, "global_step/max_steps": "2440/2513", "percentage": "97.10%", "elapsed_time": "6h 42m 14s", "remaining_time": "12m 2s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101098} +{"loss": 0.35660613, "grad_norm": 0.04019842, "learning_rate": 2.2e-07, "epoch": 0.97163897, "global_step/max_steps": "2441/2513", "percentage": "97.13%", "elapsed_time": "6h 42m 24s", "remaining_time": "11m 52s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101098} +{"loss": 0.37193772, "grad_norm": 0.04142248, "learning_rate": 2.1e-07, "epoch": 0.97203702, "global_step/max_steps": "2442/2513", "percentage": "97.17%", "elapsed_time": "6h 42m 34s", "remaining_time": "11m 42s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101097} +{"loss": 0.3974697, "grad_norm": 0.04220326, "learning_rate": 2e-07, "epoch": 0.97243507, "global_step/max_steps": "2443/2513", "percentage": "97.21%", "elapsed_time": "6h 42m 44s", "remaining_time": "11m 32s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101098} +{"loss": 0.36785996, "grad_norm": 0.04152654, "learning_rate": 2e-07, "epoch": 0.97283312, "global_step/max_steps": "2444/2513", "percentage": "97.25%", "elapsed_time": "6h 42m 54s", "remaining_time": "11m 22s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101097} +{"loss": 0.39894795, "grad_norm": 0.04257893, "learning_rate": 1.9e-07, "epoch": 0.97323117, "global_step/max_steps": "2445/2513", "percentage": "97.29%", "elapsed_time": "6h 43m 4s", "remaining_time": "11m 12s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101098} +{"loss": 0.38103253, "grad_norm": 0.04233173, "learning_rate": 1.9e-07, "epoch": 0.97362922, "global_step/max_steps": "2446/2513", "percentage": "97.33%", "elapsed_time": "6h 43m 14s", "remaining_time": "11m 2s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101098} +{"loss": 0.40024728, "grad_norm": 0.04162109, "learning_rate": 1.8e-07, "epoch": 0.97402727, "global_step/max_steps": "2447/2513", "percentage": "97.37%", "elapsed_time": "6h 43m 24s", "remaining_time": "10m 52s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101099} +{"loss": 0.39775127, "grad_norm": 0.04044888, "learning_rate": 1.8e-07, "epoch": 0.97442532, "global_step/max_steps": "2448/2513", "percentage": "97.41%", "elapsed_time": "6h 43m 33s", "remaining_time": "10m 42s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101101} +{"loss": 0.37795109, "grad_norm": 0.03877995, "learning_rate": 1.7e-07, "epoch": 0.97482337, "global_step/max_steps": "2449/2513", "percentage": "97.45%", "elapsed_time": "6h 43m 42s", "remaining_time": "10m 33s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101102} +{"loss": 0.33177349, "grad_norm": 0.03904858, "learning_rate": 1.6e-07, "epoch": 0.97522142, "global_step/max_steps": "2450/2513", "percentage": "97.49%", "elapsed_time": "6h 43m 52s", "remaining_time": "10m 23s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101103} +{"loss": 0.32672283, "grad_norm": 0.03919371, "learning_rate": 1.6e-07, "epoch": 0.97561946, "global_step/max_steps": "2451/2513", "percentage": "97.53%", "elapsed_time": "6h 44m 2s", "remaining_time": "10m 13s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101103} +{"loss": 0.37938857, "grad_norm": 0.04207526, "learning_rate": 1.5e-07, "epoch": 0.97601751, "global_step/max_steps": "2452/2513", "percentage": "97.57%", "elapsed_time": "6h 44m 12s", "remaining_time": "10m 3s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101103} +{"loss": 0.3230899, "grad_norm": 0.04176084, "learning_rate": 1.5e-07, "epoch": 0.97641556, "global_step/max_steps": "2453/2513", "percentage": "97.61%", "elapsed_time": "6h 44m 21s", "remaining_time": "9m 53s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101105} +{"loss": 0.36788899, "grad_norm": 0.03900337, "learning_rate": 1.4e-07, "epoch": 0.97681361, "global_step/max_steps": "2454/2513", "percentage": "97.65%", "elapsed_time": "6h 44m 32s", "remaining_time": "9m 43s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101103} +{"loss": 0.35815144, "grad_norm": 0.04270543, "learning_rate": 1.4e-07, "epoch": 0.97721166, "global_step/max_steps": "2455/2513", "percentage": "97.69%", "elapsed_time": "6h 44m 42s", "remaining_time": "9m 33s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101102} +{"loss": 0.31850755, "grad_norm": 0.04386235, "learning_rate": 1.3e-07, "epoch": 0.97760971, "global_step/max_steps": "2456/2513", "percentage": "97.73%", "elapsed_time": "6h 44m 52s", "remaining_time": "9m 23s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101101} +{"loss": 0.37188566, "grad_norm": 0.03907944, "learning_rate": 1.3e-07, "epoch": 0.97800776, "global_step/max_steps": "2457/2513", "percentage": "97.77%", "elapsed_time": "6h 45m 2s", "remaining_time": "9m 13s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101102} +{"loss": 0.35868877, "grad_norm": 0.03739635, "learning_rate": 1.3e-07, "epoch": 0.97840581, "global_step/max_steps": "2458/2513", "percentage": "97.81%", "elapsed_time": "6h 45m 12s", "remaining_time": "9m 4s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101101} +{"loss": 0.38630593, "grad_norm": 0.04316159, "learning_rate": 1.2e-07, "epoch": 0.97880386, "global_step/max_steps": "2459/2513", "percentage": "97.85%", "elapsed_time": "6h 45m 22s", "remaining_time": "8m 54s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101099} +{"loss": 0.34428334, "grad_norm": 0.03781468, "learning_rate": 1.2e-07, "epoch": 0.97920191, "global_step/max_steps": "2460/2513", "percentage": "97.89%", "elapsed_time": "6h 45m 32s", "remaining_time": "8m 44s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101099} +{"loss": 0.34553888, "grad_norm": 0.04069931, "learning_rate": 1.1e-07, "epoch": 0.97959996, "global_step/max_steps": "2461/2513", "percentage": "97.93%", "elapsed_time": "6h 45m 42s", "remaining_time": "8m 34s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.1011} +{"loss": 0.34011361, "grad_norm": 0.03697128, "learning_rate": 1.1e-07, "epoch": 0.97999801, "global_step/max_steps": "2462/2513", "percentage": "97.97%", "elapsed_time": "6h 45m 51s", "remaining_time": "8m 24s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101101} +{"loss": 0.33891988, "grad_norm": 0.03839596, "learning_rate": 1e-07, "epoch": 0.98039606, "global_step/max_steps": "2463/2513", "percentage": "98.01%", "elapsed_time": "6h 46m 1s", "remaining_time": "8m 14s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101102} +{"loss": 0.36484152, "grad_norm": 0.04169842, "learning_rate": 1e-07, "epoch": 0.98079411, "global_step/max_steps": "2464/2513", "percentage": "98.05%", "elapsed_time": "6h 46m 11s", "remaining_time": "8m 4s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101102} +{"loss": 0.34597328, "grad_norm": 0.03905229, "learning_rate": 1e-07, "epoch": 0.98119216, "global_step/max_steps": "2465/2513", "percentage": "98.09%", "elapsed_time": "6h 46m 21s", "remaining_time": "7m 54s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101103} +{"loss": 0.37323701, "grad_norm": 0.0406319, "learning_rate": 9e-08, "epoch": 0.98159021, "global_step/max_steps": "2466/2513", "percentage": "98.13%", "elapsed_time": "6h 46m 30s", "remaining_time": "7m 44s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101103} +{"loss": 0.35647875, "grad_norm": 0.03871121, "learning_rate": 9e-08, "epoch": 0.98198826, "global_step/max_steps": "2467/2513", "percentage": "98.17%", "elapsed_time": "6h 46m 40s", "remaining_time": "7m 34s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101104} +{"loss": 0.37880433, "grad_norm": 0.04014798, "learning_rate": 8e-08, "epoch": 0.98238631, "global_step/max_steps": "2468/2513", "percentage": "98.21%", "elapsed_time": "6h 46m 50s", "remaining_time": "7m 25s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101105} +{"loss": 0.35549933, "grad_norm": 0.04428937, "learning_rate": 8e-08, "epoch": 0.98278436, "global_step/max_steps": "2469/2513", "percentage": "98.25%", "elapsed_time": "6h 46m 59s", "remaining_time": "7m 15s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101106} +{"loss": 0.41722399, "grad_norm": 0.04042324, "learning_rate": 8e-08, "epoch": 0.98318241, "global_step/max_steps": "2470/2513", "percentage": "98.29%", "elapsed_time": "6h 47m 9s", "remaining_time": "7m 5s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101105} +{"loss": 0.35699934, "grad_norm": 0.04077589, "learning_rate": 7e-08, "epoch": 0.98358046, "global_step/max_steps": "2471/2513", "percentage": "98.33%", "elapsed_time": "6h 47m 19s", "remaining_time": "6m 55s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101106} +{"loss": 0.36560678, "grad_norm": 0.03848813, "learning_rate": 7e-08, "epoch": 0.98397851, "global_step/max_steps": "2472/2513", "percentage": "98.37%", "elapsed_time": "6h 47m 30s", "remaining_time": "6m 45s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101104} +{"loss": 0.3685624, "grad_norm": 0.03827019, "learning_rate": 7e-08, "epoch": 0.98437655, "global_step/max_steps": "2473/2513", "percentage": "98.41%", "elapsed_time": "6h 47m 39s", "remaining_time": "6m 35s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101105} +{"loss": 0.34799275, "grad_norm": 0.04139335, "learning_rate": 6e-08, "epoch": 0.9847746, "global_step/max_steps": "2474/2513", "percentage": "98.45%", "elapsed_time": "6h 47m 49s", "remaining_time": "6m 25s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101104} +{"loss": 0.29893813, "grad_norm": 0.03834161, "learning_rate": 6e-08, "epoch": 0.98517265, "global_step/max_steps": "2475/2513", "percentage": "98.49%", "elapsed_time": "6h 47m 59s", "remaining_time": "6m 15s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101105} +{"loss": 0.35114497, "grad_norm": 0.03886909, "learning_rate": 6e-08, "epoch": 0.9855707, "global_step/max_steps": "2476/2513", "percentage": "98.53%", "elapsed_time": "6h 48m 9s", "remaining_time": "6m 5s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101107} +{"loss": 0.36918023, "grad_norm": 0.03981522, "learning_rate": 5e-08, "epoch": 0.98596875, "global_step/max_steps": "2477/2513", "percentage": "98.57%", "elapsed_time": "6h 48m 18s", "remaining_time": "5m 56s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101108} +{"loss": 0.33084995, "grad_norm": 0.03809114, "learning_rate": 5e-08, "epoch": 0.9863668, "global_step/max_steps": "2478/2513", "percentage": "98.61%", "elapsed_time": "6h 48m 28s", "remaining_time": "5m 46s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101109} +{"loss": 0.38470966, "grad_norm": 0.03972008, "learning_rate": 5e-08, "epoch": 0.98676485, "global_step/max_steps": "2479/2513", "percentage": "98.65%", "elapsed_time": "6h 48m 37s", "remaining_time": "5m 36s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101111} +{"loss": 0.35900998, "grad_norm": 0.04322089, "learning_rate": 5e-08, "epoch": 0.9871629, "global_step/max_steps": "2480/2513", "percentage": "98.69%", "elapsed_time": "6h 48m 47s", "remaining_time": "5m 26s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10111} +{"loss": 0.34861931, "grad_norm": 0.04018058, "learning_rate": 4e-08, "epoch": 0.98756095, "global_step/max_steps": "2481/2513", "percentage": "98.73%", "elapsed_time": "6h 48m 57s", "remaining_time": "5m 16s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101111} +{"loss": 0.34154308, "grad_norm": 0.03533245, "learning_rate": 4e-08, "epoch": 0.987959, "global_step/max_steps": "2482/2513", "percentage": "98.77%", "elapsed_time": "6h 49m 7s", "remaining_time": "5m 6s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10111} +{"loss": 0.3592841, "grad_norm": 0.03960074, "learning_rate": 4e-08, "epoch": 0.98835705, "global_step/max_steps": "2483/2513", "percentage": "98.81%", "elapsed_time": "6h 49m 17s", "remaining_time": "4m 56s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101109} +{"loss": 0.39767349, "grad_norm": 0.04338061, "learning_rate": 3e-08, "epoch": 0.9887551, "global_step/max_steps": "2484/2513", "percentage": "98.85%", "elapsed_time": "6h 49m 27s", "remaining_time": "4m 46s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101109} +{"loss": 0.3578957, "grad_norm": 0.04041511, "learning_rate": 3e-08, "epoch": 0.98915315, "global_step/max_steps": "2485/2513", "percentage": "98.89%", "elapsed_time": "6h 49m 37s", "remaining_time": "4m 36s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10111} +{"loss": 0.40068182, "grad_norm": 0.04406558, "learning_rate": 3e-08, "epoch": 0.9895512, "global_step/max_steps": "2486/2513", "percentage": "98.93%", "elapsed_time": "6h 49m 47s", "remaining_time": "4m 27s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10111} +{"loss": 0.32634541, "grad_norm": 0.03961652, "learning_rate": 3e-08, "epoch": 0.98994925, "global_step/max_steps": "2487/2513", "percentage": "98.97%", "elapsed_time": "6h 49m 57s", "remaining_time": "4m 17s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101109} +{"loss": 0.36932218, "grad_norm": 0.0394087, "learning_rate": 3e-08, "epoch": 0.9903473, "global_step/max_steps": "2488/2513", "percentage": "99.01%", "elapsed_time": "6h 50m 7s", "remaining_time": "4m 7s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101108} +{"loss": 0.32034284, "grad_norm": 0.03645729, "learning_rate": 2e-08, "epoch": 0.99074535, "global_step/max_steps": "2489/2513", "percentage": "99.04%", "elapsed_time": "6h 50m 17s", "remaining_time": "3m 57s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101107} +{"loss": 0.38947442, "grad_norm": 0.04028374, "learning_rate": 2e-08, "epoch": 0.9911434, "global_step/max_steps": "2490/2513", "percentage": "99.08%", "elapsed_time": "6h 50m 27s", "remaining_time": "3m 47s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101108} +{"loss": 0.37237486, "grad_norm": 0.04132749, "learning_rate": 2e-08, "epoch": 0.99154145, "global_step/max_steps": "2491/2513", "percentage": "99.12%", "elapsed_time": "6h 50m 36s", "remaining_time": "3m 37s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10111} +{"loss": 0.38364145, "grad_norm": 0.04463814, "learning_rate": 2e-08, "epoch": 0.9919395, "global_step/max_steps": "2492/2513", "percentage": "99.16%", "elapsed_time": "6h 50m 46s", "remaining_time": "3m 27s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101109} +{"loss": 0.38533935, "grad_norm": 0.0394909, "learning_rate": 2e-08, "epoch": 0.99233755, "global_step/max_steps": "2493/2513", "percentage": "99.20%", "elapsed_time": "6h 50m 56s", "remaining_time": "3m 17s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101109} +{"loss": 0.39245373, "grad_norm": 0.03753519, "learning_rate": 1e-08, "epoch": 0.9927356, "global_step/max_steps": "2494/2513", "percentage": "99.24%", "elapsed_time": "6h 51m 6s", "remaining_time": "3m 7s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101109} +{"loss": 0.38976407, "grad_norm": 0.03870498, "learning_rate": 1e-08, "epoch": 0.99313365, "global_step/max_steps": "2495/2513", "percentage": "99.28%", "elapsed_time": "6h 51m 16s", "remaining_time": "2m 58s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101109} +{"loss": 0.32568976, "grad_norm": 0.03996669, "learning_rate": 1e-08, "epoch": 0.99353169, "global_step/max_steps": "2496/2513", "percentage": "99.32%", "elapsed_time": "6h 51m 26s", "remaining_time": "2m 48s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101109} +{"loss": 0.37703824, "grad_norm": 0.03986956, "learning_rate": 1e-08, "epoch": 0.99392974, "global_step/max_steps": "2497/2513", "percentage": "99.36%", "elapsed_time": "6h 51m 35s", "remaining_time": "2m 38s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10111} +{"loss": 0.36218682, "grad_norm": 0.04235032, "learning_rate": 1e-08, "epoch": 0.99432779, "global_step/max_steps": "2498/2513", "percentage": "99.40%", "elapsed_time": "6h 51m 46s", "remaining_time": "2m 28s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101109} +{"loss": 0.34616211, "grad_norm": 0.041739, "learning_rate": 1e-08, "epoch": 0.99472584, "global_step/max_steps": "2499/2513", "percentage": "99.44%", "elapsed_time": "6h 51m 55s", "remaining_time": "2m 18s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.10111} +{"loss": 0.40598312, "grad_norm": 0.04304638, "learning_rate": 1e-08, "epoch": 0.99512389, "global_step/max_steps": "2500/2513", "percentage": "99.48%", "elapsed_time": "6h 52m 5s", "remaining_time": "2m 8s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101109} +{"loss": 0.33172449, "grad_norm": 0.03585724, "learning_rate": 1e-08, "epoch": 0.99552194, "global_step/max_steps": "2501/2513", "percentage": "99.52%", "elapsed_time": "6h 52m 17s", "remaining_time": "1m 58s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101102} +{"loss": 0.39664519, "grad_norm": 0.03934396, "learning_rate": 1e-08, "epoch": 0.99591999, "global_step/max_steps": "2502/2513", "percentage": "99.56%", "elapsed_time": "6h 52m 27s", "remaining_time": "1m 48s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101103} +{"loss": 0.34626347, "grad_norm": 0.03824099, "learning_rate": 0.0, "epoch": 0.99631804, "global_step/max_steps": "2503/2513", "percentage": "99.60%", "elapsed_time": "6h 52m 37s", "remaining_time": "1m 38s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.1011} +{"loss": 0.35545126, "grad_norm": 0.04037773, "learning_rate": 0.0, "epoch": 0.99671609, "global_step/max_steps": "2504/2513", "percentage": "99.64%", "elapsed_time": "6h 52m 47s", "remaining_time": "1m 29s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101102} +{"loss": 0.35711318, "grad_norm": 0.04100077, "learning_rate": 0.0, "epoch": 0.99711414, "global_step/max_steps": "2505/2513", "percentage": "99.68%", "elapsed_time": "6h 52m 56s", "remaining_time": "1m 19s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101102} +{"loss": 0.34194177, "grad_norm": 0.03982453, "learning_rate": 0.0, "epoch": 0.99751219, "global_step/max_steps": "2506/2513", "percentage": "99.72%", "elapsed_time": "6h 53m 6s", "remaining_time": "1m 9s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101102} +{"loss": 0.3473115, "grad_norm": 0.04018227, "learning_rate": 0.0, "epoch": 0.99791024, "global_step/max_steps": "2507/2513", "percentage": "99.76%", "elapsed_time": "6h 53m 16s", "remaining_time": "59s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101104} +{"loss": 0.39743024, "grad_norm": 0.04152031, "learning_rate": 0.0, "epoch": 0.99830829, "global_step/max_steps": "2508/2513", "percentage": "99.80%", "elapsed_time": "6h 53m 25s", "remaining_time": "49s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101106} +{"loss": 0.38982216, "grad_norm": 0.04288373, "learning_rate": 0.0, "epoch": 0.99870634, "global_step/max_steps": "2509/2513", "percentage": "99.84%", "elapsed_time": "6h 53m 35s", "remaining_time": "39s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101107} +{"loss": 0.3862446, "grad_norm": 0.0415205, "learning_rate": 0.0, "epoch": 0.99910439, "global_step/max_steps": "2510/2513", "percentage": "99.88%", "elapsed_time": "6h 53m 45s", "remaining_time": "29s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101107} +{"loss": 0.37592769, "grad_norm": 0.04026971, "learning_rate": 0.0, "epoch": 0.99950244, "global_step/max_steps": "2511/2513", "percentage": "99.92%", "elapsed_time": "6h 53m 54s", "remaining_time": "19s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101107} +{"loss": 0.38199115, "grad_norm": 0.04123579, "learning_rate": 0.0, "epoch": 0.99990049, "global_step/max_steps": "2512/2513", "percentage": "99.96%", "elapsed_time": "6h 54m 4s", "remaining_time": "9s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101108} +{"loss": 0.31798542, "grad_norm": 0.1327611, "learning_rate": 0.0, "epoch": 1.0, "global_step/max_steps": "2513/2513", "percentage": "100.00%", "elapsed_time": "6h 54m 6s", "remaining_time": "0s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101141} +{"train_runtime": 24848.0711, "train_samples_per_second": 12.94, "train_steps_per_second": 0.101, "total_flos": 7708589808975872.0, "train_loss": 0.3871003, "epoch": 1.0, "global_step/max_steps": "2513/2513", "percentage": "100.00%", "elapsed_time": "6h 54m 8s", "remaining_time": "0s", "memory(GiB)": 77.65, "train_speed(iter/s)": 0.101135} +{"model_parameter_info": "PeftModelForCausalLM: 8453.6474M Params (161.4807M Trainable [1.9102%]), 0.0019M Buffers.", "last_model_checkpoint": "/home/ubuntu/workspace/ChartCon-v2/output/sft/full_data/lora_qwen2_5_vl/v0-20251210-172217/checkpoint-2513", "best_model_checkpoint": null, "best_metric": null, "global_step": 2513, "log_history": [{"loss": 0.994921863079071, "grad_norm": 0.21555180389638928, "learning_rate": 1.3157894736842106e-06, "epoch": 0.0003980495571698676, "step": 1}, {"loss": 0.9436769485473633, "grad_norm": 0.21890049894569963, "learning_rate": 2.631578947368421e-06, "epoch": 0.0007960991143397353, "step": 2}, {"loss": 1.0668821334838867, "grad_norm": 0.21416434617795874, "learning_rate": 3.9473684210526315e-06, "epoch": 0.001194148671509603, "step": 3}, {"loss": 0.9817889332771301, "grad_norm": 0.21602338858935055, "learning_rate": 5.263157894736842e-06, "epoch": 0.0015921982286794705, "step": 4}, {"loss": 1.061290979385376, "grad_norm": 0.2117792388622925, "learning_rate": 6.578947368421053e-06, "epoch": 0.0019902477858493384, "step": 5}, {"loss": 0.9953017830848694, "grad_norm": 0.20776735915518263, "learning_rate": 7.894736842105263e-06, "epoch": 0.002388297343019206, "step": 6}, {"loss": 0.9360724091529846, "grad_norm": 0.19928025214428335, "learning_rate": 9.210526315789474e-06, "epoch": 0.0027863469001890733, "step": 7}, {"loss": 1.0062780380249023, "grad_norm": 0.21057382341406877, "learning_rate": 1.0526315789473684e-05, "epoch": 0.003184396457358941, "step": 8}, {"loss": 1.0055251121520996, "grad_norm": 0.23786747745322556, "learning_rate": 1.1842105263157895e-05, "epoch": 0.0035824460145288087, "step": 9}, {"loss": 0.9317706823348999, "grad_norm": 0.2672223002296858, "learning_rate": 1.3157894736842106e-05, "epoch": 0.003980495571698677, "step": 10}, {"loss": 0.9052281975746155, "grad_norm": 0.2139976063802623, "learning_rate": 1.4473684210526317e-05, "epoch": 0.0043785451288685445, "step": 11}, {"loss": 0.8854976296424866, "grad_norm": 0.2245688125767743, "learning_rate": 1.5789473684210526e-05, "epoch": 0.004776594686038412, "step": 12}, {"loss": 0.8297715187072754, "grad_norm": 0.18524821255454882, "learning_rate": 1.7105263157894737e-05, "epoch": 0.00517464424320828, "step": 13}, {"loss": 0.8233564496040344, "grad_norm": 0.1307090971963668, "learning_rate": 1.8421052631578947e-05, "epoch": 0.005572693800378147, "step": 14}, {"loss": 0.8831678628921509, "grad_norm": 0.13550087141530553, "learning_rate": 1.9736842105263158e-05, "epoch": 0.005970743357548014, "step": 15}, {"loss": 0.6960184574127197, "grad_norm": 0.12782316055194146, "learning_rate": 2.105263157894737e-05, "epoch": 0.006368792914717882, "step": 16}, {"loss": 0.7162208557128906, "grad_norm": 0.13018635068676898, "learning_rate": 2.236842105263158e-05, "epoch": 0.00676684247188775, "step": 17}, {"loss": 0.7075245380401611, "grad_norm": 0.1281694191362279, "learning_rate": 2.368421052631579e-05, "epoch": 0.007164892029057617, "step": 18}, {"loss": 0.6991642117500305, "grad_norm": 0.13075612674457504, "learning_rate": 2.5e-05, "epoch": 0.007562941586227485, "step": 19}, {"loss": 0.6845373511314392, "grad_norm": 0.13933953524367226, "learning_rate": 2.6315789473684212e-05, "epoch": 0.007960991143397354, "step": 20}, {"loss": 0.681851327419281, "grad_norm": 0.09653551899912459, "learning_rate": 2.7631578947368426e-05, "epoch": 0.00835904070056722, "step": 21}, {"loss": 0.628661572933197, "grad_norm": 0.09195901707468233, "learning_rate": 2.8947368421052634e-05, "epoch": 0.008757090257737089, "step": 22}, {"loss": 0.6128085851669312, "grad_norm": 0.10400123710444766, "learning_rate": 3.0263157894736844e-05, "epoch": 0.009155139814906956, "step": 23}, {"loss": 0.6254985332489014, "grad_norm": 0.10874229141927291, "learning_rate": 3.157894736842105e-05, "epoch": 0.009553189372076824, "step": 24}, {"loss": 0.5829029083251953, "grad_norm": 0.10870521689828144, "learning_rate": 3.289473684210527e-05, "epoch": 0.009951238929246691, "step": 25}, {"loss": 0.642863392829895, "grad_norm": 0.10474670018357173, "learning_rate": 3.421052631578947e-05, "epoch": 0.01034928848641656, "step": 26}, {"loss": 0.6130627393722534, "grad_norm": 0.10712945957221291, "learning_rate": 3.5526315789473684e-05, "epoch": 0.010747338043586427, "step": 27}, {"loss": 0.5691463947296143, "grad_norm": 0.09571934726102369, "learning_rate": 3.6842105263157895e-05, "epoch": 0.011145387600756293, "step": 28}, {"loss": 0.5804609060287476, "grad_norm": 0.06699313843733441, "learning_rate": 3.815789473684211e-05, "epoch": 0.011543437157926162, "step": 29}, {"loss": 0.5339746475219727, "grad_norm": 0.04525255417569087, "learning_rate": 3.9473684210526316e-05, "epoch": 0.011941486715096029, "step": 30}, {"loss": 0.5287268161773682, "grad_norm": 0.040233285652640435, "learning_rate": 4.078947368421053e-05, "epoch": 0.012339536272265897, "step": 31}, {"loss": 0.5856388211250305, "grad_norm": 0.06605311415368409, "learning_rate": 4.210526315789474e-05, "epoch": 0.012737585829435764, "step": 32}, {"loss": 0.5421078205108643, "grad_norm": 0.046101435400230276, "learning_rate": 4.342105263157895e-05, "epoch": 0.013135635386605633, "step": 33}, {"loss": 0.5099552869796753, "grad_norm": 0.049271766231913754, "learning_rate": 4.473684210526316e-05, "epoch": 0.0135336849437755, "step": 34}, {"loss": 0.5219672918319702, "grad_norm": 0.04793801680373196, "learning_rate": 4.605263157894737e-05, "epoch": 0.013931734500945368, "step": 35}, {"loss": 0.4667450189590454, "grad_norm": 0.0382129702043581, "learning_rate": 4.736842105263158e-05, "epoch": 0.014329784058115235, "step": 36}, {"loss": 0.46464723348617554, "grad_norm": 0.03418195859193605, "learning_rate": 4.868421052631579e-05, "epoch": 0.014727833615285103, "step": 37}, {"loss": 0.5468077063560486, "grad_norm": 0.0480258252092818, "learning_rate": 5e-05, "epoch": 0.01512588317245497, "step": 38}, {"loss": 0.546057939529419, "grad_norm": 0.04874969504260967, "learning_rate": 5.131578947368422e-05, "epoch": 0.015523932729624839, "step": 39}, {"loss": 0.5315722227096558, "grad_norm": 0.04837356982475934, "learning_rate": 5.2631578947368424e-05, "epoch": 0.015921982286794707, "step": 40}, {"loss": 0.5174524784088135, "grad_norm": 0.0414993460655702, "learning_rate": 5.3947368421052635e-05, "epoch": 0.016320031843964574, "step": 41}, {"loss": 0.5410268902778625, "grad_norm": 0.04199228125812565, "learning_rate": 5.526315789473685e-05, "epoch": 0.01671808140113444, "step": 42}, {"loss": 0.543163537979126, "grad_norm": 0.039904992141745164, "learning_rate": 5.6578947368421056e-05, "epoch": 0.017116130958304308, "step": 43}, {"loss": 0.5215365290641785, "grad_norm": 0.036695353897362004, "learning_rate": 5.789473684210527e-05, "epoch": 0.017514180515474178, "step": 44}, {"loss": 0.5254760980606079, "grad_norm": 0.03882230639298054, "learning_rate": 5.921052631578947e-05, "epoch": 0.017912230072644045, "step": 45}, {"loss": 0.5239067077636719, "grad_norm": 0.04098727127041338, "learning_rate": 6.052631578947369e-05, "epoch": 0.01831027962981391, "step": 46}, {"loss": 0.44943276047706604, "grad_norm": 0.03812268070384536, "learning_rate": 6.18421052631579e-05, "epoch": 0.01870832918698378, "step": 47}, {"loss": 0.520372748374939, "grad_norm": 0.039380311063436216, "learning_rate": 6.31578947368421e-05, "epoch": 0.01910637874415365, "step": 48}, {"loss": 0.5280528664588928, "grad_norm": 0.03527931151562182, "learning_rate": 6.447368421052632e-05, "epoch": 0.019504428301323515, "step": 49}, {"loss": 0.5039846897125244, "grad_norm": 0.0362592366921348, "learning_rate": 6.578947368421054e-05, "epoch": 0.019902477858493382, "step": 50}, {"loss": 0.45674365758895874, "grad_norm": 0.035933716306113736, "learning_rate": 6.710526315789474e-05, "epoch": 0.02030052741566325, "step": 51}, {"loss": 0.5482999682426453, "grad_norm": 0.043697808760462954, "learning_rate": 6.842105263157895e-05, "epoch": 0.02069857697283312, "step": 52}, {"loss": 0.4810941815376282, "grad_norm": 0.03834308742045025, "learning_rate": 6.973684210526315e-05, "epoch": 0.021096626530002986, "step": 53}, {"loss": 0.47800877690315247, "grad_norm": 0.03573499956558447, "learning_rate": 7.105263157894737e-05, "epoch": 0.021494676087172853, "step": 54}, {"loss": 0.5019316673278809, "grad_norm": 0.03294038299966432, "learning_rate": 7.236842105263159e-05, "epoch": 0.02189272564434272, "step": 55}, {"loss": 0.512109637260437, "grad_norm": 0.0387869332667778, "learning_rate": 7.368421052631579e-05, "epoch": 0.022290775201512587, "step": 56}, {"loss": 0.5112982988357544, "grad_norm": 0.03802566099170539, "learning_rate": 7.500000000000001e-05, "epoch": 0.022688824758682457, "step": 57}, {"loss": 0.5098364949226379, "grad_norm": 0.037012930816362706, "learning_rate": 7.631578947368422e-05, "epoch": 0.023086874315852324, "step": 58}, {"loss": 0.5295729637145996, "grad_norm": 0.03734610450232538, "learning_rate": 7.763157894736843e-05, "epoch": 0.02348492387302219, "step": 59}, {"loss": 0.43766993284225464, "grad_norm": 0.03432748363999195, "learning_rate": 7.894736842105263e-05, "epoch": 0.023882973430192057, "step": 60}, {"loss": 0.46333521604537964, "grad_norm": 0.03595589380289613, "learning_rate": 8.026315789473685e-05, "epoch": 0.024281022987361928, "step": 61}, {"loss": 0.45054978132247925, "grad_norm": 0.03505860547021009, "learning_rate": 8.157894736842105e-05, "epoch": 0.024679072544531794, "step": 62}, {"loss": 0.4752957224845886, "grad_norm": 0.04025972825026733, "learning_rate": 8.289473684210527e-05, "epoch": 0.02507712210170166, "step": 63}, {"loss": 0.4241446554660797, "grad_norm": 0.03552579822656216, "learning_rate": 8.421052631578948e-05, "epoch": 0.025475171658871528, "step": 64}, {"loss": 0.4888819456100464, "grad_norm": 0.03537808114906161, "learning_rate": 8.552631578947369e-05, "epoch": 0.0258732212160414, "step": 65}, {"loss": 0.467548131942749, "grad_norm": 0.039065246310351506, "learning_rate": 8.68421052631579e-05, "epoch": 0.026271270773211265, "step": 66}, {"loss": 0.4422213137149811, "grad_norm": 0.03745347472617776, "learning_rate": 8.81578947368421e-05, "epoch": 0.026669320330381132, "step": 67}, {"loss": 0.419810950756073, "grad_norm": 0.03608848084943323, "learning_rate": 8.947368421052632e-05, "epoch": 0.027067369887551, "step": 68}, {"loss": 0.4485206604003906, "grad_norm": 0.03446097235363998, "learning_rate": 9.078947368421054e-05, "epoch": 0.02746541944472087, "step": 69}, {"loss": 0.42335641384124756, "grad_norm": 0.036747699082564096, "learning_rate": 9.210526315789474e-05, "epoch": 0.027863469001890736, "step": 70}, {"loss": 0.42593246698379517, "grad_norm": 0.03501366689380654, "learning_rate": 9.342105263157896e-05, "epoch": 0.028261518559060603, "step": 71}, {"loss": 0.4556373059749603, "grad_norm": 0.04298532876543945, "learning_rate": 9.473684210526316e-05, "epoch": 0.02865956811623047, "step": 72}, {"loss": 0.46544915437698364, "grad_norm": 0.03907184619644726, "learning_rate": 9.605263157894737e-05, "epoch": 0.02905761767340034, "step": 73}, {"loss": 0.5278407335281372, "grad_norm": 0.04066806491409933, "learning_rate": 9.736842105263158e-05, "epoch": 0.029455667230570207, "step": 74}, {"loss": 0.4684775769710541, "grad_norm": 0.038594663763478775, "learning_rate": 9.868421052631579e-05, "epoch": 0.029853716787740073, "step": 75}, {"loss": 0.4965198040008545, "grad_norm": 0.03836249144575907, "learning_rate": 0.0001, "epoch": 0.03025176634490994, "step": 76}, {"loss": 0.435840368270874, "grad_norm": 0.040426126739401505, "learning_rate": 9.999995845405561e-05, "epoch": 0.03064981590207981, "step": 77}, {"loss": 0.4301852583885193, "grad_norm": 0.0375686254754197, "learning_rate": 9.999983381629146e-05, "epoch": 0.031047865459249677, "step": 78}, {"loss": 0.44210824370384216, "grad_norm": 0.03331262631283744, "learning_rate": 9.999962608691468e-05, "epoch": 0.031445915016419544, "step": 79}, {"loss": 0.4545646011829376, "grad_norm": 0.04031183180311645, "learning_rate": 9.999933526627048e-05, "epoch": 0.031843964573589414, "step": 80}, {"loss": 0.5159822702407837, "grad_norm": 0.042135607043865236, "learning_rate": 9.999896135484217e-05, "epoch": 0.03224201413075928, "step": 81}, {"loss": 0.47242194414138794, "grad_norm": 0.038052009058556664, "learning_rate": 9.999850435325113e-05, "epoch": 0.03264006368792915, "step": 82}, {"loss": 0.4144842326641083, "grad_norm": 0.03853696089005584, "learning_rate": 9.99979642622568e-05, "epoch": 0.03303811324509901, "step": 83}, {"loss": 0.45854294300079346, "grad_norm": 0.04499345177181929, "learning_rate": 9.999734108275676e-05, "epoch": 0.03343616280226888, "step": 84}, {"loss": 0.465740442276001, "grad_norm": 0.04198587140750975, "learning_rate": 9.99966348157866e-05, "epoch": 0.03383421235943875, "step": 85}, {"loss": 0.4308220446109772, "grad_norm": 0.04056991188503894, "learning_rate": 9.999584546252003e-05, "epoch": 0.034232261916608615, "step": 86}, {"loss": 0.4903455674648285, "grad_norm": 0.03974547340598062, "learning_rate": 9.999497302426883e-05, "epoch": 0.034630311473778486, "step": 87}, {"loss": 0.43225497007369995, "grad_norm": 0.03939760993170123, "learning_rate": 9.999401750248287e-05, "epoch": 0.035028361030948356, "step": 88}, {"loss": 0.46949511766433716, "grad_norm": 0.03816381328312435, "learning_rate": 9.999297889875005e-05, "epoch": 0.03542641058811822, "step": 89}, {"loss": 0.4468074440956116, "grad_norm": 0.03871481490353871, "learning_rate": 9.999185721479635e-05, "epoch": 0.03582446014528809, "step": 90}, {"loss": 0.39418840408325195, "grad_norm": 0.03732983193686334, "learning_rate": 9.999065245248585e-05, "epoch": 0.03622250970245795, "step": 91}, {"loss": 0.4268158972263336, "grad_norm": 0.039328444980796566, "learning_rate": 9.998936461382066e-05, "epoch": 0.03662055925962782, "step": 92}, {"loss": 0.3905756175518036, "grad_norm": 0.03829150787787755, "learning_rate": 9.998799370094097e-05, "epoch": 0.03701860881679769, "step": 93}, {"loss": 0.4239282011985779, "grad_norm": 0.03774094777608143, "learning_rate": 9.998653971612497e-05, "epoch": 0.03741665837396756, "step": 94}, {"loss": 0.4725581407546997, "grad_norm": 0.04289900132922451, "learning_rate": 9.9985002661789e-05, "epoch": 0.03781470793113743, "step": 95}, {"loss": 0.4235262870788574, "grad_norm": 0.04366744574312237, "learning_rate": 9.99833825404874e-05, "epoch": 0.0382127574883073, "step": 96}, {"loss": 0.43142402172088623, "grad_norm": 0.040072666091526314, "learning_rate": 9.998167935491251e-05, "epoch": 0.03861080704547716, "step": 97}, {"loss": 0.45048943161964417, "grad_norm": 0.0445928733811468, "learning_rate": 9.997989310789475e-05, "epoch": 0.03900885660264703, "step": 98}, {"loss": 0.44291624426841736, "grad_norm": 0.0402534309559076, "learning_rate": 9.997802380240259e-05, "epoch": 0.039406906159816894, "step": 99}, {"loss": 0.4442812204360962, "grad_norm": 0.04076024595160122, "learning_rate": 9.997607144154253e-05, "epoch": 0.039804955716986765, "step": 100}, {"loss": 0.4487001299858093, "grad_norm": 0.03944969630915711, "learning_rate": 9.997403602855903e-05, "epoch": 0.040203005274156635, "step": 101}, {"loss": 0.4802417755126953, "grad_norm": 0.0444640927236358, "learning_rate": 9.997191756683466e-05, "epoch": 0.0406010548313265, "step": 102}, {"loss": 0.44365811347961426, "grad_norm": 0.03753712686476302, "learning_rate": 9.996971605988994e-05, "epoch": 0.04099910438849637, "step": 103}, {"loss": 0.4575066864490509, "grad_norm": 0.04935929951469436, "learning_rate": 9.996743151138342e-05, "epoch": 0.04139715394566624, "step": 104}, {"loss": 0.4320715665817261, "grad_norm": 0.042822767404489655, "learning_rate": 9.996506392511165e-05, "epoch": 0.0417952035028361, "step": 105}, {"loss": 0.42911651730537415, "grad_norm": 0.03503738500436561, "learning_rate": 9.996261330500916e-05, "epoch": 0.04219325306000597, "step": 106}, {"loss": 0.4515736401081085, "grad_norm": 0.04159520584906276, "learning_rate": 9.99600796551485e-05, "epoch": 0.042591302617175836, "step": 107}, {"loss": 0.4506798982620239, "grad_norm": 0.04353626817609942, "learning_rate": 9.995746297974018e-05, "epoch": 0.042989352174345706, "step": 108}, {"loss": 0.4510691165924072, "grad_norm": 0.043865784630964126, "learning_rate": 9.995476328313269e-05, "epoch": 0.043387401731515576, "step": 109}, {"loss": 0.5013118982315063, "grad_norm": 0.04238444056547549, "learning_rate": 9.99519805698125e-05, "epoch": 0.04378545128868544, "step": 110}, {"loss": 0.37856751680374146, "grad_norm": 0.038610099244767666, "learning_rate": 9.9949114844404e-05, "epoch": 0.04418350084585531, "step": 111}, {"loss": 0.40496110916137695, "grad_norm": 0.041405287461860805, "learning_rate": 9.994616611166958e-05, "epoch": 0.04458155040302517, "step": 112}, {"loss": 0.43783047795295715, "grad_norm": 0.04486701601475104, "learning_rate": 9.994313437650956e-05, "epoch": 0.044979599960195044, "step": 113}, {"loss": 0.4261724650859833, "grad_norm": 0.03791196376705789, "learning_rate": 9.994001964396218e-05, "epoch": 0.045377649517364914, "step": 114}, {"loss": 0.4090481400489807, "grad_norm": 0.04239538242384711, "learning_rate": 9.993682191920362e-05, "epoch": 0.04577569907453478, "step": 115}, {"loss": 0.43200939893722534, "grad_norm": 0.038921435276134585, "learning_rate": 9.9933541207548e-05, "epoch": 0.04617374863170465, "step": 116}, {"loss": 0.3549041450023651, "grad_norm": 0.03638755993986027, "learning_rate": 9.993017751444732e-05, "epoch": 0.04657179818887452, "step": 117}, {"loss": 0.43866395950317383, "grad_norm": 0.03842415345835967, "learning_rate": 9.992673084549147e-05, "epoch": 0.04696984774604438, "step": 118}, {"loss": 0.3637409806251526, "grad_norm": 0.04200447725480666, "learning_rate": 9.992320120640826e-05, "epoch": 0.04736789730321425, "step": 119}, {"loss": 0.41988104581832886, "grad_norm": 0.03868980893005599, "learning_rate": 9.99195886030634e-05, "epoch": 0.047765946860384115, "step": 120}, {"loss": 0.4532768130302429, "grad_norm": 0.042028417812286084, "learning_rate": 9.991589304146046e-05, "epoch": 0.048163996417553985, "step": 121}, {"loss": 0.4115150272846222, "grad_norm": 0.038309316748290086, "learning_rate": 9.991211452774084e-05, "epoch": 0.048562045974723855, "step": 122}, {"loss": 0.4357512593269348, "grad_norm": 0.040250037113612795, "learning_rate": 9.99082530681838e-05, "epoch": 0.04896009553189372, "step": 123}, {"loss": 0.42165613174438477, "grad_norm": 0.04027883668538545, "learning_rate": 9.99043086692065e-05, "epoch": 0.04935814508906359, "step": 124}, {"loss": 0.4236012101173401, "grad_norm": 0.041311609807131534, "learning_rate": 9.990028133736386e-05, "epoch": 0.04975619464623346, "step": 125}, {"loss": 0.4609946012496948, "grad_norm": 0.040071104699477576, "learning_rate": 9.989617107934868e-05, "epoch": 0.05015424420340332, "step": 126}, {"loss": 0.4831309914588928, "grad_norm": 0.039477844435037095, "learning_rate": 9.98919779019915e-05, "epoch": 0.05055229376057319, "step": 127}, {"loss": 0.47838130593299866, "grad_norm": 0.046207935104153046, "learning_rate": 9.988770181226075e-05, "epoch": 0.050950343317743056, "step": 128}, {"loss": 0.4385298788547516, "grad_norm": 0.04389549075427485, "learning_rate": 9.988334281726257e-05, "epoch": 0.051348392874912926, "step": 129}, {"loss": 0.43673020601272583, "grad_norm": 0.0376293610608306, "learning_rate": 9.98789009242409e-05, "epoch": 0.0517464424320828, "step": 130}, {"loss": 0.4273418188095093, "grad_norm": 0.040420903660230934, "learning_rate": 9.987437614057744e-05, "epoch": 0.05214449198925266, "step": 131}, {"loss": 0.42770153284072876, "grad_norm": 0.04365689056582912, "learning_rate": 9.986976847379168e-05, "epoch": 0.05254254154642253, "step": 132}, {"loss": 0.43427616357803345, "grad_norm": 0.04318428339622392, "learning_rate": 9.986507793154078e-05, "epoch": 0.0529405911035924, "step": 133}, {"loss": 0.43279463052749634, "grad_norm": 0.0389939039085004, "learning_rate": 9.986030452161968e-05, "epoch": 0.053338640660762264, "step": 134}, {"loss": 0.47584807872772217, "grad_norm": 0.041704231734765486, "learning_rate": 9.985544825196102e-05, "epoch": 0.053736690217932134, "step": 135}, {"loss": 0.41349175572395325, "grad_norm": 0.051310294620781904, "learning_rate": 9.98505091306351e-05, "epoch": 0.054134739775102, "step": 136}, {"loss": 0.4710317850112915, "grad_norm": 0.044770157733257654, "learning_rate": 9.984548716584995e-05, "epoch": 0.05453278933227187, "step": 137}, {"loss": 0.4455372095108032, "grad_norm": 0.04221402612430218, "learning_rate": 9.984038236595129e-05, "epoch": 0.05493083888944174, "step": 138}, {"loss": 0.4428879916667938, "grad_norm": 0.04525967890036551, "learning_rate": 9.983519473942245e-05, "epoch": 0.0553288884466116, "step": 139}, {"loss": 0.38288018107414246, "grad_norm": 0.0426238668970556, "learning_rate": 9.98299242948844e-05, "epoch": 0.05572693800378147, "step": 140}, {"loss": 0.40578493475914, "grad_norm": 0.03923056695920563, "learning_rate": 9.98245710410958e-05, "epoch": 0.056124987560951335, "step": 141}, {"loss": 0.4178992509841919, "grad_norm": 0.03766845459395343, "learning_rate": 9.981913498695285e-05, "epoch": 0.056523037118121205, "step": 142}, {"loss": 0.4165794253349304, "grad_norm": 0.04392187316225729, "learning_rate": 9.981361614148945e-05, "epoch": 0.056921086675291076, "step": 143}, {"loss": 0.42920127511024475, "grad_norm": 0.03932949575424485, "learning_rate": 9.980801451387697e-05, "epoch": 0.05731913623246094, "step": 144}, {"loss": 0.4276886582374573, "grad_norm": 0.04066640471734911, "learning_rate": 9.980233011342442e-05, "epoch": 0.05771718578963081, "step": 145}, {"loss": 0.4642341732978821, "grad_norm": 0.04011628081648322, "learning_rate": 9.979656294957837e-05, "epoch": 0.05811523534680068, "step": 146}, {"loss": 0.4901285469532013, "grad_norm": 0.04006902599958677, "learning_rate": 9.97907130319229e-05, "epoch": 0.05851328490397054, "step": 147}, {"loss": 0.4260837137699127, "grad_norm": 0.04280027147344821, "learning_rate": 9.978478037017961e-05, "epoch": 0.05891133446114041, "step": 148}, {"loss": 0.46243804693222046, "grad_norm": 0.03962510101583967, "learning_rate": 9.977876497420766e-05, "epoch": 0.05930938401831028, "step": 149}, {"loss": 0.4044482707977295, "grad_norm": 0.036826795107623114, "learning_rate": 9.977266685400361e-05, "epoch": 0.05970743357548015, "step": 150}, {"loss": 0.4035593867301941, "grad_norm": 0.041817108453242456, "learning_rate": 9.97664860197016e-05, "epoch": 0.06010548313265002, "step": 151}, {"loss": 0.3800760507583618, "grad_norm": 0.040062262003775384, "learning_rate": 9.976022248157313e-05, "epoch": 0.06050353268981988, "step": 152}, {"loss": 0.441882848739624, "grad_norm": 0.042244539283471785, "learning_rate": 9.975387625002719e-05, "epoch": 0.06090158224698975, "step": 153}, {"loss": 0.38021713495254517, "grad_norm": 0.039482088519213726, "learning_rate": 9.974744733561021e-05, "epoch": 0.06129963180415962, "step": 154}, {"loss": 0.4046151041984558, "grad_norm": 0.040345522586222066, "learning_rate": 9.974093574900601e-05, "epoch": 0.061697681361329484, "step": 155}, {"loss": 0.45719361305236816, "grad_norm": 0.04262316879215699, "learning_rate": 9.973434150103576e-05, "epoch": 0.062095730918499355, "step": 156}, {"loss": 0.4434029757976532, "grad_norm": 0.042022085522263775, "learning_rate": 9.972766460265804e-05, "epoch": 0.06249378047566922, "step": 157}, {"loss": 0.4208804965019226, "grad_norm": 0.043868979913042935, "learning_rate": 9.972090506496876e-05, "epoch": 0.06289183003283909, "step": 158}, {"loss": 0.4215505123138428, "grad_norm": 0.037144078477630876, "learning_rate": 9.97140628992012e-05, "epoch": 0.06328987959000895, "step": 159}, {"loss": 0.4108273983001709, "grad_norm": 0.044004924798779696, "learning_rate": 9.970713811672591e-05, "epoch": 0.06368792914717883, "step": 160}, {"loss": 0.40703824162483215, "grad_norm": 0.03960422525320676, "learning_rate": 9.970013072905078e-05, "epoch": 0.06408597870434869, "step": 161}, {"loss": 0.44042515754699707, "grad_norm": 0.04117990547139435, "learning_rate": 9.969304074782093e-05, "epoch": 0.06448402826151856, "step": 162}, {"loss": 0.4230448305606842, "grad_norm": 0.04577458461214273, "learning_rate": 9.968586818481877e-05, "epoch": 0.06488207781868843, "step": 163}, {"loss": 0.38179314136505127, "grad_norm": 0.03980219457098357, "learning_rate": 9.967861305196393e-05, "epoch": 0.0652801273758583, "step": 164}, {"loss": 0.45506608486175537, "grad_norm": 0.04363962377821095, "learning_rate": 9.967127536131326e-05, "epoch": 0.06567817693302816, "step": 165}, {"loss": 0.428062379360199, "grad_norm": 0.04431815398102832, "learning_rate": 9.96638551250608e-05, "epoch": 0.06607622649019802, "step": 166}, {"loss": 0.44230401515960693, "grad_norm": 0.04031154364410871, "learning_rate": 9.965635235553784e-05, "epoch": 0.0664742760473679, "step": 167}, {"loss": 0.3866247534751892, "grad_norm": 0.03662057926526026, "learning_rate": 9.964876706521268e-05, "epoch": 0.06687232560453776, "step": 168}, {"loss": 0.42420628666877747, "grad_norm": 0.0408593012380572, "learning_rate": 9.96410992666909e-05, "epoch": 0.06727037516170763, "step": 169}, {"loss": 0.38610824942588806, "grad_norm": 0.04358714748904212, "learning_rate": 9.96333489727151e-05, "epoch": 0.0676684247188775, "step": 170}, {"loss": 0.42256689071655273, "grad_norm": 0.044877131711787546, "learning_rate": 9.962551619616505e-05, "epoch": 0.06806647427604737, "step": 171}, {"loss": 0.3797181248664856, "grad_norm": 0.03857102863374109, "learning_rate": 9.961760095005753e-05, "epoch": 0.06846452383321723, "step": 172}, {"loss": 0.46156105399131775, "grad_norm": 0.045151743022884445, "learning_rate": 9.96096032475464e-05, "epoch": 0.06886257339038711, "step": 173}, {"loss": 0.3986797332763672, "grad_norm": 0.03983258242285698, "learning_rate": 9.960152310192255e-05, "epoch": 0.06926062294755697, "step": 174}, {"loss": 0.44861018657684326, "grad_norm": 0.04497158654161355, "learning_rate": 9.959336052661386e-05, "epoch": 0.06965867250472683, "step": 175}, {"loss": 0.4685113728046417, "grad_norm": 0.04425122089338003, "learning_rate": 9.958511553518521e-05, "epoch": 0.07005672206189671, "step": 176}, {"loss": 0.44791150093078613, "grad_norm": 0.04343736946579461, "learning_rate": 9.957678814133844e-05, "epoch": 0.07045477161906658, "step": 177}, {"loss": 0.3994130492210388, "grad_norm": 0.03739605360806228, "learning_rate": 9.956837835891232e-05, "epoch": 0.07085282117623644, "step": 178}, {"loss": 0.349687397480011, "grad_norm": 0.04127822912187052, "learning_rate": 9.955988620188258e-05, "epoch": 0.07125087073340632, "step": 179}, {"loss": 0.4158002734184265, "grad_norm": 0.04055145054185753, "learning_rate": 9.955131168436175e-05, "epoch": 0.07164892029057618, "step": 180}, {"loss": 0.41803792119026184, "grad_norm": 0.042327808081380376, "learning_rate": 9.954265482059932e-05, "epoch": 0.07204696984774604, "step": 181}, {"loss": 0.4892021417617798, "grad_norm": 0.045014717799306436, "learning_rate": 9.95339156249816e-05, "epoch": 0.0724450194049159, "step": 182}, {"loss": 0.3699420094490051, "grad_norm": 0.04224199500860969, "learning_rate": 9.95250941120317e-05, "epoch": 0.07284306896208578, "step": 183}, {"loss": 0.41931775212287903, "grad_norm": 0.04251162189336387, "learning_rate": 9.951619029640954e-05, "epoch": 0.07324111851925565, "step": 184}, {"loss": 0.45457687973976135, "grad_norm": 0.041301976820686676, "learning_rate": 9.950720419291184e-05, "epoch": 0.07363916807642551, "step": 185}, {"loss": 0.41679853200912476, "grad_norm": 0.039038996831424755, "learning_rate": 9.949813581647203e-05, "epoch": 0.07403721763359539, "step": 186}, {"loss": 0.42772388458251953, "grad_norm": 0.03830274226569726, "learning_rate": 9.948898518216028e-05, "epoch": 0.07443526719076525, "step": 187}, {"loss": 0.3990447521209717, "grad_norm": 0.03690682557914544, "learning_rate": 9.947975230518345e-05, "epoch": 0.07483331674793511, "step": 188}, {"loss": 0.4435923397541046, "grad_norm": 0.0451427352880959, "learning_rate": 9.947043720088511e-05, "epoch": 0.07523136630510499, "step": 189}, {"loss": 0.4158438444137573, "grad_norm": 0.041131869684936276, "learning_rate": 9.946103988474542e-05, "epoch": 0.07562941586227485, "step": 190}, {"loss": 0.43835365772247314, "grad_norm": 0.04583373352268476, "learning_rate": 9.945156037238122e-05, "epoch": 0.07602746541944472, "step": 191}, {"loss": 0.3938068151473999, "grad_norm": 0.041423941163045784, "learning_rate": 9.944199867954593e-05, "epoch": 0.0764255149766146, "step": 192}, {"loss": 0.41637110710144043, "grad_norm": 0.041513648325986044, "learning_rate": 9.94323548221295e-05, "epoch": 0.07682356453378446, "step": 193}, {"loss": 0.46362555027008057, "grad_norm": 0.04510411517818611, "learning_rate": 9.942262881615848e-05, "epoch": 0.07722161409095432, "step": 194}, {"loss": 0.42084819078445435, "grad_norm": 0.04305696923311579, "learning_rate": 9.94128206777959e-05, "epoch": 0.07761966364812418, "step": 195}, {"loss": 0.3859346807003021, "grad_norm": 0.03656732234869033, "learning_rate": 9.940293042334131e-05, "epoch": 0.07801771320529406, "step": 196}, {"loss": 0.44058340787887573, "grad_norm": 0.039978350687312786, "learning_rate": 9.939295806923071e-05, "epoch": 0.07841576276246393, "step": 197}, {"loss": 0.3958757221698761, "grad_norm": 0.038272960461252156, "learning_rate": 9.938290363203651e-05, "epoch": 0.07881381231963379, "step": 198}, {"loss": 0.3935411870479584, "grad_norm": 0.04259150334688577, "learning_rate": 9.937276712846757e-05, "epoch": 0.07921186187680367, "step": 199}, {"loss": 0.45415544509887695, "grad_norm": 0.04769524389503111, "learning_rate": 9.936254857536912e-05, "epoch": 0.07960991143397353, "step": 200}, {"loss": 0.37071096897125244, "grad_norm": 0.03537959735480652, "learning_rate": 9.935224798972273e-05, "epoch": 0.08000796099114339, "step": 201}, {"loss": 0.4470628798007965, "grad_norm": 0.04283047409966095, "learning_rate": 9.934186538864629e-05, "epoch": 0.08040601054831327, "step": 202}, {"loss": 0.4277156591415405, "grad_norm": 0.03941538699830154, "learning_rate": 9.933140078939404e-05, "epoch": 0.08080406010548313, "step": 203}, {"loss": 0.42674386501312256, "grad_norm": 0.04517798019722471, "learning_rate": 9.93208542093564e-05, "epoch": 0.081202109662653, "step": 204}, {"loss": 0.45254310965538025, "grad_norm": 0.04286422625087783, "learning_rate": 9.93102256660601e-05, "epoch": 0.08160015921982287, "step": 205}, {"loss": 0.4146169424057007, "grad_norm": 0.03787321350987435, "learning_rate": 9.929951517716804e-05, "epoch": 0.08199820877699274, "step": 206}, {"loss": 0.4352809190750122, "grad_norm": 0.043024591284465674, "learning_rate": 9.928872276047933e-05, "epoch": 0.0823962583341626, "step": 207}, {"loss": 0.40190589427948, "grad_norm": 0.038769395090287526, "learning_rate": 9.927784843392921e-05, "epoch": 0.08279430789133248, "step": 208}, {"loss": 0.44457823038101196, "grad_norm": 0.04057871433177929, "learning_rate": 9.926689221558904e-05, "epoch": 0.08319235744850234, "step": 209}, {"loss": 0.4133947789669037, "grad_norm": 0.040516515285825117, "learning_rate": 9.92558541236663e-05, "epoch": 0.0835904070056722, "step": 210}, {"loss": 0.42213231325149536, "grad_norm": 0.040946595303433087, "learning_rate": 9.924473417650446e-05, "epoch": 0.08398845656284207, "step": 211}, {"loss": 0.3774731755256653, "grad_norm": 0.04093782719766919, "learning_rate": 9.923353239258313e-05, "epoch": 0.08438650612001194, "step": 212}, {"loss": 0.41167977452278137, "grad_norm": 0.03921886012858554, "learning_rate": 9.922224879051782e-05, "epoch": 0.08478455567718181, "step": 213}, {"loss": 0.4256435036659241, "grad_norm": 0.03942346535310096, "learning_rate": 9.921088338906004e-05, "epoch": 0.08518260523435167, "step": 214}, {"loss": 0.44722259044647217, "grad_norm": 0.042764150496616575, "learning_rate": 9.919943620709727e-05, "epoch": 0.08558065479152155, "step": 215}, {"loss": 0.4617920517921448, "grad_norm": 0.042066071297651435, "learning_rate": 9.918790726365286e-05, "epoch": 0.08597870434869141, "step": 216}, {"loss": 0.3586368262767792, "grad_norm": 0.03802702969905908, "learning_rate": 9.917629657788603e-05, "epoch": 0.08637675390586128, "step": 217}, {"loss": 0.4115375876426697, "grad_norm": 0.04075949524612807, "learning_rate": 9.916460416909187e-05, "epoch": 0.08677480346303115, "step": 218}, {"loss": 0.4272756576538086, "grad_norm": 0.04357170066439433, "learning_rate": 9.915283005670126e-05, "epoch": 0.08717285302020102, "step": 219}, {"loss": 0.409174382686615, "grad_norm": 0.0418277261240607, "learning_rate": 9.914097426028088e-05, "epoch": 0.08757090257737088, "step": 220}, {"loss": 0.4309459328651428, "grad_norm": 0.04157737585145332, "learning_rate": 9.912903679953311e-05, "epoch": 0.08796895213454076, "step": 221}, {"loss": 0.37832939624786377, "grad_norm": 0.036940816837784336, "learning_rate": 9.911701769429611e-05, "epoch": 0.08836700169171062, "step": 222}, {"loss": 0.43967968225479126, "grad_norm": 0.04123129756730642, "learning_rate": 9.910491696454365e-05, "epoch": 0.08876505124888048, "step": 223}, {"loss": 0.3826737403869629, "grad_norm": 0.037793991202797785, "learning_rate": 9.90927346303852e-05, "epoch": 0.08916310080605035, "step": 224}, {"loss": 0.4298074543476105, "grad_norm": 0.041807253187454776, "learning_rate": 9.908047071206582e-05, "epoch": 0.08956115036322022, "step": 225}, {"loss": 0.4168733060359955, "grad_norm": 0.04156560399007536, "learning_rate": 9.906812522996615e-05, "epoch": 0.08995919992039009, "step": 226}, {"loss": 0.39908093214035034, "grad_norm": 0.040484345072216765, "learning_rate": 9.905569820460238e-05, "epoch": 0.09035724947755995, "step": 227}, {"loss": 0.4186454713344574, "grad_norm": 0.03815248241769041, "learning_rate": 9.90431896566262e-05, "epoch": 0.09075529903472983, "step": 228}, {"loss": 0.3910391926765442, "grad_norm": 0.03708797829611071, "learning_rate": 9.903059960682481e-05, "epoch": 0.09115334859189969, "step": 229}, {"loss": 0.4044170379638672, "grad_norm": 0.038850754125269886, "learning_rate": 9.90179280761208e-05, "epoch": 0.09155139814906955, "step": 230}, {"loss": 0.40901124477386475, "grad_norm": 0.044013329243008506, "learning_rate": 9.900517508557223e-05, "epoch": 0.09194944770623943, "step": 231}, {"loss": 0.38673049211502075, "grad_norm": 0.03772669270565151, "learning_rate": 9.899234065637247e-05, "epoch": 0.0923474972634093, "step": 232}, {"loss": 0.4128769636154175, "grad_norm": 0.039903278673923626, "learning_rate": 9.897942480985029e-05, "epoch": 0.09274554682057916, "step": 233}, {"loss": 0.45152929425239563, "grad_norm": 0.04096039762963063, "learning_rate": 9.896642756746973e-05, "epoch": 0.09314359637774904, "step": 234}, {"loss": 0.4287716746330261, "grad_norm": 0.03932694578707697, "learning_rate": 9.895334895083006e-05, "epoch": 0.0935416459349189, "step": 235}, {"loss": 0.43262046575546265, "grad_norm": 0.039807658905750956, "learning_rate": 9.894018898166584e-05, "epoch": 0.09393969549208876, "step": 236}, {"loss": 0.44375234842300415, "grad_norm": 0.03895334345530885, "learning_rate": 9.892694768184684e-05, "epoch": 0.09433774504925864, "step": 237}, {"loss": 0.39813339710235596, "grad_norm": 0.04013608878746097, "learning_rate": 9.89136250733779e-05, "epoch": 0.0947357946064285, "step": 238}, {"loss": 0.41826319694519043, "grad_norm": 0.039493013419155794, "learning_rate": 9.890022117839906e-05, "epoch": 0.09513384416359837, "step": 239}, {"loss": 0.39613017439842224, "grad_norm": 0.03941763232787639, "learning_rate": 9.88867360191854e-05, "epoch": 0.09553189372076823, "step": 240}, {"loss": 0.36194276809692383, "grad_norm": 0.04054167145224511, "learning_rate": 9.887316961814712e-05, "epoch": 0.0959299432779381, "step": 241}, {"loss": 0.4562069773674011, "grad_norm": 0.04391406750780408, "learning_rate": 9.88595219978293e-05, "epoch": 0.09632799283510797, "step": 242}, {"loss": 0.31752869486808777, "grad_norm": 0.03593024213352482, "learning_rate": 9.884579318091214e-05, "epoch": 0.09672604239227783, "step": 243}, {"loss": 0.3816753029823303, "grad_norm": 0.04358820082270423, "learning_rate": 9.883198319021066e-05, "epoch": 0.09712409194944771, "step": 244}, {"loss": 0.4183397889137268, "grad_norm": 0.037886016287282186, "learning_rate": 9.881809204867485e-05, "epoch": 0.09752214150661757, "step": 245}, {"loss": 0.46071240305900574, "grad_norm": 0.04262907120069682, "learning_rate": 9.880411977938952e-05, "epoch": 0.09792019106378744, "step": 246}, {"loss": 0.3988252282142639, "grad_norm": 0.04106831285459527, "learning_rate": 9.879006640557431e-05, "epoch": 0.09831824062095731, "step": 247}, {"loss": 0.3727218508720398, "grad_norm": 0.038945181583259623, "learning_rate": 9.877593195058367e-05, "epoch": 0.09871629017812718, "step": 248}, {"loss": 0.4156814515590668, "grad_norm": 0.03973811511653363, "learning_rate": 9.876171643790675e-05, "epoch": 0.09911433973529704, "step": 249}, {"loss": 0.43539130687713623, "grad_norm": 0.041129423140107375, "learning_rate": 9.874741989116743e-05, "epoch": 0.09951238929246692, "step": 250}, {"loss": 0.40725281834602356, "grad_norm": 0.0353112465184484, "learning_rate": 9.873304233412426e-05, "epoch": 0.09991043884963678, "step": 251}, {"loss": 0.39336836338043213, "grad_norm": 0.0383672335519336, "learning_rate": 9.871858379067041e-05, "epoch": 0.10030848840680665, "step": 252}, {"loss": 0.4119214713573456, "grad_norm": 0.03955725392291635, "learning_rate": 9.870404428483362e-05, "epoch": 0.10070653796397651, "step": 253}, {"loss": 0.46002480387687683, "grad_norm": 0.04591149127007547, "learning_rate": 9.868942384077621e-05, "epoch": 0.10110458752114639, "step": 254}, {"loss": 0.41638830304145813, "grad_norm": 0.040469233708873766, "learning_rate": 9.867472248279497e-05, "epoch": 0.10150263707831625, "step": 255}, {"loss": 0.3776403069496155, "grad_norm": 0.04061568349581623, "learning_rate": 9.865994023532115e-05, "epoch": 0.10190068663548611, "step": 256}, {"loss": 0.4319694936275482, "grad_norm": 0.0402554933121857, "learning_rate": 9.86450771229205e-05, "epoch": 0.10229873619265599, "step": 257}, {"loss": 0.4386405050754547, "grad_norm": 0.041194495178644285, "learning_rate": 9.863013317029305e-05, "epoch": 0.10269678574982585, "step": 258}, {"loss": 0.3834368586540222, "grad_norm": 0.038243530305035724, "learning_rate": 9.861510840227329e-05, "epoch": 0.10309483530699572, "step": 259}, {"loss": 0.4216727912425995, "grad_norm": 0.04346565265191586, "learning_rate": 9.860000284382987e-05, "epoch": 0.1034928848641656, "step": 260}, {"loss": 0.3837319314479828, "grad_norm": 0.043625368366783226, "learning_rate": 9.858481652006581e-05, "epoch": 0.10389093442133546, "step": 261}, {"loss": 0.4109954833984375, "grad_norm": 0.04128422556771251, "learning_rate": 9.856954945621834e-05, "epoch": 0.10428898397850532, "step": 262}, {"loss": 0.4031594395637512, "grad_norm": 0.04578854236585009, "learning_rate": 9.855420167765882e-05, "epoch": 0.1046870335356752, "step": 263}, {"loss": 0.3804945945739746, "grad_norm": 0.038882869143341374, "learning_rate": 9.853877320989277e-05, "epoch": 0.10508508309284506, "step": 264}, {"loss": 0.3643825650215149, "grad_norm": 0.03827493667206872, "learning_rate": 9.85232640785598e-05, "epoch": 0.10548313265001492, "step": 265}, {"loss": 0.45318666100502014, "grad_norm": 0.04026961548319946, "learning_rate": 9.850767430943358e-05, "epoch": 0.1058811822071848, "step": 266}, {"loss": 0.4306051433086395, "grad_norm": 0.03975171046452114, "learning_rate": 9.849200392842176e-05, "epoch": 0.10627923176435466, "step": 267}, {"loss": 0.3905956745147705, "grad_norm": 0.04206689074129096, "learning_rate": 9.847625296156598e-05, "epoch": 0.10667728132152453, "step": 268}, {"loss": 0.42062604427337646, "grad_norm": 0.04100228487417135, "learning_rate": 9.84604214350418e-05, "epoch": 0.10707533087869439, "step": 269}, {"loss": 0.42568033933639526, "grad_norm": 0.04123776055851874, "learning_rate": 9.844450937515864e-05, "epoch": 0.10747338043586427, "step": 270}, {"loss": 0.417341947555542, "grad_norm": 0.04407816851008842, "learning_rate": 9.842851680835977e-05, "epoch": 0.10787142999303413, "step": 271}, {"loss": 0.3689631223678589, "grad_norm": 0.039102899671514896, "learning_rate": 9.841244376122223e-05, "epoch": 0.108269479550204, "step": 272}, {"loss": 0.4298265278339386, "grad_norm": 0.042473998089085356, "learning_rate": 9.839629026045682e-05, "epoch": 0.10866752910737387, "step": 273}, {"loss": 0.4388338029384613, "grad_norm": 0.041231647173338916, "learning_rate": 9.838005633290804e-05, "epoch": 0.10906557866454374, "step": 274}, {"loss": 0.3878886103630066, "grad_norm": 0.04026946925424408, "learning_rate": 9.836374200555405e-05, "epoch": 0.1094636282217136, "step": 275}, {"loss": 0.3590395450592041, "grad_norm": 0.03945482847444275, "learning_rate": 9.83473473055066e-05, "epoch": 0.10986167777888348, "step": 276}, {"loss": 0.39957648515701294, "grad_norm": 0.04015244239641414, "learning_rate": 9.833087226001105e-05, "epoch": 0.11025972733605334, "step": 277}, {"loss": 0.40955787897109985, "grad_norm": 0.03867542391955392, "learning_rate": 9.831431689644621e-05, "epoch": 0.1106577768932232, "step": 278}, {"loss": 0.43313831090927124, "grad_norm": 0.03991583005736097, "learning_rate": 9.829768124232446e-05, "epoch": 0.11105582645039308, "step": 279}, {"loss": 0.45130181312561035, "grad_norm": 0.039325673020272636, "learning_rate": 9.828096532529152e-05, "epoch": 0.11145387600756294, "step": 280}, {"loss": 0.39202070236206055, "grad_norm": 0.03882094909157594, "learning_rate": 9.826416917312654e-05, "epoch": 0.11185192556473281, "step": 281}, {"loss": 0.3969845175743103, "grad_norm": 0.03905571280000889, "learning_rate": 9.824729281374201e-05, "epoch": 0.11224997512190267, "step": 282}, {"loss": 0.399698406457901, "grad_norm": 0.03772587544238222, "learning_rate": 9.823033627518367e-05, "epoch": 0.11264802467907255, "step": 283}, {"loss": 0.4441763162612915, "grad_norm": 0.04025220560919426, "learning_rate": 9.82132995856306e-05, "epoch": 0.11304607423624241, "step": 284}, {"loss": 0.37657856941223145, "grad_norm": 0.03929229479556206, "learning_rate": 9.819618277339495e-05, "epoch": 0.11344412379341227, "step": 285}, {"loss": 0.3968500792980194, "grad_norm": 0.037946520704934816, "learning_rate": 9.817898586692214e-05, "epoch": 0.11384217335058215, "step": 286}, {"loss": 0.45299655199050903, "grad_norm": 0.040212948758312855, "learning_rate": 9.816170889479058e-05, "epoch": 0.11424022290775201, "step": 287}, {"loss": 0.4282580614089966, "grad_norm": 0.04417430189424275, "learning_rate": 9.814435188571184e-05, "epoch": 0.11463827246492188, "step": 288}, {"loss": 0.4048348069190979, "grad_norm": 0.04330594326634229, "learning_rate": 9.812691486853042e-05, "epoch": 0.11503632202209176, "step": 289}, {"loss": 0.42288607358932495, "grad_norm": 0.04215585256375776, "learning_rate": 9.810939787222386e-05, "epoch": 0.11543437157926162, "step": 290}, {"loss": 0.34067124128341675, "grad_norm": 0.04109865452071034, "learning_rate": 9.809180092590254e-05, "epoch": 0.11583242113643148, "step": 291}, {"loss": 0.47663480043411255, "grad_norm": 0.04569075858780607, "learning_rate": 9.807412405880972e-05, "epoch": 0.11623047069360136, "step": 292}, {"loss": 0.41257303953170776, "grad_norm": 0.04060029180975312, "learning_rate": 9.805636730032149e-05, "epoch": 0.11662852025077122, "step": 293}, {"loss": 0.4323192238807678, "grad_norm": 0.038257369684456796, "learning_rate": 9.80385306799467e-05, "epoch": 0.11702656980794109, "step": 294}, {"loss": 0.41773083806037903, "grad_norm": 0.04086457467219757, "learning_rate": 9.802061422732692e-05, "epoch": 0.11742461936511096, "step": 295}, {"loss": 0.38483744859695435, "grad_norm": 0.0385660200032891, "learning_rate": 9.800261797223641e-05, "epoch": 0.11782266892228083, "step": 296}, {"loss": 0.4123501777648926, "grad_norm": 0.04212107942383715, "learning_rate": 9.7984541944582e-05, "epoch": 0.11822071847945069, "step": 297}, {"loss": 0.4812725782394409, "grad_norm": 0.045805154636272015, "learning_rate": 9.796638617440314e-05, "epoch": 0.11861876803662055, "step": 298}, {"loss": 0.4139779210090637, "grad_norm": 0.04175884237604085, "learning_rate": 9.794815069187176e-05, "epoch": 0.11901681759379043, "step": 299}, {"loss": 0.38807913661003113, "grad_norm": 0.03863097481139229, "learning_rate": 9.792983552729225e-05, "epoch": 0.1194148671509603, "step": 300}, {"loss": 0.4293481707572937, "grad_norm": 0.041471579992865305, "learning_rate": 9.791144071110149e-05, "epoch": 0.11981291670813016, "step": 301}, {"loss": 0.4229589104652405, "grad_norm": 0.0425802989750573, "learning_rate": 9.789296627386865e-05, "epoch": 0.12021096626530003, "step": 302}, {"loss": 0.41160041093826294, "grad_norm": 0.03980078857552689, "learning_rate": 9.787441224629525e-05, "epoch": 0.1206090158224699, "step": 303}, {"loss": 0.40149420499801636, "grad_norm": 0.0390702163909971, "learning_rate": 9.785577865921509e-05, "epoch": 0.12100706537963976, "step": 304}, {"loss": 0.4120422303676605, "grad_norm": 0.04301612532775671, "learning_rate": 9.783706554359416e-05, "epoch": 0.12140511493680964, "step": 305}, {"loss": 0.38755208253860474, "grad_norm": 0.03971665983870957, "learning_rate": 9.78182729305306e-05, "epoch": 0.1218031644939795, "step": 306}, {"loss": 0.4232228398323059, "grad_norm": 0.039044213603638606, "learning_rate": 9.779940085125471e-05, "epoch": 0.12220121405114936, "step": 307}, {"loss": 0.3719865679740906, "grad_norm": 0.03534118521844285, "learning_rate": 9.778044933712883e-05, "epoch": 0.12259926360831924, "step": 308}, {"loss": 0.3904886245727539, "grad_norm": 0.036881011086407275, "learning_rate": 9.776141841964728e-05, "epoch": 0.1229973131654891, "step": 309}, {"loss": 0.40225750207901, "grad_norm": 0.04171476685037898, "learning_rate": 9.774230813043638e-05, "epoch": 0.12339536272265897, "step": 310}, {"loss": 0.38312411308288574, "grad_norm": 0.03963149090711351, "learning_rate": 9.772311850125431e-05, "epoch": 0.12379341227982883, "step": 311}, {"loss": 0.39071497321128845, "grad_norm": 0.04415069036690034, "learning_rate": 9.770384956399112e-05, "epoch": 0.12419146183699871, "step": 312}, {"loss": 0.4006495475769043, "grad_norm": 0.03856094622664213, "learning_rate": 9.768450135066868e-05, "epoch": 0.12458951139416857, "step": 313}, {"loss": 0.3938033878803253, "grad_norm": 0.041307901892455486, "learning_rate": 9.766507389344057e-05, "epoch": 0.12498756095133844, "step": 314}, {"loss": 0.43045392632484436, "grad_norm": 0.0421087816134629, "learning_rate": 9.764556722459206e-05, "epoch": 0.1253856105085083, "step": 315}, {"loss": 0.40216904878616333, "grad_norm": 0.03882879596371879, "learning_rate": 9.762598137654009e-05, "epoch": 0.12578366006567818, "step": 316}, {"loss": 0.43507838249206543, "grad_norm": 0.043540269520803115, "learning_rate": 9.760631638183315e-05, "epoch": 0.12618170962284805, "step": 317}, {"loss": 0.37054333090782166, "grad_norm": 0.04167061256737787, "learning_rate": 9.75865722731513e-05, "epoch": 0.1265797591800179, "step": 318}, {"loss": 0.3966912627220154, "grad_norm": 0.04182671300852681, "learning_rate": 9.756674908330598e-05, "epoch": 0.12697780873718778, "step": 319}, {"loss": 0.3955579996109009, "grad_norm": 0.04321408250753292, "learning_rate": 9.754684684524016e-05, "epoch": 0.12737585829435766, "step": 320}, {"loss": 0.3710075616836548, "grad_norm": 0.04009045271960528, "learning_rate": 9.752686559202815e-05, "epoch": 0.1277739078515275, "step": 321}, {"loss": 0.45274442434310913, "grad_norm": 0.04295680210765361, "learning_rate": 9.75068053568755e-05, "epoch": 0.12817195740869738, "step": 322}, {"loss": 0.3914702236652374, "grad_norm": 0.03819731038521213, "learning_rate": 9.748666617311912e-05, "epoch": 0.12857000696586726, "step": 323}, {"loss": 0.41818931698799133, "grad_norm": 0.039872844922919454, "learning_rate": 9.746644807422702e-05, "epoch": 0.1289680565230371, "step": 324}, {"loss": 0.40861034393310547, "grad_norm": 0.03977090729630188, "learning_rate": 9.744615109379843e-05, "epoch": 0.129366106080207, "step": 325}, {"loss": 0.4417041838169098, "grad_norm": 0.042405932241218726, "learning_rate": 9.742577526556363e-05, "epoch": 0.12976415563737687, "step": 326}, {"loss": 0.4612831473350525, "grad_norm": 0.0422473634193453, "learning_rate": 9.740532062338394e-05, "epoch": 0.13016220519454672, "step": 327}, {"loss": 0.4123196601867676, "grad_norm": 0.037515949621509746, "learning_rate": 9.738478720125165e-05, "epoch": 0.1305602547517166, "step": 328}, {"loss": 0.4192107319831848, "grad_norm": 0.04015876089002835, "learning_rate": 9.736417503329001e-05, "epoch": 0.13095830430888647, "step": 329}, {"loss": 0.41410189867019653, "grad_norm": 0.039380419815895296, "learning_rate": 9.734348415375305e-05, "epoch": 0.13135635386605632, "step": 330}, {"loss": 0.44357675313949585, "grad_norm": 0.04300261640370017, "learning_rate": 9.732271459702569e-05, "epoch": 0.1317544034232262, "step": 331}, {"loss": 0.476681649684906, "grad_norm": 0.04216665192333357, "learning_rate": 9.730186639762353e-05, "epoch": 0.13215245298039605, "step": 332}, {"loss": 0.3941146731376648, "grad_norm": 0.038120575322863566, "learning_rate": 9.728093959019294e-05, "epoch": 0.13255050253756592, "step": 333}, {"loss": 0.46367087960243225, "grad_norm": 0.041172237195620916, "learning_rate": 9.725993420951086e-05, "epoch": 0.1329485520947358, "step": 334}, {"loss": 0.3861929774284363, "grad_norm": 0.036641816260330765, "learning_rate": 9.723885029048481e-05, "epoch": 0.13334660165190565, "step": 335}, {"loss": 0.34580254554748535, "grad_norm": 0.03416268550160123, "learning_rate": 9.721768786815286e-05, "epoch": 0.13374465120907553, "step": 336}, {"loss": 0.38384395837783813, "grad_norm": 0.03885616482843378, "learning_rate": 9.719644697768353e-05, "epoch": 0.1341427007662454, "step": 337}, {"loss": 0.4134787917137146, "grad_norm": 0.04123837054568253, "learning_rate": 9.717512765437571e-05, "epoch": 0.13454075032341525, "step": 338}, {"loss": 0.41235801577568054, "grad_norm": 0.03998744942370595, "learning_rate": 9.715372993365868e-05, "epoch": 0.13493879988058513, "step": 339}, {"loss": 0.36377185583114624, "grad_norm": 0.037291735304976446, "learning_rate": 9.713225385109197e-05, "epoch": 0.135336849437755, "step": 340}, {"loss": 0.39249616861343384, "grad_norm": 0.03820098196300802, "learning_rate": 9.711069944236532e-05, "epoch": 0.13573489899492486, "step": 341}, {"loss": 0.37631404399871826, "grad_norm": 0.040645715957133524, "learning_rate": 9.708906674329872e-05, "epoch": 0.13613294855209473, "step": 342}, {"loss": 0.40443962812423706, "grad_norm": 0.03852529153341787, "learning_rate": 9.706735578984215e-05, "epoch": 0.1365309981092646, "step": 343}, {"loss": 0.38541120290756226, "grad_norm": 0.03866160040120267, "learning_rate": 9.704556661807573e-05, "epoch": 0.13692904766643446, "step": 344}, {"loss": 0.42870551347732544, "grad_norm": 0.044111741288409584, "learning_rate": 9.70236992642095e-05, "epoch": 0.13732709722360434, "step": 345}, {"loss": 0.40259701013565063, "grad_norm": 0.037378218197801204, "learning_rate": 9.700175376458348e-05, "epoch": 0.13772514678077422, "step": 346}, {"loss": 0.3816920518875122, "grad_norm": 0.03968716382892475, "learning_rate": 9.697973015566752e-05, "epoch": 0.13812319633794407, "step": 347}, {"loss": 0.41498610377311707, "grad_norm": 0.04068502075701426, "learning_rate": 9.695762847406127e-05, "epoch": 0.13852124589511394, "step": 348}, {"loss": 0.44331663846969604, "grad_norm": 0.04177054981155606, "learning_rate": 9.693544875649416e-05, "epoch": 0.13891929545228382, "step": 349}, {"loss": 0.4100690484046936, "grad_norm": 0.039670726584035726, "learning_rate": 9.691319103982527e-05, "epoch": 0.13931734500945367, "step": 350}, {"loss": 0.4097104072570801, "grad_norm": 0.040591214856803705, "learning_rate": 9.689085536104334e-05, "epoch": 0.13971539456662355, "step": 351}, {"loss": 0.3780149221420288, "grad_norm": 0.04236502993821542, "learning_rate": 9.686844175726661e-05, "epoch": 0.14011344412379342, "step": 352}, {"loss": 0.3946231007575989, "grad_norm": 0.04267126849568907, "learning_rate": 9.684595026574288e-05, "epoch": 0.14051149368096327, "step": 353}, {"loss": 0.4196978211402893, "grad_norm": 0.04377444997217282, "learning_rate": 9.682338092384935e-05, "epoch": 0.14090954323813315, "step": 354}, {"loss": 0.4052705764770508, "grad_norm": 0.03753189343265348, "learning_rate": 9.68007337690926e-05, "epoch": 0.14130759279530303, "step": 355}, {"loss": 0.3883543908596039, "grad_norm": 0.040486909723284774, "learning_rate": 9.677800883910852e-05, "epoch": 0.14170564235247288, "step": 356}, {"loss": 0.4206618070602417, "grad_norm": 0.04416409056796048, "learning_rate": 9.675520617166227e-05, "epoch": 0.14210369190964275, "step": 357}, {"loss": 0.39708107709884644, "grad_norm": 0.04227871991311324, "learning_rate": 9.673232580464819e-05, "epoch": 0.14250174146681263, "step": 358}, {"loss": 0.4169766902923584, "grad_norm": 0.04237761189325662, "learning_rate": 9.670936777608975e-05, "epoch": 0.14289979102398248, "step": 359}, {"loss": 0.40371939539909363, "grad_norm": 0.046842266772483966, "learning_rate": 9.668633212413942e-05, "epoch": 0.14329784058115236, "step": 360}, {"loss": 0.40961647033691406, "grad_norm": 0.042084178761391984, "learning_rate": 9.666321888707874e-05, "epoch": 0.1436958901383222, "step": 361}, {"loss": 0.35609903931617737, "grad_norm": 0.038256220806303434, "learning_rate": 9.66400281033182e-05, "epoch": 0.14409393969549208, "step": 362}, {"loss": 0.4496447741985321, "grad_norm": 0.039132329036963474, "learning_rate": 9.661675981139707e-05, "epoch": 0.14449198925266196, "step": 363}, {"loss": 0.4229081869125366, "grad_norm": 0.04063351448491965, "learning_rate": 9.659341404998348e-05, "epoch": 0.1448900388098318, "step": 364}, {"loss": 0.3519589602947235, "grad_norm": 0.03388202213582818, "learning_rate": 9.656999085787434e-05, "epoch": 0.1452880883670017, "step": 365}, {"loss": 0.42910340428352356, "grad_norm": 0.04369249450479429, "learning_rate": 9.654649027399516e-05, "epoch": 0.14568613792417157, "step": 366}, {"loss": 0.39379632472991943, "grad_norm": 0.04124169323739096, "learning_rate": 9.65229123374001e-05, "epoch": 0.14608418748134142, "step": 367}, {"loss": 0.37762829661369324, "grad_norm": 0.03945122363681856, "learning_rate": 9.649925708727185e-05, "epoch": 0.1464822370385113, "step": 368}, {"loss": 0.4661373198032379, "grad_norm": 0.04100882883819229, "learning_rate": 9.647552456292165e-05, "epoch": 0.14688028659568117, "step": 369}, {"loss": 0.4408138692378998, "grad_norm": 0.0401529074845743, "learning_rate": 9.645171480378908e-05, "epoch": 0.14727833615285102, "step": 370}, {"loss": 0.35208356380462646, "grad_norm": 0.0385129214093715, "learning_rate": 9.642782784944206e-05, "epoch": 0.1476763857100209, "step": 371}, {"loss": 0.3908153176307678, "grad_norm": 0.038984239843856704, "learning_rate": 9.640386373957688e-05, "epoch": 0.14807443526719077, "step": 372}, {"loss": 0.3718884289264679, "grad_norm": 0.03719998790245794, "learning_rate": 9.6379822514018e-05, "epoch": 0.14847248482436062, "step": 373}, {"loss": 0.3803236484527588, "grad_norm": 0.03822905334776594, "learning_rate": 9.635570421271803e-05, "epoch": 0.1488705343815305, "step": 374}, {"loss": 0.4148911237716675, "grad_norm": 0.0402519807192828, "learning_rate": 9.633150887575764e-05, "epoch": 0.14926858393870038, "step": 375}, {"loss": 0.39321520924568176, "grad_norm": 0.04006610986112075, "learning_rate": 9.63072365433456e-05, "epoch": 0.14966663349587023, "step": 376}, {"loss": 0.43051639199256897, "grad_norm": 0.04543954377113224, "learning_rate": 9.628288725581858e-05, "epoch": 0.1500646830530401, "step": 377}, {"loss": 0.379738450050354, "grad_norm": 0.0412850144779994, "learning_rate": 9.625846105364113e-05, "epoch": 0.15046273261020998, "step": 378}, {"loss": 0.401308536529541, "grad_norm": 0.04033879590785085, "learning_rate": 9.623395797740564e-05, "epoch": 0.15086078216737983, "step": 379}, {"loss": 0.4514630436897278, "grad_norm": 0.041444622988472665, "learning_rate": 9.620937806783226e-05, "epoch": 0.1512588317245497, "step": 380}, {"loss": 0.43836653232574463, "grad_norm": 0.04119305802746497, "learning_rate": 9.618472136576882e-05, "epoch": 0.15165688128171959, "step": 381}, {"loss": 0.421159565448761, "grad_norm": 0.039949479196317195, "learning_rate": 9.615998791219072e-05, "epoch": 0.15205493083888943, "step": 382}, {"loss": 0.38707900047302246, "grad_norm": 0.03877375344249926, "learning_rate": 9.613517774820098e-05, "epoch": 0.1524529803960593, "step": 383}, {"loss": 0.41162484884262085, "grad_norm": 0.040179942774094604, "learning_rate": 9.611029091503005e-05, "epoch": 0.1528510299532292, "step": 384}, {"loss": 0.4128221571445465, "grad_norm": 0.037974730383546686, "learning_rate": 9.608532745403584e-05, "epoch": 0.15324907951039904, "step": 385}, {"loss": 0.42371782660484314, "grad_norm": 0.038721109278015586, "learning_rate": 9.606028740670354e-05, "epoch": 0.15364712906756892, "step": 386}, {"loss": 0.36211633682250977, "grad_norm": 0.04155729676353007, "learning_rate": 9.603517081464564e-05, "epoch": 0.1540451786247388, "step": 387}, {"loss": 0.40073859691619873, "grad_norm": 0.04230561934962309, "learning_rate": 9.600997771960188e-05, "epoch": 0.15444322818190864, "step": 388}, {"loss": 0.4036528468132019, "grad_norm": 0.037581152835085854, "learning_rate": 9.598470816343907e-05, "epoch": 0.15484127773907852, "step": 389}, {"loss": 0.4410549998283386, "grad_norm": 0.03985709339024297, "learning_rate": 9.595936218815111e-05, "epoch": 0.15523932729624837, "step": 390}, {"loss": 0.42898738384246826, "grad_norm": 0.039569492824587084, "learning_rate": 9.593393983585892e-05, "epoch": 0.15563737685341825, "step": 391}, {"loss": 0.3881680369377136, "grad_norm": 0.03945185925278315, "learning_rate": 9.590844114881032e-05, "epoch": 0.15603542641058812, "step": 392}, {"loss": 0.4207717776298523, "grad_norm": 0.03974059953772571, "learning_rate": 9.588286616937996e-05, "epoch": 0.15643347596775797, "step": 393}, {"loss": 0.4462347626686096, "grad_norm": 0.041487114863017364, "learning_rate": 9.585721494006935e-05, "epoch": 0.15683152552492785, "step": 394}, {"loss": 0.38908857107162476, "grad_norm": 0.03649422018635143, "learning_rate": 9.583148750350665e-05, "epoch": 0.15722957508209773, "step": 395}, {"loss": 0.40280598402023315, "grad_norm": 0.040601025054328894, "learning_rate": 9.580568390244669e-05, "epoch": 0.15762762463926758, "step": 396}, {"loss": 0.3584386706352234, "grad_norm": 0.03932205497021046, "learning_rate": 9.577980417977086e-05, "epoch": 0.15802567419643745, "step": 397}, {"loss": 0.4044828712940216, "grad_norm": 0.0444593996320978, "learning_rate": 9.57538483784871e-05, "epoch": 0.15842372375360733, "step": 398}, {"loss": 0.39401859045028687, "grad_norm": 0.03553968563065413, "learning_rate": 9.572781654172967e-05, "epoch": 0.15882177331077718, "step": 399}, {"loss": 0.4427478611469269, "grad_norm": 0.04111161368299127, "learning_rate": 9.570170871275932e-05, "epoch": 0.15921982286794706, "step": 400}, {"loss": 0.3624044954776764, "grad_norm": 0.03763436448930962, "learning_rate": 9.5675524934963e-05, "epoch": 0.15961787242511694, "step": 401}, {"loss": 0.4205958843231201, "grad_norm": 0.039737781472508484, "learning_rate": 9.564926525185392e-05, "epoch": 0.16001592198228679, "step": 402}, {"loss": 0.38174837827682495, "grad_norm": 0.03707740473538291, "learning_rate": 9.562292970707138e-05, "epoch": 0.16041397153945666, "step": 403}, {"loss": 0.4320589005947113, "grad_norm": 0.04118223341183807, "learning_rate": 9.559651834438083e-05, "epoch": 0.16081202109662654, "step": 404}, {"loss": 0.376958966255188, "grad_norm": 0.038127099554173566, "learning_rate": 9.557003120767363e-05, "epoch": 0.1612100706537964, "step": 405}, {"loss": 0.35138699412345886, "grad_norm": 0.036855576173427995, "learning_rate": 9.554346834096713e-05, "epoch": 0.16160812021096627, "step": 406}, {"loss": 0.3933624029159546, "grad_norm": 0.037536221122283345, "learning_rate": 9.551682978840448e-05, "epoch": 0.16200616976813614, "step": 407}, {"loss": 0.3943382501602173, "grad_norm": 0.03766763722030671, "learning_rate": 9.549011559425465e-05, "epoch": 0.162404219325306, "step": 408}, {"loss": 0.3700028657913208, "grad_norm": 0.03464119955075374, "learning_rate": 9.546332580291228e-05, "epoch": 0.16280226888247587, "step": 409}, {"loss": 0.42637258768081665, "grad_norm": 0.04077097493788642, "learning_rate": 9.54364604588977e-05, "epoch": 0.16320031843964575, "step": 410}, {"loss": 0.3550247550010681, "grad_norm": 0.039915044062689496, "learning_rate": 9.54095196068567e-05, "epoch": 0.1635983679968156, "step": 411}, {"loss": 0.40059107542037964, "grad_norm": 0.036616458418166624, "learning_rate": 9.538250329156064e-05, "epoch": 0.16399641755398547, "step": 412}, {"loss": 0.4029604196548462, "grad_norm": 0.04154465217908537, "learning_rate": 9.535541155790626e-05, "epoch": 0.16439446711115535, "step": 413}, {"loss": 0.4256274104118347, "grad_norm": 0.03859768969705245, "learning_rate": 9.532824445091558e-05, "epoch": 0.1647925166683252, "step": 414}, {"loss": 0.37159794569015503, "grad_norm": 0.04059910934984861, "learning_rate": 9.530100201573597e-05, "epoch": 0.16519056622549508, "step": 415}, {"loss": 0.39319607615470886, "grad_norm": 0.038267806928006666, "learning_rate": 9.527368429763994e-05, "epoch": 0.16558861578266496, "step": 416}, {"loss": 0.38508331775665283, "grad_norm": 0.03537512817654085, "learning_rate": 9.524629134202507e-05, "epoch": 0.1659866653398348, "step": 417}, {"loss": 0.4496425688266754, "grad_norm": 0.04100260634020172, "learning_rate": 9.521882319441402e-05, "epoch": 0.16638471489700468, "step": 418}, {"loss": 0.3994331955909729, "grad_norm": 0.03698952260193856, "learning_rate": 9.519127990045441e-05, "epoch": 0.16678276445417453, "step": 419}, {"loss": 0.40674349665641785, "grad_norm": 0.03828394352184475, "learning_rate": 9.516366150591871e-05, "epoch": 0.1671808140113444, "step": 420}, {"loss": 0.4244731664657593, "grad_norm": 0.042023332856833816, "learning_rate": 9.513596805670425e-05, "epoch": 0.16757886356851429, "step": 421}, {"loss": 0.3691408634185791, "grad_norm": 0.04032470672427232, "learning_rate": 9.5108199598833e-05, "epoch": 0.16797691312568414, "step": 422}, {"loss": 0.35767966508865356, "grad_norm": 0.036828998887074846, "learning_rate": 9.508035617845163e-05, "epoch": 0.168374962682854, "step": 423}, {"loss": 0.39381128549575806, "grad_norm": 0.03793851110774677, "learning_rate": 9.505243784183145e-05, "epoch": 0.1687730122400239, "step": 424}, {"loss": 0.42446526885032654, "grad_norm": 0.03970136324366445, "learning_rate": 9.502444463536815e-05, "epoch": 0.16917106179719374, "step": 425}, {"loss": 0.4132591485977173, "grad_norm": 0.042013350387359184, "learning_rate": 9.499637660558191e-05, "epoch": 0.16956911135436362, "step": 426}, {"loss": 0.40549805760383606, "grad_norm": 0.04040713849194192, "learning_rate": 9.496823379911726e-05, "epoch": 0.1699671609115335, "step": 427}, {"loss": 0.39220237731933594, "grad_norm": 0.037594239013687955, "learning_rate": 9.494001626274295e-05, "epoch": 0.17036521046870334, "step": 428}, {"loss": 0.4125314950942993, "grad_norm": 0.03873875940358716, "learning_rate": 9.491172404335197e-05, "epoch": 0.17076326002587322, "step": 429}, {"loss": 0.4112716615200043, "grad_norm": 0.03731770663570797, "learning_rate": 9.488335718796138e-05, "epoch": 0.1711613095830431, "step": 430}, {"loss": 0.40110093355178833, "grad_norm": 0.040933853965417516, "learning_rate": 9.485491574371233e-05, "epoch": 0.17155935914021295, "step": 431}, {"loss": 0.385872483253479, "grad_norm": 0.03581341096096431, "learning_rate": 9.482639975786985e-05, "epoch": 0.17195740869738282, "step": 432}, {"loss": 0.4116959273815155, "grad_norm": 0.0394504800498143, "learning_rate": 9.479780927782289e-05, "epoch": 0.1723554582545527, "step": 433}, {"loss": 0.37987077236175537, "grad_norm": 0.04387197766462701, "learning_rate": 9.47691443510842e-05, "epoch": 0.17275350781172255, "step": 434}, {"loss": 0.3973982036113739, "grad_norm": 0.04172361460068723, "learning_rate": 9.474040502529022e-05, "epoch": 0.17315155736889243, "step": 435}, {"loss": 0.46508586406707764, "grad_norm": 0.03932560134222253, "learning_rate": 9.471159134820107e-05, "epoch": 0.1735496069260623, "step": 436}, {"loss": 0.410735547542572, "grad_norm": 0.04074103650147955, "learning_rate": 9.468270336770037e-05, "epoch": 0.17394765648323215, "step": 437}, {"loss": 0.3987840414047241, "grad_norm": 0.03794933255048685, "learning_rate": 9.465374113179531e-05, "epoch": 0.17434570604040203, "step": 438}, {"loss": 0.39110249280929565, "grad_norm": 0.0390458779198587, "learning_rate": 9.46247046886164e-05, "epoch": 0.1747437555975719, "step": 439}, {"loss": 0.3875223398208618, "grad_norm": 0.03783070629046863, "learning_rate": 9.45955940864175e-05, "epoch": 0.17514180515474176, "step": 440}, {"loss": 0.41689059138298035, "grad_norm": 0.04476051253009858, "learning_rate": 9.456640937357568e-05, "epoch": 0.17553985471191164, "step": 441}, {"loss": 0.393161416053772, "grad_norm": 0.040618461154503537, "learning_rate": 9.453715059859125e-05, "epoch": 0.1759379042690815, "step": 442}, {"loss": 0.4154399633407593, "grad_norm": 0.039260096573492824, "learning_rate": 9.450781781008753e-05, "epoch": 0.17633595382625136, "step": 443}, {"loss": 0.4274646043777466, "grad_norm": 0.04104031005227378, "learning_rate": 9.447841105681084e-05, "epoch": 0.17673400338342124, "step": 444}, {"loss": 0.3844272494316101, "grad_norm": 0.04008734440316746, "learning_rate": 9.444893038763042e-05, "epoch": 0.17713205294059112, "step": 445}, {"loss": 0.41134369373321533, "grad_norm": 0.042089148980233484, "learning_rate": 9.44193758515384e-05, "epoch": 0.17753010249776097, "step": 446}, {"loss": 0.381368488073349, "grad_norm": 0.037879528058534895, "learning_rate": 9.438974749764961e-05, "epoch": 0.17792815205493084, "step": 447}, {"loss": 0.3665882349014282, "grad_norm": 0.03798917528657286, "learning_rate": 9.436004537520157e-05, "epoch": 0.1783262016121007, "step": 448}, {"loss": 0.4369085431098938, "grad_norm": 0.04013140396586054, "learning_rate": 9.433026953355437e-05, "epoch": 0.17872425116927057, "step": 449}, {"loss": 0.4016512632369995, "grad_norm": 0.038874624155480896, "learning_rate": 9.430042002219065e-05, "epoch": 0.17912230072644045, "step": 450}, {"loss": 0.40920132398605347, "grad_norm": 0.040342620159030916, "learning_rate": 9.427049689071545e-05, "epoch": 0.1795203502836103, "step": 451}, {"loss": 0.4179263710975647, "grad_norm": 0.04057186993564771, "learning_rate": 9.424050018885615e-05, "epoch": 0.17991839984078017, "step": 452}, {"loss": 0.42316848039627075, "grad_norm": 0.03979832077824595, "learning_rate": 9.421042996646241e-05, "epoch": 0.18031644939795005, "step": 453}, {"loss": 0.41397714614868164, "grad_norm": 0.038732099241145214, "learning_rate": 9.418028627350606e-05, "epoch": 0.1807144989551199, "step": 454}, {"loss": 0.3918016850948334, "grad_norm": 0.04097088353505276, "learning_rate": 9.415006916008105e-05, "epoch": 0.18111254851228978, "step": 455}, {"loss": 0.43606865406036377, "grad_norm": 0.04152085113549303, "learning_rate": 9.411977867640328e-05, "epoch": 0.18151059806945966, "step": 456}, {"loss": 0.4296274185180664, "grad_norm": 0.039069165889833564, "learning_rate": 9.408941487281064e-05, "epoch": 0.1819086476266295, "step": 457}, {"loss": 0.41636765003204346, "grad_norm": 0.03645085487877534, "learning_rate": 9.405897779976284e-05, "epoch": 0.18230669718379938, "step": 458}, {"loss": 0.4090272784233093, "grad_norm": 0.03743212483789185, "learning_rate": 9.402846750784137e-05, "epoch": 0.18270474674096926, "step": 459}, {"loss": 0.4118373394012451, "grad_norm": 0.04007658771933816, "learning_rate": 9.399788404774938e-05, "epoch": 0.1831027962981391, "step": 460}, {"loss": 0.37161269783973694, "grad_norm": 0.04007125769073327, "learning_rate": 9.396722747031163e-05, "epoch": 0.18350084585530899, "step": 461}, {"loss": 0.4092848300933838, "grad_norm": 0.04038055237675384, "learning_rate": 9.393649782647436e-05, "epoch": 0.18389889541247886, "step": 462}, {"loss": 0.42964720726013184, "grad_norm": 0.04065051635832896, "learning_rate": 9.390569516730525e-05, "epoch": 0.1842969449696487, "step": 463}, {"loss": 0.410689115524292, "grad_norm": 0.03746908233301776, "learning_rate": 9.387481954399334e-05, "epoch": 0.1846949945268186, "step": 464}, {"loss": 0.38294869661331177, "grad_norm": 0.03968377182918642, "learning_rate": 9.384387100784891e-05, "epoch": 0.18509304408398847, "step": 465}, {"loss": 0.4257342219352722, "grad_norm": 0.04005689656481967, "learning_rate": 9.381284961030338e-05, "epoch": 0.18549109364115832, "step": 466}, {"loss": 0.39892587065696716, "grad_norm": 0.04000848399341666, "learning_rate": 9.37817554029093e-05, "epoch": 0.1858891431983282, "step": 467}, {"loss": 0.43865126371383667, "grad_norm": 0.0436143372213534, "learning_rate": 9.37505884373402e-05, "epoch": 0.18628719275549807, "step": 468}, {"loss": 0.3935243487358093, "grad_norm": 0.03814779220448839, "learning_rate": 9.37193487653905e-05, "epoch": 0.18668524231266792, "step": 469}, {"loss": 0.3716939389705658, "grad_norm": 0.03982124712476835, "learning_rate": 9.36880364389755e-05, "epoch": 0.1870832918698378, "step": 470}, {"loss": 0.3581838607788086, "grad_norm": 0.03557709657472714, "learning_rate": 9.365665151013119e-05, "epoch": 0.18748134142700768, "step": 471}, {"loss": 0.4749634861946106, "grad_norm": 0.04149421627593823, "learning_rate": 9.362519403101422e-05, "epoch": 0.18787939098417752, "step": 472}, {"loss": 0.44812238216400146, "grad_norm": 0.041005865000329815, "learning_rate": 9.359366405390183e-05, "epoch": 0.1882774405413474, "step": 473}, {"loss": 0.4141015410423279, "grad_norm": 0.03690933632033149, "learning_rate": 9.356206163119172e-05, "epoch": 0.18867549009851728, "step": 474}, {"loss": 0.3977903723716736, "grad_norm": 0.038225094062033935, "learning_rate": 9.353038681540201e-05, "epoch": 0.18907353965568713, "step": 475}, {"loss": 0.3963810205459595, "grad_norm": 0.03625893391550009, "learning_rate": 9.349863965917107e-05, "epoch": 0.189471589212857, "step": 476}, {"loss": 0.39708709716796875, "grad_norm": 0.04061794528147399, "learning_rate": 9.346682021525755e-05, "epoch": 0.18986963877002686, "step": 477}, {"loss": 0.38378819823265076, "grad_norm": 0.03723516856179152, "learning_rate": 9.34349285365402e-05, "epoch": 0.19026768832719673, "step": 478}, {"loss": 0.39741992950439453, "grad_norm": 0.03875984699232092, "learning_rate": 9.340296467601782e-05, "epoch": 0.1906657378843666, "step": 479}, {"loss": 0.4517289698123932, "grad_norm": 0.04503751771632875, "learning_rate": 9.337092868680913e-05, "epoch": 0.19106378744153646, "step": 480}, {"loss": 0.36869797110557556, "grad_norm": 0.037375444181146876, "learning_rate": 9.33388206221528e-05, "epoch": 0.19146183699870634, "step": 481}, {"loss": 0.3954632580280304, "grad_norm": 0.042671628487007576, "learning_rate": 9.330664053540718e-05, "epoch": 0.1918598865558762, "step": 482}, {"loss": 0.3960896134376526, "grad_norm": 0.03936427285008176, "learning_rate": 9.327438848005036e-05, "epoch": 0.19225793611304606, "step": 483}, {"loss": 0.4135330319404602, "grad_norm": 0.03846092496465794, "learning_rate": 9.324206450968005e-05, "epoch": 0.19265598567021594, "step": 484}, {"loss": 0.3778441846370697, "grad_norm": 0.03765392343096679, "learning_rate": 9.320966867801341e-05, "epoch": 0.19305403522738582, "step": 485}, {"loss": 0.3790014982223511, "grad_norm": 0.038502696887647854, "learning_rate": 9.31772010388871e-05, "epoch": 0.19345208478455567, "step": 486}, {"loss": 0.3788261413574219, "grad_norm": 0.0370770917997647, "learning_rate": 9.314466164625703e-05, "epoch": 0.19385013434172554, "step": 487}, {"loss": 0.3218253552913666, "grad_norm": 0.036480265194914784, "learning_rate": 9.31120505541984e-05, "epoch": 0.19424818389889542, "step": 488}, {"loss": 0.45690417289733887, "grad_norm": 0.04030608526440649, "learning_rate": 9.307936781690558e-05, "epoch": 0.19464623345606527, "step": 489}, {"loss": 0.416892409324646, "grad_norm": 0.03897972412641558, "learning_rate": 9.304661348869195e-05, "epoch": 0.19504428301323515, "step": 490}, {"loss": 0.38784995675086975, "grad_norm": 0.03785495834817757, "learning_rate": 9.30137876239899e-05, "epoch": 0.19544233257040503, "step": 491}, {"loss": 0.42070573568344116, "grad_norm": 0.03988143445643504, "learning_rate": 9.298089027735068e-05, "epoch": 0.19584038212757487, "step": 492}, {"loss": 0.38060417771339417, "grad_norm": 0.034447651478455056, "learning_rate": 9.294792150344438e-05, "epoch": 0.19623843168474475, "step": 493}, {"loss": 0.441726952791214, "grad_norm": 0.04030796102556379, "learning_rate": 9.291488135705971e-05, "epoch": 0.19663648124191463, "step": 494}, {"loss": 0.36312973499298096, "grad_norm": 0.03473888271716477, "learning_rate": 9.288176989310404e-05, "epoch": 0.19703453079908448, "step": 495}, {"loss": 0.38864463567733765, "grad_norm": 0.0361158912261295, "learning_rate": 9.284858716660328e-05, "epoch": 0.19743258035625436, "step": 496}, {"loss": 0.37512272596359253, "grad_norm": 0.041479162818646935, "learning_rate": 9.281533323270171e-05, "epoch": 0.19783062991342423, "step": 497}, {"loss": 0.3529120683670044, "grad_norm": 0.0369493696654319, "learning_rate": 9.278200814666201e-05, "epoch": 0.19822867947059408, "step": 498}, {"loss": 0.424915075302124, "grad_norm": 0.03978023920549289, "learning_rate": 9.2748611963865e-05, "epoch": 0.19862672902776396, "step": 499}, {"loss": 0.39221230149269104, "grad_norm": 0.041162284658601604, "learning_rate": 9.271514473980979e-05, "epoch": 0.19902477858493384, "step": 500}, {"loss": 0.3956616520881653, "grad_norm": 0.040980848948175455, "learning_rate": 9.268160653011344e-05, "epoch": 0.1994228281421037, "step": 501}, {"loss": 0.3975604176521301, "grad_norm": 0.0380130126653583, "learning_rate": 9.264799739051102e-05, "epoch": 0.19982087769927356, "step": 502}, {"loss": 0.39644166827201843, "grad_norm": 0.039258813600085714, "learning_rate": 9.261431737685546e-05, "epoch": 0.20021892725644344, "step": 503}, {"loss": 0.44583815336227417, "grad_norm": 0.040895208569298944, "learning_rate": 9.258056654511751e-05, "epoch": 0.2006169768136133, "step": 504}, {"loss": 0.38320890069007874, "grad_norm": 0.03932423497237871, "learning_rate": 9.254674495138554e-05, "epoch": 0.20101502637078317, "step": 505}, {"loss": 0.3923526406288147, "grad_norm": 0.036874675948727974, "learning_rate": 9.251285265186556e-05, "epoch": 0.20141307592795302, "step": 506}, {"loss": 0.4181939959526062, "grad_norm": 0.040461789269802394, "learning_rate": 9.24788897028811e-05, "epoch": 0.2018111254851229, "step": 507}, {"loss": 0.39845073223114014, "grad_norm": 0.03969060818807724, "learning_rate": 9.244485616087304e-05, "epoch": 0.20220917504229277, "step": 508}, {"loss": 0.37429162859916687, "grad_norm": 0.035123055713713804, "learning_rate": 9.24107520823996e-05, "epoch": 0.20260722459946262, "step": 509}, {"loss": 0.42061781883239746, "grad_norm": 0.04448080141277121, "learning_rate": 9.237657752413626e-05, "epoch": 0.2030052741566325, "step": 510}, {"loss": 0.4075678884983063, "grad_norm": 0.03743252021305215, "learning_rate": 9.234233254287559e-05, "epoch": 0.20340332371380238, "step": 511}, {"loss": 0.36916035413742065, "grad_norm": 0.03811298253840233, "learning_rate": 9.230801719552717e-05, "epoch": 0.20380137327097222, "step": 512}, {"loss": 0.38644856214523315, "grad_norm": 0.03683859569892099, "learning_rate": 9.227363153911754e-05, "epoch": 0.2041994228281421, "step": 513}, {"loss": 0.35839933156967163, "grad_norm": 0.03550445142243576, "learning_rate": 9.22391756307901e-05, "epoch": 0.20459747238531198, "step": 514}, {"loss": 0.4136126935482025, "grad_norm": 0.039396513063068364, "learning_rate": 9.220464952780498e-05, "epoch": 0.20499552194248183, "step": 515}, {"loss": 0.3962666988372803, "grad_norm": 0.03843812142728689, "learning_rate": 9.217005328753894e-05, "epoch": 0.2053935714996517, "step": 516}, {"loss": 0.3547739088535309, "grad_norm": 0.035766823593989125, "learning_rate": 9.213538696748534e-05, "epoch": 0.20579162105682158, "step": 517}, {"loss": 0.4059969484806061, "grad_norm": 0.03784698722735649, "learning_rate": 9.210065062525398e-05, "epoch": 0.20618967061399143, "step": 518}, {"loss": 0.42336809635162354, "grad_norm": 0.037637463849399304, "learning_rate": 9.206584431857101e-05, "epoch": 0.2065877201711613, "step": 519}, {"loss": 0.4129765033721924, "grad_norm": 0.04273812105340694, "learning_rate": 9.203096810527888e-05, "epoch": 0.2069857697283312, "step": 520}, {"loss": 0.4090093970298767, "grad_norm": 0.03803043228587778, "learning_rate": 9.19960220433362e-05, "epoch": 0.20738381928550104, "step": 521}, {"loss": 0.3777045011520386, "grad_norm": 0.03523742925131892, "learning_rate": 9.196100619081764e-05, "epoch": 0.2077818688426709, "step": 522}, {"loss": 0.4249908924102783, "grad_norm": 0.03427974593061213, "learning_rate": 9.192592060591387e-05, "epoch": 0.2081799183998408, "step": 523}, {"loss": 0.444797158241272, "grad_norm": 0.03947001369903298, "learning_rate": 9.189076534693146e-05, "epoch": 0.20857796795701064, "step": 524}, {"loss": 0.40261805057525635, "grad_norm": 0.03960082243022491, "learning_rate": 9.185554047229273e-05, "epoch": 0.20897601751418052, "step": 525}, {"loss": 0.3896107077598572, "grad_norm": 0.03807428381027277, "learning_rate": 9.182024604053572e-05, "epoch": 0.2093740670713504, "step": 526}, {"loss": 0.36818450689315796, "grad_norm": 0.038316627697183823, "learning_rate": 9.178488211031402e-05, "epoch": 0.20977211662852024, "step": 527}, {"loss": 0.42621904611587524, "grad_norm": 0.039533709043752925, "learning_rate": 9.174944874039678e-05, "epoch": 0.21017016618569012, "step": 528}, {"loss": 0.4162560701370239, "grad_norm": 0.03706160272942852, "learning_rate": 9.17139459896685e-05, "epoch": 0.21056821574286, "step": 529}, {"loss": 0.3932320475578308, "grad_norm": 0.035084277299559125, "learning_rate": 9.167837391712899e-05, "epoch": 0.21096626530002985, "step": 530}, {"loss": 0.4399354159832001, "grad_norm": 0.040819003240830924, "learning_rate": 9.164273258189325e-05, "epoch": 0.21136431485719973, "step": 531}, {"loss": 0.3811988830566406, "grad_norm": 0.04233171370561854, "learning_rate": 9.160702204319143e-05, "epoch": 0.2117623644143696, "step": 532}, {"loss": 0.41985881328582764, "grad_norm": 0.038210151570938364, "learning_rate": 9.157124236036864e-05, "epoch": 0.21216041397153945, "step": 533}, {"loss": 0.3473092019557953, "grad_norm": 0.03504725248642579, "learning_rate": 9.153539359288487e-05, "epoch": 0.21255846352870933, "step": 534}, {"loss": 0.35103362798690796, "grad_norm": 0.03459599815017307, "learning_rate": 9.149947580031503e-05, "epoch": 0.21295651308587918, "step": 535}, {"loss": 0.3905608057975769, "grad_norm": 0.03639330965890202, "learning_rate": 9.146348904234861e-05, "epoch": 0.21335456264304906, "step": 536}, {"loss": 0.3902738690376282, "grad_norm": 0.037971474723538595, "learning_rate": 9.142743337878977e-05, "epoch": 0.21375261220021893, "step": 537}, {"loss": 0.39158937335014343, "grad_norm": 0.03818269788948331, "learning_rate": 9.139130886955718e-05, "epoch": 0.21415066175738878, "step": 538}, {"loss": 0.3802120089530945, "grad_norm": 0.03883035724550337, "learning_rate": 9.135511557468392e-05, "epoch": 0.21454871131455866, "step": 539}, {"loss": 0.42220890522003174, "grad_norm": 0.040753306118308366, "learning_rate": 9.131885355431738e-05, "epoch": 0.21494676087172854, "step": 540}, {"loss": 0.40389716625213623, "grad_norm": 0.039578488821235464, "learning_rate": 9.128252286871912e-05, "epoch": 0.2153448104288984, "step": 541}, {"loss": 0.4262523055076599, "grad_norm": 0.043145021585125065, "learning_rate": 9.12461235782649e-05, "epoch": 0.21574285998606826, "step": 542}, {"loss": 0.4230383336544037, "grad_norm": 0.03873019831846949, "learning_rate": 9.12096557434444e-05, "epoch": 0.21614090954323814, "step": 543}, {"loss": 0.412395715713501, "grad_norm": 0.04393616220911426, "learning_rate": 9.117311942486122e-05, "epoch": 0.216538959100408, "step": 544}, {"loss": 0.3651766777038574, "grad_norm": 0.03820521015220515, "learning_rate": 9.113651468323287e-05, "epoch": 0.21693700865757787, "step": 545}, {"loss": 0.36060643196105957, "grad_norm": 0.03855308642785342, "learning_rate": 9.109984157939042e-05, "epoch": 0.21733505821474774, "step": 546}, {"loss": 0.4345490634441376, "grad_norm": 0.04500090714658066, "learning_rate": 9.106310017427865e-05, "epoch": 0.2177331077719176, "step": 547}, {"loss": 0.3955548405647278, "grad_norm": 0.0362786996535705, "learning_rate": 9.102629052895584e-05, "epoch": 0.21813115732908747, "step": 548}, {"loss": 0.40128546953201294, "grad_norm": 0.037720216559957775, "learning_rate": 9.09894127045936e-05, "epoch": 0.21852920688625735, "step": 549}, {"loss": 0.3925573229789734, "grad_norm": 0.037524929488709856, "learning_rate": 9.095246676247693e-05, "epoch": 0.2189272564434272, "step": 550}, {"loss": 0.363197922706604, "grad_norm": 0.037098935413675725, "learning_rate": 9.091545276400397e-05, "epoch": 0.21932530600059708, "step": 551}, {"loss": 0.35345137119293213, "grad_norm": 0.03547272582948809, "learning_rate": 9.087837077068597e-05, "epoch": 0.21972335555776695, "step": 552}, {"loss": 0.39264020323753357, "grad_norm": 0.03779872926526867, "learning_rate": 9.084122084414723e-05, "epoch": 0.2201214051149368, "step": 553}, {"loss": 0.3902384042739868, "grad_norm": 0.038663915218007855, "learning_rate": 9.080400304612487e-05, "epoch": 0.22051945467210668, "step": 554}, {"loss": 0.37022390961647034, "grad_norm": 0.035982900113494434, "learning_rate": 9.076671743846883e-05, "epoch": 0.22091750422927656, "step": 555}, {"loss": 0.433847576379776, "grad_norm": 0.03781653779780315, "learning_rate": 9.072936408314176e-05, "epoch": 0.2213155537864464, "step": 556}, {"loss": 0.3591976761817932, "grad_norm": 0.035927496485301094, "learning_rate": 9.069194304221886e-05, "epoch": 0.22171360334361628, "step": 557}, {"loss": 0.36309656500816345, "grad_norm": 0.03726579306680338, "learning_rate": 9.065445437788783e-05, "epoch": 0.22211165290078616, "step": 558}, {"loss": 0.3893352746963501, "grad_norm": 0.0364873065746809, "learning_rate": 9.061689815244876e-05, "epoch": 0.222509702457956, "step": 559}, {"loss": 0.3783146142959595, "grad_norm": 0.03580397752823344, "learning_rate": 9.0579274428314e-05, "epoch": 0.2229077520151259, "step": 560}, {"loss": 0.38781294226646423, "grad_norm": 0.04111246781744364, "learning_rate": 9.054158326800808e-05, "epoch": 0.22330580157229576, "step": 561}, {"loss": 0.4295772314071655, "grad_norm": 0.03731126682248044, "learning_rate": 9.050382473416758e-05, "epoch": 0.22370385112946561, "step": 562}, {"loss": 0.39483675360679626, "grad_norm": 0.03602529444451627, "learning_rate": 9.046599888954108e-05, "epoch": 0.2241019006866355, "step": 563}, {"loss": 0.3859916925430298, "grad_norm": 0.040049593916826096, "learning_rate": 9.042810579698898e-05, "epoch": 0.22449995024380534, "step": 564}, {"loss": 0.35610121488571167, "grad_norm": 0.03546474830509761, "learning_rate": 9.039014551948346e-05, "epoch": 0.22489799980097522, "step": 565}, {"loss": 0.39040476083755493, "grad_norm": 0.037793866977665354, "learning_rate": 9.035211812010833e-05, "epoch": 0.2252960493581451, "step": 566}, {"loss": 0.45170941948890686, "grad_norm": 0.042779427454858215, "learning_rate": 9.031402366205898e-05, "epoch": 0.22569409891531494, "step": 567}, {"loss": 0.40703094005584717, "grad_norm": 0.04190017260682866, "learning_rate": 9.02758622086422e-05, "epoch": 0.22609214847248482, "step": 568}, {"loss": 0.44918861985206604, "grad_norm": 0.04270778054097073, "learning_rate": 9.023763382327614e-05, "epoch": 0.2264901980296547, "step": 569}, {"loss": 0.34330329298973083, "grad_norm": 0.03679695320993222, "learning_rate": 9.019933856949019e-05, "epoch": 0.22688824758682455, "step": 570}, {"loss": 0.4268772602081299, "grad_norm": 0.038559867807779216, "learning_rate": 9.016097651092483e-05, "epoch": 0.22728629714399443, "step": 571}, {"loss": 0.4083315432071686, "grad_norm": 0.03842542655610064, "learning_rate": 9.01225477113316e-05, "epoch": 0.2276843467011643, "step": 572}, {"loss": 0.3933855891227722, "grad_norm": 0.039600877095917, "learning_rate": 9.00840522345729e-05, "epoch": 0.22808239625833415, "step": 573}, {"loss": 0.3908676505088806, "grad_norm": 0.036241603190823084, "learning_rate": 9.004549014462199e-05, "epoch": 0.22848044581550403, "step": 574}, {"loss": 0.39349979162216187, "grad_norm": 0.03771998253824872, "learning_rate": 9.00068615055628e-05, "epoch": 0.2288784953726739, "step": 575}, {"loss": 0.3953763246536255, "grad_norm": 0.03509784311492775, "learning_rate": 8.996816638158986e-05, "epoch": 0.22927654492984376, "step": 576}, {"loss": 0.4008740484714508, "grad_norm": 0.035839066676633294, "learning_rate": 8.992940483700821e-05, "epoch": 0.22967459448701363, "step": 577}, {"loss": 0.3761654794216156, "grad_norm": 0.041871415600780275, "learning_rate": 8.989057693623321e-05, "epoch": 0.2300726440441835, "step": 578}, {"loss": 0.4256053566932678, "grad_norm": 0.040175867715994996, "learning_rate": 8.985168274379057e-05, "epoch": 0.23047069360135336, "step": 579}, {"loss": 0.4153198003768921, "grad_norm": 0.03964649168382834, "learning_rate": 8.98127223243161e-05, "epoch": 0.23086874315852324, "step": 580}, {"loss": 0.3764423429965973, "grad_norm": 0.03662604382546945, "learning_rate": 8.977369574255571e-05, "epoch": 0.23126679271569311, "step": 581}, {"loss": 0.45945823192596436, "grad_norm": 0.03921081110519695, "learning_rate": 8.973460306336526e-05, "epoch": 0.23166484227286296, "step": 582}, {"loss": 0.36577486991882324, "grad_norm": 0.03665758433724948, "learning_rate": 8.969544435171042e-05, "epoch": 0.23206289183003284, "step": 583}, {"loss": 0.3744468688964844, "grad_norm": 0.03549710719983845, "learning_rate": 8.965621967266662e-05, "epoch": 0.23246094138720272, "step": 584}, {"loss": 0.3672794699668884, "grad_norm": 0.03807237375537043, "learning_rate": 8.961692909141892e-05, "epoch": 0.23285899094437257, "step": 585}, {"loss": 0.43532416224479675, "grad_norm": 0.043065915859636116, "learning_rate": 8.957757267326188e-05, "epoch": 0.23325704050154245, "step": 586}, {"loss": 0.4014749825000763, "grad_norm": 0.03941207735189246, "learning_rate": 8.953815048359951e-05, "epoch": 0.23365509005871232, "step": 587}, {"loss": 0.38080549240112305, "grad_norm": 0.037816457801518676, "learning_rate": 8.949866258794506e-05, "epoch": 0.23405313961588217, "step": 588}, {"loss": 0.36971762776374817, "grad_norm": 0.036765640982591094, "learning_rate": 8.945910905192103e-05, "epoch": 0.23445118917305205, "step": 589}, {"loss": 0.4530743956565857, "grad_norm": 0.041338031846932535, "learning_rate": 8.941948994125897e-05, "epoch": 0.23484923873022193, "step": 590}, {"loss": 0.35865318775177, "grad_norm": 0.03641837649884965, "learning_rate": 8.937980532179943e-05, "epoch": 0.23524728828739178, "step": 591}, {"loss": 0.3391808271408081, "grad_norm": 0.031236813907833397, "learning_rate": 8.934005525949179e-05, "epoch": 0.23564533784456165, "step": 592}, {"loss": 0.4195290505886078, "grad_norm": 0.037089183945584574, "learning_rate": 8.93002398203942e-05, "epoch": 0.2360433874017315, "step": 593}, {"loss": 0.42962151765823364, "grad_norm": 0.040664665832737354, "learning_rate": 8.92603590706735e-05, "epoch": 0.23644143695890138, "step": 594}, {"loss": 0.3763956129550934, "grad_norm": 0.038687698590185, "learning_rate": 8.922041307660497e-05, "epoch": 0.23683948651607126, "step": 595}, {"loss": 0.3744131922721863, "grad_norm": 0.03781222762844978, "learning_rate": 8.918040190457242e-05, "epoch": 0.2372375360732411, "step": 596}, {"loss": 0.38545045256614685, "grad_norm": 0.036640329233150595, "learning_rate": 8.91403256210679e-05, "epoch": 0.23763558563041098, "step": 597}, {"loss": 0.3835058808326721, "grad_norm": 0.0355908278866354, "learning_rate": 8.91001842926917e-05, "epoch": 0.23803363518758086, "step": 598}, {"loss": 0.3801095485687256, "grad_norm": 0.03315098247970726, "learning_rate": 8.90599779861522e-05, "epoch": 0.2384316847447507, "step": 599}, {"loss": 0.33961427211761475, "grad_norm": 0.03633326149240091, "learning_rate": 8.901970676826576e-05, "epoch": 0.2388297343019206, "step": 600}, {"loss": 0.40508195757865906, "grad_norm": 0.03525441920097117, "learning_rate": 8.89793707059566e-05, "epoch": 0.23922778385909046, "step": 601}, {"loss": 0.39710330963134766, "grad_norm": 0.036868443064919064, "learning_rate": 8.893896986625673e-05, "epoch": 0.23962583341626031, "step": 602}, {"loss": 0.37115728855133057, "grad_norm": 0.035067470071481796, "learning_rate": 8.889850431630577e-05, "epoch": 0.2400238829734302, "step": 603}, {"loss": 0.38503238558769226, "grad_norm": 0.03669997274189583, "learning_rate": 8.88579741233509e-05, "epoch": 0.24042193253060007, "step": 604}, {"loss": 0.44429969787597656, "grad_norm": 0.041985882826313, "learning_rate": 8.881737935474677e-05, "epoch": 0.24081998208776992, "step": 605}, {"loss": 0.37965497374534607, "grad_norm": 0.03790022319809584, "learning_rate": 8.877672007795524e-05, "epoch": 0.2412180316449398, "step": 606}, {"loss": 0.33939552307128906, "grad_norm": 0.03628371115352106, "learning_rate": 8.873599636054544e-05, "epoch": 0.24161608120210967, "step": 607}, {"loss": 0.4334941804409027, "grad_norm": 0.03983184561487862, "learning_rate": 8.869520827019364e-05, "epoch": 0.24201413075927952, "step": 608}, {"loss": 0.3799220323562622, "grad_norm": 0.037120360139998976, "learning_rate": 8.865435587468296e-05, "epoch": 0.2424121803164494, "step": 609}, {"loss": 0.33312225341796875, "grad_norm": 0.035779735752197166, "learning_rate": 8.86134392419035e-05, "epoch": 0.24281022987361928, "step": 610}, {"loss": 0.39519745111465454, "grad_norm": 0.0357122511246499, "learning_rate": 8.857245843985204e-05, "epoch": 0.24320827943078913, "step": 611}, {"loss": 0.3984397053718567, "grad_norm": 0.03864844147335651, "learning_rate": 8.853141353663203e-05, "epoch": 0.243606328987959, "step": 612}, {"loss": 0.40167737007141113, "grad_norm": 0.041203126708949246, "learning_rate": 8.849030460045345e-05, "epoch": 0.24400437854512888, "step": 613}, {"loss": 0.3864404559135437, "grad_norm": 0.03540329074096251, "learning_rate": 8.844913169963269e-05, "epoch": 0.24440242810229873, "step": 614}, {"loss": 0.37281376123428345, "grad_norm": 0.03678903448097713, "learning_rate": 8.84078949025924e-05, "epoch": 0.2448004776594686, "step": 615}, {"loss": 0.44499701261520386, "grad_norm": 0.043779647404535556, "learning_rate": 8.836659427786149e-05, "epoch": 0.24519852721663848, "step": 616}, {"loss": 0.4118572473526001, "grad_norm": 0.03920707480958235, "learning_rate": 8.832522989407485e-05, "epoch": 0.24559657677380833, "step": 617}, {"loss": 0.3444732427597046, "grad_norm": 0.038823857629764716, "learning_rate": 8.828380181997341e-05, "epoch": 0.2459946263309782, "step": 618}, {"loss": 0.4050634503364563, "grad_norm": 0.03504896896700958, "learning_rate": 8.82423101244039e-05, "epoch": 0.2463926758881481, "step": 619}, {"loss": 0.3728637099266052, "grad_norm": 0.03910165946914318, "learning_rate": 8.82007548763188e-05, "epoch": 0.24679072544531794, "step": 620}, {"loss": 0.41125309467315674, "grad_norm": 0.04355651256712523, "learning_rate": 8.815913614477615e-05, "epoch": 0.24718877500248781, "step": 621}, {"loss": 0.3808304965496063, "grad_norm": 0.04083377559676517, "learning_rate": 8.811745399893958e-05, "epoch": 0.24758682455965766, "step": 622}, {"loss": 0.37254613637924194, "grad_norm": 0.035549082641097934, "learning_rate": 8.807570850807803e-05, "epoch": 0.24798487411682754, "step": 623}, {"loss": 0.39604347944259644, "grad_norm": 0.03806756199071817, "learning_rate": 8.803389974156573e-05, "epoch": 0.24838292367399742, "step": 624}, {"loss": 0.40497732162475586, "grad_norm": 0.04058396990162811, "learning_rate": 8.799202776888206e-05, "epoch": 0.24878097323116727, "step": 625}, {"loss": 0.3908602297306061, "grad_norm": 0.03966280909813803, "learning_rate": 8.79500926596115e-05, "epoch": 0.24917902278833715, "step": 626}, {"loss": 0.39297181367874146, "grad_norm": 0.03541654954761812, "learning_rate": 8.790809448344333e-05, "epoch": 0.24957707234550702, "step": 627}, {"loss": 0.39785006642341614, "grad_norm": 0.0373506625184716, "learning_rate": 8.786603331017172e-05, "epoch": 0.24997512190267687, "step": 628}, {"loss": 0.3680535554885864, "grad_norm": 0.036523594522343515, "learning_rate": 8.782390920969554e-05, "epoch": 0.2503731714598468, "step": 629}, {"loss": 0.34212344884872437, "grad_norm": 0.03670731615614106, "learning_rate": 8.778172225201822e-05, "epoch": 0.2507712210170166, "step": 630}, {"loss": 0.3185552954673767, "grad_norm": 0.03610923474964367, "learning_rate": 8.773947250724761e-05, "epoch": 0.2511692705741865, "step": 631}, {"loss": 0.4392058551311493, "grad_norm": 0.0457991690841126, "learning_rate": 8.769716004559594e-05, "epoch": 0.25156732013135635, "step": 632}, {"loss": 0.37778306007385254, "grad_norm": 0.0412506943804001, "learning_rate": 8.765478493737967e-05, "epoch": 0.25196536968852623, "step": 633}, {"loss": 0.386160671710968, "grad_norm": 0.03686642918635565, "learning_rate": 8.761234725301935e-05, "epoch": 0.2523634192456961, "step": 634}, {"loss": 0.3983137607574463, "grad_norm": 0.03867248896268702, "learning_rate": 8.75698470630395e-05, "epoch": 0.25276146880286593, "step": 635}, {"loss": 0.39499515295028687, "grad_norm": 0.039654857292842044, "learning_rate": 8.752728443806859e-05, "epoch": 0.2531595183600358, "step": 636}, {"loss": 0.3555513322353363, "grad_norm": 0.03444407318148271, "learning_rate": 8.748465944883877e-05, "epoch": 0.2535575679172057, "step": 637}, {"loss": 0.3964555263519287, "grad_norm": 0.03965600189769821, "learning_rate": 8.744197216618584e-05, "epoch": 0.25395561747437556, "step": 638}, {"loss": 0.4043642580509186, "grad_norm": 0.04249810922317728, "learning_rate": 8.739922266104918e-05, "epoch": 0.25435366703154544, "step": 639}, {"loss": 0.436888188123703, "grad_norm": 0.04144375492222758, "learning_rate": 8.73564110044715e-05, "epoch": 0.2547517165887153, "step": 640}, {"loss": 0.39327865839004517, "grad_norm": 0.06059235599553319, "learning_rate": 8.731353726759883e-05, "epoch": 0.25514976614588514, "step": 641}, {"loss": 0.37052983045578003, "grad_norm": 0.035133333027062674, "learning_rate": 8.72706015216804e-05, "epoch": 0.255547815703055, "step": 642}, {"loss": 0.3960340619087219, "grad_norm": 0.03840259684905385, "learning_rate": 8.722760383806841e-05, "epoch": 0.2559458652602249, "step": 643}, {"loss": 0.45931729674339294, "grad_norm": 0.03829338819957176, "learning_rate": 8.718454428821806e-05, "epoch": 0.25634391481739477, "step": 644}, {"loss": 0.359139621257782, "grad_norm": 0.03794221323871482, "learning_rate": 8.714142294368734e-05, "epoch": 0.25674196437456465, "step": 645}, {"loss": 0.38493674993515015, "grad_norm": 0.03515270753066405, "learning_rate": 8.709823987613691e-05, "epoch": 0.2571400139317345, "step": 646}, {"loss": 0.3837330937385559, "grad_norm": 0.03647069333595731, "learning_rate": 8.705499515733005e-05, "epoch": 0.25753806348890435, "step": 647}, {"loss": 0.3751949071884155, "grad_norm": 0.03803250272665651, "learning_rate": 8.701168885913243e-05, "epoch": 0.2579361130460742, "step": 648}, {"loss": 0.39679163694381714, "grad_norm": 0.03990880031146932, "learning_rate": 8.696832105351212e-05, "epoch": 0.2583341626032441, "step": 649}, {"loss": 0.4168563187122345, "grad_norm": 0.036824871925601736, "learning_rate": 8.692489181253937e-05, "epoch": 0.258732212160414, "step": 650}, {"loss": 0.3511507511138916, "grad_norm": 0.03625213484524976, "learning_rate": 8.688140120838652e-05, "epoch": 0.25913026171758385, "step": 651}, {"loss": 0.3915506899356842, "grad_norm": 0.03809184440322537, "learning_rate": 8.683784931332794e-05, "epoch": 0.25952831127475373, "step": 652}, {"loss": 0.3794567584991455, "grad_norm": 0.037471897818388174, "learning_rate": 8.679423619973976e-05, "epoch": 0.25992636083192355, "step": 653}, {"loss": 0.3978666067123413, "grad_norm": 0.04076172096732682, "learning_rate": 8.675056194009996e-05, "epoch": 0.26032441038909343, "step": 654}, {"loss": 0.37394654750823975, "grad_norm": 0.03720196530158318, "learning_rate": 8.670682660698801e-05, "epoch": 0.2607224599462633, "step": 655}, {"loss": 0.37431269884109497, "grad_norm": 0.038981708896965346, "learning_rate": 8.666303027308499e-05, "epoch": 0.2611205095034332, "step": 656}, {"loss": 0.39312508702278137, "grad_norm": 0.037324457384801654, "learning_rate": 8.661917301117328e-05, "epoch": 0.26151855906060306, "step": 657}, {"loss": 0.4206024408340454, "grad_norm": 0.03712810052714979, "learning_rate": 8.657525489413651e-05, "epoch": 0.26191660861777294, "step": 658}, {"loss": 0.380302757024765, "grad_norm": 0.03870429328140876, "learning_rate": 8.653127599495951e-05, "epoch": 0.26231465817494276, "step": 659}, {"loss": 0.3796495795249939, "grad_norm": 0.034275423008128636, "learning_rate": 8.648723638672806e-05, "epoch": 0.26271270773211264, "step": 660}, {"loss": 0.3661483824253082, "grad_norm": 0.03532339380036525, "learning_rate": 8.644313614262887e-05, "epoch": 0.2631107572892825, "step": 661}, {"loss": 0.3940781354904175, "grad_norm": 0.035931802867702284, "learning_rate": 8.639897533594934e-05, "epoch": 0.2635088068464524, "step": 662}, {"loss": 0.39178138971328735, "grad_norm": 0.03841859075197349, "learning_rate": 8.635475404007758e-05, "epoch": 0.26390685640362227, "step": 663}, {"loss": 0.4154405891895294, "grad_norm": 0.03424105910337299, "learning_rate": 8.631047232850223e-05, "epoch": 0.2643049059607921, "step": 664}, {"loss": 0.40125495195388794, "grad_norm": 0.03461290589729631, "learning_rate": 8.62661302748123e-05, "epoch": 0.26470295551796197, "step": 665}, {"loss": 0.3873884081840515, "grad_norm": 0.03606515414121197, "learning_rate": 8.62217279526971e-05, "epoch": 0.26510100507513185, "step": 666}, {"loss": 0.40063783526420593, "grad_norm": 0.038495248367605595, "learning_rate": 8.617726543594609e-05, "epoch": 0.2654990546323017, "step": 667}, {"loss": 0.40470701456069946, "grad_norm": 0.03991099342970727, "learning_rate": 8.613274279844873e-05, "epoch": 0.2658971041894716, "step": 668}, {"loss": 0.3570280075073242, "grad_norm": 0.03426716410791367, "learning_rate": 8.608816011419445e-05, "epoch": 0.2662951537466415, "step": 669}, {"loss": 0.3760698437690735, "grad_norm": 0.03521995561132577, "learning_rate": 8.604351745727241e-05, "epoch": 0.2666932033038113, "step": 670}, {"loss": 0.3955170810222626, "grad_norm": 0.03771663078796589, "learning_rate": 8.59988149018715e-05, "epoch": 0.2670912528609812, "step": 671}, {"loss": 0.3615563213825226, "grad_norm": 0.039811553643530825, "learning_rate": 8.59540525222801e-05, "epoch": 0.26748930241815105, "step": 672}, {"loss": 0.3844960331916809, "grad_norm": 0.037597058158181165, "learning_rate": 8.590923039288601e-05, "epoch": 0.26788735197532093, "step": 673}, {"loss": 0.40448611974716187, "grad_norm": 0.03492925471621575, "learning_rate": 8.586434858817635e-05, "epoch": 0.2682854015324908, "step": 674}, {"loss": 0.3617110252380371, "grad_norm": 0.036728853292519584, "learning_rate": 8.58194071827374e-05, "epoch": 0.2686834510896607, "step": 675}, {"loss": 0.3999275863170624, "grad_norm": 0.034638001285308806, "learning_rate": 8.577440625125447e-05, "epoch": 0.2690815006468305, "step": 676}, {"loss": 0.40268033742904663, "grad_norm": 0.038077361884221735, "learning_rate": 8.572934586851184e-05, "epoch": 0.2694795502040004, "step": 677}, {"loss": 0.3660908639431, "grad_norm": 0.03440987884830661, "learning_rate": 8.568422610939252e-05, "epoch": 0.26987759976117026, "step": 678}, {"loss": 0.4061199426651001, "grad_norm": 0.037142693425613584, "learning_rate": 8.563904704887826e-05, "epoch": 0.27027564931834014, "step": 679}, {"loss": 0.418781042098999, "grad_norm": 0.038946252694892936, "learning_rate": 8.559380876204932e-05, "epoch": 0.27067369887551, "step": 680}, {"loss": 0.39566823840141296, "grad_norm": 0.03596859397260779, "learning_rate": 8.554851132408438e-05, "epoch": 0.2710717484326799, "step": 681}, {"loss": 0.40405285358428955, "grad_norm": 0.04209323862097724, "learning_rate": 8.550315481026043e-05, "epoch": 0.2714697979898497, "step": 682}, {"loss": 0.38496607542037964, "grad_norm": 0.03402384614121257, "learning_rate": 8.545773929595268e-05, "epoch": 0.2718678475470196, "step": 683}, {"loss": 0.41825902462005615, "grad_norm": 0.03753219275179449, "learning_rate": 8.541226485663429e-05, "epoch": 0.27226589710418947, "step": 684}, {"loss": 0.394442617893219, "grad_norm": 0.035803619712785055, "learning_rate": 8.536673156787645e-05, "epoch": 0.27266394666135935, "step": 685}, {"loss": 0.36176806688308716, "grad_norm": 0.03716755284270195, "learning_rate": 8.532113950534807e-05, "epoch": 0.2730619962185292, "step": 686}, {"loss": 0.42424139380455017, "grad_norm": 0.03883287481597833, "learning_rate": 8.527548874481578e-05, "epoch": 0.2734600457756991, "step": 687}, {"loss": 0.42573267221450806, "grad_norm": 0.038762912060324214, "learning_rate": 8.522977936214373e-05, "epoch": 0.2738580953328689, "step": 688}, {"loss": 0.379724383354187, "grad_norm": 0.03618767715397237, "learning_rate": 8.518401143329348e-05, "epoch": 0.2742561448900388, "step": 689}, {"loss": 0.41350919008255005, "grad_norm": 0.03884387320494886, "learning_rate": 8.513818503432393e-05, "epoch": 0.2746541944472087, "step": 690}, {"loss": 0.39474934339523315, "grad_norm": 0.03837453276924398, "learning_rate": 8.509230024139113e-05, "epoch": 0.27505224400437855, "step": 691}, {"loss": 0.35641753673553467, "grad_norm": 0.035804918717143724, "learning_rate": 8.504635713074814e-05, "epoch": 0.27545029356154843, "step": 692}, {"loss": 0.37054699659347534, "grad_norm": 0.03648505701191637, "learning_rate": 8.500035577874494e-05, "epoch": 0.27584834311871825, "step": 693}, {"loss": 0.40173178911209106, "grad_norm": 0.03981817001085139, "learning_rate": 8.495429626182834e-05, "epoch": 0.27624639267588813, "step": 694}, {"loss": 0.38630959391593933, "grad_norm": 0.03647743662889361, "learning_rate": 8.49081786565418e-05, "epoch": 0.276644442233058, "step": 695}, {"loss": 0.33930444717407227, "grad_norm": 0.03579109358459726, "learning_rate": 8.486200303952527e-05, "epoch": 0.2770424917902279, "step": 696}, {"loss": 0.4114375114440918, "grad_norm": 0.04083903713947536, "learning_rate": 8.481576948751512e-05, "epoch": 0.27744054134739776, "step": 697}, {"loss": 0.40964269638061523, "grad_norm": 0.039700163817253956, "learning_rate": 8.476947807734407e-05, "epoch": 0.27783859090456764, "step": 698}, {"loss": 0.3947267234325409, "grad_norm": 0.03797140361369097, "learning_rate": 8.47231288859409e-05, "epoch": 0.27823664046173746, "step": 699}, {"loss": 0.3801184892654419, "grad_norm": 0.03774770813582373, "learning_rate": 8.467672199033044e-05, "epoch": 0.27863469001890734, "step": 700}, {"loss": 0.3973364233970642, "grad_norm": 0.04221170447557253, "learning_rate": 8.463025746763342e-05, "epoch": 0.2790327395760772, "step": 701}, {"loss": 0.3989131450653076, "grad_norm": 0.04200006130405829, "learning_rate": 8.458373539506638e-05, "epoch": 0.2794307891332471, "step": 702}, {"loss": 0.4226028323173523, "grad_norm": 0.0454110541401327, "learning_rate": 8.453715584994141e-05, "epoch": 0.27982883869041697, "step": 703}, {"loss": 0.3719666004180908, "grad_norm": 0.03682004628022101, "learning_rate": 8.449051890966617e-05, "epoch": 0.28022688824758685, "step": 704}, {"loss": 0.4142642915248871, "grad_norm": 0.041578761706322805, "learning_rate": 8.444382465174369e-05, "epoch": 0.28062493780475667, "step": 705}, {"loss": 0.3955291509628296, "grad_norm": 0.03769763810616539, "learning_rate": 8.439707315377228e-05, "epoch": 0.28102298736192655, "step": 706}, {"loss": 0.3641568124294281, "grad_norm": 0.037970056465557706, "learning_rate": 8.43502644934453e-05, "epoch": 0.2814210369190964, "step": 707}, {"loss": 0.3572172224521637, "grad_norm": 0.03883168708596445, "learning_rate": 8.430339874855118e-05, "epoch": 0.2818190864762663, "step": 708}, {"loss": 0.349972665309906, "grad_norm": 0.0345620874202541, "learning_rate": 8.425647599697317e-05, "epoch": 0.2822171360334362, "step": 709}, {"loss": 0.43604689836502075, "grad_norm": 0.035320747578666045, "learning_rate": 8.420949631668929e-05, "epoch": 0.28261518559060606, "step": 710}, {"loss": 0.41823774576187134, "grad_norm": 0.03706688267151927, "learning_rate": 8.416245978577213e-05, "epoch": 0.2830132351477759, "step": 711}, {"loss": 0.35273444652557373, "grad_norm": 0.03661869918885983, "learning_rate": 8.411536648238878e-05, "epoch": 0.28341128470494575, "step": 712}, {"loss": 0.4033399224281311, "grad_norm": 0.03754917660179606, "learning_rate": 8.406821648480066e-05, "epoch": 0.28380933426211563, "step": 713}, {"loss": 0.38567915558815, "grad_norm": 0.03333773498993702, "learning_rate": 8.402100987136344e-05, "epoch": 0.2842073838192855, "step": 714}, {"loss": 0.42924565076828003, "grad_norm": 0.03905861437931671, "learning_rate": 8.397374672052683e-05, "epoch": 0.2846054333764554, "step": 715}, {"loss": 0.3862330913543701, "grad_norm": 0.03676739434855674, "learning_rate": 8.392642711083454e-05, "epoch": 0.28500348293362526, "step": 716}, {"loss": 0.42254209518432617, "grad_norm": 0.03985614210670228, "learning_rate": 8.387905112092406e-05, "epoch": 0.2854015324907951, "step": 717}, {"loss": 0.4101502001285553, "grad_norm": 0.03909147164007186, "learning_rate": 8.383161882952663e-05, "epoch": 0.28579958204796496, "step": 718}, {"loss": 0.41978561878204346, "grad_norm": 0.03949571920959807, "learning_rate": 8.378413031546702e-05, "epoch": 0.28619763160513484, "step": 719}, {"loss": 0.37958356738090515, "grad_norm": 0.034991220821410624, "learning_rate": 8.37365856576634e-05, "epoch": 0.2865956811623047, "step": 720}, {"loss": 0.31989091634750366, "grad_norm": 0.036021732154875266, "learning_rate": 8.368898493512732e-05, "epoch": 0.2869937307194746, "step": 721}, {"loss": 0.3587903380393982, "grad_norm": 0.03494572619804994, "learning_rate": 8.364132822696343e-05, "epoch": 0.2873917802766444, "step": 722}, {"loss": 0.40202128887176514, "grad_norm": 0.04014835386860016, "learning_rate": 8.359361561236944e-05, "epoch": 0.2877898298338143, "step": 723}, {"loss": 0.413565456867218, "grad_norm": 0.040083789411154744, "learning_rate": 8.354584717063602e-05, "epoch": 0.28818787939098417, "step": 724}, {"loss": 0.411207914352417, "grad_norm": 0.036692953832957335, "learning_rate": 8.349802298114655e-05, "epoch": 0.28858592894815405, "step": 725}, {"loss": 0.3785688281059265, "grad_norm": 0.03877610744929285, "learning_rate": 8.345014312337706e-05, "epoch": 0.2889839785053239, "step": 726}, {"loss": 0.40101224184036255, "grad_norm": 0.03857208630712701, "learning_rate": 8.340220767689613e-05, "epoch": 0.2893820280624938, "step": 727}, {"loss": 0.37130972743034363, "grad_norm": 0.03994598261794033, "learning_rate": 8.335421672136468e-05, "epoch": 0.2897800776196636, "step": 728}, {"loss": 0.40202033519744873, "grad_norm": 0.04022467470656383, "learning_rate": 8.33061703365359e-05, "epoch": 0.2901781271768335, "step": 729}, {"loss": 0.3862348794937134, "grad_norm": 0.03825640646161951, "learning_rate": 8.325806860225507e-05, "epoch": 0.2905761767340034, "step": 730}, {"loss": 0.42868155241012573, "grad_norm": 0.039312449762897474, "learning_rate": 8.320991159845947e-05, "epoch": 0.29097422629117325, "step": 731}, {"loss": 0.3435894846916199, "grad_norm": 0.03545261418672236, "learning_rate": 8.316169940517826e-05, "epoch": 0.29137227584834313, "step": 732}, {"loss": 0.4017844796180725, "grad_norm": 0.03722584602646851, "learning_rate": 8.311343210253226e-05, "epoch": 0.291770325405513, "step": 733}, {"loss": 0.37925902009010315, "grad_norm": 0.03626353632044599, "learning_rate": 8.306510977073392e-05, "epoch": 0.29216837496268283, "step": 734}, {"loss": 0.3870329260826111, "grad_norm": 0.04367718826281978, "learning_rate": 8.301673249008708e-05, "epoch": 0.2925664245198527, "step": 735}, {"loss": 0.3951336145401001, "grad_norm": 0.0411988591534935, "learning_rate": 8.296830034098695e-05, "epoch": 0.2929644740770226, "step": 736}, {"loss": 0.4066249132156372, "grad_norm": 0.03741539150817325, "learning_rate": 8.29198134039199e-05, "epoch": 0.29336252363419246, "step": 737}, {"loss": 0.40235430002212524, "grad_norm": 0.03964093676217151, "learning_rate": 8.287127175946336e-05, "epoch": 0.29376057319136234, "step": 738}, {"loss": 0.3668045401573181, "grad_norm": 0.034484362275186926, "learning_rate": 8.282267548828568e-05, "epoch": 0.2941586227485322, "step": 739}, {"loss": 0.34866130352020264, "grad_norm": 0.03742252939838237, "learning_rate": 8.277402467114595e-05, "epoch": 0.29455667230570204, "step": 740}, {"loss": 0.3717847466468811, "grad_norm": 0.03917089211418926, "learning_rate": 8.272531938889396e-05, "epoch": 0.2949547218628719, "step": 741}, {"loss": 0.3418647050857544, "grad_norm": 0.03957530334055801, "learning_rate": 8.267655972246998e-05, "epoch": 0.2953527714200418, "step": 742}, {"loss": 0.40772080421447754, "grad_norm": 0.037362827215163194, "learning_rate": 8.262774575290467e-05, "epoch": 0.29575082097721167, "step": 743}, {"loss": 0.4175499379634857, "grad_norm": 0.03891970387672352, "learning_rate": 8.257887756131892e-05, "epoch": 0.29614887053438155, "step": 744}, {"loss": 0.4064297676086426, "grad_norm": 0.038404043708843406, "learning_rate": 8.252995522892375e-05, "epoch": 0.2965469200915514, "step": 745}, {"loss": 0.36818450689315796, "grad_norm": 0.03695842868309963, "learning_rate": 8.248097883702011e-05, "epoch": 0.29694496964872125, "step": 746}, {"loss": 0.3898788094520569, "grad_norm": 0.0364451867528145, "learning_rate": 8.243194846699887e-05, "epoch": 0.2973430192058911, "step": 747}, {"loss": 0.3393363952636719, "grad_norm": 0.03847350708901057, "learning_rate": 8.238286420034049e-05, "epoch": 0.297741068763061, "step": 748}, {"loss": 0.4068445563316345, "grad_norm": 0.03883474186015172, "learning_rate": 8.233372611861512e-05, "epoch": 0.2981391183202309, "step": 749}, {"loss": 0.37084144353866577, "grad_norm": 0.036235395845169674, "learning_rate": 8.228453430348222e-05, "epoch": 0.29853716787740076, "step": 750}, {"loss": 0.36967647075653076, "grad_norm": 0.036883481251537724, "learning_rate": 8.223528883669065e-05, "epoch": 0.2989352174345706, "step": 751}, {"loss": 0.39964574575424194, "grad_norm": 0.039130744643138794, "learning_rate": 8.218598980007837e-05, "epoch": 0.29933326699174045, "step": 752}, {"loss": 0.37190982699394226, "grad_norm": 0.03456015795286077, "learning_rate": 8.213663727557237e-05, "epoch": 0.29973131654891033, "step": 753}, {"loss": 0.3522270917892456, "grad_norm": 0.035883178642004895, "learning_rate": 8.208723134518859e-05, "epoch": 0.3001293661060802, "step": 754}, {"loss": 0.3662225008010864, "grad_norm": 0.038208675942233115, "learning_rate": 8.203777209103161e-05, "epoch": 0.3005274156632501, "step": 755}, {"loss": 0.3932841718196869, "grad_norm": 0.038706915321592046, "learning_rate": 8.198825959529473e-05, "epoch": 0.30092546522041996, "step": 756}, {"loss": 0.3310996890068054, "grad_norm": 0.03621302485555271, "learning_rate": 8.193869394025964e-05, "epoch": 0.3013235147775898, "step": 757}, {"loss": 0.43397754430770874, "grad_norm": 0.038012655571358446, "learning_rate": 8.188907520829647e-05, "epoch": 0.30172156433475966, "step": 758}, {"loss": 0.4045506417751312, "grad_norm": 0.0352982011465029, "learning_rate": 8.183940348186347e-05, "epoch": 0.30211961389192954, "step": 759}, {"loss": 0.42623281478881836, "grad_norm": 0.03800199704987554, "learning_rate": 8.178967884350699e-05, "epoch": 0.3025176634490994, "step": 760}, {"loss": 0.33651095628738403, "grad_norm": 0.03281367334427346, "learning_rate": 8.173990137586133e-05, "epoch": 0.3029157130062693, "step": 761}, {"loss": 0.4493618309497833, "grad_norm": 0.038872331693962266, "learning_rate": 8.169007116164857e-05, "epoch": 0.30331376256343917, "step": 762}, {"loss": 0.3922487199306488, "grad_norm": 0.03915822360513847, "learning_rate": 8.164018828367843e-05, "epoch": 0.303711812120609, "step": 763}, {"loss": 0.3671058714389801, "grad_norm": 0.03842421352348639, "learning_rate": 8.159025282484816e-05, "epoch": 0.30410986167777887, "step": 764}, {"loss": 0.4373021423816681, "grad_norm": 0.05072375667671702, "learning_rate": 8.154026486814239e-05, "epoch": 0.30450791123494875, "step": 765}, {"loss": 0.40010905265808105, "grad_norm": 0.038452103748122586, "learning_rate": 8.1490224496633e-05, "epoch": 0.3049059607921186, "step": 766}, {"loss": 0.32358425855636597, "grad_norm": 0.03637640413416323, "learning_rate": 8.144013179347898e-05, "epoch": 0.3053040103492885, "step": 767}, {"loss": 0.4424345791339874, "grad_norm": 0.038866756414177694, "learning_rate": 8.138998684192625e-05, "epoch": 0.3057020599064584, "step": 768}, {"loss": 0.3686906099319458, "grad_norm": 0.0356017791559051, "learning_rate": 8.133978972530761e-05, "epoch": 0.3061001094636282, "step": 769}, {"loss": 0.35894230008125305, "grad_norm": 0.03837977505155052, "learning_rate": 8.128954052704252e-05, "epoch": 0.3064981590207981, "step": 770}, {"loss": 0.39787620306015015, "grad_norm": 0.042815052988225334, "learning_rate": 8.123923933063697e-05, "epoch": 0.30689620857796795, "step": 771}, {"loss": 0.3879792094230652, "grad_norm": 0.03942709478179493, "learning_rate": 8.118888621968342e-05, "epoch": 0.30729425813513783, "step": 772}, {"loss": 0.4290284514427185, "grad_norm": 0.038984771004563025, "learning_rate": 8.113848127786056e-05, "epoch": 0.3076923076923077, "step": 773}, {"loss": 0.35802000761032104, "grad_norm": 0.035958754874988764, "learning_rate": 8.108802458893323e-05, "epoch": 0.3080903572494776, "step": 774}, {"loss": 0.3429155647754669, "grad_norm": 0.037635799793040145, "learning_rate": 8.103751623675227e-05, "epoch": 0.3084884068066474, "step": 775}, {"loss": 0.35383379459381104, "grad_norm": 0.036623432938404574, "learning_rate": 8.098695630525433e-05, "epoch": 0.3088864563638173, "step": 776}, {"loss": 0.4033557176589966, "grad_norm": 0.0393249667693365, "learning_rate": 8.093634487846186e-05, "epoch": 0.30928450592098716, "step": 777}, {"loss": 0.4063947796821594, "grad_norm": 0.040631623108410654, "learning_rate": 8.088568204048283e-05, "epoch": 0.30968255547815704, "step": 778}, {"loss": 0.408205509185791, "grad_norm": 0.039508013031920065, "learning_rate": 8.083496787551063e-05, "epoch": 0.3100806050353269, "step": 779}, {"loss": 0.37763506174087524, "grad_norm": 0.0384715549404356, "learning_rate": 8.0784202467824e-05, "epoch": 0.31047865459249674, "step": 780}, {"loss": 0.36905431747436523, "grad_norm": 0.03848230307804343, "learning_rate": 8.073338590178682e-05, "epoch": 0.3108767041496666, "step": 781}, {"loss": 0.38688647747039795, "grad_norm": 0.038549133158833314, "learning_rate": 8.068251826184794e-05, "epoch": 0.3112747537068365, "step": 782}, {"loss": 0.4038981795310974, "grad_norm": 0.03648758191847315, "learning_rate": 8.063159963254118e-05, "epoch": 0.31167280326400637, "step": 783}, {"loss": 0.391489714384079, "grad_norm": 0.03787341628706013, "learning_rate": 8.0580630098485e-05, "epoch": 0.31207085282117625, "step": 784}, {"loss": 0.4006823003292084, "grad_norm": 0.038459682309498616, "learning_rate": 8.052960974438251e-05, "epoch": 0.3124689023783461, "step": 785}, {"loss": 0.3888184130191803, "grad_norm": 0.040921174246972954, "learning_rate": 8.047853865502126e-05, "epoch": 0.31286695193551595, "step": 786}, {"loss": 0.43073025345802307, "grad_norm": 0.03700325276903266, "learning_rate": 8.042741691527313e-05, "epoch": 0.3132650014926858, "step": 787}, {"loss": 0.3789212703704834, "grad_norm": 0.0397276435390164, "learning_rate": 8.037624461009415e-05, "epoch": 0.3136630510498557, "step": 788}, {"loss": 0.41667813062667847, "grad_norm": 0.0395357196368601, "learning_rate": 8.032502182452437e-05, "epoch": 0.3140611006070256, "step": 789}, {"loss": 0.3885061740875244, "grad_norm": 0.03531028192244049, "learning_rate": 8.027374864368775e-05, "epoch": 0.31445915016419546, "step": 790}, {"loss": 0.39348304271698, "grad_norm": 0.036817458654016595, "learning_rate": 8.022242515279203e-05, "epoch": 0.31485719972136533, "step": 791}, {"loss": 0.41172081232070923, "grad_norm": 0.03976988587269681, "learning_rate": 8.017105143712853e-05, "epoch": 0.31525524927853515, "step": 792}, {"loss": 0.41270560026168823, "grad_norm": 0.03716264662999423, "learning_rate": 8.011962758207199e-05, "epoch": 0.31565329883570503, "step": 793}, {"loss": 0.37577024102211, "grad_norm": 0.040427225194676154, "learning_rate": 8.006815367308056e-05, "epoch": 0.3160513483928749, "step": 794}, {"loss": 0.37990063428878784, "grad_norm": 0.038813996233788214, "learning_rate": 8.001662979569548e-05, "epoch": 0.3164493979500448, "step": 795}, {"loss": 0.401939332485199, "grad_norm": 0.03923739530072502, "learning_rate": 7.99650560355411e-05, "epoch": 0.31684744750721466, "step": 796}, {"loss": 0.34289100766181946, "grad_norm": 0.03778835145668313, "learning_rate": 7.991343247832466e-05, "epoch": 0.31724549706438454, "step": 797}, {"loss": 0.36352071166038513, "grad_norm": 0.036688729103932546, "learning_rate": 7.986175920983611e-05, "epoch": 0.31764354662155436, "step": 798}, {"loss": 0.45587265491485596, "grad_norm": 0.0413240972611931, "learning_rate": 7.981003631594806e-05, "epoch": 0.31804159617872424, "step": 799}, {"loss": 0.36033323407173157, "grad_norm": 0.04178113305553904, "learning_rate": 7.975826388261554e-05, "epoch": 0.3184396457358941, "step": 800}, {"loss": 0.40316498279571533, "grad_norm": 0.042429050050048364, "learning_rate": 7.970644199587596e-05, "epoch": 0.318837695293064, "step": 801}, {"loss": 0.3731704354286194, "grad_norm": 0.040793164471786506, "learning_rate": 7.965457074184888e-05, "epoch": 0.31923574485023387, "step": 802}, {"loss": 0.35714197158813477, "grad_norm": 0.039418421420600855, "learning_rate": 7.960265020673592e-05, "epoch": 0.31963379440740375, "step": 803}, {"loss": 0.37631887197494507, "grad_norm": 0.03901566869773436, "learning_rate": 7.955068047682057e-05, "epoch": 0.32003184396457357, "step": 804}, {"loss": 0.37987419962882996, "grad_norm": 0.035012673795910325, "learning_rate": 7.949866163846811e-05, "epoch": 0.32042989352174345, "step": 805}, {"loss": 0.4209093749523163, "grad_norm": 0.039846851325902155, "learning_rate": 7.94465937781254e-05, "epoch": 0.3208279430789133, "step": 806}, {"loss": 0.3506585359573364, "grad_norm": 0.04165681671559149, "learning_rate": 7.939447698232077e-05, "epoch": 0.3212259926360832, "step": 807}, {"loss": 0.3692915141582489, "grad_norm": 0.03365289871488559, "learning_rate": 7.934231133766387e-05, "epoch": 0.3216240421932531, "step": 808}, {"loss": 0.3621044456958771, "grad_norm": 0.03756483200143065, "learning_rate": 7.929009693084559e-05, "epoch": 0.3220220917504229, "step": 809}, {"loss": 0.3649030923843384, "grad_norm": 0.03876079202995229, "learning_rate": 7.923783384863775e-05, "epoch": 0.3224201413075928, "step": 810}, {"loss": 0.3851947784423828, "grad_norm": 0.036825701723663486, "learning_rate": 7.918552217789314e-05, "epoch": 0.32281819086476266, "step": 811}, {"loss": 0.37716659903526306, "grad_norm": 0.036850348167822136, "learning_rate": 7.913316200554525e-05, "epoch": 0.32321624042193253, "step": 812}, {"loss": 0.36162570118904114, "grad_norm": 0.034502392945957444, "learning_rate": 7.908075341860823e-05, "epoch": 0.3236142899791024, "step": 813}, {"loss": 0.35547539591789246, "grad_norm": 0.03411849595979235, "learning_rate": 7.902829650417662e-05, "epoch": 0.3240123395362723, "step": 814}, {"loss": 0.38767582178115845, "grad_norm": 0.03877639624276806, "learning_rate": 7.89757913494253e-05, "epoch": 0.3244103890934421, "step": 815}, {"loss": 0.350908100605011, "grad_norm": 0.036155643742378674, "learning_rate": 7.892323804160932e-05, "epoch": 0.324808438650612, "step": 816}, {"loss": 0.3840307593345642, "grad_norm": 0.04262886509063164, "learning_rate": 7.887063666806378e-05, "epoch": 0.32520648820778186, "step": 817}, {"loss": 0.38940754532814026, "grad_norm": 0.03902118988508853, "learning_rate": 7.881798731620361e-05, "epoch": 0.32560453776495174, "step": 818}, {"loss": 0.3810456693172455, "grad_norm": 0.03814755174018834, "learning_rate": 7.876529007352348e-05, "epoch": 0.3260025873221216, "step": 819}, {"loss": 0.45925599336624146, "grad_norm": 0.041355914322052774, "learning_rate": 7.871254502759767e-05, "epoch": 0.3264006368792915, "step": 820}, {"loss": 0.3773195147514343, "grad_norm": 0.03688401491640676, "learning_rate": 7.86597522660799e-05, "epoch": 0.3267986864364613, "step": 821}, {"loss": 0.3715212047100067, "grad_norm": 0.036862706326631724, "learning_rate": 7.860691187670315e-05, "epoch": 0.3271967359936312, "step": 822}, {"loss": 0.41119152307510376, "grad_norm": 0.04200466675433269, "learning_rate": 7.855402394727959e-05, "epoch": 0.32759478555080107, "step": 823}, {"loss": 0.37336018681526184, "grad_norm": 0.036852857110268825, "learning_rate": 7.85010885657004e-05, "epoch": 0.32799283510797095, "step": 824}, {"loss": 0.39030003547668457, "grad_norm": 0.03846774089790718, "learning_rate": 7.844810581993557e-05, "epoch": 0.3283908846651408, "step": 825}, {"loss": 0.37967491149902344, "grad_norm": 0.035530299840471916, "learning_rate": 7.839507579803382e-05, "epoch": 0.3287889342223107, "step": 826}, {"loss": 0.3854385018348694, "grad_norm": 0.03697950650722005, "learning_rate": 7.834199858812248e-05, "epoch": 0.3291869837794805, "step": 827}, {"loss": 0.39857015013694763, "grad_norm": 0.03818137204844395, "learning_rate": 7.828887427840723e-05, "epoch": 0.3295850333366504, "step": 828}, {"loss": 0.37179261445999146, "grad_norm": 0.03797708996383726, "learning_rate": 7.823570295717206e-05, "epoch": 0.3299830828938203, "step": 829}, {"loss": 0.37895506620407104, "grad_norm": 0.03993438920820619, "learning_rate": 7.818248471277909e-05, "epoch": 0.33038113245099016, "step": 830}, {"loss": 0.43075042963027954, "grad_norm": 0.03772867075747853, "learning_rate": 7.812921963366841e-05, "epoch": 0.33077918200816003, "step": 831}, {"loss": 0.41675153374671936, "grad_norm": 0.040865617178179325, "learning_rate": 7.807590780835795e-05, "epoch": 0.3311772315653299, "step": 832}, {"loss": 0.38082051277160645, "grad_norm": 0.03694739726580126, "learning_rate": 7.802254932544327e-05, "epoch": 0.33157528112249973, "step": 833}, {"loss": 0.3585208058357239, "grad_norm": 0.03779419739207531, "learning_rate": 7.796914427359754e-05, "epoch": 0.3319733306796696, "step": 834}, {"loss": 0.33808961510658264, "grad_norm": 0.036499562298800255, "learning_rate": 7.791569274157132e-05, "epoch": 0.3323713802368395, "step": 835}, {"loss": 0.4104965329170227, "grad_norm": 0.04058657154993314, "learning_rate": 7.786219481819235e-05, "epoch": 0.33276942979400936, "step": 836}, {"loss": 0.3816071152687073, "grad_norm": 0.04103353464762891, "learning_rate": 7.78086505923655e-05, "epoch": 0.33316747935117924, "step": 837}, {"loss": 0.4404466152191162, "grad_norm": 0.03967549156344844, "learning_rate": 7.775506015307262e-05, "epoch": 0.33356552890834906, "step": 838}, {"loss": 0.36082491278648376, "grad_norm": 0.038256395030733165, "learning_rate": 7.770142358937228e-05, "epoch": 0.33396357846551894, "step": 839}, {"loss": 0.32890093326568604, "grad_norm": 0.03689550486715946, "learning_rate": 7.76477409903998e-05, "epoch": 0.3343616280226888, "step": 840}, {"loss": 0.38580718636512756, "grad_norm": 0.041538291367577126, "learning_rate": 7.759401244536691e-05, "epoch": 0.3347596775798587, "step": 841}, {"loss": 0.38952314853668213, "grad_norm": 0.039973281651351625, "learning_rate": 7.754023804356176e-05, "epoch": 0.33515772713702857, "step": 842}, {"loss": 0.3573082983493805, "grad_norm": 0.036896044108688755, "learning_rate": 7.748641787434865e-05, "epoch": 0.33555577669419845, "step": 843}, {"loss": 0.37234917283058167, "grad_norm": 0.037808490101253714, "learning_rate": 7.7432552027168e-05, "epoch": 0.33595382625136827, "step": 844}, {"loss": 0.4266722798347473, "grad_norm": 0.045156000312949716, "learning_rate": 7.73786405915361e-05, "epoch": 0.33635187580853815, "step": 845}, {"loss": 0.432503342628479, "grad_norm": 0.03775198736300346, "learning_rate": 7.732468365704502e-05, "epoch": 0.336749925365708, "step": 846}, {"loss": 0.3796793818473816, "grad_norm": 0.03737087687940775, "learning_rate": 7.727068131336241e-05, "epoch": 0.3371479749228779, "step": 847}, {"loss": 0.37069594860076904, "grad_norm": 0.03673558540243252, "learning_rate": 7.721663365023142e-05, "epoch": 0.3375460244800478, "step": 848}, {"loss": 0.3729356527328491, "grad_norm": 0.04102129853912705, "learning_rate": 7.71625407574705e-05, "epoch": 0.33794407403721766, "step": 849}, {"loss": 0.40777352452278137, "grad_norm": 0.04338142835345861, "learning_rate": 7.710840272497326e-05, "epoch": 0.3383421235943875, "step": 850}, {"loss": 0.4113582968711853, "grad_norm": 0.040917649453411205, "learning_rate": 7.705421964270834e-05, "epoch": 0.33874017315155736, "step": 851}, {"loss": 0.4037177562713623, "grad_norm": 0.03801094809034697, "learning_rate": 7.69999916007192e-05, "epoch": 0.33913822270872723, "step": 852}, {"loss": 0.40404897928237915, "grad_norm": 0.036594292455672243, "learning_rate": 7.694571868912408e-05, "epoch": 0.3395362722658971, "step": 853}, {"loss": 0.37208741903305054, "grad_norm": 0.03978645256905779, "learning_rate": 7.689140099811574e-05, "epoch": 0.339934321823067, "step": 854}, {"loss": 0.41594141721725464, "grad_norm": 0.039779274724550735, "learning_rate": 7.683703861796136e-05, "epoch": 0.34033237138023686, "step": 855}, {"loss": 0.39853957295417786, "grad_norm": 0.0389690937796472, "learning_rate": 7.678263163900243e-05, "epoch": 0.3407304209374067, "step": 856}, {"loss": 0.37848955392837524, "grad_norm": 0.03740137092627384, "learning_rate": 7.672818015165449e-05, "epoch": 0.34112847049457656, "step": 857}, {"loss": 0.38843658566474915, "grad_norm": 0.0388666354802609, "learning_rate": 7.66736842464071e-05, "epoch": 0.34152652005174644, "step": 858}, {"loss": 0.38097715377807617, "grad_norm": 0.034553760299598495, "learning_rate": 7.661914401382359e-05, "epoch": 0.3419245696089163, "step": 859}, {"loss": 0.3532256782054901, "grad_norm": 0.03473708407093839, "learning_rate": 7.6564559544541e-05, "epoch": 0.3423226191660862, "step": 860}, {"loss": 0.39355796575546265, "grad_norm": 0.03622199640095985, "learning_rate": 7.650993092926988e-05, "epoch": 0.34272066872325607, "step": 861}, {"loss": 0.38389623165130615, "grad_norm": 0.03840085274497457, "learning_rate": 7.64552582587941e-05, "epoch": 0.3431187182804259, "step": 862}, {"loss": 0.312508225440979, "grad_norm": 0.03214924330041964, "learning_rate": 7.640054162397074e-05, "epoch": 0.34351676783759577, "step": 863}, {"loss": 0.39971405267715454, "grad_norm": 0.035374712968954426, "learning_rate": 7.634578111573006e-05, "epoch": 0.34391481739476565, "step": 864}, {"loss": 0.3838176429271698, "grad_norm": 0.03700932731344093, "learning_rate": 7.629097682507503e-05, "epoch": 0.3443128669519355, "step": 865}, {"loss": 0.40065112709999084, "grad_norm": 0.03646942986061808, "learning_rate": 7.623612884308158e-05, "epoch": 0.3447109165091054, "step": 866}, {"loss": 0.3788797855377197, "grad_norm": 0.03927763179583504, "learning_rate": 7.61812372608981e-05, "epoch": 0.3451089660662752, "step": 867}, {"loss": 0.35138291120529175, "grad_norm": 0.039033279862850774, "learning_rate": 7.612630216974553e-05, "epoch": 0.3455070156234451, "step": 868}, {"loss": 0.40744274854660034, "grad_norm": 0.03821611674665472, "learning_rate": 7.607132366091709e-05, "epoch": 0.345905065180615, "step": 869}, {"loss": 0.3850864768028259, "grad_norm": 0.0361767046200033, "learning_rate": 7.601630182577809e-05, "epoch": 0.34630311473778486, "step": 870}, {"loss": 0.3871955871582031, "grad_norm": 0.03674430740288857, "learning_rate": 7.596123675576594e-05, "epoch": 0.34670116429495473, "step": 871}, {"loss": 0.36946743726730347, "grad_norm": 0.03750655223436605, "learning_rate": 7.590612854238983e-05, "epoch": 0.3470992138521246, "step": 872}, {"loss": 0.37711191177368164, "grad_norm": 0.0350763395792127, "learning_rate": 7.58509772772307e-05, "epoch": 0.34749726340929443, "step": 873}, {"loss": 0.3776005208492279, "grad_norm": 0.03891025466848445, "learning_rate": 7.579578305194097e-05, "epoch": 0.3478953129664643, "step": 874}, {"loss": 0.4556564390659332, "grad_norm": 0.04091699806576619, "learning_rate": 7.574054595824452e-05, "epoch": 0.3482933625236342, "step": 875}, {"loss": 0.4257289171218872, "grad_norm": 0.040828959249403725, "learning_rate": 7.56852660879364e-05, "epoch": 0.34869141208080406, "step": 876}, {"loss": 0.4092128574848175, "grad_norm": 0.04092671659897629, "learning_rate": 7.562994353288282e-05, "epoch": 0.34908946163797394, "step": 877}, {"loss": 0.4033675193786621, "grad_norm": 0.046173866250233585, "learning_rate": 7.557457838502089e-05, "epoch": 0.3494875111951438, "step": 878}, {"loss": 0.3832464814186096, "grad_norm": 0.03994397516776993, "learning_rate": 7.551917073635849e-05, "epoch": 0.34988556075231364, "step": 879}, {"loss": 0.3835020661354065, "grad_norm": 0.03736773552018162, "learning_rate": 7.546372067897416e-05, "epoch": 0.3502836103094835, "step": 880}, {"loss": 0.3876296877861023, "grad_norm": 0.03676004962903278, "learning_rate": 7.540822830501688e-05, "epoch": 0.3506816598666534, "step": 881}, {"loss": 0.3861679434776306, "grad_norm": 0.038048220620864735, "learning_rate": 7.535269370670598e-05, "epoch": 0.35107970942382327, "step": 882}, {"loss": 0.385206937789917, "grad_norm": 0.036225133106743135, "learning_rate": 7.529711697633097e-05, "epoch": 0.35147775898099315, "step": 883}, {"loss": 0.3487272262573242, "grad_norm": 0.034668208424429005, "learning_rate": 7.524149820625133e-05, "epoch": 0.351875808538163, "step": 884}, {"loss": 0.4048854112625122, "grad_norm": 0.039316295569050974, "learning_rate": 7.518583748889646e-05, "epoch": 0.35227385809533285, "step": 885}, {"loss": 0.37199893593788147, "grad_norm": 0.03879181553072163, "learning_rate": 7.513013491676544e-05, "epoch": 0.3526719076525027, "step": 886}, {"loss": 0.395846426486969, "grad_norm": 0.038184587451014114, "learning_rate": 7.50743905824269e-05, "epoch": 0.3530699572096726, "step": 887}, {"loss": 0.33084017038345337, "grad_norm": 0.03564367540741797, "learning_rate": 7.501860457851886e-05, "epoch": 0.3534680067668425, "step": 888}, {"loss": 0.4429159164428711, "grad_norm": 0.04019806055249296, "learning_rate": 7.496277699774863e-05, "epoch": 0.35386605632401236, "step": 889}, {"loss": 0.36531907320022583, "grad_norm": 0.03671797163792129, "learning_rate": 7.490690793289261e-05, "epoch": 0.35426410588118223, "step": 890}, {"loss": 0.36903488636016846, "grad_norm": 0.03758556305313903, "learning_rate": 7.48509974767961e-05, "epoch": 0.35466215543835206, "step": 891}, {"loss": 0.3694780766963959, "grad_norm": 0.03336416997169766, "learning_rate": 7.479504572237321e-05, "epoch": 0.35506020499552193, "step": 892}, {"loss": 0.3765166997909546, "grad_norm": 0.045334077384937686, "learning_rate": 7.473905276260669e-05, "epoch": 0.3554582545526918, "step": 893}, {"loss": 0.38363227248191833, "grad_norm": 0.033826751973785756, "learning_rate": 7.468301869054776e-05, "epoch": 0.3558563041098617, "step": 894}, {"loss": 0.3790667653083801, "grad_norm": 0.038852647470653104, "learning_rate": 7.462694359931593e-05, "epoch": 0.35625435366703156, "step": 895}, {"loss": 0.3474653959274292, "grad_norm": 0.03593923305130126, "learning_rate": 7.457082758209892e-05, "epoch": 0.3566524032242014, "step": 896}, {"loss": 0.3932214379310608, "grad_norm": 0.03644192016439116, "learning_rate": 7.451467073215245e-05, "epoch": 0.35705045278137126, "step": 897}, {"loss": 0.3368533253669739, "grad_norm": 0.03276666785802477, "learning_rate": 7.44584731428001e-05, "epoch": 0.35744850233854114, "step": 898}, {"loss": 0.3760644197463989, "grad_norm": 0.038125049699437016, "learning_rate": 7.440223490743312e-05, "epoch": 0.357846551895711, "step": 899}, {"loss": 0.3639260232448578, "grad_norm": 0.03463995876991518, "learning_rate": 7.434595611951037e-05, "epoch": 0.3582446014528809, "step": 900}, {"loss": 0.3613651692867279, "grad_norm": 0.03409097575123766, "learning_rate": 7.428963687255805e-05, "epoch": 0.3586426510100508, "step": 901}, {"loss": 0.4074559211730957, "grad_norm": 0.037627892112266026, "learning_rate": 7.42332772601696e-05, "epoch": 0.3590407005672206, "step": 902}, {"loss": 0.38606926798820496, "grad_norm": 0.03687134480827929, "learning_rate": 7.417687737600558e-05, "epoch": 0.35943875012439047, "step": 903}, {"loss": 0.39273688197135925, "grad_norm": 0.03592106952347583, "learning_rate": 7.412043731379343e-05, "epoch": 0.35983679968156035, "step": 904}, {"loss": 0.3452506959438324, "grad_norm": 0.03862758792555347, "learning_rate": 7.406395716732737e-05, "epoch": 0.3602348492387302, "step": 905}, {"loss": 0.40722784399986267, "grad_norm": 0.0400131083010212, "learning_rate": 7.400743703046826e-05, "epoch": 0.3606328987959001, "step": 906}, {"loss": 0.40479040145874023, "grad_norm": 0.04036227500275947, "learning_rate": 7.395087699714339e-05, "epoch": 0.36103094835307, "step": 907}, {"loss": 0.3996145725250244, "grad_norm": 0.036618670254471554, "learning_rate": 7.389427716134634e-05, "epoch": 0.3614289979102398, "step": 908}, {"loss": 0.45012933015823364, "grad_norm": 0.04301020871008294, "learning_rate": 7.38376376171369e-05, "epoch": 0.3618270474674097, "step": 909}, {"loss": 0.3850652575492859, "grad_norm": 0.03582937935071459, "learning_rate": 7.378095845864077e-05, "epoch": 0.36222509702457956, "step": 910}, {"loss": 0.38369613885879517, "grad_norm": 0.04281360983477801, "learning_rate": 7.372423978004952e-05, "epoch": 0.36262314658174943, "step": 911}, {"loss": 0.35413825511932373, "grad_norm": 0.03640937556578236, "learning_rate": 7.36674816756204e-05, "epoch": 0.3630211961389193, "step": 912}, {"loss": 0.37294960021972656, "grad_norm": 0.034583485062499496, "learning_rate": 7.361068423967616e-05, "epoch": 0.3634192456960892, "step": 913}, {"loss": 0.43280845880508423, "grad_norm": 0.03715080531855591, "learning_rate": 7.355384756660496e-05, "epoch": 0.363817295253259, "step": 914}, {"loss": 0.43558111786842346, "grad_norm": 0.04186750299416689, "learning_rate": 7.349697175086009e-05, "epoch": 0.3642153448104289, "step": 915}, {"loss": 0.3999122679233551, "grad_norm": 0.04160348560889167, "learning_rate": 7.344005688695996e-05, "epoch": 0.36461339436759876, "step": 916}, {"loss": 0.3833656907081604, "grad_norm": 0.03914576181387888, "learning_rate": 7.338310306948781e-05, "epoch": 0.36501144392476864, "step": 917}, {"loss": 0.420352041721344, "grad_norm": 0.03768741662088628, "learning_rate": 7.332611039309166e-05, "epoch": 0.3654094934819385, "step": 918}, {"loss": 0.40368515253067017, "grad_norm": 0.03736845610529472, "learning_rate": 7.32690789524841e-05, "epoch": 0.3658075430391084, "step": 919}, {"loss": 0.38748666644096375, "grad_norm": 0.03442677426790434, "learning_rate": 7.321200884244212e-05, "epoch": 0.3662055925962782, "step": 920}, {"loss": 0.3803667426109314, "grad_norm": 0.03756398304163133, "learning_rate": 7.315490015780699e-05, "epoch": 0.3666036421534481, "step": 921}, {"loss": 0.41424477100372314, "grad_norm": 0.03837286729303145, "learning_rate": 7.309775299348408e-05, "epoch": 0.36700169171061797, "step": 922}, {"loss": 0.3452605605125427, "grad_norm": 0.039940349685135226, "learning_rate": 7.30405674444427e-05, "epoch": 0.36739974126778785, "step": 923}, {"loss": 0.34458127617836, "grad_norm": 0.035018339487482106, "learning_rate": 7.298334360571599e-05, "epoch": 0.3677977908249577, "step": 924}, {"loss": 0.39862897992134094, "grad_norm": 0.03901135647808901, "learning_rate": 7.292608157240063e-05, "epoch": 0.36819584038212755, "step": 925}, {"loss": 0.37571847438812256, "grad_norm": 0.03585529169640429, "learning_rate": 7.286878143965689e-05, "epoch": 0.3685938899392974, "step": 926}, {"loss": 0.36580124497413635, "grad_norm": 0.0341914265422859, "learning_rate": 7.281144330270824e-05, "epoch": 0.3689919394964673, "step": 927}, {"loss": 0.32010889053344727, "grad_norm": 0.03543742206996469, "learning_rate": 7.275406725684137e-05, "epoch": 0.3693899890536372, "step": 928}, {"loss": 0.3986983895301819, "grad_norm": 0.0397419127788123, "learning_rate": 7.2696653397406e-05, "epoch": 0.36978803861080706, "step": 929}, {"loss": 0.3677726984024048, "grad_norm": 0.03906205785112285, "learning_rate": 7.263920181981461e-05, "epoch": 0.37018608816797693, "step": 930}, {"loss": 0.401081919670105, "grad_norm": 0.03816474554945266, "learning_rate": 7.258171261954245e-05, "epoch": 0.37058413772514676, "step": 931}, {"loss": 0.4217010736465454, "grad_norm": 0.041745328760315586, "learning_rate": 7.25241858921272e-05, "epoch": 0.37098218728231663, "step": 932}, {"loss": 0.3841804265975952, "grad_norm": 0.036307680959528695, "learning_rate": 7.246662173316894e-05, "epoch": 0.3713802368394865, "step": 933}, {"loss": 0.376112163066864, "grad_norm": 0.03948016488392322, "learning_rate": 7.240902023833e-05, "epoch": 0.3717782863966564, "step": 934}, {"loss": 0.40509992837905884, "grad_norm": 0.034663191371453024, "learning_rate": 7.235138150333471e-05, "epoch": 0.37217633595382626, "step": 935}, {"loss": 0.38211220502853394, "grad_norm": 0.03894168480751638, "learning_rate": 7.22937056239693e-05, "epoch": 0.37257438551099614, "step": 936}, {"loss": 0.39059779047966003, "grad_norm": 0.03636175029878986, "learning_rate": 7.223599269608172e-05, "epoch": 0.37297243506816596, "step": 937}, {"loss": 0.3957163095474243, "grad_norm": 0.03563861263634541, "learning_rate": 7.217824281558148e-05, "epoch": 0.37337048462533584, "step": 938}, {"loss": 0.38543635606765747, "grad_norm": 0.038262020736036365, "learning_rate": 7.212045607843953e-05, "epoch": 0.3737685341825057, "step": 939}, {"loss": 0.37814438343048096, "grad_norm": 0.038842837425696254, "learning_rate": 7.206263258068804e-05, "epoch": 0.3741665837396756, "step": 940}, {"loss": 0.3992766737937927, "grad_norm": 0.03769215832828293, "learning_rate": 7.200477241842029e-05, "epoch": 0.3745646332968455, "step": 941}, {"loss": 0.39600464701652527, "grad_norm": 0.03688247299764881, "learning_rate": 7.19468756877905e-05, "epoch": 0.37496268285401535, "step": 942}, {"loss": 0.3644353747367859, "grad_norm": 0.03425132264679234, "learning_rate": 7.188894248501363e-05, "epoch": 0.37536073241118517, "step": 943}, {"loss": 0.36200106143951416, "grad_norm": 0.035082412432783795, "learning_rate": 7.183097290636525e-05, "epoch": 0.37575878196835505, "step": 944}, {"loss": 0.4113442301750183, "grad_norm": 0.036815772350948316, "learning_rate": 7.177296704818142e-05, "epoch": 0.3761568315255249, "step": 945}, {"loss": 0.3457128405570984, "grad_norm": 0.03714686121249891, "learning_rate": 7.171492500685843e-05, "epoch": 0.3765548810826948, "step": 946}, {"loss": 0.3696892261505127, "grad_norm": 0.036290591281991545, "learning_rate": 7.165684687885279e-05, "epoch": 0.3769529306398647, "step": 947}, {"loss": 0.3603419065475464, "grad_norm": 0.035337858412731814, "learning_rate": 7.15987327606809e-05, "epoch": 0.37735098019703456, "step": 948}, {"loss": 0.3649914562702179, "grad_norm": 0.03734614258923698, "learning_rate": 7.154058274891898e-05, "epoch": 0.3777490297542044, "step": 949}, {"loss": 0.3882737159729004, "grad_norm": 0.03693075012175047, "learning_rate": 7.148239694020293e-05, "epoch": 0.37814707931137426, "step": 950}, {"loss": 0.38592326641082764, "grad_norm": 0.03621793781780523, "learning_rate": 7.142417543122815e-05, "epoch": 0.37854512886854413, "step": 951}, {"loss": 0.38558152318000793, "grad_norm": 0.03601803577156779, "learning_rate": 7.13659183187493e-05, "epoch": 0.378943178425714, "step": 952}, {"loss": 0.4045000970363617, "grad_norm": 0.04196598643533835, "learning_rate": 7.130762569958027e-05, "epoch": 0.3793412279828839, "step": 953}, {"loss": 0.3802696466445923, "grad_norm": 0.03897746518466152, "learning_rate": 7.124929767059395e-05, "epoch": 0.3797392775400537, "step": 954}, {"loss": 0.3695735037326813, "grad_norm": 0.03477454572616283, "learning_rate": 7.119093432872203e-05, "epoch": 0.3801373270972236, "step": 955}, {"loss": 0.3713560700416565, "grad_norm": 0.03789524419055085, "learning_rate": 7.113253577095495e-05, "epoch": 0.38053537665439346, "step": 956}, {"loss": 0.4004422724246979, "grad_norm": 0.03737127291228844, "learning_rate": 7.107410209434162e-05, "epoch": 0.38093342621156334, "step": 957}, {"loss": 0.41522103548049927, "grad_norm": 0.035989806207208985, "learning_rate": 7.101563339598934e-05, "epoch": 0.3813314757687332, "step": 958}, {"loss": 0.400961309671402, "grad_norm": 0.03774058513053316, "learning_rate": 7.09571297730636e-05, "epoch": 0.3817295253259031, "step": 959}, {"loss": 0.3933753967285156, "grad_norm": 0.03564509447004502, "learning_rate": 7.089859132278793e-05, "epoch": 0.3821275748830729, "step": 960}, {"loss": 0.395210862159729, "grad_norm": 0.0351317697422293, "learning_rate": 7.084001814244372e-05, "epoch": 0.3825256244402428, "step": 961}, {"loss": 0.36216098070144653, "grad_norm": 0.03356789693873997, "learning_rate": 7.078141032937012e-05, "epoch": 0.3829236739974127, "step": 962}, {"loss": 0.37822139263153076, "grad_norm": 0.03735407578559667, "learning_rate": 7.07227679809638e-05, "epoch": 0.38332172355458255, "step": 963}, {"loss": 0.350747287273407, "grad_norm": 0.03581619252736567, "learning_rate": 7.066409119467882e-05, "epoch": 0.3837197731117524, "step": 964}, {"loss": 0.38919591903686523, "grad_norm": 0.037812609810630136, "learning_rate": 7.06053800680265e-05, "epoch": 0.3841178226689223, "step": 965}, {"loss": 0.38271933794021606, "grad_norm": 0.03527826366443659, "learning_rate": 7.054663469857519e-05, "epoch": 0.3845158722260921, "step": 966}, {"loss": 0.38837331533432007, "grad_norm": 0.036209683428287544, "learning_rate": 7.048785518395016e-05, "epoch": 0.384913921783262, "step": 967}, {"loss": 0.3933030068874359, "grad_norm": 0.03308777764848952, "learning_rate": 7.042904162183343e-05, "epoch": 0.3853119713404319, "step": 968}, {"loss": 0.35627540946006775, "grad_norm": 0.035450787042520755, "learning_rate": 7.037019410996364e-05, "epoch": 0.38571002089760176, "step": 969}, {"loss": 0.38529136776924133, "grad_norm": 0.040237696320622574, "learning_rate": 7.031131274613575e-05, "epoch": 0.38610807045477163, "step": 970}, {"loss": 0.3560100197792053, "grad_norm": 0.03400807836610172, "learning_rate": 7.025239762820107e-05, "epoch": 0.3865061200119415, "step": 971}, {"loss": 0.3926514685153961, "grad_norm": 0.033317304495565865, "learning_rate": 7.019344885406696e-05, "epoch": 0.38690416956911133, "step": 972}, {"loss": 0.39435601234436035, "grad_norm": 0.05299295486799131, "learning_rate": 7.01344665216967e-05, "epoch": 0.3873022191262812, "step": 973}, {"loss": 0.3943886458873749, "grad_norm": 0.039715743552613605, "learning_rate": 7.007545072910939e-05, "epoch": 0.3877002686834511, "step": 974}, {"loss": 0.3824337124824524, "grad_norm": 0.035579316262043, "learning_rate": 7.001640157437968e-05, "epoch": 0.38809831824062097, "step": 975}, {"loss": 0.3742203116416931, "grad_norm": 0.0385890984229658, "learning_rate": 6.995731915563772e-05, "epoch": 0.38849636779779084, "step": 976}, {"loss": 0.34791243076324463, "grad_norm": 0.03510102241736683, "learning_rate": 6.989820357106886e-05, "epoch": 0.3888944173549607, "step": 977}, {"loss": 0.3513430655002594, "grad_norm": 0.035738942370441514, "learning_rate": 6.983905491891362e-05, "epoch": 0.38929246691213054, "step": 978}, {"loss": 0.3899417519569397, "grad_norm": 0.03673694871395556, "learning_rate": 6.977987329746749e-05, "epoch": 0.3896905164693004, "step": 979}, {"loss": 0.3628859519958496, "grad_norm": 0.038253175280249685, "learning_rate": 6.972065880508072e-05, "epoch": 0.3900885660264703, "step": 980}, {"loss": 0.3608624339103699, "grad_norm": 0.03661552993708878, "learning_rate": 6.966141154015818e-05, "epoch": 0.3904866155836402, "step": 981}, {"loss": 0.3788892924785614, "grad_norm": 0.03842705399265955, "learning_rate": 6.960213160115921e-05, "epoch": 0.39088466514081005, "step": 982}, {"loss": 0.3690430223941803, "grad_norm": 0.035848614997359765, "learning_rate": 6.954281908659745e-05, "epoch": 0.39128271469797987, "step": 983}, {"loss": 0.3716038465499878, "grad_norm": 0.03762047711826348, "learning_rate": 6.94834740950407e-05, "epoch": 0.39168076425514975, "step": 984}, {"loss": 0.41812199354171753, "grad_norm": 0.04049576975576149, "learning_rate": 6.942409672511067e-05, "epoch": 0.3920788138123196, "step": 985}, {"loss": 0.3299465775489807, "grad_norm": 0.0341345569754252, "learning_rate": 6.936468707548296e-05, "epoch": 0.3924768633694895, "step": 986}, {"loss": 0.3991384506225586, "grad_norm": 0.040137206997276416, "learning_rate": 6.930524524488674e-05, "epoch": 0.3928749129266594, "step": 987}, {"loss": 0.387466162443161, "grad_norm": 0.03426619901211974, "learning_rate": 6.92457713321047e-05, "epoch": 0.39327296248382926, "step": 988}, {"loss": 0.34796464443206787, "grad_norm": 0.03895526286145538, "learning_rate": 6.918626543597283e-05, "epoch": 0.3936710120409991, "step": 989}, {"loss": 0.38768473267555237, "grad_norm": 0.03836879385029728, "learning_rate": 6.912672765538029e-05, "epoch": 0.39406906159816896, "step": 990}, {"loss": 0.3445826768875122, "grad_norm": 0.034785128833191785, "learning_rate": 6.90671580892692e-05, "epoch": 0.39446711115533883, "step": 991}, {"loss": 0.403422087430954, "grad_norm": 0.037905833398890594, "learning_rate": 6.90075568366345e-05, "epoch": 0.3948651607125087, "step": 992}, {"loss": 0.3958730399608612, "grad_norm": 0.04092029894783264, "learning_rate": 6.894792399652383e-05, "epoch": 0.3952632102696786, "step": 993}, {"loss": 0.38266700506210327, "grad_norm": 0.03559684129994385, "learning_rate": 6.88882596680373e-05, "epoch": 0.39566125982684847, "step": 994}, {"loss": 0.3948616087436676, "grad_norm": 0.03560005180844706, "learning_rate": 6.882856395032731e-05, "epoch": 0.3960593093840183, "step": 995}, {"loss": 0.3506733179092407, "grad_norm": 0.03726057265322386, "learning_rate": 6.87688369425985e-05, "epoch": 0.39645735894118816, "step": 996}, {"loss": 0.346149206161499, "grad_norm": 0.03447243553659441, "learning_rate": 6.870907874410744e-05, "epoch": 0.39685540849835804, "step": 997}, {"loss": 0.4151867628097534, "grad_norm": 0.03898026451466907, "learning_rate": 6.864928945416256e-05, "epoch": 0.3972534580555279, "step": 998}, {"loss": 0.36745521426200867, "grad_norm": 0.04149121506146197, "learning_rate": 6.858946917212399e-05, "epoch": 0.3976515076126978, "step": 999}, {"loss": 0.3608710467815399, "grad_norm": 0.036273337714042876, "learning_rate": 6.85296179974033e-05, "epoch": 0.3980495571698677, "step": 1000}, {"loss": 0.39311105012893677, "grad_norm": 0.038570732381852046, "learning_rate": 6.846973602946348e-05, "epoch": 0.3984476067270375, "step": 1001}, {"loss": 0.38886791467666626, "grad_norm": 0.03862334122466048, "learning_rate": 6.840982336781858e-05, "epoch": 0.3988456562842074, "step": 1002}, {"loss": 0.33682191371917725, "grad_norm": 0.03479141971364623, "learning_rate": 6.834988011203375e-05, "epoch": 0.39924370584137725, "step": 1003}, {"loss": 0.3300033211708069, "grad_norm": 0.03764594160147828, "learning_rate": 6.8289906361725e-05, "epoch": 0.3996417553985471, "step": 1004}, {"loss": 0.38881731033325195, "grad_norm": 0.07310547406433644, "learning_rate": 6.822990221655892e-05, "epoch": 0.400039804955717, "step": 1005}, {"loss": 0.39167526364326477, "grad_norm": 0.038828292316288, "learning_rate": 6.816986777625269e-05, "epoch": 0.4004378545128869, "step": 1006}, {"loss": 0.363625705242157, "grad_norm": 0.03453396017928711, "learning_rate": 6.810980314057382e-05, "epoch": 0.4008359040700567, "step": 1007}, {"loss": 0.37414801120758057, "grad_norm": 0.0329042936349594, "learning_rate": 6.804970840933995e-05, "epoch": 0.4012339536272266, "step": 1008}, {"loss": 0.39009159803390503, "grad_norm": 0.03916964078752067, "learning_rate": 6.79895836824188e-05, "epoch": 0.40163200318439646, "step": 1009}, {"loss": 0.4297437071800232, "grad_norm": 0.03819191007089797, "learning_rate": 6.792942905972794e-05, "epoch": 0.40203005274156633, "step": 1010}, {"loss": 0.3687635064125061, "grad_norm": 0.038595378449432435, "learning_rate": 6.786924464123455e-05, "epoch": 0.4024281022987362, "step": 1011}, {"loss": 0.3686050772666931, "grad_norm": 0.03859009754938198, "learning_rate": 6.780903052695541e-05, "epoch": 0.40282615185590603, "step": 1012}, {"loss": 0.35071563720703125, "grad_norm": 0.03621693373186222, "learning_rate": 6.774878681695656e-05, "epoch": 0.4032242014130759, "step": 1013}, {"loss": 0.3923836648464203, "grad_norm": 0.03686261585340717, "learning_rate": 6.768851361135331e-05, "epoch": 0.4036222509702458, "step": 1014}, {"loss": 0.3629353642463684, "grad_norm": 0.03702720864090922, "learning_rate": 6.762821101030994e-05, "epoch": 0.40402030052741567, "step": 1015}, {"loss": 0.39003080129623413, "grad_norm": 0.03907849496110403, "learning_rate": 6.75678791140396e-05, "epoch": 0.40441835008458554, "step": 1016}, {"loss": 0.36020174622535706, "grad_norm": 0.03628784250293233, "learning_rate": 6.75075180228041e-05, "epoch": 0.4048163996417554, "step": 1017}, {"loss": 0.35455018281936646, "grad_norm": 0.03467604789932123, "learning_rate": 6.744712783691376e-05, "epoch": 0.40521444919892524, "step": 1018}, {"loss": 0.36921438574790955, "grad_norm": 0.03868725476243658, "learning_rate": 6.738670865672734e-05, "epoch": 0.4056124987560951, "step": 1019}, {"loss": 0.39923328161239624, "grad_norm": 0.05340121482338746, "learning_rate": 6.732626058265165e-05, "epoch": 0.406010548313265, "step": 1020}, {"loss": 0.4341929852962494, "grad_norm": 0.040435879907359466, "learning_rate": 6.726578371514163e-05, "epoch": 0.4064085978704349, "step": 1021}, {"loss": 0.3682630658149719, "grad_norm": 0.03483810660271312, "learning_rate": 6.720527815469999e-05, "epoch": 0.40680664742760475, "step": 1022}, {"loss": 0.4002925753593445, "grad_norm": 0.03897470894221226, "learning_rate": 6.714474400187716e-05, "epoch": 0.40720469698477463, "step": 1023}, {"loss": 0.372749924659729, "grad_norm": 0.03863595551043611, "learning_rate": 6.708418135727109e-05, "epoch": 0.40760274654194445, "step": 1024}, {"loss": 0.37175893783569336, "grad_norm": 0.037189951607706155, "learning_rate": 6.702359032152708e-05, "epoch": 0.4080007960991143, "step": 1025}, {"loss": 0.3932002782821655, "grad_norm": 0.042639337710452446, "learning_rate": 6.696297099533757e-05, "epoch": 0.4083988456562842, "step": 1026}, {"loss": 0.38057494163513184, "grad_norm": 0.03721487672408303, "learning_rate": 6.69023234794421e-05, "epoch": 0.4087968952134541, "step": 1027}, {"loss": 0.37353819608688354, "grad_norm": 0.03557235719604289, "learning_rate": 6.684164787462695e-05, "epoch": 0.40919494477062396, "step": 1028}, {"loss": 0.36271822452545166, "grad_norm": 0.034389203530271306, "learning_rate": 6.678094428172514e-05, "epoch": 0.40959299432779384, "step": 1029}, {"loss": 0.41627949476242065, "grad_norm": 0.038577289704132184, "learning_rate": 6.672021280161624e-05, "epoch": 0.40999104388496366, "step": 1030}, {"loss": 0.3839510679244995, "grad_norm": 0.03742490373498532, "learning_rate": 6.665945353522606e-05, "epoch": 0.41038909344213353, "step": 1031}, {"loss": 0.3530697524547577, "grad_norm": 0.03432615554968153, "learning_rate": 6.659866658352669e-05, "epoch": 0.4107871429993034, "step": 1032}, {"loss": 0.3437328636646271, "grad_norm": 0.03648109854725026, "learning_rate": 6.653785204753615e-05, "epoch": 0.4111851925564733, "step": 1033}, {"loss": 0.3552601933479309, "grad_norm": 0.034537588897155226, "learning_rate": 6.647701002831834e-05, "epoch": 0.41158324211364317, "step": 1034}, {"loss": 0.3831234574317932, "grad_norm": 0.03792814696106451, "learning_rate": 6.641614062698282e-05, "epoch": 0.41198129167081304, "step": 1035}, {"loss": 0.3620588183403015, "grad_norm": 0.038880456567051, "learning_rate": 6.635524394468468e-05, "epoch": 0.41237934122798287, "step": 1036}, {"loss": 0.422084778547287, "grad_norm": 0.04188436308209281, "learning_rate": 6.629432008262433e-05, "epoch": 0.41277739078515274, "step": 1037}, {"loss": 0.3932923674583435, "grad_norm": 0.041070981239408125, "learning_rate": 6.623336914204735e-05, "epoch": 0.4131754403423226, "step": 1038}, {"loss": 0.40946269035339355, "grad_norm": 0.03816457096743189, "learning_rate": 6.617239122424426e-05, "epoch": 0.4135734898994925, "step": 1039}, {"loss": 0.4280807673931122, "grad_norm": 0.03973761449280506, "learning_rate": 6.611138643055051e-05, "epoch": 0.4139715394566624, "step": 1040}, {"loss": 0.3864428400993347, "grad_norm": 0.04066067952011196, "learning_rate": 6.605035486234619e-05, "epoch": 0.4143695890138322, "step": 1041}, {"loss": 0.3253800868988037, "grad_norm": 0.03521776386955552, "learning_rate": 6.598929662105582e-05, "epoch": 0.4147676385710021, "step": 1042}, {"loss": 0.3592946529388428, "grad_norm": 0.035861139812713584, "learning_rate": 6.592821180814834e-05, "epoch": 0.41516568812817195, "step": 1043}, {"loss": 0.3878720998764038, "grad_norm": 0.03556655698978227, "learning_rate": 6.586710052513676e-05, "epoch": 0.4155637376853418, "step": 1044}, {"loss": 0.3721018433570862, "grad_norm": 0.03544829942124101, "learning_rate": 6.580596287357814e-05, "epoch": 0.4159617872425117, "step": 1045}, {"loss": 0.37046653032302856, "grad_norm": 0.03880814793148938, "learning_rate": 6.574479895507332e-05, "epoch": 0.4163598367996816, "step": 1046}, {"loss": 0.33823561668395996, "grad_norm": 0.04151272579092633, "learning_rate": 6.568360887126683e-05, "epoch": 0.4167578863568514, "step": 1047}, {"loss": 0.4224322438240051, "grad_norm": 0.03997838631988413, "learning_rate": 6.562239272384665e-05, "epoch": 0.4171559359140213, "step": 1048}, {"loss": 0.39314940571784973, "grad_norm": 0.0351096052976626, "learning_rate": 6.556115061454409e-05, "epoch": 0.41755398547119116, "step": 1049}, {"loss": 0.39251381158828735, "grad_norm": 0.03867108486321694, "learning_rate": 6.549988264513361e-05, "epoch": 0.41795203502836104, "step": 1050}, {"loss": 0.40018683671951294, "grad_norm": 0.036824485485118075, "learning_rate": 6.543858891743262e-05, "epoch": 0.4183500845855309, "step": 1051}, {"loss": 0.3927406966686249, "grad_norm": 0.03811367045164361, "learning_rate": 6.537726953330135e-05, "epoch": 0.4187481341427008, "step": 1052}, {"loss": 0.37567517161369324, "grad_norm": 0.03864212710206321, "learning_rate": 6.531592459464269e-05, "epoch": 0.4191461836998706, "step": 1053}, {"loss": 0.3623846471309662, "grad_norm": 0.0347638953791856, "learning_rate": 6.525455420340196e-05, "epoch": 0.4195442332570405, "step": 1054}, {"loss": 0.36084985733032227, "grad_norm": 0.036149141034747555, "learning_rate": 6.519315846156682e-05, "epoch": 0.41994228281421037, "step": 1055}, {"loss": 0.3999324142932892, "grad_norm": 0.03659557559799851, "learning_rate": 6.5131737471167e-05, "epoch": 0.42034033237138024, "step": 1056}, {"loss": 0.37827515602111816, "grad_norm": 0.03375261888888061, "learning_rate": 6.507029133427425e-05, "epoch": 0.4207383819285501, "step": 1057}, {"loss": 0.3734423816204071, "grad_norm": 0.03791124483518619, "learning_rate": 6.500882015300204e-05, "epoch": 0.42113643148572, "step": 1058}, {"loss": 0.3893612027168274, "grad_norm": 0.03696891186852855, "learning_rate": 6.494732402950554e-05, "epoch": 0.4215344810428898, "step": 1059}, {"loss": 0.39089590311050415, "grad_norm": 0.03595776611369711, "learning_rate": 6.488580306598131e-05, "epoch": 0.4219325306000597, "step": 1060}, {"loss": 0.4177897572517395, "grad_norm": 0.03978824949085779, "learning_rate": 6.482425736466724e-05, "epoch": 0.4223305801572296, "step": 1061}, {"loss": 0.3724096417427063, "grad_norm": 0.0380486300776369, "learning_rate": 6.476268702784226e-05, "epoch": 0.42272862971439945, "step": 1062}, {"loss": 0.35698190331459045, "grad_norm": 0.040222939291453076, "learning_rate": 6.470109215782631e-05, "epoch": 0.42312667927156933, "step": 1063}, {"loss": 0.35362881422042847, "grad_norm": 0.03587215210406882, "learning_rate": 6.463947285698007e-05, "epoch": 0.4235247288287392, "step": 1064}, {"loss": 0.33945679664611816, "grad_norm": 0.0357948490878811, "learning_rate": 6.457782922770481e-05, "epoch": 0.423922778385909, "step": 1065}, {"loss": 0.4060021638870239, "grad_norm": 0.04022887044866846, "learning_rate": 6.451616137244225e-05, "epoch": 0.4243208279430789, "step": 1066}, {"loss": 0.36610496044158936, "grad_norm": 0.036079403835720336, "learning_rate": 6.445446939367436e-05, "epoch": 0.4247188775002488, "step": 1067}, {"loss": 0.3544192612171173, "grad_norm": 0.0375908923462299, "learning_rate": 6.43927533939232e-05, "epoch": 0.42511692705741866, "step": 1068}, {"loss": 0.3529096841812134, "grad_norm": 0.03809512980816047, "learning_rate": 6.433101347575075e-05, "epoch": 0.42551497661458854, "step": 1069}, {"loss": 0.443853497505188, "grad_norm": 0.041014347003807114, "learning_rate": 6.426924974175874e-05, "epoch": 0.42591302617175836, "step": 1070}, {"loss": 0.39407628774642944, "grad_norm": 0.04115494677530542, "learning_rate": 6.420746229458848e-05, "epoch": 0.42631107572892823, "step": 1071}, {"loss": 0.3617252707481384, "grad_norm": 0.03800834343369183, "learning_rate": 6.414565123692068e-05, "epoch": 0.4267091252860981, "step": 1072}, {"loss": 0.3715144991874695, "grad_norm": 0.037108203438242594, "learning_rate": 6.408381667147527e-05, "epoch": 0.427107174843268, "step": 1073}, {"loss": 0.40530574321746826, "grad_norm": 0.03905117154776357, "learning_rate": 6.40219587010113e-05, "epoch": 0.42750522440043787, "step": 1074}, {"loss": 0.4036136865615845, "grad_norm": 0.042020900137383306, "learning_rate": 6.396007742832667e-05, "epoch": 0.42790327395760774, "step": 1075}, {"loss": 0.3949970304965973, "grad_norm": 0.03719840929439054, "learning_rate": 6.389817295625802e-05, "epoch": 0.42830132351477757, "step": 1076}, {"loss": 0.38722124695777893, "grad_norm": 0.03627866323445843, "learning_rate": 6.383624538768052e-05, "epoch": 0.42869937307194744, "step": 1077}, {"loss": 0.353490948677063, "grad_norm": 0.035282014510027875, "learning_rate": 6.377429482550777e-05, "epoch": 0.4290974226291173, "step": 1078}, {"loss": 0.3607896864414215, "grad_norm": 0.03721738298783189, "learning_rate": 6.371232137269154e-05, "epoch": 0.4294954721862872, "step": 1079}, {"loss": 0.3964061439037323, "grad_norm": 0.03676761495593476, "learning_rate": 6.365032513222164e-05, "epoch": 0.4298935217434571, "step": 1080}, {"loss": 0.37652555108070374, "grad_norm": 0.036171869019458056, "learning_rate": 6.35883062071258e-05, "epoch": 0.43029157130062695, "step": 1081}, {"loss": 0.3793220818042755, "grad_norm": 0.037785579943873826, "learning_rate": 6.352626470046939e-05, "epoch": 0.4306896208577968, "step": 1082}, {"loss": 0.35589808225631714, "grad_norm": 0.03550904612592203, "learning_rate": 6.346420071535533e-05, "epoch": 0.43108767041496665, "step": 1083}, {"loss": 0.37179118394851685, "grad_norm": 0.039696878072852104, "learning_rate": 6.34021143549239e-05, "epoch": 0.4314857199721365, "step": 1084}, {"loss": 0.38533806800842285, "grad_norm": 0.03770190829592167, "learning_rate": 6.334000572235256e-05, "epoch": 0.4318837695293064, "step": 1085}, {"loss": 0.34884029626846313, "grad_norm": 0.036084098808095016, "learning_rate": 6.327787492085577e-05, "epoch": 0.4322818190864763, "step": 1086}, {"loss": 0.4037531614303589, "grad_norm": 0.0396794279165848, "learning_rate": 6.321572205368488e-05, "epoch": 0.43267986864364616, "step": 1087}, {"loss": 0.3439621925354004, "grad_norm": 0.04057393185857347, "learning_rate": 6.315354722412782e-05, "epoch": 0.433077918200816, "step": 1088}, {"loss": 0.3753013610839844, "grad_norm": 0.03957297838386537, "learning_rate": 6.30913505355091e-05, "epoch": 0.43347596775798586, "step": 1089}, {"loss": 0.3856964707374573, "grad_norm": 0.036967574008413835, "learning_rate": 6.302913209118953e-05, "epoch": 0.43387401731515574, "step": 1090}, {"loss": 0.35508495569229126, "grad_norm": 0.03715165466157895, "learning_rate": 6.296689199456605e-05, "epoch": 0.4342720668723256, "step": 1091}, {"loss": 0.3745967149734497, "grad_norm": 0.0401071293195719, "learning_rate": 6.290463034907162e-05, "epoch": 0.4346701164294955, "step": 1092}, {"loss": 0.43121784925460815, "grad_norm": 0.04025778605912262, "learning_rate": 6.2842347258175e-05, "epoch": 0.43506816598666537, "step": 1093}, {"loss": 0.38961392641067505, "grad_norm": 0.03546382470570606, "learning_rate": 6.278004282538056e-05, "epoch": 0.4354662155438352, "step": 1094}, {"loss": 0.3712572455406189, "grad_norm": 0.035973175996694536, "learning_rate": 6.271771715422819e-05, "epoch": 0.43586426510100507, "step": 1095}, {"loss": 0.36575615406036377, "grad_norm": 0.03615457071796691, "learning_rate": 6.265537034829302e-05, "epoch": 0.43626231465817494, "step": 1096}, {"loss": 0.39060765504837036, "grad_norm": 0.04239420261874116, "learning_rate": 6.259300251118537e-05, "epoch": 0.4366603642153448, "step": 1097}, {"loss": 0.40133461356163025, "grad_norm": 0.038720602331590734, "learning_rate": 6.25306137465504e-05, "epoch": 0.4370584137725147, "step": 1098}, {"loss": 0.37237364053726196, "grad_norm": 0.035164370547842355, "learning_rate": 6.246820415806815e-05, "epoch": 0.4374564633296845, "step": 1099}, {"loss": 0.3960481882095337, "grad_norm": 0.03850228200878479, "learning_rate": 6.240577384945324e-05, "epoch": 0.4378545128868544, "step": 1100}, {"loss": 0.3990837335586548, "grad_norm": 0.043223618647838614, "learning_rate": 6.234332292445471e-05, "epoch": 0.4382525624440243, "step": 1101}, {"loss": 0.388821542263031, "grad_norm": 0.03469779670558362, "learning_rate": 6.228085148685586e-05, "epoch": 0.43865061200119415, "step": 1102}, {"loss": 0.37507420778274536, "grad_norm": 0.0349353211978777, "learning_rate": 6.221835964047408e-05, "epoch": 0.43904866155836403, "step": 1103}, {"loss": 0.40341347455978394, "grad_norm": 0.03607217588457412, "learning_rate": 6.21558474891607e-05, "epoch": 0.4394467111155339, "step": 1104}, {"loss": 0.4225393831729889, "grad_norm": 0.03797552717124645, "learning_rate": 6.209331513680075e-05, "epoch": 0.4398447606727037, "step": 1105}, {"loss": 0.4043247103691101, "grad_norm": 0.037409099526733505, "learning_rate": 6.203076268731288e-05, "epoch": 0.4402428102298736, "step": 1106}, {"loss": 0.3604956865310669, "grad_norm": 0.036300080016110486, "learning_rate": 6.196819024464911e-05, "epoch": 0.4406408597870435, "step": 1107}, {"loss": 0.3841654658317566, "grad_norm": 0.03547362587024723, "learning_rate": 6.190559791279467e-05, "epoch": 0.44103890934421336, "step": 1108}, {"loss": 0.39441198110580444, "grad_norm": 0.037599263225955346, "learning_rate": 6.184298579576789e-05, "epoch": 0.44143695890138324, "step": 1109}, {"loss": 0.3368271589279175, "grad_norm": 0.03514341206902375, "learning_rate": 6.178035399761992e-05, "epoch": 0.4418350084585531, "step": 1110}, {"loss": 0.3501816391944885, "grad_norm": 0.0344533820765208, "learning_rate": 6.171770262243466e-05, "epoch": 0.44223305801572294, "step": 1111}, {"loss": 0.42314302921295166, "grad_norm": 0.03970608199344407, "learning_rate": 6.165503177432855e-05, "epoch": 0.4426311075728928, "step": 1112}, {"loss": 0.34312087297439575, "grad_norm": 0.03598126446454638, "learning_rate": 6.159234155745035e-05, "epoch": 0.4430291571300627, "step": 1113}, {"loss": 0.35786616802215576, "grad_norm": 0.0358011959948908, "learning_rate": 6.152963207598103e-05, "epoch": 0.44342720668723257, "step": 1114}, {"loss": 0.4212929606437683, "grad_norm": 0.042052313903089646, "learning_rate": 6.14669034341336e-05, "epoch": 0.44382525624440244, "step": 1115}, {"loss": 0.3879357576370239, "grad_norm": 0.03444265502938135, "learning_rate": 6.140415573615285e-05, "epoch": 0.4442233058015723, "step": 1116}, {"loss": 0.3704368472099304, "grad_norm": 0.035840272726092276, "learning_rate": 6.134138908631531e-05, "epoch": 0.44462135535874214, "step": 1117}, {"loss": 0.3915570378303528, "grad_norm": 0.036632446487536496, "learning_rate": 6.127860358892895e-05, "epoch": 0.445019404915912, "step": 1118}, {"loss": 0.3519851863384247, "grad_norm": 0.0400959874508738, "learning_rate": 6.12157993483331e-05, "epoch": 0.4454174544730819, "step": 1119}, {"loss": 0.3548136353492737, "grad_norm": 0.037194246107587314, "learning_rate": 6.11529764688982e-05, "epoch": 0.4458155040302518, "step": 1120}, {"loss": 0.3668060898780823, "grad_norm": 0.03580325119752799, "learning_rate": 6.109013505502566e-05, "epoch": 0.44621355358742165, "step": 1121}, {"loss": 0.3602171242237091, "grad_norm": 0.03613036009449391, "learning_rate": 6.102727521114777e-05, "epoch": 0.44661160314459153, "step": 1122}, {"loss": 0.38199907541275024, "grad_norm": 0.033468327661395285, "learning_rate": 6.0964397041727384e-05, "epoch": 0.44700965270176135, "step": 1123}, {"loss": 0.37550538778305054, "grad_norm": 0.03640751967367033, "learning_rate": 6.090150065125778e-05, "epoch": 0.44740770225893123, "step": 1124}, {"loss": 0.348245769739151, "grad_norm": 0.03671646039984662, "learning_rate": 6.083858614426259e-05, "epoch": 0.4478057518161011, "step": 1125}, {"loss": 0.3580840528011322, "grad_norm": 0.03705851992770693, "learning_rate": 6.077565362529549e-05, "epoch": 0.448203801373271, "step": 1126}, {"loss": 0.41514286398887634, "grad_norm": 0.0411187165085972, "learning_rate": 6.071270319894015e-05, "epoch": 0.44860185093044086, "step": 1127}, {"loss": 0.4018482565879822, "grad_norm": 0.04047346955233768, "learning_rate": 6.0649734969809956e-05, "epoch": 0.4489999004876107, "step": 1128}, {"loss": 0.4158313274383545, "grad_norm": 0.0378188443353165, "learning_rate": 6.058674904254788e-05, "epoch": 0.44939795004478056, "step": 1129}, {"loss": 0.3635256588459015, "grad_norm": 0.03499690909335363, "learning_rate": 6.0523745521826324e-05, "epoch": 0.44979599960195044, "step": 1130}, {"loss": 0.3755130171775818, "grad_norm": 0.036387924518374984, "learning_rate": 6.0460724512346914e-05, "epoch": 0.4501940491591203, "step": 1131}, {"loss": 0.395702600479126, "grad_norm": 0.03906948529455254, "learning_rate": 6.039768611884034e-05, "epoch": 0.4505920987162902, "step": 1132}, {"loss": 0.39556828141212463, "grad_norm": 0.036202044358235756, "learning_rate": 6.03346304460662e-05, "epoch": 0.45099014827346007, "step": 1133}, {"loss": 0.4113815724849701, "grad_norm": 0.03742876059705795, "learning_rate": 6.027155759881279e-05, "epoch": 0.4513881978306299, "step": 1134}, {"loss": 0.39060813188552856, "grad_norm": 0.03906970207087607, "learning_rate": 6.020846768189693e-05, "epoch": 0.45178624738779977, "step": 1135}, {"loss": 0.4037550985813141, "grad_norm": 0.04188640624095433, "learning_rate": 6.014536080016384e-05, "epoch": 0.45218429694496964, "step": 1136}, {"loss": 0.3821169137954712, "grad_norm": 0.04009414666958078, "learning_rate": 6.008223705848692e-05, "epoch": 0.4525823465021395, "step": 1137}, {"loss": 0.38647016882896423, "grad_norm": 0.03627750559280531, "learning_rate": 6.0019096561767604e-05, "epoch": 0.4529803960593094, "step": 1138}, {"loss": 0.4044339656829834, "grad_norm": 0.03754392422336407, "learning_rate": 5.9955939414935135e-05, "epoch": 0.4533784456164793, "step": 1139}, {"loss": 0.3940204679965973, "grad_norm": 0.03487633124140043, "learning_rate": 5.9892765722946464e-05, "epoch": 0.4537764951736491, "step": 1140}, {"loss": 0.35458099842071533, "grad_norm": 0.03659962008417149, "learning_rate": 5.982957559078599e-05, "epoch": 0.454174544730819, "step": 1141}, {"loss": 0.3772830367088318, "grad_norm": 0.0331501423730011, "learning_rate": 5.976636912346548e-05, "epoch": 0.45457259428798885, "step": 1142}, {"loss": 0.36136186122894287, "grad_norm": 0.03818795685140892, "learning_rate": 5.970314642602384e-05, "epoch": 0.45497064384515873, "step": 1143}, {"loss": 0.37427958846092224, "grad_norm": 0.0358885850877556, "learning_rate": 5.96399076035269e-05, "epoch": 0.4553686934023286, "step": 1144}, {"loss": 0.38884955644607544, "grad_norm": 0.03632774857346843, "learning_rate": 5.9576652761067374e-05, "epoch": 0.4557667429594985, "step": 1145}, {"loss": 0.37677478790283203, "grad_norm": 0.03488243807300091, "learning_rate": 5.951338200376451e-05, "epoch": 0.4561647925166683, "step": 1146}, {"loss": 0.3801659941673279, "grad_norm": 0.03772218622360735, "learning_rate": 5.9450095436764043e-05, "epoch": 0.4565628420738382, "step": 1147}, {"loss": 0.32646745443344116, "grad_norm": 0.03444506628120715, "learning_rate": 5.938679316523801e-05, "epoch": 0.45696089163100806, "step": 1148}, {"loss": 0.36697328090667725, "grad_norm": 0.03609756632784099, "learning_rate": 5.932347529438449e-05, "epoch": 0.45735894118817794, "step": 1149}, {"loss": 0.38726741075515747, "grad_norm": 0.039651198447267326, "learning_rate": 5.9260141929427526e-05, "epoch": 0.4577569907453478, "step": 1150}, {"loss": 0.3907758295536041, "grad_norm": 0.03950307989412426, "learning_rate": 5.91967931756169e-05, "epoch": 0.4581550403025177, "step": 1151}, {"loss": 0.4165996313095093, "grad_norm": 0.03989047669063669, "learning_rate": 5.9133429138227955e-05, "epoch": 0.4585530898596875, "step": 1152}, {"loss": 0.3766500949859619, "grad_norm": 0.03775196852025751, "learning_rate": 5.907004992256144e-05, "epoch": 0.4589511394168574, "step": 1153}, {"loss": 0.3704991936683655, "grad_norm": 0.03410515016762748, "learning_rate": 5.900665563394334e-05, "epoch": 0.45934918897402727, "step": 1154}, {"loss": 0.36373770236968994, "grad_norm": 0.038065525478241775, "learning_rate": 5.894324637772466e-05, "epoch": 0.45974723853119714, "step": 1155}, {"loss": 0.33799082040786743, "grad_norm": 0.03919117658858488, "learning_rate": 5.8879822259281324e-05, "epoch": 0.460145288088367, "step": 1156}, {"loss": 0.35624778270721436, "grad_norm": 0.03425209643733996, "learning_rate": 5.8816383384013915e-05, "epoch": 0.46054333764553684, "step": 1157}, {"loss": 0.375675767660141, "grad_norm": 0.03558574819748386, "learning_rate": 5.8752929857347536e-05, "epoch": 0.4609413872027067, "step": 1158}, {"loss": 0.3375585079193115, "grad_norm": 0.03679212088083657, "learning_rate": 5.868946178473168e-05, "epoch": 0.4613394367598766, "step": 1159}, {"loss": 0.3100201487541199, "grad_norm": 0.03512711394067779, "learning_rate": 5.8625979271639976e-05, "epoch": 0.4617374863170465, "step": 1160}, {"loss": 0.38598203659057617, "grad_norm": 0.039166061341269744, "learning_rate": 5.856248242357008e-05, "epoch": 0.46213553587421635, "step": 1161}, {"loss": 0.3603900074958801, "grad_norm": 0.03672372401831359, "learning_rate": 5.849897134604341e-05, "epoch": 0.46253358543138623, "step": 1162}, {"loss": 0.36806806921958923, "grad_norm": 0.03857403301755807, "learning_rate": 5.8435446144605125e-05, "epoch": 0.46293163498855605, "step": 1163}, {"loss": 0.3368191123008728, "grad_norm": 0.036927230259655636, "learning_rate": 5.837190692482376e-05, "epoch": 0.46332968454572593, "step": 1164}, {"loss": 0.3957725167274475, "grad_norm": 0.036247379222208144, "learning_rate": 5.830835379229122e-05, "epoch": 0.4637277341028958, "step": 1165}, {"loss": 0.35047775506973267, "grad_norm": 0.03664860670738801, "learning_rate": 5.824478685262248e-05, "epoch": 0.4641257836600657, "step": 1166}, {"loss": 0.3636395037174225, "grad_norm": 0.03496392427047426, "learning_rate": 5.818120621145551e-05, "epoch": 0.46452383321723556, "step": 1167}, {"loss": 0.40208443999290466, "grad_norm": 0.039346082120528265, "learning_rate": 5.8117611974451e-05, "epoch": 0.46492188277440544, "step": 1168}, {"loss": 0.3926907479763031, "grad_norm": 0.04071296963402902, "learning_rate": 5.805400424729226e-05, "epoch": 0.46531993233157526, "step": 1169}, {"loss": 0.350517213344574, "grad_norm": 0.036065204090336714, "learning_rate": 5.799038313568501e-05, "epoch": 0.46571798188874514, "step": 1170}, {"loss": 0.297579288482666, "grad_norm": 0.03601246129178732, "learning_rate": 5.792674874535722e-05, "epoch": 0.466116031445915, "step": 1171}, {"loss": 0.39720481634140015, "grad_norm": 0.03646463899327683, "learning_rate": 5.786310118205893e-05, "epoch": 0.4665140810030849, "step": 1172}, {"loss": 0.3866811692714691, "grad_norm": 0.03678504988292268, "learning_rate": 5.779944055156207e-05, "epoch": 0.46691213056025477, "step": 1173}, {"loss": 0.376099556684494, "grad_norm": 0.03761296371616325, "learning_rate": 5.7735766959660266e-05, "epoch": 0.46731018011742465, "step": 1174}, {"loss": 0.359019935131073, "grad_norm": 0.038652114216969985, "learning_rate": 5.767208051216869e-05, "epoch": 0.46770822967459447, "step": 1175}, {"loss": 0.40018871426582336, "grad_norm": 0.03790100722627931, "learning_rate": 5.76083813149239e-05, "epoch": 0.46810627923176434, "step": 1176}, {"loss": 0.41470828652381897, "grad_norm": 0.03789650209192395, "learning_rate": 5.754466947378363e-05, "epoch": 0.4685043287889342, "step": 1177}, {"loss": 0.34036025404930115, "grad_norm": 0.03446350519444202, "learning_rate": 5.748094509462663e-05, "epoch": 0.4689023783461041, "step": 1178}, {"loss": 0.3873051106929779, "grad_norm": 0.039319395665121146, "learning_rate": 5.741720828335246e-05, "epoch": 0.469300427903274, "step": 1179}, {"loss": 0.3784462809562683, "grad_norm": 0.037701382955163364, "learning_rate": 5.735345914588138e-05, "epoch": 0.46969847746044385, "step": 1180}, {"loss": 0.3219122290611267, "grad_norm": 0.03997273579774234, "learning_rate": 5.728969778815411e-05, "epoch": 0.4700965270176137, "step": 1181}, {"loss": 0.34468066692352295, "grad_norm": 0.038608920382528396, "learning_rate": 5.7225924316131665e-05, "epoch": 0.47049457657478355, "step": 1182}, {"loss": 0.3994317054748535, "grad_norm": 0.037559206396456435, "learning_rate": 5.716213883579524e-05, "epoch": 0.47089262613195343, "step": 1183}, {"loss": 0.3311782479286194, "grad_norm": 0.03419051357298744, "learning_rate": 5.7098341453145944e-05, "epoch": 0.4712906756891233, "step": 1184}, {"loss": 0.38912349939346313, "grad_norm": 0.041145243702036295, "learning_rate": 5.7034532274204675e-05, "epoch": 0.4716887252462932, "step": 1185}, {"loss": 0.3466004729270935, "grad_norm": 0.03748944495383589, "learning_rate": 5.697071140501193e-05, "epoch": 0.472086774803463, "step": 1186}, {"loss": 0.3588666617870331, "grad_norm": 0.03697407400267784, "learning_rate": 5.690687895162764e-05, "epoch": 0.4724848243606329, "step": 1187}, {"loss": 0.37510040402412415, "grad_norm": 0.03679509324221678, "learning_rate": 5.6843035020131e-05, "epoch": 0.47288287391780276, "step": 1188}, {"loss": 0.40294697880744934, "grad_norm": 0.04071148827581889, "learning_rate": 5.677917971662027e-05, "epoch": 0.47328092347497264, "step": 1189}, {"loss": 0.4165912866592407, "grad_norm": 0.03904452847471841, "learning_rate": 5.671531314721259e-05, "epoch": 0.4736789730321425, "step": 1190}, {"loss": 0.3819833993911743, "grad_norm": 0.03918281658500507, "learning_rate": 5.665143541804384e-05, "epoch": 0.4740770225893124, "step": 1191}, {"loss": 0.44376856088638306, "grad_norm": 0.03676340306868458, "learning_rate": 5.658754663526845e-05, "epoch": 0.4744750721464822, "step": 1192}, {"loss": 0.334980309009552, "grad_norm": 0.03511262084454545, "learning_rate": 5.6523646905059225e-05, "epoch": 0.4748731217036521, "step": 1193}, {"loss": 0.33646488189697266, "grad_norm": 0.036192851317749464, "learning_rate": 5.645973633360713e-05, "epoch": 0.47527117126082197, "step": 1194}, {"loss": 0.3964965343475342, "grad_norm": 0.038945998731849854, "learning_rate": 5.639581502712117e-05, "epoch": 0.47566922081799184, "step": 1195}, {"loss": 0.395749032497406, "grad_norm": 0.03523080305053528, "learning_rate": 5.633188309182817e-05, "epoch": 0.4760672703751617, "step": 1196}, {"loss": 0.3222233057022095, "grad_norm": 0.033546152708546155, "learning_rate": 5.626794063397269e-05, "epoch": 0.4764653199323316, "step": 1197}, {"loss": 0.3928260803222656, "grad_norm": 0.0374292831448294, "learning_rate": 5.6203987759816665e-05, "epoch": 0.4768633694895014, "step": 1198}, {"loss": 0.37699949741363525, "grad_norm": 0.03668064940462698, "learning_rate": 5.614002457563943e-05, "epoch": 0.4772614190466713, "step": 1199}, {"loss": 0.37359970808029175, "grad_norm": 0.03592350271140173, "learning_rate": 5.60760511877374e-05, "epoch": 0.4776594686038412, "step": 1200}, {"loss": 0.3536718189716339, "grad_norm": 0.03754923940774018, "learning_rate": 5.601206770242399e-05, "epoch": 0.47805751816101105, "step": 1201}, {"loss": 0.35670310258865356, "grad_norm": 0.036259548576279485, "learning_rate": 5.5948074226029356e-05, "epoch": 0.47845556771818093, "step": 1202}, {"loss": 0.36814481019973755, "grad_norm": 0.034449556566499, "learning_rate": 5.5884070864900274e-05, "epoch": 0.4788536172753508, "step": 1203}, {"loss": 0.3688846528530121, "grad_norm": 0.03636791489360737, "learning_rate": 5.5820057725399965e-05, "epoch": 0.47925166683252063, "step": 1204}, {"loss": 0.38934633135795593, "grad_norm": 0.038470063816171234, "learning_rate": 5.5756034913907865e-05, "epoch": 0.4796497163896905, "step": 1205}, {"loss": 0.3635393977165222, "grad_norm": 0.036495945015898756, "learning_rate": 5.569200253681949e-05, "epoch": 0.4800477659468604, "step": 1206}, {"loss": 0.3713546395301819, "grad_norm": 0.03677517125057603, "learning_rate": 5.562796070054629e-05, "epoch": 0.48044581550403026, "step": 1207}, {"loss": 0.38271117210388184, "grad_norm": 0.03912142782110174, "learning_rate": 5.55639095115154e-05, "epoch": 0.48084386506120014, "step": 1208}, {"loss": 0.34730109572410583, "grad_norm": 0.036455325668966, "learning_rate": 5.54998490761695e-05, "epoch": 0.48124191461837, "step": 1209}, {"loss": 0.38035985827445984, "grad_norm": 0.044791348981900604, "learning_rate": 5.543577950096664e-05, "epoch": 0.48163996417553984, "step": 1210}, {"loss": 0.34143921732902527, "grad_norm": 0.03792173236396935, "learning_rate": 5.5371700892380066e-05, "epoch": 0.4820380137327097, "step": 1211}, {"loss": 0.3722090721130371, "grad_norm": 0.03521024374320186, "learning_rate": 5.530761335689801e-05, "epoch": 0.4824360632898796, "step": 1212}, {"loss": 0.3634943962097168, "grad_norm": 0.03976797544509414, "learning_rate": 5.52435170010236e-05, "epoch": 0.48283411284704947, "step": 1213}, {"loss": 0.36230939626693726, "grad_norm": 0.03785743471831599, "learning_rate": 5.517941193127456e-05, "epoch": 0.48323216240421935, "step": 1214}, {"loss": 0.3657744228839874, "grad_norm": 0.034264513629351255, "learning_rate": 5.5115298254183114e-05, "epoch": 0.48363021196138917, "step": 1215}, {"loss": 0.35658857226371765, "grad_norm": 0.03872726714259163, "learning_rate": 5.505117607629579e-05, "epoch": 0.48402826151855904, "step": 1216}, {"loss": 0.3797532021999359, "grad_norm": 0.03639129608000449, "learning_rate": 5.498704550417324e-05, "epoch": 0.4844263110757289, "step": 1217}, {"loss": 0.37077757716178894, "grad_norm": 0.04081394082876936, "learning_rate": 5.492290664439009e-05, "epoch": 0.4848243606328988, "step": 1218}, {"loss": 0.3278660476207733, "grad_norm": 0.03829842286899257, "learning_rate": 5.485875960353472e-05, "epoch": 0.4852224101900687, "step": 1219}, {"loss": 0.3726128935813904, "grad_norm": 0.03702070154363758, "learning_rate": 5.479460448820909e-05, "epoch": 0.48562045974723855, "step": 1220}, {"loss": 0.34285596013069153, "grad_norm": 0.035105576995391546, "learning_rate": 5.473044140502859e-05, "epoch": 0.4860185093044084, "step": 1221}, {"loss": 0.36556142568588257, "grad_norm": 0.037516585314601045, "learning_rate": 5.466627046062187e-05, "epoch": 0.48641655886157825, "step": 1222}, {"loss": 0.4021671414375305, "grad_norm": 0.04110837688128548, "learning_rate": 5.4602091761630626e-05, "epoch": 0.48681460841874813, "step": 1223}, {"loss": 0.3584437072277069, "grad_norm": 0.035025670737526035, "learning_rate": 5.453790541470946e-05, "epoch": 0.487212657975918, "step": 1224}, {"loss": 0.41606029868125916, "grad_norm": 0.038050010376093964, "learning_rate": 5.447371152652564e-05, "epoch": 0.4876107075330879, "step": 1225}, {"loss": 0.38591426610946655, "grad_norm": 0.03870676434488021, "learning_rate": 5.4409510203759016e-05, "epoch": 0.48800875709025776, "step": 1226}, {"loss": 0.4259459674358368, "grad_norm": 0.03865747290186669, "learning_rate": 5.4345301553101746e-05, "epoch": 0.4884068066474276, "step": 1227}, {"loss": 0.3299679756164551, "grad_norm": 0.03664633278128708, "learning_rate": 5.4281085681258215e-05, "epoch": 0.48880485620459746, "step": 1228}, {"loss": 0.4049144387245178, "grad_norm": 0.03749395971152908, "learning_rate": 5.421686269494477e-05, "epoch": 0.48920290576176734, "step": 1229}, {"loss": 0.3673116862773895, "grad_norm": 0.03696005746111006, "learning_rate": 5.415263270088962e-05, "epoch": 0.4896009553189372, "step": 1230}, {"loss": 0.4020673334598541, "grad_norm": 0.03798193969418997, "learning_rate": 5.408839580583256e-05, "epoch": 0.4899990048761071, "step": 1231}, {"loss": 0.36567723751068115, "grad_norm": 0.037239789866563816, "learning_rate": 5.402415211652492e-05, "epoch": 0.49039705443327697, "step": 1232}, {"loss": 0.36912524700164795, "grad_norm": 0.03981820567693898, "learning_rate": 5.3959901739729256e-05, "epoch": 0.4907951039904468, "step": 1233}, {"loss": 0.3823345899581909, "grad_norm": 0.03772345956572202, "learning_rate": 5.3895644782219314e-05, "epoch": 0.49119315354761667, "step": 1234}, {"loss": 0.3443232774734497, "grad_norm": 0.035405817401838305, "learning_rate": 5.383138135077971e-05, "epoch": 0.49159120310478654, "step": 1235}, {"loss": 0.33088958263397217, "grad_norm": 0.03398751923088524, "learning_rate": 5.376711155220584e-05, "epoch": 0.4919892526619564, "step": 1236}, {"loss": 0.3671081066131592, "grad_norm": 0.03466790344921965, "learning_rate": 5.370283549330368e-05, "epoch": 0.4923873022191263, "step": 1237}, {"loss": 0.3789195716381073, "grad_norm": 0.03652881819543863, "learning_rate": 5.363855328088963e-05, "epoch": 0.4927853517762962, "step": 1238}, {"loss": 0.37275630235671997, "grad_norm": 0.0348928147473413, "learning_rate": 5.357426502179027e-05, "epoch": 0.493183401333466, "step": 1239}, {"loss": 0.3873579502105713, "grad_norm": 0.03834851414882032, "learning_rate": 5.350997082284228e-05, "epoch": 0.4935814508906359, "step": 1240}, {"loss": 0.3723049759864807, "grad_norm": 0.03734789322277753, "learning_rate": 5.34456707908922e-05, "epoch": 0.49397950044780575, "step": 1241}, {"loss": 0.3835778832435608, "grad_norm": 0.038322495660667, "learning_rate": 5.338136503279623e-05, "epoch": 0.49437755000497563, "step": 1242}, {"loss": 0.3574468493461609, "grad_norm": 0.038317307063003035, "learning_rate": 5.3317053655420105e-05, "epoch": 0.4947755995621455, "step": 1243}, {"loss": 0.38007569313049316, "grad_norm": 0.03610677542400796, "learning_rate": 5.325273676563891e-05, "epoch": 0.49517364911931533, "step": 1244}, {"loss": 0.3952692747116089, "grad_norm": 0.03919041437706827, "learning_rate": 5.318841447033689e-05, "epoch": 0.4955716986764852, "step": 1245}, {"loss": 0.3889434039592743, "grad_norm": 0.03830916831021415, "learning_rate": 5.312408687640727e-05, "epoch": 0.4959697482336551, "step": 1246}, {"loss": 0.4363046884536743, "grad_norm": 0.03773645873344558, "learning_rate": 5.305975409075204e-05, "epoch": 0.49636779779082496, "step": 1247}, {"loss": 0.34351646900177, "grad_norm": 0.032094307829628785, "learning_rate": 5.2995416220281904e-05, "epoch": 0.49676584734799484, "step": 1248}, {"loss": 0.37258949875831604, "grad_norm": 0.03631428019821192, "learning_rate": 5.2931073371915916e-05, "epoch": 0.4971638969051647, "step": 1249}, {"loss": 0.41162264347076416, "grad_norm": 0.03778493776763402, "learning_rate": 5.2866725652581486e-05, "epoch": 0.49756194646233454, "step": 1250}, {"loss": 0.39077091217041016, "grad_norm": 0.03689729666690168, "learning_rate": 5.2802373169214056e-05, "epoch": 0.4979599960195044, "step": 1251}, {"loss": 0.35946282744407654, "grad_norm": 0.03895072336714278, "learning_rate": 5.2738016028757046e-05, "epoch": 0.4983580455766743, "step": 1252}, {"loss": 0.38457608222961426, "grad_norm": 0.03820880553707822, "learning_rate": 5.267365433816157e-05, "epoch": 0.49875609513384417, "step": 1253}, {"loss": 0.4018512964248657, "grad_norm": 0.040708842267369844, "learning_rate": 5.260928820438631e-05, "epoch": 0.49915414469101405, "step": 1254}, {"loss": 0.39805907011032104, "grad_norm": 0.03874367608833234, "learning_rate": 5.254491773439734e-05, "epoch": 0.4995521942481839, "step": 1255}, {"loss": 0.3691985011100769, "grad_norm": 0.03875529233047036, "learning_rate": 5.248054303516794e-05, "epoch": 0.49995024380535374, "step": 1256}, {"loss": 0.39525294303894043, "grad_norm": 0.037128481313677276, "learning_rate": 5.241616421367843e-05, "epoch": 0.5003482933625236, "step": 1257}, {"loss": 0.3915693759918213, "grad_norm": 0.036434908765100825, "learning_rate": 5.235178137691596e-05, "epoch": 0.5007463429196936, "step": 1258}, {"loss": 0.33865925669670105, "grad_norm": 0.040687100640810325, "learning_rate": 5.2287394631874355e-05, "epoch": 0.5011443924768634, "step": 1259}, {"loss": 0.4151684045791626, "grad_norm": 0.03763056839176957, "learning_rate": 5.222300408555394e-05, "epoch": 0.5015424420340332, "step": 1260}, {"loss": 0.352144718170166, "grad_norm": 0.03603199579403706, "learning_rate": 5.2158609844961346e-05, "epoch": 0.5019404915912031, "step": 1261}, {"loss": 0.40353113412857056, "grad_norm": 0.039355598864601625, "learning_rate": 5.209421201710938e-05, "epoch": 0.502338541148373, "step": 1262}, {"loss": 0.4314389228820801, "grad_norm": 0.04055528722448075, "learning_rate": 5.202981070901677e-05, "epoch": 0.5027365907055429, "step": 1263}, {"loss": 0.34920206665992737, "grad_norm": 0.03460906075268911, "learning_rate": 5.196540602770805e-05, "epoch": 0.5031346402627127, "step": 1264}, {"loss": 0.3854944407939911, "grad_norm": 0.036284193813512604, "learning_rate": 5.190099808021335e-05, "epoch": 0.5035326898198825, "step": 1265}, {"loss": 0.3588533401489258, "grad_norm": 0.03308800862010561, "learning_rate": 5.1836586973568224e-05, "epoch": 0.5039307393770525, "step": 1266}, {"loss": 0.41268259286880493, "grad_norm": 0.03837494985345541, "learning_rate": 5.177217281481348e-05, "epoch": 0.5043287889342223, "step": 1267}, {"loss": 0.3964235186576843, "grad_norm": 0.04021687537831155, "learning_rate": 5.170775571099502e-05, "epoch": 0.5047268384913922, "step": 1268}, {"loss": 0.3539178967475891, "grad_norm": 0.033680127631969974, "learning_rate": 5.1643335769163606e-05, "epoch": 0.505124888048562, "step": 1269}, {"loss": 0.35676029324531555, "grad_norm": 0.0354722149038247, "learning_rate": 5.157891309637473e-05, "epoch": 0.5055229376057319, "step": 1270}, {"loss": 0.3782637417316437, "grad_norm": 0.038694317522279614, "learning_rate": 5.1514487799688414e-05, "epoch": 0.5059209871629018, "step": 1271}, {"loss": 0.36010313034057617, "grad_norm": 0.03736696864061114, "learning_rate": 5.145005998616908e-05, "epoch": 0.5063190367200716, "step": 1272}, {"loss": 0.3926060199737549, "grad_norm": 0.039851760485098045, "learning_rate": 5.138562976288527e-05, "epoch": 0.5067170862772415, "step": 1273}, {"loss": 0.36220037937164307, "grad_norm": 0.03634654821695794, "learning_rate": 5.132119723690957e-05, "epoch": 0.5071151358344114, "step": 1274}, {"loss": 0.3423851728439331, "grad_norm": 0.037583187518825736, "learning_rate": 5.1256762515318416e-05, "epoch": 0.5075131853915813, "step": 1275}, {"loss": 0.39759135246276855, "grad_norm": 0.037109544153970324, "learning_rate": 5.119232570519182e-05, "epoch": 0.5079112349487511, "step": 1276}, {"loss": 0.35478517413139343, "grad_norm": 0.05262240383262713, "learning_rate": 5.112788691361333e-05, "epoch": 0.5083092845059209, "step": 1277}, {"loss": 0.34199661016464233, "grad_norm": 0.0356011430828162, "learning_rate": 5.106344624766975e-05, "epoch": 0.5087073340630909, "step": 1278}, {"loss": 0.38995784521102905, "grad_norm": 0.0371361722613098, "learning_rate": 5.099900381445105e-05, "epoch": 0.5091053836202607, "step": 1279}, {"loss": 0.36334723234176636, "grad_norm": 0.03812359016102496, "learning_rate": 5.093455972105006e-05, "epoch": 0.5095034331774306, "step": 1280}, {"loss": 0.3614453673362732, "grad_norm": 0.03857770103175099, "learning_rate": 5.0870114074562424e-05, "epoch": 0.5099014827346005, "step": 1281}, {"loss": 0.3352402150630951, "grad_norm": 0.03672687360883936, "learning_rate": 5.080566698208634e-05, "epoch": 0.5102995322917703, "step": 1282}, {"loss": 0.34491434693336487, "grad_norm": 0.03760310606679323, "learning_rate": 5.074121855072242e-05, "epoch": 0.5106975818489402, "step": 1283}, {"loss": 0.3156568109989166, "grad_norm": 0.03600815190148399, "learning_rate": 5.067676888757353e-05, "epoch": 0.51109563140611, "step": 1284}, {"loss": 0.3855277895927429, "grad_norm": 0.03863072763126752, "learning_rate": 5.061231809974454e-05, "epoch": 0.51149368096328, "step": 1285}, {"loss": 0.3693893849849701, "grad_norm": 0.03458490389906258, "learning_rate": 5.054786629434219e-05, "epoch": 0.5118917305204498, "step": 1286}, {"loss": 0.36609429121017456, "grad_norm": 0.03979462984240158, "learning_rate": 5.0483413578474945e-05, "epoch": 0.5122897800776197, "step": 1287}, {"loss": 0.3963834047317505, "grad_norm": 0.0387511929670458, "learning_rate": 5.041896005925275e-05, "epoch": 0.5126878296347895, "step": 1288}, {"loss": 0.3936919867992401, "grad_norm": 0.0375226839758183, "learning_rate": 5.0354505843786894e-05, "epoch": 0.5130858791919594, "step": 1289}, {"loss": 0.4116871654987335, "grad_norm": 0.037225475795901256, "learning_rate": 5.029005103918986e-05, "epoch": 0.5134839287491293, "step": 1290}, {"loss": 0.3543041944503784, "grad_norm": 0.03312449073110895, "learning_rate": 5.022559575257503e-05, "epoch": 0.5138819783062991, "step": 1291}, {"loss": 0.390561044216156, "grad_norm": 0.041138550400168034, "learning_rate": 5.016114009105667e-05, "epoch": 0.514280027863469, "step": 1292}, {"loss": 0.38059914112091064, "grad_norm": 0.03870867114138564, "learning_rate": 5.0096684161749595e-05, "epoch": 0.5146780774206389, "step": 1293}, {"loss": 0.3469569981098175, "grad_norm": 0.03360430575803803, "learning_rate": 5.003222807176915e-05, "epoch": 0.5150761269778087, "step": 1294}, {"loss": 0.33001452684402466, "grad_norm": 0.03544368349500637, "learning_rate": 4.996777192823087e-05, "epoch": 0.5154741765349786, "step": 1295}, {"loss": 0.3989471197128296, "grad_norm": 0.03892443050783798, "learning_rate": 4.990331583825041e-05, "epoch": 0.5158722260921484, "step": 1296}, {"loss": 0.38934627175331116, "grad_norm": 0.036423194704035794, "learning_rate": 4.983885990894335e-05, "epoch": 0.5162702756493184, "step": 1297}, {"loss": 0.41923901438713074, "grad_norm": 0.0430756368918295, "learning_rate": 4.977440424742499e-05, "epoch": 0.5166683252064882, "step": 1298}, {"loss": 0.36835116147994995, "grad_norm": 0.03850428437263681, "learning_rate": 4.970994896081016e-05, "epoch": 0.517066374763658, "step": 1299}, {"loss": 0.3849736750125885, "grad_norm": 0.03706642866592711, "learning_rate": 4.9645494156213104e-05, "epoch": 0.517464424320828, "step": 1300}, {"loss": 0.3392372131347656, "grad_norm": 0.040190804317424195, "learning_rate": 4.9581039940747255e-05, "epoch": 0.5178624738779978, "step": 1301}, {"loss": 0.400834858417511, "grad_norm": 0.03945491241899721, "learning_rate": 4.951658642152505e-05, "epoch": 0.5182605234351677, "step": 1302}, {"loss": 0.3227859139442444, "grad_norm": 0.03711731070331935, "learning_rate": 4.945213370565781e-05, "epoch": 0.5186585729923375, "step": 1303}, {"loss": 0.35633450746536255, "grad_norm": 0.03702311819191652, "learning_rate": 4.9387681900255466e-05, "epoch": 0.5190566225495075, "step": 1304}, {"loss": 0.3581780791282654, "grad_norm": 0.0390335261117401, "learning_rate": 4.932323111242647e-05, "epoch": 0.5194546721066773, "step": 1305}, {"loss": 0.377602219581604, "grad_norm": 0.037047804314701054, "learning_rate": 4.925878144927759e-05, "epoch": 0.5198527216638471, "step": 1306}, {"loss": 0.3524743318557739, "grad_norm": 0.03590400527673243, "learning_rate": 4.9194333017913684e-05, "epoch": 0.520250771221017, "step": 1307}, {"loss": 0.3686349093914032, "grad_norm": 0.03783801038447595, "learning_rate": 4.91298859254376e-05, "epoch": 0.5206488207781869, "step": 1308}, {"loss": 0.3280102610588074, "grad_norm": 0.031243181968277148, "learning_rate": 4.906544027894996e-05, "epoch": 0.5210468703353568, "step": 1309}, {"loss": 0.40247535705566406, "grad_norm": 0.03930139972692783, "learning_rate": 4.9000996185548974e-05, "epoch": 0.5214449198925266, "step": 1310}, {"loss": 0.36921054124832153, "grad_norm": 0.041395291845138744, "learning_rate": 4.893655375233025e-05, "epoch": 0.5218429694496964, "step": 1311}, {"loss": 0.3740151524543762, "grad_norm": 0.0361250103113098, "learning_rate": 4.8872113086386686e-05, "epoch": 0.5222410190068664, "step": 1312}, {"loss": 0.34609225392341614, "grad_norm": 0.036904370828265906, "learning_rate": 4.880767429480819e-05, "epoch": 0.5226390685640362, "step": 1313}, {"loss": 0.361050009727478, "grad_norm": 0.03833218708590021, "learning_rate": 4.8743237484681596e-05, "epoch": 0.5230371181212061, "step": 1314}, {"loss": 0.36963313817977905, "grad_norm": 0.03985428776687452, "learning_rate": 4.8678802763090435e-05, "epoch": 0.523435167678376, "step": 1315}, {"loss": 0.3771844208240509, "grad_norm": 0.035349500699215615, "learning_rate": 4.861437023711475e-05, "epoch": 0.5238332172355459, "step": 1316}, {"loss": 0.3782932758331299, "grad_norm": 0.03632031364779364, "learning_rate": 4.8549940013830934e-05, "epoch": 0.5242312667927157, "step": 1317}, {"loss": 0.4223320484161377, "grad_norm": 0.03871157498973699, "learning_rate": 4.848551220031159e-05, "epoch": 0.5246293163498855, "step": 1318}, {"loss": 0.32020434737205505, "grad_norm": 0.03422884222243228, "learning_rate": 4.8421086903625276e-05, "epoch": 0.5250273659070555, "step": 1319}, {"loss": 0.34961897134780884, "grad_norm": 0.03998562452492239, "learning_rate": 4.835666423083641e-05, "epoch": 0.5254254154642253, "step": 1320}, {"loss": 0.33273717761039734, "grad_norm": 0.03615601991983762, "learning_rate": 4.829224428900499e-05, "epoch": 0.5258234650213952, "step": 1321}, {"loss": 0.39682668447494507, "grad_norm": 0.03670787953009987, "learning_rate": 4.822782718518652e-05, "epoch": 0.526221514578565, "step": 1322}, {"loss": 0.41369327902793884, "grad_norm": 0.042220919527005435, "learning_rate": 4.816341302643178e-05, "epoch": 0.5266195641357349, "step": 1323}, {"loss": 0.3490922152996063, "grad_norm": 0.03463815287218508, "learning_rate": 4.809900191978665e-05, "epoch": 0.5270176136929048, "step": 1324}, {"loss": 0.3742249011993408, "grad_norm": 0.034186667273290255, "learning_rate": 4.803459397229196e-05, "epoch": 0.5274156632500746, "step": 1325}, {"loss": 0.3257296085357666, "grad_norm": 0.03641287142429623, "learning_rate": 4.797018929098323e-05, "epoch": 0.5278137128072445, "step": 1326}, {"loss": 0.37079960107803345, "grad_norm": 0.03642298729148534, "learning_rate": 4.7905787982890644e-05, "epoch": 0.5282117623644144, "step": 1327}, {"loss": 0.37160035967826843, "grad_norm": 0.035572086079809793, "learning_rate": 4.784139015503867e-05, "epoch": 0.5286098119215842, "step": 1328}, {"loss": 0.39504528045654297, "grad_norm": 0.0368956506363415, "learning_rate": 4.777699591444609e-05, "epoch": 0.5290078614787541, "step": 1329}, {"loss": 0.3750969469547272, "grad_norm": 0.03804504751303787, "learning_rate": 4.771260536812567e-05, "epoch": 0.5294059110359239, "step": 1330}, {"loss": 0.33185112476348877, "grad_norm": 0.03767236988600486, "learning_rate": 4.764821862308405e-05, "epoch": 0.5298039605930939, "step": 1331}, {"loss": 0.3652339577674866, "grad_norm": 0.03635044517798451, "learning_rate": 4.758383578632158e-05, "epoch": 0.5302020101502637, "step": 1332}, {"loss": 0.34177064895629883, "grad_norm": 0.03767948114441329, "learning_rate": 4.7519456964832067e-05, "epoch": 0.5306000597074336, "step": 1333}, {"loss": 0.37291818857192993, "grad_norm": 0.0392213236268448, "learning_rate": 4.745508226560267e-05, "epoch": 0.5309981092646034, "step": 1334}, {"loss": 0.43030452728271484, "grad_norm": 0.04264135321230142, "learning_rate": 4.73907117956137e-05, "epoch": 0.5313961588217733, "step": 1335}, {"loss": 0.3845194876194, "grad_norm": 0.03807440839184051, "learning_rate": 4.732634566183844e-05, "epoch": 0.5317942083789432, "step": 1336}, {"loss": 0.4053289592266083, "grad_norm": 0.04007441363671537, "learning_rate": 4.7261983971242966e-05, "epoch": 0.532192257936113, "step": 1337}, {"loss": 0.35036998987197876, "grad_norm": 0.03393676492304717, "learning_rate": 4.719762683078595e-05, "epoch": 0.532590307493283, "step": 1338}, {"loss": 0.37265121936798096, "grad_norm": 0.036601376480585195, "learning_rate": 4.713327434741853e-05, "epoch": 0.5329883570504528, "step": 1339}, {"loss": 0.38289907574653625, "grad_norm": 0.03836186203338345, "learning_rate": 4.7068926628084096e-05, "epoch": 0.5333864066076226, "step": 1340}, {"loss": 0.35110336542129517, "grad_norm": 0.04074544275978065, "learning_rate": 4.700458377971811e-05, "epoch": 0.5337844561647925, "step": 1341}, {"loss": 0.3420702815055847, "grad_norm": 0.03557877568191968, "learning_rate": 4.694024590924796e-05, "epoch": 0.5341825057219624, "step": 1342}, {"loss": 0.3989320993423462, "grad_norm": 0.037530234614611455, "learning_rate": 4.687591312359274e-05, "epoch": 0.5345805552791323, "step": 1343}, {"loss": 0.3533203899860382, "grad_norm": 0.03479154236154664, "learning_rate": 4.681158552966311e-05, "epoch": 0.5349786048363021, "step": 1344}, {"loss": 0.38157904148101807, "grad_norm": 0.03921328225273712, "learning_rate": 4.6747263234361085e-05, "epoch": 0.535376654393472, "step": 1345}, {"loss": 0.36454063653945923, "grad_norm": 0.03795063445968256, "learning_rate": 4.6682946344579893e-05, "epoch": 0.5357747039506419, "step": 1346}, {"loss": 0.4012272357940674, "grad_norm": 0.04254630133704739, "learning_rate": 4.661863496720378e-05, "epoch": 0.5361727535078117, "step": 1347}, {"loss": 0.34140267968177795, "grad_norm": 0.03414401864267894, "learning_rate": 4.655432920910782e-05, "epoch": 0.5365708030649816, "step": 1348}, {"loss": 0.34526968002319336, "grad_norm": 0.03681366074447913, "learning_rate": 4.649002917715772e-05, "epoch": 0.5369688526221514, "step": 1349}, {"loss": 0.4335389733314514, "grad_norm": 0.042253719774488795, "learning_rate": 4.642573497820975e-05, "epoch": 0.5373669021793214, "step": 1350}, {"loss": 0.3920060396194458, "grad_norm": 0.03817278391771853, "learning_rate": 4.6361446719110405e-05, "epoch": 0.5377649517364912, "step": 1351}, {"loss": 0.39699453115463257, "grad_norm": 0.0375534313682507, "learning_rate": 4.629716450669634e-05, "epoch": 0.538163001293661, "step": 1352}, {"loss": 0.36047160625457764, "grad_norm": 0.037005767444832256, "learning_rate": 4.623288844779418e-05, "epoch": 0.538561050850831, "step": 1353}, {"loss": 0.37551385164260864, "grad_norm": 0.043128405512911444, "learning_rate": 4.61686186492203e-05, "epoch": 0.5389591004080008, "step": 1354}, {"loss": 0.4084341824054718, "grad_norm": 0.039738190517709354, "learning_rate": 4.6104355217780704e-05, "epoch": 0.5393571499651707, "step": 1355}, {"loss": 0.37786972522735596, "grad_norm": 0.039816022961587864, "learning_rate": 4.604009826027075e-05, "epoch": 0.5397551995223405, "step": 1356}, {"loss": 0.382270872592926, "grad_norm": 0.03928725444856859, "learning_rate": 4.59758478834751e-05, "epoch": 0.5401532490795103, "step": 1357}, {"loss": 0.3847622275352478, "grad_norm": 0.04069302275774452, "learning_rate": 4.591160419416744e-05, "epoch": 0.5405512986366803, "step": 1358}, {"loss": 0.37604662775993347, "grad_norm": 0.034203858525197674, "learning_rate": 4.5847367299110394e-05, "epoch": 0.5409493481938501, "step": 1359}, {"loss": 0.3454027771949768, "grad_norm": 0.03461434289663709, "learning_rate": 4.578313730505524e-05, "epoch": 0.54134739775102, "step": 1360}, {"loss": 0.39446601271629333, "grad_norm": 0.037841846927994915, "learning_rate": 4.5718914318741803e-05, "epoch": 0.5417454473081899, "step": 1361}, {"loss": 0.348932683467865, "grad_norm": 0.039208686916515155, "learning_rate": 4.5654698446898266e-05, "epoch": 0.5421434968653598, "step": 1362}, {"loss": 0.3413374423980713, "grad_norm": 0.0349665675725125, "learning_rate": 4.5590489796240996e-05, "epoch": 0.5425415464225296, "step": 1363}, {"loss": 0.3963305652141571, "grad_norm": 0.04084802667290766, "learning_rate": 4.552628847347437e-05, "epoch": 0.5429395959796994, "step": 1364}, {"loss": 0.36317020654678345, "grad_norm": 0.036004579615264105, "learning_rate": 4.5462094585290554e-05, "epoch": 0.5433376455368694, "step": 1365}, {"loss": 0.3496832549571991, "grad_norm": 0.03446973038467828, "learning_rate": 4.5397908238369366e-05, "epoch": 0.5437356950940392, "step": 1366}, {"loss": 0.41146138310432434, "grad_norm": 0.04281944487523147, "learning_rate": 4.533372953937813e-05, "epoch": 0.5441337446512091, "step": 1367}, {"loss": 0.39069032669067383, "grad_norm": 0.042818913724489704, "learning_rate": 4.526955859497141e-05, "epoch": 0.5445317942083789, "step": 1368}, {"loss": 0.36535078287124634, "grad_norm": 0.035826473093154475, "learning_rate": 4.520539551179091e-05, "epoch": 0.5449298437655488, "step": 1369}, {"loss": 0.3873562812805176, "grad_norm": 0.04045545245180776, "learning_rate": 4.51412403964653e-05, "epoch": 0.5453278933227187, "step": 1370}, {"loss": 0.4022118151187897, "grad_norm": 0.038650309015719586, "learning_rate": 4.507709335560992e-05, "epoch": 0.5457259428798885, "step": 1371}, {"loss": 0.3449764549732208, "grad_norm": 0.03547087515979208, "learning_rate": 4.5012954495826773e-05, "epoch": 0.5461239924370584, "step": 1372}, {"loss": 0.4050453305244446, "grad_norm": 0.0386252547405431, "learning_rate": 4.494882392370423e-05, "epoch": 0.5465220419942283, "step": 1373}, {"loss": 0.36528563499450684, "grad_norm": 0.040128048226385334, "learning_rate": 4.4884701745816905e-05, "epoch": 0.5469200915513982, "step": 1374}, {"loss": 0.36388784646987915, "grad_norm": 0.03789471970138307, "learning_rate": 4.482058806872545e-05, "epoch": 0.547318141108568, "step": 1375}, {"loss": 0.3947349190711975, "grad_norm": 0.037667229054051045, "learning_rate": 4.47564829989764e-05, "epoch": 0.5477161906657378, "step": 1376}, {"loss": 0.4023997485637665, "grad_norm": 0.04033205913381574, "learning_rate": 4.4692386643101994e-05, "epoch": 0.5481142402229078, "step": 1377}, {"loss": 0.36376142501831055, "grad_norm": 0.03660902199358023, "learning_rate": 4.462829910761995e-05, "epoch": 0.5485122897800776, "step": 1378}, {"loss": 0.4060869812965393, "grad_norm": 0.03860856405718109, "learning_rate": 4.4564220499033375e-05, "epoch": 0.5489103393372475, "step": 1379}, {"loss": 0.40045738220214844, "grad_norm": 0.04286908991384851, "learning_rate": 4.450015092383051e-05, "epoch": 0.5493083888944174, "step": 1380}, {"loss": 0.3727560341358185, "grad_norm": 0.03828220198872983, "learning_rate": 4.443609048848461e-05, "epoch": 0.5497064384515872, "step": 1381}, {"loss": 0.3723967671394348, "grad_norm": 0.03808503888913672, "learning_rate": 4.437203929945373e-05, "epoch": 0.5501044880087571, "step": 1382}, {"loss": 0.40236538648605347, "grad_norm": 0.03928036967068712, "learning_rate": 4.430799746318052e-05, "epoch": 0.5505025375659269, "step": 1383}, {"loss": 0.3858381509780884, "grad_norm": 0.03829184449133399, "learning_rate": 4.4243965086092154e-05, "epoch": 0.5509005871230969, "step": 1384}, {"loss": 0.36716634035110474, "grad_norm": 0.04245746187063126, "learning_rate": 4.4179942274600054e-05, "epoch": 0.5512986366802667, "step": 1385}, {"loss": 0.37201574444770813, "grad_norm": 0.03953209131419996, "learning_rate": 4.4115929135099724e-05, "epoch": 0.5516966862374365, "step": 1386}, {"loss": 0.3828321099281311, "grad_norm": 0.0400474474494043, "learning_rate": 4.4051925773970656e-05, "epoch": 0.5520947357946064, "step": 1387}, {"loss": 0.36801037192344666, "grad_norm": 0.03842109536948109, "learning_rate": 4.398793229757602e-05, "epoch": 0.5524927853517763, "step": 1388}, {"loss": 0.3591698110103607, "grad_norm": 0.03464890324110814, "learning_rate": 4.3923948812262596e-05, "epoch": 0.5528908349089462, "step": 1389}, {"loss": 0.3254311680793762, "grad_norm": 0.03799956722115735, "learning_rate": 4.3859975424360574e-05, "epoch": 0.553288884466116, "step": 1390}, {"loss": 0.3468061089515686, "grad_norm": 0.03797158412374946, "learning_rate": 4.379601224018335e-05, "epoch": 0.553686934023286, "step": 1391}, {"loss": 0.37590810656547546, "grad_norm": 0.03794139065876156, "learning_rate": 4.373205936602733e-05, "epoch": 0.5540849835804558, "step": 1392}, {"loss": 0.35631313920021057, "grad_norm": 0.0385678020912351, "learning_rate": 4.366811690817183e-05, "epoch": 0.5544830331376256, "step": 1393}, {"loss": 0.3252243399620056, "grad_norm": 0.03295491947210555, "learning_rate": 4.360418497287885e-05, "epoch": 0.5548810826947955, "step": 1394}, {"loss": 0.36037033796310425, "grad_norm": 0.04268104360627898, "learning_rate": 4.354026366639289e-05, "epoch": 0.5552791322519653, "step": 1395}, {"loss": 0.3949925899505615, "grad_norm": 0.03967969795962407, "learning_rate": 4.347635309494079e-05, "epoch": 0.5556771818091353, "step": 1396}, {"loss": 0.3295280635356903, "grad_norm": 0.033554446720162744, "learning_rate": 4.3412453364731557e-05, "epoch": 0.5560752313663051, "step": 1397}, {"loss": 0.3442707359790802, "grad_norm": 0.037460341826362153, "learning_rate": 4.3348564581956163e-05, "epoch": 0.5564732809234749, "step": 1398}, {"loss": 0.35273507237434387, "grad_norm": 0.03802475806014993, "learning_rate": 4.328468685278742e-05, "epoch": 0.5568713304806449, "step": 1399}, {"loss": 0.3622961640357971, "grad_norm": 0.04146789918664561, "learning_rate": 4.322082028337975e-05, "epoch": 0.5572693800378147, "step": 1400}, {"loss": 0.37150731682777405, "grad_norm": 0.03952655532188959, "learning_rate": 4.3156964979869006e-05, "epoch": 0.5576674295949846, "step": 1401}, {"loss": 0.39060306549072266, "grad_norm": 0.03938337083929605, "learning_rate": 4.3093121048372364e-05, "epoch": 0.5580654791521544, "step": 1402}, {"loss": 0.42115986347198486, "grad_norm": 0.04020093982876788, "learning_rate": 4.302928859498808e-05, "epoch": 0.5584635287093244, "step": 1403}, {"loss": 0.36175018548965454, "grad_norm": 0.03955843501357955, "learning_rate": 4.2965467725795336e-05, "epoch": 0.5588615782664942, "step": 1404}, {"loss": 0.41246023774147034, "grad_norm": 0.03886865575141582, "learning_rate": 4.290165854685407e-05, "epoch": 0.559259627823664, "step": 1405}, {"loss": 0.3649027347564697, "grad_norm": 0.03607538735873018, "learning_rate": 4.283786116420476e-05, "epoch": 0.5596576773808339, "step": 1406}, {"loss": 0.3538016676902771, "grad_norm": 0.03726033203096025, "learning_rate": 4.277407568386833e-05, "epoch": 0.5600557269380038, "step": 1407}, {"loss": 0.3434063196182251, "grad_norm": 0.03425944061337341, "learning_rate": 4.27103022118459e-05, "epoch": 0.5604537764951737, "step": 1408}, {"loss": 0.33478331565856934, "grad_norm": 0.03358534227422084, "learning_rate": 4.2646540854118624e-05, "epoch": 0.5608518260523435, "step": 1409}, {"loss": 0.3549577295780182, "grad_norm": 0.04075748561307251, "learning_rate": 4.258279171664755e-05, "epoch": 0.5612498756095133, "step": 1410}, {"loss": 0.38263845443725586, "grad_norm": 0.03837965966723768, "learning_rate": 4.251905490537338e-05, "epoch": 0.5616479251666833, "step": 1411}, {"loss": 0.36798566579818726, "grad_norm": 0.03515039114769803, "learning_rate": 4.245533052621637e-05, "epoch": 0.5620459747238531, "step": 1412}, {"loss": 0.37892287969589233, "grad_norm": 0.03725679909745875, "learning_rate": 4.2391618685076126e-05, "epoch": 0.562444024281023, "step": 1413}, {"loss": 0.3458070158958435, "grad_norm": 0.03869077033662991, "learning_rate": 4.232791948783134e-05, "epoch": 0.5628420738381928, "step": 1414}, {"loss": 0.3665464222431183, "grad_norm": 0.03757107026126323, "learning_rate": 4.226423304033976e-05, "epoch": 0.5632401233953627, "step": 1415}, {"loss": 0.3215528130531311, "grad_norm": 0.0364144482469732, "learning_rate": 4.220055944843795e-05, "epoch": 0.5636381729525326, "step": 1416}, {"loss": 0.39481252431869507, "grad_norm": 0.03945907984901699, "learning_rate": 4.2136898817941087e-05, "epoch": 0.5640362225097024, "step": 1417}, {"loss": 0.3697317838668823, "grad_norm": 0.036798541900640205, "learning_rate": 4.207325125464279e-05, "epoch": 0.5644342720668724, "step": 1418}, {"loss": 0.401546448469162, "grad_norm": 0.04083597685286913, "learning_rate": 4.2009616864315e-05, "epoch": 0.5648323216240422, "step": 1419}, {"loss": 0.3873291611671448, "grad_norm": 0.036277982046408024, "learning_rate": 4.194599575270776e-05, "epoch": 0.5652303711812121, "step": 1420}, {"loss": 0.39931726455688477, "grad_norm": 0.03789589507864439, "learning_rate": 4.188238802554901e-05, "epoch": 0.5656284207383819, "step": 1421}, {"loss": 0.34670358896255493, "grad_norm": 0.03747210073293811, "learning_rate": 4.181879378854451e-05, "epoch": 0.5660264702955518, "step": 1422}, {"loss": 0.3369259536266327, "grad_norm": 0.03523170795848293, "learning_rate": 4.175521314737753e-05, "epoch": 0.5664245198527217, "step": 1423}, {"loss": 0.39996135234832764, "grad_norm": 0.038691252607412414, "learning_rate": 4.1691646207708794e-05, "epoch": 0.5668225694098915, "step": 1424}, {"loss": 0.37516313791275024, "grad_norm": 0.03596674583352536, "learning_rate": 4.162809307517625e-05, "epoch": 0.5672206189670614, "step": 1425}, {"loss": 0.3966641426086426, "grad_norm": 0.03910180643232596, "learning_rate": 4.1564553855394894e-05, "epoch": 0.5676186685242313, "step": 1426}, {"loss": 0.34076470136642456, "grad_norm": 0.0366793792371702, "learning_rate": 4.15010286539566e-05, "epoch": 0.5680167180814011, "step": 1427}, {"loss": 0.35089999437332153, "grad_norm": 0.03746575599093211, "learning_rate": 4.143751757642994e-05, "epoch": 0.568414767638571, "step": 1428}, {"loss": 0.3111540377140045, "grad_norm": 0.037003720522827535, "learning_rate": 4.137402072836002e-05, "epoch": 0.5688128171957408, "step": 1429}, {"loss": 0.35011768341064453, "grad_norm": 0.04035016402601048, "learning_rate": 4.1310538215268315e-05, "epoch": 0.5692108667529108, "step": 1430}, {"loss": 0.31710124015808105, "grad_norm": 0.03718733743229535, "learning_rate": 4.124707014265246e-05, "epoch": 0.5696089163100806, "step": 1431}, {"loss": 0.36210137605667114, "grad_norm": 0.0400696991436463, "learning_rate": 4.11836166159861e-05, "epoch": 0.5700069658672505, "step": 1432}, {"loss": 0.3863376975059509, "grad_norm": 0.03969402634830041, "learning_rate": 4.112017774071868e-05, "epoch": 0.5704050154244203, "step": 1433}, {"loss": 0.3838648498058319, "grad_norm": 0.039319830937184816, "learning_rate": 4.105675362227536e-05, "epoch": 0.5708030649815902, "step": 1434}, {"loss": 0.3563937842845917, "grad_norm": 0.040934112941783514, "learning_rate": 4.0993344366056686e-05, "epoch": 0.5712011145387601, "step": 1435}, {"loss": 0.40131211280822754, "grad_norm": 0.04046550883441495, "learning_rate": 4.092995007743858e-05, "epoch": 0.5715991640959299, "step": 1436}, {"loss": 0.39017724990844727, "grad_norm": 0.042784977565256045, "learning_rate": 4.0866570861772056e-05, "epoch": 0.5719972136530999, "step": 1437}, {"loss": 0.3556586503982544, "grad_norm": 0.036255673312119624, "learning_rate": 4.080320682438311e-05, "epoch": 0.5723952632102697, "step": 1438}, {"loss": 0.3811701536178589, "grad_norm": 0.0410463701652713, "learning_rate": 4.073985807057249e-05, "epoch": 0.5727933127674395, "step": 1439}, {"loss": 0.37277254462242126, "grad_norm": 0.033507386825851665, "learning_rate": 4.0676524705615526e-05, "epoch": 0.5731913623246094, "step": 1440}, {"loss": 0.3159979581832886, "grad_norm": 0.03697387185426966, "learning_rate": 4.061320683476201e-05, "epoch": 0.5735894118817793, "step": 1441}, {"loss": 0.36663687229156494, "grad_norm": 0.037873013475676454, "learning_rate": 4.054990456323596e-05, "epoch": 0.5739874614389492, "step": 1442}, {"loss": 0.3380158245563507, "grad_norm": 0.034862106916356606, "learning_rate": 4.048661799623551e-05, "epoch": 0.574385510996119, "step": 1443}, {"loss": 0.36514800786972046, "grad_norm": 0.03648360141170109, "learning_rate": 4.042334723893264e-05, "epoch": 0.5747835605532888, "step": 1444}, {"loss": 0.3839520812034607, "grad_norm": 0.03847795002861718, "learning_rate": 4.036009239647311e-05, "epoch": 0.5751816101104588, "step": 1445}, {"loss": 0.414617121219635, "grad_norm": 0.04433949061143192, "learning_rate": 4.029685357397618e-05, "epoch": 0.5755796596676286, "step": 1446}, {"loss": 0.3579922020435333, "grad_norm": 0.039972819495198715, "learning_rate": 4.0233630876534525e-05, "epoch": 0.5759777092247985, "step": 1447}, {"loss": 0.33641692996025085, "grad_norm": 0.03520916749359931, "learning_rate": 4.0170424409214014e-05, "epoch": 0.5763757587819683, "step": 1448}, {"loss": 0.39083075523376465, "grad_norm": 0.03962765073969019, "learning_rate": 4.010723427705354e-05, "epoch": 0.5767738083391383, "step": 1449}, {"loss": 0.33700448274612427, "grad_norm": 0.03714294485698637, "learning_rate": 4.004406058506487e-05, "epoch": 0.5771718578963081, "step": 1450}, {"loss": 0.35564881563186646, "grad_norm": 0.03937981890879557, "learning_rate": 3.99809034382324e-05, "epoch": 0.5775699074534779, "step": 1451}, {"loss": 0.3449694514274597, "grad_norm": 0.038723739240088144, "learning_rate": 3.991776294151307e-05, "epoch": 0.5779679570106478, "step": 1452}, {"loss": 0.36534297466278076, "grad_norm": 0.03746958031585588, "learning_rate": 3.9854639199836156e-05, "epoch": 0.5783660065678177, "step": 1453}, {"loss": 0.40701407194137573, "grad_norm": 0.04448579765574059, "learning_rate": 3.9791532318103075e-05, "epoch": 0.5787640561249876, "step": 1454}, {"loss": 0.4110798239707947, "grad_norm": 0.0381084680407692, "learning_rate": 3.972844240118724e-05, "epoch": 0.5791621056821574, "step": 1455}, {"loss": 0.368230938911438, "grad_norm": 0.037731829814779114, "learning_rate": 3.9665369553933815e-05, "epoch": 0.5795601552393272, "step": 1456}, {"loss": 0.3983006179332733, "grad_norm": 0.04293748945463655, "learning_rate": 3.960231388115968e-05, "epoch": 0.5799582047964972, "step": 1457}, {"loss": 0.40512216091156006, "grad_norm": 0.039085370578919795, "learning_rate": 3.9539275487653105e-05, "epoch": 0.580356254353667, "step": 1458}, {"loss": 0.3858243227005005, "grad_norm": 0.040241618479040064, "learning_rate": 3.947625447817369e-05, "epoch": 0.5807543039108369, "step": 1459}, {"loss": 0.36988621950149536, "grad_norm": 0.03934519380064259, "learning_rate": 3.941325095745213e-05, "epoch": 0.5811523534680068, "step": 1460}, {"loss": 0.3623926639556885, "grad_norm": 0.04040368102135461, "learning_rate": 3.9350265030190056e-05, "epoch": 0.5815504030251767, "step": 1461}, {"loss": 0.3307524621486664, "grad_norm": 0.03826055688676019, "learning_rate": 3.9287296801059856e-05, "epoch": 0.5819484525823465, "step": 1462}, {"loss": 0.36559706926345825, "grad_norm": 0.035770095144414454, "learning_rate": 3.9224346374704515e-05, "epoch": 0.5823465021395163, "step": 1463}, {"loss": 0.3890279233455658, "grad_norm": 0.0355541642372174, "learning_rate": 3.916141385573743e-05, "epoch": 0.5827445516966863, "step": 1464}, {"loss": 0.39375296235084534, "grad_norm": 0.0412614664567786, "learning_rate": 3.909849934874224e-05, "epoch": 0.5831426012538561, "step": 1465}, {"loss": 0.38048601150512695, "grad_norm": 0.035518665612685976, "learning_rate": 3.9035602958272634e-05, "epoch": 0.583540650811026, "step": 1466}, {"loss": 0.38089150190353394, "grad_norm": 0.0352052256192589, "learning_rate": 3.897272478885224e-05, "epoch": 0.5839387003681958, "step": 1467}, {"loss": 0.3688841760158539, "grad_norm": 0.034721389454075095, "learning_rate": 3.890986494497434e-05, "epoch": 0.5843367499253657, "step": 1468}, {"loss": 0.3207430839538574, "grad_norm": 0.032735798459217, "learning_rate": 3.884702353110182e-05, "epoch": 0.5847347994825356, "step": 1469}, {"loss": 0.4276953339576721, "grad_norm": 0.03967023204230433, "learning_rate": 3.878420065166691e-05, "epoch": 0.5851328490397054, "step": 1470}, {"loss": 0.3781083822250366, "grad_norm": 0.03917043445258268, "learning_rate": 3.8721396411071046e-05, "epoch": 0.5855308985968753, "step": 1471}, {"loss": 0.35458335280418396, "grad_norm": 0.039907929327235865, "learning_rate": 3.865861091368469e-05, "epoch": 0.5859289481540452, "step": 1472}, {"loss": 0.35634690523147583, "grad_norm": 0.03855930367358774, "learning_rate": 3.8595844263847146e-05, "epoch": 0.586326997711215, "step": 1473}, {"loss": 0.37370115518569946, "grad_norm": 0.03692088009253235, "learning_rate": 3.853309656586641e-05, "epoch": 0.5867250472683849, "step": 1474}, {"loss": 0.4167558550834656, "grad_norm": 0.039023002025134534, "learning_rate": 3.8470367924018965e-05, "epoch": 0.5871230968255547, "step": 1475}, {"loss": 0.37387847900390625, "grad_norm": 0.03789365988096989, "learning_rate": 3.8407658442549656e-05, "epoch": 0.5875211463827247, "step": 1476}, {"loss": 0.4038604497909546, "grad_norm": 0.03820578585998294, "learning_rate": 3.834496822567148e-05, "epoch": 0.5879191959398945, "step": 1477}, {"loss": 0.3491314947605133, "grad_norm": 0.03643610866267774, "learning_rate": 3.828229737756535e-05, "epoch": 0.5883172454970644, "step": 1478}, {"loss": 0.4069643020629883, "grad_norm": 0.038177909236334656, "learning_rate": 3.82196460023801e-05, "epoch": 0.5887152950542343, "step": 1479}, {"loss": 0.3805556297302246, "grad_norm": 0.040998752993043903, "learning_rate": 3.815701420423213e-05, "epoch": 0.5891133446114041, "step": 1480}, {"loss": 0.3198011815547943, "grad_norm": 0.0344416473442439, "learning_rate": 3.809440208720534e-05, "epoch": 0.589511394168574, "step": 1481}, {"loss": 0.3618210256099701, "grad_norm": 0.03552154775956825, "learning_rate": 3.80318097553509e-05, "epoch": 0.5899094437257438, "step": 1482}, {"loss": 0.4118311405181885, "grad_norm": 0.03896852910849183, "learning_rate": 3.796923731268712e-05, "epoch": 0.5903074932829138, "step": 1483}, {"loss": 0.3632003962993622, "grad_norm": 0.03353433459515292, "learning_rate": 3.7906684863199266e-05, "epoch": 0.5907055428400836, "step": 1484}, {"loss": 0.361624151468277, "grad_norm": 0.039017201024482466, "learning_rate": 3.784415251083932e-05, "epoch": 0.5911035923972534, "step": 1485}, {"loss": 0.36183881759643555, "grad_norm": 0.035374931390953594, "learning_rate": 3.778164035952594e-05, "epoch": 0.5915016419544233, "step": 1486}, {"loss": 0.39573267102241516, "grad_norm": 0.03544066138243923, "learning_rate": 3.771914851314417e-05, "epoch": 0.5918996915115932, "step": 1487}, {"loss": 0.34043145179748535, "grad_norm": 0.03726188186482645, "learning_rate": 3.765667707554531e-05, "epoch": 0.5922977410687631, "step": 1488}, {"loss": 0.3690953850746155, "grad_norm": 0.03257512486465629, "learning_rate": 3.759422615054678e-05, "epoch": 0.5926957906259329, "step": 1489}, {"loss": 0.3592092990875244, "grad_norm": 0.03711157742518628, "learning_rate": 3.753179584193187e-05, "epoch": 0.5930938401831028, "step": 1490}, {"loss": 0.40755990147590637, "grad_norm": 0.03792724629626637, "learning_rate": 3.7469386253449624e-05, "epoch": 0.5934918897402727, "step": 1491}, {"loss": 0.3508398234844208, "grad_norm": 0.03495211945861511, "learning_rate": 3.740699748881465e-05, "epoch": 0.5938899392974425, "step": 1492}, {"loss": 0.380648672580719, "grad_norm": 0.03959202141516316, "learning_rate": 3.7344629651706974e-05, "epoch": 0.5942879888546124, "step": 1493}, {"loss": 0.400083065032959, "grad_norm": 0.03804064786078234, "learning_rate": 3.728228284577181e-05, "epoch": 0.5946860384117822, "step": 1494}, {"loss": 0.36677661538124084, "grad_norm": 0.037914416870515034, "learning_rate": 3.7219957174619435e-05, "epoch": 0.5950840879689522, "step": 1495}, {"loss": 0.3529689908027649, "grad_norm": 0.07573885822402493, "learning_rate": 3.7157652741825e-05, "epoch": 0.595482137526122, "step": 1496}, {"loss": 0.4228888750076294, "grad_norm": 0.04235904024600779, "learning_rate": 3.709536965092837e-05, "epoch": 0.5958801870832918, "step": 1497}, {"loss": 0.3371814489364624, "grad_norm": 0.03611832149629748, "learning_rate": 3.703310800543397e-05, "epoch": 0.5962782366404618, "step": 1498}, {"loss": 0.36648690700531006, "grad_norm": 0.040105455498882114, "learning_rate": 3.6970867908810494e-05, "epoch": 0.5966762861976316, "step": 1499}, {"loss": 0.3789559304714203, "grad_norm": 0.03652614012493364, "learning_rate": 3.6908649464490915e-05, "epoch": 0.5970743357548015, "step": 1500}, {"loss": 0.33035507798194885, "grad_norm": 0.03679032384555477, "learning_rate": 3.6846452775872194e-05, "epoch": 0.5974723853119713, "step": 1501}, {"loss": 0.3347591459751129, "grad_norm": 0.0351337342945087, "learning_rate": 3.6784277946315146e-05, "epoch": 0.5978704348691412, "step": 1502}, {"loss": 0.39417311549186707, "grad_norm": 0.03567960250480377, "learning_rate": 3.672212507914423e-05, "epoch": 0.5982684844263111, "step": 1503}, {"loss": 0.3518398106098175, "grad_norm": 0.03936377407657785, "learning_rate": 3.665999427764745e-05, "epoch": 0.5986665339834809, "step": 1504}, {"loss": 0.3673250079154968, "grad_norm": 0.03648141247884105, "learning_rate": 3.6597885645076104e-05, "epoch": 0.5990645835406508, "step": 1505}, {"loss": 0.39645349979400635, "grad_norm": 0.03890665158294902, "learning_rate": 3.653579928464467e-05, "epoch": 0.5994626330978207, "step": 1506}, {"loss": 0.3527374863624573, "grad_norm": 0.03708293191783425, "learning_rate": 3.647373529953063e-05, "epoch": 0.5998606826549906, "step": 1507}, {"loss": 0.331756055355072, "grad_norm": 0.04028776910701449, "learning_rate": 3.641169379287421e-05, "epoch": 0.6002587322121604, "step": 1508}, {"loss": 0.3521445691585541, "grad_norm": 0.035231004949449626, "learning_rate": 3.634967486777836e-05, "epoch": 0.6006567817693302, "step": 1509}, {"loss": 0.34621310234069824, "grad_norm": 0.03831500255393537, "learning_rate": 3.6287678627308475e-05, "epoch": 0.6010548313265002, "step": 1510}, {"loss": 0.34019824862480164, "grad_norm": 0.03755052936650233, "learning_rate": 3.6225705174492245e-05, "epoch": 0.60145288088367, "step": 1511}, {"loss": 0.3892752528190613, "grad_norm": 0.0405593588996848, "learning_rate": 3.616375461231949e-05, "epoch": 0.6018509304408399, "step": 1512}, {"loss": 0.3945283889770508, "grad_norm": 0.03775851013133414, "learning_rate": 3.6101827043741995e-05, "epoch": 0.6022489799980097, "step": 1513}, {"loss": 0.3882424235343933, "grad_norm": 0.041772825503690124, "learning_rate": 3.603992257167334e-05, "epoch": 0.6026470295551796, "step": 1514}, {"loss": 0.35585546493530273, "grad_norm": 0.04168913647349925, "learning_rate": 3.59780412989887e-05, "epoch": 0.6030450791123495, "step": 1515}, {"loss": 0.3438390791416168, "grad_norm": 0.03643807160421366, "learning_rate": 3.5916183328524724e-05, "epoch": 0.6034431286695193, "step": 1516}, {"loss": 0.3499404788017273, "grad_norm": 0.03631478685617716, "learning_rate": 3.5854348763079336e-05, "epoch": 0.6038411782266893, "step": 1517}, {"loss": 0.3850410580635071, "grad_norm": 0.03567855698757955, "learning_rate": 3.5792537705411527e-05, "epoch": 0.6042392277838591, "step": 1518}, {"loss": 0.3375018239021301, "grad_norm": 0.03555703609457907, "learning_rate": 3.573075025824128e-05, "epoch": 0.604637277341029, "step": 1519}, {"loss": 0.3942347764968872, "grad_norm": 0.03753846993691861, "learning_rate": 3.566898652424927e-05, "epoch": 0.6050353268981988, "step": 1520}, {"loss": 0.37369218468666077, "grad_norm": 0.03871875835530215, "learning_rate": 3.560724660607681e-05, "epoch": 0.6054333764553687, "step": 1521}, {"loss": 0.38148263096809387, "grad_norm": 0.040259500832901654, "learning_rate": 3.5545530606325654e-05, "epoch": 0.6058314260125386, "step": 1522}, {"loss": 0.36272111535072327, "grad_norm": 0.043551939369641614, "learning_rate": 3.5483838627557764e-05, "epoch": 0.6062294755697084, "step": 1523}, {"loss": 0.3532043993473053, "grad_norm": 0.03628937325287193, "learning_rate": 3.542217077229521e-05, "epoch": 0.6066275251268783, "step": 1524}, {"loss": 0.36282017827033997, "grad_norm": 0.03669380211505415, "learning_rate": 3.536052714301995e-05, "epoch": 0.6070255746840482, "step": 1525}, {"loss": 0.3945884704589844, "grad_norm": 0.03902849666647598, "learning_rate": 3.52989078421737e-05, "epoch": 0.607423624241218, "step": 1526}, {"loss": 0.3780196011066437, "grad_norm": 0.037881169096742474, "learning_rate": 3.5237312972157754e-05, "epoch": 0.6078216737983879, "step": 1527}, {"loss": 0.39305299520492554, "grad_norm": 0.037709032514803904, "learning_rate": 3.5175742635332774e-05, "epoch": 0.6082197233555577, "step": 1528}, {"loss": 0.35779374837875366, "grad_norm": 0.03882350531998952, "learning_rate": 3.5114196934018703e-05, "epoch": 0.6086177729127277, "step": 1529}, {"loss": 0.39719584584236145, "grad_norm": 0.03992942096147537, "learning_rate": 3.505267597049447e-05, "epoch": 0.6090158224698975, "step": 1530}, {"loss": 0.3563001751899719, "grad_norm": 0.0378370769735722, "learning_rate": 3.499117984699797e-05, "epoch": 0.6094138720270673, "step": 1531}, {"loss": 0.3306661546230316, "grad_norm": 0.03527414884381917, "learning_rate": 3.492970866572577e-05, "epoch": 0.6098119215842372, "step": 1532}, {"loss": 0.37296903133392334, "grad_norm": 0.039090867720199006, "learning_rate": 3.4868262528833e-05, "epoch": 0.6102099711414071, "step": 1533}, {"loss": 0.36312398314476013, "grad_norm": 0.042618146079933715, "learning_rate": 3.480684153843319e-05, "epoch": 0.610608020698577, "step": 1534}, {"loss": 0.3633597195148468, "grad_norm": 0.0482188642624782, "learning_rate": 3.474544579659803e-05, "epoch": 0.6110060702557468, "step": 1535}, {"loss": 0.3678963780403137, "grad_norm": 0.044885371919144726, "learning_rate": 3.468407540535731e-05, "epoch": 0.6114041198129168, "step": 1536}, {"loss": 0.3683526813983917, "grad_norm": 0.03674157357709337, "learning_rate": 3.462273046669865e-05, "epoch": 0.6118021693700866, "step": 1537}, {"loss": 0.3654821813106537, "grad_norm": 0.040080800249770766, "learning_rate": 3.4561411082567383e-05, "epoch": 0.6122002189272564, "step": 1538}, {"loss": 0.4040085971355438, "grad_norm": 0.03726364653999491, "learning_rate": 3.450011735486641e-05, "epoch": 0.6125982684844263, "step": 1539}, {"loss": 0.39234787225723267, "grad_norm": 0.03912638895285928, "learning_rate": 3.4438849385455916e-05, "epoch": 0.6129963180415962, "step": 1540}, {"loss": 0.38392361998558044, "grad_norm": 0.03853862250208718, "learning_rate": 3.437760727615337e-05, "epoch": 0.6133943675987661, "step": 1541}, {"loss": 0.3920736312866211, "grad_norm": 0.03791830418633227, "learning_rate": 3.4316391128733197e-05, "epoch": 0.6137924171559359, "step": 1542}, {"loss": 0.3739506006240845, "grad_norm": 0.03989873595392917, "learning_rate": 3.42552010449267e-05, "epoch": 0.6141904667131057, "step": 1543}, {"loss": 0.33461636304855347, "grad_norm": 0.032932552706669456, "learning_rate": 3.419403712642188e-05, "epoch": 0.6145885162702757, "step": 1544}, {"loss": 0.36379361152648926, "grad_norm": 0.03907281049220056, "learning_rate": 3.4132899474863256e-05, "epoch": 0.6149865658274455, "step": 1545}, {"loss": 0.342647910118103, "grad_norm": 0.03938063708476495, "learning_rate": 3.407178819185167e-05, "epoch": 0.6153846153846154, "step": 1546}, {"loss": 0.3550448417663574, "grad_norm": 0.04289872150536001, "learning_rate": 3.401070337894419e-05, "epoch": 0.6157826649417852, "step": 1547}, {"loss": 0.36354321241378784, "grad_norm": 0.040161397969889295, "learning_rate": 3.394964513765383e-05, "epoch": 0.6161807144989552, "step": 1548}, {"loss": 0.3678874671459198, "grad_norm": 0.038732554412765886, "learning_rate": 3.388861356944949e-05, "epoch": 0.616578764056125, "step": 1549}, {"loss": 0.37185338139533997, "grad_norm": 0.04017127650620364, "learning_rate": 3.382760877575575e-05, "epoch": 0.6169768136132948, "step": 1550}, {"loss": 0.37030068039894104, "grad_norm": 0.03723158437192068, "learning_rate": 3.376663085795268e-05, "epoch": 0.6173748631704647, "step": 1551}, {"loss": 0.39511293172836304, "grad_norm": 0.03999366868102315, "learning_rate": 3.370567991737567e-05, "epoch": 0.6177729127276346, "step": 1552}, {"loss": 0.39594030380249023, "grad_norm": 0.041330903650697956, "learning_rate": 3.364475605531532e-05, "epoch": 0.6181709622848045, "step": 1553}, {"loss": 0.40123605728149414, "grad_norm": 0.04074311244044744, "learning_rate": 3.358385937301718e-05, "epoch": 0.6185690118419743, "step": 1554}, {"loss": 0.35023775696754456, "grad_norm": 0.03647155435852887, "learning_rate": 3.352298997168167e-05, "epoch": 0.6189670613991441, "step": 1555}, {"loss": 0.3917219638824463, "grad_norm": 0.041163558926768436, "learning_rate": 3.3462147952463865e-05, "epoch": 0.6193651109563141, "step": 1556}, {"loss": 0.39477574825286865, "grad_norm": 0.04245973016723136, "learning_rate": 3.340133341647333e-05, "epoch": 0.6197631605134839, "step": 1557}, {"loss": 0.3689468204975128, "grad_norm": 0.04321521860059687, "learning_rate": 3.334054646477395e-05, "epoch": 0.6201612100706538, "step": 1558}, {"loss": 0.3456647992134094, "grad_norm": 0.03839971885342294, "learning_rate": 3.3279787198383773e-05, "epoch": 0.6205592596278237, "step": 1559}, {"loss": 0.3678511679172516, "grad_norm": 0.040415470249992494, "learning_rate": 3.321905571827485e-05, "epoch": 0.6209573091849935, "step": 1560}, {"loss": 0.37332668900489807, "grad_norm": 0.03738370053840331, "learning_rate": 3.315835212537305e-05, "epoch": 0.6213553587421634, "step": 1561}, {"loss": 0.3904413878917694, "grad_norm": 0.03773076455174505, "learning_rate": 3.309767652055793e-05, "epoch": 0.6217534082993332, "step": 1562}, {"loss": 0.39470797777175903, "grad_norm": 0.038464671372926404, "learning_rate": 3.303702900466244e-05, "epoch": 0.6221514578565032, "step": 1563}, {"loss": 0.39453795552253723, "grad_norm": 0.03948133227955216, "learning_rate": 3.297640967847295e-05, "epoch": 0.622549507413673, "step": 1564}, {"loss": 0.3565846383571625, "grad_norm": 0.038217101541729186, "learning_rate": 3.291581864272893e-05, "epoch": 0.6229475569708429, "step": 1565}, {"loss": 0.36390218138694763, "grad_norm": 0.03891460962723128, "learning_rate": 3.2855255998122856e-05, "epoch": 0.6233456065280127, "step": 1566}, {"loss": 0.3703502416610718, "grad_norm": 0.03782075077379377, "learning_rate": 3.279472184530004e-05, "epoch": 0.6237436560851826, "step": 1567}, {"loss": 0.38220280408859253, "grad_norm": 0.04018078556748722, "learning_rate": 3.2734216284858386e-05, "epoch": 0.6241417056423525, "step": 1568}, {"loss": 0.3969442546367645, "grad_norm": 0.03852001477777684, "learning_rate": 3.267373941734836e-05, "epoch": 0.6245397551995223, "step": 1569}, {"loss": 0.3836630880832672, "grad_norm": 0.04404449723611789, "learning_rate": 3.2613291343272676e-05, "epoch": 0.6249378047566923, "step": 1570}, {"loss": 0.39173558354377747, "grad_norm": 0.04131712792533609, "learning_rate": 3.2552872163086234e-05, "epoch": 0.6253358543138621, "step": 1571}, {"loss": 0.37002724409103394, "grad_norm": 0.036511912987569305, "learning_rate": 3.249248197719592e-05, "epoch": 0.6257339038710319, "step": 1572}, {"loss": 0.36519819498062134, "grad_norm": 0.03575856932456499, "learning_rate": 3.243212088596042e-05, "epoch": 0.6261319534282018, "step": 1573}, {"loss": 0.3384925127029419, "grad_norm": 0.03416847150409553, "learning_rate": 3.2371788989690075e-05, "epoch": 0.6265300029853716, "step": 1574}, {"loss": 0.3743298649787903, "grad_norm": 0.038988564781299626, "learning_rate": 3.23114863886467e-05, "epoch": 0.6269280525425416, "step": 1575}, {"loss": 0.36584654450416565, "grad_norm": 0.03862985801273126, "learning_rate": 3.225121318304345e-05, "epoch": 0.6273261020997114, "step": 1576}, {"loss": 0.3797476589679718, "grad_norm": 0.038827042073731596, "learning_rate": 3.219096947304461e-05, "epoch": 0.6277241516568813, "step": 1577}, {"loss": 0.34756967425346375, "grad_norm": 0.0379452006146104, "learning_rate": 3.2130755358765444e-05, "epoch": 0.6281222012140512, "step": 1578}, {"loss": 0.36661046743392944, "grad_norm": 0.03903212352534353, "learning_rate": 3.2070570940272066e-05, "epoch": 0.628520250771221, "step": 1579}, {"loss": 0.40112048387527466, "grad_norm": 0.03845690405519655, "learning_rate": 3.201041631758119e-05, "epoch": 0.6289183003283909, "step": 1580}, {"loss": 0.347009539604187, "grad_norm": 0.03830872914923669, "learning_rate": 3.1950291590660055e-05, "epoch": 0.6293163498855607, "step": 1581}, {"loss": 0.37663930654525757, "grad_norm": 0.04040989568794574, "learning_rate": 3.189019685942619e-05, "epoch": 0.6297143994427307, "step": 1582}, {"loss": 0.3749157190322876, "grad_norm": 0.042622374991565716, "learning_rate": 3.1830132223747297e-05, "epoch": 0.6301124489999005, "step": 1583}, {"loss": 0.3583713471889496, "grad_norm": 0.04088215952390817, "learning_rate": 3.177009778344109e-05, "epoch": 0.6305104985570703, "step": 1584}, {"loss": 0.30990639328956604, "grad_norm": 0.04057049551919928, "learning_rate": 3.171009363827502e-05, "epoch": 0.6309085481142402, "step": 1585}, {"loss": 0.34537556767463684, "grad_norm": 0.03749521217532599, "learning_rate": 3.165011988796626e-05, "epoch": 0.6313065976714101, "step": 1586}, {"loss": 0.34633275866508484, "grad_norm": 0.03708453377063508, "learning_rate": 3.159017663218144e-05, "epoch": 0.63170464722858, "step": 1587}, {"loss": 0.3402025103569031, "grad_norm": 0.034722050956256745, "learning_rate": 3.1530263970536555e-05, "epoch": 0.6321026967857498, "step": 1588}, {"loss": 0.3734888732433319, "grad_norm": 0.038511372392692166, "learning_rate": 3.1470382002596705e-05, "epoch": 0.6325007463429196, "step": 1589}, {"loss": 0.4303601086139679, "grad_norm": 0.046042802074106576, "learning_rate": 3.141053082787602e-05, "epoch": 0.6328987959000896, "step": 1590}, {"loss": 0.41827505826950073, "grad_norm": 0.04373274725345982, "learning_rate": 3.135071054583745e-05, "epoch": 0.6332968454572594, "step": 1591}, {"loss": 0.45960158109664917, "grad_norm": 0.04118875352810413, "learning_rate": 3.129092125589258e-05, "epoch": 0.6336948950144293, "step": 1592}, {"loss": 0.389235258102417, "grad_norm": 0.04090707042322386, "learning_rate": 3.123116305740151e-05, "epoch": 0.6340929445715991, "step": 1593}, {"loss": 0.34899580478668213, "grad_norm": 0.04013298017548738, "learning_rate": 3.11714360496727e-05, "epoch": 0.6344909941287691, "step": 1594}, {"loss": 0.42754441499710083, "grad_norm": 0.039714146904581596, "learning_rate": 3.1111740331962714e-05, "epoch": 0.6348890436859389, "step": 1595}, {"loss": 0.3744359016418457, "grad_norm": 0.03831672504397032, "learning_rate": 3.105207600347617e-05, "epoch": 0.6352870932431087, "step": 1596}, {"loss": 0.3304743766784668, "grad_norm": 0.03367130975908512, "learning_rate": 3.099244316336551e-05, "epoch": 0.6356851428002787, "step": 1597}, {"loss": 0.40924203395843506, "grad_norm": 0.03982999420322531, "learning_rate": 3.093284191073082e-05, "epoch": 0.6360831923574485, "step": 1598}, {"loss": 0.3741125464439392, "grad_norm": 0.03791626740192498, "learning_rate": 3.0873272344619717e-05, "epoch": 0.6364812419146184, "step": 1599}, {"loss": 0.3318183422088623, "grad_norm": 0.03746769416217962, "learning_rate": 3.081373456402717e-05, "epoch": 0.6368792914717882, "step": 1600}, {"loss": 0.37335386872291565, "grad_norm": 0.039439051110125835, "learning_rate": 3.07542286678953e-05, "epoch": 0.637277341028958, "step": 1601}, {"loss": 0.34863734245300293, "grad_norm": 0.0405631967554156, "learning_rate": 3.069475475511326e-05, "epoch": 0.637675390586128, "step": 1602}, {"loss": 0.40071848034858704, "grad_norm": 0.04345918886950121, "learning_rate": 3.063531292451705e-05, "epoch": 0.6380734401432978, "step": 1603}, {"loss": 0.3780031204223633, "grad_norm": 0.03718189233277887, "learning_rate": 3.0575903274889325e-05, "epoch": 0.6384714897004677, "step": 1604}, {"loss": 0.3533087372779846, "grad_norm": 0.03542449672451498, "learning_rate": 3.0516525904959325e-05, "epoch": 0.6388695392576376, "step": 1605}, {"loss": 0.36983031034469604, "grad_norm": 0.041705070003804634, "learning_rate": 3.0457180913402562e-05, "epoch": 0.6392675888148075, "step": 1606}, {"loss": 0.3444669544696808, "grad_norm": 0.03670847799625615, "learning_rate": 3.03978683988408e-05, "epoch": 0.6396656383719773, "step": 1607}, {"loss": 0.378846675157547, "grad_norm": 0.03988961839932574, "learning_rate": 3.0338588459841832e-05, "epoch": 0.6400636879291471, "step": 1608}, {"loss": 0.36015260219573975, "grad_norm": 0.03672630115284777, "learning_rate": 3.02793411949193e-05, "epoch": 0.6404617374863171, "step": 1609}, {"loss": 0.3396000266075134, "grad_norm": 0.03737988618205114, "learning_rate": 3.0220126702532515e-05, "epoch": 0.6408597870434869, "step": 1610}, {"loss": 0.3678939938545227, "grad_norm": 0.03858214822973101, "learning_rate": 3.016094508108639e-05, "epoch": 0.6412578366006568, "step": 1611}, {"loss": 0.3685268759727478, "grad_norm": 0.0372878862365645, "learning_rate": 3.010179642893116e-05, "epoch": 0.6416558861578266, "step": 1612}, {"loss": 0.34108203649520874, "grad_norm": 0.03798959018372544, "learning_rate": 3.0042680844362305e-05, "epoch": 0.6420539357149965, "step": 1613}, {"loss": 0.33471959829330444, "grad_norm": 0.036655004522781885, "learning_rate": 2.998359842562033e-05, "epoch": 0.6424519852721664, "step": 1614}, {"loss": 0.36279311776161194, "grad_norm": 0.0347340233206809, "learning_rate": 2.9924549270890622e-05, "epoch": 0.6428500348293362, "step": 1615}, {"loss": 0.3443152606487274, "grad_norm": 0.03585937791347011, "learning_rate": 2.9865533478303308e-05, "epoch": 0.6432480843865062, "step": 1616}, {"loss": 0.32344427704811096, "grad_norm": 0.037109775484348226, "learning_rate": 2.980655114593306e-05, "epoch": 0.643646133943676, "step": 1617}, {"loss": 0.33458808064460754, "grad_norm": 0.038157400578462154, "learning_rate": 2.9747602371798944e-05, "epoch": 0.6440441835008458, "step": 1618}, {"loss": 0.3992919623851776, "grad_norm": 0.039216147724852664, "learning_rate": 2.9688687253864265e-05, "epoch": 0.6444422330580157, "step": 1619}, {"loss": 0.3881251811981201, "grad_norm": 0.03822767827009527, "learning_rate": 2.962980589003638e-05, "epoch": 0.6448402826151856, "step": 1620}, {"loss": 0.3968873620033264, "grad_norm": 0.04180868665847053, "learning_rate": 2.9570958378166567e-05, "epoch": 0.6452383321723555, "step": 1621}, {"loss": 0.345795214176178, "grad_norm": 0.04116907850046903, "learning_rate": 2.9512144816049846e-05, "epoch": 0.6456363817295253, "step": 1622}, {"loss": 0.37222158908843994, "grad_norm": 0.0416834265203733, "learning_rate": 2.945336530142482e-05, "epoch": 0.6460344312866952, "step": 1623}, {"loss": 0.41384387016296387, "grad_norm": 0.04367976006475424, "learning_rate": 2.939461993197351e-05, "epoch": 0.6464324808438651, "step": 1624}, {"loss": 0.3663254678249359, "grad_norm": 0.03929461378683091, "learning_rate": 2.9335908805321182e-05, "epoch": 0.6468305304010349, "step": 1625}, {"loss": 0.3564280867576599, "grad_norm": 0.036117962123283255, "learning_rate": 2.9277232019036228e-05, "epoch": 0.6472285799582048, "step": 1626}, {"loss": 0.3596256375312805, "grad_norm": 0.03878482431862578, "learning_rate": 2.9218589670629898e-05, "epoch": 0.6476266295153746, "step": 1627}, {"loss": 0.41412386298179626, "grad_norm": 0.04113202095917963, "learning_rate": 2.9159981857556285e-05, "epoch": 0.6480246790725446, "step": 1628}, {"loss": 0.35877639055252075, "grad_norm": 0.03652215542915897, "learning_rate": 2.91014086772121e-05, "epoch": 0.6484227286297144, "step": 1629}, {"loss": 0.3968539237976074, "grad_norm": 0.038875127557623256, "learning_rate": 2.9042870226936424e-05, "epoch": 0.6488207781868842, "step": 1630}, {"loss": 0.39990556240081787, "grad_norm": 0.0416288848941642, "learning_rate": 2.898436660401067e-05, "epoch": 0.6492188277440541, "step": 1631}, {"loss": 0.35807764530181885, "grad_norm": 0.03468819686023497, "learning_rate": 2.892589790565839e-05, "epoch": 0.649616877301224, "step": 1632}, {"loss": 0.38522371649742126, "grad_norm": 0.040369046334535495, "learning_rate": 2.886746422904506e-05, "epoch": 0.6500149268583939, "step": 1633}, {"loss": 0.34281712770462036, "grad_norm": 0.03903313883289965, "learning_rate": 2.8809065671277967e-05, "epoch": 0.6504129764155637, "step": 1634}, {"loss": 0.38619011640548706, "grad_norm": 0.038562812993713586, "learning_rate": 2.8750702329406066e-05, "epoch": 0.6508110259727337, "step": 1635}, {"loss": 0.41080421209335327, "grad_norm": 0.04331512500132177, "learning_rate": 2.8692374300419732e-05, "epoch": 0.6512090755299035, "step": 1636}, {"loss": 0.3204422891139984, "grad_norm": 0.03729568093283471, "learning_rate": 2.8634081681250702e-05, "epoch": 0.6516071250870733, "step": 1637}, {"loss": 0.364033579826355, "grad_norm": 0.03599924359300843, "learning_rate": 2.8575824568771857e-05, "epoch": 0.6520051746442432, "step": 1638}, {"loss": 0.33226799964904785, "grad_norm": 0.03464879160965999, "learning_rate": 2.851760305979708e-05, "epoch": 0.6524032242014131, "step": 1639}, {"loss": 0.413943350315094, "grad_norm": 0.03970817627837066, "learning_rate": 2.8459417251081034e-05, "epoch": 0.652801273758583, "step": 1640}, {"loss": 0.375465989112854, "grad_norm": 0.037484832301234755, "learning_rate": 2.840126723931912e-05, "epoch": 0.6531993233157528, "step": 1641}, {"loss": 0.34531429409980774, "grad_norm": 0.03856520096725001, "learning_rate": 2.8343153121147214e-05, "epoch": 0.6535973728729226, "step": 1642}, {"loss": 0.4065207242965698, "grad_norm": 0.04268193942213485, "learning_rate": 2.8285074993141564e-05, "epoch": 0.6539954224300926, "step": 1643}, {"loss": 0.3789072632789612, "grad_norm": 0.038092913498507, "learning_rate": 2.8227032951818587e-05, "epoch": 0.6543934719872624, "step": 1644}, {"loss": 0.34563493728637695, "grad_norm": 0.03777844420997581, "learning_rate": 2.8169027093634747e-05, "epoch": 0.6547915215444323, "step": 1645}, {"loss": 0.39303678274154663, "grad_norm": 0.04272968653855581, "learning_rate": 2.811105751498637e-05, "epoch": 0.6551895711016021, "step": 1646}, {"loss": 0.3538144826889038, "grad_norm": 0.03759883437350915, "learning_rate": 2.805312431220949e-05, "epoch": 0.655587620658772, "step": 1647}, {"loss": 0.37197113037109375, "grad_norm": 0.04234959949433025, "learning_rate": 2.799522758157971e-05, "epoch": 0.6559856702159419, "step": 1648}, {"loss": 0.3325093388557434, "grad_norm": 0.03713213123974754, "learning_rate": 2.7937367419311966e-05, "epoch": 0.6563837197731117, "step": 1649}, {"loss": 0.39621996879577637, "grad_norm": 0.04230754938805555, "learning_rate": 2.787954392156048e-05, "epoch": 0.6567817693302817, "step": 1650}, {"loss": 0.4206045866012573, "grad_norm": 0.04357387870965192, "learning_rate": 2.7821757184418552e-05, "epoch": 0.6571798188874515, "step": 1651}, {"loss": 0.37576785683631897, "grad_norm": 0.037839718832780166, "learning_rate": 2.7764007303918304e-05, "epoch": 0.6575778684446214, "step": 1652}, {"loss": 0.3686347007751465, "grad_norm": 0.03928339750997328, "learning_rate": 2.770629437603072e-05, "epoch": 0.6579759180017912, "step": 1653}, {"loss": 0.30996572971343994, "grad_norm": 0.038921125893543615, "learning_rate": 2.76486184966653e-05, "epoch": 0.658373967558961, "step": 1654}, {"loss": 0.350904643535614, "grad_norm": 0.03663348650482256, "learning_rate": 2.7590979761670003e-05, "epoch": 0.658772017116131, "step": 1655}, {"loss": 0.3881462514400482, "grad_norm": 0.039852553303285444, "learning_rate": 2.753337826683107e-05, "epoch": 0.6591700666733008, "step": 1656}, {"loss": 0.37154725193977356, "grad_norm": 0.03999609791530227, "learning_rate": 2.747581410787282e-05, "epoch": 0.6595681162304707, "step": 1657}, {"loss": 0.3547547459602356, "grad_norm": 0.03998553123098506, "learning_rate": 2.7418287380457563e-05, "epoch": 0.6599661657876406, "step": 1658}, {"loss": 0.3804132342338562, "grad_norm": 0.03905061088626077, "learning_rate": 2.736079818018538e-05, "epoch": 0.6603642153448104, "step": 1659}, {"loss": 0.387360155582428, "grad_norm": 0.036775259264980785, "learning_rate": 2.7303346602594e-05, "epoch": 0.6607622649019803, "step": 1660}, {"loss": 0.3373270034790039, "grad_norm": 0.03675431777641189, "learning_rate": 2.724593274315864e-05, "epoch": 0.6611603144591501, "step": 1661}, {"loss": 0.3647826313972473, "grad_norm": 0.04081826255313839, "learning_rate": 2.718855669729179e-05, "epoch": 0.6615583640163201, "step": 1662}, {"loss": 0.3757597804069519, "grad_norm": 0.03990384853247548, "learning_rate": 2.713121856034314e-05, "epoch": 0.6619564135734899, "step": 1663}, {"loss": 0.3916475474834442, "grad_norm": 0.03746461286938681, "learning_rate": 2.7073918427599377e-05, "epoch": 0.6623544631306598, "step": 1664}, {"loss": 0.4018011689186096, "grad_norm": 0.04268776466886185, "learning_rate": 2.7016656394284023e-05, "epoch": 0.6627525126878296, "step": 1665}, {"loss": 0.3656764626502991, "grad_norm": 0.04058705656346989, "learning_rate": 2.695943255555729e-05, "epoch": 0.6631505622449995, "step": 1666}, {"loss": 0.3515140116214752, "grad_norm": 0.0357434377227144, "learning_rate": 2.6902247006515913e-05, "epoch": 0.6635486118021694, "step": 1667}, {"loss": 0.35783851146698, "grad_norm": 0.03956662636425285, "learning_rate": 2.6845099842193007e-05, "epoch": 0.6639466613593392, "step": 1668}, {"loss": 0.39122915267944336, "grad_norm": 0.041204942257441715, "learning_rate": 2.6787991157557897e-05, "epoch": 0.6643447109165092, "step": 1669}, {"loss": 0.38768434524536133, "grad_norm": 0.040407565579399854, "learning_rate": 2.673092104751591e-05, "epoch": 0.664742760473679, "step": 1670}, {"loss": 0.35812705755233765, "grad_norm": 0.03793013409234233, "learning_rate": 2.667388960690835e-05, "epoch": 0.6651408100308488, "step": 1671}, {"loss": 0.3779442608356476, "grad_norm": 0.03875937337723923, "learning_rate": 2.6616896930512203e-05, "epoch": 0.6655388595880187, "step": 1672}, {"loss": 0.3750404417514801, "grad_norm": 0.04057793001760471, "learning_rate": 2.6559943113040054e-05, "epoch": 0.6659369091451885, "step": 1673}, {"loss": 0.3437681794166565, "grad_norm": 0.03890876778855809, "learning_rate": 2.650302824913993e-05, "epoch": 0.6663349587023585, "step": 1674}, {"loss": 0.39865416288375854, "grad_norm": 0.03767242744762373, "learning_rate": 2.6446152433395056e-05, "epoch": 0.6667330082595283, "step": 1675}, {"loss": 0.3895580768585205, "grad_norm": 0.04067006073490802, "learning_rate": 2.638931576032384e-05, "epoch": 0.6671310578166981, "step": 1676}, {"loss": 0.3503094017505646, "grad_norm": 0.04008521504447105, "learning_rate": 2.6332518324379607e-05, "epoch": 0.6675291073738681, "step": 1677}, {"loss": 0.33471161127090454, "grad_norm": 0.039136183494859, "learning_rate": 2.627576021995049e-05, "epoch": 0.6679271569310379, "step": 1678}, {"loss": 0.33756086230278015, "grad_norm": 0.037503464567359285, "learning_rate": 2.6219041541359248e-05, "epoch": 0.6683252064882078, "step": 1679}, {"loss": 0.39282456040382385, "grad_norm": 0.039823603116167075, "learning_rate": 2.616236238286311e-05, "epoch": 0.6687232560453776, "step": 1680}, {"loss": 0.3672478497028351, "grad_norm": 0.035552052851914685, "learning_rate": 2.6105722838653657e-05, "epoch": 0.6691213056025476, "step": 1681}, {"loss": 0.32578641176223755, "grad_norm": 0.036720854906170115, "learning_rate": 2.604912300285662e-05, "epoch": 0.6695193551597174, "step": 1682}, {"loss": 0.3332767188549042, "grad_norm": 0.03772152228993651, "learning_rate": 2.599256296953174e-05, "epoch": 0.6699174047168872, "step": 1683}, {"loss": 0.40556836128234863, "grad_norm": 0.042596469506744776, "learning_rate": 2.5936042832672647e-05, "epoch": 0.6703154542740571, "step": 1684}, {"loss": 0.3302568793296814, "grad_norm": 0.036112141745116254, "learning_rate": 2.5879562686206586e-05, "epoch": 0.670713503831227, "step": 1685}, {"loss": 0.38295313715934753, "grad_norm": 0.038502591278041176, "learning_rate": 2.5823122623994427e-05, "epoch": 0.6711115533883969, "step": 1686}, {"loss": 0.3812180459499359, "grad_norm": 0.03909292746018363, "learning_rate": 2.5766722739830396e-05, "epoch": 0.6715096029455667, "step": 1687}, {"loss": 0.34039992094039917, "grad_norm": 0.03808339124211496, "learning_rate": 2.5710363127441954e-05, "epoch": 0.6719076525027365, "step": 1688}, {"loss": 0.3877973258495331, "grad_norm": 0.0373043093448758, "learning_rate": 2.565404388048963e-05, "epoch": 0.6723057020599065, "step": 1689}, {"loss": 0.3427654504776001, "grad_norm": 0.0405969375200345, "learning_rate": 2.5597765092566893e-05, "epoch": 0.6727037516170763, "step": 1690}, {"loss": 0.3695580065250397, "grad_norm": 0.03816648775568835, "learning_rate": 2.5541526857199928e-05, "epoch": 0.6731018011742462, "step": 1691}, {"loss": 0.368442565202713, "grad_norm": 0.04231149733716185, "learning_rate": 2.548532926784757e-05, "epoch": 0.673499850731416, "step": 1692}, {"loss": 0.3761008381843567, "grad_norm": 0.03740900954204978, "learning_rate": 2.5429172417901093e-05, "epoch": 0.673897900288586, "step": 1693}, {"loss": 0.3637271523475647, "grad_norm": 0.03731058953622827, "learning_rate": 2.537305640068408e-05, "epoch": 0.6742959498457558, "step": 1694}, {"loss": 0.32387351989746094, "grad_norm": 0.03481709867451493, "learning_rate": 2.5316981309452252e-05, "epoch": 0.6746939994029256, "step": 1695}, {"loss": 0.37851011753082275, "grad_norm": 0.03870973098908709, "learning_rate": 2.526094723739333e-05, "epoch": 0.6750920489600956, "step": 1696}, {"loss": 0.3579971194267273, "grad_norm": 0.037454055176725995, "learning_rate": 2.5204954277626803e-05, "epoch": 0.6754900985172654, "step": 1697}, {"loss": 0.35260009765625, "grad_norm": 0.03942648350477656, "learning_rate": 2.514900252320391e-05, "epoch": 0.6758881480744353, "step": 1698}, {"loss": 0.40692758560180664, "grad_norm": 0.042844857614632935, "learning_rate": 2.5093092067107405e-05, "epoch": 0.6762861976316051, "step": 1699}, {"loss": 0.36663928627967834, "grad_norm": 0.040117412547146854, "learning_rate": 2.5037223002251374e-05, "epoch": 0.676684247188775, "step": 1700}, {"loss": 0.416032612323761, "grad_norm": 0.044639963004007244, "learning_rate": 2.4981395421481156e-05, "epoch": 0.6770822967459449, "step": 1701}, {"loss": 0.3344511389732361, "grad_norm": 0.037016211601389376, "learning_rate": 2.4925609417573127e-05, "epoch": 0.6774803463031147, "step": 1702}, {"loss": 0.3952987790107727, "grad_norm": 0.03748685949048807, "learning_rate": 2.486986508323457e-05, "epoch": 0.6778783958602846, "step": 1703}, {"loss": 0.4213525652885437, "grad_norm": 0.04364386572848637, "learning_rate": 2.4814162511103546e-05, "epoch": 0.6782764454174545, "step": 1704}, {"loss": 0.37974128127098083, "grad_norm": 0.0409673911853823, "learning_rate": 2.475850179374867e-05, "epoch": 0.6786744949746243, "step": 1705}, {"loss": 0.3649902939796448, "grad_norm": 0.03626869439064599, "learning_rate": 2.470288302366905e-05, "epoch": 0.6790725445317942, "step": 1706}, {"loss": 0.33662843704223633, "grad_norm": 0.036335782178534996, "learning_rate": 2.464730629329403e-05, "epoch": 0.679470594088964, "step": 1707}, {"loss": 0.3673788011074066, "grad_norm": 0.03855756683213281, "learning_rate": 2.4591771694983128e-05, "epoch": 0.679868643646134, "step": 1708}, {"loss": 0.3715640902519226, "grad_norm": 0.038891520489155416, "learning_rate": 2.453627932102585e-05, "epoch": 0.6802666932033038, "step": 1709}, {"loss": 0.3686440587043762, "grad_norm": 0.038820588747866085, "learning_rate": 2.448082926364151e-05, "epoch": 0.6806647427604737, "step": 1710}, {"loss": 0.33472296595573425, "grad_norm": 0.03643122844150385, "learning_rate": 2.442542161497911e-05, "epoch": 0.6810627923176436, "step": 1711}, {"loss": 0.3736708164215088, "grad_norm": 0.04131014935207968, "learning_rate": 2.437005646711719e-05, "epoch": 0.6814608418748134, "step": 1712}, {"loss": 0.41095271706581116, "grad_norm": 0.038659821742369205, "learning_rate": 2.431473391206361e-05, "epoch": 0.6818588914319833, "step": 1713}, {"loss": 0.4054298996925354, "grad_norm": 0.040969024914610734, "learning_rate": 2.4259454041755503e-05, "epoch": 0.6822569409891531, "step": 1714}, {"loss": 0.34205299615859985, "grad_norm": 0.0398918320205309, "learning_rate": 2.420421694805904e-05, "epoch": 0.6826549905463231, "step": 1715}, {"loss": 0.388895183801651, "grad_norm": 0.03857500195693195, "learning_rate": 2.414902272276931e-05, "epoch": 0.6830530401034929, "step": 1716}, {"loss": 0.3726249933242798, "grad_norm": 0.03987691714017685, "learning_rate": 2.4093871457610167e-05, "epoch": 0.6834510896606627, "step": 1717}, {"loss": 0.3942406177520752, "grad_norm": 0.03865157228738455, "learning_rate": 2.4038763244234063e-05, "epoch": 0.6838491392178326, "step": 1718}, {"loss": 0.3339681029319763, "grad_norm": 0.035798658731243126, "learning_rate": 2.3983698174221924e-05, "epoch": 0.6842471887750025, "step": 1719}, {"loss": 0.3689996600151062, "grad_norm": 0.03928338945804464, "learning_rate": 2.392867633908294e-05, "epoch": 0.6846452383321724, "step": 1720}, {"loss": 0.3709176480770111, "grad_norm": 0.04209875819920269, "learning_rate": 2.3873697830254478e-05, "epoch": 0.6850432878893422, "step": 1721}, {"loss": 0.3561776280403137, "grad_norm": 0.03772846074513767, "learning_rate": 2.3818762739101908e-05, "epoch": 0.6854413374465121, "step": 1722}, {"loss": 0.34399014711380005, "grad_norm": 0.037544456499196466, "learning_rate": 2.376387115691844e-05, "epoch": 0.685839387003682, "step": 1723}, {"loss": 0.3814377188682556, "grad_norm": 0.03641704452597759, "learning_rate": 2.3709023174924978e-05, "epoch": 0.6862374365608518, "step": 1724}, {"loss": 0.35302212834358215, "grad_norm": 0.03922260990246186, "learning_rate": 2.3654218884269963e-05, "epoch": 0.6866354861180217, "step": 1725}, {"loss": 0.3437385559082031, "grad_norm": 0.03773180179472373, "learning_rate": 2.359945837602925e-05, "epoch": 0.6870335356751915, "step": 1726}, {"loss": 0.3458910286426544, "grad_norm": 0.04049447373088136, "learning_rate": 2.3544741741205912e-05, "epoch": 0.6874315852323615, "step": 1727}, {"loss": 0.4071219563484192, "grad_norm": 0.040559784218231996, "learning_rate": 2.349006907073012e-05, "epoch": 0.6878296347895313, "step": 1728}, {"loss": 0.3845076560974121, "grad_norm": 0.0370524563680153, "learning_rate": 2.3435440455459003e-05, "epoch": 0.6882276843467011, "step": 1729}, {"loss": 0.36765724420547485, "grad_norm": 0.03788841913628884, "learning_rate": 2.3380855986176414e-05, "epoch": 0.688625733903871, "step": 1730}, {"loss": 0.33265209197998047, "grad_norm": 0.03846188176281366, "learning_rate": 2.3326315753592916e-05, "epoch": 0.6890237834610409, "step": 1731}, {"loss": 0.3394678235054016, "grad_norm": 0.03826766004460956, "learning_rate": 2.3271819848345518e-05, "epoch": 0.6894218330182108, "step": 1732}, {"loss": 0.3559405207633972, "grad_norm": 0.040018282843898925, "learning_rate": 2.32173683609976e-05, "epoch": 0.6898198825753806, "step": 1733}, {"loss": 0.39923346042633057, "grad_norm": 0.04015151757246137, "learning_rate": 2.3162961382038658e-05, "epoch": 0.6902179321325504, "step": 1734}, {"loss": 0.36918914318084717, "grad_norm": 0.04043985711404647, "learning_rate": 2.3108599001884284e-05, "epoch": 0.6906159816897204, "step": 1735}, {"loss": 0.35840827226638794, "grad_norm": 0.03745336670432929, "learning_rate": 2.305428131087594e-05, "epoch": 0.6910140312468902, "step": 1736}, {"loss": 0.4022548794746399, "grad_norm": 0.03791191187920208, "learning_rate": 2.300000839928081e-05, "epoch": 0.6914120808040601, "step": 1737}, {"loss": 0.33355429768562317, "grad_norm": 0.03628249734700951, "learning_rate": 2.2945780357291675e-05, "epoch": 0.69181013036123, "step": 1738}, {"loss": 0.38916605710983276, "grad_norm": 0.042979072513919477, "learning_rate": 2.2891597275026737e-05, "epoch": 0.6922081799183999, "step": 1739}, {"loss": 0.3490102291107178, "grad_norm": 0.04204021279440442, "learning_rate": 2.2837459242529498e-05, "epoch": 0.6926062294755697, "step": 1740}, {"loss": 0.38676226139068604, "grad_norm": 0.040507389079682384, "learning_rate": 2.278336634976859e-05, "epoch": 0.6930042790327395, "step": 1741}, {"loss": 0.35275959968566895, "grad_norm": 0.03960122132056328, "learning_rate": 2.2729318686637608e-05, "epoch": 0.6934023285899095, "step": 1742}, {"loss": 0.37321406602859497, "grad_norm": 0.038163345455713794, "learning_rate": 2.2675316342955e-05, "epoch": 0.6938003781470793, "step": 1743}, {"loss": 0.35742300748825073, "grad_norm": 0.03535690107863413, "learning_rate": 2.262135940846391e-05, "epoch": 0.6941984277042492, "step": 1744}, {"loss": 0.36755186319351196, "grad_norm": 0.037784372792109076, "learning_rate": 2.2567447972832013e-05, "epoch": 0.694596477261419, "step": 1745}, {"loss": 0.40731531381607056, "grad_norm": 0.041412666656851185, "learning_rate": 2.2513582125651356e-05, "epoch": 0.6949945268185889, "step": 1746}, {"loss": 0.4093080461025238, "grad_norm": 0.040733826922554596, "learning_rate": 2.245976195643825e-05, "epoch": 0.6953925763757588, "step": 1747}, {"loss": 0.3692610263824463, "grad_norm": 0.03627899178918495, "learning_rate": 2.240598755463309e-05, "epoch": 0.6957906259329286, "step": 1748}, {"loss": 0.3779873251914978, "grad_norm": 0.03761037017888518, "learning_rate": 2.2352259009600198e-05, "epoch": 0.6961886754900986, "step": 1749}, {"loss": 0.38563278317451477, "grad_norm": 0.0360897167800462, "learning_rate": 2.2298576410627703e-05, "epoch": 0.6965867250472684, "step": 1750}, {"loss": 0.37910592555999756, "grad_norm": 0.03897485715081174, "learning_rate": 2.2244939846927395e-05, "epoch": 0.6969847746044383, "step": 1751}, {"loss": 0.3545415997505188, "grad_norm": 0.03610626866492268, "learning_rate": 2.2191349407634497e-05, "epoch": 0.6973828241616081, "step": 1752}, {"loss": 0.3521507680416107, "grad_norm": 0.03736172066650655, "learning_rate": 2.2137805181807657e-05, "epoch": 0.697780873718778, "step": 1753}, {"loss": 0.3432219624519348, "grad_norm": 0.03900705972507811, "learning_rate": 2.208430725842869e-05, "epoch": 0.6981789232759479, "step": 1754}, {"loss": 0.40238815546035767, "grad_norm": 0.03709558842023848, "learning_rate": 2.2030855726402472e-05, "epoch": 0.6985769728331177, "step": 1755}, {"loss": 0.32320743799209595, "grad_norm": 0.03663081370514385, "learning_rate": 2.1977450674556755e-05, "epoch": 0.6989750223902876, "step": 1756}, {"loss": 0.36835384368896484, "grad_norm": 0.04275472175531619, "learning_rate": 2.1924092191642086e-05, "epoch": 0.6993730719474575, "step": 1757}, {"loss": 0.385633260011673, "grad_norm": 0.035095327675029535, "learning_rate": 2.1870780366331605e-05, "epoch": 0.6997711215046273, "step": 1758}, {"loss": 0.3805627226829529, "grad_norm": 0.03920911590671917, "learning_rate": 2.1817515287220917e-05, "epoch": 0.7001691710617972, "step": 1759}, {"loss": 0.3511718511581421, "grad_norm": 0.03540642795730413, "learning_rate": 2.1764297042827947e-05, "epoch": 0.700567220618967, "step": 1760}, {"loss": 0.389091432094574, "grad_norm": 0.04158143416325046, "learning_rate": 2.171112572159278e-05, "epoch": 0.700965270176137, "step": 1761}, {"loss": 0.4239938259124756, "grad_norm": 0.04063628404241661, "learning_rate": 2.1658001411877527e-05, "epoch": 0.7013633197333068, "step": 1762}, {"loss": 0.3346025049686432, "grad_norm": 0.03459313762102308, "learning_rate": 2.1604924201966193e-05, "epoch": 0.7017613692904766, "step": 1763}, {"loss": 0.3688023090362549, "grad_norm": 0.03827954015449633, "learning_rate": 2.1551894180064452e-05, "epoch": 0.7021594188476465, "step": 1764}, {"loss": 0.3721041977405548, "grad_norm": 0.03828960113829288, "learning_rate": 2.149891143429961e-05, "epoch": 0.7025574684048164, "step": 1765}, {"loss": 0.399004727602005, "grad_norm": 0.03912022048041857, "learning_rate": 2.1445976052720413e-05, "epoch": 0.7029555179619863, "step": 1766}, {"loss": 0.34453004598617554, "grad_norm": 0.04076133628781534, "learning_rate": 2.1393088123296858e-05, "epoch": 0.7033535675191561, "step": 1767}, {"loss": 0.38108423352241516, "grad_norm": 0.039630875683988626, "learning_rate": 2.1340247733920115e-05, "epoch": 0.703751617076326, "step": 1768}, {"loss": 0.4198358654975891, "grad_norm": 0.0421401168489537, "learning_rate": 2.1287454972402338e-05, "epoch": 0.7041496666334959, "step": 1769}, {"loss": 0.35811084508895874, "grad_norm": 0.036928757604551665, "learning_rate": 2.1234709926476526e-05, "epoch": 0.7045477161906657, "step": 1770}, {"loss": 0.34532463550567627, "grad_norm": 0.03860535384193817, "learning_rate": 2.1182012683796397e-05, "epoch": 0.7049457657478356, "step": 1771}, {"loss": 0.3601348102092743, "grad_norm": 0.039451923536120904, "learning_rate": 2.112936333193622e-05, "epoch": 0.7053438153050055, "step": 1772}, {"loss": 0.3738410472869873, "grad_norm": 0.0425816757186779, "learning_rate": 2.1076761958390666e-05, "epoch": 0.7057418648621754, "step": 1773}, {"loss": 0.39233699440956116, "grad_norm": 0.043545915739499425, "learning_rate": 2.1024208650574717e-05, "epoch": 0.7061399144193452, "step": 1774}, {"loss": 0.3491612672805786, "grad_norm": 0.036173627979836434, "learning_rate": 2.0971703495823396e-05, "epoch": 0.706537963976515, "step": 1775}, {"loss": 0.3509598970413208, "grad_norm": 0.03611675375819017, "learning_rate": 2.09192465813918e-05, "epoch": 0.706936013533685, "step": 1776}, {"loss": 0.37690383195877075, "grad_norm": 0.039943393075794954, "learning_rate": 2.0866837994454762e-05, "epoch": 0.7073340630908548, "step": 1777}, {"loss": 0.3926258087158203, "grad_norm": 0.0425037843489483, "learning_rate": 2.081447782210688e-05, "epoch": 0.7077321126480247, "step": 1778}, {"loss": 0.35077327489852905, "grad_norm": 0.048507176651028744, "learning_rate": 2.0762166151362266e-05, "epoch": 0.7081301622051945, "step": 1779}, {"loss": 0.3710358142852783, "grad_norm": 0.037294315937982915, "learning_rate": 2.070990306915443e-05, "epoch": 0.7085282117623645, "step": 1780}, {"loss": 0.3416087329387665, "grad_norm": 0.038806457691716166, "learning_rate": 2.0657688662336133e-05, "epoch": 0.7089262613195343, "step": 1781}, {"loss": 0.4095533490180969, "grad_norm": 0.04018359205610819, "learning_rate": 2.0605523017679245e-05, "epoch": 0.7093243108767041, "step": 1782}, {"loss": 0.417898952960968, "grad_norm": 0.04189720523799379, "learning_rate": 2.0553406221874617e-05, "epoch": 0.709722360433874, "step": 1783}, {"loss": 0.35929736495018005, "grad_norm": 0.03927404194383474, "learning_rate": 2.0501338361531897e-05, "epoch": 0.7101204099910439, "step": 1784}, {"loss": 0.35312706232070923, "grad_norm": 0.03834073097809812, "learning_rate": 2.0449319523179427e-05, "epoch": 0.7105184595482138, "step": 1785}, {"loss": 0.36717188358306885, "grad_norm": 0.040156954345632365, "learning_rate": 2.0397349793264105e-05, "epoch": 0.7109165091053836, "step": 1786}, {"loss": 0.3513142764568329, "grad_norm": 0.0373174360459802, "learning_rate": 2.0345429258151132e-05, "epoch": 0.7113145586625534, "step": 1787}, {"loss": 0.3299289345741272, "grad_norm": 0.03625371643912122, "learning_rate": 2.0293558004124058e-05, "epoch": 0.7117126082197234, "step": 1788}, {"loss": 0.3897024095058441, "grad_norm": 0.03708370573146644, "learning_rate": 2.0241736117384473e-05, "epoch": 0.7121106577768932, "step": 1789}, {"loss": 0.4043044447898865, "grad_norm": 0.043208804475726593, "learning_rate": 2.018996368405196e-05, "epoch": 0.7125087073340631, "step": 1790}, {"loss": 0.38373881578445435, "grad_norm": 0.04147462929288976, "learning_rate": 2.01382407901639e-05, "epoch": 0.712906756891233, "step": 1791}, {"loss": 0.39703142642974854, "grad_norm": 0.040710499153210525, "learning_rate": 2.008656752167535e-05, "epoch": 0.7133048064484028, "step": 1792}, {"loss": 0.33006203174591064, "grad_norm": 0.03695576634665283, "learning_rate": 2.003494396445889e-05, "epoch": 0.7137028560055727, "step": 1793}, {"loss": 0.38934561610221863, "grad_norm": 0.03888919313125408, "learning_rate": 1.998337020430452e-05, "epoch": 0.7141009055627425, "step": 1794}, {"loss": 0.35393673181533813, "grad_norm": 0.038188589509832234, "learning_rate": 1.9931846326919444e-05, "epoch": 0.7144989551199125, "step": 1795}, {"loss": 0.37780898809432983, "grad_norm": 0.04022615233389889, "learning_rate": 1.9880372417928018e-05, "epoch": 0.7148970046770823, "step": 1796}, {"loss": 0.39242058992385864, "grad_norm": 0.041178870385857276, "learning_rate": 1.982894856287148e-05, "epoch": 0.7152950542342522, "step": 1797}, {"loss": 0.34969937801361084, "grad_norm": 0.036614714552905554, "learning_rate": 1.977757484720798e-05, "epoch": 0.715693103791422, "step": 1798}, {"loss": 0.33093538880348206, "grad_norm": 0.03934725475947621, "learning_rate": 1.9726251356312263e-05, "epoch": 0.7160911533485919, "step": 1799}, {"loss": 0.3519941568374634, "grad_norm": 0.038570314840792604, "learning_rate": 1.9674978175475662e-05, "epoch": 0.7164892029057618, "step": 1800}, {"loss": 0.37082910537719727, "grad_norm": 0.03787124217120573, "learning_rate": 1.9623755389905885e-05, "epoch": 0.7168872524629316, "step": 1801}, {"loss": 0.4136200249195099, "grad_norm": 0.042503206100502094, "learning_rate": 1.9572583084726886e-05, "epoch": 0.7172853020201015, "step": 1802}, {"loss": 0.3855622708797455, "grad_norm": 0.044242437191368875, "learning_rate": 1.9521461344978748e-05, "epoch": 0.7176833515772714, "step": 1803}, {"loss": 0.34056082367897034, "grad_norm": 0.03953006342131965, "learning_rate": 1.94703902556175e-05, "epoch": 0.7180814011344412, "step": 1804}, {"loss": 0.38728323578834534, "grad_norm": 0.04074144171194199, "learning_rate": 1.9419369901515012e-05, "epoch": 0.7184794506916111, "step": 1805}, {"loss": 0.37379440665245056, "grad_norm": 0.039077104799344974, "learning_rate": 1.936840036745883e-05, "epoch": 0.7188775002487809, "step": 1806}, {"loss": 0.3878703713417053, "grad_norm": 0.03868653500648165, "learning_rate": 1.9317481738152053e-05, "epoch": 0.7192755498059509, "step": 1807}, {"loss": 0.38965579867362976, "grad_norm": 0.039143034936997056, "learning_rate": 1.9266614098213204e-05, "epoch": 0.7196735993631207, "step": 1808}, {"loss": 0.3495563268661499, "grad_norm": 0.03914408449396983, "learning_rate": 1.921579753217601e-05, "epoch": 0.7200716489202906, "step": 1809}, {"loss": 0.375482439994812, "grad_norm": 0.0402586386075891, "learning_rate": 1.916503212448938e-05, "epoch": 0.7204696984774605, "step": 1810}, {"loss": 0.38334280252456665, "grad_norm": 0.03697430264083074, "learning_rate": 1.9114317959517185e-05, "epoch": 0.7208677480346303, "step": 1811}, {"loss": 0.3707520365715027, "grad_norm": 0.040082340318117006, "learning_rate": 1.9063655121538137e-05, "epoch": 0.7212657975918002, "step": 1812}, {"loss": 0.3799390196800232, "grad_norm": 0.04029290603277962, "learning_rate": 1.9013043694745663e-05, "epoch": 0.72166384714897, "step": 1813}, {"loss": 0.3592240512371063, "grad_norm": 0.039803307482824665, "learning_rate": 1.8962483763247735e-05, "epoch": 0.72206189670614, "step": 1814}, {"loss": 0.31761589646339417, "grad_norm": 0.03837535549275291, "learning_rate": 1.8911975411066763e-05, "epoch": 0.7224599462633098, "step": 1815}, {"loss": 0.3517720103263855, "grad_norm": 0.03776905701883333, "learning_rate": 1.886151872213943e-05, "epoch": 0.7228579958204796, "step": 1816}, {"loss": 0.3534567356109619, "grad_norm": 0.03722289018631445, "learning_rate": 1.881111378031657e-05, "epoch": 0.7232560453776495, "step": 1817}, {"loss": 0.3693557679653168, "grad_norm": 0.037917723420624855, "learning_rate": 1.8760760669363035e-05, "epoch": 0.7236540949348194, "step": 1818}, {"loss": 0.38714873790740967, "grad_norm": 0.040070098002851905, "learning_rate": 1.8710459472957496e-05, "epoch": 0.7240521444919893, "step": 1819}, {"loss": 0.3474270701408386, "grad_norm": 0.03880935976062734, "learning_rate": 1.8660210274692397e-05, "epoch": 0.7244501940491591, "step": 1820}, {"loss": 0.3562425374984741, "grad_norm": 0.03707650973326854, "learning_rate": 1.8610013158073768e-05, "epoch": 0.7248482436063289, "step": 1821}, {"loss": 0.38619130849838257, "grad_norm": 0.038744939088323466, "learning_rate": 1.8559868206521043e-05, "epoch": 0.7252462931634989, "step": 1822}, {"loss": 0.3590298891067505, "grad_norm": 0.039934344806458716, "learning_rate": 1.850977550336701e-05, "epoch": 0.7256443427206687, "step": 1823}, {"loss": 0.37589457631111145, "grad_norm": 0.04415081363469814, "learning_rate": 1.845973513185762e-05, "epoch": 0.7260423922778386, "step": 1824}, {"loss": 0.36447322368621826, "grad_norm": 0.03700173110795849, "learning_rate": 1.8409747175151854e-05, "epoch": 0.7264404418350084, "step": 1825}, {"loss": 0.36553436517715454, "grad_norm": 0.03852185967555424, "learning_rate": 1.8359811716321577e-05, "epoch": 0.7268384913921784, "step": 1826}, {"loss": 0.39745116233825684, "grad_norm": 0.037138283931233744, "learning_rate": 1.830992883835143e-05, "epoch": 0.7272365409493482, "step": 1827}, {"loss": 0.3731403350830078, "grad_norm": 0.03959654186334968, "learning_rate": 1.8260098624138666e-05, "epoch": 0.727634590506518, "step": 1828}, {"loss": 0.4160448908805847, "grad_norm": 0.04043249414387786, "learning_rate": 1.8210321156493005e-05, "epoch": 0.728032640063688, "step": 1829}, {"loss": 0.37108802795410156, "grad_norm": 0.037752408287200764, "learning_rate": 1.8160596518136535e-05, "epoch": 0.7284306896208578, "step": 1830}, {"loss": 0.3295917510986328, "grad_norm": 0.035178710589557516, "learning_rate": 1.8110924791703554e-05, "epoch": 0.7288287391780277, "step": 1831}, {"loss": 0.40543675422668457, "grad_norm": 0.04140364141321296, "learning_rate": 1.8061306059740372e-05, "epoch": 0.7292267887351975, "step": 1832}, {"loss": 0.31942063570022583, "grad_norm": 0.03858237731461299, "learning_rate": 1.8011740404705297e-05, "epoch": 0.7296248382923674, "step": 1833}, {"loss": 0.36630699038505554, "grad_norm": 0.03760737291098369, "learning_rate": 1.79622279089684e-05, "epoch": 0.7300228878495373, "step": 1834}, {"loss": 0.3471079170703888, "grad_norm": 0.040731434446501145, "learning_rate": 1.7912768654811425e-05, "epoch": 0.7304209374067071, "step": 1835}, {"loss": 0.3689090609550476, "grad_norm": 0.04342727309979034, "learning_rate": 1.7863362724427617e-05, "epoch": 0.730818986963877, "step": 1836}, {"loss": 0.36565032601356506, "grad_norm": 0.04161928498717381, "learning_rate": 1.7814010199921632e-05, "epoch": 0.7312170365210469, "step": 1837}, {"loss": 0.37575477361679077, "grad_norm": 0.03813612949579532, "learning_rate": 1.776471116330935e-05, "epoch": 0.7316150860782168, "step": 1838}, {"loss": 0.3881016969680786, "grad_norm": 0.042073539827662675, "learning_rate": 1.771546569651778e-05, "epoch": 0.7320131356353866, "step": 1839}, {"loss": 0.34133538603782654, "grad_norm": 0.0374062578241826, "learning_rate": 1.7666273881384904e-05, "epoch": 0.7324111851925564, "step": 1840}, {"loss": 0.3387095630168915, "grad_norm": 0.03787679894792569, "learning_rate": 1.7617135799659513e-05, "epoch": 0.7328092347497264, "step": 1841}, {"loss": 0.3664672374725342, "grad_norm": 0.04051629886485529, "learning_rate": 1.7568051533001145e-05, "epoch": 0.7332072843068962, "step": 1842}, {"loss": 0.39348623156547546, "grad_norm": 0.04170314151966795, "learning_rate": 1.7519021162979903e-05, "epoch": 0.7336053338640661, "step": 1843}, {"loss": 0.32605910301208496, "grad_norm": 0.03769679606849618, "learning_rate": 1.747004477107627e-05, "epoch": 0.7340033834212359, "step": 1844}, {"loss": 0.36723554134368896, "grad_norm": 0.038319033555578225, "learning_rate": 1.7421122438681093e-05, "epoch": 0.7344014329784058, "step": 1845}, {"loss": 0.35812246799468994, "grad_norm": 0.03775706104167036, "learning_rate": 1.7372254247095344e-05, "epoch": 0.7347994825355757, "step": 1846}, {"loss": 0.3785731792449951, "grad_norm": 0.03802053200564182, "learning_rate": 1.7323440277530024e-05, "epoch": 0.7351975320927455, "step": 1847}, {"loss": 0.33334559202194214, "grad_norm": 0.039040041816339084, "learning_rate": 1.7274680611106043e-05, "epoch": 0.7355955816499155, "step": 1848}, {"loss": 0.3698333203792572, "grad_norm": 0.04055827834882695, "learning_rate": 1.7225975328854056e-05, "epoch": 0.7359936312070853, "step": 1849}, {"loss": 0.39640673995018005, "grad_norm": 0.04201367908648294, "learning_rate": 1.7177324511714326e-05, "epoch": 0.7363916807642551, "step": 1850}, {"loss": 0.33556151390075684, "grad_norm": 0.0359081308297422, "learning_rate": 1.712872824053664e-05, "epoch": 0.736789730321425, "step": 1851}, {"loss": 0.31716907024383545, "grad_norm": 0.0367565763778887, "learning_rate": 1.70801865960801e-05, "epoch": 0.7371877798785949, "step": 1852}, {"loss": 0.3456237018108368, "grad_norm": 0.038799764301518345, "learning_rate": 1.7031699659013072e-05, "epoch": 0.7375858294357648, "step": 1853}, {"loss": 0.3825717866420746, "grad_norm": 0.04097362990965387, "learning_rate": 1.6983267509912937e-05, "epoch": 0.7379838789929346, "step": 1854}, {"loss": 0.35344135761260986, "grad_norm": 0.03673315511631189, "learning_rate": 1.6934890229266098e-05, "epoch": 0.7383819285501045, "step": 1855}, {"loss": 0.33982864022254944, "grad_norm": 0.03684210063758915, "learning_rate": 1.6886567897467737e-05, "epoch": 0.7387799781072744, "step": 1856}, {"loss": 0.3302130103111267, "grad_norm": 0.03490046958998912, "learning_rate": 1.6838300594821736e-05, "epoch": 0.7391780276644442, "step": 1857}, {"loss": 0.4056750237941742, "grad_norm": 0.04064079214525029, "learning_rate": 1.679008840154052e-05, "epoch": 0.7395760772216141, "step": 1858}, {"loss": 0.3748209774494171, "grad_norm": 0.04336724846649182, "learning_rate": 1.6741931397744935e-05, "epoch": 0.7399741267787839, "step": 1859}, {"loss": 0.373457670211792, "grad_norm": 0.04169436750607853, "learning_rate": 1.669382966346411e-05, "epoch": 0.7403721763359539, "step": 1860}, {"loss": 0.37432873249053955, "grad_norm": 0.03964141017970666, "learning_rate": 1.6645783278635336e-05, "epoch": 0.7407702258931237, "step": 1861}, {"loss": 0.3696022033691406, "grad_norm": 0.037784567541119175, "learning_rate": 1.659779232310388e-05, "epoch": 0.7411682754502935, "step": 1862}, {"loss": 0.34483709931373596, "grad_norm": 0.03981555980452836, "learning_rate": 1.6549856876622937e-05, "epoch": 0.7415663250074634, "step": 1863}, {"loss": 0.379716694355011, "grad_norm": 0.039967740984410575, "learning_rate": 1.6501977018853455e-05, "epoch": 0.7419643745646333, "step": 1864}, {"loss": 0.32652440667152405, "grad_norm": 0.036018669560010146, "learning_rate": 1.6454152829363973e-05, "epoch": 0.7423624241218032, "step": 1865}, {"loss": 0.3705421984195709, "grad_norm": 0.04083525255552901, "learning_rate": 1.640638438763057e-05, "epoch": 0.742760473678973, "step": 1866}, {"loss": 0.3512694835662842, "grad_norm": 0.03797587844154195, "learning_rate": 1.63586717730366e-05, "epoch": 0.743158523236143, "step": 1867}, {"loss": 0.39321914315223694, "grad_norm": 0.03902424546332747, "learning_rate": 1.631101506487271e-05, "epoch": 0.7435565727933128, "step": 1868}, {"loss": 0.3607693314552307, "grad_norm": 0.04242645252954654, "learning_rate": 1.6263414342336618e-05, "epoch": 0.7439546223504826, "step": 1869}, {"loss": 0.3864924907684326, "grad_norm": 0.03817251765763667, "learning_rate": 1.6215869684533004e-05, "epoch": 0.7443526719076525, "step": 1870}, {"loss": 0.3860227167606354, "grad_norm": 0.03887993961146492, "learning_rate": 1.6168381170473374e-05, "epoch": 0.7447507214648224, "step": 1871}, {"loss": 0.37047940492630005, "grad_norm": 0.042011985398631005, "learning_rate": 1.6120948879075937e-05, "epoch": 0.7451487710219923, "step": 1872}, {"loss": 0.3910256624221802, "grad_norm": 0.0365370970978064, "learning_rate": 1.6073572889165462e-05, "epoch": 0.7455468205791621, "step": 1873}, {"loss": 0.3880227208137512, "grad_norm": 0.04105330732994527, "learning_rate": 1.602625327947317e-05, "epoch": 0.7459448701363319, "step": 1874}, {"loss": 0.3432193994522095, "grad_norm": 0.0384548028748152, "learning_rate": 1.5978990128636563e-05, "epoch": 0.7463429196935019, "step": 1875}, {"loss": 0.34918951988220215, "grad_norm": 0.04013680677232496, "learning_rate": 1.593178351519935e-05, "epoch": 0.7467409692506717, "step": 1876}, {"loss": 0.36736613512039185, "grad_norm": 0.03774970714504042, "learning_rate": 1.5884633517611237e-05, "epoch": 0.7471390188078416, "step": 1877}, {"loss": 0.31855344772338867, "grad_norm": 0.03744502177745149, "learning_rate": 1.583754021422788e-05, "epoch": 0.7475370683650114, "step": 1878}, {"loss": 0.36753910779953003, "grad_norm": 0.04297926702165778, "learning_rate": 1.579050368331072e-05, "epoch": 0.7479351179221813, "step": 1879}, {"loss": 0.3602948784828186, "grad_norm": 0.03820128346425322, "learning_rate": 1.5743524003026826e-05, "epoch": 0.7483331674793512, "step": 1880}, {"loss": 0.3876633644104004, "grad_norm": 0.03929847843794526, "learning_rate": 1.5696601251448823e-05, "epoch": 0.748731217036521, "step": 1881}, {"loss": 0.3520588278770447, "grad_norm": 0.04190519974827564, "learning_rate": 1.5649735506554697e-05, "epoch": 0.749129266593691, "step": 1882}, {"loss": 0.35468313097953796, "grad_norm": 0.040551195779122175, "learning_rate": 1.5602926846227738e-05, "epoch": 0.7495273161508608, "step": 1883}, {"loss": 0.3850244879722595, "grad_norm": 0.044523666221428726, "learning_rate": 1.5556175348256312e-05, "epoch": 0.7499253657080307, "step": 1884}, {"loss": 0.3527185022830963, "grad_norm": 0.039761786613983094, "learning_rate": 1.5509481090333837e-05, "epoch": 0.7503234152652005, "step": 1885}, {"loss": 0.3735174834728241, "grad_norm": 0.03798196357792949, "learning_rate": 1.5462844150058603e-05, "epoch": 0.7507214648223703, "step": 1886}, {"loss": 0.34388890862464905, "grad_norm": 0.03648869243319613, "learning_rate": 1.5416264604933633e-05, "epoch": 0.7511195143795403, "step": 1887}, {"loss": 0.33297258615493774, "grad_norm": 0.03728346354308481, "learning_rate": 1.536974253236659e-05, "epoch": 0.7515175639367101, "step": 1888}, {"loss": 0.35659751296043396, "grad_norm": 0.03883942693094905, "learning_rate": 1.532327800966958e-05, "epoch": 0.75191561349388, "step": 1889}, {"loss": 0.40059512853622437, "grad_norm": 0.03877812962716118, "learning_rate": 1.5276871114059116e-05, "epoch": 0.7523136630510499, "step": 1890}, {"loss": 0.394569993019104, "grad_norm": 0.04768570856503336, "learning_rate": 1.5230521922655932e-05, "epoch": 0.7527117126082197, "step": 1891}, {"loss": 0.35792139172554016, "grad_norm": 0.03774197757901003, "learning_rate": 1.518423051248487e-05, "epoch": 0.7531097621653896, "step": 1892}, {"loss": 0.3637424111366272, "grad_norm": 0.042340097480399484, "learning_rate": 1.5137996960474742e-05, "epoch": 0.7535078117225594, "step": 1893}, {"loss": 0.3571650981903076, "grad_norm": 0.03895366040826865, "learning_rate": 1.5091821343458207e-05, "epoch": 0.7539058612797294, "step": 1894}, {"loss": 0.3907700181007385, "grad_norm": 0.03913799864236679, "learning_rate": 1.5045703738171657e-05, "epoch": 0.7543039108368992, "step": 1895}, {"loss": 0.37506890296936035, "grad_norm": 0.038877210979804566, "learning_rate": 1.499964422125506e-05, "epoch": 0.7547019603940691, "step": 1896}, {"loss": 0.3971094489097595, "grad_norm": 0.03909973879436638, "learning_rate": 1.4953642869251872e-05, "epoch": 0.7551000099512389, "step": 1897}, {"loss": 0.3181026577949524, "grad_norm": 0.03838900691275072, "learning_rate": 1.4907699758608884e-05, "epoch": 0.7554980595084088, "step": 1898}, {"loss": 0.3467220962047577, "grad_norm": 0.03803818337559248, "learning_rate": 1.4861814965676068e-05, "epoch": 0.7558961090655787, "step": 1899}, {"loss": 0.39533063769340515, "grad_norm": 0.04070630572475413, "learning_rate": 1.4815988566706523e-05, "epoch": 0.7562941586227485, "step": 1900}, {"loss": 0.3978947699069977, "grad_norm": 0.040404773357894136, "learning_rate": 1.477022063785628e-05, "epoch": 0.7566922081799184, "step": 1901}, {"loss": 0.3476754128932953, "grad_norm": 0.03855731841731346, "learning_rate": 1.4724511255184226e-05, "epoch": 0.7570902577370883, "step": 1902}, {"loss": 0.3888607621192932, "grad_norm": 0.04176291776297895, "learning_rate": 1.4678860494651925e-05, "epoch": 0.7574883072942581, "step": 1903}, {"loss": 0.33032289147377014, "grad_norm": 0.04052974755828241, "learning_rate": 1.4633268432123565e-05, "epoch": 0.757886356851428, "step": 1904}, {"loss": 0.36374396085739136, "grad_norm": 0.039455743825533976, "learning_rate": 1.4587735143365717e-05, "epoch": 0.7582844064085978, "step": 1905}, {"loss": 0.34634697437286377, "grad_norm": 0.03944393859320339, "learning_rate": 1.4542260704047338e-05, "epoch": 0.7586824559657678, "step": 1906}, {"loss": 0.36094018816947937, "grad_norm": 0.03798658789415294, "learning_rate": 1.4496845189739573e-05, "epoch": 0.7590805055229376, "step": 1907}, {"loss": 0.31771180033683777, "grad_norm": 0.03582692076708487, "learning_rate": 1.4451488675915632e-05, "epoch": 0.7594785550801074, "step": 1908}, {"loss": 0.36276042461395264, "grad_norm": 0.04204422368293057, "learning_rate": 1.4406191237950689e-05, "epoch": 0.7598766046372774, "step": 1909}, {"loss": 0.40873658657073975, "grad_norm": 0.03915032278140954, "learning_rate": 1.4360952951121754e-05, "epoch": 0.7602746541944472, "step": 1910}, {"loss": 0.360749214887619, "grad_norm": 0.04263062475973208, "learning_rate": 1.4315773890607487e-05, "epoch": 0.7606727037516171, "step": 1911}, {"loss": 0.3942781984806061, "grad_norm": 0.04006838009771116, "learning_rate": 1.427065413148817e-05, "epoch": 0.7610707533087869, "step": 1912}, {"loss": 0.3269903063774109, "grad_norm": 0.03621234109817166, "learning_rate": 1.422559374874553e-05, "epoch": 0.7614688028659569, "step": 1913}, {"loss": 0.3788313865661621, "grad_norm": 0.03786376146870068, "learning_rate": 1.4180592817262611e-05, "epoch": 0.7618668524231267, "step": 1914}, {"loss": 0.38859760761260986, "grad_norm": 0.04029631654322693, "learning_rate": 1.4135651411823658e-05, "epoch": 0.7622649019802965, "step": 1915}, {"loss": 0.3378605842590332, "grad_norm": 0.04466377932262517, "learning_rate": 1.4090769607113996e-05, "epoch": 0.7626629515374664, "step": 1916}, {"loss": 0.33965742588043213, "grad_norm": 0.03759015376231541, "learning_rate": 1.4045947477719907e-05, "epoch": 0.7630610010946363, "step": 1917}, {"loss": 0.335244357585907, "grad_norm": 0.04003431313622291, "learning_rate": 1.4001185098128499e-05, "epoch": 0.7634590506518062, "step": 1918}, {"loss": 0.4185422658920288, "grad_norm": 0.04534606268344062, "learning_rate": 1.3956482542727584e-05, "epoch": 0.763857100208976, "step": 1919}, {"loss": 0.3237791061401367, "grad_norm": 0.03622863316988237, "learning_rate": 1.3911839885805555e-05, "epoch": 0.7642551497661458, "step": 1920}, {"loss": 0.4118348956108093, "grad_norm": 0.043438668010252576, "learning_rate": 1.386725720155128e-05, "epoch": 0.7646531993233158, "step": 1921}, {"loss": 0.3696123957633972, "grad_norm": 0.04703898544655929, "learning_rate": 1.3822734564053924e-05, "epoch": 0.7650512488804856, "step": 1922}, {"loss": 0.3679657280445099, "grad_norm": 0.042176173722960134, "learning_rate": 1.3778272047302899e-05, "epoch": 0.7654492984376555, "step": 1923}, {"loss": 0.3981359004974365, "grad_norm": 0.044749456640379624, "learning_rate": 1.3733869725187698e-05, "epoch": 0.7658473479948253, "step": 1924}, {"loss": 0.3793066740036011, "grad_norm": 0.044555624928563554, "learning_rate": 1.368952767149777e-05, "epoch": 0.7662453975519953, "step": 1925}, {"loss": 0.35652533173561096, "grad_norm": 0.04252190753118701, "learning_rate": 1.3645245959922436e-05, "epoch": 0.7666434471091651, "step": 1926}, {"loss": 0.4222603440284729, "grad_norm": 0.04326121993125026, "learning_rate": 1.3601024664050683e-05, "epoch": 0.7670414966663349, "step": 1927}, {"loss": 0.35534238815307617, "grad_norm": 0.0405449519963399, "learning_rate": 1.3556863857371155e-05, "epoch": 0.7674395462235049, "step": 1928}, {"loss": 0.3100375533103943, "grad_norm": 0.03590788762361007, "learning_rate": 1.3512763613271934e-05, "epoch": 0.7678375957806747, "step": 1929}, {"loss": 0.3640863597393036, "grad_norm": 0.044777588736526576, "learning_rate": 1.3468724005040484e-05, "epoch": 0.7682356453378446, "step": 1930}, {"loss": 0.37083399295806885, "grad_norm": 0.04041374765171152, "learning_rate": 1.3424745105863484e-05, "epoch": 0.7686336948950144, "step": 1931}, {"loss": 0.39379069209098816, "grad_norm": 0.04490259753100966, "learning_rate": 1.3380826988826734e-05, "epoch": 0.7690317444521843, "step": 1932}, {"loss": 0.37123626470565796, "grad_norm": 0.03767938842213512, "learning_rate": 1.3336969726915034e-05, "epoch": 0.7694297940093542, "step": 1933}, {"loss": 0.36597198247909546, "grad_norm": 0.039650178590655116, "learning_rate": 1.3293173393012005e-05, "epoch": 0.769827843566524, "step": 1934}, {"loss": 0.4010729193687439, "grad_norm": 0.045385466182437675, "learning_rate": 1.324943805990006e-05, "epoch": 0.7702258931236939, "step": 1935}, {"loss": 0.3256797790527344, "grad_norm": 0.03677707703107149, "learning_rate": 1.320576380026024e-05, "epoch": 0.7706239426808638, "step": 1936}, {"loss": 0.3581530451774597, "grad_norm": 0.042854645801394135, "learning_rate": 1.316215068667207e-05, "epoch": 0.7710219922380336, "step": 1937}, {"loss": 0.31480997800827026, "grad_norm": 0.03984444802876249, "learning_rate": 1.3118598791613478e-05, "epoch": 0.7714200417952035, "step": 1938}, {"loss": 0.37249231338500977, "grad_norm": 0.03982321956708193, "learning_rate": 1.3075108187460639e-05, "epoch": 0.7718180913523733, "step": 1939}, {"loss": 0.3611046373844147, "grad_norm": 0.040472652545619645, "learning_rate": 1.3031678946487886e-05, "epoch": 0.7722161409095433, "step": 1940}, {"loss": 0.3512504994869232, "grad_norm": 0.039313223490271616, "learning_rate": 1.2988311140867577e-05, "epoch": 0.7726141904667131, "step": 1941}, {"loss": 0.4016292691230774, "grad_norm": 0.044390266857785715, "learning_rate": 1.2945004842669962e-05, "epoch": 0.773012240023883, "step": 1942}, {"loss": 0.3312394618988037, "grad_norm": 0.03725912505815865, "learning_rate": 1.29017601238631e-05, "epoch": 0.7734102895810528, "step": 1943}, {"loss": 0.32399260997772217, "grad_norm": 0.03773820546993138, "learning_rate": 1.2858577056312671e-05, "epoch": 0.7738083391382227, "step": 1944}, {"loss": 0.37109437584877014, "grad_norm": 0.042467444772511756, "learning_rate": 1.2815455711781943e-05, "epoch": 0.7742063886953926, "step": 1945}, {"loss": 0.3936200737953186, "grad_norm": 0.03854192487833152, "learning_rate": 1.2772396161931594e-05, "epoch": 0.7746044382525624, "step": 1946}, {"loss": 0.3757363259792328, "grad_norm": 0.03845973628272591, "learning_rate": 1.272939847831962e-05, "epoch": 0.7750024878097324, "step": 1947}, {"loss": 0.3892839550971985, "grad_norm": 0.04048479876508048, "learning_rate": 1.2686462732401177e-05, "epoch": 0.7754005373669022, "step": 1948}, {"loss": 0.33032751083374023, "grad_norm": 0.036785178625494123, "learning_rate": 1.2643588995528515e-05, "epoch": 0.775798586924072, "step": 1949}, {"loss": 0.36823171377182007, "grad_norm": 0.04036023087166357, "learning_rate": 1.2600777338950837e-05, "epoch": 0.7761966364812419, "step": 1950}, {"loss": 0.39706525206565857, "grad_norm": 0.04211202934375449, "learning_rate": 1.2558027833814168e-05, "epoch": 0.7765946860384118, "step": 1951}, {"loss": 0.347686231136322, "grad_norm": 0.04157782365504013, "learning_rate": 1.2515340551161247e-05, "epoch": 0.7769927355955817, "step": 1952}, {"loss": 0.3771267235279083, "grad_norm": 0.03724625256338861, "learning_rate": 1.247271556193142e-05, "epoch": 0.7773907851527515, "step": 1953}, {"loss": 0.3782404065132141, "grad_norm": 0.03976940999819716, "learning_rate": 1.24301529369605e-05, "epoch": 0.7777888347099214, "step": 1954}, {"loss": 0.36917591094970703, "grad_norm": 0.03768755736767683, "learning_rate": 1.2387652746980672e-05, "epoch": 0.7781868842670913, "step": 1955}, {"loss": 0.3542528748512268, "grad_norm": 0.0379491787849783, "learning_rate": 1.234521506262034e-05, "epoch": 0.7785849338242611, "step": 1956}, {"loss": 0.3639964461326599, "grad_norm": 0.037142083662146, "learning_rate": 1.2302839954404067e-05, "epoch": 0.778982983381431, "step": 1957}, {"loss": 0.3753269612789154, "grad_norm": 0.0418883703813485, "learning_rate": 1.2260527492752399e-05, "epoch": 0.7793810329386008, "step": 1958}, {"loss": 0.325797975063324, "grad_norm": 0.034393100034308885, "learning_rate": 1.2218277747981789e-05, "epoch": 0.7797790824957708, "step": 1959}, {"loss": 0.4156484603881836, "grad_norm": 0.04085084154640922, "learning_rate": 1.2176090790304457e-05, "epoch": 0.7801771320529406, "step": 1960}, {"loss": 0.35703200101852417, "grad_norm": 0.040294861566195404, "learning_rate": 1.213396668982828e-05, "epoch": 0.7805751816101104, "step": 1961}, {"loss": 0.3322446942329407, "grad_norm": 0.03934088556081747, "learning_rate": 1.2091905516556684e-05, "epoch": 0.7809732311672803, "step": 1962}, {"loss": 0.3871491849422455, "grad_norm": 0.03974556916683614, "learning_rate": 1.2049907340388512e-05, "epoch": 0.7813712807244502, "step": 1963}, {"loss": 0.3450950086116791, "grad_norm": 0.04202911673232623, "learning_rate": 1.2007972231117926e-05, "epoch": 0.7817693302816201, "step": 1964}, {"loss": 0.39171597361564636, "grad_norm": 0.0468850351288264, "learning_rate": 1.1966100258434283e-05, "epoch": 0.7821673798387899, "step": 1965}, {"loss": 0.36350125074386597, "grad_norm": 0.04203722787844189, "learning_rate": 1.192429149192198e-05, "epoch": 0.7825654293959597, "step": 1966}, {"loss": 0.33169081807136536, "grad_norm": 0.03746528976182631, "learning_rate": 1.1882546001060424e-05, "epoch": 0.7829634789531297, "step": 1967}, {"loss": 0.34454846382141113, "grad_norm": 0.040290457101055624, "learning_rate": 1.184086385522386e-05, "epoch": 0.7833615285102995, "step": 1968}, {"loss": 0.35759562253952026, "grad_norm": 0.037680006477772386, "learning_rate": 1.1799245123681224e-05, "epoch": 0.7837595780674694, "step": 1969}, {"loss": 0.3842439353466034, "grad_norm": 0.04024762656230706, "learning_rate": 1.1757689875596112e-05, "epoch": 0.7841576276246393, "step": 1970}, {"loss": 0.3493756651878357, "grad_norm": 0.03979492678134879, "learning_rate": 1.17161981800266e-05, "epoch": 0.7845556771818092, "step": 1971}, {"loss": 0.36846765875816345, "grad_norm": 0.03927880116670257, "learning_rate": 1.167477010592516e-05, "epoch": 0.784953726738979, "step": 1972}, {"loss": 0.3737693428993225, "grad_norm": 0.03666881017511879, "learning_rate": 1.163340572213853e-05, "epoch": 0.7853517762961488, "step": 1973}, {"loss": 0.363140344619751, "grad_norm": 0.03694187953704166, "learning_rate": 1.1592105097407602e-05, "epoch": 0.7857498258533188, "step": 1974}, {"loss": 0.40818527340888977, "grad_norm": 0.03887507710092085, "learning_rate": 1.1550868300367318e-05, "epoch": 0.7861478754104886, "step": 1975}, {"loss": 0.3414795994758606, "grad_norm": 0.03829830211468501, "learning_rate": 1.150969539954655e-05, "epoch": 0.7865459249676585, "step": 1976}, {"loss": 0.3933921456336975, "grad_norm": 0.040725748945657565, "learning_rate": 1.146858646336797e-05, "epoch": 0.7869439745248283, "step": 1977}, {"loss": 0.3639180362224579, "grad_norm": 0.03780002869091761, "learning_rate": 1.1427541560147974e-05, "epoch": 0.7873420240819982, "step": 1978}, {"loss": 0.40788644552230835, "grad_norm": 0.041907205264854586, "learning_rate": 1.1386560758096515e-05, "epoch": 0.7877400736391681, "step": 1979}, {"loss": 0.3903982639312744, "grad_norm": 0.04344617943104407, "learning_rate": 1.1345644125317051e-05, "epoch": 0.7881381231963379, "step": 1980}, {"loss": 0.34266364574432373, "grad_norm": 0.03742012738989468, "learning_rate": 1.1304791729806375e-05, "epoch": 0.7885361727535078, "step": 1981}, {"loss": 0.3267175853252411, "grad_norm": 0.03599561008853057, "learning_rate": 1.1264003639454556e-05, "epoch": 0.7889342223106777, "step": 1982}, {"loss": 0.3722847104072571, "grad_norm": 0.042636326909844584, "learning_rate": 1.1223279922044776e-05, "epoch": 0.7893322718678476, "step": 1983}, {"loss": 0.37408119440078735, "grad_norm": 0.03770704488054032, "learning_rate": 1.1182620645253245e-05, "epoch": 0.7897303214250174, "step": 1984}, {"loss": 0.30473968386650085, "grad_norm": 0.0371539025878046, "learning_rate": 1.1142025876649093e-05, "epoch": 0.7901283709821872, "step": 1985}, {"loss": 0.3749767541885376, "grad_norm": 0.0386706837837524, "learning_rate": 1.1101495683694236e-05, "epoch": 0.7905264205393572, "step": 1986}, {"loss": 0.3048720061779022, "grad_norm": 0.035356951587009786, "learning_rate": 1.1061030133743272e-05, "epoch": 0.790924470096527, "step": 1987}, {"loss": 0.36047637462615967, "grad_norm": 0.0401068866188007, "learning_rate": 1.1020629294043405e-05, "epoch": 0.7913225196536969, "step": 1988}, {"loss": 0.3677241802215576, "grad_norm": 0.04356167519089655, "learning_rate": 1.0980293231734252e-05, "epoch": 0.7917205692108668, "step": 1989}, {"loss": 0.38908514380455017, "grad_norm": 0.04441491710489359, "learning_rate": 1.0940022013847817e-05, "epoch": 0.7921186187680366, "step": 1990}, {"loss": 0.3807780146598816, "grad_norm": 0.03896700291882934, "learning_rate": 1.0899815707308315e-05, "epoch": 0.7925166683252065, "step": 1991}, {"loss": 0.34488117694854736, "grad_norm": 0.040119830738432755, "learning_rate": 1.0859674378932116e-05, "epoch": 0.7929147178823763, "step": 1992}, {"loss": 0.34160879254341125, "grad_norm": 0.034950171734521376, "learning_rate": 1.0819598095427603e-05, "epoch": 0.7933127674395463, "step": 1993}, {"loss": 0.3582054674625397, "grad_norm": 0.03974383371681118, "learning_rate": 1.077958692339504e-05, "epoch": 0.7937108169967161, "step": 1994}, {"loss": 0.35152584314346313, "grad_norm": 0.03806406402383912, "learning_rate": 1.0739640929326522e-05, "epoch": 0.7941088665538859, "step": 1995}, {"loss": 0.3447744846343994, "grad_norm": 0.03726882204768152, "learning_rate": 1.0699760179605806e-05, "epoch": 0.7945069161110558, "step": 1996}, {"loss": 0.3955838680267334, "grad_norm": 0.044176578469628115, "learning_rate": 1.065994474050822e-05, "epoch": 0.7949049656682257, "step": 1997}, {"loss": 0.3750811815261841, "grad_norm": 0.038732080349584905, "learning_rate": 1.0620194678200579e-05, "epoch": 0.7953030152253956, "step": 1998}, {"loss": 0.3877999186515808, "grad_norm": 0.04076427695397619, "learning_rate": 1.0580510058741027e-05, "epoch": 0.7957010647825654, "step": 1999}, {"loss": 0.3739587068557739, "grad_norm": 0.045514647390726436, "learning_rate": 1.0540890948078979e-05, "epoch": 0.7960991143397353, "step": 2000}, {"loss": 0.39389848709106445, "grad_norm": 0.040513779052648885, "learning_rate": 1.050133741205495e-05, "epoch": 0.7964971638969052, "step": 2001}, {"loss": 0.34851813316345215, "grad_norm": 0.041648449889145615, "learning_rate": 1.0461849516400502e-05, "epoch": 0.796895213454075, "step": 2002}, {"loss": 0.37838423252105713, "grad_norm": 0.038305532928075384, "learning_rate": 1.042242732673812e-05, "epoch": 0.7972932630112449, "step": 2003}, {"loss": 0.31143978238105774, "grad_norm": 0.038689418259086845, "learning_rate": 1.0383070908581089e-05, "epoch": 0.7976913125684147, "step": 2004}, {"loss": 0.31295841932296753, "grad_norm": 0.041537644752837395, "learning_rate": 1.0343780327333385e-05, "epoch": 0.7980893621255847, "step": 2005}, {"loss": 0.3587118983268738, "grad_norm": 0.040932735218802, "learning_rate": 1.0304555648289588e-05, "epoch": 0.7984874116827545, "step": 2006}, {"loss": 0.36349570751190186, "grad_norm": 0.04055065258498555, "learning_rate": 1.0265396936634741e-05, "epoch": 0.7988854612399243, "step": 2007}, {"loss": 0.38490453362464905, "grad_norm": 0.042026741845259875, "learning_rate": 1.022630425744428e-05, "epoch": 0.7992835107970943, "step": 2008}, {"loss": 0.3561680316925049, "grad_norm": 0.03553919288899061, "learning_rate": 1.0187277675683893e-05, "epoch": 0.7996815603542641, "step": 2009}, {"loss": 0.3962414264678955, "grad_norm": 0.04174896366010658, "learning_rate": 1.0148317256209439e-05, "epoch": 0.800079609911434, "step": 2010}, {"loss": 0.3657546043395996, "grad_norm": 0.04048875159730753, "learning_rate": 1.0109423063766793e-05, "epoch": 0.8004776594686038, "step": 2011}, {"loss": 0.38026732206344604, "grad_norm": 0.04171957205207271, "learning_rate": 1.0070595162991814e-05, "epoch": 0.8008757090257738, "step": 2012}, {"loss": 0.33955860137939453, "grad_norm": 0.03896373911643131, "learning_rate": 1.0031833618410147e-05, "epoch": 0.8012737585829436, "step": 2013}, {"loss": 0.3288482427597046, "grad_norm": 0.0381321844603447, "learning_rate": 9.993138494437209e-06, "epoch": 0.8016718081401134, "step": 2014}, {"loss": 0.3373393416404724, "grad_norm": 0.037958738497026134, "learning_rate": 9.954509855378025e-06, "epoch": 0.8020698576972833, "step": 2015}, {"loss": 0.3738294243812561, "grad_norm": 0.038842547959242496, "learning_rate": 9.915947765427108e-06, "epoch": 0.8024679072544532, "step": 2016}, {"loss": 0.3634626865386963, "grad_norm": 0.040345692861874186, "learning_rate": 9.877452288668415e-06, "epoch": 0.8028659568116231, "step": 2017}, {"loss": 0.3078131377696991, "grad_norm": 0.03918279943454307, "learning_rate": 9.839023489075172e-06, "epoch": 0.8032640063687929, "step": 2018}, {"loss": 0.386392742395401, "grad_norm": 0.04205258845629215, "learning_rate": 9.80066143050981e-06, "epoch": 0.8036620559259627, "step": 2019}, {"loss": 0.38922005891799927, "grad_norm": 0.044381744346813204, "learning_rate": 9.762366176723858e-06, "epoch": 0.8040601054831327, "step": 2020}, {"loss": 0.34741276502609253, "grad_norm": 0.039665207999841516, "learning_rate": 9.724137791357802e-06, "epoch": 0.8044581550403025, "step": 2021}, {"loss": 0.3677186965942383, "grad_norm": 0.039616297513651935, "learning_rate": 9.685976337941022e-06, "epoch": 0.8048562045974724, "step": 2022}, {"loss": 0.3760112226009369, "grad_norm": 0.04187283085436985, "learning_rate": 9.647881879891679e-06, "epoch": 0.8052542541546422, "step": 2023}, {"loss": 0.3588649034500122, "grad_norm": 0.040619141147354126, "learning_rate": 9.60985448051655e-06, "epoch": 0.8056523037118121, "step": 2024}, {"loss": 0.3880789577960968, "grad_norm": 0.03904058557020496, "learning_rate": 9.571894203011028e-06, "epoch": 0.806050353268982, "step": 2025}, {"loss": 0.37364161014556885, "grad_norm": 0.04030291981861136, "learning_rate": 9.534001110458924e-06, "epoch": 0.8064484028261518, "step": 2026}, {"loss": 0.3797941505908966, "grad_norm": 0.04338404219779586, "learning_rate": 9.496175265832413e-06, "epoch": 0.8068464523833218, "step": 2027}, {"loss": 0.31285929679870605, "grad_norm": 0.03825924662991379, "learning_rate": 9.45841673199192e-06, "epoch": 0.8072445019404916, "step": 2028}, {"loss": 0.3626263439655304, "grad_norm": 0.04090073065952917, "learning_rate": 9.420725571685995e-06, "epoch": 0.8076425514976615, "step": 2029}, {"loss": 0.3535410761833191, "grad_norm": 0.04015322897999171, "learning_rate": 9.383101847551234e-06, "epoch": 0.8080406010548313, "step": 2030}, {"loss": 0.3594149947166443, "grad_norm": 0.041507876187041486, "learning_rate": 9.345545622112168e-06, "epoch": 0.8084386506120012, "step": 2031}, {"loss": 0.34741953015327454, "grad_norm": 0.03830220284880113, "learning_rate": 9.308056957781141e-06, "epoch": 0.8088367001691711, "step": 2032}, {"loss": 0.3516601622104645, "grad_norm": 0.04197060001061803, "learning_rate": 9.270635916858255e-06, "epoch": 0.8092347497263409, "step": 2033}, {"loss": 0.36289846897125244, "grad_norm": 0.041595961037113056, "learning_rate": 9.233282561531176e-06, "epoch": 0.8096327992835108, "step": 2034}, {"loss": 0.39327311515808105, "grad_norm": 0.04329556350835404, "learning_rate": 9.195996953875153e-06, "epoch": 0.8100308488406807, "step": 2035}, {"loss": 0.3644140660762787, "grad_norm": 0.04169905460236713, "learning_rate": 9.158779155852787e-06, "epoch": 0.8104288983978505, "step": 2036}, {"loss": 0.32994088530540466, "grad_norm": 0.03953179223137627, "learning_rate": 9.121629229314033e-06, "epoch": 0.8108269479550204, "step": 2037}, {"loss": 0.36862844228744507, "grad_norm": 0.04022644863117592, "learning_rate": 9.08454723599605e-06, "epoch": 0.8112249975121902, "step": 2038}, {"loss": 0.35520198941230774, "grad_norm": 0.04052184845593417, "learning_rate": 9.047533237523081e-06, "epoch": 0.8116230470693602, "step": 2039}, {"loss": 0.39048945903778076, "grad_norm": 0.04071084031157413, "learning_rate": 9.010587295406404e-06, "epoch": 0.81202109662653, "step": 2040}, {"loss": 0.4137278199195862, "grad_norm": 0.04209201796976556, "learning_rate": 8.973709471044166e-06, "epoch": 0.8124191461836999, "step": 2041}, {"loss": 0.3700346350669861, "grad_norm": 0.03981308754131415, "learning_rate": 8.93689982572134e-06, "epoch": 0.8128171957408697, "step": 2042}, {"loss": 0.39241987466812134, "grad_norm": 0.04703303880187207, "learning_rate": 8.900158420609579e-06, "epoch": 0.8132152452980396, "step": 2043}, {"loss": 0.3371703624725342, "grad_norm": 0.044714315394366606, "learning_rate": 8.863485316767139e-06, "epoch": 0.8136132948552095, "step": 2044}, {"loss": 0.36130520701408386, "grad_norm": 0.040610742407972904, "learning_rate": 8.82688057513878e-06, "epoch": 0.8140113444123793, "step": 2045}, {"loss": 0.3616398870944977, "grad_norm": 0.03905941792901784, "learning_rate": 8.790344256555627e-06, "epoch": 0.8144093939695493, "step": 2046}, {"loss": 0.36824870109558105, "grad_norm": 0.04171974932942554, "learning_rate": 8.75387642173512e-06, "epoch": 0.8148074435267191, "step": 2047}, {"loss": 0.34108200669288635, "grad_norm": 0.04030537742377159, "learning_rate": 8.717477131280882e-06, "epoch": 0.8152054930838889, "step": 2048}, {"loss": 0.3749021291732788, "grad_norm": 0.041930016586900924, "learning_rate": 8.681146445682636e-06, "epoch": 0.8156035426410588, "step": 2049}, {"loss": 0.3947913944721222, "grad_norm": 0.042090356965080566, "learning_rate": 8.644884425316085e-06, "epoch": 0.8160015921982287, "step": 2050}, {"loss": 0.39804065227508545, "grad_norm": 0.039878616329445005, "learning_rate": 8.60869113044282e-06, "epoch": 0.8163996417553986, "step": 2051}, {"loss": 0.3352276384830475, "grad_norm": 0.0393868778518014, "learning_rate": 8.572566621210232e-06, "epoch": 0.8167976913125684, "step": 2052}, {"loss": 0.34963592886924744, "grad_norm": 0.04109552752882505, "learning_rate": 8.536510957651394e-06, "epoch": 0.8171957408697382, "step": 2053}, {"loss": 0.3409014046192169, "grad_norm": 0.03813347777871461, "learning_rate": 8.500524199684983e-06, "epoch": 0.8175937904269082, "step": 2054}, {"loss": 0.3778686821460724, "grad_norm": 0.04095430004463086, "learning_rate": 8.464606407115122e-06, "epoch": 0.817991839984078, "step": 2055}, {"loss": 0.3891613185405731, "grad_norm": 0.04163813064032432, "learning_rate": 8.428757639631374e-06, "epoch": 0.8183898895412479, "step": 2056}, {"loss": 0.37232255935668945, "grad_norm": 0.03710061183847172, "learning_rate": 8.392977956808584e-06, "epoch": 0.8187879390984177, "step": 2057}, {"loss": 0.38773882389068604, "grad_norm": 0.04355094501556318, "learning_rate": 8.357267418106757e-06, "epoch": 0.8191859886555877, "step": 2058}, {"loss": 0.36635878682136536, "grad_norm": 0.04233346397631533, "learning_rate": 8.321626082871032e-06, "epoch": 0.8195840382127575, "step": 2059}, {"loss": 0.3737359941005707, "grad_norm": 0.038966140162407285, "learning_rate": 8.286054010331513e-06, "epoch": 0.8199820877699273, "step": 2060}, {"loss": 0.38649970293045044, "grad_norm": 0.04091046512158485, "learning_rate": 8.250551259603228e-06, "epoch": 0.8203801373270972, "step": 2061}, {"loss": 0.37433499097824097, "grad_norm": 0.039878336075725274, "learning_rate": 8.215117889685986e-06, "epoch": 0.8207781868842671, "step": 2062}, {"loss": 0.3342606723308563, "grad_norm": 0.04473416992798873, "learning_rate": 8.179753959464293e-06, "epoch": 0.821176236441437, "step": 2063}, {"loss": 0.315080463886261, "grad_norm": 0.038759226287069196, "learning_rate": 8.144459527707272e-06, "epoch": 0.8215742859986068, "step": 2064}, {"loss": 0.34756040573120117, "grad_norm": 0.04024163798916067, "learning_rate": 8.109234653068543e-06, "epoch": 0.8219723355557766, "step": 2065}, {"loss": 0.40478748083114624, "grad_norm": 0.04208288181223131, "learning_rate": 8.074079394086126e-06, "epoch": 0.8223703851129466, "step": 2066}, {"loss": 0.35319116711616516, "grad_norm": 0.03808355280558379, "learning_rate": 8.03899380918236e-06, "epoch": 0.8227684346701164, "step": 2067}, {"loss": 0.3721293807029724, "grad_norm": 0.036542087080397224, "learning_rate": 8.003977956663816e-06, "epoch": 0.8231664842272863, "step": 2068}, {"loss": 0.3493967652320862, "grad_norm": 0.04084611797332701, "learning_rate": 7.969031894721123e-06, "epoch": 0.8235645337844562, "step": 2069}, {"loss": 0.3527751564979553, "grad_norm": 0.036310110200215906, "learning_rate": 7.934155681428995e-06, "epoch": 0.8239625833416261, "step": 2070}, {"loss": 0.3863634467124939, "grad_norm": 0.0406695145417892, "learning_rate": 7.899349374746024e-06, "epoch": 0.8243606328987959, "step": 2071}, {"loss": 0.35045111179351807, "grad_norm": 0.04034306218455999, "learning_rate": 7.864613032514662e-06, "epoch": 0.8247586824559657, "step": 2072}, {"loss": 0.37488797307014465, "grad_norm": 0.04083955746483334, "learning_rate": 7.829946712461056e-06, "epoch": 0.8251567320131357, "step": 2073}, {"loss": 0.3826369643211365, "grad_norm": 0.04377901986912327, "learning_rate": 7.795350472195023e-06, "epoch": 0.8255547815703055, "step": 2074}, {"loss": 0.30791813135147095, "grad_norm": 0.03561269564864169, "learning_rate": 7.760824369209908e-06, "epoch": 0.8259528311274754, "step": 2075}, {"loss": 0.35002565383911133, "grad_norm": 0.04035079334103426, "learning_rate": 7.726368460882467e-06, "epoch": 0.8263508806846452, "step": 2076}, {"loss": 0.3569282293319702, "grad_norm": 0.0381856043948274, "learning_rate": 7.691982804472841e-06, "epoch": 0.8267489302418151, "step": 2077}, {"loss": 0.35949599742889404, "grad_norm": 0.038957485083444196, "learning_rate": 7.657667457124418e-06, "epoch": 0.827146979798985, "step": 2078}, {"loss": 0.35159438848495483, "grad_norm": 0.036240240480330414, "learning_rate": 7.623422475863734e-06, "epoch": 0.8275450293561548, "step": 2079}, {"loss": 0.3409033417701721, "grad_norm": 0.04048537268941456, "learning_rate": 7.589247917600406e-06, "epoch": 0.8279430789133247, "step": 2080}, {"loss": 0.37881314754486084, "grad_norm": 0.040493020832630316, "learning_rate": 7.555143839126982e-06, "epoch": 0.8283411284704946, "step": 2081}, {"loss": 0.3721542954444885, "grad_norm": 0.03859010452612747, "learning_rate": 7.521110297118922e-06, "epoch": 0.8287391780276644, "step": 2082}, {"loss": 0.3789125084877014, "grad_norm": 0.03933052797331598, "learning_rate": 7.487147348134449e-06, "epoch": 0.8291372275848343, "step": 2083}, {"loss": 0.393189013004303, "grad_norm": 0.04133895718348935, "learning_rate": 7.4532550486144714e-06, "epoch": 0.8295352771420041, "step": 2084}, {"loss": 0.3563185930252075, "grad_norm": 0.03791454124402381, "learning_rate": 7.419433454882502e-06, "epoch": 0.8299333266991741, "step": 2085}, {"loss": 0.4249532222747803, "grad_norm": 0.04431420224031585, "learning_rate": 7.385682623144541e-06, "epoch": 0.8303313762563439, "step": 2086}, {"loss": 0.38101255893707275, "grad_norm": 0.04092602101776841, "learning_rate": 7.352002609488984e-06, "epoch": 0.8307294258135138, "step": 2087}, {"loss": 0.34306684136390686, "grad_norm": 0.03732676772089911, "learning_rate": 7.318393469886564e-06, "epoch": 0.8311274753706837, "step": 2088}, {"loss": 0.359345018863678, "grad_norm": 0.04301722060269859, "learning_rate": 7.28485526019021e-06, "epoch": 0.8315255249278535, "step": 2089}, {"loss": 0.3698313236236572, "grad_norm": 0.03960834226227433, "learning_rate": 7.251388036135004e-06, "epoch": 0.8319235744850234, "step": 2090}, {"loss": 0.38930320739746094, "grad_norm": 0.041001331726682744, "learning_rate": 7.217991853338013e-06, "epoch": 0.8323216240421932, "step": 2091}, {"loss": 0.3386398255825043, "grad_norm": 0.038100308200297923, "learning_rate": 7.184666767298292e-06, "epoch": 0.8327196735993632, "step": 2092}, {"loss": 0.3782586455345154, "grad_norm": 0.04180863913108386, "learning_rate": 7.151412833396726e-06, "epoch": 0.833117723156533, "step": 2093}, {"loss": 0.36807167530059814, "grad_norm": 0.0371565492621105, "learning_rate": 7.118230106895962e-06, "epoch": 0.8335157727137028, "step": 2094}, {"loss": 0.3455932140350342, "grad_norm": 0.04225916103186877, "learning_rate": 7.085118642940303e-06, "epoch": 0.8339138222708727, "step": 2095}, {"loss": 0.34308964014053345, "grad_norm": 0.036103466485499655, "learning_rate": 7.0520784965556305e-06, "epoch": 0.8343118718280426, "step": 2096}, {"loss": 0.3700520396232605, "grad_norm": 0.04160344028162367, "learning_rate": 7.019109722649325e-06, "epoch": 0.8347099213852125, "step": 2097}, {"loss": 0.3301225006580353, "grad_norm": 0.03631332186286273, "learning_rate": 6.9862123760101095e-06, "epoch": 0.8351079709423823, "step": 2098}, {"loss": 0.35046809911727905, "grad_norm": 0.038031499672609125, "learning_rate": 6.9533865113080575e-06, "epoch": 0.8355060204995522, "step": 2099}, {"loss": 0.3394712507724762, "grad_norm": 0.03992773235181316, "learning_rate": 6.920632183094422e-06, "epoch": 0.8359040700567221, "step": 2100}, {"loss": 0.3138503134250641, "grad_norm": 0.037496349253710363, "learning_rate": 6.887949445801595e-06, "epoch": 0.8363021196138919, "step": 2101}, {"loss": 0.36834651231765747, "grad_norm": 0.03945380783063002, "learning_rate": 6.855338353742985e-06, "epoch": 0.8367001691710618, "step": 2102}, {"loss": 0.3617302179336548, "grad_norm": 0.04150949552185206, "learning_rate": 6.822798961112914e-06, "epoch": 0.8370982187282316, "step": 2103}, {"loss": 0.3393429219722748, "grad_norm": 0.03674013739937149, "learning_rate": 6.790331321986593e-06, "epoch": 0.8374962682854016, "step": 2104}, {"loss": 0.3839133083820343, "grad_norm": 0.04062900391110511, "learning_rate": 6.757935490319961e-06, "epoch": 0.8378943178425714, "step": 2105}, {"loss": 0.4101385176181793, "grad_norm": 0.043119978833270164, "learning_rate": 6.725611519949648e-06, "epoch": 0.8382923673997412, "step": 2106}, {"loss": 0.34708768129348755, "grad_norm": 0.038772655876766336, "learning_rate": 6.6933594645928355e-06, "epoch": 0.8386904169569112, "step": 2107}, {"loss": 0.38662606477737427, "grad_norm": 0.04261488624755267, "learning_rate": 6.661179377847215e-06, "epoch": 0.839088466514081, "step": 2108}, {"loss": 0.3468320965766907, "grad_norm": 0.03690624506400896, "learning_rate": 6.6290713131908715e-06, "epoch": 0.8394865160712509, "step": 2109}, {"loss": 0.3657330274581909, "grad_norm": 0.039598313504642, "learning_rate": 6.597035323982192e-06, "epoch": 0.8398845656284207, "step": 2110}, {"loss": 0.38700318336486816, "grad_norm": 0.04075491003649344, "learning_rate": 6.565071463459799e-06, "epoch": 0.8402826151855906, "step": 2111}, {"loss": 0.37660688161849976, "grad_norm": 0.041946552507964155, "learning_rate": 6.53317978474246e-06, "epoch": 0.8406806647427605, "step": 2112}, {"loss": 0.3697699308395386, "grad_norm": 0.03826377853975833, "learning_rate": 6.5013603408289385e-06, "epoch": 0.8410787142999303, "step": 2113}, {"loss": 0.3253992795944214, "grad_norm": 0.03939155423182736, "learning_rate": 6.4696131845980055e-06, "epoch": 0.8414767638571002, "step": 2114}, {"loss": 0.37705084681510925, "grad_norm": 0.0399381834383434, "learning_rate": 6.437938368808283e-06, "epoch": 0.8418748134142701, "step": 2115}, {"loss": 0.35076773166656494, "grad_norm": 0.03737572569104284, "learning_rate": 6.406335946098174e-06, "epoch": 0.84227286297144, "step": 2116}, {"loss": 0.3459762930870056, "grad_norm": 0.04266816577476429, "learning_rate": 6.374805968985781e-06, "epoch": 0.8426709125286098, "step": 2117}, {"loss": 0.3857154846191406, "grad_norm": 0.04609189037876572, "learning_rate": 6.343348489868828e-06, "epoch": 0.8430689620857796, "step": 2118}, {"loss": 0.38780248165130615, "grad_norm": 0.04044531327612765, "learning_rate": 6.311963561024509e-06, "epoch": 0.8434670116429496, "step": 2119}, {"loss": 0.3733755648136139, "grad_norm": 0.04197373803388921, "learning_rate": 6.280651234609503e-06, "epoch": 0.8438650612001194, "step": 2120}, {"loss": 0.3349369764328003, "grad_norm": 0.03953558893827741, "learning_rate": 6.2494115626598125e-06, "epoch": 0.8442631107572893, "step": 2121}, {"loss": 0.3259201943874359, "grad_norm": 0.036524464031814266, "learning_rate": 6.218244597090705e-06, "epoch": 0.8446611603144591, "step": 2122}, {"loss": 0.3505760729312897, "grad_norm": 0.038781789766596866, "learning_rate": 6.187150389696628e-06, "epoch": 0.845059209871629, "step": 2123}, {"loss": 0.3277239203453064, "grad_norm": 0.039734813635046766, "learning_rate": 6.1561289921511e-06, "epoch": 0.8454572594287989, "step": 2124}, {"loss": 0.3565792441368103, "grad_norm": 0.04207391769957494, "learning_rate": 6.125180456006668e-06, "epoch": 0.8458553089859687, "step": 2125}, {"loss": 0.3423389792442322, "grad_norm": 0.04444460580225576, "learning_rate": 6.0943048326947584e-06, "epoch": 0.8462533585431387, "step": 2126}, {"loss": 0.3828768730163574, "grad_norm": 0.042329149467356046, "learning_rate": 6.063502173525654e-06, "epoch": 0.8466514081003085, "step": 2127}, {"loss": 0.36807820200920105, "grad_norm": 0.04071359473953073, "learning_rate": 6.032772529688379e-06, "epoch": 0.8470494576574784, "step": 2128}, {"loss": 0.3334546387195587, "grad_norm": 0.036813499285255485, "learning_rate": 6.002115952250614e-06, "epoch": 0.8474475072146482, "step": 2129}, {"loss": 0.3971553444862366, "grad_norm": 0.046631523632795435, "learning_rate": 5.9715324921586264e-06, "epoch": 0.847845556771818, "step": 2130}, {"loss": 0.35019737482070923, "grad_norm": 0.03796367368847247, "learning_rate": 5.941022200237156e-06, "epoch": 0.848243606328988, "step": 2131}, {"loss": 0.3691500425338745, "grad_norm": 0.0395109316878621, "learning_rate": 5.910585127189367e-06, "epoch": 0.8486416558861578, "step": 2132}, {"loss": 0.37086227536201477, "grad_norm": 0.03840027087496094, "learning_rate": 5.880221323596729e-06, "epoch": 0.8490397054433277, "step": 2133}, {"loss": 0.3727572560310364, "grad_norm": 0.03941027869191904, "learning_rate": 5.84993083991896e-06, "epoch": 0.8494377550004976, "step": 2134}, {"loss": 0.4008180797100067, "grad_norm": 0.03895602065219097, "learning_rate": 5.819713726493942e-06, "epoch": 0.8498358045576674, "step": 2135}, {"loss": 0.3190762400627136, "grad_norm": 0.041760566266601566, "learning_rate": 5.7895700335375945e-06, "epoch": 0.8502338541148373, "step": 2136}, {"loss": 0.3614473342895508, "grad_norm": 0.038880282696065215, "learning_rate": 5.759499811143859e-06, "epoch": 0.8506319036720071, "step": 2137}, {"loss": 0.3485577702522278, "grad_norm": 0.04201546989954316, "learning_rate": 5.729503109284557e-06, "epoch": 0.8510299532291771, "step": 2138}, {"loss": 0.3288244903087616, "grad_norm": 0.03725718784680698, "learning_rate": 5.6995799778093604e-06, "epoch": 0.8514280027863469, "step": 2139}, {"loss": 0.3259902000427246, "grad_norm": 0.03876790301084644, "learning_rate": 5.669730466445639e-06, "epoch": 0.8518260523435167, "step": 2140}, {"loss": 0.33685749769210815, "grad_norm": 0.039449496953687796, "learning_rate": 5.639954624798444e-06, "epoch": 0.8522241019006866, "step": 2141}, {"loss": 0.36627787351608276, "grad_norm": 0.03849700630329188, "learning_rate": 5.610252502350394e-06, "epoch": 0.8526221514578565, "step": 2142}, {"loss": 0.3596959114074707, "grad_norm": 0.03747489317506396, "learning_rate": 5.580624148461599e-06, "epoch": 0.8530202010150264, "step": 2143}, {"loss": 0.32660895586013794, "grad_norm": 0.03719087183475522, "learning_rate": 5.55106961236958e-06, "epoch": 0.8534182505721962, "step": 2144}, {"loss": 0.3503205180168152, "grad_norm": 0.03823756888582771, "learning_rate": 5.521588943189176e-06, "epoch": 0.8538163001293662, "step": 2145}, {"loss": 0.3491063714027405, "grad_norm": 0.036637023476276835, "learning_rate": 5.492182189912482e-06, "epoch": 0.854214349686536, "step": 2146}, {"loss": 0.385436475276947, "grad_norm": 0.0409090428730271, "learning_rate": 5.46284940140876e-06, "epoch": 0.8546123992437058, "step": 2147}, {"loss": 0.3618273138999939, "grad_norm": 0.037924413568484826, "learning_rate": 5.433590626424323e-06, "epoch": 0.8550104488008757, "step": 2148}, {"loss": 0.346186101436615, "grad_norm": 0.04111122223811155, "learning_rate": 5.404405913582522e-06, "epoch": 0.8554084983580456, "step": 2149}, {"loss": 0.34431761503219604, "grad_norm": 0.03773953478472044, "learning_rate": 5.37529531138361e-06, "epoch": 0.8558065479152155, "step": 2150}, {"loss": 0.3567744791507721, "grad_norm": 0.03656834803189407, "learning_rate": 5.346258868204695e-06, "epoch": 0.8562045974723853, "step": 2151}, {"loss": 0.36865776777267456, "grad_norm": 0.036546195947752935, "learning_rate": 5.317296632299623e-06, "epoch": 0.8566026470295551, "step": 2152}, {"loss": 0.4090920090675354, "grad_norm": 0.044356008005474463, "learning_rate": 5.288408651798943e-06, "epoch": 0.8570006965867251, "step": 2153}, {"loss": 0.3344285190105438, "grad_norm": 0.03772620211893636, "learning_rate": 5.259594974709786e-06, "epoch": 0.8573987461438949, "step": 2154}, {"loss": 0.35277247428894043, "grad_norm": 0.04202267050744148, "learning_rate": 5.230855648915811e-06, "epoch": 0.8577967957010648, "step": 2155}, {"loss": 0.384965181350708, "grad_norm": 0.04203579249799897, "learning_rate": 5.202190722177114e-06, "epoch": 0.8581948452582346, "step": 2156}, {"loss": 0.3744237422943115, "grad_norm": 0.044042476298185425, "learning_rate": 5.173600242130167e-06, "epoch": 0.8585928948154046, "step": 2157}, {"loss": 0.3856009840965271, "grad_norm": 0.0420548585420761, "learning_rate": 5.14508425628768e-06, "epoch": 0.8589909443725744, "step": 2158}, {"loss": 0.3859652280807495, "grad_norm": 0.04096208653476179, "learning_rate": 5.116642812038619e-06, "epoch": 0.8593889939297442, "step": 2159}, {"loss": 0.3799474239349365, "grad_norm": 0.04237922877075631, "learning_rate": 5.088275956648042e-06, "epoch": 0.8597870434869141, "step": 2160}, {"loss": 0.3492079973220825, "grad_norm": 0.037240983327262535, "learning_rate": 5.059983737257068e-06, "epoch": 0.860185093044084, "step": 2161}, {"loss": 0.36627650260925293, "grad_norm": 0.04243734993205318, "learning_rate": 5.031766200882765e-06, "epoch": 0.8605831426012539, "step": 2162}, {"loss": 0.3834967017173767, "grad_norm": 0.037632803436175456, "learning_rate": 5.003623394418105e-06, "epoch": 0.8609811921584237, "step": 2163}, {"loss": 0.3393605649471283, "grad_norm": 0.037661580846813016, "learning_rate": 4.975555364631868e-06, "epoch": 0.8613792417155935, "step": 2164}, {"loss": 0.3736042082309723, "grad_norm": 0.0376312727721766, "learning_rate": 4.947562158168561e-06, "epoch": 0.8617772912727635, "step": 2165}, {"loss": 0.3848246932029724, "grad_norm": 0.04179161737900303, "learning_rate": 4.919643821548364e-06, "epoch": 0.8621753408299333, "step": 2166}, {"loss": 0.3670192062854767, "grad_norm": 0.04267217450648636, "learning_rate": 4.891800401167018e-06, "epoch": 0.8625733903871032, "step": 2167}, {"loss": 0.3786769509315491, "grad_norm": 0.04359602585455378, "learning_rate": 4.864031943295766e-06, "epoch": 0.862971439944273, "step": 2168}, {"loss": 0.3414278030395508, "grad_norm": 0.03888197242638018, "learning_rate": 4.836338494081283e-06, "epoch": 0.8633694895014429, "step": 2169}, {"loss": 0.35484179854393005, "grad_norm": 0.037358596782908676, "learning_rate": 4.8087200995456e-06, "epoch": 0.8637675390586128, "step": 2170}, {"loss": 0.37828534841537476, "grad_norm": 0.04071718490735491, "learning_rate": 4.781176805585991e-06, "epoch": 0.8641655886157826, "step": 2171}, {"loss": 0.36100250482559204, "grad_norm": 0.038442755085928385, "learning_rate": 4.753708657974948e-06, "epoch": 0.8645636381729526, "step": 2172}, {"loss": 0.3457539677619934, "grad_norm": 0.03758214373909756, "learning_rate": 4.726315702360079e-06, "epoch": 0.8649616877301224, "step": 2173}, {"loss": 0.3746402859687805, "grad_norm": 0.03813827470864555, "learning_rate": 4.698997984264031e-06, "epoch": 0.8653597372872923, "step": 2174}, {"loss": 0.3872326910495758, "grad_norm": 0.039505630768374315, "learning_rate": 4.671755549084422e-06, "epoch": 0.8657577868444621, "step": 2175}, {"loss": 0.34315603971481323, "grad_norm": 0.03914242745598176, "learning_rate": 4.644588442093756e-06, "epoch": 0.866155836401632, "step": 2176}, {"loss": 0.37327057123184204, "grad_norm": 0.041740094288164664, "learning_rate": 4.617496708439356e-06, "epoch": 0.8665538859588019, "step": 2177}, {"loss": 0.39304739236831665, "grad_norm": 0.04313336387393121, "learning_rate": 4.590480393143298e-06, "epoch": 0.8669519355159717, "step": 2178}, {"loss": 0.3638196885585785, "grad_norm": 0.037874702800974124, "learning_rate": 4.563539541102308e-06, "epoch": 0.8673499850731416, "step": 2179}, {"loss": 0.340057909488678, "grad_norm": 0.03794487875028405, "learning_rate": 4.53667419708772e-06, "epoch": 0.8677480346303115, "step": 2180}, {"loss": 0.315255343914032, "grad_norm": 0.039945070629809745, "learning_rate": 4.5098844057453685e-06, "epoch": 0.8681460841874813, "step": 2181}, {"loss": 0.2981036305427551, "grad_norm": 0.0369510794789778, "learning_rate": 4.483170211595544e-06, "epoch": 0.8685441337446512, "step": 2182}, {"loss": 0.36575847864151, "grad_norm": 0.04142537482376159, "learning_rate": 4.456531659032897e-06, "epoch": 0.868942183301821, "step": 2183}, {"loss": 0.40696483850479126, "grad_norm": 0.0438348247384695, "learning_rate": 4.429968792326389e-06, "epoch": 0.869340232858991, "step": 2184}, {"loss": 0.35977160930633545, "grad_norm": 0.040395306487988496, "learning_rate": 4.403481655619185e-06, "epoch": 0.8697382824161608, "step": 2185}, {"loss": 0.33971285820007324, "grad_norm": 0.03787889352806701, "learning_rate": 4.37707029292862e-06, "epoch": 0.8701363319733307, "step": 2186}, {"loss": 0.3498438596725464, "grad_norm": 0.042131573530356944, "learning_rate": 4.350734748146096e-06, "epoch": 0.8705343815305006, "step": 2187}, {"loss": 0.3857504725456238, "grad_norm": 0.05366590848868132, "learning_rate": 4.324475065037004e-06, "epoch": 0.8709324310876704, "step": 2188}, {"loss": 0.37617257237434387, "grad_norm": 0.03915642950369727, "learning_rate": 4.2982912872406855e-06, "epoch": 0.8713304806448403, "step": 2189}, {"loss": 0.3852948248386383, "grad_norm": 0.04501157739320904, "learning_rate": 4.272183458270334e-06, "epoch": 0.8717285302020101, "step": 2190}, {"loss": 0.36967819929122925, "grad_norm": 0.0403551955837067, "learning_rate": 4.246151621512917e-06, "epoch": 0.8721265797591801, "step": 2191}, {"loss": 0.367453396320343, "grad_norm": 0.034379924211269834, "learning_rate": 4.220195820229145e-06, "epoch": 0.8725246293163499, "step": 2192}, {"loss": 0.36684995889663696, "grad_norm": 0.040870901917382164, "learning_rate": 4.194316097553319e-06, "epoch": 0.8729226788735197, "step": 2193}, {"loss": 0.37281134724617004, "grad_norm": 0.04030440257240991, "learning_rate": 4.16851249649336e-06, "epoch": 0.8733207284306896, "step": 2194}, {"loss": 0.3836788833141327, "grad_norm": 0.044693794932959566, "learning_rate": 4.142785059930654e-06, "epoch": 0.8737187779878595, "step": 2195}, {"loss": 0.3933508098125458, "grad_norm": 0.04024485155115261, "learning_rate": 4.11713383062004e-06, "epoch": 0.8741168275450294, "step": 2196}, {"loss": 0.32582563161849976, "grad_norm": 0.03781102090917487, "learning_rate": 4.091558851189692e-06, "epoch": 0.8745148771021992, "step": 2197}, {"loss": 0.4082094430923462, "grad_norm": 0.040217592212707216, "learning_rate": 4.066060164141083e-06, "epoch": 0.874912926659369, "step": 2198}, {"loss": 0.4134451150894165, "grad_norm": 0.04191152776656841, "learning_rate": 4.040637811848885e-06, "epoch": 0.875310976216539, "step": 2199}, {"loss": 0.37733951210975647, "grad_norm": 0.04079166677906135, "learning_rate": 4.015291836560936e-06, "epoch": 0.8757090257737088, "step": 2200}, {"loss": 0.31417566537857056, "grad_norm": 0.03767161225041137, "learning_rate": 3.99002228039812e-06, "epoch": 0.8761070753308787, "step": 2201}, {"loss": 0.323293536901474, "grad_norm": 0.0386515158582251, "learning_rate": 3.964829185354363e-06, "epoch": 0.8765051248880485, "step": 2202}, {"loss": 0.3744819760322571, "grad_norm": 0.03996214877641686, "learning_rate": 3.939712593296474e-06, "epoch": 0.8769031744452185, "step": 2203}, {"loss": 0.3766176104545593, "grad_norm": 0.04628430030676851, "learning_rate": 3.914672545964182e-06, "epoch": 0.8773012240023883, "step": 2204}, {"loss": 0.4063642621040344, "grad_norm": 0.043263515930800134, "learning_rate": 3.889709084969956e-06, "epoch": 0.8776992735595581, "step": 2205}, {"loss": 0.37760406732559204, "grad_norm": 0.03971325042367384, "learning_rate": 3.864822251799033e-06, "epoch": 0.8780973231167281, "step": 2206}, {"loss": 0.35968220233917236, "grad_norm": 0.04036427326418243, "learning_rate": 3.840012087809292e-06, "epoch": 0.8784953726738979, "step": 2207}, {"loss": 0.3663249909877777, "grad_norm": 0.04038746783068454, "learning_rate": 3.815278634231195e-06, "epoch": 0.8788934222310678, "step": 2208}, {"loss": 0.35491669178009033, "grad_norm": 0.03932467514250291, "learning_rate": 3.7906219321677427e-06, "epoch": 0.8792914717882376, "step": 2209}, {"loss": 0.36566630005836487, "grad_norm": 0.038289907673958234, "learning_rate": 3.766042022594363e-06, "epoch": 0.8796895213454075, "step": 2210}, {"loss": 0.4094424843788147, "grad_norm": 0.04036062311663223, "learning_rate": 3.7415389463588744e-06, "epoch": 0.8800875709025774, "step": 2211}, {"loss": 0.39650723338127136, "grad_norm": 0.04189367413021079, "learning_rate": 3.717112744181428e-06, "epoch": 0.8804856204597472, "step": 2212}, {"loss": 0.3862749934196472, "grad_norm": 0.0402961715324747, "learning_rate": 3.6927634566544024e-06, "epoch": 0.8808836700169171, "step": 2213}, {"loss": 0.3480958342552185, "grad_norm": 0.03619997669461489, "learning_rate": 3.668491124242368e-06, "epoch": 0.881281719574087, "step": 2214}, {"loss": 0.3376547694206238, "grad_norm": 0.03556318626899439, "learning_rate": 3.6442957872819926e-06, "epoch": 0.8816797691312569, "step": 2215}, {"loss": 0.3432578444480896, "grad_norm": 0.04236738384963827, "learning_rate": 3.6201774859820128e-06, "epoch": 0.8820778186884267, "step": 2216}, {"loss": 0.321664035320282, "grad_norm": 0.03734673745111652, "learning_rate": 3.5961362604231218e-06, "epoch": 0.8824758682455965, "step": 2217}, {"loss": 0.3729662597179413, "grad_norm": 0.04065388900660863, "learning_rate": 3.572172150557945e-06, "epoch": 0.8828739178027665, "step": 2218}, {"loss": 0.3691394627094269, "grad_norm": 0.03709412113760359, "learning_rate": 3.548285196210943e-06, "epoch": 0.8832719673599363, "step": 2219}, {"loss": 0.34303271770477295, "grad_norm": 0.03888328289804236, "learning_rate": 3.524475437078356e-06, "epoch": 0.8836700169171062, "step": 2220}, {"loss": 0.3831624984741211, "grad_norm": 0.04107862160997579, "learning_rate": 3.5007429127281443e-06, "epoch": 0.884068066474276, "step": 2221}, {"loss": 0.3075230121612549, "grad_norm": 0.037080702292922034, "learning_rate": 3.477087662599915e-06, "epoch": 0.8844661160314459, "step": 2222}, {"loss": 0.3741074800491333, "grad_norm": 0.04173564870692419, "learning_rate": 3.45350972600485e-06, "epoch": 0.8848641655886158, "step": 2223}, {"loss": 0.3698442578315735, "grad_norm": 0.03966332631370515, "learning_rate": 3.4300091421256676e-06, "epoch": 0.8852622151457856, "step": 2224}, {"loss": 0.3586094081401825, "grad_norm": 0.04361382950942053, "learning_rate": 3.4065859500165165e-06, "epoch": 0.8856602647029556, "step": 2225}, {"loss": 0.35638415813446045, "grad_norm": 0.03765189450087353, "learning_rate": 3.383240188602943e-06, "epoch": 0.8860583142601254, "step": 2226}, {"loss": 0.3312864899635315, "grad_norm": 0.036864944154927984, "learning_rate": 3.3599718966818185e-06, "epoch": 0.8864563638172952, "step": 2227}, {"loss": 0.34662696719169617, "grad_norm": 0.044058357024151396, "learning_rate": 3.336781112921261e-06, "epoch": 0.8868544133744651, "step": 2228}, {"loss": 0.3127092719078064, "grad_norm": 0.03707827742865207, "learning_rate": 3.3136678758605933e-06, "epoch": 0.887252462931635, "step": 2229}, {"loss": 0.37119245529174805, "grad_norm": 0.03847649591452141, "learning_rate": 3.290632223910273e-06, "epoch": 0.8876505124888049, "step": 2230}, {"loss": 0.3555906116962433, "grad_norm": 0.03836597224557183, "learning_rate": 3.2676741953518106e-06, "epoch": 0.8880485620459747, "step": 2231}, {"loss": 0.3723151385784149, "grad_norm": 0.04008476558314548, "learning_rate": 3.2447938283377276e-06, "epoch": 0.8884466116031446, "step": 2232}, {"loss": 0.3298184871673584, "grad_norm": 0.042400181383731674, "learning_rate": 3.2219911608914856e-06, "epoch": 0.8888446611603145, "step": 2233}, {"loss": 0.41944241523742676, "grad_norm": 0.04900022031842721, "learning_rate": 3.1992662309074104e-06, "epoch": 0.8892427107174843, "step": 2234}, {"loss": 0.3046071529388428, "grad_norm": 0.034384081125945785, "learning_rate": 3.1766190761506554e-06, "epoch": 0.8896407602746542, "step": 2235}, {"loss": 0.4006359875202179, "grad_norm": 0.040686033657180154, "learning_rate": 3.1540497342571194e-06, "epoch": 0.890038809831824, "step": 2236}, {"loss": 0.3550195097923279, "grad_norm": 0.039109197587390926, "learning_rate": 3.13155824273339e-06, "epoch": 0.890436859388994, "step": 2237}, {"loss": 0.3683176636695862, "grad_norm": 0.042666443123632755, "learning_rate": 3.1091446389566627e-06, "epoch": 0.8908349089461638, "step": 2238}, {"loss": 0.3735935688018799, "grad_norm": 0.039719361863456475, "learning_rate": 3.0868089601747197e-06, "epoch": 0.8912329585033336, "step": 2239}, {"loss": 0.3250776529312134, "grad_norm": 0.03996240789310365, "learning_rate": 3.0645512435058368e-06, "epoch": 0.8916310080605035, "step": 2240}, {"loss": 0.33446621894836426, "grad_norm": 0.03862911734326932, "learning_rate": 3.042371525938731e-06, "epoch": 0.8920290576176734, "step": 2241}, {"loss": 0.38389548659324646, "grad_norm": 0.03831474114506597, "learning_rate": 3.0202698443324905e-06, "epoch": 0.8924271071748433, "step": 2242}, {"loss": 0.3162670135498047, "grad_norm": 0.040227587738487676, "learning_rate": 2.9982462354165243e-06, "epoch": 0.8928251567320131, "step": 2243}, {"loss": 0.38242441415786743, "grad_norm": 0.04179610251203293, "learning_rate": 2.9763007357904994e-06, "epoch": 0.8932232062891831, "step": 2244}, {"loss": 0.3429411053657532, "grad_norm": 0.046859899629874184, "learning_rate": 2.9544333819242763e-06, "epoch": 0.8936212558463529, "step": 2245}, {"loss": 0.3699416518211365, "grad_norm": 0.0420940567724311, "learning_rate": 2.9326442101578532e-06, "epoch": 0.8940193054035227, "step": 2246}, {"loss": 0.38397443294525146, "grad_norm": 0.03963103398961516, "learning_rate": 2.9109332567012928e-06, "epoch": 0.8944173549606926, "step": 2247}, {"loss": 0.3652493953704834, "grad_norm": 0.04243485087374588, "learning_rate": 2.889300557634678e-06, "epoch": 0.8948154045178625, "step": 2248}, {"loss": 0.3320420980453491, "grad_norm": 0.03635827277592062, "learning_rate": 2.8677461489080513e-06, "epoch": 0.8952134540750324, "step": 2249}, {"loss": 0.4145897626876831, "grad_norm": 0.04151410699779073, "learning_rate": 2.8462700663413333e-06, "epoch": 0.8956115036322022, "step": 2250}, {"loss": 0.38790372014045715, "grad_norm": 0.04292895341163981, "learning_rate": 2.8248723456242965e-06, "epoch": 0.896009553189372, "step": 2251}, {"loss": 0.3543241620063782, "grad_norm": 0.03812750546538635, "learning_rate": 2.8035530223164797e-06, "epoch": 0.896407602746542, "step": 2252}, {"loss": 0.42483198642730713, "grad_norm": 0.04423216733019861, "learning_rate": 2.782312131847137e-06, "epoch": 0.8968056523037118, "step": 2253}, {"loss": 0.36042818427085876, "grad_norm": 0.03988797401280174, "learning_rate": 2.761149709515193e-06, "epoch": 0.8972037018608817, "step": 2254}, {"loss": 0.3324083387851715, "grad_norm": 0.03966881282223693, "learning_rate": 2.7400657904891438e-06, "epoch": 0.8976017514180515, "step": 2255}, {"loss": 0.3456255793571472, "grad_norm": 0.04072718213047935, "learning_rate": 2.7190604098070617e-06, "epoch": 0.8979998009752214, "step": 2256}, {"loss": 0.37229758501052856, "grad_norm": 0.03867706386017719, "learning_rate": 2.6981336023764627e-06, "epoch": 0.8983978505323913, "step": 2257}, {"loss": 0.35123467445373535, "grad_norm": 0.03819756530193599, "learning_rate": 2.677285402974322e-06, "epoch": 0.8987959000895611, "step": 2258}, {"loss": 0.3269639015197754, "grad_norm": 0.038226822148878, "learning_rate": 2.656515846246965e-06, "epoch": 0.899193949646731, "step": 2259}, {"loss": 0.3439939618110657, "grad_norm": 0.040770286269410184, "learning_rate": 2.63582496671001e-06, "epoch": 0.8995919992039009, "step": 2260}, {"loss": 0.36668068170547485, "grad_norm": 0.038663928023514, "learning_rate": 2.615212798748351e-06, "epoch": 0.8999900487610708, "step": 2261}, {"loss": 0.4156007766723633, "grad_norm": 0.04148083037822881, "learning_rate": 2.594679376616066e-06, "epoch": 0.9003880983182406, "step": 2262}, {"loss": 0.3512727618217468, "grad_norm": 0.03838582106111327, "learning_rate": 2.5742247344363757e-06, "epoch": 0.9007861478754104, "step": 2263}, {"loss": 0.35800760984420776, "grad_norm": 0.03593715143462972, "learning_rate": 2.5538489062015724e-06, "epoch": 0.9011841974325804, "step": 2264}, {"loss": 0.3594546914100647, "grad_norm": 0.04078573354034262, "learning_rate": 2.5335519257729813e-06, "epoch": 0.9015822469897502, "step": 2265}, {"loss": 0.3452777564525604, "grad_norm": 0.04223142866860434, "learning_rate": 2.513333826880887e-06, "epoch": 0.9019802965469201, "step": 2266}, {"loss": 0.32971301674842834, "grad_norm": 0.04139362218329556, "learning_rate": 2.4931946431244968e-06, "epoch": 0.90237834610409, "step": 2267}, {"loss": 0.3230806291103363, "grad_norm": 0.038002648414646255, "learning_rate": 2.4731344079718666e-06, "epoch": 0.9027763956612598, "step": 2268}, {"loss": 0.3779425323009491, "grad_norm": 0.04097464125584525, "learning_rate": 2.453153154759841e-06, "epoch": 0.9031744452184297, "step": 2269}, {"loss": 0.3621421158313751, "grad_norm": 0.04060320937110503, "learning_rate": 2.43325091669403e-06, "epoch": 0.9035724947755995, "step": 2270}, {"loss": 0.335235059261322, "grad_norm": 0.041885902144499325, "learning_rate": 2.4134277268487215e-06, "epoch": 0.9039705443327695, "step": 2271}, {"loss": 0.3296283185482025, "grad_norm": 0.035997977016197454, "learning_rate": 2.3936836181668533e-06, "epoch": 0.9043685938899393, "step": 2272}, {"loss": 0.3639025390148163, "grad_norm": 0.04024254254160462, "learning_rate": 2.3740186234599117e-06, "epoch": 0.9047666434471092, "step": 2273}, {"loss": 0.4028654098510742, "grad_norm": 0.04300710158075826, "learning_rate": 2.3544327754079385e-06, "epoch": 0.905164693004279, "step": 2274}, {"loss": 0.3637426495552063, "grad_norm": 0.04241626894054134, "learning_rate": 2.3349261065594376e-06, "epoch": 0.9055627425614489, "step": 2275}, {"loss": 0.38054871559143066, "grad_norm": 0.04053233765766521, "learning_rate": 2.3154986493313214e-06, "epoch": 0.9059607921186188, "step": 2276}, {"loss": 0.3470621407032013, "grad_norm": 0.04053876205400488, "learning_rate": 2.2961504360088815e-06, "epoch": 0.9063588416757886, "step": 2277}, {"loss": 0.3927617371082306, "grad_norm": 0.0389063699223632, "learning_rate": 2.2768814987456975e-06, "epoch": 0.9067568912329586, "step": 2278}, {"loss": 0.35797739028930664, "grad_norm": 0.04116286438087469, "learning_rate": 2.2576918695636327e-06, "epoch": 0.9071549407901284, "step": 2279}, {"loss": 0.3557458221912384, "grad_norm": 0.04430362638241872, "learning_rate": 2.238581580352722e-06, "epoch": 0.9075529903472982, "step": 2280}, {"loss": 0.3393566608428955, "grad_norm": 0.04054870286175556, "learning_rate": 2.219550662871173e-06, "epoch": 0.9079510399044681, "step": 2281}, {"loss": 0.37190425395965576, "grad_norm": 0.04142747647165046, "learning_rate": 2.200599148745297e-06, "epoch": 0.908349089461638, "step": 2282}, {"loss": 0.3310416638851166, "grad_norm": 0.040362771981374584, "learning_rate": 2.181727069469408e-06, "epoch": 0.9087471390188079, "step": 2283}, {"loss": 0.3625061511993408, "grad_norm": 0.03748366618440168, "learning_rate": 2.1629344564058572e-06, "epoch": 0.9091451885759777, "step": 2284}, {"loss": 0.36618155241012573, "grad_norm": 0.05130217146692501, "learning_rate": 2.144221340784919e-06, "epoch": 0.9095432381331475, "step": 2285}, {"loss": 0.3734222650527954, "grad_norm": 0.03995246817212748, "learning_rate": 2.1255877537047496e-06, "epoch": 0.9099412876903175, "step": 2286}, {"loss": 0.3716510534286499, "grad_norm": 0.0401151488164824, "learning_rate": 2.107033726131358e-06, "epoch": 0.9103393372474873, "step": 2287}, {"loss": 0.37005189061164856, "grad_norm": 0.041978597637840585, "learning_rate": 2.0885592888985117e-06, "epoch": 0.9107373868046572, "step": 2288}, {"loss": 0.3962196707725525, "grad_norm": 0.03906064435427749, "learning_rate": 2.0701644727077528e-06, "epoch": 0.911135436361827, "step": 2289}, {"loss": 0.3607112765312195, "grad_norm": 0.04155517934835584, "learning_rate": 2.051849308128256e-06, "epoch": 0.911533485918997, "step": 2290}, {"loss": 0.3908624053001404, "grad_norm": 0.040616729229379166, "learning_rate": 2.0336138255968694e-06, "epoch": 0.9119315354761668, "step": 2291}, {"loss": 0.33939528465270996, "grad_norm": 0.03880685078048676, "learning_rate": 2.0154580554180015e-06, "epoch": 0.9123295850333366, "step": 2292}, {"loss": 0.3802681267261505, "grad_norm": 0.042361059704975855, "learning_rate": 1.9973820277635945e-06, "epoch": 0.9127276345905065, "step": 2293}, {"loss": 0.36152610182762146, "grad_norm": 0.04093774535479589, "learning_rate": 1.979385772673087e-06, "epoch": 0.9131256841476764, "step": 2294}, {"loss": 0.3776471018791199, "grad_norm": 0.041445479798217354, "learning_rate": 1.9614693200533164e-06, "epoch": 0.9135237337048463, "step": 2295}, {"loss": 0.40605324506759644, "grad_norm": 0.04297891349255284, "learning_rate": 1.9436326996785327e-06, "epoch": 0.9139217832620161, "step": 2296}, {"loss": 0.3685782551765442, "grad_norm": 0.0393683748549394, "learning_rate": 1.925875941190303e-06, "epoch": 0.9143198328191859, "step": 2297}, {"loss": 0.36139214038848877, "grad_norm": 0.04007238293916912, "learning_rate": 1.9081990740974774e-06, "epoch": 0.9147178823763559, "step": 2298}, {"loss": 0.34474703669548035, "grad_norm": 0.04181385654736887, "learning_rate": 1.8906021277761422e-06, "epoch": 0.9151159319335257, "step": 2299}, {"loss": 0.39937466382980347, "grad_norm": 0.042706199550625934, "learning_rate": 1.8730851314695708e-06, "epoch": 0.9155139814906956, "step": 2300}, {"loss": 0.3487665057182312, "grad_norm": 0.040714453181505544, "learning_rate": 1.8556481142881664e-06, "epoch": 0.9159120310478654, "step": 2301}, {"loss": 0.3393745422363281, "grad_norm": 0.039566431193620764, "learning_rate": 1.838291105209422e-06, "epoch": 0.9163100806050354, "step": 2302}, {"loss": 0.3382166922092438, "grad_norm": 0.037636984766250874, "learning_rate": 1.821014133077875e-06, "epoch": 0.9167081301622052, "step": 2303}, {"loss": 0.34459179639816284, "grad_norm": 0.042042100584876216, "learning_rate": 1.803817226605048e-06, "epoch": 0.917106179719375, "step": 2304}, {"loss": 0.36559411883354187, "grad_norm": 0.03885804259559513, "learning_rate": 1.7867004143694078e-06, "epoch": 0.917504229276545, "step": 2305}, {"loss": 0.31613850593566895, "grad_norm": 0.03910920338993747, "learning_rate": 1.7696637248163228e-06, "epoch": 0.9179022788337148, "step": 2306}, {"loss": 0.35485517978668213, "grad_norm": 0.03877635555376686, "learning_rate": 1.7527071862580014e-06, "epoch": 0.9183003283908847, "step": 2307}, {"loss": 0.32582759857177734, "grad_norm": 0.038228960995750745, "learning_rate": 1.735830826873469e-06, "epoch": 0.9186983779480545, "step": 2308}, {"loss": 0.3230430483818054, "grad_norm": 0.03978027977432753, "learning_rate": 1.7190346747084918e-06, "epoch": 0.9190964275052244, "step": 2309}, {"loss": 0.34750714898109436, "grad_norm": 0.04068441936159275, "learning_rate": 1.7023187576755528e-06, "epoch": 0.9194944770623943, "step": 2310}, {"loss": 0.4036877751350403, "grad_norm": 0.038866940463487515, "learning_rate": 1.6856831035537868e-06, "epoch": 0.9198925266195641, "step": 2311}, {"loss": 0.3632376492023468, "grad_norm": 0.0377824562205185, "learning_rate": 1.6691277399889626e-06, "epoch": 0.920290576176734, "step": 2312}, {"loss": 0.3209530711174011, "grad_norm": 0.03727115664409887, "learning_rate": 1.6526526944934007e-06, "epoch": 0.9206886257339039, "step": 2313}, {"loss": 0.36258116364479065, "grad_norm": 0.04452861446286723, "learning_rate": 1.6362579944459554e-06, "epoch": 0.9210866752910737, "step": 2314}, {"loss": 0.3180396556854248, "grad_norm": 0.03509193853430102, "learning_rate": 1.619943667091961e-06, "epoch": 0.9214847248482436, "step": 2315}, {"loss": 0.37281399965286255, "grad_norm": 0.04277602823988709, "learning_rate": 1.6037097395431798e-06, "epoch": 0.9218827744054134, "step": 2316}, {"loss": 0.3615191578865051, "grad_norm": 0.03953928856751194, "learning_rate": 1.5875562387777765e-06, "epoch": 0.9222808239625834, "step": 2317}, {"loss": 0.33866533637046814, "grad_norm": 0.03882176384347776, "learning_rate": 1.5714831916402383e-06, "epoch": 0.9226788735197532, "step": 2318}, {"loss": 0.35268422961235046, "grad_norm": 0.03559230896199223, "learning_rate": 1.5554906248413604e-06, "epoch": 0.9230769230769231, "step": 2319}, {"loss": 0.37416431307792664, "grad_norm": 0.04069170269319149, "learning_rate": 1.5395785649581995e-06, "epoch": 0.923474972634093, "step": 2320}, {"loss": 0.30533742904663086, "grad_norm": 0.037519592933063645, "learning_rate": 1.5237470384340202e-06, "epoch": 0.9238730221912628, "step": 2321}, {"loss": 0.32351526618003845, "grad_norm": 0.04068908199454139, "learning_rate": 1.5079960715782438e-06, "epoch": 0.9242710717484327, "step": 2322}, {"loss": 0.3504239320755005, "grad_norm": 0.03912884522908878, "learning_rate": 1.4923256905664263e-06, "epoch": 0.9246691213056025, "step": 2323}, {"loss": 0.3527126908302307, "grad_norm": 0.03770637622653767, "learning_rate": 1.4767359214402032e-06, "epoch": 0.9250671708627725, "step": 2324}, {"loss": 0.3802565634250641, "grad_norm": 0.04136386055217465, "learning_rate": 1.461226790107234e-06, "epoch": 0.9254652204199423, "step": 2325}, {"loss": 0.39575815200805664, "grad_norm": 0.04151048760545781, "learning_rate": 1.4457983223411798e-06, "epoch": 0.9258632699771121, "step": 2326}, {"loss": 0.3595276474952698, "grad_norm": 0.04292821723489023, "learning_rate": 1.430450543781664e-06, "epoch": 0.926261319534282, "step": 2327}, {"loss": 0.37900248169898987, "grad_norm": 0.04298764214011167, "learning_rate": 1.4151834799341845e-06, "epoch": 0.9266593690914519, "step": 2328}, {"loss": 0.3787077069282532, "grad_norm": 0.041003423654139656, "learning_rate": 1.3999971561701408e-06, "epoch": 0.9270574186486218, "step": 2329}, {"loss": 0.3516188859939575, "grad_norm": 0.03834892803359715, "learning_rate": 1.3848915977267284e-06, "epoch": 0.9274554682057916, "step": 2330}, {"loss": 0.33716994524002075, "grad_norm": 0.03806559921628288, "learning_rate": 1.3698668297069395e-06, "epoch": 0.9278535177629615, "step": 2331}, {"loss": 0.34911274909973145, "grad_norm": 0.03845935442924988, "learning_rate": 1.3549228770795118e-06, "epoch": 0.9282515673201314, "step": 2332}, {"loss": 0.3966204524040222, "grad_norm": 0.04232448504160101, "learning_rate": 1.3400597646788527e-06, "epoch": 0.9286496168773012, "step": 2333}, {"loss": 0.3252112865447998, "grad_norm": 0.04058937240823024, "learning_rate": 1.325277517205048e-06, "epoch": 0.9290476664344711, "step": 2334}, {"loss": 0.45049649477005005, "grad_norm": 0.043455711952390334, "learning_rate": 1.3105761592237976e-06, "epoch": 0.9294457159916409, "step": 2335}, {"loss": 0.35118556022644043, "grad_norm": 0.040048234581172334, "learning_rate": 1.2959557151663803e-06, "epoch": 0.9298437655488109, "step": 2336}, {"loss": 0.32746636867523193, "grad_norm": 0.04053207254800748, "learning_rate": 1.281416209329589e-06, "epoch": 0.9302418151059807, "step": 2337}, {"loss": 0.38243600726127625, "grad_norm": 0.04118386627780795, "learning_rate": 1.2669576658757342e-06, "epoch": 0.9306398646631505, "step": 2338}, {"loss": 0.31870871782302856, "grad_norm": 0.0385213814147749, "learning_rate": 1.252580108832574e-06, "epoch": 0.9310379142203205, "step": 2339}, {"loss": 0.3462691903114319, "grad_norm": 0.03714068717800277, "learning_rate": 1.238283562093262e-06, "epoch": 0.9314359637774903, "step": 2340}, {"loss": 0.38229185342788696, "grad_norm": 0.04353077655506282, "learning_rate": 1.2240680494163437e-06, "epoch": 0.9318340133346602, "step": 2341}, {"loss": 0.35440129041671753, "grad_norm": 0.03950735127672214, "learning_rate": 1.2099335944256994e-06, "epoch": 0.93223206289183, "step": 2342}, {"loss": 0.37859469652175903, "grad_norm": 0.04293373648482546, "learning_rate": 1.1958802206104957e-06, "epoch": 0.9326301124489998, "step": 2343}, {"loss": 0.3605368137359619, "grad_norm": 0.042370414493452115, "learning_rate": 1.1819079513251618e-06, "epoch": 0.9330281620061698, "step": 2344}, {"loss": 0.38432180881500244, "grad_norm": 0.04064270959976492, "learning_rate": 1.1680168097893462e-06, "epoch": 0.9334262115633396, "step": 2345}, {"loss": 0.31761372089385986, "grad_norm": 0.03972470513747297, "learning_rate": 1.1542068190878663e-06, "epoch": 0.9338242611205095, "step": 2346}, {"loss": 0.3544498682022095, "grad_norm": 0.038054182000163805, "learning_rate": 1.1404780021706917e-06, "epoch": 0.9342223106776794, "step": 2347}, {"loss": 0.3615521490573883, "grad_norm": 0.038805174612388846, "learning_rate": 1.126830381852889e-06, "epoch": 0.9346203602348493, "step": 2348}, {"loss": 0.35682618618011475, "grad_norm": 0.03925405906929878, "learning_rate": 1.1132639808145884e-06, "epoch": 0.9350184097920191, "step": 2349}, {"loss": 0.370069682598114, "grad_norm": 0.04190917905218554, "learning_rate": 1.0997788216009442e-06, "epoch": 0.9354164593491889, "step": 2350}, {"loss": 0.36935603618621826, "grad_norm": 0.04076890822447189, "learning_rate": 1.0863749266221025e-06, "epoch": 0.9358145089063589, "step": 2351}, {"loss": 0.3483363389968872, "grad_norm": 0.039017746230516064, "learning_rate": 1.073052318153167e-06, "epoch": 0.9362125584635287, "step": 2352}, {"loss": 0.3973033130168915, "grad_norm": 0.04241699510393565, "learning_rate": 1.059811018334156e-06, "epoch": 0.9366106080206986, "step": 2353}, {"loss": 0.3901476562023163, "grad_norm": 0.041912469948797784, "learning_rate": 1.04665104916995e-06, "epoch": 0.9370086575778684, "step": 2354}, {"loss": 0.34607255458831787, "grad_norm": 0.03851503821741697, "learning_rate": 1.0335724325302942e-06, "epoch": 0.9374067071350383, "step": 2355}, {"loss": 0.3396296501159668, "grad_norm": 0.03812294167112707, "learning_rate": 1.020575190149714e-06, "epoch": 0.9378047566922082, "step": 2356}, {"loss": 0.321296364068985, "grad_norm": 0.03800034152635278, "learning_rate": 1.0076593436275317e-06, "epoch": 0.938202806249378, "step": 2357}, {"loss": 0.4183874726295471, "grad_norm": 0.04114882978147291, "learning_rate": 9.948249144277833e-07, "epoch": 0.938600855806548, "step": 2358}, {"loss": 0.3402720093727112, "grad_norm": 0.04147636721441484, "learning_rate": 9.820719238792075e-07, "epoch": 0.9389989053637178, "step": 2359}, {"loss": 0.4113449454307556, "grad_norm": 0.04389985004300122, "learning_rate": 9.694003931752016e-07, "epoch": 0.9393969549208877, "step": 2360}, {"loss": 0.3492525815963745, "grad_norm": 0.037741017483427265, "learning_rate": 9.568103433738096e-07, "epoch": 0.9397950044780575, "step": 2361}, {"loss": 0.3592848777770996, "grad_norm": 0.04152104990736604, "learning_rate": 9.443017953976341e-07, "epoch": 0.9401930540352273, "step": 2362}, {"loss": 0.3726111054420471, "grad_norm": 0.04157440869763631, "learning_rate": 9.318747700338581e-07, "epoch": 0.9405911035923973, "step": 2363}, {"loss": 0.3691651225090027, "grad_norm": 0.03821759837470035, "learning_rate": 9.195292879341844e-07, "epoch": 0.9409891531495671, "step": 2364}, {"loss": 0.3608046770095825, "grad_norm": 0.04133954187093129, "learning_rate": 9.072653696148014e-07, "epoch": 0.941387202706737, "step": 2365}, {"loss": 0.35129016637802124, "grad_norm": 0.040186321376825596, "learning_rate": 8.950830354563511e-07, "epoch": 0.9417852522639069, "step": 2366}, {"loss": 0.33271390199661255, "grad_norm": 0.03571930035746617, "learning_rate": 8.829823057039e-07, "epoch": 0.9421833018210767, "step": 2367}, {"loss": 0.36158621311187744, "grad_norm": 0.03714375035456289, "learning_rate": 8.7096320046689e-07, "epoch": 0.9425813513782466, "step": 2368}, {"loss": 0.37232115864753723, "grad_norm": 0.039070568662763365, "learning_rate": 8.59025739719127e-07, "epoch": 0.9429794009354164, "step": 2369}, {"loss": 0.3363485336303711, "grad_norm": 0.03942347700399885, "learning_rate": 8.471699432987368e-07, "epoch": 0.9433774504925864, "step": 2370}, {"loss": 0.37813228368759155, "grad_norm": 0.04208194220344072, "learning_rate": 8.353958309081311e-07, "epoch": 0.9437755000497562, "step": 2371}, {"loss": 0.3492320775985718, "grad_norm": 0.04161219296217409, "learning_rate": 8.23703422113975e-07, "epoch": 0.944173549606926, "step": 2372}, {"loss": 0.3661509156227112, "grad_norm": 0.04166213255904817, "learning_rate": 8.120927363471476e-07, "epoch": 0.9445715991640959, "step": 2373}, {"loss": 0.3555467426776886, "grad_norm": 0.0411019766310233, "learning_rate": 8.005637929027309e-07, "epoch": 0.9449696487212658, "step": 2374}, {"loss": 0.3677583336830139, "grad_norm": 0.03841994764286556, "learning_rate": 7.891166109399661e-07, "epoch": 0.9453676982784357, "step": 2375}, {"loss": 0.3856412172317505, "grad_norm": 0.044722751783352506, "learning_rate": 7.777512094821915e-07, "epoch": 0.9457657478356055, "step": 2376}, {"loss": 0.3511900305747986, "grad_norm": 0.04261597426616109, "learning_rate": 7.664676074168764e-07, "epoch": 0.9461637973927755, "step": 2377}, {"loss": 0.37496575713157654, "grad_norm": 0.04053713305909003, "learning_rate": 7.552658234955379e-07, "epoch": 0.9465618469499453, "step": 2378}, {"loss": 0.39919543266296387, "grad_norm": 0.037579863703688696, "learning_rate": 7.441458763337184e-07, "epoch": 0.9469598965071151, "step": 2379}, {"loss": 0.33090728521347046, "grad_norm": 0.041062637842102796, "learning_rate": 7.331077844109635e-07, "epoch": 0.947357946064285, "step": 2380}, {"loss": 0.39112013578414917, "grad_norm": 0.04324697249412149, "learning_rate": 7.221515660707945e-07, "epoch": 0.9477559956214548, "step": 2381}, {"loss": 0.39259177446365356, "grad_norm": 0.03751465558173441, "learning_rate": 7.112772395206746e-07, "epoch": 0.9481540451786248, "step": 2382}, {"loss": 0.36136287450790405, "grad_norm": 0.04028756089298943, "learning_rate": 7.00484822831965e-07, "epoch": 0.9485520947357946, "step": 2383}, {"loss": 0.38353779911994934, "grad_norm": 0.03913161342725678, "learning_rate": 6.897743339399132e-07, "epoch": 0.9489501442929644, "step": 2384}, {"loss": 0.34877121448516846, "grad_norm": 0.03939532417681131, "learning_rate": 6.791457906436094e-07, "epoch": 0.9493481938501344, "step": 2385}, {"loss": 0.3455509543418884, "grad_norm": 0.038921788828289186, "learning_rate": 6.685992106059691e-07, "epoch": 0.9497462434073042, "step": 2386}, {"loss": 0.3918002247810364, "grad_norm": 0.04131938417873116, "learning_rate": 6.581346113537057e-07, "epoch": 0.9501442929644741, "step": 2387}, {"loss": 0.3727664351463318, "grad_norm": 0.03815891588468397, "learning_rate": 6.477520102772749e-07, "epoch": 0.9505423425216439, "step": 2388}, {"loss": 0.3610989451408386, "grad_norm": 0.042658785338190605, "learning_rate": 6.37451424630886e-07, "epoch": 0.9509403920788139, "step": 2389}, {"loss": 0.38555294275283813, "grad_norm": 0.04147407061147075, "learning_rate": 6.272328715324349e-07, "epoch": 0.9513384416359837, "step": 2390}, {"loss": 0.35941046476364136, "grad_norm": 0.03926016015410607, "learning_rate": 6.170963679634934e-07, "epoch": 0.9517364911931535, "step": 2391}, {"loss": 0.34895092248916626, "grad_norm": 0.04185753324703602, "learning_rate": 6.070419307692976e-07, "epoch": 0.9521345407503234, "step": 2392}, {"loss": 0.360871285200119, "grad_norm": 0.03904166089893308, "learning_rate": 5.970695766586876e-07, "epoch": 0.9525325903074933, "step": 2393}, {"loss": 0.32956188917160034, "grad_norm": 0.03739023511275713, "learning_rate": 5.871793222041011e-07, "epoch": 0.9529306398646632, "step": 2394}, {"loss": 0.3731682300567627, "grad_norm": 0.040316128079214586, "learning_rate": 5.773711838415241e-07, "epoch": 0.953328689421833, "step": 2395}, {"loss": 0.38456588983535767, "grad_norm": 0.040268432029881125, "learning_rate": 5.676451778705071e-07, "epoch": 0.9537267389790028, "step": 2396}, {"loss": 0.41073256731033325, "grad_norm": 0.04164356418074295, "learning_rate": 5.580013204540824e-07, "epoch": 0.9541247885361728, "step": 2397}, {"loss": 0.38050663471221924, "grad_norm": 0.04480748850862136, "learning_rate": 5.484396276187798e-07, "epoch": 0.9545228380933426, "step": 2398}, {"loss": 0.2967812120914459, "grad_norm": 0.038495347549578035, "learning_rate": 5.389601152545832e-07, "epoch": 0.9549208876505125, "step": 2399}, {"loss": 0.35244429111480713, "grad_norm": 0.039490425033786934, "learning_rate": 5.295627991149022e-07, "epoch": 0.9553189372076824, "step": 2400}, {"loss": 0.40524226427078247, "grad_norm": 0.04149273201666206, "learning_rate": 5.202476948165558e-07, "epoch": 0.9557169867648522, "step": 2401}, {"loss": 0.36410796642303467, "grad_norm": 0.03775361619891518, "learning_rate": 5.110148178397333e-07, "epoch": 0.9561150363220221, "step": 2402}, {"loss": 0.3558043837547302, "grad_norm": 0.04125267098765174, "learning_rate": 5.018641835279836e-07, "epoch": 0.9565130858791919, "step": 2403}, {"loss": 0.3205832839012146, "grad_norm": 0.03839222312256901, "learning_rate": 4.927958070881644e-07, "epoch": 0.9569111354363619, "step": 2404}, {"loss": 0.3551952838897705, "grad_norm": 0.04378365399825592, "learning_rate": 4.838097035904599e-07, "epoch": 0.9573091849935317, "step": 2405}, {"loss": 0.35669028759002686, "grad_norm": 0.04444516402577983, "learning_rate": 4.749058879683077e-07, "epoch": 0.9577072345507016, "step": 2406}, {"loss": 0.3664317727088928, "grad_norm": 0.04185405552607315, "learning_rate": 4.6608437501840517e-07, "epoch": 0.9581052841078714, "step": 2407}, {"loss": 0.36393964290618896, "grad_norm": 0.04192754923414509, "learning_rate": 4.5734517940068114e-07, "epoch": 0.9585033336650413, "step": 2408}, {"loss": 0.3866455554962158, "grad_norm": 0.041522222668048066, "learning_rate": 4.4868831563825707e-07, "epoch": 0.9589013832222112, "step": 2409}, {"loss": 0.3491613268852234, "grad_norm": 0.03669818405260546, "learning_rate": 4.4011379811743635e-07, "epoch": 0.959299432779381, "step": 2410}, {"loss": 0.40478938817977905, "grad_norm": 0.04390502647220235, "learning_rate": 4.3162164108767614e-07, "epoch": 0.9596974823365509, "step": 2411}, {"loss": 0.4001696705818176, "grad_norm": 0.04371899239660177, "learning_rate": 4.232118586615652e-07, "epoch": 0.9600955318937208, "step": 2412}, {"loss": 0.347908616065979, "grad_norm": 0.04183854886698024, "learning_rate": 4.148844648147965e-07, "epoch": 0.9604935814508906, "step": 2413}, {"loss": 0.3591424822807312, "grad_norm": 0.040172782339157345, "learning_rate": 4.066394733861445e-07, "epoch": 0.9608916310080605, "step": 2414}, {"loss": 0.3702049255371094, "grad_norm": 0.038967241283660435, "learning_rate": 3.9847689807745447e-07, "epoch": 0.9612896805652303, "step": 2415}, {"loss": 0.3660240173339844, "grad_norm": 0.0400862370041594, "learning_rate": 3.9039675245360343e-07, "epoch": 0.9616877301224003, "step": 2416}, {"loss": 0.35130783915519714, "grad_norm": 0.03829276116797206, "learning_rate": 3.823990499424723e-07, "epoch": 0.9620857796795701, "step": 2417}, {"loss": 0.40820184350013733, "grad_norm": 0.042535102075501256, "learning_rate": 3.7448380383495186e-07, "epoch": 0.96248382923674, "step": 2418}, {"loss": 0.3792387843132019, "grad_norm": 0.04272078820900196, "learning_rate": 3.6665102728490333e-07, "epoch": 0.9628818787939099, "step": 2419}, {"loss": 0.3751845061779022, "grad_norm": 0.03982388247423926, "learning_rate": 3.5890073330911433e-07, "epoch": 0.9632799283510797, "step": 2420}, {"loss": 0.33693361282348633, "grad_norm": 0.03714871827733953, "learning_rate": 3.512329347873322e-07, "epoch": 0.9636779779082496, "step": 2421}, {"loss": 0.3675621449947357, "grad_norm": 0.04006860587121641, "learning_rate": 3.436476444621806e-07, "epoch": 0.9640760274654194, "step": 2422}, {"loss": 0.36346009373664856, "grad_norm": 0.04211524733787645, "learning_rate": 3.3614487493919266e-07, "epoch": 0.9644740770225894, "step": 2423}, {"loss": 0.3682933449745178, "grad_norm": 0.043075534125626064, "learning_rate": 3.287246386867504e-07, "epoch": 0.9648721265797592, "step": 2424}, {"loss": 0.3486655354499817, "grad_norm": 0.04074363915164665, "learning_rate": 3.213869480360787e-07, "epoch": 0.965270176136929, "step": 2425}, {"loss": 0.34332460165023804, "grad_norm": 0.038763282122809266, "learning_rate": 3.141318151812345e-07, "epoch": 0.9656682256940989, "step": 2426}, {"loss": 0.31501495838165283, "grad_norm": 0.03873540405455583, "learning_rate": 3.069592521790676e-07, "epoch": 0.9660662752512688, "step": 2427}, {"loss": 0.3615865707397461, "grad_norm": 0.039564459836785186, "learning_rate": 2.998692709492212e-07, "epoch": 0.9664643248084387, "step": 2428}, {"loss": 0.344431608915329, "grad_norm": 0.03613060583585623, "learning_rate": 2.9286188327408704e-07, "epoch": 0.9668623743656085, "step": 2429}, {"loss": 0.3508828282356262, "grad_norm": 0.04041115140207347, "learning_rate": 2.859371007988054e-07, "epoch": 0.9672604239227783, "step": 2430}, {"loss": 0.35142549872398376, "grad_norm": 0.037967440633335334, "learning_rate": 2.7909493503124885e-07, "epoch": 0.9676584734799483, "step": 2431}, {"loss": 0.3711167573928833, "grad_norm": 0.039719675765815574, "learning_rate": 2.723353973419829e-07, "epoch": 0.9680565230371181, "step": 2432}, {"loss": 0.34118136763572693, "grad_norm": 0.03600189130905742, "learning_rate": 2.6565849896426073e-07, "epoch": 0.968454572594288, "step": 2433}, {"loss": 0.3486214876174927, "grad_norm": 0.03720780431572582, "learning_rate": 2.590642509940011e-07, "epoch": 0.9688526221514578, "step": 2434}, {"loss": 0.3576136827468872, "grad_norm": 0.04287850060196157, "learning_rate": 2.52552664389788e-07, "epoch": 0.9692506717086278, "step": 2435}, {"loss": 0.38516122102737427, "grad_norm": 0.04179702138201296, "learning_rate": 2.4612374997280975e-07, "epoch": 0.9696487212657976, "step": 2436}, {"loss": 0.38515520095825195, "grad_norm": 0.03913362583916956, "learning_rate": 2.3977751842688136e-07, "epoch": 0.9700467708229674, "step": 2437}, {"loss": 0.32142412662506104, "grad_norm": 0.038784842536907796, "learning_rate": 2.33513980298411e-07, "epoch": 0.9704448203801374, "step": 2438}, {"loss": 0.40969428420066833, "grad_norm": 0.04041745034211161, "learning_rate": 2.27333145996389e-07, "epoch": 0.9708428699373072, "step": 2439}, {"loss": 0.34962061047554016, "grad_norm": 0.03502026467331624, "learning_rate": 2.21235025792349e-07, "epoch": 0.9712409194944771, "step": 2440}, {"loss": 0.356606125831604, "grad_norm": 0.04019841822881757, "learning_rate": 2.1521962982038457e-07, "epoch": 0.9716389690516469, "step": 2441}, {"loss": 0.37193772196769714, "grad_norm": 0.04142247823697307, "learning_rate": 2.092869680771048e-07, "epoch": 0.9720370186088167, "step": 2442}, {"loss": 0.397469699382782, "grad_norm": 0.04220325526027369, "learning_rate": 2.0343705042162875e-07, "epoch": 0.9724350681659867, "step": 2443}, {"loss": 0.36785995960235596, "grad_norm": 0.04152654013491158, "learning_rate": 1.9766988657557994e-07, "epoch": 0.9728331177231565, "step": 2444}, {"loss": 0.39894795417785645, "grad_norm": 0.042578927361404235, "learning_rate": 1.9198548612303635e-07, "epoch": 0.9732311672803264, "step": 2445}, {"loss": 0.3810325264930725, "grad_norm": 0.04233173202253769, "learning_rate": 1.8638385851056372e-07, "epoch": 0.9736292168374963, "step": 2446}, {"loss": 0.4002472758293152, "grad_norm": 0.04162109221690288, "learning_rate": 1.8086501304714344e-07, "epoch": 0.9740272663946662, "step": 2447}, {"loss": 0.3977512717247009, "grad_norm": 0.040448876202407905, "learning_rate": 1.754289589042113e-07, "epoch": 0.974425315951836, "step": 2448}, {"loss": 0.37795108556747437, "grad_norm": 0.03877995035657585, "learning_rate": 1.7007570511560767e-07, "epoch": 0.9748233655090058, "step": 2449}, {"loss": 0.3317734897136688, "grad_norm": 0.03904857758522321, "learning_rate": 1.648052605775663e-07, "epoch": 0.9752214150661758, "step": 2450}, {"loss": 0.3267228305339813, "grad_norm": 0.03919371138531323, "learning_rate": 1.5961763404871434e-07, "epoch": 0.9756194646233456, "step": 2451}, {"loss": 0.37938857078552246, "grad_norm": 0.04207526025903996, "learning_rate": 1.545128341500446e-07, "epoch": 0.9760175141805155, "step": 2452}, {"loss": 0.3230898976325989, "grad_norm": 0.04176083636935153, "learning_rate": 1.4949086936490997e-07, "epoch": 0.9764155637376853, "step": 2453}, {"loss": 0.3678889870643616, "grad_norm": 0.039003369297261815, "learning_rate": 1.4455174803899575e-07, "epoch": 0.9768136132948552, "step": 2454}, {"loss": 0.3581514358520508, "grad_norm": 0.042705429121738026, "learning_rate": 1.3969547838031948e-07, "epoch": 0.9772116628520251, "step": 2455}, {"loss": 0.3185075521469116, "grad_norm": 0.04386234595148299, "learning_rate": 1.3492206845922005e-07, "epoch": 0.9776097124091949, "step": 2456}, {"loss": 0.37188565731048584, "grad_norm": 0.03907944422767226, "learning_rate": 1.3023152620832423e-07, "epoch": 0.9780077619663649, "step": 2457}, {"loss": 0.3586887717247009, "grad_norm": 0.03739635476258623, "learning_rate": 1.2562385942255782e-07, "epoch": 0.9784058115235347, "step": 2458}, {"loss": 0.38630592823028564, "grad_norm": 0.0431615895974549, "learning_rate": 1.2109907575911238e-07, "epoch": 0.9788038610807045, "step": 2459}, {"loss": 0.3442833423614502, "grad_norm": 0.0378146828816708, "learning_rate": 1.166571827374452e-07, "epoch": 0.9792019106378744, "step": 2460}, {"loss": 0.3455388844013214, "grad_norm": 0.04069931449293029, "learning_rate": 1.1229818773925704e-07, "epoch": 0.9795999601950443, "step": 2461}, {"loss": 0.3401136100292206, "grad_norm": 0.03697128301598013, "learning_rate": 1.0802209800849783e-07, "epoch": 0.9799980097522142, "step": 2462}, {"loss": 0.33891987800598145, "grad_norm": 0.038395962973076604, "learning_rate": 1.0382892065133321e-07, "epoch": 0.980396059309384, "step": 2463}, {"loss": 0.3648415207862854, "grad_norm": 0.04169842117636119, "learning_rate": 9.97186626361446e-08, "epoch": 0.9807941088665539, "step": 2464}, {"loss": 0.34597328305244446, "grad_norm": 0.039052289929940295, "learning_rate": 9.5691330793507e-08, "epoch": 0.9811921584237238, "step": 2465}, {"loss": 0.3732370138168335, "grad_norm": 0.04063189904334897, "learning_rate": 9.174693181620008e-08, "epoch": 0.9815902079808936, "step": 2466}, {"loss": 0.356478750705719, "grad_norm": 0.03871121192805617, "learning_rate": 8.788547225917488e-08, "epoch": 0.9819882575380635, "step": 2467}, {"loss": 0.3788043260574341, "grad_norm": 0.04014798255838189, "learning_rate": 8.41069585395482e-08, "epoch": 0.9823863070952333, "step": 2468}, {"loss": 0.3554993271827698, "grad_norm": 0.044289372126703946, "learning_rate": 8.04113969365916e-08, "epoch": 0.9827843566524033, "step": 2469}, {"loss": 0.4172239899635315, "grad_norm": 0.04042323979041821, "learning_rate": 7.679879359173691e-08, "epoch": 0.9831824062095731, "step": 2470}, {"loss": 0.35699933767318726, "grad_norm": 0.04077588647406031, "learning_rate": 7.326915450854288e-08, "epoch": 0.9835804557667429, "step": 2471}, {"loss": 0.36560678482055664, "grad_norm": 0.03848813212361684, "learning_rate": 6.982248555270076e-08, "epoch": 0.9839785053239128, "step": 2472}, {"loss": 0.36856240034103394, "grad_norm": 0.03827018715875809, "learning_rate": 6.645879245200659e-08, "epoch": 0.9843765548810827, "step": 2473}, {"loss": 0.34799274802207947, "grad_norm": 0.041393350320774076, "learning_rate": 6.317808079637777e-08, "epoch": 0.9847746044382526, "step": 2474}, {"loss": 0.298938125371933, "grad_norm": 0.03834160825171602, "learning_rate": 5.998035603782537e-08, "epoch": 0.9851726539954224, "step": 2475}, {"loss": 0.3511449694633484, "grad_norm": 0.03886909307450373, "learning_rate": 5.686562349044855e-08, "epoch": 0.9855707035525924, "step": 2476}, {"loss": 0.36918023228645325, "grad_norm": 0.039815222814762845, "learning_rate": 5.3833888330423465e-08, "epoch": 0.9859687531097622, "step": 2477}, {"loss": 0.33084994554519653, "grad_norm": 0.03809113557594861, "learning_rate": 5.088515559600881e-08, "epoch": 0.986366802666932, "step": 2478}, {"loss": 0.3847096562385559, "grad_norm": 0.039720080223766684, "learning_rate": 4.801943018751254e-08, "epoch": 0.9867648522241019, "step": 2479}, {"loss": 0.3590099811553955, "grad_norm": 0.04322088729765507, "learning_rate": 4.523671686731401e-08, "epoch": 0.9871629017812718, "step": 2480}, {"loss": 0.3486193120479584, "grad_norm": 0.04018057536868011, "learning_rate": 4.253702025983075e-08, "epoch": 0.9875609513384417, "step": 2481}, {"loss": 0.3415430784225464, "grad_norm": 0.03533245483021581, "learning_rate": 3.9920344851507306e-08, "epoch": 0.9879590008956115, "step": 2482}, {"loss": 0.35928410291671753, "grad_norm": 0.039600737114022726, "learning_rate": 3.738669499084857e-08, "epoch": 0.9883570504527813, "step": 2483}, {"loss": 0.39767348766326904, "grad_norm": 0.04338060554849439, "learning_rate": 3.493607488836426e-08, "epoch": 0.9887551000099513, "step": 2484}, {"loss": 0.35789570212364197, "grad_norm": 0.04041511084432876, "learning_rate": 3.2568488616591117e-08, "epoch": 0.9891531495671211, "step": 2485}, {"loss": 0.40068182349205017, "grad_norm": 0.04406557870869541, "learning_rate": 3.028394011006519e-08, "epoch": 0.989551199124291, "step": 2486}, {"loss": 0.32634541392326355, "grad_norm": 0.03961651799223283, "learning_rate": 2.8082433165338428e-08, "epoch": 0.9899492486814608, "step": 2487}, {"loss": 0.36932218074798584, "grad_norm": 0.03940870107884029, "learning_rate": 2.5963971440962077e-08, "epoch": 0.9903472982386307, "step": 2488}, {"loss": 0.3203428387641907, "grad_norm": 0.03645728747980246, "learning_rate": 2.3928558457475547e-08, "epoch": 0.9907453477958006, "step": 2489}, {"loss": 0.38947442173957825, "grad_norm": 0.04028374295363278, "learning_rate": 2.1976197597400884e-08, "epoch": 0.9911433973529704, "step": 2490}, {"loss": 0.37237486243247986, "grad_norm": 0.04132749441681188, "learning_rate": 2.01068921052483e-08, "epoch": 0.9915414469101403, "step": 2491}, {"loss": 0.38364145159721375, "grad_norm": 0.04463813505806349, "learning_rate": 1.8320645087499534e-08, "epoch": 0.9919394964673102, "step": 2492}, {"loss": 0.3853393495082855, "grad_norm": 0.03949089979301168, "learning_rate": 1.6617459512607846e-08, "epoch": 0.9923375460244801, "step": 2493}, {"loss": 0.39245373010635376, "grad_norm": 0.037535188963455125, "learning_rate": 1.4997338210992473e-08, "epoch": 0.9927355955816499, "step": 2494}, {"loss": 0.38976407051086426, "grad_norm": 0.038704977020486996, "learning_rate": 1.3460283875027512e-08, "epoch": 0.9931336451388197, "step": 2495}, {"loss": 0.32568976283073425, "grad_norm": 0.03996669065219372, "learning_rate": 1.2006299059047488e-08, "epoch": 0.9935316946959897, "step": 2496}, {"loss": 0.37703824043273926, "grad_norm": 0.039869557500514644, "learning_rate": 1.063538617934734e-08, "epoch": 0.9939297442531595, "step": 2497}, {"loss": 0.36218681931495667, "grad_norm": 0.04235031839391709, "learning_rate": 9.347547514160227e-09, "epoch": 0.9943277938103294, "step": 2498}, {"loss": 0.3461621105670929, "grad_norm": 0.04173899530068073, "learning_rate": 8.142785203657521e-09, "epoch": 0.9947258433674993, "step": 2499}, {"loss": 0.4059831202030182, "grad_norm": 0.043046381064133936, "learning_rate": 7.021101249959916e-09, "epoch": 0.9951238929246691, "step": 2500}, {"loss": 0.33172449469566345, "grad_norm": 0.03585723704650251, "learning_rate": 5.982497517131868e-09, "epoch": 0.995521942481839, "step": 2501}, {"loss": 0.396645188331604, "grad_norm": 0.03934396131060645, "learning_rate": 5.026975731164951e-09, "epoch": 0.9959199920390088, "step": 2502}, {"loss": 0.34626346826553345, "grad_norm": 0.03824098628591386, "learning_rate": 4.154537479966747e-09, "epoch": 0.9963180415961788, "step": 2503}, {"loss": 0.3554512560367584, "grad_norm": 0.04037773408562237, "learning_rate": 3.365184213405259e-09, "epoch": 0.9967160911533486, "step": 2504}, {"loss": 0.35711318254470825, "grad_norm": 0.04100077440305889, "learning_rate": 2.6589172432478493e-09, "epoch": 0.9971141407105185, "step": 2505}, {"loss": 0.341941773891449, "grad_norm": 0.03982453174497287, "learning_rate": 2.0357377431945435e-09, "epoch": 0.9975121902676883, "step": 2506}, {"loss": 0.34731149673461914, "grad_norm": 0.04018227051840291, "learning_rate": 1.4956467488780324e-09, "epoch": 0.9979102398248582, "step": 2507}, {"loss": 0.3974302411079407, "grad_norm": 0.04152030822934574, "learning_rate": 1.0386451578303647e-09, "epoch": 0.9983082893820281, "step": 2508}, {"loss": 0.3898221552371979, "grad_norm": 0.04288373443846666, "learning_rate": 6.647337295218048e-10, "epoch": 0.9987063389391979, "step": 2509}, {"loss": 0.38624459505081177, "grad_norm": 0.04152049564925222, "learning_rate": 3.7391308532752635e-10, "epoch": 0.9991043884963678, "step": 2510}, {"loss": 0.3759276866912842, "grad_norm": 0.040269712512132606, "learning_rate": 1.6618370854981636e-10, "epoch": 0.9995024380535377, "step": 2511}, {"loss": 0.3819911479949951, "grad_norm": 0.041235793116552864, "learning_rate": 4.1545944395871005e-11, "epoch": 0.9999004876107075, "step": 2512}, {"loss": 0.3179854154586792, "grad_norm": 0.13276110160399002, "learning_rate": 0.0, "epoch": 1.0, "step": 2513}, {"train_runtime": 24848.0711, "train_samples_per_second": 12.94, "train_steps_per_second": 0.101, "total_flos": 7708589808975872.0, "train_loss": 0.38710029800954304, "epoch": 1.0, "step": 2513}], "memory": 77.65234375}