diff --git "a/sg/antmaze-medium-diverse-v0/2/debug.log" "b/sg/antmaze-medium-diverse-v0/2/debug.log" new file mode 100644--- /dev/null +++ "b/sg/antmaze-medium-diverse-v0/2/debug.log" @@ -0,0 +1,52000 @@ +2022-05-10 13:11:23.106994 PDT | [2] Epoch -1000 finished +---------------------------------- ---------------- +epoch -1000 +replay_buffer/size 999033 +trainer/num train calls 1000 +trainer/Policy Loss 40.1482 +trainer/Log Pis Mean -5.3444 +trainer/Log Pis Std 0.600067 +trainer/Log Pis Max -3.44991 +trainer/Log Pis Min -7.22494 +trainer/policy/mean Mean -9.77952e-05 +trainer/policy/mean Std 0.000200975 +trainer/policy/mean Max 0.000558803 +trainer/policy/mean Min -0.000967621 +trainer/policy/normal/std Mean 1.00026 +trainer/policy/normal/std Std 0.000417196 +trainer/policy/normal/std Max 1.00148 +trainer/policy/normal/std Min 0.999263 +trainer/policy/normal/log_std Mean 0.000264837 +trainer/policy/normal/log_std Std 0.000417073 +trainer/policy/normal/log_std Max 0.0014827 +trainer/policy/normal/log_std Min -0.00073776 +eval/num steps total 1000 +eval/num paths total 1 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -6.40302e-05 +eval/Actions Std 0.000139387 +eval/Actions Max 0.000250722 +eval/Actions Min -0.000500003 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 5.63631 +time/logging (s) 0.0042437 +time/sampling batch (s) 0.537916 +time/saving (s) 0.00444259 +time/training (s) 7.28961 +time/epoch (s) 13.4725 +time/total (s) 49.6823 +Epoch -1000 +---------------------------------- ---------------- +2022-05-10 13:11:32.472911 PDT | [2] Epoch -999 finished +---------------------------------- --------------- +epoch -999 +replay_buffer/size 999033 +trainer/num train calls 2000 +trainer/Policy Loss -16.2075 +trainer/Log Pis Mean 21.3515 +trainer/Log Pis Std 14.0183 +trainer/Log Pis Max 88.3228 +trainer/Log Pis Min -9.48849 +trainer/policy/mean Mean -0.0429971 +trainer/policy/mean Std 0.724921 +trainer/policy/mean Max 0.994676 +trainer/policy/mean Min -0.996872 +trainer/policy/normal/std Mean 3.95378 +trainer/policy/normal/std Std 0.549754 +trainer/policy/normal/std Max 6.88326 +trainer/policy/normal/std Min 2.43765 +trainer/policy/normal/log_std Mean 1.36528 +trainer/policy/normal/log_std Std 0.136533 +trainer/policy/normal/log_std Max 1.92909 +trainer/policy/normal/log_std Min 0.891035 +eval/num steps total 2000 +eval/num paths total 2 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0697614 +eval/Actions Std 0.714989 +eval/Actions Max 0.980349 +eval/Actions Min -0.987213 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61222 +time/logging (s) 0.00395821 +time/sampling batch (s) 0.531495 +time/saving (s) 0.00399206 +time/training (s) 6.19393 +time/epoch (s) 9.34559 +time/total (s) 59.0316 +Epoch -999 +---------------------------------- --------------- +2022-05-10 13:11:41.932432 PDT | [2] Epoch -998 finished +---------------------------------- --------------- +epoch -998 +replay_buffer/size 999033 +trainer/num train calls 3000 +trainer/Policy Loss -17.8439 +trainer/Log Pis Mean 21.9818 +trainer/Log Pis Std 13.2133 +trainer/Log Pis Max 61.0078 +trainer/Log Pis Min -8.65245 +trainer/policy/mean Mean -0.0623761 +trainer/policy/mean Std 0.822229 +trainer/policy/mean Max 0.99998 +trainer/policy/mean Min -0.999969 +trainer/policy/normal/std Mean 3.51275 +trainer/policy/normal/std Std 0.461424 +trainer/policy/normal/std Max 5.56093 +trainer/policy/normal/std Min 1.82502 +trainer/policy/normal/log_std Mean 1.24752 +trainer/policy/normal/log_std Std 0.134737 +trainer/policy/normal/log_std Max 1.71576 +trainer/policy/normal/log_std Min 0.601591 +eval/num steps total 3000 +eval/num paths total 3 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0569768 +eval/Actions Std 0.751668 +eval/Actions Max 0.999773 +eval/Actions Min -0.998574 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.82445 +time/logging (s) 0.00416109 +time/sampling batch (s) 0.282338 +time/saving (s) 0.00414533 +time/training (s) 6.32498 +time/epoch (s) 9.44007 +time/total (s) 68.4749 +Epoch -998 +---------------------------------- --------------- +2022-05-10 13:11:52.629748 PDT | [2] Epoch -997 finished +---------------------------------- --------------- +epoch -997 +replay_buffer/size 999033 +trainer/num train calls 4000 +trainer/Policy Loss -17.6762 +trainer/Log Pis Mean 22.5034 +trainer/Log Pis Std 12.8992 +trainer/Log Pis Max 70.9995 +trainer/Log Pis Min -6.9035 +trainer/policy/mean Mean -0.0460457 +trainer/policy/mean Std 0.854169 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 3.36455 +trainer/policy/normal/std Std 0.456894 +trainer/policy/normal/std Max 6.04514 +trainer/policy/normal/std Min 1.61449 +trainer/policy/normal/log_std Mean 1.20368 +trainer/policy/normal/log_std Std 0.140847 +trainer/policy/normal/log_std Max 1.79925 +trainer/policy/normal/log_std Min 0.479022 +eval/num steps total 4000 +eval/num paths total 4 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.106855 +eval/Actions Std 0.734471 +eval/Actions Max 0.999511 +eval/Actions Min -0.997792 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5091 +time/logging (s) 0.00375996 +time/sampling batch (s) 0.283376 +time/saving (s) 0.00369728 +time/training (s) 7.87735 +time/epoch (s) 10.6773 +time/total (s) 79.1555 +Epoch -997 +---------------------------------- --------------- +2022-05-10 13:12:02.474864 PDT | [2] Epoch -996 finished +---------------------------------- --------------- +epoch -996 +replay_buffer/size 999033 +trainer/num train calls 5000 +trainer/Policy Loss -18.8544 +trainer/Log Pis Mean 23.7633 +trainer/Log Pis Std 12.7228 +trainer/Log Pis Max 72.6488 +trainer/Log Pis Min -10.9404 +trainer/policy/mean Mean -0.0415403 +trainer/policy/mean Std 0.86702 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 3.26299 +trainer/policy/normal/std Std 0.438324 +trainer/policy/normal/std Max 5.56421 +trainer/policy/normal/std Min 1.48155 +trainer/policy/normal/log_std Mean 1.17305 +trainer/policy/normal/log_std Std 0.141685 +trainer/policy/normal/log_std Max 1.71636 +trainer/policy/normal/log_std Min 0.393089 +eval/num steps total 5000 +eval/num paths total 5 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.210527 +eval/Actions Std 0.780688 +eval/Actions Max 0.99997 +eval/Actions Min -0.99994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.30289 +time/logging (s) 0.00370974 +time/sampling batch (s) 0.534097 +time/saving (s) 0.00349187 +time/training (s) 6.98146 +time/epoch (s) 9.82565 +time/total (s) 88.9841 +Epoch -996 +---------------------------------- --------------- +2022-05-10 13:12:12.525608 PDT | [2] Epoch -995 finished +---------------------------------- --------------- +epoch -995 +replay_buffer/size 999033 +trainer/num train calls 6000 +trainer/Policy Loss -18.0485 +trainer/Log Pis Mean 21.4366 +trainer/Log Pis Std 12.9893 +trainer/Log Pis Max 62.9294 +trainer/Log Pis Min -10.7311 +trainer/policy/mean Mean -0.0595294 +trainer/policy/mean Std 0.877267 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 3.15566 +trainer/policy/normal/std Std 0.419376 +trainer/policy/normal/std Max 5.37724 +trainer/policy/normal/std Min 1.37792 +trainer/policy/normal/log_std Mean 1.13957 +trainer/policy/normal/log_std Std 0.142736 +trainer/policy/normal/log_std Max 1.68218 +trainer/policy/normal/log_std Min 0.320577 +eval/num steps total 6000 +eval/num paths total 6 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.274431 +eval/Actions Std 0.840046 +eval/Actions Max 0.999989 +eval/Actions Min -0.999988 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.3355 +time/logging (s) 0.00419878 +time/sampling batch (s) 0.536764 +time/saving (s) 0.00424453 +time/training (s) 7.15094 +time/epoch (s) 10.0316 +time/total (s) 99.0187 +Epoch -995 +---------------------------------- --------------- +2022-05-10 13:12:23.680576 PDT | [2] Epoch -994 finished +---------------------------------- --------------- +epoch -994 +replay_buffer/size 999033 +trainer/num train calls 7000 +trainer/Policy Loss -18.616 +trainer/Log Pis Mean 23.2933 +trainer/Log Pis Std 12.971 +trainer/Log Pis Max 59.7468 +trainer/Log Pis Min -11.6601 +trainer/policy/mean Mean -0.0520734 +trainer/policy/mean Std 0.885065 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 3.08606 +trainer/policy/normal/std Std 0.402593 +trainer/policy/normal/std Max 5.63571 +trainer/policy/normal/std Min 1.31643 +trainer/policy/normal/log_std Mean 1.11755 +trainer/policy/normal/log_std Std 0.141239 +trainer/policy/normal/log_std Max 1.72912 +trainer/policy/normal/log_std Min 0.274924 +eval/num steps total 7000 +eval/num paths total 7 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.177582 +eval/Actions Std 0.888708 +eval/Actions Max 0.999999 +eval/Actions Min -0.999965 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55179 +time/logging (s) 0.0042143 +time/sampling batch (s) 0.54315 +time/saving (s) 0.00419891 +time/training (s) 8.03123 +time/epoch (s) 11.1346 +time/total (s) 110.157 +Epoch -994 +---------------------------------- --------------- +2022-05-10 13:12:33.501729 PDT | [2] Epoch -993 finished +---------------------------------- --------------- +epoch -993 +replay_buffer/size 999033 +trainer/num train calls 8000 +trainer/Policy Loss -18.4765 +trainer/Log Pis Mean 24.5651 +trainer/Log Pis Std 12.8777 +trainer/Log Pis Max 73.6987 +trainer/Log Pis Min -8.2879 +trainer/policy/mean Mean -0.0568327 +trainer/policy/mean Std 0.890803 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 3.18069 +trainer/policy/normal/std Std 0.435058 +trainer/policy/normal/std Max 4.65736 +trainer/policy/normal/std Min 1.17813 +trainer/policy/normal/log_std Mean 1.14644 +trainer/policy/normal/log_std Std 0.152219 +trainer/policy/normal/log_std Max 1.53845 +trainer/policy/normal/log_std Min 0.163927 +eval/num steps total 8000 +eval/num paths total 8 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0559076 +eval/Actions Std 0.890262 +eval/Actions Max 0.999997 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59714 +time/logging (s) 0.00396353 +time/sampling batch (s) 0.288331 +time/saving (s) 0.00423625 +time/training (s) 6.90721 +time/epoch (s) 9.80087 +time/total (s) 119.961 +Epoch -993 +---------------------------------- --------------- +2022-05-10 13:12:43.314733 PDT | [2] Epoch -992 finished +---------------------------------- --------------- +epoch -992 +replay_buffer/size 999033 +trainer/num train calls 9000 +trainer/Policy Loss -18.1601 +trainer/Log Pis Mean 22.8774 +trainer/Log Pis Std 13.1889 +trainer/Log Pis Max 69.6906 +trainer/Log Pis Min -9.91722 +trainer/policy/mean Mean -0.0560644 +trainer/policy/mean Std 0.886278 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 3.1183 +trainer/policy/normal/std Std 0.457591 +trainer/policy/normal/std Max 5.34432 +trainer/policy/normal/std Min 1.26061 +trainer/policy/normal/log_std Mean 1.12532 +trainer/policy/normal/log_std Std 0.160295 +trainer/policy/normal/log_std Max 1.67603 +trainer/policy/normal/log_std Min 0.231598 +eval/num steps total 9000 +eval/num paths total 9 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.106918 +eval/Actions Std 0.892172 +eval/Actions Max 0.999999 +eval/Actions Min -0.999988 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61515 +time/logging (s) 0.00423247 +time/sampling batch (s) 0.537127 +time/saving (s) 0.00427778 +time/training (s) 6.63263 +time/epoch (s) 9.79341 +time/total (s) 129.758 +Epoch -992 +---------------------------------- --------------- +2022-05-10 13:12:53.694418 PDT | [2] Epoch -991 finished +---------------------------------- --------------- +epoch -991 +replay_buffer/size 999033 +trainer/num train calls 10000 +trainer/Policy Loss -19.1212 +trainer/Log Pis Mean 23.7423 +trainer/Log Pis Std 12.5377 +trainer/Log Pis Max 69.0004 +trainer/Log Pis Min -5.18642 +trainer/policy/mean Mean -0.0287806 +trainer/policy/mean Std 0.892429 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 3.06541 +trainer/policy/normal/std Std 0.422556 +trainer/policy/normal/std Max 4.62017 +trainer/policy/normal/std Min 1.13114 +trainer/policy/normal/log_std Mean 1.10923 +trainer/policy/normal/log_std Std 0.154911 +trainer/policy/normal/log_std Max 1.53043 +trainer/policy/normal/log_std Min 0.123226 +eval/num steps total 10000 +eval/num paths total 10 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.115092 +eval/Actions Std 0.904419 +eval/Actions Max 0.999996 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.46272 +time/logging (s) 0.00477452 +time/sampling batch (s) 0.533549 +time/saving (s) 0.004107 +time/training (s) 7.35535 +time/epoch (s) 10.3605 +time/total (s) 140.121 +Epoch -991 +---------------------------------- --------------- +2022-05-10 13:13:04.558357 PDT | [2] Epoch -990 finished +---------------------------------- --------------- +epoch -990 +replay_buffer/size 999033 +trainer/num train calls 11000 +trainer/Policy Loss -18.4329 +trainer/Log Pis Mean 23.5207 +trainer/Log Pis Std 12.1338 +trainer/Log Pis Max 68.6735 +trainer/Log Pis Min -2.94076 +trainer/policy/mean Mean -0.0340909 +trainer/policy/mean Std 0.889638 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 3.09586 +trainer/policy/normal/std Std 0.453761 +trainer/policy/normal/std Max 5.64566 +trainer/policy/normal/std Min 1.17224 +trainer/policy/normal/log_std Mean 1.11776 +trainer/policy/normal/log_std Std 0.163914 +trainer/policy/normal/log_std Max 1.73089 +trainer/policy/normal/log_std Min 0.158917 +eval/num steps total 11000 +eval/num paths total 11 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0991149 +eval/Actions Std 0.902808 +eval/Actions Max 0.999996 +eval/Actions Min -0.999969 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.79914 +time/logging (s) 0.00392985 +time/sampling batch (s) 0.289684 +time/saving (s) 0.00382329 +time/training (s) 7.74637 +time/epoch (s) 10.8429 +time/total (s) 150.968 +Epoch -990 +---------------------------------- --------------- +2022-05-10 13:13:14.262661 PDT | [2] Epoch -989 finished +---------------------------------- --------------- +epoch -989 +replay_buffer/size 999033 +trainer/num train calls 12000 +trainer/Policy Loss -19.3115 +trainer/Log Pis Mean 23.3691 +trainer/Log Pis Std 13.4176 +trainer/Log Pis Max 65.4398 +trainer/Log Pis Min -6.60769 +trainer/policy/mean Mean -0.0360632 +trainer/policy/mean Std 0.892012 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.96414 +trainer/policy/normal/std Std 0.412971 +trainer/policy/normal/std Max 4.81993 +trainer/policy/normal/std Min 1.11147 +trainer/policy/normal/log_std Mean 1.07547 +trainer/policy/normal/log_std Std 0.155792 +trainer/policy/normal/log_std Max 1.57276 +trainer/policy/normal/log_std Min 0.105683 +eval/num steps total 12000 +eval/num paths total 12 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.143383 +eval/Actions Std 0.890851 +eval/Actions Max 0.999997 +eval/Actions Min -0.999969 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69546 +time/logging (s) 0.00372432 +time/sampling batch (s) 0.285146 +time/saving (s) 0.00350581 +time/training (s) 6.69682 +time/epoch (s) 9.68466 +time/total (s) 160.655 +Epoch -989 +---------------------------------- --------------- +2022-05-10 13:13:25.502000 PDT | [2] Epoch -988 finished +---------------------------------- --------------- +epoch -988 +replay_buffer/size 999033 +trainer/num train calls 13000 +trainer/Policy Loss -18.524 +trainer/Log Pis Mean 24.6448 +trainer/Log Pis Std 12.7654 +trainer/Log Pis Max 73.9122 +trainer/Log Pis Min -8.64885 +trainer/policy/mean Mean -0.0344594 +trainer/policy/mean Std 0.897703 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 3.02619 +trainer/policy/normal/std Std 0.435031 +trainer/policy/normal/std Max 5.35636 +trainer/policy/normal/std Min 1.09346 +trainer/policy/normal/log_std Mean 1.09519 +trainer/policy/normal/log_std Std 0.163859 +trainer/policy/normal/log_std Max 1.67828 +trainer/policy/normal/log_std Min 0.0893489 +eval/num steps total 13000 +eval/num paths total 13 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.103372 +eval/Actions Std 0.910058 +eval/Actions Max 1 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.71634 +time/logging (s) 0.00417867 +time/sampling batch (s) 0.316532 +time/saving (s) 0.00413511 +time/training (s) 8.17883 +time/epoch (s) 11.22 +time/total (s) 171.878 +Epoch -988 +---------------------------------- --------------- +2022-05-10 13:13:35.082848 PDT | [2] Epoch -987 finished +---------------------------------- --------------- +epoch -987 +replay_buffer/size 999033 +trainer/num train calls 14000 +trainer/Policy Loss -18.0758 +trainer/Log Pis Mean 24.6155 +trainer/Log Pis Std 13.134 +trainer/Log Pis Max 65.4046 +trainer/Log Pis Min -4.51072 +trainer/policy/mean Mean -0.0367955 +trainer/policy/mean Std 0.894855 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 3.10575 +trainer/policy/normal/std Std 0.483502 +trainer/policy/normal/std Max 5.98316 +trainer/policy/normal/std Min 1.03261 +trainer/policy/normal/log_std Mean 1.11928 +trainer/policy/normal/log_std Std 0.175456 +trainer/policy/normal/log_std Max 1.78895 +trainer/policy/normal/log_std Min 0.0320884 +eval/num steps total 14000 +eval/num paths total 14 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0940771 +eval/Actions Std 0.817463 +eval/Actions Max 1 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.72438 +time/logging (s) 0.00377973 +time/sampling batch (s) 0.281314 +time/saving (s) 0.00363037 +time/training (s) 6.54785 +time/epoch (s) 9.56095 +time/total (s) 181.443 +Epoch -987 +---------------------------------- --------------- +2022-05-10 13:13:44.815600 PDT | [2] Epoch -986 finished +---------------------------------- --------------- +epoch -986 +replay_buffer/size 999033 +trainer/num train calls 15000 +trainer/Policy Loss -19.2815 +trainer/Log Pis Mean 24.9388 +trainer/Log Pis Std 12.802 +trainer/Log Pis Max 61.2924 +trainer/Log Pis Min -7.56521 +trainer/policy/mean Mean -0.0399317 +trainer/policy/mean Std 0.896253 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 3.04646 +trainer/policy/normal/std Std 0.46638 +trainer/policy/normal/std Max 5.85903 +trainer/policy/normal/std Min 1.002 +trainer/policy/normal/log_std Mean 1.10036 +trainer/policy/normal/log_std Std 0.173607 +trainer/policy/normal/log_std Max 1.76798 +trainer/policy/normal/log_std Min 0.0019975 +eval/num steps total 15000 +eval/num paths total 15 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.149948 +eval/Actions Std 0.889125 +eval/Actions Max 0.999998 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58973 +time/logging (s) 0.00374423 +time/sampling batch (s) 0.281119 +time/saving (s) 0.00344051 +time/training (s) 6.83539 +time/epoch (s) 9.71342 +time/total (s) 191.159 +Epoch -986 +---------------------------------- --------------- +2022-05-10 13:13:54.659165 PDT | [2] Epoch -985 finished +---------------------------------- --------------- +epoch -985 +replay_buffer/size 999033 +trainer/num train calls 16000 +trainer/Policy Loss -17.2484 +trainer/Log Pis Mean 22.9189 +trainer/Log Pis Std 12.6753 +trainer/Log Pis Max 70.6024 +trainer/Log Pis Min -12.0014 +trainer/policy/mean Mean -0.0472915 +trainer/policy/mean Std 0.890874 +trainer/policy/mean Max 0.99998 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.95069 +trainer/policy/normal/std Std 0.467328 +trainer/policy/normal/std Max 4.88943 +trainer/policy/normal/std Min 0.927427 +trainer/policy/normal/log_std Mean 1.06695 +trainer/policy/normal/log_std Std 0.1843 +trainer/policy/normal/log_std Max 1.58708 +trainer/policy/normal/log_std Min -0.0753412 +eval/num steps total 16000 +eval/num paths total 16 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.37434 +eval/Actions Std 0.894655 +eval/Actions Max 1 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.43697 +time/logging (s) 0.00376576 +time/sampling batch (s) 0.531024 +time/saving (s) 0.00344712 +time/training (s) 6.84931 +time/epoch (s) 9.82451 +time/total (s) 200.986 +Epoch -985 +---------------------------------- --------------- +2022-05-10 13:14:04.546673 PDT | [2] Epoch -984 finished +---------------------------------- --------------- +epoch -984 +replay_buffer/size 999033 +trainer/num train calls 17000 +trainer/Policy Loss -19.1242 +trainer/Log Pis Mean 24.8877 +trainer/Log Pis Std 13.6123 +trainer/Log Pis Max 89.4047 +trainer/Log Pis Min -9.61481 +trainer/policy/mean Mean -0.0578969 +trainer/policy/mean Std 0.89739 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 3.08159 +trainer/policy/normal/std Std 0.483795 +trainer/policy/normal/std Max 4.82449 +trainer/policy/normal/std Min 0.887778 +trainer/policy/normal/log_std Mean 1.11033 +trainer/policy/normal/log_std Std 0.185543 +trainer/policy/normal/log_std Max 1.5737 +trainer/policy/normal/log_std Min -0.119034 +eval/num steps total 17000 +eval/num paths total 17 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.195047 +eval/Actions Std 0.944118 +eval/Actions Max 0.999997 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.45653 +time/logging (s) 0.00376822 +time/sampling batch (s) 0.281708 +time/saving (s) 0.00366006 +time/training (s) 7.12262 +time/epoch (s) 9.86828 +time/total (s) 210.858 +Epoch -984 +---------------------------------- --------------- +2022-05-10 13:14:14.209357 PDT | [2] Epoch -983 finished +---------------------------------- --------------- +epoch -983 +replay_buffer/size 999033 +trainer/num train calls 18000 +trainer/Policy Loss -19.1104 +trainer/Log Pis Mean 25.0319 +trainer/Log Pis Std 12.8818 +trainer/Log Pis Max 67.1483 +trainer/Log Pis Min -5.60967 +trainer/policy/mean Mean -0.00848001 +trainer/policy/mean Std 0.899873 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 3.02998 +trainer/policy/normal/std Std 0.46177 +trainer/policy/normal/std Max 4.78252 +trainer/policy/normal/std Min 0.956727 +trainer/policy/normal/log_std Mean 1.09456 +trainer/policy/normal/log_std Std 0.177591 +trainer/policy/normal/log_std Max 1.56497 +trainer/policy/normal/log_std Min -0.0442374 +eval/num steps total 18000 +eval/num paths total 18 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.13495 +eval/Actions Std 0.8794 +eval/Actions Max 0.999996 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.18496 +time/logging (s) 0.00374465 +time/sampling batch (s) 0.281394 +time/saving (s) 0.00349237 +time/training (s) 7.16974 +time/epoch (s) 9.64333 +time/total (s) 220.504 +Epoch -983 +---------------------------------- --------------- +2022-05-10 13:14:24.557875 PDT | [2] Epoch -982 finished +---------------------------------- --------------- +epoch -982 +replay_buffer/size 999033 +trainer/num train calls 19000 +trainer/Policy Loss -19.0392 +trainer/Log Pis Mean 23.7603 +trainer/Log Pis Std 12.2381 +trainer/Log Pis Max 68.2948 +trainer/Log Pis Min -8.40098 +trainer/policy/mean Mean -0.00815422 +trainer/policy/mean Std 0.897088 +trainer/policy/mean Max 1 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 3.00496 +trainer/policy/normal/std Std 0.461048 +trainer/policy/normal/std Max 4.8718 +trainer/policy/normal/std Min 0.961456 +trainer/policy/normal/log_std Mean 1.08629 +trainer/policy/normal/log_std Std 0.176404 +trainer/policy/normal/log_std Max 1.58346 +trainer/policy/normal/log_std Min -0.0393067 +eval/num steps total 19000 +eval/num paths total 19 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.309309 +eval/Actions Std 0.872432 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68291 +time/logging (s) 0.00409975 +time/sampling batch (s) 0.535649 +time/saving (s) 0.00412065 +time/training (s) 7.10245 +time/epoch (s) 10.3292 +time/total (s) 230.836 +Epoch -982 +---------------------------------- --------------- +2022-05-10 13:14:35.722606 PDT | [2] Epoch -981 finished +---------------------------------- --------------- +epoch -981 +replay_buffer/size 999033 +trainer/num train calls 20000 +trainer/Policy Loss -18.9544 +trainer/Log Pis Mean 24.0278 +trainer/Log Pis Std 12.8958 +trainer/Log Pis Max 65.027 +trainer/Log Pis Min -14.2901 +trainer/policy/mean Mean -0.0237033 +trainer/policy/mean Std 0.895324 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 3.02057 +trainer/policy/normal/std Std 0.4815 +trainer/policy/normal/std Max 5.29619 +trainer/policy/normal/std Min 0.885011 +trainer/policy/normal/log_std Mean 1.09037 +trainer/policy/normal/log_std Std 0.183702 +trainer/policy/normal/log_std Max 1.66699 +trainer/policy/normal/log_std Min -0.122156 +eval/num steps total 20000 +eval/num paths total 20 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.227051 +eval/Actions Std 0.948425 +eval/Actions Max 1 +eval/Actions Min -0.999987 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.80541 +time/logging (s) 0.00374507 +time/sampling batch (s) 0.284058 +time/saving (s) 0.00347152 +time/training (s) 8.04791 +time/epoch (s) 11.1446 +time/total (s) 241.984 +Epoch -981 +---------------------------------- --------------- +2022-05-10 13:14:46.820866 PDT | [2] Epoch -980 finished +---------------------------------- --------------- +epoch -980 +replay_buffer/size 999033 +trainer/num train calls 21000 +trainer/Policy Loss -19.496 +trainer/Log Pis Mean 24.182 +trainer/Log Pis Std 13.1484 +trainer/Log Pis Max 66.2424 +trainer/Log Pis Min -4.89254 +trainer/policy/mean Mean -0.0264471 +trainer/policy/mean Std 0.893576 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 3.04516 +trainer/policy/normal/std Std 0.493873 +trainer/policy/normal/std Max 4.96534 +trainer/policy/normal/std Min 0.902693 +trainer/policy/normal/log_std Mean 1.09728 +trainer/policy/normal/log_std Std 0.193111 +trainer/policy/normal/log_std Max 1.60248 +trainer/policy/normal/log_std Min -0.102373 +eval/num steps total 21000 +eval/num paths total 21 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0549672 +eval/Actions Std 0.926496 +eval/Actions Max 0.999999 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.85974 +time/logging (s) 0.00372983 +time/sampling batch (s) 0.284878 +time/saving (s) 0.00382495 +time/training (s) 7.92655 +time/epoch (s) 11.0787 +time/total (s) 253.066 +Epoch -980 +---------------------------------- --------------- +2022-05-10 13:14:56.790608 PDT | [2] Epoch -979 finished +---------------------------------- --------------- +epoch -979 +replay_buffer/size 999033 +trainer/num train calls 22000 +trainer/Policy Loss -19.0353 +trainer/Log Pis Mean 23.7759 +trainer/Log Pis Std 12.241 +trainer/Log Pis Max 69.7085 +trainer/Log Pis Min -10.2748 +trainer/policy/mean Mean -0.0116775 +trainer/policy/mean Std 0.893368 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 3.02191 +trainer/policy/normal/std Std 0.463952 +trainer/policy/normal/std Max 4.53496 +trainer/policy/normal/std Min 0.903663 +trainer/policy/normal/log_std Mean 1.09138 +trainer/policy/normal/log_std Std 0.181724 +trainer/policy/normal/log_std Max 1.51182 +trainer/policy/normal/log_std Min -0.101299 +eval/num steps total 22000 +eval/num paths total 22 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0837853 +eval/Actions Std 0.881703 +eval/Actions Max 0.999997 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70445 +time/logging (s) 0.00379621 +time/sampling batch (s) 0.284921 +time/saving (s) 0.00351106 +time/training (s) 6.95373 +time/epoch (s) 9.95041 +time/total (s) 263.019 +Epoch -979 +---------------------------------- --------------- +2022-05-10 13:15:07.432631 PDT | [2] Epoch -978 finished +---------------------------------- --------------- +epoch -978 +replay_buffer/size 999033 +trainer/num train calls 23000 +trainer/Policy Loss -18.9381 +trainer/Log Pis Mean 23.7936 +trainer/Log Pis Std 12.9303 +trainer/Log Pis Max 65.2149 +trainer/Log Pis Min -5.00311 +trainer/policy/mean Mean -0.045204 +trainer/policy/mean Std 0.89785 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.96725 +trainer/policy/normal/std Std 0.473126 +trainer/policy/normal/std Max 4.75603 +trainer/policy/normal/std Min 0.830955 +trainer/policy/normal/log_std Mean 1.07176 +trainer/policy/normal/log_std Std 0.191008 +trainer/policy/normal/log_std Max 1.55941 +trainer/policy/normal/log_std Min -0.185179 +eval/num steps total 23000 +eval/num paths total 23 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.135703 +eval/Actions Std 0.883423 +eval/Actions Max 0.999999 +eval/Actions Min -0.999988 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60504 +time/logging (s) 0.00371758 +time/sampling batch (s) 0.53163 +time/saving (s) 0.00343639 +time/training (s) 7.47886 +time/epoch (s) 10.6227 +time/total (s) 273.645 +Epoch -978 +---------------------------------- --------------- +2022-05-10 13:15:17.128920 PDT | [2] Epoch -977 finished +---------------------------------- --------------- +epoch -977 +replay_buffer/size 999033 +trainer/num train calls 24000 +trainer/Policy Loss -20.1449 +trainer/Log Pis Mean 25.2278 +trainer/Log Pis Std 12.7703 +trainer/Log Pis Max 64.5306 +trainer/Log Pis Min -10.4023 +trainer/policy/mean Mean -0.0417177 +trainer/policy/mean Std 0.905751 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 3.01474 +trainer/policy/normal/std Std 0.444897 +trainer/policy/normal/std Max 4.63086 +trainer/policy/normal/std Min 0.797405 +trainer/policy/normal/log_std Mean 1.09011 +trainer/policy/normal/log_std Std 0.175188 +trainer/policy/normal/log_std Max 1.53274 +trainer/policy/normal/log_std Min -0.226392 +eval/num steps total 24000 +eval/num paths total 24 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.401536 +eval/Actions Std 0.859942 +eval/Actions Max 0.999981 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58331 +time/logging (s) 0.00420812 +time/sampling batch (s) 0.27983 +time/saving (s) 0.00425396 +time/training (s) 6.80603 +time/epoch (s) 9.67762 +time/total (s) 283.325 +Epoch -977 +---------------------------------- --------------- +2022-05-10 13:15:28.272661 PDT | [2] Epoch -976 finished +---------------------------------- --------------- +epoch -976 +replay_buffer/size 999033 +trainer/num train calls 25000 +trainer/Policy Loss -18.0479 +trainer/Log Pis Mean 23.7565 +trainer/Log Pis Std 12.6934 +trainer/Log Pis Max 61.1353 +trainer/Log Pis Min -7.12105 +trainer/policy/mean Mean -0.020625 +trainer/policy/mean Std 0.897931 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.96538 +trainer/policy/normal/std Std 0.465943 +trainer/policy/normal/std Max 5.67006 +trainer/policy/normal/std Min 0.823714 +trainer/policy/normal/log_std Mean 1.07212 +trainer/policy/normal/log_std Std 0.183124 +trainer/policy/normal/log_std Max 1.7352 +trainer/policy/normal/log_std Min -0.193932 +eval/num steps total 25000 +eval/num paths total 25 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.29804 +eval/Actions Std 0.860997 +eval/Actions Max 0.999994 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73288 +time/logging (s) 0.0037662 +time/sampling batch (s) 0.530228 +time/saving (s) 0.00359222 +time/training (s) 7.85333 +time/epoch (s) 11.1238 +time/total (s) 294.452 +Epoch -976 +---------------------------------- --------------- +2022-05-10 13:15:39.221248 PDT | [2] Epoch -975 finished +---------------------------------- --------------- +epoch -975 +replay_buffer/size 999033 +trainer/num train calls 26000 +trainer/Policy Loss -19.38 +trainer/Log Pis Mean 23.5063 +trainer/Log Pis Std 13.8147 +trainer/Log Pis Max 78.4098 +trainer/Log Pis Min -7.54176 +trainer/policy/mean Mean -0.034813 +trainer/policy/mean Std 0.898227 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 3.00786 +trainer/policy/normal/std Std 0.51652 +trainer/policy/normal/std Max 4.8586 +trainer/policy/normal/std Min 0.780799 +trainer/policy/normal/log_std Mean 1.08237 +trainer/policy/normal/log_std Std 0.209564 +trainer/policy/normal/log_std Max 1.58075 +trainer/policy/normal/log_std Min -0.247437 +eval/num steps total 26000 +eval/num paths total 26 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0601696 +eval/Actions Std 0.889548 +eval/Actions Max 0.999998 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50193 +time/logging (s) 0.00384764 +time/sampling batch (s) 0.530223 +time/saving (s) 0.00347853 +time/training (s) 7.89005 +time/epoch (s) 10.9295 +time/total (s) 305.385 +Epoch -975 +---------------------------------- --------------- +2022-05-10 13:15:49.517479 PDT | [2] Epoch -974 finished +---------------------------------- --------------- +epoch -974 +replay_buffer/size 999033 +trainer/num train calls 27000 +trainer/Policy Loss -17.7069 +trainer/Log Pis Mean 23.1642 +trainer/Log Pis Std 12.1875 +trainer/Log Pis Max 60.2407 +trainer/Log Pis Min -12.9205 +trainer/policy/mean Mean -0.04173 +trainer/policy/mean Std 0.901666 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.90428 +trainer/policy/normal/std Std 0.456378 +trainer/policy/normal/std Max 4.60834 +trainer/policy/normal/std Min 0.79165 +trainer/policy/normal/log_std Mean 1.05084 +trainer/policy/normal/log_std Std 0.187605 +trainer/policy/normal/log_std Max 1.52787 +trainer/policy/normal/log_std Min -0.233636 +eval/num steps total 27000 +eval/num paths total 27 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00776359 +eval/Actions Std 0.864132 +eval/Actions Max 0.999996 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56585 +time/logging (s) 0.00373381 +time/sampling batch (s) 0.282101 +time/saving (s) 0.00353181 +time/training (s) 7.42146 +time/epoch (s) 10.2767 +time/total (s) 315.664 +Epoch -974 +---------------------------------- --------------- +2022-05-10 13:15:59.398048 PDT | [2] Epoch -973 finished +---------------------------------- --------------- +epoch -973 +replay_buffer/size 999033 +trainer/num train calls 28000 +trainer/Policy Loss -18.8087 +trainer/Log Pis Mean 23.7073 +trainer/Log Pis Std 13.1501 +trainer/Log Pis Max 64.0713 +trainer/Log Pis Min -7.07071 +trainer/policy/mean Mean -0.0350117 +trainer/policy/mean Std 0.898069 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999985 +trainer/policy/normal/std Mean 2.94186 +trainer/policy/normal/std Std 0.498523 +trainer/policy/normal/std Max 4.9239 +trainer/policy/normal/std Min 0.792757 +trainer/policy/normal/log_std Mean 1.06106 +trainer/policy/normal/log_std Std 0.203647 +trainer/policy/normal/log_std Max 1.5941 +trainer/policy/normal/log_std Min -0.232238 +eval/num steps total 28000 +eval/num paths total 28 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0177906 +eval/Actions Std 0.921729 +eval/Actions Max 0.999988 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61318 +time/logging (s) 0.00370704 +time/sampling batch (s) 0.283914 +time/saving (s) 0.00348287 +time/training (s) 6.95682 +time/epoch (s) 9.86111 +time/total (s) 325.529 +Epoch -973 +---------------------------------- --------------- +2022-05-10 13:16:09.705326 PDT | [2] Epoch -972 finished +---------------------------------- --------------- +epoch -972 +replay_buffer/size 999033 +trainer/num train calls 29000 +trainer/Policy Loss -17.9619 +trainer/Log Pis Mean 23.5309 +trainer/Log Pis Std 13.0691 +trainer/Log Pis Max 61.6135 +trainer/Log Pis Min -5.31786 +trainer/policy/mean Mean -0.0229983 +trainer/policy/mean Std 0.895883 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.94621 +trainer/policy/normal/std Std 0.480266 +trainer/policy/normal/std Max 4.63152 +trainer/policy/normal/std Min 0.720242 +trainer/policy/normal/log_std Mean 1.06394 +trainer/policy/normal/log_std Std 0.195599 +trainer/policy/normal/log_std Max 1.53289 +trainer/policy/normal/log_std Min -0.328168 +eval/num steps total 29000 +eval/num paths total 29 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.460223 +eval/Actions Std 0.873263 +eval/Actions Max 0.999992 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66362 +time/logging (s) 0.00416986 +time/sampling batch (s) 0.534119 +time/saving (s) 0.00409514 +time/training (s) 7.08239 +time/epoch (s) 10.2884 +time/total (s) 335.82 +Epoch -972 +---------------------------------- --------------- +2022-05-10 13:16:19.096516 PDT | [2] Epoch -971 finished +---------------------------------- --------------- +epoch -971 +replay_buffer/size 999033 +trainer/num train calls 30000 +trainer/Policy Loss -20.0074 +trainer/Log Pis Mean 24.8794 +trainer/Log Pis Std 12.5783 +trainer/Log Pis Max 60.2191 +trainer/Log Pis Min -8.18637 +trainer/policy/mean Mean -0.045648 +trainer/policy/mean Std 0.901548 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.98329 +trainer/policy/normal/std Std 0.498557 +trainer/policy/normal/std Max 4.93372 +trainer/policy/normal/std Min 0.761342 +trainer/policy/normal/log_std Mean 1.07531 +trainer/policy/normal/log_std Std 0.20292 +trainer/policy/normal/log_std Max 1.59609 +trainer/policy/normal/log_std Min -0.272673 +eval/num steps total 30000 +eval/num paths total 30 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.156027 +eval/Actions Std 0.86039 +eval/Actions Max 0.999996 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.4289 +time/logging (s) 0.00373147 +time/sampling batch (s) 0.533389 +time/saving (s) 0.00363321 +time/training (s) 6.40117 +time/epoch (s) 9.37082 +time/total (s) 345.194 +Epoch -971 +---------------------------------- --------------- +2022-05-10 13:16:29.184483 PDT | [2] Epoch -970 finished +---------------------------------- --------------- +epoch -970 +replay_buffer/size 999033 +trainer/num train calls 31000 +trainer/Policy Loss -19.5419 +trainer/Log Pis Mean 24.2611 +trainer/Log Pis Std 12.9875 +trainer/Log Pis Max 72.4302 +trainer/Log Pis Min -7.74817 +trainer/policy/mean Mean -0.0469878 +trainer/policy/mean Std 0.901923 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.94526 +trainer/policy/normal/std Std 0.501232 +trainer/policy/normal/std Max 5.00959 +trainer/policy/normal/std Min 0.726164 +trainer/policy/normal/log_std Mean 1.06157 +trainer/policy/normal/log_std Std 0.208693 +trainer/policy/normal/log_std Max 1.61135 +trainer/policy/normal/log_std Min -0.31998 +eval/num steps total 31000 +eval/num paths total 31 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0239695 +eval/Actions Std 0.969038 +eval/Actions Max 0.999979 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59411 +time/logging (s) 0.00374474 +time/sampling batch (s) 0.537107 +time/saving (s) 0.0035791 +time/training (s) 6.9298 +time/epoch (s) 10.0683 +time/total (s) 355.265 +Epoch -970 +---------------------------------- --------------- +2022-05-10 13:16:38.758642 PDT | [2] Epoch -969 finished +---------------------------------- --------------- +epoch -969 +replay_buffer/size 999033 +trainer/num train calls 32000 +trainer/Policy Loss -19.6868 +trainer/Log Pis Mean 23.7666 +trainer/Log Pis Std 13.3564 +trainer/Log Pis Max 81.3842 +trainer/Log Pis Min -11.3122 +trainer/policy/mean Mean -0.058081 +trainer/policy/mean Std 0.897172 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.99082 +trainer/policy/normal/std Std 0.497072 +trainer/policy/normal/std Max 4.94883 +trainer/policy/normal/std Min 0.73091 +trainer/policy/normal/log_std Mean 1.07809 +trainer/policy/normal/log_std Std 0.201497 +trainer/policy/normal/log_std Max 1.59915 +trainer/policy/normal/log_std Min -0.313465 +eval/num steps total 32000 +eval/num paths total 32 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0165403 +eval/Actions Std 0.899309 +eval/Actions Max 0.999997 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.49834 +time/logging (s) 0.00375798 +time/sampling batch (s) 0.279475 +time/saving (s) 0.00342277 +time/training (s) 6.77004 +time/epoch (s) 9.55503 +time/total (s) 364.823 +Epoch -969 +---------------------------------- --------------- +2022-05-10 13:16:48.797127 PDT | [2] Epoch -968 finished +---------------------------------- --------------- +epoch -968 +replay_buffer/size 999033 +trainer/num train calls 33000 +trainer/Policy Loss -19.3041 +trainer/Log Pis Mean 24.2645 +trainer/Log Pis Std 13.0073 +trainer/Log Pis Max 64.5833 +trainer/Log Pis Min -7.01046 +trainer/policy/mean Mean -0.0417088 +trainer/policy/mean Std 0.896256 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.93383 +trainer/policy/normal/std Std 0.495708 +trainer/policy/normal/std Max 4.63623 +trainer/policy/normal/std Min 0.660415 +trainer/policy/normal/log_std Mean 1.05787 +trainer/policy/normal/log_std Std 0.208367 +trainer/policy/normal/log_std Max 1.5339 +trainer/policy/normal/log_std Min -0.414887 +eval/num steps total 33000 +eval/num paths total 33 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0781821 +eval/Actions Std 0.911054 +eval/Actions Max 0.999995 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70817 +time/logging (s) 0.0037495 +time/sampling batch (s) 0.779895 +time/saving (s) 0.00343948 +time/training (s) 6.52415 +time/epoch (s) 10.0194 +time/total (s) 374.846 +Epoch -968 +---------------------------------- --------------- +2022-05-10 13:16:58.475588 PDT | [2] Epoch -967 finished +---------------------------------- --------------- +epoch -967 +replay_buffer/size 999033 +trainer/num train calls 34000 +trainer/Policy Loss -19.2729 +trainer/Log Pis Mean 24.0374 +trainer/Log Pis Std 12.8801 +trainer/Log Pis Max 62.1041 +trainer/Log Pis Min -8.07259 +trainer/policy/mean Mean -0.0183641 +trainer/policy/mean Std 0.905227 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.95064 +trainer/policy/normal/std Std 0.487189 +trainer/policy/normal/std Max 4.98938 +trainer/policy/normal/std Min 0.765168 +trainer/policy/normal/log_std Mean 1.06533 +trainer/policy/normal/log_std Std 0.195164 +trainer/policy/normal/log_std Max 1.60731 +trainer/policy/normal/log_std Min -0.26766 +eval/num steps total 34000 +eval/num paths total 34 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.333397 +eval/Actions Std 0.852236 +eval/Actions Max 0.999993 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64986 +time/logging (s) 0.00369123 +time/sampling batch (s) 0.529981 +time/saving (s) 0.00342839 +time/training (s) 6.47234 +time/epoch (s) 9.6593 +time/total (s) 384.508 +Epoch -967 +---------------------------------- --------------- +2022-05-10 13:17:07.961633 PDT | [2] Epoch -966 finished +---------------------------------- --------------- +epoch -966 +replay_buffer/size 999033 +trainer/num train calls 35000 +trainer/Policy Loss -17.3259 +trainer/Log Pis Mean 23.9708 +trainer/Log Pis Std 12.3977 +trainer/Log Pis Max 62.5752 +trainer/Log Pis Min -6.49975 +trainer/policy/mean Mean -0.0309477 +trainer/policy/mean Std 0.903003 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.91625 +trainer/policy/normal/std Std 0.488406 +trainer/policy/normal/std Max 5.43905 +trainer/policy/normal/std Min 0.690293 +trainer/policy/normal/log_std Mean 1.05261 +trainer/policy/normal/log_std Std 0.202506 +trainer/policy/normal/log_std Max 1.6936 +trainer/policy/normal/log_std Min -0.370639 +eval/num steps total 35000 +eval/num paths total 35 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0785125 +eval/Actions Std 0.905119 +eval/Actions Max 0.999996 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60427 +time/logging (s) 0.00417192 +time/sampling batch (s) 0.279445 +time/saving (s) 0.00400053 +time/training (s) 6.57544 +time/epoch (s) 9.46733 +time/total (s) 393.978 +Epoch -966 +---------------------------------- --------------- +2022-05-10 13:17:17.692513 PDT | [2] Epoch -965 finished +---------------------------------- --------------- +epoch -965 +replay_buffer/size 999033 +trainer/num train calls 36000 +trainer/Policy Loss -19.7743 +trainer/Log Pis Mean 23.9848 +trainer/Log Pis Std 12.551 +trainer/Log Pis Max 67.3902 +trainer/Log Pis Min -7.88159 +trainer/policy/mean Mean -0.0316301 +trainer/policy/mean Std 0.900324 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999982 +trainer/policy/normal/std Mean 2.92372 +trainer/policy/normal/std Std 0.496015 +trainer/policy/normal/std Max 5.31761 +trainer/policy/normal/std Min 0.685982 +trainer/policy/normal/log_std Mean 1.05435 +trainer/policy/normal/log_std Std 0.208239 +trainer/policy/normal/log_std Max 1.67102 +trainer/policy/normal/log_std Min -0.376904 +eval/num steps total 36000 +eval/num paths total 36 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.200405 +eval/Actions Std 0.943266 +eval/Actions Max 0.999996 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54244 +time/logging (s) 0.00373766 +time/sampling batch (s) 0.280423 +time/saving (s) 0.00361008 +time/training (s) 6.88053 +time/epoch (s) 9.71074 +time/total (s) 403.692 +Epoch -965 +---------------------------------- --------------- +2022-05-10 13:17:27.858353 PDT | [2] Epoch -964 finished +---------------------------------- --------------- +epoch -964 +replay_buffer/size 999033 +trainer/num train calls 37000 +trainer/Policy Loss -19.1519 +trainer/Log Pis Mean 23.802 +trainer/Log Pis Std 13.3045 +trainer/Log Pis Max 73.1995 +trainer/Log Pis Min -6.27325 +trainer/policy/mean Mean -0.0451031 +trainer/policy/mean Std 0.895755 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.88178 +trainer/policy/normal/std Std 0.510043 +trainer/policy/normal/std Max 4.80192 +trainer/policy/normal/std Min 0.656643 +trainer/policy/normal/log_std Mean 1.03863 +trainer/policy/normal/log_std Std 0.21429 +trainer/policy/normal/log_std Max 1.56902 +trainer/policy/normal/log_std Min -0.420615 +eval/num steps total 37000 +eval/num paths total 37 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.116403 +eval/Actions Std 0.914968 +eval/Actions Max 0.999992 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51204 +time/logging (s) 0.00382106 +time/sampling batch (s) 0.529544 +time/saving (s) 0.00362901 +time/training (s) 7.09779 +time/epoch (s) 10.1468 +time/total (s) 413.842 +Epoch -964 +---------------------------------- --------------- +2022-05-10 13:17:37.862048 PDT | [2] Epoch -963 finished +---------------------------------- --------------- +epoch -963 +replay_buffer/size 999033 +trainer/num train calls 38000 +trainer/Policy Loss -19.0019 +trainer/Log Pis Mean 23.9721 +trainer/Log Pis Std 13.1871 +trainer/Log Pis Max 62.4032 +trainer/Log Pis Min -7.43321 +trainer/policy/mean Mean -0.0493862 +trainer/policy/mean Std 0.897795 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.94503 +trainer/policy/normal/std Std 0.556446 +trainer/policy/normal/std Max 6.44382 +trainer/policy/normal/std Min 0.658172 +trainer/policy/normal/log_std Mean 1.05723 +trainer/policy/normal/log_std Std 0.232376 +trainer/policy/normal/log_std Max 1.86312 +trainer/policy/normal/log_std Min -0.418289 +eval/num steps total 38000 +eval/num paths total 38 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.359774 +eval/Actions Std 0.852105 +eval/Actions Max 0.99999 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64654 +time/logging (s) 0.00378505 +time/sampling batch (s) 0.529265 +time/saving (s) 0.00342291 +time/training (s) 6.80162 +time/epoch (s) 9.98462 +time/total (s) 423.829 +Epoch -963 +---------------------------------- --------------- +2022-05-10 13:17:47.014871 PDT | [2] Epoch -962 finished +---------------------------------- --------------- +epoch -962 +replay_buffer/size 999033 +trainer/num train calls 39000 +trainer/Policy Loss -19.4655 +trainer/Log Pis Mean 24.769 +trainer/Log Pis Std 13.189 +trainer/Log Pis Max 64.6635 +trainer/Log Pis Min -8.94479 +trainer/policy/mean Mean -0.0368842 +trainer/policy/mean Std 0.902207 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.95352 +trainer/policy/normal/std Std 0.505669 +trainer/policy/normal/std Max 5.56446 +trainer/policy/normal/std Min 0.613551 +trainer/policy/normal/log_std Mean 1.06418 +trainer/policy/normal/log_std Std 0.210686 +trainer/policy/normal/log_std Max 1.7164 +trainer/policy/normal/log_std Min -0.488492 +eval/num steps total 39000 +eval/num paths total 39 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0947483 +eval/Actions Std 0.906938 +eval/Actions Max 0.999989 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51621 +time/logging (s) 0.00375415 +time/sampling batch (s) 0.279568 +time/saving (s) 0.00346856 +time/training (s) 6.33064 +time/epoch (s) 9.13364 +time/total (s) 432.966 +Epoch -962 +---------------------------------- --------------- +2022-05-10 13:17:57.755570 PDT | [2] Epoch -961 finished +---------------------------------- --------------- +epoch -961 +replay_buffer/size 999033 +trainer/num train calls 40000 +trainer/Policy Loss -18.9083 +trainer/Log Pis Mean 25.1135 +trainer/Log Pis Std 12.5971 +trainer/Log Pis Max 56.9447 +trainer/Log Pis Min -6.2784 +trainer/policy/mean Mean -0.0515112 +trainer/policy/mean Std 0.904536 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.97109 +trainer/policy/normal/std Std 0.516976 +trainer/policy/normal/std Max 5.14852 +trainer/policy/normal/std Min 0.641908 +trainer/policy/normal/log_std Mean 1.06902 +trainer/policy/normal/log_std Std 0.217761 +trainer/policy/normal/log_std Max 1.63871 +trainer/policy/normal/log_std Min -0.44331 +eval/num steps total 40000 +eval/num paths total 40 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.473307 +eval/Actions Std 0.872986 +eval/Actions Max 0.99999 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.94603 +time/logging (s) 0.00382163 +time/sampling batch (s) 0.530689 +time/saving (s) 0.00346573 +time/training (s) 7.23759 +time/epoch (s) 10.7216 +time/total (s) 443.69 +Epoch -961 +---------------------------------- --------------- +2022-05-10 13:18:08.209512 PDT | [2] Epoch -960 finished +---------------------------------- --------------- +epoch -960 +replay_buffer/size 999033 +trainer/num train calls 41000 +trainer/Policy Loss -19.8911 +trainer/Log Pis Mean 24.0042 +trainer/Log Pis Std 12.1999 +trainer/Log Pis Max 71.6479 +trainer/Log Pis Min -2.28659 +trainer/policy/mean Mean -0.0467388 +trainer/policy/mean Std 0.90531 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.94993 +trainer/policy/normal/std Std 0.494946 +trainer/policy/normal/std Max 5.71479 +trainer/policy/normal/std Min 0.626084 +trainer/policy/normal/log_std Mean 1.06348 +trainer/policy/normal/log_std Std 0.20798 +trainer/policy/normal/log_std Max 1.74306 +trainer/policy/normal/log_std Min -0.46827 +eval/num steps total 41000 +eval/num paths total 41 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0808408 +eval/Actions Std 0.898312 +eval/Actions Max 1 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.48226 +time/logging (s) 0.00415242 +time/sampling batch (s) 0.784787 +time/saving (s) 0.00396271 +time/training (s) 7.15953 +time/epoch (s) 10.4347 +time/total (s) 454.128 +Epoch -960 +---------------------------------- --------------- +2022-05-10 13:18:18.611109 PDT | [2] Epoch -959 finished +---------------------------------- --------------- +epoch -959 +replay_buffer/size 999033 +trainer/num train calls 42000 +trainer/Policy Loss -19.143 +trainer/Log Pis Mean 24.8227 +trainer/Log Pis Std 13.125 +trainer/Log Pis Max 62.7359 +trainer/Log Pis Min -9.04561 +trainer/policy/mean Mean -0.0223482 +trainer/policy/mean Std 0.899859 +trainer/policy/mean Max 0.999981 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.98907 +trainer/policy/normal/std Std 0.511208 +trainer/policy/normal/std Max 5.10834 +trainer/policy/normal/std Min 0.679619 +trainer/policy/normal/log_std Mean 1.0768 +trainer/policy/normal/log_std Std 0.204509 +trainer/policy/normal/log_std Max 1.63087 +trainer/policy/normal/log_std Min -0.386223 +eval/num steps total 42000 +eval/num paths total 42 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.106164 +eval/Actions Std 0.914878 +eval/Actions Max 0.999996 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65862 +time/logging (s) 0.00378932 +time/sampling batch (s) 0.286475 +time/saving (s) 0.00376536 +time/training (s) 7.42861 +time/epoch (s) 10.3813 +time/total (s) 464.512 +Epoch -959 +---------------------------------- --------------- +2022-05-10 13:18:28.446389 PDT | [2] Epoch -958 finished +---------------------------------- --------------- +epoch -958 +replay_buffer/size 999033 +trainer/num train calls 43000 +trainer/Policy Loss -19.4345 +trainer/Log Pis Mean 23.7411 +trainer/Log Pis Std 12.9817 +trainer/Log Pis Max 58.8391 +trainer/Log Pis Min -4.66757 +trainer/policy/mean Mean -0.0487739 +trainer/policy/mean Std 0.901343 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.85444 +trainer/policy/normal/std Std 0.503929 +trainer/policy/normal/std Max 4.87094 +trainer/policy/normal/std Min 0.702203 +trainer/policy/normal/log_std Mean 1.02895 +trainer/policy/normal/log_std Std 0.215414 +trainer/policy/normal/log_std Max 1.58329 +trainer/policy/normal/log_std Min -0.353533 +eval/num steps total 43000 +eval/num paths total 43 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0971116 +eval/Actions Std 0.911077 +eval/Actions Max 0.999997 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67901 +time/logging (s) 0.00374999 +time/sampling batch (s) 0.78224 +time/saving (s) 0.00377206 +time/training (s) 6.34704 +time/epoch (s) 9.81581 +time/total (s) 474.331 +Epoch -958 +---------------------------------- --------------- +2022-05-10 13:18:38.965244 PDT | [2] Epoch -957 finished +---------------------------------- --------------- +epoch -957 +replay_buffer/size 999033 +trainer/num train calls 44000 +trainer/Policy Loss -17.8782 +trainer/Log Pis Mean 24.4815 +trainer/Log Pis Std 12.2798 +trainer/Log Pis Max 75.0308 +trainer/Log Pis Min -4.21468 +trainer/policy/mean Mean -0.0463552 +trainer/policy/mean Std 0.899404 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.94239 +trainer/policy/normal/std Std 0.521886 +trainer/policy/normal/std Max 5.24136 +trainer/policy/normal/std Min 0.626237 +trainer/policy/normal/log_std Mean 1.0589 +trainer/policy/normal/log_std Std 0.218834 +trainer/policy/normal/log_std Max 1.65658 +trainer/policy/normal/log_std Min -0.468026 +eval/num steps total 44000 +eval/num paths total 44 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0685434 +eval/Actions Std 0.886657 +eval/Actions Max 0.999992 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66489 +time/logging (s) 0.00370647 +time/sampling batch (s) 0.286453 +time/saving (s) 0.00349072 +time/training (s) 7.54066 +time/epoch (s) 10.4992 +time/total (s) 484.833 +Epoch -957 +---------------------------------- --------------- +2022-05-10 13:18:48.321939 PDT | [2] Epoch -956 finished +---------------------------------- --------------- +epoch -956 +replay_buffer/size 999033 +trainer/num train calls 45000 +trainer/Policy Loss -19.5633 +trainer/Log Pis Mean 24.2856 +trainer/Log Pis Std 13.6244 +trainer/Log Pis Max 63.4428 +trainer/Log Pis Min -13.9428 +trainer/policy/mean Mean -0.0236154 +trainer/policy/mean Std 0.902908 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.9436 +trainer/policy/normal/std Std 0.536101 +trainer/policy/normal/std Max 5.28206 +trainer/policy/normal/std Min 0.635958 +trainer/policy/normal/log_std Mean 1.05856 +trainer/policy/normal/log_std Std 0.221892 +trainer/policy/normal/log_std Max 1.66432 +trainer/policy/normal/log_std Min -0.452623 +eval/num steps total 45000 +eval/num paths total 45 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.036447 +eval/Actions Std 0.896038 +eval/Actions Max 0.999996 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.82237 +time/logging (s) 0.00368832 +time/sampling batch (s) 0.282942 +time/saving (s) 0.00350106 +time/training (s) 6.22426 +time/epoch (s) 9.33676 +time/total (s) 494.173 +Epoch -956 +---------------------------------- --------------- +2022-05-10 13:18:59.088437 PDT | [2] Epoch -955 finished +---------------------------------- --------------- +epoch -955 +replay_buffer/size 999033 +trainer/num train calls 46000 +trainer/Policy Loss -18.7315 +trainer/Log Pis Mean 24.6649 +trainer/Log Pis Std 13.5262 +trainer/Log Pis Max 79.6698 +trainer/Log Pis Min -2.984 +trainer/policy/mean Mean -0.0196951 +trainer/policy/mean Std 0.903721 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.9635 +trainer/policy/normal/std Std 0.527241 +trainer/policy/normal/std Max 5.13245 +trainer/policy/normal/std Min 0.62449 +trainer/policy/normal/log_std Mean 1.06598 +trainer/policy/normal/log_std Std 0.219434 +trainer/policy/normal/log_std Max 1.63558 +trainer/policy/normal/log_std Min -0.47082 +eval/num steps total 46000 +eval/num paths total 46 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.125329 +eval/Actions Std 0.917795 +eval/Actions Max 0.999995 +eval/Actions Min -0.999985 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.88374 +time/logging (s) 0.00407763 +time/sampling batch (s) 0.284139 +time/saving (s) 0.0039467 +time/training (s) 7.57117 +time/epoch (s) 10.7471 +time/total (s) 504.923 +Epoch -955 +---------------------------------- --------------- +2022-05-10 13:19:08.626195 PDT | [2] Epoch -954 finished +---------------------------------- --------------- +epoch -954 +replay_buffer/size 999033 +trainer/num train calls 47000 +trainer/Policy Loss -19.0682 +trainer/Log Pis Mean 24.8364 +trainer/Log Pis Std 13.3977 +trainer/Log Pis Max 77.0585 +trainer/Log Pis Min -10.0834 +trainer/policy/mean Mean -0.0423233 +trainer/policy/mean Std 0.899767 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.96826 +trainer/policy/normal/std Std 0.524086 +trainer/policy/normal/std Max 5.30197 +trainer/policy/normal/std Min 0.493572 +trainer/policy/normal/log_std Mean 1.06737 +trainer/policy/normal/log_std Std 0.221802 +trainer/policy/normal/log_std Max 1.66808 +trainer/policy/normal/log_std Min -0.706087 +eval/num steps total 47000 +eval/num paths total 47 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.123969 +eval/Actions Std 0.884764 +eval/Actions Max 0.999988 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59118 +time/logging (s) 0.00384283 +time/sampling batch (s) 0.282687 +time/saving (s) 0.00376161 +time/training (s) 6.63605 +time/epoch (s) 9.51752 +time/total (s) 514.444 +Epoch -954 +---------------------------------- --------------- +2022-05-10 13:19:19.010223 PDT | [2] Epoch -953 finished +---------------------------------- --------------- +epoch -953 +replay_buffer/size 999033 +trainer/num train calls 48000 +trainer/Policy Loss -18.2784 +trainer/Log Pis Mean 24.3989 +trainer/Log Pis Std 13.3986 +trainer/Log Pis Max 74.1417 +trainer/Log Pis Min -9.23449 +trainer/policy/mean Mean -0.0160884 +trainer/policy/mean Std 0.89756 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.9877 +trainer/policy/normal/std Std 0.542212 +trainer/policy/normal/std Max 5.7317 +trainer/policy/normal/std Min 0.557836 +trainer/policy/normal/log_std Mean 1.07366 +trainer/policy/normal/log_std Std 0.220886 +trainer/policy/normal/log_std Max 1.74601 +trainer/policy/normal/log_std Min -0.583691 +eval/num steps total 48000 +eval/num paths total 48 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0106769 +eval/Actions Std 0.895026 +eval/Actions Max 0.999999 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.79281 +time/logging (s) 0.00382082 +time/sampling batch (s) 0.28281 +time/saving (s) 0.00345433 +time/training (s) 7.28128 +time/epoch (s) 10.3642 +time/total (s) 524.811 +Epoch -953 +---------------------------------- --------------- +2022-05-10 13:19:29.679584 PDT | [2] Epoch -952 finished +---------------------------------- --------------- +epoch -952 +replay_buffer/size 999033 +trainer/num train calls 49000 +trainer/Policy Loss -19.5823 +trainer/Log Pis Mean 24.4146 +trainer/Log Pis Std 13.3583 +trainer/Log Pis Max 68.7555 +trainer/Log Pis Min -8.39576 +trainer/policy/mean Mean -0.0319847 +trainer/policy/mean Std 0.901306 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.9817 +trainer/policy/normal/std Std 0.522365 +trainer/policy/normal/std Max 5.08494 +trainer/policy/normal/std Min 0.653158 +trainer/policy/normal/log_std Mean 1.07265 +trainer/policy/normal/log_std Std 0.216246 +trainer/policy/normal/log_std Max 1.62628 +trainer/policy/normal/log_std Min -0.425936 +eval/num steps total 49000 +eval/num paths total 49 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0193764 +eval/Actions Std 0.897969 +eval/Actions Max 0.999995 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.74481 +time/logging (s) 0.00375105 +time/sampling batch (s) 0.283727 +time/saving (s) 0.00349556 +time/training (s) 7.61384 +time/epoch (s) 10.6496 +time/total (s) 535.464 +Epoch -952 +---------------------------------- --------------- +2022-05-10 13:19:39.572579 PDT | [2] Epoch -951 finished +---------------------------------- --------------- +epoch -951 +replay_buffer/size 999033 +trainer/num train calls 50000 +trainer/Policy Loss -19.423 +trainer/Log Pis Mean 24.6016 +trainer/Log Pis Std 12.8289 +trainer/Log Pis Max 62.0605 +trainer/Log Pis Min -11.409 +trainer/policy/mean Mean -0.0539599 +trainer/policy/mean Std 0.903161 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.90554 +trainer/policy/normal/std Std 0.529263 +trainer/policy/normal/std Max 4.86116 +trainer/policy/normal/std Min 0.575875 +trainer/policy/normal/log_std Mean 1.04491 +trainer/policy/normal/log_std Std 0.226926 +trainer/policy/normal/log_std Max 1.58128 +trainer/policy/normal/log_std Min -0.551864 +eval/num steps total 50000 +eval/num paths total 50 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.146212 +eval/Actions Std 0.889729 +eval/Actions Max 0.999978 +eval/Actions Min -0.999983 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.7396 +time/logging (s) 0.00384817 +time/sampling batch (s) 0.28258 +time/saving (s) 0.00340638 +time/training (s) 6.84411 +time/epoch (s) 9.87354 +time/total (s) 545.34 +Epoch -951 +---------------------------------- --------------- +2022-05-10 13:19:48.932592 PDT | [2] Epoch -950 finished +---------------------------------- --------------- +epoch -950 +replay_buffer/size 999033 +trainer/num train calls 51000 +trainer/Policy Loss -19.5103 +trainer/Log Pis Mean 26.2974 +trainer/Log Pis Std 13.2959 +trainer/Log Pis Max 65.9055 +trainer/Log Pis Min -6.52636 +trainer/policy/mean Mean -0.0386087 +trainer/policy/mean Std 0.901074 +trainer/policy/mean Max 1 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.98825 +trainer/policy/normal/std Std 0.569882 +trainer/policy/normal/std Max 5.4978 +trainer/policy/normal/std Min 0.630567 +trainer/policy/normal/log_std Mean 1.07127 +trainer/policy/normal/log_std Std 0.234776 +trainer/policy/normal/log_std Max 1.70435 +trainer/policy/normal/log_std Min -0.461136 +eval/num steps total 51000 +eval/num paths total 51 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.121944 +eval/Actions Std 0.88145 +eval/Actions Max 1 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54227 +time/logging (s) 0.00378193 +time/sampling batch (s) 0.277173 +time/saving (s) 0.0034355 +time/training (s) 6.51422 +time/epoch (s) 9.34088 +time/total (s) 554.684 +Epoch -950 +---------------------------------- --------------- +2022-05-10 13:20:00.031116 PDT | [2] Epoch -949 finished +---------------------------------- --------------- +epoch -949 +replay_buffer/size 999033 +trainer/num train calls 52000 +trainer/Policy Loss -19.0544 +trainer/Log Pis Mean 24.3403 +trainer/Log Pis Std 13.2632 +trainer/Log Pis Max 61.9729 +trainer/Log Pis Min -10.4806 +trainer/policy/mean Mean -0.0370887 +trainer/policy/mean Std 0.89985 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.95114 +trainer/policy/normal/std Std 0.505702 +trainer/policy/normal/std Max 5.53439 +trainer/policy/normal/std Min 0.615705 +trainer/policy/normal/log_std Mean 1.06317 +trainer/policy/normal/log_std Std 0.212202 +trainer/policy/normal/log_std Max 1.71098 +trainer/policy/normal/log_std Min -0.484988 +eval/num steps total 52000 +eval/num paths total 52 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0245944 +eval/Actions Std 0.899103 +eval/Actions Max 0.999994 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.75599 +time/logging (s) 0.00413855 +time/sampling batch (s) 0.784195 +time/saving (s) 0.00408731 +time/training (s) 7.53066 +time/epoch (s) 11.0791 +time/total (s) 565.766 +Epoch -949 +---------------------------------- --------------- +2022-05-10 13:20:09.685366 PDT | [2] Epoch -948 finished +---------------------------------- --------------- +epoch -948 +replay_buffer/size 999033 +trainer/num train calls 53000 +trainer/Policy Loss -19.981 +trainer/Log Pis Mean 24.0635 +trainer/Log Pis Std 12.6772 +trainer/Log Pis Max 63.4592 +trainer/Log Pis Min -8.55674 +trainer/policy/mean Mean -0.0441852 +trainer/policy/mean Std 0.905595 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.91699 +trainer/policy/normal/std Std 0.521623 +trainer/policy/normal/std Max 4.70574 +trainer/policy/normal/std Min 0.608485 +trainer/policy/normal/log_std Mean 1.04994 +trainer/policy/normal/log_std Std 0.219888 +trainer/policy/normal/log_std Max 1.54878 +trainer/policy/normal/log_std Min -0.496783 +eval/num steps total 53000 +eval/num paths total 53 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0671631 +eval/Actions Std 0.88595 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5892 +time/logging (s) 0.00373196 +time/sampling batch (s) 0.280142 +time/saving (s) 0.00353492 +time/training (s) 6.75766 +time/epoch (s) 9.63427 +time/total (s) 575.404 +Epoch -948 +---------------------------------- --------------- +2022-05-10 13:20:20.094139 PDT | [2] Epoch -947 finished +---------------------------------- --------------- +epoch -947 +replay_buffer/size 999033 +trainer/num train calls 54000 +trainer/Policy Loss -18.8879 +trainer/Log Pis Mean 24.2246 +trainer/Log Pis Std 12.4647 +trainer/Log Pis Max 66.9931 +trainer/Log Pis Min -3.00941 +trainer/policy/mean Mean -0.0105167 +trainer/policy/mean Std 0.900857 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.96627 +trainer/policy/normal/std Std 0.526715 +trainer/policy/normal/std Max 5.28544 +trainer/policy/normal/std Min 0.637243 +trainer/policy/normal/log_std Mean 1.06736 +trainer/policy/normal/log_std Std 0.215643 +trainer/policy/normal/log_std Max 1.66496 +trainer/policy/normal/log_std Min -0.450605 +eval/num steps total 54000 +eval/num paths total 54 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.190993 +eval/Actions Std 0.930908 +eval/Actions Max 0.999999 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.47434 +time/logging (s) 0.00371367 +time/sampling batch (s) 0.782961 +time/saving (s) 0.00358613 +time/training (s) 7.12494 +time/epoch (s) 10.3895 +time/total (s) 585.796 +Epoch -947 +---------------------------------- --------------- +2022-05-10 13:20:30.140507 PDT | [2] Epoch -946 finished +---------------------------------- --------------- +epoch -946 +replay_buffer/size 999033 +trainer/num train calls 55000 +trainer/Policy Loss -19.8115 +trainer/Log Pis Mean 23.538 +trainer/Log Pis Std 12.7723 +trainer/Log Pis Max 68.5773 +trainer/Log Pis Min -9.75791 +trainer/policy/mean Mean -0.0525182 +trainer/policy/mean Std 0.906355 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.88995 +trainer/policy/normal/std Std 0.538565 +trainer/policy/normal/std Max 4.9944 +trainer/policy/normal/std Min 0.589248 +trainer/policy/normal/log_std Mean 1.03796 +trainer/policy/normal/log_std Std 0.236086 +trainer/policy/normal/log_std Max 1.60832 +trainer/policy/normal/log_std Min -0.528908 +eval/num steps total 55000 +eval/num paths total 55 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.126531 +eval/Actions Std 0.905651 +eval/Actions Max 0.999997 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58629 +time/logging (s) 0.0037661 +time/sampling batch (s) 0.287819 +time/saving (s) 0.00340579 +time/training (s) 7.1454 +time/epoch (s) 10.0267 +time/total (s) 595.826 +Epoch -946 +---------------------------------- --------------- +2022-05-10 13:20:39.318473 PDT | [2] Epoch -945 finished +---------------------------------- --------------- +epoch -945 +replay_buffer/size 999033 +trainer/num train calls 56000 +trainer/Policy Loss -20.3967 +trainer/Log Pis Mean 24.5 +trainer/Log Pis Std 13.0277 +trainer/Log Pis Max 79.5282 +trainer/Log Pis Min -8.73444 +trainer/policy/mean Mean -0.0407262 +trainer/policy/mean Std 0.905098 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.90185 +trainer/policy/normal/std Std 0.544614 +trainer/policy/normal/std Max 5.36433 +trainer/policy/normal/std Min 0.562735 +trainer/policy/normal/log_std Mean 1.04237 +trainer/policy/normal/log_std Std 0.233655 +trainer/policy/normal/log_std Max 1.67977 +trainer/policy/normal/log_std Min -0.574946 +eval/num steps total 56000 +eval/num paths total 56 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.105539 +eval/Actions Std 0.913922 +eval/Actions Max 0.999995 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.72332 +time/logging (s) 0.00368482 +time/sampling batch (s) 0.282653 +time/saving (s) 0.00349326 +time/training (s) 6.14522 +time/epoch (s) 9.15836 +time/total (s) 604.987 +Epoch -945 +---------------------------------- --------------- +2022-05-10 13:20:49.252784 PDT | [2] Epoch -944 finished +---------------------------------- --------------- +epoch -944 +replay_buffer/size 999033 +trainer/num train calls 57000 +trainer/Policy Loss -19.4773 +trainer/Log Pis Mean 23.9878 +trainer/Log Pis Std 13.037 +trainer/Log Pis Max 62.3032 +trainer/Log Pis Min -8.79458 +trainer/policy/mean Mean -0.0310905 +trainer/policy/mean Std 0.903751 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.91703 +trainer/policy/normal/std Std 0.522917 +trainer/policy/normal/std Max 5.35366 +trainer/policy/normal/std Min 0.562983 +trainer/policy/normal/log_std Mean 1.04933 +trainer/policy/normal/log_std Std 0.225603 +trainer/policy/normal/log_std Max 1.67778 +trainer/policy/normal/log_std Min -0.574505 +eval/num steps total 57000 +eval/num paths total 57 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0248861 +eval/Actions Std 0.944382 +eval/Actions Max 0.999994 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63218 +time/logging (s) 0.00418579 +time/sampling batch (s) 0.285055 +time/saving (s) 0.00397741 +time/training (s) 6.98979 +time/epoch (s) 9.91519 +time/total (s) 614.905 +Epoch -944 +---------------------------------- --------------- +2022-05-10 13:20:59.031239 PDT | [2] Epoch -943 finished +---------------------------------- --------------- +epoch -943 +replay_buffer/size 999033 +trainer/num train calls 58000 +trainer/Policy Loss -19.8322 +trainer/Log Pis Mean 24.7505 +trainer/Log Pis Std 13.6429 +trainer/Log Pis Max 80.9956 +trainer/Log Pis Min -8.24176 +trainer/policy/mean Mean -0.0619158 +trainer/policy/mean Std 0.903417 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.9329 +trainer/policy/normal/std Std 0.526322 +trainer/policy/normal/std Max 4.97902 +trainer/policy/normal/std Min 0.472625 +trainer/policy/normal/log_std Mean 1.05439 +trainer/policy/normal/log_std Std 0.228633 +trainer/policy/normal/log_std Max 1.60523 +trainer/policy/normal/log_std Min -0.749452 +eval/num steps total 58000 +eval/num paths total 58 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0214531 +eval/Actions Std 0.899953 +eval/Actions Max 0.999999 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6311 +time/logging (s) 0.00379475 +time/sampling batch (s) 0.284673 +time/saving (s) 0.00388972 +time/training (s) 6.83474 +time/epoch (s) 9.75821 +time/total (s) 624.666 +Epoch -943 +---------------------------------- --------------- +2022-05-10 13:21:09.610832 PDT | [2] Epoch -942 finished +---------------------------------- --------------- +epoch -942 +replay_buffer/size 999033 +trainer/num train calls 59000 +trainer/Policy Loss -17.7721 +trainer/Log Pis Mean 24.2775 +trainer/Log Pis Std 13.7245 +trainer/Log Pis Max 74.9307 +trainer/Log Pis Min -6.13582 +trainer/policy/mean Mean -0.03902 +trainer/policy/mean Std 0.897335 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.96672 +trainer/policy/normal/std Std 0.551711 +trainer/policy/normal/std Max 5.37704 +trainer/policy/normal/std Min 0.623638 +trainer/policy/normal/log_std Mean 1.06457 +trainer/policy/normal/log_std Std 0.234094 +trainer/policy/normal/log_std Max 1.68214 +trainer/policy/normal/log_std Min -0.472184 +eval/num steps total 59000 +eval/num paths total 59 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.296859 +eval/Actions Std 0.699061 +eval/Actions Max 0.999994 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.41474 +time/logging (s) 0.00403826 +time/sampling batch (s) 0.530923 +time/saving (s) 0.00373737 +time/training (s) 7.60665 +time/epoch (s) 10.5601 +time/total (s) 635.23 +Epoch -942 +---------------------------------- --------------- +2022-05-10 13:21:19.346584 PDT | [2] Epoch -941 finished +---------------------------------- --------------- +epoch -941 +replay_buffer/size 999033 +trainer/num train calls 60000 +trainer/Policy Loss -18.8801 +trainer/Log Pis Mean 23.3672 +trainer/Log Pis Std 12.9604 +trainer/Log Pis Max 70.5533 +trainer/Log Pis Min -13.3535 +trainer/policy/mean Mean -0.0439067 +trainer/policy/mean Std 0.904344 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.84124 +trainer/policy/normal/std Std 0.515421 +trainer/policy/normal/std Max 5.16689 +trainer/policy/normal/std Min 0.586172 +trainer/policy/normal/log_std Mean 1.02275 +trainer/policy/normal/log_std Std 0.225573 +trainer/policy/normal/log_std Max 1.64227 +trainer/policy/normal/log_std Min -0.534142 +eval/num steps total 60000 +eval/num paths total 60 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.162069 +eval/Actions Std 0.81593 +eval/Actions Max 0.999997 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.74216 +time/logging (s) 0.00373184 +time/sampling batch (s) 0.280851 +time/saving (s) 0.00344801 +time/training (s) 6.68519 +time/epoch (s) 9.71538 +time/total (s) 644.948 +Epoch -941 +---------------------------------- --------------- +2022-05-10 13:21:30.110353 PDT | [2] Epoch -940 finished +---------------------------------- --------------- +epoch -940 +replay_buffer/size 999033 +trainer/num train calls 61000 +trainer/Policy Loss -19.5183 +trainer/Log Pis Mean 24.4383 +trainer/Log Pis Std 12.8659 +trainer/Log Pis Max 71.4746 +trainer/Log Pis Min -3.73939 +trainer/policy/mean Mean -0.0366987 +trainer/policy/mean Std 0.9019 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.92756 +trainer/policy/normal/std Std 0.522368 +trainer/policy/normal/std Max 4.95453 +trainer/policy/normal/std Min 0.556555 +trainer/policy/normal/log_std Mean 1.05331 +trainer/policy/normal/log_std Std 0.222993 +trainer/policy/normal/log_std Max 1.6003 +trainer/policy/normal/log_std Min -0.585989 +eval/num steps total 61000 +eval/num paths total 61 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00359294 +eval/Actions Std 0.884024 +eval/Actions Max 0.999994 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.8066 +time/logging (s) 0.00378047 +time/sampling batch (s) 0.280662 +time/saving (s) 0.00348608 +time/training (s) 7.6499 +time/epoch (s) 10.7444 +time/total (s) 655.695 +Epoch -940 +---------------------------------- --------------- +2022-05-10 13:21:40.857731 PDT | [2] Epoch -939 finished +---------------------------------- --------------- +epoch -939 +replay_buffer/size 999033 +trainer/num train calls 62000 +trainer/Policy Loss -18.4162 +trainer/Log Pis Mean 24.4018 +trainer/Log Pis Std 12.8691 +trainer/Log Pis Max 62.4508 +trainer/Log Pis Min -2.76632 +trainer/policy/mean Mean -0.0366441 +trainer/policy/mean Std 0.904304 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.91009 +trainer/policy/normal/std Std 0.516343 +trainer/policy/normal/std Max 5.7779 +trainer/policy/normal/std Min 0.551806 +trainer/policy/normal/log_std Mean 1.04806 +trainer/policy/normal/log_std Std 0.21748 +trainer/policy/normal/log_std Max 1.75404 +trainer/policy/normal/log_std Min -0.594559 +eval/num steps total 62000 +eval/num paths total 62 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.118892 +eval/Actions Std 0.918878 +eval/Actions Max 0.999983 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6836 +time/logging (s) 0.00372988 +time/sampling batch (s) 0.281219 +time/saving (s) 0.00345157 +time/training (s) 7.75575 +time/epoch (s) 10.7277 +time/total (s) 666.426 +Epoch -939 +---------------------------------- --------------- +2022-05-10 13:21:50.872087 PDT | [2] Epoch -938 finished +---------------------------------- --------------- +epoch -938 +replay_buffer/size 999033 +trainer/num train calls 63000 +trainer/Policy Loss -18.9496 +trainer/Log Pis Mean 24.4073 +trainer/Log Pis Std 13.4266 +trainer/Log Pis Max 63.6007 +trainer/Log Pis Min -13.8368 +trainer/policy/mean Mean -0.0487989 +trainer/policy/mean Std 0.902598 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.92691 +trainer/policy/normal/std Std 0.54223 +trainer/policy/normal/std Max 5.31469 +trainer/policy/normal/std Min 0.558102 +trainer/policy/normal/log_std Mean 1.0511 +trainer/policy/normal/log_std Std 0.234415 +trainer/policy/normal/log_std Max 1.67047 +trainer/policy/normal/log_std Min -0.583214 +eval/num steps total 63000 +eval/num paths total 63 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.578206 +eval/Actions Std 0.737868 +eval/Actions Max 0.999989 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67584 +time/logging (s) 0.00399882 +time/sampling batch (s) 0.532798 +time/saving (s) 0.00415195 +time/training (s) 6.77833 +time/epoch (s) 9.99511 +time/total (s) 676.424 +Epoch -938 +---------------------------------- --------------- +2022-05-10 13:22:02.089401 PDT | [2] Epoch -937 finished +---------------------------------- --------------- +epoch -937 +replay_buffer/size 999033 +trainer/num train calls 64000 +trainer/Policy Loss -20.3954 +trainer/Log Pis Mean 24.6941 +trainer/Log Pis Std 13.3847 +trainer/Log Pis Max 67.6359 +trainer/Log Pis Min -8.54927 +trainer/policy/mean Mean -0.0292608 +trainer/policy/mean Std 0.907626 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.92942 +trainer/policy/normal/std Std 0.515822 +trainer/policy/normal/std Max 5.03163 +trainer/policy/normal/std Min 0.658699 +trainer/policy/normal/log_std Mean 1.05519 +trainer/policy/normal/log_std Std 0.213597 +trainer/policy/normal/log_std Max 1.61574 +trainer/policy/normal/log_std Min -0.417488 +eval/num steps total 64000 +eval/num paths total 64 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.11474 +eval/Actions Std 0.901064 +eval/Actions Max 0.999988 +eval/Actions Min -0.999985 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.89187 +time/logging (s) 0.00373517 +time/sampling batch (s) 0.28576 +time/saving (s) 0.00358933 +time/training (s) 8.01208 +time/epoch (s) 11.197 +time/total (s) 687.624 +Epoch -937 +---------------------------------- --------------- +2022-05-10 13:22:11.812524 PDT | [2] Epoch -936 finished +---------------------------------- --------------- +epoch -936 +replay_buffer/size 999033 +trainer/num train calls 65000 +trainer/Policy Loss -19.5901 +trainer/Log Pis Mean 24.3367 +trainer/Log Pis Std 13.4205 +trainer/Log Pis Max 72.615 +trainer/Log Pis Min -8.18224 +trainer/policy/mean Mean -0.0347766 +trainer/policy/mean Std 0.905102 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.84607 +trainer/policy/normal/std Std 0.526874 +trainer/policy/normal/std Max 5.02308 +trainer/policy/normal/std Min 0.514068 +trainer/policy/normal/log_std Mean 1.02362 +trainer/policy/normal/log_std Std 0.229936 +trainer/policy/normal/log_std Max 1.61404 +trainer/policy/normal/log_std Min -0.6654 +eval/num steps total 65000 +eval/num paths total 65 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.146106 +eval/Actions Std 0.932161 +eval/Actions Max 0.999998 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.78825 +time/logging (s) 0.00378862 +time/sampling batch (s) 0.282935 +time/saving (s) 0.00345433 +time/training (s) 6.62496 +time/epoch (s) 9.7034 +time/total (s) 697.331 +Epoch -936 +---------------------------------- --------------- +2022-05-10 13:22:22.092483 PDT | [2] Epoch -935 finished +---------------------------------- --------------- +epoch -935 +replay_buffer/size 999033 +trainer/num train calls 66000 +trainer/Policy Loss -18.3083 +trainer/Log Pis Mean 23.5087 +trainer/Log Pis Std 13.6325 +trainer/Log Pis Max 75.0423 +trainer/Log Pis Min -5.84244 +trainer/policy/mean Mean -0.0303361 +trainer/policy/mean Std 0.895855 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999982 +trainer/policy/normal/std Mean 2.87225 +trainer/policy/normal/std Std 0.549494 +trainer/policy/normal/std Max 5.24578 +trainer/policy/normal/std Min 0.634427 +trainer/policy/normal/log_std Mean 1.03097 +trainer/policy/normal/log_std Std 0.239566 +trainer/policy/normal/log_std Max 1.65742 +trainer/policy/normal/log_std Min -0.455033 +eval/num steps total 66000 +eval/num paths total 66 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0318515 +eval/Actions Std 0.89238 +eval/Actions Max 0.999996 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.38431 +time/logging (s) 0.00382202 +time/sampling batch (s) 0.533894 +time/saving (s) 0.00349122 +time/training (s) 7.33476 +time/epoch (s) 10.2603 +time/total (s) 707.594 +Epoch -935 +---------------------------------- --------------- +2022-05-10 13:22:32.707073 PDT | [2] Epoch -934 finished +---------------------------------- --------------- +epoch -934 +replay_buffer/size 999033 +trainer/num train calls 67000 +trainer/Policy Loss -20.1374 +trainer/Log Pis Mean 25.0671 +trainer/Log Pis Std 13.2976 +trainer/Log Pis Max 78.0311 +trainer/Log Pis Min -7.01289 +trainer/policy/mean Mean -0.00877843 +trainer/policy/mean Std 0.90973 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.83509 +trainer/policy/normal/std Std 0.527468 +trainer/policy/normal/std Max 5.18958 +trainer/policy/normal/std Min 0.512037 +trainer/policy/normal/log_std Mean 1.01925 +trainer/policy/normal/log_std Std 0.233349 +trainer/policy/normal/log_std Max 1.64665 +trainer/policy/normal/log_std Min -0.669358 +eval/num steps total 67000 +eval/num paths total 67 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0916019 +eval/Actions Std 0.890394 +eval/Actions Max 0.999995 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.75532 +time/logging (s) 0.00384724 +time/sampling batch (s) 0.529239 +time/saving (s) 0.0035786 +time/training (s) 7.30325 +time/epoch (s) 10.5952 +time/total (s) 718.192 +Epoch -934 +---------------------------------- --------------- +2022-05-10 13:22:43.192279 PDT | [2] Epoch -933 finished +---------------------------------- --------------- +epoch -933 +replay_buffer/size 999033 +trainer/num train calls 68000 +trainer/Policy Loss -18.4162 +trainer/Log Pis Mean 23.5309 +trainer/Log Pis Std 13.0106 +trainer/Log Pis Max 67.3352 +trainer/Log Pis Min -7.69783 +trainer/policy/mean Mean -0.0410716 +trainer/policy/mean Std 0.904424 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.90497 +trainer/policy/normal/std Std 0.536346 +trainer/policy/normal/std Max 6.24887 +trainer/policy/normal/std Min 0.614598 +trainer/policy/normal/log_std Mean 1.0441 +trainer/policy/normal/log_std Std 0.230086 +trainer/policy/normal/log_std Max 1.8324 +trainer/policy/normal/log_std Min -0.486787 +eval/num steps total 68000 +eval/num paths total 68 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.276493 +eval/Actions Std 0.907153 +eval/Actions Max 0.999984 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65348 +time/logging (s) 0.00412653 +time/sampling batch (s) 0.283078 +time/saving (s) 0.00428038 +time/training (s) 7.52069 +time/epoch (s) 10.4657 +time/total (s) 728.661 +Epoch -933 +---------------------------------- --------------- +2022-05-10 13:22:53.352926 PDT | [2] Epoch -932 finished +---------------------------------- --------------- +epoch -932 +replay_buffer/size 999033 +trainer/num train calls 69000 +trainer/Policy Loss -19.5802 +trainer/Log Pis Mean 24.061 +trainer/Log Pis Std 12.8565 +trainer/Log Pis Max 72.298 +trainer/Log Pis Min -7.33793 +trainer/policy/mean Mean -0.0440661 +trainer/policy/mean Std 0.901875 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.87099 +trainer/policy/normal/std Std 0.517526 +trainer/policy/normal/std Max 4.83636 +trainer/policy/normal/std Min 0.523119 +trainer/policy/normal/log_std Mean 1.0337 +trainer/policy/normal/log_std Std 0.222341 +trainer/policy/normal/log_std Max 1.57616 +trainer/policy/normal/log_std Min -0.647946 +eval/num steps total 69000 +eval/num paths total 69 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0782261 +eval/Actions Std 0.934924 +eval/Actions Max 0.999993 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66313 +time/logging (s) 0.00373928 +time/sampling batch (s) 0.283622 +time/saving (s) 0.00355192 +time/training (s) 7.18597 +time/epoch (s) 10.14 +time/total (s) 738.804 +Epoch -932 +---------------------------------- --------------- +2022-05-10 13:23:03.998194 PDT | [2] Epoch -931 finished +---------------------------------- --------------- +epoch -931 +replay_buffer/size 999033 +trainer/num train calls 70000 +trainer/Policy Loss -18.327 +trainer/Log Pis Mean 24.0927 +trainer/Log Pis Std 12.6123 +trainer/Log Pis Max 64.8537 +trainer/Log Pis Min -7.04534 +trainer/policy/mean Mean -0.0273334 +trainer/policy/mean Std 0.906855 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.83991 +trainer/policy/normal/std Std 0.509832 +trainer/policy/normal/std Max 4.89344 +trainer/policy/normal/std Min 0.502333 +trainer/policy/normal/log_std Mean 1.0221 +trainer/policy/normal/log_std Std 0.229156 +trainer/policy/normal/log_std Max 1.5879 +trainer/policy/normal/log_std Min -0.688493 +eval/num steps total 70000 +eval/num paths total 70 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.10531 +eval/Actions Std 0.917171 +eval/Actions Max 0.99999 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58752 +time/logging (s) 0.00369499 +time/sampling batch (s) 0.2837 +time/saving (s) 0.00346844 +time/training (s) 7.74687 +time/epoch (s) 10.6253 +time/total (s) 749.432 +Epoch -931 +---------------------------------- --------------- +2022-05-10 13:23:15.420241 PDT | [2] Epoch -930 finished +---------------------------------- --------------- +epoch -930 +replay_buffer/size 999033 +trainer/num train calls 71000 +trainer/Policy Loss -19.1156 +trainer/Log Pis Mean 23.8695 +trainer/Log Pis Std 13.1039 +trainer/Log Pis Max 76.5213 +trainer/Log Pis Min -5.62516 +trainer/policy/mean Mean -0.037772 +trainer/policy/mean Std 0.898628 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.8835 +trainer/policy/normal/std Std 0.571466 +trainer/policy/normal/std Max 5.11669 +trainer/policy/normal/std Min 0.439915 +trainer/policy/normal/log_std Mean 1.03294 +trainer/policy/normal/log_std Std 0.250527 +trainer/policy/normal/log_std Max 1.63251 +trainer/policy/normal/log_std Min -0.821174 +eval/num steps total 71000 +eval/num paths total 71 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.199087 +eval/Actions Std 0.93564 +eval/Actions Max 0.999994 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5776 +time/logging (s) 0.00367588 +time/sampling batch (s) 0.53385 +time/saving (s) 0.00342212 +time/training (s) 8.28365 +time/epoch (s) 11.4022 +time/total (s) 760.837 +Epoch -930 +---------------------------------- --------------- +2022-05-10 13:23:25.205698 PDT | [2] Epoch -929 finished +---------------------------------- --------------- +epoch -929 +replay_buffer/size 999033 +trainer/num train calls 72000 +trainer/Policy Loss -19.6204 +trainer/Log Pis Mean 23.3759 +trainer/Log Pis Std 13.0883 +trainer/Log Pis Max 67.5899 +trainer/Log Pis Min -12.6765 +trainer/policy/mean Mean -0.0594231 +trainer/policy/mean Std 0.899965 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.88928 +trainer/policy/normal/std Std 0.546636 +trainer/policy/normal/std Max 5.01185 +trainer/policy/normal/std Min 0.428159 +trainer/policy/normal/log_std Mean 1.03733 +trainer/policy/normal/log_std Std 0.238155 +trainer/policy/normal/log_std Max 1.61181 +trainer/policy/normal/log_std Min -0.84826 +eval/num steps total 72000 +eval/num paths total 72 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0365816 +eval/Actions Std 0.954369 +eval/Actions Max 0.999999 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68652 +time/logging (s) 0.00373931 +time/sampling batch (s) 0.283621 +time/saving (s) 0.00356222 +time/training (s) 6.78814 +time/epoch (s) 9.76558 +time/total (s) 770.606 +Epoch -929 +---------------------------------- --------------- +2022-05-10 13:23:35.198253 PDT | [2] Epoch -928 finished +---------------------------------- --------------- +epoch -928 +replay_buffer/size 999033 +trainer/num train calls 73000 +trainer/Policy Loss -18.7877 +trainer/Log Pis Mean 23.4503 +trainer/Log Pis Std 13.0983 +trainer/Log Pis Max 72.7194 +trainer/Log Pis Min -7.82192 +trainer/policy/mean Mean -0.0272782 +trainer/policy/mean Std 0.904058 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.86969 +trainer/policy/normal/std Std 0.549593 +trainer/policy/normal/std Max 5.80659 +trainer/policy/normal/std Min 0.540347 +trainer/policy/normal/log_std Mean 1.03016 +trainer/policy/normal/log_std Std 0.239388 +trainer/policy/normal/log_std Max 1.75899 +trainer/policy/normal/log_std Min -0.615543 +eval/num steps total 73000 +eval/num paths total 73 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.218152 +eval/Actions Std 0.95061 +eval/Actions Max 0.999992 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66881 +time/logging (s) 0.00418673 +time/sampling batch (s) 0.281344 +time/saving (s) 0.00420145 +time/training (s) 7.01451 +time/epoch (s) 9.97305 +time/total (s) 780.582 +Epoch -928 +---------------------------------- --------------- +2022-05-10 13:23:46.016288 PDT | [2] Epoch -927 finished +---------------------------------- --------------- +epoch -927 +replay_buffer/size 999033 +trainer/num train calls 74000 +trainer/Policy Loss -19.5222 +trainer/Log Pis Mean 24.5569 +trainer/Log Pis Std 13.118 +trainer/Log Pis Max 71.5051 +trainer/Log Pis Min -5.06316 +trainer/policy/mean Mean -0.0497586 +trainer/policy/mean Std 0.902111 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.87484 +trainer/policy/normal/std Std 0.553115 +trainer/policy/normal/std Max 5.22166 +trainer/policy/normal/std Min 0.540879 +trainer/policy/normal/log_std Mean 1.0316 +trainer/policy/normal/log_std Std 0.241569 +trainer/policy/normal/log_std Max 1.65282 +trainer/policy/normal/log_std Min -0.61456 +eval/num steps total 74000 +eval/num paths total 74 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0475521 +eval/Actions Std 0.903881 +eval/Actions Max 0.999993 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65443 +time/logging (s) 0.00382598 +time/sampling batch (s) 0.781277 +time/saving (s) 0.0036009 +time/training (s) 7.35469 +time/epoch (s) 10.7978 +time/total (s) 791.383 +Epoch -927 +---------------------------------- --------------- +2022-05-10 13:23:56.516173 PDT | [2] Epoch -926 finished +---------------------------------- --------------- +epoch -926 +replay_buffer/size 999033 +trainer/num train calls 75000 +trainer/Policy Loss -18.7764 +trainer/Log Pis Mean 24.7151 +trainer/Log Pis Std 12.5817 +trainer/Log Pis Max 66.9855 +trainer/Log Pis Min -6.8017 +trainer/policy/mean Mean -0.0441191 +trainer/policy/mean Std 0.900675 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.91072 +trainer/policy/normal/std Std 0.542035 +trainer/policy/normal/std Max 5.15512 +trainer/policy/normal/std Min 0.532909 +trainer/policy/normal/log_std Mean 1.04439 +trainer/policy/normal/log_std Std 0.242946 +trainer/policy/normal/log_std Max 1.63999 +trainer/policy/normal/log_std Min -0.629404 +eval/num steps total 75000 +eval/num paths total 75 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.182219 +eval/Actions Std 0.899195 +eval/Actions Max 0.999979 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60824 +time/logging (s) 0.00374596 +time/sampling batch (s) 0.282059 +time/saving (s) 0.00360419 +time/training (s) 7.58256 +time/epoch (s) 10.4802 +time/total (s) 801.866 +Epoch -926 +---------------------------------- --------------- +2022-05-10 13:24:06.811590 PDT | [2] Epoch -925 finished +---------------------------------- --------------- +epoch -925 +replay_buffer/size 999033 +trainer/num train calls 76000 +trainer/Policy Loss -19.4849 +trainer/Log Pis Mean 23.8569 +trainer/Log Pis Std 13.2063 +trainer/Log Pis Max 74.6947 +trainer/Log Pis Min -3.57462 +trainer/policy/mean Mean -0.0548245 +trainer/policy/mean Std 0.902015 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.86495 +trainer/policy/normal/std Std 0.540344 +trainer/policy/normal/std Max 5.11256 +trainer/policy/normal/std Min 0.535227 +trainer/policy/normal/log_std Mean 1.02873 +trainer/policy/normal/log_std Std 0.239074 +trainer/policy/normal/log_std Max 1.6317 +trainer/policy/normal/log_std Min -0.625064 +eval/num steps total 76000 +eval/num paths total 76 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.31282 +eval/Actions Std 0.831494 +eval/Actions Max 0.999999 +eval/Actions Min -0.999984 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58906 +time/logging (s) 0.00386312 +time/sampling batch (s) 0.281878 +time/saving (s) 0.00367624 +time/training (s) 7.39752 +time/epoch (s) 10.276 +time/total (s) 812.145 +Epoch -925 +---------------------------------- --------------- +2022-05-10 13:24:16.561083 PDT | [2] Epoch -924 finished +---------------------------------- --------------- +epoch -924 +replay_buffer/size 999033 +trainer/num train calls 77000 +trainer/Policy Loss -19.891 +trainer/Log Pis Mean 24.0642 +trainer/Log Pis Std 13.461 +trainer/Log Pis Max 67.1618 +trainer/Log Pis Min -8.93839 +trainer/policy/mean Mean -0.0438924 +trainer/policy/mean Std 0.901613 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.87634 +trainer/policy/normal/std Std 0.534506 +trainer/policy/normal/std Max 5.06114 +trainer/policy/normal/std Min 0.611762 +trainer/policy/normal/log_std Mean 1.03382 +trainer/policy/normal/log_std Std 0.232374 +trainer/policy/normal/log_std Max 1.62159 +trainer/policy/normal/log_std Min -0.491413 +eval/num steps total 77000 +eval/num paths total 77 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.120375 +eval/Actions Std 0.900014 +eval/Actions Max 0.999993 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56634 +time/logging (s) 0.00379766 +time/sampling batch (s) 0.280184 +time/saving (s) 0.00349986 +time/training (s) 6.87576 +time/epoch (s) 9.72958 +time/total (s) 821.878 +Epoch -924 +---------------------------------- --------------- +2022-05-10 13:24:27.384411 PDT | [2] Epoch -923 finished +---------------------------------- --------------- +epoch -923 +replay_buffer/size 999033 +trainer/num train calls 78000 +trainer/Policy Loss -19.8055 +trainer/Log Pis Mean 24.4004 +trainer/Log Pis Std 13.6664 +trainer/Log Pis Max 74.0099 +trainer/Log Pis Min -7.85376 +trainer/policy/mean Mean -0.0418987 +trainer/policy/mean Std 0.911382 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.90504 +trainer/policy/normal/std Std 0.522925 +trainer/policy/normal/std Max 5.29123 +trainer/policy/normal/std Min 0.562501 +trainer/policy/normal/log_std Mean 1.04561 +trainer/policy/normal/log_std Std 0.221425 +trainer/policy/normal/log_std Max 1.66605 +trainer/policy/normal/log_std Min -0.575362 +eval/num steps total 78000 +eval/num paths total 78 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00784775 +eval/Actions Std 0.913537 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66504 +time/logging (s) 0.00405509 +time/sampling batch (s) 0.532152 +time/saving (s) 0.00390462 +time/training (s) 7.59861 +time/epoch (s) 10.8038 +time/total (s) 832.685 +Epoch -923 +---------------------------------- --------------- +2022-05-10 13:24:37.592243 PDT | [2] Epoch -922 finished +---------------------------------- --------------- +epoch -922 +replay_buffer/size 999033 +trainer/num train calls 79000 +trainer/Policy Loss -19.3268 +trainer/Log Pis Mean 24.1239 +trainer/Log Pis Std 13.1239 +trainer/Log Pis Max 70.2517 +trainer/Log Pis Min -9.71382 +trainer/policy/mean Mean -0.0281659 +trainer/policy/mean Std 0.901479 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.91694 +trainer/policy/normal/std Std 0.56109 +trainer/policy/normal/std Max 4.87202 +trainer/policy/normal/std Min 0.576179 +trainer/policy/normal/log_std Mean 1.04622 +trainer/policy/normal/log_std Std 0.24091 +trainer/policy/normal/log_std Max 1.58351 +trainer/policy/normal/log_std Min -0.551336 +eval/num steps total 79000 +eval/num paths total 79 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.071946 +eval/Actions Std 0.914757 +eval/Actions Max 0.999993 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5453 +time/logging (s) 0.00394564 +time/sampling batch (s) 0.279723 +time/saving (s) 0.00379047 +time/training (s) 7.35507 +time/epoch (s) 10.1878 +time/total (s) 842.876 +Epoch -922 +---------------------------------- --------------- +2022-05-10 13:24:47.134487 PDT | [2] Epoch -921 finished +---------------------------------- --------------- +epoch -921 +replay_buffer/size 999033 +trainer/num train calls 80000 +trainer/Policy Loss -19.6354 +trainer/Log Pis Mean 23.6297 +trainer/Log Pis Std 12.9514 +trainer/Log Pis Max 61.3318 +trainer/Log Pis Min -4.42288 +trainer/policy/mean Mean -0.0517383 +trainer/policy/mean Std 0.90311 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.88566 +trainer/policy/normal/std Std 0.54069 +trainer/policy/normal/std Max 5.51824 +trainer/policy/normal/std Min 0.580737 +trainer/policy/normal/log_std Mean 1.03672 +trainer/policy/normal/log_std Std 0.23442 +trainer/policy/normal/log_std Max 1.70806 +trainer/policy/normal/log_std Min -0.543457 +eval/num steps total 80000 +eval/num paths total 80 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.104212 +eval/Actions Std 0.904552 +eval/Actions Max 0.999996 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.32439 +time/logging (s) 0.00372322 +time/sampling batch (s) 0.279404 +time/saving (s) 0.00361402 +time/training (s) 6.91115 +time/epoch (s) 9.52228 +time/total (s) 852.401 +Epoch -921 +---------------------------------- --------------- +2022-05-10 13:24:57.625008 PDT | [2] Epoch -920 finished +---------------------------------- --------------- +epoch -920 +replay_buffer/size 999033 +trainer/num train calls 81000 +trainer/Policy Loss -19.2681 +trainer/Log Pis Mean 24.4479 +trainer/Log Pis Std 13.4464 +trainer/Log Pis Max 75.2114 +trainer/Log Pis Min -8.58308 +trainer/policy/mean Mean -0.0443323 +trainer/policy/mean Std 0.903689 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.90856 +trainer/policy/normal/std Std 0.563761 +trainer/policy/normal/std Max 5.42708 +trainer/policy/normal/std Min 0.479253 +trainer/policy/normal/log_std Mean 1.04273 +trainer/policy/normal/log_std Std 0.244736 +trainer/policy/normal/log_std Max 1.6914 +trainer/policy/normal/log_std Min -0.735528 +eval/num steps total 81000 +eval/num paths total 81 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0702655 +eval/Actions Std 0.913531 +eval/Actions Max 0.999996 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69852 +time/logging (s) 0.00383003 +time/sampling batch (s) 0.530645 +time/saving (s) 0.00343075 +time/training (s) 7.23456 +time/epoch (s) 10.471 +time/total (s) 862.875 +Epoch -920 +---------------------------------- --------------- +2022-05-10 13:25:08.592257 PDT | [2] Epoch -919 finished +---------------------------------- --------------- +epoch -919 +replay_buffer/size 999033 +trainer/num train calls 82000 +trainer/Policy Loss -18.6528 +trainer/Log Pis Mean 25.2589 +trainer/Log Pis Std 13.0701 +trainer/Log Pis Max 67.5051 +trainer/Log Pis Min -7.15507 +trainer/policy/mean Mean -0.0401364 +trainer/policy/mean Std 0.902651 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.92395 +trainer/policy/normal/std Std 0.547466 +trainer/policy/normal/std Max 5.5358 +trainer/policy/normal/std Min 0.623128 +trainer/policy/normal/log_std Mean 1.05021 +trainer/policy/normal/log_std Std 0.232157 +trainer/policy/normal/log_std Max 1.71124 +trainer/policy/normal/log_std Min -0.473004 +eval/num steps total 82000 +eval/num paths total 82 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.118462 +eval/Actions Std 0.902391 +eval/Actions Max 0.999997 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66583 +time/logging (s) 0.00378311 +time/sampling batch (s) 0.532787 +time/saving (s) 0.00346257 +time/training (s) 7.74157 +time/epoch (s) 10.9474 +time/total (s) 873.825 +Epoch -919 +---------------------------------- --------------- +2022-05-10 13:25:19.084232 PDT | [2] Epoch -918 finished +---------------------------------- --------------- +epoch -918 +replay_buffer/size 999033 +trainer/num train calls 83000 +trainer/Policy Loss -19.6547 +trainer/Log Pis Mean 23.5301 +trainer/Log Pis Std 13.8699 +trainer/Log Pis Max 78.3419 +trainer/Log Pis Min -5.97215 +trainer/policy/mean Mean -0.0369775 +trainer/policy/mean Std 0.903375 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.88921 +trainer/policy/normal/std Std 0.55686 +trainer/policy/normal/std Max 5.33855 +trainer/policy/normal/std Min 0.495254 +trainer/policy/normal/log_std Mean 1.03558 +trainer/policy/normal/log_std Std 0.248693 +trainer/policy/normal/log_std Max 1.67495 +trainer/policy/normal/log_std Min -0.702684 +eval/num steps total 83000 +eval/num paths total 83 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0126065 +eval/Actions Std 0.893125 +eval/Actions Max 0.999992 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59903 +time/logging (s) 0.00374282 +time/sampling batch (s) 0.533256 +time/saving (s) 0.00347083 +time/training (s) 7.33248 +time/epoch (s) 10.472 +time/total (s) 884.3 +Epoch -918 +---------------------------------- --------------- +2022-05-10 13:25:29.591170 PDT | [2] Epoch -917 finished +---------------------------------- --------------- +epoch -917 +replay_buffer/size 999033 +trainer/num train calls 84000 +trainer/Policy Loss -18.4189 +trainer/Log Pis Mean 23.227 +trainer/Log Pis Std 13.1601 +trainer/Log Pis Max 71.8977 +trainer/Log Pis Min -10.3871 +trainer/policy/mean Mean -0.0583742 +trainer/policy/mean Std 0.902753 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.88369 +trainer/policy/normal/std Std 0.559056 +trainer/policy/normal/std Max 5.03155 +trainer/policy/normal/std Min 0.465065 +trainer/policy/normal/log_std Mean 1.03433 +trainer/policy/normal/log_std Std 0.242892 +trainer/policy/normal/log_std Max 1.61573 +trainer/policy/normal/log_std Min -0.765579 +eval/num steps total 84000 +eval/num paths total 84 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.165048 +eval/Actions Std 0.933761 +eval/Actions Max 0.999996 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6542 +time/logging (s) 0.0039696 +time/sampling batch (s) 0.535139 +time/saving (s) 0.00415178 +time/training (s) 7.28956 +time/epoch (s) 10.487 +time/total (s) 894.79 +Epoch -917 +---------------------------------- --------------- +2022-05-10 13:25:40.337294 PDT | [2] Epoch -916 finished +---------------------------------- --------------- +epoch -916 +replay_buffer/size 999033 +trainer/num train calls 85000 +trainer/Policy Loss -19.8298 +trainer/Log Pis Mean 24.3266 +trainer/Log Pis Std 13.5981 +trainer/Log Pis Max 74.4448 +trainer/Log Pis Min -5.75165 +trainer/policy/mean Mean -0.0321042 +trainer/policy/mean Std 0.906506 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.8869 +trainer/policy/normal/std Std 0.547494 +trainer/policy/normal/std Max 4.9482 +trainer/policy/normal/std Min 0.58403 +trainer/policy/normal/log_std Mean 1.03636 +trainer/policy/normal/log_std Std 0.238571 +trainer/policy/normal/log_std Max 1.59902 +trainer/policy/normal/log_std Min -0.537803 +eval/num steps total 85000 +eval/num paths total 85 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.426176 +eval/Actions Std 0.872641 +eval/Actions Max 0.999996 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62895 +time/logging (s) 0.00374244 +time/sampling batch (s) 0.283421 +time/saving (s) 0.00357095 +time/training (s) 7.80611 +time/epoch (s) 10.7258 +time/total (s) 905.519 +Epoch -916 +---------------------------------- --------------- +2022-05-10 13:25:49.643492 PDT | [2] Epoch -915 finished +---------------------------------- --------------- +epoch -915 +replay_buffer/size 999033 +trainer/num train calls 86000 +trainer/Policy Loss -20.6943 +trainer/Log Pis Mean 25.474 +trainer/Log Pis Std 13.8202 +trainer/Log Pis Max 75.7917 +trainer/Log Pis Min -8.41898 +trainer/policy/mean Mean -0.0327558 +trainer/policy/mean Std 0.905739 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.98251 +trainer/policy/normal/std Std 0.556078 +trainer/policy/normal/std Max 5.39295 +trainer/policy/normal/std Min 0.562445 +trainer/policy/normal/log_std Mean 1.07051 +trainer/policy/normal/log_std Std 0.229057 +trainer/policy/normal/log_std Max 1.68509 +trainer/policy/normal/log_std Min -0.575463 +eval/num steps total 86000 +eval/num paths total 86 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.170728 +eval/Actions Std 0.888411 +eval/Actions Max 0.999995 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50765 +time/logging (s) 0.00421366 +time/sampling batch (s) 0.28372 +time/saving (s) 0.0035708 +time/training (s) 6.48759 +time/epoch (s) 9.28675 +time/total (s) 914.809 +Epoch -915 +---------------------------------- --------------- +2022-05-10 13:25:58.532828 PDT | [2] Epoch -914 finished +---------------------------------- --------------- +epoch -914 +replay_buffer/size 999033 +trainer/num train calls 87000 +trainer/Policy Loss -19.6456 +trainer/Log Pis Mean 23.99 +trainer/Log Pis Std 13.0737 +trainer/Log Pis Max 65.8712 +trainer/Log Pis Min -8.83087 +trainer/policy/mean Mean -0.0354043 +trainer/policy/mean Std 0.901313 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.88213 +trainer/policy/normal/std Std 0.543748 +trainer/policy/normal/std Max 5.26737 +trainer/policy/normal/std Min 0.468327 +trainer/policy/normal/log_std Mean 1.03431 +trainer/policy/normal/log_std Std 0.242932 +trainer/policy/normal/log_std Max 1.66153 +trainer/policy/normal/log_std Min -0.75859 +eval/num steps total 87000 +eval/num paths total 87 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0994911 +eval/Actions Std 0.895835 +eval/Actions Max 0.999997 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56454 +time/logging (s) 0.0037482 +time/sampling batch (s) 0.283256 +time/saving (s) 0.00348326 +time/training (s) 6.01357 +time/epoch (s) 8.8686 +time/total (s) 923.681 +Epoch -914 +---------------------------------- --------------- +2022-05-10 13:26:08.983628 PDT | [2] Epoch -913 finished +---------------------------------- --------------- +epoch -913 +replay_buffer/size 999033 +trainer/num train calls 88000 +trainer/Policy Loss -19.0394 +trainer/Log Pis Mean 24.7646 +trainer/Log Pis Std 13.4638 +trainer/Log Pis Max 71.6072 +trainer/Log Pis Min -5.35699 +trainer/policy/mean Mean -0.0137284 +trainer/policy/mean Std 0.905379 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.92664 +trainer/policy/normal/std Std 0.565948 +trainer/policy/normal/std Max 5.27981 +trainer/policy/normal/std Min 0.562609 +trainer/policy/normal/log_std Mean 1.04905 +trainer/policy/normal/log_std Std 0.243654 +trainer/policy/normal/log_std Max 1.66389 +trainer/policy/normal/log_std Min -0.575171 +eval/num steps total 88000 +eval/num paths total 88 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.128486 +eval/Actions Std 0.909044 +eval/Actions Max 0.999992 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.74889 +time/logging (s) 0.00394802 +time/sampling batch (s) 0.283329 +time/saving (s) 0.00353211 +time/training (s) 7.39148 +time/epoch (s) 10.4312 +time/total (s) 934.115 +Epoch -913 +---------------------------------- --------------- +2022-05-10 13:26:18.798657 PDT | [2] Epoch -912 finished +---------------------------------- --------------- +epoch -912 +replay_buffer/size 999033 +trainer/num train calls 89000 +trainer/Policy Loss -18.8952 +trainer/Log Pis Mean 24.9512 +trainer/Log Pis Std 12.8068 +trainer/Log Pis Max 80.2123 +trainer/Log Pis Min -5.55381 +trainer/policy/mean Mean -0.0538712 +trainer/policy/mean Std 0.907122 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.91521 +trainer/policy/normal/std Std 0.552249 +trainer/policy/normal/std Max 5.75394 +trainer/policy/normal/std Min 0.519522 +trainer/policy/normal/log_std Mean 1.04723 +trainer/policy/normal/log_std Std 0.230555 +trainer/policy/normal/log_std Max 1.74989 +trainer/policy/normal/log_std Min -0.654846 +eval/num steps total 89000 +eval/num paths total 89 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.129894 +eval/Actions Std 0.908087 +eval/Actions Max 0.999998 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.72535 +time/logging (s) 0.00411565 +time/sampling batch (s) 0.285714 +time/saving (s) 0.00412836 +time/training (s) 6.77575 +time/epoch (s) 9.79506 +time/total (s) 943.913 +Epoch -912 +---------------------------------- --------------- +2022-05-10 13:26:28.173141 PDT | [2] Epoch -911 finished +---------------------------------- --------------- +epoch -911 +replay_buffer/size 999033 +trainer/num train calls 90000 +trainer/Policy Loss -19.1146 +trainer/Log Pis Mean 26.3111 +trainer/Log Pis Std 14.205 +trainer/Log Pis Max 67.3839 +trainer/Log Pis Min -9.60057 +trainer/policy/mean Mean -0.0240966 +trainer/policy/mean Std 0.908852 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999982 +trainer/policy/normal/std Mean 2.90877 +trainer/policy/normal/std Std 0.554118 +trainer/policy/normal/std Max 4.6912 +trainer/policy/normal/std Min 0.457274 +trainer/policy/normal/log_std Mean 1.04313 +trainer/policy/normal/log_std Std 0.244704 +trainer/policy/normal/log_std Max 1.54569 +trainer/policy/normal/log_std Min -0.782473 +eval/num steps total 90000 +eval/num paths total 90 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.273787 +eval/Actions Std 0.681073 +eval/Actions Max 0.999999 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65155 +time/logging (s) 0.00369328 +time/sampling batch (s) 0.532427 +time/saving (s) 0.00359221 +time/training (s) 6.16269 +time/epoch (s) 9.35396 +time/total (s) 953.27 +Epoch -911 +---------------------------------- --------------- +2022-05-10 13:26:38.168195 PDT | [2] Epoch -910 finished +---------------------------------- --------------- +epoch -910 +replay_buffer/size 999033 +trainer/num train calls 91000 +trainer/Policy Loss -18.3444 +trainer/Log Pis Mean 24.2156 +trainer/Log Pis Std 12.7566 +trainer/Log Pis Max 67.5298 +trainer/Log Pis Min -5.60054 +trainer/policy/mean Mean -0.0192848 +trainer/policy/mean Std 0.9019 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.88692 +trainer/policy/normal/std Std 0.575557 +trainer/policy/normal/std Max 5.31125 +trainer/policy/normal/std Min 0.509397 +trainer/policy/normal/log_std Mean 1.03298 +trainer/policy/normal/log_std Std 0.258013 +trainer/policy/normal/log_std Max 1.66983 +trainer/policy/normal/log_std Min -0.674527 +eval/num steps total 91000 +eval/num paths total 91 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.105783 +eval/Actions Std 0.91734 +eval/Actions Max 0.999999 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67938 +time/logging (s) 0.00373336 +time/sampling batch (s) 0.283456 +time/saving (s) 0.00348593 +time/training (s) 7.00495 +time/epoch (s) 9.975 +time/total (s) 963.248 +Epoch -910 +---------------------------------- --------------- +2022-05-10 13:26:47.568146 PDT | [2] Epoch -909 finished +---------------------------------- --------------- +epoch -909 +replay_buffer/size 999033 +trainer/num train calls 92000 +trainer/Policy Loss -18.8898 +trainer/Log Pis Mean 23.3727 +trainer/Log Pis Std 12.2115 +trainer/Log Pis Max 61.2223 +trainer/Log Pis Min -8.17771 +trainer/policy/mean Mean -0.00912393 +trainer/policy/mean Std 0.897922 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.86156 +trainer/policy/normal/std Std 0.559052 +trainer/policy/normal/std Max 5.08727 +trainer/policy/normal/std Min 0.464293 +trainer/policy/normal/log_std Mean 1.02556 +trainer/policy/normal/log_std Std 0.250275 +trainer/policy/normal/log_std Max 1.62674 +trainer/policy/normal/log_std Min -0.767239 +eval/num steps total 92000 +eval/num paths total 92 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0131937 +eval/Actions Std 0.87589 +eval/Actions Max 0.999997 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73437 +time/logging (s) 0.00368346 +time/sampling batch (s) 0.28424 +time/saving (s) 0.00347448 +time/training (s) 6.35403 +time/epoch (s) 9.3798 +time/total (s) 972.631 +Epoch -909 +---------------------------------- --------------- +2022-05-10 13:26:57.428093 PDT | [2] Epoch -908 finished +---------------------------------- --------------- +epoch -908 +replay_buffer/size 999033 +trainer/num train calls 93000 +trainer/Policy Loss -20.1751 +trainer/Log Pis Mean 24.0804 +trainer/Log Pis Std 13.1994 +trainer/Log Pis Max 66.3689 +trainer/Log Pis Min -11.3404 +trainer/policy/mean Mean -0.0243825 +trainer/policy/mean Std 0.907077 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.84977 +trainer/policy/normal/std Std 0.543939 +trainer/policy/normal/std Max 5.42364 +trainer/policy/normal/std Min 0.494262 +trainer/policy/normal/log_std Mean 1.02308 +trainer/policy/normal/log_std Std 0.240855 +trainer/policy/normal/log_std Max 1.69077 +trainer/policy/normal/log_std Min -0.70469 +eval/num steps total 93000 +eval/num paths total 93 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.175351 +eval/Actions Std 0.893871 +eval/Actions Max 0.999996 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69125 +time/logging (s) 0.00367381 +time/sampling batch (s) 0.283778 +time/saving (s) 0.00344703 +time/training (s) 6.85801 +time/epoch (s) 9.84016 +time/total (s) 982.474 +Epoch -908 +---------------------------------- --------------- +2022-05-10 13:27:08.173432 PDT | [2] Epoch -907 finished +---------------------------------- --------------- +epoch -907 +replay_buffer/size 999033 +trainer/num train calls 94000 +trainer/Policy Loss -20.6993 +trainer/Log Pis Mean 25.049 +trainer/Log Pis Std 13.5399 +trainer/Log Pis Max 75.2707 +trainer/Log Pis Min -6.39897 +trainer/policy/mean Mean -0.0167856 +trainer/policy/mean Std 0.903231 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.87618 +trainer/policy/normal/std Std 0.55019 +trainer/policy/normal/std Max 5.04641 +trainer/policy/normal/std Min 0.504959 +trainer/policy/normal/log_std Mean 1.03151 +trainer/policy/normal/log_std Std 0.246674 +trainer/policy/normal/log_std Max 1.61868 +trainer/policy/normal/log_std Min -0.683278 +eval/num steps total 94000 +eval/num paths total 94 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.123552 +eval/Actions Std 0.918257 +eval/Actions Max 0.999982 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.47336 +time/logging (s) 0.00434872 +time/sampling batch (s) 0.296643 +time/saving (s) 0.00424086 +time/training (s) 7.94679 +time/epoch (s) 10.7254 +time/total (s) 993.203 +Epoch -907 +---------------------------------- --------------- +2022-05-10 13:27:19.770465 PDT | [2] Epoch -906 finished +---------------------------------- --------------- +epoch -906 +replay_buffer/size 999033 +trainer/num train calls 95000 +trainer/Policy Loss -17.9518 +trainer/Log Pis Mean 24.1832 +trainer/Log Pis Std 12.9637 +trainer/Log Pis Max 65.2526 +trainer/Log Pis Min -3.7382 +trainer/policy/mean Mean -0.0376054 +trainer/policy/mean Std 0.905317 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.89846 +trainer/policy/normal/std Std 0.544389 +trainer/policy/normal/std Max 5.28546 +trainer/policy/normal/std Min 0.529721 +trainer/policy/normal/log_std Mean 1.04091 +trainer/policy/normal/log_std Std 0.236256 +trainer/policy/normal/log_std Max 1.66496 +trainer/policy/normal/log_std Min -0.635405 +eval/num steps total 95000 +eval/num paths total 95 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.202097 +eval/Actions Std 0.933412 +eval/Actions Max 0.999998 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 3.02295 +time/logging (s) 0.00391827 +time/sampling batch (s) 0.337802 +time/saving (s) 0.00401584 +time/training (s) 8.20677 +time/epoch (s) 11.5755 +time/total (s) 1004.78 +Epoch -906 +---------------------------------- --------------- +2022-05-10 13:27:30.506999 PDT | [2] Epoch -905 finished +---------------------------------- --------------- +epoch -905 +replay_buffer/size 999033 +trainer/num train calls 96000 +trainer/Policy Loss -18.4724 +trainer/Log Pis Mean 25.5239 +trainer/Log Pis Std 13.2854 +trainer/Log Pis Max 62.4634 +trainer/Log Pis Min -7.54045 +trainer/policy/mean Mean -0.0426578 +trainer/policy/mean Std 0.897429 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.88513 +trainer/policy/normal/std Std 0.568132 +trainer/policy/normal/std Max 5.89936 +trainer/policy/normal/std Min 0.452847 +trainer/policy/normal/log_std Mean 1.03374 +trainer/policy/normal/log_std Std 0.249737 +trainer/policy/normal/log_std Max 1.77484 +trainer/policy/normal/log_std Min -0.792202 +eval/num steps total 96000 +eval/num paths total 96 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.48732 +eval/Actions Std 0.867162 +eval/Actions Max 0.999998 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 3.01112 +time/logging (s) 0.00371653 +time/sampling batch (s) 0.568321 +time/saving (s) 0.00344045 +time/training (s) 7.12911 +time/epoch (s) 10.7157 +time/total (s) 1015.5 +Epoch -905 +---------------------------------- --------------- +2022-05-10 13:27:39.968244 PDT | [2] Epoch -904 finished +---------------------------------- --------------- +epoch -904 +replay_buffer/size 999033 +trainer/num train calls 97000 +trainer/Policy Loss -18.3988 +trainer/Log Pis Mean 23.7638 +trainer/Log Pis Std 12.2181 +trainer/Log Pis Max 63.865 +trainer/Log Pis Min -5.03096 +trainer/policy/mean Mean -0.0289091 +trainer/policy/mean Std 0.902825 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.81991 +trainer/policy/normal/std Std 0.539967 +trainer/policy/normal/std Max 4.9279 +trainer/policy/normal/std Min 0.528796 +trainer/policy/normal/log_std Mean 1.01241 +trainer/policy/normal/log_std Std 0.241411 +trainer/policy/normal/log_std Max 1.59491 +trainer/policy/normal/log_std Min -0.637153 +eval/num steps total 97000 +eval/num paths total 97 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0232294 +eval/Actions Std 0.901621 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6169 +time/logging (s) 0.00371699 +time/sampling batch (s) 0.282945 +time/saving (s) 0.00343064 +time/training (s) 6.53431 +time/epoch (s) 9.4413 +time/total (s) 1024.94 +Epoch -904 +---------------------------------- --------------- +2022-05-10 13:27:50.299499 PDT | [2] Epoch -903 finished +---------------------------------- --------------- +epoch -903 +replay_buffer/size 999033 +trainer/num train calls 98000 +trainer/Policy Loss -18.2018 +trainer/Log Pis Mean 23.5827 +trainer/Log Pis Std 12.9408 +trainer/Log Pis Max 82.4026 +trainer/Log Pis Min -7.52784 +trainer/policy/mean Mean -0.0243411 +trainer/policy/mean Std 0.901194 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999985 +trainer/policy/normal/std Mean 2.8434 +trainer/policy/normal/std Std 0.562805 +trainer/policy/normal/std Max 5.51655 +trainer/policy/normal/std Min 0.457506 +trainer/policy/normal/log_std Mean 1.0184 +trainer/policy/normal/log_std Std 0.254913 +trainer/policy/normal/log_std Max 1.70775 +trainer/policy/normal/log_std Min -0.781966 +eval/num steps total 98000 +eval/num paths total 98 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0940604 +eval/Actions Std 0.906784 +eval/Actions Max 0.999987 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58512 +time/logging (s) 0.00378729 +time/sampling batch (s) 0.279376 +time/saving (s) 0.00345754 +time/training (s) 7.4399 +time/epoch (s) 10.3116 +time/total (s) 1035.26 +Epoch -903 +---------------------------------- --------------- +2022-05-10 13:27:59.721108 PDT | [2] Epoch -902 finished +---------------------------------- --------------- +epoch -902 +replay_buffer/size 999033 +trainer/num train calls 99000 +trainer/Policy Loss -19.5999 +trainer/Log Pis Mean 23.7014 +trainer/Log Pis Std 12.465 +trainer/Log Pis Max 69.4064 +trainer/Log Pis Min -7.30283 +trainer/policy/mean Mean -0.0390791 +trainer/policy/mean Std 0.902106 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.86195 +trainer/policy/normal/std Std 0.567244 +trainer/policy/normal/std Max 5.33026 +trainer/policy/normal/std Min 0.463824 +trainer/policy/normal/log_std Mean 1.02467 +trainer/policy/normal/log_std Std 0.255951 +trainer/policy/normal/log_std Max 1.6734 +trainer/policy/normal/log_std Min -0.76825 +eval/num steps total 99000 +eval/num paths total 99 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.196558 +eval/Actions Std 0.92393 +eval/Actions Max 0.999993 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.40766 +time/logging (s) 0.0041308 +time/sampling batch (s) 0.529176 +time/saving (s) 0.00408648 +time/training (s) 6.45706 +time/epoch (s) 9.40211 +time/total (s) 1044.66 +Epoch -902 +---------------------------------- --------------- +2022-05-10 13:28:09.746646 PDT | [2] Epoch -901 finished +---------------------------------- --------------- +epoch -901 +replay_buffer/size 999033 +trainer/num train calls 100000 +trainer/Policy Loss -19.2599 +trainer/Log Pis Mean 23.5568 +trainer/Log Pis Std 12.7855 +trainer/Log Pis Max 73.7622 +trainer/Log Pis Min -9.36845 +trainer/policy/mean Mean -0.038871 +trainer/policy/mean Std 0.905054 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.81495 +trainer/policy/normal/std Std 0.537509 +trainer/policy/normal/std Max 4.56141 +trainer/policy/normal/std Min 0.517922 +trainer/policy/normal/log_std Mean 1.01063 +trainer/policy/normal/log_std Std 0.241564 +trainer/policy/normal/log_std Max 1.51763 +trainer/policy/normal/log_std Min -0.65793 +eval/num steps total 100000 +eval/num paths total 100 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.142152 +eval/Actions Std 0.897045 +eval/Actions Max 0.999986 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63884 +time/logging (s) 0.00404311 +time/sampling batch (s) 0.284588 +time/saving (s) 0.00705831 +time/training (s) 7.07062 +time/epoch (s) 10.0051 +time/total (s) 1054.67 +Epoch -901 +---------------------------------- --------------- +2022-05-10 13:28:19.349384 PDT | [2] Epoch -900 finished +---------------------------------- --------------- +epoch -900 +replay_buffer/size 999033 +trainer/num train calls 101000 +trainer/Policy Loss -19.4346 +trainer/Log Pis Mean 24.2204 +trainer/Log Pis Std 13.1657 +trainer/Log Pis Max 61.9365 +trainer/Log Pis Min -6.31171 +trainer/policy/mean Mean -0.0456087 +trainer/policy/mean Std 0.904997 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.80488 +trainer/policy/normal/std Std 0.548689 +trainer/policy/normal/std Max 5.30792 +trainer/policy/normal/std Min 0.498611 +trainer/policy/normal/log_std Mean 1.00611 +trainer/policy/normal/log_std Std 0.245704 +trainer/policy/normal/log_std Max 1.6692 +trainer/policy/normal/log_std Min -0.695929 +eval/num steps total 101000 +eval/num paths total 101 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.106495 +eval/Actions Std 0.910204 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.77511 +time/logging (s) 0.00375622 +time/sampling batch (s) 0.278825 +time/saving (s) 0.00343887 +time/training (s) 6.52148 +time/epoch (s) 9.5826 +time/total (s) 1064.26 +Epoch -900 +---------------------------------- --------------- +2022-05-10 13:28:29.758308 PDT | [2] Epoch -899 finished +---------------------------------- --------------- +epoch -899 +replay_buffer/size 999033 +trainer/num train calls 102000 +trainer/Policy Loss -19.6793 +trainer/Log Pis Mean 24.8187 +trainer/Log Pis Std 12.7437 +trainer/Log Pis Max 63.8539 +trainer/Log Pis Min -8.53427 +trainer/policy/mean Mean -0.0394621 +trainer/policy/mean Std 0.90454 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.90439 +trainer/policy/normal/std Std 0.556244 +trainer/policy/normal/std Max 5.10236 +trainer/policy/normal/std Min 0.560448 +trainer/policy/normal/log_std Mean 1.0418 +trainer/policy/normal/log_std Std 0.241742 +trainer/policy/normal/log_std Max 1.6297 +trainer/policy/normal/log_std Min -0.57902 +eval/num steps total 101893 +eval/num paths total 102 +eval/path length Mean 893 +eval/path length Std 0 +eval/path length Max 893 +eval/path length Min 893 +eval/Rewards Mean 0.00111982 +eval/Rewards Std 0.033445 +eval/Rewards Max 1 +eval/Rewards Min 0 +eval/Returns Mean 1 +eval/Returns Std 0 +eval/Returns Max 1 +eval/Returns Min 1 +eval/Actions Mean -0.0471266 +eval/Actions Std 0.882773 +eval/Actions Max 0.999993 +eval/Actions Min -0.999984 +eval/Num Paths 1 +eval/Average Returns 1 +eval/normalized_score 100 +time/evaluation sampling (s) 2.62327 +time/logging (s) 0.00343104 +time/sampling batch (s) 0.279578 +time/saving (s) 0.003487 +time/training (s) 7.47913 +time/epoch (s) 10.3889 +time/total (s) 1074.65 +Epoch -899 +---------------------------------- --------------- +2022-05-10 13:28:39.701264 PDT | [2] Epoch -898 finished +---------------------------------- --------------- +epoch -898 +replay_buffer/size 999033 +trainer/num train calls 103000 +trainer/Policy Loss -20.43 +trainer/Log Pis Mean 24.8222 +trainer/Log Pis Std 13.1906 +trainer/Log Pis Max 68.822 +trainer/Log Pis Min -12.2923 +trainer/policy/mean Mean -0.0533306 +trainer/policy/mean Std 0.906442 +trainer/policy/mean Max 0.999964 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.87559 +trainer/policy/normal/std Std 0.578699 +trainer/policy/normal/std Max 5.4167 +trainer/policy/normal/std Min 0.517142 +trainer/policy/normal/log_std Mean 1.02921 +trainer/policy/normal/log_std Std 0.255378 +trainer/policy/normal/log_std Max 1.68949 +trainer/policy/normal/log_std Min -0.659438 +eval/num steps total 102893 +eval/num paths total 103 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.12378 +eval/Actions Std 0.911906 +eval/Actions Max 0.999989 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69218 +time/logging (s) 0.0036961 +time/sampling batch (s) 0.279915 +time/saving (s) 0.00347029 +time/training (s) 6.94425 +time/epoch (s) 9.92351 +time/total (s) 1084.58 +Epoch -898 +---------------------------------- --------------- +2022-05-10 13:28:48.945231 PDT | [2] Epoch -897 finished +---------------------------------- --------------- +epoch -897 +replay_buffer/size 999033 +trainer/num train calls 104000 +trainer/Policy Loss -20.1707 +trainer/Log Pis Mean 24.5444 +trainer/Log Pis Std 13.5248 +trainer/Log Pis Max 63.9669 +trainer/Log Pis Min -4.4056 +trainer/policy/mean Mean -0.0262848 +trainer/policy/mean Std 0.903505 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.87206 +trainer/policy/normal/std Std 0.567604 +trainer/policy/normal/std Max 5.00863 +trainer/policy/normal/std Min 0.4669 +trainer/policy/normal/log_std Mean 1.02789 +trainer/policy/normal/log_std Std 0.258943 +trainer/policy/normal/log_std Max 1.61116 +trainer/policy/normal/log_std Min -0.761641 +eval/num steps total 103893 +eval/num paths total 104 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.253804 +eval/Actions Std 0.864164 +eval/Actions Max 0.999997 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64688 +time/logging (s) 0.00371895 +time/sampling batch (s) 0.28468 +time/saving (s) 0.00338899 +time/training (s) 6.28536 +time/epoch (s) 9.22403 +time/total (s) 1093.8 +Epoch -897 +---------------------------------- --------------- +2022-05-10 13:28:58.925048 PDT | [2] Epoch -896 finished +---------------------------------- --------------- +epoch -896 +replay_buffer/size 999033 +trainer/num train calls 105000 +trainer/Policy Loss -18.4931 +trainer/Log Pis Mean 23.9153 +trainer/Log Pis Std 13.1897 +trainer/Log Pis Max 70.2144 +trainer/Log Pis Min -6.0945 +trainer/policy/mean Mean -0.0388841 +trainer/policy/mean Std 0.900069 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.79938 +trainer/policy/normal/std Std 0.561992 +trainer/policy/normal/std Max 4.60771 +trainer/policy/normal/std Min 0.524415 +trainer/policy/normal/log_std Mean 1.00182 +trainer/policy/normal/log_std Std 0.259244 +trainer/policy/normal/log_std Max 1.52773 +trainer/policy/normal/log_std Min -0.645471 +eval/num steps total 104893 +eval/num paths total 105 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0411325 +eval/Actions Std 0.904964 +eval/Actions Max 0.999993 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56273 +time/logging (s) 0.00417196 +time/sampling batch (s) 0.280716 +time/saving (s) 0.00411161 +time/training (s) 7.10871 +time/epoch (s) 9.96044 +time/total (s) 1103.77 +Epoch -896 +---------------------------------- --------------- +2022-05-10 13:29:09.377132 PDT | [2] Epoch -895 finished +---------------------------------- --------------- +epoch -895 +replay_buffer/size 999033 +trainer/num train calls 106000 +trainer/Policy Loss -19.5933 +trainer/Log Pis Mean 23.6179 +trainer/Log Pis Std 13.1633 +trainer/Log Pis Max 66.7631 +trainer/Log Pis Min -8.57367 +trainer/policy/mean Mean -0.0497995 +trainer/policy/mean Std 0.905036 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.87699 +trainer/policy/normal/std Std 0.534657 +trainer/policy/normal/std Max 5.07327 +trainer/policy/normal/std Min 0.504468 +trainer/policy/normal/log_std Mean 1.03387 +trainer/policy/normal/log_std Std 0.234352 +trainer/policy/normal/log_std Max 1.62399 +trainer/policy/normal/log_std Min -0.684251 +eval/num steps total 105893 +eval/num paths total 106 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.133478 +eval/Actions Std 0.8747 +eval/Actions Max 0.999993 +eval/Actions Min -0.999983 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56849 +time/logging (s) 0.00373835 +time/sampling batch (s) 0.529762 +time/saving (s) 0.00356073 +time/training (s) 7.3259 +time/epoch (s) 10.4315 +time/total (s) 1114.2 +Epoch -895 +---------------------------------- --------------- +2022-05-10 13:29:19.510094 PDT | [2] Epoch -894 finished +---------------------------------- --------------- +epoch -894 +replay_buffer/size 999033 +trainer/num train calls 107000 +trainer/Policy Loss -20.1051 +trainer/Log Pis Mean 24.6223 +trainer/Log Pis Std 13.8717 +trainer/Log Pis Max 69.9645 +trainer/Log Pis Min -8.35346 +trainer/policy/mean Mean -0.037294 +trainer/policy/mean Std 0.907981 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.85459 +trainer/policy/normal/std Std 0.535344 +trainer/policy/normal/std Max 4.85011 +trainer/policy/normal/std Min 0.492743 +trainer/policy/normal/log_std Mean 1.02502 +trainer/policy/normal/log_std Std 0.241694 +trainer/policy/normal/log_std Max 1.579 +trainer/policy/normal/log_std Min -0.707768 +eval/num steps total 106893 +eval/num paths total 107 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.149512 +eval/Actions Std 0.933143 +eval/Actions Max 0.999996 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64829 +time/logging (s) 0.00370746 +time/sampling batch (s) 0.279293 +time/saving (s) 0.00347805 +time/training (s) 7.17835 +time/epoch (s) 10.1131 +time/total (s) 1124.32 +Epoch -894 +---------------------------------- --------------- +2022-05-10 13:29:29.265067 PDT | [2] Epoch -893 finished +---------------------------------- -------------- +epoch -893 +replay_buffer/size 999033 +trainer/num train calls 108000 +trainer/Policy Loss -17.9595 +trainer/Log Pis Mean 24.2362 +trainer/Log Pis Std 13.7293 +trainer/Log Pis Max 74.0037 +trainer/Log Pis Min -10.8444 +trainer/policy/mean Mean -0.0418883 +trainer/policy/mean Std 0.902419 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.8551 +trainer/policy/normal/std Std 0.55928 +trainer/policy/normal/std Max 4.94425 +trainer/policy/normal/std Min 0.513703 +trainer/policy/normal/log_std Mean 1.02293 +trainer/policy/normal/log_std Std 0.252582 +trainer/policy/normal/log_std Max 1.59823 +trainer/policy/normal/log_std Min -0.666111 +eval/num steps total 107893 +eval/num paths total 108 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0875476 +eval/Actions Std 0.904585 +eval/Actions Max 0.999998 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55497 +time/logging (s) 0.0037146 +time/sampling batch (s) 0.280055 +time/saving (s) 0.0035323 +time/training (s) 6.89305 +time/epoch (s) 9.73532 +time/total (s) 1134.06 +Epoch -893 +---------------------------------- -------------- +2022-05-10 13:29:38.531917 PDT | [2] Epoch -892 finished +---------------------------------- --------------- +epoch -892 +replay_buffer/size 999033 +trainer/num train calls 109000 +trainer/Policy Loss -20.4097 +trainer/Log Pis Mean 24.5323 +trainer/Log Pis Std 13.0347 +trainer/Log Pis Max 69.2132 +trainer/Log Pis Min -6.77644 +trainer/policy/mean Mean -0.0224203 +trainer/policy/mean Std 0.907877 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.90408 +trainer/policy/normal/std Std 0.553918 +trainer/policy/normal/std Max 5.22263 +trainer/policy/normal/std Min 0.466045 +trainer/policy/normal/log_std Mean 1.04179 +trainer/policy/normal/log_std Std 0.242156 +trainer/policy/normal/log_std Max 1.653 +trainer/policy/normal/log_std Min -0.763473 +eval/num steps total 108893 +eval/num paths total 109 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0441417 +eval/Actions Std 0.898432 +eval/Actions Max 0.999999 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.43286 +time/logging (s) 0.00367953 +time/sampling batch (s) 0.283691 +time/saving (s) 0.00349717 +time/training (s) 6.52288 +time/epoch (s) 9.24661 +time/total (s) 1143.31 +Epoch -892 +---------------------------------- --------------- +2022-05-10 13:29:48.859048 PDT | [2] Epoch -891 finished +---------------------------------- --------------- +epoch -891 +replay_buffer/size 999033 +trainer/num train calls 110000 +trainer/Policy Loss -18.5558 +trainer/Log Pis Mean 24.8997 +trainer/Log Pis Std 13.5251 +trainer/Log Pis Max 74.9299 +trainer/Log Pis Min -5.10657 +trainer/policy/mean Mean -0.0493054 +trainer/policy/mean Std 0.905602 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.85272 +trainer/policy/normal/std Std 0.54493 +trainer/policy/normal/std Max 5.17369 +trainer/policy/normal/std Min 0.525597 +trainer/policy/normal/log_std Mean 1.02459 +trainer/policy/normal/log_std Std 0.236917 +trainer/policy/normal/log_std Max 1.64359 +trainer/policy/normal/log_std Min -0.64322 +eval/num steps total 109893 +eval/num paths total 110 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.151578 +eval/Actions Std 0.881197 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.75129 +time/logging (s) 0.00406388 +time/sampling batch (s) 0.290035 +time/saving (s) 0.00397442 +time/training (s) 7.25757 +time/epoch (s) 10.3069 +time/total (s) 1153.62 +Epoch -891 +---------------------------------- --------------- +2022-05-10 13:29:58.085218 PDT | [2] Epoch -890 finished +---------------------------------- --------------- +epoch -890 +replay_buffer/size 999033 +trainer/num train calls 111000 +trainer/Policy Loss -18.9172 +trainer/Log Pis Mean 25.0616 +trainer/Log Pis Std 12.8275 +trainer/Log Pis Max 63.7353 +trainer/Log Pis Min -4.74761 +trainer/policy/mean Mean -0.0116924 +trainer/policy/mean Std 0.907304 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.81705 +trainer/policy/normal/std Std 0.55053 +trainer/policy/normal/std Max 4.72875 +trainer/policy/normal/std Min 0.410905 +trainer/policy/normal/log_std Mean 1.01016 +trainer/policy/normal/log_std Std 0.248316 +trainer/policy/normal/log_std Max 1.55366 +trainer/policy/normal/log_std Min -0.889393 +eval/num steps total 110893 +eval/num paths total 111 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0848958 +eval/Actions Std 0.89272 +eval/Actions Max 0.999995 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56175 +time/logging (s) 0.0038051 +time/sampling batch (s) 0.284615 +time/saving (s) 0.00377442 +time/training (s) 6.35158 +time/epoch (s) 9.20552 +time/total (s) 1162.82 +Epoch -890 +---------------------------------- --------------- +2022-05-10 13:30:07.800780 PDT | [2] Epoch -889 finished +---------------------------------- --------------- +epoch -889 +replay_buffer/size 999033 +trainer/num train calls 112000 +trainer/Policy Loss -18.9305 +trainer/Log Pis Mean 24.4139 +trainer/Log Pis Std 12.7832 +trainer/Log Pis Max 69.4603 +trainer/Log Pis Min -7.9866 +trainer/policy/mean Mean -0.0257899 +trainer/policy/mean Std 0.909008 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.83524 +trainer/policy/normal/std Std 0.542085 +trainer/policy/normal/std Max 4.77263 +trainer/policy/normal/std Min 0.532333 +trainer/policy/normal/log_std Mean 1.018 +trainer/policy/normal/log_std Std 0.23985 +trainer/policy/normal/log_std Max 1.5629 +trainer/policy/normal/log_std Min -0.630486 +eval/num steps total 111893 +eval/num paths total 112 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.129434 +eval/Actions Std 0.889674 +eval/Actions Max 0.999961 +eval/Actions Min -0.999982 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67464 +time/logging (s) 0.00374625 +time/sampling batch (s) 0.284941 +time/saving (s) 0.00359534 +time/training (s) 6.72832 +time/epoch (s) 9.69524 +time/total (s) 1172.52 +Epoch -889 +---------------------------------- --------------- +2022-05-10 13:30:17.660322 PDT | [2] Epoch -888 finished +---------------------------------- --------------- +epoch -888 +replay_buffer/size 999033 +trainer/num train calls 113000 +trainer/Policy Loss -19.8885 +trainer/Log Pis Mean 23.8643 +trainer/Log Pis Std 13.6276 +trainer/Log Pis Max 63.9005 +trainer/Log Pis Min -9.78187 +trainer/policy/mean Mean -0.0320589 +trainer/policy/mean Std 0.905146 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.88053 +trainer/policy/normal/std Std 0.562915 +trainer/policy/normal/std Max 5.03048 +trainer/policy/normal/std Min 0.484176 +trainer/policy/normal/log_std Mean 1.03228 +trainer/policy/normal/log_std Std 0.249309 +trainer/policy/normal/log_std Max 1.61552 +trainer/policy/normal/log_std Min -0.725307 +eval/num steps total 112893 +eval/num paths total 113 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0772065 +eval/Actions Std 0.916087 +eval/Actions Max 0.999996 +eval/Actions Min -0.999983 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64943 +time/logging (s) 0.00371034 +time/sampling batch (s) 0.28203 +time/saving (s) 0.00343555 +time/training (s) 6.90092 +time/epoch (s) 9.83953 +time/total (s) 1182.36 +Epoch -888 +---------------------------------- --------------- +2022-05-10 13:30:28.062607 PDT | [2] Epoch -887 finished +---------------------------------- --------------- +epoch -887 +replay_buffer/size 999033 +trainer/num train calls 114000 +trainer/Policy Loss -19.2312 +trainer/Log Pis Mean 24.8339 +trainer/Log Pis Std 13.2998 +trainer/Log Pis Max 70.42 +trainer/Log Pis Min -6.54858 +trainer/policy/mean Mean -0.0284842 +trainer/policy/mean Std 0.904015 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999979 +trainer/policy/normal/std Mean 2.95212 +trainer/policy/normal/std Std 0.568192 +trainer/policy/normal/std Max 6.05167 +trainer/policy/normal/std Min 0.434787 +trainer/policy/normal/log_std Mean 1.057 +trainer/policy/normal/log_std Std 0.25048 +trainer/policy/normal/log_std Max 1.80033 +trainer/policy/normal/log_std Min -0.832899 +eval/num steps total 113893 +eval/num paths total 114 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.507062 +eval/Actions Std 0.737193 +eval/Actions Max 0.999963 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.35441 +time/logging (s) 0.00438533 +time/sampling batch (s) 0.27877 +time/saving (s) 0.00347172 +time/training (s) 7.74184 +time/epoch (s) 10.3829 +time/total (s) 1192.75 +Epoch -887 +---------------------------------- --------------- +2022-05-10 13:30:38.488713 PDT | [2] Epoch -886 finished +---------------------------------- --------------- +epoch -886 +replay_buffer/size 999033 +trainer/num train calls 115000 +trainer/Policy Loss -19.3819 +trainer/Log Pis Mean 25.0844 +trainer/Log Pis Std 13.585 +trainer/Log Pis Max 76.4973 +trainer/Log Pis Min -9.30738 +trainer/policy/mean Mean -0.0360796 +trainer/policy/mean Std 0.912549 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.88135 +trainer/policy/normal/std Std 0.576987 +trainer/policy/normal/std Max 5.22861 +trainer/policy/normal/std Min 0.551134 +trainer/policy/normal/log_std Mean 1.03138 +trainer/policy/normal/log_std Std 0.25521 +trainer/policy/normal/log_std Max 1.65415 +trainer/policy/normal/log_std Min -0.595777 +eval/num steps total 114893 +eval/num paths total 115 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0149541 +eval/Actions Std 0.930693 +eval/Actions Max 0.999991 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69982 +time/logging (s) 0.0037253 +time/sampling batch (s) 0.279696 +time/saving (s) 0.00344839 +time/training (s) 7.4186 +time/epoch (s) 10.4053 +time/total (s) 1203.16 +Epoch -886 +---------------------------------- --------------- +2022-05-10 13:30:48.897070 PDT | [2] Epoch -885 finished +---------------------------------- --------------- +epoch -885 +replay_buffer/size 999033 +trainer/num train calls 116000 +trainer/Policy Loss -18.9489 +trainer/Log Pis Mean 24.8729 +trainer/Log Pis Std 14.0266 +trainer/Log Pis Max 70.5278 +trainer/Log Pis Min -11.2686 +trainer/policy/mean Mean -0.0446437 +trainer/policy/mean Std 0.906003 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.91876 +trainer/policy/normal/std Std 0.567216 +trainer/policy/normal/std Max 5.92738 +trainer/policy/normal/std Min 0.53809 +trainer/policy/normal/log_std Mean 1.04622 +trainer/policy/normal/log_std Std 0.24432 +trainer/policy/normal/log_std Max 1.77958 +trainer/policy/normal/log_std Min -0.619729 +eval/num steps total 115893 +eval/num paths total 116 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.254528 +eval/Actions Std 0.8826 +eval/Actions Max 0.999989 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.47521 +time/logging (s) 0.00410934 +time/sampling batch (s) 0.285598 +time/saving (s) 0.00419462 +time/training (s) 7.61911 +time/epoch (s) 10.3882 +time/total (s) 1213.55 +Epoch -885 +---------------------------------- --------------- +2022-05-10 13:30:58.795128 PDT | [2] Epoch -884 finished +---------------------------------- --------------- +epoch -884 +replay_buffer/size 999033 +trainer/num train calls 117000 +trainer/Policy Loss -19.0687 +trainer/Log Pis Mean 24.8609 +trainer/Log Pis Std 13.2107 +trainer/Log Pis Max 74.4981 +trainer/Log Pis Min -5.31574 +trainer/policy/mean Mean -0.0209485 +trainer/policy/mean Std 0.901681 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.84324 +trainer/policy/normal/std Std 0.560141 +trainer/policy/normal/std Max 5.99758 +trainer/policy/normal/std Min 0.448879 +trainer/policy/normal/log_std Mean 1.01869 +trainer/policy/normal/log_std Std 0.252522 +trainer/policy/normal/log_std Max 1.79136 +trainer/policy/normal/log_std Min -0.801001 +eval/num steps total 116893 +eval/num paths total 117 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0274664 +eval/Actions Std 0.894294 +eval/Actions Max 0.999995 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56972 +time/logging (s) 0.00385588 +time/sampling batch (s) 0.528554 +time/saving (s) 0.00345978 +time/training (s) 6.77187 +time/epoch (s) 9.87746 +time/total (s) 1223.43 +Epoch -884 +---------------------------------- --------------- +2022-05-10 13:31:08.394170 PDT | [2] Epoch -883 finished +---------------------------------- --------------- +epoch -883 +replay_buffer/size 999033 +trainer/num train calls 118000 +trainer/Policy Loss -19.4517 +trainer/Log Pis Mean 22.9226 +trainer/Log Pis Std 13.1752 +trainer/Log Pis Max 71.2283 +trainer/Log Pis Min -7.61884 +trainer/policy/mean Mean -0.032938 +trainer/policy/mean Std 0.906913 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.79507 +trainer/policy/normal/std Std 0.567985 +trainer/policy/normal/std Max 4.73221 +trainer/policy/normal/std Min 0.490655 +trainer/policy/normal/log_std Mean 0.999159 +trainer/policy/normal/log_std Std 0.265441 +trainer/policy/normal/log_std Max 1.55439 +trainer/policy/normal/log_std Min -0.712013 +eval/num steps total 117893 +eval/num paths total 118 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.132554 +eval/Actions Std 0.911785 +eval/Actions Max 1 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.36974 +time/logging (s) 0.00372096 +time/sampling batch (s) 0.285702 +time/saving (s) 0.00349544 +time/training (s) 6.91633 +time/epoch (s) 9.57899 +time/total (s) 1233.01 +Epoch -883 +---------------------------------- --------------- +2022-05-10 13:31:18.776007 PDT | [2] Epoch -882 finished +---------------------------------- --------------- +epoch -882 +replay_buffer/size 999033 +trainer/num train calls 119000 +trainer/Policy Loss -19.2626 +trainer/Log Pis Mean 24.1615 +trainer/Log Pis Std 13.3072 +trainer/Log Pis Max 64.2619 +trainer/Log Pis Min -6.74822 +trainer/policy/mean Mean -0.0198595 +trainer/policy/mean Std 0.907225 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.87831 +trainer/policy/normal/std Std 0.562087 +trainer/policy/normal/std Max 4.71156 +trainer/policy/normal/std Min 0.504549 +trainer/policy/normal/log_std Mean 1.03178 +trainer/policy/normal/log_std Std 0.247319 +trainer/policy/normal/log_std Max 1.55002 +trainer/policy/normal/log_std Min -0.684091 +eval/num steps total 118893 +eval/num paths total 119 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.030974 +eval/Actions Std 0.896164 +eval/Actions Max 0.999995 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60842 +time/logging (s) 0.0038622 +time/sampling batch (s) 0.319298 +time/saving (s) 0.00371948 +time/training (s) 7.42614 +time/epoch (s) 10.3614 +time/total (s) 1243.38 +Epoch -882 +---------------------------------- --------------- +2022-05-10 13:31:29.401291 PDT | [2] Epoch -881 finished +---------------------------------- --------------- +epoch -881 +replay_buffer/size 999033 +trainer/num train calls 120000 +trainer/Policy Loss -19.6299 +trainer/Log Pis Mean 23.469 +trainer/Log Pis Std 13.2228 +trainer/Log Pis Max 64.7427 +trainer/Log Pis Min -7.32025 +trainer/policy/mean Mean -0.0410955 +trainer/policy/mean Std 0.906109 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.84867 +trainer/policy/normal/std Std 0.544322 +trainer/policy/normal/std Max 4.97696 +trainer/policy/normal/std Min 0.501098 +trainer/policy/normal/log_std Mean 1.02214 +trainer/policy/normal/log_std Std 0.244603 +trainer/policy/normal/log_std Max 1.60482 +trainer/policy/normal/log_std Min -0.690954 +eval/num steps total 119893 +eval/num paths total 120 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.122445 +eval/Actions Std 0.918076 +eval/Actions Max 0.999981 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51317 +time/logging (s) 0.00379596 +time/sampling batch (s) 0.334725 +time/saving (s) 0.0037706 +time/training (s) 7.74895 +time/epoch (s) 10.6044 +time/total (s) 1253.99 +Epoch -881 +---------------------------------- --------------- +2022-05-10 13:31:40.896949 PDT | [2] Epoch -880 finished +---------------------------------- --------------- +epoch -880 +replay_buffer/size 999033 +trainer/num train calls 121000 +trainer/Policy Loss -20.237 +trainer/Log Pis Mean 24.3695 +trainer/Log Pis Std 12.9627 +trainer/Log Pis Max 61.1739 +trainer/Log Pis Min -9.51186 +trainer/policy/mean Mean -0.0327403 +trainer/policy/mean Std 0.906018 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.86133 +trainer/policy/normal/std Std 0.565505 +trainer/policy/normal/std Max 4.97947 +trainer/policy/normal/std Min 0.463215 +trainer/policy/normal/log_std Mean 1.02537 +trainer/policy/normal/log_std Std 0.249269 +trainer/policy/normal/log_std Max 1.60532 +trainer/policy/normal/log_std Min -0.769565 +eval/num steps total 120893 +eval/num paths total 121 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0399647 +eval/Actions Std 0.922299 +eval/Actions Max 0.999996 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.76761 +time/logging (s) 0.0041669 +time/sampling batch (s) 0.837225 +time/saving (s) 0.00436361 +time/training (s) 7.8618 +time/epoch (s) 11.4752 +time/total (s) 1265.46 +Epoch -880 +---------------------------------- --------------- +2022-05-10 13:31:52.463313 PDT | [2] Epoch -879 finished +---------------------------------- --------------- +epoch -879 +replay_buffer/size 999033 +trainer/num train calls 122000 +trainer/Policy Loss -19.2411 +trainer/Log Pis Mean 23.9581 +trainer/Log Pis Std 13.2642 +trainer/Log Pis Max 65.1456 +trainer/Log Pis Min -5.66917 +trainer/policy/mean Mean -0.0390029 +trainer/policy/mean Std 0.901062 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.90182 +trainer/policy/normal/std Std 0.557238 +trainer/policy/normal/std Max 5.49782 +trainer/policy/normal/std Min 0.467069 +trainer/policy/normal/log_std Mean 1.04063 +trainer/policy/normal/log_std Std 0.244857 +trainer/policy/normal/log_std Max 1.70435 +trainer/policy/normal/log_std Min -0.761278 +eval/num steps total 121893 +eval/num paths total 122 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.154441 +eval/Actions Std 0.931711 +eval/Actions Max 0.999994 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.89557 +time/logging (s) 0.00380797 +time/sampling batch (s) 0.340096 +time/saving (s) 0.00366783 +time/training (s) 8.30141 +time/epoch (s) 11.5446 +time/total (s) 1277.01 +Epoch -879 +---------------------------------- --------------- +2022-05-10 13:32:02.923093 PDT | [2] Epoch -878 finished +---------------------------------- --------------- +epoch -878 +replay_buffer/size 999033 +trainer/num train calls 123000 +trainer/Policy Loss -18.8414 +trainer/Log Pis Mean 25.9853 +trainer/Log Pis Std 14.0798 +trainer/Log Pis Max 76.7439 +trainer/Log Pis Min -5.63786 +trainer/policy/mean Mean -0.0485181 +trainer/policy/mean Std 0.907943 +trainer/policy/mean Max 0.999977 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.91832 +trainer/policy/normal/std Std 0.582832 +trainer/policy/normal/std Max 5.90769 +trainer/policy/normal/std Min 0.49433 +trainer/policy/normal/log_std Mean 1.04489 +trainer/policy/normal/log_std Std 0.249758 +trainer/policy/normal/log_std Max 1.77626 +trainer/policy/normal/log_std Min -0.704552 +eval/num steps total 122893 +eval/num paths total 123 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00456783 +eval/Actions Std 0.963213 +eval/Actions Max 0.999992 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.7793 +time/logging (s) 0.00375366 +time/sampling batch (s) 0.575734 +time/saving (s) 0.0035941 +time/training (s) 7.0765 +time/epoch (s) 10.4389 +time/total (s) 1287.45 +Epoch -878 +---------------------------------- --------------- +2022-05-10 13:32:12.762440 PDT | [2] Epoch -877 finished +---------------------------------- --------------- +epoch -877 +replay_buffer/size 999033 +trainer/num train calls 124000 +trainer/Policy Loss -19.2006 +trainer/Log Pis Mean 24.1324 +trainer/Log Pis Std 12.4573 +trainer/Log Pis Max 62.5066 +trainer/Log Pis Min -6.00569 +trainer/policy/mean Mean -0.0454539 +trainer/policy/mean Std 0.90948 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.80084 +trainer/policy/normal/std Std 0.560323 +trainer/policy/normal/std Max 6.69071 +trainer/policy/normal/std Min 0.495003 +trainer/policy/normal/log_std Mean 1.00333 +trainer/policy/normal/log_std Std 0.252521 +trainer/policy/normal/log_std Max 1.90072 +trainer/policy/normal/log_std Min -0.703192 +eval/num steps total 123893 +eval/num paths total 124 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0630918 +eval/Actions Std 0.913595 +eval/Actions Max 0.999997 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.77239 +time/logging (s) 0.00376038 +time/sampling batch (s) 0.281028 +time/saving (s) 0.00347924 +time/training (s) 6.75855 +time/epoch (s) 9.81921 +time/total (s) 1297.28 +Epoch -877 +---------------------------------- --------------- +2022-05-10 13:32:23.063186 PDT | [2] Epoch -876 finished +---------------------------------- --------------- +epoch -876 +replay_buffer/size 999033 +trainer/num train calls 125000 +trainer/Policy Loss -19.7936 +trainer/Log Pis Mean 23.6456 +trainer/Log Pis Std 13.3805 +trainer/Log Pis Max 62.6081 +trainer/Log Pis Min -8.32986 +trainer/policy/mean Mean -0.0135342 +trainer/policy/mean Std 0.905383 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.90038 +trainer/policy/normal/std Std 0.586352 +trainer/policy/normal/std Max 5.4679 +trainer/policy/normal/std Min 0.523744 +trainer/policy/normal/log_std Mean 1.03733 +trainer/policy/normal/log_std Std 0.257923 +trainer/policy/normal/log_std Max 1.69889 +trainer/policy/normal/log_std Min -0.646751 +eval/num steps total 124893 +eval/num paths total 125 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.104092 +eval/Actions Std 0.919361 +eval/Actions Max 0.999996 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59947 +time/logging (s) 0.0038111 +time/sampling batch (s) 0.279044 +time/saving (s) 0.00349223 +time/training (s) 7.39517 +time/epoch (s) 10.281 +time/total (s) 1307.56 +Epoch -876 +---------------------------------- --------------- +2022-05-10 13:32:33.854239 PDT | [2] Epoch -875 finished +---------------------------------- --------------- +epoch -875 +replay_buffer/size 999033 +trainer/num train calls 126000 +trainer/Policy Loss -19.2288 +trainer/Log Pis Mean 23.8893 +trainer/Log Pis Std 13.407 +trainer/Log Pis Max 71.3615 +trainer/Log Pis Min -6.35386 +trainer/policy/mean Mean -0.0366868 +trainer/policy/mean Std 0.906764 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.8621 +trainer/policy/normal/std Std 0.567887 +trainer/policy/normal/std Max 5.45174 +trainer/policy/normal/std Min 0.40395 +trainer/policy/normal/log_std Mean 1.02488 +trainer/policy/normal/log_std Std 0.254881 +trainer/policy/normal/log_std Max 1.69593 +trainer/policy/normal/log_std Min -0.906465 +eval/num steps total 125893 +eval/num paths total 126 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.100795 +eval/Actions Std 0.903027 +eval/Actions Max 0.999995 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59287 +time/logging (s) 0.00416776 +time/sampling batch (s) 0.532153 +time/saving (s) 0.00408331 +time/training (s) 7.63805 +time/epoch (s) 10.7713 +time/total (s) 1318.33 +Epoch -875 +---------------------------------- --------------- +2022-05-10 13:32:44.202950 PDT | [2] Epoch -874 finished +---------------------------------- --------------- +epoch -874 +replay_buffer/size 999033 +trainer/num train calls 127000 +trainer/Policy Loss -18.2199 +trainer/Log Pis Mean 24.5347 +trainer/Log Pis Std 12.8607 +trainer/Log Pis Max 64.825 +trainer/Log Pis Min -11.2913 +trainer/policy/mean Mean -0.0137487 +trainer/policy/mean Std 0.901925 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.8659 +trainer/policy/normal/std Std 0.580263 +trainer/policy/normal/std Max 7.1027 +trainer/policy/normal/std Min 0.487259 +trainer/policy/normal/log_std Mean 1.02591 +trainer/policy/normal/log_std Std 0.253968 +trainer/policy/normal/log_std Max 1.96047 +trainer/policy/normal/log_std Min -0.71896 +eval/num steps total 126893 +eval/num paths total 127 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0658453 +eval/Actions Std 0.891291 +eval/Actions Max 0.999999 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64339 +time/logging (s) 0.00373037 +time/sampling batch (s) 0.286148 +time/saving (s) 0.0035874 +time/training (s) 7.39088 +time/epoch (s) 10.3277 +time/total (s) 1328.67 +Epoch -874 +---------------------------------- --------------- +2022-05-10 13:32:54.207018 PDT | [2] Epoch -873 finished +---------------------------------- --------------- +epoch -873 +replay_buffer/size 999033 +trainer/num train calls 128000 +trainer/Policy Loss -20.128 +trainer/Log Pis Mean 24.0277 +trainer/Log Pis Std 12.7023 +trainer/Log Pis Max 69.2668 +trainer/Log Pis Min -5.2002 +trainer/policy/mean Mean -0.0404789 +trainer/policy/mean Std 0.909699 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999977 +trainer/policy/normal/std Mean 2.85743 +trainer/policy/normal/std Std 0.557004 +trainer/policy/normal/std Max 5.66999 +trainer/policy/normal/std Min 0.50532 +trainer/policy/normal/log_std Mean 1.02402 +trainer/policy/normal/log_std Std 0.251847 +trainer/policy/normal/log_std Max 1.73519 +trainer/policy/normal/log_std Min -0.682563 +eval/num steps total 127893 +eval/num paths total 128 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.089743 +eval/Actions Std 0.917552 +eval/Actions Max 0.999994 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.49009 +time/logging (s) 0.00387236 +time/sampling batch (s) 0.533176 +time/saving (s) 0.00346907 +time/training (s) 6.95337 +time/epoch (s) 9.98398 +time/total (s) 1338.65 +Epoch -873 +---------------------------------- --------------- +2022-05-10 13:33:04.238540 PDT | [2] Epoch -872 finished +---------------------------------- ---------------- +epoch -872 +replay_buffer/size 999033 +trainer/num train calls 129000 +trainer/Policy Loss -20.7951 +trainer/Log Pis Mean 24.5831 +trainer/Log Pis Std 14.8887 +trainer/Log Pis Max 78.5764 +trainer/Log Pis Min -8.50285 +trainer/policy/mean Mean -0.0233597 +trainer/policy/mean Std 0.902669 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.88757 +trainer/policy/normal/std Std 0.56257 +trainer/policy/normal/std Max 4.83463 +trainer/policy/normal/std Min 0.481781 +trainer/policy/normal/log_std Mean 1.03401 +trainer/policy/normal/log_std Std 0.255593 +trainer/policy/normal/log_std Max 1.5758 +trainer/policy/normal/log_std Min -0.730266 +eval/num steps total 128893 +eval/num paths total 129 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.000150936 +eval/Actions Std 0.890021 +eval/Actions Max 0.999997 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.47675 +time/logging (s) 0.00368243 +time/sampling batch (s) 0.283193 +time/saving (s) 0.00347876 +time/training (s) 7.2439 +time/epoch (s) 10.011 +time/total (s) 1348.67 +Epoch -872 +---------------------------------- ---------------- +2022-05-10 13:33:14.248050 PDT | [2] Epoch -871 finished +---------------------------------- --------------- +epoch -871 +replay_buffer/size 999033 +trainer/num train calls 130000 +trainer/Policy Loss -19.236 +trainer/Log Pis Mean 24.3389 +trainer/Log Pis Std 13.1144 +trainer/Log Pis Max 59.9138 +trainer/Log Pis Min -4.52846 +trainer/policy/mean Mean -0.0429525 +trainer/policy/mean Std 0.907866 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.81515 +trainer/policy/normal/std Std 0.557025 +trainer/policy/normal/std Max 5.1967 +trainer/policy/normal/std Min 0.410741 +trainer/policy/normal/log_std Mean 1.00882 +trainer/policy/normal/log_std Std 0.25145 +trainer/policy/normal/log_std Max 1.64802 +trainer/policy/normal/log_std Min -0.889793 +eval/num steps total 129893 +eval/num paths total 130 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.123459 +eval/Actions Std 0.911036 +eval/Actions Max 0.99999 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73883 +time/logging (s) 0.00373694 +time/sampling batch (s) 0.535482 +time/saving (s) 0.00348073 +time/training (s) 6.70784 +time/epoch (s) 9.98936 +time/total (s) 1358.66 +Epoch -871 +---------------------------------- --------------- +2022-05-10 13:33:24.972600 PDT | [2] Epoch -870 finished +---------------------------------- --------------- +epoch -870 +replay_buffer/size 999033 +trainer/num train calls 131000 +trainer/Policy Loss -19.6096 +trainer/Log Pis Mean 24.2247 +trainer/Log Pis Std 13.7723 +trainer/Log Pis Max 72.5463 +trainer/Log Pis Min -7.9799 +trainer/policy/mean Mean -0.040402 +trainer/policy/mean Std 0.906455 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.89867 +trainer/policy/normal/std Std 0.579323 +trainer/policy/normal/std Max 6.33335 +trainer/policy/normal/std Min 0.434333 +trainer/policy/normal/log_std Mean 1.03682 +trainer/policy/normal/log_std Std 0.259882 +trainer/policy/normal/log_std Max 1.84583 +trainer/policy/normal/log_std Min -0.833944 +eval/num steps total 130893 +eval/num paths total 131 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.332622 +eval/Actions Std 0.833732 +eval/Actions Max 0.99999 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.79888 +time/logging (s) 0.00419317 +time/sampling batch (s) 0.783996 +time/saving (s) 0.00415004 +time/training (s) 7.1134 +time/epoch (s) 10.7046 +time/total (s) 1369.37 +Epoch -870 +---------------------------------- --------------- +2022-05-10 13:33:35.462298 PDT | [2] Epoch -869 finished +---------------------------------- --------------- +epoch -869 +replay_buffer/size 999033 +trainer/num train calls 132000 +trainer/Policy Loss -19.8449 +trainer/Log Pis Mean 24.6073 +trainer/Log Pis Std 12.787 +trainer/Log Pis Max 66.0008 +trainer/Log Pis Min -9.67186 +trainer/policy/mean Mean -0.0275919 +trainer/policy/mean Std 0.904786 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.90736 +trainer/policy/normal/std Std 0.561359 +trainer/policy/normal/std Max 7.14464 +trainer/policy/normal/std Min 0.517021 +trainer/policy/normal/log_std Mean 1.04244 +trainer/policy/normal/log_std Std 0.244757 +trainer/policy/normal/log_std Max 1.96636 +trainer/policy/normal/log_std Min -0.659671 +eval/num steps total 131893 +eval/num paths total 132 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.206159 +eval/Actions Std 0.919042 +eval/Actions Max 1 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.19438 +time/logging (s) 0.0037336 +time/sampling batch (s) 0.28153 +time/saving (s) 0.00363416 +time/training (s) 7.98572 +time/epoch (s) 10.469 +time/total (s) 1379.84 +Epoch -869 +---------------------------------- --------------- +2022-05-10 13:33:45.357732 PDT | [2] Epoch -868 finished +---------------------------------- --------------- +epoch -868 +replay_buffer/size 999033 +trainer/num train calls 133000 +trainer/Policy Loss -18.9242 +trainer/Log Pis Mean 24.72 +trainer/Log Pis Std 12.8861 +trainer/Log Pis Max 59.5602 +trainer/Log Pis Min -7.70233 +trainer/policy/mean Mean -0.0236349 +trainer/policy/mean Std 0.901161 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.87337 +trainer/policy/normal/std Std 0.591843 +trainer/policy/normal/std Max 6.5172 +trainer/policy/normal/std Min 0.458864 +trainer/policy/normal/log_std Mean 1.02681 +trainer/policy/normal/log_std Std 0.264272 +trainer/policy/normal/log_std Max 1.87444 +trainer/policy/normal/log_std Min -0.779001 +eval/num steps total 132893 +eval/num paths total 133 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.241177 +eval/Actions Std 0.865132 +eval/Actions Max 0.999995 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68066 +time/logging (s) 0.00372566 +time/sampling batch (s) 0.281283 +time/saving (s) 0.00341518 +time/training (s) 6.90639 +time/epoch (s) 9.87547 +time/total (s) 1389.72 +Epoch -868 +---------------------------------- --------------- +2022-05-10 13:33:55.588628 PDT | [2] Epoch -867 finished +---------------------------------- --------------- +epoch -867 +replay_buffer/size 999033 +trainer/num train calls 134000 +trainer/Policy Loss -20.5992 +trainer/Log Pis Mean 24.2083 +trainer/Log Pis Std 12.8971 +trainer/Log Pis Max 72.3464 +trainer/Log Pis Min -7.05765 +trainer/policy/mean Mean -0.0144982 +trainer/policy/mean Std 0.907497 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.88309 +trainer/policy/normal/std Std 0.579451 +trainer/policy/normal/std Max 5.15091 +trainer/policy/normal/std Min 0.472497 +trainer/policy/normal/log_std Mean 1.03128 +trainer/policy/normal/log_std Std 0.25957 +trainer/policy/normal/log_std Max 1.63917 +trainer/policy/normal/log_std Min -0.749725 +eval/num steps total 133893 +eval/num paths total 134 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0373522 +eval/Actions Std 0.893359 +eval/Actions Max 0.999998 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5713 +time/logging (s) 0.00377128 +time/sampling batch (s) 0.278646 +time/saving (s) 0.00346494 +time/training (s) 7.35399 +time/epoch (s) 10.2112 +time/total (s) 1399.93 +Epoch -867 +---------------------------------- --------------- +2022-05-10 13:34:05.764857 PDT | [2] Epoch -866 finished +---------------------------------- --------------- +epoch -866 +replay_buffer/size 999033 +trainer/num train calls 135000 +trainer/Policy Loss -19.3541 +trainer/Log Pis Mean 24.6798 +trainer/Log Pis Std 13.2982 +trainer/Log Pis Max 69.5722 +trainer/Log Pis Min -10.4 +trainer/policy/mean Mean -0.0425884 +trainer/policy/mean Std 0.910252 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.85636 +trainer/policy/normal/std Std 0.583392 +trainer/policy/normal/std Max 5.40861 +trainer/policy/normal/std Min 0.403255 +trainer/policy/normal/log_std Mean 1.02038 +trainer/policy/normal/log_std Std 0.269054 +trainer/policy/normal/log_std Max 1.68799 +trainer/policy/normal/log_std Min -0.908187 +eval/num steps total 134893 +eval/num paths total 135 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0967509 +eval/Actions Std 0.893079 +eval/Actions Max 0.999997 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67324 +time/logging (s) 0.0037348 +time/sampling batch (s) 0.283474 +time/saving (s) 0.00343931 +time/training (s) 7.192 +time/epoch (s) 10.1559 +time/total (s) 1410.09 +Epoch -866 +---------------------------------- --------------- +2022-05-10 13:34:16.076728 PDT | [2] Epoch -865 finished +---------------------------------- --------------- +epoch -865 +replay_buffer/size 999033 +trainer/num train calls 136000 +trainer/Policy Loss -19.2748 +trainer/Log Pis Mean 24.2908 +trainer/Log Pis Std 12.9856 +trainer/Log Pis Max 66.2237 +trainer/Log Pis Min -2.76932 +trainer/policy/mean Mean -0.0495878 +trainer/policy/mean Std 0.907394 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.83014 +trainer/policy/normal/std Std 0.573624 +trainer/policy/normal/std Max 5.35229 +trainer/policy/normal/std Min 0.41637 +trainer/policy/normal/log_std Mean 1.01075 +trainer/policy/normal/log_std Std 0.273198 +trainer/policy/normal/log_std Max 1.67752 +trainer/policy/normal/log_std Min -0.876182 +eval/num steps total 135893 +eval/num paths total 136 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.48288 +eval/Actions Std 0.805764 +eval/Actions Max 0.999993 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51799 +time/logging (s) 0.00410366 +time/sampling batch (s) 0.533246 +time/saving (s) 0.00403819 +time/training (s) 7.23267 +time/epoch (s) 10.292 +time/total (s) 1420.38 +Epoch -865 +---------------------------------- --------------- +2022-05-10 13:34:25.720645 PDT | [2] Epoch -864 finished +---------------------------------- --------------- +epoch -864 +replay_buffer/size 999033 +trainer/num train calls 137000 +trainer/Policy Loss -19.4536 +trainer/Log Pis Mean 23.9382 +trainer/Log Pis Std 13.1169 +trainer/Log Pis Max 63.5634 +trainer/Log Pis Min -5.9761 +trainer/policy/mean Mean -0.0448559 +trainer/policy/mean Std 0.898895 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.8684 +trainer/policy/normal/std Std 0.605082 +trainer/policy/normal/std Max 5.08334 +trainer/policy/normal/std Min 0.415403 +trainer/policy/normal/log_std Mean 1.0219 +trainer/policy/normal/log_std Std 0.282309 +trainer/policy/normal/log_std Max 1.62597 +trainer/policy/normal/log_std Min -0.878507 +eval/num steps total 136893 +eval/num paths total 137 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.133596 +eval/Actions Std 0.935141 +eval/Actions Max 0.999998 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56928 +time/logging (s) 0.0038816 +time/sampling batch (s) 0.283296 +time/saving (s) 0.00389457 +time/training (s) 6.76264 +time/epoch (s) 9.623 +time/total (s) 1430.01 +Epoch -864 +---------------------------------- --------------- +2022-05-10 13:34:35.147801 PDT | [2] Epoch -863 finished +---------------------------------- --------------- +epoch -863 +replay_buffer/size 999033 +trainer/num train calls 138000 +trainer/Policy Loss -19.8266 +trainer/Log Pis Mean 24.6087 +trainer/Log Pis Std 13.102 +trainer/Log Pis Max 71.7904 +trainer/Log Pis Min -8.51662 +trainer/policy/mean Mean -0.0314412 +trainer/policy/mean Std 0.907467 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.86714 +trainer/policy/normal/std Std 0.567572 +trainer/policy/normal/std Max 5.16828 +trainer/policy/normal/std Min 0.478347 +trainer/policy/normal/log_std Mean 1.02711 +trainer/policy/normal/log_std Std 0.251329 +trainer/policy/normal/log_std Max 1.64254 +trainer/policy/normal/log_std Min -0.73742 +eval/num steps total 137893 +eval/num paths total 138 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.106133 +eval/Actions Std 0.917546 +eval/Actions Max 0.999994 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.388 +time/logging (s) 0.00380514 +time/sampling batch (s) 0.289617 +time/saving (s) 0.0035024 +time/training (s) 6.72144 +time/epoch (s) 9.40636 +time/total (s) 1439.42 +Epoch -863 +---------------------------------- --------------- +2022-05-10 13:34:44.584598 PDT | [2] Epoch -862 finished +---------------------------------- --------------- +epoch -862 +replay_buffer/size 999033 +trainer/num train calls 139000 +trainer/Policy Loss -19.0231 +trainer/Log Pis Mean 24.3579 +trainer/Log Pis Std 12.7908 +trainer/Log Pis Max 62.7861 +trainer/Log Pis Min -8.39574 +trainer/policy/mean Mean -0.0254702 +trainer/policy/mean Std 0.903445 +trainer/policy/mean Max 0.999982 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.85625 +trainer/policy/normal/std Std 0.575541 +trainer/policy/normal/std Max 5.09295 +trainer/policy/normal/std Min 0.43927 +trainer/policy/normal/log_std Mean 1.02163 +trainer/policy/normal/log_std Std 0.261628 +trainer/policy/normal/log_std Max 1.62786 +trainer/policy/normal/log_std Min -0.822641 +eval/num steps total 138893 +eval/num paths total 139 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0715725 +eval/Actions Std 0.922223 +eval/Actions Max 0.999992 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66418 +time/logging (s) 0.00490929 +time/sampling batch (s) 0.284227 +time/saving (s) 0.00453854 +time/training (s) 6.45971 +time/epoch (s) 9.41757 +time/total (s) 1448.84 +Epoch -862 +---------------------------------- --------------- +2022-05-10 13:34:54.832402 PDT | [2] Epoch -861 finished +---------------------------------- --------------- +epoch -861 +replay_buffer/size 999033 +trainer/num train calls 140000 +trainer/Policy Loss -20.5358 +trainer/Log Pis Mean 25.8212 +trainer/Log Pis Std 13.2372 +trainer/Log Pis Max 79.6188 +trainer/Log Pis Min -6.46659 +trainer/policy/mean Mean -0.0218067 +trainer/policy/mean Std 0.908383 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.83479 +trainer/policy/normal/std Std 0.573255 +trainer/policy/normal/std Max 5.19303 +trainer/policy/normal/std Min 0.467353 +trainer/policy/normal/log_std Mean 1.01475 +trainer/policy/normal/log_std Std 0.25667 +trainer/policy/normal/log_std Max 1.64732 +trainer/policy/normal/log_std Min -0.760671 +eval/num steps total 139893 +eval/num paths total 140 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.186427 +eval/Actions Std 0.916296 +eval/Actions Max 0.999987 +eval/Actions Min -0.999975 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67553 +time/logging (s) 0.0040647 +time/sampling batch (s) 0.279742 +time/saving (s) 0.00404869 +time/training (s) 7.26252 +time/epoch (s) 10.2259 +time/total (s) 1459.07 +Epoch -861 +---------------------------------- --------------- +2022-05-10 13:35:05.457082 PDT | [2] Epoch -860 finished +---------------------------------- --------------- +epoch -860 +replay_buffer/size 999033 +trainer/num train calls 141000 +trainer/Policy Loss -20.0358 +trainer/Log Pis Mean 25.7556 +trainer/Log Pis Std 13.4539 +trainer/Log Pis Max 72.5729 +trainer/Log Pis Min -5.18541 +trainer/policy/mean Mean -0.0261757 +trainer/policy/mean Std 0.911064 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.93435 +trainer/policy/normal/std Std 0.579427 +trainer/policy/normal/std Max 5.72394 +trainer/policy/normal/std Min 0.506326 +trainer/policy/normal/log_std Mean 1.05004 +trainer/policy/normal/log_std Std 0.254486 +trainer/policy/normal/log_std Max 1.74466 +trainer/policy/normal/log_std Min -0.680575 +eval/num steps total 140893 +eval/num paths total 141 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.199251 +eval/Actions Std 0.948149 +eval/Actions Max 0.999981 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65233 +time/logging (s) 0.00463651 +time/sampling batch (s) 0.540645 +time/saving (s) 0.00439944 +time/training (s) 7.40219 +time/epoch (s) 10.6042 +time/total (s) 1469.68 +Epoch -860 +---------------------------------- --------------- +2022-05-10 13:35:15.043431 PDT | [2] Epoch -859 finished +---------------------------------- --------------- +epoch -859 +replay_buffer/size 999033 +trainer/num train calls 142000 +trainer/Policy Loss -18.9805 +trainer/Log Pis Mean 24.0402 +trainer/Log Pis Std 13.1562 +trainer/Log Pis Max 67.7298 +trainer/Log Pis Min -9.86331 +trainer/policy/mean Mean -0.0614275 +trainer/policy/mean Std 0.901116 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.90198 +trainer/policy/normal/std Std 0.573865 +trainer/policy/normal/std Max 5.05898 +trainer/policy/normal/std Min 0.492377 +trainer/policy/normal/log_std Mean 1.03942 +trainer/policy/normal/log_std Std 0.249872 +trainer/policy/normal/log_std Max 1.62117 +trainer/policy/normal/log_std Min -0.708511 +eval/num steps total 141893 +eval/num paths total 142 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.173636 +eval/Actions Std 0.87159 +eval/Actions Max 0.999992 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.71802 +time/logging (s) 0.00377123 +time/sampling batch (s) 0.290184 +time/saving (s) 0.00361614 +time/training (s) 6.54843 +time/epoch (s) 9.56403 +time/total (s) 1479.25 +Epoch -859 +---------------------------------- --------------- +2022-05-10 13:35:24.618616 PDT | [2] Epoch -858 finished +---------------------------------- --------------- +epoch -858 +replay_buffer/size 999033 +trainer/num train calls 143000 +trainer/Policy Loss -19.4071 +trainer/Log Pis Mean 24.7396 +trainer/Log Pis Std 13.2967 +trainer/Log Pis Max 62.8785 +trainer/Log Pis Min -9.50622 +trainer/policy/mean Mean -0.0316068 +trainer/policy/mean Std 0.904406 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.8833 +trainer/policy/normal/std Std 0.580611 +trainer/policy/normal/std Max 4.81164 +trainer/policy/normal/std Min 0.485583 +trainer/policy/normal/log_std Mean 1.03091 +trainer/policy/normal/log_std Std 0.262317 +trainer/policy/normal/log_std Max 1.57104 +trainer/policy/normal/log_std Min -0.722405 +eval/num steps total 142893 +eval/num paths total 143 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.098931 +eval/Actions Std 0.917026 +eval/Actions Max 0.999999 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.49807 +time/logging (s) 0.00412993 +time/sampling batch (s) 0.289964 +time/saving (s) 0.00417494 +time/training (s) 6.75855 +time/epoch (s) 9.5549 +time/total (s) 1488.8 +Epoch -858 +---------------------------------- --------------- +2022-05-10 13:35:34.510357 PDT | [2] Epoch -857 finished +---------------------------------- --------------- +epoch -857 +replay_buffer/size 999033 +trainer/num train calls 144000 +trainer/Policy Loss -19.7365 +trainer/Log Pis Mean 25.0814 +trainer/Log Pis Std 12.7663 +trainer/Log Pis Max 65.7398 +trainer/Log Pis Min -5.47932 +trainer/policy/mean Mean -0.0417924 +trainer/policy/mean Std 0.904948 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.87672 +trainer/policy/normal/std Std 0.591673 +trainer/policy/normal/std Max 6.45356 +trainer/policy/normal/std Min 0.374746 +trainer/policy/normal/log_std Mean 1.02794 +trainer/policy/normal/log_std Std 0.264912 +trainer/policy/normal/log_std Max 1.86463 +trainer/policy/normal/log_std Min -0.981506 +eval/num steps total 143893 +eval/num paths total 144 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.1534 +eval/Actions Std 0.874724 +eval/Actions Max 0.999987 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.75442 +time/logging (s) 0.00383496 +time/sampling batch (s) 0.282725 +time/saving (s) 0.00361309 +time/training (s) 6.82611 +time/epoch (s) 9.8707 +time/total (s) 1498.68 +Epoch -857 +---------------------------------- --------------- +2022-05-10 13:35:43.936572 PDT | [2] Epoch -856 finished +---------------------------------- --------------- +epoch -856 +replay_buffer/size 999033 +trainer/num train calls 145000 +trainer/Policy Loss -19.4431 +trainer/Log Pis Mean 24.9554 +trainer/Log Pis Std 12.5301 +trainer/Log Pis Max 66.9562 +trainer/Log Pis Min -8.05562 +trainer/policy/mean Mean -0.0453645 +trainer/policy/mean Std 0.908826 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.95023 +trainer/policy/normal/std Std 0.562133 +trainer/policy/normal/std Max 5.63898 +trainer/policy/normal/std Min 0.445337 +trainer/policy/normal/log_std Mean 1.05819 +trainer/policy/normal/log_std Std 0.238328 +trainer/policy/normal/log_std Max 1.7297 +trainer/policy/normal/log_std Min -0.808924 +eval/num steps total 144893 +eval/num paths total 145 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.206014 +eval/Actions Std 0.932635 +eval/Actions Max 0.999997 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67873 +time/logging (s) 0.0039782 +time/sampling batch (s) 0.278884 +time/saving (s) 0.00357599 +time/training (s) 6.44133 +time/epoch (s) 9.4065 +time/total (s) 1508.09 +Epoch -856 +---------------------------------- --------------- +2022-05-10 13:35:54.033576 PDT | [2] Epoch -855 finished +---------------------------------- --------------- +epoch -855 +replay_buffer/size 999033 +trainer/num train calls 146000 +trainer/Policy Loss -19.9505 +trainer/Log Pis Mean 24.0267 +trainer/Log Pis Std 13.3547 +trainer/Log Pis Max 66.7887 +trainer/Log Pis Min -3.00075 +trainer/policy/mean Mean -0.0182942 +trainer/policy/mean Std 0.903253 +trainer/policy/mean Max 0.999974 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.84586 +trainer/policy/normal/std Std 0.60579 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.435967 +trainer/policy/normal/log_std Mean 1.01447 +trainer/policy/normal/log_std Std 0.279162 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -0.830189 +eval/num steps total 145893 +eval/num paths total 146 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.159436 +eval/Actions Std 0.890988 +eval/Actions Max 0.999996 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54655 +time/logging (s) 0.00375308 +time/sampling batch (s) 0.280558 +time/saving (s) 0.00348492 +time/training (s) 7.24233 +time/epoch (s) 10.0767 +time/total (s) 1518.17 +Epoch -855 +---------------------------------- --------------- +2022-05-10 13:36:04.906180 PDT | [2] Epoch -854 finished +---------------------------------- --------------- +epoch -854 +replay_buffer/size 999033 +trainer/num train calls 147000 +trainer/Policy Loss -19.1853 +trainer/Log Pis Mean 23.3522 +trainer/Log Pis Std 13.7952 +trainer/Log Pis Max 76.1013 +trainer/Log Pis Min -12.1794 +trainer/policy/mean Mean -0.0345732 +trainer/policy/mean Std 0.899847 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.86312 +trainer/policy/normal/std Std 0.592629 +trainer/policy/normal/std Max 5.16383 +trainer/policy/normal/std Min 0.388332 +trainer/policy/normal/log_std Mean 1.02129 +trainer/policy/normal/log_std Std 0.277527 +trainer/policy/normal/log_std Max 1.64168 +trainer/policy/normal/log_std Min -0.945895 +eval/num steps total 146893 +eval/num paths total 147 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.313873 +eval/Actions Std 0.877048 +eval/Actions Max 0.999986 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55189 +time/logging (s) 0.00380345 +time/sampling batch (s) 0.281756 +time/saving (s) 0.00350244 +time/training (s) 8.01154 +time/epoch (s) 10.8525 +time/total (s) 1529.02 +Epoch -854 +---------------------------------- --------------- +2022-05-10 13:36:15.395752 PDT | [2] Epoch -853 finished +---------------------------------- --------------- +epoch -853 +replay_buffer/size 999033 +trainer/num train calls 148000 +trainer/Policy Loss -19.3838 +trainer/Log Pis Mean 24.3252 +trainer/Log Pis Std 13.4932 +trainer/Log Pis Max 71.8458 +trainer/Log Pis Min -5.97358 +trainer/policy/mean Mean -0.0515795 +trainer/policy/mean Std 0.897184 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.88586 +trainer/policy/normal/std Std 0.580427 +trainer/policy/normal/std Max 5.87245 +trainer/policy/normal/std Min 0.391436 +trainer/policy/normal/log_std Mean 1.03303 +trainer/policy/normal/log_std Std 0.254476 +trainer/policy/normal/log_std Max 1.77027 +trainer/policy/normal/log_std Min -0.937933 +eval/num steps total 147893 +eval/num paths total 148 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0380216 +eval/Actions Std 0.859762 +eval/Actions Max 0.999994 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50436 +time/logging (s) 0.00414741 +time/sampling batch (s) 0.283027 +time/saving (s) 0.00401486 +time/training (s) 7.67402 +time/epoch (s) 10.4696 +time/total (s) 1539.5 +Epoch -853 +---------------------------------- --------------- +2022-05-10 13:36:25.091734 PDT | [2] Epoch -852 finished +---------------------------------- --------------- +epoch -852 +replay_buffer/size 999033 +trainer/num train calls 149000 +trainer/Policy Loss -19.8315 +trainer/Log Pis Mean 24.7527 +trainer/Log Pis Std 14.171 +trainer/Log Pis Max 76.286 +trainer/Log Pis Min -8.35651 +trainer/policy/mean Mean -0.0318461 +trainer/policy/mean Std 0.906725 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.85051 +trainer/policy/normal/std Std 0.586171 +trainer/policy/normal/std Max 5.39952 +trainer/policy/normal/std Min 0.477885 +trainer/policy/normal/log_std Mean 1.01824 +trainer/policy/normal/log_std Std 0.268344 +trainer/policy/normal/log_std Max 1.68631 +trainer/policy/normal/log_std Min -0.738386 +eval/num steps total 148893 +eval/num paths total 149 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00325262 +eval/Actions Std 0.903666 +eval/Actions Max 0.999993 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.72544 +time/logging (s) 0.00383523 +time/sampling batch (s) 0.284765 +time/saving (s) 0.00389739 +time/training (s) 6.65689 +time/epoch (s) 9.67482 +time/total (s) 1549.17 +Epoch -852 +---------------------------------- --------------- +2022-05-10 13:36:34.587285 PDT | [2] Epoch -851 finished +---------------------------------- --------------- +epoch -851 +replay_buffer/size 999033 +trainer/num train calls 150000 +trainer/Policy Loss -18.1784 +trainer/Log Pis Mean 25.1157 +trainer/Log Pis Std 13.7644 +trainer/Log Pis Max 70.0718 +trainer/Log Pis Min -9.81447 +trainer/policy/mean Mean -0.043473 +trainer/policy/mean Std 0.902354 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.89033 +trainer/policy/normal/std Std 0.59401 +trainer/policy/normal/std Max 5.02005 +trainer/policy/normal/std Min 0.482256 +trainer/policy/normal/log_std Mean 1.03273 +trainer/policy/normal/log_std Std 0.263831 +trainer/policy/normal/log_std Max 1.61344 +trainer/policy/normal/log_std Min -0.72928 +eval/num steps total 149893 +eval/num paths total 150 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.178445 +eval/Actions Std 0.919072 +eval/Actions Max 0.999997 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5036 +time/logging (s) 0.0038053 +time/sampling batch (s) 0.281957 +time/saving (s) 0.00356611 +time/training (s) 6.68252 +time/epoch (s) 9.47544 +time/total (s) 1558.65 +Epoch -851 +---------------------------------- --------------- +2022-05-10 13:36:45.858905 PDT | [2] Epoch -850 finished +---------------------------------- --------------- +epoch -850 +replay_buffer/size 999033 +trainer/num train calls 151000 +trainer/Policy Loss -20.1383 +trainer/Log Pis Mean 23.9992 +trainer/Log Pis Std 13.1676 +trainer/Log Pis Max 68.7814 +trainer/Log Pis Min -7.89761 +trainer/policy/mean Mean -0.0279118 +trainer/policy/mean Std 0.905277 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.86807 +trainer/policy/normal/std Std 0.56208 +trainer/policy/normal/std Max 4.93628 +trainer/policy/normal/std Min 0.504606 +trainer/policy/normal/log_std Mean 1.02761 +trainer/policy/normal/log_std Std 0.251761 +trainer/policy/normal/log_std Max 1.59661 +trainer/policy/normal/log_std Min -0.683978 +eval/num steps total 150893 +eval/num paths total 151 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.10966 +eval/Actions Std 0.895851 +eval/Actions Max 0.999992 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.78438 +time/logging (s) 0.00371061 +time/sampling batch (s) 0.794884 +time/saving (s) 0.00367383 +time/training (s) 7.6637 +time/epoch (s) 11.2504 +time/total (s) 1569.91 +Epoch -850 +---------------------------------- --------------- +2022-05-10 13:36:55.660867 PDT | [2] Epoch -849 finished +---------------------------------- --------------- +epoch -849 +replay_buffer/size 999033 +trainer/num train calls 152000 +trainer/Policy Loss -19.3459 +trainer/Log Pis Mean 24.6903 +trainer/Log Pis Std 13.8859 +trainer/Log Pis Max 76.8067 +trainer/Log Pis Min -13.6413 +trainer/policy/mean Mean -0.0476589 +trainer/policy/mean Std 0.911258 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.83889 +trainer/policy/normal/std Std 0.547794 +trainer/policy/normal/std Max 5.02555 +trainer/policy/normal/std Min 0.461142 +trainer/policy/normal/log_std Mean 1.0189 +trainer/policy/normal/log_std Std 0.242322 +trainer/policy/normal/log_std Max 1.61453 +trainer/policy/normal/log_std Min -0.774048 +eval/num steps total 151893 +eval/num paths total 152 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0812594 +eval/Actions Std 0.914905 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.74203 +time/logging (s) 0.00377969 +time/sampling batch (s) 0.3018 +time/saving (s) 0.0034736 +time/training (s) 6.72962 +time/epoch (s) 9.7807 +time/total (s) 1579.69 +Epoch -849 +---------------------------------- --------------- +2022-05-10 13:37:05.716008 PDT | [2] Epoch -848 finished +---------------------------------- --------------- +epoch -848 +replay_buffer/size 999033 +trainer/num train calls 153000 +trainer/Policy Loss -18.8392 +trainer/Log Pis Mean 25.0936 +trainer/Log Pis Std 13.0848 +trainer/Log Pis Max 58.6105 +trainer/Log Pis Min -8.36144 +trainer/policy/mean Mean -0.0499563 +trainer/policy/mean Std 0.904684 +trainer/policy/mean Max 0.999981 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.81095 +trainer/policy/normal/std Std 0.5868 +trainer/policy/normal/std Max 5.28614 +trainer/policy/normal/std Min 0.531919 +trainer/policy/normal/log_std Mean 1.00391 +trainer/policy/normal/log_std Std 0.268375 +trainer/policy/normal/log_std Max 1.66509 +trainer/policy/normal/log_std Min -0.631265 +eval/num steps total 152893 +eval/num paths total 153 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0360197 +eval/Actions Std 0.901405 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61131 +time/logging (s) 0.00421584 +time/sampling batch (s) 0.536239 +time/saving (s) 0.00402331 +time/training (s) 6.87922 +time/epoch (s) 10.035 +time/total (s) 1589.73 +Epoch -848 +---------------------------------- --------------- +2022-05-10 13:37:14.950838 PDT | [2] Epoch -847 finished +---------------------------------- --------------- +epoch -847 +replay_buffer/size 999033 +trainer/num train calls 154000 +trainer/Policy Loss -19.1732 +trainer/Log Pis Mean 24.6937 +trainer/Log Pis Std 13.1654 +trainer/Log Pis Max 67.1266 +trainer/Log Pis Min -7.02157 +trainer/policy/mean Mean -0.0544345 +trainer/policy/mean Std 0.905494 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.84654 +trainer/policy/normal/std Std 0.574807 +trainer/policy/normal/std Max 5.55365 +trainer/policy/normal/std Min 0.438203 +trainer/policy/normal/log_std Mean 1.01874 +trainer/policy/normal/log_std Std 0.257866 +trainer/policy/normal/log_std Max 1.71446 +trainer/policy/normal/log_std Min -0.825072 +eval/num steps total 153893 +eval/num paths total 154 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0377956 +eval/Actions Std 0.921677 +eval/Actions Max 0.999993 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51964 +time/logging (s) 0.00379819 +time/sampling batch (s) 0.279626 +time/saving (s) 0.00376057 +time/training (s) 6.40685 +time/epoch (s) 9.21367 +time/total (s) 1598.94 +Epoch -847 +---------------------------------- --------------- +2022-05-10 13:37:24.180930 PDT | [2] Epoch -846 finished +---------------------------------- --------------- +epoch -846 +replay_buffer/size 999033 +trainer/num train calls 155000 +trainer/Policy Loss -20.2671 +trainer/Log Pis Mean 24.1063 +trainer/Log Pis Std 13.9611 +trainer/Log Pis Max 70.3446 +trainer/Log Pis Min -10.4707 +trainer/policy/mean Mean -0.0482508 +trainer/policy/mean Std 0.907274 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.8424 +trainer/policy/normal/std Std 0.582923 +trainer/policy/normal/std Max 5.38153 +trainer/policy/normal/std Min 0.45251 +trainer/policy/normal/log_std Mean 1.01637 +trainer/policy/normal/log_std Std 0.261765 +trainer/policy/normal/log_std Max 1.68297 +trainer/policy/normal/log_std Min -0.792946 +eval/num steps total 154893 +eval/num paths total 155 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.107347 +eval/Actions Std 0.910434 +eval/Actions Max 0.999994 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.44191 +time/logging (s) 0.00370482 +time/sampling batch (s) 0.285559 +time/saving (s) 0.00345184 +time/training (s) 6.47467 +time/epoch (s) 9.2093 +time/total (s) 1608.16 +Epoch -846 +---------------------------------- --------------- +2022-05-10 13:37:35.175474 PDT | [2] Epoch -845 finished +---------------------------------- --------------- +epoch -845 +replay_buffer/size 999033 +trainer/num train calls 156000 +trainer/Policy Loss -19.3639 +trainer/Log Pis Mean 24.7098 +trainer/Log Pis Std 12.7524 +trainer/Log Pis Max 78.9345 +trainer/Log Pis Min -6.02248 +trainer/policy/mean Mean -0.00145976 +trainer/policy/mean Std 0.905753 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.83855 +trainer/policy/normal/std Std 0.564138 +trainer/policy/normal/std Max 4.90779 +trainer/policy/normal/std Min 0.403318 +trainer/policy/normal/log_std Mean 1.01714 +trainer/policy/normal/log_std Std 0.250813 +trainer/policy/normal/log_std Max 1.59082 +trainer/policy/normal/log_std Min -0.908031 +eval/num steps total 155893 +eval/num paths total 156 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.074682 +eval/Actions Std 0.914554 +eval/Actions Max 0.999992 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.7058 +time/logging (s) 0.00372579 +time/sampling batch (s) 0.283475 +time/saving (s) 0.00347014 +time/training (s) 7.97786 +time/epoch (s) 10.9743 +time/total (s) 1619.13 +Epoch -845 +---------------------------------- --------------- +2022-05-10 13:37:44.917647 PDT | [2] Epoch -844 finished +---------------------------------- --------------- +epoch -844 +replay_buffer/size 999033 +trainer/num train calls 157000 +trainer/Policy Loss -20.8578 +trainer/Log Pis Mean 23.3706 +trainer/Log Pis Std 13.766 +trainer/Log Pis Max 79.4111 +trainer/Log Pis Min -7.73076 +trainer/policy/mean Mean -0.0390547 +trainer/policy/mean Std 0.906397 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.84132 +trainer/policy/normal/std Std 0.544948 +trainer/policy/normal/std Max 5.68835 +trainer/policy/normal/std Min 0.474666 +trainer/policy/normal/log_std Mean 1.01952 +trainer/policy/normal/log_std Std 0.244684 +trainer/policy/normal/log_std Max 1.73842 +trainer/policy/normal/log_std Min -0.745143 +eval/num steps total 156893 +eval/num paths total 157 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0155415 +eval/Actions Std 0.902817 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64908 +time/logging (s) 0.00372433 +time/sampling batch (s) 0.283276 +time/saving (s) 0.00458935 +time/training (s) 6.78097 +time/epoch (s) 9.72164 +time/total (s) 1628.86 +Epoch -844 +---------------------------------- --------------- +2022-05-10 13:37:54.196812 PDT | [2] Epoch -843 finished +---------------------------------- --------------- +epoch -843 +replay_buffer/size 999033 +trainer/num train calls 158000 +trainer/Policy Loss -19.5608 +trainer/Log Pis Mean 24.4607 +trainer/Log Pis Std 13.3778 +trainer/Log Pis Max 66.3265 +trainer/Log Pis Min -7.55418 +trainer/policy/mean Mean -0.0410236 +trainer/policy/mean Std 0.908463 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.91486 +trainer/policy/normal/std Std 0.581151 +trainer/policy/normal/std Max 6.07571 +trainer/policy/normal/std Min 0.424102 +trainer/policy/normal/log_std Mean 1.0437 +trainer/policy/normal/log_std Std 0.250344 +trainer/policy/normal/log_std Max 1.8043 +trainer/policy/normal/log_std Min -0.857781 +eval/num steps total 157893 +eval/num paths total 158 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.313706 +eval/Actions Std 0.85343 +eval/Actions Max 0.999999 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57726 +time/logging (s) 0.00375545 +time/sampling batch (s) 0.281676 +time/saving (s) 0.00344369 +time/training (s) 6.39264 +time/epoch (s) 9.25878 +time/total (s) 1638.12 +Epoch -843 +---------------------------------- --------------- +2022-05-10 13:38:04.510612 PDT | [2] Epoch -842 finished +---------------------------------- --------------- +epoch -842 +replay_buffer/size 999033 +trainer/num train calls 159000 +trainer/Policy Loss -19.657 +trainer/Log Pis Mean 24.2081 +trainer/Log Pis Std 13.0747 +trainer/Log Pis Max 64.6406 +trainer/Log Pis Min -8.97375 +trainer/policy/mean Mean -0.0411675 +trainer/policy/mean Std 0.904278 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.86211 +trainer/policy/normal/std Std 0.576503 +trainer/policy/normal/std Max 5.01071 +trainer/policy/normal/std Min 0.401701 +trainer/policy/normal/log_std Mean 1.0235 +trainer/policy/normal/log_std Std 0.263232 +trainer/policy/normal/log_std Max 1.61158 +trainer/policy/normal/log_std Min -0.912048 +eval/num steps total 158893 +eval/num paths total 159 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.141696 +eval/Actions Std 0.915385 +eval/Actions Max 0.999995 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63395 +time/logging (s) 0.00406214 +time/sampling batch (s) 0.281942 +time/saving (s) 0.00417325 +time/training (s) 7.36967 +time/epoch (s) 10.2938 +time/total (s) 1648.42 +Epoch -842 +---------------------------------- --------------- +2022-05-10 13:38:14.410820 PDT | [2] Epoch -841 finished +---------------------------------- --------------- +epoch -841 +replay_buffer/size 999033 +trainer/num train calls 160000 +trainer/Policy Loss -18.7625 +trainer/Log Pis Mean 24.0126 +trainer/Log Pis Std 13.5138 +trainer/Log Pis Max 66.8566 +trainer/Log Pis Min -13.4771 +trainer/policy/mean Mean -0.0329023 +trainer/policy/mean Std 0.906878 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999977 +trainer/policy/normal/std Mean 2.80235 +trainer/policy/normal/std Std 0.585747 +trainer/policy/normal/std Max 5.10196 +trainer/policy/normal/std Min 0.377171 +trainer/policy/normal/log_std Mean 0.999226 +trainer/policy/normal/log_std Std 0.280184 +trainer/policy/normal/log_std Max 1.62962 +trainer/policy/normal/log_std Min -0.975056 +eval/num steps total 159893 +eval/num paths total 160 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.116984 +eval/Actions Std 0.914479 +eval/Actions Max 0.999997 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.75237 +time/logging (s) 0.00375725 +time/sampling batch (s) 0.279592 +time/saving (s) 0.00346408 +time/training (s) 6.84023 +time/epoch (s) 9.87941 +time/total (s) 1658.3 +Epoch -841 +---------------------------------- --------------- +2022-05-10 13:38:24.628938 PDT | [2] Epoch -840 finished +---------------------------------- --------------- +epoch -840 +replay_buffer/size 999033 +trainer/num train calls 161000 +trainer/Policy Loss -19.0087 +trainer/Log Pis Mean 25.3502 +trainer/Log Pis Std 14.0584 +trainer/Log Pis Max 77.8213 +trainer/Log Pis Min -6.73481 +trainer/policy/mean Mean -0.0201498 +trainer/policy/mean Std 0.905703 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.85288 +trainer/policy/normal/std Std 0.572435 +trainer/policy/normal/std Max 5.55181 +trainer/policy/normal/std Min 0.398292 +trainer/policy/normal/log_std Mean 1.02052 +trainer/policy/normal/log_std Std 0.262238 +trainer/policy/normal/log_std Max 1.71412 +trainer/policy/normal/log_std Min -0.920569 +eval/num steps total 160893 +eval/num paths total 161 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.186191 +eval/Actions Std 0.893656 +eval/Actions Max 0.999987 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64061 +time/logging (s) 0.00373635 +time/sampling batch (s) 0.282737 +time/saving (s) 0.00347724 +time/training (s) 7.26714 +time/epoch (s) 10.1977 +time/total (s) 1668.5 +Epoch -840 +---------------------------------- --------------- +2022-05-10 13:38:35.052026 PDT | [2] Epoch -839 finished +---------------------------------- --------------- +epoch -839 +replay_buffer/size 999033 +trainer/num train calls 162000 +trainer/Policy Loss -19.156 +trainer/Log Pis Mean 24.8492 +trainer/Log Pis Std 13.7031 +trainer/Log Pis Max 70.7543 +trainer/Log Pis Min -7.72103 +trainer/policy/mean Mean -0.0249201 +trainer/policy/mean Std 0.906115 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.91891 +trainer/policy/normal/std Std 0.594959 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.385015 +trainer/policy/normal/log_std Mean 1.04247 +trainer/policy/normal/log_std Std 0.267059 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -0.954474 +eval/num steps total 161893 +eval/num paths total 162 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0229766 +eval/Actions Std 0.909658 +eval/Actions Max 0.999999 +eval/Actions Min -0.999988 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59576 +time/logging (s) 0.00373289 +time/sampling batch (s) 0.782361 +time/saving (s) 0.00348369 +time/training (s) 7.01743 +time/epoch (s) 10.4028 +time/total (s) 1678.91 +Epoch -839 +---------------------------------- --------------- +2022-05-10 13:38:45.331069 PDT | [2] Epoch -838 finished +---------------------------------- --------------- +epoch -838 +replay_buffer/size 999033 +trainer/num train calls 163000 +trainer/Policy Loss -20.4893 +trainer/Log Pis Mean 24.7083 +trainer/Log Pis Std 13.2359 +trainer/Log Pis Max 68.2243 +trainer/Log Pis Min -3.43567 +trainer/policy/mean Mean -0.0295624 +trainer/policy/mean Std 0.907549 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.85598 +trainer/policy/normal/std Std 0.603309 +trainer/policy/normal/std Max 6.05441 +trainer/policy/normal/std Min 0.442771 +trainer/policy/normal/log_std Mean 1.01746 +trainer/policy/normal/log_std Std 0.284278 +trainer/policy/normal/log_std Max 1.80079 +trainer/policy/normal/log_std Min -0.814702 +eval/num steps total 162893 +eval/num paths total 163 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.350657 +eval/Actions Std 0.82988 +eval/Actions Max 0.999994 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56733 +time/logging (s) 0.00382204 +time/sampling batch (s) 0.532918 +time/saving (s) 0.00349278 +time/training (s) 7.1511 +time/epoch (s) 10.2587 +time/total (s) 1689.17 +Epoch -838 +---------------------------------- --------------- +2022-05-10 13:38:56.106120 PDT | [2] Epoch -837 finished +---------------------------------- --------------- +epoch -837 +replay_buffer/size 999033 +trainer/num train calls 164000 +trainer/Policy Loss -18.6049 +trainer/Log Pis Mean 24.7406 +trainer/Log Pis Std 13.6988 +trainer/Log Pis Max 69.6954 +trainer/Log Pis Min -6.43646 +trainer/policy/mean Mean -0.0510965 +trainer/policy/mean Std 0.903012 +trainer/policy/mean Max 1 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.82741 +trainer/policy/normal/std Std 0.604213 +trainer/policy/normal/std Max 6.35101 +trainer/policy/normal/std Min 0.396477 +trainer/policy/normal/log_std Mean 1.00661 +trainer/policy/normal/log_std Std 0.287861 +trainer/policy/normal/log_std Max 1.84861 +trainer/policy/normal/log_std Min -0.925136 +eval/num steps total 163893 +eval/num paths total 164 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.197646 +eval/Actions Std 0.846344 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64117 +time/logging (s) 0.00413456 +time/sampling batch (s) 0.28318 +time/saving (s) 0.00426555 +time/training (s) 7.82201 +time/epoch (s) 10.7548 +time/total (s) 1699.93 +Epoch -837 +---------------------------------- --------------- +2022-05-10 13:39:06.086807 PDT | [2] Epoch -836 finished +---------------------------------- --------------- +epoch -836 +replay_buffer/size 999033 +trainer/num train calls 165000 +trainer/Policy Loss -20.1518 +trainer/Log Pis Mean 24.4412 +trainer/Log Pis Std 13.5281 +trainer/Log Pis Max 81.3834 +trainer/Log Pis Min -6.97239 +trainer/policy/mean Mean -0.0468267 +trainer/policy/mean Std 0.907608 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.80048 +trainer/policy/normal/std Std 0.57156 +trainer/policy/normal/std Max 5.93972 +trainer/policy/normal/std Min 0.427289 +trainer/policy/normal/log_std Mean 1.0013 +trainer/policy/normal/log_std Std 0.264375 +trainer/policy/normal/log_std Max 1.78166 +trainer/policy/normal/log_std Min -0.850294 +eval/num steps total 164893 +eval/num paths total 165 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.103037 +eval/Actions Std 0.907748 +eval/Actions Max 0.999997 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68386 +time/logging (s) 0.00379881 +time/sampling batch (s) 0.28236 +time/saving (s) 0.00359696 +time/training (s) 6.98597 +time/epoch (s) 9.95959 +time/total (s) 1709.89 +Epoch -836 +---------------------------------- --------------- +2022-05-10 13:39:17.047995 PDT | [2] Epoch -835 finished +---------------------------------- --------------- +epoch -835 +replay_buffer/size 999033 +trainer/num train calls 166000 +trainer/Policy Loss -19.1398 +trainer/Log Pis Mean 24.2453 +trainer/Log Pis Std 13.8351 +trainer/Log Pis Max 67.4091 +trainer/Log Pis Min -11.3146 +trainer/policy/mean Mean -0.0696098 +trainer/policy/mean Std 0.90181 +trainer/policy/mean Max 1 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.8569 +trainer/policy/normal/std Std 0.591306 +trainer/policy/normal/std Max 5.40728 +trainer/policy/normal/std Min 0.45095 +trainer/policy/normal/log_std Mean 1.01988 +trainer/policy/normal/log_std Std 0.27259 +trainer/policy/normal/log_std Max 1.68775 +trainer/policy/normal/log_std Min -0.7964 +eval/num steps total 165893 +eval/num paths total 166 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.249341 +eval/Actions Std 0.885511 +eval/Actions Max 0.999995 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57865 +time/logging (s) 0.00372185 +time/sampling batch (s) 0.53879 +time/saving (s) 0.00341632 +time/training (s) 7.81581 +time/epoch (s) 10.9404 +time/total (s) 1720.83 +Epoch -835 +---------------------------------- --------------- +2022-05-10 13:39:27.798545 PDT | [2] Epoch -834 finished +---------------------------------- --------------- +epoch -834 +replay_buffer/size 999033 +trainer/num train calls 167000 +trainer/Policy Loss -18.3666 +trainer/Log Pis Mean 23.8906 +trainer/Log Pis Std 13.0607 +trainer/Log Pis Max 62.4244 +trainer/Log Pis Min -11.2771 +trainer/policy/mean Mean -0.0210131 +trainer/policy/mean Std 0.904683 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999983 +trainer/policy/normal/std Mean 2.8627 +trainer/policy/normal/std Std 0.580348 +trainer/policy/normal/std Max 5.14583 +trainer/policy/normal/std Min 0.484328 +trainer/policy/normal/log_std Mean 1.02286 +trainer/policy/normal/log_std Std 0.26813 +trainer/policy/normal/log_std Max 1.63819 +trainer/policy/normal/log_std Min -0.724993 +eval/num steps total 166893 +eval/num paths total 167 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0122213 +eval/Actions Std 0.894773 +eval/Actions Max 0.999997 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62432 +time/logging (s) 0.00375125 +time/sampling batch (s) 0.281658 +time/saving (s) 0.00403441 +time/training (s) 7.81658 +time/epoch (s) 10.7303 +time/total (s) 1731.57 +Epoch -834 +---------------------------------- --------------- +2022-05-10 13:39:39.024443 PDT | [2] Epoch -833 finished +---------------------------------- --------------- +epoch -833 +replay_buffer/size 999033 +trainer/num train calls 168000 +trainer/Policy Loss -19.0863 +trainer/Log Pis Mean 23.2037 +trainer/Log Pis Std 13.4794 +trainer/Log Pis Max 78.0012 +trainer/Log Pis Min -8.37112 +trainer/policy/mean Mean -0.0451362 +trainer/policy/mean Std 0.898916 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.81592 +trainer/policy/normal/std Std 0.59677 +trainer/policy/normal/std Max 5.53902 +trainer/policy/normal/std Min 0.445538 +trainer/policy/normal/log_std Mean 1.00379 +trainer/policy/normal/log_std Std 0.279264 +trainer/policy/normal/log_std Max 1.71182 +trainer/policy/normal/log_std Min -0.808472 +eval/num steps total 167893 +eval/num paths total 168 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.223606 +eval/Actions Std 0.895602 +eval/Actions Max 0.99999 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.74591 +time/logging (s) 0.00370956 +time/sampling batch (s) 0.27886 +time/saving (s) 0.0034448 +time/training (s) 8.17322 +time/epoch (s) 11.2051 +time/total (s) 1742.77 +Epoch -833 +---------------------------------- --------------- +2022-05-10 13:39:49.010289 PDT | [2] Epoch -832 finished +---------------------------------- --------------- +epoch -832 +replay_buffer/size 999033 +trainer/num train calls 169000 +trainer/Policy Loss -19.8702 +trainer/Log Pis Mean 24.3065 +trainer/Log Pis Std 12.9195 +trainer/Log Pis Max 72.7564 +trainer/Log Pis Min -6.06017 +trainer/policy/mean Mean -0.0257806 +trainer/policy/mean Std 0.905493 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.90112 +trainer/policy/normal/std Std 0.60565 +trainer/policy/normal/std Max 5.59714 +trainer/policy/normal/std Min 0.407618 +trainer/policy/normal/log_std Mean 1.03504 +trainer/policy/normal/log_std Std 0.271877 +trainer/policy/normal/log_std Max 1.72226 +trainer/policy/normal/log_std Min -0.897426 +eval/num steps total 168893 +eval/num paths total 169 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.108951 +eval/Actions Std 0.896679 +eval/Actions Max 0.999988 +eval/Actions Min -0.999946 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64939 +time/logging (s) 0.00417792 +time/sampling batch (s) 0.281359 +time/saving (s) 0.00414277 +time/training (s) 7.02696 +time/epoch (s) 9.96603 +time/total (s) 1752.74 +Epoch -832 +---------------------------------- --------------- +2022-05-10 13:39:59.522253 PDT | [2] Epoch -831 finished +---------------------------------- --------------- +epoch -831 +replay_buffer/size 999033 +trainer/num train calls 170000 +trainer/Policy Loss -18.6554 +trainer/Log Pis Mean 24.0097 +trainer/Log Pis Std 13.1295 +trainer/Log Pis Max 81.388 +trainer/Log Pis Min -9.95667 +trainer/policy/mean Mean -0.0121126 +trainer/policy/mean Std 0.905798 +trainer/policy/mean Max 1 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.82787 +trainer/policy/normal/std Std 0.592868 +trainer/policy/normal/std Max 5.56618 +trainer/policy/normal/std Min 0.451903 +trainer/policy/normal/log_std Mean 1.00881 +trainer/policy/normal/log_std Std 0.275943 +trainer/policy/normal/log_std Max 1.71671 +trainer/policy/normal/log_std Min -0.794287 +eval/num steps total 169893 +eval/num paths total 170 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.049623 +eval/Actions Std 0.889886 +eval/Actions Max 0.999997 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68238 +time/logging (s) 0.00379759 +time/sampling batch (s) 0.530508 +time/saving (s) 0.00365033 +time/training (s) 7.27066 +time/epoch (s) 10.491 +time/total (s) 1763.24 +Epoch -831 +---------------------------------- --------------- +2022-05-10 13:40:10.383913 PDT | [2] Epoch -830 finished +---------------------------------- --------------- +epoch -830 +replay_buffer/size 999033 +trainer/num train calls 171000 +trainer/Policy Loss -17.5775 +trainer/Log Pis Mean 23.8766 +trainer/Log Pis Std 12.972 +trainer/Log Pis Max 64.8387 +trainer/Log Pis Min -6.31146 +trainer/policy/mean Mean -0.0280738 +trainer/policy/mean Std 0.900122 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.80124 +trainer/policy/normal/std Std 0.592882 +trainer/policy/normal/std Max 4.91052 +trainer/policy/normal/std Min 0.424052 +trainer/policy/normal/log_std Mean 0.998998 +trainer/policy/normal/log_std Std 0.276854 +trainer/policy/normal/log_std Max 1.59138 +trainer/policy/normal/log_std Min -0.857899 +eval/num steps total 170893 +eval/num paths total 171 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.282171 +eval/Actions Std 0.885264 +eval/Actions Max 0.999988 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.83144 +time/logging (s) 0.00389055 +time/sampling batch (s) 0.53635 +time/saving (s) 0.00365322 +time/training (s) 7.46451 +time/epoch (s) 10.8398 +time/total (s) 1774.08 +Epoch -830 +---------------------------------- --------------- +2022-05-10 13:40:20.396169 PDT | [2] Epoch -829 finished +---------------------------------- --------------- +epoch -829 +replay_buffer/size 999033 +trainer/num train calls 172000 +trainer/Policy Loss -20.2797 +trainer/Log Pis Mean 24.9181 +trainer/Log Pis Std 14.0681 +trainer/Log Pis Max 78.8779 +trainer/Log Pis Min -1.86658 +trainer/policy/mean Mean -0.0514484 +trainer/policy/mean Std 0.906008 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.86075 +trainer/policy/normal/std Std 0.60456 +trainer/policy/normal/std Max 5.48046 +trainer/policy/normal/std Min 0.34756 +trainer/policy/normal/log_std Mean 1.0195 +trainer/policy/normal/log_std Std 0.281581 +trainer/policy/normal/log_std Max 1.70119 +trainer/policy/normal/log_std Min -1.05682 +eval/num steps total 171893 +eval/num paths total 172 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0312401 +eval/Actions Std 0.82814 +eval/Actions Max 0.999998 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.72619 +time/logging (s) 0.00372973 +time/sampling batch (s) 0.286609 +time/saving (s) 0.0034728 +time/training (s) 6.97131 +time/epoch (s) 9.99131 +time/total (s) 1784.08 +Epoch -829 +---------------------------------- --------------- +2022-05-10 13:40:30.576167 PDT | [2] Epoch -828 finished +---------------------------------- --------------- +epoch -828 +replay_buffer/size 999033 +trainer/num train calls 173000 +trainer/Policy Loss -18.9786 +trainer/Log Pis Mean 24.1139 +trainer/Log Pis Std 13.0145 +trainer/Log Pis Max 76.9615 +trainer/Log Pis Min -7.90161 +trainer/policy/mean Mean -0.0225692 +trainer/policy/mean Std 0.903373 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.79752 +trainer/policy/normal/std Std 0.594759 +trainer/policy/normal/std Max 6.9435 +trainer/policy/normal/std Min 0.391918 +trainer/policy/normal/log_std Mean 0.996484 +trainer/policy/normal/log_std Std 0.285611 +trainer/policy/normal/log_std Max 1.93781 +trainer/policy/normal/log_std Min -0.936703 +eval/num steps total 172893 +eval/num paths total 173 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0582371 +eval/Actions Std 0.896669 +eval/Actions Max 0.999988 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.776 +time/logging (s) 0.0036925 +time/sampling batch (s) 0.533708 +time/saving (s) 0.00352957 +time/training (s) 6.84246 +time/epoch (s) 10.1594 +time/total (s) 1794.24 +Epoch -828 +---------------------------------- --------------- +2022-05-10 13:40:40.401625 PDT | [2] Epoch -827 finished +---------------------------------- --------------- +epoch -827 +replay_buffer/size 999033 +trainer/num train calls 174000 +trainer/Policy Loss -20.5781 +trainer/Log Pis Mean 25.9912 +trainer/Log Pis Std 13.3978 +trainer/Log Pis Max 66.0959 +trainer/Log Pis Min -9.26447 +trainer/policy/mean Mean -0.0417691 +trainer/policy/mean Std 0.91048 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.86877 +trainer/policy/normal/std Std 0.577708 +trainer/policy/normal/std Max 5.56634 +trainer/policy/normal/std Min 0.483116 +trainer/policy/normal/log_std Mean 1.02646 +trainer/policy/normal/log_std Std 0.258713 +trainer/policy/normal/log_std Max 1.71674 +trainer/policy/normal/log_std Min -0.727499 +eval/num steps total 173893 +eval/num paths total 174 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00400236 +eval/Actions Std 0.903565 +eval/Actions Max 0.999993 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.76876 +time/logging (s) 0.00411766 +time/sampling batch (s) 0.282909 +time/saving (s) 0.00398216 +time/training (s) 6.74581 +time/epoch (s) 9.80558 +time/total (s) 1804.05 +Epoch -827 +---------------------------------- --------------- +2022-05-10 13:40:50.735625 PDT | [2] Epoch -826 finished +---------------------------------- --------------- +epoch -826 +replay_buffer/size 999033 +trainer/num train calls 175000 +trainer/Policy Loss -19.6043 +trainer/Log Pis Mean 23.4743 +trainer/Log Pis Std 12.793 +trainer/Log Pis Max 61.4104 +trainer/Log Pis Min -6.48046 +trainer/policy/mean Mean -0.0124516 +trainer/policy/mean Std 0.905758 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.85045 +trainer/policy/normal/std Std 0.592149 +trainer/policy/normal/std Max 5.00631 +trainer/policy/normal/std Min 0.347077 +trainer/policy/normal/log_std Mean 1.01721 +trainer/policy/normal/log_std Std 0.274696 +trainer/policy/normal/log_std Max 1.6107 +trainer/policy/normal/log_std Min -1.05821 +eval/num steps total 174893 +eval/num paths total 175 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.639645 +eval/Actions Std 0.707803 +eval/Actions Max 0.999994 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.74795 +time/logging (s) 0.00401659 +time/sampling batch (s) 0.529485 +time/saving (s) 0.00370563 +time/training (s) 7.02811 +time/epoch (s) 10.3133 +time/total (s) 1814.36 +Epoch -826 +---------------------------------- --------------- +2022-05-10 13:40:59.864567 PDT | [2] Epoch -825 finished +---------------------------------- --------------- +epoch -825 +replay_buffer/size 999033 +trainer/num train calls 176000 +trainer/Policy Loss -20.3469 +trainer/Log Pis Mean 23.2093 +trainer/Log Pis Std 13.9129 +trainer/Log Pis Max 82.8047 +trainer/Log Pis Min -11.8559 +trainer/policy/mean Mean -0.0367724 +trainer/policy/mean Std 0.902321 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.8927 +trainer/policy/normal/std Std 0.618926 +trainer/policy/normal/std Max 5.32956 +trainer/policy/normal/std Min 0.419753 +trainer/policy/normal/log_std Mean 1.03002 +trainer/policy/normal/log_std Std 0.283167 +trainer/policy/normal/log_std Max 1.67327 +trainer/policy/normal/log_std Min -0.868089 +eval/num steps total 175893 +eval/num paths total 176 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0151947 +eval/Actions Std 0.895995 +eval/Actions Max 0.999996 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66145 +time/logging (s) 0.00375288 +time/sampling batch (s) 0.278648 +time/saving (s) 0.00344182 +time/training (s) 6.16118 +time/epoch (s) 9.10847 +time/total (s) 1823.47 +Epoch -825 +---------------------------------- --------------- +2022-05-10 13:41:09.769668 PDT | [2] Epoch -824 finished +---------------------------------- --------------- +epoch -824 +replay_buffer/size 999033 +trainer/num train calls 177000 +trainer/Policy Loss -18.6695 +trainer/Log Pis Mean 24.2448 +trainer/Log Pis Std 13.2572 +trainer/Log Pis Max 74.1372 +trainer/Log Pis Min -6.56573 +trainer/policy/mean Mean -0.0230624 +trainer/policy/mean Std 0.906349 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.82812 +trainer/policy/normal/std Std 0.592295 +trainer/policy/normal/std Max 5.14084 +trainer/policy/normal/std Min 0.421765 +trainer/policy/normal/log_std Mean 1.00881 +trainer/policy/normal/log_std Std 0.276921 +trainer/policy/normal/log_std Max 1.63722 +trainer/policy/normal/log_std Min -0.863306 +eval/num steps total 176893 +eval/num paths total 177 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.162434 +eval/Actions Std 0.900672 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53857 +time/logging (s) 0.00390038 +time/sampling batch (s) 0.27841 +time/saving (s) 0.00358061 +time/training (s) 7.06081 +time/epoch (s) 9.88527 +time/total (s) 1833.36 +Epoch -824 +---------------------------------- --------------- +2022-05-10 13:41:19.901237 PDT | [2] Epoch -823 finished +---------------------------------- --------------- +epoch -823 +replay_buffer/size 999033 +trainer/num train calls 178000 +trainer/Policy Loss -18.482 +trainer/Log Pis Mean 23.2513 +trainer/Log Pis Std 13.0317 +trainer/Log Pis Max 62.0715 +trainer/Log Pis Min -12.0425 +trainer/policy/mean Mean -0.0191038 +trainer/policy/mean Std 0.90358 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.83755 +trainer/policy/normal/std Std 0.576735 +trainer/policy/normal/std Max 5.21029 +trainer/policy/normal/std Min 0.496998 +trainer/policy/normal/log_std Mean 1.01506 +trainer/policy/normal/log_std Std 0.260355 +trainer/policy/normal/log_std Max 1.65064 +trainer/policy/normal/log_std Min -0.699169 +eval/num steps total 177893 +eval/num paths total 178 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00893565 +eval/Actions Std 0.90104 +eval/Actions Max 0.999997 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57637 +time/logging (s) 0.00369261 +time/sampling batch (s) 0.529583 +time/saving (s) 0.00354962 +time/training (s) 6.99784 +time/epoch (s) 10.111 +time/total (s) 1843.48 +Epoch -823 +---------------------------------- --------------- +2022-05-10 13:41:28.970486 PDT | [2] Epoch -822 finished +---------------------------------- --------------- +epoch -822 +replay_buffer/size 999033 +trainer/num train calls 179000 +trainer/Policy Loss -19.7789 +trainer/Log Pis Mean 25.175 +trainer/Log Pis Std 13.5638 +trainer/Log Pis Max 66.7078 +trainer/Log Pis Min -9.72641 +trainer/policy/mean Mean -0.0557648 +trainer/policy/mean Std 0.910344 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.93162 +trainer/policy/normal/std Std 0.588452 +trainer/policy/normal/std Max 5.76809 +trainer/policy/normal/std Min 0.468766 +trainer/policy/normal/log_std Mean 1.0494 +trainer/policy/normal/log_std Std 0.249947 +trainer/policy/normal/log_std Max 1.75234 +trainer/policy/normal/log_std Min -0.757653 +eval/num steps total 178893 +eval/num paths total 179 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.138942 +eval/Actions Std 0.878804 +eval/Actions Max 0.999989 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68371 +time/logging (s) 0.00371326 +time/sampling batch (s) 0.279371 +time/saving (s) 0.00341294 +time/training (s) 6.07904 +time/epoch (s) 9.04924 +time/total (s) 1852.53 +Epoch -822 +---------------------------------- --------------- +2022-05-10 13:41:38.756006 PDT | [2] Epoch -821 finished +---------------------------------- --------------- +epoch -821 +replay_buffer/size 999033 +trainer/num train calls 180000 +trainer/Policy Loss -20.4231 +trainer/Log Pis Mean 24.7478 +trainer/Log Pis Std 12.9275 +trainer/Log Pis Max 74.5633 +trainer/Log Pis Min -6.54631 +trainer/policy/mean Mean -0.0171918 +trainer/policy/mean Std 0.905365 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.85703 +trainer/policy/normal/std Std 0.582432 +trainer/policy/normal/std Max 5.81715 +trainer/policy/normal/std Min 0.359148 +trainer/policy/normal/log_std Mean 1.02227 +trainer/policy/normal/log_std Std 0.257516 +trainer/policy/normal/log_std Max 1.76081 +trainer/policy/normal/log_std Min -1.02402 +eval/num steps total 179893 +eval/num paths total 180 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0810913 +eval/Actions Std 0.91423 +eval/Actions Max 0.999996 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67895 +time/logging (s) 0.00403933 +time/sampling batch (s) 0.282914 +time/saving (s) 0.00415708 +time/training (s) 6.79546 +time/epoch (s) 9.76552 +time/total (s) 1862.3 +Epoch -821 +---------------------------------- --------------- +2022-05-10 13:41:48.969768 PDT | [2] Epoch -820 finished +---------------------------------- --------------- +epoch -820 +replay_buffer/size 999033 +trainer/num train calls 181000 +trainer/Policy Loss -19.4291 +trainer/Log Pis Mean 24.4761 +trainer/Log Pis Std 13.1992 +trainer/Log Pis Max 77.666 +trainer/Log Pis Min -5.63131 +trainer/policy/mean Mean -0.0257012 +trainer/policy/mean Std 0.904585 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999983 +trainer/policy/normal/std Mean 2.86663 +trainer/policy/normal/std Std 0.579054 +trainer/policy/normal/std Max 6.07274 +trainer/policy/normal/std Min 0.484469 +trainer/policy/normal/log_std Mean 1.02541 +trainer/policy/normal/log_std Std 0.26071 +trainer/policy/normal/log_std Max 1.80381 +trainer/policy/normal/log_std Min -0.724701 +eval/num steps total 180893 +eval/num paths total 181 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0310931 +eval/Actions Std 0.939624 +eval/Actions Max 0.999991 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64498 +time/logging (s) 0.00382084 +time/sampling batch (s) 0.283094 +time/saving (s) 0.00348179 +time/training (s) 7.25748 +time/epoch (s) 10.1929 +time/total (s) 1872.49 +Epoch -820 +---------------------------------- --------------- +2022-05-10 13:41:58.624920 PDT | [2] Epoch -819 finished +---------------------------------- --------------- +epoch -819 +replay_buffer/size 999033 +trainer/num train calls 182000 +trainer/Policy Loss -20.5887 +trainer/Log Pis Mean 23.8651 +trainer/Log Pis Std 12.5072 +trainer/Log Pis Max 58.9128 +trainer/Log Pis Min -12.8557 +trainer/policy/mean Mean -0.0424216 +trainer/policy/mean Std 0.904879 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.80824 +trainer/policy/normal/std Std 0.588059 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.445993 +trainer/policy/normal/log_std Mean 1.00289 +trainer/policy/normal/log_std Std 0.269491 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -0.807451 +eval/num steps total 181627 +eval/num paths total 182 +eval/path length Mean 734 +eval/path length Std 0 +eval/path length Max 734 +eval/path length Min 734 +eval/Rewards Mean 0.0013624 +eval/Rewards Std 0.0368855 +eval/Rewards Max 1 +eval/Rewards Min 0 +eval/Returns Mean 1 +eval/Returns Std 0 +eval/Returns Max 1 +eval/Returns Min 1 +eval/Actions Mean 0.0202656 +eval/Actions Std 0.89661 +eval/Actions Max 0.999991 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 1 +eval/normalized_score 100 +time/evaluation sampling (s) 2.7775 +time/logging (s) 0.00299084 +time/sampling batch (s) 0.784197 +time/saving (s) 0.00348561 +time/training (s) 6.06564 +time/epoch (s) 9.63381 +time/total (s) 1882.13 +Epoch -819 +---------------------------------- --------------- +2022-05-10 13:42:07.808212 PDT | [2] Epoch -818 finished +---------------------------------- --------------- +epoch -818 +replay_buffer/size 999033 +trainer/num train calls 183000 +trainer/Policy Loss -18.3142 +trainer/Log Pis Mean 23.3609 +trainer/Log Pis Std 13.4 +trainer/Log Pis Max 67.5084 +trainer/Log Pis Min -6.92086 +trainer/policy/mean Mean -0.0459565 +trainer/policy/mean Std 0.90462 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.8619 +trainer/policy/normal/std Std 0.594753 +trainer/policy/normal/std Max 5.56341 +trainer/policy/normal/std Min 0.447865 +trainer/policy/normal/log_std Mean 1.02216 +trainer/policy/normal/log_std Std 0.268223 +trainer/policy/normal/log_std Max 1.71621 +trainer/policy/normal/log_std Min -0.803262 +eval/num steps total 182627 +eval/num paths total 183 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.130894 +eval/Actions Std 0.891528 +eval/Actions Max 0.999994 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6887 +time/logging (s) 0.00366966 +time/sampling batch (s) 0.282731 +time/saving (s) 0.00351208 +time/training (s) 6.18504 +time/epoch (s) 9.16365 +time/total (s) 1891.3 +Epoch -818 +---------------------------------- --------------- +2022-05-10 13:42:18.131046 PDT | [2] Epoch -817 finished +---------------------------------- --------------- +epoch -817 +replay_buffer/size 999033 +trainer/num train calls 184000 +trainer/Policy Loss -19.6675 +trainer/Log Pis Mean 25.0058 +trainer/Log Pis Std 13.1961 +trainer/Log Pis Max 62.7856 +trainer/Log Pis Min -11.2966 +trainer/policy/mean Mean -0.0496154 +trainer/policy/mean Std 0.907708 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.90544 +trainer/policy/normal/std Std 0.611449 +trainer/policy/normal/std Max 5.19764 +trainer/policy/normal/std Min 0.466473 +trainer/policy/normal/log_std Mean 1.03553 +trainer/policy/normal/log_std Std 0.27752 +trainer/policy/normal/log_std Max 1.6482 +trainer/policy/normal/log_std Min -0.762555 +eval/num steps total 183627 +eval/num paths total 184 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0713199 +eval/Actions Std 0.911004 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.43205 +time/logging (s) 0.0043136 +time/sampling batch (s) 0.28432 +time/saving (s) 0.00441895 +time/training (s) 7.57772 +time/epoch (s) 10.3028 +time/total (s) 1901.6 +Epoch -817 +---------------------------------- --------------- +2022-05-10 13:42:28.241813 PDT | [2] Epoch -816 finished +---------------------------------- --------------- +epoch -816 +replay_buffer/size 999033 +trainer/num train calls 185000 +trainer/Policy Loss -19.6974 +trainer/Log Pis Mean 25.2315 +trainer/Log Pis Std 12.9076 +trainer/Log Pis Max 69.8703 +trainer/Log Pis Min -7.92464 +trainer/policy/mean Mean -0.0441893 +trainer/policy/mean Std 0.904595 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.89367 +trainer/policy/normal/std Std 0.614565 +trainer/policy/normal/std Max 5.19657 +trainer/policy/normal/std Min 0.418572 +trainer/policy/normal/log_std Mean 1.03068 +trainer/policy/normal/log_std Std 0.281935 +trainer/policy/normal/log_std Max 1.648 +trainer/policy/normal/log_std Min -0.870907 +eval/num steps total 184627 +eval/num paths total 185 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0117354 +eval/Actions Std 0.890852 +eval/Actions Max 0.999995 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.71169 +time/logging (s) 0.00372563 +time/sampling batch (s) 0.281795 +time/saving (s) 0.00358159 +time/training (s) 7.0883 +time/epoch (s) 10.0891 +time/total (s) 1911.7 +Epoch -816 +---------------------------------- --------------- +2022-05-10 13:42:38.088682 PDT | [2] Epoch -815 finished +---------------------------------- --------------- +epoch -815 +replay_buffer/size 999033 +trainer/num train calls 186000 +trainer/Policy Loss -18.792 +trainer/Log Pis Mean 24.4502 +trainer/Log Pis Std 12.9304 +trainer/Log Pis Max 68.8769 +trainer/Log Pis Min -9.04693 +trainer/policy/mean Mean -0.0332014 +trainer/policy/mean Std 0.908614 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.82119 +trainer/policy/normal/std Std 0.577406 +trainer/policy/normal/std Max 5.75491 +trainer/policy/normal/std Min 0.394854 +trainer/policy/normal/log_std Mean 1.00835 +trainer/policy/normal/log_std Std 0.266879 +trainer/policy/normal/log_std Max 1.75005 +trainer/policy/normal/log_std Min -0.929239 +eval/num steps total 185627 +eval/num paths total 186 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0219823 +eval/Actions Std 0.879215 +eval/Actions Max 0.999996 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62181 +time/logging (s) 0.00422664 +time/sampling batch (s) 0.283173 +time/saving (s) 0.00431982 +time/training (s) 6.91335 +time/epoch (s) 9.82687 +time/total (s) 1921.53 +Epoch -815 +---------------------------------- --------------- +2022-05-10 13:42:47.623227 PDT | [2] Epoch -814 finished +---------------------------------- --------------- +epoch -814 +replay_buffer/size 999033 +trainer/num train calls 187000 +trainer/Policy Loss -18.6167 +trainer/Log Pis Mean 23.8933 +trainer/Log Pis Std 13.0289 +trainer/Log Pis Max 78.2283 +trainer/Log Pis Min -6.38972 +trainer/policy/mean Mean -0.0333725 +trainer/policy/mean Std 0.904957 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999983 +trainer/policy/normal/std Mean 2.88406 +trainer/policy/normal/std Std 0.605002 +trainer/policy/normal/std Max 6.11456 +trainer/policy/normal/std Min 0.456508 +trainer/policy/normal/log_std Mean 1.02969 +trainer/policy/normal/log_std Std 0.267819 +trainer/policy/normal/log_std Max 1.81067 +trainer/policy/normal/log_std Min -0.784149 +eval/num steps total 186627 +eval/num paths total 187 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.216267 +eval/Actions Std 0.945533 +eval/Actions Max 0.999995 +eval/Actions Min -0.99997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51645 +time/logging (s) 0.00381699 +time/sampling batch (s) 0.2828 +time/saving (s) 0.00358459 +time/training (s) 6.70641 +time/epoch (s) 9.51307 +time/total (s) 1931.04 +Epoch -814 +---------------------------------- --------------- +2022-05-10 13:42:57.383477 PDT | [2] Epoch -813 finished +---------------------------------- --------------- +epoch -813 +replay_buffer/size 999033 +trainer/num train calls 188000 +trainer/Policy Loss -19.0226 +trainer/Log Pis Mean 24.7391 +trainer/Log Pis Std 13.4855 +trainer/Log Pis Max 73.8342 +trainer/Log Pis Min -6.10677 +trainer/policy/mean Mean -0.0183 +trainer/policy/mean Std 0.906057 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.94763 +trainer/policy/normal/std Std 0.613081 +trainer/policy/normal/std Max 6.67441 +trainer/policy/normal/std Min 0.384345 +trainer/policy/normal/log_std Mean 1.05096 +trainer/policy/normal/log_std Std 0.27352 +trainer/policy/normal/log_std Max 1.89828 +trainer/policy/normal/log_std Min -0.956215 +eval/num steps total 187627 +eval/num paths total 188 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.379772 +eval/Actions Std 0.864541 +eval/Actions Max 0.999998 +eval/Actions Min -0.999982 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63037 +time/logging (s) 0.00368964 +time/sampling batch (s) 0.287774 +time/saving (s) 0.00350923 +time/training (s) 6.81404 +time/epoch (s) 9.73938 +time/total (s) 1940.78 +Epoch -813 +---------------------------------- --------------- +2022-05-10 13:43:07.492711 PDT | [2] Epoch -812 finished +---------------------------------- --------------- +epoch -812 +replay_buffer/size 999033 +trainer/num train calls 189000 +trainer/Policy Loss -19.8368 +trainer/Log Pis Mean 22.9772 +trainer/Log Pis Std 13.1102 +trainer/Log Pis Max 77.5998 +trainer/Log Pis Min -7.00843 +trainer/policy/mean Mean -0.0191463 +trainer/policy/mean Std 0.903011 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.85187 +trainer/policy/normal/std Std 0.572791 +trainer/policy/normal/std Max 5.13937 +trainer/policy/normal/std Min 0.334273 +trainer/policy/normal/log_std Mean 1.02013 +trainer/policy/normal/log_std Std 0.262402 +trainer/policy/normal/log_std Max 1.63693 +trainer/policy/normal/log_std Min -1.0958 +eval/num steps total 188627 +eval/num paths total 189 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0836146 +eval/Actions Std 0.897172 +eval/Actions Max 0.999993 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.4849 +time/logging (s) 0.00416137 +time/sampling batch (s) 0.283385 +time/saving (s) 0.00432795 +time/training (s) 7.31219 +time/epoch (s) 10.089 +time/total (s) 1950.88 +Epoch -812 +---------------------------------- --------------- +2022-05-10 13:43:18.140924 PDT | [2] Epoch -811 finished +---------------------------------- --------------- +epoch -811 +replay_buffer/size 999033 +trainer/num train calls 190000 +trainer/Policy Loss -18.4252 +trainer/Log Pis Mean 24.2609 +trainer/Log Pis Std 13.213 +trainer/Log Pis Max 59.9983 +trainer/Log Pis Min -5.24025 +trainer/policy/mean Mean -0.0285299 +trainer/policy/mean Std 0.902874 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.87285 +trainer/policy/normal/std Std 0.600243 +trainer/policy/normal/std Max 6.2361 +trainer/policy/normal/std Min 0.450979 +trainer/policy/normal/log_std Mean 1.02538 +trainer/policy/normal/log_std Std 0.271323 +trainer/policy/normal/log_std Max 1.83035 +trainer/policy/normal/log_std Min -0.796335 +eval/num steps total 189627 +eval/num paths total 190 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.371336 +eval/Actions Std 0.88271 +eval/Actions Max 0.999985 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60461 +time/logging (s) 0.00380239 +time/sampling batch (s) 0.282387 +time/saving (s) 0.00357314 +time/training (s) 7.73242 +time/epoch (s) 10.6268 +time/total (s) 1961.51 +Epoch -811 +---------------------------------- --------------- +2022-05-10 13:43:28.213418 PDT | [2] Epoch -810 finished +---------------------------------- --------------- +epoch -810 +replay_buffer/size 999033 +trainer/num train calls 191000 +trainer/Policy Loss -18.9838 +trainer/Log Pis Mean 24.7083 +trainer/Log Pis Std 13.4334 +trainer/Log Pis Max 68.4509 +trainer/Log Pis Min -6.5337 +trainer/policy/mean Mean -0.0433897 +trainer/policy/mean Std 0.908253 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.91502 +trainer/policy/normal/std Std 0.587898 +trainer/policy/normal/std Max 6.76975 +trainer/policy/normal/std Min 0.449597 +trainer/policy/normal/log_std Mean 1.04218 +trainer/policy/normal/log_std Std 0.26066 +trainer/policy/normal/log_std Max 1.91246 +trainer/policy/normal/log_std Min -0.799403 +eval/num steps total 190627 +eval/num paths total 191 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0740131 +eval/Actions Std 0.87302 +eval/Actions Max 0.999991 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67911 +time/logging (s) 0.00381659 +time/sampling batch (s) 0.27996 +time/saving (s) 0.00340368 +time/training (s) 7.0858 +time/epoch (s) 10.0521 +time/total (s) 1971.56 +Epoch -810 +---------------------------------- --------------- +2022-05-10 13:43:38.039664 PDT | [2] Epoch -809 finished +---------------------------------- --------------- +epoch -809 +replay_buffer/size 999033 +trainer/num train calls 192000 +trainer/Policy Loss -19.5595 +trainer/Log Pis Mean 24.768 +trainer/Log Pis Std 14.3046 +trainer/Log Pis Max 70.7328 +trainer/Log Pis Min -10.0227 +trainer/policy/mean Mean -0.0605121 +trainer/policy/mean Std 0.903578 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.86295 +trainer/policy/normal/std Std 0.587555 +trainer/policy/normal/std Max 5.15114 +trainer/policy/normal/std Min 0.429654 +trainer/policy/normal/log_std Mean 1.02229 +trainer/policy/normal/log_std Std 0.271158 +trainer/policy/normal/log_std Max 1.63922 +trainer/policy/normal/log_std Min -0.844774 +eval/num steps total 191627 +eval/num paths total 192 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00304135 +eval/Actions Std 0.897088 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.71955 +time/logging (s) 0.00419813 +time/sampling batch (s) 0.281192 +time/saving (s) 0.00440179 +time/training (s) 6.79679 +time/epoch (s) 9.80614 +time/total (s) 1981.37 +Epoch -809 +---------------------------------- --------------- +2022-05-10 13:43:48.328195 PDT | [2] Epoch -808 finished +---------------------------------- --------------- +epoch -808 +replay_buffer/size 999033 +trainer/num train calls 193000 +trainer/Policy Loss -19.3355 +trainer/Log Pis Mean 23.8754 +trainer/Log Pis Std 13.554 +trainer/Log Pis Max 64.3405 +trainer/Log Pis Min -7.44229 +trainer/policy/mean Mean -0.0212148 +trainer/policy/mean Std 0.906127 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.81376 +trainer/policy/normal/std Std 0.59944 +trainer/policy/normal/std Max 5.03973 +trainer/policy/normal/std Min 0.347302 +trainer/policy/normal/log_std Mean 1.00281 +trainer/policy/normal/log_std Std 0.280568 +trainer/policy/normal/log_std Max 1.61735 +trainer/policy/normal/log_std Min -1.05756 +eval/num steps total 192627 +eval/num paths total 193 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0190266 +eval/Actions Std 0.904993 +eval/Actions Max 0.999995 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53131 +time/logging (s) 0.00372389 +time/sampling batch (s) 0.537379 +time/saving (s) 0.00343976 +time/training (s) 7.19086 +time/epoch (s) 10.2667 +time/total (s) 1991.64 +Epoch -808 +---------------------------------- --------------- +2022-05-10 13:43:57.857492 PDT | [2] Epoch -807 finished +---------------------------------- --------------- +epoch -807 +replay_buffer/size 999033 +trainer/num train calls 194000 +trainer/Policy Loss -19.0298 +trainer/Log Pis Mean 23.608 +trainer/Log Pis Std 13.091 +trainer/Log Pis Max 67.8294 +trainer/Log Pis Min -6.5664 +trainer/policy/mean Mean -0.0319075 +trainer/policy/mean Std 0.901862 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.82935 +trainer/policy/normal/std Std 0.603144 +trainer/policy/normal/std Max 5.99792 +trainer/policy/normal/std Min 0.353363 +trainer/policy/normal/log_std Mean 1.00843 +trainer/policy/normal/log_std Std 0.279617 +trainer/policy/normal/log_std Max 1.79141 +trainer/policy/normal/log_std Min -1.04026 +eval/num steps total 193627 +eval/num paths total 194 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0153111 +eval/Actions Std 0.947364 +eval/Actions Max 0.999982 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60049 +time/logging (s) 0.00378219 +time/sampling batch (s) 0.278924 +time/saving (s) 0.00346618 +time/training (s) 6.62252 +time/epoch (s) 9.50919 +time/total (s) 2001.15 +Epoch -807 +---------------------------------- --------------- +2022-05-10 13:44:08.688342 PDT | [2] Epoch -806 finished +---------------------------------- --------------- +epoch -806 +replay_buffer/size 999033 +trainer/num train calls 195000 +trainer/Policy Loss -19.8194 +trainer/Log Pis Mean 25.7273 +trainer/Log Pis Std 13.7175 +trainer/Log Pis Max 68.0124 +trainer/Log Pis Min -7.2958 +trainer/policy/mean Mean -0.0435965 +trainer/policy/mean Std 0.905328 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.89731 +trainer/policy/normal/std Std 0.58518 +trainer/policy/normal/std Max 5.32975 +trainer/policy/normal/std Min 0.394505 +trainer/policy/normal/log_std Mean 1.03535 +trainer/policy/normal/log_std Std 0.266109 +trainer/policy/normal/log_std Max 1.6733 +trainer/policy/normal/log_std Min -0.930122 +eval/num steps total 194627 +eval/num paths total 195 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.132694 +eval/Actions Std 0.879474 +eval/Actions Max 0.999988 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.7394 +time/logging (s) 0.00422714 +time/sampling batch (s) 0.530459 +time/saving (s) 0.00434835 +time/training (s) 7.53246 +time/epoch (s) 10.8109 +time/total (s) 2011.97 +Epoch -806 +---------------------------------- --------------- +2022-05-10 13:44:18.489540 PDT | [2] Epoch -805 finished +---------------------------------- --------------- +epoch -805 +replay_buffer/size 999033 +trainer/num train calls 196000 +trainer/Policy Loss -19.7969 +trainer/Log Pis Mean 25.2874 +trainer/Log Pis Std 13.0543 +trainer/Log Pis Max 74.8993 +trainer/Log Pis Min -5.20973 +trainer/policy/mean Mean -0.0215817 +trainer/policy/mean Std 0.905089 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.89214 +trainer/policy/normal/std Std 0.607177 +trainer/policy/normal/std Max 5.89183 +trainer/policy/normal/std Min 0.355708 +trainer/policy/normal/log_std Mean 1.02963 +trainer/policy/normal/log_std Std 0.288598 +trainer/policy/normal/log_std Max 1.77357 +trainer/policy/normal/log_std Min -1.03364 +eval/num steps total 195627 +eval/num paths total 196 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00378915 +eval/Actions Std 0.965188 +eval/Actions Max 0.99997 +eval/Actions Min -0.999988 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50301 +time/logging (s) 0.00372598 +time/sampling batch (s) 0.52937 +time/saving (s) 0.00356207 +time/training (s) 6.74017 +time/epoch (s) 9.77983 +time/total (s) 2021.75 +Epoch -805 +---------------------------------- --------------- +2022-05-10 13:44:29.408032 PDT | [2] Epoch -804 finished +---------------------------------- --------------- +epoch -804 +replay_buffer/size 999033 +trainer/num train calls 197000 +trainer/Policy Loss -20.0956 +trainer/Log Pis Mean 24.4628 +trainer/Log Pis Std 13.3655 +trainer/Log Pis Max 68.7522 +trainer/Log Pis Min -3.28781 +trainer/policy/mean Mean -0.0380255 +trainer/policy/mean Std 0.912414 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.87785 +trainer/policy/normal/std Std 0.573597 +trainer/policy/normal/std Max 5.84087 +trainer/policy/normal/std Min 0.410348 +trainer/policy/normal/log_std Mean 1.03008 +trainer/policy/normal/log_std Std 0.256846 +trainer/policy/normal/log_std Max 1.76488 +trainer/policy/normal/log_std Min -0.890749 +eval/num steps total 196627 +eval/num paths total 197 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.014623 +eval/Actions Std 0.952431 +eval/Actions Max 0.999989 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.83157 +time/logging (s) 0.00371989 +time/sampling batch (s) 0.280081 +time/saving (s) 0.00344533 +time/training (s) 7.77949 +time/epoch (s) 10.8983 +time/total (s) 2032.65 +Epoch -804 +---------------------------------- --------------- +2022-05-10 13:44:38.922508 PDT | [2] Epoch -803 finished +---------------------------------- --------------- +epoch -803 +replay_buffer/size 999033 +trainer/num train calls 198000 +trainer/Policy Loss -19.044 +trainer/Log Pis Mean 24.8759 +trainer/Log Pis Std 14.2517 +trainer/Log Pis Max 78.8955 +trainer/Log Pis Min -8.84209 +trainer/policy/mean Mean -0.0367311 +trainer/policy/mean Std 0.907134 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.83111 +trainer/policy/normal/std Std 0.595811 +trainer/policy/normal/std Max 5.96903 +trainer/policy/normal/std Min 0.448535 +trainer/policy/normal/log_std Mean 1.00858 +trainer/policy/normal/log_std Std 0.284613 +trainer/policy/normal/log_std Max 1.78658 +trainer/policy/normal/log_std Min -0.801769 +eval/num steps total 197627 +eval/num paths total 198 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.351476 +eval/Actions Std 0.904899 +eval/Actions Max 0.999994 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65267 +time/logging (s) 0.00417422 +time/sampling batch (s) 0.530387 +time/saving (s) 0.00421495 +time/training (s) 6.30306 +time/epoch (s) 9.49451 +time/total (s) 2042.15 +Epoch -803 +---------------------------------- --------------- +2022-05-10 13:44:48.527213 PDT | [2] Epoch -802 finished +---------------------------------- --------------- +epoch -802 +replay_buffer/size 999033 +trainer/num train calls 199000 +trainer/Policy Loss -19.8295 +trainer/Log Pis Mean 24.8716 +trainer/Log Pis Std 13.1375 +trainer/Log Pis Max 66.3814 +trainer/Log Pis Min -5.35943 +trainer/policy/mean Mean -0.0262076 +trainer/policy/mean Std 0.909132 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.87105 +trainer/policy/normal/std Std 0.582913 +trainer/policy/normal/std Max 5.78994 +trainer/policy/normal/std Min 0.43853 +trainer/policy/normal/log_std Mean 1.02697 +trainer/policy/normal/log_std Std 0.259544 +trainer/policy/normal/log_std Max 1.75612 +trainer/policy/normal/log_std Min -0.824327 +eval/num steps total 198627 +eval/num paths total 199 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.12188 +eval/Actions Std 0.914248 +eval/Actions Max 0.999992 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62227 +time/logging (s) 0.00447766 +time/sampling batch (s) 0.28608 +time/saving (s) 0.00359926 +time/training (s) 6.66745 +time/epoch (s) 9.58387 +time/total (s) 2051.74 +Epoch -802 +---------------------------------- --------------- +2022-05-10 13:44:57.984544 PDT | [2] Epoch -801 finished +---------------------------------- --------------- +epoch -801 +replay_buffer/size 999033 +trainer/num train calls 200000 +trainer/Policy Loss -19.9365 +trainer/Log Pis Mean 24.7102 +trainer/Log Pis Std 12.4354 +trainer/Log Pis Max 61.7975 +trainer/Log Pis Min -5.72665 +trainer/policy/mean Mean -0.0298113 +trainer/policy/mean Std 0.908479 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.83812 +trainer/policy/normal/std Std 0.590693 +trainer/policy/normal/std Max 5.70186 +trainer/policy/normal/std Min 0.395942 +trainer/policy/normal/log_std Mean 1.01309 +trainer/policy/normal/log_std Std 0.272495 +trainer/policy/normal/log_std Max 1.74079 +trainer/policy/normal/log_std Min -0.926488 +eval/num steps total 199627 +eval/num paths total 200 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.137059 +eval/Actions Std 0.885054 +eval/Actions Max 0.999999 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64567 +time/logging (s) 0.00376263 +time/sampling batch (s) 0.278702 +time/saving (s) 0.00659284 +time/training (s) 6.50105 +time/epoch (s) 9.43578 +time/total (s) 2061.17 +Epoch -801 +---------------------------------- --------------- +2022-05-10 13:45:07.774829 PDT | [2] Epoch -800 finished +---------------------------------- --------------- +epoch -800 +replay_buffer/size 999033 +trainer/num train calls 201000 +trainer/Policy Loss -19.411 +trainer/Log Pis Mean 24.6256 +trainer/Log Pis Std 13.2499 +trainer/Log Pis Max 69.3701 +trainer/Log Pis Min -6.90926 +trainer/policy/mean Mean -0.0419977 +trainer/policy/mean Std 0.910219 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999985 +trainer/policy/normal/std Mean 2.82503 +trainer/policy/normal/std Std 0.586658 +trainer/policy/normal/std Max 5.2695 +trainer/policy/normal/std Min 0.388048 +trainer/policy/normal/log_std Mean 1.00819 +trainer/policy/normal/log_std Std 0.275593 +trainer/policy/normal/log_std Max 1.66194 +trainer/policy/normal/log_std Min -0.946627 +eval/num steps total 200627 +eval/num paths total 201 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.400215 +eval/Actions Std 0.889485 +eval/Actions Max 0.999998 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.91786 +time/logging (s) 0.0037945 +time/sampling batch (s) 0.280194 +time/saving (s) 0.00344607 +time/training (s) 6.56452 +time/epoch (s) 9.76981 +time/total (s) 2070.95 +Epoch -800 +---------------------------------- --------------- +2022-05-10 13:45:17.340743 PDT | [2] Epoch -799 finished +---------------------------------- --------------- +epoch -799 +replay_buffer/size 999033 +trainer/num train calls 202000 +trainer/Policy Loss -20.4159 +trainer/Log Pis Mean 25.1247 +trainer/Log Pis Std 13.2431 +trainer/Log Pis Max 73.518 +trainer/Log Pis Min -9.65322 +trainer/policy/mean Mean -0.0529376 +trainer/policy/mean Std 0.904699 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.84574 +trainer/policy/normal/std Std 0.610965 +trainer/policy/normal/std Max 5.55227 +trainer/policy/normal/std Min 0.393268 +trainer/policy/normal/log_std Mean 1.01305 +trainer/policy/normal/log_std Std 0.287237 +trainer/policy/normal/log_std Max 1.71421 +trainer/policy/normal/log_std Min -0.933264 +eval/num steps total 201627 +eval/num paths total 202 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00323674 +eval/Actions Std 0.901397 +eval/Actions Max 0.999995 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.76446 +time/logging (s) 0.003758 +time/sampling batch (s) 0.278408 +time/saving (s) 0.00346196 +time/training (s) 6.49536 +time/epoch (s) 9.54545 +time/total (s) 2080.5 +Epoch -799 +---------------------------------- --------------- +2022-05-10 13:45:27.733808 PDT | [2] Epoch -798 finished +---------------------------------- --------------- +epoch -798 +replay_buffer/size 999033 +trainer/num train calls 203000 +trainer/Policy Loss -19.985 +trainer/Log Pis Mean 24.8652 +trainer/Log Pis Std 13.2557 +trainer/Log Pis Max 73.8331 +trainer/Log Pis Min -10.8595 +trainer/policy/mean Mean -0.039096 +trainer/policy/mean Std 0.907211 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.76923 +trainer/policy/normal/std Std 0.587101 +trainer/policy/normal/std Max 5.21307 +trainer/policy/normal/std Min 0.398214 +trainer/policy/normal/log_std Mean 0.986309 +trainer/policy/normal/log_std Std 0.284657 +trainer/policy/normal/log_std Max 1.65117 +trainer/policy/normal/log_std Min -0.920765 +eval/num steps total 202627 +eval/num paths total 203 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0394939 +eval/Actions Std 0.930869 +eval/Actions Max 0.999997 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68326 +time/logging (s) 0.00418996 +time/sampling batch (s) 0.281934 +time/saving (s) 0.00396752 +time/training (s) 7.39944 +time/epoch (s) 10.3728 +time/total (s) 2090.87 +Epoch -798 +---------------------------------- --------------- +2022-05-10 13:45:37.086258 PDT | [2] Epoch -797 finished +---------------------------------- --------------- +epoch -797 +replay_buffer/size 999033 +trainer/num train calls 204000 +trainer/Policy Loss -19.5854 +trainer/Log Pis Mean 24.5318 +trainer/Log Pis Std 13.0892 +trainer/Log Pis Max 62.2969 +trainer/Log Pis Min -10.0998 +trainer/policy/mean Mean -0.0179543 +trainer/policy/mean Std 0.908373 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.86436 +trainer/policy/normal/std Std 0.585416 +trainer/policy/normal/std Max 5.68654 +trainer/policy/normal/std Min 0.410749 +trainer/policy/normal/log_std Mean 1.02468 +trainer/policy/normal/log_std Std 0.258825 +trainer/policy/normal/log_std Max 1.7381 +trainer/policy/normal/log_std Min -0.889772 +eval/num steps total 203627 +eval/num paths total 204 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.303742 +eval/Actions Std 0.852212 +eval/Actions Max 0.999988 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68385 +time/logging (s) 0.00387173 +time/sampling batch (s) 0.282742 +time/saving (s) 0.00372496 +time/training (s) 6.35716 +time/epoch (s) 9.33135 +time/total (s) 2100.21 +Epoch -797 +---------------------------------- --------------- +2022-05-10 13:45:46.868156 PDT | [2] Epoch -796 finished +---------------------------------- --------------- +epoch -796 +replay_buffer/size 999033 +trainer/num train calls 205000 +trainer/Policy Loss -20.054 +trainer/Log Pis Mean 25.1692 +trainer/Log Pis Std 13.0585 +trainer/Log Pis Max 73.408 +trainer/Log Pis Min -4.45038 +trainer/policy/mean Mean -0.0415453 +trainer/policy/mean Std 0.907527 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.86627 +trainer/policy/normal/std Std 0.593976 +trainer/policy/normal/std Max 4.9917 +trainer/policy/normal/std Min 0.393523 +trainer/policy/normal/log_std Mean 1.02266 +trainer/policy/normal/log_std Std 0.275686 +trainer/policy/normal/log_std Max 1.60778 +trainer/policy/normal/log_std Min -0.932616 +eval/num steps total 204627 +eval/num paths total 205 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.106933 +eval/Actions Std 0.902557 +eval/Actions Max 0.999989 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.31886 +time/logging (s) 0.00378157 +time/sampling batch (s) 0.280626 +time/saving (s) 0.00345209 +time/training (s) 7.15448 +time/epoch (s) 9.76121 +time/total (s) 2109.97 +Epoch -796 +---------------------------------- --------------- +2022-05-10 13:45:58.008108 PDT | [2] Epoch -795 finished +---------------------------------- --------------- +epoch -795 +replay_buffer/size 999033 +trainer/num train calls 206000 +trainer/Policy Loss -19.7114 +trainer/Log Pis Mean 24.5817 +trainer/Log Pis Std 12.6725 +trainer/Log Pis Max 71.7205 +trainer/Log Pis Min -10.2191 +trainer/policy/mean Mean -0.022837 +trainer/policy/mean Std 0.913139 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.77949 +trainer/policy/normal/std Std 0.607252 +trainer/policy/normal/std Max 5.21953 +trainer/policy/normal/std Min 0.384219 +trainer/policy/normal/log_std Mean 0.987984 +trainer/policy/normal/log_std Std 0.293929 +trainer/policy/normal/log_std Max 1.65241 +trainer/policy/normal/log_std Min -0.956543 +eval/num steps total 205627 +eval/num paths total 206 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0391734 +eval/Actions Std 0.912726 +eval/Actions Max 0.999998 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.49986 +time/logging (s) 0.00379984 +time/sampling batch (s) 0.280575 +time/saving (s) 0.00348135 +time/training (s) 8.33178 +time/epoch (s) 11.1195 +time/total (s) 2121.09 +Epoch -795 +---------------------------------- --------------- +2022-05-10 13:46:08.478694 PDT | [2] Epoch -794 finished +---------------------------------- --------------- +epoch -794 +replay_buffer/size 999033 +trainer/num train calls 207000 +trainer/Policy Loss -19.0348 +trainer/Log Pis Mean 24.5216 +trainer/Log Pis Std 12.8521 +trainer/Log Pis Max 72.9491 +trainer/Log Pis Min -7.39597 +trainer/policy/mean Mean -0.0317086 +trainer/policy/mean Std 0.90657 +trainer/policy/mean Max 0.999977 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.86215 +trainer/policy/normal/std Std 0.602791 +trainer/policy/normal/std Max 5.0709 +trainer/policy/normal/std Min 0.384034 +trainer/policy/normal/log_std Mean 1.02063 +trainer/policy/normal/log_std Std 0.277172 +trainer/policy/normal/log_std Max 1.62352 +trainer/policy/normal/log_std Min -0.957025 +eval/num steps total 206627 +eval/num paths total 207 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0389429 +eval/Actions Std 0.901138 +eval/Actions Max 0.999995 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70352 +time/logging (s) 0.00368222 +time/sampling batch (s) 0.280252 +time/saving (s) 0.00343482 +time/training (s) 7.4591 +time/epoch (s) 10.45 +time/total (s) 2131.55 +Epoch -794 +---------------------------------- --------------- +2022-05-10 13:46:18.612435 PDT | [2] Epoch -793 finished +---------------------------------- --------------- +epoch -793 +replay_buffer/size 999033 +trainer/num train calls 208000 +trainer/Policy Loss -19.5659 +trainer/Log Pis Mean 24.2576 +trainer/Log Pis Std 12.6774 +trainer/Log Pis Max 70.1987 +trainer/Log Pis Min -7.64376 +trainer/policy/mean Mean -0.0337149 +trainer/policy/mean Std 0.901499 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.85049 +trainer/policy/normal/std Std 0.604641 +trainer/policy/normal/std Max 6.52882 +trainer/policy/normal/std Min 0.440399 +trainer/policy/normal/log_std Mean 1.01649 +trainer/policy/normal/log_std Std 0.277436 +trainer/policy/normal/log_std Max 1.87623 +trainer/policy/normal/log_std Min -0.820075 +eval/num steps total 207627 +eval/num paths total 208 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.227333 +eval/Actions Std 0.934746 +eval/Actions Max 1 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61578 +time/logging (s) 0.0041238 +time/sampling batch (s) 0.279203 +time/saving (s) 0.00415357 +time/training (s) 7.21038 +time/epoch (s) 10.1136 +time/total (s) 2141.66 +Epoch -793 +---------------------------------- --------------- +2022-05-10 13:46:28.612298 PDT | [2] Epoch -792 finished +---------------------------------- --------------- +epoch -792 +replay_buffer/size 999033 +trainer/num train calls 209000 +trainer/Policy Loss -19.4356 +trainer/Log Pis Mean 23.7457 +trainer/Log Pis Std 13.6316 +trainer/Log Pis Max 67.1792 +trainer/Log Pis Min -12.0068 +trainer/policy/mean Mean -0.0349958 +trainer/policy/mean Std 0.905658 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.89011 +trainer/policy/normal/std Std 0.601052 +trainer/policy/normal/std Max 5.20639 +trainer/policy/normal/std Min 0.342932 +trainer/policy/normal/log_std Mean 1.03067 +trainer/policy/normal/log_std Std 0.277926 +trainer/policy/normal/log_std Max 1.64989 +trainer/policy/normal/log_std Min -1.07022 +eval/num steps total 208627 +eval/num paths total 209 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.014607 +eval/Actions Std 0.968982 +eval/Actions Max 0.999993 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.77055 +time/logging (s) 0.00418934 +time/sampling batch (s) 0.288156 +time/saving (s) 0.00443102 +time/training (s) 6.91133 +time/epoch (s) 9.97866 +time/total (s) 2151.64 +Epoch -792 +---------------------------------- --------------- +2022-05-10 13:46:38.452282 PDT | [2] Epoch -791 finished +---------------------------------- --------------- +epoch -791 +replay_buffer/size 999033 +trainer/num train calls 210000 +trainer/Policy Loss -19.8586 +trainer/Log Pis Mean 24.8895 +trainer/Log Pis Std 13.6315 +trainer/Log Pis Max 69.8258 +trainer/Log Pis Min -6.92531 +trainer/policy/mean Mean -0.0318554 +trainer/policy/mean Std 0.905426 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.84019 +trainer/policy/normal/std Std 0.613953 +trainer/policy/normal/std Max 5.59941 +trainer/policy/normal/std Min 0.397168 +trainer/policy/normal/log_std Mean 1.00973 +trainer/policy/normal/log_std Std 0.295419 +trainer/policy/normal/log_std Max 1.72266 +trainer/policy/normal/log_std Min -0.923396 +eval/num steps total 209627 +eval/num paths total 210 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.238971 +eval/Actions Std 0.942572 +eval/Actions Max 0.99999 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68203 +time/logging (s) 0.00417264 +time/sampling batch (s) 0.286021 +time/saving (s) 0.00508217 +time/training (s) 6.84122 +time/epoch (s) 9.81852 +time/total (s) 2161.47 +Epoch -791 +---------------------------------- --------------- +2022-05-10 13:46:48.535616 PDT | [2] Epoch -790 finished +---------------------------------- --------------- +epoch -790 +replay_buffer/size 999033 +trainer/num train calls 211000 +trainer/Policy Loss -20.0915 +trainer/Log Pis Mean 25.5232 +trainer/Log Pis Std 13.2274 +trainer/Log Pis Max 66.1558 +trainer/Log Pis Min -8.33441 +trainer/policy/mean Mean -0.026128 +trainer/policy/mean Std 0.907289 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.80672 +trainer/policy/normal/std Std 0.602077 +trainer/policy/normal/std Max 5.18914 +trainer/policy/normal/std Min 0.427693 +trainer/policy/normal/log_std Mean 0.999423 +trainer/policy/normal/log_std Std 0.28513 +trainer/policy/normal/log_std Max 1.64657 +trainer/policy/normal/log_std Min -0.84935 +eval/num steps total 210627 +eval/num paths total 211 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.180544 +eval/Actions Std 0.912393 +eval/Actions Max 0.999996 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61005 +time/logging (s) 0.00382378 +time/sampling batch (s) 0.282584 +time/saving (s) 0.00365424 +time/training (s) 7.16073 +time/epoch (s) 10.0608 +time/total (s) 2171.53 +Epoch -790 +---------------------------------- --------------- +2022-05-10 13:46:58.356328 PDT | [2] Epoch -789 finished +---------------------------------- --------------- +epoch -789 +replay_buffer/size 999033 +trainer/num train calls 212000 +trainer/Policy Loss -20.1971 +trainer/Log Pis Mean 24.1101 +trainer/Log Pis Std 13.5324 +trainer/Log Pis Max 67.4014 +trainer/Log Pis Min -6.24252 +trainer/policy/mean Mean -0.0485207 +trainer/policy/mean Std 0.907135 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.87801 +trainer/policy/normal/std Std 0.612942 +trainer/policy/normal/std Max 5.57473 +trainer/policy/normal/std Min 0.419898 +trainer/policy/normal/log_std Mean 1.02587 +trainer/policy/normal/log_std Std 0.277908 +trainer/policy/normal/log_std Max 1.71824 +trainer/policy/normal/log_std Min -0.867744 +eval/num steps total 211627 +eval/num paths total 212 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0948161 +eval/Actions Std 0.903722 +eval/Actions Max 0.99999 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58743 +time/logging (s) 0.00423533 +time/sampling batch (s) 0.28432 +time/saving (s) 0.00441381 +time/training (s) 6.91976 +time/epoch (s) 9.80015 +time/total (s) 2181.33 +Epoch -789 +---------------------------------- --------------- +2022-05-10 13:47:08.396544 PDT | [2] Epoch -788 finished +---------------------------------- --------------- +epoch -788 +replay_buffer/size 999033 +trainer/num train calls 213000 +trainer/Policy Loss -19.5473 +trainer/Log Pis Mean 25.059 +trainer/Log Pis Std 13.1502 +trainer/Log Pis Max 69.1303 +trainer/Log Pis Min -11.3172 +trainer/policy/mean Mean -0.0406637 +trainer/policy/mean Std 0.906442 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.88343 +trainer/policy/normal/std Std 0.582807 +trainer/policy/normal/std Max 5.42201 +trainer/policy/normal/std Min 0.398567 +trainer/policy/normal/log_std Mean 1.03169 +trainer/policy/normal/log_std Std 0.256761 +trainer/policy/normal/log_std Max 1.69047 +trainer/policy/normal/log_std Min -0.919879 +eval/num steps total 212627 +eval/num paths total 213 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0718642 +eval/Actions Std 0.928853 +eval/Actions Max 0.999996 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.42661 +time/logging (s) 0.00375621 +time/sampling batch (s) 0.284019 +time/saving (s) 0.00357415 +time/training (s) 7.30077 +time/epoch (s) 10.0187 +time/total (s) 2191.36 +Epoch -788 +---------------------------------- --------------- +2022-05-10 13:47:19.369930 PDT | [2] Epoch -787 finished +---------------------------------- --------------- +epoch -787 +replay_buffer/size 999033 +trainer/num train calls 214000 +trainer/Policy Loss -20.5241 +trainer/Log Pis Mean 25.7462 +trainer/Log Pis Std 13.1668 +trainer/Log Pis Max 83.4424 +trainer/Log Pis Min -8.93383 +trainer/policy/mean Mean -0.0358769 +trainer/policy/mean Std 0.905754 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.91282 +trainer/policy/normal/std Std 0.596624 +trainer/policy/normal/std Max 5.55192 +trainer/policy/normal/std Min 0.357438 +trainer/policy/normal/log_std Mean 1.03942 +trainer/policy/normal/log_std Std 0.272814 +trainer/policy/normal/log_std Max 1.71414 +trainer/policy/normal/log_std Min -1.02879 +eval/num steps total 213627 +eval/num paths total 214 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0752911 +eval/Actions Std 0.913599 +eval/Actions Max 0.999999 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66461 +time/logging (s) 0.00370469 +time/sampling batch (s) 0.287406 +time/saving (s) 0.0034771 +time/training (s) 7.9931 +time/epoch (s) 10.9523 +time/total (s) 2202.31 +Epoch -787 +---------------------------------- --------------- +2022-05-10 13:47:30.513391 PDT | [2] Epoch -786 finished +---------------------------------- --------------- +epoch -786 +replay_buffer/size 999033 +trainer/num train calls 215000 +trainer/Policy Loss -19.582 +trainer/Log Pis Mean 24.5582 +trainer/Log Pis Std 12.505 +trainer/Log Pis Max 64.3133 +trainer/Log Pis Min -2.79329 +trainer/policy/mean Mean -0.0497611 +trainer/policy/mean Std 0.90199 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.87262 +trainer/policy/normal/std Std 0.619001 +trainer/policy/normal/std Max 5.51811 +trainer/policy/normal/std Min 0.414414 +trainer/policy/normal/log_std Mean 1.02321 +trainer/policy/normal/log_std Std 0.281359 +trainer/policy/normal/log_std Max 1.70804 +trainer/policy/normal/log_std Min -0.88089 +eval/num steps total 214627 +eval/num paths total 215 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0642215 +eval/Actions Std 0.902056 +eval/Actions Max 1 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6091 +time/logging (s) 0.00402501 +time/sampling batch (s) 0.282823 +time/saving (s) 0.00408279 +time/training (s) 8.22291 +time/epoch (s) 11.1229 +time/total (s) 2213.44 +Epoch -786 +---------------------------------- --------------- +2022-05-10 13:47:40.329062 PDT | [2] Epoch -785 finished +---------------------------------- --------------- +epoch -785 +replay_buffer/size 999033 +trainer/num train calls 216000 +trainer/Policy Loss -20.1477 +trainer/Log Pis Mean 24.9585 +trainer/Log Pis Std 12.9802 +trainer/Log Pis Max 66.4415 +trainer/Log Pis Min -5.84092 +trainer/policy/mean Mean -0.0362301 +trainer/policy/mean Std 0.902355 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.8181 +trainer/policy/normal/std Std 0.58747 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.35537 +trainer/policy/normal/log_std Mean 1.00663 +trainer/policy/normal/log_std Std 0.268964 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.0346 +eval/num steps total 215627 +eval/num paths total 216 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.239507 +eval/Actions Std 0.93196 +eval/Actions Max 0.999976 +eval/Actions Min -0.999966 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58602 +time/logging (s) 0.0037141 +time/sampling batch (s) 0.282054 +time/saving (s) 0.00349715 +time/training (s) 6.91907 +time/epoch (s) 9.79435 +time/total (s) 2223.24 +Epoch -785 +---------------------------------- --------------- +2022-05-10 13:47:50.908071 PDT | [2] Epoch -784 finished +---------------------------------- --------------- +epoch -784 +replay_buffer/size 999033 +trainer/num train calls 217000 +trainer/Policy Loss -19.941 +trainer/Log Pis Mean 24.0473 +trainer/Log Pis Std 12.3294 +trainer/Log Pis Max 67.4553 +trainer/Log Pis Min -7.31595 +trainer/policy/mean Mean -0.0513994 +trainer/policy/mean Std 0.906344 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.84483 +trainer/policy/normal/std Std 0.624224 +trainer/policy/normal/std Max 7.32574 +trainer/policy/normal/std Min 0.41944 +trainer/policy/normal/log_std Mean 1.01195 +trainer/policy/normal/log_std Std 0.288746 +trainer/policy/normal/log_std Max 1.99139 +trainer/policy/normal/log_std Min -0.868835 +eval/num steps total 216627 +eval/num paths total 217 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.270407 +eval/Actions Std 0.855525 +eval/Actions Max 0.999997 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6177 +time/logging (s) 0.00374009 +time/sampling batch (s) 0.785621 +time/saving (s) 0.0034738 +time/training (s) 7.14757 +time/epoch (s) 10.5581 +time/total (s) 2233.8 +Epoch -784 +---------------------------------- --------------- +2022-05-10 13:48:01.285554 PDT | [2] Epoch -783 finished +---------------------------------- --------------- +epoch -783 +replay_buffer/size 999033 +trainer/num train calls 218000 +trainer/Policy Loss -18.5746 +trainer/Log Pis Mean 24.1136 +trainer/Log Pis Std 13.3593 +trainer/Log Pis Max 84.9825 +trainer/Log Pis Min -9.78972 +trainer/policy/mean Mean -0.0506126 +trainer/policy/mean Std 0.90433 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.86079 +trainer/policy/normal/std Std 0.630439 +trainer/policy/normal/std Max 6.73762 +trainer/policy/normal/std Min 0.412641 +trainer/policy/normal/log_std Mean 1.01612 +trainer/policy/normal/log_std Std 0.297932 +trainer/policy/normal/log_std Max 1.90771 +trainer/policy/normal/log_std Min -0.885177 +eval/num steps total 217627 +eval/num paths total 218 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0774155 +eval/Actions Std 0.912981 +eval/Actions Max 0.999996 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59291 +time/logging (s) 0.00377669 +time/sampling batch (s) 0.283094 +time/saving (s) 0.00347787 +time/training (s) 7.47367 +time/epoch (s) 10.3569 +time/total (s) 2244.16 +Epoch -783 +---------------------------------- --------------- +2022-05-10 13:48:11.622554 PDT | [2] Epoch -782 finished +---------------------------------- --------------- +epoch -782 +replay_buffer/size 999033 +trainer/num train calls 219000 +trainer/Policy Loss -19.5306 +trainer/Log Pis Mean 25.1081 +trainer/Log Pis Std 13.2124 +trainer/Log Pis Max 70.6973 +trainer/Log Pis Min -5.29347 +trainer/policy/mean Mean -0.0432219 +trainer/policy/mean Std 0.909618 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.85674 +trainer/policy/normal/std Std 0.603387 +trainer/policy/normal/std Max 4.99498 +trainer/policy/normal/std Min 0.433963 +trainer/policy/normal/log_std Mean 1.01897 +trainer/policy/normal/log_std Std 0.275403 +trainer/policy/normal/log_std Max 1.60843 +trainer/policy/normal/log_std Min -0.834795 +eval/num steps total 218627 +eval/num paths total 219 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0986607 +eval/Actions Std 0.914922 +eval/Actions Max 0.99999 +eval/Actions Min -0.999988 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59046 +time/logging (s) 0.00377621 +time/sampling batch (s) 0.283089 +time/saving (s) 0.00345053 +time/training (s) 7.43537 +time/epoch (s) 10.3161 +time/total (s) 2254.48 +Epoch -782 +---------------------------------- --------------- +2022-05-10 13:48:21.844859 PDT | [2] Epoch -781 finished +---------------------------------- --------------- +epoch -781 +replay_buffer/size 999033 +trainer/num train calls 220000 +trainer/Policy Loss -18.8734 +trainer/Log Pis Mean 24.5177 +trainer/Log Pis Std 13.63 +trainer/Log Pis Max 74.5358 +trainer/Log Pis Min -6.05095 +trainer/policy/mean Mean -0.0235076 +trainer/policy/mean Std 0.906469 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.9017 +trainer/policy/normal/std Std 0.622768 +trainer/policy/normal/std Max 5.66384 +trainer/policy/normal/std Min 0.437634 +trainer/policy/normal/log_std Mean 1.03472 +trainer/policy/normal/log_std Std 0.271886 +trainer/policy/normal/log_std Max 1.7341 +trainer/policy/normal/log_std Min -0.826372 +eval/num steps total 219627 +eval/num paths total 220 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.294267 +eval/Actions Std 0.880163 +eval/Actions Max 0.999991 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.49378 +time/logging (s) 0.00397587 +time/sampling batch (s) 0.286619 +time/saving (s) 0.00418545 +time/training (s) 7.41297 +time/epoch (s) 10.2015 +time/total (s) 2264.68 +Epoch -781 +---------------------------------- --------------- +2022-05-10 13:48:32.232159 PDT | [2] Epoch -780 finished +---------------------------------- --------------- +epoch -780 +replay_buffer/size 999033 +trainer/num train calls 221000 +trainer/Policy Loss -18.8857 +trainer/Log Pis Mean 23.1891 +trainer/Log Pis Std 12.5774 +trainer/Log Pis Max 62.4582 +trainer/Log Pis Min -9.82652 +trainer/policy/mean Mean -0.0266356 +trainer/policy/mean Std 0.905726 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.85272 +trainer/policy/normal/std Std 0.587688 +trainer/policy/normal/std Max 5.44636 +trainer/policy/normal/std Min 0.3739 +trainer/policy/normal/log_std Mean 1.019 +trainer/policy/normal/log_std Std 0.269383 +trainer/policy/normal/log_std Max 1.69495 +trainer/policy/normal/log_std Min -0.983766 +eval/num steps total 220627 +eval/num paths total 221 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.479556 +eval/Actions Std 0.871291 +eval/Actions Max 1 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63165 +time/logging (s) 0.00373707 +time/sampling batch (s) 0.290426 +time/saving (s) 0.0034825 +time/training (s) 7.43601 +time/epoch (s) 10.3653 +time/total (s) 2275.05 +Epoch -780 +---------------------------------- --------------- +2022-05-10 13:48:42.175629 PDT | [2] Epoch -779 finished +---------------------------------- --------------- +epoch -779 +replay_buffer/size 999033 +trainer/num train calls 222000 +trainer/Policy Loss -19.1879 +trainer/Log Pis Mean 24.1549 +trainer/Log Pis Std 13.2003 +trainer/Log Pis Max 70.9942 +trainer/Log Pis Min -7.43782 +trainer/policy/mean Mean -0.0267299 +trainer/policy/mean Std 0.905448 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.82254 +trainer/policy/normal/std Std 0.587633 +trainer/policy/normal/std Max 5.38684 +trainer/policy/normal/std Min 0.443958 +trainer/policy/normal/log_std Mean 1.00807 +trainer/policy/normal/log_std Std 0.268946 +trainer/policy/normal/log_std Max 1.68396 +trainer/policy/normal/log_std Min -0.812025 +eval/num steps total 221627 +eval/num paths total 222 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0968982 +eval/Actions Std 0.912216 +eval/Actions Max 0.999985 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67917 +time/logging (s) 0.00377086 +time/sampling batch (s) 0.28585 +time/saving (s) 0.00348838 +time/training (s) 6.9503 +time/epoch (s) 9.92257 +time/total (s) 2284.97 +Epoch -779 +---------------------------------- --------------- +2022-05-10 13:48:52.350192 PDT | [2] Epoch -778 finished +---------------------------------- --------------- +epoch -778 +replay_buffer/size 999033 +trainer/num train calls 223000 +trainer/Policy Loss -20.1316 +trainer/Log Pis Mean 24.3043 +trainer/Log Pis Std 13.8519 +trainer/Log Pis Max 68.4126 +trainer/Log Pis Min -7.17609 +trainer/policy/mean Mean -0.0572041 +trainer/policy/mean Std 0.903372 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.90125 +trainer/policy/normal/std Std 0.652665 +trainer/policy/normal/std Max 5.92032 +trainer/policy/normal/std Min 0.387103 +trainer/policy/normal/log_std Mean 1.02845 +trainer/policy/normal/log_std Std 0.306105 +trainer/policy/normal/log_std Max 1.77839 +trainer/policy/normal/log_std Min -0.949064 +eval/num steps total 222627 +eval/num paths total 223 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0440157 +eval/Actions Std 0.944296 +eval/Actions Max 0.999994 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51473 +time/logging (s) 0.00378702 +time/sampling batch (s) 0.285042 +time/saving (s) 0.00348601 +time/training (s) 7.34665 +time/epoch (s) 10.1537 +time/total (s) 2295.13 +Epoch -778 +---------------------------------- --------------- +2022-05-10 13:49:02.059879 PDT | [2] Epoch -777 finished +---------------------------------- --------------- +epoch -777 +replay_buffer/size 999033 +trainer/num train calls 224000 +trainer/Policy Loss -19.7056 +trainer/Log Pis Mean 24.2994 +trainer/Log Pis Std 13.4965 +trainer/Log Pis Max 84.9275 +trainer/Log Pis Min -4.38837 +trainer/policy/mean Mean -0.0445576 +trainer/policy/mean Std 0.90703 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.806 +trainer/policy/normal/std Std 0.588546 +trainer/policy/normal/std Max 5.41085 +trainer/policy/normal/std Min 0.4088 +trainer/policy/normal/log_std Mean 1.00215 +trainer/policy/normal/log_std Std 0.268351 +trainer/policy/normal/log_std Max 1.68841 +trainer/policy/normal/log_std Min -0.894529 +eval/num steps total 223627 +eval/num paths total 224 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0995763 +eval/Actions Std 0.906642 +eval/Actions Max 0.999996 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57935 +time/logging (s) 0.00385486 +time/sampling batch (s) 0.542196 +time/saving (s) 0.00350907 +time/training (s) 6.55968 +time/epoch (s) 9.68858 +time/total (s) 2304.82 +Epoch -777 +---------------------------------- --------------- +2022-05-10 13:49:11.676822 PDT | [2] Epoch -776 finished +---------------------------------- --------------- +epoch -776 +replay_buffer/size 999033 +trainer/num train calls 225000 +trainer/Policy Loss -19.5915 +trainer/Log Pis Mean 24.408 +trainer/Log Pis Std 13.6323 +trainer/Log Pis Max 66.4335 +trainer/Log Pis Min -14.084 +trainer/policy/mean Mean -0.0311448 +trainer/policy/mean Std 0.899926 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.83116 +trainer/policy/normal/std Std 0.608999 +trainer/policy/normal/std Max 5.76451 +trainer/policy/normal/std Min 0.381626 +trainer/policy/normal/log_std Mean 1.00778 +trainer/policy/normal/log_std Std 0.287892 +trainer/policy/normal/log_std Max 1.75172 +trainer/policy/normal/log_std Min -0.963314 +eval/num steps total 224627 +eval/num paths total 225 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.119536 +eval/Actions Std 0.918733 +eval/Actions Max 0.99999 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.74016 +time/logging (s) 0.00475009 +time/sampling batch (s) 0.283583 +time/saving (s) 0.004469 +time/training (s) 6.56419 +time/epoch (s) 9.59716 +time/total (s) 2314.42 +Epoch -776 +---------------------------------- --------------- +2022-05-10 13:49:22.184898 PDT | [2] Epoch -775 finished +---------------------------------- --------------- +epoch -775 +replay_buffer/size 999033 +trainer/num train calls 226000 +trainer/Policy Loss -20.438 +trainer/Log Pis Mean 24.3186 +trainer/Log Pis Std 13.205 +trainer/Log Pis Max 62.8172 +trainer/Log Pis Min -5.53645 +trainer/policy/mean Mean -0.0393621 +trainer/policy/mean Std 0.907046 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.88979 +trainer/policy/normal/std Std 0.590781 +trainer/policy/normal/std Max 5.3441 +trainer/policy/normal/std Min 0.373333 +trainer/policy/normal/log_std Mean 1.03233 +trainer/policy/normal/log_std Std 0.26752 +trainer/policy/normal/log_std Max 1.67599 +trainer/policy/normal/log_std Min -0.985283 +eval/num steps total 225627 +eval/num paths total 226 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.267435 +eval/Actions Std 0.905914 +eval/Actions Max 0.999996 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.90667 +time/logging (s) 0.00442093 +time/sampling batch (s) 0.538524 +time/saving (s) 0.00372998 +time/training (s) 7.03265 +time/epoch (s) 10.486 +time/total (s) 2324.91 +Epoch -775 +---------------------------------- --------------- +2022-05-10 13:49:32.665708 PDT | [2] Epoch -774 finished +---------------------------------- --------------- +epoch -774 +replay_buffer/size 999033 +trainer/num train calls 227000 +trainer/Policy Loss -18.7069 +trainer/Log Pis Mean 24.5481 +trainer/Log Pis Std 13.433 +trainer/Log Pis Max 69.3768 +trainer/Log Pis Min -10.0262 +trainer/policy/mean Mean -0.0388348 +trainer/policy/mean Std 0.907617 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.81439 +trainer/policy/normal/std Std 0.625668 +trainer/policy/normal/std Max 5.77229 +trainer/policy/normal/std Min 0.443663 +trainer/policy/normal/log_std Mean 1.00039 +trainer/policy/normal/log_std Std 0.291901 +trainer/policy/normal/log_std Max 1.75307 +trainer/policy/normal/log_std Min -0.81269 +eval/num steps total 226627 +eval/num paths total 227 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0919893 +eval/Actions Std 0.906284 +eval/Actions Max 0.999999 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.81342 +time/logging (s) 0.00430209 +time/sampling batch (s) 0.540482 +time/saving (s) 0.0034976 +time/training (s) 7.09793 +time/epoch (s) 10.4596 +time/total (s) 2335.38 +Epoch -774 +---------------------------------- --------------- +2022-05-10 13:49:41.993458 PDT | [2] Epoch -773 finished +---------------------------------- --------------- +epoch -773 +replay_buffer/size 999033 +trainer/num train calls 228000 +trainer/Policy Loss -19.3869 +trainer/Log Pis Mean 25.5048 +trainer/Log Pis Std 13.1611 +trainer/Log Pis Max 79.9896 +trainer/Log Pis Min -13.7742 +trainer/policy/mean Mean -0.0371011 +trainer/policy/mean Std 0.904928 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.81029 +trainer/policy/normal/std Std 0.616295 +trainer/policy/normal/std Max 5.15928 +trainer/policy/normal/std Min 0.414308 +trainer/policy/normal/log_std Mean 0.999697 +trainer/policy/normal/log_std Std 0.288829 +trainer/policy/normal/log_std Max 1.6408 +trainer/policy/normal/log_std Min -0.881146 +eval/num steps total 227627 +eval/num paths total 228 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0975731 +eval/Actions Std 0.915212 +eval/Actions Max 0.999999 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73074 +time/logging (s) 0.00398823 +time/sampling batch (s) 0.282088 +time/saving (s) 0.00353286 +time/training (s) 6.28576 +time/epoch (s) 9.30611 +time/total (s) 2344.68 +Epoch -773 +---------------------------------- --------------- +2022-05-10 13:49:52.474568 PDT | [2] Epoch -772 finished +---------------------------------- --------------- +epoch -772 +replay_buffer/size 999033 +trainer/num train calls 229000 +trainer/Policy Loss -19.3129 +trainer/Log Pis Mean 24.1432 +trainer/Log Pis Std 12.613 +trainer/Log Pis Max 62.4006 +trainer/Log Pis Min -8.0113 +trainer/policy/mean Mean -0.0304714 +trainer/policy/mean Std 0.908983 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.85022 +trainer/policy/normal/std Std 0.57223 +trainer/policy/normal/std Max 4.74953 +trainer/policy/normal/std Min 0.338727 +trainer/policy/normal/log_std Mean 1.01993 +trainer/policy/normal/log_std Std 0.259379 +trainer/policy/normal/log_std Max 1.55805 +trainer/policy/normal/log_std Min -1.08256 +eval/num steps total 228627 +eval/num paths total 229 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.278505 +eval/Actions Std 0.903413 +eval/Actions Max 0.99999 +eval/Actions Min -0.999982 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63369 +time/logging (s) 0.00370299 +time/sampling batch (s) 0.289097 +time/saving (s) 0.00348315 +time/training (s) 7.52948 +time/epoch (s) 10.4594 +time/total (s) 2355.15 +Epoch -772 +---------------------------------- --------------- +2022-05-10 13:50:03.057606 PDT | [2] Epoch -771 finished +---------------------------------- --------------- +epoch -771 +replay_buffer/size 999033 +trainer/num train calls 230000 +trainer/Policy Loss -19.5787 +trainer/Log Pis Mean 24.3906 +trainer/Log Pis Std 12.7876 +trainer/Log Pis Max 68.2994 +trainer/Log Pis Min -10.523 +trainer/policy/mean Mean -0.0146249 +trainer/policy/mean Std 0.907384 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.789 +trainer/policy/normal/std Std 0.606406 +trainer/policy/normal/std Max 6.72718 +trainer/policy/normal/std Min 0.337996 +trainer/policy/normal/log_std Mean 0.992959 +trainer/policy/normal/log_std Std 0.285392 +trainer/policy/normal/log_std Max 1.90616 +trainer/policy/normal/log_std Min -1.08472 +eval/num steps total 229627 +eval/num paths total 230 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.412314 +eval/Actions Std 0.886213 +eval/Actions Max 0.999995 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.80431 +time/logging (s) 0.0041655 +time/sampling batch (s) 0.535951 +time/saving (s) 0.00404907 +time/training (s) 7.21383 +time/epoch (s) 10.5623 +time/total (s) 2365.71 +Epoch -771 +---------------------------------- --------------- +2022-05-10 13:50:13.114796 PDT | [2] Epoch -770 finished +---------------------------------- --------------- +epoch -770 +replay_buffer/size 999033 +trainer/num train calls 231000 +trainer/Policy Loss -19.753 +trainer/Log Pis Mean 24.6539 +trainer/Log Pis Std 13.4379 +trainer/Log Pis Max 72.4856 +trainer/Log Pis Min -7.90464 +trainer/policy/mean Mean -0.0387652 +trainer/policy/mean Std 0.901329 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.86526 +trainer/policy/normal/std Std 0.619918 +trainer/policy/normal/std Max 5.14408 +trainer/policy/normal/std Min 0.344979 +trainer/policy/normal/log_std Mean 1.01933 +trainer/policy/normal/log_std Std 0.289633 +trainer/policy/normal/log_std Max 1.63785 +trainer/policy/normal/log_std Min -1.06427 +eval/num steps total 230627 +eval/num paths total 231 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0385578 +eval/Actions Std 0.900275 +eval/Actions Max 0.999989 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65457 +time/logging (s) 0.00380645 +time/sampling batch (s) 0.286778 +time/saving (s) 0.00384425 +time/training (s) 7.0861 +time/epoch (s) 10.0351 +time/total (s) 2375.75 +Epoch -770 +---------------------------------- --------------- +2022-05-10 13:50:22.527438 PDT | [2] Epoch -769 finished +---------------------------------- --------------- +epoch -769 +replay_buffer/size 999033 +trainer/num train calls 232000 +trainer/Policy Loss -20.3485 +trainer/Log Pis Mean 25.0701 +trainer/Log Pis Std 13.5405 +trainer/Log Pis Max 70.4671 +trainer/Log Pis Min -9.45479 +trainer/policy/mean Mean -0.031614 +trainer/policy/mean Std 0.907785 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.87409 +trainer/policy/normal/std Std 0.59687 +trainer/policy/normal/std Max 5.52323 +trainer/policy/normal/std Min 0.366476 +trainer/policy/normal/log_std Mean 1.0261 +trainer/policy/normal/log_std Std 0.270788 +trainer/policy/normal/log_std Max 1.70896 +trainer/policy/normal/log_std Min -1.00382 +eval/num steps total 231627 +eval/num paths total 232 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0674214 +eval/Actions Std 0.914597 +eval/Actions Max 0.999995 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.52328 +time/logging (s) 0.00370081 +time/sampling batch (s) 0.283601 +time/saving (s) 0.00346664 +time/training (s) 6.57746 +time/epoch (s) 9.39151 +time/total (s) 2385.15 +Epoch -769 +---------------------------------- --------------- +2022-05-10 13:50:32.509272 PDT | [2] Epoch -768 finished +---------------------------------- --------------- +epoch -768 +replay_buffer/size 999033 +trainer/num train calls 233000 +trainer/Policy Loss -19.1271 +trainer/Log Pis Mean 23.9058 +trainer/Log Pis Std 12.4693 +trainer/Log Pis Max 72.9934 +trainer/Log Pis Min -10.8805 +trainer/policy/mean Mean -0.0558068 +trainer/policy/mean Std 0.902785 +trainer/policy/mean Max 0.999981 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.84275 +trainer/policy/normal/std Std 0.607995 +trainer/policy/normal/std Max 5.72 +trainer/policy/normal/std Min 0.321977 +trainer/policy/normal/log_std Mean 1.0122 +trainer/policy/normal/log_std Std 0.287255 +trainer/policy/normal/log_std Max 1.74397 +trainer/policy/normal/log_std Min -1.13327 +eval/num steps total 232627 +eval/num paths total 233 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0697579 +eval/Actions Std 0.913827 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66525 +time/logging (s) 0.00371524 +time/sampling batch (s) 0.284229 +time/saving (s) 0.00349546 +time/training (s) 7.00426 +time/epoch (s) 9.96095 +time/total (s) 2395.11 +Epoch -768 +---------------------------------- --------------- +2022-05-10 13:50:42.704717 PDT | [2] Epoch -767 finished +---------------------------------- --------------- +epoch -767 +replay_buffer/size 999033 +trainer/num train calls 234000 +trainer/Policy Loss -20.0964 +trainer/Log Pis Mean 23.6377 +trainer/Log Pis Std 12.2877 +trainer/Log Pis Max 67.1615 +trainer/Log Pis Min -9.09764 +trainer/policy/mean Mean -0.0195804 +trainer/policy/mean Std 0.907667 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.80994 +trainer/policy/normal/std Std 0.609289 +trainer/policy/normal/std Max 5.22182 +trainer/policy/normal/std Min 0.332267 +trainer/policy/normal/log_std Mean 0.999356 +trainer/policy/normal/log_std Std 0.292816 +trainer/policy/normal/log_std Max 1.65285 +trainer/policy/normal/log_std Min -1.10182 +eval/num steps total 233627 +eval/num paths total 234 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0850364 +eval/Actions Std 0.912636 +eval/Actions Max 0.999989 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.74828 +time/logging (s) 0.00370965 +time/sampling batch (s) 0.283679 +time/saving (s) 0.00346817 +time/training (s) 7.1355 +time/epoch (s) 10.1746 +time/total (s) 2405.29 +Epoch -767 +---------------------------------- --------------- +2022-05-10 13:50:51.874527 PDT | [2] Epoch -766 finished +---------------------------------- --------------- +epoch -766 +replay_buffer/size 999033 +trainer/num train calls 235000 +trainer/Policy Loss -19.0063 +trainer/Log Pis Mean 22.2908 +trainer/Log Pis Std 13.2147 +trainer/Log Pis Max 64.9619 +trainer/Log Pis Min -4.65732 +trainer/policy/mean Mean -0.0336201 +trainer/policy/mean Std 0.90495 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.86862 +trainer/policy/normal/std Std 0.599243 +trainer/policy/normal/std Max 5.52369 +trainer/policy/normal/std Min 0.32433 +trainer/policy/normal/log_std Mean 1.0243 +trainer/policy/normal/log_std Std 0.2688 +trainer/policy/normal/log_std Max 1.70905 +trainer/policy/normal/log_std Min -1.126 +eval/num steps total 234627 +eval/num paths total 235 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.110011 +eval/Actions Std 0.880216 +eval/Actions Max 0.99999 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.78278 +time/logging (s) 0.00444807 +time/sampling batch (s) 0.286328 +time/saving (s) 0.0036118 +time/training (s) 6.07249 +time/epoch (s) 9.14967 +time/total (s) 2414.44 +Epoch -766 +---------------------------------- --------------- +2022-05-10 13:51:02.336465 PDT | [2] Epoch -765 finished +---------------------------------- --------------- +epoch -765 +replay_buffer/size 999033 +trainer/num train calls 236000 +trainer/Policy Loss -19.1984 +trainer/Log Pis Mean 24.2967 +trainer/Log Pis Std 13.7064 +trainer/Log Pis Max 94.3808 +trainer/Log Pis Min -6.08427 +trainer/policy/mean Mean -0.0308877 +trainer/policy/mean Std 0.909671 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.86834 +trainer/policy/normal/std Std 0.603954 +trainer/policy/normal/std Max 6.09061 +trainer/policy/normal/std Min 0.383826 +trainer/policy/normal/log_std Mean 1.02268 +trainer/policy/normal/log_std Std 0.278158 +trainer/policy/normal/log_std Max 1.80675 +trainer/policy/normal/log_std Min -0.957567 +eval/num steps total 235627 +eval/num paths total 236 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0521492 +eval/Actions Std 0.909141 +eval/Actions Max 1 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50661 +time/logging (s) 0.00418576 +time/sampling batch (s) 0.284823 +time/saving (s) 0.00422932 +time/training (s) 7.64012 +time/epoch (s) 10.44 +time/total (s) 2424.88 +Epoch -765 +---------------------------------- --------------- +2022-05-10 13:51:12.796815 PDT | [2] Epoch -764 finished +---------------------------------- --------------- +epoch -764 +replay_buffer/size 999033 +trainer/num train calls 237000 +trainer/Policy Loss -18.9654 +trainer/Log Pis Mean 24.263 +trainer/Log Pis Std 13.3877 +trainer/Log Pis Max 68.6156 +trainer/Log Pis Min -7.67336 +trainer/policy/mean Mean -0.0331101 +trainer/policy/mean Std 0.90175 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.88294 +trainer/policy/normal/std Std 0.601 +trainer/policy/normal/std Max 5.19124 +trainer/policy/normal/std Min 0.433527 +trainer/policy/normal/log_std Mean 1.02901 +trainer/policy/normal/log_std Std 0.270565 +trainer/policy/normal/log_std Max 1.64697 +trainer/policy/normal/log_std Min -0.835802 +eval/num steps total 236627 +eval/num paths total 237 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.248655 +eval/Actions Std 0.928588 +eval/Actions Max 0.999981 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.89834 +time/logging (s) 0.00377447 +time/sampling batch (s) 0.536208 +time/saving (s) 0.00361373 +time/training (s) 6.99669 +time/epoch (s) 10.4386 +time/total (s) 2435.32 +Epoch -764 +---------------------------------- --------------- +2022-05-10 13:51:22.241069 PDT | [2] Epoch -763 finished +---------------------------------- --------------- +epoch -763 +replay_buffer/size 999033 +trainer/num train calls 238000 +trainer/Policy Loss -19.1396 +trainer/Log Pis Mean 25.025 +trainer/Log Pis Std 13.2939 +trainer/Log Pis Max 71.1518 +trainer/Log Pis Min -7.74215 +trainer/policy/mean Mean -0.0222043 +trainer/policy/mean Std 0.905265 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.89243 +trainer/policy/normal/std Std 0.597355 +trainer/policy/normal/std Max 5.32126 +trainer/policy/normal/std Min 0.397469 +trainer/policy/normal/log_std Mean 1.03131 +trainer/policy/normal/log_std Std 0.279206 +trainer/policy/normal/log_std Max 1.67171 +trainer/policy/normal/log_std Min -0.922639 +eval/num steps total 237627 +eval/num paths total 238 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.191264 +eval/Actions Std 0.874493 +eval/Actions Max 1 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5077 +time/logging (s) 0.00368594 +time/sampling batch (s) 0.281964 +time/saving (s) 0.00347736 +time/training (s) 6.62656 +time/epoch (s) 9.42338 +time/total (s) 2444.75 +Epoch -763 +---------------------------------- --------------- +2022-05-10 13:51:32.618915 PDT | [2] Epoch -762 finished +---------------------------------- --------------- +epoch -762 +replay_buffer/size 999033 +trainer/num train calls 239000 +trainer/Policy Loss -19.3103 +trainer/Log Pis Mean 24.7103 +trainer/Log Pis Std 12.7054 +trainer/Log Pis Max 63.921 +trainer/Log Pis Min -2.60665 +trainer/policy/mean Mean -0.0292845 +trainer/policy/mean Std 0.907195 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.85397 +trainer/policy/normal/std Std 0.595198 +trainer/policy/normal/std Max 5.00912 +trainer/policy/normal/std Min 0.387968 +trainer/policy/normal/log_std Mean 1.01758 +trainer/policy/normal/log_std Std 0.280417 +trainer/policy/normal/log_std Max 1.61126 +trainer/policy/normal/log_std Min -0.946833 +eval/num steps total 238627 +eval/num paths total 239 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.151463 +eval/Actions Std 0.799961 +eval/Actions Max 0.999995 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.83442 +time/logging (s) 0.00434545 +time/sampling batch (s) 0.288344 +time/saving (s) 0.00438807 +time/training (s) 7.22569 +time/epoch (s) 10.3572 +time/total (s) 2455.11 +Epoch -762 +---------------------------------- --------------- +2022-05-10 13:51:42.503558 PDT | [2] Epoch -761 finished +---------------------------------- --------------- +epoch -761 +replay_buffer/size 999033 +trainer/num train calls 240000 +trainer/Policy Loss -19.4684 +trainer/Log Pis Mean 25.267 +trainer/Log Pis Std 13.4647 +trainer/Log Pis Max 72.2838 +trainer/Log Pis Min -8.55307 +trainer/policy/mean Mean -0.0671246 +trainer/policy/mean Std 0.904684 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.8784 +trainer/policy/normal/std Std 0.602306 +trainer/policy/normal/std Max 6.47259 +trainer/policy/normal/std Min 0.399242 +trainer/policy/normal/log_std Mean 1.02735 +trainer/policy/normal/log_std Std 0.271008 +trainer/policy/normal/log_std Max 1.86758 +trainer/policy/normal/log_std Min -0.918188 +eval/num steps total 239627 +eval/num paths total 240 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.170722 +eval/Actions Std 0.896792 +eval/Actions Max 0.999995 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.89569 +time/logging (s) 0.00408883 +time/sampling batch (s) 0.292391 +time/saving (s) 0.0041825 +time/training (s) 6.66588 +time/epoch (s) 9.86223 +time/total (s) 2464.98 +Epoch -761 +---------------------------------- --------------- +2022-05-10 13:51:51.880401 PDT | [2] Epoch -760 finished +---------------------------------- --------------- +epoch -760 +replay_buffer/size 999033 +trainer/num train calls 241000 +trainer/Policy Loss -20.1149 +trainer/Log Pis Mean 24.6048 +trainer/Log Pis Std 13.3117 +trainer/Log Pis Max 72.8471 +trainer/Log Pis Min -12.1862 +trainer/policy/mean Mean -0.0308351 +trainer/policy/mean Std 0.906262 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.84313 +trainer/policy/normal/std Std 0.612982 +trainer/policy/normal/std Max 5.5285 +trainer/policy/normal/std Min 0.314868 +trainer/policy/normal/log_std Mean 1.01215 +trainer/policy/normal/log_std Std 0.287652 +trainer/policy/normal/log_std Max 1.70992 +trainer/policy/normal/log_std Min -1.1556 +eval/num steps total 240502 +eval/num paths total 241 +eval/path length Mean 875 +eval/path length Std 0 +eval/path length Max 875 +eval/path length Min 875 +eval/Rewards Mean 0.00114286 +eval/Rewards Std 0.0337868 +eval/Rewards Max 1 +eval/Rewards Min 0 +eval/Returns Mean 1 +eval/Returns Std 0 +eval/Returns Max 1 +eval/Returns Min 1 +eval/Actions Mean 0.0112276 +eval/Actions Std 0.896655 +eval/Actions Max 0.999987 +eval/Actions Min -0.999986 +eval/Num Paths 1 +eval/Average Returns 1 +eval/normalized_score 100 +time/evaluation sampling (s) 2.62131 +time/logging (s) 0.00365126 +time/sampling batch (s) 0.28244 +time/saving (s) 0.00417966 +time/training (s) 6.44361 +time/epoch (s) 9.3552 +time/total (s) 2474.34 +Epoch -760 +---------------------------------- --------------- +2022-05-10 13:52:03.133870 PDT | [2] Epoch -759 finished +---------------------------------- --------------- +epoch -759 +replay_buffer/size 999033 +trainer/num train calls 242000 +trainer/Policy Loss -19.5459 +trainer/Log Pis Mean 23.7129 +trainer/Log Pis Std 12.4614 +trainer/Log Pis Max 65.5025 +trainer/Log Pis Min -6.52311 +trainer/policy/mean Mean -0.0533942 +trainer/policy/mean Std 0.906761 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.77399 +trainer/policy/normal/std Std 0.606078 +trainer/policy/normal/std Max 5.40174 +trainer/policy/normal/std Min 0.339346 +trainer/policy/normal/log_std Mean 0.986996 +trainer/policy/normal/log_std Std 0.28815 +trainer/policy/normal/log_std Max 1.68672 +trainer/policy/normal/log_std Min -1.08074 +eval/num steps total 241502 +eval/num paths total 242 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00116947 +eval/Actions Std 0.902522 +eval/Actions Max 0.999996 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6677 +time/logging (s) 0.00413275 +time/sampling batch (s) 0.535533 +time/saving (s) 0.0042981 +time/training (s) 8.02079 +time/epoch (s) 11.2325 +time/total (s) 2485.57 +Epoch -759 +---------------------------------- --------------- +2022-05-10 13:52:13.504047 PDT | [2] Epoch -758 finished +---------------------------------- --------------- +epoch -758 +replay_buffer/size 999033 +trainer/num train calls 243000 +trainer/Policy Loss -21.623 +trainer/Log Pis Mean 24.0948 +trainer/Log Pis Std 12.6709 +trainer/Log Pis Max 60.2029 +trainer/Log Pis Min -8.24603 +trainer/policy/mean Mean -0.0611833 +trainer/policy/mean Std 0.906529 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.75402 +trainer/policy/normal/std Std 0.594663 +trainer/policy/normal/std Max 6.14264 +trainer/policy/normal/std Min 0.309582 +trainer/policy/normal/log_std Mean 0.979809 +trainer/policy/normal/log_std Std 0.290097 +trainer/policy/normal/log_std Max 1.81525 +trainer/policy/normal/log_std Min -1.17253 +eval/num steps total 242502 +eval/num paths total 243 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0802595 +eval/Actions Std 0.905498 +eval/Actions Max 0.999992 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62097 +time/logging (s) 0.00371169 +time/sampling batch (s) 0.283193 +time/saving (s) 0.00347631 +time/training (s) 7.43705 +time/epoch (s) 10.3484 +time/total (s) 2495.92 +Epoch -758 +---------------------------------- --------------- +2022-05-10 13:52:23.463630 PDT | [2] Epoch -757 finished +---------------------------------- --------------- +epoch -757 +replay_buffer/size 999033 +trainer/num train calls 244000 +trainer/Policy Loss -19.9018 +trainer/Log Pis Mean 24.4624 +trainer/Log Pis Std 13.2599 +trainer/Log Pis Max 84.6713 +trainer/Log Pis Min -6.74281 +trainer/policy/mean Mean -0.0330148 +trainer/policy/mean Std 0.907192 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999983 +trainer/policy/normal/std Mean 2.87129 +trainer/policy/normal/std Std 0.60461 +trainer/policy/normal/std Max 5.29366 +trainer/policy/normal/std Min 0.410518 +trainer/policy/normal/log_std Mean 1.02376 +trainer/policy/normal/log_std Std 0.277346 +trainer/policy/normal/log_std Max 1.66651 +trainer/policy/normal/log_std Min -0.890336 +eval/num steps total 243502 +eval/num paths total 244 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.320092 +eval/Actions Std 0.835126 +eval/Actions Max 0.999994 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50203 +time/logging (s) 0.00422235 +time/sampling batch (s) 0.28163 +time/saving (s) 0.00403113 +time/training (s) 7.14738 +time/epoch (s) 9.93929 +time/total (s) 2505.86 +Epoch -757 +---------------------------------- --------------- +2022-05-10 13:52:33.134320 PDT | [2] Epoch -756 finished +---------------------------------- --------------- +epoch -756 +replay_buffer/size 999033 +trainer/num train calls 245000 +trainer/Policy Loss -19.0943 +trainer/Log Pis Mean 23.6438 +trainer/Log Pis Std 12.872 +trainer/Log Pis Max 72.1225 +trainer/Log Pis Min -12.1992 +trainer/policy/mean Mean -0.0307627 +trainer/policy/mean Std 0.905208 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.86627 +trainer/policy/normal/std Std 0.628691 +trainer/policy/normal/std Max 5.76703 +trainer/policy/normal/std Min 0.422391 +trainer/policy/normal/log_std Mean 1.0199 +trainer/policy/normal/log_std Std 0.28609 +trainer/policy/normal/log_std Max 1.75216 +trainer/policy/normal/log_std Min -0.861825 +eval/num steps total 244502 +eval/num paths total 245 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.133603 +eval/Actions Std 0.895015 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69908 +time/logging (s) 0.00381735 +time/sampling batch (s) 0.282323 +time/saving (s) 0.00373991 +time/training (s) 6.66017 +time/epoch (s) 9.64913 +time/total (s) 2515.52 +Epoch -756 +---------------------------------- --------------- +2022-05-10 13:52:43.341204 PDT | [2] Epoch -755 finished +---------------------------------- --------------- +epoch -755 +replay_buffer/size 999033 +trainer/num train calls 246000 +trainer/Policy Loss -19.023 +trainer/Log Pis Mean 24.3863 +trainer/Log Pis Std 13.2872 +trainer/Log Pis Max 65.9258 +trainer/Log Pis Min -7.9091 +trainer/policy/mean Mean -0.0331548 +trainer/policy/mean Std 0.908319 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.84051 +trainer/policy/normal/std Std 0.608329 +trainer/policy/normal/std Max 6.01785 +trainer/policy/normal/std Min 0.334252 +trainer/policy/normal/log_std Mean 1.01227 +trainer/policy/normal/log_std Std 0.280746 +trainer/policy/normal/log_std Max 1.79473 +trainer/policy/normal/log_std Min -1.09586 +eval/num steps total 245502 +eval/num paths total 246 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.216601 +eval/Actions Std 0.95392 +eval/Actions Max 0.99999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60282 +time/logging (s) 0.0043288 +time/sampling batch (s) 0.279617 +time/saving (s) 0.00346709 +time/training (s) 7.29644 +time/epoch (s) 10.1867 +time/total (s) 2525.71 +Epoch -755 +---------------------------------- --------------- +2022-05-10 13:52:53.314228 PDT | [2] Epoch -754 finished +---------------------------------- --------------- +epoch -754 +replay_buffer/size 999033 +trainer/num train calls 247000 +trainer/Policy Loss -20.6193 +trainer/Log Pis Mean 23.9102 +trainer/Log Pis Std 12.8092 +trainer/Log Pis Max 76.5252 +trainer/Log Pis Min -4.10648 +trainer/policy/mean Mean -0.0464752 +trainer/policy/mean Std 0.903119 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.84547 +trainer/policy/normal/std Std 0.596651 +trainer/policy/normal/std Max 5.50622 +trainer/policy/normal/std Min 0.416089 +trainer/policy/normal/log_std Mean 1.01587 +trainer/policy/normal/log_std Std 0.270261 +trainer/policy/normal/log_std Max 1.70588 +trainer/policy/normal/log_std Min -0.876855 +eval/num steps total 246502 +eval/num paths total 247 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0597233 +eval/Actions Std 0.903457 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.52972 +time/logging (s) 0.00365179 +time/sampling batch (s) 0.279518 +time/saving (s) 0.00347622 +time/training (s) 7.13514 +time/epoch (s) 9.9515 +time/total (s) 2535.66 +Epoch -754 +---------------------------------- --------------- +2022-05-10 13:53:03.943579 PDT | [2] Epoch -753 finished +---------------------------------- --------------- +epoch -753 +replay_buffer/size 999033 +trainer/num train calls 248000 +trainer/Policy Loss -18.7208 +trainer/Log Pis Mean 23.8582 +trainer/Log Pis Std 13.2714 +trainer/Log Pis Max 81.9723 +trainer/Log Pis Min -7.47931 +trainer/policy/mean Mean -0.0114723 +trainer/policy/mean Std 0.904395 +trainer/policy/mean Max 1 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.85845 +trainer/policy/normal/std Std 0.661334 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.356452 +trainer/policy/normal/log_std Mean 1.01192 +trainer/policy/normal/log_std Std 0.312171 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.03156 +eval/num steps total 247502 +eval/num paths total 248 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.223451 +eval/Actions Std 0.91872 +eval/Actions Max 0.999996 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63221 +time/logging (s) 0.00367517 +time/sampling batch (s) 0.279154 +time/saving (s) 0.00347422 +time/training (s) 7.69007 +time/epoch (s) 10.6086 +time/total (s) 2546.27 +Epoch -753 +---------------------------------- --------------- +2022-05-10 13:53:13.931942 PDT | [2] Epoch -752 finished +---------------------------------- --------------- +epoch -752 +replay_buffer/size 999033 +trainer/num train calls 249000 +trainer/Policy Loss -20.098 +trainer/Log Pis Mean 25.2354 +trainer/Log Pis Std 13.4921 +trainer/Log Pis Max 71.465 +trainer/Log Pis Min -7.45118 +trainer/policy/mean Mean -0.0428547 +trainer/policy/mean Std 0.904378 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.84965 +trainer/policy/normal/std Std 0.619165 +trainer/policy/normal/std Max 6.09033 +trainer/policy/normal/std Min 0.342117 +trainer/policy/normal/log_std Mean 1.01335 +trainer/policy/normal/log_std Std 0.293276 +trainer/policy/normal/log_std Max 1.8067 +trainer/policy/normal/log_std Min -1.0726 +eval/num steps total 248502 +eval/num paths total 249 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0735071 +eval/Actions Std 0.914804 +eval/Actions Max 0.999992 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59001 +time/logging (s) 0.00410819 +time/sampling batch (s) 0.281483 +time/saving (s) 0.00400526 +time/training (s) 7.08826 +time/epoch (s) 9.96786 +time/total (s) 2556.24 +Epoch -752 +---------------------------------- --------------- +2022-05-10 13:53:24.105183 PDT | [2] Epoch -751 finished +---------------------------------- --------------- +epoch -751 +replay_buffer/size 999033 +trainer/num train calls 250000 +trainer/Policy Loss -20.7966 +trainer/Log Pis Mean 24.6986 +trainer/Log Pis Std 13.2417 +trainer/Log Pis Max 75.9197 +trainer/Log Pis Min -8.46856 +trainer/policy/mean Mean -0.0514732 +trainer/policy/mean Std 0.907203 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.87349 +trainer/policy/normal/std Std 0.620707 +trainer/policy/normal/std Max 5.89704 +trainer/policy/normal/std Min 0.325696 +trainer/policy/normal/log_std Mean 1.02346 +trainer/policy/normal/log_std Std 0.281303 +trainer/policy/normal/log_std Max 1.77445 +trainer/policy/normal/log_std Min -1.12179 +eval/num steps total 249502 +eval/num paths total 250 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0064979 +eval/Actions Std 0.899288 +eval/Actions Max 0.999997 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.49021 +time/logging (s) 0.0039323 +time/sampling batch (s) 0.278781 +time/saving (s) 0.00376749 +time/training (s) 7.3752 +time/epoch (s) 10.1519 +time/total (s) 2566.4 +Epoch -751 +---------------------------------- --------------- +2022-05-10 13:53:33.695845 PDT | [2] Epoch -750 finished +---------------------------------- --------------- +epoch -750 +replay_buffer/size 999033 +trainer/num train calls 251000 +trainer/Policy Loss -19.6261 +trainer/Log Pis Mean 24.3739 +trainer/Log Pis Std 13.7992 +trainer/Log Pis Max 66.6691 +trainer/Log Pis Min -11.0476 +trainer/policy/mean Mean -0.0771034 +trainer/policy/mean Std 0.90767 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.85295 +trainer/policy/normal/std Std 0.601223 +trainer/policy/normal/std Max 5.9004 +trainer/policy/normal/std Min 0.394238 +trainer/policy/normal/log_std Mean 1.01851 +trainer/policy/normal/log_std Std 0.269196 +trainer/policy/normal/log_std Max 1.77502 +trainer/policy/normal/log_std Min -0.930801 +eval/num steps total 250502 +eval/num paths total 251 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.111146 +eval/Actions Std 0.915294 +eval/Actions Max 0.999992 +eval/Actions Min -0.999983 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57533 +time/logging (s) 0.00371894 +time/sampling batch (s) 0.279513 +time/saving (s) 0.00346715 +time/training (s) 6.70744 +time/epoch (s) 9.56947 +time/total (s) 2575.97 +Epoch -750 +---------------------------------- --------------- +2022-05-10 13:53:43.740568 PDT | [2] Epoch -749 finished +---------------------------------- --------------- +epoch -749 +replay_buffer/size 999033 +trainer/num train calls 252000 +trainer/Policy Loss -18.3927 +trainer/Log Pis Mean 24.6514 +trainer/Log Pis Std 13.3197 +trainer/Log Pis Max 71.7579 +trainer/Log Pis Min -6.63906 +trainer/policy/mean Mean -0.043146 +trainer/policy/mean Std 0.905418 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.892 +trainer/policy/normal/std Std 0.59435 +trainer/policy/normal/std Max 5.35979 +trainer/policy/normal/std Min 0.309754 +trainer/policy/normal/log_std Mean 1.03232 +trainer/policy/normal/log_std Std 0.2727 +trainer/policy/normal/log_std Max 1.67892 +trainer/policy/normal/log_std Min -1.17198 +eval/num steps total 251502 +eval/num paths total 252 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.125088 +eval/Actions Std 0.912392 +eval/Actions Max 0.999988 +eval/Actions Min -0.99998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.76649 +time/logging (s) 0.00372836 +time/sampling batch (s) 0.282002 +time/saving (s) 0.00360452 +time/training (s) 6.96806 +time/epoch (s) 10.0239 +time/total (s) 2586 +Epoch -749 +---------------------------------- --------------- +2022-05-10 13:53:53.883036 PDT | [2] Epoch -748 finished +---------------------------------- --------------- +epoch -748 +replay_buffer/size 999033 +trainer/num train calls 253000 +trainer/Policy Loss -20.0682 +trainer/Log Pis Mean 24.3007 +trainer/Log Pis Std 12.7716 +trainer/Log Pis Max 60.5765 +trainer/Log Pis Min -8.7228 +trainer/policy/mean Mean -0.0205409 +trainer/policy/mean Std 0.906848 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.90639 +trainer/policy/normal/std Std 0.621593 +trainer/policy/normal/std Max 5.45808 +trainer/policy/normal/std Min 0.355662 +trainer/policy/normal/log_std Mean 1.03543 +trainer/policy/normal/log_std Std 0.279584 +trainer/policy/normal/log_std Max 1.6971 +trainer/policy/normal/log_std Min -1.03378 +eval/num steps total 252502 +eval/num paths total 253 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0106302 +eval/Actions Std 0.951449 +eval/Actions Max 0.999999 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.78439 +time/logging (s) 0.00376705 +time/sampling batch (s) 0.281328 +time/saving (s) 0.00347978 +time/training (s) 7.0488 +time/epoch (s) 10.1218 +time/total (s) 2596.12 +Epoch -748 +---------------------------------- --------------- +2022-05-10 13:54:03.229083 PDT | [2] Epoch -747 finished +---------------------------------- --------------- +epoch -747 +replay_buffer/size 999033 +trainer/num train calls 254000 +trainer/Policy Loss -19.2109 +trainer/Log Pis Mean 23.6923 +trainer/Log Pis Std 13.3892 +trainer/Log Pis Max 71.9738 +trainer/Log Pis Min -9.64963 +trainer/policy/mean Mean -0.0337372 +trainer/policy/mean Std 0.905721 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.83265 +trainer/policy/normal/std Std 0.605055 +trainer/policy/normal/std Max 5.19058 +trainer/policy/normal/std Min 0.398682 +trainer/policy/normal/log_std Mean 1.00996 +trainer/policy/normal/log_std Std 0.27797 +trainer/policy/normal/log_std Max 1.64685 +trainer/policy/normal/log_std Min -0.91959 +eval/num steps total 253502 +eval/num paths total 254 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0602478 +eval/Actions Std 0.913268 +eval/Actions Max 0.999994 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.41221 +time/logging (s) 0.00378604 +time/sampling batch (s) 0.284717 +time/saving (s) 0.0035051 +time/training (s) 6.62076 +time/epoch (s) 9.32498 +time/total (s) 2605.45 +Epoch -747 +---------------------------------- --------------- +2022-05-10 13:54:12.136375 PDT | [2] Epoch -746 finished +---------------------------------- --------------- +epoch -746 +replay_buffer/size 999033 +trainer/num train calls 255000 +trainer/Policy Loss -20.419 +trainer/Log Pis Mean 24.2155 +trainer/Log Pis Std 13.3816 +trainer/Log Pis Max 73.2154 +trainer/Log Pis Min -8.90649 +trainer/policy/mean Mean -0.0132591 +trainer/policy/mean Std 0.911732 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999973 +trainer/policy/normal/std Mean 2.84928 +trainer/policy/normal/std Std 0.600631 +trainer/policy/normal/std Max 5.8687 +trainer/policy/normal/std Min 0.396815 +trainer/policy/normal/log_std Mean 1.01613 +trainer/policy/normal/log_std Std 0.277648 +trainer/policy/normal/log_std Max 1.76963 +trainer/policy/normal/log_std Min -0.924284 +eval/num steps total 254502 +eval/num paths total 255 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0467289 +eval/Actions Std 0.893605 +eval/Actions Max 0.99998 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.52582 +time/logging (s) 0.00410504 +time/sampling batch (s) 0.284369 +time/saving (s) 0.00401693 +time/training (s) 6.06824 +time/epoch (s) 8.88656 +time/total (s) 2614.34 +Epoch -746 +---------------------------------- --------------- +2022-05-10 13:54:22.841176 PDT | [2] Epoch -745 finished +---------------------------------- --------------- +epoch -745 +replay_buffer/size 999033 +trainer/num train calls 256000 +trainer/Policy Loss -19.8296 +trainer/Log Pis Mean 23.707 +trainer/Log Pis Std 12.8348 +trainer/Log Pis Max 78.3101 +trainer/Log Pis Min -5.47129 +trainer/policy/mean Mean -0.0465889 +trainer/policy/mean Std 0.903785 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.85088 +trainer/policy/normal/std Std 0.604913 +trainer/policy/normal/std Max 5.39309 +trainer/policy/normal/std Min 0.349864 +trainer/policy/normal/log_std Mean 1.01618 +trainer/policy/normal/log_std Std 0.279798 +trainer/policy/normal/log_std Max 1.68512 +trainer/policy/normal/log_std Min -1.05021 +eval/num steps total 255502 +eval/num paths total 256 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0332053 +eval/Actions Std 0.904057 +eval/Actions Max 0.999998 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.8407 +time/logging (s) 0.00376437 +time/sampling batch (s) 0.302892 +time/saving (s) 0.00364236 +time/training (s) 7.53201 +time/epoch (s) 10.683 +time/total (s) 2625.03 +Epoch -745 +---------------------------------- --------------- +2022-05-10 13:54:32.647315 PDT | [2] Epoch -744 finished +---------------------------------- --------------- +epoch -744 +replay_buffer/size 999033 +trainer/num train calls 257000 +trainer/Policy Loss -19.3008 +trainer/Log Pis Mean 25.1344 +trainer/Log Pis Std 13.3059 +trainer/Log Pis Max 72.7644 +trainer/Log Pis Min -4.54189 +trainer/policy/mean Mean -0.0267168 +trainer/policy/mean Std 0.905932 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.81741 +trainer/policy/normal/std Std 0.629879 +trainer/policy/normal/std Max 5.68888 +trainer/policy/normal/std Min 0.348308 +trainer/policy/normal/log_std Mean 0.999532 +trainer/policy/normal/log_std Std 0.303757 +trainer/policy/normal/log_std Max 1.73851 +trainer/policy/normal/log_std Min -1.05467 +eval/num steps total 256502 +eval/num paths total 257 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0457073 +eval/Actions Std 0.906418 +eval/Actions Max 0.999996 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53294 +time/logging (s) 0.00371183 +time/sampling batch (s) 0.530849 +time/saving (s) 0.00346103 +time/training (s) 6.71417 +time/epoch (s) 9.78513 +time/total (s) 2634.81 +Epoch -744 +---------------------------------- --------------- +2022-05-10 13:54:42.625337 PDT | [2] Epoch -743 finished +---------------------------------- --------------- +epoch -743 +replay_buffer/size 999033 +trainer/num train calls 258000 +trainer/Policy Loss -20.704 +trainer/Log Pis Mean 23.5821 +trainer/Log Pis Std 13.3943 +trainer/Log Pis Max 71.8652 +trainer/Log Pis Min -7.55443 +trainer/policy/mean Mean -0.0478968 +trainer/policy/mean Std 0.905086 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.82486 +trainer/policy/normal/std Std 0.610504 +trainer/policy/normal/std Max 6.45396 +trainer/policy/normal/std Min 0.385396 +trainer/policy/normal/log_std Mean 1.00605 +trainer/policy/normal/log_std Std 0.28366 +trainer/policy/normal/log_std Max 1.86469 +trainer/policy/normal/log_std Min -0.953484 +eval/num steps total 257502 +eval/num paths total 258 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0470871 +eval/Actions Std 0.883374 +eval/Actions Max 0.999984 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50482 +time/logging (s) 0.00371975 +time/sampling batch (s) 0.529812 +time/saving (s) 0.00356365 +time/training (s) 6.91543 +time/epoch (s) 9.95735 +time/total (s) 2644.78 +Epoch -743 +---------------------------------- --------------- +2022-05-10 13:54:52.667407 PDT | [2] Epoch -742 finished +---------------------------------- --------------- +epoch -742 +replay_buffer/size 999033 +trainer/num train calls 259000 +trainer/Policy Loss -19.3225 +trainer/Log Pis Mean 24.0628 +trainer/Log Pis Std 13.3547 +trainer/Log Pis Max 64.4574 +trainer/Log Pis Min -10.2817 +trainer/policy/mean Mean -0.04478 +trainer/policy/mean Std 0.903815 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.84557 +trainer/policy/normal/std Std 0.607669 +trainer/policy/normal/std Max 5.93859 +trainer/policy/normal/std Min 0.360351 +trainer/policy/normal/log_std Mean 1.01484 +trainer/policy/normal/log_std Std 0.276244 +trainer/policy/normal/log_std Max 1.78147 +trainer/policy/normal/log_std Min -1.02068 +eval/num steps total 258502 +eval/num paths total 259 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.167145 +eval/Actions Std 0.871374 +eval/Actions Max 0.999997 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68498 +time/logging (s) 0.00367432 +time/sampling batch (s) 0.282056 +time/saving (s) 0.00352615 +time/training (s) 7.04689 +time/epoch (s) 10.0211 +time/total (s) 2654.8 +Epoch -742 +---------------------------------- --------------- +2022-05-10 13:55:04.070741 PDT | [2] Epoch -741 finished +---------------------------------- --------------- +epoch -741 +replay_buffer/size 999033 +trainer/num train calls 260000 +trainer/Policy Loss -19.8693 +trainer/Log Pis Mean 23.8599 +trainer/Log Pis Std 12.8664 +trainer/Log Pis Max 64.3213 +trainer/Log Pis Min -10.963 +trainer/policy/mean Mean -0.0461545 +trainer/policy/mean Std 0.90331 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.84727 +trainer/policy/normal/std Std 0.60655 +trainer/policy/normal/std Max 5.39781 +trainer/policy/normal/std Min 0.39501 +trainer/policy/normal/log_std Mean 1.01564 +trainer/policy/normal/log_std Std 0.273479 +trainer/policy/normal/log_std Max 1.68599 +trainer/policy/normal/log_std Min -0.928843 +eval/num steps total 259502 +eval/num paths total 260 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0864653 +eval/Actions Std 0.926061 +eval/Actions Max 0.999999 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68214 +time/logging (s) 0.00406927 +time/sampling batch (s) 0.538483 +time/saving (s) 0.00401801 +time/training (s) 8.15363 +time/epoch (s) 11.3824 +time/total (s) 2666.18 +Epoch -741 +---------------------------------- --------------- +2022-05-10 13:55:13.810077 PDT | [2] Epoch -740 finished +---------------------------------- --------------- +epoch -740 +replay_buffer/size 999033 +trainer/num train calls 261000 +trainer/Policy Loss -20.3951 +trainer/Log Pis Mean 24.2005 +trainer/Log Pis Std 12.4376 +trainer/Log Pis Max 65.8432 +trainer/Log Pis Min -7.9248 +trainer/policy/mean Mean -0.0473461 +trainer/policy/mean Std 0.905714 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.84891 +trainer/policy/normal/std Std 0.579585 +trainer/policy/normal/std Max 5.06959 +trainer/policy/normal/std Min 0.386898 +trainer/policy/normal/log_std Mean 1.0174 +trainer/policy/normal/log_std Std 0.272887 +trainer/policy/normal/log_std Max 1.62326 +trainer/policy/normal/log_std Min -0.949593 +eval/num steps total 260502 +eval/num paths total 261 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0255432 +eval/Actions Std 0.939625 +eval/Actions Max 0.999996 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.8091 +time/logging (s) 0.00420189 +time/sampling batch (s) 0.536534 +time/saving (s) 0.00404522 +time/training (s) 6.36385 +time/epoch (s) 9.71773 +time/total (s) 2675.91 +Epoch -740 +---------------------------------- --------------- +2022-05-10 13:55:23.401997 PDT | [2] Epoch -739 finished +---------------------------------- --------------- +epoch -739 +replay_buffer/size 999033 +trainer/num train calls 262000 +trainer/Policy Loss -19.3534 +trainer/Log Pis Mean 24.8867 +trainer/Log Pis Std 13.5505 +trainer/Log Pis Max 69.4132 +trainer/Log Pis Min -9.88501 +trainer/policy/mean Mean -0.0184991 +trainer/policy/mean Std 0.906575 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.87423 +trainer/policy/normal/std Std 0.617871 +trainer/policy/normal/std Max 5.30592 +trainer/policy/normal/std Min 0.424143 +trainer/policy/normal/log_std Mean 1.0238 +trainer/policy/normal/log_std Std 0.281769 +trainer/policy/normal/log_std Max 1.66882 +trainer/policy/normal/log_std Min -0.857685 +eval/num steps total 261502 +eval/num paths total 262 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.222634 +eval/Actions Std 0.909962 +eval/Actions Max 0.999997 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.79365 +time/logging (s) 0.00370824 +time/sampling batch (s) 0.28393 +time/saving (s) 0.00344772 +time/training (s) 6.48483 +time/epoch (s) 9.56956 +time/total (s) 2685.48 +Epoch -739 +---------------------------------- --------------- +2022-05-10 13:55:33.271162 PDT | [2] Epoch -738 finished +---------------------------------- --------------- +epoch -738 +replay_buffer/size 999033 +trainer/num train calls 263000 +trainer/Policy Loss -19.832 +trainer/Log Pis Mean 24.7006 +trainer/Log Pis Std 13.5743 +trainer/Log Pis Max 72.1566 +trainer/Log Pis Min -9.6064 +trainer/policy/mean Mean -0.0400906 +trainer/policy/mean Std 0.907637 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.80249 +trainer/policy/normal/std Std 0.605092 +trainer/policy/normal/std Max 5.42284 +trainer/policy/normal/std Min 0.361973 +trainer/policy/normal/log_std Mean 0.996596 +trainer/policy/normal/log_std Std 0.294298 +trainer/policy/normal/log_std Max 1.69062 +trainer/policy/normal/log_std Min -1.01619 +eval/num steps total 262502 +eval/num paths total 263 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.227484 +eval/Actions Std 0.877229 +eval/Actions Max 0.999989 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67232 +time/logging (s) 0.00370213 +time/sampling batch (s) 0.533756 +time/saving (s) 0.00351021 +time/training (s) 6.63487 +time/epoch (s) 9.84816 +time/total (s) 2695.33 +Epoch -738 +---------------------------------- --------------- +2022-05-10 13:55:43.641682 PDT | [2] Epoch -737 finished +---------------------------------- --------------- +epoch -737 +replay_buffer/size 999033 +trainer/num train calls 264000 +trainer/Policy Loss -20.4908 +trainer/Log Pis Mean 24.2705 +trainer/Log Pis Std 13.5038 +trainer/Log Pis Max 62.0359 +trainer/Log Pis Min -10.2148 +trainer/policy/mean Mean -0.0386886 +trainer/policy/mean Std 0.908029 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.77885 +trainer/policy/normal/std Std 0.609097 +trainer/policy/normal/std Max 5.21929 +trainer/policy/normal/std Min 0.366482 +trainer/policy/normal/log_std Mean 0.988275 +trainer/policy/normal/log_std Std 0.290349 +trainer/policy/normal/log_std Max 1.65236 +trainer/policy/normal/log_std Min -1.00381 +eval/num steps total 263502 +eval/num paths total 264 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.111996 +eval/Actions Std 0.923306 +eval/Actions Max 0.999993 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.72821 +time/logging (s) 0.00377931 +time/sampling batch (s) 0.283989 +time/saving (s) 0.00346123 +time/training (s) 7.33006 +time/epoch (s) 10.3495 +time/total (s) 2705.68 +Epoch -737 +---------------------------------- --------------- +2022-05-10 13:55:53.399045 PDT | [2] Epoch -736 finished +---------------------------------- --------------- +epoch -736 +replay_buffer/size 999033 +trainer/num train calls 265000 +trainer/Policy Loss -20.074 +trainer/Log Pis Mean 24.2906 +trainer/Log Pis Std 13.1185 +trainer/Log Pis Max 69.6555 +trainer/Log Pis Min -7.45117 +trainer/policy/mean Mean -0.0306146 +trainer/policy/mean Std 0.9058 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.8796 +trainer/policy/normal/std Std 0.612332 +trainer/policy/normal/std Max 5.23366 +trainer/policy/normal/std Min 0.39775 +trainer/policy/normal/log_std Mean 1.02644 +trainer/policy/normal/log_std Std 0.2779 +trainer/policy/normal/log_std Max 1.65511 +trainer/policy/normal/log_std Min -0.921932 +eval/num steps total 264502 +eval/num paths total 265 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.010396 +eval/Actions Std 0.903999 +eval/Actions Max 0.999991 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62628 +time/logging (s) 0.003679 +time/sampling batch (s) 0.283628 +time/saving (s) 0.00348906 +time/training (s) 6.81904 +time/epoch (s) 9.73612 +time/total (s) 2715.42 +Epoch -736 +---------------------------------- --------------- +2022-05-10 13:56:02.208787 PDT | [2] Epoch -735 finished +---------------------------------- --------------- +epoch -735 +replay_buffer/size 999033 +trainer/num train calls 266000 +trainer/Policy Loss -19.2393 +trainer/Log Pis Mean 25.5041 +trainer/Log Pis Std 14.4723 +trainer/Log Pis Max 68.4864 +trainer/Log Pis Min -6.81372 +trainer/policy/mean Mean -0.0402272 +trainer/policy/mean Std 0.902898 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.8264 +trainer/policy/normal/std Std 0.615569 +trainer/policy/normal/std Max 5.56254 +trainer/policy/normal/std Min 0.39773 +trainer/policy/normal/log_std Mean 1.00586 +trainer/policy/normal/log_std Std 0.287396 +trainer/policy/normal/log_std Max 1.71605 +trainer/policy/normal/log_std Min -0.921981 +eval/num steps total 265502 +eval/num paths total 266 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.107183 +eval/Actions Std 0.910765 +eval/Actions Max 0.999993 +eval/Actions Min -0.999985 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.44436 +time/logging (s) 0.00400564 +time/sampling batch (s) 0.283177 +time/saving (s) 0.00401067 +time/training (s) 6.05327 +time/epoch (s) 8.78882 +time/total (s) 2724.21 +Epoch -735 +---------------------------------- --------------- +2022-05-10 13:56:12.288499 PDT | [2] Epoch -734 finished +---------------------------------- --------------- +epoch -734 +replay_buffer/size 999033 +trainer/num train calls 267000 +trainer/Policy Loss -20.1623 +trainer/Log Pis Mean 24.6082 +trainer/Log Pis Std 13.3866 +trainer/Log Pis Max 71.8521 +trainer/Log Pis Min -9.46833 +trainer/policy/mean Mean -0.0285137 +trainer/policy/mean Std 0.906731 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.88121 +trainer/policy/normal/std Std 0.642015 +trainer/policy/normal/std Max 5.57854 +trainer/policy/normal/std Min 0.396964 +trainer/policy/normal/log_std Mean 1.02213 +trainer/policy/normal/log_std Std 0.303527 +trainer/policy/normal/log_std Max 1.71893 +trainer/policy/normal/log_std Min -0.92391 +eval/num steps total 266502 +eval/num paths total 267 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.192241 +eval/Actions Std 0.732676 +eval/Actions Max 0.99998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.85133 +time/logging (s) 0.00381821 +time/sampling batch (s) 0.281332 +time/saving (s) 0.00355908 +time/training (s) 6.91813 +time/epoch (s) 10.0582 +time/total (s) 2734.27 +Epoch -734 +---------------------------------- --------------- +2022-05-10 13:56:21.446658 PDT | [2] Epoch -733 finished +---------------------------------- --------------- +epoch -733 +replay_buffer/size 999033 +trainer/num train calls 268000 +trainer/Policy Loss -19.9306 +trainer/Log Pis Mean 23.8135 +trainer/Log Pis Std 12.7627 +trainer/Log Pis Max 73.6838 +trainer/Log Pis Min -4.38272 +trainer/policy/mean Mean -0.0393818 +trainer/policy/mean Std 0.900215 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.85102 +trainer/policy/normal/std Std 0.633904 +trainer/policy/normal/std Max 6.1665 +trainer/policy/normal/std Min 0.394649 +trainer/policy/normal/log_std Mean 1.01418 +trainer/policy/normal/log_std Std 0.287121 +trainer/policy/normal/log_std Max 1.81913 +trainer/policy/normal/log_std Min -0.929759 +eval/num steps total 267502 +eval/num paths total 268 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.208789 +eval/Actions Std 0.871253 +eval/Actions Max 0.999996 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.49847 +time/logging (s) 0.00382316 +time/sampling batch (s) 0.281924 +time/saving (s) 0.0035132 +time/training (s) 6.34942 +time/epoch (s) 9.13715 +time/total (s) 2743.41 +Epoch -733 +---------------------------------- --------------- +2022-05-10 13:56:32.141265 PDT | [2] Epoch -732 finished +---------------------------------- --------------- +epoch -732 +replay_buffer/size 999033 +trainer/num train calls 269000 +trainer/Policy Loss -20.018 +trainer/Log Pis Mean 24.8276 +trainer/Log Pis Std 13.0262 +trainer/Log Pis Max 66.9536 +trainer/Log Pis Min -4.90786 +trainer/policy/mean Mean -0.0440855 +trainer/policy/mean Std 0.904138 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.83602 +trainer/policy/normal/std Std 0.619992 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.409005 +trainer/policy/normal/log_std Mean 1.00968 +trainer/policy/normal/log_std Std 0.284165 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -0.894028 +eval/num steps total 268502 +eval/num paths total 269 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.118893 +eval/Actions Std 0.884597 +eval/Actions Max 0.999998 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.49606 +time/logging (s) 0.00369324 +time/sampling batch (s) 0.285423 +time/saving (s) 0.00346971 +time/training (s) 7.88432 +time/epoch (s) 10.673 +time/total (s) 2754.09 +Epoch -732 +---------------------------------- --------------- +2022-05-10 13:56:42.375039 PDT | [2] Epoch -731 finished +---------------------------------- --------------- +epoch -731 +replay_buffer/size 999033 +trainer/num train calls 270000 +trainer/Policy Loss -19.493 +trainer/Log Pis Mean 24.566 +trainer/Log Pis Std 12.5443 +trainer/Log Pis Max 61.1844 +trainer/Log Pis Min -6.41877 +trainer/policy/mean Mean -0.0271458 +trainer/policy/mean Std 0.909003 +trainer/policy/mean Max 0.999981 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.82242 +trainer/policy/normal/std Std 0.611921 +trainer/policy/normal/std Max 5.85252 +trainer/policy/normal/std Min 0.406674 +trainer/policy/normal/log_std Mean 1.00335 +trainer/policy/normal/log_std Std 0.295586 +trainer/policy/normal/log_std Max 1.76687 +trainer/policy/normal/log_std Min -0.899742 +eval/num steps total 269502 +eval/num paths total 270 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.282214 +eval/Actions Std 0.875207 +eval/Actions Max 0.999987 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.72528 +time/logging (s) 0.00377761 +time/sampling batch (s) 0.289576 +time/saving (s) 0.00349751 +time/training (s) 7.1902 +time/epoch (s) 10.2123 +time/total (s) 2764.31 +Epoch -731 +---------------------------------- --------------- +2022-05-10 13:56:51.807170 PDT | [2] Epoch -730 finished +---------------------------------- --------------- +epoch -730 +replay_buffer/size 999033 +trainer/num train calls 271000 +trainer/Policy Loss -18.9781 +trainer/Log Pis Mean 24.4017 +trainer/Log Pis Std 13.6193 +trainer/Log Pis Max 60.5377 +trainer/Log Pis Min -6.1027 +trainer/policy/mean Mean -0.0426954 +trainer/policy/mean Std 0.90549 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.78276 +trainer/policy/normal/std Std 0.605299 +trainer/policy/normal/std Max 5.5536 +trainer/policy/normal/std Min 0.42309 +trainer/policy/normal/log_std Mean 0.990481 +trainer/policy/normal/log_std Std 0.285821 +trainer/policy/normal/log_std Max 1.71445 +trainer/policy/normal/log_std Min -0.860171 +eval/num steps total 270502 +eval/num paths total 271 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.336451 +eval/Actions Std 0.833185 +eval/Actions Max 0.999995 +eval/Actions Min -0.999986 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.78186 +time/logging (s) 0.00413525 +time/sampling batch (s) 0.283302 +time/saving (s) 0.00405962 +time/training (s) 6.33794 +time/epoch (s) 9.4113 +time/total (s) 2773.72 +Epoch -730 +---------------------------------- --------------- +2022-05-10 13:57:02.107803 PDT | [2] Epoch -729 finished +---------------------------------- --------------- +epoch -729 +replay_buffer/size 999033 +trainer/num train calls 272000 +trainer/Policy Loss -19.5613 +trainer/Log Pis Mean 23.8488 +trainer/Log Pis Std 13.8892 +trainer/Log Pis Max 67.8475 +trainer/Log Pis Min -7.81614 +trainer/policy/mean Mean -0.0262783 +trainer/policy/mean Std 0.902918 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.81942 +trainer/policy/normal/std Std 0.627433 +trainer/policy/normal/std Max 5.71321 +trainer/policy/normal/std Min 0.369187 +trainer/policy/normal/log_std Mean 1.0013 +trainer/policy/normal/log_std Std 0.298284 +trainer/policy/normal/log_std Max 1.74278 +trainer/policy/normal/log_std Min -0.996451 +eval/num steps total 271502 +eval/num paths total 272 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.104946 +eval/Actions Std 0.911193 +eval/Actions Max 0.999992 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59986 +time/logging (s) 0.00378272 +time/sampling batch (s) 0.285629 +time/saving (s) 0.00388256 +time/training (s) 7.38537 +time/epoch (s) 10.2785 +time/total (s) 2784 +Epoch -729 +---------------------------------- --------------- +2022-05-10 13:57:12.595916 PDT | [2] Epoch -728 finished +---------------------------------- --------------- +epoch -728 +replay_buffer/size 999033 +trainer/num train calls 273000 +trainer/Policy Loss -20.9756 +trainer/Log Pis Mean 24.3231 +trainer/Log Pis Std 13.7365 +trainer/Log Pis Max 68.7943 +trainer/Log Pis Min -12.7815 +trainer/policy/mean Mean -0.0281559 +trainer/policy/mean Std 0.911428 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.85016 +trainer/policy/normal/std Std 0.59707 +trainer/policy/normal/std Max 6.74261 +trainer/policy/normal/std Min 0.346047 +trainer/policy/normal/log_std Mean 1.01719 +trainer/policy/normal/log_std Std 0.273788 +trainer/policy/normal/log_std Max 1.90845 +trainer/policy/normal/log_std Min -1.06118 +eval/num steps total 272502 +eval/num paths total 273 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.127329 +eval/Actions Std 0.887652 +eval/Actions Max 0.999992 +eval/Actions Min -0.999987 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.75779 +time/logging (s) 0.00371676 +time/sampling batch (s) 0.283683 +time/saving (s) 0.00349096 +time/training (s) 7.41803 +time/epoch (s) 10.4667 +time/total (s) 2794.47 +Epoch -728 +---------------------------------- --------------- +2022-05-10 13:57:22.161517 PDT | [2] Epoch -727 finished +---------------------------------- -------------- +epoch -727 +replay_buffer/size 999033 +trainer/num train calls 274000 +trainer/Policy Loss -19.1844 +trainer/Log Pis Mean 24.2733 +trainer/Log Pis Std 13.0007 +trainer/Log Pis Max 61.1988 +trainer/Log Pis Min -6.12002 +trainer/policy/mean Mean -0.0335704 +trainer/policy/mean Std 0.905107 +trainer/policy/mean Max 0.999976 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.75964 +trainer/policy/normal/std Std 0.604768 +trainer/policy/normal/std Max 5.17936 +trainer/policy/normal/std Min 0.327722 +trainer/policy/normal/log_std Mean 0.981519 +trainer/policy/normal/log_std Std 0.288952 +trainer/policy/normal/log_std Max 1.64468 +trainer/policy/normal/log_std Min -1.11559 +eval/num steps total 273502 +eval/num paths total 274 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.157378 +eval/Actions Std 0.868498 +eval/Actions Max 0.999992 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.83213 +time/logging (s) 0.0038172 +time/sampling batch (s) 0.284754 +time/saving (s) 0.0034674 +time/training (s) 6.42037 +time/epoch (s) 9.54454 +time/total (s) 2804.02 +Epoch -727 +---------------------------------- -------------- +2022-05-10 13:57:31.949540 PDT | [2] Epoch -726 finished +---------------------------------- --------------- +epoch -726 +replay_buffer/size 999033 +trainer/num train calls 275000 +trainer/Policy Loss -19.0437 +trainer/Log Pis Mean 23.7375 +trainer/Log Pis Std 12.3595 +trainer/Log Pis Max 64.5459 +trainer/Log Pis Min -6.00447 +trainer/policy/mean Mean -0.0480243 +trainer/policy/mean Std 0.904847 +trainer/policy/mean Max 0.999972 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.78143 +trainer/policy/normal/std Std 0.604451 +trainer/policy/normal/std Max 6.45767 +trainer/policy/normal/std Min 0.348858 +trainer/policy/normal/log_std Mean 0.990197 +trainer/policy/normal/log_std Std 0.285187 +trainer/policy/normal/log_std Max 1.86527 +trainer/policy/normal/log_std Min -1.05309 +eval/num steps total 274502 +eval/num paths total 275 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.177939 +eval/Actions Std 0.89882 +eval/Actions Max 0.999998 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.85894 +time/logging (s) 0.00372158 +time/sampling batch (s) 0.283657 +time/saving (s) 0.00348019 +time/training (s) 6.61696 +time/epoch (s) 9.76676 +time/total (s) 2813.79 +Epoch -726 +---------------------------------- --------------- +2022-05-10 13:57:41.479431 PDT | [2] Epoch -725 finished +---------------------------------- --------------- +epoch -725 +replay_buffer/size 999033 +trainer/num train calls 276000 +trainer/Policy Loss -20.1886 +trainer/Log Pis Mean 24.5249 +trainer/Log Pis Std 14.2784 +trainer/Log Pis Max 79.7669 +trainer/Log Pis Min -10.361 +trainer/policy/mean Mean -0.0558961 +trainer/policy/mean Std 0.906145 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.80936 +trainer/policy/normal/std Std 0.632039 +trainer/policy/normal/std Max 5.61317 +trainer/policy/normal/std Min 0.351968 +trainer/policy/normal/log_std Mean 0.995105 +trainer/policy/normal/log_std Std 0.313899 +trainer/policy/normal/log_std Max 1.72512 +trainer/policy/normal/log_std Min -1.04421 +eval/num steps total 275502 +eval/num paths total 276 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.137794 +eval/Actions Std 0.887832 +eval/Actions Max 0.999993 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.87932 +time/logging (s) 0.00414111 +time/sampling batch (s) 0.284258 +time/saving (s) 0.00401465 +time/training (s) 6.33727 +time/epoch (s) 9.509 +time/total (s) 2823.3 +Epoch -725 +---------------------------------- --------------- +2022-05-10 13:57:51.070300 PDT | [2] Epoch -724 finished +---------------------------------- --------------- +epoch -724 +replay_buffer/size 999033 +trainer/num train calls 277000 +trainer/Policy Loss -19.2512 +trainer/Log Pis Mean 23.724 +trainer/Log Pis Std 13.7236 +trainer/Log Pis Max 86.0262 +trainer/Log Pis Min -3.92501 +trainer/policy/mean Mean -0.0355893 +trainer/policy/mean Std 0.904941 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999985 +trainer/policy/normal/std Mean 2.83146 +trainer/policy/normal/std Std 0.599141 +trainer/policy/normal/std Max 5.16376 +trainer/policy/normal/std Min 0.303221 +trainer/policy/normal/log_std Mean 1.00933 +trainer/policy/normal/log_std Std 0.28055 +trainer/policy/normal/log_std Max 1.64166 +trainer/policy/normal/log_std Min -1.19329 +eval/num steps total 276502 +eval/num paths total 277 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.231207 +eval/Actions Std 0.915281 +eval/Actions Max 0.999995 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63951 +time/logging (s) 0.00378763 +time/sampling batch (s) 0.283245 +time/saving (s) 0.00379245 +time/training (s) 6.63851 +time/epoch (s) 9.56885 +time/total (s) 2832.87 +Epoch -724 +---------------------------------- --------------- +2022-05-10 13:58:01.065599 PDT | [2] Epoch -723 finished +---------------------------------- --------------- +epoch -723 +replay_buffer/size 999033 +trainer/num train calls 278000 +trainer/Policy Loss -19.5727 +trainer/Log Pis Mean 23.9892 +trainer/Log Pis Std 13.369 +trainer/Log Pis Max 65.8616 +trainer/Log Pis Min -9.7088 +trainer/policy/mean Mean -0.0302523 +trainer/policy/mean Std 0.906147 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.85079 +trainer/policy/normal/std Std 0.628168 +trainer/policy/normal/std Max 6.72953 +trainer/policy/normal/std Min 0.388107 +trainer/policy/normal/log_std Mean 1.01179 +trainer/policy/normal/log_std Std 0.303581 +trainer/policy/normal/log_std Max 1.9065 +trainer/policy/normal/log_std Min -0.946475 +eval/num steps total 277502 +eval/num paths total 278 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0805635 +eval/Actions Std 0.873787 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.48664 +time/logging (s) 0.00366 +time/sampling batch (s) 0.282987 +time/saving (s) 0.00341881 +time/training (s) 7.19738 +time/epoch (s) 9.97408 +time/total (s) 2842.85 +Epoch -723 +---------------------------------- --------------- +2022-05-10 13:58:11.455149 PDT | [2] Epoch -722 finished +---------------------------------- --------------- +epoch -722 +replay_buffer/size 999033 +trainer/num train calls 279000 +trainer/Policy Loss -19.5593 +trainer/Log Pis Mean 24.4516 +trainer/Log Pis Std 13.3291 +trainer/Log Pis Max 61.831 +trainer/Log Pis Min -6.77863 +trainer/policy/mean Mean -0.0322815 +trainer/policy/mean Std 0.907264 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.81197 +trainer/policy/normal/std Std 0.614597 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.384519 +trainer/policy/normal/log_std Mean 1.0006 +trainer/policy/normal/log_std Std 0.288092 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -0.955761 +eval/num steps total 278502 +eval/num paths total 279 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.231331 +eval/Actions Std 0.889644 +eval/Actions Max 1 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73803 +time/logging (s) 0.00373115 +time/sampling batch (s) 0.283083 +time/saving (s) 0.00347179 +time/training (s) 7.34048 +time/epoch (s) 10.3688 +time/total (s) 2853.22 +Epoch -722 +---------------------------------- --------------- +2022-05-10 13:58:22.451262 PDT | [2] Epoch -721 finished +---------------------------------- --------------- +epoch -721 +replay_buffer/size 999033 +trainer/num train calls 280000 +trainer/Policy Loss -20.6979 +trainer/Log Pis Mean 25.2714 +trainer/Log Pis Std 13.8526 +trainer/Log Pis Max 70.2683 +trainer/Log Pis Min -6.70331 +trainer/policy/mean Mean -0.0292188 +trainer/policy/mean Std 0.908288 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.88737 +trainer/policy/normal/std Std 0.605279 +trainer/policy/normal/std Max 5.52782 +trainer/policy/normal/std Min 0.327442 +trainer/policy/normal/log_std Mean 1.02962 +trainer/policy/normal/log_std Std 0.27725 +trainer/policy/normal/log_std Max 1.70979 +trainer/policy/normal/log_std Min -1.11644 +eval/num steps total 279502 +eval/num paths total 280 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0418528 +eval/Actions Std 0.907925 +eval/Actions Max 0.999994 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5836 +time/logging (s) 0.00368805 +time/sampling batch (s) 0.290486 +time/saving (s) 0.00343236 +time/training (s) 8.09342 +time/epoch (s) 10.9746 +time/total (s) 2864.2 +Epoch -721 +---------------------------------- --------------- +2022-05-10 13:58:34.071250 PDT | [2] Epoch -720 finished +---------------------------------- --------------- +epoch -720 +replay_buffer/size 999033 +trainer/num train calls 281000 +trainer/Policy Loss -20.1361 +trainer/Log Pis Mean 23.8142 +trainer/Log Pis Std 13.3877 +trainer/Log Pis Max 71.3858 +trainer/Log Pis Min -8.81386 +trainer/policy/mean Mean -0.0429249 +trainer/policy/mean Std 0.903677 +trainer/policy/mean Max 0.999982 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.81059 +trainer/policy/normal/std Std 0.61502 +trainer/policy/normal/std Max 4.82217 +trainer/policy/normal/std Min 0.335189 +trainer/policy/normal/log_std Mean 0.999163 +trainer/policy/normal/log_std Std 0.293818 +trainer/policy/normal/log_std Max 1.57322 +trainer/policy/normal/log_std Min -1.09306 +eval/num steps total 280502 +eval/num paths total 281 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0679928 +eval/Actions Std 0.890628 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5695 +time/logging (s) 0.00371873 +time/sampling batch (s) 0.536922 +time/saving (s) 0.00346811 +time/training (s) 8.48511 +time/epoch (s) 11.5987 +time/total (s) 2875.8 +Epoch -720 +---------------------------------- --------------- +2022-05-10 13:58:44.347593 PDT | [2] Epoch -719 finished +---------------------------------- --------------- +epoch -719 +replay_buffer/size 999033 +trainer/num train calls 282000 +trainer/Policy Loss -19.3128 +trainer/Log Pis Mean 25.3534 +trainer/Log Pis Std 13.8473 +trainer/Log Pis Max 76.2554 +trainer/Log Pis Min -5.84524 +trainer/policy/mean Mean -0.0319828 +trainer/policy/mean Std 0.909838 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.85805 +trainer/policy/normal/std Std 0.597549 +trainer/policy/normal/std Max 5.1927 +trainer/policy/normal/std Min 0.396991 +trainer/policy/normal/log_std Mean 1.01968 +trainer/policy/normal/log_std Std 0.275585 +trainer/policy/normal/log_std Max 1.64725 +trainer/policy/normal/log_std Min -0.923841 +eval/num steps total 281502 +eval/num paths total 282 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0445276 +eval/Actions Std 0.903632 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61042 +time/logging (s) 0.00409492 +time/sampling batch (s) 0.282255 +time/saving (s) 0.00397442 +time/training (s) 7.35487 +time/epoch (s) 10.2556 +time/total (s) 2886.06 +Epoch -719 +---------------------------------- --------------- +2022-05-10 13:58:54.127215 PDT | [2] Epoch -718 finished +---------------------------------- --------------- +epoch -718 +replay_buffer/size 999033 +trainer/num train calls 283000 +trainer/Policy Loss -18.2314 +trainer/Log Pis Mean 24.0935 +trainer/Log Pis Std 13.7308 +trainer/Log Pis Max 63.3183 +trainer/Log Pis Min -8.39554 +trainer/policy/mean Mean -0.0291593 +trainer/policy/mean Std 0.902223 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.86441 +trainer/policy/normal/std Std 0.625396 +trainer/policy/normal/std Max 5.59443 +trainer/policy/normal/std Min 0.434147 +trainer/policy/normal/log_std Mean 1.01944 +trainer/policy/normal/log_std Std 0.285175 +trainer/policy/normal/log_std Max 1.72177 +trainer/policy/normal/log_std Min -0.834372 +eval/num steps total 282502 +eval/num paths total 283 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.11077 +eval/Actions Std 0.885857 +eval/Actions Max 0.999982 +eval/Actions Min -0.999966 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.82635 +time/logging (s) 0.00383084 +time/sampling batch (s) 0.279664 +time/saving (s) 0.0036986 +time/training (s) 6.64422 +time/epoch (s) 9.75777 +time/total (s) 2895.82 +Epoch -718 +---------------------------------- --------------- +2022-05-10 13:59:03.985774 PDT | [2] Epoch -717 finished +---------------------------------- --------------- +epoch -717 +replay_buffer/size 999033 +trainer/num train calls 284000 +trainer/Policy Loss -19.0558 +trainer/Log Pis Mean 25.6334 +trainer/Log Pis Std 13.648 +trainer/Log Pis Max 74.7795 +trainer/Log Pis Min -7.47447 +trainer/policy/mean Mean -0.0494593 +trainer/policy/mean Std 0.905772 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.8439 +trainer/policy/normal/std Std 0.631955 +trainer/policy/normal/std Max 5.16565 +trainer/policy/normal/std Min 0.318266 +trainer/policy/normal/log_std Mean 1.00923 +trainer/policy/normal/log_std Std 0.302176 +trainer/policy/normal/log_std Max 1.64203 +trainer/policy/normal/log_std Min -1.14487 +eval/num steps total 283502 +eval/num paths total 284 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0939978 +eval/Actions Std 0.925024 +eval/Actions Max 0.999988 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.44201 +time/logging (s) 0.00370753 +time/sampling batch (s) 0.286915 +time/saving (s) 0.00350084 +time/training (s) 7.1008 +time/epoch (s) 9.83693 +time/total (s) 2905.66 +Epoch -717 +---------------------------------- --------------- +2022-05-10 13:59:13.926657 PDT | [2] Epoch -716 finished +---------------------------------- --------------- +epoch -716 +replay_buffer/size 999033 +trainer/num train calls 285000 +trainer/Policy Loss -18.9677 +trainer/Log Pis Mean 25.1481 +trainer/Log Pis Std 13.8407 +trainer/Log Pis Max 69.049 +trainer/Log Pis Min -10.1418 +trainer/policy/mean Mean -0.0395941 +trainer/policy/mean Std 0.906731 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.90059 +trainer/policy/normal/std Std 0.621999 +trainer/policy/normal/std Max 5.39376 +trainer/policy/normal/std Min 0.3788 +trainer/policy/normal/log_std Mean 1.03372 +trainer/policy/normal/log_std Std 0.276651 +trainer/policy/normal/log_std Max 1.68524 +trainer/policy/normal/log_std Min -0.970748 +eval/num steps total 284502 +eval/num paths total 285 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0957893 +eval/Actions Std 0.899508 +eval/Actions Max 0.999996 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.84417 +time/logging (s) 0.00370879 +time/sampling batch (s) 0.285419 +time/saving (s) 0.00360311 +time/training (s) 6.78234 +time/epoch (s) 9.91925 +time/total (s) 2915.58 +Epoch -716 +---------------------------------- --------------- +2022-05-10 13:59:23.680012 PDT | [2] Epoch -715 finished +---------------------------------- --------------- +epoch -715 +replay_buffer/size 999033 +trainer/num train calls 286000 +trainer/Policy Loss -18.7411 +trainer/Log Pis Mean 24.486 +trainer/Log Pis Std 13.1449 +trainer/Log Pis Max 76.7284 +trainer/Log Pis Min -10.3108 +trainer/policy/mean Mean -0.0248187 +trainer/policy/mean Std 0.9052 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.84861 +trainer/policy/normal/std Std 0.605412 +trainer/policy/normal/std Max 5.50474 +trainer/policy/normal/std Min 0.30651 +trainer/policy/normal/log_std Mean 1.01512 +trainer/policy/normal/log_std Std 0.281576 +trainer/policy/normal/log_std Max 1.70561 +trainer/policy/normal/log_std Min -1.1825 +eval/num steps total 285502 +eval/num paths total 286 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.186735 +eval/Actions Std 0.936663 +eval/Actions Max 0.999988 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59118 +time/logging (s) 0.0039068 +time/sampling batch (s) 0.288318 +time/saving (s) 0.00407663 +time/training (s) 6.84451 +time/epoch (s) 9.73199 +time/total (s) 2925.32 +Epoch -715 +---------------------------------- --------------- +2022-05-10 13:59:33.507329 PDT | [2] Epoch -714 finished +---------------------------------- --------------- +epoch -714 +replay_buffer/size 999033 +trainer/num train calls 287000 +trainer/Policy Loss -19.8555 +trainer/Log Pis Mean 24.2385 +trainer/Log Pis Std 13.0912 +trainer/Log Pis Max 66.2441 +trainer/Log Pis Min -6.76961 +trainer/policy/mean Mean -0.0188396 +trainer/policy/mean Std 0.906317 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.81925 +trainer/policy/normal/std Std 0.609168 +trainer/policy/normal/std Max 6.49809 +trainer/policy/normal/std Min 0.302503 +trainer/policy/normal/log_std Mean 1.00317 +trainer/policy/normal/log_std Std 0.290273 +trainer/policy/normal/log_std Max 1.87151 +trainer/policy/normal/log_std Min -1.19566 +eval/num steps total 286502 +eval/num paths total 287 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0366053 +eval/Actions Std 0.900615 +eval/Actions Max 0.999996 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.88655 +time/logging (s) 0.00422994 +time/sampling batch (s) 0.30466 +time/saving (s) 0.00436618 +time/training (s) 6.60606 +time/epoch (s) 9.80586 +time/total (s) 2935.13 +Epoch -714 +---------------------------------- --------------- +2022-05-10 13:59:43.665423 PDT | [2] Epoch -713 finished +---------------------------------- --------------- +epoch -713 +replay_buffer/size 999033 +trainer/num train calls 288000 +trainer/Policy Loss -19.1709 +trainer/Log Pis Mean 25.3158 +trainer/Log Pis Std 13.171 +trainer/Log Pis Max 61.9902 +trainer/Log Pis Min -5.04779 +trainer/policy/mean Mean -0.0325887 +trainer/policy/mean Std 0.906253 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.82448 +trainer/policy/normal/std Std 0.63201 +trainer/policy/normal/std Max 5.20049 +trainer/policy/normal/std Min 0.303128 +trainer/policy/normal/log_std Mean 1.00136 +trainer/policy/normal/log_std Std 0.308497 +trainer/policy/normal/log_std Max 1.64875 +trainer/policy/normal/log_std Min -1.1936 +eval/num steps total 287502 +eval/num paths total 288 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0790364 +eval/Actions Std 0.902101 +eval/Actions Max 0.999998 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6449 +time/logging (s) 0.00403155 +time/sampling batch (s) 0.530723 +time/saving (s) 0.00409438 +time/training (s) 6.95266 +time/epoch (s) 10.1364 +time/total (s) 2945.27 +Epoch -713 +---------------------------------- --------------- +2022-05-10 13:59:53.393134 PDT | [2] Epoch -712 finished +---------------------------------- --------------- +epoch -712 +replay_buffer/size 999033 +trainer/num train calls 289000 +trainer/Policy Loss -20.1853 +trainer/Log Pis Mean 24.5593 +trainer/Log Pis Std 12.886 +trainer/Log Pis Max 74.2024 +trainer/Log Pis Min -9.45311 +trainer/policy/mean Mean -0.0291172 +trainer/policy/mean Std 0.904205 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.8119 +trainer/policy/normal/std Std 0.617823 +trainer/policy/normal/std Max 5.21473 +trainer/policy/normal/std Min 0.316295 +trainer/policy/normal/log_std Mean 0.998898 +trainer/policy/normal/log_std Std 0.298644 +trainer/policy/normal/log_std Max 1.65149 +trainer/policy/normal/log_std Min -1.15108 +eval/num steps total 288502 +eval/num paths total 289 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.111107 +eval/Actions Std 0.917434 +eval/Actions Max 0.999993 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.48412 +time/logging (s) 0.00377688 +time/sampling batch (s) 0.289962 +time/saving (s) 0.0040622 +time/training (s) 6.92397 +time/epoch (s) 9.70589 +time/total (s) 2954.98 +Epoch -712 +---------------------------------- --------------- +2022-05-10 14:00:03.215346 PDT | [2] Epoch -711 finished +---------------------------------- --------------- +epoch -711 +replay_buffer/size 999033 +trainer/num train calls 290000 +trainer/Policy Loss -19.1276 +trainer/Log Pis Mean 24.1085 +trainer/Log Pis Std 14.1516 +trainer/Log Pis Max 89.9707 +trainer/Log Pis Min -7.36349 +trainer/policy/mean Mean -0.0227805 +trainer/policy/mean Std 0.910587 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.89027 +trainer/policy/normal/std Std 0.625105 +trainer/policy/normal/std Max 5.33436 +trainer/policy/normal/std Min 0.315718 +trainer/policy/normal/log_std Mean 1.02885 +trainer/policy/normal/log_std Std 0.284977 +trainer/policy/normal/log_std Max 1.67417 +trainer/policy/normal/log_std Min -1.15291 +eval/num steps total 289502 +eval/num paths total 290 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0499787 +eval/Actions Std 0.892384 +eval/Actions Max 0.999992 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5001 +time/logging (s) 0.00405931 +time/sampling batch (s) 0.536749 +time/saving (s) 0.00410147 +time/training (s) 6.75558 +time/epoch (s) 9.80059 +time/total (s) 2964.78 +Epoch -711 +---------------------------------- --------------- +2022-05-10 14:00:13.073010 PDT | [2] Epoch -710 finished +---------------------------------- --------------- +epoch -710 +replay_buffer/size 999033 +trainer/num train calls 291000 +trainer/Policy Loss -19.5082 +trainer/Log Pis Mean 24.5716 +trainer/Log Pis Std 13.0272 +trainer/Log Pis Max 89.0788 +trainer/Log Pis Min -10.3334 +trainer/policy/mean Mean -0.0371833 +trainer/policy/mean Std 0.905256 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999985 +trainer/policy/normal/std Mean 2.81665 +trainer/policy/normal/std Std 0.621007 +trainer/policy/normal/std Max 5.67171 +trainer/policy/normal/std Min 0.344151 +trainer/policy/normal/log_std Mean 1.00167 +trainer/policy/normal/log_std Std 0.290192 +trainer/policy/normal/log_std Max 1.73549 +trainer/policy/normal/log_std Min -1.06668 +eval/num steps total 290502 +eval/num paths total 291 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0330228 +eval/Actions Std 0.906204 +eval/Actions Max 0.999993 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.7752 +time/logging (s) 0.00367299 +time/sampling batch (s) 0.285207 +time/saving (s) 0.00357854 +time/training (s) 6.76791 +time/epoch (s) 9.83557 +time/total (s) 2974.62 +Epoch -710 +---------------------------------- --------------- +2022-05-10 14:00:23.270438 PDT | [2] Epoch -709 finished +---------------------------------- --------------- +epoch -709 +replay_buffer/size 999033 +trainer/num train calls 292000 +trainer/Policy Loss -20.7095 +trainer/Log Pis Mean 24.6434 +trainer/Log Pis Std 13.2762 +trainer/Log Pis Max 66.0209 +trainer/Log Pis Min -3.64543 +trainer/policy/mean Mean -0.0354184 +trainer/policy/mean Std 0.904954 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.86169 +trainer/policy/normal/std Std 0.627992 +trainer/policy/normal/std Max 5.66235 +trainer/policy/normal/std Min 0.315719 +trainer/policy/normal/log_std Mean 1.01795 +trainer/policy/normal/log_std Std 0.289492 +trainer/policy/normal/log_std Max 1.73384 +trainer/policy/normal/log_std Min -1.1529 +eval/num steps total 291502 +eval/num paths total 292 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.129361 +eval/Actions Std 0.882849 +eval/Actions Max 0.99999 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6646 +time/logging (s) 0.00453958 +time/sampling batch (s) 0.287907 +time/saving (s) 0.00409596 +time/training (s) 7.21552 +time/epoch (s) 10.1767 +time/total (s) 2984.8 +Epoch -709 +---------------------------------- --------------- +2022-05-10 14:00:33.995881 PDT | [2] Epoch -708 finished +---------------------------------- --------------- +epoch -708 +replay_buffer/size 999033 +trainer/num train calls 293000 +trainer/Policy Loss -20.2244 +trainer/Log Pis Mean 24.7299 +trainer/Log Pis Std 13.2825 +trainer/Log Pis Max 77.1677 +trainer/Log Pis Min -6.85519 +trainer/policy/mean Mean -0.0318692 +trainer/policy/mean Std 0.90837 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.85066 +trainer/policy/normal/std Std 0.573991 +trainer/policy/normal/std Max 6.49572 +trainer/policy/normal/std Min 0.42676 +trainer/policy/normal/log_std Mean 1.02086 +trainer/policy/normal/log_std Std 0.253432 +trainer/policy/normal/log_std Max 1.87114 +trainer/policy/normal/log_std Min -0.851534 +eval/num steps total 292502 +eval/num paths total 293 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.156305 +eval/Actions Std 0.889293 +eval/Actions Max 0.999981 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6583 +time/logging (s) 0.00379774 +time/sampling batch (s) 0.532954 +time/saving (s) 0.00373768 +time/training (s) 7.50338 +time/epoch (s) 10.7022 +time/total (s) 2995.5 +Epoch -708 +---------------------------------- --------------- +2022-05-10 14:00:44.700610 PDT | [2] Epoch -707 finished +---------------------------------- --------------- +epoch -707 +replay_buffer/size 999033 +trainer/num train calls 294000 +trainer/Policy Loss -20.0605 +trainer/Log Pis Mean 25.1929 +trainer/Log Pis Std 12.8231 +trainer/Log Pis Max 61.8538 +trainer/Log Pis Min -8.17633 +trainer/policy/mean Mean -0.0593533 +trainer/policy/mean Std 0.906874 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.83107 +trainer/policy/normal/std Std 0.606069 +trainer/policy/normal/std Max 5.88863 +trainer/policy/normal/std Min 0.302476 +trainer/policy/normal/log_std Mean 1.00868 +trainer/policy/normal/log_std Std 0.283896 +trainer/policy/normal/log_std Max 1.77302 +trainer/policy/normal/log_std Min -1.19575 +eval/num steps total 293502 +eval/num paths total 294 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.014625 +eval/Actions Std 0.853657 +eval/Actions Max 0.999947 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73966 +time/logging (s) 0.00385672 +time/sampling batch (s) 0.534719 +time/saving (s) 0.00351034 +time/training (s) 7.40154 +time/epoch (s) 10.6833 +time/total (s) 3006.19 +Epoch -707 +---------------------------------- --------------- +2022-05-10 14:00:54.487409 PDT | [2] Epoch -706 finished +---------------------------------- --------------- +epoch -706 +replay_buffer/size 999033 +trainer/num train calls 295000 +trainer/Policy Loss -20.4652 +trainer/Log Pis Mean 24.7178 +trainer/Log Pis Std 13.538 +trainer/Log Pis Max 69.5615 +trainer/Log Pis Min -7.65033 +trainer/policy/mean Mean -0.049616 +trainer/policy/mean Std 0.908503 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999984 +trainer/policy/normal/std Mean 2.83317 +trainer/policy/normal/std Std 0.632968 +trainer/policy/normal/std Max 6.23624 +trainer/policy/normal/std Min 0.412533 +trainer/policy/normal/log_std Mean 1.00492 +trainer/policy/normal/log_std Std 0.304989 +trainer/policy/normal/log_std Max 1.83038 +trainer/policy/normal/log_std Min -0.885438 +eval/num steps total 294502 +eval/num paths total 295 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0608987 +eval/Actions Std 0.912091 +eval/Actions Max 0.999994 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.75727 +time/logging (s) 0.00372567 +time/sampling batch (s) 0.284828 +time/saving (s) 0.00347907 +time/training (s) 6.71482 +time/epoch (s) 9.76412 +time/total (s) 3015.96 +Epoch -706 +---------------------------------- --------------- +2022-05-10 14:01:04.879970 PDT | [2] Epoch -705 finished +---------------------------------- --------------- +epoch -705 +replay_buffer/size 999033 +trainer/num train calls 296000 +trainer/Policy Loss -20.6942 +trainer/Log Pis Mean 24.7893 +trainer/Log Pis Std 13.9766 +trainer/Log Pis Max 67.7721 +trainer/Log Pis Min -3.97683 +trainer/policy/mean Mean -0.0513415 +trainer/policy/mean Std 0.912089 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.89798 +trainer/policy/normal/std Std 0.605319 +trainer/policy/normal/std Max 5.53266 +trainer/policy/normal/std Min 0.36047 +trainer/policy/normal/log_std Mean 1.03282 +trainer/policy/normal/log_std Std 0.280572 +trainer/policy/normal/log_std Max 1.71067 +trainer/policy/normal/log_std Min -1.02035 +eval/num steps total 295502 +eval/num paths total 296 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0128925 +eval/Actions Std 0.903004 +eval/Actions Max 0.999998 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.45199 +time/logging (s) 0.00380563 +time/sampling batch (s) 0.280595 +time/saving (s) 0.00347631 +time/training (s) 7.63185 +time/epoch (s) 10.3717 +time/total (s) 3026.33 +Epoch -705 +---------------------------------- --------------- +2022-05-10 14:01:14.853570 PDT | [2] Epoch -704 finished +---------------------------------- --------------- +epoch -704 +replay_buffer/size 999033 +trainer/num train calls 297000 +trainer/Policy Loss -19.8084 +trainer/Log Pis Mean 24.6878 +trainer/Log Pis Std 12.8186 +trainer/Log Pis Max 66.8164 +trainer/Log Pis Min -9.24352 +trainer/policy/mean Mean -0.0307663 +trainer/policy/mean Std 0.908235 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.90696 +trainer/policy/normal/std Std 0.594805 +trainer/policy/normal/std Max 5.58577 +trainer/policy/normal/std Min 0.351176 +trainer/policy/normal/log_std Mean 1.03818 +trainer/policy/normal/log_std Std 0.268732 +trainer/policy/normal/log_std Max 1.72022 +trainer/policy/normal/log_std Min -1.04647 +eval/num steps total 296502 +eval/num paths total 297 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.105585 +eval/Actions Std 0.824596 +eval/Actions Max 0.99999 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.47464 +time/logging (s) 0.00416453 +time/sampling batch (s) 0.282545 +time/saving (s) 0.00419454 +time/training (s) 7.18724 +time/epoch (s) 9.95279 +time/total (s) 3036.29 +Epoch -704 +---------------------------------- --------------- +2022-05-10 14:01:24.610095 PDT | [2] Epoch -703 finished +---------------------------------- --------------- +epoch -703 +replay_buffer/size 999033 +trainer/num train calls 298000 +trainer/Policy Loss -18.7759 +trainer/Log Pis Mean 23.8049 +trainer/Log Pis Std 13.3068 +trainer/Log Pis Max 70.9887 +trainer/Log Pis Min -8.95257 +trainer/policy/mean Mean -0.036407 +trainer/policy/mean Std 0.911403 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999985 +trainer/policy/normal/std Mean 2.8559 +trainer/policy/normal/std Std 0.624315 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.329421 +trainer/policy/normal/log_std Mean 1.01578 +trainer/policy/normal/log_std Std 0.290386 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.11042 +eval/num steps total 297502 +eval/num paths total 298 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.199543 +eval/Actions Std 0.897057 +eval/Actions Max 0.99999 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.7124 +time/logging (s) 0.00383463 +time/sampling batch (s) 0.280503 +time/saving (s) 0.00375907 +time/training (s) 6.73427 +time/epoch (s) 9.73477 +time/total (s) 3046.03 +Epoch -703 +---------------------------------- --------------- +2022-05-10 14:01:35.780009 PDT | [2] Epoch -702 finished +---------------------------------- --------------- +epoch -702 +replay_buffer/size 999033 +trainer/num train calls 299000 +trainer/Policy Loss -18.4437 +trainer/Log Pis Mean 24.5226 +trainer/Log Pis Std 13.2526 +trainer/Log Pis Max 64.7491 +trainer/Log Pis Min -4.00526 +trainer/policy/mean Mean -0.0440161 +trainer/policy/mean Std 0.903659 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.86065 +trainer/policy/normal/std Std 0.609291 +trainer/policy/normal/std Max 6.3859 +trainer/policy/normal/std Min 0.35987 +trainer/policy/normal/log_std Mean 1.02007 +trainer/policy/normal/log_std Std 0.276481 +trainer/policy/normal/log_std Max 1.85409 +trainer/policy/normal/log_std Min -1.02201 +eval/num steps total 298502 +eval/num paths total 299 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.053889 +eval/Actions Std 0.909531 +eval/Actions Max 0.999996 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59852 +time/logging (s) 0.00382036 +time/sampling batch (s) 0.281415 +time/saving (s) 0.00344046 +time/training (s) 8.2614 +time/epoch (s) 11.1486 +time/total (s) 3057.18 +Epoch -702 +---------------------------------- --------------- +2022-05-10 14:01:45.850450 PDT | [2] Epoch -701 finished +---------------------------------- --------------- +epoch -701 +replay_buffer/size 999033 +trainer/num train calls 300000 +trainer/Policy Loss -20.0491 +trainer/Log Pis Mean 24.1761 +trainer/Log Pis Std 13.4337 +trainer/Log Pis Max 69.3899 +trainer/Log Pis Min -7.6014 +trainer/policy/mean Mean -0.039774 +trainer/policy/mean Std 0.904882 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.88946 +trainer/policy/normal/std Std 0.628422 +trainer/policy/normal/std Max 6.2937 +trainer/policy/normal/std Min 0.353579 +trainer/policy/normal/log_std Mean 1.02793 +trainer/policy/normal/log_std Std 0.288531 +trainer/policy/normal/log_std Max 1.83955 +trainer/policy/normal/log_std Min -1.03965 +eval/num steps total 299502 +eval/num paths total 300 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.264604 +eval/Actions Std 0.893375 +eval/Actions Max 0.99999 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6941 +time/logging (s) 0.00393855 +time/sampling batch (s) 0.279197 +time/saving (s) 0.00642497 +time/training (s) 7.06604 +time/epoch (s) 10.0497 +time/total (s) 3067.23 +Epoch -701 +---------------------------------- --------------- +2022-05-10 14:01:56.035560 PDT | [2] Epoch -700 finished +---------------------------------- --------------- +epoch -700 +replay_buffer/size 999033 +trainer/num train calls 301000 +trainer/Policy Loss -19.6013 +trainer/Log Pis Mean 25.5094 +trainer/Log Pis Std 14.0673 +trainer/Log Pis Max 71.6823 +trainer/Log Pis Min -10.2383 +trainer/policy/mean Mean -0.0340287 +trainer/policy/mean Std 0.908627 +trainer/policy/mean Max 1 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.87556 +trainer/policy/normal/std Std 0.627695 +trainer/policy/normal/std Max 6.30713 +trainer/policy/normal/std Min 0.305651 +trainer/policy/normal/log_std Mean 1.0229 +trainer/policy/normal/log_std Std 0.289313 +trainer/policy/normal/log_std Max 1.84168 +trainer/policy/normal/log_std Min -1.18531 +eval/num steps total 300502 +eval/num paths total 301 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0183601 +eval/Actions Std 0.892694 +eval/Actions Max 0.999997 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.39379 +time/logging (s) 0.00371953 +time/sampling batch (s) 0.286859 +time/saving (s) 0.00345092 +time/training (s) 7.47575 +time/epoch (s) 10.1636 +time/total (s) 3077.4 +Epoch -700 +---------------------------------- --------------- +2022-05-10 14:02:06.524414 PDT | [2] Epoch -699 finished +---------------------------------- --------------- +epoch -699 +replay_buffer/size 999033 +trainer/num train calls 302000 +trainer/Policy Loss -19.5392 +trainer/Log Pis Mean 24.9429 +trainer/Log Pis Std 13.3487 +trainer/Log Pis Max 70.724 +trainer/Log Pis Min -12.1863 +trainer/policy/mean Mean -0.0503171 +trainer/policy/mean Std 0.904901 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.84944 +trainer/policy/normal/std Std 0.62393 +trainer/policy/normal/std Max 5.42009 +trainer/policy/normal/std Min 0.311636 +trainer/policy/normal/log_std Mean 1.01248 +trainer/policy/normal/log_std Std 0.296819 +trainer/policy/normal/log_std Max 1.69011 +trainer/policy/normal/log_std Min -1.16592 +eval/num steps total 301502 +eval/num paths total 302 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.144705 +eval/Actions Std 0.874 +eval/Actions Max 0.999984 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62563 +time/logging (s) 0.00416174 +time/sampling batch (s) 0.541569 +time/saving (s) 0.00405187 +time/training (s) 7.29193 +time/epoch (s) 10.4673 +time/total (s) 3087.87 +Epoch -699 +---------------------------------- --------------- +2022-05-10 14:02:16.715432 PDT | [2] Epoch -698 finished +---------------------------------- --------------- +epoch -698 +replay_buffer/size 999033 +trainer/num train calls 303000 +trainer/Policy Loss -18.8901 +trainer/Log Pis Mean 24.7591 +trainer/Log Pis Std 12.9575 +trainer/Log Pis Max 67.3323 +trainer/Log Pis Min -5.79467 +trainer/policy/mean Mean -0.0231176 +trainer/policy/mean Std 0.903597 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.8736 +trainer/policy/normal/std Std 0.608366 +trainer/policy/normal/std Max 5.98771 +trainer/policy/normal/std Min 0.415435 +trainer/policy/normal/log_std Mean 1.02529 +trainer/policy/normal/log_std Std 0.272283 +trainer/policy/normal/log_std Max 1.78971 +trainer/policy/normal/log_std Min -0.878429 +eval/num steps total 302502 +eval/num paths total 303 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0377634 +eval/Actions Std 0.917708 +eval/Actions Max 0.999994 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6241 +time/logging (s) 0.00408998 +time/sampling batch (s) 0.284826 +time/saving (s) 0.00396468 +time/training (s) 7.25212 +time/epoch (s) 10.1691 +time/total (s) 3098.04 +Epoch -698 +---------------------------------- --------------- +2022-05-10 14:02:27.329474 PDT | [2] Epoch -697 finished +---------------------------------- --------------- +epoch -697 +replay_buffer/size 999033 +trainer/num train calls 304000 +trainer/Policy Loss -18.8184 +trainer/Log Pis Mean 24.2445 +trainer/Log Pis Std 13.2105 +trainer/Log Pis Max 70.8482 +trainer/Log Pis Min -5.97178 +trainer/policy/mean Mean -0.0450129 +trainer/policy/mean Std 0.899311 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.86636 +trainer/policy/normal/std Std 0.647746 +trainer/policy/normal/std Max 5.56653 +trainer/policy/normal/std Min 0.389104 +trainer/policy/normal/log_std Mean 1.01704 +trainer/policy/normal/log_std Std 0.299964 +trainer/policy/normal/log_std Max 1.71677 +trainer/policy/normal/log_std Min -0.943908 +eval/num steps total 303502 +eval/num paths total 304 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.195588 +eval/Actions Std 0.918875 +eval/Actions Max 0.999988 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61964 +time/logging (s) 0.00376499 +time/sampling batch (s) 0.284151 +time/saving (s) 0.00361772 +time/training (s) 7.68086 +time/epoch (s) 10.592 +time/total (s) 3108.64 +Epoch -697 +---------------------------------- --------------- +2022-05-10 14:02:37.095395 PDT | [2] Epoch -696 finished +---------------------------------- --------------- +epoch -696 +replay_buffer/size 999033 +trainer/num train calls 305000 +trainer/Policy Loss -18.5739 +trainer/Log Pis Mean 25.1999 +trainer/Log Pis Std 14.2516 +trainer/Log Pis Max 72.3705 +trainer/Log Pis Min -4.86501 +trainer/policy/mean Mean -0.0445903 +trainer/policy/mean Std 0.905546 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.85597 +trainer/policy/normal/std Std 0.634468 +trainer/policy/normal/std Max 5.88755 +trainer/policy/normal/std Min 0.442885 +trainer/policy/normal/log_std Mean 1.01505 +trainer/policy/normal/log_std Std 0.291679 +trainer/policy/normal/log_std Max 1.77284 +trainer/policy/normal/log_std Min -0.814444 +eval/num steps total 304502 +eval/num paths total 305 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.104264 +eval/Actions Std 0.91103 +eval/Actions Max 0.999989 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59613 +time/logging (s) 0.00385859 +time/sampling batch (s) 0.290925 +time/saving (s) 0.0035262 +time/training (s) 6.85056 +time/epoch (s) 9.745 +time/total (s) 3118.38 +Epoch -696 +---------------------------------- --------------- +2022-05-10 14:02:47.304691 PDT | [2] Epoch -695 finished +---------------------------------- --------------- +epoch -695 +replay_buffer/size 999033 +trainer/num train calls 306000 +trainer/Policy Loss -20.4482 +trainer/Log Pis Mean 24.7003 +trainer/Log Pis Std 12.8731 +trainer/Log Pis Max 71.3564 +trainer/Log Pis Min -7.90962 +trainer/policy/mean Mean -0.0313681 +trainer/policy/mean Std 0.905723 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.8741 +trainer/policy/normal/std Std 0.614418 +trainer/policy/normal/std Max 6.37567 +trainer/policy/normal/std Min 0.358124 +trainer/policy/normal/log_std Mean 1.02374 +trainer/policy/normal/log_std Std 0.283839 +trainer/policy/normal/log_std Max 1.85249 +trainer/policy/normal/log_std Min -1.02688 +eval/num steps total 305502 +eval/num paths total 306 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.138554 +eval/Actions Std 0.913823 +eval/Actions Max 0.999995 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58531 +time/logging (s) 0.00374848 +time/sampling batch (s) 0.28307 +time/saving (s) 0.00365054 +time/training (s) 7.31201 +time/epoch (s) 10.1878 +time/total (s) 3128.57 +Epoch -695 +---------------------------------- --------------- +2022-05-10 14:02:57.696617 PDT | [2] Epoch -694 finished +---------------------------------- --------------- +epoch -694 +replay_buffer/size 999033 +trainer/num train calls 307000 +trainer/Policy Loss -19.425 +trainer/Log Pis Mean 24.5629 +trainer/Log Pis Std 13.6842 +trainer/Log Pis Max 68.1917 +trainer/Log Pis Min -11.2769 +trainer/policy/mean Mean -0.0468875 +trainer/policy/mean Std 0.902579 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.86273 +trainer/policy/normal/std Std 0.648271 +trainer/policy/normal/std Max 5.33162 +trainer/policy/normal/std Min 0.377632 +trainer/policy/normal/log_std Mean 1.01618 +trainer/policy/normal/log_std Std 0.29655 +trainer/policy/normal/log_std Max 1.67365 +trainer/policy/normal/log_std Min -0.973836 +eval/num steps total 306502 +eval/num paths total 307 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.182575 +eval/Actions Std 0.910162 +eval/Actions Max 0.999995 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60633 +time/logging (s) 0.00367137 +time/sampling batch (s) 0.284193 +time/saving (s) 0.00346551 +time/training (s) 7.47284 +time/epoch (s) 10.3705 +time/total (s) 3138.95 +Epoch -694 +---------------------------------- --------------- +2022-05-10 14:03:07.887481 PDT | [2] Epoch -693 finished +---------------------------------- --------------- +epoch -693 +replay_buffer/size 999033 +trainer/num train calls 308000 +trainer/Policy Loss -20.9009 +trainer/Log Pis Mean 24.8044 +trainer/Log Pis Std 13.2273 +trainer/Log Pis Max 71.8326 +trainer/Log Pis Min -9.40483 +trainer/policy/mean Mean -0.0547849 +trainer/policy/mean Std 0.904721 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.89251 +trainer/policy/normal/std Std 0.615901 +trainer/policy/normal/std Max 5.17243 +trainer/policy/normal/std Min 0.384079 +trainer/policy/normal/log_std Mean 1.03117 +trainer/policy/normal/log_std Std 0.275439 +trainer/policy/normal/log_std Max 1.64334 +trainer/policy/normal/log_std Min -0.956906 +eval/num steps total 307502 +eval/num paths total 308 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0147429 +eval/Actions Std 0.904074 +eval/Actions Max 0.999994 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5871 +time/logging (s) 0.00397856 +time/sampling batch (s) 0.284845 +time/saving (s) 0.0041949 +time/training (s) 7.28996 +time/epoch (s) 10.1701 +time/total (s) 3149.12 +Epoch -693 +---------------------------------- --------------- +2022-05-10 14:03:18.378074 PDT | [2] Epoch -692 finished +---------------------------------- --------------- +epoch -692 +replay_buffer/size 999033 +trainer/num train calls 309000 +trainer/Policy Loss -19.176 +trainer/Log Pis Mean 25.0909 +trainer/Log Pis Std 13.4672 +trainer/Log Pis Max 63.5143 +trainer/Log Pis Min -9.47672 +trainer/policy/mean Mean -0.0367206 +trainer/policy/mean Std 0.909996 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.84931 +trainer/policy/normal/std Std 0.631691 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.347834 +trainer/policy/normal/log_std Mean 1.01296 +trainer/policy/normal/log_std Std 0.291332 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.05603 +eval/num steps total 308502 +eval/num paths total 309 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.212101 +eval/Actions Std 0.743272 +eval/Actions Max 0.999991 +eval/Actions Min -0.999986 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54618 +time/logging (s) 0.0037632 +time/sampling batch (s) 0.280181 +time/saving (s) 0.00349021 +time/training (s) 7.63541 +time/epoch (s) 10.469 +time/total (s) 3159.59 +Epoch -692 +---------------------------------- --------------- +2022-05-10 14:03:28.396595 PDT | [2] Epoch -691 finished +---------------------------------- --------------- +epoch -691 +replay_buffer/size 999033 +trainer/num train calls 310000 +trainer/Policy Loss -19.1637 +trainer/Log Pis Mean 24.0004 +trainer/Log Pis Std 12.9451 +trainer/Log Pis Max 62.5148 +trainer/Log Pis Min -6.70863 +trainer/policy/mean Mean -0.0484013 +trainer/policy/mean Std 0.903078 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.783 +trainer/policy/normal/std Std 0.621689 +trainer/policy/normal/std Max 5.12585 +trainer/policy/normal/std Min 0.349403 +trainer/policy/normal/log_std Mean 0.987025 +trainer/policy/normal/log_std Std 0.305312 +trainer/policy/normal/log_std Max 1.6343 +trainer/policy/normal/log_std Min -1.05153 +eval/num steps total 309502 +eval/num paths total 310 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0464996 +eval/Actions Std 0.922681 +eval/Actions Max 0.999979 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5181 +time/logging (s) 0.00372545 +time/sampling batch (s) 0.282198 +time/saving (s) 0.00345145 +time/training (s) 7.18987 +time/epoch (s) 9.99735 +time/total (s) 3169.59 +Epoch -691 +---------------------------------- --------------- +2022-05-10 14:03:39.400955 PDT | [2] Epoch -690 finished +---------------------------------- --------------- +epoch -690 +replay_buffer/size 999033 +trainer/num train calls 311000 +trainer/Policy Loss -19.965 +trainer/Log Pis Mean 24.5298 +trainer/Log Pis Std 12.9024 +trainer/Log Pis Max 64.815 +trainer/Log Pis Min -3.73819 +trainer/policy/mean Mean -0.0242458 +trainer/policy/mean Std 0.908219 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.86945 +trainer/policy/normal/std Std 0.634912 +trainer/policy/normal/std Max 5.61543 +trainer/policy/normal/std Min 0.351904 +trainer/policy/normal/log_std Mean 1.01945 +trainer/policy/normal/log_std Std 0.295219 +trainer/policy/normal/log_std Max 1.72552 +trainer/policy/normal/log_std Min -1.0444 +eval/num steps total 310502 +eval/num paths total 311 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.240971 +eval/Actions Std 0.930426 +eval/Actions Max 0.999993 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.36919 +time/logging (s) 0.00370543 +time/sampling batch (s) 0.287346 +time/saving (s) 0.00345115 +time/training (s) 8.31893 +time/epoch (s) 10.9826 +time/total (s) 3180.58 +Epoch -690 +---------------------------------- --------------- +2022-05-10 14:03:49.805411 PDT | [2] Epoch -689 finished +---------------------------------- --------------- +epoch -689 +replay_buffer/size 999033 +trainer/num train calls 312000 +trainer/Policy Loss -20.6116 +trainer/Log Pis Mean 25.077 +trainer/Log Pis Std 12.7582 +trainer/Log Pis Max 67.2341 +trainer/Log Pis Min -3.882 +trainer/policy/mean Mean -0.0234645 +trainer/policy/mean Std 0.909416 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.83658 +trainer/policy/normal/std Std 0.618706 +trainer/policy/normal/std Max 5.19813 +trainer/policy/normal/std Min 0.348854 +trainer/policy/normal/log_std Mean 1.00865 +trainer/policy/normal/log_std Std 0.293162 +trainer/policy/normal/log_std Max 1.6483 +trainer/policy/normal/log_std Min -1.0531 +eval/num steps total 311502 +eval/num paths total 312 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0501504 +eval/Actions Std 0.891345 +eval/Actions Max 0.999994 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56145 +time/logging (s) 0.00381299 +time/sampling batch (s) 0.285332 +time/saving (s) 0.00348553 +time/training (s) 7.52902 +time/epoch (s) 10.3831 +time/total (s) 3190.97 +Epoch -689 +---------------------------------- --------------- +2022-05-10 14:04:00.463054 PDT | [2] Epoch -688 finished +---------------------------------- --------------- +epoch -688 +replay_buffer/size 999033 +trainer/num train calls 313000 +trainer/Policy Loss -19.1395 +trainer/Log Pis Mean 23.9933 +trainer/Log Pis Std 13.1506 +trainer/Log Pis Max 65.3171 +trainer/Log Pis Min -3.41647 +trainer/policy/mean Mean -0.0482693 +trainer/policy/mean Std 0.907535 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.82398 +trainer/policy/normal/std Std 0.62777 +trainer/policy/normal/std Max 5.71758 +trainer/policy/normal/std Min 0.327155 +trainer/policy/normal/log_std Mean 1.00233 +trainer/policy/normal/log_std Std 0.30211 +trainer/policy/normal/log_std Max 1.74355 +trainer/policy/normal/log_std Min -1.11732 +eval/num steps total 312502 +eval/num paths total 313 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0637395 +eval/Actions Std 0.911702 +eval/Actions Max 0.999996 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66254 +time/logging (s) 0.00396309 +time/sampling batch (s) 0.285297 +time/saving (s) 0.00398188 +time/training (s) 7.68055 +time/epoch (s) 10.6363 +time/total (s) 3201.6 +Epoch -688 +---------------------------------- --------------- +2022-05-10 14:04:10.684853 PDT | [2] Epoch -687 finished +---------------------------------- --------------- +epoch -687 +replay_buffer/size 999033 +trainer/num train calls 314000 +trainer/Policy Loss -19.409 +trainer/Log Pis Mean 24.9173 +trainer/Log Pis Std 13.4936 +trainer/Log Pis Max 72.0995 +trainer/Log Pis Min -8.46138 +trainer/policy/mean Mean -0.0276451 +trainer/policy/mean Std 0.907939 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999977 +trainer/policy/normal/std Mean 2.76051 +trainer/policy/normal/std Std 0.587139 +trainer/policy/normal/std Max 5.84852 +trainer/policy/normal/std Min 0.391841 +trainer/policy/normal/log_std Mean 0.983632 +trainer/policy/normal/log_std Std 0.281794 +trainer/policy/normal/log_std Max 1.76619 +trainer/policy/normal/log_std Min -0.936899 +eval/num steps total 313502 +eval/num paths total 314 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.292614 +eval/Actions Std 0.868909 +eval/Actions Max 0.999993 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67341 +time/logging (s) 0.00370826 +time/sampling batch (s) 0.283073 +time/saving (s) 0.00353031 +time/training (s) 7.23642 +time/epoch (s) 10.2001 +time/total (s) 3211.81 +Epoch -687 +---------------------------------- --------------- +2022-05-10 14:04:20.400267 PDT | [2] Epoch -686 finished +---------------------------------- --------------- +epoch -686 +replay_buffer/size 999033 +trainer/num train calls 315000 +trainer/Policy Loss -19.5712 +trainer/Log Pis Mean 25.5402 +trainer/Log Pis Std 12.7433 +trainer/Log Pis Max 68.5181 +trainer/Log Pis Min -4.75041 +trainer/policy/mean Mean -0.0382838 +trainer/policy/mean Std 0.908069 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.80761 +trainer/policy/normal/std Std 0.622086 +trainer/policy/normal/std Max 5.07978 +trainer/policy/normal/std Min 0.284551 +trainer/policy/normal/log_std Mean 0.996161 +trainer/policy/normal/log_std Std 0.305609 +trainer/policy/normal/log_std Max 1.62527 +trainer/policy/normal/log_std Min -1.25684 +eval/num steps total 314502 +eval/num paths total 315 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.079045 +eval/Actions Std 0.902886 +eval/Actions Max 0.999994 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57766 +time/logging (s) 0.00371321 +time/sampling batch (s) 0.283695 +time/saving (s) 0.003621 +time/training (s) 6.82535 +time/epoch (s) 9.69405 +time/total (s) 3221.5 +Epoch -686 +---------------------------------- --------------- +2022-05-10 14:04:31.184835 PDT | [2] Epoch -685 finished +---------------------------------- --------------- +epoch -685 +replay_buffer/size 999033 +trainer/num train calls 316000 +trainer/Policy Loss -20.0475 +trainer/Log Pis Mean 23.6878 +trainer/Log Pis Std 12.972 +trainer/Log Pis Max 65.2587 +trainer/Log Pis Min -8.70647 +trainer/policy/mean Mean -0.020844 +trainer/policy/mean Std 0.90799 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.81419 +trainer/policy/normal/std Std 0.614636 +trainer/policy/normal/std Max 5.97459 +trainer/policy/normal/std Min 0.403214 +trainer/policy/normal/log_std Mean 1.00177 +trainer/policy/normal/log_std Std 0.285021 +trainer/policy/normal/log_std Max 1.78751 +trainer/policy/normal/log_std Min -0.908287 +eval/num steps total 315502 +eval/num paths total 316 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.167801 +eval/Actions Std 0.924684 +eval/Actions Max 0.999994 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5009 +time/logging (s) 0.00368615 +time/sampling batch (s) 0.278877 +time/saving (s) 0.00343198 +time/training (s) 7.9767 +time/epoch (s) 10.7636 +time/total (s) 3232.27 +Epoch -685 +---------------------------------- --------------- +2022-05-10 14:04:42.864737 PDT | [2] Epoch -684 finished +---------------------------------- --------------- +epoch -684 +replay_buffer/size 999033 +trainer/num train calls 317000 +trainer/Policy Loss -19.2205 +trainer/Log Pis Mean 25.0144 +trainer/Log Pis Std 13.3611 +trainer/Log Pis Max 64.056 +trainer/Log Pis Min -11.9372 +trainer/policy/mean Mean -0.0463102 +trainer/policy/mean Std 0.907731 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.82918 +trainer/policy/normal/std Std 0.630378 +trainer/policy/normal/std Max 5.56308 +trainer/policy/normal/std Min 0.366461 +trainer/policy/normal/log_std Mean 1.00452 +trainer/policy/normal/log_std Std 0.299307 +trainer/policy/normal/log_std Max 1.71615 +trainer/policy/normal/log_std Min -1.00386 +eval/num steps total 316502 +eval/num paths total 317 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0557892 +eval/Actions Std 0.894014 +eval/Actions Max 0.999997 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61597 +time/logging (s) 0.00368153 +time/sampling batch (s) 0.528766 +time/saving (s) 0.00343153 +time/training (s) 8.50721 +time/epoch (s) 11.6591 +time/total (s) 3243.93 +Epoch -684 +---------------------------------- --------------- +2022-05-10 14:04:52.268387 PDT | [2] Epoch -683 finished +---------------------------------- --------------- +epoch -683 +replay_buffer/size 999033 +trainer/num train calls 318000 +trainer/Policy Loss -19.3475 +trainer/Log Pis Mean 24.6957 +trainer/Log Pis Std 12.9566 +trainer/Log Pis Max 72.8233 +trainer/Log Pis Min -8.4291 +trainer/policy/mean Mean -0.0125854 +trainer/policy/mean Std 0.904294 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.84766 +trainer/policy/normal/std Std 0.632445 +trainer/policy/normal/std Max 5.51802 +trainer/policy/normal/std Min 0.273631 +trainer/policy/normal/log_std Mean 1.01063 +trainer/policy/normal/log_std Std 0.303469 +trainer/policy/normal/log_std Max 1.70802 +trainer/policy/normal/log_std Min -1.29597 +eval/num steps total 317502 +eval/num paths total 318 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.190053 +eval/Actions Std 0.932799 +eval/Actions Max 1 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.28915 +time/logging (s) 0.00414242 +time/sampling batch (s) 0.278031 +time/saving (s) 0.00412847 +time/training (s) 6.80779 +time/epoch (s) 9.38324 +time/total (s) 3253.32 +Epoch -683 +---------------------------------- --------------- +2022-05-10 14:05:03.620811 PDT | [2] Epoch -682 finished +---------------------------------- --------------- +epoch -682 +replay_buffer/size 999033 +trainer/num train calls 319000 +trainer/Policy Loss -18.9392 +trainer/Log Pis Mean 23.7223 +trainer/Log Pis Std 13.0934 +trainer/Log Pis Max 67.3227 +trainer/Log Pis Min -3.74167 +trainer/policy/mean Mean -0.0474753 +trainer/policy/mean Std 0.904054 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.78373 +trainer/policy/normal/std Std 0.616662 +trainer/policy/normal/std Max 5.46361 +trainer/policy/normal/std Min 0.366398 +trainer/policy/normal/log_std Mean 0.989508 +trainer/policy/normal/log_std Std 0.292282 +trainer/policy/normal/log_std Max 1.69811 +trainer/policy/normal/log_std Min -1.00404 +eval/num steps total 318502 +eval/num paths total 319 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0908776 +eval/Actions Std 0.886391 +eval/Actions Max 0.999988 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61275 +time/logging (s) 0.00371741 +time/sampling batch (s) 0.529923 +time/saving (s) 0.00350642 +time/training (s) 8.18067 +time/epoch (s) 11.3306 +time/total (s) 3264.65 +Epoch -682 +---------------------------------- --------------- +2022-05-10 14:05:14.478387 PDT | [2] Epoch -681 finished +---------------------------------- --------------- +epoch -681 +replay_buffer/size 999033 +trainer/num train calls 320000 +trainer/Policy Loss -18.9784 +trainer/Log Pis Mean 24.2158 +trainer/Log Pis Std 13.7587 +trainer/Log Pis Max 86.9486 +trainer/Log Pis Min -12.5975 +trainer/policy/mean Mean -0.0511054 +trainer/policy/mean Std 0.900145 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.82476 +trainer/policy/normal/std Std 0.597212 +trainer/policy/normal/std Max 5.11423 +trainer/policy/normal/std Min 0.307665 +trainer/policy/normal/log_std Mean 1.0061 +trainer/policy/normal/log_std Std 0.287725 +trainer/policy/normal/log_std Max 1.63203 +trainer/policy/normal/log_std Min -1.17874 +eval/num steps total 319502 +eval/num paths total 320 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.082115 +eval/Actions Std 0.914076 +eval/Actions Max 0.999991 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.52291 +time/logging (s) 0.00369662 +time/sampling batch (s) 0.528689 +time/saving (s) 0.003545 +time/training (s) 7.77774 +time/epoch (s) 10.8366 +time/total (s) 3275.49 +Epoch -681 +---------------------------------- --------------- +2022-05-10 14:05:24.762393 PDT | [2] Epoch -680 finished +---------------------------------- --------------- +epoch -680 +replay_buffer/size 999033 +trainer/num train calls 321000 +trainer/Policy Loss -19.1247 +trainer/Log Pis Mean 24.3748 +trainer/Log Pis Std 12.6719 +trainer/Log Pis Max 66.4922 +trainer/Log Pis Min -12.9282 +trainer/policy/mean Mean -0.0157538 +trainer/policy/mean Std 0.908435 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.78791 +trainer/policy/normal/std Std 0.642438 +trainer/policy/normal/std Max 6.31426 +trainer/policy/normal/std Min 0.268844 +trainer/policy/normal/log_std Mean 0.985915 +trainer/policy/normal/log_std Std 0.319146 +trainer/policy/normal/log_std Max 1.84281 +trainer/policy/normal/log_std Min -1.31363 +eval/num steps total 320502 +eval/num paths total 321 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.228412 +eval/Actions Std 0.90856 +eval/Actions Max 0.999981 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73848 +time/logging (s) 0.00415557 +time/sampling batch (s) 0.288147 +time/saving (s) 0.00435542 +time/training (s) 7.22777 +time/epoch (s) 10.2629 +time/total (s) 3285.76 +Epoch -680 +---------------------------------- --------------- +2022-05-10 14:05:34.880021 PDT | [2] Epoch -679 finished +---------------------------------- --------------- +epoch -679 +replay_buffer/size 999033 +trainer/num train calls 322000 +trainer/Policy Loss -19.2464 +trainer/Log Pis Mean 23.523 +trainer/Log Pis Std 13.9352 +trainer/Log Pis Max 72.144 +trainer/Log Pis Min -10.3778 +trainer/policy/mean Mean -0.0503834 +trainer/policy/mean Std 0.905596 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.80893 +trainer/policy/normal/std Std 0.612417 +trainer/policy/normal/std Max 5.46813 +trainer/policy/normal/std Min 0.380154 +trainer/policy/normal/log_std Mean 0.999445 +trainer/policy/normal/log_std Std 0.288323 +trainer/policy/normal/log_std Max 1.69894 +trainer/policy/normal/log_std Min -0.96718 +eval/num steps total 321502 +eval/num paths total 322 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.243422 +eval/Actions Std 0.902053 +eval/Actions Max 0.999998 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.48986 +time/logging (s) 0.00427407 +time/sampling batch (s) 0.529986 +time/saving (s) 0.00433516 +time/training (s) 7.06754 +time/epoch (s) 10.096 +time/total (s) 3295.86 +Epoch -679 +---------------------------------- --------------- +2022-05-10 14:05:44.252796 PDT | [2] Epoch -678 finished +---------------------------------- --------------- +epoch -678 +replay_buffer/size 999033 +trainer/num train calls 323000 +trainer/Policy Loss -19.2446 +trainer/Log Pis Mean 24.3112 +trainer/Log Pis Std 13.0244 +trainer/Log Pis Max 64.78 +trainer/Log Pis Min -4.27977 +trainer/policy/mean Mean -0.0463252 +trainer/policy/mean Std 0.903063 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.79809 +trainer/policy/normal/std Std 0.617139 +trainer/policy/normal/std Max 6.24641 +trainer/policy/normal/std Min 0.405078 +trainer/policy/normal/log_std Mean 0.994218 +trainer/policy/normal/log_std Std 0.296588 +trainer/policy/normal/log_std Max 1.83201 +trainer/policy/normal/log_std Min -0.903675 +eval/num steps total 322502 +eval/num paths total 323 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.225849 +eval/Actions Std 0.804546 +eval/Actions Max 0.999997 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59735 +time/logging (s) 0.00395561 +time/sampling batch (s) 0.282363 +time/saving (s) 0.00393468 +time/training (s) 6.46307 +time/epoch (s) 9.35068 +time/total (s) 3305.21 +Epoch -678 +---------------------------------- --------------- +2022-05-10 14:05:55.388236 PDT | [2] Epoch -677 finished +---------------------------------- --------------- +epoch -677 +replay_buffer/size 999033 +trainer/num train calls 324000 +trainer/Policy Loss -19.3819 +trainer/Log Pis Mean 25.5402 +trainer/Log Pis Std 12.9964 +trainer/Log Pis Max 75.8124 +trainer/Log Pis Min -8.41832 +trainer/policy/mean Mean -0.0322045 +trainer/policy/mean Std 0.904492 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.83487 +trainer/policy/normal/std Std 0.620743 +trainer/policy/normal/std Max 5.66648 +trainer/policy/normal/std Min 0.331847 +trainer/policy/normal/log_std Mean 1.00723 +trainer/policy/normal/log_std Std 0.297532 +trainer/policy/normal/log_std Max 1.73457 +trainer/policy/normal/log_std Min -1.10308 +eval/num steps total 323502 +eval/num paths total 324 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.10116 +eval/Actions Std 0.90665 +eval/Actions Max 0.999993 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67785 +time/logging (s) 0.00439385 +time/sampling batch (s) 1.08509 +time/saving (s) 0.00451204 +time/training (s) 7.34169 +time/epoch (s) 11.1135 +time/total (s) 3316.33 +Epoch -677 +---------------------------------- --------------- +2022-05-10 14:06:06.583548 PDT | [2] Epoch -676 finished +---------------------------------- --------------- +epoch -676 +replay_buffer/size 999033 +trainer/num train calls 325000 +trainer/Policy Loss -18.7481 +trainer/Log Pis Mean 24.125 +trainer/Log Pis Std 13.203 +trainer/Log Pis Max 67.6446 +trainer/Log Pis Min -5.40138 +trainer/policy/mean Mean -0.0440477 +trainer/policy/mean Std 0.904403 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.88327 +trainer/policy/normal/std Std 0.621688 +trainer/policy/normal/std Max 7.28233 +trainer/policy/normal/std Min 0.335485 +trainer/policy/normal/log_std Mean 1.02743 +trainer/policy/normal/log_std Std 0.278155 +trainer/policy/normal/log_std Max 1.98545 +trainer/policy/normal/log_std Min -1.09218 +eval/num steps total 324502 +eval/num paths total 325 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.116262 +eval/Actions Std 0.904951 +eval/Actions Max 0.999999 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.87167 +time/logging (s) 0.00381974 +time/sampling batch (s) 0.584461 +time/saving (s) 0.00387662 +time/training (s) 7.70822 +time/epoch (s) 11.172 +time/total (s) 3327.5 +Epoch -676 +---------------------------------- --------------- +2022-05-10 14:06:17.221032 PDT | [2] Epoch -675 finished +---------------------------------- --------------- +epoch -675 +replay_buffer/size 999033 +trainer/num train calls 326000 +trainer/Policy Loss -19.422 +trainer/Log Pis Mean 25.7035 +trainer/Log Pis Std 13.6778 +trainer/Log Pis Max 69.4183 +trainer/Log Pis Min -12.0241 +trainer/policy/mean Mean -0.0358544 +trainer/policy/mean Std 0.906311 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.80324 +trainer/policy/normal/std Std 0.617669 +trainer/policy/normal/std Max 5.64792 +trainer/policy/normal/std Min 0.341627 +trainer/policy/normal/log_std Mean 0.996259 +trainer/policy/normal/log_std Std 0.294365 +trainer/policy/normal/log_std Max 1.73129 +trainer/policy/normal/log_std Min -1.07404 +eval/num steps total 325502 +eval/num paths total 326 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0540636 +eval/Actions Std 0.936935 +eval/Actions Max 0.999988 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.90108 +time/logging (s) 0.00413595 +time/sampling batch (s) 0.325021 +time/saving (s) 0.00421385 +time/training (s) 7.38112 +time/epoch (s) 10.6156 +time/total (s) 3338.12 +Epoch -675 +---------------------------------- --------------- +2022-05-10 14:06:26.375877 PDT | [2] Epoch -674 finished +---------------------------------- --------------- +epoch -674 +replay_buffer/size 999033 +trainer/num train calls 327000 +trainer/Policy Loss -19.9168 +trainer/Log Pis Mean 24.3667 +trainer/Log Pis Std 13.7136 +trainer/Log Pis Max 63.6877 +trainer/Log Pis Min -10.0974 +trainer/policy/mean Mean -0.0331372 +trainer/policy/mean Std 0.908935 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.83481 +trainer/policy/normal/std Std 0.622815 +trainer/policy/normal/std Max 5.52066 +trainer/policy/normal/std Min 0.309331 +trainer/policy/normal/log_std Mean 1.00709 +trainer/policy/normal/log_std Std 0.298007 +trainer/policy/normal/log_std Max 1.7085 +trainer/policy/normal/log_std Min -1.17334 +eval/num steps total 326502 +eval/num paths total 327 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0375105 +eval/Actions Std 0.916844 +eval/Actions Max 1 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.39317 +time/logging (s) 0.00377128 +time/sampling batch (s) 0.281121 +time/saving (s) 0.00368853 +time/training (s) 6.45109 +time/epoch (s) 9.13285 +time/total (s) 3347.26 +Epoch -674 +---------------------------------- --------------- +2022-05-10 14:06:37.116694 PDT | [2] Epoch -673 finished +---------------------------------- --------------- +epoch -673 +replay_buffer/size 999033 +trainer/num train calls 328000 +trainer/Policy Loss -18.8901 +trainer/Log Pis Mean 23.8135 +trainer/Log Pis Std 13.6908 +trainer/Log Pis Max 67.2311 +trainer/Log Pis Min -6.51942 +trainer/policy/mean Mean -0.0474657 +trainer/policy/mean Std 0.904782 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999984 +trainer/policy/normal/std Mean 2.71841 +trainer/policy/normal/std Std 0.607691 +trainer/policy/normal/std Max 5.19456 +trainer/policy/normal/std Min 0.298642 +trainer/policy/normal/log_std Mean 0.964029 +trainer/policy/normal/log_std Std 0.302517 +trainer/policy/normal/log_std Max 1.64761 +trainer/policy/normal/log_std Min -1.20851 +eval/num steps total 327502 +eval/num paths total 328 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.253649 +eval/Actions Std 0.897136 +eval/Actions Max 0.999991 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64318 +time/logging (s) 0.00370324 +time/sampling batch (s) 0.532553 +time/saving (s) 0.00349864 +time/training (s) 7.53641 +time/epoch (s) 10.7193 +time/total (s) 3357.98 +Epoch -673 +---------------------------------- --------------- +2022-05-10 14:06:46.733058 PDT | [2] Epoch -672 finished +---------------------------------- --------------- +epoch -672 +replay_buffer/size 999033 +trainer/num train calls 329000 +trainer/Policy Loss -18.9281 +trainer/Log Pis Mean 24.2424 +trainer/Log Pis Std 13.4102 +trainer/Log Pis Max 57.6215 +trainer/Log Pis Min -10.3269 +trainer/policy/mean Mean -0.0278153 +trainer/policy/mean Std 0.903363 +trainer/policy/mean Max 0.999978 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.81703 +trainer/policy/normal/std Std 0.639774 +trainer/policy/normal/std Max 7.13259 +trainer/policy/normal/std Min 0.349923 +trainer/policy/normal/log_std Mean 0.999511 +trainer/policy/normal/log_std Std 0.300652 +trainer/policy/normal/log_std Max 1.96467 +trainer/policy/normal/log_std Min -1.05004 +eval/num steps total 328502 +eval/num paths total 329 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.108175 +eval/Actions Std 0.888358 +eval/Actions Max 0.999995 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67158 +time/logging (s) 0.00380108 +time/sampling batch (s) 0.535401 +time/saving (s) 0.00350558 +time/training (s) 6.38059 +time/epoch (s) 9.59488 +time/total (s) 3367.58 +Epoch -672 +---------------------------------- --------------- +2022-05-10 14:06:57.240180 PDT | [2] Epoch -671 finished +---------------------------------- --------------- +epoch -671 +replay_buffer/size 999033 +trainer/num train calls 330000 +trainer/Policy Loss -20.7273 +trainer/Log Pis Mean 25.2724 +trainer/Log Pis Std 13.1999 +trainer/Log Pis Max 76.4149 +trainer/Log Pis Min -5.56363 +trainer/policy/mean Mean -0.0391875 +trainer/policy/mean Std 0.911512 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.81977 +trainer/policy/normal/std Std 0.620515 +trainer/policy/normal/std Max 5.19873 +trainer/policy/normal/std Min 0.297968 +trainer/policy/normal/log_std Mean 1.00243 +trainer/policy/normal/log_std Std 0.292689 +trainer/policy/normal/log_std Max 1.64841 +trainer/policy/normal/log_std Min -1.21077 +eval/num steps total 329502 +eval/num paths total 330 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.118495 +eval/Actions Std 0.915637 +eval/Actions Max 0.999998 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59008 +time/logging (s) 0.00375524 +time/sampling batch (s) 0.551613 +time/saving (s) 0.00347081 +time/training (s) 7.3364 +time/epoch (s) 10.4853 +time/total (s) 3378.07 +Epoch -671 +---------------------------------- --------------- +2022-05-10 14:07:05.985284 PDT | [2] Epoch -670 finished +---------------------------------- --------------- +epoch -670 +replay_buffer/size 999033 +trainer/num train calls 331000 +trainer/Policy Loss -19.7184 +trainer/Log Pis Mean 25.4113 +trainer/Log Pis Std 13.4601 +trainer/Log Pis Max 68.0293 +trainer/Log Pis Min -5.23752 +trainer/policy/mean Mean -0.0410039 +trainer/policy/mean Std 0.904727 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.93523 +trainer/policy/normal/std Std 0.622623 +trainer/policy/normal/std Max 5.11465 +trainer/policy/normal/std Min 0.357306 +trainer/policy/normal/log_std Mean 1.04492 +trainer/policy/normal/log_std Std 0.283192 +trainer/policy/normal/log_std Max 1.63211 +trainer/policy/normal/log_std Min -1.02916 +eval/num steps total 330502 +eval/num paths total 331 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.226482 +eval/Actions Std 0.889726 +eval/Actions Max 0.999993 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.48 +time/logging (s) 0.0037023 +time/sampling batch (s) 0.280108 +time/saving (s) 0.00355095 +time/training (s) 5.95633 +time/epoch (s) 8.72369 +time/total (s) 3386.79 +Epoch -670 +---------------------------------- --------------- +2022-05-10 14:07:16.333322 PDT | [2] Epoch -669 finished +---------------------------------- --------------- +epoch -669 +replay_buffer/size 999033 +trainer/num train calls 332000 +trainer/Policy Loss -18.7845 +trainer/Log Pis Mean 24.7011 +trainer/Log Pis Std 13.103 +trainer/Log Pis Max 85.9536 +trainer/Log Pis Min -6.34115 +trainer/policy/mean Mean -0.0444304 +trainer/policy/mean Std 0.905857 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.83589 +trainer/policy/normal/std Std 0.636159 +trainer/policy/normal/std Max 5.55382 +trainer/policy/normal/std Min 0.37446 +trainer/policy/normal/log_std Mean 1.00654 +trainer/policy/normal/log_std Std 0.300502 +trainer/policy/normal/log_std Max 1.71449 +trainer/policy/normal/log_std Min -0.982271 +eval/num steps total 331502 +eval/num paths total 332 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0375558 +eval/Actions Std 0.862765 +eval/Actions Max 0.999997 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60779 +time/logging (s) 0.00399806 +time/sampling batch (s) 0.529956 +time/saving (s) 0.00399815 +time/training (s) 7.18123 +time/epoch (s) 10.327 +time/total (s) 3397.12 +Epoch -669 +---------------------------------- --------------- +2022-05-10 14:07:26.905320 PDT | [2] Epoch -668 finished +---------------------------------- --------------- +epoch -668 +replay_buffer/size 999033 +trainer/num train calls 333000 +trainer/Policy Loss -18.8121 +trainer/Log Pis Mean 24.209 +trainer/Log Pis Std 13.426 +trainer/Log Pis Max 69.6119 +trainer/Log Pis Min -5.99159 +trainer/policy/mean Mean -0.0228813 +trainer/policy/mean Std 0.905434 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.8389 +trainer/policy/normal/std Std 0.631187 +trainer/policy/normal/std Max 5.55285 +trainer/policy/normal/std Min 0.264007 +trainer/policy/normal/log_std Mean 1.00839 +trainer/policy/normal/log_std Std 0.297328 +trainer/policy/normal/log_std Max 1.71431 +trainer/policy/normal/log_std Min -1.33178 +eval/num steps total 332470 +eval/num paths total 333 +eval/path length Mean 968 +eval/path length Std 0 +eval/path length Max 968 +eval/path length Min 968 +eval/Rewards Mean 0.00103306 +eval/Rewards Std 0.0321246 +eval/Rewards Max 1 +eval/Rewards Min 0 +eval/Returns Mean 1 +eval/Returns Std 0 +eval/Returns Max 1 +eval/Returns Min 1 +eval/Actions Mean 0.0137925 +eval/Actions Std 0.897507 +eval/Actions Max 0.999996 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 1 +eval/normalized_score 100 +time/evaluation sampling (s) 2.59159 +time/logging (s) 0.00367992 +time/sampling batch (s) 0.529963 +time/saving (s) 0.00367615 +time/training (s) 7.42127 +time/epoch (s) 10.5502 +time/total (s) 3407.68 +Epoch -668 +---------------------------------- --------------- +2022-05-10 14:07:37.120670 PDT | [2] Epoch -667 finished +---------------------------------- --------------- +epoch -667 +replay_buffer/size 999033 +trainer/num train calls 334000 +trainer/Policy Loss -19.4014 +trainer/Log Pis Mean 25.3631 +trainer/Log Pis Std 13.4604 +trainer/Log Pis Max 69.6528 +trainer/Log Pis Min -11.4728 +trainer/policy/mean Mean -0.00581651 +trainer/policy/mean Std 0.906097 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.79142 +trainer/policy/normal/std Std 0.655225 +trainer/policy/normal/std Max 5.65517 +trainer/policy/normal/std Min 0.361093 +trainer/policy/normal/log_std Mean 0.986061 +trainer/policy/normal/log_std Std 0.322535 +trainer/policy/normal/log_std Max 1.73257 +trainer/policy/normal/log_std Min -1.01862 +eval/num steps total 333470 +eval/num paths total 334 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.104698 +eval/Actions Std 0.912711 +eval/Actions Max 0.999974 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.31742 +time/logging (s) 0.00371494 +time/sampling batch (s) 0.280237 +time/saving (s) 0.00342496 +time/training (s) 7.58935 +time/epoch (s) 10.1941 +time/total (s) 3417.88 +Epoch -667 +---------------------------------- --------------- +2022-05-10 14:07:47.231862 PDT | [2] Epoch -666 finished +---------------------------------- --------------- +epoch -666 +replay_buffer/size 999033 +trainer/num train calls 335000 +trainer/Policy Loss -19.5995 +trainer/Log Pis Mean 24.2104 +trainer/Log Pis Std 13.3092 +trainer/Log Pis Max 75.1631 +trainer/Log Pis Min -8.30241 +trainer/policy/mean Mean -0.025361 +trainer/policy/mean Std 0.904945 +trainer/policy/mean Max 0.999978 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.83171 +trainer/policy/normal/std Std 0.619485 +trainer/policy/normal/std Max 5.60207 +trainer/policy/normal/std Min 0.361007 +trainer/policy/normal/log_std Mean 1.00614 +trainer/policy/normal/log_std Std 0.297848 +trainer/policy/normal/log_std Max 1.72314 +trainer/policy/normal/log_std Min -1.01886 +eval/num steps total 334470 +eval/num paths total 335 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0294117 +eval/Actions Std 0.900449 +eval/Actions Max 0.999997 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6255 +time/logging (s) 0.00376089 +time/sampling batch (s) 0.530842 +time/saving (s) 0.00344009 +time/training (s) 6.92645 +time/epoch (s) 10.09 +time/total (s) 3427.97 +Epoch -666 +---------------------------------- --------------- +2022-05-10 14:07:57.483772 PDT | [2] Epoch -665 finished +---------------------------------- --------------- +epoch -665 +replay_buffer/size 999033 +trainer/num train calls 336000 +trainer/Policy Loss -18.9333 +trainer/Log Pis Mean 24.8439 +trainer/Log Pis Std 13.2469 +trainer/Log Pis Max 70.6413 +trainer/Log Pis Min -10.4801 +trainer/policy/mean Mean -0.0442037 +trainer/policy/mean Std 0.905577 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999984 +trainer/policy/normal/std Mean 2.79497 +trainer/policy/normal/std Std 0.612474 +trainer/policy/normal/std Max 4.85033 +trainer/policy/normal/std Min 0.332759 +trainer/policy/normal/log_std Mean 0.993164 +trainer/policy/normal/log_std Std 0.296418 +trainer/policy/normal/log_std Max 1.57905 +trainer/policy/normal/log_std Min -1.10034 +eval/num steps total 335470 +eval/num paths total 336 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.346449 +eval/Actions Std 0.78683 +eval/Actions Max 0.999994 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61934 +time/logging (s) 0.00367433 +time/sampling batch (s) 0.282376 +time/saving (s) 0.00344131 +time/training (s) 7.32158 +time/epoch (s) 10.2304 +time/total (s) 3438.2 +Epoch -665 +---------------------------------- --------------- +2022-05-10 14:08:07.428602 PDT | [2] Epoch -664 finished +---------------------------------- --------------- +epoch -664 +replay_buffer/size 999033 +trainer/num train calls 337000 +trainer/Policy Loss -19.8207 +trainer/Log Pis Mean 25.5968 +trainer/Log Pis Std 13.4461 +trainer/Log Pis Max 70.1728 +trainer/Log Pis Min -11.582 +trainer/policy/mean Mean -0.0375999 +trainer/policy/mean Std 0.90663 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.79833 +trainer/policy/normal/std Std 0.621671 +trainer/policy/normal/std Max 7.21092 +trainer/policy/normal/std Min 0.327692 +trainer/policy/normal/log_std Mean 0.993521 +trainer/policy/normal/log_std Std 0.300198 +trainer/policy/normal/log_std Max 1.9756 +trainer/policy/normal/log_std Min -1.11568 +eval/num steps total 336470 +eval/num paths total 337 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00688385 +eval/Actions Std 0.917753 +eval/Actions Max 0.999985 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.44953 +time/logging (s) 0.00408933 +time/sampling batch (s) 0.531235 +time/saving (s) 0.00411374 +time/training (s) 6.93487 +time/epoch (s) 9.92384 +time/total (s) 3448.13 +Epoch -664 +---------------------------------- --------------- +2022-05-10 14:08:18.242471 PDT | [2] Epoch -663 finished +---------------------------------- --------------- +epoch -663 +replay_buffer/size 999033 +trainer/num train calls 338000 +trainer/Policy Loss -20.2594 +trainer/Log Pis Mean 24.1358 +trainer/Log Pis Std 13.1893 +trainer/Log Pis Max 71.6414 +trainer/Log Pis Min -8.39807 +trainer/policy/mean Mean -0.0425393 +trainer/policy/mean Std 0.908284 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.85431 +trainer/policy/normal/std Std 0.62381 +trainer/policy/normal/std Max 6.19218 +trainer/policy/normal/std Min 0.325625 +trainer/policy/normal/log_std Mean 1.01528 +trainer/policy/normal/log_std Std 0.290545 +trainer/policy/normal/log_std Max 1.82329 +trainer/policy/normal/log_std Min -1.12201 +eval/num steps total 337470 +eval/num paths total 338 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.312873 +eval/Actions Std 0.879641 +eval/Actions Max 0.999988 +eval/Actions Min -0.999985 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.76456 +time/logging (s) 0.00376198 +time/sampling batch (s) 0.279888 +time/saving (s) 0.00363937 +time/training (s) 7.73991 +time/epoch (s) 10.7918 +time/total (s) 3458.92 +Epoch -663 +---------------------------------- --------------- +2022-05-10 14:08:28.541847 PDT | [2] Epoch -662 finished +---------------------------------- --------------- +epoch -662 +replay_buffer/size 999033 +trainer/num train calls 339000 +trainer/Policy Loss -19.3894 +trainer/Log Pis Mean 24.9339 +trainer/Log Pis Std 13.8056 +trainer/Log Pis Max 63.3681 +trainer/Log Pis Min -9.10345 +trainer/policy/mean Mean -0.0192519 +trainer/policy/mean Std 0.905095 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.8693 +trainer/policy/normal/std Std 0.641468 +trainer/policy/normal/std Max 5.99238 +trainer/policy/normal/std Min 0.315928 +trainer/policy/normal/log_std Mean 1.01821 +trainer/policy/normal/log_std Std 0.300906 +trainer/policy/normal/log_std Max 1.79049 +trainer/policy/normal/log_std Min -1.15224 +eval/num steps total 338470 +eval/num paths total 339 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0710669 +eval/Actions Std 0.918591 +eval/Actions Max 0.999988 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.7013 +time/logging (s) 0.00375623 +time/sampling batch (s) 0.529189 +time/saving (s) 0.00344214 +time/training (s) 7.04036 +time/epoch (s) 10.278 +time/total (s) 3469.2 +Epoch -662 +---------------------------------- --------------- +2022-05-10 14:08:39.143013 PDT | [2] Epoch -661 finished +---------------------------------- --------------- +epoch -661 +replay_buffer/size 999033 +trainer/num train calls 340000 +trainer/Policy Loss -19.4239 +trainer/Log Pis Mean 25.4554 +trainer/Log Pis Std 14.0201 +trainer/Log Pis Max 72.6222 +trainer/Log Pis Min -6.75071 +trainer/policy/mean Mean -0.0253177 +trainer/policy/mean Std 0.912597 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.82083 +trainer/policy/normal/std Std 0.59354 +trainer/policy/normal/std Max 5.39106 +trainer/policy/normal/std Min 0.361228 +trainer/policy/normal/log_std Mean 1.00685 +trainer/policy/normal/log_std Std 0.272303 +trainer/policy/normal/log_std Max 1.68474 +trainer/policy/normal/log_std Min -1.01825 +eval/num steps total 339470 +eval/num paths total 340 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0212679 +eval/Actions Std 0.908255 +eval/Actions Max 0.999996 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62529 +time/logging (s) 0.00378339 +time/sampling batch (s) 0.779772 +time/saving (s) 0.00346805 +time/training (s) 7.16774 +time/epoch (s) 10.5801 +time/total (s) 3479.79 +Epoch -661 +---------------------------------- --------------- +2022-05-10 14:08:48.532544 PDT | [2] Epoch -660 finished +---------------------------------- --------------- +epoch -660 +replay_buffer/size 999033 +trainer/num train calls 341000 +trainer/Policy Loss -18.7089 +trainer/Log Pis Mean 23.9244 +trainer/Log Pis Std 13.2795 +trainer/Log Pis Max 62.9168 +trainer/Log Pis Min -4.58524 +trainer/policy/mean Mean -0.0405687 +trainer/policy/mean Std 0.90369 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.77551 +trainer/policy/normal/std Std 0.627625 +trainer/policy/normal/std Max 5.49089 +trainer/policy/normal/std Min 0.255884 +trainer/policy/normal/log_std Mean 0.98427 +trainer/policy/normal/log_std Std 0.30434 +trainer/policy/normal/log_std Max 1.70309 +trainer/policy/normal/log_std Min -1.36303 +eval/num steps total 340470 +eval/num paths total 341 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.149154 +eval/Actions Std 0.843027 +eval/Actions Max 0.999993 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54978 +time/logging (s) 0.00372375 +time/sampling batch (s) 0.281178 +time/saving (s) 0.00358706 +time/training (s) 6.52988 +time/epoch (s) 9.36815 +time/total (s) 3489.16 +Epoch -660 +---------------------------------- --------------- +2022-05-10 14:08:58.919393 PDT | [2] Epoch -659 finished +---------------------------------- --------------- +epoch -659 +replay_buffer/size 999033 +trainer/num train calls 342000 +trainer/Policy Loss -20.5861 +trainer/Log Pis Mean 23.4674 +trainer/Log Pis Std 13.6234 +trainer/Log Pis Max 66.7321 +trainer/Log Pis Min -10.1046 +trainer/policy/mean Mean -0.0300396 +trainer/policy/mean Std 0.907144 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.83844 +trainer/policy/normal/std Std 0.608084 +trainer/policy/normal/std Max 5.25882 +trainer/policy/normal/std Min 0.35024 +trainer/policy/normal/log_std Mean 1.01164 +trainer/policy/normal/log_std Std 0.279058 +trainer/policy/normal/log_std Max 1.65991 +trainer/policy/normal/log_std Min -1.04914 +eval/num steps total 341470 +eval/num paths total 342 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.122357 +eval/Actions Std 0.89282 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61186 +time/logging (s) 0.00372398 +time/sampling batch (s) 0.281364 +time/saving (s) 0.00342491 +time/training (s) 7.46511 +time/epoch (s) 10.3655 +time/total (s) 3499.53 +Epoch -659 +---------------------------------- --------------- +2022-05-10 14:09:08.586507 PDT | [2] Epoch -658 finished +---------------------------------- --------------- +epoch -658 +replay_buffer/size 999033 +trainer/num train calls 343000 +trainer/Policy Loss -18.0924 +trainer/Log Pis Mean 23.9551 +trainer/Log Pis Std 13.43 +trainer/Log Pis Max 71.4848 +trainer/Log Pis Min -12.5095 +trainer/policy/mean Mean -0.0362537 +trainer/policy/mean Std 0.904294 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999982 +trainer/policy/normal/std Mean 2.77214 +trainer/policy/normal/std Std 0.637638 +trainer/policy/normal/std Max 5.11191 +trainer/policy/normal/std Min 0.354097 +trainer/policy/normal/log_std Mean 0.980037 +trainer/policy/normal/log_std Std 0.320051 +trainer/policy/normal/log_std Max 1.63157 +trainer/policy/normal/log_std Min -1.03818 +eval/num steps total 342470 +eval/num paths total 343 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.129984 +eval/Actions Std 0.909459 +eval/Actions Max 1 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.49933 +time/logging (s) 0.00411503 +time/sampling batch (s) 0.280487 +time/saving (s) 0.00398981 +time/training (s) 6.85829 +time/epoch (s) 9.64621 +time/total (s) 3509.18 +Epoch -658 +---------------------------------- --------------- +2022-05-10 14:09:19.243535 PDT | [2] Epoch -657 finished +---------------------------------- --------------- +epoch -657 +replay_buffer/size 999033 +trainer/num train calls 344000 +trainer/Policy Loss -19.853 +trainer/Log Pis Mean 24.7139 +trainer/Log Pis Std 14.5473 +trainer/Log Pis Max 74.1509 +trainer/Log Pis Min -4.46861 +trainer/policy/mean Mean -0.0253937 +trainer/policy/mean Std 0.908232 +trainer/policy/mean Max 1 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.88101 +trainer/policy/normal/std Std 0.636586 +trainer/policy/normal/std Max 6.41056 +trainer/policy/normal/std Min 0.400151 +trainer/policy/normal/log_std Mean 1.0237 +trainer/policy/normal/log_std Std 0.294288 +trainer/policy/normal/log_std Max 1.85795 +trainer/policy/normal/log_std Min -0.915913 +eval/num steps total 343470 +eval/num paths total 344 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.108237 +eval/Actions Std 0.894564 +eval/Actions Max 0.999998 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59711 +time/logging (s) 0.00385749 +time/sampling batch (s) 0.531242 +time/saving (s) 0.00361987 +time/training (s) 7.49938 +time/epoch (s) 10.6352 +time/total (s) 3519.81 +Epoch -657 +---------------------------------- --------------- +2022-05-10 14:09:29.225568 PDT | [2] Epoch -656 finished +---------------------------------- --------------- +epoch -656 +replay_buffer/size 999033 +trainer/num train calls 345000 +trainer/Policy Loss -19.5873 +trainer/Log Pis Mean 25.0528 +trainer/Log Pis Std 13.7356 +trainer/Log Pis Max 79.1803 +trainer/Log Pis Min -7.58778 +trainer/policy/mean Mean -0.0284408 +trainer/policy/mean Std 0.908118 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.79409 +trainer/policy/normal/std Std 0.632184 +trainer/policy/normal/std Max 5.37522 +trainer/policy/normal/std Min 0.358452 +trainer/policy/normal/log_std Mean 0.991551 +trainer/policy/normal/log_std Std 0.299289 +trainer/policy/normal/log_std Max 1.6818 +trainer/policy/normal/log_std Min -1.02596 +eval/num steps total 344470 +eval/num paths total 345 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0241029 +eval/Actions Std 0.899889 +eval/Actions Max 0.999998 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64789 +time/logging (s) 0.00382275 +time/sampling batch (s) 0.280919 +time/saving (s) 0.00344984 +time/training (s) 7.02437 +time/epoch (s) 9.96046 +time/total (s) 3529.78 +Epoch -656 +---------------------------------- --------------- +2022-05-10 14:09:40.077911 PDT | [2] Epoch -655 finished +---------------------------------- --------------- +epoch -655 +replay_buffer/size 999033 +trainer/num train calls 346000 +trainer/Policy Loss -20.3782 +trainer/Log Pis Mean 24.1805 +trainer/Log Pis Std 13.4996 +trainer/Log Pis Max 74.5106 +trainer/Log Pis Min -4.3754 +trainer/policy/mean Mean -0.0247891 +trainer/policy/mean Std 0.907553 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.7801 +trainer/policy/normal/std Std 0.61768 +trainer/policy/normal/std Max 5.98585 +trainer/policy/normal/std Min 0.280562 +trainer/policy/normal/log_std Mean 0.987031 +trainer/policy/normal/log_std Std 0.300756 +trainer/policy/normal/log_std Max 1.7894 +trainer/policy/normal/log_std Min -1.27096 +eval/num steps total 345470 +eval/num paths total 346 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.2569 +eval/Actions Std 0.880362 +eval/Actions Max 0.999995 +eval/Actions Min -0.999988 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.81008 +time/logging (s) 0.00380157 +time/sampling batch (s) 0.531205 +time/saving (s) 0.00343176 +time/training (s) 7.48256 +time/epoch (s) 10.8311 +time/total (s) 3540.61 +Epoch -655 +---------------------------------- --------------- +2022-05-10 14:09:50.639220 PDT | [2] Epoch -654 finished +---------------------------------- --------------- +epoch -654 +replay_buffer/size 999033 +trainer/num train calls 347000 +trainer/Policy Loss -18.4726 +trainer/Log Pis Mean 24.1672 +trainer/Log Pis Std 13.4236 +trainer/Log Pis Max 69.5845 +trainer/Log Pis Min -12.0081 +trainer/policy/mean Mean -0.0267767 +trainer/policy/mean Std 0.906219 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.82013 +trainer/policy/normal/std Std 0.612138 +trainer/policy/normal/std Max 5.2869 +trainer/policy/normal/std Min 0.363099 +trainer/policy/normal/log_std Mean 1.0036 +trainer/policy/normal/log_std Std 0.288074 +trainer/policy/normal/log_std Max 1.66523 +trainer/policy/normal/log_std Min -1.01308 +eval/num steps total 346470 +eval/num paths total 347 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.101087 +eval/Actions Std 0.917043 +eval/Actions Max 0.999999 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50232 +time/logging (s) 0.0037196 +time/sampling batch (s) 0.28768 +time/saving (s) 0.00361609 +time/training (s) 7.74212 +time/epoch (s) 10.5395 +time/total (s) 3551.15 +Epoch -654 +---------------------------------- --------------- +2022-05-10 14:10:00.589250 PDT | [2] Epoch -653 finished +---------------------------------- --------------- +epoch -653 +replay_buffer/size 999033 +trainer/num train calls 348000 +trainer/Policy Loss -19.9453 +trainer/Log Pis Mean 26.4692 +trainer/Log Pis Std 13.787 +trainer/Log Pis Max 69.5987 +trainer/Log Pis Min -8.24947 +trainer/policy/mean Mean -0.0245078 +trainer/policy/mean Std 0.908276 +trainer/policy/mean Max 1 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.85432 +trainer/policy/normal/std Std 0.613591 +trainer/policy/normal/std Max 5.35172 +trainer/policy/normal/std Min 0.350563 +trainer/policy/normal/log_std Mean 1.01676 +trainer/policy/normal/log_std Std 0.283081 +trainer/policy/normal/log_std Max 1.67742 +trainer/policy/normal/log_std Min -1.04821 +eval/num steps total 347470 +eval/num paths total 348 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0976485 +eval/Actions Std 0.910353 +eval/Actions Max 0.999996 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.44925 +time/logging (s) 0.00373961 +time/sampling batch (s) 0.282919 +time/saving (s) 0.00348439 +time/training (s) 7.18903 +time/epoch (s) 9.92843 +time/total (s) 3561.09 +Epoch -653 +---------------------------------- --------------- +2022-05-10 14:10:09.897868 PDT | [2] Epoch -652 finished +---------------------------------- --------------- +epoch -652 +replay_buffer/size 999033 +trainer/num train calls 349000 +trainer/Policy Loss -19.5135 +trainer/Log Pis Mean 24.1823 +trainer/Log Pis Std 13.1503 +trainer/Log Pis Max 81.8925 +trainer/Log Pis Min -10.4862 +trainer/policy/mean Mean -0.0246559 +trainer/policy/mean Std 0.906959 +trainer/policy/mean Max 0.999982 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.82609 +trainer/policy/normal/std Std 0.615812 +trainer/policy/normal/std Max 6.29196 +trainer/policy/normal/std Min 0.324489 +trainer/policy/normal/log_std Mean 1.00553 +trainer/policy/normal/log_std Std 0.289725 +trainer/policy/normal/log_std Max 1.83927 +trainer/policy/normal/log_std Min -1.12551 +eval/num steps total 348470 +eval/num paths total 349 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.103147 +eval/Actions Std 0.915427 +eval/Actions Max 0.999996 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56346 +time/logging (s) 0.00367325 +time/sampling batch (s) 0.534915 +time/saving (s) 0.00348741 +time/training (s) 6.18133 +time/epoch (s) 9.28687 +time/total (s) 3570.38 +Epoch -652 +---------------------------------- --------------- +2022-05-10 14:10:19.168242 PDT | [2] Epoch -651 finished +---------------------------------- --------------- +epoch -651 +replay_buffer/size 999033 +trainer/num train calls 350000 +trainer/Policy Loss -19.0574 +trainer/Log Pis Mean 25.7458 +trainer/Log Pis Std 12.9294 +trainer/Log Pis Max 64.0221 +trainer/Log Pis Min -7.99308 +trainer/policy/mean Mean -0.0222801 +trainer/policy/mean Std 0.906921 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.90188 +trainer/policy/normal/std Std 0.630039 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.36938 +trainer/policy/normal/log_std Mean 1.033 +trainer/policy/normal/log_std Std 0.283862 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -0.995928 +eval/num steps total 349470 +eval/num paths total 350 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.116437 +eval/Actions Std 0.881516 +eval/Actions Max 0.999993 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67524 +time/logging (s) 0.00406988 +time/sampling batch (s) 0.28347 +time/saving (s) 0.00405686 +time/training (s) 6.28225 +time/epoch (s) 9.24908 +time/total (s) 3579.63 +Epoch -651 +---------------------------------- --------------- +2022-05-10 14:10:29.980361 PDT | [2] Epoch -650 finished +---------------------------------- --------------- +epoch -650 +replay_buffer/size 999033 +trainer/num train calls 351000 +trainer/Policy Loss -19.7566 +trainer/Log Pis Mean 23.9573 +trainer/Log Pis Std 13.2545 +trainer/Log Pis Max 62.7112 +trainer/Log Pis Min -7.96487 +trainer/policy/mean Mean -0.0344585 +trainer/policy/mean Std 0.909483 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.86926 +trainer/policy/normal/std Std 0.616161 +trainer/policy/normal/std Max 6.43047 +trainer/policy/normal/std Min 0.382935 +trainer/policy/normal/log_std Mean 1.02292 +trainer/policy/normal/log_std Std 0.275859 +trainer/policy/normal/log_std Max 1.86105 +trainer/policy/normal/log_std Min -0.959889 +eval/num steps total 350470 +eval/num paths total 351 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0318643 +eval/Actions Std 0.911535 +eval/Actions Max 0.999997 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.52326 +time/logging (s) 0.00385398 +time/sampling batch (s) 1.03469 +time/saving (s) 0.00375473 +time/training (s) 7.22435 +time/epoch (s) 10.7899 +time/total (s) 3590.42 +Epoch -650 +---------------------------------- --------------- +2022-05-10 14:10:40.346164 PDT | [2] Epoch -649 finished +---------------------------------- --------------- +epoch -649 +replay_buffer/size 999033 +trainer/num train calls 352000 +trainer/Policy Loss -20.4068 +trainer/Log Pis Mean 25.7038 +trainer/Log Pis Std 13.8408 +trainer/Log Pis Max 64.9583 +trainer/Log Pis Min -15.0634 +trainer/policy/mean Mean -0.0500925 +trainer/policy/mean Std 0.909337 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.87747 +trainer/policy/normal/std Std 0.633518 +trainer/policy/normal/std Max 5.39333 +trainer/policy/normal/std Min 0.260973 +trainer/policy/normal/log_std Mean 1.02225 +trainer/policy/normal/log_std Std 0.296443 +trainer/policy/normal/log_std Max 1.68516 +trainer/policy/normal/log_std Min -1.34334 +eval/num steps total 351470 +eval/num paths total 352 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00235174 +eval/Actions Std 0.904164 +eval/Actions Max 0.999995 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62389 +time/logging (s) 0.00367441 +time/sampling batch (s) 0.279132 +time/saving (s) 0.00348033 +time/training (s) 7.43402 +time/epoch (s) 10.3442 +time/total (s) 3600.77 +Epoch -649 +---------------------------------- --------------- +2022-05-10 14:10:50.444575 PDT | [2] Epoch -648 finished +---------------------------------- --------------- +epoch -648 +replay_buffer/size 999033 +trainer/num train calls 353000 +trainer/Policy Loss -19.9365 +trainer/Log Pis Mean 24.0044 +trainer/Log Pis Std 13.3126 +trainer/Log Pis Max 67.8653 +trainer/Log Pis Min -7.47556 +trainer/policy/mean Mean -0.0264619 +trainer/policy/mean Std 0.90263 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.85293 +trainer/policy/normal/std Std 0.641487 +trainer/policy/normal/std Max 5.91176 +trainer/policy/normal/std Min 0.279534 +trainer/policy/normal/log_std Mean 1.01256 +trainer/policy/normal/log_std Std 0.301046 +trainer/policy/normal/log_std Max 1.77694 +trainer/policy/normal/log_std Min -1.27463 +eval/num steps total 352470 +eval/num paths total 353 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.158381 +eval/Actions Std 0.930595 +eval/Actions Max 0.999998 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62681 +time/logging (s) 0.00391568 +time/sampling batch (s) 0.528181 +time/saving (s) 0.00346321 +time/training (s) 6.91521 +time/epoch (s) 10.0776 +time/total (s) 3610.85 +Epoch -648 +---------------------------------- --------------- +2022-05-10 14:11:00.642625 PDT | [2] Epoch -647 finished +---------------------------------- --------------- +epoch -647 +replay_buffer/size 999033 +trainer/num train calls 354000 +trainer/Policy Loss -19.6745 +trainer/Log Pis Mean 22.746 +trainer/Log Pis Std 13.4913 +trainer/Log Pis Max 68.8757 +trainer/Log Pis Min -13.016 +trainer/policy/mean Mean -0.0286458 +trainer/policy/mean Std 0.902384 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.76264 +trainer/policy/normal/std Std 0.62334 +trainer/policy/normal/std Max 4.86005 +trainer/policy/normal/std Min 0.327083 +trainer/policy/normal/log_std Mean 0.979934 +trainer/policy/normal/log_std Std 0.301586 +trainer/policy/normal/log_std Max 1.58105 +trainer/policy/normal/log_std Min -1.11754 +eval/num steps total 353470 +eval/num paths total 354 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0170319 +eval/Actions Std 0.906409 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.75445 +time/logging (s) 0.00371843 +time/sampling batch (s) 0.528316 +time/saving (s) 0.00338987 +time/training (s) 6.88668 +time/epoch (s) 10.1766 +time/total (s) 3621.03 +Epoch -647 +---------------------------------- --------------- +2022-05-10 14:11:10.403319 PDT | [2] Epoch -646 finished +---------------------------------- --------------- +epoch -646 +replay_buffer/size 999033 +trainer/num train calls 355000 +trainer/Policy Loss -19.1846 +trainer/Log Pis Mean 24.1931 +trainer/Log Pis Std 13.0896 +trainer/Log Pis Max 69.8548 +trainer/Log Pis Min -7.24357 +trainer/policy/mean Mean -0.030318 +trainer/policy/mean Std 0.908353 +trainer/policy/mean Max 0.99998 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.88451 +trainer/policy/normal/std Std 0.639759 +trainer/policy/normal/std Max 5.97538 +trainer/policy/normal/std Min 0.323757 +trainer/policy/normal/log_std Mean 1.0251 +trainer/policy/normal/log_std Std 0.293595 +trainer/policy/normal/log_std Max 1.78765 +trainer/policy/normal/log_std Min -1.12776 +eval/num steps total 354470 +eval/num paths total 355 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0584043 +eval/Actions Std 0.914034 +eval/Actions Max 0.999995 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51567 +time/logging (s) 0.00370625 +time/sampling batch (s) 0.278512 +time/saving (s) 0.00345852 +time/training (s) 6.93812 +time/epoch (s) 9.73946 +time/total (s) 3630.77 +Epoch -646 +---------------------------------- --------------- +2022-05-10 14:11:20.391792 PDT | [2] Epoch -645 finished +---------------------------------- --------------- +epoch -645 +replay_buffer/size 999033 +trainer/num train calls 356000 +trainer/Policy Loss -19.0526 +trainer/Log Pis Mean 25.9251 +trainer/Log Pis Std 13.5043 +trainer/Log Pis Max 74.4623 +trainer/Log Pis Min -11.9605 +trainer/policy/mean Mean -0.0257712 +trainer/policy/mean Std 0.906576 +trainer/policy/mean Max 0.999982 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.87168 +trainer/policy/normal/std Std 0.625882 +trainer/policy/normal/std Max 5.29659 +trainer/policy/normal/std Min 0.402087 +trainer/policy/normal/log_std Mean 1.02171 +trainer/policy/normal/log_std Std 0.287825 +trainer/policy/normal/log_std Max 1.66706 +trainer/policy/normal/log_std Min -0.911086 +eval/num steps total 355470 +eval/num paths total 356 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00712557 +eval/Actions Std 0.910904 +eval/Actions Max 1 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6522 +time/logging (s) 0.00424239 +time/sampling batch (s) 0.278137 +time/saving (s) 0.00423731 +time/training (s) 7.02877 +time/epoch (s) 9.96758 +time/total (s) 3640.74 +Epoch -645 +---------------------------------- --------------- +2022-05-10 14:11:31.435782 PDT | [2] Epoch -644 finished +---------------------------------- --------------- +epoch -644 +replay_buffer/size 999033 +trainer/num train calls 357000 +trainer/Policy Loss -20.1847 +trainer/Log Pis Mean 24.241 +trainer/Log Pis Std 12.2303 +trainer/Log Pis Max 62.0231 +trainer/Log Pis Min -5.90326 +trainer/policy/mean Mean -0.0250646 +trainer/policy/mean Std 0.911952 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.77432 +trainer/policy/normal/std Std 0.622194 +trainer/policy/normal/std Max 5.45608 +trainer/policy/normal/std Min 0.315142 +trainer/policy/normal/log_std Mean 0.983642 +trainer/policy/normal/log_std Std 0.307998 +trainer/policy/normal/log_std Max 1.69673 +trainer/policy/normal/log_std Min -1.15473 +eval/num steps total 356470 +eval/num paths total 357 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.214087 +eval/Actions Std 0.930083 +eval/Actions Max 0.999996 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.75581 +time/logging (s) 0.00377764 +time/sampling batch (s) 0.291196 +time/saving (s) 0.00370681 +time/training (s) 7.96675 +time/epoch (s) 11.0212 +time/total (s) 3651.77 +Epoch -644 +---------------------------------- --------------- +2022-05-10 14:11:41.212363 PDT | [2] Epoch -643 finished +---------------------------------- --------------- +epoch -643 +replay_buffer/size 999033 +trainer/num train calls 358000 +trainer/Policy Loss -18.5405 +trainer/Log Pis Mean 24.3931 +trainer/Log Pis Std 13.4588 +trainer/Log Pis Max 67.6478 +trainer/Log Pis Min -3.32516 +trainer/policy/mean Mean -0.0380218 +trainer/policy/mean Std 0.905233 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.8205 +trainer/policy/normal/std Std 0.62063 +trainer/policy/normal/std Max 5.25716 +trainer/policy/normal/std Min 0.35954 +trainer/policy/normal/log_std Mean 1.00276 +trainer/policy/normal/log_std Std 0.292297 +trainer/policy/normal/log_std Max 1.65959 +trainer/policy/normal/log_std Min -1.02293 +eval/num steps total 357470 +eval/num paths total 358 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0294646 +eval/Actions Std 0.910984 +eval/Actions Max 0.999995 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.84683 +time/logging (s) 0.00381001 +time/sampling batch (s) 0.286371 +time/saving (s) 0.00361737 +time/training (s) 6.61445 +time/epoch (s) 9.75508 +time/total (s) 3661.52 +Epoch -643 +---------------------------------- --------------- +2022-05-10 14:11:51.256098 PDT | [2] Epoch -642 finished +---------------------------------- --------------- +epoch -642 +replay_buffer/size 999033 +trainer/num train calls 359000 +trainer/Policy Loss -20.1495 +trainer/Log Pis Mean 24.7679 +trainer/Log Pis Std 13.4541 +trainer/Log Pis Max 67.5875 +trainer/Log Pis Min -5.99436 +trainer/policy/mean Mean -0.0352018 +trainer/policy/mean Std 0.90799 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.85896 +trainer/policy/normal/std Std 0.65095 +trainer/policy/normal/std Max 5.58255 +trainer/policy/normal/std Min 0.309198 +trainer/policy/normal/log_std Mean 1.01285 +trainer/policy/normal/log_std Std 0.310076 +trainer/policy/normal/log_std Max 1.71965 +trainer/policy/normal/log_std Min -1.17377 +eval/num steps total 358470 +eval/num paths total 359 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.212427 +eval/Actions Std 0.905227 +eval/Actions Max 0.999994 +eval/Actions Min -0.999985 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.52948 +time/logging (s) 0.00375546 +time/sampling batch (s) 0.278777 +time/saving (s) 0.00381195 +time/training (s) 7.20666 +time/epoch (s) 10.0225 +time/total (s) 3671.55 +Epoch -642 +---------------------------------- --------------- +2022-05-10 14:12:01.864073 PDT | [2] Epoch -641 finished +---------------------------------- --------------- +epoch -641 +replay_buffer/size 999033 +trainer/num train calls 360000 +trainer/Policy Loss -19.9522 +trainer/Log Pis Mean 24.926 +trainer/Log Pis Std 12.9668 +trainer/Log Pis Max 64.8916 +trainer/Log Pis Min -4.02124 +trainer/policy/mean Mean -0.0426131 +trainer/policy/mean Std 0.904206 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.86245 +trainer/policy/normal/std Std 0.618879 +trainer/policy/normal/std Max 5.60258 +trainer/policy/normal/std Min 0.348732 +trainer/policy/normal/log_std Mean 1.02006 +trainer/policy/normal/log_std Std 0.278634 +trainer/policy/normal/log_std Max 1.72323 +trainer/policy/normal/log_std Min -1.05345 +eval/num steps total 359470 +eval/num paths total 360 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.245865 +eval/Actions Std 0.907574 +eval/Actions Max 0.999987 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.94208 +time/logging (s) 0.00372441 +time/sampling batch (s) 0.286302 +time/saving (s) 0.00349819 +time/training (s) 7.35009 +time/epoch (s) 10.5857 +time/total (s) 3682.14 +Epoch -641 +---------------------------------- --------------- +2022-05-10 14:12:11.642843 PDT | [2] Epoch -640 finished +---------------------------------- --------------- +epoch -640 +replay_buffer/size 999033 +trainer/num train calls 361000 +trainer/Policy Loss -19.2316 +trainer/Log Pis Mean 24.1046 +trainer/Log Pis Std 12.7041 +trainer/Log Pis Max 66.5543 +trainer/Log Pis Min -9.82427 +trainer/policy/mean Mean -0.0217832 +trainer/policy/mean Std 0.902934 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.82312 +trainer/policy/normal/std Std 0.637143 +trainer/policy/normal/std Max 6.20368 +trainer/policy/normal/std Min 0.322374 +trainer/policy/normal/log_std Mean 1.00206 +trainer/policy/normal/log_std Std 0.299161 +trainer/policy/normal/log_std Max 1.82514 +trainer/policy/normal/log_std Min -1.13204 +eval/num steps total 360470 +eval/num paths total 361 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.169209 +eval/Actions Std 0.810778 +eval/Actions Max 0.999995 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66929 +time/logging (s) 0.00370508 +time/sampling batch (s) 0.284431 +time/saving (s) 0.00343508 +time/training (s) 6.79608 +time/epoch (s) 9.75694 +time/total (s) 3691.9 +Epoch -640 +---------------------------------- --------------- +2022-05-10 14:12:22.664744 PDT | [2] Epoch -639 finished +---------------------------------- --------------- +epoch -639 +replay_buffer/size 999033 +trainer/num train calls 362000 +trainer/Policy Loss -18.9118 +trainer/Log Pis Mean 24.5445 +trainer/Log Pis Std 12.3713 +trainer/Log Pis Max 62.9106 +trainer/Log Pis Min -3.80805 +trainer/policy/mean Mean -0.023392 +trainer/policy/mean Std 0.902997 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999976 +trainer/policy/normal/std Mean 2.79243 +trainer/policy/normal/std Std 0.62989 +trainer/policy/normal/std Max 5.36155 +trainer/policy/normal/std Min 0.321595 +trainer/policy/normal/log_std Mean 0.989117 +trainer/policy/normal/log_std Std 0.312326 +trainer/policy/normal/log_std Max 1.67925 +trainer/policy/normal/log_std Min -1.13446 +eval/num steps total 361470 +eval/num paths total 362 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.288546 +eval/Actions Std 0.859534 +eval/Actions Max 0.99999 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73858 +time/logging (s) 0.00421735 +time/sampling batch (s) 0.282873 +time/saving (s) 0.00475187 +time/training (s) 7.97013 +time/epoch (s) 11.0005 +time/total (s) 3702.9 +Epoch -639 +---------------------------------- --------------- +2022-05-10 14:12:33.581262 PDT | [2] Epoch -638 finished +---------------------------------- --------------- +epoch -638 +replay_buffer/size 999033 +trainer/num train calls 363000 +trainer/Policy Loss -19.2132 +trainer/Log Pis Mean 24.1597 +trainer/Log Pis Std 12.7996 +trainer/Log Pis Max 70.1131 +trainer/Log Pis Min -11.7711 +trainer/policy/mean Mean -0.0470617 +trainer/policy/mean Std 0.907689 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.81076 +trainer/policy/normal/std Std 0.623692 +trainer/policy/normal/std Max 5.52483 +trainer/policy/normal/std Min 0.344042 +trainer/policy/normal/log_std Mean 0.998214 +trainer/policy/normal/log_std Std 0.298748 +trainer/policy/normal/log_std Max 1.70925 +trainer/policy/normal/log_std Min -1.06699 +eval/num steps total 362470 +eval/num paths total 363 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.360666 +eval/Actions Std 0.887034 +eval/Actions Max 0.999994 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65228 +time/logging (s) 0.00374311 +time/sampling batch (s) 0.289396 +time/saving (s) 0.00353775 +time/training (s) 7.94425 +time/epoch (s) 10.8932 +time/total (s) 3713.8 +Epoch -638 +---------------------------------- --------------- +2022-05-10 14:12:44.762329 PDT | [2] Epoch -637 finished +---------------------------------- --------------- +epoch -637 +replay_buffer/size 999033 +trainer/num train calls 364000 +trainer/Policy Loss -19.6024 +trainer/Log Pis Mean 24.6983 +trainer/Log Pis Std 13.7703 +trainer/Log Pis Max 79.9531 +trainer/Log Pis Min -3.87209 +trainer/policy/mean Mean -0.0491373 +trainer/policy/mean Std 0.904816 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.82991 +trainer/policy/normal/std Std 0.611857 +trainer/policy/normal/std Max 5.50671 +trainer/policy/normal/std Min 0.38787 +trainer/policy/normal/log_std Mean 1.00812 +trainer/policy/normal/log_std Std 0.281532 +trainer/policy/normal/log_std Max 1.70597 +trainer/policy/normal/log_std Min -0.947086 +eval/num steps total 363470 +eval/num paths total 364 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0351462 +eval/Actions Std 0.897943 +eval/Actions Max 0.999997 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63169 +time/logging (s) 0.00372396 +time/sampling batch (s) 0.283128 +time/saving (s) 0.00357095 +time/training (s) 8.23733 +time/epoch (s) 11.1594 +time/total (s) 3724.96 +Epoch -637 +---------------------------------- --------------- +2022-05-10 14:12:55.335648 PDT | [2] Epoch -636 finished +---------------------------------- --------------- +epoch -636 +replay_buffer/size 999033 +trainer/num train calls 365000 +trainer/Policy Loss -18.059 +trainer/Log Pis Mean 24.3507 +trainer/Log Pis Std 12.9979 +trainer/Log Pis Max 69.7622 +trainer/Log Pis Min -6.12948 +trainer/policy/mean Mean -0.0421426 +trainer/policy/mean Std 0.901043 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.82011 +trainer/policy/normal/std Std 0.613518 +trainer/policy/normal/std Max 5.31342 +trainer/policy/normal/std Min 0.405188 +trainer/policy/normal/log_std Mean 1.00342 +trainer/policy/normal/log_std Std 0.289479 +trainer/policy/normal/log_std Max 1.67024 +trainer/policy/normal/log_std Min -0.903403 +eval/num steps total 364470 +eval/num paths total 365 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.357003 +eval/Actions Std 0.853759 +eval/Actions Max 0.999988 +eval/Actions Min -0.999973 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.79819 +time/logging (s) 0.00384655 +time/sampling batch (s) 0.296472 +time/saving (s) 0.00363863 +time/training (s) 7.44954 +time/epoch (s) 10.5517 +time/total (s) 3735.51 +Epoch -636 +---------------------------------- --------------- +2022-05-10 14:13:05.867490 PDT | [2] Epoch -635 finished +---------------------------------- --------------- +epoch -635 +replay_buffer/size 999033 +trainer/num train calls 366000 +trainer/Policy Loss -20.2545 +trainer/Log Pis Mean 23.8124 +trainer/Log Pis Std 13.1425 +trainer/Log Pis Max 65.3417 +trainer/Log Pis Min -6.41026 +trainer/policy/mean Mean -0.0435721 +trainer/policy/mean Std 0.905527 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.83329 +trainer/policy/normal/std Std 0.617207 +trainer/policy/normal/std Max 5.59761 +trainer/policy/normal/std Min 0.381261 +trainer/policy/normal/log_std Mean 1.00908 +trainer/policy/normal/log_std Std 0.281902 +trainer/policy/normal/log_std Max 1.72234 +trainer/policy/normal/log_std Min -0.964271 +eval/num steps total 365470 +eval/num paths total 366 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.267117 +eval/Actions Std 0.820256 +eval/Actions Max 0.999983 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.89698 +time/logging (s) 0.00369225 +time/sampling batch (s) 0.290894 +time/saving (s) 0.00345194 +time/training (s) 7.31508 +time/epoch (s) 10.5101 +time/total (s) 3746.03 +Epoch -635 +---------------------------------- --------------- +2022-05-10 14:13:16.365285 PDT | [2] Epoch -634 finished +---------------------------------- --------------- +epoch -634 +replay_buffer/size 999033 +trainer/num train calls 367000 +trainer/Policy Loss -18.9169 +trainer/Log Pis Mean 24.1107 +trainer/Log Pis Std 12.6508 +trainer/Log Pis Max 63.9707 +trainer/Log Pis Min -7.43729 +trainer/policy/mean Mean -0.053018 +trainer/policy/mean Std 0.903718 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.82188 +trainer/policy/normal/std Std 0.635595 +trainer/policy/normal/std Max 5.59852 +trainer/policy/normal/std Min 0.31675 +trainer/policy/normal/log_std Mean 1.00238 +trainer/policy/normal/log_std Std 0.294721 +trainer/policy/normal/log_std Max 1.7225 +trainer/policy/normal/log_std Min -1.14964 +eval/num steps total 366470 +eval/num paths total 367 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.184313 +eval/Actions Std 0.746033 +eval/Actions Max 0.99997 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66736 +time/logging (s) 0.00408243 +time/sampling batch (s) 0.532544 +time/saving (s) 0.00408948 +time/training (s) 7.26855 +time/epoch (s) 10.4766 +time/total (s) 3756.51 +Epoch -634 +---------------------------------- --------------- +2022-05-10 14:13:26.366954 PDT | [2] Epoch -633 finished +---------------------------------- --------------- +epoch -633 +replay_buffer/size 999033 +trainer/num train calls 368000 +trainer/Policy Loss -18.9315 +trainer/Log Pis Mean 23.4664 +trainer/Log Pis Std 12.6856 +trainer/Log Pis Max 62.9983 +trainer/Log Pis Min -12.3683 +trainer/policy/mean Mean -0.0243346 +trainer/policy/mean Std 0.9067 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.78135 +trainer/policy/normal/std Std 0.591948 +trainer/policy/normal/std Max 5.15418 +trainer/policy/normal/std Min 0.338278 +trainer/policy/normal/log_std Mean 0.991331 +trainer/policy/normal/log_std Std 0.280913 +trainer/policy/normal/log_std Max 1.63981 +trainer/policy/normal/log_std Min -1.08389 +eval/num steps total 367470 +eval/num paths total 368 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.243327 +eval/Actions Std 0.897536 +eval/Actions Max 0.999999 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.77079 +time/logging (s) 0.00375765 +time/sampling batch (s) 0.284652 +time/saving (s) 0.00360841 +time/training (s) 6.91673 +time/epoch (s) 9.97954 +time/total (s) 3766.49 +Epoch -633 +---------------------------------- --------------- +2022-05-10 14:13:36.493283 PDT | [2] Epoch -632 finished +---------------------------------- --------------- +epoch -632 +replay_buffer/size 999033 +trainer/num train calls 369000 +trainer/Policy Loss -19.59 +trainer/Log Pis Mean 24.1905 +trainer/Log Pis Std 13.3697 +trainer/Log Pis Max 73.2771 +trainer/Log Pis Min -5.73337 +trainer/policy/mean Mean -0.0342374 +trainer/policy/mean Std 0.902446 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.7869 +trainer/policy/normal/std Std 0.641423 +trainer/policy/normal/std Max 5.38072 +trainer/policy/normal/std Min 0.35639 +trainer/policy/normal/log_std Mean 0.986432 +trainer/policy/normal/log_std Std 0.313222 +trainer/policy/normal/log_std Max 1.68282 +trainer/policy/normal/log_std Min -1.03173 +eval/num steps total 368470 +eval/num paths total 369 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.355696 +eval/Actions Std 0.83786 +eval/Actions Max 0.999998 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73379 +time/logging (s) 0.00368188 +time/sampling batch (s) 0.537085 +time/saving (s) 0.00357827 +time/training (s) 6.82625 +time/epoch (s) 10.1044 +time/total (s) 3776.6 +Epoch -632 +---------------------------------- --------------- +2022-05-10 14:13:46.635610 PDT | [2] Epoch -631 finished +---------------------------------- --------------- +epoch -631 +replay_buffer/size 999033 +trainer/num train calls 370000 +trainer/Policy Loss -20.2441 +trainer/Log Pis Mean 23.8533 +trainer/Log Pis Std 13.4633 +trainer/Log Pis Max 77.3668 +trainer/Log Pis Min -7.8142 +trainer/policy/mean Mean -0.0183008 +trainer/policy/mean Std 0.906039 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.86259 +trainer/policy/normal/std Std 0.641399 +trainer/policy/normal/std Max 6.4993 +trainer/policy/normal/std Min 0.339282 +trainer/policy/normal/log_std Mean 1.01685 +trainer/policy/normal/log_std Std 0.294725 +trainer/policy/normal/log_std Max 1.87169 +trainer/policy/normal/log_std Min -1.08092 +eval/num steps total 369470 +eval/num paths total 370 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0715442 +eval/Actions Std 0.923535 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.71764 +time/logging (s) 0.00370191 +time/sampling batch (s) 0.28676 +time/saving (s) 0.00347397 +time/training (s) 7.1089 +time/epoch (s) 10.1205 +time/total (s) 3786.72 +Epoch -631 +---------------------------------- --------------- +2022-05-10 14:13:57.573368 PDT | [2] Epoch -630 finished +---------------------------------- --------------- +epoch -630 +replay_buffer/size 999033 +trainer/num train calls 371000 +trainer/Policy Loss -21.0503 +trainer/Log Pis Mean 25.6604 +trainer/Log Pis Std 13.0953 +trainer/Log Pis Max 70.2452 +trainer/Log Pis Min -8.62615 +trainer/policy/mean Mean -0.0240864 +trainer/policy/mean Std 0.911312 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.79059 +trainer/policy/normal/std Std 0.615701 +trainer/policy/normal/std Max 5.46267 +trainer/policy/normal/std Min 0.355639 +trainer/policy/normal/log_std Mean 0.992373 +trainer/policy/normal/log_std Std 0.290643 +trainer/policy/normal/log_std Max 1.69794 +trainer/policy/normal/log_std Min -1.03384 +eval/num steps total 370470 +eval/num paths total 371 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.360752 +eval/Actions Std 0.816897 +eval/Actions Max 0.999993 +eval/Actions Min -0.999987 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.48712 +time/logging (s) 0.00375471 +time/sampling batch (s) 0.283529 +time/saving (s) 0.00349992 +time/training (s) 8.13839 +time/epoch (s) 10.9163 +time/total (s) 3797.64 +Epoch -630 +---------------------------------- --------------- +2022-05-10 14:14:07.601402 PDT | [2] Epoch -629 finished +---------------------------------- --------------- +epoch -629 +replay_buffer/size 999033 +trainer/num train calls 372000 +trainer/Policy Loss -18.3136 +trainer/Log Pis Mean 24.7248 +trainer/Log Pis Std 13.9463 +trainer/Log Pis Max 78.9347 +trainer/Log Pis Min -5.35128 +trainer/policy/mean Mean -0.0320495 +trainer/policy/mean Std 0.904222 +trainer/policy/mean Max 0.999974 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.8201 +trainer/policy/normal/std Std 0.648106 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.281309 +trainer/policy/normal/log_std Mean 0.999729 +trainer/policy/normal/log_std Std 0.305417 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.2683 +eval/num steps total 371470 +eval/num paths total 372 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.424781 +eval/Actions Std 0.888872 +eval/Actions Max 0.999998 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.76969 +time/logging (s) 0.00410093 +time/sampling batch (s) 0.530629 +time/saving (s) 0.00395942 +time/training (s) 6.69843 +time/epoch (s) 10.0068 +time/total (s) 3807.65 +Epoch -629 +---------------------------------- --------------- +2022-05-10 14:14:18.898188 PDT | [2] Epoch -628 finished +---------------------------------- --------------- +epoch -628 +replay_buffer/size 999033 +trainer/num train calls 373000 +trainer/Policy Loss -19.7471 +trainer/Log Pis Mean 24.1906 +trainer/Log Pis Std 12.4969 +trainer/Log Pis Max 73.4559 +trainer/Log Pis Min -10.1837 +trainer/policy/mean Mean -0.0778409 +trainer/policy/mean Std 0.904053 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.80083 +trainer/policy/normal/std Std 0.654623 +trainer/policy/normal/std Max 5.89274 +trainer/policy/normal/std Min 0.302253 +trainer/policy/normal/log_std Mean 0.990107 +trainer/policy/normal/log_std Std 0.319292 +trainer/policy/normal/log_std Max 1.77372 +trainer/policy/normal/log_std Min -1.19649 +eval/num steps total 372470 +eval/num paths total 373 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.283766 +eval/Actions Std 0.68002 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.71712 +time/logging (s) 0.00394678 +time/sampling batch (s) 0.533163 +time/saving (s) 0.00390653 +time/training (s) 8.01666 +time/epoch (s) 11.2748 +time/total (s) 3818.93 +Epoch -628 +---------------------------------- --------------- +2022-05-10 14:14:28.235612 PDT | [2] Epoch -627 finished +---------------------------------- --------------- +epoch -627 +replay_buffer/size 999033 +trainer/num train calls 374000 +trainer/Policy Loss -19.907 +trainer/Log Pis Mean 25.3155 +trainer/Log Pis Std 13.0508 +trainer/Log Pis Max 86.1795 +trainer/Log Pis Min -7.48054 +trainer/policy/mean Mean -0.0320619 +trainer/policy/mean Std 0.907378 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.8509 +trainer/policy/normal/std Std 0.633633 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.328813 +trainer/policy/normal/log_std Mean 1.01376 +trainer/policy/normal/log_std Std 0.289427 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.11227 +eval/num steps total 373470 +eval/num paths total 374 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00331676 +eval/Actions Std 0.94365 +eval/Actions Max 0.999996 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70952 +time/logging (s) 0.00380061 +time/sampling batch (s) 0.28449 +time/saving (s) 0.00346695 +time/training (s) 6.31429 +time/epoch (s) 9.31557 +time/total (s) 3828.25 +Epoch -627 +---------------------------------- --------------- +2022-05-10 14:14:38.905149 PDT | [2] Epoch -626 finished +---------------------------------- --------------- +epoch -626 +replay_buffer/size 999033 +trainer/num train calls 375000 +trainer/Policy Loss -18.7624 +trainer/Log Pis Mean 24.4526 +trainer/Log Pis Std 12.9931 +trainer/Log Pis Max 80.9452 +trainer/Log Pis Min -4.48336 +trainer/policy/mean Mean -0.040256 +trainer/policy/mean Std 0.907313 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.81189 +trainer/policy/normal/std Std 0.633294 +trainer/policy/normal/std Max 5.79717 +trainer/policy/normal/std Min 0.265333 +trainer/policy/normal/log_std Mean 0.997217 +trainer/policy/normal/log_std Std 0.305527 +trainer/policy/normal/log_std Max 1.75737 +trainer/policy/normal/log_std Min -1.32677 +eval/num steps total 374470 +eval/num paths total 375 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.360089 +eval/Actions Std 0.86363 +eval/Actions Max 0.99999 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.74238 +time/logging (s) 0.00371105 +time/sampling batch (s) 0.278903 +time/saving (s) 0.00342625 +time/training (s) 7.61975 +time/epoch (s) 10.6482 +time/total (s) 3838.9 +Epoch -626 +---------------------------------- --------------- +2022-05-10 14:14:48.983344 PDT | [2] Epoch -625 finished +---------------------------------- --------------- +epoch -625 +replay_buffer/size 999033 +trainer/num train calls 376000 +trainer/Policy Loss -20.4044 +trainer/Log Pis Mean 25.0469 +trainer/Log Pis Std 14.24 +trainer/Log Pis Max 71.7993 +trainer/Log Pis Min -9.33482 +trainer/policy/mean Mean -0.0248744 +trainer/policy/mean Std 0.907045 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.82182 +trainer/policy/normal/std Std 0.636775 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.37655 +trainer/policy/normal/log_std Mean 1.00192 +trainer/policy/normal/log_std Std 0.29758 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -0.976704 +eval/num steps total 375470 +eval/num paths total 376 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.099495 +eval/Actions Std 0.918411 +eval/Actions Max 0.999983 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50956 +time/logging (s) 0.00377978 +time/sampling batch (s) 0.279115 +time/saving (s) 0.0035576 +time/training (s) 7.26095 +time/epoch (s) 10.057 +time/total (s) 3848.96 +Epoch -625 +---------------------------------- --------------- +2022-05-10 14:14:59.586067 PDT | [2] Epoch -624 finished +---------------------------------- --------------- +epoch -624 +replay_buffer/size 999033 +trainer/num train calls 377000 +trainer/Policy Loss -20.1809 +trainer/Log Pis Mean 25.6775 +trainer/Log Pis Std 13.8121 +trainer/Log Pis Max 74.6437 +trainer/Log Pis Min -9.22999 +trainer/policy/mean Mean -0.0609135 +trainer/policy/mean Std 0.908764 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999985 +trainer/policy/normal/std Mean 2.84065 +trainer/policy/normal/std Std 0.629797 +trainer/policy/normal/std Max 5.05647 +trainer/policy/normal/std Min 0.384542 +trainer/policy/normal/log_std Mean 1.00998 +trainer/policy/normal/log_std Std 0.289932 +trainer/policy/normal/log_std Max 1.62067 +trainer/policy/normal/log_std Min -0.955703 +eval/num steps total 376470 +eval/num paths total 377 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0461142 +eval/Actions Std 0.940963 +eval/Actions Max 0.999991 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63374 +time/logging (s) 0.00380703 +time/sampling batch (s) 0.780819 +time/saving (s) 0.00359483 +time/training (s) 7.15927 +time/epoch (s) 10.5812 +time/total (s) 3859.54 +Epoch -624 +---------------------------------- --------------- +2022-05-10 14:15:09.595895 PDT | [2] Epoch -623 finished +---------------------------------- --------------- +epoch -623 +replay_buffer/size 999033 +trainer/num train calls 378000 +trainer/Policy Loss -19.7339 +trainer/Log Pis Mean 24.3645 +trainer/Log Pis Std 13.2479 +trainer/Log Pis Max 65.9905 +trainer/Log Pis Min -6.34899 +trainer/policy/mean Mean -0.0405551 +trainer/policy/mean Std 0.903954 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.81104 +trainer/policy/normal/std Std 0.648826 +trainer/policy/normal/std Max 5.41767 +trainer/policy/normal/std Min 0.292575 +trainer/policy/normal/log_std Mean 0.995817 +trainer/policy/normal/log_std Std 0.308322 +trainer/policy/normal/log_std Max 1.68967 +trainer/policy/normal/log_std Min -1.22903 +eval/num steps total 377470 +eval/num paths total 378 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.222106 +eval/Actions Std 0.877256 +eval/Actions Max 0.999997 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51126 +time/logging (s) 0.00425731 +time/sampling batch (s) 0.298474 +time/saving (s) 0.00431367 +time/training (s) 7.17001 +time/epoch (s) 9.98832 +time/total (s) 3869.53 +Epoch -623 +---------------------------------- --------------- +2022-05-10 14:15:20.101951 PDT | [2] Epoch -622 finished +---------------------------------- ---------------- +epoch -622 +replay_buffer/size 999033 +trainer/num train calls 379000 +trainer/Policy Loss -19.513 +trainer/Log Pis Mean 25.2039 +trainer/Log Pis Std 13.5589 +trainer/Log Pis Max 64.7629 +trainer/Log Pis Min -6.46191 +trainer/policy/mean Mean -0.0610084 +trainer/policy/mean Std 0.905129 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.79556 +trainer/policy/normal/std Std 0.612961 +trainer/policy/normal/std Max 5.2964 +trainer/policy/normal/std Min 0.376832 +trainer/policy/normal/log_std Mean 0.993813 +trainer/policy/normal/log_std Std 0.293909 +trainer/policy/normal/log_std Max 1.66703 +trainer/policy/normal/log_std Min -0.975957 +eval/num steps total 378470 +eval/num paths total 379 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.000778984 +eval/Actions Std 0.903292 +eval/Actions Max 0.999998 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64171 +time/logging (s) 0.00420122 +time/sampling batch (s) 0.53771 +time/saving (s) 0.00491478 +time/training (s) 7.29471 +time/epoch (s) 10.4833 +time/total (s) 3880.02 +Epoch -622 +---------------------------------- ---------------- +2022-05-10 14:15:30.480708 PDT | [2] Epoch -621 finished +---------------------------------- --------------- +epoch -621 +replay_buffer/size 999033 +trainer/num train calls 380000 +trainer/Policy Loss -19.185 +trainer/Log Pis Mean 24.6548 +trainer/Log Pis Std 13.7718 +trainer/Log Pis Max 68.0851 +trainer/Log Pis Min -9.68834 +trainer/policy/mean Mean -0.0449569 +trainer/policy/mean Std 0.904971 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999983 +trainer/policy/normal/std Mean 2.91431 +trainer/policy/normal/std Std 0.662738 +trainer/policy/normal/std Max 5.46068 +trainer/policy/normal/std Min 0.307861 +trainer/policy/normal/log_std Mean 1.03193 +trainer/policy/normal/log_std Std 0.31058 +trainer/policy/normal/log_std Max 1.69757 +trainer/policy/normal/log_std Min -1.17811 +eval/num steps total 379470 +eval/num paths total 380 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.184968 +eval/Actions Std 0.918516 +eval/Actions Max 0.999994 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.79854 +time/logging (s) 0.00423243 +time/sampling batch (s) 0.295174 +time/saving (s) 0.00432349 +time/training (s) 7.25277 +time/epoch (s) 10.355 +time/total (s) 3890.38 +Epoch -621 +---------------------------------- --------------- +2022-05-10 14:15:41.188810 PDT | [2] Epoch -620 finished +---------------------------------- --------------- +epoch -620 +replay_buffer/size 999033 +trainer/num train calls 381000 +trainer/Policy Loss -19.2641 +trainer/Log Pis Mean 24.421 +trainer/Log Pis Std 12.6495 +trainer/Log Pis Max 71.5374 +trainer/Log Pis Min -1.65785 +trainer/policy/mean Mean -0.0291398 +trainer/policy/mean Std 0.905204 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.77565 +trainer/policy/normal/std Std 0.629385 +trainer/policy/normal/std Max 6.29244 +trainer/policy/normal/std Min 0.275112 +trainer/policy/normal/log_std Mean 0.984791 +trainer/policy/normal/log_std Std 0.300754 +trainer/policy/normal/log_std Max 1.83935 +trainer/policy/normal/log_std Min -1.29058 +eval/num steps total 380470 +eval/num paths total 381 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0498439 +eval/Actions Std 0.903701 +eval/Actions Max 0.999996 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67122 +time/logging (s) 0.00388994 +time/sampling batch (s) 0.309778 +time/saving (s) 0.00488706 +time/training (s) 7.69504 +time/epoch (s) 10.6848 +time/total (s) 3901.07 +Epoch -620 +---------------------------------- --------------- +2022-05-10 14:15:52.768678 PDT | [2] Epoch -619 finished +---------------------------------- --------------- +epoch -619 +replay_buffer/size 999033 +trainer/num train calls 382000 +trainer/Policy Loss -19.9854 +trainer/Log Pis Mean 24.892 +trainer/Log Pis Std 12.1393 +trainer/Log Pis Max 68.7987 +trainer/Log Pis Min -9.45271 +trainer/policy/mean Mean -0.0211254 +trainer/policy/mean Std 0.910205 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.83657 +trainer/policy/normal/std Std 0.624558 +trainer/policy/normal/std Max 5.29696 +trainer/policy/normal/std Min 0.338007 +trainer/policy/normal/log_std Mean 1.00727 +trainer/policy/normal/log_std Std 0.300419 +trainer/policy/normal/log_std Max 1.66713 +trainer/policy/normal/log_std Min -1.08469 +eval/num steps total 381470 +eval/num paths total 382 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0675575 +eval/Actions Std 0.899037 +eval/Actions Max 0.999992 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68292 +time/logging (s) 0.00377003 +time/sampling batch (s) 0.585257 +time/saving (s) 0.00345203 +time/training (s) 8.28007 +time/epoch (s) 11.5555 +time/total (s) 3912.63 +Epoch -619 +---------------------------------- --------------- +2022-05-10 14:16:03.730070 PDT | [2] Epoch -618 finished +---------------------------------- --------------- +epoch -618 +replay_buffer/size 999033 +trainer/num train calls 383000 +trainer/Policy Loss -19.153 +trainer/Log Pis Mean 24.3295 +trainer/Log Pis Std 13.4324 +trainer/Log Pis Max 66.8711 +trainer/Log Pis Min -7.89242 +trainer/policy/mean Mean -0.0113224 +trainer/policy/mean Std 0.900834 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.81709 +trainer/policy/normal/std Std 0.666049 +trainer/policy/normal/std Max 6.98976 +trainer/policy/normal/std Min 0.350664 +trainer/policy/normal/log_std Mean 0.995428 +trainer/policy/normal/log_std Std 0.319855 +trainer/policy/normal/log_std Max 1.94445 +trainer/policy/normal/log_std Min -1.04793 +eval/num steps total 382470 +eval/num paths total 383 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0241297 +eval/Actions Std 0.897859 +eval/Actions Max 0.999996 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6447 +time/logging (s) 0.00472064 +time/sampling batch (s) 0.301161 +time/saving (s) 0.00408422 +time/training (s) 7.98526 +time/epoch (s) 10.9399 +time/total (s) 3923.57 +Epoch -618 +---------------------------------- --------------- +2022-05-10 14:16:14.088486 PDT | [2] Epoch -617 finished +---------------------------------- --------------- +epoch -617 +replay_buffer/size 999033 +trainer/num train calls 384000 +trainer/Policy Loss -20.1797 +trainer/Log Pis Mean 24.2981 +trainer/Log Pis Std 12.5932 +trainer/Log Pis Max 67.4548 +trainer/Log Pis Min -6.0446 +trainer/policy/mean Mean -0.0244774 +trainer/policy/mean Std 0.90354 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.8235 +trainer/policy/normal/std Std 0.628064 +trainer/policy/normal/std Max 6.09671 +trainer/policy/normal/std Min 0.223165 +trainer/policy/normal/log_std Mean 1.00245 +trainer/policy/normal/log_std Std 0.30083 +trainer/policy/normal/log_std Max 1.80775 +trainer/policy/normal/log_std Min -1.49984 +eval/num steps total 383470 +eval/num paths total 384 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.087137 +eval/Actions Std 0.914221 +eval/Actions Max 0.999997 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.83597 +time/logging (s) 0.00394977 +time/sampling batch (s) 0.536741 +time/saving (s) 0.00414998 +time/training (s) 6.95387 +time/epoch (s) 10.3347 +time/total (s) 3933.91 +Epoch -617 +---------------------------------- --------------- +2022-05-10 14:16:23.522182 PDT | [2] Epoch -616 finished +---------------------------------- --------------- +epoch -616 +replay_buffer/size 999033 +trainer/num train calls 385000 +trainer/Policy Loss -20.4957 +trainer/Log Pis Mean 24.9316 +trainer/Log Pis Std 12.7918 +trainer/Log Pis Max 66.7369 +trainer/Log Pis Min -6.71266 +trainer/policy/mean Mean -0.0393996 +trainer/policy/mean Std 0.905097 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.81792 +trainer/policy/normal/std Std 0.644645 +trainer/policy/normal/std Max 6.32947 +trainer/policy/normal/std Min 0.392233 +trainer/policy/normal/log_std Mean 0.998165 +trainer/policy/normal/log_std Std 0.310595 +trainer/policy/normal/log_std Max 1.84522 +trainer/policy/normal/log_std Min -0.935899 +eval/num steps total 384470 +eval/num paths total 385 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00563335 +eval/Actions Std 0.900406 +eval/Actions Max 0.999996 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67797 +time/logging (s) 0.00376174 +time/sampling batch (s) 0.280148 +time/saving (s) 0.00344199 +time/training (s) 6.44663 +time/epoch (s) 9.41195 +time/total (s) 3943.32 +Epoch -616 +---------------------------------- --------------- +2022-05-10 14:16:34.393936 PDT | [2] Epoch -615 finished +---------------------------------- --------------- +epoch -615 +replay_buffer/size 999033 +trainer/num train calls 386000 +trainer/Policy Loss -19.7894 +trainer/Log Pis Mean 25.3865 +trainer/Log Pis Std 13.5063 +trainer/Log Pis Max 79.5085 +trainer/Log Pis Min -10.2409 +trainer/policy/mean Mean -0.0498722 +trainer/policy/mean Std 0.907353 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.82592 +trainer/policy/normal/std Std 0.63321 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.309944 +trainer/policy/normal/log_std Mean 1.00371 +trainer/policy/normal/log_std Std 0.296303 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.17136 +eval/num steps total 385470 +eval/num paths total 386 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.11698 +eval/Actions Std 0.912029 +eval/Actions Max 0.999992 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.86056 +time/logging (s) 0.0038078 +time/sampling batch (s) 0.279794 +time/saving (s) 0.00356788 +time/training (s) 7.70259 +time/epoch (s) 10.8503 +time/total (s) 3954.18 +Epoch -615 +---------------------------------- --------------- +2022-05-10 14:16:44.183765 PDT | [2] Epoch -614 finished +---------------------------------- --------------- +epoch -614 +replay_buffer/size 999033 +trainer/num train calls 387000 +trainer/Policy Loss -18.8838 +trainer/Log Pis Mean 23.7197 +trainer/Log Pis Std 13.4814 +trainer/Log Pis Max 71.8175 +trainer/Log Pis Min -7.77399 +trainer/policy/mean Mean -0.0399815 +trainer/policy/mean Std 0.901518 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.79027 +trainer/policy/normal/std Std 0.623127 +trainer/policy/normal/std Max 6.52406 +trainer/policy/normal/std Min 0.339343 +trainer/policy/normal/log_std Mean 0.991064 +trainer/policy/normal/log_std Std 0.295261 +trainer/policy/normal/log_std Max 1.8755 +trainer/policy/normal/log_std Min -1.08074 +eval/num steps total 386470 +eval/num paths total 387 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.123483 +eval/Actions Std 0.910899 +eval/Actions Max 0.999999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64486 +time/logging (s) 0.00409493 +time/sampling batch (s) 0.280671 +time/saving (s) 0.00398517 +time/training (s) 6.83492 +time/epoch (s) 9.76853 +time/total (s) 3963.95 +Epoch -614 +---------------------------------- --------------- +2022-05-10 14:16:54.689156 PDT | [2] Epoch -613 finished +---------------------------------- --------------- +epoch -613 +replay_buffer/size 999033 +trainer/num train calls 388000 +trainer/Policy Loss -19.7008 +trainer/Log Pis Mean 25.3393 +trainer/Log Pis Std 12.8441 +trainer/Log Pis Max 62.5915 +trainer/Log Pis Min -5.40517 +trainer/policy/mean Mean -0.0438693 +trainer/policy/mean Std 0.911456 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.82872 +trainer/policy/normal/std Std 0.624694 +trainer/policy/normal/std Max 5.19983 +trainer/policy/normal/std Min 0.390983 +trainer/policy/normal/log_std Mean 1.00515 +trainer/policy/normal/log_std Std 0.295245 +trainer/policy/normal/log_std Max 1.64863 +trainer/policy/normal/log_std Min -0.939092 +eval/num steps total 387470 +eval/num paths total 388 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.218887 +eval/Actions Std 0.826265 +eval/Actions Max 0.999997 +eval/Actions Min -0.999981 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56547 +time/logging (s) 0.00385156 +time/sampling batch (s) 0.53075 +time/saving (s) 0.00372075 +time/training (s) 7.37952 +time/epoch (s) 10.4833 +time/total (s) 3974.43 +Epoch -613 +---------------------------------- --------------- +2022-05-10 14:17:05.539176 PDT | [2] Epoch -612 finished +---------------------------------- --------------- +epoch -612 +replay_buffer/size 999033 +trainer/num train calls 389000 +trainer/Policy Loss -20.3043 +trainer/Log Pis Mean 24.8469 +trainer/Log Pis Std 14.3488 +trainer/Log Pis Max 71.3885 +trainer/Log Pis Min -8.78087 +trainer/policy/mean Mean -0.0294734 +trainer/policy/mean Std 0.908853 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.81218 +trainer/policy/normal/std Std 0.648383 +trainer/policy/normal/std Max 5.82727 +trainer/policy/normal/std Min 0.306608 +trainer/policy/normal/log_std Mean 0.993204 +trainer/policy/normal/log_std Std 0.327948 +trainer/policy/normal/log_std Max 1.76255 +trainer/policy/normal/log_std Min -1.18219 +eval/num steps total 388470 +eval/num paths total 389 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.258458 +eval/Actions Std 0.754455 +eval/Actions Max 0.999993 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.47656 +time/logging (s) 0.00381765 +time/sampling batch (s) 0.280122 +time/saving (s) 0.00359651 +time/training (s) 8.0642 +time/epoch (s) 10.8283 +time/total (s) 3985.26 +Epoch -612 +---------------------------------- --------------- +2022-05-10 14:17:16.370050 PDT | [2] Epoch -611 finished +---------------------------------- --------------- +epoch -611 +replay_buffer/size 999033 +trainer/num train calls 390000 +trainer/Policy Loss -19.3817 +trainer/Log Pis Mean 24.3357 +trainer/Log Pis Std 12.8701 +trainer/Log Pis Max 72.1322 +trainer/Log Pis Min -8.01116 +trainer/policy/mean Mean -0.0336885 +trainer/policy/mean Std 0.904704 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.86552 +trainer/policy/normal/std Std 0.619459 +trainer/policy/normal/std Max 5.21208 +trainer/policy/normal/std Min 0.305183 +trainer/policy/normal/log_std Mean 1.01999 +trainer/policy/normal/log_std Std 0.286124 +trainer/policy/normal/log_std Max 1.65098 +trainer/policy/normal/log_std Min -1.18684 +eval/num steps total 389470 +eval/num paths total 390 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0964163 +eval/Actions Std 0.906532 +eval/Actions Max 0.999991 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.71356 +time/logging (s) 0.00364035 +time/sampling batch (s) 0.282887 +time/saving (s) 0.00345156 +time/training (s) 7.80549 +time/epoch (s) 10.809 +time/total (s) 3996.08 +Epoch -611 +---------------------------------- --------------- +2022-05-10 14:17:25.761781 PDT | [2] Epoch -610 finished +---------------------------------- --------------- +epoch -610 +replay_buffer/size 999033 +trainer/num train calls 391000 +trainer/Policy Loss -19.5529 +trainer/Log Pis Mean 24.0708 +trainer/Log Pis Std 13.0603 +trainer/Log Pis Max 65.7358 +trainer/Log Pis Min -5.6619 +trainer/policy/mean Mean -0.0313297 +trainer/policy/mean Std 0.906471 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.82623 +trainer/policy/normal/std Std 0.627844 +trainer/policy/normal/std Max 5.64132 +trainer/policy/normal/std Min 0.365841 +trainer/policy/normal/log_std Mean 1.00422 +trainer/policy/normal/log_std Std 0.294829 +trainer/policy/normal/log_std Max 1.73012 +trainer/policy/normal/log_std Min -1.00556 +eval/num steps total 390470 +eval/num paths total 391 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0444515 +eval/Actions Std 0.892243 +eval/Actions Max 0.999996 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68271 +time/logging (s) 0.00376775 +time/sampling batch (s) 0.279306 +time/saving (s) 0.00344774 +time/training (s) 6.40123 +time/epoch (s) 9.37046 +time/total (s) 4005.45 +Epoch -610 +---------------------------------- --------------- +2022-05-10 14:17:36.481631 PDT | [2] Epoch -609 finished +---------------------------------- --------------- +epoch -609 +replay_buffer/size 999033 +trainer/num train calls 392000 +trainer/Policy Loss -18.5671 +trainer/Log Pis Mean 23.9306 +trainer/Log Pis Std 13.6752 +trainer/Log Pis Max 69.1633 +trainer/Log Pis Min -11.1952 +trainer/policy/mean Mean -0.0219633 +trainer/policy/mean Std 0.905562 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.83528 +trainer/policy/normal/std Std 0.622282 +trainer/policy/normal/std Max 5.07749 +trainer/policy/normal/std Min 0.303246 +trainer/policy/normal/log_std Mean 1.00752 +trainer/policy/normal/log_std Std 0.296977 +trainer/policy/normal/log_std Max 1.62482 +trainer/policy/normal/log_std Min -1.19321 +eval/num steps total 391470 +eval/num paths total 392 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.219776 +eval/Actions Std 0.899815 +eval/Actions Max 0.999992 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.75201 +time/logging (s) 0.00380537 +time/sampling batch (s) 0.536476 +time/saving (s) 0.00351236 +time/training (s) 7.40215 +time/epoch (s) 10.698 +time/total (s) 4016.15 +Epoch -609 +---------------------------------- --------------- +2022-05-10 14:17:46.042543 PDT | [2] Epoch -608 finished +---------------------------------- --------------- +epoch -608 +replay_buffer/size 999033 +trainer/num train calls 393000 +trainer/Policy Loss -19.8785 +trainer/Log Pis Mean 23.2949 +trainer/Log Pis Std 12.6698 +trainer/Log Pis Max 69.9654 +trainer/Log Pis Min -8.81421 +trainer/policy/mean Mean -0.0311303 +trainer/policy/mean Std 0.90201 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.79793 +trainer/policy/normal/std Std 0.628239 +trainer/policy/normal/std Max 5.77735 +trainer/policy/normal/std Min 0.382059 +trainer/policy/normal/log_std Mean 0.992434 +trainer/policy/normal/log_std Std 0.303874 +trainer/policy/normal/log_std Max 1.75395 +trainer/policy/normal/log_std Min -0.962182 +eval/num steps total 392470 +eval/num paths total 393 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.109641 +eval/Actions Std 0.913785 +eval/Actions Max 0.999985 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.52918 +time/logging (s) 0.00401182 +time/sampling batch (s) 0.281662 +time/saving (s) 0.00419966 +time/training (s) 6.72021 +time/epoch (s) 9.53927 +time/total (s) 4025.69 +Epoch -608 +---------------------------------- --------------- +2022-05-10 14:17:55.745446 PDT | [2] Epoch -607 finished +---------------------------------- --------------- +epoch -607 +replay_buffer/size 999033 +trainer/num train calls 394000 +trainer/Policy Loss -18.3856 +trainer/Log Pis Mean 24.7947 +trainer/Log Pis Std 13.137 +trainer/Log Pis Max 73.2321 +trainer/Log Pis Min -11.121 +trainer/policy/mean Mean -0.0394487 +trainer/policy/mean Std 0.909738 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.8086 +trainer/policy/normal/std Std 0.628862 +trainer/policy/normal/std Max 6.44955 +trainer/policy/normal/std Min 0.327179 +trainer/policy/normal/log_std Mean 0.996513 +trainer/policy/normal/log_std Std 0.304308 +trainer/policy/normal/log_std Max 1.86401 +trainer/policy/normal/log_std Min -1.11725 +eval/num steps total 393470 +eval/num paths total 394 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00949811 +eval/Actions Std 0.883566 +eval/Actions Max 0.999987 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.4419 +time/logging (s) 0.00372721 +time/sampling batch (s) 0.53066 +time/saving (s) 0.00357797 +time/training (s) 6.70069 +time/epoch (s) 9.68055 +time/total (s) 4035.38 +Epoch -607 +---------------------------------- --------------- +2022-05-10 14:18:05.906682 PDT | [2] Epoch -606 finished +---------------------------------- --------------- +epoch -606 +replay_buffer/size 999033 +trainer/num train calls 395000 +trainer/Policy Loss -19.0982 +trainer/Log Pis Mean 24.0575 +trainer/Log Pis Std 13.6566 +trainer/Log Pis Max 79.355 +trainer/Log Pis Min -7.22855 +trainer/policy/mean Mean -0.0374533 +trainer/policy/mean Std 0.903466 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.79917 +trainer/policy/normal/std Std 0.635157 +trainer/policy/normal/std Max 5.46821 +trainer/policy/normal/std Min 0.275539 +trainer/policy/normal/log_std Mean 0.992134 +trainer/policy/normal/log_std Std 0.306958 +trainer/policy/normal/log_std Max 1.69895 +trainer/policy/normal/log_std Min -1.28903 +eval/num steps total 394470 +eval/num paths total 395 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0196499 +eval/Actions Std 0.910939 +eval/Actions Max 0.999994 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63752 +time/logging (s) 0.00384072 +time/sampling batch (s) 0.534128 +time/saving (s) 0.00344407 +time/training (s) 6.96066 +time/epoch (s) 10.1396 +time/total (s) 4045.52 +Epoch -606 +---------------------------------- --------------- +2022-05-10 14:18:16.152120 PDT | [2] Epoch -605 finished +---------------------------------- --------------- +epoch -605 +replay_buffer/size 999033 +trainer/num train calls 396000 +trainer/Policy Loss -20.1071 +trainer/Log Pis Mean 24.3481 +trainer/Log Pis Std 12.9368 +trainer/Log Pis Max 66.7939 +trainer/Log Pis Min -7.37516 +trainer/policy/mean Mean -0.0230452 +trainer/policy/mean Std 0.904123 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.77371 +trainer/policy/normal/std Std 0.633412 +trainer/policy/normal/std Max 5.30924 +trainer/policy/normal/std Min 0.278623 +trainer/policy/normal/log_std Mean 0.982196 +trainer/policy/normal/log_std Std 0.311896 +trainer/policy/normal/log_std Max 1.66945 +trainer/policy/normal/log_std Min -1.2779 +eval/num steps total 395470 +eval/num paths total 396 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0553597 +eval/Actions Std 0.903849 +eval/Actions Max 0.999998 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63912 +time/logging (s) 0.00367883 +time/sampling batch (s) 0.284108 +time/saving (s) 0.00349537 +time/training (s) 7.2932 +time/epoch (s) 10.2236 +time/total (s) 4055.75 +Epoch -605 +---------------------------------- --------------- +2022-05-10 14:18:26.596582 PDT | [2] Epoch -604 finished +---------------------------------- --------------- +epoch -604 +replay_buffer/size 999033 +trainer/num train calls 397000 +trainer/Policy Loss -19.4623 +trainer/Log Pis Mean 25.9977 +trainer/Log Pis Std 13.364 +trainer/Log Pis Max 78.3112 +trainer/Log Pis Min -4.12552 +trainer/policy/mean Mean -0.0267239 +trainer/policy/mean Std 0.906619 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.81559 +trainer/policy/normal/std Std 0.63427 +trainer/policy/normal/std Max 5.89513 +trainer/policy/normal/std Min 0.303476 +trainer/policy/normal/log_std Mean 0.999362 +trainer/policy/normal/log_std Std 0.299822 +trainer/policy/normal/log_std Max 1.77413 +trainer/policy/normal/log_std Min -1.19245 +eval/num steps total 396470 +eval/num paths total 397 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.134826 +eval/Actions Std 0.888107 +eval/Actions Max 0.999993 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70211 +time/logging (s) 0.00376703 +time/sampling batch (s) 0.532224 +time/saving (s) 0.00346802 +time/training (s) 7.18134 +time/epoch (s) 10.4229 +time/total (s) 4066.17 +Epoch -604 +---------------------------------- --------------- +2022-05-10 14:18:37.542210 PDT | [2] Epoch -603 finished +---------------------------------- --------------- +epoch -603 +replay_buffer/size 999033 +trainer/num train calls 398000 +trainer/Policy Loss -20.3525 +trainer/Log Pis Mean 23.5823 +trainer/Log Pis Std 13.4023 +trainer/Log Pis Max 63.518 +trainer/Log Pis Min -12.7914 +trainer/policy/mean Mean -0.0261391 +trainer/policy/mean Std 0.90853 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.79798 +trainer/policy/normal/std Std 0.637332 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.359514 +trainer/policy/normal/log_std Mean 0.99282 +trainer/policy/normal/log_std Std 0.299883 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.023 +eval/num steps total 397470 +eval/num paths total 398 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00879203 +eval/Actions Std 0.899227 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63642 +time/logging (s) 0.00433782 +time/sampling batch (s) 0.544147 +time/saving (s) 0.0045204 +time/training (s) 7.73438 +time/epoch (s) 10.9238 +time/total (s) 4077.1 +Epoch -603 +---------------------------------- --------------- +2022-05-10 14:18:48.157608 PDT | [2] Epoch -602 finished +---------------------------------- --------------- +epoch -602 +replay_buffer/size 999033 +trainer/num train calls 399000 +trainer/Policy Loss -21.3857 +trainer/Log Pis Mean 25.6808 +trainer/Log Pis Std 13.0414 +trainer/Log Pis Max 62.4607 +trainer/Log Pis Min -7.63153 +trainer/policy/mean Mean -0.0502629 +trainer/policy/mean Std 0.911182 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.86562 +trainer/policy/normal/std Std 0.621654 +trainer/policy/normal/std Max 6.83868 +trainer/policy/normal/std Min 0.317495 +trainer/policy/normal/log_std Mean 1.0204 +trainer/policy/normal/log_std Std 0.283704 +trainer/policy/normal/log_std Max 1.92259 +trainer/policy/normal/log_std Min -1.14729 +eval/num steps total 398470 +eval/num paths total 399 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0755452 +eval/Actions Std 0.918029 +eval/Actions Max 0.999995 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64707 +time/logging (s) 0.00375468 +time/sampling batch (s) 0.285497 +time/saving (s) 0.00355391 +time/training (s) 7.65271 +time/epoch (s) 10.5926 +time/total (s) 4087.69 +Epoch -602 +---------------------------------- --------------- +2022-05-10 14:18:59.381307 PDT | [2] Epoch -601 finished +---------------------------------- --------------- +epoch -601 +replay_buffer/size 999033 +trainer/num train calls 400000 +trainer/Policy Loss -18.7087 +trainer/Log Pis Mean 24.8085 +trainer/Log Pis Std 13.4719 +trainer/Log Pis Max 72.2448 +trainer/Log Pis Min -7.79202 +trainer/policy/mean Mean -0.0672481 +trainer/policy/mean Std 0.904846 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.82551 +trainer/policy/normal/std Std 0.632488 +trainer/policy/normal/std Max 6.83244 +trainer/policy/normal/std Min 0.300896 +trainer/policy/normal/log_std Mean 1.00274 +trainer/policy/normal/log_std Std 0.302139 +trainer/policy/normal/log_std Max 1.92168 +trainer/policy/normal/log_std Min -1.20099 +eval/num steps total 399470 +eval/num paths total 400 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0958309 +eval/Actions Std 0.903669 +eval/Actions Max 0.999987 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62554 +time/logging (s) 0.00390141 +time/sampling batch (s) 0.5432 +time/saving (s) 0.00650827 +time/training (s) 8.02228 +time/epoch (s) 11.2014 +time/total (s) 4098.9 +Epoch -601 +---------------------------------- --------------- +2022-05-10 14:19:09.981036 PDT | [2] Epoch -600 finished +---------------------------------- --------------- +epoch -600 +replay_buffer/size 999033 +trainer/num train calls 401000 +trainer/Policy Loss -19.7431 +trainer/Log Pis Mean 25.1542 +trainer/Log Pis Std 13.0905 +trainer/Log Pis Max 67.0026 +trainer/Log Pis Min -9.58934 +trainer/policy/mean Mean -0.029346 +trainer/policy/mean Std 0.907792 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.81542 +trainer/policy/normal/std Std 0.638239 +trainer/policy/normal/std Max 5.50192 +trainer/policy/normal/std Min 0.292567 +trainer/policy/normal/log_std Mean 0.997703 +trainer/policy/normal/log_std Std 0.309966 +trainer/policy/normal/log_std Max 1.7051 +trainer/policy/normal/log_std Min -1.22906 +eval/num steps total 400470 +eval/num paths total 401 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.37603 +eval/Actions Std 0.892652 +eval/Actions Max 0.999991 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65254 +time/logging (s) 0.00382165 +time/sampling batch (s) 0.539898 +time/saving (s) 0.00399065 +time/training (s) 7.37573 +time/epoch (s) 10.576 +time/total (s) 4109.48 +Epoch -600 +---------------------------------- --------------- +2022-05-10 14:19:20.051539 PDT | [2] Epoch -599 finished +---------------------------------- --------------- +epoch -599 +replay_buffer/size 999033 +trainer/num train calls 402000 +trainer/Policy Loss -20.305 +trainer/Log Pis Mean 24.4782 +trainer/Log Pis Std 12.6373 +trainer/Log Pis Max 73.3594 +trainer/Log Pis Min -4.76662 +trainer/policy/mean Mean -0.0420839 +trainer/policy/mean Std 0.909308 +trainer/policy/mean Max 0.999975 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.81173 +trainer/policy/normal/std Std 0.599177 +trainer/policy/normal/std Max 5.07693 +trainer/policy/normal/std Min 0.333067 +trainer/policy/normal/log_std Mean 1.00176 +trainer/policy/normal/log_std Std 0.282869 +trainer/policy/normal/log_std Max 1.62471 +trainer/policy/normal/log_std Min -1.09941 +eval/num steps total 401470 +eval/num paths total 402 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.281943 +eval/Actions Std 0.879623 +eval/Actions Max 0.999992 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63914 +time/logging (s) 0.00428176 +time/sampling batch (s) 0.288126 +time/saving (s) 0.0044924 +time/training (s) 7.11216 +time/epoch (s) 10.0482 +time/total (s) 4119.53 +Epoch -599 +---------------------------------- --------------- +2022-05-10 14:19:29.876380 PDT | [2] Epoch -598 finished +---------------------------------- --------------- +epoch -598 +replay_buffer/size 999033 +trainer/num train calls 403000 +trainer/Policy Loss -19.1958 +trainer/Log Pis Mean 23.968 +trainer/Log Pis Std 12.9568 +trainer/Log Pis Max 75.2837 +trainer/Log Pis Min -11.1179 +trainer/policy/mean Mean -0.0263244 +trainer/policy/mean Std 0.907875 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.81646 +trainer/policy/normal/std Std 0.637187 +trainer/policy/normal/std Max 5.55987 +trainer/policy/normal/std Min 0.252452 +trainer/policy/normal/log_std Mean 0.997173 +trainer/policy/normal/log_std Std 0.315702 +trainer/policy/normal/log_std Max 1.71557 +trainer/policy/normal/log_std Min -1.37653 +eval/num steps total 402470 +eval/num paths total 403 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.105183 +eval/Actions Std 0.873164 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58832 +time/logging (s) 0.00415343 +time/sampling batch (s) 0.284758 +time/saving (s) 0.00428779 +time/training (s) 6.92003 +time/epoch (s) 9.80156 +time/total (s) 4129.33 +Epoch -598 +---------------------------------- --------------- +2022-05-10 14:19:39.864773 PDT | [2] Epoch -597 finished +---------------------------------- --------------- +epoch -597 +replay_buffer/size 999033 +trainer/num train calls 404000 +trainer/Policy Loss -19.0721 +trainer/Log Pis Mean 24.9482 +trainer/Log Pis Std 13.7028 +trainer/Log Pis Max 80.666 +trainer/Log Pis Min -7.43821 +trainer/policy/mean Mean -0.0613646 +trainer/policy/mean Std 0.906829 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.82783 +trainer/policy/normal/std Std 0.635911 +trainer/policy/normal/std Max 5.98474 +trainer/policy/normal/std Min 0.254107 +trainer/policy/normal/log_std Mean 1.00265 +trainer/policy/normal/log_std Std 0.308458 +trainer/policy/normal/log_std Max 1.78921 +trainer/policy/normal/log_std Min -1.37 +eval/num steps total 403470 +eval/num paths total 404 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.139946 +eval/Actions Std 0.889899 +eval/Actions Max 0.999989 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.52562 +time/logging (s) 0.0041664 +time/sampling batch (s) 0.28694 +time/saving (s) 0.00414818 +time/training (s) 7.14465 +time/epoch (s) 9.96553 +time/total (s) 4139.3 +Epoch -597 +---------------------------------- --------------- +2022-05-10 14:19:49.637498 PDT | [2] Epoch -596 finished +---------------------------------- --------------- +epoch -596 +replay_buffer/size 999033 +trainer/num train calls 405000 +trainer/Policy Loss -19.5278 +trainer/Log Pis Mean 24.6985 +trainer/Log Pis Std 14.0739 +trainer/Log Pis Max 77.2214 +trainer/Log Pis Min -5.63128 +trainer/policy/mean Mean -0.0414891 +trainer/policy/mean Std 0.909213 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.80459 +trainer/policy/normal/std Std 0.617376 +trainer/policy/normal/std Max 5.24417 +trainer/policy/normal/std Min 0.350302 +trainer/policy/normal/log_std Mean 0.997371 +trainer/policy/normal/log_std Std 0.290592 +trainer/policy/normal/log_std Max 1.65712 +trainer/policy/normal/log_std Min -1.04896 +eval/num steps total 404470 +eval/num paths total 405 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0473203 +eval/Actions Std 0.907339 +eval/Actions Max 0.999998 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56314 +time/logging (s) 0.00416772 +time/sampling batch (s) 0.28214 +time/saving (s) 0.00406118 +time/training (s) 6.89685 +time/epoch (s) 9.75036 +time/total (s) 4149.06 +Epoch -596 +---------------------------------- --------------- +2022-05-10 14:19:59.783274 PDT | [2] Epoch -595 finished +---------------------------------- --------------- +epoch -595 +replay_buffer/size 999033 +trainer/num train calls 406000 +trainer/Policy Loss -19.7803 +trainer/Log Pis Mean 24.8684 +trainer/Log Pis Std 13.0143 +trainer/Log Pis Max 68.6635 +trainer/Log Pis Min -4.7395 +trainer/policy/mean Mean -0.0177665 +trainer/policy/mean Std 0.907573 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.85838 +trainer/policy/normal/std Std 0.644398 +trainer/policy/normal/std Max 6.80083 +trainer/policy/normal/std Min 0.377778 +trainer/policy/normal/log_std Mean 1.01477 +trainer/policy/normal/log_std Std 0.298081 +trainer/policy/normal/log_std Max 1.91704 +trainer/policy/normal/log_std Min -0.97345 +eval/num steps total 405470 +eval/num paths total 406 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0562476 +eval/Actions Std 0.904611 +eval/Actions Max 0.999996 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54747 +time/logging (s) 0.00430291 +time/sampling batch (s) 0.287109 +time/saving (s) 0.00393659 +time/training (s) 7.28042 +time/epoch (s) 10.1232 +time/total (s) 4159.18 +Epoch -595 +---------------------------------- --------------- +2022-05-10 14:20:09.213466 PDT | [2] Epoch -594 finished +---------------------------------- --------------- +epoch -594 +replay_buffer/size 999033 +trainer/num train calls 407000 +trainer/Policy Loss -19.5898 +trainer/Log Pis Mean 24.6901 +trainer/Log Pis Std 13.1562 +trainer/Log Pis Max 66.6563 +trainer/Log Pis Min -9.8458 +trainer/policy/mean Mean -0.0272474 +trainer/policy/mean Std 0.90676 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.80807 +trainer/policy/normal/std Std 0.628292 +trainer/policy/normal/std Max 5.51631 +trainer/policy/normal/std Min 0.347231 +trainer/policy/normal/log_std Mean 0.998547 +trainer/policy/normal/log_std Std 0.288413 +trainer/policy/normal/log_std Max 1.70771 +trainer/policy/normal/log_std Min -1.05777 +eval/num steps total 406470 +eval/num paths total 407 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.118655 +eval/Actions Std 0.910131 +eval/Actions Max 0.999986 +eval/Actions Min -0.999982 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64076 +time/logging (s) 0.00376424 +time/sampling batch (s) 0.280146 +time/saving (s) 0.00350089 +time/training (s) 6.47893 +time/epoch (s) 9.4071 +time/total (s) 4168.59 +Epoch -594 +---------------------------------- --------------- +2022-05-10 14:20:18.120085 PDT | [2] Epoch -593 finished +---------------------------------- --------------- +epoch -593 +replay_buffer/size 999033 +trainer/num train calls 408000 +trainer/Policy Loss -18.8098 +trainer/Log Pis Mean 24.4684 +trainer/Log Pis Std 13.2201 +trainer/Log Pis Max 67.4931 +trainer/Log Pis Min -10.9297 +trainer/policy/mean Mean -0.0530577 +trainer/policy/mean Std 0.904793 +trainer/policy/mean Max 1 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.84244 +trainer/policy/normal/std Std 0.619951 +trainer/policy/normal/std Max 6.11528 +trainer/policy/normal/std Min 0.329234 +trainer/policy/normal/log_std Mean 1.01183 +trainer/policy/normal/log_std Std 0.28499 +trainer/policy/normal/log_std Max 1.81079 +trainer/policy/normal/log_std Min -1.11099 +eval/num steps total 407470 +eval/num paths total 408 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0325643 +eval/Actions Std 0.88509 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56406 +time/logging (s) 0.00407544 +time/sampling batch (s) 0.27991 +time/saving (s) 0.00398712 +time/training (s) 6.03314 +time/epoch (s) 8.88518 +time/total (s) 4177.48 +Epoch -593 +---------------------------------- --------------- +2022-05-10 14:20:27.359524 PDT | [2] Epoch -592 finished +---------------------------------- --------------- +epoch -592 +replay_buffer/size 999033 +trainer/num train calls 409000 +trainer/Policy Loss -19.2498 +trainer/Log Pis Mean 24.9145 +trainer/Log Pis Std 12.9358 +trainer/Log Pis Max 63.3993 +trainer/Log Pis Min -8.20078 +trainer/policy/mean Mean -0.0372237 +trainer/policy/mean Std 0.908762 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.77641 +trainer/policy/normal/std Std 0.634685 +trainer/policy/normal/std Max 6.17549 +trainer/policy/normal/std Min 0.340927 +trainer/policy/normal/log_std Mean 0.983922 +trainer/policy/normal/log_std Std 0.306654 +trainer/policy/normal/log_std Max 1.82059 +trainer/policy/normal/log_std Min -1.07609 +eval/num steps total 408470 +eval/num paths total 409 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.157246 +eval/Actions Std 0.899398 +eval/Actions Max 0.999983 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.37128 +time/logging (s) 0.00376302 +time/sampling batch (s) 0.530512 +time/saving (s) 0.00369766 +time/training (s) 6.30745 +time/epoch (s) 9.2167 +time/total (s) 4186.7 +Epoch -592 +---------------------------------- --------------- +2022-05-10 14:20:38.652761 PDT | [2] Epoch -591 finished +---------------------------------- --------------- +epoch -591 +replay_buffer/size 999033 +trainer/num train calls 410000 +trainer/Policy Loss -19.5668 +trainer/Log Pis Mean 25.2657 +trainer/Log Pis Std 12.9884 +trainer/Log Pis Max 68.2745 +trainer/Log Pis Min -3.81961 +trainer/policy/mean Mean -0.0400399 +trainer/policy/mean Std 0.908918 +trainer/policy/mean Max 0.999982 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.86973 +trainer/policy/normal/std Std 0.62664 +trainer/policy/normal/std Max 5.42515 +trainer/policy/normal/std Min 0.315195 +trainer/policy/normal/log_std Mean 1.02013 +trainer/policy/normal/log_std Std 0.293222 +trainer/policy/normal/log_std Max 1.69104 +trainer/policy/normal/log_std Min -1.15456 +eval/num steps total 409470 +eval/num paths total 410 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0461633 +eval/Actions Std 0.893503 +eval/Actions Max 0.999998 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57495 +time/logging (s) 0.00380745 +time/sampling batch (s) 0.541079 +time/saving (s) 0.00347308 +time/training (s) 8.14763 +time/epoch (s) 11.2709 +time/total (s) 4197.97 +Epoch -591 +---------------------------------- --------------- +2022-05-10 14:20:48.550410 PDT | [2] Epoch -590 finished +---------------------------------- --------------- +epoch -590 +replay_buffer/size 999033 +trainer/num train calls 411000 +trainer/Policy Loss -19.2645 +trainer/Log Pis Mean 24.961 +trainer/Log Pis Std 13.5514 +trainer/Log Pis Max 71.7964 +trainer/Log Pis Min -6.9662 +trainer/policy/mean Mean -0.046831 +trainer/policy/mean Std 0.908942 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.83369 +trainer/policy/normal/std Std 0.634821 +trainer/policy/normal/std Max 6.59702 +trainer/policy/normal/std Min 0.327582 +trainer/policy/normal/log_std Mean 1.0059 +trainer/policy/normal/log_std Std 0.30002 +trainer/policy/normal/log_std Max 1.88662 +trainer/policy/normal/log_std Min -1.11602 +eval/num steps total 410470 +eval/num paths total 411 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.243002 +eval/Actions Std 0.897887 +eval/Actions Max 0.999995 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.78327 +time/logging (s) 0.00368673 +time/sampling batch (s) 0.28303 +time/saving (s) 0.00345669 +time/training (s) 6.80215 +time/epoch (s) 9.8756 +time/total (s) 4207.85 +Epoch -590 +---------------------------------- --------------- +2022-05-10 14:20:58.397757 PDT | [2] Epoch -589 finished +---------------------------------- --------------- +epoch -589 +replay_buffer/size 999033 +trainer/num train calls 412000 +trainer/Policy Loss -18.3218 +trainer/Log Pis Mean 24.0403 +trainer/Log Pis Std 13.3169 +trainer/Log Pis Max 77.5848 +trainer/Log Pis Min -10.3821 +trainer/policy/mean Mean -0.0286557 +trainer/policy/mean Std 0.901376 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.82456 +trainer/policy/normal/std Std 0.62355 +trainer/policy/normal/std Max 6.02554 +trainer/policy/normal/std Min 0.3867 +trainer/policy/normal/log_std Mean 1.00404 +trainer/policy/normal/log_std Std 0.293239 +trainer/policy/normal/log_std Max 1.79601 +trainer/policy/normal/log_std Min -0.950107 +eval/num steps total 411470 +eval/num paths total 412 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.286782 +eval/Actions Std 0.849415 +eval/Actions Max 0.999987 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.35911 +time/logging (s) 0.00371898 +time/sampling batch (s) 0.290728 +time/saving (s) 0.00351658 +time/training (s) 7.16811 +time/epoch (s) 9.82518 +time/total (s) 4217.68 +Epoch -589 +---------------------------------- --------------- +2022-05-10 14:21:08.423350 PDT | [2] Epoch -588 finished +---------------------------------- --------------- +epoch -588 +replay_buffer/size 999033 +trainer/num train calls 413000 +trainer/Policy Loss -19.4905 +trainer/Log Pis Mean 23.9938 +trainer/Log Pis Std 12.5387 +trainer/Log Pis Max 70.8308 +trainer/Log Pis Min -9.77498 +trainer/policy/mean Mean -0.0313664 +trainer/policy/mean Std 0.9071 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.82361 +trainer/policy/normal/std Std 0.617169 +trainer/policy/normal/std Max 5.13681 +trainer/policy/normal/std Min 0.297305 +trainer/policy/normal/log_std Mean 1.00435 +trainer/policy/normal/log_std Std 0.290476 +trainer/policy/normal/log_std Max 1.63643 +trainer/policy/normal/log_std Min -1.213 +eval/num steps total 412470 +eval/num paths total 413 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0889932 +eval/Actions Std 0.914306 +eval/Actions Max 0.999989 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.4613 +time/logging (s) 0.00414431 +time/sampling batch (s) 0.283423 +time/saving (s) 0.00411607 +time/training (s) 7.25098 +time/epoch (s) 10.004 +time/total (s) 4227.69 +Epoch -588 +---------------------------------- --------------- +2022-05-10 14:21:18.473050 PDT | [2] Epoch -587 finished +---------------------------------- --------------- +epoch -587 +replay_buffer/size 999033 +trainer/num train calls 414000 +trainer/Policy Loss -18.8884 +trainer/Log Pis Mean 23.2189 +trainer/Log Pis Std 12.8477 +trainer/Log Pis Max 61.9144 +trainer/Log Pis Min -8.06233 +trainer/policy/mean Mean -0.0283353 +trainer/policy/mean Std 0.904475 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.79288 +trainer/policy/normal/std Std 0.647852 +trainer/policy/normal/std Max 5.99951 +trainer/policy/normal/std Min 0.302848 +trainer/policy/normal/log_std Mean 0.988613 +trainer/policy/normal/log_std Std 0.312048 +trainer/policy/normal/log_std Max 1.79168 +trainer/policy/normal/log_std Min -1.19452 +eval/num steps total 413470 +eval/num paths total 414 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0831271 +eval/Actions Std 0.919359 +eval/Actions Max 0.999998 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62933 +time/logging (s) 0.00371142 +time/sampling batch (s) 0.282959 +time/saving (s) 0.00347559 +time/training (s) 7.10732 +time/epoch (s) 10.0268 +time/total (s) 4237.72 +Epoch -587 +---------------------------------- --------------- +2022-05-10 14:21:27.850380 PDT | [2] Epoch -586 finished +---------------------------------- --------------- +epoch -586 +replay_buffer/size 999033 +trainer/num train calls 415000 +trainer/Policy Loss -19.8644 +trainer/Log Pis Mean 24.6924 +trainer/Log Pis Std 12.6145 +trainer/Log Pis Max 61.4788 +trainer/Log Pis Min -8.40388 +trainer/policy/mean Mean -0.0381329 +trainer/policy/mean Std 0.908107 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.8438 +trainer/policy/normal/std Std 0.618333 +trainer/policy/normal/std Max 6.28337 +trainer/policy/normal/std Min 0.39943 +trainer/policy/normal/log_std Mean 1.01294 +trainer/policy/normal/log_std Std 0.281869 +trainer/policy/normal/log_std Max 1.83791 +trainer/policy/normal/log_std Min -0.917717 +eval/num steps total 414470 +eval/num paths total 415 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.240878 +eval/Actions Std 0.886189 +eval/Actions Max 0.999995 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66031 +time/logging (s) 0.00370782 +time/sampling batch (s) 0.283653 +time/saving (s) 0.00341765 +time/training (s) 6.40409 +time/epoch (s) 9.35517 +time/total (s) 4247.08 +Epoch -586 +---------------------------------- --------------- +2022-05-10 14:21:37.214683 PDT | [2] Epoch -585 finished +---------------------------------- --------------- +epoch -585 +replay_buffer/size 999033 +trainer/num train calls 416000 +trainer/Policy Loss -18.954 +trainer/Log Pis Mean 24.4376 +trainer/Log Pis Std 13.4306 +trainer/Log Pis Max 73.82 +trainer/Log Pis Min -5.7661 +trainer/policy/mean Mean -0.0212564 +trainer/policy/mean Std 0.911 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.87163 +trainer/policy/normal/std Std 0.635915 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.279412 +trainer/policy/normal/log_std Mean 1.02088 +trainer/policy/normal/log_std Std 0.292303 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.27507 +eval/num steps total 415470 +eval/num paths total 416 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0792579 +eval/Actions Std 0.880876 +eval/Actions Max 0.999977 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73114 +time/logging (s) 0.0037727 +time/sampling batch (s) 0.28227 +time/saving (s) 0.00339668 +time/training (s) 6.32182 +time/epoch (s) 9.34239 +time/total (s) 4256.42 +Epoch -585 +---------------------------------- --------------- +2022-05-10 14:21:48.452664 PDT | [2] Epoch -584 finished +---------------------------------- --------------- +epoch -584 +replay_buffer/size 999033 +trainer/num train calls 417000 +trainer/Policy Loss -19.434 +trainer/Log Pis Mean 25.7515 +trainer/Log Pis Std 13.057 +trainer/Log Pis Max 74.4569 +trainer/Log Pis Min -5.19939 +trainer/policy/mean Mean -0.0414502 +trainer/policy/mean Std 0.908955 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.83376 +trainer/policy/normal/std Std 0.641232 +trainer/policy/normal/std Max 5.24439 +trainer/policy/normal/std Min 0.368304 +trainer/policy/normal/log_std Mean 1.00477 +trainer/policy/normal/log_std Std 0.306106 +trainer/policy/normal/log_std Max 1.65716 +trainer/policy/normal/log_std Min -0.998846 +eval/num steps total 416470 +eval/num paths total 417 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.050532 +eval/Actions Std 0.836619 +eval/Actions Max 0.999994 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62563 +time/logging (s) 0.00371777 +time/sampling batch (s) 0.282057 +time/saving (s) 0.00352681 +time/training (s) 8.30113 +time/epoch (s) 11.2161 +time/total (s) 4267.64 +Epoch -584 +---------------------------------- --------------- +2022-05-10 14:21:57.764089 PDT | [2] Epoch -583 finished +---------------------------------- --------------- +epoch -583 +replay_buffer/size 999033 +trainer/num train calls 418000 +trainer/Policy Loss -18.4493 +trainer/Log Pis Mean 24.3284 +trainer/Log Pis Std 12.5906 +trainer/Log Pis Max 61.9636 +trainer/Log Pis Min -10.7455 +trainer/policy/mean Mean -0.0378918 +trainer/policy/mean Std 0.909502 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999982 +trainer/policy/normal/std Mean 2.86627 +trainer/policy/normal/std Std 0.643461 +trainer/policy/normal/std Max 5.9741 +trainer/policy/normal/std Min 0.277077 +trainer/policy/normal/log_std Mean 1.01648 +trainer/policy/normal/log_std Std 0.306104 +trainer/policy/normal/log_std Max 1.78743 +trainer/policy/normal/log_std Min -1.28346 +eval/num steps total 417470 +eval/num paths total 418 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.120846 +eval/Actions Std 0.89403 +eval/Actions Max 0.999996 +eval/Actions Min -0.999985 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51378 +time/logging (s) 0.00415742 +time/sampling batch (s) 0.28331 +time/saving (s) 0.00397375 +time/training (s) 6.4847 +time/epoch (s) 9.28992 +time/total (s) 4276.93 +Epoch -583 +---------------------------------- --------------- +2022-05-10 14:22:07.674293 PDT | [2] Epoch -582 finished +---------------------------------- --------------- +epoch -582 +replay_buffer/size 999033 +trainer/num train calls 419000 +trainer/Policy Loss -19.0053 +trainer/Log Pis Mean 25.6869 +trainer/Log Pis Std 13.6575 +trainer/Log Pis Max 69.8887 +trainer/Log Pis Min -5.91205 +trainer/policy/mean Mean -0.0410711 +trainer/policy/mean Std 0.907676 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.82702 +trainer/policy/normal/std Std 0.624684 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.268235 +trainer/policy/normal/log_std Mean 1.00456 +trainer/policy/normal/log_std Std 0.295619 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.31589 +eval/num steps total 418470 +eval/num paths total 419 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.105531 +eval/Actions Std 0.91265 +eval/Actions Max 0.999997 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65538 +time/logging (s) 0.00382887 +time/sampling batch (s) 0.280972 +time/saving (s) 0.00374187 +time/training (s) 6.94379 +time/epoch (s) 9.88771 +time/total (s) 4286.82 +Epoch -582 +---------------------------------- --------------- +2022-05-10 14:22:17.199498 PDT | [2] Epoch -581 finished +---------------------------------- --------------- +epoch -581 +replay_buffer/size 999033 +trainer/num train calls 420000 +trainer/Policy Loss -20.8347 +trainer/Log Pis Mean 24.1166 +trainer/Log Pis Std 13.8227 +trainer/Log Pis Max 68.2776 +trainer/Log Pis Min -9.23733 +trainer/policy/mean Mean -0.0353403 +trainer/policy/mean Std 0.906484 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.82656 +trainer/policy/normal/std Std 0.650044 +trainer/policy/normal/std Max 5.28879 +trainer/policy/normal/std Min 0.268315 +trainer/policy/normal/log_std Mean 1.0011 +trainer/policy/normal/log_std Std 0.310134 +trainer/policy/normal/log_std Max 1.66559 +trainer/policy/normal/log_std Min -1.31559 +eval/num steps total 419470 +eval/num paths total 420 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00810365 +eval/Actions Std 0.906561 +eval/Actions Max 0.999995 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.38727 +time/logging (s) 0.00375905 +time/sampling batch (s) 0.280739 +time/saving (s) 0.00342708 +time/training (s) 6.82797 +time/epoch (s) 9.50316 +time/total (s) 4296.33 +Epoch -581 +---------------------------------- --------------- +2022-05-10 14:22:26.642323 PDT | [2] Epoch -580 finished +---------------------------------- --------------- +epoch -580 +replay_buffer/size 999033 +trainer/num train calls 421000 +trainer/Policy Loss -19.6381 +trainer/Log Pis Mean 24.2782 +trainer/Log Pis Std 12.9451 +trainer/Log Pis Max 60.1414 +trainer/Log Pis Min -5.93665 +trainer/policy/mean Mean -0.0358352 +trainer/policy/mean Std 0.907867 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.8069 +trainer/policy/normal/std Std 0.617667 +trainer/policy/normal/std Max 6.94016 +trainer/policy/normal/std Min 0.351446 +trainer/policy/normal/log_std Mean 0.998294 +trainer/policy/normal/log_std Std 0.290671 +trainer/policy/normal/log_std Max 1.93733 +trainer/policy/normal/log_std Min -1.0457 +eval/num steps total 420470 +eval/num paths total 421 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.293629 +eval/Actions Std 0.890276 +eval/Actions Max 0.999996 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.7129 +time/logging (s) 0.00367126 +time/sampling batch (s) 0.279782 +time/saving (s) 0.00338365 +time/training (s) 6.42143 +time/epoch (s) 9.42116 +time/total (s) 4305.76 +Epoch -580 +---------------------------------- --------------- +2022-05-10 14:22:36.588990 PDT | [2] Epoch -579 finished +---------------------------------- --------------- +epoch -579 +replay_buffer/size 999033 +trainer/num train calls 422000 +trainer/Policy Loss -18.6675 +trainer/Log Pis Mean 25.2007 +trainer/Log Pis Std 13.6915 +trainer/Log Pis Max 80.5978 +trainer/Log Pis Min -7.14072 +trainer/policy/mean Mean -0.0457664 +trainer/policy/mean Std 0.909678 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.80912 +trainer/policy/normal/std Std 0.642547 +trainer/policy/normal/std Max 6.00744 +trainer/policy/normal/std Min 0.292119 +trainer/policy/normal/log_std Mean 0.996376 +trainer/policy/normal/log_std Std 0.302536 +trainer/policy/normal/log_std Max 1.793 +trainer/policy/normal/log_std Min -1.23059 +eval/num steps total 421470 +eval/num paths total 422 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.156542 +eval/Actions Std 0.911364 +eval/Actions Max 0.999997 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.45208 +time/logging (s) 0.00379899 +time/sampling batch (s) 0.52937 +time/saving (s) 0.00342916 +time/training (s) 6.93655 +time/epoch (s) 9.92522 +time/total (s) 4315.68 +Epoch -579 +---------------------------------- --------------- +2022-05-10 14:22:45.797588 PDT | [2] Epoch -578 finished +---------------------------------- --------------- +epoch -578 +replay_buffer/size 999033 +trainer/num train calls 423000 +trainer/Policy Loss -18.9766 +trainer/Log Pis Mean 25.0767 +trainer/Log Pis Std 13.3271 +trainer/Log Pis Max 67.6262 +trainer/Log Pis Min -8.4195 +trainer/policy/mean Mean -0.025669 +trainer/policy/mean Std 0.908332 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.87242 +trainer/policy/normal/std Std 0.647498 +trainer/policy/normal/std Max 6.5041 +trainer/policy/normal/std Min 0.284761 +trainer/policy/normal/log_std Mean 1.01873 +trainer/policy/normal/log_std Std 0.303978 +trainer/policy/normal/log_std Max 1.87243 +trainer/policy/normal/log_std Min -1.25611 +eval/num steps total 422470 +eval/num paths total 423 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.037955 +eval/Actions Std 0.902783 +eval/Actions Max 1 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.48008 +time/logging (s) 0.00375441 +time/sampling batch (s) 0.275307 +time/saving (s) 0.00337401 +time/training (s) 6.42461 +time/epoch (s) 9.18712 +time/total (s) 4324.87 +Epoch -578 +---------------------------------- --------------- +2022-05-10 14:22:55.208625 PDT | [2] Epoch -577 finished +---------------------------------- --------------- +epoch -577 +replay_buffer/size 999033 +trainer/num train calls 424000 +trainer/Policy Loss -20.3824 +trainer/Log Pis Mean 25.5211 +trainer/Log Pis Std 13.4012 +trainer/Log Pis Max 67.6798 +trainer/Log Pis Min -9.30116 +trainer/policy/mean Mean -0.0346731 +trainer/policy/mean Std 0.911242 +trainer/policy/mean Max 0.999982 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.87088 +trainer/policy/normal/std Std 0.649391 +trainer/policy/normal/std Max 6.41341 +trainer/policy/normal/std Min 0.343243 +trainer/policy/normal/log_std Mean 1.01905 +trainer/policy/normal/log_std Std 0.297755 +trainer/policy/normal/log_std Max 1.85839 +trainer/policy/normal/log_std Min -1.06932 +eval/num steps total 423470 +eval/num paths total 424 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.14213 +eval/Actions Std 0.929416 +eval/Actions Max 0.99999 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65971 +time/logging (s) 0.00408078 +time/sampling batch (s) 0.278976 +time/saving (s) 0.00398932 +time/training (s) 6.44285 +time/epoch (s) 9.3896 +time/total (s) 4334.27 +Epoch -577 +---------------------------------- --------------- +2022-05-10 14:23:05.848428 PDT | [2] Epoch -576 finished +---------------------------------- --------------- +epoch -576 +replay_buffer/size 999033 +trainer/num train calls 425000 +trainer/Policy Loss -20.6378 +trainer/Log Pis Mean 24.4678 +trainer/Log Pis Std 12.3131 +trainer/Log Pis Max 68.3553 +trainer/Log Pis Min -8.58266 +trainer/policy/mean Mean -0.0319087 +trainer/policy/mean Std 0.909099 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.82092 +trainer/policy/normal/std Std 0.632603 +trainer/policy/normal/std Max 6.02876 +trainer/policy/normal/std Min 0.331083 +trainer/policy/normal/log_std Mean 1.00086 +trainer/policy/normal/log_std Std 0.303517 +trainer/policy/normal/log_std Max 1.79654 +trainer/policy/normal/log_std Min -1.10538 +eval/num steps total 424470 +eval/num paths total 425 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.280254 +eval/Actions Std 0.909166 +eval/Actions Max 0.999995 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64783 +time/logging (s) 0.00383936 +time/sampling batch (s) 0.527431 +time/saving (s) 0.00351687 +time/training (s) 7.43501 +time/epoch (s) 10.6176 +time/total (s) 4344.89 +Epoch -576 +---------------------------------- --------------- +2022-05-10 14:23:16.382293 PDT | [2] Epoch -575 finished +---------------------------------- --------------- +epoch -575 +replay_buffer/size 999033 +trainer/num train calls 426000 +trainer/Policy Loss -19.2172 +trainer/Log Pis Mean 23.0615 +trainer/Log Pis Std 13.359 +trainer/Log Pis Max 60.5237 +trainer/Log Pis Min -6.82991 +trainer/policy/mean Mean -0.0379839 +trainer/policy/mean Std 0.900714 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.86311 +trainer/policy/normal/std Std 0.640621 +trainer/policy/normal/std Max 5.95016 +trainer/policy/normal/std Min 0.338863 +trainer/policy/normal/log_std Mean 1.01707 +trainer/policy/normal/log_std Std 0.29449 +trainer/policy/normal/log_std Max 1.78342 +trainer/policy/normal/log_std Min -1.08216 +eval/num steps total 425470 +eval/num paths total 426 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.181306 +eval/Actions Std 0.845603 +eval/Actions Max 0.999995 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69713 +time/logging (s) 0.00377714 +time/sampling batch (s) 0.780535 +time/saving (s) 0.00340077 +time/training (s) 7.02735 +time/epoch (s) 10.5122 +time/total (s) 4355.4 +Epoch -575 +---------------------------------- --------------- +2022-05-10 14:23:26.121793 PDT | [2] Epoch -574 finished +---------------------------------- --------------- +epoch -574 +replay_buffer/size 999033 +trainer/num train calls 427000 +trainer/Policy Loss -20.3227 +trainer/Log Pis Mean 25.9476 +trainer/Log Pis Std 13.3936 +trainer/Log Pis Max 72.8155 +trainer/Log Pis Min -8.18011 +trainer/policy/mean Mean -0.0576693 +trainer/policy/mean Std 0.906526 +trainer/policy/mean Max 0.999973 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.8172 +trainer/policy/normal/std Std 0.630219 +trainer/policy/normal/std Max 6.66042 +trainer/policy/normal/std Min 0.267452 +trainer/policy/normal/log_std Mean 0.999907 +trainer/policy/normal/log_std Std 0.301265 +trainer/policy/normal/log_std Max 1.89618 +trainer/policy/normal/log_std Min -1.31881 +eval/num steps total 426470 +eval/num paths total 427 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.101642 +eval/Actions Std 0.87892 +eval/Actions Max 0.99999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.74481 +time/logging (s) 0.00379036 +time/sampling batch (s) 0.282738 +time/saving (s) 0.00340663 +time/training (s) 6.68295 +time/epoch (s) 9.71769 +time/total (s) 4365.12 +Epoch -574 +---------------------------------- --------------- +2022-05-10 14:23:36.731639 PDT | [2] Epoch -573 finished +---------------------------------- --------------- +epoch -573 +replay_buffer/size 999033 +trainer/num train calls 428000 +trainer/Policy Loss -18.7092 +trainer/Log Pis Mean 24.3089 +trainer/Log Pis Std 12.4545 +trainer/Log Pis Max 59.9785 +trainer/Log Pis Min -6.75506 +trainer/policy/mean Mean -0.036027 +trainer/policy/mean Std 0.9032 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.77728 +trainer/policy/normal/std Std 0.629492 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.27177 +trainer/policy/normal/log_std Mean 0.984469 +trainer/policy/normal/log_std Std 0.307422 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.3028 +eval/num steps total 427470 +eval/num paths total 428 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00562894 +eval/Actions Std 0.979168 +eval/Actions Max 0.999994 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.87413 +time/logging (s) 0.00370717 +time/sampling batch (s) 0.279558 +time/saving (s) 0.0033952 +time/training (s) 7.42665 +time/epoch (s) 10.5874 +time/total (s) 4375.71 +Epoch -573 +---------------------------------- --------------- +2022-05-10 14:23:47.040648 PDT | [2] Epoch -572 finished +---------------------------------- --------------- +epoch -572 +replay_buffer/size 999033 +trainer/num train calls 429000 +trainer/Policy Loss -19.2069 +trainer/Log Pis Mean 25.0645 +trainer/Log Pis Std 13.7125 +trainer/Log Pis Max 69.1659 +trainer/Log Pis Min -6.14475 +trainer/policy/mean Mean -0.024323 +trainer/policy/mean Std 0.90529 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.89557 +trainer/policy/normal/std Std 0.665971 +trainer/policy/normal/std Max 5.84932 +trainer/policy/normal/std Min 0.337721 +trainer/policy/normal/log_std Mean 1.02554 +trainer/policy/normal/log_std Std 0.308671 +trainer/policy/normal/log_std Max 1.76632 +trainer/policy/normal/log_std Min -1.08554 +eval/num steps total 428470 +eval/num paths total 429 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0186919 +eval/Actions Std 0.837149 +eval/Actions Max 0.999992 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.49017 +time/logging (s) 0.0039775 +time/sampling batch (s) 0.781004 +time/saving (s) 0.00373731 +time/training (s) 7.00843 +time/epoch (s) 10.2873 +time/total (s) 4386 +Epoch -572 +---------------------------------- --------------- +2022-05-10 14:23:57.210531 PDT | [2] Epoch -571 finished +---------------------------------- --------------- +epoch -571 +replay_buffer/size 999033 +trainer/num train calls 430000 +trainer/Policy Loss -18.6907 +trainer/Log Pis Mean 24.519 +trainer/Log Pis Std 14.052 +trainer/Log Pis Max 68.4538 +trainer/Log Pis Min -9.25717 +trainer/policy/mean Mean -0.0215537 +trainer/policy/mean Std 0.905266 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.8237 +trainer/policy/normal/std Std 0.65088 +trainer/policy/normal/std Max 5.14274 +trainer/policy/normal/std Min 0.316735 +trainer/policy/normal/log_std Mean 1.00003 +trainer/policy/normal/log_std Std 0.309772 +trainer/policy/normal/log_std Max 1.63759 +trainer/policy/normal/log_std Min -1.14969 +eval/num steps total 429470 +eval/num paths total 430 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0862698 +eval/Actions Std 0.908708 +eval/Actions Max 0.999995 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.74529 +time/logging (s) 0.00394735 +time/sampling batch (s) 0.530717 +time/saving (s) 0.00378366 +time/training (s) 6.86392 +time/epoch (s) 10.1477 +time/total (s) 4396.15 +Epoch -571 +---------------------------------- --------------- +2022-05-10 14:24:06.652896 PDT | [2] Epoch -570 finished +---------------------------------- --------------- +epoch -570 +replay_buffer/size 999033 +trainer/num train calls 431000 +trainer/Policy Loss -19.5179 +trainer/Log Pis Mean 24.2492 +trainer/Log Pis Std 13.3817 +trainer/Log Pis Max 69.425 +trainer/Log Pis Min -6.2708 +trainer/policy/mean Mean -0.0285993 +trainer/policy/mean Std 0.907115 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.78298 +trainer/policy/normal/std Std 0.626218 +trainer/policy/normal/std Max 7.35682 +trainer/policy/normal/std Min 0.316335 +trainer/policy/normal/log_std Mean 0.987974 +trainer/policy/normal/log_std Std 0.29976 +trainer/policy/normal/log_std Max 1.99563 +trainer/policy/normal/log_std Min -1.15095 +eval/num steps total 430470 +eval/num paths total 431 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.36399 +eval/Actions Std 0.882795 +eval/Actions Max 0.999999 +eval/Actions Min -0.999986 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.75074 +time/logging (s) 0.00386545 +time/sampling batch (s) 0.281329 +time/saving (s) 0.00348682 +time/training (s) 6.38064 +time/epoch (s) 9.42006 +time/total (s) 4405.58 +Epoch -570 +---------------------------------- --------------- +2022-05-10 14:24:17.478392 PDT | [2] Epoch -569 finished +---------------------------------- --------------- +epoch -569 +replay_buffer/size 999033 +trainer/num train calls 432000 +trainer/Policy Loss -20.6674 +trainer/Log Pis Mean 23.4793 +trainer/Log Pis Std 12.5712 +trainer/Log Pis Max 72.7567 +trainer/Log Pis Min -10.9227 +trainer/policy/mean Mean -0.0260606 +trainer/policy/mean Std 0.90614 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.78566 +trainer/policy/normal/std Std 0.643784 +trainer/policy/normal/std Max 5.87051 +trainer/policy/normal/std Min 0.295457 +trainer/policy/normal/log_std Mean 0.986441 +trainer/policy/normal/log_std Std 0.309871 +trainer/policy/normal/log_std Max 1.76994 +trainer/policy/normal/log_std Min -1.21923 +eval/num steps total 431470 +eval/num paths total 432 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.202517 +eval/Actions Std 0.937354 +eval/Actions Max 0.999993 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.71708 +time/logging (s) 0.00380736 +time/sampling batch (s) 0.533976 +time/saving (s) 0.00343137 +time/training (s) 7.54498 +time/epoch (s) 10.8033 +time/total (s) 4416.38 +Epoch -569 +---------------------------------- --------------- +2022-05-10 14:24:27.515521 PDT | [2] Epoch -568 finished +---------------------------------- --------------- +epoch -568 +replay_buffer/size 999033 +trainer/num train calls 433000 +trainer/Policy Loss -19.684 +trainer/Log Pis Mean 25.7133 +trainer/Log Pis Std 13.3333 +trainer/Log Pis Max 66.8674 +trainer/Log Pis Min -6.99183 +trainer/policy/mean Mean -0.0258309 +trainer/policy/mean Std 0.908314 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.84337 +trainer/policy/normal/std Std 0.632981 +trainer/policy/normal/std Max 5.68603 +trainer/policy/normal/std Min 0.306164 +trainer/policy/normal/log_std Mean 1.00927 +trainer/policy/normal/log_std Std 0.301418 +trainer/policy/normal/log_std Max 1.73801 +trainer/policy/normal/log_std Min -1.18363 +eval/num steps total 432470 +eval/num paths total 433 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0808557 +eval/Actions Std 0.84512 +eval/Actions Max 0.999998 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67612 +time/logging (s) 0.00368531 +time/sampling batch (s) 0.293139 +time/saving (s) 0.0035414 +time/training (s) 7.03824 +time/epoch (s) 10.0147 +time/total (s) 4426.4 +Epoch -568 +---------------------------------- --------------- +2022-05-10 14:24:36.790347 PDT | [2] Epoch -567 finished +---------------------------------- --------------- +epoch -567 +replay_buffer/size 999033 +trainer/num train calls 434000 +trainer/Policy Loss -19.4119 +trainer/Log Pis Mean 24.2125 +trainer/Log Pis Std 13.0943 +trainer/Log Pis Max 70.2742 +trainer/Log Pis Min -11.2547 +trainer/policy/mean Mean -0.0147288 +trainer/policy/mean Std 0.900093 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.83056 +trainer/policy/normal/std Std 0.647219 +trainer/policy/normal/std Max 6.04629 +trainer/policy/normal/std Min 0.355781 +trainer/policy/normal/log_std Mean 1.00408 +trainer/policy/normal/log_std Std 0.300591 +trainer/policy/normal/log_std Max 1.79944 +trainer/policy/normal/log_std Min -1.03344 +eval/num steps total 433470 +eval/num paths total 434 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.353132 +eval/Actions Std 0.845226 +eval/Actions Max 0.999995 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.43566 +time/logging (s) 0.00377715 +time/sampling batch (s) 0.280447 +time/saving (s) 0.00343455 +time/training (s) 6.52963 +time/epoch (s) 9.25295 +time/total (s) 4435.66 +Epoch -567 +---------------------------------- --------------- +2022-05-10 14:24:46.100976 PDT | [2] Epoch -566 finished +---------------------------------- --------------- +epoch -566 +replay_buffer/size 999033 +trainer/num train calls 435000 +trainer/Policy Loss -19.503 +trainer/Log Pis Mean 24.3835 +trainer/Log Pis Std 13.248 +trainer/Log Pis Max 74.9375 +trainer/Log Pis Min -9.67673 +trainer/policy/mean Mean -0.0229613 +trainer/policy/mean Std 0.907605 +trainer/policy/mean Max 1 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.84565 +trainer/policy/normal/std Std 0.624426 +trainer/policy/normal/std Max 5.16068 +trainer/policy/normal/std Min 0.358091 +trainer/policy/normal/log_std Mean 1.01059 +trainer/policy/normal/log_std Std 0.29992 +trainer/policy/normal/log_std Max 1.64107 +trainer/policy/normal/log_std Min -1.02697 +eval/num steps total 434470 +eval/num paths total 435 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.105784 +eval/Actions Std 0.915512 +eval/Actions Max 0.999992 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57882 +time/logging (s) 0.00428728 +time/sampling batch (s) 0.280957 +time/saving (s) 0.00429108 +time/training (s) 6.42077 +time/epoch (s) 9.28912 +time/total (s) 4444.95 +Epoch -566 +---------------------------------- --------------- +2022-05-10 14:24:56.043045 PDT | [2] Epoch -565 finished +---------------------------------- --------------- +epoch -565 +replay_buffer/size 999033 +trainer/num train calls 436000 +trainer/Policy Loss -20.5431 +trainer/Log Pis Mean 25.8745 +trainer/Log Pis Std 13.6546 +trainer/Log Pis Max 75.6975 +trainer/Log Pis Min -7.99584 +trainer/policy/mean Mean -0.0462269 +trainer/policy/mean Std 0.912461 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.86271 +trainer/policy/normal/std Std 0.633078 +trainer/policy/normal/std Max 5.64921 +trainer/policy/normal/std Min 0.334274 +trainer/policy/normal/log_std Mean 1.01699 +trainer/policy/normal/log_std Std 0.295371 +trainer/policy/normal/log_std Max 1.73152 +trainer/policy/normal/log_std Min -1.0958 +eval/num steps total 435470 +eval/num paths total 436 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0875848 +eval/Actions Std 0.916387 +eval/Actions Max 0.999987 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66478 +time/logging (s) 0.00367784 +time/sampling batch (s) 0.278225 +time/saving (s) 0.00344836 +time/training (s) 6.96907 +time/epoch (s) 9.9192 +time/total (s) 4454.87 +Epoch -565 +---------------------------------- --------------- +2022-05-10 14:25:06.582611 PDT | [2] Epoch -564 finished +---------------------------------- --------------- +epoch -564 +replay_buffer/size 999033 +trainer/num train calls 437000 +trainer/Policy Loss -19.1219 +trainer/Log Pis Mean 24.0222 +trainer/Log Pis Std 13.205 +trainer/Log Pis Max 67.6169 +trainer/Log Pis Min -6.37227 +trainer/policy/mean Mean -0.0376066 +trainer/policy/mean Std 0.909552 +trainer/policy/mean Max 0.999977 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.75871 +trainer/policy/normal/std Std 0.624329 +trainer/policy/normal/std Max 5.07734 +trainer/policy/normal/std Min 0.32621 +trainer/policy/normal/log_std Mean 0.978111 +trainer/policy/normal/log_std Std 0.303749 +trainer/policy/normal/log_std Max 1.62479 +trainer/policy/normal/log_std Min -1.12021 +eval/num steps total 436470 +eval/num paths total 437 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0546937 +eval/Actions Std 0.914023 +eval/Actions Max 0.999996 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59342 +time/logging (s) 0.0037701 +time/sampling batch (s) 0.280787 +time/saving (s) 0.00339898 +time/training (s) 7.6362 +time/epoch (s) 10.5176 +time/total (s) 4465.39 +Epoch -564 +---------------------------------- --------------- +2022-05-10 14:25:17.270251 PDT | [2] Epoch -563 finished +---------------------------------- --------------- +epoch -563 +replay_buffer/size 999033 +trainer/num train calls 438000 +trainer/Policy Loss -19.675 +trainer/Log Pis Mean 24.9731 +trainer/Log Pis Std 13.055 +trainer/Log Pis Max 63.605 +trainer/Log Pis Min -10.1657 +trainer/policy/mean Mean -0.0751506 +trainer/policy/mean Std 0.90742 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.80251 +trainer/policy/normal/std Std 0.626485 +trainer/policy/normal/std Max 7.18278 +trainer/policy/normal/std Min 0.337232 +trainer/policy/normal/log_std Mean 0.995967 +trainer/policy/normal/log_std Std 0.293276 +trainer/policy/normal/log_std Max 1.97169 +trainer/policy/normal/log_std Min -1.08698 +eval/num steps total 437470 +eval/num paths total 438 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.249555 +eval/Actions Std 0.880316 +eval/Actions Max 0.999998 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59105 +time/logging (s) 0.00390397 +time/sampling batch (s) 0.281045 +time/saving (s) 0.0037747 +time/training (s) 7.78598 +time/epoch (s) 10.6658 +time/total (s) 4476.06 +Epoch -563 +---------------------------------- --------------- +2022-05-10 14:25:27.731408 PDT | [2] Epoch -562 finished +---------------------------------- --------------- +epoch -562 +replay_buffer/size 999033 +trainer/num train calls 439000 +trainer/Policy Loss -20.2101 +trainer/Log Pis Mean 22.877 +trainer/Log Pis Std 12.7612 +trainer/Log Pis Max 66.665 +trainer/Log Pis Min -4.47109 +trainer/policy/mean Mean -0.038451 +trainer/policy/mean Std 0.904758 +trainer/policy/mean Max 0.999974 +trainer/policy/mean Min -0.999979 +trainer/policy/normal/std Mean 2.78342 +trainer/policy/normal/std Std 0.635783 +trainer/policy/normal/std Max 5.55832 +trainer/policy/normal/std Min 0.302546 +trainer/policy/normal/log_std Mean 0.985705 +trainer/policy/normal/log_std Std 0.312307 +trainer/policy/normal/log_std Max 1.7153 +trainer/policy/normal/log_std Min -1.19552 +eval/num steps total 438470 +eval/num paths total 439 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.216988 +eval/Actions Std 0.908184 +eval/Actions Max 0.999999 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66886 +time/logging (s) 0.00376019 +time/sampling batch (s) 0.532868 +time/saving (s) 0.0036968 +time/training (s) 7.22935 +time/epoch (s) 10.4385 +time/total (s) 4486.5 +Epoch -562 +---------------------------------- --------------- +2022-05-10 14:25:38.289862 PDT | [2] Epoch -561 finished +---------------------------------- --------------- +epoch -561 +replay_buffer/size 999033 +trainer/num train calls 440000 +trainer/Policy Loss -19.3242 +trainer/Log Pis Mean 23.2773 +trainer/Log Pis Std 13.3841 +trainer/Log Pis Max 66.5154 +trainer/Log Pis Min -9.87076 +trainer/policy/mean Mean -0.0529487 +trainer/policy/mean Std 0.905176 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.82163 +trainer/policy/normal/std Std 0.640331 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.292806 +trainer/policy/normal/log_std Mean 1.00173 +trainer/policy/normal/log_std Std 0.29745 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.22825 +eval/num steps total 439470 +eval/num paths total 440 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.106002 +eval/Actions Std 0.904939 +eval/Actions Max 0.999998 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59644 +time/logging (s) 0.00377899 +time/sampling batch (s) 0.278732 +time/saving (s) 0.00339689 +time/training (s) 7.65406 +time/epoch (s) 10.5364 +time/total (s) 4497.04 +Epoch -561 +---------------------------------- --------------- +2022-05-10 14:25:48.531966 PDT | [2] Epoch -560 finished +---------------------------------- --------------- +epoch -560 +replay_buffer/size 999033 +trainer/num train calls 441000 +trainer/Policy Loss -19.4468 +trainer/Log Pis Mean 24.6877 +trainer/Log Pis Std 13.4627 +trainer/Log Pis Max 70.9736 +trainer/Log Pis Min -7.31674 +trainer/policy/mean Mean -0.0373098 +trainer/policy/mean Std 0.905582 +trainer/policy/mean Max 1 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.87688 +trainer/policy/normal/std Std 0.634599 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.341987 +trainer/policy/normal/log_std Mean 1.02446 +trainer/policy/normal/log_std Std 0.280017 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.07298 +eval/num steps total 440470 +eval/num paths total 441 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.249855 +eval/Actions Std 0.920659 +eval/Actions Max 0.999999 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70933 +time/logging (s) 0.00387975 +time/sampling batch (s) 0.526347 +time/saving (s) 0.0034663 +time/training (s) 6.97745 +time/epoch (s) 10.2205 +time/total (s) 4507.27 +Epoch -560 +---------------------------------- --------------- +2022-05-10 14:25:58.110551 PDT | [2] Epoch -559 finished +---------------------------------- -------------- +epoch -559 +replay_buffer/size 999033 +trainer/num train calls 442000 +trainer/Policy Loss -19.2081 +trainer/Log Pis Mean 23.3576 +trainer/Log Pis Std 12.8833 +trainer/Log Pis Max 62.4866 +trainer/Log Pis Min -9.88116 +trainer/policy/mean Mean -0.0343441 +trainer/policy/mean Std 0.907612 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.83607 +trainer/policy/normal/std Std 0.630451 +trainer/policy/normal/std Max 6.00187 +trainer/policy/normal/std Min 0.305194 +trainer/policy/normal/log_std Mean 1.00752 +trainer/policy/normal/log_std Std 0.296652 +trainer/policy/normal/log_std Max 1.79207 +trainer/policy/normal/log_std Min -1.18681 +eval/num steps total 441470 +eval/num paths total 442 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.373058 +eval/Actions Std 0.873769 +eval/Actions Max 0.999989 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66297 +time/logging (s) 0.0041905 +time/sampling batch (s) 0.275484 +time/saving (s) 0.0041051 +time/training (s) 6.61044 +time/epoch (s) 9.55719 +time/total (s) 4516.83 +Epoch -559 +---------------------------------- -------------- +2022-05-10 14:26:08.607391 PDT | [2] Epoch -558 finished +---------------------------------- --------------- +epoch -558 +replay_buffer/size 999033 +trainer/num train calls 443000 +trainer/Policy Loss -19.183 +trainer/Log Pis Mean 24.2704 +trainer/Log Pis Std 13.4443 +trainer/Log Pis Max 71.8387 +trainer/Log Pis Min -7.58286 +trainer/policy/mean Mean -0.0273581 +trainer/policy/mean Std 0.90589 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.79531 +trainer/policy/normal/std Std 0.633604 +trainer/policy/normal/std Max 5.83549 +trainer/policy/normal/std Min 0.358392 +trainer/policy/normal/log_std Mean 0.991386 +trainer/policy/normal/log_std Std 0.304299 +trainer/policy/normal/log_std Max 1.76396 +trainer/policy/normal/log_std Min -1.02613 +eval/num steps total 442470 +eval/num paths total 443 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.197661 +eval/Actions Std 0.93846 +eval/Actions Max 0.999986 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.52601 +time/logging (s) 0.00381317 +time/sampling batch (s) 0.277083 +time/saving (s) 0.00356433 +time/training (s) 7.66392 +time/epoch (s) 10.4744 +time/total (s) 4527.3 +Epoch -558 +---------------------------------- --------------- +2022-05-10 14:26:18.976444 PDT | [2] Epoch -557 finished +---------------------------------- --------------- +epoch -557 +replay_buffer/size 999033 +trainer/num train calls 444000 +trainer/Policy Loss -19.7087 +trainer/Log Pis Mean 24.6495 +trainer/Log Pis Std 13.2648 +trainer/Log Pis Max 65.3656 +trainer/Log Pis Min -9.18014 +trainer/policy/mean Mean -0.0519256 +trainer/policy/mean Std 0.904053 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.82016 +trainer/policy/normal/std Std 0.632424 +trainer/policy/normal/std Max 5.73543 +trainer/policy/normal/std Min 0.284271 +trainer/policy/normal/log_std Mean 1.00207 +trainer/policy/normal/log_std Std 0.293554 +trainer/policy/normal/log_std Max 1.74666 +trainer/policy/normal/log_std Min -1.25783 +eval/num steps total 443470 +eval/num paths total 444 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.175242 +eval/Actions Std 0.880569 +eval/Actions Max 0.999998 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.82652 +time/logging (s) 0.00377524 +time/sampling batch (s) 0.276092 +time/saving (s) 0.00337592 +time/training (s) 7.23748 +time/epoch (s) 10.3472 +time/total (s) 4537.65 +Epoch -557 +---------------------------------- --------------- +2022-05-10 14:26:28.992927 PDT | [2] Epoch -556 finished +---------------------------------- --------------- +epoch -556 +replay_buffer/size 999033 +trainer/num train calls 445000 +trainer/Policy Loss -19.7532 +trainer/Log Pis Mean 23.958 +trainer/Log Pis Std 12.3994 +trainer/Log Pis Max 65.4766 +trainer/Log Pis Min -8.6714 +trainer/policy/mean Mean -0.0463814 +trainer/policy/mean Std 0.903938 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.81211 +trainer/policy/normal/std Std 0.64092 +trainer/policy/normal/std Max 5.60271 +trainer/policy/normal/std Min 0.303035 +trainer/policy/normal/log_std Mean 0.997571 +trainer/policy/normal/log_std Std 0.301746 +trainer/policy/normal/log_std Max 1.72325 +trainer/policy/normal/log_std Min -1.19391 +eval/num steps total 444470 +eval/num paths total 445 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0765341 +eval/Actions Std 0.937091 +eval/Actions Max 0.99999 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65227 +time/logging (s) 0.00374811 +time/sampling batch (s) 0.279781 +time/saving (s) 0.00343729 +time/training (s) 7.05534 +time/epoch (s) 9.99457 +time/total (s) 4547.65 +Epoch -556 +---------------------------------- --------------- +2022-05-10 14:26:39.597139 PDT | [2] Epoch -555 finished +---------------------------------- --------------- +epoch -555 +replay_buffer/size 999033 +trainer/num train calls 446000 +trainer/Policy Loss -20.0978 +trainer/Log Pis Mean 24.1765 +trainer/Log Pis Std 13.0429 +trainer/Log Pis Max 67.5302 +trainer/Log Pis Min -9.36161 +trainer/policy/mean Mean -0.0358301 +trainer/policy/mean Std 0.90156 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.75591 +trainer/policy/normal/std Std 0.667143 +trainer/policy/normal/std Max 5.46357 +trainer/policy/normal/std Min 0.308455 +trainer/policy/normal/log_std Mean 0.971448 +trainer/policy/normal/log_std Std 0.328001 +trainer/policy/normal/log_std Max 1.6981 +trainer/policy/normal/log_std Min -1.17618 +eval/num steps total 445470 +eval/num paths total 446 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.117874 +eval/Actions Std 0.927094 +eval/Actions Max 0.999998 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59891 +time/logging (s) 0.00376274 +time/sampling batch (s) 0.526339 +time/saving (s) 0.00334676 +time/training (s) 7.45032 +time/epoch (s) 10.5827 +time/total (s) 4558.24 +Epoch -555 +---------------------------------- --------------- +2022-05-10 14:26:50.109301 PDT | [2] Epoch -554 finished +---------------------------------- --------------- +epoch -554 +replay_buffer/size 999033 +trainer/num train calls 447000 +trainer/Policy Loss -20.1316 +trainer/Log Pis Mean 24.5342 +trainer/Log Pis Std 13.0514 +trainer/Log Pis Max 70.802 +trainer/Log Pis Min -8.65027 +trainer/policy/mean Mean -0.0271563 +trainer/policy/mean Std 0.907811 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999985 +trainer/policy/normal/std Mean 2.83192 +trainer/policy/normal/std Std 0.639683 +trainer/policy/normal/std Max 5.52983 +trainer/policy/normal/std Min 0.275621 +trainer/policy/normal/log_std Mean 1.00347 +trainer/policy/normal/log_std Std 0.310467 +trainer/policy/normal/log_std Max 1.71016 +trainer/policy/normal/log_std Min -1.28873 +eval/num steps total 446470 +eval/num paths total 447 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.150425 +eval/Actions Std 0.855702 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5691 +time/logging (s) 0.00409071 +time/sampling batch (s) 0.277395 +time/saving (s) 0.00416996 +time/training (s) 7.63594 +time/epoch (s) 10.4907 +time/total (s) 4568.73 +Epoch -554 +---------------------------------- --------------- +2022-05-10 14:26:59.991171 PDT | [2] Epoch -553 finished +---------------------------------- --------------- +epoch -553 +replay_buffer/size 999033 +trainer/num train calls 448000 +trainer/Policy Loss -19.0082 +trainer/Log Pis Mean 25.0092 +trainer/Log Pis Std 13.6502 +trainer/Log Pis Max 71.555 +trainer/Log Pis Min -6.98503 +trainer/policy/mean Mean -0.0464413 +trainer/policy/mean Std 0.902559 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.89884 +trainer/policy/normal/std Std 0.6416 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.312261 +trainer/policy/normal/log_std Mean 1.03231 +trainer/policy/normal/log_std Std 0.278937 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.16392 +eval/num steps total 447470 +eval/num paths total 448 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00203702 +eval/Actions Std 0.908847 +eval/Actions Max 1 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51912 +time/logging (s) 0.00379287 +time/sampling batch (s) 0.280428 +time/saving (s) 0.00361601 +time/training (s) 7.05237 +time/epoch (s) 9.85933 +time/total (s) 4578.59 +Epoch -553 +---------------------------------- --------------- +2022-05-10 14:27:10.203003 PDT | [2] Epoch -552 finished +---------------------------------- --------------- +epoch -552 +replay_buffer/size 999033 +trainer/num train calls 449000 +trainer/Policy Loss -18.9702 +trainer/Log Pis Mean 25.8328 +trainer/Log Pis Std 13.3818 +trainer/Log Pis Max 64.2671 +trainer/Log Pis Min -7.71333 +trainer/policy/mean Mean -0.0401836 +trainer/policy/mean Std 0.908456 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.92171 +trainer/policy/normal/std Std 0.655673 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.344088 +trainer/policy/normal/log_std Mean 1.0377 +trainer/policy/normal/log_std Std 0.292606 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.06686 +eval/num steps total 448470 +eval/num paths total 449 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0982954 +eval/Actions Std 0.92439 +eval/Actions Max 0.99999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65778 +time/logging (s) 0.0037281 +time/sampling batch (s) 0.283074 +time/saving (s) 0.00349376 +time/training (s) 7.24148 +time/epoch (s) 10.1896 +time/total (s) 4588.78 +Epoch -552 +---------------------------------- --------------- +2022-05-10 14:27:20.943752 PDT | [2] Epoch -551 finished +---------------------------------- --------------- +epoch -551 +replay_buffer/size 999033 +trainer/num train calls 450000 +trainer/Policy Loss -18.809 +trainer/Log Pis Mean 23.5791 +trainer/Log Pis Std 13.2628 +trainer/Log Pis Max 64.4833 +trainer/Log Pis Min -9.7943 +trainer/policy/mean Mean -0.0578372 +trainer/policy/mean Std 0.904714 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.80366 +trainer/policy/normal/std Std 0.616427 +trainer/policy/normal/std Max 5.90207 +trainer/policy/normal/std Min 0.380736 +trainer/policy/normal/log_std Mean 0.998558 +trainer/policy/normal/log_std Std 0.281326 +trainer/policy/normal/log_std Max 1.7753 +trainer/policy/normal/log_std Min -0.96565 +eval/num steps total 449470 +eval/num paths total 450 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.296289 +eval/Actions Std 0.839062 +eval/Actions Max 0.999999 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5404 +time/logging (s) 0.00372856 +time/sampling batch (s) 0.28195 +time/saving (s) 0.00336089 +time/training (s) 7.88925 +time/epoch (s) 10.7187 +time/total (s) 4599.51 +Epoch -551 +---------------------------------- --------------- +2022-05-10 14:27:31.413540 PDT | [2] Epoch -550 finished +---------------------------------- --------------- +epoch -550 +replay_buffer/size 999033 +trainer/num train calls 451000 +trainer/Policy Loss -19.3471 +trainer/Log Pis Mean 24.3977 +trainer/Log Pis Std 13.0125 +trainer/Log Pis Max 63.4068 +trainer/Log Pis Min -11.8509 +trainer/policy/mean Mean -0.0289921 +trainer/policy/mean Std 0.902683 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.83853 +trainer/policy/normal/std Std 0.622687 +trainer/policy/normal/std Max 5.43265 +trainer/policy/normal/std Min 0.366623 +trainer/policy/normal/log_std Mean 1.01035 +trainer/policy/normal/log_std Std 0.285581 +trainer/policy/normal/log_std Max 1.69243 +trainer/policy/normal/log_std Min -1.00342 +eval/num steps total 450470 +eval/num paths total 451 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.135302 +eval/Actions Std 0.886775 +eval/Actions Max 0.999996 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58739 +time/logging (s) 0.00367837 +time/sampling batch (s) 0.529752 +time/saving (s) 0.00346357 +time/training (s) 7.32346 +time/epoch (s) 10.4477 +time/total (s) 4609.96 +Epoch -550 +---------------------------------- --------------- +2022-05-10 14:27:41.580669 PDT | [2] Epoch -549 finished +---------------------------------- --------------- +epoch -549 +replay_buffer/size 999033 +trainer/num train calls 452000 +trainer/Policy Loss -19.3118 +trainer/Log Pis Mean 25.156 +trainer/Log Pis Std 13.0952 +trainer/Log Pis Max 73.4678 +trainer/Log Pis Min -7.3333 +trainer/policy/mean Mean -0.0500462 +trainer/policy/mean Std 0.909038 +trainer/policy/mean Max 0.99998 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.80161 +trainer/policy/normal/std Std 0.644514 +trainer/policy/normal/std Max 6.77256 +trainer/policy/normal/std Min 0.252957 +trainer/policy/normal/log_std Mean 0.992974 +trainer/policy/normal/log_std Std 0.306357 +trainer/policy/normal/log_std Max 1.91288 +trainer/policy/normal/log_std Min -1.37454 +eval/num steps total 451470 +eval/num paths total 452 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.173261 +eval/Actions Std 0.930101 +eval/Actions Max 0.999992 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65045 +time/logging (s) 0.00369726 +time/sampling batch (s) 0.281133 +time/saving (s) 0.0033948 +time/training (s) 7.20626 +time/epoch (s) 10.1449 +time/total (s) 4620.1 +Epoch -549 +---------------------------------- --------------- +2022-05-10 14:27:52.737588 PDT | [2] Epoch -548 finished +---------------------------------- --------------- +epoch -548 +replay_buffer/size 999033 +trainer/num train calls 453000 +trainer/Policy Loss -19.708 +trainer/Log Pis Mean 24.1925 +trainer/Log Pis Std 13.2701 +trainer/Log Pis Max 68.8191 +trainer/Log Pis Min -4.31157 +trainer/policy/mean Mean -0.0446628 +trainer/policy/mean Std 0.906247 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.72461 +trainer/policy/normal/std Std 0.64013 +trainer/policy/normal/std Max 6.12108 +trainer/policy/normal/std Min 0.239934 +trainer/policy/normal/log_std Mean 0.960227 +trainer/policy/normal/log_std Std 0.333413 +trainer/policy/normal/log_std Max 1.81174 +trainer/policy/normal/log_std Min -1.42739 +eval/num steps total 452470 +eval/num paths total 453 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0548506 +eval/Actions Std 0.918873 +eval/Actions Max 0.999997 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67519 +time/logging (s) 0.0041668 +time/sampling batch (s) 0.28522 +time/saving (s) 0.00394603 +time/training (s) 8.16664 +time/epoch (s) 11.1352 +time/total (s) 4631.24 +Epoch -548 +---------------------------------- --------------- +2022-05-10 14:28:02.323474 PDT | [2] Epoch -547 finished +---------------------------------- --------------- +epoch -547 +replay_buffer/size 999033 +trainer/num train calls 454000 +trainer/Policy Loss -19.0231 +trainer/Log Pis Mean 24.8038 +trainer/Log Pis Std 13.7065 +trainer/Log Pis Max 80.883 +trainer/Log Pis Min -7.54723 +trainer/policy/mean Mean -0.00729711 +trainer/policy/mean Std 0.910309 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.83305 +trainer/policy/normal/std Std 0.627772 +trainer/policy/normal/std Max 5.59074 +trainer/policy/normal/std Min 0.31501 +trainer/policy/normal/log_std Mean 1.00705 +trainer/policy/normal/log_std Std 0.293066 +trainer/policy/normal/log_std Max 1.72111 +trainer/policy/normal/log_std Min -1.15515 +eval/num steps total 453470 +eval/num paths total 454 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0222644 +eval/Actions Std 0.896207 +eval/Actions Max 0.999992 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.31783 +time/logging (s) 0.00378936 +time/sampling batch (s) 0.280886 +time/saving (s) 0.00371981 +time/training (s) 6.9569 +time/epoch (s) 9.56312 +time/total (s) 4640.81 +Epoch -547 +---------------------------------- --------------- +2022-05-10 14:28:13.826755 PDT | [2] Epoch -546 finished +---------------------------------- --------------- +epoch -546 +replay_buffer/size 999033 +trainer/num train calls 455000 +trainer/Policy Loss -20.9134 +trainer/Log Pis Mean 24.1227 +trainer/Log Pis Std 12.917 +trainer/Log Pis Max 72.1234 +trainer/Log Pis Min -9.85878 +trainer/policy/mean Mean -0.0435282 +trainer/policy/mean Std 0.901739 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.77564 +trainer/policy/normal/std Std 0.67038 +trainer/policy/normal/std Max 6.30871 +trainer/policy/normal/std Min 0.287234 +trainer/policy/normal/log_std Mean 0.977503 +trainer/policy/normal/log_std Std 0.335483 +trainer/policy/normal/log_std Max 1.84193 +trainer/policy/normal/log_std Min -1.24746 +eval/num steps total 454470 +eval/num paths total 455 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.196644 +eval/Actions Std 0.813311 +eval/Actions Max 0.999992 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73096 +time/logging (s) 0.00369361 +time/sampling batch (s) 0.282128 +time/saving (s) 0.00341065 +time/training (s) 8.46026 +time/epoch (s) 11.4804 +time/total (s) 4652.29 +Epoch -546 +---------------------------------- --------------- +2022-05-10 14:28:24.306945 PDT | [2] Epoch -545 finished +---------------------------------- --------------- +epoch -545 +replay_buffer/size 999033 +trainer/num train calls 456000 +trainer/Policy Loss -20.242 +trainer/Log Pis Mean 24.3329 +trainer/Log Pis Std 13.613 +trainer/Log Pis Max 66.33 +trainer/Log Pis Min -6.39152 +trainer/policy/mean Mean -0.0417835 +trainer/policy/mean Std 0.907166 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.81842 +trainer/policy/normal/std Std 0.625447 +trainer/policy/normal/std Max 6.15187 +trainer/policy/normal/std Min 0.278307 +trainer/policy/normal/log_std Mean 1.00231 +trainer/policy/normal/log_std Std 0.2905 +trainer/policy/normal/log_std Max 1.81676 +trainer/policy/normal/log_std Min -1.27903 +eval/num steps total 455470 +eval/num paths total 456 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.216786 +eval/Actions Std 0.950602 +eval/Actions Max 0.999984 +eval/Actions Min -0.99998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.76803 +time/logging (s) 0.00369783 +time/sampling batch (s) 0.530746 +time/saving (s) 0.00346942 +time/training (s) 7.15221 +time/epoch (s) 10.4582 +time/total (s) 4662.75 +Epoch -545 +---------------------------------- --------------- +2022-05-10 14:28:34.281665 PDT | [2] Epoch -544 finished +---------------------------------- --------------- +epoch -544 +replay_buffer/size 999033 +trainer/num train calls 457000 +trainer/Policy Loss -20.7564 +trainer/Log Pis Mean 24.1657 +trainer/Log Pis Std 12.9891 +trainer/Log Pis Max 70.6422 +trainer/Log Pis Min -4.97761 +trainer/policy/mean Mean -0.0431193 +trainer/policy/mean Std 0.909594 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.8072 +trainer/policy/normal/std Std 0.617352 +trainer/policy/normal/std Max 5.98285 +trainer/policy/normal/std Min 0.355984 +trainer/policy/normal/log_std Mean 0.997819 +trainer/policy/normal/log_std Std 0.2946 +trainer/policy/normal/log_std Max 1.7889 +trainer/policy/normal/log_std Min -1.03287 +eval/num steps total 456470 +eval/num paths total 457 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.100553 +eval/Actions Std 0.876361 +eval/Actions Max 0.99999 +eval/Actions Min -0.999985 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67564 +time/logging (s) 0.0041719 +time/sampling batch (s) 0.313277 +time/saving (s) 0.00414622 +time/training (s) 6.9554 +time/epoch (s) 9.95263 +time/total (s) 4672.71 +Epoch -544 +---------------------------------- --------------- +2022-05-10 14:28:44.349229 PDT | [2] Epoch -543 finished +---------------------------------- --------------- +epoch -543 +replay_buffer/size 999033 +trainer/num train calls 458000 +trainer/Policy Loss -19.209 +trainer/Log Pis Mean 25.17 +trainer/Log Pis Std 13.5811 +trainer/Log Pis Max 70.0376 +trainer/Log Pis Min -6.24824 +trainer/policy/mean Mean -0.0282331 +trainer/policy/mean Std 0.910574 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.85765 +trainer/policy/normal/std Std 0.640973 +trainer/policy/normal/std Max 6.78316 +trainer/policy/normal/std Min 0.331923 +trainer/policy/normal/log_std Mean 1.0136 +trainer/policy/normal/log_std Std 0.304263 +trainer/policy/normal/log_std Max 1.91444 +trainer/policy/normal/log_std Min -1.10285 +eval/num steps total 457470 +eval/num paths total 458 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0518816 +eval/Actions Std 0.906271 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65848 +time/logging (s) 0.00414284 +time/sampling batch (s) 0.280369 +time/saving (s) 0.00424747 +time/training (s) 7.09769 +time/epoch (s) 10.0449 +time/total (s) 4682.76 +Epoch -543 +---------------------------------- --------------- +2022-05-10 14:28:55.153327 PDT | [2] Epoch -542 finished +---------------------------------- -------------- +epoch -542 +replay_buffer/size 999033 +trainer/num train calls 459000 +trainer/Policy Loss -19.3807 +trainer/Log Pis Mean 23.586 +trainer/Log Pis Std 13.3602 +trainer/Log Pis Max 63.9579 +trainer/Log Pis Min -11.0245 +trainer/policy/mean Mean -0.0280169 +trainer/policy/mean Std 0.907825 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.76439 +trainer/policy/normal/std Std 0.644567 +trainer/policy/normal/std Max 5.54019 +trainer/policy/normal/std Min 0.247819 +trainer/policy/normal/log_std Mean 0.976026 +trainer/policy/normal/log_std Std 0.326424 +trainer/policy/normal/log_std Max 1.71203 +trainer/policy/normal/log_std Min -1.39506 +eval/num steps total 458470 +eval/num paths total 459 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.119889 +eval/Actions Std 0.893277 +eval/Actions Max 0.999988 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.47481 +time/logging (s) 0.004812 +time/sampling batch (s) 0.283278 +time/saving (s) 0.0044519 +time/training (s) 8.01475 +time/epoch (s) 10.7821 +time/total (s) 4693.54 +Epoch -542 +---------------------------------- -------------- +2022-05-10 14:29:04.278662 PDT | [2] Epoch -541 finished +---------------------------------- --------------- +epoch -541 +replay_buffer/size 999033 +trainer/num train calls 460000 +trainer/Policy Loss -19.6889 +trainer/Log Pis Mean 24.5271 +trainer/Log Pis Std 13.2086 +trainer/Log Pis Max 63.3738 +trainer/Log Pis Min -5.87152 +trainer/policy/mean Mean -0.0428024 +trainer/policy/mean Std 0.905292 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.81035 +trainer/policy/normal/std Std 0.653881 +trainer/policy/normal/std Max 5.21174 +trainer/policy/normal/std Min 0.322689 +trainer/policy/normal/log_std Mean 0.993635 +trainer/policy/normal/log_std Std 0.319634 +trainer/policy/normal/log_std Max 1.65091 +trainer/policy/normal/log_std Min -1.13107 +eval/num steps total 459470 +eval/num paths total 460 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0336741 +eval/Actions Std 0.92134 +eval/Actions Max 0.999999 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.49768 +time/logging (s) 0.00377637 +time/sampling batch (s) 0.274882 +time/saving (s) 0.00340593 +time/training (s) 6.32217 +time/epoch (s) 9.10192 +time/total (s) 4702.65 +Epoch -541 +---------------------------------- --------------- +2022-05-10 14:29:14.547337 PDT | [2] Epoch -540 finished +---------------------------------- --------------- +epoch -540 +replay_buffer/size 999033 +trainer/num train calls 461000 +trainer/Policy Loss -18.5321 +trainer/Log Pis Mean 24.3517 +trainer/Log Pis Std 12.9318 +trainer/Log Pis Max 68.8974 +trainer/Log Pis Min -4.16239 +trainer/policy/mean Mean -0.0552245 +trainer/policy/mean Std 0.90491 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.79804 +trainer/policy/normal/std Std 0.652019 +trainer/policy/normal/std Max 7.34695 +trainer/policy/normal/std Min 0.300497 +trainer/policy/normal/log_std Mean 0.989895 +trainer/policy/normal/log_std Std 0.315504 +trainer/policy/normal/log_std Max 1.99429 +trainer/policy/normal/log_std Min -1.20232 +eval/num steps total 460470 +eval/num paths total 461 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0437698 +eval/Actions Std 0.912012 +eval/Actions Max 0.999997 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61347 +time/logging (s) 0.00374079 +time/sampling batch (s) 0.277165 +time/saving (s) 0.00340309 +time/training (s) 7.3491 +time/epoch (s) 10.2469 +time/total (s) 4712.9 +Epoch -540 +---------------------------------- --------------- +2022-05-10 14:29:24.854525 PDT | [2] Epoch -539 finished +---------------------------------- --------------- +epoch -539 +replay_buffer/size 999033 +trainer/num train calls 462000 +trainer/Policy Loss -18.9533 +trainer/Log Pis Mean 24.4835 +trainer/Log Pis Std 13.3315 +trainer/Log Pis Max 72.0714 +trainer/Log Pis Min -6.54116 +trainer/policy/mean Mean -0.0435386 +trainer/policy/mean Std 0.904724 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.87073 +trainer/policy/normal/std Std 0.661473 +trainer/policy/normal/std Max 5.79554 +trainer/policy/normal/std Min 0.335238 +trainer/policy/normal/log_std Mean 1.01733 +trainer/policy/normal/log_std Std 0.305257 +trainer/policy/normal/log_std Max 1.75709 +trainer/policy/normal/log_std Min -1.09291 +eval/num steps total 461470 +eval/num paths total 462 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0979148 +eval/Actions Std 0.915794 +eval/Actions Max 0.999995 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58711 +time/logging (s) 0.00462719 +time/sampling batch (s) 0.286459 +time/saving (s) 0.00439944 +time/training (s) 7.4031 +time/epoch (s) 10.2857 +time/total (s) 4723.19 +Epoch -539 +---------------------------------- --------------- +2022-05-10 14:29:34.287458 PDT | [2] Epoch -538 finished +---------------------------------- --------------- +epoch -538 +replay_buffer/size 999033 +trainer/num train calls 463000 +trainer/Policy Loss -19.9513 +trainer/Log Pis Mean 24.5846 +trainer/Log Pis Std 13.2539 +trainer/Log Pis Max 66.7143 +trainer/Log Pis Min -10.3769 +trainer/policy/mean Mean -0.0138084 +trainer/policy/mean Std 0.904658 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.78663 +trainer/policy/normal/std Std 0.647849 +trainer/policy/normal/std Max 6.12222 +trainer/policy/normal/std Min 0.303942 +trainer/policy/normal/log_std Mean 0.98549 +trainer/policy/normal/log_std Std 0.31743 +trainer/policy/normal/log_std Max 1.81193 +trainer/policy/normal/log_std Min -1.19092 +eval/num steps total 462470 +eval/num paths total 463 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.209408 +eval/Actions Std 0.859609 +eval/Actions Max 0.999995 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.8917 +time/logging (s) 0.00438599 +time/sampling batch (s) 0.279738 +time/saving (s) 0.0041561 +time/training (s) 6.23006 +time/epoch (s) 9.41005 +time/total (s) 4732.6 +Epoch -538 +---------------------------------- --------------- +2022-05-10 14:29:43.775102 PDT | [2] Epoch -537 finished +---------------------------------- --------------- +epoch -537 +replay_buffer/size 999033 +trainer/num train calls 464000 +trainer/Policy Loss -19.4001 +trainer/Log Pis Mean 24.163 +trainer/Log Pis Std 13.2172 +trainer/Log Pis Max 65.8824 +trainer/Log Pis Min -9.40427 +trainer/policy/mean Mean -0.0289508 +trainer/policy/mean Std 0.902564 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.80799 +trainer/policy/normal/std Std 0.65613 +trainer/policy/normal/std Max 5.07308 +trainer/policy/normal/std Min 0.31688 +trainer/policy/normal/log_std Mean 0.991328 +trainer/policy/normal/log_std Std 0.327376 +trainer/policy/normal/log_std Max 1.62395 +trainer/policy/normal/log_std Min -1.14923 +eval/num steps total 463470 +eval/num paths total 464 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.085749 +eval/Actions Std 0.910526 +eval/Actions Max 0.999995 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.71743 +time/logging (s) 0.00371474 +time/sampling batch (s) 0.529244 +time/saving (s) 0.00345997 +time/training (s) 6.21078 +time/epoch (s) 9.46464 +time/total (s) 4742.07 +Epoch -537 +---------------------------------- --------------- +2022-05-10 14:29:55.407704 PDT | [2] Epoch -536 finished +---------------------------------- --------------- +epoch -536 +replay_buffer/size 999033 +trainer/num train calls 465000 +trainer/Policy Loss -18.5976 +trainer/Log Pis Mean 24.9892 +trainer/Log Pis Std 12.5897 +trainer/Log Pis Max 65.837 +trainer/Log Pis Min -5.64672 +trainer/policy/mean Mean -0.0351649 +trainer/policy/mean Std 0.913189 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.76012 +trainer/policy/normal/std Std 0.631728 +trainer/policy/normal/std Max 5.63862 +trainer/policy/normal/std Min 0.341492 +trainer/policy/normal/log_std Mean 0.977481 +trainer/policy/normal/log_std Std 0.310443 +trainer/policy/normal/log_std Max 1.72964 +trainer/policy/normal/log_std Min -1.07443 +eval/num steps total 464470 +eval/num paths total 465 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00125632 +eval/Actions Std 0.908531 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64171 +time/logging (s) 0.00382639 +time/sampling batch (s) 0.567589 +time/saving (s) 0.00375713 +time/training (s) 8.39347 +time/epoch (s) 11.6103 +time/total (s) 4753.68 +Epoch -536 +---------------------------------- --------------- +2022-05-10 14:30:06.183760 PDT | [2] Epoch -535 finished +---------------------------------- --------------- +epoch -535 +replay_buffer/size 999033 +trainer/num train calls 466000 +trainer/Policy Loss -18.5351 +trainer/Log Pis Mean 23.7533 +trainer/Log Pis Std 13.3939 +trainer/Log Pis Max 72.5539 +trainer/Log Pis Min -8.35511 +trainer/policy/mean Mean -0.0359594 +trainer/policy/mean Std 0.904121 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.84443 +trainer/policy/normal/std Std 0.646697 +trainer/policy/normal/std Max 5.52161 +trainer/policy/normal/std Min 0.329604 +trainer/policy/normal/log_std Mean 1.00747 +trainer/policy/normal/log_std Std 0.312265 +trainer/policy/normal/log_std Max 1.70867 +trainer/policy/normal/log_std Min -1.10986 +eval/num steps total 465470 +eval/num paths total 466 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.567722 +eval/Actions Std 0.712151 +eval/Actions Max 0.999987 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.76145 +time/logging (s) 0.00422908 +time/sampling batch (s) 0.331515 +time/saving (s) 0.00413112 +time/training (s) 7.65232 +time/epoch (s) 10.7536 +time/total (s) 4764.44 +Epoch -535 +---------------------------------- --------------- +2022-05-10 14:30:17.031784 PDT | [2] Epoch -534 finished +---------------------------------- --------------- +epoch -534 +replay_buffer/size 999033 +trainer/num train calls 467000 +trainer/Policy Loss -19.5962 +trainer/Log Pis Mean 24.787 +trainer/Log Pis Std 13.2049 +trainer/Log Pis Max 64.0242 +trainer/Log Pis Min -7.33821 +trainer/policy/mean Mean -0.0436622 +trainer/policy/mean Std 0.911531 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.8554 +trainer/policy/normal/std Std 0.616591 +trainer/policy/normal/std Max 5.902 +trainer/policy/normal/std Min 0.279748 +trainer/policy/normal/log_std Mean 1.01752 +trainer/policy/normal/log_std Std 0.279951 +trainer/policy/normal/log_std Max 1.77529 +trainer/policy/normal/log_std Min -1.27386 +eval/num steps total 466470 +eval/num paths total 467 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.201216 +eval/Actions Std 0.911325 +eval/Actions Max 0.999999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.74219 +time/logging (s) 0.00388323 +time/sampling batch (s) 0.579834 +time/saving (s) 0.00377755 +time/training (s) 7.49502 +time/epoch (s) 10.8247 +time/total (s) 4775.27 +Epoch -534 +---------------------------------- --------------- +2022-05-10 14:30:28.248492 PDT | [2] Epoch -533 finished +---------------------------------- -------------- +epoch -533 +replay_buffer/size 999033 +trainer/num train calls 468000 +trainer/Policy Loss -19.5665 +trainer/Log Pis Mean 23.8766 +trainer/Log Pis Std 13.3592 +trainer/Log Pis Max 71.0074 +trainer/Log Pis Min -11.1154 +trainer/policy/mean Mean -0.0347557 +trainer/policy/mean Std 0.905675 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.77127 +trainer/policy/normal/std Std 0.657917 +trainer/policy/normal/std Max 5.17939 +trainer/policy/normal/std Min 0.287765 +trainer/policy/normal/log_std Mean 0.977757 +trainer/policy/normal/log_std Std 0.326473 +trainer/policy/normal/log_std Max 1.64469 +trainer/policy/normal/log_std Min -1.24561 +eval/num steps total 467470 +eval/num paths total 468 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.084545 +eval/Actions Std 0.913211 +eval/Actions Max 0.999995 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 3.00551 +time/logging (s) 0.0037539 +time/sampling batch (s) 0.581706 +time/saving (s) 0.0034951 +time/training (s) 7.5991 +time/epoch (s) 11.1936 +time/total (s) 4786.46 +Epoch -533 +---------------------------------- -------------- +2022-05-10 14:30:38.646900 PDT | [2] Epoch -532 finished +---------------------------------- --------------- +epoch -532 +replay_buffer/size 999033 +trainer/num train calls 469000 +trainer/Policy Loss -19.9273 +trainer/Log Pis Mean 24.3103 +trainer/Log Pis Std 13.3198 +trainer/Log Pis Max 64.4538 +trainer/Log Pis Min -12.3969 +trainer/policy/mean Mean -0.0354444 +trainer/policy/mean Std 0.905065 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.8172 +trainer/policy/normal/std Std 0.637338 +trainer/policy/normal/std Max 5.92028 +trainer/policy/normal/std Min 0.264669 +trainer/policy/normal/log_std Mean 0.999105 +trainer/policy/normal/log_std Std 0.305177 +trainer/policy/normal/log_std Max 1.77838 +trainer/policy/normal/log_std Min -1.32927 +eval/num steps total 468470 +eval/num paths total 469 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0786488 +eval/Actions Std 0.929877 +eval/Actions Max 0.999996 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60695 +time/logging (s) 0.00374588 +time/sampling batch (s) 0.279979 +time/saving (s) 0.00342113 +time/training (s) 7.48223 +time/epoch (s) 10.3763 +time/total (s) 4796.84 +Epoch -532 +---------------------------------- --------------- +2022-05-10 14:30:48.357627 PDT | [2] Epoch -531 finished +---------------------------------- --------------- +epoch -531 +replay_buffer/size 999033 +trainer/num train calls 470000 +trainer/Policy Loss -19.1224 +trainer/Log Pis Mean 24.8536 +trainer/Log Pis Std 13.401 +trainer/Log Pis Max 68.2351 +trainer/Log Pis Min -5.72157 +trainer/policy/mean Mean -0.040029 +trainer/policy/mean Std 0.909507 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.90318 +trainer/policy/normal/std Std 0.633826 +trainer/policy/normal/std Max 6.14586 +trainer/policy/normal/std Min 0.347248 +trainer/policy/normal/log_std Mean 1.03164 +trainer/policy/normal/log_std Std 0.294713 +trainer/policy/normal/log_std Max 1.81578 +trainer/policy/normal/log_std Min -1.05772 +eval/num steps total 469470 +eval/num paths total 470 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0036944 +eval/Actions Std 0.952377 +eval/Actions Max 0.999997 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54102 +time/logging (s) 0.00402535 +time/sampling batch (s) 0.281844 +time/saving (s) 0.00403467 +time/training (s) 6.8579 +time/epoch (s) 9.68882 +time/total (s) 4806.53 +Epoch -531 +---------------------------------- --------------- +2022-05-10 14:30:59.586094 PDT | [2] Epoch -530 finished +---------------------------------- --------------- +epoch -530 +replay_buffer/size 999033 +trainer/num train calls 471000 +trainer/Policy Loss -20.6099 +trainer/Log Pis Mean 24.1019 +trainer/Log Pis Std 13.3715 +trainer/Log Pis Max 75.8204 +trainer/Log Pis Min -4.38849 +trainer/policy/mean Mean -0.0473782 +trainer/policy/mean Std 0.906231 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.77817 +trainer/policy/normal/std Std 0.63483 +trainer/policy/normal/std Max 5.36656 +trainer/policy/normal/std Min 0.233089 +trainer/policy/normal/log_std Mean 0.983852 +trainer/policy/normal/log_std Std 0.311448 +trainer/policy/normal/log_std Max 1.68019 +trainer/policy/normal/log_std Min -1.45633 +eval/num steps total 470470 +eval/num paths total 471 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.166441 +eval/Actions Std 0.903651 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.83279 +time/logging (s) 0.00370093 +time/sampling batch (s) 0.281991 +time/saving (s) 0.00343006 +time/training (s) 8.08379 +time/epoch (s) 11.2057 +time/total (s) 4817.74 +Epoch -530 +---------------------------------- --------------- +2022-05-10 14:31:10.534489 PDT | [2] Epoch -529 finished +---------------------------------- --------------- +epoch -529 +replay_buffer/size 999033 +trainer/num train calls 472000 +trainer/Policy Loss -20.0987 +trainer/Log Pis Mean 23.8628 +trainer/Log Pis Std 13.3091 +trainer/Log Pis Max 73.7199 +trainer/Log Pis Min -6.88264 +trainer/policy/mean Mean -0.0323343 +trainer/policy/mean Std 0.908481 +trainer/policy/mean Max 0.999975 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.79188 +trainer/policy/normal/std Std 0.645795 +trainer/policy/normal/std Max 5.60201 +trainer/policy/normal/std Min 0.276502 +trainer/policy/normal/log_std Mean 0.987192 +trainer/policy/normal/log_std Std 0.319463 +trainer/policy/normal/log_std Max 1.72313 +trainer/policy/normal/log_std Min -1.28554 +eval/num steps total 471470 +eval/num paths total 472 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.374023 +eval/Actions Std 0.844824 +eval/Actions Max 0.999986 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68451 +time/logging (s) 0.00422729 +time/sampling batch (s) 0.529581 +time/saving (s) 0.0042241 +time/training (s) 7.70433 +time/epoch (s) 10.9269 +time/total (s) 4828.67 +Epoch -529 +---------------------------------- --------------- +2022-05-10 14:31:20.084108 PDT | [2] Epoch -528 finished +---------------------------------- --------------- +epoch -528 +replay_buffer/size 999033 +trainer/num train calls 473000 +trainer/Policy Loss -20.3544 +trainer/Log Pis Mean 24.5668 +trainer/Log Pis Std 13.1324 +trainer/Log Pis Max 63.188 +trainer/Log Pis Min -4.04518 +trainer/policy/mean Mean -0.0445882 +trainer/policy/mean Std 0.909159 +trainer/policy/mean Max 0.999968 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.80165 +trainer/policy/normal/std Std 0.632292 +trainer/policy/normal/std Max 5.53165 +trainer/policy/normal/std Min 0.33725 +trainer/policy/normal/log_std Mean 0.994693 +trainer/policy/normal/log_std Std 0.297396 +trainer/policy/normal/log_std Max 1.71049 +trainer/policy/normal/log_std Min -1.08693 +eval/num steps total 472470 +eval/num paths total 473 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.558189 +eval/Actions Std 0.752792 +eval/Actions Max 0.999991 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59381 +time/logging (s) 0.00368816 +time/sampling batch (s) 0.28049 +time/saving (s) 0.00340658 +time/training (s) 6.64524 +time/epoch (s) 9.52664 +time/total (s) 4838.2 +Epoch -528 +---------------------------------- --------------- +2022-05-10 14:31:30.449570 PDT | [2] Epoch -527 finished +---------------------------------- --------------- +epoch -527 +replay_buffer/size 999033 +trainer/num train calls 474000 +trainer/Policy Loss -19.5717 +trainer/Log Pis Mean 24.0731 +trainer/Log Pis Std 14.1884 +trainer/Log Pis Max 72.0018 +trainer/Log Pis Min -11.6893 +trainer/policy/mean Mean -0.0469954 +trainer/policy/mean Std 0.90131 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.83811 +trainer/policy/normal/std Std 0.65544 +trainer/policy/normal/std Max 6.33889 +trainer/policy/normal/std Min 0.364887 +trainer/policy/normal/log_std Mean 1.00703 +trainer/policy/normal/log_std Std 0.298161 +trainer/policy/normal/log_std Max 1.8467 +trainer/policy/normal/log_std Min -1.00817 +eval/num steps total 473470 +eval/num paths total 474 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.069676 +eval/Actions Std 0.914575 +eval/Actions Max 0.999993 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54538 +time/logging (s) 0.00379235 +time/sampling batch (s) 0.278995 +time/saving (s) 0.0035226 +time/training (s) 7.51194 +time/epoch (s) 10.3436 +time/total (s) 4848.55 +Epoch -527 +---------------------------------- --------------- +2022-05-10 14:31:40.192759 PDT | [2] Epoch -526 finished +---------------------------------- --------------- +epoch -526 +replay_buffer/size 999033 +trainer/num train calls 475000 +trainer/Policy Loss -18.9619 +trainer/Log Pis Mean 23.4621 +trainer/Log Pis Std 12.8053 +trainer/Log Pis Max 61.0209 +trainer/Log Pis Min -8.17562 +trainer/policy/mean Mean -0.0234974 +trainer/policy/mean Std 0.905881 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999983 +trainer/policy/normal/std Mean 2.8882 +trainer/policy/normal/std Std 0.663987 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.287058 +trainer/policy/normal/log_std Mean 1.0241 +trainer/policy/normal/log_std Std 0.302999 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.24807 +eval/num steps total 474470 +eval/num paths total 475 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.434414 +eval/Actions Std 0.858243 +eval/Actions Max 0.999998 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66982 +time/logging (s) 0.00415026 +time/sampling batch (s) 0.277924 +time/saving (s) 0.00387294 +time/training (s) 6.76567 +time/epoch (s) 9.72144 +time/total (s) 4858.27 +Epoch -526 +---------------------------------- --------------- +2022-05-10 14:31:49.686672 PDT | [2] Epoch -525 finished +---------------------------------- --------------- +epoch -525 +replay_buffer/size 999033 +trainer/num train calls 476000 +trainer/Policy Loss -19.0565 +trainer/Log Pis Mean 23.974 +trainer/Log Pis Std 13.1243 +trainer/Log Pis Max 63.6875 +trainer/Log Pis Min -8.49401 +trainer/policy/mean Mean -0.0199219 +trainer/policy/mean Std 0.900416 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.82285 +trainer/policy/normal/std Std 0.651 +trainer/policy/normal/std Max 5.74997 +trainer/policy/normal/std Min 0.252376 +trainer/policy/normal/log_std Mean 1.00005 +trainer/policy/normal/log_std Std 0.308668 +trainer/policy/normal/log_std Max 1.7492 +trainer/policy/normal/log_std Min -1.37683 +eval/num steps total 475470 +eval/num paths total 476 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.119528 +eval/Actions Std 0.881701 +eval/Actions Max 0.999989 +eval/Actions Min -0.999986 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.47167 +time/logging (s) 0.00390314 +time/sampling batch (s) 0.276175 +time/saving (s) 0.0035434 +time/training (s) 6.71622 +time/epoch (s) 9.47151 +time/total (s) 4867.75 +Epoch -525 +---------------------------------- --------------- +2022-05-10 14:31:59.598762 PDT | [2] Epoch -524 finished +---------------------------------- --------------- +epoch -524 +replay_buffer/size 999033 +trainer/num train calls 477000 +trainer/Policy Loss -20.7907 +trainer/Log Pis Mean 23.6996 +trainer/Log Pis Std 12.7265 +trainer/Log Pis Max 65.5099 +trainer/Log Pis Min -7.72084 +trainer/policy/mean Mean -0.0441796 +trainer/policy/mean Std 0.906258 +trainer/policy/mean Max 0.999982 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.78455 +trainer/policy/normal/std Std 0.628698 +trainer/policy/normal/std Max 6.48921 +trainer/policy/normal/std Min 0.313361 +trainer/policy/normal/log_std Mean 0.987693 +trainer/policy/normal/log_std Std 0.303707 +trainer/policy/normal/log_std Max 1.87014 +trainer/policy/normal/log_std Min -1.1604 +eval/num steps total 476470 +eval/num paths total 477 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.16762 +eval/Actions Std 0.949009 +eval/Actions Max 0.999991 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.4839 +time/logging (s) 0.00380279 +time/sampling batch (s) 0.274633 +time/saving (s) 0.00341293 +time/training (s) 7.12445 +time/epoch (s) 9.8902 +time/total (s) 4877.64 +Epoch -524 +---------------------------------- --------------- +2022-05-10 14:32:09.188584 PDT | [2] Epoch -523 finished +---------------------------------- --------------- +epoch -523 +replay_buffer/size 999033 +trainer/num train calls 478000 +trainer/Policy Loss -18.9984 +trainer/Log Pis Mean 24.7818 +trainer/Log Pis Std 13.1452 +trainer/Log Pis Max 69.704 +trainer/Log Pis Min -5.69337 +trainer/policy/mean Mean -0.0350723 +trainer/policy/mean Std 0.905253 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.73191 +trainer/policy/normal/std Std 0.637153 +trainer/policy/normal/std Max 5.57307 +trainer/policy/normal/std Min 0.296953 +trainer/policy/normal/log_std Mean 0.964959 +trainer/policy/normal/log_std Std 0.322038 +trainer/policy/normal/log_std Max 1.71795 +trainer/policy/normal/log_std Min -1.21418 +eval/num steps total 477470 +eval/num paths total 478 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0131845 +eval/Actions Std 0.963366 +eval/Actions Max 0.999993 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70919 +time/logging (s) 0.00380721 +time/sampling batch (s) 0.274564 +time/saving (s) 0.00341478 +time/training (s) 6.57723 +time/epoch (s) 9.56821 +time/total (s) 4887.21 +Epoch -523 +---------------------------------- --------------- +2022-05-10 14:32:19.262334 PDT | [2] Epoch -522 finished +---------------------------------- --------------- +epoch -522 +replay_buffer/size 999033 +trainer/num train calls 479000 +trainer/Policy Loss -19.8379 +trainer/Log Pis Mean 24.8951 +trainer/Log Pis Std 13.6307 +trainer/Log Pis Max 80.7976 +trainer/Log Pis Min -10.006 +trainer/policy/mean Mean -0.0517184 +trainer/policy/mean Std 0.906982 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.87817 +trainer/policy/normal/std Std 0.654889 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.270895 +trainer/policy/normal/log_std Mean 1.02093 +trainer/policy/normal/log_std Std 0.302121 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.30603 +eval/num steps total 478470 +eval/num paths total 479 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0565582 +eval/Actions Std 0.914925 +eval/Actions Max 0.999996 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63459 +time/logging (s) 0.00371674 +time/sampling batch (s) 0.277355 +time/saving (s) 0.00336962 +time/training (s) 7.13278 +time/epoch (s) 10.0518 +time/total (s) 4897.27 +Epoch -522 +---------------------------------- --------------- +2022-05-10 14:32:29.107040 PDT | [2] Epoch -521 finished +---------------------------------- --------------- +epoch -521 +replay_buffer/size 999033 +trainer/num train calls 480000 +trainer/Policy Loss -19.9052 +trainer/Log Pis Mean 25.5797 +trainer/Log Pis Std 13.6815 +trainer/Log Pis Max 74.1349 +trainer/Log Pis Min -9.44794 +trainer/policy/mean Mean -0.0517805 +trainer/policy/mean Std 0.910885 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.84079 +trainer/policy/normal/std Std 0.644766 +trainer/policy/normal/std Max 5.9939 +trainer/policy/normal/std Min 0.248177 +trainer/policy/normal/log_std Mean 1.00742 +trainer/policy/normal/log_std Std 0.30538 +trainer/policy/normal/log_std Max 1.79074 +trainer/policy/normal/log_std Min -1.39361 +eval/num steps total 479470 +eval/num paths total 480 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.163185 +eval/Actions Std 0.905272 +eval/Actions Max 0.999988 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.47115 +time/logging (s) 0.00394049 +time/sampling batch (s) 0.281495 +time/saving (s) 0.00404735 +time/training (s) 7.0622 +time/epoch (s) 9.82283 +time/total (s) 4907.09 +Epoch -521 +---------------------------------- --------------- +2022-05-10 14:32:39.250763 PDT | [2] Epoch -520 finished +---------------------------------- --------------- +epoch -520 +replay_buffer/size 999033 +trainer/num train calls 481000 +trainer/Policy Loss -19.7379 +trainer/Log Pis Mean 25.755 +trainer/Log Pis Std 13.2452 +trainer/Log Pis Max 69.627 +trainer/Log Pis Min -3.7668 +trainer/policy/mean Mean -0.030619 +trainer/policy/mean Std 0.91078 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.75185 +trainer/policy/normal/std Std 0.628634 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.260084 +trainer/policy/normal/log_std Mean 0.973611 +trainer/policy/normal/log_std Std 0.316639 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.34675 +eval/num steps total 480470 +eval/num paths total 481 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.406424 +eval/Actions Std 0.770265 +eval/Actions Max 0.999999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65544 +time/logging (s) 0.00369246 +time/sampling batch (s) 0.777542 +time/saving (s) 0.00337502 +time/training (s) 6.68125 +time/epoch (s) 10.1213 +time/total (s) 4917.22 +Epoch -520 +---------------------------------- --------------- +2022-05-10 14:32:49.977763 PDT | [2] Epoch -519 finished +---------------------------------- --------------- +epoch -519 +replay_buffer/size 999033 +trainer/num train calls 482000 +trainer/Policy Loss -19.1333 +trainer/Log Pis Mean 25.6328 +trainer/Log Pis Std 13.2333 +trainer/Log Pis Max 66.6906 +trainer/Log Pis Min -5.40979 +trainer/policy/mean Mean -0.0425205 +trainer/policy/mean Std 0.910712 +trainer/policy/mean Max 0.999978 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.84372 +trainer/policy/normal/std Std 0.667423 +trainer/policy/normal/std Max 5.2477 +trainer/policy/normal/std Min 0.341621 +trainer/policy/normal/log_std Mean 1.00374 +trainer/policy/normal/log_std Std 0.328093 +trainer/policy/normal/log_std Max 1.65779 +trainer/policy/normal/log_std Min -1.07405 +eval/num steps total 481470 +eval/num paths total 482 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.288155 +eval/Actions Std 0.860741 +eval/Actions Max 0.99999 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68631 +time/logging (s) 0.00369184 +time/sampling batch (s) 0.280871 +time/saving (s) 0.00337252 +time/training (s) 7.73051 +time/epoch (s) 10.7048 +time/total (s) 4927.92 +Epoch -519 +---------------------------------- --------------- +2022-05-10 14:32:59.690199 PDT | [2] Epoch -518 finished +---------------------------------- --------------- +epoch -518 +replay_buffer/size 999033 +trainer/num train calls 483000 +trainer/Policy Loss -18.5593 +trainer/Log Pis Mean 25.6193 +trainer/Log Pis Std 13.7958 +trainer/Log Pis Max 71.0218 +trainer/Log Pis Min -10.9936 +trainer/policy/mean Mean -0.0469125 +trainer/policy/mean Std 0.907411 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.83652 +trainer/policy/normal/std Std 0.650258 +trainer/policy/normal/std Max 5.74343 +trainer/policy/normal/std Min 0.320061 +trainer/policy/normal/log_std Mean 1.00504 +trainer/policy/normal/log_std Std 0.308552 +trainer/policy/normal/log_std Max 1.74806 +trainer/policy/normal/log_std Min -1.13925 +eval/num steps total 482470 +eval/num paths total 483 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.195667 +eval/Actions Std 0.873521 +eval/Actions Max 0.999993 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.4292 +time/logging (s) 0.00367279 +time/sampling batch (s) 0.280012 +time/saving (s) 0.00334571 +time/training (s) 6.974 +time/epoch (s) 9.69024 +time/total (s) 4937.62 +Epoch -518 +---------------------------------- --------------- +2022-05-10 14:33:10.531640 PDT | [2] Epoch -517 finished +---------------------------------- --------------- +epoch -517 +replay_buffer/size 999033 +trainer/num train calls 484000 +trainer/Policy Loss -19.6936 +trainer/Log Pis Mean 24.496 +trainer/Log Pis Std 13.2297 +trainer/Log Pis Max 67.2032 +trainer/Log Pis Min -6.69559 +trainer/policy/mean Mean -0.026049 +trainer/policy/mean Std 0.905618 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.76181 +trainer/policy/normal/std Std 0.644897 +trainer/policy/normal/std Max 6.20006 +trainer/policy/normal/std Min 0.312907 +trainer/policy/normal/log_std Mean 0.975148 +trainer/policy/normal/log_std Std 0.325477 +trainer/policy/normal/log_std Max 1.82456 +trainer/policy/normal/log_std Min -1.16185 +eval/num steps total 483470 +eval/num paths total 484 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00959064 +eval/Actions Std 0.909624 +eval/Actions Max 0.999995 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59201 +time/logging (s) 0.00429114 +time/sampling batch (s) 0.777795 +time/saving (s) 0.00346564 +time/training (s) 7.44255 +time/epoch (s) 10.8201 +time/total (s) 4948.44 +Epoch -517 +---------------------------------- --------------- +2022-05-10 14:33:20.745188 PDT | [2] Epoch -516 finished +---------------------------------- --------------- +epoch -516 +replay_buffer/size 999033 +trainer/num train calls 485000 +trainer/Policy Loss -19.6447 +trainer/Log Pis Mean 24.242 +trainer/Log Pis Std 13.7807 +trainer/Log Pis Max 66.3132 +trainer/Log Pis Min -12.9273 +trainer/policy/mean Mean -0.012626 +trainer/policy/mean Std 0.904281 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.82175 +trainer/policy/normal/std Std 0.689399 +trainer/policy/normal/std Max 6.69515 +trainer/policy/normal/std Min 0.297428 +trainer/policy/normal/log_std Mean 0.991894 +trainer/policy/normal/log_std Std 0.345428 +trainer/policy/normal/log_std Max 1.90138 +trainer/policy/normal/log_std Min -1.21258 +eval/num steps total 484470 +eval/num paths total 485 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.38252 +eval/Actions Std 0.886906 +eval/Actions Max 0.999988 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.74509 +time/logging (s) 0.003678 +time/sampling batch (s) 0.281622 +time/saving (s) 0.00341231 +time/training (s) 7.15626 +time/epoch (s) 10.1901 +time/total (s) 4958.63 +Epoch -516 +---------------------------------- --------------- +2022-05-10 14:33:29.901150 PDT | [2] Epoch -515 finished +---------------------------------- --------------- +epoch -515 +replay_buffer/size 999033 +trainer/num train calls 486000 +trainer/Policy Loss -20.1585 +trainer/Log Pis Mean 24.0268 +trainer/Log Pis Std 13.768 +trainer/Log Pis Max 67.1322 +trainer/Log Pis Min -9.68329 +trainer/policy/mean Mean -0.0445025 +trainer/policy/mean Std 0.903427 +trainer/policy/mean Max 0.99998 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.7526 +trainer/policy/normal/std Std 0.640792 +trainer/policy/normal/std Max 5.86273 +trainer/policy/normal/std Min 0.267151 +trainer/policy/normal/log_std Mean 0.973143 +trainer/policy/normal/log_std Std 0.317159 +trainer/policy/normal/log_std Max 1.76862 +trainer/policy/normal/log_std Min -1.31994 +eval/num steps total 485470 +eval/num paths total 486 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.377199 +eval/Actions Std 0.866427 +eval/Actions Max 0.999988 +eval/Actions Min -0.999988 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.78004 +time/logging (s) 0.00399172 +time/sampling batch (s) 0.279707 +time/saving (s) 0.00405887 +time/training (s) 6.06644 +time/epoch (s) 9.13424 +time/total (s) 4967.77 +Epoch -515 +---------------------------------- --------------- +2022-05-10 14:33:39.189034 PDT | [2] Epoch -514 finished +---------------------------------- --------------- +epoch -514 +replay_buffer/size 999033 +trainer/num train calls 487000 +trainer/Policy Loss -20.1528 +trainer/Log Pis Mean 24.2368 +trainer/Log Pis Std 12.3635 +trainer/Log Pis Max 62.5128 +trainer/Log Pis Min -6.34667 +trainer/policy/mean Mean -0.0179125 +trainer/policy/mean Std 0.906073 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.83928 +trainer/policy/normal/std Std 0.64493 +trainer/policy/normal/std Max 5.5212 +trainer/policy/normal/std Min 0.336952 +trainer/policy/normal/log_std Mean 1.00735 +trainer/policy/normal/log_std Std 0.301674 +trainer/policy/normal/log_std Max 1.7086 +trainer/policy/normal/log_std Min -1.08782 +eval/num steps total 486470 +eval/num paths total 487 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0390772 +eval/Actions Std 0.902 +eval/Actions Max 0.999997 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.45155 +time/logging (s) 0.00380212 +time/sampling batch (s) 0.278699 +time/saving (s) 0.00346162 +time/training (s) 6.52776 +time/epoch (s) 9.26527 +time/total (s) 4977.04 +Epoch -514 +---------------------------------- --------------- +2022-05-10 14:33:49.743045 PDT | [2] Epoch -513 finished +---------------------------------- --------------- +epoch -513 +replay_buffer/size 999033 +trainer/num train calls 488000 +trainer/Policy Loss -19.1583 +trainer/Log Pis Mean 23.7824 +trainer/Log Pis Std 13.3597 +trainer/Log Pis Max 69.2227 +trainer/Log Pis Min -6.95811 +trainer/policy/mean Mean -0.0573435 +trainer/policy/mean Std 0.905257 +trainer/policy/mean Max 0.999978 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.84645 +trainer/policy/normal/std Std 0.629587 +trainer/policy/normal/std Max 6.17315 +trainer/policy/normal/std Min 0.281084 +trainer/policy/normal/log_std Mean 1.0112 +trainer/policy/normal/log_std Std 0.298526 +trainer/policy/normal/log_std Max 1.82021 +trainer/policy/normal/log_std Min -1.2691 +eval/num steps total 487470 +eval/num paths total 488 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0734791 +eval/Actions Std 0.900326 +eval/Actions Max 0.999997 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65478 +time/logging (s) 0.00379633 +time/sampling batch (s) 0.29554 +time/saving (s) 0.00345858 +time/training (s) 7.57411 +time/epoch (s) 10.5317 +time/total (s) 4987.57 +Epoch -513 +---------------------------------- --------------- +2022-05-10 14:33:59.267189 PDT | [2] Epoch -512 finished +---------------------------------- --------------- +epoch -512 +replay_buffer/size 999033 +trainer/num train calls 489000 +trainer/Policy Loss -19.3134 +trainer/Log Pis Mean 24.2522 +trainer/Log Pis Std 13.4137 +trainer/Log Pis Max 64.3076 +trainer/Log Pis Min -5.68834 +trainer/policy/mean Mean -0.0382521 +trainer/policy/mean Std 0.908976 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.77346 +trainer/policy/normal/std Std 0.62773 +trainer/policy/normal/std Max 5.31749 +trainer/policy/normal/std Min 0.315703 +trainer/policy/normal/log_std Mean 0.983864 +trainer/policy/normal/log_std Std 0.302603 +trainer/policy/normal/log_std Max 1.671 +trainer/policy/normal/log_std Min -1.15295 +eval/num steps total 488470 +eval/num paths total 489 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.233511 +eval/Actions Std 0.940608 +eval/Actions Max 0.999993 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.72636 +time/logging (s) 0.00378531 +time/sampling batch (s) 0.279958 +time/saving (s) 0.00344407 +time/training (s) 6.48846 +time/epoch (s) 9.50201 +time/total (s) 4997.08 +Epoch -512 +---------------------------------- --------------- +2022-05-10 14:34:09.296859 PDT | [2] Epoch -511 finished +---------------------------------- --------------- +epoch -511 +replay_buffer/size 999033 +trainer/num train calls 490000 +trainer/Policy Loss -18.8936 +trainer/Log Pis Mean 23.907 +trainer/Log Pis Std 13.0587 +trainer/Log Pis Max 69.3454 +trainer/Log Pis Min -7.521 +trainer/policy/mean Mean -0.0411593 +trainer/policy/mean Std 0.904489 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.81728 +trainer/policy/normal/std Std 0.651312 +trainer/policy/normal/std Max 6.09679 +trainer/policy/normal/std Min 0.308777 +trainer/policy/normal/log_std Mean 0.998362 +trainer/policy/normal/log_std Std 0.306056 +trainer/policy/normal/log_std Max 1.80776 +trainer/policy/normal/log_std Min -1.17513 +eval/num steps total 489470 +eval/num paths total 490 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.223464 +eval/Actions Std 0.919815 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.34307 +time/logging (s) 0.00375481 +time/sampling batch (s) 0.279667 +time/saving (s) 0.00343389 +time/training (s) 7.37752 +time/epoch (s) 10.0074 +time/total (s) 5007.09 +Epoch -511 +---------------------------------- --------------- +2022-05-10 14:34:18.804326 PDT | [2] Epoch -510 finished +---------------------------------- --------------- +epoch -510 +replay_buffer/size 999033 +trainer/num train calls 491000 +trainer/Policy Loss -19.9753 +trainer/Log Pis Mean 24.4217 +trainer/Log Pis Std 13.4111 +trainer/Log Pis Max 74.7953 +trainer/Log Pis Min -5.65312 +trainer/policy/mean Mean -0.0383619 +trainer/policy/mean Std 0.90962 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.82257 +trainer/policy/normal/std Std 0.654509 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.292585 +trainer/policy/normal/log_std Mean 0.998854 +trainer/policy/normal/log_std Std 0.314763 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.229 +eval/num steps total 490470 +eval/num paths total 491 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0874479 +eval/Actions Std 0.917529 +eval/Actions Max 0.999996 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.37867 +time/logging (s) 0.00388068 +time/sampling batch (s) 0.529489 +time/saving (s) 0.00404877 +time/training (s) 6.56923 +time/epoch (s) 9.48532 +time/total (s) 5016.58 +Epoch -510 +---------------------------------- --------------- +2022-05-10 14:34:29.016904 PDT | [2] Epoch -509 finished +---------------------------------- --------------- +epoch -509 +replay_buffer/size 999033 +trainer/num train calls 492000 +trainer/Policy Loss -19.8794 +trainer/Log Pis Mean 24.8357 +trainer/Log Pis Std 13.4411 +trainer/Log Pis Max 71.9687 +trainer/Log Pis Min -11.1231 +trainer/policy/mean Mean -0.0312165 +trainer/policy/mean Std 0.904389 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.88073 +trainer/policy/normal/std Std 0.649466 +trainer/policy/normal/std Max 5.82129 +trainer/policy/normal/std Min 0.27002 +trainer/policy/normal/log_std Mean 1.02169 +trainer/policy/normal/log_std Std 0.304432 +trainer/policy/normal/log_std Max 1.76152 +trainer/policy/normal/log_std Min -1.30926 +eval/num steps total 491330 +eval/num paths total 492 +eval/path length Mean 860 +eval/path length Std 0 +eval/path length Max 860 +eval/path length Min 860 +eval/Rewards Mean 0.00116279 +eval/Rewards Std 0.0340799 +eval/Rewards Max 1 +eval/Rewards Min 0 +eval/Returns Mean 1 +eval/Returns Std 0 +eval/Returns Max 1 +eval/Returns Min 1 +eval/Actions Mean -0.0117187 +eval/Actions Std 0.898242 +eval/Actions Max 0.99999 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 1 +eval/normalized_score 100 +time/evaluation sampling (s) 2.55107 +time/logging (s) 0.00334431 +time/sampling batch (s) 0.278465 +time/saving (s) 0.00347858 +time/training (s) 7.35327 +time/epoch (s) 10.1896 +time/total (s) 5026.77 +Epoch -509 +---------------------------------- --------------- +2022-05-10 14:34:38.627859 PDT | [2] Epoch -508 finished +---------------------------------- --------------- +epoch -508 +replay_buffer/size 999033 +trainer/num train calls 493000 +trainer/Policy Loss -18.3385 +trainer/Log Pis Mean 23.2063 +trainer/Log Pis Std 13.962 +trainer/Log Pis Max 72.4713 +trainer/Log Pis Min -7.86024 +trainer/policy/mean Mean -0.0154512 +trainer/policy/mean Std 0.901018 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.84333 +trainer/policy/normal/std Std 0.655107 +trainer/policy/normal/std Max 6.47543 +trainer/policy/normal/std Min 0.332997 +trainer/policy/normal/log_std Mean 1.00804 +trainer/policy/normal/log_std Std 0.303845 +trainer/policy/normal/log_std Max 1.86802 +trainer/policy/normal/log_std Min -1.09962 +eval/num steps total 492330 +eval/num paths total 493 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.100657 +eval/Actions Std 0.893322 +eval/Actions Max 0.999992 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.44962 +time/logging (s) 0.00369701 +time/sampling batch (s) 0.277975 +time/saving (s) 0.00342784 +time/training (s) 6.85443 +time/epoch (s) 9.58915 +time/total (s) 5036.36 +Epoch -508 +---------------------------------- --------------- +2022-05-10 14:34:48.451644 PDT | [2] Epoch -507 finished +---------------------------------- --------------- +epoch -507 +replay_buffer/size 999033 +trainer/num train calls 494000 +trainer/Policy Loss -17.9001 +trainer/Log Pis Mean 24.2274 +trainer/Log Pis Std 13.816 +trainer/Log Pis Max 66.1784 +trainer/Log Pis Min -6.95184 +trainer/policy/mean Mean -0.0435505 +trainer/policy/mean Std 0.903882 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.80702 +trainer/policy/normal/std Std 0.641943 +trainer/policy/normal/std Max 5.58666 +trainer/policy/normal/std Min 0.354893 +trainer/policy/normal/log_std Mean 0.994661 +trainer/policy/normal/log_std Std 0.308441 +trainer/policy/normal/log_std Max 1.72038 +trainer/policy/normal/log_std Min -1.03594 +eval/num steps total 493330 +eval/num paths total 494 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.335107 +eval/Actions Std 0.893118 +eval/Actions Max 0.999995 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59677 +time/logging (s) 0.00368756 +time/sampling batch (s) 0.278336 +time/saving (s) 0.00342985 +time/training (s) 6.91936 +time/epoch (s) 9.80158 +time/total (s) 5046.17 +Epoch -507 +---------------------------------- --------------- +2022-05-10 14:34:58.730654 PDT | [2] Epoch -506 finished +---------------------------------- --------------- +epoch -506 +replay_buffer/size 999033 +trainer/num train calls 495000 +trainer/Policy Loss -17.9601 +trainer/Log Pis Mean 24.3548 +trainer/Log Pis Std 13.5575 +trainer/Log Pis Max 68.8889 +trainer/Log Pis Min -10.9344 +trainer/policy/mean Mean -0.0319017 +trainer/policy/mean Std 0.90337 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.8118 +trainer/policy/normal/std Std 0.66504 +trainer/policy/normal/std Max 6.23924 +trainer/policy/normal/std Min 0.223962 +trainer/policy/normal/log_std Mean 0.992584 +trainer/policy/normal/log_std Std 0.326727 +trainer/policy/normal/log_std Max 1.83086 +trainer/policy/normal/log_std Min -1.49628 +eval/num steps total 494330 +eval/num paths total 495 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.120702 +eval/Actions Std 0.907656 +eval/Actions Max 0.99999 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64862 +time/logging (s) 0.0037302 +time/sampling batch (s) 0.529349 +time/saving (s) 0.00343526 +time/training (s) 7.07177 +time/epoch (s) 10.2569 +time/total (s) 5056.43 +Epoch -506 +---------------------------------- --------------- +2022-05-10 14:35:09.280175 PDT | [2] Epoch -505 finished +---------------------------------- --------------- +epoch -505 +replay_buffer/size 999033 +trainer/num train calls 496000 +trainer/Policy Loss -19.1099 +trainer/Log Pis Mean 24.3922 +trainer/Log Pis Std 13.2758 +trainer/Log Pis Max 64.9215 +trainer/Log Pis Min -12.8326 +trainer/policy/mean Mean -0.0311244 +trainer/policy/mean Std 0.903252 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.80906 +trainer/policy/normal/std Std 0.629305 +trainer/policy/normal/std Max 5.11703 +trainer/policy/normal/std Min 0.32305 +trainer/policy/normal/log_std Mean 0.995944 +trainer/policy/normal/log_std Std 0.307633 +trainer/policy/normal/log_std Max 1.63257 +trainer/policy/normal/log_std Min -1.12995 +eval/num steps total 495330 +eval/num paths total 496 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.102302 +eval/Actions Std 0.915628 +eval/Actions Max 0.999984 +eval/Actions Min -0.999986 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65193 +time/logging (s) 0.00409183 +time/sampling batch (s) 0.279125 +time/saving (s) 0.00391627 +time/training (s) 7.58856 +time/epoch (s) 10.5276 +time/total (s) 5066.96 +Epoch -505 +---------------------------------- --------------- +2022-05-10 14:35:19.479022 PDT | [2] Epoch -504 finished +---------------------------------- --------------- +epoch -504 +replay_buffer/size 999033 +trainer/num train calls 497000 +trainer/Policy Loss -20.2069 +trainer/Log Pis Mean 25.0477 +trainer/Log Pis Std 12.7165 +trainer/Log Pis Max 63.9076 +trainer/Log Pis Min -7.96344 +trainer/policy/mean Mean -0.0225782 +trainer/policy/mean Std 0.910297 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.79785 +trainer/policy/normal/std Std 0.631769 +trainer/policy/normal/std Max 5.26358 +trainer/policy/normal/std Min 0.294084 +trainer/policy/normal/log_std Mean 0.99292 +trainer/policy/normal/log_std Std 0.300379 +trainer/policy/normal/log_std Max 1.66081 +trainer/policy/normal/log_std Min -1.22389 +eval/num steps total 496330 +eval/num paths total 497 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.105366 +eval/Actions Std 0.916357 +eval/Actions Max 0.999993 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6391 +time/logging (s) 0.00374798 +time/sampling batch (s) 0.278901 +time/saving (s) 0.00358711 +time/training (s) 7.24947 +time/epoch (s) 10.1748 +time/total (s) 5077.13 +Epoch -504 +---------------------------------- --------------- +2022-05-10 14:35:30.712310 PDT | [2] Epoch -503 finished +---------------------------------- --------------- +epoch -503 +replay_buffer/size 999033 +trainer/num train calls 498000 +trainer/Policy Loss -19.7771 +trainer/Log Pis Mean 25.1038 +trainer/Log Pis Std 12.9745 +trainer/Log Pis Max 62.3749 +trainer/Log Pis Min -8.7027 +trainer/policy/mean Mean -0.047324 +trainer/policy/mean Std 0.909562 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999978 +trainer/policy/normal/std Mean 2.89575 +trainer/policy/normal/std Std 0.64819 +trainer/policy/normal/std Max 5.94214 +trainer/policy/normal/std Min 0.308498 +trainer/policy/normal/log_std Mean 1.02782 +trainer/policy/normal/log_std Std 0.299789 +trainer/policy/normal/log_std Max 1.78207 +trainer/policy/normal/log_std Min -1.17604 +eval/num steps total 497330 +eval/num paths total 498 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0233239 +eval/Actions Std 0.910919 +eval/Actions Max 0.999993 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64636 +time/logging (s) 0.00371263 +time/sampling batch (s) 0.279008 +time/saving (s) 0.00339256 +time/training (s) 8.27851 +time/epoch (s) 11.211 +time/total (s) 5088.35 +Epoch -503 +---------------------------------- --------------- +2022-05-10 14:35:40.843788 PDT | [2] Epoch -502 finished +---------------------------------- --------------- +epoch -502 +replay_buffer/size 999033 +trainer/num train calls 499000 +trainer/Policy Loss -19.8957 +trainer/Log Pis Mean 24.7024 +trainer/Log Pis Std 13.5816 +trainer/Log Pis Max 61.6537 +trainer/Log Pis Min -11.7427 +trainer/policy/mean Mean -0.047627 +trainer/policy/mean Std 0.908279 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.75481 +trainer/policy/normal/std Std 0.640303 +trainer/policy/normal/std Max 5.4498 +trainer/policy/normal/std Min 0.231974 +trainer/policy/normal/log_std Mean 0.973592 +trainer/policy/normal/log_std Std 0.319994 +trainer/policy/normal/log_std Max 1.69558 +trainer/policy/normal/log_std Min -1.46113 +eval/num steps total 498330 +eval/num paths total 499 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.174365 +eval/Actions Std 0.872008 +eval/Actions Max 0.999995 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65849 +time/logging (s) 0.0037641 +time/sampling batch (s) 0.276562 +time/saving (s) 0.00348096 +time/training (s) 7.16722 +time/epoch (s) 10.1095 +time/total (s) 5098.46 +Epoch -502 +---------------------------------- --------------- +2022-05-10 14:35:51.067470 PDT | [2] Epoch -501 finished +---------------------------------- --------------- +epoch -501 +replay_buffer/size 999033 +trainer/num train calls 500000 +trainer/Policy Loss -17.7178 +trainer/Log Pis Mean 24.2642 +trainer/Log Pis Std 13.2932 +trainer/Log Pis Max 69.1473 +trainer/Log Pis Min -4.05376 +trainer/policy/mean Mean -0.0340054 +trainer/policy/mean Std 0.906276 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.81782 +trainer/policy/normal/std Std 0.620092 +trainer/policy/normal/std Max 5.24732 +trainer/policy/normal/std Min 0.340067 +trainer/policy/normal/log_std Mean 1.00111 +trainer/policy/normal/log_std Std 0.297327 +trainer/policy/normal/log_std Max 1.65772 +trainer/policy/normal/log_std Min -1.07861 +eval/num steps total 499330 +eval/num paths total 500 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0989558 +eval/Actions Std 0.906832 +eval/Actions Max 0.999992 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.78007 +time/logging (s) 0.00375636 +time/sampling batch (s) 0.274914 +time/saving (s) 0.00626755 +time/training (s) 7.13673 +time/epoch (s) 10.2017 +time/total (s) 5108.67 +Epoch -501 +---------------------------------- --------------- +2022-05-10 14:36:01.797030 PDT | [2] Epoch -500 finished +---------------------------------- --------------- +epoch -500 +replay_buffer/size 999033 +trainer/num train calls 501000 +trainer/Policy Loss -19.2216 +trainer/Log Pis Mean 24.5282 +trainer/Log Pis Std 13.1808 +trainer/Log Pis Max 62.2192 +trainer/Log Pis Min -6.15841 +trainer/policy/mean Mean -0.053896 +trainer/policy/mean Std 0.905542 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.88313 +trainer/policy/normal/std Std 0.65547 +trainer/policy/normal/std Max 6.11717 +trainer/policy/normal/std Min 0.351367 +trainer/policy/normal/log_std Mean 1.02186 +trainer/policy/normal/log_std Std 0.306663 +trainer/policy/normal/log_std Max 1.8111 +trainer/policy/normal/log_std Min -1.04592 +eval/num steps total 500330 +eval/num paths total 501 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.168212 +eval/Actions Std 0.919081 +eval/Actions Max 0.999995 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62593 +time/logging (s) 0.00412761 +time/sampling batch (s) 0.274574 +time/saving (s) 0.00404054 +time/training (s) 7.7991 +time/epoch (s) 10.7078 +time/total (s) 5119.38 +Epoch -500 +---------------------------------- --------------- +2022-05-10 14:36:12.911963 PDT | [2] Epoch -499 finished +---------------------------------- --------------- +epoch -499 +replay_buffer/size 999033 +trainer/num train calls 502000 +trainer/Policy Loss -20.0589 +trainer/Log Pis Mean 24.5934 +trainer/Log Pis Std 13.1721 +trainer/Log Pis Max 75.4237 +trainer/Log Pis Min -3.60974 +trainer/policy/mean Mean -0.0243326 +trainer/policy/mean Std 0.90745 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.79147 +trainer/policy/normal/std Std 0.640311 +trainer/policy/normal/std Max 5.74235 +trainer/policy/normal/std Min 0.345688 +trainer/policy/normal/log_std Mean 0.987712 +trainer/policy/normal/log_std Std 0.316456 +trainer/policy/normal/log_std Max 1.74787 +trainer/policy/normal/log_std Min -1.06222 +eval/num steps total 501330 +eval/num paths total 502 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.147098 +eval/Actions Std 0.91888 +eval/Actions Max 0.999979 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.72 +time/logging (s) 0.00373175 +time/sampling batch (s) 0.525917 +time/saving (s) 0.00338174 +time/training (s) 7.83915 +time/epoch (s) 11.0922 +time/total (s) 5130.47 +Epoch -499 +---------------------------------- --------------- +2022-05-10 14:36:24.151960 PDT | [2] Epoch -498 finished +---------------------------------- --------------- +epoch -498 +replay_buffer/size 999033 +trainer/num train calls 503000 +trainer/Policy Loss -20.1913 +trainer/Log Pis Mean 24.734 +trainer/Log Pis Std 13.349 +trainer/Log Pis Max 62.3673 +trainer/Log Pis Min -3.81305 +trainer/policy/mean Mean -0.0333047 +trainer/policy/mean Std 0.903652 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.85625 +trainer/policy/normal/std Std 0.699006 +trainer/policy/normal/std Max 6.85822 +trainer/policy/normal/std Min 0.342819 +trainer/policy/normal/log_std Mean 1.00558 +trainer/policy/normal/log_std Std 0.3372 +trainer/policy/normal/log_std Max 1.92545 +trainer/policy/normal/log_std Min -1.07055 +eval/num steps total 502330 +eval/num paths total 503 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.123439 +eval/Actions Std 0.912261 +eval/Actions Max 0.999993 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.83129 +time/logging (s) 0.00374453 +time/sampling batch (s) 0.525457 +time/saving (s) 0.00339749 +time/training (s) 7.85421 +time/epoch (s) 11.2181 +time/total (s) 5141.69 +Epoch -498 +---------------------------------- --------------- +2022-05-10 14:36:33.403310 PDT | [2] Epoch -497 finished +---------------------------------- --------------- +epoch -497 +replay_buffer/size 999033 +trainer/num train calls 504000 +trainer/Policy Loss -20.3576 +trainer/Log Pis Mean 25.2494 +trainer/Log Pis Std 14.2084 +trainer/Log Pis Max 74.3448 +trainer/Log Pis Min -9.98593 +trainer/policy/mean Mean -0.0571557 +trainer/policy/mean Std 0.908529 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.85105 +trainer/policy/normal/std Std 0.641088 +trainer/policy/normal/std Max 4.98398 +trainer/policy/normal/std Min 0.245399 +trainer/policy/normal/log_std Mean 1.01102 +trainer/policy/normal/log_std Std 0.306478 +trainer/policy/normal/log_std Max 1.60623 +trainer/policy/normal/log_std Min -1.40487 +eval/num steps total 503085 +eval/num paths total 504 +eval/path length Mean 755 +eval/path length Std 0 +eval/path length Max 755 +eval/path length Min 755 +eval/Rewards Mean 0.0013245 +eval/Rewards Std 0.0363696 +eval/Rewards Max 1 +eval/Rewards Min 0 +eval/Returns Mean 1 +eval/Returns Std 0 +eval/Returns Max 1 +eval/Returns Min 1 +eval/Actions Mean 0.00568375 +eval/Actions Std 0.901963 +eval/Actions Max 0.999998 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 1 +eval/normalized_score 100 +time/evaluation sampling (s) 2.65209 +time/logging (s) 0.00305557 +time/sampling batch (s) 0.275016 +time/saving (s) 0.00353451 +time/training (s) 6.29486 +time/epoch (s) 9.22856 +time/total (s) 5150.92 +Epoch -497 +---------------------------------- --------------- +2022-05-10 14:36:42.651757 PDT | [2] Epoch -496 finished +---------------------------------- --------------- +epoch -496 +replay_buffer/size 999033 +trainer/num train calls 505000 +trainer/Policy Loss -20.7966 +trainer/Log Pis Mean 25.2539 +trainer/Log Pis Std 13.8047 +trainer/Log Pis Max 75.3817 +trainer/Log Pis Min -8.73347 +trainer/policy/mean Mean -0.0431827 +trainer/policy/mean Std 0.907686 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.77815 +trainer/policy/normal/std Std 0.63941 +trainer/policy/normal/std Max 6.68853 +trainer/policy/normal/std Min 0.282515 +trainer/policy/normal/log_std Mean 0.982575 +trainer/policy/normal/log_std Std 0.318399 +trainer/policy/normal/log_std Max 1.90039 +trainer/policy/normal/log_std Min -1.26402 +eval/num steps total 504085 +eval/num paths total 505 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0892525 +eval/Actions Std 0.91569 +eval/Actions Max 0.999999 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54412 +time/logging (s) 0.00410346 +time/sampling batch (s) 0.27555 +time/saving (s) 0.00394743 +time/training (s) 6.39972 +time/epoch (s) 9.22744 +time/total (s) 5160.15 +Epoch -496 +---------------------------------- --------------- +2022-05-10 14:36:51.894935 PDT | [2] Epoch -495 finished +---------------------------------- --------------- +epoch -495 +replay_buffer/size 999033 +trainer/num train calls 506000 +trainer/Policy Loss -19.8733 +trainer/Log Pis Mean 24.8137 +trainer/Log Pis Std 13.1537 +trainer/Log Pis Max 62.3222 +trainer/Log Pis Min -12.7807 +trainer/policy/mean Mean -0.0396265 +trainer/policy/mean Std 0.90361 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.82069 +trainer/policy/normal/std Std 0.658315 +trainer/policy/normal/std Max 7.11633 +trainer/policy/normal/std Min 0.283224 +trainer/policy/normal/log_std Mean 0.997257 +trainer/policy/normal/log_std Std 0.318994 +trainer/policy/normal/log_std Max 1.96239 +trainer/policy/normal/log_std Min -1.26152 +eval/num steps total 505085 +eval/num paths total 506 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0708406 +eval/Actions Std 0.912451 +eval/Actions Max 0.999996 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.72649 +time/logging (s) 0.00389877 +time/sampling batch (s) 0.528361 +time/saving (s) 0.00355293 +time/training (s) 5.95806 +time/epoch (s) 9.22036 +time/total (s) 5169.38 +Epoch -495 +---------------------------------- --------------- +2022-05-10 14:37:01.218888 PDT | [2] Epoch -494 finished +---------------------------------- --------------- +epoch -494 +replay_buffer/size 999033 +trainer/num train calls 507000 +trainer/Policy Loss -19.2485 +trainer/Log Pis Mean 24.8462 +trainer/Log Pis Std 13.1146 +trainer/Log Pis Max 68.8691 +trainer/Log Pis Min -6.11568 +trainer/policy/mean Mean -0.024981 +trainer/policy/mean Std 0.907029 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.79853 +trainer/policy/normal/std Std 0.636192 +trainer/policy/normal/std Max 6.77491 +trainer/policy/normal/std Min 0.304996 +trainer/policy/normal/log_std Mean 0.993774 +trainer/policy/normal/log_std Std 0.295525 +trainer/policy/normal/log_std Max 1.91323 +trainer/policy/normal/log_std Min -1.18746 +eval/num steps total 506085 +eval/num paths total 507 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0821183 +eval/Actions Std 0.912941 +eval/Actions Max 0.999997 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6986 +time/logging (s) 0.00380564 +time/sampling batch (s) 0.276147 +time/saving (s) 0.00347708 +time/training (s) 6.31968 +time/epoch (s) 9.30171 +time/total (s) 5178.68 +Epoch -494 +---------------------------------- --------------- +2022-05-10 14:37:11.661201 PDT | [2] Epoch -493 finished +---------------------------------- --------------- +epoch -493 +replay_buffer/size 999033 +trainer/num train calls 508000 +trainer/Policy Loss -20.0924 +trainer/Log Pis Mean 23.9572 +trainer/Log Pis Std 13.2012 +trainer/Log Pis Max 68.4257 +trainer/Log Pis Min -7.76863 +trainer/policy/mean Mean -0.0348016 +trainer/policy/mean Std 0.904216 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.82905 +trainer/policy/normal/std Std 0.649072 +trainer/policy/normal/std Max 5.74901 +trainer/policy/normal/std Min 0.383423 +trainer/policy/normal/log_std Mean 1.00302 +trainer/policy/normal/log_std Std 0.303576 +trainer/policy/normal/log_std Max 1.74903 +trainer/policy/normal/log_std Min -0.958616 +eval/num steps total 507085 +eval/num paths total 508 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.440302 +eval/Actions Std 0.801508 +eval/Actions Max 0.999988 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64089 +time/logging (s) 0.00384486 +time/sampling batch (s) 0.530662 +time/saving (s) 0.00344883 +time/training (s) 7.24111 +time/epoch (s) 10.42 +time/total (s) 5189.11 +Epoch -493 +---------------------------------- --------------- +2022-05-10 14:37:21.944812 PDT | [2] Epoch -492 finished +---------------------------------- --------------- +epoch -492 +replay_buffer/size 999033 +trainer/num train calls 509000 +trainer/Policy Loss -20.1552 +trainer/Log Pis Mean 24.0206 +trainer/Log Pis Std 12.7118 +trainer/Log Pis Max 71.8997 +trainer/Log Pis Min -7.34496 +trainer/policy/mean Mean -0.0288423 +trainer/policy/mean Std 0.909054 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.81142 +trainer/policy/normal/std Std 0.647896 +trainer/policy/normal/std Max 6.7102 +trainer/policy/normal/std Min 0.326699 +trainer/policy/normal/log_std Mean 0.996745 +trainer/policy/normal/log_std Std 0.304489 +trainer/policy/normal/log_std Max 1.90363 +trainer/policy/normal/log_std Min -1.11872 +eval/num steps total 508085 +eval/num paths total 509 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0541936 +eval/Actions Std 0.910141 +eval/Actions Max 0.999995 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6847 +time/logging (s) 0.00380176 +time/sampling batch (s) 0.281163 +time/saving (s) 0.0033895 +time/training (s) 7.28772 +time/epoch (s) 10.2608 +time/total (s) 5199.37 +Epoch -492 +---------------------------------- --------------- +2022-05-10 14:37:32.291365 PDT | [2] Epoch -491 finished +---------------------------------- --------------- +epoch -491 +replay_buffer/size 999033 +trainer/num train calls 510000 +trainer/Policy Loss -19.4819 +trainer/Log Pis Mean 24.6858 +trainer/Log Pis Std 14.171 +trainer/Log Pis Max 70.4591 +trainer/Log Pis Min -11.0426 +trainer/policy/mean Mean -0.0344934 +trainer/policy/mean Std 0.904111 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.83517 +trainer/policy/normal/std Std 0.677017 +trainer/policy/normal/std Max 5.52964 +trainer/policy/normal/std Min 0.242653 +trainer/policy/normal/log_std Mean 1.00007 +trainer/policy/normal/log_std Std 0.328992 +trainer/policy/normal/log_std Max 1.71012 +trainer/policy/normal/log_std Min -1.41612 +eval/num steps total 509085 +eval/num paths total 510 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.129009 +eval/Actions Std 0.900796 +eval/Actions Max 0.999996 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.7589 +time/logging (s) 0.00387295 +time/sampling batch (s) 0.55877 +time/saving (s) 0.00364512 +time/training (s) 6.99876 +time/epoch (s) 10.3239 +time/total (s) 5209.7 +Epoch -491 +---------------------------------- --------------- +2022-05-10 14:37:42.908386 PDT | [2] Epoch -490 finished +---------------------------------- --------------- +epoch -490 +replay_buffer/size 999033 +trainer/num train calls 511000 +trainer/Policy Loss -20.2745 +trainer/Log Pis Mean 24.4596 +trainer/Log Pis Std 13.2793 +trainer/Log Pis Max 76.1614 +trainer/Log Pis Min -9.5501 +trainer/policy/mean Mean -0.0216618 +trainer/policy/mean Std 0.908707 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.8008 +trainer/policy/normal/std Std 0.638937 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.304718 +trainer/policy/normal/log_std Mean 0.993335 +trainer/policy/normal/log_std Std 0.303364 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.18837 +eval/num steps total 510085 +eval/num paths total 511 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.180451 +eval/Actions Std 0.955236 +eval/Actions Max 0.999995 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 3.0092 +time/logging (s) 0.00368051 +time/sampling batch (s) 0.785032 +time/saving (s) 0.00345028 +time/training (s) 6.79263 +time/epoch (s) 10.594 +time/total (s) 5220.29 +Epoch -490 +---------------------------------- --------------- +2022-05-10 14:37:53.898188 PDT | [2] Epoch -489 finished +---------------------------------- --------------- +epoch -489 +replay_buffer/size 999033 +trainer/num train calls 512000 +trainer/Policy Loss -19.3008 +trainer/Log Pis Mean 25.2398 +trainer/Log Pis Std 13.1985 +trainer/Log Pis Max 66.1036 +trainer/Log Pis Min -7.09204 +trainer/policy/mean Mean -0.031597 +trainer/policy/mean Std 0.90671 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.78313 +trainer/policy/normal/std Std 0.627015 +trainer/policy/normal/std Max 5.71709 +trainer/policy/normal/std Min 0.263829 +trainer/policy/normal/log_std Mean 0.986831 +trainer/policy/normal/log_std Std 0.306156 +trainer/policy/normal/log_std Max 1.74346 +trainer/policy/normal/log_std Min -1.33245 +eval/num steps total 511085 +eval/num paths total 512 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0916522 +eval/Actions Std 0.89759 +eval/Actions Max 0.999999 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.7368 +time/logging (s) 0.00392505 +time/sampling batch (s) 0.282261 +time/saving (s) 0.00386342 +time/training (s) 7.94061 +time/epoch (s) 10.9675 +time/total (s) 5231.26 +Epoch -489 +---------------------------------- --------------- +2022-05-10 14:38:04.703368 PDT | [2] Epoch -488 finished +---------------------------------- --------------- +epoch -488 +replay_buffer/size 999033 +trainer/num train calls 513000 +trainer/Policy Loss -19.7672 +trainer/Log Pis Mean 24.1405 +trainer/Log Pis Std 13.4901 +trainer/Log Pis Max 67.3585 +trainer/Log Pis Min -9.24603 +trainer/policy/mean Mean -0.047792 +trainer/policy/mean Std 0.903355 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.8087 +trainer/policy/normal/std Std 0.629489 +trainer/policy/normal/std Max 5.65102 +trainer/policy/normal/std Min 0.284035 +trainer/policy/normal/log_std Mean 0.99786 +trainer/policy/normal/log_std Std 0.295351 +trainer/policy/normal/log_std Max 1.73184 +trainer/policy/normal/log_std Min -1.25866 +eval/num steps total 512085 +eval/num paths total 513 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.217585 +eval/Actions Std 0.848781 +eval/Actions Max 0.999989 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69048 +time/logging (s) 0.00370068 +time/sampling batch (s) 0.281092 +time/saving (s) 0.00344657 +time/training (s) 7.80338 +time/epoch (s) 10.7821 +time/total (s) 5242.05 +Epoch -488 +---------------------------------- --------------- +2022-05-10 14:38:15.461360 PDT | [2] Epoch -487 finished +---------------------------------- --------------- +epoch -487 +replay_buffer/size 999033 +trainer/num train calls 514000 +trainer/Policy Loss -19.8327 +trainer/Log Pis Mean 24.4373 +trainer/Log Pis Std 13.9407 +trainer/Log Pis Max 74.1313 +trainer/Log Pis Min -12.6333 +trainer/policy/mean Mean -0.0590967 +trainer/policy/mean Std 0.906466 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.84648 +trainer/policy/normal/std Std 0.67932 +trainer/policy/normal/std Max 6.62641 +trainer/policy/normal/std Min 0.237281 +trainer/policy/normal/log_std Mean 1.00495 +trainer/policy/normal/log_std Std 0.324988 +trainer/policy/normal/log_std Max 1.89106 +trainer/policy/normal/log_std Min -1.43851 +eval/num steps total 513085 +eval/num paths total 514 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.101787 +eval/Actions Std 0.911034 +eval/Actions Max 0.999993 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6634 +time/logging (s) 0.00369194 +time/sampling batch (s) 0.279648 +time/saving (s) 0.00342102 +time/training (s) 7.78545 +time/epoch (s) 10.7356 +time/total (s) 5252.79 +Epoch -487 +---------------------------------- --------------- +2022-05-10 14:38:26.314423 PDT | [2] Epoch -486 finished +---------------------------------- --------------- +epoch -486 +replay_buffer/size 999033 +trainer/num train calls 515000 +trainer/Policy Loss -19.6724 +trainer/Log Pis Mean 24.8095 +trainer/Log Pis Std 14.0617 +trainer/Log Pis Max 74.2459 +trainer/Log Pis Min -8.76159 +trainer/policy/mean Mean -0.0495381 +trainer/policy/mean Std 0.910738 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.84497 +trainer/policy/normal/std Std 0.664397 +trainer/policy/normal/std Max 7.29382 +trainer/policy/normal/std Min 0.310138 +trainer/policy/normal/log_std Mean 1.00709 +trainer/policy/normal/log_std Std 0.312614 +trainer/policy/normal/log_std Max 1.98703 +trainer/policy/normal/log_std Min -1.17074 +eval/num steps total 514085 +eval/num paths total 515 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.235358 +eval/Actions Std 0.934479 +eval/Actions Max 0.999961 +eval/Actions Min -0.999985 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55125 +time/logging (s) 0.00369678 +time/sampling batch (s) 0.527591 +time/saving (s) 0.00337548 +time/training (s) 7.74514 +time/epoch (s) 10.831 +time/total (s) 5263.62 +Epoch -486 +---------------------------------- --------------- +2022-05-10 14:38:36.978169 PDT | [2] Epoch -485 finished +---------------------------------- --------------- +epoch -485 +replay_buffer/size 999033 +trainer/num train calls 516000 +trainer/Policy Loss -19.8319 +trainer/Log Pis Mean 24.9651 +trainer/Log Pis Std 13.1503 +trainer/Log Pis Max 62.2453 +trainer/Log Pis Min -7.61659 +trainer/policy/mean Mean -0.0545888 +trainer/policy/mean Std 0.903765 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.82564 +trainer/policy/normal/std Std 0.629106 +trainer/policy/normal/std Max 5.91906 +trainer/policy/normal/std Min 0.335687 +trainer/policy/normal/log_std Mean 1.00391 +trainer/policy/normal/log_std Std 0.295295 +trainer/policy/normal/log_std Max 1.77818 +trainer/policy/normal/log_std Min -1.09158 +eval/num steps total 515085 +eval/num paths total 516 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.134482 +eval/Actions Std 0.906305 +eval/Actions Max 0.999995 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63198 +time/logging (s) 0.00371587 +time/sampling batch (s) 0.531311 +time/saving (s) 0.00355219 +time/training (s) 7.47065 +time/epoch (s) 10.6412 +time/total (s) 5274.27 +Epoch -485 +---------------------------------- --------------- +2022-05-10 14:38:47.209138 PDT | [2] Epoch -484 finished +---------------------------------- --------------- +epoch -484 +replay_buffer/size 999033 +trainer/num train calls 517000 +trainer/Policy Loss -19.7659 +trainer/Log Pis Mean 23.2676 +trainer/Log Pis Std 13.3209 +trainer/Log Pis Max 77.3626 +trainer/Log Pis Min -9.06743 +trainer/policy/mean Mean -0.0222998 +trainer/policy/mean Std 0.900519 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999985 +trainer/policy/normal/std Mean 2.77928 +trainer/policy/normal/std Std 0.639593 +trainer/policy/normal/std Max 5.1833 +trainer/policy/normal/std Min 0.316137 +trainer/policy/normal/log_std Mean 0.984446 +trainer/policy/normal/log_std Std 0.308632 +trainer/policy/normal/log_std Max 1.64544 +trainer/policy/normal/log_std Min -1.15158 +eval/num steps total 516085 +eval/num paths total 517 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.432733 +eval/Actions Std 0.886965 +eval/Actions Max 0.999995 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56659 +time/logging (s) 0.00367576 +time/sampling batch (s) 0.27653 +time/saving (s) 0.00342534 +time/training (s) 7.35842 +time/epoch (s) 10.2086 +time/total (s) 5284.48 +Epoch -484 +---------------------------------- --------------- +2022-05-10 14:38:57.531931 PDT | [2] Epoch -483 finished +---------------------------------- --------------- +epoch -483 +replay_buffer/size 999033 +trainer/num train calls 518000 +trainer/Policy Loss -19.3625 +trainer/Log Pis Mean 24.1682 +trainer/Log Pis Std 13.1459 +trainer/Log Pis Max 70.953 +trainer/Log Pis Min -10.748 +trainer/policy/mean Mean -0.0332846 +trainer/policy/mean Std 0.905456 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.75241 +trainer/policy/normal/std Std 0.640789 +trainer/policy/normal/std Max 6.25726 +trainer/policy/normal/std Min 0.313194 +trainer/policy/normal/log_std Mean 0.973053 +trainer/policy/normal/log_std Std 0.318539 +trainer/policy/normal/log_std Max 1.83374 +trainer/policy/normal/log_std Min -1.16093 +eval/num steps total 517085 +eval/num paths total 518 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0413144 +eval/Actions Std 0.85828 +eval/Actions Max 0.99999 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66561 +time/logging (s) 0.00413822 +time/sampling batch (s) 0.274971 +time/saving (s) 0.00394464 +time/training (s) 7.35202 +time/epoch (s) 10.3007 +time/total (s) 5294.78 +Epoch -483 +---------------------------------- --------------- +2022-05-10 14:39:07.723665 PDT | [2] Epoch -482 finished +---------------------------------- --------------- +epoch -482 +replay_buffer/size 999033 +trainer/num train calls 519000 +trainer/Policy Loss -18.9025 +trainer/Log Pis Mean 23.9777 +trainer/Log Pis Std 13.0452 +trainer/Log Pis Max 64.7421 +trainer/Log Pis Min -7.5634 +trainer/policy/mean Mean -0.0325919 +trainer/policy/mean Std 0.906575 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.82648 +trainer/policy/normal/std Std 0.64966 +trainer/policy/normal/std Max 6.12481 +trainer/policy/normal/std Min 0.216049 +trainer/policy/normal/log_std Mean 1.00032 +trainer/policy/normal/log_std Std 0.315533 +trainer/policy/normal/log_std Max 1.81235 +trainer/policy/normal/log_std Min -1.53225 +eval/num steps total 518085 +eval/num paths total 519 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.230723 +eval/Actions Std 0.929115 +eval/Actions Max 0.999995 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5664 +time/logging (s) 0.00387753 +time/sampling batch (s) 0.275955 +time/saving (s) 0.00360595 +time/training (s) 7.31908 +time/epoch (s) 10.1689 +time/total (s) 5304.95 +Epoch -482 +---------------------------------- --------------- +2022-05-10 14:39:17.975040 PDT | [2] Epoch -481 finished +---------------------------------- --------------- +epoch -481 +replay_buffer/size 999033 +trainer/num train calls 520000 +trainer/Policy Loss -20.5987 +trainer/Log Pis Mean 25.2201 +trainer/Log Pis Std 13.0491 +trainer/Log Pis Max 62.6401 +trainer/Log Pis Min -5.08897 +trainer/policy/mean Mean -0.0480551 +trainer/policy/mean Std 0.909714 +trainer/policy/mean Max 0.999979 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.84411 +trainer/policy/normal/std Std 0.632971 +trainer/policy/normal/std Max 5.72215 +trainer/policy/normal/std Min 0.249011 +trainer/policy/normal/log_std Mean 1.00961 +trainer/policy/normal/log_std Std 0.301692 +trainer/policy/normal/log_std Max 1.74434 +trainer/policy/normal/log_std Min -1.39026 +eval/num steps total 519085 +eval/num paths total 520 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0444408 +eval/Actions Std 0.914928 +eval/Actions Max 0.999991 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59952 +time/logging (s) 0.00370079 +time/sampling batch (s) 0.274338 +time/saving (s) 0.00342493 +time/training (s) 7.34809 +time/epoch (s) 10.2291 +time/total (s) 5315.19 +Epoch -481 +---------------------------------- --------------- +2022-05-10 14:39:28.153990 PDT | [2] Epoch -480 finished +---------------------------------- --------------- +epoch -480 +replay_buffer/size 999033 +trainer/num train calls 521000 +trainer/Policy Loss -20.0752 +trainer/Log Pis Mean 24.5445 +trainer/Log Pis Std 13.7186 +trainer/Log Pis Max 66.2495 +trainer/Log Pis Min -9.41356 +trainer/policy/mean Mean -0.0206953 +trainer/policy/mean Std 0.908012 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.84294 +trainer/policy/normal/std Std 0.62748 +trainer/policy/normal/std Max 5.98819 +trainer/policy/normal/std Min 0.31581 +trainer/policy/normal/log_std Mean 1.01196 +trainer/policy/normal/log_std Std 0.283798 +trainer/policy/normal/log_std Max 1.78979 +trainer/policy/normal/log_std Min -1.15262 +eval/num steps total 520085 +eval/num paths total 521 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00630211 +eval/Actions Std 0.907494 +eval/Actions Max 0.999995 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.27686 +time/logging (s) 0.00416643 +time/sampling batch (s) 0.286592 +time/saving (s) 0.0035907 +time/training (s) 7.58546 +time/epoch (s) 10.1567 +time/total (s) 5325.34 +Epoch -480 +---------------------------------- --------------- +2022-05-10 14:39:38.932705 PDT | [2] Epoch -479 finished +---------------------------------- --------------- +epoch -479 +replay_buffer/size 999033 +trainer/num train calls 522000 +trainer/Policy Loss -20.5725 +trainer/Log Pis Mean 24.3524 +trainer/Log Pis Std 12.9626 +trainer/Log Pis Max 74.9104 +trainer/Log Pis Min -6.54307 +trainer/policy/mean Mean -0.0122544 +trainer/policy/mean Std 0.910081 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.83188 +trainer/policy/normal/std Std 0.660901 +trainer/policy/normal/std Max 6.92462 +trainer/policy/normal/std Min 0.253206 +trainer/policy/normal/log_std Mean 1.00143 +trainer/policy/normal/log_std Std 0.318656 +trainer/policy/normal/log_std Max 1.93508 +trainer/policy/normal/log_std Min -1.37355 +eval/num steps total 521085 +eval/num paths total 522 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0617521 +eval/Actions Std 0.90964 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65 +time/logging (s) 0.00380734 +time/sampling batch (s) 0.285573 +time/saving (s) 0.00342396 +time/training (s) 7.81171 +time/epoch (s) 10.7545 +time/total (s) 5336.1 +Epoch -479 +---------------------------------- --------------- +2022-05-10 14:39:48.627836 PDT | [2] Epoch -478 finished +---------------------------------- --------------- +epoch -478 +replay_buffer/size 999033 +trainer/num train calls 523000 +trainer/Policy Loss -19.934 +trainer/Log Pis Mean 23.95 +trainer/Log Pis Std 12.5674 +trainer/Log Pis Max 70.6466 +trainer/Log Pis Min -5.4852 +trainer/policy/mean Mean -0.0415504 +trainer/policy/mean Std 0.903057 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.79684 +trainer/policy/normal/std Std 0.642413 +trainer/policy/normal/std Max 6.45473 +trainer/policy/normal/std Min 0.333122 +trainer/policy/normal/log_std Mean 0.991303 +trainer/policy/normal/log_std Std 0.30586 +trainer/policy/normal/log_std Max 1.86481 +trainer/policy/normal/log_std Min -1.09925 +eval/num steps total 522085 +eval/num paths total 523 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.356284 +eval/Actions Std 0.826666 +eval/Actions Max 0.999999 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.7135 +time/logging (s) 0.00430547 +time/sampling batch (s) 0.535081 +time/saving (s) 0.00414033 +time/training (s) 6.41582 +time/epoch (s) 9.67285 +time/total (s) 5345.78 +Epoch -478 +---------------------------------- --------------- +2022-05-10 14:39:58.367776 PDT | [2] Epoch -477 finished +---------------------------------- --------------- +epoch -477 +replay_buffer/size 999033 +trainer/num train calls 524000 +trainer/Policy Loss -19.7835 +trainer/Log Pis Mean 23.8716 +trainer/Log Pis Std 13.4023 +trainer/Log Pis Max 62.3214 +trainer/Log Pis Min -12.9469 +trainer/policy/mean Mean -0.0236921 +trainer/policy/mean Std 0.907176 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.79472 +trainer/policy/normal/std Std 0.637436 +trainer/policy/normal/std Max 6.74534 +trainer/policy/normal/std Min 0.266219 +trainer/policy/normal/log_std Mean 0.991141 +trainer/policy/normal/log_std Std 0.303707 +trainer/policy/normal/log_std Max 1.90885 +trainer/policy/normal/log_std Min -1.32343 +eval/num steps total 523085 +eval/num paths total 524 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0663162 +eval/Actions Std 0.906824 +eval/Actions Max 0.99999 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59279 +time/logging (s) 0.00418081 +time/sampling batch (s) 0.28021 +time/saving (s) 0.004109 +time/training (s) 6.83536 +time/epoch (s) 9.71666 +time/total (s) 5355.5 +Epoch -477 +---------------------------------- --------------- +2022-05-10 14:40:07.538353 PDT | [2] Epoch -476 finished +---------------------------------- --------------- +epoch -476 +replay_buffer/size 999033 +trainer/num train calls 525000 +trainer/Policy Loss -19.2436 +trainer/Log Pis Mean 25.6649 +trainer/Log Pis Std 13.6598 +trainer/Log Pis Max 72.7306 +trainer/Log Pis Min -13.4247 +trainer/policy/mean Mean -0.0399727 +trainer/policy/mean Std 0.906483 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.76248 +trainer/policy/normal/std Std 0.657159 +trainer/policy/normal/std Max 4.96188 +trainer/policy/normal/std Min 0.283833 +trainer/policy/normal/log_std Mean 0.973244 +trainer/policy/normal/log_std Std 0.33499 +trainer/policy/normal/log_std Max 1.60179 +trainer/policy/normal/log_std Min -1.25937 +eval/num steps total 524085 +eval/num paths total 525 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.206195 +eval/Actions Std 0.95346 +eval/Actions Max 0.999994 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.72241 +time/logging (s) 0.00371189 +time/sampling batch (s) 0.5275 +time/saving (s) 0.00342189 +time/training (s) 5.89028 +time/epoch (s) 9.14732 +time/total (s) 5364.65 +Epoch -476 +---------------------------------- --------------- +2022-05-10 14:40:17.495019 PDT | [2] Epoch -475 finished +---------------------------------- --------------- +epoch -475 +replay_buffer/size 999033 +trainer/num train calls 526000 +trainer/Policy Loss -19.7651 +trainer/Log Pis Mean 25.0756 +trainer/Log Pis Std 13.5624 +trainer/Log Pis Max 78.1491 +trainer/Log Pis Min -7.71573 +trainer/policy/mean Mean -0.0487204 +trainer/policy/mean Std 0.903278 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.81698 +trainer/policy/normal/std Std 0.667197 +trainer/policy/normal/std Max 5.19782 +trainer/policy/normal/std Min 0.214154 +trainer/policy/normal/log_std Mean 0.99267 +trainer/policy/normal/log_std Std 0.337651 +trainer/policy/normal/log_std Max 1.64824 +trainer/policy/normal/log_std Min -1.54106 +eval/num steps total 525085 +eval/num paths total 526 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.167704 +eval/Actions Std 0.887057 +eval/Actions Max 0.999999 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.31405 +time/logging (s) 0.00388355 +time/sampling batch (s) 0.527348 +time/saving (s) 0.00353873 +time/training (s) 7.08579 +time/epoch (s) 9.93461 +time/total (s) 5374.59 +Epoch -475 +---------------------------------- --------------- +2022-05-10 14:40:26.616804 PDT | [2] Epoch -474 finished +---------------------------------- --------------- +epoch -474 +replay_buffer/size 999033 +trainer/num train calls 527000 +trainer/Policy Loss -20.0017 +trainer/Log Pis Mean 25.0421 +trainer/Log Pis Std 13.2574 +trainer/Log Pis Max 72.1462 +trainer/Log Pis Min -10.3346 +trainer/policy/mean Mean -0.0453407 +trainer/policy/mean Std 0.905679 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.82085 +trainer/policy/normal/std Std 0.651988 +trainer/policy/normal/std Max 6.22199 +trainer/policy/normal/std Min 0.311134 +trainer/policy/normal/log_std Mean 0.998433 +trainer/policy/normal/log_std Std 0.314669 +trainer/policy/normal/log_std Max 1.82809 +trainer/policy/normal/log_std Min -1.16753 +eval/num steps total 526085 +eval/num paths total 527 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.133823 +eval/Actions Std 0.891192 +eval/Actions Max 0.999993 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55632 +time/logging (s) 0.00405049 +time/sampling batch (s) 0.280539 +time/saving (s) 0.00399676 +time/training (s) 6.25435 +time/epoch (s) 9.09926 +time/total (s) 5383.69 +Epoch -474 +---------------------------------- --------------- +2022-05-10 14:40:36.074198 PDT | [2] Epoch -473 finished +---------------------------------- --------------- +epoch -473 +replay_buffer/size 999033 +trainer/num train calls 528000 +trainer/Policy Loss -19.6418 +trainer/Log Pis Mean 23.8462 +trainer/Log Pis Std 13.6805 +trainer/Log Pis Max 64.5566 +trainer/Log Pis Min -4.95276 +trainer/policy/mean Mean -0.0576615 +trainer/policy/mean Std 0.902869 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.80465 +trainer/policy/normal/std Std 0.666365 +trainer/policy/normal/std Max 6.2025 +trainer/policy/normal/std Min 0.306312 +trainer/policy/normal/log_std Mean 0.991133 +trainer/policy/normal/log_std Std 0.317092 +trainer/policy/normal/log_std Max 1.82495 +trainer/policy/normal/log_std Min -1.18315 +eval/num steps total 527085 +eval/num paths total 528 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.242788 +eval/Actions Std 0.862193 +eval/Actions Max 0.999999 +eval/Actions Min -0.999983 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51162 +time/logging (s) 0.00372183 +time/sampling batch (s) 0.280494 +time/saving (s) 0.00350649 +time/training (s) 6.6351 +time/epoch (s) 9.43444 +time/total (s) 5393.13 +Epoch -473 +---------------------------------- --------------- +2022-05-10 14:40:46.391958 PDT | [2] Epoch -472 finished +---------------------------------- --------------- +epoch -472 +replay_buffer/size 999033 +trainer/num train calls 529000 +trainer/Policy Loss -19.368 +trainer/Log Pis Mean 24.9874 +trainer/Log Pis Std 13.328 +trainer/Log Pis Max 72.2371 +trainer/Log Pis Min -5.56404 +trainer/policy/mean Mean -0.0500515 +trainer/policy/mean Std 0.901717 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.80466 +trainer/policy/normal/std Std 0.646031 +trainer/policy/normal/std Max 5.53031 +trainer/policy/normal/std Min 0.339299 +trainer/policy/normal/log_std Mean 0.992379 +trainer/policy/normal/log_std Std 0.315985 +trainer/policy/normal/log_std Max 1.71024 +trainer/policy/normal/log_std Min -1.08087 +eval/num steps total 528085 +eval/num paths total 529 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.099753 +eval/Actions Std 0.911068 +eval/Actions Max 0.999993 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53915 +time/logging (s) 0.00370526 +time/sampling batch (s) 0.529986 +time/saving (s) 0.00345975 +time/training (s) 7.21908 +time/epoch (s) 10.2954 +time/total (s) 5403.42 +Epoch -472 +---------------------------------- --------------- +2022-05-10 14:40:56.248015 PDT | [2] Epoch -471 finished +---------------------------------- --------------- +epoch -471 +replay_buffer/size 999033 +trainer/num train calls 530000 +trainer/Policy Loss -20.376 +trainer/Log Pis Mean 24.2085 +trainer/Log Pis Std 12.843 +trainer/Log Pis Max 69.9594 +trainer/Log Pis Min -8.29251 +trainer/policy/mean Mean -0.0275109 +trainer/policy/mean Std 0.906841 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.85766 +trainer/policy/normal/std Std 0.66192 +trainer/policy/normal/std Max 5.53482 +trainer/policy/normal/std Min 0.346516 +trainer/policy/normal/log_std Mean 1.01224 +trainer/policy/normal/log_std Std 0.307602 +trainer/policy/normal/log_std Max 1.71106 +trainer/policy/normal/log_std Min -1.05983 +eval/num steps total 529085 +eval/num paths total 530 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0341456 +eval/Actions Std 0.909823 +eval/Actions Max 0.999999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66208 +time/logging (s) 0.00421202 +time/sampling batch (s) 0.28684 +time/saving (s) 0.00412253 +time/training (s) 6.8764 +time/epoch (s) 9.83365 +time/total (s) 5413.26 +Epoch -471 +---------------------------------- --------------- +2022-05-10 14:41:07.111545 PDT | [2] Epoch -470 finished +---------------------------------- --------------- +epoch -470 +replay_buffer/size 999033 +trainer/num train calls 531000 +trainer/Policy Loss -19.3793 +trainer/Log Pis Mean 24.7532 +trainer/Log Pis Std 13.0741 +trainer/Log Pis Max 70.7472 +trainer/Log Pis Min -8.92021 +trainer/policy/mean Mean -0.0304059 +trainer/policy/mean Std 0.910284 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999984 +trainer/policy/normal/std Mean 2.83579 +trainer/policy/normal/std Std 0.644947 +trainer/policy/normal/std Max 5.61757 +trainer/policy/normal/std Min 0.298821 +trainer/policy/normal/log_std Mean 1.00542 +trainer/policy/normal/log_std Std 0.305654 +trainer/policy/normal/log_std Max 1.7259 +trainer/policy/normal/log_std Min -1.20791 +eval/num steps total 530085 +eval/num paths total 531 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0579563 +eval/Actions Std 0.911318 +eval/Actions Max 0.999997 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66549 +time/logging (s) 0.00374542 +time/sampling batch (s) 0.276232 +time/saving (s) 0.00344599 +time/training (s) 7.8914 +time/epoch (s) 10.8403 +time/total (s) 5424.1 +Epoch -470 +---------------------------------- --------------- +2022-05-10 14:41:17.452553 PDT | [2] Epoch -469 finished +---------------------------------- --------------- +epoch -469 +replay_buffer/size 999033 +trainer/num train calls 532000 +trainer/Policy Loss -19.205 +trainer/Log Pis Mean 25.2293 +trainer/Log Pis Std 13.4287 +trainer/Log Pis Max 75.4957 +trainer/Log Pis Min -7.63491 +trainer/policy/mean Mean -0.0434818 +trainer/policy/mean Std 0.905903 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.76247 +trainer/policy/normal/std Std 0.652849 +trainer/policy/normal/std Max 6.90914 +trainer/policy/normal/std Min 0.304887 +trainer/policy/normal/log_std Mean 0.976264 +trainer/policy/normal/log_std Std 0.316719 +trainer/policy/normal/log_std Max 1.93285 +trainer/policy/normal/log_std Min -1.18782 +eval/num steps total 531085 +eval/num paths total 532 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.119762 +eval/Actions Std 0.918371 +eval/Actions Max 0.999991 +eval/Actions Min -0.999985 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54899 +time/logging (s) 0.00382214 +time/sampling batch (s) 0.277436 +time/saving (s) 0.00352253 +time/training (s) 7.48494 +time/epoch (s) 10.3187 +time/total (s) 5434.43 +Epoch -469 +---------------------------------- --------------- +2022-05-10 14:41:27.469625 PDT | [2] Epoch -468 finished +---------------------------------- --------------- +epoch -468 +replay_buffer/size 999033 +trainer/num train calls 533000 +trainer/Policy Loss -20.181 +trainer/Log Pis Mean 23.9863 +trainer/Log Pis Std 13.6613 +trainer/Log Pis Max 70.9492 +trainer/Log Pis Min -6.5374 +trainer/policy/mean Mean -0.0454465 +trainer/policy/mean Std 0.905557 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.80856 +trainer/policy/normal/std Std 0.636172 +trainer/policy/normal/std Max 5.70247 +trainer/policy/normal/std Min 0.325262 +trainer/policy/normal/log_std Mean 0.996531 +trainer/policy/normal/log_std Std 0.302588 +trainer/policy/normal/log_std Max 1.7409 +trainer/policy/normal/log_std Min -1.12312 +eval/num steps total 532085 +eval/num paths total 533 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0518663 +eval/Actions Std 0.894426 +eval/Actions Max 0.999998 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68445 +time/logging (s) 0.00377423 +time/sampling batch (s) 0.281016 +time/saving (s) 0.00340142 +time/training (s) 7.02203 +time/epoch (s) 9.99467 +time/total (s) 5444.42 +Epoch -468 +---------------------------------- --------------- +2022-05-10 14:41:37.686172 PDT | [2] Epoch -467 finished +---------------------------------- --------------- +epoch -467 +replay_buffer/size 999033 +trainer/num train calls 534000 +trainer/Policy Loss -19.1683 +trainer/Log Pis Mean 24.7794 +trainer/Log Pis Std 13.6083 +trainer/Log Pis Max 70.8541 +trainer/Log Pis Min -10.0932 +trainer/policy/mean Mean -0.0364062 +trainer/policy/mean Std 0.909703 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999985 +trainer/policy/normal/std Mean 2.84548 +trainer/policy/normal/std Std 0.64383 +trainer/policy/normal/std Max 5.37791 +trainer/policy/normal/std Min 0.308125 +trainer/policy/normal/log_std Mean 1.00993 +trainer/policy/normal/log_std Std 0.299037 +trainer/policy/normal/log_std Max 1.6823 +trainer/policy/normal/log_std Min -1.17725 +eval/num steps total 533085 +eval/num paths total 534 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0478225 +eval/Actions Std 0.906971 +eval/Actions Max 0.999987 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66322 +time/logging (s) 0.00374655 +time/sampling batch (s) 0.279614 +time/saving (s) 0.00355112 +time/training (s) 7.24401 +time/epoch (s) 10.1941 +time/total (s) 5454.62 +Epoch -467 +---------------------------------- --------------- +2022-05-10 14:41:48.106680 PDT | [2] Epoch -466 finished +---------------------------------- --------------- +epoch -466 +replay_buffer/size 999033 +trainer/num train calls 535000 +trainer/Policy Loss -19.8248 +trainer/Log Pis Mean 24.6753 +trainer/Log Pis Std 13.4219 +trainer/Log Pis Max 77.9767 +trainer/Log Pis Min -10.6787 +trainer/policy/mean Mean -0.0247983 +trainer/policy/mean Std 0.909473 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.82448 +trainer/policy/normal/std Std 0.647668 +trainer/policy/normal/std Max 5.79779 +trainer/policy/normal/std Min 0.295334 +trainer/policy/normal/log_std Mean 0.999686 +trainer/policy/normal/log_std Std 0.31553 +trainer/policy/normal/log_std Max 1.75748 +trainer/policy/normal/log_std Min -1.21965 +eval/num steps total 534085 +eval/num paths total 535 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.119753 +eval/Actions Std 0.939085 +eval/Actions Max 0.999985 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54877 +time/logging (s) 0.00383719 +time/sampling batch (s) 0.283587 +time/saving (s) 0.00338907 +time/training (s) 7.55842 +time/epoch (s) 10.398 +time/total (s) 5465.02 +Epoch -466 +---------------------------------- --------------- +2022-05-10 14:41:59.230673 PDT | [2] Epoch -465 finished +---------------------------------- --------------- +epoch -465 +replay_buffer/size 999033 +trainer/num train calls 536000 +trainer/Policy Loss -20.1996 +trainer/Log Pis Mean 25.0519 +trainer/Log Pis Std 13.256 +trainer/Log Pis Max 62.4745 +trainer/Log Pis Min -9.61762 +trainer/policy/mean Mean -0.0414603 +trainer/policy/mean Std 0.909589 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.84244 +trainer/policy/normal/std Std 0.652594 +trainer/policy/normal/std Max 6.56318 +trainer/policy/normal/std Min 0.351506 +trainer/policy/normal/log_std Mean 1.00813 +trainer/policy/normal/log_std Std 0.301313 +trainer/policy/normal/log_std Max 1.88148 +trainer/policy/normal/log_std Min -1.04553 +eval/num steps total 535085 +eval/num paths total 536 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0141707 +eval/Actions Std 0.974884 +eval/Actions Max 0.999988 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64243 +time/logging (s) 0.00412021 +time/sampling batch (s) 0.528783 +time/saving (s) 0.00407264 +time/training (s) 7.92235 +time/epoch (s) 11.1018 +time/total (s) 5476.13 +Epoch -465 +---------------------------------- --------------- +2022-05-10 14:42:09.445802 PDT | [2] Epoch -464 finished +---------------------------------- --------------- +epoch -464 +replay_buffer/size 999033 +trainer/num train calls 537000 +trainer/Policy Loss -19.8757 +trainer/Log Pis Mean 24.265 +trainer/Log Pis Std 13.3151 +trainer/Log Pis Max 75.5598 +trainer/Log Pis Min -10.754 +trainer/policy/mean Mean -0.0244782 +trainer/policy/mean Std 0.904345 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.80731 +trainer/policy/normal/std Std 0.665926 +trainer/policy/normal/std Max 5.74173 +trainer/policy/normal/std Min 0.28125 +trainer/policy/normal/log_std Mean 0.991734 +trainer/policy/normal/log_std Std 0.320462 +trainer/policy/normal/log_std Max 1.74776 +trainer/policy/normal/log_std Min -1.26851 +eval/num steps total 536085 +eval/num paths total 537 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.190427 +eval/Actions Std 0.896084 +eval/Actions Max 0.999984 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60394 +time/logging (s) 0.00373943 +time/sampling batch (s) 0.275809 +time/saving (s) 0.0034305 +time/training (s) 7.30509 +time/epoch (s) 10.192 +time/total (s) 5486.32 +Epoch -464 +---------------------------------- --------------- +2022-05-10 14:42:19.711772 PDT | [2] Epoch -463 finished +---------------------------------- --------------- +epoch -463 +replay_buffer/size 999033 +trainer/num train calls 538000 +trainer/Policy Loss -18.5359 +trainer/Log Pis Mean 25.1205 +trainer/Log Pis Std 12.8886 +trainer/Log Pis Max 68.6851 +trainer/Log Pis Min -9.54974 +trainer/policy/mean Mean -0.0219058 +trainer/policy/mean Std 0.907707 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.80901 +trainer/policy/normal/std Std 0.645177 +trainer/policy/normal/std Max 5.21583 +trainer/policy/normal/std Min 0.31729 +trainer/policy/normal/log_std Mean 0.994215 +trainer/policy/normal/log_std Std 0.314851 +trainer/policy/normal/log_std Max 1.6517 +trainer/policy/normal/log_std Min -1.14794 +eval/num steps total 537085 +eval/num paths total 538 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.124799 +eval/Actions Std 0.888613 +eval/Actions Max 0.999995 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.4879 +time/logging (s) 0.00379412 +time/sampling batch (s) 0.276035 +time/saving (s) 0.00350768 +time/training (s) 7.4724 +time/epoch (s) 10.2436 +time/total (s) 5496.57 +Epoch -463 +---------------------------------- --------------- +2022-05-10 14:42:30.358681 PDT | [2] Epoch -462 finished +---------------------------------- --------------- +epoch -462 +replay_buffer/size 999033 +trainer/num train calls 539000 +trainer/Policy Loss -19.9308 +trainer/Log Pis Mean 26.0454 +trainer/Log Pis Std 13.7069 +trainer/Log Pis Max 83.2706 +trainer/Log Pis Min -10.2238 +trainer/policy/mean Mean -0.0507193 +trainer/policy/mean Std 0.910096 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.82845 +trainer/policy/normal/std Std 0.629321 +trainer/policy/normal/std Max 5.42704 +trainer/policy/normal/std Min 0.271772 +trainer/policy/normal/log_std Mean 1.00436 +trainer/policy/normal/log_std Std 0.299358 +trainer/policy/normal/log_std Max 1.69139 +trainer/policy/normal/log_std Min -1.30279 +eval/num steps total 538085 +eval/num paths total 539 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.342349 +eval/Actions Std 0.902957 +eval/Actions Max 0.999995 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67687 +time/logging (s) 0.00371785 +time/sampling batch (s) 0.525884 +time/saving (s) 0.00334348 +time/training (s) 7.41478 +time/epoch (s) 10.6246 +time/total (s) 5507.2 +Epoch -462 +---------------------------------- --------------- +2022-05-10 14:42:40.397563 PDT | [2] Epoch -461 finished +---------------------------------- --------------- +epoch -461 +replay_buffer/size 999033 +trainer/num train calls 540000 +trainer/Policy Loss -20.1068 +trainer/Log Pis Mean 25.2572 +trainer/Log Pis Std 13.3616 +trainer/Log Pis Max 81.3271 +trainer/Log Pis Min -6.77958 +trainer/policy/mean Mean -0.0360783 +trainer/policy/mean Std 0.904308 +trainer/policy/mean Max 1 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.84378 +trainer/policy/normal/std Std 0.666481 +trainer/policy/normal/std Max 6.51649 +trainer/policy/normal/std Min 0.336647 +trainer/policy/normal/log_std Mean 1.00655 +trainer/policy/normal/log_std Std 0.312353 +trainer/policy/normal/log_std Max 1.87434 +trainer/policy/normal/log_std Min -1.08872 +eval/num steps total 539085 +eval/num paths total 540 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.211231 +eval/Actions Std 0.954869 +eval/Actions Max 0.999992 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60974 +time/logging (s) 0.00367931 +time/sampling batch (s) 0.277023 +time/saving (s) 0.00337163 +time/training (s) 7.12283 +time/epoch (s) 10.0166 +time/total (s) 5517.22 +Epoch -461 +---------------------------------- --------------- +2022-05-10 14:42:51.505097 PDT | [2] Epoch -460 finished +---------------------------------- --------------- +epoch -460 +replay_buffer/size 999033 +trainer/num train calls 541000 +trainer/Policy Loss -19.9279 +trainer/Log Pis Mean 24.7305 +trainer/Log Pis Std 13.2601 +trainer/Log Pis Max 71.0467 +trainer/Log Pis Min -7.23177 +trainer/policy/mean Mean -0.0420497 +trainer/policy/mean Std 0.906324 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999984 +trainer/policy/normal/std Mean 2.82283 +trainer/policy/normal/std Std 0.651525 +trainer/policy/normal/std Max 5.58221 +trainer/policy/normal/std Min 0.288417 +trainer/policy/normal/log_std Mean 0.998669 +trainer/policy/normal/log_std Std 0.317549 +trainer/policy/normal/log_std Max 1.71959 +trainer/policy/normal/log_std Min -1.24335 +eval/num steps total 540085 +eval/num paths total 541 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.251385 +eval/Actions Std 0.893327 +eval/Actions Max 0.999992 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51289 +time/logging (s) 0.00380419 +time/sampling batch (s) 0.279477 +time/saving (s) 0.00345622 +time/training (s) 8.2857 +time/epoch (s) 11.0853 +time/total (s) 5528.3 +Epoch -460 +---------------------------------- --------------- +2022-05-10 14:43:00.925964 PDT | [2] Epoch -459 finished +---------------------------------- --------------- +epoch -459 +replay_buffer/size 999033 +trainer/num train calls 542000 +trainer/Policy Loss -18.8116 +trainer/Log Pis Mean 24.936 +trainer/Log Pis Std 13.1902 +trainer/Log Pis Max 66.4965 +trainer/Log Pis Min -7.45377 +trainer/policy/mean Mean -0.0227991 +trainer/policy/mean Std 0.906172 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.82967 +trainer/policy/normal/std Std 0.647526 +trainer/policy/normal/std Max 5.41304 +trainer/policy/normal/std Min 0.207667 +trainer/policy/normal/log_std Mean 1.00178 +trainer/policy/normal/log_std Std 0.314204 +trainer/policy/normal/log_std Max 1.68881 +trainer/policy/normal/log_std Min -1.57182 +eval/num steps total 541085 +eval/num paths total 542 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0199728 +eval/Actions Std 0.969418 +eval/Actions Max 0.999975 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55144 +time/logging (s) 0.00373951 +time/sampling batch (s) 0.287762 +time/saving (s) 0.00344072 +time/training (s) 6.55147 +time/epoch (s) 9.39785 +time/total (s) 5537.7 +Epoch -459 +---------------------------------- --------------- +2022-05-10 14:43:10.396475 PDT | [2] Epoch -458 finished +---------------------------------- --------------- +epoch -458 +replay_buffer/size 999033 +trainer/num train calls 543000 +trainer/Policy Loss -21.5427 +trainer/Log Pis Mean 25.418 +trainer/Log Pis Std 13.6296 +trainer/Log Pis Max 68.258 +trainer/Log Pis Min -8.65415 +trainer/policy/mean Mean -0.0225773 +trainer/policy/mean Std 0.908213 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.88263 +trainer/policy/normal/std Std 0.671968 +trainer/policy/normal/std Max 6.1384 +trainer/policy/normal/std Min 0.317077 +trainer/policy/normal/log_std Mean 1.02061 +trainer/policy/normal/log_std Std 0.30978 +trainer/policy/normal/log_std Max 1.81456 +trainer/policy/normal/log_std Min -1.14861 +eval/num steps total 542085 +eval/num paths total 543 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0774225 +eval/Actions Std 0.907473 +eval/Actions Max 1 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.80814 +time/logging (s) 0.00368711 +time/sampling batch (s) 0.279064 +time/saving (s) 0.00338811 +time/training (s) 6.35371 +time/epoch (s) 9.44799 +time/total (s) 5547.16 +Epoch -458 +---------------------------------- --------------- +2022-05-10 14:43:20.085393 PDT | [2] Epoch -457 finished +---------------------------------- --------------- +epoch -457 +replay_buffer/size 999033 +trainer/num train calls 544000 +trainer/Policy Loss -19.4565 +trainer/Log Pis Mean 24.8133 +trainer/Log Pis Std 13.3882 +trainer/Log Pis Max 62.4499 +trainer/Log Pis Min -10.6935 +trainer/policy/mean Mean -0.0454926 +trainer/policy/mean Std 0.906235 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.78514 +trainer/policy/normal/std Std 0.64708 +trainer/policy/normal/std Max 6.33368 +trainer/policy/normal/std Min 0.299886 +trainer/policy/normal/log_std Mean 0.985578 +trainer/policy/normal/log_std Std 0.314542 +trainer/policy/normal/log_std Max 1.84588 +trainer/policy/normal/log_std Min -1.20435 +eval/num steps total 543085 +eval/num paths total 544 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.114602 +eval/Actions Std 0.912305 +eval/Actions Max 0.999996 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70839 +time/logging (s) 0.00393984 +time/sampling batch (s) 0.278258 +time/saving (s) 0.0039113 +time/training (s) 6.67228 +time/epoch (s) 9.66678 +time/total (s) 5556.83 +Epoch -457 +---------------------------------- --------------- +2022-05-10 14:43:30.233129 PDT | [2] Epoch -456 finished +---------------------------------- --------------- +epoch -456 +replay_buffer/size 999033 +trainer/num train calls 545000 +trainer/Policy Loss -20.3701 +trainer/Log Pis Mean 24.7792 +trainer/Log Pis Std 13.9425 +trainer/Log Pis Max 66.668 +trainer/Log Pis Min -9.04991 +trainer/policy/mean Mean -0.0474425 +trainer/policy/mean Std 0.906476 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.88236 +trainer/policy/normal/std Std 0.626611 +trainer/policy/normal/std Max 5.56137 +trainer/policy/normal/std Min 0.23947 +trainer/policy/normal/log_std Mean 1.0261 +trainer/policy/normal/log_std Std 0.284159 +trainer/policy/normal/log_std Max 1.71584 +trainer/policy/normal/log_std Min -1.42933 +eval/num steps total 544085 +eval/num paths total 545 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.258464 +eval/Actions Std 0.828105 +eval/Actions Max 0.999998 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.21471 +time/logging (s) 0.00372689 +time/sampling batch (s) 0.27959 +time/saving (s) 0.0034639 +time/training (s) 7.62334 +time/epoch (s) 10.1248 +time/total (s) 5566.95 +Epoch -456 +---------------------------------- --------------- +2022-05-10 14:43:40.048058 PDT | [2] Epoch -455 finished +---------------------------------- --------------- +epoch -455 +replay_buffer/size 999033 +trainer/num train calls 546000 +trainer/Policy Loss -19.2774 +trainer/Log Pis Mean 25.0874 +trainer/Log Pis Std 13.5034 +trainer/Log Pis Max 72.2274 +trainer/Log Pis Min -11.1522 +trainer/policy/mean Mean -0.0466562 +trainer/policy/mean Std 0.908426 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.86667 +trainer/policy/normal/std Std 0.650251 +trainer/policy/normal/std Max 5.60394 +trainer/policy/normal/std Min 0.267406 +trainer/policy/normal/log_std Mean 1.01751 +trainer/policy/normal/log_std Std 0.298273 +trainer/policy/normal/log_std Max 1.72347 +trainer/policy/normal/log_std Min -1.31899 +eval/num steps total 545085 +eval/num paths total 546 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0220885 +eval/Actions Std 0.906611 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.75099 +time/logging (s) 0.00389265 +time/sampling batch (s) 0.280068 +time/saving (s) 0.00405495 +time/training (s) 6.75366 +time/epoch (s) 9.79267 +time/total (s) 5576.75 +Epoch -455 +---------------------------------- --------------- +2022-05-10 14:43:50.123592 PDT | [2] Epoch -454 finished +---------------------------------- --------------- +epoch -454 +replay_buffer/size 999033 +trainer/num train calls 547000 +trainer/Policy Loss -19.4243 +trainer/Log Pis Mean 23.7401 +trainer/Log Pis Std 13.5057 +trainer/Log Pis Max 66.6112 +trainer/Log Pis Min -12.8408 +trainer/policy/mean Mean -0.0368546 +trainer/policy/mean Std 0.908789 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999985 +trainer/policy/normal/std Mean 2.79906 +trainer/policy/normal/std Std 0.637492 +trainer/policy/normal/std Max 5.75953 +trainer/policy/normal/std Min 0.324121 +trainer/policy/normal/log_std Mean 0.992204 +trainer/policy/normal/log_std Std 0.305752 +trainer/policy/normal/log_std Max 1.75086 +trainer/policy/normal/log_std Min -1.12664 +eval/num steps total 546085 +eval/num paths total 547 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0376979 +eval/Actions Std 0.888281 +eval/Actions Max 0.999991 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64019 +time/logging (s) 0.00373095 +time/sampling batch (s) 0.282613 +time/saving (s) 0.0034109 +time/training (s) 7.12225 +time/epoch (s) 10.0522 +time/total (s) 5586.8 +Epoch -454 +---------------------------------- --------------- +2022-05-10 14:43:59.958167 PDT | [2] Epoch -453 finished +---------------------------------- --------------- +epoch -453 +replay_buffer/size 999033 +trainer/num train calls 548000 +trainer/Policy Loss -18.7216 +trainer/Log Pis Mean 24.201 +trainer/Log Pis Std 13.0144 +trainer/Log Pis Max 65.367 +trainer/Log Pis Min -6.40199 +trainer/policy/mean Mean -0.0220966 +trainer/policy/mean Std 0.909332 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.73413 +trainer/policy/normal/std Std 0.620573 +trainer/policy/normal/std Max 5.30423 +trainer/policy/normal/std Min 0.293319 +trainer/policy/normal/log_std Mean 0.968942 +trainer/policy/normal/log_std Std 0.305868 +trainer/policy/normal/log_std Max 1.6685 +trainer/policy/normal/log_std Min -1.2265 +eval/num steps total 547085 +eval/num paths total 548 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.100371 +eval/Actions Std 0.918443 +eval/Actions Max 0.999997 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64037 +time/logging (s) 0.00367355 +time/sampling batch (s) 0.282082 +time/saving (s) 0.00353471 +time/training (s) 6.8821 +time/epoch (s) 9.81176 +time/total (s) 5596.62 +Epoch -453 +---------------------------------- --------------- +2022-05-10 14:44:10.560726 PDT | [2] Epoch -452 finished +---------------------------------- --------------- +epoch -452 +replay_buffer/size 999033 +trainer/num train calls 549000 +trainer/Policy Loss -19.9821 +trainer/Log Pis Mean 24.0629 +trainer/Log Pis Std 12.8598 +trainer/Log Pis Max 62.7373 +trainer/Log Pis Min -7.46552 +trainer/policy/mean Mean -0.0603777 +trainer/policy/mean Std 0.910528 +trainer/policy/mean Max 0.999981 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.79827 +trainer/policy/normal/std Std 0.649987 +trainer/policy/normal/std Max 5.30624 +trainer/policy/normal/std Min 0.29787 +trainer/policy/normal/log_std Mean 0.989395 +trainer/policy/normal/log_std Std 0.319173 +trainer/policy/normal/log_std Max 1.66888 +trainer/policy/normal/log_std Min -1.2111 +eval/num steps total 548085 +eval/num paths total 549 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0539849 +eval/Actions Std 0.894766 +eval/Actions Max 0.999996 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63426 +time/logging (s) 0.00368173 +time/sampling batch (s) 0.278706 +time/saving (s) 0.00336119 +time/training (s) 7.66012 +time/epoch (s) 10.5801 +time/total (s) 5607.2 +Epoch -452 +---------------------------------- --------------- +2022-05-10 14:44:21.336864 PDT | [2] Epoch -451 finished +---------------------------------- --------------- +epoch -451 +replay_buffer/size 999033 +trainer/num train calls 550000 +trainer/Policy Loss -19.9295 +trainer/Log Pis Mean 24.0249 +trainer/Log Pis Std 13.0394 +trainer/Log Pis Max 67.69 +trainer/Log Pis Min -8.72052 +trainer/policy/mean Mean -0.0347268 +trainer/policy/mean Std 0.908767 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.8507 +trainer/policy/normal/std Std 0.659955 +trainer/policy/normal/std Max 5.95259 +trainer/policy/normal/std Min 0.352253 +trainer/policy/normal/log_std Mean 1.00929 +trainer/policy/normal/log_std Std 0.311124 +trainer/policy/normal/log_std Max 1.78383 +trainer/policy/normal/log_std Min -1.04341 +eval/num steps total 549085 +eval/num paths total 550 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00129612 +eval/Actions Std 0.903995 +eval/Actions Max 0.999991 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58282 +time/logging (s) 0.00373327 +time/sampling batch (s) 0.524476 +time/saving (s) 0.00335528 +time/training (s) 7.63979 +time/epoch (s) 10.7542 +time/total (s) 5617.96 +Epoch -451 +---------------------------------- --------------- +2022-05-10 14:44:31.508253 PDT | [2] Epoch -450 finished +---------------------------------- --------------- +epoch -450 +replay_buffer/size 999033 +trainer/num train calls 551000 +trainer/Policy Loss -20.2169 +trainer/Log Pis Mean 24.3341 +trainer/Log Pis Std 13.0045 +trainer/Log Pis Max 61.5428 +trainer/Log Pis Min -2.77526 +trainer/policy/mean Mean -0.0250008 +trainer/policy/mean Std 0.907234 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.7507 +trainer/policy/normal/std Std 0.628955 +trainer/policy/normal/std Max 6.14754 +trainer/policy/normal/std Min 0.262663 +trainer/policy/normal/log_std Mean 0.974153 +trainer/policy/normal/log_std Std 0.310153 +trainer/policy/normal/log_std Max 1.81605 +trainer/policy/normal/log_std Min -1.33689 +eval/num steps total 550085 +eval/num paths total 551 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0366283 +eval/Actions Std 0.937558 +eval/Actions Max 0.999973 +eval/Actions Min -0.999984 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.43322 +time/logging (s) 0.00373949 +time/sampling batch (s) 0.275422 +time/saving (s) 0.003391 +time/training (s) 7.43322 +time/epoch (s) 10.149 +time/total (s) 5628.11 +Epoch -450 +---------------------------------- --------------- +2022-05-10 14:44:42.037423 PDT | [2] Epoch -449 finished +---------------------------------- --------------- +epoch -449 +replay_buffer/size 999033 +trainer/num train calls 552000 +trainer/Policy Loss -19.9013 +trainer/Log Pis Mean 24.7039 +trainer/Log Pis Std 13.5472 +trainer/Log Pis Max 76.4346 +trainer/Log Pis Min -11.3134 +trainer/policy/mean Mean -0.0219196 +trainer/policy/mean Std 0.907674 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.78983 +trainer/policy/normal/std Std 0.675676 +trainer/policy/normal/std Max 5.58176 +trainer/policy/normal/std Min 0.29195 +trainer/policy/normal/log_std Mean 0.981978 +trainer/policy/normal/log_std Std 0.338294 +trainer/policy/normal/log_std Max 1.7195 +trainer/policy/normal/log_std Min -1.23117 +eval/num steps total 551085 +eval/num paths total 552 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00473644 +eval/Actions Std 0.942057 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.39858 +time/logging (s) 0.00459458 +time/sampling batch (s) 0.279907 +time/saving (s) 0.00410012 +time/training (s) 7.82024 +time/epoch (s) 10.5074 +time/total (s) 5638.62 +Epoch -449 +---------------------------------- --------------- +2022-05-10 14:44:52.437632 PDT | [2] Epoch -448 finished +---------------------------------- --------------- +epoch -448 +replay_buffer/size 999033 +trainer/num train calls 553000 +trainer/Policy Loss -19.2906 +trainer/Log Pis Mean 24.2135 +trainer/Log Pis Std 12.97 +trainer/Log Pis Max 67.3744 +trainer/Log Pis Min -5.26724 +trainer/policy/mean Mean -0.0440736 +trainer/policy/mean Std 0.907192 +trainer/policy/mean Max 0.999982 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.86041 +trainer/policy/normal/std Std 0.65902 +trainer/policy/normal/std Max 5.39973 +trainer/policy/normal/std Min 0.33924 +trainer/policy/normal/log_std Mean 1.01173 +trainer/policy/normal/log_std Std 0.319065 +trainer/policy/normal/log_std Max 1.68635 +trainer/policy/normal/log_std Min -1.08105 +eval/num steps total 552085 +eval/num paths total 553 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.391451 +eval/Actions Std 0.826043 +eval/Actions Max 0.999992 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64851 +time/logging (s) 0.00370036 +time/sampling batch (s) 0.528686 +time/saving (s) 0.0034417 +time/training (s) 7.19128 +time/epoch (s) 10.3756 +time/total (s) 5649 +Epoch -448 +---------------------------------- --------------- +2022-05-10 14:45:03.094897 PDT | [2] Epoch -447 finished +---------------------------------- --------------- +epoch -447 +replay_buffer/size 999033 +trainer/num train calls 554000 +trainer/Policy Loss -21.7102 +trainer/Log Pis Mean 24.7951 +trainer/Log Pis Std 12.5815 +trainer/Log Pis Max 60.8071 +trainer/Log Pis Min -7.25083 +trainer/policy/mean Mean -0.0524259 +trainer/policy/mean Std 0.910991 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.73656 +trainer/policy/normal/std Std 0.62844 +trainer/policy/normal/std Max 6.3457 +trainer/policy/normal/std Min 0.323921 +trainer/policy/normal/log_std Mean 0.969133 +trainer/policy/normal/log_std Std 0.308632 +trainer/policy/normal/log_std Max 1.84778 +trainer/policy/normal/log_std Min -1.12725 +eval/num steps total 553085 +eval/num paths total 554 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.229553 +eval/Actions Std 0.895839 +eval/Actions Max 0.999996 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68187 +time/logging (s) 0.00379963 +time/sampling batch (s) 0.278209 +time/saving (s) 0.00345624 +time/training (s) 7.66751 +time/epoch (s) 10.6348 +time/total (s) 5659.64 +Epoch -447 +---------------------------------- --------------- +2022-05-10 14:45:13.033174 PDT | [2] Epoch -446 finished +---------------------------------- --------------- +epoch -446 +replay_buffer/size 999033 +trainer/num train calls 555000 +trainer/Policy Loss -20.516 +trainer/Log Pis Mean 25.1307 +trainer/Log Pis Std 13.3083 +trainer/Log Pis Max 77.4483 +trainer/Log Pis Min -9.18118 +trainer/policy/mean Mean -0.0395223 +trainer/policy/mean Std 0.909732 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.81509 +trainer/policy/normal/std Std 0.654269 +trainer/policy/normal/std Max 5.94827 +trainer/policy/normal/std Min 0.274943 +trainer/policy/normal/log_std Mean 0.994994 +trainer/policy/normal/log_std Std 0.322394 +trainer/policy/normal/log_std Max 1.7831 +trainer/policy/normal/log_std Min -1.29119 +eval/num steps total 554085 +eval/num paths total 555 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0692126 +eval/Actions Std 0.911098 +eval/Actions Max 0.999991 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60819 +time/logging (s) 0.00373442 +time/sampling batch (s) 0.529623 +time/saving (s) 0.00341427 +time/training (s) 6.77064 +time/epoch (s) 9.9156 +time/total (s) 5669.56 +Epoch -446 +---------------------------------- --------------- +2022-05-10 14:45:23.710074 PDT | [2] Epoch -445 finished +---------------------------------- --------------- +epoch -445 +replay_buffer/size 999033 +trainer/num train calls 556000 +trainer/Policy Loss -18.8714 +trainer/Log Pis Mean 24.888 +trainer/Log Pis Std 13.5259 +trainer/Log Pis Max 69.7905 +trainer/Log Pis Min -6.63372 +trainer/policy/mean Mean -0.049051 +trainer/policy/mean Std 0.909991 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.87222 +trainer/policy/normal/std Std 0.674196 +trainer/policy/normal/std Max 6.17136 +trainer/policy/normal/std Min 0.291741 +trainer/policy/normal/log_std Mean 1.01501 +trainer/policy/normal/log_std Std 0.321143 +trainer/policy/normal/log_std Max 1.81992 +trainer/policy/normal/log_std Min -1.23189 +eval/num steps total 555085 +eval/num paths total 556 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0263817 +eval/Actions Std 0.872761 +eval/Actions Max 0.999999 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62554 +time/logging (s) 0.00401376 +time/sampling batch (s) 0.303397 +time/saving (s) 0.00363024 +time/training (s) 7.71772 +time/epoch (s) 10.6543 +time/total (s) 5680.21 +Epoch -445 +---------------------------------- --------------- +2022-05-10 14:45:34.658242 PDT | [2] Epoch -444 finished +---------------------------------- --------------- +epoch -444 +replay_buffer/size 999033 +trainer/num train calls 557000 +trainer/Policy Loss -19.5519 +trainer/Log Pis Mean 23.956 +trainer/Log Pis Std 12.8272 +trainer/Log Pis Max 64.5371 +trainer/Log Pis Min -12.3223 +trainer/policy/mean Mean -0.0246759 +trainer/policy/mean Std 0.905361 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.82382 +trainer/policy/normal/std Std 0.655778 +trainer/policy/normal/std Max 6.34192 +trainer/policy/normal/std Min 0.314207 +trainer/policy/normal/log_std Mean 0.999448 +trainer/policy/normal/log_std Std 0.313516 +trainer/policy/normal/log_std Max 1.84718 +trainer/policy/normal/log_std Min -1.1577 +eval/num steps total 556085 +eval/num paths total 557 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0091228 +eval/Actions Std 0.905442 +eval/Actions Max 0.999996 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60431 +time/logging (s) 0.00388397 +time/sampling batch (s) 0.328494 +time/saving (s) 0.00362313 +time/training (s) 7.98439 +time/epoch (s) 10.9247 +time/total (s) 5691.14 +Epoch -444 +---------------------------------- --------------- +2022-05-10 14:45:45.277141 PDT | [2] Epoch -443 finished +---------------------------------- --------------- +epoch -443 +replay_buffer/size 999033 +trainer/num train calls 558000 +trainer/Policy Loss -19.5461 +trainer/Log Pis Mean 24.1312 +trainer/Log Pis Std 13.8055 +trainer/Log Pis Max 67.0552 +trainer/Log Pis Min -4.98899 +trainer/policy/mean Mean -0.0318017 +trainer/policy/mean Std 0.904994 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.79267 +trainer/policy/normal/std Std 0.667189 +trainer/policy/normal/std Max 7.10656 +trainer/policy/normal/std Min 0.309063 +trainer/policy/normal/log_std Mean 0.984167 +trainer/policy/normal/log_std Std 0.334229 +trainer/policy/normal/log_std Max 1.96102 +trainer/policy/normal/log_std Min -1.17421 +eval/num steps total 557085 +eval/num paths total 558 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0875199 +eval/Actions Std 0.91754 +eval/Actions Max 0.999971 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65447 +time/logging (s) 0.00403434 +time/sampling batch (s) 0.280983 +time/saving (s) 0.00394375 +time/training (s) 7.6525 +time/epoch (s) 10.5959 +time/total (s) 5701.74 +Epoch -443 +---------------------------------- --------------- +2022-05-10 14:45:55.081561 PDT | [2] Epoch -442 finished +---------------------------------- --------------- +epoch -442 +replay_buffer/size 999033 +trainer/num train calls 559000 +trainer/Policy Loss -18.7056 +trainer/Log Pis Mean 24.4243 +trainer/Log Pis Std 13.2107 +trainer/Log Pis Max 64.8763 +trainer/Log Pis Min -6.5386 +trainer/policy/mean Mean -0.0468568 +trainer/policy/mean Std 0.906794 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.82729 +trainer/policy/normal/std Std 0.641701 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.307562 +trainer/policy/normal/log_std Mean 1.00307 +trainer/policy/normal/log_std Std 0.302777 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.17908 +eval/num steps total 558085 +eval/num paths total 559 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0827719 +eval/Actions Std 0.908666 +eval/Actions Max 0.999995 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.45107 +time/logging (s) 0.00372405 +time/sampling batch (s) 0.279941 +time/saving (s) 0.00357614 +time/training (s) 7.04265 +time/epoch (s) 9.78096 +time/total (s) 5711.52 +Epoch -442 +---------------------------------- --------------- +2022-05-10 14:46:04.756751 PDT | [2] Epoch -441 finished +---------------------------------- --------------- +epoch -441 +replay_buffer/size 999033 +trainer/num train calls 560000 +trainer/Policy Loss -19.3679 +trainer/Log Pis Mean 23.8684 +trainer/Log Pis Std 12.7 +trainer/Log Pis Max 64.537 +trainer/Log Pis Min -7.56181 +trainer/policy/mean Mean -0.0467737 +trainer/policy/mean Std 0.905951 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.82721 +trainer/policy/normal/std Std 0.659705 +trainer/policy/normal/std Max 5.57606 +trainer/policy/normal/std Min 0.265855 +trainer/policy/normal/log_std Mean 0.999539 +trainer/policy/normal/log_std Std 0.319697 +trainer/policy/normal/log_std Max 1.71848 +trainer/policy/normal/log_std Min -1.3248 +eval/num steps total 559085 +eval/num paths total 560 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.519354 +eval/Actions Std 0.719662 +eval/Actions Max 1 +eval/Actions Min -0.999972 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63439 +time/logging (s) 0.00369282 +time/sampling batch (s) 0.277721 +time/saving (s) 0.00343174 +time/training (s) 6.73339 +time/epoch (s) 9.65262 +time/total (s) 5721.18 +Epoch -441 +---------------------------------- --------------- +2022-05-10 14:46:16.092164 PDT | [2] Epoch -440 finished +---------------------------------- --------------- +epoch -440 +replay_buffer/size 999033 +trainer/num train calls 561000 +trainer/Policy Loss -21.4673 +trainer/Log Pis Mean 26.0777 +trainer/Log Pis Std 14.2057 +trainer/Log Pis Max 70.9351 +trainer/Log Pis Min -6.00285 +trainer/policy/mean Mean -0.0200973 +trainer/policy/mean Std 0.910893 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.85991 +trainer/policy/normal/std Std 0.642489 +trainer/policy/normal/std Max 5.56847 +trainer/policy/normal/std Min 0.370258 +trainer/policy/normal/log_std Mean 1.01453 +trainer/policy/normal/log_std Std 0.304029 +trainer/policy/normal/log_std Max 1.71712 +trainer/policy/normal/log_std Min -0.993555 +eval/num steps total 560085 +eval/num paths total 561 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0614622 +eval/Actions Std 0.901846 +eval/Actions Max 0.999995 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50295 +time/logging (s) 0.00386472 +time/sampling batch (s) 0.532897 +time/saving (s) 0.00343222 +time/training (s) 8.26968 +time/epoch (s) 11.3128 +time/total (s) 5732.5 +Epoch -440 +---------------------------------- --------------- +2022-05-10 14:46:26.297258 PDT | [2] Epoch -439 finished +---------------------------------- --------------- +epoch -439 +replay_buffer/size 999033 +trainer/num train calls 562000 +trainer/Policy Loss -19.1021 +trainer/Log Pis Mean 24.0389 +trainer/Log Pis Std 13.5434 +trainer/Log Pis Max 84.1076 +trainer/Log Pis Min -10.7834 +trainer/policy/mean Mean -0.0376947 +trainer/policy/mean Std 0.907209 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.84416 +trainer/policy/normal/std Std 0.65885 +trainer/policy/normal/std Max 6.32564 +trainer/policy/normal/std Min 0.29077 +trainer/policy/normal/log_std Mean 1.00747 +trainer/policy/normal/log_std Std 0.309067 +trainer/policy/normal/log_std Max 1.84461 +trainer/policy/normal/log_std Min -1.23522 +eval/num steps total 561085 +eval/num paths total 562 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0151665 +eval/Actions Std 0.970303 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.89474 +time/logging (s) 0.00377445 +time/sampling batch (s) 0.283505 +time/saving (s) 0.00337624 +time/training (s) 6.99701 +time/epoch (s) 10.1824 +time/total (s) 5742.68 +Epoch -439 +---------------------------------- --------------- +2022-05-10 14:46:36.137064 PDT | [2] Epoch -438 finished +---------------------------------- --------------- +epoch -438 +replay_buffer/size 999033 +trainer/num train calls 563000 +trainer/Policy Loss -18.4037 +trainer/Log Pis Mean 24.0152 +trainer/Log Pis Std 13.6525 +trainer/Log Pis Max 64.3424 +trainer/Log Pis Min -11.2229 +trainer/policy/mean Mean -0.0149618 +trainer/policy/mean Std 0.903558 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.8169 +trainer/policy/normal/std Std 0.662406 +trainer/policy/normal/std Max 5.77086 +trainer/policy/normal/std Min 0.29134 +trainer/policy/normal/log_std Mean 0.995074 +trainer/policy/normal/log_std Std 0.323449 +trainer/policy/normal/log_std Max 1.75282 +trainer/policy/normal/log_std Min -1.23327 +eval/num steps total 562085 +eval/num paths total 563 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.499452 +eval/Actions Std 0.664928 +eval/Actions Max 0.999998 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.34329 +time/logging (s) 0.00379532 +time/sampling batch (s) 0.274882 +time/saving (s) 0.00343651 +time/training (s) 7.19217 +time/epoch (s) 9.81757 +time/total (s) 5752.5 +Epoch -438 +---------------------------------- --------------- +2022-05-10 14:46:46.357606 PDT | [2] Epoch -437 finished +---------------------------------- --------------- +epoch -437 +replay_buffer/size 999033 +trainer/num train calls 564000 +trainer/Policy Loss -20.2389 +trainer/Log Pis Mean 24.0005 +trainer/Log Pis Std 13.3989 +trainer/Log Pis Max 65.4998 +trainer/Log Pis Min -7.21099 +trainer/policy/mean Mean -0.0614148 +trainer/policy/mean Std 0.90961 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.79576 +trainer/policy/normal/std Std 0.621294 +trainer/policy/normal/std Max 5.6494 +trainer/policy/normal/std Min 0.303419 +trainer/policy/normal/log_std Mean 0.993158 +trainer/policy/normal/log_std Std 0.297242 +trainer/policy/normal/log_std Max 1.73155 +trainer/policy/normal/log_std Min -1.19264 +eval/num steps total 563085 +eval/num paths total 564 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0351665 +eval/Actions Std 0.898639 +eval/Actions Max 0.999998 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54991 +time/logging (s) 0.00391401 +time/sampling batch (s) 0.527248 +time/saving (s) 0.00339677 +time/training (s) 7.1136 +time/epoch (s) 10.1981 +time/total (s) 5762.7 +Epoch -437 +---------------------------------- --------------- +2022-05-10 14:46:57.319659 PDT | [2] Epoch -436 finished +---------------------------------- --------------- +epoch -436 +replay_buffer/size 999033 +trainer/num train calls 565000 +trainer/Policy Loss -19.529 +trainer/Log Pis Mean 25.1845 +trainer/Log Pis Std 13.6242 +trainer/Log Pis Max 71.593 +trainer/Log Pis Min -7.36552 +trainer/policy/mean Mean -0.0260625 +trainer/policy/mean Std 0.906306 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.85711 +trainer/policy/normal/std Std 0.650109 +trainer/policy/normal/std Max 6.04613 +trainer/policy/normal/std Min 0.269983 +trainer/policy/normal/log_std Mean 1.01255 +trainer/policy/normal/log_std Std 0.308503 +trainer/policy/normal/log_std Max 1.79942 +trainer/policy/normal/log_std Min -1.3094 +eval/num steps total 564085 +eval/num paths total 565 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.168761 +eval/Actions Std 0.886796 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61996 +time/logging (s) 0.00380317 +time/sampling batch (s) 0.531024 +time/saving (s) 0.00344401 +time/training (s) 7.78059 +time/epoch (s) 10.9388 +time/total (s) 5773.64 +Epoch -436 +---------------------------------- --------------- +2022-05-10 14:47:07.607993 PDT | [2] Epoch -435 finished +---------------------------------- --------------- +epoch -435 +replay_buffer/size 999033 +trainer/num train calls 566000 +trainer/Policy Loss -18.7557 +trainer/Log Pis Mean 26.0073 +trainer/Log Pis Std 13.5353 +trainer/Log Pis Max 67.6012 +trainer/Log Pis Min -15.2055 +trainer/policy/mean Mean -0.0545957 +trainer/policy/mean Std 0.909546 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.89443 +trainer/policy/normal/std Std 0.660551 +trainer/policy/normal/std Max 6.69405 +trainer/policy/normal/std Min 0.335673 +trainer/policy/normal/log_std Mean 1.02648 +trainer/policy/normal/log_std Std 0.301536 +trainer/policy/normal/log_std Max 1.90122 +trainer/policy/normal/log_std Min -1.09162 +eval/num steps total 565085 +eval/num paths total 566 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.394693 +eval/Actions Std 0.782726 +eval/Actions Max 0.999981 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5063 +time/logging (s) 0.003703 +time/sampling batch (s) 0.277057 +time/saving (s) 0.00341002 +time/training (s) 7.47507 +time/epoch (s) 10.2655 +time/total (s) 5783.91 +Epoch -435 +---------------------------------- --------------- +2022-05-10 14:47:17.577218 PDT | [2] Epoch -434 finished +---------------------------------- --------------- +epoch -434 +replay_buffer/size 999033 +trainer/num train calls 567000 +trainer/Policy Loss -19.6111 +trainer/Log Pis Mean 23.895 +trainer/Log Pis Std 12.6772 +trainer/Log Pis Max 67.4663 +trainer/Log Pis Min -6.59124 +trainer/policy/mean Mean -0.0426397 +trainer/policy/mean Std 0.905852 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.83148 +trainer/policy/normal/std Std 0.623077 +trainer/policy/normal/std Max 5.92158 +trainer/policy/normal/std Min 0.335525 +trainer/policy/normal/log_std Mean 1.00666 +trainer/policy/normal/log_std Std 0.29283 +trainer/policy/normal/log_std Max 1.7786 +trainer/policy/normal/log_std Min -1.09206 +eval/num steps total 566085 +eval/num paths total 567 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0993487 +eval/Actions Std 0.923592 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6151 +time/logging (s) 0.00409406 +time/sampling batch (s) 0.528641 +time/saving (s) 0.00395637 +time/training (s) 6.79515 +time/epoch (s) 9.94694 +time/total (s) 5793.86 +Epoch -434 +---------------------------------- --------------- +2022-05-10 14:47:27.134536 PDT | [2] Epoch -433 finished +---------------------------------- --------------- +epoch -433 +replay_buffer/size 999033 +trainer/num train calls 568000 +trainer/Policy Loss -19.6064 +trainer/Log Pis Mean 24.9478 +trainer/Log Pis Std 13.978 +trainer/Log Pis Max 75.5925 +trainer/Log Pis Min -9.05019 +trainer/policy/mean Mean -0.0340896 +trainer/policy/mean Std 0.90413 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.90537 +trainer/policy/normal/std Std 0.677327 +trainer/policy/normal/std Max 6.35021 +trainer/policy/normal/std Min 0.288781 +trainer/policy/normal/log_std Mean 1.0276 +trainer/policy/normal/log_std Std 0.314287 +trainer/policy/normal/log_std Max 1.84849 +trainer/policy/normal/log_std Min -1.24209 +eval/num steps total 567085 +eval/num paths total 568 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.219946 +eval/Actions Std 0.913446 +eval/Actions Max 0.999995 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55839 +time/logging (s) 0.00375356 +time/sampling batch (s) 0.27643 +time/saving (s) 0.00354353 +time/training (s) 6.69195 +time/epoch (s) 9.53407 +time/total (s) 5803.4 +Epoch -433 +---------------------------------- --------------- +2022-05-10 14:47:36.942737 PDT | [2] Epoch -432 finished +---------------------------------- --------------- +epoch -432 +replay_buffer/size 999033 +trainer/num train calls 569000 +trainer/Policy Loss -20.682 +trainer/Log Pis Mean 23.7661 +trainer/Log Pis Std 13.1595 +trainer/Log Pis Max 65.605 +trainer/Log Pis Min -10.2143 +trainer/policy/mean Mean -0.0310652 +trainer/policy/mean Std 0.907874 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.75042 +trainer/policy/normal/std Std 0.632878 +trainer/policy/normal/std Max 6.65365 +trainer/policy/normal/std Min 0.353004 +trainer/policy/normal/log_std Mean 0.973234 +trainer/policy/normal/log_std Std 0.313351 +trainer/policy/normal/log_std Max 1.89517 +trainer/policy/normal/log_std Min -1.04128 +eval/num steps total 568085 +eval/num paths total 569 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0484426 +eval/Actions Std 0.899398 +eval/Actions Max 0.999995 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64092 +time/logging (s) 0.00372061 +time/sampling batch (s) 0.274978 +time/saving (s) 0.00339419 +time/training (s) 6.86272 +time/epoch (s) 9.78573 +time/total (s) 5813.19 +Epoch -432 +---------------------------------- --------------- +2022-05-10 14:47:48.904643 PDT | [2] Epoch -431 finished +---------------------------------- --------------- +epoch -431 +replay_buffer/size 999033 +trainer/num train calls 570000 +trainer/Policy Loss -18.9828 +trainer/Log Pis Mean 24.0682 +trainer/Log Pis Std 13.9682 +trainer/Log Pis Max 80.1029 +trainer/Log Pis Min -8.0032 +trainer/policy/mean Mean -0.0313995 +trainer/policy/mean Std 0.907717 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.76769 +trainer/policy/normal/std Std 0.650066 +trainer/policy/normal/std Max 5.09784 +trainer/policy/normal/std Min 0.308931 +trainer/policy/normal/log_std Mean 0.976991 +trainer/policy/normal/log_std Std 0.325597 +trainer/policy/normal/log_std Max 1.62882 +trainer/policy/normal/log_std Min -1.17464 +eval/num steps total 569085 +eval/num paths total 570 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.310872 +eval/Actions Std 0.807948 +eval/Actions Max 0.99999 +eval/Actions Min -0.999988 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69054 +time/logging (s) 0.00373529 +time/sampling batch (s) 0.782087 +time/saving (s) 0.00341381 +time/training (s) 8.45922 +time/epoch (s) 11.939 +time/total (s) 5825.13 +Epoch -431 +---------------------------------- --------------- +2022-05-10 14:47:59.360687 PDT | [2] Epoch -430 finished +---------------------------------- --------------- +epoch -430 +replay_buffer/size 999033 +trainer/num train calls 571000 +trainer/Policy Loss -18.894 +trainer/Log Pis Mean 24.1513 +trainer/Log Pis Std 13.2115 +trainer/Log Pis Max 64.4436 +trainer/Log Pis Min -4.47518 +trainer/policy/mean Mean -0.0276434 +trainer/policy/mean Std 0.897952 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.84375 +trainer/policy/normal/std Std 0.677254 +trainer/policy/normal/std Max 7.28378 +trainer/policy/normal/std Min 0.36469 +trainer/policy/normal/log_std Mean 1.00413 +trainer/policy/normal/log_std Std 0.323306 +trainer/policy/normal/log_std Max 1.98565 +trainer/policy/normal/log_std Min -1.00871 +eval/num steps total 570085 +eval/num paths total 571 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0134675 +eval/Actions Std 0.907939 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.38226 +time/logging (s) 0.00363487 +time/sampling batch (s) 0.279009 +time/saving (s) 0.00343548 +time/training (s) 7.76367 +time/epoch (s) 10.432 +time/total (s) 5835.57 +Epoch -430 +---------------------------------- --------------- +2022-05-10 14:48:09.859147 PDT | [2] Epoch -429 finished +---------------------------------- ---------------- +epoch -429 +replay_buffer/size 999033 +trainer/num train calls 572000 +trainer/Policy Loss -19.9353 +trainer/Log Pis Mean 24.3789 +trainer/Log Pis Std 13.314 +trainer/Log Pis Max 81.073 +trainer/Log Pis Min -9.1795 +trainer/policy/mean Mean -0.0552526 +trainer/policy/mean Std 0.908312 +trainer/policy/mean Max 0.999977 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.77473 +trainer/policy/normal/std Std 0.672059 +trainer/policy/normal/std Max 5.66815 +trainer/policy/normal/std Min 0.25012 +trainer/policy/normal/log_std Mean 0.975116 +trainer/policy/normal/log_std Std 0.347623 +trainer/policy/normal/log_std Max 1.73486 +trainer/policy/normal/log_std Min -1.38581 +eval/num steps total 571085 +eval/num paths total 572 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.000933509 +eval/Actions Std 0.868593 +eval/Actions Max 0.999991 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.77483 +time/logging (s) 0.00374131 +time/sampling batch (s) 0.276069 +time/saving (s) 0.00347124 +time/training (s) 7.41793 +time/epoch (s) 10.476 +time/total (s) 5846.05 +Epoch -429 +---------------------------------- ---------------- +2022-05-10 14:48:20.464852 PDT | [2] Epoch -428 finished +---------------------------------- ---------------- +epoch -428 +replay_buffer/size 999033 +trainer/num train calls 573000 +trainer/Policy Loss -18.8849 +trainer/Log Pis Mean 24.6415 +trainer/Log Pis Std 12.6168 +trainer/Log Pis Max 68.1325 +trainer/Log Pis Min -8.85736 +trainer/policy/mean Mean -0.0285759 +trainer/policy/mean Std 0.905194 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.73584 +trainer/policy/normal/std Std 0.65182 +trainer/policy/normal/std Max 5.95942 +trainer/policy/normal/std Min 0.272769 +trainer/policy/normal/log_std Mean 0.965982 +trainer/policy/normal/log_std Std 0.319827 +trainer/policy/normal/log_std Max 1.78497 +trainer/policy/normal/log_std Min -1.29913 +eval/num steps total 572085 +eval/num paths total 573 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 8.27704e-05 +eval/Actions Std 0.960265 +eval/Actions Max 0.999982 +eval/Actions Min -0.999985 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.78294 +time/logging (s) 0.0037601 +time/sampling batch (s) 0.278824 +time/saving (s) 0.0034036 +time/training (s) 7.51389 +time/epoch (s) 10.5828 +time/total (s) 5856.63 +Epoch -428 +---------------------------------- ---------------- +2022-05-10 14:48:30.966217 PDT | [2] Epoch -427 finished +---------------------------------- -------------- +epoch -427 +replay_buffer/size 999033 +trainer/num train calls 574000 +trainer/Policy Loss -19.6668 +trainer/Log Pis Mean 24.4024 +trainer/Log Pis Std 13.1439 +trainer/Log Pis Max 74.3515 +trainer/Log Pis Min -7.91697 +trainer/policy/mean Mean -0.0395613 +trainer/policy/mean Std 0.907711 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.76961 +trainer/policy/normal/std Std 0.645503 +trainer/policy/normal/std Max 5.36047 +trainer/policy/normal/std Min 0.294203 +trainer/policy/normal/log_std Mean 0.980657 +trainer/policy/normal/log_std Std 0.308947 +trainer/policy/normal/log_std Max 1.67905 +trainer/policy/normal/log_std Min -1.22349 +eval/num steps total 573085 +eval/num paths total 574 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0739782 +eval/Actions Std 0.905859 +eval/Actions Max 0.999994 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.74618 +time/logging (s) 0.0037338 +time/sampling batch (s) 0.279344 +time/saving (s) 0.0033737 +time/training (s) 7.44581 +time/epoch (s) 10.4784 +time/total (s) 5867.11 +Epoch -427 +---------------------------------- -------------- +2022-05-10 14:48:40.546251 PDT | [2] Epoch -426 finished +---------------------------------- --------------- +epoch -426 +replay_buffer/size 999033 +trainer/num train calls 575000 +trainer/Policy Loss -20.3379 +trainer/Log Pis Mean 24.6373 +trainer/Log Pis Std 13.7083 +trainer/Log Pis Max 62.0003 +trainer/Log Pis Min -9.85923 +trainer/policy/mean Mean -0.0589349 +trainer/policy/mean Std 0.907016 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.83219 +trainer/policy/normal/std Std 0.654894 +trainer/policy/normal/std Max 5.69991 +trainer/policy/normal/std Min 0.339926 +trainer/policy/normal/log_std Mean 1.00166 +trainer/policy/normal/log_std Std 0.318711 +trainer/policy/normal/log_std Max 1.74045 +trainer/policy/normal/log_std Min -1.07903 +eval/num steps total 574085 +eval/num paths total 575 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.405244 +eval/Actions Std 0.83478 +eval/Actions Max 0.999996 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.45194 +time/logging (s) 0.0039244 +time/sampling batch (s) 0.275838 +time/saving (s) 0.00395338 +time/training (s) 6.82199 +time/epoch (s) 9.55765 +time/total (s) 5876.67 +Epoch -426 +---------------------------------- --------------- +2022-05-10 14:48:50.475847 PDT | [2] Epoch -425 finished +---------------------------------- --------------- +epoch -425 +replay_buffer/size 999033 +trainer/num train calls 576000 +trainer/Policy Loss -19.1745 +trainer/Log Pis Mean 23.799 +trainer/Log Pis Std 13.0762 +trainer/Log Pis Max 62.7613 +trainer/Log Pis Min -5.45181 +trainer/policy/mean Mean -0.0169404 +trainer/policy/mean Std 0.90231 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.79802 +trainer/policy/normal/std Std 0.645203 +trainer/policy/normal/std Max 6.56276 +trainer/policy/normal/std Min 0.317296 +trainer/policy/normal/log_std Mean 0.991611 +trainer/policy/normal/log_std Std 0.306583 +trainer/policy/normal/log_std Max 1.88141 +trainer/policy/normal/log_std Min -1.14792 +eval/num steps total 575085 +eval/num paths total 576 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.111896 +eval/Actions Std 0.912986 +eval/Actions Max 0.999984 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67767 +time/logging (s) 0.00377343 +time/sampling batch (s) 0.275543 +time/saving (s) 0.00349112 +time/training (s) 6.94617 +time/epoch (s) 9.90665 +time/total (s) 5886.58 +Epoch -425 +---------------------------------- --------------- +2022-05-10 14:49:00.050250 PDT | [2] Epoch -424 finished +---------------------------------- --------------- +epoch -424 +replay_buffer/size 999033 +trainer/num train calls 577000 +trainer/Policy Loss -19.2915 +trainer/Log Pis Mean 24.3055 +trainer/Log Pis Std 13.2907 +trainer/Log Pis Max 62.0731 +trainer/Log Pis Min -5.47798 +trainer/policy/mean Mean -0.00848408 +trainer/policy/mean Std 0.908285 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.7899 +trainer/policy/normal/std Std 0.64381 +trainer/policy/normal/std Max 5.51455 +trainer/policy/normal/std Min 0.248744 +trainer/policy/normal/log_std Mean 0.987766 +trainer/policy/normal/log_std Std 0.311504 +trainer/policy/normal/log_std Max 1.70739 +trainer/policy/normal/log_std Min -1.39133 +eval/num steps total 576085 +eval/num paths total 577 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0171253 +eval/Actions Std 0.932379 +eval/Actions Max 0.999996 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.38101 +time/logging (s) 0.00393306 +time/sampling batch (s) 0.525898 +time/saving (s) 0.00340407 +time/training (s) 6.63773 +time/epoch (s) 9.55197 +time/total (s) 5896.14 +Epoch -424 +---------------------------------- --------------- +2022-05-10 14:49:09.591673 PDT | [2] Epoch -423 finished +---------------------------------- --------------- +epoch -423 +replay_buffer/size 999033 +trainer/num train calls 578000 +trainer/Policy Loss -20.5471 +trainer/Log Pis Mean 24.602 +trainer/Log Pis Std 13.0893 +trainer/Log Pis Max 65.5662 +trainer/Log Pis Min -8.14286 +trainer/policy/mean Mean -0.0496654 +trainer/policy/mean Std 0.904888 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.78048 +trainer/policy/normal/std Std 0.628882 +trainer/policy/normal/std Max 5.69137 +trainer/policy/normal/std Min 0.307431 +trainer/policy/normal/log_std Mean 0.986913 +trainer/policy/normal/log_std Std 0.299005 +trainer/policy/normal/log_std Max 1.73895 +trainer/policy/normal/log_std Min -1.1795 +eval/num steps total 577085 +eval/num paths total 578 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.10324 +eval/Actions Std 0.894024 +eval/Actions Max 0.999996 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.52677 +time/logging (s) 0.00372532 +time/sampling batch (s) 0.526265 +time/saving (s) 0.00352518 +time/training (s) 6.45834 +time/epoch (s) 9.51863 +time/total (s) 5905.66 +Epoch -423 +---------------------------------- --------------- +2022-05-10 14:49:19.846455 PDT | [2] Epoch -422 finished +---------------------------------- --------------- +epoch -422 +replay_buffer/size 999033 +trainer/num train calls 579000 +trainer/Policy Loss -20.4704 +trainer/Log Pis Mean 24.8304 +trainer/Log Pis Std 13.597 +trainer/Log Pis Max 74.7969 +trainer/Log Pis Min -8.46854 +trainer/policy/mean Mean -0.0104365 +trainer/policy/mean Std 0.909947 +trainer/policy/mean Max 0.99998 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.7688 +trainer/policy/normal/std Std 0.655722 +trainer/policy/normal/std Max 6.67636 +trainer/policy/normal/std Min 0.28468 +trainer/policy/normal/log_std Mean 0.977968 +trainer/policy/normal/log_std Std 0.321147 +trainer/policy/normal/log_std Max 1.89857 +trainer/policy/normal/log_std Min -1.25639 +eval/num steps total 578085 +eval/num paths total 579 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0938262 +eval/Actions Std 0.915276 +eval/Actions Max 0.999993 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.41312 +time/logging (s) 0.00391574 +time/sampling batch (s) 0.280452 +time/saving (s) 0.00383107 +time/training (s) 7.53087 +time/epoch (s) 10.2322 +time/total (s) 5915.89 +Epoch -422 +---------------------------------- --------------- +2022-05-10 14:49:30.274447 PDT | [2] Epoch -421 finished +---------------------------------- --------------- +epoch -421 +replay_buffer/size 999033 +trainer/num train calls 580000 +trainer/Policy Loss -19.1031 +trainer/Log Pis Mean 25.3268 +trainer/Log Pis Std 13.3676 +trainer/Log Pis Max 78.8042 +trainer/Log Pis Min -5.4767 +trainer/policy/mean Mean -0.0529592 +trainer/policy/mean Std 0.909131 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.84018 +trainer/policy/normal/std Std 0.64529 +trainer/policy/normal/std Max 5.88619 +trainer/policy/normal/std Min 0.344855 +trainer/policy/normal/log_std Mean 1.007 +trainer/policy/normal/log_std Std 0.305203 +trainer/policy/normal/log_std Max 1.77261 +trainer/policy/normal/log_std Min -1.06463 +eval/num steps total 579085 +eval/num paths total 580 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0720221 +eval/Actions Std 0.913665 +eval/Actions Max 0.999994 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.52881 +time/logging (s) 0.00381207 +time/sampling batch (s) 0.531289 +time/saving (s) 0.00343502 +time/training (s) 7.33685 +time/epoch (s) 10.4042 +time/total (s) 5926.3 +Epoch -421 +---------------------------------- --------------- +2022-05-10 14:49:40.206813 PDT | [2] Epoch -420 finished +---------------------------------- --------------- +epoch -420 +replay_buffer/size 999033 +trainer/num train calls 581000 +trainer/Policy Loss -18.9582 +trainer/Log Pis Mean 24.9439 +trainer/Log Pis Std 13.3342 +trainer/Log Pis Max 76.2933 +trainer/Log Pis Min -5.44609 +trainer/policy/mean Mean -0.0432221 +trainer/policy/mean Std 0.906807 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.81207 +trainer/policy/normal/std Std 0.658762 +trainer/policy/normal/std Max 5.71465 +trainer/policy/normal/std Min 0.298323 +trainer/policy/normal/log_std Mean 0.993508 +trainer/policy/normal/log_std Std 0.323039 +trainer/policy/normal/log_std Max 1.74303 +trainer/policy/normal/log_std Min -1.20958 +eval/num steps total 580085 +eval/num paths total 581 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.113606 +eval/Actions Std 0.884978 +eval/Actions Max 0.999982 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.47339 +time/logging (s) 0.00427597 +time/sampling batch (s) 0.280167 +time/saving (s) 0.00344669 +time/training (s) 7.14855 +time/epoch (s) 9.90983 +time/total (s) 5936.21 +Epoch -420 +---------------------------------- --------------- +2022-05-10 14:49:49.844619 PDT | [2] Epoch -419 finished +---------------------------------- --------------- +epoch -419 +replay_buffer/size 999033 +trainer/num train calls 582000 +trainer/Policy Loss -18.9426 +trainer/Log Pis Mean 23.4731 +trainer/Log Pis Std 12.9169 +trainer/Log Pis Max 71.5876 +trainer/Log Pis Min -13.5482 +trainer/policy/mean Mean -0.0455011 +trainer/policy/mean Std 0.903299 +trainer/policy/mean Max 0.99998 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.78037 +trainer/policy/normal/std Std 0.644502 +trainer/policy/normal/std Max 6.38869 +trainer/policy/normal/std Min 0.285889 +trainer/policy/normal/log_std Mean 0.982678 +trainer/policy/normal/log_std Std 0.322571 +trainer/policy/normal/log_std Max 1.85453 +trainer/policy/normal/log_std Min -1.25215 +eval/num steps total 581085 +eval/num paths total 582 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.346053 +eval/Actions Std 0.836177 +eval/Actions Max 0.999963 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.78642 +time/logging (s) 0.00411673 +time/sampling batch (s) 0.283701 +time/saving (s) 0.00410135 +time/training (s) 6.53548 +time/epoch (s) 9.61381 +time/total (s) 5945.83 +Epoch -419 +---------------------------------- --------------- +2022-05-10 14:49:59.348263 PDT | [2] Epoch -418 finished +---------------------------------- --------------- +epoch -418 +replay_buffer/size 999033 +trainer/num train calls 583000 +trainer/Policy Loss -19.0092 +trainer/Log Pis Mean 24.5565 +trainer/Log Pis Std 12.9849 +trainer/Log Pis Max 68.1167 +trainer/Log Pis Min -7.14445 +trainer/policy/mean Mean -0.0305403 +trainer/policy/mean Std 0.904366 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.81337 +trainer/policy/normal/std Std 0.660917 +trainer/policy/normal/std Max 5.7903 +trainer/policy/normal/std Min 0.287434 +trainer/policy/normal/log_std Mean 0.994832 +trainer/policy/normal/log_std Std 0.317378 +trainer/policy/normal/log_std Max 1.75618 +trainer/policy/normal/log_std Min -1.24676 +eval/num steps total 582085 +eval/num paths total 583 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0295597 +eval/Actions Std 0.965948 +eval/Actions Max 0.999981 +eval/Actions Min -0.999987 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64105 +time/logging (s) 0.00372858 +time/sampling batch (s) 0.279085 +time/saving (s) 0.00347119 +time/training (s) 6.55251 +time/epoch (s) 9.47985 +time/total (s) 5955.31 +Epoch -418 +---------------------------------- --------------- +2022-05-10 14:50:08.945721 PDT | [2] Epoch -417 finished +---------------------------------- --------------- +epoch -417 +replay_buffer/size 999033 +trainer/num train calls 584000 +trainer/Policy Loss -18.4624 +trainer/Log Pis Mean 23.2493 +trainer/Log Pis Std 12.9825 +trainer/Log Pis Max 80.7914 +trainer/Log Pis Min -7.02059 +trainer/policy/mean Mean -0.0493111 +trainer/policy/mean Std 0.901918 +trainer/policy/mean Max 0.999981 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.81116 +trainer/policy/normal/std Std 0.671395 +trainer/policy/normal/std Max 5.16408 +trainer/policy/normal/std Min 0.301425 +trainer/policy/normal/log_std Mean 0.991357 +trainer/policy/normal/log_std Std 0.330227 +trainer/policy/normal/log_std Max 1.64173 +trainer/policy/normal/log_std Min -1.19923 +eval/num steps total 583085 +eval/num paths total 584 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.233197 +eval/Actions Std 0.938018 +eval/Actions Max 0.999994 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54557 +time/logging (s) 0.00376071 +time/sampling batch (s) 0.278957 +time/saving (s) 0.00338413 +time/training (s) 6.74306 +time/epoch (s) 9.57474 +time/total (s) 5964.89 +Epoch -417 +---------------------------------- --------------- +2022-05-10 14:50:18.796515 PDT | [2] Epoch -416 finished +---------------------------------- --------------- +epoch -416 +replay_buffer/size 999033 +trainer/num train calls 585000 +trainer/Policy Loss -20.2405 +trainer/Log Pis Mean 24.6945 +trainer/Log Pis Std 12.6959 +trainer/Log Pis Max 68.1259 +trainer/Log Pis Min -8.74984 +trainer/policy/mean Mean -0.0245849 +trainer/policy/mean Std 0.906184 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.803 +trainer/policy/normal/std Std 0.654597 +trainer/policy/normal/std Max 5.71958 +trainer/policy/normal/std Min 0.336303 +trainer/policy/normal/log_std Mean 0.990587 +trainer/policy/normal/log_std Std 0.321755 +trainer/policy/normal/log_std Max 1.7439 +trainer/policy/normal/log_std Min -1.08974 +eval/num steps total 584085 +eval/num paths total 585 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.26357 +eval/Actions Std 0.887304 +eval/Actions Max 0.999972 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.71221 +time/logging (s) 0.00378017 +time/sampling batch (s) 0.276659 +time/saving (s) 0.00339006 +time/training (s) 6.83211 +time/epoch (s) 9.82815 +time/total (s) 5974.72 +Epoch -416 +---------------------------------- --------------- +2022-05-10 14:50:28.937043 PDT | [2] Epoch -415 finished +---------------------------------- --------------- +epoch -415 +replay_buffer/size 999033 +trainer/num train calls 586000 +trainer/Policy Loss -20.3081 +trainer/Log Pis Mean 23.8138 +trainer/Log Pis Std 12.822 +trainer/Log Pis Max 65.2299 +trainer/Log Pis Min -9.23702 +trainer/policy/mean Mean -0.0375197 +trainer/policy/mean Std 0.912753 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.803 +trainer/policy/normal/std Std 0.649876 +trainer/policy/normal/std Max 6.522 +trainer/policy/normal/std Min 0.278887 +trainer/policy/normal/log_std Mean 0.991539 +trainer/policy/normal/log_std Std 0.317643 +trainer/policy/normal/log_std Max 1.87518 +trainer/policy/normal/log_std Min -1.27695 +eval/num steps total 585085 +eval/num paths total 586 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0818703 +eval/Actions Std 0.903877 +eval/Actions Max 0.999993 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.72728 +time/logging (s) 0.00376272 +time/sampling batch (s) 0.275455 +time/saving (s) 0.00341282 +time/training (s) 7.10784 +time/epoch (s) 10.1177 +time/total (s) 5984.84 +Epoch -415 +---------------------------------- --------------- +2022-05-10 14:50:39.205406 PDT | [2] Epoch -414 finished +---------------------------------- --------------- +epoch -414 +replay_buffer/size 999033 +trainer/num train calls 587000 +trainer/Policy Loss -18.6949 +trainer/Log Pis Mean 23.88 +trainer/Log Pis Std 13.2359 +trainer/Log Pis Max 65.7285 +trainer/Log Pis Min -7.81481 +trainer/policy/mean Mean -0.0467955 +trainer/policy/mean Std 0.907876 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.79208 +trainer/policy/normal/std Std 0.648544 +trainer/policy/normal/std Max 5.29315 +trainer/policy/normal/std Min 0.256113 +trainer/policy/normal/log_std Mean 0.986297 +trainer/policy/normal/log_std Std 0.324757 +trainer/policy/normal/log_std Max 1.66641 +trainer/policy/normal/log_std Min -1.36214 +eval/num steps total 586085 +eval/num paths total 587 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0635866 +eval/Actions Std 0.921729 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59 +time/logging (s) 0.0038413 +time/sampling batch (s) 0.274812 +time/saving (s) 0.00350739 +time/training (s) 7.37363 +time/epoch (s) 10.2458 +time/total (s) 5995.09 +Epoch -414 +---------------------------------- --------------- +2022-05-10 14:50:48.846149 PDT | [2] Epoch -413 finished +---------------------------------- --------------- +epoch -413 +replay_buffer/size 999033 +trainer/num train calls 588000 +trainer/Policy Loss -20.096 +trainer/Log Pis Mean 24.8831 +trainer/Log Pis Std 13.3163 +trainer/Log Pis Max 65.0585 +trainer/Log Pis Min -11.2719 +trainer/policy/mean Mean -0.0297757 +trainer/policy/mean Std 0.904572 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.79724 +trainer/policy/normal/std Std 0.663885 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.299292 +trainer/policy/normal/log_std Mean 0.987983 +trainer/policy/normal/log_std Std 0.3227 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.20634 +eval/num steps total 587085 +eval/num paths total 588 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.206734 +eval/Actions Std 0.888166 +eval/Actions Max 0.999993 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61515 +time/logging (s) 0.00410894 +time/sampling batch (s) 0.524286 +time/saving (s) 0.0038952 +time/training (s) 6.47117 +time/epoch (s) 9.61861 +time/total (s) 6004.71 +Epoch -413 +---------------------------------- --------------- +2022-05-10 14:50:59.071014 PDT | [2] Epoch -412 finished +---------------------------------- --------------- +epoch -412 +replay_buffer/size 999033 +trainer/num train calls 589000 +trainer/Policy Loss -20.759 +trainer/Log Pis Mean 24.9378 +trainer/Log Pis Std 13.3277 +trainer/Log Pis Max 71.6617 +trainer/Log Pis Min -10.9727 +trainer/policy/mean Mean -0.0310283 +trainer/policy/mean Std 0.90531 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.80968 +trainer/policy/normal/std Std 0.650957 +trainer/policy/normal/std Max 6.01723 +trainer/policy/normal/std Min 0.276539 +trainer/policy/normal/log_std Mean 0.993838 +trainer/policy/normal/log_std Std 0.318418 +trainer/policy/normal/log_std Max 1.79463 +trainer/policy/normal/log_std Min -1.2854 +eval/num steps total 588085 +eval/num paths total 589 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.196547 +eval/Actions Std 0.860683 +eval/Actions Max 0.999978 +eval/Actions Min -0.999988 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56623 +time/logging (s) 0.00382102 +time/sampling batch (s) 0.529287 +time/saving (s) 0.00356276 +time/training (s) 7.09836 +time/epoch (s) 10.2013 +time/total (s) 6014.92 +Epoch -412 +---------------------------------- --------------- +2022-05-10 14:51:08.801325 PDT | [2] Epoch -411 finished +---------------------------------- --------------- +epoch -411 +replay_buffer/size 999033 +trainer/num train calls 590000 +trainer/Policy Loss -19.4903 +trainer/Log Pis Mean 24.0169 +trainer/Log Pis Std 12.8267 +trainer/Log Pis Max 72.2598 +trainer/Log Pis Min -5.1928 +trainer/policy/mean Mean -0.0308074 +trainer/policy/mean Std 0.907439 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.7476 +trainer/policy/normal/std Std 0.66478 +trainer/policy/normal/std Max 5.44646 +trainer/policy/normal/std Min 0.278014 +trainer/policy/normal/log_std Mean 0.967634 +trainer/policy/normal/log_std Std 0.332731 +trainer/policy/normal/log_std Max 1.69496 +trainer/policy/normal/log_std Min -1.28008 +eval/num steps total 589085 +eval/num paths total 590 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.270926 +eval/Actions Std 0.840405 +eval/Actions Max 0.999993 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61941 +time/logging (s) 0.00387259 +time/sampling batch (s) 0.274448 +time/saving (s) 0.00337933 +time/training (s) 6.80658 +time/epoch (s) 9.70769 +time/total (s) 6024.63 +Epoch -411 +---------------------------------- --------------- +2022-05-10 14:51:18.686992 PDT | [2] Epoch -410 finished +---------------------------------- --------------- +epoch -410 +replay_buffer/size 999033 +trainer/num train calls 591000 +trainer/Policy Loss -19.8978 +trainer/Log Pis Mean 24.4879 +trainer/Log Pis Std 13.5818 +trainer/Log Pis Max 90.3363 +trainer/Log Pis Min -6.30178 +trainer/policy/mean Mean -0.0414113 +trainer/policy/mean Std 0.909046 +trainer/policy/mean Max 0.999982 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.79276 +trainer/policy/normal/std Std 0.644469 +trainer/policy/normal/std Max 5.11131 +trainer/policy/normal/std Min 0.202307 +trainer/policy/normal/log_std Mean 0.987665 +trainer/policy/normal/log_std Std 0.319324 +trainer/policy/normal/log_std Max 1.63146 +trainer/policy/normal/log_std Min -1.59797 +eval/num steps total 590085 +eval/num paths total 591 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0410264 +eval/Actions Std 0.927464 +eval/Actions Max 0.999996 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.43929 +time/logging (s) 0.00377831 +time/sampling batch (s) 1.03012 +time/saving (s) 0.00338274 +time/training (s) 6.38624 +time/epoch (s) 9.86281 +time/total (s) 6034.49 +Epoch -410 +---------------------------------- --------------- +2022-05-10 14:51:28.903538 PDT | [2] Epoch -409 finished +---------------------------------- --------------- +epoch -409 +replay_buffer/size 999033 +trainer/num train calls 592000 +trainer/Policy Loss -20.4511 +trainer/Log Pis Mean 26.0825 +trainer/Log Pis Std 13.9068 +trainer/Log Pis Max 69.4168 +trainer/Log Pis Min -5.18507 +trainer/policy/mean Mean -0.0544369 +trainer/policy/mean Std 0.907133 +trainer/policy/mean Max 0.999981 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.85288 +trainer/policy/normal/std Std 0.671635 +trainer/policy/normal/std Max 6.78057 +trainer/policy/normal/std Min 0.304857 +trainer/policy/normal/log_std Mean 1.0087 +trainer/policy/normal/log_std Std 0.318332 +trainer/policy/normal/log_std Max 1.91406 +trainer/policy/normal/log_std Min -1.18791 +eval/num steps total 591085 +eval/num paths total 592 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.244393 +eval/Actions Std 0.814422 +eval/Actions Max 0.999995 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59144 +time/logging (s) 0.00382477 +time/sampling batch (s) 0.278313 +time/saving (s) 0.00336192 +time/training (s) 7.31672 +time/epoch (s) 10.1937 +time/total (s) 6044.69 +Epoch -409 +---------------------------------- --------------- +2022-05-10 14:51:38.898216 PDT | [2] Epoch -408 finished +---------------------------------- --------------- +epoch -408 +replay_buffer/size 999033 +trainer/num train calls 593000 +trainer/Policy Loss -18.4267 +trainer/Log Pis Mean 24.8799 +trainer/Log Pis Std 13.2439 +trainer/Log Pis Max 64.1713 +trainer/Log Pis Min -5.21843 +trainer/policy/mean Mean -0.028651 +trainer/policy/mean Std 0.90672 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.85547 +trainer/policy/normal/std Std 0.651321 +trainer/policy/normal/std Max 6.30787 +trainer/policy/normal/std Min 0.382056 +trainer/policy/normal/log_std Mean 1.01273 +trainer/policy/normal/log_std Std 0.303066 +trainer/policy/normal/log_std Max 1.8418 +trainer/policy/normal/log_std Min -0.962187 +eval/num steps total 592085 +eval/num paths total 593 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.116187 +eval/Actions Std 0.897694 +eval/Actions Max 0.999983 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.42584 +time/logging (s) 0.00373879 +time/sampling batch (s) 0.279543 +time/saving (s) 0.00344839 +time/training (s) 7.2591 +time/epoch (s) 9.97167 +time/total (s) 6054.67 +Epoch -408 +---------------------------------- --------------- +2022-05-10 14:51:48.887278 PDT | [2] Epoch -407 finished +---------------------------------- --------------- +epoch -407 +replay_buffer/size 999033 +trainer/num train calls 594000 +trainer/Policy Loss -18.1011 +trainer/Log Pis Mean 23.8881 +trainer/Log Pis Std 13.5631 +trainer/Log Pis Max 82.8182 +trainer/Log Pis Min -10.018 +trainer/policy/mean Mean -0.0197673 +trainer/policy/mean Std 0.904658 +trainer/policy/mean Max 0.999979 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.81928 +trainer/policy/normal/std Std 0.650897 +trainer/policy/normal/std Max 6.29999 +trainer/policy/normal/std Min 0.302697 +trainer/policy/normal/log_std Mean 0.998335 +trainer/policy/normal/log_std Std 0.311837 +trainer/policy/normal/log_std Max 1.84055 +trainer/policy/normal/log_std Min -1.19502 +eval/num steps total 593085 +eval/num paths total 594 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0476037 +eval/Actions Std 0.891745 +eval/Actions Max 0.999996 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64922 +time/logging (s) 0.00405711 +time/sampling batch (s) 0.280671 +time/saving (s) 0.00397627 +time/training (s) 7.02834 +time/epoch (s) 9.96627 +time/total (s) 6064.64 +Epoch -407 +---------------------------------- --------------- +2022-05-10 14:51:59.448448 PDT | [2] Epoch -406 finished +---------------------------------- --------------- +epoch -406 +replay_buffer/size 999033 +trainer/num train calls 595000 +trainer/Policy Loss -18.791 +trainer/Log Pis Mean 24.1615 +trainer/Log Pis Std 13.2435 +trainer/Log Pis Max 61.6346 +trainer/Log Pis Min -12.0255 +trainer/policy/mean Mean -0.0320059 +trainer/policy/mean Std 0.907301 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.81242 +trainer/policy/normal/std Std 0.648321 +trainer/policy/normal/std Max 6.0993 +trainer/policy/normal/std Min 0.285965 +trainer/policy/normal/log_std Mean 0.99633 +trainer/policy/normal/log_std Std 0.308295 +trainer/policy/normal/log_std Max 1.80817 +trainer/policy/normal/log_std Min -1.25189 +eval/num steps total 594085 +eval/num paths total 595 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.102489 +eval/Actions Std 0.887986 +eval/Actions Max 0.999991 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60887 +time/logging (s) 0.0041333 +time/sampling batch (s) 0.279292 +time/saving (s) 0.00367522 +time/training (s) 7.64171 +time/epoch (s) 10.5377 +time/total (s) 6075.18 +Epoch -406 +---------------------------------- --------------- +2022-05-10 14:52:10.240544 PDT | [2] Epoch -405 finished +---------------------------------- --------------- +epoch -405 +replay_buffer/size 999033 +trainer/num train calls 596000 +trainer/Policy Loss -19.7309 +trainer/Log Pis Mean 24.0149 +trainer/Log Pis Std 13.0458 +trainer/Log Pis Max 61.9179 +trainer/Log Pis Min -11.0721 +trainer/policy/mean Mean -0.0490642 +trainer/policy/mean Std 0.904089 +trainer/policy/mean Max 0.99998 +trainer/policy/mean Min -0.999983 +trainer/policy/normal/std Mean 2.77975 +trainer/policy/normal/std Std 0.65971 +trainer/policy/normal/std Max 5.61972 +trainer/policy/normal/std Min 0.29159 +trainer/policy/normal/log_std Mean 0.981725 +trainer/policy/normal/log_std Std 0.32223 +trainer/policy/normal/log_std Max 1.72628 +trainer/policy/normal/log_std Min -1.23241 +eval/num steps total 595085 +eval/num paths total 596 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0108736 +eval/Actions Std 0.902196 +eval/Actions Max 0.999994 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59414 +time/logging (s) 0.00383305 +time/sampling batch (s) 0.775637 +time/saving (s) 0.00342031 +time/training (s) 7.39159 +time/epoch (s) 10.7686 +time/total (s) 6085.95 +Epoch -405 +---------------------------------- --------------- +2022-05-10 14:52:20.610516 PDT | [2] Epoch -404 finished +---------------------------------- --------------- +epoch -404 +replay_buffer/size 999033 +trainer/num train calls 597000 +trainer/Policy Loss -20.188 +trainer/Log Pis Mean 24.6514 +trainer/Log Pis Std 13.0706 +trainer/Log Pis Max 74.7866 +trainer/Log Pis Min -5.51112 +trainer/policy/mean Mean -0.0410304 +trainer/policy/mean Std 0.901493 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.73185 +trainer/policy/normal/std Std 0.658317 +trainer/policy/normal/std Max 5.79534 +trainer/policy/normal/std Min 0.220342 +trainer/policy/normal/log_std Mean 0.961242 +trainer/policy/normal/log_std Std 0.338702 +trainer/policy/normal/log_std Max 1.75705 +trainer/policy/normal/log_std Min -1.51257 +eval/num steps total 596085 +eval/num paths total 597 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.234411 +eval/Actions Std 0.957568 +eval/Actions Max 0.999984 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61643 +time/logging (s) 0.00381866 +time/sampling batch (s) 0.276121 +time/saving (s) 0.00342989 +time/training (s) 7.4473 +time/epoch (s) 10.3471 +time/total (s) 6096.3 +Epoch -404 +---------------------------------- --------------- +2022-05-10 14:52:30.835076 PDT | [2] Epoch -403 finished +---------------------------------- --------------- +epoch -403 +replay_buffer/size 999033 +trainer/num train calls 598000 +trainer/Policy Loss -19.9187 +trainer/Log Pis Mean 24.9691 +trainer/Log Pis Std 13.4075 +trainer/Log Pis Max 74.769 +trainer/Log Pis Min -7.07906 +trainer/policy/mean Mean -0.0437127 +trainer/policy/mean Std 0.909834 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.83348 +trainer/policy/normal/std Std 0.642841 +trainer/policy/normal/std Max 5.95726 +trainer/policy/normal/std Min 0.330562 +trainer/policy/normal/log_std Mean 1.00507 +trainer/policy/normal/log_std Std 0.302276 +trainer/policy/normal/log_std Max 1.78461 +trainer/policy/normal/log_std Min -1.10696 +eval/num steps total 597085 +eval/num paths total 598 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.307344 +eval/Actions Std 0.878239 +eval/Actions Max 0.999993 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61285 +time/logging (s) 0.00369734 +time/sampling batch (s) 0.273947 +time/saving (s) 0.00336842 +time/training (s) 7.30816 +time/epoch (s) 10.202 +time/total (s) 6106.5 +Epoch -403 +---------------------------------- --------------- +2022-05-10 14:52:41.457540 PDT | [2] Epoch -402 finished +---------------------------------- --------------- +epoch -402 +replay_buffer/size 999033 +trainer/num train calls 599000 +trainer/Policy Loss -18.7235 +trainer/Log Pis Mean 24.0238 +trainer/Log Pis Std 12.991 +trainer/Log Pis Max 64.5727 +trainer/Log Pis Min -4.99388 +trainer/policy/mean Mean -0.0222998 +trainer/policy/mean Std 0.905541 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999974 +trainer/policy/normal/std Mean 2.82275 +trainer/policy/normal/std Std 0.645891 +trainer/policy/normal/std Max 5.26528 +trainer/policy/normal/std Min 0.273605 +trainer/policy/normal/log_std Mean 1.00005 +trainer/policy/normal/log_std Std 0.309723 +trainer/policy/normal/log_std Max 1.66113 +trainer/policy/normal/log_std Min -1.29607 +eval/num steps total 598085 +eval/num paths total 599 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.440907 +eval/Actions Std 0.7476 +eval/Actions Max 0.999998 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.46803 +time/logging (s) 0.00380477 +time/sampling batch (s) 0.527393 +time/saving (s) 0.00336982 +time/training (s) 7.59742 +time/epoch (s) 10.6 +time/total (s) 6117.11 +Epoch -402 +---------------------------------- --------------- +2022-05-10 14:52:51.625563 PDT | [2] Epoch -401 finished +---------------------------------- --------------- +epoch -401 +replay_buffer/size 999033 +trainer/num train calls 600000 +trainer/Policy Loss -18.9936 +trainer/Log Pis Mean 24.134 +trainer/Log Pis Std 13.005 +trainer/Log Pis Max 67.8079 +trainer/Log Pis Min -8.91141 +trainer/policy/mean Mean -0.0408578 +trainer/policy/mean Std 0.905997 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.79547 +trainer/policy/normal/std Std 0.656208 +trainer/policy/normal/std Max 5.9775 +trainer/policy/normal/std Min 0.247006 +trainer/policy/normal/log_std Mean 0.988816 +trainer/policy/normal/log_std Std 0.315729 +trainer/policy/normal/log_std Max 1.788 +trainer/policy/normal/log_std Min -1.39834 +eval/num steps total 599085 +eval/num paths total 600 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0893056 +eval/Actions Std 0.91384 +eval/Actions Max 0.999995 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54527 +time/logging (s) 0.00410728 +time/sampling batch (s) 0.276784 +time/saving (s) 0.00703884 +time/training (s) 7.31214 +time/epoch (s) 10.1453 +time/total (s) 6127.25 +Epoch -401 +---------------------------------- --------------- +2022-05-10 14:53:02.701669 PDT | [2] Epoch -400 finished +---------------------------------- --------------- +epoch -400 +replay_buffer/size 999033 +trainer/num train calls 601000 +trainer/Policy Loss -19.2541 +trainer/Log Pis Mean 25.9729 +trainer/Log Pis Std 13.3559 +trainer/Log Pis Max 65.3505 +trainer/Log Pis Min -7.53818 +trainer/policy/mean Mean -0.0491195 +trainer/policy/mean Std 0.906796 +trainer/policy/mean Max 0.999982 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.81289 +trainer/policy/normal/std Std 0.65704 +trainer/policy/normal/std Max 5.33173 +trainer/policy/normal/std Min 0.296113 +trainer/policy/normal/log_std Mean 0.992475 +trainer/policy/normal/log_std Std 0.331732 +trainer/policy/normal/log_std Max 1.67368 +trainer/policy/normal/log_std Min -1.21701 +eval/num steps total 600085 +eval/num paths total 601 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.319528 +eval/Actions Std 0.868073 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68078 +time/logging (s) 0.00368105 +time/sampling batch (s) 0.279131 +time/saving (s) 0.00338129 +time/training (s) 8.08543 +time/epoch (s) 11.0524 +time/total (s) 6138.31 +Epoch -400 +---------------------------------- --------------- +2022-05-10 14:53:13.024589 PDT | [2] Epoch -399 finished +---------------------------------- --------------- +epoch -399 +replay_buffer/size 999033 +trainer/num train calls 602000 +trainer/Policy Loss -18.6946 +trainer/Log Pis Mean 23.9356 +trainer/Log Pis Std 13.4324 +trainer/Log Pis Max 65.0594 +trainer/Log Pis Min -6.24303 +trainer/policy/mean Mean -0.0391577 +trainer/policy/mean Std 0.90272 +trainer/policy/mean Max 0.999978 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.77828 +trainer/policy/normal/std Std 0.646853 +trainer/policy/normal/std Max 6.61379 +trainer/policy/normal/std Min 0.326891 +trainer/policy/normal/log_std Mean 0.983513 +trainer/policy/normal/log_std Std 0.310736 +trainer/policy/normal/log_std Max 1.88916 +trainer/policy/normal/log_std Min -1.11813 +eval/num steps total 601085 +eval/num paths total 602 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.138726 +eval/Actions Std 0.869788 +eval/Actions Max 0.999995 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.75561 +time/logging (s) 0.00373804 +time/sampling batch (s) 0.526253 +time/saving (s) 0.0033914 +time/training (s) 7.01141 +time/epoch (s) 10.3004 +time/total (s) 6148.61 +Epoch -399 +---------------------------------- --------------- +2022-05-10 14:53:23.379956 PDT | [2] Epoch -398 finished +---------------------------------- --------------- +epoch -398 +replay_buffer/size 999033 +trainer/num train calls 603000 +trainer/Policy Loss -19.4234 +trainer/Log Pis Mean 24.2438 +trainer/Log Pis Std 12.0891 +trainer/Log Pis Max 71.5027 +trainer/Log Pis Min -5.21604 +trainer/policy/mean Mean -0.0390143 +trainer/policy/mean Std 0.911402 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.78738 +trainer/policy/normal/std Std 0.664855 +trainer/policy/normal/std Max 5.47653 +trainer/policy/normal/std Min 0.296874 +trainer/policy/normal/log_std Mean 0.983123 +trainer/policy/normal/log_std Std 0.329682 +trainer/policy/normal/log_std Max 1.70047 +trainer/policy/normal/log_std Min -1.21445 +eval/num steps total 602085 +eval/num paths total 603 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.112701 +eval/Actions Std 0.926303 +eval/Actions Max 0.999995 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61708 +time/logging (s) 0.0037276 +time/sampling batch (s) 0.274735 +time/saving (s) 0.00343523 +time/training (s) 7.43372 +time/epoch (s) 10.3327 +time/total (s) 6158.95 +Epoch -398 +---------------------------------- --------------- +2022-05-10 14:53:33.523775 PDT | [2] Epoch -397 finished +---------------------------------- --------------- +epoch -397 +replay_buffer/size 999033 +trainer/num train calls 604000 +trainer/Policy Loss -19.4569 +trainer/Log Pis Mean 24.1084 +trainer/Log Pis Std 12.4642 +trainer/Log Pis Max 62.8154 +trainer/Log Pis Min -6.72714 +trainer/policy/mean Mean -0.0468527 +trainer/policy/mean Std 0.903046 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.77831 +trainer/policy/normal/std Std 0.66999 +trainer/policy/normal/std Max 5.58922 +trainer/policy/normal/std Min 0.267055 +trainer/policy/normal/log_std Mean 0.97866 +trainer/policy/normal/log_std Std 0.334737 +trainer/policy/normal/log_std Max 1.72084 +trainer/policy/normal/log_std Min -1.3203 +eval/num steps total 603085 +eval/num paths total 604 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0843572 +eval/Actions Std 0.907423 +eval/Actions Max 0.999997 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60155 +time/logging (s) 0.0038472 +time/sampling batch (s) 0.275846 +time/saving (s) 0.00350376 +time/training (s) 7.23651 +time/epoch (s) 10.1213 +time/total (s) 6169.07 +Epoch -397 +---------------------------------- --------------- +2022-05-10 14:53:43.467807 PDT | [2] Epoch -396 finished +---------------------------------- --------------- +epoch -396 +replay_buffer/size 999033 +trainer/num train calls 605000 +trainer/Policy Loss -19.743 +trainer/Log Pis Mean 25.2767 +trainer/Log Pis Std 12.6592 +trainer/Log Pis Max 63.7588 +trainer/Log Pis Min -8.11579 +trainer/policy/mean Mean -0.0441018 +trainer/policy/mean Std 0.902985 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999983 +trainer/policy/normal/std Mean 2.73383 +trainer/policy/normal/std Std 0.644947 +trainer/policy/normal/std Max 7.07759 +trainer/policy/normal/std Min 0.275829 +trainer/policy/normal/log_std Mean 0.964025 +trainer/policy/normal/log_std Std 0.329487 +trainer/policy/normal/log_std Max 1.95693 +trainer/policy/normal/log_std Min -1.28797 +eval/num steps total 604085 +eval/num paths total 605 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.144698 +eval/Actions Std 0.884593 +eval/Actions Max 0.999996 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.46408 +time/logging (s) 0.00411184 +time/sampling batch (s) 0.274556 +time/saving (s) 0.00403325 +time/training (s) 7.17477 +time/epoch (s) 9.92155 +time/total (s) 6179 +Epoch -396 +---------------------------------- --------------- +2022-05-10 14:53:54.424779 PDT | [2] Epoch -395 finished +---------------------------------- --------------- +epoch -395 +replay_buffer/size 999033 +trainer/num train calls 606000 +trainer/Policy Loss -19.9604 +trainer/Log Pis Mean 24.4171 +trainer/Log Pis Std 13.6635 +trainer/Log Pis Max 79.1784 +trainer/Log Pis Min -8.71371 +trainer/policy/mean Mean -0.0330072 +trainer/policy/mean Std 0.909695 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.84407 +trainer/policy/normal/std Std 0.647377 +trainer/policy/normal/std Max 5.89119 +trainer/policy/normal/std Min 0.306646 +trainer/policy/normal/log_std Mean 1.00931 +trainer/policy/normal/log_std Std 0.29926 +trainer/policy/normal/log_std Max 1.77346 +trainer/policy/normal/log_std Min -1.18206 +eval/num steps total 605085 +eval/num paths total 606 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.074204 +eval/Actions Std 0.885524 +eval/Actions Max 0.999978 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.37833 +time/logging (s) 0.0037076 +time/sampling batch (s) 0.525584 +time/saving (s) 0.00345779 +time/training (s) 8.02246 +time/epoch (s) 10.9335 +time/total (s) 6189.93 +Epoch -395 +---------------------------------- --------------- +2022-05-10 14:54:04.560873 PDT | [2] Epoch -394 finished +---------------------------------- --------------- +epoch -394 +replay_buffer/size 999033 +trainer/num train calls 607000 +trainer/Policy Loss -20.2979 +trainer/Log Pis Mean 25.1848 +trainer/Log Pis Std 13.4964 +trainer/Log Pis Max 64.6675 +trainer/Log Pis Min -7.07951 +trainer/policy/mean Mean -0.0237276 +trainer/policy/mean Std 0.913155 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.82955 +trainer/policy/normal/std Std 0.659498 +trainer/policy/normal/std Max 5.67057 +trainer/policy/normal/std Min 0.328536 +trainer/policy/normal/log_std Mean 1.00089 +trainer/policy/normal/log_std Std 0.315887 +trainer/policy/normal/log_std Max 1.73529 +trainer/policy/normal/log_std Min -1.11311 +eval/num steps total 606085 +eval/num paths total 607 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0704207 +eval/Actions Std 0.892743 +eval/Actions Max 0.999994 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66471 +time/logging (s) 0.00368342 +time/sampling batch (s) 0.274687 +time/saving (s) 0.00339361 +time/training (s) 7.16692 +time/epoch (s) 10.1134 +time/total (s) 6200.05 +Epoch -394 +---------------------------------- --------------- +2022-05-10 14:54:13.978429 PDT | [2] Epoch -393 finished +---------------------------------- --------------- +epoch -393 +replay_buffer/size 999033 +trainer/num train calls 608000 +trainer/Policy Loss -19.1407 +trainer/Log Pis Mean 23.8054 +trainer/Log Pis Std 12.9546 +trainer/Log Pis Max 59.1234 +trainer/Log Pis Min -11.6482 +trainer/policy/mean Mean -0.0178091 +trainer/policy/mean Std 0.904089 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.85835 +trainer/policy/normal/std Std 0.660986 +trainer/policy/normal/std Max 5.72209 +trainer/policy/normal/std Min 0.323407 +trainer/policy/normal/log_std Mean 1.01335 +trainer/policy/normal/log_std Std 0.303432 +trainer/policy/normal/log_std Max 1.74433 +trainer/policy/normal/log_std Min -1.12884 +eval/num steps total 607085 +eval/num paths total 608 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.136634 +eval/Actions Std 0.905142 +eval/Actions Max 0.999997 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50774 +time/logging (s) 0.00380258 +time/sampling batch (s) 0.274639 +time/saving (s) 0.00338318 +time/training (s) 6.60542 +time/epoch (s) 9.39499 +time/total (s) 6209.45 +Epoch -393 +---------------------------------- --------------- +2022-05-10 14:54:24.201357 PDT | [2] Epoch -392 finished +---------------------------------- --------------- +epoch -392 +replay_buffer/size 999033 +trainer/num train calls 609000 +trainer/Policy Loss -19.1185 +trainer/Log Pis Mean 24.1985 +trainer/Log Pis Std 13.308 +trainer/Log Pis Max 66.9319 +trainer/Log Pis Min -15.362 +trainer/policy/mean Mean -0.0334267 +trainer/policy/mean Std 0.906591 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.88002 +trainer/policy/normal/std Std 0.658193 +trainer/policy/normal/std Max 5.62993 +trainer/policy/normal/std Min 0.298892 +trainer/policy/normal/log_std Mean 1.02081 +trainer/policy/normal/log_std Std 0.305616 +trainer/policy/normal/log_std Max 1.7281 +trainer/policy/normal/log_std Min -1.20767 +eval/num steps total 608085 +eval/num paths total 609 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.316306 +eval/Actions Std 0.865035 +eval/Actions Max 0.999994 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50048 +time/logging (s) 0.00371086 +time/sampling batch (s) 0.525547 +time/saving (s) 0.0034602 +time/training (s) 7.16706 +time/epoch (s) 10.2003 +time/total (s) 6219.65 +Epoch -392 +---------------------------------- --------------- +2022-05-10 14:54:35.032615 PDT | [2] Epoch -391 finished +---------------------------------- --------------- +epoch -391 +replay_buffer/size 999033 +trainer/num train calls 610000 +trainer/Policy Loss -20.0331 +trainer/Log Pis Mean 25.1934 +trainer/Log Pis Std 13.8373 +trainer/Log Pis Max 72.0796 +trainer/Log Pis Min -7.98589 +trainer/policy/mean Mean -0.0315997 +trainer/policy/mean Std 0.90263 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.82729 +trainer/policy/normal/std Std 0.681335 +trainer/policy/normal/std Max 5.9882 +trainer/policy/normal/std Min 0.309258 +trainer/policy/normal/log_std Mean 0.997018 +trainer/policy/normal/log_std Std 0.329149 +trainer/policy/normal/log_std Max 1.78979 +trainer/policy/normal/log_std Min -1.17358 +eval/num steps total 609085 +eval/num paths total 610 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.436309 +eval/Actions Std 0.860789 +eval/Actions Max 0.999994 +eval/Actions Min -0.999988 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6385 +time/logging (s) 0.00369607 +time/sampling batch (s) 0.2744 +time/saving (s) 0.00339275 +time/training (s) 7.88868 +time/epoch (s) 10.8087 +time/total (s) 6230.46 +Epoch -391 +---------------------------------- --------------- +2022-05-10 14:54:45.398804 PDT | [2] Epoch -390 finished +---------------------------------- --------------- +epoch -390 +replay_buffer/size 999033 +trainer/num train calls 611000 +trainer/Policy Loss -18.6115 +trainer/Log Pis Mean 25.0553 +trainer/Log Pis Std 13.3485 +trainer/Log Pis Max 69.1572 +trainer/Log Pis Min -7.39581 +trainer/policy/mean Mean -0.0590918 +trainer/policy/mean Std 0.906504 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.81357 +trainer/policy/normal/std Std 0.665075 +trainer/policy/normal/std Max 6.58054 +trainer/policy/normal/std Min 0.322226 +trainer/policy/normal/log_std Mean 0.994391 +trainer/policy/normal/log_std Std 0.319791 +trainer/policy/normal/log_std Max 1.88412 +trainer/policy/normal/log_std Min -1.1325 +eval/num steps total 610085 +eval/num paths total 611 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.139483 +eval/Actions Std 0.896499 +eval/Actions Max 0.999993 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63421 +time/logging (s) 0.00397587 +time/sampling batch (s) 0.274978 +time/saving (s) 0.00389115 +time/training (s) 7.42684 +time/epoch (s) 10.3439 +time/total (s) 6240.81 +Epoch -390 +---------------------------------- --------------- +2022-05-10 14:54:55.509545 PDT | [2] Epoch -389 finished +---------------------------------- --------------- +epoch -389 +replay_buffer/size 999033 +trainer/num train calls 612000 +trainer/Policy Loss -19.4493 +trainer/Log Pis Mean 23.8619 +trainer/Log Pis Std 13.138 +trainer/Log Pis Max 70.8858 +trainer/Log Pis Min -4.84161 +trainer/policy/mean Mean -0.035858 +trainer/policy/mean Std 0.907583 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.80971 +trainer/policy/normal/std Std 0.649583 +trainer/policy/normal/std Max 6.29008 +trainer/policy/normal/std Min 0.281319 +trainer/policy/normal/log_std Mean 0.993546 +trainer/policy/normal/log_std Std 0.319046 +trainer/policy/normal/log_std Max 1.83897 +trainer/policy/normal/log_std Min -1.26826 +eval/num steps total 611085 +eval/num paths total 612 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0700964 +eval/Actions Std 0.901366 +eval/Actions Max 0.999998 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60726 +time/logging (s) 0.00374802 +time/sampling batch (s) 0.274279 +time/saving (s) 0.0034318 +time/training (s) 7.19886 +time/epoch (s) 10.0876 +time/total (s) 6250.9 +Epoch -389 +---------------------------------- --------------- +2022-05-10 14:55:05.414938 PDT | [2] Epoch -388 finished +---------------------------------- --------------- +epoch -388 +replay_buffer/size 999033 +trainer/num train calls 613000 +trainer/Policy Loss -18.824 +trainer/Log Pis Mean 24.1165 +trainer/Log Pis Std 12.1784 +trainer/Log Pis Max 66.6297 +trainer/Log Pis Min -4.76826 +trainer/policy/mean Mean -0.022815 +trainer/policy/mean Std 0.904655 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.80518 +trainer/policy/normal/std Std 0.647474 +trainer/policy/normal/std Max 6.32034 +trainer/policy/normal/std Min 0.273983 +trainer/policy/normal/log_std Mean 0.992374 +trainer/policy/normal/log_std Std 0.317543 +trainer/policy/normal/log_std Max 1.84377 +trainer/policy/normal/log_std Min -1.29469 +eval/num steps total 612085 +eval/num paths total 613 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0727297 +eval/Actions Std 0.8909 +eval/Actions Max 0.999997 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63828 +time/logging (s) 0.00379326 +time/sampling batch (s) 0.274443 +time/saving (s) 0.00340224 +time/training (s) 6.96252 +time/epoch (s) 9.88244 +time/total (s) 6260.79 +Epoch -388 +---------------------------------- --------------- +2022-05-10 14:55:15.748751 PDT | [2] Epoch -387 finished +---------------------------------- --------------- +epoch -387 +replay_buffer/size 999033 +trainer/num train calls 614000 +trainer/Policy Loss -18.508 +trainer/Log Pis Mean 24.0223 +trainer/Log Pis Std 13.2642 +trainer/Log Pis Max 62.4494 +trainer/Log Pis Min -10.2 +trainer/policy/mean Mean -0.017028 +trainer/policy/mean Std 0.908078 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.77217 +trainer/policy/normal/std Std 0.623795 +trainer/policy/normal/std Max 4.91924 +trainer/policy/normal/std Min 0.286073 +trainer/policy/normal/log_std Mean 0.983539 +trainer/policy/normal/log_std Std 0.302343 +trainer/policy/normal/log_std Max 1.59315 +trainer/policy/normal/log_std Min -1.25151 +eval/num steps total 613085 +eval/num paths total 614 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.181964 +eval/Actions Std 0.950861 +eval/Actions Max 0.999996 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55759 +time/logging (s) 0.00370292 +time/sampling batch (s) 0.273756 +time/saving (s) 0.00338973 +time/training (s) 7.47256 +time/epoch (s) 10.311 +time/total (s) 6271.1 +Epoch -387 +---------------------------------- --------------- +2022-05-10 14:55:26.310849 PDT | [2] Epoch -386 finished +---------------------------------- --------------- +epoch -386 +replay_buffer/size 999033 +trainer/num train calls 615000 +trainer/Policy Loss -19.9975 +trainer/Log Pis Mean 23.8558 +trainer/Log Pis Std 13.2711 +trainer/Log Pis Max 64.9857 +trainer/Log Pis Min -8.13557 +trainer/policy/mean Mean -0.0456054 +trainer/policy/mean Std 0.902149 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.79994 +trainer/policy/normal/std Std 0.6773 +trainer/policy/normal/std Max 5.2601 +trainer/policy/normal/std Min 0.23949 +trainer/policy/normal/log_std Mean 0.985596 +trainer/policy/normal/log_std Std 0.339242 +trainer/policy/normal/log_std Max 1.66015 +trainer/policy/normal/log_std Min -1.42924 +eval/num steps total 614085 +eval/num paths total 615 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0496648 +eval/Actions Std 0.913429 +eval/Actions Max 0.999995 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59536 +time/logging (s) 0.00367169 +time/sampling batch (s) 0.276288 +time/saving (s) 0.00339723 +time/training (s) 7.66055 +time/epoch (s) 10.5393 +time/total (s) 6281.64 +Epoch -386 +---------------------------------- --------------- +2022-05-10 14:55:37.251789 PDT | [2] Epoch -385 finished +---------------------------------- --------------- +epoch -385 +replay_buffer/size 999033 +trainer/num train calls 616000 +trainer/Policy Loss -19.3811 +trainer/Log Pis Mean 23.7264 +trainer/Log Pis Std 12.9079 +trainer/Log Pis Max 71.3242 +trainer/Log Pis Min -9.28861 +trainer/policy/mean Mean -0.0437812 +trainer/policy/mean Std 0.906427 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.81644 +trainer/policy/normal/std Std 0.64438 +trainer/policy/normal/std Max 5.64284 +trainer/policy/normal/std Min 0.347153 +trainer/policy/normal/log_std Mean 0.99917 +trainer/policy/normal/log_std Std 0.301294 +trainer/policy/normal/log_std Max 1.73039 +trainer/policy/normal/log_std Min -1.05799 +eval/num steps total 615085 +eval/num paths total 616 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.109127 +eval/Actions Std 0.90544 +eval/Actions Max 0.999997 +eval/Actions Min -0.999986 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68616 +time/logging (s) 0.00402006 +time/sampling batch (s) 0.526682 +time/saving (s) 0.0039228 +time/training (s) 7.69748 +time/epoch (s) 10.9183 +time/total (s) 6292.56 +Epoch -385 +---------------------------------- --------------- +2022-05-10 14:55:47.254207 PDT | [2] Epoch -384 finished +---------------------------------- --------------- +epoch -384 +replay_buffer/size 999033 +trainer/num train calls 617000 +trainer/Policy Loss -20.6481 +trainer/Log Pis Mean 23.9886 +trainer/Log Pis Std 13.3312 +trainer/Log Pis Max 69.6923 +trainer/Log Pis Min -3.76875 +trainer/policy/mean Mean -0.022561 +trainer/policy/mean Std 0.910079 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.78263 +trainer/policy/normal/std Std 0.641579 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.291971 +trainer/policy/normal/log_std Mean 0.985794 +trainer/policy/normal/log_std Std 0.308242 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.2311 +eval/num steps total 616085 +eval/num paths total 617 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.105186 +eval/Actions Std 0.901038 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60866 +time/logging (s) 0.00386852 +time/sampling batch (s) 0.526197 +time/saving (s) 0.00356903 +time/training (s) 6.83682 +time/epoch (s) 9.97912 +time/total (s) 6302.55 +Epoch -384 +---------------------------------- --------------- +2022-05-10 14:55:57.394673 PDT | [2] Epoch -383 finished +---------------------------------- --------------- +epoch -383 +replay_buffer/size 999033 +trainer/num train calls 618000 +trainer/Policy Loss -19.1298 +trainer/Log Pis Mean 24.9855 +trainer/Log Pis Std 13.2989 +trainer/Log Pis Max 77.1834 +trainer/Log Pis Min -10.7638 +trainer/policy/mean Mean -0.0214801 +trainer/policy/mean Std 0.906579 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.79055 +trainer/policy/normal/std Std 0.702412 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.291508 +trainer/policy/normal/log_std Mean 0.980829 +trainer/policy/normal/log_std Std 0.340705 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.23269 +eval/num steps total 617085 +eval/num paths total 618 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.321432 +eval/Actions Std 0.895621 +eval/Actions Max 0.999981 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68758 +time/logging (s) 0.00385609 +time/sampling batch (s) 1.02662 +time/saving (s) 0.00350281 +time/training (s) 6.39608 +time/epoch (s) 10.1176 +time/total (s) 6312.67 +Epoch -383 +---------------------------------- --------------- +2022-05-10 14:56:07.226818 PDT | [2] Epoch -382 finished +---------------------------------- --------------- +epoch -382 +replay_buffer/size 999033 +trainer/num train calls 619000 +trainer/Policy Loss -20.2546 +trainer/Log Pis Mean 23.8315 +trainer/Log Pis Std 13.5345 +trainer/Log Pis Max 73.0209 +trainer/Log Pis Min -8.23409 +trainer/policy/mean Mean -0.0326602 +trainer/policy/mean Std 0.904746 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.86435 +trainer/policy/normal/std Std 0.664416 +trainer/policy/normal/std Max 5.60507 +trainer/policy/normal/std Min 0.280339 +trainer/policy/normal/log_std Mean 1.01207 +trainer/policy/normal/log_std Std 0.324352 +trainer/policy/normal/log_std Max 1.72367 +trainer/policy/normal/log_std Min -1.27176 +eval/num steps total 618085 +eval/num paths total 619 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.215522 +eval/Actions Std 0.890112 +eval/Actions Max 0.999997 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.72149 +time/logging (s) 0.00372773 +time/sampling batch (s) 0.278455 +time/saving (s) 0.0033829 +time/training (s) 6.80201 +time/epoch (s) 9.80906 +time/total (s) 6322.48 +Epoch -382 +---------------------------------- --------------- +2022-05-10 14:56:17.151798 PDT | [2] Epoch -381 finished +---------------------------------- --------------- +epoch -381 +replay_buffer/size 999033 +trainer/num train calls 620000 +trainer/Policy Loss -19.3309 +trainer/Log Pis Mean 24.2874 +trainer/Log Pis Std 13.6642 +trainer/Log Pis Max 74.2016 +trainer/Log Pis Min -5.4366 +trainer/policy/mean Mean -0.02292 +trainer/policy/mean Std 0.91283 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.84245 +trainer/policy/normal/std Std 0.649806 +trainer/policy/normal/std Max 6.54522 +trainer/policy/normal/std Min 0.297044 +trainer/policy/normal/log_std Mean 1.00822 +trainer/policy/normal/log_std Std 0.3039 +trainer/policy/normal/log_std Max 1.87874 +trainer/policy/normal/log_std Min -1.21387 +eval/num steps total 619085 +eval/num paths total 620 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0978982 +eval/Actions Std 0.900817 +eval/Actions Max 0.999996 +eval/Actions Min -0.999986 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58736 +time/logging (s) 0.00435924 +time/sampling batch (s) 0.28042 +time/saving (s) 0.00343744 +time/training (s) 7.02696 +time/epoch (s) 9.90254 +time/total (s) 6332.38 +Epoch -381 +---------------------------------- --------------- +2022-05-10 14:56:27.340082 PDT | [2] Epoch -380 finished +---------------------------------- --------------- +epoch -380 +replay_buffer/size 999033 +trainer/num train calls 621000 +trainer/Policy Loss -20.0331 +trainer/Log Pis Mean 25.1554 +trainer/Log Pis Std 13.036 +trainer/Log Pis Max 61.8818 +trainer/Log Pis Min -5.74962 +trainer/policy/mean Mean -0.0381458 +trainer/policy/mean Std 0.908121 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.96259 +trainer/policy/normal/std Std 0.698735 +trainer/policy/normal/std Max 7.22633 +trainer/policy/normal/std Min 0.309042 +trainer/policy/normal/log_std Mean 1.04735 +trainer/policy/normal/log_std Std 0.311335 +trainer/policy/normal/log_std Max 1.97773 +trainer/policy/normal/log_std Min -1.17428 +eval/num steps total 620085 +eval/num paths total 621 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.194101 +eval/Actions Std 0.930693 +eval/Actions Max 0.999986 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.84842 +time/logging (s) 0.00379449 +time/sampling batch (s) 0.282956 +time/saving (s) 0.00344287 +time/training (s) 7.02537 +time/epoch (s) 10.164 +time/total (s) 6342.55 +Epoch -380 +---------------------------------- --------------- +2022-05-10 14:56:37.311381 PDT | [2] Epoch -379 finished +---------------------------------- --------------- +epoch -379 +replay_buffer/size 999033 +trainer/num train calls 622000 +trainer/Policy Loss -20.3823 +trainer/Log Pis Mean 23.5064 +trainer/Log Pis Std 13.1313 +trainer/Log Pis Max 73.3751 +trainer/Log Pis Min -8.1752 +trainer/policy/mean Mean -0.0268004 +trainer/policy/mean Std 0.90711 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.77632 +trainer/policy/normal/std Std 0.655291 +trainer/policy/normal/std Max 5.62974 +trainer/policy/normal/std Min 0.28554 +trainer/policy/normal/log_std Mean 0.979972 +trainer/policy/normal/log_std Std 0.326122 +trainer/policy/normal/log_std Max 1.72806 +trainer/policy/normal/log_std Min -1.25337 +eval/num steps total 621085 +eval/num paths total 622 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.280235 +eval/Actions Std 0.896697 +eval/Actions Max 0.999996 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56606 +time/logging (s) 0.00392501 +time/sampling batch (s) 0.280817 +time/saving (s) 0.00353708 +time/training (s) 7.09398 +time/epoch (s) 9.94832 +time/total (s) 6352.5 +Epoch -379 +---------------------------------- --------------- +2022-05-10 14:56:47.376081 PDT | [2] Epoch -378 finished +---------------------------------- --------------- +epoch -378 +replay_buffer/size 999033 +trainer/num train calls 623000 +trainer/Policy Loss -20.0577 +trainer/Log Pis Mean 24.8698 +trainer/Log Pis Std 13.3205 +trainer/Log Pis Max 74.4254 +trainer/Log Pis Min -8.62043 +trainer/policy/mean Mean -0.054384 +trainer/policy/mean Std 0.905595 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.77493 +trainer/policy/normal/std Std 0.667057 +trainer/policy/normal/std Max 5.60373 +trainer/policy/normal/std Min 0.245568 +trainer/policy/normal/log_std Mean 0.978621 +trainer/policy/normal/log_std Std 0.328448 +trainer/policy/normal/log_std Max 1.72343 +trainer/policy/normal/log_std Min -1.40418 +eval/num steps total 622085 +eval/num paths total 623 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0953979 +eval/Actions Std 0.900041 +eval/Actions Max 0.999996 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.71097 +time/logging (s) 0.0036551 +time/sampling batch (s) 0.282259 +time/saving (s) 0.00340446 +time/training (s) 7.04077 +time/epoch (s) 10.0411 +time/total (s) 6362.55 +Epoch -378 +---------------------------------- --------------- +2022-05-10 14:56:58.270801 PDT | [2] Epoch -377 finished +---------------------------------- --------------- +epoch -377 +replay_buffer/size 999033 +trainer/num train calls 624000 +trainer/Policy Loss -19.2698 +trainer/Log Pis Mean 24.8093 +trainer/Log Pis Std 13.9374 +trainer/Log Pis Max 77.3985 +trainer/Log Pis Min -10.5285 +trainer/policy/mean Mean -0.0405538 +trainer/policy/mean Std 0.903975 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.79241 +trainer/policy/normal/std Std 0.646064 +trainer/policy/normal/std Max 5.06956 +trainer/policy/normal/std Min 0.312531 +trainer/policy/normal/log_std Mean 0.988411 +trainer/policy/normal/log_std Std 0.312441 +trainer/policy/normal/log_std Max 1.62325 +trainer/policy/normal/log_std Min -1.16305 +eval/num steps total 623085 +eval/num paths total 624 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.108773 +eval/Actions Std 0.91851 +eval/Actions Max 0.999994 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.72457 +time/logging (s) 0.00365194 +time/sampling batch (s) 0.27741 +time/saving (s) 0.00340082 +time/training (s) 7.86294 +time/epoch (s) 10.872 +time/total (s) 6373.42 +Epoch -377 +---------------------------------- --------------- +2022-05-10 14:57:08.560281 PDT | [2] Epoch -376 finished +---------------------------------- --------------- +epoch -376 +replay_buffer/size 999033 +trainer/num train calls 625000 +trainer/Policy Loss -18.8009 +trainer/Log Pis Mean 24.4946 +trainer/Log Pis Std 12.6396 +trainer/Log Pis Max 65.3541 +trainer/Log Pis Min -8.93468 +trainer/policy/mean Mean -0.0502637 +trainer/policy/mean Std 0.907167 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.79806 +trainer/policy/normal/std Std 0.648204 +trainer/policy/normal/std Max 5.71529 +trainer/policy/normal/std Min 0.35684 +trainer/policy/normal/log_std Mean 0.991138 +trainer/policy/normal/log_std Std 0.308202 +trainer/policy/normal/log_std Max 1.74314 +trainer/policy/normal/log_std Min -1.03047 +eval/num steps total 624085 +eval/num paths total 625 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.295576 +eval/Actions Std 0.84197 +eval/Actions Max 0.999981 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.75336 +time/logging (s) 0.0037422 +time/sampling batch (s) 0.526681 +time/saving (s) 0.00335801 +time/training (s) 6.97976 +time/epoch (s) 10.2669 +time/total (s) 6383.69 +Epoch -376 +---------------------------------- --------------- +2022-05-10 14:57:18.260475 PDT | [2] Epoch -375 finished +---------------------------------- --------------- +epoch -375 +replay_buffer/size 999033 +trainer/num train calls 626000 +trainer/Policy Loss -20.1969 +trainer/Log Pis Mean 25.5236 +trainer/Log Pis Std 13.8043 +trainer/Log Pis Max 61.5455 +trainer/Log Pis Min -3.52271 +trainer/policy/mean Mean -0.0259229 +trainer/policy/mean Std 0.90861 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.7994 +trainer/policy/normal/std Std 0.64516 +trainer/policy/normal/std Max 7.07146 +trainer/policy/normal/std Min 0.280131 +trainer/policy/normal/log_std Mean 0.991731 +trainer/policy/normal/log_std Std 0.308927 +trainer/policy/normal/log_std Max 1.95607 +trainer/policy/normal/log_std Min -1.2725 +eval/num steps total 625085 +eval/num paths total 626 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.148361 +eval/Actions Std 0.906397 +eval/Actions Max 0.999991 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59341 +time/logging (s) 0.00378791 +time/sampling batch (s) 0.523118 +time/saving (s) 0.00340678 +time/training (s) 6.55403 +time/epoch (s) 9.67775 +time/total (s) 6393.37 +Epoch -375 +---------------------------------- --------------- +2022-05-10 14:57:28.793287 PDT | [2] Epoch -374 finished +---------------------------------- --------------- +epoch -374 +replay_buffer/size 999033 +trainer/num train calls 627000 +trainer/Policy Loss -19.3811 +trainer/Log Pis Mean 23.9407 +trainer/Log Pis Std 13.2243 +trainer/Log Pis Max 67.3943 +trainer/Log Pis Min -11.0156 +trainer/policy/mean Mean -0.0250184 +trainer/policy/mean Std 0.905305 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.78948 +trainer/policy/normal/std Std 0.653221 +trainer/policy/normal/std Max 6.82514 +trainer/policy/normal/std Min 0.322587 +trainer/policy/normal/log_std Mean 0.987913 +trainer/policy/normal/log_std Std 0.307053 +trainer/policy/normal/log_std Max 1.92061 +trainer/policy/normal/log_std Min -1.13138 +eval/num steps total 626085 +eval/num paths total 627 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0952541 +eval/Actions Std 0.912879 +eval/Actions Max 0.999993 +eval/Actions Min -0.999984 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67816 +time/logging (s) 0.00403949 +time/sampling batch (s) 0.27807 +time/saving (s) 0.00398682 +time/training (s) 7.5456 +time/epoch (s) 10.5099 +time/total (s) 6403.88 +Epoch -374 +---------------------------------- --------------- +2022-05-10 14:57:39.388943 PDT | [2] Epoch -373 finished +---------------------------------- --------------- +epoch -373 +replay_buffer/size 999033 +trainer/num train calls 628000 +trainer/Policy Loss -20.3137 +trainer/Log Pis Mean 24.586 +trainer/Log Pis Std 13.5244 +trainer/Log Pis Max 73.6871 +trainer/Log Pis Min -7.37762 +trainer/policy/mean Mean -0.0331361 +trainer/policy/mean Std 0.910704 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.85969 +trainer/policy/normal/std Std 0.681684 +trainer/policy/normal/std Max 7.23003 +trainer/policy/normal/std Min 0.227664 +trainer/policy/normal/log_std Mean 1.00828 +trainer/policy/normal/log_std Std 0.333242 +trainer/policy/normal/log_std Max 1.97824 +trainer/policy/normal/log_std Min -1.47988 +eval/num steps total 627085 +eval/num paths total 628 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00326327 +eval/Actions Std 0.90632 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.7211 +time/logging (s) 0.0037285 +time/sampling batch (s) 0.528059 +time/saving (s) 0.00407168 +time/training (s) 7.31503 +time/epoch (s) 10.572 +time/total (s) 6414.46 +Epoch -373 +---------------------------------- --------------- +2022-05-10 14:57:50.232349 PDT | [2] Epoch -372 finished +---------------------------------- --------------- +epoch -372 +replay_buffer/size 999033 +trainer/num train calls 629000 +trainer/Policy Loss -21.0396 +trainer/Log Pis Mean 23.8613 +trainer/Log Pis Std 13.0217 +trainer/Log Pis Max 75.6892 +trainer/Log Pis Min -8.7106 +trainer/policy/mean Mean -0.0203281 +trainer/policy/mean Std 0.911094 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.79422 +trainer/policy/normal/std Std 0.648121 +trainer/policy/normal/std Max 6.96307 +trainer/policy/normal/std Min 0.295846 +trainer/policy/normal/log_std Mean 0.98926 +trainer/policy/normal/log_std Std 0.311264 +trainer/policy/normal/log_std Max 1.94062 +trainer/policy/normal/log_std Min -1.21792 +eval/num steps total 628085 +eval/num paths total 629 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.126231 +eval/Actions Std 0.849025 +eval/Actions Max 0.999993 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64093 +time/logging (s) 0.0037134 +time/sampling batch (s) 0.278821 +time/saving (s) 0.00341882 +time/training (s) 7.89323 +time/epoch (s) 10.8201 +time/total (s) 6425.28 +Epoch -372 +---------------------------------- --------------- +2022-05-10 14:58:00.664756 PDT | [2] Epoch -371 finished +---------------------------------- --------------- +epoch -371 +replay_buffer/size 999033 +trainer/num train calls 630000 +trainer/Policy Loss -19.6261 +trainer/Log Pis Mean 24.1893 +trainer/Log Pis Std 13.4956 +trainer/Log Pis Max 70.5912 +trainer/Log Pis Min -10.4602 +trainer/policy/mean Mean -0.0452215 +trainer/policy/mean Std 0.910368 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.78338 +trainer/policy/normal/std Std 0.658784 +trainer/policy/normal/std Max 5.59866 +trainer/policy/normal/std Min 0.295931 +trainer/policy/normal/log_std Mean 0.981628 +trainer/policy/normal/log_std Std 0.3316 +trainer/policy/normal/log_std Max 1.72253 +trainer/policy/normal/log_std Min -1.21763 +eval/num steps total 629085 +eval/num paths total 630 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00201032 +eval/Actions Std 0.953434 +eval/Actions Max 0.999993 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.46304 +time/logging (s) 0.00369795 +time/sampling batch (s) 0.275807 +time/saving (s) 0.00344805 +time/training (s) 7.66354 +time/epoch (s) 10.4095 +time/total (s) 6435.7 +Epoch -371 +---------------------------------- --------------- +2022-05-10 14:58:11.077812 PDT | [2] Epoch -370 finished +---------------------------------- --------------- +epoch -370 +replay_buffer/size 999033 +trainer/num train calls 631000 +trainer/Policy Loss -19.5983 +trainer/Log Pis Mean 24.7597 +trainer/Log Pis Std 13.2451 +trainer/Log Pis Max 78.3868 +trainer/Log Pis Min -3.47614 +trainer/policy/mean Mean -0.0421218 +trainer/policy/mean Std 0.903753 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.78226 +trainer/policy/normal/std Std 0.661208 +trainer/policy/normal/std Max 5.51038 +trainer/policy/normal/std Min 0.269368 +trainer/policy/normal/log_std Mean 0.98263 +trainer/policy/normal/log_std Std 0.321939 +trainer/policy/normal/log_std Max 1.70663 +trainer/policy/normal/log_std Min -1.31168 +eval/num steps total 630085 +eval/num paths total 631 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0895164 +eval/Actions Std 0.922962 +eval/Actions Max 0.999999 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57901 +time/logging (s) 0.00367357 +time/sampling batch (s) 0.276116 +time/saving (s) 0.00343755 +time/training (s) 7.52787 +time/epoch (s) 10.3901 +time/total (s) 6446.09 +Epoch -370 +---------------------------------- --------------- +2022-05-10 14:58:21.895423 PDT | [2] Epoch -369 finished +---------------------------------- --------------- +epoch -369 +replay_buffer/size 999033 +trainer/num train calls 632000 +trainer/Policy Loss -18.8149 +trainer/Log Pis Mean 23.9125 +trainer/Log Pis Std 13.2042 +trainer/Log Pis Max 61.6522 +trainer/Log Pis Min -14.111 +trainer/policy/mean Mean -0.0409021 +trainer/policy/mean Std 0.909031 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.78647 +trainer/policy/normal/std Std 0.65513 +trainer/policy/normal/std Max 6.11484 +trainer/policy/normal/std Min 0.316099 +trainer/policy/normal/log_std Mean 0.986046 +trainer/policy/normal/log_std Std 0.31115 +trainer/policy/normal/log_std Max 1.81072 +trainer/policy/normal/log_std Min -1.1517 +eval/num steps total 631085 +eval/num paths total 632 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.110497 +eval/Actions Std 0.918134 +eval/Actions Max 0.999993 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55816 +time/logging (s) 0.00369754 +time/sampling batch (s) 0.276122 +time/saving (s) 0.00347938 +time/training (s) 7.95331 +time/epoch (s) 10.7948 +time/total (s) 6456.89 +Epoch -369 +---------------------------------- --------------- +2022-05-10 14:58:32.182935 PDT | [2] Epoch -368 finished +---------------------------------- --------------- +epoch -368 +replay_buffer/size 999033 +trainer/num train calls 633000 +trainer/Policy Loss -20.1291 +trainer/Log Pis Mean 23.8643 +trainer/Log Pis Std 12.7293 +trainer/Log Pis Max 60.2617 +trainer/Log Pis Min -10.8496 +trainer/policy/mean Mean -0.0470817 +trainer/policy/mean Std 0.908701 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.83284 +trainer/policy/normal/std Std 0.658333 +trainer/policy/normal/std Max 6.72867 +trainer/policy/normal/std Min 0.312526 +trainer/policy/normal/log_std Mean 1.0022 +trainer/policy/normal/log_std Std 0.315962 +trainer/policy/normal/log_std Max 1.90638 +trainer/policy/normal/log_std Min -1.16307 +eval/num steps total 632085 +eval/num paths total 633 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0155669 +eval/Actions Std 0.906062 +eval/Actions Max 0.999997 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60177 +time/logging (s) 0.00370591 +time/sampling batch (s) 0.27564 +time/saving (s) 0.00347194 +time/training (s) 7.37999 +time/epoch (s) 10.2646 +time/total (s) 6467.15 +Epoch -368 +---------------------------------- --------------- +2022-05-10 14:58:42.058858 PDT | [2] Epoch -367 finished +---------------------------------- --------------- +epoch -367 +replay_buffer/size 999033 +trainer/num train calls 634000 +trainer/Policy Loss -19.1434 +trainer/Log Pis Mean 23.2976 +trainer/Log Pis Std 13.5071 +trainer/Log Pis Max 68.167 +trainer/Log Pis Min -11.0134 +trainer/policy/mean Mean -0.0127687 +trainer/policy/mean Std 0.90392 +trainer/policy/mean Max 0.999977 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.79991 +trainer/policy/normal/std Std 0.662351 +trainer/policy/normal/std Max 5.57462 +trainer/policy/normal/std Min 0.280432 +trainer/policy/normal/log_std Mean 0.988762 +trainer/policy/normal/log_std Std 0.32354 +trainer/policy/normal/log_std Max 1.71822 +trainer/policy/normal/log_std Min -1.27143 +eval/num steps total 633085 +eval/num paths total 634 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.060195 +eval/Actions Std 0.91377 +eval/Actions Max 0.999994 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62999 +time/logging (s) 0.00372861 +time/sampling batch (s) 0.277573 +time/saving (s) 0.0034863 +time/training (s) 6.93824 +time/epoch (s) 9.85302 +time/total (s) 6477.01 +Epoch -367 +---------------------------------- --------------- +2022-05-10 14:58:52.971695 PDT | [2] Epoch -366 finished +---------------------------------- --------------- +epoch -366 +replay_buffer/size 999033 +trainer/num train calls 635000 +trainer/Policy Loss -20.5733 +trainer/Log Pis Mean 22.8475 +trainer/Log Pis Std 12.5125 +trainer/Log Pis Max 65.3744 +trainer/Log Pis Min -5.71222 +trainer/policy/mean Mean -0.0305052 +trainer/policy/mean Std 0.902855 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.75411 +trainer/policy/normal/std Std 0.663884 +trainer/policy/normal/std Max 6.27737 +trainer/policy/normal/std Min 0.322778 +trainer/policy/normal/log_std Mean 0.97021 +trainer/policy/normal/log_std Std 0.332998 +trainer/policy/normal/log_std Max 1.83695 +trainer/policy/normal/log_std Min -1.13079 +eval/num steps total 634085 +eval/num paths total 635 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0326162 +eval/Actions Std 0.900224 +eval/Actions Max 0.99999 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59153 +time/logging (s) 0.00377849 +time/sampling batch (s) 0.276143 +time/saving (s) 0.00341444 +time/training (s) 8.01458 +time/epoch (s) 10.8894 +time/total (s) 6487.9 +Epoch -366 +---------------------------------- --------------- +2022-05-10 14:59:03.580355 PDT | [2] Epoch -365 finished +---------------------------------- --------------- +epoch -365 +replay_buffer/size 999033 +trainer/num train calls 636000 +trainer/Policy Loss -19.8818 +trainer/Log Pis Mean 25.1239 +trainer/Log Pis Std 14.2881 +trainer/Log Pis Max 88.5637 +trainer/Log Pis Min -9.32091 +trainer/policy/mean Mean -0.0313909 +trainer/policy/mean Std 0.909292 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.86008 +trainer/policy/normal/std Std 0.653176 +trainer/policy/normal/std Max 6.97369 +trainer/policy/normal/std Min 0.33171 +trainer/policy/normal/log_std Mean 1.01575 +trainer/policy/normal/log_std Std 0.29451 +trainer/policy/normal/log_std Max 1.94214 +trainer/policy/normal/log_std Min -1.10349 +eval/num steps total 635085 +eval/num paths total 636 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0900111 +eval/Actions Std 0.900645 +eval/Actions Max 0.999995 +eval/Actions Min -0.999986 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68421 +time/logging (s) 0.00376888 +time/sampling batch (s) 0.527704 +time/saving (s) 0.00338153 +time/training (s) 7.3664 +time/epoch (s) 10.5855 +time/total (s) 6498.49 +Epoch -365 +---------------------------------- --------------- +2022-05-10 14:59:13.770574 PDT | [2] Epoch -364 finished +---------------------------------- --------------- +epoch -364 +replay_buffer/size 999033 +trainer/num train calls 637000 +trainer/Policy Loss -20.311 +trainer/Log Pis Mean 24.2023 +trainer/Log Pis Std 13.51 +trainer/Log Pis Max 68.111 +trainer/Log Pis Min -10.4931 +trainer/policy/mean Mean -0.0497598 +trainer/policy/mean Std 0.908007 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999981 +trainer/policy/normal/std Mean 2.81184 +trainer/policy/normal/std Std 0.654241 +trainer/policy/normal/std Max 6.88759 +trainer/policy/normal/std Min 0.328228 +trainer/policy/normal/log_std Mean 0.994322 +trainer/policy/normal/log_std Std 0.318223 +trainer/policy/normal/log_std Max 1.92972 +trainer/policy/normal/log_std Min -1.11405 +eval/num steps total 636085 +eval/num paths total 637 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0495817 +eval/Actions Std 0.915228 +eval/Actions Max 0.999993 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.605 +time/logging (s) 0.00367622 +time/sampling batch (s) 0.275762 +time/saving (s) 0.0034142 +time/training (s) 7.27928 +time/epoch (s) 10.1671 +time/total (s) 6508.66 +Epoch -364 +---------------------------------- --------------- +2022-05-10 14:59:24.560504 PDT | [2] Epoch -363 finished +---------------------------------- --------------- +epoch -363 +replay_buffer/size 999033 +trainer/num train calls 638000 +trainer/Policy Loss -18.6966 +trainer/Log Pis Mean 23.4917 +trainer/Log Pis Std 13.4928 +trainer/Log Pis Max 70.6725 +trainer/Log Pis Min -5.73638 +trainer/policy/mean Mean -0.0442791 +trainer/policy/mean Std 0.906764 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.75587 +trainer/policy/normal/std Std 0.648747 +trainer/policy/normal/std Max 5.58757 +trainer/policy/normal/std Min 0.287181 +trainer/policy/normal/log_std Mean 0.973275 +trainer/policy/normal/log_std Std 0.322833 +trainer/policy/normal/log_std Max 1.72054 +trainer/policy/normal/log_std Min -1.24764 +eval/num steps total 637085 +eval/num paths total 638 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0996362 +eval/Actions Std 0.912354 +eval/Actions Max 0.999995 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60383 +time/logging (s) 0.00420348 +time/sampling batch (s) 0.321719 +time/saving (s) 0.00424548 +time/training (s) 7.83217 +time/epoch (s) 10.7662 +time/total (s) 6519.43 +Epoch -363 +---------------------------------- --------------- +2022-05-10 14:59:35.793719 PDT | [2] Epoch -362 finished +---------------------------------- --------------- +epoch -362 +replay_buffer/size 999033 +trainer/num train calls 639000 +trainer/Policy Loss -20.1253 +trainer/Log Pis Mean 25.495 +trainer/Log Pis Std 13.8635 +trainer/Log Pis Max 63.9834 +trainer/Log Pis Min -10.0773 +trainer/policy/mean Mean -0.0153325 +trainer/policy/mean Std 0.908352 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.84253 +trainer/policy/normal/std Std 0.665339 +trainer/policy/normal/std Max 5.69416 +trainer/policy/normal/std Min 0.377863 +trainer/policy/normal/log_std Mean 1.00665 +trainer/policy/normal/log_std Std 0.308524 +trainer/policy/normal/log_std Max 1.73944 +trainer/policy/normal/log_std Min -0.973223 +eval/num steps total 638085 +eval/num paths total 639 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.211784 +eval/Actions Std 0.850642 +eval/Actions Max 0.999995 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.91482 +time/logging (s) 0.00411511 +time/sampling batch (s) 0.583047 +time/saving (s) 0.00404107 +time/training (s) 7.70262 +time/epoch (s) 11.2086 +time/total (s) 6530.64 +Epoch -362 +---------------------------------- --------------- +2022-05-10 14:59:47.053223 PDT | [2] Epoch -361 finished +---------------------------------- --------------- +epoch -361 +replay_buffer/size 999033 +trainer/num train calls 640000 +trainer/Policy Loss -20.1579 +trainer/Log Pis Mean 25.3966 +trainer/Log Pis Std 13.6753 +trainer/Log Pis Max 69.9239 +trainer/Log Pis Min -6.19823 +trainer/policy/mean Mean -0.024134 +trainer/policy/mean Std 0.904447 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.79461 +trainer/policy/normal/std Std 0.660793 +trainer/policy/normal/std Max 5.90574 +trainer/policy/normal/std Min 0.263337 +trainer/policy/normal/log_std Mean 0.987454 +trainer/policy/normal/log_std Std 0.320491 +trainer/policy/normal/log_std Max 1.77592 +trainer/policy/normal/log_std Min -1.33432 +eval/num steps total 639085 +eval/num paths total 640 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.19365 +eval/Actions Std 0.923812 +eval/Actions Max 0.999997 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.89906 +time/logging (s) 0.00386612 +time/sampling batch (s) 0.579885 +time/saving (s) 0.00373002 +time/training (s) 7.74838 +time/epoch (s) 11.2349 +time/total (s) 6541.88 +Epoch -361 +---------------------------------- --------------- +2022-05-10 14:59:58.636075 PDT | [2] Epoch -360 finished +---------------------------------- --------------- +epoch -360 +replay_buffer/size 999033 +trainer/num train calls 641000 +trainer/Policy Loss -19.7791 +trainer/Log Pis Mean 24.982 +trainer/Log Pis Std 13.1786 +trainer/Log Pis Max 62.0877 +trainer/Log Pis Min -7.16332 +trainer/policy/mean Mean -0.04848 +trainer/policy/mean Std 0.910794 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.78955 +trainer/policy/normal/std Std 0.670568 +trainer/policy/normal/std Max 5.81489 +trainer/policy/normal/std Min 0.278 +trainer/policy/normal/log_std Mean 0.983952 +trainer/policy/normal/log_std Std 0.328235 +trainer/policy/normal/log_std Max 1.76042 +trainer/policy/normal/log_std Min -1.28014 +eval/num steps total 640085 +eval/num paths total 641 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.173157 +eval/Actions Std 0.919667 +eval/Actions Max 0.999998 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.77118 +time/logging (s) 0.00379154 +time/sampling batch (s) 1.0785 +time/saving (s) 0.00365942 +time/training (s) 7.70169 +time/epoch (s) 11.5588 +time/total (s) 6553.44 +Epoch -360 +---------------------------------- --------------- +2022-05-10 15:00:09.586474 PDT | [2] Epoch -359 finished +---------------------------------- --------------- +epoch -359 +replay_buffer/size 999033 +trainer/num train calls 642000 +trainer/Policy Loss -19.031 +trainer/Log Pis Mean 24.7036 +trainer/Log Pis Std 13.0419 +trainer/Log Pis Max 75.7177 +trainer/Log Pis Min -5.68226 +trainer/policy/mean Mean -0.0273748 +trainer/policy/mean Std 0.909517 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.83494 +trainer/policy/normal/std Std 0.667901 +trainer/policy/normal/std Max 6.44059 +trainer/policy/normal/std Min 0.23551 +trainer/policy/normal/log_std Mean 1.00172 +trainer/policy/normal/log_std Std 0.322506 +trainer/policy/normal/log_std Max 1.86262 +trainer/policy/normal/log_std Min -1.446 +eval/num steps total 641085 +eval/num paths total 642 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.068843 +eval/Actions Std 0.912618 +eval/Actions Max 0.999993 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.74363 +time/logging (s) 0.0037023 +time/sampling batch (s) 0.280652 +time/saving (s) 0.00332893 +time/training (s) 7.89575 +time/epoch (s) 10.9271 +time/total (s) 6564.37 +Epoch -359 +---------------------------------- --------------- +2022-05-10 15:00:19.652694 PDT | [2] Epoch -358 finished +---------------------------------- --------------- +epoch -358 +replay_buffer/size 999033 +trainer/num train calls 643000 +trainer/Policy Loss -18.0662 +trainer/Log Pis Mean 24.2918 +trainer/Log Pis Std 12.936 +trainer/Log Pis Max 66.7063 +trainer/Log Pis Min -9.22586 +trainer/policy/mean Mean -0.0301099 +trainer/policy/mean Std 0.905912 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.79735 +trainer/policy/normal/std Std 0.666909 +trainer/policy/normal/std Max 6.3895 +trainer/policy/normal/std Min 0.235759 +trainer/policy/normal/log_std Mean 0.987124 +trainer/policy/normal/log_std Std 0.328273 +trainer/policy/normal/log_std Max 1.85466 +trainer/policy/normal/log_std Min -1.44494 +eval/num steps total 642085 +eval/num paths total 643 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0615847 +eval/Actions Std 0.916663 +eval/Actions Max 0.999995 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.7182 +time/logging (s) 0.00376858 +time/sampling batch (s) 0.284611 +time/saving (s) 0.00362603 +time/training (s) 7.03291 +time/epoch (s) 10.0431 +time/total (s) 6574.42 +Epoch -358 +---------------------------------- --------------- +2022-05-10 15:00:30.941285 PDT | [2] Epoch -357 finished +---------------------------------- --------------- +epoch -357 +replay_buffer/size 999033 +trainer/num train calls 644000 +trainer/Policy Loss -19.5806 +trainer/Log Pis Mean 25.6844 +trainer/Log Pis Std 13.3452 +trainer/Log Pis Max 64.1207 +trainer/Log Pis Min -10.7237 +trainer/policy/mean Mean -0.0390397 +trainer/policy/mean Std 0.906969 +trainer/policy/mean Max 0.999981 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.82049 +trainer/policy/normal/std Std 0.667064 +trainer/policy/normal/std Max 5.55341 +trainer/policy/normal/std Min 0.307434 +trainer/policy/normal/log_std Mean 0.995785 +trainer/policy/normal/log_std Std 0.325947 +trainer/policy/normal/log_std Max 1.71441 +trainer/policy/normal/log_std Min -1.1795 +eval/num steps total 643085 +eval/num paths total 644 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0904282 +eval/Actions Std 0.919484 +eval/Actions Max 0.999987 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.89529 +time/logging (s) 0.00377086 +time/sampling batch (s) 0.57608 +time/saving (s) 0.00361241 +time/training (s) 7.78597 +time/epoch (s) 11.2647 +time/total (s) 6585.69 +Epoch -357 +---------------------------------- --------------- +2022-05-10 15:00:41.986647 PDT | [2] Epoch -356 finished +---------------------------------- --------------- +epoch -356 +replay_buffer/size 999033 +trainer/num train calls 645000 +trainer/Policy Loss -19.3602 +trainer/Log Pis Mean 24.5686 +trainer/Log Pis Std 13.0412 +trainer/Log Pis Max 62.9645 +trainer/Log Pis Min -6.42073 +trainer/policy/mean Mean -0.0265117 +trainer/policy/mean Std 0.905187 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.78989 +trainer/policy/normal/std Std 0.64601 +trainer/policy/normal/std Max 5.52366 +trainer/policy/normal/std Min 0.290791 +trainer/policy/normal/log_std Mean 0.986264 +trainer/policy/normal/log_std Std 0.321294 +trainer/policy/normal/log_std Max 1.70904 +trainer/policy/normal/log_std Min -1.23515 +eval/num steps total 644085 +eval/num paths total 645 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.301212 +eval/Actions Std 0.868441 +eval/Actions Max 0.999989 +eval/Actions Min -0.999981 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67721 +time/logging (s) 0.00378939 +time/sampling batch (s) 0.57791 +time/saving (s) 0.0036472 +time/training (s) 7.75909 +time/epoch (s) 11.0217 +time/total (s) 6596.71 +Epoch -356 +---------------------------------- --------------- +2022-05-10 15:00:52.592061 PDT | [2] Epoch -355 finished +---------------------------------- --------------- +epoch -355 +replay_buffer/size 999033 +trainer/num train calls 646000 +trainer/Policy Loss -19.9378 +trainer/Log Pis Mean 23.4364 +trainer/Log Pis Std 13.0363 +trainer/Log Pis Max 74.5316 +trainer/Log Pis Min -7.36757 +trainer/policy/mean Mean -0.042501 +trainer/policy/mean Std 0.906192 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.7883 +trainer/policy/normal/std Std 0.690019 +trainer/policy/normal/std Max 6.15573 +trainer/policy/normal/std Min 0.254776 +trainer/policy/normal/log_std Mean 0.97776 +trainer/policy/normal/log_std Std 0.357364 +trainer/policy/normal/log_std Max 1.81738 +trainer/policy/normal/log_std Min -1.36737 +eval/num steps total 645085 +eval/num paths total 646 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.250122 +eval/Actions Std 0.876476 +eval/Actions Max 0.999998 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66484 +time/logging (s) 0.00391601 +time/sampling batch (s) 0.575323 +time/saving (s) 0.0035794 +time/training (s) 7.33423 +time/epoch (s) 10.5819 +time/total (s) 6607.3 +Epoch -355 +---------------------------------- --------------- +2022-05-10 15:01:02.902557 PDT | [2] Epoch -354 finished +---------------------------------- --------------- +epoch -354 +replay_buffer/size 999033 +trainer/num train calls 647000 +trainer/Policy Loss -18.7891 +trainer/Log Pis Mean 24.2168 +trainer/Log Pis Std 14.0579 +trainer/Log Pis Max 68.4306 +trainer/Log Pis Min -11.8488 +trainer/policy/mean Mean -0.051408 +trainer/policy/mean Std 0.906473 +trainer/policy/mean Max 0.999978 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.80247 +trainer/policy/normal/std Std 0.654379 +trainer/policy/normal/std Max 5.74913 +trainer/policy/normal/std Min 0.282942 +trainer/policy/normal/log_std Mean 0.99081 +trainer/policy/normal/log_std Std 0.319362 +trainer/policy/normal/log_std Max 1.74905 +trainer/policy/normal/log_std Min -1.26251 +eval/num steps total 646085 +eval/num paths total 647 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.17941 +eval/Actions Std 0.90894 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64201 +time/logging (s) 0.00416476 +time/sampling batch (s) 0.576465 +time/saving (s) 0.00403251 +time/training (s) 7.0581 +time/epoch (s) 10.2848 +time/total (s) 6617.58 +Epoch -354 +---------------------------------- --------------- +2022-05-10 15:01:14.117894 PDT | [2] Epoch -353 finished +---------------------------------- --------------- +epoch -353 +replay_buffer/size 999033 +trainer/num train calls 648000 +trainer/Policy Loss -19.512 +trainer/Log Pis Mean 23.9409 +trainer/Log Pis Std 12.9774 +trainer/Log Pis Max 69.8139 +trainer/Log Pis Min -9.17021 +trainer/policy/mean Mean -0.0290699 +trainer/policy/mean Std 0.90695 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.75428 +trainer/policy/normal/std Std 0.656545 +trainer/policy/normal/std Max 5.56364 +trainer/policy/normal/std Min 0.233028 +trainer/policy/normal/log_std Mean 0.970145 +trainer/policy/normal/log_std Std 0.336667 +trainer/policy/normal/log_std Max 1.71625 +trainer/policy/normal/log_std Min -1.4566 +eval/num steps total 647085 +eval/num paths total 648 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.25763 +eval/Actions Std 0.768335 +eval/Actions Max 0.999993 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.85956 +time/logging (s) 0.0037291 +time/sampling batch (s) 0.790847 +time/saving (s) 0.00351934 +time/training (s) 7.53325 +time/epoch (s) 11.1909 +time/total (s) 6628.78 +Epoch -353 +---------------------------------- --------------- +2022-05-10 15:01:24.578507 PDT | [2] Epoch -352 finished +---------------------------------- --------------- +epoch -352 +replay_buffer/size 999033 +trainer/num train calls 649000 +trainer/Policy Loss -20.5622 +trainer/Log Pis Mean 24.3854 +trainer/Log Pis Std 12.9858 +trainer/Log Pis Max 67.1933 +trainer/Log Pis Min -13.0003 +trainer/policy/mean Mean -0.0349119 +trainer/policy/mean Std 0.908416 +trainer/policy/mean Max 0.999978 +trainer/policy/mean Min -0.999984 +trainer/policy/normal/std Mean 2.8762 +trainer/policy/normal/std Std 0.664423 +trainer/policy/normal/std Max 5.34268 +trainer/policy/normal/std Min 0.261462 +trainer/policy/normal/log_std Mean 1.01666 +trainer/policy/normal/log_std Std 0.321307 +trainer/policy/normal/log_std Max 1.67573 +trainer/policy/normal/log_std Min -1.34147 +eval/num steps total 648085 +eval/num paths total 649 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.113951 +eval/Actions Std 0.919344 +eval/Actions Max 0.999981 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63951 +time/logging (s) 0.00375234 +time/sampling batch (s) 0.776594 +time/saving (s) 0.00341585 +time/training (s) 7.01432 +time/epoch (s) 10.4376 +time/total (s) 6639.22 +Epoch -352 +---------------------------------- --------------- +2022-05-10 15:01:35.349382 PDT | [2] Epoch -351 finished +---------------------------------- --------------- +epoch -351 +replay_buffer/size 999033 +trainer/num train calls 650000 +trainer/Policy Loss -19.4347 +trainer/Log Pis Mean 23.9885 +trainer/Log Pis Std 12.6381 +trainer/Log Pis Max 63.2661 +trainer/Log Pis Min -8.36584 +trainer/policy/mean Mean -0.0217649 +trainer/policy/mean Std 0.908354 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.80091 +trainer/policy/normal/std Std 0.637001 +trainer/policy/normal/std Max 5.12217 +trainer/policy/normal/std Min 0.246957 +trainer/policy/normal/log_std Mean 0.992747 +trainer/policy/normal/log_std Std 0.308907 +trainer/policy/normal/log_std Max 1.63358 +trainer/policy/normal/log_std Min -1.39854 +eval/num steps total 649085 +eval/num paths total 650 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.189407 +eval/Actions Std 0.879038 +eval/Actions Max 0.999994 +eval/Actions Min -0.999988 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57856 +time/logging (s) 0.0036779 +time/sampling batch (s) 0.276262 +time/saving (s) 0.00342081 +time/training (s) 7.88563 +time/epoch (s) 10.7475 +time/total (s) 6649.97 +Epoch -351 +---------------------------------- --------------- +2022-05-10 15:01:45.134290 PDT | [2] Epoch -350 finished +---------------------------------- --------------- +epoch -350 +replay_buffer/size 999033 +trainer/num train calls 651000 +trainer/Policy Loss -20.2167 +trainer/Log Pis Mean 23.6095 +trainer/Log Pis Std 13.42 +trainer/Log Pis Max 74.3186 +trainer/Log Pis Min -10.2665 +trainer/policy/mean Mean -0.0479711 +trainer/policy/mean Std 0.90729 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.78971 +trainer/policy/normal/std Std 0.625693 +trainer/policy/normal/std Max 6.65574 +trainer/policy/normal/std Min 0.301075 +trainer/policy/normal/log_std Mean 0.991139 +trainer/policy/normal/log_std Std 0.29402 +trainer/policy/normal/log_std Max 1.89548 +trainer/policy/normal/log_std Min -1.2004 +eval/num steps total 650085 +eval/num paths total 651 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.108387 +eval/Actions Std 0.911073 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57751 +time/logging (s) 0.00370034 +time/sampling batch (s) 0.274589 +time/saving (s) 0.0033744 +time/training (s) 6.90292 +time/epoch (s) 9.76209 +time/total (s) 6659.74 +Epoch -350 +---------------------------------- --------------- +2022-05-10 15:01:55.806607 PDT | [2] Epoch -349 finished +---------------------------------- --------------- +epoch -349 +replay_buffer/size 999033 +trainer/num train calls 652000 +trainer/Policy Loss -20.3937 +trainer/Log Pis Mean 25.025 +trainer/Log Pis Std 14.3982 +trainer/Log Pis Max 80.529 +trainer/Log Pis Min -6.47789 +trainer/policy/mean Mean -0.0101798 +trainer/policy/mean Std 0.905975 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.81227 +trainer/policy/normal/std Std 0.666241 +trainer/policy/normal/std Max 7.03598 +trainer/policy/normal/std Min 0.262805 +trainer/policy/normal/log_std Mean 0.992794 +trainer/policy/normal/log_std Std 0.326862 +trainer/policy/normal/log_std Max 1.95104 +trainer/policy/normal/log_std Min -1.33634 +eval/num steps total 651085 +eval/num paths total 652 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00267602 +eval/Actions Std 0.909944 +eval/Actions Max 0.999992 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59911 +time/logging (s) 0.00381114 +time/sampling batch (s) 0.273694 +time/saving (s) 0.0034457 +time/training (s) 7.76952 +time/epoch (s) 10.6496 +time/total (s) 6670.39 +Epoch -349 +---------------------------------- --------------- +2022-05-10 15:02:06.414473 PDT | [2] Epoch -348 finished +---------------------------------- --------------- +epoch -348 +replay_buffer/size 999033 +trainer/num train calls 653000 +trainer/Policy Loss -19.6354 +trainer/Log Pis Mean 25.3267 +trainer/Log Pis Std 12.9724 +trainer/Log Pis Max 63.1485 +trainer/Log Pis Min -5.85561 +trainer/policy/mean Mean -0.0349946 +trainer/policy/mean Std 0.907198 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.83089 +trainer/policy/normal/std Std 0.652075 +trainer/policy/normal/std Max 5.65243 +trainer/policy/normal/std Min 0.312279 +trainer/policy/normal/log_std Mean 1.00344 +trainer/policy/normal/log_std Std 0.304771 +trainer/policy/normal/log_std Max 1.73209 +trainer/policy/normal/log_std Min -1.16386 +eval/num steps total 652085 +eval/num paths total 653 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0709581 +eval/Actions Std 0.831215 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.49241 +time/logging (s) 0.00370549 +time/sampling batch (s) 0.276466 +time/saving (s) 0.00332207 +time/training (s) 7.80872 +time/epoch (s) 10.5846 +time/total (s) 6680.98 +Epoch -348 +---------------------------------- --------------- +2022-05-10 15:02:17.237247 PDT | [2] Epoch -347 finished +---------------------------------- --------------- +epoch -347 +replay_buffer/size 999033 +trainer/num train calls 654000 +trainer/Policy Loss -19.6513 +trainer/Log Pis Mean 24.2551 +trainer/Log Pis Std 13.5666 +trainer/Log Pis Max 75.8002 +trainer/Log Pis Min -5.35925 +trainer/policy/mean Mean -0.0402155 +trainer/policy/mean Std 0.907926 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999982 +trainer/policy/normal/std Mean 2.75167 +trainer/policy/normal/std Std 0.645584 +trainer/policy/normal/std Max 6.19116 +trainer/policy/normal/std Min 0.340882 +trainer/policy/normal/log_std Mean 0.972964 +trainer/policy/normal/log_std Std 0.3156 +trainer/policy/normal/log_std Max 1.82312 +trainer/policy/normal/log_std Min -1.07622 +eval/num steps total 653085 +eval/num paths total 654 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.110718 +eval/Actions Std 0.76156 +eval/Actions Max 1 +eval/Actions Min -0.999969 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53274 +time/logging (s) 0.00408199 +time/sampling batch (s) 0.274869 +time/saving (s) 0.00388301 +time/training (s) 7.98465 +time/epoch (s) 10.8002 +time/total (s) 6691.78 +Epoch -347 +---------------------------------- --------------- +2022-05-10 15:02:27.682794 PDT | [2] Epoch -346 finished +---------------------------------- --------------- +epoch -346 +replay_buffer/size 999033 +trainer/num train calls 655000 +trainer/Policy Loss -18.4697 +trainer/Log Pis Mean 23.4159 +trainer/Log Pis Std 13.0416 +trainer/Log Pis Max 72.8251 +trainer/Log Pis Min -13.512 +trainer/policy/mean Mean -0.0403843 +trainer/policy/mean Std 0.906936 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.78768 +trainer/policy/normal/std Std 0.638556 +trainer/policy/normal/std Max 5.60511 +trainer/policy/normal/std Min 0.258057 +trainer/policy/normal/log_std Mean 0.987431 +trainer/policy/normal/log_std Std 0.310035 +trainer/policy/normal/log_std Max 1.72368 +trainer/policy/normal/log_std Min -1.35457 +eval/num steps total 654085 +eval/num paths total 655 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0881955 +eval/Actions Std 0.913079 +eval/Actions Max 0.999997 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51913 +time/logging (s) 0.00438269 +time/sampling batch (s) 0.525617 +time/saving (s) 0.00358963 +time/training (s) 7.36994 +time/epoch (s) 10.4227 +time/total (s) 6702.2 +Epoch -346 +---------------------------------- --------------- +2022-05-10 15:02:38.086910 PDT | [2] Epoch -345 finished +---------------------------------- --------------- +epoch -345 +replay_buffer/size 999033 +trainer/num train calls 656000 +trainer/Policy Loss -19.3105 +trainer/Log Pis Mean 24.4115 +trainer/Log Pis Std 13.5931 +trainer/Log Pis Max 65.0126 +trainer/Log Pis Min -13.0843 +trainer/policy/mean Mean -0.0450767 +trainer/policy/mean Std 0.906633 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.78245 +trainer/policy/normal/std Std 0.670348 +trainer/policy/normal/std Max 5.92685 +trainer/policy/normal/std Min 0.282594 +trainer/policy/normal/log_std Mean 0.98025 +trainer/policy/normal/log_std Std 0.333557 +trainer/policy/normal/log_std Max 1.77949 +trainer/policy/normal/log_std Min -1.26374 +eval/num steps total 655085 +eval/num paths total 656 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0340335 +eval/Actions Std 0.907899 +eval/Actions Max 0.999994 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5884 +time/logging (s) 0.00371598 +time/sampling batch (s) 0.526382 +time/saving (s) 0.00335976 +time/training (s) 7.25858 +time/epoch (s) 10.3804 +time/total (s) 6712.59 +Epoch -345 +---------------------------------- --------------- +2022-05-10 15:02:48.281881 PDT | [2] Epoch -344 finished +---------------------------------- --------------- +epoch -344 +replay_buffer/size 999033 +trainer/num train calls 657000 +trainer/Policy Loss -21.0258 +trainer/Log Pis Mean 24.497 +trainer/Log Pis Std 13.2377 +trainer/Log Pis Max 71.29 +trainer/Log Pis Min -5.84042 +trainer/policy/mean Mean -0.0333977 +trainer/policy/mean Std 0.901251 +trainer/policy/mean Max 0.999977 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.86998 +trainer/policy/normal/std Std 0.722121 +trainer/policy/normal/std Max 6.07208 +trainer/policy/normal/std Min 0.278195 +trainer/policy/normal/log_std Mean 1.00803 +trainer/policy/normal/log_std Std 0.345547 +trainer/policy/normal/log_std Max 1.8037 +trainer/policy/normal/log_std Min -1.27943 +eval/num steps total 656085 +eval/num paths total 657 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.200055 +eval/Actions Std 0.85875 +eval/Actions Max 0.999998 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53295 +time/logging (s) 0.00375833 +time/sampling batch (s) 0.275965 +time/saving (s) 0.00341646 +time/training (s) 7.35576 +time/epoch (s) 10.1718 +time/total (s) 6722.76 +Epoch -344 +---------------------------------- --------------- +2022-05-10 15:02:58.494149 PDT | [2] Epoch -343 finished +---------------------------------- --------------- +epoch -343 +replay_buffer/size 999033 +trainer/num train calls 658000 +trainer/Policy Loss -20.5327 +trainer/Log Pis Mean 24.5814 +trainer/Log Pis Std 13.9042 +trainer/Log Pis Max 69.9206 +trainer/Log Pis Min -12.256 +trainer/policy/mean Mean -0.0212929 +trainer/policy/mean Std 0.90927 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.78253 +trainer/policy/normal/std Std 0.64947 +trainer/policy/normal/std Max 5.60662 +trainer/policy/normal/std Min 0.378138 +trainer/policy/normal/log_std Mean 0.98429 +trainer/policy/normal/log_std Std 0.314784 +trainer/policy/normal/log_std Max 1.72395 +trainer/policy/normal/log_std Min -0.972497 +eval/num steps total 657085 +eval/num paths total 658 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0925615 +eval/Actions Std 0.895733 +eval/Actions Max 0.999995 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.46713 +time/logging (s) 0.00374566 +time/sampling batch (s) 0.276204 +time/saving (s) 0.00343971 +time/training (s) 7.4386 +time/epoch (s) 10.1891 +time/total (s) 6732.95 +Epoch -343 +---------------------------------- --------------- +2022-05-10 15:03:09.284409 PDT | [2] Epoch -342 finished +---------------------------------- --------------- +epoch -342 +replay_buffer/size 999033 +trainer/num train calls 659000 +trainer/Policy Loss -19.5001 +trainer/Log Pis Mean 25.2151 +trainer/Log Pis Std 12.9247 +trainer/Log Pis Max 59.7446 +trainer/Log Pis Min -9.08871 +trainer/policy/mean Mean -0.0254861 +trainer/policy/mean Std 0.908565 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.82325 +trainer/policy/normal/std Std 0.677601 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.299972 +trainer/policy/normal/log_std Mean 0.996199 +trainer/policy/normal/log_std Std 0.327646 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.20407 +eval/num steps total 658085 +eval/num paths total 659 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0105872 +eval/Actions Std 0.951242 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68934 +time/logging (s) 0.00369934 +time/sampling batch (s) 0.528576 +time/saving (s) 0.00340914 +time/training (s) 7.54182 +time/epoch (s) 10.7668 +time/total (s) 6743.72 +Epoch -342 +---------------------------------- --------------- +2022-05-10 15:03:20.062470 PDT | [2] Epoch -341 finished +---------------------------------- --------------- +epoch -341 +replay_buffer/size 999033 +trainer/num train calls 660000 +trainer/Policy Loss -20.469 +trainer/Log Pis Mean 23.9946 +trainer/Log Pis Std 13.6704 +trainer/Log Pis Max 69.2736 +trainer/Log Pis Min -8.53799 +trainer/policy/mean Mean -0.0470888 +trainer/policy/mean Std 0.90629 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.73751 +trainer/policy/normal/std Std 0.631601 +trainer/policy/normal/std Max 5.75088 +trainer/policy/normal/std Min 0.289618 +trainer/policy/normal/log_std Mean 0.967499 +trainer/policy/normal/log_std Std 0.319876 +trainer/policy/normal/log_std Max 1.74935 +trainer/policy/normal/log_std Min -1.23919 +eval/num steps total 659085 +eval/num paths total 660 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0457869 +eval/Actions Std 0.916384 +eval/Actions Max 0.999995 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.3827 +time/logging (s) 0.00400762 +time/sampling batch (s) 0.282 +time/saving (s) 0.003698 +time/training (s) 8.08255 +time/epoch (s) 10.755 +time/total (s) 6754.48 +Epoch -341 +---------------------------------- --------------- +2022-05-10 15:03:30.368816 PDT | [2] Epoch -340 finished +---------------------------------- --------------- +epoch -340 +replay_buffer/size 999033 +trainer/num train calls 661000 +trainer/Policy Loss -20.0558 +trainer/Log Pis Mean 24.7738 +trainer/Log Pis Std 13.4688 +trainer/Log Pis Max 75.3017 +trainer/Log Pis Min -7.57532 +trainer/policy/mean Mean -0.0286254 +trainer/policy/mean Std 0.903686 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.79042 +trainer/policy/normal/std Std 0.699561 +trainer/policy/normal/std Max 5.81901 +trainer/policy/normal/std Min 0.241888 +trainer/policy/normal/log_std Mean 0.978951 +trainer/policy/normal/log_std Std 0.351525 +trainer/policy/normal/log_std Max 1.76113 +trainer/policy/normal/log_std Min -1.41928 +eval/num steps total 660085 +eval/num paths total 661 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0299923 +eval/Actions Std 0.908977 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66095 +time/logging (s) 0.00372233 +time/sampling batch (s) 0.279489 +time/saving (s) 0.00351052 +time/training (s) 7.33494 +time/epoch (s) 10.2826 +time/total (s) 6764.77 +Epoch -340 +---------------------------------- --------------- +2022-05-10 15:03:41.241560 PDT | [2] Epoch -339 finished +---------------------------------- --------------- +epoch -339 +replay_buffer/size 999033 +trainer/num train calls 662000 +trainer/Policy Loss -20.6433 +trainer/Log Pis Mean 25.2383 +trainer/Log Pis Std 13.5266 +trainer/Log Pis Max 80.2978 +trainer/Log Pis Min -6.01921 +trainer/policy/mean Mean -0.0252708 +trainer/policy/mean Std 0.906835 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.8107 +trainer/policy/normal/std Std 0.670123 +trainer/policy/normal/std Max 5.28256 +trainer/policy/normal/std Min 0.297884 +trainer/policy/normal/log_std Mean 0.990545 +trainer/policy/normal/log_std Std 0.333691 +trainer/policy/normal/log_std Max 1.66441 +trainer/policy/normal/log_std Min -1.21105 +eval/num steps total 661085 +eval/num paths total 662 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.130954 +eval/Actions Std 0.891407 +eval/Actions Max 0.999995 +eval/Actions Min -0.999982 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62886 +time/logging (s) 0.00368289 +time/sampling batch (s) 0.521871 +time/saving (s) 0.00335122 +time/training (s) 7.69235 +time/epoch (s) 10.8501 +time/total (s) 6775.62 +Epoch -339 +---------------------------------- --------------- +2022-05-10 15:03:51.577521 PDT | [2] Epoch -338 finished +---------------------------------- --------------- +epoch -338 +replay_buffer/size 999033 +trainer/num train calls 663000 +trainer/Policy Loss -19.0786 +trainer/Log Pis Mean 24.168 +trainer/Log Pis Std 12.3701 +trainer/Log Pis Max 69.3022 +trainer/Log Pis Min -10.1149 +trainer/policy/mean Mean -0.0179167 +trainer/policy/mean Std 0.907514 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.77096 +trainer/policy/normal/std Std 0.650831 +trainer/policy/normal/std Max 5.89253 +trainer/policy/normal/std Min 0.309211 +trainer/policy/normal/log_std Mean 0.978149 +trainer/policy/normal/log_std Std 0.326604 +trainer/policy/normal/log_std Max 1.77369 +trainer/policy/normal/log_std Min -1.17373 +eval/num steps total 662085 +eval/num paths total 663 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.109592 +eval/Actions Std 0.904283 +eval/Actions Max 0.99999 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.46692 +time/logging (s) 0.00369251 +time/sampling batch (s) 0.523056 +time/saving (s) 0.00336187 +time/training (s) 7.31624 +time/epoch (s) 10.3133 +time/total (s) 6785.94 +Epoch -338 +---------------------------------- --------------- +2022-05-10 15:04:01.627830 PDT | [2] Epoch -337 finished +---------------------------------- --------------- +epoch -337 +replay_buffer/size 999033 +trainer/num train calls 664000 +trainer/Policy Loss -19.2724 +trainer/Log Pis Mean 25.4068 +trainer/Log Pis Std 13.757 +trainer/Log Pis Max 71.9707 +trainer/Log Pis Min -6.05794 +trainer/policy/mean Mean -0.029077 +trainer/policy/mean Std 0.904989 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.78744 +trainer/policy/normal/std Std 0.684595 +trainer/policy/normal/std Max 7.14798 +trainer/policy/normal/std Min 0.213003 +trainer/policy/normal/log_std Mean 0.979487 +trainer/policy/normal/log_std Std 0.347313 +trainer/policy/normal/log_std Max 1.96683 +trainer/policy/normal/log_std Min -1.54645 +eval/num steps total 663085 +eval/num paths total 664 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.196265 +eval/Actions Std 0.943201 +eval/Actions Max 0.999995 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5103 +time/logging (s) 0.00372448 +time/sampling batch (s) 0.523941 +time/saving (s) 0.00339359 +time/training (s) 6.98613 +time/epoch (s) 10.0275 +time/total (s) 6795.97 +Epoch -337 +---------------------------------- --------------- +2022-05-10 15:04:11.612372 PDT | [2] Epoch -336 finished +---------------------------------- --------------- +epoch -336 +replay_buffer/size 999033 +trainer/num train calls 665000 +trainer/Policy Loss -20.4886 +trainer/Log Pis Mean 25.1576 +trainer/Log Pis Std 13.1666 +trainer/Log Pis Max 71.9139 +trainer/Log Pis Min -5.75122 +trainer/policy/mean Mean -0.0338414 +trainer/policy/mean Std 0.907526 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.88327 +trainer/policy/normal/std Std 0.687229 +trainer/policy/normal/std Max 6.89973 +trainer/policy/normal/std Min 0.308627 +trainer/policy/normal/log_std Mean 1.0178 +trainer/policy/normal/log_std Std 0.324393 +trainer/policy/normal/log_std Max 1.93148 +trainer/policy/normal/log_std Min -1.17562 +eval/num steps total 664085 +eval/num paths total 665 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.193311 +eval/Actions Std 0.798088 +eval/Actions Max 0.999998 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.37726 +time/logging (s) 0.00417129 +time/sampling batch (s) 0.525029 +time/saving (s) 0.00375116 +time/training (s) 7.05165 +time/epoch (s) 9.96186 +time/total (s) 6805.93 +Epoch -336 +---------------------------------- --------------- +2022-05-10 15:04:21.851951 PDT | [2] Epoch -335 finished +---------------------------------- --------------- +epoch -335 +replay_buffer/size 999033 +trainer/num train calls 666000 +trainer/Policy Loss -19.7664 +trainer/Log Pis Mean 23.8907 +trainer/Log Pis Std 12.8123 +trainer/Log Pis Max 70.5101 +trainer/Log Pis Min -8.17402 +trainer/policy/mean Mean -0.0294531 +trainer/policy/mean Std 0.905126 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.79953 +trainer/policy/normal/std Std 0.66292 +trainer/policy/normal/std Max 5.64036 +trainer/policy/normal/std Min 0.299735 +trainer/policy/normal/log_std Mean 0.990587 +trainer/policy/normal/log_std Std 0.311318 +trainer/policy/normal/log_std Max 1.72995 +trainer/policy/normal/log_std Min -1.20486 +eval/num steps total 665085 +eval/num paths total 666 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.129347 +eval/Actions Std 0.893522 +eval/Actions Max 0.999987 +eval/Actions Min -0.999983 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55622 +time/logging (s) 0.00372667 +time/sampling batch (s) 0.277689 +time/saving (s) 0.00358829 +time/training (s) 7.37411 +time/epoch (s) 10.2153 +time/total (s) 6816.15 +Epoch -335 +---------------------------------- --------------- +2022-05-10 15:04:31.396208 PDT | [2] Epoch -334 finished +---------------------------------- --------------- +epoch -334 +replay_buffer/size 999033 +trainer/num train calls 667000 +trainer/Policy Loss -19.5794 +trainer/Log Pis Mean 24.9993 +trainer/Log Pis Std 13.527 +trainer/Log Pis Max 68.753 +trainer/Log Pis Min -11.0729 +trainer/policy/mean Mean -0.0558427 +trainer/policy/mean Std 0.908801 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.87177 +trainer/policy/normal/std Std 0.667163 +trainer/policy/normal/std Max 5.78758 +trainer/policy/normal/std Min 0.286968 +trainer/policy/normal/log_std Mean 1.01633 +trainer/policy/normal/log_std Std 0.313349 +trainer/policy/normal/log_std Max 1.75571 +trainer/policy/normal/log_std Min -1.24838 +eval/num steps total 666085 +eval/num paths total 667 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0679056 +eval/Actions Std 0.906093 +eval/Actions Max 0.999989 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65439 +time/logging (s) 0.00375327 +time/sampling batch (s) 0.285835 +time/saving (s) 0.00347892 +time/training (s) 6.57299 +time/epoch (s) 9.52045 +time/total (s) 6825.67 +Epoch -334 +---------------------------------- --------------- +2022-05-10 15:04:40.807503 PDT | [2] Epoch -333 finished +---------------------------------- --------------- +epoch -333 +replay_buffer/size 999033 +trainer/num train calls 668000 +trainer/Policy Loss -20.0059 +trainer/Log Pis Mean 24.8531 +trainer/Log Pis Std 12.9227 +trainer/Log Pis Max 64.7998 +trainer/Log Pis Min -5.19224 +trainer/policy/mean Mean -0.0436987 +trainer/policy/mean Std 0.903478 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.80118 +trainer/policy/normal/std Std 0.658436 +trainer/policy/normal/std Max 5.81502 +trainer/policy/normal/std Min 0.3413 +trainer/policy/normal/log_std Mean 0.990408 +trainer/policy/normal/log_std Std 0.316821 +trainer/policy/normal/log_std Max 1.76044 +trainer/policy/normal/log_std Min -1.07499 +eval/num steps total 667085 +eval/num paths total 668 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00436295 +eval/Actions Std 0.904809 +eval/Actions Max 0.999999 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66748 +time/logging (s) 0.00371533 +time/sampling batch (s) 0.291869 +time/saving (s) 0.00334818 +time/training (s) 6.42092 +time/epoch (s) 9.38733 +time/total (s) 6835.06 +Epoch -333 +---------------------------------- --------------- +2022-05-10 15:04:51.414729 PDT | [2] Epoch -332 finished +---------------------------------- --------------- +epoch -332 +replay_buffer/size 999033 +trainer/num train calls 669000 +trainer/Policy Loss -18.9829 +trainer/Log Pis Mean 24.4421 +trainer/Log Pis Std 13.954 +trainer/Log Pis Max 74.1834 +trainer/Log Pis Min -10.7541 +trainer/policy/mean Mean -0.0471729 +trainer/policy/mean Std 0.901107 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.80614 +trainer/policy/normal/std Std 0.666582 +trainer/policy/normal/std Max 6.85926 +trainer/policy/normal/std Min 0.213276 +trainer/policy/normal/log_std Mean 0.990418 +trainer/policy/normal/log_std Std 0.32729 +trainer/policy/normal/log_std Max 1.9256 +trainer/policy/normal/log_std Min -1.54517 +eval/num steps total 668085 +eval/num paths total 669 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.18993 +eval/Actions Std 0.884229 +eval/Actions Max 0.999994 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.83668 +time/logging (s) 0.00387802 +time/sampling batch (s) 0.52583 +time/saving (s) 0.00335251 +time/training (s) 7.21476 +time/epoch (s) 10.5845 +time/total (s) 6845.65 +Epoch -332 +---------------------------------- --------------- +2022-05-10 15:05:01.175157 PDT | [2] Epoch -331 finished +---------------------------------- --------------- +epoch -331 +replay_buffer/size 999033 +trainer/num train calls 670000 +trainer/Policy Loss -19.9423 +trainer/Log Pis Mean 23.9711 +trainer/Log Pis Std 13.0918 +trainer/Log Pis Max 66.1287 +trainer/Log Pis Min -12.6302 +trainer/policy/mean Mean -0.0449226 +trainer/policy/mean Std 0.904984 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.81781 +trainer/policy/normal/std Std 0.672148 +trainer/policy/normal/std Max 5.94143 +trainer/policy/normal/std Min 0.295195 +trainer/policy/normal/log_std Mean 0.994285 +trainer/policy/normal/log_std Std 0.327475 +trainer/policy/normal/log_std Max 1.78195 +trainer/policy/normal/log_std Min -1.22012 +eval/num steps total 669085 +eval/num paths total 670 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.143776 +eval/Actions Std 0.927016 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.49499 +time/logging (s) 0.00371675 +time/sampling batch (s) 0.279222 +time/saving (s) 0.00339363 +time/training (s) 6.95523 +time/epoch (s) 9.73655 +time/total (s) 6855.39 +Epoch -331 +---------------------------------- --------------- +2022-05-10 15:05:10.898120 PDT | [2] Epoch -330 finished +---------------------------------- --------------- +epoch -330 +replay_buffer/size 999033 +trainer/num train calls 671000 +trainer/Policy Loss -19.7051 +trainer/Log Pis Mean 24.62 +trainer/Log Pis Std 13.0845 +trainer/Log Pis Max 81.5972 +trainer/Log Pis Min -3.62483 +trainer/policy/mean Mean -0.0584864 +trainer/policy/mean Std 0.907512 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.74669 +trainer/policy/normal/std Std 0.649666 +trainer/policy/normal/std Max 5.72203 +trainer/policy/normal/std Min 0.244745 +trainer/policy/normal/log_std Mean 0.968569 +trainer/policy/normal/log_std Std 0.329623 +trainer/policy/normal/log_std Max 1.74432 +trainer/policy/normal/log_std Min -1.40754 +eval/num steps total 670085 +eval/num paths total 671 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0827163 +eval/Actions Std 0.913048 +eval/Actions Max 0.999999 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.4881 +time/logging (s) 0.00387822 +time/sampling batch (s) 0.275291 +time/saving (s) 0.0034474 +time/training (s) 6.92939 +time/epoch (s) 9.70011 +time/total (s) 6865.09 +Epoch -330 +---------------------------------- --------------- +2022-05-10 15:05:21.420469 PDT | [2] Epoch -329 finished +---------------------------------- --------------- +epoch -329 +replay_buffer/size 999033 +trainer/num train calls 672000 +trainer/Policy Loss -19.6604 +trainer/Log Pis Mean 24.4507 +trainer/Log Pis Std 13.8999 +trainer/Log Pis Max 72.301 +trainer/Log Pis Min -5.55588 +trainer/policy/mean Mean -0.0163231 +trainer/policy/mean Std 0.905957 +trainer/policy/mean Max 0.999982 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.86995 +trainer/policy/normal/std Std 0.67489 +trainer/policy/normal/std Max 6.80352 +trainer/policy/normal/std Min 0.254367 +trainer/policy/normal/log_std Mean 1.01382 +trainer/policy/normal/log_std Std 0.323666 +trainer/policy/normal/log_std Max 1.91744 +trainer/policy/normal/log_std Min -1.36898 +eval/num steps total 671085 +eval/num paths total 672 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0825405 +eval/Actions Std 0.89691 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51772 +time/logging (s) 0.00401079 +time/sampling batch (s) 0.526559 +time/saving (s) 0.00378064 +time/training (s) 7.44726 +time/epoch (s) 10.4993 +time/total (s) 6875.6 +Epoch -329 +---------------------------------- --------------- +2022-05-10 15:05:30.849382 PDT | [2] Epoch -328 finished +---------------------------------- --------------- +epoch -328 +replay_buffer/size 999033 +trainer/num train calls 673000 +trainer/Policy Loss -19.7078 +trainer/Log Pis Mean 24.6874 +trainer/Log Pis Std 13.3811 +trainer/Log Pis Max 80.0558 +trainer/Log Pis Min -10.3823 +trainer/policy/mean Mean -0.035636 +trainer/policy/mean Std 0.908496 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999984 +trainer/policy/normal/std Mean 2.80039 +trainer/policy/normal/std Std 0.691975 +trainer/policy/normal/std Max 5.56263 +trainer/policy/normal/std Min 0.30869 +trainer/policy/normal/log_std Mean 0.984516 +trainer/policy/normal/log_std Std 0.342484 +trainer/policy/normal/log_std Max 1.71607 +trainer/policy/normal/log_std Min -1.17542 +eval/num steps total 672085 +eval/num paths total 673 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.218239 +eval/Actions Std 0.93215 +eval/Actions Max 0.999979 +eval/Actions Min -0.999984 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.52283 +time/logging (s) 0.00390229 +time/sampling batch (s) 0.526293 +time/saving (s) 0.00337363 +time/training (s) 6.34915 +time/epoch (s) 9.40554 +time/total (s) 6885.01 +Epoch -328 +---------------------------------- --------------- +2022-05-10 15:05:41.323864 PDT | [2] Epoch -327 finished +---------------------------------- --------------- +epoch -327 +replay_buffer/size 999033 +trainer/num train calls 674000 +trainer/Policy Loss -18.5043 +trainer/Log Pis Mean 24.2331 +trainer/Log Pis Std 12.9648 +trainer/Log Pis Max 69.4489 +trainer/Log Pis Min -4.01844 +trainer/policy/mean Mean -0.0352734 +trainer/policy/mean Std 0.905353 +trainer/policy/mean Max 0.999982 +trainer/policy/mean Min -0.999985 +trainer/policy/normal/std Mean 2.80289 +trainer/policy/normal/std Std 0.672942 +trainer/policy/normal/std Max 6.63066 +trainer/policy/normal/std Min 0.319962 +trainer/policy/normal/log_std Mean 0.988877 +trainer/policy/normal/log_std Std 0.327594 +trainer/policy/normal/log_std Max 1.8917 +trainer/policy/normal/log_std Min -1.13955 +eval/num steps total 673085 +eval/num paths total 674 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0973002 +eval/Actions Std 0.920911 +eval/Actions Max 0.999989 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70592 +time/logging (s) 0.00372562 +time/sampling batch (s) 0.527986 +time/saving (s) 0.00339153 +time/training (s) 7.21001 +time/epoch (s) 10.451 +time/total (s) 6895.46 +Epoch -327 +---------------------------------- --------------- +2022-05-10 15:05:51.737734 PDT | [2] Epoch -326 finished +---------------------------------- --------------- +epoch -326 +replay_buffer/size 999033 +trainer/num train calls 675000 +trainer/Policy Loss -19.3547 +trainer/Log Pis Mean 24.5048 +trainer/Log Pis Std 13.3966 +trainer/Log Pis Max 65.6976 +trainer/Log Pis Min -11.9262 +trainer/policy/mean Mean -0.0334625 +trainer/policy/mean Std 0.906864 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.77654 +trainer/policy/normal/std Std 0.655168 +trainer/policy/normal/std Max 6.14298 +trainer/policy/normal/std Min 0.201526 +trainer/policy/normal/log_std Mean 0.981026 +trainer/policy/normal/log_std Std 0.320481 +trainer/policy/normal/log_std Max 1.81531 +trainer/policy/normal/log_std Min -1.60183 +eval/num steps total 674085 +eval/num paths total 675 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0664958 +eval/Actions Std 0.910931 +eval/Actions Max 0.999999 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.652 +time/logging (s) 0.00374527 +time/sampling batch (s) 0.276775 +time/saving (s) 0.0033734 +time/training (s) 7.45486 +time/epoch (s) 10.3908 +time/total (s) 6905.85 +Epoch -326 +---------------------------------- --------------- +2022-05-10 15:06:01.407146 PDT | [2] Epoch -325 finished +---------------------------------- --------------- +epoch -325 +replay_buffer/size 999033 +trainer/num train calls 676000 +trainer/Policy Loss -19.4133 +trainer/Log Pis Mean 24.2889 +trainer/Log Pis Std 13.1886 +trainer/Log Pis Max 68.9972 +trainer/Log Pis Min -7.50754 +trainer/policy/mean Mean -0.0347254 +trainer/policy/mean Std 0.90579 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.83883 +trainer/policy/normal/std Std 0.651012 +trainer/policy/normal/std Max 6.49714 +trainer/policy/normal/std Min 0.309279 +trainer/policy/normal/log_std Mean 1.00648 +trainer/policy/normal/log_std Std 0.305047 +trainer/policy/normal/log_std Max 1.87136 +trainer/policy/normal/log_std Min -1.17351 +eval/num steps total 675085 +eval/num paths total 676 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.177966 +eval/Actions Std 0.918904 +eval/Actions Max 1 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.46438 +time/logging (s) 0.00370414 +time/sampling batch (s) 0.277766 +time/saving (s) 0.00341132 +time/training (s) 6.897 +time/epoch (s) 9.64626 +time/total (s) 6915.5 +Epoch -325 +---------------------------------- --------------- +2022-05-10 15:06:11.248385 PDT | [2] Epoch -324 finished +---------------------------------- --------------- +epoch -324 +replay_buffer/size 999033 +trainer/num train calls 677000 +trainer/Policy Loss -18.9084 +trainer/Log Pis Mean 24.407 +trainer/Log Pis Std 13.6054 +trainer/Log Pis Max 80.1943 +trainer/Log Pis Min -8.80988 +trainer/policy/mean Mean -0.0240101 +trainer/policy/mean Std 0.907327 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.87333 +trainer/policy/normal/std Std 0.698887 +trainer/policy/normal/std Max 7.08712 +trainer/policy/normal/std Min 0.260837 +trainer/policy/normal/log_std Mean 1.01299 +trainer/policy/normal/log_std Std 0.329491 +trainer/policy/normal/log_std Max 1.95828 +trainer/policy/normal/log_std Min -1.34386 +eval/num steps total 676085 +eval/num paths total 677 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00956683 +eval/Actions Std 0.906835 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51168 +time/logging (s) 0.00373749 +time/sampling batch (s) 0.525256 +time/saving (s) 0.00335434 +time/training (s) 6.77431 +time/epoch (s) 9.81834 +time/total (s) 6925.32 +Epoch -324 +---------------------------------- --------------- +2022-05-10 15:06:22.237480 PDT | [2] Epoch -323 finished +---------------------------------- --------------- +epoch -323 +replay_buffer/size 999033 +trainer/num train calls 678000 +trainer/Policy Loss -19.6373 +trainer/Log Pis Mean 23.9634 +trainer/Log Pis Std 12.9861 +trainer/Log Pis Max 65.1128 +trainer/Log Pis Min -4.67513 +trainer/policy/mean Mean -0.0255418 +trainer/policy/mean Std 0.906374 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.80659 +trainer/policy/normal/std Std 0.651539 +trainer/policy/normal/std Max 6.21597 +trainer/policy/normal/std Min 0.24804 +trainer/policy/normal/log_std Mean 0.993847 +trainer/policy/normal/log_std Std 0.310845 +trainer/policy/normal/log_std Max 1.82712 +trainer/policy/normal/log_std Min -1.39416 +eval/num steps total 677085 +eval/num paths total 678 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.16801 +eval/Actions Std 0.851323 +eval/Actions Max 1 +eval/Actions Min -0.999975 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70476 +time/logging (s) 0.00372363 +time/sampling batch (s) 0.280332 +time/saving (s) 0.0034947 +time/training (s) 7.97345 +time/epoch (s) 10.9658 +time/total (s) 6936.29 +Epoch -323 +---------------------------------- --------------- +2022-05-10 15:06:32.489191 PDT | [2] Epoch -322 finished +---------------------------------- --------------- +epoch -322 +replay_buffer/size 999033 +trainer/num train calls 679000 +trainer/Policy Loss -19.6385 +trainer/Log Pis Mean 25.386 +trainer/Log Pis Std 13.9917 +trainer/Log Pis Max 79.0148 +trainer/Log Pis Min -11.6427 +trainer/policy/mean Mean -0.0333546 +trainer/policy/mean Std 0.907547 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.8068 +trainer/policy/normal/std Std 0.65809 +trainer/policy/normal/std Max 5.86413 +trainer/policy/normal/std Min 0.29781 +trainer/policy/normal/log_std Mean 0.992567 +trainer/policy/normal/log_std Std 0.31725 +trainer/policy/normal/log_std Max 1.76885 +trainer/policy/normal/log_std Min -1.2113 +eval/num steps total 678085 +eval/num paths total 679 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0763719 +eval/Actions Std 0.921646 +eval/Actions Max 0.999999 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.40936 +time/logging (s) 0.00410108 +time/sampling batch (s) 0.52958 +time/saving (s) 0.00377931 +time/training (s) 7.28172 +time/epoch (s) 10.2285 +time/total (s) 6946.52 +Epoch -322 +---------------------------------- --------------- +2022-05-10 15:06:43.527727 PDT | [2] Epoch -321 finished +---------------------------------- --------------- +epoch -321 +replay_buffer/size 999033 +trainer/num train calls 680000 +trainer/Policy Loss -18.8981 +trainer/Log Pis Mean 24.6293 +trainer/Log Pis Std 12.4318 +trainer/Log Pis Max 66.9243 +trainer/Log Pis Min -2.90586 +trainer/policy/mean Mean -0.0499683 +trainer/policy/mean Std 0.907088 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.81865 +trainer/policy/normal/std Std 0.650314 +trainer/policy/normal/std Max 6.88092 +trainer/policy/normal/std Min 0.327694 +trainer/policy/normal/log_std Mean 0.998042 +trainer/policy/normal/log_std Std 0.312806 +trainer/policy/normal/log_std Max 1.92875 +trainer/policy/normal/log_std Min -1.11567 +eval/num steps total 679081 +eval/num paths total 680 +eval/path length Mean 996 +eval/path length Std 0 +eval/path length Max 996 +eval/path length Min 996 +eval/Rewards Mean 0.00100402 +eval/Rewards Std 0.0316703 +eval/Rewards Max 1 +eval/Rewards Min 0 +eval/Returns Mean 1 +eval/Returns Std 0 +eval/Returns Max 1 +eval/Returns Min 1 +eval/Actions Mean -0.0933566 +eval/Actions Std 0.919267 +eval/Actions Max 0.999996 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 1 +eval/normalized_score 100 +time/evaluation sampling (s) 2.60835 +time/logging (s) 0.00377107 +time/sampling batch (s) 0.275694 +time/saving (s) 0.00357647 +time/training (s) 8.12322 +time/epoch (s) 11.0146 +time/total (s) 6957.54 +Epoch -321 +---------------------------------- --------------- +2022-05-10 15:06:54.453039 PDT | [2] Epoch -320 finished +---------------------------------- --------------- +epoch -320 +replay_buffer/size 999033 +trainer/num train calls 681000 +trainer/Policy Loss -19.3175 +trainer/Log Pis Mean 24.1608 +trainer/Log Pis Std 13.6906 +trainer/Log Pis Max 71.6633 +trainer/Log Pis Min -11.3315 +trainer/policy/mean Mean -0.0499023 +trainer/policy/mean Std 0.902854 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.80743 +trainer/policy/normal/std Std 0.663075 +trainer/policy/normal/std Max 6.31686 +trainer/policy/normal/std Min 0.324939 +trainer/policy/normal/log_std Mean 0.9917 +trainer/policy/normal/log_std Std 0.32226 +trainer/policy/normal/log_std Max 1.84322 +trainer/policy/normal/log_std Min -1.12412 +eval/num steps total 680081 +eval/num paths total 681 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.164315 +eval/Actions Std 0.921498 +eval/Actions Max 0.999998 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5629 +time/logging (s) 0.00368768 +time/sampling batch (s) 0.527687 +time/saving (s) 0.00341207 +time/training (s) 7.80427 +time/epoch (s) 10.902 +time/total (s) 6968.45 +Epoch -320 +---------------------------------- --------------- +2022-05-10 15:07:06.187959 PDT | [2] Epoch -319 finished +---------------------------------- --------------- +epoch -319 +replay_buffer/size 999033 +trainer/num train calls 682000 +trainer/Policy Loss -19.2547 +trainer/Log Pis Mean 25.1198 +trainer/Log Pis Std 12.9546 +trainer/Log Pis Max 64.6765 +trainer/Log Pis Min -6.6777 +trainer/policy/mean Mean -0.0415971 +trainer/policy/mean Std 0.905493 +trainer/policy/mean Max 0.999981 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.81308 +trainer/policy/normal/std Std 0.683123 +trainer/policy/normal/std Max 6.70826 +trainer/policy/normal/std Min 0.325562 +trainer/policy/normal/log_std Mean 0.991064 +trainer/policy/normal/log_std Std 0.333487 +trainer/policy/normal/log_std Max 1.90334 +trainer/policy/normal/log_std Min -1.1222 +eval/num steps total 681081 +eval/num paths total 682 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.194193 +eval/Actions Std 0.943246 +eval/Actions Max 0.999993 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60118 +time/logging (s) 0.0038011 +time/sampling batch (s) 0.550772 +time/saving (s) 0.00355169 +time/training (s) 8.5521 +time/epoch (s) 11.7114 +time/total (s) 6980.16 +Epoch -319 +---------------------------------- --------------- +2022-05-10 15:07:16.254902 PDT | [2] Epoch -318 finished +---------------------------------- --------------- +epoch -318 +replay_buffer/size 999033 +trainer/num train calls 683000 +trainer/Policy Loss -19.2842 +trainer/Log Pis Mean 23.9288 +trainer/Log Pis Std 12.7012 +trainer/Log Pis Max 62.4997 +trainer/Log Pis Min -15.388 +trainer/policy/mean Mean -0.0466151 +trainer/policy/mean Std 0.910824 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.74927 +trainer/policy/normal/std Std 0.643741 +trainer/policy/normal/std Max 5.64538 +trainer/policy/normal/std Min 0.275992 +trainer/policy/normal/log_std Mean 0.971492 +trainer/policy/normal/log_std Std 0.319735 +trainer/policy/normal/log_std Max 1.73084 +trainer/policy/normal/log_std Min -1.28738 +eval/num steps total 682081 +eval/num paths total 683 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.37549 +eval/Actions Std 0.840828 +eval/Actions Max 0.999986 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.90272 +time/logging (s) 0.00371086 +time/sampling batch (s) 0.28736 +time/saving (s) 0.00341178 +time/training (s) 6.84614 +time/epoch (s) 10.0433 +time/total (s) 6990.21 +Epoch -318 +---------------------------------- --------------- +2022-05-10 15:07:26.078573 PDT | [2] Epoch -317 finished +---------------------------------- --------------- +epoch -317 +replay_buffer/size 999033 +trainer/num train calls 684000 +trainer/Policy Loss -19.4329 +trainer/Log Pis Mean 24.1692 +trainer/Log Pis Std 13.4412 +trainer/Log Pis Max 64.3549 +trainer/Log Pis Min -4.34894 +trainer/policy/mean Mean -0.0580589 +trainer/policy/mean Std 0.901853 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.81421 +trainer/policy/normal/std Std 0.685826 +trainer/policy/normal/std Max 6.80551 +trainer/policy/normal/std Min 0.251043 +trainer/policy/normal/log_std Mean 0.990144 +trainer/policy/normal/log_std Std 0.340769 +trainer/policy/normal/log_std Max 1.91773 +trainer/policy/normal/log_std Min -1.38213 +eval/num steps total 683081 +eval/num paths total 684 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0213216 +eval/Actions Std 0.956157 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68966 +time/logging (s) 0.00411965 +time/sampling batch (s) 0.531535 +time/saving (s) 0.00376351 +time/training (s) 6.57132 +time/epoch (s) 9.8004 +time/total (s) 7000.01 +Epoch -317 +---------------------------------- --------------- +2022-05-10 15:07:36.617010 PDT | [2] Epoch -316 finished +---------------------------------- --------------- +epoch -316 +replay_buffer/size 999033 +trainer/num train calls 685000 +trainer/Policy Loss -20.3102 +trainer/Log Pis Mean 23.0748 +trainer/Log Pis Std 12.6083 +trainer/Log Pis Max 60.9809 +trainer/Log Pis Min -7.46712 +trainer/policy/mean Mean -0.0331112 +trainer/policy/mean Std 0.909086 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.80459 +trainer/policy/normal/std Std 0.661116 +trainer/policy/normal/std Max 5.93161 +trainer/policy/normal/std Min 0.245149 +trainer/policy/normal/log_std Mean 0.991137 +trainer/policy/normal/log_std Std 0.321145 +trainer/policy/normal/log_std Max 1.78029 +trainer/policy/normal/log_std Min -1.40589 +eval/num steps total 684081 +eval/num paths total 685 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0871264 +eval/Actions Std 0.908597 +eval/Actions Max 0.999997 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51205 +time/logging (s) 0.00378575 +time/sampling batch (s) 0.279639 +time/saving (s) 0.00352362 +time/training (s) 7.7152 +time/epoch (s) 10.5142 +time/total (s) 7010.53 +Epoch -316 +---------------------------------- --------------- +2022-05-10 15:07:46.863020 PDT | [2] Epoch -315 finished +---------------------------------- --------------- +epoch -315 +replay_buffer/size 999033 +trainer/num train calls 686000 +trainer/Policy Loss -20.4401 +trainer/Log Pis Mean 25.5685 +trainer/Log Pis Std 12.9737 +trainer/Log Pis Max 78.0395 +trainer/Log Pis Min -9.0388 +trainer/policy/mean Mean -0.0212839 +trainer/policy/mean Std 0.908979 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.84698 +trainer/policy/normal/std Std 0.66495 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.312514 +trainer/policy/normal/log_std Mean 1.00638 +trainer/policy/normal/log_std Std 0.320372 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.16311 +eval/num steps total 685081 +eval/num paths total 686 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.212504 +eval/Actions Std 0.77667 +eval/Actions Max 0.999991 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.72338 +time/logging (s) 0.00377671 +time/sampling batch (s) 0.286396 +time/saving (s) 0.00347081 +time/training (s) 7.20523 +time/epoch (s) 10.2223 +time/total (s) 7020.75 +Epoch -315 +---------------------------------- --------------- +2022-05-10 15:07:56.945168 PDT | [2] Epoch -314 finished +---------------------------------- --------------- +epoch -314 +replay_buffer/size 999033 +trainer/num train calls 687000 +trainer/Policy Loss -18.5725 +trainer/Log Pis Mean 24.9438 +trainer/Log Pis Std 13.3211 +trainer/Log Pis Max 82.5546 +trainer/Log Pis Min -4.69236 +trainer/policy/mean Mean -0.0278584 +trainer/policy/mean Std 0.904506 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.78348 +trainer/policy/normal/std Std 0.668413 +trainer/policy/normal/std Max 6.38133 +trainer/policy/normal/std Min 0.264188 +trainer/policy/normal/log_std Mean 0.981871 +trainer/policy/normal/log_std Std 0.327881 +trainer/policy/normal/log_std Max 1.85338 +trainer/policy/normal/log_std Min -1.3311 +eval/num steps total 686081 +eval/num paths total 687 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0499071 +eval/Actions Std 0.899502 +eval/Actions Max 0.999998 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70207 +time/logging (s) 0.00374789 +time/sampling batch (s) 0.279216 +time/saving (s) 0.00343012 +time/training (s) 7.07023 +time/epoch (s) 10.0587 +time/total (s) 7030.81 +Epoch -314 +---------------------------------- --------------- +2022-05-10 15:08:06.851024 PDT | [2] Epoch -313 finished +---------------------------------- --------------- +epoch -313 +replay_buffer/size 999033 +trainer/num train calls 688000 +trainer/Policy Loss -19.0139 +trainer/Log Pis Mean 23.5773 +trainer/Log Pis Std 12.909 +trainer/Log Pis Max 79.3118 +trainer/Log Pis Min -10.284 +trainer/policy/mean Mean -0.050343 +trainer/policy/mean Std 0.901686 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.74978 +trainer/policy/normal/std Std 0.643857 +trainer/policy/normal/std Max 5.51578 +trainer/policy/normal/std Min 0.257696 +trainer/policy/normal/log_std Mean 0.972291 +trainer/policy/normal/log_std Std 0.315208 +trainer/policy/normal/log_std Max 1.70761 +trainer/policy/normal/log_std Min -1.35597 +eval/num steps total 687081 +eval/num paths total 688 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.228432 +eval/Actions Std 0.888313 +eval/Actions Max 0.999992 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.52461 +time/logging (s) 0.00381792 +time/sampling batch (s) 0.279662 +time/saving (s) 0.00341168 +time/training (s) 7.07075 +time/epoch (s) 9.88225 +time/total (s) 7040.7 +Epoch -313 +---------------------------------- --------------- +2022-05-10 15:08:17.127782 PDT | [2] Epoch -312 finished +---------------------------------- --------------- +epoch -312 +replay_buffer/size 999033 +trainer/num train calls 689000 +trainer/Policy Loss -19.1549 +trainer/Log Pis Mean 23.8775 +trainer/Log Pis Std 13.1037 +trainer/Log Pis Max 68.554 +trainer/Log Pis Min -9.60035 +trainer/policy/mean Mean -0.0421761 +trainer/policy/mean Std 0.909308 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.7912 +trainer/policy/normal/std Std 0.6342 +trainer/policy/normal/std Max 5.79337 +trainer/policy/normal/std Min 0.298272 +trainer/policy/normal/log_std Mean 0.989136 +trainer/policy/normal/log_std Std 0.309339 +trainer/policy/normal/log_std Max 1.75671 +trainer/policy/normal/log_std Min -1.20975 +eval/num steps total 688081 +eval/num paths total 689 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0765909 +eval/Actions Std 0.903619 +eval/Actions Max 0.999995 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.74891 +time/logging (s) 0.00405866 +time/sampling batch (s) 0.529716 +time/saving (s) 0.00377209 +time/training (s) 6.967 +time/epoch (s) 10.2535 +time/total (s) 7050.96 +Epoch -312 +---------------------------------- --------------- +2022-05-10 15:08:26.609861 PDT | [2] Epoch -311 finished +---------------------------------- --------------- +epoch -311 +replay_buffer/size 999033 +trainer/num train calls 690000 +trainer/Policy Loss -19.7313 +trainer/Log Pis Mean 24.65 +trainer/Log Pis Std 12.8528 +trainer/Log Pis Max 60.0195 +trainer/Log Pis Min -4.41051 +trainer/policy/mean Mean -0.0210782 +trainer/policy/mean Std 0.908403 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.87933 +trainer/policy/normal/std Std 0.673869 +trainer/policy/normal/std Max 7.06937 +trainer/policy/normal/std Min 0.299855 +trainer/policy/normal/log_std Mean 1.02005 +trainer/policy/normal/log_std Std 0.30541 +trainer/policy/normal/log_std Max 1.95577 +trainer/policy/normal/log_std Min -1.20446 +eval/num steps total 689081 +eval/num paths total 690 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00319209 +eval/Actions Std 0.912608 +eval/Actions Max 0.999995 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70652 +time/logging (s) 0.00379094 +time/sampling batch (s) 0.281107 +time/saving (s) 0.00360143 +time/training (s) 6.46254 +time/epoch (s) 9.45756 +time/total (s) 7060.42 +Epoch -311 +---------------------------------- --------------- +2022-05-10 15:08:36.445244 PDT | [2] Epoch -310 finished +---------------------------------- --------------- +epoch -310 +replay_buffer/size 999033 +trainer/num train calls 691000 +trainer/Policy Loss -18.7505 +trainer/Log Pis Mean 24.7371 +trainer/Log Pis Std 13.8306 +trainer/Log Pis Max 80.61 +trainer/Log Pis Min -18.28 +trainer/policy/mean Mean -0.0328409 +trainer/policy/mean Std 0.907385 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.77765 +trainer/policy/normal/std Std 0.652428 +trainer/policy/normal/std Max 5.95297 +trainer/policy/normal/std Min 0.218242 +trainer/policy/normal/log_std Mean 0.981325 +trainer/policy/normal/log_std Std 0.322031 +trainer/policy/normal/log_std Max 1.78389 +trainer/policy/normal/log_std Min -1.52215 +eval/num steps total 690081 +eval/num paths total 691 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0189061 +eval/Actions Std 0.943852 +eval/Actions Max 0.999995 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57641 +time/logging (s) 0.00369423 +time/sampling batch (s) 0.276038 +time/saving (s) 0.00338843 +time/training (s) 6.95269 +time/epoch (s) 9.81223 +time/total (s) 7070.23 +Epoch -310 +---------------------------------- --------------- +2022-05-10 15:08:46.667149 PDT | [2] Epoch -309 finished +---------------------------------- --------------- +epoch -309 +replay_buffer/size 999033 +trainer/num train calls 692000 +trainer/Policy Loss -19.9396 +trainer/Log Pis Mean 23.697 +trainer/Log Pis Std 13.113 +trainer/Log Pis Max 59.2637 +trainer/Log Pis Min -6.89367 +trainer/policy/mean Mean -0.0365021 +trainer/policy/mean Std 0.903371 +trainer/policy/mean Max 0.999977 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.79585 +trainer/policy/normal/std Std 0.658783 +trainer/policy/normal/std Max 6.13336 +trainer/policy/normal/std Min 0.304639 +trainer/policy/normal/log_std Mean 0.988423 +trainer/policy/normal/log_std Std 0.317253 +trainer/policy/normal/log_std Max 1.81374 +trainer/policy/normal/log_std Min -1.18863 +eval/num steps total 691081 +eval/num paths total 692 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.117872 +eval/Actions Std 0.904976 +eval/Actions Max 0.999997 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66247 +time/logging (s) 0.00368157 +time/sampling batch (s) 0.52406 +time/saving (s) 0.0033701 +time/training (s) 7.00572 +time/epoch (s) 10.1993 +time/total (s) 7080.43 +Epoch -309 +---------------------------------- --------------- +2022-05-10 15:08:55.952946 PDT | [2] Epoch -308 finished +---------------------------------- --------------- +epoch -308 +replay_buffer/size 999033 +trainer/num train calls 693000 +trainer/Policy Loss -19.8904 +trainer/Log Pis Mean 25.3051 +trainer/Log Pis Std 12.976 +trainer/Log Pis Max 69.5014 +trainer/Log Pis Min -9.6785 +trainer/policy/mean Mean -0.0315736 +trainer/policy/mean Std 0.905138 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.75111 +trainer/policy/normal/std Std 0.673441 +trainer/policy/normal/std Max 5.99451 +trainer/policy/normal/std Min 0.2976 +trainer/policy/normal/log_std Mean 0.968626 +trainer/policy/normal/log_std Std 0.332695 +trainer/policy/normal/log_std Max 1.79084 +trainer/policy/normal/log_std Min -1.21201 +eval/num steps total 692081 +eval/num paths total 693 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.086641 +eval/Actions Std 0.915258 +eval/Actions Max 0.99998 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5819 +time/logging (s) 0.00369749 +time/sampling batch (s) 0.522366 +time/saving (s) 0.00333571 +time/training (s) 6.15195 +time/epoch (s) 9.26325 +time/total (s) 7089.7 +Epoch -308 +---------------------------------- --------------- +2022-05-10 15:09:05.394482 PDT | [2] Epoch -307 finished +---------------------------------- --------------- +epoch -307 +replay_buffer/size 999033 +trainer/num train calls 694000 +trainer/Policy Loss -20.1829 +trainer/Log Pis Mean 24.3614 +trainer/Log Pis Std 13.4025 +trainer/Log Pis Max 76.7142 +trainer/Log Pis Min -10.7899 +trainer/policy/mean Mean -0.0528596 +trainer/policy/mean Std 0.903182 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.80499 +trainer/policy/normal/std Std 0.661323 +trainer/policy/normal/std Max 6.05056 +trainer/policy/normal/std Min 0.249893 +trainer/policy/normal/log_std Mean 0.99079 +trainer/policy/normal/log_std Std 0.322663 +trainer/policy/normal/log_std Max 1.80015 +trainer/policy/normal/log_std Min -1.38672 +eval/num steps total 693081 +eval/num paths total 694 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.123686 +eval/Actions Std 0.883708 +eval/Actions Max 0.999983 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.29425 +time/logging (s) 0.00369307 +time/sampling batch (s) 0.273662 +time/saving (s) 0.00334108 +time/training (s) 6.84396 +time/epoch (s) 9.4189 +time/total (s) 7099.12 +Epoch -307 +---------------------------------- --------------- +2022-05-10 15:09:16.124535 PDT | [2] Epoch -306 finished +---------------------------------- --------------- +epoch -306 +replay_buffer/size 999033 +trainer/num train calls 695000 +trainer/Policy Loss -20.4513 +trainer/Log Pis Mean 24.5528 +trainer/Log Pis Std 12.6141 +trainer/Log Pis Max 64.5745 +trainer/Log Pis Min -7.5039 +trainer/policy/mean Mean -0.0480829 +trainer/policy/mean Std 0.909939 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.8075 +trainer/policy/normal/std Std 0.660357 +trainer/policy/normal/std Max 5.43361 +trainer/policy/normal/std Min 0.281811 +trainer/policy/normal/log_std Mean 0.993084 +trainer/policy/normal/log_std Std 0.314414 +trainer/policy/normal/log_std Max 1.6926 +trainer/policy/normal/log_std Min -1.26652 +eval/num steps total 694081 +eval/num paths total 695 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.346485 +eval/Actions Std 0.849757 +eval/Actions Max 0.999996 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55758 +time/logging (s) 0.00373052 +time/sampling batch (s) 0.271737 +time/saving (s) 0.0033829 +time/training (s) 7.87105 +time/epoch (s) 10.7075 +time/total (s) 7109.83 +Epoch -306 +---------------------------------- --------------- +2022-05-10 15:09:26.397998 PDT | [2] Epoch -305 finished +---------------------------------- --------------- +epoch -305 +replay_buffer/size 999033 +trainer/num train calls 696000 +trainer/Policy Loss -20.4447 +trainer/Log Pis Mean 24.7755 +trainer/Log Pis Std 13.7543 +trainer/Log Pis Max 70.6349 +trainer/Log Pis Min -12.3174 +trainer/policy/mean Mean -0.0341421 +trainer/policy/mean Std 0.905663 +trainer/policy/mean Max 0.999979 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.82876 +trainer/policy/normal/std Std 0.665412 +trainer/policy/normal/std Max 6.47496 +trainer/policy/normal/std Min 0.299901 +trainer/policy/normal/log_std Mean 0.998658 +trainer/policy/normal/log_std Std 0.327421 +trainer/policy/normal/log_std Max 1.86794 +trainer/policy/normal/log_std Min -1.2043 +eval/num steps total 695081 +eval/num paths total 696 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0906124 +eval/Actions Std 0.908749 +eval/Actions Max 0.999995 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55757 +time/logging (s) 0.0040822 +time/sampling batch (s) 0.278665 +time/saving (s) 0.00370551 +time/training (s) 7.40655 +time/epoch (s) 10.2506 +time/total (s) 7120.09 +Epoch -305 +---------------------------------- --------------- +2022-05-10 15:09:36.614624 PDT | [2] Epoch -304 finished +---------------------------------- --------------- +epoch -304 +replay_buffer/size 999033 +trainer/num train calls 697000 +trainer/Policy Loss -20.7947 +trainer/Log Pis Mean 25.0472 +trainer/Log Pis Std 13.6069 +trainer/Log Pis Max 75.7782 +trainer/Log Pis Min -8.24714 +trainer/policy/mean Mean -0.0318902 +trainer/policy/mean Std 0.908794 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.85756 +trainer/policy/normal/std Std 0.671138 +trainer/policy/normal/std Max 6.23934 +trainer/policy/normal/std Min 0.347556 +trainer/policy/normal/log_std Mean 1.00902 +trainer/policy/normal/log_std Std 0.325219 +trainer/policy/normal/log_std Max 1.83087 +trainer/policy/normal/log_std Min -1.05683 +eval/num steps total 696081 +eval/num paths total 697 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.289105 +eval/Actions Std 0.889216 +eval/Actions Max 0.999989 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60247 +time/logging (s) 0.00386123 +time/sampling batch (s) 0.279426 +time/saving (s) 0.00338445 +time/training (s) 7.30394 +time/epoch (s) 10.1931 +time/total (s) 7130.28 +Epoch -304 +---------------------------------- --------------- +2022-05-10 15:09:47.018482 PDT | [2] Epoch -303 finished +---------------------------------- --------------- +epoch -303 +replay_buffer/size 999033 +trainer/num train calls 698000 +trainer/Policy Loss -19.8481 +trainer/Log Pis Mean 24.5188 +trainer/Log Pis Std 13.4106 +trainer/Log Pis Max 62.2569 +trainer/Log Pis Min -10.145 +trainer/policy/mean Mean -0.0275563 +trainer/policy/mean Std 0.906263 +trainer/policy/mean Max 0.999974 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.80393 +trainer/policy/normal/std Std 0.688093 +trainer/policy/normal/std Max 5.59425 +trainer/policy/normal/std Min 0.295438 +trainer/policy/normal/log_std Mean 0.985971 +trainer/policy/normal/log_std Std 0.342678 +trainer/policy/normal/log_std Max 1.72174 +trainer/policy/normal/log_std Min -1.2193 +eval/num steps total 697081 +eval/num paths total 698 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0109204 +eval/Actions Std 0.959727 +eval/Actions Max 0.999993 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60593 +time/logging (s) 0.00408342 +time/sampling batch (s) 0.279002 +time/saving (s) 0.00383244 +time/training (s) 7.48805 +time/epoch (s) 10.3809 +time/total (s) 7140.67 +Epoch -303 +---------------------------------- --------------- +2022-05-10 15:09:57.349713 PDT | [2] Epoch -302 finished +---------------------------------- --------------- +epoch -302 +replay_buffer/size 999033 +trainer/num train calls 699000 +trainer/Policy Loss -19.5531 +trainer/Log Pis Mean 24.4457 +trainer/Log Pis Std 13.1455 +trainer/Log Pis Max 65.2848 +trainer/Log Pis Min -4.94248 +trainer/policy/mean Mean -0.0475076 +trainer/policy/mean Std 0.906916 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.74101 +trainer/policy/normal/std Std 0.650734 +trainer/policy/normal/std Max 5.07211 +trainer/policy/normal/std Min 0.284089 +trainer/policy/normal/log_std Mean 0.967008 +trainer/policy/normal/log_std Std 0.325937 +trainer/policy/normal/log_std Max 1.62376 +trainer/policy/normal/log_std Min -1.25847 +eval/num steps total 698081 +eval/num paths total 699 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.192052 +eval/Actions Std 0.77771 +eval/Actions Max 0.999994 +eval/Actions Min -0.999983 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54413 +time/logging (s) 0.00372016 +time/sampling batch (s) 0.526234 +time/saving (s) 0.0033854 +time/training (s) 7.22991 +time/epoch (s) 10.3074 +time/total (s) 7150.98 +Epoch -302 +---------------------------------- --------------- +2022-05-10 15:10:06.948971 PDT | [2] Epoch -301 finished +---------------------------------- --------------- +epoch -301 +replay_buffer/size 999033 +trainer/num train calls 700000 +trainer/Policy Loss -21.0886 +trainer/Log Pis Mean 24.7804 +trainer/Log Pis Std 14.2782 +trainer/Log Pis Max 70.3002 +trainer/Log Pis Min -9.00788 +trainer/policy/mean Mean -0.0175678 +trainer/policy/mean Std 0.907299 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.83941 +trainer/policy/normal/std Std 0.702975 +trainer/policy/normal/std Max 5.87396 +trainer/policy/normal/std Min 0.282084 +trainer/policy/normal/log_std Mean 0.997921 +trainer/policy/normal/log_std Std 0.34467 +trainer/policy/normal/log_std Max 1.77053 +trainer/policy/normal/log_std Min -1.26555 +eval/num steps total 699081 +eval/num paths total 700 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0713384 +eval/Actions Std 0.918269 +eval/Actions Max 0.999991 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.33726 +time/logging (s) 0.00372953 +time/sampling batch (s) 0.274263 +time/saving (s) 0.00616502 +time/training (s) 6.95521 +time/epoch (s) 9.57663 +time/total (s) 7160.56 +Epoch -301 +---------------------------------- --------------- +2022-05-10 15:10:17.861626 PDT | [2] Epoch -300 finished +---------------------------------- --------------- +epoch -300 +replay_buffer/size 999033 +trainer/num train calls 701000 +trainer/Policy Loss -19.5298 +trainer/Log Pis Mean 23.6438 +trainer/Log Pis Std 12.6152 +trainer/Log Pis Max 67.8385 +trainer/Log Pis Min -9.01106 +trainer/policy/mean Mean -0.0462247 +trainer/policy/mean Std 0.906524 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.80534 +trainer/policy/normal/std Std 0.658677 +trainer/policy/normal/std Max 6.21215 +trainer/policy/normal/std Min 0.256311 +trainer/policy/normal/log_std Mean 0.99165 +trainer/policy/normal/log_std Std 0.320297 +trainer/policy/normal/log_std Max 1.82651 +trainer/policy/normal/log_std Min -1.36136 +eval/num steps total 700081 +eval/num paths total 701 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0958822 +eval/Actions Std 0.913922 +eval/Actions Max 0.999997 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65987 +time/logging (s) 0.00410501 +time/sampling batch (s) 0.530858 +time/saving (s) 0.00385233 +time/training (s) 7.69119 +time/epoch (s) 10.8899 +time/total (s) 7171.45 +Epoch -300 +---------------------------------- --------------- +2022-05-10 15:10:28.818043 PDT | [2] Epoch -299 finished +---------------------------------- --------------- +epoch -299 +replay_buffer/size 999033 +trainer/num train calls 702000 +trainer/Policy Loss -19.4613 +trainer/Log Pis Mean 24.9281 +trainer/Log Pis Std 13.6618 +trainer/Log Pis Max 79.3255 +trainer/Log Pis Min -8.84172 +trainer/policy/mean Mean -0.0395348 +trainer/policy/mean Std 0.905622 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.79011 +trainer/policy/normal/std Std 0.708424 +trainer/policy/normal/std Max 6.34316 +trainer/policy/normal/std Min 0.279547 +trainer/policy/normal/log_std Mean 0.975963 +trainer/policy/normal/log_std Std 0.36559 +trainer/policy/normal/log_std Max 1.84738 +trainer/policy/normal/log_std Min -1.27458 +eval/num steps total 701081 +eval/num paths total 702 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.307264 +eval/Actions Std 0.883141 +eval/Actions Max 0.999995 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67841 +time/logging (s) 0.00405987 +time/sampling batch (s) 0.528522 +time/saving (s) 0.00367742 +time/training (s) 7.71812 +time/epoch (s) 10.9328 +time/total (s) 7182.39 +Epoch -299 +---------------------------------- --------------- +2022-05-10 15:10:39.092705 PDT | [2] Epoch -298 finished +---------------------------------- --------------- +epoch -298 +replay_buffer/size 999033 +trainer/num train calls 703000 +trainer/Policy Loss -18.9191 +trainer/Log Pis Mean 24.7584 +trainer/Log Pis Std 13.0147 +trainer/Log Pis Max 77.3512 +trainer/Log Pis Min -3.08477 +trainer/policy/mean Mean -0.028263 +trainer/policy/mean Std 0.906832 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.80537 +trainer/policy/normal/std Std 0.667151 +trainer/policy/normal/std Max 5.9656 +trainer/policy/normal/std Min 0.342883 +trainer/policy/normal/log_std Mean 0.991104 +trainer/policy/normal/log_std Std 0.319515 +trainer/policy/normal/log_std Max 1.78601 +trainer/policy/normal/log_std Min -1.07037 +eval/num steps total 702081 +eval/num paths total 703 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.166985 +eval/Actions Std 0.819751 +eval/Actions Max 0.999957 +eval/Actions Min -0.999986 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6128 +time/logging (s) 0.00390004 +time/sampling batch (s) 0.524267 +time/saving (s) 0.00357865 +time/training (s) 7.10693 +time/epoch (s) 10.2515 +time/total (s) 7192.64 +Epoch -298 +---------------------------------- --------------- +2022-05-10 15:10:50.291833 PDT | [2] Epoch -297 finished +---------------------------------- --------------- +epoch -297 +replay_buffer/size 999033 +trainer/num train calls 704000 +trainer/Policy Loss -20.9414 +trainer/Log Pis Mean 25.7214 +trainer/Log Pis Std 12.8398 +trainer/Log Pis Max 69.0184 +trainer/Log Pis Min -9.21445 +trainer/policy/mean Mean -0.0255874 +trainer/policy/mean Std 0.907341 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.8067 +trainer/policy/normal/std Std 0.672093 +trainer/policy/normal/std Max 6.43663 +trainer/policy/normal/std Min 0.274687 +trainer/policy/normal/log_std Mean 0.991293 +trainer/policy/normal/log_std Std 0.321461 +trainer/policy/normal/log_std Max 1.86201 +trainer/policy/normal/log_std Min -1.29212 +eval/num steps total 703081 +eval/num paths total 704 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00596895 +eval/Actions Std 0.902538 +eval/Actions Max 0.999994 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60257 +time/logging (s) 0.00403329 +time/sampling batch (s) 0.526605 +time/saving (s) 0.00367651 +time/training (s) 8.03925 +time/epoch (s) 11.1761 +time/total (s) 7203.82 +Epoch -297 +---------------------------------- --------------- +2022-05-10 15:11:01.702015 PDT | [2] Epoch -296 finished +---------------------------------- --------------- +epoch -296 +replay_buffer/size 999033 +trainer/num train calls 705000 +trainer/Policy Loss -20.1834 +trainer/Log Pis Mean 24.2448 +trainer/Log Pis Std 12.6356 +trainer/Log Pis Max 62.5386 +trainer/Log Pis Min -10.5587 +trainer/policy/mean Mean -0.0287453 +trainer/policy/mean Std 0.910993 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.79029 +trainer/policy/normal/std Std 0.655339 +trainer/policy/normal/std Max 5.63529 +trainer/policy/normal/std Min 0.287695 +trainer/policy/normal/log_std Mean 0.985314 +trainer/policy/normal/log_std Std 0.32496 +trainer/policy/normal/log_std Max 1.72905 +trainer/policy/normal/log_std Min -1.24586 +eval/num steps total 704081 +eval/num paths total 705 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.362288 +eval/Actions Std 0.852612 +eval/Actions Max 0.999993 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55899 +time/logging (s) 0.00375069 +time/sampling batch (s) 1.27872 +time/saving (s) 0.00330976 +time/training (s) 7.54204 +time/epoch (s) 11.3868 +time/total (s) 7215.21 +Epoch -296 +---------------------------------- --------------- +2022-05-10 15:11:12.975160 PDT | [2] Epoch -295 finished +---------------------------------- --------------- +epoch -295 +replay_buffer/size 999033 +trainer/num train calls 706000 +trainer/Policy Loss -18.7832 +trainer/Log Pis Mean 25.0203 +trainer/Log Pis Std 13.4196 +trainer/Log Pis Max 75.4609 +trainer/Log Pis Min -6.83226 +trainer/policy/mean Mean -0.0272598 +trainer/policy/mean Std 0.906705 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.83114 +trainer/policy/normal/std Std 0.659034 +trainer/policy/normal/std Max 6.44543 +trainer/policy/normal/std Min 0.310285 +trainer/policy/normal/log_std Mean 1.00271 +trainer/policy/normal/log_std Std 0.30935 +trainer/policy/normal/log_std Max 1.86337 +trainer/policy/normal/log_std Min -1.17026 +eval/num steps total 705081 +eval/num paths total 706 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.173168 +eval/Actions Std 0.950527 +eval/Actions Max 0.999996 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50958 +time/logging (s) 0.003665 +time/sampling batch (s) 0.530152 +time/saving (s) 0.00340572 +time/training (s) 8.20332 +time/epoch (s) 11.2501 +time/total (s) 7226.46 +Epoch -295 +---------------------------------- --------------- +2022-05-10 15:11:23.400873 PDT | [2] Epoch -294 finished +---------------------------------- --------------- +epoch -294 +replay_buffer/size 999033 +trainer/num train calls 707000 +trainer/Policy Loss -19.1542 +trainer/Log Pis Mean 25.144 +trainer/Log Pis Std 13.8881 +trainer/Log Pis Max 70.598 +trainer/Log Pis Min -6.74347 +trainer/policy/mean Mean -0.0381225 +trainer/policy/mean Std 0.908924 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.75188 +trainer/policy/normal/std Std 0.661978 +trainer/policy/normal/std Max 6.27824 +trainer/policy/normal/std Min 0.259979 +trainer/policy/normal/log_std Mean 0.969235 +trainer/policy/normal/log_std Std 0.334922 +trainer/policy/normal/log_std Max 1.83709 +trainer/policy/normal/log_std Min -1.34716 +eval/num steps total 706081 +eval/num paths total 707 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.115889 +eval/Actions Std 0.876503 +eval/Actions Max 0.999987 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66219 +time/logging (s) 0.00372945 +time/sampling batch (s) 0.527761 +time/saving (s) 0.00338237 +time/training (s) 7.20565 +time/epoch (s) 10.4027 +time/total (s) 7236.87 +Epoch -294 +---------------------------------- --------------- +2022-05-10 15:11:33.097582 PDT | [2] Epoch -293 finished +---------------------------------- --------------- +epoch -293 +replay_buffer/size 999033 +trainer/num train calls 708000 +trainer/Policy Loss -19.9494 +trainer/Log Pis Mean 23.7974 +trainer/Log Pis Std 13.3887 +trainer/Log Pis Max 82.9631 +trainer/Log Pis Min -7.44005 +trainer/policy/mean Mean -0.0559819 +trainer/policy/mean Std 0.907415 +trainer/policy/mean Max 0.999982 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.82375 +trainer/policy/normal/std Std 0.663396 +trainer/policy/normal/std Max 5.59819 +trainer/policy/normal/std Min 0.29758 +trainer/policy/normal/log_std Mean 0.999107 +trainer/policy/normal/log_std Std 0.313902 +trainer/policy/normal/log_std Max 1.72244 +trainer/policy/normal/log_std Min -1.21207 +eval/num steps total 707081 +eval/num paths total 708 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.102968 +eval/Actions Std 0.921414 +eval/Actions Max 0.999994 +eval/Actions Min -0.999982 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.4282 +time/logging (s) 0.00389749 +time/sampling batch (s) 0.276137 +time/saving (s) 0.00374472 +time/training (s) 6.96193 +time/epoch (s) 9.6739 +time/total (s) 7246.54 +Epoch -293 +---------------------------------- --------------- +2022-05-10 15:11:42.839957 PDT | [2] Epoch -292 finished +---------------------------------- --------------- +epoch -292 +replay_buffer/size 999033 +trainer/num train calls 709000 +trainer/Policy Loss -18.9367 +trainer/Log Pis Mean 24.591 +trainer/Log Pis Std 13.6004 +trainer/Log Pis Max 70.7681 +trainer/Log Pis Min -12.1527 +trainer/policy/mean Mean -0.0312935 +trainer/policy/mean Std 0.906267 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.76178 +trainer/policy/normal/std Std 0.664246 +trainer/policy/normal/std Max 6.9851 +trainer/policy/normal/std Min 0.313323 +trainer/policy/normal/log_std Mean 0.97287 +trainer/policy/normal/log_std Std 0.33384 +trainer/policy/normal/log_std Max 1.94378 +trainer/policy/normal/log_std Min -1.16052 +eval/num steps total 708081 +eval/num paths total 709 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.345667 +eval/Actions Std 0.853437 +eval/Actions Max 0.999998 +eval/Actions Min -0.999977 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.76368 +time/logging (s) 0.00370452 +time/sampling batch (s) 0.27895 +time/saving (s) 0.0033359 +time/training (s) 6.66917 +time/epoch (s) 9.71885 +time/total (s) 7256.27 +Epoch -292 +---------------------------------- --------------- +2022-05-10 15:11:52.217832 PDT | [2] Epoch -291 finished +---------------------------------- --------------- +epoch -291 +replay_buffer/size 999033 +trainer/num train calls 710000 +trainer/Policy Loss -19.0893 +trainer/Log Pis Mean 25.245 +trainer/Log Pis Std 14.1053 +trainer/Log Pis Max 72.8208 +trainer/Log Pis Min -11.2845 +trainer/policy/mean Mean -0.0708139 +trainer/policy/mean Std 0.909983 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.79129 +trainer/policy/normal/std Std 0.671132 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.334847 +trainer/policy/normal/log_std Mean 0.984373 +trainer/policy/normal/log_std Std 0.328734 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.09408 +eval/num steps total 709081 +eval/num paths total 710 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.374995 +eval/Actions Std 0.87288 +eval/Actions Max 0.999997 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58028 +time/logging (s) 0.00370285 +time/sampling batch (s) 0.276545 +time/saving (s) 0.00337201 +time/training (s) 6.49106 +time/epoch (s) 9.35495 +time/total (s) 7265.62 +Epoch -291 +---------------------------------- --------------- +2022-05-10 15:12:02.903459 PDT | [2] Epoch -290 finished +---------------------------------- --------------- +epoch -290 +replay_buffer/size 999033 +trainer/num train calls 711000 +trainer/Policy Loss -18.6588 +trainer/Log Pis Mean 25.9838 +trainer/Log Pis Std 12.7042 +trainer/Log Pis Max 68.9193 +trainer/Log Pis Min -3.90056 +trainer/policy/mean Mean -0.0315368 +trainer/policy/mean Std 0.900788 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.79515 +trainer/policy/normal/std Std 0.689209 +trainer/policy/normal/std Max 5.64304 +trainer/policy/normal/std Min 0.260646 +trainer/policy/normal/log_std Mean 0.983556 +trainer/policy/normal/log_std Std 0.337115 +trainer/policy/normal/log_std Max 1.73042 +trainer/policy/normal/log_std Min -1.34459 +eval/num steps total 710081 +eval/num paths total 711 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.387175 +eval/Actions Std 0.878147 +eval/Actions Max 0.999988 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.75682 +time/logging (s) 0.00368011 +time/sampling batch (s) 0.777102 +time/saving (s) 0.00336481 +time/training (s) 7.12163 +time/epoch (s) 10.6626 +time/total (s) 7276.29 +Epoch -290 +---------------------------------- --------------- +2022-05-10 15:12:13.057926 PDT | [2] Epoch -289 finished +---------------------------------- --------------- +epoch -289 +replay_buffer/size 999033 +trainer/num train calls 712000 +trainer/Policy Loss -18.4205 +trainer/Log Pis Mean 24.5111 +trainer/Log Pis Std 13.0395 +trainer/Log Pis Max 63.3072 +trainer/Log Pis Min -8.13248 +trainer/policy/mean Mean -0.0469307 +trainer/policy/mean Std 0.905449 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.82871 +trainer/policy/normal/std Std 0.683708 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.265516 +trainer/policy/normal/log_std Mean 0.997621 +trainer/policy/normal/log_std Std 0.329444 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.32608 +eval/num steps total 711081 +eval/num paths total 712 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.171442 +eval/Actions Std 0.950968 +eval/Actions Max 0.999994 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62048 +time/logging (s) 0.00369099 +time/sampling batch (s) 0.280912 +time/saving (s) 0.00337924 +time/training (s) 7.22298 +time/epoch (s) 10.1314 +time/total (s) 7286.42 +Epoch -289 +---------------------------------- --------------- +2022-05-10 15:12:23.128638 PDT | [2] Epoch -288 finished +---------------------------------- --------------- +epoch -288 +replay_buffer/size 999033 +trainer/num train calls 713000 +trainer/Policy Loss -18.7182 +trainer/Log Pis Mean 23.7923 +trainer/Log Pis Std 12.5889 +trainer/Log Pis Max 70.3882 +trainer/Log Pis Min -8.01935 +trainer/policy/mean Mean -0.0265278 +trainer/policy/mean Std 0.907334 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.81712 +trainer/policy/normal/std Std 0.667251 +trainer/policy/normal/std Max 5.77346 +trainer/policy/normal/std Min 0.28515 +trainer/policy/normal/log_std Mean 0.995721 +trainer/policy/normal/log_std Std 0.31933 +trainer/policy/normal/log_std Max 1.75327 +trainer/policy/normal/log_std Min -1.25474 +eval/num steps total 712081 +eval/num paths total 713 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.249603 +eval/Actions Std 0.883888 +eval/Actions Max 0.999992 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.85222 +time/logging (s) 0.00395914 +time/sampling batch (s) 0.278926 +time/saving (s) 0.00365447 +time/training (s) 6.90918 +time/epoch (s) 10.0479 +time/total (s) 7296.47 +Epoch -288 +---------------------------------- --------------- +2022-05-10 15:12:34.291564 PDT | [2] Epoch -287 finished +---------------------------------- --------------- +epoch -287 +replay_buffer/size 999033 +trainer/num train calls 714000 +trainer/Policy Loss -19.6466 +trainer/Log Pis Mean 24.7755 +trainer/Log Pis Std 12.7731 +trainer/Log Pis Max 63.9651 +trainer/Log Pis Min -6.62362 +trainer/policy/mean Mean -0.0354258 +trainer/policy/mean Std 0.904229 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.77759 +trainer/policy/normal/std Std 0.684064 +trainer/policy/normal/std Max 5.16296 +trainer/policy/normal/std Min 0.265 +trainer/policy/normal/log_std Mean 0.97668 +trainer/policy/normal/log_std Std 0.340405 +trainer/policy/normal/log_std Max 1.64151 +trainer/policy/normal/log_std Min -1.32802 +eval/num steps total 713081 +eval/num paths total 714 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.479139 +eval/Actions Std 0.811922 +eval/Actions Max 0.999996 +eval/Actions Min -0.999985 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57715 +time/logging (s) 0.00377905 +time/sampling batch (s) 0.528265 +time/saving (s) 0.00339248 +time/training (s) 8.02683 +time/epoch (s) 11.1394 +time/total (s) 7307.62 +Epoch -287 +---------------------------------- --------------- +2022-05-10 15:12:44.265868 PDT | [2] Epoch -286 finished +---------------------------------- --------------- +epoch -286 +replay_buffer/size 999033 +trainer/num train calls 715000 +trainer/Policy Loss -19.2865 +trainer/Log Pis Mean 24.6532 +trainer/Log Pis Std 14.1315 +trainer/Log Pis Max 76.9971 +trainer/Log Pis Min -8.93531 +trainer/policy/mean Mean -0.0488372 +trainer/policy/mean Std 0.907623 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.76405 +trainer/policy/normal/std Std 0.668613 +trainer/policy/normal/std Max 6.82009 +trainer/policy/normal/std Min 0.276993 +trainer/policy/normal/log_std Mean 0.973272 +trainer/policy/normal/log_std Std 0.334593 +trainer/policy/normal/log_std Max 1.91987 +trainer/policy/normal/log_std Min -1.28376 +eval/num steps total 714081 +eval/num paths total 715 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0974574 +eval/Actions Std 0.917815 +eval/Actions Max 0.999996 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53787 +time/logging (s) 0.00372428 +time/sampling batch (s) 0.521249 +time/saving (s) 0.00333945 +time/training (s) 6.88561 +time/epoch (s) 9.95178 +time/total (s) 7317.57 +Epoch -286 +---------------------------------- --------------- +2022-05-10 15:12:54.924484 PDT | [2] Epoch -285 finished +---------------------------------- --------------- +epoch -285 +replay_buffer/size 999033 +trainer/num train calls 716000 +trainer/Policy Loss -19.5993 +trainer/Log Pis Mean 24.7032 +trainer/Log Pis Std 13.2009 +trainer/Log Pis Max 66.7583 +trainer/Log Pis Min -5.76765 +trainer/policy/mean Mean -0.0303373 +trainer/policy/mean Std 0.906736 +trainer/policy/mean Max 0.999982 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.84347 +trainer/policy/normal/std Std 0.691941 +trainer/policy/normal/std Max 6.17625 +trainer/policy/normal/std Min 0.261262 +trainer/policy/normal/log_std Mean 1.00123 +trainer/policy/normal/log_std Std 0.337441 +trainer/policy/normal/log_std Max 1.82071 +trainer/policy/normal/log_std Min -1.34223 +eval/num steps total 715081 +eval/num paths total 716 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.361781 +eval/Actions Std 0.789541 +eval/Actions Max 0.999994 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61924 +time/logging (s) 0.00392993 +time/sampling batch (s) 0.271032 +time/saving (s) 0.00335016 +time/training (s) 7.73878 +time/epoch (s) 10.6363 +time/total (s) 7328.21 +Epoch -285 +---------------------------------- --------------- +2022-05-10 15:13:05.888382 PDT | [2] Epoch -284 finished +---------------------------------- --------------- +epoch -284 +replay_buffer/size 999033 +trainer/num train calls 717000 +trainer/Policy Loss -19.9232 +trainer/Log Pis Mean 24.2541 +trainer/Log Pis Std 13.2317 +trainer/Log Pis Max 70.3322 +trainer/Log Pis Min -9.63027 +trainer/policy/mean Mean -0.0289543 +trainer/policy/mean Std 0.908709 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.78565 +trainer/policy/normal/std Std 0.676561 +trainer/policy/normal/std Max 5.80148 +trainer/policy/normal/std Min 0.269306 +trainer/policy/normal/log_std Mean 0.981255 +trainer/policy/normal/log_std Std 0.333232 +trainer/policy/normal/log_std Max 1.75811 +trainer/policy/normal/log_std Min -1.31191 +eval/num steps total 716081 +eval/num paths total 717 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0236246 +eval/Actions Std 0.906796 +eval/Actions Max 0.999995 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70976 +time/logging (s) 0.00374296 +time/sampling batch (s) 0.272245 +time/saving (s) 0.00334223 +time/training (s) 7.95158 +time/epoch (s) 10.9407 +time/total (s) 7339.15 +Epoch -284 +---------------------------------- --------------- +2022-05-10 15:13:15.408344 PDT | [2] Epoch -283 finished +---------------------------------- --------------- +epoch -283 +replay_buffer/size 999033 +trainer/num train calls 718000 +trainer/Policy Loss -19.6945 +trainer/Log Pis Mean 24.0176 +trainer/Log Pis Std 12.8058 +trainer/Log Pis Max 63.1682 +trainer/Log Pis Min -8.80247 +trainer/policy/mean Mean -0.035845 +trainer/policy/mean Std 0.905634 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.78423 +trainer/policy/normal/std Std 0.657514 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.23943 +trainer/policy/normal/log_std Mean 0.983244 +trainer/policy/normal/log_std Std 0.324296 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.42949 +eval/num steps total 717081 +eval/num paths total 718 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00469067 +eval/Actions Std 0.905173 +eval/Actions Max 0.999997 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.23818 +time/logging (s) 0.0037755 +time/sampling batch (s) 0.272552 +time/saving (s) 0.00338121 +time/training (s) 6.97932 +time/epoch (s) 9.4972 +time/total (s) 7348.65 +Epoch -283 +---------------------------------- --------------- +2022-05-10 15:13:25.657315 PDT | [2] Epoch -282 finished +---------------------------------- --------------- +epoch -282 +replay_buffer/size 999033 +trainer/num train calls 719000 +trainer/Policy Loss -20.0135 +trainer/Log Pis Mean 24.7547 +trainer/Log Pis Std 13.5937 +trainer/Log Pis Max 92.5014 +trainer/Log Pis Min -6.52409 +trainer/policy/mean Mean -0.0512842 +trainer/policy/mean Std 0.90657 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.76158 +trainer/policy/normal/std Std 0.674365 +trainer/policy/normal/std Max 6.38511 +trainer/policy/normal/std Min 0.232561 +trainer/policy/normal/log_std Mean 0.970247 +trainer/policy/normal/log_std Std 0.346462 +trainer/policy/normal/log_std Max 1.85397 +trainer/policy/normal/log_std Min -1.4586 +eval/num steps total 718081 +eval/num paths total 719 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.237058 +eval/Actions Std 0.908513 +eval/Actions Max 0.999979 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55146 +time/logging (s) 0.00388478 +time/sampling batch (s) 0.272366 +time/saving (s) 0.00357653 +time/training (s) 7.39516 +time/epoch (s) 10.2264 +time/total (s) 7358.88 +Epoch -282 +---------------------------------- --------------- +2022-05-10 15:13:36.168078 PDT | [2] Epoch -281 finished +---------------------------------- --------------- +epoch -281 +replay_buffer/size 999033 +trainer/num train calls 720000 +trainer/Policy Loss -20.7784 +trainer/Log Pis Mean 24.2214 +trainer/Log Pis Std 12.9784 +trainer/Log Pis Max 65.5401 +trainer/Log Pis Min -8.82686 +trainer/policy/mean Mean -0.0380534 +trainer/policy/mean Std 0.91299 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.81827 +trainer/policy/normal/std Std 0.657425 +trainer/policy/normal/std Max 7.24318 +trainer/policy/normal/std Min 0.331884 +trainer/policy/normal/log_std Mean 0.997687 +trainer/policy/normal/log_std Std 0.312324 +trainer/policy/normal/log_std Max 1.98006 +trainer/policy/normal/log_std Min -1.10297 +eval/num steps total 719081 +eval/num paths total 720 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.291888 +eval/Actions Std 0.715325 +eval/Actions Max 0.999984 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.72119 +time/logging (s) 0.00382628 +time/sampling batch (s) 0.272427 +time/saving (s) 0.00347722 +time/training (s) 7.48698 +time/epoch (s) 10.4879 +time/total (s) 7369.38 +Epoch -281 +---------------------------------- --------------- +2022-05-10 15:13:47.493377 PDT | [2] Epoch -280 finished +---------------------------------- --------------- +epoch -280 +replay_buffer/size 999033 +trainer/num train calls 721000 +trainer/Policy Loss -20.3416 +trainer/Log Pis Mean 24.9577 +trainer/Log Pis Std 13.3441 +trainer/Log Pis Max 64.8675 +trainer/Log Pis Min -8.12572 +trainer/policy/mean Mean -0.0362546 +trainer/policy/mean Std 0.907208 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.80446 +trainer/policy/normal/std Std 0.676208 +trainer/policy/normal/std Max 5.6491 +trainer/policy/normal/std Min 0.265551 +trainer/policy/normal/log_std Mean 0.987286 +trainer/policy/normal/log_std Std 0.339243 +trainer/policy/normal/log_std Max 1.7315 +trainer/policy/normal/log_std Min -1.32595 +eval/num steps total 720081 +eval/num paths total 721 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.179394 +eval/Actions Std 0.85723 +eval/Actions Max 0.999984 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61522 +time/logging (s) 0.0039113 +time/sampling batch (s) 1.27265 +time/saving (s) 0.00341694 +time/training (s) 7.40717 +time/epoch (s) 11.3024 +time/total (s) 7380.68 +Epoch -280 +---------------------------------- --------------- +2022-05-10 15:14:00.138244 PDT | [2] Epoch -279 finished +---------------------------------- --------------- +epoch -279 +replay_buffer/size 999033 +trainer/num train calls 722000 +trainer/Policy Loss -20.0538 +trainer/Log Pis Mean 23.6728 +trainer/Log Pis Std 13.7185 +trainer/Log Pis Max 67.3706 +trainer/Log Pis Min -7.97184 +trainer/policy/mean Mean -0.0496177 +trainer/policy/mean Std 0.908425 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.77895 +trainer/policy/normal/std Std 0.627421 +trainer/policy/normal/std Max 5.70041 +trainer/policy/normal/std Min 0.323522 +trainer/policy/normal/log_std Mean 0.987152 +trainer/policy/normal/log_std Std 0.295128 +trainer/policy/normal/log_std Max 1.74054 +trainer/policy/normal/log_std Min -1.12849 +eval/num steps total 721081 +eval/num paths total 722 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.122684 +eval/Actions Std 0.932724 +eval/Actions Max 0.999995 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.52184 +time/logging (s) 0.00373115 +time/sampling batch (s) 1.02514 +time/saving (s) 0.00331358 +time/training (s) 9.06752 +time/epoch (s) 12.6215 +time/total (s) 7393.31 +Epoch -279 +---------------------------------- --------------- +2022-05-10 15:14:10.191546 PDT | [2] Epoch -278 finished +---------------------------------- --------------- +epoch -278 +replay_buffer/size 999033 +trainer/num train calls 723000 +trainer/Policy Loss -18.4131 +trainer/Log Pis Mean 24.3049 +trainer/Log Pis Std 12.8962 +trainer/Log Pis Max 71.4516 +trainer/Log Pis Min -5.44956 +trainer/policy/mean Mean -0.0442624 +trainer/policy/mean Std 0.903327 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.78756 +trainer/policy/normal/std Std 0.657808 +trainer/policy/normal/std Max 6.38089 +trainer/policy/normal/std Min 0.273631 +trainer/policy/normal/log_std Mean 0.983435 +trainer/policy/normal/log_std Std 0.33087 +trainer/policy/normal/log_std Max 1.85331 +trainer/policy/normal/log_std Min -1.29597 +eval/num steps total 722081 +eval/num paths total 723 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0976352 +eval/Actions Std 0.920361 +eval/Actions Max 0.999986 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70844 +time/logging (s) 0.00377431 +time/sampling batch (s) 0.271204 +time/saving (s) 0.00332612 +time/training (s) 7.04413 +time/epoch (s) 10.0309 +time/total (s) 7403.34 +Epoch -278 +---------------------------------- --------------- +2022-05-10 15:14:19.604595 PDT | [2] Epoch -277 finished +---------------------------------- --------------- +epoch -277 +replay_buffer/size 999033 +trainer/num train calls 724000 +trainer/Policy Loss -19.4362 +trainer/Log Pis Mean 25.1866 +trainer/Log Pis Std 13.0529 +trainer/Log Pis Max 79.785 +trainer/Log Pis Min -4.12081 +trainer/policy/mean Mean -0.0508072 +trainer/policy/mean Std 0.907145 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.83392 +trainer/policy/normal/std Std 0.676849 +trainer/policy/normal/std Max 6.35938 +trainer/policy/normal/std Min 0.270085 +trainer/policy/normal/log_std Mean 1.00049 +trainer/policy/normal/log_std Std 0.325941 +trainer/policy/normal/log_std Max 1.84993 +trainer/policy/normal/log_std Min -1.30902 +eval/num steps total 723081 +eval/num paths total 724 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0623492 +eval/Actions Std 0.891494 +eval/Actions Max 0.999991 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.48578 +time/logging (s) 0.00394348 +time/sampling batch (s) 0.271108 +time/saving (s) 0.00360339 +time/training (s) 6.62627 +time/epoch (s) 9.39071 +time/total (s) 7412.73 +Epoch -277 +---------------------------------- --------------- +2022-05-10 15:14:30.849028 PDT | [2] Epoch -276 finished +---------------------------------- --------------- +epoch -276 +replay_buffer/size 999033 +trainer/num train calls 725000 +trainer/Policy Loss -18.6035 +trainer/Log Pis Mean 23.9986 +trainer/Log Pis Std 13.2048 +trainer/Log Pis Max 71.6979 +trainer/Log Pis Min -7.32727 +trainer/policy/mean Mean -0.0349988 +trainer/policy/mean Std 0.90223 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.80053 +trainer/policy/normal/std Std 0.695961 +trainer/policy/normal/std Max 6.65063 +trainer/policy/normal/std Min 0.237167 +trainer/policy/normal/log_std Mean 0.984393 +trainer/policy/normal/log_std Std 0.342369 +trainer/policy/normal/log_std Max 1.89471 +trainer/policy/normal/log_std Min -1.43899 +eval/num steps total 724081 +eval/num paths total 725 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.108889 +eval/Actions Std 0.911346 +eval/Actions Max 0.999979 +eval/Actions Min -0.999965 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.7608 +time/logging (s) 0.00371444 +time/sampling batch (s) 0.270967 +time/saving (s) 0.00334663 +time/training (s) 8.18261 +time/epoch (s) 11.2214 +time/total (s) 7423.96 +Epoch -276 +---------------------------------- --------------- +2022-05-10 15:14:40.526991 PDT | [2] Epoch -275 finished +---------------------------------- --------------- +epoch -275 +replay_buffer/size 999033 +trainer/num train calls 726000 +trainer/Policy Loss -19.0094 +trainer/Log Pis Mean 24.2992 +trainer/Log Pis Std 13.0149 +trainer/Log Pis Max 67.7889 +trainer/Log Pis Min -12.0215 +trainer/policy/mean Mean -0.0369857 +trainer/policy/mean Std 0.909514 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.81441 +trainer/policy/normal/std Std 0.65696 +trainer/policy/normal/std Max 5.6444 +trainer/policy/normal/std Min 0.29695 +trainer/policy/normal/log_std Mean 0.995633 +trainer/policy/normal/log_std Std 0.3153 +trainer/policy/normal/log_std Max 1.73066 +trainer/policy/normal/log_std Min -1.21419 +eval/num steps total 725081 +eval/num paths total 726 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00989665 +eval/Actions Std 0.910708 +eval/Actions Max 0.999997 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55179 +time/logging (s) 0.00373241 +time/sampling batch (s) 0.270001 +time/saving (s) 0.00334058 +time/training (s) 6.82658 +time/epoch (s) 9.65544 +time/total (s) 7433.62 +Epoch -275 +---------------------------------- --------------- +2022-05-10 15:14:50.846013 PDT | [2] Epoch -274 finished +---------------------------------- --------------- +epoch -274 +replay_buffer/size 999033 +trainer/num train calls 727000 +trainer/Policy Loss -19.638 +trainer/Log Pis Mean 25.1831 +trainer/Log Pis Std 12.9721 +trainer/Log Pis Max 74.0671 +trainer/Log Pis Min -8.70332 +trainer/policy/mean Mean -0.0374236 +trainer/policy/mean Std 0.907808 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.86461 +trainer/policy/normal/std Std 0.693355 +trainer/policy/normal/std Max 6.36837 +trainer/policy/normal/std Min 0.284737 +trainer/policy/normal/log_std Mean 1.01121 +trainer/policy/normal/log_std Std 0.321481 +trainer/policy/normal/log_std Max 1.85134 +trainer/policy/normal/log_std Min -1.25619 +eval/num steps total 726081 +eval/num paths total 727 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.470962 +eval/Actions Std 0.873931 +eval/Actions Max 0.999988 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.49808 +time/logging (s) 0.00374914 +time/sampling batch (s) 0.521947 +time/saving (s) 0.00331504 +time/training (s) 7.26916 +time/epoch (s) 10.2963 +time/total (s) 7443.91 +Epoch -274 +---------------------------------- --------------- +2022-05-10 15:15:01.631758 PDT | [2] Epoch -273 finished +---------------------------------- --------------- +epoch -273 +replay_buffer/size 999033 +trainer/num train calls 728000 +trainer/Policy Loss -20.2486 +trainer/Log Pis Mean 24.2954 +trainer/Log Pis Std 12.9258 +trainer/Log Pis Max 67.6836 +trainer/Log Pis Min -5.98118 +trainer/policy/mean Mean -0.0516665 +trainer/policy/mean Std 0.90669 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.82655 +trainer/policy/normal/std Std 0.679405 +trainer/policy/normal/std Max 6.61779 +trainer/policy/normal/std Min 0.287691 +trainer/policy/normal/log_std Mean 0.997416 +trainer/policy/normal/log_std Std 0.32663 +trainer/policy/normal/log_std Max 1.88976 +trainer/policy/normal/log_std Min -1.24587 +eval/num steps total 727081 +eval/num paths total 728 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.177692 +eval/Actions Std 0.922548 +eval/Actions Max 0.999994 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.46368 +time/logging (s) 0.00395278 +time/sampling batch (s) 0.547965 +time/saving (s) 0.00371759 +time/training (s) 7.74354 +time/epoch (s) 10.7628 +time/total (s) 7454.68 +Epoch -273 +---------------------------------- --------------- +2022-05-10 15:15:12.888272 PDT | [2] Epoch -272 finished +---------------------------------- --------------- +epoch -272 +replay_buffer/size 999033 +trainer/num train calls 729000 +trainer/Policy Loss -18.992 +trainer/Log Pis Mean 24.2747 +trainer/Log Pis Std 13.4992 +trainer/Log Pis Max 68.5721 +trainer/Log Pis Min -7.12517 +trainer/policy/mean Mean -0.025709 +trainer/policy/mean Std 0.913006 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.82958 +trainer/policy/normal/std Std 0.662064 +trainer/policy/normal/std Max 5.9012 +trainer/policy/normal/std Min 0.215426 +trainer/policy/normal/log_std Mean 1.0014 +trainer/policy/normal/log_std Std 0.312681 +trainer/policy/normal/log_std Max 1.77516 +trainer/policy/normal/log_std Min -1.53514 +eval/num steps total 728081 +eval/num paths total 729 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00242156 +eval/Actions Std 0.901598 +eval/Actions Max 0.999994 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.99615 +time/logging (s) 0.00418304 +time/sampling batch (s) 0.324004 +time/saving (s) 0.00395442 +time/training (s) 7.90456 +time/epoch (s) 11.2328 +time/total (s) 7465.92 +Epoch -272 +---------------------------------- --------------- +2022-05-10 15:15:23.181945 PDT | [2] Epoch -271 finished +---------------------------------- --------------- +epoch -271 +replay_buffer/size 999033 +trainer/num train calls 730000 +trainer/Policy Loss -20.0848 +trainer/Log Pis Mean 22.8536 +trainer/Log Pis Std 12.5997 +trainer/Log Pis Max 69.5747 +trainer/Log Pis Min -9.33936 +trainer/policy/mean Mean -0.0160635 +trainer/policy/mean Std 0.904829 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.7915 +trainer/policy/normal/std Std 0.688571 +trainer/policy/normal/std Max 7.23568 +trainer/policy/normal/std Min 0.298083 +trainer/policy/normal/log_std Mean 0.981352 +trainer/policy/normal/log_std Std 0.343093 +trainer/policy/normal/log_std Max 1.97902 +trainer/policy/normal/log_std Min -1.21038 +eval/num steps total 729081 +eval/num paths total 730 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0442815 +eval/Actions Std 0.802364 +eval/Actions Max 0.999994 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.85015 +time/logging (s) 0.00393473 +time/sampling batch (s) 0.322646 +time/saving (s) 0.00356894 +time/training (s) 7.08912 +time/epoch (s) 10.2694 +time/total (s) 7476.19 +Epoch -271 +---------------------------------- --------------- +2022-05-10 15:15:33.557202 PDT | [2] Epoch -270 finished +---------------------------------- --------------- +epoch -270 +replay_buffer/size 999033 +trainer/num train calls 731000 +trainer/Policy Loss -19.2614 +trainer/Log Pis Mean 24.2672 +trainer/Log Pis Std 12.5712 +trainer/Log Pis Max 59.4116 +trainer/Log Pis Min -6.3291 +trainer/policy/mean Mean -0.0399799 +trainer/policy/mean Std 0.910514 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.77316 +trainer/policy/normal/std Std 0.663609 +trainer/policy/normal/std Max 5.70667 +trainer/policy/normal/std Min 0.310708 +trainer/policy/normal/log_std Mean 0.97901 +trainer/policy/normal/log_std Std 0.322132 +trainer/policy/normal/log_std Max 1.74164 +trainer/policy/normal/log_std Min -1.1689 +eval/num steps total 730081 +eval/num paths total 731 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.204361 +eval/Actions Std 0.877346 +eval/Actions Max 0.999997 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.74098 +time/logging (s) 0.00365813 +time/sampling batch (s) 0.533054 +time/saving (s) 0.00334634 +time/training (s) 7.07088 +time/epoch (s) 10.3519 +time/total (s) 7486.54 +Epoch -270 +---------------------------------- --------------- +2022-05-10 15:15:44.279796 PDT | [2] Epoch -269 finished +---------------------------------- --------------- +epoch -269 +replay_buffer/size 999033 +trainer/num train calls 732000 +trainer/Policy Loss -18.6608 +trainer/Log Pis Mean 25.0302 +trainer/Log Pis Std 13.3192 +trainer/Log Pis Max 67.2313 +trainer/Log Pis Min -9.91055 +trainer/policy/mean Mean -0.0264363 +trainer/policy/mean Std 0.911025 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.8335 +trainer/policy/normal/std Std 0.682981 +trainer/policy/normal/std Max 6.60626 +trainer/policy/normal/std Min 0.327808 +trainer/policy/normal/log_std Mean 0.998048 +trainer/policy/normal/log_std Std 0.33689 +trainer/policy/normal/log_std Max 1.88802 +trainer/policy/normal/log_std Min -1.11533 +eval/num steps total 731081 +eval/num paths total 732 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0195518 +eval/Actions Std 0.898436 +eval/Actions Max 0.999988 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60281 +time/logging (s) 0.00432988 +time/sampling batch (s) 0.533086 +time/saving (s) 0.00334781 +time/training (s) 7.55657 +time/epoch (s) 10.7001 +time/total (s) 7497.25 +Epoch -269 +---------------------------------- --------------- +2022-05-10 15:15:53.935181 PDT | [2] Epoch -268 finished +---------------------------------- --------------- +epoch -268 +replay_buffer/size 999033 +trainer/num train calls 733000 +trainer/Policy Loss -19.7553 +trainer/Log Pis Mean 23.7021 +trainer/Log Pis Std 13.1885 +trainer/Log Pis Max 72.3568 +trainer/Log Pis Min -9.48614 +trainer/policy/mean Mean -0.0411323 +trainer/policy/mean Std 0.906922 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.78521 +trainer/policy/normal/std Std 0.654818 +trainer/policy/normal/std Max 6.2684 +trainer/policy/normal/std Min 0.283363 +trainer/policy/normal/log_std Mean 0.985539 +trainer/policy/normal/log_std Std 0.312085 +trainer/policy/normal/log_std Max 1.83552 +trainer/policy/normal/log_std Min -1.26103 +eval/num steps total 732081 +eval/num paths total 733 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0714081 +eval/Actions Std 0.910475 +eval/Actions Max 0.999992 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.46153 +time/logging (s) 0.00373392 +time/sampling batch (s) 0.280419 +time/saving (s) 0.00330974 +time/training (s) 6.88203 +time/epoch (s) 9.63103 +time/total (s) 7506.88 +Epoch -268 +---------------------------------- --------------- +2022-05-10 15:16:04.280581 PDT | [2] Epoch -267 finished +---------------------------------- --------------- +epoch -267 +replay_buffer/size 999033 +trainer/num train calls 734000 +trainer/Policy Loss -20.3903 +trainer/Log Pis Mean 25.4652 +trainer/Log Pis Std 13.7249 +trainer/Log Pis Max 69.9346 +trainer/Log Pis Min -8.35937 +trainer/policy/mean Mean -0.0460738 +trainer/policy/mean Std 0.909404 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.91146 +trainer/policy/normal/std Std 0.667764 +trainer/policy/normal/std Max 5.66544 +trainer/policy/normal/std Min 0.273667 +trainer/policy/normal/log_std Mean 1.03074 +trainer/policy/normal/log_std Std 0.311793 +trainer/policy/normal/log_std Max 1.73438 +trainer/policy/normal/log_std Min -1.29584 +eval/num steps total 733081 +eval/num paths total 734 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.209395 +eval/Actions Std 0.877887 +eval/Actions Max 1 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.40344 +time/logging (s) 0.00370303 +time/sampling batch (s) 0.521465 +time/saving (s) 0.00331415 +time/training (s) 7.39071 +time/epoch (s) 10.3226 +time/total (s) 7517.21 +Epoch -267 +---------------------------------- --------------- +2022-05-10 15:16:13.695359 PDT | [2] Epoch -266 finished +---------------------------------- --------------- +epoch -266 +replay_buffer/size 999033 +trainer/num train calls 735000 +trainer/Policy Loss -19.9378 +trainer/Log Pis Mean 25.6568 +trainer/Log Pis Std 13.5564 +trainer/Log Pis Max 67.1692 +trainer/Log Pis Min -5.74585 +trainer/policy/mean Mean -0.0522607 +trainer/policy/mean Std 0.91131 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.82799 +trainer/policy/normal/std Std 0.67374 +trainer/policy/normal/std Max 5.74683 +trainer/policy/normal/std Min 0.291815 +trainer/policy/normal/log_std Mean 0.997117 +trainer/policy/normal/log_std Std 0.332401 +trainer/policy/normal/log_std Max 1.74865 +trainer/policy/normal/log_std Min -1.23163 +eval/num steps total 734081 +eval/num paths total 735 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.124642 +eval/Actions Std 0.890698 +eval/Actions Max 0.999985 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.47967 +time/logging (s) 0.0036919 +time/sampling batch (s) 0.271847 +time/saving (s) 0.00330952 +time/training (s) 6.63363 +time/epoch (s) 9.39214 +time/total (s) 7526.6 +Epoch -266 +---------------------------------- --------------- +2022-05-10 15:16:24.814217 PDT | [2] Epoch -265 finished +---------------------------------- --------------- +epoch -265 +replay_buffer/size 999033 +trainer/num train calls 736000 +trainer/Policy Loss -19.9436 +trainer/Log Pis Mean 23.8402 +trainer/Log Pis Std 13.1462 +trainer/Log Pis Max 72.94 +trainer/Log Pis Min -5.00639 +trainer/policy/mean Mean -0.028457 +trainer/policy/mean Std 0.908369 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.84066 +trainer/policy/normal/std Std 0.681885 +trainer/policy/normal/std Max 6.09678 +trainer/policy/normal/std Min 0.307185 +trainer/policy/normal/log_std Mean 1.00366 +trainer/policy/normal/log_std Std 0.319013 +trainer/policy/normal/log_std Max 1.80776 +trainer/policy/normal/log_std Min -1.18031 +eval/num steps total 735081 +eval/num paths total 736 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.248525 +eval/Actions Std 0.7488 +eval/Actions Max 0.999998 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62349 +time/logging (s) 0.00374049 +time/sampling batch (s) 0.772049 +time/saving (s) 0.00335773 +time/training (s) 7.69373 +time/epoch (s) 11.0964 +time/total (s) 7537.7 +Epoch -265 +---------------------------------- --------------- +2022-05-10 15:16:35.346073 PDT | [2] Epoch -264 finished +---------------------------------- --------------- +epoch -264 +replay_buffer/size 999033 +trainer/num train calls 737000 +trainer/Policy Loss -19.9898 +trainer/Log Pis Mean 23.7473 +trainer/Log Pis Std 13.049 +trainer/Log Pis Max 63.0922 +trainer/Log Pis Min -7.99682 +trainer/policy/mean Mean -0.0354208 +trainer/policy/mean Std 0.911175 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.99997 +trainer/policy/normal/std Mean 2.76993 +trainer/policy/normal/std Std 0.670436 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.230464 +trainer/policy/normal/log_std Mean 0.974744 +trainer/policy/normal/log_std Std 0.34077 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.46766 +eval/num steps total 736081 +eval/num paths total 737 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.41298 +eval/Actions Std 0.873309 +eval/Actions Max 0.999991 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68924 +time/logging (s) 0.00400157 +time/sampling batch (s) 0.270712 +time/saving (s) 0.00360852 +time/training (s) 7.54198 +time/epoch (s) 10.5095 +time/total (s) 7548.21 +Epoch -264 +---------------------------------- --------------- +2022-05-10 15:16:45.778151 PDT | [2] Epoch -263 finished +---------------------------------- --------------- +epoch -263 +replay_buffer/size 999033 +trainer/num train calls 738000 +trainer/Policy Loss -19.6164 +trainer/Log Pis Mean 23.9941 +trainer/Log Pis Std 13.1811 +trainer/Log Pis Max 66.6029 +trainer/Log Pis Min -10.4411 +trainer/policy/mean Mean -0.0402725 +trainer/policy/mean Std 0.902871 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.83513 +trainer/policy/normal/std Std 0.644016 +trainer/policy/normal/std Max 5.41502 +trainer/policy/normal/std Min 0.351007 +trainer/policy/normal/log_std Mean 1.00545 +trainer/policy/normal/log_std Std 0.304125 +trainer/policy/normal/log_std Max 1.68918 +trainer/policy/normal/log_std Min -1.04695 +eval/num steps total 737081 +eval/num paths total 738 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0121384 +eval/Actions Std 0.906756 +eval/Actions Max 1 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.38598 +time/logging (s) 0.00377591 +time/sampling batch (s) 0.771369 +time/saving (s) 0.00346266 +time/training (s) 7.24394 +time/epoch (s) 10.4085 +time/total (s) 7558.63 +Epoch -263 +---------------------------------- --------------- +2022-05-10 15:16:55.551939 PDT | [2] Epoch -262 finished +---------------------------------- --------------- +epoch -262 +replay_buffer/size 999033 +trainer/num train calls 739000 +trainer/Policy Loss -20.9811 +trainer/Log Pis Mean 24.113 +trainer/Log Pis Std 13.4631 +trainer/Log Pis Max 71.7301 +trainer/Log Pis Min -9.6258 +trainer/policy/mean Mean -0.0467167 +trainer/policy/mean Std 0.906044 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.77896 +trainer/policy/normal/std Std 0.630498 +trainer/policy/normal/std Max 5.63665 +trainer/policy/normal/std Min 0.313629 +trainer/policy/normal/log_std Mean 0.986178 +trainer/policy/normal/log_std Std 0.299767 +trainer/policy/normal/log_std Max 1.72929 +trainer/policy/normal/log_std Min -1.15954 +eval/num steps total 738081 +eval/num paths total 739 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0189388 +eval/Actions Std 0.968606 +eval/Actions Max 0.999996 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.36628 +time/logging (s) 0.00374725 +time/sampling batch (s) 0.521462 +time/saving (s) 0.00332197 +time/training (s) 6.85614 +time/epoch (s) 9.75094 +time/total (s) 7568.38 +Epoch -262 +---------------------------------- --------------- +2022-05-10 15:17:05.622090 PDT | [2] Epoch -261 finished +---------------------------------- --------------- +epoch -261 +replay_buffer/size 999033 +trainer/num train calls 740000 +trainer/Policy Loss -19.2999 +trainer/Log Pis Mean 24.3106 +trainer/Log Pis Std 13.0421 +trainer/Log Pis Max 70.4119 +trainer/Log Pis Min -8.03014 +trainer/policy/mean Mean -0.0531188 +trainer/policy/mean Std 0.907529 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.77406 +trainer/policy/normal/std Std 0.642402 +trainer/policy/normal/std Max 6.34307 +trainer/policy/normal/std Min 0.286198 +trainer/policy/normal/log_std Mean 0.981565 +trainer/policy/normal/log_std Std 0.314962 +trainer/policy/normal/log_std Max 1.84736 +trainer/policy/normal/log_std Min -1.25107 +eval/num steps total 739081 +eval/num paths total 740 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0627932 +eval/Actions Std 0.905903 +eval/Actions Max 0.999995 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64041 +time/logging (s) 0.00378805 +time/sampling batch (s) 0.523218 +time/saving (s) 0.00331094 +time/training (s) 6.8766 +time/epoch (s) 10.0473 +time/total (s) 7578.43 +Epoch -261 +---------------------------------- --------------- +2022-05-10 15:17:14.758260 PDT | [2] Epoch -260 finished +---------------------------------- --------------- +epoch -260 +replay_buffer/size 999033 +trainer/num train calls 741000 +trainer/Policy Loss -19.681 +trainer/Log Pis Mean 25.4927 +trainer/Log Pis Std 13.8547 +trainer/Log Pis Max 91.4489 +trainer/Log Pis Min -8.39606 +trainer/policy/mean Mean -0.0460891 +trainer/policy/mean Std 0.904105 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.82487 +trainer/policy/normal/std Std 0.659149 +trainer/policy/normal/std Max 6.54867 +trainer/policy/normal/std Min 0.305808 +trainer/policy/normal/log_std Mean 0.999539 +trainer/policy/normal/log_std Std 0.315494 +trainer/policy/normal/log_std Max 1.87926 +trainer/policy/normal/log_std Min -1.1848 +eval/num steps total 740081 +eval/num paths total 741 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.111171 +eval/Actions Std 0.911881 +eval/Actions Max 0.999991 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.4087 +time/logging (s) 0.00374365 +time/sampling batch (s) 0.270687 +time/saving (s) 0.00337889 +time/training (s) 6.42693 +time/epoch (s) 9.11344 +time/total (s) 7587.55 +Epoch -260 +---------------------------------- --------------- +2022-05-10 15:17:25.145385 PDT | [2] Epoch -259 finished +---------------------------------- --------------- +epoch -259 +replay_buffer/size 999033 +trainer/num train calls 742000 +trainer/Policy Loss -19.8977 +trainer/Log Pis Mean 24.4894 +trainer/Log Pis Std 13.5834 +trainer/Log Pis Max 71.7888 +trainer/Log Pis Min -8.47188 +trainer/policy/mean Mean -0.028688 +trainer/policy/mean Std 0.905542 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.8315 +trainer/policy/normal/std Std 0.689025 +trainer/policy/normal/std Max 6.00552 +trainer/policy/normal/std Min 0.285371 +trainer/policy/normal/log_std Mean 0.997601 +trainer/policy/normal/log_std Std 0.334133 +trainer/policy/normal/log_std Max 1.79268 +trainer/policy/normal/log_std Min -1.25396 +eval/num steps total 741081 +eval/num paths total 742 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0189071 +eval/Actions Std 0.90478 +eval/Actions Max 0.999993 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.72309 +time/logging (s) 0.00379606 +time/sampling batch (s) 0.521536 +time/saving (s) 0.0033797 +time/training (s) 7.11253 +time/epoch (s) 10.3643 +time/total (s) 7597.91 +Epoch -259 +---------------------------------- --------------- +2022-05-10 15:17:35.504789 PDT | [2] Epoch -258 finished +---------------------------------- --------------- +epoch -258 +replay_buffer/size 999033 +trainer/num train calls 743000 +trainer/Policy Loss -19.4979 +trainer/Log Pis Mean 23.1403 +trainer/Log Pis Std 13.2827 +trainer/Log Pis Max 66.9831 +trainer/Log Pis Min -7.91661 +trainer/policy/mean Mean -0.0305513 +trainer/policy/mean Std 0.904146 +trainer/policy/mean Max 0.999981 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.76656 +trainer/policy/normal/std Std 0.672554 +trainer/policy/normal/std Max 5.97583 +trainer/policy/normal/std Min 0.242755 +trainer/policy/normal/log_std Mean 0.97368 +trainer/policy/normal/log_std Std 0.339123 +trainer/policy/normal/log_std Max 1.78772 +trainer/policy/normal/log_std Min -1.4157 +eval/num steps total 742081 +eval/num paths total 743 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0671561 +eval/Actions Std 0.91477 +eval/Actions Max 0.999993 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.45119 +time/logging (s) 0.00373707 +time/sampling batch (s) 0.522674 +time/saving (s) 0.00334196 +time/training (s) 7.35549 +time/epoch (s) 10.3364 +time/total (s) 7608.25 +Epoch -258 +---------------------------------- --------------- +2022-05-10 15:17:44.826748 PDT | [2] Epoch -257 finished +---------------------------------- --------------- +epoch -257 +replay_buffer/size 999033 +trainer/num train calls 744000 +trainer/Policy Loss -19.1395 +trainer/Log Pis Mean 24.399 +trainer/Log Pis Std 12.7916 +trainer/Log Pis Max 70.0237 +trainer/Log Pis Min -8.79752 +trainer/policy/mean Mean -0.0153829 +trainer/policy/mean Std 0.909771 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.76228 +trainer/policy/normal/std Std 0.652367 +trainer/policy/normal/std Max 5.22768 +trainer/policy/normal/std Min 0.315763 +trainer/policy/normal/log_std Mean 0.975045 +trainer/policy/normal/log_std Std 0.324997 +trainer/policy/normal/log_std Max 1.65397 +trainer/policy/normal/log_std Min -1.15276 +eval/num steps total 743081 +eval/num paths total 744 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.136839 +eval/Actions Std 0.895325 +eval/Actions Max 0.999984 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70647 +time/logging (s) 0.00369199 +time/sampling batch (s) 0.271398 +time/saving (s) 0.00332971 +time/training (s) 6.31419 +time/epoch (s) 9.29908 +time/total (s) 7617.55 +Epoch -257 +---------------------------------- --------------- +2022-05-10 15:17:56.377076 PDT | [2] Epoch -256 finished +---------------------------------- --------------- +epoch -256 +replay_buffer/size 999033 +trainer/num train calls 745000 +trainer/Policy Loss -20.3282 +trainer/Log Pis Mean 24.6692 +trainer/Log Pis Std 13.3416 +trainer/Log Pis Max 74.0864 +trainer/Log Pis Min -10.583 +trainer/policy/mean Mean -0.0449211 +trainer/policy/mean Std 0.907204 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.83786 +trainer/policy/normal/std Std 0.684206 +trainer/policy/normal/std Max 6.94991 +trainer/policy/normal/std Min 0.268372 +trainer/policy/normal/log_std Mean 1.00183 +trainer/policy/normal/log_std Std 0.324123 +trainer/policy/normal/log_std Max 1.93873 +trainer/policy/normal/log_std Min -1.31538 +eval/num steps total 744081 +eval/num paths total 745 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0386931 +eval/Actions Std 0.897134 +eval/Actions Max 0.999999 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50177 +time/logging (s) 0.00385173 +time/sampling batch (s) 0.773225 +time/saving (s) 0.00341455 +time/training (s) 8.24539 +time/epoch (s) 11.5277 +time/total (s) 7629.08 +Epoch -256 +---------------------------------- --------------- +2022-05-10 15:18:07.872463 PDT | [2] Epoch -255 finished +---------------------------------- --------------- +epoch -255 +replay_buffer/size 999033 +trainer/num train calls 746000 +trainer/Policy Loss -20.1871 +trainer/Log Pis Mean 25.5119 +trainer/Log Pis Std 13.6586 +trainer/Log Pis Max 81.1904 +trainer/Log Pis Min -10.7441 +trainer/policy/mean Mean -0.0359369 +trainer/policy/mean Std 0.910121 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.83634 +trainer/policy/normal/std Std 0.670876 +trainer/policy/normal/std Max 6.93872 +trainer/policy/normal/std Min 0.223779 +trainer/policy/normal/log_std Mean 1.0018 +trainer/policy/normal/log_std Std 0.326131 +trainer/policy/normal/log_std Max 1.93712 +trainer/policy/normal/log_std Min -1.4971 +eval/num steps total 745081 +eval/num paths total 746 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.152444 +eval/Actions Std 0.867815 +eval/Actions Max 0.999987 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59409 +time/logging (s) 0.00456045 +time/sampling batch (s) 0.77913 +time/saving (s) 0.00335065 +time/training (s) 8.09171 +time/epoch (s) 11.4728 +time/total (s) 7640.56 +Epoch -255 +---------------------------------- --------------- +2022-05-10 15:18:18.378870 PDT | [2] Epoch -254 finished +---------------------------------- --------------- +epoch -254 +replay_buffer/size 999033 +trainer/num train calls 747000 +trainer/Policy Loss -18.7069 +trainer/Log Pis Mean 23.5937 +trainer/Log Pis Std 13.1222 +trainer/Log Pis Max 63.5325 +trainer/Log Pis Min -8.30307 +trainer/policy/mean Mean -0.0388617 +trainer/policy/mean Std 0.908228 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.74957 +trainer/policy/normal/std Std 0.642817 +trainer/policy/normal/std Max 7.27483 +trainer/policy/normal/std Min 0.267283 +trainer/policy/normal/log_std Mean 0.972234 +trainer/policy/normal/log_std Std 0.315799 +trainer/policy/normal/log_std Max 1.98442 +trainer/policy/normal/log_std Min -1.31945 +eval/num steps total 746081 +eval/num paths total 747 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0430159 +eval/Actions Std 0.91219 +eval/Actions Max 0.999998 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50183 +time/logging (s) 0.0039647 +time/sampling batch (s) 0.276896 +time/saving (s) 0.00362307 +time/training (s) 7.6955 +time/epoch (s) 10.4818 +time/total (s) 7651.05 +Epoch -254 +---------------------------------- --------------- +2022-05-10 15:18:28.320659 PDT | [2] Epoch -253 finished +---------------------------------- --------------- +epoch -253 +replay_buffer/size 999033 +trainer/num train calls 748000 +trainer/Policy Loss -19.3177 +trainer/Log Pis Mean 26.0099 +trainer/Log Pis Std 12.9493 +trainer/Log Pis Max 77.5951 +trainer/Log Pis Min -2.63832 +trainer/policy/mean Mean -0.0356812 +trainer/policy/mean Std 0.90499 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.79551 +trainer/policy/normal/std Std 0.656445 +trainer/policy/normal/std Max 5.68789 +trainer/policy/normal/std Min 0.251184 +trainer/policy/normal/log_std Mean 0.986734 +trainer/policy/normal/log_std Std 0.328521 +trainer/policy/normal/log_std Max 1.73834 +trainer/policy/normal/log_std Min -1.38157 +eval/num steps total 747081 +eval/num paths total 748 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.114465 +eval/Actions Std 0.913655 +eval/Actions Max 0.999997 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56347 +time/logging (s) 0.00379531 +time/sampling batch (s) 0.276266 +time/saving (s) 0.00349185 +time/training (s) 7.071 +time/epoch (s) 9.91802 +time/total (s) 7660.97 +Epoch -253 +---------------------------------- --------------- +2022-05-10 15:18:39.169496 PDT | [2] Epoch -252 finished +---------------------------------- --------------- +epoch -252 +replay_buffer/size 999033 +trainer/num train calls 749000 +trainer/Policy Loss -19.938 +trainer/Log Pis Mean 25.2735 +trainer/Log Pis Std 13.488 +trainer/Log Pis Max 72.6033 +trainer/Log Pis Min -14.5217 +trainer/policy/mean Mean -0.0303687 +trainer/policy/mean Std 0.909776 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.81313 +trainer/policy/normal/std Std 0.664056 +trainer/policy/normal/std Max 5.6708 +trainer/policy/normal/std Min 0.299895 +trainer/policy/normal/log_std Mean 0.993102 +trainer/policy/normal/log_std Std 0.326551 +trainer/policy/normal/log_std Max 1.73533 +trainer/policy/normal/log_std Min -1.20432 +eval/num steps total 748081 +eval/num paths total 749 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.198443 +eval/Actions Std 0.889163 +eval/Actions Max 0.999999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59365 +time/logging (s) 0.00369571 +time/sampling batch (s) 0.277368 +time/saving (s) 0.00335077 +time/training (s) 7.9472 +time/epoch (s) 10.8253 +time/total (s) 7671.79 +Epoch -252 +---------------------------------- --------------- +2022-05-10 15:18:48.887099 PDT | [2] Epoch -251 finished +---------------------------------- --------------- +epoch -251 +replay_buffer/size 999033 +trainer/num train calls 750000 +trainer/Policy Loss -18.7716 +trainer/Log Pis Mean 23.4774 +trainer/Log Pis Std 13.0314 +trainer/Log Pis Max 67.5213 +trainer/Log Pis Min -6.49058 +trainer/policy/mean Mean -0.0501274 +trainer/policy/mean Std 0.905578 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.80367 +trainer/policy/normal/std Std 0.666359 +trainer/policy/normal/std Max 6.26894 +trainer/policy/normal/std Min 0.302005 +trainer/policy/normal/log_std Mean 0.991295 +trainer/policy/normal/log_std Std 0.314921 +trainer/policy/normal/log_std Max 1.83561 +trainer/policy/normal/log_std Min -1.19731 +eval/num steps total 749081 +eval/num paths total 750 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.263147 +eval/Actions Std 0.853367 +eval/Actions Max 0.999981 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67489 +time/logging (s) 0.00377094 +time/sampling batch (s) 0.27665 +time/saving (s) 0.00334134 +time/training (s) 6.73583 +time/epoch (s) 9.69448 +time/total (s) 7681.49 +Epoch -251 +---------------------------------- --------------- +2022-05-10 15:18:59.113715 PDT | [2] Epoch -250 finished +---------------------------------- --------------- +epoch -250 +replay_buffer/size 999033 +trainer/num train calls 751000 +trainer/Policy Loss -19.4082 +trainer/Log Pis Mean 24.3045 +trainer/Log Pis Std 13.2339 +trainer/Log Pis Max 67.5762 +trainer/Log Pis Min -11.5901 +trainer/policy/mean Mean -0.0364815 +trainer/policy/mean Std 0.904539 +trainer/policy/mean Max 0.99998 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.78995 +trainer/policy/normal/std Std 0.685661 +trainer/policy/normal/std Max 5.60174 +trainer/policy/normal/std Min 0.24215 +trainer/policy/normal/log_std Mean 0.979772 +trainer/policy/normal/log_std Std 0.350585 +trainer/policy/normal/log_std Max 1.72308 +trainer/policy/normal/log_std Min -1.4182 +eval/num steps total 750081 +eval/num paths total 751 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.105849 +eval/Actions Std 0.91004 +eval/Actions Max 0.999995 +eval/Actions Min -0.999984 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60529 +time/logging (s) 0.00380216 +time/sampling batch (s) 0.277001 +time/saving (s) 0.00337555 +time/training (s) 7.31386 +time/epoch (s) 10.2033 +time/total (s) 7691.7 +Epoch -250 +---------------------------------- --------------- +2022-05-10 15:19:09.783872 PDT | [2] Epoch -249 finished +---------------------------------- --------------- +epoch -249 +replay_buffer/size 999033 +trainer/num train calls 752000 +trainer/Policy Loss -19.4617 +trainer/Log Pis Mean 24.5796 +trainer/Log Pis Std 13.2074 +trainer/Log Pis Max 64.9596 +trainer/Log Pis Min -6.10819 +trainer/policy/mean Mean -0.0269773 +trainer/policy/mean Std 0.907798 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.78639 +trainer/policy/normal/std Std 0.652735 +trainer/policy/normal/std Max 5.95649 +trainer/policy/normal/std Min 0.217749 +trainer/policy/normal/log_std Mean 0.98465 +trainer/policy/normal/log_std Std 0.323085 +trainer/policy/normal/log_std Max 1.78448 +trainer/policy/normal/log_std Min -1.52441 +eval/num steps total 751081 +eval/num paths total 752 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0934075 +eval/Actions Std 0.915797 +eval/Actions Max 0.999996 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.74184 +time/logging (s) 0.00374418 +time/sampling batch (s) 0.274408 +time/saving (s) 0.00334345 +time/training (s) 7.62369 +time/epoch (s) 10.647 +time/total (s) 7702.35 +Epoch -249 +---------------------------------- --------------- +2022-05-10 15:19:20.106514 PDT | [2] Epoch -248 finished +---------------------------------- --------------- +epoch -248 +replay_buffer/size 999033 +trainer/num train calls 753000 +trainer/Policy Loss -20.295 +trainer/Log Pis Mean 24.2791 +trainer/Log Pis Std 13.0344 +trainer/Log Pis Max 66.4573 +trainer/Log Pis Min -4.63255 +trainer/policy/mean Mean -0.0256951 +trainer/policy/mean Std 0.90633 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.76477 +trainer/policy/normal/std Std 0.668003 +trainer/policy/normal/std Max 5.9603 +trainer/policy/normal/std Min 0.322278 +trainer/policy/normal/log_std Mean 0.975286 +trainer/policy/normal/log_std Std 0.324829 +trainer/policy/normal/log_std Max 1.78512 +trainer/policy/normal/log_std Min -1.13234 +eval/num steps total 752081 +eval/num paths total 753 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0647251 +eval/Actions Std 0.916417 +eval/Actions Max 0.99999 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.27123 +time/logging (s) 0.00371417 +time/sampling batch (s) 0.773116 +time/saving (s) 0.00333751 +time/training (s) 7.24824 +time/epoch (s) 10.2996 +time/total (s) 7712.65 +Epoch -248 +---------------------------------- --------------- +2022-05-10 15:19:30.184079 PDT | [2] Epoch -247 finished +---------------------------------- --------------- +epoch -247 +replay_buffer/size 999033 +trainer/num train calls 754000 +trainer/Policy Loss -18.8542 +trainer/Log Pis Mean 24.7985 +trainer/Log Pis Std 13.6723 +trainer/Log Pis Max 65.4284 +trainer/Log Pis Min -8.4642 +trainer/policy/mean Mean -0.0182365 +trainer/policy/mean Std 0.906509 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.80714 +trainer/policy/normal/std Std 0.681118 +trainer/policy/normal/std Max 6.22008 +trainer/policy/normal/std Min 0.237421 +trainer/policy/normal/log_std Mean 0.989506 +trainer/policy/normal/log_std Std 0.331378 +trainer/policy/normal/log_std Max 1.82778 +trainer/policy/normal/log_std Min -1.43792 +eval/num steps total 753081 +eval/num paths total 754 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0823362 +eval/Actions Std 0.912543 +eval/Actions Max 0.999993 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63933 +time/logging (s) 0.00373403 +time/sampling batch (s) 0.272219 +time/saving (s) 0.00347102 +time/training (s) 7.13578 +time/epoch (s) 10.0545 +time/total (s) 7722.71 +Epoch -247 +---------------------------------- --------------- +2022-05-10 15:19:41.014220 PDT | [2] Epoch -246 finished +---------------------------------- --------------- +epoch -246 +replay_buffer/size 999033 +trainer/num train calls 755000 +trainer/Policy Loss -18.8852 +trainer/Log Pis Mean 24.8916 +trainer/Log Pis Std 13.0004 +trainer/Log Pis Max 75.8097 +trainer/Log Pis Min -5.22782 +trainer/policy/mean Mean -0.025346 +trainer/policy/mean Std 0.907279 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.79509 +trainer/policy/normal/std Std 0.63784 +trainer/policy/normal/std Max 7.10761 +trainer/policy/normal/std Min 0.29273 +trainer/policy/normal/log_std Mean 0.992445 +trainer/policy/normal/log_std Std 0.295843 +trainer/policy/normal/log_std Max 1.96117 +trainer/policy/normal/log_std Min -1.2285 +eval/num steps total 754081 +eval/num paths total 755 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0763575 +eval/Actions Std 0.838974 +eval/Actions Max 0.999946 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.75514 +time/logging (s) 0.00377838 +time/sampling batch (s) 0.272785 +time/saving (s) 0.00357775 +time/training (s) 7.7719 +time/epoch (s) 10.8072 +time/total (s) 7733.52 +Epoch -246 +---------------------------------- --------------- +2022-05-10 15:19:51.811736 PDT | [2] Epoch -245 finished +---------------------------------- --------------- +epoch -245 +replay_buffer/size 999033 +trainer/num train calls 756000 +trainer/Policy Loss -20.3006 +trainer/Log Pis Mean 25.421 +trainer/Log Pis Std 13.0867 +trainer/Log Pis Max 69.9058 +trainer/Log Pis Min -3.0722 +trainer/policy/mean Mean -0.0393104 +trainer/policy/mean Std 0.907243 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.76502 +trainer/policy/normal/std Std 0.671411 +trainer/policy/normal/std Max 5.84892 +trainer/policy/normal/std Min 0.245675 +trainer/policy/normal/log_std Mean 0.973239 +trainer/policy/normal/log_std Std 0.336772 +trainer/policy/normal/log_std Max 1.76626 +trainer/policy/normal/log_std Min -1.40375 +eval/num steps total 755081 +eval/num paths total 756 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.318486 +eval/Actions Std 0.868703 +eval/Actions Max 0.999968 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58222 +time/logging (s) 0.003771 +time/sampling batch (s) 0.275382 +time/saving (s) 0.00336858 +time/training (s) 7.90924 +time/epoch (s) 10.774 +time/total (s) 7744.3 +Epoch -245 +---------------------------------- --------------- +2022-05-10 15:20:02.042315 PDT | [2] Epoch -244 finished +---------------------------------- --------------- +epoch -244 +replay_buffer/size 999033 +trainer/num train calls 757000 +trainer/Policy Loss -21.0747 +trainer/Log Pis Mean 25.3392 +trainer/Log Pis Std 13.3696 +trainer/Log Pis Max 69.1327 +trainer/Log Pis Min -7.5045 +trainer/policy/mean Mean -0.0323051 +trainer/policy/mean Std 0.908947 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.79822 +trainer/policy/normal/std Std 0.673806 +trainer/policy/normal/std Max 6.308 +trainer/policy/normal/std Min 0.203064 +trainer/policy/normal/log_std Mean 0.985857 +trainer/policy/normal/log_std Std 0.335127 +trainer/policy/normal/log_std Max 1.84182 +trainer/policy/normal/log_std Min -1.59423 +eval/num steps total 756081 +eval/num paths total 757 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00947954 +eval/Actions Std 0.904851 +eval/Actions Max 0.999997 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.82162 +time/logging (s) 0.00372124 +time/sampling batch (s) 0.27506 +time/saving (s) 0.00350201 +time/training (s) 7.1034 +time/epoch (s) 10.2073 +time/total (s) 7754.51 +Epoch -244 +---------------------------------- --------------- +2022-05-10 15:20:11.261791 PDT | [2] Epoch -243 finished +---------------------------------- --------------- +epoch -243 +replay_buffer/size 999033 +trainer/num train calls 758000 +trainer/Policy Loss -18.871 +trainer/Log Pis Mean 24.9629 +trainer/Log Pis Std 13.0805 +trainer/Log Pis Max 65.0198 +trainer/Log Pis Min -9.61134 +trainer/policy/mean Mean -0.0296243 +trainer/policy/mean Std 0.905535 +trainer/policy/mean Max 0.999976 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.80724 +trainer/policy/normal/std Std 0.651454 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.295817 +trainer/policy/normal/log_std Mean 0.994779 +trainer/policy/normal/log_std Std 0.307486 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.21801 +eval/num steps total 757081 +eval/num paths total 758 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.259958 +eval/Actions Std 0.897813 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5442 +time/logging (s) 0.00369269 +time/sampling batch (s) 0.278543 +time/saving (s) 0.00335263 +time/training (s) 6.3661 +time/epoch (s) 9.19588 +time/total (s) 7763.7 +Epoch -243 +---------------------------------- --------------- +2022-05-10 15:20:22.021962 PDT | [2] Epoch -242 finished +---------------------------------- --------------- +epoch -242 +replay_buffer/size 999033 +trainer/num train calls 759000 +trainer/Policy Loss -19.5767 +trainer/Log Pis Mean 24.4878 +trainer/Log Pis Std 13.0765 +trainer/Log Pis Max 78.8704 +trainer/Log Pis Min -9.57641 +trainer/policy/mean Mean -0.0537036 +trainer/policy/mean Std 0.903315 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.81758 +trainer/policy/normal/std Std 0.689251 +trainer/policy/normal/std Max 6.04541 +trainer/policy/normal/std Min 0.249034 +trainer/policy/normal/log_std Mean 0.990414 +trainer/policy/normal/log_std Std 0.347689 +trainer/policy/normal/log_std Max 1.7993 +trainer/policy/normal/log_std Min -1.39017 +eval/num steps total 758081 +eval/num paths total 759 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.232945 +eval/Actions Std 0.914105 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6404 +time/logging (s) 0.00372589 +time/sampling batch (s) 0.278382 +time/saving (s) 0.00337702 +time/training (s) 7.81082 +time/epoch (s) 10.7367 +time/total (s) 7774.44 +Epoch -242 +---------------------------------- --------------- +2022-05-10 15:20:31.569636 PDT | [2] Epoch -241 finished +---------------------------------- --------------- +epoch -241 +replay_buffer/size 999033 +trainer/num train calls 760000 +trainer/Policy Loss -19.3671 +trainer/Log Pis Mean 24.4317 +trainer/Log Pis Std 13.734 +trainer/Log Pis Max 67.7465 +trainer/Log Pis Min -4.88636 +trainer/policy/mean Mean -0.0292832 +trainer/policy/mean Std 0.908606 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.77017 +trainer/policy/normal/std Std 0.655912 +trainer/policy/normal/std Max 6.96991 +trainer/policy/normal/std Min 0.302205 +trainer/policy/normal/log_std Mean 0.978877 +trainer/policy/normal/log_std Std 0.319315 +trainer/policy/normal/log_std Max 1.9416 +trainer/policy/normal/log_std Min -1.19665 +eval/num steps total 759081 +eval/num paths total 760 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.119797 +eval/Actions Std 0.89608 +eval/Actions Max 0.999991 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73421 +time/logging (s) 0.0036816 +time/sampling batch (s) 0.275967 +time/saving (s) 0.00342868 +time/training (s) 6.50695 +time/epoch (s) 9.52423 +time/total (s) 7783.97 +Epoch -241 +---------------------------------- --------------- +2022-05-10 15:20:41.476163 PDT | [2] Epoch -240 finished +---------------------------------- --------------- +epoch -240 +replay_buffer/size 999033 +trainer/num train calls 761000 +trainer/Policy Loss -19.3775 +trainer/Log Pis Mean 24.2571 +trainer/Log Pis Std 13.2742 +trainer/Log Pis Max 70.8714 +trainer/Log Pis Min -6.00662 +trainer/policy/mean Mean -0.0221303 +trainer/policy/mean Std 0.906409 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.75453 +trainer/policy/normal/std Std 0.681038 +trainer/policy/normal/std Max 7.17163 +trainer/policy/normal/std Min 0.266517 +trainer/policy/normal/log_std Mean 0.968734 +trainer/policy/normal/log_std Std 0.338941 +trainer/policy/normal/log_std Max 1.97013 +trainer/policy/normal/log_std Min -1.32232 +eval/num steps total 760081 +eval/num paths total 761 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0963655 +eval/Actions Std 0.902485 +eval/Actions Max 0.999996 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.76602 +time/logging (s) 0.00368016 +time/sampling batch (s) 0.527491 +time/saving (s) 0.00339129 +time/training (s) 6.58245 +time/epoch (s) 9.88304 +time/total (s) 7793.86 +Epoch -240 +---------------------------------- --------------- +2022-05-10 15:20:51.799516 PDT | [2] Epoch -239 finished +---------------------------------- --------------- +epoch -239 +replay_buffer/size 999033 +trainer/num train calls 762000 +trainer/Policy Loss -20.4671 +trainer/Log Pis Mean 23.9377 +trainer/Log Pis Std 13.3398 +trainer/Log Pis Max 65.9663 +trainer/Log Pis Min -10.5634 +trainer/policy/mean Mean -0.0471047 +trainer/policy/mean Std 0.90637 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.84204 +trainer/policy/normal/std Std 0.690572 +trainer/policy/normal/std Max 6.08236 +trainer/policy/normal/std Min 0.252235 +trainer/policy/normal/log_std Mean 1.0014 +trainer/policy/normal/log_std Std 0.334066 +trainer/policy/normal/log_std Max 1.80539 +trainer/policy/normal/log_std Min -1.37739 +eval/num steps total 761081 +eval/num paths total 762 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.114213 +eval/Actions Std 0.915899 +eval/Actions Max 0.999991 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.78956 +time/logging (s) 0.00369333 +time/sampling batch (s) 0.527931 +time/saving (s) 0.00334539 +time/training (s) 6.97526 +time/epoch (s) 10.2998 +time/total (s) 7804.16 +Epoch -239 +---------------------------------- --------------- +2022-05-10 15:21:02.740558 PDT | [2] Epoch -238 finished +---------------------------------- --------------- +epoch -238 +replay_buffer/size 999033 +trainer/num train calls 763000 +trainer/Policy Loss -19.6854 +trainer/Log Pis Mean 25.193 +trainer/Log Pis Std 14.1826 +trainer/Log Pis Max 75.0239 +trainer/Log Pis Min -8.16068 +trainer/policy/mean Mean -0.0455757 +trainer/policy/mean Std 0.907177 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.79765 +trainer/policy/normal/std Std 0.686269 +trainer/policy/normal/std Max 5.54203 +trainer/policy/normal/std Min 0.196119 +trainer/policy/normal/log_std Mean 0.983617 +trainer/policy/normal/log_std Std 0.344573 +trainer/policy/normal/log_std Max 1.71236 +trainer/policy/normal/log_std Min -1.62904 +eval/num steps total 762081 +eval/num paths total 763 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0491835 +eval/Actions Std 0.928558 +eval/Actions Max 0.999998 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65103 +time/logging (s) 0.00369801 +time/sampling batch (s) 0.527919 +time/saving (s) 0.0033188 +time/training (s) 7.73136 +time/epoch (s) 10.9173 +time/total (s) 7815.08 +Epoch -238 +---------------------------------- --------------- +2022-05-10 15:21:13.538825 PDT | [2] Epoch -237 finished +---------------------------------- --------------- +epoch -237 +replay_buffer/size 999033 +trainer/num train calls 764000 +trainer/Policy Loss -19.1342 +trainer/Log Pis Mean 24.7027 +trainer/Log Pis Std 14.3475 +trainer/Log Pis Max 79.2644 +trainer/Log Pis Min -6.8823 +trainer/policy/mean Mean -0.0325489 +trainer/policy/mean Std 0.907276 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.86865 +trainer/policy/normal/std Std 0.696536 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.313991 +trainer/policy/normal/log_std Mean 1.01157 +trainer/policy/normal/log_std Std 0.329275 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.15839 +eval/num steps total 763081 +eval/num paths total 764 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.161012 +eval/Actions Std 0.914266 +eval/Actions Max 0.999995 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64616 +time/logging (s) 0.00396026 +time/sampling batch (s) 0.5323 +time/saving (s) 0.0036079 +time/training (s) 7.33809 +time/epoch (s) 10.5241 +time/total (s) 7825.86 +Epoch -237 +---------------------------------- --------------- +2022-05-10 15:21:23.652993 PDT | [2] Epoch -236 finished +---------------------------------- --------------- +epoch -236 +replay_buffer/size 999033 +trainer/num train calls 765000 +trainer/Policy Loss -21.3156 +trainer/Log Pis Mean 24.7835 +trainer/Log Pis Std 13.5095 +trainer/Log Pis Max 82.6413 +trainer/Log Pis Min -4.1414 +trainer/policy/mean Mean -0.0418154 +trainer/policy/mean Std 0.909071 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.81425 +trainer/policy/normal/std Std 0.662832 +trainer/policy/normal/std Max 6.39775 +trainer/policy/normal/std Min 0.263529 +trainer/policy/normal/log_std Mean 0.99356 +trainer/policy/normal/log_std Std 0.327649 +trainer/policy/normal/log_std Max 1.85595 +trainer/policy/normal/log_std Min -1.33359 +eval/num steps total 764081 +eval/num paths total 765 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0596274 +eval/Actions Std 0.925114 +eval/Actions Max 0.999997 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66144 +time/logging (s) 0.00379161 +time/sampling batch (s) 0.276173 +time/saving (s) 0.00345143 +time/training (s) 7.14532 +time/epoch (s) 10.0902 +time/total (s) 7835.95 +Epoch -236 +---------------------------------- --------------- +2022-05-10 15:21:34.072450 PDT | [2] Epoch -235 finished +---------------------------------- --------------- +epoch -235 +replay_buffer/size 999033 +trainer/num train calls 766000 +trainer/Policy Loss -19.9801 +trainer/Log Pis Mean 24.394 +trainer/Log Pis Std 13.3859 +trainer/Log Pis Max 63.6617 +trainer/Log Pis Min -6.96943 +trainer/policy/mean Mean -0.0456685 +trainer/policy/mean Std 0.908389 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.82202 +trainer/policy/normal/std Std 0.670418 +trainer/policy/normal/std Max 5.97142 +trainer/policy/normal/std Min 0.259301 +trainer/policy/normal/log_std Mean 0.996575 +trainer/policy/normal/log_std Std 0.324 +trainer/policy/normal/log_std Max 1.78698 +trainer/policy/normal/log_std Min -1.34977 +eval/num steps total 765081 +eval/num paths total 766 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.119234 +eval/Actions Std 0.91721 +eval/Actions Max 0.999994 +eval/Actions Min -0.999988 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.46612 +time/logging (s) 0.00369912 +time/sampling batch (s) 0.27961 +time/saving (s) 0.00332378 +time/training (s) 7.64289 +time/epoch (s) 10.3956 +time/total (s) 7846.35 +Epoch -235 +---------------------------------- --------------- +2022-05-10 15:21:43.806544 PDT | [2] Epoch -234 finished +---------------------------------- --------------- +epoch -234 +replay_buffer/size 999033 +trainer/num train calls 767000 +trainer/Policy Loss -19.1813 +trainer/Log Pis Mean 25.071 +trainer/Log Pis Std 14.4902 +trainer/Log Pis Max 68.706 +trainer/Log Pis Min -7.3336 +trainer/policy/mean Mean -0.0331338 +trainer/policy/mean Std 0.905723 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.83431 +trainer/policy/normal/std Std 0.689352 +trainer/policy/normal/std Max 6.95567 +trainer/policy/normal/std Min 0.310912 +trainer/policy/normal/log_std Mean 0.998 +trainer/policy/normal/log_std Std 0.336491 +trainer/policy/normal/log_std Max 1.93956 +trainer/policy/normal/log_std Min -1.16824 +eval/num steps total 766081 +eval/num paths total 767 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.132748 +eval/Actions Std 0.872945 +eval/Actions Max 0.999995 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56731 +time/logging (s) 0.0036623 +time/sampling batch (s) 0.274249 +time/saving (s) 0.00337862 +time/training (s) 6.86227 +time/epoch (s) 9.71086 +time/total (s) 7856.06 +Epoch -234 +---------------------------------- --------------- +2022-05-10 15:21:55.155136 PDT | [2] Epoch -233 finished +---------------------------------- --------------- +epoch -233 +replay_buffer/size 999033 +trainer/num train calls 768000 +trainer/Policy Loss -19.5686 +trainer/Log Pis Mean 24.6609 +trainer/Log Pis Std 13.4553 +trainer/Log Pis Max 76.6665 +trainer/Log Pis Min -4.11195 +trainer/policy/mean Mean -0.0289694 +trainer/policy/mean Std 0.90712 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.79906 +trainer/policy/normal/std Std 0.660429 +trainer/policy/normal/std Max 5.29697 +trainer/policy/normal/std Min 0.295937 +trainer/policy/normal/log_std Mean 0.989327 +trainer/policy/normal/log_std Std 0.319444 +trainer/policy/normal/log_std Max 1.66714 +trainer/policy/normal/log_std Min -1.21761 +eval/num steps total 767081 +eval/num paths total 768 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.511489 +eval/Actions Std 0.701593 +eval/Actions Max 0.999997 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.52359 +time/logging (s) 0.00383733 +time/sampling batch (s) 0.278604 +time/saving (s) 0.00356377 +time/training (s) 8.51564 +time/epoch (s) 11.3252 +time/total (s) 7867.39 +Epoch -233 +---------------------------------- --------------- +2022-05-10 15:22:05.147363 PDT | [2] Epoch -232 finished +---------------------------------- --------------- +epoch -232 +replay_buffer/size 999033 +trainer/num train calls 769000 +trainer/Policy Loss -18.9865 +trainer/Log Pis Mean 23.4106 +trainer/Log Pis Std 13.0904 +trainer/Log Pis Max 63.7367 +trainer/Log Pis Min -10.2493 +trainer/policy/mean Mean -0.0168158 +trainer/policy/mean Std 0.903733 +trainer/policy/mean Max 0.999967 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.80278 +trainer/policy/normal/std Std 0.638291 +trainer/policy/normal/std Max 6.03065 +trainer/policy/normal/std Min 0.302349 +trainer/policy/normal/log_std Mean 0.992961 +trainer/policy/normal/log_std Std 0.311465 +trainer/policy/normal/log_std Max 1.79685 +trainer/policy/normal/log_std Min -1.19617 +eval/num steps total 768081 +eval/num paths total 769 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0759309 +eval/Actions Std 0.912303 +eval/Actions Max 0.999997 +eval/Actions Min -0.999988 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54331 +time/logging (s) 0.0037435 +time/sampling batch (s) 0.52948 +time/saving (s) 0.00341679 +time/training (s) 6.88735 +time/epoch (s) 9.9673 +time/total (s) 7877.36 +Epoch -232 +---------------------------------- --------------- +2022-05-10 15:22:15.541448 PDT | [2] Epoch -231 finished +---------------------------------- --------------- +epoch -231 +replay_buffer/size 999033 +trainer/num train calls 770000 +trainer/Policy Loss -21.0574 +trainer/Log Pis Mean 24.5533 +trainer/Log Pis Std 13.0157 +trainer/Log Pis Max 71.28 +trainer/Log Pis Min -8.1206 +trainer/policy/mean Mean -0.0314178 +trainer/policy/mean Std 0.910938 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.79634 +trainer/policy/normal/std Std 0.675722 +trainer/policy/normal/std Max 5.71722 +trainer/policy/normal/std Min 0.299604 +trainer/policy/normal/log_std Mean 0.986129 +trainer/policy/normal/log_std Std 0.328258 +trainer/policy/normal/log_std Max 1.74348 +trainer/policy/normal/log_std Min -1.20529 +eval/num steps total 769081 +eval/num paths total 770 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.11323 +eval/Actions Std 0.887417 +eval/Actions Max 0.999983 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59202 +time/logging (s) 0.00372947 +time/sampling batch (s) 0.278979 +time/saving (s) 0.0033602 +time/training (s) 7.49192 +time/epoch (s) 10.37 +time/total (s) 7887.74 +Epoch -231 +---------------------------------- --------------- +2022-05-10 15:22:25.938285 PDT | [2] Epoch -230 finished +---------------------------------- --------------- +epoch -230 +replay_buffer/size 999033 +trainer/num train calls 771000 +trainer/Policy Loss -19.3242 +trainer/Log Pis Mean 23.9538 +trainer/Log Pis Std 13.2546 +trainer/Log Pis Max 64.8445 +trainer/Log Pis Min -11.8667 +trainer/policy/mean Mean -0.038733 +trainer/policy/mean Std 0.9037 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999979 +trainer/policy/normal/std Mean 2.81961 +trainer/policy/normal/std Std 0.682237 +trainer/policy/normal/std Max 5.84547 +trainer/policy/normal/std Min 0.286583 +trainer/policy/normal/log_std Mean 0.994114 +trainer/policy/normal/log_std Std 0.329808 +trainer/policy/normal/log_std Max 1.76567 +trainer/policy/normal/log_std Min -1.24973 +eval/num steps total 770081 +eval/num paths total 771 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.358155 +eval/Actions Std 0.833697 +eval/Actions Max 0.999987 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56838 +time/logging (s) 0.00415559 +time/sampling batch (s) 0.529137 +time/saving (s) 0.00361979 +time/training (s) 7.26813 +time/epoch (s) 10.3734 +time/total (s) 7898.11 +Epoch -230 +---------------------------------- --------------- +2022-05-10 15:22:36.383521 PDT | [2] Epoch -229 finished +---------------------------------- --------------- +epoch -229 +replay_buffer/size 999033 +trainer/num train calls 772000 +trainer/Policy Loss -20.0663 +trainer/Log Pis Mean 24.0965 +trainer/Log Pis Std 12.8894 +trainer/Log Pis Max 70.7539 +trainer/Log Pis Min -8.85703 +trainer/policy/mean Mean -0.0235581 +trainer/policy/mean Std 0.910687 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.78961 +trainer/policy/normal/std Std 0.656685 +trainer/policy/normal/std Max 6.09203 +trainer/policy/normal/std Min 0.26336 +trainer/policy/normal/log_std Mean 0.985095 +trainer/policy/normal/log_std Std 0.324381 +trainer/policy/normal/log_std Max 1.80698 +trainer/policy/normal/log_std Min -1.33423 +eval/num steps total 771081 +eval/num paths total 772 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.200587 +eval/Actions Std 0.935387 +eval/Actions Max 0.999981 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54936 +time/logging (s) 0.00374901 +time/sampling batch (s) 0.530776 +time/saving (s) 0.00344485 +time/training (s) 7.33289 +time/epoch (s) 10.4202 +time/total (s) 7908.54 +Epoch -229 +---------------------------------- --------------- +2022-05-10 15:22:46.349569 PDT | [2] Epoch -228 finished +---------------------------------- --------------- +epoch -228 +replay_buffer/size 999033 +trainer/num train calls 773000 +trainer/Policy Loss -19.4061 +trainer/Log Pis Mean 24.5159 +trainer/Log Pis Std 12.9326 +trainer/Log Pis Max 68.6655 +trainer/Log Pis Min -6.48701 +trainer/policy/mean Mean -0.0710225 +trainer/policy/mean Std 0.902802 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.8565 +trainer/policy/normal/std Std 0.663086 +trainer/policy/normal/std Max 5.35765 +trainer/policy/normal/std Min 0.290437 +trainer/policy/normal/log_std Mean 1.01104 +trainer/policy/normal/log_std Std 0.312204 +trainer/policy/normal/log_std Max 1.67852 +trainer/policy/normal/log_std Min -1.23637 +eval/num steps total 772081 +eval/num paths total 773 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0383954 +eval/Actions Std 0.911928 +eval/Actions Max 0.999998 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55253 +time/logging (s) 0.0041852 +time/sampling batch (s) 0.278304 +time/saving (s) 0.00346809 +time/training (s) 7.10465 +time/epoch (s) 9.94314 +time/total (s) 7918.48 +Epoch -228 +---------------------------------- --------------- +2022-05-10 15:22:57.123069 PDT | [2] Epoch -227 finished +---------------------------------- --------------- +epoch -227 +replay_buffer/size 999033 +trainer/num train calls 774000 +trainer/Policy Loss -19.7374 +trainer/Log Pis Mean 24.3461 +trainer/Log Pis Std 13.1795 +trainer/Log Pis Max 73.0209 +trainer/Log Pis Min -6.57993 +trainer/policy/mean Mean -0.0478074 +trainer/policy/mean Std 0.905932 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.77345 +trainer/policy/normal/std Std 0.636814 +trainer/policy/normal/std Max 5.23886 +trainer/policy/normal/std Min 0.247282 +trainer/policy/normal/log_std Mean 0.981212 +trainer/policy/normal/log_std Std 0.317075 +trainer/policy/normal/log_std Max 1.6561 +trainer/policy/normal/log_std Min -1.39723 +eval/num steps total 773081 +eval/num paths total 774 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.195606 +eval/Actions Std 0.90616 +eval/Actions Max 0.999989 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53848 +time/logging (s) 0.00378651 +time/sampling batch (s) 0.527255 +time/saving (s) 0.00339745 +time/training (s) 7.67653 +time/epoch (s) 10.7494 +time/total (s) 7929.24 +Epoch -227 +---------------------------------- --------------- +2022-05-10 15:23:07.574631 PDT | [2] Epoch -226 finished +---------------------------------- --------------- +epoch -226 +replay_buffer/size 999033 +trainer/num train calls 775000 +trainer/Policy Loss -19.2388 +trainer/Log Pis Mean 24.7858 +trainer/Log Pis Std 13.0713 +trainer/Log Pis Max 72.5896 +trainer/Log Pis Min -6.77878 +trainer/policy/mean Mean -0.0379892 +trainer/policy/mean Std 0.907636 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.78136 +trainer/policy/normal/std Std 0.635921 +trainer/policy/normal/std Max 5.06298 +trainer/policy/normal/std Min 0.315515 +trainer/policy/normal/log_std Mean 0.985007 +trainer/policy/normal/log_std Std 0.311117 +trainer/policy/normal/log_std Max 1.62196 +trainer/policy/normal/log_std Min -1.15355 +eval/num steps total 774081 +eval/num paths total 775 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0320753 +eval/Actions Std 0.904255 +eval/Actions Max 0.999994 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.48312 +time/logging (s) 0.00370997 +time/sampling batch (s) 0.289554 +time/saving (s) 0.00335563 +time/training (s) 7.64823 +time/epoch (s) 10.428 +time/total (s) 7939.67 +Epoch -226 +---------------------------------- --------------- +2022-05-10 15:23:18.319114 PDT | [2] Epoch -225 finished +---------------------------------- --------------- +epoch -225 +replay_buffer/size 999033 +trainer/num train calls 776000 +trainer/Policy Loss -19.6979 +trainer/Log Pis Mean 23.2168 +trainer/Log Pis Std 13.6625 +trainer/Log Pis Max 69.9432 +trainer/Log Pis Min -10.6414 +trainer/policy/mean Mean -0.0183975 +trainer/policy/mean Std 0.9043 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.75687 +trainer/policy/normal/std Std 0.685928 +trainer/policy/normal/std Max 5.67934 +trainer/policy/normal/std Min 0.280571 +trainer/policy/normal/log_std Mean 0.968534 +trainer/policy/normal/log_std Std 0.342502 +trainer/policy/normal/log_std Max 1.73683 +trainer/policy/normal/log_std Min -1.27093 +eval/num steps total 775081 +eval/num paths total 776 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.107797 +eval/Actions Std 0.921548 +eval/Actions Max 0.999998 +eval/Actions Min -0.999987 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54483 +time/logging (s) 0.00373318 +time/sampling batch (s) 0.277206 +time/saving (s) 0.00335464 +time/training (s) 7.89196 +time/epoch (s) 10.7211 +time/total (s) 7950.39 +Epoch -225 +---------------------------------- --------------- +2022-05-10 15:23:29.430870 PDT | [2] Epoch -224 finished +---------------------------------- --------------- +epoch -224 +replay_buffer/size 999033 +trainer/num train calls 777000 +trainer/Policy Loss -19.506 +trainer/Log Pis Mean 26.2789 +trainer/Log Pis Std 13.735 +trainer/Log Pis Max 87.3643 +trainer/Log Pis Min -8.6219 +trainer/policy/mean Mean -0.0403783 +trainer/policy/mean Std 0.91118 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.8044 +trainer/policy/normal/std Std 0.686001 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.272022 +trainer/policy/normal/log_std Mean 0.987342 +trainer/policy/normal/log_std Std 0.336598 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.30187 +eval/num steps total 776081 +eval/num paths total 777 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.110376 +eval/Actions Std 0.846353 +eval/Actions Max 0.999989 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65428 +time/logging (s) 0.00372496 +time/sampling batch (s) 0.527526 +time/saving (s) 0.00336279 +time/training (s) 7.89932 +time/epoch (s) 11.0882 +time/total (s) 7961.48 +Epoch -224 +---------------------------------- --------------- +2022-05-10 15:23:38.987351 PDT | [2] Epoch -223 finished +---------------------------------- --------------- +epoch -223 +replay_buffer/size 999033 +trainer/num train calls 778000 +trainer/Policy Loss -19.197 +trainer/Log Pis Mean 24.3641 +trainer/Log Pis Std 13.6765 +trainer/Log Pis Max 71.3898 +trainer/Log Pis Min -5.70184 +trainer/policy/mean Mean -0.0274866 +trainer/policy/mean Std 0.904579 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.80678 +trainer/policy/normal/std Std 0.65551 +trainer/policy/normal/std Max 6.10944 +trainer/policy/normal/std Min 0.254281 +trainer/policy/normal/log_std Mean 0.993608 +trainer/policy/normal/log_std Std 0.31152 +trainer/policy/normal/log_std Max 1.80984 +trainer/policy/normal/log_std Min -1.36932 +eval/num steps total 777081 +eval/num paths total 778 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0394394 +eval/Actions Std 0.911723 +eval/Actions Max 0.999994 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.74833 +time/logging (s) 0.00366054 +time/sampling batch (s) 0.277957 +time/saving (s) 0.00338501 +time/training (s) 6.49919 +time/epoch (s) 9.53252 +time/total (s) 7971.02 +Epoch -223 +---------------------------------- --------------- +2022-05-10 15:23:49.473910 PDT | [2] Epoch -222 finished +---------------------------------- --------------- +epoch -222 +replay_buffer/size 999033 +trainer/num train calls 779000 +trainer/Policy Loss -20.6277 +trainer/Log Pis Mean 24.7382 +trainer/Log Pis Std 13.2547 +trainer/Log Pis Max 66.9716 +trainer/Log Pis Min -12.4203 +trainer/policy/mean Mean -0.0290287 +trainer/policy/mean Std 0.909272 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.79962 +trainer/policy/normal/std Std 0.687636 +trainer/policy/normal/std Max 6.12961 +trainer/policy/normal/std Min 0.287799 +trainer/policy/normal/log_std Mean 0.984043 +trainer/policy/normal/log_std Std 0.345404 +trainer/policy/normal/log_std Max 1.81313 +trainer/policy/normal/log_std Min -1.24549 +eval/num steps total 778081 +eval/num paths total 779 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.115373 +eval/Actions Std 0.913223 +eval/Actions Max 0.999992 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.46039 +time/logging (s) 0.00373593 +time/sampling batch (s) 0.278281 +time/saving (s) 0.00345596 +time/training (s) 7.71705 +time/epoch (s) 10.4629 +time/total (s) 7981.48 +Epoch -222 +---------------------------------- --------------- +2022-05-10 15:24:00.219097 PDT | [2] Epoch -221 finished +---------------------------------- --------------- +epoch -221 +replay_buffer/size 999033 +trainer/num train calls 780000 +trainer/Policy Loss -19.9784 +trainer/Log Pis Mean 25.3068 +trainer/Log Pis Std 13.7595 +trainer/Log Pis Max 71.835 +trainer/Log Pis Min -8.1026 +trainer/policy/mean Mean -0.0478132 +trainer/policy/mean Std 0.908453 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.8202 +trainer/policy/normal/std Std 0.661726 +trainer/policy/normal/std Max 6.04643 +trainer/policy/normal/std Min 0.299537 +trainer/policy/normal/log_std Mean 0.997245 +trainer/policy/normal/log_std Std 0.31756 +trainer/policy/normal/log_std Max 1.79947 +trainer/policy/normal/log_std Min -1.20552 +eval/num steps total 779081 +eval/num paths total 780 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00751744 +eval/Actions Std 0.969477 +eval/Actions Max 0.999991 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.41252 +time/logging (s) 0.00395446 +time/sampling batch (s) 0.280919 +time/saving (s) 0.00365159 +time/training (s) 8.02065 +time/epoch (s) 10.7217 +time/total (s) 7992.21 +Epoch -221 +---------------------------------- --------------- +2022-05-10 15:24:10.500876 PDT | [2] Epoch -220 finished +---------------------------------- --------------- +epoch -220 +replay_buffer/size 999033 +trainer/num train calls 781000 +trainer/Policy Loss -19.2817 +trainer/Log Pis Mean 23.6116 +trainer/Log Pis Std 13.7355 +trainer/Log Pis Max 69.4408 +trainer/Log Pis Min -6.3647 +trainer/policy/mean Mean -0.0451681 +trainer/policy/mean Std 0.90803 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.76369 +trainer/policy/normal/std Std 0.676169 +trainer/policy/normal/std Max 5.83245 +trainer/policy/normal/std Min 0.271105 +trainer/policy/normal/log_std Mean 0.973037 +trainer/policy/normal/log_std Std 0.333392 +trainer/policy/normal/log_std Max 1.76344 +trainer/policy/normal/log_std Min -1.30525 +eval/num steps total 780081 +eval/num paths total 781 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0854439 +eval/Actions Std 0.870226 +eval/Actions Max 0.999995 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61229 +time/logging (s) 0.00376584 +time/sampling batch (s) 0.278459 +time/saving (s) 0.00341005 +time/training (s) 7.35979 +time/epoch (s) 10.2577 +time/total (s) 8002.47 +Epoch -220 +---------------------------------- --------------- +2022-05-10 15:24:20.466050 PDT | [2] Epoch -219 finished +---------------------------------- --------------- +epoch -219 +replay_buffer/size 999033 +trainer/num train calls 782000 +trainer/Policy Loss -19.7569 +trainer/Log Pis Mean 24.5663 +trainer/Log Pis Std 14.1167 +trainer/Log Pis Max 76.7383 +trainer/Log Pis Min -7.05362 +trainer/policy/mean Mean -0.0295129 +trainer/policy/mean Std 0.902543 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.8868 +trainer/policy/normal/std Std 0.711101 +trainer/policy/normal/std Max 6.54803 +trainer/policy/normal/std Min 0.172707 +trainer/policy/normal/log_std Mean 1.01511 +trainer/policy/normal/log_std Std 0.343607 +trainer/policy/normal/log_std Max 1.87916 +trainer/policy/normal/log_std Min -1.75616 +eval/num steps total 781081 +eval/num paths total 782 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0452147 +eval/Actions Std 0.886881 +eval/Actions Max 0.999977 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70919 +time/logging (s) 0.003686 +time/sampling batch (s) 0.278039 +time/saving (s) 0.00333712 +time/training (s) 6.94729 +time/epoch (s) 9.94154 +time/total (s) 8012.41 +Epoch -219 +---------------------------------- --------------- +2022-05-10 15:24:30.793034 PDT | [2] Epoch -218 finished +---------------------------------- --------------- +epoch -218 +replay_buffer/size 999033 +trainer/num train calls 783000 +trainer/Policy Loss -19.8702 +trainer/Log Pis Mean 25.0988 +trainer/Log Pis Std 13.0706 +trainer/Log Pis Max 65.7411 +trainer/Log Pis Min -9.47707 +trainer/policy/mean Mean -0.0290347 +trainer/policy/mean Std 0.910192 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.78565 +trainer/policy/normal/std Std 0.692924 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.260475 +trainer/policy/normal/log_std Mean 0.978457 +trainer/policy/normal/log_std Std 0.347478 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.34525 +eval/num steps total 782081 +eval/num paths total 783 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.113099 +eval/Actions Std 0.88564 +eval/Actions Max 0.999991 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.682 +time/logging (s) 0.0037386 +time/sampling batch (s) 0.278147 +time/saving (s) 0.00336807 +time/training (s) 7.33624 +time/epoch (s) 10.3035 +time/total (s) 8022.72 +Epoch -218 +---------------------------------- --------------- +2022-05-10 15:24:40.481718 PDT | [2] Epoch -217 finished +---------------------------------- --------------- +epoch -217 +replay_buffer/size 999033 +trainer/num train calls 784000 +trainer/Policy Loss -19.7815 +trainer/Log Pis Mean 24.603 +trainer/Log Pis Std 13.3555 +trainer/Log Pis Max 71.0372 +trainer/Log Pis Min -5.43779 +trainer/policy/mean Mean -0.0391358 +trainer/policy/mean Std 0.907848 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.7802 +trainer/policy/normal/std Std 0.673362 +trainer/policy/normal/std Max 6.47309 +trainer/policy/normal/std Min 0.22271 +trainer/policy/normal/log_std Mean 0.980532 +trainer/policy/normal/log_std Std 0.326729 +trainer/policy/normal/log_std Max 1.86765 +trainer/policy/normal/log_std Min -1.50189 +eval/num steps total 783081 +eval/num paths total 784 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.194785 +eval/Actions Std 0.901199 +eval/Actions Max 0.99999 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64401 +time/logging (s) 0.0037477 +time/sampling batch (s) 0.279335 +time/saving (s) 0.00338946 +time/training (s) 6.73455 +time/epoch (s) 9.66503 +time/total (s) 8032.39 +Epoch -217 +---------------------------------- --------------- +2022-05-10 15:24:51.365730 PDT | [2] Epoch -216 finished +---------------------------------- --------------- +epoch -216 +replay_buffer/size 999033 +trainer/num train calls 785000 +trainer/Policy Loss -19.988 +trainer/Log Pis Mean 25.5065 +trainer/Log Pis Std 13.2411 +trainer/Log Pis Max 63.9941 +trainer/Log Pis Min -7.7394 +trainer/policy/mean Mean -0.0164645 +trainer/policy/mean Std 0.904925 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.769 +trainer/policy/normal/std Std 0.652147 +trainer/policy/normal/std Max 5.428 +trainer/policy/normal/std Min 0.284863 +trainer/policy/normal/log_std Mean 0.978118 +trainer/policy/normal/log_std Std 0.321758 +trainer/policy/normal/log_std Max 1.69157 +trainer/policy/normal/log_std Min -1.25575 +eval/num steps total 784081 +eval/num paths total 785 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.502351 +eval/Actions Std 0.808832 +eval/Actions Max 0.999985 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.72872 +time/logging (s) 0.00378886 +time/sampling batch (s) 0.77927 +time/saving (s) 0.00348072 +time/training (s) 7.34511 +time/epoch (s) 10.8604 +time/total (s) 8043.25 +Epoch -216 +---------------------------------- --------------- +2022-05-10 15:25:03.525593 PDT | [2] Epoch -215 finished +---------------------------------- --------------- +epoch -215 +replay_buffer/size 999033 +trainer/num train calls 786000 +trainer/Policy Loss -19.5309 +trainer/Log Pis Mean 24.5427 +trainer/Log Pis Std 13.327 +trainer/Log Pis Max 66.5382 +trainer/Log Pis Min -9.48414 +trainer/policy/mean Mean -0.0402813 +trainer/policy/mean Std 0.908866 +trainer/policy/mean Max 0.999975 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.76958 +trainer/policy/normal/std Std 0.669139 +trainer/policy/normal/std Max 6.05164 +trainer/policy/normal/std Min 0.190987 +trainer/policy/normal/log_std Mean 0.974884 +trainer/policy/normal/log_std Std 0.339633 +trainer/policy/normal/log_std Max 1.80033 +trainer/policy/normal/log_std Min -1.65555 +eval/num steps total 785081 +eval/num paths total 786 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.226709 +eval/Actions Std 0.883171 +eval/Actions Max 0.999999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68636 +time/logging (s) 0.00406382 +time/sampling batch (s) 0.780651 +time/saving (s) 0.00369039 +time/training (s) 8.6616 +time/epoch (s) 12.1364 +time/total (s) 8055.39 +Epoch -215 +---------------------------------- --------------- +2022-05-10 15:25:13.923562 PDT | [2] Epoch -214 finished +---------------------------------- --------------- +epoch -214 +replay_buffer/size 999033 +trainer/num train calls 787000 +trainer/Policy Loss -18.5622 +trainer/Log Pis Mean 25.0005 +trainer/Log Pis Std 14.1765 +trainer/Log Pis Max 74.9337 +trainer/Log Pis Min -4.99321 +trainer/policy/mean Mean -0.0394324 +trainer/policy/mean Std 0.906705 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.80145 +trainer/policy/normal/std Std 0.698665 +trainer/policy/normal/std Max 6.93886 +trainer/policy/normal/std Min 0.283949 +trainer/policy/normal/log_std Mean 0.983222 +trainer/policy/normal/log_std Std 0.3505 +trainer/policy/normal/log_std Max 1.93714 +trainer/policy/normal/log_std Min -1.25896 +eval/num steps total 786081 +eval/num paths total 787 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.272302 +eval/Actions Std 0.888973 +eval/Actions Max 0.999994 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.4444 +time/logging (s) 0.0037751 +time/sampling batch (s) 0.280714 +time/saving (s) 0.00348654 +time/training (s) 7.64121 +time/epoch (s) 10.3736 +time/total (s) 8065.77 +Epoch -214 +---------------------------------- --------------- +2022-05-10 15:25:24.089554 PDT | [2] Epoch -213 finished +---------------------------------- --------------- +epoch -213 +replay_buffer/size 999033 +trainer/num train calls 788000 +trainer/Policy Loss -19.8155 +trainer/Log Pis Mean 23.3051 +trainer/Log Pis Std 12.8283 +trainer/Log Pis Max 65.0411 +trainer/Log Pis Min -4.04886 +trainer/policy/mean Mean -0.017845 +trainer/policy/mean Std 0.905102 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.76762 +trainer/policy/normal/std Std 0.684696 +trainer/policy/normal/std Max 6.23365 +trainer/policy/normal/std Min 0.305319 +trainer/policy/normal/log_std Mean 0.972652 +trainer/policy/normal/log_std Std 0.343113 +trainer/policy/normal/log_std Max 1.82996 +trainer/policy/normal/log_std Min -1.1864 +eval/num steps total 787081 +eval/num paths total 788 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0802669 +eval/Actions Std 0.905218 +eval/Actions Max 0.999991 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54269 +time/logging (s) 0.00368423 +time/sampling batch (s) 0.278305 +time/saving (s) 0.00339188 +time/training (s) 7.31427 +time/epoch (s) 10.1423 +time/total (s) 8075.91 +Epoch -213 +---------------------------------- --------------- +2022-05-10 15:25:33.566123 PDT | [2] Epoch -212 finished +---------------------------------- --------------- +epoch -212 +replay_buffer/size 999033 +trainer/num train calls 789000 +trainer/Policy Loss -19.3121 +trainer/Log Pis Mean 24.2707 +trainer/Log Pis Std 13.1757 +trainer/Log Pis Max 74.4415 +trainer/Log Pis Min -8.66046 +trainer/policy/mean Mean -0.0246932 +trainer/policy/mean Std 0.906655 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999985 +trainer/policy/normal/std Mean 2.83313 +trainer/policy/normal/std Std 0.658789 +trainer/policy/normal/std Max 6.12001 +trainer/policy/normal/std Min 0.272509 +trainer/policy/normal/log_std Mean 1.00345 +trainer/policy/normal/log_std Std 0.310167 +trainer/policy/normal/log_std Max 1.81156 +trainer/policy/normal/log_std Min -1.30009 +eval/num steps total 788081 +eval/num paths total 789 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.251636 +eval/Actions Std 0.922036 +eval/Actions Max 0.999997 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62599 +time/logging (s) 0.00367831 +time/sampling batch (s) 0.277305 +time/saving (s) 0.00337199 +time/training (s) 6.54271 +time/epoch (s) 9.45305 +time/total (s) 8085.37 +Epoch -212 +---------------------------------- --------------- +2022-05-10 15:25:43.471397 PDT | [2] Epoch -211 finished +---------------------------------- --------------- +epoch -211 +replay_buffer/size 999033 +trainer/num train calls 790000 +trainer/Policy Loss -19.8294 +trainer/Log Pis Mean 23.3413 +trainer/Log Pis Std 13.4226 +trainer/Log Pis Max 65.113 +trainer/Log Pis Min -7.18913 +trainer/policy/mean Mean -0.0145918 +trainer/policy/mean Std 0.903532 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.81098 +trainer/policy/normal/std Std 0.664889 +trainer/policy/normal/std Max 5.57015 +trainer/policy/normal/std Min 0.281656 +trainer/policy/normal/log_std Mean 0.993864 +trainer/policy/normal/log_std Std 0.317021 +trainer/policy/normal/log_std Max 1.71742 +trainer/policy/normal/log_std Min -1.26707 +eval/num steps total 789081 +eval/num paths total 790 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.490576 +eval/Actions Std 0.867988 +eval/Actions Max 0.999987 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.46045 +time/logging (s) 0.00373893 +time/sampling batch (s) 0.277148 +time/saving (s) 0.00339223 +time/training (s) 7.1372 +time/epoch (s) 9.88192 +time/total (s) 8095.25 +Epoch -211 +---------------------------------- --------------- +2022-05-10 15:25:53.396316 PDT | [2] Epoch -210 finished +---------------------------------- --------------- +epoch -210 +replay_buffer/size 999033 +trainer/num train calls 791000 +trainer/Policy Loss -20.0403 +trainer/Log Pis Mean 24.7695 +trainer/Log Pis Std 12.3842 +trainer/Log Pis Max 66.6673 +trainer/Log Pis Min -6.2231 +trainer/policy/mean Mean -0.0483798 +trainer/policy/mean Std 0.909008 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.83447 +trainer/policy/normal/std Std 0.667866 +trainer/policy/normal/std Max 6.2754 +trainer/policy/normal/std Min 0.347526 +trainer/policy/normal/log_std Mean 1.0023 +trainer/policy/normal/log_std Std 0.316927 +trainer/policy/normal/log_std Max 1.83664 +trainer/policy/normal/log_std Min -1.05692 +eval/num steps total 790081 +eval/num paths total 791 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.175985 +eval/Actions Std 0.824613 +eval/Actions Max 0.999997 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55441 +time/logging (s) 0.00369158 +time/sampling batch (s) 0.278257 +time/saving (s) 0.0033568 +time/training (s) 7.06165 +time/epoch (s) 9.90136 +time/total (s) 8105.16 +Epoch -210 +---------------------------------- --------------- +2022-05-10 15:26:03.869123 PDT | [2] Epoch -209 finished +---------------------------------- --------------- +epoch -209 +replay_buffer/size 999033 +trainer/num train calls 792000 +trainer/Policy Loss -19.883 +trainer/Log Pis Mean 23.2753 +trainer/Log Pis Std 13.0136 +trainer/Log Pis Max 62.2808 +trainer/Log Pis Min -3.65758 +trainer/policy/mean Mean -0.0471643 +trainer/policy/mean Std 0.904466 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.75327 +trainer/policy/normal/std Std 0.687282 +trainer/policy/normal/std Max 5.31854 +trainer/policy/normal/std Min 0.194673 +trainer/policy/normal/log_std Mean 0.964265 +trainer/policy/normal/log_std Std 0.360808 +trainer/policy/normal/log_std Max 1.6712 +trainer/policy/normal/log_std Min -1.63643 +eval/num steps total 791081 +eval/num paths total 792 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.120676 +eval/Actions Std 0.921389 +eval/Actions Max 0.999995 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70065 +time/logging (s) 0.00370884 +time/sampling batch (s) 0.278543 +time/saving (s) 0.00342256 +time/training (s) 7.46305 +time/epoch (s) 10.4494 +time/total (s) 8115.61 +Epoch -209 +---------------------------------- --------------- +2022-05-10 15:26:13.417686 PDT | [2] Epoch -208 finished +---------------------------------- --------------- +epoch -208 +replay_buffer/size 999033 +trainer/num train calls 793000 +trainer/Policy Loss -18.9553 +trainer/Log Pis Mean 24.3042 +trainer/Log Pis Std 12.623 +trainer/Log Pis Max 57.7716 +trainer/Log Pis Min -5.02417 +trainer/policy/mean Mean -0.036682 +trainer/policy/mean Std 0.905933 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.80247 +trainer/policy/normal/std Std 0.639316 +trainer/policy/normal/std Max 5.91392 +trainer/policy/normal/std Min 0.244323 +trainer/policy/normal/log_std Mean 0.993423 +trainer/policy/normal/log_std Std 0.30629 +trainer/policy/normal/log_std Max 1.77731 +trainer/policy/normal/log_std Min -1.40926 +eval/num steps total 792081 +eval/num paths total 793 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.221765 +eval/Actions Std 0.944751 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60225 +time/logging (s) 0.00373543 +time/sampling batch (s) 0.277869 +time/saving (s) 0.00335454 +time/training (s) 6.63779 +time/epoch (s) 9.52501 +time/total (s) 8125.14 +Epoch -208 +---------------------------------- --------------- +2022-05-10 15:26:23.836837 PDT | [2] Epoch -207 finished +---------------------------------- --------------- +epoch -207 +replay_buffer/size 999033 +trainer/num train calls 794000 +trainer/Policy Loss -19.5385 +trainer/Log Pis Mean 24.383 +trainer/Log Pis Std 13.5683 +trainer/Log Pis Max 72.7247 +trainer/Log Pis Min -12.5966 +trainer/policy/mean Mean -0.0300634 +trainer/policy/mean Std 0.906046 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.86195 +trainer/policy/normal/std Std 0.681894 +trainer/policy/normal/std Max 6.62285 +trainer/policy/normal/std Min 0.327251 +trainer/policy/normal/log_std Mean 1.0127 +trainer/policy/normal/log_std Std 0.311029 +trainer/policy/normal/log_std Max 1.89053 +trainer/policy/normal/log_std Min -1.11703 +eval/num steps total 793081 +eval/num paths total 794 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.228708 +eval/Actions Std 0.790096 +eval/Actions Max 0.999991 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.72076 +time/logging (s) 0.00629524 +time/sampling batch (s) 0.280788 +time/saving (s) 0.00526583 +time/training (s) 7.38483 +time/epoch (s) 10.3979 +time/total (s) 8135.54 +Epoch -207 +---------------------------------- --------------- +2022-05-10 15:26:33.644892 PDT | [2] Epoch -206 finished +---------------------------------- --------------- +epoch -206 +replay_buffer/size 999033 +trainer/num train calls 795000 +trainer/Policy Loss -19.5532 +trainer/Log Pis Mean 23.8475 +trainer/Log Pis Std 12.531 +trainer/Log Pis Max 63.0415 +trainer/Log Pis Min -4.72589 +trainer/policy/mean Mean -0.0243679 +trainer/policy/mean Std 0.905168 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.80077 +trainer/policy/normal/std Std 0.668265 +trainer/policy/normal/std Max 5.21846 +trainer/policy/normal/std Min 0.29183 +trainer/policy/normal/log_std Mean 0.988888 +trainer/policy/normal/log_std Std 0.322757 +trainer/policy/normal/log_std Max 1.6522 +trainer/policy/normal/log_std Min -1.23158 +eval/num steps total 794081 +eval/num paths total 795 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.260759 +eval/Actions Std 0.882118 +eval/Actions Max 0.999991 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.45168 +time/logging (s) 0.00371639 +time/sampling batch (s) 0.274135 +time/saving (s) 0.00345692 +time/training (s) 7.04465 +time/epoch (s) 9.77764 +time/total (s) 8145.32 +Epoch -206 +---------------------------------- --------------- +2022-05-10 15:26:44.050839 PDT | [2] Epoch -205 finished +---------------------------------- --------------- +epoch -205 +replay_buffer/size 999033 +trainer/num train calls 796000 +trainer/Policy Loss -19.2961 +trainer/Log Pis Mean 23.3802 +trainer/Log Pis Std 12.9223 +trainer/Log Pis Max 71.4436 +trainer/Log Pis Min -6.61515 +trainer/policy/mean Mean -0.0301821 +trainer/policy/mean Std 0.906327 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.7543 +trainer/policy/normal/std Std 0.662278 +trainer/policy/normal/std Max 5.24252 +trainer/policy/normal/std Min 0.24699 +trainer/policy/normal/log_std Mean 0.968471 +trainer/policy/normal/log_std Std 0.345144 +trainer/policy/normal/log_std Max 1.6568 +trainer/policy/normal/log_std Min -1.39841 +eval/num steps total 795081 +eval/num paths total 796 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0510617 +eval/Actions Std 0.913696 +eval/Actions Max 0.999989 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.43333 +time/logging (s) 0.00370996 +time/sampling batch (s) 0.77555 +time/saving (s) 0.00332959 +time/training (s) 7.16661 +time/epoch (s) 10.3825 +time/total (s) 8155.71 +Epoch -205 +---------------------------------- --------------- +2022-05-10 15:26:54.333822 PDT | [2] Epoch -204 finished +---------------------------------- --------------- +epoch -204 +replay_buffer/size 999033 +trainer/num train calls 797000 +trainer/Policy Loss -19.5789 +trainer/Log Pis Mean 24.2621 +trainer/Log Pis Std 13.1988 +trainer/Log Pis Max 68.7589 +trainer/Log Pis Min -9.48644 +trainer/policy/mean Mean -0.051182 +trainer/policy/mean Std 0.905945 +trainer/policy/mean Max 0.99998 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.81177 +trainer/policy/normal/std Std 0.655925 +trainer/policy/normal/std Max 6.04683 +trainer/policy/normal/std Min 0.228424 +trainer/policy/normal/log_std Mean 0.995684 +trainer/policy/normal/log_std Std 0.309949 +trainer/policy/normal/log_std Max 1.79953 +trainer/policy/normal/log_std Min -1.47655 +eval/num steps total 796081 +eval/num paths total 797 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0214998 +eval/Actions Std 0.864383 +eval/Actions Max 0.999993 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64218 +time/logging (s) 0.00372358 +time/sampling batch (s) 0.524366 +time/saving (s) 0.00335444 +time/training (s) 7.08608 +time/epoch (s) 10.2597 +time/total (s) 8165.97 +Epoch -204 +---------------------------------- --------------- +2022-05-10 15:27:04.052478 PDT | [2] Epoch -203 finished +---------------------------------- --------------- +epoch -203 +replay_buffer/size 999033 +trainer/num train calls 798000 +trainer/Policy Loss -19.9728 +trainer/Log Pis Mean 25.036 +trainer/Log Pis Std 13.179 +trainer/Log Pis Max 67.0292 +trainer/Log Pis Min -7.43571 +trainer/policy/mean Mean -0.0521383 +trainer/policy/mean Std 0.907297 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.73192 +trainer/policy/normal/std Std 0.672903 +trainer/policy/normal/std Max 5.67202 +trainer/policy/normal/std Min 0.274341 +trainer/policy/normal/log_std Mean 0.959339 +trainer/policy/normal/log_std Std 0.345303 +trainer/policy/normal/log_std Max 1.73555 +trainer/policy/normal/log_std Min -1.29338 +eval/num steps total 797081 +eval/num paths total 798 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0756044 +eval/Actions Std 0.906717 +eval/Actions Max 0.999998 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6437 +time/logging (s) 0.00382812 +time/sampling batch (s) 0.273421 +time/saving (s) 0.00343515 +time/training (s) 6.77115 +time/epoch (s) 9.69553 +time/total (s) 8175.67 +Epoch -203 +---------------------------------- --------------- +2022-05-10 15:27:14.080763 PDT | [2] Epoch -202 finished +---------------------------------- --------------- +epoch -202 +replay_buffer/size 999033 +trainer/num train calls 799000 +trainer/Policy Loss -20.255 +trainer/Log Pis Mean 24.7419 +trainer/Log Pis Std 13.9942 +trainer/Log Pis Max 76.4116 +trainer/Log Pis Min -11.2458 +trainer/policy/mean Mean -0.0309632 +trainer/policy/mean Std 0.906947 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.76506 +trainer/policy/normal/std Std 0.674837 +trainer/policy/normal/std Max 5.37772 +trainer/policy/normal/std Min 0.206192 +trainer/policy/normal/log_std Mean 0.973328 +trainer/policy/normal/log_std Std 0.336721 +trainer/policy/normal/log_std Max 1.68226 +trainer/policy/normal/log_std Min -1.57895 +eval/num steps total 798081 +eval/num paths total 799 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.101105 +eval/Actions Std 0.918443 +eval/Actions Max 0.999976 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.49778 +time/logging (s) 0.00373556 +time/sampling batch (s) 0.775651 +time/saving (s) 0.00335431 +time/training (s) 6.72427 +time/epoch (s) 10.0048 +time/total (s) 8185.67 +Epoch -202 +---------------------------------- --------------- +2022-05-10 15:27:24.519943 PDT | [2] Epoch -201 finished +---------------------------------- --------------- +epoch -201 +replay_buffer/size 999033 +trainer/num train calls 800000 +trainer/Policy Loss -18.1635 +trainer/Log Pis Mean 24.6685 +trainer/Log Pis Std 12.8751 +trainer/Log Pis Max 71.4387 +trainer/Log Pis Min -10.0602 +trainer/policy/mean Mean -0.0355408 +trainer/policy/mean Std 0.904516 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.82742 +trainer/policy/normal/std Std 0.678047 +trainer/policy/normal/std Max 5.85036 +trainer/policy/normal/std Min 0.287716 +trainer/policy/normal/log_std Mean 0.997969 +trainer/policy/normal/log_std Std 0.324702 +trainer/policy/normal/log_std Max 1.7665 +trainer/policy/normal/log_std Min -1.24578 +eval/num steps total 799081 +eval/num paths total 800 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.255051 +eval/Actions Std 0.86422 +eval/Actions Max 0.999984 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.4444 +time/logging (s) 0.0037096 +time/sampling batch (s) 0.526171 +time/saving (s) 0.00619805 +time/training (s) 7.43534 +time/epoch (s) 10.4158 +time/total (s) 8196.09 +Epoch -201 +---------------------------------- --------------- +2022-05-10 15:27:35.407821 PDT | [2] Epoch -200 finished +---------------------------------- -------------- +epoch -200 +replay_buffer/size 999033 +trainer/num train calls 801000 +trainer/Policy Loss -18.8123 +trainer/Log Pis Mean 24.855 +trainer/Log Pis Std 13.8165 +trainer/Log Pis Max 78.4501 +trainer/Log Pis Min -12.6106 +trainer/policy/mean Mean -0.045889 +trainer/policy/mean Std 0.904465 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.76936 +trainer/policy/normal/std Std 0.659686 +trainer/policy/normal/std Max 5.42055 +trainer/policy/normal/std Min 0.213173 +trainer/policy/normal/log_std Mean 0.975817 +trainer/policy/normal/log_std Std 0.33602 +trainer/policy/normal/log_std Max 1.6902 +trainer/policy/normal/log_std Min -1.54565 +eval/num steps total 800081 +eval/num paths total 801 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.435289 +eval/Actions Std 0.887765 +eval/Actions Max 0.999989 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68647 +time/logging (s) 0.0037017 +time/sampling batch (s) 0.526274 +time/saving (s) 0.0033525 +time/training (s) 7.64477 +time/epoch (s) 10.8646 +time/total (s) 8206.96 +Epoch -200 +---------------------------------- -------------- +2022-05-10 15:27:44.673727 PDT | [2] Epoch -199 finished +---------------------------------- --------------- +epoch -199 +replay_buffer/size 999033 +trainer/num train calls 802000 +trainer/Policy Loss -18.59 +trainer/Log Pis Mean 23.807 +trainer/Log Pis Std 13.2457 +trainer/Log Pis Max 73.967 +trainer/Log Pis Min -8.1151 +trainer/policy/mean Mean -0.0494063 +trainer/policy/mean Std 0.904685 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.80124 +trainer/policy/normal/std Std 0.691941 +trainer/policy/normal/std Max 5.97778 +trainer/policy/normal/std Min 0.344082 +trainer/policy/normal/log_std Mean 0.985679 +trainer/policy/normal/log_std Std 0.338362 +trainer/policy/normal/log_std Max 1.78805 +trainer/policy/normal/log_std Min -1.06687 +eval/num steps total 801081 +eval/num paths total 802 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.119123 +eval/Actions Std 0.886442 +eval/Actions Max 0.999992 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.32227 +time/logging (s) 0.00368544 +time/sampling batch (s) 0.274751 +time/saving (s) 0.00332582 +time/training (s) 6.6386 +time/epoch (s) 9.24263 +time/total (s) 8216.21 +Epoch -199 +---------------------------------- --------------- +2022-05-10 15:27:54.010734 PDT | [2] Epoch -198 finished +---------------------------------- --------------- +epoch -198 +replay_buffer/size 999033 +trainer/num train calls 803000 +trainer/Policy Loss -19.2839 +trainer/Log Pis Mean 24.9024 +trainer/Log Pis Std 14.5025 +trainer/Log Pis Max 74.8556 +trainer/Log Pis Min -10.6739 +trainer/policy/mean Mean -0.0273225 +trainer/policy/mean Std 0.904632 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.79341 +trainer/policy/normal/std Std 0.68696 +trainer/policy/normal/std Max 6.24754 +trainer/policy/normal/std Min 0.287745 +trainer/policy/normal/log_std Mean 0.982586 +trainer/policy/normal/log_std Std 0.340787 +trainer/policy/normal/log_std Max 1.83219 +trainer/policy/normal/log_std Min -1.24568 +eval/num steps total 802081 +eval/num paths total 803 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.104142 +eval/Actions Std 0.872572 +eval/Actions Max 0.999993 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59725 +time/logging (s) 0.00365661 +time/sampling batch (s) 0.27572 +time/saving (s) 0.00337204 +time/training (s) 6.43371 +time/epoch (s) 9.31371 +time/total (s) 8225.52 +Epoch -198 +---------------------------------- --------------- +2022-05-10 15:28:04.417971 PDT | [2] Epoch -197 finished +---------------------------------- --------------- +epoch -197 +replay_buffer/size 999033 +trainer/num train calls 804000 +trainer/Policy Loss -19.4936 +trainer/Log Pis Mean 24.6754 +trainer/Log Pis Std 13.0355 +trainer/Log Pis Max 63.3207 +trainer/Log Pis Min -5.91072 +trainer/policy/mean Mean -0.0291544 +trainer/policy/mean Std 0.906328 +trainer/policy/mean Max 0.999976 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.82952 +trainer/policy/normal/std Std 0.703127 +trainer/policy/normal/std Max 6.34 +trainer/policy/normal/std Min 0.275044 +trainer/policy/normal/log_std Mean 0.994725 +trainer/policy/normal/log_std Std 0.343027 +trainer/policy/normal/log_std Max 1.84688 +trainer/policy/normal/log_std Min -1.29082 +eval/num steps total 803081 +eval/num paths total 804 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.313044 +eval/Actions Std 0.722782 +eval/Actions Max 0.999989 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.88196 +time/logging (s) 0.00370662 +time/sampling batch (s) 0.274088 +time/saving (s) 0.00334801 +time/training (s) 7.22074 +time/epoch (s) 10.3838 +time/total (s) 8235.91 +Epoch -197 +---------------------------------- --------------- +2022-05-10 15:28:14.754347 PDT | [2] Epoch -196 finished +---------------------------------- --------------- +epoch -196 +replay_buffer/size 999033 +trainer/num train calls 805000 +trainer/Policy Loss -20.6392 +trainer/Log Pis Mean 23.7696 +trainer/Log Pis Std 12.6959 +trainer/Log Pis Max 66.8381 +trainer/Log Pis Min -9.40287 +trainer/policy/mean Mean -0.0246298 +trainer/policy/mean Std 0.908131 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.79514 +trainer/policy/normal/std Std 0.653168 +trainer/policy/normal/std Max 6.24425 +trainer/policy/normal/std Min 0.288129 +trainer/policy/normal/log_std Mean 0.989217 +trainer/policy/normal/log_std Std 0.31309 +trainer/policy/normal/log_std Max 1.83166 +trainer/policy/normal/log_std Min -1.24435 +eval/num steps total 804081 +eval/num paths total 805 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.159401 +eval/Actions Std 0.907261 +eval/Actions Max 0.999994 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59197 +time/logging (s) 0.00387581 +time/sampling batch (s) 0.275833 +time/saving (s) 0.00359753 +time/training (s) 7.43784 +time/epoch (s) 10.3131 +time/total (s) 8246.22 +Epoch -196 +---------------------------------- --------------- +2022-05-10 15:28:25.238414 PDT | [2] Epoch -195 finished +---------------------------------- --------------- +epoch -195 +replay_buffer/size 999033 +trainer/num train calls 806000 +trainer/Policy Loss -19.0519 +trainer/Log Pis Mean 24.2203 +trainer/Log Pis Std 13.9076 +trainer/Log Pis Max 79.8393 +trainer/Log Pis Min -10.5981 +trainer/policy/mean Mean -0.0509334 +trainer/policy/mean Std 0.907044 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.82216 +trainer/policy/normal/std Std 0.67224 +trainer/policy/normal/std Max 6.91323 +trainer/policy/normal/std Min 0.288205 +trainer/policy/normal/log_std Mean 0.996906 +trainer/policy/normal/log_std Std 0.322353 +trainer/policy/normal/log_std Max 1.93344 +trainer/policy/normal/log_std Min -1.24409 +eval/num steps total 805081 +eval/num paths total 806 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0286658 +eval/Actions Std 0.902626 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.545 +time/logging (s) 0.0036757 +time/sampling batch (s) 0.774649 +time/saving (s) 0.00337118 +time/training (s) 7.1338 +time/epoch (s) 10.4605 +time/total (s) 8256.69 +Epoch -195 +---------------------------------- --------------- +2022-05-10 15:28:36.439996 PDT | [2] Epoch -194 finished +---------------------------------- --------------- +epoch -194 +replay_buffer/size 999033 +trainer/num train calls 807000 +trainer/Policy Loss -20.1958 +trainer/Log Pis Mean 25.6566 +trainer/Log Pis Std 12.9546 +trainer/Log Pis Max 73.7142 +trainer/Log Pis Min -11.7509 +trainer/policy/mean Mean -0.0212542 +trainer/policy/mean Std 0.910663 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.76462 +trainer/policy/normal/std Std 0.662626 +trainer/policy/normal/std Max 5.69791 +trainer/policy/normal/std Min 0.23044 +trainer/policy/normal/log_std Mean 0.973491 +trainer/policy/normal/log_std Std 0.337196 +trainer/policy/normal/log_std Max 1.7401 +trainer/policy/normal/log_std Min -1.46776 +eval/num steps total 806081 +eval/num paths total 807 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0627412 +eval/Actions Std 0.919701 +eval/Actions Max 0.999997 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6143 +time/logging (s) 0.00377119 +time/sampling batch (s) 0.524258 +time/saving (s) 0.00330424 +time/training (s) 8.03288 +time/epoch (s) 11.1785 +time/total (s) 8267.87 +Epoch -194 +---------------------------------- --------------- +2022-05-10 15:28:47.271534 PDT | [2] Epoch -193 finished +---------------------------------- --------------- +epoch -193 +replay_buffer/size 999033 +trainer/num train calls 808000 +trainer/Policy Loss -20.6862 +trainer/Log Pis Mean 22.9411 +trainer/Log Pis Std 12.845 +trainer/Log Pis Max 63.5754 +trainer/Log Pis Min -7.87367 +trainer/policy/mean Mean -0.0358762 +trainer/policy/mean Std 0.905012 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.80307 +trainer/policy/normal/std Std 0.657747 +trainer/policy/normal/std Max 5.19088 +trainer/policy/normal/std Min 0.208374 +trainer/policy/normal/log_std Mean 0.989692 +trainer/policy/normal/log_std Std 0.326661 +trainer/policy/normal/log_std Max 1.6469 +trainer/policy/normal/log_std Min -1.56842 +eval/num steps total 807081 +eval/num paths total 808 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.100535 +eval/Actions Std 0.917233 +eval/Actions Max 0.999987 +eval/Actions Min -0.999988 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57186 +time/logging (s) 0.00372567 +time/sampling batch (s) 0.275405 +time/saving (s) 0.0033256 +time/training (s) 7.95384 +time/epoch (s) 10.8081 +time/total (s) 8278.68 +Epoch -193 +---------------------------------- --------------- +2022-05-10 15:28:57.337333 PDT | [2] Epoch -192 finished +---------------------------------- --------------- +epoch -192 +replay_buffer/size 999033 +trainer/num train calls 809000 +trainer/Policy Loss -20.2108 +trainer/Log Pis Mean 23.4827 +trainer/Log Pis Std 13.0509 +trainer/Log Pis Max 69.9782 +trainer/Log Pis Min -5.90039 +trainer/policy/mean Mean -0.0224866 +trainer/policy/mean Std 0.909513 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999979 +trainer/policy/normal/std Mean 2.76327 +trainer/policy/normal/std Std 0.655876 +trainer/policy/normal/std Max 7.30901 +trainer/policy/normal/std Min 0.230739 +trainer/policy/normal/log_std Mean 0.975449 +trainer/policy/normal/log_std Std 0.324936 +trainer/policy/normal/log_std Max 1.98911 +trainer/policy/normal/log_std Min -1.46647 +eval/num steps total 808081 +eval/num paths total 809 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00963702 +eval/Actions Std 0.908435 +eval/Actions Max 0.999994 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61592 +time/logging (s) 0.00371459 +time/sampling batch (s) 0.276219 +time/saving (s) 0.00338774 +time/training (s) 7.14315 +time/epoch (s) 10.0424 +time/total (s) 8288.73 +Epoch -192 +---------------------------------- --------------- +2022-05-10 15:29:08.490133 PDT | [2] Epoch -191 finished +---------------------------------- --------------- +epoch -191 +replay_buffer/size 999033 +trainer/num train calls 810000 +trainer/Policy Loss -20.1654 +trainer/Log Pis Mean 25.033 +trainer/Log Pis Std 13.7358 +trainer/Log Pis Max 75.5969 +trainer/Log Pis Min -10.656 +trainer/policy/mean Mean -0.0368036 +trainer/policy/mean Std 0.90185 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.89513 +trainer/policy/normal/std Std 0.697222 +trainer/policy/normal/std Max 5.68621 +trainer/policy/normal/std Min 0.235021 +trainer/policy/normal/log_std Mean 1.02146 +trainer/policy/normal/log_std Std 0.324818 +trainer/policy/normal/log_std Max 1.73804 +trainer/policy/normal/log_std Min -1.44808 +eval/num steps total 809081 +eval/num paths total 810 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.089855 +eval/Actions Std 0.905036 +eval/Actions Max 0.999994 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59609 +time/logging (s) 0.00370147 +time/sampling batch (s) 0.776716 +time/saving (s) 0.00336593 +time/training (s) 7.74936 +time/epoch (s) 11.1292 +time/total (s) 8299.86 +Epoch -191 +---------------------------------- --------------- +2022-05-10 15:29:19.638062 PDT | [2] Epoch -190 finished +---------------------------------- --------------- +epoch -190 +replay_buffer/size 999033 +trainer/num train calls 811000 +trainer/Policy Loss -18.6545 +trainer/Log Pis Mean 24.074 +trainer/Log Pis Std 13.5915 +trainer/Log Pis Max 82.3714 +trainer/Log Pis Min -4.52667 +trainer/policy/mean Mean -0.0437258 +trainer/policy/mean Std 0.90514 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.84785 +trainer/policy/normal/std Std 0.689541 +trainer/policy/normal/std Max 7.31818 +trainer/policy/normal/std Min 0.352689 +trainer/policy/normal/log_std Mean 1.00495 +trainer/policy/normal/log_std Std 0.324149 +trainer/policy/normal/log_std Max 1.99036 +trainer/policy/normal/log_std Min -1.04217 +eval/num steps total 810081 +eval/num paths total 811 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.250631 +eval/Actions Std 0.877587 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50436 +time/logging (s) 0.00367692 +time/sampling batch (s) 0.776766 +time/saving (s) 0.00336055 +time/training (s) 7.83612 +time/epoch (s) 11.1243 +time/total (s) 8310.99 +Epoch -190 +---------------------------------- --------------- +2022-05-10 15:29:30.336747 PDT | [2] Epoch -189 finished +---------------------------------- --------------- +epoch -189 +replay_buffer/size 999033 +trainer/num train calls 812000 +trainer/Policy Loss -20.319 +trainer/Log Pis Mean 24.5027 +trainer/Log Pis Std 13.292 +trainer/Log Pis Max 67.4805 +trainer/Log Pis Min -8.91418 +trainer/policy/mean Mean -0.0257722 +trainer/policy/mean Std 0.909892 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.76138 +trainer/policy/normal/std Std 0.658159 +trainer/policy/normal/std Max 6.75312 +trainer/policy/normal/std Min 0.252626 +trainer/policy/normal/log_std Mean 0.973943 +trainer/policy/normal/log_std Std 0.329235 +trainer/policy/normal/log_std Max 1.91 +trainer/policy/normal/log_std Min -1.37584 +eval/num steps total 811081 +eval/num paths total 812 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.217138 +eval/Actions Std 0.764834 +eval/Actions Max 0.999999 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6435 +time/logging (s) 0.00368473 +time/sampling batch (s) 0.274805 +time/saving (s) 0.00331291 +time/training (s) 7.75005 +time/epoch (s) 10.6754 +time/total (s) 8321.66 +Epoch -189 +---------------------------------- --------------- +2022-05-10 15:29:41.249099 PDT | [2] Epoch -188 finished +---------------------------------- --------------- +epoch -188 +replay_buffer/size 999033 +trainer/num train calls 813000 +trainer/Policy Loss -20.6638 +trainer/Log Pis Mean 23.5653 +trainer/Log Pis Std 13.0073 +trainer/Log Pis Max 66.301 +trainer/Log Pis Min -4.60579 +trainer/policy/mean Mean -0.0247561 +trainer/policy/mean Std 0.903168 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999984 +trainer/policy/normal/std Mean 2.7877 +trainer/policy/normal/std Std 0.698367 +trainer/policy/normal/std Max 6.75855 +trainer/policy/normal/std Min 0.276258 +trainer/policy/normal/log_std Mean 0.977919 +trainer/policy/normal/log_std Std 0.352895 +trainer/policy/normal/log_std Max 1.91081 +trainer/policy/normal/log_std Min -1.28642 +eval/num steps total 812081 +eval/num paths total 813 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.176586 +eval/Actions Std 0.927956 +eval/Actions Max 0.999986 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73564 +time/logging (s) 0.0038004 +time/sampling batch (s) 0.275486 +time/saving (s) 0.00344172 +time/training (s) 7.87085 +time/epoch (s) 10.8892 +time/total (s) 8332.56 +Epoch -188 +---------------------------------- --------------- +2022-05-10 15:29:51.673802 PDT | [2] Epoch -187 finished +---------------------------------- --------------- +epoch -187 +replay_buffer/size 999033 +trainer/num train calls 814000 +trainer/Policy Loss -19.2339 +trainer/Log Pis Mean 23.568 +trainer/Log Pis Std 13.238 +trainer/Log Pis Max 64.3482 +trainer/Log Pis Min -7.56695 +trainer/policy/mean Mean -0.0336375 +trainer/policy/mean Std 0.906915 +trainer/policy/mean Max 0.999978 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.88636 +trainer/policy/normal/std Std 0.694991 +trainer/policy/normal/std Max 6.94397 +trainer/policy/normal/std Min 0.294259 +trainer/policy/normal/log_std Mean 1.02089 +trainer/policy/normal/log_std Std 0.310784 +trainer/policy/normal/log_std Max 1.93787 +trainer/policy/normal/log_std Min -1.2233 +eval/num steps total 813081 +eval/num paths total 814 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0350531 +eval/Actions Std 0.954164 +eval/Actions Max 0.999999 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.8557 +time/logging (s) 0.00391121 +time/sampling batch (s) 0.275102 +time/saving (s) 0.00357514 +time/training (s) 7.26296 +time/epoch (s) 10.4013 +time/total (s) 8342.96 +Epoch -187 +---------------------------------- --------------- +2022-05-10 15:30:01.229974 PDT | [2] Epoch -186 finished +---------------------------------- --------------- +epoch -186 +replay_buffer/size 999033 +trainer/num train calls 815000 +trainer/Policy Loss -20.0675 +trainer/Log Pis Mean 24.2026 +trainer/Log Pis Std 13.5547 +trainer/Log Pis Max 72.8438 +trainer/Log Pis Min -8.53774 +trainer/policy/mean Mean -0.0272374 +trainer/policy/mean Std 0.905228 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.73129 +trainer/policy/normal/std Std 0.655352 +trainer/policy/normal/std Max 5.43929 +trainer/policy/normal/std Min 0.288433 +trainer/policy/normal/log_std Mean 0.961975 +trainer/policy/normal/log_std Std 0.333135 +trainer/policy/normal/log_std Max 1.69365 +trainer/policy/normal/log_std Min -1.24329 +eval/num steps total 814081 +eval/num paths total 815 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.077018 +eval/Actions Std 0.911746 +eval/Actions Max 0.999991 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64166 +time/logging (s) 0.00373415 +time/sampling batch (s) 0.273978 +time/saving (s) 0.0033581 +time/training (s) 6.60954 +time/epoch (s) 9.53227 +time/total (s) 8352.5 +Epoch -186 +---------------------------------- --------------- +2022-05-10 15:30:12.376585 PDT | [2] Epoch -185 finished +---------------------------------- --------------- +epoch -185 +replay_buffer/size 999033 +trainer/num train calls 816000 +trainer/Policy Loss -20.5302 +trainer/Log Pis Mean 25.6582 +trainer/Log Pis Std 13.6042 +trainer/Log Pis Max 79.5605 +trainer/Log Pis Min -14.534 +trainer/policy/mean Mean -0.0414749 +trainer/policy/mean Std 0.911082 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.783 +trainer/policy/normal/std Std 0.645017 +trainer/policy/normal/std Max 6.54089 +trainer/policy/normal/std Min 0.234002 +trainer/policy/normal/log_std Mean 0.98607 +trainer/policy/normal/log_std Std 0.306469 +trainer/policy/normal/log_std Max 1.87807 +trainer/policy/normal/log_std Min -1.45243 +eval/num steps total 815081 +eval/num paths total 816 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0288536 +eval/Actions Std 0.885801 +eval/Actions Max 0.999992 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59705 +time/logging (s) 0.00377141 +time/sampling batch (s) 0.274026 +time/saving (s) 0.00333201 +time/training (s) 8.24481 +time/epoch (s) 11.123 +time/total (s) 8363.62 +Epoch -185 +---------------------------------- --------------- +2022-05-10 15:30:24.617314 PDT | [2] Epoch -184 finished +---------------------------------- --------------- +epoch -184 +replay_buffer/size 999033 +trainer/num train calls 817000 +trainer/Policy Loss -20.0014 +trainer/Log Pis Mean 25.5039 +trainer/Log Pis Std 13.4072 +trainer/Log Pis Max 72.4627 +trainer/Log Pis Min -5.65946 +trainer/policy/mean Mean -0.0566192 +trainer/policy/mean Std 0.907744 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.99998 +trainer/policy/normal/std Mean 2.79219 +trainer/policy/normal/std Std 0.685694 +trainer/policy/normal/std Max 5.46802 +trainer/policy/normal/std Min 0.277568 +trainer/policy/normal/log_std Mean 0.982647 +trainer/policy/normal/log_std Std 0.337112 +trainer/policy/normal/log_std Max 1.69892 +trainer/policy/normal/log_std Min -1.28169 +eval/num steps total 816081 +eval/num paths total 817 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0523262 +eval/Actions Std 0.902891 +eval/Actions Max 0.999999 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58521 +time/logging (s) 0.00378461 +time/sampling batch (s) 0.280484 +time/saving (s) 0.00334717 +time/training (s) 9.34419 +time/epoch (s) 12.217 +time/total (s) 8375.84 +Epoch -184 +---------------------------------- --------------- +2022-05-10 15:30:36.463119 PDT | [2] Epoch -183 finished +---------------------------------- --------------- +epoch -183 +replay_buffer/size 999033 +trainer/num train calls 818000 +trainer/Policy Loss -19.6894 +trainer/Log Pis Mean 24.9647 +trainer/Log Pis Std 13.5324 +trainer/Log Pis Max 67.6523 +trainer/Log Pis Min -4.93519 +trainer/policy/mean Mean -0.0349266 +trainer/policy/mean Std 0.907501 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999984 +trainer/policy/normal/std Mean 2.77082 +trainer/policy/normal/std Std 0.662027 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.320605 +trainer/policy/normal/log_std Mean 0.977963 +trainer/policy/normal/log_std Std 0.324214 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.13755 +eval/num steps total 817081 +eval/num paths total 818 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.120796 +eval/Actions Std 0.779543 +eval/Actions Max 0.999979 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61978 +time/logging (s) 0.00392875 +time/sampling batch (s) 1.03142 +time/saving (s) 0.00334761 +time/training (s) 8.16365 +time/epoch (s) 11.8221 +time/total (s) 8387.67 +Epoch -183 +---------------------------------- --------------- +2022-05-10 15:30:46.626071 PDT | [2] Epoch -182 finished +---------------------------------- --------------- +epoch -182 +replay_buffer/size 999033 +trainer/num train calls 819000 +trainer/Policy Loss -19.1561 +trainer/Log Pis Mean 24.154 +trainer/Log Pis Std 13.4211 +trainer/Log Pis Max 97.3598 +trainer/Log Pis Min -8.52209 +trainer/policy/mean Mean -0.0329935 +trainer/policy/mean Std 0.906206 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.75487 +trainer/policy/normal/std Std 0.660237 +trainer/policy/normal/std Max 5.81282 +trainer/policy/normal/std Min 0.234731 +trainer/policy/normal/log_std Mean 0.971568 +trainer/policy/normal/log_std Std 0.328593 +trainer/policy/normal/log_std Max 1.76007 +trainer/policy/normal/log_std Min -1.44932 +eval/num steps total 818081 +eval/num paths total 819 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.162668 +eval/Actions Std 0.952667 +eval/Actions Max 0.999993 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51956 +time/logging (s) 0.00381723 +time/sampling batch (s) 0.319342 +time/saving (s) 0.00355794 +time/training (s) 7.2925 +time/epoch (s) 10.1388 +time/total (s) 8397.81 +Epoch -182 +---------------------------------- --------------- +2022-05-10 15:30:57.985123 PDT | [2] Epoch -181 finished +---------------------------------- --------------- +epoch -181 +replay_buffer/size 999033 +trainer/num train calls 820000 +trainer/Policy Loss -19.2599 +trainer/Log Pis Mean 24.8253 +trainer/Log Pis Std 13.3903 +trainer/Log Pis Max 64.3163 +trainer/Log Pis Min -9.85062 +trainer/policy/mean Mean -0.0293782 +trainer/policy/mean Std 0.902272 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.78509 +trainer/policy/normal/std Std 0.661412 +trainer/policy/normal/std Max 5.92571 +trainer/policy/normal/std Min 0.230774 +trainer/policy/normal/log_std Mean 0.98253 +trainer/policy/normal/log_std Std 0.330137 +trainer/policy/normal/log_std Max 1.7793 +trainer/policy/normal/log_std Min -1.46632 +eval/num steps total 819081 +eval/num paths total 820 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0828811 +eval/Actions Std 0.896355 +eval/Actions Max 0.999994 +eval/Actions Min -0.999983 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70905 +time/logging (s) 0.0037668 +time/sampling batch (s) 0.326829 +time/saving (s) 0.00354392 +time/training (s) 8.29117 +time/epoch (s) 11.3344 +time/total (s) 8409.15 +Epoch -181 +---------------------------------- --------------- +2022-05-10 15:31:09.270746 PDT | [2] Epoch -180 finished +---------------------------------- --------------- +epoch -180 +replay_buffer/size 999033 +trainer/num train calls 821000 +trainer/Policy Loss -18.4691 +trainer/Log Pis Mean 23.905 +trainer/Log Pis Std 13.5587 +trainer/Log Pis Max 64.443 +trainer/Log Pis Min -11.5181 +trainer/policy/mean Mean -0.0224104 +trainer/policy/mean Std 0.908055 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.77677 +trainer/policy/normal/std Std 0.687128 +trainer/policy/normal/std Max 6.08449 +trainer/policy/normal/std Min 0.248302 +trainer/policy/normal/log_std Mean 0.977361 +trainer/policy/normal/log_std Std 0.334867 +trainer/policy/normal/log_std Max 1.80574 +trainer/policy/normal/log_std Min -1.39311 +eval/num steps total 820081 +eval/num paths total 821 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.218348 +eval/Actions Std 0.92601 +eval/Actions Max 0.999984 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62446 +time/logging (s) 0.00380602 +time/sampling batch (s) 1.07485 +time/saving (s) 0.00361514 +time/training (s) 7.55453 +time/epoch (s) 11.2613 +time/total (s) 8420.41 +Epoch -180 +---------------------------------- --------------- +2022-05-10 15:31:20.557379 PDT | [2] Epoch -179 finished +---------------------------------- --------------- +epoch -179 +replay_buffer/size 999033 +trainer/num train calls 822000 +trainer/Policy Loss -19.5003 +trainer/Log Pis Mean 24.7611 +trainer/Log Pis Std 12.897 +trainer/Log Pis Max 65.2603 +trainer/Log Pis Min -6.41537 +trainer/policy/mean Mean -0.0274235 +trainer/policy/mean Std 0.906768 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.83047 +trainer/policy/normal/std Std 0.686686 +trainer/policy/normal/std Max 5.85944 +trainer/policy/normal/std Min 0.278441 +trainer/policy/normal/log_std Mean 0.996094 +trainer/policy/normal/log_std Std 0.340595 +trainer/policy/normal/log_std Max 1.76805 +trainer/policy/normal/log_std Min -1.27855 +eval/num steps total 821081 +eval/num paths total 822 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0521533 +eval/Actions Std 0.901173 +eval/Actions Max 0.999998 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.7995 +time/logging (s) 0.00408472 +time/sampling batch (s) 0.575478 +time/saving (s) 0.00382494 +time/training (s) 7.8798 +time/epoch (s) 11.2627 +time/total (s) 8431.68 +Epoch -179 +---------------------------------- --------------- +2022-05-10 15:31:30.821952 PDT | [2] Epoch -178 finished +---------------------------------- --------------- +epoch -178 +replay_buffer/size 999033 +trainer/num train calls 823000 +trainer/Policy Loss -20.907 +trainer/Log Pis Mean 25.7804 +trainer/Log Pis Std 13.4568 +trainer/Log Pis Max 62.9126 +trainer/Log Pis Min -7.51596 +trainer/policy/mean Mean -0.0404981 +trainer/policy/mean Std 0.903595 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999984 +trainer/policy/normal/std Mean 2.82648 +trainer/policy/normal/std Std 0.673673 +trainer/policy/normal/std Max 6.82301 +trainer/policy/normal/std Min 0.263954 +trainer/policy/normal/log_std Mean 0.997583 +trainer/policy/normal/log_std Std 0.327385 +trainer/policy/normal/log_std Max 1.9203 +trainer/policy/normal/log_std Min -1.33198 +eval/num steps total 822081 +eval/num paths total 823 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.136772 +eval/Actions Std 0.877848 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.75971 +time/logging (s) 0.00389174 +time/sampling batch (s) 0.524528 +time/saving (s) 0.00346608 +time/training (s) 6.94875 +time/epoch (s) 10.2404 +time/total (s) 8441.92 +Epoch -178 +---------------------------------- --------------- +2022-05-10 15:31:41.136496 PDT | [2] Epoch -177 finished +---------------------------------- -------------- +epoch -177 +replay_buffer/size 999033 +trainer/num train calls 824000 +trainer/Policy Loss -19.9826 +trainer/Log Pis Mean 24.9896 +trainer/Log Pis Std 13.1189 +trainer/Log Pis Max 74.4181 +trainer/Log Pis Min -8.42143 +trainer/policy/mean Mean -0.0193312 +trainer/policy/mean Std 0.905918 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.81534 +trainer/policy/normal/std Std 0.66117 +trainer/policy/normal/std Max 6.85468 +trainer/policy/normal/std Min 0.236151 +trainer/policy/normal/log_std Mean 0.995509 +trainer/policy/normal/log_std Std 0.31805 +trainer/policy/normal/log_std Max 1.92493 +trainer/policy/normal/log_std Min -1.44329 +eval/num steps total 823081 +eval/num paths total 824 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.081122 +eval/Actions Std 0.917046 +eval/Actions Max 0.999998 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54892 +time/logging (s) 0.0037226 +time/sampling batch (s) 0.274396 +time/saving (s) 0.0033876 +time/training (s) 7.45997 +time/epoch (s) 10.2904 +time/total (s) 8452.21 +Epoch -177 +---------------------------------- -------------- +2022-05-10 15:31:51.708119 PDT | [2] Epoch -176 finished +---------------------------------- --------------- +epoch -176 +replay_buffer/size 999033 +trainer/num train calls 825000 +trainer/Policy Loss -19.8711 +trainer/Log Pis Mean 25.6471 +trainer/Log Pis Std 13.6373 +trainer/Log Pis Max 82.6026 +trainer/Log Pis Min -8.24879 +trainer/policy/mean Mean -0.0284384 +trainer/policy/mean Std 0.904782 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.84454 +trainer/policy/normal/std Std 0.677532 +trainer/policy/normal/std Max 6.94831 +trainer/policy/normal/std Min 0.283133 +trainer/policy/normal/log_std Mean 1.0038 +trainer/policy/normal/log_std Std 0.328246 +trainer/policy/normal/log_std Max 1.9385 +trainer/policy/normal/log_std Min -1.26184 +eval/num steps total 824081 +eval/num paths total 825 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.115831 +eval/Actions Std 0.891725 +eval/Actions Max 0.999996 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56555 +time/logging (s) 0.00424013 +time/sampling batch (s) 0.79862 +time/saving (s) 0.00401652 +time/training (s) 7.1758 +time/epoch (s) 10.5482 +time/total (s) 8462.76 +Epoch -176 +---------------------------------- --------------- +2022-05-10 15:32:02.261177 PDT | [2] Epoch -175 finished +---------------------------------- --------------- +epoch -175 +replay_buffer/size 999033 +trainer/num train calls 826000 +trainer/Policy Loss -19.4614 +trainer/Log Pis Mean 24.4622 +trainer/Log Pis Std 13.7618 +trainer/Log Pis Max 63.1449 +trainer/Log Pis Min -8.763 +trainer/policy/mean Mean -0.0304188 +trainer/policy/mean Std 0.905233 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.80331 +trainer/policy/normal/std Std 0.710059 +trainer/policy/normal/std Max 6.29118 +trainer/policy/normal/std Min 0.284783 +trainer/policy/normal/log_std Mean 0.982667 +trainer/policy/normal/log_std Std 0.354641 +trainer/policy/normal/log_std Max 1.83915 +trainer/policy/normal/log_std Min -1.25603 +eval/num steps total 825081 +eval/num paths total 826 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0784083 +eval/Actions Std 0.916888 +eval/Actions Max 0.999996 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.98314 +time/logging (s) 0.00404419 +time/sampling batch (s) 0.325791 +time/saving (s) 0.00373081 +time/training (s) 7.21136 +time/epoch (s) 10.5281 +time/total (s) 8473.3 +Epoch -175 +---------------------------------- --------------- +2022-05-10 15:32:12.774391 PDT | [2] Epoch -174 finished +---------------------------------- --------------- +epoch -174 +replay_buffer/size 999033 +trainer/num train calls 827000 +trainer/Policy Loss -19.9087 +trainer/Log Pis Mean 24.2536 +trainer/Log Pis Std 13.2973 +trainer/Log Pis Max 80.086 +trainer/Log Pis Min -8.17459 +trainer/policy/mean Mean -0.0301187 +trainer/policy/mean Std 0.91 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.77381 +trainer/policy/normal/std Std 0.665721 +trainer/policy/normal/std Max 5.47418 +trainer/policy/normal/std Min 0.247854 +trainer/policy/normal/log_std Mean 0.978563 +trainer/policy/normal/log_std Std 0.326626 +trainer/policy/normal/log_std Max 1.70004 +trainer/policy/normal/log_std Min -1.39491 +eval/num steps total 826081 +eval/num paths total 827 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0225322 +eval/Actions Std 0.895205 +eval/Actions Max 0.999994 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59429 +time/logging (s) 0.00400429 +time/sampling batch (s) 0.525437 +time/saving (s) 0.00370292 +time/training (s) 7.362 +time/epoch (s) 10.4894 +time/total (s) 8483.79 +Epoch -174 +---------------------------------- --------------- +2022-05-10 15:32:22.649358 PDT | [2] Epoch -173 finished +---------------------------------- --------------- +epoch -173 +replay_buffer/size 999033 +trainer/num train calls 828000 +trainer/Policy Loss -19.9165 +trainer/Log Pis Mean 23.3903 +trainer/Log Pis Std 13.4438 +trainer/Log Pis Max 66.4541 +trainer/Log Pis Min -8.49681 +trainer/policy/mean Mean -0.0339036 +trainer/policy/mean Std 0.909674 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.84274 +trainer/policy/normal/std Std 0.686478 +trainer/policy/normal/std Max 6.3752 +trainer/policy/normal/std Min 0.273448 +trainer/policy/normal/log_std Mean 1.00334 +trainer/policy/normal/log_std Std 0.324976 +trainer/policy/normal/log_std Max 1.85242 +trainer/policy/normal/log_std Min -1.29664 +eval/num steps total 827081 +eval/num paths total 828 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.165209 +eval/Actions Std 0.897511 +eval/Actions Max 0.999996 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.77058 +time/logging (s) 0.0037488 +time/sampling batch (s) 0.525651 +time/saving (s) 0.00356148 +time/training (s) 6.54751 +time/epoch (s) 9.85105 +time/total (s) 8493.64 +Epoch -173 +---------------------------------- --------------- +2022-05-10 15:32:31.924647 PDT | [2] Epoch -172 finished +---------------------------------- --------------- +epoch -172 +replay_buffer/size 999033 +trainer/num train calls 829000 +trainer/Policy Loss -19.4556 +trainer/Log Pis Mean 23.6224 +trainer/Log Pis Std 14.1414 +trainer/Log Pis Max 77.4869 +trainer/Log Pis Min -10.9676 +trainer/policy/mean Mean -0.0429912 +trainer/policy/mean Std 0.901546 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.75415 +trainer/policy/normal/std Std 0.67933 +trainer/policy/normal/std Max 5.69657 +trainer/policy/normal/std Min 0.275167 +trainer/policy/normal/log_std Mean 0.969356 +trainer/policy/normal/log_std Std 0.333454 +trainer/policy/normal/log_std Max 1.73986 +trainer/policy/normal/log_std Min -1.29038 +eval/num steps total 828081 +eval/num paths total 829 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0932449 +eval/Actions Std 0.917813 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59744 +time/logging (s) 0.00410028 +time/sampling batch (s) 0.276643 +time/saving (s) 0.00375417 +time/training (s) 6.37002 +time/epoch (s) 9.25196 +time/total (s) 8502.9 +Epoch -172 +---------------------------------- --------------- +2022-05-10 15:32:42.142808 PDT | [2] Epoch -171 finished +---------------------------------- --------------- +epoch -171 +replay_buffer/size 999033 +trainer/num train calls 830000 +trainer/Policy Loss -19.7999 +trainer/Log Pis Mean 24.4531 +trainer/Log Pis Std 13.6108 +trainer/Log Pis Max 70.2156 +trainer/Log Pis Min -15.7864 +trainer/policy/mean Mean -0.0506529 +trainer/policy/mean Std 0.906933 +trainer/policy/mean Max 0.999976 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.78398 +trainer/policy/normal/std Std 0.687815 +trainer/policy/normal/std Max 7.30266 +trainer/policy/normal/std Min 0.291229 +trainer/policy/normal/log_std Mean 0.978209 +trainer/policy/normal/log_std Std 0.34541 +trainer/policy/normal/log_std Max 1.98824 +trainer/policy/normal/log_std Min -1.23365 +eval/num steps total 829081 +eval/num paths total 830 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.202318 +eval/Actions Std 0.932466 +eval/Actions Max 0.999996 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.78538 +time/logging (s) 0.00376826 +time/sampling batch (s) 0.277334 +time/saving (s) 0.00337778 +time/training (s) 7.12392 +time/epoch (s) 10.1938 +time/total (s) 8513.09 +Epoch -171 +---------------------------------- --------------- +2022-05-10 15:32:51.722109 PDT | [2] Epoch -170 finished +---------------------------------- --------------- +epoch -170 +replay_buffer/size 999033 +trainer/num train calls 831000 +trainer/Policy Loss -21.0015 +trainer/Log Pis Mean 24.9885 +trainer/Log Pis Std 13.6323 +trainer/Log Pis Max 77.558 +trainer/Log Pis Min -7.74336 +trainer/policy/mean Mean -0.0204259 +trainer/policy/mean Std 0.912452 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.75504 +trainer/policy/normal/std Std 0.628842 +trainer/policy/normal/std Max 5.99411 +trainer/policy/normal/std Min 0.314758 +trainer/policy/normal/log_std Mean 0.976101 +trainer/policy/normal/log_std Std 0.30838 +trainer/policy/normal/log_std Max 1.79078 +trainer/policy/normal/log_std Min -1.15595 +eval/num steps total 830081 +eval/num paths total 831 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.237096 +eval/Actions Std 0.893052 +eval/Actions Max 0.99998 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60073 +time/logging (s) 0.00379027 +time/sampling batch (s) 0.276112 +time/saving (s) 0.00338157 +time/training (s) 6.67165 +time/epoch (s) 9.55566 +time/total (s) 8522.65 +Epoch -170 +---------------------------------- --------------- +2022-05-10 15:33:01.789252 PDT | [2] Epoch -169 finished +---------------------------------- --------------- +epoch -169 +replay_buffer/size 999033 +trainer/num train calls 832000 +trainer/Policy Loss -19.6804 +trainer/Log Pis Mean 23.6132 +trainer/Log Pis Std 14.0613 +trainer/Log Pis Max 78.954 +trainer/Log Pis Min -6.76725 +trainer/policy/mean Mean -0.0104319 +trainer/policy/mean Std 0.899962 +trainer/policy/mean Max 0.999982 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.83337 +trainer/policy/normal/std Std 0.699267 +trainer/policy/normal/std Max 6.21121 +trainer/policy/normal/std Min 0.222536 +trainer/policy/normal/log_std Mean 0.996745 +trainer/policy/normal/log_std Std 0.340283 +trainer/policy/normal/log_std Max 1.82636 +trainer/policy/normal/log_std Min -1.50267 +eval/num steps total 831081 +eval/num paths total 832 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00537688 +eval/Actions Std 0.902169 +eval/Actions Max 0.999996 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.48413 +time/logging (s) 0.0037975 +time/sampling batch (s) 0.280963 +time/saving (s) 0.0034074 +time/training (s) 7.27102 +time/epoch (s) 10.0433 +time/total (s) 8532.7 +Epoch -169 +---------------------------------- --------------- +2022-05-10 15:33:11.209670 PDT | [2] Epoch -168 finished +---------------------------------- --------------- +epoch -168 +replay_buffer/size 999033 +trainer/num train calls 833000 +trainer/Policy Loss -20.3025 +trainer/Log Pis Mean 24.5896 +trainer/Log Pis Std 12.7851 +trainer/Log Pis Max 68.4324 +trainer/Log Pis Min -10.4439 +trainer/policy/mean Mean -0.0223279 +trainer/policy/mean Std 0.907165 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.80337 +trainer/policy/normal/std Std 0.676069 +trainer/policy/normal/std Max 5.73074 +trainer/policy/normal/std Min 0.284445 +trainer/policy/normal/log_std Mean 0.987163 +trainer/policy/normal/log_std Std 0.338145 +trainer/policy/normal/log_std Max 1.74585 +trainer/policy/normal/log_std Min -1.25721 +eval/num steps total 832081 +eval/num paths total 833 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.098023 +eval/Actions Std 0.907205 +eval/Actions Max 0.999993 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.34654 +time/logging (s) 0.00395777 +time/sampling batch (s) 0.525588 +time/saving (s) 0.00361591 +time/training (s) 6.51724 +time/epoch (s) 9.39695 +time/total (s) 8542.1 +Epoch -168 +---------------------------------- --------------- +2022-05-10 15:33:20.752318 PDT | [2] Epoch -167 finished +---------------------------------- --------------- +epoch -167 +replay_buffer/size 999033 +trainer/num train calls 834000 +trainer/Policy Loss -20.8192 +trainer/Log Pis Mean 24.0558 +trainer/Log Pis Std 12.8736 +trainer/Log Pis Max 62.8978 +trainer/Log Pis Min -9.4228 +trainer/policy/mean Mean -0.0239508 +trainer/policy/mean Std 0.905942 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.77743 +trainer/policy/normal/std Std 0.684377 +trainer/policy/normal/std Max 6.62741 +trainer/policy/normal/std Min 0.247908 +trainer/policy/normal/log_std Mean 0.97575 +trainer/policy/normal/log_std Std 0.346858 +trainer/policy/normal/log_std Max 1.89121 +trainer/policy/normal/log_std Min -1.3947 +eval/num steps total 833081 +eval/num paths total 834 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.535806 +eval/Actions Std 0.768862 +eval/Actions Max 0.999987 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70219 +time/logging (s) 0.00381828 +time/sampling batch (s) 0.279672 +time/saving (s) 0.00348978 +time/training (s) 6.52915 +time/epoch (s) 9.51832 +time/total (s) 8551.62 +Epoch -167 +---------------------------------- --------------- +2022-05-10 15:33:31.623282 PDT | [2] Epoch -166 finished +---------------------------------- --------------- +epoch -166 +replay_buffer/size 999033 +trainer/num train calls 835000 +trainer/Policy Loss -19.6963 +trainer/Log Pis Mean 23.8754 +trainer/Log Pis Std 12.9793 +trainer/Log Pis Max 67.3043 +trainer/Log Pis Min -7.77245 +trainer/policy/mean Mean -0.0431579 +trainer/policy/mean Std 0.907679 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.79894 +trainer/policy/normal/std Std 0.656908 +trainer/policy/normal/std Max 6.11296 +trainer/policy/normal/std Min 0.229681 +trainer/policy/normal/log_std Mean 0.989361 +trainer/policy/normal/log_std Std 0.320485 +trainer/policy/normal/log_std Max 1.81041 +trainer/policy/normal/log_std Min -1.47106 +eval/num steps total 834081 +eval/num paths total 835 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0203985 +eval/Actions Std 0.930597 +eval/Actions Max 0.999994 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51648 +time/logging (s) 0.00398227 +time/sampling batch (s) 0.302737 +time/saving (s) 0.00357828 +time/training (s) 8.01992 +time/epoch (s) 10.8467 +time/total (s) 8562.47 +Epoch -166 +---------------------------------- --------------- +2022-05-10 15:33:42.750958 PDT | [2] Epoch -165 finished +---------------------------------- --------------- +epoch -165 +replay_buffer/size 999033 +trainer/num train calls 836000 +trainer/Policy Loss -20.4346 +trainer/Log Pis Mean 24.4255 +trainer/Log Pis Std 12.6715 +trainer/Log Pis Max 64.6754 +trainer/Log Pis Min -6.24234 +trainer/policy/mean Mean -0.0272172 +trainer/policy/mean Std 0.906848 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.84335 +trainer/policy/normal/std Std 0.684119 +trainer/policy/normal/std Max 5.95925 +trainer/policy/normal/std Min 0.304877 +trainer/policy/normal/log_std Mean 1.004 +trainer/policy/normal/log_std Std 0.322121 +trainer/policy/normal/log_std Max 1.78494 +trainer/policy/normal/log_std Min -1.18785 +eval/num steps total 835081 +eval/num paths total 836 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0139099 +eval/Actions Std 0.907506 +eval/Actions Max 0.999998 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.71836 +time/logging (s) 0.00374478 +time/sampling batch (s) 0.820259 +time/saving (s) 0.00335661 +time/training (s) 7.55745 +time/epoch (s) 11.1032 +time/total (s) 8573.58 +Epoch -165 +---------------------------------- --------------- +2022-05-10 15:33:53.359705 PDT | [2] Epoch -164 finished +---------------------------------- --------------- +epoch -164 +replay_buffer/size 999033 +trainer/num train calls 837000 +trainer/Policy Loss -18.7754 +trainer/Log Pis Mean 24.4574 +trainer/Log Pis Std 13.0583 +trainer/Log Pis Max 73.7678 +trainer/Log Pis Min -7.29324 +trainer/policy/mean Mean -0.0296722 +trainer/policy/mean Std 0.908257 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.74282 +trainer/policy/normal/std Std 0.66357 +trainer/policy/normal/std Max 6.11585 +trainer/policy/normal/std Min 0.310835 +trainer/policy/normal/log_std Mean 0.965982 +trainer/policy/normal/log_std Std 0.333276 +trainer/policy/normal/log_std Max 1.81088 +trainer/policy/normal/log_std Min -1.16849 +eval/num steps total 836081 +eval/num paths total 837 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0686165 +eval/Actions Std 0.912548 +eval/Actions Max 0.999997 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57313 +time/logging (s) 0.0037163 +time/sampling batch (s) 0.278221 +time/saving (s) 0.00332429 +time/training (s) 7.72674 +time/epoch (s) 10.5851 +time/total (s) 8584.16 +Epoch -164 +---------------------------------- --------------- +2022-05-10 15:34:03.605439 PDT | [2] Epoch -163 finished +---------------------------------- --------------- +epoch -163 +replay_buffer/size 999033 +trainer/num train calls 838000 +trainer/Policy Loss -18.9269 +trainer/Log Pis Mean 24.5759 +trainer/Log Pis Std 12.7629 +trainer/Log Pis Max 61.6729 +trainer/Log Pis Min -7.59333 +trainer/policy/mean Mean -0.0366718 +trainer/policy/mean Std 0.907393 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.84492 +trainer/policy/normal/std Std 0.68388 +trainer/policy/normal/std Max 5.82688 +trainer/policy/normal/std Min 0.249816 +trainer/policy/normal/log_std Mean 1.00317 +trainer/policy/normal/log_std Std 0.33078 +trainer/policy/normal/log_std Max 1.76248 +trainer/policy/normal/log_std Min -1.38703 +eval/num steps total 837081 +eval/num paths total 838 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0139728 +eval/Actions Std 0.901916 +eval/Actions Max 0.999992 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63167 +time/logging (s) 0.00400347 +time/sampling batch (s) 0.270519 +time/saving (s) 0.00362521 +time/training (s) 7.31289 +time/epoch (s) 10.2227 +time/total (s) 8594.39 +Epoch -163 +---------------------------------- --------------- +2022-05-10 15:34:15.274215 PDT | [2] Epoch -162 finished +---------------------------------- --------------- +epoch -162 +replay_buffer/size 999033 +trainer/num train calls 839000 +trainer/Policy Loss -19.9315 +trainer/Log Pis Mean 23.8318 +trainer/Log Pis Std 12.5291 +trainer/Log Pis Max 68.5679 +trainer/Log Pis Min -4.27384 +trainer/policy/mean Mean -0.0395323 +trainer/policy/mean Std 0.907719 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.79859 +trainer/policy/normal/std Std 0.667732 +trainer/policy/normal/std Max 5.37056 +trainer/policy/normal/std Min 0.27578 +trainer/policy/normal/log_std Mean 0.986749 +trainer/policy/normal/log_std Std 0.331519 +trainer/policy/normal/log_std Max 1.68093 +trainer/policy/normal/log_std Min -1.28815 +eval/num steps total 838081 +eval/num paths total 839 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.290911 +eval/Actions Std 0.799526 +eval/Actions Max 0.999993 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57426 +time/logging (s) 0.00364236 +time/sampling batch (s) 0.276673 +time/saving (s) 0.00335956 +time/training (s) 8.78651 +time/epoch (s) 11.6445 +time/total (s) 8606.04 +Epoch -162 +---------------------------------- --------------- +2022-05-10 15:34:26.275244 PDT | [2] Epoch -161 finished +---------------------------------- --------------- +epoch -161 +replay_buffer/size 999033 +trainer/num train calls 840000 +trainer/Policy Loss -20.9306 +trainer/Log Pis Mean 25.2491 +trainer/Log Pis Std 13.0646 +trainer/Log Pis Max 65.158 +trainer/Log Pis Min -7.36756 +trainer/policy/mean Mean -0.0324249 +trainer/policy/mean Std 0.909466 +trainer/policy/mean Max 0.999979 +trainer/policy/mean Min -0.999982 +trainer/policy/normal/std Mean 2.82745 +trainer/policy/normal/std Std 0.682815 +trainer/policy/normal/std Max 6.27277 +trainer/policy/normal/std Min 0.26611 +trainer/policy/normal/log_std Mean 0.995322 +trainer/policy/normal/log_std Std 0.341069 +trainer/policy/normal/log_std Max 1.83622 +trainer/policy/normal/log_std Min -1.32385 +eval/num steps total 839081 +eval/num paths total 840 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.110674 +eval/Actions Std 0.886396 +eval/Actions Max 0.999995 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.76459 +time/logging (s) 0.00368568 +time/sampling batch (s) 0.524452 +time/saving (s) 0.00331467 +time/training (s) 7.68152 +time/epoch (s) 10.9776 +time/total (s) 8617.02 +Epoch -161 +---------------------------------- --------------- +2022-05-10 15:34:38.533682 PDT | [2] Epoch -160 finished +---------------------------------- --------------- +epoch -160 +replay_buffer/size 999033 +trainer/num train calls 841000 +trainer/Policy Loss -20.1682 +trainer/Log Pis Mean 23.7551 +trainer/Log Pis Std 12.8582 +trainer/Log Pis Max 64.5321 +trainer/Log Pis Min -3.67301 +trainer/policy/mean Mean -0.0228609 +trainer/policy/mean Std 0.905501 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.81236 +trainer/policy/normal/std Std 0.679257 +trainer/policy/normal/std Max 6.31213 +trainer/policy/normal/std Min 0.227487 +trainer/policy/normal/log_std Mean 0.990718 +trainer/policy/normal/log_std Std 0.336225 +trainer/policy/normal/log_std Max 1.84247 +trainer/policy/normal/log_std Min -1.48066 +eval/num steps total 840030 +eval/num paths total 841 +eval/path length Mean 949 +eval/path length Std 0 +eval/path length Max 949 +eval/path length Min 949 +eval/Rewards Mean 0.00105374 +eval/Rewards Std 0.0324443 +eval/Rewards Max 1 +eval/Rewards Min 0 +eval/Returns Mean 1 +eval/Returns Std 0 +eval/Returns Max 1 +eval/Returns Min 1 +eval/Actions Mean 0.0156874 +eval/Actions Std 0.90777 +eval/Actions Max 0.999998 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 1 +eval/normalized_score 100 +time/evaluation sampling (s) 2.62922 +time/logging (s) 0.00356815 +time/sampling batch (s) 0.776781 +time/saving (s) 0.00334693 +time/training (s) 8.82167 +time/epoch (s) 12.2346 +time/total (s) 8629.26 +Epoch -160 +---------------------------------- --------------- +2022-05-10 15:34:49.793813 PDT | [2] Epoch -159 finished +---------------------------------- --------------- +epoch -159 +replay_buffer/size 999033 +trainer/num train calls 842000 +trainer/Policy Loss -19.9557 +trainer/Log Pis Mean 24.5775 +trainer/Log Pis Std 13.0922 +trainer/Log Pis Max 63.01 +trainer/Log Pis Min -7.24596 +trainer/policy/mean Mean -0.037683 +trainer/policy/mean Std 0.911105 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.78807 +trainer/policy/normal/std Std 0.647228 +trainer/policy/normal/std Max 5.80541 +trainer/policy/normal/std Min 0.304291 +trainer/policy/normal/log_std Mean 0.986895 +trainer/policy/normal/log_std Std 0.312743 +trainer/policy/normal/log_std Max 1.75879 +trainer/policy/normal/log_std Min -1.18977 +eval/num steps total 841030 +eval/num paths total 842 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.435647 +eval/Actions Std 0.861909 +eval/Actions Max 0.999995 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62371 +time/logging (s) 0.00372841 +time/sampling batch (s) 0.272027 +time/saving (s) 0.00332475 +time/training (s) 8.33413 +time/epoch (s) 11.2369 +time/total (s) 8640.5 +Epoch -159 +---------------------------------- --------------- +2022-05-10 15:34:59.579682 PDT | [2] Epoch -158 finished +---------------------------------- --------------- +epoch -158 +replay_buffer/size 999033 +trainer/num train calls 843000 +trainer/Policy Loss -19.8862 +trainer/Log Pis Mean 24.3635 +trainer/Log Pis Std 13.3332 +trainer/Log Pis Max 70.341 +trainer/Log Pis Min -13.8301 +trainer/policy/mean Mean -0.0347182 +trainer/policy/mean Std 0.908431 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.81807 +trainer/policy/normal/std Std 0.659572 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.285726 +trainer/policy/normal/log_std Mean 0.996906 +trainer/policy/normal/log_std Std 0.317004 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.25272 +eval/num steps total 842030 +eval/num paths total 843 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0149172 +eval/Actions Std 0.908706 +eval/Actions Max 0.999996 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55142 +time/logging (s) 0.00403215 +time/sampling batch (s) 0.271542 +time/saving (s) 0.00372441 +time/training (s) 6.9321 +time/epoch (s) 9.76282 +time/total (s) 8650.26 +Epoch -158 +---------------------------------- --------------- +2022-05-10 15:35:10.811214 PDT | [2] Epoch -157 finished +---------------------------------- --------------- +epoch -157 +replay_buffer/size 999033 +trainer/num train calls 844000 +trainer/Policy Loss -19.309 +trainer/Log Pis Mean 25.5586 +trainer/Log Pis Std 13.6904 +trainer/Log Pis Max 69.0906 +trainer/Log Pis Min -8.44634 +trainer/policy/mean Mean -0.0512052 +trainer/policy/mean Std 0.906236 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.87448 +trainer/policy/normal/std Std 0.682321 +trainer/policy/normal/std Max 7.11397 +trainer/policy/normal/std Min 0.375188 +trainer/policy/normal/log_std Mean 1.01583 +trainer/policy/normal/log_std Std 0.318465 +trainer/policy/normal/log_std Max 1.96206 +trainer/policy/normal/log_std Min -0.980327 +eval/num steps total 843030 +eval/num paths total 844 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.202087 +eval/Actions Std 0.944471 +eval/Actions Max 0.999998 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.36481 +time/logging (s) 0.00375425 +time/sampling batch (s) 0.52683 +time/saving (s) 0.00339987 +time/training (s) 8.30854 +time/epoch (s) 11.2073 +time/total (s) 8661.47 +Epoch -157 +---------------------------------- --------------- +2022-05-10 15:35:21.367282 PDT | [2] Epoch -156 finished +---------------------------------- --------------- +epoch -156 +replay_buffer/size 999033 +trainer/num train calls 845000 +trainer/Policy Loss -19.4062 +trainer/Log Pis Mean 24.594 +trainer/Log Pis Std 13.4923 +trainer/Log Pis Max 77.9342 +trainer/Log Pis Min -11.7693 +trainer/policy/mean Mean -0.0127653 +trainer/policy/mean Std 0.907763 +trainer/policy/mean Max 0.999981 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.79274 +trainer/policy/normal/std Std 0.655708 +trainer/policy/normal/std Max 5.62838 +trainer/policy/normal/std Min 0.325065 +trainer/policy/normal/log_std Mean 0.98657 +trainer/policy/normal/log_std Std 0.321939 +trainer/policy/normal/log_std Max 1.72782 +trainer/policy/normal/log_std Min -1.12373 +eval/num steps total 844030 +eval/num paths total 845 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.139869 +eval/Actions Std 0.87432 +eval/Actions Max 0.999998 +eval/Actions Min -0.999987 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53644 +time/logging (s) 0.00391339 +time/sampling batch (s) 0.275824 +time/saving (s) 0.00344926 +time/training (s) 7.71301 +time/epoch (s) 10.5326 +time/total (s) 8672.01 +Epoch -156 +---------------------------------- --------------- +2022-05-10 15:35:32.296300 PDT | [2] Epoch -155 finished +---------------------------------- --------------- +epoch -155 +replay_buffer/size 999033 +trainer/num train calls 846000 +trainer/Policy Loss -19.5288 +trainer/Log Pis Mean 23.3187 +trainer/Log Pis Std 13.2456 +trainer/Log Pis Max 60.6685 +trainer/Log Pis Min -7.57146 +trainer/policy/mean Mean -0.0114657 +trainer/policy/mean Std 0.904059 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.82497 +trainer/policy/normal/std Std 0.680551 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.294833 +trainer/policy/normal/log_std Mean 0.997054 +trainer/policy/normal/log_std Std 0.324678 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.22135 +eval/num steps total 845030 +eval/num paths total 846 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.136826 +eval/Actions Std 0.881218 +eval/Actions Max 0.999983 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59563 +time/logging (s) 0.00380229 +time/sampling batch (s) 0.527378 +time/saving (s) 0.00336425 +time/training (s) 7.77488 +time/epoch (s) 10.905 +time/total (s) 8682.91 +Epoch -155 +---------------------------------- --------------- +2022-05-10 15:35:42.174965 PDT | [2] Epoch -154 finished +---------------------------------- --------------- +epoch -154 +replay_buffer/size 999033 +trainer/num train calls 847000 +trainer/Policy Loss -19.4623 +trainer/Log Pis Mean 25.0682 +trainer/Log Pis Std 14.0458 +trainer/Log Pis Max 85.3714 +trainer/Log Pis Min -12.5207 +trainer/policy/mean Mean -0.0384441 +trainer/policy/mean Std 0.903222 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.83066 +trainer/policy/normal/std Std 0.679639 +trainer/policy/normal/std Max 5.38649 +trainer/policy/normal/std Min 0.218937 +trainer/policy/normal/log_std Mean 0.996673 +trainer/policy/normal/log_std Std 0.340716 +trainer/policy/normal/log_std Max 1.68389 +trainer/policy/normal/log_std Min -1.51897 +eval/num steps total 846030 +eval/num paths total 847 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.389224 +eval/Actions Std 0.856422 +eval/Actions Max 0.999985 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58768 +time/logging (s) 0.00387047 +time/sampling batch (s) 0.773453 +time/saving (s) 0.00349031 +time/training (s) 6.48693 +time/epoch (s) 9.85543 +time/total (s) 8692.77 +Epoch -154 +---------------------------------- --------------- +2022-05-10 15:35:53.102246 PDT | [2] Epoch -153 finished +---------------------------------- --------------- +epoch -153 +replay_buffer/size 999033 +trainer/num train calls 848000 +trainer/Policy Loss -19.3025 +trainer/Log Pis Mean 25.3864 +trainer/Log Pis Std 13.6436 +trainer/Log Pis Max 83.7637 +trainer/Log Pis Min -7.57651 +trainer/policy/mean Mean -0.0175014 +trainer/policy/mean Std 0.901765 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.77336 +trainer/policy/normal/std Std 0.66116 +trainer/policy/normal/std Max 5.22797 +trainer/policy/normal/std Min 0.30283 +trainer/policy/normal/log_std Mean 0.978818 +trainer/policy/normal/log_std Std 0.324161 +trainer/policy/normal/log_std Max 1.65402 +trainer/policy/normal/log_std Min -1.19458 +eval/num steps total 847030 +eval/num paths total 848 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.161401 +eval/Actions Std 0.915066 +eval/Actions Max 0.999992 +eval/Actions Min -0.99998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56839 +time/logging (s) 0.00411553 +time/sampling batch (s) 0.276951 +time/saving (s) 0.00368674 +time/training (s) 8.05045 +time/epoch (s) 10.9036 +time/total (s) 8703.68 +Epoch -153 +---------------------------------- --------------- +2022-05-10 15:36:02.128402 PDT | [2] Epoch -152 finished +---------------------------------- --------------- +epoch -152 +replay_buffer/size 999033 +trainer/num train calls 849000 +trainer/Policy Loss -19.8644 +trainer/Log Pis Mean 24.5189 +trainer/Log Pis Std 13.5726 +trainer/Log Pis Max 71.8389 +trainer/Log Pis Min -10.9971 +trainer/policy/mean Mean -0.0190603 +trainer/policy/mean Std 0.906929 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.80992 +trainer/policy/normal/std Std 0.669091 +trainer/policy/normal/std Max 5.34816 +trainer/policy/normal/std Min 0.305962 +trainer/policy/normal/log_std Mean 0.990974 +trainer/policy/normal/log_std Std 0.329912 +trainer/policy/normal/log_std Max 1.67675 +trainer/policy/normal/log_std Min -1.18429 +eval/num steps total 848030 +eval/num paths total 849 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.139793 +eval/Actions Std 0.944527 +eval/Actions Max 0.999993 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61194 +time/logging (s) 0.00376191 +time/sampling batch (s) 0.27488 +time/saving (s) 0.00353728 +time/training (s) 6.1077 +time/epoch (s) 9.00182 +time/total (s) 8712.68 +Epoch -152 +---------------------------------- --------------- +2022-05-10 15:36:12.409516 PDT | [2] Epoch -151 finished +---------------------------------- --------------- +epoch -151 +replay_buffer/size 999033 +trainer/num train calls 850000 +trainer/Policy Loss -19.8599 +trainer/Log Pis Mean 25.1618 +trainer/Log Pis Std 12.9404 +trainer/Log Pis Max 70.0511 +trainer/Log Pis Min -8.19027 +trainer/policy/mean Mean -0.0412814 +trainer/policy/mean Std 0.903846 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.76322 +trainer/policy/normal/std Std 0.658924 +trainer/policy/normal/std Max 6.04993 +trainer/policy/normal/std Min 0.319393 +trainer/policy/normal/log_std Mean 0.974576 +trainer/policy/normal/log_std Std 0.32867 +trainer/policy/normal/log_std Max 1.80005 +trainer/policy/normal/log_std Min -1.14133 +eval/num steps total 849030 +eval/num paths total 850 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.343968 +eval/Actions Std 0.866511 +eval/Actions Max 0.999993 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60291 +time/logging (s) 0.00365983 +time/sampling batch (s) 0.525158 +time/saving (s) 0.00337068 +time/training (s) 7.12211 +time/epoch (s) 10.2572 +time/total (s) 8722.95 +Epoch -151 +---------------------------------- --------------- +2022-05-10 15:36:22.573611 PDT | [2] Epoch -150 finished +---------------------------------- --------------- +epoch -150 +replay_buffer/size 999033 +trainer/num train calls 851000 +trainer/Policy Loss -19.8885 +trainer/Log Pis Mean 24.2442 +trainer/Log Pis Std 13.1299 +trainer/Log Pis Max 64.1881 +trainer/Log Pis Min -7.66062 +trainer/policy/mean Mean -0.0487984 +trainer/policy/mean Std 0.9135 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.85157 +trainer/policy/normal/std Std 0.652693 +trainer/policy/normal/std Max 5.47455 +trainer/policy/normal/std Min 0.275628 +trainer/policy/normal/log_std Mean 1.009 +trainer/policy/normal/log_std Std 0.317636 +trainer/policy/normal/log_std Max 1.70011 +trainer/policy/normal/log_std Min -1.2887 +eval/num steps total 850030 +eval/num paths total 851 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.107616 +eval/Actions Std 0.89994 +eval/Actions Max 0.999992 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73521 +time/logging (s) 0.00371041 +time/sampling batch (s) 0.274589 +time/saving (s) 0.00334351 +time/training (s) 7.12348 +time/epoch (s) 10.1403 +time/total (s) 8733.09 +Epoch -150 +---------------------------------- --------------- +2022-05-10 15:36:32.950287 PDT | [2] Epoch -149 finished +---------------------------------- --------------- +epoch -149 +replay_buffer/size 999033 +trainer/num train calls 852000 +trainer/Policy Loss -19.5191 +trainer/Log Pis Mean 25.2969 +trainer/Log Pis Std 13.4074 +trainer/Log Pis Max 69.1822 +trainer/Log Pis Min -7.42639 +trainer/policy/mean Mean -0.0256347 +trainer/policy/mean Std 0.906188 +trainer/policy/mean Max 0.999981 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.77925 +trainer/policy/normal/std Std 0.672952 +trainer/policy/normal/std Max 6.77749 +trainer/policy/normal/std Min 0.300293 +trainer/policy/normal/log_std Mean 0.979002 +trainer/policy/normal/log_std Std 0.333409 +trainer/policy/normal/log_std Max 1.91361 +trainer/policy/normal/log_std Min -1.203 +eval/num steps total 851030 +eval/num paths total 852 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.356883 +eval/Actions Std 0.889504 +eval/Actions Max 0.999995 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.72465 +time/logging (s) 0.00367966 +time/sampling batch (s) 0.275146 +time/saving (s) 0.00337665 +time/training (s) 7.34603 +time/epoch (s) 10.3529 +time/total (s) 8743.44 +Epoch -149 +---------------------------------- --------------- +2022-05-10 15:36:42.059523 PDT | [2] Epoch -148 finished +---------------------------------- --------------- +epoch -148 +replay_buffer/size 999033 +trainer/num train calls 853000 +trainer/Policy Loss -19.3972 +trainer/Log Pis Mean 24.2161 +trainer/Log Pis Std 13.5571 +trainer/Log Pis Max 70.1743 +trainer/Log Pis Min -6.89868 +trainer/policy/mean Mean -0.0191136 +trainer/policy/mean Std 0.905949 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.81896 +trainer/policy/normal/std Std 0.678136 +trainer/policy/normal/std Max 7.37631 +trainer/policy/normal/std Min 0.33023 +trainer/policy/normal/log_std Mean 0.994842 +trainer/policy/normal/log_std Std 0.325644 +trainer/policy/normal/log_std Max 1.99827 +trainer/policy/normal/log_std Min -1.10797 +eval/num steps total 852030 +eval/num paths total 853 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.161576 +eval/Actions Std 0.837615 +eval/Actions Max 0.999996 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.52495 +time/logging (s) 0.00426722 +time/sampling batch (s) 0.529105 +time/saving (s) 0.003361 +time/training (s) 6.02438 +time/epoch (s) 9.08606 +time/total (s) 8752.53 +Epoch -148 +---------------------------------- --------------- +2022-05-10 15:36:52.523974 PDT | [2] Epoch -147 finished +---------------------------------- --------------- +epoch -147 +replay_buffer/size 999033 +trainer/num train calls 854000 +trainer/Policy Loss -18.3661 +trainer/Log Pis Mean 25.0254 +trainer/Log Pis Std 13.4001 +trainer/Log Pis Max 80.917 +trainer/Log Pis Min -10.9242 +trainer/policy/mean Mean -0.0258912 +trainer/policy/mean Std 0.91205 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.81194 +trainer/policy/normal/std Std 0.67152 +trainer/policy/normal/std Max 6.41418 +trainer/policy/normal/std Min 0.304109 +trainer/policy/normal/log_std Mean 0.992563 +trainer/policy/normal/log_std Std 0.325211 +trainer/policy/normal/log_std Max 1.85851 +trainer/policy/normal/log_std Min -1.19037 +eval/num steps total 853030 +eval/num paths total 854 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0434438 +eval/Actions Std 0.89781 +eval/Actions Max 0.999996 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60367 +time/logging (s) 0.00389947 +time/sampling batch (s) 0.276213 +time/saving (s) 0.00360069 +time/training (s) 7.5521 +time/epoch (s) 10.4395 +time/total (s) 8762.98 +Epoch -147 +---------------------------------- --------------- +2022-05-10 15:37:02.871765 PDT | [2] Epoch -146 finished +---------------------------------- --------------- +epoch -146 +replay_buffer/size 999033 +trainer/num train calls 855000 +trainer/Policy Loss -19.5701 +trainer/Log Pis Mean 23.4257 +trainer/Log Pis Std 12.6766 +trainer/Log Pis Max 76.8884 +trainer/Log Pis Min -6.14351 +trainer/policy/mean Mean -0.0249109 +trainer/policy/mean Std 0.905775 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.84819 +trainer/policy/normal/std Std 0.700661 +trainer/policy/normal/std Max 6.22912 +trainer/policy/normal/std Min 0.301399 +trainer/policy/normal/log_std Mean 1.00216 +trainer/policy/normal/log_std Std 0.340052 +trainer/policy/normal/log_std Max 1.82924 +trainer/policy/normal/log_std Min -1.19932 +eval/num steps total 854030 +eval/num paths total 855 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.100284 +eval/Actions Std 0.904635 +eval/Actions Max 0.999998 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.726 +time/logging (s) 0.00368682 +time/sampling batch (s) 0.273902 +time/saving (s) 0.00336822 +time/training (s) 7.31667 +time/epoch (s) 10.3236 +time/total (s) 8773.3 +Epoch -146 +---------------------------------- --------------- +2022-05-10 15:37:12.952339 PDT | [2] Epoch -145 finished +---------------------------------- --------------- +epoch -145 +replay_buffer/size 999033 +trainer/num train calls 856000 +trainer/Policy Loss -20.3786 +trainer/Log Pis Mean 24.8289 +trainer/Log Pis Std 14.1471 +trainer/Log Pis Max 66.3596 +trainer/Log Pis Min -9.20745 +trainer/policy/mean Mean -0.0244498 +trainer/policy/mean Std 0.906638 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.81292 +trainer/policy/normal/std Std 0.67361 +trainer/policy/normal/std Max 5.90395 +trainer/policy/normal/std Min 0.314378 +trainer/policy/normal/log_std Mean 0.992848 +trainer/policy/normal/log_std Std 0.324993 +trainer/policy/normal/log_std Max 1.77562 +trainer/policy/normal/log_std Min -1.15716 +eval/num steps total 855030 +eval/num paths total 856 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.200239 +eval/Actions Std 0.860939 +eval/Actions Max 0.999987 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56243 +time/logging (s) 0.00371029 +time/sampling batch (s) 0.274262 +time/saving (s) 0.00339145 +time/training (s) 7.21288 +time/epoch (s) 10.0567 +time/total (s) 8783.36 +Epoch -145 +---------------------------------- --------------- +2022-05-10 15:37:23.436600 PDT | [2] Epoch -144 finished +---------------------------------- --------------- +epoch -144 +replay_buffer/size 999033 +trainer/num train calls 857000 +trainer/Policy Loss -20.1708 +trainer/Log Pis Mean 23.5376 +trainer/Log Pis Std 13.4267 +trainer/Log Pis Max 69.5366 +trainer/Log Pis Min -7.9636 +trainer/policy/mean Mean -0.0401038 +trainer/policy/mean Std 0.905501 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.73577 +trainer/policy/normal/std Std 0.654227 +trainer/policy/normal/std Max 5.64265 +trainer/policy/normal/std Min 0.305597 +trainer/policy/normal/log_std Mean 0.964548 +trainer/policy/normal/log_std Std 0.327546 +trainer/policy/normal/log_std Max 1.73035 +trainer/policy/normal/log_std Min -1.18549 +eval/num steps total 856030 +eval/num paths total 857 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0776459 +eval/Actions Std 0.922652 +eval/Actions Max 0.999997 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67684 +time/logging (s) 0.00369795 +time/sampling batch (s) 0.274843 +time/saving (s) 0.00339566 +time/training (s) 7.50156 +time/epoch (s) 10.4603 +time/total (s) 8793.83 +Epoch -144 +---------------------------------- --------------- +2022-05-10 15:37:33.127005 PDT | [2] Epoch -143 finished +---------------------------------- --------------- +epoch -143 +replay_buffer/size 999033 +trainer/num train calls 858000 +trainer/Policy Loss -20.5241 +trainer/Log Pis Mean 24.161 +trainer/Log Pis Std 12.8749 +trainer/Log Pis Max 70.7448 +trainer/Log Pis Min -3.6241 +trainer/policy/mean Mean -0.0376522 +trainer/policy/mean Std 0.902344 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.76754 +trainer/policy/normal/std Std 0.684276 +trainer/policy/normal/std Max 6.03277 +trainer/policy/normal/std Min 0.283137 +trainer/policy/normal/log_std Mean 0.971724 +trainer/policy/normal/log_std Std 0.34884 +trainer/policy/normal/log_std Max 1.79721 +trainer/policy/normal/log_std Min -1.26182 +eval/num steps total 857030 +eval/num paths total 858 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.390813 +eval/Actions Std 0.826808 +eval/Actions Max 0.999995 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.46061 +time/logging (s) 0.00434556 +time/sampling batch (s) 0.530464 +time/saving (s) 0.00340497 +time/training (s) 6.66819 +time/epoch (s) 9.66702 +time/total (s) 8803.5 +Epoch -143 +---------------------------------- --------------- +2022-05-10 15:37:43.073264 PDT | [2] Epoch -142 finished +---------------------------------- --------------- +epoch -142 +replay_buffer/size 999033 +trainer/num train calls 859000 +trainer/Policy Loss -20.2632 +trainer/Log Pis Mean 24.9899 +trainer/Log Pis Std 13.5447 +trainer/Log Pis Max 73.8392 +trainer/Log Pis Min -9.48743 +trainer/policy/mean Mean -0.0516069 +trainer/policy/mean Std 0.904284 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.85211 +trainer/policy/normal/std Std 0.705583 +trainer/policy/normal/std Max 6.58186 +trainer/policy/normal/std Min 0.282559 +trainer/policy/normal/log_std Mean 1.00427 +trainer/policy/normal/log_std Std 0.334418 +trainer/policy/normal/log_std Max 1.88432 +trainer/policy/normal/log_std Min -1.26387 +eval/num steps total 858030 +eval/num paths total 859 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.035093 +eval/Actions Std 0.911944 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65062 +time/logging (s) 0.00402627 +time/sampling batch (s) 0.281304 +time/saving (s) 0.00363082 +time/training (s) 6.98147 +time/epoch (s) 9.92105 +time/total (s) 8813.42 +Epoch -142 +---------------------------------- --------------- +2022-05-10 15:37:52.528539 PDT | [2] Epoch -141 finished +---------------------------------- --------------- +epoch -141 +replay_buffer/size 999033 +trainer/num train calls 860000 +trainer/Policy Loss -19.9635 +trainer/Log Pis Mean 23.9059 +trainer/Log Pis Std 12.7554 +trainer/Log Pis Max 62.8157 +trainer/Log Pis Min -12.404 +trainer/policy/mean Mean -0.0282662 +trainer/policy/mean Std 0.910512 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999973 +trainer/policy/normal/std Mean 2.83786 +trainer/policy/normal/std Std 0.686759 +trainer/policy/normal/std Max 7.0531 +trainer/policy/normal/std Min 0.259945 +trainer/policy/normal/log_std Mean 1.00129 +trainer/policy/normal/log_std Std 0.32633 +trainer/policy/normal/log_std Max 1.95347 +trainer/policy/normal/log_std Min -1.34728 +eval/num steps total 859030 +eval/num paths total 860 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0529251 +eval/Actions Std 0.89671 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.4838 +time/logging (s) 0.00382923 +time/sampling batch (s) 0.776117 +time/saving (s) 0.00350595 +time/training (s) 6.16384 +time/epoch (s) 9.4311 +time/total (s) 8822.85 +Epoch -141 +---------------------------------- --------------- +2022-05-10 15:38:02.395864 PDT | [2] Epoch -140 finished +---------------------------------- --------------- +epoch -140 +replay_buffer/size 999033 +trainer/num train calls 861000 +trainer/Policy Loss -19.827 +trainer/Log Pis Mean 23.913 +trainer/Log Pis Std 13.1039 +trainer/Log Pis Max 73.9641 +trainer/Log Pis Min -9.77637 +trainer/policy/mean Mean -0.0365328 +trainer/policy/mean Std 0.905312 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.73013 +trainer/policy/normal/std Std 0.654641 +trainer/policy/normal/std Max 5.44084 +trainer/policy/normal/std Min 0.258391 +trainer/policy/normal/log_std Mean 0.961833 +trainer/policy/normal/log_std Std 0.330817 +trainer/policy/normal/log_std Max 1.69393 +trainer/policy/normal/log_std Min -1.35328 +eval/num steps total 860030 +eval/num paths total 861 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.17219 +eval/Actions Std 0.913922 +eval/Actions Max 0.999999 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61843 +time/logging (s) 0.00369292 +time/sampling batch (s) 0.276427 +time/saving (s) 0.00332442 +time/training (s) 6.94144 +time/epoch (s) 9.84331 +time/total (s) 8832.7 +Epoch -140 +---------------------------------- --------------- +2022-05-10 15:38:12.727511 PDT | [2] Epoch -139 finished +---------------------------------- --------------- +epoch -139 +replay_buffer/size 999033 +trainer/num train calls 862000 +trainer/Policy Loss -20.9091 +trainer/Log Pis Mean 25.3163 +trainer/Log Pis Std 13.6492 +trainer/Log Pis Max 69.4046 +trainer/Log Pis Min -11.0646 +trainer/policy/mean Mean -0.0461707 +trainer/policy/mean Std 0.906588 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.78426 +trainer/policy/normal/std Std 0.668775 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.274511 +trainer/policy/normal/log_std Mean 0.982266 +trainer/policy/normal/log_std Std 0.327091 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.29277 +eval/num steps total 861030 +eval/num paths total 862 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0390811 +eval/Actions Std 0.942012 +eval/Actions Max 0.999999 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.75427 +time/logging (s) 0.00365974 +time/sampling batch (s) 0.526399 +time/saving (s) 0.00336023 +time/training (s) 7.02035 +time/epoch (s) 10.308 +time/total (s) 8843.01 +Epoch -139 +---------------------------------- --------------- +2022-05-10 15:38:23.002528 PDT | [2] Epoch -138 finished +---------------------------------- --------------- +epoch -138 +replay_buffer/size 999033 +trainer/num train calls 863000 +trainer/Policy Loss -19.5322 +trainer/Log Pis Mean 25.2949 +trainer/Log Pis Std 13.4807 +trainer/Log Pis Max 72.2232 +trainer/Log Pis Min -6.98066 +trainer/policy/mean Mean -0.0507162 +trainer/policy/mean Std 0.909262 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.82009 +trainer/policy/normal/std Std 0.671062 +trainer/policy/normal/std Max 6.44368 +trainer/policy/normal/std Min 0.308369 +trainer/policy/normal/log_std Mean 0.997236 +trainer/policy/normal/log_std Std 0.314228 +trainer/policy/normal/log_std Max 1.8631 +trainer/policy/normal/log_std Min -1.17646 +eval/num steps total 862030 +eval/num paths total 863 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0362606 +eval/Actions Std 0.911261 +eval/Actions Max 0.999998 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.43381 +time/logging (s) 0.00368448 +time/sampling batch (s) 0.276316 +time/saving (s) 0.00335634 +time/training (s) 7.53416 +time/epoch (s) 10.2513 +time/total (s) 8853.26 +Epoch -138 +---------------------------------- --------------- +2022-05-10 15:38:33.177736 PDT | [2] Epoch -137 finished +---------------------------------- --------------- +epoch -137 +replay_buffer/size 999033 +trainer/num train calls 864000 +trainer/Policy Loss -18.521 +trainer/Log Pis Mean 24.8103 +trainer/Log Pis Std 13.0047 +trainer/Log Pis Max 69.4858 +trainer/Log Pis Min -6.95355 +trainer/policy/mean Mean -0.0497348 +trainer/policy/mean Std 0.905537 +trainer/policy/mean Max 1 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.72781 +trainer/policy/normal/std Std 0.697839 +trainer/policy/normal/std Max 5.89916 +trainer/policy/normal/std Min 0.229679 +trainer/policy/normal/log_std Mean 0.954891 +trainer/policy/normal/log_std Std 0.355878 +trainer/policy/normal/log_std Max 1.77481 +trainer/policy/normal/log_std Min -1.47107 +eval/num steps total 863030 +eval/num paths total 864 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.186827 +eval/Actions Std 0.848695 +eval/Actions Max 0.999996 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53091 +time/logging (s) 0.00370885 +time/sampling batch (s) 0.274722 +time/saving (s) 0.00335287 +time/training (s) 7.33875 +time/epoch (s) 10.1514 +time/total (s) 8863.42 +Epoch -137 +---------------------------------- --------------- +2022-05-10 15:38:43.426513 PDT | [2] Epoch -136 finished +---------------------------------- --------------- +epoch -136 +replay_buffer/size 999033 +trainer/num train calls 865000 +trainer/Policy Loss -19.8215 +trainer/Log Pis Mean 24.9626 +trainer/Log Pis Std 13.8822 +trainer/Log Pis Max 71.2607 +trainer/Log Pis Min -11.4747 +trainer/policy/mean Mean -0.042147 +trainer/policy/mean Std 0.905975 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.82691 +trainer/policy/normal/std Std 0.655334 +trainer/policy/normal/std Max 5.72347 +trainer/policy/normal/std Min 0.308644 +trainer/policy/normal/log_std Mean 1.00056 +trainer/policy/normal/log_std Std 0.313216 +trainer/policy/normal/log_std Max 1.74457 +trainer/policy/normal/log_std Min -1.17557 +eval/num steps total 864030 +eval/num paths total 865 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0316546 +eval/Actions Std 0.90999 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.47651 +time/logging (s) 0.0039309 +time/sampling batch (s) 0.52888 +time/saving (s) 0.00361073 +time/training (s) 7.21214 +time/epoch (s) 10.2251 +time/total (s) 8873.65 +Epoch -136 +---------------------------------- --------------- +2022-05-10 15:38:53.201352 PDT | [2] Epoch -135 finished +---------------------------------- --------------- +epoch -135 +replay_buffer/size 999033 +trainer/num train calls 866000 +trainer/Policy Loss -20.0388 +trainer/Log Pis Mean 23.9529 +trainer/Log Pis Std 12.5961 +trainer/Log Pis Max 62.6123 +trainer/Log Pis Min -10.0784 +trainer/policy/mean Mean -0.0297146 +trainer/policy/mean Std 0.905327 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.74125 +trainer/policy/normal/std Std 0.659755 +trainer/policy/normal/std Max 6.20164 +trainer/policy/normal/std Min 0.228905 +trainer/policy/normal/log_std Mean 0.967494 +trainer/policy/normal/log_std Std 0.321568 +trainer/policy/normal/log_std Max 1.82481 +trainer/policy/normal/log_std Min -1.47445 +eval/num steps total 865030 +eval/num paths total 866 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0276945 +eval/Actions Std 0.788455 +eval/Actions Max 0.999992 +eval/Actions Min -0.999978 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68251 +time/logging (s) 0.00373214 +time/sampling batch (s) 0.525159 +time/saving (s) 0.003518 +time/training (s) 6.53559 +time/epoch (s) 9.75051 +time/total (s) 8883.4 +Epoch -135 +---------------------------------- --------------- +2022-05-10 15:39:02.850556 PDT | [2] Epoch -134 finished +---------------------------------- --------------- +epoch -134 +replay_buffer/size 999033 +trainer/num train calls 867000 +trainer/Policy Loss -18.9113 +trainer/Log Pis Mean 25.083 +trainer/Log Pis Std 13.3082 +trainer/Log Pis Max 66.9748 +trainer/Log Pis Min -7.24884 +trainer/policy/mean Mean -0.0368847 +trainer/policy/mean Std 0.906248 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.79448 +trainer/policy/normal/std Std 0.656489 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.301127 +trainer/policy/normal/log_std Mean 0.988364 +trainer/policy/normal/log_std Std 0.315297 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.20022 +eval/num steps total 866030 +eval/num paths total 867 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.284769 +eval/Actions Std 0.885521 +eval/Actions Max 0.999995 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60512 +time/logging (s) 0.00369904 +time/sampling batch (s) 0.275685 +time/saving (s) 0.00335545 +time/training (s) 6.73737 +time/epoch (s) 9.62523 +time/total (s) 8893.03 +Epoch -134 +---------------------------------- --------------- +2022-05-10 15:39:12.669970 PDT | [2] Epoch -133 finished +---------------------------------- --------------- +epoch -133 +replay_buffer/size 999033 +trainer/num train calls 868000 +trainer/Policy Loss -20.1161 +trainer/Log Pis Mean 23.9901 +trainer/Log Pis Std 13.166 +trainer/Log Pis Max 71.3704 +trainer/Log Pis Min -4.08916 +trainer/policy/mean Mean -0.0210973 +trainer/policy/mean Std 0.907756 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.74519 +trainer/policy/normal/std Std 0.702296 +trainer/policy/normal/std Max 5.68069 +trainer/policy/normal/std Min 0.215196 +trainer/policy/normal/log_std Mean 0.960143 +trainer/policy/normal/log_std Std 0.362048 +trainer/policy/normal/log_std Max 1.73707 +trainer/policy/normal/log_std Min -1.53621 +eval/num steps total 867030 +eval/num paths total 868 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0836074 +eval/Actions Std 0.886446 +eval/Actions Max 0.99999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51593 +time/logging (s) 0.00364883 +time/sampling batch (s) 0.781356 +time/saving (s) 0.00337721 +time/training (s) 6.49075 +time/epoch (s) 9.79506 +time/total (s) 8902.83 +Epoch -133 +---------------------------------- --------------- +2022-05-10 15:39:24.074992 PDT | [2] Epoch -132 finished +---------------------------------- --------------- +epoch -132 +replay_buffer/size 999033 +trainer/num train calls 869000 +trainer/Policy Loss -20.3302 +trainer/Log Pis Mean 25.0029 +trainer/Log Pis Std 13.6677 +trainer/Log Pis Max 68.3517 +trainer/Log Pis Min -8.06214 +trainer/policy/mean Mean -0.0621931 +trainer/policy/mean Std 0.904289 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.83264 +trainer/policy/normal/std Std 0.674553 +trainer/policy/normal/std Max 6.28919 +trainer/policy/normal/std Min 0.271187 +trainer/policy/normal/log_std Mean 1.00045 +trainer/policy/normal/log_std Std 0.32362 +trainer/policy/normal/log_std Max 1.83883 +trainer/policy/normal/log_std Min -1.30495 +eval/num steps total 868030 +eval/num paths total 869 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.110132 +eval/Actions Std 0.919112 +eval/Actions Max 0.999983 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69789 +time/logging (s) 0.00380678 +time/sampling batch (s) 1.02682 +time/saving (s) 0.00338763 +time/training (s) 7.64941 +time/epoch (s) 11.3813 +time/total (s) 8914.21 +Epoch -132 +---------------------------------- --------------- +2022-05-10 15:39:35.058126 PDT | [2] Epoch -131 finished +---------------------------------- --------------- +epoch -131 +replay_buffer/size 999033 +trainer/num train calls 870000 +trainer/Policy Loss -19.3333 +trainer/Log Pis Mean 23.8711 +trainer/Log Pis Std 13.2765 +trainer/Log Pis Max 67.0205 +trainer/Log Pis Min -12.9125 +trainer/policy/mean Mean -0.0304326 +trainer/policy/mean Std 0.902141 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.79964 +trainer/policy/normal/std Std 0.688797 +trainer/policy/normal/std Max 6.89041 +trainer/policy/normal/std Min 0.271583 +trainer/policy/normal/log_std Mean 0.986818 +trainer/policy/normal/log_std Std 0.32985 +trainer/policy/normal/log_std Max 1.93013 +trainer/policy/normal/log_std Min -1.30349 +eval/num steps total 869030 +eval/num paths total 870 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.103377 +eval/Actions Std 0.913045 +eval/Actions Max 0.999987 +eval/Actions Min -0.999987 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50546 +time/logging (s) 0.00378004 +time/sampling batch (s) 0.820676 +time/saving (s) 0.00355578 +time/training (s) 7.62535 +time/epoch (s) 10.9588 +time/total (s) 8925.17 +Epoch -131 +---------------------------------- --------------- +2022-05-10 15:39:46.160136 PDT | [2] Epoch -130 finished +---------------------------------- --------------- +epoch -130 +replay_buffer/size 999033 +trainer/num train calls 871000 +trainer/Policy Loss -19.6334 +trainer/Log Pis Mean 24.5989 +trainer/Log Pis Std 12.8982 +trainer/Log Pis Max 65.3805 +trainer/Log Pis Min -8.86897 +trainer/policy/mean Mean -0.0521709 +trainer/policy/mean Std 0.907212 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.8094 +trainer/policy/normal/std Std 0.682118 +trainer/policy/normal/std Max 6.87855 +trainer/policy/normal/std Min 0.289896 +trainer/policy/normal/log_std Mean 0.989889 +trainer/policy/normal/log_std Std 0.33229 +trainer/policy/normal/log_std Max 1.92841 +trainer/policy/normal/log_std Min -1.23823 +eval/num steps total 870030 +eval/num paths total 871 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.367124 +eval/Actions Std 0.853603 +eval/Actions Max 0.999972 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.75621 +time/logging (s) 0.00409401 +time/sampling batch (s) 0.540382 +time/saving (s) 0.0037515 +time/training (s) 7.77377 +time/epoch (s) 11.0782 +time/total (s) 8936.25 +Epoch -130 +---------------------------------- --------------- +2022-05-10 15:39:56.960265 PDT | [2] Epoch -129 finished +---------------------------------- --------------- +epoch -129 +replay_buffer/size 999033 +trainer/num train calls 872000 +trainer/Policy Loss -20.2122 +trainer/Log Pis Mean 24.096 +trainer/Log Pis Std 13.4182 +trainer/Log Pis Max 59.0578 +trainer/Log Pis Min -10.4117 +trainer/policy/mean Mean -0.0286304 +trainer/policy/mean Std 0.906803 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.81396 +trainer/policy/normal/std Std 0.690924 +trainer/policy/normal/std Max 6.16131 +trainer/policy/normal/std Min 0.227211 +trainer/policy/normal/log_std Mean 0.989284 +trainer/policy/normal/log_std Std 0.345088 +trainer/policy/normal/log_std Max 1.81829 +trainer/policy/normal/log_std Min -1.48188 +eval/num steps total 871030 +eval/num paths total 872 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.375374 +eval/Actions Std 0.790108 +eval/Actions Max 0.999995 +eval/Actions Min -0.999985 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61624 +time/logging (s) 0.00370936 +time/sampling batch (s) 0.526462 +time/saving (s) 0.00340529 +time/training (s) 7.6256 +time/epoch (s) 10.7754 +time/total (s) 8947.03 +Epoch -129 +---------------------------------- --------------- +2022-05-10 15:40:07.812809 PDT | [2] Epoch -128 finished +---------------------------------- --------------- +epoch -128 +replay_buffer/size 999033 +trainer/num train calls 873000 +trainer/Policy Loss -19.1736 +trainer/Log Pis Mean 24.401 +trainer/Log Pis Std 12.8456 +trainer/Log Pis Max 71.341 +trainer/Log Pis Min -8.86197 +trainer/policy/mean Mean -0.0445545 +trainer/policy/mean Std 0.909451 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.83759 +trainer/policy/normal/std Std 0.663242 +trainer/policy/normal/std Max 5.67528 +trainer/policy/normal/std Min 0.236981 +trainer/policy/normal/log_std Mean 1.00503 +trainer/policy/normal/log_std Std 0.308375 +trainer/policy/normal/log_std Max 1.73612 +trainer/policy/normal/log_std Min -1.43978 +eval/num steps total 872030 +eval/num paths total 873 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0574928 +eval/Actions Std 0.907859 +eval/Actions Max 0.999988 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6643 +time/logging (s) 0.00372631 +time/sampling batch (s) 0.527636 +time/saving (s) 0.00335523 +time/training (s) 7.62932 +time/epoch (s) 10.8283 +time/total (s) 8957.86 +Epoch -128 +---------------------------------- --------------- +2022-05-10 15:40:18.543563 PDT | [2] Epoch -127 finished +---------------------------------- --------------- +epoch -127 +replay_buffer/size 999033 +trainer/num train calls 874000 +trainer/Policy Loss -19.1557 +trainer/Log Pis Mean 24.78 +trainer/Log Pis Std 13.6981 +trainer/Log Pis Max 74.9004 +trainer/Log Pis Min -5.84686 +trainer/policy/mean Mean -0.0446278 +trainer/policy/mean Std 0.907487 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.82642 +trainer/policy/normal/std Std 0.70144 +trainer/policy/normal/std Max 5.27018 +trainer/policy/normal/std Min 0.25862 +trainer/policy/normal/log_std Mean 0.991523 +trainer/policy/normal/log_std Std 0.355038 +trainer/policy/normal/log_std Max 1.66206 +trainer/policy/normal/log_std Min -1.35239 +eval/num steps total 873030 +eval/num paths total 874 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.10155 +eval/Actions Std 0.915409 +eval/Actions Max 0.999984 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60341 +time/logging (s) 0.00386553 +time/sampling batch (s) 0.278719 +time/saving (s) 0.00349061 +time/training (s) 7.81718 +time/epoch (s) 10.7067 +time/total (s) 8968.57 +Epoch -127 +---------------------------------- --------------- +2022-05-10 15:40:28.438165 PDT | [2] Epoch -126 finished +---------------------------------- --------------- +epoch -126 +replay_buffer/size 999033 +trainer/num train calls 875000 +trainer/Policy Loss -19.4226 +trainer/Log Pis Mean 24.3309 +trainer/Log Pis Std 13.568 +trainer/Log Pis Max 73.7502 +trainer/Log Pis Min -5.28022 +trainer/policy/mean Mean -0.0328852 +trainer/policy/mean Std 0.908158 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.74433 +trainer/policy/normal/std Std 0.660747 +trainer/policy/normal/std Max 5.86651 +trainer/policy/normal/std Min 0.193701 +trainer/policy/normal/log_std Mean 0.967479 +trainer/policy/normal/log_std Std 0.329159 +trainer/policy/normal/log_std Max 1.76926 +trainer/policy/normal/log_std Min -1.64144 +eval/num steps total 874030 +eval/num paths total 875 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.107219 +eval/Actions Std 0.919055 +eval/Actions Max 0.99999 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53135 +time/logging (s) 0.00372268 +time/sampling batch (s) 0.277356 +time/saving (s) 0.0033703 +time/training (s) 7.0542 +time/epoch (s) 9.87 +time/total (s) 8978.45 +Epoch -126 +---------------------------------- --------------- +2022-05-10 15:40:39.632646 PDT | [2] Epoch -125 finished +---------------------------------- --------------- +epoch -125 +replay_buffer/size 999033 +trainer/num train calls 876000 +trainer/Policy Loss -20.2729 +trainer/Log Pis Mean 24.3809 +trainer/Log Pis Std 13.3708 +trainer/Log Pis Max 63.9925 +trainer/Log Pis Min -9.48279 +trainer/policy/mean Mean -0.0568986 +trainer/policy/mean Std 0.90848 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.7443 +trainer/policy/normal/std Std 0.658536 +trainer/policy/normal/std Max 5.19079 +trainer/policy/normal/std Min 0.289335 +trainer/policy/normal/log_std Mean 0.967041 +trainer/policy/normal/log_std Std 0.33013 +trainer/policy/normal/log_std Max 1.64689 +trainer/policy/normal/log_std Min -1.24017 +eval/num steps total 875030 +eval/num paths total 876 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.19689 +eval/Actions Std 0.898477 +eval/Actions Max 0.999998 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50976 +time/logging (s) 0.00397862 +time/sampling batch (s) 0.781733 +time/saving (s) 0.0035954 +time/training (s) 7.87111 +time/epoch (s) 11.1702 +time/total (s) 8989.62 +Epoch -125 +---------------------------------- --------------- +2022-05-10 15:40:49.730265 PDT | [2] Epoch -124 finished +---------------------------------- --------------- +epoch -124 +replay_buffer/size 999033 +trainer/num train calls 877000 +trainer/Policy Loss -19.9962 +trainer/Log Pis Mean 24.5498 +trainer/Log Pis Std 13.1383 +trainer/Log Pis Max 70.4635 +trainer/Log Pis Min -7.01979 +trainer/policy/mean Mean -0.0453466 +trainer/policy/mean Std 0.907628 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.81484 +trainer/policy/normal/std Std 0.694027 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.279155 +trainer/policy/normal/log_std Mean 0.990967 +trainer/policy/normal/log_std Std 0.335309 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.27599 +eval/num steps total 876030 +eval/num paths total 877 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.123271 +eval/Actions Std 0.883682 +eval/Actions Max 0.99999 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.71305 +time/logging (s) 0.00375906 +time/sampling batch (s) 0.279103 +time/saving (s) 0.00351155 +time/training (s) 7.0734 +time/epoch (s) 10.0728 +time/total (s) 8999.7 +Epoch -124 +---------------------------------- --------------- +2022-05-10 15:41:00.212072 PDT | [2] Epoch -123 finished +---------------------------------- --------------- +epoch -123 +replay_buffer/size 999033 +trainer/num train calls 878000 +trainer/Policy Loss -19.4685 +trainer/Log Pis Mean 25.0147 +trainer/Log Pis Std 13.7905 +trainer/Log Pis Max 77.0628 +trainer/Log Pis Min -6.42684 +trainer/policy/mean Mean -0.0304224 +trainer/policy/mean Std 0.908194 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.825 +trainer/policy/normal/std Std 0.702883 +trainer/policy/normal/std Max 6.08239 +trainer/policy/normal/std Min 0.29045 +trainer/policy/normal/log_std Mean 0.992142 +trainer/policy/normal/log_std Std 0.348859 +trainer/policy/normal/log_std Max 1.8054 +trainer/policy/normal/log_std Min -1.23632 +eval/num steps total 877030 +eval/num paths total 878 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0611101 +eval/Actions Std 0.921434 +eval/Actions Max 0.999997 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50029 +time/logging (s) 0.00372007 +time/sampling batch (s) 0.275899 +time/saving (s) 0.00335986 +time/training (s) 7.6746 +time/epoch (s) 10.4579 +time/total (s) 9010.16 +Epoch -123 +---------------------------------- --------------- +2022-05-10 15:41:11.076018 PDT | [2] Epoch -122 finished +---------------------------------- --------------- +epoch -122 +replay_buffer/size 999033 +trainer/num train calls 879000 +trainer/Policy Loss -19.6632 +trainer/Log Pis Mean 23.572 +trainer/Log Pis Std 14.3641 +trainer/Log Pis Max 67.5512 +trainer/Log Pis Min -7.60456 +trainer/policy/mean Mean -0.0338434 +trainer/policy/mean Std 0.907309 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999968 +trainer/policy/normal/std Mean 2.78177 +trainer/policy/normal/std Std 0.664223 +trainer/policy/normal/std Max 6.62425 +trainer/policy/normal/std Min 0.320083 +trainer/policy/normal/log_std Mean 0.981432 +trainer/policy/normal/log_std Std 0.326703 +trainer/policy/normal/log_std Max 1.89074 +trainer/policy/normal/log_std Min -1.13918 +eval/num steps total 878030 +eval/num paths total 879 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.108917 +eval/Actions Std 0.876711 +eval/Actions Max 0.999997 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6988 +time/logging (s) 0.0037595 +time/sampling batch (s) 0.775409 +time/saving (s) 0.00337414 +time/training (s) 7.35865 +time/epoch (s) 10.84 +time/total (s) 9021 +Epoch -122 +---------------------------------- --------------- +2022-05-10 15:41:20.833214 PDT | [2] Epoch -121 finished +---------------------------------- --------------- +epoch -121 +replay_buffer/size 999033 +trainer/num train calls 880000 +trainer/Policy Loss -19.5273 +trainer/Log Pis Mean 23.2736 +trainer/Log Pis Std 12.6794 +trainer/Log Pis Max 72.3384 +trainer/Log Pis Min -8.75727 +trainer/policy/mean Mean -0.036486 +trainer/policy/mean Std 0.901152 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.77873 +trainer/policy/normal/std Std 0.674184 +trainer/policy/normal/std Max 5.64563 +trainer/policy/normal/std Min 0.31447 +trainer/policy/normal/log_std Mean 0.979918 +trainer/policy/normal/log_std Std 0.32627 +trainer/policy/normal/log_std Max 1.73088 +trainer/policy/normal/log_std Min -1.15687 +eval/num steps total 879030 +eval/num paths total 880 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0317693 +eval/Actions Std 0.887579 +eval/Actions Max 0.999996 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56992 +time/logging (s) 0.00369493 +time/sampling batch (s) 0.271965 +time/saving (s) 0.00331554 +time/training (s) 6.88453 +time/epoch (s) 9.73342 +time/total (s) 9030.74 +Epoch -121 +---------------------------------- --------------- +2022-05-10 15:41:32.126781 PDT | [2] Epoch -120 finished +---------------------------------- --------------- +epoch -120 +replay_buffer/size 999033 +trainer/num train calls 881000 +trainer/Policy Loss -20.527 +trainer/Log Pis Mean 24.9844 +trainer/Log Pis Std 13.1983 +trainer/Log Pis Max 61.5191 +trainer/Log Pis Min -6.27656 +trainer/policy/mean Mean -0.0309362 +trainer/policy/mean Std 0.909016 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.83734 +trainer/policy/normal/std Std 0.68614 +trainer/policy/normal/std Max 5.93862 +trainer/policy/normal/std Min 0.287093 +trainer/policy/normal/log_std Mean 0.999624 +trainer/policy/normal/log_std Std 0.33473 +trainer/policy/normal/log_std Max 1.78148 +trainer/policy/normal/log_std Min -1.24795 +eval/num steps total 880030 +eval/num paths total 881 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0104335 +eval/Actions Std 0.855764 +eval/Actions Max 0.999993 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67629 +time/logging (s) 0.00373688 +time/sampling batch (s) 0.524934 +time/saving (s) 0.00335383 +time/training (s) 8.06134 +time/epoch (s) 11.2697 +time/total (s) 9042.01 +Epoch -120 +---------------------------------- --------------- +2022-05-10 15:41:42.717450 PDT | [2] Epoch -119 finished +---------------------------------- --------------- +epoch -119 +replay_buffer/size 999033 +trainer/num train calls 882000 +trainer/Policy Loss -21.0149 +trainer/Log Pis Mean 23.9552 +trainer/Log Pis Std 12.5985 +trainer/Log Pis Max 76.9892 +trainer/Log Pis Min -7.50876 +trainer/policy/mean Mean -0.0112688 +trainer/policy/mean Std 0.905822 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.80752 +trainer/policy/normal/std Std 0.677901 +trainer/policy/normal/std Max 5.76695 +trainer/policy/normal/std Min 0.263468 +trainer/policy/normal/log_std Mean 0.988502 +trainer/policy/normal/log_std Std 0.338853 +trainer/policy/normal/log_std Max 1.75214 +trainer/policy/normal/log_std Min -1.33382 +eval/num steps total 881030 +eval/num paths total 882 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.139088 +eval/Actions Std 0.800518 +eval/Actions Max 0.999983 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67641 +time/logging (s) 0.00399486 +time/sampling batch (s) 0.522425 +time/saving (s) 0.00363419 +time/training (s) 7.36075 +time/epoch (s) 10.5672 +time/total (s) 9052.58 +Epoch -119 +---------------------------------- --------------- +2022-05-10 15:41:53.680043 PDT | [2] Epoch -118 finished +---------------------------------- --------------- +epoch -118 +replay_buffer/size 999033 +trainer/num train calls 883000 +trainer/Policy Loss -18.0901 +trainer/Log Pis Mean 24.5874 +trainer/Log Pis Std 13.3718 +trainer/Log Pis Max 69.6692 +trainer/Log Pis Min -8.60797 +trainer/policy/mean Mean -0.0473501 +trainer/policy/mean Std 0.903917 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.78365 +trainer/policy/normal/std Std 0.662121 +trainer/policy/normal/std Max 6.9052 +trainer/policy/normal/std Min 0.312712 +trainer/policy/normal/log_std Mean 0.982624 +trainer/policy/normal/log_std Std 0.32536 +trainer/policy/normal/log_std Max 1.93227 +trainer/policy/normal/log_std Min -1.16247 +eval/num steps total 882030 +eval/num paths total 883 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.145054 +eval/Actions Std 0.914442 +eval/Actions Max 0.999995 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62551 +time/logging (s) 0.00436076 +time/sampling batch (s) 0.777515 +time/saving (s) 0.00330804 +time/training (s) 7.52784 +time/epoch (s) 10.9385 +time/total (s) 9063.52 +Epoch -118 +---------------------------------- --------------- +2022-05-10 15:42:04.335670 PDT | [2] Epoch -117 finished +---------------------------------- --------------- +epoch -117 +replay_buffer/size 999033 +trainer/num train calls 884000 +trainer/Policy Loss -19.1149 +trainer/Log Pis Mean 24.4399 +trainer/Log Pis Std 13.7182 +trainer/Log Pis Max 78.268 +trainer/Log Pis Min -7.02412 +trainer/policy/mean Mean -0.0540237 +trainer/policy/mean Std 0.905384 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999981 +trainer/policy/normal/std Mean 2.81948 +trainer/policy/normal/std Std 0.657122 +trainer/policy/normal/std Max 5.81535 +trainer/policy/normal/std Min 0.204747 +trainer/policy/normal/log_std Mean 0.997697 +trainer/policy/normal/log_std Std 0.315061 +trainer/policy/normal/log_std Max 1.7605 +trainer/policy/normal/log_std Min -1.58598 +eval/num steps total 883030 +eval/num paths total 884 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.059891 +eval/Actions Std 0.893099 +eval/Actions Max 0.999994 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54839 +time/logging (s) 0.00376316 +time/sampling batch (s) 0.527379 +time/saving (s) 0.00346101 +time/training (s) 7.54748 +time/epoch (s) 10.6305 +time/total (s) 9074.15 +Epoch -117 +---------------------------------- --------------- +2022-05-10 15:42:15.478525 PDT | [2] Epoch -116 finished +---------------------------------- --------------- +epoch -116 +replay_buffer/size 999033 +trainer/num train calls 885000 +trainer/Policy Loss -19.7361 +trainer/Log Pis Mean 24.3192 +trainer/Log Pis Std 13.1689 +trainer/Log Pis Max 64.772 +trainer/Log Pis Min -5.30492 +trainer/policy/mean Mean -0.0345007 +trainer/policy/mean Std 0.908546 +trainer/policy/mean Max 0.999976 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.78669 +trainer/policy/normal/std Std 0.663212 +trainer/policy/normal/std Max 6.12333 +trainer/policy/normal/std Min 0.28999 +trainer/policy/normal/log_std Mean 0.982865 +trainer/policy/normal/log_std Std 0.330635 +trainer/policy/normal/log_std Max 1.81211 +trainer/policy/normal/log_std Min -1.23791 +eval/num steps total 884030 +eval/num paths total 885 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.149188 +eval/Actions Std 0.851213 +eval/Actions Max 0.999991 +eval/Actions Min -0.999987 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69498 +time/logging (s) 0.00374704 +time/sampling batch (s) 0.521831 +time/saving (s) 0.00344136 +time/training (s) 7.89483 +time/epoch (s) 11.1188 +time/total (s) 9085.28 +Epoch -116 +---------------------------------- --------------- +2022-05-10 15:42:25.808514 PDT | [2] Epoch -115 finished +---------------------------------- --------------- +epoch -115 +replay_buffer/size 999033 +trainer/num train calls 886000 +trainer/Policy Loss -19.8797 +trainer/Log Pis Mean 25.4105 +trainer/Log Pis Std 12.7074 +trainer/Log Pis Max 70.6893 +trainer/Log Pis Min -3.00549 +trainer/policy/mean Mean -0.0429377 +trainer/policy/mean Std 0.908313 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.81021 +trainer/policy/normal/std Std 0.67606 +trainer/policy/normal/std Max 6.17663 +trainer/policy/normal/std Min 0.317548 +trainer/policy/normal/log_std Mean 0.991519 +trainer/policy/normal/log_std Std 0.326693 +trainer/policy/normal/log_std Max 1.82077 +trainer/policy/normal/log_std Min -1.14713 +eval/num steps total 885030 +eval/num paths total 886 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0586538 +eval/Actions Std 0.914951 +eval/Actions Max 0.999997 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.46694 +time/logging (s) 0.00397184 +time/sampling batch (s) 0.271163 +time/saving (s) 0.00355041 +time/training (s) 7.56056 +time/epoch (s) 10.3062 +time/total (s) 9095.59 +Epoch -115 +---------------------------------- --------------- +2022-05-10 15:42:35.728871 PDT | [2] Epoch -114 finished +---------------------------------- --------------- +epoch -114 +replay_buffer/size 999033 +trainer/num train calls 887000 +trainer/Policy Loss -19.7097 +trainer/Log Pis Mean 24.4432 +trainer/Log Pis Std 12.6639 +trainer/Log Pis Max 66.8162 +trainer/Log Pis Min -7.25888 +trainer/policy/mean Mean -0.0500591 +trainer/policy/mean Std 0.905241 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.7441 +trainer/policy/normal/std Std 0.662602 +trainer/policy/normal/std Max 5.95325 +trainer/policy/normal/std Min 0.282457 +trainer/policy/normal/log_std Mean 0.967645 +trainer/policy/normal/log_std Std 0.326103 +trainer/policy/normal/log_std Max 1.78394 +trainer/policy/normal/log_std Min -1.26423 +eval/num steps total 886030 +eval/num paths total 887 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0381068 +eval/Actions Std 0.910278 +eval/Actions Max 0.999993 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.566 +time/logging (s) 0.00411521 +time/sampling batch (s) 0.271277 +time/saving (s) 0.00381555 +time/training (s) 7.0511 +time/epoch (s) 9.89631 +time/total (s) 9105.48 +Epoch -114 +---------------------------------- --------------- +2022-05-10 15:42:45.646532 PDT | [2] Epoch -113 finished +---------------------------------- --------------- +epoch -113 +replay_buffer/size 999033 +trainer/num train calls 888000 +trainer/Policy Loss -19.3623 +trainer/Log Pis Mean 23.4899 +trainer/Log Pis Std 12.9208 +trainer/Log Pis Max 68.4406 +trainer/Log Pis Min -6.9319 +trainer/policy/mean Mean -0.0270092 +trainer/policy/mean Std 0.905012 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.80045 +trainer/policy/normal/std Std 0.661951 +trainer/policy/normal/std Max 5.98552 +trainer/policy/normal/std Min 0.284165 +trainer/policy/normal/log_std Mean 0.988922 +trainer/policy/normal/log_std Std 0.323691 +trainer/policy/normal/log_std Max 1.78934 +trainer/policy/normal/log_std Min -1.2582 +eval/num steps total 887030 +eval/num paths total 888 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00782267 +eval/Actions Std 0.90205 +eval/Actions Max 0.999995 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.43505 +time/logging (s) 0.00376341 +time/sampling batch (s) 0.521885 +time/saving (s) 0.00339227 +time/training (s) 6.9288 +time/epoch (s) 9.89289 +time/total (s) 9115.38 +Epoch -113 +---------------------------------- --------------- +2022-05-10 15:42:55.675283 PDT | [2] Epoch -112 finished +---------------------------------- --------------- +epoch -112 +replay_buffer/size 999033 +trainer/num train calls 889000 +trainer/Policy Loss -20.1043 +trainer/Log Pis Mean 24.5578 +trainer/Log Pis Std 13.5173 +trainer/Log Pis Max 80.4084 +trainer/Log Pis Min -4.79811 +trainer/policy/mean Mean -0.033329 +trainer/policy/mean Std 0.907812 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.79978 +trainer/policy/normal/std Std 0.679003 +trainer/policy/normal/std Max 5.83431 +trainer/policy/normal/std Min 0.245687 +trainer/policy/normal/log_std Mean 0.985939 +trainer/policy/normal/log_std Std 0.335971 +trainer/policy/normal/log_std Max 1.76376 +trainer/policy/normal/log_std Min -1.4037 +eval/num steps total 888030 +eval/num paths total 889 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.102846 +eval/Actions Std 0.87103 +eval/Actions Max 0.999994 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.76308 +time/logging (s) 0.00371021 +time/sampling batch (s) 0.525511 +time/saving (s) 0.00336535 +time/training (s) 6.70894 +time/epoch (s) 10.0046 +time/total (s) 9125.39 +Epoch -112 +---------------------------------- --------------- +2022-05-10 15:43:05.995225 PDT | [2] Epoch -111 finished +---------------------------------- --------------- +epoch -111 +replay_buffer/size 999033 +trainer/num train calls 890000 +trainer/Policy Loss -19.3506 +trainer/Log Pis Mean 24.6819 +trainer/Log Pis Std 13.2755 +trainer/Log Pis Max 72.4562 +trainer/Log Pis Min -7.66095 +trainer/policy/mean Mean -0.0356221 +trainer/policy/mean Std 0.90948 +trainer/policy/mean Max 0.999982 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.82426 +trainer/policy/normal/std Std 0.641875 +trainer/policy/normal/std Max 5.63979 +trainer/policy/normal/std Min 0.247712 +trainer/policy/normal/log_std Mean 1.00114 +trainer/policy/normal/log_std Std 0.308278 +trainer/policy/normal/log_std Max 1.72985 +trainer/policy/normal/log_std Min -1.39549 +eval/num steps total 889030 +eval/num paths total 890 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0243847 +eval/Actions Std 0.9517 +eval/Actions Max 0.999989 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.4783 +time/logging (s) 0.00371051 +time/sampling batch (s) 0.525449 +time/saving (s) 0.00337482 +time/training (s) 7.285 +time/epoch (s) 10.2958 +time/total (s) 9135.69 +Epoch -111 +---------------------------------- --------------- +2022-05-10 15:43:16.173936 PDT | [2] Epoch -110 finished +---------------------------------- --------------- +epoch -110 +replay_buffer/size 999033 +trainer/num train calls 891000 +trainer/Policy Loss -20.1623 +trainer/Log Pis Mean 25.4439 +trainer/Log Pis Std 12.7905 +trainer/Log Pis Max 72.1109 +trainer/Log Pis Min -7.38959 +trainer/policy/mean Mean -0.0681125 +trainer/policy/mean Std 0.904094 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.8113 +trainer/policy/normal/std Std 0.672991 +trainer/policy/normal/std Max 5.51064 +trainer/policy/normal/std Min 0.311297 +trainer/policy/normal/log_std Mean 0.992293 +trainer/policy/normal/log_std Std 0.324355 +trainer/policy/normal/log_std Max 1.70668 +trainer/policy/normal/log_std Min -1.16701 +eval/num steps total 890030 +eval/num paths total 891 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.130514 +eval/Actions Std 0.825682 +eval/Actions Max 0.999991 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.72352 +time/logging (s) 0.00414976 +time/sampling batch (s) 0.550309 +time/saving (s) 0.00397819 +time/training (s) 6.87279 +time/epoch (s) 10.1547 +time/total (s) 9145.85 +Epoch -110 +---------------------------------- --------------- +2022-05-10 15:43:26.432911 PDT | [2] Epoch -109 finished +---------------------------------- --------------- +epoch -109 +replay_buffer/size 999033 +trainer/num train calls 892000 +trainer/Policy Loss -20.2486 +trainer/Log Pis Mean 24.9968 +trainer/Log Pis Std 12.9923 +trainer/Log Pis Max 65.4219 +trainer/Log Pis Min -10.7205 +trainer/policy/mean Mean -0.0241609 +trainer/policy/mean Std 0.908498 +trainer/policy/mean Max 0.999978 +trainer/policy/mean Min -0.999985 +trainer/policy/normal/std Mean 2.80151 +trainer/policy/normal/std Std 0.6729 +trainer/policy/normal/std Max 6.19498 +trainer/policy/normal/std Min 0.278792 +trainer/policy/normal/log_std Mean 0.988216 +trainer/policy/normal/log_std Std 0.328107 +trainer/policy/normal/log_std Max 1.82374 +trainer/policy/normal/log_std Min -1.27729 +eval/num steps total 891030 +eval/num paths total 892 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.113254 +eval/Actions Std 0.882734 +eval/Actions Max 0.999979 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.92351 +time/logging (s) 0.00421362 +time/sampling batch (s) 0.323163 +time/saving (s) 0.00404876 +time/training (s) 6.979 +time/epoch (s) 10.2339 +time/total (s) 9156.08 +Epoch -109 +---------------------------------- --------------- +2022-05-10 15:43:37.216335 PDT | [2] Epoch -108 finished +---------------------------------- --------------- +epoch -108 +replay_buffer/size 999033 +trainer/num train calls 893000 +trainer/Policy Loss -20.8076 +trainer/Log Pis Mean 25.543 +trainer/Log Pis Std 13.4607 +trainer/Log Pis Max 68.7103 +trainer/Log Pis Min -8.88384 +trainer/policy/mean Mean -0.0391338 +trainer/policy/mean Std 0.909285 +trainer/policy/mean Max 0.99998 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.82812 +trainer/policy/normal/std Std 0.685599 +trainer/policy/normal/std Max 6.13212 +trainer/policy/normal/std Min 0.280849 +trainer/policy/normal/log_std Mean 0.996321 +trainer/policy/normal/log_std Std 0.334445 +trainer/policy/normal/log_std Max 1.81354 +trainer/policy/normal/log_std Min -1.26994 +eval/num steps total 892030 +eval/num paths total 893 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.116799 +eval/Actions Std 0.915053 +eval/Actions Max 0.99999 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.79502 +time/logging (s) 0.00375852 +time/sampling batch (s) 0.558991 +time/saving (s) 0.00350063 +time/training (s) 7.3966 +time/epoch (s) 10.7579 +time/total (s) 9166.84 +Epoch -108 +---------------------------------- --------------- +2022-05-10 15:43:46.475951 PDT | [2] Epoch -107 finished +---------------------------------- --------------- +epoch -107 +replay_buffer/size 999033 +trainer/num train calls 894000 +trainer/Policy Loss -17.7736 +trainer/Log Pis Mean 25.0839 +trainer/Log Pis Std 14.1783 +trainer/Log Pis Max 87.2379 +trainer/Log Pis Min -6.70603 +trainer/policy/mean Mean -0.0243393 +trainer/policy/mean Std 0.905815 +trainer/policy/mean Max 0.999981 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.84015 +trainer/policy/normal/std Std 0.694866 +trainer/policy/normal/std Max 5.64141 +trainer/policy/normal/std Min 0.291199 +trainer/policy/normal/log_std Mean 1.0004 +trainer/policy/normal/log_std Std 0.334235 +trainer/policy/normal/log_std Max 1.73013 +trainer/policy/normal/log_std Min -1.23375 +eval/num steps total 893030 +eval/num paths total 894 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.031778 +eval/Actions Std 0.928348 +eval/Actions Max 0.999992 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63298 +time/logging (s) 0.00412905 +time/sampling batch (s) 0.276234 +time/saving (s) 0.00374454 +time/training (s) 6.31872 +time/epoch (s) 9.23581 +time/total (s) 9176.08 +Epoch -107 +---------------------------------- --------------- +2022-05-10 15:43:56.277589 PDT | [2] Epoch -106 finished +---------------------------------- --------------- +epoch -106 +replay_buffer/size 999033 +trainer/num train calls 895000 +trainer/Policy Loss -19.6292 +trainer/Log Pis Mean 23.8634 +trainer/Log Pis Std 13.8446 +trainer/Log Pis Max 66.4988 +trainer/Log Pis Min -7.61818 +trainer/policy/mean Mean -0.0298982 +trainer/policy/mean Std 0.899766 +trainer/policy/mean Max 0.999978 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.76362 +trainer/policy/normal/std Std 0.697442 +trainer/policy/normal/std Max 6.25384 +trainer/policy/normal/std Min 0.259243 +trainer/policy/normal/log_std Mean 0.969155 +trainer/policy/normal/log_std Std 0.351537 +trainer/policy/normal/log_std Max 1.8332 +trainer/policy/normal/log_std Min -1.34999 +eval/num steps total 894030 +eval/num paths total 895 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.400077 +eval/Actions Std 0.759181 +eval/Actions Max 1 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.77496 +time/logging (s) 0.00405832 +time/sampling batch (s) 0.288697 +time/saving (s) 0.00370061 +time/training (s) 6.70504 +time/epoch (s) 9.77645 +time/total (s) 9185.86 +Epoch -106 +---------------------------------- --------------- +2022-05-10 15:44:06.231329 PDT | [2] Epoch -105 finished +---------------------------------- --------------- +epoch -105 +replay_buffer/size 999033 +trainer/num train calls 896000 +trainer/Policy Loss -19.3109 +trainer/Log Pis Mean 24.1154 +trainer/Log Pis Std 13.5024 +trainer/Log Pis Max 64.6534 +trainer/Log Pis Min -9.12283 +trainer/policy/mean Mean -0.0344752 +trainer/policy/mean Std 0.906391 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.83933 +trainer/policy/normal/std Std 0.652173 +trainer/policy/normal/std Max 7.21401 +trainer/policy/normal/std Min 0.28823 +trainer/policy/normal/log_std Mean 1.00641 +trainer/policy/normal/log_std Std 0.306793 +trainer/policy/normal/log_std Max 1.97602 +trainer/policy/normal/log_std Min -1.244 +eval/num steps total 895030 +eval/num paths total 896 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.227745 +eval/Actions Std 0.841891 +eval/Actions Max 0.999999 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.24257 +time/logging (s) 0.00418108 +time/sampling batch (s) 0.522042 +time/saving (s) 0.00386305 +time/training (s) 7.15705 +time/epoch (s) 9.92971 +time/total (s) 9195.8 +Epoch -105 +---------------------------------- --------------- +2022-05-10 15:44:17.129339 PDT | [2] Epoch -104 finished +---------------------------------- --------------- +epoch -104 +replay_buffer/size 999033 +trainer/num train calls 897000 +trainer/Policy Loss -19.9759 +trainer/Log Pis Mean 24.3049 +trainer/Log Pis Std 13.4759 +trainer/Log Pis Max 84.0458 +trainer/Log Pis Min -11.9348 +trainer/policy/mean Mean -0.043227 +trainer/policy/mean Std 0.907126 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.75875 +trainer/policy/normal/std Std 0.674757 +trainer/policy/normal/std Max 6.00349 +trainer/policy/normal/std Min 0.286758 +trainer/policy/normal/log_std Mean 0.970628 +trainer/policy/normal/log_std Std 0.3373 +trainer/policy/normal/log_std Max 1.79234 +trainer/policy/normal/log_std Min -1.24912 +eval/num steps total 896030 +eval/num paths total 897 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.261507 +eval/Actions Std 0.767705 +eval/Actions Max 0.99999 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51862 +time/logging (s) 0.00378465 +time/sampling batch (s) 0.282031 +time/saving (s) 0.00335184 +time/training (s) 8.06469 +time/epoch (s) 10.8725 +time/total (s) 9206.67 +Epoch -104 +---------------------------------- --------------- +2022-05-10 15:44:28.233389 PDT | [2] Epoch -103 finished +---------------------------------- --------------- +epoch -103 +replay_buffer/size 999033 +trainer/num train calls 898000 +trainer/Policy Loss -20.4788 +trainer/Log Pis Mean 23.5835 +trainer/Log Pis Std 12.5449 +trainer/Log Pis Max 59.4192 +trainer/Log Pis Min -8.0759 +trainer/policy/mean Mean -0.0434627 +trainer/policy/mean Std 0.907199 +trainer/policy/mean Max 0.999981 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.82976 +trainer/policy/normal/std Std 0.689278 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.300597 +trainer/policy/normal/log_std Mean 0.997393 +trainer/policy/normal/log_std Std 0.330857 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.20198 +eval/num steps total 897030 +eval/num paths total 898 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00680943 +eval/Actions Std 0.898999 +eval/Actions Max 0.999996 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6144 +time/logging (s) 0.00410101 +time/sampling batch (s) 1.27238 +time/saving (s) 0.0037131 +time/training (s) 7.186 +time/epoch (s) 11.0806 +time/total (s) 9217.75 +Epoch -103 +---------------------------------- --------------- +2022-05-10 15:44:39.858677 PDT | [2] Epoch -102 finished +---------------------------------- --------------- +epoch -102 +replay_buffer/size 999033 +trainer/num train calls 899000 +trainer/Policy Loss -19.469 +trainer/Log Pis Mean 24.6883 +trainer/Log Pis Std 12.7348 +trainer/Log Pis Max 63.0714 +trainer/Log Pis Min -3.97225 +trainer/policy/mean Mean -0.0367592 +trainer/policy/mean Std 0.908901 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.84086 +trainer/policy/normal/std Std 0.677303 +trainer/policy/normal/std Max 6.27942 +trainer/policy/normal/std Min 0.309735 +trainer/policy/normal/log_std Mean 1.004 +trainer/policy/normal/log_std Std 0.318451 +trainer/policy/normal/log_std Max 1.83728 +trainer/policy/normal/log_std Min -1.17204 +eval/num steps total 898030 +eval/num paths total 899 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.225025 +eval/Actions Std 0.892814 +eval/Actions Max 0.999994 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59918 +time/logging (s) 0.00373879 +time/sampling batch (s) 1.02242 +time/saving (s) 0.00340005 +time/training (s) 7.97181 +time/epoch (s) 11.6006 +time/total (s) 9229.36 +Epoch -102 +---------------------------------- --------------- +2022-05-10 15:44:50.014809 PDT | [2] Epoch -101 finished +---------------------------------- --------------- +epoch -101 +replay_buffer/size 999033 +trainer/num train calls 900000 +trainer/Policy Loss -19.8259 +trainer/Log Pis Mean 24.4325 +trainer/Log Pis Std 13.1692 +trainer/Log Pis Max 64.7979 +trainer/Log Pis Min -8.78707 +trainer/policy/mean Mean -0.0435641 +trainer/policy/mean Std 0.906524 +trainer/policy/mean Max 0.999974 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.74976 +trainer/policy/normal/std Std 0.637401 +trainer/policy/normal/std Max 5.93037 +trainer/policy/normal/std Min 0.26409 +trainer/policy/normal/log_std Mean 0.972384 +trainer/policy/normal/log_std Std 0.317207 +trainer/policy/normal/log_std Max 1.78009 +trainer/policy/normal/log_std Min -1.33147 +eval/num steps total 899030 +eval/num paths total 900 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.014813 +eval/Actions Std 0.90495 +eval/Actions Max 0.999995 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.46307 +time/logging (s) 0.00408588 +time/sampling batch (s) 0.779232 +time/saving (s) 0.00650496 +time/training (s) 6.87939 +time/epoch (s) 10.1323 +time/total (s) 9239.49 +Epoch -101 +---------------------------------- --------------- +2022-05-10 15:44:59.887142 PDT | [2] Epoch -100 finished +---------------------------------- --------------- +epoch -100 +replay_buffer/size 999033 +trainer/num train calls 901000 +trainer/Policy Loss -19.6873 +trainer/Log Pis Mean 23.9489 +trainer/Log Pis Std 12.6492 +trainer/Log Pis Max 74.2303 +trainer/Log Pis Min -7.25717 +trainer/policy/mean Mean -0.0346028 +trainer/policy/mean Std 0.901467 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.77105 +trainer/policy/normal/std Std 0.685105 +trainer/policy/normal/std Max 6.06169 +trainer/policy/normal/std Min 0.307454 +trainer/policy/normal/log_std Mean 0.973743 +trainer/policy/normal/log_std Std 0.343816 +trainer/policy/normal/log_std Max 1.80199 +trainer/policy/normal/log_std Min -1.17943 +eval/num steps total 900030 +eval/num paths total 901 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.300603 +eval/Actions Std 0.716945 +eval/Actions Max 0.999996 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70803 +time/logging (s) 0.00371365 +time/sampling batch (s) 0.521311 +time/saving (s) 0.00336423 +time/training (s) 6.61123 +time/epoch (s) 9.84764 +time/total (s) 9249.34 +Epoch -100 +---------------------------------- --------------- +2022-05-10 15:45:09.236944 PDT | [2] Epoch -99 finished +---------------------------------- --------------- +epoch -99 +replay_buffer/size 999033 +trainer/num train calls 902000 +trainer/Policy Loss -19.0332 +trainer/Log Pis Mean 25.0317 +trainer/Log Pis Std 13.6511 +trainer/Log Pis Max 71.1265 +trainer/Log Pis Min -9.35933 +trainer/policy/mean Mean -0.0242494 +trainer/policy/mean Std 0.909782 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999979 +trainer/policy/normal/std Mean 2.78364 +trainer/policy/normal/std Std 0.667746 +trainer/policy/normal/std Max 5.12026 +trainer/policy/normal/std Min 0.286912 +trainer/policy/normal/log_std Mean 0.982419 +trainer/policy/normal/log_std Std 0.32454 +trainer/policy/normal/log_std Max 1.6332 +trainer/policy/normal/log_std Min -1.24858 +eval/num steps total 901030 +eval/num paths total 902 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.205905 +eval/Actions Std 0.900117 +eval/Actions Max 0.999993 +eval/Actions Min -0.999988 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.76063 +time/logging (s) 0.00372594 +time/sampling batch (s) 0.273735 +time/saving (s) 0.00332929 +time/training (s) 6.28436 +time/epoch (s) 9.32578 +time/total (s) 9258.67 +Epoch -99 +---------------------------------- --------------- +2022-05-10 15:45:19.654491 PDT | [2] Epoch -98 finished +---------------------------------- --------------- +epoch -98 +replay_buffer/size 999033 +trainer/num train calls 903000 +trainer/Policy Loss -19.3602 +trainer/Log Pis Mean 24.92 +trainer/Log Pis Std 14.3219 +trainer/Log Pis Max 73.2638 +trainer/Log Pis Min -11.0751 +trainer/policy/mean Mean -0.0329069 +trainer/policy/mean Std 0.911078 +trainer/policy/mean Max 0.999978 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.74922 +trainer/policy/normal/std Std 0.647691 +trainer/policy/normal/std Max 5.62385 +trainer/policy/normal/std Min 0.292552 +trainer/policy/normal/log_std Mean 0.971179 +trainer/policy/normal/log_std Std 0.319908 +trainer/policy/normal/log_std Max 1.72702 +trainer/policy/normal/log_std Min -1.22911 +eval/num steps total 902030 +eval/num paths total 903 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.248179 +eval/Actions Std 0.764747 +eval/Actions Max 0.999997 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70536 +time/logging (s) 0.00379544 +time/sampling batch (s) 0.525535 +time/saving (s) 0.00338631 +time/training (s) 7.15553 +time/epoch (s) 10.3936 +time/total (s) 9269.07 +Epoch -98 +---------------------------------- --------------- +2022-05-10 15:45:29.264786 PDT | [2] Epoch -97 finished +---------------------------------- --------------- +epoch -97 +replay_buffer/size 999033 +trainer/num train calls 904000 +trainer/Policy Loss -20.3634 +trainer/Log Pis Mean 23.9936 +trainer/Log Pis Std 13.3371 +trainer/Log Pis Max 73.7963 +trainer/Log Pis Min -7.27629 +trainer/policy/mean Mean -0.0443987 +trainer/policy/mean Std 0.908119 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.84318 +trainer/policy/normal/std Std 0.667185 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.278369 +trainer/policy/normal/log_std Mean 1.00443 +trainer/policy/normal/log_std Std 0.324579 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.27881 +eval/num steps total 903030 +eval/num paths total 904 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.304187 +eval/Actions Std 0.881116 +eval/Actions Max 0.999991 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53377 +time/logging (s) 0.00410115 +time/sampling batch (s) 0.275858 +time/saving (s) 0.00374034 +time/training (s) 6.7691 +time/epoch (s) 9.58657 +time/total (s) 9278.66 +Epoch -97 +---------------------------------- --------------- +2022-05-10 15:45:39.972817 PDT | [2] Epoch -96 finished +---------------------------------- --------------- +epoch -96 +replay_buffer/size 999033 +trainer/num train calls 905000 +trainer/Policy Loss -19.7491 +trainer/Log Pis Mean 23.7602 +trainer/Log Pis Std 13.3179 +trainer/Log Pis Max 65.5692 +trainer/Log Pis Min -8.33992 +trainer/policy/mean Mean -0.039735 +trainer/policy/mean Std 0.908309 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.74451 +trainer/policy/normal/std Std 0.648707 +trainer/policy/normal/std Max 5.52054 +trainer/policy/normal/std Min 0.226365 +trainer/policy/normal/log_std Mean 0.96856 +trainer/policy/normal/log_std Std 0.325466 +trainer/policy/normal/log_std Max 1.70848 +trainer/policy/normal/log_std Min -1.48561 +eval/num steps total 904030 +eval/num paths total 905 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00239216 +eval/Actions Std 0.832093 +eval/Actions Max 0.999999 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.78856 +time/logging (s) 0.00368726 +time/sampling batch (s) 0.284656 +time/saving (s) 0.00336218 +time/training (s) 7.60246 +time/epoch (s) 10.6827 +time/total (s) 9289.35 +Epoch -96 +---------------------------------- --------------- +2022-05-10 15:45:49.522459 PDT | [2] Epoch -95 finished +---------------------------------- --------------- +epoch -95 +replay_buffer/size 999033 +trainer/num train calls 906000 +trainer/Policy Loss -19.2494 +trainer/Log Pis Mean 22.9715 +trainer/Log Pis Std 12.9013 +trainer/Log Pis Max 74.323 +trainer/Log Pis Min -10.1257 +trainer/policy/mean Mean -0.0132993 +trainer/policy/mean Std 0.901079 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.76121 +trainer/policy/normal/std Std 0.698444 +trainer/policy/normal/std Max 5.86552 +trainer/policy/normal/std Min 0.285837 +trainer/policy/normal/log_std Mean 0.968571 +trainer/policy/normal/log_std Std 0.348333 +trainer/policy/normal/log_std Max 1.76909 +trainer/policy/normal/log_std Min -1.25233 +eval/num steps total 905030 +eval/num paths total 906 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.122914 +eval/Actions Std 0.878871 +eval/Actions Max 0.999993 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70847 +time/logging (s) 0.00369794 +time/sampling batch (s) 0.287793 +time/saving (s) 0.00338157 +time/training (s) 6.52136 +time/epoch (s) 9.5247 +time/total (s) 9298.87 +Epoch -95 +---------------------------------- --------------- +2022-05-10 15:45:59.665619 PDT | [2] Epoch -94 finished +---------------------------------- --------------- +epoch -94 +replay_buffer/size 999033 +trainer/num train calls 907000 +trainer/Policy Loss -20.7195 +trainer/Log Pis Mean 24.4824 +trainer/Log Pis Std 13.7315 +trainer/Log Pis Max 80.2799 +trainer/Log Pis Min -13.4015 +trainer/policy/mean Mean -0.03061 +trainer/policy/mean Std 0.90837 +trainer/policy/mean Max 1 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.7709 +trainer/policy/normal/std Std 0.676409 +trainer/policy/normal/std Max 5.58734 +trainer/policy/normal/std Min 0.240586 +trainer/policy/normal/log_std Mean 0.973742 +trainer/policy/normal/log_std Std 0.347182 +trainer/policy/normal/log_std Max 1.7205 +trainer/policy/normal/log_std Min -1.42468 +eval/num steps total 906030 +eval/num paths total 907 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0346921 +eval/Actions Std 0.906416 +eval/Actions Max 0.999995 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.71406 +time/logging (s) 0.00373329 +time/sampling batch (s) 0.777185 +time/saving (s) 0.00344512 +time/training (s) 6.62045 +time/epoch (s) 10.1189 +time/total (s) 9308.99 +Epoch -94 +---------------------------------- --------------- +2022-05-10 15:46:10.222810 PDT | [2] Epoch -93 finished +---------------------------------- --------------- +epoch -93 +replay_buffer/size 999033 +trainer/num train calls 908000 +trainer/Policy Loss -20.1618 +trainer/Log Pis Mean 24.7001 +trainer/Log Pis Std 13.4366 +trainer/Log Pis Max 71.7881 +trainer/Log Pis Min -6.53023 +trainer/policy/mean Mean -0.0377526 +trainer/policy/mean Std 0.90708 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.77058 +trainer/policy/normal/std Std 0.678696 +trainer/policy/normal/std Max 5.88327 +trainer/policy/normal/std Min 0.250109 +trainer/policy/normal/log_std Mean 0.975087 +trainer/policy/normal/log_std Std 0.336514 +trainer/policy/normal/log_std Max 1.77211 +trainer/policy/normal/log_std Min -1.38586 +eval/num steps total 907030 +eval/num paths total 908 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0327025 +eval/Actions Std 0.945947 +eval/Actions Max 0.999991 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65387 +time/logging (s) 0.00370838 +time/sampling batch (s) 0.280022 +time/saving (s) 0.00338274 +time/training (s) 7.59178 +time/epoch (s) 10.5328 +time/total (s) 9319.53 +Epoch -93 +---------------------------------- --------------- +2022-05-10 15:46:21.203966 PDT | [2] Epoch -92 finished +---------------------------------- --------------- +epoch -92 +replay_buffer/size 999033 +trainer/num train calls 909000 +trainer/Policy Loss -19.6354 +trainer/Log Pis Mean 25.164 +trainer/Log Pis Std 12.8309 +trainer/Log Pis Max 68.7706 +trainer/Log Pis Min -8.21875 +trainer/policy/mean Mean -0.0431009 +trainer/policy/mean Std 0.907398 +trainer/policy/mean Max 0.999978 +trainer/policy/mean Min -0.999984 +trainer/policy/normal/std Mean 2.83531 +trainer/policy/normal/std Std 0.659589 +trainer/policy/normal/std Max 5.39179 +trainer/policy/normal/std Min 0.361802 +trainer/policy/normal/log_std Mean 1.00344 +trainer/policy/normal/log_std Std 0.313478 +trainer/policy/normal/log_std Max 1.68488 +trainer/policy/normal/log_std Min -1.01666 +eval/num steps total 908030 +eval/num paths total 909 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.202252 +eval/Actions Std 0.944091 +eval/Actions Max 0.999995 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.47589 +time/logging (s) 0.00402226 +time/sampling batch (s) 0.529038 +time/saving (s) 0.0037377 +time/training (s) 7.94445 +time/epoch (s) 10.9571 +time/total (s) 9330.49 +Epoch -92 +---------------------------------- --------------- +2022-05-10 15:46:30.970366 PDT | [2] Epoch -91 finished +---------------------------------- --------------- +epoch -91 +replay_buffer/size 999033 +trainer/num train calls 910000 +trainer/Policy Loss -21.6444 +trainer/Log Pis Mean 25.8832 +trainer/Log Pis Std 13.3885 +trainer/Log Pis Max 68.5667 +trainer/Log Pis Min -6.07587 +trainer/policy/mean Mean -0.0266286 +trainer/policy/mean Std 0.912223 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.84895 +trainer/policy/normal/std Std 0.691163 +trainer/policy/normal/std Max 6.26492 +trainer/policy/normal/std Min 0.285039 +trainer/policy/normal/log_std Mean 1.00386 +trainer/policy/normal/log_std Std 0.334195 +trainer/policy/normal/log_std Max 1.83497 +trainer/policy/normal/log_std Min -1.25513 +eval/num steps total 909030 +eval/num paths total 910 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0401429 +eval/Actions Std 0.914839 +eval/Actions Max 0.999997 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.45856 +time/logging (s) 0.00370763 +time/sampling batch (s) 0.52591 +time/saving (s) 0.00341486 +time/training (s) 6.7501 +time/epoch (s) 9.74169 +time/total (s) 9340.24 +Epoch -91 +---------------------------------- --------------- +2022-05-10 15:46:41.103620 PDT | [2] Epoch -90 finished +---------------------------------- --------------- +epoch -90 +replay_buffer/size 999033 +trainer/num train calls 911000 +trainer/Policy Loss -20.6276 +trainer/Log Pis Mean 23.3143 +trainer/Log Pis Std 13.1719 +trainer/Log Pis Max 66.4699 +trainer/Log Pis Min -8.55172 +trainer/policy/mean Mean -0.026406 +trainer/policy/mean Std 0.90459 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.84432 +trainer/policy/normal/std Std 0.691208 +trainer/policy/normal/std Max 6.39404 +trainer/policy/normal/std Min 0.283129 +trainer/policy/normal/log_std Mean 1.00473 +trainer/policy/normal/log_std Std 0.31804 +trainer/policy/normal/log_std Max 1.85537 +trainer/policy/normal/log_std Min -1.26185 +eval/num steps total 910030 +eval/num paths total 911 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0673268 +eval/Actions Std 0.843122 +eval/Actions Max 0.999997 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65021 +time/logging (s) 0.00368355 +time/sampling batch (s) 0.529029 +time/saving (s) 0.00335602 +time/training (s) 6.92256 +time/epoch (s) 10.1088 +time/total (s) 9350.35 +Epoch -90 +---------------------------------- --------------- +2022-05-10 15:46:51.237540 PDT | [2] Epoch -89 finished +---------------------------------- --------------- +epoch -89 +replay_buffer/size 999033 +trainer/num train calls 912000 +trainer/Policy Loss -20.6533 +trainer/Log Pis Mean 24.1634 +trainer/Log Pis Std 13.7978 +trainer/Log Pis Max 74.5758 +trainer/Log Pis Min -9.24541 +trainer/policy/mean Mean -0.0279166 +trainer/policy/mean Std 0.903828 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999974 +trainer/policy/normal/std Mean 2.81181 +trainer/policy/normal/std Std 0.677155 +trainer/policy/normal/std Max 5.40268 +trainer/policy/normal/std Min 0.31639 +trainer/policy/normal/log_std Mean 0.992078 +trainer/policy/normal/log_std Std 0.325831 +trainer/policy/normal/log_std Max 1.6869 +trainer/policy/normal/log_std Min -1.15078 +eval/num steps total 911030 +eval/num paths total 912 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.474752 +eval/Actions Std 0.872656 +eval/Actions Max 0.999981 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55479 +time/logging (s) 0.00377818 +time/sampling batch (s) 0.275651 +time/saving (s) 0.00333953 +time/training (s) 7.2726 +time/epoch (s) 10.1102 +time/total (s) 9360.46 +Epoch -89 +---------------------------------- --------------- +2022-05-10 15:47:01.226078 PDT | [2] Epoch -88 finished +---------------------------------- --------------- +epoch -88 +replay_buffer/size 999033 +trainer/num train calls 913000 +trainer/Policy Loss -20.3329 +trainer/Log Pis Mean 24.7428 +trainer/Log Pis Std 13.7692 +trainer/Log Pis Max 67.1773 +trainer/Log Pis Min -15.1563 +trainer/policy/mean Mean -0.0172965 +trainer/policy/mean Std 0.901469 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.86382 +trainer/policy/normal/std Std 0.714513 +trainer/policy/normal/std Max 6.08657 +trainer/policy/normal/std Min 0.251052 +trainer/policy/normal/log_std Mean 1.00627 +trainer/policy/normal/log_std Std 0.345136 +trainer/policy/normal/log_std Max 1.80608 +trainer/policy/normal/log_std Min -1.38209 +eval/num steps total 912030 +eval/num paths total 913 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0617985 +eval/Actions Std 0.90095 +eval/Actions Max 1 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70483 +time/logging (s) 0.00371639 +time/sampling batch (s) 0.274362 +time/saving (s) 0.00338607 +time/training (s) 6.978 +time/epoch (s) 9.9643 +time/total (s) 9370.43 +Epoch -88 +---------------------------------- --------------- +2022-05-10 15:47:11.997734 PDT | [2] Epoch -87 finished +---------------------------------- --------------- +epoch -87 +replay_buffer/size 999033 +trainer/num train calls 914000 +trainer/Policy Loss -21.1352 +trainer/Log Pis Mean 25.3176 +trainer/Log Pis Std 13.2539 +trainer/Log Pis Max 69.389 +trainer/Log Pis Min -6.78748 +trainer/policy/mean Mean -0.00361047 +trainer/policy/mean Std 0.906915 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.80072 +trainer/policy/normal/std Std 0.674077 +trainer/policy/normal/std Max 5.94798 +trainer/policy/normal/std Min 0.320476 +trainer/policy/normal/log_std Mean 0.987097 +trainer/policy/normal/log_std Std 0.33196 +trainer/policy/normal/log_std Max 1.78305 +trainer/policy/normal/log_std Min -1.13795 +eval/num steps total 913030 +eval/num paths total 914 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0220142 +eval/Actions Std 0.9045 +eval/Actions Max 0.999996 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62955 +time/logging (s) 0.00397212 +time/sampling batch (s) 0.777041 +time/saving (s) 0.0036611 +time/training (s) 7.33316 +time/epoch (s) 10.7474 +time/total (s) 9381.18 +Epoch -87 +---------------------------------- --------------- +2022-05-10 15:47:23.094359 PDT | [2] Epoch -86 finished +---------------------------------- --------------- +epoch -86 +replay_buffer/size 999033 +trainer/num train calls 915000 +trainer/Policy Loss -18.7248 +trainer/Log Pis Mean 23.9542 +trainer/Log Pis Std 13.3704 +trainer/Log Pis Max 76.3839 +trainer/Log Pis Min -4.1016 +trainer/policy/mean Mean -0.028366 +trainer/policy/mean Std 0.906026 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.81302 +trainer/policy/normal/std Std 0.690538 +trainer/policy/normal/std Max 7.17524 +trainer/policy/normal/std Min 0.291993 +trainer/policy/normal/log_std Mean 0.990814 +trainer/policy/normal/log_std Std 0.33304 +trainer/policy/normal/log_std Max 1.97064 +trainer/policy/normal/log_std Min -1.23102 +eval/num steps total 914030 +eval/num paths total 915 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0680315 +eval/Actions Std 0.863002 +eval/Actions Max 0.999999 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55698 +time/logging (s) 0.00377212 +time/sampling batch (s) 0.28072 +time/saving (s) 0.00351452 +time/training (s) 8.22646 +time/epoch (s) 11.0714 +time/total (s) 9392.25 +Epoch -86 +---------------------------------- --------------- +2022-05-10 15:47:33.748675 PDT | [2] Epoch -85 finished +---------------------------------- --------------- +epoch -85 +replay_buffer/size 999033 +trainer/num train calls 916000 +trainer/Policy Loss -20.055 +trainer/Log Pis Mean 24.4977 +trainer/Log Pis Std 12.9328 +trainer/Log Pis Max 65.7719 +trainer/Log Pis Min -5.6957 +trainer/policy/mean Mean -0.0267658 +trainer/policy/mean Std 0.903745 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.74036 +trainer/policy/normal/std Std 0.667283 +trainer/policy/normal/std Max 5.84907 +trainer/policy/normal/std Min 0.256511 +trainer/policy/normal/log_std Mean 0.964131 +trainer/policy/normal/log_std Std 0.338393 +trainer/policy/normal/log_std Max 1.76628 +trainer/policy/normal/log_std Min -1.36058 +eval/num steps total 915030 +eval/num paths total 916 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.136085 +eval/Actions Std 0.914255 +eval/Actions Max 1 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58075 +time/logging (s) 0.00371378 +time/sampling batch (s) 0.525384 +time/saving (s) 0.0033859 +time/training (s) 7.51666 +time/epoch (s) 10.6299 +time/total (s) 9402.89 +Epoch -85 +---------------------------------- --------------- +2022-05-10 15:47:44.488797 PDT | [2] Epoch -84 finished +---------------------------------- --------------- +epoch -84 +replay_buffer/size 999033 +trainer/num train calls 917000 +trainer/Policy Loss -19.6223 +trainer/Log Pis Mean 25.6305 +trainer/Log Pis Std 13.6633 +trainer/Log Pis Max 77.6221 +trainer/Log Pis Min -8.16633 +trainer/policy/mean Mean -0.0300984 +trainer/policy/mean Std 0.909162 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.83727 +trainer/policy/normal/std Std 0.673608 +trainer/policy/normal/std Max 5.5593 +trainer/policy/normal/std Min 0.328817 +trainer/policy/normal/log_std Mean 1.00201 +trainer/policy/normal/log_std Std 0.323684 +trainer/policy/normal/log_std Max 1.71547 +trainer/policy/normal/log_std Min -1.11225 +eval/num steps total 916030 +eval/num paths total 917 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.369202 +eval/Actions Std 0.667951 +eval/Actions Max 0.999996 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.47946 +time/logging (s) 0.00371701 +time/sampling batch (s) 0.273875 +time/saving (s) 0.00335425 +time/training (s) 7.95538 +time/epoch (s) 10.7158 +time/total (s) 9413.6 +Epoch -84 +---------------------------------- --------------- +2022-05-10 15:47:55.222282 PDT | [2] Epoch -83 finished +---------------------------------- --------------- +epoch -83 +replay_buffer/size 999033 +trainer/num train calls 918000 +trainer/Policy Loss -19.566 +trainer/Log Pis Mean 25.0125 +trainer/Log Pis Std 13.7057 +trainer/Log Pis Max 67.7889 +trainer/Log Pis Min -8.00612 +trainer/policy/mean Mean -0.00267036 +trainer/policy/mean Std 0.904895 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.82777 +trainer/policy/normal/std Std 0.685709 +trainer/policy/normal/std Max 6.63614 +trainer/policy/normal/std Min 0.301506 +trainer/policy/normal/log_std Mean 0.998974 +trainer/policy/normal/log_std Std 0.318072 +trainer/policy/normal/log_std Max 1.89253 +trainer/policy/normal/log_std Min -1.19897 +eval/num steps total 917030 +eval/num paths total 918 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0934045 +eval/Actions Std 0.900991 +eval/Actions Max 0.999994 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58505 +time/logging (s) 0.0037229 +time/sampling batch (s) 0.524148 +time/saving (s) 0.0033663 +time/training (s) 7.59305 +time/epoch (s) 10.7093 +time/total (s) 9424.32 +Epoch -83 +---------------------------------- --------------- +2022-05-10 15:48:05.338942 PDT | [2] Epoch -82 finished +---------------------------------- --------------- +epoch -82 +replay_buffer/size 999033 +trainer/num train calls 919000 +trainer/Policy Loss -19.4177 +trainer/Log Pis Mean 24.9964 +trainer/Log Pis Std 13.5844 +trainer/Log Pis Max 69.0097 +trainer/Log Pis Min -6.53837 +trainer/policy/mean Mean -0.0466626 +trainer/policy/mean Std 0.908227 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.87492 +trainer/policy/normal/std Std 0.697809 +trainer/policy/normal/std Max 5.98466 +trainer/policy/normal/std Min 0.238374 +trainer/policy/normal/log_std Mean 1.01175 +trainer/policy/normal/log_std Std 0.341584 +trainer/policy/normal/log_std Max 1.7892 +trainer/policy/normal/log_std Min -1.43391 +eval/num steps total 918030 +eval/num paths total 919 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0916946 +eval/Actions Std 0.887779 +eval/Actions Max 0.999996 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59733 +time/logging (s) 0.00371735 +time/sampling batch (s) 0.274236 +time/saving (s) 0.00334859 +time/training (s) 7.21389 +time/epoch (s) 10.0925 +time/total (s) 9434.41 +Epoch -82 +---------------------------------- --------------- +2022-05-10 15:48:15.813449 PDT | [2] Epoch -81 finished +---------------------------------- --------------- +epoch -81 +replay_buffer/size 999033 +trainer/num train calls 920000 +trainer/Policy Loss -19.0175 +trainer/Log Pis Mean 24.5359 +trainer/Log Pis Std 13.829 +trainer/Log Pis Max 77.6033 +trainer/Log Pis Min -9.39757 +trainer/policy/mean Mean -0.0381804 +trainer/policy/mean Std 0.905916 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.72563 +trainer/policy/normal/std Std 0.653133 +trainer/policy/normal/std Max 5.75664 +trainer/policy/normal/std Min 0.299505 +trainer/policy/normal/log_std Mean 0.960376 +trainer/policy/normal/log_std Std 0.330253 +trainer/policy/normal/log_std Max 1.75035 +trainer/policy/normal/log_std Min -1.20562 +eval/num steps total 919030 +eval/num paths total 920 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00436634 +eval/Actions Std 0.944494 +eval/Actions Max 0.999995 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58838 +time/logging (s) 0.00368636 +time/sampling batch (s) 0.275988 +time/saving (s) 0.00334938 +time/training (s) 7.57893 +time/epoch (s) 10.4503 +time/total (s) 9444.86 +Epoch -81 +---------------------------------- --------------- +2022-05-10 15:48:25.656237 PDT | [2] Epoch -80 finished +---------------------------------- --------------- +epoch -80 +replay_buffer/size 999033 +trainer/num train calls 921000 +trainer/Policy Loss -19.5644 +trainer/Log Pis Mean 24.2249 +trainer/Log Pis Std 13.3815 +trainer/Log Pis Max 76.66 +trainer/Log Pis Min -11.6722 +trainer/policy/mean Mean -0.0377899 +trainer/policy/mean Std 0.907497 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.7589 +trainer/policy/normal/std Std 0.655968 +trainer/policy/normal/std Max 6.38229 +trainer/policy/normal/std Min 0.277745 +trainer/policy/normal/log_std Mean 0.97258 +trainer/policy/normal/log_std Std 0.332101 +trainer/policy/normal/log_std Max 1.85353 +trainer/policy/normal/log_std Min -1.28105 +eval/num steps total 920030 +eval/num paths total 921 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.324792 +eval/Actions Std 0.695093 +eval/Actions Max 0.999996 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61144 +time/logging (s) 0.00448996 +time/sampling batch (s) 0.275311 +time/saving (s) 0.00376624 +time/training (s) 6.92417 +time/epoch (s) 9.81917 +time/total (s) 9454.69 +Epoch -80 +---------------------------------- --------------- +2022-05-10 15:48:36.536871 PDT | [2] Epoch -79 finished +---------------------------------- --------------- +epoch -79 +replay_buffer/size 999033 +trainer/num train calls 922000 +trainer/Policy Loss -19.4427 +trainer/Log Pis Mean 24.2568 +trainer/Log Pis Std 13.2625 +trainer/Log Pis Max 65.7704 +trainer/Log Pis Min -6.1916 +trainer/policy/mean Mean -0.0106278 +trainer/policy/mean Std 0.904585 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.7768 +trainer/policy/normal/std Std 0.684523 +trainer/policy/normal/std Max 6.00452 +trainer/policy/normal/std Min 0.208867 +trainer/policy/normal/log_std Mean 0.975968 +trainer/policy/normal/log_std Std 0.345176 +trainer/policy/normal/log_std Max 1.79251 +trainer/policy/normal/log_std Min -1.56606 +eval/num steps total 921030 +eval/num paths total 922 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.209183 +eval/Actions Std 0.866849 +eval/Actions Max 0.999998 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.46494 +time/logging (s) 0.00380532 +time/sampling batch (s) 0.277683 +time/saving (s) 0.00351188 +time/training (s) 8.10457 +time/epoch (s) 10.8545 +time/total (s) 9465.54 +Epoch -79 +---------------------------------- --------------- +2022-05-10 15:48:47.204572 PDT | [2] Epoch -78 finished +---------------------------------- --------------- +epoch -78 +replay_buffer/size 999033 +trainer/num train calls 923000 +trainer/Policy Loss -19.3821 +trainer/Log Pis Mean 25.516 +trainer/Log Pis Std 13.9558 +trainer/Log Pis Max 74.599 +trainer/Log Pis Min -6.36219 +trainer/policy/mean Mean -0.0377849 +trainer/policy/mean Std 0.909407 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.81479 +trainer/policy/normal/std Std 0.688109 +trainer/policy/normal/std Max 5.81995 +trainer/policy/normal/std Min 0.278788 +trainer/policy/normal/log_std Mean 0.991356 +trainer/policy/normal/log_std Std 0.335037 +trainer/policy/normal/log_std Max 1.76129 +trainer/policy/normal/log_std Min -1.2773 +eval/num steps total 922030 +eval/num paths total 923 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.106925 +eval/Actions Std 0.912278 +eval/Actions Max 0.999989 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70828 +time/logging (s) 0.00369828 +time/sampling batch (s) 0.275805 +time/saving (s) 0.00336339 +time/training (s) 7.65214 +time/epoch (s) 10.6433 +time/total (s) 9476.19 +Epoch -78 +---------------------------------- --------------- +2022-05-10 15:48:57.307357 PDT | [2] Epoch -77 finished +---------------------------------- --------------- +epoch -77 +replay_buffer/size 999033 +trainer/num train calls 924000 +trainer/Policy Loss -18.9728 +trainer/Log Pis Mean 24.1442 +trainer/Log Pis Std 13.1911 +trainer/Log Pis Max 67.0469 +trainer/Log Pis Min -7.20974 +trainer/policy/mean Mean -0.0416736 +trainer/policy/mean Std 0.907579 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.79654 +trainer/policy/normal/std Std 0.66724 +trainer/policy/normal/std Max 6.22716 +trainer/policy/normal/std Min 0.252309 +trainer/policy/normal/log_std Mean 0.987919 +trainer/policy/normal/log_std Std 0.321265 +trainer/policy/normal/log_std Max 1.82892 +trainer/policy/normal/log_std Min -1.3771 +eval/num steps total 923030 +eval/num paths total 924 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.329517 +eval/Actions Std 0.885225 +eval/Actions Max 0.999992 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69201 +time/logging (s) 0.0036839 +time/sampling batch (s) 0.52592 +time/saving (s) 0.00337037 +time/training (s) 6.85362 +time/epoch (s) 10.0786 +time/total (s) 9486.27 +Epoch -77 +---------------------------------- --------------- +2022-05-10 15:49:08.232816 PDT | [2] Epoch -76 finished +---------------------------------- --------------- +epoch -76 +replay_buffer/size 999033 +trainer/num train calls 925000 +trainer/Policy Loss -19.5682 +trainer/Log Pis Mean 24.35 +trainer/Log Pis Std 13.7505 +trainer/Log Pis Max 75.6675 +trainer/Log Pis Min -8.96576 +trainer/policy/mean Mean -0.0258277 +trainer/policy/mean Std 0.907324 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.79238 +trainer/policy/normal/std Std 0.679023 +trainer/policy/normal/std Max 5.59779 +trainer/policy/normal/std Min 0.272004 +trainer/policy/normal/log_std Mean 0.981018 +trainer/policy/normal/log_std Std 0.349702 +trainer/policy/normal/log_std Max 1.72237 +trainer/policy/normal/log_std Min -1.30194 +eval/num steps total 924030 +eval/num paths total 925 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.216802 +eval/Actions Std 0.822219 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6937 +time/logging (s) 0.00372499 +time/sampling batch (s) 0.274939 +time/saving (s) 0.00333841 +time/training (s) 7.92546 +time/epoch (s) 10.9012 +time/total (s) 9497.18 +Epoch -76 +---------------------------------- --------------- +2022-05-10 15:49:17.682651 PDT | [2] Epoch -75 finished +---------------------------------- --------------- +epoch -75 +replay_buffer/size 999033 +trainer/num train calls 926000 +trainer/Policy Loss -18.8515 +trainer/Log Pis Mean 23.5727 +trainer/Log Pis Std 12.8548 +trainer/Log Pis Max 61.0266 +trainer/Log Pis Min -9.84004 +trainer/policy/mean Mean -0.0267892 +trainer/policy/mean Std 0.906468 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.75938 +trainer/policy/normal/std Std 0.675487 +trainer/policy/normal/std Max 5.6428 +trainer/policy/normal/std Min 0.268702 +trainer/policy/normal/log_std Mean 0.971012 +trainer/policy/normal/log_std Std 0.336889 +trainer/policy/normal/log_std Max 1.73038 +trainer/policy/normal/log_std Min -1.31415 +eval/num steps total 924392 +eval/num paths total 926 +eval/path length Mean 362 +eval/path length Std 0 +eval/path length Max 362 +eval/path length Min 362 +eval/Rewards Mean 0.00276243 +eval/Rewards Std 0.0524862 +eval/Rewards Max 1 +eval/Rewards Min 0 +eval/Returns Mean 1 +eval/Returns Std 0 +eval/Returns Max 1 +eval/Returns Min 1 +eval/Actions Mean 0.0306852 +eval/Actions Std 0.912706 +eval/Actions Max 0.999996 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 1 +eval/normalized_score 100 +time/evaluation sampling (s) 2.45833 +time/logging (s) 0.00193898 +time/sampling batch (s) 0.523747 +time/saving (s) 0.00338705 +time/training (s) 6.4366 +time/epoch (s) 9.424 +time/total (s) 9506.6 +Epoch -75 +---------------------------------- --------------- +2022-05-10 15:49:28.558008 PDT | [2] Epoch -74 finished +---------------------------------- --------------- +epoch -74 +replay_buffer/size 999033 +trainer/num train calls 927000 +trainer/Policy Loss -21.0062 +trainer/Log Pis Mean 25.5033 +trainer/Log Pis Std 13.6066 +trainer/Log Pis Max 71.5455 +trainer/Log Pis Min -9.71478 +trainer/policy/mean Mean -0.0418081 +trainer/policy/mean Std 0.909582 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.82598 +trainer/policy/normal/std Std 0.694262 +trainer/policy/normal/std Max 5.99154 +trainer/policy/normal/std Min 0.235903 +trainer/policy/normal/log_std Mean 0.992721 +trainer/policy/normal/log_std Std 0.350394 +trainer/policy/normal/log_std Max 1.79035 +trainer/policy/normal/log_std Min -1.44433 +eval/num steps total 925392 +eval/num paths total 927 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.166872 +eval/Actions Std 0.805284 +eval/Actions Max 0.999997 +eval/Actions Min -0.999986 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62278 +time/logging (s) 0.00368262 +time/sampling batch (s) 0.274622 +time/saving (s) 0.00337307 +time/training (s) 7.94866 +time/epoch (s) 10.8531 +time/total (s) 9517.46 +Epoch -74 +---------------------------------- --------------- +2022-05-10 15:49:38.638491 PDT | [2] Epoch -73 finished +---------------------------------- --------------- +epoch -73 +replay_buffer/size 999033 +trainer/num train calls 928000 +trainer/Policy Loss -20.6883 +trainer/Log Pis Mean 24.776 +trainer/Log Pis Std 13.2112 +trainer/Log Pis Max 66.8776 +trainer/Log Pis Min -4.8814 +trainer/policy/mean Mean -0.0370583 +trainer/policy/mean Std 0.909467 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.77357 +trainer/policy/normal/std Std 0.652161 +trainer/policy/normal/std Max 5.44239 +trainer/policy/normal/std Min 0.287085 +trainer/policy/normal/log_std Mean 0.978871 +trainer/policy/normal/log_std Std 0.328065 +trainer/policy/normal/log_std Max 1.69422 +trainer/policy/normal/log_std Min -1.24798 +eval/num steps total 926392 +eval/num paths total 928 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0120449 +eval/Actions Std 0.904521 +eval/Actions Max 0.999997 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.48352 +time/logging (s) 0.00403706 +time/sampling batch (s) 0.529233 +time/saving (s) 0.00393229 +time/training (s) 7.03611 +time/epoch (s) 10.0568 +time/total (s) 9527.52 +Epoch -73 +---------------------------------- --------------- +2022-05-10 15:49:48.655192 PDT | [2] Epoch -72 finished +---------------------------------- --------------- +epoch -72 +replay_buffer/size 999033 +trainer/num train calls 929000 +trainer/Policy Loss -18.7871 +trainer/Log Pis Mean 25.2434 +trainer/Log Pis Std 13.5869 +trainer/Log Pis Max 71.7495 +trainer/Log Pis Min -9.33513 +trainer/policy/mean Mean -0.022616 +trainer/policy/mean Std 0.902362 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.83148 +trainer/policy/normal/std Std 0.699366 +trainer/policy/normal/std Max 6.43501 +trainer/policy/normal/std Min 0.288291 +trainer/policy/normal/log_std Mean 0.995989 +trainer/policy/normal/log_std Std 0.340384 +trainer/policy/normal/log_std Max 1.86175 +trainer/policy/normal/log_std Min -1.24378 +eval/num steps total 927392 +eval/num paths total 929 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.264312 +eval/Actions Std 0.825981 +eval/Actions Max 0.999997 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.36536 +time/logging (s) 0.00371208 +time/sampling batch (s) 1.02327 +time/saving (s) 0.00339384 +time/training (s) 6.59632 +time/epoch (s) 9.99205 +time/total (s) 9537.51 +Epoch -72 +---------------------------------- --------------- +2022-05-10 15:49:58.261584 PDT | [2] Epoch -71 finished +---------------------------------- --------------- +epoch -71 +replay_buffer/size 999033 +trainer/num train calls 930000 +trainer/Policy Loss -19.1531 +trainer/Log Pis Mean 24.9283 +trainer/Log Pis Std 13.2335 +trainer/Log Pis Max 63.2247 +trainer/Log Pis Min -8.05437 +trainer/policy/mean Mean -0.0197534 +trainer/policy/mean Std 0.911177 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.79635 +trainer/policy/normal/std Std 0.671621 +trainer/policy/normal/std Max 6.34861 +trainer/policy/normal/std Min 0.239302 +trainer/policy/normal/log_std Mean 0.985765 +trainer/policy/normal/log_std Std 0.332947 +trainer/policy/normal/log_std Max 1.84824 +trainer/policy/normal/log_std Min -1.43003 +eval/num steps total 928392 +eval/num paths total 930 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.224512 +eval/Actions Std 0.931587 +eval/Actions Max 0.999997 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66019 +time/logging (s) 0.00373884 +time/sampling batch (s) 0.27175 +time/saving (s) 0.0033325 +time/training (s) 6.64348 +time/epoch (s) 9.58249 +time/total (s) 9547.1 +Epoch -71 +---------------------------------- --------------- +2022-05-10 15:50:09.389263 PDT | [2] Epoch -70 finished +---------------------------------- --------------- +epoch -70 +replay_buffer/size 999033 +trainer/num train calls 931000 +trainer/Policy Loss -20.4068 +trainer/Log Pis Mean 23.9671 +trainer/Log Pis Std 12.9257 +trainer/Log Pis Max 68.0748 +trainer/Log Pis Min -9.10685 +trainer/policy/mean Mean -0.0211101 +trainer/policy/mean Std 0.905897 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.82336 +trainer/policy/normal/std Std 0.698049 +trainer/policy/normal/std Max 5.74458 +trainer/policy/normal/std Min 0.231656 +trainer/policy/normal/log_std Mean 0.99265 +trainer/policy/normal/log_std Std 0.343983 +trainer/policy/normal/log_std Max 1.74826 +trainer/policy/normal/log_std Min -1.4625 +eval/num steps total 929392 +eval/num paths total 931 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.153043 +eval/Actions Std 0.870544 +eval/Actions Max 0.999999 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.7054 +time/logging (s) 0.00371354 +time/sampling batch (s) 0.276055 +time/saving (s) 0.00335323 +time/training (s) 8.11504 +time/epoch (s) 11.1036 +time/total (s) 9558.21 +Epoch -70 +---------------------------------- --------------- +2022-05-10 15:50:22.201066 PDT | [2] Epoch -69 finished +---------------------------------- --------------- +epoch -69 +replay_buffer/size 999033 +trainer/num train calls 932000 +trainer/Policy Loss -19.0075 +trainer/Log Pis Mean 24.4514 +trainer/Log Pis Std 13.4671 +trainer/Log Pis Max 68.9053 +trainer/Log Pis Min -9.1008 +trainer/policy/mean Mean -0.0399649 +trainer/policy/mean Std 0.902064 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.78656 +trainer/policy/normal/std Std 0.682175 +trainer/policy/normal/std Max 6.92375 +trainer/policy/normal/std Min 0.249584 +trainer/policy/normal/log_std Mean 0.981668 +trainer/policy/normal/log_std Std 0.333237 +trainer/policy/normal/log_std Max 1.93496 +trainer/policy/normal/log_std Min -1.38796 +eval/num steps total 930392 +eval/num paths total 932 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0430668 +eval/Actions Std 0.913249 +eval/Actions Max 0.999995 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.88469 +time/logging (s) 0.00407784 +time/sampling batch (s) 0.279154 +time/saving (s) 0.00370498 +time/training (s) 9.61621 +time/epoch (s) 12.7878 +time/total (s) 9571 +Epoch -69 +---------------------------------- --------------- +2022-05-10 15:50:32.477436 PDT | [2] Epoch -68 finished +---------------------------------- --------------- +epoch -68 +replay_buffer/size 999033 +trainer/num train calls 933000 +trainer/Policy Loss -20.6771 +trainer/Log Pis Mean 25.6769 +trainer/Log Pis Std 12.9839 +trainer/Log Pis Max 66.0497 +trainer/Log Pis Min -4.70305 +trainer/policy/mean Mean -0.0661512 +trainer/policy/mean Std 0.907051 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.8238 +trainer/policy/normal/std Std 0.674121 +trainer/policy/normal/std Max 5.45555 +trainer/policy/normal/std Min 0.303369 +trainer/policy/normal/log_std Mean 0.996179 +trainer/policy/normal/log_std Std 0.329366 +trainer/policy/normal/log_std Max 1.69663 +trainer/policy/normal/log_std Min -1.19281 +eval/num steps total 931392 +eval/num paths total 933 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.076893 +eval/Actions Std 0.924528 +eval/Actions Max 0.999995 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67223 +time/logging (s) 0.00373659 +time/sampling batch (s) 0.274794 +time/saving (s) 0.00338647 +time/training (s) 7.29737 +time/epoch (s) 10.2515 +time/total (s) 9581.25 +Epoch -68 +---------------------------------- --------------- +2022-05-10 15:50:43.366928 PDT | [2] Epoch -67 finished +---------------------------------- --------------- +epoch -67 +replay_buffer/size 999033 +trainer/num train calls 934000 +trainer/Policy Loss -18.5574 +trainer/Log Pis Mean 25.668 +trainer/Log Pis Std 13.3357 +trainer/Log Pis Max 61.4003 +trainer/Log Pis Min -11.0403 +trainer/policy/mean Mean -0.0385892 +trainer/policy/mean Std 0.905553 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.87384 +trainer/policy/normal/std Std 0.698213 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.31438 +trainer/policy/normal/log_std Mean 1.01424 +trainer/policy/normal/log_std Std 0.323227 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.15715 +eval/num steps total 932392 +eval/num paths total 934 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.143875 +eval/Actions Std 0.88622 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66167 +time/logging (s) 0.00373927 +time/sampling batch (s) 0.526751 +time/saving (s) 0.00336351 +time/training (s) 7.66952 +time/epoch (s) 10.865 +time/total (s) 9592.12 +Epoch -67 +---------------------------------- --------------- +2022-05-10 15:50:53.725061 PDT | [2] Epoch -66 finished +---------------------------------- --------------- +epoch -66 +replay_buffer/size 999033 +trainer/num train calls 935000 +trainer/Policy Loss -20.0362 +trainer/Log Pis Mean 23.8702 +trainer/Log Pis Std 12.8595 +trainer/Log Pis Max 61.798 +trainer/Log Pis Min -7.60553 +trainer/policy/mean Mean -0.0436966 +trainer/policy/mean Std 0.906676 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.85068 +trainer/policy/normal/std Std 0.685776 +trainer/policy/normal/std Max 6.25703 +trainer/policy/normal/std Min 0.262287 +trainer/policy/normal/log_std Mean 1.00467 +trainer/policy/normal/log_std Std 0.335179 +trainer/policy/normal/log_std Max 1.83371 +trainer/policy/normal/log_std Min -1.33832 +eval/num steps total 933392 +eval/num paths total 935 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.23825 +eval/Actions Std 0.900486 +eval/Actions Max 0.999999 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.47689 +time/logging (s) 0.00377349 +time/sampling batch (s) 0.290433 +time/saving (s) 0.00383868 +time/training (s) 7.55803 +time/epoch (s) 10.333 +time/total (s) 9602.46 +Epoch -66 +---------------------------------- --------------- +2022-05-10 15:51:03.507615 PDT | [2] Epoch -65 finished +---------------------------------- --------------- +epoch -65 +replay_buffer/size 999033 +trainer/num train calls 936000 +trainer/Policy Loss -19.4724 +trainer/Log Pis Mean 24.6243 +trainer/Log Pis Std 13.1096 +trainer/Log Pis Max 67.2896 +trainer/Log Pis Min -7.32814 +trainer/policy/mean Mean -0.0202918 +trainer/policy/mean Std 0.912703 +trainer/policy/mean Max 0.999981 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.75756 +trainer/policy/normal/std Std 0.654382 +trainer/policy/normal/std Max 6.09136 +trainer/policy/normal/std Min 0.273683 +trainer/policy/normal/log_std Mean 0.973456 +trainer/policy/normal/log_std Std 0.32279 +trainer/policy/normal/log_std Max 1.80687 +trainer/policy/normal/log_std Min -1.29579 +eval/num steps total 934392 +eval/num paths total 936 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.341934 +eval/Actions Std 0.854134 +eval/Actions Max 0.999981 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59742 +time/logging (s) 0.00404327 +time/sampling batch (s) 0.284383 +time/saving (s) 0.00374449 +time/training (s) 6.86775 +time/epoch (s) 9.75734 +time/total (s) 9612.22 +Epoch -65 +---------------------------------- --------------- +2022-05-10 15:51:13.173955 PDT | [2] Epoch -64 finished +---------------------------------- --------------- +epoch -64 +replay_buffer/size 999033 +trainer/num train calls 937000 +trainer/Policy Loss -20.7738 +trainer/Log Pis Mean 25.3926 +trainer/Log Pis Std 13.4925 +trainer/Log Pis Max 63.2348 +trainer/Log Pis Min -13.53 +trainer/policy/mean Mean -0.0442872 +trainer/policy/mean Std 0.907663 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.80608 +trainer/policy/normal/std Std 0.671087 +trainer/policy/normal/std Max 6.35041 +trainer/policy/normal/std Min 0.310456 +trainer/policy/normal/log_std Mean 0.990604 +trainer/policy/normal/log_std Std 0.324117 +trainer/policy/normal/log_std Max 1.84852 +trainer/policy/normal/log_std Min -1.16971 +eval/num steps total 935392 +eval/num paths total 937 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.225267 +eval/Actions Std 0.898603 +eval/Actions Max 0.999996 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.44945 +time/logging (s) 0.00395639 +time/sampling batch (s) 0.52631 +time/saving (s) 0.00363918 +time/training (s) 6.65817 +time/epoch (s) 9.64152 +time/total (s) 9621.86 +Epoch -64 +---------------------------------- --------------- +2022-05-10 15:51:23.803821 PDT | [2] Epoch -63 finished +---------------------------------- --------------- +epoch -63 +replay_buffer/size 999033 +trainer/num train calls 938000 +trainer/Policy Loss -19.9104 +trainer/Log Pis Mean 23.4317 +trainer/Log Pis Std 12.9045 +trainer/Log Pis Max 65.5611 +trainer/Log Pis Min -10.8691 +trainer/policy/mean Mean -0.0222206 +trainer/policy/mean Std 0.90838 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.80509 +trainer/policy/normal/std Std 0.641719 +trainer/policy/normal/std Max 5.85864 +trainer/policy/normal/std Min 0.25226 +trainer/policy/normal/log_std Mean 0.994067 +trainer/policy/normal/log_std Std 0.308301 +trainer/policy/normal/log_std Max 1.76792 +trainer/policy/normal/log_std Min -1.37729 +eval/num steps total 936392 +eval/num paths total 938 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0204002 +eval/Actions Std 0.904051 +eval/Actions Max 0.999999 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68977 +time/logging (s) 0.00391885 +time/sampling batch (s) 0.52644 +time/saving (s) 0.00361872 +time/training (s) 7.38139 +time/epoch (s) 10.6051 +time/total (s) 9632.47 +Epoch -63 +---------------------------------- --------------- +2022-05-10 15:51:35.676820 PDT | [2] Epoch -62 finished +---------------------------------- --------------- +epoch -62 +replay_buffer/size 999033 +trainer/num train calls 939000 +trainer/Policy Loss -19.8151 +trainer/Log Pis Mean 24.4385 +trainer/Log Pis Std 13.9497 +trainer/Log Pis Max 70.246 +trainer/Log Pis Min -5.5359 +trainer/policy/mean Mean -0.0228939 +trainer/policy/mean Std 0.90616 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.8341 +trainer/policy/normal/std Std 0.68242 +trainer/policy/normal/std Max 5.99588 +trainer/policy/normal/std Min 0.283709 +trainer/policy/normal/log_std Mean 1.00036 +trainer/policy/normal/log_std Std 0.323525 +trainer/policy/normal/log_std Max 1.79107 +trainer/policy/normal/log_std Min -1.25981 +eval/num steps total 937392 +eval/num paths total 939 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0870631 +eval/Actions Std 0.916197 +eval/Actions Max 0.999997 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68007 +time/logging (s) 0.00368751 +time/sampling batch (s) 1.28556 +time/saving (s) 0.00332574 +time/training (s) 7.87553 +time/epoch (s) 11.8482 +time/total (s) 9644.32 +Epoch -62 +---------------------------------- --------------- +2022-05-10 15:51:46.735887 PDT | [2] Epoch -61 finished +---------------------------------- --------------- +epoch -61 +replay_buffer/size 999033 +trainer/num train calls 940000 +trainer/Policy Loss -19.8269 +trainer/Log Pis Mean 24.217 +trainer/Log Pis Std 13.308 +trainer/Log Pis Max 64.8658 +trainer/Log Pis Min -12.2907 +trainer/policy/mean Mean -0.0402794 +trainer/policy/mean Std 0.912471 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.78261 +trainer/policy/normal/std Std 0.671317 +trainer/policy/normal/std Max 6.07841 +trainer/policy/normal/std Min 0.266217 +trainer/policy/normal/log_std Mean 0.980107 +trainer/policy/normal/log_std Std 0.335447 +trainer/policy/normal/log_std Max 1.80474 +trainer/policy/normal/log_std Min -1.32344 +eval/num steps total 938392 +eval/num paths total 940 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00338211 +eval/Actions Std 0.936291 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61984 +time/logging (s) 0.00446264 +time/sampling batch (s) 0.772819 +time/saving (s) 0.0035927 +time/training (s) 7.6353 +time/epoch (s) 11.036 +time/total (s) 9655.36 +Epoch -61 +---------------------------------- --------------- +2022-05-10 15:51:58.938876 PDT | [2] Epoch -60 finished +---------------------------------- --------------- +epoch -60 +replay_buffer/size 999033 +trainer/num train calls 941000 +trainer/Policy Loss -19.7541 +trainer/Log Pis Mean 24.5852 +trainer/Log Pis Std 13.1031 +trainer/Log Pis Max 91.5159 +trainer/Log Pis Min -8.33075 +trainer/policy/mean Mean -0.0145084 +trainer/policy/mean Std 0.907662 +trainer/policy/mean Max 0.999978 +trainer/policy/mean Min -0.999982 +trainer/policy/normal/std Mean 2.80285 +trainer/policy/normal/std Std 0.681424 +trainer/policy/normal/std Max 6.34088 +trainer/policy/normal/std Min 0.330795 +trainer/policy/normal/log_std Mean 0.987927 +trainer/policy/normal/log_std Std 0.330728 +trainer/policy/normal/log_std Max 1.84702 +trainer/policy/normal/log_std Min -1.10626 +eval/num steps total 939392 +eval/num paths total 941 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.280924 +eval/Actions Std 0.864371 +eval/Actions Max 0.999979 +eval/Actions Min -0.999985 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.71436 +time/logging (s) 0.00369519 +time/sampling batch (s) 0.772345 +time/saving (s) 0.00331954 +time/training (s) 8.6837 +time/epoch (s) 12.1774 +time/total (s) 9667.54 +Epoch -60 +---------------------------------- --------------- +2022-05-10 15:52:08.048805 PDT | [2] Epoch -59 finished +---------------------------------- --------------- +epoch -59 +replay_buffer/size 999033 +trainer/num train calls 942000 +trainer/Policy Loss -20.1254 +trainer/Log Pis Mean 24.1534 +trainer/Log Pis Std 13.027 +trainer/Log Pis Max 80.002 +trainer/Log Pis Min -6.81139 +trainer/policy/mean Mean -0.0743815 +trainer/policy/mean Std 0.903793 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.81088 +trainer/policy/normal/std Std 0.669449 +trainer/policy/normal/std Max 7.03808 +trainer/policy/normal/std Min 0.225226 +trainer/policy/normal/log_std Mean 0.991255 +trainer/policy/normal/log_std Std 0.332306 +trainer/policy/normal/log_std Max 1.95134 +trainer/policy/normal/log_std Min -1.49065 +eval/num steps total 940392 +eval/num paths total 942 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.152393 +eval/Actions Std 0.887809 +eval/Actions Max 0.999986 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.4105 +time/logging (s) 0.00384859 +time/sampling batch (s) 0.520511 +time/saving (s) 0.00335015 +time/training (s) 6.14813 +time/epoch (s) 9.08635 +time/total (s) 9676.63 +Epoch -59 +---------------------------------- --------------- +2022-05-10 15:52:17.728868 PDT | [2] Epoch -58 finished +---------------------------------- --------------- +epoch -58 +replay_buffer/size 999033 +trainer/num train calls 943000 +trainer/Policy Loss -20.9184 +trainer/Log Pis Mean 25.1368 +trainer/Log Pis Std 13.6182 +trainer/Log Pis Max 74.6058 +trainer/Log Pis Min -5.78136 +trainer/policy/mean Mean -0.055723 +trainer/policy/mean Std 0.910715 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.79242 +trainer/policy/normal/std Std 0.670521 +trainer/policy/normal/std Max 5.92079 +trainer/policy/normal/std Min 0.269593 +trainer/policy/normal/log_std Mean 0.984116 +trainer/policy/normal/log_std Std 0.333538 +trainer/policy/normal/log_std Max 1.77847 +trainer/policy/normal/log_std Min -1.31084 +eval/num steps total 941392 +eval/num paths total 943 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.237435 +eval/Actions Std 0.934816 +eval/Actions Max 0.999981 +eval/Actions Min -0.999981 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.86922 +time/logging (s) 0.00390019 +time/sampling batch (s) 0.270857 +time/saving (s) 0.00361572 +time/training (s) 6.50856 +time/epoch (s) 9.65615 +time/total (s) 9686.29 +Epoch -58 +---------------------------------- --------------- +2022-05-10 15:52:29.413274 PDT | [2] Epoch -57 finished +---------------------------------- --------------- +epoch -57 +replay_buffer/size 999033 +trainer/num train calls 944000 +trainer/Policy Loss -19.4198 +trainer/Log Pis Mean 24.5584 +trainer/Log Pis Std 13.215 +trainer/Log Pis Max 74.0549 +trainer/Log Pis Min -8.98965 +trainer/policy/mean Mean -0.0333117 +trainer/policy/mean Std 0.911805 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.85333 +trainer/policy/normal/std Std 0.675404 +trainer/policy/normal/std Max 7.15274 +trainer/policy/normal/std Min 0.245087 +trainer/policy/normal/log_std Mean 1.00831 +trainer/policy/normal/log_std Std 0.320565 +trainer/policy/normal/log_std Max 1.9675 +trainer/policy/normal/log_std Min -1.40614 +eval/num steps total 942392 +eval/num paths total 944 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.115471 +eval/Actions Std 0.908026 +eval/Actions Max 0.999985 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68052 +time/logging (s) 0.00368857 +time/sampling batch (s) 0.774048 +time/saving (s) 0.00337489 +time/training (s) 8.19811 +time/epoch (s) 11.6597 +time/total (s) 9697.95 +Epoch -57 +---------------------------------- --------------- +2022-05-10 15:52:38.969542 PDT | [2] Epoch -56 finished +---------------------------------- --------------- +epoch -56 +replay_buffer/size 999033 +trainer/num train calls 945000 +trainer/Policy Loss -20.1607 +trainer/Log Pis Mean 24.079 +trainer/Log Pis Std 13.0171 +trainer/Log Pis Max 63.2456 +trainer/Log Pis Min -8.35562 +trainer/policy/mean Mean -0.029575 +trainer/policy/mean Std 0.902126 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.7354 +trainer/policy/normal/std Std 0.668681 +trainer/policy/normal/std Max 6.04805 +trainer/policy/normal/std Min 0.267225 +trainer/policy/normal/log_std Mean 0.963447 +trainer/policy/normal/log_std Std 0.330139 +trainer/policy/normal/log_std Max 1.79974 +trainer/policy/normal/log_std Min -1.31966 +eval/num steps total 943392 +eval/num paths total 945 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0473485 +eval/Actions Std 0.883032 +eval/Actions Max 0.999994 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60925 +time/logging (s) 0.00370792 +time/sampling batch (s) 0.274573 +time/saving (s) 0.00335222 +time/training (s) 6.64099 +time/epoch (s) 9.53188 +time/total (s) 9707.48 +Epoch -56 +---------------------------------- --------------- +2022-05-10 15:52:48.803063 PDT | [2] Epoch -55 finished +---------------------------------- --------------- +epoch -55 +replay_buffer/size 999033 +trainer/num train calls 946000 +trainer/Policy Loss -19.6118 +trainer/Log Pis Mean 23.5178 +trainer/Log Pis Std 13.4518 +trainer/Log Pis Max 72.0156 +trainer/Log Pis Min -10.2052 +trainer/policy/mean Mean -0.0441903 +trainer/policy/mean Std 0.90121 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.76872 +trainer/policy/normal/std Std 0.70756 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.25722 +trainer/policy/normal/log_std Mean 0.970644 +trainer/policy/normal/log_std Std 0.350865 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.35782 +eval/num steps total 944392 +eval/num paths total 946 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.110014 +eval/Actions Std 0.914495 +eval/Actions Max 0.999987 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.41221 +time/logging (s) 0.00370402 +time/sampling batch (s) 0.523389 +time/saving (s) 0.00336284 +time/training (s) 6.8668 +time/epoch (s) 9.80947 +time/total (s) 9717.3 +Epoch -55 +---------------------------------- --------------- +2022-05-10 15:52:59.341101 PDT | [2] Epoch -54 finished +---------------------------------- --------------- +epoch -54 +replay_buffer/size 999033 +trainer/num train calls 947000 +trainer/Policy Loss -19.9941 +trainer/Log Pis Mean 24.4415 +trainer/Log Pis Std 12.9494 +trainer/Log Pis Max 72.8645 +trainer/Log Pis Min -5.74626 +trainer/policy/mean Mean -0.0481255 +trainer/policy/mean Std 0.907906 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.7976 +trainer/policy/normal/std Std 0.6551 +trainer/policy/normal/std Max 5.89675 +trainer/policy/normal/std Min 0.311754 +trainer/policy/normal/log_std Mean 0.989975 +trainer/policy/normal/log_std Std 0.312424 +trainer/policy/normal/log_std Max 1.7744 +trainer/policy/normal/log_std Min -1.16554 +eval/num steps total 945392 +eval/num paths total 947 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.200442 +eval/Actions Std 0.91786 +eval/Actions Max 0.999992 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.46096 +time/logging (s) 0.0036851 +time/sampling batch (s) 0.523707 +time/saving (s) 0.00335349 +time/training (s) 7.52218 +time/epoch (s) 10.5139 +time/total (s) 9727.81 +Epoch -54 +---------------------------------- --------------- +2022-05-10 15:53:10.515333 PDT | [2] Epoch -53 finished +---------------------------------- --------------- +epoch -53 +replay_buffer/size 999033 +trainer/num train calls 948000 +trainer/Policy Loss -19.036 +trainer/Log Pis Mean 25.2032 +trainer/Log Pis Std 13.1644 +trainer/Log Pis Max 73.438 +trainer/Log Pis Min -5.68768 +trainer/policy/mean Mean -0.0177083 +trainer/policy/mean Std 0.908287 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.78158 +trainer/policy/normal/std Std 0.704437 +trainer/policy/normal/std Max 5.69471 +trainer/policy/normal/std Min 0.225512 +trainer/policy/normal/log_std Mean 0.973611 +trainer/policy/normal/log_std Std 0.362638 +trainer/policy/normal/log_std Max 1.73954 +trainer/policy/normal/log_std Min -1.48938 +eval/num steps total 946392 +eval/num paths total 948 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0104473 +eval/Actions Std 0.909354 +eval/Actions Max 0.999996 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5377 +time/logging (s) 0.0038202 +time/sampling batch (s) 0.527883 +time/saving (s) 0.00360242 +time/training (s) 8.0768 +time/epoch (s) 11.1498 +time/total (s) 9738.97 +Epoch -53 +---------------------------------- --------------- +2022-05-10 15:53:20.779826 PDT | [2] Epoch -52 finished +---------------------------------- --------------- +epoch -52 +replay_buffer/size 999033 +trainer/num train calls 949000 +trainer/Policy Loss -19.8824 +trainer/Log Pis Mean 25.4827 +trainer/Log Pis Std 13.2892 +trainer/Log Pis Max 71.9812 +trainer/Log Pis Min -12.3343 +trainer/policy/mean Mean -0.0389214 +trainer/policy/mean Std 0.907249 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.82043 +trainer/policy/normal/std Std 0.678824 +trainer/policy/normal/std Max 5.56543 +trainer/policy/normal/std Min 0.272217 +trainer/policy/normal/log_std Mean 0.994976 +trainer/policy/normal/log_std Std 0.327455 +trainer/policy/normal/log_std Max 1.71657 +trainer/policy/normal/log_std Min -1.30116 +eval/num steps total 947392 +eval/num paths total 949 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.113937 +eval/Actions Std 0.877866 +eval/Actions Max 0.999993 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63955 +time/logging (s) 0.00378698 +time/sampling batch (s) 0.276791 +time/saving (s) 0.00343362 +time/training (s) 7.31613 +time/epoch (s) 10.2397 +time/total (s) 9749.21 +Epoch -52 +---------------------------------- --------------- +2022-05-10 15:53:31.238171 PDT | [2] Epoch -51 finished +---------------------------------- --------------- +epoch -51 +replay_buffer/size 999033 +trainer/num train calls 950000 +trainer/Policy Loss -20.081 +trainer/Log Pis Mean 24.141 +trainer/Log Pis Std 13.2775 +trainer/Log Pis Max 82.3783 +trainer/Log Pis Min -9.27873 +trainer/policy/mean Mean -0.0322998 +trainer/policy/mean Std 0.908629 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.76616 +trainer/policy/normal/std Std 0.675081 +trainer/policy/normal/std Max 6.26124 +trainer/policy/normal/std Min 0.260002 +trainer/policy/normal/log_std Mean 0.974099 +trainer/policy/normal/log_std Std 0.333914 +trainer/policy/normal/log_std Max 1.83438 +trainer/policy/normal/log_std Min -1.34706 +eval/num steps total 948392 +eval/num paths total 950 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00993367 +eval/Actions Std 0.906928 +eval/Actions Max 0.999996 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.49807 +time/logging (s) 0.00369265 +time/sampling batch (s) 0.277855 +time/saving (s) 0.00337728 +time/training (s) 7.65053 +time/epoch (s) 10.4335 +time/total (s) 9759.65 +Epoch -51 +---------------------------------- --------------- +2022-05-10 15:53:40.855217 PDT | [2] Epoch -50 finished +---------------------------------- --------------- +epoch -50 +replay_buffer/size 999033 +trainer/num train calls 951000 +trainer/Policy Loss -18.908 +trainer/Log Pis Mean 24.1459 +trainer/Log Pis Std 13.701 +trainer/Log Pis Max 62.5037 +trainer/Log Pis Min -14.5153 +trainer/policy/mean Mean -0.0244307 +trainer/policy/mean Std 0.909656 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.83689 +trainer/policy/normal/std Std 0.675269 +trainer/policy/normal/std Max 6.43747 +trainer/policy/normal/std Min 0.240513 +trainer/policy/normal/log_std Mean 1.00167 +trainer/policy/normal/log_std Std 0.324233 +trainer/policy/normal/log_std Max 1.86214 +trainer/policy/normal/log_std Min -1.42498 +eval/num steps total 949392 +eval/num paths total 951 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.183363 +eval/Actions Std 0.864962 +eval/Actions Max 0.999999 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.52597 +time/logging (s) 0.00368148 +time/sampling batch (s) 0.276902 +time/saving (s) 0.00350688 +time/training (s) 6.78226 +time/epoch (s) 9.59233 +time/total (s) 9769.24 +Epoch -50 +---------------------------------- --------------- +2022-05-10 15:53:50.917937 PDT | [2] Epoch -49 finished +---------------------------------- --------------- +epoch -49 +replay_buffer/size 999033 +trainer/num train calls 952000 +trainer/Policy Loss -19.5367 +trainer/Log Pis Mean 25.2738 +trainer/Log Pis Std 13.3091 +trainer/Log Pis Max 73.3602 +trainer/Log Pis Min -4.8474 +trainer/policy/mean Mean -0.0462785 +trainer/policy/mean Std 0.904726 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.79699 +trainer/policy/normal/std Std 0.669863 +trainer/policy/normal/std Max 5.62192 +trainer/policy/normal/std Min 0.272255 +trainer/policy/normal/log_std Mean 0.98619 +trainer/policy/normal/log_std Std 0.330921 +trainer/policy/normal/log_std Max 1.72667 +trainer/policy/normal/log_std Min -1.30101 +eval/num steps total 950392 +eval/num paths total 952 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.232437 +eval/Actions Std 0.753671 +eval/Actions Max 0.999992 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53272 +time/logging (s) 0.00374657 +time/sampling batch (s) 0.5264 +time/saving (s) 0.0033767 +time/training (s) 6.9721 +time/epoch (s) 10.0384 +time/total (s) 9779.28 +Epoch -49 +---------------------------------- --------------- +2022-05-10 15:54:00.962480 PDT | [2] Epoch -48 finished +---------------------------------- --------------- +epoch -48 +replay_buffer/size 999033 +trainer/num train calls 953000 +trainer/Policy Loss -19.8375 +trainer/Log Pis Mean 24.3586 +trainer/Log Pis Std 12.5651 +trainer/Log Pis Max 75.0873 +trainer/Log Pis Min -6.46911 +trainer/policy/mean Mean -0.0469934 +trainer/policy/mean Std 0.903211 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.78676 +trainer/policy/normal/std Std 0.718787 +trainer/policy/normal/std Max 6.4258 +trainer/policy/normal/std Min 0.202808 +trainer/policy/normal/log_std Mean 0.972158 +trainer/policy/normal/log_std Std 0.378304 +trainer/policy/normal/log_std Max 1.86032 +trainer/policy/normal/log_std Min -1.5955 +eval/num steps total 951392 +eval/num paths total 953 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.364203 +eval/Actions Std 0.808656 +eval/Actions Max 0.999998 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66371 +time/logging (s) 0.00377689 +time/sampling batch (s) 0.275483 +time/saving (s) 0.00334776 +time/training (s) 7.074 +time/epoch (s) 10.0203 +time/total (s) 9789.31 +Epoch -48 +---------------------------------- --------------- +2022-05-10 15:54:10.005708 PDT | [2] Epoch -47 finished +---------------------------------- --------------- +epoch -47 +replay_buffer/size 999033 +trainer/num train calls 954000 +trainer/Policy Loss -20.5394 +trainer/Log Pis Mean 24.767 +trainer/Log Pis Std 12.5784 +trainer/Log Pis Max 71.6486 +trainer/Log Pis Min -10.8042 +trainer/policy/mean Mean -0.02854 +trainer/policy/mean Std 0.908306 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.79093 +trainer/policy/normal/std Std 0.666578 +trainer/policy/normal/std Max 6.40142 +trainer/policy/normal/std Min 0.265208 +trainer/policy/normal/log_std Mean 0.984769 +trainer/policy/normal/log_std Std 0.328146 +trainer/policy/normal/log_std Max 1.85652 +trainer/policy/normal/log_std Min -1.32724 +eval/num steps total 952392 +eval/num paths total 954 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.268776 +eval/Actions Std 0.891873 +eval/Actions Max 0.999993 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53445 +time/logging (s) 0.00386267 +time/sampling batch (s) 0.274375 +time/saving (s) 0.00342023 +time/training (s) 6.20292 +time/epoch (s) 9.01903 +time/total (s) 9798.33 +Epoch -47 +---------------------------------- --------------- +2022-05-10 15:54:20.396044 PDT | [2] Epoch -46 finished +---------------------------------- --------------- +epoch -46 +replay_buffer/size 999033 +trainer/num train calls 955000 +trainer/Policy Loss -19.8442 +trainer/Log Pis Mean 24.2762 +trainer/Log Pis Std 13.8117 +trainer/Log Pis Max 75.9309 +trainer/Log Pis Min -7.70012 +trainer/policy/mean Mean -0.03854 +trainer/policy/mean Std 0.90545 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.7548 +trainer/policy/normal/std Std 0.649063 +trainer/policy/normal/std Max 5.63712 +trainer/policy/normal/std Min 0.301251 +trainer/policy/normal/log_std Mean 0.973474 +trainer/policy/normal/log_std Std 0.31801 +trainer/policy/normal/log_std Max 1.72937 +trainer/policy/normal/log_std Min -1.19981 +eval/num steps total 953392 +eval/num paths total 955 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.254153 +eval/Actions Std 0.915381 +eval/Actions Max 0.99998 +eval/Actions Min -0.999981 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5164 +time/logging (s) 0.00396335 +time/sampling batch (s) 0.528319 +time/saving (s) 0.00358641 +time/training (s) 7.31364 +time/epoch (s) 10.3659 +time/total (s) 9808.7 +Epoch -46 +---------------------------------- --------------- +2022-05-10 15:54:30.874162 PDT | [2] Epoch -45 finished +---------------------------------- --------------- +epoch -45 +replay_buffer/size 999033 +trainer/num train calls 956000 +trainer/Policy Loss -20.7487 +trainer/Log Pis Mean 25.5007 +trainer/Log Pis Std 13.4587 +trainer/Log Pis Max 73.7398 +trainer/Log Pis Min -6.87712 +trainer/policy/mean Mean -0.065529 +trainer/policy/mean Std 0.909255 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.8106 +trainer/policy/normal/std Std 0.686191 +trainer/policy/normal/std Max 6.72236 +trainer/policy/normal/std Min 0.290169 +trainer/policy/normal/log_std Mean 0.988917 +trainer/policy/normal/log_std Std 0.340248 +trainer/policy/normal/log_std Max 1.90544 +trainer/policy/normal/log_std Min -1.23729 +eval/num steps total 954392 +eval/num paths total 956 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.200095 +eval/Actions Std 0.948424 +eval/Actions Max 0.999992 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.46909 +time/logging (s) 0.00382781 +time/sampling batch (s) 1.02697 +time/saving (s) 0.00347335 +time/training (s) 6.94979 +time/epoch (s) 10.4532 +time/total (s) 9819.15 +Epoch -45 +---------------------------------- --------------- +2022-05-10 15:54:40.192861 PDT | [2] Epoch -44 finished +---------------------------------- --------------- +epoch -44 +replay_buffer/size 999033 +trainer/num train calls 957000 +trainer/Policy Loss -20.5877 +trainer/Log Pis Mean 24.7623 +trainer/Log Pis Std 13.456 +trainer/Log Pis Max 72.95 +trainer/Log Pis Min -10.5331 +trainer/policy/mean Mean -0.0417759 +trainer/policy/mean Std 0.907718 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.78515 +trainer/policy/normal/std Std 0.700878 +trainer/policy/normal/std Max 5.52969 +trainer/policy/normal/std Min 0.229464 +trainer/policy/normal/log_std Mean 0.977127 +trainer/policy/normal/log_std Std 0.350604 +trainer/policy/normal/log_std Max 1.71013 +trainer/policy/normal/log_std Min -1.47201 +eval/num steps total 955392 +eval/num paths total 957 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0985981 +eval/Actions Std 0.912595 +eval/Actions Max 0.999998 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51812 +time/logging (s) 0.00382555 +time/sampling batch (s) 0.527173 +time/saving (s) 0.00354726 +time/training (s) 6.24169 +time/epoch (s) 9.29435 +time/total (s) 9828.45 +Epoch -44 +---------------------------------- --------------- +2022-05-10 15:54:51.725439 PDT | [2] Epoch -43 finished +---------------------------------- --------------- +epoch -43 +replay_buffer/size 999033 +trainer/num train calls 958000 +trainer/Policy Loss -20.7297 +trainer/Log Pis Mean 24.1746 +trainer/Log Pis Std 13.0643 +trainer/Log Pis Max 84.4646 +trainer/Log Pis Min -4.25576 +trainer/policy/mean Mean -0.0411876 +trainer/policy/mean Std 0.905091 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.80578 +trainer/policy/normal/std Std 0.705005 +trainer/policy/normal/std Max 6.539 +trainer/policy/normal/std Min 0.304252 +trainer/policy/normal/log_std Mean 0.984493 +trainer/policy/normal/log_std Std 0.350924 +trainer/policy/normal/log_std Max 1.87778 +trainer/policy/normal/log_std Min -1.1899 +eval/num steps total 956392 +eval/num paths total 958 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.213388 +eval/Actions Std 0.95137 +eval/Actions Max 0.999998 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 3.02751 +time/logging (s) 0.00374069 +time/sampling batch (s) 0.788545 +time/saving (s) 0.00334485 +time/training (s) 7.68459 +time/epoch (s) 11.5077 +time/total (s) 9839.96 +Epoch -43 +---------------------------------- --------------- +2022-05-10 15:55:01.606814 PDT | [2] Epoch -42 finished +---------------------------------- --------------- +epoch -42 +replay_buffer/size 999033 +trainer/num train calls 959000 +trainer/Policy Loss -19.0086 +trainer/Log Pis Mean 23.7652 +trainer/Log Pis Std 12.7465 +trainer/Log Pis Max 66.2923 +trainer/Log Pis Min -5.83609 +trainer/policy/mean Mean -0.00665219 +trainer/policy/mean Std 0.904143 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.79291 +trainer/policy/normal/std Std 0.686774 +trainer/policy/normal/std Max 6.58294 +trainer/policy/normal/std Min 0.292757 +trainer/policy/normal/log_std Mean 0.984318 +trainer/policy/normal/log_std Std 0.328667 +trainer/policy/normal/log_std Max 1.88448 +trainer/policy/normal/log_std Min -1.22841 +eval/num steps total 957392 +eval/num paths total 959 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0349123 +eval/Actions Std 0.898651 +eval/Actions Max 0.999998 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59827 +time/logging (s) 0.00371003 +time/sampling batch (s) 0.274038 +time/saving (s) 0.00332889 +time/training (s) 6.97765 +time/epoch (s) 9.857 +time/total (s) 9849.82 +Epoch -42 +---------------------------------- --------------- +2022-05-10 15:55:11.489575 PDT | [2] Epoch -41 finished +---------------------------------- --------------- +epoch -41 +replay_buffer/size 999033 +trainer/num train calls 960000 +trainer/Policy Loss -19.1781 +trainer/Log Pis Mean 23.6604 +trainer/Log Pis Std 12.9418 +trainer/Log Pis Max 59.738 +trainer/Log Pis Min -5.77476 +trainer/policy/mean Mean -0.0330824 +trainer/policy/mean Std 0.908159 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.82273 +trainer/policy/normal/std Std 0.671444 +trainer/policy/normal/std Max 5.82118 +trainer/policy/normal/std Min 0.298117 +trainer/policy/normal/log_std Mean 0.996411 +trainer/policy/normal/log_std Std 0.326552 +trainer/policy/normal/log_std Max 1.7615 +trainer/policy/normal/log_std Min -1.21027 +eval/num steps total 958392 +eval/num paths total 960 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.234647 +eval/Actions Std 0.812797 +eval/Actions Max 0.999996 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67918 +time/logging (s) 0.0037132 +time/sampling batch (s) 0.270667 +time/saving (s) 0.00333497 +time/training (s) 6.90173 +time/epoch (s) 9.85862 +time/total (s) 9859.68 +Epoch -41 +---------------------------------- --------------- +2022-05-10 15:55:21.509021 PDT | [2] Epoch -40 finished +---------------------------------- --------------- +epoch -40 +replay_buffer/size 999033 +trainer/num train calls 961000 +trainer/Policy Loss -19.5197 +trainer/Log Pis Mean 24.2681 +trainer/Log Pis Std 13.4476 +trainer/Log Pis Max 72.6005 +trainer/Log Pis Min -10.8732 +trainer/policy/mean Mean -0.0275565 +trainer/policy/mean Std 0.910139 +trainer/policy/mean Max 0.999975 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.83561 +trainer/policy/normal/std Std 0.675777 +trainer/policy/normal/std Max 5.63841 +trainer/policy/normal/std Min 0.324413 +trainer/policy/normal/log_std Mean 1.00133 +trainer/policy/normal/log_std Std 0.322778 +trainer/policy/normal/log_std Max 1.7296 +trainer/policy/normal/log_std Min -1.12574 +eval/num steps total 959392 +eval/num paths total 961 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.173276 +eval/Actions Std 0.877882 +eval/Actions Max 0.999997 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68166 +time/logging (s) 0.00380987 +time/sampling batch (s) 0.277027 +time/saving (s) 0.00333915 +time/training (s) 7.0293 +time/epoch (s) 9.99514 +time/total (s) 9869.68 +Epoch -40 +---------------------------------- --------------- +2022-05-10 15:55:31.235331 PDT | [2] Epoch -39 finished +---------------------------------- --------------- +epoch -39 +replay_buffer/size 999033 +trainer/num train calls 962000 +trainer/Policy Loss -19.1884 +trainer/Log Pis Mean 24.9977 +trainer/Log Pis Std 13.1814 +trainer/Log Pis Max 64.4035 +trainer/Log Pis Min -11.4929 +trainer/policy/mean Mean -0.0462391 +trainer/policy/mean Std 0.904834 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.8264 +trainer/policy/normal/std Std 0.675269 +trainer/policy/normal/std Max 6.12582 +trainer/policy/normal/std Min 0.27734 +trainer/policy/normal/log_std Mean 0.997166 +trainer/policy/normal/log_std Std 0.328308 +trainer/policy/normal/log_std Max 1.81251 +trainer/policy/normal/log_std Min -1.28251 +eval/num steps total 960392 +eval/num paths total 962 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.229173 +eval/Actions Std 0.766771 +eval/Actions Max 0.999991 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65926 +time/logging (s) 0.00400952 +time/sampling batch (s) 0.272226 +time/saving (s) 0.00359164 +time/training (s) 6.76327 +time/epoch (s) 9.70236 +time/total (s) 9879.39 +Epoch -39 +---------------------------------- --------------- +2022-05-10 15:55:41.586583 PDT | [2] Epoch -38 finished +---------------------------------- --------------- +epoch -38 +replay_buffer/size 999033 +trainer/num train calls 963000 +trainer/Policy Loss -20.0589 +trainer/Log Pis Mean 24.4354 +trainer/Log Pis Std 13.0452 +trainer/Log Pis Max 78.3448 +trainer/Log Pis Min -7.07066 +trainer/policy/mean Mean -0.0241374 +trainer/policy/mean Std 0.906477 +trainer/policy/mean Max 0.999979 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.75122 +trainer/policy/normal/std Std 0.681874 +trainer/policy/normal/std Max 5.49492 +trainer/policy/normal/std Min 0.26757 +trainer/policy/normal/log_std Mean 0.965627 +trainer/policy/normal/log_std Std 0.349308 +trainer/policy/normal/log_std Max 1.70382 +trainer/policy/normal/log_std Min -1.31837 +eval/num steps total 961392 +eval/num paths total 963 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.363541 +eval/Actions Std 0.892083 +eval/Actions Max 0.999994 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.32268 +time/logging (s) 0.00405421 +time/sampling batch (s) 0.773038 +time/saving (s) 0.0036263 +time/training (s) 7.22343 +time/epoch (s) 10.3268 +time/total (s) 9889.72 +Epoch -38 +---------------------------------- --------------- +2022-05-10 15:55:52.762757 PDT | [2] Epoch -37 finished +---------------------------------- --------------- +epoch -37 +replay_buffer/size 999033 +trainer/num train calls 964000 +trainer/Policy Loss -19.8064 +trainer/Log Pis Mean 25.0419 +trainer/Log Pis Std 12.9905 +trainer/Log Pis Max 58.8302 +trainer/Log Pis Min -9.19535 +trainer/policy/mean Mean -0.0429656 +trainer/policy/mean Std 0.906394 +trainer/policy/mean Max 0.999977 +trainer/policy/mean Min -0.999982 +trainer/policy/normal/std Mean 2.78814 +trainer/policy/normal/std Std 0.700809 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.288554 +trainer/policy/normal/log_std Mean 0.978662 +trainer/policy/normal/log_std Std 0.347863 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.24287 +eval/num steps total 962392 +eval/num paths total 964 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0656336 +eval/Actions Std 0.91984 +eval/Actions Max 0.999997 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62988 +time/logging (s) 0.00369139 +time/sampling batch (s) 0.775263 +time/saving (s) 0.00347438 +time/training (s) 7.7386 +time/epoch (s) 11.1509 +time/total (s) 9900.87 +Epoch -37 +---------------------------------- --------------- +2022-05-10 15:56:04.481781 PDT | [2] Epoch -36 finished +---------------------------------- --------------- +epoch -36 +replay_buffer/size 999033 +trainer/num train calls 965000 +trainer/Policy Loss -20.5079 +trainer/Log Pis Mean 25.0252 +trainer/Log Pis Std 14.0078 +trainer/Log Pis Max 69.4704 +trainer/Log Pis Min -11.9484 +trainer/policy/mean Mean -0.0810659 +trainer/policy/mean Std 0.908908 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.81249 +trainer/policy/normal/std Std 0.67049 +trainer/policy/normal/std Max 6.7965 +trainer/policy/normal/std Min 0.275533 +trainer/policy/normal/log_std Mean 0.991896 +trainer/policy/normal/log_std Std 0.330905 +trainer/policy/normal/log_std Max 1.91641 +trainer/policy/normal/log_std Min -1.28905 +eval/num steps total 963392 +eval/num paths total 965 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.295928 +eval/Actions Std 0.857803 +eval/Actions Max 0.999991 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6525 +time/logging (s) 0.00370932 +time/sampling batch (s) 0.773176 +time/saving (s) 0.0033517 +time/training (s) 8.26206 +time/epoch (s) 11.6948 +time/total (s) 9912.57 +Epoch -36 +---------------------------------- --------------- +2022-05-10 15:56:15.415734 PDT | [2] Epoch -35 finished +---------------------------------- --------------- +epoch -35 +replay_buffer/size 999033 +trainer/num train calls 966000 +trainer/Policy Loss -19.2843 +trainer/Log Pis Mean 25.0321 +trainer/Log Pis Std 13.0353 +trainer/Log Pis Max 76.0535 +trainer/Log Pis Min -9.18467 +trainer/policy/mean Mean -0.0342766 +trainer/policy/mean Std 0.905221 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.84613 +trainer/policy/normal/std Std 0.713109 +trainer/policy/normal/std Max 6.12996 +trainer/policy/normal/std Min 0.237945 +trainer/policy/normal/log_std Mean 0.99815 +trainer/policy/normal/log_std Std 0.355753 +trainer/policy/normal/log_std Max 1.81319 +trainer/policy/normal/log_std Min -1.43571 +eval/num steps total 964392 +eval/num paths total 966 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.29147 +eval/Actions Std 0.72101 +eval/Actions Max 0.999998 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63977 +time/logging (s) 0.00373182 +time/sampling batch (s) 0.531822 +time/saving (s) 0.00333705 +time/training (s) 7.73069 +time/epoch (s) 10.9093 +time/total (s) 9923.48 +Epoch -35 +---------------------------------- --------------- +2022-05-10 15:56:25.462614 PDT | [2] Epoch -34 finished +---------------------------------- --------------- +epoch -34 +replay_buffer/size 999033 +trainer/num train calls 967000 +trainer/Policy Loss -19.3534 +trainer/Log Pis Mean 24.1884 +trainer/Log Pis Std 13.2127 +trainer/Log Pis Max 81.073 +trainer/Log Pis Min -4.40961 +trainer/policy/mean Mean -0.0505362 +trainer/policy/mean Std 0.90619 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.81391 +trainer/policy/normal/std Std 0.708895 +trainer/policy/normal/std Max 5.96379 +trainer/policy/normal/std Min 0.219381 +trainer/policy/normal/log_std Mean 0.987215 +trainer/policy/normal/log_std Std 0.35258 +trainer/policy/normal/log_std Max 1.78571 +trainer/policy/normal/log_std Min -1.51695 +eval/num steps total 965392 +eval/num paths total 967 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0663572 +eval/Actions Std 0.909012 +eval/Actions Max 0.999997 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.423 +time/logging (s) 0.00379948 +time/sampling batch (s) 0.773974 +time/saving (s) 0.00333118 +time/training (s) 6.81879 +time/epoch (s) 10.0229 +time/total (s) 9933.51 +Epoch -34 +---------------------------------- --------------- +2022-05-10 15:56:35.814789 PDT | [2] Epoch -33 finished +---------------------------------- --------------- +epoch -33 +replay_buffer/size 999033 +trainer/num train calls 968000 +trainer/Policy Loss -20.0236 +trainer/Log Pis Mean 24.1772 +trainer/Log Pis Std 13.7883 +trainer/Log Pis Max 66.3485 +trainer/Log Pis Min -7.34094 +trainer/policy/mean Mean -0.031385 +trainer/policy/mean Std 0.906063 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999985 +trainer/policy/normal/std Mean 2.77736 +trainer/policy/normal/std Std 0.66904 +trainer/policy/normal/std Max 6.15602 +trainer/policy/normal/std Min 0.276266 +trainer/policy/normal/log_std Mean 0.977895 +trainer/policy/normal/log_std Std 0.337835 +trainer/policy/normal/log_std Max 1.81743 +trainer/policy/normal/log_std Min -1.28639 +eval/num steps total 966392 +eval/num paths total 968 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.14098 +eval/Actions Std 0.818265 +eval/Actions Max 0.999993 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.49958 +time/logging (s) 0.00411074 +time/sampling batch (s) 0.523416 +time/saving (s) 0.00375008 +time/training (s) 7.29745 +time/epoch (s) 10.3283 +time/total (s) 9943.84 +Epoch -33 +---------------------------------- --------------- +2022-05-10 15:56:46.239607 PDT | [2] Epoch -32 finished +---------------------------------- --------------- +epoch -32 +replay_buffer/size 999033 +trainer/num train calls 969000 +trainer/Policy Loss -19.8393 +trainer/Log Pis Mean 25.3629 +trainer/Log Pis Std 13.3895 +trainer/Log Pis Max 72.0377 +trainer/Log Pis Min -5.89142 +trainer/policy/mean Mean -0.0447927 +trainer/policy/mean Std 0.907467 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.74196 +trainer/policy/normal/std Std 0.644152 +trainer/policy/normal/std Max 5.97586 +trainer/policy/normal/std Min 0.265516 +trainer/policy/normal/log_std Mean 0.969007 +trainer/policy/normal/log_std Std 0.317397 +trainer/policy/normal/log_std Max 1.78773 +trainer/policy/normal/log_std Min -1.32608 +eval/num steps total 967392 +eval/num paths total 969 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0863657 +eval/Actions Std 0.920307 +eval/Actions Max 0.999995 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.72623 +time/logging (s) 0.00368316 +time/sampling batch (s) 0.279018 +time/saving (s) 0.00342528 +time/training (s) 7.38696 +time/epoch (s) 10.3993 +time/total (s) 9954.24 +Epoch -32 +---------------------------------- --------------- +2022-05-10 15:56:56.864289 PDT | [2] Epoch -31 finished +---------------------------------- --------------- +epoch -31 +replay_buffer/size 999033 +trainer/num train calls 970000 +trainer/Policy Loss -19.2685 +trainer/Log Pis Mean 25.1883 +trainer/Log Pis Std 13.8111 +trainer/Log Pis Max 67.9833 +trainer/Log Pis Min -7.68315 +trainer/policy/mean Mean -0.0330012 +trainer/policy/mean Std 0.904427 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.80193 +trainer/policy/normal/std Std 0.678414 +trainer/policy/normal/std Max 5.50314 +trainer/policy/normal/std Min 0.280193 +trainer/policy/normal/log_std Mean 0.986835 +trainer/policy/normal/log_std Std 0.336553 +trainer/policy/normal/log_std Max 1.70532 +trainer/policy/normal/log_std Min -1.27228 +eval/num steps total 968392 +eval/num paths total 970 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.237627 +eval/Actions Std 0.930225 +eval/Actions Max 0.99998 +eval/Actions Min -0.999986 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51569 +time/logging (s) 0.00375245 +time/sampling batch (s) 0.779338 +time/saving (s) 0.00336534 +time/training (s) 7.29826 +time/epoch (s) 10.6004 +time/total (s) 9964.84 +Epoch -31 +---------------------------------- --------------- +2022-05-10 15:57:07.492415 PDT | [2] Epoch -30 finished +---------------------------------- --------------- +epoch -30 +replay_buffer/size 999033 +trainer/num train calls 971000 +trainer/Policy Loss -19.9016 +trainer/Log Pis Mean 24.974 +trainer/Log Pis Std 13.0093 +trainer/Log Pis Max 69.7921 +trainer/Log Pis Min -6.15417 +trainer/policy/mean Mean -0.0639365 +trainer/policy/mean Std 0.90843 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.80877 +trainer/policy/normal/std Std 0.664174 +trainer/policy/normal/std Max 6.14119 +trainer/policy/normal/std Min 0.34411 +trainer/policy/normal/log_std Mean 0.992583 +trainer/policy/normal/log_std Std 0.320097 +trainer/policy/normal/log_std Max 1.81502 +trainer/policy/normal/log_std Min -1.06679 +eval/num steps total 969392 +eval/num paths total 971 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.265544 +eval/Actions Std 0.917653 +eval/Actions Max 0.999998 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.43994 +time/logging (s) 0.00367318 +time/sampling batch (s) 1.02429 +time/saving (s) 0.00333123 +time/training (s) 7.13266 +time/epoch (s) 10.6039 +time/total (s) 9975.45 +Epoch -30 +---------------------------------- --------------- +2022-05-10 15:57:18.563168 PDT | [2] Epoch -29 finished +---------------------------------- --------------- +epoch -29 +replay_buffer/size 999033 +trainer/num train calls 972000 +trainer/Policy Loss -20.3278 +trainer/Log Pis Mean 24.9734 +trainer/Log Pis Std 13.4964 +trainer/Log Pis Max 64.7091 +trainer/Log Pis Min -8.44012 +trainer/policy/mean Mean -0.0241348 +trainer/policy/mean Std 0.909338 +trainer/policy/mean Max 1 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.79971 +trainer/policy/normal/std Std 0.67917 +trainer/policy/normal/std Max 5.41049 +trainer/policy/normal/std Min 0.250117 +trainer/policy/normal/log_std Mean 0.984984 +trainer/policy/normal/log_std Std 0.342135 +trainer/policy/normal/log_std Max 1.68834 +trainer/policy/normal/log_std Min -1.38583 +eval/num steps total 970392 +eval/num paths total 972 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0559386 +eval/Actions Std 0.899586 +eval/Actions Max 1 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.46071 +time/logging (s) 0.00376997 +time/sampling batch (s) 0.77544 +time/saving (s) 0.00336763 +time/training (s) 7.80317 +time/epoch (s) 11.0465 +time/total (s) 9986.5 +Epoch -29 +---------------------------------- --------------- +2022-05-10 15:57:29.992206 PDT | [2] Epoch -28 finished +---------------------------------- --------------- +epoch -28 +replay_buffer/size 999033 +trainer/num train calls 973000 +trainer/Policy Loss -21.0958 +trainer/Log Pis Mean 25.0202 +trainer/Log Pis Std 14.3154 +trainer/Log Pis Max 67.254 +trainer/Log Pis Min -8.65535 +trainer/policy/mean Mean -0.00252409 +trainer/policy/mean Std 0.907053 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.82435 +trainer/policy/normal/std Std 0.668003 +trainer/policy/normal/std Max 5.90478 +trainer/policy/normal/std Min 0.235278 +trainer/policy/normal/log_std Mean 0.998242 +trainer/policy/normal/log_std Std 0.31936 +trainer/policy/normal/log_std Max 1.77576 +trainer/policy/normal/log_std Min -1.44699 +eval/num steps total 971392 +eval/num paths total 973 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0127805 +eval/Actions Std 0.884444 +eval/Actions Max 0.999996 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64746 +time/logging (s) 0.00406573 +time/sampling batch (s) 1.02575 +time/saving (s) 0.00367449 +time/training (s) 7.72397 +time/epoch (s) 11.4049 +time/total (s) 9997.91 +Epoch -28 +---------------------------------- --------------- +2022-05-10 15:57:39.943977 PDT | [2] Epoch -27 finished +---------------------------------- --------------- +epoch -27 +replay_buffer/size 999033 +trainer/num train calls 974000 +trainer/Policy Loss -20.0814 +trainer/Log Pis Mean 24.642 +trainer/Log Pis Std 14.1927 +trainer/Log Pis Max 66.9692 +trainer/Log Pis Min -11.415 +trainer/policy/mean Mean -0.0268277 +trainer/policy/mean Std 0.90541 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.82431 +trainer/policy/normal/std Std 0.689928 +trainer/policy/normal/std Max 5.61668 +trainer/policy/normal/std Min 0.250421 +trainer/policy/normal/log_std Mean 0.994766 +trainer/policy/normal/log_std Std 0.334979 +trainer/policy/normal/log_std Max 1.72574 +trainer/policy/normal/log_std Min -1.38461 +eval/num steps total 972392 +eval/num paths total 974 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.133032 +eval/Actions Std 0.867507 +eval/Actions Max 0.999993 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.77413 +time/logging (s) 0.0037813 +time/sampling batch (s) 0.276712 +time/saving (s) 0.00349064 +time/training (s) 6.8685 +time/epoch (s) 9.92662 +time/total (s) 10007.8 +Epoch -27 +---------------------------------- --------------- +2022-05-10 15:57:49.787377 PDT | [2] Epoch -26 finished +---------------------------------- --------------- +epoch -26 +replay_buffer/size 999033 +trainer/num train calls 975000 +trainer/Policy Loss -20.6181 +trainer/Log Pis Mean 24.4633 +trainer/Log Pis Std 13.6654 +trainer/Log Pis Max 73.5589 +trainer/Log Pis Min -10.7216 +trainer/policy/mean Mean -0.0588438 +trainer/policy/mean Std 0.909946 +trainer/policy/mean Max 0.999979 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.81673 +trainer/policy/normal/std Std 0.680961 +trainer/policy/normal/std Max 7.02509 +trainer/policy/normal/std Min 0.279607 +trainer/policy/normal/log_std Mean 0.992669 +trainer/policy/normal/log_std Std 0.332233 +trainer/policy/normal/log_std Max 1.94949 +trainer/policy/normal/log_std Min -1.27437 +eval/num steps total 973392 +eval/num paths total 975 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.168902 +eval/Actions Std 0.839939 +eval/Actions Max 0.999992 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54593 +time/logging (s) 0.00384816 +time/sampling batch (s) 0.530674 +time/saving (s) 0.00333192 +time/training (s) 6.73491 +time/epoch (s) 9.8187 +time/total (s) 10017.7 +Epoch -26 +---------------------------------- --------------- +2022-05-10 15:58:00.311380 PDT | [2] Epoch -25 finished +---------------------------------- --------------- +epoch -25 +replay_buffer/size 999033 +trainer/num train calls 976000 +trainer/Policy Loss -19.3052 +trainer/Log Pis Mean 23.6879 +trainer/Log Pis Std 13.8345 +trainer/Log Pis Max 69.9508 +trainer/Log Pis Min -10.4073 +trainer/policy/mean Mean -0.0166443 +trainer/policy/mean Std 0.900547 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.83652 +trainer/policy/normal/std Std 0.683905 +trainer/policy/normal/std Max 6.79156 +trainer/policy/normal/std Min 0.261869 +trainer/policy/normal/log_std Mean 1.0001 +trainer/policy/normal/log_std Std 0.33066 +trainer/policy/normal/log_std Max 1.91568 +trainer/policy/normal/log_std Min -1.33991 +eval/num steps total 974392 +eval/num paths total 976 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.291372 +eval/Actions Std 0.722181 +eval/Actions Max 0.999978 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.7309 +time/logging (s) 0.0038315 +time/sampling batch (s) 0.274953 +time/saving (s) 0.00327614 +time/training (s) 7.48661 +time/epoch (s) 10.4996 +time/total (s) 10028.2 +Epoch -25 +---------------------------------- --------------- +2022-05-10 15:58:11.273979 PDT | [2] Epoch -24 finished +---------------------------------- --------------- +epoch -24 +replay_buffer/size 999033 +trainer/num train calls 977000 +trainer/Policy Loss -20.271 +trainer/Log Pis Mean 24.6679 +trainer/Log Pis Std 12.7277 +trainer/Log Pis Max 66.901 +trainer/Log Pis Min -9.58946 +trainer/policy/mean Mean -0.0456119 +trainer/policy/mean Std 0.909921 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.72243 +trainer/policy/normal/std Std 0.64069 +trainer/policy/normal/std Max 5.35121 +trainer/policy/normal/std Min 0.197209 +trainer/policy/normal/log_std Mean 0.961483 +trainer/policy/normal/log_std Std 0.319928 +trainer/policy/normal/log_std Max 1.67732 +trainer/policy/normal/log_std Min -1.62349 +eval/num steps total 975392 +eval/num paths total 977 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00628137 +eval/Actions Std 0.884387 +eval/Actions Max 0.999995 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55068 +time/logging (s) 0.00387311 +time/sampling batch (s) 0.773304 +time/saving (s) 0.00332474 +time/training (s) 7.60703 +time/epoch (s) 10.9382 +time/total (s) 10039.1 +Epoch -24 +---------------------------------- --------------- +2022-05-10 15:58:21.509778 PDT | [2] Epoch -23 finished +---------------------------------- --------------- +epoch -23 +replay_buffer/size 999033 +trainer/num train calls 978000 +trainer/Policy Loss -19.8691 +trainer/Log Pis Mean 25.2633 +trainer/Log Pis Std 13.1658 +trainer/Log Pis Max 65.945 +trainer/Log Pis Min -6.01898 +trainer/policy/mean Mean -0.0375647 +trainer/policy/mean Std 0.907185 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.78853 +trainer/policy/normal/std Std 0.686974 +trainer/policy/normal/std Max 6.60446 +trainer/policy/normal/std Min 0.291115 +trainer/policy/normal/log_std Mean 0.982667 +trainer/policy/normal/log_std Std 0.328571 +trainer/policy/normal/log_std Max 1.88774 +trainer/policy/normal/log_std Min -1.23404 +eval/num steps total 976392 +eval/num paths total 978 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.2068 +eval/Actions Std 0.912248 +eval/Actions Max 0.999993 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.24023 +time/logging (s) 0.00396095 +time/sampling batch (s) 0.521967 +time/saving (s) 0.003347 +time/training (s) 7.44237 +time/epoch (s) 10.2119 +time/total (s) 10049.3 +Epoch -23 +---------------------------------- --------------- +2022-05-10 15:58:32.759627 PDT | [2] Epoch -22 finished +---------------------------------- --------------- +epoch -22 +replay_buffer/size 999033 +trainer/num train calls 979000 +trainer/Policy Loss -19.6933 +trainer/Log Pis Mean 23.9802 +trainer/Log Pis Std 13.1033 +trainer/Log Pis Max 68.7304 +trainer/Log Pis Min -14.1461 +trainer/policy/mean Mean -0.0306252 +trainer/policy/mean Std 0.90775 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.78897 +trainer/policy/normal/std Std 0.676958 +trainer/policy/normal/std Max 5.46542 +trainer/policy/normal/std Min 0.308503 +trainer/policy/normal/log_std Mean 0.98313 +trainer/policy/normal/log_std Std 0.328587 +trainer/policy/normal/log_std Max 1.69844 +trainer/policy/normal/log_std Min -1.17602 +eval/num steps total 977392 +eval/num paths total 979 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0249491 +eval/Actions Std 0.929915 +eval/Actions Max 0.999996 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55387 +time/logging (s) 0.0038997 +time/sampling batch (s) 0.771127 +time/saving (s) 0.00362959 +time/training (s) 7.89287 +time/epoch (s) 11.2254 +time/total (s) 10060.5 +Epoch -22 +---------------------------------- --------------- +2022-05-10 15:58:43.316753 PDT | [2] Epoch -21 finished +---------------------------------- --------------- +epoch -21 +replay_buffer/size 999033 +trainer/num train calls 980000 +trainer/Policy Loss -20.216 +trainer/Log Pis Mean 24.0735 +trainer/Log Pis Std 12.8366 +trainer/Log Pis Max 63.3835 +trainer/Log Pis Min -7.82495 +trainer/policy/mean Mean -0.0251351 +trainer/policy/mean Std 0.908265 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.85722 +trainer/policy/normal/std Std 0.667728 +trainer/policy/normal/std Max 5.65195 +trainer/policy/normal/std Min 0.252089 +trainer/policy/normal/log_std Mean 1.0099 +trainer/policy/normal/log_std Std 0.320634 +trainer/policy/normal/log_std Max 1.732 +trainer/policy/normal/log_std Min -1.37797 +eval/num steps total 978392 +eval/num paths total 980 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.197821 +eval/Actions Std 0.837455 +eval/Actions Max 1 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70588 +time/logging (s) 0.0037399 +time/sampling batch (s) 0.281098 +time/saving (s) 0.00406031 +time/training (s) 7.53736 +time/epoch (s) 10.5321 +time/total (s) 10071.1 +Epoch -21 +---------------------------------- --------------- +2022-05-10 15:58:54.777304 PDT | [2] Epoch -20 finished +---------------------------------- --------------- +epoch -20 +replay_buffer/size 999033 +trainer/num train calls 981000 +trainer/Policy Loss -20.7552 +trainer/Log Pis Mean 25.134 +trainer/Log Pis Std 13.7537 +trainer/Log Pis Max 83.9763 +trainer/Log Pis Min -5.23068 +trainer/policy/mean Mean -0.0133866 +trainer/policy/mean Std 0.911109 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.83553 +trainer/policy/normal/std Std 0.676594 +trainer/policy/normal/std Max 6.83442 +trainer/policy/normal/std Min 0.27039 +trainer/policy/normal/log_std Mean 1.00089 +trainer/policy/normal/log_std Std 0.326814 +trainer/policy/normal/log_std Max 1.92197 +trainer/policy/normal/log_std Min -1.30789 +eval/num steps total 979392 +eval/num paths total 981 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.3146 +eval/Actions Std 0.852725 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.74894 +time/logging (s) 0.00378598 +time/sampling batch (s) 1.284 +time/saving (s) 0.0034311 +time/training (s) 7.39445 +time/epoch (s) 11.4346 +time/total (s) 10082.5 +Epoch -20 +---------------------------------- --------------- +2022-05-10 15:59:04.848219 PDT | [2] Epoch -19 finished +---------------------------------- --------------- +epoch -19 +replay_buffer/size 999033 +trainer/num train calls 982000 +trainer/Policy Loss -19.6554 +trainer/Log Pis Mean 23.9219 +trainer/Log Pis Std 12.555 +trainer/Log Pis Max 60.6572 +trainer/Log Pis Min -13.0023 +trainer/policy/mean Mean -0.0233376 +trainer/policy/mean Std 0.907253 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.78803 +trainer/policy/normal/std Std 0.665585 +trainer/policy/normal/std Max 5.6841 +trainer/policy/normal/std Min 0.249852 +trainer/policy/normal/log_std Mean 0.983943 +trainer/policy/normal/log_std Std 0.325115 +trainer/policy/normal/log_std Max 1.73767 +trainer/policy/normal/log_std Min -1.38689 +eval/num steps total 980392 +eval/num paths total 982 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.103047 +eval/Actions Std 0.904194 +eval/Actions Max 0.999996 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56353 +time/logging (s) 0.00381802 +time/sampling batch (s) 0.271861 +time/saving (s) 0.00335203 +time/training (s) 7.20424 +time/epoch (s) 10.0468 +time/total (s) 10092.6 +Epoch -19 +---------------------------------- --------------- +2022-05-10 15:59:15.178613 PDT | [2] Epoch -18 finished +---------------------------------- --------------- +epoch -18 +replay_buffer/size 999033 +trainer/num train calls 983000 +trainer/Policy Loss -18.7551 +trainer/Log Pis Mean 23.8304 +trainer/Log Pis Std 13.014 +trainer/Log Pis Max 70.6045 +trainer/Log Pis Min -11.8887 +trainer/policy/mean Mean -0.0382189 +trainer/policy/mean Std 0.906327 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.73587 +trainer/policy/normal/std Std 0.654097 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.256573 +trainer/policy/normal/log_std Mean 0.964386 +trainer/policy/normal/log_std Std 0.331115 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.36034 +eval/num steps total 981392 +eval/num paths total 983 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.080789 +eval/Actions Std 0.909198 +eval/Actions Max 0.999997 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56217 +time/logging (s) 0.00375924 +time/sampling batch (s) 0.271229 +time/saving (s) 0.00333649 +time/training (s) 7.46566 +time/epoch (s) 10.3061 +time/total (s) 10102.9 +Epoch -18 +---------------------------------- --------------- +2022-05-10 15:59:25.940301 PDT | [2] Epoch -17 finished +---------------------------------- --------------- +epoch -17 +replay_buffer/size 999033 +trainer/num train calls 984000 +trainer/Policy Loss -20.1198 +trainer/Log Pis Mean 24.2191 +trainer/Log Pis Std 13.0262 +trainer/Log Pis Max 63.6639 +trainer/Log Pis Min -8.9855 +trainer/policy/mean Mean -0.0525539 +trainer/policy/mean Std 0.908055 +trainer/policy/mean Max 0.999978 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.76702 +trainer/policy/normal/std Std 0.657366 +trainer/policy/normal/std Max 5.41426 +trainer/policy/normal/std Min 0.238722 +trainer/policy/normal/log_std Mean 0.976604 +trainer/policy/normal/log_std Std 0.325472 +trainer/policy/normal/log_std Max 1.68904 +trainer/policy/normal/log_std Min -1.43245 +eval/num steps total 982392 +eval/num paths total 984 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.367983 +eval/Actions Std 0.868279 +eval/Actions Max 1 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5567 +time/logging (s) 0.00397491 +time/sampling batch (s) 0.772567 +time/saving (s) 0.00351556 +time/training (s) 7.40091 +time/epoch (s) 10.7377 +time/total (s) 10113.6 +Epoch -17 +---------------------------------- --------------- +2022-05-10 15:59:36.144793 PDT | [2] Epoch -16 finished +---------------------------------- --------------- +epoch -16 +replay_buffer/size 999033 +trainer/num train calls 985000 +trainer/Policy Loss -18.9693 +trainer/Log Pis Mean 24.3566 +trainer/Log Pis Std 13.4194 +trainer/Log Pis Max 70.0609 +trainer/Log Pis Min -5.5325 +trainer/policy/mean Mean -0.023578 +trainer/policy/mean Std 0.907512 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.8047 +trainer/policy/normal/std Std 0.677466 +trainer/policy/normal/std Max 6.06346 +trainer/policy/normal/std Min 0.295788 +trainer/policy/normal/log_std Mean 0.988162 +trainer/policy/normal/log_std Std 0.334311 +trainer/policy/normal/log_std Max 1.80228 +trainer/policy/normal/log_std Min -1.21811 +eval/num steps total 983392 +eval/num paths total 985 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.279481 +eval/Actions Std 0.910701 +eval/Actions Max 1 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.42586 +time/logging (s) 0.00409442 +time/sampling batch (s) 0.525412 +time/saving (s) 0.00357142 +time/training (s) 7.22112 +time/epoch (s) 10.1801 +time/total (s) 10123.8 +Epoch -16 +---------------------------------- --------------- +2022-05-10 15:59:47.618062 PDT | [2] Epoch -15 finished +---------------------------------- --------------- +epoch -15 +replay_buffer/size 999033 +trainer/num train calls 986000 +trainer/Policy Loss -19.9273 +trainer/Log Pis Mean 24.3687 +trainer/Log Pis Std 13.194 +trainer/Log Pis Max 70.978 +trainer/Log Pis Min -6.71081 +trainer/policy/mean Mean -0.0371254 +trainer/policy/mean Std 0.91029 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.77536 +trainer/policy/normal/std Std 0.66671 +trainer/policy/normal/std Max 6.02284 +trainer/policy/normal/std Min 0.271033 +trainer/policy/normal/log_std Mean 0.979399 +trainer/policy/normal/log_std Std 0.324171 +trainer/policy/normal/log_std Max 1.79556 +trainer/policy/normal/log_std Min -1.30551 +eval/num steps total 984392 +eval/num paths total 986 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.287099 +eval/Actions Std 0.896451 +eval/Actions Max 0.999997 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54787 +time/logging (s) 0.00373595 +time/sampling batch (s) 0.273808 +time/saving (s) 0.00342021 +time/training (s) 8.61938 +time/epoch (s) 11.4482 +time/total (s) 10135.3 +Epoch -15 +---------------------------------- --------------- +2022-05-10 15:59:57.682773 PDT | [2] Epoch -14 finished +---------------------------------- --------------- +epoch -14 +replay_buffer/size 999033 +trainer/num train calls 987000 +trainer/Policy Loss -19.5423 +trainer/Log Pis Mean 24.9308 +trainer/Log Pis Std 12.8597 +trainer/Log Pis Max 63.6814 +trainer/Log Pis Min -7.3483 +trainer/policy/mean Mean -0.0135441 +trainer/policy/mean Std 0.902526 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.78391 +trainer/policy/normal/std Std 0.71509 +trainer/policy/normal/std Max 5.90837 +trainer/policy/normal/std Min 0.283262 +trainer/policy/normal/log_std Mean 0.973826 +trainer/policy/normal/log_std Std 0.363426 +trainer/policy/normal/log_std Max 1.77637 +trainer/policy/normal/log_std Min -1.26138 +eval/num steps total 985392 +eval/num paths total 987 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.160642 +eval/Actions Std 0.917091 +eval/Actions Max 0.999999 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.7488 +time/logging (s) 0.00376205 +time/sampling batch (s) 0.772589 +time/saving (s) 0.00333889 +time/training (s) 6.51188 +time/epoch (s) 10.0404 +time/total (s) 10145.3 +Epoch -14 +---------------------------------- --------------- +2022-05-10 16:00:07.005424 PDT | [2] Epoch -13 finished +---------------------------------- --------------- +epoch -13 +replay_buffer/size 999033 +trainer/num train calls 988000 +trainer/Policy Loss -19.5026 +trainer/Log Pis Mean 24.8389 +trainer/Log Pis Std 12.9875 +trainer/Log Pis Max 58.3128 +trainer/Log Pis Min -8.90757 +trainer/policy/mean Mean -0.0423668 +trainer/policy/mean Std 0.902209 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.78507 +trainer/policy/normal/std Std 0.687645 +trainer/policy/normal/std Max 5.73858 +trainer/policy/normal/std Min 0.199457 +trainer/policy/normal/log_std Mean 0.978829 +trainer/policy/normal/log_std Std 0.34493 +trainer/policy/normal/log_std Max 1.74721 +trainer/policy/normal/log_std Min -1.61216 +eval/num steps total 986392 +eval/num paths total 988 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.17385 +eval/Actions Std 0.856739 +eval/Actions Max 0.999999 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.81029 +time/logging (s) 0.00370075 +time/sampling batch (s) 0.523949 +time/saving (s) 0.00337874 +time/training (s) 5.957 +time/epoch (s) 9.29831 +time/total (s) 10154.6 +Epoch -13 +---------------------------------- --------------- +2022-05-10 16:00:15.707913 PDT | [2] Epoch -12 finished +---------------------------------- --------------- +epoch -12 +replay_buffer/size 999033 +trainer/num train calls 989000 +trainer/Policy Loss -19.6204 +trainer/Log Pis Mean 23.6754 +trainer/Log Pis Std 13.9714 +trainer/Log Pis Max 74.2769 +trainer/Log Pis Min -8.64834 +trainer/policy/mean Mean -0.0139397 +trainer/policy/mean Std 0.903575 +trainer/policy/mean Max 0.999977 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.7938 +trainer/policy/normal/std Std 0.703825 +trainer/policy/normal/std Max 6.98122 +trainer/policy/normal/std Min 0.302456 +trainer/policy/normal/log_std Mean 0.980503 +trainer/policy/normal/log_std Std 0.348177 +trainer/policy/normal/log_std Max 1.94322 +trainer/policy/normal/log_std Min -1.19582 +eval/num steps total 987392 +eval/num paths total 989 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.270159 +eval/Actions Std 0.732927 +eval/Actions Max 0.999991 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 1.80791 +time/logging (s) 0.00385881 +time/sampling batch (s) 0.778654 +time/saving (s) 0.0040537 +time/training (s) 6.08365 +time/epoch (s) 8.67813 +time/total (s) 10163.3 +Epoch -12 +---------------------------------- --------------- +2022-05-10 16:00:23.490885 PDT | [2] Epoch -11 finished +---------------------------------- --------------- +epoch -11 +replay_buffer/size 999033 +trainer/num train calls 990000 +trainer/Policy Loss -19.372 +trainer/Log Pis Mean 25.2822 +trainer/Log Pis Std 13.8811 +trainer/Log Pis Max 68.6707 +trainer/Log Pis Min -9.25024 +trainer/policy/mean Mean -0.0408179 +trainer/policy/mean Std 0.912878 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.77798 +trainer/policy/normal/std Std 0.674015 +trainer/policy/normal/std Max 7.24301 +trainer/policy/normal/std Min 0.277006 +trainer/policy/normal/log_std Mean 0.979463 +trainer/policy/normal/log_std Std 0.328727 +trainer/policy/normal/log_std Max 1.98004 +trainer/policy/normal/log_std Min -1.28372 +eval/num steps total 988392 +eval/num paths total 990 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00226061 +eval/Actions Std 0.905373 +eval/Actions Max 0.999996 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 1.88827 +time/logging (s) 0.00374016 +time/sampling batch (s) 0.279446 +time/saving (s) 0.00334036 +time/training (s) 5.58215 +time/epoch (s) 7.75695 +time/total (s) 10171 +Epoch -11 +---------------------------------- --------------- +2022-05-10 16:00:31.929819 PDT | [2] Epoch -10 finished +---------------------------------- --------------- +epoch -10 +replay_buffer/size 999033 +trainer/num train calls 991000 +trainer/Policy Loss -19.7817 +trainer/Log Pis Mean 24.1393 +trainer/Log Pis Std 13.405 +trainer/Log Pis Max 72.1258 +trainer/Log Pis Min -7.62007 +trainer/policy/mean Mean -0.0215541 +trainer/policy/mean Std 0.907051 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.78245 +trainer/policy/normal/std Std 0.688224 +trainer/policy/normal/std Max 5.87243 +trainer/policy/normal/std Min 0.237128 +trainer/policy/normal/log_std Mean 0.976614 +trainer/policy/normal/log_std Std 0.352136 +trainer/policy/normal/log_std Max 1.77027 +trainer/policy/normal/log_std Min -1.43915 +eval/num steps total 989392 +eval/num paths total 991 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.100356 +eval/Actions Std 0.917002 +eval/Actions Max 0.999986 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 1.90573 +time/logging (s) 0.00375437 +time/sampling batch (s) 0.273987 +time/saving (s) 0.00336825 +time/training (s) 6.2278 +time/epoch (s) 8.41464 +time/total (s) 10179.5 +Epoch -10 +---------------------------------- --------------- +2022-05-10 16:00:42.513256 PDT | [2] Epoch -9 finished +---------------------------------- --------------- +epoch -9 +replay_buffer/size 999033 +trainer/num train calls 992000 +trainer/Policy Loss -20.0363 +trainer/Log Pis Mean 25.0959 +trainer/Log Pis Std 13.1424 +trainer/Log Pis Max 70.7466 +trainer/Log Pis Min -6.61039 +trainer/policy/mean Mean -0.0229517 +trainer/policy/mean Std 0.903249 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.78528 +trainer/policy/normal/std Std 0.684202 +trainer/policy/normal/std Max 7.18866 +trainer/policy/normal/std Min 0.29422 +trainer/policy/normal/log_std Mean 0.979299 +trainer/policy/normal/log_std Std 0.344726 +trainer/policy/normal/log_std Max 1.9725 +trainer/policy/normal/log_std Min -1.22343 +eval/num steps total 990392 +eval/num paths total 992 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0151016 +eval/Actions Std 0.90644 +eval/Actions Max 0.999987 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.52963 +time/logging (s) 0.00366352 +time/sampling batch (s) 0.522599 +time/saving (s) 0.00332026 +time/training (s) 7.49961 +time/epoch (s) 10.5588 +time/total (s) 10190 +Epoch -9 +---------------------------------- --------------- +2022-05-10 16:00:53.172323 PDT | [2] Epoch -8 finished +---------------------------------- --------------- +epoch -8 +replay_buffer/size 999033 +trainer/num train calls 993000 +trainer/Policy Loss -19.2728 +trainer/Log Pis Mean 23.8267 +trainer/Log Pis Std 13.0093 +trainer/Log Pis Max 68.7325 +trainer/Log Pis Min -5.5886 +trainer/policy/mean Mean -0.0366673 +trainer/policy/mean Std 0.907009 +trainer/policy/mean Max 0.999981 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.83218 +trainer/policy/normal/std Std 0.698843 +trainer/policy/normal/std Max 5.90852 +trainer/policy/normal/std Min 0.187584 +trainer/policy/normal/log_std Mean 0.996458 +trainer/policy/normal/log_std Std 0.339311 +trainer/policy/normal/log_std Max 1.7764 +trainer/policy/normal/log_std Min -1.67353 +eval/num steps total 991392 +eval/num paths total 993 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0256407 +eval/Actions Std 0.907356 +eval/Actions Max 0.999988 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68241 +time/logging (s) 0.00409203 +time/sampling batch (s) 0.523139 +time/saving (s) 0.00384947 +time/training (s) 7.42156 +time/epoch (s) 10.635 +time/total (s) 10200.7 +Epoch -8 +---------------------------------- --------------- +2022-05-10 16:01:03.196669 PDT | [2] Epoch -7 finished +---------------------------------- ---------------- +epoch -7 +replay_buffer/size 999033 +trainer/num train calls 994000 +trainer/Policy Loss -19.7803 +trainer/Log Pis Mean 24.3591 +trainer/Log Pis Std 12.7791 +trainer/Log Pis Max 64.0645 +trainer/Log Pis Min -5.78687 +trainer/policy/mean Mean 0.000171553 +trainer/policy/mean Std 0.908905 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.85947 +trainer/policy/normal/std Std 0.684572 +trainer/policy/normal/std Max 7.37199 +trainer/policy/normal/std Min 0.251752 +trainer/policy/normal/log_std Mean 1.00907 +trainer/policy/normal/log_std Std 0.327821 +trainer/policy/normal/log_std Max 1.99769 +trainer/policy/normal/log_std Min -1.37931 +eval/num steps total 992392 +eval/num paths total 994 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0344663 +eval/Actions Std 0.919101 +eval/Actions Max 0.999999 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6118 +time/logging (s) 0.00374246 +time/sampling batch (s) 0.275199 +time/saving (s) 0.00338756 +time/training (s) 7.10468 +time/epoch (s) 9.99881 +time/total (s) 10210.7 +Epoch -7 +---------------------------------- ---------------- +2022-05-10 16:01:12.601093 PDT | [2] Epoch -6 finished +---------------------------------- --------------- +epoch -6 +replay_buffer/size 999033 +trainer/num train calls 995000 +trainer/Policy Loss -18.8147 +trainer/Log Pis Mean 23.2577 +trainer/Log Pis Std 13.0431 +trainer/Log Pis Max 70.7538 +trainer/Log Pis Min -5.35949 +trainer/policy/mean Mean -0.0451511 +trainer/policy/mean Std 0.904296 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.79561 +trainer/policy/normal/std Std 0.694244 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.260434 +trainer/policy/normal/log_std Mean 0.982147 +trainer/policy/normal/log_std Std 0.347599 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.34541 +eval/num steps total 993392 +eval/num paths total 995 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00330268 +eval/Actions Std 0.906497 +eval/Actions Max 0.999997 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.77505 +time/logging (s) 0.00376092 +time/sampling batch (s) 0.525054 +time/saving (s) 0.00338256 +time/training (s) 6.07238 +time/epoch (s) 9.37963 +time/total (s) 10220 +Epoch -6 +---------------------------------- --------------- +2022-05-10 16:01:23.170843 PDT | [2] Epoch -5 finished +---------------------------------- --------------- +epoch -5 +replay_buffer/size 999033 +trainer/num train calls 996000 +trainer/Policy Loss -19.443 +trainer/Log Pis Mean 24.726 +trainer/Log Pis Std 13.099 +trainer/Log Pis Max 78.514 +trainer/Log Pis Min -7.73461 +trainer/policy/mean Mean -0.0437124 +trainer/policy/mean Std 0.90999 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.75955 +trainer/policy/normal/std Std 0.676613 +trainer/policy/normal/std Max 6.40843 +trainer/policy/normal/std Min 0.245933 +trainer/policy/normal/log_std Mean 0.970593 +trainer/policy/normal/log_std Std 0.33985 +trainer/policy/normal/log_std Max 1.85761 +trainer/policy/normal/log_std Min -1.4027 +eval/num steps total 994392 +eval/num paths total 996 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.29813 +eval/Actions Std 0.859326 +eval/Actions Max 1 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.79014 +time/logging (s) 0.00380927 +time/sampling batch (s) 0.557743 +time/saving (s) 0.00354417 +time/training (s) 7.18907 +time/epoch (s) 10.5443 +time/total (s) 10230.6 +Epoch -5 +---------------------------------- --------------- +2022-05-10 16:01:34.120552 PDT | [2] Epoch -4 finished +---------------------------------- --------------- +epoch -4 +replay_buffer/size 999033 +trainer/num train calls 997000 +trainer/Policy Loss -20.6346 +trainer/Log Pis Mean 24.2954 +trainer/Log Pis Std 13.6533 +trainer/Log Pis Max 65.9181 +trainer/Log Pis Min -10.1074 +trainer/policy/mean Mean -0.0467416 +trainer/policy/mean Std 0.909076 +trainer/policy/mean Max 0.99998 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.82178 +trainer/policy/normal/std Std 0.681018 +trainer/policy/normal/std Max 5.99711 +trainer/policy/normal/std Min 0.293566 +trainer/policy/normal/log_std Mean 0.995198 +trainer/policy/normal/log_std Std 0.327871 +trainer/policy/normal/log_std Max 1.79128 +trainer/policy/normal/log_std Min -1.22565 +eval/num steps total 995392 +eval/num paths total 997 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0812603 +eval/Actions Std 0.915296 +eval/Actions Max 0.999993 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.78811 +time/logging (s) 0.003715 +time/sampling batch (s) 0.284127 +time/saving (s) 0.00380484 +time/training (s) 7.84514 +time/epoch (s) 10.9249 +time/total (s) 10241.5 +Epoch -4 +---------------------------------- --------------- +2022-05-10 16:01:43.856721 PDT | [2] Epoch -3 finished +---------------------------------- --------------- +epoch -3 +replay_buffer/size 999033 +trainer/num train calls 998000 +trainer/Policy Loss -19.3598 +trainer/Log Pis Mean 24.7561 +trainer/Log Pis Std 13.2492 +trainer/Log Pis Max 71.9031 +trainer/Log Pis Min -8.28238 +trainer/policy/mean Mean -0.0213569 +trainer/policy/mean Std 0.904437 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.7571 +trainer/policy/normal/std Std 0.682225 +trainer/policy/normal/std Max 6.86762 +trainer/policy/normal/std Min 0.270258 +trainer/policy/normal/log_std Mean 0.967461 +trainer/policy/normal/log_std Std 0.351383 +trainer/policy/normal/log_std Max 1.92682 +trainer/policy/normal/log_std Min -1.30838 +eval/num steps total 996392 +eval/num paths total 998 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0213719 +eval/Actions Std 0.913394 +eval/Actions Max 0.999995 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68875 +time/logging (s) 0.00368385 +time/sampling batch (s) 0.523693 +time/saving (s) 0.00331428 +time/training (s) 6.49108 +time/epoch (s) 9.71053 +time/total (s) 10251.2 +Epoch -3 +---------------------------------- --------------- +2022-05-10 16:01:55.321570 PDT | [2] Epoch -2 finished +---------------------------------- --------------- +epoch -2 +replay_buffer/size 999033 +trainer/num train calls 999000 +trainer/Policy Loss -18.6484 +trainer/Log Pis Mean 24.4185 +trainer/Log Pis Std 13.8371 +trainer/Log Pis Max 73.7874 +trainer/Log Pis Min -7.12206 +trainer/policy/mean Mean -0.0333359 +trainer/policy/mean Std 0.903919 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.81175 +trainer/policy/normal/std Std 0.688922 +trainer/policy/normal/std Max 5.75497 +trainer/policy/normal/std Min 0.238993 +trainer/policy/normal/log_std Mean 0.9908 +trainer/policy/normal/log_std Std 0.330695 +trainer/policy/normal/log_std Max 1.75006 +trainer/policy/normal/log_std Min -1.43132 +eval/num steps total 997392 +eval/num paths total 999 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0616786 +eval/Actions Std 0.906293 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.28906 +time/logging (s) 0.0037716 +time/sampling batch (s) 0.527704 +time/saving (s) 0.00340826 +time/training (s) 8.61647 +time/epoch (s) 11.4404 +time/total (s) 10262.7 +Epoch -2 +---------------------------------- --------------- +2022-05-10 16:02:05.564838 PDT | [2] Epoch -1 finished +---------------------------------- --------------- +epoch -1 +replay_buffer/size 999033 +trainer/num train calls 1e+06 +trainer/Policy Loss -19.6051 +trainer/Log Pis Mean 24.4883 +trainer/Log Pis Std 13.7913 +trainer/Log Pis Max 69.8601 +trainer/Log Pis Min -8.21345 +trainer/policy/mean Mean -0.0494025 +trainer/policy/mean Std 0.902791 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.75534 +trainer/policy/normal/std Std 0.693681 +trainer/policy/normal/std Max 6.25657 +trainer/policy/normal/std Min 0.247179 +trainer/policy/normal/log_std Mean 0.966813 +trainer/policy/normal/log_std Std 0.348917 +trainer/policy/normal/log_std Max 1.83363 +trainer/policy/normal/log_std Min -1.39764 +eval/num steps total 998392 +eval/num paths total 1000 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0901913 +eval/Actions Std 0.920978 +eval/Actions Max 0.999986 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.41169 +time/logging (s) 0.00395282 +time/sampling batch (s) 0.528575 +time/saving (s) 0.00674875 +time/training (s) 7.26744 +time/epoch (s) 10.2184 +time/total (s) 10272.9 +Epoch -1 +---------------------------------- ---------------